: I'm currently testing a large index with more than 10 million
: documents and 24 fields, using the example installation with
: Jetty.
: When deleting or updateing documents from the index or doing
: search queries I get "Java heap space" error messages like
: this (in this case while trying to delete documents):
        ...
: In both cases obviously the server ran out of heap space.
: I'm wondering what I can do to prevent this. I started the
: server using the java options "-server -Xmx1024m".
:
: Does anybody else have problems with heap space using a large
: index? Is there anything I can do against this?

How big is your physical index directory on disk?

Generally speaking, there isn't much you can do to improve upon the memory
needs of a Lucene index that Solr isn't allready doing: Reuse a single
IndexSearcher as much as possible.

Your best bet is to allocate as much ram to the server as you can.
Depending on how full your caches are, and what hitratios you are getting
(the "STATISTICS" link from the Admin screen will tell you) you might want
to make some of them smaller to reduce the amount of RAM Solr uses for
them.

>From an acctual index standpoint, if you don't care about doc/field boosts
of lengthNorms, then the omitNorm="true" option on your fields (or
fieldtypes) will help save one byte per document per field you use it on.


-Hoss

Reply via email to