In a word, "no". I once doubled the JVM requirements
by changing just the query. You have to prototype. Here's
a blog on the subject:

https://lucidworks.com/blog/2012/07/23/sizing-hardware-in-the-abstract-why-we-dont-have-a-definitive-answer/



On Wed, Oct 28, 2015 at 11:06 AM, Salman Ansari <salman.rah...@gmail.com> wrote:
> I have already indexed all the documents in Solr and not indexing anymore.
> So the problem I am running in is after all the documents are indexed. I am
> using Solr cloud with two shards and two replicas for each shard but on the
> same machine. Is there anywhere I can look at the relation between index
> size and machine specs and its effect on Solr query performance?
>
> Regards,
> Salman
>
> On Mon, Oct 26, 2015 at 5:55 PM, Upayavira <u...@odoko.co.uk> wrote:
>
>>
>>
>> On Sun, Oct 25, 2015, at 05:43 PM, Salman Ansari wrote:
>> > Thanks guys for your responses.
>> >
>> > That's a very very large cache size.  It is likely to use a VERY large
>> > amount of heap, and autowarming up to 4096 entries at commit time might
>> > take many *minutes*.  Each filterCache entry is maxDoc/8 bytes.  On an
>> > index core with 70 million documents, each filterCache entry is at least
>> > 8.75 million bytes.  Multiply that by 16384, and a completely full cache
>> > would need about 140GB of heap memory.  4096 entries will require 35GB.
>> >  I don't think this cache is actually storing that many entries, or you
>> > would most certainly be running into OutOfMemoryError exceptions.
>> >
>> > True, however, I have tried with the default filtercache at the beginning
>> > but the problem was still there. So, I don't think that is how I should
>> > increase the performance of my Solr. Moreover, as you mentioned, when I
>> > change the configuration, I should be running out of memory but that did
>> > not happen. Do you think my Solr has not taken the latest configs? I have
>> > restarted the Solr btw.
>> >
>> > Lately I have been trying different ways to improve this and I have
>> > created
>> > a brand new index on the same machine using 2 shards and it had few
>> > entries
>> > (about 5) and the performance was booming, I got the results back in 42
>> > ms
>> > sometimes. What concerns me is that may be I am loading too much into one
>> > index so that is why this is killing the performance. Is there a
>> > recommended index size/document number and size that I should be looking
>> > at
>> > to tune this? Any other ideas other than increasing the memory size as I
>> > have already tried this?
>>
>> The optimal index size is down to the size of segments on disk. New
>> segments are created when hard commits occur, and existing on-disk
>> segments may get merged in the background when the segment count gets
>> too high. Now, if those on-disk segments get too large, copying them
>> around at merge time can get prohibitive, especially if your index is
>> changing frequently.
>>
>> Splitting such an index into shards is one approach to dealing with this
>> issue.
>>
>> Upayavira
>>

Reply via email to