Hi Shawn,

Thanks for the helpful and thorough response.  While I understand all of the
factors that you've outlined for memory requirements (in fact, I'd
previously read your page on Solr performance problems), it is baffling to
me why two identical SolrCloud instances, each sharded across 3 machines
with identical hardware, would run into these memory issues at such
differently memory limits (one SolrCloud instance started seeing OOM issues
at 2 million indexed documents, the other started seeing OOM issues between
20 and 30 million indexed documents). 

When I stated that approximately 1.5GB, I mean that this is how much heap
space I allocated when launching java with -Xmx, and I can see the java
process using that full amount of RAM.  

>From a usage perspective, the usage doesn't seem all that heavy.  I'm
indexing about 600k documents an hour (each of which have ~20 short numeric
or string fields).  I have the autoSoftCommit parameter set for once a
second, and the autoCommit time set for every 5 minutes, with openSearcher
set to false.  Finally, I have maxWarmingSearchers at 2.  Besides indexing
those documents, I've been doing a few small queries just to check how many
documents have been indexed, and a few other small queries, sorting by a
single attribute.  These searches are very infrequent though, maybe 5 or 6
an hour.

Seems like a strange issue indeed.  My expectation is that Solr would hit a
point where it becomes horribly slow after some threshold where things don't
fit in the cache, but I'd never expect it to simply crash like it's doing.



--
View this message in context: 
http://lucene.472066.n3.nabble.com/SolrException-Error-when-indexing-new-documents-at-scale-in-SolrCloud-tp4111551p4111680.html
Sent from the Solr - User mailing list archive at Nabble.com.

Reply via email to