Inline below. On Feb 2, 2010, at 8:40 PM, Vauthrin, Laurent wrote:
> Hello, > > > > I'm trying to troubleshoot a problem that occurred on a few Solr slave > Tomcat instances and wanted to run it by the list to see if I'm on the > right track. > > > > The setup involves 1 master replicating to three slaves (I don't know > what the replication interval is at this time). These instances have > been running fine for a while (from what I understand) but ran into > problems just today during peak site usage. > > > > The following two exceptions were observed (partially stripped stack > traces): > > > > WARNING: [] Error opening new searcher. exceeded limit of > maxWarmingSearchers=2, try again later. > > Feb 1, 2010 10:00:31 AM org.apache.solr.common.SolrException log > > SEVERE: org.apache.solr.common.SolrException: Error opening new > searcher. exceeded limit of maxWarmingSearchers=2, try again later. > > at org.apache.solr.core.SolrCore.getSearcher(SolrCore.java:941) > > at > org.apache.solr.update.DirectUpdateHandler2.commit(DirectUpdateHandler2. > java:368) > > at > org.apache.solr.update.processor.RunUpdateProcessor.processCommit(RunUpd > ateProcessorFactory.java:77) > > > > Feb 1, 2010 10:29:36 AM org.apache.solr.common.SolrException log > > SEVERE: java.lang.OutOfMemoryError: GC overhead limit exceeded > > at > org.apache.lucene.index.SegmentReader.termDocs(SegmentReader.java:734) > > at > org.apache.lucene.index.MultiSegmentReader$MultiTermDocs.termDocs(MultiS > egmentReader.java:612) > > at > org.apache.lucene.index.MultiSegmentReader$MultiTermDocs.termDocs(MultiS > egmentReader.java:605) > > at > org.apache.lucene.index.MultiSegmentReader$MultiTermDocs.read(MultiSegme > ntReader.java:570) > > at org.apache.lucene.search.TermScorer.next(TermScorer.java:106) > > at > org.apache.lucene.search.DisjunctionSumScorer.initScorerDocQueue(Disjunc > tionSumScorer.java:105) > > at > org.apache.lucene.search.DisjunctionSumScorer.next(DisjunctionSumScorer. > java:144) > > at > org.apache.lucene.search.BooleanScorer2.next(BooleanScorer2.java:352) > > at > org.apache.lucene.search.DisjunctionSumScorer.initScorerDocQueue(Disjunc > tionSumScorer.java:105) > > at > org.apache.lucene.search.DisjunctionSumScorer.next(DisjunctionSumScorer. > java:144) > > at > org.apache.lucene.search.BooleanScorer2.next(BooleanScorer2.java:352) > > at > org.apache.lucene.search.ConjunctionScorer.init(ConjunctionScorer.java:8 > 0) > > at > org.apache.lucene.search.ConjunctionScorer.next(ConjunctionScorer.java:4 > 8) > > at > org.apache.lucene.search.BooleanScorer2.score(BooleanScorer2.java:319) > > at > org.apache.lucene.search.IndexSearcher.search(IndexSearcher.java:137) > > at org.apache.lucene.search.Searcher.search(Searcher.java:126) > > at org.apache.lucene.search.Searcher.search(Searcher.java:105) > > at > org.apache.solr.search.SolrIndexSearcher.getDocListNC(SolrIndexSearcher. > java:920) > > at > org.apache.solr.search.SolrIndexSearcher.getDocListC(SolrIndexSearcher.j > ava:838) > > at > org.apache.solr.search.SolrIndexSearcher.search(SolrIndexSearcher.java:2 > 69) > > > > Here's the config for the caches: > > > > filterCache: size="15000" initialSize="5000" autowarmCount="5000" > > queryResultCache: size="15000" initialSize="5000" autowarmCount="15000" > > documentCache: size="15000" initialSize="5000" > > > > From what I understand, the first exception indicates that multiple > replications are being processed at the same time. Is that correct or > could it be something else? You are probably committing/replicating faster than Solr can open up the new index and warm the new searcher. > > Does the second exception indicate that Solr is having problems handling > the query load (possibly due to a commit happening at the same time)? This is likely caused by the first problem b/c you are running out of memory > > > > Does anyone have any insight that might help here? I sort of suspect > that the autowarm counts are too large but I may be off there. I can > provide more details (as I get them) about this if needed. You probably should start smaller, yes. Bigger is not always better when it comes to caches, especially when GC is factored in. -------------------------- Grant Ingersoll http://www.lucidimagination.com/ Search the Lucene ecosystem using Solr/Lucene: http://www.lucidimagination.com/search