On Mon, Oct 5, 2009 at 4:54 PM, Jeff Newburn <jnewb...@zappos.com> wrote:
> Ok we have done some more testing on this issue.  When I only have the 1
> core the reindex completes fine.  However, when I added a second core with
> no documents it runs out of heap again.  This time the heap was 322Mb of
> LRUCache.  The 1 query that warms returns exactly 2 documents so I have no
> idea where the LRUCache is getting its information or what is even in there.

I guess the obvious thing to check would be the custom search component.
Does it access documents?  I don't see how else the document cache
could self populate with so many entries (assuming it is the document
cache again).

-Yonik
http://www.lucidimagination.com




>
> --
> Jeff Newburn
> Software Engineer, Zappos.com
> jnewb...@zappos.com - 702-943-7562
>
>
>> From: Yonik Seeley <yo...@lucidimagination.com>
>> Reply-To: <solr-user@lucene.apache.org>
>> Date: Mon, 5 Oct 2009 13:32:32 -0400
>> To: <solr-user@lucene.apache.org>
>> Subject: Re: Solr Trunk Heap Space Issues
>>
>> On Mon, Oct 5, 2009 at 1:00 PM, Jeff Newburn <jnewb...@zappos.com> wrote:
>>> Ok I have eliminated all queries for warming and am still getting the heap
>>> space dump.  Any ideas at this point what could be wrong?  This seems like a
>>> huge increase in memory to go from indexing without issues to not being able
>>> to even with warming off.
>>
>> Do you have any custom Analyzers, Tokenizers, TokenFilters?
>> Another change is that token streams are reused by caching in a
>> thread-local, so every thread in your server could potentially have a
>> copy of an analysis chain (token stream) per field that you have used.
>>  This normally shouldn't be an issue since these will be small.  Also,
>> how many unique fields do you have?
>>
>> -Yonik
>> http://www.lucidimagination.com
>>
>>
>>
>>> Jeff Newburn
>>> Software Engineer, Zappos.com
>>> jnewb...@zappos.com - 702-943-7562
>>>
>>>
>>>> From: Jeff Newburn <jnewb...@zappos.com>
>>>> Reply-To: <solr-user@lucene.apache.org>
>>>> Date: Thu, 01 Oct 2009 08:41:18 -0700
>>>> To: "solr-user@lucene.apache.org" <solr-user@lucene.apache.org>
>>>> Subject: Solr Trunk Heap Space Issues
>>>>
>>>> I am trying to update to the newest version of solr from trunk as of May
>>>> 5th.  I updated and compiled from trunk as of yesterday (09/30/2009).  When
>>>> I try to do a full import I am receiving a GC heap error after changing
>>>> nothing in the configuration files.  Why would this happen in the most
>>>> recent versions but not in the version from a few months ago.  The stack
>>>> trace is below.
>>>>
>>>> Oct 1, 2009 8:34:32 AM org.apache.solr.update.processor.LogUpdateProcessor
>>>> finish
>>>> INFO: {add=[166400, 166608, 166698, 166800, 166811, 167097, 167316, 167353,
>>>> ...(83 more)]} 0 35991
>>>> Oct 1, 2009 8:34:32 AM org.apache.solr.common.SolrException log
>>>> SEVERE: java.lang.OutOfMemoryError: GC overhead limit exceeded
>>>>     at java.util.Arrays.copyOfRange(Arrays.java:3209)
>>>>     at java.lang.String.<init>(String.java:215)
>>>>     at com.ctc.wstx.util.TextBuffer.contentsAsString(TextBuffer.java:384)
>>>>     at 
>>>> com.ctc.wstx.sr.BasicStreamReader.getText(BasicStreamReader.java:821)
>>>>     at org.apache.solr.handler.XMLLoader.readDoc(XMLLoader.java:280)
>>>>     at org.apache.solr.handler.XMLLoader.processUpdate(XMLLoader.java:139)
>>>>     at org.apache.solr.handler.XMLLoader.load(XMLLoader.java:69)
>>>>     at
>>>> org.apache.solr.handler.ContentStreamHandlerBase.handleRequestBody(ContentSt
>>>> reamHandlerBase.java:54)
>>>>     at
>>>> org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.
>>>> java:131)
>>>>     at org.apache.solr.core.SolrCore.execute(SolrCore.java:1316)
>>>>     at
>>>> org.apache.solr.servlet.SolrDispatchFilter.execute(SolrDispatchFilter.java:3
>>>> 38)
>>>>     at
>>>> org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:
>>>> 241)
>>>>     at
>>>> org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(Application
>>>> FilterChain.java:235)
>>>>     at
>>>> org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterCh
>>>> ain.java:206)
>>>>     at
>>>> org.apache.catalina.core.StandardWrapperValve.invoke(StandardWrapperValve.ja
>>>> va:233)
>>>>     at
>>>> org.apache.catalina.core.StandardContextValve.invoke(StandardContextValve.ja
>>>> va:175)
>>>>     at
>>>> org.apache.catalina.core.StandardHostValve.invoke(StandardHostValve.java:128
>>>> )
>>>>     at
>>>> org.apache.catalina.valves.ErrorReportValve.invoke(ErrorReportValve.java:102
>>>> )
>>>>     at
>>>> org.apache.catalina.core.StandardEngineValve.invoke(StandardEngineValve.java
>>>> :109)
>>>>     at
>>>> org.apache.catalina.connector.CoyoteAdapter.service(CoyoteAdapter.java:286)
>>>>     at
>>>> org.apache.coyote.http11.Http11NioProcessor.process(Http11NioProcessor.java:
>>>> 879)
>>>>     at
>>>> org.apache.coyote.http11.Http11NioProtocol$Http11ConnectionHandler.process(H
>>>> ttp11NioProtocol.java:719)
>>>>     at
>>>> org.apache.tomcat.util.net.NioEndpoint$SocketProcessor.run(NioEndpoint.java:
>>>> 2080)
>>>>     at
>>>> java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.ja
>>>> va:886)
>>>>     at
>>>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:9
>>>> 08)
>>>>     at java.lang.Thread.run(Thread.java:619)
>>>>
>>>> Oct 1, 2009 8:40:06 AM org.apache.solr.core.SolrCore execute
>>>> INFO: [zeta-main] webapp=/solr path=/update params={} status=500 QTime=5265
>>>> Oct 1, 2009 8:40:12 AM org.apache.solr.common.SolrException log
>>>> SEVERE: java.lang.OutOfMemoryError: GC overhead limit exceeded
>>>>
>>>> --
>>>> Jeff Newburn
>>>> Software Engineer, Zappos.com
>>>> jnewb...@zappos.com - 702-943-7562
>>>>
>>>
>>>
>
>

Reply via email to