TestIndexingPerformance?
What the heck... that's not even multi-threaded!

-Yonik
http://www.lucidimagination.com



On Tue, Oct 6, 2009 at 12:17 PM, Mark Miller <markrmil...@gmail.com> wrote:
> Darnit - didn't finish that email. This is after running your old short
> doc perf test for 10,000 iterations. You see the same thing with 1000
> iterations but much less pronounced eg gettin' worse with more iterations.
>
> Mark Miller wrote:
>> A little before and after. The before is around may 5th'is - the after
>> is trunk.
>>
>> http://myhardshadow.com/memanalysis/before.png
>> http://myhardshadow.com/memanalysis/after.png
>>
>> Mark Miller wrote:
>>
>>> Took a peak at the checkout around the time he says he's using.
>>>
>>> CharTokenizer appears to be holding onto much large char[] arrays now
>>> than before. Same with snowball.Among - used to be almost nothing, now
>>> its largio.
>>>
>>> The new TokenStream stuff appears to be clinging. Needs to find some
>>> inner peace.
>>>
>>> Yonik Seeley wrote:
>>>
>>>
>>>> On Mon, Oct 5, 2009 at 4:54 PM, Jeff Newburn <jnewb...@zappos.com> wrote:
>>>>
>>>>
>>>>
>>>>> Ok we have done some more testing on this issue.  When I only have the 1
>>>>> core the reindex completes fine.  However, when I added a second core with
>>>>> no documents it runs out of heap again.  This time the heap was 322Mb of
>>>>> LRUCache.  The 1 query that warms returns exactly 2 documents so I have no
>>>>> idea where the LRUCache is getting its information or what is even in 
>>>>> there.
>>>>>
>>>>>
>>>>>
>>>> I guess the obvious thing to check would be the custom search component.
>>>> Does it access documents?  I don't see how else the document cache
>>>> could self populate with so many entries (assuming it is the document
>>>> cache again).
>>>>
>>>> -Yonik
>>>> http://www.lucidimagination.com
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>> --
>>>>> Jeff Newburn
>>>>> Software Engineer, Zappos.com
>>>>> jnewb...@zappos.com - 702-943-7562
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>> From: Yonik Seeley <yo...@lucidimagination.com>
>>>>>> Reply-To: <solr-user@lucene.apache.org>
>>>>>> Date: Mon, 5 Oct 2009 13:32:32 -0400
>>>>>> To: <solr-user@lucene.apache.org>
>>>>>> Subject: Re: Solr Trunk Heap Space Issues
>>>>>>
>>>>>> On Mon, Oct 5, 2009 at 1:00 PM, Jeff Newburn <jnewb...@zappos.com> wrote:
>>>>>>
>>>>>>
>>>>>>
>>>>>>> Ok I have eliminated all queries for warming and am still getting the 
>>>>>>> heap
>>>>>>> space dump.  Any ideas at this point what could be wrong?  This seems 
>>>>>>> like a
>>>>>>> huge increase in memory to go from indexing without issues to not being 
>>>>>>> able
>>>>>>> to even with warming off.
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>> Do you have any custom Analyzers, Tokenizers, TokenFilters?
>>>>>> Another change is that token streams are reused by caching in a
>>>>>> thread-local, so every thread in your server could potentially have a
>>>>>> copy of an analysis chain (token stream) per field that you have used.
>>>>>>  This normally shouldn't be an issue since these will be small.  Also,
>>>>>> how many unique fields do you have?
>>>>>>
>>>>>> -Yonik
>>>>>> http://www.lucidimagination.com
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>>> Jeff Newburn
>>>>>>> Software Engineer, Zappos.com
>>>>>>> jnewb...@zappos.com - 702-943-7562
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>> From: Jeff Newburn <jnewb...@zappos.com>
>>>>>>>> Reply-To: <solr-user@lucene.apache.org>
>>>>>>>> Date: Thu, 01 Oct 2009 08:41:18 -0700
>>>>>>>> To: "solr-user@lucene.apache.org" <solr-user@lucene.apache.org>
>>>>>>>> Subject: Solr Trunk Heap Space Issues
>>>>>>>>
>>>>>>>> I am trying to update to the newest version of solr from trunk as of 
>>>>>>>> May
>>>>>>>> 5th.  I updated and compiled from trunk as of yesterday (09/30/2009).  
>>>>>>>> When
>>>>>>>> I try to do a full import I am receiving a GC heap error after changing
>>>>>>>> nothing in the configuration files.  Why would this happen in the most
>>>>>>>> recent versions but not in the version from a few months ago.  The 
>>>>>>>> stack
>>>>>>>> trace is below.
>>>>>>>>
>>>>>>>> Oct 1, 2009 8:34:32 AM 
>>>>>>>> org.apache.solr.update.processor.LogUpdateProcessor
>>>>>>>> finish
>>>>>>>> INFO: {add=[166400, 166608, 166698, 166800, 166811, 167097, 167316, 
>>>>>>>> 167353,
>>>>>>>> ...(83 more)]} 0 35991
>>>>>>>> Oct 1, 2009 8:34:32 AM org.apache.solr.common.SolrException log
>>>>>>>> SEVERE: java.lang.OutOfMemoryError: GC overhead limit exceeded
>>>>>>>>     at java.util.Arrays.copyOfRange(Arrays.java:3209)
>>>>>>>>     at java.lang.String.<init>(String.java:215)
>>>>>>>>     at 
>>>>>>>> com.ctc.wstx.util.TextBuffer.contentsAsString(TextBuffer.java:384)
>>>>>>>>     at 
>>>>>>>> com.ctc.wstx.sr.BasicStreamReader.getText(BasicStreamReader.java:821)
>>>>>>>>     at org.apache.solr.handler.XMLLoader.readDoc(XMLLoader.java:280)
>>>>>>>>     at 
>>>>>>>> org.apache.solr.handler.XMLLoader.processUpdate(XMLLoader.java:139)
>>>>>>>>     at org.apache.solr.handler.XMLLoader.load(XMLLoader.java:69)
>>>>>>>>     at
>>>>>>>> org.apache.solr.handler.ContentStreamHandlerBase.handleRequestBody(ContentSt
>>>>>>>> reamHandlerBase.java:54)
>>>>>>>>     at
>>>>>>>> org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.
>>>>>>>> java:131)
>>>>>>>>     at org.apache.solr.core.SolrCore.execute(SolrCore.java:1316)
>>>>>>>>     at
>>>>>>>> org.apache.solr.servlet.SolrDispatchFilter.execute(SolrDispatchFilter.java:3
>>>>>>>> 38)
>>>>>>>>     at
>>>>>>>> org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:
>>>>>>>> 241)
>>>>>>>>     at
>>>>>>>> org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(Application
>>>>>>>> FilterChain.java:235)
>>>>>>>>     at
>>>>>>>> org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterCh
>>>>>>>> ain.java:206)
>>>>>>>>     at
>>>>>>>> org.apache.catalina.core.StandardWrapperValve.invoke(StandardWrapperValve.ja
>>>>>>>> va:233)
>>>>>>>>     at
>>>>>>>> org.apache.catalina.core.StandardContextValve.invoke(StandardContextValve.ja
>>>>>>>> va:175)
>>>>>>>>     at
>>>>>>>> org.apache.catalina.core.StandardHostValve.invoke(StandardHostValve.java:128
>>>>>>>> )
>>>>>>>>     at
>>>>>>>> org.apache.catalina.valves.ErrorReportValve.invoke(ErrorReportValve.java:102
>>>>>>>> )
>>>>>>>>     at
>>>>>>>> org.apache.catalina.core.StandardEngineValve.invoke(StandardEngineValve.java
>>>>>>>> :109)
>>>>>>>>     at
>>>>>>>> org.apache.catalina.connector.CoyoteAdapter.service(CoyoteAdapter.java:286)
>>>>>>>>     at
>>>>>>>> org.apache.coyote.http11.Http11NioProcessor.process(Http11NioProcessor.java:
>>>>>>>> 879)
>>>>>>>>     at
>>>>>>>> org.apache.coyote.http11.Http11NioProtocol$Http11ConnectionHandler.process(H
>>>>>>>> ttp11NioProtocol.java:719)
>>>>>>>>     at
>>>>>>>> org.apache.tomcat.util.net.NioEndpoint$SocketProcessor.run(NioEndpoint.java:
>>>>>>>> 2080)
>>>>>>>>     at
>>>>>>>> java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.ja
>>>>>>>> va:886)
>>>>>>>>     at
>>>>>>>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:9
>>>>>>>> 08)
>>>>>>>>     at java.lang.Thread.run(Thread.java:619)
>>>>>>>>
>>>>>>>> Oct 1, 2009 8:40:06 AM org.apache.solr.core.SolrCore execute
>>>>>>>> INFO: [zeta-main] webapp=/solr path=/update params={} status=500 
>>>>>>>> QTime=5265
>>>>>>>> Oct 1, 2009 8:40:12 AM org.apache.solr.common.SolrException log
>>>>>>>> SEVERE: java.lang.OutOfMemoryError: GC overhead limit exceeded
>>>>>>>>
>>>>>>>> --
>>>>>>>> Jeff Newburn
>>>>>>>> Software Engineer, Zappos.com
>>>>>>>> jnewb...@zappos.com - 702-943-7562
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>
>>>>>
>>>>>
>>>
>>>
>>
>>
>>
>
>
> --
> - Mark
>
> http://www.lucidimagination.com
>
>
>
>

Reply via email to