Yeah - I was wondering about that ... not sure how these guys are
stacking up ...

Yonik Seeley wrote:
> TestIndexingPerformance?
> What the heck... that's not even multi-threaded!
>
> -Yonik
> http://www.lucidimagination.com
>
>
>
> On Tue, Oct 6, 2009 at 12:17 PM, Mark Miller <markrmil...@gmail.com> wrote:
>   
>> Darnit - didn't finish that email. This is after running your old short
>> doc perf test for 10,000 iterations. You see the same thing with 1000
>> iterations but much less pronounced eg gettin' worse with more iterations.
>>
>> Mark Miller wrote:
>>     
>>> A little before and after. The before is around may 5th'is - the after
>>> is trunk.
>>>
>>> http://myhardshadow.com/memanalysis/before.png
>>> http://myhardshadow.com/memanalysis/after.png
>>>
>>> Mark Miller wrote:
>>>
>>>       
>>>> Took a peak at the checkout around the time he says he's using.
>>>>
>>>> CharTokenizer appears to be holding onto much large char[] arrays now
>>>> than before. Same with snowball.Among - used to be almost nothing, now
>>>> its largio.
>>>>
>>>> The new TokenStream stuff appears to be clinging. Needs to find some
>>>> inner peace.
>>>>
>>>> Yonik Seeley wrote:
>>>>
>>>>
>>>>         
>>>>> On Mon, Oct 5, 2009 at 4:54 PM, Jeff Newburn <jnewb...@zappos.com> wrote:
>>>>>
>>>>>
>>>>>
>>>>>           
>>>>>> Ok we have done some more testing on this issue.  When I only have the 1
>>>>>> core the reindex completes fine.  However, when I added a second core 
>>>>>> with
>>>>>> no documents it runs out of heap again.  This time the heap was 322Mb of
>>>>>> LRUCache.  The 1 query that warms returns exactly 2 documents so I have 
>>>>>> no
>>>>>> idea where the LRUCache is getting its information or what is even in 
>>>>>> there.
>>>>>>
>>>>>>
>>>>>>
>>>>>>             
>>>>> I guess the obvious thing to check would be the custom search component.
>>>>> Does it access documents?  I don't see how else the document cache
>>>>> could self populate with so many entries (assuming it is the document
>>>>> cache again).
>>>>>
>>>>> -Yonik
>>>>> http://www.lucidimagination.com
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>           
>>>>>> --
>>>>>> Jeff Newburn
>>>>>> Software Engineer, Zappos.com
>>>>>> jnewb...@zappos.com - 702-943-7562
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>>             
>>>>>>> From: Yonik Seeley <yo...@lucidimagination.com>
>>>>>>> Reply-To: <solr-user@lucene.apache.org>
>>>>>>> Date: Mon, 5 Oct 2009 13:32:32 -0400
>>>>>>> To: <solr-user@lucene.apache.org>
>>>>>>> Subject: Re: Solr Trunk Heap Space Issues
>>>>>>>
>>>>>>> On Mon, Oct 5, 2009 at 1:00 PM, Jeff Newburn <jnewb...@zappos.com> 
>>>>>>> wrote:
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>               
>>>>>>>> Ok I have eliminated all queries for warming and am still getting the 
>>>>>>>> heap
>>>>>>>> space dump.  Any ideas at this point what could be wrong?  This seems 
>>>>>>>> like a
>>>>>>>> huge increase in memory to go from indexing without issues to not 
>>>>>>>> being able
>>>>>>>> to even with warming off.
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>>                 
>>>>>>> Do you have any custom Analyzers, Tokenizers, TokenFilters?
>>>>>>> Another change is that token streams are reused by caching in a
>>>>>>> thread-local, so every thread in your server could potentially have a
>>>>>>> copy of an analysis chain (token stream) per field that you have used.
>>>>>>>  This normally shouldn't be an issue since these will be small.  Also,
>>>>>>> how many unique fields do you have?
>>>>>>>
>>>>>>> -Yonik
>>>>>>> http://www.lucidimagination.com
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>               
>>>>>>>> Jeff Newburn
>>>>>>>> Software Engineer, Zappos.com
>>>>>>>> jnewb...@zappos.com - 702-943-7562
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>>                 
>>>>>>>>> From: Jeff Newburn <jnewb...@zappos.com>
>>>>>>>>> Reply-To: <solr-user@lucene.apache.org>
>>>>>>>>> Date: Thu, 01 Oct 2009 08:41:18 -0700
>>>>>>>>> To: "solr-user@lucene.apache.org" <solr-user@lucene.apache.org>
>>>>>>>>> Subject: Solr Trunk Heap Space Issues
>>>>>>>>>
>>>>>>>>> I am trying to update to the newest version of solr from trunk as of 
>>>>>>>>> May
>>>>>>>>> 5th.  I updated and compiled from trunk as of yesterday (09/30/2009). 
>>>>>>>>>  When
>>>>>>>>> I try to do a full import I am receiving a GC heap error after 
>>>>>>>>> changing
>>>>>>>>> nothing in the configuration files.  Why would this happen in the most
>>>>>>>>> recent versions but not in the version from a few months ago.  The 
>>>>>>>>> stack
>>>>>>>>> trace is below.
>>>>>>>>>
>>>>>>>>> Oct 1, 2009 8:34:32 AM 
>>>>>>>>> org.apache.solr.update.processor.LogUpdateProcessor
>>>>>>>>> finish
>>>>>>>>> INFO: {add=[166400, 166608, 166698, 166800, 166811, 167097, 167316, 
>>>>>>>>> 167353,
>>>>>>>>> ...(83 more)]} 0 35991
>>>>>>>>> Oct 1, 2009 8:34:32 AM org.apache.solr.common.SolrException log
>>>>>>>>> SEVERE: java.lang.OutOfMemoryError: GC overhead limit exceeded
>>>>>>>>>     at java.util.Arrays.copyOfRange(Arrays.java:3209)
>>>>>>>>>     at java.lang.String.<init>(String.java:215)
>>>>>>>>>     at 
>>>>>>>>> com.ctc.wstx.util.TextBuffer.contentsAsString(TextBuffer.java:384)
>>>>>>>>>     at 
>>>>>>>>> com.ctc.wstx.sr.BasicStreamReader.getText(BasicStreamReader.java:821)
>>>>>>>>>     at org.apache.solr.handler.XMLLoader.readDoc(XMLLoader.java:280)
>>>>>>>>>     at 
>>>>>>>>> org.apache.solr.handler.XMLLoader.processUpdate(XMLLoader.java:139)
>>>>>>>>>     at org.apache.solr.handler.XMLLoader.load(XMLLoader.java:69)
>>>>>>>>>     at
>>>>>>>>> org.apache.solr.handler.ContentStreamHandlerBase.handleRequestBody(ContentSt
>>>>>>>>> reamHandlerBase.java:54)
>>>>>>>>>     at
>>>>>>>>> org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.
>>>>>>>>> java:131)
>>>>>>>>>     at org.apache.solr.core.SolrCore.execute(SolrCore.java:1316)
>>>>>>>>>     at
>>>>>>>>> org.apache.solr.servlet.SolrDispatchFilter.execute(SolrDispatchFilter.java:3
>>>>>>>>> 38)
>>>>>>>>>     at
>>>>>>>>> org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:
>>>>>>>>> 241)
>>>>>>>>>     at
>>>>>>>>> org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(Application
>>>>>>>>> FilterChain.java:235)
>>>>>>>>>     at
>>>>>>>>> org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterCh
>>>>>>>>> ain.java:206)
>>>>>>>>>     at
>>>>>>>>> org.apache.catalina.core.StandardWrapperValve.invoke(StandardWrapperValve.ja
>>>>>>>>> va:233)
>>>>>>>>>     at
>>>>>>>>> org.apache.catalina.core.StandardContextValve.invoke(StandardContextValve.ja
>>>>>>>>> va:175)
>>>>>>>>>     at
>>>>>>>>> org.apache.catalina.core.StandardHostValve.invoke(StandardHostValve.java:128
>>>>>>>>> )
>>>>>>>>>     at
>>>>>>>>> org.apache.catalina.valves.ErrorReportValve.invoke(ErrorReportValve.java:102
>>>>>>>>> )
>>>>>>>>>     at
>>>>>>>>> org.apache.catalina.core.StandardEngineValve.invoke(StandardEngineValve.java
>>>>>>>>> :109)
>>>>>>>>>     at
>>>>>>>>> org.apache.catalina.connector.CoyoteAdapter.service(CoyoteAdapter.java:286)
>>>>>>>>>     at
>>>>>>>>> org.apache.coyote.http11.Http11NioProcessor.process(Http11NioProcessor.java:
>>>>>>>>> 879)
>>>>>>>>>     at
>>>>>>>>> org.apache.coyote.http11.Http11NioProtocol$Http11ConnectionHandler.process(H
>>>>>>>>> ttp11NioProtocol.java:719)
>>>>>>>>>     at
>>>>>>>>> org.apache.tomcat.util.net.NioEndpoint$SocketProcessor.run(NioEndpoint.java:
>>>>>>>>> 2080)
>>>>>>>>>     at
>>>>>>>>> java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.ja
>>>>>>>>> va:886)
>>>>>>>>>     at
>>>>>>>>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:9
>>>>>>>>> 08)
>>>>>>>>>     at java.lang.Thread.run(Thread.java:619)
>>>>>>>>>
>>>>>>>>> Oct 1, 2009 8:40:06 AM org.apache.solr.core.SolrCore execute
>>>>>>>>> INFO: [zeta-main] webapp=/solr path=/update params={} status=500 
>>>>>>>>> QTime=5265
>>>>>>>>> Oct 1, 2009 8:40:12 AM org.apache.solr.common.SolrException log
>>>>>>>>> SEVERE: java.lang.OutOfMemoryError: GC overhead limit exceeded
>>>>>>>>>
>>>>>>>>> --
>>>>>>>>> Jeff Newburn
>>>>>>>>> Software Engineer, Zappos.com
>>>>>>>>> jnewb...@zappos.com - 702-943-7562
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>                   
>>>>>>>>
>>>>>>>>                 
>>>>>>
>>>>>>             
>>>>         
>>>
>>>       
>> --
>> - Mark
>>
>> http://www.lucidimagination.com
>>
>>
>>
>>
>>     


-- 
- Mark

http://www.lucidimagination.com



Reply via email to