Whoops. There is my lazy brain for you - march, may, august - all the
same ;)

Okay - forgot Solr went straight down and used FieldSortedHitQueue.

So it all still makes sense ;)

Still interested in seeing his field sanity output to see whats possibly
being doubled.

Yonik Seeley wrote:
> On Thu, Oct 1, 2009 at 3:14 PM, Mark Miller <markrmil...@gmail.com> wrote:
>   
>> bq. Tons of changes since... including the per-segment
>> searching/sorting/function queries (I think).
>>
>> Yup. I actually didn't think so, because that was committed to Lucene in
>> Feburary - but it didn't come into Solr till March 10th. March 5th just
>> ducked it.
>>     
>
> Jeff said May 5th
>
> But it wasn't until the end of May that Solr started using Lucene's
> new sorting facilities that worked per-segment.
>
> -Yonik
> http://www.lucidimagination.com
>
>
>   
>> Yonik Seeley wrote:
>>     
>>> On Thu, Oct 1, 2009 at 11:41 AM, Jeff Newburn <jnewb...@zappos.com> wrote:
>>>
>>>       
>>>> I am trying to update to the newest version of solr from trunk as of May
>>>> 5th.
>>>>
>>>>         
>>> Tons of changes since... including the per-segment
>>> searching/sorting/function queries (I think).
>>>
>>> Do you sort on any single valued fields that you also facet on?
>>> Do you use ord() or rord() in any function queries?
>>>
>>> Unfortunately, some of these things will take up more memory because
>>> some things still cache FieldCache elements with the top-level reader,
>>> while some use segment readers.  The direction is going toward all
>>> segment readers, but we're not there yet (and won't be for 1.4).
>>> ord() rord() will never be fixed... people need to migrate to
>>> something else.
>>>
>>> http://issues.apache.org/jira/browse/SOLR-1111 is the main issue for this.
>>>
>>> If course, I've really only been talking about search related changes.
>>>  Nothing on the indexing side should cause greater memory usage....
>>> but perhaps the indexing side could run out of memory due to the
>>> search side taking up more.
>>>
>>> -Yonik
>>> http://www.lucidimagination.com
>>>
>>>
>>>       
>>>>  I updated and compiled from trunk as of yesterday (09/30/2009).  When
>>>> I try to do a full import I am receiving a GC heap error after changing
>>>> nothing in the configuration files.  Why would this happen in the most
>>>> recent versions but not in the version from a few months ago.  The stack
>>>> trace is below.
>>>>
>>>> Oct 1, 2009 8:34:32 AM org.apache.solr.update.processor.LogUpdateProcessor
>>>> finish
>>>> INFO: {add=[166400, 166608, 166698, 166800, 166811, 167097, 167316, 167353,
>>>> ...(83 more)]} 0 35991
>>>> Oct 1, 2009 8:34:32 AM org.apache.solr.common.SolrException log
>>>> SEVERE: java.lang.OutOfMemoryError: GC overhead limit exceeded
>>>>    at java.util.Arrays.copyOfRange(Arrays.java:3209)
>>>>    at java.lang.String.<init>(String.java:215)
>>>>    at com.ctc.wstx.util.TextBuffer.contentsAsString(TextBuffer.java:384)
>>>>    at com.ctc.wstx.sr.BasicStreamReader.getText(BasicStreamReader.java:821)
>>>>    at org.apache.solr.handler.XMLLoader.readDoc(XMLLoader.java:280)
>>>>    at org.apache.solr.handler.XMLLoader.processUpdate(XMLLoader.java:139)
>>>>    at org.apache.solr.handler.XMLLoader.load(XMLLoader.java:69)
>>>>    at
>>>> org.apache.solr.handler.ContentStreamHandlerBase.handleRequestBody(ContentSt
>>>> reamHandlerBase.java:54)
>>>>    at
>>>> org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHandlerBase.
>>>> java:131)
>>>>    at org.apache.solr.core.SolrCore.execute(SolrCore.java:1316)
>>>>    at
>>>> org.apache.solr.servlet.SolrDispatchFilter.execute(SolrDispatchFilter.java:3
>>>> 38)
>>>>    at
>>>> org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilter.java:
>>>> 241)
>>>>    at
>>>> org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(Application
>>>> FilterChain.java:235)
>>>>    at
>>>> org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterCh
>>>> ain.java:206)
>>>>    at
>>>> org.apache.catalina.core.StandardWrapperValve.invoke(StandardWrapperValve.ja
>>>> va:233)
>>>>    at
>>>> org.apache.catalina.core.StandardContextValve.invoke(StandardContextValve.ja
>>>> va:175)
>>>>    at
>>>> org.apache.catalina.core.StandardHostValve.invoke(StandardHostValve.java:128
>>>> )
>>>>    at
>>>> org.apache.catalina.valves.ErrorReportValve.invoke(ErrorReportValve.java:102
>>>> )
>>>>    at
>>>> org.apache.catalina.core.StandardEngineValve.invoke(StandardEngineValve.java
>>>> :109)
>>>>    at
>>>> org.apache.catalina.connector.CoyoteAdapter.service(CoyoteAdapter.java:286)
>>>>    at
>>>> org.apache.coyote.http11.Http11NioProcessor.process(Http11NioProcessor.java:
>>>> 879)
>>>>    at
>>>> org.apache.coyote.http11.Http11NioProtocol$Http11ConnectionHandler.process(H
>>>> ttp11NioProtocol.java:719)
>>>>    at
>>>> org.apache.tomcat.util.net.NioEndpoint$SocketProcessor.run(NioEndpoint.java:
>>>> 2080)
>>>>    at
>>>> java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.ja
>>>> va:886)
>>>>    at
>>>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:9
>>>> 08)
>>>>    at java.lang.Thread.run(Thread.java:619)
>>>>
>>>> Oct 1, 2009 8:40:06 AM org.apache.solr.core.SolrCore execute
>>>> INFO: [zeta-main] webapp=/solr path=/update params={} status=500 QTime=5265
>>>> Oct 1, 2009 8:40:12 AM org.apache.solr.common.SolrException log
>>>> SEVERE: java.lang.OutOfMemoryError: GC overhead limit exceeded
>>>>
>>>> --
>>>> Jeff Newburn
>>>> Software Engineer, Zappos.com
>>>> jnewb...@zappos.com - 702-943-7562
>>>>
>>>>
>>>>
>>>>         
>> --
>> - Mark
>>
>> http://www.lucidimagination.com
>>
>>
>>
>>
>>     


-- 
- Mark

http://www.lucidimagination.com



Reply via email to