[image: Immagine incorporata 1]

2016-02-11 16:05 GMT+01:00 Matteo Grolla <matteo.gro...@gmail.com>:

> I see a lot of time spent in splitOnTokens
>
> which is called by (last part of stack trace)
>
> BinaryResponseWriter$Resolver.writeResultsBody()
> ...
> solr.search.ReturnsField.wantsField()
> commons.io.FileNameUtils.wildcardmatch()
> commons.io.FileNameUtils.splitOnTokens()
>
>
>
> 2016-02-11 15:42 GMT+01:00 Matteo Grolla <matteo.gro...@gmail.com>:
>
>> Hi Yonic,
>>      after the first query I find 1000 docs in the document cache.
>> I'm using curl to send the request and requesting javabin format to mimic
>> the application.
>> gc activity is low
>> I managed to load the entire 50GB index in the filesystem cache, after
>> that queries don't cause disk activity anymore.
>> Time improves now queries that took ~30s take <10s. But I hoped better
>> I'm going to use jvisualvm's sampler to analyze where time is spent
>>
>>
>> 2016-02-11 15:25 GMT+01:00 Yonik Seeley <ysee...@gmail.com>:
>>
>>> On Thu, Feb 11, 2016 at 7:45 AM, Matteo Grolla <matteo.gro...@gmail.com>
>>> wrote:
>>> > Thanks Toke, yes, they are long times, and solr qtime (to execute the
>>> > query) is a fraction of a second.
>>> > The response in javabin format is around 300k.
>>>
>>> OK, That tells us a lot.
>>> And if you actually tested so that all the docs would be in the cache
>>> (can you verify this by looking at the cache stats after you
>>> re-execute?) then it seems like the slowness is down to any of:
>>> a) serializing the response (it doesn't seem like a 300K response
>>> should take *that* long to serialize)
>>> b) reading/processing the response (how fast the client can do
>>> something with each doc is also a factor...)
>>> c) other (GC, network, etc)
>>>
>>> You can try taking client processing out of the equation by trying a
>>> curl request.
>>>
>>> -Yonik
>>>
>>
>>
>

Reply via email to