As many results will be cached as you ask. See solrconfig.xml,
the queryResultCache. This cache is essentially a map of queries
and result document IDs. The number of doc IDs cached for
each query is controlled by queryResultWindowSize in
solrconfig.xml

Best
Erick

On Sun, Aug 14, 2011 at 8:35 AM, jame vaalet <jamevaa...@gmail.com> wrote:
> thanks erick ... that means it depends upon the memory allocated to the JVM
> .
>
> going back queryCacheResults factor i have got this doubt ..
> say, i have got 10 threads with 10 different queries ..and each of them in
> parallel are searching the same index with millions of docs in it
> (multisharded ) .
> now each of the queries have large number of results in it hence got to page
> them all..
> which all thread's (query ) result-set will be cached ? so that subsequent
> pages can be retrieved quickly ..?
>
> On 14 August 2011 17:40, Erick Erickson <erickerick...@gmail.com> wrote:
>
>> There isn't an "optimum" page size that I know of, it'll vary with lots of
>> stuff, not the least of which is whatever servlet container limits there
>> are.
>>
>> But I suspect you can get quite a few (1000s) without
>> too much problem, and you can always use the JSON response
>> writer to pack in more pages with less overhead.
>>
>> You pretty much have to try it and see.
>>
>> Best
>> Erick
>>
>> On Sun, Aug 14, 2011 at 5:42 AM, jame vaalet <jamevaa...@gmail.com> wrote:
>> > speaking about pagesizes, what is the optimum page size that should be
>> > retrieved each time ??
>> > i understand it depends upon the data you are fetching back fromeach hit
>> > document ... but lets say when ever a document is hit am fetching back
>> 100
>> > bytes worth data from each of those docs in indexes (along with solr
>> > response statements ) .
>> > this will make 100*x bytes worth data in each page if x is the page size
>> ..
>> > what is the optimum value of this x that solr can return each time
>> without
>> > going into exceptions ....
>> >
>> > On 13 August 2011 19:59, Erick Erickson <erickerick...@gmail.com> wrote:
>> >
>> >> Jame:
>> >>
>> >> You control the number via settings in solrconfig.xml, so it's
>> >> up to you.
>> >>
>> >> Jonathan:
>> >> Hmmm, that's seems right, after all the "deep paging" penalty is really
>> >> about keeping a large sorted array in memory.... but at least you only
>> >> pay it once per 10,000, rather than 100 times (assuming page size is
>> >> 100)...
>> >>
>> >> Best
>> >> Erick
>> >>
>> >> On Wed, Aug 10, 2011 at 10:58 AM, jame vaalet <jamevaa...@gmail.com>
>> >> wrote:
>> >> > when you say queryResultCache, does it only cache n number of result
>> for
>> >> the
>> >> > last one query or more than one queries?
>> >> >
>> >> >
>> >> > On 10 August 2011 20:14, simon <mtnes...@gmail.com> wrote:
>> >> >
>> >> >> Worth remembering there are some performance penalties with deep
>> >> >> paging, if you use the page-by-page approach. may not be too much of
>> a
>> >> >> problem if you really are only looking to retrieve 10K docs.
>> >> >>
>> >> >> -Simon
>> >> >>
>> >> >> On Wed, Aug 10, 2011 at 10:32 AM, Erick Erickson
>> >> >> <erickerick...@gmail.com> wrote:
>> >> >> > Well, if you really want to you can specify start=0 and rows=10000
>> and
>> >> >> > get them all back at once.
>> >> >> >
>> >> >> > You can do page-by-page by incrementing the "start" parameter as
>> you
>> >> >> > indicated.
>> >> >> >
>> >> >> > You can keep from re-executing the search by setting your
>> >> >> queryResultCache
>> >> >> > appropriately, but this affects all searches so might be an issue.
>> >> >> >
>> >> >> > Best
>> >> >> > Erick
>> >> >> >
>> >> >> > On Wed, Aug 10, 2011 at 9:09 AM, jame vaalet <jamevaa...@gmail.com
>> >
>> >> >> wrote:
>> >> >> >> hi,
>> >> >> >> i want to retrieve all the data from solr (say 10,000 ids ) and my
>> >> page
>> >> >> size
>> >> >> >> is 1000 .
>> >> >> >> how do i get back the data (pages) one after other ?do i have to
>> >> >> increment
>> >> >> >> the "start" value each time by the page size from 0 and do the
>> >> iteration
>> >> >> ?
>> >> >> >> In this case am i querying the index 10 time instead of one or
>> after
>> >> >> first
>> >> >> >> query the result will be cached somewhere for the subsequent pages
>> ?
>> >> >> >>
>> >> >> >>
>> >> >> >> JAME VAALET
>> >> >> >>
>> >> >> >
>> >> >>
>> >> >
>> >> >
>> >> >
>> >> > --
>> >> >
>> >> > -JAME
>> >> >
>> >>
>> >
>> >
>> >
>> > --
>> >
>> > -JAME
>> >
>>
>
>
>
> --
>
> -JAME
>

Reply via email to