1 .what does this specify ?

<queryResultCache class="*solr.LRUCache*"
size="*${queryResultCacheSize:0}*"initialSize
="*${queryResultCacheInitialSize:0}*" autowarmCount="*
${queryResultCacheRows:0}*" />

2.

when i say *queryResultCacheSize : 512 *, does it mean 512 queries can be
cached or 512 bytes are reserved for caching ?

can some please give me an answer ?



On 14 August 2011 21:41, Erick Erickson <erickerick...@gmail.com> wrote:

> Yep.
>
> ResultWindowSize in
> >> solrconfig.xml
> >>
> >> Best
> >> Erick
> >>
> >> On Sun, Aug 14, 2011 at 8:35 AM, jame vaalet <jamevaa...@gmail.com>
> wrote:
> >> > thanks erick ... that means it depends upon the memory allocated to
> the
> >> JVM
> >> > .
> >> >
> >> > going back queryCacheResults factor i have got this doubt ..
> >> > say, i have got 10 threads with 10 different queries ..and each of
> them
> >> in
> >> > parallel are searching the same index with millions of docs in it
> >> > (multisharded ) .
> >> > now each of the queries have large number of results in it hence got
> to
> >> page
> >> > them all..
> >> > which all thread's (query ) result-set will be cached ? so that
> >> subsequent
> >> > pages can be retrieved quickly ..?
> >> >
> >> > On 14 August 2011 17:40, Erick Erickson <erickerick...@gmail.com>
> wrote:
> >> >
> >> >> There isn't an "optimum" page size that I know of, it'll vary with
> lots
> >> of
> >> >> stuff, not the least of which is whatever servlet container limits
> there
> >> >> are.
> >> >>
> >> >> But I suspect you can get quite a few (1000s) without
> >> >> too much problem, and you can always use the JSON response
> >> >> writer to pack in more pages with less overhead.
> >> >>
> >> >> You pretty much have to try it and see.
> >> >>
> >> >> Best
> >> >> Erick
> >> >>
> >> >> On Sun, Aug 14, 2011 at 5:42 AM, jame vaalet <jamevaa...@gmail.com>
> >> wrote:
> >> >> > speaking about pagesizes, what is the optimum page size that should
> be
> >> >> > retrieved each time ??
> >> >> > i understand it depends upon the data you are fetching back
> fromeach
> >> hit
> >> >> > document ... but lets say when ever a document is hit am fetching
> back
> >> >> 100
> >> >> > bytes worth data from each of those docs in indexes (along with
> solr
> >> >> > response statements ) .
> >> >> > this will make 100*x bytes worth data in each page if x is the page
> >> size
> >> >> ..
> >> >> > what is the optimum value of this x that solr can return each time
> >> >> without
> >> >> > going into exceptions ....
> >> >> >
> >> >> > On 13 August 2011 19:59, Erick Erickson <erickerick...@gmail.com>
> >> wrote:
> >> >> >
> >> >> >> Jame:
> >> >> >>
> >> >> >> You control the number via settings in solrconfig.xml, so it's
> >> >> >> up to you.
> >> >> >>
> >> >> >> Jonathan:
> >> >> >> Hmmm, that's seems right, after all the "deep paging" penalty is
> >> really
> >> >> >> about keeping a large sorted array in memory.... but at least you
> >> only
> >> >> >> pay it once per 10,000, rather than 100 times (assuming page size
> is
> >> >> >> 100)...
> >> >> >>
> >> >> >> Best
> >> >> >> Erick
> >> >> >>
> >> >> >> On Wed, Aug 10, 2011 at 10:58 AM, jame vaalet <
> jamevaa...@gmail.com>
> >> >> >> wrote:
> >> >> >> > when you say queryResultCache, does it only cache n number of
> >> result
> >> >> for
> >> >> >> the
> >> >> >> > last one query or more than one queries?
> >> >> >> >
> >> >> >> >
> >> >> >> > On 10 August 2011 20:14, simon <mtnes...@gmail.com> wrote:
> >> >> >> >
> >> >> >> >> Worth remembering there are some performance penalties with
> deep
> >> >> >> >> paging, if you use the page-by-page approach. may not be too
> much
> >> of
> >> >> a
> >> >> >> >> problem if you really are only looking to retrieve 10K docs.
> >> >> >> >>
> >> >> >> >> -Simon
> >> >> >> >>
> >> >> >> >> On Wed, Aug 10, 2011 at 10:32 AM, Erick Erickson
> >> >> >> >> <erickerick...@gmail.com> wrote:
> >> >> >> >> > Well, if you really want to you can specify start=0 and
> >> rows=10000
> >> >> and
> >> >> >> >> > get them all back at once.
> >> >> >> >> >
> >> >> >> >> > You can do page-by-page by incrementing the "start" parameter
> as
> >> >> you
> >> >> >> >> > indicated.
> >> >> >> >> >
> >> >> >> >> > You can keep from re-executing the search by setting your
> >> >> >> >> queryResultCache
> >> >> >> >> > appropriately, but this affects all searches so might be an
> >> issue.
> >> >> >> >> >
> >> >> >> >> > Best
> >> >> >> >> > Erick
> >> >> >> >> >
> >> >> >> >> > On Wed, Aug 10, 2011 at 9:09 AM, jame vaalet <
> >> jamevaa...@gmail.com
> >> >> >
> >> >> >> >> wrote:
> >> >> >> >> >> hi,
> >> >> >> >> >> i want to retrieve all the data from solr (say 10,000 ids )
> and
> >> my
> >> >> >> page
> >> >> >> >> size
> >> >> >> >> >> is 1000 .
> >> >> >> >> >> how do i get back the data (pages) one after other ?do i
> have
> >> to
> >> >> >> >> increment
> >> >> >> >> >> the "start" value each time by the page size from 0 and do
> the
> >> >> >> iteration
> >> >> >> >> ?
> >> >> >> >> >> In this case am i querying the index 10 time instead of one
> or
> >> >> after
> >> >> >> >> first
> >> >> >> >> >> query the result will be cached somewhere for the subsequent
> >> pages
> >> >> ?
> >> >> >> >> >>
> >> >> >> >> >>
> >> >> >> >> >> JAME VAALET
> >> >> >> >> >>
> >> >> >> >> >
> >> >> >> >>
> >> >> >> >
> >> >> >> >
> >> >> >> >
> >> >> >> > --
> >> >> >> >
> >> >> >> > -JAME
> >> >> >> >
> >> >> >>
> >> >> >
> >> >> >
> >> >> >
> >> >> > --
> >> >> >
> >> >> > -JAME
> >> >> >
> >> >>
> >> >
> >> >
> >> >
> >> > --
> >> >
> >> > -JAME
> >> >
> >>
> >
> >
> >
> > --
> >
> > -JAME
> >
>



-- 

-JAME

Reply via email to