I routinely see hit rates over 75% on the document cache. Perhaps yours is too 
small. Mine is set at 10240 entries.

wunder

On Jan 20, 2013, at 8:08 AM, Erick Erickson wrote:

> About your question about document cache: Typically the document cache
> has a pretty low hit-ratio. I've rarely, if ever, seen it get hit very
> often. And remember that this cache is only hit when assembling the
> response for a few documents (your page size).
> 
> Bottom line: I wouldn't worry about this cache much. It's quite useful
> for processing a particular query faster, but not really intended for
> cross-query use.
> 
> Really, I think you're getting the cart before the horse here. Run it
> up the flagpole and try it. Rely on the OS to do its job
> (http://blog.thetaphi.de/2012/07/use-lucenes-mmapdirectory-on-64bit.html).
> Find  a bottleneck _then_ tune. Premature optimization and all
> that....
> 
> Several tens of millions of docs isn't that large unless the text
> fields are enormous.
> 
> Best
> Erick
> 
> On Sat, Jan 19, 2013 at 2:32 PM, Isaac Hebsh <isaac.he...@gmail.com> wrote:
>> Ok. Thank you everyone for your helpful answers.
>> I understand that fieldValueCache is not used for resolving queries.
>> Is there any cache that can help this basic scenario (a lot of different
>> queries, on a small set of fields)?
>> Does Lucene's FieldCache help (implicitly)?
>> How can I use RAM to reduce I/O in this type of queries?
>> 
>> 
>> On Fri, Jan 18, 2013 at 4:09 PM, Tomás Fernández Löbbe <
>> tomasflo...@gmail.com> wrote:
>> 
>>> No, the fieldValueCache is not used for resolving queries. Only for
>>> multi-token faceting and apparently for the stats component too. The
>>> document cache maintains in memory the stored content of the fields you are
>>> retrieving or highlighting on. It'll hit if the same document matches the
>>> query multiple times and the same fields are requested, but as Eirck said,
>>> it is important for cases when multiple components in the same request need
>>> to access the same data.
>>> 
>>> I think soft committing every 10 minutes is totally fine, but you should
>>> hard commit more often if you are going to be using transaction log.
>>> openSearcher=false will essentially tell Solr not to open a new searcher
>>> after the (hard) commit, so you won't see the new indexed data and caches
>>> wont be flushed. openSearcher=false makes sense when you are using
>>> hard-commits together with soft-commits, as the "soft-commit" is dealing
>>> with opening/closing searchers, you don't need hard commits to do it.
>>> 
>>> Tomás
>>> 
>>> 
>>> On Fri, Jan 18, 2013 at 2:20 AM, Isaac Hebsh <isaac.he...@gmail.com>
>>> wrote:
>>> 
>>>> Unfortunately, it seems (
>>>> http://lucene.472066.n3.nabble.com/Nrt-and-caching-td3993612.html) that
>>>> these caches are not per-segment. In this case, I want to (soft) commit
>>>> less frequently. Am I right?
>>>> 
>>>> Tomás, as the fieldValueCache is very similar to lucene's FieldCache, I
>>>> guess it has a big contribution to standard (not only faceted) queries
>>>> time. SolrWiki claims that it primarily used by faceting. What that says
>>>> about complex textual queries?
>>>> 
>>>> documentCache:
>>>> Erick, After a query processing is finished, doesn't some documents stay
>>> in
>>>> the documentCache? can't I use it to accelerate queries that should
>>>> retrieve stored fields of documents? In this case, a big documentCache
>>> can
>>>> hold more documents..
>>>> 
>>>> About commit frequency:
>>>> HardCommit: "openSearch=false" seems as a nice solution. Where can I read
>>>> about this? (found nothing but one unexplained sentence in SolrWiki).
>>>> SoftCommit: In my case, the required index freshness is 10 minutes. The
>>>> plan to soft commit every 10 minutes is similar to storing all of the
>>>> documents in a queue (outside to Solr), an indexing a bulk every 10
>>>> minutes.
>>>> 
>>>> Thanks.
>>>> 
>>>> 
>>>> On Fri, Jan 18, 2013 at 2:15 AM, Tomás Fernández Löbbe <
>>>> tomasflo...@gmail.com> wrote:
>>>> 
>>>>> I think fieldValueCache is not per segment, only fieldCache is.
>>> However,
>>>>> unless I'm missing something, this cache is only used for faceting on
>>>>> multivalued fields
>>>>> 
>>>>> 
>>>>> On Thu, Jan 17, 2013 at 8:58 PM, Erick Erickson <
>>> erickerick...@gmail.com
>>>>>> wrote:
>>>>> 
>>>>>> filterCache: This is bounded by 1M * (maxDoc) / 8 * (num filters in
>>>>>> cache). Notice the /8. This reflects the fact that the filters are
>>>>>> represented by a bitset on the _internal_ Lucene ID. UniqueId has no
>>>>>> bearing here whatsoever. This is, in a nutshell, why warming is
>>>>>> required, the internal Lucene IDs may change. Note also that it's
>>>>>> maxDoc, the internal arrays have "holes" for deleted documents.
>>>>>> 
>>>>>> Note this is an _upper_ bound, if there are only a few docs that
>>>>>> match, the size will be (num of matching docs) * sizeof(int)).
>>>>>> 
>>>>>> fieldValueCache. I don't think so, although I'm a bit fuzzy on this.
>>>>>> It depends on whether these are "per-segment" caches or not. Any "per
>>>>>> segment" cache is still valid.
>>>>>> 
>>>>>> Think of documentCache as intended to hold the stored fields while
>>>>>> various components operate on it, thus avoiding repeatedly fetching
>>>>>> the data from disk. It's _usually_ not too big a worry.
>>>>>> 
>>>>>> About hard-commits once a day. That's _extremely_ long. Think instead
>>>>>> of committing more frequently with openSearcher=false. If nothing
>>>>>> else, you transaction log will grow lots and lots and lots. I'm
>>>>>> thinking on the order of 15 minutes, or possibly even much less. With
>>>>>> softCommits happening more often, maybe every 15 seconds. In fact,
>>> I'd
>>>>>> start out with soft commits every 15 seconds and hard commits
>>>>>> (openSearcher=false) every 5 minutes. The problem with hard commits
>>>>>> being once a day is that, if for any reason the server is
>>> interrupted,
>>>>>> on startup Solr will try to replay the entire transaction log to
>>>>>> assure index integrity. Not to mention that your tlog will be huge.
>>>>>> Not to mention that there is some memory usage for each document in
>>>>>> the tlog. Hard commits roll over the tlog, flush the in-memory tlog
>>>>>> pointers, close index segments, etc.
>>>>>> 
>>>>>> Best
>>>>>> Erick
>>>>>> 
>>>>>> On Thu, Jan 17, 2013 at 1:29 PM, Isaac Hebsh <isaac.he...@gmail.com>
>>>>>> wrote:
>>>>>>> Hi,
>>>>>>> 
>>>>>>> I am going to build a big Solr (4.0?) index, which holds some
>>> dozens
>>>> of
>>>>>>> millions of documents. Each document has some dozens of fields, and
>>>> one
>>>>>> big
>>>>>>> textual field.
>>>>>>> The queries on the index are non-trivial, and a little-bit long
>>>> (might
>>>>> be
>>>>>>> hundreds of terms). No query is identical to another.
>>>>>>> 
>>>>>>> Now, I want to analyze the cache performance (before setting up the
>>>>> whole
>>>>>>> environment), in order to estimate how much RAM will I need.
>>>>>>> 
>>>>>>> filterCache:
>>>>>>> In my scenariom, every query has some filters. let's say that each
>>>>> filter
>>>>>>> matches 1M documents, out of 10M. Does the estimated memory usage
>>>>> should
>>>>>> be
>>>>>>> 1M * sizeof(uniqueId) * num-of-filters-in-cache?
>>>>>>> 
>>>>>>> fieldValueCache:
>>>>>>> Due to the difference between queries, I guess that fieldValueCache
>>>> is
>>>>>> the
>>>>>>> most important factor on query performance. Here comes a generic
>>>>>> question:
>>>>>>> I'm indexing new documents to the index constantly. Soft commits
>>> will
>>>>> be
>>>>>>> performed every 10 mins. Does it say that the cache is meaningless,
>>>>> after
>>>>>>> every 10 minutes?
>>>>>>> 
>>>>>>> documentCache:
>>>>>>> enableLazyFieldLoading will be enabled, and "fl" contains a very
>>>> small
>>>>>> set
>>>>>>> of fields. BUT, I need to return highlighting on about (possibly)
>>> 20
>>>>>>> fields. Does the highlighting component use the documentCache? I
>>>> guess
>>>>>> that
>>>>>>> highlighting requires the whole field to be loaded into the
>>>>>> documentCache.
>>>>>>> Will it happen only for fields that matched a term from the query?
>>>>>>> 
>>>>>>> And one more question: I'm planning to hard-commit once a day.
>>>> Should I
>>>>>>> prepare to a significant RAM usage growth between hard-commits?
>>>>>> (consider a
>>>>>>> lot of new documents in this period...)
>>>>>>> Does this RAM comes from the same pool as the caches? An
>>> OutOfMemory
>>>>>>> exception can happen is this scenario?
>>>>>>> 
>>>>>>> Thanks a lot.
>>>>>> 
>>>>> 
>>>> 
>>> 

--
Walter Underwood
wun...@wunderwood.org



Reply via email to