Oleg, you can't explicitly say "N GB for index". Wunder was just saying how much you can imagine how much RAM each piece might need and be happy with. Otis -- Sematext -- http://sematext.com/ -- Lucene - Solr - Nutch
----- Original Message ---- From: oleg_gnatovskiy <[EMAIL PROTECTED]> To: solr-user@lucene.apache.org Sent: Wednesday, April 16, 2008 2:05:23 PM Subject: Re: too many queries? Hello. I am having a similar problem as the OP. I see that you recommended setting 4GB for the index, and 2 for Solr. How do I allocate memory for the index? I was under the impression that Solr did not support a RAMIndex. Walter Underwood wrote: > > Do it. 32-bit OS's went out of style five years ago in server-land. > > I would start with 8GB of RAM. 4GB for your index, 2 for Solr, 1 for > the OS and 1 for other processes. That might be tight. 12GB would > be a lot better. > > wunder > > On 4/16/08 7:50 AM, "Jonathan Ariel" <[EMAIL PROTECTED]> wrote: > >> In order to do that I have to change to a 64 bits OS so I can have more >> than >> 4 GB of RAM.Is there any way to see how long does it takes to Solr to >> warmup >> the searcher? >> >> On Wed, Apr 16, 2008 at 11:40 AM, Walter Underwood >> <[EMAIL PROTECTED]> >> wrote: >> >>> A commit every two minutes means that the Solr caches are flushed >>> before they even start to stabilize. Two things to try: >>> >>> * commit less often, 5 minutes or 10 minutes >>> * have enough RAM that your entire index can fit in OS file buffers >>> >>> wunder >>> >>> On 4/16/08 6:27 AM, "Jonathan Ariel" <[EMAIL PROTECTED]> wrote: >>> >>>> So I counted the number if distinct values that I have for each field >>> that I >>>> want a facet on. In total it's around 100,000. I tried with a >>> filterCache >>>> of 120,000 but it seems like too much because the server went down. I >>> will >>>> try with less, around 75,000 and let you know. >>>> >>>> How do you to partition the data to a static set and a dynamic set, and >>> then >>>> combining them at query time? Do you have a link to read about that? >>>> >>>> >>>> >>>> On Tue, Apr 15, 2008 at 7:21 PM, Mike Klaas <[EMAIL PROTECTED]> >>> wrote: >>>> >>>>> On 15-Apr-08, at 5:38 AM, Jonathan Ariel wrote: >>>>> >>>>>> My index is 4GB on disk. My servers has 8 GB of RAM each (the OS is >>>>>> 32 >>>>>> bits). >>>>>> It is optimized twice a day, it takes around 15 minutes to optimize. >>>>>> The index is updated (commits) every two minutes. There are between >>>>>> 10 >>>>>> and >>>>>> 100 inserts/updates every 2 minutes. >>>>>> >>>>> >>>>> Caching could help--you should definitely start there. >>>>> >>>>> The commit every 2 minutes could end up being an unsurmountable >>> problem. >>>>> You may have to partition your data into a large, mostly static set >>> and a >>>>> small dynamic set, combining the results at query time. >>>>> >>>>> -Mike >>>>> >>> >>> > > > -- View this message in context: http://www.nabble.com/too-many-queries--tp16690870p16727264.html Sent from the Solr - User mailing list archive at Nabble.com.