Otis,

 We are not running master-slave configuration. We get very few
searches(admin only) in a day so we didn't see the need of
replication/snapshot. This problem is with one Solr instance managing
4 cores (each core 200 million records). Both indexing and searching
is performed by the same Solr instance.

What are .tii files used for? I see this file under only one core.

Still looking for what gets loaded in heap by Solr (during load time,
indexing, and searching) and stays there. I see most of these are
tenured objects and not getting released by GC - will post profile
records tomorrow.

Thanks,
-vivek





On Wed, May 13, 2009 at 6:34 PM, Otis Gospodnetic
<otis_gospodne...@yahoo.com> wrote:
>
> There is constant mixing of indexing concepts and searching concepts in this 
> thread.  Are you having problems on the master (indexing) or on the slave 
> (searching)?
>
>
> That .tii is only 20K and you said this is a large index?  That doesn't smell 
> right...
>
> Otis
> --
> Sematext -- http://sematext.com/ -- Lucene - Solr - Nutch
>
>
>
> ----- Original Message ----
>> From: vivek sar <vivex...@gmail.com>
>> To: solr-user@lucene.apache.org
>> Sent: Wednesday, May 13, 2009 5:12:00 PM
>> Subject: Re: Solr memory requirements?
>>
>> Otis,
>>
>> In that case, I'm not sure why Solr is taking up so much memory as
>> soon as we start it up. I checked for .tii file and there is only one,
>>
>> -rw-r--r--  1 search  staff  20306 May 11 21:47 
>> ./20090510_1/data/index/_3au.tii
>>
>> I have all the cache disabled - so that shouldn't be a problem too. My
>> ramBuffer size is only 64MB.
>>
>> I read note on sorting,
>> http://wiki.apache.org/solr/SchemaDesign?highlight=(sort), and see
>> something related to FieldCache. I don't see this as parameter defined
>> in either solrconfig.xml or schema.xml. Could this be something that
>> can load things in memory at startup? How can we disable it?
>>
>> I'm trying to find out if there is a way to tell how much memory Solr
>> would consume and way to cap it.
>>
>> Thanks,
>> -vivek
>>
>>
>>
>>
>> On Wed, May 13, 2009 at 1:50 PM, Otis Gospodnetic
>> wrote:
>> >
>> > Hi,
>> >
>> > Sorting is triggered by the sort parameter in the URL, not a 
>> > characteristic of
>> a field. :)
>> >
>> > Otis
>> > --
>> > Sematext -- http://sematext.com/ -- Lucene - Solr - Nutch
>> >
>> >
>> >
>> > ----- Original Message ----
>> >> From: vivek sar
>> >> To: solr-user@lucene.apache.org
>> >> Sent: Wednesday, May 13, 2009 4:42:16 PM
>> >> Subject: Re: Solr memory requirements?
>> >>
>> >> Thanks Otis.
>> >>
>> >> Our use case doesn't require any sorting or faceting. I'm wondering if
>> >> I've configured anything wrong.
>> >>
>> >> I got total of 25 fields (15 are indexed and stored, other 10 are just
>> >> stored). All my fields are basic data type - which I thought are not
>> >> sorted. My id field is unique key.
>> >>
>> >> Is there any field here that might be getting sorted?
>> >>
>> >>
>> >> required="true" omitNorms="true" compressed="false"/>
>> >>
>> >>
>> >> compressed="false"/>
>> >>
>> >> omitNorms="true" compressed="false"/>
>> >>
>> >> omitNorms="true" compressed="false"/>
>> >>
>> >> omitNorms="true" compressed="false"/>
>> >>
>> >> default="NOW/HOUR"  compressed="false"/>
>> >>
>> >> omitNorms="true" compressed="false"/>
>> >>
>> >> omitNorms="true" compressed="false"/>
>> >>
>> >> compressed="false"/>
>> >>
>> >> compressed="false"/>
>> >>
>> >> omitNorms="true" compressed="false"/>
>> >>
>> >> omitNorms="true" compressed="false"/>
>> >>
>> >> omitNorms="true" compressed="false"/>
>> >>
>> >> omitNorms="true" compressed="false"/>
>> >>
>> >> omitNorms="true" compressed="false"/>
>> >>
>> >> compressed="false"/>
>> >>
>> >> compressed="false"/>
>> >>
>> >> compressed="false"/>
>> >>
>> >> omitNorms="true" compressed="false"/>
>> >>
>> >> compressed="false"/>
>> >>
>> >> default="NOW/HOUR" omitNorms="true"/>
>> >>
>> >>
>> >>
>> >>
>> >> omitNorms="true" multiValued="true"/>
>> >>
>> >> Thanks,
>> >> -vivek
>> >>
>> >> On Wed, May 13, 2009 at 1:10 PM, Otis Gospodnetic
>> >> wrote:
>> >> >
>> >> > Hi,
>> >> > Some answers:
>> >> > 1) .tii files in the Lucene index.  When you sort, all distinct values 
>> >> > for
>> the
>> >> field(s) used for sorting.  Similarly for facet fields.  Solr caches.
>> >> > 2) ramBufferSizeMB dictates, more or less, how much Lucene/Solr will
>> consume
>> >> during indexing.  There is no need to commit every 50K docs unless you 
>> >> want
>> to
>> >> trigger snapshot creation.
>> >> > 3) see 1) above
>> >> >
>> >> > 1.5 billion docs per instance where each doc is cca 1KB?  I doubt that's
>> going
>> >> to fly. :)
>> >> >
>> >> > Otis
>> >> > --
>> >> > Sematext -- http://sematext.com/ -- Lucene - Solr - Nutch
>> >> >
>> >> >
>> >> >
>> >> > ----- Original Message ----
>> >> >> From: vivek sar
>> >> >> To: solr-user@lucene.apache.org
>> >> >> Sent: Wednesday, May 13, 2009 3:04:46 PM
>> >> >> Subject: Solr memory requirements?
>> >> >>
>> >> >> Hi,
>> >> >>
>> >> >>   I'm pretty sure this has been asked before, but I couldn't find a
>> >> >> complete answer in the forum archive. Here are my questions,
>> >> >>
>> >> >> 1) When solr starts up what does it loads up in the memory? Let's say
>> >> >> I've 4 cores with each core 50G in size. When Solr comes up how much
>> >> >> of it would be loaded in memory?
>> >> >>
>> >> >> 2) How much memory is required during index time? If I'm committing
>> >> >> 50K records at a time (1 record = 1KB) using solrj, how much memory do
>> >> >> I need to give to Solr.
>> >> >>
>> >> >> 3) Is there a minimum memory requirement by Solr to maintain a certain
>> >> >> size index? Is there any benchmark on this?
>> >> >>
>> >> >> Here are some of my configuration from solrconfig.xml,
>> >> >>
>> >> >> 1) 64
>> >> >> 2) All the caches (under query tag) are commented out
>> >> >> 3) Few others,
>> >> >>       a)  true    ==>
>> >> >> would this require memory?
>> >> >>       b)  50
>> >> >>       c) 200
>> >> >>       d)
>> >> >>       e) false
>> >> >>       f)  2
>> >> >>
>> >> >> The problem we are having is following,
>> >> >>
>> >> >> I've given Solr RAM of 6G. As the total index size (all cores
>> >> >> combined) start growing the Solr memory consumption  goes up. With 800
>> >> >> million documents, I see Solr already taking up all the memory at
>> >> >> startup. After that the commits, searches everything become slow. We
>> >> >> will be having distributed setup with multiple Solr instances (around
>> >> >> 8) on four boxes, but our requirement is to have each Solr instance at
>> >> >> least maintain around 1.5 billion documents.
>> >> >>
>> >> >> We are trying to see if we can somehow reduce the Solr memory
>> >> >> footprint. If someone can provide a pointer on what parameters affect
>> >> >> memory and what effects it has we can then decide whether we want that
>> >> >> parameter or not. I'm not sure if there is any minimum Solr
>> >> >> requirement for it to be able maintain large indexes. I've used Lucene
>> >> >> before and that didn't require anything by default - it used up memory
>> >> >> only during index and search times - not otherwise.
>> >> >>
>> >> >> Any help is very much appreciated.
>> >> >>
>> >> >> Thanks,
>> >> >> -vivek
>> >> >
>> >> >
>> >
>> >
>
>

Reply via email to