Make it 10: <mergeFactor>10</mergeFactor> -Fuad
> -----Original Message----- > From: Ranganathan, Sharmila [mailto:sranganat...@library.rochester.edu] > Sent: October-23-09 1:08 PM > To: solr-user@lucene.apache.org > Subject: Too many open files > > Hi, > > I am getting too many open files error. > > Usually I test on a server that has 4GB RAM and assigned 1GB for > tomcat(set JAVA_OPTS=-Xms256m -Xmx1024m), ulimit -n is 256 for this > server and has following setting for SolrConfig.xml > > > > <useCompoundFile>true</useCompoundFile> > > <ramBufferSizeMB>1024</ramBufferSizeMB> > > <mergeFactor>100</mergeFactor> > > <maxMergeDocs>2147483647</maxMergeDocs> > > <maxFieldLength>10000</maxFieldLength> > > > > In my case 200,000 documents is of 1024MB size and in this testing, I am > indexing total of million documents. We have high setting because we are > expected to index about 10+ million records in production. It works fine > in this server. > > > > When I deploy same solr configuration on a server with 32GB RAM, I get > "too many open files" error. The ulimit -n is 1024 for this server. Any > idea? Is this because 2nd server has 32GB RAM? Is 1024 open files limit > too low? Also I don't find any documentation for <ramBufferSizeMB>. > I checked Solr 'Solr 1.4 Enterprise Search Server' book, wiki, etc. I am > using Solr 1.3. > > > > Is it good idea to use ramBufferSizeMB? Vs maxBufferedDocs? What does > ramBufferSizeMB mean? My understanding is that when documents added to > index which are initially stored in memory reaches size > 1024MB(ramBufferSizeMB), it flushes data to disk. Or is it when total > memory used(by tomcat, etc) reaches 1024, it flushed data to disk? > > > > Thanks, > > Sharmila > > > > > > > > > > > > > >