try ulimit -n50000 or something

On Mon, Apr 6, 2009 at 6:28 PM, Jarek Zgoda <jarek.zg...@redefine.pl> wrote:
> I'm indexing a set of 500000 small documents. I'm adding documents in
> batches of 1000. At the beginning I had a setup that optimized the index
> each 10000 documents, but quickly I had to optimize after adding each batch
> of documents. Unfortunately, I'm still getting the "Too many open files" IO
> error on optimize. I went from mergeFactor of 25 down to 10, but I'm still
> unable to optimize the index.
>
> I have configuration:
>    <useCompoundFile>false</useCompoundFile>
>    <ramBufferSizeMB>256</ramBufferSizeMB>
>    <mergeFactor>2</mergeFactor>
>    <maxMergeDocs>2147483647</maxMergeDocs>
>    <maxFieldLength>10000</maxFieldLength>
>
> The machine (2 core AMD64, 4GB RAM) is running Debian Linux, Java is
> 1.6.0_11 64-Bit, Solr is nightly build (2009-04-02). And no, I can not
> change the limit of file descriptors (currently: 1024). What more can I do?
>
> --
> We read Knuth so you don't have to. - Tim Peters
>
> Jarek Zgoda, R&D, Redefine
> jarek.zg...@redefine.pl
>
>



-- 

+1 510 277-0891 (o)
+91 9999 33 7458 (m)

web: http://pajamadesign.com

Skype: pajamadesign
Yahoo: jacobsingh
AIM: jacobsingh
gTalk: jacobsi...@gmail.com

Reply via email to