Erik,

Not answering your question directly, but how many cores does your Solr machine 
have?  If it has 2 cores, for example, then running 6 indexers against it 
likely doesn't make indexing faster.

Otis
--
Sematext -- http://sematext.com/ -- Lucene - Solr - Nutch



----- Original Message ----
> From: Erik Holstad <[EMAIL PROTECTED]>
> To: solr-user@lucene.apache.org
> Sent: Wednesday, September 24, 2008 3:24:51 PM
> Subject: java.io.IOException: cannot read directory 
> org.apache.lucene.store.FSDirectory@/home/solr/src/apache-solr-nightly/example/solr/data/index:
>  list() returned null
> 
> We are using Solr out of the box, with only a couple of changes in the
> solconfig file.
> 
> We are running a MapReduce job to import into Solr. Every map creates one
> document and used to
> add and commit it to Solr. We got org.apache.solr.common.SolrException:
> Error_opening_new_searcher_exceeded_limit_of_maxWarmingSearchers4_try_again_later,
> which we solved by removing the commit statment from the MR job and added
> auto-commit in solrconfig.
> 
> We reran the job and got another exception: java.io.IOException: cannot read
> directory 
> org.apache.lucene.store.FSDirectory@/home/solr/src/apache-solr-nightly/example/solr/data/index:
> list() returned null
> followed by: SEVERE: org.apache.lucene.store.LockObtainFailedException: Lock
> obtain timed out: SingleInstanceLock: write.lock
> 
> This was happening when the number of mappers writing to solr was 6, we
> lowered the number of inputters to 3 and everything worked fine.
> 
> Does anyone know what happens, and how we can use more than 3 input sources
> at the same time?
> 
> Regards Erik

Reply via email to