Can you share more about your setup, specifically what you changed in
your solrconfig file? What version of Solr (looks like a nightly, but
from when)? What did you set auto-commit to be? Can you provide the
full stack trace? Also, were you starting fresh when you got the
second exception?
On Sep 24, 2008, at 3:24 PM, Erik Holstad wrote:
We are using Solr out of the box, with only a couple of changes in the
solconfig file.
We are running a MapReduce job to import into Solr. Every map
creates one
document and used to
add and commit it to Solr. We got
org.apache.solr.common.SolrException:
Error_opening_new_searcher_exceeded_limit_of_maxWarmingSearchers4_try_again_later
,
which we solved by removing the commit statment from the MR job and
added
auto-commit in solrconfig.
We reran the job and got another exception: java.io.IOException:
cannot read
directory org.apache.lucene.store.FSDirectory@/home/solr/src/apache-
solr-nightly/example/solr/data/index:
list() returned null
followed by: SEVERE:
org.apache.lucene.store.LockObtainFailedException: Lock
obtain timed out: SingleInstanceLock: write.lock
This was happening when the number of mappers writing to solr was 6,
we
lowered the number of inputters to 3 and everything worked fine.
Does anyone know what happens, and how we can use more than 3 input
sources
at the same time?
Regards Erik