That is exactly what we are doing now added all the documents to the server in the Map phase of the job and send them all to on reducer, which commits them all. Seems to be working.
Thanks Erik On Wed, Sep 24, 2008 at 2:27 PM, Otis Gospodnetic < [EMAIL PROTECTED]> wrote: > Erik, > There is little benefit from having more indexer threads than cores. > You have multiple indexers calling commit? I suggest you make only one of > them call commit. Or use autoCommit. > > Otis > -- > Sematext -- http://sematext.com/ -- Lucene - Solr - Nutch > > > > ----- Original Message ---- > > From: Erik Holstad <[EMAIL PROTECTED]> > > To: solr-user@lucene.apache.org > > Sent: Wednesday, September 24, 2008 4:16:06 PM > > Subject: Re: java.io.IOException: cannot read directory > org.apache.lucene.store.FSDirectory@/home/solr/src/apache-solr-nightly/example/solr/data/index: > list() returned null > > > > Otis, > > > > The machine we are running on has 4 cores, and that seems to make sense, > > since running for inserters also > > failed. So what you are saying is that one inserter uses 1 core? So we > can > > only have as many methods calling the commit() > > as we have cores? > > > > Regards Erik > > > > On Wed, Sep 24, 2008 at 12:48 PM, Otis Gospodnetic < > > [EMAIL PROTECTED]> wrote: > > > > > Erik, > > > > > > Not answering your question directly, but how many cores does your Solr > > > machine have? If it has 2 cores, for example, then running 6 indexers > > > against it likely doesn't make indexing faster. > > > > > > Otis > > > -- > > > Sematext -- http://sematext.com/ -- Lucene - Solr - Nutch > > > > > > > > > > > > ----- Original Message ---- > > > > From: Erik Holstad > > > > To: solr-user@lucene.apache.org > > > > Sent: Wednesday, September 24, 2008 3:24:51 PM > > > > Subject: java.io.IOException: cannot read directory > > > > > org.apache.lucene.store.FSDirectory@ > /home/solr/src/apache-solr-nightly/example/solr/data/index: > > > list() returned null > > > > > > > > We are using Solr out of the box, with only a couple of changes in > the > > > > solconfig file. > > > > > > > > We are running a MapReduce job to import into Solr. Every map creates > one > > > > document and used to > > > > add and commit it to Solr. We got > org.apache.solr.common.SolrException: > > > > > > > > > > Error_opening_new_searcher_exceeded_limit_of_maxWarmingSearchers4_try_again_later, > > > > which we solved by removing the commit statment from the MR job and > added > > > > auto-commit in solrconfig. > > > > > > > > We reran the job and got another exception: java.io.IOException: > cannot > > > read > > > > directory > > > > org.apache.lucene.store.FSDirectory@ > > > /home/solr/src/apache-solr-nightly/example/solr/data/index: > > > > list() returned null > > > > followed by: SEVERE: > org.apache.lucene.store.LockObtainFailedException: > > > Lock > > > > obtain timed out: SingleInstanceLock: write.lock > > > > > > > > This was happening when the number of mappers writing to solr was 6, > we > > > > lowered the number of inputters to 3 and everything worked fine. > > > > > > > > Does anyone know what happens, and how we can use more than 3 input > > > sources > > > > at the same time? > > > > > > > > Regards Erik > > > > > > > >