Jason, What you read is valid advice. Just don't commit that often, or even at all until the very end if you can wait. :) And make sure you are indexing to a machine that doesn't warm up caches and searcher every time you commit.
Otis ---- Sematext :: http://sematext.com/ :: Solr - Lucene - Nutch Lucene ecosystem search :: http://search-lucene.com/ >________________________________ >From: Jason Toy <jason...@gmail.com> >To: solr-user@lucene.apache.org >Sent: Tuesday, November 15, 2011 3:41 PM >Subject: getting lots of errors doing bulk insertion > >I've written a script that does bulk insertion from my database, it >grabs chunks of 500 docs (out of 100 million ) and inserts them into >solr over http. I have 5 threads that are inserting from a queue. >After each insert I issue a commit. >Every 20 or so inserts I get this error message: >Error: Error opening new searcher. exceeded limit of >maxWarmingSearchers=2, try again later. > >I saw that people suggest to reduce the commit frequency to fix this. >Is this really the way to need to fix this? The reason I was >committing after every insert of 500 docs is so there would not be too >much uncommmited data. > > > > >-- >- sent from my mobile > > >