Re: Good practices on indexing larger amount of documents at once using SolrJ

2018-07-24 Thread Arunan Sugunakumar
Dear Erick, Unfortunately I deleted the original Solr logs, so I couldn't post it here. But removing the hard commit from the loop solved my problem and made indexing faster. Now there are no errors thrown from the client side. Thanks Arunan On 22 July 2018 at 04:45, Erick Erickson wrote: > c

Re: Good practices on indexing larger amount of documents at once using SolrJ

2018-07-21 Thread Erick Erickson
commitWithin parameter. Well, what I usually do is set my autocommit interval in my solrconfig.xml file and forget about it. For searching, set your autosoftcommit in solrconfig.xml and forget about _that_. Here's more than you want to know about the topic. https://lucidworks.com/2013/08/23/under

Re: Good practices on indexing larger amount of documents at once using SolrJ

2018-07-20 Thread Arunan Sugunakumar
Dear Erick, Thank you for your reply. I initialize the arraylist variable with a new Array List after I add and commit the solrDocumentList into the solrClient. So I dont think I have the problem of ever increasing ArrayList. (I hope the add method in solrClient flushes the previous documents adde

Re: Good practices on indexing larger amount of documents at once using SolrJ

2018-07-20 Thread Erick Erickson
I do this all the time with batches of 1,000 and don't see this problem. one thing that sometimes bites people is to fail to clear the doclist after every call to add. So you send ever-increasing batches to Solr. Assuming when you talk about batch size meaning the size of the solrDocunentList, inc