that's pretty much my strategy.

I'll add parenthetically that I often see the bottleneck for indexing
to be acquiring the data from the system of record in the first place
rather than Solr. Assuming you're using SolrJ, an easy test is to
comment out the line that sends to Solr. There's usually some kind of
loop like:

while (more docs) {
    gather 1,000 docs into a list
    cloudSolrClient.add(docList);
    docList.clear()
}

So just comment out the cloudSolrClient.add line. I've seen situations
where the program still takes 95% of the time it takes to actually
index to Solr, in which case you need to focus on getting the data in
the first place.

And you need to batch updates, see:
https://lucidworks.com/2015/10/05/really-batch-updates-solr-2/

Good Luck!
Erick

On Fri, Jun 2, 2017 at 2:59 AM, gigo314 <gigo...@gmail.com> wrote:
> Thanks for the replies. Just to confirm that I got it right:
> 1. Since there is no setting to control index writers, is it fair to assume
> that Solr always indexes at maximum possible speed?
> 2. The way to control write speed is to control number of clients that are
> simultaneously posting data, right?
>
>
>
> --
> View this message in context: 
> http://lucene.472066.n3.nabble.com/Configuration-of-parallel-indexing-threads-tp4338466p4338599.html
> Sent from the Solr - User mailing list archive at Nabble.com.

Reply via email to