90+% of the time when I see this question it’s a problem with the client not 
being able to push docs at Solr fast enough. This is particularly true if 
databases are involved.

In addition to Jörn’s comment, I’d ask whether your Solr CPUs are running flat 
out. If your CPUs aren’t maxed out, you need to send more docs to Solr.

It looks like you have a single indexing Solrj client. The bottleneck might be 
there, i.e. at 30 threads the _client_ is acquiring docs as fast as it can, due 
to CPU utilization, disk I/O, whatever and Solr is just idling along.

Certainly Jörn’s comment about batching docs is germane. Here’s some 
quantification: https://lucidworks.com/2015/10/05/really-batch-updates-solr-2/

Best,
Erick

> On Jun 2, 2019, at 1:15 AM, Jörn Franke <jornfra...@gmail.com> wrote:
> 
> And to send multiple documents in one add document step
> 
>> Am 02.06.2019 um 08:45 schrieb calamita.agost...@libero.it.invalid:
>> 
>> 
>> 
>> Ho all,
>> I've an ingestion  application  that reads many files and indexes  records  
>> in a SolrCloud  7.4 with SolrJ using addDocument.
>> The application is  multithreading, every thread reads a file and send 
>> addDocument.
>> I see that from 10 to 30 threads executing addDocument in parallel, number 
>> of files indexed increase from 120 to 360 files per minute.
>> If  I use 50 threads, the number of files per minute remain the same.
>> Is there a parameter  to customize in Solr to increase parallelism relative 
>> to ingestion with addDocument http interface ? 
>> Thanks.
>> --
>> Inviato da Libero Mail per Android

Reply via email to