Mutliple threads to the same index ? And how many concurrent threads?

Our case is not merely multiple threads but actually large scale spark indexer 
jobs that index 1B records at a time with a concurrency of 400.
In this case multiple such jobs were indexing into the same index. 


> On Apr 2, 2019, at 7:25 AM, Walter Underwood <wun...@wunderwood.org> wrote:
> 
> We run multiple threads indexing to Solr all the time and have been doing so 
> for years.
> 
> How big are your documents and how big are your batches?
> 
> wunder
> Walter Underwood
> wun...@wunderwood.org
> http://observer.wunderwood.org/  (my blog)
> 
>> On Apr 1, 2019, at 10:51 PM, Aroop Ganguly <aroopgang...@icloud.com> wrote:
>> 
>> Turns out the cause was multiple indexing jobs indexing into the index 
>> simultaneously, which one can imagine can cause jvm loads on certain 
>> replicas for sure.
>> Once this was found and only one job ran at a time, things were back to 
>> normal.
>> 
>> Your comments seem right on no correlation to the stack trace! 
>> 
>>> On Apr 1, 2019, at 5:32 PM, Shawn Heisey <apa...@elyograg.org> wrote:
>>> 
>>> 4/1/2019 5:40 PM, Aroop Ganguly wrote:
>>>> Thanks Shawn, for the initial response.
>>>> Digging into a bit, I was wondering if we’d care to read the inner most 
>>>> stack.
>>>> From the inner most stack it seems to be telling us something about what 
>>>> trigger it ?
>>>> Ofcourse, the system could have been overloaded as well, but is the 
>>>> exception telling us something or its of no use to consider this stack
>>> 
>>> The stacktrace on OOME is rarely useful.  The memory allocation where the 
>>> error is thrown probably has absolutely no connection to the part of the 
>>> program where major amounts of memory are being used.  It could be ANY 
>>> memory allocation that actually causes the error.
>>> 
>>> Thanks,
>>> Shawn
>> 
> 

Reply via email to