Sending batches in parallel is perfectly fine. _However_,
if you’re updating the same document, there’s no
guarantee which would win.
Imagine you have two processes sending batches. The
order of execution depends on way too many variables.
If nothing else, if process 1 sends a document then some
Hi!
On 2019-02-18T20:36:35, Erick Erickson wrote:
> Typically, people set their autocommit (hard) settings in
> solrconfig.xml and forget about it. I usually use a time-based trigger
> and don’t use documents as a trigger.
I added a timed autoCommit and it seems to work out nicely. Thank you!
> U
Typically, people set their autocommit (hard) settings in solrconfig.xml and
forget about it. I usually use a time-based trigger and don’t use documents as
a trigger.
If you were waiting until the end of your batch run (all 46M docs) to issue a
commit, that’s an anit-pattern. Until you do a har
Hello!
Another question I could not find an answer to:
is there a best-practice / recommendation for pushing several million
documents into a new index?
I'm currently splittig my documents into batches of 10,000 json-line
payloads into the update request handler, with commit set to 'true'
(yes, f