Any document that has the same value in the "id" field (or whatever you've defined in <uniqueKey> in your schema) will replace any older documents with the same value. So my guess is that your data has some duplicate keys.
A simple way to check is to watch maxDoc .vs. numDocs in the admin UI for a particular replica. If those numbers diverge as your indexing job is running, then you're overwriting documents. Do note that background merging will purge deleted docs from the segments being merged, so it's not enough to just look once. Best, Erick On Fri, Apr 27, 2018 at 5:22 AM, LOPEZ-CORTES Mariano-ext <mariano.lopez-cortes-...@pole-emploi.fr> wrote: > Hi > > We've finished the data import of 40 millions data into a 3 node Solr cluster. > > After injecting all data via a Java program, we've noticed that the number of > documents was less than expected (in 100000 rows). > No exception, no error. > > Some config details: > > <autoCommit> > <maxTime>15000</maxTime> > > <openSearcher>false</openSearcher> > </autoCommit> > <autoSoftCommit> > <maxTime>15000</maxTime> > </autoSoftCommit> > > We have no commits in the client application. > > But also, when consulting via admin, we've noticed that the number total of > rows in Solr increase slowly (numFound). > > It's a normal behaviour? What's the problem? > > Thanks! > > > >