It’s the "Deleted Docs” metric in the statistic core.
I now that eventually the merges will expunge this deletes but I will run out of space soon and I want to know the _real_ space that I have. Actually I have space enough (about 3.5x the size of the index) to do the optimize. Other question that I have is if I can optimize one shard at a time instead of do an optimize over the full collection (this give me more control about space used, I have more than one shard of the same collection in each node of the cluster). — /Yago Riveiro On Fri, Nov 21, 2014 at 4:29 PM, Erick Erickson <erickerick...@gmail.com> wrote: > Yes, should be no problem. > Although this should be happening automatically, the percentage > of documents in a segment weighs quite heavily when the decision > is made to merge segments in the background. > You say you have "millions of deletes". Is this the difference between > numDocs and maxDoc on the admin page for the core in question? > Or is it just that you've issued millions of updates (or deletes)? Because > if the latter, I'd advise monitoring the numDocs/maxDoc pair to see > if the problem goes away on its own. > bq: ...and need free space > This is a red flag. If you're talking about disk space, before you get the > free space forceMerge will copy the _entire_ index so you'll need at > least 2x the current index size. > Best, > Erick > On Fri, Nov 21, 2014 at 6:40 AM, yriveiro <yago.rive...@gmail.com> wrote: >> Hi, >> >> It´s possible perform an optimize operation and continuing indexing over a >> collection? >> >> I need to force expunge deletes from the index I have millions os deletes >> and need free space. >> >> >> >> ----- >> Best regards >> -- >> View this message in context: >> http://lucene.472066.n3.nabble.com/Optimize-during-indexing-tp4170261.html >> Sent from the Solr - User mailing list archive at Nabble.com.