I have a situation in which I can't safely optimize, the index is on a
machine that doesn't have enough disk space. I have no control over the
hardware. There are some raw statistics on this index that don't make
sense, it's roughly twice the size of a similar index (could be legit, but
it seems out of whack). And the number of deleted documents isn't big
enough to account for the discrepancy.

So is there a good way, without optimizing, to purge any segments not
referenced in the segments file? Actually I doubt that optimizing would
even do it if I _could_, any phantom segments aren't visible from the
segments file anyway...

I can generate a list of active segments using CheckIndex and compare that
against the contents of the index directory right enough, and remove the
segments that aren't listed, feels feels awkward and error-prone though (of
course the solr server should be killed while doing _any_ of this).

Hmmm, I suppose I could replicate the index to another fresh Solr
instance.....

Thanks,
Erick

Reply via email to