Hi!

On Wed, Feb 29, 2012 at 22:21, Emmanuel Espina <espinaemman...@gmail.com> wrote:
> No. But probably we can find another way to do what you want. Please
> describe the problem and include some "numbers" to give us an idea of
> the sizes that you are handling. Number of documents, size of the
> index, etc.

Thank you! Our Solr holds currently about 168Mio documents. From each
of these documents we extract the most important keywords and store
them in a multivalued field (topic). Our goal is to provide faceted
navigation through these topics. The topic field holds roughly 5
values per doc, but I wasn't able to compute the correct number right
now.

The use cases require that the facets have to be calculated fast
enough so that they can be answered in reasonable time (1-2 secs)
which we were able to do with a 192GB RAM machine and regular warming.

Splitting the Solr into a few smaller ones (even on the same machine)
seems to be the most promising way, but I've been shying away from it
for some reasons: higher complexity, a huge reimport (though, I could
split the current index), some components didn't support it when we
were starting (Grouping was only introduced with 3.5 IIRC). I've
tested the sharding approach and it was a bit slower than the one huge
index approach.

I'd be happy to hear some suggestions,
Michael

Reply via email to