On 10-Dec-07, at 12:50 PM, Doug T wrote:
I have been using parallelmultisearches on multi-CPU machines, and
seen
sizable benefit over a single large index (even if all of the
fragments are
on 1 disk). Is there a way to quickly enable this on a solr
server? Or do
I need to go into the so
in context:
http://www.nabble.com/distributing-indexes-via-solr-tp4112411p14261467.html
Sent from the Solr - User mailing list archive at Nabble.com.
Each indexed document will represent an email, consisting of the typical
fields to/from/subject/cc/bcc/body/attachment/mailheaders where the body and
attachment texts will be indexed and tokenized but not stored. It's
difficult to give an estimate of the # of such documents, other than to say
that
If you are after faster disks, it might just be easier to use RAID.
If you want real scalability with a single-index view, you want
multiple machines (which Solr doesn't support yet).
If you can partition your data such that queries can be run against
single partitions, then use separate Solr serv
So the thinking here was to divide the total indexed data among N partitions
since the amount of data will be massive. Each partition would probably be
using a separate physical disk(s), and then for searching I could use
ParallelMultiSearcher to dispatch searches to each of these partitions as a
: Suppose I want the xml input submitted to solr to be distributed among a
: fixed set of partitions; basically, something like round-robin among each of
: them, so that each directory has a relatively equal size in terms of # of
: segments. Is there an easy way to do this? I took a quick look a
Hi,
Suppose I want the xml input submitted to solr to be distributed among a
fixed set of partitions; basically, something like round-robin among each of
them, so that each directory has a relatively equal size in terms of # of
segments. Is there an easy way to do this? I took a quick look at th