Personally I see no benefit to have more than one JVM per node, cores can handle it. I would say that splitting a 20m index into 25 shards strikes me as serious overkill, unless you expect to expand significantly. 20m would likely be okay with two or three shards. You can store the indexes for each core on different disks which can give ome performance benefit.
Just some thoughts. Upayavira On Thu, Nov 15, 2012, at 11:04 PM, Buttler, David wrote: > Hi, > I have a question about the optimal way to distribute solr indexes across > a cloud. I have a small number of collections (less than 10). And a > small cluster (6 nodes), but each node has several disks - 5 of which I > am using for my solr indexes. The cluster is also a hadoop cluster, so > the disks are not RAIDed, they are JBOD. So, on my 5 slave nodes, each > with 5 disks, I was thinking of putting one shard per collection. This > means I end up with 25 shards per collection. If I had 10 collections, > that would make it 250 shards total. Given that Solr 4 supports > multi-core, my first thought was to try one JVM for each node: for 10 > collections per node, that means that each JVM would contain 50 shards. > > So, I set up my first collection, with a modest 20M documents, and > everything seems to work fine. But, now my subsequent collections that I > have added are having issues. The first one is that every time I query > for the document count (*:* with rows=0), a different number of documents > is returned. The number can differ by as much as 10%. Now if I query > each shard individually (setting distrib=false), the number returned is > always consistent. > > I am not entirely sure this is related as I may have missed a step in my > setup of subsequent collections (bootstrapping the config) > > But, more related to the architecture question: is it better to have one > JVM per disk, one JVM per shard, or one JVM per node. Given the MMap of > the indexes, how does memory play into the question? There is a blog > post > (http://blog.thetaphi.de/2012/07/use-lucenes-mmapdirectory-on-64bit.html) > that recommends minimizing the amount of JVM memory and maximizing the > amount of OS-level file cache, but how does that impact sorting / > boosting? > > Sorry if I have missed some documentation: I have been through the cloud > tutorial a couple of times, and I didn't see any discussion of these > issues > > Thanks, > Dave