Hi Mark, So I tried this: started up one instance w/ zookeeper, and started a second instance defining a shard name in solr.xml -- it worked, searching would search both indices, and looking at the zookeeper ui, I'd see the second shard. However, when I brought the second server down -- the first one stopped working: it didn't kick the second shard out of the cluster.
Any way to do this? Thanks, Ranjan > From: Mark Miller <markrmil...@gmail.com> > To: solr-user@lucene.apache.org > Cc: > Date: Wed, 29 Feb 2012 22:57:26 -0500 > Subject: Re: Building a resilient cluster > Doh! Sorry - this was broken - I need to fix the doc or add it back. > > The shard id is actually set in solr.xml since its per core - the sys prop > was a sugar option we had setup. So either add 'shard' to the core in > solr.xml, or to make it work like it does in the doc, do: > > <core name="collection1" shard="${shard:}" instanceDir="." /> > > That sets shard to the 'shard' system property if its set, or as a default, > act as if it wasn't set. > > I've been working with custom shard ids mainly through solrj, so I hadn't > noticed this. > > - Mark > > On Wed, Feb 29, 2012 at 10:36 AM, Ranjan Bagchi <ranjan.bag...@gmail.com > >wrote: > > > Hi, > > > > At this point I'm ok with one zk instance being a point of failure, I > just > > want to create sharded solr instances, bring them into the cluster, and > be > > able to shut them down without bringing down the whole cluster. > > > > According to the wiki page, I should be able to bring up new shard by > using > > shardId [-D shardId], but when I did that, the logs showed it replicating > > an existing shard. > > > > Ranjan > > Andre Bois-Crettez wrote: > > > > > You have to run ZK on a at least 3 different machines for fault > > > tolerance (a ZK ensemble). > > > > > > > > > http://wiki.apache.org/solr/SolrCloud#Example_C:_Two_shard_cluster_with_sha= > > > rd_replicas_and_zookeeper_ensemble > > > > > > Ranjan Bagchi wrote: > > > > Hi, > > > > > > > > I'm interested in setting up a solr cluster where each machine [at > > least > > > > initially] hosts a separate shard of a big index [too big to sit on > the > > > > machine]. I'm able to put a cloud together by telling it that I have > > (to > > > > start out with) 4 nodes, and then starting up nodes on 3 machines > > > pointin= > > > g > > > > at the zkInstance. I'm able to load my sharded data onto each > machine > > > > individually and it seems to work. > > > > > > > > My concern is that it's not fault tolerant: if one of the > > non-zookeeper > > > > machines falls over, the whole cluster won't work. Also, I can't > > create > > > = > > > a > > > > shard with more data, and have it work within the existing cloud. > > > > > > > > I tried using -DshardId=3Dshard5 [on an existing 4-shard cluster], > but > > it > > > > just started replicating, which doesn't seem right. > > > > > > > > Are there ways around this? > > > > > > > > Thanks, > > > > Ranjan Bagchi > > > > > > > > > > > > > > -- > - Mark > > http://www.lucidimagination.com > >