Can you also let me know what version of Solr are you on?
On Sat, Jul 27, 2013 at 8:26 AM, Anshum Gupta <ans...@anshumgupta.net>wrote: > Hi Katie, > > 1. First things first, I would strongly advice to manually update/remove > zk or any other info when you're running things in the SolrCloud mode > unless you are sure of what you're doing. > > 2. Also, your node could be currently recovering from the transaction > log(did you issue a hard commit after indexing?). > The mailing list doesn't allow long texts inline so it'd be good if you > could use something like http://pastebin.com/ to share the log in detail. > > 3. If you had replicas, you wouldn't need to manually switch. It get's > taken care of automatically. > > > On Sat, Jul 27, 2013 at 4:16 AM, Katie McCorkell <katiemccork...@gmail.com > > wrote: > >> Hello, >> >> I am using the SolrCloud with a zookeeper ensemble like on example C from >> the wiki except with total of 3 shards and no replicas (oops). After >> indexing a whole bunch of documents, shard 2 went down and I'm not sure >> why. I tried restarting it with the jar command and I tried deleting >> shard1 >> 's zoo_data folder and then restarting but it is still down, and I'm not >> sure what to do. >> >> 1) Is there anyway to avoid reindexing all the data? It's no good to >> proceed without shard 2 because I don't know which documents are there vs. >> the other shards, and indexing and querying don't work when one shard is >> down. >> >> I can't exactly tell why restarting it is failing, all I can see is on the >> admin tool webpage the shard is yellow in the little cloud diagram. On the >> console is messages that I will copy and paste below. 2) How can I tell >> the >> exact problem? >> >> 3) If I had had replicas, I could have just switched to shard 2's replica >> at this point, correct? >> >> Thanks! >> Katie >> >> Console message from start.jar >> >> ----------------------------------------------------------------------------------------------------------------------- >> 2325 [coreLoadExecutor-4-thread-1] INFO >> org.apache.solr.cloud.ZkController >> – We are http://172.16.2.182:5555/solr/collection1/ and leader is >> http://172.16.2.182:5555/solr/collection1/ >> 12329 [recoveryExecutor-6-thread-1] WARN org.apache.solr.update.UpdateLog >> – Starting log replay >> >> tlog{file=/opt/solr-4.3.1/example/solr/collection1/data/tlog/tlog.0000000000000005179 >> refcount=2} active=false starting pos=0 >> 12534 [recoveryExecutor-6-thread-1] INFO org.apache.solr.core.SolrCore – >> SolrDeletionPolicy.onInit: commits:num=1 >> >> commit{dir=NRTCachingDirectory(org.apache.lucene.store.MMapDirectory@ >> /opt/solr-4.3.1/example/solr/collection1/data/index >> lockFactory=org.apache.lucene.store.NativeFSLockFactory@5f99ea3c; >> maxCacheMB=48.0 >> >> maxMergeSizeMB=4.0),segFN=segments_404,generation=5188,filenames=[_1gqo.fdx, >> _1h1q.nvm, _1h8x.fdt, _1gmi_Lucene41_0.pos, _1gqo.fdt, _1h8s.nvd, _ >> 1gmi.si, >> _1h1q.nvd, _1h6l.fnm, _1h8q.nvm, _1h6l_Lucene41_0.tim, >> _1h6l_Lucene41_0.tip, _1h8o_Lucene41_0.tim, _1h8o_Lucene41_0.tip, >> _1aq9_67.del, _1gqo.nvm, _1aq9_Lucene41_0.pos, _1h8q.fdx, _1h1q.fdt, >> _1h8r.fdt, _1h8q.fdt, _1h8p_Lucene41_0.pos, _1h8s_Lucene41_0.pos, >> _1h8r.fdx, _1gqo.nvd, _1h8s.fdx, _1h8s.fdt, _1h8x_Lucene41_..... >> > > > > -- > > Anshum Gupta > http://www.anshumgupta.net > -- Anshum Gupta http://www.anshumgupta.net