I don't kill the solr instance forcefully using "kill -9". I checked the core.properties file for that shard. The content is different from the core.properties file for all the other shards. It has the following two lines which are different
config=solrconfig.xml schema=schema.xml In other shards, it is collection.configName=v4 (name I have given to the config) name=collectionName_shardNumber_replica1 Should I modify this file before restarting the Cloud? There is a strange thing I just observed about the data dir of the shard that is not coming up. There is an addition index dir that has been created hdfs://Ingest/solr53/collection/core_node32/data/index/index/ The size and content is same as of hdfs://Ingest/solr53/collection/core_node32/data/index/ What could be the reason of this extra dir? Should I delete it? Thanks! On Wed, Nov 16, 2016 at 1:51 PM, Erick Erickson <erickerick...@gmail.com> wrote: > bq: Before restarting, I delete all the write.lock files from the data > dir. But > every time I restart I get the same exception. > > First, this shouldn't be necessary. Are you by any chance killing the > Solr instances with > the equivalent of "kill -9"? Allow them to shut down gracefully. That > said, until recently > the bin/solr script would kill them forcefully after 5 seconds which > is too short an interval. > > But the error really is telling you that somehow two or more Solr > cores are pointing at the > same data directory. Whichever one gets there first will block any > later cores with the > message you see. So check your core.properties files and your HDFS magic > to see > how this is occurring would be my first guess. > > Best, > Erick > > On Wed, Nov 16, 2016 at 1:38 PM, Chetas Joshi <chetas.jo...@gmail.com> > wrote: > > Hi, > > > > I have a SolrCloud (on HDFS) of 52 nodes. I have 3 collections each with > 50 > > shards and maxShards per node for every collection is 1. > > > > I am having problem restarting a solr shard for a collection. > > > > When I restart, there is always a particular shard of a particular > > collection that remains down. The 2 shards on the same host for the rest > of > > the collections are up and running. > > > > Before restarting, I delete all the write.lock files from the data dir. > But > > every time I restart I get the same exception. > > > > index dir yyy of core xxx is already locked. The most likely cause is > > another Solr server (or another solr core in this server) also configured > > to use this directory; other possible causes may be specific to lockType: > > hdfs > > > > Thanks! >