Re: SolrCloud graph status is out of date
thanks Mark. will further dig into the logs. there is another problem related. we have collections with 3 shards (2 nodes in one shard), the collection have about 1000 records in it. but unfortunately that after the leader is down, replica node failed to become the leader.the detail is : after the leader node is down, replica node try to become the new leader, but it said === ShardLeaderElectionContext.runLeaderProcess(131) - Running the leader process. ShardLeaderElectionContext.shouldIBeLeader(331) - Checking if I should try and be the leader. ShardLeaderElectionContext.shouldIBeLeader(339) - My last published State was Active, it's okay to be the leader. ShardLeaderElectionContext.runLeaderProcess(164) - I may be the new leader - try and sync SyncStrategy.sync(89) - Sync replicas to http://localhost:8486/solr/exception/ PeerSync.sync(182) - PeerSync: core=exception url=http://localhost:8486/solr START replicas=[http://localhost:8483/solr/exception/] nUpdates=100 PeerSync.sync(250) - PeerSync: core=exception url=http://localhost:8486/solr DONE. We have no versions. sync failed. SyncStrategy.log(114) - Sync Failed ShardLeaderElectionContext.rejoinLeaderElection(311) - There is a better leader candidate than us - going back into recovery DefaultSolrCoreState.doRecovery(214) - Running recovery - first canceling any ongoing recovery after that, it try to recovery from the leader node, which is already down. then recovery + failed + recovery. is it related to SOLR-3939 and SOLR-3940? but the index data isn't empty. On Thu, Jan 10, 2013 at 10:09 AM, Mark Miller wrote: > It may be able to do that because it's forwarding requests to other nodes > that are up? > > Would be good to dig into the logs to see if you can narrow in on the > reason for the recovery_failed. > > - Mark > > On Jan 9, 2013, at 8:52 PM, Zeng Lames wrote: > > > Hi , > > > > we meet below strange case in production environment. from the Solr Admin > > Console -> Cloud -> Graph, we can find that one node is in > recovery_failed > > status. but at the same time, we found that the recovery_failed node can > > server query/update request normally. > > > > any idea about it? thanks! > > > > -- > > Best Wishes! > > Lames > > -- Best Wishes! Lames
Re: SolrCloud graph status is out of date
thanks Mark. may I know the target release date of 4.1? On Thu, Jan 10, 2013 at 10:13 PM, Mark Miller wrote: > It may still be related. Even a non empty index can have no versions (eg > one that was just replicated). Should behave better in this case in 4.1. > > - Mark > > On Jan 10, 2013, at 12:41 AM, Zeng Lames wrote: > > > thanks Mark. will further dig into the logs. there is another problem > > related. > > > > we have collections with 3 shards (2 nodes in one shard), the collection > > have about 1000 records in it. but unfortunately that after the leader is > > down, replica node failed to become the leader.the detail is : after the > > leader node is down, replica node try to become the new leader, but it > said > > > > === > > ShardLeaderElectionContext.runLeaderProcess(131) - Running the leader > > process. > > ShardLeaderElectionContext.shouldIBeLeader(331) - Checking if I should > try > > and be the leader. > > ShardLeaderElectionContext.shouldIBeLeader(339) - My last published State > > was Active, it's okay to be the leader. > > ShardLeaderElectionContext.runLeaderProcess(164) - I may be the new > leader > > - try and sync > > SyncStrategy.sync(89) - Sync replicas to > > http://localhost:8486/solr/exception/ > > PeerSync.sync(182) - PeerSync: core=exception > > url=http://localhost:8486/solr START > > replicas=[http://localhost:8483/solr/exception/] nUpdates=100 > > PeerSync.sync(250) - PeerSync: core=exception > > url=http://localhost:8486/solr DONE. > > We have no versions. sync failed. > > SyncStrategy.log(114) - Sync Failed > > ShardLeaderElectionContext.rejoinLeaderElection(311) - There is a better > > leader candidate than us - going back into recovery > > DefaultSolrCoreState.doRecovery(214) - Running recovery - first canceling > > any ongoing recovery > > > > > > after that, it try to recovery from the leader node, which is already > down. > > then recovery + failed + recovery. > > > > is it related to SOLR-3939 and SOLR-3940? but the index data isn't empty. > > > > > > On Thu, Jan 10, 2013 at 10:09 AM, Mark Miller > wrote: > > > >> It may be able to do that because it's forwarding requests to other > nodes > >> that are up? > >> > >> Would be good to dig into the logs to see if you can narrow in on the > >> reason for the recovery_failed. > >> > >> - Mark > >> > >> On Jan 9, 2013, at 8:52 PM, Zeng Lames wrote: > >> > >>> Hi , > >>> > >>> we meet below strange case in production environment. from the Solr > Admin > >>> Console -> Cloud -> Graph, we can find that one node is in > >> recovery_failed > >>> status. but at the same time, we found that the recovery_failed node > can > >>> server query/update request normally. > >>> > >>> any idea about it? thanks! > >>> > >>> -- > >>> Best Wishes! > >>> Lames > >> > >> > > > > > > -- > > Best Wishes! > > Lames > > -- Best Wishes! Lames
Re: SolrCloud graph status is out of date
thanks Mark. looking forward it On Fri, Jan 11, 2013 at 9:28 AM, Mark Miller wrote: > Looks like we are talking about making a release candidate next week. > > Mark > > Sent from my iPhone > > On Jan 10, 2013, at 7:50 PM, Zeng Lames wrote: > > > thanks Mark. may I know the target release date of 4.1? > > > > > > On Thu, Jan 10, 2013 at 10:13 PM, Mark Miller > wrote: > > > >> It may still be related. Even a non empty index can have no versions (eg > >> one that was just replicated). Should behave better in this case in 4.1. > >> > >> - Mark > >> > >> On Jan 10, 2013, at 12:41 AM, Zeng Lames wrote: > >> > >>> thanks Mark. will further dig into the logs. there is another problem > >>> related. > >>> > >>> we have collections with 3 shards (2 nodes in one shard), the > collection > >>> have about 1000 records in it. but unfortunately that after the leader > is > >>> down, replica node failed to become the leader.the detail is : after > the > >>> leader node is down, replica node try to become the new leader, but it > >> said > >>> > >>> === > >>> ShardLeaderElectionContext.runLeaderProcess(131) - Running the leader > >>> process. > >>> ShardLeaderElectionContext.shouldIBeLeader(331) - Checking if I should > >> try > >>> and be the leader. > >>> ShardLeaderElectionContext.shouldIBeLeader(339) - My last published > State > >>> was Active, it's okay to be the leader. > >>> ShardLeaderElectionContext.runLeaderProcess(164) - I may be the new > >> leader > >>> - try and sync > >>> SyncStrategy.sync(89) - Sync replicas to > >>> http://localhost:8486/solr/exception/ > >>> PeerSync.sync(182) - PeerSync: core=exception > >>> url=http://localhost:8486/solr START > >>> replicas=[http://localhost:8483/solr/exception/] nUpdates=100 > >>> PeerSync.sync(250) - PeerSync: core=exception > >>> url=http://localhost:8486/solr DONE. > >>> We have no versions. sync failed. > >>> SyncStrategy.log(114) - Sync Failed > >>> ShardLeaderElectionContext.rejoinLeaderElection(311) - There is a > better > >>> leader candidate than us - going back into recovery > >>> DefaultSolrCoreState.doRecovery(214) - Running recovery - first > canceling > >>> any ongoing recovery > >>> > >>> > >>> after that, it try to recovery from the leader node, which is already > >> down. > >>> then recovery + failed + recovery. > >>> > >>> is it related to SOLR-3939 and SOLR-3940? but the index data isn't > empty. > >>> > >>> > >>> On Thu, Jan 10, 2013 at 10:09 AM, Mark Miller > >> wrote: > >>> > >>>> It may be able to do that because it's forwarding requests to other > >> nodes > >>>> that are up? > >>>> > >>>> Would be good to dig into the logs to see if you can narrow in on the > >>>> reason for the recovery_failed. > >>>> > >>>> - Mark > >>>> > >>>> On Jan 9, 2013, at 8:52 PM, Zeng Lames wrote: > >>>> > >>>>> Hi , > >>>>> > >>>>> we meet below strange case in production environment. from the Solr > >> Admin > >>>>> Console -> Cloud -> Graph, we can find that one node is in > >>>> recovery_failed > >>>>> status. but at the same time, we found that the recovery_failed node > >> can > >>>>> server query/update request normally. > >>>>> > >>>>> any idea about it? thanks! > >>>>> > >>>>> -- > >>>>> Best Wishes! > >>>>> Lames > >>> > >>> > >>> -- > >>> Best Wishes! > >>> Lames > > > > > > -- > > Best Wishes! > > Lames > -- Best Wishes! Lames
Re: Can we retrieve deleted records before optimized
Thanks kan for your prompt help. It is really a great solution to recovery those deleted records. Another question is about Solr history data housekeep problem. the scenario is as below: we have a solr core to store biz records, which is large volume that the index files is more than 50GB in one month. Due to disk space limitation, we need to delete one month ago records from solr. but in the other hand, we need to keep those data in the cheaper disk for analysis. the problem is how to keep those data one month ago into cheaper disk in quickly solution. one simple but so slow solution is that we search out those records and add into the solr in the cheaper disk. wanna to know is there any other solution for such kind of problem. e.g. move the index files directly? thanks a lot! On Wed, Oct 17, 2012 at 12:31 AM, Dmitry Kan wrote: > Hello, > > One approach (not a solrish one, but still) would be to use Lucene API and > set up an IndexReader onto the solr index in question. You can then do: > > [code] > Directory indexDir = FSDirectory.open(new File(pathToDir)); > IndexReader input = IndexReader.open(indexDir, true); > > FieldSelector fieldSelector = new SetBasedFieldSelector( > null, // to retrive all stored fields > Collections.emptySet()); > > int maxDoc = input.maxDoc(); > for (int i = 0; i < maxDoc; i++) { > if (input.isDeleted(i)) { > // deleted document found, retrieve it > Document document = input.document(i, fieldSelector); > // analyze its field values here... > } > } > [/code] > > I haven't compiled this code myself, you'll need to experiment with it. > > Dmitry > > On Tue, Oct 16, 2012 at 11:06 AM, Zeng Lames wrote: > > > Hi, > > > > as we know, when we delete document from solr, it will add .del file to > > related segment index files, then delete them from disk after optimize. > > Now, the question is that before optimized, can we retrieve those deleted > > records? if yes, how to? > > > > thanks a lot! > > > > Best Wishes > > Lames > > > > > > -- > Regards, > > Dmitry Kan >
How to migrate index from 3.6 to 4.0 with solrcloud
Dear all, we have an existing index with solr 3.6, and now we want to migrate it to Solr4.0 with shard (2 shards, 2 nodes in a shard). The question are: 1.which node should i copy the existing index files to? any node in any shard ? 2. if copy the index files into any one of nodes, can it be replicated to the 'right' shard according hash code? 3. if above steps can't fulfill a index migrate to solrcloud, how should we do? thanks a lot Lames
Re: How to migrate index from 3.6 to 4.0 with solrcloud
thanks all for your prompt response. i think i know what should i do now. thank you so much again. On Fri, Nov 2, 2012 at 2:36 PM, Erick Erickson wrote: > It's not clear whether your index is already sharded or not. But it doesn't > matter because: > 1> if it's not sharded, there's no shard-splitter (yet). So you can't copy > the right parts of your single index to the right shard. > 2> if your 3.6 index _is_ sharded already, I pretty much guarantee that it > wasn't created with the same hashing algorithm that SolrCloud uses, so just > copying the shards to some node on the cloud won't work. > > In either case, you'll have to re-indexing everything fresh. > > Best > Erick > > > On Thu, Nov 1, 2012 at 9:34 PM, Zeng Lames wrote: > > > Dear all, > > > > we have an existing index with solr 3.6, and now we want to migrate it to > > Solr4.0 with shard (2 shards, 2 nodes in a shard). The question are: > > > > 1.which node should i copy the existing index files to? any node in any > > shard ? > > > > 2. if copy the index files into any one of nodes, can it be replicated to > > the 'right' shard according hash code? > > > > 3. if above steps can't fulfill a index migrate to solrcloud, how should > we > > do? > > > > thanks a lot > > Lames > > >
Re: Add new shard will be treated as replicas in Solr4.0?
hi Erick, thanks for your kindly response. hv got the information from the SolrCloud wiki. think we may need to defined the shard numbers when we really rollout it. thanks again On Mon, Nov 5, 2012 at 8:40 PM, Erick Erickson wrote: > Not at present. What you're interested in is "shard splitting" which is > certainly on the roadmap but not implemented yet. To expand the > number of shards you'll have to reconfigure, then re-index. > > Best > Erick > > > On Mon, Nov 5, 2012 at 4:09 AM, Zeng Lames wrote: > > > Dear All, > > > > we have an existing solr collection, 2 shards, numOfShard is 2. and there > > are already records in the index files. now we start another solr > instance > > with ShardId= shard3, and found that Solr treat it as replicas. > > > > check the zookeeper data, found the range of shard doesn't > > change correspondingly. shard 1 is 0-7fff, while shard 2 is > > 8000-. > > > > is there any way to increase new shard for existing collection? > > > > thanks a lot! > > Lames > > >
Re: Add new shard will be treated as replicas in Solr4.0?
btw, where can i find all the items in the road map? thanks! On Tue, Nov 6, 2012 at 8:55 AM, Zeng Lames wrote: > hi Erick, thanks for your kindly response. hv got the information from the > SolrCloud wiki. > think we may need to defined the shard numbers when we really rollout it. > > thanks again > > > On Mon, Nov 5, 2012 at 8:40 PM, Erick Erickson wrote: > >> Not at present. What you're interested in is "shard splitting" which is >> certainly on the roadmap but not implemented yet. To expand the >> number of shards you'll have to reconfigure, then re-index. >> >> Best >> Erick >> >> >> On Mon, Nov 5, 2012 at 4:09 AM, Zeng Lames wrote: >> >> > Dear All, >> > >> > we have an existing solr collection, 2 shards, numOfShard is 2. and >> there >> > are already records in the index files. now we start another solr >> instance >> > with ShardId= shard3, and found that Solr treat it as replicas. >> > >> > check the zookeeper data, found the range of shard doesn't >> > change correspondingly. shard 1 is 0-7fff, while shard 2 is >> > 8000-. >> > >> > is there any way to increase new shard for existing collection? >> > >> > thanks a lot! >> > Lames >> > >> > >
Re: Add new shard will be treated as replicas in Solr4.0?
got it. thanks a lot On Tue, Nov 6, 2012 at 8:43 PM, Erick Erickson wrote: > bq: where can i find all the items on the road map? > > Well, you really can't ... There's no "official" roadmap. I happen to > know this since I follow the developer's list and I've seen references to > this being important to the folks doing SolrCloud development work and it's > been a recurring theme on the user's list. It's one of those things that > _everybody_ understands would be useful in certain circumstances, but > haven't had time to actually implement yet. > > You can track this at: https://issues.apache.org/jira/browse/SOLR-2592 > > Best > Erick > > > > On Mon, Nov 5, 2012 at 7:57 PM, Zeng Lames wrote: > > > btw, where can i find all the items in the road map? thanks! > > > > > > On Tue, Nov 6, 2012 at 8:55 AM, Zeng Lames wrote: > > > > > hi Erick, thanks for your kindly response. hv got the information from > > the > > > SolrCloud wiki. > > > think we may need to defined the shard numbers when we really rollout > it. > > > > > > thanks again > > > > > > > > > On Mon, Nov 5, 2012 at 8:40 PM, Erick Erickson < > erickerick...@gmail.com > > >wrote: > > > > > >> Not at present. What you're interested in is "shard splitting" which > is > > >> certainly on the roadmap but not implemented yet. To expand the > > >> number of shards you'll have to reconfigure, then re-index. > > >> > > >> Best > > >> Erick > > >> > > >> > > >> On Mon, Nov 5, 2012 at 4:09 AM, Zeng Lames > > wrote: > > >> > > >> > Dear All, > > >> > > > >> > we have an existing solr collection, 2 shards, numOfShard is 2. and > > >> there > > >> > are already records in the index files. now we start another solr > > >> instance > > >> > with ShardId= shard3, and found that Solr treat it as replicas. > > >> > > > >> > check the zookeeper data, found the range of shard doesn't > > >> > change correspondingly. shard 1 is 0-7fff, while shard 2 is > > >> > 8000-. > > >> > > > >> > is there any way to increase new shard for existing collection? > > >> > > > >> > thanks a lot! > > >> > Lames > > >> > > > >> > > > > > > > > >