Hi,

I experienced this from time to time with older releases of ceph, but
haven't stumbled upon it for some time.

Often I had to revert to the older state by using:
http://ceph.com/docs/master/rados/operations/add-or-rm-mons/#removing-monitors-from-an-unhealthy-cluster

and dump the monlist, find the original monitor - kill the newest addition,
inject and restart it - then it should get online again.

Cheers,
Martin


On Wed, Mar 26, 2014 at 11:40 AM, <[email protected]> wrote:

>
> Hi,
>         I just add a new mon to a health cluster by following website
> manual "http://ceph.com/docs/master/rados/operations/add-or-rm-mons/"; "ADDING
> MONITORS" step by step,
>
> but when i execute step 6:
> ceph mon add <mon-id> <ip>[:<port>]
>
> the command didn't return, then i execute "ceph -s" on health mon node,
> this command didn't return either.
>
> so i try to restart mon to recover the whole cluster, but it seems never
> recover.
>
> Please anyone tell me how to deal with it?
>
>
> === mon.storage1 ===
> Starting Ceph mon.storage1 on storage1...
> Starting ceph-create-keys on storage1...
>
> [root@storage1 ~]# ceph -s   //after restart mon , "ceph -s" still have
> no output
>
>
>
>
> [root@storage1 ceph]# tail ceph-mon.storage1.log
> 2014-03-26 18:20:33.338554 7f60dbb967a0  0 ceph version 0.72.2
> (a913ded2ff138aefb8cb84d347d72164099cfd60), process ceph-mon, pid 24214
> 2014-03-26 18:20:33.460282 7f60dbb967a0  1 mon.storage1@-1(probing) e2
> preinit fsid 3429fd17-4a92-4d3b-a7fa-04adedb0da82
> 2014-03-26 18:20:33.460694 7f60dbb967a0  1 mon.storage1@-1(probing).pg v0
> on_upgrade discarding in-core PGMap
> 2014-03-26 18:20:33.487899 7f60dbb967a0  0 mon.storage1@-1(probing) e2
>  my rank is now 0 (was -1)
> 2014-03-26 18:20:33.488575 7f60d6854700  0 -- 193.168.1.100:6789/0 >>
> 193.168.1.133:6789/0 pipe(0x3f38280 sd=21 :0 s=1 pgs=0 cs=0 l=0
> c=0x3f19600).fault
> 2014-03-26 18:21:33.487686 7f60d8657700  0 
> mon.storage1@0(probing).data_health(0)
> update_stats avail 86% total 51606140 used 4324004 avail 44660696
> 2014-03-26 18:22:33.488091 7f60d8657700  0 
> mon.storage1@0(probing).data_health(0)
> update_stats avail 86% total 51606140 used 4324004 avail 44660696
> 2014-03-26 18:23:33.488500 7f60d8657700  0 
> mon.storage1@0(probing).data_health(0)
> update_stats avail 86% total 51606140 used 4324004 avail 44660696
>
>
> --------------------------------------------------------
> ZTE Information Security Notice: The information contained in this mail (and 
> any attachment transmitted herewith) is privileged and confidential and is 
> intended for the exclusive use of the addressee(s).  If you are not an 
> intended recipient, any disclosure, reproduction, distribution or other 
> dissemination or use of the information contained is strictly prohibited.  If 
> you have received this mail in error, please delete it and notify us 
> immediately.
>
>
>
>
> _______________________________________________
> ceph-users mailing list
> [email protected]
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
>
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to