Hi,
I have removed a host (hvs004) that was in maintenance.
The system disk of this host had failed, so removed the host hvs004 in ceph;
replaced the system disk; erased all the osd-disks and reinstalled the host as
hvs005.
Resulting a cluster status in waring that doesn’t goes away:
health: HEALTH_WARN
1 host is in maintenance mode
Removal is done by “ceph orch host rm hvs004 --offline –force” in cephadm
shell.
How can I correct this false warning?
Some more info:
root@hvs001:/# ceph orch host ls
HOST ADDR LABELS STATUS
hvs001 xxx.xxx.xxx.xxx _admin
hvs002 xxx.xxx.xxx.xxx _admin
hvs003 xxx.xxx.xxx.xxx _admin
hvs005 xxx.xxx.xxx.xxx _admin
4 hosts in cluster
root@hvs001:/# ceph health detail
HEALTH_WARN 1 host is in maintenance mode
[WRN] HOST_IN_MAINTENANCE: 1 host is in maintenance mode
hvs004 is in maintenance :-/
Help is greatly apreciated…
_______________________________________________
ceph-users mailing list -- [email protected]
To unsubscribe send an email to [email protected]