Hello,
i seem to not have removed old osd. Now i have:
root@ceph07:/tmp# ceph orch ps |grep -e error -e stopped |grep ceph07
_osd.33 ceph07 stopped 2h ago 2y
<unknown> quay.io/ceph/ceph:v15.2.17 <unknown> <unknown>
mon.ceph01 ceph07 error 2h ago 2y
<unknown> quay.io/ceph/ceph:v15.2.17 <unknown> <unknown>
osd.0 ceph07 error 2h ago 2y
<unknown> quay.io/ceph/ceph:v15.2.17 <unknown> <unknown>
osd.1 ceph07 error 2h ago 2y
<unknown> quay.io/ceph/ceph:v15.2.17 <unknown> <unknown>
osd.11 ceph07 error 2h ago 2y
<unknown> quay.io/ceph/ceph:v15.2.17 <unknown> <unknown>
osd.12 ceph07 error 2h ago 2y
<unknown> quay.io/ceph/ceph:v15.2.17 <unknown> <unknown>
osd.14 ceph07 error 2h ago 2y
<unknown> quay.io/ceph/ceph:v15.2.17 <unknown> <unknown>
osd.18 ceph07 error 2h ago 2y
<unknown> quay.io/ceph/ceph:v15.2.17 <unknown> <unknown>
osd.22 ceph07 error 2h ago 2y
<unknown> quay.io/ceph/ceph:v15.2.17 <unknown> <unknown>
osd.30 ceph07 error 2h ago 2y
<unknown> quay.io/ceph/ceph:v15.2.17 <unknown> <unknown>
osd.4 ceph07 error 2h ago 2y
<unknown> quay.io/ceph/ceph:v15.2.17 <unknown> <unknown>
osd.64 ceph07 error 2h ago 2y
<unknown> quay.io/ceph/ceph:v15.2.17 <unknown> <unknown>
osd.8 ceph07 error 2h ago 2y
<unknown> quay.io/ceph/ceph:v15.2.17 <unknown> <unknown>
Which are non-existing Daemons on that node "ceph07". And i can not remove them:
cephadm rm-daemon --fsid=5436dd5d-83d4-4dc8-a93b-60ab5db145df
--name=osd.0 --force
cephadm rm-daemon --fsid=5436dd5d-83d4-4dc8-a93b-60ab5db145df
--name=osd.1 --force
cephadm rm-daemon --fsid=5436dd5d-83d4-4dc8-a93b-60ab5db145df
--name=osd.12 --force
cephadm rm-daemon --fsid=5436dd5d-83d4-4dc8-a93b-60ab5db145df
--name=osd.14 --force
cephadm rm-daemon --fsid=5436dd5d-83d4-4dc8-a93b-60ab5db145df
--name=osd.18 --force
cephadm rm-daemon --fsid=5436dd5d-83d4-4dc8-a93b-60ab5db145df
--name=osd.30 --force
cephadm rm-daemon --fsid=5436dd5d-83d4-4dc8-a93b-60ab5db145df
--name=osd.4 --force
cephadm rm-daemon --fsid=5436dd5d-83d4-4dc8-a93b-60ab5db145df
--name=osd.64 --force
cephadm rm-daemon --fsid=5436dd5d-83d4-4dc8-a93b-60ab5db145df
--name=osd.8 --force
root@ceph07:/tmp# ls /var/lib/ceph/5436dd5d-83d4-4dc8-a93b-60ab5db145df/
crash home osd.66 osd.67 osd.68 osd.69 osd.999 removed
=> thats correct.
root@ceph07:/tmp# ls /var/lib/ceph/5436dd5d-83d4-4dc8-a93b-60ab5db145df/removed/
mon.ceph01_2020-09-02T07:11:30.232540
mon.ceph07_2020-11-20T14:17:56.122749
osd.12_2022-12-29T13:17:47.855132 osd.22_2022-12-29T13:13:47.233379
osd.64_2022-12-29T13:17:50.732467 osd.73_2022-12-29T09:54:58.009039Z
mon.ceph01_2022-12-29T13:18:33.702553
osd.0_2022-12-29T13:17:46.661637
osd.14_2022-12-29T13:17:48.485548 osd.30_2022-12-29T13:17:49.685540
osd.70_2022-12-29T09:56:15.014346Z osd.74_2022-12-29T09:54:59.529058Z
mon.ceph02_2020-09-01T12:07:11.808391
osd.11_2022-12-29T13:15:39.944974
osd.18_2022-12-29T13:17:49.145034 osd.32_2020-07-30T09:44:23.252102
osd.71_2022-12-29T09:54:55.157744Z osd.75_2022-12-29T09:55:02.647709Z
mon.ceph03_2020-09-01T13:26:34.704724
osd.1_2022-12-29T13:17:47.233991
osd.20_2022-12-29T12:58:27.511277 osd.4_2022-12-29T13:17:50.199486
osd.72_2022-12-29T09:54:56.537846Z osd.8_2022-12-29T13:17:51.372638
my first try was to rename the old/non-active OSD from osd.33 to
_osd.33, but now i have a dangling module here:
root@ceph07:/tmp# ceph -s
cluster:
id: 5436dd5d-83d4-4dc8-a93b-60ab5db145df
health: HEALTH_ERR
mons are allowing insecure global_id reclaim
20 failed cephadm daemon(s)
Module 'cephadm' has failed: '_osd'
Any hints on how to clean up my node? :)
Cheers,
Mario
_______________________________________________
ceph-users mailing list -- [email protected]
To unsubscribe send an email to [email protected]