I think i was luky...
```sh
[root@ceph1 ~]# cephfs-journal-tool --rank=cephfs:0 event recover_dentries
summary
Events by type:
OPEN: 34407
PURGED: 2
SEGMENT: 125
SESSION: 15
SUBTREEMAP: 9
UPDATE: 75836
Errors: 0
```
Do i interpret this correctly, that `Errors: 0` means all journal events had
been able to be recovered? According to that output i didn't dig deeper for a
`cephfs-data-scan` and mds map reset.
After running the above procedure i followed with the following:
Mark as repaired:
```sh
ceph mds repaired cephfs:0
```
Allow clients:
```sh
ceph config rm mds mds_deny_all_reconnect
ceph fs set cephfs refuse_client_session false
```
Start mds scrub:
```sh
ceph tell mds.cephfs:0 scrub start / recursive
```
After that i got a error:
```
[ERR] MDS_DAMAGE: 1 MDSs report damaged metadata
mds.cephfs.ceph1.yzqmuo(mds.0): Metadata damage detected
```
Repair metadata:
```sh
ceph tell mds.cephfs:0 scrub start / recursive,repair,force
```
Now everything seems fine.
_______________________________________________
ceph-users mailing list -- [email protected]
To unsubscribe send an email to [email protected]