I am working to develop some monitoring for our File clusters and as part of
the check I inspect `ceph mds stat` for damaged,failed,stopped MDS/Ranks.
Initially I set my check to Alarm if any of these states was discovered but as
I distributed it out I noticed that one of our clusters had the following:
"failed": [],
"damaged": [],
"stopped": [
2
],
However the cluster health is good and the mds state is: cephfs-2/2/2 up
{0=p3plcephmds001=up:active,1=p3plcephmds002=up:active}, 1 up:standby
A little further digging and I found that a stopped state doesnt apply to an
MDS but rather a rank and may indicate that max_mds was previously set higher
than its current setting of 2, and the "Stopped" ranks are simply ranks which
were active and simply offloaded their state to other ranks.
My question is, how can I inspect further which ranks are "stopped" and would
it be appropriate to "clear" those stopped ranks if possible or should I modify
my check to ignore stopped ranks and only focus on damaged/failed ranks.
The cluster is running 12.2.12
Thanks.
Respectfully,
Wes Dillingham
[email protected]
Site Reliability Engineer IV - Platform Storage
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com