Hello Stefan,
The status was "HEALTH_OK" before i ran those commands.
I removed the OSDs with:
ceph osd out osd.10
ceph auth del osd.10
systemctl stop ceph-osd@10
ceph osd rm 10
umount /var/lib/ceph/osd/ceph-10
ceph osd crush remove osd.10
dd if=/dev/zero of=/dev/sdc
ceph osd out osd.9
ceph auth del osd.9
systemctl stop ceph-osd@9
ceph osd rm 9
umount /var/lib/ceph/osd/ceph-9
ceph osd crush remove osd.9
root@ceph01:~# ceph osd crush rule dump
[
{
"rule_id": 0,
"rule_name": "replicated_ruleset",
"ruleset": 0,
"type": 1,
"min_size": 1,
"max_size": 10,
"steps": [
{
"op": "take",
"item": -1,
"item_name": "default"
},
{
"op": "chooseleaf_firstn",
"num": 0,
"type": "host"
},
{
"op": "emit"
}
]
}
]
Thanks,
Mario
On Sun, Dec 29, 2019 at 2:16 PM Stefan Kooman <[email protected]> wrote:
>
> Quoting Ml Ml ([email protected]):
> > Hello List,
> > i have size = 3 and min_size = 2 with 3 Nodes.
>
> That's good.
>
> >
> >
> > I replaced two osds on node ceph01 and ran into "HEALTH_ERR".
> > My problem: it waits for the backfilling process?
> > Why did i run into HEALTH_ERR? I thought all data will be available on
> > at least one more node. or even two:
>
> How did you replace them? Did you first set them "out" and waited for
> the data to be repicated elsewhere before you removed them?
>
> It *might* be because your CRUSH rule set is replicating over "OSD" and
> not host. What does a "ceph osd crush rule dump" shows?
>
> Gr. Stefan
>
> --
> | BIT BV https://www.bit.nl/ Kamer van Koophandel 09090351
> | GPG: 0xD14839C6 +31 318 648 688 / [email protected]
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com