Dears,

maybe someon can explain/answer my question.

We are working on setting up a new ceph cluster (underneath a proxmox cluster), 
the setup is as follows

3 datacenters
3 hosts per datacenter (all in one rack)
8 physical disks per host

This is the profile configutation

ceph osd erasure-code-profile set ec_profile plugin=isa technique=reed_sol_van 
k=4 m=2 crush-root=default crush-failure-domain=datacenter 
crush-device-class=nvme crush-osds-per-failure-domain=2 
crush-num-failure-domains=3

We have a hierarchy - created with a custom_location_hook that contains


  *
datacenter 01
     *
rack 01
        *
host 01
        *
host 02
        *
host 03
  *
datacenter 02
     *
rack 01
        *
host 01
        *
host 02
        *
host 03
  *
datacenter 03
     *
rack 01
        *
host 01
        *
host 02
        *
host 03

So this means (my understanding) that we are placing 2 shards of data in each 
datacenter.

We want to make sure that we only have max one share of data per host, and in 
future if possible also distributed across racks once we should ahve more per 
DC.

Is the bucket hierarchy automatically taken into account when choosing the OSD 
? Or if not, where do i start, with CRUSH rules ? (I am hesitant to manually 
modify CRUSH rules.

Thanks in advance

Soeren

_______________________________________________
ceph-users mailing list -- [email protected]
To unsubscribe send an email to [email protected]

Reply via email to