Hi Nico,
No 2 data centers.
- We use size=4
- our ceph map is configured with OSD's assigned to 2 separate data center
locations, so we end up with 2 OSD's in use from in each DC
- min_size=2.
- we have (1) monitor in each DC
- we have a 3rd monitor that is in a 3rd DC and has a VPN connection to each
of the other 2 DC's to provide a tie-breaker for the monitors.
You would also need to change your ceph rules to indicate you want
datacenter and host redundancy, that part of ours looks like this:
# rules
rule replicated_ruleset {
id 0
type replicated
min_size 1
max_size 10
step take default
step choose firstn 2 type datacenter
step chooseleaf firstn 2 type host
step emit
}
We run dedicated fiber between the 2 DC's, and so with this config we can
lose an entire DC (or have a fiber cut) and our ceph cluster will still be
able to serve data to VM's located either on each side (if there is a fiber
cut), or we can restart VM's that were running in the DC that is down on
spare hardware in the other location - so long as we don't also lose access
to our tie-breaker ceph monitor. If that happens.. I think we are screwed :/
.. but it uses the internet for those VPN connections.. so we would also
have to have no internet access at the same time and both sides are
redundant internet wise.
If anyone sees I hole in my config here, please feel free to correct me :)
I'd hate to find out the hard way lol.
Cheers,
D.
-----Original Message-----
From: Nico Schottelius [mailto:[email protected]]
Sent: Friday, March 25, 2022 12:58 PM
To: Wolfpaw - Dale Corse <[email protected]>
Cc: [email protected]
Subject: [ceph-users] Re: Even number of replicas?
Hey Dale,
are you distributing your clusters over 4 DCs via dark fiber or can you
elaborate on your setup?
We are currently running 3/2, but each cluster is isolated in its own DC.
Best,
Nico
"Wolfpaw - Dale Corse" <[email protected]> writes:
> Hi George,
>
> We use 4/2 for our deployment and it works fine - but it's a huge
> waste of space :)
>
> Our reason is because we want to be able to lose a data center and
> still have ceph running. You could accomplish that with size=1 on an
> emergency basis, but we didn't like the redundancy loss.
>
> Cheers,
> D.
>
> -----Original Message-----
> From: Kyriazis, George [mailto:[email protected]]
> Sent: Friday, March 25, 2022 9:54 AM
> To: ceph-users <[email protected]>
> Subject: [ceph-users] Even number of replicas?
>
> Hello ceph-users,
>
> I was wondering if it is good practice to have an even number of
> replicas in a replicated pool. For example, have size=4 and min_size=2.
>
> Thank you!
>
> George
>
> _______________________________________________
> ceph-users mailing list -- [email protected] To unsubscribe send an
> email to [email protected]
>
> _______________________________________________
> ceph-users mailing list -- [email protected] To unsubscribe send an
> email to [email protected]
--
Sustainable and modern Infrastructures by ungleich.ch
_______________________________________________
ceph-users mailing list -- [email protected] To unsubscribe send an email
to [email protected]
_______________________________________________
ceph-users mailing list -- [email protected]
To unsubscribe send an email to [email protected]