Hi Anthony,
Thanks for the reply!
We will be using CephFS to access Ceph Storage from clients. So, this
will need MDS daemon also.
So, based on your advice, I am thinking of having 4 Dell PowerEdge servers
. 3 of them will run 3 Monitor daemons and one of them will run MDS
daemon.
These Dell Servers will have following hardware :-
1. 4 cores ( 8 threads ) ( Can go for 8 core and 16 threads )
2. 64G RAM
3. 2x4T Samsung SSD with RA!D 1 to install OS and run monitor and
metadata services.
OSD nodes will be upgraded to have 32 cores ( 64 threads ). Disk and RAM
will remain same ( 128G and 22X8T Samsung SSD )
Actually , I want to use OSD nodes to run OSD damons and not any
other demons and which is why I am thinking of having 4 additional Dell
servers as mentioned above.
Please advise if this plan will be better.
Thanks,
Gagan
On Wed, Apr 9, 2025 at 8:12 PM Anthony D'Atri <[email protected]>
wrote:
>
> >
> > We would start deploying Ceph with 4 hosts ( HP Proliant servers ) each
> > running RockyLinux 9.
> >
> > One of the hosts called ceph-adm will be smaller one and will have
> > following hardware :-
> >
> > 2x4T SSD with raid 1 to install OS on.
> >
> > 8 Core with 3600MHz freq.
> >
> > 64G RAM
> >
> > We are planning to run all Ceph daemons except OSD daemon like monitor ,
> > metadata ,etc on this host.
>
> 8 core == 16 threads? Are you provisioning this node because you have it
> laying around idle?
>
> Note that you will want *at least* 3 Monitor (monitors) daemons, which
> must be on different nodes. 5 is better, but at least 3. You’ll also have
> Grafana, Prometheus, MDS (if you’re going to CephFS vs using S3 object
> storage or RBD block)
>
> 8c is likely on the light side for all of that. You would also benefit
> from not having that node be a single point of failure. I would suggest if
> you can raising this node to the spec of the planned 3x OSD nodes so you
> have 4x equivalent nodes, and spread that non-OSD daemons across them.
>
> Note also that your OSD nodes will also have node_exporter, crash, and
> other boilerplate daemons.
>
>
> > We will have 3 hosts to run OSD which will store actual data.
> >
> > Each OSD host will have following hardware
> >
> > 2x4T SSD with raid 1 to install OS on.
> >
> > 22X8T SSD to store data ( OSDs ) ( without partition ). We will use
> entire
> > disk without partitions
>
> SAS, SATA, or NVMe SSDs? Which specific model? You really want to avoid
> client (desktop) models for Ceph, but you likely do not need to pay for
> higher endurance mixed-use SKUs.
>
> > Each OSD host will have 128G RAM ( No swap space )
>
> Thank you for skipping swap. Some people are really stuck in the past in
> that regard.
>
> > Each OSD host will have 16 cores.
>
> So 32 threads total? That is very light for 22 OSDs + other daemons. For
> HDD OSDs a common rule of thumb is at minimum 2x threads per, for SAS/SATA
> SSDs, 4, for NVMe SSDs 6. Plus margin for the OS and other processes.
>
> > All 4 hosts will connect to each via 10G nic.
>
> Two ports with bonding? Redundant switches?
>
> > The 500T data
>
> The specs you list above include 528 TB of *raw* space. Be advised that
> with three OSD nodes, you will necessarily be doing replication. For
> safety replication with size=3. Taking into consideration TB vs TiB and
> headroom, you’re looking at 133TiB of usable space. You could go with
> size=2 to get 300TB of usable space, but at increased risk of data
> unavailability or loss when drives/hosts fail or reboot.
>
> With at least 4 OSD nodes - even if they aren’t fully populated with
> capacity drives — you could do EC for a more favorable raw:usable ratio, at
> the expense of slower writes and recovery. With 4 nodes you could in
> theory do 2,2 EC for 200 TiB of usable space, with 5 you could do 3,2 for
> 240 TiB usable, etc.
>
> > will be accessed by the clients. We need to have
> > read performance as fast as possible.
>
> Hope your SSDs are enterprise NVMe.
>
> > We can't afford data loss and downtime.
>
> Then no size=2 for you.
>
> > So, we want to have a Ceph
> > deployment which serves our purpose.
> >
> > So, please advise me if the plan that I have designed will serve our
> > purpose.
> > Or is there a better way , please advise that.
> >
> > Thanks,
> > Gagan
> >
> >
> >
> >
> >
> >
> > We have a HP storage server with 12 SDD of 5T each and have set-up
> hardware
> > RAID6 on these disks.
> >
> > HP storage server has 64G RAM and 18 cores.
> >
> > So, please advise how I should go about setting up Ceph on it to have
> best
> > read performance. We need fastest read performance.
> >
> >
> > Thanks,
> > Gagan
> > _______________________________________________
> > ceph-users mailing list -- [email protected]
> > To unsubscribe send an email to [email protected]
>
>
_______________________________________________
ceph-users mailing list -- [email protected]
To unsubscribe send an email to [email protected]