Hi all,

I have a similar question regarding a cluster configuration consisting of HDDs, SSDs and NVMes. Let's say I would setup a OSD configuration in a yaml file like this:

service_type:osd
service_id:osd_spec_default
placement:
host_pattern:'*'
spec:
data_devices:
model:HDD-Model-XY
db_devices:
model:NVME-Model-A
wal_devices:
model:NVME-Model-A Afaik that would result in having the db and wal on the nvmes, whereas all data is put on the HDD. I assume I would get significant IOPS performance when using Block device in Openstack (or Proxmox), but I have to put up with wearing out the NVMes, right? In the Ceph documentation, I saw that it would be also sufficient to define the 2,5" SSDs as db_devices. Nevertheless, I need some SSDs as OSDs since the Ceph FS metadata pool and NFS pool needs to be put on a SSD OSD. I might use 3 out of 4 SSDs each node as db_devices, but I need some for the metadata pools. Any suggestions? Regards, Mevludin

Am 02.01.2023 um 15:03 schrieb Erik Lindahl:
Depends.

In theory, each OSD will have access to 1/4 of the separate WAL/DB device, so 
to get better performance you need to find an NVMe device that delivers 
significantly more than 4x the IOPS rate of the pm1643 drives, which is not 
common.

That assumes the pm1643 devices are connected to a high-quality well-configured 
12Gb SAS controller that really can deliver the full IOPS rate of 4 drives 
combined. The only way to find that out is likely to benchmark.

Having said that, for a storage cluster where write performance is expected to 
be the main bottleneck, I would be hesitant to use drives that only have 1DWPD 
endurance since Ceph has fairly high write amplification factors. If you use 
3-fold replication, this cluster might only be able to handle a few TB of 
writes per day without wearing out the drives prematurely.

In practice we've been quite happy with Samsung drives that have often far 
exceeded their warranty endurance, but that's not something I would like to 
rely on when providing a commercial service.

Cheers,

Erik




--
Erik Lindahl<[email protected]>
On 2 Jan 2023 at 10:25 +0100,[email protected]  <[email protected]>, 
wrote:
Hi Experts,I am seeking for if there is achievable significant write 
performance improvements when separating WAL/DB in a ceph cluster with all SSD 
type OSD.I have a cluster with 40 SSD (PM1643 1.8 TB SSD Enterprise Samsung). I 
have 10 Storage node each with 4 OSD. I want to know that can I get better 
write IOPs and throughput if I add one NVMe OSD per node and separate WAL/DB on 
it?Is the result of this separation, meaningful performance improvement or not?
My ceph cluster is block storage back-end of Openstack cinder in a public cloud 
service.

Thanks in advance.
_______________________________________________
ceph-users mailing list [email protected]
To unsubscribe send an email [email protected]
_______________________________________________
ceph-users mailing list [email protected]
To unsubscribe send an email [email protected]

_______________________________________________
ceph-users mailing list -- [email protected]
To unsubscribe send an email to [email protected]

Reply via email to