On 3/10/21 5:43 PM, Ignazio Cassano wrote:
Hello, what do you think about of ceph cluster made up of 6 nodes each one
with the following configuration ?
A+ Server 1113S-WN10RT
Barebone
Supermicro A+ Server 1113S-WN10RT - 1U - 10x U.2 NVMe - 2x M.2 - Dual
10-Gigabit LAN - 750W Redundant
Processor
AMD EPYC™ 7272 Processor 12-core 2.90GHz 64MB Cache (120W)
Memory
8 x 8GB PC4-25600 3200MHz DDR4 ECC RDIMM
^^ I would double that amount of RAM, especially (see below) if you plan
on adding more NVMe drives.
U.2/U.3 NVMe Drive
5 x 8.0TB Intel® SSD DC P4510 Series U.2 PCIe 3.1 x4 NVMe Solid State Drive
Hard Drive
^^ Why 5 * 8.0 TB instead of 10 * 4.0 TB? Are you planning on upgrading
later? Ceph likes more OSDs better than fewer larger ones. Recovery will
be faster as well, and the impact of one NVMe dying will be lower.
2 x 240GB Intel® SSD D3-S4610 Series 2.5" SATA 6.0Gb/s Solid State Drive
^^ Do you sacrifce two NVMe ports for two SATA OS disks? If so, I would
advise for getting (redundant, optional) U.2 NVMe or SATADOM or similar.
Network Card
2 x Intel® 10-Gigabit Ethernet Converged Network Adapter X710-DA2 (2x SFP+)
Server Management
^ Why two? One for "public" and one for "cluster"? Than most probably
you won't need that, and one bond would suffice (see current Ceph best
practices). If you need 40 Gb/s in one LACP trunk: perfectly fine.
Gr. Stefan
_______________________________________________
ceph-users mailing list -- [email protected]
To unsubscribe send an email to [email protected]