Hi Laszlo, I think it depends on the type of cluster you're trying to build. If made of HDD/SSD OSDs, then 2 cores per OSD is probably still valid. I believe the 5-6 cores per OSDs recommendation you mentioned relates to all flash (NVMe) clusters where CPUs and especially memory bandwidth can't always keep up with all flash storage and small IO workloads. "Ceph can easily utilize five or six cores on real clusters and up to about fourteen cores on single OSDs in isolation" [1] I'd say 'real clusters' here stands for all flash NVMe clusters with multiples OSDs on multiple hosts with x3 replication, while 'a single OSDs in isolation' refers to a single NVMe OSD, maybe on a single host cluster with no replication at all. But this may need further confirmation. At the end of day, if your intention is to build an all flash cluster then I think you should multiply 5-6 cores by the number of NVMe OSDs and chose CPU accordingly (with the highest frequency you can get for the bucks). You might want to check Mark's talks [2][3] and studies [4][5][6] about all flash Ceph clusters. It explains it all and suggests some modern hardware for all flash storage, if that's what you're building. Cheers, Frédéric. [1] https://github.com/ceph/ceph/pull/44466#discussion_r779650295 [2] https://youtu.be/S2rPA7qlSYY [3] https://youtu.be/pGwwlaCXfzo [4] https://ceph.io/en/news/blog/2024/ceph-a-journey-to-1tibps/ [5] https://docs.clyso.com/blog/ceph-a-journey-to-1tibps [6] https://docs.clyso.com/blog/clyso-enterprise-storage-all-flash-ceph-deployment-guide-preview/ ----- Le 18 Sep 24, à 10:15, Laszlo Budai laszlo@xxxxxxxxxxxxxxxx a écrit : > Hello everyone, > > I'm trying to understand the CPU requirements for recent versions of CEPH. > Reading the documentation > (https://docs.ceph.com/en/latest/start/hardware-recommendations/) I cannot get > any conclusion about how to plan CPUs for ceph. There is the following > statement: > > "With earlier releases of Ceph, we would make hardware recommendations based on > the number of cores per OSD, but this cores-per-osd metric is no longer as > useful a metric as the number of cycles per IOP and the number of IOPS per OSD. > For example, with NVMe OSD drives, Ceph can easily utilize five or six cores on > real clusters and up to about fourteen cores on single OSDs in isolation. So > cores per OSD are no longer as pressing a concern as they were. When selecting > hardware, select for IOPS per core." > > How should I understand this? On one side it's saying "with NVMe OSD drives, > Ceph can easily utilize five or six cores on real clusters" > and then continues: "and up to about fourteen cores on single OSDs in > isolation." so should I cont 5/6 cores per NVMe OSD or 14cores? > > And after all that there's the next sentence that says: " So cores per OSD are > no longer as pressing a concern as they were.", But if a single OSD can consume > 5/6 cpu cures (that could go up to 14) then I would assume that cores per OSD > is still an important concern. > > > Can anyone explain these CPU requirements? Or point me to some other documents > that describes in more details the resources required for Ceph? > > > Thank you, > Laszlo > _______________________________________________ > ceph-users mailing list -- ceph-users@xxxxxxx > To unsubscribe send an email to ceph-users-leave@xxxxxxx _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx