Hi, I can remember that it took with the tools before cephadm somehow 8 hours to deploy a ceph cluster with more than 2000 osds. But I also know that CBT has a much faster approach to installing a Ceph cluster. Just an idea: maybe you can look at the approach at CBT to make cephadm faster. Regards, Joachim joachim.kraftmayer@xxxxxxxxx www.clyso.com Hohenzollernstr. 27, 80801 Munich Utting | HR: Augsburg | HRB: 25866 | USt. ID-Nr.: DE275430677 Am Fr., 22. Nov. 2024 um 15:46 Uhr schrieb Eugen Block <eblock@xxxxxx>: > Hi, > > I don't see how it would be currently possible. The OSD creation is > handled by ceph-volume, which activates each OSD separately: > > [2024-11-22 14:03:08,415][ceph_volume.main][INFO ] Running command: > ceph-volume activate --osd-id 0 --osd-uuid > aacabeca-9adb-465c-88ee-935f06fa45f7 --no-systemd --no-tmpfs > > [2024-11-22 14:03:09,343][ceph_volume.devices.raw.activate][INFO ] > Activating osd.0 uuid aacabeca-9adb-465c-88ee-935f06fa45f7 cluster > e57f7b6a-a8d9-11ef-af3c-fa163e2ad8c5 > > The ceph-volume lvm activate description [0] states: > > > It is possible to activate all existing OSDs at once by using the > > --all flag. For example: > > > > ceph-volume lvm activate --all > > > > This call will inspect all the OSDs created by ceph-volume that are > > inactive and will activate them one by one. > > I assume that even if the OSD creation process could be tweaked in a > way that all OSDs are created first without separate activation, and > then cephadm would issue "ceph-volume lvm activate --all", the OSDs > would still be activated one by one. > > But as Tim already stated, an hour for almost 200 OSDs is not that > bad. ;-) I guess you could create a tracker issue for an enhancement, > maybe some of the devs can clarify why the OSDs need to be activated > one by one. > > Regards, > Eugen > > [0] https://docs.ceph.com/en/latest/ceph-volume/lvm/activate/ > > > Zitat von YuFan Chen <wiz.chen@xxxxxxxxx>: > > > Hi, > > > > I’m setting up a 6-node Ceph cluster using Ceph Squid. > > Each node is configured with 32 OSDs (32 HDDs and 8 NVMe SSDs for > > db_devices). > > > > I’ve created an OSD service specification and am using cephadm to > > apply the configuration. > > The deployment of all 192 OSDs takes about an hour to complete. > > > > However, I’ve noticed that cephadm creates the OSDs sequentially. > > Then, on each node, it starts a single OSD and waits for it to become > > ready before moving on to the next. > > > > Is there a way to speed up the OSD deployment process? > > Thanks in advance for your help! > > > > Best regards, > > Yufan Chen > > _______________________________________________ > > ceph-users mailing list -- ceph-users@xxxxxxx > > To unsubscribe send an email to ceph-users-leave@xxxxxxx > > > _______________________________________________ > ceph-users mailing list -- ceph-users@xxxxxxx > To unsubscribe send an email to ceph-users-leave@xxxxxxx > _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx