Hi,
I don't see how it would be currently possible. The OSD creation is
handled by ceph-volume, which activates each OSD separately:
[2024-11-22 14:03:08,415][ceph_volume.main][INFO ] Running command:
ceph-volume activate --osd-id 0 --osd-uuid
aacabeca-9adb-465c-88ee-935f06fa45f7 --no-systemd --no-tmpfs
[2024-11-22 14:03:09,343][ceph_volume.devices.raw.activate][INFO ]
Activating osd.0 uuid aacabeca-9adb-465c-88ee-935f06fa45f7 cluster
e57f7b6a-a8d9-11ef-af3c-fa163e2ad8c5
The ceph-volume lvm activate description [0] states:
It is possible to activate all existing OSDs at once by using the
--all flag. For example:
ceph-volume lvm activate --all
This call will inspect all the OSDs created by ceph-volume that are
inactive and will activate them one by one.
I assume that even if the OSD creation process could be tweaked in a
way that all OSDs are created first without separate activation, and
then cephadm would issue "ceph-volume lvm activate --all", the OSDs
would still be activated one by one.
But as Tim already stated, an hour for almost 200 OSDs is not that
bad. ;-) I guess you could create a tracker issue for an enhancement,
maybe some of the devs can clarify why the OSDs need to be activated
one by one.
Regards,
Eugen
[0] https://docs.ceph.com/en/latest/ceph-volume/lvm/activate/
Zitat von YuFan Chen <wiz.chen@xxxxxxxxx>:
Hi,
I’m setting up a 6-node Ceph cluster using Ceph Squid.
Each node is configured with 32 OSDs (32 HDDs and 8 NVMe SSDs for
db_devices).
I’ve created an OSD service specification and am using cephadm to
apply the configuration.
The deployment of all 192 OSDs takes about an hour to complete.
However, I’ve noticed that cephadm creates the OSDs sequentially.
Then, on each node, it starts a single OSD and waits for it to become
ready before moving on to the next.
Is there a way to speed up the OSD deployment process?
Thanks in advance for your help!
Best regards,
Yufan Chen
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx