The OSD node is the same as the monitor and manager node at the moment and has the ceph.conf file: user@node1:~$ ls -l /etc/ceph/ total 15 -rw------- 1 root root 151 Aug 13 15:50 ceph.client.admin.keyring -rw-r--r-- 1 root root 432 Aug 13 16:09 ceph.conf -rw-r--r-- 1 root root 92 Jun 30 16:44 rbdmap On 8/14/20 10:19 AM, Eugen Block wrote: > The OSD node also needs the ceph.conf, it seems that is not the case in your setup. > > > Zitat von Joshua Schaeffer <jschaeffer@xxxxxxxxxxxxxxx>: > >> Hey all, >> >> I'm trying to deploy Ceph 15.2.4 on Ubuntu 20.04 and I am going through the manual deploy process [1]. I was able to successfully bootstrap the monitor and manager and am now trying to add the OSD's, but the `ceph-volume` command is hanging when running `ceph osd new`. It appears the command hangs for about 30 minutes and then times out with this error: >> >> user@node1:~$ sudo ceph-volume lvm create --data /dev/sdc >> Running command: /usr/bin/ceph-authtool --gen-print-key >> Running command: /usr/bin/ceph --cluster ceph --name client.bootstrap-osd --keyring /var/lib/ceph/bootstrap-osd/ceph.keyring -i - osd new 9a305232-6b0a-4fde-9685-68fe6150da0e >> stderr: [errno 110] RADOS timed out (error connecting to the cluster) >> --> RuntimeError: Unable to create a new OSD id >> >> I have also tried to manually run the `ceph osd new` command, but it results in the same error. My cluster shows as up (at least the monitor and manager): >> >> user@node1:~$ sudo ceph -s >> cluster: >> id: 6aacdfe1-42e4-4106-9503-3e16e9812195 >> health: HEALTH_WARN >> 1 monitors have not enabled msgr2 >> Reduced data availability: 1 pg inactive >> OSD count 0 < osd_pool_default_size 3 >> >> services: >> mon: 1 daemons, quorum bllcloudceph01 (age 23h) >> mgr: bllcloudceph01(active, since 23h) >> osd: 0 osds: 0 up, 0 in >> >> data: >> pools: 1 pools, 1 pgs >> objects: 0 objects, 0 B >> usage: 0 B used, 0 B / 0 B avail >> pgs: 100.000% pgs unknown >> 1 unknown >> >> And ceph appears to be listening: >> >> user@node1:~$ sudo ss -tupanl | grep ceph >> tcp LISTEN 0 512 10.2.22.1:6789 0.0.0.0:* users:(("ceph-mon",pid=136244,fd=25)) >> tcp LISTEN 0 512 10.2.22.1:6800 0.0.0.0:* users:(("ceph-mgr",pid=138125,fd=25)) >> tcp LISTEN 0 512 10.2.22.1:6801 0.0.0.0:* users:(("ceph-mgr",pid=138125,fd=26)) >> >> It seems like I'm missing something simple, but I can't spot it. Can anybody point me in the right direction? >> >> [1] https://docs.ceph.com/docs/master/install/manual-deployment/ >> >> -- >> Thanks, >> Joshua Schaeffer >> >> _______________________________________________ >> ceph-users mailing list -- ceph-users@xxxxxxx >> To unsubscribe send an email to ceph-users-leave@xxxxxxx > > > _______________________________________________ > ceph-users mailing list -- ceph-users@xxxxxxx > To unsubscribe send an email to ceph-users-leave@xxxxxxx -- Thanks, Joshua Schaeffer _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx