Hi, i like to answer to myself :-) I finally found the rest of my documentation... So after reinstalling the OS also the osd config must be created. Here is what i have done, maybe this helps someone: ------------------ Get the informations: ``` cephadm ceph-volume lvm list ceph config generate-minimal-conf ceph auth get osd.[ID] ``` Now create a minimal osd config: ``` vi osd.[ID].json ``` ``` { "config": "# minimal ceph.conf for 6d0ecf22-9155-4684-971a-2f6cde8628c8\n[global]\n\tfsid = 6d0ecf22-9155-4684-971a-2f6cde8628c8\n\tmon_host = [v2:192.168.6.21:3300/0,v1:192.168.6.21:6789/0] [v2:192.168.6.22:3300/0,v1:192.168.6.22:6789/0] [v2:192.168.6.23:3300/0,v1:192.168.6.23:6789/0] [v2:192.168.6.24:3300/0,v1:192.168.6.24:6789/0] [v2:192.168.6.25:3300/0,v1:192.168.6.25:6789/0]\n", "keyring": "[osd.XXX]\n\tkey = XXXXXXXXXXXXXXXXXXXX\n" } ``` Deploy the OSD daemon: ``` cephadm deploy --fsid 6d0ecf22-9155-4684-971a-2f6cde8628c8 --osd-fsid [ID] --name osd.[ID] --config-json osd.[ID].json ``` Yours, bbk On Thu, 2021-12-09 at 18:35 +0100, bbk wrote: > After reading my mail it may not be clear that i reinstalled the OS of > a node with OSDs. > > On Thu, 2021-12-09 at 18:10 +0100, bbk wrote: > > Hi, > > > > the last time i have reinstalled a node with OSDs, i added the disks > > with the following command. But unfortunatly this time i ran into a > > error. > > > > It seems like this time the command doesn't create the container, i > > am able to run `cephadm shell`, and other daemons (mon,mgr,mds) are > > running. > > > > I don't know if that is the right way to do it? > > > > > > ~# cephadm deploy --fsid 6d0ecf22-9155-4684-971a-2f6cde8628c8 --osd- > > fsid 941c6cb6-6898-4aa2-a33a-cec3b6a95cf1 --name osd.9 > > > > Non-zero exit code 125 from /usr/bin/podman container inspect -- > > format {{.State.Status}} ceph-6d0ecf22-9155-4684-971a-2f6cde8628c8- > > osd-9 > > /usr/bin/podman: stderr Error: error inspecting object: no such > > container ceph-6d0ecf22-9155-4684-971a-2f6cde8628c8-osd-9 > > Non-zero exit code 125 from /usr/bin/podman container inspect -- > > format {{.State.Status}} ceph-6d0ecf22-9155-4684-971a-2f6cde8628c8- > > osd.9 > > /usr/bin/podman: stderr Error: error inspecting object: no such > > container ceph-6d0ecf22-9155-4684-971a-2f6cde8628c8-osd.9 > > Deploy daemon osd.9 ... > > Non-zero exit code 1 from systemctl start > > ceph-6d0ecf22-9155-4684-971a-2f6cde8628c8@osd.9 > > systemctl: stderr Job for > > ceph-6d0ecf22-9155-4684-971a-2f6cde8628c8@osd.9.service failed > > because the control process exited with error code. > > systemctl: stderr See "systemctl status > > ceph-6d0ecf22-9155-4684-971a-2f6cde8628c8@osd.9.service" and > > "journalctl -xe" for details. > > Traceback (most recent call last): > > File "/usr/sbin/cephadm", line 8571, in <module> > > main() > > File "/usr/sbin/cephadm", line 8559, in main > > r = ctx.func(ctx) > > File "/usr/sbin/cephadm", line 1787, in _default_image > > return func(ctx) > > File "/usr/sbin/cephadm", line 4549, in command_deploy > > ports=daemon_ports) > > File "/usr/sbin/cephadm", line 2677, in deploy_daemon > > c, osd_fsid=osd_fsid, ports=ports) > > File "/usr/sbin/cephadm", line 2906, in deploy_daemon_units > > call_throws(ctx, ['systemctl', 'start', unit_name]) > > File "/usr/sbin/cephadm", line 1467, in call_throws > > raise RuntimeError('Failed command: %s' % ' '.join(command)) > > RuntimeError: Failed command: systemctl start > > ceph-6d0ecf22-9155-4684-971a-2f6cde8628c8@osd.9 > > > > > > ~# cephadm ceph-volume lvm list > > > > ====== osd.9 ======= > > > > [block] /dev/ceph-07fa2bb7-628f-40c0-8725-0266926371c0/osd- > > block-941c6cb6-6898-4aa2-a33a-cec3b6a95cf1 > > > > block device /dev/ceph-07fa2bb7-628f-40c0-8725- > > 0266926371c0/osd-block-941c6cb6-6898-4aa2-a33a-cec3b6a95cf1 > > block uuid mVEhfF-LK4E-Dtmb-Jj23-tn8x-lpLy- > > KiUy1a > > cephx lockbox secret > > cluster fsid 6d0ecf22-9155-4684-971a-2f6cde8628c8 > > cluster name ceph > > crush device class None > > encrypted 0 > > osd fsid 941c6cb6-6898-4aa2-a33a-cec3b6a95cf1 > > osd id 9 > > type block > > vdo 0 > > devices /dev/sdd > > > > > > ~# podman --version > > podman version 3.2.3 > > > > > > ~# cephadm version > > Using recent ceph image > > quay.io/ceph/ceph@sha256:2f7f0af8663e73a422f797de605e769ae44eb0297f2a > > 79324739404cc1765728 > > ceph version 16.2.7 (dd0603118f56ab514f133c8d2e3adfc983942503) > > pacific (stable) > > > > > > ~# lsb_release -a > > LSB Version: :core-4.1-amd64:core-4.1-noarch > > Distributor ID: RedHatEnterprise > > Description: Red Hat Enterprise Linux release 8.5 (Ootpa) > > Release: 8.5 > > Codename: Ootpa > > > > > > ~# cephadm shell > > Inferring fsid 6d0ecf22-9155-4684-971a-2f6cde8628c8 > > Using recent ceph image > > quay.io/ceph/ceph@sha256:2f7f0af8663e73a422f797de605e769ae44eb0297f2a > > 79324739404cc1765728 > > [ceph: root@hobro /]# > > > > > > Yours, > > bbk > > > > _______________________________________________ > > ceph-users mailing list -- ceph-users@xxxxxxx > > To unsubscribe send an email to ceph-users-leave@xxxxxxx > > > _______________________________________________ > ceph-users mailing list -- ceph-users@xxxxxxx > To unsubscribe send an email to ceph-users-leave@xxxxxxx _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx