You can either provide an image with the adopt command (—image) or you
configure it globally with ceph config set (I don’t have the exact
command right now). Which image does it fail to pull? You should see
that in cephadm.log. Does that node with osd.17 have access to the
image repo?
Zitat von Patrick Vranckx <patrick.vranckx@xxxxxxxxxxxx>:
Hi,
We've already convert two PRODUCTION storage nodes on Octopus to
cephadm without problem.
On the third one, we succeeded to convert only one OSD.
[root@server4 osd]# cephadm adopt --style legacy --name osd.0
Found online OSD at //var/lib/ceph/osd/ceph-0/fsid
objectstore_type is bluestore
Stopping old systemd unit ceph-osd@0...
Disabling old systemd unit ceph-osd@0...
Moving data...
Chowning content...
Chowning /var/lib/ceph/fsid replaced/osd.0/block...
Renaming /etc/ceph/osd/0-2d973f03-82f3-499f-b5dc-d4c28dbe1b3d.json
->
/etc/ceph/osd/0-2d973f03-82f3-499f-b5dc-d4c28dbe1b3d.json.adopted-by-cephadm
Disabling host unit ceph-volume@ simple unit...
Moving logs...
Creating new units...
For the others, we have this error:
[root@server4 osd]# cephadm adopt --style legacy --name osd.17
Found online OSD at //var/lib/ceph/osd/ceph-17/fsid
objectstore_type is bluestore
Stopping old systemd unit ceph-osd@17...
Disabling old systemd unit ceph-osd@17...
Moving data...
Traceback (most recent call last):
File "/sbin/cephadm", line 6251, in <module>
r = args.func()
File "/sbin/cephadm", line 1458, in _default_image
return func()
File "/sbin/cephadm", line 4027, in command_adopt
command_adopt_ceph(daemon_type, daemon_id, fsid);
File "/sbin/cephadm", line 4170, in command_adopt_ceph
os.rmdir(data_dir_src)
OSError: [Errno 16] Device or resource busy: '//var/lib/ceph/osd/ceph-17'
The directory /var/lib/ceph/osd/ceph-17 is now empty.
The directory /var/lib/ceph/<fsid>/osd.17 contains:
[root@server4 osd.17]# ls -l
total 72
-rw-r--r-- 1 ceph ceph 411 Jan 29 2018 activate.monmap
-rw-r--r-- 1 ceph ceph 3 Jan 29 2018 active
lrwxrwxrwx 1 root root 10 Nov 8 15:54 block -> /dev/sdad2
-rw-r--r-- 1 ceph ceph 37 Jan 29 2018 block_uuid
-rw-r--r-- 1 ceph ceph 2 Jan 29 2018 bluefs
-rw-r--r-- 1 ceph ceph 37 Jan 29 2018 ceph_fsid
-rw-r--r-- 1 ceph ceph 1226 Nov 8 15:53 config
-rw-r--r-- 1 ceph ceph 37 Jan 29 2018 fsid
-rw------- 1 ceph ceph 57 Jan 29 2018 keyring
-rw-r--r-- 1 ceph ceph 8 Jan 29 2018 kv_backend
-rw-r--r-- 1 ceph ceph 21 Jan 29 2018 magic
-rw-r--r-- 1 ceph ceph 4 Jan 29 2018 mkfs_done
-rw-r--r-- 1 ceph ceph 6 Jan 29 2018 ready
-rw------- 1 ceph ceph 3 Nov 8 14:47 require_osd_release
-rw-r--r-- 1 ceph ceph 0 Jan 13 2020 systemd
-rw-r--r-- 1 ceph ceph 10 Jan 29 2018 type
-rw------- 1 root root 22 Nov 8 15:53 unit.image
-rw------- 1 root root 1042 Nov 8 16:30 unit.poststop
-rw------- 1 root root 1851 Nov 8 16:30 unit.run
-rw-r--r-- 1 ceph ceph 3 Jan 29 2018 whoami
When trying to start or redeploy osd.17, podman inspect complains
about non-existent image:
2022-11-08 16:58:58,503 7f930fab3740 DEBUG Running command:
/bin/podman inspect --format
{{.Id}},{{.Config.Image}},{{.Image}},{{.Created}},{{index
.Config.Labels "io.ceph.version"}} ceph-<fsid replaced>-osd.17
2022-11-08 16:58:58,591 7f930fab3740 DEBUG /bin/podman: stderr
Error: error getting image "ceph-<fsid replaced>-osd.17": unable to
find a name and tag match for ceph-<fsid replaced>-osd.17 in
repotags: no such image
Is there a way to save osd.17 and create manually the podman image ?
Thanks in advance,
Patrick
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx