Re: Reinstalling OSD node managed by cephadm

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Well, at least it reports that the OSDs were activated successfully:

/bin/docker: --> ceph-volume lvm activate successful for osd ID: 12
/bin/docker: --> Activating OSD ID 25 FSID 3f3d61f8-6964-4922-98cb-6620aff5cb6f

Now you need to get the pods up, I'm not sure if cephadm will manage that automatically or if you need to create anything yourself. Can you try to start it via 'systemctl start ceph-d221bc3c-8ff4-11ec-b4ba-b02628267680@osd.12' ?

Zitat von Manuel Holtgrewe <zyklenfrei@xxxxxxxxx>:

Hm, not much more luck:

# cephadm --verbose ceph-volume lvm activate --all --no-systemd
--------------------------------------------------------------------------------
cephadm ['--verbose', 'ceph-volume', 'lvm', 'activate', '--all',
'--no-systemd']
Using default config: /etc/ceph/ceph.conf
/bin/docker: 920d72f71171,79MiB / 251.5GiB
/bin/docker: 72570e59f28e,77.43MiB / 251.5GiB
/bin/docker: b1a0de17c02d,854.8MiB / 251.5GiB
/bin/docker: cce740906d36,1.158GiB / 251.5GiB
/bin/docker: 4fba9fb86122,209.3MiB / 251.5GiB
/bin/docker: 9b06ce4f696b,96.83MiB / 251.5GiB
/bin/docker: 35269d267dfc,91.12MiB / 251.5GiB
/bin/docker: 54a2a92d68fa,278.1MiB / 251.5GiB
/bin/docker: cee7af4d922d,871.8MiB / 251.5GiB
/bin/docker: f821b8d5e473,627.8MiB / 251.5GiB
/bin/docker: 805061ed9471,473.3MiB / 251.5GiB
/bin/docker: 52e339e96d47,451.7MiB / 251.5GiB
/bin/docker: 542bda96ee76,409.9MiB / 251.5GiB
/bin/docker: 1d76652acb43,192MiB / 251.5GiB
/bin/docker: 0934566689b5,120.9MiB / 251.5GiB
/bin/docker: 17c036fbb15d,284.9MiB / 251.5GiB
/bin/docker: 63d0925cfc81,1.646GiB / 251.5GiB
/bin/docker: 06e96a3e5cbd,572.8MiB / 251.5GiB
/bin/docker: 2d2fc47ec89a,19.59MiB / 251.5GiB
/bin/docker: 26adae09c237,233.6MiB / 251.5GiB
/bin/docker: 86cf7b732b13,90.63MiB / 251.5GiB
/bin/docker: 92d366bd19aa,2.008MiB / 251.5GiB
/bin/docker: 01e9dc38faab,208.1MiB / 251.5GiB
/bin/docker: da22e3f7811c,655MiB / 251.5GiB
/bin/docker: d88137996826,783.1MiB / 251.5GiB
/bin/docker: a9f1e48f4a04,585.5MiB / 251.5GiB
/bin/docker: 49742307a5b4,478.1MiB / 251.5GiB
/bin/docker: 2bb7d623ed93,205.7MiB / 251.5GiB
/bin/docker: 0b323f254217,243.5MiB / 251.5GiB
/bin/docker: 8835104b0515,123.5MiB / 251.5GiB
/bin/docker: 67472592f1b4,658.9MiB / 251.5GiB
/bin/docker: 5aef3941d33b,474.4MiB / 251.5GiB
/bin/docker: 0d571f24c0e9,100.5MiB / 251.5GiB
/bin/docker: 7e706a832e4a,2.11GiB / 251.5GiB
/bin/docker: 23a669b4fd7c,1.164GiB / 251.5GiB
/bin/docker: 74581172b6fb,1.91MiB / 251.5GiB
/bin/docker: 37e980f9f83d,2.258MiB / 251.5GiB
/bin/docker: 858a9b82375d,178.5MiB / 251.5GiB
/bin/docker: dc97abdfc393,16.77MiB / 251.5GiB
/bin/docker: d561d2aa6053,5.379MiB / 251.5GiB
/bin/docker: 804b450b057f,655.4MiB / 251.5GiB
/bin/docker: 011180b66a7a,2.633MiB / 251.5GiB
/bin/docker: 21b984cdc711,56.98MiB / 251.5GiB
/bin/docker: c197971ee2ba,1.754MiB / 251.5GiB
/bin/docker: a10112ce1fa7,1.887MiB / 251.5GiB
/bin/docker: 546536f82ada,160.5MiB / 251.5GiB
/bin/docker: 5822da42e73b,38.15MiB / 251.5GiB
/bin/docker: 38e044b7d42f,6.645MiB / 251.5GiB
Inferring fsid d221bc3c-8ff4-11ec-b4ba-b02628267680
/bin/docker:
quay.io/ceph/ceph@sha256:0d927ccbd8892180ee09894c2b2c26d07c938bf96a56eaee9b80fc9f26083ddb
/bin/docker: quay.io/ceph/ceph@
Using recent ceph image
quay.io/ceph/ceph@sha256:0d927ccbd8892180ee09894c2b2c26d07c938bf96a56eaee9b80fc9f26083ddb
stat: 167 167
Acquiring lock 140627186853536 on
/run/cephadm/d221bc3c-8ff4-11ec-b4ba-b02628267680.lock
Lock 140627186853536 acquired on
/run/cephadm/d221bc3c-8ff4-11ec-b4ba-b02628267680.lock
sestatus: SELinux status:                 disabled
sestatus: SELinux status:                 disabled
/bin/docker: --> Activating OSD ID 12 FSID
e2ebb627-28aa-45a3-9261-d7c27bc08448
/bin/docker: Running command: /usr/bin/mount -t tmpfs tmpfs
/var/lib/ceph/osd/ceph-12
/bin/docker: Running command: /usr/bin/chown -R ceph:ceph
/var/lib/ceph/osd/ceph-12
/bin/docker: Running command: /usr/bin/ceph-bluestore-tool --cluster=ceph
prime-osd-dir --dev
/dev/ceph-2cbf3973-13a3-444a-b335-a0262cff6074/osd-block-e2ebb627-28aa-45a3-9261-d7c27bc08448
--path /var/lib/ceph/osd/ceph-12 --no-mon-config
/bin/docker: Running command: /usr/bin/ln -snf
/dev/ceph-2cbf3973-13a3-444a-b335-a0262cff6074/osd-block-e2ebb627-28aa-45a3-9261-d7c27bc08448
/var/lib/ceph/osd/ceph-12/block
/bin/docker: Running command: /usr/bin/chown -h ceph:ceph
/var/lib/ceph/osd/ceph-12/block
/bin/docker: Running command: /usr/bin/chown -R ceph:ceph /dev/dm-3
/bin/docker: Running command: /usr/bin/chown -R ceph:ceph
/var/lib/ceph/osd/ceph-12
/bin/docker: --> ceph-volume lvm activate successful for osd ID: 12
/bin/docker: --> Activating OSD ID 25 FSID
3f3d61f8-6964-4922-98cb-6620aff5cb6f
/bin/docker: Running command: /usr/bin/mount -t tmpfs tmpfs
/var/lib/ceph/osd/ceph-25
/bin/docker: Running command: /usr/bin/chown -R ceph:ceph
/var/lib/ceph/osd/ceph-25
/bin/docker: Running command: /usr/bin/ceph-bluestore-tool --cluster=ceph
prime-osd-dir --dev
/dev/ceph-6f69c24e-2930-48ff-a18f-278470e558e1/osd-block-3f3d61f8-6964-4922-98cb-6620aff5cb6f
--path /var/lib/ceph/osd/ceph-25 --no-mon-config
/bin/docker: Running command: /usr/bin/ln -snf
/dev/ceph-6f69c24e-2930-48ff-a18f-278470e558e1/osd-block-3f3d61f8-6964-4922-98cb-6620aff5cb6f
/var/lib/ceph/osd/ceph-25/block
/bin/docker: Running command: /usr/bin/chown -h ceph:ceph
/var/lib/ceph/osd/ceph-25/block
/bin/docker: Running command: /usr/bin/chown -R ceph:ceph /dev/dm-6
/bin/docker: Running command: /usr/bin/chown -R ceph:ceph
/var/lib/ceph/osd/ceph-25
/bin/docker: --> ceph-volume lvm activate successful for osd ID: 25
/bin/docker: --> Activating OSD ID 0 FSID
8d0b1bad-069a-4acf-b13b-982fab58f285
/bin/docker: Running command: /usr/bin/mount -t tmpfs tmpfs
/var/lib/ceph/osd/ceph-0
/bin/docker: Running command: /usr/bin/chown -R ceph:ceph
/var/lib/ceph/osd/ceph-0
/bin/docker: Running command: /usr/bin/ceph-bluestore-tool --cluster=ceph
prime-osd-dir --dev
/dev/ceph-8d9bd6eb-ce97-4940-8fed-f57f7bed3f5a/osd-block-8d0b1bad-069a-4acf-b13b-982fab58f285
--path /var/lib/ceph/osd/ceph-0 --no-mon-config
/bin/docker: Running command: /usr/bin/ln -snf
/dev/ceph-8d9bd6eb-ce97-4940-8fed-f57f7bed3f5a/osd-block-8d0b1bad-069a-4acf-b13b-982fab58f285
/var/lib/ceph/osd/ceph-0/block
/bin/docker: Running command: /usr/bin/chown -h ceph:ceph
/var/lib/ceph/osd/ceph-0/block
/bin/docker: Running command: /usr/bin/chown -R ceph:ceph /dev/dm-0
/bin/docker: Running command: /usr/bin/chown -R ceph:ceph
/var/lib/ceph/osd/ceph-0
/bin/docker: --> ceph-volume lvm activate successful for osd ID: 0
/bin/docker: --> Activating OSD ID 8 FSID
ff82d0d0-6d55-49c2-85cc-bb8a0a74ae89
/bin/docker: Running command: /usr/bin/mount -t tmpfs tmpfs
/var/lib/ceph/osd/ceph-8
/bin/docker: Running command: /usr/bin/chown -R ceph:ceph
/var/lib/ceph/osd/ceph-8
/bin/docker: Running command: /usr/bin/ceph-bluestore-tool --cluster=ceph
prime-osd-dir --dev
/dev/ceph-a8c3f39b-ec8d-4dd5-a85a-13a5773b99fa/osd-block-ff82d0d0-6d55-49c2-85cc-bb8a0a74ae89
--path /var/lib/ceph/osd/ceph-8 --no-mon-config
/bin/docker: Running command: /usr/bin/ln -snf
/dev/ceph-a8c3f39b-ec8d-4dd5-a85a-13a5773b99fa/osd-block-ff82d0d0-6d55-49c2-85cc-bb8a0a74ae89
/var/lib/ceph/osd/ceph-8/block
/bin/docker: Running command: /usr/bin/chown -h ceph:ceph
/var/lib/ceph/osd/ceph-8/block
/bin/docker: Running command: /usr/bin/chown -R ceph:ceph /dev/dm-2
/bin/docker: Running command: /usr/bin/chown -R ceph:ceph
/var/lib/ceph/osd/ceph-8
/bin/docker: --> ceph-volume lvm activate successful for osd ID: 8
/bin/docker: --> Activating OSD ID 4 FSID
313160a8-594d-4384-9640-68c4d8c1b6da
/bin/docker: Running command: /usr/bin/mount -t tmpfs tmpfs
/var/lib/ceph/osd/ceph-4
/bin/docker: Running command: /usr/bin/chown -R ceph:ceph
/var/lib/ceph/osd/ceph-4
/bin/docker: Running command: /usr/bin/ceph-bluestore-tool --cluster=ceph
prime-osd-dir --dev
/dev/ceph-bf94d5a5-eb04-42e3-867a-dee2886daa62/osd-block-313160a8-594d-4384-9640-68c4d8c1b6da
--path /var/lib/ceph/osd/ceph-4 --no-mon-config
/bin/docker: Running command: /usr/bin/ln -snf
/dev/ceph-bf94d5a5-eb04-42e3-867a-dee2886daa62/osd-block-313160a8-594d-4384-9640-68c4d8c1b6da
/var/lib/ceph/osd/ceph-4/block
/bin/docker: Running command: /usr/bin/chown -h ceph:ceph
/var/lib/ceph/osd/ceph-4/block
/bin/docker: Running command: /usr/bin/chown -R ceph:ceph /dev/dm-1
/bin/docker: Running command: /usr/bin/chown -R ceph:ceph
/var/lib/ceph/osd/ceph-4
/bin/docker: --> ceph-volume lvm activate successful for osd ID: 4
/bin/docker: --> Activating OSD ID 20 FSID
f7e67343-4fde-4e45-bc70-f44c92a178bd
/bin/docker: Running command: /usr/bin/mount -t tmpfs tmpfs
/var/lib/ceph/osd/ceph-20
/bin/docker: Running command: /usr/bin/chown -R ceph:ceph
/var/lib/ceph/osd/ceph-20
/bin/docker: Running command: /usr/bin/ceph-bluestore-tool --cluster=ceph
prime-osd-dir --dev
/dev/ceph-ee2517dc-ecb6-4c14-ab22-59d9c54f0952/osd-block-f7e67343-4fde-4e45-bc70-f44c92a178bd
--path /var/lib/ceph/osd/ceph-20 --no-mon-config
/bin/docker: Running command: /usr/bin/ln -snf
/dev/ceph-ee2517dc-ecb6-4c14-ab22-59d9c54f0952/osd-block-f7e67343-4fde-4e45-bc70-f44c92a178bd
/var/lib/ceph/osd/ceph-20/block
/bin/docker: Running command: /usr/bin/chown -h ceph:ceph
/var/lib/ceph/osd/ceph-20/block
/bin/docker: Running command: /usr/bin/chown -R ceph:ceph /dev/dm-5
/bin/docker: Running command: /usr/bin/chown -R ceph:ceph
/var/lib/ceph/osd/ceph-20
/bin/docker: --> ceph-volume lvm activate successful for osd ID: 20
/bin/docker: --> Activating OSD ID 16 FSID
078ece8f-1251-4482-b664-8f7068bc1685
/bin/docker: Running command: /usr/bin/mount -t tmpfs tmpfs
/var/lib/ceph/osd/ceph-16
/bin/docker: Running command: /usr/bin/chown -R ceph:ceph
/var/lib/ceph/osd/ceph-16
/bin/docker: Running command: /usr/bin/ceph-bluestore-tool --cluster=ceph
prime-osd-dir --dev
/dev/ceph-efb79afa-5c60-4a48-8419-259b72a5d73e/osd-block-078ece8f-1251-4482-b664-8f7068bc1685
--path /var/lib/ceph/osd/ceph-16 --no-mon-config
/bin/docker: Running command: /usr/bin/ln -snf
/dev/ceph-efb79afa-5c60-4a48-8419-259b72a5d73e/osd-block-078ece8f-1251-4482-b664-8f7068bc1685
/var/lib/ceph/osd/ceph-16/block
/bin/docker: Running command: /usr/bin/chown -h ceph:ceph
/var/lib/ceph/osd/ceph-16/block
/bin/docker: Running command: /usr/bin/chown -R ceph:ceph /dev/dm-4
/bin/docker: Running command: /usr/bin/chown -R ceph:ceph
/var/lib/ceph/osd/ceph-16
/bin/docker: --> ceph-volume lvm activate successful for osd ID: 16

On Wed, Apr 20, 2022 at 2:28 PM Eugen Block <eblock@xxxxxx> wrote:

IIUC it's just the arrow that can't be displayed when the systemd-unit
is enabled and the symlinks in
/etc/systemd/system/multi-user.target.wants/ are created. When OSDs
are created by cephadm the flag --no-systemd is usually used, can you
try this command?

cephadm ceph-volume lvm activate --all --no-systemd

Zitat von Manuel Holtgrewe <zyklenfrei@xxxxxxxxx>:

> Hi,
>
> thank you for your reply.
>
> That really is all. I tried to call `cephadm ceph-volume lvm activate
> --all`, see below, and this apparently crashes because of some unicode
> problem... might that be the root cause?
>
> Cheers,
> Manuel
>
> [root@dmz-host-4 rocky]# cephadm ceph-volume lvm activate --all
> Inferring fsid d221bc3c-8ff4-11ec-b4ba-b02628267680
> Using recent ceph image
>
quay.io/ceph/ceph@sha256:0d927ccbd8892180ee09894c2b2c26d07c938bf96a56eaee9b80fc9f26083ddb
> Non-zero exit code 1 from /bin/docker run --rm --ipc=host
> --stop-signal=SIGTERM --net=host --entrypoint /usr/sbin/ceph-volume
> --privileged --group-add=disk --init -e CONTAINER_IMAGE=
>
quay.io/ceph/ceph@sha256:0d927ccbd8892180ee09894c2b2c26d07c938bf96a56eaee9b80fc9f26083ddb
> -e NODE_NAME=dmz-host-4 -e CEPH_USE_RANDOM_NONCE=1 -v
> /var/run/ceph/d221bc3c-8ff4-11ec-b4ba-b02628267680:/var/run/ceph:z -v
> /var/log/ceph/d221bc3c-8ff4-11ec-b4ba-b02628267680:/var/log/ceph:z -v
>
/var/lib/ceph/d221bc3c-8ff4-11ec-b4ba-b02628267680/crash:/var/lib/ceph/crash:z
> -v /dev:/dev -v /run/udev:/run/udev -v /sys:/sys -v /run/lvm:/run/lvm -v
> /run/lock/lvm:/run/lock/lvm -v
/tmp/ceph-tmp1rkweygp:/etc/ceph/ceph.conf:z
>
quay.io/ceph/ceph@sha256:0d927ccbd8892180ee09894c2b2c26d07c938bf96a56eaee9b80fc9f26083ddb
> lvm activate --all
> /bin/docker: stderr --> Activating OSD ID 12 FSID
> e2ebb627-28aa-45a3-9261-d7c27bc08448
> /bin/docker: stderr Running command: /usr/bin/mount -t tmpfs tmpfs
> /var/lib/ceph/osd/ceph-12
> /bin/docker: stderr Running command: /usr/bin/chown -R ceph:ceph
> /var/lib/ceph/osd/ceph-12
> /bin/docker: stderr Running command: /usr/bin/ceph-bluestore-tool
> --cluster=ceph prime-osd-dir --dev
>
/dev/ceph-2cbf3973-13a3-444a-b335-a0262cff6074/osd-block-e2ebb627-28aa-45a3-9261-d7c27bc08448
> --path /var/lib/ceph/osd/ceph-12 --no-mon-config
> /bin/docker: stderr Running command: /usr/bin/ln -snf
>
/dev/ceph-2cbf3973-13a3-444a-b335-a0262cff6074/osd-block-e2ebb627-28aa-45a3-9261-d7c27bc08448
> /var/lib/ceph/osd/ceph-12/block
> /bin/docker: stderr Running command: /usr/bin/chown -h ceph:ceph
> /var/lib/ceph/osd/ceph-12/block
> /bin/docker: stderr Running command: /usr/bin/chown -R ceph:ceph
/dev/dm-3
> /bin/docker: stderr Running command: /usr/bin/chown -R ceph:ceph
> /var/lib/ceph/osd/ceph-12
> /bin/docker: stderr Running command: /usr/bin/systemctl enable
> ceph-volume@lvm-12-e2ebb627-28aa-45a3-9261-d7c27bc08448
> /bin/docker: stderr  stderr: Created symlink
>
/etc/systemd/system/multi-user.target.wants/ceph-volume@lvm-12-e2ebb627-28aa-45a3-9261-d7c27bc08448.service
> \u2192 /usr/lib/systemd/system/ceph-volume@.service.
> /bin/docker: stderr --- Logging error ---
> /bin/docker: stderr Traceback (most recent call last):
> /bin/docker: stderr   File "/usr/lib64/python3.6/logging/__init__.py",
line
> 996, in emit
> /bin/docker: stderr     stream.write(msg)
> /bin/docker: stderr UnicodeEncodeError: 'ascii' codec can't encode
> character '\u2192' in position 186: ordinal not in range(128)
> /bin/docker: stderr Call stack:
> /bin/docker: stderr   File "/usr/sbin/ceph-volume", line 11, in <module>
> /bin/docker: stderr     load_entry_point('ceph-volume==1.0.0',
> 'console_scripts', 'ceph-volume')()
> /bin/docker: stderr   File
> "/usr/lib/python3.6/site-packages/ceph_volume/main.py", line 40, in
__init__
> /bin/docker: stderr     self.main(self.argv)
> /bin/docker: stderr   File
> "/usr/lib/python3.6/site-packages/ceph_volume/decorators.py", line 59, in
> newfunc
> /bin/docker: stderr     return f(*a, **kw)
> /bin/docker: stderr   File
> "/usr/lib/python3.6/site-packages/ceph_volume/main.py", line 152, in main
> /bin/docker: stderr     terminal.dispatch(self.mapper, subcommand_args)
> /bin/docker: stderr   File
> "/usr/lib/python3.6/site-packages/ceph_volume/terminal.py", line 194, in
> dispatch
> /bin/docker: stderr     instance.main()
> /bin/docker: stderr   File
> "/usr/lib/python3.6/site-packages/ceph_volume/devices/lvm/main.py", line
> 46, in main
> /bin/docker: stderr     terminal.dispatch(self.mapper, self.argv)
> /bin/docker: stderr   File
> "/usr/lib/python3.6/site-packages/ceph_volume/terminal.py", line 194, in
> dispatch
> /bin/docker: stderr     instance.main()
> /bin/docker: stderr   File
> "/usr/lib/python3.6/site-packages/ceph_volume/devices/lvm/activate.py",
> line 373, in main
> /bin/docker: stderr     self.activate_all(args)
> /bin/docker: stderr   File
> "/usr/lib/python3.6/site-packages/ceph_volume/decorators.py", line 16, in
> is_root
> /bin/docker: stderr     return func(*a, **kw)
> /bin/docker: stderr   File
> "/usr/lib/python3.6/site-packages/ceph_volume/devices/lvm/activate.py",
> line 254, in activate_all
> /bin/docker: stderr     self.activate(args, osd_id=osd_id,
> osd_fsid=osd_fsid)
> /bin/docker: stderr   File
> "/usr/lib/python3.6/site-packages/ceph_volume/decorators.py", line 16, in
> is_root
> /bin/docker: stderr     return func(*a, **kw)
> /bin/docker: stderr   File
> "/usr/lib/python3.6/site-packages/ceph_volume/devices/lvm/activate.py",
> line 299, in activate
> /bin/docker: stderr     activate_bluestore(lvs, args.no_systemd)
> /bin/docker: stderr   File
> "/usr/lib/python3.6/site-packages/ceph_volume/devices/lvm/activate.py",
> line 214, in activate_bluestore
> /bin/docker: stderr     systemctl.enable_volume(osd_id, osd_fsid, 'lvm')
> /bin/docker: stderr   File
> "/usr/lib/python3.6/site-packages/ceph_volume/systemd/systemctl.py", line
> 82, in enable_volume
> /bin/docker: stderr     return enable(volume_unit % (device_type, id_,
> fsid))
> /bin/docker: stderr   File
> "/usr/lib/python3.6/site-packages/ceph_volume/systemd/systemctl.py", line
> 22, in enable
> /bin/docker: stderr     process.run(['systemctl', 'enable', unit])
> /bin/docker: stderr   File
> "/usr/lib/python3.6/site-packages/ceph_volume/process.py", line 137, in
run
> /bin/docker: stderr     log_descriptors(reads, process, terminal_logging)
> /bin/docker: stderr   File
> "/usr/lib/python3.6/site-packages/ceph_volume/process.py", line 59, in
> log_descriptors
> /bin/docker: stderr     log_output(descriptor_name, message,
> terminal_logging, True)
> /bin/docker: stderr   File
> "/usr/lib/python3.6/site-packages/ceph_volume/process.py", line 34, in
> log_output
> /bin/docker: stderr     logger.info(line)
> /bin/docker: stderr Message: 'stderr Created symlink
>
/etc/systemd/system/multi-user.target.wants/ceph-volume@lvm-12-e2ebb627-28aa-45a3-9261-d7c27bc08448.service
> \u2192 /usr/lib/systemd/system/ceph-volume@.service.'
> /bin/docker: stderr Arguments: ()
> /bin/docker: stderr Running command: /usr/bin/systemctl enable --runtime
> ceph-osd@12
> /bin/docker: stderr  stderr: Created symlink
> /run/systemd/system/ceph-osd.target.wants/ceph-osd@12.service \u2192
> /usr/lib/systemd/system/ceph-osd@.service.
> /bin/docker: stderr --- Logging error ---
> /bin/docker: stderr Traceback (most recent call last):
> /bin/docker: stderr   File "/usr/lib64/python3.6/logging/__init__.py",
line
> 996, in emit
> /bin/docker: stderr     stream.write(msg)
> /bin/docker: stderr UnicodeEncodeError: 'ascii' codec can't encode
> character '\u2192' in position 140: ordinal not in range(128)
> /bin/docker: stderr Call stack:
> /bin/docker: stderr   File "/usr/sbin/ceph-volume", line 11, in <module>
> /bin/docker: stderr     load_entry_point('ceph-volume==1.0.0',
> 'console_scripts', 'ceph-volume')()
> /bin/docker: stderr   File
> "/usr/lib/python3.6/site-packages/ceph_volume/main.py", line 40, in
__init__
> /bin/docker: stderr     self.main(self.argv)
> /bin/docker: stderr   File
> "/usr/lib/python3.6/site-packages/ceph_volume/decorators.py", line 59, in
> newfunc
> /bin/docker: stderr     return f(*a, **kw)
> /bin/docker: stderr   File
> "/usr/lib/python3.6/site-packages/ceph_volume/main.py", line 152, in main
> /bin/docker: stderr     terminal.dispatch(self.mapper, subcommand_args)
> /bin/docker: stderr   File
> "/usr/lib/python3.6/site-packages/ceph_volume/terminal.py", line 194, in
> dispatch
> /bin/docker: stderr     instance.main()
> /bin/docker: stderr   File
> "/usr/lib/python3.6/site-packages/ceph_volume/devices/lvm/main.py", line
> 46, in main
> /bin/docker: stderr     terminal.dispatch(self.mapper, self.argv)
> /bin/docker: stderr   File
> "/usr/lib/python3.6/site-packages/ceph_volume/terminal.py", line 194, in
> dispatch
> /bin/docker: stderr     instance.main()
> /bin/docker: stderr   File
> "/usr/lib/python3.6/site-packages/ceph_volume/devices/lvm/activate.py",
> line 373, in main
> /bin/docker: stderr     self.activate_all(args)
> /bin/docker: stderr   File
> "/usr/lib/python3.6/site-packages/ceph_volume/decorators.py", line 16, in
> is_root
> /bin/docker: stderr     return func(*a, **kw)
> /bin/docker: stderr   File
> "/usr/lib/python3.6/site-packages/ceph_volume/devices/lvm/activate.py",
> line 254, in activate_all
> /bin/docker: stderr     self.activate(args, osd_id=osd_id,
> osd_fsid=osd_fsid)
> /bin/docker: stderr   File
> "/usr/lib/python3.6/site-packages/ceph_volume/decorators.py", line 16, in
> is_root
> /bin/docker: stderr     return func(*a, **kw)
> /bin/docker: stderr   File
> "/usr/lib/python3.6/site-packages/ceph_volume/devices/lvm/activate.py",
> line 299, in activate
> /bin/docker: stderr     activate_bluestore(lvs, args.no_systemd)
> /bin/docker: stderr   File
> "/usr/lib/python3.6/site-packages/ceph_volume/devices/lvm/activate.py",
> line 217, in activate_bluestore
> /bin/docker: stderr     systemctl.enable_osd(osd_id)
> /bin/docker: stderr   File
> "/usr/lib/python3.6/site-packages/ceph_volume/systemd/systemctl.py", line
> 70, in enable_osd
> /bin/docker: stderr     return enable(osd_unit % id_, runtime=True)
> /bin/docker: stderr   File
> "/usr/lib/python3.6/site-packages/ceph_volume/systemd/systemctl.py", line
> 20, in enable
> /bin/docker: stderr     process.run(['systemctl', 'enable', '--runtime',
> unit])
> /bin/docker: stderr   File
> "/usr/lib/python3.6/site-packages/ceph_volume/process.py", line 137, in
run
> /bin/docker: stderr     log_descriptors(reads, process, terminal_logging)
> /bin/docker: stderr   File
> "/usr/lib/python3.6/site-packages/ceph_volume/process.py", line 59, in
> log_descriptors
> /bin/docker: stderr     log_output(descriptor_name, message,
> terminal_logging, True)
> /bin/docker: stderr   File
> "/usr/lib/python3.6/site-packages/ceph_volume/process.py", line 34, in
> log_output
> /bin/docker: stderr     logger.info(line)
> /bin/docker: stderr Message: 'stderr Created symlink
> /run/systemd/system/ceph-osd.target.wants/ceph-osd@12.service \u2192
> /usr/lib/systemd/system/ceph-osd@.service.'
> /bin/docker: stderr Arguments: ()
> /bin/docker: stderr Running command: /usr/bin/systemctl start ceph-osd@12
> /bin/docker: stderr  stderr: Failed to connect to bus: No such file or
> directory
> /bin/docker: stderr -->  RuntimeError: command returned non-zero exit
> status: 1
> Traceback (most recent call last):
>   File "/sbin/cephadm", line 8571, in <module>
>     main()
>   File "/sbin/cephadm", line 8559, in main
>     r = ctx.func(ctx)
>   File "/sbin/cephadm", line 1737, in _infer_config
>     return func(ctx)
>   File "/sbin/cephadm", line 1707, in _infer_fsid
>     return func(ctx)
>   File "/sbin/cephadm", line 1765, in _infer_image
>     return func(ctx)
>   File "/sbin/cephadm", line 1665, in _validate_fsid
>     return func(ctx)
>   File "/sbin/cephadm", line 4822, in command_ceph_volume
>     out, err, code = call_throws(ctx, c.run_cmd())
>   File "/sbin/cephadm", line 1467, in call_throws
>     raise RuntimeError('Failed command: %s' % ' '.join(command))
> RuntimeError: Failed command: /bin/docker run --rm --ipc=host
> --stop-signal=SIGTERM --net=host --entrypoint /usr/sbin/ceph-volume
> --privileged --group-add=disk --init -e CONTAINER_IMAGE=
>
quay.io/ceph/ceph@sha256:0d927ccbd8892180ee09894c2b2c26d07c938bf96a56eaee9b80fc9f26083ddb
> -e NODE_NAME=dmz-host-4 -e CEPH_USE_RANDOM_NONCE=1 -v
> /var/run/ceph/d221bc3c-8ff4-11ec-b4ba-b02628267680:/var/run/ceph:z -v
> /var/log/ceph/d221bc3c-8ff4-11ec-b4ba-b02628267680:/var/log/ceph:z -v
>
/var/lib/ceph/d221bc3c-8ff4-11ec-b4ba-b02628267680/crash:/var/lib/ceph/crash:z
> -v /dev:/dev -v /run/udev:/run/udev -v /sys:/sys -v /run/lvm:/run/lvm -v
> /run/lock/lvm:/run/lock/lvm -v
/tmp/ceph-tmp1rkweygp:/etc/ceph/ceph.conf:z
>
quay.io/ceph/ceph@sha256:0d927ccbd8892180ee09894c2b2c26d07c938bf96a56eaee9b80fc9f26083ddb
> lvm activate --all
>
> On Wed, Apr 20, 2022 at 1:26 PM Eugen Block <eblock@xxxxxx> wrote:
>
>> Hi,
>>
>> > and the relevant log output is here:
>> >
>>
https://privatepastebin.com/?f95c66924a7ddda9#ADEposX5DCo5fb5wGv42czrxaHscnwoHB7igc3eNQMwc
>>
>> This is just the output of 'ceph-volume lvm list', is that really all?
>> I haven't had the chance to test 'ceph cephadm osd activate' myself so
>> I can't really tell what to expect but apparently the LVs and OSDs are
>> present, cephadm just doesn't seem to try to do anything with them. Do
>> you see in the host's logs any attempt to start container for the OSDs?
>>
>> Zitat von Manuel Holtgrewe <zyklenfrei@xxxxxxxxx>:
>>
>> > Dear Eugen,
>> >
>> > thanks for the hint. The output is pasted below. I can't gather any
>> useful
>> > information from that.
>> >
>> > I also followed the instructions from
>> >
>> >
>>
https://docs.ceph.com/en/latest/cephadm/operations/#watching-cephadm-log-messages
>> >
>> > ```
>> > ceph config set mgr mgr/cephadm/log_to_cluster_level debug
>> > ceph -W cephadm --watch-debug
>> > ```
>> >
>> > and the relevant log output is here:
>> >
>> > -
>> >
>>
https://privatepastebin.com/?f95c66924a7ddda9#ADEposX5DCo5fb5wGv42czrxaHscnwoHB7igc3eNQMwc
>> >
>> > Cheers,
>> > Manuel
>> >
>> >
>> > [2022-04-20 10:38:01,909][ceph_volume.main][INFO  ] Running command:
>> > ceph-volume  lvm list --format json
>> > [2022-04-20 10:38:01,911][ceph_volume.process][INFO  ] Running
command:
>> > /usr/sbin/lvs --noheadings --readonly --separator=";" -a --units=b
>> > --nosuffix -S  -o lv_tags,lv_path,lv_name,vg_name,lv_uuid,lv_size
>> > [2022-04-20 10:38:01,975][ceph_volume.process][INFO  ] stdout
>> >
>>
ceph.block_device=/dev/ceph-2cbf3973-13a3-444a-b335-a0262cff6074/osd-block-e2ebb627-28aa-45a3-9261-d7c27bc08448,ceph.block_uuid=AbCFEt-JrjE-bfnc-HiHu-XlyH-G72j-0jwec7,ceph.cephx_lockbox_secret=,ceph.cluster_fsid=d221bc3c-8ff4-11ec-b4ba-b02628267680,ceph.cluster_name=ceph,ceph.crush_device_class=None,ceph.encrypted=0,ceph.osd_fsid=e2ebb627-28aa-45a3-9261-d7c27bc08448,ceph.osd_id=12,ceph.osdspec_affinity=all,ceph.type=block,ceph.vdo=0";"/dev/ceph-2cbf3973-13a3-444a-b335-a0262cff6074/osd-block-e2ebb627-28aa-45a3-9261-d7c27bc08448";"osd-block-e2ebb627-28aa-45a3-9261-d7c27bc08448";"ceph-2cbf3973-13a3-444a-b335-a0262cff6074";"AbCFEt-JrjE-bfnc-HiHu-XlyH-G72j-0jwec7";"1920378863616
>> > [2022-04-20 10:38:01,975][ceph_volume.process][INFO  ] stdout
>> >
>>
ceph.block_device=/dev/ceph-6f69c24e-2930-48ff-a18f-278470e558e1/osd-block-3f3d61f8-6964-4922-98cb-6620aff5cb6f,ceph.block_uuid=MS5ExP-ROK9-nVRG-l6wA-eHLL-Wjic-lMsYPM,ceph.cephx_lockbox_secret=,ceph.cluster_fsid=d221bc3c-8ff4-11ec-b4ba-b02628267680,ceph.cluster_name=ceph,ceph.crush_device_class=None,ceph.encrypted=0,ceph.osd_fsid=3f3d61f8-6964-4922-98cb-6620aff5cb6f,ceph.osd_id=25,ceph.osdspec_affinity=all,ceph.type=block,ceph.vdo=0";"/dev/ceph-6f69c24e-2930-48ff-a18f-278470e558e1/osd-block-3f3d61f8-6964-4922-98cb-6620aff5cb6f";"osd-block-3f3d61f8-6964-4922-98cb-6620aff5cb6f";"ceph-6f69c24e-2930-48ff-a18f-278470e558e1";"MS5ExP-ROK9-nVRG-l6wA-eHLL-Wjic-lMsYPM";"1920378863616
>> > [2022-04-20 10:38:01,975][ceph_volume.process][INFO  ] stdout
>> >
>>
ceph.block_device=/dev/ceph-8d9bd6eb-ce97-4940-8fed-f57f7bed3f5a/osd-block-8d0b1bad-069a-4acf-b13b-982fab58f285,ceph.block_uuid=rrJbzA-JSuc-GrSf-KsXd-VWi6-ffTW-LS5PB3,ceph.cephx_lockbox_secret=,ceph.cluster_fsid=d221bc3c-8ff4-11ec-b4ba-b02628267680,ceph.cluster_name=ceph,ceph.crush_device_class=None,ceph.encrypted=0,ceph.osd_fsid=8d0b1bad-069a-4acf-b13b-982fab58f285,ceph.osd_id=0,ceph.osdspec_affinity=all,ceph.type=block,ceph.vdo=0";"/dev/ceph-8d9bd6eb-ce97-4940-8fed-f57f7bed3f5a/osd-block-8d0b1bad-069a-4acf-b13b-982fab58f285";"osd-block-8d0b1bad-069a-4acf-b13b-982fab58f285";"ceph-8d9bd6eb-ce97-4940-8fed-f57f7bed3f5a";"rrJbzA-JSuc-GrSf-KsXd-VWi6-ffTW-LS5PB3";"1920378863616
>> > [2022-04-20 10:38:01,975][ceph_volume.process][INFO  ] stdout
>> >
>>
ceph.block_device=/dev/ceph-a8c3f39b-ec8d-4dd5-a85a-13a5773b99fa/osd-block-ff82d0d0-6d55-49c2-85cc-bb8a0a74ae89,ceph.block_uuid=CB1rmb-Onhq-Vdrh-9Ide-XpA0-IBg3-TFrSNX,ceph.cephx_lockbox_secret=,ceph.cluster_fsid=d221bc3c-8ff4-11ec-b4ba-b02628267680,ceph.cluster_name=ceph,ceph.crush_device_class=None,ceph.encrypted=0,ceph.osd_fsid=ff82d0d0-6d55-49c2-85cc-bb8a0a74ae89,ceph.osd_id=8,ceph.osdspec_affinity=all,ceph.type=block,ceph.vdo=0";"/dev/ceph-a8c3f39b-ec8d-4dd5-a85a-13a5773b99fa/osd-block-ff82d0d0-6d55-49c2-85cc-bb8a0a74ae89";"osd-block-ff82d0d0-6d55-49c2-85cc-bb8a0a74ae89";"ceph-a8c3f39b-ec8d-4dd5-a85a-13a5773b99fa";"CB1rmb-Onhq-Vdrh-9Ide-XpA0-IBg3-TFrSNX";"1920378863616
>> > [2022-04-20 10:38:01,975][ceph_volume.process][INFO  ] stdout
>> >
>>
ceph.block_device=/dev/ceph-bf94d5a5-eb04-42e3-867a-dee2886daa62/osd-block-313160a8-594d-4384-9640-68c4d8c1b6da,ceph.block_uuid=CgvBFd-wk3r-7a0y-XPiH-WoKS-VGmX-yntCse,ceph.cephx_lockbox_secret=,ceph.cluster_fsid=d221bc3c-8ff4-11ec-b4ba-b02628267680,ceph.cluster_name=ceph,ceph.crush_device_class=None,ceph.encrypted=0,ceph.osd_fsid=313160a8-594d-4384-9640-68c4d8c1b6da,ceph.osd_id=4,ceph.osdspec_affinity=all,ceph.type=block,ceph.vdo=0";"/dev/ceph-bf94d5a5-eb04-42e3-867a-dee2886daa62/osd-block-313160a8-594d-4384-9640-68c4d8c1b6da";"osd-block-313160a8-594d-4384-9640-68c4d8c1b6da";"ceph-bf94d5a5-eb04-42e3-867a-dee2886daa62";"CgvBFd-wk3r-7a0y-XPiH-WoKS-VGmX-yntCse";"1920378863616
>> > [2022-04-20 10:38:01,975][ceph_volume.process][INFO  ] stdout
>> >
>>
ceph.block_device=/dev/ceph-ee2517dc-ecb6-4c14-ab22-59d9c54f0952/osd-block-f7e67343-4fde-4e45-bc70-f44c92a178bd,ceph.block_uuid=pfWtmF-6Xlc-R2LO-kzeV-2jIw-3Ki8-gCOMwZ,ceph.cephx_lockbox_secret=,ceph.cluster_fsid=d221bc3c-8ff4-11ec-b4ba-b02628267680,ceph.cluster_name=ceph,ceph.crush_device_class=None,ceph.encrypted=0,ceph.osd_fsid=f7e67343-4fde-4e45-bc70-f44c92a178bd,ceph.osd_id=20,ceph.osdspec_affinity=all,ceph.type=block,ceph.vdo=0";"/dev/ceph-ee2517dc-ecb6-4c14-ab22-59d9c54f0952/osd-block-f7e67343-4fde-4e45-bc70-f44c92a178bd";"osd-block-f7e67343-4fde-4e45-bc70-f44c92a178bd";"ceph-ee2517dc-ecb6-4c14-ab22-59d9c54f0952";"pfWtmF-6Xlc-R2LO-kzeV-2jIw-3Ki8-gCOMwZ";"1920378863616
>> > [2022-04-20 10:38:01,975][ceph_volume.process][INFO  ] stdout
>> >
>>
ceph.block_device=/dev/ceph-efb79afa-5c60-4a48-8419-259b72a5d73e/osd-block-078ece8f-1251-4482-b664-8f7068bc1685,ceph.block_uuid=ufliE0-MMk9-nURa-GtC0-kcaP-mR8c-iHp7Wl,ceph.cephx_lockbox_secret=,ceph.cluster_fsid=d221bc3c-8ff4-11ec-b4ba-b02628267680,ceph.cluster_name=ceph,ceph.crush_device_class=None,ceph.encrypted=0,ceph.osd_fsid=078ece8f-1251-4482-b664-8f7068bc1685,ceph.osd_id=16,ceph.osdspec_affinity=all,ceph.type=block,ceph.vdo=0";"/dev/ceph-efb79afa-5c60-4a48-8419-259b72a5d73e/osd-block-078ece8f-1251-4482-b664-8f7068bc1685";"osd-block-078ece8f-1251-4482-b664-8f7068bc1685";"ceph-efb79afa-5c60-4a48-8419-259b72a5d73e";"ufliE0-MMk9-nURa-GtC0-kcaP-mR8c-iHp7Wl";"1920378863616
>> > [2022-04-20 10:38:01,976][ceph_volume.process][INFO  ] Running
command:
>> > /usr/sbin/pvs --noheadings --readonly --separator=";" -S
>> > lv_uuid=AbCFEt-JrjE-bfnc-HiHu-XlyH-G72j-0jwec7 -o
>> > pv_name,pv_tags,pv_uuid,vg_name,lv_uuid
>> > [2022-04-20 10:38:02,025][ceph_volume.process][INFO  ] stdout
>> >
>>
/dev/sdd";"";"hHdDtY-umyH-78lt-K3XC-qo7s-vGfj-SQiDLp";"ceph-2cbf3973-13a3-444a-b335-a0262cff6074";"AbCFEt-JrjE-bfnc-HiHu-XlyH-G72j-0jwec7
>> > [2022-04-20 10:38:02,026][ceph_volume.process][INFO  ] Running
command:
>> > /usr/sbin/pvs --noheadings --readonly --separator=";" -S
>> > lv_uuid=MS5ExP-ROK9-nVRG-l6wA-eHLL-Wjic-lMsYPM -o
>> > pv_name,pv_tags,pv_uuid,vg_name,lv_uuid
>> > [2022-04-20 10:38:02,088][ceph_volume.process][INFO  ] stdout
>> >
>>
/dev/sdg";"";"J5nuJt-ylF4-wG3Q-Twnj-Rw5h-PoIh-VHnJJq";"ceph-6f69c24e-2930-48ff-a18f-278470e558e1";"MS5ExP-ROK9-nVRG-l6wA-eHLL-Wjic-lMsYPM
>> > [2022-04-20 10:38:02,089][ceph_volume.process][INFO  ] Running
command:
>> > /usr/sbin/pvs --noheadings --readonly --separator=";" -S
>> > lv_uuid=rrJbzA-JSuc-GrSf-KsXd-VWi6-ffTW-LS5PB3 -o
>> > pv_name,pv_tags,pv_uuid,vg_name,lv_uuid
>> > [2022-04-20 10:38:02,146][ceph_volume.process][INFO  ] stdout
>> >
>>
/dev/sda";"";"EfXlSa-jzab-toFc-sim7-wh67-TY4D-vXwQyk";"ceph-8d9bd6eb-ce97-4940-8fed-f57f7bed3f5a";"rrJbzA-JSuc-GrSf-KsXd-VWi6-ffTW-LS5PB3
>> > [2022-04-20 10:38:02,147][ceph_volume.process][INFO  ] Running
command:
>> > /usr/sbin/pvs --noheadings --readonly --separator=";" -S
>> > lv_uuid=CB1rmb-Onhq-Vdrh-9Ide-XpA0-IBg3-TFrSNX -o
>> > pv_name,pv_tags,pv_uuid,vg_name,lv_uuid
>> > [2022-04-20 10:38:02,196][ceph_volume.process][INFO  ] stdout
>> >
>>
/dev/sdc";"";"1KJrej-zCY8-4Qaq-Ab10-MFQe-LRln-wjQeUj";"ceph-a8c3f39b-ec8d-4dd5-a85a-13a5773b99fa";"CB1rmb-Onhq-Vdrh-9Ide-XpA0-IBg3-TFrSNX
>> > [2022-04-20 10:38:02,197][ceph_volume.process][INFO  ] Running
command:
>> > /usr/sbin/pvs --noheadings --readonly --separator=";" -S
>> > lv_uuid=CgvBFd-wk3r-7a0y-XPiH-WoKS-VGmX-yntCse -o
>> > pv_name,pv_tags,pv_uuid,vg_name,lv_uuid
>> > [2022-04-20 10:38:02,252][ceph_volume.process][INFO  ] stdout
>> >
>>
/dev/sdb";"";"ceOUeN-zG4y-JMhj-Gi82-yo88-2fBY-dMD8Go";"ceph-bf94d5a5-eb04-42e3-867a-dee2886daa62";"CgvBFd-wk3r-7a0y-XPiH-WoKS-VGmX-yntCse
>> > [2022-04-20 10:38:02,252][ceph_volume.process][INFO  ] Running
command:
>> > /usr/sbin/pvs --noheadings --readonly --separator=";" -S
>> > lv_uuid=pfWtmF-6Xlc-R2LO-kzeV-2jIw-3Ki8-gCOMwZ -o
>> > pv_name,pv_tags,pv_uuid,vg_name,lv_uuid
>> > [2022-04-20 10:38:02,301][ceph_volume.process][INFO  ] stdout
>> >
>>
/dev/sdf";"";"j2Ilk4-12ZW-qR9u-3n5Y-gn6B-80yO-vWgdAn";"ceph-ee2517dc-ecb6-4c14-ab22-59d9c54f0952";"pfWtmF-6Xlc-R2LO-kzeV-2jIw-3Ki8-gCOMwZ
>> > [2022-04-20 10:38:02,302][ceph_volume.process][INFO  ] Running
command:
>> > /usr/sbin/pvs --noheadings --readonly --separator=";" -S
>> > lv_uuid=ufliE0-MMk9-nURa-GtC0-kcaP-mR8c-iHp7Wl -o
>> > pv_name,pv_tags,pv_uuid,vg_name,lv_uuid
>> > [2022-04-20 10:38:02,355][ceph_volume.process][INFO  ] stdout
>> >
>>
/dev/sde";"";"Ym7N4c-IBxm-cK3l-EOdZ-oN8c-gFjz-Nj0rVT";"ceph-efb79afa-5c60-4a48-8419-259b72a5d73e";"ufliE0-MMk9-nURa-GtC0-kcaP-mR8c-iHp7Wl
>> >
>> > On Wed, Apr 20, 2022 at 12:28 PM Eugen Block <eblock@xxxxxx> wrote:
>> >
>> >> Hi,
>> >>
>> >> have you checked /var/log/ceph/cephadm.log for any hints?
>> >> ceph-volume.log may also provide some information
>> >> (/var/log/ceph/<FSID>/ceph-volume.log) what might be going on.
>> >>
>> >> Zitat von Manuel Holtgrewe <zyklenfrei@xxxxxxxxx>:
>> >>
>> >> > Dear all,
>> >> >
>> >> > I now attempted this and my host is back in the cluster but the
`ceph
>> >> > cephadm osd activate` does not work.
>> >> >
>> >> > # ceph cephadm osd activate HOST
>> >> > Created no osd(s) on host HOST; already created?
>> >> >
>> >> > Using --verbose is not too helpful either:
>> >> >
>> >> > bestcmds_sorted:
>> >> > [{'flags': 8,
>> >> >   'help': 'Start OSD containers for existing OSDs',
>> >> >   'module': 'mgr',
>> >> >   'perm': 'rw',
>> >> >   'sig': [argdesc(<class 'ceph_argparse.CephPrefix'>, req=True,
>> >> > name=prefix, n=1, numseen=0, prefix=cephadm),
>> >> >           argdesc(<class 'ceph_argparse.CephPrefix'>, req=True,
>> >> > name=prefix, n=1, numseen=0, prefix=osd),
>> >> >           argdesc(<class 'ceph_argparse.CephPrefix'>, req=True,
>> >> > name=prefix, n=1, numseen=0, prefix=activate),
>> >> >           argdesc(<class 'ceph_argparse.CephString'>, req=True,
>> >> name=host,
>> >> > n=N, numseen=0)]}]
>> >> > Submitting command:  {'prefix': 'cephadm osd activate', 'host':
>> ['HOST'],
>> >> > 'target': ('mon-mgr', '')}
>> >> > submit {"prefix": "cephadm osd activate", "host": ["HOST"],
"target":
>> >> > ["mon-mgr", ""]} to mon-mgr
>> >> >
>> >> > How can I find out what goes wrong here?
>> >> >
>> >> > Thanks!
>> >> > Manuel
>> >> >
>> >> > On Wed, Feb 2, 2022 at 12:43 PM Robert Sander <
>> >> r.sander@xxxxxxxxxxxxxxxxxxx>
>> >> > wrote:
>> >> >
>> >> >> On 02.02.22 12:15, Manuel Holtgrewe wrote:
>> >> >> >
>> >> >> > Would this also work when renaming hosts at the same time?
>> >> >> >
>> >> >> > - remove host from ceph orch
>> >> >> > - reinstall host with different name/IP
>> >> >> > - add back host into ceph orch
>> >> >> > - use ceph osd activate
>> >> >> >
>> >> >> > as above?
>> >> >>
>> >> >> That could also work as long as the OSDs are still in the CRUSH
map.
>> >> >> Keep in mind that the last command is
>> >> >>
>> >> >>    ceph cephadm osd activate $HOSTNAME
>> >> >>
>> >> >> Regards
>> >> >> --
>> >> >> Robert Sander
>> >> >> Heinlein Consulting GmbH
>> >> >> Schwedter Str. 8/9b, 10119 Berlin
>> >> >>
>> >> >> https://www.heinlein-support.de
>> >> >>
>> >> >> Tel: 030 / 405051-43
>> >> >> Fax: 030 / 405051-19
>> >> >>
>> >> >> Amtsgericht Berlin-Charlottenburg - HRB 220009 B
>> >> >> Geschäftsführer: Peer Heinlein - Sitz: Berlin
>> >> >> _______________________________________________
>> >> >> ceph-users mailing list -- ceph-users@xxxxxxx
>> >> >> To unsubscribe send an email to ceph-users-leave@xxxxxxx
>> >> >>
>> >> > _______________________________________________
>> >> > ceph-users mailing list -- ceph-users@xxxxxxx
>> >> > To unsubscribe send an email to ceph-users-leave@xxxxxxx
>> >>
>> >>
>> >>
>> >> _______________________________________________
>> >> ceph-users mailing list -- ceph-users@xxxxxxx
>> >> To unsubscribe send an email to ceph-users-leave@xxxxxxx
>> >>
>>
>>
>>
>>








_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux