Re: Failed to probe daemons or devices

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



I've created an issue: https://tracker.ceph.com/issues/57918
What can I do more the get to fix this issue?

And the output of the requested commands
[cephadm@mdshost2 ~]$ sudo lvs -a
  LV VG Attr LSize Pool Origin Data% Meta% Move Log Cpy%Sync Convert
  lv_home vg_sys -wi-ao---- 256.00m
  lv_opt vg_sys -wi-ao---- 3.00g
  lv_root vg_sys -wi-ao---- 5.00g
  lv_swap vg_sys -wi-ao---- 7.56g
  lv_tmp vg_sys -wi-ao---- 1.00g
  lv_var vg_sys -wi-ao---- 15.00g
  lv_var_log vg_sys -wi-ao---- 5.00g
  lv_var_log_audit vg_sys -wi-ao---- 512.00m

[cephadm@mdshost2 ~]$ sudo vgs -a
  VG #PV #LV #SN Attr VSize VFree
  vg_sys 1 8 0 wz--n- <49.00g 11.68g

[cephadm@mdshost2 ~]$ sudo parted --list
Model: VMware Virtual disk (scsi)
Disk /dev/sda: 53.7GB
Sector size (logical/physical): 512B/512B
Partition Table: msdos
Disk Flags:

Number Start End Size Type File system Flags
 1 1049kB 1075MB 1074MB primary xfs boot
 2 1075MB 53.7GB 52.6GB primary lvm

Error: /dev/sdb: unrecognised disk label
Model: VMware Virtual disk (scsi)
Disk /dev/sdb: 53.7GB
Sector size (logical/physical): 512B/512B
Partition Table: unknown
Disk Flags:
________________________________
From: Guillaume Abrioux <gabrioux@xxxxxxxxxx>
Sent: Monday, October 24, 2022 5:50:20 PM
To: Sake Paulusma <sake1989@xxxxxxxxxxx>
Cc: ceph-users@xxxxxxx <ceph-users@xxxxxxx>
Subject: Re:  Failed to probe daemons or devices

Hello Sake,

Could you share the output of vgs / lvs commands?
Also, I would suggest you to open a tracker [1]

Thanks!

[1] https://tracker.ceph.com/projects/ceph-volume<https://nam12.safelinks.protection.outlook.com/?url=https%3A%2F%2Ftracker.ceph.com%2Fprojects%2Fceph-volume&data=05%7C01%7C%7C8cf7f4f7348f4560917308dab5d77e21%7C84df9e7fe9f640afb435aaaaaaaaaaaa%7C1%7C0%7C638022234378380450%7CUnknown%7CTWFpbGZsb3d8eyJWIjoiMC4wLjAwMDAiLCJQIjoiV2luMzIiLCJBTiI6Ik1haWwiLCJXVCI6Mn0%3D%7C3000%7C%7C%7C&sdata=QZh7xliaxqBOtNGgJsfUo2OGZV%2FzOI9cdNotTNT%2Bs%2BU%3D&reserved=0>

On Mon, 24 Oct 2022 at 10:51, Sake Paulusma <sake1989@xxxxxxxxxxx<mailto:sake1989@xxxxxxxxxxx>> wrote:
Last friday I upgrade the Ceph cluster from 17.2.3 to 17.2.5 with "ceph orch upgrade start --image localcontainerregistry.local.com:5000/ceph/ceph:v17.2.5-20221017<https://nam12.safelinks.protection.outlook.com/?url=http%3A%2F%2Flocalcontainerregistry.local.com%3A5000%2Fceph%2Fceph%3Av17.2.5-20221017&data=05%7C01%7C%7C8cf7f4f7348f4560917308dab5d77e21%7C84df9e7fe9f640afb435aaaaaaaaaaaa%7C1%7C0%7C638022234378380450%7CUnknown%7CTWFpbGZsb3d8eyJWIjoiMC4wLjAwMDAiLCJQIjoiV2luMzIiLCJBTiI6Ik1haWwiLCJXVCI6Mn0%3D%7C3000%7C%7C%7C&sdata=GKhA934ODD%2FQEHdd97vLskLkEhJuekIfUjBqx4wRlGU%3D&reserved=0>". After sometime, an hour?, I've got a health warning: CEPHADM_REFRESH_FAILED: failed to probe daemons or devices. I'm using only Cephfs on the cluster and it's still working correctly.
Checking the running services, everything is up and running; mon, osd and mds. But on the hosts running mon and mds services I get errors in the cephadm.log, see the loglines below.

I look likes cephadm tries to start a container for checking something? What could be wrong?


On mon nodes I got the following:
2022-10-24 10:31:43,880 7f179e5bfb80 DEBUG --------------------------------------------------------------------------------
cephadm ['gather-facts']
2022-10-24 10:31:44,333 7fc2d52b6b80 DEBUG --------------------------------------------------------------------------------
cephadm ['--image', 'localcontainerregistry.local.com:5000/ceph/ceph@sha256:122436e2f1df0c803666c5591db4a9b6c9196a71b4d44c6bd5d18102509dfca0<https://nam12.safelinks.protection.outlook.com/?url=http%3A%2F%2Flocalcontainerregistry.local.com%3A5000%2Fceph%2Fceph%40sha256%3A122436e2f1df0c803666c5591db4a9b6c9196a71b4d44c6bd5d18102509dfca0&data=05%7C01%7C%7C8cf7f4f7348f4560917308dab5d77e21%7C84df9e7fe9f640afb435aaaaaaaaaaaa%7C1%7C0%7C638022234378380450%7CUnknown%7CTWFpbGZsb3d8eyJWIjoiMC4wLjAwMDAiLCJQIjoiV2luMzIiLCJBTiI6Ik1haWwiLCJXVCI6Mn0%3D%7C3000%7C%7C%7C&sdata=uNhTCkSwBrcq%2FZNFF6TgKHy5nmC%2FDM5ly3QMIEl8Z8I%3D&reserved=0>', 'ceph-volume', '--fsid', '8909ef90-22ea-11ed-8df1-0050569ee1b1', '--', 'inventory', '--format=json-pretty', '--filter-for-batch']
2022-10-24 10:31:44,663 7fc2d52b6b80 INFO Inferring config /var/lib/ceph/8909ef90-22ea-11ed-8df1-0050569ee1b1/mon.oqsoel24332/config
2022-10-24 10:31:44,663 7fc2d52b6b80 DEBUG Using specified fsid: 8909ef90-22ea-11ed-8df1-0050569ee1b1
2022-10-24 10:31:45,574 7fc2d52b6b80 INFO Non-zero exit code 1 from /bin/podman run --rm --ipc=host --stop-signal=SIGTERM --authfile=/etc/ceph/podman-auth.json --net=host --entrypoint /usr/sbin/ceph-volume --privileged --group-add=disk --init -e CONTAINER_IMAGE=localcontainerregistry.local.com:5000/ceph/ceph@sha256:122436e2f1df0c803666c5591db4a9b6c9196a71b4d44c6bd5d18102509dfca0<https://nam12.safelinks.protection.outlook.com/?url=http%3A%2F%2Flocalcontainerregistry.local.com%3A5000%2Fceph%2Fceph%40sha256%3A122436e2f1df0c803666c5591db4a9b6c9196a71b4d44c6bd5d18102509dfca0&data=05%7C01%7C%7C8cf7f4f7348f4560917308dab5d77e21%7C84df9e7fe9f640afb435aaaaaaaaaaaa%7C1%7C0%7C638022234378380450%7CUnknown%7CTWFpbGZsb3d8eyJWIjoiMC4wLjAwMDAiLCJQIjoiV2luMzIiLCJBTiI6Ik1haWwiLCJXVCI6Mn0%3D%7C3000%7C%7C%7C&sdata=uNhTCkSwBrcq%2FZNFF6TgKHy5nmC%2FDM5ly3QMIEl8Z8I%3D&reserved=0> -e NODE_NAME=monnode2.local.com<https://nam12.safelinks.protection.outlook.com/?url=http%3A%2F%2Fmonnode2.local.com%2F&data=05%7C0
 1%7C%7C8cf7f4f7348f4560917308dab5d77e21%7C84df9e7fe9f640afb435aaaaaaaaaaaa%7C1%7C0%7C638022234378692872%7CUnknown%7CTWFpbGZsb3d8eyJWIjoiMC4wLjAwMDAiLCJQIjoiV2luMzIiLCJBTiI6Ik1haWwiLCJXVCI6Mn0%3D%7C3000%7C%7C%7C&sdata=DhsHCxFBck3TkzM%2FBnKFArbX13pWP0x2fowL9NoC7c0%3D&reserved=0> -e CEPH_USE_RANDOM_NONCE=1 -e CEPH_VOLUME_SKIP_RESTORECON=yes -e CEPH_VOLUME_DEBUG=1 -v /var/run/ceph/8909ef90-22ea-11ed-8df1-0050569ee1b1:/var/run/ceph:z -v /var/log/ceph/8909ef90-22ea-11ed-8df1-0050569ee1b1:/var/log/ceph:z -v /var/lib/ceph/8909ef90-22ea-11ed-8df1-0050569ee1b1/crash:/var/lib/ceph/crash:z -v /run/systemd/journal:/run/systemd/journal -v /dev:/dev -v /run/udev:/run/udev -v /sys:/sys -v /run/lvm:/run/lvm -v /run/lock/lvm:/run/lock/lvm -v /var/lib/ceph/8909ef90-22ea-11ed-8df1-0050569ee1b1/selinux:/sys/fs/selinux:ro -v /:/rootfs -v /tmp/ceph-tmp31tx1iy2:/etc/ceph/ce
 ph.conf:z localcontainerregistry.local.com:5000/ceph/ceph@sha256:122436e2f1df0c803666c5591db4a9b6c9196a71b4d44c6bd5d18102509dfca0<https://nam12.safelinks.protection.outlook.com/?url=http%3A%2F%2Flocalcontainerregistry.local.com%3A5000%2Fceph%2Fceph%40sha256%3A122436e2f1df0c803666c5591db4a9b6c9196a71b4d44c6bd5d18102509dfca0&data=05%7C01%7C%7C8cf7f4f7348f4560917308dab5d77e21%7C84df9e7fe9f640afb435aaaaaaaaaaaa%7C1%7C0%7C638022234378692872%7CUnknown%7CTWFpbGZsb3d8eyJWIjoiMC4wLjAwMDAiLCJQIjoiV2luMzIiLCJBTiI6Ik1haWwiLCJXVCI6Mn0%3D%7C3000%7C%7C%7C&sdata=pVgOytiFU6r13Pn1pr30uTRiYAoJNBv%2F9yQtaeNvXwQ%3D&reserved=0> inventory --format=json-pretty --filter-for-batch
2022-10-24 10:31:45,575 7fc2d52b6b80 INFO /bin/podman: stderr Traceback (most recent call last):
2022-10-24 10:31:45,575 7fc2d52b6b80 INFO /bin/podman: stderr File "/usr/sbin/ceph-volume", line 11, in <module>
2022-10-24 10:31:45,575 7fc2d52b6b80 INFO /bin/podman: stderr load_entry_point('ceph-volume==1.0.0', 'console_scripts', 'ceph-volume')()
2022-10-24 10:31:45,575 7fc2d52b6b80 INFO /bin/podman: stderr File "/usr/lib/python3.6/site-packages/ceph_volume/main.py", line 41, in __init__
2022-10-24 10:31:45,575 7fc2d52b6b80 INFO /bin/podman: stderr self.main(self.argv)
2022-10-24 10:31:45,575 7fc2d52b6b80 INFO /bin/podman: stderr File "/usr/lib/python3.6/site-packages/ceph_volume/decorators.py", line 59, in newfunc
2022-10-24 10:31:45,575 7fc2d52b6b80 INFO /bin/podman: stderr return f(*a, **kw)
2022-10-24 10:31:45,575 7fc2d52b6b80 INFO /bin/podman: stderr File "/usr/lib/python3.6/site-packages/ceph_volume/main.py", line 153, in main
2022-10-24 10:31:45,575 7fc2d52b6b80 INFO /bin/podman: stderr terminal.dispatch(self.mapper, subcommand_args)
2022-10-24 10:31:45,575 7fc2d52b6b80 INFO /bin/podman: stderr File "/usr/lib/python3.6/site-packages/ceph_volume/terminal.py", line 194, in dispatch
2022-10-24 10:31:45,575 7fc2d52b6b80 INFO /bin/podman: stderr instance.main()
2022-10-24 10:31:45,575 7fc2d52b6b80 INFO /bin/podman: stderr File "/usr/lib/python3.6/site-packages/ceph_volume/inventory/main.py", line 53, in main
2022-10-24 10:31:45,575 7fc2d52b6b80 INFO /bin/podman: stderr with_lsm=self.args.with_lsm))
2022-10-24 10:31:45,575 7fc2d52b6b80 INFO /bin/podman: stderr File "/usr/lib/python3.6/site-packages/ceph_volume/util/device.py", line 39, in __init__
2022-10-24 10:31:45,575 7fc2d52b6b80 INFO /bin/podman: stderr all_devices_vgs = lvm.get_all_devices_vgs()
2022-10-24 10:31:45,575 7fc2d52b6b80 INFO /bin/podman: stderr File "/usr/lib/python3.6/site-packages/ceph_volume/api/lvm.py", line 797, in get_all_devices_vgs
2022-10-24 10:31:45,575 7fc2d52b6b80 INFO /bin/podman: stderr return [VolumeGroup(**vg) for vg in vgs]
2022-10-24 10:31:45,575 7fc2d52b6b80 INFO /bin/podman: stderr File "/usr/lib/python3.6/site-packages/ceph_volume/api/lvm.py", line 797, in <listcomp>
2022-10-24 10:31:45,575 7fc2d52b6b80 INFO /bin/podman: stderr return [VolumeGroup(**vg) for vg in vgs]
2022-10-24 10:31:45,575 7fc2d52b6b80 INFO /bin/podman: stderr File "/usr/lib/python3.6/site-packages/ceph_volume/api/lvm.py", line 517, in __init__
2022-10-24 10:31:45,575 7fc2d52b6b80 INFO /bin/podman: stderr raise ValueError('VolumeGroup must have a non-empty name')
2022-10-24 10:31:45,575 7fc2d52b6b80 INFO /bin/podman: stderr ValueError: VolumeGroup must have a non-empty name


On mds nodes I got the following:
2022-10-24 10:25:18,506 7f613f6fdb80 DEBUG --------------------------------------------------------------------------------
cephadm ['gather-facts']
2022-10-24 10:25:19,047 7fd9b0d92b80 DEBUG --------------------------------------------------------------------------------
cephadm ['--image', 'localcontainerregistry.local.com:5000/ceph/ceph@sha256:122436e2f1df0c803666c5591db4a9b6c9196a71b4d44c6bd5d18102509dfca0<https://nam12.safelinks.protection.outlook.com/?url=http%3A%2F%2Flocalcontainerregistry.local.com%3A5000%2Fceph%2Fceph%40sha256%3A122436e2f1df0c803666c5591db4a9b6c9196a71b4d44c6bd5d18102509dfca0&data=05%7C01%7C%7C8cf7f4f7348f4560917308dab5d77e21%7C84df9e7fe9f640afb435aaaaaaaaaaaa%7C1%7C0%7C638022234378692872%7CUnknown%7CTWFpbGZsb3d8eyJWIjoiMC4wLjAwMDAiLCJQIjoiV2luMzIiLCJBTiI6Ik1haWwiLCJXVCI6Mn0%3D%7C3000%7C%7C%7C&sdata=pVgOytiFU6r13Pn1pr30uTRiYAoJNBv%2F9yQtaeNvXwQ%3D&reserved=0>', 'ceph-volume', '--fsid', '8909ef90-22ea-11ed-8df1-0050569ee1b1', '--', 'inventory', '--format=json-pretty', '--filter-for-batch']
2022-10-24 10:25:19,388 7fd9b0d92b80 DEBUG Using specified fsid: 8909ef90-22ea-11ed-8df1-0050569ee1b1
2022-10-24 10:25:20,306 7fd9b0d92b80 INFO Non-zero exit code 1 from /bin/podman run --rm --ipc=host --stop-signal=SIGTERM --authfile=/etc/ceph/podman-auth.json --net=host --entrypoint /usr/sbin/ceph-volume --privileged --group-add=disk --init -e CONTAINER_IMAGE=localcontainerregistry.local.com:5000/ceph/ceph@sha256:122436e2f1df0c803666c5591db4a9b6c9196a71b4d44c6bd5d18102509dfca0<https://nam12.safelinks.protection.outlook.com/?url=http%3A%2F%2Flocalcontainerregistry.local.com%3A5000%2Fceph%2Fceph%40sha256%3A122436e2f1df0c803666c5591db4a9b6c9196a71b4d44c6bd5d18102509dfca0&data=05%7C01%7C%7C8cf7f4f7348f4560917308dab5d77e21%7C84df9e7fe9f640afb435aaaaaaaaaaaa%7C1%7C0%7C638022234378849100%7CUnknown%7CTWFpbGZsb3d8eyJWIjoiMC4wLjAwMDAiLCJQIjoiV2luMzIiLCJBTiI6Ik1haWwiLCJXVCI6Mn0%3D%7C3000%7C%7C%7C&sdata=1BYancA6xzUHVEJtP9dX8%2B75BXdYsDE173TgxsXzBO8%3D&reserved=0> -e NODE_NAME=mdsnode1.local.com<https://nam12.safelinks.protection.outlook.com/?url=http%3A%2F%2Fmdsnode1.local.com%2F&data=05%7C01%
 7C%7C8cf7f4f7348f4560917308dab5d77e21%7C84df9e7fe9f640afb435aaaaaaaaaaaa%7C1%7C0%7C638022234378849100%7CUnknown%7CTWFpbGZsb3d8eyJWIjoiMC4wLjAwMDAiLCJQIjoiV2luMzIiLCJBTiI6Ik1haWwiLCJXVCI6Mn0%3D%7C3000%7C%7C%7C&sdata=sJgN8wZU6s21P2RObYsWsWkBuz8N9cIYNU6y3JNr05s%3D&reserved=0> -e CEPH_USE_RANDOM_NONCE=1 -e CEPH_VOLUME_SKIP_RESTORECON=yes -e CEPH_VOLUME_DEBUG=1 -v /var/run/ceph/8909ef90-22ea-11ed-8df1-0050569ee1b1:/var/run/ceph:z -v /var/log/ceph/8909ef90-22ea-11ed-8df1-0050569ee1b1:/var/log/ceph:z -v /var/lib/ceph/8909ef90-22ea-11ed-8df1-0050569ee1b1/crash:/var/lib/ceph/crash:z -v /run/systemd/journal:/run/systemd/journal -v /dev:/dev -v /run/udev:/run/udev -v /sys:/sys -v /run/lvm:/run/lvm -v /run/lock/lvm:/run/lock/lvm -v /var/lib/ceph/8909ef90-22ea-11ed-8df1-0050569ee1b1/selinux:/sys/fs/selinux:ro -v /:/rootfs localcontainerregistry.local.com:5000<https://nam12.safelinks.protection.outlook.com/?url=http%3A%2F%2Flocalcontainerregistry.local.com%3A5000%2F&data=05%7C01%7C%7C8cf7f4f7348f
 4560917308dab5d77e21%7C84df9e7fe9f640afb435aaaaaaaaaaaa%7C1%7C0%7C638022234378849100%7CUnknown%7CTWFpbGZsb3d8eyJWIjoiMC4wLjAwMDAiLCJQIjoiV2luMzIiLCJBTiI6Ik1haWwiLCJXVCI6Mn0%3D%7C3000%7C%7C%7C&sdata=g6X8qiHxuAFnVN7dWYUF%2BvzJlxlfJdLTBMewDjaVX5A%3D&reserved=0>
 /ceph/ceph@sha256:122436e2f1df0c803666c5591db4a9b6c9196a71b4d44c6bd5d18102509dfca0 inventory --format=json-pretty --filter-for-batch
2022-10-24 10:25:20,306 7fd9b0d92b80 INFO /bin/podman: stderr Traceback (most recent call last):
2022-10-24 10:25:20,306 7fd9b0d92b80 INFO /bin/podman: stderr File "/usr/sbin/ceph-volume", line 11, in <module>
2022-10-24 10:25:20,306 7fd9b0d92b80 INFO /bin/podman: stderr load_entry_point('ceph-volume==1.0.0', 'console_scripts', 'ceph-volume')()
2022-10-24 10:25:20,306 7fd9b0d92b80 INFO /bin/podman: stderr File "/usr/lib/python3.6/site-packages/ceph_volume/main.py", line 41, in __init__
2022-10-24 10:25:20,307 7fd9b0d92b80 INFO /bin/podman: stderr self.main(self.argv)
2022-10-24 10:25:20,307 7fd9b0d92b80 INFO /bin/podman: stderr File "/usr/lib/python3.6/site-packages/ceph_volume/decorators.py", line 59, in newfunc
2022-10-24 10:25:20,307 7fd9b0d92b80 INFO /bin/podman: stderr return f(*a, **kw)
2022-10-24 10:25:20,307 7fd9b0d92b80 INFO /bin/podman: stderr File "/usr/lib/python3.6/site-packages/ceph_volume/main.py", line 153, in main
2022-10-24 10:25:20,307 7fd9b0d92b80 INFO /bin/podman: stderr terminal.dispatch(self.mapper, subcommand_args)
2022-10-24 10:25:20,307 7fd9b0d92b80 INFO /bin/podman: stderr File "/usr/lib/python3.6/site-packages/ceph_volume/terminal.py", line 194, in dispatch
2022-10-24 10:25:20,307 7fd9b0d92b80 INFO /bin/podman: stderr instance.main()
2022-10-24 10:25:20,307 7fd9b0d92b80 INFO /bin/podman: stderr File "/usr/lib/python3.6/site-packages/ceph_volume/inventory/main.py", line 53, in main
2022-10-24 10:25:20,307 7fd9b0d92b80 INFO /bin/podman: stderr with_lsm=self.args.with_lsm))
2022-10-24 10:25:20,307 7fd9b0d92b80 INFO /bin/podman: stderr File "/usr/lib/python3.6/site-packages/ceph_volume/util/device.py", line 39, in __init__
2022-10-24 10:25:20,307 7fd9b0d92b80 INFO /bin/podman: stderr all_devices_vgs = lvm.get_all_devices_vgs()
2022-10-24 10:25:20,307 7fd9b0d92b80 INFO /bin/podman: stderr File "/usr/lib/python3.6/site-packages/ceph_volume/api/lvm.py", line 797, in get_all_devices_vgs
2022-10-24 10:25:20,307 7fd9b0d92b80 INFO /bin/podman: stderr return [VolumeGroup(**vg) for vg in vgs]
2022-10-24 10:25:20,307 7fd9b0d92b80 INFO /bin/podman: stderr File "/usr/lib/python3.6/site-packages/ceph_volume/api/lvm.py", line 797, in <listcomp>
2022-10-24 10:25:20,307 7fd9b0d92b80 INFO /bin/podman: stderr return [VolumeGroup(**vg) for vg in vgs]
2022-10-24 10:25:20,307 7fd9b0d92b80 INFO /bin/podman: stderr File "/usr/lib/python3.6/site-packages/ceph_volume/api/lvm.py", line 517, in __init__
2022-10-24 10:25:20,307 7fd9b0d92b80 INFO /bin/podman: stderr raise ValueError('VolumeGroup must have a non-empty name')
2022-10-24 10:25:20,307 7fd9b0d92b80 INFO /bin/podman: stderr ValueError: VolumeGroup must have a non-empty name

_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx<mailto:ceph-users@xxxxxxx>
To unsubscribe send an email to ceph-users-leave@xxxxxxx<mailto:ceph-users-leave@xxxxxxx>



--
Guillaume Abrioux
Senior Software Engineer
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux