cephadm ls / ceph orch ps => here does it get its information?

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hello,

i seem to not have removed old osd. Now i have:

root@ceph07:/tmp# ceph orch ps |grep -e error -e stopped |grep ceph07
_osd.33                   ceph07  stopped       2h ago     2y
<unknown>  quay.io/ceph/ceph:v15.2.17  <unknown>     <unknown>
mon.ceph01                ceph07  error         2h ago     2y
<unknown>  quay.io/ceph/ceph:v15.2.17  <unknown>     <unknown>
osd.0                     ceph07  error         2h ago     2y
<unknown>  quay.io/ceph/ceph:v15.2.17  <unknown>     <unknown>
osd.1                     ceph07  error         2h ago     2y
<unknown>  quay.io/ceph/ceph:v15.2.17  <unknown>     <unknown>
osd.11                    ceph07  error         2h ago     2y
<unknown>  quay.io/ceph/ceph:v15.2.17  <unknown>     <unknown>
osd.12                    ceph07  error         2h ago     2y
<unknown>  quay.io/ceph/ceph:v15.2.17  <unknown>     <unknown>
osd.14                    ceph07  error         2h ago     2y
<unknown>  quay.io/ceph/ceph:v15.2.17  <unknown>     <unknown>
osd.18                    ceph07  error         2h ago     2y
<unknown>  quay.io/ceph/ceph:v15.2.17  <unknown>     <unknown>
osd.22                    ceph07  error         2h ago     2y
<unknown>  quay.io/ceph/ceph:v15.2.17  <unknown>     <unknown>
osd.30                    ceph07  error         2h ago     2y
<unknown>  quay.io/ceph/ceph:v15.2.17  <unknown>     <unknown>
osd.4                     ceph07  error         2h ago     2y
<unknown>  quay.io/ceph/ceph:v15.2.17  <unknown>     <unknown>
osd.64                    ceph07  error         2h ago     2y
<unknown>  quay.io/ceph/ceph:v15.2.17  <unknown>     <unknown>
osd.8                     ceph07  error         2h ago     2y
<unknown>  quay.io/ceph/ceph:v15.2.17  <unknown>     <unknown>


Which are non-existing Daemons on that node "ceph07". And i can not remove them:
cephadm rm-daemon --fsid=5436dd5d-83d4-4dc8-a93b-60ab5db145df
--name=osd.0 --force
cephadm rm-daemon --fsid=5436dd5d-83d4-4dc8-a93b-60ab5db145df
--name=osd.1 --force
cephadm rm-daemon --fsid=5436dd5d-83d4-4dc8-a93b-60ab5db145df
--name=osd.12 --force
cephadm rm-daemon --fsid=5436dd5d-83d4-4dc8-a93b-60ab5db145df
--name=osd.14 --force
cephadm rm-daemon --fsid=5436dd5d-83d4-4dc8-a93b-60ab5db145df
--name=osd.18 --force
cephadm rm-daemon --fsid=5436dd5d-83d4-4dc8-a93b-60ab5db145df
--name=osd.30 --force
cephadm rm-daemon --fsid=5436dd5d-83d4-4dc8-a93b-60ab5db145df
--name=osd.4 --force
cephadm rm-daemon --fsid=5436dd5d-83d4-4dc8-a93b-60ab5db145df
--name=osd.64 --force
cephadm rm-daemon --fsid=5436dd5d-83d4-4dc8-a93b-60ab5db145df
--name=osd.8 --force

root@ceph07:/tmp# ls /var/lib/ceph/5436dd5d-83d4-4dc8-a93b-60ab5db145df/
crash  home  osd.66  osd.67  osd.68  osd.69  osd.999  removed
  => thats correct.

root@ceph07:/tmp# ls /var/lib/ceph/5436dd5d-83d4-4dc8-a93b-60ab5db145df/removed/
mon.ceph01_2020-09-02T07:11:30.232540
mon.ceph07_2020-11-20T14:17:56.122749
osd.12_2022-12-29T13:17:47.855132  osd.22_2022-12-29T13:13:47.233379
osd.64_2022-12-29T13:17:50.732467   osd.73_2022-12-29T09:54:58.009039Z
mon.ceph01_2022-12-29T13:18:33.702553
osd.0_2022-12-29T13:17:46.661637
osd.14_2022-12-29T13:17:48.485548  osd.30_2022-12-29T13:17:49.685540
osd.70_2022-12-29T09:56:15.014346Z  osd.74_2022-12-29T09:54:59.529058Z
mon.ceph02_2020-09-01T12:07:11.808391
osd.11_2022-12-29T13:15:39.944974
osd.18_2022-12-29T13:17:49.145034  osd.32_2020-07-30T09:44:23.252102
osd.71_2022-12-29T09:54:55.157744Z  osd.75_2022-12-29T09:55:02.647709Z
mon.ceph03_2020-09-01T13:26:34.704724
osd.1_2022-12-29T13:17:47.233991
osd.20_2022-12-29T12:58:27.511277  osd.4_2022-12-29T13:17:50.199486
osd.72_2022-12-29T09:54:56.537846Z  osd.8_2022-12-29T13:17:51.372638


my first try was to rename the old/non-active OSD from osd.33 to
_osd.33, but now i have a dangling module here:

root@ceph07:/tmp# ceph -s
  cluster:
    id:     5436dd5d-83d4-4dc8-a93b-60ab5db145df
    health: HEALTH_ERR
            mons are allowing insecure global_id reclaim
            20 failed cephadm daemon(s)
            Module 'cephadm' has failed: '_osd'

Any hints on how to clean up my node? :)

Cheers,
Mario
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux