Re: Ceph orchestrator not refreshing device list

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



I’ve been away on vacation and just got back to this. I’m happy to report that manually recreating the OSD with ceph-volume and then adopting it with cephadm fixed the problem.

Thanks again for your help Eugen!

Cheers,
/rjg

> On Sep 29, 2024, at 10:40 AM, Eugen Block <eblock@xxxxxx> wrote:
> 
> EXTERNAL EMAIL | USE CAUTION
> 
> Okay, apparently this is not what I was facing. I see two other
> options right now. The first would be to purge osd.88 from the crush
> tree entirely.
> The second approach would be to create an osd manually with
> ceph-volume, not cephadm ceph-volume, to create a legacy osd (you'd
> get warnings about a stray daemon). If that works, adopt the osd with
> cephadm.
> I don't have a better idea right now.
> 
> Zitat von Bob Gibson <rjg@xxxxxxxxxx>:
> 
>> Here are the contents from the same directory on our osd node:
>> 
>> ceph-osd31.prod.os:/var/lib/ceph/9b3b3539-59a9-4338-8bab-3badfab6e855# ls -l
>> total 412
>> -rw-r--r--  1 root root 366903 Sep 14 14:53
>> cephadm.8b92cafd937eb89681ee011f9e70f85937fd09c4bd61ed4a59981d275a1f255b
>> drwx------  3  167  167   4096 Sep 14 15:01 crash
>> drwxr-xr-x 12 root root   4096 Sep 15 12:06 custom_config_files
>> drw-rw----  2 root root   4096 Sep 23 17:00 home
>> drwx------  2  167  167   4096 Sep 26 12:47 osd.84
>> drwx------  2  167  167   4096 Sep 26 12:47 osd.85
>> drwx------  2  167  167   4096 Sep 26 12:47 osd.86
>> drwx------  2  167  167   4096 Sep 26 12:47 osd.87
>> drwx------  2  167  167   4096 Sep 26 12:47 osd.89
>> drwx------  2  167  167   4096 Sep 26 12:47 osd.90
>> drwx------  2  167  167   4096 Sep 26 12:47 osd.91
>> drwx------  2  167  167   4096 Sep 26 12:47 osd.92
>> drwx------  2  167  167   4096 Sep 26 12:47 osd.93
>> drwx------  6 root root   4096 Sep 23 15:59 removed
>> 
>> In our case the osd.88 directory is under the subdirectory named
>> “removed”, the same as the other odds which have been converted.
>> 
>> ceph-osd31.prod.os:/var/lib/ceph/9b3b3539-59a9-4338-8bab-3badfab6e855# ls -l
>> removed/osd.88_2024-09-23T19\:59\:42.162302Z/
>> total 64
>> lrwxrwxrwx 1 167 167   93 Sep 15 12:10 block ->
>> /dev/ceph-2a13ec6a-a5f0-4773-8254-c38b915c824a/osd-block-7f8f9778-5ae2-47c1-bd03-a92a3a7a1db1
>> -rw------- 1 167 167   37 Sep 15 12:10 ceph_fsid
>> -rw------- 1 167 167  259 Sep 14 15:14 config
>> -rw------- 1 167 167   37 Sep 15 12:10 fsid
>> -rw------- 1 167 167   56 Sep 15 12:10 keyring
>> -rw------- 1 167 167    6 Sep 15 12:10 ready
>> -rw------- 1 167 167    3 Sep 14 11:11 require_osd_release
>> -rw------- 1 167 167   10 Sep 15 12:10 type
>> -rw------- 1 167 167   38 Sep 14 15:14 unit.configured
>> -rw------- 1 167 167   48 Sep 14 15:14 unit.created
>> -rw------- 1 167 167   26 Sep 14 15:06 unit.image
>> -rw------- 1 167 167   76 Sep 14 15:06 unit.meta
>> -rw------- 1 167 167 1527 Sep 14 15:06 unit.poststop
>> -rw------- 1 167 167 2586 Sep 14 15:06 unit.run
>> -rw------- 1 167 167  334 Sep 14 15:06 unit.stop
>> -rw------- 1 167 167    3 Sep 15 12:10 whoami
>> 
>> On Sep 27, 2024, at 9:30 AM, Eugen Block <eblock@xxxxxx> wrote:
>> 
>> EXTERNAL EMAIL | USE CAUTION
>> 
>> Oh interesting, I just got into the same situation (I believe) on a
>> test cluster:
>> 
>> host1:~ # ceph orch ps | grep unknown
>> osd.1                              host6
>> stopped          72s ago  36m        -    4096M  <unknown>  <unknown>
>>  <unknown>
>> osd.13                             host6
>> error            72s ago  36m        -    4096M  <unknown>  <unknown>
>>  <unknown>
>> 
>> I still had the remainders on the filesystem:
>> 
>> host6:~ # ll /var/lib/ceph/543967bc-e586-32b8-bd2c-2d8b8b168f02/osd.1
>> insgesamt 68
>> lrwxrwxrwx 1 ceph ceph  111 27. Sep 14:43 block ->
>> /dev/mapper/ceph--0e90997f--456e--4a9b--a8f9--a6f1038c1216-osd--block--81e7f32a--a728--4848--b14d--0b86bb7e1c69
>> lrwxrwxrwx 1 ceph ceph  108 27. Sep 14:43 block.db ->
>> /dev/mapper/ceph--9ea6e95f--ad43--4e40--8920--2e772b2efa2f-osd--db--f9c57ec1--77c8--4d9a--85df--1dc053a24000
>> 
>> I just removed those two directories to clear the warning, now my
>> orchestrator can deploy OSDs again on that node.
>> 
>> Hope that helps!
>> 
>> Zitat von Eugen Block <eblock@xxxxxx>:
>> 
>> Right, if you need encryption, a rebuild is required. Your procedure
>> has already worked 4 times, so I'd say nothing seems wrong with that
>> per se.
>> Regarding the stuck device list, do you see the mgr logging anything
>> suspicious? Especially when you say that it only returns output
>> after a failover. Those two osd specs are not conflicting since the
>> first is "unmanaged" after adoption.
>> Is there something in 'ceph orch osd rm status'? Can you run
>> 'cephadm ceph-volume inventory' locally on that node? Do you see any
>> hints in the node's syslog? Maybe try a reboot or something?
>> 
>> 
>> Zitat von Bob Gibson <rjg@xxxxxxxxxx>:
>> 
>> Thanks for your reply Eugen. I’m fairly new to cephadm so I wasn’t
>> aware that we could manage the drives without rebuilding them.
>> However, we thought we’d take advantage of this opportunity to also
>> encrypt the drives, and that does require a rebuild.
>> 
>> I have a theory on why the orchestrator is confused. I want to
>> create an osd service for each osd node so I can manage drives on a
>> per-node basis.
>> 
>> I started by creating a spec for the first node:
>> 
>> service_type: osd
>> service_id: ceph-osd31
>> placement:
>> hosts:
>> - ceph-osd31
>> spec:
>> data_devices:
>>  rotational: 0
>>  size: '3TB:'
>> encrypted: true
>> filter_logic: AND
>> objectstore: bluestore
>> 
>> But I also see a default spec, “osd”, which has placement set to
>> “unmanaged”.
>> 
>> `ceph orch ls osd —export` shows the following:
>> 
>> service_type: osd
>> service_name: osd
>> unmanaged: true
>> spec:
>> filter_logic: AND
>> objectstore: bluestore
>> ---
>> service_type: osd
>> service_id: ceph-osd31
>> service_name: osd.ceph-osd31
>> placement:
>> hosts:
>> - ceph-osd31
>> spec:
>> data_devices:
>>  rotational: 0
>>  size: '3TB:'
>> encrypted: true
>> filter_logic: AND
>> objectstore: bluestore
>> 
>> `ceph orch ls osd` shows that I was able to convert 4 drives using my spec:
>> 
>> NAME            PORTS  RUNNING  REFRESHED  AGE  PLACEMENT
>> osd                         95  10m ago    -    <unmanaged>
>> osd.ceph-osd31               4  10m ago    43m  ceph-osd31
>> 
>> Despite being able to convert 4 drives, I’m wondering if these
>> specs are conflicting with one another, and that has confused the
>> orchestrator. If so, how do I safely get from where I am now to
>> where I want to be? :-)
>> 
>> Cheers,
>> /rjg
>> 
>> On Sep 26, 2024, at 3:31 PM, Eugen Block <eblock@xxxxxx> wrote:
>> 
>> EXTERNAL EMAIL | USE CAUTION
>> 
>> Hi,
>> 
>> this seems a bit unnecessary to rebuild OSDs just to get them managed.
>> If you apply a spec file that targets your hosts/OSDs, they will
>> appear as managed. So when you would need to replace a drive, you
>> could already utilize the orchestrator to remove and zap the drive.
>> That works just fine.
>> How to get out of your current situation is not entirely clear to me
>> yet. I’ll reread your post tomorrow.
>> 
>> Regards,
>> Eugen
>> 
>> Zitat von Bob Gibson <rjg@xxxxxxxxxx>:
>> 
>> Hi,
>> 
>> We recently converted a legacy cluster running Quincy v17.2.7 to
>> cephadm. The conversion went smoothly and left all osds unmanaged by
>> the orchestrator as expected. We’re now in the process of converting
>> the osds to be managed by the orchestrator. We successfully
>> converted a few of them, but then the orchestrator somehow got
>> confused. `ceph health detail` reports a “stray daemon” for the osd
>> we’re trying to convert, and the orchestrator is unable to refresh
>> its device list so it doesn’t see any available devices.
>> 
>> From the perspective of the osd node, the osd has been wiped and is
>> ready to be reinstalled. We’ve also rebooted the node for good
>> measure. `ceph osd tree` shows that the osd has been destroyed, but
>> the orchestrator won’t reinstall it because it thinks the device is
>> still active. The orchestrator device information is stale, but
>> we’re unable to refresh it. The usual recommended workaround of
>> failing over the mgr hasn’t helped. We’ve also tried `ceph orch
>> device ls —refresh` to no avail. In fact after running that command
>> subsequent runs of `ceph orch device ls` produce no output until the
>> mgr is failed over again.
>> 
>> Is there a way to force the orchestrator to refresh its list of
>> devices when in this state? If not, can anyone offer any suggestions
>> on how to fix this problem?
>> 
>> Cheers,
>> /rjg
>> 
>> P.S. Some additional information in case it’s helpful...
>> 
>> We’re using the following command to replace existing devices so
>> that they’re managed by the orchestrator:
>> 
>> ```
>> ceph orch osd rm <osd> --replace —zap
>> ```
>> 
>> and we’re currently stuck on osd 88.
>> 
>> ```
>> ceph health detail
>> HEALTH_WARN 1 stray daemon(s) not managed by cephadm
>> [WRN] CEPHADM_STRAY_DAEMON: 1 stray daemon(s) not managed by cephadm
>> stray daemon osd.88 on host ceph-osd31 not managed by cephadm
>> ```
>> 
>> `ceph osd tree` shows that the osd has been destroyed and is ready
>> to be replaced:
>> 
>> ```
>> ceph osd tree-from ceph-osd31
>> ID   CLASS  WEIGHT    TYPE NAME        STATUS     REWEIGHT  PRI-AFF
>> -46         34.93088  host ceph-osd31
>> 84    ssd   3.49309      osd.84              up   1.00000  1.00000
>> 85    ssd   3.49309      osd.85              up   1.00000  1.00000
>> 86    ssd   3.49309      osd.86              up   1.00000  1.00000
>> 87    ssd   3.49309      osd.87              up   1.00000  1.00000
>> 88    ssd   3.49309      osd.88       destroyed         0  1.00000
>> 89    ssd   3.49309      osd.89              up   1.00000  1.00000
>> 90    ssd   3.49309      osd.90              up   1.00000  1.00000
>> 91    ssd   3.49309      osd.91              up   1.00000  1.00000
>> 92    ssd   3.49309      osd.92              up   1.00000  1.00000
>> 93    ssd   3.49309      osd.93              up   1.00000  1.00000
>> ```
>> 
>> The cephadm log shows a claim on node `ceph-osd31` for that osd:
>> 
>> ```
>> 2024-09-25T14:15:45.699348-0400 mgr.ceph-mon3.qzjgws [INF] Found osd
>> claims -> {'ceph-osd31': ['88']}
>> 2024-09-25T14:15:45.699534-0400 mgr.ceph-mon3.qzjgws [INF] Found osd
>> claims for drivegroup ceph-osd31 -> {'ceph-osd31': ['88']}
>> ```
>> 
>> `ceph orch device ls` shows that the device list isn’t refreshing:
>> 
>> ```
>> ceph orch device ls ceph-osd31
>> HOST        PATH      TYPE  DEVICE ID
>> SIZE  AVAILABLE  REFRESHED  REJECT REASONS
>> ceph-osd31  /dev/sdc  ssd   INTEL_SSDSC2KG038T8_PHYG039603PE3P8EGN
>> 3576G  No         22h ago    Insufficient space (<10 extents) on
>> vgs, LVM detected, locked
>> ceph-osd31  /dev/sdd  ssd   INTEL_SSDSC2KG038T8_PHYG039600AY3P8EGN
>> 3576G  No         22h ago    Insufficient space (<10 extents) on
>> vgs, LVM detected, locked
>> ceph-osd31  /dev/sde  ssd   INTEL_SSDSC2KG038T8_PHYG039600CW3P8EGN
>> 3576G  No         22h ago    Insufficient space (<10 extents) on
>> vgs, LVM detected, locked
>> ceph-osd31  /dev/sdf  ssd   INTEL_SSDSC2KG038T8_PHYG039600CM3P8EGN
>> 3576G  No         22h ago    Insufficient space (<10 extents) on
>> vgs, LVM detected, locked
>> ceph-osd31  /dev/sdg  ssd   INTEL_SSDSC2KG038T8_PHYG039600UB3P8EGN
>> 3576G  No         22h ago    Insufficient space (<10 extents) on
>> vgs, LVM detected, locked
>> ceph-osd31  /dev/sdh  ssd   INTEL_SSDSC2KG038T8_PHYG039603753P8EGN
>> 3576G  No         22h ago    Insufficient space (<10 extents) on
>> vgs, LVM detected, locked
>> ceph-osd31  /dev/sdi  ssd   INTEL_SSDSC2KG038T8_PHYG039603R63P8EGN
>> 3576G  No         22h ago    Insufficient space (<10 extents) on
>> vgs, LVM detected, locked
>> ceph-osd31  /dev/sdj  ssd   INTEL_SSDSC2KG038TZ_PHYJ4011032M3P8DGN
>> 3576G  No         22h ago    Insufficient space (<10 extents) on
>> vgs, LVM detected, locked
>> ceph-osd31  /dev/sdk  ssd   INTEL_SSDSC2KG038TZ_PHYJ3234010J3P8DGN
>> 3576G  No         22h ago    Insufficient space (<10 extents) on
>> vgs, LVM detected, locked
>> ceph-osd31  /dev/sdl  ssd   INTEL_SSDSC2KG038T8_PHYG039603NS3P8EGN
>> 3576G  No         22h ago    Insufficient space (<10 extents) on
>> vgs, LVM detected, locked
>> ```
>> 
>> `ceph node ls` thinks the osd still exists
>> 
>> ```
>> ceph node ls osd | jq -r '."ceph-osd31"'
>> [
>> 84,
>> 85,
>> 86,
>> 87,
>> 88, <— this shouldn’t exist
>> 89,
>> 90,
>> 91,
>> 92,
>> 93
>> ]
>> ```
>> 
>> Each osd node has 10x 3.8 TB ssd drives for osds. On `ceph-osd31`,
>> cephadm doesn’t see osd.88 as expected:
>> 
>> ```
>> cephadm ls --no-detail
>> [
>> {
>>     "style": "cephadm:v1",
>>     "name": "osd.93",
>>     "fsid": "9b3b3539-59a9-4338-8bab-3badfab6e855",
>>     "systemd_unit": "ceph-9b3b3539-59a9-4338-8bab-3badfab6e855@osd.93"
>> },
>> {
>>     "style": "cephadm:v1",
>>     "name": "osd.85",
>>     "fsid": "9b3b3539-59a9-4338-8bab-3badfab6e855",
>>     "systemd_unit": "ceph-9b3b3539-59a9-4338-8bab-3badfab6e855@osd.85"
>> },
>> {
>>     "style": "cephadm:v1",
>>     "name": "osd.90",
>>     "fsid": "9b3b3539-59a9-4338-8bab-3badfab6e855",
>>     "systemd_unit": "ceph-9b3b3539-59a9-4338-8bab-3badfab6e855@osd.90"
>> },
>> {
>>     "style": "cephadm:v1",
>>     "name": "osd.92",
>>     "fsid": "9b3b3539-59a9-4338-8bab-3badfab6e855",
>>     "systemd_unit": "ceph-9b3b3539-59a9-4338-8bab-3badfab6e855@osd.92"
>> },
>> {
>>     "style": "cephadm:v1",
>>     "name": "osd.89",
>>     "fsid": "9b3b3539-59a9-4338-8bab-3badfab6e855",
>>     "systemd_unit": "ceph-9b3b3539-59a9-4338-8bab-3badfab6e855@osd.89"
>> },
>> {
>>     "style": "cephadm:v1",
>>     "name": "osd.87",
>>     "fsid": "9b3b3539-59a9-4338-8bab-3badfab6e855",
>>     "systemd_unit": "ceph-9b3b3539-59a9-4338-8bab-3badfab6e855@osd.87"
>> },
>> {
>>     "style": "cephadm:v1",
>>     "name": "osd.86",
>>     "fsid": "9b3b3539-59a9-4338-8bab-3badfab6e855",
>>     "systemd_unit": "ceph-9b3b3539-59a9-4338-8bab-3badfab6e855@osd.86"
>> },
>> {
>>     "style": "cephadm:v1",
>>     "name": "osd.84",
>>     "fsid": "9b3b3539-59a9-4338-8bab-3badfab6e855",
>>     "systemd_unit": "ceph-9b3b3539-59a9-4338-8bab-3badfab6e855@osd.84"
>> },
>> {
>>     "style": "cephadm:v1",
>>     "name": "osd.91",
>>     "fsid": "9b3b3539-59a9-4338-8bab-3badfab6e855",
>>     "systemd_unit": "ceph-9b3b3539-59a9-4338-8bab-3badfab6e855@osd.91"
>> }
>> ]
>> ```
>> 
>> `lsblk` shows that `/dev/sdg` has been wiped.
>> 
>> ```
>> NAME
>>                              MAJ:MIN RM   SIZE RO TYPE  MOUNTPOINT
>> sda
>>                                8:0    0 223.6G  0 disk
>> |-sda1
>>                                8:1    0    94M  0 part
>> `-sda2
>>                                8:2    0 223.5G  0 part
>> `-md0
>>                                9:0    0 223.4G  0 raid1 /
>> sdb
>>                                8:16   0 223.6G  0 disk
>> |-sdb1
>>                                8:17   0    94M  0 part
>> `-sdb2
>>                                8:18   0 223.5G  0 part
>> `-md0
>>                                9:0    0 223.4G  0 raid1 /
>> sdc
>>                                8:32   1   3.5T  0 disk
>> `-ceph--03782b4c--9faa--49f5--b554--98e7b8515834-osd--block--ba272724--daa6--45f5--9f69--789cc0bda077 253:3    0
>> 3.5T
>> lvm
>> `-keCkP2-o6h8-jKkw-RKiD-UBFf-A8EL-JDJGPR
>>                              253:9    0   3.5T  0 crypt
>> sdd
>>                                8:48   1   3.5T  0 disk
>> `-ceph--c07907d8--4a75--4ba3--b5e1--2ebf49ecbdf6-osd--block--58d1d50d--6228--4e6f--9a52--2a305ba00700 253:7    0
>> 3.5T
>> lvm
>> `-WB8Mxn-qCHI-4T01-imiG-hNBR-by60-YuxgfD
>>                              253:11   0   3.5T  0 crypt
>> sde
>>                                8:64   1   3.5T  0 disk
>> `-ceph--6f9d4df4--7ce6--44a4--a7b1--62c85af8cfe0-osd--block--aabcb30d--0084--490a--969b--78f7af6e94da 253:8    0
>> 3.5T
>> lvm
>> `-g9qErH-vTXY-JQbs-eh61-W0Mn-TAV8-gof4zy
>>                              253:12   0   3.5T  0 crypt
>> sdf
>>                                8:80   1   3.5T  0 disk
>> `-ceph--d6b728f8--e365--46db--b30f--6c00805c752b-osd--block--88426db7--2322--4807--ac2e--b49929e170d6 253:6    0
>> 3.5T
>> lvm
>> `-LNG2gB-pa0w-gl2v-hVQ3-6qTd-aXsR-Lenri3
>>                              253:10   0   3.5T  0 crypt
>> sdg
>>                                8:96   1   3.5T  0 disk
>> sdh
>>                                8:112  1   3.5T  0 disk
>> `-ceph--de2cfee6--8e0a--4aa0--9e6b--90c09025768c-osd--block--a3b86251--2799--4243--a857--f218fa90f29a 253:2    0
>> 3.5T
>> lvm
>> sdi
>>                                8:128  1   3.5T  0 disk
>> `-ceph--30dee450--0fdd--46ea--9eec--6a4c7706df9c-osd--block--bfc090db--dde4--47dd--a1c9--1cd838ea43b3 253:4    0
>> 3.5T
>> lvm
>> sdj
>>                                8:144  1   3.5T  0 disk
>> `-ceph--78febcf5--43f4--4820--8dc7--0f6c22816c9f-osd--block--da1e69c7--6427--4562--8290--90bcb9526747 253:0    0
>> 3.5T
>> lvm
>> sdk
>>                                8:160  1   3.5T  0 disk
>> `-ceph--fe210281--b1f5--4d5e--9ab0--2f226612af00-osd--block--6bb9f308--e853--4303--83ea--553c3a3513e1 253:1    0
>> 3.5T
>> lvm
>> sdl
>>                                8:176  1   3.5T  0 disk
>> `-ceph--9f21c916--f211--4d1b--8214--6ad1cecac810-osd--block--572d850c--c201--4af4--ac42--0ed2a6ed73ed 253:5    0
>> 3.5T
>> lvm
>> ```
>> 
>> _______________________________________________
>> ceph-users mailing list -- ceph-users@xxxxxxx
>> To unsubscribe send an email to ceph-users-leave@xxxxxxx
>> 
>> 
>> _______________________________________________
>> ceph-users mailing list -- ceph-users@xxxxxxx
>> To unsubscribe send an email to ceph-users-leave@xxxxxxx
> 
> 
> 

_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux