Re: Replacing OSD with containerized deployment

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hello,

I'm sorry for not getting back to you sooner.

[2023-01-26 16:25:00,785][ceph_volume.process][INFO  ] stdout
ceph.block_device=/dev/ceph-808efc2a-54fd-47cc-90e2-c5cc96bdd825/osd-block-
2a1d1bf0-300e-4160-ac55-047837a5af0b,ceph.block_uuid=
b4WDQQ-eMTb-AN1U-D7dk-yD2q-4dPZ-KyFrHi,ceph.cephx_
lockbox_secret=,ceph.cluster_fsid=8038f09a-27a0-11ed-8de8-
55262cdd5a37,ceph.cluster_name=ceph,ceph.crush_device_
class=,ceph.encrypted=0,ceph.osd_fsid=2a1d1bf0-300e-4160-
ac55-047837a5af0b,ceph.osd_id=232,ceph.osdspec_affinity=
dashboard-admin-1661788934732,ceph.type=wal,ceph.vdo=0,ceph.
wal_device=/dev/ceph-3a336b8e-ed39-4532-a199-ac6a3730840b/
osd-wal-5d845dba-8b55-4984-890b-547fbdaff10c,ceph.wal_
uuid=dquBMJ-s8ou-Wp6M-NY8Z-QoFh-6L4b-9Lwqm0";"/dev/ceph-
3a336b8e-ed39-4532-a199-ac6a3730840b/osd-wal-5d845dba-
8b55-4984-890b-547fbdaff10c";"osd-wal-5d845dba-8b55-4984-
890b-547fbdaff10c";"ceph-3a336b8e-ed39-4532-a199-ac6a3730840b";"dquBMJ-s8ou-
Wp6M-NY8Z-QoFh-6L4b-9Lwqm0";"355622453248

>From that line, I read that osd.232 has block on /dev/ceph-
808efc2a-54fd-47cc-90e2-c5cc96bdd825/osd-block-2a1d1bf0-300e-4160-ac55-047837a5af0b
and block.wal on /dev/ceph-3a336b8e-ed39-4532-a199-ac6a3730840b/
osd-wal-5d845dba-8b55-4984-890b-547fbdaff10c from there, check if that
device is well an LV member of the NVME device.
Can you share the full output of lsblk ?

Thanks,

On Wed, 1 Feb 2023 at 17:22, mailing-lists <mailing-lists@xxxxxxxxx> wrote:

> I've  pulled a few lines from the log and i've attached this to this mail.
> (I hope this works for this mailinglist?)
>
>
> I found the line 135
>
> [2023-01-26 16:25:00,785][ceph_volume.process][INFO  ] stdout
> ceph.block_device=/dev/ceph-808efc2a-54fd-47cc-90e2-c5cc96bdd825/osd-block-2a1d1bf0-300e-4160-ac55-047837a5af0b,ceph.block_uuid=b4WDQQ-eMTb-AN1U-D7dk-yD2q-4dPZ-KyFrHi,ceph.cephx_lockbox_secret=,ceph.cluster_fsid=8038f09a-27a0-11ed-8de8-55262cdd5a37,ceph.cluster_name=ceph,ceph.crush_device_class=,ceph.encrypted=0,ceph.osd_fsid=2a1d1bf0-300e-4160-ac55-047837a5af0b,ceph.osd_id=232,ceph.osdspec_affinity=dashboard-admin-1661788934732,ceph.type=wal,ceph.vdo=0,ceph.wal_device=/dev/ceph-3a336b8e-ed39-4532-a199-ac6a3730840b/osd-wal-5d845dba-8b55-4984-890b-547fbdaff10c,ceph.wal_uuid=dquBMJ-s8ou-Wp6M-NY8Z-QoFh-6L4b-9Lwqm0";"/dev/ceph-3a336b8e-ed39-4532-a199-ac6a3730840b/osd-wal-5d845dba-8b55-4984-890b-547fbdaff10c";"osd-wal-5d845dba-8b55-4984-890b-547fbdaff10c";"ceph-3a336b8e-ed39-4532-a199-ac6a3730840b";"dquBMJ-s8ou-Wp6M-NY8Z-QoFh-6L4b-9Lwqm0";"355622453248
>
>
> Which indicates that this OSD is in fact using a WAL, since WAL and DB
> should both be on the NVME, i would guess it is just a visual bug in the
> dashboard?
>
>
> From line 135:
>
>
> ceph.wal_device=/dev/ceph-3a336b8e-ed39-4532-a199-ac6a3730840b/osd-wal-5d845dba-8b55-4984-890b-547fbdaff10c
>
> From lsblk:
>
> ├─ceph--3a336b8e--ed39--4532--a199--ac6a3730840b-osd--wal--5d845dba--8b55--4984--890b--547fbdaff10c
> 253:12   0 331.2G  0 lvm
>
>
> So it looks like it is using that lvm group right there. Yet, the
> dashboard doesn't show a nvme. (please compare screenshot osd_232.png and
> osd_218.png)
>
>
> Can I somehow confirm, that my osd 232 is really using the nvme as wal/db?
>
>
> Thanks and best regards
>
> Ken
>
>
>
> On 01.02.23 10:35, Guillaume Abrioux wrote:
>
> Any chance you can share the ceph-volume.log (from the corresponding host)?
> It should be in /var/log/ceph/<cluster fsid>/ceph-volume.log. Note that
> there might be several log files (log rotation). Ideally, the one that
> includes the recreation steps.
>
> Thanks,
>
> On Wed, 1 Feb 2023 at 10:13, mailing-lists <mailing-lists@xxxxxxxxx>
> wrote:
>
>> Ah, nice.
>>
>> service_type: osd
>> service_id: dashboard-admin-1661788934732
>> service_name: osd.dashboard-admin-1661788934732
>> placement:
>>   host_pattern: '*'
>> spec:
>>   data_devices:
>>     model: MG08SCA16TEY
>>   db_devices:
>>     model: Dell Ent NVMe AGN MU AIC 6.4TB
>>   filter_logic: AND
>>   objectstore: bluestore
>>   wal_devices:
>>     model: Dell Ent NVMe AGN MU AIC 6.4TB
>> status:
>>   created: '2022-08-29T16:02:22.822027Z'
>>   last_refresh: '2023-02-01T09:03:22.853860Z'
>>   running: 306
>>   size: 306
>>
>>
>> Best
>>
>> Ken
>> On 31.01.23 23:51, Guillaume Abrioux wrote:
>>
>> On Tue, 31 Jan 2023 at 22:31, mailing-lists <mailing-lists@xxxxxxxxx>
>> wrote:
>>
>>> I am not sure. I didn't find it... It should be somewhere, right? I used
>>> the dashboard to create the osd service.
>>>
>>
>> what does a `cephadm shell -- ceph orch ls osd --format yaml` say?
>>
>> --
>>
>> *Guillaume Abrioux *Senior Software Engineer
>>
>>
>
> --
>
> *Guillaume Abrioux *Senior Software Engineer
>
>

-- 

*Guillaume Abrioux*Senior Software Engineer
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux