Re: cephadm: How to replace failed HDD where DB is on SSD

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



We've found that after doing the osd rm, you can use: "ceph-volume lvm
zap --osd-id 178 --destroy" on the server with that OSD as per:
https://docs.ceph.com/en/latest/ceph-volume/lvm/zap/#removing-devices
and it will clean things up so they work as expected.

On Tue, May 25, 2021 at 6:51 AM Kai Stian Olstad <ceph+list@xxxxxxxxxx> wrote:
>
> Hi
>
> The server run 15.2.9 and has 15 HDD and 3 SSD.
> The OSDs was created with this YAML file
>
> hdd.yml
> --------
> service_type: osd
> service_id: hdd
> placement:
>    host_pattern: 'pech-hd-*'
> data_devices:
>    rotational: 1
> db_devices:
>    rotational: 0
>
>
> The result was that the 3 SSD is added to 1 VG with 15 LV on it.
>
> # vgs | egrep "VG|dbs"
>    VG                                                  #PV #LV #SN Attr
> VSize  VFree
>    ceph-block-dbs-563432b7-f52d-4cfe-b952-11542594843b   3  15   0 wz--n-
> <5.24t 48.00m
>
>
> One of the osd failed and I run rm with replace
>
> # ceph orch osd rm 178 --replace
>
> and the result is
>
> # ceph osd tree | grep "ID|destroyed"
> ID   CLASS  WEIGHT      TYPE NAME             STATUS     REWEIGHT
> PRI-AFF
> 178    hdd    12.82390          osd.178       destroyed         0
> 1.00000
>
>
> But I'm not able to replace the disk with the same YAML file as shown
> above.
>
>
> # ceph orch apply osd -i hdd.yml --dry-run
> ################
> OSDSPEC PREVIEWS
> ################
> +---------+------+------+------+----+-----+
> |SERVICE  |NAME  |HOST  |DATA  |DB  |WAL  |
> +---------+------+------+------+----+-----+
> +---------+------+------+------+----+-----+
>
> I guess this is the wrong way to do it, but I can't find the answer in
> the documentation.
> So how can I replace this failed disk in Cephadm?
>
>
> --
> Kai Stian Olstad
> _______________________________________________
> ceph-users mailing list -- ceph-users@xxxxxxx
> To unsubscribe send an email to ceph-users-leave@xxxxxxx
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux