Re: Replacing OSD with containerized deployment

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



The 'down' status is why it's not being replaced, vs. destroyed, which would allow the replacement. I'm not sure why --replace lead to that scenario, but you will probably need to mark it destroyed for it to be replaced.

https://docs.ceph.com/en/latest/rados/operations/add-or-rm-osds/#replacing-an-osd has instructions on the non-orch way of doing that. You only need 1/2.

You should look through your logs to see what happened that the OSD was marked down and not destroyed. Obviously, make sure you understand ramifications before running any commands. :)

David

On Mon, Jan 30, 2023, at 04:24, mailing-lists wrote:
> # ceph orch osd rm status
> No OSD remove/replace operations reported
> # ceph orch osd rm 232 --replace
> Unable to find OSDs: ['232']
>
> It is not finding 232 anymore. It is still shown as down and out in the 
> Ceph-Dashboard.
>
>
>      pgs:     3236 active+clean
>
>
> This is the new disk shown as locked (because unzapped at the moment).
>
> # ceph orch device ls
>
> ceph-a1-06  /dev/sdm      hdd   TOSHIBA_X_X 16.0T             9m ago     
> locked
>
>
> Best
>
> Ken
>
>
> On 29.01.23 18:19, David Orman wrote:
>> What does "ceph orch osd rm status" show before you try the zap? Is 
>> your cluster still backfilling to the other OSDs for the PGs that were 
>> on the failed disk?
>>
>> David
>>
>> On Fri, Jan 27, 2023, at 03:25, mailing-lists wrote:
>>> Dear Ceph-Users,
>>>
>>> i am struggling to replace a disk. My ceph-cluster is not replacing the
>>> old OSD even though I did:
>>>
>>> ceph orch osd rm 232 --replace
>>>
>>> The OSD 232 is still shown in the osd list, but the new hdd will be
>>> placed as a new OSD. This wouldnt mind me much, if the OSD was also
>>> placed on the bluestoreDB / NVME, but it doesn't.
>>>
>>>
>>> My steps:
>>>
>>> "ceph orch osd rm 232 --replace"
>>>
>>> remove the failed hdd.
>>>
>>> add the new one.
>>>
>>> Convert the disk within the servers bios, so that the node can have
>>> direct access on it.
>>>
>>> It shows up as /dev/sdt,
>>>
>>> enter maintenance mode
>>>
>>> reboot server
>>>
>>> drive is now /dev/sdm (which the old drive had)
>>>
>>> "ceph orch device zap node-x /dev/sdm"
>>>
>>> A new OSD is placed on the cluster.
>>>
>>>
>>> Can you give me a hint, where did I take a wrong turn? Why is the disk
>>> not being used as OSD 232?
>>>
>>>
>>> Best
>>>
>>> Ken
>>>
>>> _______________________________________________
>>> ceph-users mailing list -- ceph-users@xxxxxxx
>>> To unsubscribe send an email to ceph-users-leave@xxxxxxx
>> _______________________________________________
>> ceph-users mailing list -- ceph-users@xxxxxxx
>> To unsubscribe send an email to ceph-users-leave@xxxxxxx
> _______________________________________________
> ceph-users mailing list -- ceph-users@xxxxxxx
> To unsubscribe send an email to ceph-users-leave@xxxxxxx
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux