Is the content of OSD.3 still available in the filesystem? If the
answer is yes you can get the OSD's keyring from
/var/lib/ceph/osd/ceph-3/keyring
Then update your osd.3.export file with the correct keyring and then
import the correct back to ceph.
Zitat von Abdelillah Asraoui <aasraoui@xxxxxxxxx>:
I must have imported osd.2 key instead, now osd.3 has the same key as osd.2
ceph auth import -i osd.3.export
How do we update this ?
thanks!
On Wed, Sep 29, 2021 at 2:13 AM Eugen Block <eblock@xxxxxx> wrote:
Just to clarify, you didn't simply import the unchanged keyring but
modified it to reflect the actual key of OSD.3, correct? If not, run
'ceph auth get osd.3' first and set the key in the osd.3.export file
before importing it to ceph.
Zitat von Abdelillah Asraoui <aasraoui@xxxxxxxxx>:
> i have created keyring for the osd3 but still pod is not booting up..
>
> As outlined:
> https://access.redhat.com/solutions/3524771
>
> ceph auth export osd.2 -o osd.2.export
> cp osd.2.export osd.3.export
> ceph auth import -i osd.3.export
> imported keyring
>
>
> Any suggestions ?
>
> Thanks!
>
> On Tue, Sep 21, 2021 at 8:34 AM Abdelillah Asraoui <aasraoui@xxxxxxxxx>
> wrote:
>
>> Hi,
>>
>> one of the osd in the cluster went down, is there a workaround to bring
>> back this osd?
>>
>>
>> logs from ceph osd pod shows the following:
>>
>> kubectl -n rook-ceph logs rook-ceph-osd-3-6497bdc65b-pn7mg
>>
>> debug 2021-09-20T14:32:46.388+0000 7f930fe9cf00 -1 auth: unable to find
a
>> keyring on /var/lib/ceph/osd/ceph-3/keyring: (13) Permission denied
>>
>> debug 2021-09-20T14:32:46.389+0000 7f930fe9cf00 -1 auth: unable to find
a
>> keyring on /var/lib/ceph/osd/ceph-3/keyring: (13) Permission denied
>>
>> debug 2021-09-20T14:32:46.389+0000 7f930fe9cf00 -1 auth: unable to find
a
>> keyring on /var/lib/ceph/osd/ceph-3/keyring: (13) Permission denied
>>
>> debug 2021-09-20T14:32:46.389+0000 7f930fe9cf00 -1 monclient: keyring
not
>> found
>>
>> failed to fetch mon config (--no-mon-config to skip)
>>
>>
>>
>>
>>
>> kubectl -n rook-ceph describe pod rook-ceph-osd-3-64
>>
>>
>>
>>
>>
>> Events:
>>
>> Type Reason Age From Message
>>
>> ---- ------ ---- ---- -------
>>
>> Normal Pulled 50m (x749 over 2d16h) kubelet Container image
>> "ceph/ceph:v15.2.13" already present on machine
>>
>> Warning BackOff 19s (x18433 over 2d16h) kubelet Back-off
restarting
>> failed container
>>
>>
>>
>> ceph health detail | more
>>
>> HEALTH_WARN noout flag(s) set; 1 osds down; 1 host (1 osds) down;
Degraded
>> data redundancy: 180969/542907 objects degraded (33.333%), 225 pgs degra
>>
>> ded, 225 pgs undersized
>>
>> [WRN] OSDMAP_FLAGS: noout flag(s) set
>>
>> [WRN] OSD_DOWN: 1 osds down
>>
>> osd.3 (root=default,host=ab-test) is down
>>
>> [WRN] OSD_HOST_DOWN: 1 host (1 osds) down
>>
>> host ab-test-mstr-1-cwan-net (root=default) (1 osds) is down
>>
>> [WRN] PG_DEGRADED: Degraded data redundancy: 180969/542907 objects
>> degraded (33.333%), 225 pgs degraded, 225 pgs undersized
>>
>> pg 3.4d is active+undersized+degraded, acting [2,0]
>>
>> pg 3.4e is stuck undersized for 3d, current state
>> active+undersized+degraded, last acting [0,2]
>>
>>
>> Thanks!
>>
> _______________________________________________
> ceph-users mailing list -- ceph-users@xxxxxxx
> To unsubscribe send an email to ceph-users-leave@xxxxxxx
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx