Re: External RGW always down

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi Eugen,

Below is the log output,

2022-09-07T12:03:42.893+0000 7fdd23fdc5c0  0 framework: beast
2022-09-07T12:03:42.893+0000 7fdd23fdc5c0  0 framework conf key: port, val:
80
2022-09-07T12:03:42.893+0000 7fdd23fdc5c0  1 radosgw_Main not setting numa
affinity
2022-09-07T12:03:42.893+0000 7fdd23fdc5c0  1 rgw_d3n:
rgw_d3n_l1_local_datacache_enabled=0
2022-09-07T12:03:42.893+0000 7fdd23fdc5c0  1 D3N datacache enabled: 0
2022-09-07T12:03:53.313+0000 7fdd23fdc5c0  1 rgw main: int
RGWSI_Notify::robust_notify(const DoutPrefixProvider*, RGWSI_RADOS::Obj&,
const RGWCacheNotifyInfo&, optional_yi>
2022-09-07T12:03:53.313+0000 7fdd23fdc5c0  1 rgw main: int
RGWSI_Notify::robust_notify(const DoutPrefixProvider*, RGWSI_RADOS::Obj&,
const RGWCacheNotifyInfo&, optional_yi>
2022-09-07T12:08:42.891+0000 7fdd1661c700 -1 Initialization timeout, failed
to initialize
2022-09-07T12:08:53.395+0000 7f69017095c0  0 deferred set uid:gid to
167:167 (ceph:ceph)
2022-09-07T12:08:53.395+0000 7f69017095c0  0 ceph version 17.2.0
(43e2e60a7559d3f46c9d53f1ca875fd499a1e35e) quincy (stable), process
radosgw, pid 7
2022-09-07T12:08:53.395+0000 7f69017095c0  0 framework: beast
2022-09-07T12:08:53.395+0000 7f69017095c0  0 framework conf key: port, val:
80
2022-09-07T12:08:53.395+0000 7f69017095c0  1 radosgw_Main not setting numa
affinity
2022-09-07T12:08:53.395+0000 7f69017095c0  1 rgw_d3n:
rgw_d3n_l1_local_datacache_enabled=0
2022-09-07T12:08:53.395+0000 7f69017095c0  1 D3N datacache enabled: 0
2022-09-07T12:09:03.747+0000 7f69017095c0  1 rgw main: int
RGWSI_Notify::robust_notify(const DoutPrefixProvider*, RGWSI_RADOS::Obj&,
const RGWCacheNotifyInfo&, optional_yi>
2022-09-07T12:09:03.747+0000 7f69017095c0  1 rgw main: int
RGWSI_Notify::robust_notify(const DoutPrefixProvider*, RGWSI_RADOS::Obj&,
const RGWCacheNotifyInfo&, optional_yi>
2022-09-07T12:13:53.397+0000 7f68f3d49700 -1 Initialization timeout, failed
to initialize

I installed the cluster in quincy.


On Sat, Sep 10, 2022 at 4:02 PM Eugen Block <eblock@xxxxxx> wrote:

> What troubleshooting have you tried? You don’t provide any log output
> or information about the cluster setup, for example the ceph osd tree,
> ceph status, are the failing OSDs random or do they all belong to the
> same pool? Any log output from failing OSDs and the RGWs might help,
> otherwise it’s just wild guessing. Is the cluster a new installation
> with cephadm or an older cluster upgraded to Quincy?
>
> Zitat von Monish Selvaraj <monish@xxxxxxxxxxxxxxx>:
>
> > Hi all,
> >
> > I have one critical issue in my prod cluster. When the customer's data
> > comes from 600 MiB .
> >
> > My Osds are down *8 to 20 from 238* . Then I manually up my osds . After
> a
> > few minutes, my all rgw crashes.
> >
> > We did some troubleshooting but nothing works. When we upgrade ceph to
> > 17.2.0. to 17.2.1 is resolved. Also we faced the issue two times. But
> both
> > times we upgraded the ceph.
> >
> > *Node schema :*
> >
> > *Node 1 to node 5 --> mon,mgr and osds*
> > *Node 6 to Node15 --> only osds*
> > *Node 16 to Node 20 --> only rgws.*
> >
> > Kindly, check this issue and let me know the correct troubleshooting
> method.
> > _______________________________________________
> > ceph-users mailing list -- ceph-users@xxxxxxx
> > To unsubscribe send an email to ceph-users-leave@xxxxxxx
>
>
>
> _______________________________________________
> ceph-users mailing list -- ceph-users@xxxxxxx
> To unsubscribe send an email to ceph-users-leave@xxxxxxx
>
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux