Re: EC 8+3 Pool PGs stuck in remapped+incomplete

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hello Nino / Users,

After some initial analysis, I had increased max_pg_per_osd to 480, but
we're out of luck. Also tried force-backfill and force-repair as well.
On querying PG using *# ceph pg **<pg.ID> query* the output says blocked_by
3 to 4 OSDs which are out of the cluster already. Guessing if these have to
do something with the recovery.

Thanks,
Jayanth Reddy

On Sat, Jun 17, 2023 at 12:31 PM Jayanth Reddy <jayanthreddy5666@xxxxxxxxx>
wrote:

> Thanks, Nino.
>
> Would give these initial suggestions a try and let you know at the
> earliest.
>
> Regards,
> Jayanth Reddy
> ------------------------------
> *From:* Nino Kotur <ninokotur@xxxxxxxxx>
> *Sent:* Saturday, June 17, 2023 12:16:09 PM
> *To:* Jayanth Reddy <jayanthreddy5666@xxxxxxxxx>
> *Cc:* ceph-users@xxxxxxx <ceph-users@xxxxxxx>
> *Subject:* Re:  EC 8+3 Pool PGs stuck in remapped+incomplete
>
> problem is just that some of your OSDs have too much PGs and pool cannot
> recover as it cannot create more PGs
>
> [osd.214,osd.223,osd.548,osd.584] have slow ops.
>             too many PGs per OSD (330 > max 250)
>
> I'd have to guess that the safest thing would be permanently or
> temporarily adding more storage so that PGs drop below 250, another option
> is just dropping down the total number of PGs but I don't know if I would
> perform that action before my pool was healthy!
>
> in case that there is only one OSD that has this number of OSDs but all
> other OSDs have less than 100-150 than you can just reweight problematic
> OSD so it rebalances those "too many PGs"
>
> But it looks to me that you have way too many PGs which is also super
> negatively impacting performance.
>
> Another option is to increase max allowed PGs per OSD to say 350 this
> should also allow cluster to rebuild honestly even tho this may be easiest
> option, i'd never do this, performance cost of having over 150 PGs per OSD
> suffer greatly.
>
>
> kind regards,
> Nino
>
>
> On Sat, Jun 17, 2023 at 8:23 AM Jayanth Reddy <jayanthreddy5666@xxxxxxxxx>
> wrote:
>
> Hello Users,
> Greetings. We've a Ceph Cluster with the version
> *ceph version 14.2.5-382-g8881d33957
> (8881d33957b54b101eae9c7627b351af10e87ee8) nautilus (stable)*
>
> 5 PGs belonging to our RGW 8+3 EC Pool are stuck in incomplete and
> incomplete+remapped states. Below are the PGs,
>
> # ceph pg dump_stuck inactive
> ok
> PG_STAT STATE               UP
>  UP_PRIMARY ACTING
>                  ACTING_PRIMARY
> 15.251e          incomplete    [151,464,146,503,166,41,555,542,9,565,268]
>      151
>  [151,464,146,503,166,41,555,542,9,565,268]            151
> 15.3f3           incomplete [584,281,672,699,199,224,239,430,355,504,196]
>      584
> [584,281,672,699,199,224,239,430,355,504,196]            584
> 15.985  remapped+incomplete  [396,690,493,214,319,209,546,91,599,237,352]
>      396
>
> [2147483647,2147483647,2147483647,214,319,2147483647,546,91,599,2147483647,352]
>            214
> 15.39d3 remapped+incomplete  [404,221,223,585,38,102,533,471,568,451,195]
>      404
>  [2147483647,2147483647,223,585,38,102,533,2147483647,231,451,2147483647]
>          223
> 15.d46  remapped+incomplete [297,646,212,254,110,169,500,372,623,470,678]
>      297
> [2147483647,548,2147483647,2147483647,110,169,500,372,2147483647,470,678]
>          548
>
> Some of the OSDs had gone down on the cluster. Below is the # ceph status
>
> # ceph -s
>   cluster:
>     id:     30d6f7ee-fa02-4ab3-8a09-9321c8002794
>     health: HEALTH_WARN
>             noscrub,nodeep-scrub flag(s) set
>             1 pools have many more objects per pg than average
>             Reduced data availability: 5 pgs inactive, 5 pgs incomplete
>             Degraded data redundancy: 44798/8718528059 objects degraded
> (0.001%), 1 pg degraded, 1 pg undersized
>             22726 pgs not deep-scrubbed in time
>             23552 pgs not scrubbed in time
>             77 slow ops, oldest one blocked for 56400 sec, daemons
> [osd.214,osd.223,osd.548,osd.584] have slow ops.
>             too many PGs per OSD (330 > max 250)
>
>   services:
>     mon: 3 daemons, quorum brc1mon2,brc1mon3,brc1mon1 (age 2y)
>     mgr: brc1mon2(active, since 8d), standbys: brc1mon1, brc1mon3
>     mds: cephfs:1 {0=brc1mds2=up:active} 1 up:standby
>     osd: 1012 osds: 698 up (since 14h), 698 in (since 2d); 3 remapped pgs
>          flags noscrub,nodeep-scrub
>     rgw: 2 daemons active (brc1rgw1, brc1rgw2)
>
>   data:
>     pools:   17 pools, 23552 pgs
>     objects: 863.74M objects, 1.2 PiB
>     usage:   2.4 PiB used, 6.2 PiB / 8.6 PiB avail
>     pgs:     0.021% pgs not active
>              44798/8718528059 objects degraded (0.001%)
>              23546 active+clean
>              3     remapped+incomplete
>              2     incomplete
>              1     active+undersized+degraded
>
>   io:
>     client:   24 MiB/s rd, 3.2 KiB/s wr, 56 op/s rd, 4 op/s wr
>
> And the health detail shows as
>
> # ceph health detail
> HEALTH_WARN noscrub,nodeep-scrub flag(s) set; 1 pools have many more
> objects per pg than average; Reduced data availability: 5 pgs inactive, 5
> pgs incomplete; Degraded data redundancy: 44798/8718528081 objects degraded
> (0.001%), 1 pg degraded, 1 pg undersized; 22726 pgs not deep-scrubbed in
> time; 23552 pgs not scrubbed in time; 77 slow ops, oldest one blocked for
> 56440 sec, daemons [osd.214,osd.223,osd.548,osd.584] have slow ops.; too
> many PGs per OSD (330 > max 250)
> OSDMAP_FLAGS noscrub,nodeep-scrub flag(s) set
> MANY_OBJECTS_PER_PG 1 pools have many more objects per pg than average
>     pool iscsi-images objects per pg (540004) is more than 14.7248 times
> cluster average (36673)
> PG_AVAILABILITY Reduced data availability: 5 pgs inactive, 5 pgs incomplete
>     pg 15.3f3 is incomplete, acting
> [584,281,672,699,199,224,239,430,355,504,196] (reducing pool
> default.rgw.buckets.data min_size from 9 may help; search ceph.com/docs
> for
> 'incomplete')
>     pg 15.985 is remapped+incomplete, acting
>
> [2147483647,2147483647,2147483647,214,319,2147483647,546,91,599,2147483647,352]
> (reducing pool default.rgw.buckets.data min_size from 9 may help; search
> ceph.com/docs for 'incomplete')
>     pg 15.d46 is remapped+incomplete, acting
> [2147483647,548,2147483647,2147483647,110,169,500,372,2147483647,470,678]
> (reducing pool default.rgw.buckets.data min_size from 9 may help; search
> ceph.com/docs for 'incomplete')
>     pg 15.251e is incomplete, acting
> [151,464,146,503,166,41,555,542,9,565,268] (reducing pool
> default.rgw.buckets.data min_size from 9 may help; search ceph.com/docs
> for
> 'incomplete')
>     pg 15.39d3 is remapped+incomplete, acting
> [2147483647,2147483647,223,585,38,102,533,2147483647,231,451,2147483647]
> (reducing pool default.rgw.buckets.data min_size from 9 may help; search
> ceph.com/docs for 'incomplete')
> PG_DEGRADED Degraded data redundancy: 44798/8718528081 objects degraded
> (0.001%), 1 pg degraded, 1 pg undersized
>     pg 15.28f0 is stuck undersized for 67359238.592403, current state
> active+undersized+degraded, last acting
> [2147483647,343,355,415,426,640,302,392,78,202,607]
> PG_NOT_DEEP_SCRUBBED 22726 pgs not deep-scrubbed in time
>
> We've the pools as below
>
> # ceph osd lspools
> 1 iscsi-images
> 2 cephfs_data
> 3 cephfs_metadata
> 4 .rgw.root
> 5 default.rgw.control
> 6 default.rgw.meta
> 7 default.rgw.log
> 8 default.rgw.buckets.index
> 13 rbd
> 15 default.rgw.buckets.data
> 16 default.rgw.buckets.non-ec
> 19 cephfs_data-ec
> 22 rbd-ec
> 23 iscsi-images-ec
> 24 hpecpool
> 25 hpec.rgw.buckets.index
> 26 hpec.rgw.buckets.non-ec
>
>
> We've been struggling for a long time to fix this but out of luck! Our RGW
> daemons hosted on dedicated machines are continuously failing to respond,
> being behind a load balancer, LB throws 504 Gateway Timeout as the daemons
> are failing to respond in the expected time. We perform active health
> checks from the LB on '/' by HTTP HEAD but these are failing as well, very
> frequently. Currently we're surviving by writing a script that restarts RGW
> daemons whenever the LB responds with HTTP status code 504. Any help is
> highly appreciated!
>
> Regards,
> Jayanth Reddy
> _______________________________________________
> ceph-users mailing list -- ceph-users@xxxxxxx
> To unsubscribe send an email to ceph-users-leave@xxxxxxx
>
>
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux