Here you go. It doesn't format very well, so I'll summarize what I'm seeing. 5.c has 78051 OMAP_BYTES and 398 OMAP_KEYS 5.16 has 80186950 OMAP_BYTES and 401505 OMAP_KEYS The remaining 30 PGS have zero of both. However, the BYTES for each PG is very much the same at around 8900000 for each. # ceph pg ls-by-pool default.rgw.meta PG OBJECTS DEGRADED MISPLACED UNFOUND BYTES OMAP_BYTES* OMAP_KEYS* LOG STATE SINCE VERSION REPORTED UP ACTING SCRUB_STAMP DEEP_SCRUB_STAMP LAST_SCRUB_DURATION SCRUB_SCHEDULING 5.0 26240 0 0 0 8909864 0 0 10076 active+clean 10h 8093'54176 8093:5396520 [21,4,12]p21 [21,4,12]p21 2023-07- 31T21:13:20.554485+0000 2023-07-26T03:40:27.457946+0000 5 periodic scrub scheduled @ 2023-08-01T23:55:14.134653+0000 5.1 26065 0 0 0 8840849 0 0 10029 active+clean 10h 8093'56529 8093:4891333 [14,7,23]p14 [14,7,23]p14 2023-07- 31T20:37:34.920128+0000 2023-07-30T10:55:16.529046+0000 5 periodic scrub scheduled @ 2023-08-01T21:12:04.440688+0000 5.2 26406 0 0 0 8943783 0 0 10076 active+clean 20h 8093'56776 8093:5022283 [26,8,25]p26 [26,8,25]p26 2023-07- 31T11:08:00.886979+0000 2023-07-30T06:03:44.341435+0000 5 periodic scrub scheduled @ 2023-08-01T16:02:32.076634+0000 5.3 26250 0 0 0 8932714 0 0 10086 active+clean 20h 8093'56786 8093:5109316 [0,26,32]p0 [0,26,32]p0 2023-07- 31T11:02:35.864452+0000 2023-07-30T04:25:56.495524+0000 5 periodic scrub scheduled @ 2023-08-01T20:18:30.975924+0000 5.4 26071 0 0 0 8874237 0 0 10024 active+clean 6h 8092'53824 8093:5146409 [15,7,34]p15 [15,7,34]p15 2023-08- 01T01:16:48.361184+0000 2023-07-25T15:47:10.627594+0000 5 periodic scrub scheduled @ 2023-08-02T12:32:06.359395+0000 5.5 26160 0 0 0 8870317 0 0 10073 active+clean 12h 8093'56173 8093:4706658 [9,31,16]p9 [9,31,16]p9 2023-07- 31T18:52:26.301525+0000 2023-07-29T08:19:00.537322+0000 5 periodic scrub scheduled @ 2023-08-02T02:11:16.267794+0000 5.6 26186 0 0 0 8904446 0 0 10084 active+clean 44m 8093'57584 8093:5032349 [7,10,38]p7 [7,10,38]p7 2023-08- 01T06:37:45.184419+0000 2023-08-01T06:37:45.184419+0000 313 periodic scrub scheduled @ 2023-08-02T10:01:32.285716+0000 5.7 26292 0 0 0 8908213 0 0 9695 active+clean 87m 8093'56896 8093:4969718 [36,1,13]p36 [36,1,13]p36 2023-08- 01T05:55:06.016287+0000 2023-07-30T21:49:33.028594+0000 5 periodic scrub scheduled @ 2023-08-02T13:54:38.778542+0000 5.8 26323 0 0 0 8911110 0 0 9747 active+clean 3h 8093'56448 8093:4981465 [36,15,2]p36 [36,15,2]p36 2023-08- 01T04:21:06.360778+0000 2023-07-29T14:46:02.363530+0000 5 periodic scrub scheduled @ 2023-08-02T10:36:48.764085+0000 5.9 26035 0 0 0 8829430 0 0 10034 active+clean 20h 8093'56335 8093:4829155 [37,21,24]p37 [37,21,24]p37 2023-07- 31T11:07:39.961751+0000 2023-07-31T11:07:39.961751+0000 309 periodic scrub scheduled @ 2023-08-01T22:42:18.862879+0000 5.a 26052 0 0 0 8859067 0 0 10087 active+clean 27h 8092'56087 8093:5022933 [2,23,10]p2 [2,23,10]p2 2023-07- 31T03:28:44.433360+0000 2023-07-31T03:28:44.433360+0000 248 periodic scrub scheduled @ 2023-08-01T13:36:39.897693+0000 5.b 25759 0 0 0 8739834 0 0 9693 active+clean 15h 8090'56293 8093:4837010 [36,7,28]p36 [36,7,28]p36 2023-07- 31T15:55:00.415967+0000 2023-07-31T15:55:00.415967+0000 323 periodic scrub scheduled @ 2023-08-01T23:41:03.756058+0000 5.c 25927 0 0 0 8788271 78051 398 10051 active+clean 24h 8093'174851 8093:4982667 [5,36,18]p5 [5,36,18]p5 2023-07- 31T07:20:32.208533+0000 2023-07-31T07:20:32.208533+0000 315 periodic scrub scheduled @ 2023-08-01T18:18:15.292651+0000 5.d 25995 0 0 0 8815306 0 0 10070 active+clean 2h 8093'57270 8093:4994478 [32,13,16]p32 [32,13,16]p32 2023-08- 01T04:27:55.863933+0000 2023-08-01T04:27:55.863933+0000 294 periodic scrub scheduled @ 2023-08-02T05:55:30.108279+0000 5.e 26253 0 0 0 8939984 0 0 10018 active+clean 5h 8092'56919 8093:5135033 [37,19,4]p37 [37,19,4]p37 2023-08- 01T01:38:15.740983+0000 2023-07-30T21:55:45.349878+0000 5 periodic scrub scheduled @ 2023-08-02T04:40:09.172157+0000 5.f 26025 0 0 0 8821973 0 0 10020 active+clean 25h 8093'53120 8093:4794909 [14,11,8]p14 [14,11,8]p14 2023-07- 31T06:22:22.849194+0000 2023-07-25T02:42:24.135997+0000 5 periodic deep scrub scheduled @ 2023-08-01T13:50:38.914828+0000 5.10 25999 0 0 0 8821303 0 0 10048 active+clean 3h 8092'55848 8093:5151525 [39,27,14]p39 [39,27,14]p39 2023-08- 01T03:40:44.355521+0000 2023-07-29T23:58:49.567904+0000 5 periodic scrub scheduled @ 2023-08-02T04:44:52.459615+0000 5.11 26200 0 0 0 8897148 0 0 8909 active+clean 23h 8093'56309 8093:4858657 [35,24,23]p35 [35,24,23]p35 2023-07- 31T08:19:21.090885+0000 2023-07-30T07:21:21.135342+0000 5 periodic scrub scheduled @ 2023-08-01T13:09:38.620801+0000 5.12 26043 0 0 0 8803496 0 0 10016 active+clean 33h 8093'52716 8093:5090415 [21,35,3]p21 [21,35,3]p21 2023-07- 30T22:10:31.308788+0000 2023-07-24T14:40:34.453392+0000 5 periodic deep scrub scheduled @ 2023-08-01T09:36:02.058413+0000 5.13 25929 0 0 0 8785411 0 0 10029 active+clean 16h 8090'54629 8093:5096641 [32,17,9]p32 [32,17,9]p32 2023-07- 31T15:19:04.119491+0000 2023-07-27T16:36:53.401620+0000 5 periodic scrub scheduled @ 2023-08-01T22:19:03.573063+0000 5.14 26102 0 0 0 8858274 0 0 10069 active+clean 89m 8093'54671 8093:4958083 [3,29,12]p3 [3,29,12]p3 2023-08- 01T05:53:20.722831+0000 2023-07-27T11:29:26.930179+0000 5 periodic scrub scheduled @ 2023-08-02T14:03:11.521746+0000 5.15 26122 0 0 0 8850708 0 0 9419 active+clean 14h 8093'57119 8093:4854254 [28,8,29]p28 [28,8,29]p28 2023-07- 31T17:04:04.790500+0000 2023-07-31T17:04:04.790500+0000 309 periodic scrub scheduled @ 2023-08-02T04:29:01.728903+0000 5.16 26168 0 0 0 8869093 80186950 401505 10031 active+clean 32h 8093'122396977 8093:127435157 [26,39,9]p26 [26,39,9]p26 2023-07- 30T23:19:12.784044+0000 2023-07-30T23:19:12.784044+0000 258 periodic scrub scheduled @ 2023-08-01T10:29:11.462563+0000 5.17 25504 0 0 0 8634818 0 0 10081 active+clean 23h 8093'55481 8093:4742014 [4,25,34]p4 [4,25,34]p4 2023-07- 31T08:11:23.105601+0000 2023-07-31T08:11:23.105601+0000 309 periodic scrub scheduled @ 2023-08-01T15:11:19.240302+0000 5.18 26143 0 0 0 8846680 0 0 10014 active+clean 23h 8093'55015 8093:4927120 [22,11,36]p22 [22,11,36]p22 2023-07- 31T08:18:27.757381+0000 2023-07-27T19:04:02.036522+0000 5 periodic scrub scheduled @ 2023-08-01T16:25:12.741673+0000 5.19 26117 0 0 0 8864860 0 0 10073 active+clean 27h 8093'55173 8093:5001362 [1,28,27]p1 [1,28,27]p1 2023-07- 31T03:37:40.525594+0000 2023-07-28T10:50:18.232627+0000 5 periodic scrub scheduled @ 2023-08-01T15:06:48.300456+0000 5.1a 26186 0 0 0 8870466 0 0 10025 active+clean 31h 8093'54025 8093:4991279 [34,17,22]p34 [34,17,22]p34 2023-07- 31T00:20:28.853158+0000 2023-07-25T21:31:55.045662+0000 5 periodic scrub scheduled @ 2023-08-01T09:42:20.646380+0000 5.1b 26070 0 0 0 8854703 0 0 10087 active+clean 8h 8093'56487 8093:4886996 [22,7,37]p22 [22,7,37]p22 2023-07- 31T22:39:17.793412+0000 2023-07-30T16:07:40.211725+0000 6 periodic scrub scheduled @ 2023-08-01T23:23:58.360930+0000 5.1c 26302 0 0 0 8925675 0 0 10015 active+clean 16h 8093'54915 8093:4986627 [33,11,2]p33 [33,11,2]p33 2023-07- 31T15:07:27.474683+0000 2023-07-27T15:11:42.794360+0000 5 periodic scrub scheduled @ 2023-08-01T15:51:01.248015+0000 5.1d 26075 0 0 0 8839499 0 0 10075 active+clean 21h 8093'52575 8093:4857107 [33,16,17]p33 [33,16,17]p33 2023-07- 31T09:23:24.577919+0000 2023-07-24T16:52:37.965968+0000 5 periodic deep scrub scheduled @ 2023-08-01T11:33:35.067584+0000 5.1e 26182 0 0 0 8905099 0 0 8905 active+clean+scrubbing+deep 4m 8093'56905 8093:4936987 [35,20,15]p35 [35,20,15]p35 2023-07- 31T04:25:51.171690+0000 2023-07-24T14:28:19.531701+0000 5 deep scrubbing for 259s 5.1f 25978 0 0 0 8796478 0 0 10068 active+clean 13h 8092'56868 8093:4813791 [26,30,13]p26 [26,30,13]p26 2023-07- 31T17:50:40.349450+0000 2023-07-31T17:50:40.349450+0000 311 periodic scrub scheduled @ 2023-08-02T04:39:41.913504+0000 On Tue, 2023-08-01 at 06:14 +0000, Eugen Block wrote: > Yeah, regarding data distribution increasing the pg_num of the data > pool is recommended. But could you also share the output of: > > ceph pg ls-by-pool default.rgw.meta > > That's where the large omap was reported, maybe you'll need to > increase the pg_num for that pool as well. Personally, I always > disable the autoscaler. > > Zitat von Mark Johnson <markj@xxxxxxxxx>: > > > Thanks Bailey, > > > > With regards to the PG count, we've been relying on PG autoscale > > and > > it is currently enabled. I figure I'd need to disable autoscale > > and > > manually increase the PG count on the default.rgw.buckets.data > > pool, > > correct? We're coming from our existing clusters running Jewel to > > this new Quincy cluster and have no prior experience with > > autoscale > > so we were just assuming autoscale would manage PG counts better > > than us doing it manually. As you can probably guess, we don't > > have > > much experience with Ceph. > > > > Regards, > > Mark Johnson > > > > > > On Mon, 2023-07-31 at 21:54 -0300, Bailey Allison wrote: > > [You don't often get email from > > ballison@xxxxxxxxxxxx<mailto:ballison@xxxxxxxxxxxx>. Learn why > > this > > is important at https://aka.ms/LearnAboutSenderIdentification ;] > > > > Hi, > > > > It appears you have quite a low PG count on your cluster (approx. > > 20 > > PGs per each OSD). > > > > Usually is recommended to have about 100-150 per each OSD. With a > > lower PG count you can have issues with balancing data and cause > > errors such as large OMAP objects. > > > > Might not be the fix in this case but either way would still > > recommend increasing PGs on your pools. > > > > If you look at the OMAP value in your ceph osd df you can see that > > some OSDs have 2GB while some have 500MB. Even for data some > > drives > > are holding 900GB while others 2TB. > > > > You will have to issue a deep-scrub on the PGs as well to get > > updated OMAP data once the PGs are increased. > > > > Regards, > > > > Bailey > > > > -----Original Message----- > > From: Mark Johnson <markj@xxxxxxxxx<mailto:markj@xxxxxxxxx>> > > Sent: July 31, 2023 9:01 PM > > To: eblock@xxxxxx<mailto:eblock@xxxxxx>; > > ceph-users@xxxxxxx<mailto:ceph-users@xxxxxxx> > > Subject: Re: 1 Large omap object found > > > > Sure thing. Thanks for the reply. > > > > ceph df > > > > --- RAW STORAGE --- > > CLASS SIZE AVAIL USED RAW USED %RAW USED > > hdd 291 TiB 244 TiB 47 TiB 47 TiB 16.02 > > TOTAL 291 TiB 244 TiB 47 TiB 47 TiB 16.02 > > > > --- POOLS --- > > POOL ID PGS STORED OBJECTS USED > > %USED > > MAX AVAIL > > .mgr 1 1 459 MiB 116 1.3 GiB > > 0 > > 65 TiB > > .rgw.root 2 32 1.3 KiB 4 48 KiB > > 0 > > 65 TiB > > default.rgw.log 3 32 5.3 KiB 209 468 KiB > > 0 > > 65 TiB > > default.rgw.control 4 32 0 B 8 0 B > > 0 > > 65 TiB > > default.rgw.meta 5 32 452 MiB 828.75k 10 GiB > > 0 > > 65 TiB > > default.rgw.buckets.index 6 32 17 GiB 4.56M 51 GiB > > 0.03 > > 65 TiB > > default.rgw.buckets.data 7 128 15 TiB 54.51M 46 TiB > > 19.24 > > 65 TiB > > cephfs_metadata 8 16 258 MiB 98 775 MiB > > 0 > > 65 TiB > > cephfs_data 9 32 1.9 GiB 998 5.6 GiB > > 0 > > 65 TiB > > > > > > ceph osd df > > > > ID CLASS WEIGHT REWEIGHT SIZE RAW USE DATA OMAP > > META > > AVAIL %USE VAR PGS STATUS > > 0 hdd 7.27739 1.00000 7.3 TiB 1.6 TiB 1.6 TiB 550 MiB > > 12 > > GiB 5.7 TiB 21.70 1.35 21 up > > 1 hdd 7.27739 1.00000 7.3 TiB 995 GiB 986 GiB 1.1 GiB > > 7.6 > > GiB 6.3 TiB 13.35 0.83 28 up > > 2 hdd 7.27739 1.00000 7.3 TiB 996 GiB 986 GiB 2.1 GiB > > 7.9 > > GiB 6.3 TiB 13.37 0.83 22 up > > 3 hdd 7.27739 1.00000 7.3 TiB 1.3 TiB 1.3 TiB 513 MiB > > 10 > > GiB 5.9 TiB 18.35 1.15 28 up > > 4 hdd 7.27739 1.00000 7.3 TiB 1.1 TiB 1.1 TiB 527 MiB > > 8.3 > > GiB 6.2 TiB 15.02 0.94 22 up > > 5 hdd 7.27739 1.00000 7.3 TiB 1.8 TiB 1.8 TiB 1.5 GiB > > 14 > > GiB 5.5 TiB 25.01 1.56 28 up > > 6 hdd 7.27739 1.00000 7.3 TiB 746 GiB 739 GiB 1.0 GiB > > 5.8 > > GiB 6.5 TiB 10.01 0.63 20 up > > 7 hdd 7.27739 1.00000 7.3 TiB 1.1 TiB 1.1 TiB 1.1 GiB > > 8.8 > > GiB 6.2 TiB 15.04 0.94 20 up > > 8 hdd 7.27739 1.00000 7.3 TiB 871 GiB 864 GiB 544 MiB > > 6.7 > > GiB 6.4 TiB 11.69 0.73 27 up > > 9 hdd 7.27739 1.00000 7.3 TiB 1.3 TiB 1.3 TiB 3.3 GiB > > 11 > > GiB 5.9 TiB 18.37 1.15 28 up > > 30 hdd 7.27739 1.00000 7.3 TiB 1.8 TiB 1.8 TiB 1.6 GiB > > 14 > > GiB 5.5 TiB 25.01 1.56 35 up > > 31 hdd 7.27739 1.00000 7.3 TiB 747 GiB 739 GiB 2.2 GiB > > 6.2 > > GiB 6.5 TiB 10.03 0.63 20 up > > 32 hdd 7.27739 1.00000 7.3 TiB 996 GiB 987 GiB 1.5 GiB > > 7.9 > > GiB 6.3 TiB 13.37 0.83 26 up > > 33 hdd 7.27739 1.00000 7.3 TiB 995 GiB 985 GiB 1.5 GiB > > 7.7 > > GiB 6.3 TiB 13.35 0.83 25 up > > 34 hdd 7.27739 1.00000 7.3 TiB 750 GiB 742 GiB 2.1 GiB > > 5.7 > > GiB 6.5 TiB 10.07 0.63 25 up > > 35 hdd 7.27739 1.00000 7.3 TiB 2.1 TiB 2.0 TiB 571 MiB > > 15 > > GiB 5.2 TiB 28.36 1.77 34 up > > 36 hdd 7.27739 1.00000 7.3 TiB 1.3 TiB 1.3 TiB 1.5 GiB > > 10 > > GiB 5.9 TiB 18.37 1.15 31 up > > 37 hdd 7.27739 1.00000 7.3 TiB 1.1 TiB 1.1 TiB 524 MiB > > 8.2 > > GiB 6.2 TiB 14.99 0.94 26 up > > 38 hdd 7.27739 1.00000 7.3 TiB 1.6 TiB 1.6 TiB 1.6 GiB > > 12 > > GiB 5.7 TiB 21.70 1.35 28 up > > 39 hdd 7.27739 1.00000 7.3 TiB 1.5 TiB 1.4 TiB 2.4 GiB > > 11 > > GiB 5.8 TiB 20.04 1.25 30 up > > 10 hdd 7.27739 1.00000 7.3 TiB 1.3 TiB 1.3 TiB 1.6 GiB > > 10 > > GiB 5.9 TiB 18.34 1.14 26 up > > 12 hdd 7.27739 1.00000 7.3 TiB 1.3 TiB 1.3 TiB 1 KiB > > 9.9 > > GiB 5.9 TiB 18.37 1.15 25 up > > 14 hdd 7.27739 1.00000 7.3 TiB 1.5 TiB 1.4 TiB 593 MiB > > 10 > > GiB 5.8 TiB 19.98 1.25 22 up > > 16 hdd 7.27739 1.00000 7.3 TiB 997 GiB 987 GiB 2.2 GiB > > 7.5 > > GiB 6.3 TiB 13.38 0.84 19 up > > 18 hdd 7.27739 1.00000 7.3 TiB 1.1 TiB 1.1 TiB 1.1 GiB > > 8.7 > > GiB 6.2 TiB 15.02 0.94 26 up > > 20 hdd 7.27739 1.00000 7.3 TiB 1.6 TiB 1.6 TiB 1.1 GiB > > 12 > > GiB 5.7 TiB 21.68 1.35 26 up > > 22 hdd 7.27739 1.00000 7.3 TiB 1.3 TiB 1.3 TiB 559 MiB > > 10 > > GiB 5.9 TiB 18.34 1.14 22 up > > 24 hdd 7.27739 1.00000 7.3 TiB 872 GiB 864 GiB 1020 MiB > > 6.8 > > GiB 6.4 TiB 11.70 0.73 23 up > > 26 hdd 7.27739 1.00000 7.3 TiB 749 GiB 741 GiB 1.8 GiB > > 6.3 > > GiB 6.5 TiB 10.05 0.63 25 up > > 28 hdd 7.27739 1.00000 7.3 TiB 1.3 TiB 1.3 TiB 1.5 GiB > > 10 > > GiB 5.9 TiB 18.36 1.15 32 up > > 11 hdd 7.27739 1.00000 7.3 TiB 1.1 TiB 1.1 TiB 2.6 GiB > > 8.5 > > GiB 6.2 TiB 15.02 0.94 23 up > > 13 hdd 7.27739 1.00000 7.3 TiB 1.3 TiB 1.3 TiB 2.2 GiB > > 10 > > GiB 5.9 TiB 18.38 1.15 36 up > > 15 hdd 7.27739 1.00000 7.3 TiB 995 GiB 986 GiB 1.1 GiB > > 7.7 > > GiB 6.3 TiB 13.35 0.83 25 up > > 17 hdd 7.27739 1.00000 7.3 TiB 623 GiB 618 GiB 419 KiB > > 5.0 > > GiB 6.7 TiB 8.35 0.52 23 up > > 19 hdd 7.27739 1.00000 7.3 TiB 870 GiB 863 GiB 513 MiB > > 6.6 > > GiB 6.4 TiB 11.67 0.73 21 up > > 21 hdd 7.27739 1.00000 7.3 TiB 1.1 TiB 1.1 TiB 1.5 GiB > > 8.6 > > GiB 6.2 TiB 15.02 0.94 25 up > > 23 hdd 7.27739 1.00000 7.3 TiB 746 GiB 739 GiB 564 MiB > > 5.8 > > GiB 6.5 TiB 10.01 0.62 22 up > > 25 hdd 7.27739 1.00000 7.3 TiB 1.1 TiB 1.1 TiB 2.1 GiB > > 8.4 > > GiB 6.2 TiB 15.03 0.94 24 up > > 27 hdd 7.27739 1.00000 7.3 TiB 1.2 TiB 1.2 TiB 532 MiB > > 9.1 > > GiB 6.1 TiB 16.68 1.04 23 up > > 29 hdd 7.27739 1.00000 7.3 TiB 1.1 TiB 1.1 TiB 1.1 GiB > > 8.4 > > GiB 6.2 TiB 14.99 0.94 19 up > > TOTAL 291 TiB 47 TiB 46 TiB 51 GiB > > 359 > > GiB 244 TiB 16.02 > > MIN/MAX VAR: 0.52/1.77 STDDEV: 4.56 > > > > > > On Mon, 2023-07-31 at 09:22 +0000, Eugen Block wrote: > > Hi, > > > > can you share some more details like 'ceph df' and 'ceph osd df'? I > > don't have too much advice yet, but to see all entries in your meta > > pool you need add the --all flag because those objects are stored > > in > > namespaces: > > > > rados -p default.rgw.meta ls --all > > > > That pool contains user and bucket information (example): > > > > # rados -p default.rgw.meta ls --all > > users.uid admin.buckets > > users.keys c0fba3ea7d9c4321b5205752c85baa85 users.uid > > admin > > users.keys JBWPRAPP1AQG471AMGC4 users.uid > > e434b82737cf4138b899c0785b49112d.buckets > > users.uid e434b82737cf4138b899c0785b49112d > > > > > > > > Zitat von Mark Johnson <markj@xxxxxxxxx<mailto:markj@xxxxxxxxx>>: > > > > I've been going round and round in circles trying to work this one > > out but I'm getting nowhere. We're running a 4 node quincy cluster > > (17.2.6) which recently reported the following: > > > > ceph.log-20230729.gz:2023-07-28T08:31:42.390003+0000 osd.26 > > (osd.26) > > 13834 : cluster [WRN] Large omap object found. Object: > > 5:6c65dd84:users.uid::callrecordings$callrecordings_rw.buckets:head > > > > PG: 5.21bba636 (5.16) Key count: 378454 Size (bytes): 75565579 > > > > This happened a week or so ago (only the key count was only just > > over the 200000 threshold on that occasion) and after much > > searching > > around, I found an article that suggested a deep scrub on the pg > > would likely resolve the issue, so I forced a deep scrub and > > shortly > > after, the warning cleared. Came into the office today to discover > > the above. It's on the same PG as before which is in the > > default.rgw.meta pool. This time, after forcing a deep-scrub on > > that PG, nothing changed. I did it a second time just to be sure > > but got the same result. > > > > I keep finding a suse article that simply suggests increasing the > > threshold to the previous default of 2,000,000, but other articles > > I > > read say it was lowered for a reason and that by the time it hits > > that figure, it's too late so I don't want to just mask it. > > Problem > > is that I don't really understand it. I found a thread here from > > a > > bit over two years ago but their issue was in the > > default.rgw.buckets.index pool. A step in the solution was to list > > out the problematic object id and check the objects per shard > > however, if I issue the command "rados -p default.rgw.meta ls" it > > returns nothing. I get a big list from "rados -p > > default.rgw.buckets.index ls" just nothing from the first pool. I > > think it may be because the meta pool isn't indexed based on > > something I read, but I really don't know what I'm talking about > > tbh. > > > > I don't know if this is helpful, but if I list out all the PGs for > > that pool, there are 32 PGs and 5.16 shows 80186950 bytes and > > 401505 > > keys. PG 5.c has 75298 and 384 keys. The remaining 30 PGs show > > zero bytes and zero keys. I'm really not sure how to troubleshoot > > and resolve from here. For the record, dynamic resharding is > > enabled in that no options have been set in the config and that is > > the default setting. > > > > Based on the suse article I mentioned which also references the > > default.rgw.meta pool, I'm gathering our issue is because we have > > so > > many buckets that are all owned by the one user and the solution is > > either: > > > > * delete unused buckets > > * create multiple users and spread buckets evenly across all users > > (not something we can do) > > * increase the threshold to stop the warning > > > > Problem is that I'm having trouble verifying this is the issue. > > I've > > tried dumping out bucket stats to a file (radosgw-admin bucket > > stats > > bucket_stats.txt) but after three hours this is still running with > > no output. > > > > Thanks for your time, > > Mark > > _______________________________________________ > > ceph-users mailing list -- > > ceph-users@xxxxxxx<mailto:ceph-users@xxxxxxx> To unsubscribe send > > an > > email to ceph-users-leave@xxxxxxx<mailto:ceph-users-leave@xxxxxxx> > > > > > > _______________________________________________ > > ceph-users mailing list -- > > ceph-users@xxxxxxx<mailto:ceph-users@xxxxxxx> To unsubscribe send > > an > > email to ceph-users-leave@xxxxxxx<mailto:ceph-users-leave@xxxxxxx> > > > > _______________________________________________ > > ceph-users mailing list -- > > ceph-users@xxxxxxx<mailto:ceph-users@xxxxxxx> To unsubscribe send > > an > > email to ceph-users-leave@xxxxxxx<mailto:ceph-users-leave@xxxxxxx> > > > _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx