Re: 1 Large omap object found

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Here you go.  It doesn't format very well, so I'll summarize what I'm
seeing.  

5.c has 78051 OMAP_BYTES and 398 OMAP_KEYS
5.16 has 80186950 OMAP_BYTES and 401505 OMAP_KEYS

The remaining 30 PGS have zero of both.  However, the BYTES for each PG
is very much the same at around 8900000 for each.


# ceph pg ls-by-pool default.rgw.meta

PG    OBJECTS  DEGRADED  MISPLACED  UNFOUND  BYTES    OMAP_BYTES* 
OMAP_KEYS*  LOG    STATE                        SINCE  VERSION        
REPORTED        UP             ACTING         SCRUB_STAMP             
DEEP_SCRUB_STAMP                 LAST_SCRUB_DURATION  SCRUB_SCHEDULING
5.0     26240         0          0        0  8909864            0     
0  10076                 active+clean    10h      8093'54176   
8093:5396520   [21,4,12]p21   [21,4,12]p21  2023-07-
31T21:13:20.554485+0000  2023-07-26T03:40:27.457946+0000              
5  periodic scrub scheduled @ 2023-08-01T23:55:14.134653+0000     
5.1     26065         0          0        0  8840849            0     
0  10029                 active+clean    10h      8093'56529   
8093:4891333   [14,7,23]p14   [14,7,23]p14  2023-07-
31T20:37:34.920128+0000  2023-07-30T10:55:16.529046+0000              
5  periodic scrub scheduled @ 2023-08-01T21:12:04.440688+0000     
5.2     26406         0          0        0  8943783            0     
0  10076                 active+clean    20h      8093'56776   
8093:5022283   [26,8,25]p26   [26,8,25]p26  2023-07-
31T11:08:00.886979+0000  2023-07-30T06:03:44.341435+0000              
5  periodic scrub scheduled @ 2023-08-01T16:02:32.076634+0000     
5.3     26250         0          0        0  8932714            0     
0  10086                 active+clean    20h      8093'56786   
8093:5109316    [0,26,32]p0    [0,26,32]p0  2023-07-
31T11:02:35.864452+0000  2023-07-30T04:25:56.495524+0000              
5  periodic scrub scheduled @ 2023-08-01T20:18:30.975924+0000     
5.4     26071         0          0        0  8874237            0     
0  10024                 active+clean     6h      8092'53824   
8093:5146409   [15,7,34]p15   [15,7,34]p15  2023-08-
01T01:16:48.361184+0000  2023-07-25T15:47:10.627594+0000              
5  periodic scrub scheduled @ 2023-08-02T12:32:06.359395+0000     
5.5     26160         0          0        0  8870317            0     
0  10073                 active+clean    12h      8093'56173   
8093:4706658    [9,31,16]p9    [9,31,16]p9  2023-07-
31T18:52:26.301525+0000  2023-07-29T08:19:00.537322+0000              
5  periodic scrub scheduled @ 2023-08-02T02:11:16.267794+0000     
5.6     26186         0          0        0  8904446            0     
0  10084                 active+clean    44m      8093'57584   
8093:5032349    [7,10,38]p7    [7,10,38]p7  2023-08-
01T06:37:45.184419+0000  2023-08-01T06:37:45.184419+0000              
313  periodic scrub scheduled @ 2023-08-02T10:01:32.285716+0000     
5.7     26292         0          0        0  8908213            0     
0   9695                 active+clean    87m      8093'56896   
8093:4969718   [36,1,13]p36   [36,1,13]p36  2023-08-
01T05:55:06.016287+0000  2023-07-30T21:49:33.028594+0000              
5  periodic scrub scheduled @ 2023-08-02T13:54:38.778542+0000     
5.8     26323         0          0        0  8911110            0     
0   9747                 active+clean     3h      8093'56448   
8093:4981465   [36,15,2]p36   [36,15,2]p36  2023-08-
01T04:21:06.360778+0000  2023-07-29T14:46:02.363530+0000              
5  periodic scrub scheduled @ 2023-08-02T10:36:48.764085+0000     
5.9     26035         0          0        0  8829430            0     
0  10034                 active+clean    20h      8093'56335   
8093:4829155  [37,21,24]p37  [37,21,24]p37  2023-07-
31T11:07:39.961751+0000  2023-07-31T11:07:39.961751+0000              
309  periodic scrub scheduled @ 2023-08-01T22:42:18.862879+0000     
5.a     26052         0          0        0  8859067            0     
0  10087                 active+clean    27h      8092'56087   
8093:5022933    [2,23,10]p2    [2,23,10]p2  2023-07-
31T03:28:44.433360+0000  2023-07-31T03:28:44.433360+0000              
248  periodic scrub scheduled @ 2023-08-01T13:36:39.897693+0000     
5.b     25759         0          0        0  8739834            0     
0   9693                 active+clean    15h      8090'56293   
8093:4837010   [36,7,28]p36   [36,7,28]p36  2023-07-
31T15:55:00.415967+0000  2023-07-31T15:55:00.415967+0000              
323  periodic scrub scheduled @ 2023-08-01T23:41:03.756058+0000     
5.c     25927         0          0        0  8788271        78051     
398  10051                 active+clean    24h     8093'174851   
8093:4982667    [5,36,18]p5    [5,36,18]p5  2023-07-
31T07:20:32.208533+0000  2023-07-31T07:20:32.208533+0000              
315  periodic scrub scheduled @ 2023-08-01T18:18:15.292651+0000     
5.d     25995         0          0        0  8815306            0     
0  10070                 active+clean     2h      8093'57270   
8093:4994478  [32,13,16]p32  [32,13,16]p32  2023-08-
01T04:27:55.863933+0000  2023-08-01T04:27:55.863933+0000              
294  periodic scrub scheduled @ 2023-08-02T05:55:30.108279+0000     
5.e     26253         0          0        0  8939984            0     
0  10018                 active+clean     5h      8092'56919   
8093:5135033   [37,19,4]p37   [37,19,4]p37  2023-08-
01T01:38:15.740983+0000  2023-07-30T21:55:45.349878+0000              
5  periodic scrub scheduled @ 2023-08-02T04:40:09.172157+0000     
5.f     26025         0          0        0  8821973            0     
0  10020                 active+clean    25h      8093'53120   
8093:4794909   [14,11,8]p14   [14,11,8]p14  2023-07-
31T06:22:22.849194+0000  2023-07-25T02:42:24.135997+0000              
5  periodic deep scrub scheduled @ 2023-08-01T13:50:38.914828+0000
5.10    25999         0          0        0  8821303            0     
0  10048                 active+clean     3h      8092'55848   
8093:5151525  [39,27,14]p39  [39,27,14]p39  2023-08-
01T03:40:44.355521+0000  2023-07-29T23:58:49.567904+0000              
5  periodic scrub scheduled @ 2023-08-02T04:44:52.459615+0000     
5.11    26200         0          0        0  8897148            0     
0   8909                 active+clean    23h      8093'56309   
8093:4858657  [35,24,23]p35  [35,24,23]p35  2023-07-
31T08:19:21.090885+0000  2023-07-30T07:21:21.135342+0000              
5  periodic scrub scheduled @ 2023-08-01T13:09:38.620801+0000     
5.12    26043         0          0        0  8803496            0     
0  10016                 active+clean    33h      8093'52716   
8093:5090415   [21,35,3]p21   [21,35,3]p21  2023-07-
30T22:10:31.308788+0000  2023-07-24T14:40:34.453392+0000              
5  periodic deep scrub scheduled @ 2023-08-01T09:36:02.058413+0000
5.13    25929         0          0        0  8785411            0     
0  10029                 active+clean    16h      8090'54629   
8093:5096641   [32,17,9]p32   [32,17,9]p32  2023-07-
31T15:19:04.119491+0000  2023-07-27T16:36:53.401620+0000              
5  periodic scrub scheduled @ 2023-08-01T22:19:03.573063+0000     
5.14    26102         0          0        0  8858274            0     
0  10069                 active+clean    89m      8093'54671   
8093:4958083    [3,29,12]p3    [3,29,12]p3  2023-08-
01T05:53:20.722831+0000  2023-07-27T11:29:26.930179+0000              
5  periodic scrub scheduled @ 2023-08-02T14:03:11.521746+0000     
5.15    26122         0          0        0  8850708            0     
0   9419                 active+clean    14h      8093'57119   
8093:4854254   [28,8,29]p28   [28,8,29]p28  2023-07-
31T17:04:04.790500+0000  2023-07-31T17:04:04.790500+0000              
309  periodic scrub scheduled @ 2023-08-02T04:29:01.728903+0000     
5.16    26168         0          0        0  8869093     80186950     
401505  10031                 active+clean    32h  8093'122396977 
8093:127435157   [26,39,9]p26   [26,39,9]p26  2023-07-
30T23:19:12.784044+0000  2023-07-30T23:19:12.784044+0000              
258  periodic scrub scheduled @ 2023-08-01T10:29:11.462563+0000     
5.17    25504         0          0        0  8634818            0     
0  10081                 active+clean    23h      8093'55481   
8093:4742014    [4,25,34]p4    [4,25,34]p4  2023-07-
31T08:11:23.105601+0000  2023-07-31T08:11:23.105601+0000              
309  periodic scrub scheduled @ 2023-08-01T15:11:19.240302+0000     
5.18    26143         0          0        0  8846680            0     
0  10014                 active+clean    23h      8093'55015   
8093:4927120  [22,11,36]p22  [22,11,36]p22  2023-07-
31T08:18:27.757381+0000  2023-07-27T19:04:02.036522+0000              
5  periodic scrub scheduled @ 2023-08-01T16:25:12.741673+0000     
5.19    26117         0          0        0  8864860            0     
0  10073                 active+clean    27h      8093'55173   
8093:5001362    [1,28,27]p1    [1,28,27]p1  2023-07-
31T03:37:40.525594+0000  2023-07-28T10:50:18.232627+0000              
5  periodic scrub scheduled @ 2023-08-01T15:06:48.300456+0000     
5.1a    26186         0          0        0  8870466            0     
0  10025                 active+clean    31h      8093'54025   
8093:4991279  [34,17,22]p34  [34,17,22]p34  2023-07-
31T00:20:28.853158+0000  2023-07-25T21:31:55.045662+0000              
5  periodic scrub scheduled @ 2023-08-01T09:42:20.646380+0000     
5.1b    26070         0          0        0  8854703            0     
0  10087                 active+clean     8h      8093'56487   
8093:4886996   [22,7,37]p22   [22,7,37]p22  2023-07-
31T22:39:17.793412+0000  2023-07-30T16:07:40.211725+0000              
6  periodic scrub scheduled @ 2023-08-01T23:23:58.360930+0000     
5.1c    26302         0          0        0  8925675            0     
0  10015                 active+clean    16h      8093'54915   
8093:4986627   [33,11,2]p33   [33,11,2]p33  2023-07-
31T15:07:27.474683+0000  2023-07-27T15:11:42.794360+0000              
5  periodic scrub scheduled @ 2023-08-01T15:51:01.248015+0000     
5.1d    26075         0          0        0  8839499            0     
0  10075                 active+clean    21h      8093'52575   
8093:4857107  [33,16,17]p33  [33,16,17]p33  2023-07-
31T09:23:24.577919+0000  2023-07-24T16:52:37.965968+0000              
5  periodic deep scrub scheduled @ 2023-08-01T11:33:35.067584+0000
5.1e    26182         0          0        0  8905099            0     
0   8905  active+clean+scrubbing+deep     4m      8093'56905   
8093:4936987  [35,20,15]p35  [35,20,15]p35  2023-07-
31T04:25:51.171690+0000  2023-07-24T14:28:19.531701+0000              
5  deep scrubbing for 259s                                        
5.1f    25978         0          0        0  8796478            0     
0  10068                 active+clean    13h      8092'56868   
8093:4813791  [26,30,13]p26  [26,30,13]p26  2023-07-
31T17:50:40.349450+0000  2023-07-31T17:50:40.349450+0000              
311  periodic scrub scheduled @ 2023-08-02T04:39:41.913504+0000     


On Tue, 2023-08-01 at 06:14 +0000, Eugen Block wrote:
> Yeah, regarding data distribution increasing the pg_num of the data  
> pool is recommended. But could you also share the output of:
> 
> ceph pg ls-by-pool default.rgw.meta
> 
> That's where the large omap was reported, maybe you'll need to  
> increase the pg_num for that pool as well. Personally, I always  
> disable the autoscaler.
> 
> Zitat von Mark Johnson <markj@xxxxxxxxx>:
> 
> > Thanks Bailey,
> > 
> > With regards to the PG count, we've been relying on PG autoscale
> > and  
> > it is currently enabled.  I figure I'd need to disable autoscale
> > and  
> > manually increase the PG count on the default.rgw.buckets.data
> > pool,  
> > correct?  We're coming from our existing clusters running Jewel to 
> > this new Quincy cluster and have no prior experience with
> > autoscale  
> > so we were just assuming autoscale would manage PG counts better  
> > than us doing it manually.  As you can probably guess, we don't
> > have  
> > much experience with Ceph.
> > 
> > Regards,
> > Mark Johnson
> > 
> > 
> > On Mon, 2023-07-31 at 21:54 -0300, Bailey Allison wrote:
> > [You don't often get email from  
> > ballison@xxxxxxxxxxxx<mailto:ballison@xxxxxxxxxxxx>. Learn why
> > this  
> > is important at https://aka.ms/LearnAboutSenderIdentification ;]
> > 
> > Hi,
> > 
> > It appears you have quite a low PG count on your cluster (approx.
> > 20  
> > PGs per each OSD).
> > 
> > Usually is recommended to have about 100-150 per each OSD. With a  
> > lower PG count you can have issues with balancing data and cause  
> > errors such as large OMAP objects.
> > 
> > Might not be the fix in this case but either way would still  
> > recommend increasing PGs on your pools.
> > 
> > If you look at the OMAP value in your ceph osd df you can see that 
> > some OSDs have 2GB while some have 500MB. Even for data some
> > drives  
> > are holding 900GB while others 2TB.
> > 
> > You will have to issue a deep-scrub on the PGs as well to get  
> > updated OMAP data once the PGs are increased.
> > 
> > Regards,
> > 
> > Bailey
> > 
> > -----Original Message-----
> > From: Mark Johnson <markj@xxxxxxxxx<mailto:markj@xxxxxxxxx>>
> > Sent: July 31, 2023 9:01 PM
> > To: eblock@xxxxxx<mailto:eblock@xxxxxx>;  
> > ceph-users@xxxxxxx<mailto:ceph-users@xxxxxxx>
> > Subject:  Re: 1 Large omap object found
> > 
> > Sure thing.  Thanks for the reply.
> > 
> > ceph df
> > 
> > --- RAW STORAGE ---
> > CLASS     SIZE    AVAIL    USED  RAW USED  %RAW USED
> > hdd    291 TiB  244 TiB  47 TiB    47 TiB      16.02
> > TOTAL  291 TiB  244 TiB  47 TiB    47 TiB      16.02
> > 
> > --- POOLS ---
> > POOL                       ID  PGS   STORED  OBJECTS     USED 
> > %USED
> > MAX AVAIL
> > .mgr                        1    1  459 MiB      116  1.3 GiB     
> > 0
> > 65 TiB
> > .rgw.root                   2   32  1.3 KiB        4   48 KiB     
> > 0
> > 65 TiB
> > default.rgw.log             3   32  5.3 KiB      209  468 KiB     
> > 0
> > 65 TiB
> > default.rgw.control         4   32      0 B        8      0 B     
> > 0
> > 65 TiB
> > default.rgw.meta            5   32  452 MiB  828.75k   10 GiB     
> > 0
> > 65 TiB
> > default.rgw.buckets.index   6   32   17 GiB    4.56M   51 GiB  
> > 0.03
> > 65 TiB
> > default.rgw.buckets.data    7  128   15 TiB   54.51M   46 TiB 
> > 19.24
> > 65 TiB
> > cephfs_metadata             8   16  258 MiB       98  775 MiB     
> > 0
> > 65 TiB
> > cephfs_data                 9   32  1.9 GiB      998  5.6 GiB     
> > 0
> > 65 TiB
> > 
> > 
> > ceph osd df
> > 
> > ID  CLASS  WEIGHT   REWEIGHT  SIZE     RAW USE  DATA     OMAP     
> > META
> > AVAIL    %USE   VAR   PGS  STATUS
> > 0    hdd  7.27739   1.00000  7.3 TiB  1.6 TiB  1.6 TiB   550 MiB  
> > 12
> > GiB  5.7 TiB  21.70  1.35   21      up
> > 1    hdd  7.27739   1.00000  7.3 TiB  995 GiB  986 GiB   1.1 GiB 
> > 7.6
> > GiB  6.3 TiB  13.35  0.83   28      up
> > 2    hdd  7.27739   1.00000  7.3 TiB  996 GiB  986 GiB   2.1 GiB 
> > 7.9
> > GiB  6.3 TiB  13.37  0.83   22      up
> > 3    hdd  7.27739   1.00000  7.3 TiB  1.3 TiB  1.3 TiB   513 MiB  
> > 10
> > GiB  5.9 TiB  18.35  1.15   28      up
> > 4    hdd  7.27739   1.00000  7.3 TiB  1.1 TiB  1.1 TiB   527 MiB 
> > 8.3
> > GiB  6.2 TiB  15.02  0.94   22      up
> > 5    hdd  7.27739   1.00000  7.3 TiB  1.8 TiB  1.8 TiB   1.5 GiB  
> > 14
> > GiB  5.5 TiB  25.01  1.56   28      up
> > 6    hdd  7.27739   1.00000  7.3 TiB  746 GiB  739 GiB   1.0 GiB 
> > 5.8
> > GiB  6.5 TiB  10.01  0.63   20      up
> > 7    hdd  7.27739   1.00000  7.3 TiB  1.1 TiB  1.1 TiB   1.1 GiB 
> > 8.8
> > GiB  6.2 TiB  15.04  0.94   20      up
> > 8    hdd  7.27739   1.00000  7.3 TiB  871 GiB  864 GiB   544 MiB 
> > 6.7
> > GiB  6.4 TiB  11.69  0.73   27      up
> > 9    hdd  7.27739   1.00000  7.3 TiB  1.3 TiB  1.3 TiB   3.3 GiB  
> > 11
> > GiB  5.9 TiB  18.37  1.15   28      up
> > 30    hdd  7.27739   1.00000  7.3 TiB  1.8 TiB  1.8 TiB   1.6 GiB  
> > 14
> > GiB  5.5 TiB  25.01  1.56   35      up
> > 31    hdd  7.27739   1.00000  7.3 TiB  747 GiB  739 GiB   2.2 GiB 
> > 6.2
> > GiB  6.5 TiB  10.03  0.63   20      up
> > 32    hdd  7.27739   1.00000  7.3 TiB  996 GiB  987 GiB   1.5 GiB 
> > 7.9
> > GiB  6.3 TiB  13.37  0.83   26      up
> > 33    hdd  7.27739   1.00000  7.3 TiB  995 GiB  985 GiB   1.5 GiB 
> > 7.7
> > GiB  6.3 TiB  13.35  0.83   25      up
> > 34    hdd  7.27739   1.00000  7.3 TiB  750 GiB  742 GiB   2.1 GiB 
> > 5.7
> > GiB  6.5 TiB  10.07  0.63   25      up
> > 35    hdd  7.27739   1.00000  7.3 TiB  2.1 TiB  2.0 TiB   571 MiB  
> > 15
> > GiB  5.2 TiB  28.36  1.77   34      up
> > 36    hdd  7.27739   1.00000  7.3 TiB  1.3 TiB  1.3 TiB   1.5 GiB  
> > 10
> > GiB  5.9 TiB  18.37  1.15   31      up
> > 37    hdd  7.27739   1.00000  7.3 TiB  1.1 TiB  1.1 TiB   524 MiB 
> > 8.2
> > GiB  6.2 TiB  14.99  0.94   26      up
> > 38    hdd  7.27739   1.00000  7.3 TiB  1.6 TiB  1.6 TiB   1.6 GiB  
> > 12
> > GiB  5.7 TiB  21.70  1.35   28      up
> > 39    hdd  7.27739   1.00000  7.3 TiB  1.5 TiB  1.4 TiB   2.4 GiB  
> > 11
> > GiB  5.8 TiB  20.04  1.25   30      up
> > 10    hdd  7.27739   1.00000  7.3 TiB  1.3 TiB  1.3 TiB   1.6 GiB  
> > 10
> > GiB  5.9 TiB  18.34  1.14   26      up
> > 12    hdd  7.27739   1.00000  7.3 TiB  1.3 TiB  1.3 TiB     1 KiB 
> > 9.9
> > GiB  5.9 TiB  18.37  1.15   25      up
> > 14    hdd  7.27739   1.00000  7.3 TiB  1.5 TiB  1.4 TiB   593 MiB  
> > 10
> > GiB  5.8 TiB  19.98  1.25   22      up
> > 16    hdd  7.27739   1.00000  7.3 TiB  997 GiB  987 GiB   2.2 GiB 
> > 7.5
> > GiB  6.3 TiB  13.38  0.84   19      up
> > 18    hdd  7.27739   1.00000  7.3 TiB  1.1 TiB  1.1 TiB   1.1 GiB 
> > 8.7
> > GiB  6.2 TiB  15.02  0.94   26      up
> > 20    hdd  7.27739   1.00000  7.3 TiB  1.6 TiB  1.6 TiB   1.1 GiB  
> > 12
> > GiB  5.7 TiB  21.68  1.35   26      up
> > 22    hdd  7.27739   1.00000  7.3 TiB  1.3 TiB  1.3 TiB   559 MiB  
> > 10
> > GiB  5.9 TiB  18.34  1.14   22      up
> > 24    hdd  7.27739   1.00000  7.3 TiB  872 GiB  864 GiB  1020 MiB 
> > 6.8
> > GiB  6.4 TiB  11.70  0.73   23      up
> > 26    hdd  7.27739   1.00000  7.3 TiB  749 GiB  741 GiB   1.8 GiB 
> > 6.3
> > GiB  6.5 TiB  10.05  0.63   25      up
> > 28    hdd  7.27739   1.00000  7.3 TiB  1.3 TiB  1.3 TiB   1.5 GiB  
> > 10
> > GiB  5.9 TiB  18.36  1.15   32      up
> > 11    hdd  7.27739   1.00000  7.3 TiB  1.1 TiB  1.1 TiB   2.6 GiB 
> > 8.5
> > GiB  6.2 TiB  15.02  0.94   23      up
> > 13    hdd  7.27739   1.00000  7.3 TiB  1.3 TiB  1.3 TiB   2.2 GiB  
> > 10
> > GiB  5.9 TiB  18.38  1.15   36      up
> > 15    hdd  7.27739   1.00000  7.3 TiB  995 GiB  986 GiB   1.1 GiB 
> > 7.7
> > GiB  6.3 TiB  13.35  0.83   25      up
> > 17    hdd  7.27739   1.00000  7.3 TiB  623 GiB  618 GiB   419 KiB 
> > 5.0
> > GiB  6.7 TiB   8.35  0.52   23      up
> > 19    hdd  7.27739   1.00000  7.3 TiB  870 GiB  863 GiB   513 MiB 
> > 6.6
> > GiB  6.4 TiB  11.67  0.73   21      up
> > 21    hdd  7.27739   1.00000  7.3 TiB  1.1 TiB  1.1 TiB   1.5 GiB 
> > 8.6
> > GiB  6.2 TiB  15.02  0.94   25      up
> > 23    hdd  7.27739   1.00000  7.3 TiB  746 GiB  739 GiB   564 MiB 
> > 5.8
> > GiB  6.5 TiB  10.01  0.62   22      up
> > 25    hdd  7.27739   1.00000  7.3 TiB  1.1 TiB  1.1 TiB   2.1 GiB 
> > 8.4
> > GiB  6.2 TiB  15.03  0.94   24      up
> > 27    hdd  7.27739   1.00000  7.3 TiB  1.2 TiB  1.2 TiB   532 MiB 
> > 9.1
> > GiB  6.1 TiB  16.68  1.04   23      up
> > 29    hdd  7.27739   1.00000  7.3 TiB  1.1 TiB  1.1 TiB   1.1 GiB 
> > 8.4
> > GiB  6.2 TiB  14.99  0.94   19      up
> >                       TOTAL  291 TiB   47 TiB   46 TiB    51 GiB 
> > 359
> > GiB  244 TiB  16.02
> > MIN/MAX VAR: 0.52/1.77  STDDEV: 4.56
> > 
> > 
> > On Mon, 2023-07-31 at 09:22 +0000, Eugen Block wrote:
> > Hi,
> > 
> > can you share some more details like 'ceph df' and 'ceph osd df'? I
> > don't have too much advice yet, but to see all entries in your meta
> > pool you need add the --all flag because those objects are stored
> > in
> > namespaces:
> > 
> > rados -p default.rgw.meta ls --all
> > 
> > That pool contains user and bucket information (example):
> > 
> > # rados -p default.rgw.meta ls --all
> > users.uid       admin.buckets
> > users.keys      c0fba3ea7d9c4321b5205752c85baa85 users.uid      
> > admin
> > users.keys      JBWPRAPP1AQG471AMGC4 users.uid
> > e434b82737cf4138b899c0785b49112d.buckets
> > users.uid       e434b82737cf4138b899c0785b49112d
> > 
> > 
> > 
> > Zitat von Mark Johnson <markj@xxxxxxxxx<mailto:markj@xxxxxxxxx>>:
> > 
> > I've been going round and round in circles trying to work this one
> > out but I'm getting nowhere.  We're running a 4 node quincy cluster
> > (17.2.6) which recently reported the following:
> > 
> > ceph.log-20230729.gz:2023-07-28T08:31:42.390003+0000 osd.26
> > (osd.26)
> > 13834 : cluster [WRN] Large omap object found. Object:
> > 5:6c65dd84:users.uid::callrecordings$callrecordings_rw.buckets:head
> > 
> > PG: 5.21bba636 (5.16) Key count: 378454 Size (bytes): 75565579
> > 
> > This happened a week or so ago (only the key count was only just
> > over the 200000 threshold on that occasion) and after much
> > searching
> > around, I found an article that suggested a deep scrub on the pg
> > would likely resolve the issue, so I forced a deep scrub and
> > shortly
> > after, the warning cleared.  Came into the office today to discover
> > the above.  It's on the same PG as before which is in the
> > default.rgw.meta pool.  This time, after forcing a deep-scrub on
> > that PG, nothing changed.  I did it a second time just to be sure
> > but got the same result.
> > 
> > I keep finding a suse article that simply suggests increasing the
> > threshold to the previous default of 2,000,000, but other articles
> > I
> > read say it was lowered for a reason and that by the time it hits
> > that figure, it's too late so I don't want to just mask it.
> > Problem
> > is that I don't really understand it.   I found a thread here from
> > a
> > bit over two years ago but their issue was in the
> > default.rgw.buckets.index pool.  A step in the solution was to list
> > out the problematic object id and check the objects per shard
> > however, if I issue the command "rados -p default.rgw.meta ls" it
> > returns nothing.  I get a big list from "rados -p
> > default.rgw.buckets.index ls" just nothing from the first pool.  I
> > think it may be because the meta pool isn't indexed based on
> > something I read, but I really don't know what I'm talking about
> > tbh.
> > 
> > I don't know if this is helpful, but if I list out all the PGs for
> > that pool, there are 32 PGs and 5.16 shows 80186950 bytes and
> > 401505
> > keys.  PG 5.c has 75298 and 384 keys.  The remaining 30 PGs show
> > zero bytes and zero keys.  I'm really not sure how to troubleshoot
> > and resolve from here.  For the record, dynamic resharding is
> > enabled in that no options have been set in the config and that is
> > the default setting.
> > 
> > Based on the suse article I mentioned which also references the
> > default.rgw.meta pool, I'm gathering our issue is because we have
> > so
> > many buckets that are all owned by the one user and the solution is
> > either:
> > 
> > * delete unused buckets
> > * create multiple users and spread buckets evenly across all users
> > (not something we can do)
> > * increase the threshold to stop the warning
> > 
> > Problem is that I'm having trouble verifying this is the issue.
> > I've
> > tried dumping out bucket stats to a file (radosgw-admin bucket
> > stats
> > bucket_stats.txt) but after three hours this is still running with
> > no output.
> > 
> > Thanks for your time,
> > Mark
> > _______________________________________________
> > ceph-users mailing list --  
> > ceph-users@xxxxxxx<mailto:ceph-users@xxxxxxx> To unsubscribe send
> > an
> > email to ceph-users-leave@xxxxxxx<mailto:ceph-users-leave@xxxxxxx>
> > 
> > 
> > _______________________________________________
> > ceph-users mailing list --  
> > ceph-users@xxxxxxx<mailto:ceph-users@xxxxxxx> To unsubscribe send
> > an
> > email to ceph-users-leave@xxxxxxx<mailto:ceph-users-leave@xxxxxxx>
> > 
> > _______________________________________________
> > ceph-users mailing list --  
> > ceph-users@xxxxxxx<mailto:ceph-users@xxxxxxx> To unsubscribe send
> > an  
> > email to ceph-users-leave@xxxxxxx<mailto:ceph-users-leave@xxxxxxx>
> 
> 
> 

_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux