Hi Ceph’s We got an issue that we’re still looking the cause after more than 60 hour searching a misconfiguration. After cheking a lot of documentation and Questions&Answer we find that bucket id and bucket marker are not the same. We compared all our other bucket and all got the same id and marker. Also found some bucket with the rgw.none section an another not. This bucket is unable to be listed in a fashionable time. Customer relaxed usage from 120TB to 93TB , from 7Million objects to 5.8M. We isolated a single petition in a RGW server and check some metric, just try to list this bucket generate 2-3Gbps traffic from RGW to OSD/MON’s. I asked at IRC if there’re any problem about index pool be in other root in the same site at crushmap and we think that shouldn’t be. Any idea or suggestion, however crazy, will be proven. Our relevant configuration that may help : CEPH DF: ceph df GLOBAL: SIZE AVAIL RAW USED %RAW USED 684 TiB 139 TiB 545 TiB 79.70 POOLS: NAME ID USED %USED MAX AVAIL OBJECTS volumes 21 3.3 TiB 63.90 1.9 TiB 831300 backups 22 0 B 0 1.9 TiB 0 images 23 1.8 TiB 49.33 1.9 TiB 237066 vms 24 3.4 TiB 64.85 1.9 TiB 811534 openstack-volumes-archive 25 30 TiB 47.92 32 TiB 7748864 .rgw.root 26 1.6 KiB 0 1.9 TiB 4 default.rgw.control 27 0 B 0 1.9 TiB 100 default.rgw.data.root 28 56 KiB 0 1.9 TiB 186 default.rgw.gc 29 0 B 0 1.9 TiB 32 default.rgw.log 30 0 B 0 1.9 TiB 175 default.rgw.users.uid 31 4.9 KiB 0 1.9 TiB 26 default.rgw.users.email 36 12 B 0 1.9 TiB 1 default.rgw.users.keys 37 243 B 0 1.9 TiB 14 default.rgw.buckets.index 38 0 B 0 1.9 TiB 1056 default.rgw.buckets.data 39 245 TiB 93.84 16 TiB 102131428 default.rgw.buckets.non-ec 40 0 B 0 1.9 TiB 23046 default.rgw.usage 43 0 B 0 1.9 TiB 6 CEPH OSD Distribution: ceph osd tree ID CLASS WEIGHT TYPE NAME STATUS REWEIGHT PRI-AFF -41 654.84045 root archive -37 130.96848 host CEPH-ARCH-R03-07 100 archive 10.91399 osd.100 up 1.00000 1.00000 101 archive 10.91399 osd.101 up 1.00000 1.00000 102 archive 10.91399 osd.102 up 1.00000 1.00000 103 archive 10.91399 osd.103 up 1.00000 1.00000 104 archive 10.91399 osd.104 up 1.00000 1.00000 105 archive 10.91399 osd.105 up 1.00000 1.00000 106 archive 10.91409 osd.106 up 1.00000 1.00000 107 archive 10.91409 osd.107 up 1.00000 1.00000 108 archive 10.91409 osd.108 up 1.00000 1.00000 109 archive 10.91409 osd.109 up 1.00000 1.00000 110 archive 10.91409 osd.110 up 1.00000 1.00000 111 archive 10.91409 osd.111 up 1.00000 1.00000 -23 130.96800 host CEPH005 4 archive 10.91399 osd.4 up 1.00000 1.00000 41 archive 10.91399 osd.41 up 1.00000 1.00000 74 archive 10.91399 osd.74 up 1.00000 1.00000 75 archive 10.91399 osd.75 up 1.00000 1.00000 81 archive 10.91399 osd.81 up 1.00000 1.00000 82 archive 10.91399 osd.82 up 1.00000 1.00000 83 archive 10.91399 osd.83 up 1.00000 1.00000 84 archive 10.91399 osd.84 up 1.00000 1.00000 85 archive 10.91399 osd.85 up 1.00000 1.00000 86 archive 10.91399 osd.86 up 1.00000 1.00000 87 archive 10.91399 osd.87 up 1.00000 1.00000 88 archive 10.91399 osd.88 up 1.00000 1.00000 -17 130.96800 host CEPH006 7 archive 10.91399 osd.7 up 1.00000 1.00000 8 archive 10.91399 osd.8 up 1.00000 1.00000 9 archive 10.91399 osd.9 up 1.00000 1.00000 10 archive 10.91399 osd.10 up 1.00000 1.00000 12 archive 10.91399 osd.12 up 1.00000 1.00000 13 archive 10.91399 osd.13 up 1.00000 1.00000 42 archive 10.91399 osd.42 up 1.00000 1.00000 43 archive 10.91399 osd.43 up 1.00000 1.00000 51 archive 10.91399 osd.51 up 1.00000 1.00000 53 archive 10.91399 osd.53 up 1.00000 1.00000 76 archive 10.91399 osd.76 up 1.00000 1.00000 80 archive 10.91399 osd.80 up 1.00000 1.00000 -26 130.96800 host CEPH007 14 archive 10.91399 osd.14 up 1.00000 1.00000 15 archive 10.91399 osd.15 up 1.00000 1.00000 16 archive 10.91399 osd.16 up 1.00000 1.00000 39 archive 10.91399 osd.39 up 1.00000 1.00000 40 archive 10.91399 osd.40 up 1.00000 1.00000 44 archive 10.91399 osd.44 up 1.00000 1.00000 48 archive 10.91399 osd.48 up 1.00000 1.00000 49 archive 10.91399 osd.49 up 1.00000 1.00000 52 archive 10.91399 osd.52 up 1.00000 1.00000 77 archive 10.91399 osd.77 up 1.00000 1.00000 89 archive 10.91399 osd.89 up 1.00000 1.00000 90 archive 10.91399 osd.90 up 1.00000 1.00000 -31 130.96800 host CEPH008 5 archive 10.91399 osd.5 up 1.00000 1.00000 6 archive 10.91399 osd.6 up 1.00000 1.00000 11 archive 10.91399 osd.11 up 1.00000 1.00000 45 archive 10.91399 osd.45 up 1.00000 1.00000 46 archive 10.91399 osd.46 up 1.00000 1.00000 47 archive 10.91399 osd.47 up 1.00000 1.00000 55 archive 10.91399 osd.55 up 1.00000 1.00000 70 archive 10.91399 osd.70 up 1.00000 1.00000 71 archive 10.91399 osd.71 up 1.00000 1.00000 78 archive 10.91399 osd.78 up 1.00000 1.00000 79 archive 10.91399 osd.79 up 1.00000 1.00000 91 archive 10.91399 osd.91 up 1.00000 1.00000 -1 27.91296 root default -30 6.98199 host CEPH-SSD-004 92 ssd 0.87299 osd.92 up 1.00000 1.00000 93 ssd 0.87299 osd.93 up 1.00000 1.00000 94 ssd 0.87299 osd.94 up 1.00000 1.00000 95 ssd 0.87299 osd.95 up 1.00000 1.00000 96 ssd 0.87299 osd.96 up 1.00000 1.00000 97 ssd 0.87299 osd.97 up 1.00000 1.00000 98 ssd 0.87299 osd.98 up 1.00000 1.00000 99 ssd 0.87299 osd.99 up 1.00000 1.00000 -3 6.97699 host CEPH001 1 ssd 0.43599 osd.1 up 1.00000 1.00000 17 ssd 0.43599 osd.17 up 1.00000 1.00000 18 ssd 0.43599 osd.18 up 1.00000 1.00000 19 ssd 0.43599 osd.19 up 1.00000 1.00000 20 ssd 0.43599 osd.20 up 1.00000 1.00000 21 ssd 0.43599 osd.21 up 1.00000 1.00000 22 ssd 0.43599 osd.22 up 1.00000 1.00000 23 ssd 0.43599 osd.23 up 1.00000 1.00000 37 ssd 0.87299 osd.37 up 1.00000 1.00000 54 ssd 0.87299 osd.54 up 1.00000 1.00000 56 ssd 0.43599 osd.56 up 1.00000 1.00000 60 ssd 0.43599 osd.60 up 1.00000 1.00000 61 ssd 0.43599 osd.61 up 1.00000 1.00000 62 ssd 0.43599 osd.62 up 1.00000 1.00000 -5 6.97699 host CEPH002 2 ssd 0.43599 osd.2 up 1.00000 1.00000 24 ssd 0 osd.24 up 1.00000 1.00000 25 ssd 0.43599 osd.25 up 1.00000 1.00000 26 ssd 0.43599 osd.26 up 1.00000 1.00000 27 ssd 0.43599 osd.27 up 1.00000 1.00000 28 ssd 0.43599 osd.28 up 1.00000 1.00000 29 ssd 0.43599 osd.29 up 1.00000 1.00000 30 ssd 0.43599 osd.30 up 1.00000 1.00000 38 ssd 0.87299 osd.38 up 1.00000 1.00000 57 ssd 0.43599 osd.57 up 1.00000 1.00000 63 ssd 0.43599 osd.63 up 1.00000 1.00000 64 ssd 0.43599 osd.64 up 1.00000 1.00000 65 ssd 0.43599 osd.65 up 1.00000 1.00000 66 ssd 0.43599 osd.66 up 1.00000 1.00000 72 ssd 0.87299 osd.72 up 1.00000 1.00000 -7 6.97699 host CEPH003 0 ssd 0.43599 osd.0 up 1.00000 1.00000 3 ssd 0.43599 osd.3 up 1.00000 1.00000 31 ssd 0 osd.31 up 1.00000 1.00000 32 ssd 0.43599 osd.32 up 1.00000 1.00000 33 ssd 0.43599 osd.33 up 1.00000 1.00000 34 ssd 0.43599 osd.34 up 1.00000 1.00000 35 ssd 0.43599 osd.35 up 1.00000 1.00000 36 ssd 0.43599 osd.36 up 1.00000 1.00000 50 ssd 0.87299 osd.50 up 1.00000 1.00000 58 ssd 0.43599 osd.58 up 1.00000 1.00000 59 ssd 0.43599 osd.59 up 1.00000 1.00000 67 ssd 0.43599 osd.67 up 1.00000 1.00000 68 ssd 0.43599 osd.68 up 1.00000 1.00000 69 ssd 0.43599 osd.69 up 1.00000 1.00000 73 ssd 0.87299 osd.73 up 1.00000 1.00000 CEPH CONF: [global] #Normal-Memory 1/5 debug rgw = 1 #Disable debug osd = 0 debug journal = 0 debug ms = 0 fsid = e1ee8086-7cce-43fd-a252-3d677af22428 mon_initial_members = CEPH001, CEPH002, CEPH003 mon_host = 172.16.2.10,172.16.2.11,172.16.2.12 auth_cluster_required = cephx auth_service_required = cephx auth_client_required = cephx osd pool default pg num = 1024 osd pool default pgp num = 1024 public network = 172.16.2.0/24 cluster network = 172.16.1.0/24 osd pool default size = 2 osd pool default min size = 1 rgw dynamic resharding = true [osd] osd mkfs type = xfs osd op threads = 12 osd disk threads = 12 osd recovery threads = 4 osd recovery op priority = 1 osd recovery max active = 2 osd recovery max single start = 1 osd max backfills = 4 osd backfill scan max = 16 osd backfill scan min = 4 osd client op priority = 63 osd memory target = 2147483648 osd scrub begin hour = 23 osd scrub end hour = 6 osd scrub load threshold = 1 #low load scrubbing osd scrub during recovery = false #scrub during recovery [mon] mon allow pool delete = true mon osd min down reporters = 3 [mon.a] host = CEPH001 public bind addr = 172.16.2.10 mon addr = 172.16.2.10:6789 mon allow pool delete = true [mon.b] host = CEPH002 public bind addr = 172.16.2.11 mon addr = 172.16.2.11:6789 mon allow pool delete = true [mon.c] host = CEPH003 public bind addr = 172.16.2.12 mon addr = 172.16.2.12:6789 mon allow pool delete = true [client.rgw.ceph-rgw01] host = ceph-rgw01 rgw dns name = eu-es-s3gateway.edh-services.com rgw frontends = "beast endpoint=172.16.2.6 port=8080" rgw resolve cname = false rgw thread pool size = 4096 rgw op thread timeout = 600 rgw num rados handles = 1 rgw num control oids = 8 rgw cache enabled = true rgw cache lru size = 10000 rgw enable usage log = true rgw usage log tick interval = 30 rgw usage log flush threshold = 1024 rgw usage max shards = 32 rgw usage max user shards = 1 rgw log http headers = "http_x_forwarded_for" [client.rgw.ceph-rgw03] host = ceph-rgw03 rgw dns name = XXXXXXXXXXXXXX rgw frontends = "beast endpoint=172.16.2.8 port=8080" rgw resolve cname = false rgw thread pool size = 4096 rgw op thread timeout = 600 rgw num rados handles = 1 rgw num control oids = 8 rgw cache enabled = true rgw cache lru size = 100000 rgw enable usage log = true rgw usage log tick interval = 30 rgw usage log flush threshold = 1024 rgw usage max shards = 32 rgw usage max user shards = 1 rgw log http headers = "http_x_forwarded_for" [OSD HOST CROP] BUCKET: radosgw-admin bucket stats --bucket=XXXX { "bucket": "XXXX", "zonegroup": "4d8c7c5f-ca40-4ee3-b5bb-b2cad90bd007", "placement_rule": "default-placement", "explicit_placement": { "data_pool": "default.rgw.buckets.data", "data_extra_pool": "default.rgw.buckets.non-ec", "index_pool": "default.rgw.buckets.index" }, "id": "48efb8c3-693c-4fe0-bbe4-fdc16f590a82.14987264.1", "marker": "48efb8c3-693c-4fe0-bbe4-fdc16f590a82.3856921.7", "index_type": "Normal", "owner": "XXXXX", "ver": "0#38096,1#37913,2#37954,3#37836,4#38134,5#38236,6#37911,7#37879,8#37804,9#38005,10#37584,11#38446,12#37957,13#37537,14#37775,15#37794,16#38399,17#38162,18#37834,19#37633,20#37403,21#38173,22#37651,23#37303,24#37992,25#38228,26#38441,27#37975,28#38095,29#37835,30#38264,31#37958,32#37666,33#37517,34#38260,35#38168,36#37689,37#37551,38#37700,39#38056,40#38175,41#37765,42#37721,43#38472,44#37928,45#38451,46#37491,47#37875,48#38405,49#38011,50#38025,51#37983,52#37940,53#38306,54#37908,55#38181,56#37721,57#38366,58#37834,59#38392,60#37928,61#38235,62#37837,63#37940,64#38294,65#37610,66#37974,67#38304,68#37725,69#37301,70#37155,71#37681,72#37463,73#37603,74#37323,75#37717,76#37111,77#37688,78#37473,79#37052,80#37413,81#37758,82#36971,83#37327,84#37056,85#37302,86#37492,87#37250,88#37708,89#36891,90#38019,91#37485,92#37335,93#37712,94#37754,95#37117,96#37085,97#37694,98#37386,99#37384,100#37668,101#37329,102#37177,103#37494,104#37296,105#37366,106#37853,107#37234,108#36945,109#37040,110#37389,111#37973,112#37092,113#37327,114#37505,115#37545,116#37884,117#37325,118#37401,119#37208,120#37277,121#37087,122#37664,123#37649,124#37517,125#37618,126#37145,127#37300", "master_ver": "0#0,1#0,2#0,3#0,4#0,5#0,6#0,7#0,8#0,9#0,10#0,11#0,12#0,13#0,14#0,15#0,16#0,17#0,18#0,19#0,20#0,21#0,22#0,23#0,24#0,25#0,26#0,27#0,28#0,29#0,30#0,31#0,32#0,33#0,34#0,35#0,36#0,37#0,38#0,39#0,40#0,41#0,42#0,43#0,44#0,45#0,46#0,47#0,48#0,49#0,50#0,51#0,52#0,53#0,54#0,55#0,56#0,57#0,58#0,59#0,60#0,61#0,62#0,63#0,64#0,65#0,66#0,67#0,68#0,69#0,70#0,71#0,72#0,73#0,74#0,75#0,76#0,77#0,78#0,79#0,80#0,81#0,82#0,83#0,84#0,85#0,86#0,87#0,88#0,89#0,90#0,91#0,92#0,93#0,94#0,95#0,96#0,97#0,98#0,99#0,100#0,101#0,102#0,103#0,104#0,105#0,106#0,107#0,108#0,109#0,110#0,111#0,112#0,113#0,114#0,115#0,116#0,117#0,118#0,119#0,120#0,121#0,122#0,123#0,124#0,125#0,126#0,127#0", "mtime": "2019-04-30 15:14:20.152747", "max_marker": "0#,1#,2#,3#,4#,5#,6#,7#,8#,9#,10#,11#,12#,13#,14#,15#,16#,17#,18#,19#,20#,21#,22#,23#,24#,25#,26#,27#,28#,29#,30#,31#,32#,33#,34#,35#,36#,37#,38#,39#,40#,41#,42#,43#,44#,45#,46#,47#,48#,49#,50#,51#,52#,53#,54#,55#,56#,57#,58#,59#,60#,61#,62#,63#,64#,65#,66#,67#,68#,69#,70#,71#,72#,73#,74#,75#,76#,77#,78#,79#,80#,81#,82#,83#,84#,85#,86#,87#,88#,89#,90#,91#,92#,93#,94#,95#,96#,97#,98#,99#,100#,101#,102#,103#,104#,105#,106#,107#,108#,109#,110#,111#,112#,113#,114#,115#,116#,117#,118#,119#,120#,121#,122#,123#,124#,125#,126#,127#", "usage": { "rgw.none": { "size": 0, "size_actual": 0, "size_utilized": 0, "size_kb": 0, "size_kb_actual": 0, "size_kb_utilized": 0, "num_objects": 0 }, "rgw.main": { "size": 95368837327307, "size_actual": 95379421843456, "size_utilized": 95370098064318, "size_kb": 93133630203, "size_kb_actual": 93143966644, "size_kb_utilized": 93134861391, "num_objects": 5872260 }, "rgw.multimeta": { "size": 0, "size_actual": 0, "size_utilized": 0, "size_kb": 0, "size_kb_actual": 0, "size_kb_utilized": 0, "num_objects": 467 } }, "bucket_quota": { "enabled": false, "check_on_raw": false, "max_size": -1024, "max_size_kb": 0, "max_objects": -1 } } |
_______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com