Re: why rgw generates large quantities orphan objects?

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi,

I’m wondering if this problem will ever get fixed?

This multipart-orphan-problem has made it now multiple times to the list, the tickets are up to 6 years old … and nothing changes. 
It screws up per-user space accounting and uses up space for nothing.

I’d open another ticket with easy steps to reproduce, but can’t get an account on tracker.ceph.com (“pending administrator approval”)

Ciao, Uli

> On 13. 10 2022, at 17:44, Haas, Josh <jhaas@xxxxxxxxxx> wrote:
> 
> Hi Liang,
> 
> My guess would be this bug:
> 
> https://tracker.ceph.com/issues/44660
> https://www.spinics.net/lists/ceph-users/msg30151.html
> 
> 
> It's actually existed for at least 6 years:
> https://tracker.ceph.com/issues/16767
> 
> 
> Which occurs any time you reupload the same *part* in a single Multipart Upload multiple times. For example, if my Multipart upload consists of 3 parts, if I upload part #2 twice, then the first upload of part #2 becomes orphaned.
> 
> 
> If this was indeed the cause, you should have multiple "_multipart_" rados objects for the same part in "rados ls". For example, here's all the rados objects associated with a bugged bucket before I deleted it:
> 
> cc79b188-89d1-4f47-acb1-ab90513e9bc9.23325574.228__multipart_file.txt.4vkWzU4C5XLd2R6unFgbQ6aZM26vPuq8.1
> cc79b188-89d1-4f47-acb1-ab90513e9bc9.23325574.228__multipart_file.txt.2~4zogSe4Ep0xvSC8j6aX71x_96cOgvQN.1
> cc79b188-89d1-4f47-acb1-ab90513e9bc9.23325574.228__shadow_file.txt.4vkWzU4C5XLd2R6unFgbQ6aZM26vPuq8.1_1
> cc79b188-89d1-4f47-acb1-ab90513e9bc9.23325574.228__shadow_file.txt.2~4zogSe4Ep0xvSC8j6aX71x_96cOgvQN.1_1
> 
> 
> If we look at just these two:
> 
> cc79b188-89d1-4f47-acb1-ab90513e9bc9.23325574.228__multipart_file.txt.4vkWzU4C5XLd2R6unFgbQ6aZM26vPuq8.1
> cc79b188-89d1-4f47-acb1-ab90513e9bc9.23325574.228__multipart_file.txt.2~4zogSe4Ep0xvSC8j6aX71x_96cOgvQN.1
> 
> 
> They are in the format:
> 
> $BUCKETID__multipart_$S3KEY.$PARTUID.$PARTNUM
> 
> Because everything matches ($BUCKETID, $S3KEY, $PARTNUM) except for $PARTUID, this S3 object has been affected by the bug. If you find instances of rados keys that match on everything except $PARTUID, then this bug is probably the cause.
> 
> 
> Josh
> 
> ________________________________
> From: 郑亮 <zhengliang0901@xxxxxxxxx>
> Sent: Wednesday, October 12, 2022 1:34:31 AM
> To: ceph-users@xxxxxxx
> Subject:  why rgw generates large quantities orphan objects?
> 
> Hi all,
> Description of problem: [RGW] Buckets/objects deletion is causing large
> quantities orphan raods objects
> 
> The cluster was running a cosbench workload, then remove the partial data
> by deleting objects from the cosbench client, then we have deleted all the
> buckets with the help of `s3cmd rb --recursive --force` command that
> removed all the buckets, but that did not help in the space reclaimation.
> 
> ```
> [root@node01 /]# rgw-orphan-list
> 
> Available pools:
> 
>    device_health_metrics
> 
>    .rgw.root
> 
>    os-test.rgw.buckets.non-ec
> 
>    os-test.rgw.log
> 
>    os-test.rgw.control
> 
>    os-test.rgw.buckets.index
> 
>    os-test.rgw.meta
> 
>    os-test.rgw.buckets.data
> 
>    deeproute-replica-hdd-pool
> 
>    deeproute-replica-ssd-pool
> 
>    cephfs-metadata
> 
>    cephfs-replicated-pool
> 
>    .nfs
> 
> Which pool do you want to search for orphans (for multiple, use
> space-separated list)? os-test.rgw.buckets.data
> 
> Pool is "os-test.rgw.buckets.data".
> 
> Note: output files produced will be tagged with the current timestamp --
> 20221008062356.
> running 'rados ls' at Sat Oct  8 06:24:05 UTC 2022
> 
> running 'rados ls' on pool os-test.rgw.buckets.data.
> 
> 
> 
> running 'radosgw-admin bucket radoslist' at Sat Oct  8 06:43:21 UTC 2022
> computing delta at Sat Oct  8 06:47:17 UTC 2022
> 
> 39662551 potential orphans found out of a possible 39844453 (99%).
> The results can be found in './orphan-list-20221008062356.out'.
> 
>    Intermediate files are './rados-20221008062356.intermediate' and
> './radosgw-admin-20221008062356.intermediate'.
> ***
> 
> *** WARNING: This is EXPERIMENTAL code and the results should be used
> ***          only with CAUTION!
> ***
> Done at Sat Oct  8 06:48:07 UTC 2022.
> 
> [root@node01 /]# radosgw-admin gc list
> []
> 
> [root@node01 /]# cat orphan-list-20221008062356.out | wc -l
> 39662551
> 
> [root@node01 /]# rados df
> POOL_NAME                       USED   OBJECTS  CLONES     COPIES
> MISSING_ON_PRIMARY  UNFOUND  DEGRADED     RD_OPS       RD     WR_OPS
> WR  USED COMPR  UNDER COMPR
> .nfs                         4.3 MiB         4       0         12
>        0        0         0      77398   76 MiB        146   79
> KiB         0 B          0 B
> .rgw.root                    180 KiB        16       0         48
>        0        0         0      28749   28 MiB          0
> 0 B         0 B          0 B
> cephfs-metadata              932 MiB     14772       0      44316
>        0        0         0    1569690  3.8 GiB    1258651  3.4
> GiB         0 B          0 B
> cephfs-replicated-pool       738 GiB    300962       0     902886
>        0        0         0     794612  470 GiB     770689  245
> GiB         0 B          0 B
> deeproute-replica-hdd-pool  1016 GiB    104276       0     312828
>        0        0         0   18176216  298 GiB  441783780  6.7
> TiB         0 B          0 B
> deeproute-replica-ssd-pool    30 GiB      3691       0      11073
>        0        0         0    2466079  2.1 GiB    8416232  221
> GiB         0 B          0 B
> device_health_metrics         50 MiB       108       0        324
>        0        0         0       1836  1.8 MiB       1944   18
> MiB         0 B          0 B
> os-test.rgw.buckets.data     5.6 TiB  39844453       0  239066718
>        0        0         0  552896177  3.0 TiB  999441015   60
> TiB         0 B          0 B
> os-test.rgw.buckets.index    1.8 GiB        33       0         99
>        0        0         0  153600295  154 GiB  110916573   62
> GiB         0 B          0 B
> os-test.rgw.buckets.non-ec   2.1 MiB        45       0        135
>        0        0         0     574240  349 MiB     153725  139
> MiB         0 B          0 B
> os-test.rgw.control              0 B         8       0         24
>        0        0         0          0      0 B          0
> 0 B         0 B          0 B
> os-test.rgw.log              3.7 MiB       346       0       1038
>        0        0         0   83877803   80 GiB    6306730  7.6
> GiB         0 B          0 B
> os-test.rgw.meta             220 KiB        23       0         69
>        0        0         0     640854  506 MiB     108229   53
> MiB         0 B          0 B
> 
> total_objects    40268737
> total_used       7.8 TiB
> total_avail      1.1 PiB
> total_space      1.1 PiB
> ```
> ceph verison:
> ```
> [root@node01 /]# ceph versions
> {
>    "mon": {
>        "ceph version 16.2.10 (45fa1a083152e41a408d15505f594ec5f1b4fe17)
> pacific (stable)": 3
>    },
>    "mgr": {
>        "ceph version 16.2.10 (45fa1a083152e41a408d15505f594ec5f1b4fe17)
> pacific (stable)": 2
>    },
>    "osd": {
>        "ceph version 16.2.10 (45fa1a083152e41a408d15505f594ec5f1b4fe17)
> pacific (stable)": 108
>    },
>    "mds": {
>        "ceph version 16.2.10 (45fa1a083152e41a408d15505f594ec5f1b4fe17)
> pacific (stable)": 2
>    },
>    "rgw": {
>        "ceph version 16.2.10 (45fa1a083152e41a408d15505f594ec5f1b4fe17)
> pacific (stable)": 9
>    },
>    "overall": {
>        "ceph version 16.2.10 (45fa1a083152e41a408d15505f594ec5f1b4fe17)
> pacific (stable)": 124
>    }
> }
> ```
> 
> Thanks,
> Best regards
> Liang Zheng
> _______________________________________________
> ceph-users mailing list -- ceph-users@xxxxxxx
> To unsubscribe send an email to ceph-users-leave@xxxxxxx
> _______________________________________________
> ceph-users mailing list -- ceph-users@xxxxxxx
> To unsubscribe send an email to ceph-users-leave@xxxxxxx

_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux