Hi, sorry forgot to add this. Ceph release: 18.2.4 (e7ad5345525c7aa95470c26863873b581076945d) reef (stable) The cluster consists out of 360 OSDs and the index pool has 128 PGs and auto-resharding is enabled. Cheers, Florian > On 11. Sep 2024, at 18:18, Anthony D'Atri <aad@xxxxxxxxxxxxxx> wrote: > > Which Ceph release? How many OSDs and how many PBs in the index pool? Auto-resharding enabled? > >> On Sep 11, 2024, at 10:24 AM, Florian Schwab <fschwab@xxxxxxxxxxxxxxxxxxx> wrote: >> >> Hi everyone, >> >> I hope maybe someone here has an idea what is happening here or can give some pointers how to debug it further. >> >> We currently have a bucket which has large omap objects. Following this guide (https://access.redhat.com/solutions/6450561) we are able to identify the bucket etc. >> >> $ ceph health detail >> HEALTH_WARN 25 large omap objects >> [WRN] LARGE_OMAP_OBJECTS: 25 large omap objects >> 25 large objects found in pool 'ceph-objectstore.rgw.buckets.index' >> Search the cluster log for 'Large omap object found' for more details. >> >> $ radosgw-admin metadata list --metadata-key bucket.instance | grep -i XXX >> “YYY:XXX”, >> >> $ radosgw-admin bilog list --bucket-id=“XXX" --bucket=“YYY" --max-entries=600000 | grep -c op_id >> 600000 >> >> So far so good! But when try to trim the bilog we get the following error: >> >> $ radosgw-admin bilog trim --bucket-id=“XXX" --bucket=“YYY" >> ERROR: trim_bi_log_entries(): (2) No such file or directory >> >> >> The bucket itself doesn’t show any issues - all S3 operations are working. >> >> >> Thanks for any input! >> >> >> Cheers, >> Florian >> _______________________________________________ >> ceph-users mailing list -- ceph-users@xxxxxxx >> To unsubscribe send an email to ceph-users-leave@xxxxxxx > _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx