Re: rgw index shard much larger than others

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi Dan,

One way to tell would be to do a:

	radosgw-admin bi list —bucket=<bucket>

And see if any of the lines output contains (perhaps using `grep`):

	"type": "olh",

That would tell you if there were any versioned objects in the bucket.

The “fix” we currently have only prevents this from happening in the future. We currently do not have a “fix” that cleans up the bucket index. Like I mentioned — an automated clean-up is non-trivial but feasible; it would have to take into account that an object with the same name as the previously deleted one was re-created in the versioned bucket.

I hope that’s informative, if not what you were hoping to hear.

Eric
--
J. Eric Ivancich
he / him / his
Red Hat Storage
Ann Arbor, Michigan, USA

> On Oct 1, 2020, at 10:53 AM, Dan van der Ster <dan@xxxxxxxxxxxxxx> wrote:
> 
> Thanks Matt and Eric,
> 
> Sorry for the basic question, but how can I as a ceph operator tell if
> a bucket is versioned?
> 
> And for fixing this current situation, I would wait for the fix then reshard?
> (We want to reshard this bucket anyway because listing perf is way too
> slow for the user with 512 shards).
> 
> -- Dan
> 
> 
> On Thu, Oct 1, 2020 at 4:36 PM Eric Ivancich <ivancich@xxxxxxxxxx> wrote:
>> 
>> Hi Matt and Dan,
>> 
>> I too suspect it’s the issue Matt linked to. That bug only affects versioned buckets, so I’m guessing your bucket is versioned, Dan.
>> 
>> This bug is triggered when the final instance of an object in a versioned bucket is deleted, but for reasons we do not yet understand, the object was not fully deleted from the bucket index. And then a reshard moves part of the object index to shard 0.
>> 
>> Upgrading to a version that included Casey’s fix would mean this situation is not re-created in the future.
>> 
>> An automated clean-up is non-trivial but feasible. It would have to take into account that an object with the same name as the previously deleted one was re-created in the versioned bucket.
>> 
>> Eric
>> 
>>> On Oct 1, 2020, at 8:46 AM, Matt Benjamin <mbenjami@xxxxxxxxxx> wrote:
>>> 
>>> Hi Dan,
>>> 
>>> Possibly you're reproducing https://tracker.ceph.com/issues/46456.
>>> 
>>> That explains how the underlying issue worked, I don't remember how a
>>> bucked exhibiting this is repaired.
>>> 
>>> Eric?
>>> 
>>> Matt
>>> 
>>> 
>>> On Thu, Oct 1, 2020 at 8:41 AM Dan van der Ster <dan@xxxxxxxxxxxxxx> wrote:
>>>> 
>>>> Dear friends,
>>>> 
>>>> Running 14.2.11, we have one particularly large bucket with a very
>>>> strange distribution of objects among the shards. The bucket has 512
>>>> shards, and most shards have ~75k entries, but shard 0 has 1.75M
>>>> entries:
>>>> 
>>>> # rados -p default.rgw.buckets.index listomapkeys
>>>> .dir.61c59385-085d-4caa-9070-63a3868dccb6.272652427.1.0 | wc -l
>>>> 1752085
>>>> 
>>>> # rados -p default.rgw.buckets.index listomapkeys
>>>> .dir.61c59385-085d-4caa-9070-63a3868dccb6.272652427.1.1 | wc -l
>>>> 78388
>>>> 
>>>> # rados -p default.rgw.buckets.index listomapkeys
>>>> .dir.61c59385-085d-4caa-9070-63a3868dccb6.272652427.1.2 | wc -l
>>>> 78764
>>>> 
>>>> We had resharded this bucket (manually) from 32 up to 512 shards just
>>>> before upgrading from 12.2.12 to 14.2.11 a couple weeks ago.
>>>> 
>>>> Any idea why shard .0 is getting such an imbalance of entries?
>>>> Should we manually reshard this bucket again?
>>>> 
>>>> Thanks!
>>>> 
>>>> Dan
>>>> _______________________________________________
>>>> ceph-users mailing list -- ceph-users@xxxxxxx
>>>> To unsubscribe send an email to ceph-users-leave@xxxxxxx
>>>> 
>>>> 
>>> 
>>> 
>>> --
>>> 
>>> Matt Benjamin
>>> Red Hat, Inc.
>>> 315 West Huron Street, Suite 140A
>>> Ann Arbor, Michigan 48103
>>> 
>>> http://www.redhat.com/en/technologies/storage
>>> 
>>> tel.  734-821-5101
>>> fax.  734-769-8938
>>> cel.  734-216-5309
>>> 
>> 
> 

_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux