Re: large bucket index in multisite environement (how to deal with large omap objects warning)?

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



When resharding is performed I believe its considered as bucket operation and undergoes through updating the bucket stats. Like new bucket shard is created and it may increase the number of objects within the bucket stats.  If it was broken during resharding, you could check the current bucket id from: radosgw-admin metadata get "bucket:BUCKET_NAME". That would hive an idea which bucket index objects to keep. Then you could remove corrupted bucket shards (not the ones with the bucket id from the previous command) .dir.corrupted_bucket_index.SHARD_NUM objects from bucket.index pool:rados -p bucket.index .dir.corrupted_bucket_index.SHARD_NUM Where SHARD_NUM is the shard number you want to delete. And then running "radosgw-admin bucket check --fix --bucket=BUCKET_NAME"   That should have resolved your issue with the number of objects. As for slow object deletion. Do you run your metadata pools for rgw on nvme drives ? Specifically bucket.index pool. The problem is that you have a lot of objects and probably not enough shards. Radosgw retrieves the list of objects from bucket.index and if I remember correct it retrieves them as ordered list which is very expensive operation. Hence handful of time might be spent just on getting the object list. We get 1000 objects per second deleted  inside our storage. I would not recommend using "--inconsistent-index" to avoid more consitency issues.Надіслано з пристрою Galaxy
-------- Оригінальне повідомлення --------Від: mhnx <morphinwithyou@xxxxxxxxx> Дата: 08.11.21  13:28  (GMT+02:00) Кому: Сергей Процун <prosergey07@xxxxxxxxx> Копія: "Szabo, Istvan (Agoda)" <Istvan.Szabo@xxxxxxxxx>, Boris Behrens <bb@xxxxxxxxx>, Ceph Users <ceph-users@xxxxxxx> Тема: Re:  Re: large bucket index in multisite environement (how to deal with large omap objects warning)? (There should not be any issues using rgw for other buckets while re-sharding.)If it is then disabling the bucket access will work right? Also sync should be disabled.Yes, after the manual reshard it should clear the leftovers but in my situation resharding failed and I got double entries for that bucket. I didn't push further, instead I divide the bucket to new buckets and reduce object count with a new bucket tree. Copied all of the objects with rclone and started bucket remove "radosgw-admin bucket rm --bucket=mybucket --bypass-gc --purge-objects --max-concurrent-ios=128" it has been very very long time "started at Sep08" and it is still working. There was 250M objects in that bucket and after the manual reshard faiI I got 500M object count when I check with bucket stats num_objects. Now I have;"size_kb": 10648067645, "num_objects": 132270190 Remove speed is 50-60 objects in a second. It's not because of the cluster speed. Cluster is fine. I have space so I let it go. When I see stable object count I will stop the remove process and start again with the " --inconsistent-index" parameter. I wonder is it safe to use the parameter with referenced objects? I want to learn how "--inconsistent-index" works and what it does. Сергей Процун <prosergey07@xxxxxxxxx>, 5 Kas 2021 Cum, 17:46 tarihinde şunu yazdı:There should not be any issues using rgw for other buckets while re-sharding.As for doubling number of objects after reshard is an interesting situation. After the manual reshard is done, there might be leftover from the old bucket index. A
s during reshard new .dir.new_bucket_index objects are created. They contain all data related to the objects which are stored in buckets.data pool. Just wondering if the issue with the doubled number of objects was related to old bucket index. If so its save to delete old bucket index. In the perfect world, it would be ideal to know the eventoal number of objects inside the bucket and set number of shards to the corresponding setting initially. In the real world when the client re-purpose the usage of the bucket, we have to deal with reshards.пт, 5 лист. 2021, 14:43 користувач mhnx <morphinwithyou@xxxxxxxxx> пише:I also use this method and I hate it.

Stopping all of the RGW clients is never an option! It shouldn't be.
Sharding is hell. I was have 250M objects in a bucket and reshard failed
after 2days and object count doubled somehow! 2 days of downtime is not an
option.

I wonder if I stop the write-read on a bucket and while resharding it is
there any problem of using RGW's with all other buckets?

Nowadays I advise splitting buckets as much as you can! That means changing
your apps directory tree but this design requires it.
You need to plan object count at least for 5 years and create ones.
Usually I use 101 shards which means 10.100.000 objects.
Also If I need to use versioning I use 2x101 or 3x101 because versions are
hard to predict. You need to predict how many versions you need and set a
lifecycle even before using the bucket!
The max shard that I use 1999. I'm not happy about it but sometimes you
gotta do what you need to do.
Fighting with customers is not an option, you can only advise changing
their apps folder tree but I've never seen someone accept the deal without
arguing.

My offers usually like this:
1- Core files bucket: no need to change or very limited changes. "calculate
the object count and multiply with 2"
2- Hot data bucket: There will be daily changes and versioning. "calculate
the object count and multiply with 3"
3- Cold data bucket[s]: There will be no daily changes. You should open new
buckets every Year or Month. This is good to keep it clean and steady. No
need for versioning and Multisite Will not suffer due to barely changes.
4- Temp files bucket[s]: This is so important. If you're crawling millions
of millions objects everyday and delete it at the end of the week or month
then you should definitely use a temp bucket.  No versioning, No multisite,
No index if it's possible.



Szabo, Istvan (Agoda) <Istvan.Szabo@xxxxxxxxx>, 5 Kas 2021 Cum, 12:30
tarihinde şunu yazdı:

> You mean prepare or reshard?
> Prepare:
> I collect as much information for the users before onboarding so I can
> prepare for their use case in the future and set things up.
>
> Preshard:
> After created the bucket:
> radosgw-admin bucket reshard --bucket=ex-bucket --num-shards=101
>
> Also when you shard the buckets, you need to use prime numbers.
>
> Istvan Szabo
> Senior Infrastructure Engineer
> ---------------------------------------------------
> Agoda Services Co., Ltd.
> e: istvan.szabo@xxxxxxxxx<mailto:istvan.szabo@xxxxxxxxx>
> ---------------------------------------------------
>
> From: Boris Behrens <bb@xxxxxxxxx>
> Sent: Friday, November 5, 2021 4:22 PM
> To: Szabo, Istvan (Agoda) <Istvan.Szabo@xxxxxxxxx>; ceph-users@xxxxxxx
> Subject: Re:  large bucket index in multisite environement
> (how to deal with large omap objects warning)?
>
> Email received from the internet. If in doubt, don't click any link nor
> open any attachment !
> ________________________________
> Cheers Istvan,
>
> how do you do this?
>
> Am Do., 4. Nov. 2021 um 19:45 Uhr schrieb Szabo, Istvan (Agoda) <
> Istvan.Szabo@xxxxxxxxx<mailto:Istvan.Szabo@xxxxxxxxx>>:
> This one you need to prepare, you beed to preshard the bucket which you
> know that will hold more than millions of objects.
>
> I have a bucket where we store 1.2 billions of objects with 24xxx shard.
> No omap issue.
> Istvan Szabo
> Senior Infrastructure Engineer
> ---------------------------------------------------
> Agoda Services Co., Ltd.
> e: istvan.szabo@xxxxxxxxx<mailto:istvan.szabo@xxxxxxxxx>
> ---------------------------------------------------
>
>
>
> --
> Die Selbsthilfegruppe "UTF-8-Probleme" trifft sich diesmal abweichend im
> groüen Saal.
> _______________________________________________
> ceph-users mailing list -- ceph-users@xxxxxxx
> To unsubscribe send an email to ceph-users-leave@xxxxxxx
>
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx


_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux