Re: Delete pool nicely

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Also, upgrade to luminous and migrate your OSDs to bluestore before using erasure coding.
Luminous + Bluestore performs so much better for erasure coding than any of the old configurations.

Also, I've found that deleting a large number of objects is far less stressfull on a Bluestore OSD than on a Filestore OSD.

Paul


2018-05-22 19:28 GMT+02:00 David Turner <drakonstein@xxxxxxxxx>:
From my experience, that would cause you some troubles as it would throw the entire pool into the deletion queue to be processed as it cleans up the disks and everything.  I would suggest using a pool listing from `rados -p .rgw.buckets ls` and iterate on that using some scripts around the `rados -p .rgw.buckest rm <obj-name>` command that you could stop, restart at a faster pace, slow down, etc.  Once the objects in the pool are gone, you can delete the empty pool without any problems.  I like this option because it makes it simple to stop it if you're impacting your VM traffic.


On Tue, May 22, 2018 at 11:05 AM Simon Ironside <sironside@xxxxxxxxxxxxx> wrote:
Hi Everyone,

I have an older cluster (Hammer 0.94.7) with a broken radosgw service
that I'd just like to blow away before upgrading to Jewel after which
I'll start again with EC pools.

I don't need the data but I'm worried that deleting the .rgw.buckets
pool will cause performance degradation for the production RBD pool used
by VMs. .rgw.buckets is a replicated pool (size=3) with ~14TB data in
5.3M objects. A little over half the data in the whole cluster.

Is deleting this pool simply using ceph osd pool delete likely to cause
me a performance problem? If so, is there a way I can do it better?

Thanks,
Simon.
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com




--
Paul Emmerich

Looking for help with your Ceph cluster? Contact us at https://croit.io

croit GmbH
Freseniusstr. 31h
81247 München
www.croit.io
Tel: +49 89 1896585 90
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux