Thanks.
We have that set
rgw gc max objs = 997
The problem we have is we have Commvault connected to the cluster. What
commvault does is every sunday does a full backup of an entire network
(around 40TB) then ages out the old data, which is basically deleting
all the old objects. Then all weel commvault makes incrimental backups
and then at the end of the week does the same full backup and then purge
of old data. Simple and standard backup.
The problem is, that over the space of a week, 80TB gets put into the
cloud and should also be purged deleted. We are not seeing the delete
complete in an entire week, so we are basically slowly running out of
space as the data is never cleaned up in time.
Any thoughts anyone?
I might try a larger prime number in like the 7919 range...
On 2014-11-15 20:42, Jean-Charles LOPEZ wrote:
Hi,
this is an old thing I remember of and it may not be exactly related
but just in case, let’s try it out.
Just verify the value of the following parameter rgw_gc_max_objs in
the RGW configuration via the admin socket (cep daemon {row.id} config
get rgw_gc_max_objs
.
If the value is 32, update the ceph.conf file for your RGW and set it
to 37 or any higher prime number such as 997 and restart your RGW.
Make sure you do the restart before your next big batch of object
removal. It will take some time to clear the overloaded buckets, may
be, but it should at least distribute the new ones evenly onto all gc
buckets after this and make the removal of the newly deleted objects
way quicker.
Keep us posted to tell us if it has improved anything.
JC
On Nov 14, 2014, at 01:20, Daniel Hoffman <daniel@xxxxxxxxxx> wrote:
Hi All.
Running a Ceph Cluster (firefly) ceph version 0.80.5
We use ceph mainly for backups via the radosGW at the moment.
There had to be an account deleted/bucket removed which had a very
large number of objects and was about 60TB in space.
We have been monitoring it for days now, and the data is purging but
very very slowly. We are actually putting new backups in much faster
than the old data is being removed.
As we are doing a lot of work with backups and aging out data we need
to find a way to improve the cleanup process.
Is there a way to improve the purge/clean performance?
Is the clean/purge performance impacted by disk thread ioprio class
setting?
Any advice or tunables to improve the removal of data is appreciated.
Thanks
Daniel
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com