"rbd snap rm" overload my cluster (during backups)

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi,

I have a backup script, which every night :
* create a snapshot of each RBD image
* then delete all snapshot that have more than 15 days

The problem is that "rbd snap rm XXX" will overload my cluster for hours
(6 hours today...).

Here I see several problems :
#1 "rbd snap rm XXX" is not blocking. The erase is done in background,
and I know no way to verify if it was completed. So I add "sleeps"
between rm, but I have to estimate the time it will take

#2 "rbd (snap) rm" are sometimes very very slow. I don't know if it's
because of XFS or not, but all my OSD are at 100% IO usage (reported by
iostat)



So :
* is there a way to reduce priority of "snap rm", to avoid overloading
of the cluster ?
* is there a way to have a blocking "snap rm" which will wait until it's
completed
* is there a way to speedup "snap rm" ?


Note that I have a too low PG number on my cluster (200 PG for 40 active
OSD ; but I'm trying to progressivly migrate data to a newer pool). Can
it be the source of the problem ?

Thanks,

Olivier

_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux