Re: Migrating to new pools

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi *,

Zitat von ceph-users-request@xxxxxxxxxxxxxx:
Hi *,

facing the problem to reduce the number of PGs for a pool, I've found
various information and suggestions, but no "definite guide" to handle
pool migration with Ceph 12.2.x. This seems to be a fairly common
problem when having to deal with "teen-age clusters", so consolidated
information would be a real help. I'm willing to start writing things
up, but don't want to duplicate information. So:

Are there any documented "operational procedures" on how to migrate

- an RBD pool (with snapshots created by Openstack)

- a CephFS data pool

- a CephFS metadata pool

to a different volume, in order to be able to utilize pool settings
that cannot be changed on an existing pool?

---

RBD pools: From what I've read, RBD snapshots are "broken" after using
"rados cppool" to move the content of an "RBD pool" to a new pool.

after having read up on "rbd-mirror" and having had a glimpse at its code, it seems that preserving clones, snapshots and their relationships has been solved for cluster-to-cluster migration.

Is this really correct? If so, it might be possible to extend the code in a fashion that will allow a one-shot, intracluster pool-to-pool migration as a spin-off to rbd-mirror.

I was thinking along the following lines:

- run rbd-mirror in a stand-alone fashion, just stating source and destination pool

- leave it to the cluster admin to take RDB "offline", so that the pool content does not change during the copy (no RBD journaling involved)

- check that the destination pool is empty. In a first version, cumulative migrates (joining multiple source pools with distinct image names) would complicate things ;)

- sync all content from source to destination pool, in a one-shot fashion

- done

Anyone out there who can judge the chances of that approach, better than me? I'd be willing to spend development time on this, but starting from scratch will be rather hard, so pointers at where to look at within the rbd-mirror code would be more than welcome.

Regards,
Jens

_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux