RBD clone to change data pool

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hello Cephers,

I'm disappointed. I thought I'v had found a good way to migrate from one data pool to 
another, without too much downtime.

I use XFS on RBD, via KRBD to store backups (see another thread). XFS with reflink and crc 
(accelerate Veeam merges).
Also, I want to migrate from a EC k3m2 to EC k8m2 on 13 nodes / 130 osds.

I cannot use rbd migration without downtime, due to krbd. Even, I tried and it was really 
slow.

But I saw the layering capacity offered by cloning, and thought of this method :
- unmount filesystem / unmap rbd device
- take snapshot / protect it
- clone snapshot to new image, changing data pool
- rename old image, new image for the new image ti have the name of the original one
- map new image and mount filesystem

This is fast, because it's only COW.

Starting from here, I thought that all new write will be on the new data pool.
And waiting for the backup retentions, I can migrate without doing anything.
To finalize, when there will not be too much data to move, I would do a rbd flatten, and be 
able to delete the source image on old data pool.

But...

I have problems.
On one image, my backups constantly fail, or a retry is triggered. No clear message, but I/
O seems slow, perhaps timeouts.

On another image, I had XFS kernel errors (on metadata) and the filesystem shuts down 
during the night (and backups).

Jul 10 01:32:28 fidcl-mrs4-vbr-repo-02 kernel: [6108174.156744] XFS (rbd1): metadata I/O 
error in "xfs_buf_iodone_callback_error" at daddr 0x8a2d4a4f8 len 8 error 5
Jul 10 01:32:47 fidcl-mrs4-vbr-repo-02 kernel: [6108193.510924] XFS (rbd1): metadata I/O 
error in "xfs_buf_iodone_callback_error" at daddr 0x8a2d4a4f8 len 8 error 5
Jul 10 01:32:58 fidcl-mrs4-vbr-repo-02 kernel: [6108204.696929] XFS (rbd1): metadata I/O 
error in "xfs_buf_iodone_callback_error" at daddr 0x8a2d4a4f8 len 8 error 5
Jul 10 01:33:13 fidcl-mrs4-vbr-repo-02 kernel: [6108219.857228] XFS (rbd1): metadata I/O 
error in "xfs_buf_iodone_callback_error" at daddr 0x8a2d4a4f8 len 8 error 5

I  unmount it and try to remount without success. xfs_repairs tells me I had to mount it to 
replay journal, and if I cannot, ignore it and... loose data.

Last unattended thing. On one image that is still mounted, and seems to work, I tries to 
launch a flatten operation, to see how long it can last and if it manage to finish, if my 
backups are doing better on it.

But her are the output I have, thought it seems to continue...

Image flatten: 2% complete...2021-07-12T23:22:55.998+0200 7f511a7fc700 -1 
librbd::operation::FlattenRequest: 0x7f50fc000f20 should_complete: encountered error: 
(85) Interrupted system call should be restarted
Image flatten: 0% complete...2021-07-12T23:23:32.142+0200 7f5119ffb700 -1 
librbd::operation::FlattenRequest: 0x7f50fc0015d0 should_complete: encountered error: 
(85) Interrupted system call should be restarted
2021-07-12T23:23:47.382+0200 7f5119ffb700 -1 librbd::operation::FlattenRequest: 
0x7f50fc0015d0 should_complete: encountered error: (85) Interrupted system call should 
be restarted
Image flatten: 2% complete...2021-07-12T23:23:58.926+0200 7f5119ffb700 -1 
librbd::operation::FlattenRequest: 0x7f50fc0015d0 should_complete: encountered error: 
(85) Interrupted system call should be restarted
Image flatten: 0% complete...2021-07-12T23:24:01.318+0200 7f5119ffb700 -1 
librbd::operation::FlattenRequest: 0x7f50fc0015d0 should_complete: encountered error: 
(85) Interrupted system call should be restarted
2021-07-12T23:24:07.422+0200 7f5119ffb700 -1 librbd::operation::FlattenRequest: 
0x7f50fc0015d0 should_complete: encountered error: (85) Interrupted system call should 
be restarted

So, either I it bugs, or layering, cloning, flattening is not supposed to work in my context... 
Perhaps due to changing data pool ? Erasure Coding ?

I'm now stuck.
I have ~400To of data to move from an EC3+2 to EC8+2 pool, and I'm only seeing one 
solution : stopping my backups during the copy, that will last weeks...

(No, I can't stay on EC3+2, I'v sold to my management and my colleagues that we'll have 
near 1PB usable on that cluster).

Thanx for reading, if you're still there !
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux