Hi, Are you exporting rbd image while a VM is running upon it ? As far as I know, rbd export is not consistent You should not export an image, but only snapshots: - create a snapshot of the image - export the snapshot (rbd export pool/image@snap - | ..) - drop the snapshot Regards, On 3/10/20 8:31 PM, Matt Dunavant wrote: > Hello, > > I think I've been running into an rbd export/import bug and wanted to see if anybody else had any experience. > > We're using rbd images for VM drives both with and without custom stripe sizes. When we try to export/import the drive to another ceph cluster, the VM always comes up in a busted state it can't recover from. This happens both when doing this export/import through stdin/stdout and when using a middle machine as a temp space. I remember doing this a few times in previous versions without error, so I'm not sure if this is a regression or I'm doing something different. I'm still testing this to try and track down where the issue is but wanted to post this here to see if anybody else has any experience. > > Example command: rbd -c /etc/ceph/cluster1.conf export pool/testvm.boot - | rbd -c /etc/ceph/cluster2.conf import - pool/testvm.boot > > Current cluster is on 14.2.8 and using Ubuntu 18.04 w/ 5.3.0-40-generic. > > Let me know if I can provide any more details to help track this down. > > Thanks, > _______________________________________________ > ceph-users mailing list -- ceph-users@xxxxxxx > To unsubscribe send an email to ceph-users-leave@xxxxxxx > _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx