Thanks, I don't have any snapshot or clone, only some standalone images. An follow the linked thread, I can change my id's, and I can restore all the images. Thanks a lot! 2017. 10. 16. 17:21 keltezéssel, Jason Dillaman írta: > The pool ids can be updated to point to the correct pool [1] with > enough patience. The larger issue is that the snapshots are not > preserved and thus your cloned images can be corrupted if the parent > image was modified after the creation of the protected snapshot. > > [1] http://lists.ceph.com/pipermail/ceph-users-ceph.com/2015-May/001398.html > > On Mon, Oct 16, 2017 at 8:11 AM, Wido den Hollander <wido@xxxxxxxx> wrote: >>> Op 16 oktober 2017 om 13:00 schreef Nagy Ákos <nagy.akos@xxxxxxxxxxxxxx>: >>> >>> >>> Thanks, >>> >>> but I erase all of the data, I have only this backup. >> I hate to bring the bad news, but it will not work. The pools have different IDs and that will make it very difficult to get this working again. >> >> Wido >> >>> If the restore work for 3 pools, I can do it for the remainig 2? >>> >>> What can I try to set, to import it or how I can find this IDs? >>> >>> 2017. 10. 16. 13:39 keltezéssel, John Spray írta: >>>> On Mon, Oct 16, 2017 at 11:35 AM, Nagy Ákos <nagy.akos@xxxxxxxxxxxxxx> wrote: >>>>> Hi, >>>>> >>>>> I want to upgrade my ceph from jewel to luminous, and switch to bluestore. >>>>> >>>>> For that I export the pools from old cluster: >>>> This is not the way to do it. You should convert your OSDs from >>>> filestore to bluestore one by one, and let the data re-replicate to >>>> the new OSDs. >>>> >>>> Dumping data out of one Ceph cluster and into another will not work, >>>> because things like RBD images record things like the ID of the pool >>>> where their parent image is, and pool IDs are usually different >>>> between clusters. >>>> >>>> John >>>> >>>>> rados export -p pool1 pool1.ceph >>>>> >>>>> and after upgrade and osd recreation: >>>>> >>>>> rados --create -p pool1 import pool1.ceph >>>>> >>>>> I can import the backup without error, but when I want to map an image, I >>>>> got error: >>>>> >>>>> rbd --image container1 --pool pool1 map >>>>> >>>>> rbd: sysfs write failed >>>>> In some cases useful info is found in syslog - try "dmesg | tail". >>>>> rbd: map failed: (2) No such file or directory >>>>> >>>>> dmesg | tail >>>>> >>>>> [160606.729840] rbd: image container1 : WARNING: kernel layering is >>>>> EXPERIMENTAL! >>>>> [160606.730675] libceph: tid 86731 pool does not exist >>>>> >>>>> >>>>> When I try to get info about the image: >>>>> >>>>> rbd info pool1/container1 >>>>> >>>>> 2017-10-16 13:18:17.404858 7f35a37fe700 -1 >>>>> librbd::image::RefreshParentRequest: failed to open parent image: (2) No >>>>> such file or directory >>>>> 2017-10-16 13:18:17.404903 7f35a37fe700 -1 librbd::image::RefreshRequest: >>>>> failed to refresh parent image: (2) No such file or directory >>>>> 2017-10-16 13:18:17.404930 7f35a37fe700 -1 librbd::image::OpenRequest: >>>>> failed to refresh image: (2) No such file or directory >>>>> rbd: error opening image container1: (2) No such file or directory >>>>> >>>>> >>>>> I check to exported image checksum after export and before import, and it's >>>>> match, and I can restore three pools with one with 60 MB one with 1.2 GB and >>>>> one with 25 GB of data. >>>>> >>>>> The problematic has 60 GB data. >>>>> >>>>> The pool store LXD container images. >>>>> >>>>> Any help is highly appreciated. >>>>> >>>>> -- >>>>> Ákos >>>>> >>>>> >>>>> _______________________________________________ >>>>> ceph-users mailing list >>>>> ceph-users@xxxxxxxxxxxxxx >>>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >>>>> >>> -- >>> Ákos >>> >>> >>> _______________________________________________ >>> ceph-users mailing list >>> ceph-users@xxxxxxxxxxxxxx >>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >> _______________________________________________ >> ceph-users mailing list >> ceph-users@xxxxxxxxxxxxxx >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > > -- Ákos _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com