Your error message indicates it failed to open a parent image of a clone: 2017-10-16 13:18:17.404858 7f35a37fe700 -1 librbd::image::RefreshParentRequest: failed to open parent image: (2) No such file or directory That means that the parent image has a snapshot that the clone is linked against. On Mon, Oct 16, 2017 at 4:10 PM, Nagy Ákos <nagy.akos@xxxxxxxxxxxxxx> wrote: > Thanks, > I don't have any snapshot or clone, only some standalone images. > > An follow the linked thread, I can change my id's, and I can restore all > the images. > > Thanks a lot! > > > 2017. 10. 16. 17:21 keltezéssel, Jason Dillaman írta: >> The pool ids can be updated to point to the correct pool [1] with >> enough patience. The larger issue is that the snapshots are not >> preserved and thus your cloned images can be corrupted if the parent >> image was modified after the creation of the protected snapshot. >> >> [1] http://lists.ceph.com/pipermail/ceph-users-ceph.com/2015-May/001398.html >> >> On Mon, Oct 16, 2017 at 8:11 AM, Wido den Hollander <wido@xxxxxxxx> wrote: >>>> Op 16 oktober 2017 om 13:00 schreef Nagy Ákos <nagy.akos@xxxxxxxxxxxxxx>: >>>> >>>> >>>> Thanks, >>>> >>>> but I erase all of the data, I have only this backup. >>> I hate to bring the bad news, but it will not work. The pools have different IDs and that will make it very difficult to get this working again. >>> >>> Wido >>> >>>> If the restore work for 3 pools, I can do it for the remainig 2? >>>> >>>> What can I try to set, to import it or how I can find this IDs? >>>> >>>> 2017. 10. 16. 13:39 keltezéssel, John Spray írta: >>>>> On Mon, Oct 16, 2017 at 11:35 AM, Nagy Ákos <nagy.akos@xxxxxxxxxxxxxx> wrote: >>>>>> Hi, >>>>>> >>>>>> I want to upgrade my ceph from jewel to luminous, and switch to bluestore. >>>>>> >>>>>> For that I export the pools from old cluster: >>>>> This is not the way to do it. You should convert your OSDs from >>>>> filestore to bluestore one by one, and let the data re-replicate to >>>>> the new OSDs. >>>>> >>>>> Dumping data out of one Ceph cluster and into another will not work, >>>>> because things like RBD images record things like the ID of the pool >>>>> where their parent image is, and pool IDs are usually different >>>>> between clusters. >>>>> >>>>> John >>>>> >>>>>> rados export -p pool1 pool1.ceph >>>>>> >>>>>> and after upgrade and osd recreation: >>>>>> >>>>>> rados --create -p pool1 import pool1.ceph >>>>>> >>>>>> I can import the backup without error, but when I want to map an image, I >>>>>> got error: >>>>>> >>>>>> rbd --image container1 --pool pool1 map >>>>>> >>>>>> rbd: sysfs write failed >>>>>> In some cases useful info is found in syslog - try "dmesg | tail". >>>>>> rbd: map failed: (2) No such file or directory >>>>>> >>>>>> dmesg | tail >>>>>> >>>>>> [160606.729840] rbd: image container1 : WARNING: kernel layering is >>>>>> EXPERIMENTAL! >>>>>> [160606.730675] libceph: tid 86731 pool does not exist >>>>>> >>>>>> >>>>>> When I try to get info about the image: >>>>>> >>>>>> rbd info pool1/container1 >>>>>> >>>>>> 2017-10-16 13:18:17.404858 7f35a37fe700 -1 >>>>>> librbd::image::RefreshParentRequest: failed to open parent image: (2) No >>>>>> such file or directory >>>>>> 2017-10-16 13:18:17.404903 7f35a37fe700 -1 librbd::image::RefreshRequest: >>>>>> failed to refresh parent image: (2) No such file or directory >>>>>> 2017-10-16 13:18:17.404930 7f35a37fe700 -1 librbd::image::OpenRequest: >>>>>> failed to refresh image: (2) No such file or directory >>>>>> rbd: error opening image container1: (2) No such file or directory >>>>>> >>>>>> >>>>>> I check to exported image checksum after export and before import, and it's >>>>>> match, and I can restore three pools with one with 60 MB one with 1.2 GB and >>>>>> one with 25 GB of data. >>>>>> >>>>>> The problematic has 60 GB data. >>>>>> >>>>>> The pool store LXD container images. >>>>>> >>>>>> Any help is highly appreciated. >>>>>> >>>>>> -- >>>>>> Ákos >>>>>> >>>>>> >>>>>> _______________________________________________ >>>>>> ceph-users mailing list >>>>>> ceph-users@xxxxxxxxxxxxxx >>>>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >>>>>> >>>> -- >>>> Ákos >>>> >>>> >>>> _______________________________________________ >>>> ceph-users mailing list >>>> ceph-users@xxxxxxxxxxxxxx >>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >>> _______________________________________________ >>> ceph-users mailing list >>> ceph-users@xxxxxxxxxxxxxx >>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >> >> > > -- > Ákos > > -- Jason _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com