Thanks, but I erase all of the data, I have only this backup. If the restore work for 3 pools, I can do it for the remainig 2? What can I try to set, to import it or how I can find this IDs? 2017. 10. 16. 13:39 keltezéssel, John Spray írta: > On Mon, Oct 16, 2017 at 11:35 AM, Nagy Ákos <nagy.akos@xxxxxxxxxxxxxx> wrote: >> Hi, >> >> I want to upgrade my ceph from jewel to luminous, and switch to bluestore. >> >> For that I export the pools from old cluster: > This is not the way to do it. You should convert your OSDs from > filestore to bluestore one by one, and let the data re-replicate to > the new OSDs. > > Dumping data out of one Ceph cluster and into another will not work, > because things like RBD images record things like the ID of the pool > where their parent image is, and pool IDs are usually different > between clusters. > > John > >> rados export -p pool1 pool1.ceph >> >> and after upgrade and osd recreation: >> >> rados --create -p pool1 import pool1.ceph >> >> I can import the backup without error, but when I want to map an image, I >> got error: >> >> rbd --image container1 --pool pool1 map >> >> rbd: sysfs write failed >> In some cases useful info is found in syslog - try "dmesg | tail". >> rbd: map failed: (2) No such file or directory >> >> dmesg | tail >> >> [160606.729840] rbd: image container1 : WARNING: kernel layering is >> EXPERIMENTAL! >> [160606.730675] libceph: tid 86731 pool does not exist >> >> >> When I try to get info about the image: >> >> rbd info pool1/container1 >> >> 2017-10-16 13:18:17.404858 7f35a37fe700 -1 >> librbd::image::RefreshParentRequest: failed to open parent image: (2) No >> such file or directory >> 2017-10-16 13:18:17.404903 7f35a37fe700 -1 librbd::image::RefreshRequest: >> failed to refresh parent image: (2) No such file or directory >> 2017-10-16 13:18:17.404930 7f35a37fe700 -1 librbd::image::OpenRequest: >> failed to refresh image: (2) No such file or directory >> rbd: error opening image container1: (2) No such file or directory >> >> >> I check to exported image checksum after export and before import, and it's >> match, and I can restore three pools with one with 60 MB one with 1.2 GB and >> one with 25 GB of data. >> >> The problematic has 60 GB data. >> >> The pool store LXD container images. >> >> Any help is highly appreciated. >> >> -- >> Ákos >> >> >> _______________________________________________ >> ceph-users mailing list >> ceph-users@xxxxxxxxxxxxxx >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >> -- Ákos _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com