On Mon, Oct 16, 2017 at 11:35 AM, Nagy Ákos <nagy.akos@xxxxxxxxxxxxxx> wrote: > Hi, > > I want to upgrade my ceph from jewel to luminous, and switch to bluestore. > > For that I export the pools from old cluster: This is not the way to do it. You should convert your OSDs from filestore to bluestore one by one, and let the data re-replicate to the new OSDs. Dumping data out of one Ceph cluster and into another will not work, because things like RBD images record things like the ID of the pool where their parent image is, and pool IDs are usually different between clusters. John > rados export -p pool1 pool1.ceph > > and after upgrade and osd recreation: > > rados --create -p pool1 import pool1.ceph > > I can import the backup without error, but when I want to map an image, I > got error: > > rbd --image container1 --pool pool1 map > > rbd: sysfs write failed > In some cases useful info is found in syslog - try "dmesg | tail". > rbd: map failed: (2) No such file or directory > > dmesg | tail > > [160606.729840] rbd: image container1 : WARNING: kernel layering is > EXPERIMENTAL! > [160606.730675] libceph: tid 86731 pool does not exist > > > When I try to get info about the image: > > rbd info pool1/container1 > > 2017-10-16 13:18:17.404858 7f35a37fe700 -1 > librbd::image::RefreshParentRequest: failed to open parent image: (2) No > such file or directory > 2017-10-16 13:18:17.404903 7f35a37fe700 -1 librbd::image::RefreshRequest: > failed to refresh parent image: (2) No such file or directory > 2017-10-16 13:18:17.404930 7f35a37fe700 -1 librbd::image::OpenRequest: > failed to refresh image: (2) No such file or directory > rbd: error opening image container1: (2) No such file or directory > > > I check to exported image checksum after export and before import, and it's > match, and I can restore three pools with one with 60 MB one with 1.2 GB and > one with 25 GB of data. > > The problematic has 60 GB data. > > The pool store LXD container images. > > Any help is highly appreciated. > > -- > Ákos > > > _______________________________________________ > ceph-users mailing list > ceph-users@xxxxxxxxxxxxxx > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com