Re: rados export/import fail

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Mon, Oct 16, 2017 at 11:35 AM, Nagy Ákos <nagy.akos@xxxxxxxxxxxxxx> wrote:
> Hi,
>
> I want to upgrade my ceph from jewel to luminous, and switch to bluestore.
>
> For that I export the pools from old cluster:

This is not the way to do it.  You should convert your OSDs from
filestore to bluestore one by one, and let the data re-replicate to
the new OSDs.

Dumping data out of one Ceph cluster and into another will not work,
because things like RBD images record things like the ID of the pool
where their parent image is, and pool IDs are usually different
between clusters.

John

> rados export -p pool1 pool1.ceph
>
> and after upgrade and osd recreation:
>
> rados --create -p pool1 import pool1.ceph
>
> I can import the backup without error, but when I want  to map an image, I
> got error:
>
> rbd --image container1 --pool pool1 map
>
> rbd: sysfs write failed
> In some cases useful info is found in syslog - try "dmesg | tail".
> rbd: map failed: (2) No such file or directory
>
> dmesg | tail
>
> [160606.729840] rbd: image container1 : WARNING: kernel layering is
> EXPERIMENTAL!
> [160606.730675] libceph: tid 86731 pool does not exist
>
>
> When I try to get info about the image:
>
> rbd info pool1/container1
>
> 2017-10-16 13:18:17.404858 7f35a37fe700 -1
> librbd::image::RefreshParentRequest: failed to open parent image: (2) No
> such file or directory
> 2017-10-16 13:18:17.404903 7f35a37fe700 -1 librbd::image::RefreshRequest:
> failed to refresh parent image: (2) No such file or directory
> 2017-10-16 13:18:17.404930 7f35a37fe700 -1 librbd::image::OpenRequest:
> failed to refresh image: (2) No such file or directory
> rbd: error opening image container1: (2) No such file or directory
>
>
> I check to exported image checksum after export and before import, and it's
> match, and I can restore three pools with one with 60 MB one with 1.2 GB and
> one with 25 GB of data.
>
> The problematic has 60 GB data.
>
> The pool store LXD container images.
>
> Any help is highly appreciated.
>
> --
> Ákos
>
>
> _______________________________________________
> ceph-users mailing list
> ceph-users@xxxxxxxxxxxxxx
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux