Re: rados export/import fail

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Thanks,
I don't have any snapshot or clone, only some standalone images.

An follow the linked thread, I can change my id's, and I can restore all
the images.

Thanks a lot!


2017. 10. 16. 17:21 keltezéssel, Jason Dillaman írta:
> The pool ids can be updated to point to the correct pool [1] with
> enough patience. The larger issue is that the snapshots are not
> preserved and thus your cloned images can be corrupted if the parent
> image was modified after the creation of the protected snapshot.
>
> [1] http://lists.ceph.com/pipermail/ceph-users-ceph.com/2015-May/001398.html
>
> On Mon, Oct 16, 2017 at 8:11 AM, Wido den Hollander <wido@xxxxxxxx> wrote:
>>> Op 16 oktober 2017 om 13:00 schreef Nagy Ákos <nagy.akos@xxxxxxxxxxxxxx>:
>>>
>>>
>>> Thanks,
>>>
>>> but I erase all of the data, I have only this backup.
>> I hate to bring the bad news, but it will not work. The pools have different IDs and that will make it very difficult to get this working again.
>>
>> Wido
>>
>>> If the restore work for 3 pools, I can do it for the remainig 2?
>>>
>>> What can I try to set, to import it or how I can find this IDs?
>>>
>>> 2017. 10. 16. 13:39 keltezéssel, John Spray írta:
>>>> On Mon, Oct 16, 2017 at 11:35 AM, Nagy Ákos <nagy.akos@xxxxxxxxxxxxxx> wrote:
>>>>> Hi,
>>>>>
>>>>> I want to upgrade my ceph from jewel to luminous, and switch to bluestore.
>>>>>
>>>>> For that I export the pools from old cluster:
>>>> This is not the way to do it.  You should convert your OSDs from
>>>> filestore to bluestore one by one, and let the data re-replicate to
>>>> the new OSDs.
>>>>
>>>> Dumping data out of one Ceph cluster and into another will not work,
>>>> because things like RBD images record things like the ID of the pool
>>>> where their parent image is, and pool IDs are usually different
>>>> between clusters.
>>>>
>>>> John
>>>>
>>>>> rados export -p pool1 pool1.ceph
>>>>>
>>>>> and after upgrade and osd recreation:
>>>>>
>>>>> rados --create -p pool1 import pool1.ceph
>>>>>
>>>>> I can import the backup without error, but when I want  to map an image, I
>>>>> got error:
>>>>>
>>>>> rbd --image container1 --pool pool1 map
>>>>>
>>>>> rbd: sysfs write failed
>>>>> In some cases useful info is found in syslog - try "dmesg | tail".
>>>>> rbd: map failed: (2) No such file or directory
>>>>>
>>>>> dmesg | tail
>>>>>
>>>>> [160606.729840] rbd: image container1 : WARNING: kernel layering is
>>>>> EXPERIMENTAL!
>>>>> [160606.730675] libceph: tid 86731 pool does not exist
>>>>>
>>>>>
>>>>> When I try to get info about the image:
>>>>>
>>>>> rbd info pool1/container1
>>>>>
>>>>> 2017-10-16 13:18:17.404858 7f35a37fe700 -1
>>>>> librbd::image::RefreshParentRequest: failed to open parent image: (2) No
>>>>> such file or directory
>>>>> 2017-10-16 13:18:17.404903 7f35a37fe700 -1 librbd::image::RefreshRequest:
>>>>> failed to refresh parent image: (2) No such file or directory
>>>>> 2017-10-16 13:18:17.404930 7f35a37fe700 -1 librbd::image::OpenRequest:
>>>>> failed to refresh image: (2) No such file or directory
>>>>> rbd: error opening image container1: (2) No such file or directory
>>>>>
>>>>>
>>>>> I check to exported image checksum after export and before import, and it's
>>>>> match, and I can restore three pools with one with 60 MB one with 1.2 GB and
>>>>> one with 25 GB of data.
>>>>>
>>>>> The problematic has 60 GB data.
>>>>>
>>>>> The pool store LXD container images.
>>>>>
>>>>> Any help is highly appreciated.
>>>>>
>>>>> --
>>>>> Ákos
>>>>>
>>>>>
>>>>> _______________________________________________
>>>>> ceph-users mailing list
>>>>> ceph-users@xxxxxxxxxxxxxx
>>>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>>>>
>>> --
>>> Ákos
>>>
>>>
>>> _______________________________________________
>>> ceph-users mailing list
>>> ceph-users@xxxxxxxxxxxxxx
>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>> _______________________________________________
>> ceph-users mailing list
>> ceph-users@xxxxxxxxxxxxxx
>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
>

-- 
Ákos


_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux