Re: How to recover/mount mirrored rbd image for file recovery

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



okay, so i have ceph version 14.2.6 nautilus on my source cluster and
ceph version 12.2.13 luminous on my backup clouster.

To be able to mount the mirrored rbd image (without a protected snapshot):
  rbd-nbd --read-only map cluster5-rbd/vm-114-disk-1
--cluster backup

I just need to upgrade my backup cluster?


Thanks,
Michael

On Thu, Mar 19, 2020 at 1:06 PM Jason Dillaman <jdillama@xxxxxxxxxx> wrote:
>
> On Thu, Mar 19, 2020 at 6:19 AM Eugen Block <eblock@xxxxxx> wrote:
> >
> > Hi,
> >
> > one workaround would be to create a protected snapshot on the primary
> > image which is also mirrored, and then clone that snapshot on the
> > remote site. That clone can be accessed as required.
>
> +1. This is the correct approach. If you are using a Mimic+ cluster
> (i.e. require OSD release >= Mimic), you can use skip the protect
> step.
>
> > I'm not sure if there's a way to directly access the remote image
> > since it's read-only.
> >
> > Regards,
> > Eugen
> >
> >
> > Zitat von Ml Ml <mliebherr99@xxxxxxxxxxxxxx>:
> >
> > > Hello,
> > >
> > > my goal is to back up a proxmox cluster with rbd-mirror for desaster
> > > recovery. Promoting/Demoting, etc.. works great.
> > >
> > > But how can i access a single file on the mirrored cluster? I tried:
> > >
> > >    root@ceph01:~# rbd-nbd --read-only map cluster5-rbd/vm-114-disk-1
> > > --cluster backup
> > >    /dev/nbd1
> > >
> > > But i get:
> > >    root@ceph01:~# fdisk -l /dev/nbd1
> > >    fdisk: cannot open /dev/nbd1: Input/output error
> > >
> > > dmesg shows stuff like:
> > >    [Thu Mar 19 09:29:55 2020]  nbd1: unable to read partition table
> > >    [Thu Mar 19 09:29:55 2020] block nbd1: Other side returned error (30)
> > >    [Thu Mar 19 09:29:55 2020] block nbd1: Other side returned error (30)
> > >
> > > Here is my state:
> > >
> > > root@ceph01:~# rbd --cluster backup mirror pool status cluster5-rbd --verbose
> > > health: OK
> > > images: 3 total
> > >     3 replaying
> > >
> > > vm-106-disk-0:
> > >   global_id:   0bc18ee1-1749-4787-a45d-01c7e946ff06
> > >   state:       up+replaying
> > >   description: replaying, master_position=[object_number=3, tag_tid=2,
> > > entry_tid=3], mirror_position=[object_number=3, tag_tid=2,
> > > entry_tid=3], entries_behind_master=0
> > >   last_update: 2020-03-19 09:29:17
> > >
> > > vm-114-disk-1:
> > >   global_id:   2219ffa9-a4e0-4f89-b352-ff30b1ffe9b9
> > >   state:       up+replaying
> > >   description: replaying, master_position=[object_number=390,
> > > tag_tid=6, entry_tid=334290], mirror_position=[object_number=382,
> > > tag_tid=6, entry_tid=328526], entries_behind_master=5764
> > >   last_update: 2020-03-19 09:29:17
> > >
> > > vm-115-disk-0:
> > >   global_id:   2b0af493-14c1-4b10-b557-84928dc37dd1
> > >   state:       up+replaying
> > >   description: replaying, master_position=[object_number=72,
> > > tag_tid=1, entry_tid=67796], mirror_position=[object_number=72,
> > > tag_tid=1, entry_tid=67796], entries_behind_master=0
> > >   last_update: 2020-03-19 09:29:17
> > >
> > > More dmesg stuff:
> > > [Thu Mar 19 09:29:55 2020] block nbd1: Other side returned error (30)
> > > [Thu Mar 19 09:29:55 2020] block nbd1: Other side returned error (30)
> > > [Thu Mar 19 09:29:55 2020] block nbd1: Other side returned error (30)
> > > [Thu Mar 19 09:29:55 2020] block nbd1: Other side returned error (30)
> > > [Thu Mar 19 09:29:55 2020] block nbd1: Other side returned error (30)
> > > [Thu Mar 19 09:29:55 2020] block nbd1: Other side returned error (30)
> > > [Thu Mar 19 09:29:55 2020] block nbd1: Other side returned error (30)
> > > [Thu Mar 19 09:30:02 2020] block nbd1: Other side returned error (30)
> > > [Thu Mar 19 09:30:02 2020] blk_update_request: 95 callbacks suppressed
> > > [Thu Mar 19 09:30:02 2020] blk_update_request: I/O error, dev nbd1, sector 0
> > > [Thu Mar 19 09:30:02 2020] buffer_io_error: 94 callbacks suppressed
> > > [Thu Mar 19 09:30:02 2020] Buffer I/O error on dev nbd1, logical block
> > > 0, async page read
> > > [Thu Mar 19 09:30:02 2020] block nbd1: Other side returned error (30)
> > > [Thu Mar 19 09:30:02 2020] blk_update_request: I/O error, dev nbd1, sector 1
> > > [Thu Mar 19 09:30:02 2020] Buffer I/O error on dev nbd1, logical block
> > > 1, async page read
> > > [Thu Mar 19 09:30:02 2020] block nbd1: Other side returned error (30)
> > > [Thu Mar 19 09:30:02 2020] blk_update_request: I/O error, dev nbd1, sector 2
> > > [Thu Mar 19 09:30:02 2020] Buffer I/O error on dev nbd1, logical block
> > > 2, async page read
> > > [Thu Mar 19 09:30:02 2020] block nbd1: Other side returned error (30)
> > > [Thu Mar 19 09:30:02 2020] blk_update_request: I/O error, dev nbd1, sector 3
> > > [Thu Mar 19 09:30:02 2020] Buffer I/O error on dev nbd1, logical block
> > > 3, async page read
> > > [Thu Mar 19 09:30:02 2020] block nbd1: Other side returned error (30)
> > > [Thu Mar 19 09:30:02 2020] blk_update_request: I/O error, dev nbd1, sector 4
> > > [Thu Mar 19 09:30:02 2020] Buffer I/O error on dev nbd1, logical block
> > > 4, async page read
> > > [Thu Mar 19 09:30:02 2020] block nbd1: Other side returned error (30)
> > > [Thu Mar 19 09:30:02 2020] blk_update_request: I/O error, dev nbd1, sector 5
> > > [Thu Mar 19 09:30:02 2020] Buffer I/O error on dev nbd1, logical block
> > > 5, async page read
> > > [Thu Mar 19 09:30:02 2020] block nbd1: Other side returned error (30)
> > > [Thu Mar 19 09:30:02 2020] blk_update_request: I/O error, dev nbd1, sector 6
> > > [Thu Mar 19 09:30:02 2020] Buffer I/O error on dev nbd1, logical block
> > > 6, async page read
> > > [Thu Mar 19 09:30:02 2020] block nbd1: Other side returned error (30)
> > > [Thu Mar 19 09:30:02 2020] blk_update_request: I/O error, dev nbd1, sector 7
> > > [Thu Mar 19 09:30:02 2020] Buffer I/O error on dev nbd1, logical block
> > > 7, async page read
> > >
> > > Do i have to stop the replaying or how can i mount the image on the
> > > backup cluster?
> > >
> > > Thanks,
> > > Michael
> > > _______________________________________________
> > > ceph-users mailing list -- ceph-users@xxxxxxx
> > > To unsubscribe send an email to ceph-users-leave@xxxxxxx
> >
> >
> > _______________________________________________
> > ceph-users mailing list -- ceph-users@xxxxxxx
> > To unsubscribe send an email to ceph-users-leave@xxxxxxx
> >
>
>
> --
> Jason
> _______________________________________________
> ceph-users mailing list -- ceph-users@xxxxxxx
> To unsubscribe send an email to ceph-users-leave@xxxxxxx
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux