On Tue, Mar 12, 2019 at 11:09 PM Vikas Rana <vrana@xxxxxxxxxxxx> wrote: > > Hi there, > > > > We are replicating a RBD image from Primary to DR site using RBD mirroring. > > On Primary, we were using 10.2.10. Just a note that Jewel is end-of-life upstream. > DR site is luminous and we promoted the DR copy to test the failure. Everything checked out good. > > > > Now we are trying to restart the replication and we did the demote and then resync the image but it stuck in “starting_replay” state for last 3 days. It’s a 200TB RBD image You would need to run "rbd --cluster <primary-site> mirror image resync nfs/dir_research" and wait for that to complete *before* demoting the primary image on cluster "cephdr". Without a primary image, there is nothing to resync against. > > > :~# rbd --cluster cephdr mirror pool status nfs --verbose > > health: WARNING > > images: 1 total > > 1 starting_replay > > > > dir_research: > > global_id: 3ad67d0c-e06b-406a-9469-4e5faedd09a4 > > state: down+unknown > > description: status not found > > last_update: > > > > > > #rbd info nfs/dir_research > > rbd image 'dir_research': > > size 200TiB in 52428800 objects > > order 22 (4MiB objects) > > block_name_prefix: rbd_data.652186b8b4567 > > format: 2 > > features: layering, exclusive-lock, journaling > > flags: > > create_timestamp: Thu Feb 7 11:53:36 2019 > > journal: 652186b8b4567 > > mirroring state: disabling > > mirroring global id: 3ad67d0c-e06b-406a-9469-4e5faedd09a4 > > mirroring primary: false > > > So the question is, how do we know the progress of the replay and how much its already completed and any ETA estimation on when it will go back to OK state? > > > > > Thanks, > > -Vikas > > > > _______________________________________________ > ceph-users mailing list > ceph-users@xxxxxxxxxxxxxx > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com -- Jason _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com