Re: rbd-mirror sync image continuously or only sync once

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Tue, Jun 9, 2020 at 7:26 AM Zhenshi Zhou <deaderzzs@xxxxxxxxx> wrote:
>
> I did promote the non-primary image, or I couldn't disable the image mirror.

OK, that means that 100% of the data was properly transferred since it
needs to replay previous events before it can get to the demotion
event, replay that, so that you could non-force promote. How are you
writing to the original primary image? Are you flushing your data?

> Jason Dillaman <jdillama@xxxxxxxxxx> 于2020年6月9日周二 下午7:19写道:
>>
>> On Mon, Jun 8, 2020 at 11:42 PM Zhenshi Zhou <deaderzzs@xxxxxxxxx> wrote:
>> >
>> > I have just done a test on rbd-mirror. Follow the steps:
>> > 1. deploy two new clusters, clusterA and clusterB
>> > 2. configure one-way replication from clusterA to clusterB with rbd-mirror
>> > 3. write data to rbd_blk on clusterA once every 5 seconds
>> > 4. get information with 'rbd mirror image status rbd_blk', the "state" is "up+replaying"
>> > 5. demote image on clusterA(just wanna stop syncing and switch client connection to clusterB)
>> >
>> > The result is:
>> > 1. I find that the "last_update" from "rbd mirror image status" updates every 30 seconds, which means I will lose at most 30s of data
>>
>> The status updates are throttled but not the data transfer.
>>
>> > 2. I stopped syncing on 11:02, while the data from rbd_blk on the clusterB is not newer than 10:50.
>>
>> After demotion, you need to promote on the original non-primary side
>> (w/o the --force) option. It won't let you non-force promote unless it
>> has all the data copied. It will continue to copy data while the other
>> side has been demoted until it fully catches up.
>>
>> >
>> > Did I have the wrong steps in the switching progress?
>> >
>> >
>> >
>> > Zhenshi Zhou <deaderzzs@xxxxxxxxx> 于2020年6月9日周二 上午8:57写道:
>> >>
>> >> Well, I'm afraid that the image didn't replay continuously, which means I have some data lost.
>> >> The "rbd mirror image status" shows the image is replayed and its time is just before I demote
>> >> the primary image. I lost about 24 hours' data and I'm not sure whether there is an interval
>> >> between the synchronization.
>> >>
>> >> I use version 14.2.9 and I deployed a one direction mirror.
>> >>
>> >> Zhenshi Zhou <deaderzzs@xxxxxxxxx> 于2020年6月5日周五 上午10:22写道:
>> >>>
>> >>> Thank you for the clarification. That's very clear.
>> >>>
>> >>> Jason Dillaman <jdillama@xxxxxxxxxx> 于2020年6月5日周五 上午12:46写道:
>> >>>>
>> >>>> On Thu, Jun 4, 2020 at 3:43 AM Zhenshi Zhou <deaderzzs@xxxxxxxxx> wrote:
>> >>>> >
>> >>>> > My condition is that the primary image being used while rbd-mirror sync.
>> >>>> > I want to get the period between the two times of rbd-mirror transfer the
>> >>>> > increased data.
>> >>>> > I will search those options you provided, thanks a lot :)
>> >>>>
>> >>>> When using the original (pre-Octopus) journal-based mirroring, once
>> >>>> the initial sync completes to transfer the bulk of the image data from
>> >>>> a point-in-time dynamic snapshot, any changes post sync will be
>> >>>> replayed continuously from the stream of events written to the journal
>> >>>> on the primary image. The "rbd mirror image status" against the
>> >>>> non-primary image will provide more details about the current state of
>> >>>> the journal replay.
>> >>>>
>> >>>> With the Octopus release, we now also support snapshot-based mirroring
>> >>>> where we transfer any image deltas between two mirroring snapshots.
>> >>>> These mirroring snapshots are different from user-created snapshots
>> >>>> and their life-time is managed by RBD mirroring (i.e. they are
>> >>>> automatically pruned when no longer needed). This version of mirroring
>> >>>> probably more closely relates to your line of questioning since the
>> >>>> period of replication is at whatever period you create new mirroring
>> >>>> snapshots (provided your two clusters can keep up).
>> >>>>
>> >>>> >
>> >>>> > Eugen Block <eblock@xxxxxx> 于2020年6月4日周四 下午3:28写道:
>> >>>> >
>> >>>> > > The initial sync is a full image sync, the rest is based on the object
>> >>>> > > sets created. There are several options to control the mirroring, for
>> >>>> > > example:
>> >>>> > >
>> >>>> > > rbd_journal_max_concurrent_object_sets
>> >>>> > > rbd_mirror_concurrent_image_syncs
>> >>>> > > rbd_mirror_leader_max_missed_heartbeats
>> >>>> > >
>> >>>> > > and many more. I'm not sure I fully understand what you're asking,
>> >>>> > > maybe you could rephrase your question?
>> >>>> > >
>> >>>> > >
>> >>>> > > Zitat von Zhenshi Zhou <deaderzzs@xxxxxxxxx>:
>> >>>> > >
>> >>>> > > > Hi Eugen,
>> >>>> > > >
>> >>>> > > > Thanks for the reply. If rbd-mirror constantly synchronize changes,
>> >>>> > > > what frequency to replay once? I don't find any options I can config.
>> >>>> > > >
>> >>>> > > > Eugen Block <eblock@xxxxxx> 于2020年6月4日周四 下午2:54写道:
>> >>>> > > >
>> >>>> > > >> Hi,
>> >>>> > > >>
>> >>>> > > >> that's the point of rbd-mirror, to constantly replay changes from the
>> >>>> > > >> primary image to the remote image (if the rbd journal feature is
>> >>>> > > >> enabled).
>> >>>> > > >>
>> >>>> > > >>
>> >>>> > > >> Zitat von Zhenshi Zhou <deaderzzs@xxxxxxxxx>:
>> >>>> > > >>
>> >>>> > > >> > Hi all,
>> >>>> > > >> >
>> >>>> > > >> > I'm gonna deploy a rbd-mirror in order to sync image from clusterA to
>> >>>> > > >> > clusterB.
>> >>>> > > >> > The image will be used while syncing. I'm not sure if the rbd-mirror
>> >>>> > > will
>> >>>> > > >> > sync image
>> >>>> > > >> > continuously or not. If not, I will inform clients not to write data
>> >>>> > > in
>> >>>> > > >> it.
>> >>>> > > >> >
>> >>>> > > >> > Thanks. Regards
>> >>>> > > >> > _______________________________________________
>> >>>> > > >> > ceph-users mailing list -- ceph-users@xxxxxxx
>> >>>> > > >> > To unsubscribe send an email to ceph-users-leave@xxxxxxx
>> >>>> > > >>
>> >>>> > > >>
>> >>>> > > >> _______________________________________________
>> >>>> > > >> ceph-users mailing list -- ceph-users@xxxxxxx
>> >>>> > > >> To unsubscribe send an email to ceph-users-leave@xxxxxxx
>> >>>> > > >>
>> >>>> > >
>> >>>> > >
>> >>>> > >
>> >>>> > >
>> >>>> > _______________________________________________
>> >>>> > ceph-users mailing list -- ceph-users@xxxxxxx
>> >>>> > To unsubscribe send an email to ceph-users-leave@xxxxxxx
>> >>>>
>> >>>>
>> >>>>
>> >>>> --
>> >>>> Jason
>> >>>>
>>
>>
>> --
>> Jason
>>


-- 
Jason
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux