Re: rbd-mirror replay is very slow - but initial bootstrap is fast

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Tue, Mar 10, 2020 at 2:31 PM Anthony D'Atri <anthony.datri@xxxxxxxxx> wrote:
>
> FWIW when using rbd-mirror to migrate volumes between SATA SSD clusters, I found that
>
>
>    rbd_mirror_journal_max_fetch_bytes:
>     section: "client"
>     value: "33554432"
>
>   rbd_journal_max_payload_bytes:
>     section: "client"
>     value: “8388608"

Indeed, that's a good tweak that applies to the primary-side librbd
client for the mirrored image for IO workloads that routinely issue
large (> than the 16KiB default), sequential writes. This was another
compromise configuration setting to reduce the potential memory
footprint of the rbd-mirror daemon.

> Made a world of difference in expediting journal reply on Luminous 12.2.2.  With defaults, some active voumes would take hours to converge, and a couple were falling even more behind.
>
> This was mirroring 1 to 2 volumes at a time.  YMMV.
>
>
>
>
> > On Mar 10, 2020, at 7:36 AM, Ml Ml <mliebherr99@xxxxxxxxxxxxxx> wrote:
> >
> > Hello Jason,
> >
> > thanks for that fast reply.
> >
> > This is now my /etc/ceph/ceph.conf
> >
> > [client]
> > rbd_mirror_journal_max_fetch_bytes = 4194304
> >
> >
> > I stopped and started my rbd-mirror manually with:
> > rbd-mirror -d -c /etc/ceph/ceph.conf
> >
> > Still same result. Slow speed shown by iftop and entries_behind_master
> > keeps increasing a lot if i produce 20MB/sec traffic on that
> > replication image.
> >
> > The latency is like:
> > --- 10.10.50.1 ping statistics ---
> > 100 packets transmitted, 100 received, 0% packet loss, time 20199ms
> > rtt min/avg/max/mdev = 0.067/0.286/1.418/0.215 ms
> >
> > iperf from the source node to the destination node (where the
> > rbd-mirror runs): 8.92 Gbits/sec
> >
> > Any other idea?
> >
> > Thanks,
> > Michael
> >
> >
> >
> > On Tue, Mar 10, 2020 at 2:19 PM Jason Dillaman <jdillama@xxxxxxxxxx> wrote:
> >>
> >> On Tue, Mar 10, 2020 at 6:47 AM Ml Ml <mliebherr99@xxxxxxxxxxxxxx> wrote:
> >>>
> >>> Hello List,
> >>>
> >>> when i initially enable journal/mirror on an image it gets
> >>> bootstrapped to my site-b pretty quickly with 250MB/sec which is about
> >>> the IO Write limit.
> >>>
> >>> Once its up2date, the replay is very slow. About 15KB/sec and the
> >>> entries_behind_maste is just running away:
> >>>
> >>> root@ceph01:~# rbd --cluster backup mirror pool status rbd-cluster6 --verbose
> >>> health: OK
> >>> images: 3 total
> >>>    3 replaying
> >>>
> >>> ...
> >>>
> >>> vm-112-disk-0:
> >>>  global_id:   60a795c3-9f5d-4be3-b9bd-3df971e531fa
> >>>  state:       up+replaying
> >>>  description: replaying, master_position=[object_number=623,
> >>> tag_tid=3, entry_tid=345567], mirror_position=[object_number=35,
> >>> tag_tid=3, entry_tid=18371], entries_behind_master=327196
> >>>  last_update: 2020-03-10 11:36:44
> >>>
> >>> ...
> >>>
> >>> Write traffic on the source is about 20/25MB/sec.
> >>>
> >>> On the Source i run 14.2.6 and on the destination 12.2.13.
> >>>
> >>> Any idea why the replaying is sooo slow?
> >>
> >> What is the latency between the two clusters?
> >>
> >> I would recommend increasing the "rbd_mirror_journal_max_fetch_bytes"
> >> config setting (defaults to 32KiB) on your destination cluster. i.e.
> >> try adding add "rbd_mirror_journal_max_fetch_bytes = 4194304" to the
> >> "[client]" section of your Ceph configuration file on the node where
> >> "rbd-mirror" daemon is running, and restart it. It defaults to a very
> >> small read size from the remote cluster in a primitive attempt to
> >> reduce the potential memory usage of the rbd-mirror daemon, but it has
> >> the side-effect of slowing down mirroring for links with higher
> >> latencies.
> >>
> >>>
> >>> Thanks,
> >>> Michael
> >>> _______________________________________________
> >>> ceph-users mailing list -- ceph-users@xxxxxxx
> >>> To unsubscribe send an email to ceph-users-leave@xxxxxxx
> >>>
> >>
> >>
> >> --
> >> Jason
> >>
> > _______________________________________________
> > ceph-users mailing list -- ceph-users@xxxxxxx
> > To unsubscribe send an email to ceph-users-leave@xxxxxxx
> _______________________________________________
> ceph-users mailing list -- ceph-users@xxxxxxx
> To unsubscribe send an email to ceph-users-leave@xxxxxxx



-- 
Jason
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux