Hello Linh, On Tue, Apr 24, 2018 at 12:34 AM, Linh Vu <vul@xxxxxxxxxxxxxx> wrote: > However, on our production cluster, with more powerful MDSes (10 cores > 3.4GHz, 256GB RAM, much faster networking), I get this in the logs > constantly: > > 2018-04-24 16:29:21.998261 7f02d1af9700 0 mds.1.migrator nicely exporting > to mds.0 [dir 0x1000010cd91.1110* /home/ [2,head] auth{0=1017} v=5632699 > cv=5632651/5632651 dir_auth=1 state=1611923458|complete|auxsubtree f(v84 > 55=0+55) n(v245771 rc2018-04-24 16:28:32.830971 b233439385711 > 423085=383063+40022) hs=55+0,ss=0+0 dirty=1 | child=1 frozen=0 subtree=1 > replicated=1 dirty=1 authpin=0 0x55691ccf1c00] > > To clarify, /home is pinned to mds.1, so there is no reason it should export > this to mds.0, and the loads on both MDSes (req/s, network load, CPU load) > are fairly low, lower than those on the test MDS VMs. As Dan said, this is simply a spurious log message. Nothing is being exported. This will be fixed in 12.2.6 as part of several fixes to the load balancer: https://github.com/ceph/ceph/pull/21412/commits/cace918dd044b979cd0d54b16a6296094c8a9f90 -- Patrick Donnelly _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com