You're probably seeing this bug: https://tracker.ceph.com/issues/48673 Sorry I've not had time to finish a fix for it yet. Hopefully soon... On Tue, Jul 19, 2022 at 5:43 PM Bryan Stillwell <bstillwell@xxxxxxxxxxx> wrote: > > We have a cluster using multiple filesystems on Pacific (16.2.7) and even though we have mds_cache_memory_limit set to 80 GiB one of the MDS daemons is using 123.1 GiB. This MDS is actually the standby-replay MDS and I'm wondering if it's because it's using more dns/inos/dirs than the active MDS?: > > $ sudo ceph fs status cephfs19 > cephfs19 - 28 clients > ======== > RANK STATE MDS ACTIVITY DNS INOS DIRS CAPS > 0 active ceph006b Reqs: 2879 /s 27.8M 27.8M 3490k 7767k > 0-s standby-replay ceph008a Evts: 1446 /s 40.1M 40.0M 6259k 0 > > Shouldn't the standby-replay MDS daemons have similar stats to the active MDS they're protecting? What could be causing this to happen? > > Thanks, > Bryan > _______________________________________________ > ceph-users mailing list -- ceph-users@xxxxxxx > To unsubscribe send an email to ceph-users-leave@xxxxxxx > -- Patrick Donnelly, Ph.D. He / Him / His Principal Software Engineer Red Hat, Inc. GPG: 19F28A586F808C2402351B93C3301A3E258DD79D _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx