I think this bug has already been worked on in https://tracker.ceph.com/issues/63364, can you tell which version you're on? -- *Dhairya Parmar* Associate Software Engineer, CephFS IBM, Inc. On Tue, Mar 26, 2024 at 2:32 AM Erich Weiler <weiler@xxxxxxxxxxxx> wrote: > Hi Y'all, > > I'm seeing this warning via 'ceph -s' (this is on Reef): > > # ceph -s > cluster: > id: 58bde08a-d7ed-11ee-9098-506b4b4da440 > health: HEALTH_WARN > 3 clients failing to advance oldest client/flush tid > 1 MDSs report slow requests > 1 MDSs behind on trimming > > services: > mon: 5 daemons, quorum > pr-md-01,pr-md-02,pr-store-01,pr-store-02,pr-md-03 (age 3d) > mgr: pr-md-01.jemmdf(active, since 3w), standbys: pr-md-02.emffhz > mds: 1/1 daemons up, 1 standby > osd: 46 osds: 46 up (since 3d), 46 in (since 2w) > > data: > volumes: 1/1 healthy > pools: 4 pools, 1313 pgs > objects: 258.13M objects, 454 TiB > usage: 688 TiB used, 441 TiB / 1.1 PiB avail > pgs: 1303 active+clean > 8 active+clean+scrubbing > 2 active+clean+scrubbing+deep > > io: > client: 131 MiB/s rd, 111 MiB/s wr, 41 op/s rd, 613 op/s wr > > I googled around and looked at the docs and it seems like this isn't a > critical problem, but I couldn't find a clear path to resolution. Does > anyone have any advice on what I can do to resolve the health issues up > top? > > My CephFS filesystem is incredibly busy so I have a feeling that has > some impact here, but not 100% sure... > > Thanks as always for the help! > > cheers, > erich > _______________________________________________ > ceph-users mailing list -- ceph-users@xxxxxxx > To unsubscribe send an email to ceph-users-leave@xxxxxxx > > _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx