The osds are continuously flapping up/down due to the slow MDS metadata IOs .. what is causing the slow MDS metadata IOs ? currently, there are 2 mds and 3 monitors deployed .. would it help to just one mds and one monitor ? thanks! On Tue, Oct 5, 2021 at 1:42 PM Eugen Block <eblock@xxxxxx> wrote: > All your PGs are inactive, if two of four OSDs are down and you > probably have a pool size of 3 then no IO can be served. You’d need at > least three up ODSs to resolve that. > > > Zitat von Abdelillah Asraoui <aasraoui@xxxxxxxxx>: > > > Ceph is reporting warning on slow metdataIOs on one of the MDS server, > > this is > > > > a new cluster with no upgrade.. > > > > Anyone has encountered this and is there a workaround .. > > > > ceph -s > > > > cluster: > > > > id: 801691e6xx-x-xx-xx-xx > > > > health: HEALTH_WARN > > > > 1 MDSs report slow metadata IOs > > > > noscrub,nodeep-scrub flag(s) set > > > > 2 osds down > > > > 2 hosts (2 osds) down > > > > Reduced data availability: 97 pgs inactive, 66 pgs peering, > 53 > > pgs stale > > > > Degraded data redundancy: 31 pgs undersized > > > > 2 slow ops, oldest one blocked for 30 sec, osd.0 has slow ops > > > > > > > > services: > > > > mon: 3 daemons, quorum a,c,f (age 15h) > > > > mgr: a(active, since 17h) > > > > mds: myfs:1 {0=myfs-a=up:creating} 1 up:standby > > > > osd: 4 osds: 2 up (since 36s), 4 in (since 10h) > > > > flags noscrub,nodeep-scrub > > > > > > > > data: > > > > pools: 4 pools, 97 pgs > > > > objects: 0 objects, 0 B > > > > usage: 1.0 GiB used, 1.8 TiB / 1.8 TiB avail > > > > pgs: 100.000% pgs not active > > > > 44 creating+peering > > > > 31 stale+undersized+peered > > > > 22 stale+creating+peering > > > > > > > > progress: > > > > Rebalancing after osd.2 marked in (10h) > > > > [............................] > > > > Rebalancing after osd.3 marked in (10h) > > > > [............................] > > > > > > Thanks! > > _______________________________________________ > > ceph-users mailing list -- ceph-users@xxxxxxx > > To unsubscribe send an email to ceph-users-leave@xxxxxxx > > > > _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx