OSD flapping was due to ports blocked by firewall. while mounting the file system, the directory structure shows csi volumes subfolder folders as in /tmp/cephFS/csi/csi-vol-xxxx/container-name-log is there a way to not show the csi volumes in the path to the container log as an example: /tmp/cephFS/container-name-log this would make it easier to update the containers configuration without dealing the volumes sub folders in the path .. Thanks! On Wed, Oct 6, 2021 at 2:58 AM Eugen Block <eblock@xxxxxx> wrote: > > what is causing the slow MDS metadata IOs ? > > Your flapping OSDs. > > > currently, there are 2 mds and 3 monitors deployed .. > > would it help to just one mds and one monitor ? > > No, you need to figure out why your OSDs crash. More details about > your setup (ceph version, deployment method, hardware resources) and > the logs from a crashing OSD could help identify the issue. > > > Zitat von Abdelillah Asraoui <aasraoui@xxxxxxxxx>: > > > The osds are continuously flapping up/down due to the slow MDS metadata > IOs > > .. > > what is causing the slow MDS metadata IOs ? > > currently, there are 2 mds and 3 monitors deployed .. > > would it help to just one mds and one monitor ? > > > > thanks! > > > > On Tue, Oct 5, 2021 at 1:42 PM Eugen Block <eblock@xxxxxx> wrote: > > > >> All your PGs are inactive, if two of four OSDs are down and you > >> probably have a pool size of 3 then no IO can be served. You’d need at > >> least three up ODSs to resolve that. > >> > >> > >> Zitat von Abdelillah Asraoui <aasraoui@xxxxxxxxx>: > >> > >> > Ceph is reporting warning on slow metdataIOs on one of the MDS server, > >> > this is > >> > > >> > a new cluster with no upgrade.. > >> > > >> > Anyone has encountered this and is there a workaround .. > >> > > >> > ceph -s > >> > > >> > cluster: > >> > > >> > id: 801691e6xx-x-xx-xx-xx > >> > > >> > health: HEALTH_WARN > >> > > >> > 1 MDSs report slow metadata IOs > >> > > >> > noscrub,nodeep-scrub flag(s) set > >> > > >> > 2 osds down > >> > > >> > 2 hosts (2 osds) down > >> > > >> > Reduced data availability: 97 pgs inactive, 66 pgs > peering, > >> 53 > >> > pgs stale > >> > > >> > Degraded data redundancy: 31 pgs undersized > >> > > >> > 2 slow ops, oldest one blocked for 30 sec, osd.0 has slow > ops > >> > > >> > > >> > > >> > services: > >> > > >> > mon: 3 daemons, quorum a,c,f (age 15h) > >> > > >> > mgr: a(active, since 17h) > >> > > >> > mds: myfs:1 {0=myfs-a=up:creating} 1 up:standby > >> > > >> > osd: 4 osds: 2 up (since 36s), 4 in (since 10h) > >> > > >> > flags noscrub,nodeep-scrub > >> > > >> > > >> > > >> > data: > >> > > >> > pools: 4 pools, 97 pgs > >> > > >> > objects: 0 objects, 0 B > >> > > >> > usage: 1.0 GiB used, 1.8 TiB / 1.8 TiB avail > >> > > >> > pgs: 100.000% pgs not active > >> > > >> > 44 creating+peering > >> > > >> > 31 stale+undersized+peered > >> > > >> > 22 stale+creating+peering > >> > > >> > > >> > > >> > progress: > >> > > >> > Rebalancing after osd.2 marked in (10h) > >> > > >> > [............................] > >> > > >> > Rebalancing after osd.3 marked in (10h) > >> > > >> > [............................] > >> > > >> > > >> > Thanks! > >> > _______________________________________________ > >> > ceph-users mailing list -- ceph-users@xxxxxxx > >> > To unsubscribe send an email to ceph-users-leave@xxxxxxx > >> > >> > >> > >> > > > > _______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx