Re: 1 MDS report slow metadata IOs

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



what is causing the slow MDS metadata IOs ?

Your flapping OSDs.

currently, there are 2 mds and 3 monitors deployed ..
would it help to just one mds and one monitor ?

No, you need to figure out why your OSDs crash. More details about your setup (ceph version, deployment method, hardware resources) and the logs from a crashing OSD could help identify the issue.


Zitat von Abdelillah Asraoui <aasraoui@xxxxxxxxx>:

The osds are continuously flapping up/down due to the slow MDS metadata IOs
..
what is causing the slow MDS metadata IOs ?
currently, there are 2 mds and 3 monitors deployed ..
would it help to just one mds and one monitor ?

thanks!

On Tue, Oct 5, 2021 at 1:42 PM Eugen Block <eblock@xxxxxx> wrote:

All your PGs are inactive, if two of four OSDs are down and you
probably have a pool size of 3 then no IO can be served. You’d need at
least three up ODSs to resolve that.


Zitat von Abdelillah Asraoui <aasraoui@xxxxxxxxx>:

> Ceph is reporting warning on slow metdataIOs on one of the MDS server,
> this is
>
> a new cluster with no upgrade..
>
> Anyone has encountered this and is there a workaround ..
>
> ceph -s
>
>   cluster:
>
>     id:     801691e6xx-x-xx-xx-xx
>
>     health: HEALTH_WARN
>
>             1 MDSs report slow metadata IOs
>
>             noscrub,nodeep-scrub flag(s) set
>
>             2 osds down
>
>             2 hosts (2 osds) down
>
>             Reduced data availability: 97 pgs inactive, 66 pgs peering,
53
> pgs stale
>
>             Degraded data redundancy: 31 pgs undersized
>
>             2 slow ops, oldest one blocked for 30 sec, osd.0 has slow ops
>
>
>
>   services:
>
>     mon: 3 daemons, quorum a,c,f (age 15h)
>
>     mgr: a(active, since 17h)
>
>     mds: myfs:1 {0=myfs-a=up:creating} 1 up:standby
>
>     osd: 4 osds: 2 up (since 36s), 4 in (since 10h)
>
>          flags noscrub,nodeep-scrub
>
>
>
>   data:
>
>     pools:   4 pools, 97 pgs
>
>     objects: 0 objects, 0 B
>
>     usage:   1.0 GiB used, 1.8 TiB / 1.8 TiB avail
>
>     pgs:     100.000% pgs not active
>
>              44 creating+peering
>
>              31 stale+undersized+peered
>
>              22 stale+creating+peering
>
>
>
>   progress:
>
>     Rebalancing after osd.2 marked in (10h)
>
>       [............................]
>
>     Rebalancing after osd.3 marked in (10h)
>
>       [............................]
>
>
> Thanks!
> _______________________________________________
> ceph-users mailing list -- ceph-users@xxxxxxx
> To unsubscribe send an email to ceph-users-leave@xxxxxxx







_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux