Re: 1 MDS report slow metadata IOs

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



OSD flapping was due to ports blocked by firewall.
while mounting the file system, the directory structure shows csi volumes
subfolder folders
as in /tmp/cephFS/csi/csi-vol-xxxx/container-name-log
is there a way to not show the csi volumes in the path to the container log
as an example:
/tmp/cephFS/container-name-log
this would make it easier to update the containers configuration without
dealing the volumes sub folders in the path ..

Thanks!

On Wed, Oct 6, 2021 at 2:58 AM Eugen Block <eblock@xxxxxx> wrote:

> > what is causing the slow MDS metadata IOs ?
>
> Your flapping OSDs.
>
> > currently, there are 2 mds and 3 monitors deployed ..
> > would it help to just one mds and one monitor ?
>
> No, you need to figure out why your OSDs crash. More details about
> your setup (ceph version, deployment method, hardware resources) and
> the logs from a crashing OSD could help identify the issue.
>
>
> Zitat von Abdelillah Asraoui <aasraoui@xxxxxxxxx>:
>
> > The osds are continuously flapping up/down due to the slow MDS metadata
> IOs
> > ..
> > what is causing the slow MDS metadata IOs ?
> > currently, there are 2 mds and 3 monitors deployed ..
> > would it help to just one mds and one monitor ?
> >
> > thanks!
> >
> > On Tue, Oct 5, 2021 at 1:42 PM Eugen Block <eblock@xxxxxx> wrote:
> >
> >> All your PGs are inactive, if two of four OSDs are down and you
> >> probably have a pool size of 3 then no IO can be served. You’d need at
> >> least three up ODSs to resolve that.
> >>
> >>
> >> Zitat von Abdelillah Asraoui <aasraoui@xxxxxxxxx>:
> >>
> >> > Ceph is reporting warning on slow metdataIOs on one of the MDS server,
> >> > this is
> >> >
> >> > a new cluster with no upgrade..
> >> >
> >> > Anyone has encountered this and is there a workaround ..
> >> >
> >> > ceph -s
> >> >
> >> >   cluster:
> >> >
> >> >     id:     801691e6xx-x-xx-xx-xx
> >> >
> >> >     health: HEALTH_WARN
> >> >
> >> >             1 MDSs report slow metadata IOs
> >> >
> >> >             noscrub,nodeep-scrub flag(s) set
> >> >
> >> >             2 osds down
> >> >
> >> >             2 hosts (2 osds) down
> >> >
> >> >             Reduced data availability: 97 pgs inactive, 66 pgs
> peering,
> >> 53
> >> > pgs stale
> >> >
> >> >             Degraded data redundancy: 31 pgs undersized
> >> >
> >> >             2 slow ops, oldest one blocked for 30 sec, osd.0 has slow
> ops
> >> >
> >> >
> >> >
> >> >   services:
> >> >
> >> >     mon: 3 daemons, quorum a,c,f (age 15h)
> >> >
> >> >     mgr: a(active, since 17h)
> >> >
> >> >     mds: myfs:1 {0=myfs-a=up:creating} 1 up:standby
> >> >
> >> >     osd: 4 osds: 2 up (since 36s), 4 in (since 10h)
> >> >
> >> >          flags noscrub,nodeep-scrub
> >> >
> >> >
> >> >
> >> >   data:
> >> >
> >> >     pools:   4 pools, 97 pgs
> >> >
> >> >     objects: 0 objects, 0 B
> >> >
> >> >     usage:   1.0 GiB used, 1.8 TiB / 1.8 TiB avail
> >> >
> >> >     pgs:     100.000% pgs not active
> >> >
> >> >              44 creating+peering
> >> >
> >> >              31 stale+undersized+peered
> >> >
> >> >              22 stale+creating+peering
> >> >
> >> >
> >> >
> >> >   progress:
> >> >
> >> >     Rebalancing after osd.2 marked in (10h)
> >> >
> >> >       [............................]
> >> >
> >> >     Rebalancing after osd.3 marked in (10h)
> >> >
> >> >       [............................]
> >> >
> >> >
> >> > Thanks!
> >> > _______________________________________________
> >> > ceph-users mailing list -- ceph-users@xxxxxxx
> >> > To unsubscribe send an email to ceph-users-leave@xxxxxxx
> >>
> >>
> >>
> >>
>
>
>
>
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux