All your PGs are inactive, if two of four OSDs are down and you
probably have a pool size of 3 then no IO can be served. You’d need at
least three up ODSs to resolve that.
Zitat von Abdelillah Asraoui <aasraoui@xxxxxxxxx>:
Ceph is reporting warning on slow metdataIOs on one of the MDS server,
this is
a new cluster with no upgrade..
Anyone has encountered this and is there a workaround ..
ceph -s
cluster:
id: 801691e6xx-x-xx-xx-xx
health: HEALTH_WARN
1 MDSs report slow metadata IOs
noscrub,nodeep-scrub flag(s) set
2 osds down
2 hosts (2 osds) down
Reduced data availability: 97 pgs inactive, 66 pgs peering, 53
pgs stale
Degraded data redundancy: 31 pgs undersized
2 slow ops, oldest one blocked for 30 sec, osd.0 has slow ops
services:
mon: 3 daemons, quorum a,c,f (age 15h)
mgr: a(active, since 17h)
mds: myfs:1 {0=myfs-a=up:creating} 1 up:standby
osd: 4 osds: 2 up (since 36s), 4 in (since 10h)
flags noscrub,nodeep-scrub
data:
pools: 4 pools, 97 pgs
objects: 0 objects, 0 B
usage: 1.0 GiB used, 1.8 TiB / 1.8 TiB avail
pgs: 100.000% pgs not active
44 creating+peering
31 stale+undersized+peered
22 stale+creating+peering
progress:
Rebalancing after osd.2 marked in (10h)
[............................]
Rebalancing after osd.3 marked in (10h)
[............................]
Thanks!
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx
_______________________________________________
ceph-users mailing list -- ceph-users@xxxxxxx
To unsubscribe send an email to ceph-users-leave@xxxxxxx