output from "ceph osd pool ls detail"
pool 2 'cephfs_metadata' replicated size 3 min_size 2 crush_rule 0 object_hash rjenkins pg_num 32 pgp_num 32 last_change 75 lfor 0/67 flags hashpspool stripe_width 0 application cephfs
On Thu, Aug 29, 2019 at 6:13 PM Heðin Ejdesgaard Møller <hej@xxxxxxxxx> wrote:
What's the output of
ceph osd pool ls detail
On hós, 2019-08-29 at 18:06 +0530, Amudhan P wrote:
> output from "ceph -s "
>
> cluster:
> id: 7c138e13-7b98-4309-b591-d4091a1742b4
> health: HEALTH_WARN
> Degraded data redundancy: 1141587/7723191 objects
> degraded (14.781%), 15 pgs degraded, 16 pgs undersized
>
> services:
> mon: 1 daemons, quorum mon01
> mgr: mon01(active)
> mds: cephfs-tst-1/1/1 up {0=mon01=up:active}
> osd: 3 osds: 3 up, 3 in; 16 remapped pgs
>
> data:
> pools: 2 pools, 64 pgs
> objects: 2.57 M objects, 59 GiB
> usage: 190 GiB used, 5.3 TiB / 5.5 TiB avail
> pgs: 1141587/7723191 objects degraded (14.781%)
> 48 active+clean
> 15 active+undersized+degraded+remapped+backfill_wait
> 1 active+undersized+remapped+backfilling
>
> io:
> recovery: 0 B/s, 10 objects/s
>
> output from "ceph osd tree"
> ID CLASS WEIGHT TYPE NAME STATUS REWEIGHT PRI-AFF
> -1 5.45819 root default
> -3 1.81940 host test-node1
> 0 hdd 1.81940 osd.0 up 1.00000 1.00000
> -5 1.81940 host test-node2
> 1 hdd 1.81940 osd.1 up 1.00000 1.00000
> -7 1.81940 host test-node3
> 2 hdd 1.81940 osd.2 up 1.00000 1.00000
>
> failure domain not configured yet, setup is 3 OSD node each with a
> single disk, 1 node with mon&mds&mgr running.
> the cluster was healthy until I run a script for creating multiple
> folders.
>
> regards
> Amudhan
>
> On Thu, Aug 29, 2019 at 5:33 PM Heðin Ejdesgaard Møller <
> hej@xxxxxxxxx> wrote:
> > In adition to ceph -s, could you provide the output of
> > ceph osd tree
> > and specify what your failure domain is ?
> >
> > /Heðin
> >
> >
> > On hós, 2019-08-29 at 13:55 +0200, Janne Johansson wrote:
> > >
> > >
> > > Den tors 29 aug. 2019 kl 13:50 skrev Amudhan P <
> > amudhan83@xxxxxxxxx>:
> > > > Hi,
> > > >
> > > > I am using ceph version 13.2.6 (mimic) on test setup trying
> > with
> > > > cephfs.
> > > > my ceph health status showing warning .
> > > >
> > > > "ceph health"
> > > > HEALTH_WARN Degraded data redundancy: 1197023/7723191 objects
> > > > degraded (15.499%)
> > > >
> > > > "ceph health detail"
> > > > HEALTH_WARN Degraded data redundancy: 1197128/7723191 objects
> > > > degraded (15.500%)
> > > > PG_DEGRADED Degraded data redundancy: 1197128/7723191 objects
> > > > degraded (15.500%)
> > > > pg 2.0 is stuck undersized for 1076.454929, current state
> > > > active+undersized+
> > > > pg 2.2 is stuck undersized for 1076.456639, current state
> > > > active+undersized+
> > > >
> > >
> > > How does "ceph -s" look?
> > > It should have more info on what else is wrong.
> > >
> > > --
> > > May the most significant bit of your life be positive.
> > > _______________________________________________
> > > ceph-users mailing list -- ceph-users@xxxxxxx
> > > To unsubscribe send an email to ceph-users-leave@xxxxxxx
> >
_______________________________________________ ceph-users mailing list -- ceph-users@xxxxxxx To unsubscribe send an email to ceph-users-leave@xxxxxxx