PG's Degraded on disk failure not remapped.

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi Cephers,

We've been testing drive failures and we're just trying to see if the
behaviour of our cluster is normal,  or if we've setup something wrong.

In summary; the OSD is down and out, but the PGs are showing as degraded
and don't seem to want to remap. We'd have assumed once the OSD was marked
out, that a re-map should have happened and we'd see misplaced rather than
degraded PGs.

  cluster bfb7e824-f37d-45c0-a4fc-a98182fed985
     health HEALTH_WARN
            43 pgs degraded
            43 pgs stuck degraded
            44 pgs stuck unclean
            43 pgs stuck undersized
            43 pgs undersized
            recovery 36899/6822836 objects degraded (0.541%)
            recovery 813/6822836 objects misplaced (0.012%)
     monmap e3: 3 mons at
{ceph-admin1=10.66.8.1:6789/0,ceph-store1=10.66.8.2:6789/0,ceph-store2=10.
66.8.3:6789/0}
            election epoch 950, quorum 0,1,2
ceph-admin1,ceph-store1,ceph-store2
     osdmap e6342: 36 osds: 35 up, 35 in; 1 remapped pgs
      pgmap v11805515: 1700 pgs, 3 pools, 13165 GB data, 3331 kobjects
            25941 GB used, 30044 GB / 55986 GB avail
            36899/6822836 objects degraded (0.541%)
            813/6822836 objects misplaced (0.012%)
                1656 active+clean
                  43 active+undersized+degraded
                   1 active+remapped
  client io 491 kB/s rd, 3998 kB/s wr, 480 op/s


# id	weight	type name	up/down	reweight
-6	43.56	root hdd
-2	21.78		host ceph-store1-hdd
0	3.63			osd.0	up	1
2	3.63			osd.2	up	1
4	3.63			osd.4	up	1
6	3.63			osd.6	up	1
8	3.63			osd.8	up	1
10	3.63			osd.10	up	1
-3	21.78		host ceph-store2-hdd
1	3.63			osd.1	up	1
3	3.63			osd.3	up	1
5	3.63			osd.5	up	1
7	3.63			osd.7	up	1
9	3.63			osd.9	up	1
11	3.63			osd.11	up	1
-1	11.48	root ssd
-4	5.74		host ceph-store1-ssd
12	0.43			osd.12	up	1
13	0.43			osd.13	up	1
14	0.43			osd.14	up	1
16	0.43			osd.16	up	1
18	0.43			osd.18	down	0
19	0.43			osd.19	up	1
20	0.43			osd.20	up	1
21	0.43			osd.21	up	1
32	0.72			osd.32	up	1
33	0.72			osd.33	up	1
17	0.43			osd.17	up	1
15	0.43			osd.15	up	1
-5	5.74		host ceph-store2-ssd
22	0.43			osd.22	up	1
23	0.43			osd.23	up	1
24	0.43			osd.24	up	1
25	0.43			osd.25	up	1
26	0.43			osd.26	up	1
27	0.43			osd.27	up	1
28	0.43			osd.28	up	1
29	0.43			osd.29	up	1
30	0.43			osd.30	up	1
31	0.43			osd.31	up	1
34	0.72			osd.34	up	1
35	0.72			osd.35	up	1

Are we misunderstanding the default behaviour? Any help you can provide
will be very much appreciated.

Regards,
Daniel

W: www.3ca.com.au
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Ceph Dev]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux