peering PGs

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Another kernel crash another invalid ceph state...

A memory allocation failure in the kernel (ixgbe) of one OSD-Server
lead to a domino effect in our ceph cluster with "0 up, 0 in".

When I restarted the cluster everything came up again. But I still
have 6 peering PGs:

pg v5898472: 3712 pgs: 3706 active+clean, 6 peering; 745 GB data, 775
GB used, 57642 GB / 59615 GB avail

# ceph pg dump -o - | grep peering
0.190   22      0       0       0       90112   92274688        200
 200     peering 6500'1256       7167'1063       [15,1]  [15,1]
6500'1256       2011-07-22 11:22:55.798745
3.18d   385     0       0       0       1529498 1566204928      300
 300     peering 7013'134376     7167'20162      [15,1]  [15,1]
6933'132427     2011-07-22 11:22:56.488471
0.4c    9       0       0       0       36864   37748736        200
 200     peering 6500'673        7163'1095       [12,1]  [12,1]
6500'673        2011-07-22 11:22:20.226119
3.49    171     0       0       0       671467  687580272       295
 295     peering 7013'10276      7163'2879       [12,1]  [12,1]
6933'9455       2011-07-22 11:22:20.701854
0.35e   6       0       0       0       24576   25165824        200
 200     peering 6500'628        7163'1142       [12,1]  [12,1]
6500'628        2011-07-22 11:22:19.267804
3.35b   198     0       0       0       791800  810803200       297
 297     peering 7013'66727      7163'5759       [12,1]  [12,1]
6933'65715      2011-07-22 11:22:20.035265


"ceph pg map" is consistent with "ceph pg dump":

# ceph pg map 0.190
2011-07-26 08:46:19.330981 mon <- [pg,map,0.190]
2011-07-26 08:46:19.331981 mon1 -> 'osdmap e7273 pg 0.190 (0.190) ->
up [15,1] acting [15,1]' (0)


But directorys of the PGs are present on multiple nodes (for example
on osd005 for 0.190):

/ceph/osd.001/current/0.190_head
/ceph/osd.001/snap_1650435/0.190_head
/ceph/osd.001/snap_1650445/0.190_head
/ceph/osd.005/current/0.190_head
/ceph/osd.005/snap_1572317/0.190_head
/ceph/osd.005/snap_1572323/0.190_head
/ceph/osd.015/current/0.190_head
/ceph/osd.015/snap_1467152/0.190_head

Any hint on how to proceed yould be great.

Thanks,
Christian
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html


[Index of Archives]     [CEPH Users]     [Ceph Large]     [Information on CEPH]     [Linux BTRFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]
  Powered by Linux