I just tried again with kernel 3.8.1 and it exhibits the same problem. The memory utilization seems fine. This was taken from the node doing the cp (and the one the OSDs go down on): [root@node-172-17-0-13 ~]# free total used free shared buffers cached Mem: 16447312 15430748 1016564 0 3028 13227728 -/+ buffers/cache: 2199992 14247320 Swap: 0 0 0 I didn't see anything that looked particularly useful in the kernel messages. We do have one flaky disk in this cluster (intentionally left there to make sure our system can cope with it), however that disk is not one of the ones getting marked down in this case. It seems like it's only the ones on the machine where the I/O is heavy due to the cp operation. Also, this time I noticed some cluster services timing out before the OSDs were actually marked down. On Thu, Feb 28, 2013 at 1:55 PM, Gregory Farnum <greg@xxxxxxxxxxx> wrote: > I haven't checked the logs at all, but my first question is what the > memory utilization on that node looked like during that time, and if > there are any dmesg warnings. > (Other than that I should leave this to probably Alex and Sam.) > -Greg > > On Thu, Feb 28, 2013 at 1:44 PM, Nick Bartos <nick@xxxxxxxxxxxxxxx> wrote: >> When a single high I/O event (in this case a cp of a 10G file on a >> filesystem mounted on an rbd) occurs, I'm having the 2 OSDs that >> reside on the same system where the rbd is mounted being marked down >> when it appears that they shouldn't be. Additionally, other cluster >> services start timing out just after the OSDs are marked down (which >> appears to be mysql which is rbd backed becoming unresponsive). >> Previous to the OSDs being marked down things are running slow, but do >> not appear to actually be failing until they are actually marked down. >> >> Here is an interesting snip in the logs: >> >> Feb 28 21:12:11 172.17.0.13 ceph-mon: 2013-02-28 21:12:11.081003 >> 7f377687d700 1 mon.0@0(leader).osd e14 we have enough >> reports/reporters to mark osd.2 down >> Feb 28 21:12:11 172.17.0.13 [ 663.241832] libceph: osd2 down >> Feb 28 21:12:11 172.17.0.14 [ 655.577185] libceph: osd2 down >> Feb 28 21:12:11 172.17.0.13 [ 663.242064] libceph: osd5 down >> Feb 28 21:12:11 172.17.0.13 kernel: [ 663.241832] libceph: osd2 down >> Feb 28 21:12:11 172.17.0.13 kernel: [ 663.242064] libceph: osd5 down >> Feb 28 21:12:11 172.17.0.14 [ 655.577434] libceph: osd5 down >> Feb 28 21:12:11 172.17.0.14 kernel: [ 655.577185] libceph: osd2 down >> Feb 28 21:12:11 172.17.0.14 kernel: [ 655.577434] libceph: osd5 down >> Feb 28 21:12:12 172.17.0.13 ceph-osd: 2013-02-28 21:12:12.423178 osd.5 >> 172.17.0.13:6803/2015 126 : [WRN] map e16 wrongly marked me down >> Feb 28 21:12:12 172.17.0.13 ceph-osd: 2013-02-28 21:12:12.423177 >> 7f4c10a0e700 0 log [WRN] : map e16 wrongly marked me down >> Feb 28 21:12:17 172.17.0.13 ceph-osd: 2013-02-28 21:12:17.208466 >> 7f01aa894700 0 log [WRN] : map e16 wrongly marked me down >> Feb 28 21:12:17 172.17.0.13 ceph-osd: 2013-02-28 21:12:17.208468 osd.2 >> 172.17.0.13:6800/1924 187 : [WRN] map e16 wrongly marked me down >> >> The full log is available here: http://download.pistoncloud.com/p/ceph-2.log.xz >> Note: The compressed log is only about 8MB, but uncompressed it's >> about 160MB. I've added libceph and rbd kernel debugging in as well. >> >> Here's a brief run-down of the configuration: >> ceph 0.56.3 >> kernel 3.5.7 with 116 patches I got from Alex Elder >> 4 x nodes with 2 x OSDs per node (each is a 120GB SSD, using xfs) >> 1gbit networking between nodes >> rbds mapped on all nodes running OSDs, and xfs filesystems running on >> top of them >> >> I'm not sure if it would help, but are there any updated ceph patches >> for the 3.5.x kernel? >> _______________________________________________ >> ceph-users mailing list >> ceph-users@xxxxxxxxxxxxxx >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com _______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com