tons of "failed lossy con, dropping message" => root cause for bad performance ?

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hello All,

I have a simple test setup with 2 osd servers each 3 NICs (1Gb each):

* One for management (ssh and such)
* One for the public network (connected to ceph clients)
* One for the cluster (osd inter-connection)

I keep seeing this messages:


Aug 26 18:43:31 ceph01 ceph-osd: 2013-08-26 18:43:31.040038 7f1afe5b6700 0 -- 192.168.113.115:6801/14629 submit_message osd_op_reply(88713 rb.0.1133.74b0dc51.0000000003cd [write 2297856~4096] ondisk = 0) v4 remote, 192.168.113.1:0/607109564, failed lossy con, dropping message 0xaf83680 Aug 26 18:43:32 ceph01 ceph-osd: 2013-08-26 18:43:32.578875 7f1afe5b6700 0 -- 192.168.113.115:6801/14629 submit_message osd_op_reply(88870 rb.0.1133.74b0dc51.000000000345 [write 3145728~524288] ondisk = 0) v4 remote, 192.168.113.1:0/607109564, failed lossy con, dropping message 0x5c2e1e0

And also:
Aug 26 18:27:08 ceph01 ceph-osd: 2013-08-26 18:27:08.211604 7f3cf738f700 0 bad crc in data 1545773059 != exp 878537506 Aug 26 18:27:08 ceph01 ceph-osd: 2013-08-26 18:27:08.225121 7f3cf738f700 0 bad crc in data 1929463652 != exp 2083940607

Any idea on the problem ?

Matthieu.
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com




[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux