Re: failed lossy con, dropping message

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi Alex,

I saw your thread, but I think mine is a little bit different.
I have only one message so far, and I want to better understand the issue. I would like to see whether there are any tunable parameters that could be adjusted to have an influence on this behavior.

Kind regards,
Laszlo

On 12.04.2017 22:19, Alex Gorbachev wrote:
Hi Laszlo,

On Wed, Apr 12, 2017 at 6:26 AM Laszlo Budai <laszlo@xxxxxxxxxxxxxxxx <mailto:laszlo@xxxxxxxxxxxxxxxx>> wrote:

    Hello,

    yesterday one of our compute nodes has recorded the following message for one of the ceph connections:

    submit_message osd_op(client.28817736.0:690186 rbd_data.15c046b11ab57b7.00000000000000c4 [read 2097152~380928] 3.6f81364a ack+read+known_if_redirected e3617) v5 remote, 10.12.68.71:6818/6623 <http://10.12.68.71:6818/6623>, failed lossy con, dropping message

    Can someone "decode" the above message, or direct me to some document where I could read more about it?

    We have ceph 0.94.10.


I am researching the same issue, but on 11 OSD nodes.  You should be able to find my thread in this list.  Looks like this could be a bug in kernels 4.4+, or a network issue.

Regards,
Alex




    Thank you,
    Laszlo
    _______________________________________________
    ceph-users mailing list
    ceph-users@xxxxxxxxxxxxxx <mailto:ceph-users@xxxxxxxxxxxxxx>
    http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

--
--
Alex Gorbachev
Storcium
_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux