Re: Regarding loss of heartbeats

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 




> -----Original Message-----
> From: ceph-users [mailto:ceph-users-bounces@xxxxxxxxxxxxxx] On Behalf Of Trygve Vea
> Sent: 29 November 2016 14:07
> To: ceph-users <ceph-users@xxxxxxxx>
> Subject:  Regarding loss of heartbeats
> 
> Since Jewel, we've seen quite a bit of funky behaviour in Ceph.  I've written about it a few times to the mailing list.
> 
> Higher CPU utilization after the upgrade / Loss of heartbeats.  We've looked at our network setup, and we've optimized some
> potential bottlenecks some places.
> 
> Interesting thing regarding loss of heartbeats.  We have observed OSDs running on the same host losing heartbeats against
> eachother.  I'm not sure why they are connected at all (we have had some remapped/degraded placement groups over the weekend,
> maybe that's why) - but I have a hard time pointing the finger at our network when the heartbeat is lost between two osds on the
> same server.
> 
> 
> I've been staring myself blind at this problem for a while, and just now noticed a pretty new bug report that I want to believe is
related
> to what I am experiencing: http://tracker.ceph.com/issues/18042
> 
> We had one OSD hit a suicide timeout value and kill itself off last night, and one can see that several of these heartbeats are
between
> osds on the same node.  (zgrep '10.22.9.21.*10.22.9.21' ceph-osd.2.gz)
> 
> http://employee.tv.situla.bitbit.net/ceph-osd.2.gz
> 
> 
> Does anyone have any thoughts about this?  Are we stumbling on a known, or unknown bug in Ceph?

Hi Trygve,

I was getting similar things to you after upgrading to 10.2.3, definitely seeing problems where OSD's on the same nodes were marking
each other out and the cluster was fairly idle. I found that it seemed to being caused by Kernel 4.7, nodes in the same cluster that
were on 4.4 were unaffected. After downgrading all nodes to 4.4, everything has been really stable for me.

Nick

> 
> 
> Regards
> --
> Trygve Vea
> _______________________________________________
> ceph-users mailing list
> ceph-users@xxxxxxxxxxxxxx
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

_______________________________________________
ceph-users mailing list
ceph-users@xxxxxxxxxxxxxx
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com



[Index of Archives]     [Information on CEPH]     [Linux Filesystem Development]     [Ceph Development]     [Ceph Large]     [Linux USB Development]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]     [xfs]


  Powered by Linux