Re: Repeated messages of "heartbeat_check: no heartbeat from"

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi,

I'm seeing this behavior again. In the monitor messages this started to
appear:

2011-10-14 16:22:19.853752   log 2011-10-14 16:22:14.887881 mon.0
[2a00:f10:113:1:230:48ff:fed3:b086]:6789/0 3281 : [INF] osd.12
[2a00:f10:113:1:225:90ff:fe33:49b0]:6800/8023 failed (by osd.0
[2a00:f10:113:1:225:90ff:fe33:49fe]:6800/10678)

So, osd.0 is telling us that osd.12 is down, but the proces is running
and seems alive.

The logs are saying:

Oct 14 16:20:52 atom0 osd.0[10679]: 7f81deeac700 osd.0 511
heartbeat_check: no heartbeat from osd.12 since 2011-10-14
16:00:07.636627 (cutoff 2011-10-14 16:20:32.025154)
Oct 14 16:20:52 atom0 osd.0[10679]: 7f81d2d93700 osd.0 511
heartbeat_check: no heartbeat from osd.12 since 2011-10-14
16:00:07.636627 (cutoff 2011-10-14 16:20:32.108369)
Oct 14 16:20:53 atom0 osd.0[10679]: 7f81deeac700 osd.0 511
heartbeat_check: no heartbeat from osd.12 since 2011-10-14
16:00:07.636627 (cutoff 2011-10-14 16:20:33.025462)

While osd.12 reports getting messages from osd.0

Oct 14 16:20:54 atom3 osd.12[8024]: 7fa3621bb700 --
[2a00:f10:113:1:225:90ff:fe33:49b0]:0/8024 <== osd.0
[2a00:f10:113:1:225:90ff:fe33:49fe]:6813/10678 1317 ====
osd_ping(heartbeat e511 as_of 355) v1 ==== 61+0+0 (2126910847 0 0)
0x1e281c0 con 0x1f02c80

I see osd.0 sending pg_info messages:

Oct 14 16:20:58 atom0 osd.0[10679]: 7f81d769d700 --
[2a00:f10:113:1:225:90ff:fe33:49fe]:6812/10678 --> osd.12
[2a00:f10:113:1:225:90ff:fe33:49b0]:6801/8023 -- pg_info(1 pgs e512) v1
-- ?+0 0x3bf1e00

And osd.12 is also sending stuff back:

Oct 14 16:21:06 atom3 osd.12[8024]: 7fa3629bc700 --
[2a00:f10:113:1:225:90ff:fe33:49b0]:6801/8023 --> osd.0
[2a00:f10:113:1:225:90ff:fe33:49fe]:6812/10678 -- pg_info(1 pgs e518) v1
-- ?+0 0x6f3e540

I'm running the latest version, but I did do a complete powercycle of my
cluster this morning. It's after the powercycle these messages started
appearing.

Any ideas?

Wido

On Wed, 2011-08-10 at 12:30 +0200, Wido den Hollander wrote:
> Hi,
> 
> I tried reproducing this today with the new heartbeat code, couldn't.
> 
> The new heartbeat code seems to work much better under load, I haven't
> seen any bouncing like I saw before.
> 
> Wido
> 
> On Fri, 2011-08-05 at 14:11 +0200, Wido den Hollander wrote:
> > Hi,
> > 
> > I'm seeing something weird which I don't want to report in a issue yet
> > in the tracker.
> > 
> > I suffered a pretty big crash which caused me to loose 15 of my 40 OSDs.
> > Some died because my WD Green 2TB disks failed, others since btrfs
> > didn't survive the hard reboot I had to give them.
> > 
> > When watching 'ceph -w' I noticed this:
> > 
> > 2011-08-05 13:32:03.209958    pg v504313: 8008 pgs: 1 inactive, 334
> > active+clean, 10 peering, 1 crashed+peering, 14 down+peering, 7 active
> > +clean+degraded, 7624 crashed+down+peering, 8 down+degraded+peering, 9
> > crashed+down+degraded+peering; 4417 GB data, 13121 GB used, 32504 GB /
> > 46575 GB avail; 8255/3391107 degraded (0.243%)
> > 2011-08-05 13:32:03.225712   mds e6033: 1/1/1 up {0=dga=up:replay(laggy
> > or crashed)}
> > 2011-08-05 13:32:03.225747   osd e9934: 40 osds: 25 up, 25 in
> > 2011-08-05 13:32:03.225874   log 2011-08-05 13:32:01.646596 mon0
> > [2a00:f10:113:1:230:48ff:fed3:b086]:6789/0 29796 : [INF] osd37
> > [2a00:f10:113:1:225:90ff:fe33:49a4]:6803/17881 failed (by osd7
> > [2a00:f10:113:1:225:90ff:fe32:cf64]:6806/2746)
> > 2011-08-05 13:32:03.225964   mon e1: 1 mons at
> > {dga=[2a00:f10:113:1:230:48ff:fed3:b086]:6789/0}
> > 2011-08-05 13:32:03.929650   log 2011-08-05 13:32:03.603681 mon0
> > [2a00:f10:113:1:230:48ff:fed3:b086]:6789/0 29797 : [INF] osd19
> > [2a00:f10:113:1:225:90ff:fe33:49cc]:6806/2740 failed (by osd5
> > [2a00:f10:113:1:225:90ff:fe32:cf64]:6800/2506)
> > 
> > Here you see that osd5 is reporting that osd37 is down and osd7 is
> > reporting that osd19 is down.
> > 
> > These messages keep being repeated, over and over. My 'debug ms' is at
> > 1, so I took a look at the logs:
> > 
> > Aug  5 13:32:41 atom1 osd.5[2507]: 7f75227fc700 osd5 9963
> > heartbeat_check: no heartbeat from osd19 since 2011-08-04
> > 17:01:02.504942 (cutoff 2011-08-05 13:32:21.916364)
> > Aug  5 13:32:42 atom1 osd.5[2507]: 7f7531579700 osd5 9963
> > heartbeat_check: no heartbeat from osd19 since 2011-08-04
> > 17:01:02.504942 (cutoff 2011-08-05 13:32:22.381632)
> > Aug  5 13:32:43 atom1 osd.5[2507]: 7f7531579700 osd5 9965
> > heartbeat_check: no heartbeat from osd19 since 2011-08-04
> > 17:01:02.504942 (cutoff 2011-08-05 13:32:23.474198)
> > Aug  5 13:32:44 atom1 osd.5[2507]: 7f75227fc700 osd5 9965
> > heartbeat_check: no heartbeat from osd19 since 2011-08-04
> > 17:01:02.504942 (cutoff 2011-08-05 13:32:24.027161)
> > 
> > osd5 keeps claiming that osd19 is down, but when I check the logs I see
> > that osd19 is running and exchanging heartbeat messages with the other
> > OSDs in the cluster.
> > 
> > What seems odd to me is that there are no log messages about osd7
> > claiming that osd37 is down, although it keeps being repeated by 'ceph
> > -w'.
> > 
> > $ ceph osd injectargs 7 '--debug-ms 1'
> > $ ceph osd injectargs 37 '--debug-ms 1'
> > 
> > I made sure that was injected, but still, no log messages.
> > 
> > Going through the logs a bit further (after setting the logging to 20) I
> > saw osd.19 reporting:
> > 
> > Aug  5 13:48:23 atom4 osd.19[2741]: 7f634ca8e700 --
> > [2a00:f10:113:1:225:90ff:fe33:49cc]:6810/2740 >>
> > [2a00:f10:113:1:225:90ff:fe32:cf64]:6808/2506 pipe(0x7f64e4193ec0 sd=56
> > pgs=1530 cs=1 l=0).reader got envelope type=70 src osd5 front=61 data=0
> > off 0
> > 
> > It seems that it is communicating with osd5, but osd5 does not seem to
> > communicate with osd19?
> > 
> > The log also confirms that:
> > 
> > Aug  5 14:10:05 atom1 osd.5[2507]: 7f7531579700 --
> > [2a00:f10:113:1:225:90ff:fe32:cf64]:6800/2506 -->
> > [2a00:f10:113:1:230:48ff:fed3:b086]:6789/0 -- osd_failure(osd19
> > [2a00:f10:113:1:225:90ff:fe33:49cc]:6806/2740 e10020 v10020) v1 -- ?+0
> > 0x6948000 con 0x1cd7710
> > Aug  5 14:10:05 atom1 osd.5[2507]: 7f7531579700 --
> > [2a00:f10:113:1:225:90ff:fe32:cf64]:6800/2506 submit_message
> > osd_failure(osd19 [2a00:f10:113:1:225:90ff:fe33:49cc]:6806/2740 e10020
> > v10020) v1 remote [2a00:f10:113:1:230:48ff:fed3:b086]:6789/0
> > Aug  5 14:10:05 atom1 osd.5[2507]: 7f75365a1700 --
> > [2a00:f10:113:1:225:90ff:fe32:cf64]:6800/2506 >>
> > [2a00:f10:113:1:230:48ff:fed3:b086]:6789/0 pipe(0x1cd74a0 sd=13 pgs=178
> > cs=1 l=1).writer encoding 17093 0x6948000 osd_failure(osd19
> > [2a00:f10:113:1:225:90ff:fe33:49cc]:6806/2740 e10020 v10020) v1
> > 
> > Now, to prevent (I tried...) my cluster from bouncing around I placed
> > this in the OSD section:
> > 
> >     osd min down reporters = 3
> >     osd min down reports = 5
> > 
> > Since only osd5 and osd7 are sending out these messages osd19 nor osd37
> > get marked down, but what is going on here?
> > 
> > I haven't restarted the OSDs yet.
> > 
> > I noticed that the OSD heartbeat process is being revised currently, but
> > I thought it might be helpful to report this.
> > 
> > Wido
> > 
> > --
> > To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
> > the body of a message to majordomo@xxxxxxxxxxxxxxx
> > More majordomo info at  http://vger.kernel.org/majordomo-info.html
> 
> 
> --
> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
> the body of a message to majordomo@xxxxxxxxxxxxxxx
> More majordomo info at  http://vger.kernel.org/majordomo-info.html


--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html


[Index of Archives]     [CEPH Users]     [Ceph Large]     [Information on CEPH]     [Linux BTRFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]
  Powered by Linux