Re: "hit suicide timeout" message after upgrade to 0.56

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 





On 01/03/2013 10:05 PM, Sage Weil wrote:
On Thu, 3 Jan 2013, Wido den Hollander wrote:
Hi,

On 01/03/2013 05:52 PM, Sage Weil wrote:
Hi Wido,

On Thu, 3 Jan 2013, Wido den Hollander wrote:
Hi,

I updated my 10 node 40 OSD cluster from 0.48 to 0.56 yesterday evening
and
found out this morning that I had 23 OSDs still up and in.

Investigating some logs I found these messages:

This sounds quite a bit #3714.  You might give wip-3714 a try...


I'll give it a try and see if I can reproduce. Probably just have to kill the
whole cluster and restart everything to see if it behaves the same way.

I'll give it a try and let you know.

You mean restart every ceph-osd with the updated code?  I hope/suspect
that will do the trick.


I need to restart for the new code indeed. But to trigger the heavy CPU usage I need to restart all the OSDs at once to make sure they are all very busy with recovery.

Not restart them with 5 minute intervals between each OSD.

That's my assumption.

Wido

BTW, these patches are now in 'testing'.. I'd run that code.

Thanks!
sage


Wido

sage



*********************************************************************
      -8> 2013-01-02 21:13:40.528936 7f9eb177a700  1 heartbeat_map
is_healthy
'OSD::op_tp thread 0x7f9ea2f5d700' had timed out after 30
      -7> 2013-01-02 21:13:40.528985 7f9eb177a700  1 heartbeat_map
is_healthy
'OSD::op_tp thread 0x7f9ea375e700' had timed out after 30
      -6> 2013-01-02 21:13:41.311088 7f9eaff77700 10 monclient:
_send_mon_message to mon.pri at
[2a00:f10:11b:cef0:230:48ff:fed3:b086]:6789/0
      -5> 2013-01-02 21:13:45.047220 7f9e92282700  0 --
[2a00:f10:11b:cef0:225:90ff:fe32:cf64]:0/2882 >>
[2a00:f10:11b:cef0:225:90ff:fe33:49fe]:6805/2373 pipe(0x9d7ad80 sd=135 :0
pgs=0 cs=0 l=1).fault
      -4> 2013-01-02 21:13:45.049225 7f9e962c2700  0 --
[2a00:f10:11b:cef0:225:90ff:fe32:cf64]:6801/2882 >>
[2a00:f10:11b:cef0:225:90ff:fe33:49fe]:6804/2373 pipe(0x9d99000 sd=104
:44363
pgs=99 cs=1 l=0).fault with nothing to send, going to standby
      -3> 2013-01-02 21:13:45.529075 7f9eb177a700  1 heartbeat_map
is_healthy
'OSD::op_tp thread 0x7f9ea2f5d700' had timed out after 30
      -2> 2013-01-02 21:13:45.529115 7f9eb177a700  1 heartbeat_map
is_healthy
'OSD::op_tp thread 0x7f9ea2f5d700' had suicide timed out after 300
      -1> 2013-01-02 21:13:45.531952 7f9eb177a700 -1
common/HeartbeatMap.cc: In
function 'bool ceph::HeartbeatMap::_check(ceph::heartbeat_handle_d*, const
char*, time_t)' thread 7f9eb177a700 time 2013-01-02 21:13:45.529176
common/HeartbeatMap.cc: 78: FAILED assert(0 == "hit suicide timeout")

   ceph version 0.56 (1a32f0a0b42f169a7b55ed48ec3208f6d4edc1e8)
   1: (ceph::HeartbeatMap::_check(ceph::heartbeat_handle_d*, char const*,
long)+0x107) [0x796877]
   2: (ceph::HeartbeatMap::is_healthy()+0x87) [0x797207]
   3: (ceph::HeartbeatMap::check_touch_file()+0x23) [0x797453]
   4: (CephContextServiceThread::entry()+0x55) [0x8338d5]
   5: (()+0x7e9a) [0x7f9eb4571e9a]
   6: (clone()+0x6d) [0x7f9eb2ff5cbd]
   NOTE: a copy of the executable, or `objdump -rdS <executable>` is needed
to
interpret this.

       0> 2013-01-02 21:13:46.314478 7f9eaff77700 10 monclient:
_send_mon_message to mon.pri at
[2a00:f10:11b:cef0:230:48ff:fed3:b086]:6789/0
*********************************************************************

Reading these messages I'm trying to figure out why those messages came
along.

Am I understanding this correctly that the heartbeat updates didn't come
along
in time and the OSDs committed suicide?

I read the code in common/HeartbeatMap.cc and it seems like that.

During the restart of the cluster the Atom CPUs were very busy, so could
it be
that the CPUs were just to busy and the OSDs weren't responding to
heartbeats
in time?

In total 16 of the 17 crashed OSDs are down with these log messages.

I'm now starting the 16 crashed OSDs one by one and that seems to go just
fine.

I've set "osd recovery max active = 1" to prevent overloading the CPUs to
much
since I know Atoms are not that powerful. I'm just still trying to get it
all
working on them :)

Am I right this is probably a lack of CPU power during the heavy recovery
which causes them to not respond to heartbeat updates in time?

Wido
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html


--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html

--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html


--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html

--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html


[Index of Archives]     [CEPH Users]     [Ceph Large]     [Information on CEPH]     [Linux BTRFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]
  Powered by Linux