Re: OSD crash after Upgrade to Jewel

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Adding some information here.

While most of the stuff is in the ticket, just as a reference.

The crash here is happening because the OSDs have a too small journal. After looking further these systems were by accident deployed with a 100M journal. Seems that somebody made a typo. Not all OSDs (864) had this 100M journal, 45 in total.

The OSDs were far behind on the OSDMaps, so I increased 'osd_map_message_max' on the MONs (!) to 5k so it would share more maps.

The OSDs were about 15k maps behind and wouldn't catch up, they were just stuck (that seems to be the true bug).

The 5k maps were to much for the 100M journal. So I flushed the journal and created a new journal which was 8G in size.

osd_map_message_max was decreased to 1k and that made the OSDs advance 1k maps at a time. However, I had to keep restarting them to make them catch up.

You can see this with:

$ ceph daemon osd.X status

It is stuck in 'prebooting' and stays there. You have to restart it again, it will advance 1k maps and stay there. Restart, etc. Painfull and time-consuming task.

This bug seems to be present in both Hammer and Jewel and is in the tracker: http://tracker.ceph.com/issues/16982

Wido

> Op 10 augustus 2016 om 19:52 schreef Wido den Hollander <wido@xxxxxxxx>:
> 
> 
> Hi,
> 
> I'm seeing this crash on multiple OSDs while upgrading to Jewel:
> 
>      0> 2016-08-10 19:41:40.686944 7f20d37f1700 -1 common/buffer.cc: In function 'void ceph::buffer::ptr::copy_in(unsigned int, unsigned int, const char*, bool)' thread 7f20d37f1700 time 2016-08-10 19:41:40.682052
> common/buffer.cc: 977: FAILED assert(o+l <= _len)
> 
> It started to happen when the OSDs came under load due to a node failing (hardware) which brought a lot of PGs in a unclean state.
> 
> I already filed a tracker issue for it: http://tracker.ceph.com/issues/16982
> 
> All machines are running Ubuntu 14.04 with the 4.4 kernel.
> 
> Wondering if somebody has seen this before. This cluster is very down at the moment..
> 
> 
> Wido
> --
> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
> the body of a message to majordomo@xxxxxxxxxxxxxxx
> More majordomo info at  http://vger.kernel.org/majordomo-info.html
--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html



[Index of Archives]     [CEPH Users]     [Ceph Large]     [Information on CEPH]     [Linux BTRFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]
  Powered by Linux