Re: Suicide

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Friday, April 15, 2011 at 4:29 PM, Zenon Panoussis wrote:

> On 04/16/2011 01:06 AM, Gregory Farnum wrote:
> 
> > Hmm, that timeline doesn't quite make sense -- node01 takes over the MDS 
> > duties at 4:33 and crashes, but then it starts up again at 4:50. But it's 
> > possible that node02 took over in the interval there and we just don't see 
> > it because the log disk was full (I had erroneously thought that a filled 
> > disk would hang the daemon but that turns out not to be the case). So I'd 
> > guess you shut everything down sometime after 5:08, and that would make 
> > sense.
> 
> Indeed, you're probably right.
> 
> > Unfortunately what we're really interested in is what caused the assert 
> > failure on node01 at 4:35 and the reasons for that aren't available in the 
> > logs we have. :(
> 
> > This is the second time we've seen that assert but we've not been able to 
> > reproduce it or figure out how the invariant that it's checking against got 
> > broken. If you like we can come up with a hacky fix that should let your 
> > cluster come back up, but it's possible that you'd lose some data and this 
> > is a very rare condition so if it's not a big deal I'd just re-create your 
> > cluster.
> 
> My data has been safe elsewhere all along and I have already re-created the
> cluster. In other words I don't need the hacky fix, but someone else might
> be desperate for it in the future, so creating it could be a good idea anyway.

Well, we'd like to develop a proper fix in the form of our fsck tools. That, however, will take some time and a hacky fix isn't really safe so we don't want to put it out there and have people use it while thinking it's safe. ;) 


> However, the cause of the corruption is still an open issue that ought to be
> understood and solved. The most likely place to be able to reproduce it at is
> right here, so if you think it's useful, I'm willing to try to crash it again.
> If you want me to, let's make a plan for it. These are just test boxes and
> I have no problem even giving you root on them, if that can help pinpoint
> the cause of the corruption.
We'd love to reproduce it and track it down! Unfortunately, the two times we've seen it so far (you're the second one) are on external users who have very sparse logging. :(

If you want to try to reproduce it for us in a useful fashion we'd love that. You'll need to add debug output to your MDS config. At a minimum we will need "debug journaler = 20". You should also add "debug ms = 1" and probably "debug mds = 10". Be warned that this will use a LOT of disk space, though. If you ran out before you're going to do so again and we will really need the logs that generated the journal and the logs that were replaying the journal to figure out what happened, so you'll need to come up with some way of handling them (writing to a big NFS disk -- though that'll impact networking, different disk, log rotation, etc). Then try and reproduce your previous conditions as exactly as possible and see if you run into that assert again.

Thanks!
-Greg



--
To unsubscribe from this list: send the line "unsubscribe ceph-devel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html


[Index of Archives]     [CEPH Users]     [Ceph Large]     [Information on CEPH]     [Linux BTRFS]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]

  Powered by Linux