On Sat, May 31, 2014 at 01:41:56AM -0700, H. Peter Anvin wrote: > On 05/30/2014 10:54 PM, Dave Chinner wrote: > > > > If we are changing the in-kernel timestamp to have a greater dynamic > > range that anything we current support on disk, then we need support > > for all filesystems for similar translation and constraint. The > > filesystems need to be able to tell the kernel what they timestamp > > range they support, and then the kernel needs to follow those > > guidelines. And if the filesystem is mounted on a kernel that > > doesn't support the current filesystem's timestamp format, then at > > minimum that filesystem cannot do anything that writes a > > timestamp.... > > > > Put simply: the filesystem defines the timestamp range that can be > > used safely, not the userspace API. If the filesystem can't support > > the date it is handed then that is an out-of-range error. Since > > when have we accepted that it's OK to handle out-of-range data with > > silent overflows or corruption of the data that we are attempting to > > store? We're defining a new API to support a wider date range - > > there is nothing that prevents us from saying ERANGE can be returned > > to a timestamp that the file cannot store correctly.... > > > > I'm still puzzled. > > Are you saying that you want a program that does: > > /* Deliberately simplified */ > gettimeofdayns(&now ...); > utimensat(... now); > > ... to suddenly start failing on Jan 19, 2038 (for a filesystem with > 32-bit timestamps), Yes. Hard fail so overflows are in your face and we know exactly what is going to cause silent timestamp screwups when the epoch > or would you propose some ways for the filesystems > in question to extend the range of the timestamps? Filesystems are going to have to change their on-disk formats, so we'd do that just like we do every other on-disk format change. With feature bits and translation layers, new ioctl structures, etc. Depending on the amount of work necessary, some filesystems could do this in 3.16, others it might be 3.20 before everything is sorted out across the kernel and userspace code... Either way, the hard fail problem goes away as each filesystem is converted. Further, if we have regression tests then new filesystems are guaranteed to be designed to handle 2038 epoch rollover, and so in a year of two this "hard fail" is effectively a non-problem. If someone breaks something in future, then we'll know about it pretty quickly. > What you seem to propose also seems to imply that on Jan 19, 2038 > anything that writes a timestamp with the current date (which logically > ends up being almost every write operation) would be dead and frozen on > such a filesystem -- pretty much meaning the filesystem would become > readonly if not in reality than in practice. Yup. If we can't do what the user wants without the user thinking corruption has occurred, then the only thing we are left with is "shut down the filesystem" error handling. Kind of like using BUG() rather than returning an error. That's why we need to be able to hard fail and return an error. However, we've got 20+ years to fix our current filesystems and all their support code to ensure this doesn't happen. In the mean time, having stuff hard fail is a great way to ensure that filesystems get fixed sooner rather than later... > I strongly suspect that that would be a more catastrophic failure than > incorrect timestamps, as you suddenly have all kinds of machines > embedded in $DEITY knows what places just stop and refuse to run. Yup, that's a great way of flushing out problems 20 years before they really matter. Cheers, Dave. -- Dave Chinner david@xxxxxxxxxxxxx -- To unsubscribe from this list: send the line "unsubscribe linux-fsdevel" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html