> For clarity, most ATA class disk drives are spec'ed to have one > non-recoverable error per 150TB or so of writes. Disk drives do blind > writes. (ie. They are not verified). So we should all expect to have > the occasional silent data corruption on write. The problem is > compounded with bad cables, controllers, RAM, etc. Cable errors should only be a PATA issue, SATA protects the command block and the data. > The only way for the linux kernel even attempt to fix that is for it > to do a read verify on everything it writes. For the vast majority of > uses that is just not acceptable for performance reasons. It's also the wrong layer > OTOH, if data integrity is of the utmost for you, then you should > maintain a md5hash or similar for your critical files and verify them > any time you make a copy. btrfs may offer a auto read-verify. I > don't know much about btrfs. If you deal with utterly enormous amounts of data (as some clusters and the like do) you protect your data from application to application layer. It leaves ECC protected memory and it comes back to ECC protected memory with a hash. That covers a lot of the errors by the OS, hardware, cables, busses - even drive memory. -- To unsubscribe, send a message with 'unsubscribe linux-mm' in the body to majordomo@xxxxxxxxxx For more info on Linux MM, see: http://www.linux-mm.org/ . Don't email: <a href=mailto:"dont@xxxxxxxxx"> email@xxxxxxxxx </a>