On Fri, March 21, 2008 5:02 am, Theodore Tso wrote: > On Thu, Mar 20, 2008 at 06:39:06PM +0100, Andre Noll wrote: >> On 12:35, Theodore Tso wrote: >> >> > If a mismatch is detected in a RAID-6 configuration, it should be >> > possible to figure out what should be fixed >> >> It can be figured out under the assumption that exactly one drive has >> bad data and all other ones have good data. But that seems to be an >> assumption that is hard to verify in reality. > > True, but it's what ECC memory does. :-) And most people agree that > it's a useful thing to do with memory. > > If you do ECC syndrome checking on every read, and follow that up with > periodic scrubbing so that you catch (and correct) errors quickly, it > is a reasonable assumption to make. My problem with this is that I don't have a good model for what might cause the error, so I cannot reason about what responses are justifiable. The analogy with ECC memory is, I think, poor. With ECC memory there are electro/physical processes which can cause a bit to change independently of any other bit with very low probability, so treating an ECC error as a single bit error is reasonable. The analogy with a disk drive would be a media error. However disk drives record CRC (or similar) checks so that media errors get reported as errors, not as incorrect data. So the analogy doesn't hold. Where else could the error come from? Presumably a bit-flip on some transfer bus between main memory and the media. There are several of these busses (mem to controller, controller to device, internal to device). The corruption could happen on the write or on the read. When you write to a RAID6 you often write several blocks to different devices at the same time. Are these really likely to be independent events wrt whatever is causing the corruption? I don't know. But without a clear model, it isn't clear to me that any particular action will be certain to improve the situation in all cases. And how often does silent corruption happen on modern hard drives? How often do you write something and later successfully read something else when it isn't due to a major hardware problem that is causing much more that just occasional errors? The ZFS people seem to say that their checksumming of all data shows up a lot of these cases. If that is true, how come people who don't use ZFS aren't reporting lots of data corruption? So yes: there are lots of things that *could* be done. But without a model for the "threat", an analysis of how the remedy would actually affect every different possible scenario, and some idea of the probability of the remedy being needed, it is very hard to justify a change of this sort. And there are plenty of other things to be coded that are genuinely useful - like converting a RAID5 to a RAID6 while online... NeilBrown -- To unsubscribe from this list: send the line "unsubscribe linux-raid" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html