On 9 May 2017, NeilBrown outgrape: > On Tue, May 09 2017, Nix wrote: >> Neil decided not to do any repair work in this case on the grounds that >> if the drive is misdirecting one write it might misdirect the repair as >> well > > My justification was a bit broader than that. I noticed your trailing comment on the blog post only after sending all these emails out :( bah! > If you get a consistency error on RAID6, there is not one model to > explain it which is significantly more likely than any other model. Yeah, I'm quite satisfied with "we don't have enough data to know if repairing is safe" as reasoning: among other things it suggests that mismatches are really rare, which is reassuring! This certainly suggests that repairing should be, at the very least, off by default, and I'm not terribly unhappy for it to not exist. ... but I do want to at least report the location of stripes that fail checks, as in my earlier ugly patch. That's useful for any array with >1 partition or LVM LV on it. ("Oh, that mismatch is harmless, it's in swap. That one is in small_but_crucial_lv, I'll restore it from backup, without affecting the massive_messy_lv which had no mismatches and would take weeks to restore.") (As far as I'm concerned, if you don't *have* a backup of some fs, you deserve what's coming to you! Good backups are easy and with md you can even make them as resilient as the main RAID arrays. I'm interested in maximizing availability here: having to take a big array with many LVs down for ages for a restore because you don't know which bit is corrupted just seems *wrong*.) -- To unsubscribe from this list: send the line "unsubscribe linux-raid" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html