best common practice in case of degraded array with read errors

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 




Hello.

I have a 6 drive raid5. One of the drives failed on me (totally), and when I replaced it (-add a new working drive) I had several sectors on another drive give me UNC errors, which made md kick that drive as well, and left me with a non-working array (with only 4 drives).

What is the best common practice to handle this scenario? Right now I'm dd_rescue:ing the drive with read errors to a (hopefully) working drive, and then when I plan to --assemble --force the array to get 5 working drives (with a few zero:ed sectors where I guess I'll have corrupted files, hopefully no important metadata), and then I plan --add a 6th drive and have everything sync up and be back to "normal".

Is there a better way? I don't really understand why kicking drives out of the array when there aren't enough of them to keep going makes sense, is there some rationale I'm missing?

I've also heard recommendations to write to the bad sectors on the existing drive, but that scares me as well in case I write to the wrong place, which is why I went the dd_rescue route (I'm also hoping that it'll retry a bit more and might be able to read the bad blocks...)

--
Mikael Abrahamsson    email: swmike@xxxxxxxxx
--
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html

[Index of Archives]     [Linux RAID Wiki]     [ATA RAID]     [Linux SCSI Target Infrastructure]     [Linux Block]     [Linux IDE]     [Linux SCSI]     [Linux Hams]     [Device Mapper]     [Device Mapper Cryptographics]     [Kernel]     [Linux Admin]     [Linux Net]     [GFS]     [RPM]     [git]     [Yosemite Forum]


  Powered by Linux