FW: Multiple Disk Failure Recovery

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



One could remove the spare drive from the system.  Than do the mdadm
--assemble --force to get it start and keep it from trying to
resync/recover.

Once you get the array up and 'limping' carefully pick the most important
stuff and copy it off the array and hope the bad sectors don't did not
affect that data.  As you mentioned, you have bad sectors.  If you try to do
it all or even as you pick the important stuff (I know it is all important
that why it was on the RAID), you will eventually hit data that has bad
sectors and mdadm will fail the effected drive and deactivate the array.  At
that point accept that the data in that area is most likely gone. Than do
the mdadm --assemble --force to get it started again and move on to the next
areas of most important data.  Could be a long cycle...

Aside from that, I am curious, was your spare disk shared by another array.
If it was not than I would recommend you don't do a RAID5 with hot spare
next time, and do a RAID6.  But this is my person feeling and you can take
it or leave it.

I too at one point did a RAID5 with hot spare and I was using eight drives.
So yes you can have two drives fall as long as the delta between failures is
long enough to allow the raid to resync the spare in and during the process
there are no be unknown bad sectors on the remaining drives.  But I got to
thinking, if I am going to be spinning/powering that "hot spare" I may as
well as do a RAID6.  As long as the hot spare is not share with other
arrays, I see no downside and it would protect you in the future from this
problem.

-----Original Message-----
From: linux-raid-owner@xxxxxxxxxxxxxxx
[mailto:linux-raid-owner@xxxxxxxxxxxxxxx] On Behalf Of Lane Brooks
Sent: Saturday, October 14, 2006 9:29 PM
To: linux-raid@xxxxxxxxxxxxxxx
Subject: Multiple Disk Failure Recovery

I have a RAID 5 setup with one spare disk.  One of the disks failed, so 
I replaced it, and while it was recovering, it found some bad sectors on 
another drive (unreadable and uncorrectable SMART errors).  This 
generated a fail event and shut down the array.

When I restart the array with the force command, it starts the recovery 
process and dies when it gets to the bad sectors, so I am stuck in an 
infinite loop.

I am wondering if there is a way to cut my losses with these bad sectors 
and have it recover what it can so that I can get my raid array back to 
functioning.  Right now I cannot get a spare disk recovery to finish 
because these bad sectors.  Is there a way to force as much recovery as 
possible so that I can replace this newly faulty drive?

Thanks
Lane
-
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html



-
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html

[Index of Archives]     [Linux RAID Wiki]     [ATA RAID]     [Linux SCSI Target Infrastructure]     [Linux Block]     [Linux IDE]     [Linux SCSI]     [Linux Hams]     [Device Mapper]     [Device Mapper Cryptographics]     [Kernel]     [Linux Admin]     [Linux Net]     [GFS]     [RPM]     [git]     [Yosemite Forum]


  Powered by Linux