RAID becomes un-bootable after failure

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



I have a 2-disk RAID-1 that I set up using the Red Hat 9 installer. My problem is that pretty much any simulated drive failure makes the system unbootable. If I power down the machine and remove drive 1, the system can't find any bootable device. If I set a drive faulty using the RAIDtools, re-add it, and let the recovery process run the boot loader seems to get overwritten and I end up with a system hung at "GRUB loading stage2". Can anyone shed some light on what's going on here? My guess is that GRUB isn't installed on drive 2, so that removing drive 1 or recovering drive 1 from drive 2 leads to no boot loader, but shouldn't GRUB have been copied to drive 2 in the mirroring process? How can I configure my system so that it will still be bootable after a drive failure?

Thanks for your help.
-
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html

[Index of Archives]     [Linux RAID Wiki]     [ATA RAID]     [Linux SCSI Target Infrastructure]     [Linux Block]     [Linux IDE]     [Linux SCSI]     [Linux Hams]     [Device Mapper]     [Device Mapper Cryptographics]     [Kernel]     [Linux Admin]     [Linux Net]     [GFS]     [RPM]     [git]     [Yosemite Forum]


  Powered by Linux