On Mon, August 3, 2009 5:30 pm, Leslie Rhorer wrote: > > >> >> NeilBrown >> > >> > I have exactly the same situation, except there are two "failed" >> > disks on a RAID5 array. As for the OP, the "failures" are spurious. >> > Running the remove and then the add command puts the disks back in as >> > spare >> > disks, not live ones, and then the array just sits there, doing >> nothing. >> > I >> > tried the trick of doing >> > >> > echo repair > /sys/block/md0/md/sync_action >> > >> > but the array still just sits there saying it is "clean, degraded", >> with >> 2 >> > spare and 5 working devices. >> > >> >> No such a good option when you have two failures. >> If you have two failures you need to stop the array, then assemble >> it again using --force. >> It is now too late for that: adding them with "--add" will have erased >> the old metadata. >> >> Your only option is to re-create the array. Make sure you use the >> same parameters (e.g. chunk size) as when you first created the array. >> You can check the correct parameters by looking at a device with >> "--examine". >> Also make sure you put the devices in the correct order. >> >> The best thing to do is to try creating the array, using >> "--assume-clean" so it won't trigger a resync. >> Then use "fsck", the "mount" to make sure the data is good. >> Once you are satisfied that the data is good and that you created >> the array with the right parameters, use "echo repair > ...." >> to make sure the array really is 'clean'. >> >> I guess I should stop mdadm from trashing the superblock when you >> add a spare to an array which has failed.... >> >> NeilBrown > > OK, Neil, I've had this occur again. A prolonged power failure took one > of > the systems offline, and now it's convicting 3 of 7 disks in a RAID5 > array. > I've done nothing but stop the array. Prior to stopping the array, mdadm > reported: > > Backup:/# mdadm -Dt /dev/md0 > /dev/md0: > Version : 01.02 > Creation Time : Sun Jul 12 20:44:02 2009 > Raid Level : raid5 > Array Size : 8790830592 (8383.59 GiB 9001.81 GB) > Used Dev Size : 2930276864 (2794.53 GiB 3000.60 GB) > Raid Devices : 7 > Total Devices : 7 > Preferred Minor : 0 > Persistence : Superblock is persistent > > Update Time : Sun Aug 2 04:01:52 2009 > State : clean, degraded > Active Devices : 4 > Working Devices : 4 > Failed Devices : 3 > Spare Devices : 0 > > Layout : left-symmetric > Chunk Size : 256K > > Name : 'Backup':0 > UUID : 940ae4e4:04057ffc:5e92d2fb:63e3efb7 > Events : 14 > > Number Major Minor RaidDevice State > 0 8 80 0 active sync /dev/sdf > 1 8 96 1 active sync /dev/sdg > 2 8 0 2 active sync /dev/sda > 3 8 16 3 active sync /dev/sdb > 4 0 0 4 removed > 5 0 0 5 removed > 6 0 0 6 removed > > 4 8 32 - faulty spare /dev/sdc > 5 8 48 - faulty spare /dev/sdd > 6 8 64 - faulty spare /dev/sde > > The array did have a bitmap, but mdadm didn't report it. When I > attempt using --assemble and --force, I get: > > Backup:/# mdadm --assemble --force /dev/md0 > mdadm: /dev/md0 not identified in config file. So try mdadm --assemble --force /dev/md0 /dev/sd[abcdefg] NeilBrown > > -- > To unsubscribe from this list: send the line "unsubscribe linux-raid" in > the body of a message to majordomo@xxxxxxxxxxxxxxx > More majordomo info at http://vger.kernel.org/majordomo-info.html > -- To unsubscribe from this list: send the line "unsubscribe linux-raid" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html