RE: How to un-degrade an array after a totally spurious failure?

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Mon, August 3, 2009 5:30 pm, Leslie Rhorer wrote:
>
>
>> >> NeilBrown
>> >
>> > 	I have exactly the same situation, except there are two "failed"
>> > disks on a RAID5 array.  As for the OP, the "failures" are spurious.
>> > Running the remove and then the add command puts the disks back in as
>> > spare
>> > disks, not live ones, and then the array just sits there, doing
>> nothing.
>> > I
>> > tried the trick of doing
>> >
>> > echo repair > /sys/block/md0/md/sync_action
>> >
>> > but the array still just sits there saying it is "clean, degraded",
>> with
>> 2
>> > spare and 5 working devices.
>> >
>>
>> No such a good option when you have two failures.
>> If you have two failures you need to stop the array, then assemble
>> it again using --force.
>> It is now too late for that:  adding them with "--add" will have erased
>> the old metadata.
>>
>> Your only option is to re-create the array.  Make sure you use the
>> same parameters (e.g. chunk size) as when you first created the array.
>> You can check the correct parameters by looking at a device with
>> "--examine".
>> Also make sure you put the devices in the correct order.
>>
>> The best thing to do is to try creating the array, using
>> "--assume-clean" so it won't trigger a resync.
>>  Then use "fsck", the "mount" to make sure the data is good.
>> Once you are satisfied that the data is good and that you created
>> the array with the right parameters, use "echo repair > ...."
>> to make sure the array really is 'clean'.
>>
>> I guess I should stop mdadm from trashing the superblock when you
>> add a spare to an array which has failed....
>>
>> NeilBrown
>
> OK, Neil, I've had this occur again.  A prolonged power failure took one
> of
> the systems offline, and now it's convicting 3 of 7 disks in a RAID5
> array.
> I've done nothing but stop the array.  Prior to stopping the array, mdadm
> reported:
>
> Backup:/# mdadm -Dt /dev/md0
> /dev/md0:
>         Version : 01.02
>   Creation Time : Sun Jul 12 20:44:02 2009
>      Raid Level : raid5
>      Array Size : 8790830592 (8383.59 GiB 9001.81 GB)
>   Used Dev Size : 2930276864 (2794.53 GiB 3000.60 GB)
>    Raid Devices : 7
>   Total Devices : 7
> Preferred Minor : 0
>     Persistence : Superblock is persistent
>
>     Update Time : Sun Aug  2 04:01:52 2009
>           State : clean, degraded
>  Active Devices : 4
> Working Devices : 4
>  Failed Devices : 3
>   Spare Devices : 0
>
>          Layout : left-symmetric
>      Chunk Size : 256K
>
>            Name : 'Backup':0
>            UUID : 940ae4e4:04057ffc:5e92d2fb:63e3efb7
>          Events : 14
>
>     Number   Major   Minor   RaidDevice State
>        0       8       80        0      active sync   /dev/sdf
>        1       8       96        1      active sync   /dev/sdg
>        2       8        0        2      active sync   /dev/sda
>        3       8       16        3      active sync   /dev/sdb
>        4       0        0        4      removed
>        5       0        0        5      removed
>        6       0        0        6      removed
>
>        4       8       32        -      faulty spare   /dev/sdc
>        5       8       48        -      faulty spare   /dev/sdd
>        6       8       64        -      faulty spare   /dev/sde
>
> 	The array did have a bitmap, but mdadm didn't report it.  When I
> attempt using --assemble and --force, I get:
>
> Backup:/# mdadm --assemble --force /dev/md0
> mdadm: /dev/md0 not identified in config file.

So try
  mdadm --assemble --force /dev/md0 /dev/sd[abcdefg]

NeilBrown


>
> --
> To unsubscribe from this list: send the line "unsubscribe linux-raid" in
> the body of a message to majordomo@xxxxxxxxxxxxxxx
> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>

--
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html

[Index of Archives]     [Linux RAID Wiki]     [ATA RAID]     [Linux SCSI Target Infrastructure]     [Linux Block]     [Linux IDE]     [Linux SCSI]     [Linux Hams]     [Device Mapper]     [Device Mapper Cryptographics]     [Kernel]     [Linux Admin]     [Linux Net]     [GFS]     [RPM]     [git]     [Yosemite Forum]


  Powered by Linux