Re: RAID5 disk failure during rebuild of spare, any chance of recovery when one of the failed devices is suspected to be intact?

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Mon, Aug 16, 2010 at 10:43 AM, Tim Small <tim@xxxxxxxxxxx> wrote:
> On 16/08/10 07:12, Nicolas Jungers wrote:
>>
>> On 08/16/2010 07:54 AM, Tor Arne Vestbø wrote:
>>>
>>> You mean you sdc and sde plus either sdb or sdd, depending on which
>>> one I think is more sane a this point?
>>
>> I'd try both.  Do a ddrescue of the failing one and try that (with copy of
>> the others) and check what's coming out.
>
> As an alternative to using ddrescue, you could quickly prototype various
> arrangements (without writing anything to the drives) using a device-mapper
> copy-on-write mapping - I posted some details to the list a while back when
> I was trying to use this to reconstruct a hw raid array...  Check the list
> archives for details.

Cool, here's what I tried:

Created spares files for each of the devices

  dd if=/dev/zero of=sdb_cow bs=1 count=0 seek=2GB

Mapped that to a loop device

  losetup /dev/loop1 sdb_cow

Then ran the following for each device:

  cow_size=`blockdev --getsize /dev/sdb1`
  chunk_size=64
  echo "0 $cow_size snapshot /dev/sdb1 /dev/loop1 p $chunk_size" |
dmsetup create sdb1_cow

After these were created I tried the following:

# mdadm -v -C /dev/md0 -l5 -n4 /dev/mapper/sdb1_cow
/dev/mapper/sdc1_cow missing /dev/mapper/sde1_cow
mdadm: layout defaults to left-symmetric
mdadm: chunk size defaults to 64K
mdadm: /dev/mapper/sdb1_cow appears to be part of a raid array:
    level=raid5 devices=4 ctime=Sun Mar  2 22:52:53 2008
mdadm: /dev/mapper/sdc1_cow appears to be part of a raid array:
    level=raid5 devices=4 ctime=Sun Mar  2 22:52:53 2008
mdadm: /dev/mapper/sde1_cow appears to be part of a raid array:
    level=raid5 devices=4 ctime=Sun Mar  2 22:52:53 2008
mdadm: size set to 732571904K
Continue creating array? Y
mdadm: array /dev/md0 started.

# mdadm --detail /dev/md0
/dev/md0:
        Version : 00.90
  Creation Time : Mon Aug 16 18:20:06 2010
     Raid Level : raid5
     Array Size : 2197715712 (2095.91 GiB 2250.46 GB)
  Used Dev Size : 732571904 (698.64 GiB 750.15 GB)
   Raid Devices : 4
  Total Devices : 3
Preferred Minor : 0
    Persistence : Superblock is persistent

    Update Time : Mon Aug 16 18:20:06 2010
          State : clean, degraded
 Active Devices : 3
Working Devices : 3
 Failed Devices : 0
  Spare Devices : 0

         Layout : left-symmetric
     Chunk Size : 64K

           UUID : 916ceaa2:b877a3cc:3973abef:31f2d600 (local to host monstre)
         Events : 0.1

    Number   Major   Minor   RaidDevice State
       0     251        9        0      active sync   /dev/block/251:9
       1     251       10        1      active sync   /dev/block/251:10
       2       0        0        2      removed
       3     251       12        3      active sync   /dev/block/251:12

And I can now mount /dev/mapper/raid-home !

The question now is, what next? Should I start copying things off to a
backup, or run fsck first or something else to try to repair errors?
Or perhaps are the 2GB sparse files to small for anything like that?

Thanks!

Tor Arne
--
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html


[Index of Archives]     [Linux RAID Wiki]     [ATA RAID]     [Linux SCSI Target Infrastructure]     [Linux Block]     [Linux IDE]     [Linux SCSI]     [Linux Hams]     [Device Mapper]     [Device Mapper Cryptographics]     [Kernel]     [Linux Admin]     [Linux Net]     [GFS]     [RPM]     [git]     [Yosemite Forum]


  Powered by Linux