Re: trouble repairing raid10

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Wed, 02 Jun 2010 18:25:58 +0200
Nicolas Jungers <nicolas@xxxxxxxxxxx> wrote:

> I've a 4 HD raid10 with to failed drive.  Any attempt I made to add 2 
> replacement disks fail consistently.
> 
> mdadm -Af /dev/md1 /dev/sdm2 /dev/sdp2   /dev/sdb2  /dev/sdd2
> mdadm: failed to add /dev/sdd2 to /dev/md1: Device or resource busy

Any idea why sdd2 is busy??


> mdadm: /dev/md1 assembled from 2 drives and 1 spare - not enough to 
> start the array.
> 
> or
> 
> root@disk:~# mdadm -AR /dev/md1 /dev/sdm2 /dev/sdp2
> mdadm: failed to RUN_ARRAY /dev/md1: Input/output error
> mdadm: Not enough devices to start the array.
> root@disk:~# mdadm --add /dev/md1 /dev/sdb2
> mdadm: add new device failed for /dev/sdb2 as 4: Invalid argument
> 
> 
> The array is in near mode and I lost disk 0 and 1.  Does it mean that my 
> data are toasted?

Yes.  RAID10 can survive the failure of 2 non-adjacent devices and sometimes
2 adjacent devices.  But not 0 and 1 of a near=2 array.

So if those devices are really dead, so is your data.

If one of these is actually usable and just had a transient failure then you
could try re-creating the array with the drives, or 'missing' in the right
order and with the write layout/chunksize set.
You would need to be user the 'Data Offset' was the same, which unfortunately
can require using exactly the same version of mdadm as created the array in
the first place.

NeilBrown

> 
> 
> 
> mdadm --examine /dev/sdm2
> /dev/sdm2:
>            Magic : a92b4efc
>          Version : 1.2
>      Feature Map : 0x0
>       Array UUID : d90ad6fe:1355134f:f83ffadc:a4fe7859
>             Name : m1:1
>    Creation Time : Thu Apr  1 21:28:58 2010
>       Raid Level : raid10
>     Raid Devices : 4
> 
>   Avail Dev Size : 3907026909 (1863.02 GiB 2000.40 GB)
>       Array Size : 7814049792 (3726.03 GiB 4000.79 GB)
>    Used Dev Size : 3907024896 (1863.01 GiB 2000.40 GB)
>      Data Offset : 272 sectors
>     Super Offset : 8 sectors
>            State : clean
>      Device UUID : e217355e:632ac2f0:8120e55e:3878bd88
> 
>      Update Time : Wed Jun  2 12:31:39 2010
>         Checksum : feef2809 - correct
>           Events : 1377156
> 
>           Layout : near=2, far=1
>       Chunk Size : 1024K
> 
>      Array Slot : 3 (failed, failed, 2, 3)
>     Array State : __uU 2 failed
> --
> To unsubscribe from this list: send the line "unsubscribe linux-raid" in
> the body of a message to majordomo@xxxxxxxxxxxxxxx
> More majordomo info at  http://vger.kernel.org/majordomo-info.html

--
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html


[Index of Archives]     [Linux RAID Wiki]     [ATA RAID]     [Linux SCSI Target Infrastructure]     [Linux Block]     [Linux IDE]     [Linux SCSI]     [Linux Hams]     [Device Mapper]     [Device Mapper Cryptographics]     [Kernel]     [Linux Admin]     [Linux Net]     [GFS]     [RPM]     [git]     [Yosemite Forum]


  Powered by Linux