Re: Problem recovering failed Intel Rapid Storage raid5 volume

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



raid.status contents:

/dev/sdb:
          Magic : Intel Raid ISM Cfg Sig.
        Version : 1.2.02
    Orig Family : 00000000
         Family : 6eb404da
     Generation : 002308e9
     Attributes : All supported
           UUID : 51c75501:a307676f:d2d6e547:dfcb2476
       Checksum : 06cf5ff9 correct
    MPB Sectors : 2
          Disks : 3
   RAID Devices : 1

  Disk01 Serial : 5VMLEGC6
          State : active
             Id : 00030000
    Usable Size : 976768264 (465.76 GiB 500.11 GB)

[VolumeData500:1]:
           UUID : a0865f28:57b7246b:ff43fa76:5531f5ca
     RAID Level : 5
        Members : 3
          Slots : [___]
    Failed disk : 1
      This Slot : 1 (out-of-sync)
     Array Size : 1953536000 (931.52 GiB 1000.21 GB)
   Per Dev Size : 976768264 (465.76 GiB 500.11 GB)
  Sector Offset : 0
    Num Stripes : 3815500
     Chunk Size : 128 KiB
       Reserved : 0
  Migrate State : idle
      Map State : failed
    Dirty State : clean

  Disk00 Serial : 9VM1GGJK:1
          State : active failed
             Id : ffffffff
    Usable Size : 976768264 (465.76 GiB 500.11 GB)

  Disk02 Serial : 6VM4EGHC
          State : active
             Id : 00040000
    Usable Size : 976768264 (465.76 GiB 500.11 GB)
/dev/sdc:
          Magic : Intel Raid ISM Cfg Sig.
        Version : 1.2.02
    Orig Family : 00000000
         Family : 6eb404da
     Generation : 002308e9
     Attributes : All supported
           UUID : 51c75501:a307676f:d2d6e547:dfcb2476
       Checksum : 06cf5ff9 correct
    MPB Sectors : 2
          Disks : 3
   RAID Devices : 1

  Disk02 Serial : 6VM4EGHC
          State : active
             Id : 00040000
    Usable Size : 976768264 (465.76 GiB 500.11 GB)

[VolumeData500:1]:
           UUID : a0865f28:57b7246b:ff43fa76:5531f5ca
     RAID Level : 5
        Members : 3
          Slots : [___]
    Failed disk : 1
      This Slot : 2 (out-of-sync)
     Array Size : 1953536000 (931.52 GiB 1000.21 GB)
   Per Dev Size : 976768264 (465.76 GiB 500.11 GB)
  Sector Offset : 0
    Num Stripes : 3815500
     Chunk Size : 128 KiB
       Reserved : 0
  Migrate State : idle
      Map State : failed
    Dirty State : clean

  Disk00 Serial : 9VM1GGJK:1
          State : active failed
             Id : ffffffff
    Usable Size : 976768264 (465.76 GiB 500.11 GB)

  Disk01 Serial : 5VMLEGC6
          State : active
             Id : 00030000
    Usable Size : 976768264 (465.76 GiB 500.11 GB)


I hope you can figure it out as I am quite lost here.

Thanks,
Khurram


On Mon, Jul 23, 2012 at 4:31 PM, Khurram Hassan <kfhassan@xxxxxxxxx> wrote:
> raid.status contents:
>
> /dev/sdb:
>           Magic : Intel Raid ISM Cfg Sig.
>         Version : 1.2.02
>     Orig Family : 00000000
>          Family : 6eb404da
>      Generation : 002308e9
>      Attributes : All supported
>            UUID : 51c75501:a307676f:d2d6e547:dfcb2476
>        Checksum : 06cf5ff9 correct
>     MPB Sectors : 2
>           Disks : 3
>    RAID Devices : 1
>
>   Disk01 Serial : 5VMLEGC6
>           State : active
>              Id : 00030000
>     Usable Size : 976768264 (465.76 GiB 500.11 GB)
>
> [VolumeData500:1]:
>            UUID : a0865f28:57b7246b:ff43fa76:5531f5ca
>      RAID Level : 5
>         Members : 3
>           Slots : [___]
>     Failed disk : 1
>       This Slot : 1 (out-of-sync)
>      Array Size : 1953536000 (931.52 GiB 1000.21 GB)
>    Per Dev Size : 976768264 (465.76 GiB 500.11 GB)
>   Sector Offset : 0
>     Num Stripes : 3815500
>      Chunk Size : 128 KiB
>        Reserved : 0
>   Migrate State : idle
>       Map State : failed
>     Dirty State : clean
>
>   Disk00 Serial : 9VM1GGJK:1
>           State : active failed
>              Id : ffffffff
>     Usable Size : 976768264 (465.76 GiB 500.11 GB)
>
>   Disk02 Serial : 6VM4EGHC
>           State : active
>              Id : 00040000
>     Usable Size : 976768264 (465.76 GiB 500.11 GB)
> /dev/sdc:
>           Magic : Intel Raid ISM Cfg Sig.
>         Version : 1.2.02
>     Orig Family : 00000000
>          Family : 6eb404da
>      Generation : 002308e9
>      Attributes : All supported
>            UUID : 51c75501:a307676f:d2d6e547:dfcb2476
>        Checksum : 06cf5ff9 correct
>     MPB Sectors : 2
>           Disks : 3
>    RAID Devices : 1
>
>   Disk02 Serial : 6VM4EGHC
>           State : active
>              Id : 00040000
>     Usable Size : 976768264 (465.76 GiB 500.11 GB)
>
> [VolumeData500:1]:
>            UUID : a0865f28:57b7246b:ff43fa76:5531f5ca
>      RAID Level : 5
>         Members : 3
>           Slots : [___]
>     Failed disk : 1
>       This Slot : 2 (out-of-sync)
>      Array Size : 1953536000 (931.52 GiB 1000.21 GB)
>    Per Dev Size : 976768264 (465.76 GiB 500.11 GB)
>   Sector Offset : 0
>     Num Stripes : 3815500
>      Chunk Size : 128 KiB
>        Reserved : 0
>   Migrate State : idle
>       Map State : failed
>     Dirty State : clean
>
>   Disk00 Serial : 9VM1GGJK:1
>           State : active failed
>              Id : ffffffff
>     Usable Size : 976768264 (465.76 GiB 500.11 GB)
>
>   Disk01 Serial : 5VMLEGC6
>           State : active
>              Id : 00030000
>     Usable Size : 976768264 (465.76 GiB 500.11 GB)
>
>
> I hope you can figure it out as I am quite lost here.
>
> Thanks,
> Khurram
>
> On Mon, Jul 23, 2012 at 4:08 AM, NeilBrown <neilb@xxxxxxx> wrote:
>> On Sat, 21 Jul 2012 21:00:19 +0500 Khurram Hassan <kfhassan@xxxxxxxxx> wrote:
>>
>>> I have this 3 disk raid5 volumne on an Asus motherboard sporting an
>>> Intel Rapid Storage chipset. The problem began when I noticed in
>>> windows that one of the hard disks (the first one in the array) was
>>> marked as failed in the Intel raid utility. I shutdown the system to
>>> remove the hard disk and removed the cables for the faulty hard disk.
>>> But I made a mistake and remove the cables for one of the working hard
>>> disks. So when I booted, it showed the raid volume as failed. I
>>> quickly shutdown the system and corrected the mistake. But it
>>> completely hosed my raid volume. When I booted the system up again,
>>> both of the remaining 2 hard disks were showed as offline.
>>>
>>> I read the raid recovery section in the wiki and installed ubuntu
>>> 12.04 on a separate non-raid hard disk (after completely disconnecting
>>> the offline raid5 volume). Then I reconnected the 2 hard disks and
>>> booted ubuntu. Then I gave the following commands:
>>>
>>> 1) mdadm --examine /dev/sd[bc] > raid.status
>>> 2) mdadm --create --assume-clean -c 128 --level=5 --raid-devices=3
>>> /dev/md1 missing /dev/sdb /dev/sdc
>>>
>>> It gave the following output:
>>>     mdadm: /dev/sdb appears to be part of a raid array:
>>>         level=container devices=0 ctime=Thu Jan  1 05:00:00 1970
>>>     mdadm: /dev/sdc appears to be part of a raid array:
>>>         level=container devices=0 ctime=Thu Jan  1 05:00:00 1970
>>>     Continue creating array? y
>>>     mdadm: Defaulting to version 1.2 metadata
>>>     mdadm: array /dev/md1 started.
>>>
>>> But the raid volume is not accessible. mdadm --examine /dev/md1 gives:
>>>
>>>     mdadm: No md superblock detected on /dev/md1.
>>>
>>> Worse, upon booting the system, the raid chipset message says the 2
>>> hard disk are non-raid hard disks. Have I completely messed up the
>>> raid volume? Is it not recoverable at all?
>>
>> Possibly :-(
>>
>> You had an array with Intel-specific metadata.  This metadata is stored at
>> the end of the device.
>>
>> When you tried to "--create" the array, you did not ask for intel metadata so
>> you got the default v1.2 metadata.  This metadata is stored at the beginning
>> of the device (a 1K block, 4K from the start).
>> So this would have over-written a small amount of filesystem data.
>>
>> Also when you --create an array, mdadm erases any other metadata that it
>> finds to avoid confusion.  So it will have erased the Intel metadata from the
>> end.
>>
>> Your best hope is to recreate the array correctly with intel metadata.  The
>> filesystem will quite possibly be corrupted, but you might get some or even
>> all of your data back.
>>
>> Can you post the "raid.status".  That would help be certain we are doing the
>> right thing.
>> Something like
>>   mdadm --create /dev/md/imsm -e imsm -n 3 missing /dev/sdb /dev/sdc
>>   mdadm --create /dev/md1 -c 128 -l 5 -n 3 /dev/md/imsm
>>
>> might do it  ... or might not.  I'm not sure about creating imsm arrays with
>> missing devices.  Maybe you still list the 3 devices rather than just the
>> container.  I'd need to experiment.  If you post the raid.status I'll see if
>> I can work out the best way forward.
>>
>> NeilBrown
>>
--
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html


[Index of Archives]     [Linux RAID Wiki]     [ATA RAID]     [Linux SCSI Target Infrastructure]     [Linux Block]     [Linux IDE]     [Linux SCSI]     [Linux Hams]     [Device Mapper]     [Device Mapper Cryptographics]     [Kernel]     [Linux Admin]     [Linux Net]     [GFS]     [RPM]     [git]     [Yosemite Forum]


  Powered by Linux