Thanks, that recovered the array in degraded mode :) I have added another hard disk and the array is rebuilding now. Thanks again. On Tue, Jul 24, 2012 at 3:13 AM, NeilBrown <neilb@xxxxxxx> wrote: > On Mon, 23 Jul 2012 21:54:24 +0500 Khurram Hassan <kfhassan@xxxxxxxxx> wrote: > >> raid.status contents: >> >> /dev/sdb: >> Magic : Intel Raid ISM Cfg Sig. >> Version : 1.2.02 >> Orig Family : 00000000 >> Family : 6eb404da >> Generation : 002308e9 >> Attributes : All supported >> UUID : 51c75501:a307676f:d2d6e547:dfcb2476 >> Checksum : 06cf5ff9 correct >> MPB Sectors : 2 >> Disks : 3 >> RAID Devices : 1 >> >> Disk01 Serial : 5VMLEGC6 >> State : active >> Id : 00030000 >> Usable Size : 976768264 (465.76 GiB 500.11 GB) >> >> [VolumeData500:1]: >> UUID : a0865f28:57b7246b:ff43fa76:5531f5ca >> RAID Level : 5 >> Members : 3 >> Slots : [___] >> Failed disk : 1 >> This Slot : 1 (out-of-sync) >> Array Size : 1953536000 (931.52 GiB 1000.21 GB) >> Per Dev Size : 976768264 (465.76 GiB 500.11 GB) >> Sector Offset : 0 >> Num Stripes : 3815500 >> Chunk Size : 128 KiB >> Reserved : 0 >> Migrate State : idle >> Map State : failed >> Dirty State : clean >> >> Disk00 Serial : 9VM1GGJK:1 >> State : active failed >> Id : ffffffff >> Usable Size : 976768264 (465.76 GiB 500.11 GB) >> >> Disk02 Serial : 6VM4EGHC >> State : active >> Id : 00040000 >> Usable Size : 976768264 (465.76 GiB 500.11 GB) > > You'll need to start out with > > echo 1 > /sys/module/md_mod/parameters/start_dirty_degraded > > otherwise creating the degraded raid5 won't work - I need to fix that. > Then > > mdadm -C /dev/md/imsm -e imsm -n 2 /dev/sdb /dev/sdc > mdadm -C /dev/md0 -l5 -n3 -c 128 missing /dev/sdb /dev/sdc > > so you create an IMSM container, then create the RAID5 inside that. > > You should then check the filesystem to make sure it looks right. > If not, you might need to stop the arrays and start again, using a different > order of devices in the second command. > > Good luck, > > NeilBrown > > > > >> /dev/sdc: >> Magic : Intel Raid ISM Cfg Sig. >> Version : 1.2.02 >> Orig Family : 00000000 >> Family : 6eb404da >> Generation : 002308e9 >> Attributes : All supported >> UUID : 51c75501:a307676f:d2d6e547:dfcb2476 >> Checksum : 06cf5ff9 correct >> MPB Sectors : 2 >> Disks : 3 >> RAID Devices : 1 >> >> Disk02 Serial : 6VM4EGHC >> State : active >> Id : 00040000 >> Usable Size : 976768264 (465.76 GiB 500.11 GB) >> >> [VolumeData500:1]: >> UUID : a0865f28:57b7246b:ff43fa76:5531f5ca >> RAID Level : 5 >> Members : 3 >> Slots : [___] >> Failed disk : 1 >> This Slot : 2 (out-of-sync) >> Array Size : 1953536000 (931.52 GiB 1000.21 GB) >> Per Dev Size : 976768264 (465.76 GiB 500.11 GB) >> Sector Offset : 0 >> Num Stripes : 3815500 >> Chunk Size : 128 KiB >> Reserved : 0 >> Migrate State : idle >> Map State : failed >> Dirty State : clean >> >> Disk00 Serial : 9VM1GGJK:1 >> State : active failed >> Id : ffffffff >> Usable Size : 976768264 (465.76 GiB 500.11 GB) >> >> Disk01 Serial : 5VMLEGC6 >> State : active >> Id : 00030000 >> Usable Size : 976768264 (465.76 GiB 500.11 GB) >> >> >> I hope you can figure it out as I am quite lost here. >> >> Thanks, >> Khurram >> >> >> On Mon, Jul 23, 2012 at 4:31 PM, Khurram Hassan <kfhassan@xxxxxxxxx> wrote: >> > raid.status contents: >> > >> > /dev/sdb: >> > Magic : Intel Raid ISM Cfg Sig. >> > Version : 1.2.02 >> > Orig Family : 00000000 >> > Family : 6eb404da >> > Generation : 002308e9 >> > Attributes : All supported >> > UUID : 51c75501:a307676f:d2d6e547:dfcb2476 >> > Checksum : 06cf5ff9 correct >> > MPB Sectors : 2 >> > Disks : 3 >> > RAID Devices : 1 >> > >> > Disk01 Serial : 5VMLEGC6 >> > State : active >> > Id : 00030000 >> > Usable Size : 976768264 (465.76 GiB 500.11 GB) >> > >> > [VolumeData500:1]: >> > UUID : a0865f28:57b7246b:ff43fa76:5531f5ca >> > RAID Level : 5 >> > Members : 3 >> > Slots : [___] >> > Failed disk : 1 >> > This Slot : 1 (out-of-sync) >> > Array Size : 1953536000 (931.52 GiB 1000.21 GB) >> > Per Dev Size : 976768264 (465.76 GiB 500.11 GB) >> > Sector Offset : 0 >> > Num Stripes : 3815500 >> > Chunk Size : 128 KiB >> > Reserved : 0 >> > Migrate State : idle >> > Map State : failed >> > Dirty State : clean >> > >> > Disk00 Serial : 9VM1GGJK:1 >> > State : active failed >> > Id : ffffffff >> > Usable Size : 976768264 (465.76 GiB 500.11 GB) >> > >> > Disk02 Serial : 6VM4EGHC >> > State : active >> > Id : 00040000 >> > Usable Size : 976768264 (465.76 GiB 500.11 GB) >> > /dev/sdc: >> > Magic : Intel Raid ISM Cfg Sig. >> > Version : 1.2.02 >> > Orig Family : 00000000 >> > Family : 6eb404da >> > Generation : 002308e9 >> > Attributes : All supported >> > UUID : 51c75501:a307676f:d2d6e547:dfcb2476 >> > Checksum : 06cf5ff9 correct >> > MPB Sectors : 2 >> > Disks : 3 >> > RAID Devices : 1 >> > >> > Disk02 Serial : 6VM4EGHC >> > State : active >> > Id : 00040000 >> > Usable Size : 976768264 (465.76 GiB 500.11 GB) >> > >> > [VolumeData500:1]: >> > UUID : a0865f28:57b7246b:ff43fa76:5531f5ca >> > RAID Level : 5 >> > Members : 3 >> > Slots : [___] >> > Failed disk : 1 >> > This Slot : 2 (out-of-sync) >> > Array Size : 1953536000 (931.52 GiB 1000.21 GB) >> > Per Dev Size : 976768264 (465.76 GiB 500.11 GB) >> > Sector Offset : 0 >> > Num Stripes : 3815500 >> > Chunk Size : 128 KiB >> > Reserved : 0 >> > Migrate State : idle >> > Map State : failed >> > Dirty State : clean >> > >> > Disk00 Serial : 9VM1GGJK:1 >> > State : active failed >> > Id : ffffffff >> > Usable Size : 976768264 (465.76 GiB 500.11 GB) >> > >> > Disk01 Serial : 5VMLEGC6 >> > State : active >> > Id : 00030000 >> > Usable Size : 976768264 (465.76 GiB 500.11 GB) >> > >> > >> > I hope you can figure it out as I am quite lost here. >> > >> > Thanks, >> > Khurram >> > >> > On Mon, Jul 23, 2012 at 4:08 AM, NeilBrown <neilb@xxxxxxx> wrote: >> >> On Sat, 21 Jul 2012 21:00:19 +0500 Khurram Hassan <kfhassan@xxxxxxxxx> wrote: >> >> >> >>> I have this 3 disk raid5 volumne on an Asus motherboard sporting an >> >>> Intel Rapid Storage chipset. The problem began when I noticed in >> >>> windows that one of the hard disks (the first one in the array) was >> >>> marked as failed in the Intel raid utility. I shutdown the system to >> >>> remove the hard disk and removed the cables for the faulty hard disk. >> >>> But I made a mistake and remove the cables for one of the working hard >> >>> disks. So when I booted, it showed the raid volume as failed. I >> >>> quickly shutdown the system and corrected the mistake. But it >> >>> completely hosed my raid volume. When I booted the system up again, >> >>> both of the remaining 2 hard disks were showed as offline. >> >>> >> >>> I read the raid recovery section in the wiki and installed ubuntu >> >>> 12.04 on a separate non-raid hard disk (after completely disconnecting >> >>> the offline raid5 volume). Then I reconnected the 2 hard disks and >> >>> booted ubuntu. Then I gave the following commands: >> >>> >> >>> 1) mdadm --examine /dev/sd[bc] > raid.status >> >>> 2) mdadm --create --assume-clean -c 128 --level=5 --raid-devices=3 >> >>> /dev/md1 missing /dev/sdb /dev/sdc >> >>> >> >>> It gave the following output: >> >>> mdadm: /dev/sdb appears to be part of a raid array: >> >>> level=container devices=0 ctime=Thu Jan 1 05:00:00 1970 >> >>> mdadm: /dev/sdc appears to be part of a raid array: >> >>> level=container devices=0 ctime=Thu Jan 1 05:00:00 1970 >> >>> Continue creating array? y >> >>> mdadm: Defaulting to version 1.2 metadata >> >>> mdadm: array /dev/md1 started. >> >>> >> >>> But the raid volume is not accessible. mdadm --examine /dev/md1 gives: >> >>> >> >>> mdadm: No md superblock detected on /dev/md1. >> >>> >> >>> Worse, upon booting the system, the raid chipset message says the 2 >> >>> hard disk are non-raid hard disks. Have I completely messed up the >> >>> raid volume? Is it not recoverable at all? >> >> >> >> Possibly :-( >> >> >> >> You had an array with Intel-specific metadata. This metadata is stored at >> >> the end of the device. >> >> >> >> When you tried to "--create" the array, you did not ask for intel metadata so >> >> you got the default v1.2 metadata. This metadata is stored at the beginning >> >> of the device (a 1K block, 4K from the start). >> >> So this would have over-written a small amount of filesystem data. >> >> >> >> Also when you --create an array, mdadm erases any other metadata that it >> >> finds to avoid confusion. So it will have erased the Intel metadata from the >> >> end. >> >> >> >> Your best hope is to recreate the array correctly with intel metadata. The >> >> filesystem will quite possibly be corrupted, but you might get some or even >> >> all of your data back. >> >> >> >> Can you post the "raid.status". That would help be certain we are doing the >> >> right thing. >> >> Something like >> >> mdadm --create /dev/md/imsm -e imsm -n 3 missing /dev/sdb /dev/sdc >> >> mdadm --create /dev/md1 -c 128 -l 5 -n 3 /dev/md/imsm >> >> >> >> might do it ... or might not. I'm not sure about creating imsm arrays with >> >> missing devices. Maybe you still list the 3 devices rather than just the >> >> container. I'd need to experiment. If you post the raid.status I'll see if >> >> I can work out the best way forward. >> >> >> >> NeilBrown >> >> >> -- >> To unsubscribe from this list: send the line "unsubscribe linux-raid" in >> the body of a message to majordomo@xxxxxxxxxxxxxxx >> More majordomo info at http://vger.kernel.org/majordomo-info.html > -- To unsubscribe from this list: send the line "unsubscribe linux-raid" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html