mdadm degraded RAID5 failure

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi all..

I had one of the disks in my 3 disk RAID5 die on me this week. When
attempting to replace the disk via a hot swap (USB), the RAID didn't
like it. It decided to mark one of my remaining 2 disks as faulty.

Can someone *please* help me get the raid back!?

More details -

Drives are /dev/sdb1, /dev/sdc1 & /dev/sdd1

sdc1 was the one that died earlier this week
sdb1 appears to be the one that was marked as faulty

mdadm detail before sdc1 was plugged in -

root@imp[~]:11 # mdadm --detail /dev/md1
/dev/md1:
Version : 00.90.01
Creation Time : Fri Nov 17 21:28:44 2006
Raid Level : raid5
Array Size : 586067072 (558.92 GiB 600.13 GB)
Device Size : 293033536 (279.46 GiB 300.07 GB)
Raid Devices : 3
Total Devices : 2
Preferred Minor : 1
Persistence : Superblock is persistent

Update Time : Sat Oct 18 20:06:34 2008
State : clean, degraded
Active Devices : 2
Working Devices : 2
Failed Devices : 0
Spare Devices : 0

Layout : left-symmetric
Chunk Size : 64K

UUID : bed40ee2:98523fdd:e4d010fb:894c0966
Events : 0.1474312

Number Major Minor RaidDevice State
0 8 17 0 active sync /dev/sdb1
1 0 0 - removed
2 8 49 2 active sync /dev/sdd1


then after plugging in the replacement sdc1 -

root@imp[~]:13 # mdadm --add /dev/md1 /dev/sdc1
mdadm: hot added /dev/sdc1
root@imp[~]:14 #
root@imp[~]:14 #
root@imp[~]:14 # mdadm --detail /dev/md1
/dev/md1:
Version : 00.90.01
Creation Time : Fri Nov 17 21:28:44 2006
Raid Level : raid5
Array Size : 586067072 (558.92 GiB 600.13 GB)
Device Size : 293033536 (279.46 GiB 300.07 GB)
Raid Devices : 3
Total Devices : 3
Preferred Minor : 1
Persistence : Superblock is persistent

Update Time : Sat Oct 18 22:13:13 2008
State : clean, degraded
Active Devices : 1
Working Devices : 2
Failed Devices : 1
Spare Devices : 1

Layout : left-symmetric
Chunk Size : 64K

UUID : bed40ee2:98523fdd:e4d010fb:894c0966
Events : 0.1480366

Number Major Minor RaidDevice State
0 0 0 - removed
1 0 0 - removed
2 8 49 2 active sync /dev/sdd1

3 8 33 0 spare rebuilding /dev/sdc1
4 8 17 - faulty /dev/sdb1

Shortly after this, subsequent mdadm --details stopped responding.. So
I rebooted in the hope I could reset and problems with the hot add..

Now, I'm unable to assemble the raid with the 2 working drives -

mdadm --assemble /dev/md1 /dev/sdb1 /dev/sdd1

doesn't work -

mdadm: /dev/md1 assembled from 1 drive and 1 spare - not enough to
start the array.

mdadm --assemble --force /dev/md1 /dev/sdb1 /dev/sdd1

doesn't' work either

This -

mdadm --assemble --force --run /dev/md1 /dev/sdb1 /dev/sdd1

Did work partially -

/dev/md1:
Version : 00.90.01
Creation Time : Fri Nov 17 21:28:44 2006
Raid Level : raid5
Device Size : 293033536 (279.46 GiB 300.07 GB)
Raid Devices : 3
Total Devices : 2
Preferred Minor : 1
Persistence : Superblock is persistent

Update Time : Sat Oct 18 22:14:48 2008
State : active, degraded
Active Devices : 1
Working Devices : 2
Failed Devices : 0
Spare Devices : 1

Layout : left-symmetric
Chunk Size : 64K

UUID : bed40ee2:98523fdd:e4d010fb:894c0966
Events : 0.1521614

Number Major Minor RaidDevice State
0 0 0 - removed
1 0 0 - removed
2 8 49 2 active sync /dev/sdd1

3 8 17 - spare /dev/sdb1

Here's the output from mdadm -E on each of the 2 drives -

/dev/sdb1:
Magic : a92b4efc
Version : 00.90.00
UUID : bed40ee2:98523fdd:e4d010fb:894c0966
Creation Time : Fri Nov 17 21:28:44 2006
Raid Level : raid5
Raid Devices : 3
Total Devices : 3
Preferred Minor : 1

Update Time : Sat Oct 18 22:14:48 2008
State : clean
Active Devices : 1
Working Devices : 2
Failed Devices : 2
Spare Devices : 1
Checksum : e6dbf75 - correct
Events : 0.1521614

Layout : left-symmetric
Chunk Size : 64K

Number Major Minor RaidDevice State
this 3 8 33 3 spare /dev/sdc1

0 0 0 0 0 removed
1 1 0 0 1 faulty removed
2 2 8 49 2 active sync /dev/sdd1
3 3 8 33 3 spare /dev/sdc1
/dev/sdd1:
Magic : a92b4efc
Version : 00.90.00
UUID : bed40ee2:98523fdd:e4d010fb:894c0966
Creation Time : Fri Nov 17 21:28:44 2006
Raid Level : raid5
Raid Devices : 3
Total Devices : 3
Preferred Minor : 1

Update Time : Sat Oct 18 22:14:48 2008
State : clean
Active Devices : 1
Working Devices : 2
Failed Devices : 2
Spare Devices : 1
Checksum : e6dbf86 - correct
Events : 0.1521614

Layout : left-symmetric
Chunk Size : 64K

Number Major Minor RaidDevice State
this 2 8 49 2 active sync /dev/sdd1

0 0 0 0 0 removed
1 1 0 0 1 faulty removed
2 2 8 49 2 active sync /dev/sdd1
3 3 8 33 0 spare /dev/sdc1

root@imp[~]:28 # mdadm --version
mdadm - v1.9.0 - 04 February 2005
root@imp[~]:29 # uname -a
Linux imp 2.6.8-3-686 #1 Tue Dec 5 21:26:38 UTC 2006 i686 GNU/Linux


Is all the data lost, or can I recover from this?

Thanks so much!
Steve..
--
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html

[Index of Archives]     [Linux RAID Wiki]     [ATA RAID]     [Linux SCSI Target Infrastructure]     [Linux Block]     [Linux IDE]     [Linux SCSI]     [Linux Hams]     [Device Mapper]     [Device Mapper Cryptographics]     [Kernel]     [Linux Admin]     [Linux Net]     [GFS]     [RPM]     [git]     [Yosemite Forum]


  Powered by Linux