So it would appear the drive number has been remapped somehow on /dev/sdb1 to be 3 instead of 0? Is there a way to reset/set these to specific values? 3 8 17 - spare /dev/sdb1 Thanks Steve.. On Wed, Oct 22, 2008 at 2:52 PM, Steve Evans <jeeping@xxxxxxxxx> wrote: > > Hi all.. > > I had one of the disks in my 3 disk RAID5 die on me this week. When > attempting to replace the disk via a hot swap (USB), the RAID didn't > like it. It decided to mark one of my remaining 2 disks as faulty. > > Can someone *please* help me get the raid back!? > > More details - > > Drives are /dev/sdb1, /dev/sdc1 & /dev/sdd1 > > sdc1 was the one that died earlier this week > sdb1 appears to be the one that was marked as faulty > > mdadm detail before sdc1 was plugged in - > > root@imp[~]:11 # mdadm --detail /dev/md1 > /dev/md1: > Version : 00.90.01 > Creation Time : Fri Nov 17 21:28:44 2006 > Raid Level : raid5 > Array Size : 586067072 (558.92 GiB 600.13 GB) > Device Size : 293033536 (279.46 GiB 300.07 GB) > Raid Devices : 3 > Total Devices : 2 > Preferred Minor : 1 > Persistence : Superblock is persistent > > Update Time : Sat Oct 18 20:06:34 2008 > State : clean, degraded > Active Devices : 2 > Working Devices : 2 > Failed Devices : 0 > Spare Devices : 0 > > Layout : left-symmetric > Chunk Size : 64K > > UUID : bed40ee2:98523fdd:e4d010fb:894c0966 > Events : 0.1474312 > > Number Major Minor RaidDevice State > 0 8 17 0 active sync /dev/sdb1 > 1 0 0 - removed > 2 8 49 2 active sync /dev/sdd1 > > > then after plugging in the replacement sdc1 - > > root@imp[~]:13 # mdadm --add /dev/md1 /dev/sdc1 > mdadm: hot added /dev/sdc1 > root@imp[~]:14 # > root@imp[~]:14 # > root@imp[~]:14 # mdadm --detail /dev/md1 > /dev/md1: > Version : 00.90.01 > Creation Time : Fri Nov 17 21:28:44 2006 > Raid Level : raid5 > Array Size : 586067072 (558.92 GiB 600.13 GB) > Device Size : 293033536 (279.46 GiB 300.07 GB) > Raid Devices : 3 > Total Devices : 3 > Preferred Minor : 1 > Persistence : Superblock is persistent > > Update Time : Sat Oct 18 22:13:13 2008 > State : clean, degraded > Active Devices : 1 > Working Devices : 2 > Failed Devices : 1 > Spare Devices : 1 > > Layout : left-symmetric > Chunk Size : 64K > > UUID : bed40ee2:98523fdd:e4d010fb:894c0966 > Events : 0.1480366 > > Number Major Minor RaidDevice State > 0 0 0 - removed > 1 0 0 - removed > 2 8 49 2 active sync /dev/sdd1 > > 3 8 33 0 spare rebuilding /dev/sdc1 > 4 8 17 - faulty /dev/sdb1 > > Shortly after this, subsequent mdadm --details stopped responding.. So > I rebooted in the hope I could reset and problems with the hot add.. > > Now, I'm unable to assemble the raid with the 2 working drives - > > mdadm --assemble /dev/md1 /dev/sdb1 /dev/sdd1 > > doesn't work - > > mdadm: /dev/md1 assembled from 1 drive and 1 spare - not enough to > start the array. > > mdadm --assemble --force /dev/md1 /dev/sdb1 /dev/sdd1 > > doesn't' work either > > This - > > mdadm --assemble --force --run /dev/md1 /dev/sdb1 /dev/sdd1 > > Did work partially - > > /dev/md1: > Version : 00.90.01 > Creation Time : Fri Nov 17 21:28:44 2006 > Raid Level : raid5 > Device Size : 293033536 (279.46 GiB 300.07 GB) > Raid Devices : 3 > Total Devices : 2 > Preferred Minor : 1 > Persistence : Superblock is persistent > > Update Time : Sat Oct 18 22:14:48 2008 > State : active, degraded > Active Devices : 1 > Working Devices : 2 > Failed Devices : 0 > Spare Devices : 1 > > Layout : left-symmetric > Chunk Size : 64K > > UUID : bed40ee2:98523fdd:e4d010fb:894c0966 > Events : 0.1521614 > > Number Major Minor RaidDevice State > 0 0 0 - removed > 1 0 0 - removed > 2 8 49 2 active sync /dev/sdd1 > > 3 8 17 - spare /dev/sdb1 > > Here's the output from mdadm -E on each of the 2 drives - > > /dev/sdb1: > Magic : a92b4efc > Version : 00.90.00 > UUID : bed40ee2:98523fdd:e4d010fb:894c0966 > Creation Time : Fri Nov 17 21:28:44 2006 > Raid Level : raid5 > Raid Devices : 3 > Total Devices : 3 > Preferred Minor : 1 > > Update Time : Sat Oct 18 22:14:48 2008 > State : clean > Active Devices : 1 > Working Devices : 2 > Failed Devices : 2 > Spare Devices : 1 > Checksum : e6dbf75 - correct > Events : 0.1521614 > > Layout : left-symmetric > Chunk Size : 64K > > Number Major Minor RaidDevice State > this 3 8 33 3 spare /dev/sdc1 > > 0 0 0 0 0 removed > 1 1 0 0 1 faulty removed > 2 2 8 49 2 active sync /dev/sdd1 > 3 3 8 33 3 spare /dev/sdc1 > /dev/sdd1: > Magic : a92b4efc > Version : 00.90.00 > UUID : bed40ee2:98523fdd:e4d010fb:894c0966 > Creation Time : Fri Nov 17 21:28:44 2006 > Raid Level : raid5 > Raid Devices : 3 > Total Devices : 3 > Preferred Minor : 1 > > Update Time : Sat Oct 18 22:14:48 2008 > State : clean > Active Devices : 1 > Working Devices : 2 > Failed Devices : 2 > Spare Devices : 1 > Checksum : e6dbf86 - correct > Events : 0.1521614 > > Layout : left-symmetric > Chunk Size : 64K > > Number Major Minor RaidDevice State > this 2 8 49 2 active sync /dev/sdd1 > > 0 0 0 0 0 removed > 1 1 0 0 1 faulty removed > 2 2 8 49 2 active sync /dev/sdd1 > 3 3 8 33 0 spare /dev/sdc1 > > root@imp[~]:28 # mdadm --version > mdadm - v1.9.0 - 04 February 2005 > root@imp[~]:29 # uname -a > Linux imp 2.6.8-3-686 #1 Tue Dec 5 21:26:38 UTC 2006 i686 GNU/Linux > > > Is all the data lost, or can I recover from this? > > Thanks so much! > Steve.. -- To unsubscribe from this list: send the line "unsubscribe linux-raid" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html