Weird, weird... I had some inkling that it may have been the
kernel/mdadm versions... I upgraded the whole thing to Ubuntu 12.04,
readded the second drive and they're now both active. For anyone out
there having this issue, try upgrading everything.
On 06/17/2012 01:13 AM, NeilBrown wrote:
On Fri, 15 Jun 2012 08:04:52 -0700 Roberto Leibman<roberto@xxxxxxxxxxx>
wrote:
I must be missing something completely obvious, but I've read the man
page, and went through the archive for this list.
One of the hard drives in my raid array failed... I have taken the hard
drive out, replaced it with a new one, copied the partition table (using
gdisk) and then added the drive to the raid array with:
mdadm --add /dev/md0 /dev/sdb3
I then monitor it with "mdadm --detail /dev/md0" or "cat /proc/mdstat"
until it synchronizes
After an ungodly number of hours, the thing finishes synchronizing, but
the new drive only shows up as a spare. So the RAID is still degraded....
The only explanation for this that I can think of is that the drive reported
an error near the end of the recovery process.
There could be some kernel bug, but you didn't say what kernel you are
running so it is hard to check.
I have not been able to get the new drive to become part of the array as
active, web searches have proved useless (people with the same problem
and no resolution). I've even failed/removed the active drive, at which
point the spare becomes active, but when I add the original drive it
still adds it as a spare)
That sounds wrong. If you have an array with one working drive and one
spare, and you fail the working drive, then you end up with no drive. There
is no way that the spare will suddenly become active.
Maybe you are misinterpreting something and thinking it is spare when it
isn't.
The below looks perfectly normal. What does it look like when the recovery
stops? Are there any messages in the kernel logs when it stops?
NeilBrown
So how do I make it active???
(it's in the middle of trying again, but here's what I have)
--------------
root@frogstar:~# cat /proc/mdstat
Personalities : [linear] [multipath] [raid0] [raid1] [raid6] [raid5]
[raid4] [raid10]
md0 : active raid1 sda3[2] sdb3[0]
1943454796 blocks super 1.2 [2/1] [U_]
[>....................] recovery = 1.0% (20096128/1943454796)
finish=737.0min speed=43493K/sec
unused devices:<none>
--------------
and
root@frogstar:~# mdadm --detail /dev/md0
/dev/md0:
Version : 1.2
Creation Time : Sat Apr 14 13:52:25 2012
Raid Level : raid1
Array Size : 1943454796 (1853.42 GiB 1990.10 GB)
Used Dev Size : 1943454796 (1853.42 GiB 1990.10 GB)
Raid Devices : 2
Total Devices : 2
Persistence : Superblock is persistent
Update Time : Thu Jun 14 13:13:54 2012
State : clean, degraded, recovering
Active Devices : 1
Working Devices : 2
Failed Devices : 0
Spare Devices : 1
Rebuild Status : 1% complete
Name : frogstar:0 (local to host frogstar)
UUID : 88ed6cd4:de463005:31ed764c:2b23a266
Events : 47610
Number Major Minor RaidDevice State
0 8 19 0 active sync /dev/sdb3
2 8 3 1 spare rebuilding /dev/sda3
The version of mdadm I'm using is the stock on ubuntu 10.10 (v3.1.4)
--
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at http://vger.kernel.org/majordomo-info.html
--
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at http://vger.kernel.org/majordomo-info.html