Re: Raid failing, which command to remove the bad drive?

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On 26 August 2011 21:13, Timothy D. Lenz <tlenz@xxxxxxxxxx> wrote:
> I have 4 drives set up as 2 pairs.  The first part has 3 partitions on it
> and it seems 1 of those drives is failing (going to have to figure out which
> drive it is too so I don't pull the wrong one out of the case)
>
> It's been awhile since I had to replace a drive in the array and my notes
> are a bit confusing. I'm not sure which I need to use to remove the drive:
>
>
>        sudo mdadm --manage /dev/md0 --fail /dev/sdb
>        sudo mdadm --manage /dev/md0 --remove /dev/sdb
>        sudo mdadm --manage /dev/md1 --fail /dev/sdb
>        sudo mdadm --manage /dev/md1 --remove /dev/sdb
>        sudo mdadm --manage /dev/md2 --fail /dev/sdb
>        sudo mdadm --manage /dev/md2 --remove /dev/sdb
>
> or
>
> sudo mdadm /dev/md0 --fail /dev/sdb1 --remove /dev/sdb1
> sudo mdadm /dev/md1 --fail /dev/sdb2 --remove /dev/sdb2
> sudo mdadm /dev/md2 --fail /dev/sdb3 --remove /dev/sdb3
>
> I'm not sure if I fail the drive partition or whole drive for each.
>
> -------------------------------------
> The mails I got are:
> -------------------------------------
> A Fail event had been detected on md device /dev/md0.
>
> It could be related to component device /dev/sdb1.
>
> Faithfully yours, etc.
>
> P.S. The /proc/mdstat file currently contains the following:
>
> Personalities : [raid1] [raid6] [raid5] [raid4] [multipath]
> md1 : active raid1 sdb2[2](F) sda2[0]
>      4891712 blocks [2/1] [U_]
>
> md2 : active raid1 sdb3[1] sda3[0]
>      459073344 blocks [2/2] [UU]
>
> md3 : active raid1 sdd1[1] sdc1[0]
>      488383936 blocks [2/2] [UU]
>
> md0 : active raid1 sdb1[2](F) sda1[0]
>      24418688 blocks [2/1] [U_]
>
> unused devices: <none>
> -------------------------------------
> A Fail event had been detected on md device /dev/md1.
>
> It could be related to component device /dev/sdb2.
>
> Faithfully yours, etc.
>
> P.S. The /proc/mdstat file currently contains the following:
>
> Personalities : [raid1] [raid6] [raid5] [raid4] [multipath]
> md1 : active raid1 sdb2[2](F) sda2[0]
>      4891712 blocks [2/1] [U_]
>
> md2 : active raid1 sdb3[1] sda3[0]
>      459073344 blocks [2/2] [UU]
>
> md3 : active raid1 sdd1[1] sdc1[0]
>      488383936 blocks [2/2] [UU]
>
> md0 : active raid1 sdb1[2](F) sda1[0]
>      24418688 blocks [2/1] [U_]
>
> unused devices: <none>
> -------------------------------------
> A Fail event had been detected on md device /dev/md2.
>
> It could be related to component device /dev/sdb3.
>
> Faithfully yours, etc.
>
> P.S. The /proc/mdstat file currently contains the following:
>
> Personalities : [raid1] [raid6] [raid5] [raid4] [multipath]
> md1 : active raid1 sdb2[2](F) sda2[0]
>      4891712 blocks [2/1] [U_]
>
> md2 : active raid1 sdb3[2](F) sda3[0]
>      459073344 blocks [2/1] [U_]
>
> md3 : active raid1 sdd1[1] sdc1[0]
>      488383936 blocks [2/2] [UU]
>
> md0 : active raid1 sdb1[2](F) sda1[0]
>      24418688 blocks [2/1] [U_]
>
> unused devices: <none>
> -------------------------------------
> --
> To unsubscribe from this list: send the line "unsubscribe linux-raid" in
> the body of a message to majordomo@xxxxxxxxxxxxxxx
> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>

Looks like your sda is failing, that's the smartctl -a /dev/sda output?

/Mathias
--
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html


[Index of Archives]     [Linux RAID Wiki]     [ATA RAID]     [Linux SCSI Target Infrastructure]     [Linux Block]     [Linux IDE]     [Linux SCSI]     [Linux Hams]     [Device Mapper]     [Device Mapper Cryptographics]     [Kernel]     [Linux Admin]     [Linux Net]     [GFS]     [RPM]     [git]     [Yosemite Forum]


  Powered by Linux