Raid failing, which command to remove the bad drive?

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



I have 4 drives set up as 2 pairs. The first part has 3 partitions on it and it seems 1 of those drives is failing (going to have to figure out which drive it is too so I don't pull the wrong one out of the case)

It's been awhile since I had to replace a drive in the array and my notes are a bit confusing. I'm not sure which I need to use to remove the drive:


	sudo mdadm --manage /dev/md0 --fail /dev/sdb
	sudo mdadm --manage /dev/md0 --remove /dev/sdb
	sudo mdadm --manage /dev/md1 --fail /dev/sdb
	sudo mdadm --manage /dev/md1 --remove /dev/sdb
	sudo mdadm --manage /dev/md2 --fail /dev/sdb
	sudo mdadm --manage /dev/md2 --remove /dev/sdb

or

sudo mdadm /dev/md0 --fail /dev/sdb1 --remove /dev/sdb1
sudo mdadm /dev/md1 --fail /dev/sdb2 --remove /dev/sdb2
sudo mdadm /dev/md2 --fail /dev/sdb3 --remove /dev/sdb3

I'm not sure if I fail the drive partition or whole drive for each.

-------------------------------------
The mails I got are:
-------------------------------------
A Fail event had been detected on md device /dev/md0.

It could be related to component device /dev/sdb1.

Faithfully yours, etc.

P.S. The /proc/mdstat file currently contains the following:

Personalities : [raid1] [raid6] [raid5] [raid4] [multipath]
md1 : active raid1 sdb2[2](F) sda2[0]
      4891712 blocks [2/1] [U_]

md2 : active raid1 sdb3[1] sda3[0]
      459073344 blocks [2/2] [UU]

md3 : active raid1 sdd1[1] sdc1[0]
      488383936 blocks [2/2] [UU]

md0 : active raid1 sdb1[2](F) sda1[0]
      24418688 blocks [2/1] [U_]

unused devices: <none>
-------------------------------------
A Fail event had been detected on md device /dev/md1.

It could be related to component device /dev/sdb2.

Faithfully yours, etc.

P.S. The /proc/mdstat file currently contains the following:

Personalities : [raid1] [raid6] [raid5] [raid4] [multipath]
md1 : active raid1 sdb2[2](F) sda2[0]
      4891712 blocks [2/1] [U_]

md2 : active raid1 sdb3[1] sda3[0]
      459073344 blocks [2/2] [UU]

md3 : active raid1 sdd1[1] sdc1[0]
      488383936 blocks [2/2] [UU]

md0 : active raid1 sdb1[2](F) sda1[0]
      24418688 blocks [2/1] [U_]

unused devices: <none>
-------------------------------------
A Fail event had been detected on md device /dev/md2.

It could be related to component device /dev/sdb3.

Faithfully yours, etc.

P.S. The /proc/mdstat file currently contains the following:

Personalities : [raid1] [raid6] [raid5] [raid4] [multipath]
md1 : active raid1 sdb2[2](F) sda2[0]
      4891712 blocks [2/1] [U_]

md2 : active raid1 sdb3[2](F) sda3[0]
      459073344 blocks [2/1] [U_]

md3 : active raid1 sdd1[1] sdc1[0]
      488383936 blocks [2/2] [UU]

md0 : active raid1 sdb1[2](F) sda1[0]
      24418688 blocks [2/1] [U_]

unused devices: <none>
-------------------------------------
--
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html


[Index of Archives]     [Linux RAID Wiki]     [ATA RAID]     [Linux SCSI Target Infrastructure]     [Linux Block]     [Linux IDE]     [Linux SCSI]     [Linux Hams]     [Device Mapper]     [Device Mapper Cryptographics]     [Kernel]     [Linux Admin]     [Linux Net]     [GFS]     [RPM]     [git]     [Yosemite Forum]


  Powered by Linux