Re: flagging a disk faulty during raid-5 reshape/grow operation

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi avoozl ;)

>From a message from Alex on 6/2/08:

If a drive failes during reshape the reshape will just continue.
The blocks which were on the failed drive are calculated from the the other
disks and writes to the failed disk are simply omitted.
The result is a raid5 with a failed drive.
You should get a new drive asap to restore the redundancy.
Also it's kinda important that you don't run 2.6.23 because it has a nasty bug
which would be triggered in this scenario.
The reshape probably increased in speed after the system was no longer actively
used and io bandwidth freed up.


So mark it as faulty and add a new RMA'd disk ASAP.

David

max power wrote:
> Hi list,
> 
> This morning, I tried to resize our raid-5 array from 3 to 4 disks.
> Immediately after issuing the 'mdadm --grow -n 4' command, I found out
> that the new disk is actually 'faulty'.  There are no unrecoverable
> errors yet, but my smartctl shows a lot of hardware ECC and seek
> errors going on. The read errors are causing the whole array to
> perform very poorly, and the resize is said to take another 2-3 days
> (according to mdstat).
> 
> Is it safe to mark the new drive as faulty before the reshape
> operation finishes, or is there a risk that this will corrupt the
> raid?
> 
> I'm using kernel version 2.6.24-21-generic #1 SMP Mon Aug 25 17:32:09
> UTC 2008 i686 GNU/Linux (Ubuntu Intrepid), and mdadm version mdadm -
> v2.6.7 - 6th June 2008.
> 
> output from mdadm  --detail:
> 
> /dev/md2:
>         Version : 00.91
>   Creation Time : Mon Nov 19 14:04:02 2007
>      Raid Level : raid5
>      Array Size : 535173120 (510.38 GiB 548.02 GB)
>   Used Dev Size : 267586560 (255.19 GiB 274.01 GB)
>    Raid Devices : 4
>   Total Devices : 4
> Preferred Minor : 2
>     Persistence : Superblock is persistent
> 
>     Update Time : Mon Nov  3 16:34:08 2008
>           State : clean, recovering
>  Active Devices : 4
> Working Devices : 4
>  Failed Devices : 0
>   Spare Devices : 0
> 
>          Layout : left-symmetric
>      Chunk Size : 64K
> 
>  Reshape Status : 1% complete
>   Delta Devices : 1, (3->4)
> 
>            UUID : 36b972e8:26bae0a4:de6b0f36:
> 485dee50
>          Events : 0.6496
> 
>     Number   Major   Minor   RaidDevice State
>        0       8        7        0      active sync   /dev/sda7
>        1       8       39        1      active sync   /dev/sdc7
>        2       8       55        2      active sync   /dev/sdd7
>        3       8       23        3      active sync   /dev/sdb7
> 
> 
> 
> Thanks in advance,
> Jorik
> --
> To unsubscribe from this list: send the line "unsubscribe linux-raid" in
> the body of a message to majordomo@xxxxxxxxxxxxxxx
> More majordomo info at  http://vger.kernel.org/majordomo-info.html


-- 
"Don't worry, you'll be fine; I saw it work in a cartoon once..."
--
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html

[Index of Archives]     [Linux RAID Wiki]     [ATA RAID]     [Linux SCSI Target Infrastructure]     [Linux Block]     [Linux IDE]     [Linux SCSI]     [Linux Hams]     [Device Mapper]     [Device Mapper Cryptographics]     [Kernel]     [Linux Admin]     [Linux Net]     [GFS]     [RPM]     [git]     [Yosemite Forum]


  Powered by Linux