RAID6 reshape, 2 disk failures

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi list,

I started a reshape from 64K chunk size to 512K (now default IIRC).
During this time 2 disks failed with some time in between. The first
one was removed by MD, so I shut down and removed the HDD, continued
the reshape. After a while the second HDD failed. This is what it
looks liek right now, the second failed HDD still in as you can see:

 $ iostat -m
Linux 3.5.5-1-ck (ion)  10/16/2012      _x86_64_        (4 CPU)

avg-cpu:  %user   %nice %system %iowait  %steal   %idle
           8.93    7.81    5.40   15.57    0.00   62.28

Device:            tps    MB_read/s    MB_wrtn/s    MB_read    MB_wrtn
sda              38.93         0.00        13.09        939    8134936
sdb              59.37         5.19         2.60    3224158    1613418
sdf              59.37         5.19         2.60    3224136    1613418
sdc              59.37         5.19         2.60    3224134    1613418
sdd              59.37         5.19         2.60    3224151    1613418
sde              42.17         3.68         1.84    2289332    1145595
sdg              59.37         5.19         2.60    3224061    1613418
sdh               0.00         0.00         0.00          9          0
md0               0.06         0.00         0.00       2023          0
dm-0              0.06         0.00         0.00       2022          0

 $ cat /proc/mdstat
Personalities : [raid6] [raid5] [raid4]
md0 : active raid6 sde1[0](F) sdg1[8] sdc1[5] sdd1[3] sdb1[4] sdf1[9]
      9751756800 blocks super 1.2 level 6, 64k chunk, algorithm 2
[7/5] [_UUUUU_]
      [================>....]  reshape = 84.6% (1650786304/1950351360)
finish=2089.2min speed=2389K/sec

unused devices: <none>

 $ sudo mdadm -D /dev/md0
[sudo] password for x:
/dev/md0:
        Version : 1.2
  Creation Time : Tue Oct 19 08:58:41 2010
     Raid Level : raid6
     Array Size : 9751756800 (9300.00 GiB 9985.80 GB)
  Used Dev Size : 1950351360 (1860.00 GiB 1997.16 GB)
   Raid Devices : 7
  Total Devices : 6
    Persistence : Superblock is persistent

    Update Time : Tue Oct 16 23:55:28 2012
          State : clean, degraded, reshaping
 Active Devices : 5
Working Devices : 5
 Failed Devices : 1
  Spare Devices : 0

         Layout : left-symmetric
     Chunk Size : 64K

 Reshape Status : 84% complete
  New Chunksize : 512K

           Name : ion:0  (local to host ion)
           UUID : e6595c64:b3ae90b3:f01133ac:3f402d20
         Events : 8386010

    Number   Major   Minor   RaidDevice State
       0       8       65        0      faulty spare rebuilding   /dev/sde1
       9       8       81        1      active sync   /dev/sdf1
       4       8       17        2      active sync   /dev/sdb1
       3       8       49        3      active sync   /dev/sdd1
       5       8       33        4      active sync   /dev/sdc1
       8       8       97        5      active sync   /dev/sdg1
       6       0        0        6      removed


What is confusing to me is that /dev/sde1 (which is failing) is
currently marked as rebuilding. But when I check iostat, it's far
behind the other drives in total I/O since the reshape started, and
the I/O hasn't actually changed for a few hours. This together with _
instead of U leads me to believe that it's not actually being used. So
why does it say rebuilding?

I guess my question is if it's possible for me to remove the drive, or
would I mess the array up? I am not going to anything until the
reshape finishes though.

Thanks,
Mathias
--
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html


[Index of Archives]     [Linux RAID Wiki]     [ATA RAID]     [Linux SCSI Target Infrastructure]     [Linux Block]     [Linux IDE]     [Linux SCSI]     [Linux Hams]     [Device Mapper]     [Device Mapper Cryptographics]     [Kernel]     [Linux Admin]     [Linux Net]     [GFS]     [RPM]     [git]     [Yosemite Forum]


  Powered by Linux