''force'' continutation of a rebuild?

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hello all,

I have another seminewbie question.  I had an issue, likely hardware
related, which forced me to reboot a machine with a RAID6 during a
rebuild after a previous drive failure.  Now, after some other hardware
issues, I've been able to successfully assemble the array, but it
seems to be in an odd state:

# mdadm -D /dev/md0
/dev/md0:
        Version : 1.01
  Creation Time : Thu Sep 29 21:26:35 2011
     Raid Level : raid6
     Array Size : 13671797440 (13038.44 GiB 13999.92 GB)
  Used Dev Size : 1953113920 (1862.63 GiB 1999.99 GB)
   Raid Devices : 9
  Total Devices : 11
Preferred Minor : 0
    Persistence : Superblock is persistent

    Update Time : Thu Dec 15 12:19:41 2011
          State : clean, degraded
 Active Devices : 8
Working Devices : 11
 Failed Devices : 0
  Spare Devices : 3

     Chunk Size : 64K

           Name : 0
           UUID : 24363b01:90deb9b5:4b51e5df:68b8b6ea
         Events : 102730

    Number   Major   Minor   RaidDevice State
       0       8       17        0      active sync   /dev/sdb1
       6       8      113        1      active sync   /dev/sdh1
      11       8      177        2      spare rebuilding   /dev/sdl1
       3       8       65        3      active sync   /dev/sde1
       4       8       81        4      active sync   /dev/sdf1
       9       8      145        5      active sync   /dev/sdj1
      10       8       97        6      active sync   /dev/sdg1
       7       8      129        7      active sync   /dev/sdi1
       8       8      161        8      active sync   /dev/sdk1

      12       8      225        -      spare   /dev/sdo1
      13       8       49        -      spare   /dev/sdd1

# cat /proc/mdstat 
Personalities : [raid6] [raid5] [raid4] 
md0 : active raid6 sdd1[13](S) sdb1[0] sdo1[12](S) sdk1[8] sdi1[7]
sdg1[10] sdj1[9] sdf1[4] sde1[3] sdl1[11] sdh1[6]
      13671797440 blocks super 1.1 level 6, 64k chunk, algorithm 2 [9/8]
[UU_UUUUUU]
      
unused devices: <none>

I'm interpreting this as that a member is missing, but for some reason
the rebuild on sdl1 has not restarted.  What would be the next logical
step to take?  I've found some posts which imply that setting sync_action
to repair will work, but I'm a little wary of doing that without knowing
how risky that is.  Or, reading Documentation/md.txt, perhaps I should
set it to "recover"?  Or "resync", since it's possible the array was not
shut down cleanly?

FWIW, I have started the array, activated the LVM volume, and am running
xfs_repair -n (which is not supposed to do any writes), but otherwise
haven't risked modifying the filesystem (e.g., by mounting it).  So far
the xfs_repair seems fine, and has not reported any errors.

Thanks for your help (and patience).

--keith

-- 
kkeller@xxxxxxxxxxxxxxxxxxxxxxxxxx


--
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html


[Index of Archives]     [Linux RAID Wiki]     [ATA RAID]     [Linux SCSI Target Infrastructure]     [Linux Block]     [Linux IDE]     [Linux SCSI]     [Linux Hams]     [Device Mapper]     [Device Mapper Cryptographics]     [Kernel]     [Linux Admin]     [Linux Net]     [GFS]     [RPM]     [git]     [Yosemite Forum]


  Powered by Linux