Re: drives failed during reshape, array won't even force-assemble

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Holy cow, I poked it with a stick and I think I did it!

As I've wrote before after a reboot the array was there but didn't start, and I've noticed the same thing happend with the overlay files right after I created them:

# /cat /proc/mdstat
Personalities : [raid1] [raid6] [raid5] [raid4] [linear] [multipath] [raid0] [raid10] md1 : inactive dm-0[8](S) dm-1[6](S) dm-7[4](S) dm-6[2](S) dm-5[0](S) dm-3[1](S) dm-4[5](S) dm-2[3](S)
      23429580800 blocks super 0.91

# mdadm --detail /dev/md1
/dev/md1:
        Version : 0.91
     Raid Level : raid0
  Total Devices : 8
Preferred Minor : 0
    Persistence : Superblock is persistent

          State : inactive

      New Level : raid6
     New Layout : left-symmetric
  New Chunksize : 64K

           UUID : 7a58ed4f:baf1934e:a2963c6e:a542ed71
         Events : 0.12370980

    Number   Major   Minor   RaidDevice

       -     252        0        -        /dev/dm-0
       -     252        1        -        /dev/dm-1
       -     252        2        -        /dev/dm-2
       -     252        3        -        /dev/dm-3
       -     252        4        -        /dev/dm-4
       -     252        5        -        /dev/dm-5
       -     252        6        -        /dev/dm-6
       -     252        7        -        /dev/dm-7

	
Now I tried

# mdadm --run /dev/md1
mdadm: failed to start array /dev/md1: Input/output error


and something interesting happend:

# mdadm --detail /dev/md1
/dev/md1:
        Version : 0.91
  Creation Time : Thu Sep  1 22:23:00 2011
     Raid Level : raid6
  Used Dev Size : 18446744073709551615
   Raid Devices : 7
  Total Devices : 6
Preferred Minor : 1
    Persistence : Superblock is persistent

    Update Time : Tue Jan 24 21:10:19 2017
          State : active, FAILED, Not Started
 Active Devices : 4
Working Devices : 6
 Failed Devices : 0
  Spare Devices : 2

         Layout : left-symmetric-6
     Chunk Size : 64K

     New Layout : left-symmetric

           UUID : 7a58ed4f:baf1934e:a2963c6e:a542ed71
         Events : 0.12370980

    Number   Major   Minor   RaidDevice State
       0     252        5        0      active sync   /dev/dm-5
       1     252        3        1      active sync   /dev/dm-3
       2     252        6        2      active sync   /dev/dm-6
       3     252        2        3      active sync   /dev/dm-2
       -       0        0        4      removed
       -       0        0        5      removed
       6     252        1        6      spare rebuilding   /dev/dm-1

       8     252        0        -      spare   /dev/dm-0
	
	
let's try to add the missing drives:

# mdadm --manage /dev/md1 --add /dev/mapper/sdc3
mdadm: re-added /dev/mapper/sdc3
	
# mdadm --manage /dev/md1 --add /dev/mapper/sdd3
mdadm: re-added /dev/mapper/sdd3
	
# mdadm --detail /dev/md1
detail /dev/md1
/dev/md1:
        Version : 0.91
  Creation Time : Thu Sep  1 22:23:00 2011
     Raid Level : raid6
  Used Dev Size : 18446744073709551615
   Raid Devices : 7
  Total Devices : 8
Preferred Minor : 1
    Persistence : Superblock is persistent

    Update Time : Tue Jan 24 21:10:19 2017
          State : active, degraded, Not Started
 Active Devices : 6
Working Devices : 8
 Failed Devices : 0
  Spare Devices : 2

         Layout : left-symmetric-6
     Chunk Size : 64K

     New Layout : left-symmetric

           UUID : 7a58ed4f:baf1934e:a2963c6e:a542ed71
         Events : 0.12370980

    Number   Major   Minor   RaidDevice State
       0     252        5        0      active sync   /dev/dm-5
       1     252        3        1      active sync   /dev/dm-3
       2     252        6        2      active sync   /dev/dm-6
       3     252        2        3      active sync   /dev/dm-2
       4     252        7        4      active sync   /dev/dm-7
       5     252        4        5      active sync   /dev/dm-4
       6     252        1        6      spare rebuilding   /dev/dm-1

       8     252        0        -      spare   /dev/dm-0
	

Not bad at all! But it still won't start, even with --run. Maybe if I wait long enough for the rebuild to finish? But I still don't see it in /proc/mdstat and I don't want to wait for several days to see if it really rebuilds in the background.

So I poke it with a stick...

# echo "clean" > /sys/block/md1/md/array_state
-bash: echo: write error: Invalid argument	

nope

# echo "active" > /sys/block/md1/md/array_state
-bash: echo: write error: Invalid argument	

nope

# echo "readonly" > /sys/block/md1/md/array_state

wait, no error?

# cat /proc/mdstat
Personalities : [raid1] [raid6] [raid5] [raid4] [linear] [multipath] [raid0] [raid10] md1 : active (read-only) raid6 dm-0[5] dm-2[4] dm-7[6] dm-6[3] dm-4[0] dm-1[2] dm-5[1] dm-3[8](S) 14643488000 blocks super 0.91 level 6, 64k chunk, algorithm 18 [7/6] [UUUUUU_]
      resync=PENDING
      bitmap: 175/175 pages [700KB], 8192KB chunk

# mdadm --detail /dev/md1
/dev/md1:
        Version : 0.91
  Creation Time : Thu Sep  1 22:23:00 2011
     Raid Level : raid6
     Array Size : 14643488000 (13965.12 GiB 14994.93 GB)
  Used Dev Size : 18446744073709551615
   Raid Devices : 7
  Total Devices : 8
Preferred Minor : 1
    Persistence : Superblock is persistent

  Intent Bitmap : Internal

    Update Time : Tue Jan 24 21:10:19 2017
          State : clean, degraded, resyncing (PENDING)
 Active Devices : 6
Working Devices : 8
 Failed Devices : 0
  Spare Devices : 2

         Layout : left-symmetric-6
     Chunk Size : 64K

     New Layout : left-symmetric

           UUID : 7a58ed4f:baf1934e:a2963c6e:a542ed71
         Events : 0.12370980

    Number   Major   Minor   RaidDevice State
       0     252        4        0      active sync   /dev/dm-4
       1     252        5        1      active sync   /dev/dm-5
       2     252        1        2      active sync   /dev/dm-1
       3     252        6        3      active sync   /dev/dm-6
       4     252        2        4      active sync   /dev/dm-2
       5     252        0        5      active sync   /dev/dm-0
       6     252        7        6      spare rebuilding   /dev/dm-7

       8     252        3        -      spare   /dev/dm-3


still no error
	
# echo "clean" > /sys/block/md1/md/array_state

# cat /proc/mdstat
Personalities : [raid1] [raid6] [raid5] [raid4] [linear] [multipath] [raid0] [raid10] md1 : active raid6 raid6 dm-0[5] dm-2[4] dm-7[6] dm-6[3] dm-4[0] dm-1[2] dm-5[1] dm-3[8](S) 14643488000 blocks super 0.91 level 6, 64k chunk, algorithm 18 [7/6] [UUUUUU_] [==============>......] reshape = 74.6% (2185464448/2928697600) finish=7719.3min speed=1603K/sec
      bitmap: 175/175 pages [700KB], 8192KB chunk
14643488000 blocks super 0.91 level 6, 64k chunk, algorithm 18 [7/6] [UUUUUU_]
      resync=PENDING
      bitmap: 175/175 pages [700KB], 8192KB chunk

# mdadm --detail /dev/md1
/dev/md1:
        Version : 0.91
  Creation Time : Thu Sep  1 22:23:00 2011
     Raid Level : raid6
     Array Size : 14643488000 (13965.12 GiB 14994.93 GB)
  Used Dev Size : 18446744073709551615
   Raid Devices : 7
  Total Devices : 8
Preferred Minor : 1
    Persistence : Superblock is persistent

  Intent Bitmap : Internal

    Update Time : Tue Jan 31 20:09:30 2017
          State : clean, degraded, reshaping
 Active Devices : 6
Working Devices : 8
 Failed Devices : 0
  Spare Devices : 2

         Layout : left-symmetric-6
     Chunk Size : 64K

 Reshape Status : 74% complete
     New Layout : left-symmetric

           UUID : 7a58ed4f:baf1934e:a2963c6e:a542ed71
         Events : 0.12370982

    Number   Major   Minor   RaidDevice State
       0     252        4        0      active sync   /dev/dm-4
       1     252        5        1      active sync   /dev/dm-5
       2     252        1        2      active sync   /dev/dm-1
       3     252        6        3      active sync   /dev/dm-6
       4     252        2        4      active sync   /dev/dm-2
       5     252        0        5      active sync   /dev/dm-0
       6     252        7        6      spare rebuilding   /dev/dm-7

       8     252        3        -      spare   /dev/dm-3

	
Looks good! fsck shows no errors, nothing in lost+found, so I've stopped the reshape (so the overlays won't fill the disk), mounted it readonly and backed up the more important data. That finished today, so I rebooted and did it for real. Reshape is finished, resync at 24% (6 hours to go), fsck still looks good. w00t!

	
--
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html



[Index of Archives]     [Linux RAID Wiki]     [ATA RAID]     [Linux SCSI Target Infrastructure]     [Linux Block]     [Linux IDE]     [Linux SCSI]     [Linux Hams]     [Device Mapper]     [Device Mapper Cryptographics]     [Kernel]     [Linux Admin]     [Linux Net]     [GFS]     [RPM]     [git]     [Yosemite Forum]


  Powered by Linux