Re: md RAID5: Disk wrongly marked "spare", need to force re-add it

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On 04/12/13 22:08, Ben Bucksch wrote:
I have a RAID5 with 8 disks. It worked fine.
Raid5 on 8 disks, EEP. Raid6! Really, raid6.

I made an update from ubuntu 10.04 to 12.04 using do-release-upgrade,
and rebooted.
before: kernel 2.6.32-41, mdadm v2.6.7.1, after: kernel 3.2.0-41, mdadm
3.2.5 - and I made a copy of the root partition before I updated, so I
can still boot both OSs

After I rebooted, one drive dropped from the array. I don't know why: it
was fine hardware-wise, and the same happened with my other RAID5 array,
where the dropped drive was fine, too. This seems like an MD or Ubuntu
bug to me.
I've seen that happen to, but with all disks marked as spare. This apparently could be race condition with udev etc.

You should have tried mdadm -A /dev/md0 and see if it would reassemble it.

So, I readded it and resynced. When the resync was around or over 80%
done, another harddrive failed (maybe because the resync is now using
all sectors the first time since a long time), this one with a real
hardware failure, so I had to permanently removed it. Now, the first
drive, which was dropped although it was fine, is marked as spare,
although it should have mostly good data on it. mdadm refuses to re-add
it, no matter what I try. mdadm 3.2.5 says "not possible" (see [1]),
while mdadm 2.5.7.1 re-adds it, but as spare, not as real member, so the
array still won't restart.

I need to forcibly re-add the drive that's marked as spare, because it
should have good data on it. I understand that some blocks may be out of
sync or corrupted, but the array has many TB and I want to get to the
rest of all the data that's still good. Even if I get 80% recovered,
that's still better than 0%.
Firstly, have you written anything TOO the array while resyncing? If not, chances are your array is in a reasonable shape still.

The 'spare' drive, I don't know what its status is. Theoretically, I would assume that the resync the data written to the disk is exactly the same as it was before, so keep that in mind as a last resort. But basically, you should ignore this drive, its data is not to be trusted.

Now the broken drive. Check your cables!! and run smartctl on it to give smart a chance to 'fix' the drive somewhat and check its status/health.

Now check the event count for all your drivers and compare. If the 'broken' drive is only a few off (1 or 2 I think i spotted below, try the following)

mdadm --run --force -A /dev/md0 /dev/sd[1-7] (leave out the earlier 'spare')

You should have your array back up and running. mount -o ro /md0 /mnt and copy anything you need off. IF can't recover any files (due to not having enough free space) and if smartctl said your broken drive was somewhat sane, you can try re-adding it again and hope it'll work this time. If it won't let you, mdadm --zero-superblock /dev/brokendisk.

But do try to copy the most important stuff off, continuing may make things worse.

If it fails again (at 80% because of hardware failure) you can't re-use the broken disk. It really is broken :p
Re-force the assembly as above and copy the rest, it's all you can do.

That said, if all else fails, your very last hope, is to not use the broken drive, and 'force' the above using the earlier marked spare. Maybe you can get more data off the array then.

After recovering your data and replacing your broken disk, make it an 8 disk raid6 instead! (or if you need the space, 10disk raid6. Raid5 while awesome, on big big arrays it's asking for trouble still.


NB 1:
Please do NOT respond with

  * restore backup - the backup doesn't have the new data, which I
    really need
  * re-"create" the array - unless you can give me the exact --create
    command that would recover it with data - other people tried this
    based on suggestions in forums and they lost all data


Appendix 1:

dmesg during resync:

...
[45345.341865] XFS (dm-13): xfs_imap_to_bp: xfs_trans_read_buf()
returned error 5.
[45345.341909] XFS (dm-13): metadata I/O error: block 0x45986fc0
("xfs_trans_read_buf") error 5 buf count 8192

(many times, about the same handful of blocks)

[45345.610858] XFS (dm-13): metadata I/O error: block 0x7a434f20
("xfs_trans_read_buf") error 5 buf count 4096

(many times, about the same handful of blocks)

(And then probably MD shut down the disk, and the RAID, and so all the
other filesystems went in the trash, too:)

[45353.184049] XFS (dm-10): xfs_log_force: error 5 returned.
[45472.694283] XFS (dm-8): metadata I/O error: block 0x29e89be8
("xfs_trans_read_buf") error 5 buf count 4096
[45473.504044] XFS (dm-10): xfs_log_force: error 5 returned.
[45485.246757] XFS (dm-8): metadata I/O error: block 0x19010c29
("xlog_iodone") error 5 buf count 2560
[45485.248966] XFS (dm-8): xfs_do_force_shutdown(0x2) called from line
1007 of file /build/buildd/linux-3.2.0/fs/xfs/xfs_log.c. Return address
= 0xffffffffa031ede1
[45485.249011] XFS (dm-8): Log I/O Error Detected.  Shutting down
filesystem
[45485.251126] XFS (dm-8): Please umount the filesystem and rectify the
problem(s)
[45503.584037] XFS (dm-10): xfs_log_force: error 5 returned.
[45514.848040] XFS (dm-8): xfs_log_force: error 5 returned.


Appendix 2:

1. During resync, i.e. after sdl was wrongly dropped, and before sdk
failed:

# cat /proc/mdstat
md0 : active raid5 sdl[8] sdp[5] sdq[7] sdk[1] sdj[0] sdo[4] sdn[6] sdm[3]
       6837337472 blocks level 5, 64k chunk, algorithm 2 [8/7] [UU_UUUUU]
       [>....................]  recovery =  0.1% (1075328/976762496)
finish=468.6min speed=34700K/sec

# mdadm --detail /dev/md0
/dev/md0:
         Version : 0.90
   Creation Time : Sun Mar 22 15:51:17 2009
      Raid Level : raid5
      Array Size : 6837337472 (6520.59 GiB 7001.43 GB)
   Used Dev Size : 976762496 (931.51 GiB 1000.20 GB)
    Raid Devices : 8
   Total Devices : 8
Preferred Minor : 0
     Persistence : Superblock is persistent

     Update Time : Thu Apr 11 16:59:55 2013
           State : clean, degraded, recovering
  Active Devices : 7
Working Devices : 8
  Failed Devices : 0
   Spare Devices : 1

          Layout : left-symmetric
      Chunk Size : 64K

  Rebuild Status : 0% complete

            UUID : c71c4168:de3a9b44:5ac2d0d1:4a2cd41c
          Events : 0.13240512

     Number   Major   Minor   RaidDevice State
        0       8      144        0      active sync   /dev/sdj
        1       8      160        1      active sync   /dev/sdk
        8       8      176        2      spare rebuilding /dev/sdl
        3       8      192        3      active sync   /dev/sdm
        4       8      224        4      active sync   /dev/sdo
        5       8      240        5      active sync   /dev/sdp
        6       8      208        6      active sync   /dev/sdn
        7      65        0        7      active sync   /dev/sdq

2. after sdk had a hardware fault during resync:

after resync, towards the end (between 78 and 100%), again:

md0 : active raid5 sdj[0] sdl[8](S) sdq[7] sdn[6] sdp[5] sdo[4] sdm[3]
sdk[9](F)
       6837337472 blocks level 5, 64k chunk, algorithm 2 [8/6] [U__UUUUU]

   *-disk:1
        description: ATA Disk
        product: WDC WD10EACS-00D
        vendor: Western Digital
        physical id: 0.1.0
        bus info: scsi@7:0.1.0
        logical name: /dev/sdk
        version: 1A01
        serial: WD-...8520
        size: 931GiB (1TB)
        capacity: 931GiB (1TB)
        capabilities: 15000rpm
        configuration: ansiversion=5

   *-disk:2
        description: ATA Disk
        product: WDC WD10EACS-00D
        vendor: Western Digital
        physical id: 0.2.0
        bus info: scsi@7:0.2.0
        logical name: /dev/sdl
        version: 1A01
        serial: WD-WCAU45964913
        size: 931GiB (1TB)
        capacity: 931GiB (1TB)
        capabilities: 15000rpm
        configuration: ansiversion=5




3. Current state, after fix attempts:

(sdk has hardware failure
sdl is probably good, but marked spare)

# cat /proc/mdstat
md0 : inactive sdk[9](S) sdl[8](S) sdj[0] sdq[7] sdn[6] sdp[5] sdo[4]
sdm[3]
       7814099968 blocks

# mdadm --detail /dev/md0
/dev/md0:
         Version : 00.90
   Creation Time : Sun Mar 22 15:51:17 2009
      Raid Level : raid5
   Used Dev Size : 976762496 (931.51 GiB 1000.20 GB)
    Raid Devices : 8
   Total Devices : 8
Preferred Minor : 0
     Persistence : Superblock is persistent

     Update Time : Fri Apr 12 17:09:30 2013
           State : active, degraded, Not Started
  Active Devices : 6
Working Devices : 8
  Failed Devices : 0
   Spare Devices : 2

          Layout : left-symmetric
      Chunk Size : 64K

            UUID : c71c4168:de3a9b44:5ac2d0d1:4a2cd41c
          Events : 0.13274865

     Number   Major   Minor   RaidDevice State
        0       8      144        0      active sync   /dev/sdj
        1       0        0        1      removed
        2       0        0        2      removed
        3       8      192        3      active sync   /dev/sdm
        4       8      224        4      active sync   /dev/sdo
        5       8      240        5      active sync   /dev/sdp
        6       8      208        6      active sync   /dev/sdn
        7      65        0        7      active sync   /dev/sdq

        8       8      176        -      spare   /dev/sdl
        9       8      160        -      spare   /dev/sdk

# mdadm -E /dev/sd[jlmnopqk]

(sdl is the one I need to add:)
/dev/sdl:
           Magic : a92b4efc
         Version : 00.90.00
            UUID : c71c4168:de3a9b44:5ac2d0d1:4a2cd41c
   Creation Time : Sun Mar 22 15:51:17 2009
      Raid Level : raid5
   Used Dev Size : 976762496 (931.51 GiB 1000.20 GB)
      Array Size : 6837337472 (6520.59 GiB 7001.43 GB)
    Raid Devices : 8
   Total Devices : 7
Preferred Minor : 0

     Update Time : Fri Apr 12 15:00:38 2013
           State : clean
  Active Devices : 6
Working Devices : 7
  Failed Devices : 2
   Spare Devices : 1
        Checksum : ca6e81a9 - correct
          Events : 13274863

          Layout : left-symmetric
      Chunk Size : 64K

       Number   Major   Minor   RaidDevice State
this     8       8      176        8      spare   /dev/sdl

    0     0       8      144        0      active sync   /dev/sdj
    1     1       0        0        1      faulty removed
    2     2       0        0        2      faulty removed
    3     3       8      192        3      active sync   /dev/sdm
    4     4       8      224        4      active sync   /dev/sdo
    5     5       8      240        5      active sync   /dev/sdp
    6     6       8      208        6      active sync   /dev/sdn
    7     7      65        0        7      active sync   /dev/sdq
    8     8       8      176        8      spare   /dev/sdl

/dev/sdj:
           Magic : a92b4efc
         Version : 00.90.00
            UUID : c71c4168:de3a9b44:5ac2d0d1:4a2cd41c
   Creation Time : Sun Mar 22 15:51:17 2009
      Raid Level : raid5
   Used Dev Size : 976762496 (931.51 GiB 1000.20 GB)
      Array Size : 6837337472 (6520.59 GiB 7001.43 GB)
    Raid Devices : 8
   Total Devices : 6
Preferred Minor : 0

     Update Time : Fri Apr 12 17:09:30 2013
           State : clean
  Active Devices : 6
Working Devices : 6
  Failed Devices : 2
   Spare Devices : 0
        Checksum : c9a40ffb - correct
          Events : 13274865

          Layout : left-symmetric
      Chunk Size : 64K

       Number   Major   Minor   RaidDevice State
this     0       8      144        0      active sync   /dev/sdj

    0     0       8      144        0      active sync   /dev/sdj
    1     1       0        0        1      faulty removed
    2     2       0        0        2      faulty removed
    3     3       8      192        3      active sync   /dev/sdm
    4     4       8      224        4      active sync   /dev/sdo
    5     5       8      240        5      active sync   /dev/sdp
    6     6       8      208        6      active sync   /dev/sdn
    7     7      65        0        7      active sync   /dev/sdq

/dev/sdm:
           Magic : a92b4efc
         Version : 00.90.00
            UUID : c71c4168:de3a9b44:5ac2d0d1:4a2cd41c
   Creation Time : Sun Mar 22 15:51:17 2009
      Raid Level : raid5
   Used Dev Size : 976762496 (931.51 GiB 1000.20 GB)
      Array Size : 6837337472 (6520.59 GiB 7001.43 GB)
    Raid Devices : 8
   Total Devices : 6
Preferred Minor : 0

     Update Time : Fri Apr 12 17:09:30 2013
           State : active
  Active Devices : 6
Working Devices : 6
  Failed Devices : 2
   Spare Devices : 0
        Checksum : c9a41030 - correct
          Events : 13274865

          Layout : left-symmetric
      Chunk Size : 64K

       Number   Major   Minor   RaidDevice State
this     3       8      192        3      active sync   /dev/sdm

    0     0       8      144        0      active sync   /dev/sdj
    1     1       0        0        1      faulty removed
    2     2       0        0        2      faulty removed
    3     3       8      192        3      active sync   /dev/sdm
    4     4       8      224        4      active sync   /dev/sdo
    5     5       8      240        5      active sync   /dev/sdp
    6     6       8      208        6      active sync   /dev/sdn
    7     7      65        0        7      active sync   /dev/sdq

/dev/sdn:
           Magic : a92b4efc
         Version : 00.90.00
            UUID : c71c4168:de3a9b44:5ac2d0d1:4a2cd41c
   Creation Time : Sun Mar 22 15:51:17 2009
      Raid Level : raid5
   Used Dev Size : 976762496 (931.51 GiB 1000.20 GB)
      Array Size : 6837337472 (6520.59 GiB 7001.43 GB)
    Raid Devices : 8
   Total Devices : 6
Preferred Minor : 0

     Update Time : Fri Apr 12 17:09:30 2013
           State : active
  Active Devices : 6
Working Devices : 6
  Failed Devices : 2
   Spare Devices : 0
        Checksum : c9a41046 - correct
          Events : 13274865

          Layout : left-symmetric
      Chunk Size : 64K

       Number   Major   Minor   RaidDevice State
this     6       8      208        6      active sync   /dev/sdn

    0     0       8      144        0      active sync   /dev/sdj
    1     1       0        0        1      faulty removed
    2     2       0        0        2      faulty removed
    3     3       8      192        3      active sync   /dev/sdm
    4     4       8      224        4      active sync   /dev/sdo
    5     5       8      240        5      active sync   /dev/sdp
    6     6       8      208        6      active sync   /dev/sdn
    7     7      65        0        7      active sync   /dev/sdq

/dev/sdo:
           Magic : a92b4efc
         Version : 00.90.00
            UUID : c71c4168:de3a9b44:5ac2d0d1:4a2cd41c
   Creation Time : Sun Mar 22 15:51:17 2009
      Raid Level : raid5
   Used Dev Size : 976762496 (931.51 GiB 1000.20 GB)
      Array Size : 6837337472 (6520.59 GiB 7001.43 GB)
    Raid Devices : 8
   Total Devices : 6
Preferred Minor : 0

     Update Time : Fri Apr 12 17:09:30 2013
           State : active
  Active Devices : 6
Working Devices : 6
  Failed Devices : 2
   Spare Devices : 0
        Checksum : c9a41052 - correct
          Events : 13274865

          Layout : left-symmetric
      Chunk Size : 64K

       Number   Major   Minor   RaidDevice State
this     4       8      224        4      active sync   /dev/sdo

    0     0       8      144        0      active sync   /dev/sdj
    1     1       0        0        1      faulty removed
    2     2       0        0        2      faulty removed
    3     3       8      192        3      active sync   /dev/sdm
    4     4       8      224        4      active sync   /dev/sdo
    5     5       8      240        5      active sync   /dev/sdp
    6     6       8      208        6      active sync   /dev/sdn
    7     7      65        0        7      active sync   /dev/sdq

/dev/sdp:
           Magic : a92b4efc
         Version : 00.90.00
            UUID : c71c4168:de3a9b44:5ac2d0d1:4a2cd41c
   Creation Time : Sun Mar 22 15:51:17 2009
      Raid Level : raid5
   Used Dev Size : 976762496 (931.51 GiB 1000.20 GB)
      Array Size : 6837337472 (6520.59 GiB 7001.43 GB)
    Raid Devices : 8
   Total Devices : 6
Preferred Minor : 0

     Update Time : Fri Apr 12 17:09:30 2013
           State : active
  Active Devices : 6
Working Devices : 6
  Failed Devices : 2
   Spare Devices : 0
        Checksum : c9a41064 - correct
          Events : 13274865

          Layout : left-symmetric
      Chunk Size : 64K

       Number   Major   Minor   RaidDevice State
this     5       8      240        5      active sync   /dev/sdp

    0     0       8      144        0      active sync   /dev/sdj
    1     1       0        0        1      faulty removed
    2     2       0        0        2      faulty removed
    3     3       8      192        3      active sync   /dev/sdm
    4     4       8      224        4      active sync   /dev/sdo
    5     5       8      240        5      active sync   /dev/sdp
    6     6       8      208        6      active sync   /dev/sdn
    7     7      65        0        7      active sync   /dev/sdq

/dev/sdq:
           Magic : a92b4efc
         Version : 00.90.00
            UUID : c71c4168:de3a9b44:5ac2d0d1:4a2cd41c
   Creation Time : Sun Mar 22 15:51:17 2009
      Raid Level : raid5
   Used Dev Size : 976762496 (931.51 GiB 1000.20 GB)
      Array Size : 6837337472 (6520.59 GiB 7001.43 GB)
    Raid Devices : 8
   Total Devices : 6
Preferred Minor : 0

     Update Time : Fri Apr 12 17:09:30 2013
           State : active
  Active Devices : 6
Working Devices : 6
  Failed Devices : 2
   Spare Devices : 0
        Checksum : c9a40fb1 - correct
          Events : 13274865

          Layout : left-symmetric
      Chunk Size : 64K

       Number   Major   Minor   RaidDevice State
this     7      65        0        7      active sync   /dev/sdq

    0     0       8      144        0      active sync   /dev/sdj
    1     1       0        0        1      faulty removed
    2     2       0        0        2      faulty removed
    3     3       8      192        3      active sync   /dev/sdm
    4     4       8      224        4      active sync   /dev/sdo
    5     5       8      240        5      active sync   /dev/sdp
    6     6       8      208        6      active sync   /dev/sdn
    7     7      65        0        7      active sync   /dev/sdq



(sdk with serial ...8520 had the hardware fault:)

/dev/sdk:
           Magic : a92b4efc
         Version : 00.90.00
            UUID : c71c4168:de3a9b44:5ac2d0d1:4a2cd41c
   Creation Time : Sun Mar 22 15:51:17 2009
      Raid Level : raid5
   Used Dev Size : 976762496 (931.51 GiB 1000.20 GB)
      Array Size : 6837337472 (6520.59 GiB 7001.43 GB)
    Raid Devices : 8
   Total Devices : 6
Preferred Minor : 0

     Update Time : Fri Apr 12 17:09:30 2013
           State : clean
  Active Devices : 6
Working Devices : 6
  Failed Devices : 2
   Spare Devices : 0
        Checksum : c9a410bf - correct
          Events : 13274865

          Layout : left-symmetric
      Chunk Size : 64K

       Number   Major   Minor   RaidDevice State
this     9       8      160       -1      spare   /dev/sdk

    0     0       8      144        0      active sync   /dev/sdj
    1     1       0        0        1      faulty removed
    2     2       0        0        2      faulty removed
    3     3       8      192        3      active sync   /dev/sdm
    4     4       8      224        4      active sync   /dev/sdo
    5     5       8      240        5      active sync   /dev/sdp
    6     6       8      208        6      active sync   /dev/sdn
    7     7      65        0        7      active sync   /dev/sdq


NB 2:
The original problem is that md dropped a perfectly good drive from the
array, just because I upgraded the OS. It seems to me that Linux MD is
all too happy and quick to kick out drives from the array, and then
refuses to readd them without a resync. This might be fine approach on
paper, but not in reality, where the resync is probably when another
drive fails, and then you have no parity left and you're told that your
data is gone.
1. It shouldn't drop drives so quickly
2. It should allow me to re-add them, if I think the data is good
3. There must be a recovery mechanism, to at least partially recover
data. Arrays can easily have 10+ TB, and just because a few
blocks/sectors in one filesystem are bad doesn't mean that I need to
throw away all filesystems that are on that LVM, and all data in the
broken filesystem.


NB 3: Seems like other people have the exact same problem:
http://www.linuxquestions.org/questions/linux-server-73/mdadm-re-added-disk-treated-as-spare-750739/

http://forums.gentoo.org/viewtopic-t-716757.html
https://raid.wiki.kernel.org/index.php/RAID_Recovery#Recreating_an_array

NB 4: Last time I upgraded the OS on the RAID server, I ended up with a
similar mess, due to another md bug:
https://bugs.launchpad.net/ubuntu/+source/mdadm/+bug/136252 )


--
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html

--
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html




[Index of Archives]     [Linux RAID Wiki]     [ATA RAID]     [Linux SCSI Target Infrastructure]     [Linux Block]     [Linux IDE]     [Linux SCSI]     [Linux Hams]     [Device Mapper]     [Device Mapper Cryptographics]     [Kernel]     [Linux Admin]     [Linux Net]     [GFS]     [RPM]     [git]     [Yosemite Forum]


  Powered by Linux