Trouble recovering raid5 array

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi everyone,
I'm having some strange problems putting one of my raid5 backtogether. Here is back ground story:
I have 4 drives paritioned into a bunch of raid arrays. One of thedrives failed and I replaced it with a new one. I was able to getmdadm to recover all arrays, except one raid5 array. The array withtroubles is /dev/md8 and it is supposed to have /dev/sd[abcd]13 underit.
This command started the recovery process (same thing that worked formy other raid5 arrays):mdadm --manage --add /dev/md8 /dev/sdc13
md8 : active raid5 sdc13[4] sdd13[3] sdb13[1] sda13[0]      117185856 blocks level 5, 64k chunk, algorithm 2 [4/3] [UU_U]      [>....................]  recovery =  1.6% (634084/39061952)finish=12.1min speed=52840K/sec
However sometime after 1.6% into recovery, I did a "cat /proc/mdstat" and saw:
md8 : active raid5 sdc13[4](S) sdd13[3] sdb13[1] sda13[5](F)      117185856 blocks level 5, 64k chunk, algorithm 2 [4/2] [_U_U]
/dev/sda only "failed" for this array and not on any of the otherarrays. I proceeded trying to remove and re-add /dev/sda13 and/dev/sdc13, however that did not work. I ran the following:
# mdadm --examine /dev/sda13/dev/sda13:          Magic : a92b4efc        Version : 00.90.00           UUID : ec0fab43:4fb8d991:e6a58c12:482d89e4  Creation Time : Sat Sep 15 00:55:37 2007     Raid Level : raid5  Used Dev Size : 39061952 (37.25 GiB 40.00 GB)     Array Size : 117185856 (111.76 GiB 120.00 GB)   Raid Devices : 4  Total Devices : 4Preferred Minor : 8
    Update Time : Sat Feb 28 00:49:48 2009          State : clean Active Devices : 2Working Devices : 4 Failed Devices : 1  Spare Devices : 2       Checksum : a6b02b9d - correct         Events : 0.36
         Layout : left-symmetric     Chunk Size : 64K
      Number   Major   Minor   RaidDevice Statethis     4       8       13        4      spare   /dev/sda13
   0     0       0        0        0      removed   1     1       8       29        1      active sync   /dev/sdb13   2     2       0        0        2      faulty removed   3     3       8       61        3      active sync   /dev/sdd13   4     4       8       13        4      spare   /dev/sda13   5     5       8       45        5      spare   /dev/sdc13# mdadm --examine /dev/sdb13/dev/sdb13:          Magic : a92b4efc        Version : 00.90.00           UUID : ec0fab43:4fb8d991:e6a58c12:482d89e4  Creation Time : Sat Sep 15 00:55:37 2007     Raid Level : raid5  Used Dev Size : 39061952 (37.25 GiB 40.00 GB)     Array Size : 117185856 (111.76 GiB 120.00 GB)   Raid Devices : 4  Total Devices : 4Preferred Minor : 8
    Update Time : Sat Feb 28 00:49:48 2009          State : clean Active Devices : 2Working Devices : 4 Failed Devices : 1  Spare Devices : 2       Checksum : a6b02bad - correct         Events : 0.36
         Layout : left-symmetric     Chunk Size : 64K
      Number   Major   Minor   RaidDevice Statethis     1       8       29        1      active sync   /dev/sdb13
   0     0       0        0        0      removed   1     1       8       29        1      active sync   /dev/sdb13   2     2       0        0        2      faulty removed   3     3       8       61        3      active sync   /dev/sdd13   4     4       8       13        4      spare   /dev/sda13   5     5       8       45        5      spare   /dev/sdc13# mdadm --examine /dev/sdc13/dev/sdc13:          Magic : a92b4efc        Version : 00.90.00           UUID : ec0fab43:4fb8d991:e6a58c12:482d89e4  Creation Time : Sat Sep 15 00:55:37 2007     Raid Level : raid5  Used Dev Size : 39061952 (37.25 GiB 40.00 GB)     Array Size : 117185856 (111.76 GiB 120.00 GB)   Raid Devices : 4  Total Devices : 4Preferred Minor : 8
    Update Time : Sat Feb 28 00:49:48 2009          State : clean Active Devices : 2Working Devices : 4 Failed Devices : 1  Spare Devices : 2       Checksum : a6b02bbf - correct         Events : 0.36
         Layout : left-symmetric     Chunk Size : 64K
      Number   Major   Minor   RaidDevice Statethis     5       8       45        5      spare   /dev/sdc13
   0     0       0        0        0      removed   1     1       8       29        1      active sync   /dev/sdb13   2     2       0        0        2      faulty removed   3     3       8       61        3      active sync   /dev/sdd13   4     4       8       13        4      spare   /dev/sda13   5     5       8       45        5      spare   /dev/sdc13# mdadm --examine /dev/sdd13/dev/sdd13:          Magic : a92b4efc        Version : 00.90.00           UUID : ec0fab43:4fb8d991:e6a58c12:482d89e4  Creation Time : Sat Sep 15 00:55:37 2007     Raid Level : raid5  Used Dev Size : 39061952 (37.25 GiB 40.00 GB)     Array Size : 117185856 (111.76 GiB 120.00 GB)   Raid Devices : 4  Total Devices : 4Preferred Minor : 8
    Update Time : Sat Feb 28 00:49:48 2009          State : clean Active Devices : 2Working Devices : 4 Failed Devices : 1  Spare Devices : 2       Checksum : a6b02bd1 - correct         Events : 0.36
         Layout : left-symmetric     Chunk Size : 64K
      Number   Major   Minor   RaidDevice Statethis     3       8       61        3      active sync   /dev/sdd13
   0     0       0        0        0      removed   1     1       8       29        1      active sync   /dev/sdb13   2     2       0        0        2      faulty removed   3     3       8       61        3      active sync   /dev/sdd13   4     4       8       13        4      spare   /dev/sda13   5     5       8       45        5      spare   /dev/sdc13

At this point I started googling and ended up doing this:
# mdadm --stop /dev/md8# mdadm --zero-superblock /dev/sda13# mdadm --zero-superblock /dev/sdb13# mdadm --zero-superblock /dev/sdc13# mdadm --zero-superblock /dev/sdd13# mdadm -A /dev/md8 /dev/sda13 /dev/sdb13 /dev/sdc13 /dev/sdd13 --forcemdadm: no recogniseable superblock on /dev/sda13mdadm: /dev/sda13 has no superblock - assembly aborted# mdadm --create /dev/md8 -l 5 -n 4 /dev/sda13 /dev/sdb13 /dev/sdc13 /dev/sdd13mdadm: /dev/sda13 appears to contain an ext2fs file system    size=117185856K  mtime=Fri Feb 27 17:22:46 2009mdadm: /dev/sdd13 appears to contain an ext2fs file system    size=117185856K  mtime=Fri Feb 27 17:22:46 2009Continue creating array? ymdadm: array /dev/md8 started.# cat /proc/mdstatmd8 : active raid5 sdd13[4] sdc13[2] sdb13[1] sda13[0]      117185856 blocks level 5, 64k chunk, algorithm 2 [4/3] [UUU_]      [>....................]  recovery =  0.7% (294244/39061952)finish=10.9min speed=58848K/sec
Again very shortly after 1.6% into recovery, it failed. Now I see the following:
md8 : active raid5 sdd13[4](S) sdc13[2] sdb13[1] sda13[5](F)      117185856 blocks level 5, 64k chunk, algorithm 2 [4/2] [_UU_]
This surely does not look good. I hope that /dev/sda13 did not getwrongly synced. Does anyone have any suggestions about what I shoulddo to recover this array? Does anyone have any ideas on what couldhave possibly caused these issues? How can /dev/sda13, a healthy partof the array lose its superblock?
Let me know if you'd like any more info. mdadm version is 2.6.4.Kernel version 2.6.24.3.
Thanks in advance for any help,Andrey��.n��������+%������w��{.n�����{����w��ܨ}���Ơz�j:+v�����w����ޙ��&�)ߡ�a����z�ޗ���ݢj��w�f


[Index of Archives]     [Linux RAID Wiki]     [ATA RAID]     [Linux SCSI Target Infrastructure]     [Linux Block]     [Linux IDE]     [Linux SCSI]     [Linux Hams]     [Device Mapper]     [Device Mapper Cryptographics]     [Kernel]     [Linux Admin]     [Linux Net]     [GFS]     [RPM]     [git]     [Yosemite Forum]


  Powered by Linux