Re: Problems recovering from a raid1 failure

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Fri, Mar 12, 2010 at 12:17 AM, Michael Evans <mjevans1983@xxxxxxxxx> wrote:
> On Thu, Mar 11, 2010 at 11:51 PM, Jonathan Gordon
> <jonathan.kinobe@xxxxxxxxx> wrote:
>> Upon reboot, my machine began recovering from a raid1 failure.
>> Querying mdadm yielded the following:
>>
>> jgordon@kubuntu:~$ sudo mdadm --detail /dev/md0
>> [sudo] password for jgordon:
>> /dev/md0:
>>       Version : 00.90
>>  Creation Time : Mon Sep 11 06:35:17 2006
>>    Raid Level : raid1
>>    Array Size : 242187776 (230.97 GiB 248.00 GB)
>>  Used Dev Size : 242187776 (230.97 GiB 248.00 GB)
>>  Raid Devices : 2
>>  Total Devices : 2
>> Preferred Minor : 0
>>   Persistence : Superblock is persistent
>>
>>   Update Time : Thu Mar 11 18:09:25 2010
>>         State : clean, degraded, recovering
>>  Active Devices : 1
>> Working Devices : 2
>>  Failed Devices : 0
>>  Spare Devices : 1
>>
>>  Rebuild Status : 26% complete
>>
>>          UUID : 7fd22081:c39cb3e4:21109eec:10ecdf10
>>        Events : 0.5260272
>>
>>   Number   Major   Minor   RaidDevice State
>>      2       8        1        0      spare rebuilding   /dev/sda1
>>      1       8       17        1      active sync   /dev/sdb1
>>
>> After some time, the rebuild seemed to complete, but the State seemed
>> to switch alternately between "active, degraded" and "clean,
>> degraded". Addiontally, the state for /dev/sda1 seems to continue to
>> stay in "spare rebuilding". This is the current output:
>>
>> jgordon@kubuntu:~$ sudo mdadm -D /dev/md0
>> [sudo] password for jgordon:
>> /dev/md0:
>>       Version : 00.90
>>  Creation Time : Mon Sep 11 06:35:17 2006
>>    Raid Level : raid1
>>    Array Size : 242187776 (230.97 GiB 248.00 GB)
>>  Used Dev Size : 242187776 (230.97 GiB 248.00 GB)
>>  Raid Devices : 2
>>  Total Devices : 2
>> Preferred Minor : 0
>>   Persistence : Superblock is persistent
>>
>>   Update Time : Thu Mar 11 23:07:59 2010
>>         State : clean, degraded
>>  Active Devices : 1
>> Working Devices : 2
>>  Failed Devices : 0
>>  Spare Devices : 1
>>
>>          UUID : 7fd22081:c39cb3e4:21109eec:10ecdf10
>>        Events : 0.5273340
>>
>>   Number   Major   Minor   RaidDevice State
>>      2       8        1        0      spare rebuilding   /dev/sda1
>>      1       8       17        1      active sync   /dev/sdb1
>>
>> Additionally, /var/log/kern.log is getting filled with the following:
>>
>> Mar 11 19:19:14 jigme kernel: [ 6596.236366] ata4: EH complete
>> Mar 11 19:19:16 jigme kernel: [ 6598.104676] ata4.00: exception Emask
>> 0x0 SAct 0x0 SErr 0x0 action 0x0
>> Mar 11 19:19:16 jigme kernel: [ 6598.104683] ata4.00: BMDMA stat 0x24
>> Mar 11 19:19:16 jigme kernel: [ 6598.104692] ata4.00: cmd
>> 25/00:08:ff:b0:e0/00:00:15:00:00/e0 tag 0 dma 4096 in
>> Mar 11 19:19:16 jigme kernel: [ 6598.104694]          res
>> 51/40:00:04:b1:e0/40:00:15:00:00/e0 Emask 0x9 (media error)
>> Mar 11 19:19:16 jigme kernel: [ 6598.104698] ata4.00: status: { DRDY ERR }
>> Mar 11 19:19:16 jigme kernel: [ 6598.104702] ata4.00: error: { UNC }
>> Mar 11 19:19:16 jigme kernel: [ 6598.120352] ata4.00: configured for UDMA/133
>> Mar 11 19:19:16 jigme kernel: [ 6598.120371] sd 3:0:0:0: [sdb]
>> Unhandled sense code
>> Mar 11 19:19:16 jigme kernel: [ 6598.120375] sd 3:0:0:0: [sdb] Result:
>> hostbyte=DID_OK driverbyte=DRIVER_SENSE
>> Mar 11 19:19:16 jigme kernel: [ 6598.120380] sd 3:0:0:0: [sdb] Sense
>> Key : Medium Error [current] [descriptor]
>> Mar 11 19:19:16 jigme kernel: [ 6598.120388] Descriptor sense data
>> with sense descriptors (in hex):
>> Mar 11 19:19:16 jigme kernel: [ 6598.120392]         72 03 11 04 00 00
>> 00 0c 00 0a 80 00 00 00 00 00
>> Mar 11 19:19:16 jigme kernel: [ 6598.120412]         15 e0 b1 04
>> Mar 11 19:19:16 jigme kernel: [ 6598.120420] sd 3:0:0:0: [sdb] Add.
>> Sense: Unrecovered read error - auto reallocate failed
>> Mar 11 19:19:16 jigme kernel: [ 6598.120428] end_request: I/O error,
>> dev sdb, sector 367046916
>> Mar 11 19:19:16 jigme kernel: [ 6598.120446] ata4: EH complete
>> Mar 11 19:19:16 jigme kernel: [ 6598.120744] raid1: sdb: unrecoverable
>> I/O read error for block 367046784
>> Mar 11 19:19:17 jigme kernel: [ 6599.164052] md: md0: recovery done.
>> Mar 11 19:19:17 jigme kernel: [ 6599.460124] RAID1 conf printout:
>> Mar 11 19:19:17 jigme kernel: [ 6599.460145]  --- wd:1 rd:2
>> Mar 11 19:19:17 jigme kernel: [ 6599.460160]  disk 0, wo:1, o:1, dev:sda1
>> Mar 11 19:19:17 jigme kernel: [ 6599.460170]  disk 1, wo:0, o:1, dev:sdb1
>> Mar 11 19:19:17 jigme kernel: [ 6599.460178] RAID1 conf printout:
>> Mar 11 19:19:17 jigme kernel: [ 6599.460185]  --- wd:1 rd:2
>> Mar 11 19:19:17 jigme kernel: [ 6599.460195]  disk 0, wo:1, o:1, dev:sda1
>> Mar 11 19:19:17 jigme kernel: [ 6599.460204]  disk 1, wo:0, o:1, dev:sdb1
>> Mar 11 19:19:22 jigme kernel: [ 6604.165111] RAID1 conf printout:
>> Mar 11 19:19:22 jigme kernel: [ 6604.165117]  --- wd:1 rd:2
>> Mar 11 19:19:22 jigme kernel: [ 6604.165122]  disk 0, wo:1, o:1, dev:sda1
>> Mar 11 19:19:22 jigme kernel: [ 6604.165125]  disk 1, wo:0, o:1, dev:sdb1
>> Mar 11 19:19:22 jigme kernel: [ 6604.165128] RAID1 conf printout:
>> Mar 11 19:19:22 jigme kernel: [ 6604.165131]  --- wd:1 rd:2
>> Mar 11 19:19:22 jigme kernel: [ 6604.165134]  disk 0, wo:1, o:1, dev:sda1
>> Mar 11 19:19:22 jigme kernel: [ 6604.165137]  disk 1, wo:0, o:1, dev:sdb1
>> ...
>> Mar 11 23:16:28 jigme kernel: [20830.889380] RAID1 conf printout:
>> Mar 11 23:16:28 jigme kernel: [20830.889386]  --- wd:1 rd:2
>> Mar 11 23:16:28 jigme kernel: [20830.889391]  disk 0, wo:1, o:1, dev:sda1
>> Mar 11 23:16:28 jigme kernel: [20830.889394]  disk 1, wo:0, o:1, dev:sdb1
>> Mar 11 23:16:28 jigme kernel: [20830.889397] RAID1 conf printout:
>> Mar 11 23:16:28 jigme kernel: [20830.889399]  --- wd:1 rd:2
>> Mar 11 23:16:28 jigme kernel: [20830.889403]  disk 0, wo:1, o:1, dev:sda1
>> Mar 11 23:16:28 jigme kernel: [20830.889406]  disk 1, wo:0, o:1, dev:sdb1
>>
>> The "RAID1 conf printout:" messages appear every few seconds or so.
>>
>> Machine info:
>>
>> jgordon@kubuntu:~$ uname -a
>> Linux kubuntu 2.6.31-20-386 #57-Ubuntu SMP Mon Feb 8 11:42:49 UTC 2010
>> i686 GNU/Linux
>>
>> Any idea what I can do to resolve this?
>>
>> Thanks!
>> --
>> To unsubscribe from this list: send the line "unsubscribe linux-raid" in
>> the body of a message to majordomo@xxxxxxxxxxxxxxx
>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>>
>
> Replace your failing disk; from the look of the kernel log and the
> description of the issue I'd say your drive is out of spare sectors
> and would fail a S.M.A.R.T. test.
>
> If you require more proof start reading up on how to use the smartctl
> command from the smartmontools package (may have dashes/etc in your
> package manager).
>
> http://sourceforge.net/apps/trac/smartmontools/wiki/TocDoc
>

Reading more carefully, I notice you're in a most unfortunate
situation.  It's failing while trying to READ your 'ACTIVE' disc.

You should buy two NEW drives, attempt to copy the current active
member of the array to a NEW drive using something like
http://www.gnu.org/software/ddrescue/ddrescue.html

For anything that won't copy you /MIGHT/ try the other member of the
array if it used to be in place.  Chances are good that stale data
from a section MAY be better than no data at all.  You MUST perform an
FSCK prior to mounting the newly re-created array; you should try to
determine which files occupy any 'recovered' sectors and validate that
they are intact or replace them from other duplicates.

How to do that depends very much on what filesystem you have on the array.
--
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html

[Index of Archives]     [Linux RAID Wiki]     [ATA RAID]     [Linux SCSI Target Infrastructure]     [Linux Block]     [Linux IDE]     [Linux SCSI]     [Linux Hams]     [Device Mapper]     [Device Mapper Cryptographics]     [Kernel]     [Linux Admin]     [Linux Net]     [GFS]     [RPM]     [git]     [Yosemite Forum]


  Powered by Linux