Re: raid 1 errors then I failed and removed the drive. now cant tell which one it was?

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Tue Apr 16, 2013 at 12:27:42AM -0400, Mitchell Laks wrote:

> Hi,
> 
> I store lots of data on a raid1 created with mdadm on debian sid  using kernel 
> Linux  3.2.0-2-amd64 #1 SMP Fri Apr 6 05:01:55 UTC 2012 x86_64 GNU/Linux.
> 
> Now I was backing up the data from the raid to another external drive
> and the errors began
> 
> [730636.445918] ata1.00: error: { UNC }
> [730636.464576] ata1.00: configured for UDMA/33
> [730636.464584] ata1: EH complete
> [730638.110558] ata1.00: exception Emask 0x0 SAct 0x0 SErr 0x0 action 0x0
> [730638.115052] ata1.00: port_status 0x20200000
> [730638.119441] ata1.00: failed command: READ DMA
> [730638.123848] ata1.00: cmd c8/00:08:ef:9f:90/00:00:00:00:00/e1 tag 0 dma 4096 in
> [730638.123850]          res 51/40:00:f4:9f:90/40:00:01:00:00/e1 Emask 0x9 (media error)
> [730638.132821] ata1.00: status: { DRDY ERR }
> [730638.137305] ata1.00: error: { UNC }
> [730638.157256] ata1.00: configured for UDMA/33
> [730638.157262] ata1: EH complete
> [730639.802239] ata1.00: exception Emask 0x0 SAct 0x0 SErr 0x0 action 0x0
> [730639.806730] ata1.00: port_status 0x20200000
> [730639.811111] ata1.00: failed command: READ DMA
> [730639.815511] ata1.00: cmd c8/00:08:ef:9f:90/00:00:00:00:00/e1 tag 0 dma 4096 in
> [730639.815513]          res 51/40:00:f4:9f:90/40:00:01:00:00/e1 Emask 0x9 (media error)
> [730639.824457] ata1.00: status: { DRDY ERR }
> [730639.828930] ata1.00: error: { UNC }
> [730639.848936] ata1.00: configured for UDMA/33
> 
> they seemed to be coming from drive /dev/sda1 of the raid while
> /dev/sdb1 was ok
> so i did
> mdadm /dev/md0 -f /dev/sda1 
> mdadm /dev/md0 -r /dev/sda1
> 
> then I rsynced from the remaining drive to /dev/sdd1 an external
> drive. No more errors.
> 
> However I forgot to label the usable drive by creating a file on it or
> editing a file on it. 
> 
> But now I shut down and unplug one of the two drives, then run 
> mdadm -E /dev/sda1
> it seems to be the good (unfailed) drive
> 
> But similarly when  I unplug the other drive and put this one back
> i still get it listed as an unfailed drive
> 
> how can i figure out which is the failed drive and which is the
> remaining one????
> 
Normally, the event count and update time will indicate which was
failed, but if you've restarted with each drive in separately then this
may have updated both. The obvious way to check in this case would be to
do a read test of the drive (dd if=/dev/sda1 of=/dev/null bs=1M) or a
SMART test - if you get errors then it's the failed one.

Cheers,
    Robin
-- 
     ___        
    ( ' }     |       Robin Hill        <robin@xxxxxxxxxxxxxxx> |
   / / )      | Little Jim says ....                            |
  // !!       |      "He fallen in de water !!"                 |

Attachment: signature.asc
Description: Digital signature


[Index of Archives]     [Linux RAID Wiki]     [ATA RAID]     [Linux SCSI Target Infrastructure]     [Linux Block]     [Linux IDE]     [Linux SCSI]     [Linux Hams]     [Device Mapper]     [Device Mapper Cryptographics]     [Kernel]     [Linux Admin]     [Linux Net]     [GFS]     [RPM]     [git]     [Yosemite Forum]


  Powered by Linux