Re: raid5 with 2 disks not working correctly. Need help!

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi again,

Can anybody please help me out here?

Roy

On Tue, Dec 16, 2008 at 2:36 PM, Roy Lim <cableroy@xxxxxxxxx> wrote:
> Hi
> I have ran into a serious problem with my raid5 array and need expert
> help on this one.
>
> My array is with 5x500GB disks none spare, I've accidentally did a
> unclean shutdown of the server. When it came up it gave me this error:
>
> Dec 15 09:07:11 ares kernel: md: kicking non-fresh hdk1 from array!
> Dec 15 09:07:11 ares kernel: md: unbind<hdk1>
> Dec 15 09:07:11 ares kernel: md: export_rdev(hdk1)
> Dec 15 09:07:11 ares kernel: md: md0: raid array is not clean --
> starting background reconstruction
> Dec 15 09:07:11 ares kernel: raid5: device hde1 operational as raid disk 0
> Dec 15 09:07:11 ares kernel: raid5: device sda1 operational as raid disk 4
> Dec 15 09:07:11 ares kernel: raid5: device hdi1 operational as raid disk 2
> Dec 15 09:07:11 ares kernel: raid5: device hdg1 operational as raid disk 1
> Dec 15 09:07:11 ares kernel: raid5: cannot start dirty degraded array for md0
>
> However i managed to get this back up with remove and add, the system
> was rebuilding the array. around 70-80% in the process i was going to
> prepare to decrypt it (luks crypted) so I'm going to mount the usb
> stick witch has the encryption key however when i was going to mount
> it i tried to mount wrong device, resulting trying to mount a device
> in the array (sda1) i hit repeatedly crtl+c, now taking a look at the
> details on the array it looks like this:
>
> /dev/md0:
>        Version : 00.90.03
>  Creation Time : Mon Feb  4 18:25:28 2008
>     Raid Level : raid5
>     Array Size : 1953535744 (1863.04 GiB 2000.42 GB)
>    Device Size : 488383936 (465.76 GiB 500.11 GB)
>   Raid Devices : 5
>  Total Devices : 5
> Preferred Minor : 0
>    Persistence : Superblock is persistent
>
>    Update Time : Mon Dec 15 16:02:54 2008
>          State : clean, degraded
>  Active Devices : 3
> Working Devices : 4
>  Failed Devices : 1
>  Spare Devices : 1
>
>         Layout : left-symmetric
>     Chunk Size : 64K
>
>           UUID : 856a6e4e:a1663e98:9efbd2b1:7507133b
>         Events : 0.74
>
>    Number   Major   Minor   RaidDevice State
>       0      33        1        0      active sync   /dev/hde1
>       1      34        1        1      active sync   /dev/hdg1
>       2      56        1        2      active sync   /dev/hdi1
>       3       0        0        3      removed
>       4       0        0        4      removed
>
>       5      57        1        -      spare   /dev/hdk1
>       6       8        1        -      faulty spare   /dev/sda1
>
> a quick summary: hdk1 was the one failing in the first place, witch
> was about to get rebuild, sda1 was the device i tried to mount while
> reconstruction the data. As you can see, hdk1 is now marked as spare
> and sda1 a faulty spare.  I have not touch the array after this. Can
> anyone help me out? How can i force mdadm to set the sda1 to active
> sync so i can mount the array and start a backup? Can i use dd_rescue
> to file of all of the disks and play with it?
>
> All help is appreciated!
>
> /Roy
>
--
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html

[Index of Archives]     [Linux RAID Wiki]     [ATA RAID]     [Linux SCSI Target Infrastructure]     [Linux Block]     [Linux IDE]     [Linux SCSI]     [Linux Hams]     [Device Mapper]     [Device Mapper Cryptographics]     [Kernel]     [Linux Admin]     [Linux Net]     [GFS]     [RPM]     [git]     [Yosemite Forum]


  Powered by Linux