Re: RAID 6 Not Mounting (Block device is empty)

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On 11/07/2015 12:05 PM, Francisco Parada wrote:
> Hello,
> 
> I’m not sure if this is the right way to go about it, so let me give
> you my story.  I had a 7 x 3TB RAID array (12TB in total), 6 drives
> in the array and 1 spare that was also part of the array but simply a
> spare waiting at the ready in case of a drive failure, is what I had
> running before my array broke.  I added two new arrays to my system
> last night, in order to back up my current RAID 6.  The first array
> was a 3 x 3TB RAID 0, for a total of 9TB.  Then a 2 x 1TB RAID 0
> array, for a total of 2TB.  9+2 = 11 and although I’m 1TB shy, I knew
> I had a bunch of crap and redundancy to get rid of, I just really
> needed a solid backup after I was going to clean up.

Yes, this is the right way to go about it.  Missed just a few items that
would help.  Good report.

> After creating the new arrays, I started transferring from my 12TB
> array, 2TB worth of data to the 2TB RAID 0 array.  At some point
> during the transfer, rsync complained of an I/O error.  It seemed to
> have transferred 500GB worth of data before this mishap.  The
> following morning, I noticed that error, and saw that I couldn’t
> mount my 12TB array anymore.  Mind you, I didn’t touch this original
> array, but I think what happened was that the I/O error blew 2 of my
> drives.
> 
> What I’m thinking of doing is the following, but I’m just looking for
> some advice in case I’m missing anything:
> 
> sudo mdadm create --assume-clean --level=6 --raid-devices=7 --size=2930135040 /dev/md127 /dev/sde /dev/sdf /dev/sdg /dev/sdh missing
missing /dev/sdd

Absolutely not!

Your array went from running to dead in minutes, so the variation of
event counts doesn't matter that much.  You should forcibly re-assemble
with all devices.

However, before you do *anything*, you need to figure out why so many
devices were ejected from your array.  Was it a controller glitch?  A
power supply failure?  Or, most likely, Unrecoverable Read Errors being
exposed by your first-ever backup, combined with timeout mismatch?

Any attempt to reassemble/recreate/recovery will simply blow up again if
the root cause isn't addressed.

In your next reply, please paste:

1) the dmesg from the time around the event, +/- a few minutes.

2) the output of the following drive diagnostics:

for x in /dev/sd[a-z] ; do echo $x ; smartctl -i -A -l scterc $x ; done

Do *not* perform any --create operation on your array.

*Do* read the list archives linked below -- if any part of it is
unclear, please ask in your next reply.

Phil

[1] http://marc.info/?l=linux-raid&m=139050322510249&w=2
[2] http://marc.info/?l=linux-raid&m=135863964624202&w=2
[3] http://marc.info/?l=linux-raid&m=135811522817345&w=1
[4] http://marc.info/?l=linux-raid&m=133761065622164&w=2
[5] http://marc.info/?l=linux-raid&m=132477199207506
[6] http://marc.info/?l=linux-raid&m=133665797115876&w=2
[7] http://marc.info/?l=linux-raid&m=142487508806844&w=3
--
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html



[Index of Archives]     [Linux RAID Wiki]     [ATA RAID]     [Linux SCSI Target Infrastructure]     [Linux Block]     [Linux IDE]     [Linux SCSI]     [Linux Hams]     [Device Mapper]     [Device Mapper Cryptographics]     [Kernel]     [Linux Admin]     [Linux Net]     [GFS]     [RPM]     [git]     [Yosemite Forum]


  Powered by Linux