Re: accidently pulled to many devices, raid6 wont start.

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Thu, 2013-12-05 at 14:36 +0100, Bernd Schubert wrote:
> On 12/05/2013 01:30 AM, Wilson Jonathan wrote:
> > mdadm: /dev/sdf6 is identified as a member of /dev/md5, slot 2.
> > mdadm: /dev/sde6 is identified as a member of /dev/md5, slot 1.
> > mdadm: /dev/sdd6 is identified as a member of /dev/md5, slot 0.
> > mdadm: /dev/sdb6 is identified as a member of /dev/md5, slot 4.
> > mdadm: /dev/sda6 is identified as a member of /dev/md5, slot 3.
> > mdadm: ignoring /dev/sde6 as it reports /dev/sdf6 as failed
> 
> [...]
> 
> > So tried again, with a different (valid) chunk...
> >
> >
> > root@BorgCUBE:/mnt/datastore/wilsonjonathan# mdadm --create
> > --assume-clean --level=6 --raid-devices=6
> > --chunk=64 /dev/md5 /dev/sdd6 /dev/sde6
> > missing /dev/sdf6 /dev/sda6 /dev/sdb6
> 
> Why are you using this order? Missing seems to be at the wrong place?
> 
> 
> Cheers,
> Bernd

Arr yes, I seem to have got the order wrong... before I continue and do
further damage; I put the original pulled drive (sdc) back in and re-ran
examine. Obviously I have corrupted the raid layout data on the other
disks (a,b,d,e,f) but here is the sdc examine output.

/dev/sdc6:
          Magic : a92b4efc
        Version : 1.2
    Feature Map : 0x1
     Array UUID : 9ff95b8b:ba34ad95:aa7cb806:169246f2
           Name : PartedMagic:6
  Creation Time : Fri Nov 25 16:30:19 2011
     Raid Level : raid6
   Raid Devices : 6

 Avail Dev Size : 1833569679 (874.31 GiB 938.79 GB)
     Array Size : 3667138560 (3497.26 GiB 3755.15 GB)
  Used Dev Size : 1833569280 (874.31 GiB 938.79 GB)
    Data Offset : 2048 sectors
   Super Offset : 8 sectors
          State : active
    Device UUID : 4ce632fb:d506da60:120cac8e:4433efc4

Internal Bitmap : 8 sectors from superblock
    Update Time : Wed Dec  4 23:04:40 2013
       Checksum : fe4e7b1f - correct
         Events : 260243

         Layout : left-symmetric
     Chunk Size : 64K

   Device Role : Active device 5
   Array State : AAAAAA ('A' == active, '.' == missing)


> 
> 

I mistook the order (a-b-c-d-) with the position number in the array.

If I'm reading this correctly, Active device 5 means it should be at
position 5, or the sixth disk in the array.

I also note it has a bitmap on it, something the other drives now lack
as it was not stated in the original forced create, so not sure how that
will affect things and the event count is lower, 260243, than the other
devices, 260321, before the array became corrupted (that said I know
exactly which files were being updated on the file system, so if I can
get it back I will just delete them and re-create).

Hopefully I have not totally killed the array by my mistakes.



--
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html




[Index of Archives]     [Linux RAID Wiki]     [ATA RAID]     [Linux SCSI Target Infrastructure]     [Linux Block]     [Linux IDE]     [Linux SCSI]     [Linux Hams]     [Device Mapper]     [Device Mapper Cryptographics]     [Kernel]     [Linux Admin]     [Linux Net]     [GFS]     [RPM]     [git]     [Yosemite Forum]


  Powered by Linux