Re: Self inflicted reshape catastrophe

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



It was this part of the original post

> The raid didn't automatically assemble so I did
> `mdadm --assemble` but really screwed up and put the 5 new disks in a
> different array

Basically, I did an `mdadm --assemble /dev/md1 <new disks for md0>`
instead of `mdadm --assemble /dev/md0 <new disks for md0>`. Further
complicated by the fact the md1 was missing a disk, so I let 1 of the
5 disks become a full member md1 since I didn't catch my error in time
and enough recovery on md1 had occurred to wipe out any data transfer
from the reshape on md0. The other 4 became hot spares. This wiped the
super block on those 5 new disks, the super blocks no longer contain
the correct information showing the original reshape attempt on md0.

I have yet to dive into the code but it seems likely that I can
manually reconstruct the appropriate super blocks for these 4 disks
that still contain valid data as a result of the reshape with a worst
case of ~1/5th data loss.

Here are the `mdadm --examine` for the 4 remaining drives destined for
md0, they believe they belong to md1 (a 6 x 4tb raid 5)

/dev/sdl1:
          Magic : a92b4efc
        Version : 1.2
    Feature Map : 0x1
     Array UUID : a6914f4a:14a64337:c3546c24:42930ff9
           Name : any:1
  Creation Time : Sun Apr  6 05:23:45 2014
     Raid Level : raid5
   Raid Devices : 6

 Avail Dev Size : 19532619776 (9313.88 GiB 10000.70 GB)
     Array Size : 19534425600 (18629.48 GiB 20003.25 GB)
  Used Dev Size : 7813770240 (3725.90 GiB 4000.65 GB)
    Data Offset : 249856 sectors
   Super Offset : 8 sectors
   Unused Space : before=249776 sectors, after=11718849536 sectors
          State : clean
    Device UUID : 2e3e6281:d6735073:c92ffd33:cb6d1413

Internal Bitmap : 8 sectors from superblock
    Update Time : Thu Jan 14 03:18:04 2021
  Bad Block Log : 512 entries available at offset 24 sectors
       Checksum : e87847d5 - correct
         Events : 55628

         Layout : left-symmetric
     Chunk Size : 512K

   Device Role : spare
   Array State : AAAAAA ('A' == active, '.' == missing, 'R' == replacing)

/dev/sdc1:
          Magic : a92b4efc
        Version : 1.2
    Feature Map : 0x1
     Array UUID : a6914f4a:14a64337:c3546c24:42930ff9
           Name : any:1
  Creation Time : Sun Apr  6 05:23:45 2014
     Raid Level : raid5
   Raid Devices : 6

 Avail Dev Size : 19532619776 (9313.88 GiB 10000.70 GB)
     Array Size : 19534425600 (18629.48 GiB 20003.25 GB)
  Used Dev Size : 7813770240 (3725.90 GiB 4000.65 GB)
    Data Offset : 249856 sectors
   Super Offset : 8 sectors
   Unused Space : before=249776 sectors, after=11718849536 sectors
          State : clean
    Device UUID : 268fa8cf:0299a907:5f76dd12:94f78cf2

Internal Bitmap : 8 sectors from superblock
    Update Time : Thu Jan 14 03:18:03 2021
  Bad Block Log : 512 entries available at offset 24 sectors
       Checksum : 89708e55 - correct
         Events : 55627

         Layout : left-symmetric
     Chunk Size : 512K

   Device Role : spare
   Array State : AAAAAA ('A' == active, '.' == missing, 'R' == replacing)

/dev/sdm1:
          Magic : a92b4efc
        Version : 1.2
    Feature Map : 0x1
     Array UUID : a6914f4a:14a64337:c3546c24:42930ff9
           Name : any:1
  Creation Time : Sun Apr  6 05:23:45 2014
     Raid Level : raid5
   Raid Devices : 6

 Avail Dev Size : 19532619776 (9313.88 GiB 10000.70 GB)
     Array Size : 19534425600 (18629.48 GiB 20003.25 GB)
  Used Dev Size : 7813770240 (3725.90 GiB 4000.65 GB)
    Data Offset : 249856 sectors
   Super Offset : 8 sectors
   Unused Space : before=249776 sectors, after=11718849536 sectors
          State : clean
    Device UUID : d8fa02f7:8dd22cea:06d0de1d:32d4039d

Internal Bitmap : 8 sectors from superblock
    Update Time : Thu Jan 14 03:37:53 2021
  Bad Block Log : 512 entries available at offset 24 sectors
       Checksum : 48c66f6b - correct
         Events : 55861

         Layout : left-symmetric
     Chunk Size : 512K

   Device Role : spare
   Array State : AAAAAA ('A' == active, '.' == missing, 'R' == replacing)

/dev/sdn1:
          Magic : a92b4efc
        Version : 1.2
    Feature Map : 0x1
     Array UUID : f464edaa:6338d020:7541f151:678234d2
           Name : big-nas:1
  Creation Time : Sun Apr  6 05:23:45 2014
     Raid Level : raid5
   Raid Devices : 6

 Avail Dev Size : 19532619776 (9313.88 GiB 10000.70 GB)
     Array Size : 19534425600 (18629.48 GiB 20003.25 GB)
  Used Dev Size : 7813770240 (3725.90 GiB 4000.65 GB)
    Data Offset : 249856 sectors
   Super Offset : 8 sectors
   Unused Space : before=249776 sectors, after=11718849536 sectors
          State : clean
    Device UUID : f6c7b906:7aa7bbe7:aec61998:d3183973

Internal Bitmap : 8 sectors from superblock
    Update Time : Thu Jan 14 13:04:45 2021
  Bad Block Log : 512 entries available at offset 24 sectors
       Checksum : 24b2884e - correct
         Events : 62212

         Layout : left-symmetric
     Chunk Size : 512K

   Device Role : spare
   Array State : AAAAAA ('A' == active, '.' == missing, 'R' == replacing)

The 5th and final 10tb disk I allowed to become the 6th member of the
6 x 4tb raid 5 after my screwup

/dev/sdo1:
          Magic : a92b4efc
        Version : 1.2
    Feature Map : 0x9
     Array UUID : f464edaa:6338d020:7541f151:678234d2
           Name : big-nas:1
  Creation Time : Sun Apr  6 05:23:45 2014
     Raid Level : raid5
   Raid Devices : 6

 Avail Dev Size : 19532619776 (9313.88 GiB 10000.70 GB)
     Array Size : 19534425600 (18629.48 GiB 20003.25 GB)
  Used Dev Size : 7813770240 (3725.90 GiB 4000.65 GB)
    Data Offset : 249856 sectors
   Super Offset : 8 sectors
   Unused Space : before=249776 sectors, after=11718849536 sectors
          State : clean
    Device UUID : 527db0c9:1636b7c4:06afabaf:35d6db73

Internal Bitmap : 8 sectors from superblock
    Update Time : Sat Jan 16 06:11:30 2021
  Bad Block Log : 512 entries available at offset 24 sectors - bad
blocks present.
       Checksum : dcdbb6cf - correct
         Events : 62885

         Layout : left-symmetric
     Chunk Size : 512K

   Device Role : Active device 4
   Array State : AAAAAA ('A' == active, '.' == missing, 'R' == replacing)

On Sun, Jan 17, 2021 at 6:49 PM NeilBrown <neil@xxxxxxxxxx> wrote:
>
> On Thu, Jan 14 2021, Nathan Brown wrote:
>
> > Scenario:
> >
> > I had a 4 by 10TB raid 5 array and was adding 5 more disks and
> > reshaping it to a raid6. This was working just fine until I got a
> > little too aggressive with perf tuning and caused `mdadm` to
> > completely hang. I froze the rebuild and rebooted the server to wipe
> > away my tuning mess. The raid didn't automatically assemble so I did
> > `mdadm --assemble` but really screwed up and put the 5 new disks in a
> > different array. Not sure why the superblock on those disks didn't
> > stop `mdadm` from putting them into service but the end result was the
> > superblock on those 5 new drives got wiped. That array was missing a
>
> What does this mean "the superblock on those 5 new drives got wiped"?
> How do you think that happened.
> Can you please report the current super blocks (mdadm --examine) anyway
> please.
>
> NeilBrown



[Index of Archives]     [Linux RAID Wiki]     [ATA RAID]     [Linux SCSI Target Infrastructure]     [Linux Block]     [Linux IDE]     [Linux SCSI]     [Linux Hams]     [Device Mapper]     [Device Mapper Cryptographics]     [Kernel]     [Linux Admin]     [Linux Net]     [GFS]     [RPM]     [git]     [Yosemite Forum]


  Powered by Linux