Re: RAID-6 aborted reshape

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Tue, Jun 11, 2019 at 5:21 AM Xiao Ni <xni@xxxxxxxxxx> wrote:
>
>
>
> On 06/11/2019 10:53 AM, Colt Boyd wrote:
> > Is there anything that can be done?
> >
> > -----Original Message-----
> > From: Colt Boyd <coltboyd@xxxxxxxxx>
> > Sent: Saturday, June 8, 2019 10:48 AM
> > To: linux-raid@xxxxxxxxxxxxxxx
> > Subject: RAID-6 aborted reshape
> >
> > I was resizing a raid6 array with a internal write intent bitmap from
> > 5 3TB drives (in RAID6) to 6 drives. It was aborted very early in reshape via reboot and then reassembled with:
> > 'mdadm -A /dev/md0 --force --verbose --update=revert-reshape --invalid-backup /dev/sdb1 /dev/sdc1 /dev/sdd1 /dev/sde1 /dev/sdf1 /dev/sdg1'
>
> Does this command finish? I tried this myself with this command. The
> filesystem is good after this command.
> I interrupt the reshape by `mdadm -S /dev/md0` before assembling it.

It did finish, but did not leave a mountable XFS file system.

>
> Then I did another test. I interrupted the reshape by `echo b >
> /proc/sysrq-trigger`. Then tried to assemble the raid
> by your command. It gave the error message:
>
> [root@dell-per720-08 home]# mdadm -A /dev/md0 --force --verbose
> --update=revert-reshape --invalid-backup /dev/loop[0-4]
> mdadm: looking for devices for /dev/md0
> mdadm: Reshape position is not suitably aligned.
> mdadm: Try normal assembly and stop again
>
>
> Then I used this command to try to assemble it:
> mdadm -A /dev/md0  --verbose  --invalid-backup /dev/loop[0-4]
>
> The filesystem is good too.
>
> By the way I used the latest upstream version.
I was using:
root@OMV1:/tmp# mdadm --version
mdadm - v3.4 - 28th January 2016
root@OMV1:/tmp# uname -r
4.19.0-0.bpo.2-amd64

I can try the upstream version.

I was not using overlay devices at first and initially tried the above
create on the actual raid devices. I still have the superblock intact
on the 6th device. Is there a way to reconstruct the super blocks on
devices 1-5 (raid devices 0-4) based on the superblock on the 6th
device and/or the backup file?

Thanks,
Colt

>
> Regards
> Xiao
> >
> > When I reassembled it this way I incorrectly thought the backup file was zero bytes. It wasn't. I still have the intact backup file.
> >
> > I’ve also since tried to reassemble it with the following create but the XFS fs is not accessible:
> > 'mdadm --create /dev/md0 --data-offset=1024 --level=6 --raid-devices=5 --chunk=1024K --name=OMV:0 /dev/sdb1 /dev/sde1 /dev/sdc1 /dev/sdd1
> > /dev/sdf1 --assume-clean --readonly'
> >
> > I can see the XFS FS on the drives, example:
> > root@OMV1:~# dd if=/dev/sde1 bs=512k count=5 | hexdump -C <snip>
> > 00200000  58 46 53 42 00 00 10 00  00 00 00 00 82 f2 c3 00  |XFSB............|
> > 00200010  00 00 00 00 00 00 00 00  00 00 00 00 00 00 00 00  |................|
> > 00200020  4e 2b 04 64 e8 1b 49 d9  a5 20 b5 74 79 94 52 f8  |N+.d..I.. .ty.R.| <snip>
> >
> > This is what it looked like immediately following the aborted reshape and before attempting to recreate it. This is from the drive that was being added at the time.
> >
> > /dev/sdg1:
> >            Magic : a92b4efc
> >          Version : 1.2
> >      Feature Map : 0x1
> >       Array UUID : f8fdf8d4:d173da32:eaa97186:eaf88ded
> >             Name : OMV:0
> >    Creation Time : Mon Feb 24 18:19:36 2014
> >       Raid Level : raid6
> >     Raid Devices : 6
> >
> >   Avail Dev Size : 5858529280 (2793.56 GiB 2999.57 GB)
> >       Array Size : 11717054464 (11174.25 GiB 11998.26 GB)
> >    Used Dev Size : 5858527232 (2793.56 GiB 2999.57 GB)
> >      Data Offset : 2048 sectors
> >     Super Offset : 8 sectors
> >     Unused Space : before=1960 sectors, after=2048 sectors
> >            State : clean
> >      Device UUID : 92e022c9:ee6fbc26:74da4bcc:5d0e0409
> >
> > Internal Bitmap : 8 sectors from superblock
> >      Update Time : Thu Jun  6 10:24:34 2019
> >    Bad Block Log : 512 entries available at offset 72 sectors
> >         Checksum : 8f0d9eb9 - correct
> >           Events : 1010399
> >
> >           Layout : left-symmetric
> >       Chunk Size : 1024K
> >
> >     Device Role : Active device 5
> >     Array State : AAAAAA ('A' == active, '.' == missing, 'R' == replacing)
> >
> > Where can I go from here to get this back?
> >
>




[Index of Archives]     [Linux RAID Wiki]     [ATA RAID]     [Linux SCSI Target Infrastructure]     [Linux Block]     [Linux IDE]     [Linux SCSI]     [Linux Hams]     [Device Mapper]     [Device Mapper Cryptographics]     [Kernel]     [Linux Admin]     [Linux Net]     [GFS]     [RPM]     [git]     [Yosemite Forum]


  Powered by Linux