Re: [PATCH 0/3] [RESEND]dm-raid1: several fixes about writing on out of sync mirror device

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi Heinz,
The test idea is basically the same as before, except mounting the lv with ext3
and iozone with the following script.
#! /bin/sh
var=0
while [ $var -lt 10 ] 
do
        let "var += 1"
        date
        iozone  -i 0 -i 1 -i 2 -l 3 -u 3 -r 4k -s 100m -F /mnt/lv/f1 /mnt/lv/f2 /mnt/lv/f3
        date
done

My machine is KVM virtual machine with 2G memory and 4 CPUs.
And the test result is the same. After the failed device comes back, dm-raid1 could do
the recovery and get full synced.(`dmsetup status vg-lv`show that all the regions are sync)
And we can see that data reading from the default mirror and writing to the other legs with 
iostat while doing the recovery.

We are still working on the userspace part. Once it's done, we will also send a 
test case to cover this feature.

Regards,
Lidong

>>> On 4/16/2015 at 05:58 PM, in message <552F87AE.2010405@xxxxxxxxxx>, Heinz
Mauelshagen <heinzm@xxxxxxxxxx> wrote: 

> Lidong, 
>  
> tests need to happen under heavy load, i.e. worst 
> case scenario failures. 
>  
> E.g. an fs is mounted and being updated whilst 
> you're tacking offline/bringing back mirror legs 
> to cause them to be get resynchronized. 
>  
> Heinz 
>  
> On 04/16/2015 05:43 AM, Lidong Zhong wrote: 
> > Hi List/Heinz, 
> > 
> > These three patches are done based on last patch series that replied on  
> April 8. 
> > The following is the test I did about this feature. My test environment: 
> > linux-klqg:~ # dmsetup ls --tree 
> > vg-lv (253:4) 
> >   ├─vg-lv_mimage_2 (253:3) 
> >   │  └─ (8:48) 
> >   ├─vg-lv_mimage_1 (253:2) 
> >   │  └─ (8:32) 
> >   ├─vg-lv_mimage_0 (253:1) 
> >   │  └─ (8:16) 
> >   └─vg-lv_mlog (253:0) 
> > 	└─ (8:64) 
> > nux-klqg:~ # dmsetup status 
> > vg-lv_mimage_2: 0 614400 linear 
> > vg-lv: 0 614400 mirror 3 253:1 253:2 253:3 600/600 1 AAA 3 disk 253:0 A 
> > vg-lv_mimage_1: 0 614400 linear 
> > vg-lv_mimage_0: 0 614400 linear 
> > vg-lv_mlog: 0 8192 linear 
> > linux-klqg:~ # dmsetup table 
> > vg-lv_mimage_2: 0 614400 linear 8:48 2048 
> > vg-lv: 0 614400 mirror disk 2 253:0 1024 3 253:1 0 253:2 0 253:3 0 2  
> handle_errors keep_log 
> > vg-lv_mimage_1: 0 614400 linear 8:32 2048 
> > vg-lv_mimage_0: 0 614400 linear 8:16 2048 
> > vg-lv_mlog: 0 8192 linear 8:64 2048 
> > 
> > 
> > 1\, single data device failure 
> > After make one of the data legs failed, writing data to the first three  
> regions. 
> > linux-klqg:~ # echo "a" |dd of=/dev/vg/lv bs=1K count=1 seek=0 
> > 0+1 records in 
> > 0+1 records out 
> > 2 bytes (2 B) copied, 0.0103211 s, 0.2 kB/s 
> > linux-klqg:~ # echo "b" |dd of=/dev/vg/lv bs=1K count=1 seek=512 
> > 0+1 records in 
> > 0+1 records out 
> > 2 bytes (2 B) copied, 0.00428962 s, 0.5 kB/s 
> > linux-klqg:~ # echo "c" |dd of=/dev/vg/lv bs=1K count=1 seek=1024 
> > 0+1 records in 
> > 0+1 records out 
> > 2 bytes (2 B) copied, 0.00282482 s, 0.7 kB/s 
> > linux-klqg:~ # dmsetup status 
> > vg-lv_mimage_2: 0 614400 linear 
> > vg-lv: 0 614400 mirror 3 253:1 253:2 253:3 597/600 1 ADA 3 disk 253:0 A 
> > vg-lv_mimage_1: 0 614400 linear 
> > vg-lv_mimage_0: 0 614400 linear 
> > vg-lv_mlog: 0 8192 linear 
> > 
> > Now the failed device comes back, its major/minor number may changes,  
> replace the table as needed. 
> > (The devices I tested on are iscsi devices and the minor number changed  
> after each attach/detach) 
> > Then start the recovery 
> > linux-klqg:~ # dmsetup suspend vg-lv 
> > linux-klqg:~ # dmsetup resume vg-lv 
> > linux-klqg:~ # dmsetup status 
> > vg-lv_mimage_2: 0 614400 linear 
> > vg-lv: 0 614400 mirror 3 253:1 253:2 253:3 600/600 1 AAA 3 disk 253:0 A 
> > vg-lv_mimage_1: 0 614400 linear 
> > vg-lv_mimage_0: 0 614400 linear 
> > vg-lv_mlog: 0 8192 linear 
> > 
> > We can see that all the regions are in sync now. 
> > 
> > 2\, two or more data device failure 
> > After detaching the first device(mine is /dev/sdb), write data to the first  
> and second region 
> > linux-klqg:~ # echo "1111111" | dd of=/dev/vg/lv bs=1K count=1 seek=0 
> > 0+1 records in 
> > 0+1 records out 
> > 8 bytes (8 B) copied, 0.00209451 s, 3.8 kB/s 
> > linux-klqg:~ # echo "222222" | dd of=/dev/vg/lv bs=1K count=1 seek=512 
> > 0+1 records in 
> > 0+1 records out 
> > 7 bytes (7 B) copied, 0.00259999 s, 2.7 kB/s 
> > linux-klqg:~ # dmsetup status 
> > vg-lv_mimage_2: 0 614400 linear 
> > vg-lv: 0 614400 mirror 3 253:1 253:2 253:3 598/600 1 ADA 3 disk 253:0 A 
> > vg-lv_mimage_1: 0 614400 linear 
> > vg-lv_mimage_0: 0 614400 linear 
> > vg-lv_mlog: 0 8192 linear 
> > 
> > Now the first and second region are marked as no sync. Then detach the  
> second device 
> > (mine is /dev/sdd) and write data to the third and fourth region 
> > linux-klqg:~ # echo "333333" | dd of=/dev/vg/lv bs=1K count=1 seek=1024 
> > 0+1 records in 
> > 0+1 records out 
> > 7 bytes (7 B) copied, 0.00178031 s, 3.9 kB/s 
> > linux-klqg:~ # echo "444444" | dd of=/dev/vg/lv bs=1K count=1 seek=1536 
> > 0+1 records in 
> > 0+1 records out 
> > 7 bytes (7 B) copied, 0.00256491 s, 2.7 kB/s 
> > linux-klqg:~ # dmsetup status 
> > vg-lv_mimage_2: 0 614400 linear 
> > vg-lv: 0 614400 mirror 3 253:1 253:2 253:3 596/600 1 DDA 3 disk 253:0 A 
> > vg-lv_mimage_1: 0 614400 linear 
> > vg-lv_mimage_0: 0 614400 linear 
> > vg-lv_mlog: 0 8192 linear 
> > 
> > Now there are 4 regions are marked as no sync. Then the first failed device  
> comes back, we try to 
> > do the recovery. 
> > linux-klqg:~ # dmsetup suspend vg-lv 
> > linux-klqg:~ # dmsetup resume vg-lv 
> > linux-klqg:~ # 
> > linux-klqg:~ # 
> > linux-klqg:~ # dmsetup status 
> > vg-lv_mimage_2: 0 614400 linear 
> > vg-lv: 0 614400 mirror 3 253:1 253:2 253:3 596/600 1 DDA 3 disk 253:0 A 
> > vg-lv_mimage_1: 0 614400 linear 
> > vg-lv_mimage_0: 0 614400 linear 
> > vg-lv_mlog: 0 8192 linear 
> > 
> > And it shows there are still 4 regions are marked as no resync, because  
> there is still 
> > a missing device. And we keep writing to the fifth region 
> > linux-klqg:~ # echo "5555555" | dd of=/dev/vg/lv bs=1K count=1 seek=2048 
> > 0+1 records in 
> > 0+1 records out 
> > 8 bytes (8 B) copied, 0.00213449 s, 3.7 kB/s 
> > 
> > And now the second missing device comes back. We try to do the recovery 
> > linux-klqg:~ # dmsetup suspend vg-lv 
> > linux-klqg:~ # dmsetup resume  vg-lv 
> > linux-klqg:~ # dmsetup status 
> > vg-lv_mimage_2: 0 614400 linear 
> > vg-lv: 0 614400 mirror 3 253:1 253:2 253:3 600/600 1 AAA 3 disk 253:0 A 
> > vg-lv_mimage_1: 0 614400 linear 
> > vg-lv_mimage_0: 0 614400 linear 
> > vg-lv_mlog: 0 8192 linear 
> > 
> > It shows all the legs are sync now. we read data from each leg and get the 
> > same result. 
> > 3\, log device failure 
> > After make the log device failed, we tried to write on this lv 
> > linux-klqg:~ # echo "test" |dd of=/dev/vg/lv bs=1K count=1 seek=0 
> > 0+1 records in 
> > 0+1 records out 
> > 21 bytes (21 B) copied, 0.00470523 s, 4.5 kB/s 
> > linux-klqg:~ # dmsetup status 
> > vg-lv_mimage_2: 0 614400 linear 
> > vg-lv: 0 614400 mirror 3 253:1 253:2 253:3 600/600 1 AAA 3 disk 253:0 D 
> > vg-lv_mimage_1: 0 614400 linear 
> > vg-lv_mimage_0: 0 614400 linear 
> > vg-lv_mlog: 0 8192 linear 
> > And we can see that the log device is marked as failed. 
> > And the bio is not written to the data legs because we can't read new data  
> our of 
> > the leg 
> > 
> > Is the test enough? or is there corner case that is not covered in the  
> patch? 
> > Any advice is appreciated. 
> > 
> > Regards, 
> > Lidong 
> > 
> > Lidong Zhong (3): 
> >    dm-raid1: fix the parameter passed into the kernel 
> >    dm-raid1: remove the error flags in the mirror set when it's in sync 
> >    dm-raid1: change default mirror when it's not in sync 
> > 
> >   drivers/md/dm-raid1.c | 38 +++++++++++++++++++++++++------------- 
> >   1 file changed, 25 insertions(+), 13 deletions(-) 
>  
>  
>  

--
dm-devel mailing list
dm-devel@xxxxxxxxxx
https://www.redhat.com/mailman/listinfo/dm-devel





[Index of Archives]     [DM Crypt]     [Fedora Desktop]     [ATA RAID]     [Fedora Marketing]     [Fedora Packaging]     [Fedora SELinux]     [Yosemite Discussion]     [KDE Users]     [Fedora Docs]

  Powered by Linux