Re: md/raid5: report one failure scenario about growing array from mirror to level 5

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 




On 03/24/2017 11:51 PM, Shaohua Li wrote:
On Thu, Mar 16, 2017 at 10:14:04PM +0800, Zhilong Liu wrote:
hi, list;

   I just trace the code of the following scenarios, grow array from mirror
   to level 5,  it would be stuck at the beginning of reshape_progress when
   using the loop devices, the same testing steps work well via to hard disk
   devices.  refer to the detail steps as follow.
   Is this issue only happened on my site?

*Steps for loop devices:*
1). create one mirror array via to mdadm source code.
#  cd mdadm/
#  ./test setup
#  ./mdadm -CR /dev/md0 --level 1 -b internal -n2 /dev/loop[0-1]
#  dmesg -c,    clean the dmesg.
2). triggers the reshape_request, then it would be stuck when start
reshape_progress.
... ...
linux-x4lv:~/mdadm-test # ./mdadm --grow /dev/md0 -l5 -n3 -a /dev/loop2
mdadm: level of /dev/md0 changed to raid5
mdadm: added /dev/loop2
mdadm: Need to backup 128K of critical section..

linux-x4lv:~/mdadm-test # cat /proc/mdstat
Personalities : [raid1] [raid6] [raid5] [raid4]
md0 : active raid5 loop2[2] loop1[1] loop0[0]
       19968 blocks super 1.2 level 5, 64k chunk, algorithm 2 [3/3] [UUU]
       [>....................]  reshape =  0.0% (1/19968) finish=41.5min
speed=0K/sec
       bitmap: 0/1 pages [0KB], 65536KB chunk

unused devices: <none>

linux-x4lv:~/mdadm-test # dmesg -c
[111544.283359] md/raid:md0: device loop1 operational as raid disk 1
[111544.283362] md/raid:md0: device loop0 operational as raid disk 0
[111544.296161] md/raid:md0: raid level 5 active with 2 out of 2 devices,
algorithm 2
[111544.296178] md/raid456: discard support disabled due to uncertainty.
[111544.296178] Set raid456.devices_handle_discard_safely=Y to override.
[111544.932238] md: reshape of RAID array md0

*Steps for hard disks:*
# ./mdadm --grow /dev/md0 -l5 -n3  -a /dev/sdd
mdadm: level of /dev/md0 changed to raid5
mdadm: added /dev/sdd
mdadm: Need to backup 128K of critical section..

# cat /proc/mdstat
Personalities : [raid1] [raid0] [raid6] [raid5] [raid4]
md0 : active raid5 sdd[2] sdc[1] sdb[0]
       102272 blocks super 1.2 level 5, 64k chunk, algorithm 2 [3/3] [UUU]
       [====>................]  reshape = 22.5% (23168/102272) finish=0.5min
speed=2316K/sec
       bitmap: 0/1 pages [0KB], 65536KB chunk

unused devices: <none>
can't reproduce locally. does this only exist in specific mdadm/kernel version?

I have noticed that it causes by the failure of "mdadm-grow-continue@%s.service",
I have sent a patch to mailing-list named:
mdadm/grow: reshape would be stuck from raid1 to raid5

I tested it under built the latest linux source code(4.11.0-rc4) and latest mdadm.
I would continue to look at this issue.

Thanks,
-Zhilong
Thanks,
Shaohua
--
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html


--
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html



[Index of Archives]     [Linux RAID Wiki]     [ATA RAID]     [Linux SCSI Target Infrastructure]     [Linux Block]     [Linux IDE]     [Linux SCSI]     [Linux Hams]     [Device Mapper]     [Device Mapper Cryptographics]     [Kernel]     [Linux Admin]     [Linux Net]     [GFS]     [RPM]     [git]     [Yosemite Forum]


  Powered by Linux