Hi, I have problem with reshaping a RAID6. I had a drive failure in 8 disk RAID6, and since I don't need that much space anymore I decided to shrink array instead of buying replacement disk. I executed following commands: e2fsck -f /dev/md2 mdadm --grow -n7 /dev/md2 mdadm: this change will reduce the size of the array. use --grow --array-size first to truncate array. e.g. mdadm --grow /dev/md2 --array-size 14650664960 resize2fs /dev/md2 3500000000 mdadm /dev/md2 --grow --array-size=14650664960 e2fsck -f /dev/md2 mdadm --grow -n7 /dev/md2 --backup-file /root/mdadm-md2.backup There was no errors and 'cat /proc/mdstat' reports reshape in progress: Personalities : [raid6] [raid5] [raid4] md2 : active raid6 sde1[1] sdn1[8] sdb1[9] sdr1[11] sdp1[10] sdl1[4] sdi1[3] 14650664960 blocks super 1.2 level 6, 512k chunk, algorithm 2 [7/6] [_UUUUUU] [>....................] reshape = 0.0% (1/2930132992) finish=9641968871.8min speed=0K/sec bitmap: 22/22 pages [88KB], 65536KB chunk unused devices: <none> The problem is that there was no progress for more than an hour, the reshaping has stopped at the fist chunk. Is this a bug or is it not possible to reshape degraded RAID? What shall I do with the array, can I abort the reshape or is it going to reshape eventually? Output of "mdadm --detail /dev/md2": /dev/md2: Version : 1.2 Creation Time : Sun Oct 19 22:10:51 2014 Raid Level : raid6 Array Size : 14650664960 (13971.96 GiB 15002.28 GB) Used Dev Size : 2930132992 (2794.39 GiB 3000.46 GB) Raid Devices : 7 Total Devices : 7 Persistence : Superblock is persistent Intent Bitmap : Internal Update Time : Fri Apr 7 08:38:29 2017 State : clean, degraded, reshaping Active Devices : 7 Working Devices : 7 Failed Devices : 0 Spare Devices : 0 Layout : left-symmetric Chunk Size : 512K Reshape Status : 0% complete Delta Devices : -1, (8->7) Name : borox:2 (local to host borox) UUID : 216515ea:4a08e3b7:022786cd:534b5f0f Events : 148046 Number Major Minor RaidDevice State 0 0 0 0 removed 1 8 65 1 active sync /dev/sde1 3 8 129 2 active sync /dev/sdi1 4 8 177 3 active sync /dev/sdl1 10 8 241 4 active sync /dev/sdp1 11 65 17 5 active sync /dev/sdr1 9 8 17 6 active sync /dev/sdb1 8 8 209 7 active sync /dev/sdn1 Thanks, Victor -- To unsubscribe from this list: send the line "unsubscribe linux-raid" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html