Over the past weekend, I added 2 more drives to my /dev/md0 array: sudo mdadm --detail /dev/md0 /dev/md0: Version : 1.2 Creation Time : Sat Dec 16 18:32:08 2017 Raid Level : raid6 Array Size : 54697266176 (52163.38 GiB 56010.00 GB) Used Dev Size : 7813895168 (7451.91 GiB 8001.43 GB) Raid Devices : 9 Total Devices : 9 Persistence : Superblock is persistent Intent Bitmap : Internal Update Time : Mon Mar 11 05:13:12 2019 State : clean Active Devices : 9 Working Devices : 9 Failed Devices : 0 Spare Devices : 0 Layout : left-symmetric Chunk Size : 512K Name : powerhouse:0 (local to host powerhouse) UUID : 19b5c7a5:59e4bd00:b4b1c18c:089df9bd Events : 45981 Number Major Minor RaidDevice State 0 8 0 0 active sync /dev/sda 1 8 16 1 active sync /dev/sdb 2 8 32 2 active sync /dev/sdc 3 8 48 3 active sync /dev/sdd 5 8 144 4 active sync /dev/sdj 4 8 128 5 active sync /dev/sdi 6 8 112 6 active sync /dev/sdh 8 8 80 7 active sync /dev/sdf 7 8 64 8 active sync /dev/sde Afterwards I did an fsck: sudo fsck.ext4 -f /dev/md0 e2fsck 1.42.13 (17-May-2015) Pass 1: Checking inodes, blocks, and sizes Pass 2: Checking directory structure Pass 3: Checking directory connectivity Pass 4: Checking reference counts Pass 5: Checking group summary information /dev/md0: 70089/1220923392 files (3.6% non-contiguous), 7726498938/9767368960 blocks Following that, I tried to perform an offline resize: sudo resize2fs /dev/md0 resize2fs 1.42.13 (17-May-2015) Resizing the filesystem on /dev/md0 to 13674316544 (4k) blocks. Should never happen: resize inode corrupt! After having done that, it looks like it should have been an online resize from reading a thread on here from 2015. After trying the resize I tried to do another fsck: sudo fsck.ext4 -f /dev/md0 e2fsck 1.42.13 (17-May-2015) ext2fs_check_desc: Corrupt group descriptor: bad block for inode table fsck.ext4: Group descriptors look bad... trying backup blocks... Superblock has an invalid journal (inode 8). Clear<y>? yes *** ext3 journal has been deleted - filesystem is now ext2 only *** Resize inode not valid. Recreate<y>? yes It's been stuck here for days, with: 14827 root 20 0 141796 121044 2688 R 93.8 0.4 5546:26 fsck.ext4 It's been running at around 100% the whole time. I don't see any disk io happening either. sudo dumpe2fs -h /dev/md0 dumpe2fs 1.42.13 (17-May-2015) Filesystem volume name: <none> Last mounted on: /Media10 Filesystem UUID: d36119d5-e3ec-47f7-b93e-124eb4598367 Filesystem magic number: 0xEF53 Filesystem revision #: 1 (dynamic) Filesystem features: has_journal ext_attr resize_inode dir_index filetype extent 64bit flex_bg sparse_super large_file huge_file uninit_bg dir_nlink extra_isize Filesystem flags: signed_directory_hash Default mount options: user_xattr acl Filesystem state: clean with errors Errors behavior: Continue Filesystem OS type: Linux Inode count: 1709293568 Block count: 13674316544 Reserved block count: 683715825 Free blocks: 5886063280 Free inodes: 1709223479 First block: 0 Block size: 4096 Fragment size: 4096 Group descriptor size: 64 Blocks per group: 32768 Fragments per group: 32768 Inodes per group: 4096 Inode blocks per group: 256 RAID stride: 128 RAID stripe width: 256 Flex block group size: 16 Filesystem created: Sun Dec 17 10:10:08 2017 Last mount time: Sat Mar 9 17:58:06 2019 Last write time: Mon Mar 11 05:48:14 2019 Mount count: 0 Maximum mount count: -1 Last checked: Mon Mar 11 05:16:14 2019 Check interval: 0 (<none>) Lifetime writes: 29 TB Reserved blocks uid: 0 (user root) Reserved blocks gid: 0 (group root) First inode: 11 Inode size: 256 Required extra isize: 28 Desired extra isize: 28 Journal inode: 8 Default directory hash: half_md4 Directory Hash Seed: 23fd4260-aee9-4f36-8406-240f3b7a39d2 Journal backup: inode blocks Journal superblock magic number invalid! Should I let the fsck continue, or is it safe to exit to try something else. I recently did an offline resize a few weeks ago on the same array and it worked out just fine. I'm not sure what happened this time, I followed the same steps. Thanks for any help.