Array size dropped from 40TB to 7TB when upgrading to 5.10

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hello,

After a clean reboot to the new kernel 5.10.0 my 40TB md raid5 array size
droped to 7TB.
The previous kernel was 5.9.5. Rebooting back to the 5.9.5 didn't fix the
issue.

# cat /proc/mdstat 
Personalities : [raid6] [raid5] [raid4] 
md0 : active raid5 sdf[9] sdd[10] sda[7] sdb[6] sdc[11] sde[8]
      6857871360 blocks super 1.2 level 5, 512k chunk, algorithm 2 [6/6]
[UUUUUU]
      
unused devices: <none>


journalctl -oshort-iso --no-hostname -b -6|grep md0
2020-12-04T02:30:47+0100 kernel: md/raid:md0: device sdf operational as raid
disk 0
2020-12-04T02:30:47+0100 kernel: md/raid:md0: device sda operational as raid
disk 5
2020-12-04T02:30:47+0100 kernel: md/raid:md0: device sdd operational as raid
disk 4
2020-12-04T02:30:47+0100 kernel: md/raid:md0: device sde operational as raid
disk 2
2020-12-04T02:30:47+0100 kernel: md/raid:md0: device sdc operational as raid
disk 1
2020-12-04T02:30:47+0100 kernel: md/raid:md0: device sdb operational as raid
disk 3
2020-12-04T02:30:47+0100 kernel: md/raid:md0: raid level 5 active with 6 out of
6 devices, algorithm 2
2020-12-04T02:30:47+0100 kernel: md0: detected capacity change from 0 to
40007809105920
2020-12-04T02:31:47+0100 kernel: EXT4-fs (md0): mounted filesystem with ordered
data mode. Opts: (null)

# journalctl -oshort-iso --no-hostname -b -5|grep md0
2020-12-15T03:53:00+0100 kernel: md/raid:md0: device sdf operational as raid
disk 0
2020-12-15T03:53:00+0100 kernel: md/raid:md0: device sda operational as raid
disk 5
2020-12-15T03:53:00+0100 kernel: md/raid:md0: device sde operational as raid
disk 2
2020-12-15T03:53:00+0100 kernel: md/raid:md0: device sdd operational as raid
disk 4
2020-12-15T03:53:00+0100 kernel: md/raid:md0: device sdc operational as raid
disk 1
2020-12-15T03:53:00+0100 kernel: md/raid:md0: device sdb operational as raid
disk 3
2020-12-15T03:53:00+0100 kernel: md/raid:md0: raid level 5 active with 6 out of
6 devices, algorithm 2
2020-12-15T03:53:00+0100 kernel: md0: detected capacity change from 0 to
7022460272640
2020-12-15T03:54:20+0100 systemd-fsck[1009]: fsck.ext4: Invalid argument while
trying to open /dev/md0

There is no log of hardware errors or unclean unmounting.

# mdadm -D /dev/md0
/dev/md0:
           Version : 1.2
     Creation Time : Mon Jan 24 02:53:21 2011
        Raid Level : raid5
        Array Size : 6857871360 (6540.18 GiB 7022.46 GB)
     Used Dev Size : 1371574272 (1308.04 GiB 1404.49 GB)
      Raid Devices : 6
     Total Devices : 6
       Persistence : Superblock is persistent

       Update Time : Tue Dec 15 17:53:13 2020
             State : clean 
    Active Devices : 6
   Working Devices : 6
    Failed Devices : 0
     Spare Devices : 0

            Layout : left-symmetric
        Chunk Size : 512K

Consistency Policy : resync

              Name : white:0  (local to host white)
              UUID : affd87df:da503e3b:52a8b97f:77b80c0c
            Events : 1791763

    Number   Major   Minor   RaidDevice State
       9       8       80        0      active sync   /dev/sdf
      11       8       32        1      active sync   /dev/sdc
       8       8       64        2      active sync   /dev/sde
       6       8       16        3      active sync   /dev/sdb
      10       8       48        4      active sync   /dev/sdd
       7       8        0        5      active sync   /dev/sda

The mdadm userspace as not been updated.
# mdadm -V
mdadm - v4.1 - 2018-10-01

An `mdadm --action check /dev/md0` was run without errors.

1) What's the best option to restore the size without loosing the data?
2) Is this issue can be related to the kernel upgrade or it's fortuitous?

Regards,

Sébastien "Seblu" Luttringer

Attachment: signature.asc
Description: This is a digitally signed message part


[Index of Archives]     [Linux RAID Wiki]     [ATA RAID]     [Linux SCSI Target Infrastructure]     [Linux Block]     [Linux IDE]     [Linux SCSI]     [Linux Hams]     [Device Mapper]     [Device Mapper Cryptographics]     [Kernel]     [Linux Admin]     [Linux Net]     [GFS]     [RPM]     [git]     [Yosemite Forum]


  Powered by Linux