I have an older system with Intel RAID (striped) I use to test updates to my linux boot disk to ensure the RAID is still working. I just went through a big update of all the utilities and shared libraries along with moving from 5.10 to 5.15 kernel. There was a stage where the kernel was updated to 5.15 but all the utilities and shared libraries were not (and md raid was working). The problem is now there are no md devices created, but there are a bunch of dm devices created that never were before. This group has been very helpful in the past. I hope you can help with what is going on here? Does it have something to do with the updated devmapper ? Should I revert? I put both the old working information and the new information below - separated by the 3 lines of equals. ============================================================ ============================================================ ============================================================ Here's the old boot information where it works (5.10.x kernel): Contents of /sys/block: fd0 loop1 loop3 loop5 loop7 md127 sdb sr0 loop0 loop2 loop4 loop6 md126 sda sdc Contents of /dev: block md126p4 tty19 tty57 bsg md126p5 tty2 tty58 btrfs-control md127 tty20 tty59 bus mem tty21 tty6 cdrom null tty22 tty60 cdrw port tty23 tty61 char psaux tty24 tty62 console ptmx tty25 tty63 core pts tty26 tty7 cpu_dma_latency random tty27 tty8 disk rfkill tty28 tty9 dvd rtc tty29 ttyS0 fb0 rtc0 tty3 ttyS1 fd sda tty30 ttyS2 fd0 sda1 tty31 ttyS3 full sda2 tty32 urandom fuse sda3 tty33 vcs fw0 sda5 tty34 vcs1 hidraw0 sdb tty35 vcs2 hidraw1 sdc tty36 vcs3 hidraw2 sdc1 tty37 vcs4 hidraw3 sg0 tty38 vcs5 hpet sg1 tty39 vcs6 input sg2 tty4 vcsa kmsg sg3 tty40 vcsa1 lightnvm sr0 tty41 vcsa2 log stderr tty42 vcsa3 loop-control stdin tty43 vcsa4 loop0 stdout tty44 vcsa5 loop1 synth tty45 vcsa6 loop2 tty tty46 vcsu loop3 tty0 tty47 vcsu1 loop4 tty1 tty48 vcsu2 loop5 tty10 tty49 vcsu3 loop6 tty11 tty5 vcsu4 loop7 tty12 tty50 vcsu5 mapper tty13 tty51 vcsu6 md tty14 tty52 vga_arbiter md126 tty15 tty53 zero md126p1 tty16 tty54 md126p2 tty17 tty55 md126p3 tty18 tty56 Contents of /proc/partitions: major minor #blocks name 2 0 4 fd0 11 0 1048575 sr0 8 0 244198584 sda 8 1 32684242 sda1 8 2 1 sda2 8 3 514048 sda3 8 5 2088418 sda5 8 16 732574584 sdb 8 32 7897088 sdc 8 33 7895916 sdc1 9 126 488391680 md126 259 0 32684242 md126p1 259 1 1 md126p2 259 2 514048 md126p3 259 3 8032 md126p4 259 4 10040593 md126p5 Contents of /proc/mdstat (Linux software RAID status): ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ Personalities : [linear] [raid0] [raid1] [raid10] [raid6] [raid5] [raid4] [multipath] md126 : active raid0 sda[1] sdb[0] 488391680 blocks super external:/md127/0 128k chunks md127 : inactive sda[1](S) sdb[0](S) 7945 blocks super external:imsm unused devices: <none> Contents of /run/mdadm/map (Linux software RAID arrays): ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ md126 /md127/0 835de710:3d35bfb1:d159af46:6570f120 /dev/md/TestRAID md127 imsm 788c3635:2e37de4b:87d08323:987f57e5 /dev/md/imsm0 Contents of /etc/mdadm/mdadm.conf (Linux software RAID config file): ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ # mdadm.conf # # Please refer to mdadm.conf(5) for information about this file. # # by default (built-in), scan all partitions (/proc/partitions) and all # containers for MD superblocks. alternatively, specify devices to scan, using # wildcards if desired. DEVICE partitions containers # automatically tag new arrays as belonging to the local system HOMEHOST <system> ARRAY metadata=imsm UUID=788c3635:2e37de4b:87d08323:987f57e5 ARRAY /dev/md/TestRAID container=788c3635:2e37de4b:87d08323:987f57e5 member=0 UUID=835de710:3d35bfb1:d159af46:6570f120 Long listing of /dev/md directory and /dev/md* device files (Linux software RAID devices): brw-rw---- 1 root disk 9, 126 Aug 13 04:17 /dev/md126 brw-rw---- 1 root disk 259, 0 Aug 13 04:17 /dev/md126p1 brw-rw---- 1 root disk 259, 1 Aug 13 04:17 /dev/md126p2 brw-rw---- 1 root disk 259, 2 Aug 13 04:17 /dev/md126p3 brw-rw---- 1 root disk 259, 3 Aug 13 04:17 /dev/md126p4 brw-rw---- 1 root disk 259, 4 Aug 13 04:17 /dev/md126p5 brw-rw---- 1 root disk 9, 127 Aug 13 04:17 /dev/md127 /dev/md: lrwxrwxrwx 1 root root 8 Aug 13 04:17 TestRAID -> ../md126 lrwxrwxrwx 1 root root 10 Aug 13 04:17 TestRAID1 -> ../md126p1 lrwxrwxrwx 1 root root 10 Aug 13 04:17 TestRAID2 -> ../md126p2 lrwxrwxrwx 1 root root 10 Aug 13 04:17 TestRAID3 -> ../md126p3 lrwxrwxrwx 1 root root 10 Aug 13 04:17 TestRAID4 -> ../md126p4 lrwxrwxrwx 1 root root 10 Aug 13 04:17 TestRAID5 -> ../md126p5 lrwxrwxrwx 1 root root 8 Aug 13 04:17 imsm0 -> ../md127 Contents of /tbu/utility/mdadm.txt (mdadm troubleshooting data captured when 'start-md' is executed): mdadm - v4.1 - 2018-10-01 Output of 'mdadm --examine --scan' ARRAY metadata=imsm UUID=788c3635:2e37de4b:87d08323:987f57e5 ARRAY /dev/md/TestRAID container=788c3635:2e37de4b:87d08323:987f57e5 member=0 UUID=835de710:3d35bfb1:d159af46:6570f120 Output of 'mdadm --assemble --scan --no-degraded -v' mdadm: looking for devices for further assembly mdadm: no RAID superblock on /dev/sdc1 mdadm: /dev/sdc is not attached to Intel(R) RAID controller. mdadm: No OROM/EFI properties for /dev/sdc mdadm: no RAID superblock on /dev/sdc mdadm: no RAID superblock on /dev/sda5 mdadm: no RAID superblock on /dev/sda3 mdadm: no RAID superblock on /dev/sda2 mdadm: no RAID superblock on /dev/sda1 mdadm: /dev/sdb is identified as a member of /dev/md/imsm0, slot 1. mdadm: /dev/sda is identified as a member of /dev/md/imsm0, slot 0. mdadm: added /dev/sdb to /dev/md/imsm0 as 1 mdadm: added /dev/sda to /dev/md/imsm0 as 0 mdadm: Container /dev/md/imsm0 has been assembled with 2 drives mdadm: looking for devices for /dev/md/TestRAID mdadm: looking in container /dev/md/imsm0 mdadm: found match on member /md127/0 in /dev/md/imsm0 mdadm: Started /dev/md/TestRAID with 2 devices Output of 'dmesg | grep md:' md: md127 stopped. md: md126 stopped. Output of 'mdadm --examine /dev/sda' /dev/sda: Magic : Intel Raid ISM Cfg Sig. Version : 1.0.00 Orig Family : a209f7f5 Family : a209f7f5 Generation : 0000016d Attributes : All supported UUID : 788c3635:2e37de4b:87d08323:987f57e5 Checksum : 7cb36503 correct MPB Sectors : 1 Disks : 2 RAID Devices : 1 Disk00 Serial : WD-WCANK2656465 State : active Id : 00000000 Usable Size : 488391680 (232.88 GiB 250.06 GB) [TestRAID]: UUID : 835de710:3d35bfb1:d159af46:6570f120 RAID Level : 0 Members : 2 Slots : [UU] Failed disk : none This Slot : 0 Sector Size : 512 Array Size : 976783360 (465.77 GiB 500.11 GB) Per Dev Size : 488391944 (232.88 GiB 250.06 GB) Sector Offset : 0 Num Stripes : 1907780 Chunk Size : 128 KiB Reserved : 0 Migrate State : idle Map State : normal Dirty State : clean RWH Policy : off Disk01 Serial : 3QK01SE5 State : active Id : 00010000 Usable Size : 1465143680 (698.63 GiB 750.15 GB) Output of 'mdadm --examine /dev/sdb' /dev/sdb: Magic : Intel Raid ISM Cfg Sig. Version : 1.0.00 Orig Family : a209f7f5 Family : a209f7f5 Generation : 0000016d Attributes : All supported UUID : 788c3635:2e37de4b:87d08323:987f57e5 Checksum : 7cb36503 correct MPB Sectors : 1 Disks : 2 RAID Devices : 1 Disk01 Serial : 3QK01SE5 State : active Id : 00010000 Usable Size : 1465138766 (698.63 GiB 750.15 GB) [TestRAID]: UUID : 835de710:3d35bfb1:d159af46:6570f120 RAID Level : 0 Members : 2 Slots : [UU] Failed disk : none This Slot : 1 Sector Size : 512 Array Size : 976783360 (465.77 GiB 500.11 GB) Per Dev Size : 488391944 (232.88 GiB 250.06 GB) Sector Offset : 0 Num Stripes : 1907780 Chunk Size : 128 KiB Reserved : 0 Migrate State : idle Map State : normal Dirty State : clean RWH Policy : off Disk00 Serial : WD-WCANK2656465 State : active Id : 00000000 Usable Size : 488386766 (232.88 GiB 250.05 GB) Output of 'mdadm --examine /dev/sdc' /dev/sdc: MBR Magic : aa55 Partition[0] : 15791832 sectors at 63 (type 0b) Output of 'mdadm --detail /dev/md*', if any: mdadm: /dev/md does not appear to be an md device /dev/md126: Container : /dev/md/imsm0, member 0 Raid Level : raid0 Array Size : 488391680 (465.77 GiB 500.11 GB) Raid Devices : 2 Total Devices : 2 State : clean Active Devices : 2 Working Devices : 2 Failed Devices : 0 Spare Devices : 0 Chunk Size : 128K Consistency Policy : none UUID : 835de710:3d35bfb1:d159af46:6570f120 Number Major Minor RaidDevice State 1 8 0 0 active sync /dev/sda 0 8 16 1 active sync /dev/sdb /dev/md126p1: Container : /dev/md/imsm0, member 0 Raid Level : raid0 Array Size : 32684242 (31.17 GiB 33.47 GB) Raid Devices : 2 Total Devices : 2 State : clean Active Devices : 2 Working Devices : 2 Failed Devices : 0 Spare Devices : 0 Chunk Size : 128K Consistency Policy : none UUID : 835de710:3d35bfb1:d159af46:6570f120 Number Major Minor RaidDevice State 1 8 0 0 active sync /dev/sda 0 8 16 1 active sync /dev/sdb /dev/md126p2: Container : /dev/md/imsm0, member 0 Raid Level : raid0 Array Size : 1 Raid Devices : 2 Total Devices : 2 State : clean Active Devices : 2 Working Devices : 2 Failed Devices : 0 Spare Devices : 0 Chunk Size : 128K Consistency Policy : none UUID : 835de710:3d35bfb1:d159af46:6570f120 Number Major Minor RaidDevice State 1 8 0 0 active sync /dev/sda 0 8 16 1 active sync /dev/sdb /dev/md126p3: Container : /dev/md/imsm0, member 0 Raid Level : raid0 Array Size : 514048 (502.00 MiB 526.39 MB) Raid Devices : 2 Total Devices : 2 State : clean Active Devices : 2 Working Devices : 2 Failed Devices : 0 Spare Devices : 0 Chunk Size : 128K Consistency Policy : none UUID : 835de710:3d35bfb1:d159af46:6570f120 Number Major Minor RaidDevice State 1 8 0 0 active sync /dev/sda 0 8 16 1 active sync /dev/sdb /dev/md126p4: Container : /dev/md/imsm0, member 0 Raid Level : raid0 Array Size : 8032 (7.84 MiB 8.23 MB) Raid Devices : 2 Total Devices : 2 State : clean Active Devices : 2 Working Devices : 2 Failed Devices : 0 Spare Devices : 0 Chunk Size : 128K Consistency Policy : none UUID : 835de710:3d35bfb1:d159af46:6570f120 Number Major Minor RaidDevice State 1 8 0 0 active sync /dev/sda 0 8 16 1 active sync /dev/sdb /dev/md126p5: Container : /dev/md/imsm0, member 0 Raid Level : raid0 Array Size : 10040593 (9.58 GiB 10.28 GB) Raid Devices : 2 Total Devices : 2 State : clean Active Devices : 2 Working Devices : 2 Failed Devices : 0 Spare Devices : 0 Chunk Size : 128K Consistency Policy : none UUID : 835de710:3d35bfb1:d159af46:6570f120 Number Major Minor RaidDevice State 1 8 0 0 active sync /dev/sda 0 8 16 1 active sync /dev/sdb /dev/md127: Version : imsm Raid Level : container Total Devices : 2 Working Devices : 2 UUID : 788c3635:2e37de4b:87d08323:987f57e5 Member Arrays : /dev/md/TestRAID Number Major Minor RaidDevice - 8 0 - /dev/sda - 8 16 - /dev/sdb Contents of /dev/mapper directory: crw------- 1 root root 10, 236 Aug 13 04:17 control ============================================================ ============================================================ ============================================================ ============================================================ Here is the new version that doesn't work (5.15.x kernel): Contents of /sys/block: dm-0 dm-2 dm-5 loop0 loop2 loop4 loop6 sda sdc dm-1 dm-4 fd0 loop1 loop3 loop5 loop7 sdb sr0 Contents of /dev: block mem tty20 tty58 bsg null tty21 tty59 btrfs-control port tty22 tty6 bus psaux tty23 tty60 cdrom ptmx tty24 tty61 cdrw pts tty25 tty62 char random tty26 tty63 console rfkill tty27 tty7 core rtc tty28 tty8 cpu_dma_latency rtc0 tty29 tty9 disk sda tty3 ttyS0 dm-0 sda1 tty30 ttyS1 dm-1 sda2 tty31 ttyS2 dm-2 sda3 tty32 ttyS3 dm-4 sda5 tty33 urandom dm-5 sdb tty34 vcs dvd sdc tty35 vcs1 fb0 sdc1 tty36 vcs2 fd sg0 tty37 vcs3 fd0 sg1 tty38 vcs4 full sg2 tty39 vcs5 fuse sg3 tty4 vcs6 fw0 sr0 tty40 vcsa hidraw0 stderr tty41 vcsa1 hidraw1 stdin tty42 vcsa2 hidraw2 stdout tty43 vcsa3 hidraw3 synth tty44 vcsa4 hpet tty tty45 vcsa5 input tty0 tty46 vcsa6 kmsg tty1 tty47 vcsu log tty10 tty48 vcsu1 loop-control tty11 tty49 vcsu2 loop0 tty12 tty5 vcsu3 loop1 tty13 tty50 vcsu4 loop2 tty14 tty51 vcsu5 loop3 tty15 tty52 vcsu6 loop4 tty16 tty53 vga_arbiter loop5 tty17 tty54 zero loop6 tty18 tty55 loop7 tty19 tty56 mapper tty2 tty57 Contents of /proc/partitions: major minor #blocks name 2 0 4 fd0 11 0 1048575 sr0 8 16 732574584 sdb 8 0 244198584 sda 8 1 32684242 sda1 8 2 1 sda2 8 3 514048 sda3 8 5 2088418 sda5 252 0 488391936 dm-0 252 1 32684242 dm-1 252 2 514048 dm-2 252 4 8032 dm-4 252 5 10040593 dm-5 8 32 15687680 sdc 8 33 15687441 sdc1 Listing of /dev/dm-x devices (LVM, dmraid, etc.): ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ brw-rw---- 1 root disk 252, 0 Aug 13 04:27 /dev/dm-0 brw-rw---- 1 root disk 252, 1 Aug 13 04:27 /dev/dm-1 brw-rw---- 1 root disk 252, 2 Aug 13 04:27 /dev/dm-2 brw-rw---- 1 root disk 252, 4 Aug 13 04:27 /dev/dm-4 brw-rw---- 1 root disk 252, 5 Aug 13 04:27 /dev/dm-5 Listing of /sys/block/dm-x/slaves: ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ /sys/block/dm-0/slaves: lrwxrwxrwx 1 root root 0 Aug 13 11:28 sda -> ../../../../pci0000:00/0000:00:1f.2/ata3/host2/target2:0:0/2:0:0:0/block/sda lrwxrwxrwx 1 root root 0 Aug 13 04:28 sdb -> ../../../../pci0000:00/0000:00:1f.2/ata4/host3/target3:0:0/3:0:0:0/block/sdb /sys/block/dm-1/slaves: lrwxrwxrwx 1 root root 0 Aug 13 04:28 dm-0 -> ../../dm-0 /sys/block/dm-2/slaves: lrwxrwxrwx 1 root root 0 Aug 13 04:28 dm-0 -> ../../dm-0 /sys/block/dm-4/slaves: lrwxrwxrwx 1 root root 0 Aug 13 04:28 dm-0 -> ../../dm-0 /sys/block/dm-5/slaves: lrwxrwxrwx 1 root root 0 Aug 13 04:28 dm-0 -> ../../dm-0 Contents of /proc/mdstat (Linux software RAID status): ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ Personalities : [linear] [raid0] [raid1] [raid10] [raid6] [raid5] [raid4] [multipath] unused devices: <none> Contents of /run/mdadm/map (Linux software RAID arrays): ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ Contents of /etc/mdadm/mdadm.conf (Linux software RAID config file): ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ # mdadm.conf # # Please refer to mdadm.conf(5) for information about this file. # # by default (built-in), scan all partitions (/proc/partitions) and all # containers for MD superblocks. alternatively, specify devices to scan, using # wildcards if desired. DEVICE partitions containers # automatically tag new arrays as belonging to the local system HOMEHOST <system> ARRAY metadata=imsm UUID=788c3635:2e37de4b:87d08323:987f57e5 ARRAY /dev/md/TestRAID container=788c3635:2e37de4b:87d08323:987f57e5 member=0 UUID=835de710:3d35bfb1:d159af46:6570f120 Contents of /tbu/utility/mdadm.txt (mdadm troubleshooting data captured when 'start-md' is executed): mdadm - v4.1 - 2018-10-01 Output of 'mdadm --examine --scan' ARRAY metadata=imsm UUID=788c3635:2e37de4b:87d08323:987f57e5 ARRAY /dev/md/TestRAID container=788c3635:2e37de4b:87d08323:987f57e5 member=0 UUID=835de710:3d35bfb1:d159af46:6570f120 Output of 'mdadm --assemble --scan --no-degraded -v' mdadm: looking for devices for further assembly mdadm: no RAID superblock on /dev/sdc1 mdadm: /dev/sdc is not attached to Intel(R) RAID controller. mdadm: No OROM/EFI properties for /dev/sdc mdadm: no RAID superblock on /dev/sdc mdadm: /dev/dm-5 is not attached to Intel(R) RAID controller. mdadm: No OROM/EFI properties for /dev/dm-5 mdadm: no RAID superblock on /dev/dm-5 mdadm: /dev/dm-4 is not attached to Intel(R) RAID controller. mdadm: No OROM/EFI properties for /dev/dm-4 mdadm: no RAID superblock on /dev/dm-4 mdadm: /dev/dm-2 is not attached to Intel(R) RAID controller. mdadm: No OROM/EFI properties for /dev/dm-2 mdadm: no RAID superblock on /dev/dm-2 mdadm: /dev/dm-1 is not attached to Intel(R) RAID controller. mdadm: No OROM/EFI properties for /dev/dm-1 mdadm: no RAID superblock on /dev/dm-1 mdadm: /dev/dm-0 is not attached to Intel(R) RAID controller. mdadm: No OROM/EFI properties for /dev/dm-0 mdadm: no RAID superblock on /dev/dm-0 mdadm: no RAID superblock on /dev/sda5 mdadm: no RAID superblock on /dev/sda3 mdadm: no RAID superblock on /dev/sda2 mdadm: no RAID superblock on /dev/sda1 mdadm: /dev/sda is busy - skipping mdadm: /dev/sdb is busy - skipping mdadm: looking for devices for /dev/md/TestRAID mdadm: Cannot assemble mbr metadata on /dev/sdc1 mdadm: Cannot assemble mbr metadata on /dev/sdc mdadm: Cannot assemble mbr metadata on /dev/dm-5 mdadm: Cannot assemble mbr metadata on /dev/dm-4 mdadm: Cannot assemble mbr metadata on /dev/dm-2 mdadm: Cannot assemble mbr metadata on /dev/dm-1 mdadm: Cannot assemble mbr metadata on /dev/dm-0 mdadm: Cannot assemble mbr metadata on /dev/sda5 mdadm: Cannot assemble mbr metadata on /dev/sda3 mdadm: Cannot assemble mbr metadata on /dev/sda2 mdadm: no recogniseable superblock on /dev/sda1 mdadm: /dev/sda has wrong uuid. mdadm: /dev/sdb has wrong uuid. Output of 'dmesg | grep md:' Output of 'mdadm --examine /dev/sda' /dev/sda: Magic : Intel Raid ISM Cfg Sig. Version : 1.0.00 Orig Family : a209f7f5 Family : a209f7f5 Generation : 0000016e Attributes : All supported UUID : 788c3635:2e37de4b:87d08323:987f57e5 Checksum : 7cb36504 correct MPB Sectors : 1 Disks : 2 RAID Devices : 1 Disk00 Serial : WD-WCANK2656465 State : active Id : 00000000 Usable Size : 488391680 (232.88 GiB 250.06 GB) [TestRAID]: UUID : 835de710:3d35bfb1:d159af46:6570f120 RAID Level : 0 Members : 2 Slots : [UU] Failed disk : none This Slot : 0 Sector Size : 512 Array Size : 976783360 (465.77 GiB 500.11 GB) Per Dev Size : 488391944 (232.88 GiB 250.06 GB) Sector Offset : 0 Num Stripes : 1907780 Chunk Size : 128 KiB Reserved : 0 Migrate State : idle Map State : normal Dirty State : clean RWH Policy : off Disk01 Serial : 3QK01SE5 State : active Id : 00010000 Usable Size : 1465143680 (698.63 GiB 750.15 GB) Output of 'mdadm --examine /dev/sdb' /dev/sdb: Magic : Intel Raid ISM Cfg Sig. Version : 1.0.00 Orig Family : a209f7f5 Family : a209f7f5 Generation : 0000016e Attributes : All supported UUID : 788c3635:2e37de4b:87d08323:987f57e5 Checksum : 7cb36504 correct MPB Sectors : 1 Disks : 2 RAID Devices : 1 Disk01 Serial : 3QK01SE5 State : active Id : 00010000 Usable Size : 1465138766 (698.63 GiB 750.15 GB) [TestRAID]: UUID : 835de710:3d35bfb1:d159af46:6570f120 RAID Level : 0 Members : 2 Slots : [UU] Failed disk : none This Slot : 1 Sector Size : 512 Array Size : 976783360 (465.77 GiB 500.11 GB) Per Dev Size : 488391944 (232.88 GiB 250.06 GB) Sector Offset : 0 Num Stripes : 1907780 Chunk Size : 128 KiB Reserved : 0 Migrate State : idle Map State : normal Dirty State : clean RWH Policy : off Disk00 Serial : WD-WCANK2656465 State : active Id : 00000000 Usable Size : 488386766 (232.88 GiB 250.05 GB) Output of 'mdadm --examine /dev/sdc' /dev/sdc: MBR Magic : aa55 Partition[0] : 31374882 sectors at 63 (type 0b) Output of 'mdadm --detail /dev/md*', if any: mdadm: cannot open /dev/md*: No such file or directory Contents of /dev/mapper directory: crw------- 1 root root 10, 236 Aug 13 04:27 control brw-rw---- 1 root disk 252, 0 Aug 13 04:27 isw_chbifgccjd_TestRAID brw-rw---- 1 root disk 252, 1 Aug 13 04:27 isw_chbifgccjd_TestRAID1 brw-rw---- 1 root disk 252, 2 Aug 13 04:27 isw_chbifgccjd_TestRAID3 brw-rw---- 1 root disk 252, 4 Aug 13 04:27 isw_chbifgccjd_TestRAID4 brw-rw---- 1 root disk 252, 5 Aug 13 04:27 isw_chbifgccjd_TestRAID5