Hi, Would it be possible to somehow manually assemble array under Arch? Since the Arch kernel doesn't see /dev/sda1 and /dev/sdb1 (fdisk sees them but they do not appear in /dev) I can't assemble arrays. partprobe doesn't help. Any ideas why the partitions wouldn't populate in /dev ? Thanks Mathias On 25 September 2015 at 09:47, Alexander Afonyashin <a.afonyashin@xxxxxxxxxxxxxx> wrote: > Hi, > > I can suppose that 2 superblocks exist on disk /dev/sda (since you > used disks at start and switch to partitions later). One is recorded > for raid built with /dev/sdX devices and 2nd - for raid with /dev/sdX1 > partitions. The only idea so far is to assemble raid on Ubuntu, remove > /dev/sda1 from raid (leave it in degraded state), then zero superblock > both for /dev/sda and /dev/sda1, and, finally, add /dev/sda1 back as > 'new' drive, wait for raid rebuild, perform these steps with other > drives. Not sure if it's optimal, though. > > Regards, > Alexander > > On Thu, Sep 24, 2015 at 9:17 PM, Mathias Burén <mathias.buren@xxxxxxxxx> wrote: >> Yeah it's odd. I believe that ages ago I had an array on the full >> disks (not partitions). I've since switched to RAID based on >> partitions (this array that works in Ubuntu but not Arch). Maybe that >> is where the confusion comes from. >> >> I just realized I provided the examine data for the old unused array >> within Arch, and you're right, sda1 and sdb1 don't exist under Arch: >> >> $ sudo partprobe --summary >> /dev/sda: msdos partitions 1 >> /dev/sdb: msdos partitions 1 >> /dev/sdc: msdos partitions 1 2 >> /dev/sdd: msdos partitions 1 2 >> /dev/sde: msdos partitions 1 2 >> /dev/sdf: msdos partitions 1 2 3 >> >> $ sudo fdisk -l /dev/sda >> Disk /dev/sda: 1.8 TiB, 1998998994944 bytes, 3904294912 sectors >> Units: sectors of 1 * 512 = 512 bytes >> Sector size (logical/physical): 512 bytes / 512 bytes >> I/O size (minimum/optimal): 512 bytes / 512 bytes >> Disklabel type: dos >> Disk identifier: 0x056cc1c0 >> >> Device Boot Start End Sectors Size Id Type >> /dev/sda1 2048 3904294911 3904292864 1.8T fd Linux raid autodetect >> >> $ sudo fdisk -l /dev/sdb >> Disk /dev/sdb: 1.8 TiB, 1998998994944 bytes, 3904294912 sectors >> Units: sectors of 1 * 512 = 512 bytes >> Sector size (logical/physical): 512 bytes / 512 bytes >> I/O size (minimum/optimal): 512 bytes / 512 bytes >> Disklabel type: dos >> Disk identifier: 0x9fb14a5c >> >> Device Boot Start End Sectors Size Id Type >> /dev/sdb1 2048 3904294911 3904292864 1.8T fd Linux raid autodetect >> >> $ ls -la /dev/sda* >> brw-rw---- 1 root disk 8, 0 Sep 24 20:14 /dev/sda >> $ ls -la /dev/sdb* >> brw-rw---- 1 root disk 8, 16 Sep 24 20:14 /dev/sdb >> >> >> So for some reason the kernel does not see the partitions even though >> partprobe and fdisk show them. >> >> Regards, >> Mathias >> >> On 24 September 2015 at 10:17, Alexander Afonyashin >> <a.afonyashin@xxxxxxxxxxxxxx> wrote: >>> Hi Matias, >>> >>> I wonder why Ubuntu detects raid6 parts as /dev/sdX1 (and counts only >>> 5 disks) while Arch detects /dev/sdX (and counts 6 disks)? Try to >>> assemble raid6 on Arch manually again by issuing: >>> >>> mdadm -S /dev/md0 >>> mdadm -A /dev/md0 /dev/sdb /dev/sda /dev/sde /dev/sdd /dev/sdc - >>> exactly as shown in mdadm's output on Ubuntu >>> >>> P.S. Still confused why Arch doesn't recognize /dev/sda1 etc. >>> >>> Regards, >>> Alexander >>> >>> On Wed, Sep 23, 2015 at 9:35 PM, Mathias Burén <mathias.buren@xxxxxxxxx> wrote: >>>> Hi Alexander, >>>> >>>> I didn't try to grow the array or anything, just swapped out the OS. >>>> When I boot back into Ubuntu it assembles fine. >>>> >>>> This is from within Ubuntu (I did a mdadm --assemble --scan using >>>> mdadm - v3.3.2-7-g21dc471 - 03th November 2014) >>>> >>>> mdadm -D on RAID6 >>>> /dev/md0: >>>> Version : 1.2 >>>> Creation Time : Thu Nov 20 23:52:58 2014 >>>> Raid Level : raid6 >>>> Array Size : 5856046080 (5584.76 GiB 5996.59 GB) >>>> Used Dev Size : 1952015360 (1861.59 GiB 1998.86 GB) >>>> Raid Devices : 5 >>>> Total Devices : 5 >>>> Persistence : Superblock is persistent >>>> >>>> Intent Bitmap : Internal >>>> >>>> Update Time : Sun Sep 20 19:47:50 2015 >>>> State : clean >>>> Active Devices : 5 >>>> Working Devices : 5 >>>> Failed Devices : 0 >>>> Spare Devices : 0 >>>> >>>> Layout : left-symmetric >>>> Chunk Size : 512K >>>> >>>> Name : ion:0 (local to host ion) >>>> UUID : 4cae433f:a40afcf5:f9aba91d:d8217b69 >>>> Events : 59736 >>>> >>>> Number Major Minor RaidDevice State >>>> 0 8 17 0 active sync /dev/sdb1 >>>> 1 8 33 1 active sync /dev/sdc1 >>>> 2 8 49 2 active sync /dev/sdd1 >>>> 3 8 65 3 active sync /dev/sde1 >>>> 4 8 1 4 active sync /dev/sda1 >>>> >>>> >>>> mdadm -D on RAID0 >>>> /dev/md1: >>>> Version : 1.2 >>>> Creation Time : Thu Nov 20 23:53:48 2014 >>>> Raid Level : raid0 >>>> Array Size : 4098048 (3.91 GiB 4.20 GB) >>>> Raid Devices : 3 >>>> Total Devices : 3 >>>> Persistence : Superblock is persistent >>>> >>>> Update Time : Thu Nov 20 23:53:48 2014 >>>> State : clean >>>> Active Devices : 3 >>>> Working Devices : 3 >>>> Failed Devices : 0 >>>> Spare Devices : 0 >>>> >>>> Chunk Size : 512K >>>> >>>> Name : ion:1 (local to host ion) >>>> UUID : a20b70d4:7ee17e3f:abab74f8:dadb8cd8 >>>> Events : 0 >>>> >>>> Number Major Minor RaidDevice State >>>> 0 8 34 0 active sync /dev/sdc2 >>>> 1 8 50 1 active sync /dev/sdd2 >>>> 2 8 66 2 active sync /dev/sde2 >>>> >>>> >>>> >>>> lsdrv from within Ubuntu >>>> PCI [megaraid_sas] 02:0e.0 RAID bus controller: LSI Logic / Symbios >>>> Logic MegaRAID SAS 1068 >>>> ├scsi 0:2:0:0 LSI MegaRAID 84016E {00ede206d4a731011c50ff5e02b00506} >>>> │└sda 1.82t [8:0] MD raid6 (6) inactive 'ion:md0' >>>> {0ad2603e-e432-83ee-0218-077398e716ef} >>>> │ └sda1 1.82t [8:1] MD raid6 (4/5) (w/ sdb1,sdc1,sdd1,sde1) in_sync >>>> 'ion:0' {4cae433f-a40a-fcf5-f9ab-a91dd8217b69} >>>> │ └md0 5.45t [9:0] MD v1.2 raid6 (5) clean, 512k Chunk >>>> {4cae433f:a40afcf5:f9aba91d:d8217b69} >>>> │ ext4 '6TB_RAID6' {9e3c1fbe-8228-4b38-9047-66a5e2429e5f} >>>> └scsi 0:2:1:0 LSI MegaRAID 84016E {0036936cc4270000ff50ff5e02b00506} >>>> └sdb 1.82t [8:16] MD raid6 (6) inactive 'ion:md0' >>>> {0ad2603e-e432-83ee-0218-077398e716ef} >>>> └sdb1 1.82t [8:17] MD raid6 (0/5) (w/ sda1,sdc1,sdd1,sde1) in_sync >>>> 'ion:0' {4cae433f-a40a-fcf5-f9ab-a91dd8217b69} >>>> └md0 5.45t [9:0] MD v1.2 raid6 (5) clean, 512k Chunk >>>> {4cae433f:a40afcf5:f9aba91d:d8217b69} >>>> ext4 '6TB_RAID6' {9e3c1fbe-8228-4b38-9047-66a5e2429e5f} >>>> PCI [ahci] 00:1f.2 SATA controller: Intel Corporation 8 Series/C220 >>>> Series Chipset Family 6-port SATA Controller 1 [AHCI mode] (rev 05) >>>> ├scsi 1:0:0:0 ATA Corsair CSSD-F60 {10326505580009990027} >>>> │└sdf 55.90g [8:80] Partitioned (dos) >>>> │ ├sdf1 243.00m [8:81] ext2 {b45c13c8-4246-43ee-ac2c-f9bb5de099f8} >>>> │ │└Mounted as /dev/sdf1 @ /boot >>>> │ ├sdf2 1.00k [8:82] Partitioned (dos) >>>> │ └sdf5 55.66g [8:85] PV LVM2_member 55.66g used, 0 free >>>> {zZ5Dgy-E7v8-Y1Mi-EuqG-uXUb-DseW-s7goxG} >>>> │ └VG ion 55.66g 0 free {XSaCyP-phLN-m4IH-2cJ7-1XXw-b3uh-qWBl8N} >>>> │ ├dm-0 52.41g [252:0] LV root ext4 {63342e0d-1b4f-4475-9835-d3a2a6610e8f} >>>> │ │└Mounted as /dev/mapper/ion-root @ / >>>> │ └dm-1 3.25g [252:1] LV swap_1 Empty/Unknown >>>> │ └dm-2 3.25g [252:2] swap {0c81625d-ee9e-4a7a-9613-410c1cf53e72} >>>> ├scsi 2:0:0:0 ATA SAMSUNG HD204UI {S2H7JR0B501861} >>>> │└sdc 1.82t [8:32] MD raid6 (6) inactive 'ion:md0' >>>> {0ad2603e-e432-83ee-0218-077398e716ef} >>>> │ ├sdc1 1.82t [8:33] MD raid6 (1/5) (w/ sda1,sdb1,sdd1,sde1) in_sync >>>> 'ion:0' {4cae433f-a40a-fcf5-f9ab-a91dd8217b69} >>>> │ │└md0 5.45t [9:0] MD v1.2 raid6 (5) clean, 512k Chunk >>>> {4cae433f:a40afcf5:f9aba91d:d8217b69} >>>> │ │ ext4 '6TB_RAID6' {9e3c1fbe-8228-4b38-9047-66a5e2429e5f} >>>> │ └sdc2 1.30g [8:34] MD raid0 (0/3) (w/ sdd2,sde2) in_sync 'ion:1' >>>> {a20b70d4-7ee1-7e3f-abab-74f8dadb8cd8} >>>> │ └md1 3.91g [9:1] MD v1.2 raid0 (3) clean, 512k Chunk, None (None) >>>> None {a20b70d4:7ee17e3f:abab74f8:dadb8cd8} >>>> │ ext4 '4GB_RAID0' {c8327dd0-9d3f-4457-a73a-c9c7d5a0ee3f} >>>> ├scsi 3:x:x:x [Empty] >>>> ├scsi 4:x:x:x [Empty] >>>> ├scsi 5:0:0:0 ATA WDC WD20EARS-00J {WD-WCAWZ2036074} >>>> │└sdd 1.82t [8:48] MD raid6 (6) inactive 'ion:md0' >>>> {0ad2603e-e432-83ee-0218-077398e716ef} >>>> │ ├sdd1 1.82t [8:49] MD raid6 (2/5) (w/ sda1,sdb1,sdc1,sde1) in_sync >>>> 'ion:0' {4cae433f-a40a-fcf5-f9ab-a91dd8217b69} >>>> │ │└md0 5.45t [9:0] MD v1.2 raid6 (5) clean, 512k Chunk >>>> {4cae433f:a40afcf5:f9aba91d:d8217b69} >>>> │ │ ext4 '6TB_RAID6' {9e3c1fbe-8228-4b38-9047-66a5e2429e5f} >>>> │ └sdd2 1.30g [8:50] MD raid0 (1/3) (w/ sdc2,sde2) in_sync 'ion:1' >>>> {a20b70d4-7ee1-7e3f-abab-74f8dadb8cd8} >>>> │ └md1 3.91g [9:1] MD v1.2 raid0 (3) clean, 512k Chunk, None (None) >>>> None {a20b70d4:7ee17e3f:abab74f8:dadb8cd8} >>>> │ ext4 '4GB_RAID0' {c8327dd0-9d3f-4457-a73a-c9c7d5a0ee3f} >>>> └scsi 6:0:0:0 ATA ST3000DM001-1CH1 {W1F2PZGH} >>>> └sde 2.73t [8:64] Partitioned (dos) >>>> ├sde1 1.82t [8:65] MD raid6 (3/5) (w/ sda1,sdb1,sdc1,sdd1) in_sync >>>> 'ion:0' {4cae433f-a40a-fcf5-f9ab-a91dd8217b69} >>>> │└md0 5.45t [9:0] MD v1.2 raid6 (5) clean, 512k Chunk >>>> {4cae433f:a40afcf5:f9aba91d:d8217b69} >>>> │ ext4 '6TB_RAID6' {9e3c1fbe-8228-4b38-9047-66a5e2429e5f} >>>> ├sde2 1.30g [8:66] MD raid0 (2/3) (w/ sdc2,sdd2) in_sync 'ion:1' >>>> {a20b70d4-7ee1-7e3f-abab-74f8dadb8cd8} >>>> │└md1 3.91g [9:1] MD v1.2 raid0 (3) clean, 512k Chunk, None (None) >>>> None {a20b70d4:7ee17e3f:abab74f8:dadb8cd8} >>>> │ ext4 '4GB_RAID0' {c8327dd0-9d3f-4457-a73a-c9c7d5a0ee3f} >>>> └sde3 184.98g [8:67] ext4 '198GB' {e4fc427a-4ec1-48dc-8d9b-bfcffe06d42f} >>>> └Mounted as /dev/sde3 @ /media/198GB >>>> Other Block Devices >>>> ├loop0 0.00k [7:0] Empty/Unknown >>>> ├loop1 0.00k [7:1] Empty/Unknown >>>> ├loop2 0.00k [7:2] Empty/Unknown >>>> ├loop3 0.00k [7:3] Empty/Unknown >>>> ├loop4 0.00k [7:4] Empty/Unknown >>>> ├loop5 0.00k [7:5] Empty/Unknown >>>> ├loop6 0.00k [7:6] Empty/Unknown >>>> ├loop7 0.00k [7:7] Empty/Unknown >>>> ├md127 0.00k [9:127] MD vnone () clear, None (None) None {None} >>>> │ Empty/Unknown >>>> ├ram0 64.00m [1:0] Empty/Unknown >>>> ├ram1 64.00m [1:1] Empty/Unknown >>>> ├ram2 64.00m [1:2] Empty/Unknown >>>> ├ram3 64.00m [1:3] Empty/Unknown >>>> ├ram4 64.00m [1:4] Empty/Unknown >>>> ├ram5 64.00m [1:5] Empty/Unknown >>>> ├ram6 64.00m [1:6] Empty/Unknown >>>> ├ram7 64.00m [1:7] Empty/Unknown >>>> ├ram8 64.00m [1:8] Empty/Unknown >>>> ├ram9 64.00m [1:9] Empty/Unknown >>>> ├ram10 64.00m [1:10] Empty/Unknown >>>> ├ram11 64.00m [1:11] Empty/Unknown >>>> ├ram12 64.00m [1:12] Empty/Unknown >>>> ├ram13 64.00m [1:13] Empty/Unknown >>>> ├ram14 64.00m [1:14] Empty/Unknown >>>> └ram15 64.00m [1:15] Empty/Unknown >>>> >>>> >>>> /proc/mdstat on Ubuntu >>>> Personalities : [raid6] [raid5] [raid4] [raid0] >>>> md0 : active raid6 sdb1[0] sda1[4] sde1[3] sdd1[2] sdc1[1] >>>> 5856046080 blocks super 1.2 level 6, 512k chunk, algorithm 2 [5/5] [UUUUU] >>>> bitmap: 0/15 pages [0KB], 65536KB chunk >>>> >>>> md1 : active raid0 sdc2[0] sde2[2] sdd2[1] >>>> 4098048 blocks super 1.2 512k chunks >>>> >>>> Regards, >>>> Mathias >>>> >>>> On 23 September 2015 at 07:24, Alexander Afonyashin >>>> <a.afonyashin@xxxxxxxxxxxxxx> wrote: >>>>> Hi, >>>>> >>>>> I wonder why metainfo from /dev/sdf1 thinks that there're only 5 >>>>> devices in raid6. Did you try to grow your array recently? >>>>> Show the output of mdadm -D /dev/mdX when array is assembled on Ubuntu. >>>>> Can you assemble the array with 4 disks: /dev/sd[a-d]? >>>>> >>>>> P.S. According to mdadm output metainfo on /dev/sdf1 claims that it's >>>>> from another md-array (check Array UUID field). You may need to zero >>>>> metadata on /dev/sdf1 and re-add in to existing raid6 array. >>>>> >>>>> Regards, >>>>> Alexander >>>>> >>>>> On Wed, Sep 23, 2015 at 12:30 AM, Mathias Burén <mathias.buren@xxxxxxxxx> wrote: >>>>>> Hi (please reply-all) >>>>>> >>>>>> I've a RAID6 array (sda sdb sdd sde sdf1) that I can't assemble under >>>>>> Arch, but it worked fine under Ubuntu. mdadm - v3.3.4 - 3rd August >>>>>> 2015, kernel 4.1.6 >>>>>> >>>>>> Here is the mdadm --examine for each drive: >>>>>> >>>>>> >>>>>> >>>>>> [root@ion ~]# mdadm --examine /dev/sda >>>>>> /dev/sda: >>>>>> Magic : a92b4efc >>>>>> Version : 1.2 >>>>>> Feature Map : 0x0 >>>>>> Array UUID : 0ad2603e:e43283ee:02180773:98e716ef >>>>>> Name : ion:md0 (local to host ion) >>>>>> Creation Time : Tue Feb 5 17:33:27 2013 >>>>>> Raid Level : raid6 >>>>>> Raid Devices : 6 >>>>>> >>>>>> Avail Dev Size : 3906767024 (1862.89 GiB 2000.26 GB) >>>>>> Array Size : 7813531648 (7451.56 GiB 8001.06 GB) >>>>>> Used Dev Size : 3906765824 (1862.89 GiB 2000.26 GB) >>>>>> Data Offset : 262144 sectors >>>>>> Super Offset : 8 sectors >>>>>> Unused Space : before=262064 sectors, after=18446744073706818560 sectors >>>>>> State : clean >>>>>> Device UUID : a09fc60d:5c4a27a5:4b89bc33:29b01582 >>>>>> >>>>>> Update Time : Tue Nov 4 21:43:49 2014 >>>>>> Checksum : 528563ee - correct >>>>>> Events : 97557 >>>>>> >>>>>> Layout : left-symmetric >>>>>> Chunk Size : 512K >>>>>> >>>>>> Device Role : Active device 3 >>>>>> Array State : AA.AAA ('A' == active, '.' == missing, 'R' == replacing) >>>>>> >>>>>> [root@ion ~]# mdadm --examine /dev/sdb >>>>>> /dev/sdb: >>>>>> Magic : a92b4efc >>>>>> Version : 1.2 >>>>>> Feature Map : 0x0 >>>>>> Array UUID : 0ad2603e:e43283ee:02180773:98e716ef >>>>>> Name : ion:md0 (local to host ion) >>>>>> Creation Time : Tue Feb 5 17:33:27 2013 >>>>>> Raid Level : raid6 >>>>>> Raid Devices : 6 >>>>>> >>>>>> Avail Dev Size : 3906767024 (1862.89 GiB 2000.26 GB) >>>>>> Array Size : 7813531648 (7451.56 GiB 8001.06 GB) >>>>>> Used Dev Size : 3906765824 (1862.89 GiB 2000.26 GB) >>>>>> Data Offset : 262144 sectors >>>>>> Super Offset : 8 sectors >>>>>> Unused Space : before=262064 sectors, after=18446744073706818560 sectors >>>>>> State : clean >>>>>> Device UUID : 93568b01:632395bf:7d0082a5:db9b6ff9 >>>>>> >>>>>> Update Time : Tue Nov 4 21:43:49 2014 >>>>>> Checksum : 49d756ca - correct >>>>>> Events : 97557 >>>>>> >>>>>> Layout : left-symmetric >>>>>> Chunk Size : 512K >>>>>> >>>>>> Device Role : Active device 0 >>>>>> Array State : AA.AAA ('A' == active, '.' == missing, 'R' == replacing) >>>>>> >>>>>> >>>>>> [root@ion ~]# mdadm --examine /dev/sdd >>>>>> /dev/sdd: >>>>>> Magic : a92b4efc >>>>>> Version : 1.2 >>>>>> Feature Map : 0x0 >>>>>> Array UUID : 0ad2603e:e43283ee:02180773:98e716ef >>>>>> Name : ion:md0 (local to host ion) >>>>>> Creation Time : Tue Feb 5 17:33:27 2013 >>>>>> Raid Level : raid6 >>>>>> Raid Devices : 6 >>>>>> >>>>>> Avail Dev Size : 3906767024 (1862.89 GiB 2000.26 GB) >>>>>> Array Size : 7813531648 (7451.56 GiB 8001.06 GB) >>>>>> Used Dev Size : 3906765824 (1862.89 GiB 2000.26 GB) >>>>>> Data Offset : 262144 sectors >>>>>> Super Offset : 8 sectors >>>>>> Unused Space : before=262064 sectors, after=1200 sectors >>>>>> State : clean >>>>>> Device UUID : 78df2586:cb5649aa:e0b6d211:d92dc224 >>>>>> >>>>>> Update Time : Tue Nov 4 21:43:49 2014 >>>>>> Checksum : 50c95b7c - correct >>>>>> Events : 97557 >>>>>> >>>>>> Layout : left-symmetric >>>>>> Chunk Size : 512K >>>>>> >>>>>> Device Role : Active device 5 >>>>>> Array State : AA.AAA ('A' == active, '.' == missing, 'R' == replacing) >>>>>> >>>>>> [root@ion ~]# mdadm --examine /dev/sde >>>>>> /dev/sde: >>>>>> Magic : a92b4efc >>>>>> Version : 1.2 >>>>>> Feature Map : 0x0 >>>>>> Array UUID : 0ad2603e:e43283ee:02180773:98e716ef >>>>>> Name : ion:md0 (local to host ion) >>>>>> Creation Time : Tue Feb 5 17:33:27 2013 >>>>>> Raid Level : raid6 >>>>>> Raid Devices : 6 >>>>>> >>>>>> Avail Dev Size : 3906767024 (1862.89 GiB 2000.26 GB) >>>>>> Array Size : 7813531648 (7451.56 GiB 8001.06 GB) >>>>>> Used Dev Size : 3906765824 (1862.89 GiB 2000.26 GB) >>>>>> Data Offset : 262144 sectors >>>>>> Super Offset : 8 sectors >>>>>> Unused Space : before=262064 sectors, after=1200 sectors >>>>>> State : clean >>>>>> Device UUID : 41712f8c:255b0f3e:0e345f7b:e1504e42 >>>>>> >>>>>> Update Time : Tue Nov 4 21:43:49 2014 >>>>>> Checksum : 71b191d6 - correct >>>>>> Events : 97557 >>>>>> >>>>>> Layout : left-symmetric >>>>>> Chunk Size : 512K >>>>>> >>>>>> Device Role : Active device 1 >>>>>> Array State : AA.AAA ('A' == active, '.' == missing, 'R' == replacing) >>>>>> >>>>>> [root@ion ~]# mdadm --examine /dev/sdf1 >>>>>> /dev/sdf1: >>>>>> Magic : a92b4efc >>>>>> Version : 1.2 >>>>>> Feature Map : 0x1 >>>>>> Array UUID : 4cae433f:a40afcf5:f9aba91d:d8217b69 >>>>>> Name : ion:0 (local to host ion) >>>>>> Creation Time : Thu Nov 20 23:52:58 2014 >>>>>> Raid Level : raid6 >>>>>> Raid Devices : 5 >>>>>> >>>>>> Avail Dev Size : 3904030720 (1861.59 GiB 1998.86 GB) >>>>>> Array Size : 5856046080 (5584.76 GiB 5996.59 GB) >>>>>> Data Offset : 262144 sectors >>>>>> Super Offset : 8 sectors >>>>>> Unused Space : before=262056 sectors, after=0 sectors >>>>>> State : clean >>>>>> Device UUID : eaa0dcba:d04a7c16:6c256916:67a491ae >>>>>> >>>>>> Internal Bitmap : 8 sectors from superblock >>>>>> Update Time : Sun Sep 20 19:47:50 2015 >>>>>> Bad Block Log : 512 entries available at offset 72 sectors >>>>>> Checksum : b1f6f725 - correct >>>>>> Events : 59736 >>>>>> >>>>>> Layout : left-symmetric >>>>>> Chunk Size : 512K >>>>>> >>>>>> Device Role : Active device 3 >>>>>> Array State : AAAAA ('A' == active, '.' == missing, 'R' == replacing) >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> Here is ldrv: >>>>>> >>>>>> [root@ion ~]# python2 lsdrv >>>>>> PCI [megaraid_sas] 02:0e.0 RAID bus controller: LSI Logic / Symbios >>>>>> Logic MegaRAID SAS 1068 >>>>>> ├scsi 0:2:0:0 LSI MegaRAID 84016E {00ede206d4a731011c50ff5e02b00506} >>>>>> │└sda 1.82t [8:0] MD raid6 (6) inactive 'ion:md0' >>>>>> {0ad2603e-e432-83ee-0218-077398e716ef} >>>>>> └scsi 0:2:1:0 LSI MegaRAID 84016E {0036936cc4270000ff50ff5e02b00506} >>>>>> └sdb 1.82t [8:16] MD raid6 (6) inactive 'ion:md0' >>>>>> {0ad2603e-e432-83ee-0218-077398e716ef} >>>>>> PCI [ahci] 00:1f.2 SATA controller: Intel Corporation 8 Series/C220 >>>>>> Series Chipset Family 6-port SATA Controller 1 [AHCI mode] (rev 05) >>>>>> ├scsi 1:0:0:0 ATA INTEL SSDSA2BW16 {BTPR2062006T160DGN} >>>>>> │└sdc 149.05g [8:32] Partitioned (dos) >>>>>> │ ├sdc1 256.00m [8:33] ext4 'boot' {f7727d21-646d-42f9-844b-50591b7f8358} >>>>>> │ │└Mounted as /dev/sdc1 @ /boot >>>>>> │ └sdc2 140.00g [8:34] PV LVM2_member 40.00g used, 100.00g free >>>>>> {KGQ5TJ-mVvL-1Ccd-CKae-1QvS-6AFn-8jFQAw} >>>>>> │ └VG ArchVG 140.00g 100.00g free {F2vaKY-XO4m-UEih-f9Fh-sHgX-cRsN-SLi3Oo} >>>>>> │ ├dm-1 32.00g [254:1] LV root ext4 'root' >>>>>> {82fa2951-25c8-4e36-bf36-9f4f7747ac46} >>>>>> │ │└Mounted as /dev/mapper/ArchVG-root @ / >>>>>> │ └dm-0 8.00g [254:0] LV swap swap {0e3f4596-fa50-4fed-875f-f8427084d9b5} >>>>>> ├scsi 2:0:0:0 ATA SAMSUNG HD204UI {S2H7JR0B501861} >>>>>> │└sdd 1.82t [8:48] MD raid6 (6) inactive 'ion:md0' >>>>>> {0ad2603e-e432-83ee-0218-077398e716ef} >>>>>> ├scsi 3:x:x:x [Empty] >>>>>> ├scsi 4:x:x:x [Empty] >>>>>> ├scsi 5:0:0:0 ATA WDC WD20EARS-00J {WD-WCAWZ2036074} >>>>>> │└sde 1.82t [8:64] MD raid6 (6) inactive 'ion:md0' >>>>>> {0ad2603e-e432-83ee-0218-077398e716ef} >>>>>> └scsi 6:0:0:0 ATA ST3000DM001-1CH1 {W1F2PZGH} >>>>>> └sdf 2.73t [8:80] Partitioned (dos) >>>>>> ├sdf1 1.82t [8:81] MD raid6 (5) inactive 'ion:0' >>>>>> {4cae433f-a40a-fcf5-f9ab-a91dd8217b69} >>>>>> ├sdf2 1.30g [8:82] MD raid0 (3) inactive 'ion:1' >>>>>> {a20b70d4-7ee1-7e3f-abab-74f8dadb8cd8} >>>>>> └sdf3 184.98g [8:83] ext4 '198GB' {e4fc427a-4ec1-48dc-8d9b-bfcffe06d42f} >>>>>> >>>>>> >>>>>> If I try: >>>>>> >>>>>> [root@ion ~]# mdadm --assemble --scan >>>>>> mdadm: /dev/md/0 assembled from 1 drive - not enough to start the array. >>>>>> mdadm: /dev/md/1 assembled from 1 drive - not enough to start the array. >>>>>> >>>>>> From dmesg: >>>>>> >>>>>> [ 1227.671344] sda: sda1 >>>>>> [ 1227.680392] md: sda does not have a valid v1.2 superblock, not importing! >>>>>> [ 1227.680414] md: md_import_device returned -22 >>>>>> [ 1227.680462] md: md0 stopped. >>>>>> [ 1227.707969] sdb: sdb1 >>>>>> [ 1227.718598] md: sdb does not have a valid v1.2 superblock, not importing! >>>>>> [ 1227.718611] md: md_import_device returned -22 >>>>>> [ 1227.718631] md: md0 stopped. >>>>>> [ 1286.334542] md: md0 stopped. >>>>>> [ 1286.338250] md: bind<sdf1> >>>>>> [ 1286.338390] md: md0 stopped. >>>>>> [ 1286.338400] md: unbind<sdf1> >>>>>> [ 1286.348350] md: export_rdev(sdf1) >>>>>> [ 1286.372268] md: bind<sdf1> >>>>>> [ 1286.373390] md: md1 stopped. >>>>>> [ 1286.373936] md: bind<sdf2> >>>>>> [ 1286.373977] md: md1 stopped. >>>>>> [ 1286.373983] md: unbind<sdf2> >>>>>> [ 1286.388061] md: export_rdev(sdf2) >>>>>> [ 1286.405140] md: bind<sdf2> >>>>>> >>>>>> [root@ion ~]# cat /proc/mdstat >>>>>> Personalities : >>>>>> md1 : inactive sdf2[2](S) >>>>>> 1366104 blocks super 1.2 >>>>>> >>>>>> md0 : inactive sdf1[3](S) >>>>>> 1952015360 blocks super 1.2 >>>>>> >>>>>> unused devices: <none> >>>>>> >>>>>> >>>>>> If I try manually (I'm not sure of the order though): >>>>>> >>>>>> [root@ion ~]# mdadm --assemble --readonly --verbose /dev/md0 /dev/sda >>>>>> /dev/sdb /dev/sdd /dev/sde >>>>>> mdadm: looking for devices for /dev/md0 >>>>>> mdadm: /dev/sda is identified as a member of /dev/md0, slot 3. >>>>>> mdadm: /dev/sdb is identified as a member of /dev/md0, slot 0. >>>>>> mdadm: /dev/sdd is identified as a member of /dev/md0, slot 5. >>>>>> mdadm: /dev/sde is identified as a member of /dev/md0, slot 1. >>>>>> mdadm: added /dev/sde to /dev/md0 as 1 >>>>>> mdadm: no uptodate device for slot 2 of /dev/md0 >>>>>> mdadm: failed to add /dev/sda to /dev/md0: Invalid argument >>>>>> mdadm: no uptodate device for slot 4 of /dev/md0 >>>>>> mdadm: added /dev/sdd to /dev/md0 as 5 >>>>>> mdadm: failed to add /dev/sdb to /dev/md0: Invalid argument >>>>>> mdadm: /dev/md0 assembled from 2 drives - need 5 to start (use --run to insist). >>>>>> >>>>>> >>>>>> Any idea where I should start? >>>>>> >>>>>> Thanks >>>>>> Mathias >>>>>> -- >>>>>> To unsubscribe from this list: send the line "unsubscribe linux-raid" in >>>>>> the body of a message to majordomo@xxxxxxxxxxxxxxx >>>>>> More majordomo info at http://vger.kernel.org/majordomo-info.html -- To unsubscribe from this list: send the line "unsubscribe linux-raid" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html