Re: Unable to assemble RAID6 after Ubuntu>Arch switch

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi,

I can suppose that 2 superblocks exist on disk /dev/sda (since you
used disks at start and switch to partitions later). One is recorded
for raid built with /dev/sdX devices and 2nd - for raid with /dev/sdX1
partitions. The only idea so far is to assemble raid on Ubuntu, remove
/dev/sda1 from raid (leave it in degraded state), then zero superblock
both for /dev/sda and /dev/sda1, and, finally, add /dev/sda1 back as
'new' drive, wait for raid rebuild, perform these steps with other
drives. Not sure if it's optimal, though.

Regards,
Alexander

On Thu, Sep 24, 2015 at 9:17 PM, Mathias Burén <mathias.buren@xxxxxxxxx> wrote:
> Yeah it's odd. I believe that ages ago I had an array on the full
> disks (not partitions). I've since switched to RAID based on
> partitions (this array that works in Ubuntu but not Arch). Maybe that
> is where the confusion comes from.
>
> I just realized I provided the examine data for the old unused array
> within Arch, and you're right, sda1 and sdb1 don't exist under Arch:
>
> $ sudo partprobe --summary
> /dev/sda: msdos partitions 1
> /dev/sdb: msdos partitions 1
> /dev/sdc: msdos partitions 1 2
> /dev/sdd: msdos partitions 1 2
> /dev/sde: msdos partitions 1 2
> /dev/sdf: msdos partitions 1 2 3
>
> $ sudo fdisk -l /dev/sda
> Disk /dev/sda: 1.8 TiB, 1998998994944 bytes, 3904294912 sectors
> Units: sectors of 1 * 512 = 512 bytes
> Sector size (logical/physical): 512 bytes / 512 bytes
> I/O size (minimum/optimal): 512 bytes / 512 bytes
> Disklabel type: dos
> Disk identifier: 0x056cc1c0
>
> Device     Boot Start        End    Sectors  Size Id Type
> /dev/sda1        2048 3904294911 3904292864  1.8T fd Linux raid autodetect
>
> $ sudo fdisk -l /dev/sdb
> Disk /dev/sdb: 1.8 TiB, 1998998994944 bytes, 3904294912 sectors
> Units: sectors of 1 * 512 = 512 bytes
> Sector size (logical/physical): 512 bytes / 512 bytes
> I/O size (minimum/optimal): 512 bytes / 512 bytes
> Disklabel type: dos
> Disk identifier: 0x9fb14a5c
>
> Device     Boot Start        End    Sectors  Size Id Type
> /dev/sdb1        2048 3904294911 3904292864  1.8T fd Linux raid autodetect
>
> $ ls -la /dev/sda*
> brw-rw---- 1 root disk 8, 0 Sep 24 20:14 /dev/sda
> $ ls -la /dev/sdb*
> brw-rw---- 1 root disk 8, 16 Sep 24 20:14 /dev/sdb
>
>
> So for some reason the kernel does not see the partitions even though
> partprobe and fdisk show them.
>
> Regards,
> Mathias
>
> On 24 September 2015 at 10:17, Alexander Afonyashin
> <a.afonyashin@xxxxxxxxxxxxxx> wrote:
>> Hi Matias,
>>
>> I wonder why Ubuntu detects raid6 parts as /dev/sdX1 (and counts only
>> 5 disks) while Arch detects /dev/sdX (and counts 6 disks)? Try to
>> assemble raid6 on Arch manually again by issuing:
>>
>> mdadm -S /dev/md0
>> mdadm -A /dev/md0 /dev/sdb /dev/sda /dev/sde /dev/sdd /dev/sdc -
>> exactly as shown in mdadm's output on Ubuntu
>>
>> P.S. Still confused why Arch doesn't recognize /dev/sda1 etc.
>>
>> Regards,
>> Alexander
>>
>> On Wed, Sep 23, 2015 at 9:35 PM, Mathias Burén <mathias.buren@xxxxxxxxx> wrote:
>>> Hi Alexander,
>>>
>>> I didn't try to grow the array or anything, just swapped out the OS.
>>> When I boot back into Ubuntu it assembles fine.
>>>
>>> This is from within Ubuntu (I did a mdadm --assemble --scan using
>>> mdadm - v3.3.2-7-g21dc471 - 03th November 2014)
>>>
>>> mdadm -D on RAID6
>>> /dev/md0:
>>>         Version : 1.2
>>>   Creation Time : Thu Nov 20 23:52:58 2014
>>>      Raid Level : raid6
>>>      Array Size : 5856046080 (5584.76 GiB 5996.59 GB)
>>>   Used Dev Size : 1952015360 (1861.59 GiB 1998.86 GB)
>>>    Raid Devices : 5
>>>   Total Devices : 5
>>>     Persistence : Superblock is persistent
>>>
>>>   Intent Bitmap : Internal
>>>
>>>     Update Time : Sun Sep 20 19:47:50 2015
>>>           State : clean
>>>  Active Devices : 5
>>> Working Devices : 5
>>>  Failed Devices : 0
>>>   Spare Devices : 0
>>>
>>>          Layout : left-symmetric
>>>      Chunk Size : 512K
>>>
>>>            Name : ion:0  (local to host ion)
>>>            UUID : 4cae433f:a40afcf5:f9aba91d:d8217b69
>>>          Events : 59736
>>>
>>>     Number   Major   Minor   RaidDevice State
>>>        0       8       17        0      active sync   /dev/sdb1
>>>        1       8       33        1      active sync   /dev/sdc1
>>>        2       8       49        2      active sync   /dev/sdd1
>>>        3       8       65        3      active sync   /dev/sde1
>>>        4       8        1        4      active sync   /dev/sda1
>>>
>>>
>>> mdadm -D on RAID0
>>> /dev/md1:
>>>         Version : 1.2
>>>   Creation Time : Thu Nov 20 23:53:48 2014
>>>      Raid Level : raid0
>>>      Array Size : 4098048 (3.91 GiB 4.20 GB)
>>>    Raid Devices : 3
>>>   Total Devices : 3
>>>     Persistence : Superblock is persistent
>>>
>>>     Update Time : Thu Nov 20 23:53:48 2014
>>>           State : clean
>>>  Active Devices : 3
>>> Working Devices : 3
>>>  Failed Devices : 0
>>>   Spare Devices : 0
>>>
>>>      Chunk Size : 512K
>>>
>>>            Name : ion:1  (local to host ion)
>>>            UUID : a20b70d4:7ee17e3f:abab74f8:dadb8cd8
>>>          Events : 0
>>>
>>>     Number   Major   Minor   RaidDevice State
>>>        0       8       34        0      active sync   /dev/sdc2
>>>        1       8       50        1      active sync   /dev/sdd2
>>>        2       8       66        2      active sync   /dev/sde2
>>>
>>>
>>>
>>> lsdrv from within Ubuntu
>>> PCI [megaraid_sas] 02:0e.0 RAID bus controller: LSI Logic / Symbios
>>> Logic MegaRAID SAS 1068
>>> ├scsi 0:2:0:0 LSI      MegaRAID 84016E  {00ede206d4a731011c50ff5e02b00506}
>>> │└sda 1.82t [8:0] MD raid6 (6) inactive 'ion:md0'
>>> {0ad2603e-e432-83ee-0218-077398e716ef}
>>> │ └sda1 1.82t [8:1] MD raid6 (4/5) (w/ sdb1,sdc1,sdd1,sde1) in_sync
>>> 'ion:0' {4cae433f-a40a-fcf5-f9ab-a91dd8217b69}
>>> │  └md0 5.45t [9:0] MD v1.2 raid6 (5) clean, 512k Chunk
>>> {4cae433f:a40afcf5:f9aba91d:d8217b69}
>>> │                   ext4 '6TB_RAID6' {9e3c1fbe-8228-4b38-9047-66a5e2429e5f}
>>> └scsi 0:2:1:0 LSI      MegaRAID 84016E  {0036936cc4270000ff50ff5e02b00506}
>>>  └sdb 1.82t [8:16] MD raid6 (6) inactive 'ion:md0'
>>> {0ad2603e-e432-83ee-0218-077398e716ef}
>>>   └sdb1 1.82t [8:17] MD raid6 (0/5) (w/ sda1,sdc1,sdd1,sde1) in_sync
>>> 'ion:0' {4cae433f-a40a-fcf5-f9ab-a91dd8217b69}
>>>    └md0 5.45t [9:0] MD v1.2 raid6 (5) clean, 512k Chunk
>>> {4cae433f:a40afcf5:f9aba91d:d8217b69}
>>>                     ext4 '6TB_RAID6' {9e3c1fbe-8228-4b38-9047-66a5e2429e5f}
>>> PCI [ahci] 00:1f.2 SATA controller: Intel Corporation 8 Series/C220
>>> Series Chipset Family 6-port SATA Controller 1 [AHCI mode] (rev 05)
>>> ├scsi 1:0:0:0 ATA      Corsair CSSD-F60 {10326505580009990027}
>>> │└sdf 55.90g [8:80] Partitioned (dos)
>>> │ ├sdf1 243.00m [8:81] ext2 {b45c13c8-4246-43ee-ac2c-f9bb5de099f8}
>>> │ │└Mounted as /dev/sdf1 @ /boot
>>> │ ├sdf2 1.00k [8:82] Partitioned (dos)
>>> │ └sdf5 55.66g [8:85] PV LVM2_member 55.66g used, 0 free
>>> {zZ5Dgy-E7v8-Y1Mi-EuqG-uXUb-DseW-s7goxG}
>>> │  └VG ion 55.66g 0 free {XSaCyP-phLN-m4IH-2cJ7-1XXw-b3uh-qWBl8N}
>>> │   ├dm-0 52.41g [252:0] LV root ext4 {63342e0d-1b4f-4475-9835-d3a2a6610e8f}
>>> │   │└Mounted as /dev/mapper/ion-root @ /
>>> │   └dm-1 3.25g [252:1] LV swap_1 Empty/Unknown
>>> │    └dm-2 3.25g [252:2] swap {0c81625d-ee9e-4a7a-9613-410c1cf53e72}
>>> ├scsi 2:0:0:0 ATA      SAMSUNG HD204UI  {S2H7JR0B501861}
>>> │└sdc 1.82t [8:32] MD raid6 (6) inactive 'ion:md0'
>>> {0ad2603e-e432-83ee-0218-077398e716ef}
>>> │ ├sdc1 1.82t [8:33] MD raid6 (1/5) (w/ sda1,sdb1,sdd1,sde1) in_sync
>>> 'ion:0' {4cae433f-a40a-fcf5-f9ab-a91dd8217b69}
>>> │ │└md0 5.45t [9:0] MD v1.2 raid6 (5) clean, 512k Chunk
>>> {4cae433f:a40afcf5:f9aba91d:d8217b69}
>>> │ │                 ext4 '6TB_RAID6' {9e3c1fbe-8228-4b38-9047-66a5e2429e5f}
>>> │ └sdc2 1.30g [8:34] MD raid0 (0/3) (w/ sdd2,sde2) in_sync 'ion:1'
>>> {a20b70d4-7ee1-7e3f-abab-74f8dadb8cd8}
>>> │  └md1 3.91g [9:1] MD v1.2 raid0 (3) clean, 512k Chunk, None (None)
>>> None {a20b70d4:7ee17e3f:abab74f8:dadb8cd8}
>>> │                   ext4 '4GB_RAID0' {c8327dd0-9d3f-4457-a73a-c9c7d5a0ee3f}
>>> ├scsi 3:x:x:x [Empty]
>>> ├scsi 4:x:x:x [Empty]
>>> ├scsi 5:0:0:0 ATA      WDC WD20EARS-00J {WD-WCAWZ2036074}
>>> │└sdd 1.82t [8:48] MD raid6 (6) inactive 'ion:md0'
>>> {0ad2603e-e432-83ee-0218-077398e716ef}
>>> │ ├sdd1 1.82t [8:49] MD raid6 (2/5) (w/ sda1,sdb1,sdc1,sde1) in_sync
>>> 'ion:0' {4cae433f-a40a-fcf5-f9ab-a91dd8217b69}
>>> │ │└md0 5.45t [9:0] MD v1.2 raid6 (5) clean, 512k Chunk
>>> {4cae433f:a40afcf5:f9aba91d:d8217b69}
>>> │ │                 ext4 '6TB_RAID6' {9e3c1fbe-8228-4b38-9047-66a5e2429e5f}
>>> │ └sdd2 1.30g [8:50] MD raid0 (1/3) (w/ sdc2,sde2) in_sync 'ion:1'
>>> {a20b70d4-7ee1-7e3f-abab-74f8dadb8cd8}
>>> │  └md1 3.91g [9:1] MD v1.2 raid0 (3) clean, 512k Chunk, None (None)
>>> None {a20b70d4:7ee17e3f:abab74f8:dadb8cd8}
>>> │                   ext4 '4GB_RAID0' {c8327dd0-9d3f-4457-a73a-c9c7d5a0ee3f}
>>> └scsi 6:0:0:0 ATA      ST3000DM001-1CH1 {W1F2PZGH}
>>>  └sde 2.73t [8:64] Partitioned (dos)
>>>   ├sde1 1.82t [8:65] MD raid6 (3/5) (w/ sda1,sdb1,sdc1,sdd1) in_sync
>>> 'ion:0' {4cae433f-a40a-fcf5-f9ab-a91dd8217b69}
>>>   │└md0 5.45t [9:0] MD v1.2 raid6 (5) clean, 512k Chunk
>>> {4cae433f:a40afcf5:f9aba91d:d8217b69}
>>>   │                 ext4 '6TB_RAID6' {9e3c1fbe-8228-4b38-9047-66a5e2429e5f}
>>>   ├sde2 1.30g [8:66] MD raid0 (2/3) (w/ sdc2,sdd2) in_sync 'ion:1'
>>> {a20b70d4-7ee1-7e3f-abab-74f8dadb8cd8}
>>>   │└md1 3.91g [9:1] MD v1.2 raid0 (3) clean, 512k Chunk, None (None)
>>> None {a20b70d4:7ee17e3f:abab74f8:dadb8cd8}
>>>   │                 ext4 '4GB_RAID0' {c8327dd0-9d3f-4457-a73a-c9c7d5a0ee3f}
>>>   └sde3 184.98g [8:67] ext4 '198GB' {e4fc427a-4ec1-48dc-8d9b-bfcffe06d42f}
>>>    └Mounted as /dev/sde3 @ /media/198GB
>>> Other Block Devices
>>> ├loop0 0.00k [7:0] Empty/Unknown
>>> ├loop1 0.00k [7:1] Empty/Unknown
>>> ├loop2 0.00k [7:2] Empty/Unknown
>>> ├loop3 0.00k [7:3] Empty/Unknown
>>> ├loop4 0.00k [7:4] Empty/Unknown
>>> ├loop5 0.00k [7:5] Empty/Unknown
>>> ├loop6 0.00k [7:6] Empty/Unknown
>>> ├loop7 0.00k [7:7] Empty/Unknown
>>> ├md127 0.00k [9:127] MD vnone  () clear, None (None) None {None}
>>> │                    Empty/Unknown
>>> ├ram0 64.00m [1:0] Empty/Unknown
>>> ├ram1 64.00m [1:1] Empty/Unknown
>>> ├ram2 64.00m [1:2] Empty/Unknown
>>> ├ram3 64.00m [1:3] Empty/Unknown
>>> ├ram4 64.00m [1:4] Empty/Unknown
>>> ├ram5 64.00m [1:5] Empty/Unknown
>>> ├ram6 64.00m [1:6] Empty/Unknown
>>> ├ram7 64.00m [1:7] Empty/Unknown
>>> ├ram8 64.00m [1:8] Empty/Unknown
>>> ├ram9 64.00m [1:9] Empty/Unknown
>>> ├ram10 64.00m [1:10] Empty/Unknown
>>> ├ram11 64.00m [1:11] Empty/Unknown
>>> ├ram12 64.00m [1:12] Empty/Unknown
>>> ├ram13 64.00m [1:13] Empty/Unknown
>>> ├ram14 64.00m [1:14] Empty/Unknown
>>> └ram15 64.00m [1:15] Empty/Unknown
>>>
>>>
>>> /proc/mdstat on Ubuntu
>>> Personalities : [raid6] [raid5] [raid4] [raid0]
>>> md0 : active raid6 sdb1[0] sda1[4] sde1[3] sdd1[2] sdc1[1]
>>>       5856046080 blocks super 1.2 level 6, 512k chunk, algorithm 2 [5/5] [UUUUU]
>>>       bitmap: 0/15 pages [0KB], 65536KB chunk
>>>
>>> md1 : active raid0 sdc2[0] sde2[2] sdd2[1]
>>>       4098048 blocks super 1.2 512k chunks
>>>
>>> Regards,
>>> Mathias
>>>
>>> On 23 September 2015 at 07:24, Alexander Afonyashin
>>> <a.afonyashin@xxxxxxxxxxxxxx> wrote:
>>>> Hi,
>>>>
>>>> I wonder why metainfo from /dev/sdf1 thinks that there're only 5
>>>> devices in raid6. Did you try to grow your array recently?
>>>> Show the output of mdadm -D /dev/mdX when array is assembled on Ubuntu.
>>>> Can you assemble the array with 4 disks: /dev/sd[a-d]?
>>>>
>>>> P.S. According to mdadm output metainfo on /dev/sdf1 claims that it's
>>>> from another md-array (check Array UUID field). You may need to zero
>>>> metadata on /dev/sdf1 and re-add in to existing raid6 array.
>>>>
>>>> Regards,
>>>> Alexander
>>>>
>>>> On Wed, Sep 23, 2015 at 12:30 AM, Mathias Burén <mathias.buren@xxxxxxxxx> wrote:
>>>>> Hi (please reply-all)
>>>>>
>>>>> I've a RAID6 array (sda sdb sdd sde sdf1) that I can't assemble under
>>>>> Arch, but it worked fine under Ubuntu. mdadm - v3.3.4 - 3rd August
>>>>> 2015, kernel 4.1.6
>>>>>
>>>>> Here is the mdadm --examine for each drive:
>>>>>
>>>>>
>>>>>
>>>>> [root@ion ~]# mdadm --examine /dev/sda
>>>>> /dev/sda:
>>>>>           Magic : a92b4efc
>>>>>         Version : 1.2
>>>>>     Feature Map : 0x0
>>>>>      Array UUID : 0ad2603e:e43283ee:02180773:98e716ef
>>>>>            Name : ion:md0  (local to host ion)
>>>>>   Creation Time : Tue Feb  5 17:33:27 2013
>>>>>      Raid Level : raid6
>>>>>    Raid Devices : 6
>>>>>
>>>>>  Avail Dev Size : 3906767024 (1862.89 GiB 2000.26 GB)
>>>>>      Array Size : 7813531648 (7451.56 GiB 8001.06 GB)
>>>>>   Used Dev Size : 3906765824 (1862.89 GiB 2000.26 GB)
>>>>>     Data Offset : 262144 sectors
>>>>>    Super Offset : 8 sectors
>>>>>    Unused Space : before=262064 sectors, after=18446744073706818560 sectors
>>>>>           State : clean
>>>>>     Device UUID : a09fc60d:5c4a27a5:4b89bc33:29b01582
>>>>>
>>>>>     Update Time : Tue Nov  4 21:43:49 2014
>>>>>        Checksum : 528563ee - correct
>>>>>          Events : 97557
>>>>>
>>>>>          Layout : left-symmetric
>>>>>      Chunk Size : 512K
>>>>>
>>>>>    Device Role : Active device 3
>>>>>    Array State : AA.AAA ('A' == active, '.' == missing, 'R' == replacing)
>>>>>
>>>>> [root@ion ~]# mdadm --examine /dev/sdb
>>>>> /dev/sdb:
>>>>>           Magic : a92b4efc
>>>>>         Version : 1.2
>>>>>     Feature Map : 0x0
>>>>>      Array UUID : 0ad2603e:e43283ee:02180773:98e716ef
>>>>>            Name : ion:md0  (local to host ion)
>>>>>   Creation Time : Tue Feb  5 17:33:27 2013
>>>>>      Raid Level : raid6
>>>>>    Raid Devices : 6
>>>>>
>>>>>  Avail Dev Size : 3906767024 (1862.89 GiB 2000.26 GB)
>>>>>      Array Size : 7813531648 (7451.56 GiB 8001.06 GB)
>>>>>   Used Dev Size : 3906765824 (1862.89 GiB 2000.26 GB)
>>>>>     Data Offset : 262144 sectors
>>>>>    Super Offset : 8 sectors
>>>>>    Unused Space : before=262064 sectors, after=18446744073706818560 sectors
>>>>>           State : clean
>>>>>     Device UUID : 93568b01:632395bf:7d0082a5:db9b6ff9
>>>>>
>>>>>     Update Time : Tue Nov  4 21:43:49 2014
>>>>>        Checksum : 49d756ca - correct
>>>>>          Events : 97557
>>>>>
>>>>>          Layout : left-symmetric
>>>>>      Chunk Size : 512K
>>>>>
>>>>>    Device Role : Active device 0
>>>>>    Array State : AA.AAA ('A' == active, '.' == missing, 'R' == replacing)
>>>>>
>>>>>
>>>>> [root@ion ~]# mdadm --examine /dev/sdd
>>>>> /dev/sdd:
>>>>>           Magic : a92b4efc
>>>>>         Version : 1.2
>>>>>     Feature Map : 0x0
>>>>>      Array UUID : 0ad2603e:e43283ee:02180773:98e716ef
>>>>>            Name : ion:md0  (local to host ion)
>>>>>   Creation Time : Tue Feb  5 17:33:27 2013
>>>>>      Raid Level : raid6
>>>>>    Raid Devices : 6
>>>>>
>>>>>  Avail Dev Size : 3906767024 (1862.89 GiB 2000.26 GB)
>>>>>      Array Size : 7813531648 (7451.56 GiB 8001.06 GB)
>>>>>   Used Dev Size : 3906765824 (1862.89 GiB 2000.26 GB)
>>>>>     Data Offset : 262144 sectors
>>>>>    Super Offset : 8 sectors
>>>>>    Unused Space : before=262064 sectors, after=1200 sectors
>>>>>           State : clean
>>>>>     Device UUID : 78df2586:cb5649aa:e0b6d211:d92dc224
>>>>>
>>>>>     Update Time : Tue Nov  4 21:43:49 2014
>>>>>        Checksum : 50c95b7c - correct
>>>>>          Events : 97557
>>>>>
>>>>>          Layout : left-symmetric
>>>>>      Chunk Size : 512K
>>>>>
>>>>>    Device Role : Active device 5
>>>>>    Array State : AA.AAA ('A' == active, '.' == missing, 'R' == replacing)
>>>>>
>>>>> [root@ion ~]# mdadm --examine /dev/sde
>>>>> /dev/sde:
>>>>>           Magic : a92b4efc
>>>>>         Version : 1.2
>>>>>     Feature Map : 0x0
>>>>>      Array UUID : 0ad2603e:e43283ee:02180773:98e716ef
>>>>>            Name : ion:md0  (local to host ion)
>>>>>   Creation Time : Tue Feb  5 17:33:27 2013
>>>>>      Raid Level : raid6
>>>>>    Raid Devices : 6
>>>>>
>>>>>  Avail Dev Size : 3906767024 (1862.89 GiB 2000.26 GB)
>>>>>      Array Size : 7813531648 (7451.56 GiB 8001.06 GB)
>>>>>   Used Dev Size : 3906765824 (1862.89 GiB 2000.26 GB)
>>>>>     Data Offset : 262144 sectors
>>>>>    Super Offset : 8 sectors
>>>>>    Unused Space : before=262064 sectors, after=1200 sectors
>>>>>           State : clean
>>>>>     Device UUID : 41712f8c:255b0f3e:0e345f7b:e1504e42
>>>>>
>>>>>     Update Time : Tue Nov  4 21:43:49 2014
>>>>>        Checksum : 71b191d6 - correct
>>>>>          Events : 97557
>>>>>
>>>>>          Layout : left-symmetric
>>>>>      Chunk Size : 512K
>>>>>
>>>>>    Device Role : Active device 1
>>>>>    Array State : AA.AAA ('A' == active, '.' == missing, 'R' == replacing)
>>>>>
>>>>> [root@ion ~]# mdadm --examine /dev/sdf1
>>>>> /dev/sdf1:
>>>>>           Magic : a92b4efc
>>>>>         Version : 1.2
>>>>>     Feature Map : 0x1
>>>>>      Array UUID : 4cae433f:a40afcf5:f9aba91d:d8217b69
>>>>>            Name : ion:0  (local to host ion)
>>>>>   Creation Time : Thu Nov 20 23:52:58 2014
>>>>>      Raid Level : raid6
>>>>>    Raid Devices : 5
>>>>>
>>>>>  Avail Dev Size : 3904030720 (1861.59 GiB 1998.86 GB)
>>>>>      Array Size : 5856046080 (5584.76 GiB 5996.59 GB)
>>>>>     Data Offset : 262144 sectors
>>>>>    Super Offset : 8 sectors
>>>>>    Unused Space : before=262056 sectors, after=0 sectors
>>>>>           State : clean
>>>>>     Device UUID : eaa0dcba:d04a7c16:6c256916:67a491ae
>>>>>
>>>>> Internal Bitmap : 8 sectors from superblock
>>>>>     Update Time : Sun Sep 20 19:47:50 2015
>>>>>   Bad Block Log : 512 entries available at offset 72 sectors
>>>>>        Checksum : b1f6f725 - correct
>>>>>          Events : 59736
>>>>>
>>>>>          Layout : left-symmetric
>>>>>      Chunk Size : 512K
>>>>>
>>>>>    Device Role : Active device 3
>>>>>    Array State : AAAAA ('A' == active, '.' == missing, 'R' == replacing)
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>> Here is ldrv:
>>>>>
>>>>> [root@ion ~]# python2 lsdrv
>>>>> PCI [megaraid_sas] 02:0e.0 RAID bus controller: LSI Logic / Symbios
>>>>> Logic MegaRAID SAS 1068
>>>>> ├scsi 0:2:0:0 LSI      MegaRAID 84016E  {00ede206d4a731011c50ff5e02b00506}
>>>>> │└sda 1.82t [8:0] MD raid6 (6) inactive 'ion:md0'
>>>>> {0ad2603e-e432-83ee-0218-077398e716ef}
>>>>> └scsi 0:2:1:0 LSI      MegaRAID 84016E  {0036936cc4270000ff50ff5e02b00506}
>>>>>  └sdb 1.82t [8:16] MD raid6 (6) inactive 'ion:md0'
>>>>> {0ad2603e-e432-83ee-0218-077398e716ef}
>>>>> PCI [ahci] 00:1f.2 SATA controller: Intel Corporation 8 Series/C220
>>>>> Series Chipset Family 6-port SATA Controller 1 [AHCI mode] (rev 05)
>>>>> ├scsi 1:0:0:0 ATA      INTEL SSDSA2BW16 {BTPR2062006T160DGN}
>>>>> │└sdc 149.05g [8:32] Partitioned (dos)
>>>>> │ ├sdc1 256.00m [8:33] ext4 'boot' {f7727d21-646d-42f9-844b-50591b7f8358}
>>>>> │ │└Mounted as /dev/sdc1 @ /boot
>>>>> │ └sdc2 140.00g [8:34] PV LVM2_member 40.00g used, 100.00g free
>>>>> {KGQ5TJ-mVvL-1Ccd-CKae-1QvS-6AFn-8jFQAw}
>>>>> │  └VG ArchVG 140.00g 100.00g free {F2vaKY-XO4m-UEih-f9Fh-sHgX-cRsN-SLi3Oo}
>>>>> │   ├dm-1 32.00g [254:1] LV root ext4 'root'
>>>>> {82fa2951-25c8-4e36-bf36-9f4f7747ac46}
>>>>> │   │└Mounted as /dev/mapper/ArchVG-root @ /
>>>>> │   └dm-0 8.00g [254:0] LV swap swap {0e3f4596-fa50-4fed-875f-f8427084d9b5}
>>>>> ├scsi 2:0:0:0 ATA      SAMSUNG HD204UI  {S2H7JR0B501861}
>>>>> │└sdd 1.82t [8:48] MD raid6 (6) inactive 'ion:md0'
>>>>> {0ad2603e-e432-83ee-0218-077398e716ef}
>>>>> ├scsi 3:x:x:x [Empty]
>>>>> ├scsi 4:x:x:x [Empty]
>>>>> ├scsi 5:0:0:0 ATA      WDC WD20EARS-00J {WD-WCAWZ2036074}
>>>>> │└sde 1.82t [8:64] MD raid6 (6) inactive 'ion:md0'
>>>>> {0ad2603e-e432-83ee-0218-077398e716ef}
>>>>> └scsi 6:0:0:0 ATA      ST3000DM001-1CH1 {W1F2PZGH}
>>>>>  └sdf 2.73t [8:80] Partitioned (dos)
>>>>>   ├sdf1 1.82t [8:81] MD raid6 (5) inactive 'ion:0'
>>>>> {4cae433f-a40a-fcf5-f9ab-a91dd8217b69}
>>>>>   ├sdf2 1.30g [8:82] MD raid0 (3) inactive 'ion:1'
>>>>> {a20b70d4-7ee1-7e3f-abab-74f8dadb8cd8}
>>>>>   └sdf3 184.98g [8:83] ext4 '198GB' {e4fc427a-4ec1-48dc-8d9b-bfcffe06d42f}
>>>>>
>>>>>
>>>>> If I try:
>>>>>
>>>>> [root@ion ~]# mdadm --assemble --scan
>>>>> mdadm: /dev/md/0 assembled from 1 drive - not enough to start the array.
>>>>> mdadm: /dev/md/1 assembled from 1 drive - not enough to start the array.
>>>>>
>>>>> From dmesg:
>>>>>
>>>>> [ 1227.671344]  sda: sda1
>>>>> [ 1227.680392] md: sda does not have a valid v1.2 superblock, not importing!
>>>>> [ 1227.680414] md: md_import_device returned -22
>>>>> [ 1227.680462] md: md0 stopped.
>>>>> [ 1227.707969]  sdb: sdb1
>>>>> [ 1227.718598] md: sdb does not have a valid v1.2 superblock, not importing!
>>>>> [ 1227.718611] md: md_import_device returned -22
>>>>> [ 1227.718631] md: md0 stopped.
>>>>> [ 1286.334542] md: md0 stopped.
>>>>> [ 1286.338250] md: bind<sdf1>
>>>>> [ 1286.338390] md: md0 stopped.
>>>>> [ 1286.338400] md: unbind<sdf1>
>>>>> [ 1286.348350] md: export_rdev(sdf1)
>>>>> [ 1286.372268] md: bind<sdf1>
>>>>> [ 1286.373390] md: md1 stopped.
>>>>> [ 1286.373936] md: bind<sdf2>
>>>>> [ 1286.373977] md: md1 stopped.
>>>>> [ 1286.373983] md: unbind<sdf2>
>>>>> [ 1286.388061] md: export_rdev(sdf2)
>>>>> [ 1286.405140] md: bind<sdf2>
>>>>>
>>>>> [root@ion ~]# cat /proc/mdstat
>>>>> Personalities :
>>>>> md1 : inactive sdf2[2](S)
>>>>>       1366104 blocks super 1.2
>>>>>
>>>>> md0 : inactive sdf1[3](S)
>>>>>       1952015360 blocks super 1.2
>>>>>
>>>>> unused devices: <none>
>>>>>
>>>>>
>>>>> If I try manually (I'm not sure of the order though):
>>>>>
>>>>> [root@ion ~]# mdadm --assemble --readonly --verbose /dev/md0 /dev/sda
>>>>> /dev/sdb /dev/sdd /dev/sde
>>>>> mdadm: looking for devices for /dev/md0
>>>>> mdadm: /dev/sda is identified as a member of /dev/md0, slot 3.
>>>>> mdadm: /dev/sdb is identified as a member of /dev/md0, slot 0.
>>>>> mdadm: /dev/sdd is identified as a member of /dev/md0, slot 5.
>>>>> mdadm: /dev/sde is identified as a member of /dev/md0, slot 1.
>>>>> mdadm: added /dev/sde to /dev/md0 as 1
>>>>> mdadm: no uptodate device for slot 2 of /dev/md0
>>>>> mdadm: failed to add /dev/sda to /dev/md0: Invalid argument
>>>>> mdadm: no uptodate device for slot 4 of /dev/md0
>>>>> mdadm: added /dev/sdd to /dev/md0 as 5
>>>>> mdadm: failed to add /dev/sdb to /dev/md0: Invalid argument
>>>>> mdadm: /dev/md0 assembled from 2 drives - need 5 to start (use --run to insist).
>>>>>
>>>>>
>>>>> Any idea where I should start?
>>>>>
>>>>> Thanks
>>>>> Mathias
>>>>> --
>>>>> To unsubscribe from this list: send the line "unsubscribe linux-raid" in
>>>>> the body of a message to majordomo@xxxxxxxxxxxxxxx
>>>>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
--
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html



[Index of Archives]     [Linux RAID Wiki]     [ATA RAID]     [Linux SCSI Target Infrastructure]     [Linux Block]     [Linux IDE]     [Linux SCSI]     [Linux Hams]     [Device Mapper]     [Device Mapper Cryptographics]     [Kernel]     [Linux Admin]     [Linux Net]     [GFS]     [RPM]     [git]     [Yosemite Forum]


  Powered by Linux