Re: Cannot add replacement hard drive to mdadm RAID5 array

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hello Artur,

I followed your recommended steps using the --force flag on run of my /dev/md126 array, in addition to stopping the extra container /dev/md125 and adding the new disk /dev/sdb to the parent container /dev/md127. It seems that it still does not work yet.

Output of mdstat prior to running the commands:

cat /proc/mdstat
Personalities : [raid6] [raid5] [raid4] [linear] [multipath] [raid0] [raid1] [raid10]
md125 : inactive sdb[0](S)
      1105 blocks super external:imsm

md126 : inactive sda[2] sdc[1] sdd[0]
      5567507712 blocks super external:/md127/0

md127 : inactive sdc[2](S) sdd[1](S) sda[0](S)
      9459 blocks super external:imsm

unused devices: <none>

Output of the recommended commands for adding the new disk to the RAID5 array:

sudo mdadm -R --force /dev/md126
mdadm: array /dev/md/Data now has 3 devices (0 new)

sudo mdadm -S /dev/md125
mdadm: stopped /dev/md125

sudo mdadm -a /dev/md127 /dev/sdb
mdadm: added /dev/sdb

Output of mdstat after running the commands. Shows that both md126 and md127 are inactive, and there is no RAID resync happening.

cat /proc/mdstat
Personalities : [raid6] [raid5] [raid4] [linear] [multipath] [raid0] [raid1] [raid10]
md126 : inactive sda[2] sdc[1] sdd[0]
      5567507712 blocks super external:/md127/0

md127 : inactive sdb[3](S) sdc[2](S) sdd[1](S) sda[0](S)
      10564 blocks super external:imsm

unused devices: <none>

Output of examination of the new disk /dev/sdb. Still shows that it does not have a UUID and is a spare disk, even after being added to the md127 parent container.

sudo mdadm -E /dev/sdb
/dev/sdb:
          Magic : Intel Raid ISM Cfg Sig.
        Version : 1.0.00
    Orig Family : 00000000
         Family : e4cd8601
     Generation : 00000001
  Creation Time : Unknown
     Attributes : All supported
           UUID : 00000000:00000000:00000000:00000000
       Checksum : c99b0c02 correct
    MPB Sectors : 1
          Disks : 1
   RAID Devices : 0

  Disk00 Serial : W462MZ0R
          State : spare
             Id : 03000000
    Usable Size : 3907026958 (1863.02 GiB 2000.40 GB)

    Disk Serial : W462MZ0R
          State : spare
             Id : 03000000
    Usable Size : 3907026958 (1863.02 GiB 2000.40 GB)
	
Output of details of parent container md127, which does show that the new disk /dev/sdb is in the container after being added.
	
sudo mdadm -D /dev/md127
/dev/md127:
           Version : imsm
        Raid Level : container
     Total Devices : 4

   Working Devices : 4


              UUID : 72360627:bb745f4c:aedafaab:e25d3123
     Member Arrays : /dev/md126

    Number   Major   Minor   RaidDevice

       -       8        0        -        /dev/sda
       -       8       16        -        /dev/sdb
       -       8       32        -        /dev/sdc
       -       8       48        -        /dev/sdd
	   
	   
Output of details on the real RAID array md126 which does not show the new disk sdb, only containing the other three devices sda, sdc, and sdd. Also the state shows active, FAILED, Not Started. Also, it shows that four devices were removed.
	   
sudo mdadm -D /dev/md126
/dev/md126:
         Container : /dev/md/imsm0, member 0
        Raid Level : raid5
     Used Dev Size : 1855835904 (1769.86 GiB 1900.38 GB)
      Raid Devices : 4
     Total Devices : 3

             State : active, FAILED, Not Started
    Active Devices : 3
   Working Devices : 3
    Failed Devices : 0
     Spare Devices : 0

            Layout : left-asymmetric
        Chunk Size : 64K

Consistency Policy : unknown


              UUID : 764aa814:831953a1:06cf2a07:1ca42b2e
    Number   Major   Minor   RaidDevice State
       -       0        0        0      removed
       -       0        0        1      removed
       -       0        0        2      removed
       -       0        0        3      removed

       -       8        0        0      sync   /dev/sda
       -       8       32        2      sync   /dev/sdc
       -       8       48        3      sync   /dev/sdd
	   
Here is an example of the details of a healthy RAID array on another system. I want to recreate this result if possible.

sudo mdadm -D /dev/md126
/dev/md126:
         Container : /dev/md/imsm0, member 0
        Raid Level : raid5
        Array Size : 5567516672 (5309.60 GiB 5701.14 GB)
     Used Dev Size : 1855838976 (1769.87 GiB 1900.38 GB)
      Raid Devices : 4
     Total Devices : 4

             State : clean
    Active Devices : 4
   Working Devices : 4
    Failed Devices : 0
     Spare Devices : 0

            Layout : left-asymmetric
        Chunk Size : 64K

Consistency Policy : resync
​

              UUID : 4541a34d:4a84d4dc:3d1b0695:96f58987
    Number   Major   Minor   RaidDevice State
       3       8        0        0      active sync   /dev/sda
       2       8       16        1      active sync   /dev/sdb
       1       8       32        2      active sync   /dev/sdc
       0       8       48        3      active sync   /dev/sdd

Please let me know if there is any additional information I can provide to help troubleshooting.

Thanks!

Devon Beets
 
  
 










From: Tkaczyk, Mariusz <mariusz.tkaczyk@xxxxxxxxxxxxxxx>
Sent: Tuesday, March 16, 2021 2:15 AM
To: Artur Paszkiewicz <artur.paszkiewicz@xxxxxxxxx>; Devon Beets <devon@xxxxxxxxxxxxxxxx>; linux-raid@xxxxxxxxxxxxxxx <linux-raid@xxxxxxxxxxxxxxx>
Cc: Glenn Wikle <gwikle@xxxxxxxxxxxxxxxx>
Subject: Re: Cannot add replacement hard drive to mdadm RAID5 array 
 
On 15.03.2021 15:38, Artur Paszkiewicz wrote:
> On 3/12/21 11:55 PM, Devon Beets wrote:
>> Hello,
>>
>> My colleague and I have been trying to replace a failed hard drive in a four-drive RAID5 array (/dev/sda to /dev/sdd). The failed drive is sdb. We have physically removed the hard drive and replaced it with a new drive that has identical specifications. We did not first use mdadm to set the failed hard drive with --fail.
>>
>> Upon booting the system with the new /dev/sdb drive installed, we see that instead of the usual two md entries (/dev/md127 which is an IMSM container and /dev/md126 which is the actual array) there are now three entries: md125 to md127. md127 is the IMSM container for sda, sdc, and sdd. md125 is a new container for sdb that we do not want. md126 is the actual array and only contains sda, sdc, and sdd. We tried to use --stop and --remove to get rid of md125, then add sdb to md127, and assemble to see if it adds to md126. It does not.
>>
>> Below is the output of some commands for additional diagnostic information. Please let me know if you need more.
>>
>> Note: The output of these commands is after a fresh reboot, without/before all the other commands we tried to fix it. It gets reset to this state after every reboot we tried so far.
>>
>> uname -a
>> Linux aerospace-pr3d-app 4.4.0-194-generic #226-Ubuntu SMP Wed Oct 21 10:19:36 UTC 2020 x86_64 x86_64 x86_64 GNU/Linux
>>
>> sudo mdadm --version
>> mdadm - v4.1-126-gbdbe7f8 - 2021-03-09
>>
>> sudo smartctl -H -i -l scterc /dev/sda
>> smartctl 6.5 2016-01-24 r4214 [x86_64-linux-4.4.0-194-generic] (local build)
>> Copyright (C) 2002-16, Bruce Allen, Christian Franke, www.smartmontools.org
>>
>> === START OF INFORMATION SECTION ===
>> Device Model:     ST2000NX0253
>> Serial Number:    W461SCHM
>> LU WWN Device Id: 5 000c50 0b426d2d0
>> Firmware Version: SN05
>> User Capacity:    2,000,398,934,016 bytes [2.00 TB]
>> Sector Sizes:     512 bytes logical, 4096 bytes physical
>> Rotation Rate:    7200 rpm
>> Form Factor:      2.5 inches
>> Device is:        Not in smartctl database [for details use: -P showall]
>> ATA Version is:   ACS-3 (minor revision not indicated)
>> SATA Version is:  SATA 3.1, 6.0 Gb/s (current: 6.0 Gb/s)
>> Local Time is:    Thu Mar 11 15:07:30 2021 MST
>> SMART support is: Available - device has SMART capability.
>> SMART support is: Enabled
>>
>> === START OF READ SMART DATA SECTION ===
>> SMART overall-health self-assessment test result: PASSED
>>
>> SCT Error Recovery Control:
>>             Read:    100 (10.0 seconds)
>>            Write:    100 (10.0 seconds)
>>   
>>
>> sudo smartctl -H -i -l scterc /dev/sdb
>> smartctl 6.5 2016-01-24 r4214 [x86_64-linux-4.4.0-194-generic] (local build)
>> Copyright (C) 2002-16, Bruce Allen, Christian Franke, www.smartmontools.org
>>
>> === START OF INFORMATION SECTION ===
>> Device Model:     ST2000NX0253
>> Serial Number:    W462MZ0R
>> LU WWN Device Id: 5 000c50 0c569b51c
>> Firmware Version: SN05
>> User Capacity:    2,000,398,934,016 bytes [2.00 TB]
>> Sector Sizes:     512 bytes logical, 4096 bytes physical
>> Rotation Rate:    7200 rpm
>> Form Factor:      2.5 inches
>> Device is:        Not in smartctl database [for details use: -P showall]
>> ATA Version is:   ACS-3 (minor revision not indicated)
>> SATA Version is:  SATA 3.1, 6.0 Gb/s (current: 6.0 Gb/s)
>> Local Time is:    Thu Mar 11 15:09:34 2021 MST
>> SMART support is: Available - device has SMART capability.
>> SMART support is: Enabled
>>
>> === START OF READ SMART DATA SECTION ===
>> SMART overall-health self-assessment test result: PASSED
>>
>> SCT Error Recovery Control:
>>             Read:    100 (10.0 seconds)
>>            Write:    100 (10.0 seconds)
>>   
>> sudo smartctl -H -i -l scterc /dev/sdc
>> smartctl 6.5 2016-01-24 r4214 [x86_64-linux-4.4.0-194-generic] (local build)
>> Copyright (C) 2002-16, Bruce Allen, Christian Franke, www.smartmontools.org
>>
>> === START OF INFORMATION SECTION ===
>> Device Model:     ST2000NX0253
>> Serial Number:    W461NLPM
>> LU WWN Device Id: 5 000c50 0b426f335
>> Firmware Version: SN05
>> User Capacity:    2,000,398,934,016 bytes [2.00 TB]
>> Sector Sizes:     512 bytes logical, 4096 bytes physical
>> Rotation Rate:    7200 rpm
>> Form Factor:      2.5 inches
>> Device is:        Not in smartctl database [for details use: -P showall]
>> ATA Version is:   ACS-3 (minor revision not indicated)
>> SATA Version is:  SATA 3.1, 6.0 Gb/s (current: 6.0 Gb/s)
>> Local Time is:    Thu Mar 11 15:14:38 2021 MST
>> SMART support is: Available - device has SMART capability.
>> SMART support is: Enabled
>>
>> === START OF READ SMART DATA SECTION ===
>> SMART overall-health self-assessment test result: PASSED
>>
>> SCT Error Recovery Control:
>>             Read:    100 (10.0 seconds)
>>            Write:    100 (10.0 seconds)
>>
>> sudo smartctl -H -i -l scterc /dev/sdd
>> smartctl 6.5 2016-01-24 r4214 [x86_64-linux-4.4.0-194-generic] (local build)
>> Copyright (C) 2002-16, Bruce Allen, Christian Franke, www.smartmontools.org
>>
>> === START OF INFORMATION SECTION ===
>> Device Model:     ST2000NX0253
>> Serial Number:    W461NHAB
>> LU WWN Device Id: 5 000c50 0b426f8a4
>> Firmware Version: SN05
>> User Capacity:    2,000,398,934,016 bytes [2.00 TB]
>> Sector Sizes:     512 bytes logical, 4096 bytes physical
>> Rotation Rate:    7200 rpm
>> Form Factor:      2.5 inches
>> Device is:        Not in smartctl database [for details use: -P showall]
>> ATA Version is:   ACS-3 (minor revision not indicated)
>> SATA Version is:  SATA 3.1, 6.0 Gb/s (current: 6.0 Gb/s)
>> Local Time is:    Thu Mar 11 15:16:24 2021 MST
>> SMART support is: Available - device has SMART capability.
>> SMART support is: Enabled
>>
>> === START OF READ SMART DATA SECTION ===
>> SMART overall-health self-assessment test result: PASSED
>>
>> SCT Error Recovery Control:
>>             Read:    100 (10.0 seconds)
>>            Write:    100 (10.0 seconds)
>>   
>> sudo mdadm --examine /dev/sda
>> /dev/sda:
>>            Magic : Intel Raid ISM Cfg Sig.
>>          Version : 1.3.00
>>      Orig Family : 154b243e
>>           Family : 154b243e
>>       Generation : 000003aa
>>    Creation Time : Unknown
>>       Attributes : All supported
>>             UUID : 72360627:bb745f4c:aedafaab:e25d3123
>>         Checksum : 21ae5a2a correct
>>      MPB Sectors : 2
>>            Disks : 4
>>     RAID Devices : 1
>>
>>    Disk00 Serial : W461SCHM
>>            State : active
>>               Id : 00000000
>>      Usable Size : 3907018766 (1863.01 GiB 2000.39 GB)
>>
>> [Data]:
>>         Subarray : 0
>>             UUID : 764aa814:831953a1:06cf2a07:1ca42b2e
>>       RAID Level : 5 <-- 5
>>          Members : 4 <-- 4
>>            Slots : [U_UU] <-- [U_UU]
>>      Failed disk : 1
>>        This Slot : 0
>>      Sector Size : 512
>>       Array Size : 11135008768 (5.19 TiB 5.70 TB)
>>     Per Dev Size : 3711671808 (1769.86 GiB 1900.38 GB)
>>    Sector Offset : 0
>>      Num Stripes : 28997420
>>       Chunk Size : 64 KiB <-- 64 KiB
>>         Reserved : 0
>>    Migrate State : repair
>>        Map State : degraded <-- degraded
>>       Checkpoint : 462393 (512)
>>      Dirty State : dirty
>>       RWH Policy : off
>>        Volume ID : 1
>>
>>    Disk01 Serial : W461S13X:0
>>            State : active
>>               Id : ffffffff
>>      Usable Size : 3907018766 (1863.01 GiB 2000.39 GB)
>>
>>    Disk02 Serial : W461NLPM
>>            State : active
>>               Id : 00000002
>>      Usable Size : 3907018766 (1863.01 GiB 2000.39 GB)
>>
>>    Disk03 Serial : W461NHAB
>>            State : active
>>               Id : 00000003
>>      Usable Size : 3907018766 (1863.01 GiB 2000.39 GB)
>>
>> sudo mdadm --examine /dev/sdb
>> /dev/sdb:
>>            Magic : Intel Raid ISM Cfg Sig.
>>          Version : 1.0.00
>>      Orig Family : 00000000
>>           Family : e5cd8601
>>       Generation : 00000001
>>    Creation Time : Unknown
>>       Attributes : All supported
>>             UUID : 00000000:00000000:00000000:00000000
>>         Checksum : cb9b0c02 correct
>>      MPB Sectors : 1
>>            Disks : 1
>>     RAID Devices : 0
>>
>>    Disk00 Serial : W462MZ0R
>>            State : spare
>>               Id : 04000000
>>      Usable Size : 3907026958 (1863.02 GiB 2000.40 GB)
>>
>>      Disk Serial : W462MZ0R
>>            State : spare
>>               Id : 04000000
>>      Usable Size : 3907026958 (1863.02 GiB 2000.40 GB)
>>
>> sudo mdadm --examine /dev/sdc
>> /dev/sdc:
>>            Magic : Intel Raid ISM Cfg Sig.
>>          Version : 1.3.00
>>      Orig Family : 154b243e
>>           Family : 154b243e
>>       Generation : 000003aa
>>    Creation Time : Unknown
>>       Attributes : All supported
>>             UUID : 72360627:bb745f4c:aedafaab:e25d3123
>>         Checksum : 21ae5a2a correct
>>      MPB Sectors : 2
>>            Disks : 4
>>     RAID Devices : 1
>>
>>    Disk02 Serial : W461NLPM
>>            State : active
>>               Id : 00000002
>>      Usable Size : 3907018766 (1863.01 GiB 2000.39 GB)
>>
>> [Data]:
>>         Subarray : 0
>>             UUID : 764aa814:831953a1:06cf2a07:1ca42b2e
>>       RAID Level : 5 <-- 5
>>          Members : 4 <-- 4
>>            Slots : [U_UU] <-- [U_UU]
>>      Failed disk : 1
>>        This Slot : 2
>>      Sector Size : 512
>>       Array Size : 11135008768 (5.19 TiB 5.70 TB)
>>     Per Dev Size : 3711671808 (1769.86 GiB 1900.38 GB)
>>    Sector Offset : 0
>>      Num Stripes : 28997420
>>       Chunk Size : 64 KiB <-- 64 KiB
>>         Reserved : 0
>>    Migrate State : repair
>>        Map State : degraded <-- degraded
>>       Checkpoint : 462393 (512)
>>      Dirty State : dirty
>>       RWH Policy : off
>>        Volume ID : 1
>>
>>    Disk00 Serial : W461SCHM
>>            State : active
>>               Id : 00000000
>>      Usable Size : 3907018766 (1863.01 GiB 2000.39 GB)
>>
>>    Disk01 Serial : W461S13X:0
>>            State : active
>>               Id : ffffffff
>>      Usable Size : 3907018766 (1863.01 GiB 2000.39 GB)
>>
>>    Disk03 Serial : W461NHAB
>>            State : active
>>               Id : 00000003
>>      Usable Size : 3907018766 (1863.01 GiB 2000.39 GB)
>>
>> sudo mdadm --examine /dev/sdd
>> /dev/sdd:
>>            Magic : Intel Raid ISM Cfg Sig.
>>          Version : 1.3.00
>>      Orig Family : 154b243e
>>           Family : 154b243e
>>       Generation : 000003aa
>>    Creation Time : Unknown
>>       Attributes : All supported
>>             UUID : 72360627:bb745f4c:aedafaab:e25d3123
>>         Checksum : 21ae5a2a correct
>>      MPB Sectors : 2
>>            Disks : 4
>>     RAID Devices : 1
>>
>>    Disk03 Serial : W461NHAB
>>            State : active
>>               Id : 00000003
>>      Usable Size : 3907018766 (1863.01 GiB 2000.39 GB)
>>
>> [Data]:
>>         Subarray : 0
>>             UUID : 764aa814:831953a1:06cf2a07:1ca42b2e
>>       RAID Level : 5 <-- 5
>>          Members : 4 <-- 4
>>            Slots : [U_UU] <-- [U_UU]
>>      Failed disk : 1
>>        This Slot : 3
>>      Sector Size : 512
>>       Array Size : 11135008768 (5.19 TiB 5.70 TB)
>>     Per Dev Size : 3711671808 (1769.86 GiB 1900.38 GB)
>>    Sector Offset : 0
>>      Num Stripes : 28997420
>>       Chunk Size : 64 KiB <-- 64 KiB
>>         Reserved : 0
>>    Migrate State : repair
>>        Map State : degraded <-- degraded
>>       Checkpoint : 462393 (512)
>>      Dirty State : dirty
>>       RWH Policy : off
>>        Volume ID : 1
>>
>>    Disk00 Serial : W461SCHM
>>            State : active
>>               Id : 00000000
>>      Usable Size : 3907018766 (1863.01 GiB 2000.39 GB)
>>
>>    Disk01 Serial : W461S13X:0
>>            State : active
>>               Id : ffffffff
>>      Usable Size : 3907018766 (1863.01 GiB 2000.39 GB)
>>
>>    Disk02 Serial : W461NLPM
>>            State : active
>>               Id : 00000002
>>      Usable Size : 3907018766 (1863.01 GiB 2000.39 GB)
>> sudo mdadm --detail /dev/md125
>> /dev/md125:
>>             Version : imsm
>>          Raid Level : container
>>       Total Devices : 1
>>
>>     Working Devices : 1
>>
>>       Member Arrays :
>>
>>      Number   Major   Minor   RaidDevice
>>
>>         -       8       16        -        /dev/sdb
>>
>> sudo mdadm --detail /dev/md126
>> /dev/md126:
>>           Container : /dev/md/imsm0, member 0
>>          Raid Level : raid5
>>       Used Dev Size : 1855835904 (1769.86 GiB 1900.38 GB)
>>        Raid Devices : 4
>>       Total Devices : 3
>>
>>               State : active, FAILED, Not Started
>>      Active Devices : 3
>>     Working Devices : 3
>>      Failed Devices : 0
>>       Spare Devices : 0
>>
>>              Layout : left-asymmetric
>>          Chunk Size : 64K
>>
>> Consistency Policy : unknown
>>
>>
>>                UUID : 764aa814:831953a1:06cf2a07:1ca42b2e
>>      Number   Major   Minor   RaidDevice State
>>         -       0        0        0      removed
>>         -       0        0        1      removed
>>         -       0        0        2      removed
>>         -       0        0        3      removed
>>
>>         -       8        0        0      sync   /dev/sda
>>         -       8       32        2      sync   /dev/sdc
>>         -       8       48        3      sync   /dev/sdd
>>
>> sudo mdadm --detail /dev/md127
>> /dev/md127:
>>             Version : imsm
>>          Raid Level : container
>>       Total Devices : 3
>>
>>     Working Devices : 3
>>
>>
>>                UUID : 72360627:bb745f4c:aedafaab:e25d3123
>>       Member Arrays : /dev/md126
>>
>>      Number   Major   Minor   RaidDevice
>>
>>         -       8        0        -        /dev/sda
>>         -       8       32        -        /dev/sdc
>>         -       8       48        -        /dev/sdd
>>   
>> lsdrv
>> **Warning** The following utility(ies) failed to execute:
>>    sginfo
>> Some information may be missing.
>>
>> PCI [nvme] 04:00.0 Non-Volatile memory controller: Samsung Electronics Co Ltd NVMe SSD Controller SM981/PM981
>> └nvme nvme0 Samsung SSD 970 EVO Plus 500GB           {S4P2NF0M501223D}
>>   └nvme0n1 465.76g [259:0] Empty/Unknown
>>    ├nvme0n1p1 512.00m [259:1] Empty/Unknown
>>    │└Mounted as /dev/nvme0n1p1 @ /boot/efi
>>    ├nvme0n1p2 732.00m [259:2] Empty/Unknown
>>    │└Mounted as /dev/nvme0n1p2 @ /boot
>>    └nvme0n1p3 464.54g [259:3] Empty/Unknown
>>     ├dm-0 463.59g [252:0] Empty/Unknown
>>     │└Mounted as /dev/mapper/customer--pr3d--app--vg-root @ /
>>     └dm-1 980.00m [252:1] Empty/Unknown
>> PCI [ahci] 00:11.5 SATA controller: Intel Corporation C620 Series Chipset Family SSATA Controller [AHCI mode] (rev 09)
>> └scsi 0:x:x:x [Empty]
>> PCI [ahci] 00:17.0 RAID bus controller: Intel Corporation C600/X79 series chipset SATA RAID Controller (rev 09)
>> ├scsi 2:0:0:0 ATA      ST2000NX0253
>> │└sda 1.82t [8:0] Empty/Unknown
>> │ ├md126 0.00k [9:126] MD vexternal:/md127/0 raid5 (4) inactive, 64k Chunk, None (None) None {None}
>> │ │                    Empty/Unknown
>> │ └md127 0.00k [9:127] MD vexternal:imsm  () inactive, None (None) None {None}
>> │                      Empty/Unknown
>> ├scsi 3:0:0:0 ATA      ST2000NX0253
>> │└sdb 1.82t [8:16] Empty/Unknown
>> │ └md125 0.00k [9:125] MD vexternal:imsm  () inactive, None (None) None {None}
>> │                      Empty/Unknown
>> ├scsi 4:0:0:0 ATA      ST2000NX0253
>> │└sdc 1.82t [8:32] Empty/Unknown
>> │ ├md126 0.00k [9:126] MD vexternal:/md127/0 raid5 (4) inactive, 64k Chunk, None (None) None {None}
>> │ │                    Empty/Unknown
>> ├scsi 5:0:0:0 ATA      ST2000NX0253
>> │└sdd 1.82t [8:48] Empty/Unknown
>> │ ├md126 0.00k [9:126] MD vexternal:/md127/0 raid5 (4) inactive, 64k Chunk, None (None) None {None}
>> │ │                    Empty/Unknown
>> └scsi 6:0:0:0 Slimtype DVD A  DS8ACSH
>>   └sr0 1.00g [11:0] Empty/Unknown
>> Other Block Devices
>> ├loop0 0.00k [7:0] Empty/Unknown
>> ├loop1 0.00k [7:1] Empty/Unknown
>> ├loop2 0.00k [7:2] Empty/Unknown
>> ├loop3 0.00k [7:3] Empty/Unknown
>> ├loop4 0.00k [7:4] Empty/Unknown
>> ├loop5 0.00k [7:5] Empty/Unknown
>> ├loop6 0.00k [7:6] Empty/Unknown
>> └loop7 0.00k [7:7] Empty/Unknown
>>
>> cat /proc/mdstat
>> Personalities : [raid6] [raid5] [raid4] [linear] [multipath] [raid0] [raid1] [raid10]
>> md125 : inactive sdb[0](S)
>>        1105 blocks super external:imsm
>>
>> md126 : inactive sda[2] sdc[1] sdd[0]
>>        5567507712 blocks super external:/md127/0
>>
>> md127 : inactive sdc[2](S) sdd[1](S) sda[0](S)
>>        9459 blocks super external:imsm
>>
>> unused devices: <none>
>>
>>
>> Thanks for the help!
>> Devon Beets
> 
> Hi Devon,
> 
> The array is in dirty degraded state. It does not start automatically because
> there is a risk of silent data corruption, i.e. RAID write hole. You can force
> it to start with:
> 
> # mdadm -R --force /dev/md126
> 
> You will need mdadm built with this commit for it to work:
> 
> https://git.kernel.org/pub/scm/utils/mdadm/mdadm.git/commit/?id=7b99edab2834d5d08ef774b4cff784caaa1a186f
> 
> It may be a good idea to copy the array contents with dd before you fsck or
> mount the filesystem in case the recovery goes wrong.
> 
> Then stop the second container and add the new drive to the array:
> 
> # mdadm -S /dev/md125
> # mdadm -a /dev/md127 /dev/sdb
> 
> Rebuild should begin at this point.
> 
> Regards,
> Artur
> 
>

Hello,
also see:
https://lore.kernel.org/linux-raid/ac370d79-95e8-d0a1-0991-
fb12b128818c@xxxxxxxxxxxxxxx/T/#t

Mariusz




[Index of Archives]     [Linux RAID Wiki]     [ATA RAID]     [Linux SCSI Target Infrastructure]     [Linux Block]     [Linux IDE]     [Linux SCSI]     [Linux Hams]     [Device Mapper]     [Device Mapper Cryptographics]     [Kernel]     [Linux Admin]     [Linux Net]     [GFS]     [RPM]     [git]     [Yosemite Forum]


  Powered by Linux