Revive a dead md raid5 array

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi,

New to this list, I understand that I can post a problem with an md raid array here. If I'm wrong in that supposition, please let me know, and accept my apologies!

I have a 4-disk rad5 array, of which two of the disks are kicked out because of read errors. The disks are WD Blue 4TB disks, which are still under guarantee.

I have reasonable recent backups, but yet I would like to try to get the array alive again.

Funny thing is, mdadm -- examine states the array as being raid0:

   /dev/md0:
               Version : 1.2
            Raid Level : raid0
         Total Devices : 4
           Persistence : Superblock is persistent

                 State : inactive
       Working Devices : 4

                  Name : souder:0  (local to host souder)
                  UUID : e9c5ac38:affff54a:730669a1:e2462279
                Events : 1273316

        Number   Major   Minor   RaidDevice

           -       8        0        -        /dev/sda
           -       8       80        -        /dev/sdf
           -       8       48        -        /dev/sdd
           -       8       16        -        /dev/sdb

while the individual disk are, correctly, stated as being part of an raid5 array:

   /dev/sda:
              Magic : a92b4efc
            Version : 1.2
        Feature Map : 0x9
         Array UUID : e9c5ac38:affff54a:730669a1:e2462279
               Name : souder:0  (local to host souder)
      Creation Time : Fri Apr 28 09:59:07 2017
         Raid Level : raid5
       Raid Devices : 4

     Avail Dev Size : 7813775024 (3725.90 GiB 4000.65 GB)
         Array Size : 11720649216 (11177.68 GiB 12001.94 GB)
      Used Dev Size : 7813766144 (3725.89 GiB 4000.65 GB)
        Data Offset : 262144 sectors
       Super Offset : 8 sectors
       Unused Space : before=262064 sectors, after=8880 sectors
              State : clean
        Device UUID : 0c6e6682:c99f679e:64256d78:b8a4188d

   Internal Bitmap : 8 sectors from superblock
        Update Time : Mon Oct 29 22:08:59 2018
      Bad Block Log : 512 entries available at offset 24 sectors - bad
   blocks present.
           Checksum : 5ad156f1 - correct
             Events : 1273316

             Layout : left-symmetric
         Chunk Size : 512K

       Device Role : Active device 0
       Array State : AA.. ('A' == active, '.' == missing, 'R' == replacing)

   /dev/sdb:

              Magic : a92b4efc
            Version : 1.2
        Feature Map : 0x1
         Array UUID : e9c5ac38:affff54a:730669a1:e2462279
               Name : souder:0  (local to host souder)
      Creation Time : Fri Apr 28 09:59:07 2017
         Raid Level : raid5
       Raid Devices : 4

     Avail Dev Size : 7813775024 (3725.90 GiB 4000.65 GB)
         Array Size : 11720649216 (11177.68 GiB 12001.94 GB)
      Used Dev Size : 7813766144 (3725.89 GiB 4000.65 GB)
        Data Offset : 262144 sectors
       Super Offset : 8 sectors
       Unused Space : before=262064 sectors, after=8880 sectors
              State : clean
        Device UUID : fac8aa17:ab74864c:faa738fc:062f3311

   Internal Bitmap : 8 sectors from superblock
        Update Time : Mon Oct 29 22:08:59 2018
      Bad Block Log : 512 entries available at offset 24 sectors
           Checksum : a61a92ed - correct
             Events : 1273316

             Layout : left-symmetric
         Chunk Size : 512K

       Device Role : Active device 1
       Array State : AA.. ('A' == active, '.' == missing, 'R' == replacing)

   /dev/sdd:

              Magic : a92b4efc
            Version : 1.2
        Feature Map : 0x9
         Array UUID : e9c5ac38:affff54a:730669a1:e2462279
               Name : souder:0  (local to host souder)
      Creation Time : Fri Apr 28 09:59:07 2017
         Raid Level : raid5
       Raid Devices : 4

     Avail Dev Size : 7813775024 (3725.90 GiB 4000.65 GB)
         Array Size : 11720649216 (11177.68 GiB 12001.94 GB)
      Used Dev Size : 7813766144 (3725.89 GiB 4000.65 GB)
        Data Offset : 262144 sectors
       Super Offset : 8 sectors
       Unused Space : before=262064 sectors, after=8880 sectors
              State : clean
        Device UUID : 28cea517:e4a1033c:5a7d638e:7d48011a

   Internal Bitmap : 8 sectors from superblock
        Update Time : Mon Oct 29 22:08:59 2018
      Bad Block Log : 512 entries available at offset 24 sectors - bad
   blocks present.
           Checksum : 308bb45c - correct
             Events : 1273316

             Layout : left-symmetric
         Chunk Size : 512K

       Device Role : spare
       Array State : AA.. ('A' == active, '.' == missing, 'R' == replacing)

   /dev/sdf:

              Magic : a92b4efc
            Version : 1.2
        Feature Map : 0x1
         Array UUID : e9c5ac38:affff54a:730669a1:e2462279
               Name : souder:0  (local to host souder)
      Creation Time : Fri Apr 28 09:59:07 2017
         Raid Level : raid5
       Raid Devices : 4

     Avail Dev Size : 7813766767 (3725.89 GiB 4000.65 GB)
         Array Size : 11720649216 (11177.68 GiB 12001.94 GB)
      Used Dev Size : 7813766144 (3725.89 GiB 4000.65 GB)
        Data Offset : 262144 sectors
       Super Offset : 8 sectors
       Unused Space : before=262064 sectors, after=623 sectors
              State : active
        Device UUID : 2126db94:d32bc3da:716a1b20:35446904

   Internal Bitmap : 8 sectors from superblock
        Update Time : Mon Oct 29 22:01:42 2018
      Bad Block Log : 512 entries available at offset 24 sectors
           Checksum : c89e599c - correct
             Events : 1272457

             Layout : left-symmetric
         Chunk Size : 512K

       Device Role : Active device 3
       Array State : AAAA ('A' == active, '.' == missing, 'R' == replacing)

=================================================================

Maybe you noticed the fact that all disks are marked as spare, and that the event count of one of the disks /dev/sdf is different from the other's

I found some more occurrences of a raid5 being recognized as a raid0 device, but not a real solution to this.

lsdrv gives

   lsdrv:

   PCI [ahci] 00:11.0 SATA controller: Advanced Micro Devices, Inc.
   [AMD/ATI] SB7x0/SB8x0/SB9x0 SATA Controller [IDE mode]
   ├scsi 0:0:0:0 ATA      WDC WD40EZRZ-00W {WD-WCC4E2RNX8LD}
   │└sda 3.64t [8:0] MD  (none/) (w/ sdf,sdd,sdb) spare 'souder:0'
   {e9c5ac38-afff-f54a-7306-69a1e2462279}
   │ └md0 0.00k [9:0] MD v1.2  () inactive, None (None) None
   {e9c5ac38:affff54a:730669a1:e2462279}
   │                  Empty/Unknown
   ├scsi 1:0:0:0 ATA      WDC WD40EZRZ-00W {WD-WCC4E3UK9RV2}
   │└sdb 3.64t [8:16] MD  (none/) (w/ sdf,sdd,sda) spare 'souder:0'
   {e9c5ac38-afff-f54a-7306-69a1e2462279}
   │ └md0 0.00k [9:0] MD v1.2  () inactive, None (None) None
   {e9c5ac38:affff54a:730669a1:e2462279}
   │                  Empty/Unknown
   ├scsi 2:0:0:0 ATA      SAMSUNG SSD 830 {S0Z3NEAC826655}
   │└sdc 119.24g [8:32] Partitioned (dos)
   │ └sdc1 119.24g [8:33] Partitioned (dos)
   {1a310d0c-e47f-408e-8f93-f77cbf694324}
   │  └Mounted as /dev/sdc1 @ /
   └scsi 3:0:0:0 ATA      WDC WD40EZRZ-00W {WD-WCC4E1PC25R4}
     └sdd 3.64t [8:48] MD  (none/) (w/ sdf,sdb,sda) spare 'souder:0'
   {e9c5ac38-afff-f54a-7306-69a1e2462279}
      └md0 0.00k [9:0] MD v1.2  () inactive, None (None) None
   {e9c5ac38:affff54a:730669a1:e2462279}
                       Empty/Unknown
   PCI [ahci] 04:00.0 SATA controller: JMicron Technology Corp. JMB363
   SATA/IDE Controller (rev 02)
   └scsi 4:x:x:x [Empty]
   PCI [pata_jmicron] 04:00.1 IDE interface: JMicron Technology Corp.
   JMB363 SATA/IDE Controller (rev 02)
   └scsi 6:x:x:x [Empty]
   PCI [pata_atiixp] 00:14.1 IDE interface: Advanced Micro Devices,
   Inc. [AMD/ATI] SB7x0/SB8x0/SB9x0 IDE Controller
   ├scsi 8:0:0:0 PLEXTOR  DVDR   PX-740A   {508084151561}
   │└sr0 1.00g [11:0] Empty/Unknown
   └scsi 9:0:1:0 ATA      WDC WD40EZRZ-00W {WD-WCC4E2EH8NZS}
     └sdf 3.64t [8:80] MD  (none/) (w/ sdd,sdb,sda) spare 'souder:0'
   {e9c5ac38-afff-f54a-7306-69a1e2462279}
      └md0 0.00k [9:0] MD v1.2  () inactive, None (None) None
   {e9c5ac38:affff54a:730669a1:e2462279}
                       Empty/Unknown
   USB [usb-storage] Bus 002 Device 005: ID 0930:6545 Toshiba Corp.
   Kingston DataTraveler 102/2.0 / HEMA Flash Drive 2 GB / PNY Attache
   4GB Stick {0240897052B3795A}
   └scsi 10:0:0:0          USB Flash Memory
     └sdg 3.84g [8:96] Partitioned (dos)
      └sdg1 3.84g [8:97] ext4 {76746e9b-0564-4c74-938a-06e5789154db}
       └Mounted as /dev/sdg1 @ /mnt/sdg1

=============================================================================================

The faulty disks are /dev/sda an /dev/sdd, and I copied the contents to new WD RED 4TB disks, with

ddrescue -d -s <size-of-target-disk> -f /dev/sd<source> /dev/sd<target> sd<source>.map

The size argument because the new disks are some 4MB smaller than the original.

ddrescue saw one one disk 14, on the other 54 read errors, and copied 99.99% of the source.

Is the a way to revive the array, and if yes, how can I do that?

Many thanks for reading this post, and for any answers!

regards, Jogchum Reitsma





[Index of Archives]     [Linux RAID Wiki]     [ATA RAID]     [Linux SCSI Target Infrastructure]     [Linux Block]     [Linux IDE]     [Linux SCSI]     [Linux Hams]     [Device Mapper]     [Device Mapper Cryptographics]     [Kernel]     [Linux Admin]     [Linux Net]     [GFS]     [RPM]     [git]     [Yosemite Forum]


  Powered by Linux