Re: Help restoring a raid10 Array (4 disk + one spare) after a hard disk failure at power on

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



after the sdd failure, you rebooted your server?

Do you have /etc/mdadm.conf(or in other location) that explicitly list
the sdb1 , sdd1 etc as the md0? If so, maybe you can comment it and
then rescan the devices. If the spare disk is defined at the
beginning, or added afterward as "spare", it should start rebuilding
right away after a disk failure.

The comands to rescan  may look like:
mdadm --stop /dev/md0
mdadm --assemble --scan

You can use UUID instead of sdb, or sdc, etc.

Other experts on the list may give you more helpful info.

Best,

Feng

On Tue, May 14, 2019 at 11:49 AM Eric Valette <eric.valette@xxxxxxx> wrote:
>
> I have a dedicated hardware nas that runs a self maintained debian 10.
>
> before the hardware disk problem (before/after)
>
> sda : system disk OK/OK no raid
> sdb : first disk of the raid10 array OK/OK
> sdc : second disk of the raid10 array OK/OK
> sdd : third disk of the raid10 array OK/KO
> sde : fourth disk of the raid10 array OK/OK but is now sdd
> sdf : spare disk for the array is now sde
>
> After the failure the BIOS does not detect the original third disk. Disk
> are renamed and I think sde has become sdd and sdf -> sde
>
> Below are more detailed info. Feel free to ask for other things as I can
> log into the machine via ssh
>
> So I have several questions :
>
>         1) How to I repair the raid10 array using the spare disk without
> replacing the faulty one immediately?
>         2) What should I do once I receive the new disk (hopefully soon)
>         3) Is there a way to use persistent naming for disk array?
>
> Sorry to annoy you but my kid wants to see a film on the nas and annoys
> me badly. And I prefer to ask rather than doing mistakes.
>
> Thanks for any
>
>
>
> mdadm --examine /dev/sdb
> /dev/sdb:
>     MBR Magic : aa55
> Partition[0] :   4294967295 sectors at            1 (type ee)
> root@nas2:~# mdadm --examine /dev/sdb
> /dev/sdb:
>     MBR Magic : aa55
> Partition[0] :   4294967295 sectors at            1 (type ee)
> root@nas2:~# mdadm --examine /dev/sdb1
> /dev/sdb1:
>            Magic : a92b4efc
>          Version : 1.2
>      Feature Map : 0x0
>       Array UUID : 6abe1f20:90c629de:fadd8dc0:ca14c928
>             Name : nas2:0  (local to host nas2)
>    Creation Time : Wed Jun 20 23:56:59 2012
>       Raid Level : raid10
>     Raid Devices : 4
>
>   Avail Dev Size : 5860268943 (2794.39 GiB 3000.46 GB)
>       Array Size : 5860268032 (5588.79 GiB 6000.91 GB)
>    Used Dev Size : 5860268032 (2794.39 GiB 3000.46 GB)
>      Data Offset : 262144 sectors
>     Super Offset : 8 sectors
>     Unused Space : before=262064 sectors, after=911 sectors
>            State : clean
>      Device UUID : ce9d878a:37a4f3a3:936bd905:c4ed9970
>
>      Update Time : Wed May  8 11:39:40 2019
>         Checksum : cf841c9f - correct
>           Events : 1193
>
>           Layout : near=2
>       Chunk Size : 512K
>
>     Device Role : Active device 0
>     Array State : AAAA ('A' == active, '.' == missing, 'R' == replacing)
> root@nas2:~# mdadm --examine /dev/sdc
> /dev/sdc:
>     MBR Magic : aa55
> Partition[0] :   4294967295 sectors at            1 (type ee)
> root@nas2:~# mdadm --examine /dev/sdc1
> /dev/sdc1:
>            Magic : a92b4efc
>          Version : 1.2
>      Feature Map : 0x0
>       Array UUID : 6abe1f20:90c629de:fadd8dc0:ca14c928
>             Name : nas2:0  (local to host nas2)
>    Creation Time : Wed Jun 20 23:56:59 2012
>       Raid Level : raid10
>     Raid Devices : 4
>
>   Avail Dev Size : 5860268943 (2794.39 GiB 3000.46 GB)
>       Array Size : 5860268032 (5588.79 GiB 6000.91 GB)
>    Used Dev Size : 5860268032 (2794.39 GiB 3000.46 GB)
>      Data Offset : 262144 sectors
>     Super Offset : 8 sectors
>     Unused Space : before=262064 sectors, after=911 sectors
>            State : clean
>      Device UUID : 8c89bdf8:4f3f8ace:c15b5634:7a874071
>
>      Update Time : Wed May  8 11:39:40 2019
>         Checksum : 97744edb - correct
>           Events : 1193
>
>           Layout : near=2
>       Chunk Size : 512K
>
>     Device Role : Active device 1
>     Array State : AAAA ('A' == active, '.' == missing, 'R' == replacing)
> root@nas2:~# mdadm --examine /dev/sdd
> /dev/sdd:
>     MBR Magic : aa55
> Partition[0] :   4294967295 sectors at            1 (type ee)
> root@nas2:~# mdadm --examine /dev/sdd1
> /dev/sdd1:
>            Magic : a92b4efc
>          Version : 1.2
>      Feature Map : 0x0
>       Array UUID : 6abe1f20:90c629de:fadd8dc0:ca14c928
>             Name : nas2:0  (local to host nas2)
>    Creation Time : Wed Jun 20 23:56:59 2012
>       Raid Level : raid10
>     Raid Devices : 4
>
>   Avail Dev Size : 5860268943 (2794.39 GiB 3000.46 GB)
>       Array Size : 5860268032 (5588.79 GiB 6000.91 GB)
>    Used Dev Size : 5860268032 (2794.39 GiB 3000.46 GB)
>      Data Offset : 262144 sectors
>     Super Offset : 8 sectors
>     Unused Space : before=262064 sectors, after=911 sectors
>            State : clean
>      Device UUID : c97b767a:84d2e7e2:52557d30:51c39784
>
>      Update Time : Wed May  8 11:39:40 2019
>         Checksum : 3d08e837 - correct
>           Events : 1193
>
>           Layout : near=2
>       Chunk Size : 512K
>
>     Device Role : Active device 3
>     Array State : AAAA ('A' == active, '.' == missing, 'R' == replacing)
> root@nas2:~# mdadm --examine /dev/sde
> /dev/sde:
>     MBR Magic : aa55
> Partition[0] :   4294967295 sectors at            1 (type ee)
> root@nas2:~# mdadm --examine /dev/sde1
> /dev/sde1:
>            Magic : a92b4efc
>          Version : 1.2
>      Feature Map : 0x0
>       Array UUID : 6abe1f20:90c629de:fadd8dc0:ca14c928
>             Name : nas2:0  (local to host nas2)
>    Creation Time : Wed Jun 20 23:56:59 2012
>       Raid Level : raid10
>     Raid Devices : 4
>
>   Avail Dev Size : 5860268943 (2794.39 GiB 3000.46 GB)
>       Array Size : 5860268032 (5588.79 GiB 6000.91 GB)
>    Used Dev Size : 5860268032 (2794.39 GiB 3000.46 GB)
>      Data Offset : 262144 sectors
>     Super Offset : 8 sectors
>     Unused Space : before=262064 sectors, after=911 sectors
>            State : clean
>      Device UUID : 82667e81:a6158319:85e0282e:845eec1c
>
>      Update Time : Wed May  8 11:00:29 2019
>         Checksum : 10ac3349 - correct
>           Events : 1193
>
>           Layout : near=2
>       Chunk Size : 512K
>
>     Device Role : spare
>     Array State : AAAA ('A' == active, '.' == missing, 'R' == replacing)
> root@nas2:~#
>
> mdadm --detail /dev/md0
> /dev/md0:
>             Version : 1.2
>          Raid Level : raid0
>       Total Devices : 4
>         Persistence : Superblock is persistent
>
>               State : inactive
>     Working Devices : 4
>
>                Name : nas2:0  (local to host nas2)
>                UUID : 6abe1f20:90c629de:fadd8dc0:ca14c928
>              Events : 1193
>
>      Number   Major   Minor   RaidDevice
>
>         -       8       65        -        /dev/sde1
>         -       8       49        -        /dev/sdd1
>         -       8       33        -        /dev/sdc1
>         -       8       17        -        /dev/sdb1
>
> cat /proc/mdstat
> Personalities : [raid10]
> md0 : inactive sdc1[1](S) sdb1[0](S) sde1[4](S) sdd1[3](S)
>        11720537886 blocks super 1.2
>
> unused devices: <none>
>
> --
>     __
>    /  `                         Eric Valette
>   /--   __  o _.                6 rue Paul Le Flem
> (___, / (_(_(__                 35740 Pace
>
> Tel: +33 (0)2 99 85 26 76       Fax: +33 (0)2 99 85 26 76
> E-mail: eric.valette@xxxxxxx
>
>
>



[Index of Archives]     [Linux RAID Wiki]     [ATA RAID]     [Linux SCSI Target Infrastructure]     [Linux Block]     [Linux IDE]     [Linux SCSI]     [Linux Hams]     [Device Mapper]     [Device Mapper Cryptographics]     [Kernel]     [Linux Admin]     [Linux Net]     [GFS]     [RPM]     [git]     [Yosemite Forum]


  Powered by Linux