after the sdd failure, you rebooted your server? Do you have /etc/mdadm.conf(or in other location) that explicitly list the sdb1 , sdd1 etc as the md0? If so, maybe you can comment it and then rescan the devices. If the spare disk is defined at the beginning, or added afterward as "spare", it should start rebuilding right away after a disk failure. The comands to rescan may look like: mdadm --stop /dev/md0 mdadm --assemble --scan You can use UUID instead of sdb, or sdc, etc. Other experts on the list may give you more helpful info. Best, Feng On Tue, May 14, 2019 at 11:49 AM Eric Valette <eric.valette@xxxxxxx> wrote: > > I have a dedicated hardware nas that runs a self maintained debian 10. > > before the hardware disk problem (before/after) > > sda : system disk OK/OK no raid > sdb : first disk of the raid10 array OK/OK > sdc : second disk of the raid10 array OK/OK > sdd : third disk of the raid10 array OK/KO > sde : fourth disk of the raid10 array OK/OK but is now sdd > sdf : spare disk for the array is now sde > > After the failure the BIOS does not detect the original third disk. Disk > are renamed and I think sde has become sdd and sdf -> sde > > Below are more detailed info. Feel free to ask for other things as I can > log into the machine via ssh > > So I have several questions : > > 1) How to I repair the raid10 array using the spare disk without > replacing the faulty one immediately? > 2) What should I do once I receive the new disk (hopefully soon) > 3) Is there a way to use persistent naming for disk array? > > Sorry to annoy you but my kid wants to see a film on the nas and annoys > me badly. And I prefer to ask rather than doing mistakes. > > Thanks for any > > > > mdadm --examine /dev/sdb > /dev/sdb: > MBR Magic : aa55 > Partition[0] : 4294967295 sectors at 1 (type ee) > root@nas2:~# mdadm --examine /dev/sdb > /dev/sdb: > MBR Magic : aa55 > Partition[0] : 4294967295 sectors at 1 (type ee) > root@nas2:~# mdadm --examine /dev/sdb1 > /dev/sdb1: > Magic : a92b4efc > Version : 1.2 > Feature Map : 0x0 > Array UUID : 6abe1f20:90c629de:fadd8dc0:ca14c928 > Name : nas2:0 (local to host nas2) > Creation Time : Wed Jun 20 23:56:59 2012 > Raid Level : raid10 > Raid Devices : 4 > > Avail Dev Size : 5860268943 (2794.39 GiB 3000.46 GB) > Array Size : 5860268032 (5588.79 GiB 6000.91 GB) > Used Dev Size : 5860268032 (2794.39 GiB 3000.46 GB) > Data Offset : 262144 sectors > Super Offset : 8 sectors > Unused Space : before=262064 sectors, after=911 sectors > State : clean > Device UUID : ce9d878a:37a4f3a3:936bd905:c4ed9970 > > Update Time : Wed May 8 11:39:40 2019 > Checksum : cf841c9f - correct > Events : 1193 > > Layout : near=2 > Chunk Size : 512K > > Device Role : Active device 0 > Array State : AAAA ('A' == active, '.' == missing, 'R' == replacing) > root@nas2:~# mdadm --examine /dev/sdc > /dev/sdc: > MBR Magic : aa55 > Partition[0] : 4294967295 sectors at 1 (type ee) > root@nas2:~# mdadm --examine /dev/sdc1 > /dev/sdc1: > Magic : a92b4efc > Version : 1.2 > Feature Map : 0x0 > Array UUID : 6abe1f20:90c629de:fadd8dc0:ca14c928 > Name : nas2:0 (local to host nas2) > Creation Time : Wed Jun 20 23:56:59 2012 > Raid Level : raid10 > Raid Devices : 4 > > Avail Dev Size : 5860268943 (2794.39 GiB 3000.46 GB) > Array Size : 5860268032 (5588.79 GiB 6000.91 GB) > Used Dev Size : 5860268032 (2794.39 GiB 3000.46 GB) > Data Offset : 262144 sectors > Super Offset : 8 sectors > Unused Space : before=262064 sectors, after=911 sectors > State : clean > Device UUID : 8c89bdf8:4f3f8ace:c15b5634:7a874071 > > Update Time : Wed May 8 11:39:40 2019 > Checksum : 97744edb - correct > Events : 1193 > > Layout : near=2 > Chunk Size : 512K > > Device Role : Active device 1 > Array State : AAAA ('A' == active, '.' == missing, 'R' == replacing) > root@nas2:~# mdadm --examine /dev/sdd > /dev/sdd: > MBR Magic : aa55 > Partition[0] : 4294967295 sectors at 1 (type ee) > root@nas2:~# mdadm --examine /dev/sdd1 > /dev/sdd1: > Magic : a92b4efc > Version : 1.2 > Feature Map : 0x0 > Array UUID : 6abe1f20:90c629de:fadd8dc0:ca14c928 > Name : nas2:0 (local to host nas2) > Creation Time : Wed Jun 20 23:56:59 2012 > Raid Level : raid10 > Raid Devices : 4 > > Avail Dev Size : 5860268943 (2794.39 GiB 3000.46 GB) > Array Size : 5860268032 (5588.79 GiB 6000.91 GB) > Used Dev Size : 5860268032 (2794.39 GiB 3000.46 GB) > Data Offset : 262144 sectors > Super Offset : 8 sectors > Unused Space : before=262064 sectors, after=911 sectors > State : clean > Device UUID : c97b767a:84d2e7e2:52557d30:51c39784 > > Update Time : Wed May 8 11:39:40 2019 > Checksum : 3d08e837 - correct > Events : 1193 > > Layout : near=2 > Chunk Size : 512K > > Device Role : Active device 3 > Array State : AAAA ('A' == active, '.' == missing, 'R' == replacing) > root@nas2:~# mdadm --examine /dev/sde > /dev/sde: > MBR Magic : aa55 > Partition[0] : 4294967295 sectors at 1 (type ee) > root@nas2:~# mdadm --examine /dev/sde1 > /dev/sde1: > Magic : a92b4efc > Version : 1.2 > Feature Map : 0x0 > Array UUID : 6abe1f20:90c629de:fadd8dc0:ca14c928 > Name : nas2:0 (local to host nas2) > Creation Time : Wed Jun 20 23:56:59 2012 > Raid Level : raid10 > Raid Devices : 4 > > Avail Dev Size : 5860268943 (2794.39 GiB 3000.46 GB) > Array Size : 5860268032 (5588.79 GiB 6000.91 GB) > Used Dev Size : 5860268032 (2794.39 GiB 3000.46 GB) > Data Offset : 262144 sectors > Super Offset : 8 sectors > Unused Space : before=262064 sectors, after=911 sectors > State : clean > Device UUID : 82667e81:a6158319:85e0282e:845eec1c > > Update Time : Wed May 8 11:00:29 2019 > Checksum : 10ac3349 - correct > Events : 1193 > > Layout : near=2 > Chunk Size : 512K > > Device Role : spare > Array State : AAAA ('A' == active, '.' == missing, 'R' == replacing) > root@nas2:~# > > mdadm --detail /dev/md0 > /dev/md0: > Version : 1.2 > Raid Level : raid0 > Total Devices : 4 > Persistence : Superblock is persistent > > State : inactive > Working Devices : 4 > > Name : nas2:0 (local to host nas2) > UUID : 6abe1f20:90c629de:fadd8dc0:ca14c928 > Events : 1193 > > Number Major Minor RaidDevice > > - 8 65 - /dev/sde1 > - 8 49 - /dev/sdd1 > - 8 33 - /dev/sdc1 > - 8 17 - /dev/sdb1 > > cat /proc/mdstat > Personalities : [raid10] > md0 : inactive sdc1[1](S) sdb1[0](S) sde1[4](S) sdd1[3](S) > 11720537886 blocks super 1.2 > > unused devices: <none> > > -- > __ > / ` Eric Valette > /-- __ o _. 6 rue Paul Le Flem > (___, / (_(_(__ 35740 Pace > > Tel: +33 (0)2 99 85 26 76 Fax: +33 (0)2 99 85 26 76 > E-mail: eric.valette@xxxxxxx > > >