thanks for your help. i took time to answer because unlucky me, the power supply of my laptop fried so no laptop and no way to work on my raid :-( any way i got a new one :-) for the dmesg i paste it here : http://pastebin.com/whUHs256 root@serveur:~# uname -a Linux serveur 3.2.0-4-amd64 #1 SMP Debian 3.2.63-2+deb7u1 x86_64 GNU/Linux root@serveur:~# mdadm -V mdadm - v3.3-78-gf43f5b3 - 02nd avril 2014 about zero the superblock on the wrong device, i hope i didn't do that, but also i really don't think i did that, because i took care and at that time the raid was working i don't know what to do, if i use dd_rescue and can't get back 100% of the data could i be able to start the raid anyway ? what are my risk if i try something like : mdadm --assemble --force /dev/md0 /dev/sd[bcde]1 thank you very much for your time Mathieu On Wed, May 11, 2016 at 3:15 PM, Robin Hill <robin@xxxxxxxxxxxxxxx> wrote: > On Tue May 10, 2016 at 11:28:31PM +0200, bobzer wrote: > >> hi everyone, >> >> I'm in panic mode :-( because i got a raid 5 with 4 disk but 2 removed >> yesterday i got a power outage which removed one disk. the disks >> sd[bcd]1 was ok and saying that sde1 is removed but sde1 said that >> everything is fine. >> so i stop the raid, zero the superblock of sde1, start the raid and >> add sde1 to the raid. then it start to reconstruct, i think it had >> time to finish before this problem (i'm not 100% sure that it finish >> but i think so) >> the data was accessible so i went to sleep >> today i discovered the raid in this state : >> >> root@serveur:/home/math# mdadm -D /dev/md0 >> /dev/md0: >> Version : 1.2 >> Creation Time : Sun Mar 4 22:49:14 2012 >> Raid Level : raid5 >> Array Size : 5860532352 (5589.04 GiB 6001.19 GB) >> Used Dev Size : 1953510784 (1863.01 GiB 2000.40 GB) >> Raid Devices : 4 >> Total Devices : 4 >> Persistence : Superblock is persistent >> >> Update Time : Fri May 6 17:44:02 2016 >> State : clean, FAILED >> Active Devices : 2 >> Working Devices : 3 >> Failed Devices : 1 >> Spare Devices : 1 >> >> Layout : left-symmetric >> Chunk Size : 128K >> >> Name : debian:0 >> UUID : bf3c605b:9699aa55:d45119a2:7ba58d56 >> Events : 892482 >> >> Number Major Minor RaidDevice State >> 3 8 33 0 active sync /dev/sdc1 >> 1 8 49 1 active sync /dev/sdd1 >> 4 0 0 4 removed >> 6 0 0 6 removed >> >> 4 8 17 - faulty /dev/sdb1 >> 5 8 65 - spare /dev/sde1 >> > So this reports /dev/sdb1 faulty and /dev/sde1 spare. That would > indicate that the rebuild hadn't finished. > >> root@serveur:/home/math# mdadm --examine /dev/sdb1 >> /dev/sdb1: >> Magic : a92b4efc >> Version : 1.2 >> Feature Map : 0x0 >> Array UUID : bf3c605b:9699aa55:d45119a2:7ba58d56 >> Name : debian:0 >> Creation Time : Sun Mar 4 22:49:14 2012 >> Raid Level : raid5 >> Raid Devices : 4 >> >> Avail Dev Size : 3907021954 (1863.01 GiB 2000.40 GB) >> Array Size : 5860532352 (5589.04 GiB 6001.19 GB) >> Used Dev Size : 3907021568 (1863.01 GiB 2000.40 GB) >> Data Offset : 2048 sectors >> Super Offset : 8 sectors >> Unused Space : before=1960 sectors, after=386 sectors >> State : clean >> Device UUID : 9bececcb:d520ca38:fd88d956:5718e361 >> >> Update Time : Fri May 6 02:07:00 2016 >> Bad Block Log : 512 entries available at offset 72 sectors >> Checksum : dc2a133a - correct >> Events : 892215 >> >> Layout : left-symmetric >> Chunk Size : 128K >> >> Device Role : Active device 2 >> Array State : AAAA ('A' == active, '.' == missing, 'R' == replacing) >> > We can see /dev/sdb1 has a lower event count than the others and also > that it indicates all the drives in the array were active when it was > last running. That would strongly suggest that it was not in the array > when /dev/sde1 was added to rebuild. The update time is also nearly 16 > hours earlier than that of the other drives. > >> root@serveur:/home/math# mdadm --examine /dev/sdc1 >> /dev/sdc1: >> Magic : a92b4efc >> Version : 1.2 >> Feature Map : 0x0 >> Array UUID : bf3c605b:9699aa55:d45119a2:7ba58d56 >> Name : debian:0 >> Creation Time : Sun Mar 4 22:49:14 2012 >> Raid Level : raid5 >> Raid Devices : 4 >> >> Avail Dev Size : 3907021954 (1863.01 GiB 2000.40 GB) >> Array Size : 5860532352 (5589.04 GiB 6001.19 GB) >> Used Dev Size : 3907021568 (1863.01 GiB 2000.40 GB) >> Data Offset : 2048 sectors >> Super Offset : 8 sectors >> Unused Space : before=1960 sectors, after=386 sectors >> State : clean >> Device UUID : 1ecaf51c:3289a902:7bb71a93:237c68e8 >> >> Update Time : Fri May 6 17:58:27 2016 >> Bad Block Log : 512 entries available at offset 72 sectors >> Checksum : b9d6aa84 - correct >> Events : 892484 >> >> Layout : left-symmetric >> Chunk Size : 128K >> >> Device Role : Active device 0 >> Array State : AA.. ('A' == active, '.' == missing, 'R' == replacing) >> >> root@serveur:/home/math# mdadm --examine /dev/sdd1 >> /dev/sdd1: >> Magic : a92b4efc >> Version : 1.2 >> Feature Map : 0x0 >> Array UUID : bf3c605b:9699aa55:d45119a2:7ba58d56 >> Name : debian:0 >> Creation Time : Sun Mar 4 22:49:14 2012 >> Raid Level : raid5 >> Raid Devices : 4 >> >> Avail Dev Size : 3907021954 (1863.01 GiB 2000.40 GB) >> Array Size : 5860532352 (5589.04 GiB 6001.19 GB) >> Used Dev Size : 3907021568 (1863.01 GiB 2000.40 GB) >> Data Offset : 2048 sectors >> Super Offset : 8 sectors >> Unused Space : before=0 sectors, after=386 sectors >> State : clean >> Device UUID : 406c4cb5:c188e4a9:7ed8be9f:14a49b16 >> >> Update Time : Fri May 6 17:58:27 2016 >> Bad Block Log : 512 entries available at offset 2032 sectors >> Checksum : 343f9cd0 - correct >> Events : 892484 >> >> Layout : left-symmetric >> Chunk Size : 128K >> >> Device Role : Active device 1 >> Array State : AA.. ('A' == active, '.' == missing, 'R' == replacing) >> > These two drives contain the same information. They indicate that they > were the only 2 running members in the array when they were last updated. > >> root@serveur:/home/math# mdadm --examine /dev/sde1 >> /dev/sde1: >> Magic : a92b4efc >> Version : 1.2 >> Feature Map : 0x8 >> Array UUID : bf3c605b:9699aa55:d45119a2:7ba58d56 >> Name : debian:0 >> Creation Time : Sun Mar 4 22:49:14 2012 >> Raid Level : raid5 >> Raid Devices : 4 >> >> Avail Dev Size : 3907025072 (1863.01 GiB 2000.40 GB) >> Array Size : 5860532352 (5589.04 GiB 6001.19 GB) >> Used Dev Size : 3907021568 (1863.01 GiB 2000.40 GB) >> Data Offset : 2048 sectors >> Super Offset : 8 sectors >> Unused Space : before=1960 sectors, after=3504 sectors >> State : clean >> Device UUID : f2e9c1ec:2852cf21:1a588581:b9f49a8b >> >> Update Time : Fri May 6 17:58:27 2016 >> Bad Block Log : 512 entries available at offset 72 sectors - bad >> blocks present. >> Checksum : 3a65b8bc - correct >> Events : 892484 >> >> Layout : left-symmetric >> Chunk Size : 128K >> >> Device Role : spare >> Array State : AA.. ('A' == active, '.' == missing, 'R' == replacing) >> > And finally /dev/sde1 shows as a spare, with the rest of the data > matching /dev/sdc1 and /dev/sde1. > >> PLEASE help me :-) i don't know what to do so i did nothing to not do >> any stupid things >> 1000 thank you >> >> ps i just saw this, i hope it not mak y case worst >> root@serveur:/home/math# cat /etc/mdadm/mdadm.conf >> DEVICE /dev/sd[bcd]1 >> ARRAY /dev/md0 metadata=1.2 name=debian:0 >> UUID=bf3c605b:9699aa55:d45119a2:7ba58d56 >> > > From the data here, if looks to me as though /dev/sdb1 failed originally > (hence it thinks the array was complete). Either then /dev/sde1 also > failed, or you've proceeded to zero the superblock on the wrong drive. > You really need to look through the system logs and verify what happened > when and to what disk (if you rebooted at any point, the drive ordering > may have changed, so don't take for granted that the drive names are > consistent throughout). > > Cheers, > Robin > -- > ___ > ( ' } | Robin Hill <robin@xxxxxxxxxxxxxxx> | > / / ) | Little Jim says .... | > // !! | "He fallen in de water !!" | -- To unsubscribe from this list: send the line "unsubscribe linux-raid" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html