On Mon, 31 Mar 2014 10:08:30 -0400 bobzer <bobzer@xxxxxxxxx> wrote: > Hi, > > My raid 5 is in a strange state. the mdstat tell me is degraded, but > when i check the disk that should be in the raid anymore it tell me > everything is right ... I'm lost > > #cat /proc/mdstat > Personalities : [raid6] [raid5] [raid4] > md0 : active raid5 sdc1[3] sdd1[1] > 3907021568 blocks super 1.2 level 5, 128k chunk, algorithm 2 [3/2] [UU_] > > unused devices: <none> > > > so my third disk is unused i check with : > > #mdadm -D /dev/md0 > /dev/md0: > Version : 1.2 > Creation Time : Sun Mar 4 22:49:14 2012 > Raid Level : raid5 > Array Size : 3907021568 (3726.03 GiB 4000.79 GB) > Used Dev Size : 1953510784 (1863.01 GiB 2000.40 GB) > Raid Devices : 3 > Total Devices : 2 > Persistence : Superblock is persistent > > Update Time : Sun Mar 30 23:01:35 2014 > State : clean, degraded > Active Devices : 2 > Working Devices : 2 > Failed Devices : 0 > Spare Devices : 0 > > Layout : left-symmetric > Chunk Size : 128K > Name : debian:0 > UUID : bf3c605b:9699aa55:d45119a2:7ba58d56 > Events : 255801 > > Number Major Minor RaidDevice State > 3 8 33 0 active sync /dev/sdc1 > 1 8 49 1 active sync /dev/sdd1 > 2 0 0 2 removed > > > after verify what says the disk i got confuse : > > > #mdadm --examine /dev/sdb1 > /dev/sdb1: > Magic : a92b4efc > Version : 1.2 > Feature Map : 0x0 > Array UUID : bf3c605b:9699aa55:d45119a2:7ba58d56 > Name : debian:0 > Creation Time : Sun Mar 4 22:49:14 2012 > Raid Level : raid5 > Raid Devices : 3 > > Avail Dev Size : 3907021954 (1863.01 GiB 2000.40 GB) > Array Size : 7814043136 (3726.03 GiB 4000.79 GB) > Used Dev Size : 3907021568 (1863.01 GiB 2000.40 GB) > Data Offset : 2048 sectors > Super Offset : 8 sectors > State : clean > Device UUID : f9059dfb:74af1ab7:bc1465b1:e2ff30ba > > Update Time : Sun Jan 5 04:11:41 2014 > Bad Block Log : 512 entries available at offset 2032 sectors > Checksum : 7df1fefc - correct > Events : 436 > Layout : left-symmetric > Chunk Size : 128K > > Device Role : Active device 2 > Array State : AAA ('A' == active, '.' == missing) sdb1 thinks it is OK, but that is normal. When a device fails the fact that it is failed isn't recorded on that device, only on the other devices. > > > so there i tried to re-add the disk that doesn't works > so i tred to stop the raid and assemble it again but it doesn't work either > > #mdadm --stop /dev/md0 > #mdadm --assemble --force /dev/md0 /dev/sd[bcd]1 > mdadm: /dev/md0 has been started with 2 drives (out of 3). > > > can you help me guys ? If mdadm /dev/md0 --add /dev/sdb1 doesn't work, them run mdadm --zero-super /dev/sdb1 first. > > > Thanks by advance > > ps: i've got mdadm 3.3-devel i would update it but don't how to do ... git clone git://neil.brown.name/mdadm cd mdadm make make install > -- > To unsubscribe from this list: send the line "unsubscribe linux-raid" in > the body of a message to majordomo@xxxxxxxxxxxxxxx > More majordomo info at http://vger.kernel.org/majordomo-info.html NeilBrown
Attachment:
signature.asc
Description: PGP signature