Re: strange status raid 5

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Thank you very much, it's recovering right now :-)

I'll update mdadm after

On Tue, Apr 1, 2014 at 6:41 PM, NeilBrown <neilb@xxxxxxx> wrote:
> On Mon, 31 Mar 2014 10:08:30 -0400 bobzer <bobzer@xxxxxxxxx> wrote:
>
>> Hi,
>>
>> My raid 5 is in a strange state. the mdstat tell me is degraded, but
>> when i check the disk that should be in the raid anymore it tell me
>> everything is right ... I'm lost
>>
>> #cat /proc/mdstat
>> Personalities : [raid6] [raid5] [raid4]
>> md0 : active raid5 sdc1[3] sdd1[1]
>>      3907021568 blocks super 1.2 level 5, 128k chunk, algorithm 2 [3/2] [UU_]
>>
>> unused devices: <none>
>>
>>
>> so my third disk is unused i check with :
>>
>> #mdadm -D /dev/md0
>> /dev/md0:
>>        Version : 1.2
>>  Creation Time : Sun Mar  4 22:49:14 2012
>>     Raid Level : raid5
>>     Array Size : 3907021568 (3726.03 GiB 4000.79 GB)
>>  Used Dev Size : 1953510784 (1863.01 GiB 2000.40 GB)
>>   Raid Devices : 3
>>  Total Devices : 2
>>    Persistence : Superblock is persistent
>>
>>    Update Time : Sun Mar 30 23:01:35 2014
>>          State : clean, degraded
>> Active Devices : 2
>> Working Devices : 2
>> Failed Devices : 0
>>  Spare Devices : 0
>>
>>         Layout : left-symmetric
>>     Chunk Size : 128K
>>           Name : debian:0
>>           UUID : bf3c605b:9699aa55:d45119a2:7ba58d56
>>         Events : 255801
>>
>>    Number   Major   Minor   RaidDevice State
>>       3       8       33        0      active sync   /dev/sdc1
>>       1       8       49        1      active sync   /dev/sdd1
>>       2       0        0        2      removed
>>
>>
>> after verify what says the disk i got confuse :
>>
>>
>> #mdadm --examine /dev/sdb1
>> /dev/sdb1:
>>          Magic : a92b4efc
>>        Version : 1.2
>>    Feature Map : 0x0
>>     Array UUID : bf3c605b:9699aa55:d45119a2:7ba58d56
>>           Name : debian:0
>>  Creation Time : Sun Mar  4 22:49:14 2012
>>     Raid Level : raid5
>>   Raid Devices : 3
>>
>> Avail Dev Size : 3907021954 (1863.01 GiB 2000.40 GB)
>>     Array Size : 7814043136 (3726.03 GiB 4000.79 GB)
>>  Used Dev Size : 3907021568 (1863.01 GiB 2000.40 GB)
>>    Data Offset : 2048 sectors
>>   Super Offset : 8 sectors
>>          State : clean
>>    Device UUID : f9059dfb:74af1ab7:bc1465b1:e2ff30ba
>>
>>    Update Time : Sun Jan  5 04:11:41 2014
>>  Bad Block Log : 512 entries available at offset 2032 sectors
>>       Checksum : 7df1fefc - correct
>>         Events : 436
>>         Layout : left-symmetric
>>     Chunk Size : 128K
>>
>>   Device Role : Active device 2
>>   Array State : AAA ('A' == active, '.' == missing)
>
> sdb1 thinks it is OK, but that is normal.  When a device fails the fact that
> it is failed isn't recorded on that device, only on the other devices.
>
>>
>>
>> so there i tried to re-add the disk that doesn't works
>> so i tred to stop the raid and assemble it again but it doesn't work either
>>
>> #mdadm --stop /dev/md0
>> #mdadm --assemble --force /dev/md0 /dev/sd[bcd]1
>> mdadm: /dev/md0 has been started with 2 drives (out of 3).
>>
>>
>> can you help me guys ?
>
> If
>    mdadm /dev/md0 --add /dev/sdb1
> doesn't work, them run
>    mdadm --zero-super /dev/sdb1
> first.
>
>>
>>
>> Thanks by advance
>>
>> ps: i've got mdadm 3.3-devel i would update it but don't how to do ...
>
>  git clone git://neil.brown.name/mdadm
>  cd mdadm
>  make
>  make install
>
>
>> --
>> To unsubscribe from this list: send the line "unsubscribe linux-raid" in
>> the body of a message to majordomo@xxxxxxxxxxxxxxx
>> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>
> NeilBrown
--
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html




[Index of Archives]     [Linux RAID Wiki]     [ATA RAID]     [Linux SCSI Target Infrastructure]     [Linux Block]     [Linux IDE]     [Linux SCSI]     [Linux Hams]     [Device Mapper]     [Device Mapper Cryptographics]     [Kernel]     [Linux Admin]     [Linux Net]     [GFS]     [RPM]     [git]     [Yosemite Forum]


  Powered by Linux