Re: RAID 10 problems, two disks marked as spare

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi all,

I decided to take the plunge and recreate the array with "mdadm --create --assume-clean --level=10 --raid-devices=4 --size=974722176 /dev/md1 /dev/sda2 /dev/sdb2 /dev/sdc2 /dev/sdd2”. The array is working correctly again, however for now I mounted it read only and I’m going to take some backups to another raid system just in case something terribly goes wrong (I have to reboot the device so that the Iomega NAS system can detect the drives again).

Greetings and thanks,
Jim



On 28 Feb 2014, at 20:40, jim@xxxxxxx wrote:

> Hi all,
> 
> I'm having a bit of trouble with my NAS and I was hoping some of you lads would be able to help my out.
> 
> First of all my setup:
> 
> The NAS itself is an Iomega StorCenter ix4-200d. It has 4x 1TiB drives, configured to use RAID10. I have already replaced one drive once. The NAS itself doesn't come with shell access but it's fairly easy to 'root' it anyway (which I did).
> 
> Several days ago the NAS send me an email that a certain drive was degraded:
> 
>> The Iomega StorCenter device is degraded and data protection is at risk. A drive may have either failed or been removed from your Iomega StorCenter device. Visit the Dashboard on the management interface for details. To prevent possible data loss, this issue should be repaired as soon as possible.
> 
> I decided to first to try and reboot the device, maybe it was a simple error. After reboot I received the following:
> 
>> Data protection is being reconstructed on your Iomega StorCenter device
> 
> So I was happy until several hours later I received the following messages (all at the same time)
> 
>> The Iomega StorCenter device has completed data protection reconstruction.
> 
>> The Iomega StorCenter device has failed and some data loss may have occurred. Multiple drives may have either failed or been removed from your storage system. Visit the Dashboard on the management interface for details.
> 
>> Drive number 4 encountered a recoverable error.
> 
> No data was accessible anymore. After that I opened a shell to the device and try to trouble shout it. But I didn’t manage to get it working. The only solution I currently see it to try and rebuild the RAID array but as I have hardly any experience with the mdadm tool I decided to ask the opinions of the people here.
> 
> Here is some information regarding the setup:
> 
> root@BauwensNAS:/# mdadm -D /dev/md1
> /dev/md1:
>        Version : 01.00
>  Creation Time : Mon Jan 24 20:57:43 2011
>     Raid Level : raid10
>  Used Dev Size : 974722176 (929.57 GiB 998.12 GB)
>   Raid Devices : 4
>  Total Devices : 4
> Preferred Minor : 1
>    Persistence : Superblock is persistent
> 
>    Update Time : Wed Feb 26 02:44:57 2014
>          State : active, degraded, Not Started
> Active Devices : 2
> Working Devices : 4
> Failed Devices : 0
>  Spare Devices : 2
> 
>         Layout : near=2, far=1
>     Chunk Size : 64K
> 
>           Name : bwns:1
>           UUID : 43a1e240:956c3131:df9f6e66:bd9a071e
>         Events : 133470
> 
>    Number   Major   Minor   RaidDevice State
>       0       8        2        0      active sync   /dev/sda2
>       4       8       18        1      active sync   /dev/sdb2
>       2       0        0        2      removed
>       3       0        0        3      removed
> 
>       2       8       34        -      spare   /dev/sdc2
>       3       8       50        -      spare   /dev/sdd2
> 
> As you can see, the two last drives are marked as spare. My multiple attempts to get the drive running with all the disks have only been failures (but I assume that's also due to me not having experience with the tools).
> 
> Also, the disks themself appear to be fine (also because the md0 device that hosts /boot works properly)
> 
> Some more info:
> 
> root@BauwensNAS:/# mdadm --examine /dev/sd[abcd]2
> /dev/sda2:
>          Magic : a92b4efc
>        Version : 1.0
>    Feature Map : 0x0
>     Array UUID : 43a1e240:956c3131:df9f6e66:bd9a071e
>           Name : bwns:1
>  Creation Time : Mon Jan 24 20:57:43 2011
>     Raid Level : raid10
>   Raid Devices : 4
> 
> Avail Dev Size : 1949444384 (929.57 GiB 998.12 GB)
>     Array Size : 3898888704 (1859.14 GiB 1996.23 GB)
>  Used Dev Size : 1949444352 (929.57 GiB 998.12 GB)
>   Super Offset : 1949444640 sectors
>          State : clean
>    Device UUID : b05f4c40:819ddbef:76872d9f:abacf3c9
> 
>    Update Time : Wed Feb 26 02:44:57 2014
>       Checksum : a94e1ae6 - correct
>         Events : 133470
> 
>         Layout : near=2, far=1
>     Chunk Size : 64K
> 
>    Array Slot : 0 (0, failed, empty, empty, 1)
>   Array State : Uu__ 1 failed
> /dev/sdb2:
>          Magic : a92b4efc
>        Version : 1.0
>    Feature Map : 0x0
>     Array UUID : 43a1e240:956c3131:df9f6e66:bd9a071e
>           Name : bwns:1
>  Creation Time : Mon Jan 24 20:57:43 2011
>     Raid Level : raid10
>   Raid Devices : 4
> 
> Avail Dev Size : 1949444384 (929.57 GiB 998.12 GB)
>     Array Size : 3898888704 (1859.14 GiB 1996.23 GB)
>  Used Dev Size : 1949444352 (929.57 GiB 998.12 GB)
>   Super Offset : 1949444640 sectors
>          State : clean
>    Device UUID : 6c34331a:0fda7f73:a1f76d41:a826ac1f
> 
>    Update Time : Wed Feb 26 02:44:57 2014
>       Checksum : fed2165a - correct
>         Events : 133470
> 
>         Layout : near=2, far=1
>     Chunk Size : 64K
> 
>    Array Slot : 4 (0, failed, empty, empty, 1)
>   Array State : uU__ 1 failed
> /dev/sdc2:
>          Magic : a92b4efc
>        Version : 1.0
>    Feature Map : 0x0
>     Array UUID : 43a1e240:956c3131:df9f6e66:bd9a071e
>           Name : bwns:1
>  Creation Time : Mon Jan 24 20:57:43 2011
>     Raid Level : raid10
>   Raid Devices : 4
> 
> Avail Dev Size : 1949444384 (929.57 GiB 998.12 GB)
>     Array Size : 3898888704 (1859.14 GiB 1996.23 GB)
>  Used Dev Size : 1949444352 (929.57 GiB 998.12 GB)
>   Super Offset : 1949444640 sectors
>          State : clean
>    Device UUID : 773dbfec:07467e62:de7be59b:5c680df5
> 
>    Update Time : Wed Feb 26 02:44:57 2014
>       Checksum : f035517e - correct
>         Events : 133470
> 
>         Layout : near=2, far=1
>     Chunk Size : 64K
> 
>    Array Slot : 2 (0, failed, empty, empty, 1)
>   Array State : uu__ 1 failed
> /dev/sdd2:
>          Magic : a92b4efc
>        Version : 1.0
>    Feature Map : 0x0
>     Array UUID : 43a1e240:956c3131:df9f6e66:bd9a071e
>           Name : bwns:1
>  Creation Time : Mon Jan 24 20:57:43 2011
>     Raid Level : raid10
>   Raid Devices : 4
> 
> Avail Dev Size : 1949444384 (929.57 GiB 998.12 GB)
>     Array Size : 3898888704 (1859.14 GiB 1996.23 GB)
>  Used Dev Size : 1949444352 (929.57 GiB 998.12 GB)
>   Super Offset : 1949444640 sectors
>          State : clean
>    Device UUID : dbd27546:6b623b53:8f887960:b7cbf424
> 
>    Update Time : Wed Feb 26 02:44:57 2014
>       Checksum : 2f247322 - correct
>         Events : 133470
> 
>         Layout : near=2, far=1
>     Chunk Size : 64K
> 
>    Array Slot : 3 (0, failed, empty, empty, 1)
>   Array State : uu__ 1 failed
> 
> 
> Taking a look at the event count they seem to be synchronized, so I'm not really sure what's going on here.
> 
> root@BauwensNAS:/# cat /proc/mdstat
> Personalities : [linear] [raid0] [raid1] [raid10] [raid6] [raid5] [raid4]
> md1 : inactive sda2[0] sdd2[3](S) sdc2[2](S) sdb2[4]
>      3898888704 blocks super 1.0
> 
> md0 : active raid1 sda1[0] sdd1[3] sdc1[2] sdb1[1]
>      2040128 blocks [4/4] [UUUU]
> 
> unused devices: <none>
> 
> Anyone have an idea how I could resolve this problem (hoping that I don't have any data loss...)? Any help is greatly appreciated. I sure regret rebooting the device without taking some extra backups.
> 
> TIA!
> Jim
> --
> To unsubscribe from this list: send the line "unsubscribe linux-raid" in
> the body of a message to majordomo@xxxxxxxxxxxxxxx
> More majordomo info at  http://vger.kernel.org/majordomo-info.html

--
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html




[Index of Archives]     [Linux RAID Wiki]     [ATA RAID]     [Linux SCSI Target Infrastructure]     [Linux Block]     [Linux IDE]     [Linux SCSI]     [Linux Hams]     [Device Mapper]     [Device Mapper Cryptographics]     [Kernel]     [Linux Admin]     [Linux Net]     [GFS]     [RPM]     [git]     [Yosemite Forum]


  Powered by Linux