On Mon, 25 Jun 2012 13:44:35 +0200 "Matthias Herrmanny" <Matthias.Herrmanny@xxxxxx> wrote: > Good day, sorry for my English schelchtes. > > I have a little problem with my two raid 5 system. > I have been in use Ubuntu 11.xx and then I lost 2 hard drives with a hardware defect, whereupon I have my new hard drives 2TB concerned. Since Ubuntu can not handle drives larger than 2TB, I've also changed Centos 6.2. > One of my two Raid 5 5x (250Gb) system, I can write off. Since there are two disc finally broken (hardware failure) > The other raid 5 system (3x 80G) I have then taken. Now on a disc that has no partition disappears, the second one was a spare Displayed and the third has no superblock more. > [root@sonne ~]# fdisk -l > Disk /dev/sdg: 82.0 GB, 81964302336 bytes > 255 heads, 63 sectors/track, 9964 cylinders > Units = cylinders of 16065 * 512 = 8225280 bytes > Sector size (logical/physical): 512 bytes / 512 bytes > I/O size (minimum/optimal): 512 bytes / 512 bytes > Disk identifier: 0x00032373 > > Device Boot Start End Blocks Id System > > Disk /dev/sde: 120.1 GB, 120060444672 bytes > 255 heads, 63 sectors/track, 14596 cylinders > Units = cylinders of 16065 * 512 = 8225280 bytes > Sector size (logical/physical): 512 bytes / 512 bytes > I/O size (minimum/optimal): 512 bytes / 512 bytes > Disk identifier: 0x00037e0a > > Device Boot Start End Blocks Id System > /dev/sde1 1 9970 80076800 fd Linux raid autodetect > /dev/sde2 9970 14597 37167105 5 Extended > /dev/sde5 9970 14597 37167104 fd Linux raid autodetect > > Disk /dev/sdf: 82.0 GB, 81964302336 bytes > 255 heads, 63 sectors/track, 9964 cylinders > Units = cylinders of 16065 * 512 = 8225280 bytes > Sector size (logical/physical): 512 bytes / 512 bytes > I/O size (minimum/optimal): 512 bytes / 512 bytes > Disk identifier: 0x00050d72 > > Device Boot Start End Blocks Id System > /dev/sdf1 1 9965 80041984 fd Linux raid autodetect > > > [root@sonne ~]# mdadm --examine /dev/sdf1 > /dev/sdf1: > Magic : a92b4efc > Version : 1.2 > Feature Map : 0x0 > Array UUID : 5874972e:5326304f:37228c78:dd15d965 > Name : sonne:5 > Creation Time : Sun Nov 27 03:44:57 2011 > Raid Level : -unknown- > Raid Devices : 0 > > Avail Dev Size : 160081920 (76.33 GiB 81.96 GB) > Data Offset : 2048 sectors > Super Offset : 8 sectors > State : active > Device UUID : 8bc40fb6:73fe82b4:c1c001f6:046775d4 > > Update Time : Mon Jun 18 01:35:12 2012 > Checksum : 59fcdac3 - correct > Events : 2 > > > Device Role : spare > Array State : ('A' == active, '.' == missing) > > > [root@sonne ~]# mdadm --examine /dev/sde1 > /dev/sde1: > Magic : a92b4efc > Version : 1.2 > Feature Map : 0x0 > Array UUID : 5874972e:5326304f:37228c78:dd15d965 > Name : sonne:5 > Creation Time : Sun Nov 27 03:44:57 2011 > Raid Level : -unknown- > Raid Devices : 0 > > Avail Dev Size : 160151552 (76.37 GiB 82.00 GB) > Used Dev Size : 160081920 (76.33 GiB 81.96 GB) > Data Offset : 2048 sectors > Super Offset : 8 sectors > State : active > Device UUID : eab5a0b6:f7034d85:b143bc4c:4f3b5054 > > Update Time : Mon Jun 18 01:35:12 2012 > Checksum : 1ee38df - correct > Events : 2 > > > Device Role : spare > Array State : ('A' == active, '.' == missing) > > > Now I've put some try to run the RAID system again. Unfortunately, all to no avail. > mdadm -A /dev/md125 /dev/sde1 /dev/sdf1 /dev/sdg1 > mdadm: cannot open device /dev/sde1: Device or resource busy > mdadm: /dev/sde1 has no superblock - assembly aborted > > mdadm -Av /md/md5 --uuid5874972e:5326304f:37228c78:dd15d965 /dev/* > > mdadm --manage /dev/md125 --re-add /dev/sdg1 > mdadm: cannot get array info for /dev/md125 > mdadm --assemble /dev/md125 --auto=yes --scan --update=summaries --verbose > mdadm: /dev/md125 not identified in config file. > mdadm -Cv /dev/md125 -e 1.20 --assume-clean -n3 -l5 /dev/sde1 /dev/sdf1 /dev/sdg1 > mdadm: unrecognised metadata identifier: 1.20 > mdadm -Cv /dev/md125 -e 1.20 --assume-clean -n3 -l5 /dev/sde1 /dev/sdf1 /dev/sdg1 --force > mdadm: unrecognised metadata identifier: 1.20 > > > I hope you still have an idea how I got the raid again and running quickly. > Thanks in advance. Hi. You have suffered from this bug: http://neil.brown.name/blog/20120615073245 http://www.heise.de/newsticker/meldung/Fehler-im-Linux-Kernel-kann-Software-RAIDs-zerstoeren-1620896.html is a brief article in German referring to it. You need to mdadm --stop /dev/md.... any array that is listed as "inactive" in /proc/mdstat, then try something like: mdadm -C /dev/md125 -e 1.2 --assume-clean -n 3 -l 5 /dev/sde1 /dev/sdf1 missing Note that I didn't include sdg as something looks strange with sdg - it doesn't have any partitions. Maybe you included "/dev/sdg" in the array rather than "/dev/sdg1" ?? What does "mdadm -E /dev/sdg" show? If that looks like an array member, then you might try mdadm -C /dev/md125 -e 1.2 --assume-clean -n 3 -l 5 \ /dev/sde1 /dev/sdf1 /dev/sdg Then try "fsck -n" or similar. If that fails, stop the array and try again with a different ordering of devices. Feel free to ask if something isn't clear or if you want to confirm your next step before you take it. NeilBrown
Attachment:
signature.asc
Description: PGP signature