Re: Help! I killed my mdadm raid 5

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Thu, 18 Nov 2010 14:51:09 -0500
<p3-500@xxxxxxxxxxxxxx> wrote:

> Arrg, Major goof. I managed to kill my raid 5 through my own stupidity. Here is the chain of events, Your help is greatly appreciated.
> 
> I added disk #6 to my working 5 disk array, after about 24 hours it was done but i did not see the extra space (I now know all I needed to do was fsck and resize).
> 
> I failed and removed the new disk from the array then somehow also failed and removed drive # 5 from the array. At this point the array was still running so I added drive #5 & 6 back to the array but they got added as spares instead of active components. next I rebooted and could not assemble the array.
> 
> I tried --assemble and --assemble --force which results with "mdadm: /dev/md0 assembled from 4 drives and 1 spare - not enough to start the array". I am naming all 6 drives on the command line.
> 
> Several posts have suggested I --create the array again but I am hesitant to do this as I do not want to lose my data. The wiki suggests mkraid --force but I believe this is deprecated? It also requires an up to date /etc/raidtab which I do not have. Before I mess it up any more here's where I am.

Yes, forget mkraid.

mdadm --create is fairly safe as long as you think about what you are doing.
In particular, make sure the same metadata type is used and importantly use
--assume-clean.


so something like

 mdadm --create /dev/md0 -e 0.90 --assume-clean --level=5 --n=6 \
 --chunk=64 /dev/sda2 /dev/sdb2 /dev/sdc2 /dev/sdd2 /dev/sdd /dev/sdf

may well work.  Use 'fsck -n' and the 'mount -o ro' to check.
All it will do is update the metadat.  If you have the devices in the wrong
order, then simple
   mdadm -Ss
and try again.  The data wont be changed until you trigger a resync, or mount
the filesystem read-write.

And next time, think before you fail a drive in an array !!!

NeilBrown


> 
> mdadam --examine /dev/sda2,b2,c2,d2,e,f
> server2@server2:~$ sudo mdadm --examine /dev/sda2
> [sudo] password for server2: 
> /dev/sda2:
>           Magic : a92b4efc
>         Version : 00.90.00
>            UUID : b7390ef0:9c9ecbe0:abcb3798:f53e3ccc
>   Creation Time : Sat Feb  6 17:37:20 2010
>      Raid Level : raid5
>   Used Dev Size : 1952537920 (1862.09 GiB 1999.40 GB)
>      Array Size : 9762689600 (9310.43 GiB 9996.99 GB)
>    Raid Devices : 6
>   Total Devices : 4
> Preferred Minor : 0
> 
>     Update Time : Tue Nov 16 14:29:15 2010
>           State : clean
>  Active Devices : 4
> Working Devices : 4
>  Failed Devices : 2
>   Spare Devices : 0
>        Checksum : aacb71b5 - correct
>          Events : 382240
> 
>          Layout : left-symmetric
>      Chunk Size : 64K
> 
>       Number   Major   Minor   RaidDevice State
> this     0       8        2        0      active sync   /dev/sda2
> 
>    0     0       8        2        0      active sync   /dev/sda2
>    1     1       8       18        1      active sync   /dev/sdb2
>    2     2       8       34        2      active sync   /dev/sdc2
>    3     3       8       50        3      active sync   /dev/sdd2
>    4     4       0        0        4      faulty removed
>    5     5       0        0        5      faulty removed
> server2@server2:~$ sudo mdadm --examine /dev/sdb2
> /dev/sdb2:
>           Magic : a92b4efc
>         Version : 00.90.00
>            UUID : b7390ef0:9c9ecbe0:abcb3798:f53e3ccc
>   Creation Time : Sat Feb  6 17:37:20 2010
>      Raid Level : raid5
>   Used Dev Size : 1952537920 (1862.09 GiB 1999.40 GB)
>      Array Size : 9762689600 (9310.43 GiB 9996.99 GB)
>    Raid Devices : 6
>   Total Devices : 4
> Preferred Minor : 0
> 
>     Update Time : Tue Nov 16 14:29:15 2010
>           State : clean
>  Active Devices : 4
> Working Devices : 4
>  Failed Devices : 2
>   Spare Devices : 0
>        Checksum : aacb71c7 - correct
>          Events : 382240
> 
>          Layout : left-symmetric
>      Chunk Size : 64K
> 
>       Number   Major   Minor   RaidDevice State
> this     1       8       18        1      active sync   /dev/sdb2
> 
>    0     0       8        2        0      active sync   /dev/sda2
>    1     1       8       18        1      active sync   /dev/sdb2
>    2     2       8       34        2      active sync   /dev/sdc2
>    3     3       8       50        3      active sync   /dev/sdd2
>    4     4       0        0        4      faulty removed
>    5     5       0        0        5      faulty removed
> server2@server2:~$ sudo mdadm --examine /dev/sdc2
> /dev/sdc2:
>           Magic : a92b4efc
>         Version : 00.90.00
>            UUID : b7390ef0:9c9ecbe0:abcb3798:f53e3ccc
>   Creation Time : Sat Feb  6 17:37:20 2010
>      Raid Level : raid5
>   Used Dev Size : 1952537920 (1862.09 GiB 1999.40 GB)
>      Array Size : 9762689600 (9310.43 GiB 9996.99 GB)
>    Raid Devices : 6
>   Total Devices : 4
> Preferred Minor : 0
> 
>     Update Time : Tue Nov 16 14:29:15 2010
>           State : clean
>  Active Devices : 4
> Working Devices : 4
>  Failed Devices : 2
>   Spare Devices : 0
>        Checksum : aacb71d9 - correct
>          Events : 382240
> 
>          Layout : left-symmetric
>      Chunk Size : 64K
> 
>       Number   Major   Minor   RaidDevice State
> this     2       8       34        2      active sync   /dev/sdc2
> 
>    0     0       8        2        0      active sync   /dev/sda2
>    1     1       8       18        1      active sync   /dev/sdb2
>    2     2       8       34        2      active sync   /dev/sdc2
>    3     3       8       50        3      active sync   /dev/sdd2
>    4     4       0        0        4      faulty removed
>    5     5       0        0        5      faulty removed
> server2@server2:~$ sudo mdadm --examine /dev/sdd2
> /dev/sdd2:
>           Magic : a92b4efc
>         Version : 00.90.00
>            UUID : b7390ef0:9c9ecbe0:abcb3798:f53e3ccc
>   Creation Time : Sat Feb  6 17:37:20 2010
>      Raid Level : raid5
>   Used Dev Size : 1952537920 (1862.09 GiB 1999.40 GB)
>      Array Size : 9762689600 (9310.43 GiB 9996.99 GB)
>    Raid Devices : 6
>   Total Devices : 4
> Preferred Minor : 0
> 
>     Update Time : Tue Nov 16 14:29:15 2010
>           State : clean
>  Active Devices : 4
> Working Devices : 4
>  Failed Devices : 2
>   Spare Devices : 0
>        Checksum : aacb71eb - correct
>          Events : 382240
> 
>          Layout : left-symmetric
>      Chunk Size : 64K
> 
>       Number   Major   Minor   RaidDevice State
> this     3       8       50        3      active sync   /dev/sdd2
> 
>    0     0       8        2        0      active sync   /dev/sda2
>    1     1       8       18        1      active sync   /dev/sdb2
>    2     2       8       34        2      active sync   /dev/sdc2
>    3     3       8       50        3      active sync   /dev/sdd2
>    4     4       0        0        4      faulty removed
>    5     5       0        0        5      faulty removed
> server2@server2:~$ sudo mdadm --examine /dev/sde
> /dev/sde:
>           Magic : a92b4efc
>         Version : 00.90.00
>            UUID : b7390ef0:9c9ecbe0:abcb3798:f53e3ccc
>   Creation Time : Sat Feb  6 17:37:20 2010
>      Raid Level : raid5
>   Used Dev Size : 1952537920 (1862.09 GiB 1999.40 GB)
>      Array Size : 9762689600 (9310.43 GiB 9996.99 GB)
>    Raid Devices : 6
>   Total Devices : 5
> Preferred Minor : 0
> 
>     Update Time : Tue Nov 16 14:22:42 2010
>           State : clean
>  Active Devices : 4
> Working Devices : 5
>  Failed Devices : 2
>   Spare Devices : 1
>        Checksum : aacb70b7 - correct
>          Events : 382232
> 
>          Layout : left-symmetric
>      Chunk Size : 64K
> 
>       Number   Major   Minor   RaidDevice State
> this     6       8       64        6      spare   /dev/sde
> 
>    0     0       8        2        0      active sync   /dev/sda2
>    1     1       8       18        1      active sync   /dev/sdb2
>    2     2       8       34        2      active sync   /dev/sdc2
>    3     3       8       50        3      active sync   /dev/sdd2
>    4     4       0        0        4      faulty removed
>    5     5       0        0        5      faulty removed
>    6     6       8       64        6      spare   /dev/sde
> server2@server2:~$ sudo mdadm --examine /dev/sdf
> /dev/sdf:
>           Magic : a92b4efc
>         Version : 00.90.00
>            UUID : b7390ef0:9c9ecbe0:abcb3798:f53e3ccc
>   Creation Time : Sat Feb  6 17:37:20 2010
>      Raid Level : raid5
>   Used Dev Size : 1952537920 (1862.09 GiB 1999.40 GB)
>      Array Size : 9762689600 (9310.43 GiB 9996.99 GB)
>    Raid Devices : 6
>   Total Devices : 6
> Preferred Minor : 0
> 
>     Update Time : Tue Nov 16 14:20:02 2010
>           State : clean
>  Active Devices : 4
> Working Devices : 6
>  Failed Devices : 2
>   Spare Devices : 2
>        Checksum : aacb7088 - correct
>          Events : 382228
> 
>          Layout : left-symmetric
>      Chunk Size : 64K
> 
>       Number   Major   Minor   RaidDevice State
> this     6       8       80        6      spare   /dev/sdf
> 
>    0     0       8        2        0      active sync   /dev/sda2
>    1     1       8       18        1      active sync   /dev/sdb2
>    2     2       8       34        2      active sync   /dev/sdc2
>    3     3       8       50        3      active sync   /dev/sdd2
>    4     4       0        0        4      faulty removed
>    5     5       0        0        5      faulty removed
>    6     6       8       80        6      spare   /dev/sdf
>    7     7       8       64        7      spare   /dev/sde
> server2@server2:~$
> 
> # mdadm.conf
> #
> # Please refer to mdadm.conf(5) for information about this file.
> #
> 
> # by default, scan all partitions (/proc/partitions) for MD superblocks.
> # alternatively, specify devices to scan, using wildcards if desired.
> DEVICE partitions
> 
> # auto-create devices with Debian standard permissions
> CREATE owner=root group=disk mode=0660 auto=yes
> 
> # automatically tag new arrays as belonging to the local system
> HOMEHOST <system>
> 
> # instruct the monitoring daemon where to send mail alerts
> MAILADDR xxxxxxxxxxxx
> # definitions of existing MD arrays
> #ARRAY /dev/md0 level=raid5 num-devices=5 UUID=b7390ef0:9c9ecbe0:abcb3798:f53e3ccc
> ARRAY /dev/md1 level=raid0 num-devices=2 UUID=559552c6:b54f0dff:0dc726d6:bae63db2
> 
> # This file was auto-generated on Mon, 15 Mar 2010 22:25:29 -0400
> # by mkconf $Id$
> MAILFROM xxxxxxxxxxx
> 
> fdisk -l
> WARNING: GPT (GUID Partition Table) detected on '/dev/sda'! The util fdisk doesn't support GPT. Use GNU Parted.
> 
> 
> Disk /dev/sda: 2000.4 GB, 2000398934016 bytes
> 255 heads, 63 sectors/track, 243201 cylinders
> Units = cylinders of 16065 * 512 = 8225280 bytes
> Sector size (logical/physical): 512 bytes / 512 bytes
> I/O size (minimum/optimal): 512 bytes / 512 bytes
> Disk identifier: 0x00000000
> 
>    Device Boot      Start         End      Blocks   Id  System
> /dev/sda1   *           1      243202  1953514583+  ee  GPT
> 
> WARNING: GPT (GUID Partition Table) detected on '/dev/sdc'! The util fdisk doesn't support GPT. Use GNU Parted.
> 
> 
> Disk /dev/sdc: 2000.4 GB, 2000398934016 bytes
> 255 heads, 63 sectors/track, 243201 cylinders
> Units = cylinders of 16065 * 512 = 8225280 bytes
> Sector size (logical/physical): 512 bytes / 512 bytes
> I/O size (minimum/optimal): 512 bytes / 512 bytes
> Disk identifier: 0x00000000
> 
>    Device Boot      Start         End      Blocks   Id  System
> /dev/sdc1               1      243202  1953514583+  ee  GPT
> 
> Disk /dev/sde: 2000.4 GB, 2000398934016 bytes
> 255 heads, 63 sectors/track, 243201 cylinders
> Units = cylinders of 16065 * 512 = 8225280 bytes
> Sector size (logical/physical): 512 bytes / 512 bytes
> I/O size (minimum/optimal): 512 bytes / 512 bytes
> Disk identifier: 0x0009e37a
> 
>    Device Boot      Start         End      Blocks   Id  System
> 
> Disk /dev/sdf: 2000.4 GB, 2000398934016 bytes
> 255 heads, 63 sectors/track, 243201 cylinders
> Units = cylinders of 16065 * 512 = 8225280 bytes
> Sector size (logical/physical): 512 bytes / 512 bytes
> I/O size (minimum/optimal): 512 bytes / 512 bytes
> Disk identifier: 0x00048a02
> 
>    Device Boot      Start         End      Blocks   Id  System
> 
> WARNING: GPT (GUID Partition Table) detected on '/dev/sdb'! The util fdisk doesn't support GPT. Use GNU Parted.
> 
> 
> Disk /dev/sdb: 2000.4 GB, 2000398934016 bytes
> 255 heads, 63 sectors/track, 243201 cylinders
> Units = cylinders of 16065 * 512 = 8225280 bytes
> Sector size (logical/physical): 512 bytes / 512 bytes
> I/O size (minimum/optimal): 512 bytes / 512 bytes
> Disk identifier: 0x00000000
> 
>    Device Boot      Start         End      Blocks   Id  System
> /dev/sdb1               1      243202  1953514583+  ee  GPT
> 
> WARNING: GPT (GUID Partition Table) detected on '/dev/sdd'! The util fdisk doesn't support GPT. Use GNU Parted.
> 
> 
> Disk /dev/sdd: 2000.4 GB, 2000398934016 bytes
> 255 heads, 63 sectors/track, 243201 cylinders
> Units = cylinders of 16065 * 512 = 8225280 bytes
> Sector size (logical/physical): 512 bytes / 512 bytes
> I/O size (minimum/optimal): 512 bytes / 512 bytes
> Disk identifier: 0x00000000
> 
>    Device Boot      Start         End      Blocks   Id  System
> /dev/sdd1               1      243202  1953514583+  ee  GPT
> 
> Disk /dev/sdg: 250.1 GB, 250059350016 bytes
> 255 heads, 63 sectors/track, 30401 cylinders
> Units = cylinders of 16065 * 512 = 8225280 bytes
> Sector size (logical/physical): 512 bytes / 512 bytes
> I/O size (minimum/optimal): 512 bytes / 512 bytes
> Disk identifier: 0x5b4a57df
> 
>    Device Boot      Start         End      Blocks   Id  System
> /dev/sdg1   *           1       30401   244196001   fd  Linux raid autodetect
> 
> Disk /dev/sdh: 250.1 GB, 250059350016 bytes
> 255 heads, 63 sectors/track, 30401 cylinders
> Units = cylinders of 16065 * 512 = 8225280 bytes
> Sector size (logical/physical): 512 bytes / 512 bytes
> I/O size (minimum/optimal): 512 bytes / 512 bytes
> Disk identifier: 0x174b0f87
> 
>    Device Boot      Start         End      Blocks   Id  System
> /dev/sdh1               1       30401   244196001   fd  Linux raid autodetect
> 
> Disk /dev/md1: 500.1 GB, 500113211392 bytes
> 2 heads, 4 sectors/track, 122097952 cylinders
> Units = cylinders of 8 * 512 = 4096 bytes
> Sector size (logical/physical): 512 bytes / 512 bytes
> I/O size (minimum/optimal): 65536 bytes / 131072 bytes
> Disk identifier: 0x00000000
> 
> Disk /dev/md1 doesn't contain a valid partition table
> --
> To unsubscribe from this list: send the line "unsubscribe linux-raid" in
> the body of a message to majordomo@xxxxxxxxxxxxxxx
> More majordomo info at  http://vger.kernel.org/majordomo-info.html

--
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html


[Index of Archives]     [Linux RAID Wiki]     [ATA RAID]     [Linux SCSI Target Infrastructure]     [Linux Block]     [Linux IDE]     [Linux SCSI]     [Linux Hams]     [Device Mapper]     [Device Mapper Cryptographics]     [Kernel]     [Linux Admin]     [Linux Net]     [GFS]     [RPM]     [git]     [Yosemite Forum]


  Powered by Linux