mdadm --run required to start raid1 device /dev/md0 after removal of one drive. why?

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi,

My system has 4 250GB hard drives:

/dev/md0  /dev/hdb1, /dev/hdg1
/dev/md1  /dev/hdc1,  /dev/hde1

My data is static, so 
I decided to take a single device from each raid off line and keep it safe.

This was the content of the /etc/mdadm/mdadm.conf file at that moment.

# cat mdadm.conf.6.25.2005
DEVICE /dev/hdb1 /dev/hdc1  /dev/hdg1 /dev/hde1

ARRAY /dev/md1 level=raid1 num-devices=2 
UUID=19def19e:3995fe1e:5470d62a:c44b06c
   devices=/dev/hdc1,/dev/hde1
ARRAY /dev/md0 level=raid1 num-devices=2 
UUID=ca17eac7:01b4ccb7:c768ca76:7b9febd
   devices=/dev/hdb1,/dev/hdg1


So I edited mdadm.conf to  changing  num-devices=2 to num-devices=1:

A1:~# cat /etc/mdadm/mdadm.conf
DEVICE /dev/hdb1 /dev/hdc1

ARRAY /dev/md1 level=raid1 num-devices=1 
UUID=19def19e:3995fe1e:5470d62a:c44b06c
   devices=/dev/hdc1
ARRAY /dev/md0 level=raid1 num-devices=1 
UUID=ca17eac7:01b4ccb7:c768ca76:7b9febd
   devices=/dev/hdb1


And then I turned off the system, and pulled out both of  the drives /dev/hdg1 
and /dev/hde1.

Then I rebooted the sytem, and the system stopped during the standard boot 
point where it runs fsck.ext3 on the mounted drives. 
because the system was unable to start the devices /dev/md0 or /dev/md1

It refused to boot until I removed the mount points /dev/md0 and /dev/md1 from 
/etc/fstab.

Now I naively thought that /dev/md0 and /dev/md1  would both start naturally,
with the new mdadm.conf and simply indicate a failed device in 
cat /proc/mdstat?

That is what hardware raid does ... I think ...
  
Well after I commented out /dev/md* from /etc/fstab
I  got the system to boot, and  I ran

mdadm -A /dev/md0 /dev/hdb1

and the system said to me:

mdadm: /dev/md0 assembled from 1 drive, need all 2 to start it.
use --run to insist.

Now if I do 
mmdadm -A --run /dev/md0 /dev/hdb1 then it starts.

similarly with /dev/md1.

But what can I do to restore the previous situation with the old /etc/fstab
so that I can reboot and have the system 
remount itself without doing this --run by hand each time? 

and also what is rhe --run needed here for??

this is the current content of 
mdadm --detail --scan

ARRAY /dev/md0 level=raid1 num-devices=2 
UUID=ca17eac7:01b4ccb7:c768ca76:7b9febd
   devices=/dev/hdb1
ARRAY /dev/md1 level=raid1 num-devices=2 
UUID=19def19e:3995fe1e:5470d62a:c44b06c
   devices=/dev/hdc1

while i had put num-devices=1 in the new mdadm.conf . Is this the problem?

I am running debian sarge 
A1:~# uname -a
Linux A1 2.6.8-1-386 #1 Thu Nov 25 04:24:08 UTC 2004 i686 GNU/Linux
A1:~# mdadm -V
mdadm - v1.9.0 - 04 February 2005

also here is the output of 

A1:~# mdadm -E /dev/hdb1
/dev/hdb1:
          Magic : a92b4efc
        Version : 00.90.00
           UUID : ca17eac7:01b4ccb7:c768ca76:7b9febad
  Creation Time : Mon Jan 10 13:55:58 2005
     Raid Level : raid1
   Raid Devices : 2
  Total Devices : 1
Preferred Minor : 0

    Update Time : Sun Jun 26 01:15:21 2005
          State : clean
 Active Devices : 1
Working Devices : 1
 Failed Devices : 1
  Spare Devices : 0
       Checksum : 4b3c0cec - correct
         Events : 0.364978


      Number   Major   Minor   RaidDevice State
this     0       3       65        0      active sync   /dev/hdb1

   0     0       3       65        0      active sync   /dev/hdb1
   1     1       0        0        1      faulty removed



Thanks!!!

Mitchell Laks
-
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html

[Index of Archives]     [Linux RAID Wiki]     [ATA RAID]     [Linux SCSI Target Infrastructure]     [Linux Block]     [Linux IDE]     [Linux SCSI]     [Linux Hams]     [Device Mapper]     [Device Mapper Cryptographics]     [Kernel]     [Linux Admin]     [Linux Net]     [GFS]     [RPM]     [git]     [Yosemite Forum]


  Powered by Linux