Re: Problem with auto-assembling raid1 on system start

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Does mdadm.conf in initrd image contains valid uuids/array names? (you
can ungzip && extract cpio archive to check this)

On Fri, May 8, 2009 at 6:11 PM, Tobias Gunkel <tobias.gunkel@xxxxxxxxx> wrote:
> Hello everyone!
>
> After rebooting one of our Debian servers yesterday (under normal
> conditions), mdadm was not able to assemble /dev/md0 automaticly any more.
> System: Debian Lenny, mdmadm v2.5.6, Kernel 2.6.26-preemptive-cpuset (from
> Debian testing sources)
>
> This is what I get during boot:
>
>  [...]
>  Begin: Mounting root file system... ...
>  Begin: Running /scripts/local-top ...
>  Begin: Loading MD modules ...
>  md: raid1 personality registered for level 1
>  Success: loaded module raid1.
>  Done.
>  Begin: Assembling all MD arrays ...
>  [...]
>  md: md0 stopped.
>  mdadm: no devices found for /dev/md0
>  Failure: failed to assemble all arrays.
>  [...]
>
> Then the system falls back to BusyBox shell from initramfs, because the root
> fs  - which is located on /dev/md0 - could not be mounted.
> But from the initramfs shell, it is possible to cleanly assemble and mount
> the md0 array:
>
>  (initramfs) mdadm -A /dev/md0 /dev/sda2 /dev/sdb2
>  md: md0 stopped.
>  md: bind<sdb2>
>  md: bind<sda2>
>  raid1: raid set md0 active with 2 out of 2 mirrors
>  mdadm: /dev/md0 has been started with 2 drives.
>
>  (initramfs) mount /dev/md0 root
>  kjournald starting. Commit interval 5 seconds
>  EXT3 FS on md0, internal journal
>  EXT3-fs: mounted filesystem with ordered data mode.
>
> After leaving the initramfs shell with 'exit', the system continues to boot
> normally.
>
> Strange: /dev/md1 (swap) which is the first array in assembling order, gets
> assembled and started correctly.
> I also played around with ROOTDELAY=60, but this did not changed anything.
>
> I'm grateful for any help.
> Best regards, Tobias
>
>
> PS: Maybe some helpful output (after starting the system the way described
> above):
>
> $ cat /proc/mdstat
> Personalities : [raid1]
> md0 : active raid1 sda2[0] sdb2[1]
>     487331648 blocks [2/2] [UU]
>
> md1 : active raid1 sda1[0] sdb1[1]
>     1052160 blocks [2/2] [UU]
>
> unused devices: <none>
>
>
> $ mdadm --detail --scan
> ARRAY /dev/md1 level=raid1 num-devices=2
> UUID=c3838888:50dbed72:15a9bffb:d0e83d23
> ARRAY /dev/md0 level=raid1 num-devices=2
> UUID=0d0a0c79:70adae03:f802952b:2b58c14d
>
>
> $ grep -v ^# /etc/mdadm/mdadm.conf
>
> DEVICE /dev/sd*[0-9] /dev/sd*[0-9]
>
> CREATE owner=root group=disk mode=0660 auto=yes
>
> HOMEHOST <system>
>
> MAILADDR root
>
> ARRAY /dev/md1 level=raid1 num-devices=2
> UUID=c3838888:50dbed72:15a9bffb:d0e83d23
> ARRAY /dev/md0 level=raid1 num-devices=2
> UUID=0d0a0c79:70adae03:f802952b:2b58c14d
>
>
> $ mdadm --detail /dev/md0
> /dev/md0:
>       Version : 00.90.03
>  Creation Time : Thu Dec 11 14:18:44 2008
>    Raid Level : raid1
>    Array Size : 487331648 (464.76 GiB 499.03 GB)
>   Device Size : 487331648 (464.76 GiB 499.03 GB)
>  Raid Devices : 2
>  Total Devices : 2
> Preferred Minor : 0
>   Persistence : Superblock is persistent
>
>   Update Time : Fri May  8 15:45:32 2009
>         State : clean
> Active Devices : 2
> Working Devices : 2
> Failed Devices : 0
>  Spare Devices : 0
>
>          UUID : 0d0a0c79:70adae03:f802952b:2b58c14d
>        Events : 0.900
>
>   Number   Major   Minor   RaidDevice State
>      0       8        2        0      active sync   /dev/sda2
>      1       8       18        1      active sync   /dev/sdb2
>
>
> --
> To unsubscribe from this list: send the line "unsubscribe linux-raid" in
> the body of a message to majordomo@xxxxxxxxxxxxxxx
> More majordomo info at  http://vger.kernel.org/majordomo-info.html
>



-- 
-- 
Best regards,
[COOLCOLD-RIPN]
--
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html

[Index of Archives]     [Linux RAID Wiki]     [ATA RAID]     [Linux SCSI Target Infrastructure]     [Linux Block]     [Linux IDE]     [Linux SCSI]     [Linux Hams]     [Device Mapper]     [Device Mapper Cryptographics]     [Kernel]     [Linux Admin]     [Linux Net]     [GFS]     [RPM]     [git]     [Yosemite Forum]


  Powered by Linux