raid1 with 1.2 superblock never marked healthy?

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



Hi!

Recently, I've updated one raid1 array to version 1.2 superblocks, in
the (apparently) misguided hope, that it's needed to use bitmaps for
faster resync after failure.

In the process, I've stumbled across a few roadblocks.

First, I've been unable to find out, how to specify the bitmap file
without hardcoding its name in a startup file.  Up until now, there was
only a generic "mdadm -As" doing all the work.  The other problem, even
if I run this by hand, specifying the bitmap file, the array always
comes up degraded, regardless of any bitmap specified.

Granted, on reboot, some of the times may be because of unclean
shutdowns, other times one of the component is marked as failed (F) in
/proc/mdstat.  After a remove/add, and waiting after the resync, all
seems correct, but what irks me, is that mdadm -E always shows some
components as failed even after a completely successful resync.

I *think* that seeing "Array State : uu 1 failed" after the resync
implies that I have to endure the resync after the next reboot.

I remember stopping/starting the array correctly does a resync again,
even without a reboot.

It also seems to me, mdadm does not really store the bitmap file in the
superblock for some reason, and/or there is no way to update this field.

But what I find the most irksome, is that the array 

Details follow, sorry for the lengthy dump, I'm not sure if anything is
useless from this.

Kernel is 2.6.16-rc3, mdadm is 2.3.1.

One time when it's started degraded for no apparent reason:

10:21:20 lame kernel: md: md0 stopped.
10:21:20 lame kernel: md: bind<hda3>
10:21:20 lame kernel: md: bind<hdc3>
10:21:20 lame kernel: raid1: raid set md0 active with 1 out of 2 mirrors

Another time, which may result from a somehow unclean shutdown:

10:11:44 lame kernel: md: raid1 personality registered for level 1
10:11:44 lame kernel: md: md driver 0.90.3 MAX_MD_DEVS=256, MD_SB_DISKS=27
10:11:44 lame kernel: md: bitmap version 4.39
10:11:44 lame kernel: md: Skipping autodetection of RAID arrays.  (raid=noautodetect)
10:11:44 lame kernel: md: md0 stopped.
10:11:44 lame kernel: md: bind<hda3>
10:11:44 lame kernel: md: bind<hdc3>
10:11:44 lame kernel: md: kicking non-fresh hda3 from array!
10:11:44 lame kernel: md: unbind<hda3>
10:11:44 lame kernel: md: export_rdev(hda3)
16:20:24 lame kernel: md: bind<hda3>
16:20:24 lame kernel: md: syncing RAID array md0
16:20:24 lame kernel: md: minimum _guaranteed_ reconstruction speed: 1000 KB/sec/disc.
16:20:24 lame kernel: md: using maximum available idle IO bandwidth (but not more than 200000 KB/sec) for reconstruction.
16:20:24 lame kernel: md: using 128k window, over a total of 17552232 blocks.
16:58:07 lame kernel: md: md0: sync done.

After the resync, however, the components do not get updated to show
correctness?

------------------------
# cat /proc/mdstat
Personalities : [raid1]
md0 : active raid1 hda3[2] hdc3[1]
      17552232 blocks super 1.2 [2/2] [UU]

unused devices: <none>
------------------------
# mdadm -D /dev/md0; mdadm -E /dev/hd[ac]3
/dev/md0:
        Version : 01.02.03
  Creation Time : Wed Feb  8 21:24:33 2006
     Raid Level : raid1
     Array Size : 17552232 (16.74 GiB 17.97 GB)
    Device Size : unknown
   Raid Devices : 2
  Total Devices : 2
Preferred Minor : 0
    Persistence : Superblock is persistent

    Update Time : Sun Feb 19 17:12:09 2006
          State : clean
 Active Devices : 2
Working Devices : 2
 Failed Devices : 0
  Spare Devices : 0

           Name :
           UUID : 4dd4...
         Events : 467156

    Number   Major   Minor   RaidDevice State
       2       3        3        0      active sync   /dev/hda3
       1      22        3        1      active sync   /dev/hdc3
/dev/hda3:
          Magic : a92b4efc
        Version : 01
    Feature Map : 0x0
     Array UUID : 4dd4...
           Name :
  Creation Time : Wed Feb  8 21:24:33 2006
     Raid Level : raid1
   Raid Devices : 2

    Device Size : 35104472 (16.74 GiB 17.97 GB)
     Array Size : 35104464 (16.74 GiB 17.97 GB)
      Used Size : 35104464 (16.74 GiB 17.97 GB)
    Data Offset : 136 sectors
   Super Offset : 8 sectors
          State : active
    Device UUID : aca9...

    Update Time : Sun Feb 19 17:12:09 2006
       Checksum : dd1ac0a1 - correct
         Events : 467156


   Array State : uu 1 failed
/dev/hdc3:
          Magic : a92b4efc
        Version : 01
    Feature Map : 0x0
     Array UUID : 4dd4...
           Name :
  Creation Time : Wed Feb  8 21:24:33 2006
     Raid Level : raid1
   Raid Devices : 2

    Device Size : 35104472 (16.74 GiB 17.97 GB)
     Array Size : 35104464 (16.74 GiB 17.97 GB)
      Used Size : 35104464 (16.74 GiB 17.97 GB)
    Data Offset : 136 sectors
   Super Offset : 8 sectors
          State : active
    Device UUID : 66d7...

    Update Time : Sun Feb 19 17:12:09 2006
       Checksum : d990f459 - correct
         Events : 467156


   Array State : uU 1 failed
------------------------
-
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html

[Index of Archives]     [Linux RAID Wiki]     [ATA RAID]     [Linux SCSI Target Infrastructure]     [Linux Block]     [Linux IDE]     [Linux SCSI]     [Linux Hams]     [Device Mapper]     [Device Mapper Cryptographics]     [Kernel]     [Linux Admin]     [Linux Net]     [GFS]     [RPM]     [git]     [Yosemite Forum]


  Powered by Linux