Hi! Recently, I've updated one raid1 array to version 1.2 superblocks, in the (apparently) misguided hope, that it's needed to use bitmaps for faster resync after failure. In the process, I've stumbled across a few roadblocks. First, I've been unable to find out, how to specify the bitmap file without hardcoding its name in a startup file. Up until now, there was only a generic "mdadm -As" doing all the work. The other problem, even if I run this by hand, specifying the bitmap file, the array always comes up degraded, regardless of any bitmap specified. Granted, on reboot, some of the times may be because of unclean shutdowns, other times one of the component is marked as failed (F) in /proc/mdstat. After a remove/add, and waiting after the resync, all seems correct, but what irks me, is that mdadm -E always shows some components as failed even after a completely successful resync. I *think* that seeing "Array State : uu 1 failed" after the resync implies that I have to endure the resync after the next reboot. I remember stopping/starting the array correctly does a resync again, even without a reboot. It also seems to me, mdadm does not really store the bitmap file in the superblock for some reason, and/or there is no way to update this field. But what I find the most irksome, is that the array Details follow, sorry for the lengthy dump, I'm not sure if anything is useless from this. Kernel is 2.6.16-rc3, mdadm is 2.3.1. One time when it's started degraded for no apparent reason: 10:21:20 lame kernel: md: md0 stopped. 10:21:20 lame kernel: md: bind<hda3> 10:21:20 lame kernel: md: bind<hdc3> 10:21:20 lame kernel: raid1: raid set md0 active with 1 out of 2 mirrors Another time, which may result from a somehow unclean shutdown: 10:11:44 lame kernel: md: raid1 personality registered for level 1 10:11:44 lame kernel: md: md driver 0.90.3 MAX_MD_DEVS=256, MD_SB_DISKS=27 10:11:44 lame kernel: md: bitmap version 4.39 10:11:44 lame kernel: md: Skipping autodetection of RAID arrays. (raid=noautodetect) 10:11:44 lame kernel: md: md0 stopped. 10:11:44 lame kernel: md: bind<hda3> 10:11:44 lame kernel: md: bind<hdc3> 10:11:44 lame kernel: md: kicking non-fresh hda3 from array! 10:11:44 lame kernel: md: unbind<hda3> 10:11:44 lame kernel: md: export_rdev(hda3) 16:20:24 lame kernel: md: bind<hda3> 16:20:24 lame kernel: md: syncing RAID array md0 16:20:24 lame kernel: md: minimum _guaranteed_ reconstruction speed: 1000 KB/sec/disc. 16:20:24 lame kernel: md: using maximum available idle IO bandwidth (but not more than 200000 KB/sec) for reconstruction. 16:20:24 lame kernel: md: using 128k window, over a total of 17552232 blocks. 16:58:07 lame kernel: md: md0: sync done. After the resync, however, the components do not get updated to show correctness? ------------------------ # cat /proc/mdstat Personalities : [raid1] md0 : active raid1 hda3[2] hdc3[1] 17552232 blocks super 1.2 [2/2] [UU] unused devices: <none> ------------------------ # mdadm -D /dev/md0; mdadm -E /dev/hd[ac]3 /dev/md0: Version : 01.02.03 Creation Time : Wed Feb 8 21:24:33 2006 Raid Level : raid1 Array Size : 17552232 (16.74 GiB 17.97 GB) Device Size : unknown Raid Devices : 2 Total Devices : 2 Preferred Minor : 0 Persistence : Superblock is persistent Update Time : Sun Feb 19 17:12:09 2006 State : clean Active Devices : 2 Working Devices : 2 Failed Devices : 0 Spare Devices : 0 Name : UUID : 4dd4... Events : 467156 Number Major Minor RaidDevice State 2 3 3 0 active sync /dev/hda3 1 22 3 1 active sync /dev/hdc3 /dev/hda3: Magic : a92b4efc Version : 01 Feature Map : 0x0 Array UUID : 4dd4... Name : Creation Time : Wed Feb 8 21:24:33 2006 Raid Level : raid1 Raid Devices : 2 Device Size : 35104472 (16.74 GiB 17.97 GB) Array Size : 35104464 (16.74 GiB 17.97 GB) Used Size : 35104464 (16.74 GiB 17.97 GB) Data Offset : 136 sectors Super Offset : 8 sectors State : active Device UUID : aca9... Update Time : Sun Feb 19 17:12:09 2006 Checksum : dd1ac0a1 - correct Events : 467156 Array State : uu 1 failed /dev/hdc3: Magic : a92b4efc Version : 01 Feature Map : 0x0 Array UUID : 4dd4... Name : Creation Time : Wed Feb 8 21:24:33 2006 Raid Level : raid1 Raid Devices : 2 Device Size : 35104472 (16.74 GiB 17.97 GB) Array Size : 35104464 (16.74 GiB 17.97 GB) Used Size : 35104464 (16.74 GiB 17.97 GB) Data Offset : 136 sectors Super Offset : 8 sectors State : active Device UUID : 66d7... Update Time : Sun Feb 19 17:12:09 2006 Checksum : d990f459 - correct Events : 467156 Array State : uU 1 failed ------------------------ - To unsubscribe from this list: send the line "unsubscribe linux-raid" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html