On Sat, Feb 27, 2010 at 1:10 AM, Piergiorgio Sartor <piergiorgio.sartor@xxxxxxxx> wrote: > Hi, > >> Ok, please run this for each disk in the array: >> >> mdadm --examine /dev/(DEVICE) >> >> The output would be most readable if you did each array's devices in >> order, and you can list them on the same command (- - examine takes >> multiple inputs) >> >> If you still think the situation isn't as I described above, post the results. > > Well, here it is: > > $> mdadm -E /dev/sd[ab]2 > /dev/sda2: > Magic : a92b4efc > Version : 1.1 > Feature Map : 0x1 > Array UUID : 54db81a7:b47e9253:7291055e:4953c163 > Name : lvm > Creation Time : Fri Feb 6 20:17:13 2009 > Raid Level : raid10 > Raid Devices : 2 > > Avail Dev Size : 624928236 (297.99 GiB 319.96 GB) > Array Size : 624928000 (297.99 GiB 319.96 GB) > Used Dev Size : 624928000 (297.99 GiB 319.96 GB) > Data Offset : 264 sectors > Super Offset : 0 sectors > State : clean > Device UUID : 8f6cd2c4:0efc8286:09ec91c6:bc5014bf > > Internal Bitmap : 8 sectors from superblock > Update Time : Sat Feb 27 10:08:22 2010 > Checksum : 1703ded0 - correct > Events : 161646 > > Layout : far=2 > Chunk Size : 64K > > Device Role : spare > Array State : AA ('A' == active, '.' == missing) > /dev/sdb2: > Magic : a92b4efc > Version : 1.1 > Feature Map : 0x1 > Array UUID : 54db81a7:b47e9253:7291055e:4953c163 > Name : lvm > Creation Time : Fri Feb 6 20:17:13 2009 > Raid Level : raid10 > Raid Devices : 2 > > Avail Dev Size : 624928236 (297.99 GiB 319.96 GB) > Array Size : 624928000 (297.99 GiB 319.96 GB) > Used Dev Size : 624928000 (297.99 GiB 319.96 GB) > Data Offset : 264 sectors > Super Offset : 0 sectors > State : clean > Device UUID : 6e2763b5:9415b181:e41a9964:b0c21ca6 > > Internal Bitmap : 8 sectors from superblock > Update Time : Sat Feb 27 10:08:22 2010 > Checksum : 87d25401 - correct > Events : 161646 > > Layout : far=2 > Chunk Size : 64K > > Device Role : Active device 0 > Array State : AA ('A' == active, '.' == missing) > > And the details too: > > $> mdadm -D /dev/md1 > /dev/md1: > Version : 1.1 > Creation Time : Fri Feb 6 20:17:13 2009 > Raid Level : raid10 > Array Size : 312464000 (297.99 GiB 319.96 GB) > Used Dev Size : 312464000 (297.99 GiB 319.96 GB) > Raid Devices : 2 > Total Devices : 2 > Persistence : Superblock is persistent > > Intent Bitmap : Internal > > Update Time : Sat Feb 27 10:09:24 2010 > State : active > Active Devices : 2 > Working Devices : 2 > Failed Devices : 0 > Spare Devices : 0 > > Layout : far=2 > Chunk Size : 64K > > Name : lvm > UUID : 54db81a7:b47e9253:7291055e:4953c163 > Events : 161646 > > Number Major Minor RaidDevice State > 0 8 18 0 active sync /dev/sdb2 > 2 8 2 1 active sync /dev/sda2 > > bye, > > -- > > piergiorgio > I've checked my arrays and my only RAID-10 array has a single spare (hot spare) as part of the set with several other members. All current members storing DATA are listed as active members. What's confusing is that /proc/mdadm lists it as an active member (synced to data) but that the device does not match. Maybe you can stop/restart the array? -- To unsubscribe from this list: send the line "unsubscribe linux-raid" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html