For a period of a year I've had power loss to more than one drive on more than one ocasion (y cables SUCK). a month ago I had power loss after I dist-upgrade my debian server with a raid6 array of 8 drives. I then proceeded to re-create the raid array only to discover to my horror that 2 things were seriously wrong. #1 The 2.6.32 kernel discovered the sata HDDs in a different order than the 2.6.26 kernel. So /dev/sdc was now /dev/sdg type scenario. Horror #2 was that I forgot to specify --assume-clean on the re-creation. Horror #3 is that I discovered that the dist-upgrade had also updated mdadm so that it used superblock version 1.2 when the raid array had been created with superblock version 0.9. Now I'm in the process of writing a C app that is assisting me to identify which chunks in each stripe of the raid array is data and which is parity. I was successful at discovering the original ordering of the drives using my app and have recreated the array with --assume-clean under 2.6.26 with superblock ver 0.9. So far so good. I now find the ReIsEr2Fs and ReIsErLB tags at the proper offsets based on the LVM2 metadata that I had a backup of. However I'm still seeing data corruption in chunks that appears to be semi-random. so, assuming the following parameters used to run mdadm with the proper flags for chunk size, pairity, etc, what should the ordering of the data, P and Q chunks be on the stripes? I attempted to work this out by reading wikipedia's raid6, and the kernel code and various pages on the net, but none of them seem to agree. 8-( Can anyone tell me what that ordering or chunks should be for the first 16 stripes so that I can finally work out the bad ordering under 2.6.32 and rebuild my raid array "by hand"? Yes, I permanently "fixed" the Y cable problem by manufacturing a power backplane for the 12 drives in the system. :) Thanks! #mdadm --examine /dev/sdc /dev/sdc: Magic : a92b4efc Version : 0.90.00 UUID : 5108f419:93f471c4:36668fe2:a3f57f0d (local to host iscsi1) Creation Time : Tue Dec 21 00:55:29 2010 Raid Level : raid6 Used Dev Size : 976762496 (931.51 GiB 1000.20 GB) Array Size : 5860574976 (5589.08 GiB 6001.23 GB) Raid Devices : 8 Total Devices : 8 Preferred Minor : 0 Update Time : Tue Dec 21 00:55:29 2010 State : clean Active Devices : 8 Working Devices : 8 Failed Devices : 0 Spare Devices : 0 Checksum : 3cdf9d27 - correct Events : 1 Layout : left-symmetric Chunk Size : 128K Number Major Minor RaidDevice State this 0 8 32 0 active sync /dev/sdc 0 0 8 32 0 active sync /dev/sdc 1 1 8 96 1 active sync /dev/sdg 2 2 8 128 2 active sync /dev/sdi 3 3 8 144 3 active sync /dev/sdj 4 4 8 64 4 active sync /dev/sde 5 5 8 80 5 active sync /dev/sdf 6 6 8 48 6 active sync /dev/sdd 7 7 8 112 7 active sync /dev/sdh i -- To unsubscribe from this list: send the line "unsubscribe linux-raid" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html