Hi We tried the same steps with some other device with following configuration. root@p2041rdb:~# uname -a Linux p2041rdb 3.0.34-rt55 #67 SMP Sat Jan 19 14:48:21 KST 2013 ppc GNU/Linux root@p2041rdb:~# parted -v parted (GNU parted) 3.1 root@p2041rdb:~# mdadm -V mdadm - v3.2.2 - 17th June 2011 But we didn't get any problem either with partitioning or with the RAID building. We are facing the problem only with the device having the following configuration. root@mpc8536ds ~ # uname -a Linux mpc8536ds 2.6.32-rc5 #9 PREEMPT Wed Feb 6 20:54:44 KST 2013 ppc unknown root@mpc8536ds ~ # parted -v parted (GNU parted) 2.3 root@mpc8536ds ~ # mdadm -V mdadm - v3.2.5 - 18th May 2012 I think the problem is because of the compatibility. Anyone is having the idea about the compatibility issues on any of the above combination? Thanks Tarak Anumolu ------- Original Message ------- Sender : Robin Hill<robin@xxxxxxxxxxxxxxx> Date : Mar 28, 2013 19:25 (GMT+09:00) Title : Re: Problem with mdadm 3.2.5 On Thu Mar 28, 2013 at 06:36:03AM +0000, Tarak Anumolu wrote: > > Hi > > FYI, We followed the below steps and At the end you can see the > problem with the file system. > > RAID operation on 8 harddisks each of size 1TB with 7 harddisks as raid devices and 1 hard disk as spare device got succeed. > > #parted -s /dev/md0 print > Model: Linux Software RAID Array (md) > Disk /dev/md0: 6001GB > Sector size (logical/physical): 512B/512B > Partition Table: gpt > Number Start End Size File system Name Flags > 1 1049kB 60.0GB 60.0GB xfs primary > 2 60.0GB 6001GB 5941GB xfs primary > > > Then We create 2 partitions md0p1 and md0p2. > > #cat /proc/partitions > major minor #blocks name > 31 0 8192 mtdblock0 > 31 1 131072 mtdblock1 > 8 0 976762584 sda > 8 1 976760832 sda1 > 8 16 976762584 sdb > 8 17 976760832 sdb1 > 8 32 976762584 sdc > 8 33 976760832 sdc1 > 8 48 976762584 sdd > 8 49 976760832 sdd1 > 8 64 976762584 sde > 8 65 976760832 sde1 > 8 80 976762584 sdf > 8 81 976760832 sdf1 > 8 96 976762584 sdg > 8 97 976760832 sdg1 > 8 112 976762584 sdh > 8 113 976760832 sdh1 > 9 0 5860563456 md0 > 259 0 58604544 md0p1 > 259 1 5801957376 md0p2 > > *************************************************************************************************** > IT'S FINE UPTO HERE > *************************************************************************************************** > > Now we failed harddisk-1 > > # mdadm -f /dev/md0 /dev/sda1 > > # mdadm -D /dev/md0 > /dev/md0: > Version : 0.90 > Creation Time : Wed Mar 27 11:10:24 2013 > Raid Level : raid5 > Array Size : 5860563456 (5589.07 GiB 6001.22 GB) > Used Dev Size : 976760576 (931.51 GiB 1000.20 GB) > Raid Devices : 7 > Total Devices : 7 > Preferred Minor : 0 > Persistence : Superblock is persistent > Intent Bitmap : Internal > Update Time : Thu Mar 28 01:03:57 2013 > State : active, degraded, recovering > Active Devices : 6 > Working Devices : 7 > Failed Devices : 0 > Spare Devices : 1 > Layout : left-symmetric > Chunk Size : 256K > Rebuild Status : 0% complete > UUID : debadbe0:49b4fe90:24472787:29621eca (local to host mpc8536ds) > Events : 0.15 > Number Major Minor RaidDevice State > 7 8 113 0 spare rebuilding /dev/sdh1 > 1 8 17 1 active sync /dev/sdb1 > 2 8 33 2 active sync /dev/sdc1 > 3 8 49 3 active sync /dev/sdd1 > 4 8 65 4 active sync /dev/sde1 > 5 8 81 5 active sync /dev/sdf1 > 6 8 97 6 active sync /dev/sdg1 > > Now harddisk-1 is revovering > > #cat /proc/mdstat > Personalities : [raid6] [raid5] [raid4] > md0 : active raid5 sdh1[7] sdg1[6] sdf1[5] sde1[4] sdd1[3] sdc1[2] sdb1[1] > 5860563456 blocks level 5, 256k chunk, algorithm 2 [7/6] [_UUUUUU] > [>....................] recovery = 0.1% (1604164/976760576) finish=324.2min speed=50130K/sec > bitmap: 0/8 pages [0KB], 65536KB chunk > > > #parted -s /dev/md0 print > Model: Linux Software RAID Array (md) > Disk /dev/md0: 6001GB > Sector size (logical/physical): 512B/512B > Partition Table: gpt > Number Start End Size File system Name Flags > 1 1049kB 60.0GB 60.0GB xfs primary > 2 60.0GB 6001GB 5941GB xfs primary > > > While recovering the harddisk, to test the power failure/ restarting situation, we unmount the partitions. > > #umount /dev/md0p[12] > > > Again try to mount the partitions but failed. > > > #mount /dev/md0p1 /mnt/md0p1 > UDF-fs: No partition found (1) > Filesystem "md0p1": Disabling barriers, trial barrier write failed > > # mount /dev/md0p2 /mnt/md0p2 > grow_buffers: requested out-of-range block 18446744072428564479 for device md0p2 > grow_buffers: requested out-of-range block 18446744072428564223 for device md0p2 > grow_buffers: requested out-of-range block 18446744072428564478 for device md0p2 > grow_buffers: requested out-of-range block 18446744072428564222 for device md0p2 > grow_buffers: requested out-of-range block 18446744072428564480 for device md0p2 > grow_buffers: requested out-of-range block 18446744072428564224 for device md0p2 > grow_buffers: requested out-of-range block 18446744072428564477 for device md0p2 > > > #parted -s /dev/md0 print > Model: Linux Software RAID Array (md) > Disk /dev/md0: 6001GB > Sector size (logical/physical): 512B/512B > Partition Table: gpt > > Number Start End Size File system Name Flags > 1 1049kB 60.0GB 60.0GB xfs primary > 2 60.0GB 6001GB 5941GB primary > > Filesystem is not shown. > > > Harddisk Recovery is completed > > # cat /proc/mdstat > Personalities : [raid6] [raid5] [raid4] > md0 : active raid5 sdh1[0] sdg1[6] sdf1[5] sde1[4] sdd1[3] sdc1[2] sdb1[1] > 5860563456 blocks level 5, 256k chunk, algorithm 2 [7/7] [UUUUUUU] > bitmap: 1/8 pages [4KB], 65536KB chunk > > #parted -s /dev/md0 print > Model: Linux Software RAID Array (md) > Disk /dev/md0: 6001GB > Sector size (logical/physical): 512B/512B > Partition Table: gpt > > Number Start End Size File system Name Flags > 1 1049kB 60.0GB 60.0GB xfs primary > 2 60.0GB 6001GB 5941GB primary > > Filesystem is empty......... > > > Please tell me if I did any thing wrong. > That all looks perfectly valid to me. You're obviously getting some sort of data corruption during the rebuild (the fact you're using a partitioned RAID array is irrelevant - it's just highlighting the issue). You are using the old 0.9 metadata, but you're not hitting any of the limitations of that here. I doubt mdadm itself has much to do with this though, as it's just passing high-level instructions on to the kernel, which will perform the low-level recovery process. What kernel version are you using? Cheers, Robin -- ___ ( ' } | Robin Hill | / / ) | Little Jim says .... | // !! | "He fallen in de water !!" |ÿôèº{.nÇ+‰·Ÿ®‰†+%ŠËÿ±éݶ¥Šwÿº{.nÇ+‰·¥Š{±þ¶¢wø§¶›¡Ü¨}©ž²Æ zÚ&j:+v‰¨þø¯ù®w¥þŠà2ŠÞ™¨èÚ&¢)ß¡«a¶Úÿÿûàz¿äz¹Þ—ú+ƒùšŽŠÝ¢jÿŠwèþf