Hi folks, While attempting to grow a RAID1 array after replacing the disks, it looks like I've messed things up a bit. I started with this: ~# mdadm --detail /dev/md1 /dev/md1: Version : 0.90 Creation Time : Sat Jun 19 23:34:39 2010 Raid Level : raid1 Array Size : 767130560 (731.59 GiB 785.54 GB) Used Dev Size : 767130560 (731.59 GiB 785.54 GB) Raid Devices : 2 Total Devices : 2 Preferred Minor : 1 Persistence : Superblock is persistent Update Time : Mon Dec 24 21:02:13 2012 State : clean Active Devices : 2 Working Devices : 2 Failed Devices : 0 Spare Devices : 0 UUID : 9be57e97:2c46675a:b5a3dfee:de98bb27 Events : 0.2246 Number Major Minor RaidDevice State 0 8 66 0 active sync /dev/sde2 1 8 82 1 active sync /dev/sdf2 Growing the array at this point wasn't working, because for some reason the size of one of the two partitions (/dev/sdf2) used for the array was still being reported with its previous smaller size (767130560) in /sys/ devices/virtual/block/md1/md/dev-sde2/size. So, I stopped the array to experiment with the "--update=devicesize" option, but then stupidly issued these commands: ~# mdadm --assemble /dev/md1 --update=devicesize /dev/sdf2 mdadm: /dev/md1 assembled from 1 drive - need all 2 to start it (use -- run to insist). ~# mdadm --assemble /dev/md1 --update=devicesize /dev/sdf2 /dev/sde2 mdadm: cannot open device /dev/sdf2: Device or resource busy mdadm: /dev/sdf2 has no superblock - assembly aborted After this, the array refused to respond to any commands until I stopped it once more. However, it then refused to start up again (assemble). So I saw somewhere that I might be able to recover it with a create command: ~# mdadm --create /dev/md1 --level=1 --raid-devices=2 /dev/sd[ef]2 mdadm: /dev/sde2 appears to be part of a raid array: level=raid1 devices=2 ctime=Sat Jun 19 23:34:39 2010 mdadm: Note: this array has metadata at the start and may not be suitable as a boot device. If you plan to store '/boot' on this device please ensure that your boot-loader understands md/v1.x metadata, or use --metadata=0.90 mdadm: /dev/sdf2 appears to be part of a raid array: level=raid1 devices=2 ctime=Sat Jun 19 23:34:39 2010 mdadm: largest drive (/dev/sde2) exceeds size (767129464K) by more than 1% Continue creating array? y mdadm: Defaulting to version 1.2 metadata mdadm: array /dev/md1 started. Perhaps this was an even more destructive thing to do. Anyway, the array now looks like this: ~# mdadm --detail /dev/md1 /dev/md1: Version : 1.2 Creation Time : Tue Dec 25 02:07:11 2012 Raid Level : raid1 Array Size : 767129464 (731.59 GiB 785.54 GB) Used Dev Size : 767129464 (731.59 GiB 785.54 GB) Raid Devices : 2 Total Devices : 2 Persistence : Superblock is persistent Update Time : Tue Dec 25 02:07:11 2012 State : clean Active Devices : 2 Working Devices : 2 Failed Devices : 0 Spare Devices : 0 Name : bitis:1 (local to host bitis) UUID : a8eca0e2:dd9941a3:03d8b0f0:d11c9994 Events : 0 Number Major Minor RaidDevice State 0 8 66 0 active sync /dev/sde2 1 8 82 1 active sync /dev/sdf2 It now has a different UUID (that was expected), but it also seems to contain no data. I was hoping to find an LVM physical volume, but pvscan is not detecting anything on the array. Oops. Luckily the data wasn't too important, but the loss is nevertheless irritating. Might there still be a way to recover my lost LVM data, or is this situation hopeless? Thanks, Jaap -- To unsubscribe from this list: send the line "unsubscribe linux-raid" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html