yesterday I had a problem with the SATA bus one of my drives is on, which i've since fixed. I have a raid5 on four 20GB partitions (each on it's own drive, of course). The one that failed was sdc1. when i rebooted, the raid was no longer active (according to `mdadm -Q /dev/md0`). I ran `mdadm --examine ` on each drive. For sdc1 it said everything was fine, but for the other three drives sdc1 was marked as failed. I've never recovered from a drive failure before and I don't think I did it correctly. I removed sdc1 from the raid and then incrementally added it again. (I now realize I should've started the raid with it removed and done a backup.) I started the raid again with all four drives, and put it in readonly mode and tried to mount it, but it presumably still isn't set up right since it refuses to mount and `e2fsck -n` returns myriad errors. Did starting the raid with the bad disk destroy everything? did it only destory a little (assuming fsck can get the filesystem back into a useable state)? How can I even find out what's wrong? I do have a separate terabyte disk that i've copied images of the disks to, so I can perform experiment on them, but I'm not sure what to do. Help! Please! -Morgan -- To unsubscribe from this list: send the line "unsubscribe linux-raid" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html