On Tue, 20 Nov 2012 15:41:57 -0500 Stephen Haran <steveharan@xxxxxxxxx> wrote: > Hi, I'm trying to recover a Western Digital Share Space NAS. > I'm able to assemble the RAID5 and restore the LVM but it can't see > any filesystem. > > Below is a raid.log file that shows how the raid was configured when > it was working. > And also the output of mdadm -D showing the raid in it's current state. > Note the Version difference 00.90.01 vs. 0.90. And the array size > difference 2925293760 vs. 2925894144 > I'm thinking this difference may be the reason Linux can not see a filesystem. Probably not - losing a few blocks from the end might make 'fsck' complain, but it should still be able to see the filesystem. How did you test if you could see a filesystem? 'mount' or 'fsck -n' ? It looks like you re-created the array recently (Nov 18 12:07:53 2012) Why did you do that? It has been created slightly smaller - not sure why. Maybe if you explicitly request the old per-device size with "--size=975097920" it might get it right. Are you sure the dm cow devices show exactly the same size and content as the originals? The stray '.01' at the end of the version number is not relevant. It just indicates a different version of mdadm in use to report the array. NeilBrown > > My question is would the version difference explain the array size difference? > And is it possible to create a version 00.90.01 array? I do not see > that in the mdadm docs. > > ....original working raid config.... > /dev/md2: > Version : 00.90.01 > Creation Time : Wed Jun 24 19:00:59 2009 > Raid Level : raid5 > Array Size : 2925293760 (2789.78 GiB 2995.50 GB) > Device Size : 975097920 (929.93 GiB 998.50 GB) > Raid Devices : 4 > Total Devices : 4 > Preferred Minor : 2 > Persistence : Superblock is persistent > > Update Time : Thu Jun 25 02:36:31 2009 > State : clean > Active Devices : 4 > Working Devices : 4 > Failed Devices : 0 > Spare Devices : 0 > > Layout : left-symmetric > Chunk Size : 64K > > UUID : 6860a291:a5479bc6:e782da22:90dbd792 > Events : 0.45705 > > Number Major Minor RaidDevice State > 0 8 4 0 active sync /dev/sda4 > 1 8 20 1 active sync /dev/sdb4 > 2 8 36 2 active sync /dev/sdc4 > 3 8 52 3 active sync /dev/sdd4 > > > ....and here is the raid as it stands now. Note the end user I'm > helping tried to rebuild back on Sunday... > > % mdadm -D /dev/md2 > /dev/md2: > Version : 0.90 > Creation Time : Sun > Raid Level : raid5 > Array Size : 2925894144 (2790.35 GiB 2996.12 GB) > Used Dev Size : 975298048 (930.12 GiB 998.71 GB) > Raid Devices : 4 > Total Devices : 4 > Preferred Minor : 2 > Persistence : Superblock is persistent > > Update Time : Tue Nov 20 16:06:10 2012 > State : clean > Active Devices : 4 > Working Devices : 4 > Failed Devices : 0 > Spare Devices : 0 > > Layout : left-symmetric > Chunk Size : 64K > > UUID : 2ac5bacd:b40dc3f5:cb031839:58437670 > Events : 0.1 > > Number Major Minor RaidDevice State > 0 253 19 0 active sync /dev/dm-19 <<< > Note I am using cow devices via dmsetup > 1 253 11 1 active sync /dev/dm-11 > 2 253 15 2 active sync /dev/dm-15 > 3 253 7 3 active sync /dev/dm-7 > > Thank you for any and all help. > > Regards, > Stephen > -- > To unsubscribe from this list: send the line "unsubscribe linux-raid" in > the body of a message to majordomo@xxxxxxxxxxxxxxx > More majordomo info at http://vger.kernel.org/majordomo-info.html
Attachment:
signature.asc
Description: PGP signature