Re: RAID5 NAS Recovery...00.90.01 vs 00.90

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Tue, 20 Nov 2012 15:41:57 -0500 Stephen Haran <steveharan@xxxxxxxxx> wrote:

> Hi, I'm trying to recover a Western Digital Share Space NAS.
> I'm able to assemble the RAID5 and restore the LVM but it can't see
> any filesystem.
> 
> Below is a raid.log file that shows how the raid was configured when
> it was working.
> And also the output of mdadm -D showing the raid in it's current state.
> Note the Version difference 00.90.01 vs. 0.90. And the array size
> difference 2925293760  vs. 2925894144
> I'm thinking this difference may be the reason Linux can not see a filesystem.

Probably not - losing a few blocks from the end might make 'fsck' complain,
but it should still be able to see the filesystem.

How did you test if you could see a filesystem?  'mount' or 'fsck -n' ?

It looks like you re-created the array recently (Nov 18 12:07:53 2012)  Why
did you do that?
It has been created slightly smaller - not sure why.  Maybe if you explicitly
request the old per-device size with "--size=975097920" it might get it right.

Are you sure the dm cow devices show exactly the same size and content as the
originals?

The stray '.01' at the end of the version number is not relevant.  It just
indicates a different version of mdadm in use to report the array.

NeilBrown


> 
> My question is would the version difference explain the array size difference?
> And is it possible to create a version 00.90.01 array? I do not see
> that in the mdadm docs.
> 
> ....original working raid config....
> /dev/md2:
>         Version : 00.90.01
>   Creation Time : Wed Jun 24 19:00:59 2009
>      Raid Level : raid5
>      Array Size : 2925293760 (2789.78 GiB 2995.50 GB)
>     Device Size : 975097920 (929.93 GiB 998.50 GB)
>    Raid Devices : 4
>   Total Devices : 4
> Preferred Minor : 2
>     Persistence : Superblock is persistent
> 
>     Update Time : Thu Jun 25 02:36:31 2009
>           State : clean
>  Active Devices : 4
> Working Devices : 4
>  Failed Devices : 0
>   Spare Devices : 0
> 
>          Layout : left-symmetric
>      Chunk Size : 64K
> 
>            UUID : 6860a291:a5479bc6:e782da22:90dbd792
>          Events : 0.45705
> 
>     Number   Major   Minor   RaidDevice State
>        0       8        4        0      active sync   /dev/sda4
>        1       8       20        1      active sync   /dev/sdb4
>        2       8       36        2      active sync   /dev/sdc4
>        3       8       52        3      active sync   /dev/sdd4
> 
> 
> ....and here is the raid as it stands now. Note the end user I'm
> helping tried to rebuild back on Sunday...
> 
>  % mdadm -D /dev/md2
> /dev/md2:
>         Version : 0.90
>   Creation Time : Sun 
>      Raid Level : raid5
>      Array Size : 2925894144 (2790.35 GiB 2996.12 GB)
>   Used Dev Size : 975298048 (930.12 GiB 998.71 GB)
>    Raid Devices : 4
>   Total Devices : 4
> Preferred Minor : 2
>     Persistence : Superblock is persistent
> 
>     Update Time : Tue Nov 20 16:06:10 2012
>           State : clean
>  Active Devices : 4
> Working Devices : 4
>  Failed Devices : 0
>   Spare Devices : 0
> 
>          Layout : left-symmetric
>      Chunk Size : 64K
> 
>            UUID : 2ac5bacd:b40dc3f5:cb031839:58437670
>          Events : 0.1
> 
>     Number   Major   Minor   RaidDevice State
>        0     253       19        0      active sync   /dev/dm-19  <<<
> Note I am using cow devices via dmsetup
>        1     253       11        1      active sync   /dev/dm-11
>        2     253       15        2      active sync   /dev/dm-15
>        3     253        7        3      active sync   /dev/dm-7
> 
> Thank you for any and all help.
> 
> Regards,
> Stephen
> --
> To unsubscribe from this list: send the line "unsubscribe linux-raid" in
> the body of a message to majordomo@xxxxxxxxxxxxxxx
> More majordomo info at  http://vger.kernel.org/majordomo-info.html

Attachment: signature.asc
Description: PGP signature


[Index of Archives]     [Linux RAID Wiki]     [ATA RAID]     [Linux SCSI Target Infrastructure]     [Linux Block]     [Linux IDE]     [Linux SCSI]     [Linux Hams]     [Device Mapper]     [Device Mapper Cryptographics]     [Kernel]     [Linux Admin]     [Linux Net]     [GFS]     [RPM]     [git]     [Yosemite Forum]


  Powered by Linux