Re: mdadm: mdNN: pN size extends beyond EOD, truncated (cannot access imsm windows raid from within linux)

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



2012/12/4 Joshua C. <joshuacov@xxxxxxxxx>:
> I have two 1TB disk that have been running in raid1 with the
> intel-p67-built-in controller. Those were created in bios (uefi;
> imsm-raids) and are actually fake(software) raids. Windows sees only
> his own (ntfs) raid and everything is fine. Under linux I set the
> partition type to 0xfd (linux auto raid) and mdadm can successfully
> see and access the linux raid partion.
>
> However I cannot access my windows-raid-partition (/dev/md126p2 on
> /dev/sd[bc]) from within linux and get the following error: "md126: p2
> size 209711104 extends beyond EOD, truncated"
>
> Looking into the data, I see that my windows-raid-partition has
> actually 4! partitions on it. This should be wrong. This what
> mdadm-3.2.6 says about my disks under kernel-3.6.8.fc17.x86_64:
>
> ===========================================
> [root@localhost]# mdadm -E /dev/md127
> /dev/md127:
>           Magic : Intel Raid ISM Cfg Sig.
>         Version : 1.1.00
>     Orig Family : 88e429ad
>          Family : 1e27d1ba
>      Generation : 0001ad23
>      Attributes : All supported
>            UUID : 516f8fc8:bd5e663f:c254ea22:d30bb530
>        Checksum : 2438ebf9 correct
>     MPB Sectors : 1
>           Disks : 2
>    RAID Devices : 1
>
>   Disk00 Serial : WD-XXXXXXXXXXXX
>           State : active
>              Id : 00000002
>     Usable Size : 1953519880 (931.51 GiB 1000.20 GB)
>
> [Volume0]:
>            UUID : fa85ec01:25e9c177:da74f1e6:03284c35
>      RAID Level : 1
>         Members : 2
>           Slots : [UU]
>     Failed disk : none
>       This Slot : 0
>      Array Size : 1953519616 (931.51 GiB 1000.20 GB)
>    Per Dev Size : 1953519880 (931.51 GiB 1000.20 GB)
>   Sector Offset : 0
>     Num Stripes : 7630936
>      Chunk Size : 64 KiB
>        Reserved : 0
>   Migrate State : idle
>       Map State : normal
>     Dirty State : clean
>
>   Disk01 Serial : WD-XXXXXXXXXXXX
>           State : active
>              Id : 00000003
>     Usable Size : 1953519880 (931.51 GiB 1000.20 GB)
> ===========================================
> [root@localhost]# mdadm -D /dev/md127
> /dev/md127:
>         Version : imsm
>      Raid Level : container
>   Total Devices : 2
>
> Working Devices : 2
>
>
>            UUID : 516f8fc8:bd5e663f:c254ea22:d30bb530
>   Member Arrays : /dev/md/Volume0
>
>     Number   Major   Minor   RaidDevice
>
>        0       8       32        -        /dev/sdc
>        1       8       16        -        /dev/sdb
> ===========================================
> [root@localhost]# mdadm -E /dev/md126
> /dev/md126:
>    MBR Magic : aa55
> Partition[0] :   1743810560 sectors at         2048 (type fd)
> Partition[1] :    209711104 sectors at   1743812608 (type 07)
> ===========================================
> [root@localhost]# mdadm -D /dev/md126
> /dev/md126:
>       Container : /dev/md127, member 0
>      Raid Level : raid1
>      Array Size : 976759808 (931.51 GiB 1000.20 GB)
>   Used Dev Size : 976759940 (931.51 GiB 1000.20 GB)
>    Raid Devices : 2
>   Total Devices : 2
>
>           State : clean
>  Active Devices : 2
> Working Devices : 2
>  Failed Devices : 0
>   Spare Devices : 0
>
>
>            UUID : fa85ec01:25e9c177:da74f1e6:03284c35
>     Number   Major   Minor   RaidDevice State
>        1       8       16        0      active sync   /dev/sdb
>        0       8       32        1      active sync   /dev/sdc
> ===========================================
> [root@localhost]# mdadm -E /dev/md126p1
> ===========================================
> [root@localhost]# mdadm -D /dev/md126p1
> /dev/md126p1:
>       Container : /dev/md127, member 0
>      Raid Level : raid1
>      Array Size : 871905280 (831.51 GiB 892.83 GB)
>   Used Dev Size : 976759940 (931.51 GiB 1000.20 GB)
>    Raid Devices : 2
>   Total Devices : 2
>
>           State : clean
>  Active Devices : 2
> Working Devices : 2
>  Failed Devices : 0
>   Spare Devices : 0
>
>
>            UUID : fa85ec01:25e9c177:da74f1e6:03284c35
>     Number   Major   Minor   RaidDevice State
>        1       8       16        0      active sync   /dev/sdb
>        0       8       32        1      active sync   /dev/sdc
> ===========================================
> [root@localhost]# mdadm -E /dev/md126p2
> /dev/md126p2:
>    MBR Magic : aa55
> Partition[0] :   1917848077 sectors at      6579571 (type 70)
> Partition[1] :   1818575915 sectors at   1953251627 (type 43)
> Partition[2] :           10 sectors at    225735265 (type 72)
> Partition[3] :        51890 sectors at   2642411520 (type 00)
> ===========================================
> [root@localhost]# mdadm -D /dev/md126p2
> /dev/md126p2:
>       Container : /dev/md127, member 0
>      Raid Level : raid1
>      Array Size : 104853504 (100.00 GiB 107.37 GB)
>   Used Dev Size : 976759940 (931.51 GiB 1000.20 GB)
>    Raid Devices : 2
>   Total Devices : 2
>
>           State : clean
>  Active Devices : 2
> Working Devices : 2
>  Failed Devices : 0
>   Spare Devices : 0
>
>
>            UUID : fa85ec01:25e9c177:da74f1e6:03284c35
>     Number   Major   Minor   RaidDevice State
>        1       8       16        0      active sync   /dev/sdb
>        0       8       32        1      active sync   /dev/sdc
> ===========================================
> [root@localhost]# mdadm -E /dev/sdb
> /dev/sdb:
>           Magic : Intel Raid ISM Cfg Sig.
>         Version : 1.1.00
>     Orig Family : 88e429ad
>          Family : 1e27d1ba
>      Generation : 0001ad29
>      Attributes : All supported
>            UUID : 516f8fc8:bd5e663f:c254ea22:d30bb530
>        Checksum : 2438ebff correct
>     MPB Sectors : 1
>           Disks : 2
>    RAID Devices : 1
>
>   Disk00 Serial : WD-XXXXXXXXXXXX
>           State : active
>              Id : 00000002
>     Usable Size : 1953519880 (931.51 GiB 1000.20 GB)
>
> [Volume0]:
>            UUID : fa85ec01:25e9c177:da74f1e6:03284c35
>      RAID Level : 1
>         Members : 2
>           Slots : [UU]
>     Failed disk : none
>       This Slot : 0
>      Array Size : 1953519616 (931.51 GiB 1000.20 GB)
>    Per Dev Size : 1953519880 (931.51 GiB 1000.20 GB)
>   Sector Offset : 0
>     Num Stripes : 7630936
>      Chunk Size : 64 KiB
>        Reserved : 0
>   Migrate State : idle
>       Map State : normal
>     Dirty State : clean
>
>   Disk01 Serial : WD-XXXXXXXXXXXX
>           State : active
>              Id : 00000003
>     Usable Size : 1953519880 (931.51 GiB 1000.20 GB)
> ===========================================
> [root@localhost]# mdadm -E /dev/sdc
> /dev/sdc:
>           Magic : Intel Raid ISM Cfg Sig.
>         Version : 1.1.00
>     Orig Family : 88e429ad
>          Family : 1e27d1ba
>      Generation : 0001ad29
>      Attributes : All supported
>            UUID : 516f8fc8:bd5e663f:c254ea22:d30bb530
>        Checksum : 2438ebff correct
>     MPB Sectors : 1
>           Disks : 2
>    RAID Devices : 1
>
>   Disk01 Serial : WD-XXXXXXXXXXXX
>           State : active
>              Id : 00000003
>     Usable Size : 1953519880 (931.51 GiB 1000.20 GB)
>
> [Volume0]:
>            UUID : fa85ec01:25e9c177:da74f1e6:03284c35
>      RAID Level : 1
>         Members : 2
>           Slots : [UU]
>     Failed disk : none
>       This Slot : 1
>      Array Size : 1953519616 (931.51 GiB 1000.20 GB)
>    Per Dev Size : 1953519880 (931.51 GiB 1000.20 GB)
>   Sector Offset : 0
>     Num Stripes : 7630936
>      Chunk Size : 64 KiB
>        Reserved : 0
>   Migrate State : idle
>       Map State : normal
>     Dirty State : clean
>
>   Disk00 Serial : WD-XXXXXXXXXXXX
>           State : active
>              Id : 00000002
>     Usable Size : 1953519880 (931.51 GiB 1000.20 GB)
> ===========================================
> [root@localhost]# cat /proc/mdstat
> Personalities : [raid1]
> md126 : active raid1 sdb[1] sdc[0]
>       976759808 blocks super external:/md127/0 [2/2] [UU]
>
> md127 : inactive sdb[1](S) sdc[0](S)
>       5288 blocks super external:imsm
>
> unused devices: <none>
> ===========================================
> [root@localhost]# cat /proc/partitions
> major minor  #blocks  name
>
>    7        0         16 loop0
>    7        1       2796 loop1
>    7        2     900352 loop2
>    7        3   10485760 loop3
>    7        4     524288 loop4
>    8        0   62522712 sda
>    8        1   18350080 sda1
>    8        2   44171264 sda2
>    8       16  976762584 sdb
>    8       32  976762584 sdc
>   11        0    1048575 sr0
>    8       48    3903488 sdd
>    8       49    3903456 sdd1
>  253        0   10485760 dm-0
>  253        1   10485760 dm-1
>    9      126  976759808 md126
>  259        0  871905280 md126p1
>  259        1  104853504 md126p2
>    8       64    2002943 sde
>    8       65    2002942 sde1
> ===========================================
> [root@localhost]# mdadm --examine --scan
> ARRAY metadata=imsm UUID=516f8fc8:bd5e663f:c254ea22:d30bb530
> ARRAY /dev/md/Volume0 container=516f8fc8:bd5e663f:c254ea22:d30bb530
> member=0 UUID=fa85ec01:25e9c177:da74f1e6:03284c35
> ===========================================
> [root@localhost]# mdadm --detail --scan
> ARRAY /dev/md127 metadata=imsm UUID=516f8fc8:bd5e663f:c254ea22:d30bb530
> ARRAY /dev/md/Volume0 container=/dev/md127 member=0
> UUID=fa85ec01:25e9c177:da74f1e6:03284c35
> ===========================================
> [root@localhost]# dmesg  |grep md
> [    0.000000] Command line: initrd=initrd0.img
> root=live:UUID=406D-2540 rootfstype=vfat ro liveimg quiet  rhgb
> rd.luks=0 rd.md=0 rd.dm=0  BOOT_IMAGE=vmlinuz0
> [    0.000000] Kernel command line: initrd=initrd0.img
> root=live:UUID=406D-2540 rootfstype=vfat ro liveimg quiet  rhgb
> rd.luks=0 rd.md=0 rd.dm=0  BOOT_IMAGE=vmlinuz0
> [    1.277784] dracut: rd.md=0: removing MD RAID activation
> [    6.231138] md: bind<sdb>
> [    6.434862] md: bind<sdc>
> [    6.440260] md: bind<sdc>
> [    6.440383] md: bind<sdb>
> [    6.453438] md: raid1 personality registered for level 1
> [    6.453654] md/raid1:md126: active with 2 out of 2 mirrors
> [    6.453678] md126: detected capacity change from 0 to 1000202043392
> [    6.460137]  md126: p1 p2
> [    6.460268] md126: p2 size 209711104 extends beyond EOD, truncated
> [    6.464890] md: md126 switched to read-write mode.
> [    6.675396] md: export_rdev(sdc)
> [    6.675454] md: export_rdev(sdb)
> [   84.923953] md: export_rdev(sdc)
> [   84.924017] md: export_rdev(sdb)
> [   93.473151] md126: detected capacity change from 1000202043392 to 0
> [   93.473228] md: md126 stopped.
> [   93.473236] md: unbind<sdb>
> [   93.479847] md: export_rdev(sdb)
> [   93.479858] md: unbind<sdc>
> [   93.486826] md: export_rdev(sdc)
> [   93.493682] md: md126 stopped.
> [   95.923141] md: md127 stopped.
> [   95.923151] md: unbind<sdc>
> [   95.927344] md: export_rdev(sdc)
> [   95.927504] md: unbind<sdb>
> [   95.936313] md: export_rdev(sdb)
> [  824.206594] md: md127 stopped.
> [  824.210398] md: bind<sdc>
> [  824.210840] md: bind<sdb>
> [  824.241537] md: bind<sdc>
> [  824.241612] md: bind<sdb>
> [  824.245817] md/raid1:md126: active with 2 out of 2 mirrors
> [  824.245829] md126: detected capacity change from 0 to 1000202043392
> [  824.247377]  md126: p1 p2
> [  824.247545] md126: p2 size 209711104 extends beyond EOD, truncated
> [  824.261049] md: md126 switched to read-write mode.
> [  824.304567] md: export_rdev(sdc)
> [  824.304594] md: export_rdev(sdb)
> [  824.320417] md: export_rdev(sdc)
> [  824.320444] md: export_rdev(sdb)
> [ 1161.533446] EXT4-fs (md126p1): warning: maximal mount count
> reached, running e2fsck is recommended
> [ 1161.573164] EXT4-fs (md126p1): mounted filesystem with ordered data
> mode. Opts: (null)
> [ 1161.573169] SELinux: initialized (dev md126p1, type ext4), uses xattr
> ===========================================
>
> Can someone take a look at why mdadm cannot access the windows-raid-partition?
>
>
> --joshua

Anyone? Any idea why I cannot access my windows raid from linux?


-- 
--joshua
--
To unsubscribe from this list: send the line "unsubscribe linux-raid" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at  http://vger.kernel.org/majordomo-info.html


[Index of Archives]     [Linux RAID Wiki]     [ATA RAID]     [Linux SCSI Target Infrastructure]     [Linux Block]     [Linux IDE]     [Linux SCSI]     [Linux Hams]     [Device Mapper]     [Device Mapper Cryptographics]     [Kernel]     [Linux Admin]     [Linux Net]     [GFS]     [RPM]     [git]     [Yosemite Forum]


  Powered by Linux