On 01/10/2013 09:23 AM, chris wrote:
Hello, I have a machine which was running a imsm raid volume, where the motherboard failed and I do not have access to another system with imsm. I remember noticing some time ago that mdadm could recognize these arrays, so I decided to try recovery in a spare machine with the disks from the array. I guess my questions are: Is this the right forum for help with this? Am I even going down a feasible path here or is this array dependent on the HBA in some way? If it is possible any ideas of anything else I can do to debug this further?
Typically mdadm probes the OROM and look for platform details. But you can try overriding with:
export IMSM_NO_PLATFORM=1 See if that works for you.
The original array was a raid 5 of 4x2TB sata disks When I examine the first disk, things look good: mdadm --examine /dev/sdb mdadm: /dev/sdb is not attached to Intel(R) RAID controller. mdadm: /dev/sdb is not attached to Intel(R) RAID controller. /dev/sdb: Magic : Intel Raid ISM Cfg Sig. Version : 1.3.00 Orig Family : 226cc5df Family : 226cc5df Generation : 000019dc Attributes : All supported UUID : b67ea029:aaea7da2:2540c1eb:ebe98af1 Checksum : 651263bf correct MPB Sectors : 2 Disks : 4 RAID Devices : 1 Disk02 Serial : Z1E1RPA9 State : active Id : 00030000 Usable Size : 3907024648 (1863.01 GiB 2000.40 GB) [Volume0]: UUID : 51a415ba:dc9c8cd7:5b3ea8de:465b4630 RAID Level : 5 Members : 4 Slots : [__U_] Failed disk : 0 This Slot : 2 Array Size : 11721072640 (5589.04 GiB 6001.19 GB) Per Dev Size : 3907024648 (1863.01 GiB 2000.40 GB) Sector Offset : 0 Num Stripes : 15261814 Chunk Size : 128 KiB Reserved : 0 Migrate State : idle Map State : failed Dirty State : clean Disk00 Serial : Z1E1AKPH:0 State : active Id : ffffffff Usable Size : 3907024648 (1863.01 GiB 2000.40 GB) Disk01 Serial : Z24091Q5:0 State : active Id : ffffffff Usable Size : 3907024648 (1863.01 GiB 2000.40 GB) Disk03 Serial : Z1E19E4K:0 State : active Id : ffffffff Usable Size : 3907024648 (1863.01 GiB 2000.40 GB) When I try to scan for arrays I get this: # mdadm --examine --scan HBAs of devices does not match (null) != (null) ARRAY metadata=imsm UUID=b67ea029:aaea7da2:2540c1eb:ebe98af1 ARRAY /dev/md/Volume0 container=b67ea029:aaea7da2:2540c1eb:ebe98af1 member=0 UUID=51a415ba:dc9c8cd7:5b3ea8de:465b4630 ARRAY metadata=imsm UUID=b67ea029:aaea7da2:2540c1eb:ebe98af1 ARRAY /dev/md/Volume0 container=b67ea029:aaea7da2:2540c1eb:ebe98af1 member=0 UUID=51a415ba:dc9c8cd7:5b3ea8de:465b4630 My first concern is the warning that the HBA is missing, the whole reason I am going at it this way is because I don't have the HBA. Second concern is duplicate detection of the same array. If i try to run # mdadm -As: mdadm: No arrays found in config file or automatically I also tried adding the output from --examine --scan to /etc/mdadm/mdadm.conf but after trying that I now get blank output: # mdadm --assemble /dev/md/Volume0 # # mdadm --assemble --scan # full examine of all disks involved: /dev/sdb: Magic : Intel Raid ISM Cfg Sig. Version : 1.3.00 Orig Family : 226cc5df Family : 226cc5df Generation : 000019dc Attributes : All supported UUID : b67ea029:aaea7da2:2540c1eb:ebe98af1 Checksum : 651263bf correct MPB Sectors : 2 Disks : 4 RAID Devices : 1 Disk02 Serial : Z1E1RPA9 State : active Id : 00030000 Usable Size : 3907024648 (1863.01 GiB 2000.40 GB) [Volume0]: UUID : 51a415ba:dc9c8cd7:5b3ea8de:465b4630 RAID Level : 5 Members : 4 Slots : [__U_] Failed disk : 0 This Slot : 2 Array Size : 11721072640 (5589.04 GiB 6001.19 GB) Per Dev Size : 3907024648 (1863.01 GiB 2000.40 GB) Sector Offset : 0 Num Stripes : 15261814 Chunk Size : 128 KiB Reserved : 0 Migrate State : idle Map State : failed Dirty State : clean Disk00 Serial : Z1E1AKPH:0 State : active Id : ffffffff Usable Size : 3907024648 (1863.01 GiB 2000.40 GB) Disk01 Serial : Z24091Q5:0 State : active Id : ffffffff Usable Size : 3907024648 (1863.01 GiB 2000.40 GB) Disk03 Serial : Z1E19E4K:0 State : active Id : ffffffff Usable Size : 3907024648 (1863.01 GiB 2000.40 GB) /dev/sdc: MBR Magic : aa55 Partition[0] : 3907027057 sectors at 63 (type 42) /dev/sdd: Magic : Intel Raid ISM Cfg Sig. Version : 1.3.00 Orig Family : 226cc5df Family : 226cc5df Generation : 000019d9 Attributes : All supported UUID : b67ea029:aaea7da2:2540c1eb:ebe98af1 Checksum : 641438ba correct MPB Sectors : 2 Disks : 4 RAID Devices : 1 Disk03 Serial : Z1E19E4K State : active Id : 00020000 Usable Size : 3907024648 (1863.01 GiB 2000.40 GB) [Volume0]: UUID : 51a415ba:dc9c8cd7:5b3ea8de:465b4630 RAID Level : 5 Members : 4 Slots : [__UU] Failed disk : 0 This Slot : 3 Array Size : 11721072640 (5589.04 GiB 6001.19 GB) Per Dev Size : 3907024648 (1863.01 GiB 2000.40 GB) Sector Offset : 0 Num Stripes : 15261814 Chunk Size : 128 KiB Reserved : 0 Migrate State : idle Map State : failed Dirty State : clean Disk00 Serial : Z1E1AKPH:0 State : active Id : ffffffff Usable Size : 3907024648 (1863.01 GiB 2000.40 GB) Disk01 Serial : Z24091Q5:0 State : active Id : ffffffff Usable Size : 3907024648 (1863.01 GiB 2000.40 GB) Disk02 Serial : Z1E1RPA9 State : active Id : 00030000 Usable Size : 3907024648 (1863.01 GiB 2000.40 GB) /dev/sde: MBR Magic : aa55 Partition[0] : 4294967295 sectors at 1 (type ee) # dpkg -l | grep mdadm ii mdadm 3.2.5-1+b1 thanks chris -- To unsubscribe from this list: send the line "unsubscribe linux-raid" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html
-- To unsubscribe from this list: send the line "unsubscribe linux-raid" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html