I have two 1tb disk that have been running in raid1 with the intel-p67-built-in controller. Those were created in bios (uefi) and are actually fake(software) raids. Windows sees only his own (ntfs) raid and everything is fine. Under linux I set the partition type to 0xfd (linux auto raid) and mdadm can successfully see and access the linux raid partion. However mdadm cannot access the windows raid and I have each partition three(!) times under KDE-Dolphin. So each time I click (unintentionally) on the real hdd-partition instead of the mdadm-partition I trigger the check of the raid, which takes about 3 hours. How to make mdadm access the windows raid and ONLY show the raid partitions in the kde-dolphin??? When I omit "rd.dm=0 and rd.md=0" on the command line then I see only the raid partitions under kde-dolphin but for some reason I cannot access them (so this is not a solution). Here is what mdadm says about my disks (done with kernel-3.6.8.fc17.x86_64): ============================= [root@localhost]# cat /proc/mdstat Personalities : [raid1] md126 : active raid1 sdb[1] sdc[0] 976759808 blocks super external:/md127/0 [2/2] [UU] md127 : inactive sdb[1](S) sdc[0](S) 5288 blocks super external:imsm unused devices: <none> ============================= [root@localhost]# mdadm --examine /dev/md126 /dev/md126: MBR Magic : aa55 Partition[0] : 1743810560 sectors at 2048 (type fd) Partition[1] : 209711104 sectors at 1743812608 (type 07) ============================= [root@localhost]# mdadm --examine /dev/md127 /dev/md127: Magic : Intel Raid ISM Cfg Sig. Version : 1.1.00 Orig Family : 88e429ad Family : 1e27d1ba Generation : 0001acff Attributes : All supported UUID : 516f8fc8:bd5e663f:c254ea22:d30bb530 Checksum : 2438ebd3 correct MPB Sectors : 1 Disks : 2 RAID Devices : 1 Disk00 Serial : WD-xxxxxxxxxxxx State : active Id : 00000002 Usable Size : 1953519880 (931.51 GiB 1000.20 GB) [Volume0]: UUID : fa85ec01:25e9c177:da74f1e6:03284c35 RAID Level : 1 Members : 2 Slots : [UU] Failed disk : none This Slot : 0 Array Size : 1953519616 (931.51 GiB 1000.20 GB) Per Dev Size : 1953519880 (931.51 GiB 1000.20 GB) Sector Offset : 0 Num Stripes : 7630936 Chunk Size : 64 KiB Reserved : 0 Migrate State : idle Map State : normal Dirty State : clean Disk01 Serial : WD-xxxxxxxxxxxx State : active Id : 00000003 Usable Size : 1953519880 (931.51 GiB 1000.20 GB) ============================= [root@localhost]# cat /proc/partitions major minor #blocks name 7 0 16 loop0 7 1 2796 loop1 7 2 900352 loop2 7 3 10485760 loop3 7 4 524288 loop4 8 0 62522712 sda 8 1 18350080 sda1 8 2 44171264 sda2 8 16 976762584 sdb 8 17 871905280 sdb1 8 18 104855552 sdb2 8 32 976762584 sdc 8 33 871905280 sdc1 8 34 104855552 sdc2 11 0 1048575 sr0 8 48 3903488 sdd 8 49 3903456 sdd1 253 0 10485760 dm-0 253 1 10485760 dm-1 9 126 976759808 md126 259 0 871905280 md126p1 259 1 104853504 md126p2 ============================= [root@localhost]# mdadm --examine /dev/md126p2 /dev/md126p2: MBR Magic : aa55 Partition[0] : 1917848077 sectors at 6579571 (type 70) Partition[1] : 1818575915 sectors at 1953251627 (type 43) Partition[2] : 10 sectors at 225735265 (type 72) Partition[3] : 51890 sectors at 2642411520 (type 00) ============================= [root@localhost]# dmesg | grep md [ 5.966318] md: bind<sdc> [ 6.168403] md: bind<sdb> [ 6.171261] md: bind<sdc> [ 6.171463] md: bind<sdb> [ 6.213554] md: raid1 personality registered for level 1 [ 6.215861] md/raid1:md126: active with 2 out of 2 mirrors [ 6.216006] md126: detected capacity change from 0 to 1000202043392 [ 6.216775] md126: p1 p2 [ 6.216955] md126: p2 size 209711104 extends beyond EOD, truncated [ 6.226735] md: md126 switched to read-write mode. [ 6.436537] md: export_rdev(sdc) [ 6.436733] md: export_rdev(sdb) [ 227.889293] EXT4-fs (md126p1): warning: maximal mount count reached, running e2fsck is recommended [ 227.929448] EXT4-fs (md126p1): mounted filesystem with ordered data mode. Opts: (null) [ 227.929454] SELinux: initialized (dev md126p1, type ext4), uses xattr ============================= [root@localhost]# mdadm --detail --scan ARRAY /dev/md/imsm0 metadata=imsm UUID=516f8fc8:bd5e663f:c254ea22:d30bb530 ARRAY /dev/md/Volume0_0 container=/dev/md/imsm0 member=0 UUID=fa85ec01:25e9c177:da74f1e6:03284c35 ============================= --joshua -- users mailing list users@xxxxxxxxxxxxxxxxxxxxxxx To unsubscribe or change subscription options: https://admin.fedoraproject.org/mailman/listinfo/users Guidelines: http://fedoraproject.org/wiki/Mailing_list_guidelines Have a question? Ask away: http://ask.fedoraproject.org