On 8/29/2013 4:20 AM, Albert Pauw wrote: ... > OS: Oracle Linux 5.9 (effectively RHEL 5.9), kernel 2.6.32-400.29.2.el5uek. > All utilities updates, mdadm (2.6.9 latest through updates). ... > Two Fusion IO Duo cards, each Fusion IO device 640 GB, so four in total. ... > mdadm --create --verbose /dev/md0 --level=10 --metadata=1.2 > --chunk=512 --raid-devices=4 /dev/fioa /dev/fioc /dev/fiob /dev/fiod > --assume-clean -N md0 > > When the performance turned out bad, after about 20 minutes, the > process was stopped. I broke the mirror, so the md0 device is only > striped, but the performance hit after 20 minutes happened again. > > The status of all cards are fine, no problems there. Then I created a > fs on only one device and have it run again. This time it worked fine. > The fs was in all cases ext3, no TRIM. You've presented insufficient information to allow a definitive answer. That said, it's very likely that you're hitting the same wall many folks do with SSDs. All md/RAID personalities are limited to a single write thread which limits you to one CPU of IO throughput. When writing to a single device without md/RAID, block IOs can be processed by all CPUs in parallel. The Fusion IO device is likely sufficiently fast that a single md/RAID10 thread can't saturate the device, so you run out of CPU before IOPS. This is very common with SSD and md/RAID. Shaohua Li has been busily working on patches for quite some time now to eliminate this CPU bottleneck in md. The fact that a single Fusion IO device with EXT3 on it is faster than md/RAID10 strongly suggests this may be the cause. If you have multiple application threads or processes writing to a single device the IOs will be processed on the same CPU (core) as the thread, so you can have IOs in flight from all CPUs in parallel. When using md/RAID all of that IO must be shuttled to the md driver which can only execute on a single CPU (core). To verify this, simply run your tests again and monitor CPU burn of the md/RAID10 thread. If that CPU is 100% at any time then this is the problem. If this is true, you can immediately mitigate it by using a layered md/RAID0 over md/RAID1 setup. Doing this will give you two md/RAID1 write threads, doubling the number of CPU cores you can put into play. To do this and maintain the card<->card mirror layout you described, you will create an md/RAID1 with fioa and fioc, and another md/RAID1 with fiob and fiod. Then you'll create an md/RAID0 across these two md/RAID1 devices. The md/RAID0 and linear personalities don't use write threads and are thus not limited to a single CPU core. One final suggestion. Use XFS instead of EXT3/4. You should get significantly better performance with a parallel database workload. But I'd strongly suggest moving up to a RHEL 6.2+ clone if you do. 5.9 is ancient, and there are tons of performance and stability enhancements in newer kernels, specifically related to XFS. -- Stan -- To unsubscribe from this list: send the line "unsubscribe linux-raid" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html