> The PCI bus is only capable of 133MB/s max. Unless you have dedicated > SATA ports, each on its own PCI-e bus, you will not get speeds in excess > of 133MB/s, 200MB/s+ I have read reports of someone using 4-5 SATA > controllers (SiI 3112 cards on PCI-e x1 ports) and they got around 200MB/s > or so on a RAID5, I assume that was read performance. Um, I *do* have dedicated SATA ports, each pair on its own PCIe bus. You might recall I wrote: >> This is 6 ST3400832AS 400 GB SATA drives, each capable of 60 MB/s >> sustained, on Sil3132 PCIe controllers with NCQ enabled. I measured >> > 300 MB/sec sustained aggregate off a temporary RAID-0 device during >> installation. I also included examples of reading 60 MB/s off 6 drives in parallel (360 MB/s aggregate), and reading off RAID-5 at 277 MB/s. To be specific, the narrowest bottleneck between drive and RAM is the 250 MB/s PCIe link shared by each pair of drives. To quote approximate one-way bandwidths: 1.5 Gb/s SATA 2.5 Gb/s PCIe Drive <--------> Sil3132 Drive <--------> Dual SATA <--------\ \ Drive <--------> Sil3132 \ 16x HyperTransport 2x DDR Drive <--------> Dual SATA <-------> nForce4 <-----> CPU <-----> RAM / 4000 MB/s 6400 MB/s Drive <--------> Sil3132 / Drive <--------> Dual SATA <--------/ 150 MB/s each 250 MB/s each - To unsubscribe from this list: send the line "unsubscribe linux-raid" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html