On 8 May 2011 17:09, Arnd Bergmann <arnd@xxxxxxxx> wrote: > On Saturday 07 May 2011, Per Forlin wrote: >> > The mmc queue never runs empty until end of transfer.. The requests >> > are 128 blocks (64k limit set in mmc host driver) compared to 256 >> > blocks before. This will not improve performance much since the >> > transfer now are smaller than before. The latency is minimal but >> > instead there extra number of transfer cause more mmc cmd overhead. >> > I added prints to print the wait time in lock_page_killable too. >> > I wonder if I can achieve a none empty mmc block queue without >> > compromising the mmc host driver performance. >> > >> There is actually a performance increase from 16.5 MB/s to 18.4 MB/s >> when lowering the max_req_size to 64k. >> I run a dd test on a pandaboard using 2.6.39-rc5 kernel. > > I've noticed with a number of cards that using 64k writes is faster > than any other size. What I could not figure out yet is whether this > is a common hardware optimization for MS Windows (which always uses > 64K I/O when it can), or if it's a software effect and we can actually > make it go faster with Linux by tuning for other sizes. > Thanks for the tip I will keep that in mind. In this case the increase in performance is due to parallel cache handling. I did a test and set the mmc_max_req to 128k (same size as the first test with low performance) and increase the read_ahead to 256k. root@(none):/ echo 256 > sys/devices/platform/omap/omap_hsmmc.0/mmc_host/mmc0/mmc0:80ca/block/mmcblk0/queue/read_ahead_kb root@(none):/ dd if=/dev/mmcblk0p3 of=/dev/null bs=4k count=25600 25600+0 records in 25600+0 records out 104857600 bytes (100.0MB) copied, 5.138585 seconds, 19.5MB/s Regards, Per -- To unsubscribe from this list: send the line "unsubscribe linux-mmc" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html