On Thu, Dec 22, 2011 at 10:45:06AM -0500, Vivek Goyal wrote: > On Thu, Dec 22, 2011 at 10:05:38AM +0900, Minchan Kim wrote: > > [..] > > > May be using deadline or noop in guest is better to benchmark against > > > PCI-E based flash. > > > > Good suggestion. > > I tested it by deadline on guest side. > > > > The result is not good. > > Although gap is within noise, Batch BIO's random performance is regressed > > compared to CFQ. > > > > Request Batch BIO > > > > (MB/s) stddev (MB/s) stddev > > w 787.030 31.494 w 748.714 68.490 > > rw 216.044 29.734 rw 216.977 40.635 > > r 771.765 3.327 r 771.107 4.299 > > rr 280.096 25.135 rr 258.067 43.916 > > > > I did some small test for only Batch BIO with deadline and cfq. > > to see I/O scheduler's effect. > > I think result is very strange, deadline :149MB, CFQ : 87M > > Because Batio BIO patch uses make_request_fn instead of request_rfn. > > So I think we should not affect by I/O scheduler.(I mean we issue I/O > > before I/O scheduler handles it) > > > > What do you think about it? > > Do I miss something? > > This indeed is very strange. In case of bio based drivers, changing IO > scheduler on the queue should not change anything. > > Trying running blktrace on the vda devie and see if you notice something > odd. > > Also you seem to be reporting contracdicting results for batch bio. > > Initially you mention that random IO is regressing with deadline as > comapred to CFQ. (It dropped from 325.976 MB/s to 258.067 MB/s). > > In this second test you are reporting that CFQ performs badly as > compared to deadline. (149MB/s vs 87MB/s). First test is with 64 thread with 512M and second test is only 1 thread with 128M for finishing quick. I think so result is very stange. This is data on test of deadline. In summary, data is very fluctuated. Sometime it's very stable but sometime data is very fluctuated like this. I don't know it's aio-stress problem or fusion I/O stuff problem 1) [root@RHEL-6 ~]# ./aio-stress -c 1 -t 1 -s 128 -r 8 -O -o 3 -d 512 /dev/vda num_thread 1 adding stage random read starting with random read file size 128MB, record size 8KB, depth 512, ios per iteration 8 max io_submit 8, buffer alignment set to 4KB threads 1 files 1 contexts 1 context offset 2MB verification off Running single thread version random read on /dev/vda (151.71 MB/s) 128.00 MB in 0.84s thread 0 random read totals (151.55 MB/s) 128.00 MB in 0.84s 2) [root@RHEL-6 ~]# ./aio-stress -c 1 -t 1 -s 128 -r 8 -O -o 3 -d 512 /dev/vda num_thread 1 adding stage random read starting with random read file size 128MB, record size 8KB, depth 512, ios per iteration 8 max io_submit 8, buffer alignment set to 4KB threads 1 files 1 contexts 1 context offset 2MB verification off Running single thread version random read on /dev/vda (201.04 MB/s) 128.00 MB in 0.64s thread 0 random read totals (200.76 MB/s) 128.00 MB in 0.64s 3) [root@RHEL-6 ~]# ./aio-stress -c 1 -t 1 -s 128 -r 8 -O -o 3 -d 512 /dev/vda num_thread 1 adding stage random read starting with random read file size 128MB, record size 8KB, depth 512, ios per iteration 8 max io_submit 8, buffer alignment set to 4KB threads 1 files 1 contexts 1 context offset 2MB verification off Running single thread version random read on /dev/vda (135.31 MB/s) 128.00 MB in 0.95s thread 0 random read totals (135.19 MB/s) 128.00 MB in 0.95s 4) [root@RHEL-6 ~]# ./aio-stress -c 1 -t 1 -s 128 -r 8 -O -o 3 -d 512 /dev/vda num_thread 1 adding stage random read starting with random read file size 128MB, record size 8KB, depth 512, ios per iteration 8 max io_submit 8, buffer alignment set to 4KB threads 1 files 1 contexts 1 context offset 2MB verification off Running single thread version random read on /dev/vda (116.93 MB/s) 128.00 MB in 1.09s thread 0 random read totals (116.82 MB/s) 128.00 MB in 1.10s 5) [root@RHEL-6 ~]# ./aio-stress -c 1 -t 1 -s 128 -r 8 -O -o 3 -d 512 /dev/vda num_thread 1 adding stage random read starting with random read file size 128MB, record size 8KB, depth 512, ios per iteration 8 max io_submit 8, buffer alignment set to 4KB threads 1 files 1 contexts 1 context offset 2MB verification off Running single thread version random read on /dev/vda (130.79 MB/s) 128.00 MB in 0.98s thread 0 random read totals (130.65 MB/s) 128.00 MB in 0.98s > > Two contradicting results? Hmm, I need test in /tmp/ to prevent it. > > Thanks > Vivek -- Kind regards, Minchan Kim -- To unsubscribe from this list: send the line "unsubscribe kvm" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html