On Thu, Jan 1, 2015 at 8:18 AM, Sedat Dilek <sedat.dilek@xxxxxxxxx> wrote: > On Thu, Jan 1, 2015 at 1:01 AM, Ming Lei <tom.leiming@xxxxxxxxx> wrote: >> Hi Sedat, >> >> On Thu, Jan 1, 2015 at 6:32 AM, Sedat Dilek <sedat.dilek@xxxxxxxxx> wrote: >>> Forgot to CC LKML and linux-fsdevel. >>> >>> - Sedat - >> >>> >>> OK, I have installed fio (1.59-1) and libaio1 (0.3.109-2ubuntu1) here. >>> >>> You say in [1]: >>> >>> "In the following test: >>> - base: v3.19-rc2-2041231 >>> - loop over file in ext4 file system on SSD disk >>> - bs: 4k, libaio, io depth: 64, O_DIRECT, num of jobs: 1 >>> - throughput: IOPS" >>> >>> I tried to reproduce that inspired by [2]... >>> >>> root# fio --name=randread --rw=randread --bs=4k --ioengine=libaio >>> --iodepth=64 --direct=1 --numjobs=1 --size=1G >>> >>> ...you had no size given (here: 1GiB) - fio requires that parameter to run. >>> >>> This results in 165 VS. 515 IOPS here. >> >> Thanks for your test. >> >> Also if your disk is quick enough, you will observe improvement on >> read test too. >> > > This is no SSD here. > > # dmesg | egrep -i 'hitachi|ata1|sda' > [ 0.457892] ata1: SATA max UDMA/133 abar m2048@0xf0708000 port > 0xf0708100 irq 25 > [ 0.777445] ata1: SATA link up 3.0 Gbps (SStatus 123 SControl 300) > [ 0.778759] ata1.00: ATA-8: Hitachi HTS545050A7E380, GG2OA6C0, max UDMA/133 > [ 0.778778] ata1.00: 976773168 sectors, multi 16: LBA48 NCQ (depth 31/32), AA > [ 0.780154] ata1.00: configured for UDMA/133 > [ 0.780970] scsi 0:0:0:0: Direct-Access ATA Hitachi > HTS54505 A6C0 PQ: 0 ANSI: 5 > [ 0.782050] sd 0:0:0:0: [sda] 976773168 512-byte logical blocks: > (500 GB/465 GiB) > [ 0.782058] sd 0:0:0:0: [sda] 4096-byte physical blocks > [ 0.782255] sd 0:0:0:0: [sda] Write Protect is off > [ 0.782262] sd 0:0:0:0: [sda] Mode Sense: 00 3a 00 00 > [ 0.782339] sd 0:0:0:0: [sda] Write cache: enabled, read cache: > enabled, doesn't support DPO or FUA > [ 0.800644] sda: sda1 sda2 sda3 > [ 0.802029] sd 0:0:0:0: [sda] Attached SCSI disk > > How did you test with fio (your fio lines)? Your fio command line is basically same with my fio config, and you can attach one image to loop via: losetup -f file_name. Looks your randread result is good, and I can observe ~80 IOPS vs. ~200 IOPS on my slow HDD. in the randread test too. #################fio config########################## [global] direct=1 size=128G bsrange=4k-4k timeout=30 numjobs=1 ioengine=libaio iodepth=64 filename=/dev/loop0 group_reporting=1 [f] rw=${RW} Thanks, Ming Lei -- To unsubscribe from this list: send the line "unsubscribe linux-fsdevel" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html