Hi Blairo,
> fio shows 70 MB/s seq read with 4M blocks, libaio, 1 thread, direct.
> fio seq write 200 MB/s
The fio numbers are from fio running on a CephFS mount I take it?
exactly.
> # rados bench -t 1 -p test 60 write --no-cleanup
I don't see an rbd test anywhere here...?
I suggest comparing fio on CephFS with fio on rbd (as in using fio's
rbd ioengine), then at least the application side of your tests is
constant.
I tested four access types:
1. rbd kernel module, xfs, fio-libaio
1. rbd kernel module, xfs, fio-libaio
2. rados bench seq
3. fio-librbdDone tests, no big diference, all methods give 50 to 80 MB/s single threaded bandwidth.
fio-librbd having 1.5 times lower bandwidth (maybe because it runs in userspace?)
What can be tuned to improve sequential read?
readahead settings are:
"client_readahead_min": "131072",
"client_readahead_max_bytes": "2097152",
"client_readahead_max_periods": "4",
"rbd_readahead_trigger_requests": "10",
"rbd_readahead_max_bytes": "524288",
"rbd_readahead_disable_after_bytes": "52428800",
--
WBR,
Ilja.
WBR,
Ilja.
_______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com