Hi all,
We're about to go live with some qemu rate limiting to RBD, and I wanted to crosscheck our values with this list, in case someone can chime in with their experience or known best practices.
The only reasonable, non test-suite, values I found on the web are:
iops_wr 200
iops_rd 400
bps_wr 40000000
bps_rd 80000000
and those seem (to me) to offer a "pretty good" service level, with more iops than a typical disk yet lower throughput (which is good considering our single gigabit NICs on the hypervisors).
Our main goal for the rate limiting is to protect the cluster from abusive users running fio, etc., while not overly restricting our varied legitimate applications.
Any opinions here?
Cheers, Dan
_______________________________________________ ceph-users mailing list ceph-users@xxxxxxxxxxxxxx http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com