Hi Haomai, Yes, let take a look into it and see what we can do. Cheng Cheng On Thu, Jan 15, 2015 at 12:57 PM, Haomai Wang <haomaiwang@xxxxxxxxx> wrote: > Hi Cheng cheng, > > There already exists a > bp(wiki.ceph.com/Planning/Blueprints/Giant/Add_QoS_capacity_to_librbd) > but still no one to implement it. > > If you are interested in this, I think we can make it father. If not, > I would like to pick it up later. :-) > > > On Fri, Jan 16, 2015 at 1:53 AM, Cheng Cheng <ccheng.leo@xxxxxxxxx> wrote: >> Hi Ceph, >> >> I am wondering is there a mechanism to prioritize the rbd_aio_write/rbd_aio_read I/Os? Currently all RBD I/Os are issued in FIFO to rados layer, and there is NO QoS mechanism to control the priority of these I/Os. >> >> A QoS mechanism will be beneficial when performing certain management operations, such as flatten. When flatten a image, the outstanding I/Os do get throttled by “rbd_concurrent_management_ops”. However this won't guarantee normal I/Os are not affected, as outstanding normal I/Os are still competing with concurrent management ops. >> >> Anyone know how/where to implement this QoS mechanism? >> >> Thanks! >> Cheng >> >> >> -- >> To unsubscribe from this list: send the line "unsubscribe ceph-devel" in >> the body of a message to majordomo@xxxxxxxxxxxxxxx >> More majordomo info at http://vger.kernel.org/majordomo-info.html > > > > -- > Best Regards, > > Wheat -- To unsubscribe from this list: send the line "unsubscribe ceph-devel" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html