On Thu, Jan 02, 2014 at 11:01:02AM -0500, tj@xxxxxxxxxx wrote: > Hello, Chris, Jan. > > On Thu, Jan 02, 2014 at 03:21:15PM +0000, Chris Mason wrote: > > On Thu, 2014-01-02 at 07:46 +0100, Jan Kara wrote: > > > In an ideal world you could compute writeback throughput for each memcg > > > (and writeback from a memcg would be accounted in a proper blkcg - we would > > > need unified memcg & blkcg hieararchy for that), take into account number of > > > dirty pages in each memcg, and compute dirty rate according to these two > > > numbers. But whether this can work in practice heavily depends on the memcg > > > size and how smooth / fair can the writeback from different memcgs be so > > > that we don't have excessive stalls and throughput estimation errors... > > > > [ Adding Tejun, Vivek and Li from another thread ] > > > > I do agree that a basket of knobs is confusing and it doesn't really > > help the admin. > > > > My first idea was a complex system where the controller in the block > > layer and the BDI flushers all communicated about current usage and > > cooperated on a single set of reader/writer rates. I think it could > > work, but it'll be fragile. > > One thing I do agree is that bdi would have to play some role. So is this a separate configuration which can be done per bdi as opposed to per device? IOW throttling offered per per cgroup per bdi. This will help with the case of throttling over NFS too, which some people have been asking for. > > > But there are a limited number of non-pagecache methods to do IO. Why > > not just push the accounting and throttling for O_DIRECT into a new BDI > > controller idea? Tejun was just telling me how he'd rather fix the > > existing controllers than add a new one, but I think we can have a much > > better admin experience by having a having a single entry point based on > > BDIs. > > But if we'll have to make bdis blkcg-aware, I think the better way to > do is splitting it per cgroup. That's what's being don in the lower > layer anyway. We split request queues to multiple queues according to > cgroup configuration. Things which can affect request issue and > completion, such as request allocation, are also split and each such > split queue is used for resource provisioning. > So is this a separate configuration which can be done per bdi as opposed to per device? IOW throttling offered per per cgroup per bdi. This will help with the case of throttling over NFS too, which some people have been asking for. So it sounds like re-implementing throttling infrastructure at bdi level now (Similar to what has been done at device level)? Of course use as much code as possible. But IIUC, proposal is that effectively there will can be two throttling controllers. One operating at bdi level and one operating below it at device level? And then writeback logic needs to be modified so that it can calculate per bdi per cgroup throughput (as opposed to bdi throughput only) and throttle writeback accordingly. Chris, in the past multiple implementations were proposed for a separate knob for writeback. A separate knob was frowned upon so I also proposed a separate implementation which throttle task in balance_dirty_pages() based on the write limit configured on device. https://lkml.org/lkml/2011/6/28/243 This approach assumed though that there is a block device associated. Building a back pressure from device/bdi and adjusting writeback accordingly (by making it cgroup aware) is complicated but once it works, I guess in long term might turn out to be a reasonable approach. Thanks Vivek -- To unsubscribe from this list: send the line "unsubscribe linux-fsdevel" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html