On Tue, Aug 09, 2011 at 10:54:38PM +0800, Vivek Goyal wrote: > On Sat, Aug 06, 2011 at 04:44:50PM +0800, Wu Fengguang wrote: > > It's all about bdi->dirty_ratelimit, which aims to be (write_bw / N) > > when there are N dd tasks. > > > > On write() syscall, use bdi->dirty_ratelimit > > ============================================ > > > > balance_dirty_pages(pages_dirtied) > > { > > pos_bw = bdi->dirty_ratelimit * bdi_position_ratio(); > > pause = pages_dirtied / pos_bw; > > sleep(pause); > > } > > > > On every 200ms, update bdi->dirty_ratelimit > > =========================================== > > > > bdi_update_dirty_ratelimit() > > { > > bw = bdi->dirty_ratelimit; > > ref_bw = bw * bdi_position_ratio() * write_bw / dirty_bw; > > if (dirty pages unbalanced) > > bdi->dirty_ratelimit = (bw * 3 + ref_bw) / 4; > > } > > > > Estimation of balanced bdi->dirty_ratelimit > > =========================================== > > > > When started N dd, throttle each dd at > > > > task_ratelimit = pos_bw (any non-zero initial value is OK) > > > > After 200ms, we got > > > > dirty_bw = # of pages dirtied by app / 200ms > > write_bw = # of pages written to disk / 200ms > > > > For aggressive dirtiers, the equality holds > > > > dirty_bw == N * task_ratelimit > > == N * pos_bw (1) > > > > The balanced throttle bandwidth can be estimated by > > > > ref_bw = pos_bw * write_bw / dirty_bw (2) > > > > >From (1) and (2), we get equality > > > > ref_bw == write_bw / N (3) > > > > If the N dd's are all throttled at ref_bw, the dirty/writeback rates > > will match. So ref_bw is the balanced dirty rate. > > Hi Fengguang, Hi Vivek, > So how much work it is to extend all this to handle the case of cgroups? Here is the simplest form. writeback: async write IO controllers http://git.kernel.org/?p=linux/kernel/git/wfg/writeback.git;a=blobdiff;f=mm/page-writeback.c;h=0b579e7fd338fd1f59cc36bf15fda06ff6260634;hp=34dff9f0d28d0f4f0794eb41187f71b4ade6b8a2;hb=1a58ad99ce1f6a9df6618a4b92fa4859cc3e7e90;hpb=5b6fcb3125ea52ff04a2fad27a51307842deb1a0 And an old email on this topic: https://lkml.org/lkml/2011/4/28/229 > IOW, I would imagine that you shall have to keep track of per cgroup/per > bdi state of many of the variables. For example, write_bw will become > per cgroup/per bdi entity instead of per bdi entity only. Same should > be true for position ratio, dirty_bw etc? The dirty_bw, write_bw and dirty_ratelimit should be replicated, but not necessarily dirty pages and position ratio. The cgroup can just rely on the root cgroup's dirty pages position control if it does not care about its own dirty pages consumptions. > I am assuming that if some cgroup is low weight on end device, then > WRITE bandwidth of that cgroup should go down and that should be > accounted for at per bdi state and task throttling should happen > accordingly so that a lower weight cgroup tasks get throttled more > as compared to higher weight cgroup tasks? Sorry I don't quite catch your meaning, but the current ->dirty_ratelimit adaptation scheme (detailed in another email) should handle all such rate/bw allocation issues automatically? Thanks, Fengguang -- To unsubscribe, send a message with 'unsubscribe linux-mm' in the body to majordomo@xxxxxxxxx. For more info on Linux MM, see: http://www.linux-mm.org/ . Fight unfair telecom internet charges in Canada: sign http://stopthemeter.ca/ Don't email: <a href=mailto:"dont@xxxxxxxxx"> email@xxxxxxxxx </a>