Hi Wu, As you know, I am not a expert in this area. So I hope my review can help understanding other newbie like me and make clear this document. :) I didn't look into the code. before it, I would like to clear your concept. On Wed, Nov 17, 2010 at 1:27 PM, Wu Fengguang <fengguang.wu@xxxxxxxxx> wrote: > As proposed by Chris, Dave and Jan, don't start foreground writeback IO > inside balance_dirty_pages(). Instead, simply let it idle sleep for some > time to throttle the dirtying task. In the mean while, kick off the > per-bdi flusher thread to do background writeback IO. > > This patch introduces the basic framework, which will be further > consolidated by the next patches. > > RATIONALS > ========= > > The current balance_dirty_pages() is rather IO inefficient. > > - concurrent writeback of multiple inodes (Dave Chinner) > > If every thread doing writes and being throttled start foreground > writeback, it leads to N IO submitters from at least N different > inodes at the same time, end up with N different sets of IO being > issued with potentially zero locality to each other, resulting in > much lower elevator sort/merge efficiency and hence we seek the disk > all over the place to service the different sets of IO. > OTOH, if there is only one submission thread, it doesn't jump between > inodes in the same way when congestion clears - it keeps writing to > the same inode, resulting in large related chunks of sequential IOs > being issued to the disk. This is more efficient than the above > foreground writeback because the elevator works better and the disk > seeks less. > > - IO size too small for fast arrays and too large for slow USB sticks > > The write_chunk used by current balance_dirty_pages() cannot be > directly set to some large value (eg. 128MB) for better IO efficiency. > Because it could lead to more than 1 second user perceivable stalls. > Even the current 4MB write size may be too large for slow USB sticks. > The fact that balance_dirty_pages() starts IO on itself couples the > IO size to wait time, which makes it hard to do suitable IO size while > keeping the wait time under control. > > For the above two reasons, it's much better to shift IO to the flusher > threads and let balance_dirty_pages() just wait for enough time or progress. > > Jan Kara, Dave Chinner and me explored the scheme to let > balance_dirty_pages() wait for enough writeback IO completions to > safeguard the dirty limit. However it's found to have two problems: > > - in large NUMA systems, the per-cpu counters may have big accounting > errors, leading to big throttle wait time and jitters. > > - NFS may kill large amount of unstable pages with one single COMMIT. > Because NFS server serves COMMIT with expensive fsync() IOs, it is > desirable to delay and reduce the number of COMMITs. So it's not > likely to optimize away such kind of bursty IO completions, and the > resulted large (and tiny) stall times in IO completion based throttling. > > So here is a pause time oriented approach, which tries to control the > pause time in each balance_dirty_pages() invocations, by controlling > the number of pages dirtied before calling balance_dirty_pages(), for > smooth and efficient dirty throttling: > > - avoid useless (eg. zero pause time) balance_dirty_pages() calls > - avoid too small pause time (less than 10ms, which burns CPU power) > - avoid too large pause time (more than 100ms, which hurts responsiveness) > - avoid big fluctuations of pause times > > For example, when doing a simple cp on ext4 with mem=4G HZ=250. > > before patch, the pause time fluctuates from 0 to 324ms > (and the stall time may grow very large for slow devices) > > [ 1237.139962] balance_dirty_pages: write_chunk=1536 pages_written=0 pause=56 > [ 1237.207489] balance_dirty_pages: write_chunk=1536 pages_written=0 pause=0 > [ 1237.225190] balance_dirty_pages: write_chunk=1536 pages_written=0 pause=0 > [ 1237.234488] balance_dirty_pages: write_chunk=1536 pages_written=0 pause=0 > [ 1237.244692] balance_dirty_pages: write_chunk=1536 pages_written=0 pause=0 > [ 1237.375231] balance_dirty_pages: write_chunk=1536 pages_written=0 pause=31 > [ 1237.443035] balance_dirty_pages: write_chunk=1536 pages_written=0 pause=15 > [ 1237.574630] balance_dirty_pages: write_chunk=1536 pages_written=0 pause=31 > [ 1237.642394] balance_dirty_pages: write_chunk=1536 pages_written=0 pause=15 > [ 1237.666320] balance_dirty_pages: write_chunk=1536 pages_written=57 pause=5 > [ 1237.973365] balance_dirty_pages: write_chunk=1536 pages_written=0 pause=81 > [ 1238.212626] balance_dirty_pages: write_chunk=1536 pages_written=0 pause=56 > [ 1238.280431] balance_dirty_pages: write_chunk=1536 pages_written=0 pause=15 > [ 1238.412029] balance_dirty_pages: write_chunk=1536 pages_written=0 pause=31 > [ 1238.412791] balance_dirty_pages: write_chunk=1536 pages_written=0 pause=0 > > after patch, the pause time remains stable around 32ms > > cp-2687 [002] 1452.237012: balance_dirty_pages: weight=56% dirtied=128 pause=8 > cp-2687 [002] 1452.246157: balance_dirty_pages: weight=56% dirtied=128 pause=8 > cp-2687 [006] 1452.253043: balance_dirty_pages: weight=56% dirtied=128 pause=8 > cp-2687 [006] 1452.261899: balance_dirty_pages: weight=57% dirtied=128 pause=8 > cp-2687 [006] 1452.268939: balance_dirty_pages: weight=57% dirtied=128 pause=8 > cp-2687 [002] 1452.276932: balance_dirty_pages: weight=57% dirtied=128 pause=8 > cp-2687 [002] 1452.285889: balance_dirty_pages: weight=57% dirtied=128 pause=8 > > CONTROL SYSTEM > ============== > > The current task_dirty_limit() adjusts bdi_dirty_limit to get > task_dirty_limit according to the dirty "weight" of the current task, > which is the percent of pages recently dirtied by the task. If 100% > pages are recently dirtied by the task, it will lower bdi_dirty_limit by > 1/8. If only 1% pages are dirtied by the task, it will return almost > unmodified bdi_dirty_limit. In this way, a heavy dirtier will get > blocked at task_dirty_limit=(bdi_dirty_limit-bdi_dirty_limit/8) while > allowing a light dirtier to progress (the latter won't be blocked > because R << B in fig.1). > > Fig.1 before patch, a heavy dirtier and a light dirtier > R > ----------------------------------------------+-o---------------------------*--| > L A B T > T: bdi_dirty_limit, as returned by bdi_dirty_limit() > L: T - T/8 > > R: bdi_reclaimable + bdi_writeback > > A: task_dirty_limit for a heavy dirtier ~= R ~= L > B: task_dirty_limit for a light dirtier ~= T > > Since each process has its own dirty limit, we reuse A/B for the tasks as > well as their dirty limits. > > If B is a newly started heavy dirtier, then it will slowly gain weight > and A will lose weight. The task_dirty_limit for A and B will be > approaching the center of region (L, T) and eventually stabilize there. > > Fig.2 before patch, two heavy dirtiers converging to the same threshold > R > ----------------------------------------------+--------------o-*---------------| > L A B T Seems good until now. So, What's the problem if two heavy dirtiers have a same threshold? > > Fig.3 after patch, one heavy dirtier > | > throttle_bandwidth ~= bdi_bandwidth => o > | o > | o > | o > | o > | o > La| o > ----------------------------------------------+-+-------------o----------------| > R A T > T: bdi_dirty_limit > A: task_dirty_limit = T - Wa * T/16 > La: task_throttle_thresh = A - A/16 > > R: bdi_dirty_pages = bdi_reclaimable + bdi_writeback ~= La > > Now for IO-less balance_dirty_pages(), let's do it in a "bandwidth control" > way. In fig.3, a soft dirty limit region (La, A) is introduced. When R enters > this region, the task may be throttled for J jiffies on every N pages it dirtied. > Let's call (N/J) the "throttle bandwidth". It is computed by the following formula: > > throttle_bandwidth = bdi_bandwidth * (A - R) / (A - La) > where > A = T - Wa * T/16 > La = A - A/16 > where Wa is task weight for A. It's 0 for very light dirtier and 1 for > the one heavy dirtier (that consumes 100% bdi write bandwidth). The > task weight will be updated independently by task_dirty_inc() at > set_page_dirty() time. Dumb question. I can't see the difference between old and new, La depends on A. A depends on Wa. T is constant? Then, throttle_bandwidth depends on Wa. Wa depends on the number of dirtied pages during some interval. So if light dirtier become heavy, at last light dirtier and heavy dirtier will have a same weight. It means throttle_bandwidth is same. It's a same with old result. Please, open my eyes. :) Thanks for the great work. > > When R < La, we don't throttle it at all. > When R > A, the code will detect the negativeness and choose to pause > 100ms (the upper pause boundary), then loop over again. -- Kind regards, Minchan Kim -- To unsubscribe, send a message with 'unsubscribe linux-mm' in the body to majordomo@xxxxxxxxxx For more info on Linux MM, see: http://www.linux-mm.org/ . Fight unfair telecom policy in Canada: sign http://dissolvethecrtc.ca/ Don't email: <a href