On Wed, Jun 30, 2010 at 2:24 PM, Andrew Morton <akpm@xxxxxxxxxxxxxxxxxxxx> wrote: > Well... why are these useful? In what operational scenario would > someone use these and get goodness from the experience? Where is the > value? Sell it to us! OK here it is in email before I add it to the commit description. Before when users are trying to track their IO activity there has always been a gap in the flow from user app to disk for buffered IO. With page_dirtied and page_entered_writeback the user can now track IO from buffered writes as they are indicated to the block layer. pages_dirtied helps storage workloads generating buffered writes that need to see over time how much memory the app is able to dirty. It can help trace app issues where iostat won't. In mixed workloads where an appserver is writing via DIRECT_IO it can help root cause issues where other apps are giving bursts of io behavior. pages_entered_writeback is useful to help grant visibility into the writeback subsystem. By tracking pages_entered_writeback with pages_dirtied app developers can learn about the performance and/or stability of the writeback subsystem. Comparing the rates of change between the two allow developers to see when writeback is not able to keep up with incoming traffic and the rate of dirty memory being sent to the IO back end. > It's hard to see how any future implementation could have a problem > implementing pages_dirtied and pages_entered_writeback, however > dirty_threshold and dirty_background_threshold are, I think, somewhat > specific to the current implementation and may be hard to maintain next > time we rip up and rewrite everything. We already expose these thresholds in /proc/sys/vm with dirty_background_ratio and background_ratio. What's frustrating about the ratio variables and the need for these are that they are not honored by the kernel. Instead the kernel may alter the number requested without giving the user any indication that is the case. An app developer can set the ratio to 2% but end up with 5% as get_dirty_limits makes sure it is never lower than 5% when set from the ratio. Arguably that can be fixed too but the limits which decide whether writeback is invoked to aggressively clean dirty pages is dependent on changing page state retrieved in determine_dirtyable_memory. It makes understanding when the kernel decides to writeback data a moving target that no app can ever determine. With these thresholds visible and collected over time it gives apps a chance to know why writeback happened, or why it did not. As systems get larger and larger RAM developers use the ratios to predict when their workloads will see writeback invoked. Today there is no way to accurately predict this. > Documentation doesn't describe the units. Pages? kbytes? bytes? Ouch. Thanks. That will be fixed. > I think it's best to encode the units in the procfs filename > (eg: dirty_expire_centisecs, min_free_kbytes). I agree that will be fixed. > units? They will all get units > We're very very interested in knowing how many pages entered writeback > via mm/vmscan.c however this procfs file lumps those together with the > pages which entered writeback via the regular writeback paths, I assume. Yes and I think that's ok. It describes how the whole system is moving dirty memory to writeback state and sending it to the I/O path. TO me trying to distinguish between fs/fs-writeback.c code doing this or vmscan.c code doing this is exposing implementation that we may change in the future. > But we need EXPORT_SYMBOL(account_page_dirtied), methinks. Ouch thanks. Will be fixed. > This should be a separate patch IMO. I will split these into two patches. One with the fix and then the other with the counters. mrubin -- To unsubscribe from this list: send the line "unsubscribe linux-fsdevel" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html