On Fri, Mar 25, 2011 at 03:00:54PM +0800, Wu Fengguang wrote: > Hi Jan, > > On Fri, Mar 25, 2011 at 09:28:03AM +0800, Jan Kara wrote: > > Hi, > > > > while working on changes to balance_dirty_pages() I was investigating why > > NFS writeback is *so* bumpy when I do not call writeback_inodes_wb() from > > balance_dirty_pages(). Take a single dd writing to NFS. What I can > > see is that we quickly accumulate dirty pages upto limit - ~700 MB on that > > machine. So flusher thread starts working and in an instant all these ~700 > > MB transition from Dirty state to Writeback state. Then, as server acks > > That can be fixed by the following patch: > > [PATCH 09/27] nfs: writeback pages wait queue > https://lkml.org/lkml/2011/3/3/79 I don't think this is a good definition of write congestion for a NFS (or any other network fs) client. Firstly, writeback congestion is really dependent on the size of the network send window remaining. That is, if you've filled the socket buffer with writes and would block trying to queue more pages on the socket, then you are congested. i.e. the measure of congestion is the rate at which write request can be sent to the server and processed by the server. Secondly, the big problem that causes the lumpiness is that we only send commits when we reach at large threshold of unstable pages. Because most servers tend to cache large writes in RAM, the server might have a long commit latency because it has to write hundred of MB of data to disk to complete the commit. IOWs, the client sends the commit only when it really needs the pages the be cleaned, and then we have the latency of the server write before it responds that they are clean. Hence commits can take a long time to complete and mark pages clean on the client side. A solution that IRIX used for this problem was the concept of a background commit. While doing writeback on an inode, if it sent more than than a certain threshold of data (typically in the range of 0.5-2s worth of data) to the server without a commit being issued, it would send an _asynchronous_ commit with the current dirty range to the server. That way the server starts writing the data before it hits dirty thresholds (i.e. prevents GBs of dirty data being cached on the server so commit lantecy is kept low). When the background commit completes the NFS client can then convert pages in the commit range to clean. Hence we keep the number of unstable pages under control without needing to wait for a certain number of unstable pages to build up before commits are triggered. This allows the process of writing dirty pages to clean unstable pages at roughly the same rate as the write rate without needing any magic thresholds to be configured.... Cheers, Dave. -- Dave Chinner david@xxxxxxxxxxxxx -- To unsubscribe from this list: send the line "unsubscribe linux-fsdevel" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html