On Tue, Sep 08, 2009 at 06:06:23PM +0200, Peter Zijlstra wrote: > On Tue, 2009-09-08 at 13:37 +0300, Artem Bityutskiy wrote: > > Hi, > > > > On 09/08/2009 12:23 PM, Jens Axboe wrote: > > > From: Theodore Ts'o<tytso@xxxxxxx> > > > > > > Originally, MAX_WRITEBACK_PAGES was hard-coded to 1024 because of a > > > concern of not holding I_SYNC for too long. (At least, that was the > > > comment previously.) This doesn't make sense now because the only > > > time we wait for I_SYNC is if we are calling sync or fsync, and in > > > that case we need to write out all of the data anyway. Previously > > > there may have been other code paths that waited on I_SYNC, but not > > > any more. > > > > > > According to Christoph, the current writeback size is way too small, > > > and XFS had a hack that bumped out nr_to_write to four times the value > > > sent by the VM to be able to saturate medium-sized RAID arrays. This > > > value was also problematic for ext4 as well, as it caused large files > > > to be come interleaved on disk by in 8 megabyte chunks (we bumped up > > > the nr_to_write by a factor of two). > > > > > > So, in this patch, we make the MAX_WRITEBACK_PAGES a tunable, > > > max_writeback_mb, and set it to a default value of 128 megabytes. > > > > > > http://bugzilla.kernel.org/show_bug.cgi?id=13930 > > > > > > Signed-off-by: "Theodore Ts'o"<tytso@xxxxxxx> > > > Signed-off-by: Jens Axboe<jens.axboe@xxxxxxxxxx> > > > > Would be nice to update doc files like > > > > Documentation/sysctl/vm.txt > > Documentation/filesystems/proc.txt > > I'm still not convinced this knob is worth the patch and I'm inclined to > flat out NAK it.. > > The whole point of MAX_WRITEBACK_PAGES seems to occasionally check the > dirty stats again and not write out too much. The problem is that 'too much' is a very abstract thing. When a process is stuck in balance_dirty_pages, we want them to do the minimal amount of work (or waiting) required to get them safely back inside file_write(). > > Clearly the current limit isn't sufficient for some people, > - xfs/btrfs seem generally stuck in balance_dirty_pages()'s > congestion_wait() > - ext4 generates inconveniently small extents This is actually two different side of the same problem. The filesystem knows that bytes 0-N in the file are setup for delayed allocation. Writepage is called on byte 0, and now the filesystem gets to decide how big an extent to make. It could decide to make an extent based on the total number of bytes under delayed allocation, and hope the caller of writepage will be kind enough to send down the pages contiguously afterward (xfs), or it could make a smaller extent based on something closer to the total number of bytes this particular writepages() call plans on writing (I guess what ext4 is doing). Either way, if pdflush or the bdi thread or whoever ends up switching to another file during a big streaming write, the end result is that we fragment. We may fragment the file (ext4) or we may fragment the writeback (xfs), but the end result isn't good. Looking at two xfs examples, this is the IO for two concurrent streaming writers (two different files) on 2.6.31-rc8 (pdflush is doing all the IO in this graph, sorry the legend colors wrapped on me). If you squint, you can kind of see the fingers of IO as pdflush switches between files. http://oss.oracle.com/~mason/seekwatcher/xfs-tag.png And here is the IO when XFS forces nr_to_write much higher with a patch from Christoph: http://oss.oracle.com/~mason/seekwatcher/xfs-extend-tag.png These graphs would look the same no matter what I did with congestion_wait(). The first graph is slower just because pdflush switches from one file to another. > > > The first seems to suggest to me the number isn't well balanced against > whatever drives congestion_wait() (that thing still gives me a > head-ache). > > # git grep clear_bdi_congested > drivers/block/pktcdvd.c: clear_bdi_congested(&pd->disk->queue->backing_dev_info, > fs/fuse/dev.c: clear_bdi_congested(&fc->bdi, BLK_RW_SYNC); > fs/fuse/dev.c: clear_bdi_congested(&fc->bdi, BLK_RW_ASYNC); > fs/nfs/write.c: clear_bdi_congested(&nfss->backing_dev_info, BLK_RW_ASYNC); > include/linux/backing-dev.h:void clear_bdi_congested(struct backing_dev_info *bdi, int sync); > include/linux/blkdev.h: clear_bdi_congested(&q->backing_dev_info, sync); > mm/backing-dev.c:void clear_bdi_congested(struct backing_dev_info *bdi, int sync) > mm/backing-dev.c:EXPORT_SYMBOL(clear_bdi_congested); > > Suggests that regular block devices don't even manage device congestion > and it reverts to a simple timeout -- should we fix that? Look for blk_clear_queue_congested(). It is managed, I personally don't think it is very useful. But, that's a different thread ;) > > Now, suppose it were to do something useful, I'd think we'd want to > limit write-out to whatever it takes so saturate the BDI. If we don't want a blanket increase, I'd suggest that we just give the FS a way to say: 'I know nr_to_write is only 32, but if you just write a few blocks more, the system will be better off'. Something like wbc->fs_write_hint This way, when the FS allocates a great big contiguous delalloc extent, it can set the wbc to reflect that we've got cheap and easy IO here. > > > As to the extends, shouldn't ext4 allocate extends based on the amount > of dirty pages in the file instead of however much we're going to write > out now? It probably does a mixture of both. -chris -- To unsubscribe from this list: send the line "unsubscribe linux-fsdevel" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html