On Sat 19-09-09 23:03:51, Wu Fengguang wrote: > On Sat, Sep 19, 2009 at 12:26:07PM +0800, Wu Fengguang wrote: > > On Sat, Sep 19, 2009 at 12:00:51PM +0800, Wu Fengguang wrote: > > > On Sat, Sep 19, 2009 at 11:58:35AM +0800, Wu Fengguang wrote: > > > > On Sat, Sep 19, 2009 at 01:52:52AM +0800, Theodore Tso wrote: > > > > > On Fri, Sep 11, 2009 at 10:39:29PM +0800, Wu Fengguang wrote: > > > > > > > > > > > > That would be good. Sorry for the late work. I'll allocate some time > > > > > > in mid next week to help review and benchmark recent writeback works, > > > > > > and hope to get things done in this merge window. > > > > > > > > > > Did you have some chance to get more work done on the your writeback > > > > > patches? > > > > > > > > Sorry for the delay, I'm now testing the patches with commands > > > > > > > > cp /dev/zero /mnt/test/zero0 & > > > > dd if=/dev/zero of=/mnt/test/zero1 & > > > > > > > > and the attached debug patch. > > > > > > > > One problem I found with ext3/4 is, redirty_tail() is called repeatedly > > > > in the traces, which could slow down the inode writeback significantly. > > > > > > FYI, it's this redirty_tail() called in writeback_single_inode(): > > > > > > /* > > > * Someone redirtied the inode while were writing back > > > * the pages. > > > */ > > > redirty_tail(inode); > > > > Hmm, this looks like an old fashioned problem get blew up by the > > 128MB MAX_WRITEBACK_PAGES. > > > > The inode was redirtied by the busy cp/dd processes. Now it takes much > > more time to sync 128MB, so that a heavy dirtier can easily redirty > > the inode in that time window. > > > > One single invocation of redirty_tail() could hold up the writeback of > > current inode for up to 30 seconds. > > It seems that this patch helps. However I'm afraid it's too late to > risk merging such kind of patches now.. Fenguang, could we maybe write down how the logic should look like and then look at the code and modify it as needed to fit the logic? Because I couldn't find a compact description of the logic anywhere in the code. Here is how I'd imaging the writeout logic should work: We would have just two lists - b_dirty and b_more_io. Both would be ordered by dirtied_when. A thread doing WB_SYNC_ALL writeback will just walk the list and cleanup everything (we should be resistant against livelocks because we stop at inode which has been dirtied after the sync has started). A thread doing WB_SYNC_NONE writeback will start walking the list. If the inode has I_SYNC set, it puts it on b_more_io. Otherwise it takes I_SYNC and writes as much as it finds necessary from the first inode. If it stopped before it wrote everything, it puts the inode at the end of b_more_io. If it wrote everything (writeback_index cycled or scanned the whole range) but inode is dirty, it puts the inode at the end of b_dirty and resets dirtied_when to the current time. Then it continues with the next inode. kupdate style writeback stops scanning dirty list when dirtied_when is new enough. Then if b_more_io is nonempty, it splices it into the beginning of the dirty list and restarts. Other types of writeback splice b_more_io to b_dirty when b_dirty gets empty. pdflush style writeback writes until we drop below background dirty limit. Other kinds of writeback (throttled threads, writeback submitted by filesystem itself) write while nr_to_write > 0. If we didn't write anything during the b_dirty scan, we wait until I_SYNC of the first inode on b_more_io gets cleared before starting the next scan. Does this look reasonably complete and cover all the cases? Honza -- Jan Kara <jack@xxxxxxx> SUSE Labs, CR -- To unsubscribe from this list: send the line "unsubscribe linux-fsdevel" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html