On Wed, Dec 13, 2017 at 09:20:04AM +1100, Dave Chinner wrote: > On Tue, Dec 12, 2017 at 01:05:35PM -0500, Josef Bacik wrote: > > On Tue, Dec 12, 2017 at 10:36:19AM +1100, Dave Chinner wrote: > > > On Mon, Dec 11, 2017 at 04:55:31PM -0500, Josef Bacik wrote: > > > > From: Josef Bacik <jbacik@xxxxxx> > > > > > > > > Now that we have metadata counters in the VM, we need to provide a way to kick > > > > writeback on dirty metadata. Introduce super_operations->write_metadata. This > > > > allows file systems to deal with writing back any dirty metadata we need based > > > > on the writeback needs of the system. Since there is no inode to key off of we > > > > need a list in the bdi for dirty super blocks to be added. From there we can > > > > find any dirty sb's on the bdi we are currently doing writeback on and call into > > > > their ->write_metadata callback. > > > > > > > > Signed-off-by: Josef Bacik <jbacik@xxxxxx> > > > > Reviewed-by: Jan Kara <jack@xxxxxxx> > > > > Reviewed-by: Tejun Heo <tj@xxxxxxxxxx> > > > > --- > > > > fs/fs-writeback.c | 72 ++++++++++++++++++++++++++++++++++++---- > > > > fs/super.c | 6 ++++ > > > > include/linux/backing-dev-defs.h | 2 ++ > > > > include/linux/fs.h | 4 +++ > > > > mm/backing-dev.c | 2 ++ > > > > 5 files changed, 80 insertions(+), 6 deletions(-) > > > > > > > > diff --git a/fs/fs-writeback.c b/fs/fs-writeback.c > > > > index 987448ed7698..fba703dff678 100644 > > > > --- a/fs/fs-writeback.c > > > > +++ b/fs/fs-writeback.c > > > > @@ -1479,6 +1479,31 @@ static long writeback_chunk_size(struct bdi_writeback *wb, > > > > return pages; > > > > } > > > > > > > > +static long writeback_sb_metadata(struct super_block *sb, > > > > + struct bdi_writeback *wb, > > > > + struct wb_writeback_work *work) > > > > +{ > > > > + struct writeback_control wbc = { > > > > + .sync_mode = work->sync_mode, > > > > + .tagged_writepages = work->tagged_writepages, > > > > + .for_kupdate = work->for_kupdate, > > > > + .for_background = work->for_background, > > > > + .for_sync = work->for_sync, > > > > + .range_cyclic = work->range_cyclic, > > > > + .range_start = 0, > > > > + .range_end = LLONG_MAX, > > > > + }; > > > > + long write_chunk; > > > > + > > > > + write_chunk = writeback_chunk_size(wb, work); > > > > + wbc.nr_to_write = write_chunk; > > > > + sb->s_op->write_metadata(sb, &wbc); > > > > + work->nr_pages -= write_chunk - wbc.nr_to_write; > > > > + > > > > + return write_chunk - wbc.nr_to_write; > > > > > > Ok, writeback_chunk_size() returns a page count. We've already gone > > > through the "metadata is not page sized" dance on the dirty > > > accounting side, so how are we supposed to use pages to account for > > > metadata writeback? > > > > > > > This is just one of those things that's going to be slightly shitty. It's the > > same for memory reclaim, all of those places use pages so we just take > > METADATA_*_BYTES >> PAGE_SHIFT to get pages and figure it's close enough. > > Ok, so that isn't exactly easy to deal with, because all our > metadata writeback is based on log sequence number targets (i.e. how > far to push the tail of the log towards the current head). We've > actually got no idea how pages/bytes actually map to a LSN target > because while we might account a full buffer as dirty for memory > reclaim purposes (up to 64k in size), we might have only logged 128 > bytes of it. > > i.e. if we are asked to push 2MB of metadata and we treat that as > 2MB of log space (i.e. push target of tail LSN + 2MB) we could have > logged several tens of megabytes of dirty metadata in that LSN > range and have to flush it all. OTOH, if the buffers are fully > logged, then that same target might only flush 1.5MB of metadata > once all the log overhead is taken into account. > > So there's a fairly large disconnect between the "flush N bytes of > metadata" API and the "push to a target LSN" that XFS uses for > flushing metadata in aged order. I'm betting that extN and otehr > filesystems might have similar mismatches with their journal > flushing... > If there's not a correlation then there's no sense in xfs using this. If btrfs has 16gib of dirty metadata then that's exactly how much we have to write out, which is what this is designed for. > > > And, from what I can tell, if work->sync_mode = WB_SYNC_ALL or > > > work->tagged_writepages is set, this will basically tell us to flush > > > the entire dirty metadata cache because write_chunk will get set to > > > LONG_MAX. > > > > > > IOWs, this would appear to me to change sync() behaviour quite > > > dramatically on filesystems where ->write_metadata is implemented. > > > That is, instead of leaving all the metadata dirty in memory and > > > just forcing the journal to stable storage, filesystems will be told > > > to also write back all their dirty metadata before sync() returns, > > > even though it is not necessary to provide correct sync() > > > semantics.... > > > > Well for btrfs that's exactly what we have currently since it's just backed by > > an inode. > > Hmmmm. That explains a lot. > > Seems to me that btrfs is the odd one out here, so I'm not sure a > mechanism primarily designed for btrfs is going to work > generically.... > The generic stuff is very lightweight specifically because we don't need a whole lot, just a way to get all of the balance dirty pages logic without duplicating it internally in btrfs. > > Obviously this is different for journaled fs'es, but I assumed that > > in your case you would either not use this part of the infrastructure or simply > > ignore WB_SYNC_ALL and use WB_SYNC_NONE as a way to be nice under memory > > pressure or whatever. > > I don't think that designing an interface with the assumption other > filesystems will abuse it until it works for them is a great process > to follow... > Again not really designing it with your stuff in mind. ext* and xfs already handle dirty metadata fine, btrfs is the odd man out so we need a little extra. It would be cool to at least use the accounting part of it in xfs and ext* so we could see how much of the system memory is in use by metadata, but I imagine the dirty metadata tracking is going to be mostly useless for you guys. > > > Mind you, writeback invocation is so convoluted now I could easily > > > be mis-interpretting this code, but it does seem to me like this > > > code is going to have some unintended behaviours.... > > > > > > > I don't think so, because right now this behavior is exactly what btrfs has > > currently with it's inode setup. I didn't really think the journaled use case > > out since you guys are already rate limited by the journal. > > We are? > > XFS is rate limited by metadata writeback, not journal throughput. > Yes, journal space is limited by the metadata writeback rate, but > journalling itself is not the bottleneck. > I'm not saying "rate limited" as in xfs sucks because journal. I'm saying your dirty metadata foot print is limited by your journal size, so you aren't going to have gigabytes of dirty metadata sitting around needing to be flushed (I assume, I'm going on previous discussions with you about this.) > > If you would want > > to start using this stuff what would you like to see done instead? Thanks, > > If this is all about reacting to memory pressure, then writeback is > not the mechanism that should drive this writeback. Reacting to > memory pressure is what shrinkers are for, and XFS already triggers > metadata writeback on memory pressure. Hence I don't see how this > writeback mechanism would help us if we have to abuse it to infer > "memory pressure occurring" > This isn't reacting to memory pressure, it's reacting to dirty pressure. Btrfs is only limited by system memory for its metadata, so I want all the benefits of years of work on balance_dirty_pages() without having to duplicate the effort internally in btrfs. This is how I'm going about doing it. > What I was hoping for was this interface to be a mechanism to drive > periodic background metadata writeback from the VFS so that when we > start to run out of memory the VFS has already started to ramp up > the rate of metadata writeback so we don't have huge amounts of dirty > metadata to write back during superblock shrinker based reclaim. > > i.e. it works more like dirty background data writeback, get's the > amount of work to do from the amount of dirty metadata associated > with the bdi and doesn't actually do anything when operations like > sync() are run because there isn't a need to writeback metadata in > those operations. > > IOWs, treating metadata like it's one great big data inode doesn't > seem to me to be the right abstraction to use for this - in most > fileystems it's a bunch of objects with a complex dependency tree > and unknown write ordering, not an inode full of data that can be > sequentially written. But this isn't dictating what to write out, just how much we need to undirty. How the fs wants to write stuff out is completely up to the file system, I specifically made it as generic as possible so we could do whatever we felt like with the numbers we got. This work gives you exactly what you want, a callback when balance dirty pages is telling us that hey we have too much dirty memory in use on the system. > > Maybe we need multiple ops with well defined behaviours. e.g. > ->writeback_metadata() for background writeback, ->sync_metadata() for > sync based operations. That way different filesystems can ignore the > parts they don't need simply by not implementing those operations, > and the writeback code doesn't need to try to cater for all > operations through the one op. The writeback code should be cleaner, > the filesystem code should be cleaner, and we can tailor the work > guidelines for each operation separately so there's less mismatch > between what writeback is asking and how filesystems track dirty > metadata... > So I don't mind adding new things or changing around, but this is just getting us the same behavior that I mentioned before, only at a higher level. We want the balance_dirty_pages() stuff to be able to dip into metadata writeback via the method that I've implemented here. Basically do data writeback, and if we didn't do enough do some metadata writeback. With what you've proposed we would keep that and instead of doing ->write_metadata() when we have SYNC_ALL we'd just do ->sync_metadata() and let the fs figure out what to do, which is what I was suggesting fs'es do. The problem is there's a disconnect between what btrfs and ext4 do with their dirty metadata and what xfs does. Ext4 is going to log entire blocks into the journal, so there's a 1:1 mapping of dirty metadata to what's going to be written out. So telling it "write x pages" worth of metadata is going to be somewhat useful. That's not the case for xfs, and I'm not sure what a good way to accommodate you would look like. My first thought is a ratio, but man trying to change how we dealt with slab ratios made me want to suck start a shotgun so I don't really want to do something like that again. How would you prefer to get information to act on from upper layers? Personally I feel like the generic writeback stuff already gives us enough info and we can figure out what we want to do from there. Thanks, Josef ps: I'm going to try and stay up for a while so we can hash this out now instead of switching back and forth through our timezones.