On Mon, Feb 19, 2018 at 08:21:04AM -0500, Brian Foster wrote: > On Mon, Feb 19, 2018 at 01:16:36PM +1100, Dave Chinner wrote: > > On Fri, Feb 16, 2018 at 07:56:25AM -0500, Brian Foster wrote: > > > On Fri, Feb 16, 2018 at 09:31:38AM +1100, Dave Chinner wrote: > > > > On Fri, Feb 09, 2018 at 11:12:41AM -0500, Brian Foster wrote: > > > > > On Thu, Feb 01, 2018 at 05:42:02PM +1100, Dave Chinner wrote: > > > > > > + bp = xfs_growfs_get_hdr_buf(mp, > > > > > > + XFS_AG_DADDR(mp, agno, XFS_SB_DADDR), > > > > > > + XFS_FSS_TO_BB(mp, 1), 0, &xfs_sb_buf_ops); > > > > > > > > > > This all seems fine to me up until the point where we use uncached > > > > > buffers for pre-existing secondary superblocks. This may all be fine now > > > > > if nothing else happens to access/use secondary supers, but it seems > > > > > like this essentially enforces that going forward. > > > > > > > > > > Hmm, I see that scrub does appear to look at secondary superblocks via > > > > > cached buffers. Shouldn't we expect this path to maintain coherency with > > > > > an sb buffer that may have been read/cached from there? > > > > > > > > Good catch! I wrote this before scrub started looking at secondary > > > > superblocks. As a general rulle, we don't want to cache secondary > > > > superblocks as they should never be used by the kernel except in > > > > exceptional situations like grow or scrub. > > > > > > > > I'll have a look at making this use cached buffers that get freed > > > > immediately after we release them (i.e. don't go onto the LRU) and > > > > that should solve the problem. > > > > > > > > > > Ok. Though that sounds a bit odd. What is the purpose of a cached buffer > > > that is not cached? > > > > Serialisation of concurrent access to what is normal a single-use > > access code path while it is in memory. i.e. exactly the reason we > > have XFS_IGET_DONTCACHE and use it for things like bulkstat lookups. > > > > Well, that's the purpose of looking up a cached instance of an uncached > buffer. That makes sense, but that's only half the question... > > > > Isn't the behavior you're after here (perhaps > > > analogous to pagecache coherency management between buffered/direct I/O) > > > more cleanly implemented using a cache invalidation mechanism? E.g., > > > invalidate cache, use uncached buffer (then perhaps invalidate again). > > > > Invalidation as a mechanism for non-coherent access sycnhronisation > > is completely broken model when it comes to concurrent access. We > > explicitly tell app developers not ot mix cached + uncached IO to > > the same file for exactly this reason. Using a cached buffer and > > using the existing xfs_buf_find/lock serialisation avoids this > > problem, and by freeing them immediately after we've used them we > > also minimise the memory footprint of single-use access patterns. > > > > Ok.. > > > > I guess I'm also a little curious why we couldn't continue to use cached > > > buffers here, > > > > As I said, we will continue to use cached buffers here. I'll just > > call xfs_buf_set_ref(bp, 0) on them so they are reclaimed when > > released. That means concurrent access will serialise correctly > > through _xfs_buf_find(), otherwise we won't keep them in memory. > > > > Ok, but what's the purpose/motivation for doing that here? Purely to > save on memory? Partly, but mainly because they are single use buffers and accesses are so rare that it's a waste of resources to cache them because they'll be reclaimed long before they are ever accessed again. > Is that really an impactful enough change in behavior > for (pre-existing) secondary superblocks? Yes. We know that there are people out there doing "create tiny, deploy, grow to thousands of AGs" as part of their crazy, screwed up container deployment scripts. THat's thousands of secondary superblocks that will be cached and generate unnecessary memory pressure when cached, > This seems a clear enough > decision when growfs was the only consumer of these buffers, but having > another cached accessor kind of clouds the logic. Scrub is not something that runs often enough we should be trying to cache it's metadata to speed up the next run. The whole point of scrub is that it reads metadata that hasn't been accessed in a long time to verify it hasn't degraded. Caching secondary superblocks for either growfs or scrub makes no sense. However, we have to make sure if the two occur at the same time, their actions are coherent and correctly serialised. > E.g., if task A reads a set of buffers cached, it's made a decision that > it's potentially beneficial to leave them around. Now we have task B > that has decided it doesn't want to cache the buffers, but what bearing > does that have on task A? It certainly makes sense for task B to drop > any buffer that wasn't already cached, but for already cached buffers it > doesn't really make sense for task B to decide there is no further > advantage to caching for task A. > > FWIW, I think this is how IGET_DONTCACHE works: don't cache the inode > unless it was actually found in cache. I presume that is so a bulkstat > or whatever doesn't toss the existing cached inode working set. Yes, precisely the point of this inode cache behaviour. However, that's not a concern for secondary superblocks because they are never part of the working set of metadata ongoing user workloads require to be cached. They only get brought into memory as a result of admin operations, and those are very, very rare. Cheers, Dave. -- Dave Chinner david@xxxxxxxxxxxxx -- To unsubscribe from this list: send the line "unsubscribe linux-xfs" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html