On Wed, Jun 12, 2019 at 08:55:06AM -0400, Brian Foster wrote: > On Tue, Jun 11, 2019 at 05:32:19PM -0700, Darrick J. Wong wrote: > > On Wed, Jun 12, 2019 at 09:23:47AM +1000, Dave Chinner wrote: > Since we're already discussing tweaks to readahead, another approach to > this problem could be to try and readahead all the way into the inode > cache. For example, consider a mechanism where a cluster buffer > readahead sets a flag on the buffer that effectively triggers an iget of > each allocated inode in the buffer. Darrick has already shown that the > inode memory allocation and iget itself has considerable overhead even > when the cluster buffer is already cached. We know that's not due to > btree lookups because quotacheck isn't using IGET_UNTRUSTED, so perhaps > we could amortize more of this cost via readahead. The DONTCACHE inode caching semantics of bulkstat tend to conflict with "readahead all the way to the inode cache". > The caveats are that would probably be more involved than something that > just caches the current cluster buffer and passes it into the iget path. > We'd have to rectify readahead in-core inodes against DONTCACHE inodes > used by bulkstat, for example, though I don't think that would be too > difficult to address via a new inode readahead flag or some such > preserve existing DONTCACHE behavior. I did try that once, the cache thrashing was .... difficult to contain under memory pressure. bulkstat pushes hundreds of thousands of inodes a second through the inode cache, and under memory pressure it will still cause working set perturbation with DONTCACHE being set. Holding DONTCACHE inodes for some time in the cache kinda defeats the simple purpose it has, and relying on cache hits to convert "readahead" to "dont cache" becomes really nasty when we try to use inode readahead for other things (like speeding up directory traversals by having xfs_readdir() issue inode readahead). The largest delay in bulkstat is the inode cluster IO latency. Getting rid of that is where the biggest win is (hence cluster read-ahead). The second largest overhead is the CPU burnt doing inode lookups, and on filesystems with lots of inodes, a significant amount of that is in the IGET_UNTRUSTED inobt lookup. IOWs, avoiding GET_UNTRUSTED is relatively low hanging fruit. The next limitation for bulkstat is the superblock inode list lock contention. Getting rid of the IGET_UNTRUSTED overhead is likely to push the lock contention into the severe range (the lock is already the largest CPU consumer at 16 threads bulkstating 600,000 inodes/s on a 16p machine) so until we get rid of that lock contention, there isn't much point in doing major rework to the bulkstat algorithm as it doesn't address the limitations that the current algorithm has. > It's also likely that passing the buffer into iget would already address > most of the overhead associated with the buffer lookup, so there might > not be enough tangible benefit at that point. The positive is that it's > probably an incremental step on top of an "iget from an existing cluster > buffer" mechanism and so could be easily prototyped by hacking in a read > side b_iodone handler or something. We don't want to put inode cache insertion into a IO completion routine. Tried it, caused horrible problems with metadata read IO latency and substantially increased inode cache lock contention by bouncing the radix trees around both submission and completion CPU contexts... /me has spent many, many years trying lots of different ways to make the inode cache in XFS go faster and has failed most of the time.... Cheers, Dave. -- Dave Chinner david@xxxxxxxxxxxxx