On Wed, Jun 12, 2019 at 09:23:47AM +1000, Dave Chinner wrote: > On Mon, Jun 10, 2019 at 09:58:52AM -0400, Brian Foster wrote: > > On Tue, Jun 04, 2019 at 02:49:40PM -0700, Darrick J. Wong wrote: > > > From: Darrick J. Wong <darrick.wong@xxxxxxxxxx> > > > > > > Convert quotacheck to use the new iwalk iterator to dig through the > > > inodes. > > > > > > Signed-off-by: Darrick J. Wong <darrick.wong@xxxxxxxxxx> > > > Reviewed-by: Dave Chinner <dchinner@xxxxxxxxxx> > > > --- > > > fs/xfs/xfs_qm.c | 62 ++++++++++++++++++------------------------------------- > > > 1 file changed, 20 insertions(+), 42 deletions(-) > > > > > > > > > diff --git a/fs/xfs/xfs_qm.c b/fs/xfs/xfs_qm.c > > > index aa6b6db3db0e..a5b2260406a8 100644 > > > --- a/fs/xfs/xfs_qm.c > > > +++ b/fs/xfs/xfs_qm.c > > ... > > > @@ -1136,20 +1135,18 @@ xfs_qm_dqusage_adjust( > > > * rootino must have its resources accounted for, not so with the quota > > > * inodes. > > > */ > > > - if (xfs_is_quota_inode(&mp->m_sb, ino)) { > > > - *res = BULKSTAT_RV_NOTHING; > > > - return -EINVAL; > > > - } > > > + if (xfs_is_quota_inode(&mp->m_sb, ino)) > > > + return 0; > > > > > > /* > > > * We don't _need_ to take the ilock EXCL here because quotacheck runs > > > * at mount time and therefore nobody will be racing chown/chproj. > > > */ > > > - error = xfs_iget(mp, NULL, ino, XFS_IGET_DONTCACHE, 0, &ip); > > > - if (error) { > > > - *res = BULKSTAT_RV_NOTHING; > > > + error = xfs_iget(mp, tp, ino, XFS_IGET_DONTCACHE, 0, &ip); > > > > I was wondering if we should start using IGET_UNTRUSTED here, but I > > guess we're 1.) protected by quotacheck context and 2.) have the same > > record validity semantics as the existing bulkstat walker. LGTM: > > FWIW, I'd be wanting to go the other way with bulkstat. i.e. finding > ways of reducing IGET_UNTRUSTED in bulkstat because it adds > substantial CPU overhead during inode lookup because it has to look > up the inobt to validate the inode number. i.e. we are locking the > AGI and doing an inobt lookup on every inode we bulkstat because > there is some time between the initial inobt lookup and the > xfs_iget() call and that's when the inode chunk can get removed. > > IOWs, we only need to validate that the inode buffer still contains > inodes before we start instantiating inodes from it, but because we > don't hold any locks across individual inode processing in bulkstat > we have to revalidate that buffer contains inodes for every > allocated inode in that buffer. If we had a way of passing a locked > cluster buffer into xfs_iget to avoid having to look it up and read > it, we could do a single inode cluster read after validating the > inobt record is still valid, we could cycle all the remaining inodes > through xfs_iget() without having to use IGET_UNTRUSTED to validate > the inode cluster still contains valid inodes on every inode.... > > We still need to cycle inodes through the cache (so bulkstat is > coherent with other inode operations), but this would substantially > reduce the per-inode bulkstat CPU overhead, I think.... I'll think about this as an addendum to the series, because I suspect that remodelling the existing users is going to be an entire series on its own. (IOWs, my brain is too tired for today) --D > Cheers, > > Dave. > -- > Dave Chinner > david@xxxxxxxxxxxxx