On Fri, Jun 18, 2021 at 10:00:17AM -0400, Brian Foster wrote: > On Thu, Jun 17, 2021 at 11:41:11AM -0700, Darrick J. Wong wrote: > > On Thu, Jun 17, 2021 at 10:23:41AM -0400, Brian Foster wrote: > > > On Tue, Jun 15, 2021 at 01:53:24PM -0700, Darrick J. Wong wrote: > > > > On Tue, Jun 15, 2021 at 10:43:28AM -0400, Brian Foster wrote: > > > > > On Mon, Jun 14, 2021 at 12:27:20PM -0700, Darrick J. Wong wrote: > > > > > > On Mon, Jun 14, 2021 at 12:17:03PM -0400, Brian Foster wrote: > > > > > > > On Sun, Jun 13, 2021 at 10:20:29AM -0700, Darrick J. Wong wrote: > > > > > > > > From: Darrick J. Wong <djwong@xxxxxxxxxx> > > > > > > > > > > > > > > > > Instead of calling xfs_inactive directly from xfs_fs_destroy_inode, > > > > > > > > defer the inactivation phase to a separate workqueue. With this change, > > > > > > > > we can speed up directory tree deletions by reducing the duration of > > > > > > > > unlink() calls to the directory and unlinked list updates. > > > > > > > > > > > > > > > > By moving the inactivation work to the background, we can reduce the > > > > > > > > total cost of deleting a lot of files by performing the file deletions > > > > > > > > in disk order instead of directory entry order, which can be arbitrary. > > > > > > > > > > > > > > > > We introduce two new inode flags -- NEEDS_INACTIVE and INACTIVATING. > > > > > > > > The first flag helps our worker find inodes needing inactivation, and > > > > > > > > the second flag marks inodes that are in the process of being > > > > > > > > inactivated. A concurrent xfs_iget on the inode can still resurrect the > > > > > > > > inode by clearing NEEDS_INACTIVE (or bailing if INACTIVATING is set). > > > > > > > > > > > > > > > > Unfortunately, deferring the inactivation has one huge downside -- > > > > > > > > eventual consistency. Since all the freeing is deferred to a worker > > > > > > > > thread, one can rm a file but the space doesn't come back immediately. > > > > > > > > This can cause some odd side effects with quota accounting and statfs, > > > > > > > > so we flush inactivation work during syncfs in order to maintain the > > > > > > > > existing behaviors, at least for callers that unlink() and sync(). > > > > > > > > > > > > > > > > For this patch we'll set the delay to zero to mimic the old timing as > > > > > > > > much as possible; in the next patch we'll play with different delay > > > > > > > > settings. > > > > > > > > > > > > > > > > Signed-off-by: Darrick J. Wong <djwong@xxxxxxxxxx> > > > > > > > > --- > > > > > > > > > > > > > > Just some high level questions on a first/cursory pass.. > > > > > > > > > > > > > > > Documentation/admin-guide/xfs.rst | 3 > > > > > > > > fs/xfs/scrub/common.c | 7 + > > > > > > > > fs/xfs/xfs_icache.c | 332 ++++++++++++++++++++++++++++++++++--- > > > > > > > > fs/xfs/xfs_icache.h | 5 + > > > > > > > > fs/xfs/xfs_inode.h | 19 ++ > > > > > > > > fs/xfs/xfs_log_recover.c | 7 + > > > > > > > > fs/xfs/xfs_mount.c | 26 +++ > > > > > > > > fs/xfs/xfs_mount.h | 21 ++ > > > > > > > > fs/xfs/xfs_super.c | 53 ++++++ > > > > > > > > fs/xfs/xfs_trace.h | 50 +++++- > > > > > > > > 10 files changed, 490 insertions(+), 33 deletions(-) > > > > > > > > > > > > > > > > > > > > > > > > diff --git a/Documentation/admin-guide/xfs.rst b/Documentation/admin-guide/xfs.rst > > > > > > > > index 8de008c0c5ad..f9b109bfc6a6 100644 > > > > > > > > --- a/Documentation/admin-guide/xfs.rst > > > > > > > > +++ b/Documentation/admin-guide/xfs.rst > > > > > > > > @@ -524,7 +524,8 @@ and the short name of the data device. They all can be found in: > > > > > > > > mount time quotacheck. > > > > > > > > xfs-gc Background garbage collection of disk space that have been > > > > > > > > speculatively allocated beyond EOF or for staging copy on > > > > > > > > - write operations. > > > > > > > > + write operations; and files that are no longer linked into > > > > > > > > + the directory tree. > > > > > > > > ================ =========== > > > > > > > > > > > > > > > > For example, the knobs for the quotacheck workqueue for /dev/nvme0n1 would be > > > > > > > ... > > > > > > > > diff --git a/fs/xfs/xfs_icache.c b/fs/xfs/xfs_icache.c > > > > > > > > index 4002f0b84401..e094c16aa8c5 100644 > > > > > > > > --- a/fs/xfs/xfs_icache.c > > > > > > > > +++ b/fs/xfs/xfs_icache.c > > > > > > > ... > > > > > > > > @@ -262,6 +285,9 @@ xfs_perag_set_inode_tag( > > > > > > > > case XFS_ICI_BLOCKGC_TAG: > > > > > > > > xfs_blockgc_queue(pag); > > > > > > > > break; > > > > > > > > + case XFS_ICI_INODEGC_TAG: > > > > > > > > + xfs_inodegc_queue(mp); > > > > > > > > + break; > > > > > > > > } > > > > > > > > > > > > > > > > trace_xfs_perag_set_inode_tag(mp, pag->pag_agno, tag, _RET_IP_); > > > > > > > > @@ -338,28 +364,26 @@ xfs_inode_mark_reclaimable( > > > > > > > > { > > > > > > > > struct xfs_mount *mp = ip->i_mount; > > > > > > > > struct xfs_perag *pag; > > > > > > > > + unsigned int tag; > > > > > > > > bool need_inactive = xfs_inode_needs_inactive(ip); > > > > > > > > > > > > > > Nit: I think we usually try to avoid function calls in the variable > > > > > > > declarations like this.. > > > > > > > > > > > > <shrug> For boolean flags that don't change value in the function body > > > > > > I'd rather they be initialized at declaration time so that code > > > > > > maintainers don't have to remember if a variable is initialized or not. > > > > > > need_inactive has scope and a valid value wherever you are in the > > > > > > function body. > > > > > > > > > > > > > > > > I think the usual pattern is to init such variables right after the > > > > > declarations so as to not obfuscate function calls (or at least I find > > > > > that much cleaner). > > > > > > > > Very well, I'll change it back. > > > > > > > > > > > > > > > > > > > > if (!need_inactive) { > > > > > > > > /* Going straight to reclaim, so drop the dquots. */ > > > > > > > > xfs_qm_dqdetach(ip); > > > > > > > > - } else { > > > > > > > > - xfs_inactive(ip); > > > > > > > > - } > > > > > > > > > > > > > > > > > > > > > > Ok, so obviously this disconnects the unlink -> destroy path from the > > > > > > > historical inactive -> reclaimable sequence. Is there any concern for > > > > > > > example from a memory usage standpoint or anything where we might want > > > > > > > to consider waiting on outstanding work here? E.g., if we had a large > > > > > > > filesystem, limited memory and enough CPUs doing unlinks where the > > > > > > > background reclaim simply can't keep up. > > > > > > > > > > > > So far in my evaluation, deferred inactivation has improved reclaim > > > > > > behavior because evict_inodes can push all eligible inodes all the way > > > > > > to IRECLAIMABLE without getting stuck behind xfs_inactive, and the > > > > > > shrinker can free other cached inodes sooner. I think we're still > > > > > > mostly reliant on xfs_inode_alloc looping when memory is low to throttle > > > > > > the size of the inode cache, since I think mass deletions are what push > > > > > > the inodegc code the hardest. If reclaim can't free inodes fast enough, > > > > > > I think that means the AIL is running slowly, in which case both > > > > > > frontend and background workers will block on the log. > > > > > > > > > > > > > I suppose it's a similar situation that inode reclaim is in, but we do > > > > > > > have shrinker feedback to hook back in and actually free available > > > > > > > objects. Do we have or need something like that if too many objects are > > > > > > > stuck waiting on inactivation? > > > > > > > > > > > > There's no direct link between the shrinker and the inodegc worker, > > > > > > which means that it can't push the inodegc worker to get even more > > > > > > inodes to IRECLAIMABLE. I've played with racing directory tree > > > > > > deletions of a filesystem with 8M inodes against fsstress on VMs with > > > > > > 512M of memory, but so far I haven't been able to drive the system into > > > > > > OOM any more easily than I could before this patch. The unlink calls > > > > > > seem to get stuck under xfs_iget, like I expect. > > > > > > > > > > > > > > > > Ok, that's sort of the scenario I was thinking about. I have a high > > > > > memory/cpu count box I occasionally use for such purposes. I threw an > > > > > fs_mark 1k file creation workload at a large fs on that system and let > > > > > it run to -ENOSPC, then ran a 64x parallel rm -rf of the top-level > > > > > subdirs. In the baseline case (a 5.12 based distro kernel), I see fairly > > > > > stable xfs_inode slab usage throughout the removal. It mostly hovers > > > > > from 150k-300k objects, sometimes spikes into the 500k-600k range, and > > > > > the test completes as expected. > > > > > > > > Hm. Without any patches applied, I loaded up a filesystem with 10 > > > > million inodes in a VM with 512M of RAM and ran rm -rf /mnt/*. The > > > > xfs_inode slab usage hovered around 20-30k objects, and occasionally > > > > spiked to about 80k. The inode count would decrease at about 7k inodes > > > > per second... > > > > > > > > > I see much different behavior from the same test on for-next plus this > > > > > patch. xfs_inode slab usage continuously increases from the onset, > > > > > eventually gets up to over 175m objects (~295GB worth of inodes), and > > > > > the oom killer takes everything down. This occurs after only about 10% > > > > > of the total consumed space has been removed. > > > > > > > > ...but as soon as I pushed the patch stack up to this particular patch, > > > > the slab cache usage went straight up to about 200k objects and the VM > > > > OOMed shortly thereafter. Until that happened, I could see the icount > > > > going down about about 12k inodes per second, but that doesn't do users > > > > much good. I observed that if I made xfs_inode_mark_reclaimable call > > > > flush_work on the inodegc work, the behaviors went back to the steady > > > > but slow deletion, though the VM still OOMed. > > > > > > > > I bet I'm messing up the shrinker accounting here, because an inode > > > > that's queued for inactivation is no longer on the vfs s_inode_lru and > > > > it's not tagged for reclaim, so a call to super_cache_count will not > > > > know about the inode, decide there isn't any reclaim work to do for this > > > > XFS and trip the OOM killer. > > > > > > > > > > Sounds plausible. > > > > Indeed, if I bump the return value by one if there's any inactivation > > work to do, the problem seems to go away, at least if reclaim_inodes_nr > > requeues the inodegc worker immediately. > > > > > > The other thing we might want to do is throttle the frontend so it won't > > > > get too far ahead of the background workers. If xfs_reclaim_inodes_nr > > > > requeues the worker immediately and xfs_inode_mark_reclaimable calls > > > > flush_work on the background worker, the OOMs seem to go away. > > > > > > > > > > Yes, that seems like the primary disconnect. At some point and somehow, > > > I think the unlink work needs to throttle against the debt of background > > > work it creates by just speeding along dumping fully unlinked inodes on > > > the "needs inactivation" workqueue. From there, the rate of unlinks or > > > frees might not matter as much, because that could always change based > > > on cpu count, RAM, storage, etc. > > > > <nod> At least for the single-threaded inodegc case, it seems to work > > all right if reclaim_inodes_nr() can set a flag to indicate that we hit > > memory pressure, kick the inodegc worker immediately, and then > > mark_reclaimable can flush_work() if the inode needs inactivation, the > > caller is not itself in a memory reclaim context, and the pressure flag > > is set. > > > > By 'all right', I mean that with no patches and 512M of RAM, the VM will > > OOM on a deltree of my 10 million inode fs; bumping the memory to 960M > > makes the OOM go away; and it doesn't come back once I roll all the > > patches back onto the branch. > > > > I don't know that this really qualifies as winning, though the test fs > > has a 1GB log, so with DRAM < LOG I guess that implies we can OOM long > > before anyone throttles on the log space. > > > > (xfs_repair -n will also sometimes OOM on this filesystem when the VM > > only has 512M of RAM...) > > > > Yeah, I intentionally did not restrict RAM availability in my test > because I didn't want to create a situation where the fundamental > problem was just too little resources for the workload. I suspect > there's always going to be some limit/workload where the right answer to > this sort of problem is "you need more memory." Yup. > > > > > I've not dug into the behavior enough to see where things are getting > > > > > backed up during this test, and I'm sure that this is not an > > > > > ideal/typical configuration (and increased slab usage is probably to be > > > > > expected given the nature of the change), but it does seem like we might > > > > > need some kind of adjustment in place to manage this kind of breakdown > > > > > scenario. > > > > > > > > When I pushed the stack up to the parallel inactivation series, the > > > > behavior changes again -- this time it takes a lot longer to push the VM > > > > to OOM. I think what's going on is that once I switch this to per-AG > > > > workers, the workers can keep up with the frontend that's inactivating > > > > the inodes. At least until those workers start competing with the > > > > unlink() activity for log space, and then they fall behind and the box > > > > OOMs again. Unfortunately, the changes I made above didn't prevent the > > > > OOM. > > > > > > > > > > Hm, Ok. So it sounds like that's a significant optimization but not > > > necessarily a fundamental fix. I.e., an environment where CPUs greatly > > > exceeds the AG count might still run into this kind of thing..? I can > > > repeat the test with subsequent patches to explore that, but my test > > > system is currently occupied with testing Dave's latest logging patches. > > > > Hehe. I /think/ I wrote an implementation bug in the parallel inodegc > > patch when I pushed it back on the stack. Though TBH I haven't made > > that much progress with the single-threaded case since I've also been > > running around in circles with the recent logging problems. > > > > > > > > The VFS still has this (erroneous) notion that calling ->destroy_inode > > > > > > will directly reclaim inodes from XFS, but the only difference from > > > > > > today's code is that it can take longer for inodes to get to that state. > > > > > > To fix that, I've also thought about making xfs_fs_free_cached_objects > > > > > > requeue the inodegc worker immediately (but without waiting for it), but > > > > > > so far I haven't needed to do that to avoid trouble. > > > > > > > > > > > > > > - if (!XFS_FORCED_SHUTDOWN(mp) && ip->i_delayed_blks) { > > > > > > > > - xfs_check_delalloc(ip, XFS_DATA_FORK); > > > > > > > > - xfs_check_delalloc(ip, XFS_COW_FORK); > > > > > > > > - ASSERT(0); > > > > > > > > + if (!XFS_FORCED_SHUTDOWN(mp) && ip->i_delayed_blks) { > > > > > > > > + xfs_check_delalloc(ip, XFS_DATA_FORK); > > > > > > > > + xfs_check_delalloc(ip, XFS_COW_FORK); > > > > > > > > + ASSERT(0); > > > > > > > > + } > > > > > > > > } > > > > > > > > > > > > > > > > XFS_STATS_INC(mp, vn_reclaim); > > > > > > > > > > > > > > > > /* > > > > > > > > - * We should never get here with one of the reclaim flags already set. > > > > > > > > + * We should never get here with any of the reclaim flags already set. > > > > > > > > */ > > > > > > > > - ASSERT_ALWAYS(!xfs_iflags_test(ip, XFS_IRECLAIMABLE)); > > > > > > > > - ASSERT_ALWAYS(!xfs_iflags_test(ip, XFS_IRECLAIM)); > > > > > > > > + ASSERT_ALWAYS(!xfs_iflags_test(ip, XFS_ALL_IRECLAIM_FLAGS)); > > > > > > > > > > > > > > > > /* > > > > > > > > * We always use background reclaim here because even if the inode is > > > > > > > ... > > > > > > > > @@ -889,22 +936,32 @@ xfs_dqrele_igrab( > > > > > > > > > > > > > > > > /* > > > > > > > > * Skip inodes that are anywhere in the reclaim machinery because we > > > > > > > > - * drop dquots before tagging an inode for reclamation. > > > > > > > > + * drop dquots before tagging an inode for reclamation. If the inode > > > > > > > > + * is being inactivated, skip it because inactivation will drop the > > > > > > > > + * dquots for us. > > > > > > > > */ > > > > > > > > - if (ip->i_flags & (XFS_IRECLAIM | XFS_IRECLAIMABLE)) > > > > > > > > + if (ip->i_flags & (XFS_IRECLAIM | XFS_IRECLAIMABLE | XFS_INACTIVATING)) > > > > > > > > goto out_unlock; > > > > > > > > > > > > > > I was a little curious why we'd skip reclaimable inodes here but not > > > > > > > need_inactive.. > > > > > > > > > > > > > > > > > > > > > > > /* > > > > > > > > - * The inode looks alive; try to grab a VFS reference so that it won't > > > > > > > > - * get destroyed. If we got the reference, return true to say that > > > > > > > > - * we grabbed the inode. > > > > > > > > + * If the inode is queued but not undergoing inactivation, set the > > > > > > > > + * inactivating flag so everyone will leave it alone and return true > > > > > > > > + * to say that we are taking ownership of it. > > > > > > > > + * > > > > > > > > + * Otherwise, the inode looks alive; try to grab a VFS reference so > > > > > > > > + * that it won't get destroyed. If we got the reference, return true > > > > > > > > + * to say that we grabbed the inode. > > > > > > > > * > > > > > > > > * If we can't get the reference, then we know the inode had its VFS > > > > > > > > * state torn down and hasn't yet entered the reclaim machinery. Since > > > > > > > > * we also know that dquots are detached from an inode before it enters > > > > > > > > * reclaim, we can skip the inode. > > > > > > > > */ > > > > > > > > - ret = igrab(VFS_I(ip)) != NULL; > > > > > > > > + ret = true; > > > > > > > > + if (ip->i_flags & XFS_NEED_INACTIVE) > > > > > > > > + ip->i_flags |= XFS_INACTIVATING; > > > > > > > > + else if (!igrab(VFS_I(ip))) > > > > > > > > + ret = false; > > > > > > > > > > > > > > ... but I guess we process the latter (in that we release the dquots and > > > > > > > revert the just added inactivating state). Is this a functional > > > > > > > requirement or considered an optimization? FWIW, it would be helpful if > > > > > > > the comment more explained why we did this as such as opposed to just > > > > > > > reflecting what the code already does. > > > > > > > > > > > > It's an optimization to avoid stalling quotaoff on NEEDS_INACTIVE > > > > > > inodes. This means that quotaoff can now cause the background inodegc > > > > > > worker to have to loop again, but seeing as quotaoff can pin the log > > > > > > tail, I decided that was a reasonable tradeoff. > > > > > > How does this rewording sound to you? > > > > > > > > > > > > > > > > Ok. I guess it would be nice to see some separation between core > > > > > functionality and this kind of optimization. It would make things more > > > > > clear to review for one, but also provides a trail of documentation and > > > > > intent if something like the approach to handling dquots causes some > > > > > problem in the future and the poor developer who might have to analyze > > > > > it otherwise has to navigate how this fits into the broader functional > > > > > rework. But I know people tend to bristle lately over factoring and > > > > > readability feedback and this is v7, so that's just my .02. That aside, > > > > > the comment below looks fine to me. > > > > > > > > I don't mind breaking up a large patch into smaller more targeted ones, > > > > because that's directly relevant to the code changes being proposed. It > > > > also isn't a terribly large change. It's more things like being told to > > > > fix or remove quotaoff (or that whole mess over building latency qos > > > > functions to decide when to defer ioend clear_pagewriteback to a > > > > workqueue) that bother me. > > > > > > > > > > IMO, it's more important to separate things based on mechanism/policy or > > > core function vs. optimizations and optional enhancements as opposed to > > > things like patch size. I suppose sometimes a patch can just end up big > > > enough that it's just easier to see a single functional change split up > > > into smaller chunks, but it's not clear to me that is the case here. > > > *shrug* Anyways, my comment is more around the mix of content than the > > > size of the patch. > > > > <nod> Ultimately, I did kick the optimization out to a separate patch. > > I will probably end up modifying this patch to flush_work from > > mark_reclaimable, then add a new patch to adjust the behavior to what we > > really want (i.e. having a real feedback loop based on memory reclaim > > trigger) so that we can capture some performance benefit. > > > > > With regard to the quotaoff thing, it seems like this work has to > > > include some nasty logic (here and in a subsequent patch) just to > > > prevent quotaoff from becoming more deadlock prone than it already is. > > > > Yep. > > > > > Given that we have outstanding work (that both Dave and I have > > > previously agreed is viable) to make quotaoff not deadlock prone, can we > > > consider whether incorporating that work would simplify this work such > > > that it can drop much of that logic? > > > > As I've said before, I only keep dragging the quotaoff changes along in > > this series because I don't want the resolution of that question to > > become a prerequisite for landing this change... > > > > > IIRC, the last time this came up Christoph suggested turning off > > > quotaoff entirely, to which we all agreed. A patchset came, had some > > > issues and hasn't been seen again. > > > > ...because that patchset came, had issues, and hasn't been resubmitted. > > I can only build my development tree off of a stable point on the > > upstream branch, so whatever I touch in quotaoff here is more or less > > the bare minimum needed to keep fstests happy. > > > > I was referring to the patches that have been sitting on the list since > before that immediately previous one came about: > > https://lore.kernel.org/linux-xfs/20210406144238.814558-1-bfoster@xxxxxxxxxx/ > > I'm not aware of any issues with that besides the ubiquitious "why don't > we do this other thing instead?" thing about removing quotaoff that so > far hasn't come to pass. So if we were to fix the deadlock issues with > quotaoff via that patch, could we eliminate the need for deferred > inactivation to incorporate all this logic just to avoid slowing > quotaoff down? We can always still remove quotaoff sometime in the > future... <nod> I didn't have any particular objection to it... > FWIW, I also vaguely recall an impediment to making the simple "allocate > the quotaoff_end transaction early" deadlock fix being that somewhere in > that sequence could lead to quotaoff running iput from transaction > context, and thus nesting with transactions in the inactivation path. It > might be worth taking another look at the big picture to see if this > work actually facilitates fixing quotaoff rather than needing to try so > hard to work around its flaws. ...I think I will look into adding your series on to the start of mine. I was rather hoping that hch would save us all the trouble and repost the quotaoff killing patches + fstests fixes, but that clearly wasn't in the cards for 5.14. I guess I have plenty of time to play with this until the next -rc4. :) --D > > Brian > > > > So as far I'm concerned, if we're > > > willing to turn quotaoff entirely, I don't care too much about if > > > something like deferred inactivation makes quotaoff incrementally more > > > slow (as opposed to turning it into a potential deadlock monster). > > > > If we dumped quotaoff entirely, I'd gut these parts from the patchset. > > > > > > > > /* > > > > > > * If the inode is queued but not currently undergoing > > > > > > * inactivation, we want to slip in to drop the dquots ASAP > > > > > > * because quotaoff can pin the log tail and cause log livelock. > > > > > > * Avoiding that is worth potentially forcing the inodegc worker > > > > > > * to make another pass. Set INACTIVATING to prevent inodegc > > > > > > * and iget from touching the inode. > > > > > > * > > > > > > * Otherwise, the inode looks alive; try to grab a VFS reference > > > > > > * so that it won't get destroyed. If we got the reference, > > > > > > * return true to say that we grabbed the inode. > > > > > > * > > > > > > * If we can't get the reference, then we know the inode had its > > > > > > * VFS state torn down and hasn't yet entered the reclaim > > > > > > * machinery. Since we also know that dquots are detached from > > > > > > * an inode before it enters reclaim, we can skip the inode. > > > > > > */ > > > > > > > > > > > > > > > > > > > > > > out_unlock: > > > > > > > > spin_unlock(&ip->i_flags_lock); > > > > > > > > @@ -917,6 +974,8 @@ xfs_dqrele_inode( > > > > > > > > struct xfs_inode *ip, > > > > > > > > struct xfs_icwalk *icw) > > > > > > > > { > > > > > > > > + bool live_inode; > > > > > > > > + > > > > > > > > if (xfs_iflags_test(ip, XFS_INEW)) > > > > > > > > xfs_inew_wait(ip); > > > > > > > > > > > > > > > > @@ -934,7 +993,19 @@ xfs_dqrele_inode( > > > > > > > > ip->i_pdquot = NULL; > > > > > > > > } > > > > > > > > xfs_iunlock(ip, XFS_ILOCK_EXCL); > > > > > > > > - xfs_irele(ip); > > > > > > > > + > > > > > > > > + /* > > > > > > > > + * If we set INACTIVATING earlier to prevent this inode from being > > > > > > > > + * touched, clear that state to let the inodegc claim it. Otherwise, > > > > > > > > + * it's a live inode and we need to release it. > > > > > > > > + */ > > > > > > > > + spin_lock(&ip->i_flags_lock); > > > > > > > > + live_inode = !(ip->i_flags & XFS_INACTIVATING); > > > > > > > > + ip->i_flags &= ~XFS_INACTIVATING; > > > > > > > > + spin_unlock(&ip->i_flags_lock); > > > > > > > > + > > > > > > > > + if (live_inode) > > > > > > > > + xfs_irele(ip); > > > > > > > > } > > > > > > > > > > > > > > > > /* > > > > > > > ... > > > > > > > > +/* > > > > > > > > + * Force all currently queued inode inactivation work to run immediately, and > > > > > > > > + * wait for the work to finish. > > > > > > > > + */ > > > > > > > > +void > > > > > > > > +xfs_inodegc_flush( > > > > > > > > + struct xfs_mount *mp) > > > > > > > > +{ > > > > > > > > + trace_xfs_inodegc_flush(mp, 0, _RET_IP_); > > > > > > > > + flush_delayed_work(&mp->m_inodegc_work); > > > > > > > > +} > > > > > > > > + > > > > > > > > +/* Disable the inode inactivation background worker and wait for it to stop. */ > > > > > > > > +void > > > > > > > > +xfs_inodegc_stop( > > > > > > > > + struct xfs_mount *mp) > > > > > > > > +{ > > > > > > > > + if (!test_and_clear_bit(XFS_OPFLAG_INODEGC_RUNNING_BIT, &mp->m_opflags)) > > > > > > > > + return; > > > > > > > > + > > > > > > > > + cancel_delayed_work_sync(&mp->m_inodegc_work); > > > > > > > > + trace_xfs_inodegc_stop(mp, 0, _RET_IP_); > > > > > > > > +} > > > > > > > > + > > > > > > > > +/* > > > > > > > > + * Enable the inode inactivation background worker and schedule deferred inode > > > > > > > > + * inactivation work if there is any. > > > > > > > > + */ > > > > > > > > +void > > > > > > > > +xfs_inodegc_start( > > > > > > > > + struct xfs_mount *mp) > > > > > > > > +{ > > > > > > > > + if (test_and_set_bit(XFS_OPFLAG_INODEGC_RUNNING_BIT, &mp->m_opflags)) > > > > > > > > + return; > > > > > > > > + > > > > > > > > + trace_xfs_inodegc_start(mp, 0, _RET_IP_); > > > > > > > > + xfs_inodegc_queue(mp); > > > > > > > > +} > > > > > > > > + > > > > > > > > > > > > > > This seems like a decent amount of boilerplate. Is all of this really > > > > > > > necessary and/or is it worth it just for the added tracepoints? > > > > > > > > > > > > I've gotten a lot of use out of the tracepoints to debug all sorts of > > > > > > weird mis-interactions between freeze, remounts, and the inodegc worker. > > > > > > > > > > > > > > > > Is that due to heavy development activity or something you realistically > > > > > expect to see used in production? I've nothing against tracepoints, but > > > > > I almost always have custom tracepoints during development that are > > > > > temporarily useful but eventually removed, and do question the tradeoff > > > > > of creating a bunch of wrapper code just for the tracepoints. > > > > > > > > I've found myself repeatedly adding them to debug some unexpected > > > > behavior, then removing them to polish things before publishing, and > > > > finally got fed up with doing this over and over and over. > > > > > > > > > > Fair enough. That's something that can be cleaned up down the road, if > > > needed. > > > > > > > > > > ISTM > > > > > > > that the start/stop thing is already technically racy wrt to queue > > > > > > > activity and the bit check in the worker might be sufficient to avoid > > > > > > > transaction/freeze deadlocks or whatever the original intent was. Hm? > > > > > > > > > > > > The bitflag isn't racy, though why isn't obvious. There are seven > > > > > > places where we change the INODEGC_RUNNING state: > > > > > > > > > > > > - For freeze and ro remount, the VFS will take s_umount before calling > > > > > > into XFS, and XFS then calls inodegc_flush to complete all the queued > > > > > > work before changing the bit state. > > > > > > > > > > > > During the v6 review Dave wondered if we even needed the INODEGC > > > > > > RUNNING flag, and I concluded that we need it for freeze, since we > > > > > > allow inodes to sleep under xfs_inactive() while trying to allocate a > > > > > > transaction. Under the new scheme, inodes can be queued for > > > > > > inactivation on a frozen fs, but we need xfs_inodegc_queue to know not > > > > > > to kick the worker without having to mess around with VFS locks to > > > > > > check the freeze state. > > > > > > > > > > > > > > > > Not sure I'm clear on what you describe above, but I'm not really > > > > > concerned about the existence of the state bit as much. The above refers > > > > > to xfs_inodegc_queue(), which already checks the bit, whereas my comment > > > > > was just more around the need for the xfs_inodegc_start/stop() wrappers. > > > > > > > > Now I'm completely confused about what you're asking about here. Are > > > > you suggesting that I could set and clear the INODEGC RUNNING bit > > > > directly from xfs_fs_freeze and xfs_fs_unfreeze, change > > > > xfs_inodegc_worker to bail out if INODEGC_RUNNING is set, and then > > > > replace all the inodegc_stop/start calls with direct calls to > > > > cancel_work_delayed_sync/xfs_inodegc_queue? > > > > > > > > > > Eh, I'm probably getting lost in the weeds here. My primary concern was > > > the proliferation of unnecessary boilerplate code and if we could manage > > > the background/workqueue task like we manage other such background > > > workers with respect to freeze and whatnot. If the helpers are going to > > > stay around anyways, then this doesn't matter that much. Feel free to > > > disregard this feedback. > > > > Ah, ok. This is totally one of those things where I want the > > tracepoints now because they're really helpful for debugging the > > machinery, but if it lands and stabilizes for a few months I won't > > object to them disappearing in one of the "get rid of trivial helpers" > > cleanups that seem so popular in these parts. ;) > > > > --D > > > > > > > > Brian > > > > > > > --D > > > > > > > > > > > > > > Brian > > > > > > > > > > > - For thaw and rw remount, the VFS takes s_umount and calls XFS, which > > > > > > re-enables the worker. > > > > > > > > > > > > - Mount and unmount run in isolation (since we can't have any open files > > > > > > on the filesystem) so there won't be any other threads trying to call > > > > > > into the filesystem. > > > > > > > > > > > > - I think xchk_{start,stop}_reaping is racy though. That code probably > > > > > > has to sb_start_write to prevent freeze or ro remount from messing > > > > > > with the INODEGC_RUNNING state. That would still leave a hole if > > > > > > fscounters scrub races with a rw remount and immediately files start > > > > > > getting deleted, but maybe that's just not worth worrying about. > > > > > > Scrub will try again a few times, and in djwong-dev I solve that by > > > > > > freezing the filesystem if the unlocked fscounters scrub fails. > > > > > > > > > > > > --D > > > > > > > > > > > > > > > > > > > > Brian > > > > > > > > > > > > > > > /* XFS Inode Cache Walking Code */ > > > > > > > > > > > > > > > > /* > > > > > > > > @@ -1708,6 +1979,8 @@ xfs_icwalk_igrab( > > > > > > > > return xfs_blockgc_igrab(ip); > > > > > > > > case XFS_ICWALK_RECLAIM: > > > > > > > > return xfs_reclaim_igrab(ip, icw); > > > > > > > > + case XFS_ICWALK_INODEGC: > > > > > > > > + return xfs_inodegc_igrab(ip); > > > > > > > > default: > > > > > > > > return false; > > > > > > > > } > > > > > > > > @@ -1736,6 +2009,9 @@ xfs_icwalk_process_inode( > > > > > > > > case XFS_ICWALK_RECLAIM: > > > > > > > > xfs_reclaim_inode(ip, pag); > > > > > > > > break; > > > > > > > > + case XFS_ICWALK_INODEGC: > > > > > > > > + xfs_inodegc_inactivate(ip, pag, icw); > > > > > > > > + break; > > > > > > > > } > > > > > > > > return error; > > > > > > > > } > > > > > > > > diff --git a/fs/xfs/xfs_icache.h b/fs/xfs/xfs_icache.h > > > > > > > > index 00dc98a92835..840eac06a71b 100644 > > > > > > > > --- a/fs/xfs/xfs_icache.h > > > > > > > > +++ b/fs/xfs/xfs_icache.h > > > > > > > > @@ -80,4 +80,9 @@ int xfs_icache_inode_is_allocated(struct xfs_mount *mp, struct xfs_trans *tp, > > > > > > > > void xfs_blockgc_stop(struct xfs_mount *mp); > > > > > > > > void xfs_blockgc_start(struct xfs_mount *mp); > > > > > > > > > > > > > > > > +void xfs_inodegc_worker(struct work_struct *work); > > > > > > > > +void xfs_inodegc_flush(struct xfs_mount *mp); > > > > > > > > +void xfs_inodegc_stop(struct xfs_mount *mp); > > > > > > > > +void xfs_inodegc_start(struct xfs_mount *mp); > > > > > > > > + > > > > > > > > #endif > > > > > > > > diff --git a/fs/xfs/xfs_inode.h b/fs/xfs/xfs_inode.h > > > > > > > > index e3137bbc7b14..fa5be0d071ad 100644 > > > > > > > > --- a/fs/xfs/xfs_inode.h > > > > > > > > +++ b/fs/xfs/xfs_inode.h > > > > > > > > @@ -240,6 +240,7 @@ static inline bool xfs_inode_has_bigtime(struct xfs_inode *ip) > > > > > > > > #define __XFS_IPINNED_BIT 8 /* wakeup key for zero pin count */ > > > > > > > > #define XFS_IPINNED (1 << __XFS_IPINNED_BIT) > > > > > > > > #define XFS_IEOFBLOCKS (1 << 9) /* has the preallocblocks tag set */ > > > > > > > > +#define XFS_NEED_INACTIVE (1 << 10) /* see XFS_INACTIVATING below */ > > > > > > > > /* > > > > > > > > * If this unlinked inode is in the middle of recovery, don't let drop_inode > > > > > > > > * truncate and free the inode. This can happen if we iget the inode during > > > > > > > > @@ -248,6 +249,21 @@ static inline bool xfs_inode_has_bigtime(struct xfs_inode *ip) > > > > > > > > #define XFS_IRECOVERY (1 << 11) > > > > > > > > #define XFS_ICOWBLOCKS (1 << 12)/* has the cowblocks tag set */ > > > > > > > > > > > > > > > > +/* > > > > > > > > + * If we need to update on-disk metadata before this IRECLAIMABLE inode can be > > > > > > > > + * freed, then NEED_INACTIVE will be set. Once we start the updates, the > > > > > > > > + * INACTIVATING bit will be set to keep iget away from this inode. After the > > > > > > > > + * inactivation completes, both flags will be cleared and the inode is a > > > > > > > > + * plain old IRECLAIMABLE inode. > > > > > > > > + */ > > > > > > > > +#define XFS_INACTIVATING (1 << 13) > > > > > > > > + > > > > > > > > +/* All inode state flags related to inode reclaim. */ > > > > > > > > +#define XFS_ALL_IRECLAIM_FLAGS (XFS_IRECLAIMABLE | \ > > > > > > > > + XFS_IRECLAIM | \ > > > > > > > > + XFS_NEED_INACTIVE | \ > > > > > > > > + XFS_INACTIVATING) > > > > > > > > + > > > > > > > > /* > > > > > > > > * Per-lifetime flags need to be reset when re-using a reclaimable inode during > > > > > > > > * inode lookup. This prevents unintended behaviour on the new inode from > > > > > > > > @@ -255,7 +271,8 @@ static inline bool xfs_inode_has_bigtime(struct xfs_inode *ip) > > > > > > > > */ > > > > > > > > #define XFS_IRECLAIM_RESET_FLAGS \ > > > > > > > > (XFS_IRECLAIMABLE | XFS_IRECLAIM | \ > > > > > > > > - XFS_IDIRTY_RELEASE | XFS_ITRUNCATED) > > > > > > > > + XFS_IDIRTY_RELEASE | XFS_ITRUNCATED | XFS_NEED_INACTIVE | \ > > > > > > > > + XFS_INACTIVATING) > > > > > > > > > > > > > > > > /* > > > > > > > > * Flags for inode locking. > > > > > > > > diff --git a/fs/xfs/xfs_log_recover.c b/fs/xfs/xfs_log_recover.c > > > > > > > > index 1227503d2246..9d8fc85bd28d 100644 > > > > > > > > --- a/fs/xfs/xfs_log_recover.c > > > > > > > > +++ b/fs/xfs/xfs_log_recover.c > > > > > > > > @@ -2784,6 +2784,13 @@ xlog_recover_process_iunlinks( > > > > > > > > } > > > > > > > > xfs_buf_rele(agibp); > > > > > > > > } > > > > > > > > + > > > > > > > > + /* > > > > > > > > + * Flush the pending unlinked inodes to ensure that the inactivations > > > > > > > > + * are fully completed on disk and the incore inodes can be reclaimed > > > > > > > > + * before we signal that recovery is complete. > > > > > > > > + */ > > > > > > > > + xfs_inodegc_flush(mp); > > > > > > > > } > > > > > > > > > > > > > > > > STATIC void > > > > > > > > diff --git a/fs/xfs/xfs_mount.c b/fs/xfs/xfs_mount.c > > > > > > > > index c3a96fb3ad80..ab65a14e51e6 100644 > > > > > > > > --- a/fs/xfs/xfs_mount.c > > > > > > > > +++ b/fs/xfs/xfs_mount.c > > > > > > > > @@ -514,7 +514,8 @@ xfs_check_summary_counts( > > > > > > > > * Flush and reclaim dirty inodes in preparation for unmount. Inodes and > > > > > > > > * internal inode structures can be sitting in the CIL and AIL at this point, > > > > > > > > * so we need to unpin them, write them back and/or reclaim them before unmount > > > > > > > > - * can proceed. > > > > > > > > + * can proceed. In other words, callers are required to have inactivated all > > > > > > > > + * inodes. > > > > > > > > * > > > > > > > > * An inode cluster that has been freed can have its buffer still pinned in > > > > > > > > * memory because the transaction is still sitting in a iclog. The stale inodes > > > > > > > > @@ -546,6 +547,7 @@ xfs_unmount_flush_inodes( > > > > > > > > mp->m_flags |= XFS_MOUNT_UNMOUNTING; > > > > > > > > > > > > > > > > xfs_ail_push_all_sync(mp->m_ail); > > > > > > > > + xfs_inodegc_stop(mp); > > > > > > > > cancel_delayed_work_sync(&mp->m_reclaim_work); > > > > > > > > xfs_reclaim_inodes(mp); > > > > > > > > xfs_health_unmount(mp); > > > > > > > > @@ -782,6 +784,9 @@ xfs_mountfs( > > > > > > > > if (error) > > > > > > > > goto out_log_dealloc; > > > > > > > > > > > > > > > > + /* Enable background inode inactivation workers. */ > > > > > > > > + xfs_inodegc_start(mp); > > > > > > > > + > > > > > > > > /* > > > > > > > > * Get and sanity-check the root inode. > > > > > > > > * Save the pointer to it in the mount structure. > > > > > > > > @@ -936,6 +941,15 @@ xfs_mountfs( > > > > > > > > xfs_irele(rip); > > > > > > > > /* Clean out dquots that might be in memory after quotacheck. */ > > > > > > > > xfs_qm_unmount(mp); > > > > > > > > + > > > > > > > > + /* > > > > > > > > + * Inactivate all inodes that might still be in memory after a log > > > > > > > > + * intent recovery failure so that reclaim can free them. Metadata > > > > > > > > + * inodes and the root directory shouldn't need inactivation, but the > > > > > > > > + * mount failed for some reason, so pull down all the state and flee. > > > > > > > > + */ > > > > > > > > + xfs_inodegc_flush(mp); > > > > > > > > + > > > > > > > > /* > > > > > > > > * Flush all inode reclamation work and flush the log. > > > > > > > > * We have to do this /after/ rtunmount and qm_unmount because those > > > > > > > > @@ -983,6 +997,16 @@ xfs_unmountfs( > > > > > > > > uint64_t resblks; > > > > > > > > int error; > > > > > > > > > > > > > > > > + /* > > > > > > > > + * Perform all on-disk metadata updates required to inactivate inodes > > > > > > > > + * that the VFS evicted earlier in the unmount process. Freeing inodes > > > > > > > > + * and discarding CoW fork preallocations can cause shape changes to > > > > > > > > + * the free inode and refcount btrees, respectively, so we must finish > > > > > > > > + * this before we discard the metadata space reservations. Metadata > > > > > > > > + * inodes and the root directory do not require inactivation. > > > > > > > > + */ > > > > > > > > + xfs_inodegc_flush(mp); > > > > > > > > + > > > > > > > > xfs_blockgc_stop(mp); > > > > > > > > xfs_fs_unreserve_ag_blocks(mp); > > > > > > > > xfs_qm_unmount_quotas(mp); > > > > > > > > diff --git a/fs/xfs/xfs_mount.h b/fs/xfs/xfs_mount.h > > > > > > > > index c78b63fe779a..dc906b78e24c 100644 > > > > > > > > --- a/fs/xfs/xfs_mount.h > > > > > > > > +++ b/fs/xfs/xfs_mount.h > > > > > > > > @@ -154,6 +154,13 @@ typedef struct xfs_mount { > > > > > > > > uint8_t m_rt_checked; > > > > > > > > uint8_t m_rt_sick; > > > > > > > > > > > > > > > > + /* > > > > > > > > + * This atomic bitset controls flags that alter the behavior of the > > > > > > > > + * filesystem. Use only the atomic bit helper functions here; see > > > > > > > > + * XFS_OPFLAG_* for information about the actual flags. > > > > > > > > + */ > > > > > > > > + unsigned long m_opflags; > > > > > > > > + > > > > > > > > /* > > > > > > > > * End of read-mostly variables. Frequently written variables and locks > > > > > > > > * should be placed below this comment from now on. The first variable > > > > > > > > @@ -184,6 +191,7 @@ typedef struct xfs_mount { > > > > > > > > uint64_t m_resblks_avail;/* available reserved blocks */ > > > > > > > > uint64_t m_resblks_save; /* reserved blks @ remount,ro */ > > > > > > > > struct delayed_work m_reclaim_work; /* background inode reclaim */ > > > > > > > > + struct delayed_work m_inodegc_work; /* background inode inactive */ > > > > > > > > struct xfs_kobj m_kobj; > > > > > > > > struct xfs_kobj m_error_kobj; > > > > > > > > struct xfs_kobj m_error_meta_kobj; > > > > > > > > @@ -258,6 +266,19 @@ typedef struct xfs_mount { > > > > > > > > #define XFS_MOUNT_DAX_ALWAYS (1ULL << 26) > > > > > > > > #define XFS_MOUNT_DAX_NEVER (1ULL << 27) > > > > > > > > > > > > > > > > +/* > > > > > > > > + * Operation flags -- each entry here is a bit index into m_opflags and is > > > > > > > > + * not itself a flag value. Use the atomic bit functions to access. > > > > > > > > + */ > > > > > > > > +enum xfs_opflag_bits { > > > > > > > > + /* > > > > > > > > + * If set, background inactivation worker threads will be scheduled to > > > > > > > > + * process queued inodegc work. If not, queued inodes remain in memory > > > > > > > > + * waiting to be processed. > > > > > > > > + */ > > > > > > > > + XFS_OPFLAG_INODEGC_RUNNING_BIT = 0, > > > > > > > > +}; > > > > > > > > + > > > > > > > > /* > > > > > > > > * Max and min values for mount-option defined I/O > > > > > > > > * preallocation sizes. > > > > > > > > diff --git a/fs/xfs/xfs_super.c b/fs/xfs/xfs_super.c > > > > > > > > index dd1ee333dcb3..0b01d9499395 100644 > > > > > > > > --- a/fs/xfs/xfs_super.c > > > > > > > > +++ b/fs/xfs/xfs_super.c > > > > > > > > @@ -714,6 +714,8 @@ xfs_fs_sync_fs( > > > > > > > > { > > > > > > > > struct xfs_mount *mp = XFS_M(sb); > > > > > > > > > > > > > > > > + trace_xfs_fs_sync_fs(mp, wait, _RET_IP_); > > > > > > > > + > > > > > > > > /* > > > > > > > > * Doing anything during the async pass would be counterproductive. > > > > > > > > */ > > > > > > > > @@ -730,6 +732,25 @@ xfs_fs_sync_fs( > > > > > > > > flush_delayed_work(&mp->m_log->l_work); > > > > > > > > } > > > > > > > > > > > > > > > > + /* > > > > > > > > + * Flush all deferred inode inactivation work so that the free space > > > > > > > > + * counters will reflect recent deletions. Do not force the log again > > > > > > > > + * because log recovery can restart the inactivation from the info that > > > > > > > > + * we just wrote into the ondisk log. > > > > > > > > + * > > > > > > > > + * For regular operation this isn't strictly necessary since we aren't > > > > > > > > + * required to guarantee that unlinking frees space immediately, but > > > > > > > > + * that is how XFS historically behaved. > > > > > > > > + * > > > > > > > > + * If, however, the filesystem is at FREEZE_PAGEFAULTS, this is our > > > > > > > > + * last chance to complete the inactivation work before the filesystem > > > > > > > > + * freezes and the log is quiesced. The background worker will not > > > > > > > > + * activate again until the fs is thawed because the VFS won't evict > > > > > > > > + * any more inodes until freeze_super drops s_umount and we disable the > > > > > > > > + * worker in xfs_fs_freeze. > > > > > > > > + */ > > > > > > > > + xfs_inodegc_flush(mp); > > > > > > > > + > > > > > > > > return 0; > > > > > > > > } > > > > > > > > > > > > > > > > @@ -844,6 +865,17 @@ xfs_fs_freeze( > > > > > > > > */ > > > > > > > > flags = memalloc_nofs_save(); > > > > > > > > xfs_blockgc_stop(mp); > > > > > > > > + > > > > > > > > + /* > > > > > > > > + * Stop the inodegc background worker. freeze_super already flushed > > > > > > > > + * all pending inodegc work when it sync'd the filesystem after setting > > > > > > > > + * SB_FREEZE_PAGEFAULTS, and it holds s_umount, so we know that inodes > > > > > > > > + * cannot enter xfs_fs_destroy_inode until the freeze is complete. > > > > > > > > + * If the filesystem is read-write, inactivated inodes will queue but > > > > > > > > + * the worker will not run until the filesystem thaws or unmounts. > > > > > > > > + */ > > > > > > > > + xfs_inodegc_stop(mp); > > > > > > > > + > > > > > > > > xfs_save_resvblks(mp); > > > > > > > > ret = xfs_log_quiesce(mp); > > > > > > > > memalloc_nofs_restore(flags); > > > > > > > > @@ -859,6 +891,14 @@ xfs_fs_unfreeze( > > > > > > > > xfs_restore_resvblks(mp); > > > > > > > > xfs_log_work_queue(mp); > > > > > > > > xfs_blockgc_start(mp); > > > > > > > > + > > > > > > > > + /* > > > > > > > > + * Don't reactivate the inodegc worker on a readonly filesystem because > > > > > > > > + * inodes are sent directly to reclaim. > > > > > > > > + */ > > > > > > > > + if (!(mp->m_flags & XFS_MOUNT_RDONLY)) > > > > > > > > + xfs_inodegc_start(mp); > > > > > > > > + > > > > > > > > return 0; > > > > > > > > } > > > > > > > > > > > > > > > > @@ -1665,6 +1705,9 @@ xfs_remount_rw( > > > > > > > > if (error && error != -ENOSPC) > > > > > > > > return error; > > > > > > > > > > > > > > > > + /* Re-enable the background inode inactivation worker. */ > > > > > > > > + xfs_inodegc_start(mp); > > > > > > > > + > > > > > > > > return 0; > > > > > > > > } > > > > > > > > > > > > > > > > @@ -1687,6 +1730,15 @@ xfs_remount_ro( > > > > > > > > return error; > > > > > > > > } > > > > > > > > > > > > > > > > + /* > > > > > > > > + * Stop the inodegc background worker. xfs_fs_reconfigure already > > > > > > > > + * flushed all pending inodegc work when it sync'd the filesystem. > > > > > > > > + * The VFS holds s_umount, so we know that inodes cannot enter > > > > > > > > + * xfs_fs_destroy_inode during a remount operation. In readonly mode > > > > > > > > + * we send inodes straight to reclaim, so no inodes will be queued. > > > > > > > > + */ > > > > > > > > + xfs_inodegc_stop(mp); > > > > > > > > + > > > > > > > > /* Free the per-AG metadata reservation pool. */ > > > > > > > > error = xfs_fs_unreserve_ag_blocks(mp); > > > > > > > > if (error) { > > > > > > > > @@ -1810,6 +1862,7 @@ static int xfs_init_fs_context( > > > > > > > > mutex_init(&mp->m_growlock); > > > > > > > > INIT_WORK(&mp->m_flush_inodes_work, xfs_flush_inodes_worker); > > > > > > > > INIT_DELAYED_WORK(&mp->m_reclaim_work, xfs_reclaim_worker); > > > > > > > > + INIT_DELAYED_WORK(&mp->m_inodegc_work, xfs_inodegc_worker); > > > > > > > > mp->m_kobj.kobject.kset = xfs_kset; > > > > > > > > /* > > > > > > > > * We don't create the finobt per-ag space reservation until after log > > > > > > > > diff --git a/fs/xfs/xfs_trace.h b/fs/xfs/xfs_trace.h > > > > > > > > index d0b4799ad1e6..ca9bfbd28886 100644 > > > > > > > > --- a/fs/xfs/xfs_trace.h > > > > > > > > +++ b/fs/xfs/xfs_trace.h > > > > > > > > @@ -156,6 +156,45 @@ DEFINE_PERAG_REF_EVENT(xfs_perag_put); > > > > > > > > DEFINE_PERAG_REF_EVENT(xfs_perag_set_inode_tag); > > > > > > > > DEFINE_PERAG_REF_EVENT(xfs_perag_clear_inode_tag); > > > > > > > > > > > > > > > > +DECLARE_EVENT_CLASS(xfs_fs_class, > > > > > > > > + TP_PROTO(struct xfs_mount *mp, int data, unsigned long caller_ip), > > > > > > > > + TP_ARGS(mp, data, caller_ip), > > > > > > > > + TP_STRUCT__entry( > > > > > > > > + __field(dev_t, dev) > > > > > > > > + __field(unsigned long long, mflags) > > > > > > > > + __field(unsigned long, opflags) > > > > > > > > + __field(unsigned long, sbflags) > > > > > > > > + __field(int, data) > > > > > > > > + __field(unsigned long, caller_ip) > > > > > > > > + ), > > > > > > > > + TP_fast_assign( > > > > > > > > + __entry->dev = mp->m_super->s_dev; > > > > > > > > + __entry->mflags = mp->m_flags; > > > > > > > > + __entry->opflags = mp->m_opflags; > > > > > > > > + __entry->sbflags = mp->m_super->s_flags; > > > > > > > > + __entry->data = data; > > > > > > > > + __entry->caller_ip = caller_ip; > > > > > > > > + ), > > > > > > > > + TP_printk("dev %d:%d flags 0x%llx opflags 0x%lx sflags 0x%lx data %d caller %pS", > > > > > > > > + MAJOR(__entry->dev), MINOR(__entry->dev), > > > > > > > > + __entry->mflags, > > > > > > > > + __entry->opflags, > > > > > > > > + __entry->sbflags, > > > > > > > > + __entry->data, > > > > > > > > + (char *)__entry->caller_ip) > > > > > > > > +); > > > > > > > > + > > > > > > > > +#define DEFINE_FS_EVENT(name) \ > > > > > > > > +DEFINE_EVENT(xfs_fs_class, name, \ > > > > > > > > + TP_PROTO(struct xfs_mount *mp, int data, unsigned long caller_ip), \ > > > > > > > > + TP_ARGS(mp, data, caller_ip)) > > > > > > > > +DEFINE_FS_EVENT(xfs_inodegc_flush); > > > > > > > > +DEFINE_FS_EVENT(xfs_inodegc_start); > > > > > > > > +DEFINE_FS_EVENT(xfs_inodegc_stop); > > > > > > > > +DEFINE_FS_EVENT(xfs_inodegc_queue); > > > > > > > > +DEFINE_FS_EVENT(xfs_inodegc_worker); > > > > > > > > +DEFINE_FS_EVENT(xfs_fs_sync_fs); > > > > > > > > + > > > > > > > > DECLARE_EVENT_CLASS(xfs_ag_class, > > > > > > > > TP_PROTO(struct xfs_mount *mp, xfs_agnumber_t agno), > > > > > > > > TP_ARGS(mp, agno), > > > > > > > > @@ -615,14 +654,17 @@ DECLARE_EVENT_CLASS(xfs_inode_class, > > > > > > > > TP_STRUCT__entry( > > > > > > > > __field(dev_t, dev) > > > > > > > > __field(xfs_ino_t, ino) > > > > > > > > + __field(unsigned long, iflags) > > > > > > > > ), > > > > > > > > TP_fast_assign( > > > > > > > > __entry->dev = VFS_I(ip)->i_sb->s_dev; > > > > > > > > __entry->ino = ip->i_ino; > > > > > > > > + __entry->iflags = ip->i_flags; > > > > > > > > ), > > > > > > > > - TP_printk("dev %d:%d ino 0x%llx", > > > > > > > > + TP_printk("dev %d:%d ino 0x%llx iflags 0x%lx", > > > > > > > > MAJOR(__entry->dev), MINOR(__entry->dev), > > > > > > > > - __entry->ino) > > > > > > > > + __entry->ino, > > > > > > > > + __entry->iflags) > > > > > > > > ) > > > > > > > > > > > > > > > > #define DEFINE_INODE_EVENT(name) \ > > > > > > > > @@ -666,6 +708,10 @@ DEFINE_INODE_EVENT(xfs_inode_free_eofblocks_invalid); > > > > > > > > DEFINE_INODE_EVENT(xfs_inode_set_cowblocks_tag); > > > > > > > > DEFINE_INODE_EVENT(xfs_inode_clear_cowblocks_tag); > > > > > > > > DEFINE_INODE_EVENT(xfs_inode_free_cowblocks_invalid); > > > > > > > > +DEFINE_INODE_EVENT(xfs_inode_set_reclaimable); > > > > > > > > +DEFINE_INODE_EVENT(xfs_inode_reclaiming); > > > > > > > > +DEFINE_INODE_EVENT(xfs_inode_set_need_inactive); > > > > > > > > +DEFINE_INODE_EVENT(xfs_inode_inactivating); > > > > > > > > > > > > > > > > /* > > > > > > > > * ftrace's __print_symbolic requires that all enum values be wrapped in the > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > > >