On Mon, Jan 17, 2022 at 04:28:52PM +0000, Al Viro wrote: > On Mon, Jan 17, 2022 at 09:35:58AM -0500, Brian Foster wrote: > > > To Al's question, at the end of the day there is no rcu delay involved > > with inode reuse in XFS. We do use call_rcu() for eventual freeing of > > inodes (see __xfs_inode_free()), but inode reuse occurs for inodes that > > have been put into a "reclaim" state before getting to the point of > > freeing the struct inode memory. This lead to the long discussion [1] > > Ian references around ways to potentially deal with that. I think the > > TLDR of that thread is there are various potential options for > > improvement, such as to rcu wait on inode creation/reuse (either > > explicitly or via more open coded grace period cookie tracking), to rcu > > wait somewhere in the destroy sequence before inodes become reuse > > candidates, etc., but none of them seemingly agreeable for varying > > reasons (IIRC mostly stemming from either performance or compexity) [2]. > > > > The change that has been made so far in XFS is to turn rcuwalk for > > symlinks off once again, which looks like landed in Linus' tree as > > commit 7b7820b83f23 ("xfs: don't expose internal symlink metadata > > buffers to the vfs"). The hope is that between that patch and this > > prospective vfs tweak, we can have a couple incremental fixes that at > > least address the practical problem users have been running into (which > > is a crash due to a NULL ->get_link() callback pointer due to inode > > reuse). The inode reuse vs. rcu thing might still be a broader problem, > > but AFAIA that mechanism has been in place in XFS on Linux pretty much > > forever. > > My problem with that is that pathname resolution very much relies upon > the assumption that any inode it observes will *not* change its nature > until the final rcu_read_unlock(). Papering over ->i_op->get_link reads > in symlink case might be sufficient at the moment (I'm still not certain > about that, though), but that's rather brittle. E.g. if some XFS change > down the road adds ->permission() on some inodes, you'll get the same > problem in do_inode_permission(). We also have places where we rely upon > sample ->d_seq > fetch ->d_flags > fetch ->d_inode > validate ->d_seq > ... > assume that inode type matches the information in flags > > How painful would it be to make xfs_destroy_inode() a ->free_inode() instance? > IOW, how far is xfs_inode_mark_reclaimable() from being callable in RCU > callback context? Note that ->destroy_inode() is called via > As discussed on IRC, this was brought up in the earlier discussion by Miklos. Dave expressed some concern around locking, but I'm not sure I grok the details from reading back [1]. The implication seems to be the lookup side would have to rcu wait on associated inodes in the destroy side, which might be more of a concern about unconditional use of free_inode() as opposed to more selective rcu waiting for unlinked (or inactive) inodes. Dave would need to chime in further on that.. As it is, it looks to me that unlinked inodes unconditionally go to the inactive queues and thus the create side (xfs_iget_cache_hit(), if we're considering inode reuse) basically pushes on the queue and waits on the inode state to clear. Given that, ISTM it shouldn't be that much functional pain to introduce an rcu delay somewhere before an inactive inode becomes reclaimable (and thus reusable). I think the impediment to something like this has been more performance related. An inode alloc/free workload can turn into a continuous reuse of the same batch of inodes, over and over. Therefore an rcu wait on iget reuse can become effectively unconditional and slow things down quite a bit (hence my previous, untested thought around making it conditional and potentially amortizing the cost). I had played with a more selective grace period in the teardown side for inactive inodes via queue_rcu_work(), since that's an easy place to inject an rcu delay/callback, but that had some performance impact on sustained file removal that might require retuning other bits.. Brian [1] https://lore.kernel.org/linux-fsdevel/20211114231834.GM449541@xxxxxxxxxxxxxxxxxxx/#t > static void destroy_inode(struct inode *inode) > { > const struct super_operations *ops = inode->i_sb->s_op; > > BUG_ON(!list_empty(&inode->i_lru)); > __destroy_inode(inode); > if (ops->destroy_inode) { > ops->destroy_inode(inode); > if (!ops->free_inode) > return; > } > inode->free_inode = ops->free_inode; > call_rcu(&inode->i_rcu, i_callback); > } > > with > > static void i_callback(struct rcu_head *head) > { > struct inode *inode = container_of(head, struct inode, i_rcu); > if (inode->free_inode) > inode->free_inode(inode); > else > free_inode_nonrcu(inode); > } > > IOW, ->free_inode() is RCU-delayed part of ->destroy_inode(). If both > are present, ->destroy_inode() will be called synchronously, followed > by ->free_inode() from RCU callback, so you can have both - moving just > the "finally mark for reuse" part into ->free_inode() would be OK. > Any blocking stuff (if any) can be left in ->destroy_inode()... >