Re: [PATCH 0/6] kernfs: proposed locking and concurrency improvement

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On Wed, Jan 13, 2021 at 1:17 PM Ian Kent <raven@xxxxxxxxxx> wrote:
>
> On Mon, 2021-01-11 at 17:02 +0800, Fox Chen wrote:
> > On Mon, Jan 11, 2021 at 4:42 PM Ian Kent <raven@xxxxxxxxxx> wrote:
> > > On Mon, 2021-01-11 at 15:04 +0800, Fox Chen wrote:
> > > > On Mon, Jan 11, 2021 at 12:20 PM Ian Kent <raven@xxxxxxxxxx>
> > > > wrote:
> > > > > On Mon, 2021-01-11 at 11:19 +0800, Ian Kent wrote:
> > > > > > On Wed, 2021-01-06 at 10:38 +0800, Fox Chen wrote:
> > > > > > > Hi Ian,
> > > > > > >
> > > > > > > I am rethinking this problem. Can we simply use a global
> > > > > > > lock?
> > > > > > >
> > > > > > >  In your original patch 5, you have a global mutex
> > > > > > > attr_mutex
> > > > > > > to
> > > > > > > protect attr, if we change it to a rwsem, is it enough to
> > > > > > > protect
> > > > > > > both
> > > > > > > inode and attr while having the concurrent read ability?
> > > > > > >
> > > > > > > like this patch I submitted. ( clearly, I missed
> > > > > > > __kernfs_iattrs
> > > > > > > part,
> > > > > > > but just about that idea )
> > > > > > > https://lore.kernel.org/lkml/20201207084333.179132-1-foxhlchen@xxxxxxxxx/
> > > > > >
> > > > > > I don't think so.
> > > > > >
> > > > > > kernfs_refresh_inode() writes to the inode so taking a read
> > > > > > lock
> > > > > > will allow multiple processes to concurrently update it which
> > > > > > is
> > > > > > what we need to avoid.
> > > >
> > > > Oh, got it. I missed the inode part. my bad. :(
> > > >
> > > > > > It's possibly even more interesting.
> > > > > >
> > > > > > For example, kernfs_iop_rmdir() and kernfs_iop_mkdir() might
> > > > > > alter
> > > > > > the inode link count (I don't know if that would be the sort
> > > > > > of
> > > > > > thing
> > > > > > they would do but kernfs can't possibly know either). Both of
> > > > > > these
> > > > > > functions rely on the VFS locking for exclusion but the inode
> > > > > > link
> > > > > > count is updated in kernfs_refresh_inode() too.
> > > > > >
> > > > > > That's the case now, without any patches.
> > > > >
> > > > > So it's not so easy to get the inode from just the kernfs
> > > > > object
> > > > > so these probably aren't a problem ...
> > > >
> > > > IIUC only when dop->revalidate, iop->lookup being called, the
> > > > result
> > > > of rmdir/mkdir will be sync with vfs.
> > >
> > > Don't quite get what you mean here?
> > >
> > > Do you mean something like, VFS objects are created on user access
> > > to the file system. Given that user access generally means path
> > > resolution possibly followed by some operation.
> > >
> > > I guess those VFS objects will go away some time after the access
> > > but even thought the code looks like that should happen pretty
> > > quickly after I've observed that these objects stay around longer
> > > than expected. There wouldn't be any use in maintaining a least
> > > recently used list of dentry candidates eligible to discard.
> >
> > Yes, that is what I meant. I think the duration may depend on the
> > current ram pressure. though not quite sure, I'm still digging this
> > part of code.
> >
> > > > kernfs_node is detached from vfs inode/dentry to save ram.
> > > >
> > > > > > I'm not entirely sure what's going on in
> > > > > > kernfs_refresh_inode().
> > > > > >
> > > > > > It could be as simple as being called with a NULL inode
> > > > > > because
> > > > > > the dentry concerned is negative at that point. I haven't had
> > > > > > time to look closely at it TBH but I have been thinking about
> > > > > > it.
> > > >
> > > > um, It shouldn't be called with a NULL inode, right?
> > > >
> > > > inode->i_mode = kn->mode;
> > > >
> > > > otherwise will crash.
> > >
> > > Yes, you're right about that.
> > >
> > > > > Certainly this can be called without a struct iattr having been
> > > > > allocated ... and given it probably needs to remain a pointer
> > > > > rather than embedded in the node the inode link count update
> > > > > can't easily be protected from concurrent updates.
> > > > >
> > > > > If it was ok to do the allocation at inode creation the problem
> > > > > becomes much simpler to resolve but I thought there were
> > > > > concerns
> > > > > about ram consumption (although I don't think that was exactly
> > > > > what
> > > > > was said?).
> > > > >
> > > >
> > > > you meant iattr to be allocated at inode creation time??
> > > > yes, I think so. it's due to ram consumption.
> > >
> > > I did, yes.
> > >
> > > The actual problem is dealing with multiple concurrent updates to
> > > the inode link count, the rest can work.
>
> Umm ... maybe I've been trying to do this in the wrong place all
> along.
>
> You know the inode i_lock looks like the sensible thing to use to
> protect these updates.
>
> Something like this for that last patch should work:
>
> kernfs: use i_lock to protect concurrent inode updates
>
> From: Ian Kent <raven@xxxxxxxxxx>
>
> The inode operations .permission() and .getattr() use the kernfs node
> write lock but all that's needed is to keep the rb tree stable while
> updating the inode attributes as well as protecting the update itself
> against concurrent changes.
>
> And .permission() is called frequently during path walks and can cause
> quite a bit of contention between kernfs node opertations and path
> walks when the number of concurrant walks is high.
>
> To change kernfs_iop_getattr() and kernfs_iop_permission() to take
> the rw sem read lock instead of the write lock an addtional lock is
> needed to protect against multiple processes concurrently updating
> the inode attributes and link count in kernfs_refresh_inode().
>
> The inode i_lock seems like the sensible thing to use to protect these
> inode attribute updates so use it in kernfs_refresh_inode().
>
> Signed-off-by: Ian Kent <raven@xxxxxxxxxx>
> ---
>  fs/kernfs/inode.c |   10 ++++++----
>  1 file changed, 6 insertions(+), 4 deletions(-)
>
> diff --git a/fs/kernfs/inode.c b/fs/kernfs/inode.c
> index ddaf18198935..e26fa5115821 100644
> --- a/fs/kernfs/inode.c
> +++ b/fs/kernfs/inode.c
> @@ -171,6 +171,7 @@ static void kernfs_refresh_inode(struct kernfs_node *kn, struct inode *inode)
>  {
>         struct kernfs_iattrs *attrs = kn->iattr;
>
> +       spin_lock(inode->i_lock);
>         inode->i_mode = kn->mode;
>         if (attrs)
>                 /*
> @@ -181,6 +182,7 @@ static void kernfs_refresh_inode(struct kernfs_node *kn, struct inode *inode)
>
>         if (kernfs_type(kn) == KERNFS_DIR)
>                 set_nlink(inode, kn->dir.subdirs + 2);
> +       spin_unlock(inode->i_lock);
>  }
>
>  int kernfs_iop_getattr(const struct path *path, struct kstat *stat,
> @@ -189,9 +191,9 @@ int kernfs_iop_getattr(const struct path *path, struct kstat *stat,
>         struct inode *inode = d_inode(path->dentry);
>         struct kernfs_node *kn = inode->i_private;
>
> -       down_write(&kernfs_rwsem);
> +       down_read(&kernfs_rwsem);
>         kernfs_refresh_inode(kn, inode);
> -       up_write(&kernfs_rwsem);
> +       up_read(&kernfs_rwsem);
>
>         generic_fillattr(inode, stat);
>         return 0;
> @@ -281,9 +283,9 @@ int kernfs_iop_permission(struct inode *inode, int mask)
>
>         kn = inode->i_private;
>
> -       down_write(&kernfs_rwsem);
> +       down_read(&kernfs_rwsem);
>         kernfs_refresh_inode(kn, inode);
> -       up_write(&kernfs_rwsem);
> +       up_read(&kernfs_rwsem);
>
>         return generic_permission(inode, mask);
>  }
>

It looks good on my local machine, let me test my benchmark on a big machine. :)

Also, I wonder why i_lock?? what if I use a local spin_lock, will
there be any difference???

static void kernfs_refresh_inode(struct kernfs_node *kn, struct inode *inode)
{
        struct kernfs_iattrs *attrs = kn->iattr;
        static DEFINE_SPINLOCK(inode_lock);

        spin_lock(&inode_lock);
        inode->i_mode = kn->mode;
        if (attrs)
                /*
                 * kernfs_node has non-default attributes get them from
                 * persistent copy in kernfs_node.
                 */
                set_inode_attr(inode, attrs);

        if (kernfs_type(kn) == KERNFS_DIR)
                set_nlink(inode, kn->dir.subdirs + 2);
        spin_unlock(&inode_lock);
}



thanks,
fox



[Index of Archives]     [Linux Ext4 Filesystem]     [Union Filesystem]     [Filesystem Testing]     [Ceph Users]     [Ecryptfs]     [AutoFS]     [Kernel Newbies]     [Share Photos]     [Security]     [Netfilter]     [Bugtraq]     [Yosemite News]     [MIPS Linux]     [ARM Linux]     [Linux Security]     [Linux Cachefs]     [Reiser Filesystem]     [Linux RAID]     [Samba]     [Device Mapper]     [CEPH Development]

  Powered by Linux