On 8/8/23, Eric W. Biederman <ebiederm@xxxxxxxxxxxx> wrote: > > Adding a couple more people. > > Mateusz Guzik <mjguzik@xxxxxxxxx> writes: > >> Making close(2) delegate fput finalization with task_work_add() runs >> into a slowdown (atomics needed to do it) which is artificially worsened >> in presence of rseq, which glibc blindly uses if present (and it is >> normally present) -- they added a user memory-touching handler into >> resume_user_mode_work(), where the thread leaving the kernel lands after >> issuing task_work_add() from fput(). Said touching requires a SMAP >> round-trip which is quite expensive and it always executes when landing >> in the resume routine. >> >> I'm going to write a separate e-mail about the rseq problem later, but >> even if it gets sorted out there is still perf to gain (or rather, >> overhead to avoid). >> >> Numbers are below in the proposed patch, but tl;dr without CONFIG_RSEQ >> making things worse for the stock kernel I see about 7% increase in >> ops/s with open+close. >> >> Searching mailing lists for discussions explaining why close(2) was not >> already doing this I found a patch with the easiest way out (call >> __fput_sync() in filp_close()): >> https://lore.kernel.org/all/20150831120525.GA31015@xxxxxxxxxx/ > > What you need to search for is probably the opposite why is > task_work_add used in close. > You are splitting hairs here. > Taking a quick look at the history it appears that fput was always > synchronous until a decade ago when commit 4a9d4b024a31 ("switch fput to > task_work_add") was merged. > > The next two commits 3ffa3c0e3f6e ("aio: now fput() is OK from interrupt > context; get rid of manual delayed __fput()") and commit 6120d3dbb122 > ("get rid of ->scm_work_list") seem to demonstrate why fput was made > asynchronous. They rely on the new fput behavior to break recursive > calls and to allow fput from any context. That plus as Al talks about > having any lock held over fput can potentially cause a deadlock. > > All 3 issues taken together says that a synchronous fput is a > loaded foot gun that must be used very carefully. That said > close(2) does seem to be a reliably safe place to be synchronous. > Benefits of fput not taking surprise sleepable locks (and only sometimes) and whatnot were pretty obvious and I'm not proposing changing random consumers back to __fput_sync equivalent. What is not obvious is if filp_close consumers, which already suffer a lot of work, forces them to be in a spot where __fput_sync is safe to do. The question was de facto brought up by Oleg's patch and did not get a response, I don't see any explanations in other places either. Cursory reading by me suggested it is indeed dodgy thus the proposal which does not alter filp_close semantics. As you stated yourself, the close syscall itself should be the safe spot here and I was surprised to find this was not sorted out already -- it genuinely looks like an oversight. > The big question is can your loop calling open then close going 7% > faster into any real world improvements? How much can it generalize? > In this context you mean how many other spots can use it? I expect none (apart from maybe close_range). But as I mention later, close is not an obscure syscall, it is used all the time. As for real world, I don't think anyone will get a marked win as is. So happens there is perf loss all over the kernel, here is perf top from the bench with task_work out of the way: 7.07% [kernel] [k] entry_SYSCALL_64 3.59% [kernel] [k] do_dentry_open 2.95% [kernel] [k] strncpy_from_user 2.93% [kernel] [k] kmem_cache_free 2.88% [kernel] [k] init_file 2.86% [kernel] [k] __call_rcu_common.constprop.0 2.72% [kernel] [k] kmem_cache_alloc 2.70% [kernel] [k] memcg_slab_post_alloc_hook 2.46% libc.so.6 [.] __GI___libc_open 2.37% [kernel] [k] mod_objcg_state 2.28% [kernel] [k] link_path_walk.part.0.constprop.0 2.20% [kernel] [k] apparmor_current_getsecid_subj 2.19% [kernel] [k] apparmor_file_open [snip] For example memory allocation/free is taking quite a bit of CPU time and I strongly suspect with enough hackery it can be significantly less expensive (it mostly bottlenecks on cmpxchg16b). That is to say, if one was to sort all other things out, there is several more percent to recover. I would agree with the concern if the patch was complex, but it is not. > Taking a look at close_fd, it is used in autofs, cachefiles, bpf, amoung > others. I think there is a very good argument that we can not say that > filep_close is always a safe place to call __fput_close. There is just > too much going on in some of those place. A particular possibly > dangerous example is cachefiles_ondemand_daemon_read which calls > complete after close_fd. If as Oleg suggested filp_close started always > calling __fput_sync that call to complete looks like a deadlock waiting > to happen. > I did not look at cachefiles specifically, brief look at others was indeed worrisome. But then again, they are not a factor with the proposed patch. > [snip] > Which is a long way of saying that this only looks safe for close(2). > Yep. Except is a highly popular syscall, so it's not like I'm proposing a change to facilitate something which happens once a year. > > Are there any real world gains if close(2) is the only place this > optimization can be applied? Is the added maintenance burden worth the > speed up? > I don't think the patch at hand adds any particular burden and can be easily modified to appease most concerns. > I would not export any of your new _sync variants to modules. They are > all loaded foot-guns. > Specifics, including glaring warning signs and some commentary to the extent "don't use it" can be trivially added if dodging task_work by close is considered fine by people maintaining the code. Ultimately the real question I asked in my e-mail was if close(2) not being exempt from task_work was *intentional* (it does not look as such). Writing a patch which only changes semantics for this syscall (and not for anyone else) is trivial and there are numerous ways to do it, I only shipped one I almost had to write for benchmarking anyway. Frankly I expected changing the behavior for close(2) to be a no-brainer, I did expect pushback on the way the patch is implemented. That is to say, my willingness to argue about this is not particularly high, so if people insist on not patching close I'm going to drop this thread sooner than later. -- Mateusz Guzik <mjguzik gmail.com>