Re: [PATCH 13/18] io_uring: add file set registration

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On 12/02/2019 17:33, Jens Axboe wrote:
On 2/12/19 10:21 AM, Alan Jenkins wrote:
On 12/02/2019 15:17, Jens Axboe wrote:
On 2/12/19 5:29 AM, Alan Jenkins wrote:
On 08/02/2019 15:13, Jens Axboe wrote:
On 2/8/19 7:02 AM, Alan Jenkins wrote:
On 08/02/2019 12:57, Jens Axboe wrote:
On 2/8/19 5:17 AM, Alan Jenkins wrote:
+static int io_sqe_files_scm(struct io_ring_ctx *ctx)
+{
+#if defined(CONFIG_NET)
+	struct scm_fp_list *fpl = ctx->user_files;
+	struct sk_buff *skb;
+	int i;
+
+	skb =  __alloc_skb(0, GFP_KERNEL, 0, NUMA_NO_NODE);
+	if (!skb)
+		return -ENOMEM;
+
+	skb->sk = ctx->ring_sock->sk;
+	skb->destructor = unix_destruct_scm;
+
+	fpl->user = get_uid(ctx->user);
+	for (i = 0; i < fpl->count; i++) {
+		get_file(fpl->fp[i]);
+		unix_inflight(fpl->user, fpl->fp[i]);
+		fput(fpl->fp[i]);
+	}
+
+	UNIXCB(skb).fp = fpl;
+	skb_queue_head(&ctx->ring_sock->sk->sk_receive_queue, skb);
This code sounds elegant if you know about the existence of unix_gc(),
but quite mysterious if you don't.  (E.g. why "inflight"?)  Could we
have a brief comment, to comfort mortal readers on their journey?

/* A message on a unix socket can hold a reference to a file. This can
cause a reference cycle. So there is a garbage collector for unix
sockets, which we hook into here. */
Yes that's a good idea, I've added a comment as to why we go through the
trouble of doing this socket + skb dance.
Great, thanks.

I think this is bypassing too_many_unix_fds() though?  I understood that
was intended to bound kernel memory allocation, at least in principle.
As the code stands above, it'll cap it at 253. I'm just now reworking it
to NOT be limited to the SCM max fd count, but still impose a limit of
1024 on the number of registered files. This is important to cap the
memory allocation attempt as well.
I saw you were limiting to SCM_MAX_FD per io_uring.  On the other hand,
there's no specific limit on the number of io_urings you can open (only
the standard limits on fds).  So this would let you allocate hundreds of
times more files than the previous limit RLIMIT_NOFILE...
But there is, the io_uring itself is under the memlock rlimit.

static inline bool too_many_unix_fds(struct task_struct *p)
{
	struct user_struct *user = current_user();

	if (unlikely(user->unix_inflight > task_rlimit(p, RLIMIT_NOFILE)))
		return !capable(CAP_SYS_RESOURCE) && !capable(CAP_SYS_ADMIN);
	return false;
}

RLIMIT_NOFILE is technically per-task, but here it is capping
unix_inflight per-user.  So the way I look at this, the number of file
descriptors per user is bounded by NOFILE * NPROC.  Then
user->unix_inflight can have one additional process' worth (NOFILE) of
"inflight" files.  (Plus SCM_MAX_FD slop, because too_many_fds() is only
called once per SCM_RIGHTS).

Because io_uring doesn't check too_many_unix_fds(), I think it will let
you have about 253 (or 1024) more process' worth of open files. That
could be big proportionally when RLIMIT_NPROC is low.

I don't know if it matters.  It maybe reads like an oversight though.

(If it does matter, it might be cleanest to change too_many_unix_fds()
to get rid of the "slop".  Since that may be different between af_unix
and io_uring; 253 v.s. 1024 or whatever. E.g. add a parameter for the
number of inflight files we want to add.)
I don't think it matters. The files in the fixed file set have already
been opened by the application, so it counts towards the number of open
files that is allowed to have. I don't think we should impose further
limits on top of that.
A process can open one io_uring and 199 other files.  Register the 199
files in the io_uring, then close their file descriptors.  The main
NOFILE limit only counts file descriptors.  So then you can open one
io_uring, 198 other files, and repeat.

You're right, I had forgotten the memlock limit on io_uring.  That makes
it much less of a practical problem.

But it raises a second point.  It's not just that it lets users allocate
more files.  You might not want to be limited by user->unix_inflight.
But you are calling unix_inflight(), which increments it!  Then if
unix->inflight exceeds the NOFILE limit, you will avoid seeing any
errors with io_uring, but the user will not be able to send files over
unix sockets.

So I think this is confusing to read, and confusing to troubleshoot if
the limit is ever hit.

I would be happy if io_uring didn't increment user->unix_inflight.  I'm
not sure what the best way is to arrange that.
How about we just do something like the below? I think that's the saner
approach, rather than bypass user->unix_inflight. It's literally the
same thing.


diff --git a/fs/io_uring.c b/fs/io_uring.c
index a4973af1c272..5196b3aa935e 100644
--- a/fs/io_uring.c
+++ b/fs/io_uring.c
@@ -2041,6 +2041,13 @@ static int __io_sqe_files_scm(struct io_ring_ctx *ctx, int nr, int offset)
   	struct sk_buff *skb;
   	int i;
+ if (!capable(CAP_SYS_RESOURCE) && !capable(CAP_SYS_ADMIN)) {
+		struct user_struct *user = ctx->user;
+
+		if (user->unix_inflight > task_rlimit(current, RLIMIT_NOFILE))
+			return -EMFILE;
+	}
+
   	fpl = kzalloc(sizeof(*fpl), GFP_KERNEL);
   	if (!fpl)
   		return -ENOMEM;


Welp, you gave me exactly what I asked for.  So now I'd better be
positive about it :-D.
;-)

I hope this will be documented accurately, at least where the EMFILE
result is explained for this syscall.
How's this:

http://git.kernel.dk/cgit/liburing/commit/?id=37e48698a09aa1e37690f8fa6dfd8da69a48ee60

+.B EMFILE
+.BR IORING_REGISTER_FILES
+was specified and adding
+.I nr_args
+file references would exceed the maximum allowed number of files the process
+is allowed to have according to the
+.B
+RLIMIT_NOFILE
+resource limit and the caller does not have
+.B CAP_SYS_RESOURCE
+capability.
+.TP

I was struggling with this.  The POSIX part of RLIMIT_NOFILE is applied per-process.  But the part we're talking about here, the Linux-specific "unix_inflight" resource, is actually accounted per-user.  It's like RLIMIT_NPROC.  The value of RLIMIT_NPROC is per-process, but the resource it limits is counted in user->processes.

This subtlety of the NOFILE limit is not made clear in the text above, nor in unix(7), nor in getrlimit(2).  I would interpret all these docs as saying this limit is a per-process thing - I think they are misleading.

IORING_MAX_FIXED_FILES is being raised to 1024, which is the same as the (soft limit) value for RLIMIT_NOFILE which the kernel sets for the init process.  I have an unjustifiable nervousness, that there will be some `fio` command, or a test written that maxes out IORING_REGISTER_FILES.  When you do that, it will provoke unexpected failures e.g. in GUI apps.  If we can't rule that out, the next best thing is a friendly man page.

Regards
Alan

Because EMFILE is different from the errno in af_unix.c, I will add a
wish for the existing documentation of ETOOMANYREFS in unix(7) to
reference this.

I'll stop bikeshedding there.  EMFILE sounds ok.  strerror() calls
ETOOMANYREFS "Too many references: cannot splice"; it doesn't seem to be
particularly helpful or well-known.
Agree





[Index of Archives]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]

  Powered by Linux