Re: [PATCH 13/18] io_uring: add file set registration

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On 2/12/19 5:29 AM, Alan Jenkins wrote:
> On 08/02/2019 15:13, Jens Axboe wrote:
>> On 2/8/19 7:02 AM, Alan Jenkins wrote:
>>> On 08/02/2019 12:57, Jens Axboe wrote:
>>>> On 2/8/19 5:17 AM, Alan Jenkins wrote:
>>>>>> +static int io_sqe_files_scm(struct io_ring_ctx *ctx)
>>>>>> +{
>>>>>> +#if defined(CONFIG_NET)
>>>>>> +	struct scm_fp_list *fpl = ctx->user_files;
>>>>>> +	struct sk_buff *skb;
>>>>>> +	int i;
>>>>>> +
>>>>>> +	skb =  __alloc_skb(0, GFP_KERNEL, 0, NUMA_NO_NODE);
>>>>>> +	if (!skb)
>>>>>> +		return -ENOMEM;
>>>>>> +
>>>>>> +	skb->sk = ctx->ring_sock->sk;
>>>>>> +	skb->destructor = unix_destruct_scm;
>>>>>> +
>>>>>> +	fpl->user = get_uid(ctx->user);
>>>>>> +	for (i = 0; i < fpl->count; i++) {
>>>>>> +		get_file(fpl->fp[i]);
>>>>>> +		unix_inflight(fpl->user, fpl->fp[i]);
>>>>>> +		fput(fpl->fp[i]);
>>>>>> +	}
>>>>>> +
>>>>>> +	UNIXCB(skb).fp = fpl;
>>>>>> +	skb_queue_head(&ctx->ring_sock->sk->sk_receive_queue, skb);
>>>>> This code sounds elegant if you know about the existence of unix_gc(),
>>>>> but quite mysterious if you don't.  (E.g. why "inflight"?)  Could we
>>>>> have a brief comment, to comfort mortal readers on their journey?
>>>>>
>>>>> /* A message on a unix socket can hold a reference to a file. This can
>>>>> cause a reference cycle. So there is a garbage collector for unix
>>>>> sockets, which we hook into here. */
>>>> Yes that's a good idea, I've added a comment as to why we go through the
>>>> trouble of doing this socket + skb dance.
>>> Great, thanks.
>>>
>>>>> I think this is bypassing too_many_unix_fds() though?  I understood that
>>>>> was intended to bound kernel memory allocation, at least in principle.
>>>> As the code stands above, it'll cap it at 253. I'm just now reworking it
>>>> to NOT be limited to the SCM max fd count, but still impose a limit of
>>>> 1024 on the number of registered files. This is important to cap the
>>>> memory allocation attempt as well.
>>> I saw you were limiting to SCM_MAX_FD per io_uring.  On the other hand,
>>> there's no specific limit on the number of io_urings you can open (only
>>> the standard limits on fds).  So this would let you allocate hundreds of
>>> times more files than the previous limit RLIMIT_NOFILE...
>> But there is, the io_uring itself is under the memlock rlimit.
>>
>>> static inline bool too_many_unix_fds(struct task_struct *p)
>>> {
>>> 	struct user_struct *user = current_user();
>>>
>>> 	if (unlikely(user->unix_inflight > task_rlimit(p, RLIMIT_NOFILE)))
>>> 		return !capable(CAP_SYS_RESOURCE) && !capable(CAP_SYS_ADMIN);
>>> 	return false;
>>> }
>>>
>>> RLIMIT_NOFILE is technically per-task, but here it is capping
>>> unix_inflight per-user.  So the way I look at this, the number of file
>>> descriptors per user is bounded by NOFILE * NPROC.  Then
>>> user->unix_inflight can have one additional process' worth (NOFILE) of
>>> "inflight" files.  (Plus SCM_MAX_FD slop, because too_many_fds() is only
>>> called once per SCM_RIGHTS).
>>>
>>> Because io_uring doesn't check too_many_unix_fds(), I think it will let
>>> you have about 253 (or 1024) more process' worth of open files. That
>>> could be big proportionally when RLIMIT_NPROC is low.
>>>
>>> I don't know if it matters.  It maybe reads like an oversight though.
>>>
>>> (If it does matter, it might be cleanest to change too_many_unix_fds()
>>> to get rid of the "slop".  Since that may be different between af_unix
>>> and io_uring; 253 v.s. 1024 or whatever. E.g. add a parameter for the
>>> number of inflight files we want to add.)
>> I don't think it matters. The files in the fixed file set have already
>> been opened by the application, so it counts towards the number of open
>> files that is allowed to have. I don't think we should impose further
>> limits on top of that.
> 
> A process can open one io_uring and 199 other files.  Register the 199 
> files in the io_uring, then close their file descriptors.  The main 
> NOFILE limit only counts file descriptors.  So then you can open one 
> io_uring, 198 other files, and repeat.
> 
> You're right, I had forgotten the memlock limit on io_uring.  That makes 
> it much less of a practical problem.
> 
> But it raises a second point.  It's not just that it lets users allocate 
> more files.  You might not want to be limited by user->unix_inflight.  
> But you are calling unix_inflight(), which increments it!  Then if 
> unix->inflight exceeds the NOFILE limit, you will avoid seeing any 
> errors with io_uring, but the user will not be able to send files over 
> unix sockets.
> 
> So I think this is confusing to read, and confusing to troubleshoot if 
> the limit is ever hit.
> 
> I would be happy if io_uring didn't increment user->unix_inflight.  I'm 
> not sure what the best way is to arrange that.

How about we just do something like the below? I think that's the saner
approach, rather than bypass user->unix_inflight. It's literally the
same thing.


diff --git a/fs/io_uring.c b/fs/io_uring.c
index a4973af1c272..5196b3aa935e 100644
--- a/fs/io_uring.c
+++ b/fs/io_uring.c
@@ -2041,6 +2041,13 @@ static int __io_sqe_files_scm(struct io_ring_ctx *ctx, int nr, int offset)
 	struct sk_buff *skb;
 	int i;
 
+	if (!capable(CAP_SYS_RESOURCE) && !capable(CAP_SYS_ADMIN)) {
+		struct user_struct *user = ctx->user;
+
+		if (user->unix_inflight > task_rlimit(current, RLIMIT_NOFILE))
+			return -EMFILE;
+	}
+
 	fpl = kzalloc(sizeof(*fpl), GFP_KERNEL);
 	if (!fpl)
 		return -ENOMEM;


-- 
Jens Axboe




[Index of Archives]     [Linux RAID]     [Linux SCSI]     [Linux ATA RAID]     [IDE]     [Linux Wireless]     [Linux Kernel]     [ATH6KL]     [Linux Bluetooth]     [Linux Netdev]     [Kernel Newbies]     [Security]     [Git]     [Netfilter]     [Bugtraq]     [Yosemite News]     [MIPS Linux]     [ARM Linux]     [Linux Security]     [Device Mapper]

  Powered by Linux