On Tue, Nov 14, 2017 at 10:46:34PM +0800, Xin Long wrote: > On Tue, Nov 14, 2017 at 10:31 PM, Neil Horman <nhorman@xxxxxxxxxxxxx> wrote: > > On Mon, Nov 13, 2017 at 11:49:28PM +0800, Xin Long wrote: > >> On Mon, Nov 13, 2017 at 11:40 PM, Xin Long <lucien.xin@xxxxxxxxx> wrote: > >> > On Mon, Nov 13, 2017 at 11:23 PM, Neil Horman <nhorman@xxxxxxxxxxxxx> wrote: > >> >> On Mon, Nov 13, 2017 at 01:47:58PM +0800, Xin Long wrote: > >> >>> Commit dfcb9f4f99f1 ("sctp: deny peeloff operation on asocs with threads > >> >>> sleeping on it") fixed the race between peeloff and wait sndbuf by > >> >>> checking waitqueue_active(&asoc->wait) in sctp_do_peeloff(). > >> >>> > >> >>> But it actually doesn't work as even if waitqueue_active returns false > >> >>> the waiting sndbuf thread may still not yet hold sk lock. > >> >>> > >> >>> This patch is to fix this by adding wait_buf flag in asoc, and setting it > >> >>> before going the waiting loop, clearing it until the waiting loop breaks, > >> >>> and checking it in sctp_do_peeloff instead. > >> >>> > >> >>> Fixes: dfcb9f4f99f1 ("sctp: deny peeloff operation on asocs with threads sleeping on it") > >> >>> Suggested-by: Marcelo Ricardo Leitner <marcelo.leitner@xxxxxxxxx> > >> >>> Signed-off-by: Xin Long <lucien.xin@xxxxxxxxx> > >> >>> --- > >> >>> include/net/sctp/structs.h | 1 + > >> >>> net/sctp/socket.c | 4 +++- > >> >>> 2 files changed, 4 insertions(+), 1 deletion(-) > >> >>> > >> >>> diff --git a/include/net/sctp/structs.h b/include/net/sctp/structs.h > >> >>> index 0477945..446350e 100644 > >> >>> --- a/include/net/sctp/structs.h > >> >>> +++ b/include/net/sctp/structs.h > >> >>> @@ -1883,6 +1883,7 @@ struct sctp_association { > >> >>> > >> >>> __u8 need_ecne:1, /* Need to send an ECNE Chunk? */ > >> >>> temp:1, /* Is it a temporary association? */ > >> >>> + wait_buf:1, > >> >>> force_delay:1, > >> >>> prsctp_enable:1, > >> >>> reconf_enable:1; > >> >>> diff --git a/net/sctp/socket.c b/net/sctp/socket.c > >> >>> index 6f45d17..1b2c78c 100644 > >> >>> --- a/net/sctp/socket.c > >> >>> +++ b/net/sctp/socket.c > >> >>> @@ -4946,7 +4946,7 @@ int sctp_do_peeloff(struct sock *sk, sctp_assoc_t id, struct socket **sockp) > >> >>> /* If there is a thread waiting on more sndbuf space for > >> >>> * sending on this asoc, it cannot be peeled. > >> >>> */ > >> >>> - if (waitqueue_active(&asoc->wait)) > >> >>> + if (asoc->wait_buf) > >> >>> return -EBUSY; > >> >>> > >> >>> /* An association cannot be branched off from an already peeled-off > >> >>> @@ -7835,6 +7835,7 @@ static int sctp_wait_for_sndbuf(struct sctp_association *asoc, long *timeo_p, > >> >>> /* Increment the association's refcnt. */ > >> >>> sctp_association_hold(asoc); > >> >>> > >> >>> + asoc->wait_buf = 1; > >> >>> /* Wait on the association specific sndbuf space. */ > >> >>> for (;;) { > >> >>> prepare_to_wait_exclusive(&asoc->wait, &wait, > >> >>> @@ -7860,6 +7861,7 @@ static int sctp_wait_for_sndbuf(struct sctp_association *asoc, long *timeo_p, > >> >>> } > >> >>> > >> >>> out: > >> >>> + asoc->wait_buf = 0; > >> >>> finish_wait(&asoc->wait, &wait); > >> >>> > >> >>> /* Release the association's refcnt. */ > >> >>> -- > >> >>> 2.1.0 > >> >>> > >> >>> > >> >> > >> >> This doesn't make much sense to me, as it appears to be prone to aliasing. That > >> >> is to say: > >> >> > >> >> a) If multiple tasks are queued waiting in sctp_wait_for_sndbuf, the first > >> >> thread to exit that for(;;) loop will clean asoc->wait_buf, even though others > >> >> may be waiting on it, allowing sctp_do_peeloff to continue when it shouldn't be > >> > You're right, we talked about this before using waitqueue_active in > >> > earlier time. > >> > I didn't remember this somehow. Sorry. > >> > > >> >> > >> >> b) In the case of a single task blocking in sct_wait_for_sendbuf, checking > >> >> waitqueue_active is equally good, because it returns true, until such time as > >> >> finish_wait is called anyway. > >> > waitqueue_active can not work here, because in sctp_wait_for_sndbuf(): > >> > ... > >> > release_sock(sk); > >> > current_timeo = schedule_timeout(current_timeo); <-----[a] > >> > lock_sock(sk); > >> > If another thread wakes up asoc->wait, it will be removed from > >> > this wait queue, you check DEFINE_WAIT, the callback autoremove_wake_function > >> > will do this removal in wake_up(). > >> > > >> > I guess we need to think about another to fix this. > >> maybe we can use > >> DEFINE_WAIT_FUNC(wait, woken_wake_function); > >> instead of DEFINE_WAIT(wait) here ? > >> > > I'm still not sure I see the problem here. If we have the following situation: > > * Exec context A is executing in sctp_do_peeloff, about to check > > waitqueue_active() > > * Exec context B is blocking in sctp_wait_for_sndbuf(), specifically without the > > socket lock held > > > > > > Then, we have two possibilities: > > > > a) Context A executes waitqueue_active, which returns true, implying that > > context B is still on the queue, or that some other undescribed context has > > begun waiting on the queue. In either case, the behavior is correct, in that > > the peeloff is denied. > > > > b) Context B is woken up (and in the most pessimal case, has its waitq entry > > removed from queue immediately, causing context B to have waitequeue_active > > return false, allowing it to continue processing the peeloff. Since it holds > > the socket lock however, context B will block on the lock_sock operation until > > such time as the peeloff completes, so you're safe. > > > > About the only issue that I see (and as I write this, I may be seeing what you > > are actually trying to fix here) is that, during the period where context A is > > sleeping in sctp_wait_for_sendbuf, with the socket lock released, it is possible > > for an sctp_do_peeloff operation to complete, meaning that assoc->base.sk might > > point to a new socket, allowing each context to hold an independent socket lock > > and execute in parallel. To combat that, I think all you really need is some > > code in sctp_wait_for_sndbuf that looks like this: > > > > release_sock(sk); > > current_timeo = schedule_timeout(current_timeo); > > lock_sock(sk); > > > > if (sk != asoc->base.sk) { > > /* a socket peeloff occured */ > > release_sock(sk); > > sk = assoc->base.sk; > > lock_sock(sk); > > } > > > > *timeo_p = current_timeo; > > > > > > Does that make sense? This way, you lock the 'old' socket lock to ensure that > > the peeloff operation is completed, then you check to see if the socket has > > changed. If it has, you migrate your socket to the new, peeled off one and > > continue your space availability check > Yes, you got what I'm trying to fix in this patch exactly. :-) > and the fix you proposed above is doable, but incomplete, > we also need to change the sk pointer in sctp_sendmsg: > @@ -1962,7 +1962,7 @@ static int sctp_sendmsg(struct sock *sk, struct > msghdr *msg, size_t msg_len) > > timeo = sock_sndtimeo(sk, msg->msg_flags & MSG_DONTWAIT); > if (!sctp_wspace(asoc)) { > - err = sctp_wait_for_sndbuf(asoc, &timeo, msg_len); > + err = sctp_wait_for_sndbuf(asoc, &timeo, msg_len, &sk); > if (err) { > if (err == -ESRCH) { > /* asoc is already dead; */ > @@ -7828,7 +7828,7 @@ void sctp_sock_rfree(struct sk_buff *skb) > > /* Helper function to wait for space in the sndbuf. */ > static int sctp_wait_for_sndbuf(struct sctp_association *asoc, long *timeo_p, > - size_t msg_len) > + size_t msg_len, struct sock **orig_sk) > { > struct sock *sk = asoc->base.sk; > int err = 0; > @@ -7862,11 +7862,17 @@ static int sctp_wait_for_sndbuf(struct > sctp_association *asoc, long *timeo_p, > release_sock(sk); > current_timeo = schedule_timeout(current_timeo); > lock_sock(sk); > + if (sk != asoc->base.sk) { > + release_sock(sk); > + sk = asoc->base.sk; > + lock_sock(sk); > + } > > *timeo_p = current_timeo; > } > > out: > + *orig_sk = sk; > finish_wait(&asoc->wait, &wait); > > > right ? > Yes, that makes sense, post that as a proper commit please, I'll support that. Neil -- To unsubscribe from this list: send the line "unsubscribe linux-sctp" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html