On Mon, Aug 22, 2022 at 08:52 AM -07, Eric Dumazet wrote: > On Fri, Aug 19, 2022 at 3:37 AM Jakub Sitnicki <jakub@xxxxxxxxxxxxxx> wrote: >> >> >> On Fri, Aug 19, 2022 at 10:01 AM GMT, liujian (CE) wrote: >> >> -----Original Message----- >> >> From: Jakub Sitnicki [mailto:jakub@xxxxxxxxxxxxxx] >> >> Sent: Friday, August 19, 2022 4:39 PM >> >> To: liujian (CE) <liujian56@xxxxxxxxxx>; john.fastabend@xxxxxxxxx; >> >> edumazet@xxxxxxxxxx >> >> Cc: davem@xxxxxxxxxxxxx; yoshfuji@xxxxxxxxxxxxxx; dsahern@xxxxxxxxxx; >> >> kuba@xxxxxxxxxx; pabeni@xxxxxxxxxx; andrii@xxxxxxxxxx; mykolal@xxxxxx; >> >> ast@xxxxxxxxxx; daniel@xxxxxxxxxxxxx; martin.lau@xxxxxxxxx; >> >> song@xxxxxxxxxx; yhs@xxxxxx; kpsingh@xxxxxxxxxx; sdf@xxxxxxxxxx; >> >> haoluo@xxxxxxxxxx; jolsa@xxxxxxxxxx; shuah@xxxxxxxxxx; >> >> bpf@xxxxxxxxxxxxxxx >> >> Subject: Re: [PATCH bpf-next 1/2] sk_msg: Keep reference on socket file >> >> while wait_memory >> >> >> >> On Mon, Aug 15, 2022 at 10:33 AM +08, Liu Jian wrote: >> >> > Fix the below NULL pointer dereference: >> >> > >> >> > [ 14.471200] Call Trace: >> >> > [ 14.471562] <TASK> >> >> > [ 14.471882] lock_acquire+0x245/0x2e0 >> >> > [ 14.472416] ? remove_wait_queue+0x12/0x50 >> >> > [ 14.473014] ? _raw_spin_lock_irqsave+0x17/0x50 >> >> > [ 14.473681] _raw_spin_lock_irqsave+0x3d/0x50 >> >> > [ 14.474318] ? remove_wait_queue+0x12/0x50 >> >> > [ 14.474907] remove_wait_queue+0x12/0x50 >> >> > [ 14.475480] sk_stream_wait_memory+0x20d/0x340 >> >> > [ 14.476127] ? do_wait_intr_irq+0x80/0x80 >> >> > [ 14.476704] do_tcp_sendpages+0x287/0x600 >> >> > [ 14.477283] tcp_bpf_push+0xab/0x260 >> >> > [ 14.477817] tcp_bpf_sendmsg_redir+0x297/0x500 >> >> > [ 14.478461] ? __local_bh_enable_ip+0x77/0xe0 >> >> > [ 14.479096] tcp_bpf_send_verdict+0x105/0x470 >> >> > [ 14.479729] tcp_bpf_sendmsg+0x318/0x4f0 >> >> > [ 14.480311] sock_sendmsg+0x2d/0x40 >> >> > [ 14.480822] ____sys_sendmsg+0x1b4/0x1c0 >> >> > [ 14.481390] ? copy_msghdr_from_user+0x62/0x80 >> >> > [ 14.482048] ___sys_sendmsg+0x78/0xb0 >> >> > [ 14.482580] ? vmf_insert_pfn_prot+0x91/0x150 >> >> > [ 14.483215] ? __do_fault+0x2a/0x1a0 >> >> > [ 14.483738] ? do_fault+0x15e/0x5d0 >> >> > [ 14.484246] ? __handle_mm_fault+0x56b/0x1040 >> >> > [ 14.484874] ? lock_is_held_type+0xdf/0x130 >> >> > [ 14.485474] ? find_held_lock+0x2d/0x90 >> >> > [ 14.486046] ? __sys_sendmsg+0x41/0x70 >> >> > [ 14.486587] __sys_sendmsg+0x41/0x70 >> >> > [ 14.487105] ? intel_pmu_drain_pebs_core+0x350/0x350 >> >> > [ 14.487822] do_syscall_64+0x34/0x80 >> >> > [ 14.488345] entry_SYSCALL_64_after_hwframe+0x63/0xcd >> >> > >> >> > The test scene as following flow: >> >> > thread1 thread2 >> >> > ----------- --------------- >> >> > tcp_bpf_sendmsg >> >> > tcp_bpf_send_verdict >> >> > tcp_bpf_sendmsg_redir sock_close >> >> > tcp_bpf_push_locked __sock_release >> >> > tcp_bpf_push //inet_release >> >> > do_tcp_sendpages sock->ops->release >> >> > sk_stream_wait_memory // tcp_close >> >> > sk_wait_event sk->sk_prot->close >> >> > release_sock(__sk); >> >> > *** >> >> > >> >> > lock_sock(sk); >> >> > __tcp_close >> >> > sock_orphan(sk) >> >> > sk->sk_wq = NULL >> >> > release_sock >> >> > **** >> >> > lock_sock(__sk); >> >> > remove_wait_queue(sk_sleep(sk), &wait); >> >> > sk_sleep(sk) >> >> > //NULL pointer dereference >> >> > &rcu_dereference_raw(sk->sk_wq)->wait >> >> > >> >> > While waiting for memory in thread1, the socket is released with its >> >> > wait queue because thread2 has closed it. This caused by >> >> > tcp_bpf_send_verdict didn't increase the f_count of psock->sk_redir- >> >> >sk_socket->file in thread1. >> >> >> >> I'm not sure about this approach. Keeping a closed sock file alive, just so we >> >> can wakeup from sleep, seems like wasted effort. >> >> >> >> __tcp_close sets sk->sk_shutdown = RCV_SHUTDOWN | SEND_SHUTDOWN. >> >> So we will return from sk_stream_wait_memory via the do_error path. >> >> >> >> SEND_SHUTDOWN might be set because socket got closed and orphaned - >> >> dead and detached from its file, like in this case. >> >> >> >> So, IMHO, we should check if SOCK_DEAD flag is set on wakeup due to >> >> SEND_SHUTDOWN in sk_stream_wait_memory, before accessing the wait >> >> queue. >> >> >> >> [...] >> > As jakub's approach, this problem can be solved. >> > >> > diff --git a/include/net/sock.h b/include/net/sock.h >> > index a7273b289188..a3dab7140f1e 100644 >> > --- a/include/net/sock.h >> > +++ b/include/net/sock.h >> > @@ -1998,6 +1998,8 @@ static inline void sk_set_socket(struct sock *sk, struct socket *sock) >> > static inline wait_queue_head_t *sk_sleep(struct sock *sk) >> > { >> > BUILD_BUG_ON(offsetof(struct socket_wq, wait) != 0); >> > + if (sock_flag(sk, SOCK_DEAD)) >> > + return NULL; >> > return &rcu_dereference_raw(sk->sk_wq)->wait; >> > } >> > /* Detach socket from process context. >> > diff --git a/kernel/sched/wait.c b/kernel/sched/wait.c >> > index 9860bb9a847c..da1be17d0b19 100644 >> > --- a/kernel/sched/wait.c >> > +++ b/kernel/sched/wait.c >> > @@ -51,6 +51,8 @@ void remove_wait_queue(struct wait_queue_head *wq_head, struct wait_queue_entry >> > { >> > unsigned long flags; >> > >> > + if (wq_head == NULL) >> > + return; >> > spin_lock_irqsave(&wq_head->lock, flags); >> > __remove_wait_queue(wq_head, wq_entry); >> > spin_unlock_irqrestore(&wq_head->lock, flags); >> >> I don't know if we want to change the contract for sk_sleep() >> remove_wait_queue() so that they accept dead sockets or nulls. >> >> How about just: >> >> diff --git a/net/core/stream.c b/net/core/stream.c >> index ccc083cdef23..1105057ce00a 100644 >> --- a/net/core/stream.c >> +++ b/net/core/stream.c >> @@ -159,7 +159,8 @@ int sk_stream_wait_memory(struct sock *sk, long *timeo_p) >> *timeo_p = current_timeo; >> } >> out: >> - remove_wait_queue(sk_sleep(sk), &wait); >> + if (!sock_flag(sk, SOCK_DEAD)) >> + remove_wait_queue(sk_sleep(sk), &wait); >> return err; >> >> do_error: >> > > OK, but what about tcp_msg_wait_data() and udp_msg_wait_data () ? > It seems they could be vulnerable as well.. TBH haven't thought about these until now. They are reachable only via sk->sk_prot->recvmsg, so from recvmsg(2) or recvmmsg(2). syscall layer will keep the socket file and the wait queue alive.