When we write all cq items to tx, we have to wait for a new event based on poll to indicate that it is writable. But the current writability is triggered based on whether tx is full or not, and In fact, when tx is dissatisfied, the user of cq's item may not necessarily get it, because it may still be occupied by the network card. In this case, we need to know when cq is available, so this patch adds a socket option, When the user configures this option using setsockopt, when cq is available, a readable event is generated for all xsk bound to this umem. I can't find a better description of this event, I think it can also be 'readable', although it is indeed different from the 'readable' of the new data. But the overhead of xsk checking whether cq or rx is readable is small. Signed-off-by: Xuan Zhuo <xuanzhuo@xxxxxxxxxxxxxxxxx> --- include/net/xdp_sock.h | 1 + include/uapi/linux/if_xdp.h | 1 + net/xdp/xsk.c | 28 ++++++++++++++++++++++++++++ 3 files changed, 30 insertions(+) diff --git a/include/net/xdp_sock.h b/include/net/xdp_sock.h index 1a9559c..faf5b1a 100644 --- a/include/net/xdp_sock.h +++ b/include/net/xdp_sock.h @@ -49,6 +49,7 @@ struct xdp_sock { struct xsk_buff_pool *pool; u16 queue_id; bool zc; + bool cq_event; enum { XSK_READY = 0, XSK_BOUND, diff --git a/include/uapi/linux/if_xdp.h b/include/uapi/linux/if_xdp.h index a78a809..2dba3cb 100644 --- a/include/uapi/linux/if_xdp.h +++ b/include/uapi/linux/if_xdp.h @@ -63,6 +63,7 @@ struct xdp_mmap_offsets { #define XDP_UMEM_COMPLETION_RING 6 #define XDP_STATISTICS 7 #define XDP_OPTIONS 8 +#define XDP_CQ_EVENT 9 struct xdp_umem_reg { __u64 addr; /* Start of packet data area */ diff --git a/net/xdp/xsk.c b/net/xdp/xsk.c index cfbec39..0c53403 100644 --- a/net/xdp/xsk.c +++ b/net/xdp/xsk.c @@ -285,7 +285,16 @@ void __xsk_map_flush(void) void xsk_tx_completed(struct xsk_buff_pool *pool, u32 nb_entries) { + struct xdp_sock *xs; + xskq_prod_submit_n(pool->cq, nb_entries); + + rcu_read_lock(); + list_for_each_entry_rcu(xs, &pool->xsk_tx_list, tx_list) { + if (xs->cq_event) + sock_def_readable(&xs->sk); + } + rcu_read_unlock(); } EXPORT_SYMBOL(xsk_tx_completed); @@ -495,6 +504,9 @@ static __poll_t xsk_poll(struct file *file, struct socket *sock, __xsk_sendmsg(sk); } + if (xs->cq_event && pool->cq && !xskq_prod_is_empty(pool->cq)) + mask |= EPOLLIN | EPOLLRDNORM; + if (xs->rx && !xskq_prod_is_empty(xs->rx)) mask |= EPOLLIN | EPOLLRDNORM; if (xs->tx && !xskq_cons_is_full(xs->tx)) @@ -882,6 +894,22 @@ static int xsk_setsockopt(struct socket *sock, int level, int optname, mutex_unlock(&xs->mutex); return err; } + case XDP_CQ_EVENT: + { + int cq_event; + + if (optlen < sizeof(cq_event)) + return -EINVAL; + if (copy_from_sockptr(&cq_event, optval, sizeof(cq_event))) + return -EFAULT; + + if (cq_event) + xs->cq_event = true; + else + xs->cq_event = false; + + return 0; + } default: break; } -- 1.8.3.1