On Fri, Nov 1, 2024 at 2:09 PM Pavel Begunkov <asml.silence@xxxxxxxxx> wrote: > > On 11/1/24 20:06, Mina Almasry wrote: > ... > >> +__maybe_unused > >> +static const struct memory_provider_ops io_uring_pp_zc_ops; > >> + > >> +static inline struct io_zcrx_area *io_zcrx_iov_to_area(const struct net_iov *niov) > >> +{ > >> + struct net_iov_area *owner = net_iov_owner(niov); > >> + > >> + return container_of(owner, struct io_zcrx_area, nia); > >> +} > >> + > > > > We discussed this before I disappeared on vacation but I'm not too > > convinced to be honest, sorry. > > > > It's invalid to call io_zcrx_iov_to_area on a devmem niov and vice > > versa, right? So current and future code has to be very careful to > > Yes > > > call the right helpers on the right niovs. > > > > At the very least there needs to be a comment above all these > > container_of helpers: > > > > /* caller must have verified that this niov is devmem/io_zcrx */. > > > > However I feel like even a comment is extremely error prone. These > > container_of's are inside of the call stack of some helpers. I would > > say we need a check. If we're concerned about performance, the check > > can be behind DEBUG_NET_WARN_ON(), although even that is a bit iffy, > > but could be fine. Doing this without a check seems too risky to me. > > No, it doesn't need a check nor it needs a comment. The very > essence of virtual function tables is that they're coupled > together with objects for which those function make sense and > called only for those objects. The only way to get here with > invalid net_iovs is to take one page pool and feed it with > net_iovs from other another page pool that won't be sane in > the first place. > That could happen. In fact the whole devmem tcp paths are very carefully written to handle that net_iovs are allocated from the page_pool, put in skbs, and then sit in the sk receive queue. In pathological cases (user is re/misconfiguring flow steering) we can have 1 sk receive queue that has a mix of page skbs, devmem skbs, and io_uring skbs, and other skbs. Code that is processing the skbs in the receive queue has no idea whether what kind of skb it is. That's why that code needs to check whether the skb has readable frags, and that's why in this very series you needed to add a check in tcp_recvmsg_dmabuf to make sure that its a dmabuf skb, and you need to add a check to io_zcrx_recv_frag that the frag inside it is io_uring niov. The code would be wrong without it. All I'm trying to say is that it's very error prone to rely on folks writing and reviewing code to check that whenever dmabuf/io_rcrx/etc handling is done, somewhere in the call stack a type verification check has been made, and a DEBUG_NET_WARN could help avoid some subtle memory corruption bugs. > That would be an equivalent of: > > struct file *f1 = ...; > struct file *f2 = ...; > > f1->f_op->read(f2, ...); > > Maybe it looks strange for you in C, but it's same as putting > comments that a virtual function that it should be called only > for objects of that class: > > struct A { > virtual void foo() = 0; > }; > struct B: public A { > void foo() override { > // we should only be called with objects of type > // struct B (or anything inheriting it), check that > if (!reinterpret_cast<struct B*>(this)) > throw; > ... > } > } > > I'm not really sure I followed here. We do not get any type of compiler or type safety from this code because the dma-buf niovs and io_uring niovs are the same net_iov type. We can get type safety by defining new types for dmabuf_net_iov and io_uring_net_iov, then provide helpers: dmabuf_net_iov *net_iov_to_dmabuf(); io_uring_net_iov *net_iov_to_io_uring(); The helpers can check the niov is of the right type once and do a cast, then the object with the specific type can be passed to all future heplers without additional checks. This is one way to do it I guess. > >> static int io_allocate_rbuf_ring(struct io_zcrx_ifq *ifq, > >> struct io_uring_zcrx_ifq_reg *reg) > >> { > >> @@ -99,6 +114,9 @@ static int io_zcrx_create_area(struct io_ring_ctx *ctx, > >> goto err; > >> > >> for (i = 0; i < nr_pages; i++) { > >> + struct net_iov *niov = &area->nia.niovs[i]; > >> + > >> + niov->owner = &area->nia; > >> area->freelist[i] = i; > >> } > >> > >> @@ -230,3 +248,200 @@ void io_shutdown_zcrx_ifqs(struct io_ring_ctx *ctx) > >> { > >> lockdep_assert_held(&ctx->uring_lock); > >> } > >> + > >> +static bool io_zcrx_niov_put(struct net_iov *niov, int nr) > >> +{ > >> + return atomic_long_sub_and_test(nr, &niov->pp_ref_count); > >> +} > >> + > >> +static bool io_zcrx_put_niov_uref(struct net_iov *niov) > >> +{ > >> + if (atomic_long_read(&niov->pp_ref_count) < IO_ZC_RX_UREF) > >> + return false; > >> + > >> + return io_zcrx_niov_put(niov, IO_ZC_RX_UREF); > >> +} > >> + > > > > Sorry, I have to push back a bit against this. The refcounting of > > netmem is already complicated. the paged netmem has 2 refcounts and > > care needs to be taken when acquiring and dropping refcounts. net_iov > > inherited the pp_ref_count but not the paged refcount, and again need > > some special handling. skb_frag_unref takes very special care checking > > Which is why it's using net_iovs. > > > pp->recycle, is_pp_netmem, and others to figure out the correct > > pp->recycle has nothing to do with the series. We don't add > it in any special way, and if it's broken it's broken even > for non-proivder buffers. > > > refcount to put based on the type of the netmem and skb flag. > > Just same as with the ->[un]readable flag, which is not > functionally needed, and if it's screwed many things can > go very wrong. > > > This code ignores all these generic code > > skb_frag_unref/napi_pp_put_page/etc paths and uses raw access to > > I don't see the point, they are not used because they're not > needed. Instead of checking whether it came from a page pool > and whether it's net_iov or not, in the path io_uring returns > it we already apriori know that they're from a specific page > pool, net_iov and from the current provider. > > Same for optimisations provided by those helpers, they are > useful when you're transferring buffers from one context to > another, e.g. task recieve path -> napi / page_pool. In this > case they're already fetched in the right context without any > need to additionally jumping through the hoops. If anything, > it'd be odd to jump out of a window to climb a rope on the > other side of the building when you could've just walked 5 > meters to the other room. > For me, "they are not used because they're not needed." is not enough justification to ignore the generic code paths that support generic use cases and add your own freeing path and recycling that needs to work adjacent to generic paths for posterity. You need to provide concrete reasons why the current code paths don't work for you and can't be made to work for you. Is it very complicated to napi_pp_put_page() niovs as the user puts them in the refill queue without adding a new syscall? If so, is it possible to do a niov equivalent of page_pool_put_page_bulk() of the refill queue while/as you process the RX path? If you've tested the generic code paths to be performance deficient and your recycling is indeed better, you could improve the page_pool to pull netmems when it needs to like you're doing here, but in a generic way that applies to the page allocator and other providers. Not a one-off implementation that only applies to your provider. If you're absolutely set on ignoring the currently supported reffing and implementing your own reffing and recycling for your use case, sure, that could work, but please don't overload the niov->pp_ref_count reserved for the generic use cases for this. Add io_zcrx_area->io_uring_ref or something and do whatever you want with it. Since it's not sharing the pp_ref_count with the generic code paths I don't see them conflicting in the future. -- Thanks, Mina