I can confirm this fixed the problem for us. Thanks a lot of the quick turnaround (as always!). Bye Norman > On 5. Sep 2020, at 10:26, Norman Maurer <norman.maurer@xxxxxxxxxxxxxx> wrote: > > Yes … I will :) I am already compiling the kernel as we speak with the patch applied. Will report back later today. > > > >> On 5. Sep 2020, at 10:24, nick@xxxxxxxxxxxx wrote: >> >> On 2020-09-04 22:50, Pavel Begunkov wrote: >>> On 05/09/2020 07:35, Jens Axboe wrote: >>>> On 9/4/20 9:57 PM, Jens Axboe wrote: >>>>> On 9/4/20 9:53 PM, Jens Axboe wrote: >>>>>> On 9/4/20 9:22 PM, nick@xxxxxxxxxxxx wrote: >>>>>>> Hi, >>>>>>> I am helping out with the netty io_uring integration, and came across >>>>>>> some strange behaviour which seems like it might be a bug related to >>>>>>> async offload of read/write iovecs. >>>>>>> Basically a WRITEV SQE seems to fail reliably with -BADADDRESS when the >>>>>>> IOSQE_ASYNC flag is set but works fine otherwise (everything else the >>>>>>> same). This is with 5.9.0-rc3. >>>>>> Do you see it just on 5.9-rc3, or also 5.8? Just curious... But that is >>>>>> very odd in any case, ASYNC writev is even part of the regular tests. >>>>>> Any sort of deferral, be it explicit via ASYNC or implicit through >>>>>> needing to retry, saves all the needed details to retry without >>>>>> needing any of the original context. >>>>>> Can you narrow down what exactly is being written - like file type, >>>>>> buffered/O_DIRECT, etc. What file system, what device is hosting it. >>>>>> The more details the better, will help me narrow down what is going on. >>>>> Forgot, also size of the IO (both total, but also number of iovecs in >>>>> that particular request. >>>>> Essentially all the details that I would need to recreate what you're >>>>> seeing. >>>> Turns out there was a bug in the explicit handling, new in the current >>>> -rc series. Can you try and add the below? >>> Hah, absolutely the same patch was in a series I was going to send >>> today, but with a note that it works by luck so not a bug. Apparently, >>> it is :) >>> BTW, const in iter->iov is guarding from such cases, yet another proof >>> that const casts are evil. >>>> diff --git a/fs/io_uring.c b/fs/io_uring.c >>>> index 0d7be2e9d005..000ae2acfd58 100644 >>>> --- a/fs/io_uring.c >>>> +++ b/fs/io_uring.c >>>> @@ -2980,14 +2980,15 @@ static inline int io_rw_prep_async(struct io_kiocb *req, int rw, >>>> bool force_nonblock) >>>> { >>>> struct io_async_rw *iorw = &req->io->rw; >>>> + struct iovec *iov; >>>> ssize_t ret; >>>> - iorw->iter.iov = iorw->fast_iov; >>>> - ret = __io_import_iovec(rw, req, (struct iovec **) &iorw->iter.iov, >>>> - &iorw->iter, !force_nonblock); >>>> + iorw->iter.iov = iov = iorw->fast_iov; >>>> + ret = __io_import_iovec(rw, req, &iov, &iorw->iter, !force_nonblock); >>>> if (unlikely(ret < 0)) >>>> return ret; >>>> + iorw->iter.iov = iov; >>>> io_req_map_rw(req, iorw->iter.iov, iorw->fast_iov, &iorw->iter); >>>> return 0; >>>> } >> >> Thanks for the speedy replies and finding/fixing this so fast! I'm new to kernel dev and haven't built my own yet but I think Norman is going to try out your patch soon. >> >> Nick >