On Wed, Oct 27, 2021 at 05:24:51PM -0700, Darrick J. Wong wrote: > On Tue, Oct 26, 2021 at 11:49:59PM -0700, Christoph Hellwig wrote: > > On Fri, Oct 22, 2021 at 08:52:55PM +0000, Jane Chu wrote: > > > Thanks - I try to be honest. As far as I can tell, the argument > > > about the flag is a philosophical argument between two views. > > > One view assumes design based on perfect hardware, and media error > > > belongs to the category of brokenness. Another view sees media > > > error as a build-in hardware component and make design to include > > > dealing with such errors. > > > > No, I don't think so. Bit errors do happen in all media, which is > > why devices are built to handle them. It is just the Intel-style > > pmem interface to handle them which is completely broken. > > Yeah, I agree, this takes me back to learning how to use DISKEDIT to > work around a hole punched in a file (with a pen!) in the 1980s... > > ...so would you happen to know if anyone's working on solving this > problem for us by putting the memory controller in charge of dealing > with media errors? > > > > errors in mind from start. I guess I'm trying to articulate why > > > it is acceptable to include the RWF_DATA_RECOVERY flag to the > > > existing RWF_ flags. - this way, pwritev2 remain fast on fast path, > > > and its slow path (w/ error clearing) is faster than other alternative. > > > Other alternative being 1 system call to clear the poison, and > > > another system call to run the fast pwrite for recovery, what > > > happens if something happened in between? > > > > Well, my point is doing recovery from bit errors is by definition not > > the fast path. Which is why I'd rather keep it away from the pmem > > read/write fast path, which also happens to be the (much more important) > > non-pmem read/write path. > > The trouble is, we really /do/ want to be able to (re)write the failed > area, and we probably want to try to read whatever we can. Those are > reads and writes, not {pre,f}allocation activities. This is where Dave > and I arrived at a month ago. > > Unless you'd be ok with a second IO path for recovery where we're > allowed to be slow? That would probably have the same user interface > flag, just a different path into the pmem driver. I just don't see how 4 single line branches to propage RWF_RECOVERY down to the hardware is in any way an imposition on the fast path. It's no different for passing RWF_HIPRI down to the hardware *in the fast path* so that the IO runs the hardware in polling mode because it's faster for some hardware. IOWs, saying that we shouldn't implement RWF_RECOVERY because it adds a handful of branches to the fast path is like saying that we shouldn't implement RWF_HIPRI because it slows down the fast path for non-polled IO.... Just factor the actual recovery operations out into a separate function like: static void noinline pmem_media_recovery(...) { } pmem_copy_from_iter() { if ((unlikely)(flag & RECOVERY)) pmem_media_recovery(...); return _copy_from_iter_flushcache(addr, bytes, i); } .... And there's basically zero overhead in the fast paths for normal data IO operations, whilst supporting a simple, easy to use data recovery IO operations for regions that have bad media... > Ha, how about a int fd2 = recoveryfd(fd); call where you'd get whatever > speshul options (retry raid mirrors! scrape the film off the disk if > you have to!) you want that can take forever, leaving the fast paths > alone? Why wouldn't we just pass RWF_RECOVERY down to a REQ_RECOVERY bio flag and have raid devices use that to trigger scraping whatever they can if there are errors? The io path through the VFS and filesystem to get the scraped data out to the user *is exactly the same*, so we're going to have to plumb this functionality into fast paths *somewhere along the line*. Really, I think this whole "flag propagation is too much overhead for the fast path" argument is completely invalid - if 4 conditional branches is too much overhead to add to the fast path, then we can't add *anything ever again* to the IO path because it has too much overhead and impact on the fast path. Cheers, Dave. -- Dave Chinner david@xxxxxxxxxxxxx -- dm-devel mailing list dm-devel@xxxxxxxxxx https://listman.redhat.com/mailman/listinfo/dm-devel