On 12/15/23 02:22, Jaegeuk Kim wrote: > On 12/14, Christoph Hellwig wrote: >> On Wed, Dec 13, 2023 at 08:41:32AM -0800, Jaegeuk Kim wrote: >>> I don't have any >>> concern to keep the same ioprio on writes, since handheld devices are mostly >>> sensitive to reads. So, if you have other use-cases using zoned writes which >>> require different ioprio on writes, I think you can suggest a knob to control >>> it by users. >> >> Get out of your little handheld world. In Linux we need a generally usable >> I/O stack, and any feature exposed by the kernel and will be used quite >> differently than you imagine. >> >> Just like people will add reordering to the I/O stack that's not there >> right now (in addition to the ones your testing doesn't hit). That >> doensn't mean we should avoid them - you genereally get better performance >> by not reordering without a good reason (like thotting), but especially >> in error handling paths or resource constrained environment they will >> hapen all over. We've had this whole discussion with the I/O barriers >> that did not work for exactly the same reasons. >> >>> >>>> >>>>> it is essential to place the data per file to get better bandwidth. And for >>>>> NAND-based storage, filesystem is the right place to deal with the more efficient >>>>> garbage collecion based on the known data locations. >>>> >>>> And that works perfectly fine match for zone append. >>> >>> How that works, if the device gives random LBAs back to the adjacent data in >>> a file? And, how to make the LBAs into the sequential ones back? >> >> Why would your device pick random LBAs? If you send a zone append to >> zone it will be written at the write pointer, which is absolutely not >> random. All I/O written in a single write is going to be sequential, >> so just like for all other devices doing large sequential writes is >> important. Multiple writes can get reordered, but if you havily hit >> the same zone you'd get the same effect in the file system allocator >> too. > > How can you guarantee the device does not give any random LBAs? What'd> be the selling point of zone append to end users? Are you sure this can > give the better write trhought forever? Have you considered how to > implement this in device side such as FTL mapping overhead and garbage > collection leading to tail latencies? Answers to all your questions, in order: 1) Asking this is to me similar to asking how can one guarantee that the device gives back the data that was written... You are asking for guarantees that the device is not buggy. By definition, zone append will return the writen start LBA within the zone that the zone append command specified. And that start LBA will always be equal to the zone write pointer value when the device started executing the zone append command. 2) When there is an FS, the user cannot know if the FS is using zone append or not, so the user should not care at all. If by "user" you mean "the file system", then it is a design decision. We already pointed out that generally speaking, zone append will be easier to use because it does not have ordering constraints. 3) Yes, because the writes are always sequential, which is the least expensive pattern for the device internal as that only triggers minimal internal activity on the FTL, GC, weir leveling etc, at least for a decently designed device. 4) See above. If the device interface forces the device user to always write sequentially, as mandated by a zoned device, then FTL, GC and weir leveling is minimized. The actual device internal GC that may or may not happen completely depend on how the device maps zones to flash super blocks. If the mapping is 1:1, then GC will be nearly non-existent. If the mapping is not 1:1, then GC overhead may exist. The discussion should then be about the design choices of the device. The fact that the host chose zone append will not in anyway make things worse for the device. Even with regular writes the host must write sequentially, same as what zone append achieves (potentially a lot more easily). > My takeaway on the two approaches would be: > zone_append zone_write > ----------- ---------- > LBA from FTL from filesystem > FTL mapping Page-map Zone-map Not sure what you mean here. zone append always returns an LBA from within the zone specified by the LBA in the command CDB. So mapping is still per zone. zone append is *NOT* a random write command. Zone append automatically implements sequential writing within a zone for the user. In the case of regular writes, the user must fully control sentimentality. In both cases the write pattern *is* sequential. > SRAM/DRAM needs Large Small There are no differences in this area because the FTL is the same for both. No changes, nothing special for zone append. > FTL GC Required Not required Incorrect. See above. That depends on the device mapping of zones to flash superblocks. And GC requirements are the same for both because the write pattern is identical: it is sequential within each zone being written. The user still controls which zone it wants to write. Zone append is not a magic command that chooses a target zone automatically. > Tail latencies Exist Not exisit Incorrect. They are the same and because of the lack of ordering requirement with zone append, if anything, zone append can give better latency. > GC Efficience Worse Better Nope. See above. Same. > Longevity As-is Longer > Discard cmd Required Not required There is no discard with zone devices. Only zone reset. So both are "not required" here. > Block complexity Small Large > Failure cases Less exist Exist > Fsck Don't know F2FS-TOOLS support > Filesystem BTRFS support(?) F2FS support Yes, btrfs data path uses zone append. > > Given this, I took zone_write, especially for mobile devices, since we can > recover the unaligned writes in the corner cases by fsck. And, most benefit > would be getting rid of FTL mapping overhead which improves random read IOPs > significantly due to the lack of SRAM in low-end storages. And, longer lifetime > by mitigating garbage collection overhead is more important in mobile world. As mentioned, GC is not an issue, or rather, GC depends on how the device is designed, not on which type of write command the host uses. Writes are always sequential for both types ! -- Damien Le Moal Western Digital Research