On 08/05/2016 10:35 PM, Shaun Tancheff wrote: > On Tue, Aug 2, 2016 at 8:29 PM, Damien Le Moal <Damien.LeMoal@xxxxxxxx> wrote: >> Hannes, Shaun, >> >> Let me add some more comments. >> >>> On Aug 2, 2016, at 23:35, Hannes Reinecke <hare@xxxxxxx> wrote: >>> >>> On 08/01/2016 07:07 PM, Shaun Tancheff wrote: >>>> On Mon, Aug 1, 2016 at 4:41 AM, Christoph Hellwig <hch@xxxxxx> wrote: >>>>> >>>>> Can you please integrate this with Hannes series so that it uses >>>>> his cache of the zone information? >>>> >>>> Adding Hannes and Damien to Cc. >>>> >>>> Christoph, >>>> >>>> I can make a patch the marshal Hannes' RB-Tree into to a block report, that is >>>> quite simple. I can even have the open/close/reset zone commands update the >>>> RB-Tree .. the non-private parts anyway. I would prefer to do this around the >>>> CONFIG_SD_ZBC support, offering the existing type of patch for setups that do >>>> not need the RB-Tree to function with zoned media. > > I have posted patches to integrate with the zone cache, hopefully they > make sense. > [ .. ] >>> I have thought about condensing the RB tree information, but then I >>> figured that for 'real' SMR handling we cannot assume all zones are of >>> fixed size, and hence we need all the information there. >>> Any condensing method would assume a given structure of the zones, which >>> the standard just doesn't provide. >>> Or am I missing something here? >> >> Indeed, the standards do not mandate any particular zone configuration, >> constant zone size, etc. So writing code so that can be handled is certainly >> the right way of doing things. However, if we decide to go forward with >> mapping RESET WRITE POINTER command to DISCARD, then at least a constant >> zone size (minus the last zone as you said) must be assumed, and that >> information can be removed from the entries in the RB tree (as it will be >> saved for the sysfs "zone_size" file anyway. Adding a little code to handle >> that eventual last runt zone with a different size is not a big problem. > >>> As for write pointer handling: yes, the write pointer on the zones is >>> not really useful for upper-level usage. >>> Where we do need it is to detect I/O which is crossing the write pointer >>> (eg when doing reads over the entire zone). >>> As per spec you will be getting an I/O error here, so we need to split >>> the I/O on the write pointer to get valid results back. >> >> To be precise here, the I/O splitting will be handled by the block layer >> thanks to the "chunk_sectors" setting. But that relies on a constant zone >> size assumption too. >> >> The RB-tree here is most useful for reads over or after the write pointer as >> this can have different behavior on different drives (URSWRZ bit). The RB-tree >> allows us to hide these differences to upper layers and simplify support at >> those levels. > > It is unfortunate that path was chosen rather than returning Made Up Data. But how would you return made up data without knowing that you _have_ to generate some? Of course it's trivial to generate made up data once an I/O error occurred, but that's too late as the I/O error already happened. The general idea behind this is that I _really_ do want to avoid triggering an I/O error on initial access. This kind of thing really tends to freak out users (connect a new drive and getting I/O errors; not the best user experience ...) > However I don't think this is a file system or device mapper problem as > neither of those layers attempt to read blocks they have not written > (or discarded). > All I can think of something that probes the drive media for a partition table > or other identification...isn't the conventional space sufficient to cover > those cases? Upon device scan the kernel will attempt to read the partition tables, which consists of reading the first few megabytes and the last sector (for GPT shadow partition tables). And depending on the driver manufacturer you might or might not have enough conventional space at the right locations. > Anyway you could just handle the error and sense codes [CHECK CONDITION / > ILLEGAL REQUEST / ATTEMPT TO READ INVALID DATA] and return MUD > when URSWRZ is 0. It would have the same effect as using the zone cache > without the possibility of the zone cache being out-of-sync. > Yes, but then we would already have registered an I/O error. And as indicated above, I really do _not_ want to handle all the customer calls for supposed faulty new SMR drives. >> I too agree that the write pointer value is not very useful to upper layers >> (e.g. FS). What matters most of the times for these layers is an "allocation >> pointer" or "submission pointer" which indicates the LBA to use for the next >> BIO submission. That LBA will most of the time be in advance of the zone WP >> because of request queing in the block scheduler. > > Which is kind of my point. > >> Note that from what we have done already, in many cases, upper layers do not >> even need this (e.g. F2FS, btrfs) and work fine without needing *any* >> zone->private_data because that "allocation pointer" information generally >> already exists in a different form (e.g. a bit in a bitmap). > > Agreed. Log structured and copy on write file systems are a natural fit for > these types of drives > >> For these cases, not having the RB-tree of zones would force a lot >> more code to be added to file systems for SMR support. > > I don't understand how the zone cache (RB-tree) is "helping" F2FS or btrfs. > Certainly any of these could trigger the zone cache to be loaded at mkfs > and/or mount time? > >>>> This all tails into domain responsibility. With the RB-Tree doing half of the >>>> work and the ‘responsible’ domain handling the active path via private_data >>>> why have the split at all? It seems to be a double work to have second object >>>> tracking the first so that I/O scheduling can function. >>>> >>> Tracking the zone states via RB trees is mainly to handly host-managed >>> drives seamlessly. Without an RB tree we will be subjected to I/O errors >>> during boot, as eg with new drives we inevitably will try to read beyond >>> the write pointer, which in turn will generate I/O errors on certain drives. > > If the zone cache is needed to boot then clearly my objection to reading > in the zone report information on drive attach is unfounded. I was under > the impression that this was not the case. > >>> I do agree that there is no strict need to setup an RB tree for >>> host-aware drives; I can easily add an attribute/flag to disable RB >>> trees for those. >>> However, tracking zone information in the RB tree _dramatically_ speeds >>> up device initialisation; issuing a blkdev_discard() for the entire >>> drive will take _ages_ without it. >> >> And the RB-tree will also be very useful for speeding up report zones >> ioctls too. Unless we want those to force an update of the RB-tree information ? > > That is debatable. I would tend to argue for both. *If* there must be > a zone cache then reading from the zone cache is a must. Forcing and > update of the zone cache from the media seems like a reasonable thing > to be able to do. > > Also the zone report is 'slow' in that there is an overhead for the > report itself but > the number of zones per query can be quite large so 4 or 5 I/Os that > run into the > hundreds if milliseconds to cache the entire drive isn't really unworkable for > something that is used infrequently. > No, surely not. But one of the _big_ advantages for the RB tree is blkdev_discard(). Without the RB tree any mkfs program will issue a 'discard' for every sector. We will be able to coalesce those into one discard per zone, but we still need to issue one for _every_ zone. Which is (as indicated) really slow, and easily takes several minutes. With the RB tree we can short-circuit discards to empty zones, and speed up processing time dramatically. Sure we could be moving the logic into mkfs and friends, but that would require us to change the programs and agree on a library (libzbc?) which should be handling that. >>>> Finally is the error handling path when the RB-Tree encounters and error it >>>> attempts to requery the drive topology virtually guaranteeing that the >>>> private_data is now out-of-sync with the RB-Tree. Again this is something >>>> that can be better encapsulated in the bi_end_io to be informed of the >>>> failed I/O and schedule the appropriate recovery (including re-querying the >>>> zone information of the affected zone(s)). >>>> >>> Well, if we have an RB tree with write pointers of course we need to >>> re-read the zone information on failure (and I thought I did that?). >>> Plus the error information will be propagated via the usual mechanism, >>> so they need to take care of updating any information in ->private_data. >>> >>> I'm perfectly fine with integrating your patches for the various >>> blkdev_XX callouts and associated ioctls; Jens favours that approach, >>> too, so we should be combining those efforts. >> >> Can we agree on an interface ? >> Summarizing all the discussions we had, I am all in favor of the following: >> >> 1) A "zone_size" sysfs attribute to indicate that a drive is zoned: >> The already existing device type file can tell us if this is an host >> managed disk (type==20) or a host aware disk (type==0). Drive managed >> disks could also be detected, but since no information on their zone >> configuration can be obtained, lets treat those as regular drives. > > Since disk type == 0 for everything that isn't HM so I would prefer the > sysfs 'zoned' file just report if the drive is HA or HM. > Okay. So let's put in the 'zoned' attribute the device type: 'host-managed', 'host-aware', or 'device managed'. >> 2) Add ioctls for zone management: >> Report zones (get information from RB tree), reset zone (simple wrapper >> to ioctl for block discard), open zone, close zone and finish zone. That >> will allow mkfs like tools to get zone information without having to parse >> thousands of sysfs files (and can also be integrated in libzbc block backend >> driver for a unified interface with the direct SG_IO path for kernels without >> the ZBC code enabled). > > I can add finish zone ... but I really can't think of a use for it, myself. > Which is not the point. The standard defines this, so clearly someone found it a reasonable addendum. So let's add this for completeness. >> 3) Try to condense the blkzone data structure to save memory: >> I think that we can at the very least remove the zone length, and also >> may be the per zone spinlock too (a single spinlock and proper state flags can >> be used). > > I have a variant that is an array of descriptors that roughly mimics the > api from blk-zoned.c that I did a few months ago as an example. > I should be able to update that to the current kernel + patches. > Okay. If we restrict the in-kernel SMR drive handling to devices with identical zone sizes of course we can remove the zone length. And we can do away with the per-zone spinlock and use a global one instead. I will be updating my patchset accordingly. Cheers, Hannes -- Dr. Hannes Reinecke Teamlead Storage & Networking hare@xxxxxxx +49 911 74053 688 SUSE LINUX GmbH, Maxfeldstr. 5, 90409 Nürnberg GF: F. Imendörffer, J. Smithard, J. Guild, D. Upmanyu, G. Norton HRB 21284 (AG Nürnberg) -- To unsubscribe from this list: send the line "unsubscribe linux-block" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html