Matthew Auld <matthew.auld@xxxxxxxxx> writes: > Try to document the object caching related bits, like cache_coherent and > cache_dirty. > > v2(Ville): > - As pointed out by Ville, fix the completely incorrect assumptions > about the "partial" coherency on shared LLC platforms. > > Suggested-by: Daniel Vetter <daniel.vetter@xxxxxxxx> > Signed-off-by: Matthew Auld <matthew.auld@xxxxxxxxx> > Cc: Ville Syrjälä <ville.syrjala@xxxxxxxxxxxxxxx> > Cc: Mika Kuoppala <mika.kuoppala@xxxxxxxxxxxxxxx> > --- > .../gpu/drm/i915/gem/i915_gem_object_types.h | 173 +++++++++++++++++- > drivers/gpu/drm/i915/i915_drv.h | 9 - > 2 files changed, 169 insertions(+), 13 deletions(-) > > diff --git a/drivers/gpu/drm/i915/gem/i915_gem_object_types.h b/drivers/gpu/drm/i915/gem/i915_gem_object_types.h > index ef3de2ae9723..a809424bc8c1 100644 > --- a/drivers/gpu/drm/i915/gem/i915_gem_object_types.h > +++ b/drivers/gpu/drm/i915/gem/i915_gem_object_types.h > @@ -92,6 +92,76 @@ struct drm_i915_gem_object_ops { > const char *name; /* friendly name for debug, e.g. lockdep classes */ > }; > > +/** > + * enum i915_cache_level - The supported GTT caching values for system memory > + * pages. > + * > + * These translate to some special GTT PTE bits when binding pages into some > + * address space. It also determines whether an object, or rather its pages are > + * coherent with the GPU, when also reading or writing through the CPU cache > + * with those pages. > + * > + * Userspace can also control this through struct drm_i915_gem_caching. > + */ > +enum i915_cache_level { > + /** > + * @I915_CACHE_NONE: > + * > + * Not coherent with the CPU cache. If the cache is dirty and we need > + * the underlying pages to be coherent with some later GPU access then > + * we need to manually flush the pages. > + * > + * Note that on shared LLC platforms reads and writes through the CPU > + * cache are still coherent even with this setting. See also > + * &drm_i915_gem_object.cache_coherent for more details. > + * > + * Note that on platforms with a shared LLC this should ideally only be > + * used for scanout surfaces, otherwise we end up over-flushing in some > + * places. > + */ > + I915_CACHE_NONE = 0, > + /** > + * @I915_CACHE_LLC: > + * > + * Coherent with the CPU cache. If the cache is dirty, then the GPU will > + * ensure that access remains coherent, when both reading and writing > + * through the CPU cache. > + * > + * Not used for scanout surfaces. > + * > + * Applies to both platforms with shared LLC(HAS_LLC), and snooping > + * based platforms(HAS_SNOOP). > + * > + * This should be the default for platforms which share the LLC with the > + * CPU. The only exception is scanout objects, where the display engine > + * is not coherent with the LLC. For such objects I915_CACHE_NONE or > + * I915_CACHE_WT should be used. > + */ > + I915_CACHE_LLC, > + /** > + * @I915_CACHE_L3_LLC: > + * > + * Explicitly enable the Gfx L3 cache, with snooped LLC. > + * > + * The Gfx L3 sits between the domain specific caches, e.g > + * sampler/render caches, and the larger LLC. LLC is coherent with the > + * GPU, but L3 is only visible to the GPU, so likely needs to be flushed > + * when the workload completes. > + * > + * Not used for scanout surfaces. > + * > + * Only exposed on some gen7 + GGTT. More recent hardware has dropped > + * this. > + */ This is stellar. Thanks! -Mika > + I915_CACHE_L3_LLC, > + /** > + * @I915_CACHE_WT: > + * > + * hsw:gt3e Write-through for scanout buffers. > + */ > + I915_CACHE_WT, > +}; > + > enum i915_map_type { > I915_MAP_WB = 0, > I915_MAP_WC, > @@ -228,14 +298,109 @@ struct drm_i915_gem_object { > unsigned int mem_flags; > #define I915_BO_FLAG_STRUCT_PAGE BIT(0) /* Object backed by struct pages */ > #define I915_BO_FLAG_IOMEM BIT(1) /* Object backed by IO memory */ > - /* > - * Is the object to be mapped as read-only to the GPU > - * Only honoured if hardware has relevant pte bit > + /** > + * @cache_level: The desired GTT caching level. > + * > + * See enum i915_cache_level for possible values, along with what > + * each does. > */ > unsigned int cache_level:3; > - unsigned int cache_coherent:2; > + /** > + * @cache_coherent: > + * > + * Track whether the pages are coherent with the GPU if reading or > + * writing through the CPU caches. The largely depends on the > + * @cache_level setting. > + * > + * On platforms which don't have the shared LLC(HAS_SNOOP), like on Atom > + * platforms, coherency must be explicitly requested with some special > + * GTT caching bits(see enum i915_cache_level). When enabling coherency > + * it does come at a performance and power cost on such platforms. On > + * the flip side the kernel does need to manually flush any buffers > + * which need to be coherent with the GPU, if the object is not > + * coherent i.e @cache_coherent is zero. > + * > + * On platforms that share the LLC with the CPU(HAS_LLC), all GT memory > + * access will automatically snoop the CPU caches(even with CACHE_NONE). > + * The one exception is when dealing with the display engine, like with > + * scanout surfaces. To handle this the kernel will always flush the > + * surface out of the CPU caches when preparing it for scanout. Also > + * note that since scanout surfaces are only ever read by the display > + * engine we only need to care about flushing any writes through the CPU > + * cache, reads on the other hand will always be coherent. > + * > + * Something strange here is why @cache_coherent is not a simple > + * boolean, i.e coherent vs non-coherent. The reasoning for this is back > + * to the display engine not being fully coherent. As a result scanout > + * surfaces will either be marked as I915_CACHE_NONE or I915_CACHE_WT. > + * In the case of seeing I915_CACHE_NONE the kernel makes the assumption > + * that this is likely a scanout surface, and will set @cache_coherent > + * as only I915_BO_CACHE_COHERENT_FOR_READ, on platforms with the shared > + * LLC. The kernel uses this to avoid flushing reads, while then also > + * applying some optimisations to always flush writes through the CPU > + * cache as early as possible, where it can, in effect keeping > + * @cache_dirty clean, so we can potentially avoid stalling when > + * flushing the surface just before doing the scanout. This does mean > + * we might unnecessarily flush non-scanout objects in some places, but > + * the default assumption is that all normal objects should be using > + * I915_CACHE_LLC, at least on platforms with the shared LLC. > + * > + * I915_BO_CACHE_COHERENT_FOR_READ: > + * > + * When reading through the CPU cache, the GPU is still coherent. Reads > + * through the CPU cache only become a concern when writes can bypass > + * the CPU cache. > + * > + * As an example, if some object is mapped on the CPU with write-back > + * caching, and we read some page, then the cache likely now contains > + * the data from that read. At this point the cache and main memory > + * match up, so all good. But next the GPU needs to write some data to > + * that same page. Now if the @cache_level is I915_CACHE_NONE and the > + * the platform doesn't have the shared LLC, then the GPU will > + * effectively skip invalidating the cache(or however that works > + * internally) when writing the new value. This is really bad since the > + * GPU has just written some new data to main memory, but the CPU cache > + * is still valid and now contains stale data. As a result the next time > + * we do a cached read with the CPU, we are rewarded with stale data. > + * Likewise if the cache is later flushed, we might be rewarded with > + * overwriting main memory with stale data. > + * > + * I915_BO_CACHE_COHERENT_FOR_WRITE: > + * > + * When writing through the CPU cache, the GPU is still coherent. Note > + * that this also implies I915_BO_CACHE_COHERENT_FOR_READ. > + */ > #define I915_BO_CACHE_COHERENT_FOR_READ BIT(0) > #define I915_BO_CACHE_COHERENT_FOR_WRITE BIT(1) > + unsigned int cache_coherent:2; > + > + /** > + * @cache_dirty: > + * > + * Track if we are we dirty with writes through the CPU cache for this > + * object. As a result reading directly from main memory might yield > + * stale data. > + * > + * This also ties into whether the kernel is tracking the object as > + * coherent with the GPU, as per @cache_coherent, as it determines if > + * flushing might be needed at various points. > + * > + * Another part of @cache_dirty is managing flushing when first > + * acquiring the pages for system memory, at this point the pages are > + * considered foreign, so the default assumption is that the cache is > + * dirty, for example the page zeroing done by the kernel might leave > + * writes though the CPU cache, or swapping-in, while the actual data in > + * main memory is potentially stale. Note that this is a potential > + * security issue when dealing with userspace objects and zeroing. Now, > + * whether we actually need apply the big sledgehammer of flushing all > + * the pages on acquire depends on if @cache_coherent is marked as > + * I915_BO_CACHE_COHERENT_FOR_WRITE, i.e that the GPU will be coherent > + * for both reads and writes though the CPU cache. > + * > + * Note that on shared LLC platforms we still apply the heavy flush for > + * I915_CACHE_NONE objects, under the assumption that this is going to > + * be used for scanout. > + */ > unsigned int cache_dirty:1; > > /** > diff --git a/drivers/gpu/drm/i915/i915_drv.h b/drivers/gpu/drm/i915/i915_drv.h > index f99b6c0dd068..ac144d0c69a5 100644 > --- a/drivers/gpu/drm/i915/i915_drv.h > +++ b/drivers/gpu/drm/i915/i915_drv.h > @@ -394,15 +394,6 @@ struct drm_i915_display_funcs { > void (*read_luts)(struct intel_crtc_state *crtc_state); > }; > > -enum i915_cache_level { > - I915_CACHE_NONE = 0, > - I915_CACHE_LLC, /* also used for snoopable memory on non-LLC */ > - I915_CACHE_L3_LLC, /* gen7+, L3 sits between the domain specifc > - caches, eg sampler/render caches, and the > - large Last-Level-Cache. LLC is coherent with > - the CPU, but L3 is only visible to the GPU. */ > - I915_CACHE_WT, /* hsw:gt3e WriteThrough for scanouts */ > -}; > > #define I915_COLOR_UNEVICTABLE (-1) /* a non-vma sharing the address space */ > > -- > 2.26.3