From: Zbigniew Kempczyński <zbigniew.kempczynski@xxxxxxxxx> Different GENs supports different tile surfaces. Older GENs have HW fences to allow X / Y surface tiling / detiling. Newer GENs have to tile / detile such surface in software. To make test developer life easier this code adds buffer operations (short buf_ops) to use appropriate functions allowing copying linear buffer to BO and from BO to linear buffer regardless GPU generation and tiling/swizzling within BO. For GENs having fences support preference is to use them if they are available (X / Y tiling is probed on buf_ops initialization). Note: on Gen2 code supports only HW tiling at the moment. Signed-off-by: Zbigniew Kempczyński <zbigniew.kempczynski@xxxxxxxxx> Cc: Chris Wilson <chris@xxxxxxxxxxxxxxxxxx> Cc: Katarzyna Dec <katarzyna.dec@xxxxxxxxx> Reviewed-by: Chris Wilson <chris@xxxxxxxxxxxxxxxxxx> --- .../igt-gpu-tools/igt-gpu-tools-docs.xml | 1 + lib/Makefile.sources | 2 + lib/intel_bufops.c | 1157 +++++++++++++++++ lib/intel_bufops.h | 86 ++ lib/meson.build | 1 + 5 files changed, 1247 insertions(+) create mode 100644 lib/intel_bufops.c create mode 100644 lib/intel_bufops.h diff --git a/docs/reference/igt-gpu-tools/igt-gpu-tools-docs.xml b/docs/reference/igt-gpu-tools/igt-gpu-tools-docs.xml index 454d64e53..aa9fef205 100644 --- a/docs/reference/igt-gpu-tools/igt-gpu-tools-docs.xml +++ b/docs/reference/igt-gpu-tools/igt-gpu-tools-docs.xml @@ -42,6 +42,7 @@ <xi:include href="xml/igt_vgem.xml"/> <xi:include href="xml/igt_x86.xml"/> <xi:include href="xml/intel_batchbuffer.xml"/> + <xi:include href="xml/intel_bufops.xml"/> <xi:include href="xml/intel_chipset.xml"/> <xi:include href="xml/intel_io.xml"/> <xi:include href="xml/ioctl_wrappers.xml"/> diff --git a/lib/Makefile.sources b/lib/Makefile.sources index 5dd3962ee..da34eae73 100644 --- a/lib/Makefile.sources +++ b/lib/Makefile.sources @@ -67,6 +67,8 @@ lib_source_list = \ intel_aub.h \ intel_batchbuffer.c \ intel_batchbuffer.h \ + intel_bufops.c \ + intel_bufops.h \ intel_chipset.c \ intel_chipset.h \ intel_device_info.c \ diff --git a/lib/intel_bufops.c b/lib/intel_bufops.c new file mode 100644 index 000000000..804b2a0ae --- /dev/null +++ b/lib/intel_bufops.c @@ -0,0 +1,1157 @@ +/* + * Copyright © 2020 Intel Corporation + * + * Permission is hereby granted, free of charge, to any person obtaining a + * copy of this software and associated documentation files (the "Software"), + * to deal in the Software without restriction, including without limitation + * the rights to use, copy, modify, merge, publish, distribute, sublicense, + * and/or sell copies of the Software, and to permit persons to whom the + * Software is furnished to do so, subject to the following conditions: + * + * The above copyright notice and this permission notice (including the next + * paragraph) shall be included in all copies or substantial portions of the + * Software. + * + * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR + * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, + * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL + * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER + * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING + * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS + * IN THE SOFTWARE. + * + */ + +#include <sys/ioctl.h> +#include "igt.h" +#include "igt_x86.h" +#include "intel_bufops.h" + +/** + * SECTION:intel_bufops + * @short_description: Buffer operation on tiled surfaces + * @title: Buffer operations + * @include: igt.h + * + * # Buffer operations + * + * Intel GPU devices supports different set of tiled surfaces. + * Checking each time what tile formats are supports is cumbersome and + * error prone. + * + * Buffer operation (buf_ops) provide a wrapper to conditional code which + * can be used without worrying of implementation details giving: + * - copy linear to tiled buffer + * - copy tiled buffer to linear + * + * Following code order should be used (linear is plain memory with some + * image data): + * + * |[<!-- language="c" --> + * struct buf_ops *bops; + * struct intel_buf ibuf; + * ... + * bops = buf_ops_create(fd); + * intel_buf_init(bops, &ibuf, 512, 512, 32, I915_TILING_X, false); + * ... + * linear_to_intel_buf(bops, &ibuf, linear); + * ... + * intel_buf_to_linear(bops, &ibuf, linear); + * ... + * intel_buf_close(bops, &ibuf); + * ... + * buf_ops_destroy(bops); + * ]| + * + * Calling buf_ops_create(fd) probes hardware capabilities (supported fences, + * swizzling) and returns opaque pointer to buf_ops. From now on + * intel_buf_to_linear() and linear_to_intel_buf() will choose appropriate + * function to do the job. + * + * Note: bufops doesn't support SW tiling code yet. + */ + +//#define BUFOPS_DEBUGGING +#ifdef BUFOPS_DEBUGGING +#define DEBUG(...) printf(__VA_ARGS__) +#define DEBUGFN() DEBUG("\t -> %s()\n", __FUNCTION__) +#else +#define DEBUG(...) +#define DEBUGFN() +#endif + +#define TILE_DEF(x) (1 << (x)) +#define TILE_NONE TILE_DEF(I915_TILING_NONE) +#define TILE_X TILE_DEF(I915_TILING_X) +#define TILE_Y TILE_DEF(I915_TILING_Y) +#define TILE_Yf TILE_DEF(I915_TILING_Yf) +#define TILE_Ys TILE_DEF(I915_TILING_Ys) + +#define CCS_OFFSET(buf) (buf->aux.offset) +#define CCS_SIZE(gen, buf) \ + (intel_buf_aux_width(gen, buf) * intel_buf_aux_height(gen, buf)) + +typedef void (*bo_copy)(struct buf_ops *, struct intel_buf *, uint32_t *); + +struct buf_ops { + int fd; + int gen_start; + int gen_end; + int intel_gen; + uint32_t supported_tiles; + uint32_t supported_hw_tiles; + uint32_t swizzle_x; + uint32_t swizzle_y; + bo_copy linear_to; + bo_copy linear_to_x; + bo_copy linear_to_y; + bo_copy linear_to_yf; + bo_copy linear_to_ys; + bo_copy to_linear; + bo_copy x_to_linear; + bo_copy y_to_linear; + bo_copy yf_to_linear; + bo_copy ys_to_linear; +}; + +static const char *tiling_str(uint32_t tiling) +{ + switch (tiling) { + case I915_TILING_NONE: return "NONE"; + case I915_TILING_X: return "X"; + case I915_TILING_Y: return "Y"; + case I915_TILING_Yf: return "Yf"; + case I915_TILING_Ys: return "Ys"; + default: return "UNKNOWN"; + } +} + +static const char *bool_str(bool v) +{ + return v ? "yes" : "no"; +} + +static inline bool is_hw_tiling_supported(struct buf_ops *bops, uint32_t tiling) +{ + return bops->supported_hw_tiles & TILE_DEF(tiling); +} + +static inline bool is_tiling_supported(struct buf_ops *bops, uint32_t tiling) +{ + return bops->supported_tiles & TILE_DEF(tiling); +} + +static int __gem_get_tiling(int fd, struct drm_i915_gem_get_tiling *arg) +{ + int err; + + err = 0; + if (igt_ioctl(fd, DRM_IOCTL_I915_GEM_GET_TILING, arg)) { + err = -errno; + igt_assume(err); + } + errno = 0; + + return err; +} + +static bool __get_tiling(int fd, uint32_t handle, uint32_t *tiling, + uint32_t *swizzle) +{ + struct drm_i915_gem_get_tiling get_tiling = { .handle = handle }; + + if (__gem_get_tiling(fd, &get_tiling) != 0) + return false; + + *tiling = get_tiling.tiling_mode; + *swizzle = get_tiling.swizzle_mode; + igt_debug("buf tiling: %s, swizzle: %x, phys_swizzle: %x\n", + tiling_str(get_tiling.tiling_mode), + get_tiling.swizzle_mode, + get_tiling.phys_swizzle_mode); + + return get_tiling.phys_swizzle_mode == get_tiling.swizzle_mode; +} + +static int __set_tiling(int fd, uint32_t handle, uint32_t tiling, + uint32_t stride) +{ + do { + struct drm_i915_gem_set_tiling st; + int err; + + st.handle = handle; + st.tiling_mode = tiling; + st.stride = tiling ? stride : 0; + + err = 0; + if (ioctl(fd, DRM_IOCTL_I915_GEM_SET_TILING, &st)) + err = -errno; + errno = 0; + if (err != -EINTR) + return err; + } while (1); +} + +static void set_hw_tiled(struct buf_ops *bops, struct intel_buf *buf) +{ + if (buf->tiling != I915_TILING_X && buf->tiling != I915_TILING_Y) + return; + + if (!buf_ops_has_hw_fence(bops, buf->tiling)) + return; + + igt_assert_eq(__set_tiling(bops->fd, + buf->handle, buf->tiling, buf->stride), + 0); +} + +static unsigned long swizzle_bit(unsigned int bit, unsigned long offset) +{ + return (offset & (1ul << bit)) >> (bit - 6); +} + +static unsigned long swizzle_addr(void *ptr, uint32_t swizzle) +{ + unsigned long addr = to_user_pointer(ptr); + + switch (swizzle) { + case I915_BIT_6_SWIZZLE_NONE: + return addr; + case I915_BIT_6_SWIZZLE_9: + return addr ^ swizzle_bit(9, addr); + case I915_BIT_6_SWIZZLE_9_10: + return addr ^ swizzle_bit(9, addr) ^ swizzle_bit(10, addr); + case I915_BIT_6_SWIZZLE_9_11: + return addr ^ swizzle_bit(9, addr) ^ swizzle_bit(11, addr); + case I915_BIT_6_SWIZZLE_9_10_11: + return (addr ^ + swizzle_bit(9, addr) ^ + swizzle_bit(10, addr) ^ + swizzle_bit(11, addr)); + + case I915_BIT_6_SWIZZLE_UNKNOWN: + case I915_BIT_6_SWIZZLE_9_17: + case I915_BIT_6_SWIZZLE_9_10_17: + default: + igt_skip("physical swizzling mode impossible to handle in userspace\n"); + return addr; + } +} + +static void *x_ptr(void *ptr, + unsigned int x, unsigned int y, + unsigned int stride, unsigned int cpp) +{ + const int tile_width = 512; + const int tile_height = 8; + const int tile_size = tile_width * tile_height; + int offset_x, offset_y, pos; + int tile_x, tile_y; + + x *= cpp; + tile_x = x / tile_width; + tile_y = y / tile_height; + offset_x = tile_x * tile_size; + offset_y = tile_y * stride * tile_height; + + pos = (offset_y + (y % tile_height * tile_width) + + offset_x + (x % tile_width)); + + return ptr + pos; +} + +static void *y_ptr(void *ptr, + unsigned int x, unsigned int y, + unsigned int stride, unsigned int cpp) +{ + const int tile_width = 128; + const int tile_height = 32; + const int owords = 16; + const int tile_size = tile_width * tile_height; + int offset_x, offset_y, pos; + int shift_x, shift_y; + int tile_x, tile_y; + + x *= cpp; + tile_x = x / tile_width; + tile_y = y / tile_height; + offset_x = tile_x * tile_size; + offset_y = tile_y * stride * tile_height; + shift_x = x % owords + (x % tile_width) / owords * tile_width * cpp; + shift_y = y % tile_height * owords; + + pos = offset_y + offset_x + shift_x + shift_y; + + return ptr + pos; +} + +static void *yf_ptr(void *ptr, + unsigned int x, unsigned int y, + unsigned int stride, unsigned int cpp) +{ + const int tile_size = 4 * 1024; + const int tile_width = 128; + int row_size = stride / tile_width * tile_size; + + x *= cpp; /* convert to Byte offset */ + + /* + * Within a 4k Yf tile, the byte swizzling pattern is + * msb......lsb + * xyxyxyyyxxxx + * The tiles themselves are laid out in row major order. + */ + return ptr + + ((x & 0xf) * 1) + /* 4x1 pixels(32bpp) = 16B */ + ((y & 0x3) * 16) + /* 4x4 pixels = 64B */ + (((y & 0x4) >> 2) * 64) + /* 1x2 64B blocks */ + (((x & 0x10) >> 4) * 128) + /* 2x2 64B blocks = 256B block */ + (((y & 0x8) >> 3) * 256) + /* 2x1 256B blocks */ + (((x & 0x20) >> 5) * 512) + /* 2x2 256B blocks */ + (((y & 0x10) >> 4) * 1024) + /* 4x2 256 blocks */ + (((x & 0x40) >> 6) * 2048) + /* 4x4 256B blocks = 4k tile */ + (((x & ~0x7f) >> 7) * tile_size) + /* row of tiles */ + (((y & ~0x1f) >> 5) * row_size); +} + +typedef void *(*tile_fn)(void *, unsigned int, unsigned int, + unsigned int, unsigned int); +static tile_fn __get_tile_fn_ptr(int tiling) +{ + tile_fn fn = NULL; + + switch (tiling) { + case I915_TILING_X: + fn = x_ptr; + break; + case I915_TILING_Y: + fn = y_ptr; + break; + case I915_TILING_Yf: + fn = yf_ptr; + break; + case I915_TILING_Ys: + /* To be implemented */ + break; + } + + igt_require_f(fn, "Can't find tile function for tiling: %d\n", tiling); + return fn; +} + +static bool is_cache_coherent(int fd, uint32_t handle) +{ + return gem_get_caching(fd, handle) != I915_CACHING_NONE; +} + +enum ccs_copy_direction { + CCS_LINEAR_TO_BUF, + CCS_BUF_TO_LINEAR, +}; + +static void __copy_ccs(struct buf_ops *bops, struct intel_buf *buf, + uint32_t *linear, enum ccs_copy_direction dir) +{ + uint64_t size, offset, ccs_size; + void *map; + int gen; + + if (!buf->compression) + return; + + gen = bops->intel_gen; + offset = CCS_OFFSET(buf); + ccs_size = CCS_SIZE(gen, buf); + size = offset + ccs_size; + + map = __gem_mmap_offset__wc(bops->fd, buf->handle, 0, size, + PROT_READ | PROT_WRITE); + if (!map) + map = gem_mmap__wc(bops->fd, buf->handle, 0, size, + PROT_READ | PROT_WRITE); + + switch (dir) { + case CCS_LINEAR_TO_BUF: + gem_set_domain(bops->fd, buf->handle, + I915_GEM_DOMAIN_WC, I915_GEM_DOMAIN_WC); + igt_memcpy_from_wc(map + offset, (uint8_t *) linear + offset, + ccs_size); + case CCS_BUF_TO_LINEAR: + gem_set_domain(bops->fd, buf->handle, I915_GEM_DOMAIN_WC, 0); + igt_memcpy_from_wc((uint8_t *) linear + offset, map + offset, + ccs_size); + } + + munmap(map, size); +} + +static void *mmap_write(int fd, struct intel_buf *buf) +{ + void *map = NULL; + + if (is_cache_coherent(fd, buf->handle)) { + map = __gem_mmap_offset__cpu(fd, buf->handle, 0, buf->size, + PROT_READ | PROT_WRITE); + if (!map) + map = __gem_mmap__cpu(fd, buf->handle, 0, buf->size, + PROT_READ | PROT_WRITE); + + if (map) + gem_set_domain(fd, buf->handle, + I915_GEM_DOMAIN_CPU, + I915_GEM_DOMAIN_CPU); + } + + if (!map) { + map = __gem_mmap_offset__wc(fd, buf->handle, 0, buf->size, + PROT_READ | PROT_WRITE); + if (!map) + map = gem_mmap__wc(fd, buf->handle, 0, buf->size, + PROT_READ | PROT_WRITE); + + gem_set_domain(fd, buf->handle, + I915_GEM_DOMAIN_WC, I915_GEM_DOMAIN_WC); + } + + return map; +} + +static void *mmap_read(int fd, struct intel_buf *buf) +{ + void *map = NULL; + + if (gem_has_llc(fd) || is_cache_coherent(fd, buf->handle)) { + map = __gem_mmap_offset__cpu(fd, buf->handle, 0, + buf->size, PROT_READ); + if (!map) + map = __gem_mmap__cpu(fd, buf->handle, 0, buf->size, + PROT_READ); + + if (map) + gem_set_domain(fd, buf->handle, I915_GEM_DOMAIN_CPU, 0); + } + + if (!map) { + map = __gem_mmap_offset__wc(fd, buf->handle, 0, buf->size, + PROT_READ); + if (!map) + map = gem_mmap__wc(fd, buf->handle, 0, buf->size, + PROT_READ); + + gem_set_domain(fd, buf->handle, I915_GEM_DOMAIN_WC, 0); + } + + return map; +} + +static void __copy_linear_to(int fd, struct intel_buf *buf, + const uint32_t *linear, + int tiling, uint32_t swizzle) +{ + const tile_fn fn = __get_tile_fn_ptr(tiling); + int height = intel_buf_height(buf); + int width = intel_buf_width(buf); + void *map = mmap_write(fd, buf); + + for (int y = 0; y < height; y++) { + for (int x = 0; x < width; x++) { + uint32_t *ptr = fn(map, x, y, buf->stride, buf->bpp/8); + + if (swizzle) + ptr = from_user_pointer(swizzle_addr(ptr, + swizzle)); + *ptr = linear[y * width + x]; + } + } + + munmap(map, buf->size); +} + +static void copy_linear_to_x(struct buf_ops *bops, struct intel_buf *buf, + uint32_t *linear) +{ + DEBUGFN(); + __copy_linear_to(bops->fd, buf, linear, I915_TILING_X, bops->swizzle_x); +} + +static void copy_linear_to_y(struct buf_ops *bops, struct intel_buf *buf, + uint32_t *linear) +{ + DEBUGFN(); + __copy_linear_to(bops->fd, buf, linear, I915_TILING_Y, bops->swizzle_y); +} + +static void copy_linear_to_yf(struct buf_ops *bops, struct intel_buf *buf, + uint32_t *linear) +{ + DEBUGFN(); + __copy_linear_to(bops->fd, buf, linear, I915_TILING_Yf, 0); +} + +static void copy_linear_to_ys(struct buf_ops *bops, struct intel_buf *buf, + uint32_t *linear) +{ + DEBUGFN(); + __copy_linear_to(bops->fd, buf, linear, I915_TILING_Ys, 0); +} + +static void __copy_to_linear(int fd, struct intel_buf *buf, + uint32_t *linear, int tiling, uint32_t swizzle) +{ + const tile_fn fn = __get_tile_fn_ptr(tiling); + int height = intel_buf_height(buf); + int width = intel_buf_width(buf); + void *map = mmap_write(fd, buf); + + for (int y = 0; y < height; y++) { + for (int x = 0; x < width; x++) { + uint32_t *ptr = fn(map, x, y, buf->stride, buf->bpp/8); + + if (swizzle) + ptr = from_user_pointer(swizzle_addr(ptr, + swizzle)); + linear[y * width + x] = *ptr; + } + } + + munmap(map, buf->size); +} + +static void copy_x_to_linear(struct buf_ops *bops, struct intel_buf *buf, + uint32_t *linear) +{ + DEBUGFN(); + __copy_to_linear(bops->fd, buf, linear, I915_TILING_X, bops->swizzle_x); +} + +static void copy_y_to_linear(struct buf_ops *bops, struct intel_buf *buf, + uint32_t *linear) +{ + DEBUGFN(); + __copy_to_linear(bops->fd, buf, linear, I915_TILING_Y, bops->swizzle_y); +} + +static void copy_yf_to_linear(struct buf_ops *bops, struct intel_buf *buf, + uint32_t *linear) +{ + DEBUGFN(); + __copy_to_linear(bops->fd, buf, linear, I915_TILING_Yf, 0); +} + +static void copy_ys_to_linear(struct buf_ops *bops, struct intel_buf *buf, + uint32_t *linear) +{ + DEBUGFN(); + __copy_to_linear(bops->fd, buf, linear, I915_TILING_Ys, 0); +} + +static void copy_linear_to_gtt(struct buf_ops *bops, struct intel_buf *buf, + uint32_t *linear) +{ + void *map; + + DEBUGFN(); + + map = gem_mmap__gtt(bops->fd, buf->handle, + buf->size, PROT_READ | PROT_WRITE); + + gem_set_domain(bops->fd, buf->handle, + I915_GEM_DOMAIN_GTT, I915_GEM_DOMAIN_GTT); + + memcpy(map, linear, buf->size); + + munmap(map, buf->size); +} + +static void copy_gtt_to_linear(struct buf_ops *bops, struct intel_buf *buf, + uint32_t *linear) +{ + void *map; + + DEBUGFN(); + + map = gem_mmap__gtt(bops->fd, buf->handle, + buf->size, PROT_READ); + + gem_set_domain(bops->fd, buf->handle, + I915_GEM_DOMAIN_GTT, 0); + + igt_memcpy_from_wc(linear, map, buf->size); + + munmap(map, buf->size); +} + +static void copy_linear_to_wc(struct buf_ops *bops, struct intel_buf *buf, + uint32_t *linear) +{ + void *map; + + DEBUGFN(); + + map = mmap_write(bops->fd, buf); + memcpy(map, linear, buf->size); + munmap(map, buf->size); +} + +static void copy_wc_to_linear(struct buf_ops *bops, struct intel_buf *buf, + uint32_t *linear) +{ + void *map; + + DEBUGFN(); + + map = mmap_read(bops->fd, buf); + igt_memcpy_from_wc(linear, map, buf->size); + munmap(map, buf->size); +} + +void intel_buf_to_linear(struct buf_ops *bops, struct intel_buf *buf, + uint32_t *linear) +{ + igt_assert(bops); + + switch (buf->tiling) { + case I915_TILING_NONE: + igt_assert(bops->to_linear); + bops->to_linear(bops, buf, linear); + break; + case I915_TILING_X: + igt_assert(bops->x_to_linear); + bops->x_to_linear(bops, buf, linear); + break; + case I915_TILING_Y: + igt_assert(bops->y_to_linear); + bops->y_to_linear(bops, buf, linear); + break; + case I915_TILING_Yf: + igt_assert(bops->yf_to_linear); + bops->yf_to_linear(bops, buf, linear); + break; + case I915_TILING_Ys: + igt_assert(bops->ys_to_linear); + bops->ys_to_linear(bops, buf, linear); + break; + } + + if (buf->compression) + __copy_ccs(bops, buf, linear, CCS_BUF_TO_LINEAR); +} + +void linear_to_intel_buf(struct buf_ops *bops, struct intel_buf *buf, + uint32_t *linear) +{ + igt_assert(bops); + + switch (buf->tiling) { + case I915_TILING_NONE: + igt_assert(bops->linear_to); + bops->linear_to(bops, buf, linear); + break; + case I915_TILING_X: + igt_assert(bops->linear_to_x); + bops->linear_to_x(bops, buf, linear); + break; + case I915_TILING_Y: + igt_assert(bops->linear_to_y); + bops->linear_to_y(bops, buf, linear); + break; + case I915_TILING_Yf: + igt_assert(bops->linear_to_yf); + bops->linear_to_yf(bops, buf, linear); + break; + case I915_TILING_Ys: + igt_assert(bops->linear_to_ys); + bops->linear_to_ys(bops, buf, linear); + break; + } + + if (buf->compression) + __copy_ccs(bops, buf, linear, CCS_LINEAR_TO_BUF); +} + +static void __intel_buf_init(struct buf_ops *bops, + uint32_t handle, + struct intel_buf *buf, + int width, int height, int bpp, + uint32_t req_tiling, uint32_t compression) +{ + uint32_t tiling = req_tiling; + uint32_t size; + + igt_assert(bops); + igt_assert(buf); + igt_assert(width > 0 && height > 0); + igt_assert(bpp == 8 || bpp == 16 || bpp == 32); + + memset(buf, 0, sizeof(*buf)); + + if (compression) { + int aux_width, aux_height; + + igt_require(bops->intel_gen >= 9); + igt_assert(req_tiling == I915_TILING_Y || + req_tiling == I915_TILING_Yf); + + /* + * On GEN12+ we align the main surface to 4 * 4 main surface + * tiles, which is 64kB. These 16 tiles are mapped by 4 AUX + * CCS units, that is 4 * 64 bytes. These 4 CCS units are in + * turn mapped by one L1 AUX page table entry. + */ + if (bops->intel_gen >= 12) + buf->stride = ALIGN(width * (bpp / 8), 128 * 4); + else + buf->stride = ALIGN(width * (bpp / 8), 128); + + if (bops->intel_gen >= 12) + height = ALIGN(height, 4 * 32); + + buf->size = buf->stride * height; + buf->tiling = tiling; + buf->bpp = bpp; + buf->compression = compression; + + aux_width = intel_buf_aux_width(bops->intel_gen, buf); + aux_height = intel_buf_aux_height(bops->intel_gen, buf); + + buf->aux.offset = buf->stride * ALIGN(height, 32); + buf->aux.stride = aux_width; + + size = buf->aux.offset + aux_width * aux_height; + + } else { + buf->stride = ALIGN(width * (bpp / 8), 128); + buf->size = buf->stride * height; + buf->tiling = tiling; + buf->bpp = bpp; + + size = buf->stride * ALIGN(height, 32); + } + + if (handle) + buf->handle = handle; + else + buf->handle = gem_create(bops->fd, size); + + set_hw_tiled(bops, buf); +} + +/** + * intel_buf_init + * @bops: pointer to buf_ops + * @buf: pointer to intel_buf structure to be filled + * @width: surface width + * @height: surface height + * @bpp: bits-per-pixel (8 / 16 / 32) + * @tiling: surface tiling + * @compression: surface compression type + * + * Function creates new BO within intel_buf structure and fills all + * structure fields. + * + * Note. For X / Y if GPU supports fences HW tiling is configured. + */ +void intel_buf_init(struct buf_ops *bops, + struct intel_buf *buf, + int width, int height, int bpp, + uint32_t tiling, uint32_t compression) +{ + __intel_buf_init(bops, 0, buf, width, height, bpp, tiling, + compression); +} + +/** + * intel_buf_close + * @bops: pointer to buf_ops + * @buf: pointer to intel_buf structure + * + * Function closes gem BO inside intel_buf. + */ +void intel_buf_close(struct buf_ops *bops, struct intel_buf *buf) +{ + igt_assert(bops); + igt_assert(buf); + + gem_close(bops->fd, buf->handle); +} + +/** + * intel_buf_init_using_handle + * @bops: pointer to buf_ops + * @handle: BO handle created by the caller + * @buf: pointer to intel_buf structure to be filled + * @width: surface width + * @height: surface height + * @bpp: bits-per-pixel (8 / 16 / 32) + * @tiling: surface tiling + * @compression: surface compression type + * + * Function configures BO handle within intel_buf structure passed by the caller + * (with all its metadata - width, height, ...). Useful if BO was created + * outside. + * + * Note: intel_buf_close() can be used to close the BO handle, but caller + * must be aware to not close the BO twice. + */ +void intel_buf_init_using_handle(struct buf_ops *bops, + uint32_t handle, + struct intel_buf *buf, + int width, int height, int bpp, + uint32_t req_tiling, uint32_t compression) +{ + __intel_buf_init(bops, handle, buf, width, height, bpp, req_tiling, + compression); +} + +#define DEFAULT_BUFOPS(__gen_start, __gen_end) \ + .gen_start = __gen_start, \ + .gen_end = __gen_end, \ + .supported_hw_tiles = TILE_X | TILE_Y, \ + .linear_to = copy_linear_to_wc, \ + .linear_to_x = copy_linear_to_gtt, \ + .linear_to_y = copy_linear_to_gtt, \ + .linear_to_yf = copy_linear_to_yf, \ + .linear_to_ys = copy_linear_to_ys, \ + .to_linear = copy_wc_to_linear, \ + .x_to_linear = copy_gtt_to_linear, \ + .y_to_linear = copy_gtt_to_linear, \ + .yf_to_linear = copy_yf_to_linear, \ + .ys_to_linear = copy_ys_to_linear + +struct buf_ops buf_ops_arr[] = { + { + DEFAULT_BUFOPS(2, 8), + .supported_tiles = TILE_NONE | TILE_X | TILE_Y, + }, + + { + DEFAULT_BUFOPS(9, 11), + .supported_tiles = TILE_NONE | TILE_X | TILE_Y | TILE_Yf, + }, + + { + DEFAULT_BUFOPS(12, 12), + .supported_tiles = TILE_NONE | TILE_X | TILE_Y | TILE_Yf | TILE_Ys, + }, +}; + +static bool probe_hw_tiling(struct buf_ops *bops, uint32_t tiling) +{ + uint64_t size = 256 * 256; + uint32_t handle, buf_tiling, buf_swizzle; + uint32_t stride; + int ret; + bool is_set = false; + + if (tiling == I915_TILING_X) + stride = 512; + else if (tiling == I915_TILING_Y) + stride = 128; + else + return false; + + handle = gem_create(bops->fd, size); + + /* Single shot, if no fences are available we fail immediately */ + ret = __set_tiling(bops->fd, handle, tiling, stride); + if (ret) + goto end; + + is_set = __get_tiling(bops->fd, handle, &buf_tiling, &buf_swizzle); + if (is_set) { + if (tiling == I915_TILING_X) + bops->swizzle_x = buf_swizzle; + else if (tiling == I915_TILING_Y) + bops->swizzle_y = buf_swizzle; + } +end: + gem_close(bops->fd, handle); + + return is_set; +} + +static void *alloc_aligned(uint64_t size) +{ + void *p; + + igt_assert_eq(posix_memalign(&p, 16, size), 0); + + return p; +} + +/* + * Simple idempotency test between HW -> SW and SW -> HW BO. + */ +static void idempotency_selftest(struct buf_ops *bops, uint32_t tiling) +{ + struct intel_buf buf; + uint8_t *linear_in, *linear_out, *map; + int i; + const int width = 512; + const int height = 512; + const int bpp = 32; + const int size = width * height * bpp / 8; + bool software_tiling = false; + + if (!is_hw_tiling_supported(bops, tiling)) + return; + + linear_in = alloc_aligned(size); + linear_out = alloc_aligned(size); + + for (i = 0; i < size; i++) + linear_in[i] = i % 253; /* prime chosen intentionally */ + + do { + igt_debug("Checking idempotency, SW: %s, HW: %s, tiling: %s\n", + bool_str(software_tiling), + bool_str(!software_tiling), + tiling_str(tiling)); + intel_buf_init(bops, &buf, width, height, bpp, tiling, false); + buf_ops_set_software_tiling(bops, tiling, software_tiling); + + linear_to_intel_buf(bops, &buf, (uint32_t *) linear_in); + map = __gem_mmap_offset__cpu(bops->fd, buf.handle, 0, + buf.size, PROT_READ); + if (!map) + map = gem_mmap__cpu(bops->fd, buf.handle, 0, + buf.size, PROT_READ); + gem_set_domain(bops->fd, buf.handle, I915_GEM_DOMAIN_CPU, 0); + igt_assert(memcmp(linear_in, map, size)); + munmap(map, size); + + buf_ops_set_software_tiling(bops, tiling, !software_tiling); + intel_buf_to_linear(bops, &buf, (uint32_t *) linear_out); + igt_assert(!memcmp(linear_in, linear_out, size)); + + intel_buf_close(bops, &buf); + + software_tiling = !software_tiling; + } while (software_tiling); + + igt_debug("Idempotency for %s tiling OK\n", tiling_str(tiling)); + buf_ops_set_software_tiling(bops, tiling, false); +} + +/** + * buf_ops_create + * @fd: device filedescriptor + * + * Create buf_ops structure depending on fd-device capabilities. + * + * Returns: opaque pointer to buf_ops. + * + */ +struct buf_ops *buf_ops_create(int fd) +{ + struct buf_ops *bops = calloc(1, sizeof(*bops)); + uint32_t devid; + int generation; + + igt_assert(bops); + + devid = intel_get_drm_devid(fd); + generation = intel_gen(devid); + + /* Predefined settings */ + for (int i = 0; i < ARRAY_SIZE(buf_ops_arr); i++) { + if (generation >= buf_ops_arr[i].gen_start && + generation <= buf_ops_arr[i].gen_end) { + memcpy(bops, &buf_ops_arr[i], sizeof(*bops)); + bops->fd = fd; + bops->intel_gen = generation; + igt_debug("generation: %d, supported tiles: 0x%02x\n", + generation, bops->supported_tiles); + break; + } + } + + igt_assert(bops->intel_gen); + + /* + * Warning! + * + * Gen2 software tiling/detiling is not supported! (yet). + * + * If you are brave hero with an access to Gen2 you can save the world. + * Until then we're doomed to use only hardware (de)tiling. + * + * Ok, you have been warned. + */ + if (bops->intel_gen == 2) { + igt_warn("Gen2 detected. HW (de)tiling support only."); + return bops; + } + + /* Let's probe X and Y hw tiling support */ + if (is_hw_tiling_supported(bops, I915_TILING_X)) { + bool supported = probe_hw_tiling(bops, I915_TILING_X); + + igt_debug("X fence support: %s\n", bool_str(supported)); + if (!supported) { + bops->supported_hw_tiles &= ~TILE_X; + bops->linear_to_x = copy_linear_to_x; + bops->x_to_linear = copy_x_to_linear; + } + } + + if (is_hw_tiling_supported(bops, I915_TILING_Y)) { + bool supported = probe_hw_tiling(bops, I915_TILING_Y); + + igt_debug("Y fence support: %s\n", bool_str(supported)); + if (!supported) { + bops->supported_hw_tiles &= ~TILE_Y; + bops->linear_to_y = copy_linear_to_y; + bops->y_to_linear = copy_y_to_linear; + } + } + + /* Disable other tiling format functions if not supported */ + if (!is_tiling_supported(bops, I915_TILING_Yf)) { + igt_debug("Yf format not supported\n"); + bops->linear_to_yf = NULL; + bops->yf_to_linear = NULL; + } + + if (!is_tiling_supported(bops, I915_TILING_Ys)) { + igt_debug("Ys format not supported\n"); + bops->linear_to_ys = NULL; + bops->ys_to_linear = NULL; + } + + idempotency_selftest(bops, I915_TILING_X); + idempotency_selftest(bops, I915_TILING_Y); + + return bops; +} + +/** + * buf_ops_destroy + * @bops: pointer to buf_ops + * + * Function frees buf_ops structure. + */ +void buf_ops_destroy(struct buf_ops *bops) +{ + igt_assert(bops); + free(bops); +} + +/** + * buf_ops_set_software_tiling + * @bops: pointer to buf_ops + * @tiling: surface tiling + * @use_software_tiling: if true use software copying methods, otherwise + * use hardware (via gtt) + * + * Function allows switch X / Y surfaces to software / hardware copying methods + * which honors tiling and swizzling. + * + * Returns: + * false - switch wasn't possible. + * true - switch to software / hardware method succeed. + */ +bool buf_ops_set_software_tiling(struct buf_ops *bops, + uint32_t tiling, + bool use_software_tiling) +{ + bool was_changed = true; + + igt_assert(bops); + + /* Until appropriate code is added we don't support SW tiling on Gen2 */ + if (bops->intel_gen == 2) { + igt_warn("Change to software tiling on Gen2 is not supported!"); + return false; + } + + switch (tiling) { + case I915_TILING_X: + if (use_software_tiling) { + igt_debug("-> change X to SW\n"); + bops->linear_to_x = copy_linear_to_x; + bops->x_to_linear = copy_x_to_linear; + } else { + if (is_hw_tiling_supported(bops, I915_TILING_X)) { + igt_debug("-> change X to HW\n"); + bops->linear_to_x = copy_linear_to_gtt; + bops->x_to_linear = copy_gtt_to_linear; + } else { + igt_debug("-> X cannot be changed to HW\n"); + was_changed = false; + } + } + break; + + case I915_TILING_Y: + if (use_software_tiling) { + igt_debug("-> change Y to SW\n"); + bops->linear_to_y = copy_linear_to_y; + bops->y_to_linear = copy_y_to_linear; + } else { + if (is_hw_tiling_supported(bops, I915_TILING_Y)) { + igt_debug("-> change Y to HW\n"); + bops->linear_to_y = copy_linear_to_gtt; + bops->y_to_linear = copy_gtt_to_linear; + } else { + igt_debug("-> Y cannot be changed to HW\n"); + was_changed = false; + } + } + break; + + default: + igt_warn("Invalid tiling: %d\n", tiling); + was_changed = false; + } + + return was_changed; +} + +/** + * buf_ops_has_hw_fence + * @bops: pointer to buf_ops + * @tiling: surface tiling + * + * Function checks if surface with tiling has HW fences which can be used + * to copy it via gtt. + * + * Returns: + * false - fence for tiling is not supported. + * true - fence for tiling is supported. + */ +bool buf_ops_has_hw_fence(struct buf_ops *bops, uint32_t tiling) +{ + uint32_t tile_mask = TILE_DEF(tiling); + + igt_assert(bops); + + if (tile_mask & bops->supported_hw_tiles) + return true; + + return false; +} + +/** + * buf_ops_has_tiling_support + * @bops: pointer to buf_ops + * @tiling: surface tiling + * + * Function checks capabilities to handle surfaces with tiling in GPU. + * + * Returns: + * false - GPU does not support tiling. + * true - GPU supports tiling. + */ +bool buf_ops_has_tiling_support(struct buf_ops *bops, uint32_t tiling) +{ + uint32_t tile_mask = TILE_DEF(tiling); + + igt_assert(bops); + + if (tile_mask & bops->supported_tiles) + return true; + + return false; +} diff --git a/lib/intel_bufops.h b/lib/intel_bufops.h new file mode 100644 index 000000000..f3d6aed8a --- /dev/null +++ b/lib/intel_bufops.h @@ -0,0 +1,86 @@ +#ifndef __INTEL_BUFOPS_H__ +#define __INTEL_BUFOPS_H__ + +#include <stdint.h> + +struct buf_ops; + +struct intel_buf { + uint32_t handle; + uint32_t stride; + uint32_t tiling; + uint32_t bpp; + uint32_t size; + uint32_t compression; + struct { + uint32_t offset; + uint32_t stride; + } aux; +}; + +static inline unsigned int intel_buf_width(const struct intel_buf *buf) +{ + return buf->stride / (buf->bpp / 8); +} + +static inline unsigned int intel_buf_height(const struct intel_buf *buf) +{ + return buf->size / buf->stride; +} + +static inline unsigned int +intel_buf_aux_width(int gen, const struct intel_buf *buf) +{ + /* + * GEN12+: The AUX CCS unit size is 64 bytes mapping 4 main surface + * tiles. Thus the width of the CCS unit is 4*32=128 pixels on the + * main surface. + */ + if (gen >= 12) + return DIV_ROUND_UP(intel_buf_width(buf), 128) * 64; + + return DIV_ROUND_UP(intel_buf_width(buf), 1024) * 128; +} + +static inline unsigned int +intel_buf_aux_height(int gen, const struct intel_buf *buf) +{ + /* + * GEN12+: The AUX CCS unit size is 64 bytes mapping 4 main surface + * tiles. Thus the height of the CCS unit is 32 pixel rows on the main + * surface. + */ + if (gen >= 12) + return DIV_ROUND_UP(intel_buf_height(buf), 32); + + return DIV_ROUND_UP(intel_buf_height(buf), 512) * 32; +} + +struct buf_ops *buf_ops_create(int fd); +void buf_ops_destroy(struct buf_ops *bops); + +bool buf_ops_set_software_tiling(struct buf_ops *bops, + uint32_t tiling, + bool use_software_tiling); + +void intel_buf_to_linear(struct buf_ops *bops, struct intel_buf *buf, + uint32_t *linear); + +void linear_to_intel_buf(struct buf_ops *bops, struct intel_buf *buf, + uint32_t *linear); + +bool buf_ops_has_hw_fence(struct buf_ops *bops, uint32_t tiling); +bool buf_ops_has_tiling_support(struct buf_ops *bops, uint32_t tiling); + +void intel_buf_init(struct buf_ops *bops, struct intel_buf *buf, + int width, int height, int bpp, + uint32_t tiling, uint32_t compression); +void intel_buf_close(struct buf_ops *bops, struct intel_buf *buf); + +void intel_buf_init_using_handle(struct buf_ops *bops, + uint32_t handle, + struct intel_buf *buf, + int width, int height, int bpp, + uint32_t req_tiling, uint32_t compression); + +#endif diff --git a/lib/meson.build b/lib/meson.build index 57eb7d938..cf3e35dad 100644 --- a/lib/meson.build +++ b/lib/meson.build @@ -27,6 +27,7 @@ lib_sources = [ 'igt_x86.c', 'instdone.c', 'intel_batchbuffer.c', + 'intel_bufops.c', 'intel_chipset.c', 'intel_device_info.c', 'intel_os.c', -- 2.25.0 _______________________________________________ Intel-gfx mailing list Intel-gfx@xxxxxxxxxxxxxxxxxxxxx https://lists.freedesktop.org/mailman/listinfo/intel-gfx