Re: [PATCH i-g-t 1/2 v2] lib: Add media spin

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



(He Shuang on behalf of Liu Lei)
Tested-by: Lei,Liu lei.a.liu@xxxxxxxxx

I-G-T test result:
./pm_sseu
IGT-Version: 1.9-g07be8fe (x86_64) (Linux: 4.0.0-rc3_drm-intel-nightly_c09a3b_20150310+ x86_64)
Subtest full-enable: SUCCESS (0.010s)

Manually test result:
SSEU Device Info
Available Slice Total: 1
Available Subslice Total: 3
Available Subslice Per Slice: 3
Available EU Total: 23
Available EU Per Subslice: 8
Has Slice Power Gating: no
Has Subslice Power Gating: no
Has EU Power Gating: yes
SSEU Device Status
Enabled Slice Total: 1
Enabled Subslice Total: 3
Enabled Subslice Per Slice: 3
Enabled EU Total: 24
Enabled EU Per Subslice: 8

EU are enabled in pairs. Because one EU in a pair can be fused-off, it is possible to see such case where reported EU enabled is greater than reported EU available. The IGT test allows for this discrepancy and only fails if enabled is less than available, which can only happen if unwanted power gating is applied

Best wishes
Liu,Lei

> -----Original Message-----
> From: Intel-gfx [mailto:intel-gfx-bounces@xxxxxxxxxxxxxxxxxxxxx] On Behalf
> Of jeff.mcgee@xxxxxxxxx
> Sent: Friday, March 13, 2015 1:52 AM
> To: intel-gfx@xxxxxxxxxxxxxxxxxxxxx
> Subject:  [PATCH i-g-t 1/2 v2] lib: Add media spin
> 
> From: Jeff McGee <jeff.mcgee@xxxxxxxxx>
> 
> The media spin utility is derived from media fill. The purpose
> is to create a simple means to keep the render engine (media
> pipeline) busy for a controlled amount of time. It does so by
> emitting a batch with a single execution thread that spins in
> a tight loop the requested number of times. Each spin increments
> a counter whose final 32-bit value is written to the destination
> buffer on completion for checking. The implementation supports
> Gen8, Gen8lp, and Gen9.
> 
> v2: Apply the recommendations of igt.cocci.
> 
> Signed-off-by: Jeff McGee <jeff.mcgee@xxxxxxxxx>
> ---
>  lib/Makefile.sources    |   2 +
>  lib/intel_batchbuffer.c |  24 +++
>  lib/intel_batchbuffer.h |  22 ++
>  lib/media_spin.c        | 540
> ++++++++++++++++++++++++++++++++++++++++++++++++
>  lib/media_spin.h        |  39 ++++
>  5 files changed, 627 insertions(+)
>  create mode 100644 lib/media_spin.c
>  create mode 100644 lib/media_spin.h
> 
> diff --git a/lib/Makefile.sources b/lib/Makefile.sources
> index 76f353a..3d93629 100644
> --- a/lib/Makefile.sources
> +++ b/lib/Makefile.sources
> @@ -29,6 +29,8 @@ libintel_tools_la_SOURCES = 	\
>  	media_fill_gen8.c       \
>  	media_fill_gen8lp.c     \
>  	media_fill_gen9.c       \
> +	media_spin.h		\
> +	media_spin.c	\
>  	gen7_media.h            \
>  	gen8_media.h            \
>  	rendercopy_i915.c	\
> diff --git a/lib/intel_batchbuffer.c b/lib/intel_batchbuffer.c
> index 666c323..195ccc4 100644
> --- a/lib/intel_batchbuffer.c
> +++ b/lib/intel_batchbuffer.c
> @@ -40,6 +40,7 @@
>  #include "rendercopy.h"
>  #include "media_fill.h"
>  #include "ioctl_wrappers.h"
> +#include "media_spin.h"
> 
>  #include <i915_drm.h>
> 
> @@ -785,3 +786,26 @@ igt_fillfunc_t igt_get_gpgpu_fillfunc(int devid)
> 
>  	return fill;
>  }
> +
> +/**
> + * igt_get_media_spinfunc:
> + * @devid: pci device id
> + *
> + * Returns:
> + *
> + * The platform-specific media spin function pointer for the device specified
> + * with @devid. Will return NULL when no media spin function is
> implemented.
> + */
> +igt_media_spinfunc_t igt_get_media_spinfunc(int devid)
> +{
> +	igt_media_spinfunc_t spin = NULL;
> +
> +	if (IS_GEN9(devid))
> +		spin = gen9_media_spinfunc;
> +	else if (IS_BROADWELL(devid))
> +		spin = gen8_media_spinfunc;
> +	else if (IS_CHERRYVIEW(devid))
> +		spin = gen8lp_media_spinfunc;
> +
> +	return spin;
> +}
> diff --git a/lib/intel_batchbuffer.h b/lib/intel_batchbuffer.h
> index fa8875b..62c8396 100644
> --- a/lib/intel_batchbuffer.h
> +++ b/lib/intel_batchbuffer.h
> @@ -300,4 +300,26 @@ typedef void (*igt_fillfunc_t)(struct
> intel_batchbuffer *batch,
>  igt_fillfunc_t igt_get_media_fillfunc(int devid);
>  igt_fillfunc_t igt_get_gpgpu_fillfunc(int devid);
> 
> +/**
> + * igt_media_spinfunc_t:
> + * @batch: batchbuffer object
> + * @dst: destination i-g-t buffer object
> + * @spins: number of loops to execute
> + *
> + * This is the type of the per-platform media spin functions. The
> + * platform-specific implementation can be obtained by calling
> + * igt_get_media_spinfunc().
> + *
> + * The media spin function emits a batchbuffer for the render engine with
> + * the media pipeline selected. The workload consists of a single thread
> + * which spins in a tight loop the requested number of times. Each spin
> + * increments a counter whose final 32-bit value is written to the
> + * destination buffer on completion. This utility provides a simple way
> + * to keep the render engine busy for a set time for various tests.
> + */
> +typedef void (*igt_media_spinfunc_t)(struct intel_batchbuffer *batch,
> +				     struct igt_buf *dst, uint32_t spins);
> +
> +igt_media_spinfunc_t igt_get_media_spinfunc(int devid);
> +
>  #endif
> diff --git a/lib/media_spin.c b/lib/media_spin.c
> new file mode 100644
> index 0000000..580c109
> --- /dev/null
> +++ b/lib/media_spin.c
> @@ -0,0 +1,540 @@
> +/*
> + * Copyright © 2015 Intel Corporation
> + *
> + * Permission is hereby granted, free of charge, to any person obtaining a
> + * copy of this software and associated documentation files (the
> "Software"),
> + * to deal in the Software without restriction, including without limitation
> + * the rights to use, copy, modify, merge, publish, distribute, sublicense,
> + * and/or sell copies of the Software, and to permit persons to whom the
> + * Software is furnished to do so, subject to the following conditions:
> + *
> + * The above copyright notice and this permission notice (including the next
> + * paragraph) shall be included in all copies or substantial portions of the
> + * Software.
> + *
> + * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
> EXPRESS OR
> + * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
> MERCHANTABILITY,
> + * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.  IN NO
> EVENT SHALL
> + * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM,
> DAMAGES OR OTHER
> + * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
> ARISING
> + * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
> OTHER DEALINGS
> + * IN THE SOFTWARE.
> + *
> + * Authors:
> + * 	Jeff McGee <jeff.mcgee@xxxxxxxxx>
> + */
> +
> +#include <intel_bufmgr.h>
> +#include <i915_drm.h>
> +#include "intel_reg.h"
> +#include "drmtest.h"
> +#include "intel_batchbuffer.h"
> +#include "gen8_media.h"
> +#include "media_spin.h"
> +
> +static const uint32_t spin_kernel[][4] = {
> +	{ 0x00600001, 0x20800208, 0x008d0000, 0x00000000 }, /* mov
> (8)r4.0<1>:ud r0.0<8;8;1>:ud */
> +	{ 0x00200001, 0x20800208, 0x00450040, 0x00000000 }, /* mov
> (2)r4.0<1>.ud r2.0<2;2;1>:ud */
> +	{ 0x00000001, 0x20880608, 0x00000000, 0x00000003 }, /* mov
> (1)r4.8<1>:ud 0x3 */
> +	{ 0x00000001, 0x20a00608, 0x00000000, 0x00000000 }, /* mov
> (1)r5.0<1>:ud 0 */
> +	{ 0x00000040, 0x20a00208, 0x060000a0, 0x00000001 }, /* add
> (1)r5.0<1>:ud r5.0<0;1;0>:ud 1 */
> +	{ 0x01000010, 0x20000200, 0x02000020, 0x000000a0 }, /* cmp.e.f0.0
> (1)null<1> r1<0;1;0> r5<0;1;0> */
> +	{ 0x00110027, 0x00000000, 0x00000000, 0xffffffe0 }, /* ~f0.0 while (1)
> -32 */
> +	{ 0x0c800031, 0x20000a00, 0x0e000080, 0x040a8000 }, /* send.dcdp1
> (16)null<1> r4.0<0;1;0> 0x040a8000 */
> +	{ 0x00600001, 0x2e000208, 0x008d0000, 0x00000000 }, /* mov
> (8)r112<1>:ud r0.0<8;8;1>:ud */
> +	{ 0x07800031, 0x20000a40, 0x0e000e00, 0x82000010 }, /* send.ts
> (16)null<1> r112<0;1;0>:d 0x82000010 */
> +};
> +
> +static uint32_t
> +batch_used(struct intel_batchbuffer *batch)
> +{
> +	return batch->ptr - batch->buffer;
> +}
> +
> +static uint32_t
> +batch_align(struct intel_batchbuffer *batch, uint32_t align)
> +{
> +	uint32_t offset = batch_used(batch);
> +	offset = ALIGN(offset, align);
> +	batch->ptr = batch->buffer + offset;
> +	return offset;
> +}
> +
> +static void *
> +batch_alloc(struct intel_batchbuffer *batch, uint32_t size, uint32_t align)
> +{
> +	uint32_t offset = batch_align(batch, align);
> +	batch->ptr += size;
> +	return memset(batch->buffer + offset, 0, size);
> +}
> +
> +static uint32_t
> +batch_offset(struct intel_batchbuffer *batch, void *ptr)
> +{
> +	return (uint8_t *)ptr - batch->buffer;
> +}
> +
> +static uint32_t
> +batch_copy(struct intel_batchbuffer *batch, const void *ptr, uint32_t size,
> +	   uint32_t align)
> +{
> +	return batch_offset(batch, memcpy(batch_alloc(batch, size, align),
> ptr, size));
> +}
> +
> +static void
> +gen8_render_flush(struct intel_batchbuffer *batch, uint32_t batch_end)
> +{
> +	int ret;
> +
> +	ret = drm_intel_bo_subdata(batch->bo, 0, 4096, batch->buffer);
> +	if (ret == 0)
> +		ret = drm_intel_gem_bo_context_exec(batch->bo, NULL,
> +						    batch_end, 0);
> +	igt_assert_eq(ret, 0);
> +}
> +
> +static uint32_t
> +gen8_spin_curbe_buffer_data(struct intel_batchbuffer *batch,
> +			    uint32_t iters)
> +{
> +	uint32_t *curbe_buffer;
> +	uint32_t offset;
> +
> +	curbe_buffer = batch_alloc(batch, 64, 64);
> +	offset = batch_offset(batch, curbe_buffer);
> +	*curbe_buffer = iters;
> +
> +	return offset;
> +}
> +
> +static uint32_t
> +gen8_spin_surface_state(struct intel_batchbuffer *batch,
> +			struct igt_buf *buf,
> +			uint32_t format,
> +			int is_dst)
> +{
> +	struct gen8_surface_state *ss;
> +	uint32_t write_domain, read_domain, offset;
> +	int ret;
> +
> +	if (is_dst) {
> +		write_domain = read_domain =
> I915_GEM_DOMAIN_RENDER;
> +	} else {
> +		write_domain = 0;
> +		read_domain = I915_GEM_DOMAIN_SAMPLER;
> +	}
> +
> +	ss = batch_alloc(batch, sizeof(*ss), 64);
> +	offset = batch_offset(batch, ss);
> +
> +	ss->ss0.surface_type = GEN8_SURFACE_2D;
> +	ss->ss0.surface_format = format;
> +	ss->ss0.render_cache_read_write = 1;
> +	ss->ss0.vertical_alignment = 1; /* align 4 */
> +	ss->ss0.horizontal_alignment = 1; /* align 4 */
> +
> +	if (buf->tiling == I915_TILING_X)
> +		ss->ss0.tiled_mode = 2;
> +	else if (buf->tiling == I915_TILING_Y)
> +		ss->ss0.tiled_mode = 3;
> +
> +	ss->ss8.base_addr = buf->bo->offset;
> +
> +	ret = drm_intel_bo_emit_reloc(batch->bo,
> +				batch_offset(batch, ss) + 8 * 4,
> +				buf->bo, 0,
> +				read_domain, write_domain);
> +	igt_assert_eq(ret, 0);
> +
> +	ss->ss2.height = igt_buf_height(buf) - 1;
> +	ss->ss2.width  = igt_buf_width(buf) - 1;
> +	ss->ss3.pitch  = buf->stride - 1;
> +
> +	ss->ss7.shader_chanel_select_r = 4;
> +	ss->ss7.shader_chanel_select_g = 5;
> +	ss->ss7.shader_chanel_select_b = 6;
> +	ss->ss7.shader_chanel_select_a = 7;
> +
> +	return offset;
> +}
> +
> +static uint32_t
> +gen8_spin_binding_table(struct intel_batchbuffer *batch,
> +			struct igt_buf *dst)
> +{
> +	uint32_t *binding_table, offset;
> +
> +	binding_table = batch_alloc(batch, 32, 64);
> +	offset = batch_offset(batch, binding_table);
> +
> +	binding_table[0] = gen8_spin_surface_state(batch, dst,
> +
> 	GEN8_SURFACEFORMAT_R8_UNORM, 1);
> +
> +	return offset;
> +}
> +
> +static uint32_t
> +gen8_spin_media_kernel(struct intel_batchbuffer *batch,
> +		       const uint32_t kernel[][4],
> +		       size_t size)
> +{
> +	uint32_t offset;
> +
> +	offset = batch_copy(batch, kernel, size, 64);
> +
> +	return offset;
> +}
> +
> +static uint32_t
> +gen8_spin_interface_descriptor(struct intel_batchbuffer *batch,
> +			       struct igt_buf *dst)
> +{
> +	struct gen8_interface_descriptor_data *idd;
> +	uint32_t offset;
> +	uint32_t binding_table_offset, kernel_offset;
> +
> +	binding_table_offset = gen8_spin_binding_table(batch, dst);
> +	kernel_offset = gen8_spin_media_kernel(batch, spin_kernel,
> +					       sizeof(spin_kernel));
> +
> +	idd = batch_alloc(batch, sizeof(*idd), 64);
> +	offset = batch_offset(batch, idd);
> +
> +	idd->desc0.kernel_start_pointer = (kernel_offset >> 6);
> +
> +	idd->desc2.single_program_flow = 1;
> +	idd->desc2.floating_point_mode =
> GEN8_FLOATING_POINT_IEEE_754;
> +
> +	idd->desc3.sampler_count = 0;      /* 0 samplers used */
> +	idd->desc3.sampler_state_pointer = 0;
> +
> +	idd->desc4.binding_table_entry_count = 0;
> +	idd->desc4.binding_table_pointer = (binding_table_offset >> 5);
> +
> +	idd->desc5.constant_urb_entry_read_offset = 0;
> +	idd->desc5.constant_urb_entry_read_length = 1; /* grf 1 */
> +
> +	return offset;
> +}
> +
> +static void
> +gen8_emit_state_base_address(struct intel_batchbuffer *batch)
> +{
> +	OUT_BATCH(GEN8_STATE_BASE_ADDRESS | (16 - 2));
> +
> +	/* general */
> +	OUT_BATCH(0 | BASE_ADDRESS_MODIFY);
> +	OUT_BATCH(0);
> +
> +	/* stateless data port */
> +	OUT_BATCH(0 | BASE_ADDRESS_MODIFY);
> +
> +	/* surface */
> +	OUT_RELOC(batch->bo, I915_GEM_DOMAIN_SAMPLER, 0,
> BASE_ADDRESS_MODIFY);
> +
> +	/* dynamic */
> +	OUT_RELOC(batch->bo, I915_GEM_DOMAIN_RENDER |
> I915_GEM_DOMAIN_INSTRUCTION,
> +		0, BASE_ADDRESS_MODIFY);
> +
> +	/* indirect */
> +	OUT_BATCH(0);
> +	OUT_BATCH(0);
> +
> +	/* instruction */
> +	OUT_RELOC(batch->bo, I915_GEM_DOMAIN_INSTRUCTION, 0,
> BASE_ADDRESS_MODIFY);
> +
> +	/* general state buffer size */
> +	OUT_BATCH(0xfffff000 | 1);
> +	/* dynamic state buffer size */
> +	OUT_BATCH(1 << 12 | 1);
> +	/* indirect object buffer size */
> +	OUT_BATCH(0xfffff000 | 1);
> +	/* intruction buffer size, must set modify enable bit, otherwise it
> may result in GPU hang */
> +	OUT_BATCH(1 << 12 | 1);
> +}
> +
> +static void
> +gen9_emit_state_base_address(struct intel_batchbuffer *batch)
> +{
> +	OUT_BATCH(GEN8_STATE_BASE_ADDRESS | (19 - 2));
> +
> +	/* general */
> +	OUT_BATCH(0 | BASE_ADDRESS_MODIFY);
> +	OUT_BATCH(0);
> +
> +	/* stateless data port */
> +	OUT_BATCH(0 | BASE_ADDRESS_MODIFY);
> +
> +	/* surface */
> +	OUT_RELOC(batch->bo, I915_GEM_DOMAIN_SAMPLER, 0,
> BASE_ADDRESS_MODIFY);
> +
> +	/* dynamic */
> +	OUT_RELOC(batch->bo, I915_GEM_DOMAIN_RENDER |
> I915_GEM_DOMAIN_INSTRUCTION,
> +		0, BASE_ADDRESS_MODIFY);
> +
> +	/* indirect */
> +	OUT_BATCH(0);
> +	OUT_BATCH(0);
> +
> +	/* instruction */
> +	OUT_RELOC(batch->bo, I915_GEM_DOMAIN_INSTRUCTION, 0,
> BASE_ADDRESS_MODIFY);
> +
> +	/* general state buffer size */
> +	OUT_BATCH(0xfffff000 | 1);
> +	/* dynamic state buffer size */
> +	OUT_BATCH(1 << 12 | 1);
> +	/* indirect object buffer size */
> +	OUT_BATCH(0xfffff000 | 1);
> +	/* intruction buffer size, must set modify enable bit, otherwise it
> may result in GPU hang */
> +	OUT_BATCH(1 << 12 | 1);
> +
> +	/* Bindless surface state base address */
> +	OUT_BATCH(0 | BASE_ADDRESS_MODIFY);
> +	OUT_BATCH(0);
> +	OUT_BATCH(0xfffff000);
> +}
> +
> +static void
> +gen8_emit_vfe_state(struct intel_batchbuffer *batch)
> +{
> +	OUT_BATCH(GEN8_MEDIA_VFE_STATE | (9 - 2));
> +
> +	/* scratch buffer */
> +	OUT_BATCH(0);
> +	OUT_BATCH(0);
> +
> +	/* number of threads & urb entries */
> +	OUT_BATCH(2 << 8);
> +
> +	OUT_BATCH(0);
> +
> +	/* urb entry size & curbe size */
> +	OUT_BATCH(2 << 16 |
> +		2);
> +
> +	/* scoreboard */
> +	OUT_BATCH(0);
> +	OUT_BATCH(0);
> +	OUT_BATCH(0);
> +}
> +
> +static void
> +gen8_emit_curbe_load(struct intel_batchbuffer *batch, uint32_t
> curbe_buffer)
> +{
> +	OUT_BATCH(GEN8_MEDIA_CURBE_LOAD | (4 - 2));
> +	OUT_BATCH(0);
> +	/* curbe total data length */
> +	OUT_BATCH(64);
> +	/* curbe data start address, is relative to the dynamics base address
> */
> +	OUT_BATCH(curbe_buffer);
> +}
> +
> +static void
> +gen8_emit_interface_descriptor_load(struct intel_batchbuffer *batch,
> +				    uint32_t interface_descriptor)
> +{
> +	OUT_BATCH(GEN8_MEDIA_INTERFACE_DESCRIPTOR_LOAD | (4 - 2));
> +	OUT_BATCH(0);
> +	/* interface descriptor data length */
> +	OUT_BATCH(sizeof(struct gen8_interface_descriptor_data));
> +	/* interface descriptor address, is relative to the dynamics base
> address */
> +	OUT_BATCH(interface_descriptor);
> +}
> +
> +static void
> +gen8_emit_media_state_flush(struct intel_batchbuffer *batch)
> +{
> +	OUT_BATCH(GEN8_MEDIA_STATE_FLUSH | (2 - 2));
> +	OUT_BATCH(0);
> +}
> +
> +static void
> +gen8_emit_media_objects(struct intel_batchbuffer *batch)
> +{
> +	OUT_BATCH(GEN8_MEDIA_OBJECT | (8 - 2));
> +
> +	/* interface descriptor offset */
> +	OUT_BATCH(0);
> +
> +	/* without indirect data */
> +	OUT_BATCH(0);
> +	OUT_BATCH(0);
> +
> +	/* scoreboard */
> +	OUT_BATCH(0);
> +	OUT_BATCH(0);
> +
> +	/* inline data (xoffset, yoffset) */
> +	OUT_BATCH(0);
> +	OUT_BATCH(0);
> +	gen8_emit_media_state_flush(batch);
> +}
> +
> +static void
> +gen8lp_emit_media_objects(struct intel_batchbuffer *batch)
> +{
> +	OUT_BATCH(GEN8_MEDIA_OBJECT | (8 - 2));
> +
> +	/* interface descriptor offset */
> +	OUT_BATCH(0);
> +
> +	/* without indirect data */
> +	OUT_BATCH(0);
> +	OUT_BATCH(0);
> +
> +	/* scoreboard */
> +	OUT_BATCH(0);
> +	OUT_BATCH(0);
> +
> +	/* inline data (xoffset, yoffset) */
> +	OUT_BATCH(0);
> +	OUT_BATCH(0);
> +}
> +
> +/*
> + * This sets up the media pipeline,
> + *
> + * +---------------+ <---- 4096
> + * |       ^       |
> + * |       |       |
> + * |    various    |
> + * |      state    |
> + * |       |       |
> + * |_______|_______| <---- 2048 + ?
> + * |       ^       |
> + * |       |       |
> + * |   batch       |
> + * |    commands   |
> + * |       |       |
> + * |       |       |
> + * +---------------+ <---- 0 + ?
> + *
> + */
> +
> +#define BATCH_STATE_SPLIT 2048
> +
> +void
> +gen8_media_spinfunc(struct intel_batchbuffer *batch,
> +		    struct igt_buf *dst, uint32_t spins)
> +{
> +	uint32_t curbe_buffer, interface_descriptor;
> +	uint32_t batch_end;
> +
> +	intel_batchbuffer_flush_with_context(batch, NULL);
> +
> +	/* setup states */
> +	batch->ptr = &batch->buffer[BATCH_STATE_SPLIT];
> +
> +	curbe_buffer = gen8_spin_curbe_buffer_data(batch, spins);
> +	interface_descriptor = gen8_spin_interface_descriptor(batch, dst);
> +	igt_assert(batch->ptr < &batch->buffer[4095]);
> +
> +	/* media pipeline */
> +	batch->ptr = batch->buffer;
> +	OUT_BATCH(GEN8_PIPELINE_SELECT | PIPELINE_SELECT_MEDIA);
> +	gen8_emit_state_base_address(batch);
> +
> +	gen8_emit_vfe_state(batch);
> +
> +	gen8_emit_curbe_load(batch, curbe_buffer);
> +
> +	gen8_emit_interface_descriptor_load(batch, interface_descriptor);
> +
> +	gen8_emit_media_objects(batch);
> +
> +	OUT_BATCH(MI_BATCH_BUFFER_END);
> +
> +	batch_end = batch_align(batch, 8);
> +	igt_assert(batch_end < BATCH_STATE_SPLIT);
> +
> +	gen8_render_flush(batch, batch_end);
> +	intel_batchbuffer_reset(batch);
> +}
> +
> +void
> +gen8lp_media_spinfunc(struct intel_batchbuffer *batch,
> +		      struct igt_buf *dst, uint32_t spins)
> +{
> +	uint32_t curbe_buffer, interface_descriptor;
> +	uint32_t batch_end;
> +
> +	intel_batchbuffer_flush_with_context(batch, NULL);
> +
> +	/* setup states */
> +	batch->ptr = &batch->buffer[BATCH_STATE_SPLIT];
> +
> +	curbe_buffer = gen8_spin_curbe_buffer_data(batch, spins);
> +	interface_descriptor = gen8_spin_interface_descriptor(batch, dst);
> +	igt_assert(batch->ptr < &batch->buffer[4095]);
> +
> +	/* media pipeline */
> +	batch->ptr = batch->buffer;
> +	OUT_BATCH(GEN8_PIPELINE_SELECT | PIPELINE_SELECT_MEDIA);
> +	gen8_emit_state_base_address(batch);
> +
> +	gen8_emit_vfe_state(batch);
> +
> +	gen8_emit_curbe_load(batch, curbe_buffer);
> +
> +	gen8_emit_interface_descriptor_load(batch, interface_descriptor);
> +
> +	gen8lp_emit_media_objects(batch);
> +
> +	OUT_BATCH(MI_BATCH_BUFFER_END);
> +
> +	batch_end = batch_align(batch, 8);
> +	igt_assert(batch_end < BATCH_STATE_SPLIT);
> +
> +	gen8_render_flush(batch, batch_end);
> +	intel_batchbuffer_reset(batch);
> +}
> +
> +void
> +gen9_media_spinfunc(struct intel_batchbuffer *batch,
> +		    struct igt_buf *dst, uint32_t spins)
> +{
> +	uint32_t curbe_buffer, interface_descriptor;
> +	uint32_t batch_end;
> +
> +	intel_batchbuffer_flush_with_context(batch, NULL);
> +
> +	/* setup states */
> +	batch->ptr = &batch->buffer[BATCH_STATE_SPLIT];
> +
> +	curbe_buffer = gen8_spin_curbe_buffer_data(batch, spins);
> +	interface_descriptor = gen8_spin_interface_descriptor(batch, dst);
> +	igt_assert(batch->ptr < &batch->buffer[4095]);
> +
> +	/* media pipeline */
> +	batch->ptr = batch->buffer;
> +	OUT_BATCH(GEN8_PIPELINE_SELECT | PIPELINE_SELECT_MEDIA |
> +			GEN9_FORCE_MEDIA_AWAKE_ENABLE |
> +			GEN9_SAMPLER_DOP_GATE_DISABLE |
> +			GEN9_PIPELINE_SELECTION_MASK |
> +			GEN9_SAMPLER_DOP_GATE_MASK |
> +			GEN9_FORCE_MEDIA_AWAKE_MASK);
> +	gen9_emit_state_base_address(batch);
> +
> +	gen8_emit_vfe_state(batch);
> +
> +	gen8_emit_curbe_load(batch, curbe_buffer);
> +
> +	gen8_emit_interface_descriptor_load(batch, interface_descriptor);
> +
> +	gen8_emit_media_objects(batch);
> +
> +	OUT_BATCH(GEN8_PIPELINE_SELECT | PIPELINE_SELECT_MEDIA |
> +			GEN9_FORCE_MEDIA_AWAKE_DISABLE |
> +			GEN9_SAMPLER_DOP_GATE_ENABLE |
> +			GEN9_PIPELINE_SELECTION_MASK |
> +			GEN9_SAMPLER_DOP_GATE_MASK |
> +			GEN9_FORCE_MEDIA_AWAKE_MASK);
> +
> +	OUT_BATCH(MI_BATCH_BUFFER_END);
> +
> +	batch_end = batch_align(batch, 8);
> +	igt_assert(batch_end < BATCH_STATE_SPLIT);
> +
> +	gen8_render_flush(batch, batch_end);
> +	intel_batchbuffer_reset(batch);
> +}
> diff --git a/lib/media_spin.h b/lib/media_spin.h
> new file mode 100644
> index 0000000..8bc4829
> --- /dev/null
> +++ b/lib/media_spin.h
> @@ -0,0 +1,39 @@
> +/*
> + * Copyright © 2015 Intel Corporation
> + *
> + * Permission is hereby granted, free of charge, to any person obtaining a
> + * copy of this software and associated documentation files (the
> "Software"),
> + * to deal in the Software without restriction, including without limitation
> + * the rights to use, copy, modify, merge, publish, distribute, sublicense,
> + * and/or sell copies of the Software, and to permit persons to whom the
> + * Software is furnished to do so, subject to the following conditions:
> + *
> + * The above copyright notice and this permission notice (including the next
> + * paragraph) shall be included in all copies or substantial portions of the
> + * Software.
> + *
> + * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
> EXPRESS OR
> + * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
> MERCHANTABILITY,
> + * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT.  IN NO
> EVENT SHALL
> + * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM,
> DAMAGES OR OTHER
> + * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
> ARISING
> + * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
> OTHER DEALINGS
> + * IN THE SOFTWARE.
> + *
> + * Authors:
> + * 	Jeff McGee <jeff.mcgee@xxxxxxxxx>
> + */
> +
> +#ifndef MEDIA_SPIN_H
> +#define MEDIA_SPIN_H
> +
> +void gen8_media_spinfunc(struct intel_batchbuffer *batch,
> +			 struct igt_buf *dst, uint32_t spins);
> +
> +void gen8lp_media_spinfunc(struct intel_batchbuffer *batch,
> +			   struct igt_buf *dst, uint32_t spins);
> +
> +void gen9_media_spinfunc(struct intel_batchbuffer *batch,
> +			 struct igt_buf *dst, uint32_t spins);
> +
> +#endif /* MEDIA_SPIN_H */
> --
> 2.3.0
> 
> _______________________________________________
> Intel-gfx mailing list
> Intel-gfx@xxxxxxxxxxxxxxxxxxxxx
> http://lists.freedesktop.org/mailman/listinfo/intel-gfx
_______________________________________________
Intel-gfx mailing list
Intel-gfx@xxxxxxxxxxxxxxxxxxxxx
http://lists.freedesktop.org/mailman/listinfo/intel-gfx





[Index of Archives]     [Linux USB Devel]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]
  Powered by Linux