GuC interface has been redesigned (or cleaned up, rather) starting with Gen11, as a stepping stone towards a new branching strategy that helps maintain backwards compatibility with previous Gens, as well as sideward compatibility with other OSes. The interface is split in two header files: one for the KMD and one for clients of the GuC (which, in our case, happens to be the KMD as well). SLPC interface files will come at a later date. Could we get eyes on the new interface header files, to make sure the GuC team is moving in the right direction? Signed-off-by: Oscar Mateo <oscar.mateo@xxxxxxxxx> Cc: Joonas Lahtinen <joonas.lahtinen@xxxxxxxxxxxxxxx> Cc: Kevin Rogovin <kevin.rogovin@xxxxxxxxx> Cc: John A Spotswood <john.a.spotswood@xxxxxxxxx> Cc: Anusha Srivatsa <anusha.srivatsa@xxxxxxxxx> Cc: Daniele Ceraolo Spurio <daniele.ceraolospurio@xxxxxxxxx> Cc: Michal Wajdeczko <michal.wajdeczko@xxxxxxxxx> Cc: Michel Thierry <michel.thierry@xxxxxxxxx> Cc: Chris Wilson <chris@xxxxxxxxxxxxxxxxxx> Cc: Michał Winiarski <michal.winiarski@xxxxxxxxx> Cc: Tomasz Lis <tomasz.lis@xxxxxxxxx> Cc: Jon Ewins <jon.ewins@xxxxxxxxx> Cc: Sujaritha Sundaresan <sujaritha.sundaresan@xxxxxxxxx> Cc: Jalpa Patel <jalpa.patel@xxxxxxxxx> Cc: Jackie Li <yaodong.li@xxxxxxxxx> --- drivers/gpu/drm/i915/intel_guc_client_interface.h | 255 +++++++ drivers/gpu/drm/i915/intel_guc_kmd_interface.h | 847 ++++++++++++++++++++++ 2 files changed, 1102 insertions(+) create mode 100644 drivers/gpu/drm/i915/intel_guc_client_interface.h create mode 100644 drivers/gpu/drm/i915/intel_guc_kmd_interface.h diff --git a/drivers/gpu/drm/i915/intel_guc_client_interface.h b/drivers/gpu/drm/i915/intel_guc_client_interface.h new file mode 100644 index 0000000..1ef91a7 --- /dev/null +++ b/drivers/gpu/drm/i915/intel_guc_client_interface.h @@ -0,0 +1,255 @@ +/* + * SPDX-License-Identifier: MIT + * + * Copyright © 2018 Intel Corporation + */ + +#ifndef _INTEL_GUC_CLIENT_INTERFACE_H_ +#define _INTEL_GUC_CLIENT_INTERFACE_H_ + +#pragma pack(1) + +/***************************************************************************** + ********************************** Engines ********************************** + *****************************************************************************/ + +#define GUC_MAX_ENGINE_INSTANCE_PER_CLASS 4 +#define GUC_MAX_SCHEDULABLE_ENGINE_CLASS 5 +#define GUC_MAX_ENGINE_CLASS_COUNT 6 +#define GUC_ENGINE_INVALID 6 + +/* Engine Class that uKernel can schedule on. This is just a SW enumeration. + * HW configuration will depend on the Platform and SKU + */ +enum uk_engine_class { + UK_RENDER_ENGINE_CLASS = 0, + UK_VDECENC_ENGINE_CLASS = 1, + UK_VE_ENGINE_CLASS = 2, + UK_BLT_COPY_ENGINE_CLASS = 3, + UK_RESERVED_ENGINE_CLASS = 4, + UK_OTHER_ENGINE_CLASS = 5, +}; + +/* Engine Instance that uKernel can schedule on */ +enum uk_engine_instance { + UK_ENGINE_INSTANCE_0 = 0, + UK_ENGINE_INSTANCE_1 = 1, + UK_ENGINE_INSTANCE_2 = 2, + UK_ENGINE_INSTANCE_3 = 3, + UK_INVALID_ENGINE_INSTANCE = GUC_MAX_ENGINE_INSTANCE_PER_CLASS, + UK_ENGINE_ALL_INSTANCES = UK_INVALID_ENGINE_INSTANCE +}; + +/* Target Engine field used in WI header and Guc2Host */ +struct guc_target_engine { + union { + struct { + /* One of enum uk_engine_class */ + u8 engine_class : 3; + /* One of enum uk_engine_instance */ + u8 engine_instance : 4; + /* All enabled engine classes and instances */ + u8 all_engines : 1; + }; + u8 value; + }; +}; + +struct guc_engine_class_bit_map { + union { + /* Bit positions must match enum uk_engine_class value */ + struct { + u32 render_engine_class : 1; + u32 vdecenc_engine_class : 1; + u32 ve_engine_class : 1; + u32 blt_copy_engine_class : 1; + u32 reserved_engine_class : 1; + u32 other_engine_class : 1; + u32 : 26; + }; + u32 value; + }; +}; + +struct guc_engine_instance_bit_map { + union { + struct { + u32 engine0 : 1; + u32 engine1 : 1; + u32 engine2 : 1; + u32 engine3 : 1; + u32 engine4 : 1; + u32 engine5 : 1; + u32 engine6 : 1; + u32 engine7 : 1; + u32 : 24; + }; + u32 value; + }; +}; + +struct guc_engine_bit_map { + struct guc_engine_class_bit_map engine_class_bit_map; + struct guc_engine_instance_bit_map + engine_instance_bit_map[GUC_MAX_ENGINE_CLASS_COUNT]; +}; + +/***************************************************************************** + ********************* Process Descriptor and Work Queue ********************* + *****************************************************************************/ + +/* Status of a Work Queue */ +enum guc_queue_status { + GUC_QUEUE_STATUS_ACTIVE = 1, + GUC_QUEUE_STATUS_SUSPENDED = 2, + GUC_QUEUE_STATUS_CMD_ERROR = 3, + GUC_QUEUE_STATUS_ENGINE_ID_NOT_USED = 4, + GUC_QUEUE_STATUS_SUSPENDED_FROM_ENGINE_RESET = 5, + GUC_QUEUE_STATUS_INVALID_STATUS = 6, +}; + +/* Priority of guc_context_descriptor */ +enum guc_context_priority { + GUC_CONTEXT_PRIORITY_KMD_HIGH = 0, + GUC_CONTEXT_PRIORITY_HIGH = 1, + GUC_CONTEXT_PRIORITY_KMD_NORMAL = 2, + GUC_CONTEXT_PRIORITY_NORMAL = 3, + GUC_CONTEXT_PRIORITY_ABSOLUTE_MAX_COUNT = 4, + GUC_CONTEXT_PRIORITY_INVALID = GUC_CONTEXT_PRIORITY_ABSOLUTE_MAX_COUNT +}; + +/* A shared structure between app and uKernel for communication */ +struct guc_sched_process_descriptor { + /* Index in the GuC Context Descriptor Pool */ + u32 context_id; + + /* Pointer to doorbell cacheline. BSpec: 1116 */ + u64 p_doorbell; + + /* WQ Head Byte Offset - Client must not write here */ + u32 head_offset; + + /* WQ Tail Byte Offset - uKernel will not write here */ + u32 tail_offset; + + /* WQ Error Byte offset */ + u32 error_offset_byte; + + /* WQ pVirt base address in Client. For use only by Client */ + u64 wqv_base_address; + + /* WQ Size in Bytes */ + u32 wq_size_bytes; + + /* WQ Status. Read by Client. Written by uKernel/KMD */ + enum guc_queue_status wq_status; + + /* Context priority. Read only by Client */ + enum guc_context_priority priority_assigned; + + u32 future; + + struct guc_engine_class_bit_map queue_engine_error; + + u32 reserved0[3]; + + /* uKernel side tracking for debug */ + + /* Written by uKernel at the time of parsing and successful removal + * from WQ (implies ring tail was updated) + */ + u32 total_work_items_parsed_by_guc; + + /* Written by uKernel if a WI was collapsed if next WI is the same + * LRCA (optimization applies only to Secure/KMD contexts) + */ + u32 total_work_items_collapsed_by_guc; + + /* Tells if the context is affected by Engine Reset. UMD needs to + * clear it after taking appropriate Action (TBD) + */ + u32 is_context_in_engine_reset; + + /* WQ Sampled tail at Engine Reset Time. Valid only if + * is_context_in_engine_reset = true + */ + u32 engine_reset_sampled_wq_tail; + + /* Valid from engine reset until all the affected Work Items are + * processed + */ + u32 engine_reset_sampled_wq_tail_valid; + + u32 reserved1[15]; +}; + +/* Work Item for submitting KMD workloads into Work Queue for GuC */ +struct guc_sched_work_queue_kmd_element_info { + /* Execlist context descriptor's lower DW. BSpec: 12254 */ + u32 element_low_dw; + union { + struct { + /* SW Context ID. BSpec: 12254 */ + u32 sw_context_index : 11; + /* SW Counter. BSpec: 12254 */ + u32 sw_context_counter : 6; + /* If this workload needs to be synced prior + * to submission use context_submit_sync_value and + * context_submit_sync_address + */ + u32 needs_sync : 1; + /* QW Aligned, TailValue <= 2048 + * (addresses 4 pages max) + */ + u32 ring_tail_qw_index : 11; + u32 : 2; + /* Bit to indicate if POCS needs to be in FREEZE state + * for this WI submission + */ + u32 wi_freeze_pocs : 1; + }; + u32 value; + } element_high_dw; +}; + +/* Work Item instruction type */ +enum guc_sched_instruction_type { + GUC_SCHED_INSTRUCTION_BATCH_BUFFER_START = 0x1, + GUC_SCHED_INSTRUCTION_GUC_CMD_PSEUDO = 0x2, + GUC_SCHED_INSTRUCTION_GUC_CMD_KMD = 0x3, + GUC_SCHED_INSTRUCTION_GUC_CMD_NOOP = 0x4, + GUC_SCHED_INSTRUCTION_RESUME_ENGINE_WQ_PARSING = 0x5, + GUC_SCHED_INSTRUCTION_INVALID = 0x6, +}; + +/* Header for every Work Item put in the Work Queue */ +struct guc_sched_work_queue_item_header { + union { + struct { + /* One of enum guc_sched_instruction_type */ + u32 work_instruction_type : 8; + /* struct guc_target_engine */ + u32 target_engine : 8; + /* Length in number of dwords following this header */ + u32 command_length_dwords : 11; + u32 : 5; + }; + u32 value; + }; +}; + +/* Work item for submitting KMD workloads into work queue for GuC */ +struct guc_sched_work_queue_item { + struct guc_sched_work_queue_item_header header; + struct guc_sched_work_queue_kmd_element_info kmd_submit_element_info; + /* Debug only */ + u32 fence_id; +}; + +struct km_gen11_resume_work_queue_processing_item { + struct guc_sched_work_queue_item header; +}; + +#pragma pack() + +#endif diff --git a/drivers/gpu/drm/i915/intel_guc_kmd_interface.h b/drivers/gpu/drm/i915/intel_guc_kmd_interface.h new file mode 100644 index 0000000..4c643ad --- /dev/null +++ b/drivers/gpu/drm/i915/intel_guc_kmd_interface.h @@ -0,0 +1,847 @@ +/* + * SPDX-License-Identifier: MIT + * + * Copyright © 2018 Intel Corporation + */ + +#ifndef _INTEL_GUC_KMD_INTERFACE_H_ +#define _INTEL_GUC_KMD_INTERFACE_H_ + +#include "intel_guc_client_interface.h" + +#pragma pack(1) + +/* Maximum number of entries in the GuC Context Descriptor Pool. Upper limit + * restricted by number of 'SW Context ID' bits in the Context Descriptor + * (BSpec: 12254) minus some reserved entries + */ +#define GUC_MAX_GUC_CONTEXT_DESCRIPTOR_ENTRIES 2032 + +/* Limited by 'SW Counter' bits. BSpec: 12254 */ +#define GUC_MAX_SW_CONTEXT_COUNTER 64 + +/* Maximum depth of HW Execlist Submission Queue. BSpec: 18934 */ +#define GUC_MAX_SUBMISSION_Q_DEPTH 8 + +/* Minimum depth of HW Execlist Submission Queue. BSpec: 18934 */ +#define GUC_MIN_SUBMISSION_Q_DEPTH 2 + +/* Default depth of HW Execlist Submission Queue. BSpec: 18934 */ +#define GUC_DEFAULT_ELEM_IN_SUBMISSION_Q GUC_MIN_SUBMISSION_Q_DEPTH + +/* 1 Cacheline = 64 Bytes */ +#define GUC_DMA_CACHELINE_SIZE_BYTES 64 + +/***************************************************************************** + ************************** Engines and System Info ************************** + *****************************************************************************/ + +/* GT system info passed down by KMD after reading fuse registers */ +struct guc_gt_system_info { + u32 slice_enabled; + u32 rcs_enabled; + u32 future0; + u32 bcs_enabled; + u32 vd_box_enable_mask; + u32 future1; + u32 ve_box_enable_mask; + u32 future2; + u32 reserved[8]; +}; + +/***************************************************************************** + ************************ GuC Context Descriptor Pool ************************ + *****************************************************************************/ + +/* State of the context */ +struct guc_engine_context_state { + union { + struct { + u32 wait_for_display_event : 1; + u32 wait_for_semaphore : 1; + u32 re_enqueue_to_submit_queue : 1; + u32 : 29; + }; + u32 wait_value; + }; + u32 reserved; +}; + +/* To describe status and access information of current ring buffer for + * a given guc_execlist_context + */ +struct guc_execlist_ring_buffer { + u32 p_execlist_ring_context; + + /* uKernel address for the ring buffer */ + u32 p_ring_begin; + /* uKernel final byte address that is valid for this ring */ + u32 p_ring_end; + /* uKernel address for next location in ring */ + u32 p_next_free_location; + + /* Last value written by software for tracking (just in case + * HW corrupts the tail in its context) + */ + u32 current_tail_pointer_value; +}; + +/* The entire execlist context including software and HW information */ +struct guc_execlist_context { + /* 2 DWs of Context Descriptor. BSpec: 12254 */ + u32 hw_context_desc_dw[2]; + u32 reserved0; + + struct guc_execlist_ring_buffer ring_buffer_obj; + struct guc_engine_context_state state; + + /* Flag to track if execlist context exists in submit queue + * Valid values 0 or 1 + */ + u32 is_present_in_sq; + + /* If needs_sync is set in WI, sync *context_submit_sync_address == + * context_submit_sync_value before submitting the context to HW + */ + u32 context_submit_sync_value; + u32 context_submit_sync_address; + + /* Reserved for SLPC hints (currently used for GT throttle modes) */ + u32 slpc_context_hints; + + u32 reserved1[4]; +}; + +/* Bitmap to track allocated and free contexts + * context_alloct_bit_map[n] = 0; Context 'n' free + * context_alloct_bit_map[n] = 1; Context 'n' allocated + */ +struct guc_execlist_context_alloc_map { + /* Bit map for execlist contexts, bits 0 to + * (GUC_MAX_SW_CONTEXT_COUNTER - 1) are valid + */ + u64 context_alloct_bit_map; + u32 reserved; +}; + +enum guc_context_descriptor_type { + /* Work will be submitted through doorbell and WQ of a + * Proxy Submission descriptor in the context desc pool + */ + GUC_CONTEXT_DESCRIPTOR_TYPE_PROXY_ENTRY = 0x00, + + /* Work will be submitted using doorbell and workqueue + * of this descriptor on behalf of other proxy Entries + * in the context desc pool + */ + GUC_CONTEXT_DESCRIPTOR_TYPE_PROXY_SUBMISSION = 0x01, + + /* Work is submitted through its own doorbell and WQ */ + GUC_CONTEXT_DESCRIPTOR_TYPE_REAL = 0x02, +}; + +/* CPU, Graphics and physical addresses */ +struct guc_address { + /* Cpu address (virtual) */ + u64 p_cpu_address; + /* uKernel address (gfx) */ + u32 p_uk_address; + /* Physical address */ + u64 p_address_gpa; +}; + +/* Context descriptor for communication between uKernel and KMD */ +struct guc_context_descriptor { + /* CPU back pointer for general KMD usage */ + u64 assigned_guc_gpu_desc; + + /* Index in the pool */ + u32 guc_context_desc_pool_index; + + /* For a Proxy Entry, this is the index of it's proxy submission entry. + * For others this is the same as guc_context_desc_pool_index above + */ + u32 proxy_submission_guc_context_desc_pool_index; + + /* The doorbell page's trigger cacheline */ + struct guc_address doorbell_trigger_address; + + /* Assigned doorbell */ + u32 doorbell_id; + + /* Array of execlist contexts */ + struct guc_execlist_context + uk_exec_list_context[GUC_MAX_SCHEDULABLE_ENGINE_CLASS] + [GUC_MAX_SW_CONTEXT_COUNTER]; + + /* Allocation map to track which execlist contexts are in use */ + struct guc_execlist_context_alloc_map + uk_execlist_context_alloc_map[GUC_MAX_SCHEDULABLE_ENGINE_CLASS]; + + /* Number of active execlist contexts */ + u32 uk_execlist_context_alloc_count; + + /* Optimization to reduce the maximum execlist context count for + * this GuC context descriptor. Should be less than + * GUC_MAX_SW_CONTEXT_COUNTER + */ + u32 max_uk_execlist_context_per_engine_class; + + union { + struct { + /* Is this context actively assigned to an app? */ + u32 is_context_active : 1; + + /* Is this a proxy entry, principal or real entry? */ + u32 context_type : 2; + + u32 is_kmd_created_context : 1; + + /* Context was part of an engine reset. KMD must take + * appropriate action (this context will not be + * resubmitted until this bit is cleared) + */ + u32 is_context_eng_reset : 1; + + /* Set it to 1 to prevent other code paths to do work + * queue processing as we use sampled values for WQ + * processing. Allowing multiple code paths to do WQ + * processing will cause same workload to execute + * multiple times + */ + u32 wq_processing_locked : 1; + + u32 future : 1; + + /* If set to 1, the context is terminated by GuC. All + * the pending work is dropped, its doorbell is evicted + * and eventually this context will be removed + */ + u32 is_context_terminated : 1; + + u32 : 24; + }; + u32 bool_values; + }; + + enum guc_context_priority priority; + + /* WQ tail Sampled and set during doorbell ISR handler */ + u32 wq_sampled_tail_offset; + + /* Global (across all submit queues). For principals + * (proxy entry), this will be zero and true count + * will be reflected in its proxy (proxy submission) + */ + u32 total_submit_queue_enqueues; + + /* Pointer to struct guc_sched_process_descriptor */ + u32 p_process_descriptor; + + /* Secure copy of WQ address and size. uKernel can not + * trust data in struct guc_sched_process_descriptor + */ + u32 p_work_queue_address; + u32 work_queue_size_bytes; + + u32 future0; + u32 future1; + + struct guc_engine_class_bit_map queue_engine_error; + + u32 reserved0[3]; + u64 reserved1[12]; +}; + +/***************************************************************************** + *************************** Host2GuC and GuC2Host *************************** + *****************************************************************************/ + +/* Host 2 GuC actions */ +enum guc_host2guc_action { + GUC_HOST2GUC_ACTION_DEFAULT = 0x0, + GUC_HOST2GUC_ACTION_REQUEST_INIT_DONE_INTERRUPT = 0x1, + GUC_HOST2GUC_ACTION_REQUEST_PREEMPTION = 0x2, + GUC_HOST2GUC_ACTION_REQUEST_ENGINE_RESET = 0x3, + GUC_HOST2GUC_ACTION_PAUSE_SCHEDULING = 0x4, + GUC_HOST2GUC_ACTION_RESUME_SCHEDULING = 0x5, + + GUC_HOST2GUC_ACTION_ALLOCATE_DOORBELL = 0x10, + GUC_HOST2GUC_ACTION_DEALLOCATE_DOORBELL = 0x20, + GUC_HOST2GUC_ACTION_LOG_BUFFER_FILE_FLUSH_COMPLETE = 0x30, + GUC_HOST2GUC_ACTION_ENABLE_LOGGING = 0x40, + GUC_HOST2GUC_ACTION_CACHE_CRASH_DUMP = 0x200, + GUC_HOST2GUC_ACTION_DEBUG_RING_DB = 0x300, + GUC_HOST2GUC_ACTION_PERFORM_GLOBAL_DEBUG_ACTIONS = 0x301, + GUC_HOST2GUC_ACTION_FORCE_LOGBUFFERFLUSH = 0x302, + GUC_HOST2GUC_ACTION_LOG_VERBOSITY_LOGOUTPUT_SELECT = 0x400, + GUC_HOST2GUC_ACTION_ENTER_S_STATE = 0x501, + GUC_HOST2GUC_ACTION_EXIT_S_STATE = 0x502, + GUC_HOST2GUC_ACTION_SET_SCHEDULING_MODE = 0x504, + GUC_HOST2GUC_ACTION_SCHED_POLICY_CHANGE = 0x506, + + /* Actions for Powr Conservation : 0x3000-0x3FFF */ + GUC_HOST2GUC_ACTION_PC_SLPM_REQUEST = 0x3003, + GUC_HOST2GUC_ACTION_PC_SETUP_GUCRC = 0x3004, + GUC_HOST2GUC_ACTION_SAMPLE_FORCEWAKE_FEATURE_REGISTER = 0x3005, + GUC_HOST2GUC_ACTION_SETUP_GUCRC = 0x3006, + + GUC_HOST2GUC_ACTION_AUTHENTICATE_HUC = 0x4000, + + GUC_HOST2GUC_ACTION_REGISTER_COMMAND_TRANSPORT_BUFFER = 0x4505, + GUC_HOST2GUC_ACTION_DEREGISTER_COMMAND_TRANSPORT_BUFFER = 0x4506, + + GUC_HOST2GUC_ACTION_MAX = 0xFFFF +}; + +enum guc_host2guc_response_status { + GUC_HOST2GUC_RESPONSE_STATUS_SUCCESS = 0x0, + GUC_HOST2GUC_RESPONSE_STATUS_UNKNOWN_ACTION = 0x30, + GUC_HOST2GUC_RESPONSE_STATUS_LOG_HOST_ADDRESS_NOT_VALID = 0x80, + GUC_HOST2GUC_RESPONSE_STATUS_GENERIC_FAIL = 0xF000, +}; + +enum { + /* Host originating types */ + GUC_MSG_TYPE_HOST2GUC_REQUEST = 0x0, + + /* GuC originating types */ + GUC_MSG_TYPE_HOST2GUC_RESPONSE = 0xF, +} GUC_GUC_MSG_TYPE; + +/* This structure represents the various formats of values put in + * SOFT_SCRATCH_0. The "Type" field is to determine which register + * definition to use, so it must be common among all unioned + * structs. + */ +struct guc_msg_format { + union { + struct { + u32 action : 16; /* enum guc_host2guc_action */ + u32 reserved : 12; /* MBZ */ + u32 type : 4; /* GUC_MSG_TYPE_HOST2GUC_REQUEST */ + } host2guc_action; + + struct { + u32 status : 16; /* enum guc_host2guc_response_status */ + u32 return_data : 12; + u32 type : 4; /* GUC_MSG_TYPE_HOST2GUC_RESPONSE */ + } host2guc_response; + + u32 dword_value; + }; +}; + +#define GUC_MAKE_HOST2GUC_RESPONSE(_status, _return_data) \ + ((GUC_MSG_TYPE_HOST2GUC_RESPONSE << 28) | \ + ((_return_data & 0xFFF) << 16) | \ + (_status & 0xFFFF)) +#define GUC_MAKE_HOST2GUC_STATUS(a) (GUC_MAKE_HOST2GUC_RESPONSE(a, 0)) + +enum guc_cmd_transport_buffer_type { + GUC_CMD_TRANSPORT_BUFFER_HOST2GUC = 0x00, + GUC_CMD_TRANSPORT_BUFFER_GUC2HOST = 0x01, + GUC_CMD_TRANSPORT_BUFFER_MAX_TYPE = 0x02, +}; + +struct guc_cmd_transport_buffer_desc { + u32 buffer_begin_gfx_address; + u64 buffer_begin_virtual_address; + u32 buffer_size_in_bytes; + /* GuC uKernel updates this */ + u32 head_offset; + /* GuC client updates this */ + u32 tail_offset; + u32 is_in_error; + /* A DW provided by H2G item that was requested to be written */ + u32 fence_report_dw; + /* Status associated with above fence_report_dw */ + u32 status_report_dw; + /* ID associated with this buffer (assigned by GuC master) */ + u32 client_id; + /* Used & set by the client for further tracking of internal clients */ + u32 client_sub_tracking_id; + u32 reserved[5]; +}; + +/* Per client command transport buffer allocated by GuC master */ +struct guc_master_cmd_transport_buffer_alloc { + /* This is the copy that GuC trusts */ + struct guc_cmd_transport_buffer_desc buffer_desc; + u32 future; + u64 reserved0; + u32 usage_special_info; + u32 valid; + u32 associated_g2h_index; + u32 reserved1; +}; + +/* Host 2 GuC Work Item + * V-----------------------------------------------------------------------V + * ************************************************************************* + * * * DW0/ * * * * + * * H2G Item Header * ReturnDW * DW1 * ... * DWn * + * ************************************************************************* + */ + +/* Command buffer header */ +struct guc_cmd_buffer_item_header { + union { + struct { + /* Number of dwords that are parameters of this + * Host2GuC action. Max of 31. E.g.: if there are 2 DWs + * following this header, this field is set to 2 + */ + u32 num_dwords : 5; + + u32 : 3; + + /* The uKernel will write the value from DW0 (aka + * ReturnDW) to fence_report_dw in struct + * guc_cmd_transport_buffer_desc + */ + u32 write_fence_from_dw0_to_descriptor : 1; + + /* Write the status of the action to DW0 following this + * header + */ + u32 write_status_to_dw0 : 1; + + /* Send a GuC2Host with Status of the action and the + * fence ID found in DW0 via the buffer used for GuC to + * Host communication + */ + u32 send_status_with_dw0_via_guc_to_host : 1; + + u32 : 5; + + /* This is the value of the enum guc_host2guc_action + * that needs to be done by the uKernel + */ + u32 host2guc_action : 16; + } h2g_cmd_buffer_item_header; + + struct { + /* Number of dwords that are parameters of this GuC2Host + * action + */ + u32 num_dwords : 5; + + u32 : 3; + + /* Indicates that this GuC2Host action is a response of + * a Host2Guc request + */ + u32 host2_guc_response : 1; + + u32 reserved : 7; + + /* struct guc_to_host_message */ + u32 guc2host_action : 16; + } g2h_cmd_buffer_item_header; + + struct { + u32 num_dwords : 5; + u32 reserved : 3; + u32 free_for_client_use : 24; + } generic_cmd_buffer_item_header; + + u32 header_value; + }; + +}; + +struct guc_to_host_message { + union { + struct { + u32 uk_init_done : 1; + u32 crash_dump_posted : 1; + u32 : 1; + u32 flush_log_buffer_to_file : 1; + u32 preempt_request_old_preempt_pending : 1; + u32 preempt_request_target_context_bad : 1; + u32 : 1; + u32 sleep_entry_in_progress : 1; + u32 guc_in_debug_halt : 1; + u32 guc_report_engine_reset_context_id : 1; + u32 : 1; + u32 host_preemption_complete : 1; + u32 reserved0 : 4; + u32 gpa_to_hpa_xlation_error : 1; + u32 doorbell_id_allocation_error : 1; + u32 doorbell_id_allocation_invalid_ctx_id : 1; + u32 : 1; + u32 force_wake_timed_out : 1; + u32 force_wake_time_out_counter : 2; + u32 : 1; + u32 iommu_cat_page_faulted : 1; + u32 host2guc_engine_reset_complete : 1; + u32 reserved1 : 2; + u32 doorbell_selection_error : 1; + u32 doorbell_id_release_error : 1; + u32 uk_exception : 1; + u32 : 1; + }; + u32 dw; + }; + +}; + +/* Size of the buffer to save GuC's state before S3. The ddress of the buffer + * goes in guc_additional_data_structs + */ +#define GUC_MAX_GUC_S3_SAVE_SPACE_PAGES 10 + +/* MMIO Offset for status of sleep state enter request */ +#define GUC_SLEEP_STATE_ENTER_STATUS 0xC1B8 + +/* Status of sleep request. Value updated in GUC_SLEEP_STATE_ENTER_STATUS */ +enum guc_sleep_state_enter_status { + GUC_SLEEP_STATE_ENTER_SUCCESS = 1, + GUC_SLEEP_STATE_ENTER_PREEMPT_TO_IDLE_FAILED = 2, + GUC_SLEEP_STATE_ENTER_ENG_RESET_FAILED = 3, +}; + + +/* Enum to determine what mode the scheduler is in */ +enum guc_scheduler_mode { + GUC_SCHEDULER_MODE_NORMAL, + GUC_SCHEDULER_MODE_STALL_IMMEDIATE, +}; + +/***************************************************************************** + ********************************** Logging ********************************** + *****************************************************************************/ + +enum guc_log_buffer_type { + GUC_LOG_BUFFER_TYPE_ISR = 0x0, + GUC_LOG_BUFFER_TYPE_DPC = 0x1, + GUC_LOG_BUFFER_TYPE_CRASH = 0x2, + GUC_LOG_BUFFER_TYPE_MAX = 0x3, +}; + +enum guc_log_verbosity { + GUC_LOG_VERBOSITY_LOW = 0x0, + GUC_LOG_VERBOSITY_MED = 0x1, + GUC_LOG_VERBOSITY_HIGH = 0x2, + GUC_LOG_VERBOSITY_ULTRA = 0x3, + GUC_LOG_VERBOSITY_MAX = 0x4, +}; + +/* This enum controls the type of logging output. Can be changed dynamically + * using GUC_HOST2GUC_ACTION_LOG_VERBOSITY_LOGOUTPUT_SELECT + */ +enum guc_logoutput_selection { + GUC_LOGOUTPUT_LOGBUFFER_ONLY = 0x0, + GUC_LOGOUTPUT_NPK_ONLY = 0x1, + GUC_LOGOUTPUT_LOGBUFFER_AND_NPK = 0x2, + GUC_LOGOUTPUT_MAX +}; + +/* Filled by KMD except version and marker that are initialized by uKernel */ +struct guc_km_log_buffer_state { + /* Marks the beginning of Buffer Flush (set by uKernel at Log Buffer + * Init) + */ + u32 marker[2]; + + /* This is the last byte offset location that was read by KMD. KMD will + * write to this and uKernel will read it + */ + u32 log_buf_rd_ptr; + + /* This is the byte offset location that will be written by uKernel */ + u32 log_buf_wr_ptr; + + u32 log_buf_size; + + /* This is written by uKernel when it sees the log buffer becoming half + * full. KMD writes this value in the log file to avoid stale data + */ + u32 sampled_log_buf_wrptr; + + union { + struct { + /* uKernel sets this when log buffer is half full or + * when a forced flush has been requested through + * Host2Guc. uKernel will send Guc2Host only if this + * bit is cleared. This is to avoid unnecessary + * interrupts from GuC + */ + u32 log_buf_flush_to_file : 1; + + /* uKernel increments this when log buffer overflows */ + u32 buffer_full_count : 4; + + u32 : 27; + }; + u32 log_buf_flags; + }; + + u32 version; +}; + +/* Logging Parameters sent via struct sched_control_data. Maintained as separate + * structure to allow debug tools to access logs without contacting GuC (for + * when GuC is stuck in ISR) + */ +struct guc_log_init_params { + union { + struct { + u32 is_log_buffer_valid : 1; + /* Raise GuC2Host interrupt when buffer is half full */ + u32 notify_on_log_half_full : 1; + u32 : 1; + /* 0 = Pages, 1 = Megabytes */ + u32 allocated_count_units : 1; + /* No. of units allocated -1 (MAX 4 Units) */ + u32 crash_dump_log_allocated_count : 2; + /* No. of units allocated -1 (MAX 8 Units) */ + u32 dpc_log_allocated_count : 3; + /* No. of units allocated -1 (MAX 8 Units) */ + u32 isr_log_allocated_count : 3; + /* Page aligned address for log buffer */ + u32 log_buffer_gfx_address : 20; + }; + u32 log_dword_value; + }; +}; + +/* Pass info for doing a Host2GuC request (GUC_HOST2GUC_ACTION_ENABLE_LOGGING) + * in order to enable/disable GuC logging + */ +struct guc_log_enable_params { + union { + struct { + u32 logging_enabled : 1; + u32 profile_logging_enabled : 1; + u32 log_output_selection : 2; + u32 log_verbosity : 4; + u32 default_logging_enabled : 1; + u32 : 23; + }; + u32 log_enable_dword_value; + }; + +}; + +/***************************************************************************** + ************** Sched Control Data and Addtional Data Structures ************* + *****************************************************************************/ + +/* Holds the init values of various parameters used by the uKernel */ +struct guc_sched_control_data { + /* Dword 0 */ + union { + struct { + /* Num of contexts in pool in blocks of 16, + * E.g.: num_contexts_in_pool16_blocks = 1 if 16 + * contexts, 64 if 1024 contexts allocated + */ + u32 num_contexts_in_pool16_blocks : 12; + + /* Aligned bits [31:12] of the GFX address where the + * pool begins + */ + u32 context_pool_gfx_address_begin : 20; + }; + }; + + /* Dword 1 */ + struct guc_log_init_params log_init_params; + + + /* Dword 2 */ + union { + struct { + u32 reserved : 1; + u32 wa_disable_dummy_all_engine_fault_fix : 1; + u32 : 30; + }; + u32 workaround_dw; + }; + + /* Dword 3 */ + union { + struct { + u32 ftr_enable_preemption_data_logging : 1; + u32 ftr_enable_guc_pavp_control : 1; + u32 ftr_enable_guc_slpm : 1; + u32 ftr_enable_engine_reset_on_preempt_failure : 1; + u32 ftr_lite_restore : 1; + u32 ftr_driver_flr : 1; + u32 future : 1; + u32 ftr_enable_psmi_logging : 1; + u32 : 1; + u32 : 1; + u32 : 1; + u32 : 1; + u32 : 1; + u32 : 1; + u32 : 18; + }; + u32 feature_dword; + }; + + /* Dword 4 */ + union { + struct { + /* One of enum guc_log_verbosity */ + u32 logging_verbosity : 4; + /* One of enum guc_logoutput_selection */ + u32 log_output_selection : 2; + u32 logging_disabled : 1; + u32 profile_logging_enabled : 1; + u32 : 24; + }; + }; + + /* Dword 5 */ + union { + struct { + u32 : 1; + u32 gfx_address_additional_data_structs : 21; + u32 : 10; + }; + }; + +}; + +/* Structure to pass additional information and structure pointers to */ +struct guc_additional_data_structs { + /* Gfx ptr to struct guc_mmio_save_restore_list (persistent) */ + u32 gfx_address_mmio_save_restore_list; + + /* Buffer of size GUC_MAX_GUC_S3_SAVE_SPACE_PAGES (persistent) */ + u32 gfx_ptr_to_gucs_state_save_buffer; + + /* Gfx addresses of struct guc_scheduling_policies (non-persistent, may + * be released after initial load), NULL or valid = 0 flag value will + * cause default policies to be loaded + */ + u32 gfx_scheduler_policies; + + /* Gfx address of struct guc_gt_system_info */ + u32 gt_system_info; + + u32 future; + + u32 gfx_ptr_to_psmi_log_control_data; + + /* LRCA addresses and sizes of golden contexts (persistent) */ + u32 gfx_golden_context_lrca[GUC_MAX_SCHEDULABLE_ENGINE_CLASS]; + u32 golden_context_eng_state_size_in_bytes[GUC_MAX_SCHEDULABLE_ENGINE_CLASS]; + + u32 reserved[16]; +}; + +/* Max number of mmio per engine class per engine instance */ +#define GUC_MAX_MMIO_PER_SET 64 + +struct guc_mmio_flags { + union { + struct { + u32 masked : 1; + u32 : 31; + }; + u32 flags_value; + }; +}; + +struct guc_mmio { + u32 offset; + u32 value; + struct guc_mmio_flags flags; +}; + +struct guc_mmio_set { + /* Array of mmio to be saved/restored */ + struct guc_mmio mmio[GUC_MAX_MMIO_PER_SET]; + /* Set after saving mmio value, cleared after restore. */ + u32 mmio_values_valid; + /* Number of mmio in the set */ + u32 number_of_mmio; +}; + +struct guc_mmio_save_restore_list { + struct guc_mmio_set + node_mmio_set[GUC_MAX_SCHEDULABLE_ENGINE_CLASS] + [GUC_MAX_ENGINE_INSTANCE_PER_CLASS]; + u32 reserved[98]; +}; + +/* Policy flags to control scheduling decisions */ +struct guc_scheduling_policy_flags { + union { + struct { + /* Should we reset engine when preemption failed within + * its time quantum + */ + u32 reset_engine_upon_preempt_failure : 1; + + /* Should we preempt to idle unconditionally for the + * execution quantum expiry + */ + u32 preempt_to_idle_on_quantum_expiry : 1; + + u32 : 30; + }; + u32 policy_dword; + }; +}; + +/* Per-engine class and per-priority struct for scheduling policy */ +struct guc_scheduling_policy { + /* Time for one workload to execute (micro seconds) */ + u32 execution_quantum; + + /* Time to wait for a preemption request to completed before issuing a + * reset (micro seconds) + */ + u32 wait_for_preemption_completion_time; + + /* How much time to allow to run after the first fault is observed. + * Then preempt afterwards (micro seconds) + */ + u32 quantum_upon_first_fault_time; + + struct guc_scheduling_policy_flags policy_flags; + + u32 reserved[8]; +}; + +/* KMD should populate this struct and pass info through struct + * guc_additional_data_structs- If KMD does not set the scheduler policy, + * uKernel will fall back to default scheduling policies + */ +struct guc_scheduling_policies { + struct guc_scheduling_policy + per_submit_queue_policy[GUC_CONTEXT_PRIORITY_ABSOLUTE_MAX_COUNT] + [GUC_MAX_SCHEDULABLE_ENGINE_CLASS]; + + /* Submission queue depth, min 2, max 8. If outside the valid range, + * default value is used + */ + u32 submission_queue_depth[GUC_MAX_SCHEDULABLE_ENGINE_CLASS]; + + /* How much time to allow before DPC processing is called back via + * interrupt (to prevent DPC queue drain starving) IN micro seconds. + * Typically in the 1000s (example only, not granularity) + */ + u32 dpc_promote_time; + + /* Must be set to take these new values */ + u32 is_valid; + + /* Number of WIs to process per call to process single. Process single + * could have a large Max Tail value which may keep CS idle. Process + * max_num_work_items_per_dpc_call WIs and try fast schedule + */ + u32 max_num_work_items_per_dpc_call; + + u32 reserved[4]; +}; + +#pragma pack() + +#endif -- 1.9.1 _______________________________________________ Intel-gfx mailing list Intel-gfx@xxxxxxxxxxxxxxxxxxxxx https://lists.freedesktop.org/mailman/listinfo/intel-gfx