On Fri, May 09, 2014 at 01:08:36PM +0100, oscar.mateo@xxxxxxxxx wrote: > From: Oscar Mateo <oscar.mateo@xxxxxxxxx> > > In the upcoming patches, we plan to break the correlation between > engines (a.k.a. rings) and ringbuffers, so it makes sense to > refactor the code and make the change obvious. > > No functional changes. > > Signed-off-by: Oscar Mateo <oscar.mateo@xxxxxxxxx> If we rename stuff I'd vote for something close to Bspec language, like CS. So maybe intel_cs_engine? /me sucks at this naming game -Daniel > --- > drivers/gpu/drm/i915/i915_cmd_parser.c | 16 +-- > drivers/gpu/drm/i915/i915_debugfs.c | 16 +-- > drivers/gpu/drm/i915/i915_dma.c | 10 +- > drivers/gpu/drm/i915/i915_drv.h | 32 +++--- > drivers/gpu/drm/i915/i915_gem.c | 58 +++++------ > drivers/gpu/drm/i915/i915_gem_context.c | 14 +-- > drivers/gpu/drm/i915/i915_gem_execbuffer.c | 18 ++-- > drivers/gpu/drm/i915/i915_gem_gtt.c | 18 ++-- > drivers/gpu/drm/i915/i915_gem_gtt.h | 2 +- > drivers/gpu/drm/i915/i915_gpu_error.c | 6 +- > drivers/gpu/drm/i915/i915_irq.c | 28 ++--- > drivers/gpu/drm/i915/i915_trace.h | 26 ++--- > drivers/gpu/drm/i915/intel_display.c | 14 +-- > drivers/gpu/drm/i915/intel_drv.h | 4 +- > drivers/gpu/drm/i915/intel_overlay.c | 12 +-- > drivers/gpu/drm/i915/intel_pm.c | 10 +- > drivers/gpu/drm/i915/intel_ringbuffer.c | 158 ++++++++++++++--------------- > drivers/gpu/drm/i915/intel_ringbuffer.h | 76 +++++++------- > 18 files changed, 259 insertions(+), 259 deletions(-) > > diff --git a/drivers/gpu/drm/i915/i915_cmd_parser.c b/drivers/gpu/drm/i915/i915_cmd_parser.c > index 69d34e4..3234d36 100644 > --- a/drivers/gpu/drm/i915/i915_cmd_parser.c > +++ b/drivers/gpu/drm/i915/i915_cmd_parser.c > @@ -498,7 +498,7 @@ static u32 gen7_blt_get_cmd_length_mask(u32 cmd_header) > return 0; > } > > -static bool validate_cmds_sorted(struct intel_ring_buffer *ring) > +static bool validate_cmds_sorted(struct intel_engine *ring) > { > int i; > bool ret = true; > @@ -550,7 +550,7 @@ static bool check_sorted(int ring_id, const u32 *reg_table, int reg_count) > return ret; > } > > -static bool validate_regs_sorted(struct intel_ring_buffer *ring) > +static bool validate_regs_sorted(struct intel_engine *ring) > { > return check_sorted(ring->id, ring->reg_table, ring->reg_count) && > check_sorted(ring->id, ring->master_reg_table, > @@ -562,10 +562,10 @@ static bool validate_regs_sorted(struct intel_ring_buffer *ring) > * @ring: the ringbuffer to initialize > * > * Optionally initializes fields related to batch buffer command parsing in the > - * struct intel_ring_buffer based on whether the platform requires software > + * struct intel_engine based on whether the platform requires software > * command parsing. > */ > -void i915_cmd_parser_init_ring(struct intel_ring_buffer *ring) > +void i915_cmd_parser_init_ring(struct intel_engine *ring) > { > if (!IS_GEN7(ring->dev)) > return; > @@ -664,7 +664,7 @@ find_cmd_in_table(const struct drm_i915_cmd_table *table, > * ring's default length encoding and returns default_desc. > */ > static const struct drm_i915_cmd_descriptor* > -find_cmd(struct intel_ring_buffer *ring, > +find_cmd(struct intel_engine *ring, > u32 cmd_header, > struct drm_i915_cmd_descriptor *default_desc) > { > @@ -744,7 +744,7 @@ finish: > * > * Return: true if the ring requires software command parsing > */ > -bool i915_needs_cmd_parser(struct intel_ring_buffer *ring) > +bool i915_needs_cmd_parser(struct intel_engine *ring) > { > struct drm_i915_private *dev_priv = ring->dev->dev_private; > > @@ -763,7 +763,7 @@ bool i915_needs_cmd_parser(struct intel_ring_buffer *ring) > return (i915.enable_cmd_parser == 1); > } > > -static bool check_cmd(const struct intel_ring_buffer *ring, > +static bool check_cmd(const struct intel_engine *ring, > const struct drm_i915_cmd_descriptor *desc, > const u32 *cmd, > const bool is_master, > @@ -865,7 +865,7 @@ static bool check_cmd(const struct intel_ring_buffer *ring, > * > * Return: non-zero if the parser finds violations or otherwise fails > */ > -int i915_parse_cmds(struct intel_ring_buffer *ring, > +int i915_parse_cmds(struct intel_engine *ring, > struct drm_i915_gem_object *batch_obj, > u32 batch_start_offset, > bool is_master) > diff --git a/drivers/gpu/drm/i915/i915_debugfs.c b/drivers/gpu/drm/i915/i915_debugfs.c > index 103e62c..0052460 100644 > --- a/drivers/gpu/drm/i915/i915_debugfs.c > +++ b/drivers/gpu/drm/i915/i915_debugfs.c > @@ -562,7 +562,7 @@ static int i915_gem_request_info(struct seq_file *m, void *data) > struct drm_info_node *node = (struct drm_info_node *) m->private; > struct drm_device *dev = node->minor->dev; > struct drm_i915_private *dev_priv = dev->dev_private; > - struct intel_ring_buffer *ring; > + struct intel_engine *ring; > struct drm_i915_gem_request *gem_request; > int ret, count, i; > > @@ -594,7 +594,7 @@ static int i915_gem_request_info(struct seq_file *m, void *data) > } > > static void i915_ring_seqno_info(struct seq_file *m, > - struct intel_ring_buffer *ring) > + struct intel_engine *ring) > { > if (ring->get_seqno) { > seq_printf(m, "Current sequence (%s): %u\n", > @@ -607,7 +607,7 @@ static int i915_gem_seqno_info(struct seq_file *m, void *data) > struct drm_info_node *node = (struct drm_info_node *) m->private; > struct drm_device *dev = node->minor->dev; > struct drm_i915_private *dev_priv = dev->dev_private; > - struct intel_ring_buffer *ring; > + struct intel_engine *ring; > int ret, i; > > ret = mutex_lock_interruptible(&dev->struct_mutex); > @@ -630,7 +630,7 @@ static int i915_interrupt_info(struct seq_file *m, void *data) > struct drm_info_node *node = (struct drm_info_node *) m->private; > struct drm_device *dev = node->minor->dev; > struct drm_i915_private *dev_priv = dev->dev_private; > - struct intel_ring_buffer *ring; > + struct intel_engine *ring; > int ret, i, pipe; > > ret = mutex_lock_interruptible(&dev->struct_mutex); > @@ -800,7 +800,7 @@ static int i915_hws_info(struct seq_file *m, void *data) > struct drm_info_node *node = (struct drm_info_node *) m->private; > struct drm_device *dev = node->minor->dev; > struct drm_i915_private *dev_priv = dev->dev_private; > - struct intel_ring_buffer *ring; > + struct intel_engine *ring; > const u32 *hws; > int i; > > @@ -1677,7 +1677,7 @@ static int i915_context_status(struct seq_file *m, void *unused) > struct drm_info_node *node = (struct drm_info_node *) m->private; > struct drm_device *dev = node->minor->dev; > struct drm_i915_private *dev_priv = dev->dev_private; > - struct intel_ring_buffer *ring; > + struct intel_engine *ring; > struct i915_hw_context *ctx; > int ret, i; > > @@ -1826,7 +1826,7 @@ static int per_file_ctx(int id, void *ptr, void *data) > static void gen8_ppgtt_info(struct seq_file *m, struct drm_device *dev) > { > struct drm_i915_private *dev_priv = dev->dev_private; > - struct intel_ring_buffer *ring; > + struct intel_engine *ring; > struct i915_hw_ppgtt *ppgtt = dev_priv->mm.aliasing_ppgtt; > int unused, i; > > @@ -1850,7 +1850,7 @@ static void gen8_ppgtt_info(struct seq_file *m, struct drm_device *dev) > static void gen6_ppgtt_info(struct seq_file *m, struct drm_device *dev) > { > struct drm_i915_private *dev_priv = dev->dev_private; > - struct intel_ring_buffer *ring; > + struct intel_engine *ring; > struct drm_file *file; > int i; > > diff --git a/drivers/gpu/drm/i915/i915_dma.c b/drivers/gpu/drm/i915/i915_dma.c > index d02c8de..5263d63 100644 > --- a/drivers/gpu/drm/i915/i915_dma.c > +++ b/drivers/gpu/drm/i915/i915_dma.c > @@ -119,7 +119,7 @@ static void i915_write_hws_pga(struct drm_device *dev) > static void i915_free_hws(struct drm_device *dev) > { > struct drm_i915_private *dev_priv = dev->dev_private; > - struct intel_ring_buffer *ring = LP_RING(dev_priv); > + struct intel_engine *ring = LP_RING(dev_priv); > > if (dev_priv->status_page_dmah) { > drm_pci_free(dev, dev_priv->status_page_dmah); > @@ -139,7 +139,7 @@ void i915_kernel_lost_context(struct drm_device * dev) > { > struct drm_i915_private *dev_priv = dev->dev_private; > struct drm_i915_master_private *master_priv; > - struct intel_ring_buffer *ring = LP_RING(dev_priv); > + struct intel_engine *ring = LP_RING(dev_priv); > > /* > * We should never lose context on the ring with modesetting > @@ -234,7 +234,7 @@ static int i915_initialize(struct drm_device * dev, drm_i915_init_t * init) > static int i915_dma_resume(struct drm_device * dev) > { > struct drm_i915_private *dev_priv = dev->dev_private; > - struct intel_ring_buffer *ring = LP_RING(dev_priv); > + struct intel_engine *ring = LP_RING(dev_priv); > > DRM_DEBUG_DRIVER("%s\n", __func__); > > @@ -782,7 +782,7 @@ static int i915_wait_irq(struct drm_device * dev, int irq_nr) > struct drm_i915_private *dev_priv = dev->dev_private; > struct drm_i915_master_private *master_priv = dev->primary->master->driver_priv; > int ret = 0; > - struct intel_ring_buffer *ring = LP_RING(dev_priv); > + struct intel_engine *ring = LP_RING(dev_priv); > > DRM_DEBUG_DRIVER("irq_nr=%d breadcrumb=%d\n", irq_nr, > READ_BREADCRUMB(dev_priv)); > @@ -1073,7 +1073,7 @@ static int i915_set_status_page(struct drm_device *dev, void *data, > { > struct drm_i915_private *dev_priv = dev->dev_private; > drm_i915_hws_addr_t *hws = data; > - struct intel_ring_buffer *ring; > + struct intel_engine *ring; > > if (drm_core_check_feature(dev, DRIVER_MODESET)) > return -ENODEV; > diff --git a/drivers/gpu/drm/i915/i915_drv.h b/drivers/gpu/drm/i915/i915_drv.h > index b1725c6..3b7a36f9 100644 > --- a/drivers/gpu/drm/i915/i915_drv.h > +++ b/drivers/gpu/drm/i915/i915_drv.h > @@ -594,7 +594,7 @@ struct i915_hw_context { > bool is_initialized; > uint8_t remap_slice; > struct drm_i915_file_private *file_priv; > - struct intel_ring_buffer *last_ring; > + struct intel_engine *last_ring; > struct drm_i915_gem_object *obj; > struct i915_ctx_hang_stats hang_stats; > struct i915_address_space *vm; > @@ -1354,7 +1354,7 @@ struct drm_i915_private { > wait_queue_head_t gmbus_wait_queue; > > struct pci_dev *bridge_dev; > - struct intel_ring_buffer ring[I915_NUM_RINGS]; > + struct intel_engine ring[I915_NUM_RINGS]; > uint32_t last_seqno, next_seqno; > > drm_dma_handle_t *status_page_dmah; > @@ -1675,7 +1675,7 @@ struct drm_i915_gem_object { > void *dma_buf_vmapping; > int vmapping_count; > > - struct intel_ring_buffer *ring; > + struct intel_engine *ring; > > /** Breadcrumb of last rendering to the buffer. */ > uint32_t last_read_seqno; > @@ -1714,7 +1714,7 @@ struct drm_i915_gem_object { > */ > struct drm_i915_gem_request { > /** On Which ring this request was generated */ > - struct intel_ring_buffer *ring; > + struct intel_engine *ring; > > /** GEM sequence number associated with this request. */ > uint32_t seqno; > @@ -1755,7 +1755,7 @@ struct drm_i915_file_private { > > struct i915_hw_context *private_default_ctx; > atomic_t rps_wait_boost; > - struct intel_ring_buffer *bsd_ring; > + struct intel_engine *bsd_ring; > }; > > /* > @@ -2182,9 +2182,9 @@ static inline void i915_gem_object_unpin_pages(struct drm_i915_gem_object *obj) > > int __must_check i915_mutex_lock_interruptible(struct drm_device *dev); > int i915_gem_object_sync(struct drm_i915_gem_object *obj, > - struct intel_ring_buffer *to); > + struct intel_engine *to); > void i915_vma_move_to_active(struct i915_vma *vma, > - struct intel_ring_buffer *ring); > + struct intel_engine *ring); > int i915_gem_dumb_create(struct drm_file *file_priv, > struct drm_device *dev, > struct drm_mode_create_dumb *args); > @@ -2226,7 +2226,7 @@ i915_gem_object_unpin_fence(struct drm_i915_gem_object *obj) > } > > struct drm_i915_gem_request * > -i915_gem_find_active_request(struct intel_ring_buffer *ring); > +i915_gem_find_active_request(struct intel_engine *ring); > > bool i915_gem_retire_requests(struct drm_device *dev); > int __must_check i915_gem_check_wedge(struct i915_gpu_error *error, > @@ -2264,18 +2264,18 @@ bool i915_gem_clflush_object(struct drm_i915_gem_object *obj, bool force); > int __must_check i915_gem_object_finish_gpu(struct drm_i915_gem_object *obj); > int __must_check i915_gem_init(struct drm_device *dev); > int __must_check i915_gem_init_hw(struct drm_device *dev); > -int i915_gem_l3_remap(struct intel_ring_buffer *ring, int slice); > +int i915_gem_l3_remap(struct intel_engine *ring, int slice); > void i915_gem_init_swizzling(struct drm_device *dev); > void i915_gem_cleanup_ringbuffer(struct drm_device *dev); > int __must_check i915_gpu_idle(struct drm_device *dev); > int __must_check i915_gem_suspend(struct drm_device *dev); > -int __i915_add_request(struct intel_ring_buffer *ring, > +int __i915_add_request(struct intel_engine *ring, > struct drm_file *file, > struct drm_i915_gem_object *batch_obj, > u32 *seqno); > #define i915_add_request(ring, seqno) \ > __i915_add_request(ring, NULL, NULL, seqno) > -int __must_check i915_wait_seqno(struct intel_ring_buffer *ring, > +int __must_check i915_wait_seqno(struct intel_engine *ring, > uint32_t seqno); > int i915_gem_fault(struct vm_area_struct *vma, struct vm_fault *vmf); > int __must_check > @@ -2286,7 +2286,7 @@ i915_gem_object_set_to_cpu_domain(struct drm_i915_gem_object *obj, bool write); > int __must_check > i915_gem_object_pin_to_display_plane(struct drm_i915_gem_object *obj, > u32 alignment, > - struct intel_ring_buffer *pipelined); > + struct intel_engine *pipelined); > void i915_gem_object_unpin_from_display_plane(struct drm_i915_gem_object *obj); > int i915_gem_attach_phys_object(struct drm_device *dev, > struct drm_i915_gem_object *obj, > @@ -2388,7 +2388,7 @@ void i915_gem_context_reset(struct drm_device *dev); > int i915_gem_context_open(struct drm_device *dev, struct drm_file *file); > int i915_gem_context_enable(struct drm_i915_private *dev_priv); > void i915_gem_context_close(struct drm_device *dev, struct drm_file *file); > -int i915_switch_context(struct intel_ring_buffer *ring, > +int i915_switch_context(struct intel_engine *ring, > struct i915_hw_context *to); > struct i915_hw_context * > i915_gem_context_get(struct drm_i915_file_private *file_priv, u32 id); > @@ -2497,9 +2497,9 @@ const char *i915_cache_level_str(int type); > > /* i915_cmd_parser.c */ > int i915_cmd_parser_get_version(void); > -void i915_cmd_parser_init_ring(struct intel_ring_buffer *ring); > -bool i915_needs_cmd_parser(struct intel_ring_buffer *ring); > -int i915_parse_cmds(struct intel_ring_buffer *ring, > +void i915_cmd_parser_init_ring(struct intel_engine *ring); > +bool i915_needs_cmd_parser(struct intel_engine *ring); > +int i915_parse_cmds(struct intel_engine *ring, > struct drm_i915_gem_object *batch_obj, > u32 batch_start_offset, > bool is_master); > diff --git a/drivers/gpu/drm/i915/i915_gem.c b/drivers/gpu/drm/i915/i915_gem.c > index 6ef53bd..a3b697b 100644 > --- a/drivers/gpu/drm/i915/i915_gem.c > +++ b/drivers/gpu/drm/i915/i915_gem.c > @@ -64,7 +64,7 @@ static unsigned long i915_gem_inactive_scan(struct shrinker *shrinker, > static unsigned long i915_gem_purge(struct drm_i915_private *dev_priv, long target); > static unsigned long i915_gem_shrink_all(struct drm_i915_private *dev_priv); > static void i915_gem_object_truncate(struct drm_i915_gem_object *obj); > -static void i915_gem_retire_requests_ring(struct intel_ring_buffer *ring); > +static void i915_gem_retire_requests_ring(struct intel_engine *ring); > > static bool cpu_cache_is_coherent(struct drm_device *dev, > enum i915_cache_level level) > @@ -977,7 +977,7 @@ i915_gem_check_wedge(struct i915_gpu_error *error, > * equal. > */ > static int > -i915_gem_check_olr(struct intel_ring_buffer *ring, u32 seqno) > +i915_gem_check_olr(struct intel_engine *ring, u32 seqno) > { > int ret; > > @@ -996,7 +996,7 @@ static void fake_irq(unsigned long data) > } > > static bool missed_irq(struct drm_i915_private *dev_priv, > - struct intel_ring_buffer *ring) > + struct intel_engine *ring) > { > return test_bit(ring->id, &dev_priv->gpu_error.missed_irq_rings); > } > @@ -1027,7 +1027,7 @@ static bool can_wait_boost(struct drm_i915_file_private *file_priv) > * Returns 0 if the seqno was found within the alloted time. Else returns the > * errno with remaining time filled in timeout argument. > */ > -static int __wait_seqno(struct intel_ring_buffer *ring, u32 seqno, > +static int __wait_seqno(struct intel_engine *ring, u32 seqno, > unsigned reset_counter, > bool interruptible, > struct timespec *timeout, > @@ -1134,7 +1134,7 @@ static int __wait_seqno(struct intel_ring_buffer *ring, u32 seqno, > * request and object lists appropriately for that event. > */ > int > -i915_wait_seqno(struct intel_ring_buffer *ring, uint32_t seqno) > +i915_wait_seqno(struct intel_engine *ring, uint32_t seqno) > { > struct drm_device *dev = ring->dev; > struct drm_i915_private *dev_priv = dev->dev_private; > @@ -1159,7 +1159,7 @@ i915_wait_seqno(struct intel_ring_buffer *ring, uint32_t seqno) > > static int > i915_gem_object_wait_rendering__tail(struct drm_i915_gem_object *obj, > - struct intel_ring_buffer *ring) > + struct intel_engine *ring) > { > if (!obj->active) > return 0; > @@ -1184,7 +1184,7 @@ static __must_check int > i915_gem_object_wait_rendering(struct drm_i915_gem_object *obj, > bool readonly) > { > - struct intel_ring_buffer *ring = obj->ring; > + struct intel_engine *ring = obj->ring; > u32 seqno; > int ret; > > @@ -1209,7 +1209,7 @@ i915_gem_object_wait_rendering__nonblocking(struct drm_i915_gem_object *obj, > { > struct drm_device *dev = obj->base.dev; > struct drm_i915_private *dev_priv = dev->dev_private; > - struct intel_ring_buffer *ring = obj->ring; > + struct intel_engine *ring = obj->ring; > unsigned reset_counter; > u32 seqno; > int ret; > @@ -2011,7 +2011,7 @@ i915_gem_object_get_pages(struct drm_i915_gem_object *obj) > > static void > i915_gem_object_move_to_active(struct drm_i915_gem_object *obj, > - struct intel_ring_buffer *ring) > + struct intel_engine *ring) > { > struct drm_device *dev = obj->base.dev; > struct drm_i915_private *dev_priv = dev->dev_private; > @@ -2049,7 +2049,7 @@ i915_gem_object_move_to_active(struct drm_i915_gem_object *obj, > } > > void i915_vma_move_to_active(struct i915_vma *vma, > - struct intel_ring_buffer *ring) > + struct intel_engine *ring) > { > list_move_tail(&vma->mm_list, &vma->vm->active_list); > return i915_gem_object_move_to_active(vma->obj, ring); > @@ -2090,7 +2090,7 @@ i915_gem_object_move_to_inactive(struct drm_i915_gem_object *obj) > static void > i915_gem_object_retire(struct drm_i915_gem_object *obj) > { > - struct intel_ring_buffer *ring = obj->ring; > + struct intel_engine *ring = obj->ring; > > if (ring == NULL) > return; > @@ -2104,7 +2104,7 @@ static int > i915_gem_init_seqno(struct drm_device *dev, u32 seqno) > { > struct drm_i915_private *dev_priv = dev->dev_private; > - struct intel_ring_buffer *ring; > + struct intel_engine *ring; > int ret, i, j; > > /* Carefully retire all requests without writing to the rings */ > @@ -2170,7 +2170,7 @@ i915_gem_get_seqno(struct drm_device *dev, u32 *seqno) > return 0; > } > > -int __i915_add_request(struct intel_ring_buffer *ring, > +int __i915_add_request(struct intel_engine *ring, > struct drm_file *file, > struct drm_i915_gem_object *obj, > u32 *out_seqno) > @@ -2330,7 +2330,7 @@ static void i915_gem_free_request(struct drm_i915_gem_request *request) > } > > struct drm_i915_gem_request * > -i915_gem_find_active_request(struct intel_ring_buffer *ring) > +i915_gem_find_active_request(struct intel_engine *ring) > { > struct drm_i915_gem_request *request; > u32 completed_seqno; > @@ -2348,7 +2348,7 @@ i915_gem_find_active_request(struct intel_ring_buffer *ring) > } > > static void i915_gem_reset_ring_status(struct drm_i915_private *dev_priv, > - struct intel_ring_buffer *ring) > + struct intel_engine *ring) > { > struct drm_i915_gem_request *request; > bool ring_hung; > @@ -2367,7 +2367,7 @@ static void i915_gem_reset_ring_status(struct drm_i915_private *dev_priv, > } > > static void i915_gem_reset_ring_cleanup(struct drm_i915_private *dev_priv, > - struct intel_ring_buffer *ring) > + struct intel_engine *ring) > { > while (!list_empty(&ring->active_list)) { > struct drm_i915_gem_object *obj; > @@ -2426,7 +2426,7 @@ void i915_gem_restore_fences(struct drm_device *dev) > void i915_gem_reset(struct drm_device *dev) > { > struct drm_i915_private *dev_priv = dev->dev_private; > - struct intel_ring_buffer *ring; > + struct intel_engine *ring; > int i; > > /* > @@ -2449,7 +2449,7 @@ void i915_gem_reset(struct drm_device *dev) > * This function clears the request list as sequence numbers are passed. > */ > static void > -i915_gem_retire_requests_ring(struct intel_ring_buffer *ring) > +i915_gem_retire_requests_ring(struct intel_engine *ring) > { > uint32_t seqno; > > @@ -2512,7 +2512,7 @@ bool > i915_gem_retire_requests(struct drm_device *dev) > { > struct drm_i915_private *dev_priv = dev->dev_private; > - struct intel_ring_buffer *ring; > + struct intel_engine *ring; > bool idle = true; > int i; > > @@ -2606,7 +2606,7 @@ i915_gem_wait_ioctl(struct drm_device *dev, void *data, struct drm_file *file) > struct drm_i915_private *dev_priv = dev->dev_private; > struct drm_i915_gem_wait *args = data; > struct drm_i915_gem_object *obj; > - struct intel_ring_buffer *ring = NULL; > + struct intel_engine *ring = NULL; > struct timespec timeout_stack, *timeout = NULL; > unsigned reset_counter; > u32 seqno = 0; > @@ -2677,9 +2677,9 @@ out: > */ > int > i915_gem_object_sync(struct drm_i915_gem_object *obj, > - struct intel_ring_buffer *to) > + struct intel_engine *to) > { > - struct intel_ring_buffer *from = obj->ring; > + struct intel_engine *from = obj->ring; > u32 seqno; > int ret, idx; > > @@ -2800,7 +2800,7 @@ int i915_vma_unbind(struct i915_vma *vma) > int i915_gpu_idle(struct drm_device *dev) > { > struct drm_i915_private *dev_priv = dev->dev_private; > - struct intel_ring_buffer *ring; > + struct intel_engine *ring; > int ret, i; > > /* Flush everything onto the inactive list. */ > @@ -3659,7 +3659,7 @@ static bool is_pin_display(struct drm_i915_gem_object *obj) > int > i915_gem_object_pin_to_display_plane(struct drm_i915_gem_object *obj, > u32 alignment, > - struct intel_ring_buffer *pipelined) > + struct intel_engine *pipelined) > { > u32 old_read_domains, old_write_domain; > int ret; > @@ -3812,7 +3812,7 @@ i915_gem_ring_throttle(struct drm_device *dev, struct drm_file *file) > struct drm_i915_file_private *file_priv = file->driver_priv; > unsigned long recent_enough = jiffies - msecs_to_jiffies(20); > struct drm_i915_gem_request *request; > - struct intel_ring_buffer *ring = NULL; > + struct intel_engine *ring = NULL; > unsigned reset_counter; > u32 seqno = 0; > int ret; > @@ -4258,7 +4258,7 @@ static void > i915_gem_stop_ringbuffers(struct drm_device *dev) > { > struct drm_i915_private *dev_priv = dev->dev_private; > - struct intel_ring_buffer *ring; > + struct intel_engine *ring; > int i; > > for_each_active_ring(ring, dev_priv, i) > @@ -4307,7 +4307,7 @@ err: > return ret; > } > > -int i915_gem_l3_remap(struct intel_ring_buffer *ring, int slice) > +int i915_gem_l3_remap(struct intel_engine *ring, int slice) > { > struct drm_device *dev = ring->dev; > struct drm_i915_private *dev_priv = dev->dev_private; > @@ -4532,7 +4532,7 @@ void > i915_gem_cleanup_ringbuffer(struct drm_device *dev) > { > struct drm_i915_private *dev_priv = dev->dev_private; > - struct intel_ring_buffer *ring; > + struct intel_engine *ring; > int i; > > for_each_active_ring(ring, dev_priv, i) > @@ -4608,7 +4608,7 @@ i915_gem_lastclose(struct drm_device *dev) > } > > static void > -init_ring_lists(struct intel_ring_buffer *ring) > +init_ring_lists(struct intel_engine *ring) > { > INIT_LIST_HEAD(&ring->active_list); > INIT_LIST_HEAD(&ring->request_list); > diff --git a/drivers/gpu/drm/i915/i915_gem_context.c b/drivers/gpu/drm/i915/i915_gem_context.c > index 014fb8f..4d37e20 100644 > --- a/drivers/gpu/drm/i915/i915_gem_context.c > +++ b/drivers/gpu/drm/i915/i915_gem_context.c > @@ -359,7 +359,7 @@ err_destroy: > void i915_gem_context_reset(struct drm_device *dev) > { > struct drm_i915_private *dev_priv = dev->dev_private; > - struct intel_ring_buffer *ring; > + struct intel_engine *ring; > int i; > > /* Prevent the hardware from restoring the last context (which hung) on > @@ -392,7 +392,7 @@ int i915_gem_context_init(struct drm_device *dev) > { > struct drm_i915_private *dev_priv = dev->dev_private; > struct i915_hw_context *ctx; > - struct intel_ring_buffer *ring; > + struct intel_engine *ring; > int unused; > > /* Init should only be called once per module load. Eventually the > @@ -428,7 +428,7 @@ void i915_gem_context_fini(struct drm_device *dev) > { > struct drm_i915_private *dev_priv = dev->dev_private; > struct i915_hw_context *dctx = dev_priv->ring[RCS].default_context; > - struct intel_ring_buffer *ring; > + struct intel_engine *ring; > int unused; > > if (dctx->obj) { > @@ -467,7 +467,7 @@ void i915_gem_context_fini(struct drm_device *dev) > > int i915_gem_context_enable(struct drm_i915_private *dev_priv) > { > - struct intel_ring_buffer *ring; > + struct intel_engine *ring; > int ret, i; > > /* This is the only place the aliasing PPGTT gets enabled, which means > @@ -546,7 +546,7 @@ i915_gem_context_get(struct drm_i915_file_private *file_priv, u32 id) > } > > static inline int > -mi_set_context(struct intel_ring_buffer *ring, > +mi_set_context(struct intel_engine *ring, > struct i915_hw_context *new_context, > u32 hw_flags) > { > @@ -596,7 +596,7 @@ mi_set_context(struct intel_ring_buffer *ring, > return ret; > } > > -static int do_switch(struct intel_ring_buffer *ring, > +static int do_switch(struct intel_engine *ring, > struct i915_hw_context *to) > { > struct drm_i915_private *dev_priv = ring->dev->dev_private; > @@ -726,7 +726,7 @@ unpin_out: > * it will have a refoucnt > 1. This allows us to destroy the context abstract > * object while letting the normal object tracking destroy the backing BO. > */ > -int i915_switch_context(struct intel_ring_buffer *ring, > +int i915_switch_context(struct intel_engine *ring, > struct i915_hw_context *to) > { > struct drm_i915_private *dev_priv = ring->dev->dev_private; > diff --git a/drivers/gpu/drm/i915/i915_gem_execbuffer.c b/drivers/gpu/drm/i915/i915_gem_execbuffer.c > index 47fe8ec..95e797e 100644 > --- a/drivers/gpu/drm/i915/i915_gem_execbuffer.c > +++ b/drivers/gpu/drm/i915/i915_gem_execbuffer.c > @@ -541,7 +541,7 @@ need_reloc_mappable(struct i915_vma *vma) > > static int > i915_gem_execbuffer_reserve_vma(struct i915_vma *vma, > - struct intel_ring_buffer *ring, > + struct intel_engine *ring, > bool *need_reloc) > { > struct drm_i915_gem_object *obj = vma->obj; > @@ -596,7 +596,7 @@ i915_gem_execbuffer_reserve_vma(struct i915_vma *vma, > } > > static int > -i915_gem_execbuffer_reserve(struct intel_ring_buffer *ring, > +i915_gem_execbuffer_reserve(struct intel_engine *ring, > struct list_head *vmas, > bool *need_relocs) > { > @@ -711,7 +711,7 @@ static int > i915_gem_execbuffer_relocate_slow(struct drm_device *dev, > struct drm_i915_gem_execbuffer2 *args, > struct drm_file *file, > - struct intel_ring_buffer *ring, > + struct intel_engine *ring, > struct eb_vmas *eb, > struct drm_i915_gem_exec_object2 *exec) > { > @@ -827,7 +827,7 @@ err: > } > > static int > -i915_gem_execbuffer_move_to_gpu(struct intel_ring_buffer *ring, > +i915_gem_execbuffer_move_to_gpu(struct intel_engine *ring, > struct list_head *vmas) > { > struct i915_vma *vma; > @@ -912,7 +912,7 @@ validate_exec_list(struct drm_i915_gem_exec_object2 *exec, > > static struct i915_hw_context * > i915_gem_validate_context(struct drm_device *dev, struct drm_file *file, > - struct intel_ring_buffer *ring, const u32 ctx_id) > + struct intel_engine *ring, const u32 ctx_id) > { > struct i915_hw_context *ctx = NULL; > struct i915_ctx_hang_stats *hs; > @@ -935,7 +935,7 @@ i915_gem_validate_context(struct drm_device *dev, struct drm_file *file, > > static void > i915_gem_execbuffer_move_to_active(struct list_head *vmas, > - struct intel_ring_buffer *ring) > + struct intel_engine *ring) > { > struct i915_vma *vma; > > @@ -970,7 +970,7 @@ i915_gem_execbuffer_move_to_active(struct list_head *vmas, > static void > i915_gem_execbuffer_retire_commands(struct drm_device *dev, > struct drm_file *file, > - struct intel_ring_buffer *ring, > + struct intel_engine *ring, > struct drm_i915_gem_object *obj) > { > /* Unconditionally force add_request to emit a full flush. */ > @@ -982,7 +982,7 @@ i915_gem_execbuffer_retire_commands(struct drm_device *dev, > > static int > i915_reset_gen7_sol_offsets(struct drm_device *dev, > - struct intel_ring_buffer *ring) > + struct intel_engine *ring) > { > struct drm_i915_private *dev_priv = dev->dev_private; > int ret, i; > @@ -1048,7 +1048,7 @@ i915_gem_do_execbuffer(struct drm_device *dev, void *data, > struct eb_vmas *eb; > struct drm_i915_gem_object *batch_obj; > struct drm_clip_rect *cliprects = NULL; > - struct intel_ring_buffer *ring; > + struct intel_engine *ring; > struct i915_hw_context *ctx; > struct i915_address_space *vm; > const u32 ctx_id = i915_execbuffer2_get_context_id(*args); > diff --git a/drivers/gpu/drm/i915/i915_gem_gtt.c b/drivers/gpu/drm/i915/i915_gem_gtt.c > index 1dff805..31b58ee 100644 > --- a/drivers/gpu/drm/i915/i915_gem_gtt.c > +++ b/drivers/gpu/drm/i915/i915_gem_gtt.c > @@ -207,7 +207,7 @@ static gen6_gtt_pte_t iris_pte_encode(dma_addr_t addr, > } > > /* Broadwell Page Directory Pointer Descriptors */ > -static int gen8_write_pdp(struct intel_ring_buffer *ring, unsigned entry, > +static int gen8_write_pdp(struct intel_engine *ring, unsigned entry, > uint64_t val, bool synchronous) > { > struct drm_i915_private *dev_priv = ring->dev->dev_private; > @@ -237,7 +237,7 @@ static int gen8_write_pdp(struct intel_ring_buffer *ring, unsigned entry, > } > > static int gen8_mm_switch(struct i915_hw_ppgtt *ppgtt, > - struct intel_ring_buffer *ring, > + struct intel_engine *ring, > bool synchronous) > { > int i, ret; > @@ -716,7 +716,7 @@ static uint32_t get_pd_offset(struct i915_hw_ppgtt *ppgtt) > } > > static int hsw_mm_switch(struct i915_hw_ppgtt *ppgtt, > - struct intel_ring_buffer *ring, > + struct intel_engine *ring, > bool synchronous) > { > struct drm_device *dev = ppgtt->base.dev; > @@ -760,7 +760,7 @@ static int hsw_mm_switch(struct i915_hw_ppgtt *ppgtt, > } > > static int gen7_mm_switch(struct i915_hw_ppgtt *ppgtt, > - struct intel_ring_buffer *ring, > + struct intel_engine *ring, > bool synchronous) > { > struct drm_device *dev = ppgtt->base.dev; > @@ -811,7 +811,7 @@ static int gen7_mm_switch(struct i915_hw_ppgtt *ppgtt, > } > > static int gen6_mm_switch(struct i915_hw_ppgtt *ppgtt, > - struct intel_ring_buffer *ring, > + struct intel_engine *ring, > bool synchronous) > { > struct drm_device *dev = ppgtt->base.dev; > @@ -832,7 +832,7 @@ static int gen8_ppgtt_enable(struct i915_hw_ppgtt *ppgtt) > { > struct drm_device *dev = ppgtt->base.dev; > struct drm_i915_private *dev_priv = dev->dev_private; > - struct intel_ring_buffer *ring; > + struct intel_engine *ring; > int j, ret; > > for_each_active_ring(ring, dev_priv, j) { > @@ -862,7 +862,7 @@ static int gen7_ppgtt_enable(struct i915_hw_ppgtt *ppgtt) > { > struct drm_device *dev = ppgtt->base.dev; > struct drm_i915_private *dev_priv = dev->dev_private; > - struct intel_ring_buffer *ring; > + struct intel_engine *ring; > uint32_t ecochk, ecobits; > int i; > > @@ -901,7 +901,7 @@ static int gen6_ppgtt_enable(struct i915_hw_ppgtt *ppgtt) > { > struct drm_device *dev = ppgtt->base.dev; > struct drm_i915_private *dev_priv = dev->dev_private; > - struct intel_ring_buffer *ring; > + struct intel_engine *ring; > uint32_t ecochk, gab_ctl, ecobits; > int i; > > @@ -1269,7 +1269,7 @@ static void undo_idling(struct drm_i915_private *dev_priv, bool interruptible) > void i915_check_and_clear_faults(struct drm_device *dev) > { > struct drm_i915_private *dev_priv = dev->dev_private; > - struct intel_ring_buffer *ring; > + struct intel_engine *ring; > int i; > > if (INTEL_INFO(dev)->gen < 6) > diff --git a/drivers/gpu/drm/i915/i915_gem_gtt.h b/drivers/gpu/drm/i915/i915_gem_gtt.h > index cfca023..0775662 100644 > --- a/drivers/gpu/drm/i915/i915_gem_gtt.h > +++ b/drivers/gpu/drm/i915/i915_gem_gtt.h > @@ -261,7 +261,7 @@ struct i915_hw_ppgtt { > > int (*enable)(struct i915_hw_ppgtt *ppgtt); > int (*switch_mm)(struct i915_hw_ppgtt *ppgtt, > - struct intel_ring_buffer *ring, > + struct intel_engine *ring, > bool synchronous); > void (*debug_dump)(struct i915_hw_ppgtt *ppgtt, struct seq_file *m); > }; > diff --git a/drivers/gpu/drm/i915/i915_gpu_error.c b/drivers/gpu/drm/i915/i915_gpu_error.c > index 8f37238..0853db3 100644 > --- a/drivers/gpu/drm/i915/i915_gpu_error.c > +++ b/drivers/gpu/drm/i915/i915_gpu_error.c > @@ -745,7 +745,7 @@ static void i915_gem_record_fences(struct drm_device *dev, > } > > static void i915_record_ring_state(struct drm_device *dev, > - struct intel_ring_buffer *ring, > + struct intel_engine *ring, > struct drm_i915_error_ring *ering) > { > struct drm_i915_private *dev_priv = dev->dev_private; > @@ -857,7 +857,7 @@ static void i915_record_ring_state(struct drm_device *dev, > } > > > -static void i915_gem_record_active_context(struct intel_ring_buffer *ring, > +static void i915_gem_record_active_context(struct intel_engine *ring, > struct drm_i915_error_state *error, > struct drm_i915_error_ring *ering) > { > @@ -884,7 +884,7 @@ static void i915_gem_record_rings(struct drm_device *dev, > int i, count; > > for (i = 0; i < I915_NUM_RINGS; i++) { > - struct intel_ring_buffer *ring = &dev_priv->ring[i]; > + struct intel_engine *ring = &dev_priv->ring[i]; > > if (!intel_ring_initialized(ring)) > continue; > diff --git a/drivers/gpu/drm/i915/i915_irq.c b/drivers/gpu/drm/i915/i915_irq.c > index 4a8e8cb..58c8812 100644 > --- a/drivers/gpu/drm/i915/i915_irq.c > +++ b/drivers/gpu/drm/i915/i915_irq.c > @@ -1077,7 +1077,7 @@ static void ironlake_rps_change_irq_handler(struct drm_device *dev) > } > > static void notify_ring(struct drm_device *dev, > - struct intel_ring_buffer *ring) > + struct intel_engine *ring) > { > if (ring->obj == NULL) > return; > @@ -2111,7 +2111,7 @@ static irqreturn_t gen8_irq_handler(int irq, void *arg) > static void i915_error_wake_up(struct drm_i915_private *dev_priv, > bool reset_completed) > { > - struct intel_ring_buffer *ring; > + struct intel_engine *ring; > int i; > > /* > @@ -2544,14 +2544,14 @@ static void gen8_disable_vblank(struct drm_device *dev, int pipe) > } > > static u32 > -ring_last_seqno(struct intel_ring_buffer *ring) > +ring_last_seqno(struct intel_engine *ring) > { > return list_entry(ring->request_list.prev, > struct drm_i915_gem_request, list)->seqno; > } > > static bool > -ring_idle(struct intel_ring_buffer *ring, u32 seqno) > +ring_idle(struct intel_engine *ring, u32 seqno) > { > return (list_empty(&ring->request_list) || > i915_seqno_passed(seqno, ring_last_seqno(ring))); > @@ -2574,11 +2574,11 @@ ipehr_is_semaphore_wait(struct drm_device *dev, u32 ipehr) > } > } > > -static struct intel_ring_buffer * > -semaphore_wait_to_signaller_ring(struct intel_ring_buffer *ring, u32 ipehr) > +static struct intel_engine * > +semaphore_wait_to_signaller_ring(struct intel_engine *ring, u32 ipehr) > { > struct drm_i915_private *dev_priv = ring->dev->dev_private; > - struct intel_ring_buffer *signaller; > + struct intel_engine *signaller; > int i; > > if (INTEL_INFO(dev_priv->dev)->gen >= 8) { > @@ -2606,8 +2606,8 @@ semaphore_wait_to_signaller_ring(struct intel_ring_buffer *ring, u32 ipehr) > return NULL; > } > > -static struct intel_ring_buffer * > -semaphore_waits_for(struct intel_ring_buffer *ring, u32 *seqno) > +static struct intel_engine * > +semaphore_waits_for(struct intel_engine *ring, u32 *seqno) > { > struct drm_i915_private *dev_priv = ring->dev->dev_private; > u32 cmd, ipehr, head; > @@ -2649,10 +2649,10 @@ semaphore_waits_for(struct intel_ring_buffer *ring, u32 *seqno) > return semaphore_wait_to_signaller_ring(ring, ipehr); > } > > -static int semaphore_passed(struct intel_ring_buffer *ring) > +static int semaphore_passed(struct intel_engine *ring) > { > struct drm_i915_private *dev_priv = ring->dev->dev_private; > - struct intel_ring_buffer *signaller; > + struct intel_engine *signaller; > u32 seqno, ctl; > > ring->hangcheck.deadlock = true; > @@ -2671,7 +2671,7 @@ static int semaphore_passed(struct intel_ring_buffer *ring) > > static void semaphore_clear_deadlocks(struct drm_i915_private *dev_priv) > { > - struct intel_ring_buffer *ring; > + struct intel_engine *ring; > int i; > > for_each_active_ring(ring, dev_priv, i) > @@ -2679,7 +2679,7 @@ static void semaphore_clear_deadlocks(struct drm_i915_private *dev_priv) > } > > static enum intel_ring_hangcheck_action > -ring_stuck(struct intel_ring_buffer *ring, u64 acthd) > +ring_stuck(struct intel_engine *ring, u64 acthd) > { > struct drm_device *dev = ring->dev; > struct drm_i915_private *dev_priv = dev->dev_private; > @@ -2735,7 +2735,7 @@ static void i915_hangcheck_elapsed(unsigned long data) > { > struct drm_device *dev = (struct drm_device *)data; > struct drm_i915_private *dev_priv = dev->dev_private; > - struct intel_ring_buffer *ring; > + struct intel_engine *ring; > int i; > int busy_count = 0, rings_hung = 0; > bool stuck[I915_NUM_RINGS] = { 0 }; > diff --git a/drivers/gpu/drm/i915/i915_trace.h b/drivers/gpu/drm/i915/i915_trace.h > index b29d7b1..a4f9e62 100644 > --- a/drivers/gpu/drm/i915/i915_trace.h > +++ b/drivers/gpu/drm/i915/i915_trace.h > @@ -326,8 +326,8 @@ TRACE_EVENT(i915_gem_evict_vm, > ); > > TRACE_EVENT(i915_gem_ring_sync_to, > - TP_PROTO(struct intel_ring_buffer *from, > - struct intel_ring_buffer *to, > + TP_PROTO(struct intel_engine *from, > + struct intel_engine *to, > u32 seqno), > TP_ARGS(from, to, seqno), > > @@ -352,7 +352,7 @@ TRACE_EVENT(i915_gem_ring_sync_to, > ); > > TRACE_EVENT(i915_gem_ring_dispatch, > - TP_PROTO(struct intel_ring_buffer *ring, u32 seqno, u32 flags), > + TP_PROTO(struct intel_engine *ring, u32 seqno, u32 flags), > TP_ARGS(ring, seqno, flags), > > TP_STRUCT__entry( > @@ -375,7 +375,7 @@ TRACE_EVENT(i915_gem_ring_dispatch, > ); > > TRACE_EVENT(i915_gem_ring_flush, > - TP_PROTO(struct intel_ring_buffer *ring, u32 invalidate, u32 flush), > + TP_PROTO(struct intel_engine *ring, u32 invalidate, u32 flush), > TP_ARGS(ring, invalidate, flush), > > TP_STRUCT__entry( > @@ -398,7 +398,7 @@ TRACE_EVENT(i915_gem_ring_flush, > ); > > DECLARE_EVENT_CLASS(i915_gem_request, > - TP_PROTO(struct intel_ring_buffer *ring, u32 seqno), > + TP_PROTO(struct intel_engine *ring, u32 seqno), > TP_ARGS(ring, seqno), > > TP_STRUCT__entry( > @@ -418,12 +418,12 @@ DECLARE_EVENT_CLASS(i915_gem_request, > ); > > DEFINE_EVENT(i915_gem_request, i915_gem_request_add, > - TP_PROTO(struct intel_ring_buffer *ring, u32 seqno), > + TP_PROTO(struct intel_engine *ring, u32 seqno), > TP_ARGS(ring, seqno) > ); > > TRACE_EVENT(i915_gem_request_complete, > - TP_PROTO(struct intel_ring_buffer *ring), > + TP_PROTO(struct intel_engine *ring), > TP_ARGS(ring), > > TP_STRUCT__entry( > @@ -443,12 +443,12 @@ TRACE_EVENT(i915_gem_request_complete, > ); > > DEFINE_EVENT(i915_gem_request, i915_gem_request_retire, > - TP_PROTO(struct intel_ring_buffer *ring, u32 seqno), > + TP_PROTO(struct intel_engine *ring, u32 seqno), > TP_ARGS(ring, seqno) > ); > > TRACE_EVENT(i915_gem_request_wait_begin, > - TP_PROTO(struct intel_ring_buffer *ring, u32 seqno), > + TP_PROTO(struct intel_engine *ring, u32 seqno), > TP_ARGS(ring, seqno), > > TP_STRUCT__entry( > @@ -477,12 +477,12 @@ TRACE_EVENT(i915_gem_request_wait_begin, > ); > > DEFINE_EVENT(i915_gem_request, i915_gem_request_wait_end, > - TP_PROTO(struct intel_ring_buffer *ring, u32 seqno), > + TP_PROTO(struct intel_engine *ring, u32 seqno), > TP_ARGS(ring, seqno) > ); > > DECLARE_EVENT_CLASS(i915_ring, > - TP_PROTO(struct intel_ring_buffer *ring), > + TP_PROTO(struct intel_engine *ring), > TP_ARGS(ring), > > TP_STRUCT__entry( > @@ -499,12 +499,12 @@ DECLARE_EVENT_CLASS(i915_ring, > ); > > DEFINE_EVENT(i915_ring, i915_ring_wait_begin, > - TP_PROTO(struct intel_ring_buffer *ring), > + TP_PROTO(struct intel_engine *ring), > TP_ARGS(ring) > ); > > DEFINE_EVENT(i915_ring, i915_ring_wait_end, > - TP_PROTO(struct intel_ring_buffer *ring), > + TP_PROTO(struct intel_engine *ring), > TP_ARGS(ring) > ); > > diff --git a/drivers/gpu/drm/i915/intel_display.c b/drivers/gpu/drm/i915/intel_display.c > index c65e7f7..f821147 100644 > --- a/drivers/gpu/drm/i915/intel_display.c > +++ b/drivers/gpu/drm/i915/intel_display.c > @@ -1944,7 +1944,7 @@ static int intel_align_height(struct drm_device *dev, int height, bool tiled) > int > intel_pin_and_fence_fb_obj(struct drm_device *dev, > struct drm_i915_gem_object *obj, > - struct intel_ring_buffer *pipelined) > + struct intel_engine *pipelined) > { > struct drm_i915_private *dev_priv = dev->dev_private; > u32 alignment; > @@ -8424,7 +8424,7 @@ out: > } > > void intel_mark_fb_busy(struct drm_i915_gem_object *obj, > - struct intel_ring_buffer *ring) > + struct intel_engine *ring) > { > struct drm_device *dev = obj->base.dev; > struct drm_crtc *crtc; > @@ -8582,7 +8582,7 @@ static int intel_gen2_queue_flip(struct drm_device *dev, > struct drm_i915_private *dev_priv = dev->dev_private; > struct intel_crtc *intel_crtc = to_intel_crtc(crtc); > u32 flip_mask; > - struct intel_ring_buffer *ring = &dev_priv->ring[RCS]; > + struct intel_engine *ring = &dev_priv->ring[RCS]; > int ret; > > ret = intel_pin_and_fence_fb_obj(dev, obj, ring); > @@ -8627,7 +8627,7 @@ static int intel_gen3_queue_flip(struct drm_device *dev, > struct drm_i915_private *dev_priv = dev->dev_private; > struct intel_crtc *intel_crtc = to_intel_crtc(crtc); > u32 flip_mask; > - struct intel_ring_buffer *ring = &dev_priv->ring[RCS]; > + struct intel_engine *ring = &dev_priv->ring[RCS]; > int ret; > > ret = intel_pin_and_fence_fb_obj(dev, obj, ring); > @@ -8669,7 +8669,7 @@ static int intel_gen4_queue_flip(struct drm_device *dev, > struct drm_i915_private *dev_priv = dev->dev_private; > struct intel_crtc *intel_crtc = to_intel_crtc(crtc); > uint32_t pf, pipesrc; > - struct intel_ring_buffer *ring = &dev_priv->ring[RCS]; > + struct intel_engine *ring = &dev_priv->ring[RCS]; > int ret; > > ret = intel_pin_and_fence_fb_obj(dev, obj, ring); > @@ -8717,7 +8717,7 @@ static int intel_gen6_queue_flip(struct drm_device *dev, > { > struct drm_i915_private *dev_priv = dev->dev_private; > struct intel_crtc *intel_crtc = to_intel_crtc(crtc); > - struct intel_ring_buffer *ring = &dev_priv->ring[RCS]; > + struct intel_engine *ring = &dev_priv->ring[RCS]; > uint32_t pf, pipesrc; > int ret; > > @@ -8762,7 +8762,7 @@ static int intel_gen7_queue_flip(struct drm_device *dev, > { > struct drm_i915_private *dev_priv = dev->dev_private; > struct intel_crtc *intel_crtc = to_intel_crtc(crtc); > - struct intel_ring_buffer *ring; > + struct intel_engine *ring; > uint32_t plane_bit = 0; > int len, ret; > > diff --git a/drivers/gpu/drm/i915/intel_drv.h b/drivers/gpu/drm/i915/intel_drv.h > index d8b540b..23b5abf 100644 > --- a/drivers/gpu/drm/i915/intel_drv.h > +++ b/drivers/gpu/drm/i915/intel_drv.h > @@ -694,7 +694,7 @@ int intel_pch_rawclk(struct drm_device *dev); > int valleyview_cur_cdclk(struct drm_i915_private *dev_priv); > void intel_mark_busy(struct drm_device *dev); > void intel_mark_fb_busy(struct drm_i915_gem_object *obj, > - struct intel_ring_buffer *ring); > + struct intel_engine *ring); > void intel_mark_idle(struct drm_device *dev); > void intel_crtc_restore_mode(struct drm_crtc *crtc); > void intel_crtc_update_dpms(struct drm_crtc *crtc); > @@ -726,7 +726,7 @@ void intel_release_load_detect_pipe(struct drm_connector *connector, > struct intel_load_detect_pipe *old); > int intel_pin_and_fence_fb_obj(struct drm_device *dev, > struct drm_i915_gem_object *obj, > - struct intel_ring_buffer *pipelined); > + struct intel_engine *pipelined); > void intel_unpin_fb_obj(struct drm_i915_gem_object *obj); > struct drm_framebuffer * > __intel_framebuffer_create(struct drm_device *dev, > diff --git a/drivers/gpu/drm/i915/intel_overlay.c b/drivers/gpu/drm/i915/intel_overlay.c > index d8adc91..965eec1 100644 > --- a/drivers/gpu/drm/i915/intel_overlay.c > +++ b/drivers/gpu/drm/i915/intel_overlay.c > @@ -213,7 +213,7 @@ static int intel_overlay_do_wait_request(struct intel_overlay *overlay, > { > struct drm_device *dev = overlay->dev; > struct drm_i915_private *dev_priv = dev->dev_private; > - struct intel_ring_buffer *ring = &dev_priv->ring[RCS]; > + struct intel_engine *ring = &dev_priv->ring[RCS]; > int ret; > > BUG_ON(overlay->last_flip_req); > @@ -236,7 +236,7 @@ static int intel_overlay_on(struct intel_overlay *overlay) > { > struct drm_device *dev = overlay->dev; > struct drm_i915_private *dev_priv = dev->dev_private; > - struct intel_ring_buffer *ring = &dev_priv->ring[RCS]; > + struct intel_engine *ring = &dev_priv->ring[RCS]; > int ret; > > BUG_ON(overlay->active); > @@ -263,7 +263,7 @@ static int intel_overlay_continue(struct intel_overlay *overlay, > { > struct drm_device *dev = overlay->dev; > struct drm_i915_private *dev_priv = dev->dev_private; > - struct intel_ring_buffer *ring = &dev_priv->ring[RCS]; > + struct intel_engine *ring = &dev_priv->ring[RCS]; > u32 flip_addr = overlay->flip_addr; > u32 tmp; > int ret; > @@ -320,7 +320,7 @@ static int intel_overlay_off(struct intel_overlay *overlay) > { > struct drm_device *dev = overlay->dev; > struct drm_i915_private *dev_priv = dev->dev_private; > - struct intel_ring_buffer *ring = &dev_priv->ring[RCS]; > + struct intel_engine *ring = &dev_priv->ring[RCS]; > u32 flip_addr = overlay->flip_addr; > int ret; > > @@ -363,7 +363,7 @@ static int intel_overlay_recover_from_interrupt(struct intel_overlay *overlay) > { > struct drm_device *dev = overlay->dev; > struct drm_i915_private *dev_priv = dev->dev_private; > - struct intel_ring_buffer *ring = &dev_priv->ring[RCS]; > + struct intel_engine *ring = &dev_priv->ring[RCS]; > int ret; > > if (overlay->last_flip_req == 0) > @@ -389,7 +389,7 @@ static int intel_overlay_release_old_vid(struct intel_overlay *overlay) > { > struct drm_device *dev = overlay->dev; > struct drm_i915_private *dev_priv = dev->dev_private; > - struct intel_ring_buffer *ring = &dev_priv->ring[RCS]; > + struct intel_engine *ring = &dev_priv->ring[RCS]; > int ret; > > /* Only wait if there is actually an old frame to release to > diff --git a/drivers/gpu/drm/i915/intel_pm.c b/drivers/gpu/drm/i915/intel_pm.c > index acfded3..17f636e 100644 > --- a/drivers/gpu/drm/i915/intel_pm.c > +++ b/drivers/gpu/drm/i915/intel_pm.c > @@ -3379,7 +3379,7 @@ static void parse_rp_state_cap(struct drm_i915_private *dev_priv, u32 rp_state_c > static void gen8_enable_rps(struct drm_device *dev) > { > struct drm_i915_private *dev_priv = dev->dev_private; > - struct intel_ring_buffer *ring; > + struct intel_engine *ring; > uint32_t rc6_mask = 0, rp_state_cap; > int unused; > > @@ -3454,7 +3454,7 @@ static void gen8_enable_rps(struct drm_device *dev) > static void gen6_enable_rps(struct drm_device *dev) > { > struct drm_i915_private *dev_priv = dev->dev_private; > - struct intel_ring_buffer *ring; > + struct intel_engine *ring; > u32 rp_state_cap; > u32 gt_perf_status; > u32 rc6vids, pcu_mbox = 0, rc6_mask = 0; > @@ -3783,7 +3783,7 @@ static void valleyview_cleanup_gt_powersave(struct drm_device *dev) > static void valleyview_enable_rps(struct drm_device *dev) > { > struct drm_i915_private *dev_priv = dev->dev_private; > - struct intel_ring_buffer *ring; > + struct intel_engine *ring; > u32 gtfifodbg, val, rc6_mode = 0; > int i; > > @@ -3914,7 +3914,7 @@ static int ironlake_setup_rc6(struct drm_device *dev) > static void ironlake_enable_rc6(struct drm_device *dev) > { > struct drm_i915_private *dev_priv = dev->dev_private; > - struct intel_ring_buffer *ring = &dev_priv->ring[RCS]; > + struct intel_engine *ring = &dev_priv->ring[RCS]; > bool was_interruptible; > int ret; > > @@ -4426,7 +4426,7 @@ EXPORT_SYMBOL_GPL(i915_gpu_lower); > bool i915_gpu_busy(void) > { > struct drm_i915_private *dev_priv; > - struct intel_ring_buffer *ring; > + struct intel_engine *ring; > bool ret = false; > int i; > > diff --git a/drivers/gpu/drm/i915/intel_ringbuffer.c b/drivers/gpu/drm/i915/intel_ringbuffer.c > index 5d61923..4c3cc44 100644 > --- a/drivers/gpu/drm/i915/intel_ringbuffer.c > +++ b/drivers/gpu/drm/i915/intel_ringbuffer.c > @@ -40,7 +40,7 @@ > */ > #define CACHELINE_BYTES 64 > > -static inline int ring_space(struct intel_ring_buffer *ring) > +static inline int ring_space(struct intel_engine *ring) > { > int space = (ring->head & HEAD_ADDR) - (ring->tail + I915_RING_FREE_SPACE); > if (space < 0) > @@ -48,13 +48,13 @@ static inline int ring_space(struct intel_ring_buffer *ring) > return space; > } > > -static bool intel_ring_stopped(struct intel_ring_buffer *ring) > +static bool intel_ring_stopped(struct intel_engine *ring) > { > struct drm_i915_private *dev_priv = ring->dev->dev_private; > return dev_priv->gpu_error.stop_rings & intel_ring_flag(ring); > } > > -void __intel_ring_advance(struct intel_ring_buffer *ring) > +void __intel_ring_advance(struct intel_engine *ring) > { > ring->tail &= ring->size - 1; > if (intel_ring_stopped(ring)) > @@ -63,7 +63,7 @@ void __intel_ring_advance(struct intel_ring_buffer *ring) > } > > static int > -gen2_render_ring_flush(struct intel_ring_buffer *ring, > +gen2_render_ring_flush(struct intel_engine *ring, > u32 invalidate_domains, > u32 flush_domains) > { > @@ -89,7 +89,7 @@ gen2_render_ring_flush(struct intel_ring_buffer *ring, > } > > static int > -gen4_render_ring_flush(struct intel_ring_buffer *ring, > +gen4_render_ring_flush(struct intel_engine *ring, > u32 invalidate_domains, > u32 flush_domains) > { > @@ -184,7 +184,7 @@ gen4_render_ring_flush(struct intel_ring_buffer *ring, > * really our business. That leaves only stall at scoreboard. > */ > static int > -intel_emit_post_sync_nonzero_flush(struct intel_ring_buffer *ring) > +intel_emit_post_sync_nonzero_flush(struct intel_engine *ring) > { > u32 scratch_addr = ring->scratch.gtt_offset + 2 * CACHELINE_BYTES; > int ret; > @@ -219,7 +219,7 @@ intel_emit_post_sync_nonzero_flush(struct intel_ring_buffer *ring) > } > > static int > -gen6_render_ring_flush(struct intel_ring_buffer *ring, > +gen6_render_ring_flush(struct intel_engine *ring, > u32 invalidate_domains, u32 flush_domains) > { > u32 flags = 0; > @@ -271,7 +271,7 @@ gen6_render_ring_flush(struct intel_ring_buffer *ring, > } > > static int > -gen7_render_ring_cs_stall_wa(struct intel_ring_buffer *ring) > +gen7_render_ring_cs_stall_wa(struct intel_engine *ring) > { > int ret; > > @@ -289,7 +289,7 @@ gen7_render_ring_cs_stall_wa(struct intel_ring_buffer *ring) > return 0; > } > > -static int gen7_ring_fbc_flush(struct intel_ring_buffer *ring, u32 value) > +static int gen7_ring_fbc_flush(struct intel_engine *ring, u32 value) > { > int ret; > > @@ -313,7 +313,7 @@ static int gen7_ring_fbc_flush(struct intel_ring_buffer *ring, u32 value) > } > > static int > -gen7_render_ring_flush(struct intel_ring_buffer *ring, > +gen7_render_ring_flush(struct intel_engine *ring, > u32 invalidate_domains, u32 flush_domains) > { > u32 flags = 0; > @@ -374,7 +374,7 @@ gen7_render_ring_flush(struct intel_ring_buffer *ring, > } > > static int > -gen8_render_ring_flush(struct intel_ring_buffer *ring, > +gen8_render_ring_flush(struct intel_engine *ring, > u32 invalidate_domains, u32 flush_domains) > { > u32 flags = 0; > @@ -414,14 +414,14 @@ gen8_render_ring_flush(struct intel_ring_buffer *ring, > > } > > -static void ring_write_tail(struct intel_ring_buffer *ring, > +static void ring_write_tail(struct intel_engine *ring, > u32 value) > { > struct drm_i915_private *dev_priv = ring->dev->dev_private; > I915_WRITE_TAIL(ring, value); > } > > -u64 intel_ring_get_active_head(struct intel_ring_buffer *ring) > +u64 intel_ring_get_active_head(struct intel_engine *ring) > { > struct drm_i915_private *dev_priv = ring->dev->dev_private; > u64 acthd; > @@ -437,7 +437,7 @@ u64 intel_ring_get_active_head(struct intel_ring_buffer *ring) > return acthd; > } > > -static void ring_setup_phys_status_page(struct intel_ring_buffer *ring) > +static void ring_setup_phys_status_page(struct intel_engine *ring) > { > struct drm_i915_private *dev_priv = ring->dev->dev_private; > u32 addr; > @@ -448,7 +448,7 @@ static void ring_setup_phys_status_page(struct intel_ring_buffer *ring) > I915_WRITE(HWS_PGA, addr); > } > > -static bool stop_ring(struct intel_ring_buffer *ring) > +static bool stop_ring(struct intel_engine *ring) > { > struct drm_i915_private *dev_priv = to_i915(ring->dev); > > @@ -472,7 +472,7 @@ static bool stop_ring(struct intel_ring_buffer *ring) > return (I915_READ_HEAD(ring) & HEAD_ADDR) == 0; > } > > -static int init_ring_common(struct intel_ring_buffer *ring) > +static int init_ring_common(struct intel_engine *ring) > { > struct drm_device *dev = ring->dev; > struct drm_i915_private *dev_priv = dev->dev_private; > @@ -550,7 +550,7 @@ out: > } > > static int > -init_pipe_control(struct intel_ring_buffer *ring) > +init_pipe_control(struct intel_engine *ring) > { > int ret; > > @@ -591,7 +591,7 @@ err: > return ret; > } > > -static int init_render_ring(struct intel_ring_buffer *ring) > +static int init_render_ring(struct intel_engine *ring) > { > struct drm_device *dev = ring->dev; > struct drm_i915_private *dev_priv = dev->dev_private; > @@ -647,7 +647,7 @@ static int init_render_ring(struct intel_ring_buffer *ring) > return ret; > } > > -static void render_ring_cleanup(struct intel_ring_buffer *ring) > +static void render_ring_cleanup(struct intel_engine *ring) > { > struct drm_device *dev = ring->dev; > > @@ -663,12 +663,12 @@ static void render_ring_cleanup(struct intel_ring_buffer *ring) > ring->scratch.obj = NULL; > } > > -static int gen6_signal(struct intel_ring_buffer *signaller, > +static int gen6_signal(struct intel_engine *signaller, > unsigned int num_dwords) > { > struct drm_device *dev = signaller->dev; > struct drm_i915_private *dev_priv = dev->dev_private; > - struct intel_ring_buffer *useless; > + struct intel_engine *useless; > int i, ret; > > /* NB: In order to be able to do semaphore MBOX updates for varying > @@ -713,7 +713,7 @@ static int gen6_signal(struct intel_ring_buffer *signaller, > * This acts like a signal in the canonical semaphore. > */ > static int > -gen6_add_request(struct intel_ring_buffer *ring) > +gen6_add_request(struct intel_engine *ring) > { > int ret; > > @@ -745,8 +745,8 @@ static inline bool i915_gem_has_seqno_wrapped(struct drm_device *dev, > * @seqno - seqno which the waiter will block on > */ > static int > -gen6_ring_sync(struct intel_ring_buffer *waiter, > - struct intel_ring_buffer *signaller, > +gen6_ring_sync(struct intel_engine *waiter, > + struct intel_engine *signaller, > u32 seqno) > { > u32 dw1 = MI_SEMAPHORE_MBOX | > @@ -794,7 +794,7 @@ do { \ > } while (0) > > static int > -pc_render_add_request(struct intel_ring_buffer *ring) > +pc_render_add_request(struct intel_engine *ring) > { > u32 scratch_addr = ring->scratch.gtt_offset + 2 * CACHELINE_BYTES; > int ret; > @@ -842,7 +842,7 @@ pc_render_add_request(struct intel_ring_buffer *ring) > } > > static u32 > -gen6_ring_get_seqno(struct intel_ring_buffer *ring, bool lazy_coherency) > +gen6_ring_get_seqno(struct intel_engine *ring, bool lazy_coherency) > { > /* Workaround to force correct ordering between irq and seqno writes on > * ivb (and maybe also on snb) by reading from a CS register (like > @@ -856,31 +856,31 @@ gen6_ring_get_seqno(struct intel_ring_buffer *ring, bool lazy_coherency) > } > > static u32 > -ring_get_seqno(struct intel_ring_buffer *ring, bool lazy_coherency) > +ring_get_seqno(struct intel_engine *ring, bool lazy_coherency) > { > return intel_read_status_page(ring, I915_GEM_HWS_INDEX); > } > > static void > -ring_set_seqno(struct intel_ring_buffer *ring, u32 seqno) > +ring_set_seqno(struct intel_engine *ring, u32 seqno) > { > intel_write_status_page(ring, I915_GEM_HWS_INDEX, seqno); > } > > static u32 > -pc_render_get_seqno(struct intel_ring_buffer *ring, bool lazy_coherency) > +pc_render_get_seqno(struct intel_engine *ring, bool lazy_coherency) > { > return ring->scratch.cpu_page[0]; > } > > static void > -pc_render_set_seqno(struct intel_ring_buffer *ring, u32 seqno) > +pc_render_set_seqno(struct intel_engine *ring, u32 seqno) > { > ring->scratch.cpu_page[0] = seqno; > } > > static bool > -gen5_ring_get_irq(struct intel_ring_buffer *ring) > +gen5_ring_get_irq(struct intel_engine *ring) > { > struct drm_device *dev = ring->dev; > struct drm_i915_private *dev_priv = dev->dev_private; > @@ -898,7 +898,7 @@ gen5_ring_get_irq(struct intel_ring_buffer *ring) > } > > static void > -gen5_ring_put_irq(struct intel_ring_buffer *ring) > +gen5_ring_put_irq(struct intel_engine *ring) > { > struct drm_device *dev = ring->dev; > struct drm_i915_private *dev_priv = dev->dev_private; > @@ -911,7 +911,7 @@ gen5_ring_put_irq(struct intel_ring_buffer *ring) > } > > static bool > -i9xx_ring_get_irq(struct intel_ring_buffer *ring) > +i9xx_ring_get_irq(struct intel_engine *ring) > { > struct drm_device *dev = ring->dev; > struct drm_i915_private *dev_priv = dev->dev_private; > @@ -932,7 +932,7 @@ i9xx_ring_get_irq(struct intel_ring_buffer *ring) > } > > static void > -i9xx_ring_put_irq(struct intel_ring_buffer *ring) > +i9xx_ring_put_irq(struct intel_engine *ring) > { > struct drm_device *dev = ring->dev; > struct drm_i915_private *dev_priv = dev->dev_private; > @@ -948,7 +948,7 @@ i9xx_ring_put_irq(struct intel_ring_buffer *ring) > } > > static bool > -i8xx_ring_get_irq(struct intel_ring_buffer *ring) > +i8xx_ring_get_irq(struct intel_engine *ring) > { > struct drm_device *dev = ring->dev; > struct drm_i915_private *dev_priv = dev->dev_private; > @@ -969,7 +969,7 @@ i8xx_ring_get_irq(struct intel_ring_buffer *ring) > } > > static void > -i8xx_ring_put_irq(struct intel_ring_buffer *ring) > +i8xx_ring_put_irq(struct intel_engine *ring) > { > struct drm_device *dev = ring->dev; > struct drm_i915_private *dev_priv = dev->dev_private; > @@ -984,7 +984,7 @@ i8xx_ring_put_irq(struct intel_ring_buffer *ring) > spin_unlock_irqrestore(&dev_priv->irq_lock, flags); > } > > -void intel_ring_setup_status_page(struct intel_ring_buffer *ring) > +void intel_ring_setup_status_page(struct intel_engine *ring) > { > struct drm_device *dev = ring->dev; > struct drm_i915_private *dev_priv = ring->dev->dev_private; > @@ -1047,7 +1047,7 @@ void intel_ring_setup_status_page(struct intel_ring_buffer *ring) > } > > static int > -bsd_ring_flush(struct intel_ring_buffer *ring, > +bsd_ring_flush(struct intel_engine *ring, > u32 invalidate_domains, > u32 flush_domains) > { > @@ -1064,7 +1064,7 @@ bsd_ring_flush(struct intel_ring_buffer *ring, > } > > static int > -i9xx_add_request(struct intel_ring_buffer *ring) > +i9xx_add_request(struct intel_engine *ring) > { > int ret; > > @@ -1082,7 +1082,7 @@ i9xx_add_request(struct intel_ring_buffer *ring) > } > > static bool > -gen6_ring_get_irq(struct intel_ring_buffer *ring) > +gen6_ring_get_irq(struct intel_engine *ring) > { > struct drm_device *dev = ring->dev; > struct drm_i915_private *dev_priv = dev->dev_private; > @@ -1107,7 +1107,7 @@ gen6_ring_get_irq(struct intel_ring_buffer *ring) > } > > static void > -gen6_ring_put_irq(struct intel_ring_buffer *ring) > +gen6_ring_put_irq(struct intel_engine *ring) > { > struct drm_device *dev = ring->dev; > struct drm_i915_private *dev_priv = dev->dev_private; > @@ -1125,7 +1125,7 @@ gen6_ring_put_irq(struct intel_ring_buffer *ring) > } > > static bool > -hsw_vebox_get_irq(struct intel_ring_buffer *ring) > +hsw_vebox_get_irq(struct intel_engine *ring) > { > struct drm_device *dev = ring->dev; > struct drm_i915_private *dev_priv = dev->dev_private; > @@ -1145,7 +1145,7 @@ hsw_vebox_get_irq(struct intel_ring_buffer *ring) > } > > static void > -hsw_vebox_put_irq(struct intel_ring_buffer *ring) > +hsw_vebox_put_irq(struct intel_engine *ring) > { > struct drm_device *dev = ring->dev; > struct drm_i915_private *dev_priv = dev->dev_private; > @@ -1163,7 +1163,7 @@ hsw_vebox_put_irq(struct intel_ring_buffer *ring) > } > > static bool > -gen8_ring_get_irq(struct intel_ring_buffer *ring) > +gen8_ring_get_irq(struct intel_engine *ring) > { > struct drm_device *dev = ring->dev; > struct drm_i915_private *dev_priv = dev->dev_private; > @@ -1189,7 +1189,7 @@ gen8_ring_get_irq(struct intel_ring_buffer *ring) > } > > static void > -gen8_ring_put_irq(struct intel_ring_buffer *ring) > +gen8_ring_put_irq(struct intel_engine *ring) > { > struct drm_device *dev = ring->dev; > struct drm_i915_private *dev_priv = dev->dev_private; > @@ -1209,7 +1209,7 @@ gen8_ring_put_irq(struct intel_ring_buffer *ring) > } > > static int > -i965_dispatch_execbuffer(struct intel_ring_buffer *ring, > +i965_dispatch_execbuffer(struct intel_engine *ring, > u64 offset, u32 length, > unsigned flags) > { > @@ -1232,7 +1232,7 @@ i965_dispatch_execbuffer(struct intel_ring_buffer *ring, > /* Just userspace ABI convention to limit the wa batch bo to a resonable size */ > #define I830_BATCH_LIMIT (256*1024) > static int > -i830_dispatch_execbuffer(struct intel_ring_buffer *ring, > +i830_dispatch_execbuffer(struct intel_engine *ring, > u64 offset, u32 len, > unsigned flags) > { > @@ -1283,7 +1283,7 @@ i830_dispatch_execbuffer(struct intel_ring_buffer *ring, > } > > static int > -i915_dispatch_execbuffer(struct intel_ring_buffer *ring, > +i915_dispatch_execbuffer(struct intel_engine *ring, > u64 offset, u32 len, > unsigned flags) > { > @@ -1300,7 +1300,7 @@ i915_dispatch_execbuffer(struct intel_ring_buffer *ring, > return 0; > } > > -static void cleanup_status_page(struct intel_ring_buffer *ring) > +static void cleanup_status_page(struct intel_engine *ring) > { > struct drm_i915_gem_object *obj; > > @@ -1314,7 +1314,7 @@ static void cleanup_status_page(struct intel_ring_buffer *ring) > ring->status_page.obj = NULL; > } > > -static int init_status_page(struct intel_ring_buffer *ring) > +static int init_status_page(struct intel_engine *ring) > { > struct drm_i915_gem_object *obj; > > @@ -1351,7 +1351,7 @@ err_unref: > return 0; > } > > -static int init_phys_status_page(struct intel_ring_buffer *ring) > +static int init_phys_status_page(struct intel_engine *ring) > { > struct drm_i915_private *dev_priv = ring->dev->dev_private; > > @@ -1368,7 +1368,7 @@ static int init_phys_status_page(struct intel_ring_buffer *ring) > return 0; > } > > -void intel_destroy_ring_buffer(struct intel_ring_buffer *ring) > +void intel_destroy_ring_buffer(struct intel_engine *ring) > { > if (!ring->obj) > return; > @@ -1379,7 +1379,7 @@ void intel_destroy_ring_buffer(struct intel_ring_buffer *ring) > ring->obj = NULL; > } > > -int intel_allocate_ring_buffer(struct intel_ring_buffer *ring) > +int intel_allocate_ring_buffer(struct intel_engine *ring) > { > struct drm_device *dev = ring->dev; > struct drm_i915_private *dev_priv = to_i915(dev); > @@ -1424,7 +1424,7 @@ err_unref: > } > > static int intel_init_ring_buffer(struct drm_device *dev, > - struct intel_ring_buffer *ring) > + struct intel_engine *ring) > { > int ret; > > @@ -1465,7 +1465,7 @@ static int intel_init_ring_buffer(struct drm_device *dev, > return ring->init(ring); > } > > -void intel_cleanup_ring_buffer(struct intel_ring_buffer *ring) > +void intel_cleanup_ring_buffer(struct intel_engine *ring) > { > struct drm_i915_private *dev_priv = to_i915(ring->dev); > > @@ -1485,7 +1485,7 @@ void intel_cleanup_ring_buffer(struct intel_ring_buffer *ring) > cleanup_status_page(ring); > } > > -static int intel_ring_wait_request(struct intel_ring_buffer *ring, int n) > +static int intel_ring_wait_request(struct intel_engine *ring, int n) > { > struct drm_i915_gem_request *request; > u32 seqno = 0, tail; > @@ -1538,7 +1538,7 @@ static int intel_ring_wait_request(struct intel_ring_buffer *ring, int n) > return 0; > } > > -static int ring_wait_for_space(struct intel_ring_buffer *ring, int n) > +static int ring_wait_for_space(struct intel_engine *ring, int n) > { > struct drm_device *dev = ring->dev; > struct drm_i915_private *dev_priv = dev->dev_private; > @@ -1586,7 +1586,7 @@ static int ring_wait_for_space(struct intel_ring_buffer *ring, int n) > return -EBUSY; > } > > -static int intel_wrap_ring_buffer(struct intel_ring_buffer *ring) > +static int intel_wrap_ring_buffer(struct intel_engine *ring) > { > uint32_t __iomem *virt; > int rem = ring->size - ring->tail; > @@ -1608,7 +1608,7 @@ static int intel_wrap_ring_buffer(struct intel_ring_buffer *ring) > return 0; > } > > -int intel_ring_idle(struct intel_ring_buffer *ring) > +int intel_ring_idle(struct intel_engine *ring) > { > u32 seqno; > int ret; > @@ -1632,7 +1632,7 @@ int intel_ring_idle(struct intel_ring_buffer *ring) > } > > static int > -intel_ring_alloc_seqno(struct intel_ring_buffer *ring) > +intel_ring_alloc_seqno(struct intel_engine *ring) > { > if (ring->outstanding_lazy_seqno) > return 0; > @@ -1650,7 +1650,7 @@ intel_ring_alloc_seqno(struct intel_ring_buffer *ring) > return i915_gem_get_seqno(ring->dev, &ring->outstanding_lazy_seqno); > } > > -static int __intel_ring_prepare(struct intel_ring_buffer *ring, > +static int __intel_ring_prepare(struct intel_engine *ring, > int bytes) > { > int ret; > @@ -1670,7 +1670,7 @@ static int __intel_ring_prepare(struct intel_ring_buffer *ring, > return 0; > } > > -int intel_ring_begin(struct intel_ring_buffer *ring, > +int intel_ring_begin(struct intel_engine *ring, > int num_dwords) > { > struct drm_i915_private *dev_priv = ring->dev->dev_private; > @@ -1695,7 +1695,7 @@ int intel_ring_begin(struct intel_ring_buffer *ring, > } > > /* Align the ring tail to a cacheline boundary */ > -int intel_ring_cacheline_align(struct intel_ring_buffer *ring) > +int intel_ring_cacheline_align(struct intel_engine *ring) > { > int num_dwords = (ring->tail & (CACHELINE_BYTES - 1)) / sizeof(uint32_t); > int ret; > @@ -1716,7 +1716,7 @@ int intel_ring_cacheline_align(struct intel_ring_buffer *ring) > return 0; > } > > -void intel_ring_init_seqno(struct intel_ring_buffer *ring, u32 seqno) > +void intel_ring_init_seqno(struct intel_engine *ring, u32 seqno) > { > struct drm_i915_private *dev_priv = ring->dev->dev_private; > > @@ -1733,7 +1733,7 @@ void intel_ring_init_seqno(struct intel_ring_buffer *ring, u32 seqno) > ring->hangcheck.seqno = seqno; > } > > -static void gen6_bsd_ring_write_tail(struct intel_ring_buffer *ring, > +static void gen6_bsd_ring_write_tail(struct intel_engine *ring, > u32 value) > { > struct drm_i915_private *dev_priv = ring->dev->dev_private; > @@ -1766,7 +1766,7 @@ static void gen6_bsd_ring_write_tail(struct intel_ring_buffer *ring, > _MASKED_BIT_DISABLE(GEN6_BSD_SLEEP_MSG_DISABLE)); > } > > -static int gen6_bsd_ring_flush(struct intel_ring_buffer *ring, > +static int gen6_bsd_ring_flush(struct intel_engine *ring, > u32 invalidate, u32 flush) > { > uint32_t cmd; > @@ -1802,7 +1802,7 @@ static int gen6_bsd_ring_flush(struct intel_ring_buffer *ring, > } > > static int > -gen8_ring_dispatch_execbuffer(struct intel_ring_buffer *ring, > +gen8_ring_dispatch_execbuffer(struct intel_engine *ring, > u64 offset, u32 len, > unsigned flags) > { > @@ -1826,7 +1826,7 @@ gen8_ring_dispatch_execbuffer(struct intel_ring_buffer *ring, > } > > static int > -hsw_ring_dispatch_execbuffer(struct intel_ring_buffer *ring, > +hsw_ring_dispatch_execbuffer(struct intel_engine *ring, > u64 offset, u32 len, > unsigned flags) > { > @@ -1847,7 +1847,7 @@ hsw_ring_dispatch_execbuffer(struct intel_ring_buffer *ring, > } > > static int > -gen6_ring_dispatch_execbuffer(struct intel_ring_buffer *ring, > +gen6_ring_dispatch_execbuffer(struct intel_engine *ring, > u64 offset, u32 len, > unsigned flags) > { > @@ -1869,7 +1869,7 @@ gen6_ring_dispatch_execbuffer(struct intel_ring_buffer *ring, > > /* Blitter support (SandyBridge+) */ > > -static int gen6_ring_flush(struct intel_ring_buffer *ring, > +static int gen6_ring_flush(struct intel_engine *ring, > u32 invalidate, u32 flush) > { > struct drm_device *dev = ring->dev; > @@ -1912,7 +1912,7 @@ static int gen6_ring_flush(struct intel_ring_buffer *ring, > int intel_init_render_ring_buffer(struct drm_device *dev) > { > struct drm_i915_private *dev_priv = dev->dev_private; > - struct intel_ring_buffer *ring = &dev_priv->ring[RCS]; > + struct intel_engine *ring = &dev_priv->ring[RCS]; > > if (INTEL_INFO(dev)->gen >= 6) { > ring->add_request = gen6_add_request; > @@ -2018,7 +2018,7 @@ int intel_init_render_ring_buffer(struct drm_device *dev) > int intel_render_ring_init_dri(struct drm_device *dev, u64 start, u32 size) > { > struct drm_i915_private *dev_priv = dev->dev_private; > - struct intel_ring_buffer *ring = &dev_priv->ring[RCS]; > + struct intel_engine *ring = &dev_priv->ring[RCS]; > int ret; > > if (INTEL_INFO(dev)->gen >= 6) { > @@ -2081,7 +2081,7 @@ int intel_render_ring_init_dri(struct drm_device *dev, u64 start, u32 size) > int intel_init_bsd_ring_buffer(struct drm_device *dev) > { > struct drm_i915_private *dev_priv = dev->dev_private; > - struct intel_ring_buffer *ring = &dev_priv->ring[VCS]; > + struct intel_engine *ring = &dev_priv->ring[VCS]; > > ring->write_tail = ring_write_tail; > if (INTEL_INFO(dev)->gen >= 6) { > @@ -2152,7 +2152,7 @@ int intel_init_bsd_ring_buffer(struct drm_device *dev) > int intel_init_bsd2_ring_buffer(struct drm_device *dev) > { > struct drm_i915_private *dev_priv = dev->dev_private; > - struct intel_ring_buffer *ring = &dev_priv->ring[VCS2]; > + struct intel_engine *ring = &dev_priv->ring[VCS2]; > > if ((INTEL_INFO(dev)->gen != 8)) { > DRM_ERROR("No dual-BSD ring on non-BDW machine\n"); > @@ -2196,7 +2196,7 @@ int intel_init_bsd2_ring_buffer(struct drm_device *dev) > int intel_init_blt_ring_buffer(struct drm_device *dev) > { > struct drm_i915_private *dev_priv = dev->dev_private; > - struct intel_ring_buffer *ring = &dev_priv->ring[BCS]; > + struct intel_engine *ring = &dev_priv->ring[BCS]; > > ring->write_tail = ring_write_tail; > ring->flush = gen6_ring_flush; > @@ -2241,7 +2241,7 @@ int intel_init_blt_ring_buffer(struct drm_device *dev) > int intel_init_vebox_ring_buffer(struct drm_device *dev) > { > struct drm_i915_private *dev_priv = dev->dev_private; > - struct intel_ring_buffer *ring = &dev_priv->ring[VECS]; > + struct intel_engine *ring = &dev_priv->ring[VECS]; > > ring->write_tail = ring_write_tail; > ring->flush = gen6_ring_flush; > @@ -2279,7 +2279,7 @@ int intel_init_vebox_ring_buffer(struct drm_device *dev) > } > > int > -intel_ring_flush_all_caches(struct intel_ring_buffer *ring) > +intel_ring_flush_all_caches(struct intel_engine *ring) > { > int ret; > > @@ -2297,7 +2297,7 @@ intel_ring_flush_all_caches(struct intel_ring_buffer *ring) > } > > int > -intel_ring_invalidate_all_caches(struct intel_ring_buffer *ring) > +intel_ring_invalidate_all_caches(struct intel_engine *ring) > { > uint32_t flush_domains; > int ret; > @@ -2317,7 +2317,7 @@ intel_ring_invalidate_all_caches(struct intel_ring_buffer *ring) > } > > void > -intel_stop_ring_buffer(struct intel_ring_buffer *ring) > +intel_stop_ring_buffer(struct intel_engine *ring) > { > int ret; > > diff --git a/drivers/gpu/drm/i915/intel_ringbuffer.h b/drivers/gpu/drm/i915/intel_ringbuffer.h > index 680e451..50cc525 100644 > --- a/drivers/gpu/drm/i915/intel_ringbuffer.h > +++ b/drivers/gpu/drm/i915/intel_ringbuffer.h > @@ -54,7 +54,7 @@ struct intel_ring_hangcheck { > bool deadlock; > }; > > -struct intel_ring_buffer { > +struct intel_engine { > const char *name; > enum intel_ring_id { > RCS = 0x0, > @@ -90,33 +90,33 @@ struct intel_ring_buffer { > unsigned irq_refcount; /* protected by dev_priv->irq_lock */ > u32 irq_enable_mask; /* bitmask to enable ring interrupt */ > u32 trace_irq_seqno; > - bool __must_check (*irq_get)(struct intel_ring_buffer *ring); > - void (*irq_put)(struct intel_ring_buffer *ring); > + bool __must_check (*irq_get)(struct intel_engine *ring); > + void (*irq_put)(struct intel_engine *ring); > > - int (*init)(struct intel_ring_buffer *ring); > + int (*init)(struct intel_engine *ring); > > - void (*write_tail)(struct intel_ring_buffer *ring, > + void (*write_tail)(struct intel_engine *ring, > u32 value); > - int __must_check (*flush)(struct intel_ring_buffer *ring, > + int __must_check (*flush)(struct intel_engine *ring, > u32 invalidate_domains, > u32 flush_domains); > - int (*add_request)(struct intel_ring_buffer *ring); > + int (*add_request)(struct intel_engine *ring); > /* Some chipsets are not quite as coherent as advertised and need > * an expensive kick to force a true read of the up-to-date seqno. > * However, the up-to-date seqno is not always required and the last > * seen value is good enough. Note that the seqno will always be > * monotonic, even if not coherent. > */ > - u32 (*get_seqno)(struct intel_ring_buffer *ring, > + u32 (*get_seqno)(struct intel_engine *ring, > bool lazy_coherency); > - void (*set_seqno)(struct intel_ring_buffer *ring, > + void (*set_seqno)(struct intel_engine *ring, > u32 seqno); > - int (*dispatch_execbuffer)(struct intel_ring_buffer *ring, > + int (*dispatch_execbuffer)(struct intel_engine *ring, > u64 offset, u32 length, > unsigned flags); > #define I915_DISPATCH_SECURE 0x1 > #define I915_DISPATCH_PINNED 0x2 > - void (*cleanup)(struct intel_ring_buffer *ring); > + void (*cleanup)(struct intel_engine *ring); > > struct { > u32 sync_seqno[I915_NUM_RINGS-1]; > @@ -129,10 +129,10 @@ struct intel_ring_buffer { > } mbox; > > /* AKA wait() */ > - int (*sync_to)(struct intel_ring_buffer *ring, > - struct intel_ring_buffer *to, > + int (*sync_to)(struct intel_engine *ring, > + struct intel_engine *to, > u32 seqno); > - int (*signal)(struct intel_ring_buffer *signaller, > + int (*signal)(struct intel_engine *signaller, > /* num_dwords needed by caller */ > unsigned int num_dwords); > } semaphore; > @@ -210,20 +210,20 @@ struct intel_ring_buffer { > }; > > static inline bool > -intel_ring_initialized(struct intel_ring_buffer *ring) > +intel_ring_initialized(struct intel_engine *ring) > { > return ring->obj != NULL; > } > > static inline unsigned > -intel_ring_flag(struct intel_ring_buffer *ring) > +intel_ring_flag(struct intel_engine *ring) > { > return 1 << ring->id; > } > > static inline u32 > -intel_ring_sync_index(struct intel_ring_buffer *ring, > - struct intel_ring_buffer *other) > +intel_ring_sync_index(struct intel_engine *ring, > + struct intel_engine *other) > { > int idx; > > @@ -241,7 +241,7 @@ intel_ring_sync_index(struct intel_ring_buffer *ring, > } > > static inline u32 > -intel_read_status_page(struct intel_ring_buffer *ring, > +intel_read_status_page(struct intel_engine *ring, > int reg) > { > /* Ensure that the compiler doesn't optimize away the load. */ > @@ -250,7 +250,7 @@ intel_read_status_page(struct intel_ring_buffer *ring, > } > > static inline void > -intel_write_status_page(struct intel_ring_buffer *ring, > +intel_write_status_page(struct intel_engine *ring, > int reg, u32 value) > { > ring->status_page.page_addr[reg] = value; > @@ -275,27 +275,27 @@ intel_write_status_page(struct intel_ring_buffer *ring, > #define I915_GEM_HWS_SCRATCH_INDEX 0x30 > #define I915_GEM_HWS_SCRATCH_ADDR (I915_GEM_HWS_SCRATCH_INDEX << MI_STORE_DWORD_INDEX_SHIFT) > > -void intel_stop_ring_buffer(struct intel_ring_buffer *ring); > -void intel_cleanup_ring_buffer(struct intel_ring_buffer *ring); > +void intel_stop_ring_buffer(struct intel_engine *ring); > +void intel_cleanup_ring_buffer(struct intel_engine *ring); > > -int __must_check intel_ring_begin(struct intel_ring_buffer *ring, int n); > -int __must_check intel_ring_cacheline_align(struct intel_ring_buffer *ring); > -static inline void intel_ring_emit(struct intel_ring_buffer *ring, > +int __must_check intel_ring_begin(struct intel_engine *ring, int n); > +int __must_check intel_ring_cacheline_align(struct intel_engine *ring); > +static inline void intel_ring_emit(struct intel_engine *ring, > u32 data) > { > iowrite32(data, ring->virtual_start + ring->tail); > ring->tail += 4; > } > -static inline void intel_ring_advance(struct intel_ring_buffer *ring) > +static inline void intel_ring_advance(struct intel_engine *ring) > { > ring->tail &= ring->size - 1; > } > -void __intel_ring_advance(struct intel_ring_buffer *ring); > +void __intel_ring_advance(struct intel_engine *ring); > > -int __must_check intel_ring_idle(struct intel_ring_buffer *ring); > -void intel_ring_init_seqno(struct intel_ring_buffer *ring, u32 seqno); > -int intel_ring_flush_all_caches(struct intel_ring_buffer *ring); > -int intel_ring_invalidate_all_caches(struct intel_ring_buffer *ring); > +int __must_check intel_ring_idle(struct intel_engine *ring); > +void intel_ring_init_seqno(struct intel_engine *ring, u32 seqno); > +int intel_ring_flush_all_caches(struct intel_engine *ring); > +int intel_ring_invalidate_all_caches(struct intel_engine *ring); > > void intel_init_rings_early(struct drm_device *dev); > int intel_init_render_ring_buffer(struct drm_device *dev); > @@ -304,24 +304,24 @@ int intel_init_bsd2_ring_buffer(struct drm_device *dev); > int intel_init_blt_ring_buffer(struct drm_device *dev); > int intel_init_vebox_ring_buffer(struct drm_device *dev); > > -u64 intel_ring_get_active_head(struct intel_ring_buffer *ring); > -void intel_ring_setup_status_page(struct intel_ring_buffer *ring); > +u64 intel_ring_get_active_head(struct intel_engine *ring); > +void intel_ring_setup_status_page(struct intel_engine *ring); > > -void intel_destroy_ring_buffer(struct intel_ring_buffer *ring); > -int intel_allocate_ring_buffer(struct intel_ring_buffer *ring); > +void intel_destroy_ring_buffer(struct intel_engine *ring); > +int intel_allocate_ring_buffer(struct intel_engine *ring); > > -static inline u32 intel_ring_get_tail(struct intel_ring_buffer *ring) > +static inline u32 intel_ring_get_tail(struct intel_engine *ring) > { > return ring->tail; > } > > -static inline u32 intel_ring_get_seqno(struct intel_ring_buffer *ring) > +static inline u32 intel_ring_get_seqno(struct intel_engine *ring) > { > BUG_ON(ring->outstanding_lazy_seqno == 0); > return ring->outstanding_lazy_seqno; > } > > -static inline void i915_trace_irq_get(struct intel_ring_buffer *ring, u32 seqno) > +static inline void i915_trace_irq_get(struct intel_engine *ring, u32 seqno) > { > if (ring->trace_irq_seqno == 0 && ring->irq_get(ring)) > ring->trace_irq_seqno = seqno; > -- > 1.9.0 > > _______________________________________________ > Intel-gfx mailing list > Intel-gfx@xxxxxxxxxxxxxxxxxxxxx > http://lists.freedesktop.org/mailman/listinfo/intel-gfx -- Daniel Vetter Software Engineer, Intel Corporation +41 (0) 79 365 57 48 - http://blog.ffwll.ch _______________________________________________ Intel-gfx mailing list Intel-gfx@xxxxxxxxxxxxxxxxxxxxx http://lists.freedesktop.org/mailman/listinfo/intel-gfx