On Thu, Jun 13, 2019 at 04:12:16PM +0100, Tvrtko Ursulin wrote: > > On 12/06/2019 17:29, Lucas De Marchi wrote: > > We are slowly converting dev_priv to i915 everywhere, spread into > > smaller series. While this is good to avoid unrelated breakages to other > > inflight patches, it's bad because inflight patches on nearby paths keep > > breaking. Paired with other code moves and refactores this is becoming a > > nightmare. > > > > Now that I915_{READ,WRITE} are gone and implicit access to dev_priv no > > Who removed I915_READ/WRITE? In fact I see some in this patch. Well colour > me confused.. how did the patch compile? as soon as this becomes true I agree with Lucas that the big sed is the best solution. > > Plus it's not I915_READ/WRITE - some display macros have hidden dependency > on implicit dev_priv as well. > > > longer exists we can simply sed all files and be done with the > > conversion. This was generated with the following commands with no > > additional fixups: > > > > git ls-files --full-name -z -- drivers/gpu/drm/i915/ | \ > > xargs -0 sed -i 's/\bdev_priv\b/i915/g' > > > > Any pending series can apply the same search and replace when rebasing. > > I am in two minds. Converting piece meal is easier for some use cases, and > converting all at once is easier for some others. Don't know. > > Regards, > > Tvrtko > > > > > Signed-off-by: Lucas De Marchi <lucas.demarchi@xxxxxxxxx> > > --- > > drivers/gpu/drm/i915/dvo_ch7017.c | 6 +- > > drivers/gpu/drm/i915/dvo_ch7xxx.c | 10 +- > > drivers/gpu/drm/i915/dvo_ivch.c | 14 +- > > drivers/gpu/drm/i915/dvo_ns2501.c | 14 +- > > drivers/gpu/drm/i915/dvo_sil164.c | 10 +- > > drivers/gpu/drm/i915/dvo_tfp410.c | 10 +- > > drivers/gpu/drm/i915/gem/i915_gem_context.c | 48 +- > > drivers/gpu/drm/i915/gem/i915_gem_context.h | 6 +- > > .../gpu/drm/i915/gem/i915_gem_execbuffer.c | 4 +- > > drivers/gpu/drm/i915/gem/i915_gem_object.c | 4 +- > > drivers/gpu/drm/i915/gem/i915_gem_shmem.c | 4 +- > > drivers/gpu/drm/i915/gem/i915_gem_stolen.c | 168 +- > > drivers/gpu/drm/i915/gem/i915_gem_tiling.c | 8 +- > > drivers/gpu/drm/i915/gem/i915_gem_userptr.c | 40 +- > > .../gpu/drm/i915/gem/selftests/huge_pages.c | 48 +- > > .../drm/i915/gem/selftests/i915_gem_context.c | 6 +- > > drivers/gpu/drm/i915/gt/intel_engine.h | 2 +- > > drivers/gpu/drm/i915/gt/intel_engine_cs.c | 66 +- > > drivers/gpu/drm/i915/gt/intel_hangcheck.c | 36 +- > > drivers/gpu/drm/i915/gt/intel_mocs.c | 24 +- > > drivers/gpu/drm/i915/gt/intel_mocs.h | 2 +- > > drivers/gpu/drm/i915/gt/intel_reset.c | 6 +- > > drivers/gpu/drm/i915/gt/intel_ringbuffer.c | 52 +- > > drivers/gpu/drm/i915/gvt/aperture_gm.c | 60 +- > > drivers/gpu/drm/i915/gvt/cfg_space.c | 4 +- > > drivers/gpu/drm/i915/gvt/cmd_parser.c | 34 +- > > drivers/gpu/drm/i915/gvt/debugfs.c | 12 +- > > drivers/gpu/drm/i915/gvt/display.c | 48 +- > > drivers/gpu/drm/i915/gvt/dmabuf.c | 12 +- > > drivers/gpu/drm/i915/gvt/edid.c | 6 +- > > drivers/gpu/drm/i915/gvt/execlist.c | 14 +- > > drivers/gpu/drm/i915/gvt/fb_decoder.c | 12 +- > > drivers/gpu/drm/i915/gvt/firmware.c | 16 +- > > drivers/gpu/drm/i915/gvt/gtt.c | 44 +- > > drivers/gpu/drm/i915/gvt/gvt.c | 26 +- > > drivers/gpu/drm/i915/gvt/gvt.h | 20 +- > > drivers/gpu/drm/i915/gvt/handlers.c | 92 +- > > drivers/gpu/drm/i915/gvt/interrupt.c | 6 +- > > drivers/gpu/drm/i915/gvt/kvmgt.c | 6 +- > > drivers/gpu/drm/i915/gvt/mmio.c | 2 +- > > drivers/gpu/drm/i915/gvt/mmio_context.c | 38 +- > > drivers/gpu/drm/i915/gvt/sched_policy.c | 12 +- > > drivers/gpu/drm/i915/gvt/scheduler.c | 72 +- > > drivers/gpu/drm/i915/gvt/vgpu.c | 6 +- > > drivers/gpu/drm/i915/i915_cmd_parser.c | 6 +- > > drivers/gpu/drm/i915/i915_debugfs.c | 952 +++--- > > drivers/gpu/drm/i915/i915_debugfs.h | 4 +- > > drivers/gpu/drm/i915/i915_drv.c | 1092 +++--- > > drivers/gpu/drm/i915/i915_drv.h | 590 ++-- > > drivers/gpu/drm/i915/i915_gem.c | 356 +- > > drivers/gpu/drm/i915/i915_gem_evict.c | 6 +- > > drivers/gpu/drm/i915/i915_gem_gtt.c | 214 +- > > drivers/gpu/drm/i915/i915_gem_gtt.h | 20 +- > > drivers/gpu/drm/i915/i915_gpu_error.c | 44 +- > > drivers/gpu/drm/i915/i915_gpu_error.h | 6 +- > > drivers/gpu/drm/i915/i915_irq.c | 1808 +++++----- > > drivers/gpu/drm/i915/i915_irq.h | 86 +- > > drivers/gpu/drm/i915/i915_memcpy.c | 2 +- > > drivers/gpu/drm/i915/i915_oa_bdw.c | 32 +- > > drivers/gpu/drm/i915/i915_oa_bdw.h | 2 +- > > drivers/gpu/drm/i915/i915_oa_bxt.c | 32 +- > > drivers/gpu/drm/i915/i915_oa_bxt.h | 2 +- > > drivers/gpu/drm/i915/i915_oa_cflgt2.c | 32 +- > > drivers/gpu/drm/i915/i915_oa_cflgt2.h | 2 +- > > drivers/gpu/drm/i915/i915_oa_cflgt3.c | 32 +- > > drivers/gpu/drm/i915/i915_oa_cflgt3.h | 2 +- > > drivers/gpu/drm/i915/i915_oa_chv.c | 32 +- > > drivers/gpu/drm/i915/i915_oa_chv.h | 2 +- > > drivers/gpu/drm/i915/i915_oa_cnl.c | 32 +- > > drivers/gpu/drm/i915/i915_oa_cnl.h | 2 +- > > drivers/gpu/drm/i915/i915_oa_glk.c | 32 +- > > drivers/gpu/drm/i915/i915_oa_glk.h | 2 +- > > drivers/gpu/drm/i915/i915_oa_hsw.c | 32 +- > > drivers/gpu/drm/i915/i915_oa_hsw.h | 2 +- > > drivers/gpu/drm/i915/i915_oa_icl.c | 32 +- > > drivers/gpu/drm/i915/i915_oa_icl.h | 2 +- > > drivers/gpu/drm/i915/i915_oa_kblgt2.c | 32 +- > > drivers/gpu/drm/i915/i915_oa_kblgt2.h | 2 +- > > drivers/gpu/drm/i915/i915_oa_kblgt3.c | 32 +- > > drivers/gpu/drm/i915/i915_oa_kblgt3.h | 2 +- > > drivers/gpu/drm/i915/i915_oa_sklgt2.c | 32 +- > > drivers/gpu/drm/i915/i915_oa_sklgt2.h | 2 +- > > drivers/gpu/drm/i915/i915_oa_sklgt3.c | 32 +- > > drivers/gpu/drm/i915/i915_oa_sklgt3.h | 2 +- > > drivers/gpu/drm/i915/i915_oa_sklgt4.c | 32 +- > > drivers/gpu/drm/i915/i915_oa_sklgt4.h | 2 +- > > drivers/gpu/drm/i915/i915_perf.c | 814 ++--- > > drivers/gpu/drm/i915/i915_pmu.c | 42 +- > > drivers/gpu/drm/i915/i915_query.c | 10 +- > > drivers/gpu/drm/i915/i915_reg.h | 420 +-- > > drivers/gpu/drm/i915/i915_suspend.c | 106 +- > > drivers/gpu/drm/i915/i915_sysfs.c | 172 +- > > drivers/gpu/drm/i915/i915_trace.h | 46 +- > > drivers/gpu/drm/i915/i915_vgpu.c | 30 +- > > drivers/gpu/drm/i915/i915_vgpu.h | 16 +- > > drivers/gpu/drm/i915/i915_vma.c | 4 +- > > drivers/gpu/drm/i915/icl_dsi.c | 122 +- > > drivers/gpu/drm/i915/intel_atomic.c | 28 +- > > drivers/gpu/drm/i915/intel_atomic.h | 2 +- > > drivers/gpu/drm/i915/intel_audio.c | 222 +- > > drivers/gpu/drm/i915/intel_audio.h | 6 +- > > drivers/gpu/drm/i915/intel_bios.c | 552 ++-- > > drivers/gpu/drm/i915/intel_bios.h | 18 +- > > drivers/gpu/drm/i915/intel_bw.c | 68 +- > > drivers/gpu/drm/i915/intel_bw.h | 8 +- > > drivers/gpu/drm/i915/intel_cdclk.c | 882 ++--- > > drivers/gpu/drm/i915/intel_cdclk.h | 14 +- > > drivers/gpu/drm/i915/intel_color.c | 176 +- > > drivers/gpu/drm/i915/intel_combo_phy.c | 64 +- > > drivers/gpu/drm/i915/intel_combo_phy.h | 6 +- > > drivers/gpu/drm/i915/intel_connector.c | 12 +- > > drivers/gpu/drm/i915/intel_crt.c | 172 +- > > drivers/gpu/drm/i915/intel_crt.h | 4 +- > > drivers/gpu/drm/i915/intel_csr.c | 144 +- > > drivers/gpu/drm/i915/intel_ddi.c | 592 ++-- > > drivers/gpu/drm/i915/intel_ddi.h | 6 +- > > drivers/gpu/drm/i915/intel_device_info.c | 164 +- > > drivers/gpu/drm/i915/intel_device_info.h | 6 +- > > drivers/gpu/drm/i915/intel_display.c | 2916 ++++++++--------- > > drivers/gpu/drm/i915/intel_display.h | 6 +- > > drivers/gpu/drm/i915/intel_display_power.c | 1020 +++--- > > drivers/gpu/drm/i915/intel_display_power.h | 62 +- > > drivers/gpu/drm/i915/intel_dp.c | 822 ++--- > > drivers/gpu/drm/i915/intel_dp.h | 18 +- > > drivers/gpu/drm/i915/intel_dp_aux_backlight.c | 4 +- > > drivers/gpu/drm/i915/intel_dp_mst.c | 32 +- > > drivers/gpu/drm/i915/intel_dpio_phy.c | 276 +- > > drivers/gpu/drm/i915/intel_dpio_phy.h | 12 +- > > drivers/gpu/drm/i915/intel_dpll_mgr.c | 356 +- > > drivers/gpu/drm/i915/intel_dpll_mgr.h | 20 +- > > drivers/gpu/drm/i915/intel_drv.h | 72 +- > > drivers/gpu/drm/i915/intel_dsi.c | 6 +- > > drivers/gpu/drm/i915/intel_dsi.h | 6 +- > > .../gpu/drm/i915/intel_dsi_dcs_backlight.c | 4 +- > > drivers/gpu/drm/i915/intel_dsi_vbt.c | 76 +- > > drivers/gpu/drm/i915/intel_dvo.c | 32 +- > > drivers/gpu/drm/i915/intel_dvo.h | 2 +- > > drivers/gpu/drm/i915/intel_dvo_dev.h | 2 +- > > drivers/gpu/drm/i915/intel_fbc.c | 410 +-- > > drivers/gpu/drm/i915/intel_fbc.h | 20 +- > > drivers/gpu/drm/i915/intel_fbdev.c | 46 +- > > drivers/gpu/drm/i915/intel_fbdev.h | 8 +- > > drivers/gpu/drm/i915/intel_fifo_underrun.c | 160 +- > > drivers/gpu/drm/i915/intel_fifo_underrun.h | 12 +- > > drivers/gpu/drm/i915/intel_frontbuffer.c | 82 +- > > drivers/gpu/drm/i915/intel_frontbuffer.h | 6 +- > > drivers/gpu/drm/i915/intel_gmbus.c | 230 +- > > drivers/gpu/drm/i915/intel_gmbus.h | 10 +- > > drivers/gpu/drm/i915/intel_guc.c | 48 +- > > drivers/gpu/drm/i915/intel_guc_ads.c | 12 +- > > drivers/gpu/drm/i915/intel_guc_fw.c | 20 +- > > drivers/gpu/drm/i915/intel_guc_log.c | 18 +- > > drivers/gpu/drm/i915/intel_guc_submission.c | 56 +- > > drivers/gpu/drm/i915/intel_gvt.c | 36 +- > > drivers/gpu/drm/i915/intel_gvt.h | 16 +- > > drivers/gpu/drm/i915/intel_hdcp.c | 286 +- > > drivers/gpu/drm/i915/intel_hdcp.h | 6 +- > > drivers/gpu/drm/i915/intel_hdmi.c | 316 +- > > drivers/gpu/drm/i915/intel_hdmi.h | 2 +- > > drivers/gpu/drm/i915/intel_hotplug.c | 210 +- > > drivers/gpu/drm/i915/intel_hotplug.h | 16 +- > > drivers/gpu/drm/i915/intel_huc.c | 6 +- > > drivers/gpu/drm/i915/intel_huc_fw.c | 18 +- > > drivers/gpu/drm/i915/intel_lpe_audio.c | 96 +- > > drivers/gpu/drm/i915/intel_lpe_audio.h | 8 +- > > drivers/gpu/drm/i915/intel_lspcon.c | 4 +- > > drivers/gpu/drm/i915/intel_lvds.c | 96 +- > > drivers/gpu/drm/i915/intel_lvds.h | 8 +- > > drivers/gpu/drm/i915/intel_opregion.c | 114 +- > > drivers/gpu/drm/i915/intel_opregion.h | 28 +- > > drivers/gpu/drm/i915/intel_overlay.c | 98 +- > > drivers/gpu/drm/i915/intel_overlay.h | 8 +- > > drivers/gpu/drm/i915/intel_panel.c | 238 +- > > drivers/gpu/drm/i915/intel_pipe_crc.c | 120 +- > > drivers/gpu/drm/i915/intel_pipe_crc.h | 4 +- > > drivers/gpu/drm/i915/intel_pm.c | 2300 ++++++------- > > drivers/gpu/drm/i915/intel_pm.h | 74 +- > > drivers/gpu/drm/i915/intel_psr.c | 394 +-- > > drivers/gpu/drm/i915/intel_psr.h | 14 +- > > drivers/gpu/drm/i915/intel_quirks.h | 2 +- > > drivers/gpu/drm/i915/intel_runtime_pm.h | 8 +- > > drivers/gpu/drm/i915/intel_sdvo.c | 128 +- > > drivers/gpu/drm/i915/intel_sdvo.h | 4 +- > > drivers/gpu/drm/i915/intel_sprite.c | 218 +- > > drivers/gpu/drm/i915/intel_sprite.h | 8 +- > > drivers/gpu/drm/i915/intel_tv.c | 72 +- > > drivers/gpu/drm/i915/intel_tv.h | 2 +- > > drivers/gpu/drm/i915/intel_uc.c | 4 +- > > drivers/gpu/drm/i915/intel_uc.h | 22 +- > > drivers/gpu/drm/i915/intel_uc_fw.c | 12 +- > > drivers/gpu/drm/i915/intel_uc_fw.h | 2 +- > > drivers/gpu/drm/i915/intel_uncore.c | 26 +- > > drivers/gpu/drm/i915/intel_uncore.h | 4 +- > > drivers/gpu/drm/i915/intel_vbt_defs.h | 2 +- > > drivers/gpu/drm/i915/intel_vdsc.c | 10 +- > > drivers/gpu/drm/i915/intel_wopcm.c | 14 +- > > drivers/gpu/drm/i915/selftests/i915_gem_gtt.c | 26 +- > > drivers/gpu/drm/i915/selftests/intel_guc.c | 46 +- > > drivers/gpu/drm/i915/selftests/intel_uncore.c | 10 +- > > drivers/gpu/drm/i915/vlv_dsi.c | 222 +- > > drivers/gpu/drm/i915/vlv_dsi_pll.c | 80 +- > > 201 files changed, 12020 insertions(+), 12020 deletions(-) > > > > diff --git a/drivers/gpu/drm/i915/dvo_ch7017.c b/drivers/gpu/drm/i915/dvo_ch7017.c > > index 602380fe74f3..df670c8d5c31 100644 > > --- a/drivers/gpu/drm/i915/dvo_ch7017.c > > +++ b/drivers/gpu/drm/i915/dvo_ch7017.c > > @@ -210,7 +210,7 @@ static bool ch7017_init(struct intel_dvo_device *dvo, > > return false; > > dvo->i2c_bus = adapter; > > - dvo->dev_priv = priv; > > + dvo->i915 = priv; > > if (!ch7017_read(dvo, CH7017_DEVICE_ID, &val)) > > goto fail; > > @@ -395,11 +395,11 @@ do { \ > > static void ch7017_destroy(struct intel_dvo_device *dvo) > > { > > - struct ch7017_priv *priv = dvo->dev_priv; > > + struct ch7017_priv *priv = dvo->i915; > > if (priv) { > > kfree(priv); > > - dvo->dev_priv = NULL; > > + dvo->i915 = NULL; > > } > > } > > diff --git a/drivers/gpu/drm/i915/dvo_ch7xxx.c b/drivers/gpu/drm/i915/dvo_ch7xxx.c > > index e070bebee7b5..e076801d032f 100644 > > --- a/drivers/gpu/drm/i915/dvo_ch7xxx.c > > +++ b/drivers/gpu/drm/i915/dvo_ch7xxx.c > > @@ -135,7 +135,7 @@ static char *ch7xxx_get_did(u8 did) > > /** Reads an 8 bit register */ > > static bool ch7xxx_readb(struct intel_dvo_device *dvo, int addr, u8 *ch) > > { > > - struct ch7xxx_priv *ch7xxx = dvo->dev_priv; > > + struct ch7xxx_priv *ch7xxx = dvo->i915; > > struct i2c_adapter *adapter = dvo->i2c_bus; > > u8 out_buf[2]; > > u8 in_buf[2]; > > @@ -173,7 +173,7 @@ static bool ch7xxx_readb(struct intel_dvo_device *dvo, int addr, u8 *ch) > > /** Writes an 8 bit register */ > > static bool ch7xxx_writeb(struct intel_dvo_device *dvo, int addr, u8 ch) > > { > > - struct ch7xxx_priv *ch7xxx = dvo->dev_priv; > > + struct ch7xxx_priv *ch7xxx = dvo->i915; > > struct i2c_adapter *adapter = dvo->i2c_bus; > > u8 out_buf[2]; > > struct i2c_msg msg = { > > @@ -210,7 +210,7 @@ static bool ch7xxx_init(struct intel_dvo_device *dvo, > > return false; > > dvo->i2c_bus = adapter; > > - dvo->dev_priv = ch7xxx; > > + dvo->i915 = ch7xxx; > > ch7xxx->quiet = true; > > if (!ch7xxx_readb(dvo, CH7xxx_REG_VID, &vendor)) > > @@ -347,11 +347,11 @@ static void ch7xxx_dump_regs(struct intel_dvo_device *dvo) > > static void ch7xxx_destroy(struct intel_dvo_device *dvo) > > { > > - struct ch7xxx_priv *ch7xxx = dvo->dev_priv; > > + struct ch7xxx_priv *ch7xxx = dvo->i915; > > if (ch7xxx) { > > kfree(ch7xxx); > > - dvo->dev_priv = NULL; > > + dvo->i915 = NULL; > > } > > } > > diff --git a/drivers/gpu/drm/i915/dvo_ivch.c b/drivers/gpu/drm/i915/dvo_ivch.c > > index 09dba35f3ffa..c99c16211567 100644 > > --- a/drivers/gpu/drm/i915/dvo_ivch.c > > +++ b/drivers/gpu/drm/i915/dvo_ivch.c > > @@ -191,7 +191,7 @@ static void ivch_dump_regs(struct intel_dvo_device *dvo); > > */ > > static bool ivch_read(struct intel_dvo_device *dvo, int addr, u16 *data) > > { > > - struct ivch_priv *priv = dvo->dev_priv; > > + struct ivch_priv *priv = dvo->i915; > > struct i2c_adapter *adapter = dvo->i2c_bus; > > u8 out_buf[1]; > > u8 in_buf[2]; > > @@ -234,7 +234,7 @@ static bool ivch_read(struct intel_dvo_device *dvo, int addr, u16 *data) > > /* Writes a 16-bit register on the ivch */ > > static bool ivch_write(struct intel_dvo_device *dvo, int addr, u16 data) > > { > > - struct ivch_priv *priv = dvo->dev_priv; > > + struct ivch_priv *priv = dvo->i915; > > struct i2c_adapter *adapter = dvo->i2c_bus; > > u8 out_buf[3]; > > struct i2c_msg msg = { > > @@ -272,7 +272,7 @@ static bool ivch_init(struct intel_dvo_device *dvo, > > return false; > > dvo->i2c_bus = adapter; > > - dvo->dev_priv = priv; > > + dvo->i915 = priv; > > priv->quiet = true; > > if (!ivch_read(dvo, VR00, &temp)) > > @@ -328,7 +328,7 @@ static enum drm_mode_status ivch_mode_valid(struct intel_dvo_device *dvo, > > */ > > static void ivch_reset(struct intel_dvo_device *dvo) > > { > > - struct ivch_priv *priv = dvo->dev_priv; > > + struct ivch_priv *priv = dvo->i915; > > int i; > > DRM_DEBUG_KMS("Resetting the IVCH registers\n"); > > @@ -398,7 +398,7 @@ static void ivch_mode_set(struct intel_dvo_device *dvo, > > const struct drm_display_mode *mode, > > const struct drm_display_mode *adjusted_mode) > > { > > - struct ivch_priv *priv = dvo->dev_priv; > > + struct ivch_priv *priv = dvo->i915; > > u16 vr40 = 0; > > u16 vr01 = 0; > > u16 vr10; > > @@ -483,11 +483,11 @@ static void ivch_dump_regs(struct intel_dvo_device *dvo) > > static void ivch_destroy(struct intel_dvo_device *dvo) > > { > > - struct ivch_priv *priv = dvo->dev_priv; > > + struct ivch_priv *priv = dvo->i915; > > if (priv) { > > kfree(priv); > > - dvo->dev_priv = NULL; > > + dvo->i915 = NULL; > > } > > } > > diff --git a/drivers/gpu/drm/i915/dvo_ns2501.c b/drivers/gpu/drm/i915/dvo_ns2501.c > > index c83a5d88d62b..af04fcaa98b8 100644 > > --- a/drivers/gpu/drm/i915/dvo_ns2501.c > > +++ b/drivers/gpu/drm/i915/dvo_ns2501.c > > @@ -392,7 +392,7 @@ struct ns2501_priv { > > */ > > static bool ns2501_readb(struct intel_dvo_device *dvo, int addr, u8 *ch) > > { > > - struct ns2501_priv *ns = dvo->dev_priv; > > + struct ns2501_priv *ns = dvo->i915; > > struct i2c_adapter *adapter = dvo->i2c_bus; > > u8 out_buf[2]; > > u8 in_buf[2]; > > @@ -437,7 +437,7 @@ static bool ns2501_readb(struct intel_dvo_device *dvo, int addr, u8 *ch) > > */ > > static bool ns2501_writeb(struct intel_dvo_device *dvo, int addr, u8 ch) > > { > > - struct ns2501_priv *ns = dvo->dev_priv; > > + struct ns2501_priv *ns = dvo->i915; > > struct i2c_adapter *adapter = dvo->i2c_bus; > > u8 out_buf[2]; > > @@ -481,7 +481,7 @@ static bool ns2501_init(struct intel_dvo_device *dvo, > > return false; > > dvo->i2c_bus = adapter; > > - dvo->dev_priv = ns; > > + dvo->i915 = ns; > > ns->quiet = true; > > if (!ns2501_readb(dvo, NS2501_VID_LO, &ch)) > > @@ -551,7 +551,7 @@ static void ns2501_mode_set(struct intel_dvo_device *dvo, > > const struct drm_display_mode *adjusted_mode) > > { > > const struct ns2501_configuration *conf; > > - struct ns2501_priv *ns = (struct ns2501_priv *)(dvo->dev_priv); > > + struct ns2501_priv *ns = (struct ns2501_priv *)(dvo->i915); > > int mode_idx, i; > > DRM_DEBUG_KMS > > @@ -655,7 +655,7 @@ static bool ns2501_get_hw_state(struct intel_dvo_device *dvo) > > /* set the NS2501 power state */ > > static void ns2501_dpms(struct intel_dvo_device *dvo, bool enable) > > { > > - struct ns2501_priv *ns = (struct ns2501_priv *)(dvo->dev_priv); > > + struct ns2501_priv *ns = (struct ns2501_priv *)(dvo->i915); > > DRM_DEBUG_KMS("Trying set the dpms of the DVO to %i\n", enable); > > @@ -691,11 +691,11 @@ static void ns2501_dpms(struct intel_dvo_device *dvo, bool enable) > > static void ns2501_destroy(struct intel_dvo_device *dvo) > > { > > - struct ns2501_priv *ns = dvo->dev_priv; > > + struct ns2501_priv *ns = dvo->i915; > > if (ns) { > > kfree(ns); > > - dvo->dev_priv = NULL; > > + dvo->i915 = NULL; > > } > > } > > diff --git a/drivers/gpu/drm/i915/dvo_sil164.c b/drivers/gpu/drm/i915/dvo_sil164.c > > index 04698eaeb632..a452dcba179c 100644 > > --- a/drivers/gpu/drm/i915/dvo_sil164.c > > +++ b/drivers/gpu/drm/i915/dvo_sil164.c > > @@ -68,7 +68,7 @@ struct sil164_priv { > > static bool sil164_readb(struct intel_dvo_device *dvo, int addr, u8 *ch) > > { > > - struct sil164_priv *sil = dvo->dev_priv; > > + struct sil164_priv *sil = dvo->i915; > > struct i2c_adapter *adapter = dvo->i2c_bus; > > u8 out_buf[2]; > > u8 in_buf[2]; > > @@ -105,7 +105,7 @@ static bool sil164_readb(struct intel_dvo_device *dvo, int addr, u8 *ch) > > static bool sil164_writeb(struct intel_dvo_device *dvo, int addr, u8 ch) > > { > > - struct sil164_priv *sil = dvo->dev_priv; > > + struct sil164_priv *sil = dvo->i915; > > struct i2c_adapter *adapter = dvo->i2c_bus; > > u8 out_buf[2]; > > struct i2c_msg msg = { > > @@ -142,7 +142,7 @@ static bool sil164_init(struct intel_dvo_device *dvo, > > return false; > > dvo->i2c_bus = adapter; > > - dvo->dev_priv = sil; > > + dvo->i915 = sil; > > sil->quiet = true; > > if (!sil164_readb(dvo, SIL164_VID_LO, &ch)) > > @@ -260,11 +260,11 @@ static void sil164_dump_regs(struct intel_dvo_device *dvo) > > static void sil164_destroy(struct intel_dvo_device *dvo) > > { > > - struct sil164_priv *sil = dvo->dev_priv; > > + struct sil164_priv *sil = dvo->i915; > > if (sil) { > > kfree(sil); > > - dvo->dev_priv = NULL; > > + dvo->i915 = NULL; > > } > > } > > diff --git a/drivers/gpu/drm/i915/dvo_tfp410.c b/drivers/gpu/drm/i915/dvo_tfp410.c > > index 623114ee73cd..8ffed8914b49 100644 > > --- a/drivers/gpu/drm/i915/dvo_tfp410.c > > +++ b/drivers/gpu/drm/i915/dvo_tfp410.c > > @@ -93,7 +93,7 @@ struct tfp410_priv { > > static bool tfp410_readb(struct intel_dvo_device *dvo, int addr, u8 *ch) > > { > > - struct tfp410_priv *tfp = dvo->dev_priv; > > + struct tfp410_priv *tfp = dvo->i915; > > struct i2c_adapter *adapter = dvo->i2c_bus; > > u8 out_buf[2]; > > u8 in_buf[2]; > > @@ -130,7 +130,7 @@ static bool tfp410_readb(struct intel_dvo_device *dvo, int addr, u8 *ch) > > static bool tfp410_writeb(struct intel_dvo_device *dvo, int addr, u8 ch) > > { > > - struct tfp410_priv *tfp = dvo->dev_priv; > > + struct tfp410_priv *tfp = dvo->i915; > > struct i2c_adapter *adapter = dvo->i2c_bus; > > u8 out_buf[2]; > > struct i2c_msg msg = { > > @@ -178,7 +178,7 @@ static bool tfp410_init(struct intel_dvo_device *dvo, > > return false; > > dvo->i2c_bus = adapter; > > - dvo->dev_priv = tfp; > > + dvo->i915 = tfp; > > tfp->quiet = true; > > if ((id = tfp410_getid(dvo, TFP410_VID_LO)) != TFP410_VID) { > > @@ -299,11 +299,11 @@ static void tfp410_dump_regs(struct intel_dvo_device *dvo) > > static void tfp410_destroy(struct intel_dvo_device *dvo) > > { > > - struct tfp410_priv *tfp = dvo->dev_priv; > > + struct tfp410_priv *tfp = dvo->i915; > > if (tfp) { > > kfree(tfp); > > - dvo->dev_priv = NULL; > > + dvo->i915 = NULL; > > } > > } > > diff --git a/drivers/gpu/drm/i915/gem/i915_gem_context.c b/drivers/gpu/drm/i915/gem/i915_gem_context.c > > index c86ca9f21532..2fa6747c7b6e 100644 > > --- a/drivers/gpu/drm/i915/gem/i915_gem_context.c > > +++ b/drivers/gpu/drm/i915/gem/i915_gem_context.c > > @@ -495,27 +495,27 @@ static void __assign_ppgtt(struct i915_gem_context *ctx, > > } > > static struct i915_gem_context * > > -i915_gem_create_context(struct drm_i915_private *dev_priv, unsigned int flags) > > +i915_gem_create_context(struct drm_i915_private *i915, unsigned int flags) > > { > > struct i915_gem_context *ctx; > > - lockdep_assert_held(&dev_priv->drm.struct_mutex); > > + lockdep_assert_held(&i915->drm.struct_mutex); > > if (flags & I915_CONTEXT_CREATE_FLAGS_SINGLE_TIMELINE && > > - !HAS_EXECLISTS(dev_priv)) > > + !HAS_EXECLISTS(i915)) > > return ERR_PTR(-EINVAL); > > /* Reap the most stale context */ > > - contexts_free_first(dev_priv); > > + contexts_free_first(i915); > > - ctx = __create_context(dev_priv); > > + ctx = __create_context(i915); > > if (IS_ERR(ctx)) > > return ctx; > > - if (HAS_FULL_PPGTT(dev_priv)) { > > + if (HAS_FULL_PPGTT(i915)) { > > struct i915_ppgtt *ppgtt; > > - ppgtt = i915_ppgtt_create(dev_priv); > > + ppgtt = i915_ppgtt_create(i915); > > if (IS_ERR(ppgtt)) { > > DRM_DEBUG_DRIVER("PPGTT setup failed (%ld)\n", > > PTR_ERR(ppgtt)); > > @@ -530,7 +530,7 @@ i915_gem_create_context(struct drm_i915_private *dev_priv, unsigned int flags) > > if (flags & I915_CONTEXT_CREATE_FLAGS_SINGLE_TIMELINE) { > > struct i915_timeline *timeline; > > - timeline = i915_timeline_create(dev_priv, NULL); > > + timeline = i915_timeline_create(i915, NULL); > > if (IS_ERR(timeline)) { > > context_close(ctx); > > return ERR_CAST(timeline); > > @@ -649,19 +649,19 @@ static bool needs_preempt_context(struct drm_i915_private *i915) > > return HAS_EXECLISTS(i915); > > } > > -int i915_gem_contexts_init(struct drm_i915_private *dev_priv) > > +int i915_gem_contexts_init(struct drm_i915_private *i915) > > { > > struct i915_gem_context *ctx; > > /* Reassure ourselves we are only called once */ > > - GEM_BUG_ON(dev_priv->kernel_context); > > - GEM_BUG_ON(dev_priv->preempt_context); > > + GEM_BUG_ON(i915->kernel_context); > > + GEM_BUG_ON(i915->preempt_context); > > - intel_engine_init_ctx_wa(dev_priv->engine[RCS0]); > > - init_contexts(dev_priv); > > + intel_engine_init_ctx_wa(i915->engine[RCS0]); > > + init_contexts(i915); > > /* lowest priority; idle task */ > > - ctx = i915_gem_context_create_kernel(dev_priv, I915_PRIORITY_MIN); > > + ctx = i915_gem_context_create_kernel(i915, I915_PRIORITY_MIN); > > if (IS_ERR(ctx)) { > > DRM_ERROR("Failed to create default global context\n"); > > return PTR_ERR(ctx); > > @@ -675,31 +675,31 @@ int i915_gem_contexts_init(struct drm_i915_private *dev_priv) > > */ > > GEM_BUG_ON(ctx->hw_id); > > GEM_BUG_ON(!atomic_read(&ctx->hw_id_pin_count)); > > - dev_priv->kernel_context = ctx; > > + i915->kernel_context = ctx; > > /* highest priority; preempting task */ > > - if (needs_preempt_context(dev_priv)) { > > - ctx = i915_gem_context_create_kernel(dev_priv, INT_MAX); > > + if (needs_preempt_context(i915)) { > > + ctx = i915_gem_context_create_kernel(i915, INT_MAX); > > if (!IS_ERR(ctx)) > > - dev_priv->preempt_context = ctx; > > + i915->preempt_context = ctx; > > else > > DRM_ERROR("Failed to create preempt context; disabling preemption\n"); > > } > > DRM_DEBUG_DRIVER("%s context support initialized\n", > > - DRIVER_CAPS(dev_priv)->has_logical_contexts ? > > + DRIVER_CAPS(i915)->has_logical_contexts ? > > "logical" : "fake"); > > return 0; > > } > > -void i915_gem_contexts_lost(struct drm_i915_private *dev_priv) > > +void i915_gem_contexts_lost(struct drm_i915_private *i915) > > { > > struct intel_engine_cs *engine; > > enum intel_engine_id id; > > - lockdep_assert_held(&dev_priv->drm.struct_mutex); > > + lockdep_assert_held(&i915->drm.struct_mutex); > > - for_each_engine(engine, dev_priv, id) > > + for_each_engine(engine, i915, id) > > intel_engine_lost_context(engine); > > } > > @@ -2368,7 +2368,7 @@ int i915_gem_context_setparam_ioctl(struct drm_device *dev, void *data, > > int i915_gem_context_reset_stats_ioctl(struct drm_device *dev, > > void *data, struct drm_file *file) > > { > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > struct drm_i915_reset_stats *args = data; > > struct i915_gem_context *ctx; > > int ret; > > @@ -2390,7 +2390,7 @@ int i915_gem_context_reset_stats_ioctl(struct drm_device *dev, > > */ > > if (capable(CAP_SYS_ADMIN)) > > - args->reset_count = i915_reset_count(&dev_priv->gpu_error); > > + args->reset_count = i915_reset_count(&i915->gpu_error); > > else > > args->reset_count = 0; > > diff --git a/drivers/gpu/drm/i915/gem/i915_gem_context.h b/drivers/gpu/drm/i915/gem/i915_gem_context.h > > index 630392c77e48..f55cb8eae87a 100644 > > --- a/drivers/gpu/drm/i915/gem/i915_gem_context.h > > +++ b/drivers/gpu/drm/i915/gem/i915_gem_context.h > > @@ -133,9 +133,9 @@ static inline bool i915_gem_context_is_kernel(struct i915_gem_context *ctx) > > } > > /* i915_gem_context.c */ > > -int __must_check i915_gem_contexts_init(struct drm_i915_private *dev_priv); > > -void i915_gem_contexts_lost(struct drm_i915_private *dev_priv); > > -void i915_gem_contexts_fini(struct drm_i915_private *dev_priv); > > +int __must_check i915_gem_contexts_init(struct drm_i915_private *i915); > > +void i915_gem_contexts_lost(struct drm_i915_private *i915); > > +void i915_gem_contexts_fini(struct drm_i915_private *i915); > > int i915_gem_context_open(struct drm_i915_private *i915, > > struct drm_file *file); > > diff --git a/drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c b/drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c > > index 528eea44dccf..83d64580a2f2 100644 > > --- a/drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c > > +++ b/drivers/gpu/drm/i915/gem/i915_gem_execbuffer.c > > @@ -2099,7 +2099,7 @@ static int eb_submit(struct i915_execbuffer *eb) > > * The engine index is returned. > > */ > > static unsigned int > > -gen8_dispatch_bsd_engine(struct drm_i915_private *dev_priv, > > +gen8_dispatch_bsd_engine(struct drm_i915_private *i915, > > struct drm_file *file) > > { > > struct drm_i915_file_private *file_priv = file->driver_priv; > > @@ -2107,7 +2107,7 @@ gen8_dispatch_bsd_engine(struct drm_i915_private *dev_priv, > > /* Check whether the file_priv has already selected one ring. */ > > if ((int)file_priv->bsd_engine < 0) > > file_priv->bsd_engine = atomic_fetch_xor(1, > > - &dev_priv->mm.bsd_engine_dispatch_index); > > + &i915->mm.bsd_engine_dispatch_index); > > return file_priv->bsd_engine; > > } > > diff --git a/drivers/gpu/drm/i915/gem/i915_gem_object.c b/drivers/gpu/drm/i915/gem/i915_gem_object.c > > index 36b76c6a0a9d..a37f0adac16e 100644 > > --- a/drivers/gpu/drm/i915/gem/i915_gem_object.c > > +++ b/drivers/gpu/drm/i915/gem/i915_gem_object.c > > @@ -367,7 +367,7 @@ void > > i915_gem_object_flush_write_domain(struct drm_i915_gem_object *obj, > > unsigned int flush_domains) > > { > > - struct drm_i915_private *dev_priv = to_i915(obj->base.dev); > > + struct drm_i915_private *i915 = to_i915(obj->base.dev); > > struct i915_vma *vma; > > assert_object_held(obj); > > @@ -377,7 +377,7 @@ i915_gem_object_flush_write_domain(struct drm_i915_gem_object *obj, > > switch (obj->write_domain) { > > case I915_GEM_DOMAIN_GTT: > > - i915_gem_flush_ggtt_writes(dev_priv); > > + i915_gem_flush_ggtt_writes(i915); > > intel_fb_obj_flush(obj, > > fb_write_origin(obj, I915_GEM_DOMAIN_GTT)); > > diff --git a/drivers/gpu/drm/i915/gem/i915_gem_shmem.c b/drivers/gpu/drm/i915/gem/i915_gem_shmem.c > > index 19d9ecdb2894..eb4aae65f6e4 100644 > > --- a/drivers/gpu/drm/i915/gem/i915_gem_shmem.c > > +++ b/drivers/gpu/drm/i915/gem/i915_gem_shmem.c > > @@ -521,7 +521,7 @@ i915_gem_object_create_shmem(struct drm_i915_private *i915, u64 size) > > /* Allocate a new GEM object and fill it with the supplied data */ > > struct drm_i915_gem_object * > > -i915_gem_object_create_shmem_from_data(struct drm_i915_private *dev_priv, > > +i915_gem_object_create_shmem_from_data(struct drm_i915_private *i915, > > const void *data, size_t size) > > { > > struct drm_i915_gem_object *obj; > > @@ -529,7 +529,7 @@ i915_gem_object_create_shmem_from_data(struct drm_i915_private *dev_priv, > > size_t offset; > > int err; > > - obj = i915_gem_object_create_shmem(dev_priv, round_up(size, PAGE_SIZE)); > > + obj = i915_gem_object_create_shmem(i915, round_up(size, PAGE_SIZE)); > > if (IS_ERR(obj)) > > return obj; > > diff --git a/drivers/gpu/drm/i915/gem/i915_gem_stolen.c b/drivers/gpu/drm/i915/gem/i915_gem_stolen.c > > index de1fab2058ec..24e174f069f3 100644 > > --- a/drivers/gpu/drm/i915/gem/i915_gem_stolen.c > > +++ b/drivers/gpu/drm/i915/gem/i915_gem_stolen.c > > @@ -24,48 +24,48 @@ > > * for is a boon. > > */ > > -int i915_gem_stolen_insert_node_in_range(struct drm_i915_private *dev_priv, > > +int i915_gem_stolen_insert_node_in_range(struct drm_i915_private *i915, > > struct drm_mm_node *node, u64 size, > > unsigned alignment, u64 start, u64 end) > > { > > int ret; > > - if (!drm_mm_initialized(&dev_priv->mm.stolen)) > > + if (!drm_mm_initialized(&i915->mm.stolen)) > > return -ENODEV; > > /* WaSkipStolenMemoryFirstPage:bdw+ */ > > - if (INTEL_GEN(dev_priv) >= 8 && start < 4096) > > + if (INTEL_GEN(i915) >= 8 && start < 4096) > > start = 4096; > > - mutex_lock(&dev_priv->mm.stolen_lock); > > - ret = drm_mm_insert_node_in_range(&dev_priv->mm.stolen, node, > > + mutex_lock(&i915->mm.stolen_lock); > > + ret = drm_mm_insert_node_in_range(&i915->mm.stolen, node, > > size, alignment, 0, > > start, end, DRM_MM_INSERT_BEST); > > - mutex_unlock(&dev_priv->mm.stolen_lock); > > + mutex_unlock(&i915->mm.stolen_lock); > > return ret; > > } > > -int i915_gem_stolen_insert_node(struct drm_i915_private *dev_priv, > > +int i915_gem_stolen_insert_node(struct drm_i915_private *i915, > > struct drm_mm_node *node, u64 size, > > unsigned alignment) > > { > > - return i915_gem_stolen_insert_node_in_range(dev_priv, node, size, > > + return i915_gem_stolen_insert_node_in_range(i915, node, size, > > alignment, 0, U64_MAX); > > } > > -void i915_gem_stolen_remove_node(struct drm_i915_private *dev_priv, > > +void i915_gem_stolen_remove_node(struct drm_i915_private *i915, > > struct drm_mm_node *node) > > { > > - mutex_lock(&dev_priv->mm.stolen_lock); > > + mutex_lock(&i915->mm.stolen_lock); > > drm_mm_remove_node(node); > > - mutex_unlock(&dev_priv->mm.stolen_lock); > > + mutex_unlock(&i915->mm.stolen_lock); > > } > > -static int i915_adjust_stolen(struct drm_i915_private *dev_priv, > > +static int i915_adjust_stolen(struct drm_i915_private *i915, > > struct resource *dsm) > > { > > - struct i915_ggtt *ggtt = &dev_priv->ggtt; > > + struct i915_ggtt *ggtt = &i915->ggtt; > > struct resource *r; > > if (dsm->start == 0 || dsm->end <= dsm->start) > > @@ -77,14 +77,14 @@ static int i915_adjust_stolen(struct drm_i915_private *dev_priv, > > */ > > /* Make sure we don't clobber the GTT if it's within stolen memory */ > > - if (INTEL_GEN(dev_priv) <= 4 && > > - !IS_G33(dev_priv) && !IS_PINEVIEW(dev_priv) && !IS_G4X(dev_priv)) { > > + if (INTEL_GEN(i915) <= 4 && > > + !IS_G33(i915) && !IS_PINEVIEW(i915) && !IS_G4X(i915)) { > > struct resource stolen[2] = {*dsm, *dsm}; > > struct resource ggtt_res; > > resource_size_t ggtt_start; > > ggtt_start = I915_READ(PGTBL_CTL); > > - if (IS_GEN(dev_priv, 4)) > > + if (IS_GEN(i915, 4)) > > ggtt_start = (ggtt_start & PGTBL_ADDRESS_LO_MASK) | > > (ggtt_start & PGTBL_ADDRESS_HI_MASK) << 28; > > else > > @@ -118,7 +118,7 @@ static int i915_adjust_stolen(struct drm_i915_private *dev_priv, > > * kernel. So if the region is already marked as busy, something > > * is seriously wrong. > > */ > > - r = devm_request_mem_region(dev_priv->drm.dev, dsm->start, > > + r = devm_request_mem_region(i915->drm.dev, dsm->start, > > resource_size(dsm), > > "Graphics Stolen Memory"); > > if (r == NULL) { > > @@ -131,14 +131,14 @@ static int i915_adjust_stolen(struct drm_i915_private *dev_priv, > > * reservation starting from 1 instead of 0. > > * There's also BIOS with off-by-one on the other end. > > */ > > - r = devm_request_mem_region(dev_priv->drm.dev, dsm->start + 1, > > + r = devm_request_mem_region(i915->drm.dev, dsm->start + 1, > > resource_size(dsm) - 2, > > "Graphics Stolen Memory"); > > /* > > * GEN3 firmware likes to smash pci bridges into the stolen > > * range. Apparently this works. > > */ > > - if (r == NULL && !IS_GEN(dev_priv, 3)) { > > + if (r == NULL && !IS_GEN(i915, 3)) { > > DRM_ERROR("conflict detected with stolen region: %pR\n", > > dsm); > > @@ -149,25 +149,25 @@ static int i915_adjust_stolen(struct drm_i915_private *dev_priv, > > return 0; > > } > > -void i915_gem_cleanup_stolen(struct drm_i915_private *dev_priv) > > +void i915_gem_cleanup_stolen(struct drm_i915_private *i915) > > { > > - if (!drm_mm_initialized(&dev_priv->mm.stolen)) > > + if (!drm_mm_initialized(&i915->mm.stolen)) > > return; > > - drm_mm_takedown(&dev_priv->mm.stolen); > > + drm_mm_takedown(&i915->mm.stolen); > > } > > -static void g4x_get_stolen_reserved(struct drm_i915_private *dev_priv, > > +static void g4x_get_stolen_reserved(struct drm_i915_private *i915, > > resource_size_t *base, > > resource_size_t *size) > > { > > - u32 reg_val = I915_READ(IS_GM45(dev_priv) ? > > + u32 reg_val = I915_READ(IS_GM45(i915) ? > > CTG_STOLEN_RESERVED : > > ELK_STOLEN_RESERVED); > > - resource_size_t stolen_top = dev_priv->dsm.end + 1; > > + resource_size_t stolen_top = i915->dsm.end + 1; > > DRM_DEBUG_DRIVER("%s_STOLEN_RESERVED = %08x\n", > > - IS_GM45(dev_priv) ? "CTG" : "ELK", reg_val); > > + IS_GM45(i915) ? "CTG" : "ELK", reg_val); > > if ((reg_val & G4X_STOLEN_RESERVED_ENABLE) == 0) > > return; > > @@ -176,7 +176,7 @@ static void g4x_get_stolen_reserved(struct drm_i915_private *dev_priv, > > * Whether ILK really reuses the ELK register for this is unclear. > > * Let's see if we catch anyone with this supposedly enabled on ILK. > > */ > > - WARN(IS_GEN(dev_priv, 5), "ILK stolen reserved found? 0x%08x\n", > > + WARN(IS_GEN(i915, 5), "ILK stolen reserved found? 0x%08x\n", > > reg_val); > > if (!(reg_val & G4X_STOLEN_RESERVED_ADDR2_MASK)) > > @@ -188,7 +188,7 @@ static void g4x_get_stolen_reserved(struct drm_i915_private *dev_priv, > > *size = stolen_top - *base; > > } > > -static void gen6_get_stolen_reserved(struct drm_i915_private *dev_priv, > > +static void gen6_get_stolen_reserved(struct drm_i915_private *i915, > > resource_size_t *base, > > resource_size_t *size) > > { > > @@ -220,12 +220,12 @@ static void gen6_get_stolen_reserved(struct drm_i915_private *dev_priv, > > } > > } > > -static void vlv_get_stolen_reserved(struct drm_i915_private *dev_priv, > > +static void vlv_get_stolen_reserved(struct drm_i915_private *i915, > > resource_size_t *base, > > resource_size_t *size) > > { > > u32 reg_val = I915_READ(GEN6_STOLEN_RESERVED); > > - resource_size_t stolen_top = dev_priv->dsm.end + 1; > > + resource_size_t stolen_top = i915->dsm.end + 1; > > DRM_DEBUG_DRIVER("GEN6_STOLEN_RESERVED = %08x\n", reg_val); > > @@ -248,7 +248,7 @@ static void vlv_get_stolen_reserved(struct drm_i915_private *dev_priv, > > *base = stolen_top - *size; > > } > > -static void gen7_get_stolen_reserved(struct drm_i915_private *dev_priv, > > +static void gen7_get_stolen_reserved(struct drm_i915_private *i915, > > resource_size_t *base, > > resource_size_t *size) > > { > > @@ -274,7 +274,7 @@ static void gen7_get_stolen_reserved(struct drm_i915_private *dev_priv, > > } > > } > > -static void chv_get_stolen_reserved(struct drm_i915_private *dev_priv, > > +static void chv_get_stolen_reserved(struct drm_i915_private *i915, > > resource_size_t *base, > > resource_size_t *size) > > { > > @@ -306,12 +306,12 @@ static void chv_get_stolen_reserved(struct drm_i915_private *dev_priv, > > } > > } > > -static void bdw_get_stolen_reserved(struct drm_i915_private *dev_priv, > > +static void bdw_get_stolen_reserved(struct drm_i915_private *i915, > > resource_size_t *base, > > resource_size_t *size) > > { > > u32 reg_val = I915_READ(GEN6_STOLEN_RESERVED); > > - resource_size_t stolen_top = dev_priv->dsm.end + 1; > > + resource_size_t stolen_top = i915->dsm.end + 1; > > DRM_DEBUG_DRIVER("GEN6_STOLEN_RESERVED = %08x\n", reg_val); > > @@ -354,19 +354,19 @@ static void icl_get_stolen_reserved(struct drm_i915_private *i915, > > } > > } > > -int i915_gem_init_stolen(struct drm_i915_private *dev_priv) > > +int i915_gem_init_stolen(struct drm_i915_private *i915) > > { > > resource_size_t reserved_base, stolen_top; > > resource_size_t reserved_total, reserved_size; > > - mutex_init(&dev_priv->mm.stolen_lock); > > + mutex_init(&i915->mm.stolen_lock); > > - if (intel_vgpu_active(dev_priv)) { > > + if (intel_vgpu_active(i915)) { > > DRM_INFO("iGVT-g active, disabling use of stolen memory\n"); > > return 0; > > } > > - if (intel_vtd_active() && INTEL_GEN(dev_priv) < 8) { > > + if (intel_vtd_active() && INTEL_GEN(i915) < 8) { > > DRM_INFO("DMAR active, disabling use of stolen memory\n"); > > return 0; > > } > > @@ -374,55 +374,55 @@ int i915_gem_init_stolen(struct drm_i915_private *dev_priv) > > if (resource_size(&intel_graphics_stolen_res) == 0) > > return 0; > > - dev_priv->dsm = intel_graphics_stolen_res; > > + i915->dsm = intel_graphics_stolen_res; > > - if (i915_adjust_stolen(dev_priv, &dev_priv->dsm)) > > + if (i915_adjust_stolen(i915, &i915->dsm)) > > return 0; > > - GEM_BUG_ON(dev_priv->dsm.start == 0); > > - GEM_BUG_ON(dev_priv->dsm.end <= dev_priv->dsm.start); > > + GEM_BUG_ON(i915->dsm.start == 0); > > + GEM_BUG_ON(i915->dsm.end <= i915->dsm.start); > > - stolen_top = dev_priv->dsm.end + 1; > > + stolen_top = i915->dsm.end + 1; > > reserved_base = stolen_top; > > reserved_size = 0; > > - switch (INTEL_GEN(dev_priv)) { > > + switch (INTEL_GEN(i915)) { > > case 2: > > case 3: > > break; > > case 4: > > - if (!IS_G4X(dev_priv)) > > + if (!IS_G4X(i915)) > > break; > > /* fall through */ > > case 5: > > - g4x_get_stolen_reserved(dev_priv, > > + g4x_get_stolen_reserved(i915, > > &reserved_base, &reserved_size); > > break; > > case 6: > > - gen6_get_stolen_reserved(dev_priv, > > + gen6_get_stolen_reserved(i915, > > &reserved_base, &reserved_size); > > break; > > case 7: > > - if (IS_VALLEYVIEW(dev_priv)) > > - vlv_get_stolen_reserved(dev_priv, > > + if (IS_VALLEYVIEW(i915)) > > + vlv_get_stolen_reserved(i915, > > &reserved_base, &reserved_size); > > else > > - gen7_get_stolen_reserved(dev_priv, > > + gen7_get_stolen_reserved(i915, > > &reserved_base, &reserved_size); > > break; > > case 8: > > case 9: > > case 10: > > - if (IS_LP(dev_priv)) > > - chv_get_stolen_reserved(dev_priv, > > + if (IS_LP(i915)) > > + chv_get_stolen_reserved(i915, > > &reserved_base, &reserved_size); > > else > > - bdw_get_stolen_reserved(dev_priv, > > + bdw_get_stolen_reserved(i915, > > &reserved_base, &reserved_size); > > break; > > case 11: > > default: > > - icl_get_stolen_reserved(dev_priv, &reserved_base, > > + icl_get_stolen_reserved(i915, &reserved_base, > > &reserved_size); > > break; > > } > > @@ -439,12 +439,12 @@ int i915_gem_init_stolen(struct drm_i915_private *dev_priv) > > reserved_size = 0; > > } > > - dev_priv->dsm_reserved = > > + i915->dsm_reserved = > > (struct resource) DEFINE_RES_MEM(reserved_base, reserved_size); > > - if (!resource_contains(&dev_priv->dsm, &dev_priv->dsm_reserved)) { > > + if (!resource_contains(&i915->dsm, &i915->dsm_reserved)) { > > DRM_ERROR("Stolen reserved area %pR outside stolen memory %pR\n", > > - &dev_priv->dsm_reserved, &dev_priv->dsm); > > + &i915->dsm_reserved, &i915->dsm); > > return 0; > > } > > @@ -453,14 +453,14 @@ int i915_gem_init_stolen(struct drm_i915_private *dev_priv) > > reserved_total = stolen_top - reserved_base; > > DRM_DEBUG_DRIVER("Memory reserved for graphics device: %lluK, usable: %lluK\n", > > - (u64)resource_size(&dev_priv->dsm) >> 10, > > - ((u64)resource_size(&dev_priv->dsm) - reserved_total) >> 10); > > + (u64)resource_size(&i915->dsm) >> 10, > > + ((u64)resource_size(&i915->dsm) - reserved_total) >> 10); > > - dev_priv->stolen_usable_size = > > - resource_size(&dev_priv->dsm) - reserved_total; > > + i915->stolen_usable_size = > > + resource_size(&i915->dsm) - reserved_total; > > /* Basic memrange allocator for stolen space. */ > > - drm_mm_init(&dev_priv->mm.stolen, 0, dev_priv->stolen_usable_size); > > + drm_mm_init(&i915->mm.stolen, 0, i915->stolen_usable_size); > > return 0; > > } > > @@ -469,11 +469,11 @@ static struct sg_table * > > i915_pages_create_for_stolen(struct drm_device *dev, > > resource_size_t offset, resource_size_t size) > > { > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > struct sg_table *st; > > struct scatterlist *sg; > > - GEM_BUG_ON(range_overflows(offset, size, resource_size(&dev_priv->dsm))); > > + GEM_BUG_ON(range_overflows(offset, size, resource_size(&i915->dsm))); > > /* We hide that we have no struct page backing our stolen object > > * by wrapping the contiguous physical allocation with a fake > > @@ -493,7 +493,7 @@ i915_pages_create_for_stolen(struct drm_device *dev, > > sg->offset = 0; > > sg->length = size; > > - sg_dma_address(sg) = (dma_addr_t)dev_priv->dsm.start + offset; > > + sg_dma_address(sg) = (dma_addr_t)i915->dsm.start + offset; > > sg_dma_len(sg) = size; > > return st; > > @@ -524,14 +524,14 @@ static void i915_gem_object_put_pages_stolen(struct drm_i915_gem_object *obj, > > static void > > i915_gem_object_release_stolen(struct drm_i915_gem_object *obj) > > { > > - struct drm_i915_private *dev_priv = to_i915(obj->base.dev); > > + struct drm_i915_private *i915 = to_i915(obj->base.dev); > > struct drm_mm_node *stolen = fetch_and_zero(&obj->stolen); > > GEM_BUG_ON(!stolen); > > __i915_gem_object_unpin_pages(obj); > > - i915_gem_stolen_remove_node(dev_priv, stolen); > > + i915_gem_stolen_remove_node(i915, stolen); > > kfree(stolen); > > } > > @@ -542,7 +542,7 @@ static const struct drm_i915_gem_object_ops i915_gem_object_stolen_ops = { > > }; > > static struct drm_i915_gem_object * > > -_i915_gem_object_create_stolen(struct drm_i915_private *dev_priv, > > +_i915_gem_object_create_stolen(struct drm_i915_private *i915, > > struct drm_mm_node *stolen) > > { > > struct drm_i915_gem_object *obj; > > @@ -552,12 +552,12 @@ _i915_gem_object_create_stolen(struct drm_i915_private *dev_priv, > > if (obj == NULL) > > return NULL; > > - drm_gem_private_object_init(&dev_priv->drm, &obj->base, stolen->size); > > + drm_gem_private_object_init(&i915->drm, &obj->base, stolen->size); > > i915_gem_object_init(obj, &i915_gem_object_stolen_ops); > > obj->stolen = stolen; > > obj->read_domains = I915_GEM_DOMAIN_CPU | I915_GEM_DOMAIN_GTT; > > - cache_level = HAS_LLC(dev_priv) ? I915_CACHE_LLC : I915_CACHE_NONE; > > + cache_level = HAS_LLC(i915) ? I915_CACHE_LLC : I915_CACHE_NONE; > > i915_gem_object_set_cache_coherency(obj, cache_level); > > if (i915_gem_object_pin_pages(obj)) > > @@ -571,14 +571,14 @@ _i915_gem_object_create_stolen(struct drm_i915_private *dev_priv, > > } > > struct drm_i915_gem_object * > > -i915_gem_object_create_stolen(struct drm_i915_private *dev_priv, > > +i915_gem_object_create_stolen(struct drm_i915_private *i915, > > resource_size_t size) > > { > > struct drm_i915_gem_object *obj; > > struct drm_mm_node *stolen; > > int ret; > > - if (!drm_mm_initialized(&dev_priv->mm.stolen)) > > + if (!drm_mm_initialized(&i915->mm.stolen)) > > return NULL; > > if (size == 0) > > @@ -588,37 +588,37 @@ i915_gem_object_create_stolen(struct drm_i915_private *dev_priv, > > if (!stolen) > > return NULL; > > - ret = i915_gem_stolen_insert_node(dev_priv, stolen, size, 4096); > > + ret = i915_gem_stolen_insert_node(i915, stolen, size, 4096); > > if (ret) { > > kfree(stolen); > > return NULL; > > } > > - obj = _i915_gem_object_create_stolen(dev_priv, stolen); > > + obj = _i915_gem_object_create_stolen(i915, stolen); > > if (obj) > > return obj; > > - i915_gem_stolen_remove_node(dev_priv, stolen); > > + i915_gem_stolen_remove_node(i915, stolen); > > kfree(stolen); > > return NULL; > > } > > struct drm_i915_gem_object * > > -i915_gem_object_create_stolen_for_preallocated(struct drm_i915_private *dev_priv, > > +i915_gem_object_create_stolen_for_preallocated(struct drm_i915_private *i915, > > resource_size_t stolen_offset, > > resource_size_t gtt_offset, > > resource_size_t size) > > { > > - struct i915_ggtt *ggtt = &dev_priv->ggtt; > > + struct i915_ggtt *ggtt = &i915->ggtt; > > struct drm_i915_gem_object *obj; > > struct drm_mm_node *stolen; > > struct i915_vma *vma; > > int ret; > > - if (!drm_mm_initialized(&dev_priv->mm.stolen)) > > + if (!drm_mm_initialized(&i915->mm.stolen)) > > return NULL; > > - lockdep_assert_held(&dev_priv->drm.struct_mutex); > > + lockdep_assert_held(&i915->drm.struct_mutex); > > DRM_DEBUG_DRIVER("creating preallocated stolen object: stolen_offset=%pa, gtt_offset=%pa, size=%pa\n", > > &stolen_offset, >t_offset, &size); > > @@ -635,19 +635,19 @@ i915_gem_object_create_stolen_for_preallocated(struct drm_i915_private *dev_priv > > stolen->start = stolen_offset; > > stolen->size = size; > > - mutex_lock(&dev_priv->mm.stolen_lock); > > - ret = drm_mm_reserve_node(&dev_priv->mm.stolen, stolen); > > - mutex_unlock(&dev_priv->mm.stolen_lock); > > + mutex_lock(&i915->mm.stolen_lock); > > + ret = drm_mm_reserve_node(&i915->mm.stolen, stolen); > > + mutex_unlock(&i915->mm.stolen_lock); > > if (ret) { > > DRM_DEBUG_DRIVER("failed to allocate stolen space\n"); > > kfree(stolen); > > return NULL; > > } > > - obj = _i915_gem_object_create_stolen(dev_priv, stolen); > > + obj = _i915_gem_object_create_stolen(i915, stolen); > > if (obj == NULL) { > > DRM_DEBUG_DRIVER("failed to allocate stolen object\n"); > > - i915_gem_stolen_remove_node(dev_priv, stolen); > > + i915_gem_stolen_remove_node(i915, stolen); > > kfree(stolen); > > return NULL; > > } > > diff --git a/drivers/gpu/drm/i915/gem/i915_gem_tiling.c b/drivers/gpu/drm/i915/gem/i915_gem_tiling.c > > index ca0c2f451742..46b523d4ed84 100644 > > --- a/drivers/gpu/drm/i915/gem/i915_gem_tiling.c > > +++ b/drivers/gpu/drm/i915/gem/i915_gem_tiling.c > > @@ -398,7 +398,7 @@ i915_gem_get_tiling_ioctl(struct drm_device *dev, void *data, > > struct drm_file *file) > > { > > struct drm_i915_gem_get_tiling *args = data; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > struct drm_i915_gem_object *obj; > > int err = -ENOENT; > > @@ -415,10 +415,10 @@ i915_gem_get_tiling_ioctl(struct drm_device *dev, void *data, > > switch (args->tiling_mode) { > > case I915_TILING_X: > > - args->swizzle_mode = dev_priv->mm.bit_6_swizzle_x; > > + args->swizzle_mode = i915->mm.bit_6_swizzle_x; > > break; > > case I915_TILING_Y: > > - args->swizzle_mode = dev_priv->mm.bit_6_swizzle_y; > > + args->swizzle_mode = i915->mm.bit_6_swizzle_y; > > break; > > default: > > case I915_TILING_NONE: > > @@ -427,7 +427,7 @@ i915_gem_get_tiling_ioctl(struct drm_device *dev, void *data, > > } > > /* Hide bit 17 from the user -- see comment in i915_gem_set_tiling */ > > - if (dev_priv->quirks & QUIRK_PIN_SWIZZLED_PAGES) > > + if (i915->quirks & QUIRK_PIN_SWIZZLED_PAGES) > > args->phys_swizzle_mode = I915_BIT_6_SWIZZLE_UNKNOWN; > > else > > args->phys_swizzle_mode = args->swizzle_mode; > > diff --git a/drivers/gpu/drm/i915/gem/i915_gem_userptr.c b/drivers/gpu/drm/i915/gem/i915_gem_userptr.c > > index 528b61678334..f22e10e8895a 100644 > > --- a/drivers/gpu/drm/i915/gem/i915_gem_userptr.c > > +++ b/drivers/gpu/drm/i915/gem/i915_gem_userptr.c > > @@ -327,12 +327,12 @@ i915_mmu_notifier_free(struct i915_mmu_notifier *mn, > > #endif > > static struct i915_mm_struct * > > -__i915_mm_struct_find(struct drm_i915_private *dev_priv, struct mm_struct *real) > > +__i915_mm_struct_find(struct drm_i915_private *i915, struct mm_struct *real) > > { > > struct i915_mm_struct *mm; > > - /* Protected by dev_priv->mm_lock */ > > - hash_for_each_possible(dev_priv->mm_structs, mm, node, (unsigned long)real) > > + /* Protected by i915->mm_lock */ > > + hash_for_each_possible(i915->mm_structs, mm, node, (unsigned long)real) > > if (mm->mm == real) > > return mm; > > @@ -342,7 +342,7 @@ __i915_mm_struct_find(struct drm_i915_private *dev_priv, struct mm_struct *real) > > static int > > i915_gem_userptr_init__mm_struct(struct drm_i915_gem_object *obj) > > { > > - struct drm_i915_private *dev_priv = to_i915(obj->base.dev); > > + struct drm_i915_private *i915 = to_i915(obj->base.dev); > > struct i915_mm_struct *mm; > > int ret = 0; > > @@ -356,8 +356,8 @@ i915_gem_userptr_init__mm_struct(struct drm_i915_gem_object *obj) > > * struct_mutex, i.e. we need to schedule a worker to do the clean > > * up. > > */ > > - mutex_lock(&dev_priv->mm_lock); > > - mm = __i915_mm_struct_find(dev_priv, current->mm); > > + mutex_lock(&i915->mm_lock); > > + mm = __i915_mm_struct_find(i915, current->mm); > > if (mm == NULL) { > > mm = kmalloc(sizeof(*mm), GFP_KERNEL); > > if (mm == NULL) { > > @@ -373,15 +373,15 @@ i915_gem_userptr_init__mm_struct(struct drm_i915_gem_object *obj) > > mm->mn = NULL; > > - /* Protected by dev_priv->mm_lock */ > > - hash_add(dev_priv->mm_structs, > > + /* Protected by i915->mm_lock */ > > + hash_add(i915->mm_structs, > > &mm->node, (unsigned long)mm->mm); > > } else > > kref_get(&mm->kref); > > obj->userptr.mm = mm; > > out: > > - mutex_unlock(&dev_priv->mm_lock); > > + mutex_unlock(&i915->mm_lock); > > return ret; > > } > > @@ -399,7 +399,7 @@ __i915_mm_struct_free(struct kref *kref) > > { > > struct i915_mm_struct *mm = container_of(kref, typeof(*mm), kref); > > - /* Protected by dev_priv->mm_lock */ > > + /* Protected by i915->mm_lock */ > > hash_del(&mm->node); > > mutex_unlock(&mm->i915->mm_lock); > > @@ -741,13 +741,13 @@ i915_gem_userptr_ioctl(struct drm_device *dev, > > void *data, > > struct drm_file *file) > > { > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > struct drm_i915_gem_userptr *args = data; > > struct drm_i915_gem_object *obj; > > int ret; > > u32 handle; > > - if (!HAS_LLC(dev_priv) && !HAS_SNOOP(dev_priv)) { > > + if (!HAS_LLC(i915) && !HAS_SNOOP(i915)) { > > /* We cannot support coherent userptr objects on hw without > > * LLC and broken snooping. > > */ > > @@ -774,7 +774,7 @@ i915_gem_userptr_ioctl(struct drm_device *dev, > > * On almost all of the older hw, we cannot tell the GPU that > > * a page is readonly. > > */ > > - vm = dev_priv->kernel_context->vm; > > + vm = i915->kernel_context->vm; > > if (!vm || !vm->has_read_only) > > return -ENODEV; > > } > > @@ -812,22 +812,22 @@ i915_gem_userptr_ioctl(struct drm_device *dev, > > return 0; > > } > > -int i915_gem_init_userptr(struct drm_i915_private *dev_priv) > > +int i915_gem_init_userptr(struct drm_i915_private *i915) > > { > > - mutex_init(&dev_priv->mm_lock); > > - hash_init(dev_priv->mm_structs); > > + mutex_init(&i915->mm_lock); > > + hash_init(i915->mm_structs); > > - dev_priv->mm.userptr_wq = > > + i915->mm.userptr_wq = > > alloc_workqueue("i915-userptr-acquire", > > WQ_HIGHPRI | WQ_UNBOUND, > > 0); > > - if (!dev_priv->mm.userptr_wq) > > + if (!i915->mm.userptr_wq) > > return -ENOMEM; > > return 0; > > } > > -void i915_gem_cleanup_userptr(struct drm_i915_private *dev_priv) > > +void i915_gem_cleanup_userptr(struct drm_i915_private *i915) > > { > > - destroy_workqueue(dev_priv->mm.userptr_wq); > > + destroy_workqueue(i915->mm.userptr_wq); > > } > > diff --git a/drivers/gpu/drm/i915/gem/selftests/huge_pages.c b/drivers/gpu/drm/i915/gem/selftests/huge_pages.c > > index 73e667b31cc4..78797699feb0 100644 > > --- a/drivers/gpu/drm/i915/gem/selftests/huge_pages.c > > +++ b/drivers/gpu/drm/i915/gem/selftests/huge_pages.c > > @@ -1415,8 +1415,8 @@ static int igt_ppgtt_gemfs_huge(void *arg) > > static int igt_ppgtt_pin_update(void *arg) > > { > > struct i915_gem_context *ctx = arg; > > - struct drm_i915_private *dev_priv = ctx->i915; > > - unsigned long supported = INTEL_INFO(dev_priv)->page_sizes; > > + struct drm_i915_private *i915 = ctx->i915; > > + unsigned long supported = INTEL_INFO(i915)->page_sizes; > > struct i915_address_space *vm = ctx->vm; > > struct drm_i915_gem_object *obj; > > struct i915_vma *vma; > > @@ -1443,7 +1443,7 @@ static int igt_ppgtt_pin_update(void *arg) > > for_each_set_bit_from(first, &supported, last + 1) { > > unsigned int page_size = BIT(first); > > - obj = i915_gem_object_create_internal(dev_priv, page_size); > > + obj = i915_gem_object_create_internal(i915, page_size); > > if (IS_ERR(obj)) > > return PTR_ERR(obj); > > @@ -1497,7 +1497,7 @@ static int igt_ppgtt_pin_update(void *arg) > > i915_gem_object_put(obj); > > } > > - obj = i915_gem_object_create_internal(dev_priv, PAGE_SIZE); > > + obj = i915_gem_object_create_internal(i915, PAGE_SIZE); > > if (IS_ERR(obj)) > > return PTR_ERR(obj); > > @@ -1518,7 +1518,7 @@ static int igt_ppgtt_pin_update(void *arg) > > * land in the now stale 2M page. > > */ > > - err = gpu_write(vma, ctx, dev_priv->engine[RCS0], 0, 0xdeadbeaf); > > + err = gpu_write(vma, ctx, i915->engine[RCS0], 0, 0xdeadbeaf); > > if (err) > > goto out_unpin; > > @@ -1682,20 +1682,20 @@ int i915_gem_huge_page_mock_selftests(void) > > SUBTEST(igt_mock_ppgtt_huge_fill), > > SUBTEST(igt_mock_ppgtt_64K), > > }; > > - struct drm_i915_private *dev_priv; > > + struct drm_i915_private *i915; > > struct i915_ppgtt *ppgtt; > > int err; > > - dev_priv = mock_gem_device(); > > - if (!dev_priv) > > + i915 = mock_gem_device(); > > + if (!i915) > > return -ENOMEM; > > /* Pretend to be a device which supports the 48b PPGTT */ > > - mkwrite_device_info(dev_priv)->ppgtt_type = INTEL_PPGTT_FULL; > > - mkwrite_device_info(dev_priv)->ppgtt_size = 48; > > + mkwrite_device_info(i915)->ppgtt_type = INTEL_PPGTT_FULL; > > + mkwrite_device_info(i915)->ppgtt_size = 48; > > - mutex_lock(&dev_priv->drm.struct_mutex); > > - ppgtt = i915_ppgtt_create(dev_priv); > > + mutex_lock(&i915->drm.struct_mutex); > > + ppgtt = i915_ppgtt_create(i915); > > if (IS_ERR(ppgtt)) { > > err = PTR_ERR(ppgtt); > > goto out_unlock; > > @@ -1720,13 +1720,13 @@ int i915_gem_huge_page_mock_selftests(void) > > i915_vm_put(&ppgtt->vm); > > out_unlock: > > - mutex_unlock(&dev_priv->drm.struct_mutex); > > - drm_dev_put(&dev_priv->drm); > > + mutex_unlock(&i915->drm.struct_mutex); > > + drm_dev_put(&i915->drm); > > return err; > > } > > -int i915_gem_huge_page_live_selftests(struct drm_i915_private *dev_priv) > > +int i915_gem_huge_page_live_selftests(struct drm_i915_private *i915) > > { > > static const struct i915_subtest tests[] = { > > SUBTEST(igt_shrink_thp), > > @@ -1741,22 +1741,22 @@ int i915_gem_huge_page_live_selftests(struct drm_i915_private *dev_priv) > > intel_wakeref_t wakeref; > > int err; > > - if (!HAS_PPGTT(dev_priv)) { > > + if (!HAS_PPGTT(i915)) { > > pr_info("PPGTT not supported, skipping live-selftests\n"); > > return 0; > > } > > - if (i915_terminally_wedged(dev_priv)) > > + if (i915_terminally_wedged(i915)) > > return 0; > > - file = mock_file(dev_priv); > > + file = mock_file(i915); > > if (IS_ERR(file)) > > return PTR_ERR(file); > > - mutex_lock(&dev_priv->drm.struct_mutex); > > - wakeref = intel_runtime_pm_get(dev_priv); > > + mutex_lock(&i915->drm.struct_mutex); > > + wakeref = intel_runtime_pm_get(i915); > > - ctx = live_context(dev_priv, file); > > + ctx = live_context(i915, file); > > if (IS_ERR(ctx)) { > > err = PTR_ERR(ctx); > > goto out_unlock; > > @@ -1768,10 +1768,10 @@ int i915_gem_huge_page_live_selftests(struct drm_i915_private *dev_priv) > > err = i915_subtests(tests, ctx); > > out_unlock: > > - intel_runtime_pm_put(dev_priv, wakeref); > > - mutex_unlock(&dev_priv->drm.struct_mutex); > > + intel_runtime_pm_put(i915, wakeref); > > + mutex_unlock(&i915->drm.struct_mutex); > > - mock_file_free(dev_priv, file); > > + mock_file_free(i915, file); > > return err; > > } > > diff --git a/drivers/gpu/drm/i915/gem/selftests/i915_gem_context.c b/drivers/gpu/drm/i915/gem/selftests/i915_gem_context.c > > index 74b0e5871c4b..2fdcaa1667bf 100644 > > --- a/drivers/gpu/drm/i915/gem/selftests/i915_gem_context.c > > +++ b/drivers/gpu/drm/i915/gem/selftests/i915_gem_context.c > > @@ -1748,7 +1748,7 @@ int i915_gem_context_mock_selftests(void) > > return err; > > } > > -int i915_gem_context_live_selftests(struct drm_i915_private *dev_priv) > > +int i915_gem_context_live_selftests(struct drm_i915_private *i915) > > { > > static const struct i915_subtest tests[] = { > > SUBTEST(live_nop_switch), > > @@ -1759,8 +1759,8 @@ int i915_gem_context_live_selftests(struct drm_i915_private *dev_priv) > > SUBTEST(igt_vm_isolation), > > }; > > - if (i915_terminally_wedged(dev_priv)) > > + if (i915_terminally_wedged(i915)) > > return 0; > > - return i915_subtests(tests, dev_priv); > > + return i915_subtests(tests, i915); > > } > > diff --git a/drivers/gpu/drm/i915/gt/intel_engine.h b/drivers/gpu/drm/i915/gt/intel_engine.h > > index 1439fa4093ac..74f9ad6648eb 100644 > > --- a/drivers/gpu/drm/i915/gt/intel_engine.h > > +++ b/drivers/gpu/drm/i915/gt/intel_engine.h > > @@ -465,7 +465,7 @@ static inline void intel_engine_reset(struct intel_engine_cs *engine, > > } > > bool intel_engine_is_idle(struct intel_engine_cs *engine); > > -bool intel_engines_are_idle(struct drm_i915_private *dev_priv); > > +bool intel_engines_are_idle(struct drm_i915_private *i915); > > void intel_engine_lost_context(struct intel_engine_cs *engine); > > diff --git a/drivers/gpu/drm/i915/gt/intel_engine_cs.c b/drivers/gpu/drm/i915/gt/intel_engine_cs.c > > index c0d986db5a75..a048b8743ce6 100644 > > --- a/drivers/gpu/drm/i915/gt/intel_engine_cs.c > > +++ b/drivers/gpu/drm/i915/gt/intel_engine_cs.c > > @@ -160,7 +160,7 @@ static const struct engine_info intel_engines[] = { > > /** > > * intel_engine_context_size() - return the size of the context for an engine > > - * @dev_priv: i915 device private > > + * @i915: i915 device private > > * @class: engine class > > * > > * Each engine class may require a different amount of space for a context > > @@ -172,7 +172,7 @@ static const struct engine_info intel_engines[] = { > > * in LRC mode, but does not include the "shared data page" used with > > * GuC submission. The caller should account for this if using the GuC. > > */ > > -u32 intel_engine_context_size(struct drm_i915_private *dev_priv, u8 class) > > +u32 intel_engine_context_size(struct drm_i915_private *i915, u8 class) > > { > > u32 cxt_size; > > @@ -180,9 +180,9 @@ u32 intel_engine_context_size(struct drm_i915_private *dev_priv, u8 class) > > switch (class) { > > case RENDER_CLASS: > > - switch (INTEL_GEN(dev_priv)) { > > + switch (INTEL_GEN(i915)) { > > default: > > - MISSING_CASE(INTEL_GEN(dev_priv)); > > + MISSING_CASE(INTEL_GEN(i915)); > > return DEFAULT_LR_CONTEXT_RENDER_SIZE; > > case 11: > > return GEN11_LR_CONTEXT_RENDER_SIZE; > > @@ -193,7 +193,7 @@ u32 intel_engine_context_size(struct drm_i915_private *dev_priv, u8 class) > > case 8: > > return GEN8_LR_CONTEXT_RENDER_SIZE; > > case 7: > > - if (IS_HASWELL(dev_priv)) > > + if (IS_HASWELL(i915)) > > return HSW_CXT_TOTAL_SIZE; > > cxt_size = I915_READ(GEN7_CXT_SIZE); > > @@ -217,7 +217,7 @@ u32 intel_engine_context_size(struct drm_i915_private *dev_priv, u8 class) > > */ > > cxt_size = I915_READ(CXT_SIZE) + 1; > > DRM_DEBUG_DRIVER("gen%d CXT_SIZE = %d bytes [0x%08x]\n", > > - INTEL_GEN(dev_priv), > > + INTEL_GEN(i915), > > cxt_size * 64, > > cxt_size - 1); > > return round_up(cxt_size * 64, PAGE_SIZE); > > @@ -234,7 +234,7 @@ u32 intel_engine_context_size(struct drm_i915_private *dev_priv, u8 class) > > case VIDEO_DECODE_CLASS: > > case VIDEO_ENHANCEMENT_CLASS: > > case COPY_ENGINE_CLASS: > > - if (INTEL_GEN(dev_priv) < 8) > > + if (INTEL_GEN(i915) < 8) > > return 0; > > return GEN8_LR_CONTEXT_OTHER_SIZE; > > } > > @@ -284,7 +284,7 @@ static void intel_engine_sanitize_mmio(struct intel_engine_cs *engine) > > } > > static int > > -intel_engine_setup(struct drm_i915_private *dev_priv, > > +intel_engine_setup(struct drm_i915_private *i915, > > enum intel_engine_id id) > > { > > const struct engine_info *info = &intel_engines[id]; > > @@ -301,10 +301,10 @@ intel_engine_setup(struct drm_i915_private *dev_priv, > > if (GEM_DEBUG_WARN_ON(info->instance > MAX_ENGINE_INSTANCE)) > > return -EINVAL; > > - if (GEM_DEBUG_WARN_ON(dev_priv->engine_class[info->class][info->instance])) > > + if (GEM_DEBUG_WARN_ON(i915->engine_class[info->class][info->instance])) > > return -EINVAL; > > - GEM_BUG_ON(dev_priv->engine[id]); > > + GEM_BUG_ON(i915->engine[id]); > > engine = kzalloc(sizeof(*engine), GFP_KERNEL); > > if (!engine) > > return -ENOMEM; > > @@ -313,11 +313,11 @@ intel_engine_setup(struct drm_i915_private *dev_priv, > > engine->id = id; > > engine->mask = BIT(id); > > - engine->i915 = dev_priv; > > - engine->uncore = &dev_priv->uncore; > > + engine->i915 = i915; > > + engine->uncore = &i915->uncore; > > __sprint_engine_name(engine->name, info); > > engine->hw_id = engine->guc_id = info->hw_id; > > - engine->mmio_base = __engine_mmio_base(dev_priv, info->mmio_bases); > > + engine->mmio_base = __engine_mmio_base(i915, info->mmio_bases); > > engine->class = info->class; > > engine->instance = info->instance; > > @@ -329,12 +329,12 @@ intel_engine_setup(struct drm_i915_private *dev_priv, > > engine->uabi_class = intel_engine_classes[info->class].uabi_class; > > - engine->context_size = intel_engine_context_size(dev_priv, > > + engine->context_size = intel_engine_context_size(i915, > > engine->class); > > if (WARN_ON(engine->context_size > BIT(20))) > > engine->context_size = 0; > > if (engine->context_size) > > - DRIVER_CAPS(dev_priv)->has_logical_contexts = true; > > + DRIVER_CAPS(i915)->has_logical_contexts = true; > > /* Nothing to do here, execute in order of dependencies */ > > engine->schedule = NULL; > > @@ -346,8 +346,8 @@ intel_engine_setup(struct drm_i915_private *dev_priv, > > /* Scrub mmio state on takeover */ > > intel_engine_sanitize_mmio(engine); > > - dev_priv->engine_class[info->class][info->instance] = engine; > > - dev_priv->engine[id] = engine; > > + i915->engine_class[info->class][info->instance] = engine; > > + i915->engine[id] = engine; > > return 0; > > } > > @@ -953,17 +953,17 @@ const char *i915_cache_level_str(struct drm_i915_private *i915, int type) > > } > > } > > -u32 intel_calculate_mcr_s_ss_select(struct drm_i915_private *dev_priv) > > +u32 intel_calculate_mcr_s_ss_select(struct drm_i915_private *i915) > > { > > - const struct sseu_dev_info *sseu = &RUNTIME_INFO(dev_priv)->sseu; > > + const struct sseu_dev_info *sseu = &RUNTIME_INFO(i915)->sseu; > > u32 mcr_s_ss_select; > > u32 slice = fls(sseu->slice_mask); > > u32 subslice = fls(sseu->subslice_mask[slice]); > > - if (IS_GEN(dev_priv, 10)) > > + if (IS_GEN(i915, 10)) > > mcr_s_ss_select = GEN8_MCR_SLICE(slice) | > > GEN8_MCR_SUBSLICE(subslice); > > - else if (INTEL_GEN(dev_priv) >= 11) > > + else if (INTEL_GEN(i915) >= 11) > > mcr_s_ss_select = GEN11_MCR_SLICE(slice) | > > GEN11_MCR_SUBSLICE(subslice); > > else > > @@ -1095,7 +1095,7 @@ void intel_engine_get_instdone(struct intel_engine_cs *engine, > > static bool ring_is_idle(struct intel_engine_cs *engine) > > { > > - struct drm_i915_private *dev_priv = engine->i915; > > + struct drm_i915_private *i915 = engine->i915; > > intel_wakeref_t wakeref; > > bool idle = true; > > @@ -1103,7 +1103,7 @@ static bool ring_is_idle(struct intel_engine_cs *engine) > > return true; > > /* If the whole device is asleep, the engine must be idle */ > > - wakeref = intel_runtime_pm_get_if_in_use(dev_priv); > > + wakeref = intel_runtime_pm_get_if_in_use(i915); > > if (!wakeref) > > return true; > > @@ -1113,11 +1113,11 @@ static bool ring_is_idle(struct intel_engine_cs *engine) > > idle = false; > > /* No bit for gen2, so assume the CS parser is idle */ > > - if (INTEL_GEN(dev_priv) > 2 && > > + if (INTEL_GEN(i915) > 2 && > > !(ENGINE_READ(engine, RING_MI_MODE) & MODE_IDLE)) > > idle = false; > > - intel_runtime_pm_put(dev_priv, wakeref); > > + intel_runtime_pm_put(i915, wakeref); > > return idle; > > } > > @@ -1321,12 +1321,12 @@ static void hexdump(struct drm_printer *m, const void *buf, size_t len) > > static void intel_engine_print_registers(const struct intel_engine_cs *engine, > > struct drm_printer *m) > > { > > - struct drm_i915_private *dev_priv = engine->i915; > > + struct drm_i915_private *i915 = engine->i915; > > const struct intel_engine_execlists * const execlists = > > &engine->execlists; > > u64 addr; > > - if (engine->id == RCS0 && IS_GEN_RANGE(dev_priv, 4, 7)) > > + if (engine->id == RCS0 && IS_GEN_RANGE(i915, 4, 7)) > > drm_printf(m, "\tCCID: 0x%08x\n", ENGINE_READ(engine, CCID)); > > drm_printf(m, "\tRING_START: 0x%08x\n", > > ENGINE_READ(engine, RING_START)); > > @@ -1343,7 +1343,7 @@ static void intel_engine_print_registers(const struct intel_engine_cs *engine, > > ENGINE_READ(engine, RING_MI_MODE) & (MODE_IDLE) ? " [idle]" : ""); > > } > > - if (INTEL_GEN(dev_priv) >= 6) { > > + if (INTEL_GEN(i915) >= 6) { > > drm_printf(m, "\tRING_IMR: %08x\n", > > ENGINE_READ(engine, RING_IMR)); > > } > > @@ -1354,15 +1354,15 @@ static void intel_engine_print_registers(const struct intel_engine_cs *engine, > > addr = intel_engine_get_last_batch_head(engine); > > drm_printf(m, "\tBBADDR: 0x%08x_%08x\n", > > upper_32_bits(addr), lower_32_bits(addr)); > > - if (INTEL_GEN(dev_priv) >= 8) > > + if (INTEL_GEN(i915) >= 8) > > addr = ENGINE_READ64(engine, RING_DMA_FADD, RING_DMA_FADD_UDW); > > - else if (INTEL_GEN(dev_priv) >= 4) > > + else if (INTEL_GEN(i915) >= 4) > > addr = ENGINE_READ(engine, RING_DMA_FADD); > > else > > addr = ENGINE_READ(engine, DMA_FADD_I8XX); > > drm_printf(m, "\tDMA_FADDR: 0x%08x_%08x\n", > > upper_32_bits(addr), lower_32_bits(addr)); > > - if (INTEL_GEN(dev_priv) >= 4) { > > + if (INTEL_GEN(i915) >= 4) { > > drm_printf(m, "\tIPEIR: 0x%08x\n", > > ENGINE_READ(engine, RING_IPEIR)); > > drm_printf(m, "\tIPEHR: 0x%08x\n", > > @@ -1372,7 +1372,7 @@ static void intel_engine_print_registers(const struct intel_engine_cs *engine, > > drm_printf(m, "\tIPEHR: 0x%08x\n", ENGINE_READ(engine, IPEHR)); > > } > > - if (HAS_EXECLISTS(dev_priv)) { > > + if (HAS_EXECLISTS(i915)) { > > const u32 *hws = > > &engine->status_page.addr[I915_HWS_CSB_BUF0_INDEX]; > > const u8 num_entries = execlists->csb_size; > > @@ -1426,7 +1426,7 @@ static void intel_engine_print_registers(const struct intel_engine_cs *engine, > > } > > drm_printf(m, "\t\tHW active? 0x%x\n", execlists->active); > > rcu_read_unlock(); > > - } else if (INTEL_GEN(dev_priv) > 6) { > > + } else if (INTEL_GEN(i915) > 6) { > > drm_printf(m, "\tPP_DIR_BASE: 0x%08x\n", > > ENGINE_READ(engine, RING_PP_DIR_BASE)); > > drm_printf(m, "\tPP_DIR_BASE_READ: 0x%08x\n", > > diff --git a/drivers/gpu/drm/i915/gt/intel_hangcheck.c b/drivers/gpu/drm/i915/gt/intel_hangcheck.c > > index 174bb0a60309..948423c5f67d 100644 > > --- a/drivers/gpu/drm/i915/gt/intel_hangcheck.c > > +++ b/drivers/gpu/drm/i915/gt/intel_hangcheck.c > > @@ -50,7 +50,7 @@ static bool instdone_unchanged(u32 current_instdone, u32 *old_instdone) > > static bool subunits_stuck(struct intel_engine_cs *engine) > > { > > - struct drm_i915_private *dev_priv = engine->i915; > > + struct drm_i915_private *i915 = engine->i915; > > struct intel_instdone instdone; > > struct intel_instdone *accu_instdone = &engine->hangcheck.instdone; > > bool stuck; > > @@ -72,7 +72,7 @@ static bool subunits_stuck(struct intel_engine_cs *engine) > > stuck &= instdone_unchanged(instdone.slice_common, > > &accu_instdone->slice_common); > > - for_each_instdone_slice_subslice(dev_priv, slice, subslice) { > > + for_each_instdone_slice_subslice(i915, slice, subslice) { > > stuck &= instdone_unchanged(instdone.sampler[slice][subslice], > > &accu_instdone->sampler[slice][subslice]); > > stuck &= instdone_unchanged(instdone.row[slice][subslice], > > @@ -103,7 +103,7 @@ head_stuck(struct intel_engine_cs *engine, u64 acthd) > > static enum intel_engine_hangcheck_action > > engine_stuck(struct intel_engine_cs *engine, u64 acthd) > > { > > - struct drm_i915_private *dev_priv = engine->i915; > > + struct drm_i915_private *i915 = engine->i915; > > enum intel_engine_hangcheck_action ha; > > u32 tmp; > > @@ -111,7 +111,7 @@ engine_stuck(struct intel_engine_cs *engine, u64 acthd) > > if (ha != ENGINE_DEAD) > > return ha; > > - if (IS_GEN(dev_priv, 2)) > > + if (IS_GEN(i915, 2)) > > return ENGINE_DEAD; > > /* Is the chip hanging on a WAIT_FOR_EVENT? > > @@ -121,7 +121,7 @@ engine_stuck(struct intel_engine_cs *engine, u64 acthd) > > */ > > tmp = ENGINE_READ(engine, RING_CTL); > > if (tmp & RING_WAIT) { > > - i915_handle_error(dev_priv, engine->mask, 0, > > + i915_handle_error(i915, engine->mask, 0, > > "stuck wait on %s", engine->name); > > ENGINE_WRITE(engine, RING_CTL, tmp); > > return ENGINE_WAIT_KICK; > > @@ -256,8 +256,8 @@ static void hangcheck_declare_hang(struct drm_i915_private *i915, > > */ > > static void i915_hangcheck_elapsed(struct work_struct *work) > > { > > - struct drm_i915_private *dev_priv = > > - container_of(work, typeof(*dev_priv), > > + struct drm_i915_private *i915 = > > + container_of(work, typeof(*i915), > > gpu_error.hangcheck_work.work); > > intel_engine_mask_t hung = 0, stuck = 0, wedged = 0; > > struct intel_engine_cs *engine; > > @@ -267,13 +267,13 @@ static void i915_hangcheck_elapsed(struct work_struct *work) > > if (!i915_modparams.enable_hangcheck) > > return; > > - if (!READ_ONCE(dev_priv->gt.awake)) > > + if (!READ_ONCE(i915->gt.awake)) > > return; > > - if (i915_terminally_wedged(dev_priv)) > > + if (i915_terminally_wedged(i915)) > > return; > > - wakeref = intel_runtime_pm_get_if_in_use(dev_priv); > > + wakeref = intel_runtime_pm_get_if_in_use(i915); > > if (!wakeref) > > return; > > @@ -281,9 +281,9 @@ static void i915_hangcheck_elapsed(struct work_struct *work) > > * periodically arm the mmio checker to see if we are triggering > > * any invalid access. > > */ > > - intel_uncore_arm_unclaimed_mmio_detection(&dev_priv->uncore); > > + intel_uncore_arm_unclaimed_mmio_detection(&i915->uncore); > > - for_each_engine(engine, dev_priv, id) { > > + for_each_engine(engine, i915, id) { > > struct hangcheck hc; > > intel_engine_signal_breadcrumbs(engine); > > @@ -305,7 +305,7 @@ static void i915_hangcheck_elapsed(struct work_struct *work) > > if (GEM_SHOW_DEBUG() && (hung | stuck)) { > > struct drm_printer p = drm_debug_printer("hangcheck"); > > - for_each_engine(engine, dev_priv, id) { > > + for_each_engine(engine, i915, id) { > > if (intel_engine_is_idle(engine)) > > continue; > > @@ -314,20 +314,20 @@ static void i915_hangcheck_elapsed(struct work_struct *work) > > } > > if (wedged) { > > - dev_err(dev_priv->drm.dev, > > + dev_err(i915->drm.dev, > > "GPU recovery timed out," > > " cancelling all in-flight rendering.\n"); > > GEM_TRACE_DUMP(); > > - i915_gem_set_wedged(dev_priv); > > + i915_gem_set_wedged(i915); > > } > > if (hung) > > - hangcheck_declare_hang(dev_priv, hung, stuck); > > + hangcheck_declare_hang(i915, hung, stuck); > > - intel_runtime_pm_put(dev_priv, wakeref); > > + intel_runtime_pm_put(i915, wakeref); > > /* Reset timer in case GPU hangs without another request being added */ > > - i915_queue_hangcheck(dev_priv); > > + i915_queue_hangcheck(i915); > > } > > void intel_engine_init_hangcheck(struct intel_engine_cs *engine) > > diff --git a/drivers/gpu/drm/i915/gt/intel_mocs.c b/drivers/gpu/drm/i915/gt/intel_mocs.c > > index 79df66022d3a..9d6f0893f528 100644 > > --- a/drivers/gpu/drm/i915/gt/intel_mocs.c > > +++ b/drivers/gpu/drm/i915/gt/intel_mocs.c > > @@ -239,7 +239,7 @@ static const struct drm_i915_mocs_entry icelake_mocs_table[] = { > > /** > > * get_mocs_settings() > > - * @dev_priv: i915 device. > > + * @i915: i915 device. > > * @table: Output table that will be made to point at appropriate > > * MOCS values for the device. > > * > > @@ -249,33 +249,33 @@ static const struct drm_i915_mocs_entry icelake_mocs_table[] = { > > * > > * Return: true if there are applicable MOCS settings for the device. > > */ > > -static bool get_mocs_settings(struct drm_i915_private *dev_priv, > > +static bool get_mocs_settings(struct drm_i915_private *i915, > > struct drm_i915_mocs_table *table) > > { > > bool result = false; > > - if (INTEL_GEN(dev_priv) >= 11) { > > + if (INTEL_GEN(i915) >= 11) { > > table->size = ARRAY_SIZE(icelake_mocs_table); > > table->table = icelake_mocs_table; > > table->n_entries = GEN11_NUM_MOCS_ENTRIES; > > result = true; > > - } else if (IS_GEN9_BC(dev_priv) || IS_CANNONLAKE(dev_priv)) { > > + } else if (IS_GEN9_BC(i915) || IS_CANNONLAKE(i915)) { > > table->size = ARRAY_SIZE(skylake_mocs_table); > > table->n_entries = GEN9_NUM_MOCS_ENTRIES; > > table->table = skylake_mocs_table; > > result = true; > > - } else if (IS_GEN9_LP(dev_priv)) { > > + } else if (IS_GEN9_LP(i915)) { > > table->size = ARRAY_SIZE(broxton_mocs_table); > > table->n_entries = GEN9_NUM_MOCS_ENTRIES; > > table->table = broxton_mocs_table; > > result = true; > > } else { > > - WARN_ONCE(INTEL_GEN(dev_priv) >= 9, > > + WARN_ONCE(INTEL_GEN(i915) >= 9, > > "Platform that should have a MOCS table does not.\n"); > > } > > /* WaDisableSkipCaching:skl,bxt,kbl,glk */ > > - if (IS_GEN(dev_priv, 9)) { > > + if (IS_GEN(i915, 9)) { > > int i; > > for (i = 0; i < table->size; i++) > > @@ -330,12 +330,12 @@ static u32 get_entry_control(const struct drm_i915_mocs_table *table, > > */ > > void intel_mocs_init_engine(struct intel_engine_cs *engine) > > { > > - struct drm_i915_private *dev_priv = engine->i915; > > + struct drm_i915_private *i915 = engine->i915; > > struct drm_i915_mocs_table table; > > unsigned int index; > > u32 unused_value; > > - if (!get_mocs_settings(dev_priv, &table)) > > + if (!get_mocs_settings(i915, &table)) > > return; > > /* Set unused values to PTE */ > > @@ -482,7 +482,7 @@ static int emit_mocs_l3cc_table(struct i915_request *rq, > > /** > > * intel_mocs_init_l3cc_table() - program the mocs control table > > - * @dev_priv: i915 device private > > + * @i915: i915 device private > > * > > * This function simply programs the mocs registers for the given table > > * starting at the given address. This register set is programmed in pairs. > > @@ -494,13 +494,13 @@ static int emit_mocs_l3cc_table(struct i915_request *rq, > > * > > * Return: Nothing. > > */ > > -void intel_mocs_init_l3cc_table(struct drm_i915_private *dev_priv) > > +void intel_mocs_init_l3cc_table(struct drm_i915_private *i915) > > { > > struct drm_i915_mocs_table table; > > unsigned int i; > > u16 unused_value; > > - if (!get_mocs_settings(dev_priv, &table)) > > + if (!get_mocs_settings(i915, &table)) > > return; > > /* Set unused values to PTE */ > > diff --git a/drivers/gpu/drm/i915/gt/intel_mocs.h b/drivers/gpu/drm/i915/gt/intel_mocs.h > > index 0913704a1af2..6f8973fbcb3e 100644 > > --- a/drivers/gpu/drm/i915/gt/intel_mocs.h > > +++ b/drivers/gpu/drm/i915/gt/intel_mocs.h > > @@ -54,7 +54,7 @@ struct i915_request; > > struct intel_engine_cs; > > int intel_rcs_context_init_mocs(struct i915_request *rq); > > -void intel_mocs_init_l3cc_table(struct drm_i915_private *dev_priv); > > +void intel_mocs_init_l3cc_table(struct drm_i915_private *i915); > > void intel_mocs_init_engine(struct intel_engine_cs *engine); > > #endif > > diff --git a/drivers/gpu/drm/i915/gt/intel_reset.c b/drivers/gpu/drm/i915/gt/intel_reset.c > > index 6368b37f26d1..d5cd3cccb407 100644 > > --- a/drivers/gpu/drm/i915/gt/intel_reset.c > > +++ b/drivers/gpu/drm/i915/gt/intel_reset.c > > @@ -1208,13 +1208,13 @@ static void clear_error_registers(struct drm_i915_private *i915, > > } > > } > > -static void gen6_check_faults(struct drm_i915_private *dev_priv) > > +static void gen6_check_faults(struct drm_i915_private *i915) > > { > > struct intel_engine_cs *engine; > > enum intel_engine_id id; > > u32 fault; > > - for_each_engine(engine, dev_priv, id) { > > + for_each_engine(engine, i915, id) { > > fault = GEN6_RING_FAULT_REG_READ(engine); > > if (fault & RING_FAULT_VALID) { > > DRM_DEBUG_DRIVER("Unexpected fault\n" > > @@ -1230,7 +1230,7 @@ static void gen6_check_faults(struct drm_i915_private *dev_priv) > > } > > } > > -static void gen8_check_faults(struct drm_i915_private *dev_priv) > > +static void gen8_check_faults(struct drm_i915_private *i915) > > { > > u32 fault = I915_READ(GEN8_RING_FAULT_REG); > > diff --git a/drivers/gpu/drm/i915/gt/intel_ringbuffer.c b/drivers/gpu/drm/i915/gt/intel_ringbuffer.c > > index b3bf47e8162f..8ec2d332e206 100644 > > --- a/drivers/gpu/drm/i915/gt/intel_ringbuffer.c > > +++ b/drivers/gpu/drm/i915/gt/intel_ringbuffer.c > > @@ -489,11 +489,11 @@ static void set_hwstam(struct intel_engine_cs *engine, u32 mask) > > static void set_hws_pga(struct intel_engine_cs *engine, phys_addr_t phys) > > { > > - struct drm_i915_private *dev_priv = engine->i915; > > + struct drm_i915_private *i915 = engine->i915; > > u32 addr; > > addr = lower_32_bits(phys); > > - if (INTEL_GEN(dev_priv) >= 4) > > + if (INTEL_GEN(i915) >= 4) > > addr |= (phys >> 28) & 0xf0; > > I915_WRITE(HWS_PGA, addr); > > @@ -515,14 +515,14 @@ static void ring_setup_phys_status_page(struct intel_engine_cs *engine) > > static void set_hwsp(struct intel_engine_cs *engine, u32 offset) > > { > > - struct drm_i915_private *dev_priv = engine->i915; > > + struct drm_i915_private *i915 = engine->i915; > > i915_reg_t hwsp; > > /* > > * The ring status page addresses are no longer next to the rest of > > * the ring registers as of gen7. > > */ > > - if (IS_GEN(dev_priv, 7)) { > > + if (IS_GEN(i915, 7)) { > > switch (engine->id) { > > /* > > * No more rings exist on Gen7. Default case is only to shut up > > @@ -544,7 +544,7 @@ static void set_hwsp(struct intel_engine_cs *engine, u32 offset) > > hwsp = VEBOX_HWS_PGA_GEN7; > > break; > > } > > - } else if (IS_GEN(dev_priv, 6)) { > > + } else if (IS_GEN(i915, 6)) { > > hwsp = RING_HWS_PGA_GEN6(engine->mmio_base); > > } else { > > hwsp = RING_HWS_PGA(engine->mmio_base); > > @@ -556,9 +556,9 @@ static void set_hwsp(struct intel_engine_cs *engine, u32 offset) > > static void flush_cs_tlb(struct intel_engine_cs *engine) > > { > > - struct drm_i915_private *dev_priv = engine->i915; > > + struct drm_i915_private *i915 = engine->i915; > > - if (!IS_GEN_RANGE(dev_priv, 6, 7)) > > + if (!IS_GEN_RANGE(i915, 6, 7)) > > return; > > /* ring should be idle before issuing a sync flush*/ > > @@ -585,9 +585,9 @@ static void ring_setup_status_page(struct intel_engine_cs *engine) > > static bool stop_ring(struct intel_engine_cs *engine) > > { > > - struct drm_i915_private *dev_priv = engine->i915; > > + struct drm_i915_private *i915 = engine->i915; > > - if (INTEL_GEN(dev_priv) > 2) { > > + if (INTEL_GEN(i915) > 2) { > > ENGINE_WRITE(engine, > > RING_MI_MODE, _MASKED_BIT_ENABLE(STOP_RING)); > > if (intel_wait_for_register(engine->uncore, > > @@ -622,7 +622,7 @@ static bool stop_ring(struct intel_engine_cs *engine) > > static int xcs_resume(struct intel_engine_cs *engine) > > { > > - struct drm_i915_private *dev_priv = engine->i915; > > + struct drm_i915_private *i915 = engine->i915; > > struct intel_ring *ring = engine->buffer; > > int ret = 0; > > @@ -654,7 +654,7 @@ static int xcs_resume(struct intel_engine_cs *engine) > > } > > } > > - if (HWS_NEEDS_PHYSICAL(dev_priv)) > > + if (HWS_NEEDS_PHYSICAL(i915)) > > ring_setup_phys_status_page(engine); > > else > > ring_setup_status_page(engine); > > @@ -705,7 +705,7 @@ static int xcs_resume(struct intel_engine_cs *engine) > > goto out; > > } > > - if (INTEL_GEN(dev_priv) > 2) > > + if (INTEL_GEN(i915) > 2) > > ENGINE_WRITE(engine, > > RING_MI_MODE, _MASKED_BIT_DISABLE(STOP_RING)); > > @@ -815,7 +815,7 @@ static int intel_rcs_ctx_init(struct i915_request *rq) > > static int rcs_resume(struct intel_engine_cs *engine) > > { > > - struct drm_i915_private *dev_priv = engine->i915; > > + struct drm_i915_private *i915 = engine->i915; > > /* > > * Disable CONSTANT_BUFFER before it is loaded from the context > > @@ -827,12 +827,12 @@ static int rcs_resume(struct intel_engine_cs *engine) > > * they are already accustomed to from before contexts were > > * enabled. > > */ > > - if (IS_GEN(dev_priv, 4)) > > + if (IS_GEN(i915, 4)) > > I915_WRITE(ECOSKPD, > > _MASKED_BIT_ENABLE(ECO_CONSTANT_BUFFER_SR_DISABLE)); > > /* WaTimedSingleVertexDispatch:cl,bw,ctg,elk,ilk,snb */ > > - if (IS_GEN_RANGE(dev_priv, 4, 6)) > > + if (IS_GEN_RANGE(i915, 4, 6)) > > I915_WRITE(MI_MODE, _MASKED_BIT_ENABLE(VS_TIMER_DISPATCH)); > > /* We need to disable the AsyncFlip performance optimisations in order > > @@ -841,22 +841,22 @@ static int rcs_resume(struct intel_engine_cs *engine) > > * > > * WaDisableAsyncFlipPerfMode:snb,ivb,hsw,vlv > > */ > > - if (IS_GEN_RANGE(dev_priv, 6, 7)) > > + if (IS_GEN_RANGE(i915, 6, 7)) > > I915_WRITE(MI_MODE, _MASKED_BIT_ENABLE(ASYNC_FLIP_PERF_DISABLE)); > > /* Required for the hardware to program scanline values for waiting */ > > /* WaEnableFlushTlbInvalidationMode:snb */ > > - if (IS_GEN(dev_priv, 6)) > > + if (IS_GEN(i915, 6)) > > I915_WRITE(GFX_MODE, > > _MASKED_BIT_ENABLE(GFX_TLB_INVALIDATE_EXPLICIT)); > > /* WaBCSVCSTlbInvalidationMode:ivb,vlv,hsw */ > > - if (IS_GEN(dev_priv, 7)) > > + if (IS_GEN(i915, 7)) > > I915_WRITE(GFX_MODE_GEN7, > > _MASKED_BIT_ENABLE(GFX_TLB_INVALIDATE_EXPLICIT) | > > _MASKED_BIT_ENABLE(GFX_REPLAY_MODE)); > > - if (IS_GEN(dev_priv, 6)) { > > + if (IS_GEN(i915, 6)) { > > /* From the Sandybridge PRM, volume 1 part 3, page 24: > > * "If this bit is set, STCunit will have LRA as replacement > > * policy. [...] This bit must be reset. LRA replacement > > @@ -866,7 +866,7 @@ static int rcs_resume(struct intel_engine_cs *engine) > > _MASKED_BIT_DISABLE(CM0_STC_EVICT_DISABLE_LRA_SNB)); > > } > > - if (IS_GEN_RANGE(dev_priv, 6, 7)) > > + if (IS_GEN_RANGE(i915, 6, 7)) > > I915_WRITE(INSTPM, _MASKED_BIT_ENABLE(INSTPM_FORCE_ORDERING)); > > return xcs_resume(engine); > > @@ -1227,15 +1227,15 @@ void intel_ring_unpin(struct intel_ring *ring) > > } > > static struct i915_vma * > > -intel_ring_create_vma(struct drm_i915_private *dev_priv, int size) > > +intel_ring_create_vma(struct drm_i915_private *i915, int size) > > { > > - struct i915_address_space *vm = &dev_priv->ggtt.vm; > > + struct i915_address_space *vm = &i915->ggtt.vm; > > struct drm_i915_gem_object *obj; > > struct i915_vma *vma; > > - obj = i915_gem_object_create_stolen(dev_priv, size); > > + obj = i915_gem_object_create_stolen(i915, size); > > if (!obj) > > - obj = i915_gem_object_create_internal(dev_priv, size); > > + obj = i915_gem_object_create_internal(i915, size); > > if (IS_ERR(obj)) > > return ERR_CAST(obj); > > @@ -2115,9 +2115,9 @@ static void gen6_bsd_set_default_submission(struct intel_engine_cs *engine) > > static void ring_destroy(struct intel_engine_cs *engine) > > { > > - struct drm_i915_private *dev_priv = engine->i915; > > + struct drm_i915_private *i915 = engine->i915; > > - WARN_ON(INTEL_GEN(dev_priv) > 2 && > > + WARN_ON(INTEL_GEN(i915) > 2 && > > (ENGINE_READ(engine, RING_MI_MODE) & MODE_IDLE) == 0); > > intel_ring_unpin(engine->buffer); > > diff --git a/drivers/gpu/drm/i915/gvt/aperture_gm.c b/drivers/gpu/drm/i915/gvt/aperture_gm.c > > index 1fa2f65c3cd1..8807de566361 100644 > > --- a/drivers/gpu/drm/i915/gvt/aperture_gm.c > > +++ b/drivers/gpu/drm/i915/gvt/aperture_gm.c > > @@ -40,7 +40,7 @@ > > static int alloc_gm(struct intel_vgpu *vgpu, bool high_gm) > > { > > struct intel_gvt *gvt = vgpu->gvt; > > - struct drm_i915_private *dev_priv = gvt->dev_priv; > > + struct drm_i915_private *i915 = gvt->i915; > > unsigned int flags; > > u64 start, end, size; > > struct drm_mm_node *node; > > @@ -60,14 +60,14 @@ static int alloc_gm(struct intel_vgpu *vgpu, bool high_gm) > > flags = PIN_MAPPABLE; > > } > > - mutex_lock(&dev_priv->drm.struct_mutex); > > - mmio_hw_access_pre(dev_priv); > > - ret = i915_gem_gtt_insert(&dev_priv->ggtt.vm, node, > > + mutex_lock(&i915->drm.struct_mutex); > > + mmio_hw_access_pre(i915); > > + ret = i915_gem_gtt_insert(&i915->ggtt.vm, node, > > size, I915_GTT_PAGE_SIZE, > > I915_COLOR_UNEVICTABLE, > > start, end, flags); > > - mmio_hw_access_post(dev_priv); > > - mutex_unlock(&dev_priv->drm.struct_mutex); > > + mmio_hw_access_post(i915); > > + mutex_unlock(&i915->drm.struct_mutex); > > if (ret) > > gvt_err("fail to alloc %s gm space from host\n", > > high_gm ? "high" : "low"); > > @@ -78,7 +78,7 @@ static int alloc_gm(struct intel_vgpu *vgpu, bool high_gm) > > static int alloc_vgpu_gm(struct intel_vgpu *vgpu) > > { > > struct intel_gvt *gvt = vgpu->gvt; > > - struct drm_i915_private *dev_priv = gvt->dev_priv; > > + struct drm_i915_private *i915 = gvt->i915; > > int ret; > > ret = alloc_gm(vgpu, false); > > @@ -97,20 +97,20 @@ static int alloc_vgpu_gm(struct intel_vgpu *vgpu) > > return 0; > > out_free_aperture: > > - mutex_lock(&dev_priv->drm.struct_mutex); > > + mutex_lock(&i915->drm.struct_mutex); > > drm_mm_remove_node(&vgpu->gm.low_gm_node); > > - mutex_unlock(&dev_priv->drm.struct_mutex); > > + mutex_unlock(&i915->drm.struct_mutex); > > return ret; > > } > > static void free_vgpu_gm(struct intel_vgpu *vgpu) > > { > > - struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv; > > + struct drm_i915_private *i915 = vgpu->gvt->i915; > > - mutex_lock(&dev_priv->drm.struct_mutex); > > + mutex_lock(&i915->drm.struct_mutex); > > drm_mm_remove_node(&vgpu->gm.low_gm_node); > > drm_mm_remove_node(&vgpu->gm.high_gm_node); > > - mutex_unlock(&dev_priv->drm.struct_mutex); > > + mutex_unlock(&i915->drm.struct_mutex); > > } > > /** > > @@ -127,11 +127,11 @@ void intel_vgpu_write_fence(struct intel_vgpu *vgpu, > > u32 fence, u64 value) > > { > > struct intel_gvt *gvt = vgpu->gvt; > > - struct drm_i915_private *dev_priv = gvt->dev_priv; > > + struct drm_i915_private *i915 = gvt->i915; > > struct drm_i915_fence_reg *reg; > > i915_reg_t fence_reg_lo, fence_reg_hi; > > - assert_rpm_wakelock_held(dev_priv); > > + assert_rpm_wakelock_held(i915); > > if (WARN_ON(fence >= vgpu_fence_sz(vgpu))) > > return; > > @@ -162,41 +162,41 @@ static void _clear_vgpu_fence(struct intel_vgpu *vgpu) > > static void free_vgpu_fence(struct intel_vgpu *vgpu) > > { > > struct intel_gvt *gvt = vgpu->gvt; > > - struct drm_i915_private *dev_priv = gvt->dev_priv; > > + struct drm_i915_private *i915 = gvt->i915; > > struct drm_i915_fence_reg *reg; > > u32 i; > > if (WARN_ON(!vgpu_fence_sz(vgpu))) > > return; > > - intel_runtime_pm_get(dev_priv); > > + intel_runtime_pm_get(i915); > > - mutex_lock(&dev_priv->drm.struct_mutex); > > + mutex_lock(&i915->drm.struct_mutex); > > _clear_vgpu_fence(vgpu); > > for (i = 0; i < vgpu_fence_sz(vgpu); i++) { > > reg = vgpu->fence.regs[i]; > > i915_unreserve_fence(reg); > > vgpu->fence.regs[i] = NULL; > > } > > - mutex_unlock(&dev_priv->drm.struct_mutex); > > + mutex_unlock(&i915->drm.struct_mutex); > > - intel_runtime_pm_put_unchecked(dev_priv); > > + intel_runtime_pm_put_unchecked(i915); > > } > > static int alloc_vgpu_fence(struct intel_vgpu *vgpu) > > { > > struct intel_gvt *gvt = vgpu->gvt; > > - struct drm_i915_private *dev_priv = gvt->dev_priv; > > + struct drm_i915_private *i915 = gvt->i915; > > struct drm_i915_fence_reg *reg; > > int i; > > - intel_runtime_pm_get(dev_priv); > > + intel_runtime_pm_get(i915); > > /* Request fences from host */ > > - mutex_lock(&dev_priv->drm.struct_mutex); > > + mutex_lock(&i915->drm.struct_mutex); > > for (i = 0; i < vgpu_fence_sz(vgpu); i++) { > > - reg = i915_reserve_fence(dev_priv); > > + reg = i915_reserve_fence(i915); > > if (IS_ERR(reg)) > > goto out_free_fence; > > @@ -205,8 +205,8 @@ static int alloc_vgpu_fence(struct intel_vgpu *vgpu) > > _clear_vgpu_fence(vgpu); > > - mutex_unlock(&dev_priv->drm.struct_mutex); > > - intel_runtime_pm_put_unchecked(dev_priv); > > + mutex_unlock(&i915->drm.struct_mutex); > > + intel_runtime_pm_put_unchecked(i915); > > return 0; > > out_free_fence: > > gvt_vgpu_err("Failed to alloc fences\n"); > > @@ -218,8 +218,8 @@ static int alloc_vgpu_fence(struct intel_vgpu *vgpu) > > i915_unreserve_fence(reg); > > vgpu->fence.regs[i] = NULL; > > } > > - mutex_unlock(&dev_priv->drm.struct_mutex); > > - intel_runtime_pm_put_unchecked(dev_priv); > > + mutex_unlock(&i915->drm.struct_mutex); > > + intel_runtime_pm_put_unchecked(i915); > > return -ENOSPC; > > } > > @@ -313,11 +313,11 @@ void intel_vgpu_free_resource(struct intel_vgpu *vgpu) > > */ > > void intel_vgpu_reset_resource(struct intel_vgpu *vgpu) > > { > > - struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv; > > + struct drm_i915_private *i915 = vgpu->gvt->i915; > > - intel_runtime_pm_get(dev_priv); > > + intel_runtime_pm_get(i915); > > _clear_vgpu_fence(vgpu); > > - intel_runtime_pm_put_unchecked(dev_priv); > > + intel_runtime_pm_put_unchecked(i915); > > } > > /** > > diff --git a/drivers/gpu/drm/i915/gvt/cfg_space.c b/drivers/gpu/drm/i915/gvt/cfg_space.c > > index 19cf1bbe059d..d3a744e7c959 100644 > > --- a/drivers/gpu/drm/i915/gvt/cfg_space.c > > +++ b/drivers/gpu/drm/i915/gvt/cfg_space.c > > @@ -391,9 +391,9 @@ void intel_vgpu_init_cfg_space(struct intel_vgpu *vgpu, > > memset(vgpu_cfg_space(vgpu) + INTEL_GVT_PCI_OPREGION, 0, 4); > > vgpu->cfg_space.bar[INTEL_GVT_PCI_BAR_GTTMMIO].size = > > - pci_resource_len(gvt->dev_priv->drm.pdev, 0); > > + pci_resource_len(gvt->i915->drm.pdev, 0); > > vgpu->cfg_space.bar[INTEL_GVT_PCI_BAR_APERTURE].size = > > - pci_resource_len(gvt->dev_priv->drm.pdev, 2); > > + pci_resource_len(gvt->i915->drm.pdev, 2); > > memset(vgpu_cfg_space(vgpu) + PCI_ROM_ADDRESS, 0, 4); > > } > > diff --git a/drivers/gpu/drm/i915/gvt/cmd_parser.c b/drivers/gpu/drm/i915/gvt/cmd_parser.c > > index 6ea88270c818..2b09acc0502f 100644 > > --- a/drivers/gpu/drm/i915/gvt/cmd_parser.c > > +++ b/drivers/gpu/drm/i915/gvt/cmd_parser.c > > @@ -819,7 +819,7 @@ static int force_nonpriv_reg_handler(struct parser_exec_state *s, > > unsigned int data; > > u32 ring_base; > > u32 nopid; > > - struct drm_i915_private *dev_priv = s->vgpu->gvt->dev_priv; > > + struct drm_i915_private *i915 = s->vgpu->gvt->i915; > > if (!strcmp(cmd, "lri")) > > data = cmd_val(s, index + 1); > > @@ -829,7 +829,7 @@ static int force_nonpriv_reg_handler(struct parser_exec_state *s, > > return -EINVAL; > > } > > - ring_base = dev_priv->engine[s->ring_id]->mmio_base; > > + ring_base = i915->engine[s->ring_id]->mmio_base; > > nopid = i915_mmio_reg_offset(RING_NOPID(ring_base)); > > if (!intel_gvt_in_force_nonpriv_whitelist(gvt, data) && > > @@ -905,7 +905,7 @@ static int cmd_reg_handler(struct parser_exec_state *s, > > * update reg values in it into vregs, so LRIs in workload with > > * inhibit context will restore with correct values > > */ > > - if (IS_GEN(gvt->dev_priv, 9) && > > + if (IS_GEN(gvt->i915, 9) && > > intel_gvt_mmio_is_in_ctx(gvt, offset) && > > !strncmp(cmd, "lri", 3)) { > > intel_gvt_hypervisor_read_gpa(s->vgpu, > > @@ -946,7 +946,7 @@ static int cmd_handler_lri(struct parser_exec_state *s) > > struct intel_gvt *gvt = s->vgpu->gvt; > > for (i = 1; i < cmd_len; i += 2) { > > - if (IS_BROADWELL(gvt->dev_priv) && s->ring_id != RCS0) { > > + if (IS_BROADWELL(gvt->i915) && s->ring_id != RCS0) { > > if (s->ring_id == BCS0 && > > cmd_reg(s, i) == i915_mmio_reg_offset(DERRMR)) > > ret |= 0; > > @@ -968,7 +968,7 @@ static int cmd_handler_lrr(struct parser_exec_state *s) > > int cmd_len = cmd_length(s); > > for (i = 1; i < cmd_len; i += 2) { > > - if (IS_BROADWELL(s->vgpu->gvt->dev_priv)) > > + if (IS_BROADWELL(s->vgpu->gvt->i915)) > > ret |= ((cmd_reg_inhibit(s, i) || > > (cmd_reg_inhibit(s, i + 1)))) ? > > -EBADRQC : 0; > > @@ -996,7 +996,7 @@ static int cmd_handler_lrm(struct parser_exec_state *s) > > int cmd_len = cmd_length(s); > > for (i = 1; i < cmd_len;) { > > - if (IS_BROADWELL(gvt->dev_priv)) > > + if (IS_BROADWELL(gvt->i915)) > > ret |= (cmd_reg_inhibit(s, i)) ? -EBADRQC : 0; > > if (ret) > > break; > > @@ -1180,7 +1180,7 @@ struct plane_code_mapping { > > static int gen8_decode_mi_display_flip(struct parser_exec_state *s, > > struct mi_display_flip_command_info *info) > > { > > - struct drm_i915_private *dev_priv = s->vgpu->gvt->dev_priv; > > + struct drm_i915_private *i915 = s->vgpu->gvt->i915; > > struct plane_code_mapping gen8_plane_code[] = { > > [0] = {PIPE_A, PLANE_A, PRIMARY_A_FLIP_DONE}, > > [1] = {PIPE_B, PLANE_A, PRIMARY_B_FLIP_DONE}, > > @@ -1226,7 +1226,7 @@ static int gen8_decode_mi_display_flip(struct parser_exec_state *s, > > static int skl_decode_mi_display_flip(struct parser_exec_state *s, > > struct mi_display_flip_command_info *info) > > { > > - struct drm_i915_private *dev_priv = s->vgpu->gvt->dev_priv; > > + struct drm_i915_private *i915 = s->vgpu->gvt->i915; > > struct intel_vgpu *vgpu = s->vgpu; > > u32 dword0 = cmd_val(s, 0); > > u32 dword1 = cmd_val(s, 1); > > @@ -1285,13 +1285,13 @@ static int skl_decode_mi_display_flip(struct parser_exec_state *s, > > static int gen8_check_mi_display_flip(struct parser_exec_state *s, > > struct mi_display_flip_command_info *info) > > { > > - struct drm_i915_private *dev_priv = s->vgpu->gvt->dev_priv; > > + struct drm_i915_private *i915 = s->vgpu->gvt->i915; > > u32 stride, tile; > > if (!info->async_flip) > > return 0; > > - if (INTEL_GEN(dev_priv) >= 9) { > > + if (INTEL_GEN(i915) >= 9) { > > stride = vgpu_vreg_t(s->vgpu, info->stride_reg) & GENMASK(9, 0); > > tile = (vgpu_vreg_t(s->vgpu, info->ctrl_reg) & > > GENMASK(12, 10)) >> 10; > > @@ -1314,12 +1314,12 @@ static int gen8_update_plane_mmio_from_mi_display_flip( > > struct parser_exec_state *s, > > struct mi_display_flip_command_info *info) > > { > > - struct drm_i915_private *dev_priv = s->vgpu->gvt->dev_priv; > > + struct drm_i915_private *i915 = s->vgpu->gvt->i915; > > struct intel_vgpu *vgpu = s->vgpu; > > set_mask_bits(&vgpu_vreg_t(vgpu, info->surf_reg), GENMASK(31, 12), > > info->surf_val << 12); > > - if (INTEL_GEN(dev_priv) >= 9) { > > + if (INTEL_GEN(i915) >= 9) { > > set_mask_bits(&vgpu_vreg_t(vgpu, info->stride_reg), GENMASK(9, 0), > > info->stride_val); > > set_mask_bits(&vgpu_vreg_t(vgpu, info->ctrl_reg), GENMASK(12, 10), > > @@ -1345,11 +1345,11 @@ static int gen8_update_plane_mmio_from_mi_display_flip( > > static int decode_mi_display_flip(struct parser_exec_state *s, > > struct mi_display_flip_command_info *info) > > { > > - struct drm_i915_private *dev_priv = s->vgpu->gvt->dev_priv; > > + struct drm_i915_private *i915 = s->vgpu->gvt->i915; > > - if (IS_BROADWELL(dev_priv)) > > + if (IS_BROADWELL(i915)) > > return gen8_decode_mi_display_flip(s, info); > > - if (INTEL_GEN(dev_priv) >= 9) > > + if (INTEL_GEN(i915) >= 9) > > return skl_decode_mi_display_flip(s, info); > > return -ENODEV; > > @@ -1756,7 +1756,7 @@ static int perform_bb_shadow(struct parser_exec_state *s) > > if (bb->ppgtt) > > start_offset = gma & ~I915_GTT_PAGE_MASK; > > - bb->obj = i915_gem_object_create_shmem(s->vgpu->gvt->dev_priv, > > + bb->obj = i915_gem_object_create_shmem(s->vgpu->gvt->i915, > > round_up(bb_size + start_offset, > > PAGE_SIZE)); > > if (IS_ERR(bb->obj)) { > > @@ -2830,7 +2830,7 @@ static int shadow_indirect_ctx(struct intel_shadow_wa_ctx *wa_ctx) > > int ret = 0; > > void *map; > > - obj = i915_gem_object_create_shmem(workload->vgpu->gvt->dev_priv, > > + obj = i915_gem_object_create_shmem(workload->vgpu->gvt->i915, > > roundup(ctx_size + CACHELINE_BYTES, > > PAGE_SIZE)); > > if (IS_ERR(obj)) > > diff --git a/drivers/gpu/drm/i915/gvt/debugfs.c b/drivers/gpu/drm/i915/gvt/debugfs.c > > index 2fb7b73b260d..1ff7f11c89a2 100644 > > --- a/drivers/gpu/drm/i915/gvt/debugfs.c > > +++ b/drivers/gpu/drm/i915/gvt/debugfs.c > > @@ -58,7 +58,7 @@ static int mmio_offset_compare(void *priv, > > static inline int mmio_diff_handler(struct intel_gvt *gvt, > > u32 offset, void *data) > > { > > - struct drm_i915_private *i915 = gvt->dev_priv; > > + struct drm_i915_private *i915 = gvt->i915; > > struct mmio_diff_param *param = data; > > struct diff_mmio *node; > > u32 preg, vreg; > > @@ -98,10 +98,10 @@ static int vgpu_mmio_diff_show(struct seq_file *s, void *unused) > > mutex_lock(&gvt->lock); > > spin_lock_bh(&gvt->scheduler.mmio_context_lock); > > - mmio_hw_access_pre(gvt->dev_priv); > > + mmio_hw_access_pre(gvt->i915); > > /* Recognize all the diff mmios to list. */ > > intel_gvt_for_each_tracked_mmio(gvt, mmio_diff_handler, ¶m); > > - mmio_hw_access_post(gvt->dev_priv); > > + mmio_hw_access_post(gvt->i915); > > spin_unlock_bh(&gvt->scheduler.mmio_context_lock); > > mutex_unlock(&gvt->lock); > > @@ -142,7 +142,7 @@ static int > > vgpu_scan_nonprivbb_set(void *data, u64 val) > > { > > struct intel_vgpu *vgpu = (struct intel_vgpu *)data; > > - struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv; > > + struct drm_i915_private *i915 = vgpu->gvt->i915; > > enum intel_engine_id id; > > char buf[128], *s; > > int len; > > @@ -164,7 +164,7 @@ vgpu_scan_nonprivbb_set(void *data, u64 val) > > for (id = 0; id < I915_NUM_ENGINES; id++) { > > struct intel_engine_cs *engine; > > - engine = dev_priv->engine[id]; > > + engine = i915->engine[id]; > > if (engine && (val & (1 << id))) { > > len = snprintf(s, 4, "%d, ", engine->id); > > s += len; > > @@ -240,7 +240,7 @@ void intel_gvt_debugfs_remove_vgpu(struct intel_vgpu *vgpu) > > */ > > int intel_gvt_debugfs_init(struct intel_gvt *gvt) > > { > > - struct drm_minor *minor = gvt->dev_priv->drm.primary; > > + struct drm_minor *minor = gvt->i915->drm.primary; > > struct dentry *ent; > > gvt->debugfs_root = debugfs_create_dir("gvt", minor->debugfs_root); > > diff --git a/drivers/gpu/drm/i915/gvt/display.c b/drivers/gpu/drm/i915/gvt/display.c > > index e1c313da6c00..b037aeed2f40 100644 > > --- a/drivers/gpu/drm/i915/gvt/display.c > > +++ b/drivers/gpu/drm/i915/gvt/display.c > > @@ -57,7 +57,7 @@ static int get_edp_pipe(struct intel_vgpu *vgpu) > > static int edp_pipe_is_enabled(struct intel_vgpu *vgpu) > > { > > - struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv; > > + struct drm_i915_private *i915 = vgpu->gvt->i915; > > if (!(vgpu_vreg_t(vgpu, PIPECONF(_PIPE_EDP)) & PIPECONF_ENABLE)) > > return 0; > > @@ -69,7 +69,7 @@ static int edp_pipe_is_enabled(struct intel_vgpu *vgpu) > > int pipe_is_enabled(struct intel_vgpu *vgpu, int pipe) > > { > > - struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv; > > + struct drm_i915_private *i915 = vgpu->gvt->i915; > > if (WARN_ON(pipe < PIPE_A || pipe >= I915_MAX_PIPES)) > > return -EINVAL; > > @@ -168,10 +168,10 @@ static u8 dpcd_fix_data[DPCD_HEADER_SIZE] = { > > static void emulate_monitor_status_change(struct intel_vgpu *vgpu) > > { > > - struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv; > > + struct drm_i915_private *i915 = vgpu->gvt->i915; > > int pipe; > > - if (IS_BROXTON(dev_priv)) { > > + if (IS_BROXTON(i915)) { > > vgpu_vreg_t(vgpu, GEN8_DE_PORT_ISR) &= ~(BXT_DE_PORT_HP_DDIA | > > BXT_DE_PORT_HP_DDIB | > > BXT_DE_PORT_HP_DDIC); > > @@ -198,8 +198,8 @@ static void emulate_monitor_status_change(struct intel_vgpu *vgpu) > > SDE_PORTC_HOTPLUG_CPT | > > SDE_PORTD_HOTPLUG_CPT); > > - if (IS_SKYLAKE(dev_priv) || IS_KABYLAKE(dev_priv) || > > - IS_COFFEELAKE(dev_priv)) { > > + if (IS_SKYLAKE(i915) || IS_KABYLAKE(i915) || > > + IS_COFFEELAKE(i915)) { > > vgpu_vreg_t(vgpu, SDEISR) &= ~(SDE_PORTA_HOTPLUG_SPT | > > SDE_PORTE_HOTPLUG_SPT); > > vgpu_vreg_t(vgpu, SKL_FUSE_STATUS) |= > > @@ -223,7 +223,7 @@ static void emulate_monitor_status_change(struct intel_vgpu *vgpu) > > (TRANS_DDI_BPC_8 | TRANS_DDI_MODE_SELECT_DVI | > > (PORT_B << TRANS_DDI_PORT_SHIFT) | > > TRANS_DDI_FUNC_ENABLE); > > - if (IS_BROADWELL(dev_priv)) { > > + if (IS_BROADWELL(i915)) { > > vgpu_vreg_t(vgpu, PORT_CLK_SEL(PORT_B)) &= > > ~PORT_CLK_SEL_MASK; > > vgpu_vreg_t(vgpu, PORT_CLK_SEL(PORT_B)) |= > > @@ -243,7 +243,7 @@ static void emulate_monitor_status_change(struct intel_vgpu *vgpu) > > (TRANS_DDI_BPC_8 | TRANS_DDI_MODE_SELECT_DVI | > > (PORT_C << TRANS_DDI_PORT_SHIFT) | > > TRANS_DDI_FUNC_ENABLE); > > - if (IS_BROADWELL(dev_priv)) { > > + if (IS_BROADWELL(i915)) { > > vgpu_vreg_t(vgpu, PORT_CLK_SEL(PORT_C)) &= > > ~PORT_CLK_SEL_MASK; > > vgpu_vreg_t(vgpu, PORT_CLK_SEL(PORT_C)) |= > > @@ -263,7 +263,7 @@ static void emulate_monitor_status_change(struct intel_vgpu *vgpu) > > (TRANS_DDI_BPC_8 | TRANS_DDI_MODE_SELECT_DVI | > > (PORT_D << TRANS_DDI_PORT_SHIFT) | > > TRANS_DDI_FUNC_ENABLE); > > - if (IS_BROADWELL(dev_priv)) { > > + if (IS_BROADWELL(i915)) { > > vgpu_vreg_t(vgpu, PORT_CLK_SEL(PORT_D)) &= > > ~PORT_CLK_SEL_MASK; > > vgpu_vreg_t(vgpu, PORT_CLK_SEL(PORT_D)) |= > > @@ -274,14 +274,14 @@ static void emulate_monitor_status_change(struct intel_vgpu *vgpu) > > vgpu_vreg_t(vgpu, SFUSE_STRAP) |= SFUSE_STRAP_DDID_DETECTED; > > } > > - if ((IS_SKYLAKE(dev_priv) || IS_KABYLAKE(dev_priv) || > > - IS_COFFEELAKE(dev_priv)) && > > + if ((IS_SKYLAKE(i915) || IS_KABYLAKE(i915) || > > + IS_COFFEELAKE(i915)) && > > intel_vgpu_has_monitor_on_port(vgpu, PORT_E)) { > > vgpu_vreg_t(vgpu, SDEISR) |= SDE_PORTE_HOTPLUG_SPT; > > } > > if (intel_vgpu_has_monitor_on_port(vgpu, PORT_A)) { > > - if (IS_BROADWELL(dev_priv)) > > + if (IS_BROADWELL(i915)) > > vgpu_vreg_t(vgpu, GEN8_DE_PORT_ISR) |= > > GEN8_PORT_DP_A_HOTPLUG; > > else > > @@ -291,11 +291,11 @@ static void emulate_monitor_status_change(struct intel_vgpu *vgpu) > > } > > /* Clear host CRT status, so guest couldn't detect this host CRT. */ > > - if (IS_BROADWELL(dev_priv)) > > + if (IS_BROADWELL(i915)) > > vgpu_vreg_t(vgpu, PCH_ADPA) &= ~ADPA_CRT_HOTPLUG_MONITOR_MASK; > > /* Disable Primary/Sprite/Cursor plane */ > > - for_each_pipe(dev_priv, pipe) { > > + for_each_pipe(i915, pipe) { > > vgpu_vreg_t(vgpu, DSPCNTR(pipe)) &= ~DISPLAY_PLANE_ENABLE; > > vgpu_vreg_t(vgpu, SPRCTL(pipe)) &= ~SPRITE_ENABLE; > > vgpu_vreg_t(vgpu, CURCNTR(pipe)) &= ~MCURSOR_MODE; > > @@ -389,7 +389,7 @@ void intel_gvt_check_vblank_emulation(struct intel_gvt *gvt) > > static void emulate_vblank_on_pipe(struct intel_vgpu *vgpu, int pipe) > > { > > - struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv; > > + struct drm_i915_private *i915 = vgpu->gvt->i915; > > struct intel_vgpu_irq *irq = &vgpu->irq; > > int vblank_event[] = { > > [PIPE_A] = PIPE_A_VBLANK, > > @@ -421,7 +421,7 @@ static void emulate_vblank(struct intel_vgpu *vgpu) > > int pipe; > > mutex_lock(&vgpu->vgpu_lock); > > - for_each_pipe(vgpu->gvt->dev_priv, pipe) > > + for_each_pipe(vgpu->gvt->i915, pipe) > > emulate_vblank_on_pipe(vgpu, pipe); > > mutex_unlock(&vgpu->vgpu_lock); > > } > > @@ -454,10 +454,10 @@ void intel_gvt_emulate_vblank(struct intel_gvt *gvt) > > */ > > void intel_vgpu_emulate_hotplug(struct intel_vgpu *vgpu, bool connected) > > { > > - struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv; > > + struct drm_i915_private *i915 = vgpu->gvt->i915; > > /* TODO: add more platforms support */ > > - if (IS_SKYLAKE(dev_priv) || IS_KABYLAKE(dev_priv)) { > > + if (IS_SKYLAKE(i915) || IS_KABYLAKE(i915)) { > > if (connected) { > > vgpu_vreg_t(vgpu, SFUSE_STRAP) |= > > SFUSE_STRAP_DDID_DETECTED; > > @@ -483,10 +483,10 @@ void intel_vgpu_emulate_hotplug(struct intel_vgpu *vgpu, bool connected) > > */ > > void intel_vgpu_clean_display(struct intel_vgpu *vgpu) > > { > > - struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv; > > + struct drm_i915_private *i915 = vgpu->gvt->i915; > > - if (IS_SKYLAKE(dev_priv) || IS_KABYLAKE(dev_priv) || > > - IS_COFFEELAKE(dev_priv)) > > + if (IS_SKYLAKE(i915) || IS_KABYLAKE(i915) || > > + IS_COFFEELAKE(i915)) > > clean_virtual_dp_monitor(vgpu, PORT_D); > > else > > clean_virtual_dp_monitor(vgpu, PORT_B); > > @@ -505,12 +505,12 @@ void intel_vgpu_clean_display(struct intel_vgpu *vgpu) > > */ > > int intel_vgpu_init_display(struct intel_vgpu *vgpu, u64 resolution) > > { > > - struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv; > > + struct drm_i915_private *i915 = vgpu->gvt->i915; > > intel_vgpu_init_i2c_edid(vgpu); > > - if (IS_SKYLAKE(dev_priv) || IS_KABYLAKE(dev_priv) || > > - IS_COFFEELAKE(dev_priv)) > > + if (IS_SKYLAKE(i915) || IS_KABYLAKE(i915) || > > + IS_COFFEELAKE(i915)) > > return setup_virtual_dp_monitor(vgpu, PORT_D, GVT_DP_D, > > resolution); > > else > > diff --git a/drivers/gpu/drm/i915/gvt/dmabuf.c b/drivers/gpu/drm/i915/gvt/dmabuf.c > > index 41c8ebc60c63..523038d224a1 100644 > > --- a/drivers/gpu/drm/i915/gvt/dmabuf.c > > +++ b/drivers/gpu/drm/i915/gvt/dmabuf.c > > @@ -39,7 +39,7 @@ > > static int vgpu_gem_get_pages( > > struct drm_i915_gem_object *obj) > > { > > - struct drm_i915_private *dev_priv = to_i915(obj->base.dev); > > + struct drm_i915_private *i915 = to_i915(obj->base.dev); > > struct sg_table *st; > > struct scatterlist *sg; > > int i, ret; > > @@ -61,7 +61,7 @@ static int vgpu_gem_get_pages( > > kfree(st); > > return ret; > > } > > - gtt_entries = (gen8_pte_t __iomem *)dev_priv->ggtt.gsm + > > + gtt_entries = (gen8_pte_t __iomem *)i915->ggtt.gsm + > > (fb_info->start >> PAGE_SHIFT); > > for_each_sg(st->sgl, sg, page_num, i) { > > sg->offset = 0; > > @@ -152,7 +152,7 @@ static const struct drm_i915_gem_object_ops intel_vgpu_gem_ops = { > > static struct drm_i915_gem_object *vgpu_create_gem(struct drm_device *dev, > > struct intel_vgpu_fb_info *info) > > { > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > struct drm_i915_gem_object *obj; > > obj = i915_gem_object_alloc(); > > @@ -165,7 +165,7 @@ static struct drm_i915_gem_object *vgpu_create_gem(struct drm_device *dev, > > obj->read_domains = I915_GEM_DOMAIN_GTT; > > obj->write_domain = 0; > > - if (INTEL_GEN(dev_priv) >= 9) { > > + if (INTEL_GEN(i915) >= 9) { > > unsigned int tiling_mode = 0; > > unsigned int stride = 0; > > @@ -360,7 +360,7 @@ static void update_fb_info(struct vfio_device_gfx_plane_info *gvt_dmabuf, > > int intel_vgpu_query_plane(struct intel_vgpu *vgpu, void *args) > > { > > - struct drm_device *dev = &vgpu->gvt->dev_priv->drm; > > + struct drm_device *dev = &vgpu->gvt->i915->drm; > > struct vfio_device_gfx_plane_info *gfx_plane_info = args; > > struct intel_vgpu_dmabuf_obj *dmabuf_obj; > > struct intel_vgpu_fb_info fb_info; > > @@ -466,7 +466,7 @@ int intel_vgpu_query_plane(struct intel_vgpu *vgpu, void *args) > > /* To associate an exposed dmabuf with the dmabuf_obj */ > > int intel_vgpu_get_dmabuf(struct intel_vgpu *vgpu, unsigned int dmabuf_id) > > { > > - struct drm_device *dev = &vgpu->gvt->dev_priv->drm; > > + struct drm_device *dev = &vgpu->gvt->i915->drm; > > struct intel_vgpu_dmabuf_obj *dmabuf_obj; > > struct drm_i915_gem_object *obj; > > struct dma_buf *dmabuf; > > diff --git a/drivers/gpu/drm/i915/gvt/edid.c b/drivers/gpu/drm/i915/gvt/edid.c > > index 1fe6124918f1..9e45cbac5eed 100644 > > --- a/drivers/gpu/drm/i915/gvt/edid.c > > +++ b/drivers/gpu/drm/i915/gvt/edid.c > > @@ -135,7 +135,7 @@ static void reset_gmbus_controller(struct intel_vgpu *vgpu) > > static int gmbus0_mmio_write(struct intel_vgpu *vgpu, > > unsigned int offset, void *p_data, unsigned int bytes) > > { > > - struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv; > > + struct drm_i915_private *i915 = vgpu->gvt->i915; > > int port, pin_select; > > memcpy(&vgpu_vreg(vgpu, offset), p_data, bytes); > > @@ -147,9 +147,9 @@ static int gmbus0_mmio_write(struct intel_vgpu *vgpu, > > if (pin_select == 0) > > return 0; > > - if (IS_BROXTON(dev_priv)) > > + if (IS_BROXTON(i915)) > > port = bxt_get_port_from_gmbus0(pin_select); > > - else if (IS_COFFEELAKE(dev_priv)) > > + else if (IS_COFFEELAKE(i915)) > > port = cnp_get_port_from_gmbus0(pin_select); > > else > > port = get_port_from_gmbus0(pin_select); > > diff --git a/drivers/gpu/drm/i915/gvt/execlist.c b/drivers/gpu/drm/i915/gvt/execlist.c > > index f21b8fb5b37e..7a7a925565a0 100644 > > --- a/drivers/gpu/drm/i915/gvt/execlist.c > > +++ b/drivers/gpu/drm/i915/gvt/execlist.c > > @@ -40,7 +40,7 @@ > > #define _EL_OFFSET_STATUS_PTR 0x3A0 > > #define execlist_ring_mmio(gvt, ring_id, offset) \ > > - (gvt->dev_priv->engine[ring_id]->mmio_base + (offset)) > > + (gvt->i915->engine[ring_id]->mmio_base + (offset)) > > #define valid_context(ctx) ((ctx)->valid) > > #define same_context(a, b) (((a)->context_id == (b)->context_id) && \ > > @@ -133,7 +133,7 @@ static void emulate_csb_update(struct intel_vgpu_execlist *execlist, > > u32 write_pointer; > > u32 ctx_status_ptr_reg, ctx_status_buf_reg, offset; > > unsigned long hwsp_gpa; > > - struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv; > > + struct drm_i915_private *i915 = vgpu->gvt->i915; > > ctx_status_ptr_reg = execlist_ring_mmio(vgpu->gvt, ring_id, > > _EL_OFFSET_STATUS_PTR); > > @@ -169,7 +169,7 @@ static void emulate_csb_update(struct intel_vgpu_execlist *execlist, > > status, 8); > > intel_gvt_hypervisor_write_gpa(vgpu, > > hwsp_gpa + > > - intel_hws_csb_write_index(dev_priv) * 4, > > + intel_hws_csb_write_index(i915) * 4, > > &write_pointer, 4); > > } > > @@ -529,12 +529,12 @@ static void init_vgpu_execlist(struct intel_vgpu *vgpu, int ring_id) > > static void clean_execlist(struct intel_vgpu *vgpu, > > intel_engine_mask_t engine_mask) > > { > > - struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv; > > + struct drm_i915_private *i915 = vgpu->gvt->i915; > > struct intel_engine_cs *engine; > > struct intel_vgpu_submission *s = &vgpu->submission; > > intel_engine_mask_t tmp; > > - for_each_engine_masked(engine, dev_priv, engine_mask, tmp) { > > + for_each_engine_masked(engine, i915, engine_mask, tmp) { > > kfree(s->ring_scan_buffer[engine->id]); > > s->ring_scan_buffer[engine->id] = NULL; > > s->ring_scan_buffer_size[engine->id] = 0; > > @@ -544,11 +544,11 @@ static void clean_execlist(struct intel_vgpu *vgpu, > > static void reset_execlist(struct intel_vgpu *vgpu, > > intel_engine_mask_t engine_mask) > > { > > - struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv; > > + struct drm_i915_private *i915 = vgpu->gvt->i915; > > struct intel_engine_cs *engine; > > intel_engine_mask_t tmp; > > - for_each_engine_masked(engine, dev_priv, engine_mask, tmp) > > + for_each_engine_masked(engine, i915, engine_mask, tmp) > > init_vgpu_execlist(vgpu, engine->id); > > } > > diff --git a/drivers/gpu/drm/i915/gvt/fb_decoder.c b/drivers/gpu/drm/i915/gvt/fb_decoder.c > > index 65e847392aea..76c346eecda3 100644 > > --- a/drivers/gpu/drm/i915/gvt/fb_decoder.c > > +++ b/drivers/gpu/drm/i915/gvt/fb_decoder.c > > @@ -146,12 +146,12 @@ static int skl_format_to_drm(int format, bool rgb_order, bool alpha, > > static u32 intel_vgpu_get_stride(struct intel_vgpu *vgpu, int pipe, > > u32 tiled, int stride_mask, int bpp) > > { > > - struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv; > > + struct drm_i915_private *i915 = vgpu->gvt->i915; > > u32 stride_reg = vgpu_vreg_t(vgpu, DSPSTRIDE(pipe)) & stride_mask; > > u32 stride = stride_reg; > > - if (INTEL_GEN(dev_priv) >= 9) { > > + if (INTEL_GEN(i915) >= 9) { > > switch (tiled) { > > case PLANE_CTL_TILED_LINEAR: > > stride = stride_reg * 64; > > @@ -203,7 +203,7 @@ int intel_vgpu_decode_primary_plane(struct intel_vgpu *vgpu, > > struct intel_vgpu_primary_plane_format *plane) > > { > > u32 val, fmt; > > - struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv; > > + struct drm_i915_private *i915 = vgpu->gvt->i915; > > int pipe; > > pipe = get_active_pipe(vgpu); > > @@ -215,7 +215,7 @@ int intel_vgpu_decode_primary_plane(struct intel_vgpu *vgpu, > > if (!plane->enabled) > > return -ENODEV; > > - if (INTEL_GEN(dev_priv) >= 9) { > > + if (INTEL_GEN(i915) >= 9) { > > plane->tiled = val & PLANE_CTL_TILED_MASK; > > fmt = skl_format_to_drm( > > val & PLANE_CTL_FORMAT_MASK, > > @@ -256,7 +256,7 @@ int intel_vgpu_decode_primary_plane(struct intel_vgpu *vgpu, > > } > > plane->stride = intel_vgpu_get_stride(vgpu, pipe, plane->tiled, > > - (INTEL_GEN(dev_priv) >= 9) ? > > + (INTEL_GEN(i915) >= 9) ? > > (_PRI_PLANE_STRIDE_MASK >> 6) : > > _PRI_PLANE_STRIDE_MASK, plane->bpp); > > @@ -334,7 +334,7 @@ int intel_vgpu_decode_cursor_plane(struct intel_vgpu *vgpu, > > { > > u32 val, mode, index; > > u32 alpha_plane, alpha_force; > > - struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv; > > + struct drm_i915_private *i915 = vgpu->gvt->i915; > > int pipe; > > pipe = get_active_pipe(vgpu); > > diff --git a/drivers/gpu/drm/i915/gvt/firmware.c b/drivers/gpu/drm/i915/gvt/firmware.c > > index 049775e8e350..efdc9df4f731 100644 > > --- a/drivers/gpu/drm/i915/gvt/firmware.c > > +++ b/drivers/gpu/drm/i915/gvt/firmware.c > > @@ -68,7 +68,7 @@ static struct bin_attribute firmware_attr = { > > static int mmio_snapshot_handler(struct intel_gvt *gvt, u32 offset, void *data) > > { > > - struct drm_i915_private *i915 = gvt->dev_priv; > > + struct drm_i915_private *i915 = gvt->i915; > > *(u32 *)(data + offset) = intel_uncore_read_notrace(&i915->uncore, > > _MMIO(offset)); > > @@ -78,7 +78,7 @@ static int mmio_snapshot_handler(struct intel_gvt *gvt, u32 offset, void *data) > > static int expose_firmware_sysfs(struct intel_gvt *gvt) > > { > > struct intel_gvt_device_info *info = &gvt->device_info; > > - struct pci_dev *pdev = gvt->dev_priv->drm.pdev; > > + struct pci_dev *pdev = gvt->i915->drm.pdev; > > struct gvt_firmware_header *h; > > void *firmware; > > void *p; > > @@ -129,7 +129,7 @@ static int expose_firmware_sysfs(struct intel_gvt *gvt) > > static void clean_firmware_sysfs(struct intel_gvt *gvt) > > { > > - struct pci_dev *pdev = gvt->dev_priv->drm.pdev; > > + struct pci_dev *pdev = gvt->i915->drm.pdev; > > device_remove_bin_file(&pdev->dev, &firmware_attr); > > vfree(firmware_attr.private); > > @@ -153,8 +153,8 @@ static int verify_firmware(struct intel_gvt *gvt, > > const struct firmware *fw) > > { > > struct intel_gvt_device_info *info = &gvt->device_info; > > - struct drm_i915_private *dev_priv = gvt->dev_priv; > > - struct pci_dev *pdev = dev_priv->drm.pdev; > > + struct drm_i915_private *i915 = gvt->i915; > > + struct pci_dev *pdev = i915->drm.pdev; > > struct gvt_firmware_header *h; > > unsigned long id, crc32_start; > > const void *mem; > > @@ -208,8 +208,8 @@ static int verify_firmware(struct intel_gvt *gvt, > > int intel_gvt_load_firmware(struct intel_gvt *gvt) > > { > > struct intel_gvt_device_info *info = &gvt->device_info; > > - struct drm_i915_private *dev_priv = gvt->dev_priv; > > - struct pci_dev *pdev = dev_priv->drm.pdev; > > + struct drm_i915_private *i915 = gvt->i915; > > + struct pci_dev *pdev = i915->drm.pdev; > > struct intel_gvt_firmware *firmware = &gvt->firmware; > > struct gvt_firmware_header *h; > > const struct firmware *fw; > > @@ -244,7 +244,7 @@ int intel_gvt_load_firmware(struct intel_gvt *gvt) > > gvt_dbg_core("request hw state firmware %s...\n", path); > > - ret = request_firmware(&fw, path, &dev_priv->drm.pdev->dev); > > + ret = request_firmware(&fw, path, &i915->drm.pdev->dev); > > kfree(path); > > if (ret) > > diff --git a/drivers/gpu/drm/i915/gvt/gtt.c b/drivers/gpu/drm/i915/gvt/gtt.c > > index 53115bdae12b..390562fe3223 100644 > > --- a/drivers/gpu/drm/i915/gvt/gtt.c > > +++ b/drivers/gpu/drm/i915/gvt/gtt.c > > @@ -275,24 +275,24 @@ static inline int get_pse_type(int type) > > return gtt_type_table[type].pse_entry_type; > > } > > -static u64 read_pte64(struct drm_i915_private *dev_priv, unsigned long index) > > +static u64 read_pte64(struct drm_i915_private *i915, unsigned long index) > > { > > - void __iomem *addr = (gen8_pte_t __iomem *)dev_priv->ggtt.gsm + index; > > + void __iomem *addr = (gen8_pte_t __iomem *)i915->ggtt.gsm + index; > > return readq(addr); > > } > > -static void ggtt_invalidate(struct drm_i915_private *dev_priv) > > +static void ggtt_invalidate(struct drm_i915_private *i915) > > { > > - mmio_hw_access_pre(dev_priv); > > + mmio_hw_access_pre(i915); > > I915_WRITE(GFX_FLSH_CNTL_GEN6, GFX_FLSH_CNTL_EN); > > - mmio_hw_access_post(dev_priv); > > + mmio_hw_access_post(i915); > > } > > -static void write_pte64(struct drm_i915_private *dev_priv, > > +static void write_pte64(struct drm_i915_private *i915, > > unsigned long index, u64 pte) > > { > > - void __iomem *addr = (gen8_pte_t __iomem *)dev_priv->ggtt.gsm + index; > > + void __iomem *addr = (gen8_pte_t __iomem *)i915->ggtt.gsm + index; > > writeq(pte, addr); > > } > > @@ -315,7 +315,7 @@ static inline int gtt_get_entry64(void *pt, > > if (WARN_ON(ret)) > > return ret; > > } else if (!pt) { > > - e->val64 = read_pte64(vgpu->gvt->dev_priv, index); > > + e->val64 = read_pte64(vgpu->gvt->i915, index); > > } else { > > e->val64 = *((u64 *)pt + index); > > } > > @@ -340,7 +340,7 @@ static inline int gtt_set_entry64(void *pt, > > if (WARN_ON(ret)) > > return ret; > > } else if (!pt) { > > - write_pte64(vgpu->gvt->dev_priv, index, e->val64); > > + write_pte64(vgpu->gvt->i915, index, e->val64); > > } else { > > *((u64 *)pt + index) = e->val64; > > } > > @@ -734,7 +734,7 @@ static int detach_oos_page(struct intel_vgpu *vgpu, > > static void ppgtt_free_spt(struct intel_vgpu_ppgtt_spt *spt) > > { > > - struct device *kdev = &spt->vgpu->gvt->dev_priv->drm.pdev->dev; > > + struct device *kdev = &spt->vgpu->gvt->i915->drm.pdev->dev; > > trace_spt_free(spt->vgpu->id, spt, spt->guest_page.type); > > @@ -819,7 +819,7 @@ static int reclaim_one_ppgtt_mm(struct intel_gvt *gvt); > > static struct intel_vgpu_ppgtt_spt *ppgtt_alloc_spt( > > struct intel_vgpu *vgpu, enum intel_gvt_gtt_type type) > > { > > - struct device *kdev = &vgpu->gvt->dev_priv->drm.pdev->dev; > > + struct device *kdev = &vgpu->gvt->i915->drm.pdev->dev; > > struct intel_vgpu_ppgtt_spt *spt = NULL; > > dma_addr_t daddr; > > int ret; > > @@ -1044,14 +1044,14 @@ static int ppgtt_invalidate_spt(struct intel_vgpu_ppgtt_spt *spt) > > static bool vgpu_ips_enabled(struct intel_vgpu *vgpu) > > { > > - struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv; > > + struct drm_i915_private *i915 = vgpu->gvt->i915; > > - if (INTEL_GEN(dev_priv) == 9 || INTEL_GEN(dev_priv) == 10) { > > + if (INTEL_GEN(i915) == 9 || INTEL_GEN(i915) == 10) { > > u32 ips = vgpu_vreg_t(vgpu, GEN8_GAMW_ECO_DEV_RW_IA) & > > GAMW_ECO_ENABLE_64K_IPS_FIELD; > > return ips == GAMW_ECO_ENABLE_64K_IPS_FIELD; > > - } else if (INTEL_GEN(dev_priv) >= 11) { > > + } else if (INTEL_GEN(i915) >= 11) { > > /* 64K paging only controlled by IPS bit in PTE now. */ > > return true; > > } else > > @@ -1153,7 +1153,7 @@ static int is_2MB_gtt_possible(struct intel_vgpu *vgpu, > > struct intel_gvt_gtt_pte_ops *ops = vgpu->gvt->gtt.pte_ops; > > unsigned long pfn; > > - if (!HAS_PAGE_SIZES(vgpu->gvt->dev_priv, I915_GTT_PAGE_SIZE_2M)) > > + if (!HAS_PAGE_SIZES(vgpu->gvt->i915, I915_GTT_PAGE_SIZE_2M)) > > return 0; > > pfn = intel_gvt_hypervisor_gfn_to_mfn(vgpu, ops->get_pfn(entry)); > > @@ -2301,7 +2301,7 @@ static int emulate_ggtt_mmio_write(struct intel_vgpu *vgpu, unsigned int off, > > ggtt_invalidate_pte(vgpu, &e); > > ggtt_set_host_entry(ggtt_mm, &m, g_gtt_index); > > - ggtt_invalidate(gvt->dev_priv); > > + ggtt_invalidate(gvt->i915); > > return 0; > > } > > @@ -2340,7 +2340,7 @@ static int alloc_scratch_pages(struct intel_vgpu *vgpu, > > vgpu->gvt->device_info.gtt_entry_size_shift; > > void *scratch_pt; > > int i; > > - struct device *dev = &vgpu->gvt->dev_priv->drm.pdev->dev; > > + struct device *dev = &vgpu->gvt->i915->drm.pdev->dev; > > dma_addr_t daddr; > > if (WARN_ON(type < GTT_TYPE_PPGTT_PTE_PT || type >= GTT_TYPE_MAX)) > > @@ -2397,7 +2397,7 @@ static int alloc_scratch_pages(struct intel_vgpu *vgpu, > > static int release_scratch_page_tree(struct intel_vgpu *vgpu) > > { > > int i; > > - struct device *dev = &vgpu->gvt->dev_priv->drm.pdev->dev; > > + struct device *dev = &vgpu->gvt->i915->drm.pdev->dev; > > dma_addr_t daddr; > > for (i = GTT_TYPE_PPGTT_PTE_PT; i < GTT_TYPE_MAX; i++) { > > @@ -2669,7 +2669,7 @@ int intel_gvt_init_gtt(struct intel_gvt *gvt) > > { > > int ret; > > void *page; > > - struct device *dev = &gvt->dev_priv->drm.pdev->dev; > > + struct device *dev = &gvt->i915->drm.pdev->dev; > > dma_addr_t daddr; > > gvt_dbg_core("init gtt\n"); > > @@ -2718,7 +2718,7 @@ int intel_gvt_init_gtt(struct intel_gvt *gvt) > > */ > > void intel_gvt_clean_gtt(struct intel_gvt *gvt) > > { > > - struct device *dev = &gvt->dev_priv->drm.pdev->dev; > > + struct device *dev = &gvt->i915->drm.pdev->dev; > > dma_addr_t daddr = (dma_addr_t)(gvt->gtt.scratch_mfn << > > I915_GTT_PAGE_SHIFT); > > @@ -2766,7 +2766,7 @@ void intel_vgpu_invalidate_ppgtt(struct intel_vgpu *vgpu) > > void intel_vgpu_reset_ggtt(struct intel_vgpu *vgpu, bool invalidate_old) > > { > > struct intel_gvt *gvt = vgpu->gvt; > > - struct drm_i915_private *dev_priv = gvt->dev_priv; > > + struct drm_i915_private *i915 = gvt->i915; > > struct intel_gvt_gtt_pte_ops *pte_ops = vgpu->gvt->gtt.pte_ops; > > struct intel_gvt_gtt_entry entry = {.type = GTT_TYPE_GGTT_PTE}; > > struct intel_gvt_gtt_entry old_entry; > > @@ -2796,7 +2796,7 @@ void intel_vgpu_reset_ggtt(struct intel_vgpu *vgpu, bool invalidate_old) > > ggtt_set_host_entry(vgpu->gtt.ggtt_mm, &entry, index++); > > } > > - ggtt_invalidate(dev_priv); > > + ggtt_invalidate(i915); > > } > > /** > > diff --git a/drivers/gpu/drm/i915/gvt/gvt.c b/drivers/gpu/drm/i915/gvt/gvt.c > > index 43f4242062dd..b84cdd8d8c98 100644 > > --- a/drivers/gpu/drm/i915/gvt/gvt.c > > +++ b/drivers/gpu/drm/i915/gvt/gvt.c > > @@ -52,7 +52,7 @@ static struct intel_vgpu_type *intel_gvt_find_vgpu_type(struct intel_gvt *gvt, > > int i; > > struct intel_vgpu_type *t; > > const char *driver_name = dev_driver_string( > > - &gvt->dev_priv->drm.pdev->dev); > > + &gvt->i915->drm.pdev->dev); > > for (i = 0; i < gvt->num_types; i++) { > > t = &gvt->types[i]; > > @@ -191,7 +191,7 @@ static const struct intel_gvt_ops intel_gvt_ops = { > > static void init_device_info(struct intel_gvt *gvt) > > { > > struct intel_gvt_device_info *info = &gvt->device_info; > > - struct pci_dev *pdev = gvt->dev_priv->drm.pdev; > > + struct pci_dev *pdev = gvt->i915->drm.pdev; > > info->max_support_vgpus = 8; > > info->cfg_space_size = PCI_CFG_SPACE_EXP_SIZE; > > @@ -257,15 +257,15 @@ static int init_service_thread(struct intel_gvt *gvt) > > /** > > * intel_gvt_clean_device - clean a GVT device > > - * @dev_priv: i915 private > > + * @i915: i915 private > > * > > * This function is called at the driver unloading stage, to free the > > * resources owned by a GVT device. > > * > > */ > > -void intel_gvt_clean_device(struct drm_i915_private *dev_priv) > > +void intel_gvt_clean_device(struct drm_i915_private *i915) > > { > > - struct intel_gvt *gvt = to_gvt(dev_priv); > > + struct intel_gvt *gvt = to_gvt(i915); > > if (WARN_ON(!gvt)) > > return; > > @@ -285,13 +285,13 @@ void intel_gvt_clean_device(struct drm_i915_private *dev_priv) > > intel_gvt_clean_mmio_info(gvt); > > idr_destroy(&gvt->vgpu_idr); > > - kfree(dev_priv->gvt); > > - dev_priv->gvt = NULL; > > + kfree(i915->gvt); > > + i915->gvt = NULL; > > } > > /** > > * intel_gvt_init_device - initialize a GVT device > > - * @dev_priv: drm i915 private data > > + * @i915: drm i915 private data > > * > > * This function is called at the initialization stage, to initialize > > * necessary GVT components. > > @@ -300,13 +300,13 @@ void intel_gvt_clean_device(struct drm_i915_private *dev_priv) > > * Zero on success, negative error code if failed. > > * > > */ > > -int intel_gvt_init_device(struct drm_i915_private *dev_priv) > > +int intel_gvt_init_device(struct drm_i915_private *i915) > > { > > struct intel_gvt *gvt; > > struct intel_vgpu *vgpu; > > int ret; > > - if (WARN_ON(dev_priv->gvt)) > > + if (WARN_ON(i915->gvt)) > > return -EEXIST; > > gvt = kzalloc(sizeof(struct intel_gvt), GFP_KERNEL); > > @@ -319,7 +319,7 @@ int intel_gvt_init_device(struct drm_i915_private *dev_priv) > > spin_lock_init(&gvt->scheduler.mmio_context_lock); > > mutex_init(&gvt->lock); > > mutex_init(&gvt->sched_lock); > > - gvt->dev_priv = dev_priv; > > + gvt->i915 = i915; > > init_device_info(gvt); > > @@ -380,8 +380,8 @@ int intel_gvt_init_device(struct drm_i915_private *dev_priv) > > gvt_err("debugfs registration failed, go on.\n"); > > gvt_dbg_core("gvt device initialization is done\n"); > > - dev_priv->gvt = gvt; > > - intel_gvt_host.dev = &dev_priv->drm.pdev->dev; > > + i915->gvt = gvt; > > + intel_gvt_host.dev = &i915->drm.pdev->dev; > > intel_gvt_host.initialized = true; > > return 0; > > diff --git a/drivers/gpu/drm/i915/gvt/gvt.h b/drivers/gpu/drm/i915/gvt/gvt.h > > index b54f2bdc13a4..fbebe0ae4096 100644 > > --- a/drivers/gpu/drm/i915/gvt/gvt.h > > +++ b/drivers/gpu/drm/i915/gvt/gvt.h > > @@ -306,7 +306,7 @@ struct intel_gvt { > > /* scheduler scope lock, protect gvt and vgpu schedule related data */ > > struct mutex sched_lock; > > - struct drm_i915_private *dev_priv; > > + struct drm_i915_private *i915; > > struct idr vgpu_idr; /* vGPU IDR pool */ > > struct intel_gvt_device_info device_info; > > @@ -373,12 +373,12 @@ int intel_gvt_load_firmware(struct intel_gvt *gvt); > > #define HOST_FENCE 4 > > /* Aperture/GM space definitions for GVT device */ > > -#define gvt_aperture_sz(gvt) (gvt->dev_priv->ggtt.mappable_end) > > -#define gvt_aperture_pa_base(gvt) (gvt->dev_priv->ggtt.gmadr.start) > > +#define gvt_aperture_sz(gvt) (gvt->i915->ggtt.mappable_end) > > +#define gvt_aperture_pa_base(gvt) (gvt->i915->ggtt.gmadr.start) > > -#define gvt_ggtt_gm_sz(gvt) (gvt->dev_priv->ggtt.vm.total) > > +#define gvt_ggtt_gm_sz(gvt) (gvt->i915->ggtt.vm.total) > > #define gvt_ggtt_sz(gvt) \ > > - ((gvt->dev_priv->ggtt.vm.total >> PAGE_SHIFT) << 3) > > + ((gvt->i915->ggtt.vm.total >> PAGE_SHIFT) << 3) > > #define gvt_hidden_sz(gvt) (gvt_ggtt_gm_sz(gvt) - gvt_aperture_sz(gvt)) > > #define gvt_aperture_gmadr_base(gvt) (0) > > @@ -390,7 +390,7 @@ int intel_gvt_load_firmware(struct intel_gvt *gvt); > > #define gvt_hidden_gmadr_end(gvt) (gvt_hidden_gmadr_base(gvt) \ > > + gvt_hidden_sz(gvt) - 1) > > -#define gvt_fence_sz(gvt) (gvt->dev_priv->num_fence_regs) > > +#define gvt_fence_sz(gvt) (gvt->i915->num_fence_regs) > > /* Aperture/GM space definitions for vGPU */ > > #define vgpu_aperture_offset(vgpu) ((vgpu)->gm.low_gm_node.start) > > @@ -582,14 +582,14 @@ enum { > > GVT_FAILSAFE_GUEST_ERR, > > }; > > -static inline void mmio_hw_access_pre(struct drm_i915_private *dev_priv) > > +static inline void mmio_hw_access_pre(struct drm_i915_private *i915) > > { > > - intel_runtime_pm_get(dev_priv); > > + intel_runtime_pm_get(i915); > > } > > -static inline void mmio_hw_access_post(struct drm_i915_private *dev_priv) > > +static inline void mmio_hw_access_post(struct drm_i915_private *i915) > > { > > - intel_runtime_pm_put_unchecked(dev_priv); > > + intel_runtime_pm_put_unchecked(i915); > > } > > /** > > diff --git a/drivers/gpu/drm/i915/gvt/handlers.c b/drivers/gpu/drm/i915/gvt/handlers.c > > index a6ade66349bd..9acb01b0b7c7 100644 > > --- a/drivers/gpu/drm/i915/gvt/handlers.c > > +++ b/drivers/gpu/drm/i915/gvt/handlers.c > > @@ -49,15 +49,15 @@ > > unsigned long intel_gvt_get_device_type(struct intel_gvt *gvt) > > { > > - if (IS_BROADWELL(gvt->dev_priv)) > > + if (IS_BROADWELL(gvt->i915)) > > return D_BDW; > > - else if (IS_SKYLAKE(gvt->dev_priv)) > > + else if (IS_SKYLAKE(gvt->i915)) > > return D_SKL; > > - else if (IS_KABYLAKE(gvt->dev_priv)) > > + else if (IS_KABYLAKE(gvt->i915)) > > return D_KBL; > > - else if (IS_BROXTON(gvt->dev_priv)) > > + else if (IS_BROXTON(gvt->i915)) > > return D_BXT; > > - else if (IS_COFFEELAKE(gvt->dev_priv)) > > + else if (IS_COFFEELAKE(gvt->i915)) > > return D_CFL; > > return 0; > > @@ -156,7 +156,7 @@ int intel_gvt_render_mmio_to_ring_id(struct intel_gvt *gvt, > > struct intel_engine_cs *engine; > > offset &= ~GENMASK(11, 0); > > - for_each_engine(engine, gvt->dev_priv, id) { > > + for_each_engine(engine, gvt->i915, id) { > > if (engine->mmio_base == offset) > > return id; > > } > > @@ -217,7 +217,7 @@ static int gamw_echo_dev_rw_ia_write(struct intel_vgpu *vgpu, > > { > > u32 ips = (*(u32 *)p_data) & GAMW_ECO_ENABLE_64K_IPS_FIELD; > > - if (INTEL_GEN(vgpu->gvt->dev_priv) <= 10) { > > + if (INTEL_GEN(vgpu->gvt->i915) <= 10) { > > if (ips == GAMW_ECO_ENABLE_64K_IPS_FIELD) > > gvt_dbg_core("vgpu%d: ips enabled\n", vgpu->id); > > else if (!ips) > > @@ -253,7 +253,7 @@ static int fence_mmio_read(struct intel_vgpu *vgpu, unsigned int off, > > static int fence_mmio_write(struct intel_vgpu *vgpu, unsigned int off, > > void *p_data, unsigned int bytes) > > { > > - struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv; > > + struct drm_i915_private *i915 = vgpu->gvt->i915; > > unsigned int fence_num = offset_to_fence_num(off); > > int ret; > > @@ -262,10 +262,10 @@ static int fence_mmio_write(struct intel_vgpu *vgpu, unsigned int off, > > return ret; > > write_vreg(vgpu, off, p_data, bytes); > > - mmio_hw_access_pre(dev_priv); > > + mmio_hw_access_pre(i915); > > intel_vgpu_write_fence(vgpu, fence_num, > > vgpu_vreg64(vgpu, fence_num_to_offset(fence_num))); > > - mmio_hw_access_post(dev_priv); > > + mmio_hw_access_post(i915); > > return 0; > > } > > @@ -283,7 +283,7 @@ static int mul_force_wake_write(struct intel_vgpu *vgpu, > > old = vgpu_vreg(vgpu, offset); > > new = CALC_MODE_MASK_REG(old, *(u32 *)p_data); > > - if (INTEL_GEN(vgpu->gvt->dev_priv) >= 9) { > > + if (INTEL_GEN(vgpu->gvt->i915) >= 9) { > > switch (offset) { > > case FORCEWAKE_RENDER_GEN9_REG: > > ack_reg_offset = FORCEWAKE_ACK_RENDER_GEN9_REG; > > @@ -341,7 +341,7 @@ static int gdrst_mmio_write(struct intel_vgpu *vgpu, unsigned int offset, > > gvt_dbg_mmio("vgpu%d: request VCS2 Reset\n", vgpu->id); > > engine_mask |= BIT(VCS1); > > } > > - engine_mask &= INTEL_INFO(vgpu->gvt->dev_priv)->engine_mask; > > + engine_mask &= INTEL_INFO(vgpu->gvt->i915)->engine_mask; > > } > > /* vgpu_lock already hold by emulate mmio r/w */ > > @@ -511,7 +511,7 @@ static int force_nonpriv_write(struct intel_vgpu *vgpu, > > u32 reg_nonpriv = *(u32 *)p_data; > > int ring_id = intel_gvt_render_mmio_to_ring_id(vgpu->gvt, offset); > > u32 ring_base; > > - struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv; > > + struct drm_i915_private *i915 = vgpu->gvt->i915; > > int ret = -EINVAL; > > if ((bytes != 4) || ((offset & (bytes - 1)) != 0) || ring_id < 0) { > > @@ -520,7 +520,7 @@ static int force_nonpriv_write(struct intel_vgpu *vgpu, > > return ret; > > } > > - ring_base = dev_priv->engine[ring_id]->mmio_base; > > + ring_base = i915->engine[ring_id]->mmio_base; > > if (in_whitelist(reg_nonpriv) || > > reg_nonpriv == i915_mmio_reg_offset(RING_NOPID(ring_base))) { > > @@ -751,7 +751,7 @@ static int south_chicken2_mmio_write(struct intel_vgpu *vgpu, > > static int pri_surf_mmio_write(struct intel_vgpu *vgpu, unsigned int offset, > > void *p_data, unsigned int bytes) > > { > > - struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv; > > + struct drm_i915_private *i915 = vgpu->gvt->i915; > > u32 pipe = DSPSURF_TO_PIPE(offset); > > int event = SKL_FLIP_EVENT(pipe, PLANE_PRIMARY); > > @@ -792,7 +792,7 @@ static int reg50080_mmio_write(struct intel_vgpu *vgpu, > > unsigned int offset, void *p_data, > > unsigned int bytes) > > { > > - struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv; > > + struct drm_i915_private *i915 = vgpu->gvt->i915; > > enum pipe pipe = REG_50080_TO_PIPE(offset); > > enum plane_id plane = REG_50080_TO_PLANE(offset); > > int event = SKL_FLIP_EVENT(pipe, plane); > > @@ -816,7 +816,7 @@ static int reg50080_mmio_write(struct intel_vgpu *vgpu, > > static int trigger_aux_channel_interrupt(struct intel_vgpu *vgpu, > > unsigned int reg) > > { > > - struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv; > > + struct drm_i915_private *i915 = vgpu->gvt->i915; > > enum intel_gvt_event_type event; > > if (reg == _DPA_AUX_CH_CTL) > > @@ -916,11 +916,11 @@ static int dp_aux_ch_ctl_mmio_write(struct intel_vgpu *vgpu, > > write_vreg(vgpu, offset, p_data, bytes); > > data = vgpu_vreg(vgpu, offset); > > - if ((INTEL_GEN(vgpu->gvt->dev_priv) >= 9) > > + if ((INTEL_GEN(vgpu->gvt->i915) >= 9) > > && offset != _REG_SKL_DP_AUX_CH_CTL(port_index)) { > > /* SKL DPB/C/D aux ctl register changed */ > > return 0; > > - } else if (IS_BROADWELL(vgpu->gvt->dev_priv) && > > + } else if (IS_BROADWELL(vgpu->gvt->i915) && > > offset != _REG_HSW_DP_AUX_CH_CTL(port_index)) { > > /* write to the data registers */ > > return 0; > > @@ -1236,8 +1236,8 @@ static int handle_g2v_notification(struct intel_vgpu *vgpu, int notification) > > static int send_display_ready_uevent(struct intel_vgpu *vgpu, int ready) > > { > > - struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv; > > - struct kobject *kobj = &dev_priv->drm.primary->kdev->kobj; > > + struct drm_i915_private *i915 = vgpu->gvt->i915; > > + struct kobject *kobj = &i915->drm.primary->kdev->kobj; > > char *env[3] = {NULL, NULL, NULL}; > > char vmid_str[20]; > > char display_ready_str[20]; > > @@ -1416,9 +1416,9 @@ static int mailbox_write(struct intel_vgpu *vgpu, unsigned int offset, > > switch (cmd) { > > case GEN9_PCODE_READ_MEM_LATENCY: > > - if (IS_SKYLAKE(vgpu->gvt->dev_priv) > > - || IS_KABYLAKE(vgpu->gvt->dev_priv) > > - || IS_COFFEELAKE(vgpu->gvt->dev_priv)) { > > + if (IS_SKYLAKE(vgpu->gvt->i915) > > + || IS_KABYLAKE(vgpu->gvt->i915) > > + || IS_COFFEELAKE(vgpu->gvt->i915)) { > > /** > > * "Read memory latency" command on gen9. > > * Below memory latency values are read > > @@ -1428,7 +1428,7 @@ static int mailbox_write(struct intel_vgpu *vgpu, unsigned int offset, > > *data0 = 0x1e1a1100; > > else > > *data0 = 0x61514b3d; > > - } else if (IS_BROXTON(vgpu->gvt->dev_priv)) { > > + } else if (IS_BROXTON(vgpu->gvt->i915)) { > > /** > > * "Read memory latency" command on gen9. > > * Below memory latency values are read > > @@ -1441,9 +1441,9 @@ static int mailbox_write(struct intel_vgpu *vgpu, unsigned int offset, > > } > > break; > > case SKL_PCODE_CDCLK_CONTROL: > > - if (IS_SKYLAKE(vgpu->gvt->dev_priv) > > - || IS_KABYLAKE(vgpu->gvt->dev_priv) > > - || IS_COFFEELAKE(vgpu->gvt->dev_priv)) > > + if (IS_SKYLAKE(vgpu->gvt->i915) > > + || IS_KABYLAKE(vgpu->gvt->i915) > > + || IS_COFFEELAKE(vgpu->gvt->i915)) > > *data0 = SKL_CDCLK_READY_FOR_CHANGE; > > break; > > case GEN6_PCODE_READ_RC6VIDS: > > @@ -1496,7 +1496,7 @@ static int skl_power_well_ctl_write(struct intel_vgpu *vgpu, > > { > > u32 v = *(u32 *)p_data; > > - if (IS_BROXTON(vgpu->gvt->dev_priv)) > > + if (IS_BROXTON(vgpu->gvt->i915)) > > v &= (1 << 31) | (1 << 29); > > else > > v &= (1 << 31) | (1 << 29) | (1 << 9) | > > @@ -1633,7 +1633,7 @@ static int mmio_read_from_hw(struct intel_vgpu *vgpu, > > unsigned int offset, void *p_data, unsigned int bytes) > > { > > struct intel_gvt *gvt = vgpu->gvt; > > - struct drm_i915_private *dev_priv = gvt->dev_priv; > > + struct drm_i915_private *i915 = gvt->i915; > > int ring_id; > > u32 ring_base; > > @@ -1645,14 +1645,14 @@ static int mmio_read_from_hw(struct intel_vgpu *vgpu, > > * c. the offset is ring time stamp mmio > > */ > > if (ring_id >= 0) > > - ring_base = dev_priv->engine[ring_id]->mmio_base; > > + ring_base = i915->engine[ring_id]->mmio_base; > > if (ring_id < 0 || vgpu == gvt->scheduler.engine_owner[ring_id] || > > offset == i915_mmio_reg_offset(RING_TIMESTAMP(ring_base)) || > > offset == i915_mmio_reg_offset(RING_TIMESTAMP_UDW(ring_base))) { > > - mmio_hw_access_pre(dev_priv); > > + mmio_hw_access_pre(i915); > > vgpu_vreg(vgpu, offset) = I915_READ(_MMIO(offset)); > > - mmio_hw_access_post(dev_priv); > > + mmio_hw_access_post(i915); > > } > > return intel_vgpu_default_mmio_read(vgpu, offset, p_data, bytes); > > @@ -1693,7 +1693,7 @@ static int ring_mode_mmio_write(struct intel_vgpu *vgpu, unsigned int offset, > > int ret; > > (*(u32 *)p_data) &= ~_MASKED_BIT_ENABLE(1); > > - if (IS_COFFEELAKE(vgpu->gvt->dev_priv)) > > + if (IS_COFFEELAKE(vgpu->gvt->i915)) > > (*(u32 *)p_data) &= ~_MASKED_BIT_ENABLE(2); > > write_vreg(vgpu, offset, p_data, bytes); > > @@ -1702,7 +1702,7 @@ static int ring_mode_mmio_write(struct intel_vgpu *vgpu, unsigned int offset, > > return 0; > > } > > - if (IS_COFFEELAKE(vgpu->gvt->dev_priv) && > > + if (IS_COFFEELAKE(vgpu->gvt->i915) && > > data & _MASKED_BIT_ENABLE(2)) { > > enter_failsafe_mode(vgpu, GVT_FAILSAFE_UNSUPPORTED_GUEST); > > return 0; > > @@ -1834,7 +1834,7 @@ static int csfe_chicken1_mmio_write(struct intel_vgpu *vgpu, > > MMIO_F(prefix(BLT_RING_BASE), s, f, am, rm, d, r, w); \ > > MMIO_F(prefix(GEN6_BSD_RING_BASE), s, f, am, rm, d, r, w); \ > > MMIO_F(prefix(VEBOX_RING_BASE), s, f, am, rm, d, r, w); \ > > - if (HAS_ENGINE(dev_priv, VCS1)) \ > > + if (HAS_ENGINE(i915, VCS1)) \ > > MMIO_F(prefix(GEN8_BSD2_RING_BASE), s, f, am, rm, d, r, w); \ > > } while (0) > > @@ -1855,7 +1855,7 @@ static int csfe_chicken1_mmio_write(struct intel_vgpu *vgpu, > > static int init_generic_mmio_info(struct intel_gvt *gvt) > > { > > - struct drm_i915_private *dev_priv = gvt->dev_priv; > > + struct drm_i915_private *i915 = gvt->i915; > > int ret; > > MMIO_RING_DFH(RING_IMR, D_ALL, F_CMD_ACCESS, NULL, > > @@ -2670,7 +2670,7 @@ static int init_generic_mmio_info(struct intel_gvt *gvt) > > static int init_broadwell_mmio_info(struct intel_gvt *gvt) > > { > > - struct drm_i915_private *dev_priv = gvt->dev_priv; > > + struct drm_i915_private *i915 = gvt->i915; > > int ret; > > MMIO_DH(GEN8_GT_IMR(0), D_BDW_PLUS, NULL, intel_vgpu_reg_imr_handler); > > @@ -2859,7 +2859,7 @@ static int init_broadwell_mmio_info(struct intel_gvt *gvt) > > static int init_skl_mmio_info(struct intel_gvt *gvt) > > { > > - struct drm_i915_private *dev_priv = gvt->dev_priv; > > + struct drm_i915_private *i915 = gvt->i915; > > int ret; > > MMIO_DH(FORCEWAKE_RENDER_GEN9, D_SKL_PLUS, NULL, mul_force_wake_write); > > @@ -3108,7 +3108,7 @@ static int init_skl_mmio_info(struct intel_gvt *gvt) > > static int init_bxt_mmio_info(struct intel_gvt *gvt) > > { > > - struct drm_i915_private *dev_priv = gvt->dev_priv; > > + struct drm_i915_private *i915 = gvt->i915; > > int ret; > > MMIO_F(_MMIO(0x80000), 0x3000, 0, 0, 0, D_BXT, NULL, NULL); > > @@ -3344,7 +3344,7 @@ static struct gvt_mmio_block mmio_blocks[] = { > > int intel_gvt_setup_mmio_info(struct intel_gvt *gvt) > > { > > struct intel_gvt_device_info *info = &gvt->device_info; > > - struct drm_i915_private *dev_priv = gvt->dev_priv; > > + struct drm_i915_private *i915 = gvt->i915; > > int size = info->mmio_size / 4 * sizeof(*gvt->mmio.mmio_attribute); > > int ret; > > @@ -3356,20 +3356,20 @@ int intel_gvt_setup_mmio_info(struct intel_gvt *gvt) > > if (ret) > > goto err; > > - if (IS_BROADWELL(dev_priv)) { > > + if (IS_BROADWELL(i915)) { > > ret = init_broadwell_mmio_info(gvt); > > if (ret) > > goto err; > > - } else if (IS_SKYLAKE(dev_priv) > > - || IS_KABYLAKE(dev_priv) > > - || IS_COFFEELAKE(dev_priv)) { > > + } else if (IS_SKYLAKE(i915) > > + || IS_KABYLAKE(i915) > > + || IS_COFFEELAKE(i915)) { > > ret = init_broadwell_mmio_info(gvt); > > if (ret) > > goto err; > > ret = init_skl_mmio_info(gvt); > > if (ret) > > goto err; > > - } else if (IS_BROXTON(dev_priv)) { > > + } else if (IS_BROXTON(i915)) { > > ret = init_broadwell_mmio_info(gvt); > > if (ret) > > goto err; > > diff --git a/drivers/gpu/drm/i915/gvt/interrupt.c b/drivers/gpu/drm/i915/gvt/interrupt.c > > index 951681813230..37e2e72bfacc 100644 > > --- a/drivers/gpu/drm/i915/gvt/interrupt.c > > +++ b/drivers/gpu/drm/i915/gvt/interrupt.c > > @@ -536,7 +536,7 @@ static void gen8_init_irq( > > SET_BIT_INFO(irq, 4, VCS_MI_FLUSH_DW, INTEL_GVT_IRQ_INFO_GT1); > > SET_BIT_INFO(irq, 8, VCS_AS_CONTEXT_SWITCH, INTEL_GVT_IRQ_INFO_GT1); > > - if (HAS_ENGINE(gvt->dev_priv, VCS1)) { > > + if (HAS_ENGINE(gvt->i915, VCS1)) { > > SET_BIT_INFO(irq, 16, VCS2_MI_USER_INTERRUPT, > > INTEL_GVT_IRQ_INFO_GT1); > > SET_BIT_INFO(irq, 20, VCS2_MI_FLUSH_DW, > > @@ -568,7 +568,7 @@ static void gen8_init_irq( > > SET_BIT_INFO(irq, 22, DP_C_HOTPLUG, INTEL_GVT_IRQ_INFO_PCH); > > SET_BIT_INFO(irq, 23, DP_D_HOTPLUG, INTEL_GVT_IRQ_INFO_PCH); > > - if (IS_BROADWELL(gvt->dev_priv)) { > > + if (IS_BROADWELL(gvt->i915)) { > > SET_BIT_INFO(irq, 25, AUX_CHANNEL_B, INTEL_GVT_IRQ_INFO_PCH); > > SET_BIT_INFO(irq, 26, AUX_CHANNEL_C, INTEL_GVT_IRQ_INFO_PCH); > > SET_BIT_INFO(irq, 27, AUX_CHANNEL_D, INTEL_GVT_IRQ_INFO_PCH); > > @@ -581,7 +581,7 @@ static void gen8_init_irq( > > SET_BIT_INFO(irq, 4, PRIMARY_C_FLIP_DONE, INTEL_GVT_IRQ_INFO_DE_PIPE_C); > > SET_BIT_INFO(irq, 5, SPRITE_C_FLIP_DONE, INTEL_GVT_IRQ_INFO_DE_PIPE_C); > > - } else if (INTEL_GEN(gvt->dev_priv) >= 9) { > > + } else if (INTEL_GEN(gvt->i915) >= 9) { > > SET_BIT_INFO(irq, 25, AUX_CHANNEL_B, INTEL_GVT_IRQ_INFO_DE_PORT); > > SET_BIT_INFO(irq, 26, AUX_CHANNEL_C, INTEL_GVT_IRQ_INFO_DE_PORT); > > SET_BIT_INFO(irq, 27, AUX_CHANNEL_D, INTEL_GVT_IRQ_INFO_DE_PORT); > > diff --git a/drivers/gpu/drm/i915/gvt/kvmgt.c b/drivers/gpu/drm/i915/gvt/kvmgt.c > > index 144301b778df..615a7959b762 100644 > > --- a/drivers/gpu/drm/i915/gvt/kvmgt.c > > +++ b/drivers/gpu/drm/i915/gvt/kvmgt.c > > @@ -187,7 +187,7 @@ static int gvt_pin_guest_page(struct intel_vgpu *vgpu, unsigned long gfn, > > static int gvt_dma_map_page(struct intel_vgpu *vgpu, unsigned long gfn, > > dma_addr_t *dma_addr, unsigned long size) > > { > > - struct device *dev = &vgpu->gvt->dev_priv->drm.pdev->dev; > > + struct device *dev = &vgpu->gvt->i915->drm.pdev->dev; > > struct page *page = NULL; > > int ret; > > @@ -210,7 +210,7 @@ static int gvt_dma_map_page(struct intel_vgpu *vgpu, unsigned long gfn, > > static void gvt_dma_unmap_page(struct intel_vgpu *vgpu, unsigned long gfn, > > dma_addr_t dma_addr, unsigned long size) > > { > > - struct device *dev = &vgpu->gvt->dev_priv->drm.pdev->dev; > > + struct device *dev = &vgpu->gvt->i915->drm.pdev->dev; > > dma_unmap_page(dev, dma_addr, size, PCI_DMA_BIDIRECTIONAL); > > gvt_unpin_guest_page(vgpu, gfn, size); > > @@ -913,7 +913,7 @@ static int intel_vgpu_aperture_rw(struct intel_vgpu *vgpu, u64 off, > > return -EINVAL; > > } > > - aperture_va = io_mapping_map_wc(&vgpu->gvt->dev_priv->ggtt.iomap, > > + aperture_va = io_mapping_map_wc(&vgpu->gvt->i915->ggtt.iomap, > > ALIGN_DOWN(off, PAGE_SIZE), > > count + offset_in_page(off)); > > if (!aperture_va) > > diff --git a/drivers/gpu/drm/i915/gvt/mmio.c b/drivers/gpu/drm/i915/gvt/mmio.c > > index a55178884d67..572b8328af61 100644 > > --- a/drivers/gpu/drm/i915/gvt/mmio.c > > +++ b/drivers/gpu/drm/i915/gvt/mmio.c > > @@ -245,7 +245,7 @@ void intel_vgpu_reset_mmio(struct intel_vgpu *vgpu, bool dmlr) > > /* set the bit 0:2(Core C-State ) to C0 */ > > vgpu_vreg_t(vgpu, GEN6_GT_CORE_STATUS) = 0; > > - if (IS_BROXTON(vgpu->gvt->dev_priv)) { > > + if (IS_BROXTON(vgpu->gvt->i915)) { > > vgpu_vreg_t(vgpu, BXT_P_CR_GT_DISP_PWRON) &= > > ~(BIT(0) | BIT(1)); > > vgpu_vreg_t(vgpu, BXT_PORT_CL1CM_DW0(DPIO_PHY0)) &= > > diff --git a/drivers/gpu/drm/i915/gvt/mmio_context.c b/drivers/gpu/drm/i915/gvt/mmio_context.c > > index 2998999e8568..44fd6a0a698c 100644 > > --- a/drivers/gpu/drm/i915/gvt/mmio_context.c > > +++ b/drivers/gpu/drm/i915/gvt/mmio_context.c > > @@ -148,7 +148,7 @@ static struct { > > u32 l3cc_table[GEN9_MOCS_SIZE / 2]; > > } gen9_render_mocs; > > -static void load_render_mocs(struct drm_i915_private *dev_priv) > > +static void load_render_mocs(struct drm_i915_private *i915) > > { > > i915_reg_t offset; > > u32 regs[] = { > > @@ -161,7 +161,7 @@ static void load_render_mocs(struct drm_i915_private *dev_priv) > > int ring_id, i; > > for (ring_id = 0; ring_id < ARRAY_SIZE(regs); ring_id++) { > > - if (!HAS_ENGINE(dev_priv, ring_id)) > > + if (!HAS_ENGINE(i915, ring_id)) > > continue; > > offset.reg = regs[ring_id]; > > for (i = 0; i < GEN9_MOCS_SIZE; i++) { > > @@ -329,8 +329,8 @@ int intel_vgpu_restore_inhibit_context(struct intel_vgpu *vgpu, > > static void handle_tlb_pending_event(struct intel_vgpu *vgpu, int ring_id) > > { > > - struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv; > > - struct intel_uncore *uncore = &dev_priv->uncore; > > + struct drm_i915_private *i915 = vgpu->gvt->i915; > > + struct intel_uncore *uncore = &i915->uncore; > > struct intel_vgpu_submission *s = &vgpu->submission; > > enum forcewake_domains fw; > > i915_reg_t reg; > > @@ -357,7 +357,7 @@ static void handle_tlb_pending_event(struct intel_vgpu *vgpu, int ring_id) > > */ > > fw = intel_uncore_forcewake_for_reg(uncore, reg, > > FW_REG_READ | FW_REG_WRITE); > > - if (ring_id == RCS0 && INTEL_GEN(dev_priv) >= 9) > > + if (ring_id == RCS0 && INTEL_GEN(i915) >= 9) > > fw |= FORCEWAKE_RENDER; > > intel_uncore_forcewake_get(uncore, fw); > > @@ -377,7 +377,7 @@ static void handle_tlb_pending_event(struct intel_vgpu *vgpu, int ring_id) > > static void switch_mocs(struct intel_vgpu *pre, struct intel_vgpu *next, > > int ring_id) > > { > > - struct drm_i915_private *dev_priv; > > + struct drm_i915_private *i915; > > i915_reg_t offset, l3_offset; > > u32 old_v, new_v; > > @@ -390,15 +390,15 @@ static void switch_mocs(struct intel_vgpu *pre, struct intel_vgpu *next, > > }; > > int i; > > - dev_priv = pre ? pre->gvt->dev_priv : next->gvt->dev_priv; > > + i915 = pre ? pre->gvt->i915 : next->gvt->i915; > > if (WARN_ON(ring_id >= ARRAY_SIZE(regs))) > > return; > > - if (ring_id == RCS0 && IS_GEN(dev_priv, 9)) > > + if (ring_id == RCS0 && IS_GEN(i915, 9)) > > return; > > if (!pre && !gen9_render_mocs.initialized) > > - load_render_mocs(dev_priv); > > + load_render_mocs(i915); > > offset.reg = regs[ring_id]; > > for (i = 0; i < GEN9_MOCS_SIZE; i++) { > > @@ -454,16 +454,16 @@ static void switch_mmio(struct intel_vgpu *pre, > > struct intel_vgpu *next, > > int ring_id) > > { > > - struct drm_i915_private *dev_priv; > > + struct drm_i915_private *i915; > > struct intel_vgpu_submission *s; > > struct engine_mmio *mmio; > > u32 old_v, new_v; > > - dev_priv = pre ? pre->gvt->dev_priv : next->gvt->dev_priv; > > - if (INTEL_GEN(dev_priv) >= 9) > > + i915 = pre ? pre->gvt->i915 : next->gvt->i915; > > + if (INTEL_GEN(i915) >= 9) > > switch_mocs(pre, next, ring_id); > > - for (mmio = dev_priv->gvt->engine_mmio_list.mmio; > > + for (mmio = i915->gvt->engine_mmio_list.mmio; > > i915_mmio_reg_valid(mmio->reg); mmio++) { > > if (mmio->ring_id != ring_id) > > continue; > > @@ -472,7 +472,7 @@ static void switch_mmio(struct intel_vgpu *pre, > > * state image on gen9, it's initialized by lri command and > > * save or restore with context together. > > */ > > - if (IS_GEN(dev_priv, 9) && mmio->in_context) > > + if (IS_GEN(i915, 9) && mmio->in_context) > > continue; > > // save > > @@ -536,7 +536,7 @@ static void switch_mmio(struct intel_vgpu *pre, > > void intel_gvt_switch_mmio(struct intel_vgpu *pre, > > struct intel_vgpu *next, int ring_id) > > { > > - struct drm_i915_private *dev_priv; > > + struct drm_i915_private *i915; > > if (WARN_ON(!pre && !next)) > > return; > > @@ -544,16 +544,16 @@ void intel_gvt_switch_mmio(struct intel_vgpu *pre, > > gvt_dbg_render("switch ring %d from %s to %s\n", ring_id, > > pre ? "vGPU" : "host", next ? "vGPU" : "HOST"); > > - dev_priv = pre ? pre->gvt->dev_priv : next->gvt->dev_priv; > > + i915 = pre ? pre->gvt->i915 : next->gvt->i915; > > /** > > * We are using raw mmio access wrapper to improve the > > * performace for batch mmio read/write, so we need > > * handle forcewake mannually. > > */ > > - intel_uncore_forcewake_get(&dev_priv->uncore, FORCEWAKE_ALL); > > + intel_uncore_forcewake_get(&i915->uncore, FORCEWAKE_ALL); > > switch_mmio(pre, next, ring_id); > > - intel_uncore_forcewake_put(&dev_priv->uncore, FORCEWAKE_ALL); > > + intel_uncore_forcewake_put(&i915->uncore, FORCEWAKE_ALL); > > } > > /** > > @@ -565,7 +565,7 @@ void intel_gvt_init_engine_mmio_context(struct intel_gvt *gvt) > > { > > struct engine_mmio *mmio; > > - if (INTEL_GEN(gvt->dev_priv) >= 9) > > + if (INTEL_GEN(gvt->i915) >= 9) > > gvt->engine_mmio_list.mmio = gen9_engine_mmio_list; > > else > > gvt->engine_mmio_list.mmio = gen8_engine_mmio_list; > > diff --git a/drivers/gpu/drm/i915/gvt/sched_policy.c b/drivers/gpu/drm/i915/gvt/sched_policy.c > > index 1c763a27a412..64a5d4688c89 100644 > > --- a/drivers/gpu/drm/i915/gvt/sched_policy.c > > +++ b/drivers/gpu/drm/i915/gvt/sched_policy.c > > @@ -39,7 +39,7 @@ static bool vgpu_has_pending_workload(struct intel_vgpu *vgpu) > > enum intel_engine_id i; > > struct intel_engine_cs *engine; > > - for_each_engine(engine, vgpu->gvt->dev_priv, i) { > > + for_each_engine(engine, vgpu->gvt->i915, i) { > > if (!list_empty(workload_q_head(vgpu, i))) > > return true; > > } > > @@ -152,7 +152,7 @@ static void try_to_schedule_next_vgpu(struct intel_gvt *gvt) > > scheduler->need_reschedule = true; > > /* still have uncompleted workload? */ > > - for_each_engine(engine, gvt->dev_priv, i) { > > + for_each_engine(engine, gvt->i915, i) { > > if (scheduler->current_workload[i]) > > return; > > } > > @@ -169,7 +169,7 @@ static void try_to_schedule_next_vgpu(struct intel_gvt *gvt) > > scheduler->need_reschedule = false; > > /* wake up workload dispatch thread */ > > - for_each_engine(engine, gvt->dev_priv, i) > > + for_each_engine(engine, gvt->i915, i) > > wake_up(&scheduler->waitq[i]); > > } > > @@ -446,7 +446,7 @@ void intel_vgpu_stop_schedule(struct intel_vgpu *vgpu) > > &vgpu->gvt->scheduler; > > int ring_id; > > struct vgpu_sched_data *vgpu_data = vgpu->sched_data; > > - struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv; > > + struct drm_i915_private *i915 = vgpu->gvt->i915; > > if (!vgpu_data->active) > > return; > > @@ -465,7 +465,7 @@ void intel_vgpu_stop_schedule(struct intel_vgpu *vgpu) > > scheduler->current_vgpu = NULL; > > } > > - intel_runtime_pm_get(dev_priv); > > + intel_runtime_pm_get(i915); > > spin_lock_bh(&scheduler->mmio_context_lock); > > for (ring_id = 0; ring_id < I915_NUM_ENGINES; ring_id++) { > > if (scheduler->engine_owner[ring_id] == vgpu) { > > @@ -474,6 +474,6 @@ void intel_vgpu_stop_schedule(struct intel_vgpu *vgpu) > > } > > } > > spin_unlock_bh(&scheduler->mmio_context_lock); > > - intel_runtime_pm_put_unchecked(dev_priv); > > + intel_runtime_pm_put_unchecked(i915); > > mutex_unlock(&vgpu->gvt->sched_lock); > > } > > diff --git a/drivers/gpu/drm/i915/gvt/scheduler.c b/drivers/gpu/drm/i915/gvt/scheduler.c > > index e301efb18d45..ec7299d8a738 100644 > > --- a/drivers/gpu/drm/i915/gvt/scheduler.c > > +++ b/drivers/gpu/drm/i915/gvt/scheduler.c > > @@ -83,9 +83,9 @@ static void update_shadow_pdps(struct intel_vgpu_workload *workload) > > static void sr_oa_regs(struct intel_vgpu_workload *workload, > > u32 *reg_state, bool save) > > { > > - struct drm_i915_private *dev_priv = workload->vgpu->gvt->dev_priv; > > - u32 ctx_oactxctrl = dev_priv->perf.oa.ctx_oactxctrl_offset; > > - u32 ctx_flexeu0 = dev_priv->perf.oa.ctx_flexeu0_offset; > > + struct drm_i915_private *i915 = workload->vgpu->gvt->i915; > > + u32 ctx_oactxctrl = i915->perf.oa.ctx_oactxctrl_offset; > > + u32 ctx_flexeu0 = i915->perf.oa.ctx_flexeu0_offset; > > int i = 0; > > u32 flex_mmio[] = { > > i915_mmio_reg_offset(EU_PERF_CNTL0), > > @@ -177,11 +177,11 @@ static int populate_shadow_context(struct intel_vgpu_workload *workload) > > gvt_dbg_sched("ring id %d workload lrca %x", ring_id, > > workload->ctx_desc.lrca); > > - context_page_num = gvt->dev_priv->engine[ring_id]->context_size; > > + context_page_num = gvt->i915->engine[ring_id]->context_size; > > context_page_num = context_page_num >> PAGE_SHIFT; > > - if (IS_BROADWELL(gvt->dev_priv) && ring_id == RCS0) > > + if (IS_BROADWELL(gvt->i915) && ring_id == RCS0) > > context_page_num = 19; > > i = 2; > > @@ -211,8 +211,8 @@ static inline bool is_gvt_request(struct i915_request *req) > > static void save_ring_hw_state(struct intel_vgpu *vgpu, int ring_id) > > { > > - struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv; > > - u32 ring_base = dev_priv->engine[ring_id]->mmio_base; > > + struct drm_i915_private *i915 = vgpu->gvt->i915; > > + u32 ring_base = i915->engine[ring_id]->mmio_base; > > i915_reg_t reg; > > reg = RING_INSTDONE(ring_base); > > @@ -391,10 +391,10 @@ intel_gvt_workload_req_alloc(struct intel_vgpu_workload *workload) > > { > > struct intel_vgpu *vgpu = workload->vgpu; > > struct intel_vgpu_submission *s = &vgpu->submission; > > - struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv; > > + struct drm_i915_private *i915 = vgpu->gvt->i915; > > struct i915_request *rq; > > - lockdep_assert_held(&dev_priv->drm.struct_mutex); > > + lockdep_assert_held(&i915->drm.struct_mutex); > > if (workload->req) > > return 0; > > @@ -421,10 +421,10 @@ int intel_gvt_scan_and_shadow_workload(struct intel_vgpu_workload *workload) > > { > > struct intel_vgpu *vgpu = workload->vgpu; > > struct intel_vgpu_submission *s = &vgpu->submission; > > - struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv; > > + struct drm_i915_private *i915 = vgpu->gvt->i915; > > int ret; > > - lockdep_assert_held(&dev_priv->drm.struct_mutex); > > + lockdep_assert_held(&i915->drm.struct_mutex); > > if (workload->shadow) > > return 0; > > @@ -577,7 +577,7 @@ static int prepare_shadow_wa_ctx(struct intel_shadow_wa_ctx *wa_ctx) > > static void release_shadow_batch_buffer(struct intel_vgpu_workload *workload) > > { > > struct intel_vgpu *vgpu = workload->vgpu; > > - struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv; > > + struct drm_i915_private *i915 = vgpu->gvt->i915; > > struct intel_vgpu_shadow_bb *bb, *pos; > > if (list_empty(&workload->shadow_bb)) > > @@ -586,7 +586,7 @@ static void release_shadow_batch_buffer(struct intel_vgpu_workload *workload) > > bb = list_first_entry(&workload->shadow_bb, > > struct intel_vgpu_shadow_bb, list); > > - mutex_lock(&dev_priv->drm.struct_mutex); > > + mutex_lock(&i915->drm.struct_mutex); > > list_for_each_entry_safe(bb, pos, &workload->shadow_bb, list) { > > if (bb->obj) { > > @@ -606,7 +606,7 @@ static void release_shadow_batch_buffer(struct intel_vgpu_workload *workload) > > kfree(bb); > > } > > - mutex_unlock(&dev_priv->drm.struct_mutex); > > + mutex_unlock(&i915->drm.struct_mutex); > > } > > static int prepare_workload(struct intel_vgpu_workload *workload) > > @@ -671,7 +671,7 @@ static int prepare_workload(struct intel_vgpu_workload *workload) > > static int dispatch_workload(struct intel_vgpu_workload *workload) > > { > > struct intel_vgpu *vgpu = workload->vgpu; > > - struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv; > > + struct drm_i915_private *i915 = vgpu->gvt->i915; > > struct intel_vgpu_submission *s = &vgpu->submission; > > struct i915_request *rq; > > int ring_id = workload->ring_id; > > @@ -681,7 +681,7 @@ static int dispatch_workload(struct intel_vgpu_workload *workload) > > ring_id, workload); > > mutex_lock(&vgpu->vgpu_lock); > > - mutex_lock(&dev_priv->drm.struct_mutex); > > + mutex_lock(&i915->drm.struct_mutex); > > ret = set_context_ppgtt_from_shadow(workload, > > s->shadow[ring_id]->gem_context); > > @@ -723,7 +723,7 @@ static int dispatch_workload(struct intel_vgpu_workload *workload) > > err_req: > > if (ret) > > workload->status = ret; > > - mutex_unlock(&dev_priv->drm.struct_mutex); > > + mutex_unlock(&i915->drm.struct_mutex); > > mutex_unlock(&vgpu->vgpu_lock); > > return ret; > > } > > @@ -796,7 +796,7 @@ static void update_guest_context(struct intel_vgpu_workload *workload) > > void *src; > > unsigned long context_gpa, context_page_num; > > int i; > > - struct drm_i915_private *dev_priv = gvt->dev_priv; > > + struct drm_i915_private *i915 = gvt->i915; > > u32 ring_base; > > u32 head, tail; > > u16 wrap_count; > > @@ -817,14 +817,14 @@ static void update_guest_context(struct intel_vgpu_workload *workload) > > head = (wrap_count << RB_HEAD_WRAP_CNT_OFF) | tail; > > - ring_base = dev_priv->engine[workload->ring_id]->mmio_base; > > + ring_base = i915->engine[workload->ring_id]->mmio_base; > > vgpu_vreg_t(vgpu, RING_TAIL(ring_base)) = tail; > > vgpu_vreg_t(vgpu, RING_HEAD(ring_base)) = head; > > context_page_num = rq->engine->context_size; > > context_page_num = context_page_num >> PAGE_SHIFT; > > - if (IS_BROADWELL(gvt->dev_priv) && rq->engine->id == RCS0) > > + if (IS_BROADWELL(gvt->i915) && rq->engine->id == RCS0) > > context_page_num = 19; > > i = 2; > > @@ -875,13 +875,13 @@ void intel_vgpu_clean_workloads(struct intel_vgpu *vgpu, > > intel_engine_mask_t engine_mask) > > { > > struct intel_vgpu_submission *s = &vgpu->submission; > > - struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv; > > + struct drm_i915_private *i915 = vgpu->gvt->i915; > > struct intel_engine_cs *engine; > > struct intel_vgpu_workload *pos, *n; > > intel_engine_mask_t tmp; > > /* free the unsubmited workloads in the queues. */ > > - for_each_engine_masked(engine, dev_priv, engine_mask, tmp) { > > + for_each_engine_masked(engine, i915, engine_mask, tmp) { > > list_for_each_entry_safe(pos, n, > > &s->workload_q_head[engine->id], list) { > > list_del_init(&pos->list); > > @@ -986,7 +986,7 @@ static int workload_thread(void *priv) > > struct intel_vgpu_workload *workload = NULL; > > struct intel_vgpu *vgpu = NULL; > > int ret; > > - bool need_force_wake = (INTEL_GEN(gvt->dev_priv) >= 9); > > + bool need_force_wake = (INTEL_GEN(gvt->i915) >= 9); > > DEFINE_WAIT_FUNC(wait, woken_wake_function); > > kfree(p); > > @@ -1015,7 +1015,7 @@ static int workload_thread(void *priv) > > workload->ring_id, workload); > > if (need_force_wake) > > - intel_uncore_forcewake_get(&gvt->dev_priv->uncore, > > + intel_uncore_forcewake_get(&gvt->i915->uncore, > > FORCEWAKE_ALL); > > ret = dispatch_workload(workload); > > @@ -1037,7 +1037,7 @@ static int workload_thread(void *priv) > > complete_current_workload(gvt, ring_id); > > if (need_force_wake) > > - intel_uncore_forcewake_put(&gvt->dev_priv->uncore, > > + intel_uncore_forcewake_put(&gvt->i915->uncore, > > FORCEWAKE_ALL); > > if (ret && (vgpu_is_vm_unhealthy(ret))) > > @@ -1068,7 +1068,7 @@ void intel_gvt_clean_workload_scheduler(struct intel_gvt *gvt) > > gvt_dbg_core("clean workload scheduler\n"); > > - for_each_engine(engine, gvt->dev_priv, i) { > > + for_each_engine(engine, gvt->i915, i) { > > atomic_notifier_chain_unregister( > > &engine->context_status_notifier, > > &gvt->shadow_ctx_notifier_block[i]); > > @@ -1088,7 +1088,7 @@ int intel_gvt_init_workload_scheduler(struct intel_gvt *gvt) > > init_waitqueue_head(&scheduler->workload_complete_wq); > > - for_each_engine(engine, gvt->dev_priv, i) { > > + for_each_engine(engine, gvt->i915, i) { > > init_waitqueue_head(&scheduler->waitq[i]); > > param = kzalloc(sizeof(*param), GFP_KERNEL); > > @@ -1152,7 +1152,7 @@ void intel_vgpu_clean_submission(struct intel_vgpu *vgpu) > > intel_vgpu_select_submission_ops(vgpu, ALL_ENGINES, 0); > > i915_context_ppgtt_root_restore(s, i915_vm_to_ppgtt(s->shadow[0]->gem_context->vm)); > > - for_each_engine(engine, vgpu->gvt->dev_priv, id) > > + for_each_engine(engine, vgpu->gvt->i915, id) > > intel_context_unpin(s->shadow[id]); > > kmem_cache_destroy(s->workloads); > > @@ -1212,13 +1212,13 @@ int intel_vgpu_setup_submission(struct intel_vgpu *vgpu) > > enum intel_engine_id i; > > int ret; > > - ctx = i915_gem_context_create_gvt(&vgpu->gvt->dev_priv->drm); > > + ctx = i915_gem_context_create_gvt(&vgpu->gvt->i915->drm); > > if (IS_ERR(ctx)) > > return PTR_ERR(ctx); > > i915_context_ppgtt_root_save(s, i915_vm_to_ppgtt(ctx->vm)); > > - for_each_engine(engine, vgpu->gvt->dev_priv, i) { > > + for_each_engine(engine, vgpu->gvt->i915, i) { > > struct intel_context *ce; > > INIT_LIST_HEAD(&s->workload_q_head[i]); > > @@ -1260,7 +1260,7 @@ int intel_vgpu_setup_submission(struct intel_vgpu *vgpu) > > out_shadow_ctx: > > i915_context_ppgtt_root_restore(s, i915_vm_to_ppgtt(ctx->vm)); > > - for_each_engine(engine, vgpu->gvt->dev_priv, i) { > > + for_each_engine(engine, vgpu->gvt->i915, i) { > > if (IS_ERR(s->shadow[i])) > > break; > > @@ -1439,7 +1439,7 @@ intel_vgpu_create_workload(struct intel_vgpu *vgpu, int ring_id, > > struct list_head *q = workload_q_head(vgpu, ring_id); > > struct intel_vgpu_workload *last_workload = get_last_workload(q); > > struct intel_vgpu_workload *workload = NULL; > > - struct drm_i915_private *dev_priv = vgpu->gvt->dev_priv; > > + struct drm_i915_private *i915 = vgpu->gvt->i915; > > u64 ring_context_gpa; > > u32 head, tail, start, ctl, ctx_ctl, per_ctx, indirect_ctx; > > u32 guest_head; > > @@ -1526,11 +1526,11 @@ intel_vgpu_create_workload(struct intel_vgpu *vgpu, int ring_id, > > * as there is only one pre-allocated buf-obj for shadow. > > */ > > if (list_empty(workload_q_head(vgpu, ring_id))) { > > - intel_runtime_pm_get(dev_priv); > > - mutex_lock(&dev_priv->drm.struct_mutex); > > + intel_runtime_pm_get(i915); > > + mutex_lock(&i915->drm.struct_mutex); > > ret = intel_gvt_scan_and_shadow_workload(workload); > > - mutex_unlock(&dev_priv->drm.struct_mutex); > > - intel_runtime_pm_put_unchecked(dev_priv); > > + mutex_unlock(&i915->drm.struct_mutex); > > + intel_runtime_pm_put_unchecked(i915); > > } > > if (ret) { > > diff --git a/drivers/gpu/drm/i915/gvt/vgpu.c b/drivers/gpu/drm/i915/gvt/vgpu.c > > index 44ce3c2b9ac1..a311737c9633 100644 > > --- a/drivers/gpu/drm/i915/gvt/vgpu.c > > +++ b/drivers/gpu/drm/i915/gvt/vgpu.c > > @@ -148,10 +148,10 @@ int intel_gvt_init_vgpu_types(struct intel_gvt *gvt) > > gvt->types[i].avail_instance = min(low_avail / vgpu_types[i].low_mm, > > high_avail / vgpu_types[i].high_mm); > > - if (IS_GEN(gvt->dev_priv, 8)) > > + if (IS_GEN(gvt->i915, 8)) > > sprintf(gvt->types[i].name, "GVTg_V4_%s", > > vgpu_types[i].name); > > - else if (IS_GEN(gvt->dev_priv, 9)) > > + else if (IS_GEN(gvt->i915, 9)) > > sprintf(gvt->types[i].name, "GVTg_V5_%s", > > vgpu_types[i].name); > > @@ -429,7 +429,7 @@ static struct intel_vgpu *__intel_gvt_create_vgpu(struct intel_gvt *gvt, > > goto out_clean_sched_policy; > > /*TODO: add more platforms support */ > > - if (IS_SKYLAKE(gvt->dev_priv) || IS_KABYLAKE(gvt->dev_priv)) > > + if (IS_SKYLAKE(gvt->i915) || IS_KABYLAKE(gvt->i915)) > > ret = intel_gvt_hypervisor_set_edid(vgpu, PORT_D); > > if (ret) > > goto out_clean_sched_policy; > > diff --git a/drivers/gpu/drm/i915/i915_cmd_parser.c b/drivers/gpu/drm/i915/i915_cmd_parser.c > > index a28bcd2d7c09..e9d076167b24 100644 > > --- a/drivers/gpu/drm/i915/i915_cmd_parser.c > > +++ b/drivers/gpu/drm/i915/i915_cmd_parser.c > > @@ -1342,21 +1342,21 @@ int intel_engine_cmd_parser(struct intel_engine_cs *engine, > > /** > > * i915_cmd_parser_get_version() - get the cmd parser version number > > - * @dev_priv: i915 device private > > + * @i915: i915 device private > > * > > * The cmd parser maintains a simple increasing integer version number suitable > > * for passing to userspace clients to determine what operations are permitted. > > * > > * Return: the current version number of the cmd parser > > */ > > -int i915_cmd_parser_get_version(struct drm_i915_private *dev_priv) > > +int i915_cmd_parser_get_version(struct drm_i915_private *i915) > > { > > struct intel_engine_cs *engine; > > enum intel_engine_id id; > > bool active = false; > > /* If the command parser is not enabled, report 0 - unsupported */ > > - for_each_engine(engine, dev_priv, id) { > > + for_each_engine(engine, i915, id) { > > if (intel_engine_needs_cmd_parser(engine)) { > > active = true; > > break; > > diff --git a/drivers/gpu/drm/i915/i915_debugfs.c b/drivers/gpu/drm/i915/i915_debugfs.c > > index 026fb46020f6..fec01858685d 100644 > > --- a/drivers/gpu/drm/i915/i915_debugfs.c > > +++ b/drivers/gpu/drm/i915/i915_debugfs.c > > @@ -55,17 +55,17 @@ static inline struct drm_i915_private *node_to_i915(struct drm_info_node *node) > > static int i915_capabilities(struct seq_file *m, void *data) > > { > > - struct drm_i915_private *dev_priv = node_to_i915(m->private); > > - const struct intel_device_info *info = INTEL_INFO(dev_priv); > > + struct drm_i915_private *i915 = node_to_i915(m->private); > > + const struct intel_device_info *info = INTEL_INFO(i915); > > struct drm_printer p = drm_seq_file_printer(m); > > - seq_printf(m, "gen: %d\n", INTEL_GEN(dev_priv)); > > + seq_printf(m, "gen: %d\n", INTEL_GEN(i915)); > > seq_printf(m, "platform: %s\n", intel_platform_name(info->platform)); > > - seq_printf(m, "pch: %d\n", INTEL_PCH_TYPE(dev_priv)); > > + seq_printf(m, "pch: %d\n", INTEL_PCH_TYPE(i915)); > > intel_device_info_dump_flags(info, &p); > > - intel_device_info_dump_runtime(RUNTIME_INFO(dev_priv), &p); > > - intel_driver_caps_print(&dev_priv->caps, &p); > > + intel_device_info_dump_runtime(RUNTIME_INFO(i915), &p); > > + intel_driver_caps_print(&i915->caps, &p); > > kernel_param_lock(THIS_MODULE); > > i915_params_dump(&i915_modparams, &p); > > @@ -137,7 +137,7 @@ stringify_page_sizes(unsigned int page_sizes, char *buf, size_t len) > > static void > > describe_obj(struct seq_file *m, struct drm_i915_gem_object *obj) > > { > > - struct drm_i915_private *dev_priv = to_i915(obj->base.dev); > > + struct drm_i915_private *i915 = to_i915(obj->base.dev); > > struct intel_engine_cs *engine; > > struct i915_vma *vma; > > unsigned int frontbuffer_bits; > > @@ -155,7 +155,7 @@ describe_obj(struct seq_file *m, struct drm_i915_gem_object *obj) > > obj->base.size / 1024, > > obj->read_domains, > > obj->write_domain, > > - i915_cache_level_str(dev_priv, obj->cache_level), > > + i915_cache_level_str(i915, obj->cache_level), > > obj->mm.dirty ? " dirty" : "", > > obj->mm.madv == I915_MADV_DONTNEED ? " purgeable" : ""); > > if (obj->base.name) > > @@ -296,7 +296,7 @@ static int per_file_stats(int id, void *ptr, void *data) > > } while (0) > > static void print_batch_pool_stats(struct seq_file *m, > > - struct drm_i915_private *dev_priv) > > + struct drm_i915_private *i915) > > { > > struct drm_i915_gem_object *obj; > > struct intel_engine_cs *engine; > > @@ -304,7 +304,7 @@ static void print_batch_pool_stats(struct seq_file *m, > > enum intel_engine_id id; > > int j; > > - for_each_engine(engine, dev_priv, id) { > > + for_each_engine(engine, i915, id) { > > for (j = 0; j < ARRAY_SIZE(engine->batch_pool.cache_list); j++) { > > list_for_each_entry(obj, > > &engine->batch_pool.cache_list[j], > > @@ -382,8 +382,8 @@ static int i915_gem_object_info(struct seq_file *m, void *data) > > static int i915_gem_batch_pool_info(struct seq_file *m, void *data) > > { > > - struct drm_i915_private *dev_priv = node_to_i915(m->private); > > - struct drm_device *dev = &dev_priv->drm; > > + struct drm_i915_private *i915 = node_to_i915(m->private); > > + struct drm_device *dev = &i915->drm; > > struct drm_i915_gem_object *obj; > > struct intel_engine_cs *engine; > > enum intel_engine_id id; > > @@ -394,7 +394,7 @@ static int i915_gem_batch_pool_info(struct seq_file *m, void *data) > > if (ret) > > return ret; > > - for_each_engine(engine, dev_priv, id) { > > + for_each_engine(engine, i915, id) { > > for (j = 0; j < ARRAY_SIZE(engine->batch_pool.cache_list); j++) { > > int count; > > @@ -427,15 +427,15 @@ static int i915_gem_batch_pool_info(struct seq_file *m, void *data) > > static void gen8_display_interrupt_info(struct seq_file *m) > > { > > - struct drm_i915_private *dev_priv = node_to_i915(m->private); > > + struct drm_i915_private *i915 = node_to_i915(m->private); > > int pipe; > > - for_each_pipe(dev_priv, pipe) { > > + for_each_pipe(i915, pipe) { > > enum intel_display_power_domain power_domain; > > intel_wakeref_t wakeref; > > power_domain = POWER_DOMAIN_PIPE(pipe); > > - wakeref = intel_display_power_get_if_enabled(dev_priv, > > + wakeref = intel_display_power_get_if_enabled(i915, > > power_domain); > > if (!wakeref) { > > seq_printf(m, "Pipe %c power disabled\n", > > @@ -452,7 +452,7 @@ static void gen8_display_interrupt_info(struct seq_file *m) > > pipe_name(pipe), > > I915_READ(GEN8_DE_PIPE_IER(pipe))); > > - intel_display_power_put(dev_priv, power_domain, wakeref); > > + intel_display_power_put(i915, power_domain, wakeref); > > } > > seq_printf(m, "Display Engine port interrupt mask:\t%08x\n", > > @@ -479,15 +479,15 @@ static void gen8_display_interrupt_info(struct seq_file *m) > > static int i915_interrupt_info(struct seq_file *m, void *data) > > { > > - struct drm_i915_private *dev_priv = node_to_i915(m->private); > > + struct drm_i915_private *i915 = node_to_i915(m->private); > > struct intel_engine_cs *engine; > > enum intel_engine_id id; > > intel_wakeref_t wakeref; > > int i, pipe; > > - wakeref = intel_runtime_pm_get(dev_priv); > > + wakeref = intel_runtime_pm_get(i915); > > - if (IS_CHERRYVIEW(dev_priv)) { > > + if (IS_CHERRYVIEW(i915)) { > > intel_wakeref_t pref; > > seq_printf(m, "Master Interrupt Control:\t%08x\n", > > @@ -501,11 +501,11 @@ static int i915_interrupt_info(struct seq_file *m, void *data) > > I915_READ(VLV_IIR_RW)); > > seq_printf(m, "Display IMR:\t%08x\n", > > I915_READ(VLV_IMR)); > > - for_each_pipe(dev_priv, pipe) { > > + for_each_pipe(i915, pipe) { > > enum intel_display_power_domain power_domain; > > power_domain = POWER_DOMAIN_PIPE(pipe); > > - pref = intel_display_power_get_if_enabled(dev_priv, > > + pref = intel_display_power_get_if_enabled(i915, > > power_domain); > > if (!pref) { > > seq_printf(m, "Pipe %c power disabled\n", > > @@ -517,17 +517,17 @@ static int i915_interrupt_info(struct seq_file *m, void *data) > > pipe_name(pipe), > > I915_READ(PIPESTAT(pipe))); > > - intel_display_power_put(dev_priv, power_domain, pref); > > + intel_display_power_put(i915, power_domain, pref); > > } > > - pref = intel_display_power_get(dev_priv, POWER_DOMAIN_INIT); > > + pref = intel_display_power_get(i915, POWER_DOMAIN_INIT); > > seq_printf(m, "Port hotplug:\t%08x\n", > > I915_READ(PORT_HOTPLUG_EN)); > > seq_printf(m, "DPFLIPSTAT:\t%08x\n", > > I915_READ(VLV_DPFLIPSTAT)); > > seq_printf(m, "DPINVGTT:\t%08x\n", > > I915_READ(DPINVGTT)); > > - intel_display_power_put(dev_priv, POWER_DOMAIN_INIT, pref); > > + intel_display_power_put(i915, POWER_DOMAIN_INIT, pref); > > for (i = 0; i < 4; i++) { > > seq_printf(m, "GT Interrupt IMR %d:\t%08x\n", > > @@ -544,7 +544,7 @@ static int i915_interrupt_info(struct seq_file *m, void *data) > > I915_READ(GEN8_PCU_IIR)); > > seq_printf(m, "PCU interrupt enable:\t%08x\n", > > I915_READ(GEN8_PCU_IER)); > > - } else if (INTEL_GEN(dev_priv) >= 11) { > > + } else if (INTEL_GEN(i915) >= 11) { > > seq_printf(m, "Master Interrupt Control: %08x\n", > > I915_READ(GEN11_GFX_MSTR_IRQ)); > > @@ -565,7 +565,7 @@ static int i915_interrupt_info(struct seq_file *m, void *data) > > I915_READ(GEN11_DISPLAY_INT_CTL)); > > gen8_display_interrupt_info(m); > > - } else if (INTEL_GEN(dev_priv) >= 8) { > > + } else if (INTEL_GEN(i915) >= 8) { > > seq_printf(m, "Master Interrupt Control:\t%08x\n", > > I915_READ(GEN8_MASTER_IRQ)); > > @@ -579,7 +579,7 @@ static int i915_interrupt_info(struct seq_file *m, void *data) > > } > > gen8_display_interrupt_info(m); > > - } else if (IS_VALLEYVIEW(dev_priv)) { > > + } else if (IS_VALLEYVIEW(i915)) { > > seq_printf(m, "Display IER:\t%08x\n", > > I915_READ(VLV_IER)); > > seq_printf(m, "Display IIR:\t%08x\n", > > @@ -588,12 +588,12 @@ static int i915_interrupt_info(struct seq_file *m, void *data) > > I915_READ(VLV_IIR_RW)); > > seq_printf(m, "Display IMR:\t%08x\n", > > I915_READ(VLV_IMR)); > > - for_each_pipe(dev_priv, pipe) { > > + for_each_pipe(i915, pipe) { > > enum intel_display_power_domain power_domain; > > intel_wakeref_t pref; > > power_domain = POWER_DOMAIN_PIPE(pipe); > > - pref = intel_display_power_get_if_enabled(dev_priv, > > + pref = intel_display_power_get_if_enabled(i915, > > power_domain); > > if (!pref) { > > seq_printf(m, "Pipe %c power disabled\n", > > @@ -604,7 +604,7 @@ static int i915_interrupt_info(struct seq_file *m, void *data) > > seq_printf(m, "Pipe %c stat:\t%08x\n", > > pipe_name(pipe), > > I915_READ(PIPESTAT(pipe))); > > - intel_display_power_put(dev_priv, power_domain, pref); > > + intel_display_power_put(i915, power_domain, pref); > > } > > seq_printf(m, "Master IER:\t%08x\n", > > @@ -631,14 +631,14 @@ static int i915_interrupt_info(struct seq_file *m, void *data) > > seq_printf(m, "DPINVGTT:\t%08x\n", > > I915_READ(DPINVGTT)); > > - } else if (!HAS_PCH_SPLIT(dev_priv)) { > > + } else if (!HAS_PCH_SPLIT(i915)) { > > seq_printf(m, "Interrupt enable: %08x\n", > > I915_READ(GEN2_IER)); > > seq_printf(m, "Interrupt identity: %08x\n", > > I915_READ(GEN2_IIR)); > > seq_printf(m, "Interrupt mask: %08x\n", > > I915_READ(GEN2_IMR)); > > - for_each_pipe(dev_priv, pipe) > > + for_each_pipe(i915, pipe) > > seq_printf(m, "Pipe %c stat: %08x\n", > > pipe_name(pipe), > > I915_READ(PIPESTAT(pipe))); > > @@ -663,7 +663,7 @@ static int i915_interrupt_info(struct seq_file *m, void *data) > > I915_READ(GTIMR)); > > } > > - if (INTEL_GEN(dev_priv) >= 11) { > > + if (INTEL_GEN(i915) >= 11) { > > seq_printf(m, "RCS Intr Mask:\t %08x\n", > > I915_READ(GEN11_RCS0_RSVD_INTR_MASK)); > > seq_printf(m, "BCS Intr Mask:\t %08x\n", > > @@ -683,35 +683,35 @@ static int i915_interrupt_info(struct seq_file *m, void *data) > > seq_printf(m, "Gunit/CSME Intr Mask:\t %08x\n", > > I915_READ(GEN11_GUNIT_CSME_INTR_MASK)); > > - } else if (INTEL_GEN(dev_priv) >= 6) { > > - for_each_engine(engine, dev_priv, id) { > > + } else if (INTEL_GEN(i915) >= 6) { > > + for_each_engine(engine, i915, id) { > > seq_printf(m, > > "Graphics Interrupt mask (%s): %08x\n", > > engine->name, ENGINE_READ(engine, RING_IMR)); > > } > > } > > - intel_runtime_pm_put(dev_priv, wakeref); > > + intel_runtime_pm_put(i915, wakeref); > > return 0; > > } > > static int i915_gem_fence_regs_info(struct seq_file *m, void *data) > > { > > - struct drm_i915_private *dev_priv = node_to_i915(m->private); > > - struct drm_device *dev = &dev_priv->drm; > > + struct drm_i915_private *i915 = node_to_i915(m->private); > > + struct drm_device *dev = &i915->drm; > > int i, ret; > > ret = mutex_lock_interruptible(&dev->struct_mutex); > > if (ret) > > return ret; > > - seq_printf(m, "Total fences = %d\n", dev_priv->num_fence_regs); > > - for (i = 0; i < dev_priv->num_fence_regs; i++) { > > - struct i915_vma *vma = dev_priv->fence_regs[i].vma; > > + seq_printf(m, "Total fences = %d\n", i915->num_fence_regs); > > + for (i = 0; i < i915->num_fence_regs; i++) { > > + struct i915_vma *vma = i915->fence_regs[i].vma; > > seq_printf(m, "Fence %d, pin count = %d, object = ", > > - i, dev_priv->fence_regs[i].pin_count); > > + i, i915->fence_regs[i].pin_count); > > if (!vma) > > seq_puts(m, "unused"); > > else > > @@ -825,15 +825,15 @@ static const struct file_operations i915_error_state_fops = { > > static int i915_frequency_info(struct seq_file *m, void *unused) > > { > > - struct drm_i915_private *dev_priv = node_to_i915(m->private); > > - struct intel_uncore *uncore = &dev_priv->uncore; > > - struct intel_rps *rps = &dev_priv->gt_pm.rps; > > + struct drm_i915_private *i915 = node_to_i915(m->private); > > + struct intel_uncore *uncore = &i915->uncore; > > + struct intel_rps *rps = &i915->gt_pm.rps; > > intel_wakeref_t wakeref; > > int ret = 0; > > - wakeref = intel_runtime_pm_get(dev_priv); > > + wakeref = intel_runtime_pm_get(i915); > > - if (IS_GEN(dev_priv, 5)) { > > + if (IS_GEN(i915, 5)) { > > u16 rgvswctl = intel_uncore_read16(uncore, MEMSWCTL); > > u16 rgvstat = intel_uncore_read16(uncore, MEMSTAT_ILK); > > @@ -843,7 +843,7 @@ static int i915_frequency_info(struct seq_file *m, void *unused) > > MEMSTAT_VID_SHIFT); > > seq_printf(m, "Current P-state: %d\n", > > (rgvstat & MEMSTAT_PSTATE_MASK) >> MEMSTAT_PSTATE_SHIFT); > > - } else if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) { > > + } else if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) { > > u32 rpmodectl, freq_sts; > > rpmodectl = I915_READ(GEN6_RP_CONTROL); > > @@ -855,32 +855,32 @@ static int i915_frequency_info(struct seq_file *m, void *unused) > > yesno((rpmodectl & GEN6_RP_MEDIA_MODE_MASK) == > > GEN6_RP_MEDIA_SW_MODE)); > > - vlv_punit_get(dev_priv); > > - freq_sts = vlv_punit_read(dev_priv, PUNIT_REG_GPU_FREQ_STS); > > - vlv_punit_put(dev_priv); > > + vlv_punit_get(i915); > > + freq_sts = vlv_punit_read(i915, PUNIT_REG_GPU_FREQ_STS); > > + vlv_punit_put(i915); > > seq_printf(m, "PUNIT_REG_GPU_FREQ_STS: 0x%08x\n", freq_sts); > > - seq_printf(m, "DDR freq: %d MHz\n", dev_priv->mem_freq); > > + seq_printf(m, "DDR freq: %d MHz\n", i915->mem_freq); > > seq_printf(m, "actual GPU freq: %d MHz\n", > > - intel_gpu_freq(dev_priv, (freq_sts >> 8) & 0xff)); > > + intel_gpu_freq(i915, (freq_sts >> 8) & 0xff)); > > seq_printf(m, "current GPU freq: %d MHz\n", > > - intel_gpu_freq(dev_priv, rps->cur_freq)); > > + intel_gpu_freq(i915, rps->cur_freq)); > > seq_printf(m, "max GPU freq: %d MHz\n", > > - intel_gpu_freq(dev_priv, rps->max_freq)); > > + intel_gpu_freq(i915, rps->max_freq)); > > seq_printf(m, "min GPU freq: %d MHz\n", > > - intel_gpu_freq(dev_priv, rps->min_freq)); > > + intel_gpu_freq(i915, rps->min_freq)); > > seq_printf(m, "idle GPU freq: %d MHz\n", > > - intel_gpu_freq(dev_priv, rps->idle_freq)); > > + intel_gpu_freq(i915, rps->idle_freq)); > > seq_printf(m, > > "efficient (RPe) frequency: %d MHz\n", > > - intel_gpu_freq(dev_priv, rps->efficient_freq)); > > - } else if (INTEL_GEN(dev_priv) >= 6) { > > + intel_gpu_freq(i915, rps->efficient_freq)); > > + } else if (INTEL_GEN(i915) >= 6) { > > u32 rp_state_limits; > > u32 gt_perf_status; > > u32 rp_state_cap; > > @@ -892,7 +892,7 @@ static int i915_frequency_info(struct seq_file *m, void *unused) > > int max_freq; > > rp_state_limits = I915_READ(GEN6_RP_STATE_LIMITS); > > - if (IS_GEN9_LP(dev_priv)) { > > + if (IS_GEN9_LP(i915)) { > > rp_state_cap = I915_READ(BXT_RP_STATE_CAP); > > gt_perf_status = I915_READ(BXT_GT_PERF_STATUS); > > } else { > > @@ -901,19 +901,19 @@ static int i915_frequency_info(struct seq_file *m, void *unused) > > } > > /* RPSTAT1 is in the GT power well */ > > - intel_uncore_forcewake_get(&dev_priv->uncore, FORCEWAKE_ALL); > > + intel_uncore_forcewake_get(&i915->uncore, FORCEWAKE_ALL); > > reqf = I915_READ(GEN6_RPNSWREQ); > > - if (INTEL_GEN(dev_priv) >= 9) > > + if (INTEL_GEN(i915) >= 9) > > reqf >>= 23; > > else { > > reqf &= ~GEN6_TURBO_DISABLE; > > - if (IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv)) > > + if (IS_HASWELL(i915) || IS_BROADWELL(i915)) > > reqf >>= 24; > > else > > reqf >>= 25; > > } > > - reqf = intel_gpu_freq(dev_priv, reqf); > > + reqf = intel_gpu_freq(i915, reqf); > > rpmodectl = I915_READ(GEN6_RP_CONTROL); > > rpinclimit = I915_READ(GEN6_RP_UP_THRESHOLD); > > @@ -926,12 +926,12 @@ static int i915_frequency_info(struct seq_file *m, void *unused) > > rpdownei = I915_READ(GEN6_RP_CUR_DOWN_EI) & GEN6_CURIAVG_MASK; > > rpcurdown = I915_READ(GEN6_RP_CUR_DOWN) & GEN6_CURBSYTAVG_MASK; > > rpprevdown = I915_READ(GEN6_RP_PREV_DOWN) & GEN6_CURBSYTAVG_MASK; > > - cagf = intel_gpu_freq(dev_priv, > > - intel_get_cagf(dev_priv, rpstat)); > > + cagf = intel_gpu_freq(i915, > > + intel_get_cagf(i915, rpstat)); > > - intel_uncore_forcewake_put(&dev_priv->uncore, FORCEWAKE_ALL); > > + intel_uncore_forcewake_put(&i915->uncore, FORCEWAKE_ALL); > > - if (INTEL_GEN(dev_priv) >= 11) { > > + if (INTEL_GEN(i915) >= 11) { > > pm_ier = I915_READ(GEN11_GPM_WGBOXPERF_INTR_ENABLE); > > pm_imr = I915_READ(GEN11_GPM_WGBOXPERF_INTR_MASK); > > /* > > @@ -940,7 +940,7 @@ static int i915_frequency_info(struct seq_file *m, void *unused) > > */ > > pm_isr = 0; > > pm_iir = 0; > > - } else if (INTEL_GEN(dev_priv) >= 8) { > > + } else if (INTEL_GEN(i915) >= 8) { > > pm_ier = I915_READ(GEN8_GT_IER(2)); > > pm_imr = I915_READ(GEN8_GT_IMR(2)); > > pm_isr = I915_READ(GEN8_GT_ISR(2)); > > @@ -963,14 +963,14 @@ static int i915_frequency_info(struct seq_file *m, void *unused) > > seq_printf(m, "PM IER=0x%08x IMR=0x%08x, MASK=0x%08x\n", > > pm_ier, pm_imr, pm_mask); > > - if (INTEL_GEN(dev_priv) <= 10) > > + if (INTEL_GEN(i915) <= 10) > > seq_printf(m, "PM ISR=0x%08x IIR=0x%08x\n", > > pm_isr, pm_iir); > > seq_printf(m, "pm_intrmsk_mbz: 0x%08x\n", > > rps->pm_intrmsk_mbz); > > seq_printf(m, "GT_PERF_STATUS: 0x%08x\n", gt_perf_status); > > seq_printf(m, "Render p-state ratio: %d\n", > > - (gt_perf_status & (INTEL_GEN(dev_priv) >= 9 ? 0x1ff00 : 0xff00)) >> 8); > > + (gt_perf_status & (INTEL_GEN(i915) >= 9 ? 0x1ff00 : 0xff00)) >> 8); > > seq_printf(m, "Render p-state VID: %d\n", > > gt_perf_status & 0xff); > > seq_printf(m, "Render p-state limit: %d\n", > > @@ -982,72 +982,72 @@ static int i915_frequency_info(struct seq_file *m, void *unused) > > seq_printf(m, "RPNSWREQ: %dMHz\n", reqf); > > seq_printf(m, "CAGF: %dMHz\n", cagf); > > seq_printf(m, "RP CUR UP EI: %d (%dus)\n", > > - rpupei, GT_PM_INTERVAL_TO_US(dev_priv, rpupei)); > > + rpupei, GT_PM_INTERVAL_TO_US(i915, rpupei)); > > seq_printf(m, "RP CUR UP: %d (%dus)\n", > > - rpcurup, GT_PM_INTERVAL_TO_US(dev_priv, rpcurup)); > > + rpcurup, GT_PM_INTERVAL_TO_US(i915, rpcurup)); > > seq_printf(m, "RP PREV UP: %d (%dus)\n", > > - rpprevup, GT_PM_INTERVAL_TO_US(dev_priv, rpprevup)); > > + rpprevup, GT_PM_INTERVAL_TO_US(i915, rpprevup)); > > seq_printf(m, "Up threshold: %d%%\n", > > rps->power.up_threshold); > > seq_printf(m, "RP CUR DOWN EI: %d (%dus)\n", > > - rpdownei, GT_PM_INTERVAL_TO_US(dev_priv, rpdownei)); > > + rpdownei, GT_PM_INTERVAL_TO_US(i915, rpdownei)); > > seq_printf(m, "RP CUR DOWN: %d (%dus)\n", > > - rpcurdown, GT_PM_INTERVAL_TO_US(dev_priv, rpcurdown)); > > + rpcurdown, GT_PM_INTERVAL_TO_US(i915, rpcurdown)); > > seq_printf(m, "RP PREV DOWN: %d (%dus)\n", > > - rpprevdown, GT_PM_INTERVAL_TO_US(dev_priv, rpprevdown)); > > + rpprevdown, GT_PM_INTERVAL_TO_US(i915, rpprevdown)); > > seq_printf(m, "Down threshold: %d%%\n", > > rps->power.down_threshold); > > - max_freq = (IS_GEN9_LP(dev_priv) ? rp_state_cap >> 0 : > > + max_freq = (IS_GEN9_LP(i915) ? rp_state_cap >> 0 : > > rp_state_cap >> 16) & 0xff; > > - max_freq *= (IS_GEN9_BC(dev_priv) || > > - INTEL_GEN(dev_priv) >= 10 ? GEN9_FREQ_SCALER : 1); > > + max_freq *= (IS_GEN9_BC(i915) || > > + INTEL_GEN(i915) >= 10 ? GEN9_FREQ_SCALER : 1); > > seq_printf(m, "Lowest (RPN) frequency: %dMHz\n", > > - intel_gpu_freq(dev_priv, max_freq)); > > + intel_gpu_freq(i915, max_freq)); > > max_freq = (rp_state_cap & 0xff00) >> 8; > > - max_freq *= (IS_GEN9_BC(dev_priv) || > > - INTEL_GEN(dev_priv) >= 10 ? GEN9_FREQ_SCALER : 1); > > + max_freq *= (IS_GEN9_BC(i915) || > > + INTEL_GEN(i915) >= 10 ? GEN9_FREQ_SCALER : 1); > > seq_printf(m, "Nominal (RP1) frequency: %dMHz\n", > > - intel_gpu_freq(dev_priv, max_freq)); > > + intel_gpu_freq(i915, max_freq)); > > - max_freq = (IS_GEN9_LP(dev_priv) ? rp_state_cap >> 16 : > > + max_freq = (IS_GEN9_LP(i915) ? rp_state_cap >> 16 : > > rp_state_cap >> 0) & 0xff; > > - max_freq *= (IS_GEN9_BC(dev_priv) || > > - INTEL_GEN(dev_priv) >= 10 ? GEN9_FREQ_SCALER : 1); > > + max_freq *= (IS_GEN9_BC(i915) || > > + INTEL_GEN(i915) >= 10 ? GEN9_FREQ_SCALER : 1); > > seq_printf(m, "Max non-overclocked (RP0) frequency: %dMHz\n", > > - intel_gpu_freq(dev_priv, max_freq)); > > + intel_gpu_freq(i915, max_freq)); > > seq_printf(m, "Max overclocked frequency: %dMHz\n", > > - intel_gpu_freq(dev_priv, rps->max_freq)); > > + intel_gpu_freq(i915, rps->max_freq)); > > seq_printf(m, "Current freq: %d MHz\n", > > - intel_gpu_freq(dev_priv, rps->cur_freq)); > > + intel_gpu_freq(i915, rps->cur_freq)); > > seq_printf(m, "Actual freq: %d MHz\n", cagf); > > seq_printf(m, "Idle freq: %d MHz\n", > > - intel_gpu_freq(dev_priv, rps->idle_freq)); > > + intel_gpu_freq(i915, rps->idle_freq)); > > seq_printf(m, "Min freq: %d MHz\n", > > - intel_gpu_freq(dev_priv, rps->min_freq)); > > + intel_gpu_freq(i915, rps->min_freq)); > > seq_printf(m, "Boost freq: %d MHz\n", > > - intel_gpu_freq(dev_priv, rps->boost_freq)); > > + intel_gpu_freq(i915, rps->boost_freq)); > > seq_printf(m, "Max freq: %d MHz\n", > > - intel_gpu_freq(dev_priv, rps->max_freq)); > > + intel_gpu_freq(i915, rps->max_freq)); > > seq_printf(m, > > "efficient (RPe) frequency: %d MHz\n", > > - intel_gpu_freq(dev_priv, rps->efficient_freq)); > > + intel_gpu_freq(i915, rps->efficient_freq)); > > } else { > > seq_puts(m, "no P-state info available\n"); > > } > > - seq_printf(m, "Current CD clock frequency: %d kHz\n", dev_priv->cdclk.hw.cdclk); > > - seq_printf(m, "Max CD clock frequency: %d kHz\n", dev_priv->max_cdclk_freq); > > - seq_printf(m, "Max pixel clock frequency: %d kHz\n", dev_priv->max_dotclk_freq); > > + seq_printf(m, "Current CD clock frequency: %d kHz\n", i915->cdclk.hw.cdclk); > > + seq_printf(m, "Max CD clock frequency: %d kHz\n", i915->max_cdclk_freq); > > + seq_printf(m, "Max pixel clock frequency: %d kHz\n", i915->max_dotclk_freq); > > - intel_runtime_pm_put(dev_priv, wakeref); > > + intel_runtime_pm_put(i915, wakeref); > > return ret; > > } > > -static void i915_instdone_info(struct drm_i915_private *dev_priv, > > +static void i915_instdone_info(struct drm_i915_private *i915, > > struct seq_file *m, > > struct intel_instdone *instdone) > > { > > @@ -1057,37 +1057,37 @@ static void i915_instdone_info(struct drm_i915_private *dev_priv, > > seq_printf(m, "\t\tINSTDONE: 0x%08x\n", > > instdone->instdone); > > - if (INTEL_GEN(dev_priv) <= 3) > > + if (INTEL_GEN(i915) <= 3) > > return; > > seq_printf(m, "\t\tSC_INSTDONE: 0x%08x\n", > > instdone->slice_common); > > - if (INTEL_GEN(dev_priv) <= 6) > > + if (INTEL_GEN(i915) <= 6) > > return; > > - for_each_instdone_slice_subslice(dev_priv, slice, subslice) > > + for_each_instdone_slice_subslice(i915, slice, subslice) > > seq_printf(m, "\t\tSAMPLER_INSTDONE[%d][%d]: 0x%08x\n", > > slice, subslice, instdone->sampler[slice][subslice]); > > - for_each_instdone_slice_subslice(dev_priv, slice, subslice) > > + for_each_instdone_slice_subslice(i915, slice, subslice) > > seq_printf(m, "\t\tROW_INSTDONE[%d][%d]: 0x%08x\n", > > slice, subslice, instdone->row[slice][subslice]); > > } > > static int i915_hangcheck_info(struct seq_file *m, void *unused) > > { > > - struct drm_i915_private *dev_priv = node_to_i915(m->private); > > + struct drm_i915_private *i915 = node_to_i915(m->private); > > struct intel_engine_cs *engine; > > u64 acthd[I915_NUM_ENGINES]; > > struct intel_instdone instdone; > > intel_wakeref_t wakeref; > > enum intel_engine_id id; > > - seq_printf(m, "Reset flags: %lx\n", dev_priv->gpu_error.flags); > > - if (test_bit(I915_WEDGED, &dev_priv->gpu_error.flags)) > > + seq_printf(m, "Reset flags: %lx\n", i915->gpu_error.flags); > > + if (test_bit(I915_WEDGED, &i915->gpu_error.flags)) > > seq_puts(m, "\tWedged\n"); > > - if (test_bit(I915_RESET_BACKOFF, &dev_priv->gpu_error.flags)) > > + if (test_bit(I915_RESET_BACKOFF, &i915->gpu_error.flags)) > > seq_puts(m, "\tDevice (global) reset in progress\n"); > > if (!i915_modparams.enable_hangcheck) { > > @@ -1095,25 +1095,25 @@ static int i915_hangcheck_info(struct seq_file *m, void *unused) > > return 0; > > } > > - with_intel_runtime_pm(dev_priv, wakeref) { > > - for_each_engine(engine, dev_priv, id) > > + with_intel_runtime_pm(i915, wakeref) { > > + for_each_engine(engine, i915, id) > > acthd[id] = intel_engine_get_active_head(engine); > > - intel_engine_get_instdone(dev_priv->engine[RCS0], &instdone); > > + intel_engine_get_instdone(i915->engine[RCS0], &instdone); > > } > > - if (timer_pending(&dev_priv->gpu_error.hangcheck_work.timer)) > > + if (timer_pending(&i915->gpu_error.hangcheck_work.timer)) > > seq_printf(m, "Hangcheck active, timer fires in %dms\n", > > - jiffies_to_msecs(dev_priv->gpu_error.hangcheck_work.timer.expires - > > + jiffies_to_msecs(i915->gpu_error.hangcheck_work.timer.expires - > > jiffies)); > > - else if (delayed_work_pending(&dev_priv->gpu_error.hangcheck_work)) > > + else if (delayed_work_pending(&i915->gpu_error.hangcheck_work)) > > seq_puts(m, "Hangcheck active, work pending\n"); > > else > > seq_puts(m, "Hangcheck inactive\n"); > > - seq_printf(m, "GT active? %s\n", yesno(dev_priv->gt.awake)); > > + seq_printf(m, "GT active? %s\n", yesno(i915->gt.awake)); > > - for_each_engine(engine, dev_priv, id) { > > + for_each_engine(engine, i915, id) { > > seq_printf(m, "%s: %d ms ago\n", > > engine->name, > > jiffies_to_msecs(jiffies - > > @@ -1126,11 +1126,11 @@ static int i915_hangcheck_info(struct seq_file *m, void *unused) > > if (engine->id == RCS0) { > > seq_puts(m, "\tinstdone read =\n"); > > - i915_instdone_info(dev_priv, m, &instdone); > > + i915_instdone_info(i915, m, &instdone); > > seq_puts(m, "\tinstdone accu =\n"); > > - i915_instdone_info(dev_priv, m, > > + i915_instdone_info(i915, m, > > &engine->hangcheck.instdone); > > } > > } > > @@ -1140,14 +1140,14 @@ static int i915_hangcheck_info(struct seq_file *m, void *unused) > > static int i915_reset_info(struct seq_file *m, void *unused) > > { > > - struct drm_i915_private *dev_priv = node_to_i915(m->private); > > - struct i915_gpu_error *error = &dev_priv->gpu_error; > > + struct drm_i915_private *i915 = node_to_i915(m->private); > > + struct i915_gpu_error *error = &i915->gpu_error; > > struct intel_engine_cs *engine; > > enum intel_engine_id id; > > seq_printf(m, "full gpu reset = %u\n", i915_reset_count(error)); > > - for_each_engine(engine, dev_priv, id) { > > + for_each_engine(engine, i915, id) { > > seq_printf(m, "%s = %u\n", engine->name, > > i915_reset_engine_count(error, engine)); > > } > > @@ -1235,16 +1235,16 @@ static void print_rc6_res(struct seq_file *m, > > const char *title, > > const i915_reg_t reg) > > { > > - struct drm_i915_private *dev_priv = node_to_i915(m->private); > > + struct drm_i915_private *i915 = node_to_i915(m->private); > > seq_printf(m, "%s %u (%llu us)\n", > > title, I915_READ(reg), > > - intel_rc6_residency_us(dev_priv, reg)); > > + intel_rc6_residency_us(i915, reg)); > > } > > static int vlv_drpc_info(struct seq_file *m) > > { > > - struct drm_i915_private *dev_priv = node_to_i915(m->private); > > + struct drm_i915_private *i915 = node_to_i915(m->private); > > u32 rcctl1, pw_status; > > pw_status = I915_READ(VLV_GTLC_PW_STATUS); > > @@ -1266,7 +1266,7 @@ static int vlv_drpc_info(struct seq_file *m) > > static int gen6_drpc_info(struct seq_file *m) > > { > > - struct drm_i915_private *dev_priv = node_to_i915(m->private); > > + struct drm_i915_private *i915 = node_to_i915(m->private); > > u32 gt_core_status, rcctl1, rc6vids = 0; > > u32 gen9_powergate_enable = 0, gen9_powergate_status = 0; > > @@ -1274,20 +1274,20 @@ static int gen6_drpc_info(struct seq_file *m) > > trace_i915_reg_rw(false, GEN6_GT_CORE_STATUS, gt_core_status, 4, true); > > rcctl1 = I915_READ(GEN6_RC_CONTROL); > > - if (INTEL_GEN(dev_priv) >= 9) { > > + if (INTEL_GEN(i915) >= 9) { > > gen9_powergate_enable = I915_READ(GEN9_PG_ENABLE); > > gen9_powergate_status = I915_READ(GEN9_PWRGT_DOMAIN_STATUS); > > } > > - if (INTEL_GEN(dev_priv) <= 7) > > - sandybridge_pcode_read(dev_priv, GEN6_PCODE_READ_RC6VIDS, > > + if (INTEL_GEN(i915) <= 7) > > + sandybridge_pcode_read(i915, GEN6_PCODE_READ_RC6VIDS, > > &rc6vids, NULL); > > seq_printf(m, "RC1e Enabled: %s\n", > > yesno(rcctl1 & GEN6_RC_CTL_RC1e_ENABLE)); > > seq_printf(m, "RC6 Enabled: %s\n", > > yesno(rcctl1 & GEN6_RC_CTL_RC6_ENABLE)); > > - if (INTEL_GEN(dev_priv) >= 9) { > > + if (INTEL_GEN(i915) >= 9) { > > seq_printf(m, "Render Well Gating Enabled: %s\n", > > yesno(gen9_powergate_enable & GEN9_RENDER_PG_ENABLE)); > > seq_printf(m, "Media Well Gating Enabled: %s\n", > > @@ -1321,7 +1321,7 @@ static int gen6_drpc_info(struct seq_file *m) > > seq_printf(m, "Core Power Down: %s\n", > > yesno(gt_core_status & GEN6_CORE_CPD_STATE_MASK)); > > - if (INTEL_GEN(dev_priv) >= 9) { > > + if (INTEL_GEN(i915) >= 9) { > > seq_printf(m, "Render Power Well: %s\n", > > (gen9_powergate_status & > > GEN9_PWRGT_RENDER_STATUS_MASK) ? "Up" : "Down"); > > @@ -1337,7 +1337,7 @@ static int gen6_drpc_info(struct seq_file *m) > > print_rc6_res(m, "RC6+ residency since boot:", GEN6_GT_GFX_RC6p); > > print_rc6_res(m, "RC6++ residency since boot:", GEN6_GT_GFX_RC6pp); > > - if (INTEL_GEN(dev_priv) <= 7) { > > + if (INTEL_GEN(i915) <= 7) { > > seq_printf(m, "RC6 voltage: %dmV\n", > > GEN6_DECODE_RC6_VID(((rc6vids >> 0) & 0xff))); > > seq_printf(m, "RC6+ voltage: %dmV\n", > > @@ -1351,14 +1351,14 @@ static int gen6_drpc_info(struct seq_file *m) > > static int i915_drpc_info(struct seq_file *m, void *unused) > > { > > - struct drm_i915_private *dev_priv = node_to_i915(m->private); > > + struct drm_i915_private *i915 = node_to_i915(m->private); > > intel_wakeref_t wakeref; > > int err = -ENODEV; > > - with_intel_runtime_pm(dev_priv, wakeref) { > > - if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) > > + with_intel_runtime_pm(i915, wakeref) { > > + if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) > > err = vlv_drpc_info(m); > > - else if (INTEL_GEN(dev_priv) >= 6) > > + else if (INTEL_GEN(i915) >= 6) > > err = gen6_drpc_info(m); > > else > > err = ironlake_drpc_info(m); > > @@ -1369,44 +1369,44 @@ static int i915_drpc_info(struct seq_file *m, void *unused) > > static int i915_frontbuffer_tracking(struct seq_file *m, void *unused) > > { > > - struct drm_i915_private *dev_priv = node_to_i915(m->private); > > + struct drm_i915_private *i915 = node_to_i915(m->private); > > seq_printf(m, "FB tracking busy bits: 0x%08x\n", > > - dev_priv->fb_tracking.busy_bits); > > + i915->fb_tracking.busy_bits); > > seq_printf(m, "FB tracking flip bits: 0x%08x\n", > > - dev_priv->fb_tracking.flip_bits); > > + i915->fb_tracking.flip_bits); > > return 0; > > } > > static int i915_fbc_status(struct seq_file *m, void *unused) > > { > > - struct drm_i915_private *dev_priv = node_to_i915(m->private); > > - struct intel_fbc *fbc = &dev_priv->fbc; > > + struct drm_i915_private *i915 = node_to_i915(m->private); > > + struct intel_fbc *fbc = &i915->fbc; > > intel_wakeref_t wakeref; > > - if (!HAS_FBC(dev_priv)) > > + if (!HAS_FBC(i915)) > > return -ENODEV; > > - wakeref = intel_runtime_pm_get(dev_priv); > > + wakeref = intel_runtime_pm_get(i915); > > mutex_lock(&fbc->lock); > > - if (intel_fbc_is_active(dev_priv)) > > + if (intel_fbc_is_active(i915)) > > seq_puts(m, "FBC enabled\n"); > > else > > seq_printf(m, "FBC disabled: %s\n", fbc->no_fbc_reason); > > - if (intel_fbc_is_active(dev_priv)) { > > + if (intel_fbc_is_active(i915)) { > > u32 mask; > > - if (INTEL_GEN(dev_priv) >= 8) > > + if (INTEL_GEN(i915) >= 8) > > mask = I915_READ(IVB_FBC_STATUS2) & BDW_FBC_COMP_SEG_MASK; > > - else if (INTEL_GEN(dev_priv) >= 7) > > + else if (INTEL_GEN(i915) >= 7) > > mask = I915_READ(IVB_FBC_STATUS2) & IVB_FBC_COMP_SEG_MASK; > > - else if (INTEL_GEN(dev_priv) >= 5) > > + else if (INTEL_GEN(i915) >= 5) > > mask = I915_READ(ILK_DPFC_STATUS) & ILK_DPFC_COMP_SEG_MASK; > > - else if (IS_G4X(dev_priv)) > > + else if (IS_G4X(i915)) > > mask = I915_READ(DPFC_STATUS) & DPFC_COMP_SEG_MASK; > > else > > mask = I915_READ(FBC_STATUS) & (FBC_STAT_COMPRESSING | > > @@ -1416,41 +1416,41 @@ static int i915_fbc_status(struct seq_file *m, void *unused) > > } > > mutex_unlock(&fbc->lock); > > - intel_runtime_pm_put(dev_priv, wakeref); > > + intel_runtime_pm_put(i915, wakeref); > > return 0; > > } > > static int i915_fbc_false_color_get(void *data, u64 *val) > > { > > - struct drm_i915_private *dev_priv = data; > > + struct drm_i915_private *i915 = data; > > - if (INTEL_GEN(dev_priv) < 7 || !HAS_FBC(dev_priv)) > > + if (INTEL_GEN(i915) < 7 || !HAS_FBC(i915)) > > return -ENODEV; > > - *val = dev_priv->fbc.false_color; > > + *val = i915->fbc.false_color; > > return 0; > > } > > static int i915_fbc_false_color_set(void *data, u64 val) > > { > > - struct drm_i915_private *dev_priv = data; > > + struct drm_i915_private *i915 = data; > > u32 reg; > > - if (INTEL_GEN(dev_priv) < 7 || !HAS_FBC(dev_priv)) > > + if (INTEL_GEN(i915) < 7 || !HAS_FBC(i915)) > > return -ENODEV; > > - mutex_lock(&dev_priv->fbc.lock); > > + mutex_lock(&i915->fbc.lock); > > reg = I915_READ(ILK_DPFC_CONTROL); > > - dev_priv->fbc.false_color = val; > > + i915->fbc.false_color = val; > > I915_WRITE(ILK_DPFC_CONTROL, val ? > > (reg | FBC_CTL_FALSE_COLOR) : > > (reg & ~FBC_CTL_FALSE_COLOR)); > > - mutex_unlock(&dev_priv->fbc.lock); > > + mutex_unlock(&i915->fbc.lock); > > return 0; > > } > > @@ -1460,18 +1460,18 @@ DEFINE_SIMPLE_ATTRIBUTE(i915_fbc_false_color_fops, > > static int i915_ips_status(struct seq_file *m, void *unused) > > { > > - struct drm_i915_private *dev_priv = node_to_i915(m->private); > > + struct drm_i915_private *i915 = node_to_i915(m->private); > > intel_wakeref_t wakeref; > > - if (!HAS_IPS(dev_priv)) > > + if (!HAS_IPS(i915)) > > return -ENODEV; > > - wakeref = intel_runtime_pm_get(dev_priv); > > + wakeref = intel_runtime_pm_get(i915); > > seq_printf(m, "Enabled by kernel parameter: %s\n", > > yesno(i915_modparams.enable_ips)); > > - if (INTEL_GEN(dev_priv) >= 8) { > > + if (INTEL_GEN(i915) >= 8) { > > seq_puts(m, "Currently: unknown\n"); > > } else { > > if (I915_READ(IPS_CTL) & IPS_ENABLE) > > @@ -1480,34 +1480,34 @@ static int i915_ips_status(struct seq_file *m, void *unused) > > seq_puts(m, "Currently: disabled\n"); > > } > > - intel_runtime_pm_put(dev_priv, wakeref); > > + intel_runtime_pm_put(i915, wakeref); > > return 0; > > } > > static int i915_sr_status(struct seq_file *m, void *unused) > > { > > - struct drm_i915_private *dev_priv = node_to_i915(m->private); > > + struct drm_i915_private *i915 = node_to_i915(m->private); > > intel_wakeref_t wakeref; > > bool sr_enabled = false; > > - wakeref = intel_display_power_get(dev_priv, POWER_DOMAIN_INIT); > > + wakeref = intel_display_power_get(i915, POWER_DOMAIN_INIT); > > - if (INTEL_GEN(dev_priv) >= 9) > > + if (INTEL_GEN(i915) >= 9) > > /* no global SR status; inspect per-plane WM */; > > - else if (HAS_PCH_SPLIT(dev_priv)) > > + else if (HAS_PCH_SPLIT(i915)) > > sr_enabled = I915_READ(WM1_LP_ILK) & WM1_LP_SR_EN; > > - else if (IS_I965GM(dev_priv) || IS_G4X(dev_priv) || > > - IS_I945G(dev_priv) || IS_I945GM(dev_priv)) > > + else if (IS_I965GM(i915) || IS_G4X(i915) || > > + IS_I945G(i915) || IS_I945GM(i915)) > > sr_enabled = I915_READ(FW_BLC_SELF) & FW_BLC_SELF_EN; > > - else if (IS_I915GM(dev_priv)) > > + else if (IS_I915GM(i915)) > > sr_enabled = I915_READ(INSTPM) & INSTPM_SELF_EN; > > - else if (IS_PINEVIEW(dev_priv)) > > + else if (IS_PINEVIEW(i915)) > > sr_enabled = I915_READ(DSPFW3) & PINEVIEW_SELF_REFRESH_EN; > > - else if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) > > + else if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) > > sr_enabled = I915_READ(FW_BLC_SELF_VLV) & FW_CSPWRDWNEN; > > - intel_display_power_put(dev_priv, POWER_DOMAIN_INIT, wakeref); > > + intel_display_power_put(i915, POWER_DOMAIN_INIT, wakeref); > > seq_printf(m, "self-refresh: %s\n", enableddisabled(sr_enabled)); > > @@ -1540,18 +1540,18 @@ static int i915_emon_status(struct seq_file *m, void *unused) > > static int i915_ring_freq_table(struct seq_file *m, void *unused) > > { > > - struct drm_i915_private *dev_priv = node_to_i915(m->private); > > - struct intel_rps *rps = &dev_priv->gt_pm.rps; > > + struct drm_i915_private *i915 = node_to_i915(m->private); > > + struct intel_rps *rps = &i915->gt_pm.rps; > > unsigned int max_gpu_freq, min_gpu_freq; > > intel_wakeref_t wakeref; > > int gpu_freq, ia_freq; > > - if (!HAS_LLC(dev_priv)) > > + if (!HAS_LLC(i915)) > > return -ENODEV; > > min_gpu_freq = rps->min_freq; > > max_gpu_freq = rps->max_freq; > > - if (IS_GEN9_BC(dev_priv) || INTEL_GEN(dev_priv) >= 10) { > > + if (IS_GEN9_BC(i915) || INTEL_GEN(i915) >= 10) { > > /* Convert GT frequency to 50 HZ units */ > > min_gpu_freq /= GEN9_FREQ_SCALER; > > max_gpu_freq /= GEN9_FREQ_SCALER; > > @@ -1559,30 +1559,30 @@ static int i915_ring_freq_table(struct seq_file *m, void *unused) > > seq_puts(m, "GPU freq (MHz)\tEffective CPU freq (MHz)\tEffective Ring freq (MHz)\n"); > > - wakeref = intel_runtime_pm_get(dev_priv); > > + wakeref = intel_runtime_pm_get(i915); > > for (gpu_freq = min_gpu_freq; gpu_freq <= max_gpu_freq; gpu_freq++) { > > ia_freq = gpu_freq; > > - sandybridge_pcode_read(dev_priv, > > + sandybridge_pcode_read(i915, > > GEN6_PCODE_READ_MIN_FREQ_TABLE, > > &ia_freq, NULL); > > seq_printf(m, "%d\t\t%d\t\t\t\t%d\n", > > - intel_gpu_freq(dev_priv, (gpu_freq * > > - (IS_GEN9_BC(dev_priv) || > > - INTEL_GEN(dev_priv) >= 10 ? > > + intel_gpu_freq(i915, (gpu_freq * > > + (IS_GEN9_BC(i915) || > > + INTEL_GEN(i915) >= 10 ? > > GEN9_FREQ_SCALER : 1))), > > ((ia_freq >> 0) & 0xff) * 100, > > ((ia_freq >> 8) & 0xff) * 100); > > } > > - intel_runtime_pm_put(dev_priv, wakeref); > > + intel_runtime_pm_put(i915, wakeref); > > return 0; > > } > > static int i915_opregion(struct seq_file *m, void *unused) > > { > > - struct drm_i915_private *dev_priv = node_to_i915(m->private); > > - struct drm_device *dev = &dev_priv->drm; > > - struct intel_opregion *opregion = &dev_priv->opregion; > > + struct drm_i915_private *i915 = node_to_i915(m->private); > > + struct drm_device *dev = &i915->drm; > > + struct intel_opregion *opregion = &i915->opregion; > > int ret; > > ret = mutex_lock_interruptible(&dev->struct_mutex); > > @@ -1610,8 +1610,8 @@ static int i915_vbt(struct seq_file *m, void *unused) > > static int i915_gem_framebuffer_info(struct seq_file *m, void *data) > > { > > - struct drm_i915_private *dev_priv = node_to_i915(m->private); > > - struct drm_device *dev = &dev_priv->drm; > > + struct drm_i915_private *i915 = node_to_i915(m->private); > > + struct drm_device *dev = &i915->drm; > > struct intel_framebuffer *fbdev_fb = NULL; > > struct drm_framebuffer *drm_fb; > > int ret; > > @@ -1621,8 +1621,8 @@ static int i915_gem_framebuffer_info(struct seq_file *m, void *data) > > return ret; > > #ifdef CONFIG_DRM_FBDEV_EMULATION > > - if (dev_priv->fbdev && dev_priv->fbdev->helper.fb) { > > - fbdev_fb = to_intel_framebuffer(dev_priv->fbdev->helper.fb); > > + if (i915->fbdev && i915->fbdev->helper.fb) { > > + fbdev_fb = to_intel_framebuffer(i915->fbdev->helper.fb); > > seq_printf(m, "fbcon size: %d x %d, depth %d, %d bpp, modifier 0x%llx, refcount %d, obj ", > > fbdev_fb->base.width, > > @@ -1666,8 +1666,8 @@ static void describe_ctx_ring(struct seq_file *m, struct intel_ring *ring) > > static int i915_context_status(struct seq_file *m, void *unused) > > { > > - struct drm_i915_private *dev_priv = node_to_i915(m->private); > > - struct drm_device *dev = &dev_priv->drm; > > + struct drm_i915_private *i915 = node_to_i915(m->private); > > + struct drm_device *dev = &i915->drm; > > struct i915_gem_context *ctx; > > int ret; > > @@ -1675,7 +1675,7 @@ static int i915_context_status(struct seq_file *m, void *unused) > > if (ret) > > return ret; > > - list_for_each_entry(ctx, &dev_priv->contexts.list, link) { > > + list_for_each_entry(ctx, &i915->contexts.list, link) { > > struct i915_gem_engines_iter it; > > struct intel_context *ce; > > @@ -1746,18 +1746,18 @@ static const char *swizzle_string(unsigned swizzle) > > static int i915_swizzle_info(struct seq_file *m, void *data) > > { > > - struct drm_i915_private *dev_priv = node_to_i915(m->private); > > - struct intel_uncore *uncore = &dev_priv->uncore; > > + struct drm_i915_private *i915 = node_to_i915(m->private); > > + struct intel_uncore *uncore = &i915->uncore; > > intel_wakeref_t wakeref; > > - wakeref = intel_runtime_pm_get(dev_priv); > > + wakeref = intel_runtime_pm_get(i915); > > seq_printf(m, "bit6 swizzle for X-tiling = %s\n", > > - swizzle_string(dev_priv->mm.bit_6_swizzle_x)); > > + swizzle_string(i915->mm.bit_6_swizzle_x)); > > seq_printf(m, "bit6 swizzle for Y-tiling = %s\n", > > - swizzle_string(dev_priv->mm.bit_6_swizzle_y)); > > + swizzle_string(i915->mm.bit_6_swizzle_y)); > > - if (IS_GEN_RANGE(dev_priv, 3, 4)) { > > + if (IS_GEN_RANGE(i915, 3, 4)) { > > seq_printf(m, "DDC = 0x%08x\n", > > intel_uncore_read(uncore, DCC)); > > seq_printf(m, "DDC2 = 0x%08x\n", > > @@ -1766,7 +1766,7 @@ static int i915_swizzle_info(struct seq_file *m, void *data) > > intel_uncore_read16(uncore, C0DRB3)); > > seq_printf(m, "C1DRB3 = 0x%04x\n", > > intel_uncore_read16(uncore, C1DRB3)); > > - } else if (INTEL_GEN(dev_priv) >= 6) { > > + } else if (INTEL_GEN(i915) >= 6) { > > seq_printf(m, "MAD_DIMM_C0 = 0x%08x\n", > > intel_uncore_read(uncore, MAD_DIMM_C0)); > > seq_printf(m, "MAD_DIMM_C1 = 0x%08x\n", > > @@ -1775,7 +1775,7 @@ static int i915_swizzle_info(struct seq_file *m, void *data) > > intel_uncore_read(uncore, MAD_DIMM_C2)); > > seq_printf(m, "TILECTL = 0x%08x\n", > > intel_uncore_read(uncore, TILECTL)); > > - if (INTEL_GEN(dev_priv) >= 8) > > + if (INTEL_GEN(i915) >= 8) > > seq_printf(m, "GAMTARBMODE = 0x%08x\n", > > intel_uncore_read(uncore, GAMTARBMODE)); > > else > > @@ -1785,10 +1785,10 @@ static int i915_swizzle_info(struct seq_file *m, void *data) > > intel_uncore_read(uncore, DISP_ARB_CTL)); > > } > > - if (dev_priv->quirks & QUIRK_PIN_SWIZZLED_PAGES) > > + if (i915->quirks & QUIRK_PIN_SWIZZLED_PAGES) > > seq_puts(m, "L-shaped memory detected\n"); > > - intel_runtime_pm_put(dev_priv, wakeref); > > + intel_runtime_pm_put(i915, wakeref); > > return 0; > > } > > @@ -1809,54 +1809,54 @@ static const char *rps_power_to_str(unsigned int power) > > static int i915_rps_boost_info(struct seq_file *m, void *data) > > { > > - struct drm_i915_private *dev_priv = node_to_i915(m->private); > > - struct intel_rps *rps = &dev_priv->gt_pm.rps; > > + struct drm_i915_private *i915 = node_to_i915(m->private); > > + struct intel_rps *rps = &i915->gt_pm.rps; > > u32 act_freq = rps->cur_freq; > > intel_wakeref_t wakeref; > > - with_intel_runtime_pm_if_in_use(dev_priv, wakeref) { > > - if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) { > > - vlv_punit_get(dev_priv); > > - act_freq = vlv_punit_read(dev_priv, > > + with_intel_runtime_pm_if_in_use(i915, wakeref) { > > + if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) { > > + vlv_punit_get(i915); > > + act_freq = vlv_punit_read(i915, > > PUNIT_REG_GPU_FREQ_STS); > > - vlv_punit_put(dev_priv); > > + vlv_punit_put(i915); > > act_freq = (act_freq >> 8) & 0xff; > > } else { > > - act_freq = intel_get_cagf(dev_priv, > > + act_freq = intel_get_cagf(i915, > > I915_READ(GEN6_RPSTAT1)); > > } > > } > > seq_printf(m, "RPS enabled? %d\n", rps->enabled); > > - seq_printf(m, "GPU busy? %s\n", yesno(dev_priv->gt.awake)); > > + seq_printf(m, "GPU busy? %s\n", yesno(i915->gt.awake)); > > seq_printf(m, "Boosts outstanding? %d\n", > > atomic_read(&rps->num_waiters)); > > seq_printf(m, "Interactive? %d\n", READ_ONCE(rps->power.interactive)); > > seq_printf(m, "Frequency requested %d, actual %d\n", > > - intel_gpu_freq(dev_priv, rps->cur_freq), > > - intel_gpu_freq(dev_priv, act_freq)); > > + intel_gpu_freq(i915, rps->cur_freq), > > + intel_gpu_freq(i915, act_freq)); > > seq_printf(m, " min hard:%d, soft:%d; max soft:%d, hard:%d\n", > > - intel_gpu_freq(dev_priv, rps->min_freq), > > - intel_gpu_freq(dev_priv, rps->min_freq_softlimit), > > - intel_gpu_freq(dev_priv, rps->max_freq_softlimit), > > - intel_gpu_freq(dev_priv, rps->max_freq)); > > + intel_gpu_freq(i915, rps->min_freq), > > + intel_gpu_freq(i915, rps->min_freq_softlimit), > > + intel_gpu_freq(i915, rps->max_freq_softlimit), > > + intel_gpu_freq(i915, rps->max_freq)); > > seq_printf(m, " idle:%d, efficient:%d, boost:%d\n", > > - intel_gpu_freq(dev_priv, rps->idle_freq), > > - intel_gpu_freq(dev_priv, rps->efficient_freq), > > - intel_gpu_freq(dev_priv, rps->boost_freq)); > > + intel_gpu_freq(i915, rps->idle_freq), > > + intel_gpu_freq(i915, rps->efficient_freq), > > + intel_gpu_freq(i915, rps->boost_freq)); > > seq_printf(m, "Wait boosts: %d\n", atomic_read(&rps->boosts)); > > - if (INTEL_GEN(dev_priv) >= 6 && rps->enabled && dev_priv->gt.awake) { > > + if (INTEL_GEN(i915) >= 6 && rps->enabled && i915->gt.awake) { > > u32 rpup, rpupei; > > u32 rpdown, rpdownei; > > - intel_uncore_forcewake_get(&dev_priv->uncore, FORCEWAKE_ALL); > > + intel_uncore_forcewake_get(&i915->uncore, FORCEWAKE_ALL); > > rpup = I915_READ_FW(GEN6_RP_CUR_UP) & GEN6_RP_EI_MASK; > > rpupei = I915_READ_FW(GEN6_RP_CUR_UP_EI) & GEN6_RP_EI_MASK; > > rpdown = I915_READ_FW(GEN6_RP_CUR_DOWN) & GEN6_RP_EI_MASK; > > rpdownei = I915_READ_FW(GEN6_RP_CUR_DOWN_EI) & GEN6_RP_EI_MASK; > > - intel_uncore_forcewake_put(&dev_priv->uncore, FORCEWAKE_ALL); > > + intel_uncore_forcewake_put(&i915->uncore, FORCEWAKE_ALL); > > seq_printf(m, "\nRPS Autotuning (current \"%s\" window):\n", > > rps_power_to_str(rps->power.mode)); > > @@ -1875,29 +1875,29 @@ static int i915_rps_boost_info(struct seq_file *m, void *data) > > static int i915_llc(struct seq_file *m, void *data) > > { > > - struct drm_i915_private *dev_priv = node_to_i915(m->private); > > - const bool edram = INTEL_GEN(dev_priv) > 8; > > + struct drm_i915_private *i915 = node_to_i915(m->private); > > + const bool edram = INTEL_GEN(i915) > 8; > > - seq_printf(m, "LLC: %s\n", yesno(HAS_LLC(dev_priv))); > > + seq_printf(m, "LLC: %s\n", yesno(HAS_LLC(i915))); > > seq_printf(m, "%s: %uMB\n", edram ? "eDRAM" : "eLLC", > > - dev_priv->edram_size_mb); > > + i915->edram_size_mb); > > return 0; > > } > > static int i915_huc_load_status_info(struct seq_file *m, void *data) > > { > > - struct drm_i915_private *dev_priv = node_to_i915(m->private); > > + struct drm_i915_private *i915 = node_to_i915(m->private); > > intel_wakeref_t wakeref; > > struct drm_printer p; > > - if (!HAS_HUC(dev_priv)) > > + if (!HAS_HUC(i915)) > > return -ENODEV; > > p = drm_seq_file_printer(m); > > - intel_uc_fw_dump(&dev_priv->huc.fw, &p); > > + intel_uc_fw_dump(&i915->huc.fw, &p); > > - with_intel_runtime_pm(dev_priv, wakeref) > > + with_intel_runtime_pm(i915, wakeref) > > seq_printf(m, "\nHuC status 0x%08x:\n", I915_READ(HUC_STATUS2)); > > return 0; > > @@ -1905,17 +1905,17 @@ static int i915_huc_load_status_info(struct seq_file *m, void *data) > > static int i915_guc_load_status_info(struct seq_file *m, void *data) > > { > > - struct drm_i915_private *dev_priv = node_to_i915(m->private); > > + struct drm_i915_private *i915 = node_to_i915(m->private); > > intel_wakeref_t wakeref; > > struct drm_printer p; > > - if (!HAS_GUC(dev_priv)) > > + if (!HAS_GUC(i915)) > > return -ENODEV; > > p = drm_seq_file_printer(m); > > - intel_uc_fw_dump(&dev_priv->guc.fw, &p); > > + intel_uc_fw_dump(&i915->guc.fw, &p); > > - with_intel_runtime_pm(dev_priv, wakeref) { > > + with_intel_runtime_pm(i915, wakeref) { > > u32 tmp = I915_READ(GUC_STATUS); > > u32 i; > > @@ -1954,9 +1954,9 @@ stringify_guc_log_type(enum guc_log_buffer_type type) > > } > > static void i915_guc_log_info(struct seq_file *m, > > - struct drm_i915_private *dev_priv) > > + struct drm_i915_private *i915) > > { > > - struct intel_guc_log *log = &dev_priv->guc.log; > > + struct intel_guc_log *log = &i915->guc.log; > > enum guc_log_buffer_type type; > > if (!intel_guc_log_relay_enabled(log)) { > > @@ -1978,7 +1978,7 @@ static void i915_guc_log_info(struct seq_file *m, > > } > > static void i915_guc_client_info(struct seq_file *m, > > - struct drm_i915_private *dev_priv, > > + struct drm_i915_private *i915, > > struct intel_guc_client *client) > > { > > struct intel_engine_cs *engine; > > @@ -1990,7 +1990,7 @@ static void i915_guc_client_info(struct seq_file *m, > > seq_printf(m, "\tDoorbell id %d, offset: 0x%lx\n", > > client->doorbell_id, client->doorbell_offset); > > - for_each_engine(engine, dev_priv, id) { > > + for_each_engine(engine, i915, id) { > > u64 submissions = client->submissions[id]; > > tot += submissions; > > seq_printf(m, "\tSubmissions: %llu %s\n", > > @@ -2001,15 +2001,15 @@ static void i915_guc_client_info(struct seq_file *m, > > static int i915_guc_info(struct seq_file *m, void *data) > > { > > - struct drm_i915_private *dev_priv = node_to_i915(m->private); > > - const struct intel_guc *guc = &dev_priv->guc; > > + struct drm_i915_private *i915 = node_to_i915(m->private); > > + const struct intel_guc *guc = &i915->guc; > > - if (!USES_GUC(dev_priv)) > > + if (!USES_GUC(i915)) > > return -ENODEV; > > - i915_guc_log_info(m, dev_priv); > > + i915_guc_log_info(m, i915); > > - if (!USES_GUC_SUBMISSION(dev_priv)) > > + if (!USES_GUC_SUBMISSION(i915)) > > return 0; > > GEM_BUG_ON(!guc->execbuf_client); > > @@ -2019,11 +2019,11 @@ static int i915_guc_info(struct seq_file *m, void *data) > > seq_printf(m, "Doorbell next cacheline: 0x%x\n", guc->db_cacheline); > > seq_printf(m, "\nGuC execbuf client @ %p:\n", guc->execbuf_client); > > - i915_guc_client_info(m, dev_priv, guc->execbuf_client); > > + i915_guc_client_info(m, i915, guc->execbuf_client); > > if (guc->preempt_client) { > > seq_printf(m, "\nGuC preempt client @ %p:\n", > > guc->preempt_client); > > - i915_guc_client_info(m, dev_priv, guc->preempt_client); > > + i915_guc_client_info(m, i915, guc->preempt_client); > > } > > /* Add more as required ... */ > > @@ -2033,14 +2033,14 @@ static int i915_guc_info(struct seq_file *m, void *data) > > static int i915_guc_stage_pool(struct seq_file *m, void *data) > > { > > - struct drm_i915_private *dev_priv = node_to_i915(m->private); > > - const struct intel_guc *guc = &dev_priv->guc; > > + struct drm_i915_private *i915 = node_to_i915(m->private); > > + const struct intel_guc *guc = &i915->guc; > > struct guc_stage_desc *desc = guc->stage_desc_pool_vaddr; > > struct intel_guc_client *client = guc->execbuf_client; > > intel_engine_mask_t tmp; > > int index; > > - if (!USES_GUC_SUBMISSION(dev_priv)) > > + if (!USES_GUC_SUBMISSION(i915)) > > return -ENODEV; > > for (index = 0; index < GUC_MAX_STAGE_DESCRIPTORS; index++, desc++) { > > @@ -2066,7 +2066,7 @@ static int i915_guc_stage_pool(struct seq_file *m, void *data) > > desc->wq_addr, desc->wq_size); > > seq_putc(m, '\n'); > > - for_each_engine_masked(engine, dev_priv, client->engines, tmp) { > > + for_each_engine_masked(engine, i915, client->engines, tmp) { > > u32 guc_engine_id = engine->guc_id; > > struct guc_execlist_context *lrc = > > &desc->lrc[guc_engine_id]; > > @@ -2088,19 +2088,19 @@ static int i915_guc_stage_pool(struct seq_file *m, void *data) > > static int i915_guc_log_dump(struct seq_file *m, void *data) > > { > > struct drm_info_node *node = m->private; > > - struct drm_i915_private *dev_priv = node_to_i915(node); > > + struct drm_i915_private *i915 = node_to_i915(node); > > bool dump_load_err = !!node->info_ent->data; > > struct drm_i915_gem_object *obj = NULL; > > u32 *log; > > int i = 0; > > - if (!HAS_GUC(dev_priv)) > > + if (!HAS_GUC(i915)) > > return -ENODEV; > > if (dump_load_err) > > - obj = dev_priv->guc.load_err_log; > > - else if (dev_priv->guc.log.vma) > > - obj = dev_priv->guc.log.vma->obj; > > + obj = i915->guc.load_err_log; > > + else if (i915->guc.log.vma) > > + obj = i915->guc.log.vma->obj; > > if (!obj) > > return 0; > > @@ -2126,24 +2126,24 @@ static int i915_guc_log_dump(struct seq_file *m, void *data) > > static int i915_guc_log_level_get(void *data, u64 *val) > > { > > - struct drm_i915_private *dev_priv = data; > > + struct drm_i915_private *i915 = data; > > - if (!USES_GUC(dev_priv)) > > + if (!USES_GUC(i915)) > > return -ENODEV; > > - *val = intel_guc_log_get_level(&dev_priv->guc.log); > > + *val = intel_guc_log_get_level(&i915->guc.log); > > return 0; > > } > > static int i915_guc_log_level_set(void *data, u64 val) > > { > > - struct drm_i915_private *dev_priv = data; > > + struct drm_i915_private *i915 = data; > > - if (!USES_GUC(dev_priv)) > > + if (!USES_GUC(i915)) > > return -ENODEV; > > - return intel_guc_log_set_level(&dev_priv->guc.log, val); > > + return intel_guc_log_set_level(&i915->guc.log, val); > > } > > DEFINE_SIMPLE_ATTRIBUTE(i915_guc_log_level_fops, > > @@ -2152,14 +2152,14 @@ DEFINE_SIMPLE_ATTRIBUTE(i915_guc_log_level_fops, > > static int i915_guc_log_relay_open(struct inode *inode, struct file *file) > > { > > - struct drm_i915_private *dev_priv = inode->i_private; > > + struct drm_i915_private *i915 = inode->i_private; > > - if (!USES_GUC(dev_priv)) > > + if (!USES_GUC(i915)) > > return -ENODEV; > > - file->private_data = &dev_priv->guc.log; > > + file->private_data = &i915->guc.log; > > - return intel_guc_log_relay_open(&dev_priv->guc.log); > > + return intel_guc_log_relay_open(&i915->guc.log); > > } > > static ssize_t > > @@ -2177,9 +2177,9 @@ i915_guc_log_relay_write(struct file *filp, > > static int i915_guc_log_relay_release(struct inode *inode, struct file *file) > > { > > - struct drm_i915_private *dev_priv = inode->i_private; > > + struct drm_i915_private *i915 = inode->i_private; > > - intel_guc_log_relay_close(&dev_priv->guc.log); > > + intel_guc_log_relay_close(&i915->guc.log); > > return 0; > > } > > @@ -2205,12 +2205,12 @@ static int i915_psr_sink_status_show(struct seq_file *m, void *data) > > "sink internal error", > > }; > > struct drm_connector *connector = m->private; > > - struct drm_i915_private *dev_priv = to_i915(connector->dev); > > + struct drm_i915_private *i915 = to_i915(connector->dev); > > struct intel_dp *intel_dp = > > enc_to_intel_dp(&intel_attached_encoder(connector)->base); > > int ret; > > - if (!CAN_PSR(dev_priv)) { > > + if (!CAN_PSR(i915)) { > > seq_puts(m, "PSR Unsupported\n"); > > return -ENODEV; > > } > > @@ -2236,12 +2236,12 @@ static int i915_psr_sink_status_show(struct seq_file *m, void *data) > > DEFINE_SHOW_ATTRIBUTE(i915_psr_sink_status); > > static void > > -psr_source_status(struct drm_i915_private *dev_priv, struct seq_file *m) > > +psr_source_status(struct drm_i915_private *i915, struct seq_file *m) > > { > > u32 val, status_val; > > const char *status = "unknown"; > > - if (dev_priv->psr.psr2_enabled) { > > + if (i915->psr.psr2_enabled) { > > static const char * const live_status[] = { > > "IDLE", > > "CAPTURE", > > @@ -2283,14 +2283,14 @@ psr_source_status(struct drm_i915_private *dev_priv, struct seq_file *m) > > static int i915_edp_psr_status(struct seq_file *m, void *data) > > { > > - struct drm_i915_private *dev_priv = node_to_i915(m->private); > > - struct i915_psr *psr = &dev_priv->psr; > > + struct drm_i915_private *i915 = node_to_i915(m->private); > > + struct i915_psr *psr = &i915->psr; > > intel_wakeref_t wakeref; > > const char *status; > > bool enabled; > > u32 val; > > - if (!HAS_PSR(dev_priv)) > > + if (!HAS_PSR(i915)) > > return -ENODEV; > > seq_printf(m, "Sink support: %s", yesno(psr->sink_support)); > > @@ -2301,7 +2301,7 @@ static int i915_edp_psr_status(struct seq_file *m, void *data) > > if (!psr->sink_support) > > return 0; > > - wakeref = intel_runtime_pm_get(dev_priv); > > + wakeref = intel_runtime_pm_get(i915); > > mutex_lock(&psr->lock); > > if (psr->enabled) > > @@ -2322,14 +2322,14 @@ static int i915_edp_psr_status(struct seq_file *m, void *data) > > } > > seq_printf(m, "Source PSR ctl: %s [0x%08x]\n", > > enableddisabled(enabled), val); > > - psr_source_status(dev_priv, m); > > + psr_source_status(i915, m); > > seq_printf(m, "Busy frontbuffer bits: 0x%08x\n", > > psr->busy_frontbuffer_bits); > > /* > > * SKL+ Perf counter is reset to 0 everytime DC state is entered > > */ > > - if (IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv)) { > > + if (IS_HASWELL(i915) || IS_BROADWELL(i915)) { > > val = I915_READ(EDP_PSR_PERF_CNT) & EDP_PSR_PERF_CNT_MASK; > > seq_printf(m, "Performance counter: %u\n", val); > > } > > @@ -2365,7 +2365,7 @@ static int i915_edp_psr_status(struct seq_file *m, void *data) > > unlock: > > mutex_unlock(&psr->lock); > > - intel_runtime_pm_put(dev_priv, wakeref); > > + intel_runtime_pm_put(i915, wakeref); > > return 0; > > } > > @@ -2373,20 +2373,20 @@ static int i915_edp_psr_status(struct seq_file *m, void *data) > > static int > > i915_edp_psr_debug_set(void *data, u64 val) > > { > > - struct drm_i915_private *dev_priv = data; > > + struct drm_i915_private *i915 = data; > > intel_wakeref_t wakeref; > > int ret; > > - if (!CAN_PSR(dev_priv)) > > + if (!CAN_PSR(i915)) > > return -ENODEV; > > DRM_DEBUG_KMS("Setting PSR debug to %llx\n", val); > > - wakeref = intel_runtime_pm_get(dev_priv); > > + wakeref = intel_runtime_pm_get(i915); > > - ret = intel_psr_debug_set(dev_priv, val); > > + ret = intel_psr_debug_set(i915, val); > > - intel_runtime_pm_put(dev_priv, wakeref); > > + intel_runtime_pm_put(i915, wakeref); > > return ret; > > } > > @@ -2394,12 +2394,12 @@ i915_edp_psr_debug_set(void *data, u64 val) > > static int > > i915_edp_psr_debug_get(void *data, u64 *val) > > { > > - struct drm_i915_private *dev_priv = data; > > + struct drm_i915_private *i915 = data; > > - if (!CAN_PSR(dev_priv)) > > + if (!CAN_PSR(i915)) > > return -ENODEV; > > - *val = READ_ONCE(dev_priv->psr.debug); > > + *val = READ_ONCE(i915->psr.debug); > > return 0; > > } > > @@ -2409,19 +2409,19 @@ DEFINE_SIMPLE_ATTRIBUTE(i915_edp_psr_debug_fops, > > static int i915_energy_uJ(struct seq_file *m, void *data) > > { > > - struct drm_i915_private *dev_priv = node_to_i915(m->private); > > + struct drm_i915_private *i915 = node_to_i915(m->private); > > unsigned long long power; > > intel_wakeref_t wakeref; > > u32 units; > > - if (INTEL_GEN(dev_priv) < 6) > > + if (INTEL_GEN(i915) < 6) > > return -ENODEV; > > if (rdmsrl_safe(MSR_RAPL_POWER_UNIT, &power)) > > return -ENODEV; > > units = (power & 0x1f00) >> 8; > > - with_intel_runtime_pm(dev_priv, wakeref) > > + with_intel_runtime_pm(i915, wakeref) > > power = I915_READ(MCH_SECP_NRG_STTS); > > power = (1000000 * power) >> units; /* convert to uJ */ > > @@ -2432,21 +2432,21 @@ static int i915_energy_uJ(struct seq_file *m, void *data) > > static int i915_runtime_pm_status(struct seq_file *m, void *unused) > > { > > - struct drm_i915_private *dev_priv = node_to_i915(m->private); > > - struct pci_dev *pdev = dev_priv->drm.pdev; > > + struct drm_i915_private *i915 = node_to_i915(m->private); > > + struct pci_dev *pdev = i915->drm.pdev; > > - if (!HAS_RUNTIME_PM(dev_priv)) > > + if (!HAS_RUNTIME_PM(i915)) > > seq_puts(m, "Runtime power management not supported\n"); > > seq_printf(m, "Runtime power status: %s\n", > > - enableddisabled(!dev_priv->power_domains.wakeref)); > > + enableddisabled(!i915->power_domains.wakeref)); > > - seq_printf(m, "GPU idle: %s\n", yesno(!dev_priv->gt.awake)); > > + seq_printf(m, "GPU idle: %s\n", yesno(!i915->gt.awake)); > > seq_printf(m, "IRQs disabled: %s\n", > > - yesno(!intel_irqs_enabled(dev_priv))); > > + yesno(!intel_irqs_enabled(i915))); > > #ifdef CONFIG_PM > > seq_printf(m, "Usage count: %d\n", > > - atomic_read(&dev_priv->drm.dev->power.usage_count)); > > + atomic_read(&i915->drm.dev->power.usage_count)); > > #else > > seq_printf(m, "Device Power Management (CONFIG_PM) disabled\n"); > > #endif > > @@ -2457,7 +2457,7 @@ static int i915_runtime_pm_status(struct seq_file *m, void *unused) > > if (IS_ENABLED(CONFIG_DRM_I915_DEBUG_RUNTIME_PM)) { > > struct drm_printer p = drm_seq_file_printer(m); > > - print_intel_runtime_pm_wakeref(dev_priv, &p); > > + print_intel_runtime_pm_wakeref(i915, &p); > > } > > return 0; > > @@ -2465,8 +2465,8 @@ static int i915_runtime_pm_status(struct seq_file *m, void *unused) > > static int i915_power_domain_info(struct seq_file *m, void *unused) > > { > > - struct drm_i915_private *dev_priv = node_to_i915(m->private); > > - struct i915_power_domains *power_domains = &dev_priv->power_domains; > > + struct drm_i915_private *i915 = node_to_i915(m->private); > > + struct i915_power_domains *power_domains = &i915->power_domains; > > int i; > > mutex_lock(&power_domains->lock); > > @@ -2493,16 +2493,16 @@ static int i915_power_domain_info(struct seq_file *m, void *unused) > > static int i915_dmc_info(struct seq_file *m, void *unused) > > { > > - struct drm_i915_private *dev_priv = node_to_i915(m->private); > > + struct drm_i915_private *i915 = node_to_i915(m->private); > > intel_wakeref_t wakeref; > > struct intel_csr *csr; > > - if (!HAS_CSR(dev_priv)) > > + if (!HAS_CSR(i915)) > > return -ENODEV; > > - csr = &dev_priv->csr; > > + csr = &i915->csr; > > - wakeref = intel_runtime_pm_get(dev_priv); > > + wakeref = intel_runtime_pm_get(i915); > > seq_printf(m, "fw loaded: %s\n", yesno(csr->dmc_payload != NULL)); > > seq_printf(m, "path: %s\n", csr->fw_path); > > @@ -2513,13 +2513,13 @@ static int i915_dmc_info(struct seq_file *m, void *unused) > > seq_printf(m, "version: %d.%d\n", CSR_VERSION_MAJOR(csr->version), > > CSR_VERSION_MINOR(csr->version)); > > - if (WARN_ON(INTEL_GEN(dev_priv) > 11)) > > + if (WARN_ON(INTEL_GEN(i915) > 11)) > > goto out; > > seq_printf(m, "DC3 -> DC5 count: %d\n", > > - I915_READ(IS_BROXTON(dev_priv) ? BXT_CSR_DC3_DC5_COUNT : > > + I915_READ(IS_BROXTON(i915) ? BXT_CSR_DC3_DC5_COUNT : > > SKL_CSR_DC3_DC5_COUNT)); > > - if (!IS_GEN9_LP(dev_priv)) > > + if (!IS_GEN9_LP(i915)) > > seq_printf(m, "DC5 -> DC6 count: %d\n", > > I915_READ(SKL_CSR_DC5_DC6_COUNT)); > > @@ -2528,7 +2528,7 @@ static int i915_dmc_info(struct seq_file *m, void *unused) > > seq_printf(m, "ssp base: 0x%08x\n", I915_READ(CSR_SSP_BASE)); > > seq_printf(m, "htp: 0x%08x\n", I915_READ(CSR_HTP_SKL)); > > - intel_runtime_pm_put(dev_priv, wakeref); > > + intel_runtime_pm_put(i915, wakeref); > > return 0; > > } > > @@ -2548,8 +2548,8 @@ static void intel_encoder_info(struct seq_file *m, > > struct intel_crtc *intel_crtc, > > struct intel_encoder *intel_encoder) > > { > > - struct drm_i915_private *dev_priv = node_to_i915(m->private); > > - struct drm_device *dev = &dev_priv->drm; > > + struct drm_i915_private *i915 = node_to_i915(m->private); > > + struct drm_device *dev = &i915->drm; > > struct drm_crtc *crtc = &intel_crtc->base; > > struct intel_connector *intel_connector; > > struct drm_encoder *encoder; > > @@ -2575,8 +2575,8 @@ static void intel_encoder_info(struct seq_file *m, > > static void intel_crtc_info(struct seq_file *m, struct intel_crtc *intel_crtc) > > { > > - struct drm_i915_private *dev_priv = node_to_i915(m->private); > > - struct drm_device *dev = &dev_priv->drm; > > + struct drm_i915_private *i915 = node_to_i915(m->private); > > + struct drm_device *dev = &i915->drm; > > struct drm_crtc *crtc = &intel_crtc->base; > > struct intel_encoder *intel_encoder; > > struct drm_plane_state *plane_state = crtc->primary->state; > > @@ -2731,8 +2731,8 @@ static void plane_rotation(char *buf, size_t bufsize, unsigned int rotation) > > static void intel_plane_info(struct seq_file *m, struct intel_crtc *intel_crtc) > > { > > - struct drm_i915_private *dev_priv = node_to_i915(m->private); > > - struct drm_device *dev = &dev_priv->drm; > > + struct drm_i915_private *i915 = node_to_i915(m->private); > > + struct drm_device *dev = &i915->drm; > > struct intel_plane *intel_plane; > > for_each_intel_plane_on_crtc(dev, intel_crtc, intel_plane) { > > @@ -2805,14 +2805,14 @@ static void intel_scaler_info(struct seq_file *m, struct intel_crtc *intel_crtc) > > static int i915_display_info(struct seq_file *m, void *unused) > > { > > - struct drm_i915_private *dev_priv = node_to_i915(m->private); > > - struct drm_device *dev = &dev_priv->drm; > > + struct drm_i915_private *i915 = node_to_i915(m->private); > > + struct drm_device *dev = &i915->drm; > > struct intel_crtc *crtc; > > struct drm_connector *connector; > > struct drm_connector_list_iter conn_iter; > > intel_wakeref_t wakeref; > > - wakeref = intel_runtime_pm_get(dev_priv); > > + wakeref = intel_runtime_pm_get(i915); > > seq_printf(m, "CRTC info\n"); > > seq_printf(m, "---------\n"); > > @@ -2861,42 +2861,42 @@ static int i915_display_info(struct seq_file *m, void *unused) > > drm_connector_list_iter_end(&conn_iter); > > mutex_unlock(&dev->mode_config.mutex); > > - intel_runtime_pm_put(dev_priv, wakeref); > > + intel_runtime_pm_put(i915, wakeref); > > return 0; > > } > > static int i915_engine_info(struct seq_file *m, void *unused) > > { > > - struct drm_i915_private *dev_priv = node_to_i915(m->private); > > + struct drm_i915_private *i915 = node_to_i915(m->private); > > struct intel_engine_cs *engine; > > intel_wakeref_t wakeref; > > enum intel_engine_id id; > > struct drm_printer p; > > - wakeref = intel_runtime_pm_get(dev_priv); > > + wakeref = intel_runtime_pm_get(i915); > > seq_printf(m, "GT awake? %s [%d]\n", > > - yesno(dev_priv->gt.awake), > > - atomic_read(&dev_priv->gt.wakeref.count)); > > + yesno(i915->gt.awake), > > + atomic_read(&i915->gt.wakeref.count)); > > seq_printf(m, "CS timestamp frequency: %u kHz\n", > > - RUNTIME_INFO(dev_priv)->cs_timestamp_frequency_khz); > > + RUNTIME_INFO(i915)->cs_timestamp_frequency_khz); > > p = drm_seq_file_printer(m); > > - for_each_engine(engine, dev_priv, id) > > + for_each_engine(engine, i915, id) > > intel_engine_dump(engine, &p, "%s\n", engine->name); > > - intel_runtime_pm_put(dev_priv, wakeref); > > + intel_runtime_pm_put(i915, wakeref); > > return 0; > > } > > static int i915_rcs_topology(struct seq_file *m, void *unused) > > { > > - struct drm_i915_private *dev_priv = node_to_i915(m->private); > > + struct drm_i915_private *i915 = node_to_i915(m->private); > > struct drm_printer p = drm_seq_file_printer(m); > > - intel_device_info_dump_topology(&RUNTIME_INFO(dev_priv)->sseu, &p); > > + intel_device_info_dump_topology(&RUNTIME_INFO(i915)->sseu, &p); > > return 0; > > } > > @@ -2913,13 +2913,13 @@ static int i915_shrinker_info(struct seq_file *m, void *unused) > > static int i915_shared_dplls_info(struct seq_file *m, void *unused) > > { > > - struct drm_i915_private *dev_priv = node_to_i915(m->private); > > - struct drm_device *dev = &dev_priv->drm; > > + struct drm_i915_private *i915 = node_to_i915(m->private); > > + struct drm_device *dev = &i915->drm; > > int i; > > drm_modeset_lock_all(dev); > > - for (i = 0; i < dev_priv->num_shared_dpll; i++) { > > - struct intel_shared_dpll *pll = &dev_priv->shared_dplls[i]; > > + for (i = 0; i < i915->num_shared_dpll; i++) { > > + struct intel_shared_dpll *pll = &i915->shared_dplls[i]; > > seq_printf(m, "DPLL%i: %s, id: %i\n", i, pll->info->name, > > pll->info->id); > > @@ -2977,28 +2977,28 @@ static int i915_wa_registers(struct seq_file *m, void *unused) > > static int i915_ipc_status_show(struct seq_file *m, void *data) > > { > > - struct drm_i915_private *dev_priv = m->private; > > + struct drm_i915_private *i915 = m->private; > > seq_printf(m, "Isochronous Priority Control: %s\n", > > - yesno(dev_priv->ipc_enabled)); > > + yesno(i915->ipc_enabled)); > > return 0; > > } > > static int i915_ipc_status_open(struct inode *inode, struct file *file) > > { > > - struct drm_i915_private *dev_priv = inode->i_private; > > + struct drm_i915_private *i915 = inode->i_private; > > - if (!HAS_IPC(dev_priv)) > > + if (!HAS_IPC(i915)) > > return -ENODEV; > > - return single_open(file, i915_ipc_status_show, dev_priv); > > + return single_open(file, i915_ipc_status_show, i915); > > } > > static ssize_t i915_ipc_status_write(struct file *file, const char __user *ubuf, > > size_t len, loff_t *offp) > > { > > struct seq_file *m = file->private_data; > > - struct drm_i915_private *dev_priv = m->private; > > + struct drm_i915_private *i915 = m->private; > > intel_wakeref_t wakeref; > > bool enable; > > int ret; > > @@ -3007,12 +3007,12 @@ static ssize_t i915_ipc_status_write(struct file *file, const char __user *ubuf, > > if (ret < 0) > > return ret; > > - with_intel_runtime_pm(dev_priv, wakeref) { > > - if (!dev_priv->ipc_enabled && enable) > > + with_intel_runtime_pm(i915, wakeref) { > > + if (!i915->ipc_enabled && enable) > > DRM_INFO("Enabling IPC: WM will be proper only after next commit\n"); > > - dev_priv->wm.distrust_bios_wm = true; > > - dev_priv->ipc_enabled = enable; > > - intel_enable_ipc(dev_priv); > > + i915->wm.distrust_bios_wm = true; > > + i915->ipc_enabled = enable; > > + intel_enable_ipc(i915); > > } > > return len; > > @@ -3029,19 +3029,19 @@ static const struct file_operations i915_ipc_status_fops = { > > static int i915_ddb_info(struct seq_file *m, void *unused) > > { > > - struct drm_i915_private *dev_priv = node_to_i915(m->private); > > - struct drm_device *dev = &dev_priv->drm; > > + struct drm_i915_private *i915 = node_to_i915(m->private); > > + struct drm_device *dev = &i915->drm; > > struct skl_ddb_entry *entry; > > struct intel_crtc *crtc; > > - if (INTEL_GEN(dev_priv) < 9) > > + if (INTEL_GEN(i915) < 9) > > return -ENODEV; > > drm_modeset_lock_all(dev); > > seq_printf(m, "%-15s%8s%8s%8s\n", "", "Start", "End", "Size"); > > - for_each_intel_crtc(&dev_priv->drm, crtc) { > > + for_each_intel_crtc(&i915->drm, crtc) { > > struct intel_crtc_state *crtc_state = > > to_intel_crtc_state(crtc->base.state); > > enum pipe pipe = crtc->pipe; > > @@ -3070,8 +3070,8 @@ static void drrs_status_per_crtc(struct seq_file *m, > > struct drm_device *dev, > > struct intel_crtc *intel_crtc) > > { > > - struct drm_i915_private *dev_priv = to_i915(dev); > > - struct i915_drrs *drrs = &dev_priv->drrs; > > + struct drm_i915_private *i915 = to_i915(dev); > > + struct i915_drrs *drrs = &i915->drrs; > > int vrefresh = 0; > > struct drm_connector *connector; > > struct drm_connector_list_iter conn_iter; > > @@ -3085,11 +3085,11 @@ static void drrs_status_per_crtc(struct seq_file *m, > > } > > drm_connector_list_iter_end(&conn_iter); > > - if (dev_priv->vbt.drrs_type == STATIC_DRRS_SUPPORT) > > + if (i915->vbt.drrs_type == STATIC_DRRS_SUPPORT) > > seq_puts(m, "\tVBT: DRRS_type: Static"); > > - else if (dev_priv->vbt.drrs_type == SEAMLESS_DRRS_SUPPORT) > > + else if (i915->vbt.drrs_type == SEAMLESS_DRRS_SUPPORT) > > seq_puts(m, "\tVBT: DRRS_type: Seamless"); > > - else if (dev_priv->vbt.drrs_type == DRRS_NOT_SUPPORTED) > > + else if (i915->vbt.drrs_type == DRRS_NOT_SUPPORTED) > > seq_puts(m, "\tVBT: DRRS_type: None"); > > else > > seq_puts(m, "\tVBT: DRRS_type: FIXME: Unrecognized Value"); > > @@ -3106,7 +3106,7 @@ static void drrs_status_per_crtc(struct seq_file *m, > > /* disable_drrs() will make drrs->dp NULL */ > > if (!drrs->dp) { > > seq_puts(m, "Idleness DRRS: Disabled\n"); > > - if (dev_priv->psr.enabled) > > + if (i915->psr.enabled) > > seq_puts(m, > > "\tAs PSR is enabled, DRRS is not enabled\n"); > > mutex_unlock(&drrs->mutex); > > @@ -3143,8 +3143,8 @@ static void drrs_status_per_crtc(struct seq_file *m, > > static int i915_drrs_status(struct seq_file *m, void *unused) > > { > > - struct drm_i915_private *dev_priv = node_to_i915(m->private); > > - struct drm_device *dev = &dev_priv->drm; > > + struct drm_i915_private *i915 = node_to_i915(m->private); > > + struct drm_device *dev = &i915->drm; > > struct intel_crtc *intel_crtc; > > int active_crtc_cnt = 0; > > @@ -3167,8 +3167,8 @@ static int i915_drrs_status(struct seq_file *m, void *unused) > > static int i915_dp_mst_info(struct seq_file *m, void *unused) > > { > > - struct drm_i915_private *dev_priv = node_to_i915(m->private); > > - struct drm_device *dev = &dev_priv->drm; > > + struct drm_i915_private *i915 = node_to_i915(m->private); > > + struct drm_device *dev = &i915->drm; > > struct intel_encoder *intel_encoder; > > struct intel_digital_port *intel_dig_port; > > struct drm_connector *connector; > > @@ -3257,8 +3257,8 @@ static ssize_t i915_displayport_test_active_write(struct file *file, > > static int i915_displayport_test_active_show(struct seq_file *m, void *data) > > { > > - struct drm_i915_private *dev_priv = m->private; > > - struct drm_device *dev = &dev_priv->drm; > > + struct drm_i915_private *i915 = m->private; > > + struct drm_device *dev = &i915->drm; > > struct drm_connector *connector; > > struct drm_connector_list_iter conn_iter; > > struct intel_dp *intel_dp; > > @@ -3307,8 +3307,8 @@ static const struct file_operations i915_displayport_test_active_fops = { > > static int i915_displayport_test_data_show(struct seq_file *m, void *data) > > { > > - struct drm_i915_private *dev_priv = m->private; > > - struct drm_device *dev = &dev_priv->drm; > > + struct drm_i915_private *i915 = m->private; > > + struct drm_device *dev = &i915->drm; > > struct drm_connector *connector; > > struct drm_connector_list_iter conn_iter; > > struct intel_dp *intel_dp; > > @@ -3351,8 +3351,8 @@ DEFINE_SHOW_ATTRIBUTE(i915_displayport_test_data); > > static int i915_displayport_test_type_show(struct seq_file *m, void *data) > > { > > - struct drm_i915_private *dev_priv = m->private; > > - struct drm_device *dev = &dev_priv->drm; > > + struct drm_i915_private *i915 = m->private; > > + struct drm_device *dev = &i915->drm; > > struct drm_connector *connector; > > struct drm_connector_list_iter conn_iter; > > struct intel_dp *intel_dp; > > @@ -3383,19 +3383,19 @@ DEFINE_SHOW_ATTRIBUTE(i915_displayport_test_type); > > static void wm_latency_show(struct seq_file *m, const u16 wm[8]) > > { > > - struct drm_i915_private *dev_priv = m->private; > > - struct drm_device *dev = &dev_priv->drm; > > + struct drm_i915_private *i915 = m->private; > > + struct drm_device *dev = &i915->drm; > > int level; > > int num_levels; > > - if (IS_CHERRYVIEW(dev_priv)) > > + if (IS_CHERRYVIEW(i915)) > > num_levels = 3; > > - else if (IS_VALLEYVIEW(dev_priv)) > > + else if (IS_VALLEYVIEW(i915)) > > num_levels = 1; > > - else if (IS_G4X(dev_priv)) > > + else if (IS_G4X(i915)) > > num_levels = 3; > > else > > - num_levels = ilk_wm_max_level(dev_priv) + 1; > > + num_levels = ilk_wm_max_level(i915) + 1; > > drm_modeset_lock_all(dev); > > @@ -3406,10 +3406,10 @@ static void wm_latency_show(struct seq_file *m, const u16 wm[8]) > > * - WM1+ latency values in 0.5us units > > * - latencies are in us on gen9/vlv/chv > > */ > > - if (INTEL_GEN(dev_priv) >= 9 || > > - IS_VALLEYVIEW(dev_priv) || > > - IS_CHERRYVIEW(dev_priv) || > > - IS_G4X(dev_priv)) > > + if (INTEL_GEN(i915) >= 9 || > > + IS_VALLEYVIEW(i915) || > > + IS_CHERRYVIEW(i915) || > > + IS_G4X(i915)) > > latency *= 10; > > else if (level > 0) > > latency *= 5; > > @@ -3423,13 +3423,13 @@ static void wm_latency_show(struct seq_file *m, const u16 wm[8]) > > static int pri_wm_latency_show(struct seq_file *m, void *data) > > { > > - struct drm_i915_private *dev_priv = m->private; > > + struct drm_i915_private *i915 = m->private; > > const u16 *latencies; > > - if (INTEL_GEN(dev_priv) >= 9) > > - latencies = dev_priv->wm.skl_latency; > > + if (INTEL_GEN(i915) >= 9) > > + latencies = i915->wm.skl_latency; > > else > > - latencies = dev_priv->wm.pri_latency; > > + latencies = i915->wm.pri_latency; > > wm_latency_show(m, latencies); > > @@ -3438,13 +3438,13 @@ static int pri_wm_latency_show(struct seq_file *m, void *data) > > static int spr_wm_latency_show(struct seq_file *m, void *data) > > { > > - struct drm_i915_private *dev_priv = m->private; > > + struct drm_i915_private *i915 = m->private; > > const u16 *latencies; > > - if (INTEL_GEN(dev_priv) >= 9) > > - latencies = dev_priv->wm.skl_latency; > > + if (INTEL_GEN(i915) >= 9) > > + latencies = i915->wm.skl_latency; > > else > > - latencies = dev_priv->wm.spr_latency; > > + latencies = i915->wm.spr_latency; > > wm_latency_show(m, latencies); > > @@ -3453,13 +3453,13 @@ static int spr_wm_latency_show(struct seq_file *m, void *data) > > static int cur_wm_latency_show(struct seq_file *m, void *data) > > { > > - struct drm_i915_private *dev_priv = m->private; > > + struct drm_i915_private *i915 = m->private; > > const u16 *latencies; > > - if (INTEL_GEN(dev_priv) >= 9) > > - latencies = dev_priv->wm.skl_latency; > > + if (INTEL_GEN(i915) >= 9) > > + latencies = i915->wm.skl_latency; > > else > > - latencies = dev_priv->wm.cur_latency; > > + latencies = i915->wm.cur_latency; > > wm_latency_show(m, latencies); > > @@ -3468,54 +3468,54 @@ static int cur_wm_latency_show(struct seq_file *m, void *data) > > static int pri_wm_latency_open(struct inode *inode, struct file *file) > > { > > - struct drm_i915_private *dev_priv = inode->i_private; > > + struct drm_i915_private *i915 = inode->i_private; > > - if (INTEL_GEN(dev_priv) < 5 && !IS_G4X(dev_priv)) > > + if (INTEL_GEN(i915) < 5 && !IS_G4X(i915)) > > return -ENODEV; > > - return single_open(file, pri_wm_latency_show, dev_priv); > > + return single_open(file, pri_wm_latency_show, i915); > > } > > static int spr_wm_latency_open(struct inode *inode, struct file *file) > > { > > - struct drm_i915_private *dev_priv = inode->i_private; > > + struct drm_i915_private *i915 = inode->i_private; > > - if (HAS_GMCH(dev_priv)) > > + if (HAS_GMCH(i915)) > > return -ENODEV; > > - return single_open(file, spr_wm_latency_show, dev_priv); > > + return single_open(file, spr_wm_latency_show, i915); > > } > > static int cur_wm_latency_open(struct inode *inode, struct file *file) > > { > > - struct drm_i915_private *dev_priv = inode->i_private; > > + struct drm_i915_private *i915 = inode->i_private; > > - if (HAS_GMCH(dev_priv)) > > + if (HAS_GMCH(i915)) > > return -ENODEV; > > - return single_open(file, cur_wm_latency_show, dev_priv); > > + return single_open(file, cur_wm_latency_show, i915); > > } > > static ssize_t wm_latency_write(struct file *file, const char __user *ubuf, > > size_t len, loff_t *offp, u16 wm[8]) > > { > > struct seq_file *m = file->private_data; > > - struct drm_i915_private *dev_priv = m->private; > > - struct drm_device *dev = &dev_priv->drm; > > + struct drm_i915_private *i915 = m->private; > > + struct drm_device *dev = &i915->drm; > > u16 new[8] = { 0 }; > > int num_levels; > > int level; > > int ret; > > char tmp[32]; > > - if (IS_CHERRYVIEW(dev_priv)) > > + if (IS_CHERRYVIEW(i915)) > > num_levels = 3; > > - else if (IS_VALLEYVIEW(dev_priv)) > > + else if (IS_VALLEYVIEW(i915)) > > num_levels = 1; > > - else if (IS_G4X(dev_priv)) > > + else if (IS_G4X(i915)) > > num_levels = 3; > > else > > - num_levels = ilk_wm_max_level(dev_priv) + 1; > > + num_levels = ilk_wm_max_level(i915) + 1; > > if (len >= sizeof(tmp)) > > return -EINVAL; > > @@ -3546,13 +3546,13 @@ static ssize_t pri_wm_latency_write(struct file *file, const char __user *ubuf, > > size_t len, loff_t *offp) > > { > > struct seq_file *m = file->private_data; > > - struct drm_i915_private *dev_priv = m->private; > > + struct drm_i915_private *i915 = m->private; > > u16 *latencies; > > - if (INTEL_GEN(dev_priv) >= 9) > > - latencies = dev_priv->wm.skl_latency; > > + if (INTEL_GEN(i915) >= 9) > > + latencies = i915->wm.skl_latency; > > else > > - latencies = dev_priv->wm.pri_latency; > > + latencies = i915->wm.pri_latency; > > return wm_latency_write(file, ubuf, len, offp, latencies); > > } > > @@ -3561,13 +3561,13 @@ static ssize_t spr_wm_latency_write(struct file *file, const char __user *ubuf, > > size_t len, loff_t *offp) > > { > > struct seq_file *m = file->private_data; > > - struct drm_i915_private *dev_priv = m->private; > > + struct drm_i915_private *i915 = m->private; > > u16 *latencies; > > - if (INTEL_GEN(dev_priv) >= 9) > > - latencies = dev_priv->wm.skl_latency; > > + if (INTEL_GEN(i915) >= 9) > > + latencies = i915->wm.skl_latency; > > else > > - latencies = dev_priv->wm.spr_latency; > > + latencies = i915->wm.spr_latency; > > return wm_latency_write(file, ubuf, len, offp, latencies); > > } > > @@ -3576,13 +3576,13 @@ static ssize_t cur_wm_latency_write(struct file *file, const char __user *ubuf, > > size_t len, loff_t *offp) > > { > > struct seq_file *m = file->private_data; > > - struct drm_i915_private *dev_priv = m->private; > > + struct drm_i915_private *i915 = m->private; > > u16 *latencies; > > - if (INTEL_GEN(dev_priv) >= 9) > > - latencies = dev_priv->wm.skl_latency; > > + if (INTEL_GEN(i915) >= 9) > > + latencies = i915->wm.skl_latency; > > else > > - latencies = dev_priv->wm.cur_latency; > > + latencies = i915->wm.cur_latency; > > return wm_latency_write(file, ubuf, len, offp, latencies); > > } > > @@ -3752,14 +3752,14 @@ DEFINE_SIMPLE_ATTRIBUTE(i915_drop_caches_fops, > > static int > > i915_cache_sharing_get(void *data, u64 *val) > > { > > - struct drm_i915_private *dev_priv = data; > > + struct drm_i915_private *i915 = data; > > intel_wakeref_t wakeref; > > u32 snpcr = 0; > > - if (!(IS_GEN_RANGE(dev_priv, 6, 7))) > > + if (!(IS_GEN_RANGE(i915, 6, 7))) > > return -ENODEV; > > - with_intel_runtime_pm(dev_priv, wakeref) > > + with_intel_runtime_pm(i915, wakeref) > > snpcr = I915_READ(GEN6_MBCUNIT_SNPCR); > > *val = (snpcr & GEN6_MBC_SNPCR_MASK) >> GEN6_MBC_SNPCR_SHIFT; > > @@ -3770,17 +3770,17 @@ i915_cache_sharing_get(void *data, u64 *val) > > static int > > i915_cache_sharing_set(void *data, u64 val) > > { > > - struct drm_i915_private *dev_priv = data; > > + struct drm_i915_private *i915 = data; > > intel_wakeref_t wakeref; > > - if (!(IS_GEN_RANGE(dev_priv, 6, 7))) > > + if (!(IS_GEN_RANGE(i915, 6, 7))) > > return -ENODEV; > > if (val > 3) > > return -EINVAL; > > DRM_DEBUG_DRIVER("Manually setting uncore sharing to %llu\n", val); > > - with_intel_runtime_pm(dev_priv, wakeref) { > > + with_intel_runtime_pm(i915, wakeref) { > > u32 snpcr; > > /* Update the cache sharing policy here as well */ > > @@ -3797,7 +3797,7 @@ DEFINE_SIMPLE_ATTRIBUTE(i915_cache_sharing_fops, > > i915_cache_sharing_get, i915_cache_sharing_set, > > "%llu\n"); > > -static void cherryview_sseu_device_status(struct drm_i915_private *dev_priv, > > +static void cherryview_sseu_device_status(struct drm_i915_private *i915, > > struct sseu_dev_info *sseu) > > { > > #define SS_MAX 2 > > @@ -3830,11 +3830,11 @@ static void cherryview_sseu_device_status(struct drm_i915_private *dev_priv, > > #undef SS_MAX > > } > > -static void gen10_sseu_device_status(struct drm_i915_private *dev_priv, > > +static void gen10_sseu_device_status(struct drm_i915_private *i915, > > struct sseu_dev_info *sseu) > > { > > #define SS_MAX 6 > > - const struct intel_runtime_info *info = RUNTIME_INFO(dev_priv); > > + const struct intel_runtime_info *info = RUNTIME_INFO(i915); > > u32 s_reg[SS_MAX], eu_reg[2 * SS_MAX], eu_mask[2]; > > int s, ss; > > @@ -3886,11 +3886,11 @@ static void gen10_sseu_device_status(struct drm_i915_private *dev_priv, > > #undef SS_MAX > > } > > -static void gen9_sseu_device_status(struct drm_i915_private *dev_priv, > > +static void gen9_sseu_device_status(struct drm_i915_private *i915, > > struct sseu_dev_info *sseu) > > { > > #define SS_MAX 3 > > - const struct intel_runtime_info *info = RUNTIME_INFO(dev_priv); > > + const struct intel_runtime_info *info = RUNTIME_INFO(i915); > > u32 s_reg[SS_MAX], eu_reg[2 * SS_MAX], eu_mask[2]; > > int s, ss; > > @@ -3916,14 +3916,14 @@ static void gen9_sseu_device_status(struct drm_i915_private *dev_priv, > > sseu->slice_mask |= BIT(s); > > - if (IS_GEN9_BC(dev_priv)) > > + if (IS_GEN9_BC(i915)) > > sseu->subslice_mask[s] = > > - RUNTIME_INFO(dev_priv)->sseu.subslice_mask[s]; > > + RUNTIME_INFO(i915)->sseu.subslice_mask[s]; > > for (ss = 0; ss < info->sseu.max_subslices; ss++) { > > unsigned int eu_cnt; > > - if (IS_GEN9_LP(dev_priv)) { > > + if (IS_GEN9_LP(i915)) { > > if (!(s_reg[s] & (GEN9_PGCTL_SS_ACK(ss)))) > > /* skip disabled subslice */ > > continue; > > @@ -3942,7 +3942,7 @@ static void gen9_sseu_device_status(struct drm_i915_private *dev_priv, > > #undef SS_MAX > > } > > -static void broadwell_sseu_device_status(struct drm_i915_private *dev_priv, > > +static void broadwell_sseu_device_status(struct drm_i915_private *i915, > > struct sseu_dev_info *sseu) > > { > > u32 slice_info = I915_READ(GEN8_GT_SLICE_INFO); > > @@ -3952,10 +3952,10 @@ static void broadwell_sseu_device_status(struct drm_i915_private *dev_priv, > > if (sseu->slice_mask) { > > sseu->eu_per_subslice = > > - RUNTIME_INFO(dev_priv)->sseu.eu_per_subslice; > > + RUNTIME_INFO(i915)->sseu.eu_per_subslice; > > for (s = 0; s < fls(sseu->slice_mask); s++) { > > sseu->subslice_mask[s] = > > - RUNTIME_INFO(dev_priv)->sseu.subslice_mask[s]; > > + RUNTIME_INFO(i915)->sseu.subslice_mask[s]; > > } > > sseu->eu_total = sseu->eu_per_subslice * > > intel_sseu_subslice_total(sseu); > > @@ -3963,7 +3963,7 @@ static void broadwell_sseu_device_status(struct drm_i915_private *dev_priv, > > /* subtract fused off EU(s) from enabled slice(s) */ > > for (s = 0; s < fls(sseu->slice_mask); s++) { > > u8 subslice_7eu = > > - RUNTIME_INFO(dev_priv)->sseu.subslice_7eu[s]; > > + RUNTIME_INFO(i915)->sseu.subslice_7eu[s]; > > sseu->eu_total -= hweight8(subslice_7eu); > > } > > @@ -3973,7 +3973,7 @@ static void broadwell_sseu_device_status(struct drm_i915_private *dev_priv, > > static void i915_print_sseu_info(struct seq_file *m, bool is_available_info, > > const struct sseu_dev_info *sseu) > > { > > - struct drm_i915_private *dev_priv = node_to_i915(m->private); > > + struct drm_i915_private *i915 = node_to_i915(m->private); > > const char *type = is_available_info ? "Available" : "Enabled"; > > int s; > > @@ -3995,8 +3995,8 @@ static void i915_print_sseu_info(struct seq_file *m, bool is_available_info, > > if (!is_available_info) > > return; > > - seq_printf(m, " Has Pooled EU: %s\n", yesno(HAS_POOLED_EU(dev_priv))); > > - if (HAS_POOLED_EU(dev_priv)) > > + seq_printf(m, " Has Pooled EU: %s\n", yesno(HAS_POOLED_EU(i915))); > > + if (HAS_POOLED_EU(i915)) > > seq_printf(m, " Min EU in pool: %u\n", sseu->min_eu_in_pool); > > seq_printf(m, " Has Slice Power Gating: %s\n", > > @@ -4009,32 +4009,32 @@ static void i915_print_sseu_info(struct seq_file *m, bool is_available_info, > > static int i915_sseu_status(struct seq_file *m, void *unused) > > { > > - struct drm_i915_private *dev_priv = node_to_i915(m->private); > > + struct drm_i915_private *i915 = node_to_i915(m->private); > > struct sseu_dev_info sseu; > > intel_wakeref_t wakeref; > > - if (INTEL_GEN(dev_priv) < 8) > > + if (INTEL_GEN(i915) < 8) > > return -ENODEV; > > seq_puts(m, "SSEU Device Info\n"); > > - i915_print_sseu_info(m, true, &RUNTIME_INFO(dev_priv)->sseu); > > + i915_print_sseu_info(m, true, &RUNTIME_INFO(i915)->sseu); > > seq_puts(m, "SSEU Device Status\n"); > > memset(&sseu, 0, sizeof(sseu)); > > - sseu.max_slices = RUNTIME_INFO(dev_priv)->sseu.max_slices; > > - sseu.max_subslices = RUNTIME_INFO(dev_priv)->sseu.max_subslices; > > + sseu.max_slices = RUNTIME_INFO(i915)->sseu.max_slices; > > + sseu.max_subslices = RUNTIME_INFO(i915)->sseu.max_subslices; > > sseu.max_eus_per_subslice = > > - RUNTIME_INFO(dev_priv)->sseu.max_eus_per_subslice; > > - > > - with_intel_runtime_pm(dev_priv, wakeref) { > > - if (IS_CHERRYVIEW(dev_priv)) > > - cherryview_sseu_device_status(dev_priv, &sseu); > > - else if (IS_BROADWELL(dev_priv)) > > - broadwell_sseu_device_status(dev_priv, &sseu); > > - else if (IS_GEN(dev_priv, 9)) > > - gen9_sseu_device_status(dev_priv, &sseu); > > - else if (INTEL_GEN(dev_priv) >= 10) > > - gen10_sseu_device_status(dev_priv, &sseu); > > + RUNTIME_INFO(i915)->sseu.max_eus_per_subslice; > > + > > + with_intel_runtime_pm(i915, wakeref) { > > + if (IS_CHERRYVIEW(i915)) > > + cherryview_sseu_device_status(i915, &sseu); > > + else if (IS_BROADWELL(i915)) > > + broadwell_sseu_device_status(i915, &sseu); > > + else if (IS_GEN(i915, 9)) > > + gen9_sseu_device_status(i915, &sseu); > > + else if (INTEL_GEN(i915) >= 10) > > + gen10_sseu_device_status(i915, &sseu); > > } > > i915_print_sseu_info(m, false, &sseu); > > @@ -4077,15 +4077,15 @@ static const struct file_operations i915_forcewake_fops = { > > static int i915_hpd_storm_ctl_show(struct seq_file *m, void *data) > > { > > - struct drm_i915_private *dev_priv = m->private; > > - struct i915_hotplug *hotplug = &dev_priv->hotplug; > > + struct drm_i915_private *i915 = m->private; > > + struct i915_hotplug *hotplug = &i915->hotplug; > > /* Synchronize with everything first in case there's been an HPD > > * storm, but we haven't finished handling it in the kernel yet > > */ > > - synchronize_irq(dev_priv->drm.irq); > > - flush_work(&dev_priv->hotplug.dig_port_work); > > - flush_work(&dev_priv->hotplug.hotplug_work); > > + synchronize_irq(i915->drm.irq); > > + flush_work(&i915->hotplug.dig_port_work); > > + flush_work(&i915->hotplug.hotplug_work); > > seq_printf(m, "Threshold: %d\n", hotplug->hpd_storm_threshold); > > seq_printf(m, "Detected: %s\n", > > @@ -4099,8 +4099,8 @@ static ssize_t i915_hpd_storm_ctl_write(struct file *file, > > loff_t *offp) > > { > > struct seq_file *m = file->private_data; > > - struct drm_i915_private *dev_priv = m->private; > > - struct i915_hotplug *hotplug = &dev_priv->hotplug; > > + struct drm_i915_private *i915 = m->private; > > + struct i915_hotplug *hotplug = &i915->hotplug; > > unsigned int new_threshold; > > int i; > > char *newline; > > @@ -4130,15 +4130,15 @@ static ssize_t i915_hpd_storm_ctl_write(struct file *file, > > else > > DRM_DEBUG_KMS("Disabling HPD storm detection\n"); > > - spin_lock_irq(&dev_priv->irq_lock); > > + spin_lock_irq(&i915->irq_lock); > > hotplug->hpd_storm_threshold = new_threshold; > > /* Reset the HPD storm stats so we don't accidentally trigger a storm */ > > for_each_hpd_pin(i) > > hotplug->stats[i].count = 0; > > - spin_unlock_irq(&dev_priv->irq_lock); > > + spin_unlock_irq(&i915->irq_lock); > > /* Re-enable hpd immediately if we were in an irq storm */ > > - flush_delayed_work(&dev_priv->hotplug.reenable_work); > > + flush_delayed_work(&i915->hotplug.reenable_work); > > return len; > > } > > @@ -4159,10 +4159,10 @@ static const struct file_operations i915_hpd_storm_ctl_fops = { > > static int i915_hpd_short_storm_ctl_show(struct seq_file *m, void *data) > > { > > - struct drm_i915_private *dev_priv = m->private; > > + struct drm_i915_private *i915 = m->private; > > seq_printf(m, "Enabled: %s\n", > > - yesno(dev_priv->hotplug.hpd_short_storm_enabled)); > > + yesno(i915->hotplug.hpd_short_storm_enabled)); > > return 0; > > } > > @@ -4179,8 +4179,8 @@ static ssize_t i915_hpd_short_storm_ctl_write(struct file *file, > > size_t len, loff_t *offp) > > { > > struct seq_file *m = file->private_data; > > - struct drm_i915_private *dev_priv = m->private; > > - struct i915_hotplug *hotplug = &dev_priv->hotplug; > > + struct drm_i915_private *i915 = m->private; > > + struct i915_hotplug *hotplug = &i915->hotplug; > > char *newline; > > char tmp[16]; > > int i; > > @@ -4201,22 +4201,22 @@ static ssize_t i915_hpd_short_storm_ctl_write(struct file *file, > > /* Reset to the "default" state for this system */ > > if (strcmp(tmp, "reset") == 0) > > - new_state = !HAS_DP_MST(dev_priv); > > + new_state = !HAS_DP_MST(i915); > > else if (kstrtobool(tmp, &new_state) != 0) > > return -EINVAL; > > DRM_DEBUG_KMS("%sabling HPD short storm detection\n", > > new_state ? "En" : "Dis"); > > - spin_lock_irq(&dev_priv->irq_lock); > > + spin_lock_irq(&i915->irq_lock); > > hotplug->hpd_short_storm_enabled = new_state; > > /* Reset the HPD storm stats so we don't accidentally trigger a storm */ > > for_each_hpd_pin(i) > > hotplug->stats[i].count = 0; > > - spin_unlock_irq(&dev_priv->irq_lock); > > + spin_unlock_irq(&i915->irq_lock); > > /* Re-enable hpd immediately if we were in an irq storm */ > > - flush_delayed_work(&dev_priv->hotplug.reenable_work); > > + flush_delayed_work(&i915->hotplug.reenable_work); > > return len; > > } > > @@ -4232,11 +4232,11 @@ static const struct file_operations i915_hpd_short_storm_ctl_fops = { > > static int i915_drrs_ctl_set(void *data, u64 val) > > { > > - struct drm_i915_private *dev_priv = data; > > - struct drm_device *dev = &dev_priv->drm; > > + struct drm_i915_private *i915 = data; > > + struct drm_device *dev = &i915->drm; > > struct intel_crtc *crtc; > > - if (INTEL_GEN(dev_priv) < 7) > > + if (INTEL_GEN(i915) < 7) > > return -ENODEV; > > for_each_intel_crtc(dev, crtc) { > > @@ -4305,9 +4305,9 @@ i915_fifo_underrun_reset_write(struct file *filp, > > const char __user *ubuf, > > size_t cnt, loff_t *ppos) > > { > > - struct drm_i915_private *dev_priv = filp->private_data; > > + struct drm_i915_private *i915 = filp->private_data; > > struct intel_crtc *intel_crtc; > > - struct drm_device *dev = &dev_priv->drm; > > + struct drm_device *dev = &i915->drm; > > int ret; > > bool reset; > > @@ -4347,7 +4347,7 @@ i915_fifo_underrun_reset_write(struct file *filp, > > return ret; > > } > > - ret = intel_fbc_reset_underrun(dev_priv); > > + ret = intel_fbc_reset_underrun(i915); > > if (ret) > > return ret; > > @@ -4437,9 +4437,9 @@ static const struct i915_debugfs_files { > > {"i915_edp_psr_debug", &i915_edp_psr_debug_fops} > > }; > > -int i915_debugfs_register(struct drm_i915_private *dev_priv) > > +int i915_debugfs_register(struct drm_i915_private *i915) > > { > > - struct drm_minor *minor = dev_priv->drm.primary; > > + struct drm_minor *minor = i915->drm.primary; > > struct dentry *ent; > > int i; > > @@ -4692,7 +4692,7 @@ static const struct file_operations i915_dsc_fec_support_fops = { > > int i915_debugfs_connector_add(struct drm_connector *connector) > > { > > struct dentry *root = connector->debugfs_entry; > > - struct drm_i915_private *dev_priv = to_i915(connector->dev); > > + struct drm_i915_private *i915 = to_i915(connector->dev); > > /* The connector must have been registered beforehands. */ > > if (!root) > > @@ -4717,7 +4717,7 @@ int i915_debugfs_connector_add(struct drm_connector *connector) > > connector, &i915_hdcp_sink_capability_fops); > > } > > - if (INTEL_GEN(dev_priv) >= 10 && > > + if (INTEL_GEN(i915) >= 10 && > > (connector->connector_type == DRM_MODE_CONNECTOR_DisplayPort || > > connector->connector_type == DRM_MODE_CONNECTOR_eDP)) > > debugfs_create_file("i915_dsc_fec_support", S_IRUGO, root, > > diff --git a/drivers/gpu/drm/i915/i915_debugfs.h b/drivers/gpu/drm/i915/i915_debugfs.h > > index c0cd22eb916d..3e99eaf8de02 100644 > > --- a/drivers/gpu/drm/i915/i915_debugfs.h > > +++ b/drivers/gpu/drm/i915/i915_debugfs.h > > @@ -10,10 +10,10 @@ struct drm_i915_private; > > struct drm_connector; > > #ifdef CONFIG_DEBUG_FS > > -int i915_debugfs_register(struct drm_i915_private *dev_priv); > > +int i915_debugfs_register(struct drm_i915_private *i915); > > int i915_debugfs_connector_add(struct drm_connector *connector); > > #else > > -static inline int i915_debugfs_register(struct drm_i915_private *dev_priv) { return 0; } > > +static inline int i915_debugfs_register(struct drm_i915_private *i915) { return 0; } > > static inline int i915_debugfs_connector_add(struct drm_connector *connector) { return 0; } > > #endif > > diff --git a/drivers/gpu/drm/i915/i915_drv.c b/drivers/gpu/drm/i915/i915_drv.c > > index 81ff2c78fd55..a09d028ec8d8 100644 > > --- a/drivers/gpu/drm/i915/i915_drv.c > > +++ b/drivers/gpu/drm/i915/i915_drv.c > > @@ -108,11 +108,11 @@ bool i915_error_injected(void) > > "providing the dmesg log by booting with drm.debug=0xf" > > void > > -__i915_printk(struct drm_i915_private *dev_priv, const char *level, > > +__i915_printk(struct drm_i915_private *i915, const char *level, > > const char *fmt, ...) > > { > > static bool shown_bug_once; > > - struct device *kdev = dev_priv->drm.dev; > > + struct device *kdev = i915->drm.dev; > > bool is_error = level[1] <= KERN_ERR[1]; > > bool is_debug = level[1] == KERN_DEBUG[1]; > > struct va_format vaf; > > @@ -148,74 +148,74 @@ __i915_printk(struct drm_i915_private *dev_priv, const char *level, > > /* Map PCH device id to PCH type, or PCH_NONE if unknown. */ > > static enum intel_pch > > -intel_pch_type(const struct drm_i915_private *dev_priv, unsigned short id) > > +intel_pch_type(const struct drm_i915_private *i915, unsigned short id) > > { > > switch (id) { > > case INTEL_PCH_IBX_DEVICE_ID_TYPE: > > DRM_DEBUG_KMS("Found Ibex Peak PCH\n"); > > - WARN_ON(!IS_GEN(dev_priv, 5)); > > + WARN_ON(!IS_GEN(i915, 5)); > > return PCH_IBX; > > case INTEL_PCH_CPT_DEVICE_ID_TYPE: > > DRM_DEBUG_KMS("Found CougarPoint PCH\n"); > > - WARN_ON(!IS_GEN(dev_priv, 6) && !IS_IVYBRIDGE(dev_priv)); > > + WARN_ON(!IS_GEN(i915, 6) && !IS_IVYBRIDGE(i915)); > > return PCH_CPT; > > case INTEL_PCH_PPT_DEVICE_ID_TYPE: > > DRM_DEBUG_KMS("Found PantherPoint PCH\n"); > > - WARN_ON(!IS_GEN(dev_priv, 6) && !IS_IVYBRIDGE(dev_priv)); > > + WARN_ON(!IS_GEN(i915, 6) && !IS_IVYBRIDGE(i915)); > > /* PantherPoint is CPT compatible */ > > return PCH_CPT; > > case INTEL_PCH_LPT_DEVICE_ID_TYPE: > > DRM_DEBUG_KMS("Found LynxPoint PCH\n"); > > - WARN_ON(!IS_HASWELL(dev_priv) && !IS_BROADWELL(dev_priv)); > > - WARN_ON(IS_HSW_ULT(dev_priv) || IS_BDW_ULT(dev_priv)); > > + WARN_ON(!IS_HASWELL(i915) && !IS_BROADWELL(i915)); > > + WARN_ON(IS_HSW_ULT(i915) || IS_BDW_ULT(i915)); > > return PCH_LPT; > > case INTEL_PCH_LPT_LP_DEVICE_ID_TYPE: > > DRM_DEBUG_KMS("Found LynxPoint LP PCH\n"); > > - WARN_ON(!IS_HASWELL(dev_priv) && !IS_BROADWELL(dev_priv)); > > - WARN_ON(!IS_HSW_ULT(dev_priv) && !IS_BDW_ULT(dev_priv)); > > + WARN_ON(!IS_HASWELL(i915) && !IS_BROADWELL(i915)); > > + WARN_ON(!IS_HSW_ULT(i915) && !IS_BDW_ULT(i915)); > > return PCH_LPT; > > case INTEL_PCH_WPT_DEVICE_ID_TYPE: > > DRM_DEBUG_KMS("Found WildcatPoint PCH\n"); > > - WARN_ON(!IS_HASWELL(dev_priv) && !IS_BROADWELL(dev_priv)); > > - WARN_ON(IS_HSW_ULT(dev_priv) || IS_BDW_ULT(dev_priv)); > > + WARN_ON(!IS_HASWELL(i915) && !IS_BROADWELL(i915)); > > + WARN_ON(IS_HSW_ULT(i915) || IS_BDW_ULT(i915)); > > /* WildcatPoint is LPT compatible */ > > return PCH_LPT; > > case INTEL_PCH_WPT_LP_DEVICE_ID_TYPE: > > DRM_DEBUG_KMS("Found WildcatPoint LP PCH\n"); > > - WARN_ON(!IS_HASWELL(dev_priv) && !IS_BROADWELL(dev_priv)); > > - WARN_ON(!IS_HSW_ULT(dev_priv) && !IS_BDW_ULT(dev_priv)); > > + WARN_ON(!IS_HASWELL(i915) && !IS_BROADWELL(i915)); > > + WARN_ON(!IS_HSW_ULT(i915) && !IS_BDW_ULT(i915)); > > /* WildcatPoint is LPT compatible */ > > return PCH_LPT; > > case INTEL_PCH_SPT_DEVICE_ID_TYPE: > > DRM_DEBUG_KMS("Found SunrisePoint PCH\n"); > > - WARN_ON(!IS_SKYLAKE(dev_priv) && !IS_KABYLAKE(dev_priv)); > > + WARN_ON(!IS_SKYLAKE(i915) && !IS_KABYLAKE(i915)); > > return PCH_SPT; > > case INTEL_PCH_SPT_LP_DEVICE_ID_TYPE: > > DRM_DEBUG_KMS("Found SunrisePoint LP PCH\n"); > > - WARN_ON(!IS_SKYLAKE(dev_priv) && !IS_KABYLAKE(dev_priv)); > > + WARN_ON(!IS_SKYLAKE(i915) && !IS_KABYLAKE(i915)); > > return PCH_SPT; > > case INTEL_PCH_KBP_DEVICE_ID_TYPE: > > DRM_DEBUG_KMS("Found Kaby Lake PCH (KBP)\n"); > > - WARN_ON(!IS_SKYLAKE(dev_priv) && !IS_KABYLAKE(dev_priv) && > > - !IS_COFFEELAKE(dev_priv)); > > + WARN_ON(!IS_SKYLAKE(i915) && !IS_KABYLAKE(i915) && > > + !IS_COFFEELAKE(i915)); > > /* KBP is SPT compatible */ > > return PCH_SPT; > > case INTEL_PCH_CNP_DEVICE_ID_TYPE: > > DRM_DEBUG_KMS("Found Cannon Lake PCH (CNP)\n"); > > - WARN_ON(!IS_CANNONLAKE(dev_priv) && !IS_COFFEELAKE(dev_priv)); > > + WARN_ON(!IS_CANNONLAKE(i915) && !IS_COFFEELAKE(i915)); > > return PCH_CNP; > > case INTEL_PCH_CNP_LP_DEVICE_ID_TYPE: > > DRM_DEBUG_KMS("Found Cannon Lake LP PCH (CNP-LP)\n"); > > - WARN_ON(!IS_CANNONLAKE(dev_priv) && !IS_COFFEELAKE(dev_priv)); > > + WARN_ON(!IS_CANNONLAKE(i915) && !IS_COFFEELAKE(i915)); > > return PCH_CNP; > > case INTEL_PCH_CMP_DEVICE_ID_TYPE: > > DRM_DEBUG_KMS("Found Comet Lake PCH (CMP)\n"); > > - WARN_ON(!IS_COFFEELAKE(dev_priv)); > > + WARN_ON(!IS_COFFEELAKE(i915)); > > /* CometPoint is CNP Compatible */ > > return PCH_CNP; > > case INTEL_PCH_ICP_DEVICE_ID_TYPE: > > DRM_DEBUG_KMS("Found Ice Lake PCH\n"); > > - WARN_ON(!IS_ICELAKE(dev_priv)); > > + WARN_ON(!IS_ICELAKE(i915)); > > return PCH_ICP; > > default: > > return PCH_NONE; > > @@ -233,7 +233,7 @@ static bool intel_is_virt_pch(unsigned short id, > > } > > static unsigned short > > -intel_virt_detect_pch(const struct drm_i915_private *dev_priv) > > +intel_virt_detect_pch(const struct drm_i915_private *i915) > > { > > unsigned short id = 0; > > @@ -244,19 +244,19 @@ intel_virt_detect_pch(const struct drm_i915_private *dev_priv) > > * make an educated guess as to which PCH is really there. > > */ > > - if (IS_ICELAKE(dev_priv)) > > + if (IS_ICELAKE(i915)) > > id = INTEL_PCH_ICP_DEVICE_ID_TYPE; > > - else if (IS_CANNONLAKE(dev_priv) || IS_COFFEELAKE(dev_priv)) > > + else if (IS_CANNONLAKE(i915) || IS_COFFEELAKE(i915)) > > id = INTEL_PCH_CNP_DEVICE_ID_TYPE; > > - else if (IS_KABYLAKE(dev_priv) || IS_SKYLAKE(dev_priv)) > > + else if (IS_KABYLAKE(i915) || IS_SKYLAKE(i915)) > > id = INTEL_PCH_SPT_DEVICE_ID_TYPE; > > - else if (IS_HSW_ULT(dev_priv) || IS_BDW_ULT(dev_priv)) > > + else if (IS_HSW_ULT(i915) || IS_BDW_ULT(i915)) > > id = INTEL_PCH_LPT_LP_DEVICE_ID_TYPE; > > - else if (IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv)) > > + else if (IS_HASWELL(i915) || IS_BROADWELL(i915)) > > id = INTEL_PCH_LPT_DEVICE_ID_TYPE; > > - else if (IS_GEN(dev_priv, 6) || IS_IVYBRIDGE(dev_priv)) > > + else if (IS_GEN(i915, 6) || IS_IVYBRIDGE(i915)) > > id = INTEL_PCH_CPT_DEVICE_ID_TYPE; > > - else if (IS_GEN(dev_priv, 5)) > > + else if (IS_GEN(i915, 5)) > > id = INTEL_PCH_IBX_DEVICE_ID_TYPE; > > if (id) > > @@ -267,7 +267,7 @@ intel_virt_detect_pch(const struct drm_i915_private *dev_priv) > > return id; > > } > > -static void intel_detect_pch(struct drm_i915_private *dev_priv) > > +static void intel_detect_pch(struct drm_i915_private *i915) > > { > > struct pci_dev *pch = NULL; > > @@ -291,22 +291,22 @@ static void intel_detect_pch(struct drm_i915_private *dev_priv) > > id = pch->device & INTEL_PCH_DEVICE_ID_MASK; > > - pch_type = intel_pch_type(dev_priv, id); > > + pch_type = intel_pch_type(i915, id); > > if (pch_type != PCH_NONE) { > > - dev_priv->pch_type = pch_type; > > - dev_priv->pch_id = id; > > + i915->pch_type = pch_type; > > + i915->pch_id = id; > > break; > > } else if (intel_is_virt_pch(id, pch->subsystem_vendor, > > pch->subsystem_device)) { > > - id = intel_virt_detect_pch(dev_priv); > > - pch_type = intel_pch_type(dev_priv, id); > > + id = intel_virt_detect_pch(i915); > > + pch_type = intel_pch_type(i915, id); > > /* Sanity check virtual PCH id */ > > if (WARN_ON(id && pch_type == PCH_NONE)) > > id = 0; > > - dev_priv->pch_type = pch_type; > > - dev_priv->pch_id = id; > > + i915->pch_type = pch_type; > > + i915->pch_id = id; > > break; > > } > > } > > @@ -315,10 +315,10 @@ static void intel_detect_pch(struct drm_i915_private *dev_priv) > > * Use PCH_NOP (PCH but no South Display) for PCH platforms without > > * display. > > */ > > - if (pch && !HAS_DISPLAY(dev_priv)) { > > + if (pch && !HAS_DISPLAY(i915)) { > > DRM_DEBUG_KMS("Display disabled, reverting to NOP PCH\n"); > > - dev_priv->pch_type = PCH_NOP; > > - dev_priv->pch_id = 0; > > + i915->pch_type = PCH_NOP; > > + i915->pch_id = 0; > > } > > if (!pch) > > @@ -330,9 +330,9 @@ static void intel_detect_pch(struct drm_i915_private *dev_priv) > > static int i915_getparam_ioctl(struct drm_device *dev, void *data, > > struct drm_file *file_priv) > > { > > - struct drm_i915_private *dev_priv = to_i915(dev); > > - struct pci_dev *pdev = dev_priv->drm.pdev; > > - const struct sseu_dev_info *sseu = &RUNTIME_INFO(dev_priv)->sseu; > > + struct drm_i915_private *i915 = to_i915(dev); > > + struct pci_dev *pdev = i915->drm.pdev; > > + const struct sseu_dev_info *sseu = &RUNTIME_INFO(i915)->sseu; > > drm_i915_getparam_t *param = data; > > int value; > > @@ -350,40 +350,40 @@ static int i915_getparam_ioctl(struct drm_device *dev, void *data, > > value = pdev->revision; > > break; > > case I915_PARAM_NUM_FENCES_AVAIL: > > - value = dev_priv->num_fence_regs; > > + value = i915->num_fence_regs; > > break; > > case I915_PARAM_HAS_OVERLAY: > > - value = dev_priv->overlay ? 1 : 0; > > + value = i915->overlay ? 1 : 0; > > break; > > case I915_PARAM_HAS_BSD: > > - value = !!dev_priv->engine[VCS0]; > > + value = !!i915->engine[VCS0]; > > break; > > case I915_PARAM_HAS_BLT: > > - value = !!dev_priv->engine[BCS0]; > > + value = !!i915->engine[BCS0]; > > break; > > case I915_PARAM_HAS_VEBOX: > > - value = !!dev_priv->engine[VECS0]; > > + value = !!i915->engine[VECS0]; > > break; > > case I915_PARAM_HAS_BSD2: > > - value = !!dev_priv->engine[VCS1]; > > + value = !!i915->engine[VCS1]; > > break; > > case I915_PARAM_HAS_LLC: > > - value = HAS_LLC(dev_priv); > > + value = HAS_LLC(i915); > > break; > > case I915_PARAM_HAS_WT: > > - value = HAS_WT(dev_priv); > > + value = HAS_WT(i915); > > break; > > case I915_PARAM_HAS_ALIASING_PPGTT: > > - value = INTEL_PPGTT(dev_priv); > > + value = INTEL_PPGTT(i915); > > break; > > case I915_PARAM_HAS_SEMAPHORES: > > - value = !!(dev_priv->caps.scheduler & I915_SCHEDULER_CAP_SEMAPHORES); > > + value = !!(i915->caps.scheduler & I915_SCHEDULER_CAP_SEMAPHORES); > > break; > > case I915_PARAM_HAS_SECURE_BATCHES: > > value = capable(CAP_SYS_ADMIN); > > break; > > case I915_PARAM_CMD_PARSER_VERSION: > > - value = i915_cmd_parser_get_version(dev_priv); > > + value = i915_cmd_parser_get_version(i915); > > break; > > case I915_PARAM_SUBSLICE_TOTAL: > > value = intel_sseu_subslice_total(sseu); > > @@ -397,21 +397,21 @@ static int i915_getparam_ioctl(struct drm_device *dev, void *data, > > break; > > case I915_PARAM_HAS_GPU_RESET: > > value = i915_modparams.enable_hangcheck && > > - intel_has_gpu_reset(dev_priv); > > - if (value && intel_has_reset_engine(dev_priv)) > > + intel_has_gpu_reset(i915); > > + if (value && intel_has_reset_engine(i915)) > > value = 2; > > break; > > case I915_PARAM_HAS_RESOURCE_STREAMER: > > value = 0; > > break; > > case I915_PARAM_HAS_POOLED_EU: > > - value = HAS_POOLED_EU(dev_priv); > > + value = HAS_POOLED_EU(i915); > > break; > > case I915_PARAM_MIN_EU_IN_POOL: > > value = sseu->min_eu_in_pool; > > break; > > case I915_PARAM_HUC_STATUS: > > - value = intel_huc_check_status(&dev_priv->huc); > > + value = intel_huc_check_status(&i915->huc); > > if (value < 0) > > return value; > > break; > > @@ -423,7 +423,7 @@ static int i915_getparam_ioctl(struct drm_device *dev, void *data, > > value = i915_gem_mmap_gtt_version(); > > break; > > case I915_PARAM_HAS_SCHEDULER: > > - value = dev_priv->caps.scheduler; > > + value = i915->caps.scheduler; > > break; > > case I915_PARAM_MMAP_VERSION: > > @@ -456,7 +456,7 @@ static int i915_getparam_ioctl(struct drm_device *dev, void *data, > > value = 1; > > break; > > case I915_PARAM_HAS_CONTEXT_ISOLATION: > > - value = intel_engines_has_context_isolation(dev_priv); > > + value = intel_engines_has_context_isolation(i915); > > break; > > case I915_PARAM_SLICE_MASK: > > value = sseu->slice_mask; > > @@ -469,10 +469,10 @@ static int i915_getparam_ioctl(struct drm_device *dev, void *data, > > return -ENODEV; > > break; > > case I915_PARAM_CS_TIMESTAMP_FREQUENCY: > > - value = 1000 * RUNTIME_INFO(dev_priv)->cs_timestamp_frequency_khz; > > + value = 1000 * RUNTIME_INFO(i915)->cs_timestamp_frequency_khz; > > break; > > case I915_PARAM_MMAP_GTT_COHERENT: > > - value = INTEL_INFO(dev_priv)->has_coherent_ggtt; > > + value = INTEL_INFO(i915)->has_coherent_ggtt; > > break; > > default: > > DRM_DEBUG("Unknown parameter %d\n", param->param); > > @@ -485,13 +485,13 @@ static int i915_getparam_ioctl(struct drm_device *dev, void *data, > > return 0; > > } > > -static int i915_get_bridge_dev(struct drm_i915_private *dev_priv) > > +static int i915_get_bridge_dev(struct drm_i915_private *i915) > > { > > - int domain = pci_domain_nr(dev_priv->drm.pdev->bus); > > + int domain = pci_domain_nr(i915->drm.pdev->bus); > > - dev_priv->bridge_dev = > > + i915->bridge_dev = > > pci_get_domain_bus_and_slot(domain, 0, PCI_DEVFN(0, 0)); > > - if (!dev_priv->bridge_dev) { > > + if (!i915->bridge_dev) { > > DRM_ERROR("bridge device not found\n"); > > return -1; > > } > > @@ -500,16 +500,16 @@ static int i915_get_bridge_dev(struct drm_i915_private *dev_priv) > > /* Allocate space for the MCH regs if needed, return nonzero on error */ > > static int > > -intel_alloc_mchbar_resource(struct drm_i915_private *dev_priv) > > +intel_alloc_mchbar_resource(struct drm_i915_private *i915) > > { > > - int reg = INTEL_GEN(dev_priv) >= 4 ? MCHBAR_I965 : MCHBAR_I915; > > + int reg = INTEL_GEN(i915) >= 4 ? MCHBAR_I965 : MCHBAR_I915; > > u32 temp_lo, temp_hi = 0; > > u64 mchbar_addr; > > int ret; > > - if (INTEL_GEN(dev_priv) >= 4) > > - pci_read_config_dword(dev_priv->bridge_dev, reg + 4, &temp_hi); > > - pci_read_config_dword(dev_priv->bridge_dev, reg, &temp_lo); > > + if (INTEL_GEN(i915) >= 4) > > + pci_read_config_dword(i915->bridge_dev, reg + 4, &temp_hi); > > + pci_read_config_dword(i915->bridge_dev, reg, &temp_lo); > > mchbar_addr = ((u64)temp_hi << 32) | temp_lo; > > /* If ACPI doesn't have it, assume we need to allocate it ourselves */ > > @@ -520,47 +520,47 @@ intel_alloc_mchbar_resource(struct drm_i915_private *dev_priv) > > #endif > > /* Get some space for it */ > > - dev_priv->mch_res.name = "i915 MCHBAR"; > > - dev_priv->mch_res.flags = IORESOURCE_MEM; > > - ret = pci_bus_alloc_resource(dev_priv->bridge_dev->bus, > > - &dev_priv->mch_res, > > + i915->mch_res.name = "i915 MCHBAR"; > > + i915->mch_res.flags = IORESOURCE_MEM; > > + ret = pci_bus_alloc_resource(i915->bridge_dev->bus, > > + &i915->mch_res, > > MCHBAR_SIZE, MCHBAR_SIZE, > > PCIBIOS_MIN_MEM, > > 0, pcibios_align_resource, > > - dev_priv->bridge_dev); > > + i915->bridge_dev); > > if (ret) { > > DRM_DEBUG_DRIVER("failed bus alloc: %d\n", ret); > > - dev_priv->mch_res.start = 0; > > + i915->mch_res.start = 0; > > return ret; > > } > > - if (INTEL_GEN(dev_priv) >= 4) > > - pci_write_config_dword(dev_priv->bridge_dev, reg + 4, > > - upper_32_bits(dev_priv->mch_res.start)); > > + if (INTEL_GEN(i915) >= 4) > > + pci_write_config_dword(i915->bridge_dev, reg + 4, > > + upper_32_bits(i915->mch_res.start)); > > - pci_write_config_dword(dev_priv->bridge_dev, reg, > > - lower_32_bits(dev_priv->mch_res.start)); > > + pci_write_config_dword(i915->bridge_dev, reg, > > + lower_32_bits(i915->mch_res.start)); > > return 0; > > } > > /* Setup MCHBAR if possible, return true if we should disable it again */ > > static void > > -intel_setup_mchbar(struct drm_i915_private *dev_priv) > > +intel_setup_mchbar(struct drm_i915_private *i915) > > { > > - int mchbar_reg = INTEL_GEN(dev_priv) >= 4 ? MCHBAR_I965 : MCHBAR_I915; > > + int mchbar_reg = INTEL_GEN(i915) >= 4 ? MCHBAR_I965 : MCHBAR_I915; > > u32 temp; > > bool enabled; > > - if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) > > + if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) > > return; > > - dev_priv->mchbar_need_disable = false; > > + i915->mchbar_need_disable = false; > > - if (IS_I915G(dev_priv) || IS_I915GM(dev_priv)) { > > - pci_read_config_dword(dev_priv->bridge_dev, DEVEN, &temp); > > + if (IS_I915G(i915) || IS_I915GM(i915)) { > > + pci_read_config_dword(i915->bridge_dev, DEVEN, &temp); > > enabled = !!(temp & DEVEN_MCHBAR_EN); > > } else { > > - pci_read_config_dword(dev_priv->bridge_dev, mchbar_reg, &temp); > > + pci_read_config_dword(i915->bridge_dev, mchbar_reg, &temp); > > enabled = temp & 1; > > } > > @@ -568,56 +568,56 @@ intel_setup_mchbar(struct drm_i915_private *dev_priv) > > if (enabled) > > return; > > - if (intel_alloc_mchbar_resource(dev_priv)) > > + if (intel_alloc_mchbar_resource(i915)) > > return; > > - dev_priv->mchbar_need_disable = true; > > + i915->mchbar_need_disable = true; > > /* Space is allocated or reserved, so enable it. */ > > - if (IS_I915G(dev_priv) || IS_I915GM(dev_priv)) { > > - pci_write_config_dword(dev_priv->bridge_dev, DEVEN, > > + if (IS_I915G(i915) || IS_I915GM(i915)) { > > + pci_write_config_dword(i915->bridge_dev, DEVEN, > > temp | DEVEN_MCHBAR_EN); > > } else { > > - pci_read_config_dword(dev_priv->bridge_dev, mchbar_reg, &temp); > > - pci_write_config_dword(dev_priv->bridge_dev, mchbar_reg, temp | 1); > > + pci_read_config_dword(i915->bridge_dev, mchbar_reg, &temp); > > + pci_write_config_dword(i915->bridge_dev, mchbar_reg, temp | 1); > > } > > } > > static void > > -intel_teardown_mchbar(struct drm_i915_private *dev_priv) > > +intel_teardown_mchbar(struct drm_i915_private *i915) > > { > > - int mchbar_reg = INTEL_GEN(dev_priv) >= 4 ? MCHBAR_I965 : MCHBAR_I915; > > + int mchbar_reg = INTEL_GEN(i915) >= 4 ? MCHBAR_I965 : MCHBAR_I915; > > - if (dev_priv->mchbar_need_disable) { > > - if (IS_I915G(dev_priv) || IS_I915GM(dev_priv)) { > > + if (i915->mchbar_need_disable) { > > + if (IS_I915G(i915) || IS_I915GM(i915)) { > > u32 deven_val; > > - pci_read_config_dword(dev_priv->bridge_dev, DEVEN, > > + pci_read_config_dword(i915->bridge_dev, DEVEN, > > &deven_val); > > deven_val &= ~DEVEN_MCHBAR_EN; > > - pci_write_config_dword(dev_priv->bridge_dev, DEVEN, > > + pci_write_config_dword(i915->bridge_dev, DEVEN, > > deven_val); > > } else { > > u32 mchbar_val; > > - pci_read_config_dword(dev_priv->bridge_dev, mchbar_reg, > > + pci_read_config_dword(i915->bridge_dev, mchbar_reg, > > &mchbar_val); > > mchbar_val &= ~1; > > - pci_write_config_dword(dev_priv->bridge_dev, mchbar_reg, > > + pci_write_config_dword(i915->bridge_dev, mchbar_reg, > > mchbar_val); > > } > > } > > - if (dev_priv->mch_res.start) > > - release_resource(&dev_priv->mch_res); > > + if (i915->mch_res.start) > > + release_resource(&i915->mch_res); > > } > > /* true = enable decode, false = disable decoder */ > > static unsigned int i915_vga_set_decode(void *cookie, bool state) > > { > > - struct drm_i915_private *dev_priv = cookie; > > + struct drm_i915_private *i915 = cookie; > > - intel_modeset_vga_set_state(dev_priv, state); > > + intel_modeset_vga_set_state(i915, state); > > if (state) > > return VGA_RSRC_LEGACY_IO | VGA_RSRC_LEGACY_MEM | > > VGA_RSRC_NORMAL_IO | VGA_RSRC_NORMAL_MEM; > > @@ -668,21 +668,21 @@ static const struct vga_switcheroo_client_ops i915_switcheroo_ops = { > > static int i915_load_modeset_init(struct drm_device *dev) > > { > > - struct drm_i915_private *dev_priv = to_i915(dev); > > - struct pci_dev *pdev = dev_priv->drm.pdev; > > + struct drm_i915_private *i915 = to_i915(dev); > > + struct pci_dev *pdev = i915->drm.pdev; > > int ret; > > if (i915_inject_load_failure()) > > return -ENODEV; > > - if (HAS_DISPLAY(dev_priv)) { > > - ret = drm_vblank_init(&dev_priv->drm, > > - INTEL_INFO(dev_priv)->num_pipes); > > + if (HAS_DISPLAY(i915)) { > > + ret = drm_vblank_init(&i915->drm, > > + INTEL_INFO(i915)->num_pipes); > > if (ret) > > goto out; > > } > > - intel_bios_init(dev_priv); > > + intel_bios_init(i915); > > /* If we have > 1 VGA cards, then we need to arbitrate access > > * to the common VGA resources. > > @@ -691,7 +691,7 @@ static int i915_load_modeset_init(struct drm_device *dev) > > * then we do not take part in VGA arbitration and the > > * vga_client_register() fails with -ENODEV. > > */ > > - ret = vga_client_register(pdev, dev_priv, NULL, i915_vga_set_decode); > > + ret = vga_client_register(pdev, i915, NULL, i915_vga_set_decode); > > if (ret && ret != -ENODEV) > > goto out; > > @@ -702,17 +702,17 @@ static int i915_load_modeset_init(struct drm_device *dev) > > goto cleanup_vga_client; > > /* must happen before intel_power_domains_init_hw() on VLV/CHV */ > > - intel_update_rawclk(dev_priv); > > + intel_update_rawclk(i915); > > - intel_power_domains_init_hw(dev_priv, false); > > + intel_power_domains_init_hw(i915, false); > > - intel_csr_ucode_init(dev_priv); > > + intel_csr_ucode_init(i915); > > - ret = intel_irq_install(dev_priv); > > + ret = intel_irq_install(i915); > > if (ret) > > goto cleanup_csr; > > - intel_gmbus_setup(dev_priv); > > + intel_gmbus_setup(i915); > > /* Important: The output setup functions called by modeset_init need > > * working irqs for e.g. gmbus and dp aux transfers. */ > > @@ -720,13 +720,13 @@ static int i915_load_modeset_init(struct drm_device *dev) > > if (ret) > > goto cleanup_irq; > > - ret = i915_gem_init(dev_priv); > > + ret = i915_gem_init(i915); > > if (ret) > > goto cleanup_modeset; > > - intel_overlay_setup(dev_priv); > > + intel_overlay_setup(i915); > > - if (!HAS_DISPLAY(dev_priv)) > > + if (!HAS_DISPLAY(i915)) > > return 0; > > ret = intel_fbdev_init(dev); > > @@ -734,24 +734,24 @@ static int i915_load_modeset_init(struct drm_device *dev) > > goto cleanup_gem; > > /* Only enable hotplug handling once the fbdev is fully set up. */ > > - intel_hpd_init(dev_priv); > > + intel_hpd_init(i915); > > - intel_init_ipc(dev_priv); > > + intel_init_ipc(i915); > > return 0; > > cleanup_gem: > > - i915_gem_suspend(dev_priv); > > - i915_gem_fini_hw(dev_priv); > > - i915_gem_fini(dev_priv); > > + i915_gem_suspend(i915); > > + i915_gem_fini_hw(i915); > > + i915_gem_fini(i915); > > cleanup_modeset: > > intel_modeset_cleanup(dev); > > cleanup_irq: > > drm_irq_uninstall(dev); > > - intel_gmbus_teardown(dev_priv); > > + intel_gmbus_teardown(i915); > > cleanup_csr: > > - intel_csr_ucode_fini(dev_priv); > > - intel_power_domains_fini_hw(dev_priv); > > + intel_csr_ucode_fini(i915); > > + intel_power_domains_fini_hw(i915); > > vga_switcheroo_unregister_client(pdev); > > cleanup_vga_client: > > vga_client_register(pdev, NULL, NULL, NULL); > > @@ -759,11 +759,11 @@ static int i915_load_modeset_init(struct drm_device *dev) > > return ret; > > } > > -static int i915_kick_out_firmware_fb(struct drm_i915_private *dev_priv) > > +static int i915_kick_out_firmware_fb(struct drm_i915_private *i915) > > { > > struct apertures_struct *ap; > > - struct pci_dev *pdev = dev_priv->drm.pdev; > > - struct i915_ggtt *ggtt = &dev_priv->ggtt; > > + struct pci_dev *pdev = i915->drm.pdev; > > + struct i915_ggtt *ggtt = &i915->ggtt; > > bool primary; > > int ret; > > @@ -784,22 +784,22 @@ static int i915_kick_out_firmware_fb(struct drm_i915_private *dev_priv) > > return ret; > > } > > -static void intel_init_dpio(struct drm_i915_private *dev_priv) > > +static void intel_init_dpio(struct drm_i915_private *i915) > > { > > /* > > * IOSF_PORT_DPIO is used for VLV x2 PHY (DP/HDMI B and C), > > * CHV x1 PHY (DP/HDMI D) > > * IOSF_PORT_DPIO_2 is used for CHV x2 PHY (DP/HDMI B and C) > > */ > > - if (IS_CHERRYVIEW(dev_priv)) { > > + if (IS_CHERRYVIEW(i915)) { > > DPIO_PHY_IOSF_PORT(DPIO_PHY0) = IOSF_PORT_DPIO_2; > > DPIO_PHY_IOSF_PORT(DPIO_PHY1) = IOSF_PORT_DPIO; > > - } else if (IS_VALLEYVIEW(dev_priv)) { > > + } else if (IS_VALLEYVIEW(i915)) { > > DPIO_PHY_IOSF_PORT(DPIO_PHY0) = IOSF_PORT_DPIO; > > } > > } > > -static int i915_workqueues_init(struct drm_i915_private *dev_priv) > > +static int i915_workqueues_init(struct drm_i915_private *i915) > > { > > /* > > * The i915 workqueue is primarily used for batched retirement of > > @@ -815,18 +815,18 @@ static int i915_workqueues_init(struct drm_i915_private *dev_priv) > > * so there is no point in running more than one instance of the > > * workqueue at any time. Use an ordered one. > > */ > > - dev_priv->wq = alloc_ordered_workqueue("i915", 0); > > - if (dev_priv->wq == NULL) > > + i915->wq = alloc_ordered_workqueue("i915", 0); > > + if (i915->wq == NULL) > > goto out_err; > > - dev_priv->hotplug.dp_wq = alloc_ordered_workqueue("i915-dp", 0); > > - if (dev_priv->hotplug.dp_wq == NULL) > > + i915->hotplug.dp_wq = alloc_ordered_workqueue("i915-dp", 0); > > + if (i915->hotplug.dp_wq == NULL) > > goto out_free_wq; > > return 0; > > out_free_wq: > > - destroy_workqueue(dev_priv->wq); > > + destroy_workqueue(i915->wq); > > out_err: > > DRM_ERROR("Failed to allocate workqueues.\n"); > > @@ -842,10 +842,10 @@ static void i915_engines_cleanup(struct drm_i915_private *i915) > > kfree(engine); > > } > > -static void i915_workqueues_cleanup(struct drm_i915_private *dev_priv) > > +static void i915_workqueues_cleanup(struct drm_i915_private *i915) > > { > > - destroy_workqueue(dev_priv->hotplug.dp_wq); > > - destroy_workqueue(dev_priv->wq); > > + destroy_workqueue(i915->hotplug.dp_wq); > > + destroy_workqueue(i915->wq); > > } > > /* > > @@ -858,14 +858,14 @@ static void i915_workqueues_cleanup(struct drm_i915_private *dev_priv) > > * (workarounds have a habit of persisting!). Anything older than that > > * should be removed along with the complications they introduce. > > */ > > -static void intel_detect_preproduction_hw(struct drm_i915_private *dev_priv) > > +static void intel_detect_preproduction_hw(struct drm_i915_private *i915) > > { > > bool pre = false; > > - pre |= IS_HSW_EARLY_SDV(dev_priv); > > - pre |= IS_SKL_REVID(dev_priv, 0, SKL_REVID_F0); > > - pre |= IS_BXT_REVID(dev_priv, 0, BXT_REVID_B_LAST); > > - pre |= IS_KBL_REVID(dev_priv, 0, KBL_REVID_A0); > > + pre |= IS_HSW_EARLY_SDV(i915); > > + pre |= IS_SKL_REVID(i915, 0, SKL_REVID_F0); > > + pre |= IS_BXT_REVID(i915, 0, BXT_REVID_B_LAST); > > + pre |= IS_KBL_REVID(i915, 0, KBL_REVID_A0); > > if (pre) { > > DRM_ERROR("This is a pre-production stepping. " > > @@ -876,7 +876,7 @@ static void intel_detect_preproduction_hw(struct drm_i915_private *dev_priv) > > /** > > * i915_driver_init_early - setup state not requiring device access > > - * @dev_priv: device private > > + * @i915: device private > > * > > * Initialize everything that is a "SW-only" state, that is state not > > * requiring accessing the device or exposing the driver via kernel internal > > @@ -884,152 +884,152 @@ static void intel_detect_preproduction_hw(struct drm_i915_private *dev_priv) > > * system memory allocation, setting up device specific attributes and > > * function hooks not requiring accessing the device. > > */ > > -static int i915_driver_init_early(struct drm_i915_private *dev_priv) > > +static int i915_driver_init_early(struct drm_i915_private *i915) > > { > > int ret = 0; > > if (i915_inject_load_failure()) > > return -ENODEV; > > - intel_device_info_subplatform_init(dev_priv); > > + intel_device_info_subplatform_init(i915); > > - intel_uncore_init_early(&dev_priv->uncore); > > + intel_uncore_init_early(&i915->uncore); > > - spin_lock_init(&dev_priv->irq_lock); > > - spin_lock_init(&dev_priv->gpu_error.lock); > > - mutex_init(&dev_priv->backlight_lock); > > + spin_lock_init(&i915->irq_lock); > > + spin_lock_init(&i915->gpu_error.lock); > > + mutex_init(&i915->backlight_lock); > > - mutex_init(&dev_priv->sb_lock); > > - pm_qos_add_request(&dev_priv->sb_qos, > > + mutex_init(&i915->sb_lock); > > + pm_qos_add_request(&i915->sb_qos, > > PM_QOS_CPU_DMA_LATENCY, PM_QOS_DEFAULT_VALUE); > > - mutex_init(&dev_priv->av_mutex); > > - mutex_init(&dev_priv->wm.wm_mutex); > > - mutex_init(&dev_priv->pps_mutex); > > - mutex_init(&dev_priv->hdcp_comp_mutex); > > + mutex_init(&i915->av_mutex); > > + mutex_init(&i915->wm.wm_mutex); > > + mutex_init(&i915->pps_mutex); > > + mutex_init(&i915->hdcp_comp_mutex); > > - i915_memcpy_init_early(dev_priv); > > - intel_runtime_pm_init_early(dev_priv); > > + i915_memcpy_init_early(i915); > > + intel_runtime_pm_init_early(i915); > > - ret = i915_workqueues_init(dev_priv); > > + ret = i915_workqueues_init(i915); > > if (ret < 0) > > goto err_engines; > > - ret = i915_gem_init_early(dev_priv); > > + ret = i915_gem_init_early(i915); > > if (ret < 0) > > goto err_workqueues; > > /* This must be called before any calls to HAS_PCH_* */ > > - intel_detect_pch(dev_priv); > > + intel_detect_pch(i915); > > - intel_wopcm_init_early(&dev_priv->wopcm); > > - intel_uc_init_early(dev_priv); > > - intel_pm_setup(dev_priv); > > - intel_init_dpio(dev_priv); > > - ret = intel_power_domains_init(dev_priv); > > + intel_wopcm_init_early(&i915->wopcm); > > + intel_uc_init_early(i915); > > + intel_pm_setup(i915); > > + intel_init_dpio(i915); > > + ret = intel_power_domains_init(i915); > > if (ret < 0) > > goto err_uc; > > - intel_irq_init(dev_priv); > > - intel_hangcheck_init(dev_priv); > > - intel_init_display_hooks(dev_priv); > > - intel_init_clock_gating_hooks(dev_priv); > > - intel_init_audio_hooks(dev_priv); > > - intel_display_crc_init(dev_priv); > > + intel_irq_init(i915); > > + intel_hangcheck_init(i915); > > + intel_init_display_hooks(i915); > > + intel_init_clock_gating_hooks(i915); > > + intel_init_audio_hooks(i915); > > + intel_display_crc_init(i915); > > - intel_detect_preproduction_hw(dev_priv); > > + intel_detect_preproduction_hw(i915); > > return 0; > > err_uc: > > - intel_uc_cleanup_early(dev_priv); > > - i915_gem_cleanup_early(dev_priv); > > + intel_uc_cleanup_early(i915); > > + i915_gem_cleanup_early(i915); > > err_workqueues: > > - i915_workqueues_cleanup(dev_priv); > > + i915_workqueues_cleanup(i915); > > err_engines: > > - i915_engines_cleanup(dev_priv); > > + i915_engines_cleanup(i915); > > return ret; > > } > > /** > > * i915_driver_cleanup_early - cleanup the setup done in i915_driver_init_early() > > - * @dev_priv: device private > > + * @i915: device private > > */ > > -static void i915_driver_cleanup_early(struct drm_i915_private *dev_priv) > > +static void i915_driver_cleanup_early(struct drm_i915_private *i915) > > { > > - intel_irq_fini(dev_priv); > > - intel_power_domains_cleanup(dev_priv); > > - intel_uc_cleanup_early(dev_priv); > > - i915_gem_cleanup_early(dev_priv); > > - i915_workqueues_cleanup(dev_priv); > > - i915_engines_cleanup(dev_priv); > > + intel_irq_fini(i915); > > + intel_power_domains_cleanup(i915); > > + intel_uc_cleanup_early(i915); > > + i915_gem_cleanup_early(i915); > > + i915_workqueues_cleanup(i915); > > + i915_engines_cleanup(i915); > > - pm_qos_remove_request(&dev_priv->sb_qos); > > - mutex_destroy(&dev_priv->sb_lock); > > + pm_qos_remove_request(&i915->sb_qos); > > + mutex_destroy(&i915->sb_lock); > > } > > /** > > * i915_driver_init_mmio - setup device MMIO > > - * @dev_priv: device private > > + * @i915: device private > > * > > * Setup minimal device state necessary for MMIO accesses later in the > > * initialization sequence. The setup here should avoid any other device-wide > > * side effects or exposing the driver via kernel internal or user space > > * interfaces. > > */ > > -static int i915_driver_init_mmio(struct drm_i915_private *dev_priv) > > +static int i915_driver_init_mmio(struct drm_i915_private *i915) > > { > > int ret; > > if (i915_inject_load_failure()) > > return -ENODEV; > > - if (i915_get_bridge_dev(dev_priv)) > > + if (i915_get_bridge_dev(i915)) > > return -EIO; > > - ret = intel_uncore_init_mmio(&dev_priv->uncore); > > + ret = intel_uncore_init_mmio(&i915->uncore); > > if (ret < 0) > > goto err_bridge; > > /* Try to make sure MCHBAR is enabled before poking at it */ > > - intel_setup_mchbar(dev_priv); > > + intel_setup_mchbar(i915); > > - intel_device_info_init_mmio(dev_priv); > > + intel_device_info_init_mmio(i915); > > - intel_uncore_prune_mmio_domains(&dev_priv->uncore); > > + intel_uncore_prune_mmio_domains(&i915->uncore); > > - intel_uc_init_mmio(dev_priv); > > + intel_uc_init_mmio(i915); > > - ret = intel_engines_init_mmio(dev_priv); > > + ret = intel_engines_init_mmio(i915); > > if (ret) > > goto err_uncore; > > - i915_gem_init_mmio(dev_priv); > > + i915_gem_init_mmio(i915); > > return 0; > > err_uncore: > > - intel_teardown_mchbar(dev_priv); > > - intel_uncore_fini_mmio(&dev_priv->uncore); > > + intel_teardown_mchbar(i915); > > + intel_uncore_fini_mmio(&i915->uncore); > > err_bridge: > > - pci_dev_put(dev_priv->bridge_dev); > > + pci_dev_put(i915->bridge_dev); > > return ret; > > } > > /** > > * i915_driver_cleanup_mmio - cleanup the setup done in i915_driver_init_mmio() > > - * @dev_priv: device private > > + * @i915: device private > > */ > > -static void i915_driver_cleanup_mmio(struct drm_i915_private *dev_priv) > > +static void i915_driver_cleanup_mmio(struct drm_i915_private *i915) > > { > > - intel_teardown_mchbar(dev_priv); > > - intel_uncore_fini_mmio(&dev_priv->uncore); > > - pci_dev_put(dev_priv->bridge_dev); > > + intel_teardown_mchbar(i915); > > + intel_uncore_fini_mmio(&i915->uncore); > > + pci_dev_put(i915->bridge_dev); > > } > > -static void intel_sanitize_options(struct drm_i915_private *dev_priv) > > +static void intel_sanitize_options(struct drm_i915_private *i915) > > { > > - intel_gvt_sanitize_options(dev_priv); > > + intel_gvt_sanitize_options(i915); > > } > > #define DRAM_TYPE_STR(type) [INTEL_DRAM_ ## type] = #type > > @@ -1131,11 +1131,11 @@ skl_is_16gb_dimm(const struct dram_dimm_info *dimm) > > } > > static void > > -skl_dram_get_dimm_info(struct drm_i915_private *dev_priv, > > +skl_dram_get_dimm_info(struct drm_i915_private *i915, > > struct dram_dimm_info *dimm, > > int channel, char dimm_name, u16 val) > > { > > - if (INTEL_GEN(dev_priv) >= 10) { > > + if (INTEL_GEN(i915) >= 10) { > > dimm->size = cnl_get_dimm_size(val); > > dimm->width = cnl_get_dimm_width(val); > > dimm->ranks = cnl_get_dimm_ranks(val); > > @@ -1151,13 +1151,13 @@ skl_dram_get_dimm_info(struct drm_i915_private *dev_priv, > > } > > static int > > -skl_dram_get_channel_info(struct drm_i915_private *dev_priv, > > +skl_dram_get_channel_info(struct drm_i915_private *i915, > > struct dram_channel_info *ch, > > int channel, u32 val) > > { > > - skl_dram_get_dimm_info(dev_priv, &ch->dimm_l, > > + skl_dram_get_dimm_info(i915, &ch->dimm_l, > > channel, 'L', val & 0xffff); > > - skl_dram_get_dimm_info(dev_priv, &ch->dimm_s, > > + skl_dram_get_dimm_info(i915, &ch->dimm_s, > > channel, 'S', val >> 16); > > if (ch->dimm_l.size == 0 && ch->dimm_s.size == 0) { > > @@ -1192,20 +1192,20 @@ intel_is_dram_symmetric(const struct dram_channel_info *ch0, > > } > > static int > > -skl_dram_get_channels_info(struct drm_i915_private *dev_priv) > > +skl_dram_get_channels_info(struct drm_i915_private *i915) > > { > > - struct dram_info *dram_info = &dev_priv->dram_info; > > + struct dram_info *dram_info = &i915->dram_info; > > struct dram_channel_info ch0 = {}, ch1 = {}; > > u32 val; > > int ret; > > val = I915_READ(SKL_MAD_DIMM_CH0_0_0_0_MCHBAR_MCMAIN); > > - ret = skl_dram_get_channel_info(dev_priv, &ch0, 0, val); > > + ret = skl_dram_get_channel_info(i915, &ch0, 0, val); > > if (ret == 0) > > dram_info->num_channels++; > > val = I915_READ(SKL_MAD_DIMM_CH1_0_0_0_MCHBAR_MCMAIN); > > - ret = skl_dram_get_channel_info(dev_priv, &ch1, 1, val); > > + ret = skl_dram_get_channel_info(i915, &ch1, 1, val); > > if (ret == 0) > > dram_info->num_channels++; > > @@ -1239,7 +1239,7 @@ skl_dram_get_channels_info(struct drm_i915_private *dev_priv) > > } > > static enum intel_dram_type > > -skl_get_dram_type(struct drm_i915_private *dev_priv) > > +skl_get_dram_type(struct drm_i915_private *i915) > > { > > u32 val; > > @@ -1261,16 +1261,16 @@ skl_get_dram_type(struct drm_i915_private *dev_priv) > > } > > static int > > -skl_get_dram_info(struct drm_i915_private *dev_priv) > > +skl_get_dram_info(struct drm_i915_private *i915) > > { > > - struct dram_info *dram_info = &dev_priv->dram_info; > > + struct dram_info *dram_info = &i915->dram_info; > > u32 mem_freq_khz, val; > > int ret; > > - dram_info->type = skl_get_dram_type(dev_priv); > > + dram_info->type = skl_get_dram_type(i915); > > DRM_DEBUG_KMS("DRAM type: %s\n", intel_dram_type_str(dram_info->type)); > > - ret = skl_dram_get_channels_info(dev_priv); > > + ret = skl_dram_get_channels_info(i915); > > if (ret) > > return ret; > > @@ -1370,9 +1370,9 @@ static void bxt_get_dimm_info(struct dram_dimm_info *dimm, > > } > > static int > > -bxt_get_dram_info(struct drm_i915_private *dev_priv) > > +bxt_get_dram_info(struct drm_i915_private *i915) > > { > > - struct dram_info *dram_info = &dev_priv->dram_info; > > + struct dram_info *dram_info = &i915->dram_info; > > u32 dram_channels; > > u32 mem_freq_khz, val; > > u8 num_active_channels; > > @@ -1443,9 +1443,9 @@ bxt_get_dram_info(struct drm_i915_private *dev_priv) > > } > > static void > > -intel_get_dram_info(struct drm_i915_private *dev_priv) > > +intel_get_dram_info(struct drm_i915_private *i915) > > { > > - struct dram_info *dram_info = &dev_priv->dram_info; > > + struct dram_info *dram_info = &i915->dram_info; > > int ret; > > /* > > @@ -1453,15 +1453,15 @@ intel_get_dram_info(struct drm_i915_private *dev_priv) > > * This is only used for the level 0 watermark latency > > * w/a which does not apply to bxt/glk. > > */ > > - dram_info->is_16gb_dimm = !IS_GEN9_LP(dev_priv); > > + dram_info->is_16gb_dimm = !IS_GEN9_LP(i915); > > - if (INTEL_GEN(dev_priv) < 9) > > + if (INTEL_GEN(i915) < 9) > > return; > > - if (IS_GEN9_LP(dev_priv)) > > - ret = bxt_get_dram_info(dev_priv); > > + if (IS_GEN9_LP(i915)) > > + ret = bxt_get_dram_info(i915); > > else > > - ret = skl_get_dram_info(dev_priv); > > + ret = skl_get_dram_info(i915); > > if (ret) > > return; > > @@ -1473,7 +1473,7 @@ intel_get_dram_info(struct drm_i915_private *dev_priv) > > dram_info->ranks, yesno(dram_info->is_16gb_dimm)); > > } > > -static u32 gen9_edram_size_mb(struct drm_i915_private *dev_priv, u32 cap) > > +static u32 gen9_edram_size_mb(struct drm_i915_private *i915, u32 cap) > > { > > const unsigned int ways[8] = { 4, 8, 12, 16, 16, 16, 16, 16 }; > > const unsigned int sets[4] = { 1, 1, 2, 2 }; > > @@ -1483,16 +1483,16 @@ static u32 gen9_edram_size_mb(struct drm_i915_private *dev_priv, u32 cap) > > sets[EDRAM_SETS_IDX(cap)]; > > } > > -static void edram_detect(struct drm_i915_private *dev_priv) > > +static void edram_detect(struct drm_i915_private *i915) > > { > > u32 edram_cap = 0; > > - if (!(IS_HASWELL(dev_priv) || > > - IS_BROADWELL(dev_priv) || > > - INTEL_GEN(dev_priv) >= 9)) > > + if (!(IS_HASWELL(i915) || > > + IS_BROADWELL(i915) || > > + INTEL_GEN(i915) >= 9)) > > return; > > - edram_cap = __raw_uncore_read32(&dev_priv->uncore, HSW_EDRAM_CAP); > > + edram_cap = __raw_uncore_read32(&i915->uncore, HSW_EDRAM_CAP); > > /* NB: We can't write IDICR yet because we don't have gt funcs set up */ > > @@ -1503,63 +1503,63 @@ static void edram_detect(struct drm_i915_private *dev_priv) > > * The needed capability bits for size calculation are not there with > > * pre gen9 so return 128MB always. > > */ > > - if (INTEL_GEN(dev_priv) < 9) > > - dev_priv->edram_size_mb = 128; > > + if (INTEL_GEN(i915) < 9) > > + i915->edram_size_mb = 128; > > else > > - dev_priv->edram_size_mb = > > - gen9_edram_size_mb(dev_priv, edram_cap); > > + i915->edram_size_mb = > > + gen9_edram_size_mb(i915, edram_cap); > > - DRM_INFO("Found %uMB of eDRAM\n", dev_priv->edram_size_mb); > > + DRM_INFO("Found %uMB of eDRAM\n", i915->edram_size_mb); > > } > > /** > > * i915_driver_init_hw - setup state requiring device access > > - * @dev_priv: device private > > + * @i915: device private > > * > > * Setup state that requires accessing the device, but doesn't require > > * exposing the driver via kernel internal or userspace interfaces. > > */ > > -static int i915_driver_init_hw(struct drm_i915_private *dev_priv) > > +static int i915_driver_init_hw(struct drm_i915_private *i915) > > { > > - struct pci_dev *pdev = dev_priv->drm.pdev; > > + struct pci_dev *pdev = i915->drm.pdev; > > int ret; > > if (i915_inject_load_failure()) > > return -ENODEV; > > - intel_device_info_runtime_init(dev_priv); > > + intel_device_info_runtime_init(i915); > > - if (HAS_PPGTT(dev_priv)) { > > - if (intel_vgpu_active(dev_priv) && > > - !intel_vgpu_has_full_ppgtt(dev_priv)) { > > - i915_report_error(dev_priv, > > + if (HAS_PPGTT(i915)) { > > + if (intel_vgpu_active(i915) && > > + !intel_vgpu_has_full_ppgtt(i915)) { > > + i915_report_error(i915, > > "incompatible vGPU found, support for isolated ppGTT required\n"); > > return -ENXIO; > > } > > } > > - if (HAS_EXECLISTS(dev_priv)) { > > + if (HAS_EXECLISTS(i915)) { > > /* > > * Older GVT emulation depends upon intercepting CSB mmio, > > * which we no longer use, preferring to use the HWSP cache > > * instead. > > */ > > - if (intel_vgpu_active(dev_priv) && > > - !intel_vgpu_has_hwsp_emulation(dev_priv)) { > > - i915_report_error(dev_priv, > > + if (intel_vgpu_active(i915) && > > + !intel_vgpu_has_hwsp_emulation(i915)) { > > + i915_report_error(i915, > > "old vGPU host found, support for HWSP emulation required\n"); > > return -ENXIO; > > } > > } > > - intel_sanitize_options(dev_priv); > > + intel_sanitize_options(i915); > > /* needs to be done before ggtt probe */ > > - edram_detect(dev_priv); > > + edram_detect(i915); > > - i915_perf_init(dev_priv); > > + i915_perf_init(i915); > > - ret = i915_ggtt_probe_hw(dev_priv); > > + ret = i915_ggtt_probe_hw(i915); > > if (ret) > > goto err_perf; > > @@ -1567,7 +1567,7 @@ static int i915_driver_init_hw(struct drm_i915_private *dev_priv) > > * WARNING: Apparently we must kick fbdev drivers before vgacon, > > * otherwise the vga fbdev driver falls over. > > */ > > - ret = i915_kick_out_firmware_fb(dev_priv); > > + ret = i915_kick_out_firmware_fb(i915); > > if (ret) { > > DRM_ERROR("failed to remove conflicting framebuffer drivers\n"); > > goto err_ggtt; > > @@ -1579,11 +1579,11 @@ static int i915_driver_init_hw(struct drm_i915_private *dev_priv) > > goto err_ggtt; > > } > > - ret = i915_ggtt_init_hw(dev_priv); > > + ret = i915_ggtt_init_hw(i915); > > if (ret) > > goto err_ggtt; > > - ret = i915_ggtt_enable_hw(dev_priv); > > + ret = i915_ggtt_enable_hw(i915); > > if (ret) { > > DRM_ERROR("failed to enable GGTT\n"); > > goto err_ggtt; > > @@ -1592,7 +1592,7 @@ static int i915_driver_init_hw(struct drm_i915_private *dev_priv) > > pci_set_master(pdev); > > /* overlay on gen2 is broken and can't address above 1G */ > > - if (IS_GEN(dev_priv, 2)) { > > + if (IS_GEN(i915, 2)) { > > ret = dma_set_coherent_mask(&pdev->dev, DMA_BIT_MASK(30)); > > if (ret) { > > DRM_ERROR("failed to set DMA mask\n"); > > @@ -1609,7 +1609,7 @@ static int i915_driver_init_hw(struct drm_i915_private *dev_priv) > > * behaviour if any general state is accessed within a page above 4GB, > > * which also needs to be handled carefully. > > */ > > - if (IS_I965G(dev_priv) || IS_I965GM(dev_priv)) { > > + if (IS_I965G(i915) || IS_I965GM(i915)) { > > ret = dma_set_coherent_mask(&pdev->dev, DMA_BIT_MASK(32)); > > if (ret) { > > @@ -1619,13 +1619,13 @@ static int i915_driver_init_hw(struct drm_i915_private *dev_priv) > > } > > } > > - pm_qos_add_request(&dev_priv->pm_qos, PM_QOS_CPU_DMA_LATENCY, > > + pm_qos_add_request(&i915->pm_qos, PM_QOS_CPU_DMA_LATENCY, > > PM_QOS_DEFAULT_VALUE); > > - intel_uncore_sanitize(dev_priv); > > + intel_uncore_sanitize(i915); > > - intel_gt_init_workarounds(dev_priv); > > - i915_gem_load_init_fences(dev_priv); > > + intel_gt_init_workarounds(i915); > > + i915_gem_load_init_fences(i915); > > /* On the 945G/GM, the chipset reports the MSI capability on the > > * integrated graphics even though the support isn't actually there > > @@ -1646,94 +1646,94 @@ static int i915_driver_init_hw(struct drm_i915_private *dev_priv) > > * device. The kernel then disables that interrupt source and so > > * prevents the other device from working properly. > > */ > > - if (INTEL_GEN(dev_priv) >= 5) { > > + if (INTEL_GEN(i915) >= 5) { > > if (pci_enable_msi(pdev) < 0) > > DRM_DEBUG_DRIVER("can't enable MSI"); > > } > > - ret = intel_gvt_init(dev_priv); > > + ret = intel_gvt_init(i915); > > if (ret) > > goto err_msi; > > - intel_opregion_setup(dev_priv); > > + intel_opregion_setup(i915); > > /* > > * Fill the dram structure to get the system raw bandwidth and > > * dram info. This will be used for memory latency calculation. > > */ > > - intel_get_dram_info(dev_priv); > > + intel_get_dram_info(i915); > > - intel_bw_init_hw(dev_priv); > > + intel_bw_init_hw(i915); > > return 0; > > err_msi: > > if (pdev->msi_enabled) > > pci_disable_msi(pdev); > > - pm_qos_remove_request(&dev_priv->pm_qos); > > + pm_qos_remove_request(&i915->pm_qos); > > err_ggtt: > > - i915_ggtt_cleanup_hw(dev_priv); > > + i915_ggtt_cleanup_hw(i915); > > err_perf: > > - i915_perf_fini(dev_priv); > > + i915_perf_fini(i915); > > return ret; > > } > > /** > > * i915_driver_cleanup_hw - cleanup the setup done in i915_driver_init_hw() > > - * @dev_priv: device private > > + * @i915: device private > > */ > > -static void i915_driver_cleanup_hw(struct drm_i915_private *dev_priv) > > +static void i915_driver_cleanup_hw(struct drm_i915_private *i915) > > { > > - struct pci_dev *pdev = dev_priv->drm.pdev; > > + struct pci_dev *pdev = i915->drm.pdev; > > - i915_perf_fini(dev_priv); > > + i915_perf_fini(i915); > > if (pdev->msi_enabled) > > pci_disable_msi(pdev); > > - pm_qos_remove_request(&dev_priv->pm_qos); > > + pm_qos_remove_request(&i915->pm_qos); > > } > > /** > > * i915_driver_register - register the driver with the rest of the system > > - * @dev_priv: device private > > + * @i915: device private > > * > > * Perform any steps necessary to make the driver available via kernel > > * internal or userspace interfaces. > > */ > > -static void i915_driver_register(struct drm_i915_private *dev_priv) > > +static void i915_driver_register(struct drm_i915_private *i915) > > { > > - struct drm_device *dev = &dev_priv->drm; > > + struct drm_device *dev = &i915->drm; > > - i915_gem_shrinker_register(dev_priv); > > - i915_pmu_register(dev_priv); > > + i915_gem_shrinker_register(i915); > > + i915_pmu_register(i915); > > /* > > * Notify a valid surface after modesetting, > > * when running inside a VM. > > */ > > - if (intel_vgpu_active(dev_priv)) > > + if (intel_vgpu_active(i915)) > > I915_WRITE(vgtif_reg(display_ready), VGT_DRV_DISPLAY_READY); > > /* Reveal our presence to userspace */ > > if (drm_dev_register(dev, 0) == 0) { > > - i915_debugfs_register(dev_priv); > > - i915_setup_sysfs(dev_priv); > > + i915_debugfs_register(i915); > > + i915_setup_sysfs(i915); > > /* Depends on sysfs having been initialized */ > > - i915_perf_register(dev_priv); > > + i915_perf_register(i915); > > } else > > DRM_ERROR("Failed to register driver for userspace access!\n"); > > - if (HAS_DISPLAY(dev_priv)) { > > + if (HAS_DISPLAY(i915)) { > > /* Must be done after probing outputs */ > > - intel_opregion_register(dev_priv); > > + intel_opregion_register(i915); > > acpi_video_register(); > > } > > - if (IS_GEN(dev_priv, 5)) > > - intel_gpu_ips_init(dev_priv); > > + if (IS_GEN(i915, 5)) > > + intel_gpu_ips_init(i915); > > - intel_audio_init(dev_priv); > > + intel_audio_init(i915); > > /* > > * Some ports require correctly set-up hpd registers for detection to > > @@ -1748,60 +1748,60 @@ static void i915_driver_register(struct drm_i915_private *dev_priv) > > * We need to coordinate the hotplugs with the asynchronous fbdev > > * configuration, for which we use the fbdev->async_cookie. > > */ > > - if (HAS_DISPLAY(dev_priv)) > > + if (HAS_DISPLAY(i915)) > > drm_kms_helper_poll_init(dev); > > - intel_power_domains_enable(dev_priv); > > - intel_runtime_pm_enable(dev_priv); > > + intel_power_domains_enable(i915); > > + intel_runtime_pm_enable(i915); > > } > > /** > > * i915_driver_unregister - cleanup the registration done in i915_driver_regiser() > > - * @dev_priv: device private > > + * @i915: device private > > */ > > -static void i915_driver_unregister(struct drm_i915_private *dev_priv) > > +static void i915_driver_unregister(struct drm_i915_private *i915) > > { > > - intel_runtime_pm_disable(dev_priv); > > - intel_power_domains_disable(dev_priv); > > + intel_runtime_pm_disable(i915); > > + intel_power_domains_disable(i915); > > - intel_fbdev_unregister(dev_priv); > > - intel_audio_deinit(dev_priv); > > + intel_fbdev_unregister(i915); > > + intel_audio_deinit(i915); > > /* > > * After flushing the fbdev (incl. a late async config which will > > * have delayed queuing of a hotplug event), then flush the hotplug > > * events. > > */ > > - drm_kms_helper_poll_fini(&dev_priv->drm); > > + drm_kms_helper_poll_fini(&i915->drm); > > intel_gpu_ips_teardown(); > > acpi_video_unregister(); > > - intel_opregion_unregister(dev_priv); > > + intel_opregion_unregister(i915); > > - i915_perf_unregister(dev_priv); > > - i915_pmu_unregister(dev_priv); > > + i915_perf_unregister(i915); > > + i915_pmu_unregister(i915); > > - i915_teardown_sysfs(dev_priv); > > - drm_dev_unplug(&dev_priv->drm); > > + i915_teardown_sysfs(i915); > > + drm_dev_unplug(&i915->drm); > > - i915_gem_shrinker_unregister(dev_priv); > > + i915_gem_shrinker_unregister(i915); > > } > > -static void i915_welcome_messages(struct drm_i915_private *dev_priv) > > +static void i915_welcome_messages(struct drm_i915_private *i915) > > { > > if (drm_debug & DRM_UT_DRIVER) { > > struct drm_printer p = drm_debug_printer("i915 device info:"); > > drm_printf(&p, "pciid=0x%04x rev=0x%02x platform=%s (subplatform=0x%x) gen=%i\n", > > - INTEL_DEVID(dev_priv), > > - INTEL_REVID(dev_priv), > > - intel_platform_name(INTEL_INFO(dev_priv)->platform), > > - intel_subplatform(RUNTIME_INFO(dev_priv), > > - INTEL_INFO(dev_priv)->platform), > > - INTEL_GEN(dev_priv)); > > - > > - intel_device_info_dump_flags(INTEL_INFO(dev_priv), &p); > > - intel_device_info_dump_runtime(RUNTIME_INFO(dev_priv), &p); > > + INTEL_DEVID(i915), > > + INTEL_REVID(i915), > > + intel_platform_name(INTEL_INFO(i915)->platform), > > + intel_subplatform(RUNTIME_INFO(i915), > > + INTEL_INFO(i915)->platform), > > + INTEL_GEN(i915)); > > + > > + intel_device_info_dump_flags(INTEL_INFO(i915), &p); > > + intel_device_info_dump_runtime(RUNTIME_INFO(i915), &p); > > } > > if (IS_ENABLED(CONFIG_DRM_I915_DEBUG)) > > @@ -1871,126 +1871,126 @@ int i915_driver_load(struct pci_dev *pdev, const struct pci_device_id *ent) > > { > > const struct intel_device_info *match_info = > > (struct intel_device_info *)ent->driver_data; > > - struct drm_i915_private *dev_priv; > > + struct drm_i915_private *i915; > > int ret; > > - dev_priv = i915_driver_create(pdev, ent); > > - if (IS_ERR(dev_priv)) > > - return PTR_ERR(dev_priv); > > + i915 = i915_driver_create(pdev, ent); > > + if (IS_ERR(i915)) > > + return PTR_ERR(i915); > > /* Disable nuclear pageflip by default on pre-ILK */ > > if (!i915_modparams.nuclear_pageflip && match_info->gen < 5) > > - dev_priv->drm.driver_features &= ~DRIVER_ATOMIC; > > + i915->drm.driver_features &= ~DRIVER_ATOMIC; > > ret = pci_enable_device(pdev); > > if (ret) > > goto out_fini; > > - ret = i915_driver_init_early(dev_priv); > > + ret = i915_driver_init_early(i915); > > if (ret < 0) > > goto out_pci_disable; > > - disable_rpm_wakeref_asserts(dev_priv); > > + disable_rpm_wakeref_asserts(i915); > > - ret = i915_driver_init_mmio(dev_priv); > > + ret = i915_driver_init_mmio(i915); > > if (ret < 0) > > goto out_runtime_pm_put; > > - ret = i915_driver_init_hw(dev_priv); > > + ret = i915_driver_init_hw(i915); > > if (ret < 0) > > goto out_cleanup_mmio; > > - ret = i915_load_modeset_init(&dev_priv->drm); > > + ret = i915_load_modeset_init(&i915->drm); > > if (ret < 0) > > goto out_cleanup_hw; > > - i915_driver_register(dev_priv); > > + i915_driver_register(i915); > > - enable_rpm_wakeref_asserts(dev_priv); > > + enable_rpm_wakeref_asserts(i915); > > - i915_welcome_messages(dev_priv); > > + i915_welcome_messages(i915); > > return 0; > > out_cleanup_hw: > > - i915_driver_cleanup_hw(dev_priv); > > - i915_ggtt_cleanup_hw(dev_priv); > > + i915_driver_cleanup_hw(i915); > > + i915_ggtt_cleanup_hw(i915); > > out_cleanup_mmio: > > - i915_driver_cleanup_mmio(dev_priv); > > + i915_driver_cleanup_mmio(i915); > > out_runtime_pm_put: > > - enable_rpm_wakeref_asserts(dev_priv); > > - i915_driver_cleanup_early(dev_priv); > > + enable_rpm_wakeref_asserts(i915); > > + i915_driver_cleanup_early(i915); > > out_pci_disable: > > pci_disable_device(pdev); > > out_fini: > > - i915_load_error(dev_priv, "Device initialization failed (%d)\n", ret); > > - i915_driver_destroy(dev_priv); > > + i915_load_error(i915, "Device initialization failed (%d)\n", ret); > > + i915_driver_destroy(i915); > > return ret; > > } > > void i915_driver_unload(struct drm_device *dev) > > { > > - struct drm_i915_private *dev_priv = to_i915(dev); > > - struct pci_dev *pdev = dev_priv->drm.pdev; > > + struct drm_i915_private *i915 = to_i915(dev); > > + struct pci_dev *pdev = i915->drm.pdev; > > - disable_rpm_wakeref_asserts(dev_priv); > > + disable_rpm_wakeref_asserts(i915); > > - i915_driver_unregister(dev_priv); > > + i915_driver_unregister(i915); > > /* > > * After unregistering the device to prevent any new users, cancel > > * all in-flight requests so that we can quickly unbind the active > > * resources. > > */ > > - i915_gem_set_wedged(dev_priv); > > + i915_gem_set_wedged(i915); > > /* Flush any external code that still may be under the RCU lock */ > > synchronize_rcu(); > > - i915_gem_suspend(dev_priv); > > + i915_gem_suspend(i915); > > drm_atomic_helper_shutdown(dev); > > - intel_gvt_cleanup(dev_priv); > > + intel_gvt_cleanup(i915); > > intel_modeset_cleanup(dev); > > - intel_bios_cleanup(dev_priv); > > + intel_bios_cleanup(i915); > > vga_switcheroo_unregister_client(pdev); > > vga_client_register(pdev, NULL, NULL, NULL); > > - intel_csr_ucode_fini(dev_priv); > > + intel_csr_ucode_fini(i915); > > /* Free error state after interrupts are fully disabled. */ > > - cancel_delayed_work_sync(&dev_priv->gpu_error.hangcheck_work); > > - i915_reset_error_state(dev_priv); > > + cancel_delayed_work_sync(&i915->gpu_error.hangcheck_work); > > + i915_reset_error_state(i915); > > - i915_gem_fini_hw(dev_priv); > > + i915_gem_fini_hw(i915); > > - intel_power_domains_fini_hw(dev_priv); > > + intel_power_domains_fini_hw(i915); > > - i915_driver_cleanup_hw(dev_priv); > > + i915_driver_cleanup_hw(i915); > > - enable_rpm_wakeref_asserts(dev_priv); > > + enable_rpm_wakeref_asserts(i915); > > } > > static void i915_driver_release(struct drm_device *dev) > > { > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > - disable_rpm_wakeref_asserts(dev_priv); > > + disable_rpm_wakeref_asserts(i915); > > - i915_gem_fini(dev_priv); > > + i915_gem_fini(i915); > > - i915_ggtt_cleanup_hw(dev_priv); > > - i915_driver_cleanup_mmio(dev_priv); > > + i915_ggtt_cleanup_hw(i915); > > + i915_driver_cleanup_mmio(i915); > > - enable_rpm_wakeref_asserts(dev_priv); > > - intel_runtime_pm_cleanup(dev_priv); > > + enable_rpm_wakeref_asserts(i915); > > + intel_runtime_pm_cleanup(i915); > > - i915_driver_cleanup_early(dev_priv); > > - i915_driver_destroy(dev_priv); > > + i915_driver_cleanup_early(i915); > > + i915_driver_destroy(i915); > > } > > static int i915_driver_open(struct drm_device *dev, struct drm_file *file) > > @@ -2035,9 +2035,9 @@ static void i915_driver_postclose(struct drm_device *dev, struct drm_file *file) > > kfree(file_priv); > > } > > -static void intel_suspend_encoders(struct drm_i915_private *dev_priv) > > +static void intel_suspend_encoders(struct drm_i915_private *i915) > > { > > - struct drm_device *dev = &dev_priv->drm; > > + struct drm_device *dev = &i915->drm; > > struct intel_encoder *encoder; > > drm_modeset_lock_all(dev); > > @@ -2047,11 +2047,11 @@ static void intel_suspend_encoders(struct drm_i915_private *dev_priv) > > drm_modeset_unlock_all(dev); > > } > > -static int vlv_resume_prepare(struct drm_i915_private *dev_priv, > > +static int vlv_resume_prepare(struct drm_i915_private *i915, > > bool rpm_resume); > > -static int vlv_suspend_complete(struct drm_i915_private *dev_priv); > > +static int vlv_suspend_complete(struct drm_i915_private *i915); > > -static bool suspend_to_idle(struct drm_i915_private *dev_priv) > > +static bool suspend_to_idle(struct drm_i915_private *i915) > > { > > #if IS_ENABLED(CONFIG_ACPI_SLEEP) > > if (acpi_target_system_state() < ACPI_STATE_S3) > > @@ -2077,15 +2077,15 @@ static int i915_drm_prepare(struct drm_device *dev) > > static int i915_drm_suspend(struct drm_device *dev) > > { > > - struct drm_i915_private *dev_priv = to_i915(dev); > > - struct pci_dev *pdev = dev_priv->drm.pdev; > > + struct drm_i915_private *i915 = to_i915(dev); > > + struct pci_dev *pdev = i915->drm.pdev; > > pci_power_t opregion_target_state; > > - disable_rpm_wakeref_asserts(dev_priv); > > + disable_rpm_wakeref_asserts(i915); > > /* We do a lot of poking in a lot of registers, make sure they work > > * properly. */ > > - intel_power_domains_disable(dev_priv); > > + intel_power_domains_disable(i915); > > drm_kms_helper_poll_disable(dev); > > @@ -2093,40 +2093,40 @@ static int i915_drm_suspend(struct drm_device *dev) > > intel_display_suspend(dev); > > - intel_dp_mst_suspend(dev_priv); > > + intel_dp_mst_suspend(i915); > > - intel_runtime_pm_disable_interrupts(dev_priv); > > - intel_hpd_cancel_work(dev_priv); > > + intel_runtime_pm_disable_interrupts(i915); > > + intel_hpd_cancel_work(i915); > > - intel_suspend_encoders(dev_priv); > > + intel_suspend_encoders(i915); > > - intel_suspend_hw(dev_priv); > > + intel_suspend_hw(i915); > > - i915_gem_suspend_gtt_mappings(dev_priv); > > + i915_gem_suspend_gtt_mappings(i915); > > - i915_save_state(dev_priv); > > + i915_save_state(i915); > > - opregion_target_state = suspend_to_idle(dev_priv) ? PCI_D1 : PCI_D3cold; > > - intel_opregion_suspend(dev_priv, opregion_target_state); > > + opregion_target_state = suspend_to_idle(i915) ? PCI_D1 : PCI_D3cold; > > + intel_opregion_suspend(i915, opregion_target_state); > > intel_fbdev_set_suspend(dev, FBINFO_STATE_SUSPENDED, true); > > - dev_priv->suspend_count++; > > + i915->suspend_count++; > > - intel_csr_ucode_suspend(dev_priv); > > + intel_csr_ucode_suspend(i915); > > - enable_rpm_wakeref_asserts(dev_priv); > > + enable_rpm_wakeref_asserts(i915); > > return 0; > > } > > static enum i915_drm_suspend_mode > > -get_suspend_mode(struct drm_i915_private *dev_priv, bool hibernate) > > +get_suspend_mode(struct drm_i915_private *i915, bool hibernate) > > { > > if (hibernate) > > return I915_DRM_SUSPEND_HIBERNATE; > > - if (suspend_to_idle(dev_priv)) > > + if (suspend_to_idle(i915)) > > return I915_DRM_SUSPEND_IDLE; > > return I915_DRM_SUSPEND_MEM; > > @@ -2134,30 +2134,30 @@ get_suspend_mode(struct drm_i915_private *dev_priv, bool hibernate) > > static int i915_drm_suspend_late(struct drm_device *dev, bool hibernation) > > { > > - struct drm_i915_private *dev_priv = to_i915(dev); > > - struct pci_dev *pdev = dev_priv->drm.pdev; > > + struct drm_i915_private *i915 = to_i915(dev); > > + struct pci_dev *pdev = i915->drm.pdev; > > int ret; > > - disable_rpm_wakeref_asserts(dev_priv); > > + disable_rpm_wakeref_asserts(i915); > > - i915_gem_suspend_late(dev_priv); > > + i915_gem_suspend_late(i915); > > - intel_uncore_suspend(&dev_priv->uncore); > > + intel_uncore_suspend(&i915->uncore); > > - intel_power_domains_suspend(dev_priv, > > - get_suspend_mode(dev_priv, hibernation)); > > + intel_power_domains_suspend(i915, > > + get_suspend_mode(i915, hibernation)); > > ret = 0; > > - if (INTEL_GEN(dev_priv) >= 11 || IS_GEN9_LP(dev_priv)) > > - bxt_enable_dc9(dev_priv); > > - else if (IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv)) > > - hsw_enable_pc8(dev_priv); > > - else if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) > > - ret = vlv_suspend_complete(dev_priv); > > + if (INTEL_GEN(i915) >= 11 || IS_GEN9_LP(i915)) > > + bxt_enable_dc9(i915); > > + else if (IS_HASWELL(i915) || IS_BROADWELL(i915)) > > + hsw_enable_pc8(i915); > > + else if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) > > + ret = vlv_suspend_complete(i915); > > if (ret) { > > DRM_ERROR("Suspend complete failed: %d\n", ret); > > - intel_power_domains_resume(dev_priv); > > + intel_power_domains_resume(i915); > > goto out; > > } > > @@ -2175,13 +2175,13 @@ static int i915_drm_suspend_late(struct drm_device *dev, bool hibernation) > > * Fujitsu FSC S7110 > > * Acer Aspire 1830T > > */ > > - if (!(hibernation && INTEL_GEN(dev_priv) < 6)) > > + if (!(hibernation && INTEL_GEN(i915) < 6)) > > pci_set_power_state(pdev, PCI_D3hot); > > out: > > - enable_rpm_wakeref_asserts(dev_priv); > > - if (!dev_priv->uncore.user_forcewake.count) > > - intel_runtime_pm_cleanup(dev_priv); > > + enable_rpm_wakeref_asserts(i915); > > + if (!i915->uncore.user_forcewake.count) > > + intel_runtime_pm_cleanup(i915); > > return ret; > > } > > @@ -2212,24 +2212,24 @@ static int i915_suspend_switcheroo(struct drm_device *dev, pm_message_t state) > > static int i915_drm_resume(struct drm_device *dev) > > { > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > int ret; > > - disable_rpm_wakeref_asserts(dev_priv); > > - intel_sanitize_gt_powersave(dev_priv); > > + disable_rpm_wakeref_asserts(i915); > > + intel_sanitize_gt_powersave(i915); > > - i915_gem_sanitize(dev_priv); > > + i915_gem_sanitize(i915); > > - ret = i915_ggtt_enable_hw(dev_priv); > > + ret = i915_ggtt_enable_hw(i915); > > if (ret) > > DRM_ERROR("failed to re-enable GGTT\n"); > > - intel_csr_ucode_resume(dev_priv); > > + intel_csr_ucode_resume(i915); > > - i915_restore_state(dev_priv); > > - intel_pps_unlock_regs_wa(dev_priv); > > + i915_restore_state(i915); > > + intel_pps_unlock_regs_wa(i915); > > - intel_init_pch_refclk(dev_priv); > > + intel_init_pch_refclk(i915); > > /* > > * Interrupts have to be enabled before any batches are run. If not the > > @@ -2241,21 +2241,21 @@ static int i915_drm_resume(struct drm_device *dev) > > * Modeset enabling in intel_modeset_init_hw() also needs working > > * interrupts. > > */ > > - intel_runtime_pm_enable_interrupts(dev_priv); > > + intel_runtime_pm_enable_interrupts(i915); > > drm_mode_config_reset(dev); > > - i915_gem_resume(dev_priv); > > + i915_gem_resume(i915); > > intel_modeset_init_hw(dev); > > - intel_init_clock_gating(dev_priv); > > + intel_init_clock_gating(i915); > > - spin_lock_irq(&dev_priv->irq_lock); > > - if (dev_priv->display.hpd_irq_setup) > > - dev_priv->display.hpd_irq_setup(dev_priv); > > - spin_unlock_irq(&dev_priv->irq_lock); > > + spin_lock_irq(&i915->irq_lock); > > + if (i915->display.hpd_irq_setup) > > + i915->display.hpd_irq_setup(i915); > > + spin_unlock_irq(&i915->irq_lock); > > - intel_dp_mst_resume(dev_priv); > > + intel_dp_mst_resume(i915); > > intel_display_resume(dev); > > @@ -2267,23 +2267,23 @@ static int i915_drm_resume(struct drm_device *dev) > > * bother with the tiny race here where we might lose hotplug > > * notifications. > > * */ > > - intel_hpd_init(dev_priv); > > + intel_hpd_init(i915); > > - intel_opregion_resume(dev_priv); > > + intel_opregion_resume(i915); > > intel_fbdev_set_suspend(dev, FBINFO_STATE_RUNNING, false); > > - intel_power_domains_enable(dev_priv); > > + intel_power_domains_enable(i915); > > - enable_rpm_wakeref_asserts(dev_priv); > > + enable_rpm_wakeref_asserts(i915); > > return 0; > > } > > static int i915_drm_resume_early(struct drm_device *dev) > > { > > - struct drm_i915_private *dev_priv = to_i915(dev); > > - struct pci_dev *pdev = dev_priv->drm.pdev; > > + struct drm_i915_private *i915 = to_i915(dev); > > + struct pci_dev *pdev = i915->drm.pdev; > > int ret; > > /* > > @@ -2330,32 +2330,32 @@ static int i915_drm_resume_early(struct drm_device *dev) > > pci_set_master(pdev); > > - disable_rpm_wakeref_asserts(dev_priv); > > + disable_rpm_wakeref_asserts(i915); > > - if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) > > - ret = vlv_resume_prepare(dev_priv, false); > > + if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) > > + ret = vlv_resume_prepare(i915, false); > > if (ret) > > DRM_ERROR("Resume prepare failed: %d, continuing anyway\n", > > ret); > > - intel_uncore_resume_early(&dev_priv->uncore); > > + intel_uncore_resume_early(&i915->uncore); > > - i915_check_and_clear_faults(dev_priv); > > + i915_check_and_clear_faults(i915); > > - if (INTEL_GEN(dev_priv) >= 11 || IS_GEN9_LP(dev_priv)) { > > - gen9_sanitize_dc_state(dev_priv); > > - bxt_disable_dc9(dev_priv); > > - } else if (IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv)) { > > - hsw_disable_pc8(dev_priv); > > + if (INTEL_GEN(i915) >= 11 || IS_GEN9_LP(i915)) { > > + gen9_sanitize_dc_state(i915); > > + bxt_disable_dc9(i915); > > + } else if (IS_HASWELL(i915) || IS_BROADWELL(i915)) { > > + hsw_disable_pc8(i915); > > } > > - intel_uncore_sanitize(dev_priv); > > + intel_uncore_sanitize(i915); > > - intel_power_domains_resume(dev_priv); > > + intel_power_domains_resume(i915); > > - intel_gt_sanitize(dev_priv, true); > > + intel_gt_sanitize(i915, true); > > - enable_rpm_wakeref_asserts(dev_priv); > > + enable_rpm_wakeref_asserts(i915); > > return ret; > > } > > @@ -2540,9 +2540,9 @@ static int i915_pm_restore(struct device *kdev) > > * a black-box for the driver. Further investigation is needed to reduce the > > * saved/restored registers even further, by following the same 3 criteria. > > */ > > -static void vlv_save_gunit_s0ix_state(struct drm_i915_private *dev_priv) > > +static void vlv_save_gunit_s0ix_state(struct drm_i915_private *i915) > > { > > - struct vlv_s0ix_state *s = &dev_priv->vlv_s0ix_state; > > + struct vlv_s0ix_state *s = &i915->vlv_s0ix_state; > > int i; > > /* GAM 0x4000-0x4770 */ > > @@ -2621,9 +2621,9 @@ static void vlv_save_gunit_s0ix_state(struct drm_i915_private *dev_priv) > > */ > > } > > -static void vlv_restore_gunit_s0ix_state(struct drm_i915_private *dev_priv) > > +static void vlv_restore_gunit_s0ix_state(struct drm_i915_private *i915) > > { > > - struct vlv_s0ix_state *s = &dev_priv->vlv_s0ix_state; > > + struct vlv_s0ix_state *s = &i915->vlv_s0ix_state; > > u32 val; > > int i; > > @@ -2732,7 +2732,7 @@ static int vlv_wait_for_pw_status(struct drm_i915_private *i915, > > return ret; > > } > > -int vlv_force_gfx_clock(struct drm_i915_private *dev_priv, bool force_on) > > +int vlv_force_gfx_clock(struct drm_i915_private *i915, bool force_on) > > { > > u32 val; > > int err; > > @@ -2746,7 +2746,7 @@ int vlv_force_gfx_clock(struct drm_i915_private *dev_priv, bool force_on) > > if (!force_on) > > return 0; > > - err = intel_wait_for_register(&dev_priv->uncore, > > + err = intel_wait_for_register(&i915->uncore, > > VLV_GTLC_SURVIVABILITY_REG, > > VLV_GFX_CLK_STATUS_BIT, > > VLV_GFX_CLK_STATUS_BIT, > > @@ -2758,7 +2758,7 @@ int vlv_force_gfx_clock(struct drm_i915_private *dev_priv, bool force_on) > > return err; > > } > > -static int vlv_allow_gt_wake(struct drm_i915_private *dev_priv, bool allow) > > +static int vlv_allow_gt_wake(struct drm_i915_private *i915, bool allow) > > { > > u32 mask; > > u32 val; > > @@ -2774,14 +2774,14 @@ static int vlv_allow_gt_wake(struct drm_i915_private *dev_priv, bool allow) > > mask = VLV_GTLC_ALLOWWAKEACK; > > val = allow ? mask : 0; > > - err = vlv_wait_for_pw_status(dev_priv, mask, val); > > + err = vlv_wait_for_pw_status(i915, mask, val); > > if (err) > > DRM_ERROR("timeout disabling GT waking\n"); > > return err; > > } > > -static void vlv_wait_for_gt_wells(struct drm_i915_private *dev_priv, > > +static void vlv_wait_for_gt_wells(struct drm_i915_private *i915, > > bool wait_for_on) > > { > > u32 mask; > > @@ -2797,12 +2797,12 @@ static void vlv_wait_for_gt_wells(struct drm_i915_private *dev_priv, > > * This can fail to turn off the rc6 if the GPU is stuck after a failed > > * reset and we are trying to force the machine to sleep. > > */ > > - if (vlv_wait_for_pw_status(dev_priv, mask, val)) > > + if (vlv_wait_for_pw_status(i915, mask, val)) > > DRM_DEBUG_DRIVER("timeout waiting for GT wells to go %s\n", > > onoff(wait_for_on)); > > } > > -static void vlv_check_no_gt_access(struct drm_i915_private *dev_priv) > > +static void vlv_check_no_gt_access(struct drm_i915_private *i915) > > { > > if (!(I915_READ(VLV_GTLC_PW_STATUS) & VLV_GTLC_ALLOWWAKEERR)) > > return; > > @@ -2811,7 +2811,7 @@ static void vlv_check_no_gt_access(struct drm_i915_private *dev_priv) > > I915_WRITE(VLV_GTLC_PW_STATUS, VLV_GTLC_ALLOWWAKEERR); > > } > > -static int vlv_suspend_complete(struct drm_i915_private *dev_priv) > > +static int vlv_suspend_complete(struct drm_i915_private *i915) > > { > > u32 mask; > > int err; > > @@ -2820,25 +2820,25 @@ static int vlv_suspend_complete(struct drm_i915_private *dev_priv) > > * Bspec defines the following GT well on flags as debug only, so > > * don't treat them as hard failures. > > */ > > - vlv_wait_for_gt_wells(dev_priv, false); > > + vlv_wait_for_gt_wells(i915, false); > > mask = VLV_GTLC_RENDER_CTX_EXISTS | VLV_GTLC_MEDIA_CTX_EXISTS; > > WARN_ON((I915_READ(VLV_GTLC_WAKE_CTRL) & mask) != mask); > > - vlv_check_no_gt_access(dev_priv); > > + vlv_check_no_gt_access(i915); > > - err = vlv_force_gfx_clock(dev_priv, true); > > + err = vlv_force_gfx_clock(i915, true); > > if (err) > > goto err1; > > - err = vlv_allow_gt_wake(dev_priv, false); > > + err = vlv_allow_gt_wake(i915, false); > > if (err) > > goto err2; > > - if (!IS_CHERRYVIEW(dev_priv)) > > - vlv_save_gunit_s0ix_state(dev_priv); > > + if (!IS_CHERRYVIEW(i915)) > > + vlv_save_gunit_s0ix_state(i915); > > - err = vlv_force_gfx_clock(dev_priv, false); > > + err = vlv_force_gfx_clock(i915, false); > > if (err) > > goto err2; > > @@ -2846,14 +2846,14 @@ static int vlv_suspend_complete(struct drm_i915_private *dev_priv) > > err2: > > /* For safety always re-enable waking and disable gfx clock forcing */ > > - vlv_allow_gt_wake(dev_priv, true); > > + vlv_allow_gt_wake(i915, true); > > err1: > > - vlv_force_gfx_clock(dev_priv, false); > > + vlv_force_gfx_clock(i915, false); > > return err; > > } > > -static int vlv_resume_prepare(struct drm_i915_private *dev_priv, > > +static int vlv_resume_prepare(struct drm_i915_private *i915, > > bool rpm_resume) > > { > > int err; > > @@ -2864,23 +2864,23 @@ static int vlv_resume_prepare(struct drm_i915_private *dev_priv, > > * can do at this point. Return the first error code (which will also > > * leave RPM permanently disabled). > > */ > > - ret = vlv_force_gfx_clock(dev_priv, true); > > + ret = vlv_force_gfx_clock(i915, true); > > - if (!IS_CHERRYVIEW(dev_priv)) > > - vlv_restore_gunit_s0ix_state(dev_priv); > > + if (!IS_CHERRYVIEW(i915)) > > + vlv_restore_gunit_s0ix_state(i915); > > - err = vlv_allow_gt_wake(dev_priv, true); > > + err = vlv_allow_gt_wake(i915, true); > > if (!ret) > > ret = err; > > - err = vlv_force_gfx_clock(dev_priv, false); > > + err = vlv_force_gfx_clock(i915, false); > > if (!ret) > > ret = err; > > - vlv_check_no_gt_access(dev_priv); > > + vlv_check_no_gt_access(i915); > > if (rpm_resume) > > - intel_init_clock_gating(dev_priv); > > + intel_init_clock_gating(i915); > > return ret; > > } > > @@ -2889,80 +2889,80 @@ static int intel_runtime_suspend(struct device *kdev) > > { > > struct pci_dev *pdev = to_pci_dev(kdev); > > struct drm_device *dev = pci_get_drvdata(pdev); > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > int ret; > > - if (WARN_ON_ONCE(!(dev_priv->gt_pm.rc6.enabled && HAS_RC6(dev_priv)))) > > + if (WARN_ON_ONCE(!(i915->gt_pm.rc6.enabled && HAS_RC6(i915)))) > > return -ENODEV; > > - if (WARN_ON_ONCE(!HAS_RUNTIME_PM(dev_priv))) > > + if (WARN_ON_ONCE(!HAS_RUNTIME_PM(i915))) > > return -ENODEV; > > DRM_DEBUG_KMS("Suspending device\n"); > > - disable_rpm_wakeref_asserts(dev_priv); > > + disable_rpm_wakeref_asserts(i915); > > /* > > * We are safe here against re-faults, since the fault handler takes > > * an RPM reference. > > */ > > - i915_gem_runtime_suspend(dev_priv); > > + i915_gem_runtime_suspend(i915); > > - intel_uc_runtime_suspend(dev_priv); > > + intel_uc_runtime_suspend(i915); > > - intel_runtime_pm_disable_interrupts(dev_priv); > > + intel_runtime_pm_disable_interrupts(i915); > > - intel_uncore_suspend(&dev_priv->uncore); > > + intel_uncore_suspend(&i915->uncore); > > ret = 0; > > - if (INTEL_GEN(dev_priv) >= 11) { > > - icl_display_core_uninit(dev_priv); > > - bxt_enable_dc9(dev_priv); > > - } else if (IS_GEN9_LP(dev_priv)) { > > - bxt_display_core_uninit(dev_priv); > > - bxt_enable_dc9(dev_priv); > > - } else if (IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv)) { > > - hsw_enable_pc8(dev_priv); > > - } else if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) { > > - ret = vlv_suspend_complete(dev_priv); > > + if (INTEL_GEN(i915) >= 11) { > > + icl_display_core_uninit(i915); > > + bxt_enable_dc9(i915); > > + } else if (IS_GEN9_LP(i915)) { > > + bxt_display_core_uninit(i915); > > + bxt_enable_dc9(i915); > > + } else if (IS_HASWELL(i915) || IS_BROADWELL(i915)) { > > + hsw_enable_pc8(i915); > > + } else if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) { > > + ret = vlv_suspend_complete(i915); > > } > > if (ret) { > > DRM_ERROR("Runtime suspend failed, disabling it (%d)\n", ret); > > - intel_uncore_runtime_resume(&dev_priv->uncore); > > + intel_uncore_runtime_resume(&i915->uncore); > > - intel_runtime_pm_enable_interrupts(dev_priv); > > + intel_runtime_pm_enable_interrupts(i915); > > - intel_uc_resume(dev_priv); > > + intel_uc_resume(i915); > > - i915_gem_init_swizzling(dev_priv); > > - i915_gem_restore_fences(dev_priv); > > + i915_gem_init_swizzling(i915); > > + i915_gem_restore_fences(i915); > > - enable_rpm_wakeref_asserts(dev_priv); > > + enable_rpm_wakeref_asserts(i915); > > return ret; > > } > > - enable_rpm_wakeref_asserts(dev_priv); > > - intel_runtime_pm_cleanup(dev_priv); > > + enable_rpm_wakeref_asserts(i915); > > + intel_runtime_pm_cleanup(i915); > > - if (intel_uncore_arm_unclaimed_mmio_detection(&dev_priv->uncore)) > > + if (intel_uncore_arm_unclaimed_mmio_detection(&i915->uncore)) > > DRM_ERROR("Unclaimed access detected prior to suspending\n"); > > - dev_priv->runtime_pm.suspended = true; > > + i915->runtime_pm.suspended = true; > > /* > > * FIXME: We really should find a document that references the arguments > > * used below! > > */ > > - if (IS_BROADWELL(dev_priv)) { > > + if (IS_BROADWELL(i915)) { > > /* > > * On Broadwell, if we use PCI_D1 the PCH DDI ports will stop > > * being detected, and the call we do at intel_runtime_resume() > > * won't be able to restore them. Since PCI_D3hot matches the > > * actual specification and appears to be working, use it. > > */ > > - intel_opregion_notify_adapter(dev_priv, PCI_D3hot); > > + intel_opregion_notify_adapter(i915, PCI_D3hot); > > } else { > > /* > > * current versions of firmware which depend on this opregion > > @@ -2971,13 +2971,13 @@ static int intel_runtime_suspend(struct device *kdev) > > * to distinguish it from notifications that might be sent via > > * the suspend path. > > */ > > - intel_opregion_notify_adapter(dev_priv, PCI_D1); > > + intel_opregion_notify_adapter(i915, PCI_D1); > > } > > - assert_forcewakes_inactive(&dev_priv->uncore); > > + assert_forcewakes_inactive(&i915->uncore); > > - if (!IS_VALLEYVIEW(dev_priv) && !IS_CHERRYVIEW(dev_priv)) > > - intel_hpd_poll_init(dev_priv); > > + if (!IS_VALLEYVIEW(i915) && !IS_CHERRYVIEW(i915)) > > + intel_hpd_poll_init(i915); > > DRM_DEBUG_KMS("Device suspended\n"); > > return 0; > > @@ -2987,69 +2987,69 @@ static int intel_runtime_resume(struct device *kdev) > > { > > struct pci_dev *pdev = to_pci_dev(kdev); > > struct drm_device *dev = pci_get_drvdata(pdev); > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > int ret = 0; > > - if (WARN_ON_ONCE(!HAS_RUNTIME_PM(dev_priv))) > > + if (WARN_ON_ONCE(!HAS_RUNTIME_PM(i915))) > > return -ENODEV; > > DRM_DEBUG_KMS("Resuming device\n"); > > - WARN_ON_ONCE(atomic_read(&dev_priv->runtime_pm.wakeref_count)); > > - disable_rpm_wakeref_asserts(dev_priv); > > + WARN_ON_ONCE(atomic_read(&i915->runtime_pm.wakeref_count)); > > + disable_rpm_wakeref_asserts(i915); > > - intel_opregion_notify_adapter(dev_priv, PCI_D0); > > - dev_priv->runtime_pm.suspended = false; > > - if (intel_uncore_unclaimed_mmio(&dev_priv->uncore)) > > + intel_opregion_notify_adapter(i915, PCI_D0); > > + i915->runtime_pm.suspended = false; > > + if (intel_uncore_unclaimed_mmio(&i915->uncore)) > > DRM_DEBUG_DRIVER("Unclaimed access during suspend, bios?\n"); > > - if (INTEL_GEN(dev_priv) >= 11) { > > - bxt_disable_dc9(dev_priv); > > - icl_display_core_init(dev_priv, true); > > - if (dev_priv->csr.dmc_payload) { > > - if (dev_priv->csr.allowed_dc_mask & > > + if (INTEL_GEN(i915) >= 11) { > > + bxt_disable_dc9(i915); > > + icl_display_core_init(i915, true); > > + if (i915->csr.dmc_payload) { > > + if (i915->csr.allowed_dc_mask & > > DC_STATE_EN_UPTO_DC6) > > - skl_enable_dc6(dev_priv); > > - else if (dev_priv->csr.allowed_dc_mask & > > + skl_enable_dc6(i915); > > + else if (i915->csr.allowed_dc_mask & > > DC_STATE_EN_UPTO_DC5) > > - gen9_enable_dc5(dev_priv); > > + gen9_enable_dc5(i915); > > } > > - } else if (IS_GEN9_LP(dev_priv)) { > > - bxt_disable_dc9(dev_priv); > > - bxt_display_core_init(dev_priv, true); > > - if (dev_priv->csr.dmc_payload && > > - (dev_priv->csr.allowed_dc_mask & DC_STATE_EN_UPTO_DC5)) > > - gen9_enable_dc5(dev_priv); > > - } else if (IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv)) { > > - hsw_disable_pc8(dev_priv); > > - } else if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) { > > - ret = vlv_resume_prepare(dev_priv, true); > > + } else if (IS_GEN9_LP(i915)) { > > + bxt_disable_dc9(i915); > > + bxt_display_core_init(i915, true); > > + if (i915->csr.dmc_payload && > > + (i915->csr.allowed_dc_mask & DC_STATE_EN_UPTO_DC5)) > > + gen9_enable_dc5(i915); > > + } else if (IS_HASWELL(i915) || IS_BROADWELL(i915)) { > > + hsw_disable_pc8(i915); > > + } else if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) { > > + ret = vlv_resume_prepare(i915, true); > > } > > - intel_uncore_runtime_resume(&dev_priv->uncore); > > + intel_uncore_runtime_resume(&i915->uncore); > > - intel_runtime_pm_enable_interrupts(dev_priv); > > + intel_runtime_pm_enable_interrupts(i915); > > - intel_uc_resume(dev_priv); > > + intel_uc_resume(i915); > > /* > > * No point of rolling back things in case of an error, as the best > > * we can do is to hope that things will still work (and disable RPM). > > */ > > - i915_gem_init_swizzling(dev_priv); > > - i915_gem_restore_fences(dev_priv); > > + i915_gem_init_swizzling(i915); > > + i915_gem_restore_fences(i915); > > /* > > * On VLV/CHV display interrupts are part of the display > > * power well, so hpd is reinitialized from there. For > > * everyone else do it here. > > */ > > - if (!IS_VALLEYVIEW(dev_priv) && !IS_CHERRYVIEW(dev_priv)) > > - intel_hpd_init(dev_priv); > > + if (!IS_VALLEYVIEW(i915) && !IS_CHERRYVIEW(i915)) > > + intel_hpd_init(i915); > > - intel_enable_ipc(dev_priv); > > + intel_enable_ipc(i915); > > - enable_rpm_wakeref_asserts(dev_priv); > > + enable_rpm_wakeref_asserts(i915); > > if (ret) > > DRM_ERROR("Runtime resume failed, disabling it (%d)\n", ret); > > diff --git a/drivers/gpu/drm/i915/i915_drv.h b/drivers/gpu/drm/i915/i915_drv.h > > index eb98d285d290..a1adb79ca2c3 100644 > > --- a/drivers/gpu/drm/i915/i915_drv.h > > +++ b/drivers/gpu/drm/i915/i915_drv.h > > @@ -208,7 +208,7 @@ struct i915_mm_struct; > > struct i915_mmu_object; > > struct drm_i915_file_private { > > - struct drm_i915_private *dev_priv; > > + struct drm_i915_private *i915; > > struct drm_file *file; > > struct { > > @@ -278,12 +278,12 @@ struct dpll; > > struct intel_cdclk_state; > > struct drm_i915_display_funcs { > > - void (*get_cdclk)(struct drm_i915_private *dev_priv, > > + void (*get_cdclk)(struct drm_i915_private *i915, > > struct intel_cdclk_state *cdclk_state); > > - void (*set_cdclk)(struct drm_i915_private *dev_priv, > > + void (*set_cdclk)(struct drm_i915_private *i915, > > const struct intel_cdclk_state *cdclk_state, > > enum pipe pipe); > > - int (*get_fifo_size)(struct drm_i915_private *dev_priv, > > + int (*get_fifo_size)(struct drm_i915_private *i915, > > enum i9xx_plane_id i9xx_plane); > > int (*compute_pipe_wm)(struct intel_crtc_state *cstate); > > int (*compute_intermediate_wm)(struct intel_crtc_state *newstate); > > @@ -317,8 +317,8 @@ struct drm_i915_display_funcs { > > const struct drm_connector_state *old_conn_state); > > void (*fdi_link_train)(struct intel_crtc *crtc, > > const struct intel_crtc_state *crtc_state); > > - void (*init_clock_gating)(struct drm_i915_private *dev_priv); > > - void (*hpd_irq_setup)(struct drm_i915_private *dev_priv); > > + void (*init_clock_gating)(struct drm_i915_private *i915); > > + void (*hpd_irq_setup)(struct drm_i915_private *i915); > > /* clock updates for mode set */ > > /* cursor updates */ > > /* render clock increase/decrease */ > > @@ -552,7 +552,7 @@ struct intel_gmbus { > > u32 reg0; > > i915_reg_t gpio_reg; > > struct i2c_algo_bit_data bit_algo; > > - struct drm_i915_private *dev_priv; > > + struct drm_i915_private *i915; > > }; > > struct i915_suspend_saved_registers { > > @@ -641,7 +641,7 @@ struct intel_rps { > > /* > > * work, interrupts_enabled and pm_iir are protected by > > - * dev_priv->irq_lock > > + * i915->irq_lock > > */ > > struct work_struct work; > > bool interrupts_enabled; > > @@ -1240,9 +1240,9 @@ struct i915_perf_stream_ops { > > */ > > struct i915_perf_stream { > > /** > > - * @dev_priv: i915 drm device > > + * @i915: i915 drm device > > */ > > - struct drm_i915_private *dev_priv; > > + struct drm_i915_private *i915; > > /** > > * @link: Links the stream into ``&drm_i915_private->streams`` > > @@ -1302,20 +1302,20 @@ struct i915_oa_ops { > > * @is_valid_b_counter_reg: Validates register's address for > > * programming boolean counters for a particular platform. > > */ > > - bool (*is_valid_b_counter_reg)(struct drm_i915_private *dev_priv, > > + bool (*is_valid_b_counter_reg)(struct drm_i915_private *i915, > > u32 addr); > > /** > > * @is_valid_mux_reg: Validates register's address for programming mux > > * for a particular platform. > > */ > > - bool (*is_valid_mux_reg)(struct drm_i915_private *dev_priv, u32 addr); > > + bool (*is_valid_mux_reg)(struct drm_i915_private *i915, u32 addr); > > /** > > * @is_valid_flex_reg: Validates register's address for programming > > * flex EU filtering for a particular platform. > > */ > > - bool (*is_valid_flex_reg)(struct drm_i915_private *dev_priv, u32 addr); > > + bool (*is_valid_flex_reg)(struct drm_i915_private *i915, u32 addr); > > /** > > * @enable_metric_set: Selects and applies any MUX configuration to set > > @@ -1329,7 +1329,7 @@ struct i915_oa_ops { > > * @disable_metric_set: Remove system constraints associated with using > > * the OA unit. > > */ > > - void (*disable_metric_set)(struct drm_i915_private *dev_priv); > > + void (*disable_metric_set)(struct drm_i915_private *i915); > > /** > > * @oa_enable: Enable periodic sampling > > @@ -1357,7 +1357,7 @@ struct i915_oa_ops { > > * handling the OA unit tail pointer race that affects multiple > > * generations. > > */ > > - u32 (*oa_hw_tail_read)(struct drm_i915_private *dev_priv); > > + u32 (*oa_hw_tail_read)(struct drm_i915_private *i915); > > }; > > struct intel_cdclk_state { > > @@ -1750,13 +1750,13 @@ struct drm_i915_private { > > /* > > * Lock associated with adding/modifying/removing OA configs > > - * in dev_priv->perf.metrics_idr. > > + * in i915->perf.metrics_idr. > > */ > > struct mutex metrics_lock; > > /* > > * List of dynamic configurations, you need to hold > > - * dev_priv->perf.metrics_lock to access it. > > + * i915->perf.metrics_lock to access it. > > */ > > struct idr metrics_idr; > > @@ -1772,7 +1772,7 @@ struct drm_i915_private { > > * The stream currently using the OA unit. If accessed > > * outside a syscall associated to its file > > * descriptor, you need to hold > > - * dev_priv->drm.struct_mutex. > > + * i915->drm.struct_mutex. > > */ > > struct i915_perf_stream *exclusive_stream; > > @@ -2066,15 +2066,15 @@ enum hdmi_force_audio { > > GENMASK(INTEL_FRONTBUFFER_BITS_PER_PIPE * ((pipe) + 1) - 1, \ > > INTEL_FRONTBUFFER_BITS_PER_PIPE * (pipe)) > > -#define INTEL_INFO(dev_priv) (&(dev_priv)->__info) > > -#define RUNTIME_INFO(dev_priv) (&(dev_priv)->__runtime) > > -#define DRIVER_CAPS(dev_priv) (&(dev_priv)->caps) > > +#define INTEL_INFO(i915) (&(i915)->__info) > > +#define RUNTIME_INFO(i915) (&(i915)->__runtime) > > +#define DRIVER_CAPS(i915) (&(i915)->caps) > > -#define INTEL_GEN(dev_priv) (INTEL_INFO(dev_priv)->gen) > > -#define INTEL_DEVID(dev_priv) (RUNTIME_INFO(dev_priv)->device_id) > > +#define INTEL_GEN(i915) (INTEL_INFO(i915)->gen) > > +#define INTEL_DEVID(i915) (RUNTIME_INFO(i915)->device_id) > > #define REVID_FOREVER 0xff > > -#define INTEL_REVID(dev_priv) ((dev_priv)->drm.pdev->revision) > > +#define INTEL_REVID(i915) ((i915)->drm.pdev->revision) > > #define INTEL_GEN_MASK(s, e) ( \ > > BUILD_BUG_ON_ZERO(!__builtin_constant_p(s)) + \ > > @@ -2082,12 +2082,12 @@ enum hdmi_force_audio { > > GENMASK((e) - 1, (s) - 1)) > > /* Returns true if Gen is in inclusive range [Start, End] */ > > -#define IS_GEN_RANGE(dev_priv, s, e) \ > > - (!!(INTEL_INFO(dev_priv)->gen_mask & INTEL_GEN_MASK((s), (e)))) > > +#define IS_GEN_RANGE(i915, s, e) \ > > + (!!(INTEL_INFO(i915)->gen_mask & INTEL_GEN_MASK((s), (e)))) > > -#define IS_GEN(dev_priv, n) \ > > +#define IS_GEN(i915, n) \ > > (BUILD_BUG_ON_ZERO(!__builtin_constant_p(n)) + \ > > - INTEL_INFO(dev_priv)->gen == (n)) > > + INTEL_INFO(i915)->gen == (n)) > > /* > > * Return true if revision is in range [since,until] inclusive. > > @@ -2159,88 +2159,88 @@ IS_SUBPLATFORM(const struct drm_i915_private *i915, > > return ((mask << (msb - pb)) & (mask << (msb - s))) & BIT(msb); > > } > > -#define IS_MOBILE(dev_priv) (INTEL_INFO(dev_priv)->is_mobile) > > - > > -#define IS_I830(dev_priv) IS_PLATFORM(dev_priv, INTEL_I830) > > -#define IS_I845G(dev_priv) IS_PLATFORM(dev_priv, INTEL_I845G) > > -#define IS_I85X(dev_priv) IS_PLATFORM(dev_priv, INTEL_I85X) > > -#define IS_I865G(dev_priv) IS_PLATFORM(dev_priv, INTEL_I865G) > > -#define IS_I915G(dev_priv) IS_PLATFORM(dev_priv, INTEL_I915G) > > -#define IS_I915GM(dev_priv) IS_PLATFORM(dev_priv, INTEL_I915GM) > > -#define IS_I945G(dev_priv) IS_PLATFORM(dev_priv, INTEL_I945G) > > -#define IS_I945GM(dev_priv) IS_PLATFORM(dev_priv, INTEL_I945GM) > > -#define IS_I965G(dev_priv) IS_PLATFORM(dev_priv, INTEL_I965G) > > -#define IS_I965GM(dev_priv) IS_PLATFORM(dev_priv, INTEL_I965GM) > > -#define IS_G45(dev_priv) IS_PLATFORM(dev_priv, INTEL_G45) > > -#define IS_GM45(dev_priv) IS_PLATFORM(dev_priv, INTEL_GM45) > > -#define IS_G4X(dev_priv) (IS_G45(dev_priv) || IS_GM45(dev_priv)) > > -#define IS_PINEVIEW(dev_priv) IS_PLATFORM(dev_priv, INTEL_PINEVIEW) > > -#define IS_G33(dev_priv) IS_PLATFORM(dev_priv, INTEL_G33) > > -#define IS_IRONLAKE(dev_priv) IS_PLATFORM(dev_priv, INTEL_IRONLAKE) > > -#define IS_IRONLAKE_M(dev_priv) \ > > - (IS_PLATFORM(dev_priv, INTEL_IRONLAKE) && IS_MOBILE(dev_priv)) > > -#define IS_IVYBRIDGE(dev_priv) IS_PLATFORM(dev_priv, INTEL_IVYBRIDGE) > > -#define IS_IVB_GT1(dev_priv) (IS_IVYBRIDGE(dev_priv) && \ > > - INTEL_INFO(dev_priv)->gt == 1) > > -#define IS_VALLEYVIEW(dev_priv) IS_PLATFORM(dev_priv, INTEL_VALLEYVIEW) > > -#define IS_CHERRYVIEW(dev_priv) IS_PLATFORM(dev_priv, INTEL_CHERRYVIEW) > > -#define IS_HASWELL(dev_priv) IS_PLATFORM(dev_priv, INTEL_HASWELL) > > -#define IS_BROADWELL(dev_priv) IS_PLATFORM(dev_priv, INTEL_BROADWELL) > > -#define IS_SKYLAKE(dev_priv) IS_PLATFORM(dev_priv, INTEL_SKYLAKE) > > -#define IS_BROXTON(dev_priv) IS_PLATFORM(dev_priv, INTEL_BROXTON) > > -#define IS_KABYLAKE(dev_priv) IS_PLATFORM(dev_priv, INTEL_KABYLAKE) > > -#define IS_GEMINILAKE(dev_priv) IS_PLATFORM(dev_priv, INTEL_GEMINILAKE) > > -#define IS_COFFEELAKE(dev_priv) IS_PLATFORM(dev_priv, INTEL_COFFEELAKE) > > -#define IS_CANNONLAKE(dev_priv) IS_PLATFORM(dev_priv, INTEL_CANNONLAKE) > > -#define IS_ICELAKE(dev_priv) IS_PLATFORM(dev_priv, INTEL_ICELAKE) > > -#define IS_ELKHARTLAKE(dev_priv) IS_PLATFORM(dev_priv, INTEL_ELKHARTLAKE) > > -#define IS_HSW_EARLY_SDV(dev_priv) (IS_HASWELL(dev_priv) && \ > > - (INTEL_DEVID(dev_priv) & 0xFF00) == 0x0C00) > > -#define IS_BDW_ULT(dev_priv) \ > > - IS_SUBPLATFORM(dev_priv, INTEL_BROADWELL, INTEL_SUBPLATFORM_ULT) > > -#define IS_BDW_ULX(dev_priv) \ > > - IS_SUBPLATFORM(dev_priv, INTEL_BROADWELL, INTEL_SUBPLATFORM_ULX) > > -#define IS_BDW_GT3(dev_priv) (IS_BROADWELL(dev_priv) && \ > > - INTEL_INFO(dev_priv)->gt == 3) > > -#define IS_HSW_ULT(dev_priv) \ > > - IS_SUBPLATFORM(dev_priv, INTEL_HASWELL, INTEL_SUBPLATFORM_ULT) > > -#define IS_HSW_GT3(dev_priv) (IS_HASWELL(dev_priv) && \ > > - INTEL_INFO(dev_priv)->gt == 3) > > -#define IS_HSW_GT1(dev_priv) (IS_HASWELL(dev_priv) && \ > > - INTEL_INFO(dev_priv)->gt == 1) > > +#define IS_MOBILE(i915) (INTEL_INFO(i915)->is_mobile) > > + > > +#define IS_I830(i915) IS_PLATFORM(i915, INTEL_I830) > > +#define IS_I845G(i915) IS_PLATFORM(i915, INTEL_I845G) > > +#define IS_I85X(i915) IS_PLATFORM(i915, INTEL_I85X) > > +#define IS_I865G(i915) IS_PLATFORM(i915, INTEL_I865G) > > +#define IS_I915G(i915) IS_PLATFORM(i915, INTEL_I915G) > > +#define IS_I915GM(i915) IS_PLATFORM(i915, INTEL_I915GM) > > +#define IS_I945G(i915) IS_PLATFORM(i915, INTEL_I945G) > > +#define IS_I945GM(i915) IS_PLATFORM(i915, INTEL_I945GM) > > +#define IS_I965G(i915) IS_PLATFORM(i915, INTEL_I965G) > > +#define IS_I965GM(i915) IS_PLATFORM(i915, INTEL_I965GM) > > +#define IS_G45(i915) IS_PLATFORM(i915, INTEL_G45) > > +#define IS_GM45(i915) IS_PLATFORM(i915, INTEL_GM45) > > +#define IS_G4X(i915) (IS_G45(i915) || IS_GM45(i915)) > > +#define IS_PINEVIEW(i915) IS_PLATFORM(i915, INTEL_PINEVIEW) > > +#define IS_G33(i915) IS_PLATFORM(i915, INTEL_G33) > > +#define IS_IRONLAKE(i915) IS_PLATFORM(i915, INTEL_IRONLAKE) > > +#define IS_IRONLAKE_M(i915) \ > > + (IS_PLATFORM(i915, INTEL_IRONLAKE) && IS_MOBILE(i915)) > > +#define IS_IVYBRIDGE(i915) IS_PLATFORM(i915, INTEL_IVYBRIDGE) > > +#define IS_IVB_GT1(i915) (IS_IVYBRIDGE(i915) && \ > > + INTEL_INFO(i915)->gt == 1) > > +#define IS_VALLEYVIEW(i915) IS_PLATFORM(i915, INTEL_VALLEYVIEW) > > +#define IS_CHERRYVIEW(i915) IS_PLATFORM(i915, INTEL_CHERRYVIEW) > > +#define IS_HASWELL(i915) IS_PLATFORM(i915, INTEL_HASWELL) > > +#define IS_BROADWELL(i915) IS_PLATFORM(i915, INTEL_BROADWELL) > > +#define IS_SKYLAKE(i915) IS_PLATFORM(i915, INTEL_SKYLAKE) > > +#define IS_BROXTON(i915) IS_PLATFORM(i915, INTEL_BROXTON) > > +#define IS_KABYLAKE(i915) IS_PLATFORM(i915, INTEL_KABYLAKE) > > +#define IS_GEMINILAKE(i915) IS_PLATFORM(i915, INTEL_GEMINILAKE) > > +#define IS_COFFEELAKE(i915) IS_PLATFORM(i915, INTEL_COFFEELAKE) > > +#define IS_CANNONLAKE(i915) IS_PLATFORM(i915, INTEL_CANNONLAKE) > > +#define IS_ICELAKE(i915) IS_PLATFORM(i915, INTEL_ICELAKE) > > +#define IS_ELKHARTLAKE(i915) IS_PLATFORM(i915, INTEL_ELKHARTLAKE) > > +#define IS_HSW_EARLY_SDV(i915) (IS_HASWELL(i915) && \ > > + (INTEL_DEVID(i915) & 0xFF00) == 0x0C00) > > +#define IS_BDW_ULT(i915) \ > > + IS_SUBPLATFORM(i915, INTEL_BROADWELL, INTEL_SUBPLATFORM_ULT) > > +#define IS_BDW_ULX(i915) \ > > + IS_SUBPLATFORM(i915, INTEL_BROADWELL, INTEL_SUBPLATFORM_ULX) > > +#define IS_BDW_GT3(i915) (IS_BROADWELL(i915) && \ > > + INTEL_INFO(i915)->gt == 3) > > +#define IS_HSW_ULT(i915) \ > > + IS_SUBPLATFORM(i915, INTEL_HASWELL, INTEL_SUBPLATFORM_ULT) > > +#define IS_HSW_GT3(i915) (IS_HASWELL(i915) && \ > > + INTEL_INFO(i915)->gt == 3) > > +#define IS_HSW_GT1(i915) (IS_HASWELL(i915) && \ > > + INTEL_INFO(i915)->gt == 1) > > /* ULX machines are also considered ULT. */ > > -#define IS_HSW_ULX(dev_priv) \ > > - IS_SUBPLATFORM(dev_priv, INTEL_HASWELL, INTEL_SUBPLATFORM_ULX) > > -#define IS_SKL_ULT(dev_priv) \ > > - IS_SUBPLATFORM(dev_priv, INTEL_SKYLAKE, INTEL_SUBPLATFORM_ULT) > > -#define IS_SKL_ULX(dev_priv) \ > > - IS_SUBPLATFORM(dev_priv, INTEL_SKYLAKE, INTEL_SUBPLATFORM_ULX) > > -#define IS_KBL_ULT(dev_priv) \ > > - IS_SUBPLATFORM(dev_priv, INTEL_KABYLAKE, INTEL_SUBPLATFORM_ULT) > > -#define IS_KBL_ULX(dev_priv) \ > > - IS_SUBPLATFORM(dev_priv, INTEL_KABYLAKE, INTEL_SUBPLATFORM_ULX) > > -#define IS_SKL_GT2(dev_priv) (IS_SKYLAKE(dev_priv) && \ > > - INTEL_INFO(dev_priv)->gt == 2) > > -#define IS_SKL_GT3(dev_priv) (IS_SKYLAKE(dev_priv) && \ > > - INTEL_INFO(dev_priv)->gt == 3) > > -#define IS_SKL_GT4(dev_priv) (IS_SKYLAKE(dev_priv) && \ > > - INTEL_INFO(dev_priv)->gt == 4) > > -#define IS_KBL_GT2(dev_priv) (IS_KABYLAKE(dev_priv) && \ > > - INTEL_INFO(dev_priv)->gt == 2) > > -#define IS_KBL_GT3(dev_priv) (IS_KABYLAKE(dev_priv) && \ > > - INTEL_INFO(dev_priv)->gt == 3) > > -#define IS_CFL_ULT(dev_priv) \ > > - IS_SUBPLATFORM(dev_priv, INTEL_COFFEELAKE, INTEL_SUBPLATFORM_ULT) > > -#define IS_CFL_ULX(dev_priv) \ > > - IS_SUBPLATFORM(dev_priv, INTEL_COFFEELAKE, INTEL_SUBPLATFORM_ULX) > > -#define IS_CFL_GT2(dev_priv) (IS_COFFEELAKE(dev_priv) && \ > > - INTEL_INFO(dev_priv)->gt == 2) > > -#define IS_CFL_GT3(dev_priv) (IS_COFFEELAKE(dev_priv) && \ > > - INTEL_INFO(dev_priv)->gt == 3) > > -#define IS_CNL_WITH_PORT_F(dev_priv) \ > > - IS_SUBPLATFORM(dev_priv, INTEL_CANNONLAKE, INTEL_SUBPLATFORM_PORTF) > > -#define IS_ICL_WITH_PORT_F(dev_priv) \ > > - IS_SUBPLATFORM(dev_priv, INTEL_ICELAKE, INTEL_SUBPLATFORM_PORTF) > > +#define IS_HSW_ULX(i915) \ > > + IS_SUBPLATFORM(i915, INTEL_HASWELL, INTEL_SUBPLATFORM_ULX) > > +#define IS_SKL_ULT(i915) \ > > + IS_SUBPLATFORM(i915, INTEL_SKYLAKE, INTEL_SUBPLATFORM_ULT) > > +#define IS_SKL_ULX(i915) \ > > + IS_SUBPLATFORM(i915, INTEL_SKYLAKE, INTEL_SUBPLATFORM_ULX) > > +#define IS_KBL_ULT(i915) \ > > + IS_SUBPLATFORM(i915, INTEL_KABYLAKE, INTEL_SUBPLATFORM_ULT) > > +#define IS_KBL_ULX(i915) \ > > + IS_SUBPLATFORM(i915, INTEL_KABYLAKE, INTEL_SUBPLATFORM_ULX) > > +#define IS_SKL_GT2(i915) (IS_SKYLAKE(i915) && \ > > + INTEL_INFO(i915)->gt == 2) > > +#define IS_SKL_GT3(i915) (IS_SKYLAKE(i915) && \ > > + INTEL_INFO(i915)->gt == 3) > > +#define IS_SKL_GT4(i915) (IS_SKYLAKE(i915) && \ > > + INTEL_INFO(i915)->gt == 4) > > +#define IS_KBL_GT2(i915) (IS_KABYLAKE(i915) && \ > > + INTEL_INFO(i915)->gt == 2) > > +#define IS_KBL_GT3(i915) (IS_KABYLAKE(i915) && \ > > + INTEL_INFO(i915)->gt == 3) > > +#define IS_CFL_ULT(i915) \ > > + IS_SUBPLATFORM(i915, INTEL_COFFEELAKE, INTEL_SUBPLATFORM_ULT) > > +#define IS_CFL_ULX(i915) \ > > + IS_SUBPLATFORM(i915, INTEL_COFFEELAKE, INTEL_SUBPLATFORM_ULX) > > +#define IS_CFL_GT2(i915) (IS_COFFEELAKE(i915) && \ > > + INTEL_INFO(i915)->gt == 2) > > +#define IS_CFL_GT3(i915) (IS_COFFEELAKE(i915) && \ > > + INTEL_INFO(i915)->gt == 3) > > +#define IS_CNL_WITH_PORT_F(i915) \ > > + IS_SUBPLATFORM(i915, INTEL_CANNONLAKE, INTEL_SUBPLATFORM_PORTF) > > +#define IS_ICL_WITH_PORT_F(i915) \ > > + IS_SUBPLATFORM(i915, INTEL_ICELAKE, INTEL_SUBPLATFORM_PORTF) > > #define SKL_REVID_A0 0x0 > > #define SKL_REVID_B0 0x1 > > @@ -2259,8 +2259,8 @@ IS_SUBPLATFORM(const struct drm_i915_private *i915, > > #define BXT_REVID_B_LAST 0x8 > > #define BXT_REVID_C0 0x9 > > -#define IS_BXT_REVID(dev_priv, since, until) \ > > - (IS_BROXTON(dev_priv) && IS_REVID(dev_priv, since, until)) > > +#define IS_BXT_REVID(i915, since, until) \ > > + (IS_BROXTON(i915) && IS_REVID(i915, since, until)) > > #define KBL_REVID_A0 0x0 > > #define KBL_REVID_B0 0x1 > > @@ -2268,14 +2268,14 @@ IS_SUBPLATFORM(const struct drm_i915_private *i915, > > #define KBL_REVID_D0 0x3 > > #define KBL_REVID_E0 0x4 > > -#define IS_KBL_REVID(dev_priv, since, until) \ > > - (IS_KABYLAKE(dev_priv) && IS_REVID(dev_priv, since, until)) > > +#define IS_KBL_REVID(i915, since, until) \ > > + (IS_KABYLAKE(i915) && IS_REVID(i915, since, until)) > > #define GLK_REVID_A0 0x0 > > #define GLK_REVID_A1 0x1 > > -#define IS_GLK_REVID(dev_priv, since, until) \ > > - (IS_GEMINILAKE(dev_priv) && IS_REVID(dev_priv, since, until)) > > +#define IS_GLK_REVID(i915, since, until) \ > > + (IS_GEMINILAKE(i915) && IS_REVID(i915, since, until)) > > #define CNL_REVID_A0 0x0 > > #define CNL_REVID_B0 0x1 > > @@ -2293,122 +2293,122 @@ IS_SUBPLATFORM(const struct drm_i915_private *i915, > > #define IS_ICL_REVID(p, since, until) \ > > (IS_ICELAKE(p) && IS_REVID(p, since, until)) > > -#define IS_LP(dev_priv) (INTEL_INFO(dev_priv)->is_lp) > > -#define IS_GEN9_LP(dev_priv) (IS_GEN(dev_priv, 9) && IS_LP(dev_priv)) > > -#define IS_GEN9_BC(dev_priv) (IS_GEN(dev_priv, 9) && !IS_LP(dev_priv)) > > +#define IS_LP(i915) (INTEL_INFO(i915)->is_lp) > > +#define IS_GEN9_LP(i915) (IS_GEN(i915, 9) && IS_LP(i915)) > > +#define IS_GEN9_BC(i915) (IS_GEN(i915, 9) && !IS_LP(i915)) > > -#define HAS_ENGINE(dev_priv, id) (INTEL_INFO(dev_priv)->engine_mask & BIT(id)) > > +#define HAS_ENGINE(i915, id) (INTEL_INFO(i915)->engine_mask & BIT(id)) > > -#define ENGINE_INSTANCES_MASK(dev_priv, first, count) ({ \ > > +#define ENGINE_INSTANCES_MASK(i915, first, count) ({ \ > > unsigned int first__ = (first); \ > > unsigned int count__ = (count); \ > > - (INTEL_INFO(dev_priv)->engine_mask & \ > > + (INTEL_INFO(i915)->engine_mask & \ > > GENMASK(first__ + count__ - 1, first__)) >> first__; \ > > }) > > -#define VDBOX_MASK(dev_priv) \ > > - ENGINE_INSTANCES_MASK(dev_priv, VCS0, I915_MAX_VCS) > > -#define VEBOX_MASK(dev_priv) \ > > - ENGINE_INSTANCES_MASK(dev_priv, VECS0, I915_MAX_VECS) > > - > > -#define HAS_LLC(dev_priv) (INTEL_INFO(dev_priv)->has_llc) > > -#define HAS_SNOOP(dev_priv) (INTEL_INFO(dev_priv)->has_snoop) > > -#define HAS_EDRAM(dev_priv) ((dev_priv)->edram_size_mb) > > -#define HAS_WT(dev_priv) ((IS_HASWELL(dev_priv) || \ > > - IS_BROADWELL(dev_priv)) && HAS_EDRAM(dev_priv)) > > - > > -#define HWS_NEEDS_PHYSICAL(dev_priv) (INTEL_INFO(dev_priv)->hws_needs_physical) > > - > > -#define HAS_LOGICAL_RING_CONTEXTS(dev_priv) \ > > - (INTEL_INFO(dev_priv)->has_logical_ring_contexts) > > -#define HAS_LOGICAL_RING_ELSQ(dev_priv) \ > > - (INTEL_INFO(dev_priv)->has_logical_ring_elsq) > > -#define HAS_LOGICAL_RING_PREEMPTION(dev_priv) \ > > - (INTEL_INFO(dev_priv)->has_logical_ring_preemption) > > - > > -#define HAS_EXECLISTS(dev_priv) HAS_LOGICAL_RING_CONTEXTS(dev_priv) > > - > > -#define INTEL_PPGTT(dev_priv) (INTEL_INFO(dev_priv)->ppgtt_type) > > -#define HAS_PPGTT(dev_priv) \ > > - (INTEL_PPGTT(dev_priv) != INTEL_PPGTT_NONE) > > -#define HAS_FULL_PPGTT(dev_priv) \ > > - (INTEL_PPGTT(dev_priv) >= INTEL_PPGTT_FULL) > > - > > -#define HAS_PAGE_SIZES(dev_priv, sizes) ({ \ > > +#define VDBOX_MASK(i915) \ > > + ENGINE_INSTANCES_MASK(i915, VCS0, I915_MAX_VCS) > > +#define VEBOX_MASK(i915) \ > > + ENGINE_INSTANCES_MASK(i915, VECS0, I915_MAX_VECS) > > + > > +#define HAS_LLC(i915) (INTEL_INFO(i915)->has_llc) > > +#define HAS_SNOOP(i915) (INTEL_INFO(i915)->has_snoop) > > +#define HAS_EDRAM(i915) ((i915)->edram_size_mb) > > +#define HAS_WT(i915) ((IS_HASWELL(i915) || \ > > + IS_BROADWELL(i915)) && HAS_EDRAM(i915)) > > + > > +#define HWS_NEEDS_PHYSICAL(i915) (INTEL_INFO(i915)->hws_needs_physical) > > + > > +#define HAS_LOGICAL_RING_CONTEXTS(i915) \ > > + (INTEL_INFO(i915)->has_logical_ring_contexts) > > +#define HAS_LOGICAL_RING_ELSQ(i915) \ > > + (INTEL_INFO(i915)->has_logical_ring_elsq) > > +#define HAS_LOGICAL_RING_PREEMPTION(i915) \ > > + (INTEL_INFO(i915)->has_logical_ring_preemption) > > + > > +#define HAS_EXECLISTS(i915) HAS_LOGICAL_RING_CONTEXTS(i915) > > + > > +#define INTEL_PPGTT(i915) (INTEL_INFO(i915)->ppgtt_type) > > +#define HAS_PPGTT(i915) \ > > + (INTEL_PPGTT(i915) != INTEL_PPGTT_NONE) > > +#define HAS_FULL_PPGTT(i915) \ > > + (INTEL_PPGTT(i915) >= INTEL_PPGTT_FULL) > > + > > +#define HAS_PAGE_SIZES(i915, sizes) ({ \ > > GEM_BUG_ON((sizes) == 0); \ > > - ((sizes) & ~INTEL_INFO(dev_priv)->page_sizes) == 0; \ > > + ((sizes) & ~INTEL_INFO(i915)->page_sizes) == 0; \ > > }) > > -#define HAS_OVERLAY(dev_priv) (INTEL_INFO(dev_priv)->display.has_overlay) > > -#define OVERLAY_NEEDS_PHYSICAL(dev_priv) \ > > - (INTEL_INFO(dev_priv)->display.overlay_needs_physical) > > +#define HAS_OVERLAY(i915) (INTEL_INFO(i915)->display.has_overlay) > > +#define OVERLAY_NEEDS_PHYSICAL(i915) \ > > + (INTEL_INFO(i915)->display.overlay_needs_physical) > > /* Early gen2 have a totally busted CS tlb and require pinned batches. */ > > -#define HAS_BROKEN_CS_TLB(dev_priv) (IS_I830(dev_priv) || IS_I845G(dev_priv)) > > +#define HAS_BROKEN_CS_TLB(i915) (IS_I830(i915) || IS_I845G(i915)) > > /* WaRsDisableCoarsePowerGating:skl,cnl */ > > -#define NEEDS_WaRsDisableCoarsePowerGating(dev_priv) \ > > - (IS_CANNONLAKE(dev_priv) || \ > > - IS_SKL_GT3(dev_priv) || IS_SKL_GT4(dev_priv)) > > +#define NEEDS_WaRsDisableCoarsePowerGating(i915) \ > > + (IS_CANNONLAKE(i915) || \ > > + IS_SKL_GT3(i915) || IS_SKL_GT4(i915)) > > -#define HAS_GMBUS_IRQ(dev_priv) (INTEL_GEN(dev_priv) >= 4) > > -#define HAS_GMBUS_BURST_READ(dev_priv) (INTEL_GEN(dev_priv) >= 10 || \ > > - IS_GEMINILAKE(dev_priv) || \ > > - IS_KABYLAKE(dev_priv)) > > +#define HAS_GMBUS_IRQ(i915) (INTEL_GEN(i915) >= 4) > > +#define HAS_GMBUS_BURST_READ(i915) (INTEL_GEN(i915) >= 10 || \ > > + IS_GEMINILAKE(i915) || \ > > + IS_KABYLAKE(i915)) > > /* With the 945 and later, Y tiling got adjusted so that it was 32 128-byte > > * rows, which changed the alignment requirements and fence programming. > > */ > > -#define HAS_128_BYTE_Y_TILING(dev_priv) (!IS_GEN(dev_priv, 2) && \ > > - !(IS_I915G(dev_priv) || \ > > - IS_I915GM(dev_priv))) > > -#define SUPPORTS_TV(dev_priv) (INTEL_INFO(dev_priv)->display.supports_tv) > > -#define I915_HAS_HOTPLUG(dev_priv) (INTEL_INFO(dev_priv)->display.has_hotplug) > > +#define HAS_128_BYTE_Y_TILING(i915) (!IS_GEN(i915, 2) && \ > > + !(IS_I915G(i915) || \ > > + IS_I915GM(i915))) > > +#define SUPPORTS_TV(i915) (INTEL_INFO(i915)->display.supports_tv) > > +#define I915_HAS_HOTPLUG(i915) (INTEL_INFO(i915)->display.has_hotplug) > > -#define HAS_FW_BLC(dev_priv) (INTEL_GEN(dev_priv) > 2) > > -#define HAS_FBC(dev_priv) (INTEL_INFO(dev_priv)->display.has_fbc) > > -#define HAS_CUR_FBC(dev_priv) (!HAS_GMCH(dev_priv) && INTEL_GEN(dev_priv) >= 7) > > +#define HAS_FW_BLC(i915) (INTEL_GEN(i915) > 2) > > +#define HAS_FBC(i915) (INTEL_INFO(i915)->display.has_fbc) > > +#define HAS_CUR_FBC(i915) (!HAS_GMCH(i915) && INTEL_GEN(i915) >= 7) > > -#define HAS_IPS(dev_priv) (IS_HSW_ULT(dev_priv) || IS_BROADWELL(dev_priv)) > > +#define HAS_IPS(i915) (IS_HSW_ULT(i915) || IS_BROADWELL(i915)) > > -#define HAS_DP_MST(dev_priv) (INTEL_INFO(dev_priv)->display.has_dp_mst) > > +#define HAS_DP_MST(i915) (INTEL_INFO(i915)->display.has_dp_mst) > > -#define HAS_DDI(dev_priv) (INTEL_INFO(dev_priv)->display.has_ddi) > > -#define HAS_FPGA_DBG_UNCLAIMED(dev_priv) (INTEL_INFO(dev_priv)->has_fpga_dbg) > > -#define HAS_PSR(dev_priv) (INTEL_INFO(dev_priv)->display.has_psr) > > -#define HAS_TRANSCODER_EDP(dev_priv) (INTEL_INFO(dev_priv)->trans_offsets[TRANSCODER_EDP] != 0) > > +#define HAS_DDI(i915) (INTEL_INFO(i915)->display.has_ddi) > > +#define HAS_FPGA_DBG_UNCLAIMED(i915) (INTEL_INFO(i915)->has_fpga_dbg) > > +#define HAS_PSR(i915) (INTEL_INFO(i915)->display.has_psr) > > +#define HAS_TRANSCODER_EDP(i915) (INTEL_INFO(i915)->trans_offsets[TRANSCODER_EDP] != 0) > > -#define HAS_RC6(dev_priv) (INTEL_INFO(dev_priv)->has_rc6) > > -#define HAS_RC6p(dev_priv) (INTEL_INFO(dev_priv)->has_rc6p) > > -#define HAS_RC6pp(dev_priv) (false) /* HW was never validated */ > > +#define HAS_RC6(i915) (INTEL_INFO(i915)->has_rc6) > > +#define HAS_RC6p(i915) (INTEL_INFO(i915)->has_rc6p) > > +#define HAS_RC6pp(i915) (false) /* HW was never validated */ > > -#define HAS_RPS(dev_priv) (INTEL_INFO(dev_priv)->has_rps) > > +#define HAS_RPS(i915) (INTEL_INFO(i915)->has_rps) > > -#define HAS_CSR(dev_priv) (INTEL_INFO(dev_priv)->display.has_csr) > > +#define HAS_CSR(i915) (INTEL_INFO(i915)->display.has_csr) > > -#define HAS_RUNTIME_PM(dev_priv) (INTEL_INFO(dev_priv)->has_runtime_pm) > > -#define HAS_64BIT_RELOC(dev_priv) (INTEL_INFO(dev_priv)->has_64bit_reloc) > > +#define HAS_RUNTIME_PM(i915) (INTEL_INFO(i915)->has_runtime_pm) > > +#define HAS_64BIT_RELOC(i915) (INTEL_INFO(i915)->has_64bit_reloc) > > -#define HAS_IPC(dev_priv) (INTEL_INFO(dev_priv)->display.has_ipc) > > +#define HAS_IPC(i915) (INTEL_INFO(i915)->display.has_ipc) > > /* > > * For now, anything with a GuC requires uCode loading, and then supports > > * command submission once loaded. But these are logically independent > > * properties, so we have separate macros to test them. > > */ > > -#define HAS_GUC(dev_priv) (INTEL_INFO(dev_priv)->has_guc) > > -#define HAS_GUC_UCODE(dev_priv) (HAS_GUC(dev_priv)) > > -#define HAS_GUC_SCHED(dev_priv) (HAS_GUC(dev_priv)) > > +#define HAS_GUC(i915) (INTEL_INFO(i915)->has_guc) > > +#define HAS_GUC_UCODE(i915) (HAS_GUC(i915)) > > +#define HAS_GUC_SCHED(i915) (HAS_GUC(i915)) > > /* For now, anything with a GuC has also HuC */ > > -#define HAS_HUC(dev_priv) (HAS_GUC(dev_priv)) > > -#define HAS_HUC_UCODE(dev_priv) (HAS_GUC(dev_priv)) > > +#define HAS_HUC(i915) (HAS_GUC(i915)) > > +#define HAS_HUC_UCODE(i915) (HAS_GUC(i915)) > > /* Having a GuC is not the same as using a GuC */ > > -#define USES_GUC(dev_priv) intel_uc_is_using_guc(dev_priv) > > -#define USES_GUC_SUBMISSION(dev_priv) intel_uc_is_using_guc_submission(dev_priv) > > -#define USES_HUC(dev_priv) intel_uc_is_using_huc(dev_priv) > > +#define USES_GUC(i915) intel_uc_is_using_guc(i915) > > +#define USES_GUC_SUBMISSION(i915) intel_uc_is_using_guc_submission(i915) > > +#define USES_HUC(i915) intel_uc_is_using_huc(i915) > > -#define HAS_POOLED_EU(dev_priv) (INTEL_INFO(dev_priv)->has_pooled_eu) > > +#define HAS_POOLED_EU(i915) (INTEL_INFO(i915)->has_pooled_eu) > > #define INTEL_PCH_DEVICE_ID_MASK 0xff80 > > #define INTEL_PCH_IBX_DEVICE_ID_TYPE 0x3b00 > > @@ -2429,36 +2429,36 @@ IS_SUBPLATFORM(const struct drm_i915_private *i915, > > #define INTEL_PCH_P3X_DEVICE_ID_TYPE 0x7000 > > #define INTEL_PCH_QEMU_DEVICE_ID_TYPE 0x2900 /* qemu q35 has 2918 */ > > -#define INTEL_PCH_TYPE(dev_priv) ((dev_priv)->pch_type) > > -#define INTEL_PCH_ID(dev_priv) ((dev_priv)->pch_id) > > -#define HAS_PCH_ICP(dev_priv) (INTEL_PCH_TYPE(dev_priv) == PCH_ICP) > > -#define HAS_PCH_CNP(dev_priv) (INTEL_PCH_TYPE(dev_priv) == PCH_CNP) > > -#define HAS_PCH_SPT(dev_priv) (INTEL_PCH_TYPE(dev_priv) == PCH_SPT) > > -#define HAS_PCH_LPT(dev_priv) (INTEL_PCH_TYPE(dev_priv) == PCH_LPT) > > -#define HAS_PCH_LPT_LP(dev_priv) \ > > - (INTEL_PCH_ID(dev_priv) == INTEL_PCH_LPT_LP_DEVICE_ID_TYPE || \ > > - INTEL_PCH_ID(dev_priv) == INTEL_PCH_WPT_LP_DEVICE_ID_TYPE) > > -#define HAS_PCH_LPT_H(dev_priv) \ > > - (INTEL_PCH_ID(dev_priv) == INTEL_PCH_LPT_DEVICE_ID_TYPE || \ > > - INTEL_PCH_ID(dev_priv) == INTEL_PCH_WPT_DEVICE_ID_TYPE) > > -#define HAS_PCH_CPT(dev_priv) (INTEL_PCH_TYPE(dev_priv) == PCH_CPT) > > -#define HAS_PCH_IBX(dev_priv) (INTEL_PCH_TYPE(dev_priv) == PCH_IBX) > > -#define HAS_PCH_NOP(dev_priv) (INTEL_PCH_TYPE(dev_priv) == PCH_NOP) > > -#define HAS_PCH_SPLIT(dev_priv) (INTEL_PCH_TYPE(dev_priv) != PCH_NONE) > > - > > -#define HAS_GMCH(dev_priv) (INTEL_INFO(dev_priv)->display.has_gmch) > > - > > -#define HAS_LSPCON(dev_priv) (INTEL_GEN(dev_priv) >= 9) > > +#define INTEL_PCH_TYPE(i915) ((i915)->pch_type) > > +#define INTEL_PCH_ID(i915) ((i915)->pch_id) > > +#define HAS_PCH_ICP(i915) (INTEL_PCH_TYPE(i915) == PCH_ICP) > > +#define HAS_PCH_CNP(i915) (INTEL_PCH_TYPE(i915) == PCH_CNP) > > +#define HAS_PCH_SPT(i915) (INTEL_PCH_TYPE(i915) == PCH_SPT) > > +#define HAS_PCH_LPT(i915) (INTEL_PCH_TYPE(i915) == PCH_LPT) > > +#define HAS_PCH_LPT_LP(i915) \ > > + (INTEL_PCH_ID(i915) == INTEL_PCH_LPT_LP_DEVICE_ID_TYPE || \ > > + INTEL_PCH_ID(i915) == INTEL_PCH_WPT_LP_DEVICE_ID_TYPE) > > +#define HAS_PCH_LPT_H(i915) \ > > + (INTEL_PCH_ID(i915) == INTEL_PCH_LPT_DEVICE_ID_TYPE || \ > > + INTEL_PCH_ID(i915) == INTEL_PCH_WPT_DEVICE_ID_TYPE) > > +#define HAS_PCH_CPT(i915) (INTEL_PCH_TYPE(i915) == PCH_CPT) > > +#define HAS_PCH_IBX(i915) (INTEL_PCH_TYPE(i915) == PCH_IBX) > > +#define HAS_PCH_NOP(i915) (INTEL_PCH_TYPE(i915) == PCH_NOP) > > +#define HAS_PCH_SPLIT(i915) (INTEL_PCH_TYPE(i915) != PCH_NONE) > > + > > +#define HAS_GMCH(i915) (INTEL_INFO(i915)->display.has_gmch) > > + > > +#define HAS_LSPCON(i915) (INTEL_GEN(i915) >= 9) > > /* DPF == dynamic parity feature */ > > -#define HAS_L3_DPF(dev_priv) (INTEL_INFO(dev_priv)->has_l3_dpf) > > -#define NUM_L3_SLICES(dev_priv) (IS_HSW_GT3(dev_priv) ? \ > > - 2 : HAS_L3_DPF(dev_priv)) > > +#define HAS_L3_DPF(i915) (INTEL_INFO(i915)->has_l3_dpf) > > +#define NUM_L3_SLICES(i915) (IS_HSW_GT3(i915) ? \ > > + 2 : HAS_L3_DPF(i915)) > > #define GT_FREQUENCY_MULTIPLIER 50 > > #define GEN9_FREQ_SCALER 3 > > -#define HAS_DISPLAY(dev_priv) (INTEL_INFO(dev_priv)->num_pipes > 0) > > +#define HAS_DISPLAY(i915) (INTEL_INFO(i915)->num_pipes > 0) > > #include "i915_trace.h" > > @@ -2471,24 +2471,24 @@ static inline bool intel_vtd_active(void) > > return false; > > } > > -static inline bool intel_scanout_needs_vtd_wa(struct drm_i915_private *dev_priv) > > +static inline bool intel_scanout_needs_vtd_wa(struct drm_i915_private *i915) > > { > > - return INTEL_GEN(dev_priv) >= 6 && intel_vtd_active(); > > + return INTEL_GEN(i915) >= 6 && intel_vtd_active(); > > } > > static inline bool > > -intel_ggtt_update_needs_vtd_wa(struct drm_i915_private *dev_priv) > > +intel_ggtt_update_needs_vtd_wa(struct drm_i915_private *i915) > > { > > - return IS_BROXTON(dev_priv) && intel_vtd_active(); > > + return IS_BROXTON(i915) && intel_vtd_active(); > > } > > /* i915_drv.c */ > > void __printf(3, 4) > > -__i915_printk(struct drm_i915_private *dev_priv, const char *level, > > +__i915_printk(struct drm_i915_private *i915, const char *level, > > const char *fmt, ...); > > -#define i915_report_error(dev_priv, fmt, ...) \ > > - __i915_printk(dev_priv, KERN_ERR, fmt, ##__VA_ARGS__) > > +#define i915_report_error(i915, fmt, ...) \ > > + __i915_printk(i915, KERN_ERR, fmt, ##__VA_ARGS__) > > #ifdef CONFIG_COMPAT > > extern long i915_compat_ioctl(struct file *filp, unsigned int cmd, > > @@ -2503,12 +2503,12 @@ extern int i915_driver_load(struct pci_dev *pdev, > > extern void i915_driver_unload(struct drm_device *dev); > > extern void intel_engine_init_hangcheck(struct intel_engine_cs *engine); > > -extern void intel_hangcheck_init(struct drm_i915_private *dev_priv); > > -int vlv_force_gfx_clock(struct drm_i915_private *dev_priv, bool on); > > +extern void intel_hangcheck_init(struct drm_i915_private *i915); > > +int vlv_force_gfx_clock(struct drm_i915_private *i915, bool on); > > -u32 intel_calculate_mcr_s_ss_select(struct drm_i915_private *dev_priv); > > +u32 intel_calculate_mcr_s_ss_select(struct drm_i915_private *i915); > > -static inline void i915_queue_hangcheck(struct drm_i915_private *dev_priv) > > +static inline void i915_queue_hangcheck(struct drm_i915_private *i915) > > { > > unsigned long delay; > > @@ -2522,28 +2522,28 @@ static inline void i915_queue_hangcheck(struct drm_i915_private *dev_priv) > > delay = round_jiffies_up_relative(DRM_I915_HANGCHECK_JIFFIES); > > queue_delayed_work(system_long_wq, > > - &dev_priv->gpu_error.hangcheck_work, delay); > > + &i915->gpu_error.hangcheck_work, delay); > > } > > -static inline bool intel_gvt_active(struct drm_i915_private *dev_priv) > > +static inline bool intel_gvt_active(struct drm_i915_private *i915) > > { > > - return dev_priv->gvt; > > + return i915->gvt; > > } > > -static inline bool intel_vgpu_active(struct drm_i915_private *dev_priv) > > +static inline bool intel_vgpu_active(struct drm_i915_private *i915) > > { > > - return dev_priv->vgpu.active; > > + return i915->vgpu.active; > > } > > /* i915_gem.c */ > > -int i915_gem_init_userptr(struct drm_i915_private *dev_priv); > > -void i915_gem_cleanup_userptr(struct drm_i915_private *dev_priv); > > +int i915_gem_init_userptr(struct drm_i915_private *i915); > > +void i915_gem_cleanup_userptr(struct drm_i915_private *i915); > > void i915_gem_sanitize(struct drm_i915_private *i915); > > -int i915_gem_init_early(struct drm_i915_private *dev_priv); > > -void i915_gem_cleanup_early(struct drm_i915_private *dev_priv); > > -void i915_gem_load_init_fences(struct drm_i915_private *dev_priv); > > -int i915_gem_freeze(struct drm_i915_private *dev_priv); > > -int i915_gem_freeze_late(struct drm_i915_private *dev_priv); > > +int i915_gem_init_early(struct drm_i915_private *i915); > > +void i915_gem_cleanup_early(struct drm_i915_private *i915); > > +void i915_gem_load_init_fences(struct drm_i915_private *i915); > > +int i915_gem_freeze(struct drm_i915_private *i915); > > +int i915_gem_freeze_late(struct drm_i915_private *i915); > > static inline void i915_gem_drain_freed_objects(struct drm_i915_private *i915) > > { > > @@ -2591,7 +2591,7 @@ i915_gem_object_ggtt_pin(struct drm_i915_gem_object *obj, > > int i915_gem_object_unbind(struct drm_i915_gem_object *obj); > > -void i915_gem_runtime_suspend(struct drm_i915_private *dev_priv); > > +void i915_gem_runtime_suspend(struct drm_i915_private *i915); > > static inline int __must_check > > i915_mutex_lock_interruptible(struct drm_device *dev) > > @@ -2633,20 +2633,20 @@ static inline u32 i915_reset_engine_count(struct i915_gpu_error *error, > > return READ_ONCE(error->reset_engine_count[engine->id]); > > } > > -void i915_gem_set_wedged(struct drm_i915_private *dev_priv); > > -bool i915_gem_unset_wedged(struct drm_i915_private *dev_priv); > > +void i915_gem_set_wedged(struct drm_i915_private *i915); > > +bool i915_gem_unset_wedged(struct drm_i915_private *i915); > > void i915_gem_init_mmio(struct drm_i915_private *i915); > > -int __must_check i915_gem_init(struct drm_i915_private *dev_priv); > > -int __must_check i915_gem_init_hw(struct drm_i915_private *dev_priv); > > -void i915_gem_init_swizzling(struct drm_i915_private *dev_priv); > > -void i915_gem_fini_hw(struct drm_i915_private *dev_priv); > > -void i915_gem_fini(struct drm_i915_private *dev_priv); > > -int i915_gem_wait_for_idle(struct drm_i915_private *dev_priv, > > +int __must_check i915_gem_init(struct drm_i915_private *i915); > > +int __must_check i915_gem_init_hw(struct drm_i915_private *i915); > > +void i915_gem_init_swizzling(struct drm_i915_private *i915); > > +void i915_gem_fini_hw(struct drm_i915_private *i915); > > +void i915_gem_fini(struct drm_i915_private *i915); > > +int i915_gem_wait_for_idle(struct drm_i915_private *i915, > > unsigned int flags, long timeout); > > -void i915_gem_suspend(struct drm_i915_private *dev_priv); > > -void i915_gem_suspend_late(struct drm_i915_private *dev_priv); > > -void i915_gem_resume(struct drm_i915_private *dev_priv); > > +void i915_gem_suspend(struct drm_i915_private *i915); > > +void i915_gem_suspend_late(struct drm_i915_private *i915); > > +void i915_gem_resume(struct drm_i915_private *i915); > > vm_fault_t i915_gem_fault(struct vm_fault *vmf); > > int i915_gem_open(struct drm_i915_private *i915, struct drm_file *file); > > @@ -2663,12 +2663,12 @@ struct dma_buf *i915_gem_prime_export(struct drm_device *dev, > > /* i915_gem_fence_reg.c */ > > struct drm_i915_fence_reg * > > -i915_reserve_fence(struct drm_i915_private *dev_priv); > > +i915_reserve_fence(struct drm_i915_private *i915); > > void i915_unreserve_fence(struct drm_i915_fence_reg *fence); > > -void i915_gem_restore_fences(struct drm_i915_private *dev_priv); > > +void i915_gem_restore_fences(struct drm_i915_private *i915); > > -void i915_gem_detect_bit_6_swizzle(struct drm_i915_private *dev_priv); > > +void i915_gem_detect_bit_6_swizzle(struct drm_i915_private *i915); > > void i915_gem_object_do_bit_17_swizzle(struct drm_i915_gem_object *obj, > > struct sg_table *pages); > > void i915_gem_object_save_bit_17_swizzle(struct drm_i915_gem_object *obj, > > @@ -2715,40 +2715,40 @@ int __must_check i915_gem_evict_for_node(struct i915_address_space *vm, > > unsigned int flags); > > int i915_gem_evict_vm(struct i915_address_space *vm); > > -void i915_gem_flush_ggtt_writes(struct drm_i915_private *dev_priv); > > +void i915_gem_flush_ggtt_writes(struct drm_i915_private *i915); > > /* belongs in i915_gem_gtt.h */ > > -static inline void i915_gem_chipset_flush(struct drm_i915_private *dev_priv) > > +static inline void i915_gem_chipset_flush(struct drm_i915_private *i915) > > { > > wmb(); > > - if (INTEL_GEN(dev_priv) < 6) > > + if (INTEL_GEN(i915) < 6) > > intel_gtt_chipset_flush(); > > } > > /* i915_gem_stolen.c */ > > -int i915_gem_stolen_insert_node(struct drm_i915_private *dev_priv, > > +int i915_gem_stolen_insert_node(struct drm_i915_private *i915, > > struct drm_mm_node *node, u64 size, > > unsigned alignment); > > -int i915_gem_stolen_insert_node_in_range(struct drm_i915_private *dev_priv, > > +int i915_gem_stolen_insert_node_in_range(struct drm_i915_private *i915, > > struct drm_mm_node *node, u64 size, > > unsigned alignment, u64 start, > > u64 end); > > -void i915_gem_stolen_remove_node(struct drm_i915_private *dev_priv, > > +void i915_gem_stolen_remove_node(struct drm_i915_private *i915, > > struct drm_mm_node *node); > > -int i915_gem_init_stolen(struct drm_i915_private *dev_priv); > > -void i915_gem_cleanup_stolen(struct drm_i915_private *dev_priv); > > +int i915_gem_init_stolen(struct drm_i915_private *i915); > > +void i915_gem_cleanup_stolen(struct drm_i915_private *i915); > > struct drm_i915_gem_object * > > -i915_gem_object_create_stolen(struct drm_i915_private *dev_priv, > > +i915_gem_object_create_stolen(struct drm_i915_private *i915, > > resource_size_t size); > > struct drm_i915_gem_object * > > -i915_gem_object_create_stolen_for_preallocated(struct drm_i915_private *dev_priv, > > +i915_gem_object_create_stolen_for_preallocated(struct drm_i915_private *i915, > > resource_size_t stolen_offset, > > resource_size_t gtt_offset, > > resource_size_t size); > > /* i915_gem_internal.c */ > > struct drm_i915_gem_object * > > -i915_gem_object_create_internal(struct drm_i915_private *dev_priv, > > +i915_gem_object_create_internal(struct drm_i915_private *i915, > > phys_addr_t size); > > /* i915_gem_shrinker.c */ > > @@ -2771,21 +2771,21 @@ void i915_gem_shrinker_taints_mutex(struct drm_i915_private *i915, > > /* i915_gem_tiling.c */ > > static inline bool i915_gem_object_needs_bit17_swizzle(struct drm_i915_gem_object *obj) > > { > > - struct drm_i915_private *dev_priv = to_i915(obj->base.dev); > > + struct drm_i915_private *i915 = to_i915(obj->base.dev); > > - return dev_priv->mm.bit_6_swizzle_x == I915_BIT_6_SWIZZLE_9_10_17 && > > + return i915->mm.bit_6_swizzle_x == I915_BIT_6_SWIZZLE_9_10_17 && > > i915_gem_object_is_tiled(obj); > > } > > -u32 i915_gem_fence_size(struct drm_i915_private *dev_priv, u32 size, > > +u32 i915_gem_fence_size(struct drm_i915_private *i915, u32 size, > > unsigned int tiling, unsigned int stride); > > -u32 i915_gem_fence_alignment(struct drm_i915_private *dev_priv, u32 size, > > +u32 i915_gem_fence_alignment(struct drm_i915_private *i915, u32 size, > > unsigned int tiling, unsigned int stride); > > const char *i915_cache_level_str(struct drm_i915_private *i915, int type); > > /* i915_cmd_parser.c */ > > -int i915_cmd_parser_get_version(struct drm_i915_private *dev_priv); > > +int i915_cmd_parser_get_version(struct drm_i915_private *i915); > > void intel_engine_init_cmd_parser(struct intel_engine_cs *engine); > > void intel_engine_cleanup_cmd_parser(struct intel_engine_cs *engine); > > int intel_engine_cmd_parser(struct intel_engine_cs *engine, > > @@ -2796,52 +2796,52 @@ int intel_engine_cmd_parser(struct intel_engine_cs *engine, > > bool is_master); > > /* i915_perf.c */ > > -extern void i915_perf_init(struct drm_i915_private *dev_priv); > > -extern void i915_perf_fini(struct drm_i915_private *dev_priv); > > -extern void i915_perf_register(struct drm_i915_private *dev_priv); > > -extern void i915_perf_unregister(struct drm_i915_private *dev_priv); > > +extern void i915_perf_init(struct drm_i915_private *i915); > > +extern void i915_perf_fini(struct drm_i915_private *i915); > > +extern void i915_perf_register(struct drm_i915_private *i915); > > +extern void i915_perf_unregister(struct drm_i915_private *i915); > > /* i915_suspend.c */ > > -extern int i915_save_state(struct drm_i915_private *dev_priv); > > -extern int i915_restore_state(struct drm_i915_private *dev_priv); > > +extern int i915_save_state(struct drm_i915_private *i915); > > +extern int i915_restore_state(struct drm_i915_private *i915); > > /* i915_sysfs.c */ > > -void i915_setup_sysfs(struct drm_i915_private *dev_priv); > > -void i915_teardown_sysfs(struct drm_i915_private *dev_priv); > > +void i915_setup_sysfs(struct drm_i915_private *i915); > > +void i915_teardown_sysfs(struct drm_i915_private *i915); > > /* intel_device_info.c */ > > static inline struct intel_device_info * > > -mkwrite_device_info(struct drm_i915_private *dev_priv) > > +mkwrite_device_info(struct drm_i915_private *i915) > > { > > - return (struct intel_device_info *)INTEL_INFO(dev_priv); > > + return (struct intel_device_info *)INTEL_INFO(i915); > > } > > /* modesetting */ > > extern void intel_modeset_init_hw(struct drm_device *dev); > > extern int intel_modeset_init(struct drm_device *dev); > > extern void intel_modeset_cleanup(struct drm_device *dev); > > -extern int intel_modeset_vga_set_state(struct drm_i915_private *dev_priv, > > +extern int intel_modeset_vga_set_state(struct drm_i915_private *i915, > > bool state); > > extern void intel_display_resume(struct drm_device *dev); > > -extern void i915_redisable_vga(struct drm_i915_private *dev_priv); > > -extern void i915_redisable_vga_power_on(struct drm_i915_private *dev_priv); > > -extern void intel_init_pch_refclk(struct drm_i915_private *dev_priv); > > +extern void i915_redisable_vga(struct drm_i915_private *i915); > > +extern void i915_redisable_vga_power_on(struct drm_i915_private *i915); > > +extern void intel_init_pch_refclk(struct drm_i915_private *i915); > > int i915_reg_read_ioctl(struct drm_device *dev, void *data, > > struct drm_file *file); > > extern struct intel_display_error_state * > > -intel_display_capture_error_state(struct drm_i915_private *dev_priv); > > +intel_display_capture_error_state(struct drm_i915_private *i915); > > extern void intel_display_print_error_state(struct drm_i915_error_state_buf *e, > > struct intel_display_error_state *error); > > #define __I915_REG_OP(op__, dev_priv__, ...) \ > > intel_uncore_##op__(&(dev_priv__)->uncore, __VA_ARGS__) > > -#define I915_READ(reg__) __I915_REG_OP(read, dev_priv, (reg__)) > > -#define I915_WRITE(reg__, val__) __I915_REG_OP(write, dev_priv, (reg__), (val__)) > > +#define I915_READ(reg__) __I915_REG_OP(read, i915, (reg__)) > > +#define I915_WRITE(reg__, val__) __I915_REG_OP(write, i915, (reg__), (val__)) > > -#define POSTING_READ(reg__) __I915_REG_OP(posting_read, dev_priv, (reg__)) > > +#define POSTING_READ(reg__) __I915_REG_OP(posting_read, i915, (reg__)) > > /* These are untraced mmio-accessors that are only valid to be used inside > > * critical sections, such as inside IRQ handlers, where forcewake is explicitly > > @@ -2851,13 +2851,13 @@ extern void intel_display_print_error_state(struct drm_i915_error_state_buf *e, > > * > > * As an example, these accessors can possibly be used between: > > * > > - * spin_lock_irq(&dev_priv->uncore.lock); > > + * spin_lock_irq(&i915->uncore.lock); > > * intel_uncore_forcewake_get__locked(); > > * > > * and > > * > > * intel_uncore_forcewake_put__locked(); > > - * spin_unlock_irq(&dev_priv->uncore.lock); > > + * spin_unlock_irq(&i915->uncore.lock); > > * > > * > > * Note: some registers may not need forcewake held, so > > @@ -2866,18 +2866,18 @@ extern void intel_display_print_error_state(struct drm_i915_error_state_buf *e, > > * > > * Certain architectures will die if the same cacheline is concurrently accessed > > * by different clients (e.g. on Ivybridge). Access to registers should > > - * therefore generally be serialised, by either the dev_priv->uncore.lock or > > + * therefore generally be serialised, by either the i915->uncore.lock or > > * a more localised lock guarding all access to that bank of registers. > > */ > > -#define I915_READ_FW(reg__) __I915_REG_OP(read_fw, dev_priv, (reg__)) > > -#define I915_WRITE_FW(reg__, val__) __I915_REG_OP(write_fw, dev_priv, (reg__), (val__)) > > +#define I915_READ_FW(reg__) __I915_REG_OP(read_fw, i915, (reg__)) > > +#define I915_WRITE_FW(reg__, val__) __I915_REG_OP(write_fw, i915, (reg__), (val__)) > > /* "Broadcast RGB" property */ > > #define INTEL_BROADCAST_RGB_AUTO 0 > > #define INTEL_BROADCAST_RGB_FULL 1 > > #define INTEL_BROADCAST_RGB_LIMITED 2 > > -void i915_memcpy_init_early(struct drm_i915_private *dev_priv); > > +void i915_memcpy_init_early(struct drm_i915_private *i915); > > bool i915_memcpy_from_wc(void *dst, const void *src, unsigned long len); > > /* The movntdqa instructions used for memcpy-from-wc require 16-byte alignment, > > diff --git a/drivers/gpu/drm/i915/i915_gem.c b/drivers/gpu/drm/i915/i915_gem.c > > index 4017ecf561f6..1774c7604076 100644 > > --- a/drivers/gpu/drm/i915/i915_gem.c > > +++ b/drivers/gpu/drm/i915/i915_gem.c > > @@ -149,7 +149,7 @@ i915_gem_phys_pwrite(struct drm_i915_gem_object *obj, > > static int > > i915_gem_create(struct drm_file *file, > > - struct drm_i915_private *dev_priv, > > + struct drm_i915_private *i915, > > u64 *size_p, > > u32 *handle_p) > > { > > @@ -163,7 +163,7 @@ i915_gem_create(struct drm_file *file, > > return -EINVAL; > > /* Allocate the new object */ > > - obj = i915_gem_object_create_shmem(dev_priv, size); > > + obj = i915_gem_object_create_shmem(i915, size); > > if (IS_ERR(obj)) > > return PTR_ERR(obj); > > @@ -223,16 +223,16 @@ int > > i915_gem_create_ioctl(struct drm_device *dev, void *data, > > struct drm_file *file) > > { > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > struct drm_i915_gem_create *args = data; > > - i915_gem_flush_free_objects(dev_priv); > > + i915_gem_flush_free_objects(i915); > > - return i915_gem_create(file, dev_priv, > > + return i915_gem_create(file, i915, > > &args->size, &args->handle); > > } > > -void i915_gem_flush_ggtt_writes(struct drm_i915_private *dev_priv) > > +void i915_gem_flush_ggtt_writes(struct drm_i915_private *i915) > > { > > intel_wakeref_t wakeref; > > @@ -257,13 +257,13 @@ void i915_gem_flush_ggtt_writes(struct drm_i915_private *dev_priv) > > wmb(); > > - if (INTEL_INFO(dev_priv)->has_coherent_ggtt) > > + if (INTEL_INFO(i915)->has_coherent_ggtt) > > return; > > - i915_gem_chipset_flush(dev_priv); > > + i915_gem_chipset_flush(i915); > > - with_intel_runtime_pm(dev_priv, wakeref) { > > - struct intel_uncore *uncore = &dev_priv->uncore; > > + with_intel_runtime_pm(i915, wakeref) { > > + struct intel_uncore *uncore = &i915->uncore; > > spin_lock_irq(&uncore->lock); > > intel_uncore_posting_read_fw(uncore, > > @@ -884,7 +884,7 @@ i915_gem_sw_finish_ioctl(struct drm_device *dev, void *data, > > return 0; > > } > > -void i915_gem_runtime_suspend(struct drm_i915_private *dev_priv) > > +void i915_gem_runtime_suspend(struct drm_i915_private *i915) > > { > > struct drm_i915_gem_object *obj, *on; > > int i; > > @@ -897,15 +897,15 @@ void i915_gem_runtime_suspend(struct drm_i915_private *dev_priv) > > */ > > list_for_each_entry_safe(obj, on, > > - &dev_priv->mm.userfault_list, userfault_link) > > + &i915->mm.userfault_list, userfault_link) > > __i915_gem_object_release_mmap(obj); > > /* The fence will be lost when the device powers down. If any were > > * in use by hardware (i.e. they are pinned), we should not be powering > > * down! All other fences will be reacquired by the user upon waking. > > */ > > - for (i = 0; i < dev_priv->num_fence_regs; i++) { > > - struct drm_i915_fence_reg *reg = &dev_priv->fence_regs[i]; > > + for (i = 0; i < i915->num_fence_regs; i++) { > > + struct drm_i915_fence_reg *reg = &i915->fence_regs[i]; > > /* Ideally we want to assert that the fence register is not > > * live at this point (i.e. that no piece of code will be > > @@ -1020,8 +1020,8 @@ i915_gem_object_ggtt_pin(struct drm_i915_gem_object *obj, > > u64 alignment, > > u64 flags) > > { > > - struct drm_i915_private *dev_priv = to_i915(obj->base.dev); > > - struct i915_address_space *vm = &dev_priv->ggtt.vm; > > + struct drm_i915_private *i915 = to_i915(obj->base.dev); > > + struct i915_address_space *vm = &i915->ggtt.vm; > > struct i915_vma *vma; > > int ret; > > @@ -1036,7 +1036,7 @@ i915_gem_object_ggtt_pin(struct drm_i915_gem_object *obj, > > * the object in and out of the Global GTT and > > * waste a lot of cycles under the mutex. > > */ > > - if (obj->base.size > dev_priv->ggtt.mappable_end) > > + if (obj->base.size > i915->ggtt.mappable_end) > > return ERR_PTR(-E2BIG); > > /* If NONBLOCK is set the caller is optimistically > > @@ -1055,7 +1055,7 @@ i915_gem_object_ggtt_pin(struct drm_i915_gem_object *obj, > > * we could try to minimise harm to others. > > */ > > if (flags & PIN_NONBLOCK && > > - obj->base.size > dev_priv->ggtt.mappable_end / 2) > > + obj->base.size > i915->ggtt.mappable_end / 2) > > return ERR_PTR(-ENOSPC); > > } > > @@ -1069,7 +1069,7 @@ i915_gem_object_ggtt_pin(struct drm_i915_gem_object *obj, > > return ERR_PTR(-ENOSPC); > > if (flags & PIN_MAPPABLE && > > - vma->fence_size > dev_priv->ggtt.mappable_end / 2) > > + vma->fence_size > i915->ggtt.mappable_end / 2) > > return ERR_PTR(-ENOSPC); > > } > > @@ -1202,30 +1202,30 @@ void i915_gem_sanitize(struct drm_i915_private *i915) > > mutex_unlock(&i915->drm.struct_mutex); > > } > > -void i915_gem_init_swizzling(struct drm_i915_private *dev_priv) > > +void i915_gem_init_swizzling(struct drm_i915_private *i915) > > { > > - if (INTEL_GEN(dev_priv) < 5 || > > - dev_priv->mm.bit_6_swizzle_x == I915_BIT_6_SWIZZLE_NONE) > > + if (INTEL_GEN(i915) < 5 || > > + i915->mm.bit_6_swizzle_x == I915_BIT_6_SWIZZLE_NONE) > > return; > > I915_WRITE(DISP_ARB_CTL, I915_READ(DISP_ARB_CTL) | > > DISP_TILE_SURFACE_SWIZZLING); > > - if (IS_GEN(dev_priv, 5)) > > + if (IS_GEN(i915, 5)) > > return; > > I915_WRITE(TILECTL, I915_READ(TILECTL) | TILECTL_SWZCTL); > > - if (IS_GEN(dev_priv, 6)) > > + if (IS_GEN(i915, 6)) > > I915_WRITE(ARB_MODE, _MASKED_BIT_ENABLE(ARB_MODE_SWIZZLE_SNB)); > > - else if (IS_GEN(dev_priv, 7)) > > + else if (IS_GEN(i915, 7)) > > I915_WRITE(ARB_MODE, _MASKED_BIT_ENABLE(ARB_MODE_SWIZZLE_IVB)); > > - else if (IS_GEN(dev_priv, 8)) > > + else if (IS_GEN(i915, 8)) > > I915_WRITE(GAMTARBMODE, _MASKED_BIT_ENABLE(ARB_MODE_SWIZZLE_BDW)); > > else > > BUG(); > > } > > -static void init_unused_ring(struct drm_i915_private *dev_priv, u32 base) > > +static void init_unused_ring(struct drm_i915_private *i915, u32 base) > > { > > I915_WRITE(RING_CTL(base), 0); > > I915_WRITE(RING_HEAD(base), 0); > > @@ -1233,45 +1233,45 @@ static void init_unused_ring(struct drm_i915_private *dev_priv, u32 base) > > I915_WRITE(RING_START(base), 0); > > } > > -static void init_unused_rings(struct drm_i915_private *dev_priv) > > +static void init_unused_rings(struct drm_i915_private *i915) > > { > > - if (IS_I830(dev_priv)) { > > - init_unused_ring(dev_priv, PRB1_BASE); > > - init_unused_ring(dev_priv, SRB0_BASE); > > - init_unused_ring(dev_priv, SRB1_BASE); > > - init_unused_ring(dev_priv, SRB2_BASE); > > - init_unused_ring(dev_priv, SRB3_BASE); > > - } else if (IS_GEN(dev_priv, 2)) { > > - init_unused_ring(dev_priv, SRB0_BASE); > > - init_unused_ring(dev_priv, SRB1_BASE); > > - } else if (IS_GEN(dev_priv, 3)) { > > - init_unused_ring(dev_priv, PRB1_BASE); > > - init_unused_ring(dev_priv, PRB2_BASE); > > + if (IS_I830(i915)) { > > + init_unused_ring(i915, PRB1_BASE); > > + init_unused_ring(i915, SRB0_BASE); > > + init_unused_ring(i915, SRB1_BASE); > > + init_unused_ring(i915, SRB2_BASE); > > + init_unused_ring(i915, SRB3_BASE); > > + } else if (IS_GEN(i915, 2)) { > > + init_unused_ring(i915, SRB0_BASE); > > + init_unused_ring(i915, SRB1_BASE); > > + } else if (IS_GEN(i915, 3)) { > > + init_unused_ring(i915, PRB1_BASE); > > + init_unused_ring(i915, PRB2_BASE); > > } > > } > > -int i915_gem_init_hw(struct drm_i915_private *dev_priv) > > +int i915_gem_init_hw(struct drm_i915_private *i915) > > { > > int ret; > > - dev_priv->gt.last_init_time = ktime_get(); > > + i915->gt.last_init_time = ktime_get(); > > /* Double layer security blanket, see i915_gem_init() */ > > - intel_uncore_forcewake_get(&dev_priv->uncore, FORCEWAKE_ALL); > > + intel_uncore_forcewake_get(&i915->uncore, FORCEWAKE_ALL); > > - if (HAS_EDRAM(dev_priv) && INTEL_GEN(dev_priv) < 9) > > + if (HAS_EDRAM(i915) && INTEL_GEN(i915) < 9) > > I915_WRITE(HSW_IDICR, I915_READ(HSW_IDICR) | IDIHASHMSK(0xf)); > > - if (IS_HASWELL(dev_priv)) > > - I915_WRITE(MI_PREDICATE_RESULT_2, IS_HSW_GT3(dev_priv) ? > > + if (IS_HASWELL(i915)) > > + I915_WRITE(MI_PREDICATE_RESULT_2, IS_HSW_GT3(i915) ? > > LOWER_SLICE_ENABLED : LOWER_SLICE_DISABLED); > > /* Apply the GT workarounds... */ > > - intel_gt_apply_workarounds(dev_priv); > > + intel_gt_apply_workarounds(i915); > > /* ...and determine whether they are sticking. */ > > - intel_gt_verify_workarounds(dev_priv, "init"); > > + intel_gt_verify_workarounds(i915, "init"); > > - i915_gem_init_swizzling(dev_priv); > > + i915_gem_init_swizzling(i915); > > /* > > * At least 830 can leave some of the unused rings > > @@ -1279,48 +1279,48 @@ int i915_gem_init_hw(struct drm_i915_private *dev_priv) > > * will prevent c3 entry. Makes sure all unused rings > > * are totally idle. > > */ > > - init_unused_rings(dev_priv); > > + init_unused_rings(i915); > > - BUG_ON(!dev_priv->kernel_context); > > - ret = i915_terminally_wedged(dev_priv); > > + BUG_ON(!i915->kernel_context); > > + ret = i915_terminally_wedged(i915); > > if (ret) > > goto out; > > - ret = i915_ppgtt_init_hw(dev_priv); > > + ret = i915_ppgtt_init_hw(i915); > > if (ret) { > > DRM_ERROR("Enabling PPGTT failed (%d)\n", ret); > > goto out; > > } > > - ret = intel_wopcm_init_hw(&dev_priv->wopcm); > > + ret = intel_wopcm_init_hw(&i915->wopcm); > > if (ret) { > > DRM_ERROR("Enabling WOPCM failed (%d)\n", ret); > > goto out; > > } > > /* We can't enable contexts until all firmware is loaded */ > > - ret = intel_uc_init_hw(dev_priv); > > + ret = intel_uc_init_hw(i915); > > if (ret) { > > DRM_ERROR("Enabling uc failed (%d)\n", ret); > > goto out; > > } > > - intel_mocs_init_l3cc_table(dev_priv); > > + intel_mocs_init_l3cc_table(i915); > > /* Only when the HW is re-initialised, can we replay the requests */ > > - ret = intel_engines_resume(dev_priv); > > + ret = intel_engines_resume(i915); > > if (ret) > > goto cleanup_uc; > > - intel_uncore_forcewake_put(&dev_priv->uncore, FORCEWAKE_ALL); > > + intel_uncore_forcewake_put(&i915->uncore, FORCEWAKE_ALL); > > - intel_engines_set_scheduler_caps(dev_priv); > > + intel_engines_set_scheduler_caps(i915); > > return 0; > > cleanup_uc: > > - intel_uc_fini_hw(dev_priv); > > + intel_uc_fini_hw(i915); > > out: > > - intel_uncore_forcewake_put(&dev_priv->uncore, FORCEWAKE_ALL); > > + intel_uncore_forcewake_put(&i915->uncore, FORCEWAKE_ALL); > > return ret; > > } > > @@ -1505,28 +1505,28 @@ static int intel_engines_verify_workarounds(struct drm_i915_private *i915) > > return err; > > } > > -int i915_gem_init(struct drm_i915_private *dev_priv) > > +int i915_gem_init(struct drm_i915_private *i915) > > { > > int ret; > > /* We need to fallback to 4K pages if host doesn't support huge gtt. */ > > - if (intel_vgpu_active(dev_priv) && !intel_vgpu_has_huge_gtt(dev_priv)) > > - mkwrite_device_info(dev_priv)->page_sizes = > > + if (intel_vgpu_active(i915) && !intel_vgpu_has_huge_gtt(i915)) > > + mkwrite_device_info(i915)->page_sizes = > > I915_GTT_PAGE_SIZE_4K; > > - dev_priv->mm.unordered_timeline = dma_fence_context_alloc(1); > > + i915->mm.unordered_timeline = dma_fence_context_alloc(1); > > - i915_timelines_init(dev_priv); > > + i915_timelines_init(i915); > > - ret = i915_gem_init_userptr(dev_priv); > > + ret = i915_gem_init_userptr(i915); > > if (ret) > > return ret; > > - ret = intel_uc_init_misc(dev_priv); > > + ret = intel_uc_init_misc(i915); > > if (ret) > > return ret; > > - ret = intel_wopcm_init(&dev_priv->wopcm); > > + ret = intel_wopcm_init(&i915->wopcm); > > if (ret) > > goto err_uc_misc; > > @@ -1536,47 +1536,47 @@ int i915_gem_init(struct drm_i915_private *dev_priv) > > * we hold the forcewake during initialisation these problems > > * just magically go away. > > */ > > - mutex_lock(&dev_priv->drm.struct_mutex); > > - intel_uncore_forcewake_get(&dev_priv->uncore, FORCEWAKE_ALL); > > + mutex_lock(&i915->drm.struct_mutex); > > + intel_uncore_forcewake_get(&i915->uncore, FORCEWAKE_ALL); > > - ret = i915_gem_init_ggtt(dev_priv); > > + ret = i915_gem_init_ggtt(i915); > > if (ret) { > > GEM_BUG_ON(ret == -EIO); > > goto err_unlock; > > } > > - ret = i915_gem_init_scratch(dev_priv, > > - IS_GEN(dev_priv, 2) ? SZ_256K : PAGE_SIZE); > > + ret = i915_gem_init_scratch(i915, > > + IS_GEN(i915, 2) ? SZ_256K : PAGE_SIZE); > > if (ret) { > > GEM_BUG_ON(ret == -EIO); > > goto err_ggtt; > > } > > - ret = intel_engines_setup(dev_priv); > > + ret = intel_engines_setup(i915); > > if (ret) { > > GEM_BUG_ON(ret == -EIO); > > goto err_unlock; > > } > > - ret = i915_gem_contexts_init(dev_priv); > > + ret = i915_gem_contexts_init(i915); > > if (ret) { > > GEM_BUG_ON(ret == -EIO); > > goto err_scratch; > > } > > - ret = intel_engines_init(dev_priv); > > + ret = intel_engines_init(i915); > > if (ret) { > > GEM_BUG_ON(ret == -EIO); > > goto err_context; > > } > > - intel_init_gt_powersave(dev_priv); > > + intel_init_gt_powersave(i915); > > - ret = intel_uc_init(dev_priv); > > + ret = intel_uc_init(i915); > > if (ret) > > goto err_pm; > > - ret = i915_gem_init_hw(dev_priv); > > + ret = i915_gem_init_hw(i915); > > if (ret) > > goto err_uc_init; > > @@ -1589,13 +1589,13 @@ int i915_gem_init(struct drm_i915_private *dev_priv) > > * > > * FIXME: break up the workarounds and apply them at the right time! > > */ > > - intel_init_clock_gating(dev_priv); > > + intel_init_clock_gating(i915); > > - ret = intel_engines_verify_workarounds(dev_priv); > > + ret = intel_engines_verify_workarounds(i915); > > if (ret) > > goto err_init_hw; > > - ret = __intel_engines_record_defaults(dev_priv); > > + ret = __intel_engines_record_defaults(i915); > > if (ret) > > goto err_init_hw; > > @@ -1609,8 +1609,8 @@ int i915_gem_init(struct drm_i915_private *dev_priv) > > goto err_init_hw; > > } > > - intel_uncore_forcewake_put(&dev_priv->uncore, FORCEWAKE_ALL); > > - mutex_unlock(&dev_priv->drm.struct_mutex); > > + intel_uncore_forcewake_put(&i915->uncore, FORCEWAKE_ALL); > > + mutex_unlock(&i915->drm.struct_mutex); > > return 0; > > @@ -1621,107 +1621,107 @@ int i915_gem_init(struct drm_i915_private *dev_priv) > > * driver doesn't explode during runtime. > > */ > > err_init_hw: > > - mutex_unlock(&dev_priv->drm.struct_mutex); > > + mutex_unlock(&i915->drm.struct_mutex); > > - i915_gem_set_wedged(dev_priv); > > - i915_gem_suspend(dev_priv); > > - i915_gem_suspend_late(dev_priv); > > + i915_gem_set_wedged(i915); > > + i915_gem_suspend(i915); > > + i915_gem_suspend_late(i915); > > - i915_gem_drain_workqueue(dev_priv); > > + i915_gem_drain_workqueue(i915); > > - mutex_lock(&dev_priv->drm.struct_mutex); > > - intel_uc_fini_hw(dev_priv); > > + mutex_lock(&i915->drm.struct_mutex); > > + intel_uc_fini_hw(i915); > > err_uc_init: > > - intel_uc_fini(dev_priv); > > + intel_uc_fini(i915); > > err_pm: > > if (ret != -EIO) { > > - intel_cleanup_gt_powersave(dev_priv); > > - intel_engines_cleanup(dev_priv); > > + intel_cleanup_gt_powersave(i915); > > + intel_engines_cleanup(i915); > > } > > err_context: > > if (ret != -EIO) > > - i915_gem_contexts_fini(dev_priv); > > + i915_gem_contexts_fini(i915); > > err_scratch: > > - i915_gem_fini_scratch(dev_priv); > > + i915_gem_fini_scratch(i915); > > err_ggtt: > > err_unlock: > > - intel_uncore_forcewake_put(&dev_priv->uncore, FORCEWAKE_ALL); > > - mutex_unlock(&dev_priv->drm.struct_mutex); > > + intel_uncore_forcewake_put(&i915->uncore, FORCEWAKE_ALL); > > + mutex_unlock(&i915->drm.struct_mutex); > > err_uc_misc: > > - intel_uc_fini_misc(dev_priv); > > + intel_uc_fini_misc(i915); > > if (ret != -EIO) { > > - i915_gem_cleanup_userptr(dev_priv); > > - i915_timelines_fini(dev_priv); > > + i915_gem_cleanup_userptr(i915); > > + i915_timelines_fini(i915); > > } > > if (ret == -EIO) { > > - mutex_lock(&dev_priv->drm.struct_mutex); > > + mutex_lock(&i915->drm.struct_mutex); > > /* > > * Allow engine initialisation to fail by marking the GPU as > > * wedged. But we only want to do this where the GPU is angry, > > * for all other failure, such as an allocation failure, bail. > > */ > > - if (!i915_reset_failed(dev_priv)) { > > - i915_load_error(dev_priv, > > + if (!i915_reset_failed(i915)) { > > + i915_load_error(i915, > > "Failed to initialize GPU, declaring it wedged!\n"); > > - i915_gem_set_wedged(dev_priv); > > + i915_gem_set_wedged(i915); > > } > > /* Minimal basic recovery for KMS */ > > - ret = i915_ggtt_enable_hw(dev_priv); > > - i915_gem_restore_gtt_mappings(dev_priv); > > - i915_gem_restore_fences(dev_priv); > > - intel_init_clock_gating(dev_priv); > > + ret = i915_ggtt_enable_hw(i915); > > + i915_gem_restore_gtt_mappings(i915); > > + i915_gem_restore_fences(i915); > > + intel_init_clock_gating(i915); > > - mutex_unlock(&dev_priv->drm.struct_mutex); > > + mutex_unlock(&i915->drm.struct_mutex); > > } > > - i915_gem_drain_freed_objects(dev_priv); > > + i915_gem_drain_freed_objects(i915); > > return ret; > > } > > -void i915_gem_fini_hw(struct drm_i915_private *dev_priv) > > +void i915_gem_fini_hw(struct drm_i915_private *i915) > > { > > - GEM_BUG_ON(dev_priv->gt.awake); > > + GEM_BUG_ON(i915->gt.awake); > > - intel_wakeref_auto_fini(&dev_priv->mm.userfault_wakeref); > > + intel_wakeref_auto_fini(&i915->mm.userfault_wakeref); > > - i915_gem_suspend_late(dev_priv); > > - intel_disable_gt_powersave(dev_priv); > > + i915_gem_suspend_late(i915); > > + intel_disable_gt_powersave(i915); > > /* Flush any outstanding unpin_work. */ > > - i915_gem_drain_workqueue(dev_priv); > > + i915_gem_drain_workqueue(i915); > > - mutex_lock(&dev_priv->drm.struct_mutex); > > - intel_uc_fini_hw(dev_priv); > > - intel_uc_fini(dev_priv); > > - mutex_unlock(&dev_priv->drm.struct_mutex); > > + mutex_lock(&i915->drm.struct_mutex); > > + intel_uc_fini_hw(i915); > > + intel_uc_fini(i915); > > + mutex_unlock(&i915->drm.struct_mutex); > > - i915_gem_drain_freed_objects(dev_priv); > > + i915_gem_drain_freed_objects(i915); > > } > > -void i915_gem_fini(struct drm_i915_private *dev_priv) > > +void i915_gem_fini(struct drm_i915_private *i915) > > { > > - mutex_lock(&dev_priv->drm.struct_mutex); > > - intel_engines_cleanup(dev_priv); > > - i915_gem_contexts_fini(dev_priv); > > - i915_gem_fini_scratch(dev_priv); > > - mutex_unlock(&dev_priv->drm.struct_mutex); > > + mutex_lock(&i915->drm.struct_mutex); > > + intel_engines_cleanup(i915); > > + i915_gem_contexts_fini(i915); > > + i915_gem_fini_scratch(i915); > > + mutex_unlock(&i915->drm.struct_mutex); > > - intel_wa_list_free(&dev_priv->gt_wa_list); > > + intel_wa_list_free(&i915->gt_wa_list); > > - intel_cleanup_gt_powersave(dev_priv); > > + intel_cleanup_gt_powersave(i915); > > - intel_uc_fini_misc(dev_priv); > > - i915_gem_cleanup_userptr(dev_priv); > > - i915_timelines_fini(dev_priv); > > + intel_uc_fini_misc(i915); > > + i915_gem_cleanup_userptr(i915); > > + i915_timelines_fini(i915); > > - i915_gem_drain_freed_objects(dev_priv); > > + i915_gem_drain_freed_objects(i915); > > - WARN_ON(!list_empty(&dev_priv->contexts.list)); > > + WARN_ON(!list_empty(&i915->contexts.list)); > > } > > void i915_gem_init_mmio(struct drm_i915_private *i915) > > @@ -1730,35 +1730,35 @@ void i915_gem_init_mmio(struct drm_i915_private *i915) > > } > > void > > -i915_gem_load_init_fences(struct drm_i915_private *dev_priv) > > +i915_gem_load_init_fences(struct drm_i915_private *i915) > > { > > int i; > > - if (INTEL_GEN(dev_priv) >= 7 && !IS_VALLEYVIEW(dev_priv) && > > - !IS_CHERRYVIEW(dev_priv)) > > - dev_priv->num_fence_regs = 32; > > - else if (INTEL_GEN(dev_priv) >= 4 || > > - IS_I945G(dev_priv) || IS_I945GM(dev_priv) || > > - IS_G33(dev_priv) || IS_PINEVIEW(dev_priv)) > > - dev_priv->num_fence_regs = 16; > > + if (INTEL_GEN(i915) >= 7 && !IS_VALLEYVIEW(i915) && > > + !IS_CHERRYVIEW(i915)) > > + i915->num_fence_regs = 32; > > + else if (INTEL_GEN(i915) >= 4 || > > + IS_I945G(i915) || IS_I945GM(i915) || > > + IS_G33(i915) || IS_PINEVIEW(i915)) > > + i915->num_fence_regs = 16; > > else > > - dev_priv->num_fence_regs = 8; > > + i915->num_fence_regs = 8; > > - if (intel_vgpu_active(dev_priv)) > > - dev_priv->num_fence_regs = > > + if (intel_vgpu_active(i915)) > > + i915->num_fence_regs = > > I915_READ(vgtif_reg(avail_rs.fence_num)); > > /* Initialize fence registers to zero */ > > - for (i = 0; i < dev_priv->num_fence_regs; i++) { > > - struct drm_i915_fence_reg *fence = &dev_priv->fence_regs[i]; > > + for (i = 0; i < i915->num_fence_regs; i++) { > > + struct drm_i915_fence_reg *fence = &i915->fence_regs[i]; > > - fence->i915 = dev_priv; > > + fence->i915 = i915; > > fence->id = i; > > - list_add_tail(&fence->link, &dev_priv->mm.fence_list); > > + list_add_tail(&fence->link, &i915->mm.fence_list); > > } > > - i915_gem_restore_fences(dev_priv); > > + i915_gem_restore_fences(i915); > > - i915_gem_detect_bit_6_swizzle(dev_priv); > > + i915_gem_detect_bit_6_swizzle(i915); > > } > > static void i915_gem_init__mm(struct drm_i915_private *i915) > > @@ -1778,56 +1778,56 @@ static void i915_gem_init__mm(struct drm_i915_private *i915) > > i915_gem_init__objects(i915); > > } > > -int i915_gem_init_early(struct drm_i915_private *dev_priv) > > +int i915_gem_init_early(struct drm_i915_private *i915) > > { > > static struct lock_class_key reset_key; > > int err; > > - intel_gt_pm_init(dev_priv); > > + intel_gt_pm_init(i915); > > - INIT_LIST_HEAD(&dev_priv->gt.active_rings); > > - INIT_LIST_HEAD(&dev_priv->gt.closed_vma); > > - spin_lock_init(&dev_priv->gt.closed_lock); > > - lockdep_init_map(&dev_priv->gt.reset_lockmap, > > + INIT_LIST_HEAD(&i915->gt.active_rings); > > + INIT_LIST_HEAD(&i915->gt.closed_vma); > > + spin_lock_init(&i915->gt.closed_lock); > > + lockdep_init_map(&i915->gt.reset_lockmap, > > "i915.reset", &reset_key, 0); > > - i915_gem_init__mm(dev_priv); > > - i915_gem_init__pm(dev_priv); > > + i915_gem_init__mm(i915); > > + i915_gem_init__pm(i915); > > - init_waitqueue_head(&dev_priv->gpu_error.wait_queue); > > - init_waitqueue_head(&dev_priv->gpu_error.reset_queue); > > - mutex_init(&dev_priv->gpu_error.wedge_mutex); > > - init_srcu_struct(&dev_priv->gpu_error.reset_backoff_srcu); > > + init_waitqueue_head(&i915->gpu_error.wait_queue); > > + init_waitqueue_head(&i915->gpu_error.reset_queue); > > + mutex_init(&i915->gpu_error.wedge_mutex); > > + init_srcu_struct(&i915->gpu_error.reset_backoff_srcu); > > - atomic_set(&dev_priv->mm.bsd_engine_dispatch_index, 0); > > + atomic_set(&i915->mm.bsd_engine_dispatch_index, 0); > > - spin_lock_init(&dev_priv->fb_tracking.lock); > > + spin_lock_init(&i915->fb_tracking.lock); > > - err = i915_gemfs_init(dev_priv); > > + err = i915_gemfs_init(i915); > > if (err) > > DRM_NOTE("Unable to create a private tmpfs mount, hugepage support will be disabled(%d).\n", err); > > return 0; > > } > > -void i915_gem_cleanup_early(struct drm_i915_private *dev_priv) > > +void i915_gem_cleanup_early(struct drm_i915_private *i915) > > { > > - i915_gem_drain_freed_objects(dev_priv); > > - GEM_BUG_ON(!llist_empty(&dev_priv->mm.free_list)); > > - GEM_BUG_ON(atomic_read(&dev_priv->mm.free_count)); > > - WARN_ON(dev_priv->mm.shrink_count); > > + i915_gem_drain_freed_objects(i915); > > + GEM_BUG_ON(!llist_empty(&i915->mm.free_list)); > > + GEM_BUG_ON(atomic_read(&i915->mm.free_count)); > > + WARN_ON(i915->mm.shrink_count); > > - cleanup_srcu_struct(&dev_priv->gpu_error.reset_backoff_srcu); > > + cleanup_srcu_struct(&i915->gpu_error.reset_backoff_srcu); > > - i915_gemfs_fini(dev_priv); > > + i915_gemfs_fini(i915); > > } > > -int i915_gem_freeze(struct drm_i915_private *dev_priv) > > +int i915_gem_freeze(struct drm_i915_private *i915) > > { > > /* Discard all purgeable objects, let userspace recover those as > > * required after resuming. > > */ > > - i915_gem_shrink_all(dev_priv); > > + i915_gem_shrink_all(i915); > > return 0; > > } > > @@ -1895,7 +1895,7 @@ int i915_gem_open(struct drm_i915_private *i915, struct drm_file *file) > > return -ENOMEM; > > file->driver_priv = file_priv; > > - file_priv->dev_priv = i915; > > + file_priv->i915 = i915; > > file_priv->file = file; > > spin_lock_init(&file_priv->mm.lock); > > diff --git a/drivers/gpu/drm/i915/i915_gem_evict.c b/drivers/gpu/drm/i915/i915_gem_evict.c > > index a5783c4cb98b..4244d9c54d66 100644 > > --- a/drivers/gpu/drm/i915/i915_gem_evict.c > > +++ b/drivers/gpu/drm/i915/i915_gem_evict.c > > @@ -99,7 +99,7 @@ i915_gem_evict_something(struct i915_address_space *vm, > > u64 start, u64 end, > > unsigned flags) > > { > > - struct drm_i915_private *dev_priv = vm->i915; > > + struct drm_i915_private *i915 = vm->i915; > > struct drm_mm_scan scan; > > struct list_head eviction_list; > > struct i915_vma *vma, *next; > > @@ -138,7 +138,7 @@ i915_gem_evict_something(struct i915_address_space *vm, > > * retiring. > > */ > > if (!(flags & PIN_NONBLOCK)) > > - i915_retire_requests(dev_priv); > > + i915_retire_requests(i915); > > search_again: > > active = NULL; > > @@ -211,7 +211,7 @@ i915_gem_evict_something(struct i915_address_space *vm, > > if (I915_SELFTEST_ONLY(igt_evict_ctl.fail_if_busy)) > > return -EBUSY; > > - ret = ggtt_flush(dev_priv); > > + ret = ggtt_flush(i915); > > if (ret) > > return ret; > > diff --git a/drivers/gpu/drm/i915/i915_gem_gtt.c b/drivers/gpu/drm/i915/i915_gem_gtt.c > > index e70675bfb51d..42d88dae9c5d 100644 > > --- a/drivers/gpu/drm/i915/i915_gem_gtt.c > > +++ b/drivers/gpu/drm/i915/i915_gem_gtt.c > > @@ -1294,7 +1294,7 @@ static int gen8_init_scratch(struct i915_address_space *vm) > > static int gen8_ppgtt_notify_vgt(struct i915_ppgtt *ppgtt, bool create) > > { > > struct i915_address_space *vm = &ppgtt->vm; > > - struct drm_i915_private *dev_priv = vm->i915; > > + struct drm_i915_private *i915 = vm->i915; > > enum vgt_g2v_type msg; > > int i; > > @@ -1692,7 +1692,7 @@ static inline void gen6_write_pde(const struct gen6_ppgtt *ppgtt, > > ppgtt->pd_addr + pde); > > } > > -static void gen7_ppgtt_enable(struct drm_i915_private *dev_priv) > > +static void gen7_ppgtt_enable(struct drm_i915_private *i915) > > { > > struct intel_engine_cs *engine; > > u32 ecochk, ecobits; > > @@ -1702,7 +1702,7 @@ static void gen7_ppgtt_enable(struct drm_i915_private *dev_priv) > > I915_WRITE(GAC_ECO_BITS, ecobits | ECOBITS_PPGTT_CACHE64B); > > ecochk = I915_READ(GAM_ECOCHK); > > - if (IS_HASWELL(dev_priv)) { > > + if (IS_HASWELL(i915)) { > > ecochk |= ECOCHK_PPGTT_WB_HSW; > > } else { > > ecochk |= ECOCHK_PPGTT_LLC_IVB; > > @@ -1710,7 +1710,7 @@ static void gen7_ppgtt_enable(struct drm_i915_private *dev_priv) > > } > > I915_WRITE(GAM_ECOCHK, ecochk); > > - for_each_engine(engine, dev_priv, id) { > > + for_each_engine(engine, i915, id) { > > /* GFX_MODE is per-ring on gen7+ */ > > ENGINE_WRITE(engine, > > RING_MODE_GEN7, > > @@ -1718,7 +1718,7 @@ static void gen7_ppgtt_enable(struct drm_i915_private *dev_priv) > > } > > } > > -static void gen6_ppgtt_enable(struct drm_i915_private *dev_priv) > > +static void gen6_ppgtt_enable(struct drm_i915_private *i915) > > { > > u32 ecochk, gab_ctl, ecobits; > > @@ -1732,7 +1732,7 @@ static void gen6_ppgtt_enable(struct drm_i915_private *dev_priv) > > ecochk = I915_READ(GAM_ECOCHK); > > I915_WRITE(GAM_ECOCHK, ecochk | ECOCHK_SNB_BIT | ECOCHK_PPGTT_CACHE64B); > > - if (HAS_PPGTT(dev_priv)) /* may be disabled for VT-d */ > > + if (HAS_PPGTT(i915)) /* may be disabled for VT-d */ > > I915_WRITE(GFX_MODE, _MASKED_BIT_ENABLE(GFX_PPGTT_ENABLE)); > > } > > @@ -2185,20 +2185,20 @@ static struct i915_ppgtt *gen6_ppgtt_create(struct drm_i915_private *i915) > > return ERR_PTR(err); > > } > > -static void gtt_write_workarounds(struct drm_i915_private *dev_priv) > > +static void gtt_write_workarounds(struct drm_i915_private *i915) > > { > > /* This function is for gtt related workarounds. This function is > > * called on driver load and after a GPU reset, so you can place > > * workarounds here even if they get overwritten by GPU reset. > > */ > > /* WaIncreaseDefaultTLBEntries:chv,bdw,skl,bxt,kbl,glk,cfl,cnl,icl */ > > - if (IS_BROADWELL(dev_priv)) > > + if (IS_BROADWELL(i915)) > > I915_WRITE(GEN8_L3_LRA_1_GPGPU, GEN8_L3_LRA_1_GPGPU_DEFAULT_VALUE_BDW); > > - else if (IS_CHERRYVIEW(dev_priv)) > > + else if (IS_CHERRYVIEW(i915)) > > I915_WRITE(GEN8_L3_LRA_1_GPGPU, GEN8_L3_LRA_1_GPGPU_DEFAULT_VALUE_CHV); > > - else if (IS_GEN9_LP(dev_priv)) > > + else if (IS_GEN9_LP(i915)) > > I915_WRITE(GEN8_L3_LRA_1_GPGPU, GEN9_L3_LRA_1_GPGPU_DEFAULT_VALUE_BXT); > > - else if (INTEL_GEN(dev_priv) >= 9) > > + else if (INTEL_GEN(i915) >= 9) > > I915_WRITE(GEN8_L3_LRA_1_GPGPU, GEN9_L3_LRA_1_GPGPU_DEFAULT_VALUE_SKL); > > /* > > @@ -2212,21 +2212,21 @@ static void gtt_write_workarounds(struct drm_i915_private *dev_priv) > > * 32K pages, but we don't currently have any support for it in our > > * driver. > > */ > > - if (HAS_PAGE_SIZES(dev_priv, I915_GTT_PAGE_SIZE_64K) && > > - INTEL_GEN(dev_priv) <= 10) > > + if (HAS_PAGE_SIZES(i915, I915_GTT_PAGE_SIZE_64K) && > > + INTEL_GEN(i915) <= 10) > > I915_WRITE(GEN8_GAMW_ECO_DEV_RW_IA, > > I915_READ(GEN8_GAMW_ECO_DEV_RW_IA) | > > GAMW_ECO_ENABLE_64K_IPS_FIELD); > > } > > -int i915_ppgtt_init_hw(struct drm_i915_private *dev_priv) > > +int i915_ppgtt_init_hw(struct drm_i915_private *i915) > > { > > - gtt_write_workarounds(dev_priv); > > + gtt_write_workarounds(i915); > > - if (IS_GEN(dev_priv, 6)) > > - gen6_ppgtt_enable(dev_priv); > > - else if (IS_GEN(dev_priv, 7)) > > - gen7_ppgtt_enable(dev_priv); > > + if (IS_GEN(i915, 6)) > > + gen6_ppgtt_enable(i915); > > + else if (IS_GEN(i915, 7)) > > + gen7_ppgtt_enable(i915); > > return 0; > > } > > @@ -2293,29 +2293,29 @@ void i915_vm_release(struct kref *kref) > > /* Certain Gen5 chipsets require require idling the GPU before > > * unmapping anything from the GTT when VT-d is enabled. > > */ > > -static bool needs_idle_maps(struct drm_i915_private *dev_priv) > > +static bool needs_idle_maps(struct drm_i915_private *i915) > > { > > /* Query intel_iommu to see if we need the workaround. Presumably that > > * was loaded first. > > */ > > - return IS_GEN(dev_priv, 5) && IS_MOBILE(dev_priv) && intel_vtd_active(); > > + return IS_GEN(i915, 5) && IS_MOBILE(i915) && intel_vtd_active(); > > } > > -void i915_gem_suspend_gtt_mappings(struct drm_i915_private *dev_priv) > > +void i915_gem_suspend_gtt_mappings(struct drm_i915_private *i915) > > { > > - struct i915_ggtt *ggtt = &dev_priv->ggtt; > > + struct i915_ggtt *ggtt = &i915->ggtt; > > /* Don't bother messing with faults pre GEN6 as we have little > > * documentation supporting that it's a good idea. > > */ > > - if (INTEL_GEN(dev_priv) < 6) > > + if (INTEL_GEN(i915) < 6) > > return; > > - i915_check_and_clear_faults(dev_priv); > > + i915_check_and_clear_faults(i915); > > ggtt->vm.clear_range(&ggtt->vm, 0, ggtt->vm.total); > > - i915_ggtt_invalidate(dev_priv); > > + i915_ggtt_invalidate(i915); > > } > > int i915_gem_gtt_prepare_pages(struct drm_i915_gem_object *obj, > > @@ -2411,7 +2411,7 @@ static void gen6_ggtt_insert_page(struct i915_address_space *vm, > > * Binds an object into the global gtt with the specified cache level. The object > > * will be accessible to the GPU via commands whose operands reference offsets > > * within the global GTT as well as accessible by the GPU through the GMADR > > - * mapped BAR (dev_priv->mm.gtt->gtt). > > + * mapped BAR (i915->mm.gtt->gtt). > > */ > > static void gen6_ggtt_insert_entries(struct i915_address_space *vm, > > struct i915_vma *vma, > > @@ -2461,7 +2461,7 @@ static void gen8_ggtt_clear_range(struct i915_address_space *vm, > > static void bxt_vtd_ggtt_wa(struct i915_address_space *vm) > > { > > - struct drm_i915_private *dev_priv = vm->i915; > > + struct drm_i915_private *i915 = vm->i915; > > /* > > * Make sure the internal GAM fifo has been cleared of all GTT > > @@ -2705,12 +2705,12 @@ static void aliasing_gtt_unbind_vma(struct i915_vma *vma) > > void i915_gem_gtt_finish_pages(struct drm_i915_gem_object *obj, > > struct sg_table *pages) > > { > > - struct drm_i915_private *dev_priv = to_i915(obj->base.dev); > > - struct device *kdev = &dev_priv->drm.pdev->dev; > > - struct i915_ggtt *ggtt = &dev_priv->ggtt; > > + struct drm_i915_private *i915 = to_i915(obj->base.dev); > > + struct device *kdev = &i915->drm.pdev->dev; > > + struct i915_ggtt *ggtt = &i915->ggtt; > > if (unlikely(ggtt->do_idle_maps)) { > > - if (i915_gem_wait_for_idle(dev_priv, 0, MAX_SCHEDULE_TIMEOUT)) { > > + if (i915_gem_wait_for_idle(i915, 0, MAX_SCHEDULE_TIMEOUT)) { > > DRM_ERROR("Failed to wait for idle; VT'd may hang.\n"); > > /* Wait a bit, in hopes it avoids the hang */ > > udelay(10); > > @@ -2834,7 +2834,7 @@ static void ggtt_release_guc_top(struct i915_ggtt *ggtt) > > drm_mm_remove_node(&ggtt->uc_fw); > > } > > -int i915_gem_init_ggtt(struct drm_i915_private *dev_priv) > > +int i915_gem_init_ggtt(struct drm_i915_private *i915) > > { > > /* Let GEM Manage all of the aperture. > > * > > @@ -2845,7 +2845,7 @@ int i915_gem_init_ggtt(struct drm_i915_private *dev_priv) > > * aperture. One page should be enough to keep any prefetching inside > > * of the aperture. > > */ > > - struct i915_ggtt *ggtt = &dev_priv->ggtt; > > + struct i915_ggtt *ggtt = &i915->ggtt; > > unsigned long hole_start, hole_end; > > struct drm_mm_node *entry; > > int ret; > > @@ -2857,9 +2857,9 @@ int i915_gem_init_ggtt(struct drm_i915_private *dev_priv) > > * why. > > */ > > ggtt->pin_bias = max_t(u32, I915_GTT_PAGE_SIZE, > > - intel_wopcm_guc_size(&dev_priv->wopcm)); > > + intel_wopcm_guc_size(&i915->wopcm)); > > - ret = intel_vgt_balloon(dev_priv); > > + ret = intel_vgt_balloon(i915); > > if (ret) > > return ret; > > @@ -2891,8 +2891,8 @@ int i915_gem_init_ggtt(struct drm_i915_private *dev_priv) > > /* And finally clear the reserved guard page */ > > ggtt->vm.clear_range(&ggtt->vm, ggtt->vm.total - PAGE_SIZE, PAGE_SIZE); > > - if (INTEL_PPGTT(dev_priv) == INTEL_PPGTT_ALIASING) { > > - ret = init_aliasing_ppgtt(dev_priv); > > + if (INTEL_PPGTT(i915) == INTEL_PPGTT_ALIASING) { > > + ret = init_aliasing_ppgtt(i915); > > if (ret) > > goto err_appgtt; > > } > > @@ -2908,18 +2908,18 @@ int i915_gem_init_ggtt(struct drm_i915_private *dev_priv) > > /** > > * i915_ggtt_cleanup_hw - Clean up GGTT hardware initialization > > - * @dev_priv: i915 device > > + * @i915: i915 device > > */ > > -void i915_ggtt_cleanup_hw(struct drm_i915_private *dev_priv) > > +void i915_ggtt_cleanup_hw(struct drm_i915_private *i915) > > { > > - struct i915_ggtt *ggtt = &dev_priv->ggtt; > > + struct i915_ggtt *ggtt = &i915->ggtt; > > struct i915_vma *vma, *vn; > > struct pagevec *pvec; > > ggtt->vm.closed = true; > > - mutex_lock(&dev_priv->drm.struct_mutex); > > - fini_aliasing_ppgtt(dev_priv); > > + mutex_lock(&i915->drm.struct_mutex); > > + fini_aliasing_ppgtt(i915); > > list_for_each_entry_safe(vma, vn, &ggtt->vm.bound_list, vm_link) > > WARN_ON(i915_vma_unbind(vma)); > > @@ -2930,24 +2930,24 @@ void i915_ggtt_cleanup_hw(struct drm_i915_private *dev_priv) > > ggtt_release_guc_top(ggtt); > > if (drm_mm_initialized(&ggtt->vm.mm)) { > > - intel_vgt_deballoon(dev_priv); > > + intel_vgt_deballoon(i915); > > i915_address_space_fini(&ggtt->vm); > > } > > ggtt->vm.cleanup(&ggtt->vm); > > - pvec = &dev_priv->mm.wc_stash.pvec; > > + pvec = &i915->mm.wc_stash.pvec; > > if (pvec->nr) { > > set_pages_array_wb(pvec->pages, pvec->nr); > > __pagevec_release(pvec); > > } > > - mutex_unlock(&dev_priv->drm.struct_mutex); > > + mutex_unlock(&i915->drm.struct_mutex); > > arch_phys_wc_del(ggtt->mtrr); > > io_mapping_fini(&ggtt->iomap); > > - i915_gem_cleanup_stolen(dev_priv); > > + i915_gem_cleanup_stolen(i915); > > } > > static unsigned int gen6_get_total_gtt_size(u16 snb_gmch_ctl) > > @@ -2986,8 +2986,8 @@ static unsigned int chv_get_total_gtt_size(u16 gmch_ctrl) > > static int ggtt_probe_common(struct i915_ggtt *ggtt, u64 size) > > { > > - struct drm_i915_private *dev_priv = ggtt->vm.i915; > > - struct pci_dev *pdev = dev_priv->drm.pdev; > > + struct drm_i915_private *i915 = ggtt->vm.i915; > > + struct pci_dev *pdev = i915->drm.pdev; > > phys_addr_t phys_addr; > > int ret; > > @@ -3001,7 +3001,7 @@ static int ggtt_probe_common(struct i915_ggtt *ggtt, u64 size) > > * resort to an uncached mapping. The WC issue is easily caught by the > > * readback check when writing GTT PTE entries. > > */ > > - if (IS_GEN9_LP(dev_priv) || INTEL_GEN(dev_priv) >= 10) > > + if (IS_GEN9_LP(i915) || INTEL_GEN(i915) >= 10) > > ggtt->gsm = ioremap_nocache(phys_addr, size); > > else > > ggtt->gsm = ioremap_wc(phys_addr, size); > > @@ -3135,9 +3135,9 @@ void intel_ppat_put(const struct intel_ppat_entry *entry) > > kref_put(&ppat->entries[index].ref, release_ppat); > > } > > -static void cnl_private_pat_update_hw(struct drm_i915_private *dev_priv) > > +static void cnl_private_pat_update_hw(struct drm_i915_private *i915) > > { > > - struct intel_ppat *ppat = &dev_priv->ppat; > > + struct intel_ppat *ppat = &i915->ppat; > > int i; > > for_each_set_bit(i, ppat->dirty, ppat->max_entries) { > > @@ -3146,9 +3146,9 @@ static void cnl_private_pat_update_hw(struct drm_i915_private *dev_priv) > > } > > } > > -static void bdw_private_pat_update_hw(struct drm_i915_private *dev_priv) > > +static void bdw_private_pat_update_hw(struct drm_i915_private *i915) > > { > > - struct intel_ppat *ppat = &dev_priv->ppat; > > + struct intel_ppat *ppat = &i915->ppat; > > u64 pat = 0; > > int i; > > @@ -3293,16 +3293,16 @@ static void gen6_gmch_remove(struct i915_address_space *vm) > > cleanup_scratch_page(vm); > > } > > -static void setup_private_pat(struct drm_i915_private *dev_priv) > > +static void setup_private_pat(struct drm_i915_private *i915) > > { > > - struct intel_ppat *ppat = &dev_priv->ppat; > > + struct intel_ppat *ppat = &i915->ppat; > > int i; > > - ppat->i915 = dev_priv; > > + ppat->i915 = i915; > > - if (INTEL_GEN(dev_priv) >= 10) > > + if (INTEL_GEN(i915) >= 10) > > cnl_setup_private_ppat(ppat); > > - else if (IS_CHERRYVIEW(dev_priv) || IS_GEN9_LP(dev_priv)) > > + else if (IS_CHERRYVIEW(i915) || IS_GEN9_LP(i915)) > > chv_setup_private_ppat(ppat); > > else > > bdw_setup_private_ppat(ppat); > > @@ -3315,13 +3315,13 @@ static void setup_private_pat(struct drm_i915_private *dev_priv) > > set_bit(i, ppat->dirty); > > } > > - ppat->update_hw(dev_priv); > > + ppat->update_hw(i915); > > } > > static int gen8_gmch_probe(struct i915_ggtt *ggtt) > > { > > - struct drm_i915_private *dev_priv = ggtt->vm.i915; > > - struct pci_dev *pdev = dev_priv->drm.pdev; > > + struct drm_i915_private *i915 = ggtt->vm.i915; > > + struct pci_dev *pdev = i915->drm.pdev; > > unsigned int size; > > u16 snb_gmch_ctl; > > int err; > > @@ -3339,7 +3339,7 @@ static int gen8_gmch_probe(struct i915_ggtt *ggtt) > > DRM_ERROR("Can't set DMA mask/consistent mask (%d)\n", err); > > pci_read_config_word(pdev, SNB_GMCH_CTRL, &snb_gmch_ctl); > > - if (IS_CHERRYVIEW(dev_priv)) > > + if (IS_CHERRYVIEW(i915)) > > size = chv_get_total_gtt_size(snb_gmch_ctl); > > else > > size = gen8_get_total_gtt_size(snb_gmch_ctl); > > @@ -3348,23 +3348,23 @@ static int gen8_gmch_probe(struct i915_ggtt *ggtt) > > ggtt->vm.cleanup = gen6_gmch_remove; > > ggtt->vm.insert_page = gen8_ggtt_insert_page; > > ggtt->vm.clear_range = nop_clear_range; > > - if (intel_scanout_needs_vtd_wa(dev_priv)) > > + if (intel_scanout_needs_vtd_wa(i915)) > > ggtt->vm.clear_range = gen8_ggtt_clear_range; > > ggtt->vm.insert_entries = gen8_ggtt_insert_entries; > > /* Serialize GTT updates with aperture access on BXT if VT-d is on. */ > > - if (intel_ggtt_update_needs_vtd_wa(dev_priv) || > > - IS_CHERRYVIEW(dev_priv) /* fails with concurrent use/update */) { > > + if (intel_ggtt_update_needs_vtd_wa(i915) || > > + IS_CHERRYVIEW(i915) /* fails with concurrent use/update */) { > > ggtt->vm.insert_entries = bxt_vtd_ggtt_insert_entries__BKL; > > ggtt->vm.insert_page = bxt_vtd_ggtt_insert_page__BKL; > > if (ggtt->vm.clear_range != nop_clear_range) > > ggtt->vm.clear_range = bxt_vtd_ggtt_clear_range__BKL; > > /* Prevent recursively calling stop_machine() and deadlocks. */ > > - dev_info(dev_priv->drm.dev, > > + dev_info(i915->drm.dev, > > "Disabling error capture for VT-d workaround\n"); > > - i915_disable_error_state(dev_priv, -ENODEV); > > + i915_disable_error_state(i915, -ENODEV); > > } > > ggtt->invalidate = gen6_ggtt_invalidate; > > @@ -3376,15 +3376,15 @@ static int gen8_gmch_probe(struct i915_ggtt *ggtt) > > ggtt->vm.pte_encode = gen8_pte_encode; > > - setup_private_pat(dev_priv); > > + setup_private_pat(i915); > > return ggtt_probe_common(ggtt, size); > > } > > static int gen6_gmch_probe(struct i915_ggtt *ggtt) > > { > > - struct drm_i915_private *dev_priv = ggtt->vm.i915; > > - struct pci_dev *pdev = dev_priv->drm.pdev; > > + struct drm_i915_private *i915 = ggtt->vm.i915; > > + struct pci_dev *pdev = i915->drm.pdev; > > unsigned int size; > > u16 snb_gmch_ctl; > > int err; > > @@ -3413,7 +3413,7 @@ static int gen6_gmch_probe(struct i915_ggtt *ggtt) > > ggtt->vm.total = (size / sizeof(gen6_pte_t)) * I915_GTT_PAGE_SIZE; > > ggtt->vm.clear_range = nop_clear_range; > > - if (!HAS_FULL_PPGTT(dev_priv) || intel_scanout_needs_vtd_wa(dev_priv)) > > + if (!HAS_FULL_PPGTT(i915) || intel_scanout_needs_vtd_wa(i915)) > > ggtt->vm.clear_range = gen6_ggtt_clear_range; > > ggtt->vm.insert_page = gen6_ggtt_insert_page; > > ggtt->vm.insert_entries = gen6_ggtt_insert_entries; > > @@ -3421,13 +3421,13 @@ static int gen6_gmch_probe(struct i915_ggtt *ggtt) > > ggtt->invalidate = gen6_ggtt_invalidate; > > - if (HAS_EDRAM(dev_priv)) > > + if (HAS_EDRAM(i915)) > > ggtt->vm.pte_encode = iris_pte_encode; > > - else if (IS_HASWELL(dev_priv)) > > + else if (IS_HASWELL(i915)) > > ggtt->vm.pte_encode = hsw_pte_encode; > > - else if (IS_VALLEYVIEW(dev_priv)) > > + else if (IS_VALLEYVIEW(i915)) > > ggtt->vm.pte_encode = byt_pte_encode; > > - else if (INTEL_GEN(dev_priv) >= 7) > > + else if (INTEL_GEN(i915) >= 7) > > ggtt->vm.pte_encode = ivb_pte_encode; > > else > > ggtt->vm.pte_encode = snb_pte_encode; > > @@ -3447,11 +3447,11 @@ static void i915_gmch_remove(struct i915_address_space *vm) > > static int i915_gmch_probe(struct i915_ggtt *ggtt) > > { > > - struct drm_i915_private *dev_priv = ggtt->vm.i915; > > + struct drm_i915_private *i915 = ggtt->vm.i915; > > phys_addr_t gmadr_base; > > int ret; > > - ret = intel_gmch_probe(dev_priv->bridge_dev, dev_priv->drm.pdev, NULL); > > + ret = intel_gmch_probe(i915->bridge_dev, i915->drm.pdev, NULL); > > if (!ret) { > > DRM_ERROR("failed to set up gmch\n"); > > return -EIO; > > @@ -3463,7 +3463,7 @@ static int i915_gmch_probe(struct i915_ggtt *ggtt) > > (struct resource) DEFINE_RES_MEM(gmadr_base, > > ggtt->mappable_end); > > - ggtt->do_idle_maps = needs_idle_maps(dev_priv); > > + ggtt->do_idle_maps = needs_idle_maps(i915); > > ggtt->vm.insert_page = i915_ggtt_insert_page; > > ggtt->vm.insert_entries = i915_ggtt_insert_entries; > > ggtt->vm.clear_range = i915_ggtt_clear_range; > > @@ -3484,19 +3484,19 @@ static int i915_gmch_probe(struct i915_ggtt *ggtt) > > /** > > * i915_ggtt_probe_hw - Probe GGTT hardware location > > - * @dev_priv: i915 device > > + * @i915: i915 device > > */ > > -int i915_ggtt_probe_hw(struct drm_i915_private *dev_priv) > > +int i915_ggtt_probe_hw(struct drm_i915_private *i915) > > { > > - struct i915_ggtt *ggtt = &dev_priv->ggtt; > > + struct i915_ggtt *ggtt = &i915->ggtt; > > int ret; > > - ggtt->vm.i915 = dev_priv; > > - ggtt->vm.dma = &dev_priv->drm.pdev->dev; > > + ggtt->vm.i915 = i915; > > + ggtt->vm.dma = &i915->drm.pdev->dev; > > - if (INTEL_GEN(dev_priv) <= 5) > > + if (INTEL_GEN(i915) <= 5) > > ret = i915_gmch_probe(ggtt); > > - else if (INTEL_GEN(dev_priv) < 8) > > + else if (INTEL_GEN(i915) < 8) > > ret = gen6_gmch_probe(ggtt); > > else > > ret = gen8_gmch_probe(ggtt); > > @@ -3532,35 +3532,35 @@ int i915_ggtt_probe_hw(struct drm_i915_private *dev_priv) > > /** > > * i915_ggtt_init_hw - Initialize GGTT hardware > > - * @dev_priv: i915 device > > + * @i915: i915 device > > */ > > -int i915_ggtt_init_hw(struct drm_i915_private *dev_priv) > > +int i915_ggtt_init_hw(struct drm_i915_private *i915) > > { > > - struct i915_ggtt *ggtt = &dev_priv->ggtt; > > + struct i915_ggtt *ggtt = &i915->ggtt; > > int ret; > > - stash_init(&dev_priv->mm.wc_stash); > > + stash_init(&i915->mm.wc_stash); > > /* Note that we use page colouring to enforce a guard page at the > > * end of the address space. This is required as the CS may prefetch > > * beyond the end of the batch buffer, across the page boundary, > > * and beyond the end of the GTT if we do not provide a guard. > > */ > > - mutex_lock(&dev_priv->drm.struct_mutex); > > + mutex_lock(&i915->drm.struct_mutex); > > i915_address_space_init(&ggtt->vm, VM_CLASS_GGTT); > > ggtt->vm.is_ggtt = true; > > /* Only VLV supports read-only GGTT mappings */ > > - ggtt->vm.has_read_only = IS_VALLEYVIEW(dev_priv); > > + ggtt->vm.has_read_only = IS_VALLEYVIEW(i915); > > - if (!HAS_LLC(dev_priv) && !HAS_PPGTT(dev_priv)) > > + if (!HAS_LLC(i915) && !HAS_PPGTT(i915)) > > ggtt->vm.mm.color_adjust = i915_gtt_color_adjust; > > - mutex_unlock(&dev_priv->drm.struct_mutex); > > + mutex_unlock(&i915->drm.struct_mutex); > > - if (!io_mapping_init_wc(&dev_priv->ggtt.iomap, > > - dev_priv->ggtt.gmadr.start, > > - dev_priv->ggtt.mappable_end)) { > > + if (!io_mapping_init_wc(&i915->ggtt.iomap, > > + i915->ggtt.gmadr.start, > > + i915->ggtt.mappable_end)) { > > ret = -EIO; > > goto out_gtt_cleanup; > > } > > @@ -3571,7 +3571,7 @@ int i915_ggtt_init_hw(struct drm_i915_private *dev_priv) > > * Initialise stolen early so that we may reserve preallocated > > * objects for the BIOS to KMS transition. > > */ > > - ret = i915_gem_init_stolen(dev_priv); > > + ret = i915_gem_init_stolen(i915); > > if (ret) > > goto out_gtt_cleanup; > > @@ -3582,9 +3582,9 @@ int i915_ggtt_init_hw(struct drm_i915_private *dev_priv) > > return ret; > > } > > -int i915_ggtt_enable_hw(struct drm_i915_private *dev_priv) > > +int i915_ggtt_enable_hw(struct drm_i915_private *i915) > > { > > - if (INTEL_GEN(dev_priv) < 6 && !intel_enable_gtt()) > > + if (INTEL_GEN(i915) < 6 && !intel_enable_gtt()) > > return -EIO; > > return 0; > > @@ -3613,12 +3613,12 @@ void i915_ggtt_disable_guc(struct drm_i915_private *i915) > > i915_ggtt_invalidate(i915); > > } > > -void i915_gem_restore_gtt_mappings(struct drm_i915_private *dev_priv) > > +void i915_gem_restore_gtt_mappings(struct drm_i915_private *i915) > > { > > - struct i915_ggtt *ggtt = &dev_priv->ggtt; > > + struct i915_ggtt *ggtt = &i915->ggtt; > > struct i915_vma *vma, *vn; > > - i915_check_and_clear_faults(dev_priv); > > + i915_check_and_clear_faults(i915); > > mutex_lock(&ggtt->vm.mutex); > > @@ -3652,15 +3652,15 @@ void i915_gem_restore_gtt_mappings(struct drm_i915_private *dev_priv) > > } > > ggtt->vm.closed = false; > > - i915_ggtt_invalidate(dev_priv); > > + i915_ggtt_invalidate(i915); > > mutex_unlock(&ggtt->vm.mutex); > > - if (INTEL_GEN(dev_priv) >= 8) { > > - struct intel_ppat *ppat = &dev_priv->ppat; > > + if (INTEL_GEN(i915) >= 8) { > > + struct intel_ppat *ppat = &i915->ppat; > > bitmap_set(ppat->dirty, 0, ppat->max_entries); > > - dev_priv->ppat.update_hw(dev_priv); > > + i915->ppat.update_hw(i915); > > return; > > } > > } > > diff --git a/drivers/gpu/drm/i915/i915_gem_gtt.h b/drivers/gpu/drm/i915/i915_gem_gtt.h > > index 89437d0a721c..b444552ccc63 100644 > > --- a/drivers/gpu/drm/i915/i915_gem_gtt.h > > +++ b/drivers/gpu/drm/i915/i915_gem_gtt.h > > @@ -400,7 +400,7 @@ struct i915_ggtt { > > /** "Graphics Stolen Memory" holds the global PTEs */ > > void __iomem *gsm; > > - void (*invalidate)(struct drm_i915_private *dev_priv); > > + void (*invalidate)(struct drm_i915_private *i915); > > bool do_idle_maps; > > @@ -627,17 +627,17 @@ const struct intel_ppat_entry * > > intel_ppat_get(struct drm_i915_private *i915, u8 value); > > void intel_ppat_put(const struct intel_ppat_entry *entry); > > -int i915_ggtt_probe_hw(struct drm_i915_private *dev_priv); > > -int i915_ggtt_init_hw(struct drm_i915_private *dev_priv); > > -int i915_ggtt_enable_hw(struct drm_i915_private *dev_priv); > > +int i915_ggtt_probe_hw(struct drm_i915_private *i915); > > +int i915_ggtt_init_hw(struct drm_i915_private *i915); > > +int i915_ggtt_enable_hw(struct drm_i915_private *i915); > > void i915_ggtt_enable_guc(struct drm_i915_private *i915); > > void i915_ggtt_disable_guc(struct drm_i915_private *i915); > > -int i915_gem_init_ggtt(struct drm_i915_private *dev_priv); > > -void i915_ggtt_cleanup_hw(struct drm_i915_private *dev_priv); > > +int i915_gem_init_ggtt(struct drm_i915_private *i915); > > +void i915_ggtt_cleanup_hw(struct drm_i915_private *i915); > > -int i915_ppgtt_init_hw(struct drm_i915_private *dev_priv); > > +int i915_ppgtt_init_hw(struct drm_i915_private *i915); > > -struct i915_ppgtt *i915_ppgtt_create(struct drm_i915_private *dev_priv); > > +struct i915_ppgtt *i915_ppgtt_create(struct drm_i915_private *i915); > > static inline struct i915_address_space * > > i915_vm_get(struct i915_address_space *vm) > > @@ -657,8 +657,8 @@ int gen6_ppgtt_pin(struct i915_ppgtt *base); > > void gen6_ppgtt_unpin(struct i915_ppgtt *base); > > void gen6_ppgtt_unpin_all(struct i915_ppgtt *base); > > -void i915_gem_suspend_gtt_mappings(struct drm_i915_private *dev_priv); > > -void i915_gem_restore_gtt_mappings(struct drm_i915_private *dev_priv); > > +void i915_gem_suspend_gtt_mappings(struct drm_i915_private *i915); > > +void i915_gem_restore_gtt_mappings(struct drm_i915_private *i915); > > int __must_check i915_gem_gtt_prepare_pages(struct drm_i915_gem_object *obj, > > struct sg_table *pages); > > diff --git a/drivers/gpu/drm/i915/i915_gpu_error.c b/drivers/gpu/drm/i915/i915_gpu_error.c > > index d7f164c9f2eb..335da59031b0 100644 > > --- a/drivers/gpu/drm/i915/i915_gpu_error.c > > +++ b/drivers/gpu/drm/i915/i915_gpu_error.c > > @@ -1122,22 +1122,22 @@ static u32 i915_error_generate_code(struct i915_gpu_state *error, > > static void gem_record_fences(struct i915_gpu_state *error) > > { > > - struct drm_i915_private *dev_priv = error->i915; > > - struct intel_uncore *uncore = &dev_priv->uncore; > > + struct drm_i915_private *i915 = error->i915; > > + struct intel_uncore *uncore = &i915->uncore; > > int i; > > - if (INTEL_GEN(dev_priv) >= 6) { > > - for (i = 0; i < dev_priv->num_fence_regs; i++) > > + if (INTEL_GEN(i915) >= 6) { > > + for (i = 0; i < i915->num_fence_regs; i++) > > error->fence[i] = > > intel_uncore_read64(uncore, > > FENCE_REG_GEN6_LO(i)); > > - } else if (INTEL_GEN(dev_priv) >= 4) { > > - for (i = 0; i < dev_priv->num_fence_regs; i++) > > + } else if (INTEL_GEN(i915) >= 4) { > > + for (i = 0; i < i915->num_fence_regs; i++) > > error->fence[i] = > > intel_uncore_read64(uncore, > > FENCE_REG_965_LO(i)); > > } else { > > - for (i = 0; i < dev_priv->num_fence_regs; i++) > > + for (i = 0; i < i915->num_fence_regs; i++) > > error->fence[i] = > > intel_uncore_read(uncore, FENCE_REG(i)); > > } > > @@ -1148,23 +1148,23 @@ static void error_record_engine_registers(struct i915_gpu_state *error, > > struct intel_engine_cs *engine, > > struct drm_i915_error_engine *ee) > > { > > - struct drm_i915_private *dev_priv = engine->i915; > > + struct drm_i915_private *i915 = engine->i915; > > - if (INTEL_GEN(dev_priv) >= 6) { > > + if (INTEL_GEN(i915) >= 6) { > > ee->rc_psmi = ENGINE_READ(engine, RING_PSMI_CTL); > > - if (INTEL_GEN(dev_priv) >= 8) > > + if (INTEL_GEN(i915) >= 8) > > ee->fault_reg = I915_READ(GEN8_RING_FAULT_REG); > > else > > ee->fault_reg = GEN6_RING_FAULT_REG_READ(engine); > > } > > - if (INTEL_GEN(dev_priv) >= 4) { > > + if (INTEL_GEN(i915) >= 4) { > > ee->faddr = ENGINE_READ(engine, RING_DMA_FADD); > > ee->ipeir = ENGINE_READ(engine, RING_IPEIR); > > ee->ipehr = ENGINE_READ(engine, RING_IPEHR); > > ee->instps = ENGINE_READ(engine, RING_INSTPS); > > ee->bbaddr = ENGINE_READ(engine, RING_BBADDR); > > - if (INTEL_GEN(dev_priv) >= 8) { > > + if (INTEL_GEN(i915) >= 8) { > > ee->faddr |= (u64)ENGINE_READ(engine, RING_DMA_FADD_UDW) << 32; > > ee->bbaddr |= (u64)ENGINE_READ(engine, RING_BBADDR_UDW) << 32; > > } > > @@ -1183,13 +1183,13 @@ static void error_record_engine_registers(struct i915_gpu_state *error, > > ee->head = ENGINE_READ(engine, RING_HEAD); > > ee->tail = ENGINE_READ(engine, RING_TAIL); > > ee->ctl = ENGINE_READ(engine, RING_CTL); > > - if (INTEL_GEN(dev_priv) > 2) > > + if (INTEL_GEN(i915) > 2) > > ee->mode = ENGINE_READ(engine, RING_MI_MODE); > > - if (!HWS_NEEDS_PHYSICAL(dev_priv)) { > > + if (!HWS_NEEDS_PHYSICAL(i915)) { > > i915_reg_t mmio; > > - if (IS_GEN(dev_priv, 7)) { > > + if (IS_GEN(i915, 7)) { > > switch (engine->id) { > > default: > > MISSING_CASE(engine->id); > > @@ -1219,21 +1219,21 @@ static void error_record_engine_registers(struct i915_gpu_state *error, > > ee->idle = intel_engine_is_idle(engine); > > if (!ee->idle) > > ee->hangcheck_timestamp = engine->hangcheck.action_timestamp; > > - ee->reset_count = i915_reset_engine_count(&dev_priv->gpu_error, > > + ee->reset_count = i915_reset_engine_count(&i915->gpu_error, > > engine); > > - if (HAS_PPGTT(dev_priv)) { > > + if (HAS_PPGTT(i915)) { > > int i; > > ee->vm_info.gfx_mode = ENGINE_READ(engine, RING_MODE_GEN7); > > - if (IS_GEN(dev_priv, 6)) { > > + if (IS_GEN(i915, 6)) { > > ee->vm_info.pp_dir_base = > > ENGINE_READ(engine, RING_PP_DIR_BASE_READ); > > - } else if (IS_GEN(dev_priv, 7)) { > > + } else if (IS_GEN(i915, 7)) { > > ee->vm_info.pp_dir_base = > > ENGINE_READ(engine, RING_PP_DIR_BASE); > > - } else if (INTEL_GEN(dev_priv) >= 8) { > > + } else if (INTEL_GEN(i915) >= 8) { > > u32 base = engine->mmio_base; > > for (i = 0; i < 4; i++) { > > @@ -1388,7 +1388,7 @@ static void request_record_user_bo(struct i915_request *request, > > } > > static struct drm_i915_error_object * > > -capture_object(struct drm_i915_private *dev_priv, > > +capture_object(struct drm_i915_private *i915, > > struct drm_i915_gem_object *obj) > > { > > if (obj && i915_gem_object_has_pages(obj)) { > > @@ -1399,7 +1399,7 @@ capture_object(struct drm_i915_private *dev_priv, > > .obj = obj, > > }; > > - return i915_error_object_create(dev_priv, &fake); > > + return i915_error_object_create(i915, &fake); > > } else { > > return NULL; > > } > > diff --git a/drivers/gpu/drm/i915/i915_gpu_error.h b/drivers/gpu/drm/i915/i915_gpu_error.h > > index 2ecd0c6a1c94..980f95fb5578 100644 > > --- a/drivers/gpu/drm/i915/i915_gpu_error.h > > +++ b/drivers/gpu/drm/i915/i915_gpu_error.h > > @@ -234,7 +234,7 @@ struct i915_gpu_error { > > /** > > * Waitqueue to signal when the reset has completed. Used by clients > > - * that wait for dev_priv->mm.wedged to settle. > > + * that wait for i915->mm.wedged to settle. > > */ > > wait_queue_head_t reset_queue; > > @@ -259,7 +259,7 @@ __printf(2, 3) > > void i915_error_printf(struct drm_i915_error_state_buf *e, const char *f, ...); > > struct i915_gpu_state *i915_capture_gpu_state(struct drm_i915_private *i915); > > -void i915_capture_error_state(struct drm_i915_private *dev_priv, > > +void i915_capture_error_state(struct drm_i915_private *i915, > > intel_engine_mask_t engine_mask, > > const char *error_msg); > > @@ -286,7 +286,7 @@ void i915_disable_error_state(struct drm_i915_private *i915, int err); > > #else > > -static inline void i915_capture_error_state(struct drm_i915_private *dev_priv, > > +static inline void i915_capture_error_state(struct drm_i915_private *i915, > > u32 engine_mask, > > const char *error_msg) > > { > > diff --git a/drivers/gpu/drm/i915/i915_irq.c b/drivers/gpu/drm/i915/i915_irq.c > > index cd9edddd6718..81ea65d85b9f 100644 > > --- a/drivers/gpu/drm/i915/i915_irq.c > > +++ b/drivers/gpu/drm/i915/i915_irq.c > > @@ -256,18 +256,18 @@ static void gen2_irq_init(struct intel_uncore *uncore, > > #define GEN2_IRQ_INIT(uncore, imr_val, ier_val) \ > > gen2_irq_init((uncore), imr_val, ier_val) > > -static void gen6_rps_irq_handler(struct drm_i915_private *dev_priv, u32 pm_iir); > > -static void gen9_guc_irq_handler(struct drm_i915_private *dev_priv, u32 pm_iir); > > +static void gen6_rps_irq_handler(struct drm_i915_private *i915, u32 pm_iir); > > +static void gen9_guc_irq_handler(struct drm_i915_private *i915, u32 pm_iir); > > /* For display hotplug interrupt */ > > static inline void > > -i915_hotplug_interrupt_update_locked(struct drm_i915_private *dev_priv, > > +i915_hotplug_interrupt_update_locked(struct drm_i915_private *i915, > > u32 mask, > > u32 bits) > > { > > u32 val; > > - lockdep_assert_held(&dev_priv->irq_lock); > > + lockdep_assert_held(&i915->irq_lock); > > WARN_ON(bits & ~mask); > > val = I915_READ(PORT_HOTPLUG_EN); > > @@ -278,7 +278,7 @@ i915_hotplug_interrupt_update_locked(struct drm_i915_private *dev_priv, > > /** > > * i915_hotplug_interrupt_update - update hotplug interrupt enable > > - * @dev_priv: driver private > > + * @i915: driver private > > * @mask: bits to update > > * @bits: bits to enable > > * NOTE: the HPD enable bits are modified both inside and outside > > @@ -288,13 +288,13 @@ i915_hotplug_interrupt_update_locked(struct drm_i915_private *dev_priv, > > * held already, this function acquires the lock itself. A non-locking > > * version is also available. > > */ > > -void i915_hotplug_interrupt_update(struct drm_i915_private *dev_priv, > > +void i915_hotplug_interrupt_update(struct drm_i915_private *i915, > > u32 mask, > > u32 bits) > > { > > - spin_lock_irq(&dev_priv->irq_lock); > > - i915_hotplug_interrupt_update_locked(dev_priv, mask, bits); > > - spin_unlock_irq(&dev_priv->irq_lock); > > + spin_lock_irq(&i915->irq_lock); > > + i915_hotplug_interrupt_update_locked(i915, mask, bits); > > + spin_unlock_irq(&i915->irq_lock); > > } > > static u32 > > @@ -334,84 +334,84 @@ static bool gen11_reset_one_iir(struct drm_i915_private * const i915, > > /** > > * ilk_update_display_irq - update DEIMR > > - * @dev_priv: driver private > > + * @i915: driver private > > * @interrupt_mask: mask of interrupt bits to update > > * @enabled_irq_mask: mask of interrupt bits to enable > > */ > > -void ilk_update_display_irq(struct drm_i915_private *dev_priv, > > +void ilk_update_display_irq(struct drm_i915_private *i915, > > u32 interrupt_mask, > > u32 enabled_irq_mask) > > { > > u32 new_val; > > - lockdep_assert_held(&dev_priv->irq_lock); > > + lockdep_assert_held(&i915->irq_lock); > > WARN_ON(enabled_irq_mask & ~interrupt_mask); > > - if (WARN_ON(!intel_irqs_enabled(dev_priv))) > > + if (WARN_ON(!intel_irqs_enabled(i915))) > > return; > > - new_val = dev_priv->irq_mask; > > + new_val = i915->irq_mask; > > new_val &= ~interrupt_mask; > > new_val |= (~enabled_irq_mask & interrupt_mask); > > - if (new_val != dev_priv->irq_mask) { > > - dev_priv->irq_mask = new_val; > > - I915_WRITE(DEIMR, dev_priv->irq_mask); > > + if (new_val != i915->irq_mask) { > > + i915->irq_mask = new_val; > > + I915_WRITE(DEIMR, i915->irq_mask); > > POSTING_READ(DEIMR); > > } > > } > > /** > > * ilk_update_gt_irq - update GTIMR > > - * @dev_priv: driver private > > + * @i915: driver private > > * @interrupt_mask: mask of interrupt bits to update > > * @enabled_irq_mask: mask of interrupt bits to enable > > */ > > -static void ilk_update_gt_irq(struct drm_i915_private *dev_priv, > > +static void ilk_update_gt_irq(struct drm_i915_private *i915, > > u32 interrupt_mask, > > u32 enabled_irq_mask) > > { > > - lockdep_assert_held(&dev_priv->irq_lock); > > + lockdep_assert_held(&i915->irq_lock); > > WARN_ON(enabled_irq_mask & ~interrupt_mask); > > - if (WARN_ON(!intel_irqs_enabled(dev_priv))) > > + if (WARN_ON(!intel_irqs_enabled(i915))) > > return; > > - dev_priv->gt_irq_mask &= ~interrupt_mask; > > - dev_priv->gt_irq_mask |= (~enabled_irq_mask & interrupt_mask); > > - I915_WRITE(GTIMR, dev_priv->gt_irq_mask); > > + i915->gt_irq_mask &= ~interrupt_mask; > > + i915->gt_irq_mask |= (~enabled_irq_mask & interrupt_mask); > > + I915_WRITE(GTIMR, i915->gt_irq_mask); > > } > > -void gen5_enable_gt_irq(struct drm_i915_private *dev_priv, u32 mask) > > +void gen5_enable_gt_irq(struct drm_i915_private *i915, u32 mask) > > { > > - ilk_update_gt_irq(dev_priv, mask, mask); > > - intel_uncore_posting_read_fw(&dev_priv->uncore, GTIMR); > > + ilk_update_gt_irq(i915, mask, mask); > > + intel_uncore_posting_read_fw(&i915->uncore, GTIMR); > > } > > -void gen5_disable_gt_irq(struct drm_i915_private *dev_priv, u32 mask) > > +void gen5_disable_gt_irq(struct drm_i915_private *i915, u32 mask) > > { > > - ilk_update_gt_irq(dev_priv, mask, 0); > > + ilk_update_gt_irq(i915, mask, 0); > > } > > -static i915_reg_t gen6_pm_iir(struct drm_i915_private *dev_priv) > > +static i915_reg_t gen6_pm_iir(struct drm_i915_private *i915) > > { > > - WARN_ON_ONCE(INTEL_GEN(dev_priv) >= 11); > > + WARN_ON_ONCE(INTEL_GEN(i915) >= 11); > > - return INTEL_GEN(dev_priv) >= 8 ? GEN8_GT_IIR(2) : GEN6_PMIIR; > > + return INTEL_GEN(i915) >= 8 ? GEN8_GT_IIR(2) : GEN6_PMIIR; > > } > > -static void write_pm_imr(struct drm_i915_private *dev_priv) > > +static void write_pm_imr(struct drm_i915_private *i915) > > { > > i915_reg_t reg; > > - u32 mask = dev_priv->pm_imr; > > + u32 mask = i915->pm_imr; > > - if (INTEL_GEN(dev_priv) >= 11) { > > + if (INTEL_GEN(i915) >= 11) { > > reg = GEN11_GPM_WGBOXPERF_INTR_MASK; > > /* pm is in upper half */ > > mask = mask << 16; > > - } else if (INTEL_GEN(dev_priv) >= 8) { > > + } else if (INTEL_GEN(i915) >= 8) { > > reg = GEN8_GT_IMR(2); > > } else { > > reg = GEN6_PMIMR; > > @@ -421,16 +421,16 @@ static void write_pm_imr(struct drm_i915_private *dev_priv) > > POSTING_READ(reg); > > } > > -static void write_pm_ier(struct drm_i915_private *dev_priv) > > +static void write_pm_ier(struct drm_i915_private *i915) > > { > > i915_reg_t reg; > > - u32 mask = dev_priv->pm_ier; > > + u32 mask = i915->pm_ier; > > - if (INTEL_GEN(dev_priv) >= 11) { > > + if (INTEL_GEN(i915) >= 11) { > > reg = GEN11_GPM_WGBOXPERF_INTR_ENABLE; > > /* pm is in upper half */ > > mask = mask << 16; > > - } else if (INTEL_GEN(dev_priv) >= 8) { > > + } else if (INTEL_GEN(i915) >= 8) { > > reg = GEN8_GT_IER(2); > > } else { > > reg = GEN6_PMIER; > > @@ -441,11 +441,11 @@ static void write_pm_ier(struct drm_i915_private *dev_priv) > > /** > > * snb_update_pm_irq - update GEN6_PMIMR > > - * @dev_priv: driver private > > + * @i915: driver private > > * @interrupt_mask: mask of interrupt bits to update > > * @enabled_irq_mask: mask of interrupt bits to enable > > */ > > -static void snb_update_pm_irq(struct drm_i915_private *dev_priv, > > +static void snb_update_pm_irq(struct drm_i915_private *i915, > > u32 interrupt_mask, > > u32 enabled_irq_mask) > > { > > @@ -453,127 +453,127 @@ static void snb_update_pm_irq(struct drm_i915_private *dev_priv, > > WARN_ON(enabled_irq_mask & ~interrupt_mask); > > - lockdep_assert_held(&dev_priv->irq_lock); > > + lockdep_assert_held(&i915->irq_lock); > > - new_val = dev_priv->pm_imr; > > + new_val = i915->pm_imr; > > new_val &= ~interrupt_mask; > > new_val |= (~enabled_irq_mask & interrupt_mask); > > - if (new_val != dev_priv->pm_imr) { > > - dev_priv->pm_imr = new_val; > > - write_pm_imr(dev_priv); > > + if (new_val != i915->pm_imr) { > > + i915->pm_imr = new_val; > > + write_pm_imr(i915); > > } > > } > > -void gen6_unmask_pm_irq(struct drm_i915_private *dev_priv, u32 mask) > > +void gen6_unmask_pm_irq(struct drm_i915_private *i915, u32 mask) > > { > > - if (WARN_ON(!intel_irqs_enabled(dev_priv))) > > + if (WARN_ON(!intel_irqs_enabled(i915))) > > return; > > - snb_update_pm_irq(dev_priv, mask, mask); > > + snb_update_pm_irq(i915, mask, mask); > > } > > -static void __gen6_mask_pm_irq(struct drm_i915_private *dev_priv, u32 mask) > > +static void __gen6_mask_pm_irq(struct drm_i915_private *i915, u32 mask) > > { > > - snb_update_pm_irq(dev_priv, mask, 0); > > + snb_update_pm_irq(i915, mask, 0); > > } > > -void gen6_mask_pm_irq(struct drm_i915_private *dev_priv, u32 mask) > > +void gen6_mask_pm_irq(struct drm_i915_private *i915, u32 mask) > > { > > - if (WARN_ON(!intel_irqs_enabled(dev_priv))) > > + if (WARN_ON(!intel_irqs_enabled(i915))) > > return; > > - __gen6_mask_pm_irq(dev_priv, mask); > > + __gen6_mask_pm_irq(i915, mask); > > } > > -static void gen6_reset_pm_iir(struct drm_i915_private *dev_priv, u32 reset_mask) > > +static void gen6_reset_pm_iir(struct drm_i915_private *i915, u32 reset_mask) > > { > > - i915_reg_t reg = gen6_pm_iir(dev_priv); > > + i915_reg_t reg = gen6_pm_iir(i915); > > - lockdep_assert_held(&dev_priv->irq_lock); > > + lockdep_assert_held(&i915->irq_lock); > > I915_WRITE(reg, reset_mask); > > I915_WRITE(reg, reset_mask); > > POSTING_READ(reg); > > } > > -static void gen6_enable_pm_irq(struct drm_i915_private *dev_priv, u32 enable_mask) > > +static void gen6_enable_pm_irq(struct drm_i915_private *i915, u32 enable_mask) > > { > > - lockdep_assert_held(&dev_priv->irq_lock); > > + lockdep_assert_held(&i915->irq_lock); > > - dev_priv->pm_ier |= enable_mask; > > - write_pm_ier(dev_priv); > > - gen6_unmask_pm_irq(dev_priv, enable_mask); > > + i915->pm_ier |= enable_mask; > > + write_pm_ier(i915); > > + gen6_unmask_pm_irq(i915, enable_mask); > > /* unmask_pm_irq provides an implicit barrier (POSTING_READ) */ > > } > > -static void gen6_disable_pm_irq(struct drm_i915_private *dev_priv, u32 disable_mask) > > +static void gen6_disable_pm_irq(struct drm_i915_private *i915, u32 disable_mask) > > { > > - lockdep_assert_held(&dev_priv->irq_lock); > > + lockdep_assert_held(&i915->irq_lock); > > - dev_priv->pm_ier &= ~disable_mask; > > - __gen6_mask_pm_irq(dev_priv, disable_mask); > > - write_pm_ier(dev_priv); > > + i915->pm_ier &= ~disable_mask; > > + __gen6_mask_pm_irq(i915, disable_mask); > > + write_pm_ier(i915); > > /* though a barrier is missing here, but don't really need a one */ > > } > > -void gen11_reset_rps_interrupts(struct drm_i915_private *dev_priv) > > +void gen11_reset_rps_interrupts(struct drm_i915_private *i915) > > { > > - spin_lock_irq(&dev_priv->irq_lock); > > + spin_lock_irq(&i915->irq_lock); > > - while (gen11_reset_one_iir(dev_priv, 0, GEN11_GTPM)) > > + while (gen11_reset_one_iir(i915, 0, GEN11_GTPM)) > > ; > > - dev_priv->gt_pm.rps.pm_iir = 0; > > + i915->gt_pm.rps.pm_iir = 0; > > - spin_unlock_irq(&dev_priv->irq_lock); > > + spin_unlock_irq(&i915->irq_lock); > > } > > -void gen6_reset_rps_interrupts(struct drm_i915_private *dev_priv) > > +void gen6_reset_rps_interrupts(struct drm_i915_private *i915) > > { > > - spin_lock_irq(&dev_priv->irq_lock); > > - gen6_reset_pm_iir(dev_priv, GEN6_PM_RPS_EVENTS); > > - dev_priv->gt_pm.rps.pm_iir = 0; > > - spin_unlock_irq(&dev_priv->irq_lock); > > + spin_lock_irq(&i915->irq_lock); > > + gen6_reset_pm_iir(i915, GEN6_PM_RPS_EVENTS); > > + i915->gt_pm.rps.pm_iir = 0; > > + spin_unlock_irq(&i915->irq_lock); > > } > > -void gen6_enable_rps_interrupts(struct drm_i915_private *dev_priv) > > +void gen6_enable_rps_interrupts(struct drm_i915_private *i915) > > { > > - struct intel_rps *rps = &dev_priv->gt_pm.rps; > > + struct intel_rps *rps = &i915->gt_pm.rps; > > if (READ_ONCE(rps->interrupts_enabled)) > > return; > > - spin_lock_irq(&dev_priv->irq_lock); > > + spin_lock_irq(&i915->irq_lock); > > WARN_ON_ONCE(rps->pm_iir); > > - if (INTEL_GEN(dev_priv) >= 11) > > - WARN_ON_ONCE(gen11_reset_one_iir(dev_priv, 0, GEN11_GTPM)); > > + if (INTEL_GEN(i915) >= 11) > > + WARN_ON_ONCE(gen11_reset_one_iir(i915, 0, GEN11_GTPM)); > > else > > - WARN_ON_ONCE(I915_READ(gen6_pm_iir(dev_priv)) & dev_priv->pm_rps_events); > > + WARN_ON_ONCE(I915_READ(gen6_pm_iir(i915)) & i915->pm_rps_events); > > rps->interrupts_enabled = true; > > - gen6_enable_pm_irq(dev_priv, dev_priv->pm_rps_events); > > + gen6_enable_pm_irq(i915, i915->pm_rps_events); > > - spin_unlock_irq(&dev_priv->irq_lock); > > + spin_unlock_irq(&i915->irq_lock); > > } > > -void gen6_disable_rps_interrupts(struct drm_i915_private *dev_priv) > > +void gen6_disable_rps_interrupts(struct drm_i915_private *i915) > > { > > - struct intel_rps *rps = &dev_priv->gt_pm.rps; > > + struct intel_rps *rps = &i915->gt_pm.rps; > > if (!READ_ONCE(rps->interrupts_enabled)) > > return; > > - spin_lock_irq(&dev_priv->irq_lock); > > + spin_lock_irq(&i915->irq_lock); > > rps->interrupts_enabled = false; > > - I915_WRITE(GEN6_PMINTRMSK, gen6_sanitize_rps_pm_mask(dev_priv, ~0u)); > > + I915_WRITE(GEN6_PMINTRMSK, gen6_sanitize_rps_pm_mask(i915, ~0u)); > > - gen6_disable_pm_irq(dev_priv, GEN6_PM_RPS_EVENTS); > > + gen6_disable_pm_irq(i915, GEN6_PM_RPS_EVENTS); > > - spin_unlock_irq(&dev_priv->irq_lock); > > - synchronize_irq(dev_priv->drm.irq); > > + spin_unlock_irq(&i915->irq_lock); > > + synchronize_irq(i915->drm.irq); > > /* Now that we will not be generating any more work, flush any > > * outstanding tasks. As we are called on the RPS idle path, > > @@ -581,48 +581,48 @@ void gen6_disable_rps_interrupts(struct drm_i915_private *dev_priv) > > * state of the worker can be discarded. > > */ > > cancel_work_sync(&rps->work); > > - if (INTEL_GEN(dev_priv) >= 11) > > - gen11_reset_rps_interrupts(dev_priv); > > + if (INTEL_GEN(i915) >= 11) > > + gen11_reset_rps_interrupts(i915); > > else > > - gen6_reset_rps_interrupts(dev_priv); > > + gen6_reset_rps_interrupts(i915); > > } > > -void gen9_reset_guc_interrupts(struct drm_i915_private *dev_priv) > > +void gen9_reset_guc_interrupts(struct drm_i915_private *i915) > > { > > - assert_rpm_wakelock_held(dev_priv); > > + assert_rpm_wakelock_held(i915); > > - spin_lock_irq(&dev_priv->irq_lock); > > - gen6_reset_pm_iir(dev_priv, dev_priv->pm_guc_events); > > - spin_unlock_irq(&dev_priv->irq_lock); > > + spin_lock_irq(&i915->irq_lock); > > + gen6_reset_pm_iir(i915, i915->pm_guc_events); > > + spin_unlock_irq(&i915->irq_lock); > > } > > -void gen9_enable_guc_interrupts(struct drm_i915_private *dev_priv) > > +void gen9_enable_guc_interrupts(struct drm_i915_private *i915) > > { > > - assert_rpm_wakelock_held(dev_priv); > > + assert_rpm_wakelock_held(i915); > > - spin_lock_irq(&dev_priv->irq_lock); > > - if (!dev_priv->guc.interrupts.enabled) { > > - WARN_ON_ONCE(I915_READ(gen6_pm_iir(dev_priv)) & > > - dev_priv->pm_guc_events); > > - dev_priv->guc.interrupts.enabled = true; > > - gen6_enable_pm_irq(dev_priv, dev_priv->pm_guc_events); > > + spin_lock_irq(&i915->irq_lock); > > + if (!i915->guc.interrupts.enabled) { > > + WARN_ON_ONCE(I915_READ(gen6_pm_iir(i915)) & > > + i915->pm_guc_events); > > + i915->guc.interrupts.enabled = true; > > + gen6_enable_pm_irq(i915, i915->pm_guc_events); > > } > > - spin_unlock_irq(&dev_priv->irq_lock); > > + spin_unlock_irq(&i915->irq_lock); > > } > > -void gen9_disable_guc_interrupts(struct drm_i915_private *dev_priv) > > +void gen9_disable_guc_interrupts(struct drm_i915_private *i915) > > { > > - assert_rpm_wakelock_held(dev_priv); > > + assert_rpm_wakelock_held(i915); > > - spin_lock_irq(&dev_priv->irq_lock); > > - dev_priv->guc.interrupts.enabled = false; > > + spin_lock_irq(&i915->irq_lock); > > + i915->guc.interrupts.enabled = false; > > - gen6_disable_pm_irq(dev_priv, dev_priv->pm_guc_events); > > + gen6_disable_pm_irq(i915, i915->pm_guc_events); > > - spin_unlock_irq(&dev_priv->irq_lock); > > - synchronize_irq(dev_priv->drm.irq); > > + spin_unlock_irq(&i915->irq_lock); > > + synchronize_irq(i915->drm.irq); > > - gen9_reset_guc_interrupts(dev_priv); > > + gen9_reset_guc_interrupts(i915); > > } > > void gen11_reset_guc_interrupts(struct drm_i915_private *i915) > > @@ -632,53 +632,53 @@ void gen11_reset_guc_interrupts(struct drm_i915_private *i915) > > spin_unlock_irq(&i915->irq_lock); > > } > > -void gen11_enable_guc_interrupts(struct drm_i915_private *dev_priv) > > +void gen11_enable_guc_interrupts(struct drm_i915_private *i915) > > { > > - spin_lock_irq(&dev_priv->irq_lock); > > - if (!dev_priv->guc.interrupts.enabled) { > > + spin_lock_irq(&i915->irq_lock); > > + if (!i915->guc.interrupts.enabled) { > > u32 events = REG_FIELD_PREP(ENGINE1_MASK, > > GEN11_GUC_INTR_GUC2HOST); > > - WARN_ON_ONCE(gen11_reset_one_iir(dev_priv, 0, GEN11_GUC)); > > + WARN_ON_ONCE(gen11_reset_one_iir(i915, 0, GEN11_GUC)); > > I915_WRITE(GEN11_GUC_SG_INTR_ENABLE, events); > > I915_WRITE(GEN11_GUC_SG_INTR_MASK, ~events); > > - dev_priv->guc.interrupts.enabled = true; > > + i915->guc.interrupts.enabled = true; > > } > > - spin_unlock_irq(&dev_priv->irq_lock); > > + spin_unlock_irq(&i915->irq_lock); > > } > > -void gen11_disable_guc_interrupts(struct drm_i915_private *dev_priv) > > +void gen11_disable_guc_interrupts(struct drm_i915_private *i915) > > { > > - spin_lock_irq(&dev_priv->irq_lock); > > - dev_priv->guc.interrupts.enabled = false; > > + spin_lock_irq(&i915->irq_lock); > > + i915->guc.interrupts.enabled = false; > > I915_WRITE(GEN11_GUC_SG_INTR_MASK, ~0); > > I915_WRITE(GEN11_GUC_SG_INTR_ENABLE, 0); > > - spin_unlock_irq(&dev_priv->irq_lock); > > - synchronize_irq(dev_priv->drm.irq); > > + spin_unlock_irq(&i915->irq_lock); > > + synchronize_irq(i915->drm.irq); > > - gen11_reset_guc_interrupts(dev_priv); > > + gen11_reset_guc_interrupts(i915); > > } > > /** > > * bdw_update_port_irq - update DE port interrupt > > - * @dev_priv: driver private > > + * @i915: driver private > > * @interrupt_mask: mask of interrupt bits to update > > * @enabled_irq_mask: mask of interrupt bits to enable > > */ > > -static void bdw_update_port_irq(struct drm_i915_private *dev_priv, > > +static void bdw_update_port_irq(struct drm_i915_private *i915, > > u32 interrupt_mask, > > u32 enabled_irq_mask) > > { > > u32 new_val; > > u32 old_val; > > - lockdep_assert_held(&dev_priv->irq_lock); > > + lockdep_assert_held(&i915->irq_lock); > > WARN_ON(enabled_irq_mask & ~interrupt_mask); > > - if (WARN_ON(!intel_irqs_enabled(dev_priv))) > > + if (WARN_ON(!intel_irqs_enabled(i915))) > > return; > > old_val = I915_READ(GEN8_DE_PORT_IMR); > > @@ -695,43 +695,43 @@ static void bdw_update_port_irq(struct drm_i915_private *dev_priv, > > /** > > * bdw_update_pipe_irq - update DE pipe interrupt > > - * @dev_priv: driver private > > + * @i915: driver private > > * @pipe: pipe whose interrupt to update > > * @interrupt_mask: mask of interrupt bits to update > > * @enabled_irq_mask: mask of interrupt bits to enable > > */ > > -void bdw_update_pipe_irq(struct drm_i915_private *dev_priv, > > +void bdw_update_pipe_irq(struct drm_i915_private *i915, > > enum pipe pipe, > > u32 interrupt_mask, > > u32 enabled_irq_mask) > > { > > u32 new_val; > > - lockdep_assert_held(&dev_priv->irq_lock); > > + lockdep_assert_held(&i915->irq_lock); > > WARN_ON(enabled_irq_mask & ~interrupt_mask); > > - if (WARN_ON(!intel_irqs_enabled(dev_priv))) > > + if (WARN_ON(!intel_irqs_enabled(i915))) > > return; > > - new_val = dev_priv->de_irq_mask[pipe]; > > + new_val = i915->de_irq_mask[pipe]; > > new_val &= ~interrupt_mask; > > new_val |= (~enabled_irq_mask & interrupt_mask); > > - if (new_val != dev_priv->de_irq_mask[pipe]) { > > - dev_priv->de_irq_mask[pipe] = new_val; > > - I915_WRITE(GEN8_DE_PIPE_IMR(pipe), dev_priv->de_irq_mask[pipe]); > > + if (new_val != i915->de_irq_mask[pipe]) { > > + i915->de_irq_mask[pipe] = new_val; > > + I915_WRITE(GEN8_DE_PIPE_IMR(pipe), i915->de_irq_mask[pipe]); > > POSTING_READ(GEN8_DE_PIPE_IMR(pipe)); > > } > > } > > /** > > * ibx_display_interrupt_update - update SDEIMR > > - * @dev_priv: driver private > > + * @i915: driver private > > * @interrupt_mask: mask of interrupt bits to update > > * @enabled_irq_mask: mask of interrupt bits to enable > > */ > > -void ibx_display_interrupt_update(struct drm_i915_private *dev_priv, > > +void ibx_display_interrupt_update(struct drm_i915_private *i915, > > u32 interrupt_mask, > > u32 enabled_irq_mask) > > { > > @@ -741,24 +741,24 @@ void ibx_display_interrupt_update(struct drm_i915_private *dev_priv, > > WARN_ON(enabled_irq_mask & ~interrupt_mask); > > - lockdep_assert_held(&dev_priv->irq_lock); > > + lockdep_assert_held(&i915->irq_lock); > > - if (WARN_ON(!intel_irqs_enabled(dev_priv))) > > + if (WARN_ON(!intel_irqs_enabled(i915))) > > return; > > I915_WRITE(SDEIMR, sdeimr); > > POSTING_READ(SDEIMR); > > } > > -u32 i915_pipestat_enable_mask(struct drm_i915_private *dev_priv, > > +u32 i915_pipestat_enable_mask(struct drm_i915_private *i915, > > enum pipe pipe) > > { > > - u32 status_mask = dev_priv->pipestat_irq_mask[pipe]; > > + u32 status_mask = i915->pipestat_irq_mask[pipe]; > > u32 enable_mask = status_mask << 16; > > - lockdep_assert_held(&dev_priv->irq_lock); > > + lockdep_assert_held(&i915->irq_lock); > > - if (INTEL_GEN(dev_priv) < 5) > > + if (INTEL_GEN(i915) < 5) > > goto out; > > /* > > @@ -791,7 +791,7 @@ u32 i915_pipestat_enable_mask(struct drm_i915_private *dev_priv, > > return enable_mask; > > } > > -void i915_enable_pipestat(struct drm_i915_private *dev_priv, > > +void i915_enable_pipestat(struct drm_i915_private *i915, > > enum pipe pipe, u32 status_mask) > > { > > i915_reg_t reg = PIPESTAT(pipe); > > @@ -801,20 +801,20 @@ void i915_enable_pipestat(struct drm_i915_private *dev_priv, > > "pipe %c: status_mask=0x%x\n", > > pipe_name(pipe), status_mask); > > - lockdep_assert_held(&dev_priv->irq_lock); > > - WARN_ON(!intel_irqs_enabled(dev_priv)); > > + lockdep_assert_held(&i915->irq_lock); > > + WARN_ON(!intel_irqs_enabled(i915)); > > - if ((dev_priv->pipestat_irq_mask[pipe] & status_mask) == status_mask) > > + if ((i915->pipestat_irq_mask[pipe] & status_mask) == status_mask) > > return; > > - dev_priv->pipestat_irq_mask[pipe] |= status_mask; > > - enable_mask = i915_pipestat_enable_mask(dev_priv, pipe); > > + i915->pipestat_irq_mask[pipe] |= status_mask; > > + enable_mask = i915_pipestat_enable_mask(i915, pipe); > > I915_WRITE(reg, enable_mask | status_mask); > > POSTING_READ(reg); > > } > > -void i915_disable_pipestat(struct drm_i915_private *dev_priv, > > +void i915_disable_pipestat(struct drm_i915_private *i915, > > enum pipe pipe, u32 status_mask) > > { > > i915_reg_t reg = PIPESTAT(pipe); > > @@ -824,44 +824,44 @@ void i915_disable_pipestat(struct drm_i915_private *dev_priv, > > "pipe %c: status_mask=0x%x\n", > > pipe_name(pipe), status_mask); > > - lockdep_assert_held(&dev_priv->irq_lock); > > - WARN_ON(!intel_irqs_enabled(dev_priv)); > > + lockdep_assert_held(&i915->irq_lock); > > + WARN_ON(!intel_irqs_enabled(i915)); > > - if ((dev_priv->pipestat_irq_mask[pipe] & status_mask) == 0) > > + if ((i915->pipestat_irq_mask[pipe] & status_mask) == 0) > > return; > > - dev_priv->pipestat_irq_mask[pipe] &= ~status_mask; > > - enable_mask = i915_pipestat_enable_mask(dev_priv, pipe); > > + i915->pipestat_irq_mask[pipe] &= ~status_mask; > > + enable_mask = i915_pipestat_enable_mask(i915, pipe); > > I915_WRITE(reg, enable_mask | status_mask); > > POSTING_READ(reg); > > } > > -static bool i915_has_asle(struct drm_i915_private *dev_priv) > > +static bool i915_has_asle(struct drm_i915_private *i915) > > { > > - if (!dev_priv->opregion.asle) > > + if (!i915->opregion.asle) > > return false; > > - return IS_PINEVIEW(dev_priv) || IS_MOBILE(dev_priv); > > + return IS_PINEVIEW(i915) || IS_MOBILE(i915); > > } > > /** > > * i915_enable_asle_pipestat - enable ASLE pipestat for OpRegion > > - * @dev_priv: i915 device private > > + * @i915: i915 device private > > */ > > -static void i915_enable_asle_pipestat(struct drm_i915_private *dev_priv) > > +static void i915_enable_asle_pipestat(struct drm_i915_private *i915) > > { > > - if (!i915_has_asle(dev_priv)) > > + if (!i915_has_asle(i915)) > > return; > > - spin_lock_irq(&dev_priv->irq_lock); > > + spin_lock_irq(&i915->irq_lock); > > - i915_enable_pipestat(dev_priv, PIPE_B, PIPE_LEGACY_BLC_EVENT_STATUS); > > - if (INTEL_GEN(dev_priv) >= 4) > > - i915_enable_pipestat(dev_priv, PIPE_A, > > + i915_enable_pipestat(i915, PIPE_B, PIPE_LEGACY_BLC_EVENT_STATUS); > > + if (INTEL_GEN(i915) >= 4) > > + i915_enable_pipestat(i915, PIPE_A, > > PIPE_LEGACY_BLC_EVENT_STATUS); > > - spin_unlock_irq(&dev_priv->irq_lock); > > + spin_unlock_irq(&i915->irq_lock); > > } > > /* > > @@ -919,7 +919,7 @@ static void i915_enable_asle_pipestat(struct drm_i915_private *dev_priv) > > */ > > static u32 i915_get_vblank_counter(struct drm_device *dev, unsigned int pipe) > > { > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > struct drm_vblank_crtc *vblank = &dev->vblank[pipe]; > > const struct drm_display_mode *mode = &vblank->hwmode; > > i915_reg_t high_frame, low_frame; > > @@ -955,7 +955,7 @@ static u32 i915_get_vblank_counter(struct drm_device *dev, unsigned int pipe) > > high_frame = PIPEFRAME(pipe); > > low_frame = PIPEFRAMEPIXEL(pipe); > > - spin_lock_irqsave(&dev_priv->uncore.lock, irqflags); > > + spin_lock_irqsave(&i915->uncore.lock, irqflags); > > /* > > * High & low register fields aren't synchronized, so make sure > > @@ -968,7 +968,7 @@ static u32 i915_get_vblank_counter(struct drm_device *dev, unsigned int pipe) > > high2 = I915_READ_FW(high_frame) & PIPE_FRAME_HIGH_MASK; > > } while (high1 != high2); > > - spin_unlock_irqrestore(&dev_priv->uncore.lock, irqflags); > > + spin_unlock_irqrestore(&i915->uncore.lock, irqflags); > > high1 >>= PIPE_FRAME_HIGH_SHIFT; > > pixel = low & PIPE_PIXEL_MASK; > > @@ -984,7 +984,7 @@ static u32 i915_get_vblank_counter(struct drm_device *dev, unsigned int pipe) > > static u32 g4x_get_vblank_counter(struct drm_device *dev, unsigned int pipe) > > { > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > return I915_READ(PIPE_FRMCOUNT_G4X(pipe)); > > } > > @@ -999,7 +999,7 @@ static u32 g4x_get_vblank_counter(struct drm_device *dev, unsigned int pipe) > > */ > > static u32 __intel_get_crtc_scanline_from_timestamp(struct intel_crtc *crtc) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > struct drm_vblank_crtc *vblank = > > &crtc->base.dev->vblank[drm_crtc_index(&crtc->base)]; > > const struct drm_display_mode *mode = &vblank->hwmode; > > @@ -1044,7 +1044,7 @@ static u32 __intel_get_crtc_scanline_from_timestamp(struct intel_crtc *crtc) > > static int __intel_get_crtc_scanline(struct intel_crtc *crtc) > > { > > struct drm_device *dev = crtc->base.dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > const struct drm_display_mode *mode; > > struct drm_vblank_crtc *vblank; > > enum pipe pipe = crtc->pipe; > > @@ -1063,7 +1063,7 @@ static int __intel_get_crtc_scanline(struct intel_crtc *crtc) > > if (mode->flags & DRM_MODE_FLAG_INTERLACE) > > vtotal /= 2; > > - if (IS_GEN(dev_priv, 2)) > > + if (IS_GEN(i915, 2)) > > position = I915_READ_FW(PIPEDSL(pipe)) & DSL_LINEMASK_GEN2; > > else > > position = I915_READ_FW(PIPEDSL(pipe)) & DSL_LINEMASK_GEN3; > > @@ -1080,7 +1080,7 @@ static int __intel_get_crtc_scanline(struct intel_crtc *crtc) > > * problem. We may need to extend this to include other platforms, > > * but so far testing only shows the problem on HSW. > > */ > > - if (HAS_DDI(dev_priv) && !position) { > > + if (HAS_DDI(i915) && !position) { > > int i, temp; > > for (i = 0; i < 100; i++) { > > @@ -1105,14 +1105,14 @@ static bool i915_get_crtc_scanoutpos(struct drm_device *dev, unsigned int pipe, > > ktime_t *stime, ktime_t *etime, > > const struct drm_display_mode *mode) > > { > > - struct drm_i915_private *dev_priv = to_i915(dev); > > - struct intel_crtc *intel_crtc = intel_get_crtc_for_pipe(dev_priv, > > + struct drm_i915_private *i915 = to_i915(dev); > > + struct intel_crtc *intel_crtc = intel_get_crtc_for_pipe(i915, > > pipe); > > int position; > > int vbl_start, vbl_end, hsync_start, htotal, vtotal; > > unsigned long irqflags; > > - bool use_scanline_counter = INTEL_GEN(dev_priv) >= 5 || > > - IS_G4X(dev_priv) || IS_GEN(dev_priv, 2) || > > + bool use_scanline_counter = INTEL_GEN(i915) >= 5 || > > + IS_G4X(i915) || IS_GEN(i915, 2) || > > mode->private_flags & I915_MODE_FLAG_USE_SCANLINE_COUNTER; > > if (WARN_ON(!mode->crtc_clock)) { > > @@ -1138,7 +1138,7 @@ static bool i915_get_crtc_scanoutpos(struct drm_device *dev, unsigned int pipe, > > * register reads, potentially with preemption disabled, so the > > * following code must not block on uncore.lock. > > */ > > - spin_lock_irqsave(&dev_priv->uncore.lock, irqflags); > > + spin_lock_irqsave(&i915->uncore.lock, irqflags); > > /* preempt_disable_rt() should go right here in PREEMPT_RT patchset. */ > > @@ -1193,7 +1193,7 @@ static bool i915_get_crtc_scanoutpos(struct drm_device *dev, unsigned int pipe, > > /* preempt_enable_rt() should go right here in PREEMPT_RT patchset. */ > > - spin_unlock_irqrestore(&dev_priv->uncore.lock, irqflags); > > + spin_unlock_irqrestore(&i915->uncore.lock, irqflags); > > /* > > * While in vblank, position will be negative > > @@ -1219,20 +1219,20 @@ static bool i915_get_crtc_scanoutpos(struct drm_device *dev, unsigned int pipe, > > int intel_get_crtc_scanline(struct intel_crtc *crtc) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > unsigned long irqflags; > > int position; > > - spin_lock_irqsave(&dev_priv->uncore.lock, irqflags); > > + spin_lock_irqsave(&i915->uncore.lock, irqflags); > > position = __intel_get_crtc_scanline(crtc); > > - spin_unlock_irqrestore(&dev_priv->uncore.lock, irqflags); > > + spin_unlock_irqrestore(&i915->uncore.lock, irqflags); > > return position; > > } > > -static void ironlake_rps_change_irq_handler(struct drm_i915_private *dev_priv) > > +static void ironlake_rps_change_irq_handler(struct drm_i915_private *i915) > > { > > - struct intel_uncore *uncore = &dev_priv->uncore; > > + struct intel_uncore *uncore = &i915->uncore; > > u32 busy_up, busy_down, max_avg, min_avg; > > u8 new_delay; > > @@ -1242,7 +1242,7 @@ static void ironlake_rps_change_irq_handler(struct drm_i915_private *dev_priv) > > MEMINTRSTS, > > intel_uncore_read(uncore, MEMINTRSTS)); > > - new_delay = dev_priv->ips.cur_delay; > > + new_delay = i915->ips.cur_delay; > > intel_uncore_write16(uncore, MEMINTRSTS, MEMINT_EVAL_CHG); > > busy_up = intel_uncore_read(uncore, RCPREVBSYTUPAVG); > > @@ -1252,26 +1252,26 @@ static void ironlake_rps_change_irq_handler(struct drm_i915_private *dev_priv) > > /* Handle RCS change request from hw */ > > if (busy_up > max_avg) { > > - if (dev_priv->ips.cur_delay != dev_priv->ips.max_delay) > > - new_delay = dev_priv->ips.cur_delay - 1; > > - if (new_delay < dev_priv->ips.max_delay) > > - new_delay = dev_priv->ips.max_delay; > > + if (i915->ips.cur_delay != i915->ips.max_delay) > > + new_delay = i915->ips.cur_delay - 1; > > + if (new_delay < i915->ips.max_delay) > > + new_delay = i915->ips.max_delay; > > } else if (busy_down < min_avg) { > > - if (dev_priv->ips.cur_delay != dev_priv->ips.min_delay) > > - new_delay = dev_priv->ips.cur_delay + 1; > > - if (new_delay > dev_priv->ips.min_delay) > > - new_delay = dev_priv->ips.min_delay; > > + if (i915->ips.cur_delay != i915->ips.min_delay) > > + new_delay = i915->ips.cur_delay + 1; > > + if (new_delay > i915->ips.min_delay) > > + new_delay = i915->ips.min_delay; > > } > > - if (ironlake_set_drps(dev_priv, new_delay)) > > - dev_priv->ips.cur_delay = new_delay; > > + if (ironlake_set_drps(i915, new_delay)) > > + i915->ips.cur_delay = new_delay; > > spin_unlock(&mchdev_lock); > > return; > > } > > -static void vlv_c0_read(struct drm_i915_private *dev_priv, > > +static void vlv_c0_read(struct drm_i915_private *i915, > > struct intel_rps_ei *ei) > > { > > ei->ktime = ktime_get_raw(); > > @@ -1279,14 +1279,14 @@ static void vlv_c0_read(struct drm_i915_private *dev_priv, > > ei->media_c0 = I915_READ(VLV_MEDIA_C0_COUNT); > > } > > -void gen6_rps_reset_ei(struct drm_i915_private *dev_priv) > > +void gen6_rps_reset_ei(struct drm_i915_private *i915) > > { > > - memset(&dev_priv->gt_pm.rps.ei, 0, sizeof(dev_priv->gt_pm.rps.ei)); > > + memset(&i915->gt_pm.rps.ei, 0, sizeof(i915->gt_pm.rps.ei)); > > } > > -static u32 vlv_wa_c0_ei(struct drm_i915_private *dev_priv, u32 pm_iir) > > +static u32 vlv_wa_c0_ei(struct drm_i915_private *i915, u32 pm_iir) > > { > > - struct intel_rps *rps = &dev_priv->gt_pm.rps; > > + struct intel_rps *rps = &i915->gt_pm.rps; > > const struct intel_rps_ei *prev = &rps->ei; > > struct intel_rps_ei now; > > u32 events = 0; > > @@ -1294,7 +1294,7 @@ static u32 vlv_wa_c0_ei(struct drm_i915_private *dev_priv, u32 pm_iir) > > if ((pm_iir & GEN6_PM_RP_UP_EI_EXPIRED) == 0) > > return 0; > > - vlv_c0_read(dev_priv, &now); > > + vlv_c0_read(i915, &now); > > if (prev->ktime) { > > u64 time, c0; > > @@ -1302,7 +1302,7 @@ static u32 vlv_wa_c0_ei(struct drm_i915_private *dev_priv, u32 pm_iir) > > time = ktime_us_delta(now.ktime, prev->ktime); > > - time *= dev_priv->czclk_freq; > > + time *= i915->czclk_freq; > > /* Workload can be split between render + media, > > * e.g. SwapBuffers being blitted in X after being rendered in > > @@ -1326,28 +1326,28 @@ static u32 vlv_wa_c0_ei(struct drm_i915_private *dev_priv, u32 pm_iir) > > static void gen6_pm_rps_work(struct work_struct *work) > > { > > - struct drm_i915_private *dev_priv = > > + struct drm_i915_private *i915 = > > container_of(work, struct drm_i915_private, gt_pm.rps.work); > > - struct intel_rps *rps = &dev_priv->gt_pm.rps; > > + struct intel_rps *rps = &i915->gt_pm.rps; > > bool client_boost = false; > > int new_delay, adj, min, max; > > u32 pm_iir = 0; > > - spin_lock_irq(&dev_priv->irq_lock); > > + spin_lock_irq(&i915->irq_lock); > > if (rps->interrupts_enabled) { > > pm_iir = fetch_and_zero(&rps->pm_iir); > > client_boost = atomic_read(&rps->num_waiters); > > } > > - spin_unlock_irq(&dev_priv->irq_lock); > > + spin_unlock_irq(&i915->irq_lock); > > /* Make sure we didn't queue anything we're not going to process. */ > > - WARN_ON(pm_iir & ~dev_priv->pm_rps_events); > > - if ((pm_iir & dev_priv->pm_rps_events) == 0 && !client_boost) > > + WARN_ON(pm_iir & ~i915->pm_rps_events); > > + if ((pm_iir & i915->pm_rps_events) == 0 && !client_boost) > > goto out; > > mutex_lock(&rps->lock); > > - pm_iir |= vlv_wa_c0_ei(dev_priv, pm_iir); > > + pm_iir |= vlv_wa_c0_ei(i915, pm_iir); > > adj = rps->last_adj; > > new_delay = rps->cur_freq; > > @@ -1362,7 +1362,7 @@ static void gen6_pm_rps_work(struct work_struct *work) > > if (adj > 0) > > adj *= 2; > > else /* CHV needs even encode values */ > > - adj = IS_CHERRYVIEW(dev_priv) ? 2 : 1; > > + adj = IS_CHERRYVIEW(i915) ? 2 : 1; > > if (new_delay >= rps->max_freq_softlimit) > > adj = 0; > > @@ -1378,7 +1378,7 @@ static void gen6_pm_rps_work(struct work_struct *work) > > if (adj < 0) > > adj *= 2; > > else /* CHV needs even encode values */ > > - adj = IS_CHERRYVIEW(dev_priv) ? -2 : -1; > > + adj = IS_CHERRYVIEW(i915) ? -2 : -1; > > if (new_delay <= rps->min_freq_softlimit) > > adj = 0; > > @@ -1406,7 +1406,7 @@ static void gen6_pm_rps_work(struct work_struct *work) > > new_delay += adj; > > new_delay = clamp_t(int, new_delay, min, max); > > - if (intel_set_rps(dev_priv, new_delay)) { > > + if (intel_set_rps(i915, new_delay)) { > > DRM_DEBUG_DRIVER("Failed to set new GPU frequency\n"); > > rps->last_adj = 0; > > } > > @@ -1415,10 +1415,10 @@ static void gen6_pm_rps_work(struct work_struct *work) > > out: > > /* Make sure not to corrupt PMIMR state used by ringbuffer on GEN6 */ > > - spin_lock_irq(&dev_priv->irq_lock); > > + spin_lock_irq(&i915->irq_lock); > > if (rps->interrupts_enabled) > > - gen6_unmask_pm_irq(dev_priv, dev_priv->pm_rps_events); > > - spin_unlock_irq(&dev_priv->irq_lock); > > + gen6_unmask_pm_irq(i915, i915->pm_rps_events); > > + spin_unlock_irq(&i915->irq_lock); > > } > > @@ -1433,8 +1433,8 @@ static void gen6_pm_rps_work(struct work_struct *work) > > */ > > static void ivybridge_parity_work(struct work_struct *work) > > { > > - struct drm_i915_private *dev_priv = > > - container_of(work, typeof(*dev_priv), l3_parity.error_work); > > + struct drm_i915_private *i915 = > > + container_of(work, typeof(*i915), l3_parity.error_work); > > u32 error_status, row, bank, subbank; > > char *parity_event[6]; > > u32 misccpctl; > > @@ -1444,24 +1444,24 @@ static void ivybridge_parity_work(struct work_struct *work) > > * In order to prevent a get/put style interface, acquire struct mutex > > * any time we access those registers. > > */ > > - mutex_lock(&dev_priv->drm.struct_mutex); > > + mutex_lock(&i915->drm.struct_mutex); > > /* If we've screwed up tracking, just let the interrupt fire again */ > > - if (WARN_ON(!dev_priv->l3_parity.which_slice)) > > + if (WARN_ON(!i915->l3_parity.which_slice)) > > goto out; > > misccpctl = I915_READ(GEN7_MISCCPCTL); > > I915_WRITE(GEN7_MISCCPCTL, misccpctl & ~GEN7_DOP_CLOCK_GATE_ENABLE); > > POSTING_READ(GEN7_MISCCPCTL); > > - while ((slice = ffs(dev_priv->l3_parity.which_slice)) != 0) { > > + while ((slice = ffs(i915->l3_parity.which_slice)) != 0) { > > i915_reg_t reg; > > slice--; > > - if (WARN_ON_ONCE(slice >= NUM_L3_SLICES(dev_priv))) > > + if (WARN_ON_ONCE(slice >= NUM_L3_SLICES(i915))) > > break; > > - dev_priv->l3_parity.which_slice &= ~(1<<slice); > > + i915->l3_parity.which_slice &= ~(1<<slice); > > reg = GEN7_L3CDERRST1(slice); > > @@ -1480,7 +1480,7 @@ static void ivybridge_parity_work(struct work_struct *work) > > parity_event[4] = kasprintf(GFP_KERNEL, "SLICE=%d", slice); > > parity_event[5] = NULL; > > - kobject_uevent_env(&dev_priv->drm.primary->kdev->kobj, > > + kobject_uevent_env(&i915->drm.primary->kdev->kobj, > > KOBJ_CHANGE, parity_event); > > DRM_DEBUG("Parity error: Slice = %d, Row = %d, Bank = %d, Sub bank = %d.\n", > > @@ -1495,60 +1495,60 @@ static void ivybridge_parity_work(struct work_struct *work) > > I915_WRITE(GEN7_MISCCPCTL, misccpctl); > > out: > > - WARN_ON(dev_priv->l3_parity.which_slice); > > - spin_lock_irq(&dev_priv->irq_lock); > > - gen5_enable_gt_irq(dev_priv, GT_PARITY_ERROR(dev_priv)); > > - spin_unlock_irq(&dev_priv->irq_lock); > > + WARN_ON(i915->l3_parity.which_slice); > > + spin_lock_irq(&i915->irq_lock); > > + gen5_enable_gt_irq(i915, GT_PARITY_ERROR(i915)); > > + spin_unlock_irq(&i915->irq_lock); > > - mutex_unlock(&dev_priv->drm.struct_mutex); > > + mutex_unlock(&i915->drm.struct_mutex); > > } > > -static void ivybridge_parity_error_irq_handler(struct drm_i915_private *dev_priv, > > +static void ivybridge_parity_error_irq_handler(struct drm_i915_private *i915, > > u32 iir) > > { > > - if (!HAS_L3_DPF(dev_priv)) > > + if (!HAS_L3_DPF(i915)) > > return; > > - spin_lock(&dev_priv->irq_lock); > > - gen5_disable_gt_irq(dev_priv, GT_PARITY_ERROR(dev_priv)); > > - spin_unlock(&dev_priv->irq_lock); > > + spin_lock(&i915->irq_lock); > > + gen5_disable_gt_irq(i915, GT_PARITY_ERROR(i915)); > > + spin_unlock(&i915->irq_lock); > > - iir &= GT_PARITY_ERROR(dev_priv); > > + iir &= GT_PARITY_ERROR(i915); > > if (iir & GT_RENDER_L3_PARITY_ERROR_INTERRUPT_S1) > > - dev_priv->l3_parity.which_slice |= 1 << 1; > > + i915->l3_parity.which_slice |= 1 << 1; > > if (iir & GT_RENDER_L3_PARITY_ERROR_INTERRUPT) > > - dev_priv->l3_parity.which_slice |= 1 << 0; > > + i915->l3_parity.which_slice |= 1 << 0; > > - queue_work(dev_priv->wq, &dev_priv->l3_parity.error_work); > > + queue_work(i915->wq, &i915->l3_parity.error_work); > > } > > -static void ilk_gt_irq_handler(struct drm_i915_private *dev_priv, > > +static void ilk_gt_irq_handler(struct drm_i915_private *i915, > > u32 gt_iir) > > { > > if (gt_iir & GT_RENDER_USER_INTERRUPT) > > - intel_engine_breadcrumbs_irq(dev_priv->engine[RCS0]); > > + intel_engine_breadcrumbs_irq(i915->engine[RCS0]); > > if (gt_iir & ILK_BSD_USER_INTERRUPT) > > - intel_engine_breadcrumbs_irq(dev_priv->engine[VCS0]); > > + intel_engine_breadcrumbs_irq(i915->engine[VCS0]); > > } > > -static void snb_gt_irq_handler(struct drm_i915_private *dev_priv, > > +static void snb_gt_irq_handler(struct drm_i915_private *i915, > > u32 gt_iir) > > { > > if (gt_iir & GT_RENDER_USER_INTERRUPT) > > - intel_engine_breadcrumbs_irq(dev_priv->engine[RCS0]); > > + intel_engine_breadcrumbs_irq(i915->engine[RCS0]); > > if (gt_iir & GT_BSD_USER_INTERRUPT) > > - intel_engine_breadcrumbs_irq(dev_priv->engine[VCS0]); > > + intel_engine_breadcrumbs_irq(i915->engine[VCS0]); > > if (gt_iir & GT_BLT_USER_INTERRUPT) > > - intel_engine_breadcrumbs_irq(dev_priv->engine[BCS0]); > > + intel_engine_breadcrumbs_irq(i915->engine[BCS0]); > > if (gt_iir & (GT_BLT_CS_ERROR_INTERRUPT | > > GT_BSD_CS_ERROR_INTERRUPT | > > GT_RENDER_CS_MASTER_ERROR_INTERRUPT)) > > DRM_DEBUG("Command parser error, gt_iir 0x%08x\n", gt_iir); > > - if (gt_iir & GT_PARITY_ERROR(dev_priv)) > > - ivybridge_parity_error_irq_handler(dev_priv, gt_iir); > > + if (gt_iir & GT_PARITY_ERROR(i915)) > > + ivybridge_parity_error_irq_handler(i915, gt_iir); > > } > > static void > > @@ -1763,7 +1763,7 @@ static bool i9xx_port_hotplug_long_detect(enum hpd_pin pin, u32 val) > > * > > * Note that the caller is expected to zero out the masks initially. > > */ > > -static void intel_get_hpd_pins(struct drm_i915_private *dev_priv, > > +static void intel_get_hpd_pins(struct drm_i915_private *i915, > > u32 *pin_mask, u32 *long_mask, > > u32 hotplug_trigger, u32 dig_hotplug_reg, > > const u32 hpd[HPD_NUM_PINS], > > @@ -1786,25 +1786,25 @@ static void intel_get_hpd_pins(struct drm_i915_private *dev_priv, > > } > > -static void gmbus_irq_handler(struct drm_i915_private *dev_priv) > > +static void gmbus_irq_handler(struct drm_i915_private *i915) > > { > > - wake_up_all(&dev_priv->gmbus_wait_queue); > > + wake_up_all(&i915->gmbus_wait_queue); > > } > > -static void dp_aux_irq_handler(struct drm_i915_private *dev_priv) > > +static void dp_aux_irq_handler(struct drm_i915_private *i915) > > { > > - wake_up_all(&dev_priv->gmbus_wait_queue); > > + wake_up_all(&i915->gmbus_wait_queue); > > } > > #if defined(CONFIG_DEBUG_FS) > > -static void display_pipe_crc_irq_handler(struct drm_i915_private *dev_priv, > > +static void display_pipe_crc_irq_handler(struct drm_i915_private *i915, > > enum pipe pipe, > > u32 crc0, u32 crc1, > > u32 crc2, u32 crc3, > > u32 crc4) > > { > > - struct intel_pipe_crc *pipe_crc = &dev_priv->pipe_crc[pipe]; > > - struct intel_crtc *crtc = intel_get_crtc_for_pipe(dev_priv, pipe); > > + struct intel_pipe_crc *pipe_crc = &i915->pipe_crc[pipe]; > > + struct intel_crtc *crtc = intel_get_crtc_for_pipe(i915, pipe); > > u32 crcs[5] = { crc0, crc1, crc2, crc3, crc4 }; > > trace_intel_pipe_crc(crtc, crcs); > > @@ -1819,7 +1819,7 @@ static void display_pipe_crc_irq_handler(struct drm_i915_private *dev_priv, > > * don't trust that one either. > > */ > > if (pipe_crc->skipped <= 0 || > > - (INTEL_GEN(dev_priv) >= 8 && pipe_crc->skipped == 1)) { > > + (INTEL_GEN(i915) >= 8 && pipe_crc->skipped == 1)) { > > pipe_crc->skipped++; > > spin_unlock(&pipe_crc->lock); > > return; > > @@ -1832,7 +1832,7 @@ static void display_pipe_crc_irq_handler(struct drm_i915_private *dev_priv, > > } > > #else > > static inline void > > -display_pipe_crc_irq_handler(struct drm_i915_private *dev_priv, > > +display_pipe_crc_irq_handler(struct drm_i915_private *i915, > > enum pipe pipe, > > u32 crc0, u32 crc1, > > u32 crc2, u32 crc3, > > @@ -1840,18 +1840,18 @@ display_pipe_crc_irq_handler(struct drm_i915_private *dev_priv, > > #endif > > -static void hsw_pipe_crc_irq_handler(struct drm_i915_private *dev_priv, > > +static void hsw_pipe_crc_irq_handler(struct drm_i915_private *i915, > > enum pipe pipe) > > { > > - display_pipe_crc_irq_handler(dev_priv, pipe, > > + display_pipe_crc_irq_handler(i915, pipe, > > I915_READ(PIPE_CRC_RES_1_IVB(pipe)), > > 0, 0, 0, 0); > > } > > -static void ivb_pipe_crc_irq_handler(struct drm_i915_private *dev_priv, > > +static void ivb_pipe_crc_irq_handler(struct drm_i915_private *i915, > > enum pipe pipe) > > { > > - display_pipe_crc_irq_handler(dev_priv, pipe, > > + display_pipe_crc_irq_handler(i915, pipe, > > I915_READ(PIPE_CRC_RES_1_IVB(pipe)), > > I915_READ(PIPE_CRC_RES_2_IVB(pipe)), > > I915_READ(PIPE_CRC_RES_3_IVB(pipe)), > > @@ -1859,22 +1859,22 @@ static void ivb_pipe_crc_irq_handler(struct drm_i915_private *dev_priv, > > I915_READ(PIPE_CRC_RES_5_IVB(pipe))); > > } > > -static void i9xx_pipe_crc_irq_handler(struct drm_i915_private *dev_priv, > > +static void i9xx_pipe_crc_irq_handler(struct drm_i915_private *i915, > > enum pipe pipe) > > { > > u32 res1, res2; > > - if (INTEL_GEN(dev_priv) >= 3) > > + if (INTEL_GEN(i915) >= 3) > > res1 = I915_READ(PIPE_CRC_RES_RES1_I915(pipe)); > > else > > res1 = 0; > > - if (INTEL_GEN(dev_priv) >= 5 || IS_G4X(dev_priv)) > > + if (INTEL_GEN(i915) >= 5 || IS_G4X(i915)) > > res2 = I915_READ(PIPE_CRC_RES_RES2_G4X(pipe)); > > else > > res2 = 0; > > - display_pipe_crc_irq_handler(dev_priv, pipe, > > + display_pipe_crc_irq_handler(i915, pipe, > > I915_READ(PIPE_CRC_RES_RED(pipe)), > > I915_READ(PIPE_CRC_RES_GREEN(pipe)), > > I915_READ(PIPE_CRC_RES_BLUE(pipe)), > > @@ -1903,34 +1903,34 @@ static void gen11_rps_irq_handler(struct drm_i915_private *i915, u32 pm_iir) > > schedule_work(&rps->work); > > } > > -static void gen6_rps_irq_handler(struct drm_i915_private *dev_priv, u32 pm_iir) > > +static void gen6_rps_irq_handler(struct drm_i915_private *i915, u32 pm_iir) > > { > > - struct intel_rps *rps = &dev_priv->gt_pm.rps; > > + struct intel_rps *rps = &i915->gt_pm.rps; > > - if (pm_iir & dev_priv->pm_rps_events) { > > - spin_lock(&dev_priv->irq_lock); > > - gen6_mask_pm_irq(dev_priv, pm_iir & dev_priv->pm_rps_events); > > + if (pm_iir & i915->pm_rps_events) { > > + spin_lock(&i915->irq_lock); > > + gen6_mask_pm_irq(i915, pm_iir & i915->pm_rps_events); > > if (rps->interrupts_enabled) { > > - rps->pm_iir |= pm_iir & dev_priv->pm_rps_events; > > + rps->pm_iir |= pm_iir & i915->pm_rps_events; > > schedule_work(&rps->work); > > } > > - spin_unlock(&dev_priv->irq_lock); > > + spin_unlock(&i915->irq_lock); > > } > > - if (INTEL_GEN(dev_priv) >= 8) > > + if (INTEL_GEN(i915) >= 8) > > return; > > if (pm_iir & PM_VEBOX_USER_INTERRUPT) > > - intel_engine_breadcrumbs_irq(dev_priv->engine[VECS0]); > > + intel_engine_breadcrumbs_irq(i915->engine[VECS0]); > > if (pm_iir & PM_VEBOX_CS_ERROR_INTERRUPT) > > DRM_DEBUG("Command parser error, pm_iir 0x%08x\n", pm_iir); > > } > > -static void gen9_guc_irq_handler(struct drm_i915_private *dev_priv, u32 gt_iir) > > +static void gen9_guc_irq_handler(struct drm_i915_private *i915, u32 gt_iir) > > { > > if (gt_iir & GEN9_GUC_TO_HOST_INT_EVENT) > > - intel_guc_to_host_event_handler(&dev_priv->guc); > > + intel_guc_to_host_event_handler(&i915->guc); > > } > > static void gen11_guc_irq_handler(struct drm_i915_private *i915, u16 iir) > > @@ -1939,32 +1939,32 @@ static void gen11_guc_irq_handler(struct drm_i915_private *i915, u16 iir) > > intel_guc_to_host_event_handler(&i915->guc); > > } > > -static void i9xx_pipestat_irq_reset(struct drm_i915_private *dev_priv) > > +static void i9xx_pipestat_irq_reset(struct drm_i915_private *i915) > > { > > enum pipe pipe; > > - for_each_pipe(dev_priv, pipe) { > > + for_each_pipe(i915, pipe) { > > I915_WRITE(PIPESTAT(pipe), > > PIPESTAT_INT_STATUS_MASK | > > PIPE_FIFO_UNDERRUN_STATUS); > > - dev_priv->pipestat_irq_mask[pipe] = 0; > > + i915->pipestat_irq_mask[pipe] = 0; > > } > > } > > -static void i9xx_pipestat_irq_ack(struct drm_i915_private *dev_priv, > > +static void i9xx_pipestat_irq_ack(struct drm_i915_private *i915, > > u32 iir, u32 pipe_stats[I915_MAX_PIPES]) > > { > > int pipe; > > - spin_lock(&dev_priv->irq_lock); > > + spin_lock(&i915->irq_lock); > > - if (!dev_priv->display_irqs_enabled) { > > - spin_unlock(&dev_priv->irq_lock); > > + if (!i915->display_irqs_enabled) { > > + spin_unlock(&i915->irq_lock); > > return; > > } > > - for_each_pipe(dev_priv, pipe) { > > + for_each_pipe(i915, pipe) { > > i915_reg_t reg; > > u32 status_mask, enable_mask, iir_bit = 0; > > @@ -1991,14 +1991,14 @@ static void i9xx_pipestat_irq_ack(struct drm_i915_private *dev_priv, > > break; > > } > > if (iir & iir_bit) > > - status_mask |= dev_priv->pipestat_irq_mask[pipe]; > > + status_mask |= i915->pipestat_irq_mask[pipe]; > > if (!status_mask) > > continue; > > reg = PIPESTAT(pipe); > > pipe_stats[pipe] = I915_READ(reg) & status_mask; > > - enable_mask = i915_pipestat_enable_mask(dev_priv, pipe); > > + enable_mask = i915_pipestat_enable_mask(i915, pipe); > > /* > > * Clear the PIPE*STAT regs before the IIR > > @@ -2014,104 +2014,104 @@ static void i9xx_pipestat_irq_ack(struct drm_i915_private *dev_priv, > > I915_WRITE(reg, enable_mask); > > } > > } > > - spin_unlock(&dev_priv->irq_lock); > > + spin_unlock(&i915->irq_lock); > > } > > -static void i8xx_pipestat_irq_handler(struct drm_i915_private *dev_priv, > > +static void i8xx_pipestat_irq_handler(struct drm_i915_private *i915, > > u16 iir, u32 pipe_stats[I915_MAX_PIPES]) > > { > > enum pipe pipe; > > - for_each_pipe(dev_priv, pipe) { > > + for_each_pipe(i915, pipe) { > > if (pipe_stats[pipe] & PIPE_VBLANK_INTERRUPT_STATUS) > > - drm_handle_vblank(&dev_priv->drm, pipe); > > + drm_handle_vblank(&i915->drm, pipe); > > if (pipe_stats[pipe] & PIPE_CRC_DONE_INTERRUPT_STATUS) > > - i9xx_pipe_crc_irq_handler(dev_priv, pipe); > > + i9xx_pipe_crc_irq_handler(i915, pipe); > > if (pipe_stats[pipe] & PIPE_FIFO_UNDERRUN_STATUS) > > - intel_cpu_fifo_underrun_irq_handler(dev_priv, pipe); > > + intel_cpu_fifo_underrun_irq_handler(i915, pipe); > > } > > } > > -static void i915_pipestat_irq_handler(struct drm_i915_private *dev_priv, > > +static void i915_pipestat_irq_handler(struct drm_i915_private *i915, > > u32 iir, u32 pipe_stats[I915_MAX_PIPES]) > > { > > bool blc_event = false; > > enum pipe pipe; > > - for_each_pipe(dev_priv, pipe) { > > + for_each_pipe(i915, pipe) { > > if (pipe_stats[pipe] & PIPE_VBLANK_INTERRUPT_STATUS) > > - drm_handle_vblank(&dev_priv->drm, pipe); > > + drm_handle_vblank(&i915->drm, pipe); > > if (pipe_stats[pipe] & PIPE_LEGACY_BLC_EVENT_STATUS) > > blc_event = true; > > if (pipe_stats[pipe] & PIPE_CRC_DONE_INTERRUPT_STATUS) > > - i9xx_pipe_crc_irq_handler(dev_priv, pipe); > > + i9xx_pipe_crc_irq_handler(i915, pipe); > > if (pipe_stats[pipe] & PIPE_FIFO_UNDERRUN_STATUS) > > - intel_cpu_fifo_underrun_irq_handler(dev_priv, pipe); > > + intel_cpu_fifo_underrun_irq_handler(i915, pipe); > > } > > if (blc_event || (iir & I915_ASLE_INTERRUPT)) > > - intel_opregion_asle_intr(dev_priv); > > + intel_opregion_asle_intr(i915); > > } > > -static void i965_pipestat_irq_handler(struct drm_i915_private *dev_priv, > > +static void i965_pipestat_irq_handler(struct drm_i915_private *i915, > > u32 iir, u32 pipe_stats[I915_MAX_PIPES]) > > { > > bool blc_event = false; > > enum pipe pipe; > > - for_each_pipe(dev_priv, pipe) { > > + for_each_pipe(i915, pipe) { > > if (pipe_stats[pipe] & PIPE_START_VBLANK_INTERRUPT_STATUS) > > - drm_handle_vblank(&dev_priv->drm, pipe); > > + drm_handle_vblank(&i915->drm, pipe); > > if (pipe_stats[pipe] & PIPE_LEGACY_BLC_EVENT_STATUS) > > blc_event = true; > > if (pipe_stats[pipe] & PIPE_CRC_DONE_INTERRUPT_STATUS) > > - i9xx_pipe_crc_irq_handler(dev_priv, pipe); > > + i9xx_pipe_crc_irq_handler(i915, pipe); > > if (pipe_stats[pipe] & PIPE_FIFO_UNDERRUN_STATUS) > > - intel_cpu_fifo_underrun_irq_handler(dev_priv, pipe); > > + intel_cpu_fifo_underrun_irq_handler(i915, pipe); > > } > > if (blc_event || (iir & I915_ASLE_INTERRUPT)) > > - intel_opregion_asle_intr(dev_priv); > > + intel_opregion_asle_intr(i915); > > if (pipe_stats[0] & PIPE_GMBUS_INTERRUPT_STATUS) > > - gmbus_irq_handler(dev_priv); > > + gmbus_irq_handler(i915); > > } > > -static void valleyview_pipestat_irq_handler(struct drm_i915_private *dev_priv, > > +static void valleyview_pipestat_irq_handler(struct drm_i915_private *i915, > > u32 pipe_stats[I915_MAX_PIPES]) > > { > > enum pipe pipe; > > - for_each_pipe(dev_priv, pipe) { > > + for_each_pipe(i915, pipe) { > > if (pipe_stats[pipe] & PIPE_START_VBLANK_INTERRUPT_STATUS) > > - drm_handle_vblank(&dev_priv->drm, pipe); > > + drm_handle_vblank(&i915->drm, pipe); > > if (pipe_stats[pipe] & PIPE_CRC_DONE_INTERRUPT_STATUS) > > - i9xx_pipe_crc_irq_handler(dev_priv, pipe); > > + i9xx_pipe_crc_irq_handler(i915, pipe); > > if (pipe_stats[pipe] & PIPE_FIFO_UNDERRUN_STATUS) > > - intel_cpu_fifo_underrun_irq_handler(dev_priv, pipe); > > + intel_cpu_fifo_underrun_irq_handler(i915, pipe); > > } > > if (pipe_stats[0] & PIPE_GMBUS_INTERRUPT_STATUS) > > - gmbus_irq_handler(dev_priv); > > + gmbus_irq_handler(i915); > > } > > -static u32 i9xx_hpd_irq_ack(struct drm_i915_private *dev_priv) > > +static u32 i9xx_hpd_irq_ack(struct drm_i915_private *i915) > > { > > u32 hotplug_status = 0, hotplug_status_mask; > > int i; > > - if (IS_G4X(dev_priv) || > > - IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) > > + if (IS_G4X(i915) || > > + IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) > > hotplug_status_mask = HOTPLUG_INT_STATUS_G4X | > > DP_AUX_CHANNEL_MASK_INT_STATUS_G4X; > > else > > @@ -2143,35 +2143,35 @@ static u32 i9xx_hpd_irq_ack(struct drm_i915_private *dev_priv) > > return hotplug_status; > > } > > -static void i9xx_hpd_irq_handler(struct drm_i915_private *dev_priv, > > +static void i9xx_hpd_irq_handler(struct drm_i915_private *i915, > > u32 hotplug_status) > > { > > u32 pin_mask = 0, long_mask = 0; > > - if (IS_G4X(dev_priv) || IS_VALLEYVIEW(dev_priv) || > > - IS_CHERRYVIEW(dev_priv)) { > > + if (IS_G4X(i915) || IS_VALLEYVIEW(i915) || > > + IS_CHERRYVIEW(i915)) { > > u32 hotplug_trigger = hotplug_status & HOTPLUG_INT_STATUS_G4X; > > if (hotplug_trigger) { > > - intel_get_hpd_pins(dev_priv, &pin_mask, &long_mask, > > + intel_get_hpd_pins(i915, &pin_mask, &long_mask, > > hotplug_trigger, hotplug_trigger, > > hpd_status_g4x, > > i9xx_port_hotplug_long_detect); > > - intel_hpd_irq_handler(dev_priv, pin_mask, long_mask); > > + intel_hpd_irq_handler(i915, pin_mask, long_mask); > > } > > if (hotplug_status & DP_AUX_CHANNEL_MASK_INT_STATUS_G4X) > > - dp_aux_irq_handler(dev_priv); > > + dp_aux_irq_handler(i915); > > } else { > > u32 hotplug_trigger = hotplug_status & HOTPLUG_INT_STATUS_I915; > > if (hotplug_trigger) { > > - intel_get_hpd_pins(dev_priv, &pin_mask, &long_mask, > > + intel_get_hpd_pins(i915, &pin_mask, &long_mask, > > hotplug_trigger, hotplug_trigger, > > hpd_status_i915, > > i9xx_port_hotplug_long_detect); > > - intel_hpd_irq_handler(dev_priv, pin_mask, long_mask); > > + intel_hpd_irq_handler(i915, pin_mask, long_mask); > > } > > } > > } > > @@ -2179,14 +2179,14 @@ static void i9xx_hpd_irq_handler(struct drm_i915_private *dev_priv, > > static irqreturn_t valleyview_irq_handler(int irq, void *arg) > > { > > struct drm_device *dev = arg; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > irqreturn_t ret = IRQ_NONE; > > - if (!intel_irqs_enabled(dev_priv)) > > + if (!intel_irqs_enabled(i915)) > > return IRQ_NONE; > > /* IRQs are synced during runtime_suspend, we don't require a wakeref */ > > - disable_rpm_wakeref_asserts(dev_priv); > > + disable_rpm_wakeref_asserts(i915); > > do { > > u32 iir, gt_iir, pm_iir; > > @@ -2226,15 +2226,15 @@ static irqreturn_t valleyview_irq_handler(int irq, void *arg) > > I915_WRITE(GEN6_PMIIR, pm_iir); > > if (iir & I915_DISPLAY_PORT_INTERRUPT) > > - hotplug_status = i9xx_hpd_irq_ack(dev_priv); > > + hotplug_status = i9xx_hpd_irq_ack(i915); > > /* Call regardless, as some status bits might not be > > * signalled in iir */ > > - i9xx_pipestat_irq_ack(dev_priv, iir, pipe_stats); > > + i9xx_pipestat_irq_ack(i915, iir, pipe_stats); > > if (iir & (I915_LPE_PIPE_A_INTERRUPT | > > I915_LPE_PIPE_B_INTERRUPT)) > > - intel_lpe_audio_irq_handler(dev_priv); > > + intel_lpe_audio_irq_handler(i915); > > /* > > * VLV_IIR is single buffered, and reflects the level > > @@ -2247,17 +2247,17 @@ static irqreturn_t valleyview_irq_handler(int irq, void *arg) > > I915_WRITE(VLV_MASTER_IER, MASTER_INTERRUPT_ENABLE); > > if (gt_iir) > > - snb_gt_irq_handler(dev_priv, gt_iir); > > + snb_gt_irq_handler(i915, gt_iir); > > if (pm_iir) > > - gen6_rps_irq_handler(dev_priv, pm_iir); > > + gen6_rps_irq_handler(i915, pm_iir); > > if (hotplug_status) > > - i9xx_hpd_irq_handler(dev_priv, hotplug_status); > > + i9xx_hpd_irq_handler(i915, hotplug_status); > > - valleyview_pipestat_irq_handler(dev_priv, pipe_stats); > > + valleyview_pipestat_irq_handler(i915, pipe_stats); > > } while (0); > > - enable_rpm_wakeref_asserts(dev_priv); > > + enable_rpm_wakeref_asserts(i915); > > return ret; > > } > > @@ -2265,14 +2265,14 @@ static irqreturn_t valleyview_irq_handler(int irq, void *arg) > > static irqreturn_t cherryview_irq_handler(int irq, void *arg) > > { > > struct drm_device *dev = arg; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > irqreturn_t ret = IRQ_NONE; > > - if (!intel_irqs_enabled(dev_priv)) > > + if (!intel_irqs_enabled(i915)) > > return IRQ_NONE; > > /* IRQs are synced during runtime_suspend, we don't require a wakeref */ > > - disable_rpm_wakeref_asserts(dev_priv); > > + disable_rpm_wakeref_asserts(i915); > > do { > > u32 master_ctl, iir; > > @@ -2306,19 +2306,19 @@ static irqreturn_t cherryview_irq_handler(int irq, void *arg) > > ier = I915_READ(VLV_IER); > > I915_WRITE(VLV_IER, 0); > > - gen8_gt_irq_ack(dev_priv, master_ctl, gt_iir); > > + gen8_gt_irq_ack(i915, master_ctl, gt_iir); > > if (iir & I915_DISPLAY_PORT_INTERRUPT) > > - hotplug_status = i9xx_hpd_irq_ack(dev_priv); > > + hotplug_status = i9xx_hpd_irq_ack(i915); > > /* Call regardless, as some status bits might not be > > * signalled in iir */ > > - i9xx_pipestat_irq_ack(dev_priv, iir, pipe_stats); > > + i9xx_pipestat_irq_ack(i915, iir, pipe_stats); > > if (iir & (I915_LPE_PIPE_A_INTERRUPT | > > I915_LPE_PIPE_B_INTERRUPT | > > I915_LPE_PIPE_C_INTERRUPT)) > > - intel_lpe_audio_irq_handler(dev_priv); > > + intel_lpe_audio_irq_handler(i915); > > /* > > * VLV_IIR is single buffered, and reflects the level > > @@ -2330,20 +2330,20 @@ static irqreturn_t cherryview_irq_handler(int irq, void *arg) > > I915_WRITE(VLV_IER, ier); > > I915_WRITE(GEN8_MASTER_IRQ, GEN8_MASTER_IRQ_CONTROL); > > - gen8_gt_irq_handler(dev_priv, master_ctl, gt_iir); > > + gen8_gt_irq_handler(i915, master_ctl, gt_iir); > > if (hotplug_status) > > - i9xx_hpd_irq_handler(dev_priv, hotplug_status); > > + i9xx_hpd_irq_handler(i915, hotplug_status); > > - valleyview_pipestat_irq_handler(dev_priv, pipe_stats); > > + valleyview_pipestat_irq_handler(i915, pipe_stats); > > } while (0); > > - enable_rpm_wakeref_asserts(dev_priv); > > + enable_rpm_wakeref_asserts(i915); > > return ret; > > } > > -static void ibx_hpd_irq_handler(struct drm_i915_private *dev_priv, > > +static void ibx_hpd_irq_handler(struct drm_i915_private *i915, > > u32 hotplug_trigger, > > const u32 hpd[HPD_NUM_PINS]) > > { > > @@ -2368,19 +2368,19 @@ static void ibx_hpd_irq_handler(struct drm_i915_private *dev_priv, > > if (!hotplug_trigger) > > return; > > - intel_get_hpd_pins(dev_priv, &pin_mask, &long_mask, hotplug_trigger, > > + intel_get_hpd_pins(i915, &pin_mask, &long_mask, hotplug_trigger, > > dig_hotplug_reg, hpd, > > pch_port_hotplug_long_detect); > > - intel_hpd_irq_handler(dev_priv, pin_mask, long_mask); > > + intel_hpd_irq_handler(i915, pin_mask, long_mask); > > } > > -static void ibx_irq_handler(struct drm_i915_private *dev_priv, u32 pch_iir) > > +static void ibx_irq_handler(struct drm_i915_private *i915, u32 pch_iir) > > { > > int pipe; > > u32 hotplug_trigger = pch_iir & SDE_HOTPLUG_MASK; > > - ibx_hpd_irq_handler(dev_priv, hotplug_trigger, hpd_ibx); > > + ibx_hpd_irq_handler(i915, hotplug_trigger, hpd_ibx); > > if (pch_iir & SDE_AUDIO_POWER_MASK) { > > int port = ffs((pch_iir & SDE_AUDIO_POWER_MASK) >> > > @@ -2390,10 +2390,10 @@ static void ibx_irq_handler(struct drm_i915_private *dev_priv, u32 pch_iir) > > } > > if (pch_iir & SDE_AUX_MASK) > > - dp_aux_irq_handler(dev_priv); > > + dp_aux_irq_handler(i915); > > if (pch_iir & SDE_GMBUS) > > - gmbus_irq_handler(dev_priv); > > + gmbus_irq_handler(i915); > > if (pch_iir & SDE_AUDIO_HDCP_MASK) > > DRM_DEBUG_DRIVER("PCH HDCP audio interrupt\n"); > > @@ -2405,7 +2405,7 @@ static void ibx_irq_handler(struct drm_i915_private *dev_priv, u32 pch_iir) > > DRM_ERROR("PCH poison interrupt\n"); > > if (pch_iir & SDE_FDI_MASK) > > - for_each_pipe(dev_priv, pipe) > > + for_each_pipe(i915, pipe) > > DRM_DEBUG_DRIVER(" pipe %c FDI IIR: 0x%08x\n", > > pipe_name(pipe), > > I915_READ(FDI_RX_IIR(pipe))); > > @@ -2417,13 +2417,13 @@ static void ibx_irq_handler(struct drm_i915_private *dev_priv, u32 pch_iir) > > DRM_DEBUG_DRIVER("PCH transcoder CRC error interrupt\n"); > > if (pch_iir & SDE_TRANSA_FIFO_UNDER) > > - intel_pch_fifo_underrun_irq_handler(dev_priv, PIPE_A); > > + intel_pch_fifo_underrun_irq_handler(i915, PIPE_A); > > if (pch_iir & SDE_TRANSB_FIFO_UNDER) > > - intel_pch_fifo_underrun_irq_handler(dev_priv, PIPE_B); > > + intel_pch_fifo_underrun_irq_handler(i915, PIPE_B); > > } > > -static void ivb_err_int_handler(struct drm_i915_private *dev_priv) > > +static void ivb_err_int_handler(struct drm_i915_private *i915) > > { > > u32 err_int = I915_READ(GEN7_ERR_INT); > > enum pipe pipe; > > @@ -2431,22 +2431,22 @@ static void ivb_err_int_handler(struct drm_i915_private *dev_priv) > > if (err_int & ERR_INT_POISON) > > DRM_ERROR("Poison interrupt\n"); > > - for_each_pipe(dev_priv, pipe) { > > + for_each_pipe(i915, pipe) { > > if (err_int & ERR_INT_FIFO_UNDERRUN(pipe)) > > - intel_cpu_fifo_underrun_irq_handler(dev_priv, pipe); > > + intel_cpu_fifo_underrun_irq_handler(i915, pipe); > > if (err_int & ERR_INT_PIPE_CRC_DONE(pipe)) { > > - if (IS_IVYBRIDGE(dev_priv)) > > - ivb_pipe_crc_irq_handler(dev_priv, pipe); > > + if (IS_IVYBRIDGE(i915)) > > + ivb_pipe_crc_irq_handler(i915, pipe); > > else > > - hsw_pipe_crc_irq_handler(dev_priv, pipe); > > + hsw_pipe_crc_irq_handler(i915, pipe); > > } > > } > > I915_WRITE(GEN7_ERR_INT, err_int); > > } > > -static void cpt_serr_int_handler(struct drm_i915_private *dev_priv) > > +static void cpt_serr_int_handler(struct drm_i915_private *i915) > > { > > u32 serr_int = I915_READ(SERR_INT); > > enum pipe pipe; > > @@ -2454,19 +2454,19 @@ static void cpt_serr_int_handler(struct drm_i915_private *dev_priv) > > if (serr_int & SERR_INT_POISON) > > DRM_ERROR("PCH poison interrupt\n"); > > - for_each_pipe(dev_priv, pipe) > > + for_each_pipe(i915, pipe) > > if (serr_int & SERR_INT_TRANS_FIFO_UNDERRUN(pipe)) > > - intel_pch_fifo_underrun_irq_handler(dev_priv, pipe); > > + intel_pch_fifo_underrun_irq_handler(i915, pipe); > > I915_WRITE(SERR_INT, serr_int); > > } > > -static void cpt_irq_handler(struct drm_i915_private *dev_priv, u32 pch_iir) > > +static void cpt_irq_handler(struct drm_i915_private *i915, u32 pch_iir) > > { > > int pipe; > > u32 hotplug_trigger = pch_iir & SDE_HOTPLUG_MASK_CPT; > > - ibx_hpd_irq_handler(dev_priv, hotplug_trigger, hpd_cpt); > > + ibx_hpd_irq_handler(i915, hotplug_trigger, hpd_cpt); > > if (pch_iir & SDE_AUDIO_POWER_MASK_CPT) { > > int port = ffs((pch_iir & SDE_AUDIO_POWER_MASK_CPT) >> > > @@ -2476,10 +2476,10 @@ static void cpt_irq_handler(struct drm_i915_private *dev_priv, u32 pch_iir) > > } > > if (pch_iir & SDE_AUX_MASK_CPT) > > - dp_aux_irq_handler(dev_priv); > > + dp_aux_irq_handler(i915); > > if (pch_iir & SDE_GMBUS_CPT) > > - gmbus_irq_handler(dev_priv); > > + gmbus_irq_handler(i915); > > if (pch_iir & SDE_AUDIO_CP_REQ_CPT) > > DRM_DEBUG_DRIVER("Audio CP request interrupt\n"); > > @@ -2488,16 +2488,16 @@ static void cpt_irq_handler(struct drm_i915_private *dev_priv, u32 pch_iir) > > DRM_DEBUG_DRIVER("Audio CP change interrupt\n"); > > if (pch_iir & SDE_FDI_MASK_CPT) > > - for_each_pipe(dev_priv, pipe) > > + for_each_pipe(i915, pipe) > > DRM_DEBUG_DRIVER(" pipe %c FDI IIR: 0x%08x\n", > > pipe_name(pipe), > > I915_READ(FDI_RX_IIR(pipe))); > > if (pch_iir & SDE_ERROR_CPT) > > - cpt_serr_int_handler(dev_priv); > > + cpt_serr_int_handler(i915); > > } > > -static void icp_irq_handler(struct drm_i915_private *dev_priv, u32 pch_iir) > > +static void icp_irq_handler(struct drm_i915_private *i915, u32 pch_iir) > > { > > u32 ddi_hotplug_trigger = pch_iir & SDE_DDI_MASK_ICP; > > u32 tc_hotplug_trigger = pch_iir & SDE_TC_MASK_ICP; > > @@ -2509,7 +2509,7 @@ static void icp_irq_handler(struct drm_i915_private *dev_priv, u32 pch_iir) > > dig_hotplug_reg = I915_READ(SHOTPLUG_CTL_DDI); > > I915_WRITE(SHOTPLUG_CTL_DDI, dig_hotplug_reg); > > - intel_get_hpd_pins(dev_priv, &pin_mask, &long_mask, > > + intel_get_hpd_pins(i915, &pin_mask, &long_mask, > > ddi_hotplug_trigger, > > dig_hotplug_reg, hpd_icp, > > icp_ddi_port_hotplug_long_detect); > > @@ -2521,20 +2521,20 @@ static void icp_irq_handler(struct drm_i915_private *dev_priv, u32 pch_iir) > > dig_hotplug_reg = I915_READ(SHOTPLUG_CTL_TC); > > I915_WRITE(SHOTPLUG_CTL_TC, dig_hotplug_reg); > > - intel_get_hpd_pins(dev_priv, &pin_mask, &long_mask, > > + intel_get_hpd_pins(i915, &pin_mask, &long_mask, > > tc_hotplug_trigger, > > dig_hotplug_reg, hpd_icp, > > icp_tc_port_hotplug_long_detect); > > } > > if (pin_mask) > > - intel_hpd_irq_handler(dev_priv, pin_mask, long_mask); > > + intel_hpd_irq_handler(i915, pin_mask, long_mask); > > if (pch_iir & SDE_GMBUS_ICP) > > - gmbus_irq_handler(dev_priv); > > + gmbus_irq_handler(i915); > > } > > -static void spt_irq_handler(struct drm_i915_private *dev_priv, u32 pch_iir) > > +static void spt_irq_handler(struct drm_i915_private *i915, u32 pch_iir) > > { > > u32 hotplug_trigger = pch_iir & SDE_HOTPLUG_MASK_SPT & > > ~SDE_PORTE_HOTPLUG_SPT; > > @@ -2547,7 +2547,7 @@ static void spt_irq_handler(struct drm_i915_private *dev_priv, u32 pch_iir) > > dig_hotplug_reg = I915_READ(PCH_PORT_HOTPLUG); > > I915_WRITE(PCH_PORT_HOTPLUG, dig_hotplug_reg); > > - intel_get_hpd_pins(dev_priv, &pin_mask, &long_mask, > > + intel_get_hpd_pins(i915, &pin_mask, &long_mask, > > hotplug_trigger, dig_hotplug_reg, hpd_spt, > > spt_port_hotplug_long_detect); > > } > > @@ -2558,19 +2558,19 @@ static void spt_irq_handler(struct drm_i915_private *dev_priv, u32 pch_iir) > > dig_hotplug_reg = I915_READ(PCH_PORT_HOTPLUG2); > > I915_WRITE(PCH_PORT_HOTPLUG2, dig_hotplug_reg); > > - intel_get_hpd_pins(dev_priv, &pin_mask, &long_mask, > > + intel_get_hpd_pins(i915, &pin_mask, &long_mask, > > hotplug2_trigger, dig_hotplug_reg, hpd_spt, > > spt_port_hotplug2_long_detect); > > } > > if (pin_mask) > > - intel_hpd_irq_handler(dev_priv, pin_mask, long_mask); > > + intel_hpd_irq_handler(i915, pin_mask, long_mask); > > if (pch_iir & SDE_GMBUS_CPT) > > - gmbus_irq_handler(dev_priv); > > + gmbus_irq_handler(i915); > > } > > -static void ilk_hpd_irq_handler(struct drm_i915_private *dev_priv, > > +static void ilk_hpd_irq_handler(struct drm_i915_private *i915, > > u32 hotplug_trigger, > > const u32 hpd[HPD_NUM_PINS]) > > { > > @@ -2579,94 +2579,94 @@ static void ilk_hpd_irq_handler(struct drm_i915_private *dev_priv, > > dig_hotplug_reg = I915_READ(DIGITAL_PORT_HOTPLUG_CNTRL); > > I915_WRITE(DIGITAL_PORT_HOTPLUG_CNTRL, dig_hotplug_reg); > > - intel_get_hpd_pins(dev_priv, &pin_mask, &long_mask, hotplug_trigger, > > + intel_get_hpd_pins(i915, &pin_mask, &long_mask, hotplug_trigger, > > dig_hotplug_reg, hpd, > > ilk_port_hotplug_long_detect); > > - intel_hpd_irq_handler(dev_priv, pin_mask, long_mask); > > + intel_hpd_irq_handler(i915, pin_mask, long_mask); > > } > > -static void ilk_display_irq_handler(struct drm_i915_private *dev_priv, > > +static void ilk_display_irq_handler(struct drm_i915_private *i915, > > u32 de_iir) > > { > > enum pipe pipe; > > u32 hotplug_trigger = de_iir & DE_DP_A_HOTPLUG; > > if (hotplug_trigger) > > - ilk_hpd_irq_handler(dev_priv, hotplug_trigger, hpd_ilk); > > + ilk_hpd_irq_handler(i915, hotplug_trigger, hpd_ilk); > > if (de_iir & DE_AUX_CHANNEL_A) > > - dp_aux_irq_handler(dev_priv); > > + dp_aux_irq_handler(i915); > > if (de_iir & DE_GSE) > > - intel_opregion_asle_intr(dev_priv); > > + intel_opregion_asle_intr(i915); > > if (de_iir & DE_POISON) > > DRM_ERROR("Poison interrupt\n"); > > - for_each_pipe(dev_priv, pipe) { > > + for_each_pipe(i915, pipe) { > > if (de_iir & DE_PIPE_VBLANK(pipe)) > > - drm_handle_vblank(&dev_priv->drm, pipe); > > + drm_handle_vblank(&i915->drm, pipe); > > if (de_iir & DE_PIPE_FIFO_UNDERRUN(pipe)) > > - intel_cpu_fifo_underrun_irq_handler(dev_priv, pipe); > > + intel_cpu_fifo_underrun_irq_handler(i915, pipe); > > if (de_iir & DE_PIPE_CRC_DONE(pipe)) > > - i9xx_pipe_crc_irq_handler(dev_priv, pipe); > > + i9xx_pipe_crc_irq_handler(i915, pipe); > > } > > /* check event from PCH */ > > if (de_iir & DE_PCH_EVENT) { > > u32 pch_iir = I915_READ(SDEIIR); > > - if (HAS_PCH_CPT(dev_priv)) > > - cpt_irq_handler(dev_priv, pch_iir); > > + if (HAS_PCH_CPT(i915)) > > + cpt_irq_handler(i915, pch_iir); > > else > > - ibx_irq_handler(dev_priv, pch_iir); > > + ibx_irq_handler(i915, pch_iir); > > /* should clear PCH hotplug event before clear CPU irq */ > > I915_WRITE(SDEIIR, pch_iir); > > } > > - if (IS_GEN(dev_priv, 5) && de_iir & DE_PCU_EVENT) > > - ironlake_rps_change_irq_handler(dev_priv); > > + if (IS_GEN(i915, 5) && de_iir & DE_PCU_EVENT) > > + ironlake_rps_change_irq_handler(i915); > > } > > -static void ivb_display_irq_handler(struct drm_i915_private *dev_priv, > > +static void ivb_display_irq_handler(struct drm_i915_private *i915, > > u32 de_iir) > > { > > enum pipe pipe; > > u32 hotplug_trigger = de_iir & DE_DP_A_HOTPLUG_IVB; > > if (hotplug_trigger) > > - ilk_hpd_irq_handler(dev_priv, hotplug_trigger, hpd_ivb); > > + ilk_hpd_irq_handler(i915, hotplug_trigger, hpd_ivb); > > if (de_iir & DE_ERR_INT_IVB) > > - ivb_err_int_handler(dev_priv); > > + ivb_err_int_handler(i915); > > if (de_iir & DE_EDP_PSR_INT_HSW) { > > u32 psr_iir = I915_READ(EDP_PSR_IIR); > > - intel_psr_irq_handler(dev_priv, psr_iir); > > + intel_psr_irq_handler(i915, psr_iir); > > I915_WRITE(EDP_PSR_IIR, psr_iir); > > } > > if (de_iir & DE_AUX_CHANNEL_A_IVB) > > - dp_aux_irq_handler(dev_priv); > > + dp_aux_irq_handler(i915); > > if (de_iir & DE_GSE_IVB) > > - intel_opregion_asle_intr(dev_priv); > > + intel_opregion_asle_intr(i915); > > - for_each_pipe(dev_priv, pipe) { > > + for_each_pipe(i915, pipe) { > > if (de_iir & (DE_PIPE_VBLANK_IVB(pipe))) > > - drm_handle_vblank(&dev_priv->drm, pipe); > > + drm_handle_vblank(&i915->drm, pipe); > > } > > /* check event from PCH */ > > - if (!HAS_PCH_NOP(dev_priv) && (de_iir & DE_PCH_EVENT_IVB)) { > > + if (!HAS_PCH_NOP(i915) && (de_iir & DE_PCH_EVENT_IVB)) { > > u32 pch_iir = I915_READ(SDEIIR); > > - cpt_irq_handler(dev_priv, pch_iir); > > + cpt_irq_handler(i915, pch_iir); > > /* clear PCH hotplug event before clear CPU irq */ > > I915_WRITE(SDEIIR, pch_iir); > > @@ -2684,15 +2684,15 @@ static void ivb_display_irq_handler(struct drm_i915_private *dev_priv, > > static irqreturn_t ironlake_irq_handler(int irq, void *arg) > > { > > struct drm_device *dev = arg; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > u32 de_iir, gt_iir, de_ier, sde_ier = 0; > > irqreturn_t ret = IRQ_NONE; > > - if (!intel_irqs_enabled(dev_priv)) > > + if (!intel_irqs_enabled(i915)) > > return IRQ_NONE; > > /* IRQs are synced during runtime_suspend, we don't require a wakeref */ > > - disable_rpm_wakeref_asserts(dev_priv); > > + disable_rpm_wakeref_asserts(i915); > > /* disable master interrupt before clearing iir */ > > de_ier = I915_READ(DEIER); > > @@ -2703,7 +2703,7 @@ static irqreturn_t ironlake_irq_handler(int irq, void *arg) > > * able to process them after we restore SDEIER (as soon as we restore > > * it, we'll get an interrupt if SDEIIR still has something to process > > * due to its back queue). */ > > - if (!HAS_PCH_NOP(dev_priv)) { > > + if (!HAS_PCH_NOP(i915)) { > > sde_ier = I915_READ(SDEIER); > > I915_WRITE(SDEIER, 0); > > } > > @@ -2714,42 +2714,42 @@ static irqreturn_t ironlake_irq_handler(int irq, void *arg) > > if (gt_iir) { > > I915_WRITE(GTIIR, gt_iir); > > ret = IRQ_HANDLED; > > - if (INTEL_GEN(dev_priv) >= 6) > > - snb_gt_irq_handler(dev_priv, gt_iir); > > + if (INTEL_GEN(i915) >= 6) > > + snb_gt_irq_handler(i915, gt_iir); > > else > > - ilk_gt_irq_handler(dev_priv, gt_iir); > > + ilk_gt_irq_handler(i915, gt_iir); > > } > > de_iir = I915_READ(DEIIR); > > if (de_iir) { > > I915_WRITE(DEIIR, de_iir); > > ret = IRQ_HANDLED; > > - if (INTEL_GEN(dev_priv) >= 7) > > - ivb_display_irq_handler(dev_priv, de_iir); > > + if (INTEL_GEN(i915) >= 7) > > + ivb_display_irq_handler(i915, de_iir); > > else > > - ilk_display_irq_handler(dev_priv, de_iir); > > + ilk_display_irq_handler(i915, de_iir); > > } > > - if (INTEL_GEN(dev_priv) >= 6) { > > + if (INTEL_GEN(i915) >= 6) { > > u32 pm_iir = I915_READ(GEN6_PMIIR); > > if (pm_iir) { > > I915_WRITE(GEN6_PMIIR, pm_iir); > > ret = IRQ_HANDLED; > > - gen6_rps_irq_handler(dev_priv, pm_iir); > > + gen6_rps_irq_handler(i915, pm_iir); > > } > > } > > I915_WRITE(DEIER, de_ier); > > - if (!HAS_PCH_NOP(dev_priv)) > > + if (!HAS_PCH_NOP(i915)) > > I915_WRITE(SDEIER, sde_ier); > > /* IRQs are synced during runtime_suspend, we don't require a wakeref */ > > - enable_rpm_wakeref_asserts(dev_priv); > > + enable_rpm_wakeref_asserts(i915); > > return ret; > > } > > -static void bxt_hpd_irq_handler(struct drm_i915_private *dev_priv, > > +static void bxt_hpd_irq_handler(struct drm_i915_private *i915, > > u32 hotplug_trigger, > > const u32 hpd[HPD_NUM_PINS]) > > { > > @@ -2758,14 +2758,14 @@ static void bxt_hpd_irq_handler(struct drm_i915_private *dev_priv, > > dig_hotplug_reg = I915_READ(PCH_PORT_HOTPLUG); > > I915_WRITE(PCH_PORT_HOTPLUG, dig_hotplug_reg); > > - intel_get_hpd_pins(dev_priv, &pin_mask, &long_mask, hotplug_trigger, > > + intel_get_hpd_pins(i915, &pin_mask, &long_mask, hotplug_trigger, > > dig_hotplug_reg, hpd, > > bxt_port_hotplug_long_detect); > > - intel_hpd_irq_handler(dev_priv, pin_mask, long_mask); > > + intel_hpd_irq_handler(i915, pin_mask, long_mask); > > } > > -static void gen11_hpd_irq_handler(struct drm_i915_private *dev_priv, u32 iir) > > +static void gen11_hpd_irq_handler(struct drm_i915_private *i915, u32 iir) > > { > > u32 pin_mask = 0, long_mask = 0; > > u32 trigger_tc = iir & GEN11_DE_TC_HOTPLUG_MASK; > > @@ -2777,7 +2777,7 @@ static void gen11_hpd_irq_handler(struct drm_i915_private *dev_priv, u32 iir) > > dig_hotplug_reg = I915_READ(GEN11_TC_HOTPLUG_CTL); > > I915_WRITE(GEN11_TC_HOTPLUG_CTL, dig_hotplug_reg); > > - intel_get_hpd_pins(dev_priv, &pin_mask, &long_mask, trigger_tc, > > + intel_get_hpd_pins(i915, &pin_mask, &long_mask, trigger_tc, > > dig_hotplug_reg, hpd_gen11, > > gen11_port_hotplug_long_detect); > > } > > @@ -2788,30 +2788,30 @@ static void gen11_hpd_irq_handler(struct drm_i915_private *dev_priv, u32 iir) > > dig_hotplug_reg = I915_READ(GEN11_TBT_HOTPLUG_CTL); > > I915_WRITE(GEN11_TBT_HOTPLUG_CTL, dig_hotplug_reg); > > - intel_get_hpd_pins(dev_priv, &pin_mask, &long_mask, trigger_tbt, > > + intel_get_hpd_pins(i915, &pin_mask, &long_mask, trigger_tbt, > > dig_hotplug_reg, hpd_gen11, > > gen11_port_hotplug_long_detect); > > } > > if (pin_mask) > > - intel_hpd_irq_handler(dev_priv, pin_mask, long_mask); > > + intel_hpd_irq_handler(i915, pin_mask, long_mask); > > else > > DRM_ERROR("Unexpected DE HPD interrupt 0x%08x\n", iir); > > } > > -static u32 gen8_de_port_aux_mask(struct drm_i915_private *dev_priv) > > +static u32 gen8_de_port_aux_mask(struct drm_i915_private *i915) > > { > > u32 mask = GEN8_AUX_CHANNEL_A; > > - if (INTEL_GEN(dev_priv) >= 9) > > + if (INTEL_GEN(i915) >= 9) > > mask |= GEN9_AUX_CHANNEL_B | > > GEN9_AUX_CHANNEL_C | > > GEN9_AUX_CHANNEL_D; > > - if (IS_CNL_WITH_PORT_F(dev_priv)) > > + if (IS_CNL_WITH_PORT_F(i915)) > > mask |= CNL_AUX_CHANNEL_F; > > - if (INTEL_GEN(dev_priv) >= 11) > > + if (INTEL_GEN(i915) >= 11) > > mask |= ICL_AUX_CHANNEL_E | > > CNL_AUX_CHANNEL_F; > > @@ -2819,7 +2819,7 @@ static u32 gen8_de_port_aux_mask(struct drm_i915_private *dev_priv) > > } > > static irqreturn_t > > -gen8_de_irq_handler(struct drm_i915_private *dev_priv, u32 master_ctl) > > +gen8_de_irq_handler(struct drm_i915_private *i915, u32 master_ctl) > > { > > irqreturn_t ret = IRQ_NONE; > > u32 iir; > > @@ -2834,14 +2834,14 @@ gen8_de_irq_handler(struct drm_i915_private *dev_priv, u32 master_ctl) > > ret = IRQ_HANDLED; > > if (iir & GEN8_DE_MISC_GSE) { > > - intel_opregion_asle_intr(dev_priv); > > + intel_opregion_asle_intr(i915); > > found = true; > > } > > if (iir & GEN8_DE_EDP_PSR) { > > u32 psr_iir = I915_READ(EDP_PSR_IIR); > > - intel_psr_irq_handler(dev_priv, psr_iir); > > + intel_psr_irq_handler(i915, psr_iir); > > I915_WRITE(EDP_PSR_IIR, psr_iir); > > found = true; > > } > > @@ -2853,12 +2853,12 @@ gen8_de_irq_handler(struct drm_i915_private *dev_priv, u32 master_ctl) > > DRM_ERROR("The master control interrupt lied (DE MISC)!\n"); > > } > > - if (INTEL_GEN(dev_priv) >= 11 && (master_ctl & GEN11_DE_HPD_IRQ)) { > > + if (INTEL_GEN(i915) >= 11 && (master_ctl & GEN11_DE_HPD_IRQ)) { > > iir = I915_READ(GEN11_DE_HPD_IIR); > > if (iir) { > > I915_WRITE(GEN11_DE_HPD_IIR, iir); > > ret = IRQ_HANDLED; > > - gen11_hpd_irq_handler(dev_priv, iir); > > + gen11_hpd_irq_handler(i915, iir); > > } else { > > DRM_ERROR("The master control interrupt lied, (DE HPD)!\n"); > > } > > @@ -2873,29 +2873,29 @@ gen8_de_irq_handler(struct drm_i915_private *dev_priv, u32 master_ctl) > > I915_WRITE(GEN8_DE_PORT_IIR, iir); > > ret = IRQ_HANDLED; > > - if (iir & gen8_de_port_aux_mask(dev_priv)) { > > - dp_aux_irq_handler(dev_priv); > > + if (iir & gen8_de_port_aux_mask(i915)) { > > + dp_aux_irq_handler(i915); > > found = true; > > } > > - if (IS_GEN9_LP(dev_priv)) { > > + if (IS_GEN9_LP(i915)) { > > tmp_mask = iir & BXT_DE_PORT_HOTPLUG_MASK; > > if (tmp_mask) { > > - bxt_hpd_irq_handler(dev_priv, tmp_mask, > > + bxt_hpd_irq_handler(i915, tmp_mask, > > hpd_bxt); > > found = true; > > } > > - } else if (IS_BROADWELL(dev_priv)) { > > + } else if (IS_BROADWELL(i915)) { > > tmp_mask = iir & GEN8_PORT_DP_A_HOTPLUG; > > if (tmp_mask) { > > - ilk_hpd_irq_handler(dev_priv, > > + ilk_hpd_irq_handler(i915, > > tmp_mask, hpd_bdw); > > found = true; > > } > > } > > - if (IS_GEN9_LP(dev_priv) && (iir & BXT_DE_PORT_GMBUS)) { > > - gmbus_irq_handler(dev_priv); > > + if (IS_GEN9_LP(i915) && (iir & BXT_DE_PORT_GMBUS)) { > > + gmbus_irq_handler(i915); > > found = true; > > } > > @@ -2906,7 +2906,7 @@ gen8_de_irq_handler(struct drm_i915_private *dev_priv, u32 master_ctl) > > DRM_ERROR("The master control interrupt lied (DE PORT)!\n"); > > } > > - for_each_pipe(dev_priv, pipe) { > > + for_each_pipe(i915, pipe) { > > u32 fault_errors; > > if (!(master_ctl & GEN8_DE_PIPE_IRQ(pipe))) > > @@ -2922,16 +2922,16 @@ gen8_de_irq_handler(struct drm_i915_private *dev_priv, u32 master_ctl) > > I915_WRITE(GEN8_DE_PIPE_IIR(pipe), iir); > > if (iir & GEN8_PIPE_VBLANK) > > - drm_handle_vblank(&dev_priv->drm, pipe); > > + drm_handle_vblank(&i915->drm, pipe); > > if (iir & GEN8_PIPE_CDCLK_CRC_DONE) > > - hsw_pipe_crc_irq_handler(dev_priv, pipe); > > + hsw_pipe_crc_irq_handler(i915, pipe); > > if (iir & GEN8_PIPE_FIFO_UNDERRUN) > > - intel_cpu_fifo_underrun_irq_handler(dev_priv, pipe); > > + intel_cpu_fifo_underrun_irq_handler(i915, pipe); > > fault_errors = iir; > > - if (INTEL_GEN(dev_priv) >= 9) > > + if (INTEL_GEN(i915) >= 9) > > fault_errors &= GEN9_DE_PIPE_IRQ_FAULT_ERRORS; > > else > > fault_errors &= GEN8_DE_PIPE_IRQ_FAULT_ERRORS; > > @@ -2942,7 +2942,7 @@ gen8_de_irq_handler(struct drm_i915_private *dev_priv, u32 master_ctl) > > fault_errors); > > } > > - if (HAS_PCH_SPLIT(dev_priv) && !HAS_PCH_NOP(dev_priv) && > > + if (HAS_PCH_SPLIT(i915) && !HAS_PCH_NOP(i915) && > > master_ctl & GEN8_DE_PCH_IRQ) { > > /* > > * FIXME(BDW): Assume for now that the new interrupt handling > > @@ -2954,12 +2954,12 @@ gen8_de_irq_handler(struct drm_i915_private *dev_priv, u32 master_ctl) > > I915_WRITE(SDEIIR, iir); > > ret = IRQ_HANDLED; > > - if (INTEL_PCH_TYPE(dev_priv) >= PCH_ICP) > > - icp_irq_handler(dev_priv, iir); > > - else if (INTEL_PCH_TYPE(dev_priv) >= PCH_SPT) > > - spt_irq_handler(dev_priv, iir); > > + if (INTEL_PCH_TYPE(i915) >= PCH_ICP) > > + icp_irq_handler(i915, iir); > > + else if (INTEL_PCH_TYPE(i915) >= PCH_SPT) > > + spt_irq_handler(i915, iir); > > else > > - cpt_irq_handler(dev_priv, iir); > > + cpt_irq_handler(i915, iir); > > } else { > > /* > > * Like on previous PCH there seems to be something > > @@ -2992,12 +2992,12 @@ static inline void gen8_master_intr_enable(void __iomem * const regs) > > static irqreturn_t gen8_irq_handler(int irq, void *arg) > > { > > - struct drm_i915_private *dev_priv = to_i915(arg); > > - void __iomem * const regs = dev_priv->uncore.regs; > > + struct drm_i915_private *i915 = to_i915(arg); > > + void __iomem * const regs = i915->uncore.regs; > > u32 master_ctl; > > u32 gt_iir[4]; > > - if (!intel_irqs_enabled(dev_priv)) > > + if (!intel_irqs_enabled(i915)) > > return IRQ_NONE; > > master_ctl = gen8_master_intr_disable(regs); > > @@ -3007,18 +3007,18 @@ static irqreturn_t gen8_irq_handler(int irq, void *arg) > > } > > /* Find, clear, then process each source of interrupt */ > > - gen8_gt_irq_ack(dev_priv, master_ctl, gt_iir); > > + gen8_gt_irq_ack(i915, master_ctl, gt_iir); > > /* IRQs are synced during runtime_suspend, we don't require a wakeref */ > > if (master_ctl & ~GEN8_GT_IRQS) { > > - disable_rpm_wakeref_asserts(dev_priv); > > - gen8_de_irq_handler(dev_priv, master_ctl); > > - enable_rpm_wakeref_asserts(dev_priv); > > + disable_rpm_wakeref_asserts(i915); > > + gen8_de_irq_handler(i915, master_ctl); > > + enable_rpm_wakeref_asserts(i915); > > } > > gen8_master_intr_enable(regs); > > - gen8_gt_irq_handler(dev_priv, master_ctl, gt_iir); > > + gen8_gt_irq_handler(i915, master_ctl, gt_iir); > > return IRQ_HANDLED; > > } > > @@ -3149,9 +3149,9 @@ gen11_gt_irq_handler(struct drm_i915_private * const i915, > > } > > static u32 > > -gen11_gu_misc_irq_ack(struct drm_i915_private *dev_priv, const u32 master_ctl) > > +gen11_gu_misc_irq_ack(struct drm_i915_private *i915, const u32 master_ctl) > > { > > - void __iomem * const regs = dev_priv->uncore.regs; > > + void __iomem * const regs = i915->uncore.regs; > > u32 iir; > > if (!(master_ctl & GEN11_GU_MISC_IRQ)) > > @@ -3165,10 +3165,10 @@ gen11_gu_misc_irq_ack(struct drm_i915_private *dev_priv, const u32 master_ctl) > > } > > static void > > -gen11_gu_misc_irq_handler(struct drm_i915_private *dev_priv, const u32 iir) > > +gen11_gu_misc_irq_handler(struct drm_i915_private *i915, const u32 iir) > > { > > if (iir & GEN11_GU_MISC_GSE) > > - intel_opregion_asle_intr(dev_priv); > > + intel_opregion_asle_intr(i915); > > } > > static inline u32 gen11_master_intr_disable(void __iomem * const regs) > > @@ -3235,54 +3235,54 @@ static irqreturn_t gen11_irq_handler(int irq, void *arg) > > */ > > static int i8xx_enable_vblank(struct drm_device *dev, unsigned int pipe) > > { > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > unsigned long irqflags; > > - spin_lock_irqsave(&dev_priv->irq_lock, irqflags); > > - i915_enable_pipestat(dev_priv, pipe, PIPE_VBLANK_INTERRUPT_STATUS); > > - spin_unlock_irqrestore(&dev_priv->irq_lock, irqflags); > > + spin_lock_irqsave(&i915->irq_lock, irqflags); > > + i915_enable_pipestat(i915, pipe, PIPE_VBLANK_INTERRUPT_STATUS); > > + spin_unlock_irqrestore(&i915->irq_lock, irqflags); > > return 0; > > } > > static int i945gm_enable_vblank(struct drm_device *dev, unsigned int pipe) > > { > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > - if (dev_priv->i945gm_vblank.enabled++ == 0) > > - schedule_work(&dev_priv->i945gm_vblank.work); > > + if (i915->i945gm_vblank.enabled++ == 0) > > + schedule_work(&i915->i945gm_vblank.work); > > return i8xx_enable_vblank(dev, pipe); > > } > > static int i965_enable_vblank(struct drm_device *dev, unsigned int pipe) > > { > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > unsigned long irqflags; > > - spin_lock_irqsave(&dev_priv->irq_lock, irqflags); > > - i915_enable_pipestat(dev_priv, pipe, > > + spin_lock_irqsave(&i915->irq_lock, irqflags); > > + i915_enable_pipestat(i915, pipe, > > PIPE_START_VBLANK_INTERRUPT_STATUS); > > - spin_unlock_irqrestore(&dev_priv->irq_lock, irqflags); > > + spin_unlock_irqrestore(&i915->irq_lock, irqflags); > > return 0; > > } > > static int ironlake_enable_vblank(struct drm_device *dev, unsigned int pipe) > > { > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > unsigned long irqflags; > > - u32 bit = INTEL_GEN(dev_priv) >= 7 ? > > + u32 bit = INTEL_GEN(i915) >= 7 ? > > DE_PIPE_VBLANK_IVB(pipe) : DE_PIPE_VBLANK(pipe); > > - spin_lock_irqsave(&dev_priv->irq_lock, irqflags); > > - ilk_enable_display_irq(dev_priv, bit); > > - spin_unlock_irqrestore(&dev_priv->irq_lock, irqflags); > > + spin_lock_irqsave(&i915->irq_lock, irqflags); > > + ilk_enable_display_irq(i915, bit); > > + spin_unlock_irqrestore(&i915->irq_lock, irqflags); > > /* Even though there is no DMC, frame counter can get stuck when > > * PSR is active as no frames are generated. > > */ > > - if (HAS_PSR(dev_priv)) > > + if (HAS_PSR(i915)) > > drm_vblank_restore(dev, pipe); > > return 0; > > @@ -3290,17 +3290,17 @@ static int ironlake_enable_vblank(struct drm_device *dev, unsigned int pipe) > > static int gen8_enable_vblank(struct drm_device *dev, unsigned int pipe) > > { > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > unsigned long irqflags; > > - spin_lock_irqsave(&dev_priv->irq_lock, irqflags); > > - bdw_enable_pipe_irq(dev_priv, pipe, GEN8_PIPE_VBLANK); > > - spin_unlock_irqrestore(&dev_priv->irq_lock, irqflags); > > + spin_lock_irqsave(&i915->irq_lock, irqflags); > > + bdw_enable_pipe_irq(i915, pipe, GEN8_PIPE_VBLANK); > > + spin_unlock_irqrestore(&i915->irq_lock, irqflags); > > /* Even if there is no DMC, frame counter can get stuck when > > * PSR is active as no frames are generated, so check only for PSR. > > */ > > - if (HAS_PSR(dev_priv)) > > + if (HAS_PSR(i915)) > > drm_vblank_restore(dev, pipe); > > return 0; > > @@ -3311,60 +3311,60 @@ static int gen8_enable_vblank(struct drm_device *dev, unsigned int pipe) > > */ > > static void i8xx_disable_vblank(struct drm_device *dev, unsigned int pipe) > > { > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > unsigned long irqflags; > > - spin_lock_irqsave(&dev_priv->irq_lock, irqflags); > > - i915_disable_pipestat(dev_priv, pipe, PIPE_VBLANK_INTERRUPT_STATUS); > > - spin_unlock_irqrestore(&dev_priv->irq_lock, irqflags); > > + spin_lock_irqsave(&i915->irq_lock, irqflags); > > + i915_disable_pipestat(i915, pipe, PIPE_VBLANK_INTERRUPT_STATUS); > > + spin_unlock_irqrestore(&i915->irq_lock, irqflags); > > } > > static void i945gm_disable_vblank(struct drm_device *dev, unsigned int pipe) > > { > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > i8xx_disable_vblank(dev, pipe); > > - if (--dev_priv->i945gm_vblank.enabled == 0) > > - schedule_work(&dev_priv->i945gm_vblank.work); > > + if (--i915->i945gm_vblank.enabled == 0) > > + schedule_work(&i915->i945gm_vblank.work); > > } > > static void i965_disable_vblank(struct drm_device *dev, unsigned int pipe) > > { > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > unsigned long irqflags; > > - spin_lock_irqsave(&dev_priv->irq_lock, irqflags); > > - i915_disable_pipestat(dev_priv, pipe, > > + spin_lock_irqsave(&i915->irq_lock, irqflags); > > + i915_disable_pipestat(i915, pipe, > > PIPE_START_VBLANK_INTERRUPT_STATUS); > > - spin_unlock_irqrestore(&dev_priv->irq_lock, irqflags); > > + spin_unlock_irqrestore(&i915->irq_lock, irqflags); > > } > > static void ironlake_disable_vblank(struct drm_device *dev, unsigned int pipe) > > { > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > unsigned long irqflags; > > - u32 bit = INTEL_GEN(dev_priv) >= 7 ? > > + u32 bit = INTEL_GEN(i915) >= 7 ? > > DE_PIPE_VBLANK_IVB(pipe) : DE_PIPE_VBLANK(pipe); > > - spin_lock_irqsave(&dev_priv->irq_lock, irqflags); > > - ilk_disable_display_irq(dev_priv, bit); > > - spin_unlock_irqrestore(&dev_priv->irq_lock, irqflags); > > + spin_lock_irqsave(&i915->irq_lock, irqflags); > > + ilk_disable_display_irq(i915, bit); > > + spin_unlock_irqrestore(&i915->irq_lock, irqflags); > > } > > static void gen8_disable_vblank(struct drm_device *dev, unsigned int pipe) > > { > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > unsigned long irqflags; > > - spin_lock_irqsave(&dev_priv->irq_lock, irqflags); > > - bdw_disable_pipe_irq(dev_priv, pipe, GEN8_PIPE_VBLANK); > > - spin_unlock_irqrestore(&dev_priv->irq_lock, irqflags); > > + spin_lock_irqsave(&i915->irq_lock, irqflags); > > + bdw_disable_pipe_irq(i915, pipe, GEN8_PIPE_VBLANK); > > + spin_unlock_irqrestore(&i915->irq_lock, irqflags); > > } > > static void i945gm_vblank_work_func(struct work_struct *work) > > { > > - struct drm_i915_private *dev_priv = > > + struct drm_i915_private *i915 = > > container_of(work, struct drm_i915_private, i945gm_vblank.work); > > /* > > @@ -3372,9 +3372,9 @@ static void i945gm_vblank_work_func(struct work_struct *work) > > * hence we want to prevent C3 usage while vblank interrupts > > * are enabled. > > */ > > - pm_qos_update_request(&dev_priv->i945gm_vblank.pm_qos, > > - READ_ONCE(dev_priv->i945gm_vblank.enabled) ? > > - dev_priv->i945gm_vblank.c3_disable_latency : > > + pm_qos_update_request(&i915->i945gm_vblank.pm_qos, > > + READ_ONCE(i915->i945gm_vblank.enabled) ? > > + i915->i945gm_vblank.c3_disable_latency : > > PM_QOS_DEFAULT_VALUE); > > } > > @@ -3398,34 +3398,34 @@ static int cstate_disable_latency(const char *name) > > return 0; > > } > > -static void i945gm_vblank_work_init(struct drm_i915_private *dev_priv) > > +static void i945gm_vblank_work_init(struct drm_i915_private *i915) > > { > > - INIT_WORK(&dev_priv->i945gm_vblank.work, > > + INIT_WORK(&i915->i945gm_vblank.work, > > i945gm_vblank_work_func); > > - dev_priv->i945gm_vblank.c3_disable_latency = > > + i915->i945gm_vblank.c3_disable_latency = > > cstate_disable_latency("C3"); > > - pm_qos_add_request(&dev_priv->i945gm_vblank.pm_qos, > > + pm_qos_add_request(&i915->i945gm_vblank.pm_qos, > > PM_QOS_CPU_DMA_LATENCY, > > PM_QOS_DEFAULT_VALUE); > > } > > -static void i945gm_vblank_work_fini(struct drm_i915_private *dev_priv) > > +static void i945gm_vblank_work_fini(struct drm_i915_private *i915) > > { > > - cancel_work_sync(&dev_priv->i945gm_vblank.work); > > - pm_qos_remove_request(&dev_priv->i945gm_vblank.pm_qos); > > + cancel_work_sync(&i915->i945gm_vblank.work); > > + pm_qos_remove_request(&i915->i945gm_vblank.pm_qos); > > } > > -static void ibx_irq_reset(struct drm_i915_private *dev_priv) > > +static void ibx_irq_reset(struct drm_i915_private *i915) > > { > > - struct intel_uncore *uncore = &dev_priv->uncore; > > + struct intel_uncore *uncore = &i915->uncore; > > - if (HAS_PCH_NOP(dev_priv)) > > + if (HAS_PCH_NOP(i915)) > > return; > > GEN3_IRQ_RESET(uncore, SDE); > > - if (HAS_PCH_CPT(dev_priv) || HAS_PCH_LPT(dev_priv)) > > + if (HAS_PCH_CPT(i915) || HAS_PCH_LPT(i915)) > > I915_WRITE(SERR_INT, 0xffffffff); > > } > > @@ -3439,9 +3439,9 @@ static void ibx_irq_reset(struct drm_i915_private *dev_priv) > > */ > > static void ibx_irq_pre_postinstall(struct drm_device *dev) > > { > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > - if (HAS_PCH_NOP(dev_priv)) > > + if (HAS_PCH_NOP(i915)) > > return; > > WARN_ON(I915_READ(SDEIER) != 0); > > @@ -3449,36 +3449,36 @@ static void ibx_irq_pre_postinstall(struct drm_device *dev) > > POSTING_READ(SDEIER); > > } > > -static void gen5_gt_irq_reset(struct drm_i915_private *dev_priv) > > +static void gen5_gt_irq_reset(struct drm_i915_private *i915) > > { > > - struct intel_uncore *uncore = &dev_priv->uncore; > > + struct intel_uncore *uncore = &i915->uncore; > > GEN3_IRQ_RESET(uncore, GT); > > - if (INTEL_GEN(dev_priv) >= 6) > > + if (INTEL_GEN(i915) >= 6) > > GEN3_IRQ_RESET(uncore, GEN6_PM); > > } > > -static void vlv_display_irq_reset(struct drm_i915_private *dev_priv) > > +static void vlv_display_irq_reset(struct drm_i915_private *i915) > > { > > - struct intel_uncore *uncore = &dev_priv->uncore; > > + struct intel_uncore *uncore = &i915->uncore; > > - if (IS_CHERRYVIEW(dev_priv)) > > + if (IS_CHERRYVIEW(i915)) > > I915_WRITE(DPINVGTT, DPINVGTT_STATUS_MASK_CHV); > > else > > I915_WRITE(DPINVGTT, DPINVGTT_STATUS_MASK); > > - i915_hotplug_interrupt_update_locked(dev_priv, 0xffffffff, 0); > > + i915_hotplug_interrupt_update_locked(i915, 0xffffffff, 0); > > I915_WRITE(PORT_HOTPLUG_STAT, I915_READ(PORT_HOTPLUG_STAT)); > > - i9xx_pipestat_irq_reset(dev_priv); > > + i9xx_pipestat_irq_reset(i915); > > GEN3_IRQ_RESET(uncore, VLV_); > > - dev_priv->irq_mask = ~0u; > > + i915->irq_mask = ~0u; > > } > > -static void vlv_display_irq_postinstall(struct drm_i915_private *dev_priv) > > +static void vlv_display_irq_postinstall(struct drm_i915_private *i915) > > { > > - struct intel_uncore *uncore = &dev_priv->uncore; > > + struct intel_uncore *uncore = &i915->uncore; > > u32 pipestat_mask; > > u32 enable_mask; > > @@ -3486,9 +3486,9 @@ static void vlv_display_irq_postinstall(struct drm_i915_private *dev_priv) > > pipestat_mask = PIPE_CRC_DONE_INTERRUPT_STATUS; > > - i915_enable_pipestat(dev_priv, PIPE_A, PIPE_GMBUS_INTERRUPT_STATUS); > > - for_each_pipe(dev_priv, pipe) > > - i915_enable_pipestat(dev_priv, pipe, pipestat_mask); > > + i915_enable_pipestat(i915, PIPE_A, PIPE_GMBUS_INTERRUPT_STATUS); > > + for_each_pipe(i915, pipe) > > + i915_enable_pipestat(i915, pipe, pipestat_mask); > > enable_mask = I915_DISPLAY_PORT_INTERRUPT | > > I915_DISPLAY_PIPE_A_EVENT_INTERRUPT | > > @@ -3496,56 +3496,56 @@ static void vlv_display_irq_postinstall(struct drm_i915_private *dev_priv) > > I915_LPE_PIPE_A_INTERRUPT | > > I915_LPE_PIPE_B_INTERRUPT; > > - if (IS_CHERRYVIEW(dev_priv)) > > + if (IS_CHERRYVIEW(i915)) > > enable_mask |= I915_DISPLAY_PIPE_C_EVENT_INTERRUPT | > > I915_LPE_PIPE_C_INTERRUPT; > > - WARN_ON(dev_priv->irq_mask != ~0u); > > + WARN_ON(i915->irq_mask != ~0u); > > - dev_priv->irq_mask = ~enable_mask; > > + i915->irq_mask = ~enable_mask; > > - GEN3_IRQ_INIT(uncore, VLV_, dev_priv->irq_mask, enable_mask); > > + GEN3_IRQ_INIT(uncore, VLV_, i915->irq_mask, enable_mask); > > } > > /* drm_dma.h hooks > > */ > > static void ironlake_irq_reset(struct drm_device *dev) > > { > > - struct drm_i915_private *dev_priv = to_i915(dev); > > - struct intel_uncore *uncore = &dev_priv->uncore; > > + struct drm_i915_private *i915 = to_i915(dev); > > + struct intel_uncore *uncore = &i915->uncore; > > GEN3_IRQ_RESET(uncore, DE); > > - if (IS_GEN(dev_priv, 7)) > > + if (IS_GEN(i915, 7)) > > I915_WRITE(GEN7_ERR_INT, 0xffffffff); > > - if (IS_HASWELL(dev_priv)) { > > + if (IS_HASWELL(i915)) { > > I915_WRITE(EDP_PSR_IMR, 0xffffffff); > > I915_WRITE(EDP_PSR_IIR, 0xffffffff); > > } > > - gen5_gt_irq_reset(dev_priv); > > + gen5_gt_irq_reset(i915); > > - ibx_irq_reset(dev_priv); > > + ibx_irq_reset(i915); > > } > > static void valleyview_irq_reset(struct drm_device *dev) > > { > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > I915_WRITE(VLV_MASTER_IER, 0); > > POSTING_READ(VLV_MASTER_IER); > > - gen5_gt_irq_reset(dev_priv); > > + gen5_gt_irq_reset(i915); > > - spin_lock_irq(&dev_priv->irq_lock); > > - if (dev_priv->display_irqs_enabled) > > - vlv_display_irq_reset(dev_priv); > > - spin_unlock_irq(&dev_priv->irq_lock); > > + spin_lock_irq(&i915->irq_lock); > > + if (i915->display_irqs_enabled) > > + vlv_display_irq_reset(i915); > > + spin_unlock_irq(&i915->irq_lock); > > } > > -static void gen8_gt_irq_reset(struct drm_i915_private *dev_priv) > > +static void gen8_gt_irq_reset(struct drm_i915_private *i915) > > { > > - struct intel_uncore *uncore = &dev_priv->uncore; > > + struct intel_uncore *uncore = &i915->uncore; > > GEN8_IRQ_RESET_NDX(uncore, GT, 0); > > GEN8_IRQ_RESET_NDX(uncore, GT, 1); > > @@ -3555,19 +3555,19 @@ static void gen8_gt_irq_reset(struct drm_i915_private *dev_priv) > > static void gen8_irq_reset(struct drm_device *dev) > > { > > - struct drm_i915_private *dev_priv = to_i915(dev); > > - struct intel_uncore *uncore = &dev_priv->uncore; > > + struct drm_i915_private *i915 = to_i915(dev); > > + struct intel_uncore *uncore = &i915->uncore; > > int pipe; > > - gen8_master_intr_disable(dev_priv->uncore.regs); > > + gen8_master_intr_disable(i915->uncore.regs); > > - gen8_gt_irq_reset(dev_priv); > > + gen8_gt_irq_reset(i915); > > I915_WRITE(EDP_PSR_IMR, 0xffffffff); > > I915_WRITE(EDP_PSR_IIR, 0xffffffff); > > - for_each_pipe(dev_priv, pipe) > > - if (intel_display_power_is_enabled(dev_priv, > > + for_each_pipe(i915, pipe) > > + if (intel_display_power_is_enabled(i915, > > POWER_DOMAIN_PIPE(pipe))) > > GEN8_IRQ_RESET_NDX(uncore, DE_PIPE, pipe); > > @@ -3575,11 +3575,11 @@ static void gen8_irq_reset(struct drm_device *dev) > > GEN3_IRQ_RESET(uncore, GEN8_DE_MISC_); > > GEN3_IRQ_RESET(uncore, GEN8_PCU_); > > - if (HAS_PCH_SPLIT(dev_priv)) > > - ibx_irq_reset(dev_priv); > > + if (HAS_PCH_SPLIT(i915)) > > + ibx_irq_reset(i915); > > } > > -static void gen11_gt_irq_reset(struct drm_i915_private *dev_priv) > > +static void gen11_gt_irq_reset(struct drm_i915_private *i915) > > { > > /* Disable RCS, BCS, VCS and VECS class engines. */ > > I915_WRITE(GEN11_RENDER_COPY_INTR_ENABLE, 0); > > @@ -3600,21 +3600,21 @@ static void gen11_gt_irq_reset(struct drm_i915_private *dev_priv) > > static void gen11_irq_reset(struct drm_device *dev) > > { > > - struct drm_i915_private *dev_priv = dev->dev_private; > > - struct intel_uncore *uncore = &dev_priv->uncore; > > + struct drm_i915_private *i915 = dev->dev_private; > > + struct intel_uncore *uncore = &i915->uncore; > > int pipe; > > - gen11_master_intr_disable(dev_priv->uncore.regs); > > + gen11_master_intr_disable(i915->uncore.regs); > > - gen11_gt_irq_reset(dev_priv); > > + gen11_gt_irq_reset(i915); > > I915_WRITE(GEN11_DISPLAY_INT_CTL, 0); > > I915_WRITE(EDP_PSR_IMR, 0xffffffff); > > I915_WRITE(EDP_PSR_IIR, 0xffffffff); > > - for_each_pipe(dev_priv, pipe) > > - if (intel_display_power_is_enabled(dev_priv, > > + for_each_pipe(i915, pipe) > > + if (intel_display_power_is_enabled(i915, > > POWER_DOMAIN_PIPE(pipe))) > > GEN8_IRQ_RESET_NDX(uncore, DE_PIPE, pipe); > > @@ -3624,87 +3624,87 @@ static void gen11_irq_reset(struct drm_device *dev) > > GEN3_IRQ_RESET(uncore, GEN11_GU_MISC_); > > GEN3_IRQ_RESET(uncore, GEN8_PCU_); > > - if (INTEL_PCH_TYPE(dev_priv) >= PCH_ICP) > > + if (INTEL_PCH_TYPE(i915) >= PCH_ICP) > > GEN3_IRQ_RESET(uncore, SDE); > > } > > -void gen8_irq_power_well_post_enable(struct drm_i915_private *dev_priv, > > +void gen8_irq_power_well_post_enable(struct drm_i915_private *i915, > > u8 pipe_mask) > > { > > - struct intel_uncore *uncore = &dev_priv->uncore; > > + struct intel_uncore *uncore = &i915->uncore; > > u32 extra_ier = GEN8_PIPE_VBLANK | GEN8_PIPE_FIFO_UNDERRUN; > > enum pipe pipe; > > - spin_lock_irq(&dev_priv->irq_lock); > > + spin_lock_irq(&i915->irq_lock); > > - if (!intel_irqs_enabled(dev_priv)) { > > - spin_unlock_irq(&dev_priv->irq_lock); > > + if (!intel_irqs_enabled(i915)) { > > + spin_unlock_irq(&i915->irq_lock); > > return; > > } > > - for_each_pipe_masked(dev_priv, pipe, pipe_mask) > > + for_each_pipe_masked(i915, pipe, pipe_mask) > > GEN8_IRQ_INIT_NDX(uncore, DE_PIPE, pipe, > > - dev_priv->de_irq_mask[pipe], > > - ~dev_priv->de_irq_mask[pipe] | extra_ier); > > + i915->de_irq_mask[pipe], > > + ~i915->de_irq_mask[pipe] | extra_ier); > > - spin_unlock_irq(&dev_priv->irq_lock); > > + spin_unlock_irq(&i915->irq_lock); > > } > > -void gen8_irq_power_well_pre_disable(struct drm_i915_private *dev_priv, > > +void gen8_irq_power_well_pre_disable(struct drm_i915_private *i915, > > u8 pipe_mask) > > { > > - struct intel_uncore *uncore = &dev_priv->uncore; > > + struct intel_uncore *uncore = &i915->uncore; > > enum pipe pipe; > > - spin_lock_irq(&dev_priv->irq_lock); > > + spin_lock_irq(&i915->irq_lock); > > - if (!intel_irqs_enabled(dev_priv)) { > > - spin_unlock_irq(&dev_priv->irq_lock); > > + if (!intel_irqs_enabled(i915)) { > > + spin_unlock_irq(&i915->irq_lock); > > return; > > } > > - for_each_pipe_masked(dev_priv, pipe, pipe_mask) > > + for_each_pipe_masked(i915, pipe, pipe_mask) > > GEN8_IRQ_RESET_NDX(uncore, DE_PIPE, pipe); > > - spin_unlock_irq(&dev_priv->irq_lock); > > + spin_unlock_irq(&i915->irq_lock); > > /* make sure we're done processing display irqs */ > > - synchronize_irq(dev_priv->drm.irq); > > + synchronize_irq(i915->drm.irq); > > } > > static void cherryview_irq_reset(struct drm_device *dev) > > { > > - struct drm_i915_private *dev_priv = to_i915(dev); > > - struct intel_uncore *uncore = &dev_priv->uncore; > > + struct drm_i915_private *i915 = to_i915(dev); > > + struct intel_uncore *uncore = &i915->uncore; > > I915_WRITE(GEN8_MASTER_IRQ, 0); > > POSTING_READ(GEN8_MASTER_IRQ); > > - gen8_gt_irq_reset(dev_priv); > > + gen8_gt_irq_reset(i915); > > GEN3_IRQ_RESET(uncore, GEN8_PCU_); > > - spin_lock_irq(&dev_priv->irq_lock); > > - if (dev_priv->display_irqs_enabled) > > - vlv_display_irq_reset(dev_priv); > > - spin_unlock_irq(&dev_priv->irq_lock); > > + spin_lock_irq(&i915->irq_lock); > > + if (i915->display_irqs_enabled) > > + vlv_display_irq_reset(i915); > > + spin_unlock_irq(&i915->irq_lock); > > } > > -static u32 intel_hpd_enabled_irqs(struct drm_i915_private *dev_priv, > > +static u32 intel_hpd_enabled_irqs(struct drm_i915_private *i915, > > const u32 hpd[HPD_NUM_PINS]) > > { > > struct intel_encoder *encoder; > > u32 enabled_irqs = 0; > > - for_each_intel_encoder(&dev_priv->drm, encoder) > > - if (dev_priv->hotplug.stats[encoder->hpd_pin].state == HPD_ENABLED) > > + for_each_intel_encoder(&i915->drm, encoder) > > + if (i915->hotplug.stats[encoder->hpd_pin].state == HPD_ENABLED) > > enabled_irqs |= hpd[encoder->hpd_pin]; > > return enabled_irqs; > > } > > -static void ibx_hpd_detection_setup(struct drm_i915_private *dev_priv) > > +static void ibx_hpd_detection_setup(struct drm_i915_private *i915) > > { > > u32 hotplug; > > @@ -3724,29 +3724,29 @@ static void ibx_hpd_detection_setup(struct drm_i915_private *dev_priv) > > * When CPU and PCH are on the same package, port A > > * HPD must be enabled in both north and south. > > */ > > - if (HAS_PCH_LPT_LP(dev_priv)) > > + if (HAS_PCH_LPT_LP(i915)) > > hotplug |= PORTA_HOTPLUG_ENABLE; > > I915_WRITE(PCH_PORT_HOTPLUG, hotplug); > > } > > -static void ibx_hpd_irq_setup(struct drm_i915_private *dev_priv) > > +static void ibx_hpd_irq_setup(struct drm_i915_private *i915) > > { > > u32 hotplug_irqs, enabled_irqs; > > - if (HAS_PCH_IBX(dev_priv)) { > > + if (HAS_PCH_IBX(i915)) { > > hotplug_irqs = SDE_HOTPLUG_MASK; > > - enabled_irqs = intel_hpd_enabled_irqs(dev_priv, hpd_ibx); > > + enabled_irqs = intel_hpd_enabled_irqs(i915, hpd_ibx); > > } else { > > hotplug_irqs = SDE_HOTPLUG_MASK_CPT; > > - enabled_irqs = intel_hpd_enabled_irqs(dev_priv, hpd_cpt); > > + enabled_irqs = intel_hpd_enabled_irqs(i915, hpd_cpt); > > } > > - ibx_display_interrupt_update(dev_priv, hotplug_irqs, enabled_irqs); > > + ibx_display_interrupt_update(i915, hotplug_irqs, enabled_irqs); > > - ibx_hpd_detection_setup(dev_priv); > > + ibx_hpd_detection_setup(i915); > > } > > -static void icp_hpd_detection_setup(struct drm_i915_private *dev_priv) > > +static void icp_hpd_detection_setup(struct drm_i915_private *i915) > > { > > u32 hotplug; > > @@ -3763,19 +3763,19 @@ static void icp_hpd_detection_setup(struct drm_i915_private *dev_priv) > > I915_WRITE(SHOTPLUG_CTL_TC, hotplug); > > } > > -static void icp_hpd_irq_setup(struct drm_i915_private *dev_priv) > > +static void icp_hpd_irq_setup(struct drm_i915_private *i915) > > { > > u32 hotplug_irqs, enabled_irqs; > > hotplug_irqs = SDE_DDI_MASK_ICP | SDE_TC_MASK_ICP; > > - enabled_irqs = intel_hpd_enabled_irqs(dev_priv, hpd_icp); > > + enabled_irqs = intel_hpd_enabled_irqs(i915, hpd_icp); > > - ibx_display_interrupt_update(dev_priv, hotplug_irqs, enabled_irqs); > > + ibx_display_interrupt_update(i915, hotplug_irqs, enabled_irqs); > > - icp_hpd_detection_setup(dev_priv); > > + icp_hpd_detection_setup(i915); > > } > > -static void gen11_hpd_detection_setup(struct drm_i915_private *dev_priv) > > +static void gen11_hpd_detection_setup(struct drm_i915_private *i915) > > { > > u32 hotplug; > > @@ -3794,12 +3794,12 @@ static void gen11_hpd_detection_setup(struct drm_i915_private *dev_priv) > > I915_WRITE(GEN11_TBT_HOTPLUG_CTL, hotplug); > > } > > -static void gen11_hpd_irq_setup(struct drm_i915_private *dev_priv) > > +static void gen11_hpd_irq_setup(struct drm_i915_private *i915) > > { > > u32 hotplug_irqs, enabled_irqs; > > u32 val; > > - enabled_irqs = intel_hpd_enabled_irqs(dev_priv, hpd_gen11); > > + enabled_irqs = intel_hpd_enabled_irqs(i915, hpd_gen11); > > hotplug_irqs = GEN11_DE_TC_HOTPLUG_MASK | GEN11_DE_TBT_HOTPLUG_MASK; > > val = I915_READ(GEN11_DE_HPD_IMR); > > @@ -3807,18 +3807,18 @@ static void gen11_hpd_irq_setup(struct drm_i915_private *dev_priv) > > I915_WRITE(GEN11_DE_HPD_IMR, val); > > POSTING_READ(GEN11_DE_HPD_IMR); > > - gen11_hpd_detection_setup(dev_priv); > > + gen11_hpd_detection_setup(i915); > > - if (INTEL_PCH_TYPE(dev_priv) >= PCH_ICP) > > - icp_hpd_irq_setup(dev_priv); > > + if (INTEL_PCH_TYPE(i915) >= PCH_ICP) > > + icp_hpd_irq_setup(i915); > > } > > -static void spt_hpd_detection_setup(struct drm_i915_private *dev_priv) > > +static void spt_hpd_detection_setup(struct drm_i915_private *i915) > > { > > u32 val, hotplug; > > /* Display WA #1179 WaHardHangonHotPlug: cnp */ > > - if (HAS_PCH_CNP(dev_priv)) { > > + if (HAS_PCH_CNP(i915)) { > > val = I915_READ(SOUTH_CHICKEN1); > > val &= ~CHASSIS_CLK_REQ_DURATION_MASK; > > val |= CHASSIS_CLK_REQ_DURATION(0xf); > > @@ -3838,19 +3838,19 @@ static void spt_hpd_detection_setup(struct drm_i915_private *dev_priv) > > I915_WRITE(PCH_PORT_HOTPLUG2, hotplug); > > } > > -static void spt_hpd_irq_setup(struct drm_i915_private *dev_priv) > > +static void spt_hpd_irq_setup(struct drm_i915_private *i915) > > { > > u32 hotplug_irqs, enabled_irqs; > > hotplug_irqs = SDE_HOTPLUG_MASK_SPT; > > - enabled_irqs = intel_hpd_enabled_irqs(dev_priv, hpd_spt); > > + enabled_irqs = intel_hpd_enabled_irqs(i915, hpd_spt); > > - ibx_display_interrupt_update(dev_priv, hotplug_irqs, enabled_irqs); > > + ibx_display_interrupt_update(i915, hotplug_irqs, enabled_irqs); > > - spt_hpd_detection_setup(dev_priv); > > + spt_hpd_detection_setup(i915); > > } > > -static void ilk_hpd_detection_setup(struct drm_i915_private *dev_priv) > > +static void ilk_hpd_detection_setup(struct drm_i915_private *i915) > > { > > u32 hotplug; > > @@ -3866,33 +3866,33 @@ static void ilk_hpd_detection_setup(struct drm_i915_private *dev_priv) > > I915_WRITE(DIGITAL_PORT_HOTPLUG_CNTRL, hotplug); > > } > > -static void ilk_hpd_irq_setup(struct drm_i915_private *dev_priv) > > +static void ilk_hpd_irq_setup(struct drm_i915_private *i915) > > { > > u32 hotplug_irqs, enabled_irqs; > > - if (INTEL_GEN(dev_priv) >= 8) { > > + if (INTEL_GEN(i915) >= 8) { > > hotplug_irqs = GEN8_PORT_DP_A_HOTPLUG; > > - enabled_irqs = intel_hpd_enabled_irqs(dev_priv, hpd_bdw); > > + enabled_irqs = intel_hpd_enabled_irqs(i915, hpd_bdw); > > - bdw_update_port_irq(dev_priv, hotplug_irqs, enabled_irqs); > > - } else if (INTEL_GEN(dev_priv) >= 7) { > > + bdw_update_port_irq(i915, hotplug_irqs, enabled_irqs); > > + } else if (INTEL_GEN(i915) >= 7) { > > hotplug_irqs = DE_DP_A_HOTPLUG_IVB; > > - enabled_irqs = intel_hpd_enabled_irqs(dev_priv, hpd_ivb); > > + enabled_irqs = intel_hpd_enabled_irqs(i915, hpd_ivb); > > - ilk_update_display_irq(dev_priv, hotplug_irqs, enabled_irqs); > > + ilk_update_display_irq(i915, hotplug_irqs, enabled_irqs); > > } else { > > hotplug_irqs = DE_DP_A_HOTPLUG; > > - enabled_irqs = intel_hpd_enabled_irqs(dev_priv, hpd_ilk); > > + enabled_irqs = intel_hpd_enabled_irqs(i915, hpd_ilk); > > - ilk_update_display_irq(dev_priv, hotplug_irqs, enabled_irqs); > > + ilk_update_display_irq(i915, hotplug_irqs, enabled_irqs); > > } > > - ilk_hpd_detection_setup(dev_priv); > > + ilk_hpd_detection_setup(i915); > > - ibx_hpd_irq_setup(dev_priv); > > + ibx_hpd_irq_setup(i915); > > } > > -static void __bxt_hpd_detection_setup(struct drm_i915_private *dev_priv, > > +static void __bxt_hpd_detection_setup(struct drm_i915_private *i915, > > u32 enabled_irqs) > > { > > u32 hotplug; > > @@ -3911,106 +3911,106 @@ static void __bxt_hpd_detection_setup(struct drm_i915_private *dev_priv, > > * for HPD detection logic, update it based on VBT fields. > > */ > > if ((enabled_irqs & BXT_DE_PORT_HP_DDIA) && > > - intel_bios_is_port_hpd_inverted(dev_priv, PORT_A)) > > + intel_bios_is_port_hpd_inverted(i915, PORT_A)) > > hotplug |= BXT_DDIA_HPD_INVERT; > > if ((enabled_irqs & BXT_DE_PORT_HP_DDIB) && > > - intel_bios_is_port_hpd_inverted(dev_priv, PORT_B)) > > + intel_bios_is_port_hpd_inverted(i915, PORT_B)) > > hotplug |= BXT_DDIB_HPD_INVERT; > > if ((enabled_irqs & BXT_DE_PORT_HP_DDIC) && > > - intel_bios_is_port_hpd_inverted(dev_priv, PORT_C)) > > + intel_bios_is_port_hpd_inverted(i915, PORT_C)) > > hotplug |= BXT_DDIC_HPD_INVERT; > > I915_WRITE(PCH_PORT_HOTPLUG, hotplug); > > } > > -static void bxt_hpd_detection_setup(struct drm_i915_private *dev_priv) > > +static void bxt_hpd_detection_setup(struct drm_i915_private *i915) > > { > > - __bxt_hpd_detection_setup(dev_priv, BXT_DE_PORT_HOTPLUG_MASK); > > + __bxt_hpd_detection_setup(i915, BXT_DE_PORT_HOTPLUG_MASK); > > } > > -static void bxt_hpd_irq_setup(struct drm_i915_private *dev_priv) > > +static void bxt_hpd_irq_setup(struct drm_i915_private *i915) > > { > > u32 hotplug_irqs, enabled_irqs; > > - enabled_irqs = intel_hpd_enabled_irqs(dev_priv, hpd_bxt); > > + enabled_irqs = intel_hpd_enabled_irqs(i915, hpd_bxt); > > hotplug_irqs = BXT_DE_PORT_HOTPLUG_MASK; > > - bdw_update_port_irq(dev_priv, hotplug_irqs, enabled_irqs); > > + bdw_update_port_irq(i915, hotplug_irqs, enabled_irqs); > > - __bxt_hpd_detection_setup(dev_priv, enabled_irqs); > > + __bxt_hpd_detection_setup(i915, enabled_irqs); > > } > > static void ibx_irq_postinstall(struct drm_device *dev) > > { > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > u32 mask; > > - if (HAS_PCH_NOP(dev_priv)) > > + if (HAS_PCH_NOP(i915)) > > return; > > - if (HAS_PCH_IBX(dev_priv)) > > + if (HAS_PCH_IBX(i915)) > > mask = SDE_GMBUS | SDE_AUX_MASK | SDE_POISON; > > - else if (HAS_PCH_CPT(dev_priv) || HAS_PCH_LPT(dev_priv)) > > + else if (HAS_PCH_CPT(i915) || HAS_PCH_LPT(i915)) > > mask = SDE_GMBUS_CPT | SDE_AUX_MASK_CPT; > > else > > mask = SDE_GMBUS_CPT; > > - gen3_assert_iir_is_zero(&dev_priv->uncore, SDEIIR); > > + gen3_assert_iir_is_zero(&i915->uncore, SDEIIR); > > I915_WRITE(SDEIMR, ~mask); > > - if (HAS_PCH_IBX(dev_priv) || HAS_PCH_CPT(dev_priv) || > > - HAS_PCH_LPT(dev_priv)) > > - ibx_hpd_detection_setup(dev_priv); > > + if (HAS_PCH_IBX(i915) || HAS_PCH_CPT(i915) || > > + HAS_PCH_LPT(i915)) > > + ibx_hpd_detection_setup(i915); > > else > > - spt_hpd_detection_setup(dev_priv); > > + spt_hpd_detection_setup(i915); > > } > > static void gen5_gt_irq_postinstall(struct drm_device *dev) > > { > > - struct drm_i915_private *dev_priv = to_i915(dev); > > - struct intel_uncore *uncore = &dev_priv->uncore; > > + struct drm_i915_private *i915 = to_i915(dev); > > + struct intel_uncore *uncore = &i915->uncore; > > u32 pm_irqs, gt_irqs; > > pm_irqs = gt_irqs = 0; > > - dev_priv->gt_irq_mask = ~0; > > - if (HAS_L3_DPF(dev_priv)) { > > + i915->gt_irq_mask = ~0; > > + if (HAS_L3_DPF(i915)) { > > /* L3 parity interrupt is always unmasked. */ > > - dev_priv->gt_irq_mask = ~GT_PARITY_ERROR(dev_priv); > > - gt_irqs |= GT_PARITY_ERROR(dev_priv); > > + i915->gt_irq_mask = ~GT_PARITY_ERROR(i915); > > + gt_irqs |= GT_PARITY_ERROR(i915); > > } > > gt_irqs |= GT_RENDER_USER_INTERRUPT; > > - if (IS_GEN(dev_priv, 5)) { > > + if (IS_GEN(i915, 5)) { > > gt_irqs |= ILK_BSD_USER_INTERRUPT; > > } else { > > gt_irqs |= GT_BLT_USER_INTERRUPT | GT_BSD_USER_INTERRUPT; > > } > > - GEN3_IRQ_INIT(uncore, GT, dev_priv->gt_irq_mask, gt_irqs); > > + GEN3_IRQ_INIT(uncore, GT, i915->gt_irq_mask, gt_irqs); > > - if (INTEL_GEN(dev_priv) >= 6) { > > + if (INTEL_GEN(i915) >= 6) { > > /* > > * RPS interrupts will get enabled/disabled on demand when RPS > > * itself is enabled/disabled. > > */ > > - if (HAS_ENGINE(dev_priv, VECS0)) { > > + if (HAS_ENGINE(i915, VECS0)) { > > pm_irqs |= PM_VEBOX_USER_INTERRUPT; > > - dev_priv->pm_ier |= PM_VEBOX_USER_INTERRUPT; > > + i915->pm_ier |= PM_VEBOX_USER_INTERRUPT; > > } > > - dev_priv->pm_imr = 0xffffffff; > > - GEN3_IRQ_INIT(uncore, GEN6_PM, dev_priv->pm_imr, pm_irqs); > > + i915->pm_imr = 0xffffffff; > > + GEN3_IRQ_INIT(uncore, GEN6_PM, i915->pm_imr, pm_irqs); > > } > > } > > static int ironlake_irq_postinstall(struct drm_device *dev) > > { > > - struct drm_i915_private *dev_priv = to_i915(dev); > > - struct intel_uncore *uncore = &dev_priv->uncore; > > + struct drm_i915_private *i915 = to_i915(dev); > > + struct intel_uncore *uncore = &i915->uncore; > > u32 display_mask, extra_mask; > > - if (INTEL_GEN(dev_priv) >= 7) { > > + if (INTEL_GEN(i915) >= 7) { > > display_mask = (DE_MASTER_IRQ_CONTROL | DE_GSE_IVB | > > DE_PCH_EVENT_IVB | DE_AUX_CHANNEL_A_IVB); > > extra_mask = (DE_PIPEC_VBLANK_IVB | DE_PIPEB_VBLANK_IVB | > > @@ -4025,78 +4025,78 @@ static int ironlake_irq_postinstall(struct drm_device *dev) > > DE_DP_A_HOTPLUG); > > } > > - if (IS_HASWELL(dev_priv)) { > > + if (IS_HASWELL(i915)) { > > gen3_assert_iir_is_zero(uncore, EDP_PSR_IIR); > > - intel_psr_irq_control(dev_priv, dev_priv->psr.debug); > > + intel_psr_irq_control(i915, i915->psr.debug); > > display_mask |= DE_EDP_PSR_INT_HSW; > > } > > - dev_priv->irq_mask = ~display_mask; > > + i915->irq_mask = ~display_mask; > > ibx_irq_pre_postinstall(dev); > > - GEN3_IRQ_INIT(uncore, DE, dev_priv->irq_mask, > > + GEN3_IRQ_INIT(uncore, DE, i915->irq_mask, > > display_mask | extra_mask); > > gen5_gt_irq_postinstall(dev); > > - ilk_hpd_detection_setup(dev_priv); > > + ilk_hpd_detection_setup(i915); > > ibx_irq_postinstall(dev); > > - if (IS_IRONLAKE_M(dev_priv)) { > > + if (IS_IRONLAKE_M(i915)) { > > /* Enable PCU event interrupts > > * > > * spinlocking not required here for correctness since interrupt > > * setup is guaranteed to run in single-threaded context. But we > > * need it to make the assert_spin_locked happy. */ > > - spin_lock_irq(&dev_priv->irq_lock); > > - ilk_enable_display_irq(dev_priv, DE_PCU_EVENT); > > - spin_unlock_irq(&dev_priv->irq_lock); > > + spin_lock_irq(&i915->irq_lock); > > + ilk_enable_display_irq(i915, DE_PCU_EVENT); > > + spin_unlock_irq(&i915->irq_lock); > > } > > return 0; > > } > > -void valleyview_enable_display_irqs(struct drm_i915_private *dev_priv) > > +void valleyview_enable_display_irqs(struct drm_i915_private *i915) > > { > > - lockdep_assert_held(&dev_priv->irq_lock); > > + lockdep_assert_held(&i915->irq_lock); > > - if (dev_priv->display_irqs_enabled) > > + if (i915->display_irqs_enabled) > > return; > > - dev_priv->display_irqs_enabled = true; > > + i915->display_irqs_enabled = true; > > - if (intel_irqs_enabled(dev_priv)) { > > - vlv_display_irq_reset(dev_priv); > > - vlv_display_irq_postinstall(dev_priv); > > + if (intel_irqs_enabled(i915)) { > > + vlv_display_irq_reset(i915); > > + vlv_display_irq_postinstall(i915); > > } > > } > > -void valleyview_disable_display_irqs(struct drm_i915_private *dev_priv) > > +void valleyview_disable_display_irqs(struct drm_i915_private *i915) > > { > > - lockdep_assert_held(&dev_priv->irq_lock); > > + lockdep_assert_held(&i915->irq_lock); > > - if (!dev_priv->display_irqs_enabled) > > + if (!i915->display_irqs_enabled) > > return; > > - dev_priv->display_irqs_enabled = false; > > + i915->display_irqs_enabled = false; > > - if (intel_irqs_enabled(dev_priv)) > > - vlv_display_irq_reset(dev_priv); > > + if (intel_irqs_enabled(i915)) > > + vlv_display_irq_reset(i915); > > } > > static int valleyview_irq_postinstall(struct drm_device *dev) > > { > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > gen5_gt_irq_postinstall(dev); > > - spin_lock_irq(&dev_priv->irq_lock); > > - if (dev_priv->display_irqs_enabled) > > - vlv_display_irq_postinstall(dev_priv); > > - spin_unlock_irq(&dev_priv->irq_lock); > > + spin_lock_irq(&i915->irq_lock); > > + if (i915->display_irqs_enabled) > > + vlv_display_irq_postinstall(i915); > > + spin_unlock_irq(&i915->irq_lock); > > I915_WRITE(VLV_MASTER_IER, MASTER_INTERRUPT_ENABLE); > > POSTING_READ(VLV_MASTER_IER); > > @@ -4104,9 +4104,9 @@ static int valleyview_irq_postinstall(struct drm_device *dev) > > return 0; > > } > > -static void gen8_gt_irq_postinstall(struct drm_i915_private *dev_priv) > > +static void gen8_gt_irq_postinstall(struct drm_i915_private *i915) > > { > > - struct intel_uncore *uncore = &dev_priv->uncore; > > + struct intel_uncore *uncore = &i915->uncore; > > /* These are interrupts we'll toggle with the ring mask register */ > > u32 gt_interrupts[] = { > > @@ -4126,21 +4126,21 @@ static void gen8_gt_irq_postinstall(struct drm_i915_private *dev_priv) > > GT_CONTEXT_SWITCH_INTERRUPT << GEN8_VECS_IRQ_SHIFT) > > }; > > - dev_priv->pm_ier = 0x0; > > - dev_priv->pm_imr = ~dev_priv->pm_ier; > > + i915->pm_ier = 0x0; > > + i915->pm_imr = ~i915->pm_ier; > > GEN8_IRQ_INIT_NDX(uncore, GT, 0, ~gt_interrupts[0], gt_interrupts[0]); > > GEN8_IRQ_INIT_NDX(uncore, GT, 1, ~gt_interrupts[1], gt_interrupts[1]); > > /* > > * RPS interrupts will get enabled/disabled on demand when RPS itself > > * is enabled/disabled. Same wil be the case for GuC interrupts. > > */ > > - GEN8_IRQ_INIT_NDX(uncore, GT, 2, dev_priv->pm_imr, dev_priv->pm_ier); > > + GEN8_IRQ_INIT_NDX(uncore, GT, 2, i915->pm_imr, i915->pm_ier); > > GEN8_IRQ_INIT_NDX(uncore, GT, 3, ~gt_interrupts[3], gt_interrupts[3]); > > } > > -static void gen8_de_irq_postinstall(struct drm_i915_private *dev_priv) > > +static void gen8_de_irq_postinstall(struct drm_i915_private *i915) > > { > > - struct intel_uncore *uncore = &dev_priv->uncore; > > + struct intel_uncore *uncore = &i915->uncore; > > u32 de_pipe_masked = GEN8_PIPE_CDCLK_CRC_DONE; > > u32 de_pipe_enables; > > @@ -4149,84 +4149,84 @@ static void gen8_de_irq_postinstall(struct drm_i915_private *dev_priv) > > u32 de_misc_masked = GEN8_DE_EDP_PSR; > > enum pipe pipe; > > - if (INTEL_GEN(dev_priv) <= 10) > > + if (INTEL_GEN(i915) <= 10) > > de_misc_masked |= GEN8_DE_MISC_GSE; > > - if (INTEL_GEN(dev_priv) >= 9) { > > + if (INTEL_GEN(i915) >= 9) { > > de_pipe_masked |= GEN9_DE_PIPE_IRQ_FAULT_ERRORS; > > de_port_masked |= GEN9_AUX_CHANNEL_B | GEN9_AUX_CHANNEL_C | > > GEN9_AUX_CHANNEL_D; > > - if (IS_GEN9_LP(dev_priv)) > > + if (IS_GEN9_LP(i915)) > > de_port_masked |= BXT_DE_PORT_GMBUS; > > } else { > > de_pipe_masked |= GEN8_DE_PIPE_IRQ_FAULT_ERRORS; > > } > > - if (INTEL_GEN(dev_priv) >= 11) > > + if (INTEL_GEN(i915) >= 11) > > de_port_masked |= ICL_AUX_CHANNEL_E; > > - if (IS_CNL_WITH_PORT_F(dev_priv) || INTEL_GEN(dev_priv) >= 11) > > + if (IS_CNL_WITH_PORT_F(i915) || INTEL_GEN(i915) >= 11) > > de_port_masked |= CNL_AUX_CHANNEL_F; > > de_pipe_enables = de_pipe_masked | GEN8_PIPE_VBLANK | > > GEN8_PIPE_FIFO_UNDERRUN; > > de_port_enables = de_port_masked; > > - if (IS_GEN9_LP(dev_priv)) > > + if (IS_GEN9_LP(i915)) > > de_port_enables |= BXT_DE_PORT_HOTPLUG_MASK; > > - else if (IS_BROADWELL(dev_priv)) > > + else if (IS_BROADWELL(i915)) > > de_port_enables |= GEN8_PORT_DP_A_HOTPLUG; > > gen3_assert_iir_is_zero(uncore, EDP_PSR_IIR); > > - intel_psr_irq_control(dev_priv, dev_priv->psr.debug); > > + intel_psr_irq_control(i915, i915->psr.debug); > > - for_each_pipe(dev_priv, pipe) { > > - dev_priv->de_irq_mask[pipe] = ~de_pipe_masked; > > + for_each_pipe(i915, pipe) { > > + i915->de_irq_mask[pipe] = ~de_pipe_masked; > > - if (intel_display_power_is_enabled(dev_priv, > > + if (intel_display_power_is_enabled(i915, > > POWER_DOMAIN_PIPE(pipe))) > > GEN8_IRQ_INIT_NDX(uncore, DE_PIPE, pipe, > > - dev_priv->de_irq_mask[pipe], > > + i915->de_irq_mask[pipe], > > de_pipe_enables); > > } > > GEN3_IRQ_INIT(uncore, GEN8_DE_PORT_, ~de_port_masked, de_port_enables); > > GEN3_IRQ_INIT(uncore, GEN8_DE_MISC_, ~de_misc_masked, de_misc_masked); > > - if (INTEL_GEN(dev_priv) >= 11) { > > + if (INTEL_GEN(i915) >= 11) { > > u32 de_hpd_masked = 0; > > u32 de_hpd_enables = GEN11_DE_TC_HOTPLUG_MASK | > > GEN11_DE_TBT_HOTPLUG_MASK; > > GEN3_IRQ_INIT(uncore, GEN11_DE_HPD_, ~de_hpd_masked, > > de_hpd_enables); > > - gen11_hpd_detection_setup(dev_priv); > > - } else if (IS_GEN9_LP(dev_priv)) { > > - bxt_hpd_detection_setup(dev_priv); > > - } else if (IS_BROADWELL(dev_priv)) { > > - ilk_hpd_detection_setup(dev_priv); > > + gen11_hpd_detection_setup(i915); > > + } else if (IS_GEN9_LP(i915)) { > > + bxt_hpd_detection_setup(i915); > > + } else if (IS_BROADWELL(i915)) { > > + ilk_hpd_detection_setup(i915); > > } > > } > > static int gen8_irq_postinstall(struct drm_device *dev) > > { > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > - if (HAS_PCH_SPLIT(dev_priv)) > > + if (HAS_PCH_SPLIT(i915)) > > ibx_irq_pre_postinstall(dev); > > - gen8_gt_irq_postinstall(dev_priv); > > - gen8_de_irq_postinstall(dev_priv); > > + gen8_gt_irq_postinstall(i915); > > + gen8_de_irq_postinstall(i915); > > - if (HAS_PCH_SPLIT(dev_priv)) > > + if (HAS_PCH_SPLIT(i915)) > > ibx_irq_postinstall(dev); > > - gen8_master_intr_enable(dev_priv->uncore.regs); > > + gen8_master_intr_enable(i915->uncore.regs); > > return 0; > > } > > -static void gen11_gt_irq_postinstall(struct drm_i915_private *dev_priv) > > +static void gen11_gt_irq_postinstall(struct drm_i915_private *i915) > > { > > const u32 irqs = GT_RENDER_USER_INTERRUPT | GT_CONTEXT_SWITCH_INTERRUPT; > > @@ -4247,8 +4247,8 @@ static void gen11_gt_irq_postinstall(struct drm_i915_private *dev_priv) > > * RPS interrupts will get enabled/disabled on demand when RPS itself > > * is enabled/disabled. > > */ > > - dev_priv->pm_ier = 0x0; > > - dev_priv->pm_imr = ~dev_priv->pm_ier; > > + i915->pm_ier = 0x0; > > + i915->pm_imr = ~i915->pm_ier; > > I915_WRITE(GEN11_GPM_WGBOXPERF_INTR_ENABLE, 0); > > I915_WRITE(GEN11_GPM_WGBOXPERF_INTR_MASK, ~0); > > @@ -4259,36 +4259,36 @@ static void gen11_gt_irq_postinstall(struct drm_i915_private *dev_priv) > > static void icp_irq_postinstall(struct drm_device *dev) > > { > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > u32 mask = SDE_GMBUS_ICP; > > WARN_ON(I915_READ(SDEIER) != 0); > > I915_WRITE(SDEIER, 0xffffffff); > > POSTING_READ(SDEIER); > > - gen3_assert_iir_is_zero(&dev_priv->uncore, SDEIIR); > > + gen3_assert_iir_is_zero(&i915->uncore, SDEIIR); > > I915_WRITE(SDEIMR, ~mask); > > - icp_hpd_detection_setup(dev_priv); > > + icp_hpd_detection_setup(i915); > > } > > static int gen11_irq_postinstall(struct drm_device *dev) > > { > > - struct drm_i915_private *dev_priv = dev->dev_private; > > - struct intel_uncore *uncore = &dev_priv->uncore; > > + struct drm_i915_private *i915 = dev->dev_private; > > + struct intel_uncore *uncore = &i915->uncore; > > u32 gu_misc_masked = GEN11_GU_MISC_GSE; > > - if (INTEL_PCH_TYPE(dev_priv) >= PCH_ICP) > > + if (INTEL_PCH_TYPE(i915) >= PCH_ICP) > > icp_irq_postinstall(dev); > > - gen11_gt_irq_postinstall(dev_priv); > > - gen8_de_irq_postinstall(dev_priv); > > + gen11_gt_irq_postinstall(i915); > > + gen8_de_irq_postinstall(i915); > > GEN3_IRQ_INIT(uncore, GEN11_GU_MISC_, ~gu_misc_masked, gu_misc_masked); > > I915_WRITE(GEN11_DISPLAY_INT_CTL, GEN11_DISPLAY_IRQ_ENABLE); > > - gen11_master_intr_enable(dev_priv->uncore.regs); > > + gen11_master_intr_enable(i915->uncore.regs); > > POSTING_READ(GEN11_GFX_MSTR_IRQ); > > return 0; > > @@ -4296,14 +4296,14 @@ static int gen11_irq_postinstall(struct drm_device *dev) > > static int cherryview_irq_postinstall(struct drm_device *dev) > > { > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > - gen8_gt_irq_postinstall(dev_priv); > > + gen8_gt_irq_postinstall(i915); > > - spin_lock_irq(&dev_priv->irq_lock); > > - if (dev_priv->display_irqs_enabled) > > - vlv_display_irq_postinstall(dev_priv); > > - spin_unlock_irq(&dev_priv->irq_lock); > > + spin_lock_irq(&i915->irq_lock); > > + if (i915->display_irqs_enabled) > > + vlv_display_irq_postinstall(i915); > > + spin_unlock_irq(&i915->irq_lock); > > I915_WRITE(GEN8_MASTER_IRQ, GEN8_MASTER_IRQ_CONTROL); > > POSTING_READ(GEN8_MASTER_IRQ); > > @@ -4313,18 +4313,18 @@ static int cherryview_irq_postinstall(struct drm_device *dev) > > static void i8xx_irq_reset(struct drm_device *dev) > > { > > - struct drm_i915_private *dev_priv = to_i915(dev); > > - struct intel_uncore *uncore = &dev_priv->uncore; > > + struct drm_i915_private *i915 = to_i915(dev); > > + struct intel_uncore *uncore = &i915->uncore; > > - i9xx_pipestat_irq_reset(dev_priv); > > + i9xx_pipestat_irq_reset(i915); > > GEN2_IRQ_RESET(uncore); > > } > > static int i8xx_irq_postinstall(struct drm_device *dev) > > { > > - struct drm_i915_private *dev_priv = to_i915(dev); > > - struct intel_uncore *uncore = &dev_priv->uncore; > > + struct drm_i915_private *i915 = to_i915(dev); > > + struct intel_uncore *uncore = &i915->uncore; > > u16 enable_mask; > > intel_uncore_write16(uncore, > > @@ -4333,7 +4333,7 @@ static int i8xx_irq_postinstall(struct drm_device *dev) > > I915_ERROR_MEMORY_REFRESH)); > > /* Unmask the interrupts that we always want on. */ > > - dev_priv->irq_mask = > > + i915->irq_mask = > > ~(I915_DISPLAY_PIPE_A_EVENT_INTERRUPT | > > I915_DISPLAY_PIPE_B_EVENT_INTERRUPT | > > I915_MASTER_ERROR_INTERRUPT); > > @@ -4344,14 +4344,14 @@ static int i8xx_irq_postinstall(struct drm_device *dev) > > I915_MASTER_ERROR_INTERRUPT | > > I915_USER_INTERRUPT; > > - GEN2_IRQ_INIT(uncore, dev_priv->irq_mask, enable_mask); > > + GEN2_IRQ_INIT(uncore, i915->irq_mask, enable_mask); > > /* Interrupt setup is already guaranteed to be single-threaded, this is > > * just to make the assert_spin_locked check happy. */ > > - spin_lock_irq(&dev_priv->irq_lock); > > - i915_enable_pipestat(dev_priv, PIPE_A, PIPE_CRC_DONE_INTERRUPT_STATUS); > > - i915_enable_pipestat(dev_priv, PIPE_B, PIPE_CRC_DONE_INTERRUPT_STATUS); > > - spin_unlock_irq(&dev_priv->irq_lock); > > + spin_lock_irq(&i915->irq_lock); > > + i915_enable_pipestat(i915, PIPE_A, PIPE_CRC_DONE_INTERRUPT_STATUS); > > + i915_enable_pipestat(i915, PIPE_B, PIPE_CRC_DONE_INTERRUPT_STATUS); > > + spin_unlock_irq(&i915->irq_lock); > > return 0; > > } > > @@ -4386,7 +4386,7 @@ static void i8xx_error_irq_ack(struct drm_i915_private *i915, > > intel_uncore_write16(uncore, EMR, emr | *eir_stuck); > > } > > -static void i8xx_error_irq_handler(struct drm_i915_private *dev_priv, > > +static void i8xx_error_irq_handler(struct drm_i915_private *i915, > > u16 eir, u16 eir_stuck) > > { > > DRM_DEBUG("Master Error: EIR 0x%04x\n", eir); > > @@ -4395,7 +4395,7 @@ static void i8xx_error_irq_handler(struct drm_i915_private *dev_priv, > > DRM_DEBUG_DRIVER("EIR stuck: 0x%04x, masked\n", eir_stuck); > > } > > -static void i9xx_error_irq_ack(struct drm_i915_private *dev_priv, > > +static void i9xx_error_irq_ack(struct drm_i915_private *i915, > > u32 *eir, u32 *eir_stuck) > > { > > u32 emr; > > @@ -4423,7 +4423,7 @@ static void i9xx_error_irq_ack(struct drm_i915_private *dev_priv, > > I915_WRITE(EMR, emr | *eir_stuck); > > } > > -static void i9xx_error_irq_handler(struct drm_i915_private *dev_priv, > > +static void i9xx_error_irq_handler(struct drm_i915_private *i915, > > u32 eir, u32 eir_stuck) > > { > > DRM_DEBUG("Master Error, EIR 0x%08x\n", eir); > > @@ -4435,21 +4435,21 @@ static void i9xx_error_irq_handler(struct drm_i915_private *dev_priv, > > static irqreturn_t i8xx_irq_handler(int irq, void *arg) > > { > > struct drm_device *dev = arg; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > irqreturn_t ret = IRQ_NONE; > > - if (!intel_irqs_enabled(dev_priv)) > > + if (!intel_irqs_enabled(i915)) > > return IRQ_NONE; > > /* IRQs are synced during runtime_suspend, we don't require a wakeref */ > > - disable_rpm_wakeref_asserts(dev_priv); > > + disable_rpm_wakeref_asserts(i915); > > do { > > u32 pipe_stats[I915_MAX_PIPES] = {}; > > u16 eir = 0, eir_stuck = 0; > > u16 iir; > > - iir = intel_uncore_read16(&dev_priv->uncore, GEN2_IIR); > > + iir = intel_uncore_read16(&i915->uncore, GEN2_IIR); > > if (iir == 0) > > break; > > @@ -4457,53 +4457,53 @@ static irqreturn_t i8xx_irq_handler(int irq, void *arg) > > /* Call regardless, as some status bits might not be > > * signalled in iir */ > > - i9xx_pipestat_irq_ack(dev_priv, iir, pipe_stats); > > + i9xx_pipestat_irq_ack(i915, iir, pipe_stats); > > if (iir & I915_MASTER_ERROR_INTERRUPT) > > - i8xx_error_irq_ack(dev_priv, &eir, &eir_stuck); > > + i8xx_error_irq_ack(i915, &eir, &eir_stuck); > > - intel_uncore_write16(&dev_priv->uncore, GEN2_IIR, iir); > > + intel_uncore_write16(&i915->uncore, GEN2_IIR, iir); > > if (iir & I915_USER_INTERRUPT) > > - intel_engine_breadcrumbs_irq(dev_priv->engine[RCS0]); > > + intel_engine_breadcrumbs_irq(i915->engine[RCS0]); > > if (iir & I915_MASTER_ERROR_INTERRUPT) > > - i8xx_error_irq_handler(dev_priv, eir, eir_stuck); > > + i8xx_error_irq_handler(i915, eir, eir_stuck); > > - i8xx_pipestat_irq_handler(dev_priv, iir, pipe_stats); > > + i8xx_pipestat_irq_handler(i915, iir, pipe_stats); > > } while (0); > > - enable_rpm_wakeref_asserts(dev_priv); > > + enable_rpm_wakeref_asserts(i915); > > return ret; > > } > > static void i915_irq_reset(struct drm_device *dev) > > { > > - struct drm_i915_private *dev_priv = to_i915(dev); > > - struct intel_uncore *uncore = &dev_priv->uncore; > > + struct drm_i915_private *i915 = to_i915(dev); > > + struct intel_uncore *uncore = &i915->uncore; > > - if (I915_HAS_HOTPLUG(dev_priv)) { > > - i915_hotplug_interrupt_update(dev_priv, 0xffffffff, 0); > > + if (I915_HAS_HOTPLUG(i915)) { > > + i915_hotplug_interrupt_update(i915, 0xffffffff, 0); > > I915_WRITE(PORT_HOTPLUG_STAT, I915_READ(PORT_HOTPLUG_STAT)); > > } > > - i9xx_pipestat_irq_reset(dev_priv); > > + i9xx_pipestat_irq_reset(i915); > > GEN3_IRQ_RESET(uncore, GEN2_); > > } > > static int i915_irq_postinstall(struct drm_device *dev) > > { > > - struct drm_i915_private *dev_priv = to_i915(dev); > > - struct intel_uncore *uncore = &dev_priv->uncore; > > + struct drm_i915_private *i915 = to_i915(dev); > > + struct intel_uncore *uncore = &i915->uncore; > > u32 enable_mask; > > I915_WRITE(EMR, ~(I915_ERROR_PAGE_TABLE | > > I915_ERROR_MEMORY_REFRESH)); > > /* Unmask the interrupts that we always want on. */ > > - dev_priv->irq_mask = > > + i915->irq_mask = > > ~(I915_ASLE_INTERRUPT | > > I915_DISPLAY_PIPE_A_EVENT_INTERRUPT | > > I915_DISPLAY_PIPE_B_EVENT_INTERRUPT | > > @@ -4516,23 +4516,23 @@ static int i915_irq_postinstall(struct drm_device *dev) > > I915_MASTER_ERROR_INTERRUPT | > > I915_USER_INTERRUPT; > > - if (I915_HAS_HOTPLUG(dev_priv)) { > > + if (I915_HAS_HOTPLUG(i915)) { > > /* Enable in IER... */ > > enable_mask |= I915_DISPLAY_PORT_INTERRUPT; > > /* and unmask in IMR */ > > - dev_priv->irq_mask &= ~I915_DISPLAY_PORT_INTERRUPT; > > + i915->irq_mask &= ~I915_DISPLAY_PORT_INTERRUPT; > > } > > - GEN3_IRQ_INIT(uncore, GEN2_, dev_priv->irq_mask, enable_mask); > > + GEN3_IRQ_INIT(uncore, GEN2_, i915->irq_mask, enable_mask); > > /* Interrupt setup is already guaranteed to be single-threaded, this is > > * just to make the assert_spin_locked check happy. */ > > - spin_lock_irq(&dev_priv->irq_lock); > > - i915_enable_pipestat(dev_priv, PIPE_A, PIPE_CRC_DONE_INTERRUPT_STATUS); > > - i915_enable_pipestat(dev_priv, PIPE_B, PIPE_CRC_DONE_INTERRUPT_STATUS); > > - spin_unlock_irq(&dev_priv->irq_lock); > > + spin_lock_irq(&i915->irq_lock); > > + i915_enable_pipestat(i915, PIPE_A, PIPE_CRC_DONE_INTERRUPT_STATUS); > > + i915_enable_pipestat(i915, PIPE_B, PIPE_CRC_DONE_INTERRUPT_STATUS); > > + spin_unlock_irq(&i915->irq_lock); > > - i915_enable_asle_pipestat(dev_priv); > > + i915_enable_asle_pipestat(i915); > > return 0; > > } > > @@ -4540,14 +4540,14 @@ static int i915_irq_postinstall(struct drm_device *dev) > > static irqreturn_t i915_irq_handler(int irq, void *arg) > > { > > struct drm_device *dev = arg; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > irqreturn_t ret = IRQ_NONE; > > - if (!intel_irqs_enabled(dev_priv)) > > + if (!intel_irqs_enabled(i915)) > > return IRQ_NONE; > > /* IRQs are synced during runtime_suspend, we don't require a wakeref */ > > - disable_rpm_wakeref_asserts(dev_priv); > > + disable_rpm_wakeref_asserts(i915); > > do { > > u32 pipe_stats[I915_MAX_PIPES] = {}; > > @@ -4561,53 +4561,53 @@ static irqreturn_t i915_irq_handler(int irq, void *arg) > > ret = IRQ_HANDLED; > > - if (I915_HAS_HOTPLUG(dev_priv) && > > + if (I915_HAS_HOTPLUG(i915) && > > iir & I915_DISPLAY_PORT_INTERRUPT) > > - hotplug_status = i9xx_hpd_irq_ack(dev_priv); > > + hotplug_status = i9xx_hpd_irq_ack(i915); > > /* Call regardless, as some status bits might not be > > * signalled in iir */ > > - i9xx_pipestat_irq_ack(dev_priv, iir, pipe_stats); > > + i9xx_pipestat_irq_ack(i915, iir, pipe_stats); > > if (iir & I915_MASTER_ERROR_INTERRUPT) > > - i9xx_error_irq_ack(dev_priv, &eir, &eir_stuck); > > + i9xx_error_irq_ack(i915, &eir, &eir_stuck); > > I915_WRITE(GEN2_IIR, iir); > > if (iir & I915_USER_INTERRUPT) > > - intel_engine_breadcrumbs_irq(dev_priv->engine[RCS0]); > > + intel_engine_breadcrumbs_irq(i915->engine[RCS0]); > > if (iir & I915_MASTER_ERROR_INTERRUPT) > > - i9xx_error_irq_handler(dev_priv, eir, eir_stuck); > > + i9xx_error_irq_handler(i915, eir, eir_stuck); > > if (hotplug_status) > > - i9xx_hpd_irq_handler(dev_priv, hotplug_status); > > + i9xx_hpd_irq_handler(i915, hotplug_status); > > - i915_pipestat_irq_handler(dev_priv, iir, pipe_stats); > > + i915_pipestat_irq_handler(i915, iir, pipe_stats); > > } while (0); > > - enable_rpm_wakeref_asserts(dev_priv); > > + enable_rpm_wakeref_asserts(i915); > > return ret; > > } > > static void i965_irq_reset(struct drm_device *dev) > > { > > - struct drm_i915_private *dev_priv = to_i915(dev); > > - struct intel_uncore *uncore = &dev_priv->uncore; > > + struct drm_i915_private *i915 = to_i915(dev); > > + struct intel_uncore *uncore = &i915->uncore; > > - i915_hotplug_interrupt_update(dev_priv, 0xffffffff, 0); > > + i915_hotplug_interrupt_update(i915, 0xffffffff, 0); > > I915_WRITE(PORT_HOTPLUG_STAT, I915_READ(PORT_HOTPLUG_STAT)); > > - i9xx_pipestat_irq_reset(dev_priv); > > + i9xx_pipestat_irq_reset(i915); > > GEN3_IRQ_RESET(uncore, GEN2_); > > } > > static int i965_irq_postinstall(struct drm_device *dev) > > { > > - struct drm_i915_private *dev_priv = to_i915(dev); > > - struct intel_uncore *uncore = &dev_priv->uncore; > > + struct drm_i915_private *i915 = to_i915(dev); > > + struct intel_uncore *uncore = &i915->uncore; > > u32 enable_mask; > > u32 error_mask; > > @@ -4615,7 +4615,7 @@ static int i965_irq_postinstall(struct drm_device *dev) > > * Enable some error detection, note the instruction error mask > > * bit is reserved, so we leave it masked. > > */ > > - if (IS_G4X(dev_priv)) { > > + if (IS_G4X(i915)) { > > error_mask = ~(GM45_ERROR_PAGE_TABLE | > > GM45_ERROR_MEM_PRIV | > > GM45_ERROR_CP_PRIV | > > @@ -4627,7 +4627,7 @@ static int i965_irq_postinstall(struct drm_device *dev) > > I915_WRITE(EMR, error_mask); > > /* Unmask the interrupts that we always want on. */ > > - dev_priv->irq_mask = > > + i915->irq_mask = > > ~(I915_ASLE_INTERRUPT | > > I915_DISPLAY_PORT_INTERRUPT | > > I915_DISPLAY_PIPE_A_EVENT_INTERRUPT | > > @@ -4642,43 +4642,43 @@ static int i965_irq_postinstall(struct drm_device *dev) > > I915_MASTER_ERROR_INTERRUPT | > > I915_USER_INTERRUPT; > > - if (IS_G4X(dev_priv)) > > + if (IS_G4X(i915)) > > enable_mask |= I915_BSD_USER_INTERRUPT; > > - GEN3_IRQ_INIT(uncore, GEN2_, dev_priv->irq_mask, enable_mask); > > + GEN3_IRQ_INIT(uncore, GEN2_, i915->irq_mask, enable_mask); > > /* Interrupt setup is already guaranteed to be single-threaded, this is > > * just to make the assert_spin_locked check happy. */ > > - spin_lock_irq(&dev_priv->irq_lock); > > - i915_enable_pipestat(dev_priv, PIPE_A, PIPE_GMBUS_INTERRUPT_STATUS); > > - i915_enable_pipestat(dev_priv, PIPE_A, PIPE_CRC_DONE_INTERRUPT_STATUS); > > - i915_enable_pipestat(dev_priv, PIPE_B, PIPE_CRC_DONE_INTERRUPT_STATUS); > > - spin_unlock_irq(&dev_priv->irq_lock); > > + spin_lock_irq(&i915->irq_lock); > > + i915_enable_pipestat(i915, PIPE_A, PIPE_GMBUS_INTERRUPT_STATUS); > > + i915_enable_pipestat(i915, PIPE_A, PIPE_CRC_DONE_INTERRUPT_STATUS); > > + i915_enable_pipestat(i915, PIPE_B, PIPE_CRC_DONE_INTERRUPT_STATUS); > > + spin_unlock_irq(&i915->irq_lock); > > - i915_enable_asle_pipestat(dev_priv); > > + i915_enable_asle_pipestat(i915); > > return 0; > > } > > -static void i915_hpd_irq_setup(struct drm_i915_private *dev_priv) > > +static void i915_hpd_irq_setup(struct drm_i915_private *i915) > > { > > u32 hotplug_en; > > - lockdep_assert_held(&dev_priv->irq_lock); > > + lockdep_assert_held(&i915->irq_lock); > > /* Note HDMI and DP share hotplug bits */ > > /* enable bits are the same for all generations */ > > - hotplug_en = intel_hpd_enabled_irqs(dev_priv, hpd_mask_i915); > > + hotplug_en = intel_hpd_enabled_irqs(i915, hpd_mask_i915); > > /* Programming the CRT detection parameters tends > > to generate a spurious hotplug event about three > > seconds later. So just do it once. > > */ > > - if (IS_G4X(dev_priv)) > > + if (IS_G4X(i915)) > > hotplug_en |= CRT_HOTPLUG_ACTIVATION_PERIOD_64; > > hotplug_en |= CRT_HOTPLUG_VOLTAGE_COMPARE_50; > > /* Ignore TV since it's buggy */ > > - i915_hotplug_interrupt_update_locked(dev_priv, > > + i915_hotplug_interrupt_update_locked(i915, > > HOTPLUG_INT_EN_MASK | > > CRT_HOTPLUG_VOLTAGE_COMPARE_MASK | > > CRT_HOTPLUG_ACTIVATION_PERIOD_64, > > @@ -4688,14 +4688,14 @@ static void i915_hpd_irq_setup(struct drm_i915_private *dev_priv) > > static irqreturn_t i965_irq_handler(int irq, void *arg) > > { > > struct drm_device *dev = arg; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > irqreturn_t ret = IRQ_NONE; > > - if (!intel_irqs_enabled(dev_priv)) > > + if (!intel_irqs_enabled(i915)) > > return IRQ_NONE; > > /* IRQs are synced during runtime_suspend, we don't require a wakeref */ > > - disable_rpm_wakeref_asserts(dev_priv); > > + disable_rpm_wakeref_asserts(i915); > > do { > > u32 pipe_stats[I915_MAX_PIPES] = {}; > > @@ -4710,76 +4710,76 @@ static irqreturn_t i965_irq_handler(int irq, void *arg) > > ret = IRQ_HANDLED; > > if (iir & I915_DISPLAY_PORT_INTERRUPT) > > - hotplug_status = i9xx_hpd_irq_ack(dev_priv); > > + hotplug_status = i9xx_hpd_irq_ack(i915); > > /* Call regardless, as some status bits might not be > > * signalled in iir */ > > - i9xx_pipestat_irq_ack(dev_priv, iir, pipe_stats); > > + i9xx_pipestat_irq_ack(i915, iir, pipe_stats); > > if (iir & I915_MASTER_ERROR_INTERRUPT) > > - i9xx_error_irq_ack(dev_priv, &eir, &eir_stuck); > > + i9xx_error_irq_ack(i915, &eir, &eir_stuck); > > I915_WRITE(GEN2_IIR, iir); > > if (iir & I915_USER_INTERRUPT) > > - intel_engine_breadcrumbs_irq(dev_priv->engine[RCS0]); > > + intel_engine_breadcrumbs_irq(i915->engine[RCS0]); > > if (iir & I915_BSD_USER_INTERRUPT) > > - intel_engine_breadcrumbs_irq(dev_priv->engine[VCS0]); > > + intel_engine_breadcrumbs_irq(i915->engine[VCS0]); > > if (iir & I915_MASTER_ERROR_INTERRUPT) > > - i9xx_error_irq_handler(dev_priv, eir, eir_stuck); > > + i9xx_error_irq_handler(i915, eir, eir_stuck); > > if (hotplug_status) > > - i9xx_hpd_irq_handler(dev_priv, hotplug_status); > > + i9xx_hpd_irq_handler(i915, hotplug_status); > > - i965_pipestat_irq_handler(dev_priv, iir, pipe_stats); > > + i965_pipestat_irq_handler(i915, iir, pipe_stats); > > } while (0); > > - enable_rpm_wakeref_asserts(dev_priv); > > + enable_rpm_wakeref_asserts(i915); > > return ret; > > } > > /** > > * intel_irq_init - initializes irq support > > - * @dev_priv: i915 device instance > > + * @i915: i915 device instance > > * > > * This function initializes all the irq support including work items, timers > > * and all the vtables. It does not setup the interrupt itself though. > > */ > > -void intel_irq_init(struct drm_i915_private *dev_priv) > > +void intel_irq_init(struct drm_i915_private *i915) > > { > > - struct drm_device *dev = &dev_priv->drm; > > - struct intel_rps *rps = &dev_priv->gt_pm.rps; > > + struct drm_device *dev = &i915->drm; > > + struct intel_rps *rps = &i915->gt_pm.rps; > > int i; > > - if (IS_I945GM(dev_priv)) > > - i945gm_vblank_work_init(dev_priv); > > + if (IS_I945GM(i915)) > > + i945gm_vblank_work_init(i915); > > - intel_hpd_init_work(dev_priv); > > + intel_hpd_init_work(i915); > > INIT_WORK(&rps->work, gen6_pm_rps_work); > > - INIT_WORK(&dev_priv->l3_parity.error_work, ivybridge_parity_work); > > + INIT_WORK(&i915->l3_parity.error_work, ivybridge_parity_work); > > for (i = 0; i < MAX_L3_SLICES; ++i) > > - dev_priv->l3_parity.remap_info[i] = NULL; > > + i915->l3_parity.remap_info[i] = NULL; > > - if (HAS_GUC_SCHED(dev_priv) && INTEL_GEN(dev_priv) < 11) > > - dev_priv->pm_guc_events = GEN9_GUC_TO_HOST_INT_EVENT; > > + if (HAS_GUC_SCHED(i915) && INTEL_GEN(i915) < 11) > > + i915->pm_guc_events = GEN9_GUC_TO_HOST_INT_EVENT; > > /* Let's track the enabled rps events */ > > - if (IS_VALLEYVIEW(dev_priv)) > > + if (IS_VALLEYVIEW(i915)) > > /* WaGsvRC0ResidencyMethod:vlv */ > > - dev_priv->pm_rps_events = GEN6_PM_RP_UP_EI_EXPIRED; > > + i915->pm_rps_events = GEN6_PM_RP_UP_EI_EXPIRED; > > else > > - dev_priv->pm_rps_events = (GEN6_PM_RP_UP_THRESHOLD | > > + i915->pm_rps_events = (GEN6_PM_RP_UP_THRESHOLD | > > GEN6_PM_RP_DOWN_THRESHOLD | > > GEN6_PM_RP_DOWN_TIMEOUT); > > /* We share the register with other engine */ > > - if (INTEL_GEN(dev_priv) > 9) > > - GEM_WARN_ON(dev_priv->pm_rps_events & 0xffff0000); > > + if (INTEL_GEN(i915) > 9) > > + GEM_WARN_ON(i915->pm_rps_events & 0xffff0000); > > rps->pm_intrmsk_mbz = 0; > > @@ -4789,15 +4789,15 @@ void intel_irq_init(struct drm_i915_private *dev_priv) > > * > > * TODO: verify if this can be reproduced on VLV,CHV. > > */ > > - if (INTEL_GEN(dev_priv) <= 7) > > + if (INTEL_GEN(i915) <= 7) > > rps->pm_intrmsk_mbz |= GEN6_PM_RP_UP_EI_EXPIRED; > > - if (INTEL_GEN(dev_priv) >= 8) > > + if (INTEL_GEN(i915) >= 8) > > rps->pm_intrmsk_mbz |= GEN8_PMINTR_DISABLE_REDIRECT_TO_GUC; > > - if (INTEL_GEN(dev_priv) >= 5 || IS_G4X(dev_priv)) > > + if (INTEL_GEN(i915) >= 5 || IS_G4X(i915)) > > dev->driver->get_vblank_counter = g4x_get_vblank_counter; > > - else if (INTEL_GEN(dev_priv) >= 3) > > + else if (INTEL_GEN(i915) >= 3) > > dev->driver->get_vblank_counter = i915_get_vblank_counter; > > dev->vblank_disable_immediate = true; > > @@ -4808,83 +4808,83 @@ void intel_irq_init(struct drm_i915_private *dev_priv) > > * outside of the power domain. We defer setting up the display irqs > > * in this case to the runtime pm. > > */ > > - dev_priv->display_irqs_enabled = true; > > - if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) > > - dev_priv->display_irqs_enabled = false; > > + i915->display_irqs_enabled = true; > > + if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) > > + i915->display_irqs_enabled = false; > > - dev_priv->hotplug.hpd_storm_threshold = HPD_STORM_DEFAULT_THRESHOLD; > > + i915->hotplug.hpd_storm_threshold = HPD_STORM_DEFAULT_THRESHOLD; > > /* If we have MST support, we want to avoid doing short HPD IRQ storm > > * detection, as short HPD storms will occur as a natural part of > > * sideband messaging with MST. > > * On older platforms however, IRQ storms can occur with both long and > > * short pulses, as seen on some G4x systems. > > */ > > - dev_priv->hotplug.hpd_short_storm_enabled = !HAS_DP_MST(dev_priv); > > + i915->hotplug.hpd_short_storm_enabled = !HAS_DP_MST(i915); > > dev->driver->get_vblank_timestamp = drm_calc_vbltimestamp_from_scanoutpos; > > dev->driver->get_scanout_position = i915_get_crtc_scanoutpos; > > - if (IS_CHERRYVIEW(dev_priv)) { > > + if (IS_CHERRYVIEW(i915)) { > > dev->driver->irq_handler = cherryview_irq_handler; > > dev->driver->irq_preinstall = cherryview_irq_reset; > > dev->driver->irq_postinstall = cherryview_irq_postinstall; > > dev->driver->irq_uninstall = cherryview_irq_reset; > > dev->driver->enable_vblank = i965_enable_vblank; > > dev->driver->disable_vblank = i965_disable_vblank; > > - dev_priv->display.hpd_irq_setup = i915_hpd_irq_setup; > > - } else if (IS_VALLEYVIEW(dev_priv)) { > > + i915->display.hpd_irq_setup = i915_hpd_irq_setup; > > + } else if (IS_VALLEYVIEW(i915)) { > > dev->driver->irq_handler = valleyview_irq_handler; > > dev->driver->irq_preinstall = valleyview_irq_reset; > > dev->driver->irq_postinstall = valleyview_irq_postinstall; > > dev->driver->irq_uninstall = valleyview_irq_reset; > > dev->driver->enable_vblank = i965_enable_vblank; > > dev->driver->disable_vblank = i965_disable_vblank; > > - dev_priv->display.hpd_irq_setup = i915_hpd_irq_setup; > > - } else if (INTEL_GEN(dev_priv) >= 11) { > > + i915->display.hpd_irq_setup = i915_hpd_irq_setup; > > + } else if (INTEL_GEN(i915) >= 11) { > > dev->driver->irq_handler = gen11_irq_handler; > > dev->driver->irq_preinstall = gen11_irq_reset; > > dev->driver->irq_postinstall = gen11_irq_postinstall; > > dev->driver->irq_uninstall = gen11_irq_reset; > > dev->driver->enable_vblank = gen8_enable_vblank; > > dev->driver->disable_vblank = gen8_disable_vblank; > > - dev_priv->display.hpd_irq_setup = gen11_hpd_irq_setup; > > - } else if (INTEL_GEN(dev_priv) >= 8) { > > + i915->display.hpd_irq_setup = gen11_hpd_irq_setup; > > + } else if (INTEL_GEN(i915) >= 8) { > > dev->driver->irq_handler = gen8_irq_handler; > > dev->driver->irq_preinstall = gen8_irq_reset; > > dev->driver->irq_postinstall = gen8_irq_postinstall; > > dev->driver->irq_uninstall = gen8_irq_reset; > > dev->driver->enable_vblank = gen8_enable_vblank; > > dev->driver->disable_vblank = gen8_disable_vblank; > > - if (IS_GEN9_LP(dev_priv)) > > - dev_priv->display.hpd_irq_setup = bxt_hpd_irq_setup; > > - else if (INTEL_PCH_TYPE(dev_priv) >= PCH_SPT) > > - dev_priv->display.hpd_irq_setup = spt_hpd_irq_setup; > > + if (IS_GEN9_LP(i915)) > > + i915->display.hpd_irq_setup = bxt_hpd_irq_setup; > > + else if (INTEL_PCH_TYPE(i915) >= PCH_SPT) > > + i915->display.hpd_irq_setup = spt_hpd_irq_setup; > > else > > - dev_priv->display.hpd_irq_setup = ilk_hpd_irq_setup; > > - } else if (HAS_PCH_SPLIT(dev_priv)) { > > + i915->display.hpd_irq_setup = ilk_hpd_irq_setup; > > + } else if (HAS_PCH_SPLIT(i915)) { > > dev->driver->irq_handler = ironlake_irq_handler; > > dev->driver->irq_preinstall = ironlake_irq_reset; > > dev->driver->irq_postinstall = ironlake_irq_postinstall; > > dev->driver->irq_uninstall = ironlake_irq_reset; > > dev->driver->enable_vblank = ironlake_enable_vblank; > > dev->driver->disable_vblank = ironlake_disable_vblank; > > - dev_priv->display.hpd_irq_setup = ilk_hpd_irq_setup; > > + i915->display.hpd_irq_setup = ilk_hpd_irq_setup; > > } else { > > - if (IS_GEN(dev_priv, 2)) { > > + if (IS_GEN(i915, 2)) { > > dev->driver->irq_preinstall = i8xx_irq_reset; > > dev->driver->irq_postinstall = i8xx_irq_postinstall; > > dev->driver->irq_handler = i8xx_irq_handler; > > dev->driver->irq_uninstall = i8xx_irq_reset; > > dev->driver->enable_vblank = i8xx_enable_vblank; > > dev->driver->disable_vblank = i8xx_disable_vblank; > > - } else if (IS_I945GM(dev_priv)) { > > + } else if (IS_I945GM(i915)) { > > dev->driver->irq_preinstall = i915_irq_reset; > > dev->driver->irq_postinstall = i915_irq_postinstall; > > dev->driver->irq_uninstall = i915_irq_reset; > > dev->driver->irq_handler = i915_irq_handler; > > dev->driver->enable_vblank = i945gm_enable_vblank; > > dev->driver->disable_vblank = i945gm_disable_vblank; > > - } else if (IS_GEN(dev_priv, 3)) { > > + } else if (IS_GEN(i915, 3)) { > > dev->driver->irq_preinstall = i915_irq_reset; > > dev->driver->irq_postinstall = i915_irq_postinstall; > > dev->driver->irq_uninstall = i915_irq_reset; > > @@ -4899,8 +4899,8 @@ void intel_irq_init(struct drm_i915_private *dev_priv) > > dev->driver->enable_vblank = i965_enable_vblank; > > dev->driver->disable_vblank = i965_disable_vblank; > > } > > - if (I915_HAS_HOTPLUG(dev_priv)) > > - dev_priv->display.hpd_irq_setup = i915_hpd_irq_setup; > > + if (I915_HAS_HOTPLUG(i915)) > > + i915->display.hpd_irq_setup = i915_hpd_irq_setup; > > } > > } > > @@ -4923,7 +4923,7 @@ void intel_irq_fini(struct drm_i915_private *i915) > > /** > > * intel_irq_install - enables the hardware interrupt > > - * @dev_priv: i915 device instance > > + * @i915: i915 device instance > > * > > * This function enables the hardware interrupt handling, but leaves the hotplug > > * handling still disabled. It is called after intel_irq_init(). > > @@ -4932,56 +4932,56 @@ void intel_irq_fini(struct drm_i915_private *i915) > > * but don't want to deal with the hassle of concurrent probe and hotplug > > * workers. Hence the split into this two-stage approach. > > */ > > -int intel_irq_install(struct drm_i915_private *dev_priv) > > +int intel_irq_install(struct drm_i915_private *i915) > > { > > /* > > * We enable some interrupt sources in our postinstall hooks, so mark > > * interrupts as enabled _before_ actually enabling them to avoid > > * special cases in our ordering checks. > > */ > > - dev_priv->runtime_pm.irqs_enabled = true; > > + i915->runtime_pm.irqs_enabled = true; > > - return drm_irq_install(&dev_priv->drm, dev_priv->drm.pdev->irq); > > + return drm_irq_install(&i915->drm, i915->drm.pdev->irq); > > } > > /** > > * intel_irq_uninstall - finilizes all irq handling > > - * @dev_priv: i915 device instance > > + * @i915: i915 device instance > > * > > * This stops interrupt and hotplug handling and unregisters and frees all > > * resources acquired in the init functions. > > */ > > -void intel_irq_uninstall(struct drm_i915_private *dev_priv) > > +void intel_irq_uninstall(struct drm_i915_private *i915) > > { > > - drm_irq_uninstall(&dev_priv->drm); > > - intel_hpd_cancel_work(dev_priv); > > - dev_priv->runtime_pm.irqs_enabled = false; > > + drm_irq_uninstall(&i915->drm); > > + intel_hpd_cancel_work(i915); > > + i915->runtime_pm.irqs_enabled = false; > > } > > /** > > * intel_runtime_pm_disable_interrupts - runtime interrupt disabling > > - * @dev_priv: i915 device instance > > + * @i915: i915 device instance > > * > > * This function is used to disable interrupts at runtime, both in the runtime > > * pm and the system suspend/resume code. > > */ > > -void intel_runtime_pm_disable_interrupts(struct drm_i915_private *dev_priv) > > +void intel_runtime_pm_disable_interrupts(struct drm_i915_private *i915) > > { > > - dev_priv->drm.driver->irq_uninstall(&dev_priv->drm); > > - dev_priv->runtime_pm.irqs_enabled = false; > > - synchronize_irq(dev_priv->drm.irq); > > + i915->drm.driver->irq_uninstall(&i915->drm); > > + i915->runtime_pm.irqs_enabled = false; > > + synchronize_irq(i915->drm.irq); > > } > > /** > > * intel_runtime_pm_enable_interrupts - runtime interrupt enabling > > - * @dev_priv: i915 device instance > > + * @i915: i915 device instance > > * > > * This function is used to enable interrupts at runtime, both in the runtime > > * pm and the system suspend/resume code. > > */ > > -void intel_runtime_pm_enable_interrupts(struct drm_i915_private *dev_priv) > > +void intel_runtime_pm_enable_interrupts(struct drm_i915_private *i915) > > { > > - dev_priv->runtime_pm.irqs_enabled = true; > > - dev_priv->drm.driver->irq_preinstall(&dev_priv->drm); > > - dev_priv->drm.driver->irq_postinstall(&dev_priv->drm); > > + i915->runtime_pm.irqs_enabled = true; > > + i915->drm.driver->irq_preinstall(&i915->drm); > > + i915->drm.driver->irq_postinstall(&i915->drm); > > } > > diff --git a/drivers/gpu/drm/i915/i915_irq.h b/drivers/gpu/drm/i915/i915_irq.h > > index cb25dd213308..604504993513 100644 > > --- a/drivers/gpu/drm/i915/i915_irq.h > > +++ b/drivers/gpu/drm/i915/i915_irq.h > > @@ -13,77 +13,77 @@ > > struct drm_i915_private; > > struct intel_crtc; > > -extern void intel_irq_init(struct drm_i915_private *dev_priv); > > -extern void intel_irq_fini(struct drm_i915_private *dev_priv); > > -int intel_irq_install(struct drm_i915_private *dev_priv); > > -void intel_irq_uninstall(struct drm_i915_private *dev_priv); > > +extern void intel_irq_init(struct drm_i915_private *i915); > > +extern void intel_irq_fini(struct drm_i915_private *i915); > > +int intel_irq_install(struct drm_i915_private *i915); > > +void intel_irq_uninstall(struct drm_i915_private *i915); > > -u32 i915_pipestat_enable_mask(struct drm_i915_private *dev_priv, > > +u32 i915_pipestat_enable_mask(struct drm_i915_private *i915, > > enum pipe pipe); > > void > > -i915_enable_pipestat(struct drm_i915_private *dev_priv, enum pipe pipe, > > +i915_enable_pipestat(struct drm_i915_private *i915, enum pipe pipe, > > u32 status_mask); > > void > > -i915_disable_pipestat(struct drm_i915_private *dev_priv, enum pipe pipe, > > +i915_disable_pipestat(struct drm_i915_private *i915, enum pipe pipe, > > u32 status_mask); > > -void valleyview_enable_display_irqs(struct drm_i915_private *dev_priv); > > -void valleyview_disable_display_irqs(struct drm_i915_private *dev_priv); > > +void valleyview_enable_display_irqs(struct drm_i915_private *i915); > > +void valleyview_disable_display_irqs(struct drm_i915_private *i915); > > -void i915_hotplug_interrupt_update(struct drm_i915_private *dev_priv, > > +void i915_hotplug_interrupt_update(struct drm_i915_private *i915, > > u32 mask, > > u32 bits); > > -void ilk_update_display_irq(struct drm_i915_private *dev_priv, > > +void ilk_update_display_irq(struct drm_i915_private *i915, > > u32 interrupt_mask, > > u32 enabled_irq_mask); > > static inline void > > -ilk_enable_display_irq(struct drm_i915_private *dev_priv, u32 bits) > > +ilk_enable_display_irq(struct drm_i915_private *i915, u32 bits) > > { > > - ilk_update_display_irq(dev_priv, bits, bits); > > + ilk_update_display_irq(i915, bits, bits); > > } > > static inline void > > -ilk_disable_display_irq(struct drm_i915_private *dev_priv, u32 bits) > > +ilk_disable_display_irq(struct drm_i915_private *i915, u32 bits) > > { > > - ilk_update_display_irq(dev_priv, bits, 0); > > + ilk_update_display_irq(i915, bits, 0); > > } > > -void bdw_update_pipe_irq(struct drm_i915_private *dev_priv, > > +void bdw_update_pipe_irq(struct drm_i915_private *i915, > > enum pipe pipe, > > u32 interrupt_mask, > > u32 enabled_irq_mask); > > -static inline void bdw_enable_pipe_irq(struct drm_i915_private *dev_priv, > > +static inline void bdw_enable_pipe_irq(struct drm_i915_private *i915, > > enum pipe pipe, u32 bits) > > { > > - bdw_update_pipe_irq(dev_priv, pipe, bits, bits); > > + bdw_update_pipe_irq(i915, pipe, bits, bits); > > } > > -static inline void bdw_disable_pipe_irq(struct drm_i915_private *dev_priv, > > +static inline void bdw_disable_pipe_irq(struct drm_i915_private *i915, > > enum pipe pipe, u32 bits) > > { > > - bdw_update_pipe_irq(dev_priv, pipe, bits, 0); > > + bdw_update_pipe_irq(i915, pipe, bits, 0); > > } > > -void ibx_display_interrupt_update(struct drm_i915_private *dev_priv, > > +void ibx_display_interrupt_update(struct drm_i915_private *i915, > > u32 interrupt_mask, > > u32 enabled_irq_mask); > > static inline void > > -ibx_enable_display_interrupt(struct drm_i915_private *dev_priv, u32 bits) > > +ibx_enable_display_interrupt(struct drm_i915_private *i915, u32 bits) > > { > > - ibx_display_interrupt_update(dev_priv, bits, bits); > > + ibx_display_interrupt_update(i915, bits, bits); > > } > > static inline void > > -ibx_disable_display_interrupt(struct drm_i915_private *dev_priv, u32 bits) > > +ibx_disable_display_interrupt(struct drm_i915_private *i915, u32 bits) > > { > > - ibx_display_interrupt_update(dev_priv, bits, 0); > > + ibx_display_interrupt_update(i915, bits, 0); > > } > > -void gen5_enable_gt_irq(struct drm_i915_private *dev_priv, u32 mask); > > -void gen5_disable_gt_irq(struct drm_i915_private *dev_priv, u32 mask); > > -void gen6_mask_pm_irq(struct drm_i915_private *dev_priv, u32 mask); > > -void gen6_unmask_pm_irq(struct drm_i915_private *dev_priv, u32 mask); > > -void gen11_reset_rps_interrupts(struct drm_i915_private *dev_priv); > > -void gen6_reset_rps_interrupts(struct drm_i915_private *dev_priv); > > -void gen6_enable_rps_interrupts(struct drm_i915_private *dev_priv); > > -void gen6_disable_rps_interrupts(struct drm_i915_private *dev_priv); > > -void gen6_rps_reset_ei(struct drm_i915_private *dev_priv); > > +void gen5_enable_gt_irq(struct drm_i915_private *i915, u32 mask); > > +void gen5_disable_gt_irq(struct drm_i915_private *i915, u32 mask); > > +void gen6_mask_pm_irq(struct drm_i915_private *i915, u32 mask); > > +void gen6_unmask_pm_irq(struct drm_i915_private *i915, u32 mask); > > +void gen11_reset_rps_interrupts(struct drm_i915_private *i915); > > +void gen6_reset_rps_interrupts(struct drm_i915_private *i915); > > +void gen6_enable_rps_interrupts(struct drm_i915_private *i915); > > +void gen6_disable_rps_interrupts(struct drm_i915_private *i915); > > +void gen6_rps_reset_ei(struct drm_i915_private *i915); > > static inline u32 gen6_sanitize_rps_pm_mask(const struct drm_i915_private *i915, > > u32 mask) > > @@ -91,25 +91,25 @@ static inline u32 gen6_sanitize_rps_pm_mask(const struct drm_i915_private *i915, > > return mask & ~i915->gt_pm.rps.pm_intrmsk_mbz; > > } > > -void intel_runtime_pm_disable_interrupts(struct drm_i915_private *dev_priv); > > -void intel_runtime_pm_enable_interrupts(struct drm_i915_private *dev_priv); > > -static inline bool intel_irqs_enabled(struct drm_i915_private *dev_priv) > > +void intel_runtime_pm_disable_interrupts(struct drm_i915_private *i915); > > +void intel_runtime_pm_enable_interrupts(struct drm_i915_private *i915); > > +static inline bool intel_irqs_enabled(struct drm_i915_private *i915) > > { > > /* > > * We only use drm_irq_uninstall() at unload and VT switch, so > > * this is the only thing we need to check. > > */ > > - return dev_priv->runtime_pm.irqs_enabled; > > + return i915->runtime_pm.irqs_enabled; > > } > > int intel_get_crtc_scanline(struct intel_crtc *crtc); > > -void gen8_irq_power_well_post_enable(struct drm_i915_private *dev_priv, > > +void gen8_irq_power_well_post_enable(struct drm_i915_private *i915, > > u8 pipe_mask); > > -void gen8_irq_power_well_pre_disable(struct drm_i915_private *dev_priv, > > +void gen8_irq_power_well_pre_disable(struct drm_i915_private *i915, > > u8 pipe_mask); > > -void gen9_reset_guc_interrupts(struct drm_i915_private *dev_priv); > > -void gen9_enable_guc_interrupts(struct drm_i915_private *dev_priv); > > -void gen9_disable_guc_interrupts(struct drm_i915_private *dev_priv); > > +void gen9_reset_guc_interrupts(struct drm_i915_private *i915); > > +void gen9_enable_guc_interrupts(struct drm_i915_private *i915); > > +void gen9_disable_guc_interrupts(struct drm_i915_private *i915); > > void gen11_reset_guc_interrupts(struct drm_i915_private *i915); > > void gen11_enable_guc_interrupts(struct drm_i915_private *i915); > > void gen11_disable_guc_interrupts(struct drm_i915_private *i915); > > diff --git a/drivers/gpu/drm/i915/i915_memcpy.c b/drivers/gpu/drm/i915/i915_memcpy.c > > index 79f8ec756362..2bd664c16ba0 100644 > > --- a/drivers/gpu/drm/i915/i915_memcpy.c > > +++ b/drivers/gpu/drm/i915/i915_memcpy.c > > @@ -94,7 +94,7 @@ bool i915_memcpy_from_wc(void *dst, const void *src, unsigned long len) > > return false; > > } > > -void i915_memcpy_init_early(struct drm_i915_private *dev_priv) > > +void i915_memcpy_init_early(struct drm_i915_private *i915) > > { > > /* > > * Some hypervisors (e.g. KVM) don't support VEX-prefix instructions > > diff --git a/drivers/gpu/drm/i915/i915_oa_bdw.c b/drivers/gpu/drm/i915/i915_oa_bdw.c > > index 4acdb94555b7..7c97cd90e973 100644 > > --- a/drivers/gpu/drm/i915/i915_oa_bdw.c > > +++ b/drivers/gpu/drm/i915/i915_oa_bdw.c > > @@ -64,28 +64,28 @@ show_test_oa_id(struct device *kdev, struct device_attribute *attr, char *buf) > > } > > void > > -i915_perf_load_test_config_bdw(struct drm_i915_private *dev_priv) > > +i915_perf_load_test_config_bdw(struct drm_i915_private *i915) > > { > > - strlcpy(dev_priv->perf.oa.test_config.uuid, > > + strlcpy(i915->perf.oa.test_config.uuid, > > "d6de6f55-e526-4f79-a6a6-d7315c09044e", > > - sizeof(dev_priv->perf.oa.test_config.uuid)); > > - dev_priv->perf.oa.test_config.id = 1; > > + sizeof(i915->perf.oa.test_config.uuid)); > > + i915->perf.oa.test_config.id = 1; > > - dev_priv->perf.oa.test_config.mux_regs = mux_config_test_oa; > > - dev_priv->perf.oa.test_config.mux_regs_len = ARRAY_SIZE(mux_config_test_oa); > > + i915->perf.oa.test_config.mux_regs = mux_config_test_oa; > > + i915->perf.oa.test_config.mux_regs_len = ARRAY_SIZE(mux_config_test_oa); > > - dev_priv->perf.oa.test_config.b_counter_regs = b_counter_config_test_oa; > > - dev_priv->perf.oa.test_config.b_counter_regs_len = ARRAY_SIZE(b_counter_config_test_oa); > > + i915->perf.oa.test_config.b_counter_regs = b_counter_config_test_oa; > > + i915->perf.oa.test_config.b_counter_regs_len = ARRAY_SIZE(b_counter_config_test_oa); > > - dev_priv->perf.oa.test_config.flex_regs = flex_eu_config_test_oa; > > - dev_priv->perf.oa.test_config.flex_regs_len = ARRAY_SIZE(flex_eu_config_test_oa); > > + i915->perf.oa.test_config.flex_regs = flex_eu_config_test_oa; > > + i915->perf.oa.test_config.flex_regs_len = ARRAY_SIZE(flex_eu_config_test_oa); > > - dev_priv->perf.oa.test_config.sysfs_metric.name = "d6de6f55-e526-4f79-a6a6-d7315c09044e"; > > - dev_priv->perf.oa.test_config.sysfs_metric.attrs = dev_priv->perf.oa.test_config.attrs; > > + i915->perf.oa.test_config.sysfs_metric.name = "d6de6f55-e526-4f79-a6a6-d7315c09044e"; > > + i915->perf.oa.test_config.sysfs_metric.attrs = i915->perf.oa.test_config.attrs; > > - dev_priv->perf.oa.test_config.attrs[0] = &dev_priv->perf.oa.test_config.sysfs_metric_id.attr; > > + i915->perf.oa.test_config.attrs[0] = &i915->perf.oa.test_config.sysfs_metric_id.attr; > > - dev_priv->perf.oa.test_config.sysfs_metric_id.attr.name = "id"; > > - dev_priv->perf.oa.test_config.sysfs_metric_id.attr.mode = 0444; > > - dev_priv->perf.oa.test_config.sysfs_metric_id.show = show_test_oa_id; > > + i915->perf.oa.test_config.sysfs_metric_id.attr.name = "id"; > > + i915->perf.oa.test_config.sysfs_metric_id.attr.mode = 0444; > > + i915->perf.oa.test_config.sysfs_metric_id.show = show_test_oa_id; > > } > > diff --git a/drivers/gpu/drm/i915/i915_oa_bdw.h b/drivers/gpu/drm/i915/i915_oa_bdw.h > > index 0e667f1a8aa1..69479e5b436f 100644 > > --- a/drivers/gpu/drm/i915/i915_oa_bdw.h > > +++ b/drivers/gpu/drm/i915/i915_oa_bdw.h > > @@ -10,6 +10,6 @@ > > #ifndef __I915_OA_BDW_H__ > > #define __I915_OA_BDW_H__ > > -extern void i915_perf_load_test_config_bdw(struct drm_i915_private *dev_priv); > > +extern void i915_perf_load_test_config_bdw(struct drm_i915_private *i915); > > #endif > > diff --git a/drivers/gpu/drm/i915/i915_oa_bxt.c b/drivers/gpu/drm/i915/i915_oa_bxt.c > > index a44195c39923..5ec236b33ecc 100644 > > --- a/drivers/gpu/drm/i915/i915_oa_bxt.c > > +++ b/drivers/gpu/drm/i915/i915_oa_bxt.c > > @@ -62,28 +62,28 @@ show_test_oa_id(struct device *kdev, struct device_attribute *attr, char *buf) > > } > > void > > -i915_perf_load_test_config_bxt(struct drm_i915_private *dev_priv) > > +i915_perf_load_test_config_bxt(struct drm_i915_private *i915) > > { > > - strlcpy(dev_priv->perf.oa.test_config.uuid, > > + strlcpy(i915->perf.oa.test_config.uuid, > > "5ee72f5c-092f-421e-8b70-225f7c3e9612", > > - sizeof(dev_priv->perf.oa.test_config.uuid)); > > - dev_priv->perf.oa.test_config.id = 1; > > + sizeof(i915->perf.oa.test_config.uuid)); > > + i915->perf.oa.test_config.id = 1; > > - dev_priv->perf.oa.test_config.mux_regs = mux_config_test_oa; > > - dev_priv->perf.oa.test_config.mux_regs_len = ARRAY_SIZE(mux_config_test_oa); > > + i915->perf.oa.test_config.mux_regs = mux_config_test_oa; > > + i915->perf.oa.test_config.mux_regs_len = ARRAY_SIZE(mux_config_test_oa); > > - dev_priv->perf.oa.test_config.b_counter_regs = b_counter_config_test_oa; > > - dev_priv->perf.oa.test_config.b_counter_regs_len = ARRAY_SIZE(b_counter_config_test_oa); > > + i915->perf.oa.test_config.b_counter_regs = b_counter_config_test_oa; > > + i915->perf.oa.test_config.b_counter_regs_len = ARRAY_SIZE(b_counter_config_test_oa); > > - dev_priv->perf.oa.test_config.flex_regs = flex_eu_config_test_oa; > > - dev_priv->perf.oa.test_config.flex_regs_len = ARRAY_SIZE(flex_eu_config_test_oa); > > + i915->perf.oa.test_config.flex_regs = flex_eu_config_test_oa; > > + i915->perf.oa.test_config.flex_regs_len = ARRAY_SIZE(flex_eu_config_test_oa); > > - dev_priv->perf.oa.test_config.sysfs_metric.name = "5ee72f5c-092f-421e-8b70-225f7c3e9612"; > > - dev_priv->perf.oa.test_config.sysfs_metric.attrs = dev_priv->perf.oa.test_config.attrs; > > + i915->perf.oa.test_config.sysfs_metric.name = "5ee72f5c-092f-421e-8b70-225f7c3e9612"; > > + i915->perf.oa.test_config.sysfs_metric.attrs = i915->perf.oa.test_config.attrs; > > - dev_priv->perf.oa.test_config.attrs[0] = &dev_priv->perf.oa.test_config.sysfs_metric_id.attr; > > + i915->perf.oa.test_config.attrs[0] = &i915->perf.oa.test_config.sysfs_metric_id.attr; > > - dev_priv->perf.oa.test_config.sysfs_metric_id.attr.name = "id"; > > - dev_priv->perf.oa.test_config.sysfs_metric_id.attr.mode = 0444; > > - dev_priv->perf.oa.test_config.sysfs_metric_id.show = show_test_oa_id; > > + i915->perf.oa.test_config.sysfs_metric_id.attr.name = "id"; > > + i915->perf.oa.test_config.sysfs_metric_id.attr.mode = 0444; > > + i915->perf.oa.test_config.sysfs_metric_id.show = show_test_oa_id; > > } > > diff --git a/drivers/gpu/drm/i915/i915_oa_bxt.h b/drivers/gpu/drm/i915/i915_oa_bxt.h > > index 679e92cf4f1d..47d8327f38a1 100644 > > --- a/drivers/gpu/drm/i915/i915_oa_bxt.h > > +++ b/drivers/gpu/drm/i915/i915_oa_bxt.h > > @@ -10,6 +10,6 @@ > > #ifndef __I915_OA_BXT_H__ > > #define __I915_OA_BXT_H__ > > -extern void i915_perf_load_test_config_bxt(struct drm_i915_private *dev_priv); > > +extern void i915_perf_load_test_config_bxt(struct drm_i915_private *i915); > > #endif > > diff --git a/drivers/gpu/drm/i915/i915_oa_cflgt2.c b/drivers/gpu/drm/i915/i915_oa_cflgt2.c > > index 7f60d51b8761..3972cd7dcbfc 100644 > > --- a/drivers/gpu/drm/i915/i915_oa_cflgt2.c > > +++ b/drivers/gpu/drm/i915/i915_oa_cflgt2.c > > @@ -63,28 +63,28 @@ show_test_oa_id(struct device *kdev, struct device_attribute *attr, char *buf) > > } > > void > > -i915_perf_load_test_config_cflgt2(struct drm_i915_private *dev_priv) > > +i915_perf_load_test_config_cflgt2(struct drm_i915_private *i915) > > { > > - strlcpy(dev_priv->perf.oa.test_config.uuid, > > + strlcpy(i915->perf.oa.test_config.uuid, > > "74fb4902-d3d3-4237-9e90-cbdc68d0a446", > > - sizeof(dev_priv->perf.oa.test_config.uuid)); > > - dev_priv->perf.oa.test_config.id = 1; > > + sizeof(i915->perf.oa.test_config.uuid)); > > + i915->perf.oa.test_config.id = 1; > > - dev_priv->perf.oa.test_config.mux_regs = mux_config_test_oa; > > - dev_priv->perf.oa.test_config.mux_regs_len = ARRAY_SIZE(mux_config_test_oa); > > + i915->perf.oa.test_config.mux_regs = mux_config_test_oa; > > + i915->perf.oa.test_config.mux_regs_len = ARRAY_SIZE(mux_config_test_oa); > > - dev_priv->perf.oa.test_config.b_counter_regs = b_counter_config_test_oa; > > - dev_priv->perf.oa.test_config.b_counter_regs_len = ARRAY_SIZE(b_counter_config_test_oa); > > + i915->perf.oa.test_config.b_counter_regs = b_counter_config_test_oa; > > + i915->perf.oa.test_config.b_counter_regs_len = ARRAY_SIZE(b_counter_config_test_oa); > > - dev_priv->perf.oa.test_config.flex_regs = flex_eu_config_test_oa; > > - dev_priv->perf.oa.test_config.flex_regs_len = ARRAY_SIZE(flex_eu_config_test_oa); > > + i915->perf.oa.test_config.flex_regs = flex_eu_config_test_oa; > > + i915->perf.oa.test_config.flex_regs_len = ARRAY_SIZE(flex_eu_config_test_oa); > > - dev_priv->perf.oa.test_config.sysfs_metric.name = "74fb4902-d3d3-4237-9e90-cbdc68d0a446"; > > - dev_priv->perf.oa.test_config.sysfs_metric.attrs = dev_priv->perf.oa.test_config.attrs; > > + i915->perf.oa.test_config.sysfs_metric.name = "74fb4902-d3d3-4237-9e90-cbdc68d0a446"; > > + i915->perf.oa.test_config.sysfs_metric.attrs = i915->perf.oa.test_config.attrs; > > - dev_priv->perf.oa.test_config.attrs[0] = &dev_priv->perf.oa.test_config.sysfs_metric_id.attr; > > + i915->perf.oa.test_config.attrs[0] = &i915->perf.oa.test_config.sysfs_metric_id.attr; > > - dev_priv->perf.oa.test_config.sysfs_metric_id.attr.name = "id"; > > - dev_priv->perf.oa.test_config.sysfs_metric_id.attr.mode = 0444; > > - dev_priv->perf.oa.test_config.sysfs_metric_id.show = show_test_oa_id; > > + i915->perf.oa.test_config.sysfs_metric_id.attr.name = "id"; > > + i915->perf.oa.test_config.sysfs_metric_id.attr.mode = 0444; > > + i915->perf.oa.test_config.sysfs_metric_id.show = show_test_oa_id; > > } > > diff --git a/drivers/gpu/drm/i915/i915_oa_cflgt2.h b/drivers/gpu/drm/i915/i915_oa_cflgt2.h > > index 4d6025559bbe..434545c9c5d9 100644 > > --- a/drivers/gpu/drm/i915/i915_oa_cflgt2.h > > +++ b/drivers/gpu/drm/i915/i915_oa_cflgt2.h > > @@ -10,6 +10,6 @@ > > #ifndef __I915_OA_CFLGT2_H__ > > #define __I915_OA_CFLGT2_H__ > > -extern void i915_perf_load_test_config_cflgt2(struct drm_i915_private *dev_priv); > > +extern void i915_perf_load_test_config_cflgt2(struct drm_i915_private *i915); > > #endif > > diff --git a/drivers/gpu/drm/i915/i915_oa_cflgt3.c b/drivers/gpu/drm/i915/i915_oa_cflgt3.c > > index a92c38e3a0ce..c00b9d9d6ddf 100644 > > --- a/drivers/gpu/drm/i915/i915_oa_cflgt3.c > > +++ b/drivers/gpu/drm/i915/i915_oa_cflgt3.c > > @@ -63,28 +63,28 @@ show_test_oa_id(struct device *kdev, struct device_attribute *attr, char *buf) > > } > > void > > -i915_perf_load_test_config_cflgt3(struct drm_i915_private *dev_priv) > > +i915_perf_load_test_config_cflgt3(struct drm_i915_private *i915) > > { > > - strlcpy(dev_priv->perf.oa.test_config.uuid, > > + strlcpy(i915->perf.oa.test_config.uuid, > > "577e8e2c-3fa0-4875-8743-3538d585e3b0", > > - sizeof(dev_priv->perf.oa.test_config.uuid)); > > - dev_priv->perf.oa.test_config.id = 1; > > + sizeof(i915->perf.oa.test_config.uuid)); > > + i915->perf.oa.test_config.id = 1; > > - dev_priv->perf.oa.test_config.mux_regs = mux_config_test_oa; > > - dev_priv->perf.oa.test_config.mux_regs_len = ARRAY_SIZE(mux_config_test_oa); > > + i915->perf.oa.test_config.mux_regs = mux_config_test_oa; > > + i915->perf.oa.test_config.mux_regs_len = ARRAY_SIZE(mux_config_test_oa); > > - dev_priv->perf.oa.test_config.b_counter_regs = b_counter_config_test_oa; > > - dev_priv->perf.oa.test_config.b_counter_regs_len = ARRAY_SIZE(b_counter_config_test_oa); > > + i915->perf.oa.test_config.b_counter_regs = b_counter_config_test_oa; > > + i915->perf.oa.test_config.b_counter_regs_len = ARRAY_SIZE(b_counter_config_test_oa); > > - dev_priv->perf.oa.test_config.flex_regs = flex_eu_config_test_oa; > > - dev_priv->perf.oa.test_config.flex_regs_len = ARRAY_SIZE(flex_eu_config_test_oa); > > + i915->perf.oa.test_config.flex_regs = flex_eu_config_test_oa; > > + i915->perf.oa.test_config.flex_regs_len = ARRAY_SIZE(flex_eu_config_test_oa); > > - dev_priv->perf.oa.test_config.sysfs_metric.name = "577e8e2c-3fa0-4875-8743-3538d585e3b0"; > > - dev_priv->perf.oa.test_config.sysfs_metric.attrs = dev_priv->perf.oa.test_config.attrs; > > + i915->perf.oa.test_config.sysfs_metric.name = "577e8e2c-3fa0-4875-8743-3538d585e3b0"; > > + i915->perf.oa.test_config.sysfs_metric.attrs = i915->perf.oa.test_config.attrs; > > - dev_priv->perf.oa.test_config.attrs[0] = &dev_priv->perf.oa.test_config.sysfs_metric_id.attr; > > + i915->perf.oa.test_config.attrs[0] = &i915->perf.oa.test_config.sysfs_metric_id.attr; > > - dev_priv->perf.oa.test_config.sysfs_metric_id.attr.name = "id"; > > - dev_priv->perf.oa.test_config.sysfs_metric_id.attr.mode = 0444; > > - dev_priv->perf.oa.test_config.sysfs_metric_id.show = show_test_oa_id; > > + i915->perf.oa.test_config.sysfs_metric_id.attr.name = "id"; > > + i915->perf.oa.test_config.sysfs_metric_id.attr.mode = 0444; > > + i915->perf.oa.test_config.sysfs_metric_id.show = show_test_oa_id; > > } > > diff --git a/drivers/gpu/drm/i915/i915_oa_cflgt3.h b/drivers/gpu/drm/i915/i915_oa_cflgt3.h > > index 0697f4077402..2c23eda056aa 100644 > > --- a/drivers/gpu/drm/i915/i915_oa_cflgt3.h > > +++ b/drivers/gpu/drm/i915/i915_oa_cflgt3.h > > @@ -10,6 +10,6 @@ > > #ifndef __I915_OA_CFLGT3_H__ > > #define __I915_OA_CFLGT3_H__ > > -extern void i915_perf_load_test_config_cflgt3(struct drm_i915_private *dev_priv); > > +extern void i915_perf_load_test_config_cflgt3(struct drm_i915_private *i915); > > #endif > > diff --git a/drivers/gpu/drm/i915/i915_oa_chv.c b/drivers/gpu/drm/i915/i915_oa_chv.c > > index 71ec889a0114..5829d24d0f7c 100644 > > --- a/drivers/gpu/drm/i915/i915_oa_chv.c > > +++ b/drivers/gpu/drm/i915/i915_oa_chv.c > > @@ -63,28 +63,28 @@ show_test_oa_id(struct device *kdev, struct device_attribute *attr, char *buf) > > } > > void > > -i915_perf_load_test_config_chv(struct drm_i915_private *dev_priv) > > +i915_perf_load_test_config_chv(struct drm_i915_private *i915) > > { > > - strlcpy(dev_priv->perf.oa.test_config.uuid, > > + strlcpy(i915->perf.oa.test_config.uuid, > > "4a534b07-cba3-414d-8d60-874830e883aa", > > - sizeof(dev_priv->perf.oa.test_config.uuid)); > > - dev_priv->perf.oa.test_config.id = 1; > > + sizeof(i915->perf.oa.test_config.uuid)); > > + i915->perf.oa.test_config.id = 1; > > - dev_priv->perf.oa.test_config.mux_regs = mux_config_test_oa; > > - dev_priv->perf.oa.test_config.mux_regs_len = ARRAY_SIZE(mux_config_test_oa); > > + i915->perf.oa.test_config.mux_regs = mux_config_test_oa; > > + i915->perf.oa.test_config.mux_regs_len = ARRAY_SIZE(mux_config_test_oa); > > - dev_priv->perf.oa.test_config.b_counter_regs = b_counter_config_test_oa; > > - dev_priv->perf.oa.test_config.b_counter_regs_len = ARRAY_SIZE(b_counter_config_test_oa); > > + i915->perf.oa.test_config.b_counter_regs = b_counter_config_test_oa; > > + i915->perf.oa.test_config.b_counter_regs_len = ARRAY_SIZE(b_counter_config_test_oa); > > - dev_priv->perf.oa.test_config.flex_regs = flex_eu_config_test_oa; > > - dev_priv->perf.oa.test_config.flex_regs_len = ARRAY_SIZE(flex_eu_config_test_oa); > > + i915->perf.oa.test_config.flex_regs = flex_eu_config_test_oa; > > + i915->perf.oa.test_config.flex_regs_len = ARRAY_SIZE(flex_eu_config_test_oa); > > - dev_priv->perf.oa.test_config.sysfs_metric.name = "4a534b07-cba3-414d-8d60-874830e883aa"; > > - dev_priv->perf.oa.test_config.sysfs_metric.attrs = dev_priv->perf.oa.test_config.attrs; > > + i915->perf.oa.test_config.sysfs_metric.name = "4a534b07-cba3-414d-8d60-874830e883aa"; > > + i915->perf.oa.test_config.sysfs_metric.attrs = i915->perf.oa.test_config.attrs; > > - dev_priv->perf.oa.test_config.attrs[0] = &dev_priv->perf.oa.test_config.sysfs_metric_id.attr; > > + i915->perf.oa.test_config.attrs[0] = &i915->perf.oa.test_config.sysfs_metric_id.attr; > > - dev_priv->perf.oa.test_config.sysfs_metric_id.attr.name = "id"; > > - dev_priv->perf.oa.test_config.sysfs_metric_id.attr.mode = 0444; > > - dev_priv->perf.oa.test_config.sysfs_metric_id.show = show_test_oa_id; > > + i915->perf.oa.test_config.sysfs_metric_id.attr.name = "id"; > > + i915->perf.oa.test_config.sysfs_metric_id.attr.mode = 0444; > > + i915->perf.oa.test_config.sysfs_metric_id.show = show_test_oa_id; > > } > > diff --git a/drivers/gpu/drm/i915/i915_oa_chv.h b/drivers/gpu/drm/i915/i915_oa_chv.h > > index 0986eae3135f..4bb3b6be7bd2 100644 > > --- a/drivers/gpu/drm/i915/i915_oa_chv.h > > +++ b/drivers/gpu/drm/i915/i915_oa_chv.h > > @@ -10,6 +10,6 @@ > > #ifndef __I915_OA_CHV_H__ > > #define __I915_OA_CHV_H__ > > -extern void i915_perf_load_test_config_chv(struct drm_i915_private *dev_priv); > > +extern void i915_perf_load_test_config_chv(struct drm_i915_private *i915); > > #endif > > diff --git a/drivers/gpu/drm/i915/i915_oa_cnl.c b/drivers/gpu/drm/i915/i915_oa_cnl.c > > index 5c23d883d6c9..fcf08b9e7c73 100644 > > --- a/drivers/gpu/drm/i915/i915_oa_cnl.c > > +++ b/drivers/gpu/drm/i915/i915_oa_cnl.c > > @@ -75,28 +75,28 @@ show_test_oa_id(struct device *kdev, struct device_attribute *attr, char *buf) > > } > > void > > -i915_perf_load_test_config_cnl(struct drm_i915_private *dev_priv) > > +i915_perf_load_test_config_cnl(struct drm_i915_private *i915) > > { > > - strlcpy(dev_priv->perf.oa.test_config.uuid, > > + strlcpy(i915->perf.oa.test_config.uuid, > > "db41edd4-d8e7-4730-ad11-b9a2d6833503", > > - sizeof(dev_priv->perf.oa.test_config.uuid)); > > - dev_priv->perf.oa.test_config.id = 1; > > + sizeof(i915->perf.oa.test_config.uuid)); > > + i915->perf.oa.test_config.id = 1; > > - dev_priv->perf.oa.test_config.mux_regs = mux_config_test_oa; > > - dev_priv->perf.oa.test_config.mux_regs_len = ARRAY_SIZE(mux_config_test_oa); > > + i915->perf.oa.test_config.mux_regs = mux_config_test_oa; > > + i915->perf.oa.test_config.mux_regs_len = ARRAY_SIZE(mux_config_test_oa); > > - dev_priv->perf.oa.test_config.b_counter_regs = b_counter_config_test_oa; > > - dev_priv->perf.oa.test_config.b_counter_regs_len = ARRAY_SIZE(b_counter_config_test_oa); > > + i915->perf.oa.test_config.b_counter_regs = b_counter_config_test_oa; > > + i915->perf.oa.test_config.b_counter_regs_len = ARRAY_SIZE(b_counter_config_test_oa); > > - dev_priv->perf.oa.test_config.flex_regs = flex_eu_config_test_oa; > > - dev_priv->perf.oa.test_config.flex_regs_len = ARRAY_SIZE(flex_eu_config_test_oa); > > + i915->perf.oa.test_config.flex_regs = flex_eu_config_test_oa; > > + i915->perf.oa.test_config.flex_regs_len = ARRAY_SIZE(flex_eu_config_test_oa); > > - dev_priv->perf.oa.test_config.sysfs_metric.name = "db41edd4-d8e7-4730-ad11-b9a2d6833503"; > > - dev_priv->perf.oa.test_config.sysfs_metric.attrs = dev_priv->perf.oa.test_config.attrs; > > + i915->perf.oa.test_config.sysfs_metric.name = "db41edd4-d8e7-4730-ad11-b9a2d6833503"; > > + i915->perf.oa.test_config.sysfs_metric.attrs = i915->perf.oa.test_config.attrs; > > - dev_priv->perf.oa.test_config.attrs[0] = &dev_priv->perf.oa.test_config.sysfs_metric_id.attr; > > + i915->perf.oa.test_config.attrs[0] = &i915->perf.oa.test_config.sysfs_metric_id.attr; > > - dev_priv->perf.oa.test_config.sysfs_metric_id.attr.name = "id"; > > - dev_priv->perf.oa.test_config.sysfs_metric_id.attr.mode = 0444; > > - dev_priv->perf.oa.test_config.sysfs_metric_id.show = show_test_oa_id; > > + i915->perf.oa.test_config.sysfs_metric_id.attr.name = "id"; > > + i915->perf.oa.test_config.sysfs_metric_id.attr.mode = 0444; > > + i915->perf.oa.test_config.sysfs_metric_id.show = show_test_oa_id; > > } > > diff --git a/drivers/gpu/drm/i915/i915_oa_cnl.h b/drivers/gpu/drm/i915/i915_oa_cnl.h > > index e830a406aff2..c70101655049 100644 > > --- a/drivers/gpu/drm/i915/i915_oa_cnl.h > > +++ b/drivers/gpu/drm/i915/i915_oa_cnl.h > > @@ -10,6 +10,6 @@ > > #ifndef __I915_OA_CNL_H__ > > #define __I915_OA_CNL_H__ > > -extern void i915_perf_load_test_config_cnl(struct drm_i915_private *dev_priv); > > +extern void i915_perf_load_test_config_cnl(struct drm_i915_private *i915); > > #endif > > diff --git a/drivers/gpu/drm/i915/i915_oa_glk.c b/drivers/gpu/drm/i915/i915_oa_glk.c > > index 4bdda66df7d2..fdb38b9ea9bb 100644 > > --- a/drivers/gpu/drm/i915/i915_oa_glk.c > > +++ b/drivers/gpu/drm/i915/i915_oa_glk.c > > @@ -62,28 +62,28 @@ show_test_oa_id(struct device *kdev, struct device_attribute *attr, char *buf) > > } > > void > > -i915_perf_load_test_config_glk(struct drm_i915_private *dev_priv) > > +i915_perf_load_test_config_glk(struct drm_i915_private *i915) > > { > > - strlcpy(dev_priv->perf.oa.test_config.uuid, > > + strlcpy(i915->perf.oa.test_config.uuid, > > "dd3fd789-e783-4204-8cd0-b671bbccb0cf", > > - sizeof(dev_priv->perf.oa.test_config.uuid)); > > - dev_priv->perf.oa.test_config.id = 1; > > + sizeof(i915->perf.oa.test_config.uuid)); > > + i915->perf.oa.test_config.id = 1; > > - dev_priv->perf.oa.test_config.mux_regs = mux_config_test_oa; > > - dev_priv->perf.oa.test_config.mux_regs_len = ARRAY_SIZE(mux_config_test_oa); > > + i915->perf.oa.test_config.mux_regs = mux_config_test_oa; > > + i915->perf.oa.test_config.mux_regs_len = ARRAY_SIZE(mux_config_test_oa); > > - dev_priv->perf.oa.test_config.b_counter_regs = b_counter_config_test_oa; > > - dev_priv->perf.oa.test_config.b_counter_regs_len = ARRAY_SIZE(b_counter_config_test_oa); > > + i915->perf.oa.test_config.b_counter_regs = b_counter_config_test_oa; > > + i915->perf.oa.test_config.b_counter_regs_len = ARRAY_SIZE(b_counter_config_test_oa); > > - dev_priv->perf.oa.test_config.flex_regs = flex_eu_config_test_oa; > > - dev_priv->perf.oa.test_config.flex_regs_len = ARRAY_SIZE(flex_eu_config_test_oa); > > + i915->perf.oa.test_config.flex_regs = flex_eu_config_test_oa; > > + i915->perf.oa.test_config.flex_regs_len = ARRAY_SIZE(flex_eu_config_test_oa); > > - dev_priv->perf.oa.test_config.sysfs_metric.name = "dd3fd789-e783-4204-8cd0-b671bbccb0cf"; > > - dev_priv->perf.oa.test_config.sysfs_metric.attrs = dev_priv->perf.oa.test_config.attrs; > > + i915->perf.oa.test_config.sysfs_metric.name = "dd3fd789-e783-4204-8cd0-b671bbccb0cf"; > > + i915->perf.oa.test_config.sysfs_metric.attrs = i915->perf.oa.test_config.attrs; > > - dev_priv->perf.oa.test_config.attrs[0] = &dev_priv->perf.oa.test_config.sysfs_metric_id.attr; > > + i915->perf.oa.test_config.attrs[0] = &i915->perf.oa.test_config.sysfs_metric_id.attr; > > - dev_priv->perf.oa.test_config.sysfs_metric_id.attr.name = "id"; > > - dev_priv->perf.oa.test_config.sysfs_metric_id.attr.mode = 0444; > > - dev_priv->perf.oa.test_config.sysfs_metric_id.show = show_test_oa_id; > > + i915->perf.oa.test_config.sysfs_metric_id.attr.name = "id"; > > + i915->perf.oa.test_config.sysfs_metric_id.attr.mode = 0444; > > + i915->perf.oa.test_config.sysfs_metric_id.show = show_test_oa_id; > > } > > diff --git a/drivers/gpu/drm/i915/i915_oa_glk.h b/drivers/gpu/drm/i915/i915_oa_glk.h > > index 06dedf991edb..5d07986827c8 100644 > > --- a/drivers/gpu/drm/i915/i915_oa_glk.h > > +++ b/drivers/gpu/drm/i915/i915_oa_glk.h > > @@ -10,6 +10,6 @@ > > #ifndef __I915_OA_GLK_H__ > > #define __I915_OA_GLK_H__ > > -extern void i915_perf_load_test_config_glk(struct drm_i915_private *dev_priv); > > +extern void i915_perf_load_test_config_glk(struct drm_i915_private *i915); > > #endif > > diff --git a/drivers/gpu/drm/i915/i915_oa_hsw.c b/drivers/gpu/drm/i915/i915_oa_hsw.c > > index cc6526fdd2bd..640301c9fb9d 100644 > > --- a/drivers/gpu/drm/i915/i915_oa_hsw.c > > +++ b/drivers/gpu/drm/i915/i915_oa_hsw.c > > @@ -92,28 +92,28 @@ show_render_basic_id(struct device *kdev, struct device_attribute *attr, char *b > > } > > void > > -i915_perf_load_test_config_hsw(struct drm_i915_private *dev_priv) > > +i915_perf_load_test_config_hsw(struct drm_i915_private *i915) > > { > > - strlcpy(dev_priv->perf.oa.test_config.uuid, > > + strlcpy(i915->perf.oa.test_config.uuid, > > "403d8832-1a27-4aa6-a64e-f5389ce7b212", > > - sizeof(dev_priv->perf.oa.test_config.uuid)); > > - dev_priv->perf.oa.test_config.id = 1; > > + sizeof(i915->perf.oa.test_config.uuid)); > > + i915->perf.oa.test_config.id = 1; > > - dev_priv->perf.oa.test_config.mux_regs = mux_config_render_basic; > > - dev_priv->perf.oa.test_config.mux_regs_len = ARRAY_SIZE(mux_config_render_basic); > > + i915->perf.oa.test_config.mux_regs = mux_config_render_basic; > > + i915->perf.oa.test_config.mux_regs_len = ARRAY_SIZE(mux_config_render_basic); > > - dev_priv->perf.oa.test_config.b_counter_regs = b_counter_config_render_basic; > > - dev_priv->perf.oa.test_config.b_counter_regs_len = ARRAY_SIZE(b_counter_config_render_basic); > > + i915->perf.oa.test_config.b_counter_regs = b_counter_config_render_basic; > > + i915->perf.oa.test_config.b_counter_regs_len = ARRAY_SIZE(b_counter_config_render_basic); > > - dev_priv->perf.oa.test_config.flex_regs = flex_eu_config_render_basic; > > - dev_priv->perf.oa.test_config.flex_regs_len = ARRAY_SIZE(flex_eu_config_render_basic); > > + i915->perf.oa.test_config.flex_regs = flex_eu_config_render_basic; > > + i915->perf.oa.test_config.flex_regs_len = ARRAY_SIZE(flex_eu_config_render_basic); > > - dev_priv->perf.oa.test_config.sysfs_metric.name = "403d8832-1a27-4aa6-a64e-f5389ce7b212"; > > - dev_priv->perf.oa.test_config.sysfs_metric.attrs = dev_priv->perf.oa.test_config.attrs; > > + i915->perf.oa.test_config.sysfs_metric.name = "403d8832-1a27-4aa6-a64e-f5389ce7b212"; > > + i915->perf.oa.test_config.sysfs_metric.attrs = i915->perf.oa.test_config.attrs; > > - dev_priv->perf.oa.test_config.attrs[0] = &dev_priv->perf.oa.test_config.sysfs_metric_id.attr; > > + i915->perf.oa.test_config.attrs[0] = &i915->perf.oa.test_config.sysfs_metric_id.attr; > > - dev_priv->perf.oa.test_config.sysfs_metric_id.attr.name = "id"; > > - dev_priv->perf.oa.test_config.sysfs_metric_id.attr.mode = 0444; > > - dev_priv->perf.oa.test_config.sysfs_metric_id.show = show_render_basic_id; > > + i915->perf.oa.test_config.sysfs_metric_id.attr.name = "id"; > > + i915->perf.oa.test_config.sysfs_metric_id.attr.mode = 0444; > > + i915->perf.oa.test_config.sysfs_metric_id.show = show_render_basic_id; > > } > > diff --git a/drivers/gpu/drm/i915/i915_oa_hsw.h b/drivers/gpu/drm/i915/i915_oa_hsw.h > > index 3d0c870cd0bd..7812a9df8647 100644 > > --- a/drivers/gpu/drm/i915/i915_oa_hsw.h > > +++ b/drivers/gpu/drm/i915/i915_oa_hsw.h > > @@ -10,6 +10,6 @@ > > #ifndef __I915_OA_HSW_H__ > > #define __I915_OA_HSW_H__ > > -extern void i915_perf_load_test_config_hsw(struct drm_i915_private *dev_priv); > > +extern void i915_perf_load_test_config_hsw(struct drm_i915_private *i915); > > #endif > > diff --git a/drivers/gpu/drm/i915/i915_oa_icl.c b/drivers/gpu/drm/i915/i915_oa_icl.c > > index baa51427a543..2a8c6de619ba 100644 > > --- a/drivers/gpu/drm/i915/i915_oa_icl.c > > +++ b/drivers/gpu/drm/i915/i915_oa_icl.c > > @@ -72,28 +72,28 @@ show_test_oa_id(struct device *kdev, struct device_attribute *attr, char *buf) > > } > > void > > -i915_perf_load_test_config_icl(struct drm_i915_private *dev_priv) > > +i915_perf_load_test_config_icl(struct drm_i915_private *i915) > > { > > - strlcpy(dev_priv->perf.oa.test_config.uuid, > > + strlcpy(i915->perf.oa.test_config.uuid, > > "a291665e-244b-4b76-9b9a-01de9d3c8068", > > - sizeof(dev_priv->perf.oa.test_config.uuid)); > > - dev_priv->perf.oa.test_config.id = 1; > > + sizeof(i915->perf.oa.test_config.uuid)); > > + i915->perf.oa.test_config.id = 1; > > - dev_priv->perf.oa.test_config.mux_regs = mux_config_test_oa; > > - dev_priv->perf.oa.test_config.mux_regs_len = ARRAY_SIZE(mux_config_test_oa); > > + i915->perf.oa.test_config.mux_regs = mux_config_test_oa; > > + i915->perf.oa.test_config.mux_regs_len = ARRAY_SIZE(mux_config_test_oa); > > - dev_priv->perf.oa.test_config.b_counter_regs = b_counter_config_test_oa; > > - dev_priv->perf.oa.test_config.b_counter_regs_len = ARRAY_SIZE(b_counter_config_test_oa); > > + i915->perf.oa.test_config.b_counter_regs = b_counter_config_test_oa; > > + i915->perf.oa.test_config.b_counter_regs_len = ARRAY_SIZE(b_counter_config_test_oa); > > - dev_priv->perf.oa.test_config.flex_regs = flex_eu_config_test_oa; > > - dev_priv->perf.oa.test_config.flex_regs_len = ARRAY_SIZE(flex_eu_config_test_oa); > > + i915->perf.oa.test_config.flex_regs = flex_eu_config_test_oa; > > + i915->perf.oa.test_config.flex_regs_len = ARRAY_SIZE(flex_eu_config_test_oa); > > - dev_priv->perf.oa.test_config.sysfs_metric.name = "a291665e-244b-4b76-9b9a-01de9d3c8068"; > > - dev_priv->perf.oa.test_config.sysfs_metric.attrs = dev_priv->perf.oa.test_config.attrs; > > + i915->perf.oa.test_config.sysfs_metric.name = "a291665e-244b-4b76-9b9a-01de9d3c8068"; > > + i915->perf.oa.test_config.sysfs_metric.attrs = i915->perf.oa.test_config.attrs; > > - dev_priv->perf.oa.test_config.attrs[0] = &dev_priv->perf.oa.test_config.sysfs_metric_id.attr; > > + i915->perf.oa.test_config.attrs[0] = &i915->perf.oa.test_config.sysfs_metric_id.attr; > > - dev_priv->perf.oa.test_config.sysfs_metric_id.attr.name = "id"; > > - dev_priv->perf.oa.test_config.sysfs_metric_id.attr.mode = 0444; > > - dev_priv->perf.oa.test_config.sysfs_metric_id.show = show_test_oa_id; > > + i915->perf.oa.test_config.sysfs_metric_id.attr.name = "id"; > > + i915->perf.oa.test_config.sysfs_metric_id.attr.mode = 0444; > > + i915->perf.oa.test_config.sysfs_metric_id.show = show_test_oa_id; > > } > > diff --git a/drivers/gpu/drm/i915/i915_oa_icl.h b/drivers/gpu/drm/i915/i915_oa_icl.h > > index 24eaa97d61ba..a7c50d31e678 100644 > > --- a/drivers/gpu/drm/i915/i915_oa_icl.h > > +++ b/drivers/gpu/drm/i915/i915_oa_icl.h > > @@ -10,6 +10,6 @@ > > #ifndef __I915_OA_ICL_H__ > > #define __I915_OA_ICL_H__ > > -extern void i915_perf_load_test_config_icl(struct drm_i915_private *dev_priv); > > +extern void i915_perf_load_test_config_icl(struct drm_i915_private *i915); > > #endif > > diff --git a/drivers/gpu/drm/i915/i915_oa_kblgt2.c b/drivers/gpu/drm/i915/i915_oa_kblgt2.c > > index 168e49ab0d4d..975311df3d62 100644 > > --- a/drivers/gpu/drm/i915/i915_oa_kblgt2.c > > +++ b/drivers/gpu/drm/i915/i915_oa_kblgt2.c > > @@ -63,28 +63,28 @@ show_test_oa_id(struct device *kdev, struct device_attribute *attr, char *buf) > > } > > void > > -i915_perf_load_test_config_kblgt2(struct drm_i915_private *dev_priv) > > +i915_perf_load_test_config_kblgt2(struct drm_i915_private *i915) > > { > > - strlcpy(dev_priv->perf.oa.test_config.uuid, > > + strlcpy(i915->perf.oa.test_config.uuid, > > "baa3c7e4-52b6-4b85-801e-465a94b746dd", > > - sizeof(dev_priv->perf.oa.test_config.uuid)); > > - dev_priv->perf.oa.test_config.id = 1; > > + sizeof(i915->perf.oa.test_config.uuid)); > > + i915->perf.oa.test_config.id = 1; > > - dev_priv->perf.oa.test_config.mux_regs = mux_config_test_oa; > > - dev_priv->perf.oa.test_config.mux_regs_len = ARRAY_SIZE(mux_config_test_oa); > > + i915->perf.oa.test_config.mux_regs = mux_config_test_oa; > > + i915->perf.oa.test_config.mux_regs_len = ARRAY_SIZE(mux_config_test_oa); > > - dev_priv->perf.oa.test_config.b_counter_regs = b_counter_config_test_oa; > > - dev_priv->perf.oa.test_config.b_counter_regs_len = ARRAY_SIZE(b_counter_config_test_oa); > > + i915->perf.oa.test_config.b_counter_regs = b_counter_config_test_oa; > > + i915->perf.oa.test_config.b_counter_regs_len = ARRAY_SIZE(b_counter_config_test_oa); > > - dev_priv->perf.oa.test_config.flex_regs = flex_eu_config_test_oa; > > - dev_priv->perf.oa.test_config.flex_regs_len = ARRAY_SIZE(flex_eu_config_test_oa); > > + i915->perf.oa.test_config.flex_regs = flex_eu_config_test_oa; > > + i915->perf.oa.test_config.flex_regs_len = ARRAY_SIZE(flex_eu_config_test_oa); > > - dev_priv->perf.oa.test_config.sysfs_metric.name = "baa3c7e4-52b6-4b85-801e-465a94b746dd"; > > - dev_priv->perf.oa.test_config.sysfs_metric.attrs = dev_priv->perf.oa.test_config.attrs; > > + i915->perf.oa.test_config.sysfs_metric.name = "baa3c7e4-52b6-4b85-801e-465a94b746dd"; > > + i915->perf.oa.test_config.sysfs_metric.attrs = i915->perf.oa.test_config.attrs; > > - dev_priv->perf.oa.test_config.attrs[0] = &dev_priv->perf.oa.test_config.sysfs_metric_id.attr; > > + i915->perf.oa.test_config.attrs[0] = &i915->perf.oa.test_config.sysfs_metric_id.attr; > > - dev_priv->perf.oa.test_config.sysfs_metric_id.attr.name = "id"; > > - dev_priv->perf.oa.test_config.sysfs_metric_id.attr.mode = 0444; > > - dev_priv->perf.oa.test_config.sysfs_metric_id.show = show_test_oa_id; > > + i915->perf.oa.test_config.sysfs_metric_id.attr.name = "id"; > > + i915->perf.oa.test_config.sysfs_metric_id.attr.mode = 0444; > > + i915->perf.oa.test_config.sysfs_metric_id.show = show_test_oa_id; > > } > > diff --git a/drivers/gpu/drm/i915/i915_oa_kblgt2.h b/drivers/gpu/drm/i915/i915_oa_kblgt2.h > > index a55398a904de..53655de7b976 100644 > > --- a/drivers/gpu/drm/i915/i915_oa_kblgt2.h > > +++ b/drivers/gpu/drm/i915/i915_oa_kblgt2.h > > @@ -10,6 +10,6 @@ > > #ifndef __I915_OA_KBLGT2_H__ > > #define __I915_OA_KBLGT2_H__ > > -extern void i915_perf_load_test_config_kblgt2(struct drm_i915_private *dev_priv); > > +extern void i915_perf_load_test_config_kblgt2(struct drm_i915_private *i915); > > #endif > > diff --git a/drivers/gpu/drm/i915/i915_oa_kblgt3.c b/drivers/gpu/drm/i915/i915_oa_kblgt3.c > > index 6ffa553c388e..ed39f8dd9af2 100644 > > --- a/drivers/gpu/drm/i915/i915_oa_kblgt3.c > > +++ b/drivers/gpu/drm/i915/i915_oa_kblgt3.c > > @@ -63,28 +63,28 @@ show_test_oa_id(struct device *kdev, struct device_attribute *attr, char *buf) > > } > > void > > -i915_perf_load_test_config_kblgt3(struct drm_i915_private *dev_priv) > > +i915_perf_load_test_config_kblgt3(struct drm_i915_private *i915) > > { > > - strlcpy(dev_priv->perf.oa.test_config.uuid, > > + strlcpy(i915->perf.oa.test_config.uuid, > > "f1792f32-6db2-4b50-b4b2-557128f1688d", > > - sizeof(dev_priv->perf.oa.test_config.uuid)); > > - dev_priv->perf.oa.test_config.id = 1; > > + sizeof(i915->perf.oa.test_config.uuid)); > > + i915->perf.oa.test_config.id = 1; > > - dev_priv->perf.oa.test_config.mux_regs = mux_config_test_oa; > > - dev_priv->perf.oa.test_config.mux_regs_len = ARRAY_SIZE(mux_config_test_oa); > > + i915->perf.oa.test_config.mux_regs = mux_config_test_oa; > > + i915->perf.oa.test_config.mux_regs_len = ARRAY_SIZE(mux_config_test_oa); > > - dev_priv->perf.oa.test_config.b_counter_regs = b_counter_config_test_oa; > > - dev_priv->perf.oa.test_config.b_counter_regs_len = ARRAY_SIZE(b_counter_config_test_oa); > > + i915->perf.oa.test_config.b_counter_regs = b_counter_config_test_oa; > > + i915->perf.oa.test_config.b_counter_regs_len = ARRAY_SIZE(b_counter_config_test_oa); > > - dev_priv->perf.oa.test_config.flex_regs = flex_eu_config_test_oa; > > - dev_priv->perf.oa.test_config.flex_regs_len = ARRAY_SIZE(flex_eu_config_test_oa); > > + i915->perf.oa.test_config.flex_regs = flex_eu_config_test_oa; > > + i915->perf.oa.test_config.flex_regs_len = ARRAY_SIZE(flex_eu_config_test_oa); > > - dev_priv->perf.oa.test_config.sysfs_metric.name = "f1792f32-6db2-4b50-b4b2-557128f1688d"; > > - dev_priv->perf.oa.test_config.sysfs_metric.attrs = dev_priv->perf.oa.test_config.attrs; > > + i915->perf.oa.test_config.sysfs_metric.name = "f1792f32-6db2-4b50-b4b2-557128f1688d"; > > + i915->perf.oa.test_config.sysfs_metric.attrs = i915->perf.oa.test_config.attrs; > > - dev_priv->perf.oa.test_config.attrs[0] = &dev_priv->perf.oa.test_config.sysfs_metric_id.attr; > > + i915->perf.oa.test_config.attrs[0] = &i915->perf.oa.test_config.sysfs_metric_id.attr; > > - dev_priv->perf.oa.test_config.sysfs_metric_id.attr.name = "id"; > > - dev_priv->perf.oa.test_config.sysfs_metric_id.attr.mode = 0444; > > - dev_priv->perf.oa.test_config.sysfs_metric_id.show = show_test_oa_id; > > + i915->perf.oa.test_config.sysfs_metric_id.attr.name = "id"; > > + i915->perf.oa.test_config.sysfs_metric_id.attr.mode = 0444; > > + i915->perf.oa.test_config.sysfs_metric_id.show = show_test_oa_id; > > } > > diff --git a/drivers/gpu/drm/i915/i915_oa_kblgt3.h b/drivers/gpu/drm/i915/i915_oa_kblgt3.h > > index 3ddd3483b7cc..60ebe2b20ada 100644 > > --- a/drivers/gpu/drm/i915/i915_oa_kblgt3.h > > +++ b/drivers/gpu/drm/i915/i915_oa_kblgt3.h > > @@ -10,6 +10,6 @@ > > #ifndef __I915_OA_KBLGT3_H__ > > #define __I915_OA_KBLGT3_H__ > > -extern void i915_perf_load_test_config_kblgt3(struct drm_i915_private *dev_priv); > > +extern void i915_perf_load_test_config_kblgt3(struct drm_i915_private *i915); > > #endif > > diff --git a/drivers/gpu/drm/i915/i915_oa_sklgt2.c b/drivers/gpu/drm/i915/i915_oa_sklgt2.c > > index 7ce6ee851d43..60b4dbcd4c59 100644 > > --- a/drivers/gpu/drm/i915/i915_oa_sklgt2.c > > +++ b/drivers/gpu/drm/i915/i915_oa_sklgt2.c > > @@ -62,28 +62,28 @@ show_test_oa_id(struct device *kdev, struct device_attribute *attr, char *buf) > > } > > void > > -i915_perf_load_test_config_sklgt2(struct drm_i915_private *dev_priv) > > +i915_perf_load_test_config_sklgt2(struct drm_i915_private *i915) > > { > > - strlcpy(dev_priv->perf.oa.test_config.uuid, > > + strlcpy(i915->perf.oa.test_config.uuid, > > "1651949f-0ac0-4cb1-a06f-dafd74a407d1", > > - sizeof(dev_priv->perf.oa.test_config.uuid)); > > - dev_priv->perf.oa.test_config.id = 1; > > + sizeof(i915->perf.oa.test_config.uuid)); > > + i915->perf.oa.test_config.id = 1; > > - dev_priv->perf.oa.test_config.mux_regs = mux_config_test_oa; > > - dev_priv->perf.oa.test_config.mux_regs_len = ARRAY_SIZE(mux_config_test_oa); > > + i915->perf.oa.test_config.mux_regs = mux_config_test_oa; > > + i915->perf.oa.test_config.mux_regs_len = ARRAY_SIZE(mux_config_test_oa); > > - dev_priv->perf.oa.test_config.b_counter_regs = b_counter_config_test_oa; > > - dev_priv->perf.oa.test_config.b_counter_regs_len = ARRAY_SIZE(b_counter_config_test_oa); > > + i915->perf.oa.test_config.b_counter_regs = b_counter_config_test_oa; > > + i915->perf.oa.test_config.b_counter_regs_len = ARRAY_SIZE(b_counter_config_test_oa); > > - dev_priv->perf.oa.test_config.flex_regs = flex_eu_config_test_oa; > > - dev_priv->perf.oa.test_config.flex_regs_len = ARRAY_SIZE(flex_eu_config_test_oa); > > + i915->perf.oa.test_config.flex_regs = flex_eu_config_test_oa; > > + i915->perf.oa.test_config.flex_regs_len = ARRAY_SIZE(flex_eu_config_test_oa); > > - dev_priv->perf.oa.test_config.sysfs_metric.name = "1651949f-0ac0-4cb1-a06f-dafd74a407d1"; > > - dev_priv->perf.oa.test_config.sysfs_metric.attrs = dev_priv->perf.oa.test_config.attrs; > > + i915->perf.oa.test_config.sysfs_metric.name = "1651949f-0ac0-4cb1-a06f-dafd74a407d1"; > > + i915->perf.oa.test_config.sysfs_metric.attrs = i915->perf.oa.test_config.attrs; > > - dev_priv->perf.oa.test_config.attrs[0] = &dev_priv->perf.oa.test_config.sysfs_metric_id.attr; > > + i915->perf.oa.test_config.attrs[0] = &i915->perf.oa.test_config.sysfs_metric_id.attr; > > - dev_priv->perf.oa.test_config.sysfs_metric_id.attr.name = "id"; > > - dev_priv->perf.oa.test_config.sysfs_metric_id.attr.mode = 0444; > > - dev_priv->perf.oa.test_config.sysfs_metric_id.show = show_test_oa_id; > > + i915->perf.oa.test_config.sysfs_metric_id.attr.name = "id"; > > + i915->perf.oa.test_config.sysfs_metric_id.attr.mode = 0444; > > + i915->perf.oa.test_config.sysfs_metric_id.show = show_test_oa_id; > > } > > diff --git a/drivers/gpu/drm/i915/i915_oa_sklgt2.h b/drivers/gpu/drm/i915/i915_oa_sklgt2.h > > index be6256037239..ee6f747502ea 100644 > > --- a/drivers/gpu/drm/i915/i915_oa_sklgt2.h > > +++ b/drivers/gpu/drm/i915/i915_oa_sklgt2.h > > @@ -10,6 +10,6 @@ > > #ifndef __I915_OA_SKLGT2_H__ > > #define __I915_OA_SKLGT2_H__ > > -extern void i915_perf_load_test_config_sklgt2(struct drm_i915_private *dev_priv); > > +extern void i915_perf_load_test_config_sklgt2(struct drm_i915_private *i915); > > #endif > > diff --git a/drivers/gpu/drm/i915/i915_oa_sklgt3.c b/drivers/gpu/drm/i915/i915_oa_sklgt3.c > > index 086ca2631e1c..8c23cf712a7a 100644 > > --- a/drivers/gpu/drm/i915/i915_oa_sklgt3.c > > +++ b/drivers/gpu/drm/i915/i915_oa_sklgt3.c > > @@ -63,28 +63,28 @@ show_test_oa_id(struct device *kdev, struct device_attribute *attr, char *buf) > > } > > void > > -i915_perf_load_test_config_sklgt3(struct drm_i915_private *dev_priv) > > +i915_perf_load_test_config_sklgt3(struct drm_i915_private *i915) > > { > > - strlcpy(dev_priv->perf.oa.test_config.uuid, > > + strlcpy(i915->perf.oa.test_config.uuid, > > "2b985803-d3c9-4629-8a4f-634bfecba0e8", > > - sizeof(dev_priv->perf.oa.test_config.uuid)); > > - dev_priv->perf.oa.test_config.id = 1; > > + sizeof(i915->perf.oa.test_config.uuid)); > > + i915->perf.oa.test_config.id = 1; > > - dev_priv->perf.oa.test_config.mux_regs = mux_config_test_oa; > > - dev_priv->perf.oa.test_config.mux_regs_len = ARRAY_SIZE(mux_config_test_oa); > > + i915->perf.oa.test_config.mux_regs = mux_config_test_oa; > > + i915->perf.oa.test_config.mux_regs_len = ARRAY_SIZE(mux_config_test_oa); > > - dev_priv->perf.oa.test_config.b_counter_regs = b_counter_config_test_oa; > > - dev_priv->perf.oa.test_config.b_counter_regs_len = ARRAY_SIZE(b_counter_config_test_oa); > > + i915->perf.oa.test_config.b_counter_regs = b_counter_config_test_oa; > > + i915->perf.oa.test_config.b_counter_regs_len = ARRAY_SIZE(b_counter_config_test_oa); > > - dev_priv->perf.oa.test_config.flex_regs = flex_eu_config_test_oa; > > - dev_priv->perf.oa.test_config.flex_regs_len = ARRAY_SIZE(flex_eu_config_test_oa); > > + i915->perf.oa.test_config.flex_regs = flex_eu_config_test_oa; > > + i915->perf.oa.test_config.flex_regs_len = ARRAY_SIZE(flex_eu_config_test_oa); > > - dev_priv->perf.oa.test_config.sysfs_metric.name = "2b985803-d3c9-4629-8a4f-634bfecba0e8"; > > - dev_priv->perf.oa.test_config.sysfs_metric.attrs = dev_priv->perf.oa.test_config.attrs; > > + i915->perf.oa.test_config.sysfs_metric.name = "2b985803-d3c9-4629-8a4f-634bfecba0e8"; > > + i915->perf.oa.test_config.sysfs_metric.attrs = i915->perf.oa.test_config.attrs; > > - dev_priv->perf.oa.test_config.attrs[0] = &dev_priv->perf.oa.test_config.sysfs_metric_id.attr; > > + i915->perf.oa.test_config.attrs[0] = &i915->perf.oa.test_config.sysfs_metric_id.attr; > > - dev_priv->perf.oa.test_config.sysfs_metric_id.attr.name = "id"; > > - dev_priv->perf.oa.test_config.sysfs_metric_id.attr.mode = 0444; > > - dev_priv->perf.oa.test_config.sysfs_metric_id.show = show_test_oa_id; > > + i915->perf.oa.test_config.sysfs_metric_id.attr.name = "id"; > > + i915->perf.oa.test_config.sysfs_metric_id.attr.mode = 0444; > > + i915->perf.oa.test_config.sysfs_metric_id.show = show_test_oa_id; > > } > > diff --git a/drivers/gpu/drm/i915/i915_oa_sklgt3.h b/drivers/gpu/drm/i915/i915_oa_sklgt3.h > > index 650beb068e56..35ccb4125d71 100644 > > --- a/drivers/gpu/drm/i915/i915_oa_sklgt3.h > > +++ b/drivers/gpu/drm/i915/i915_oa_sklgt3.h > > @@ -10,6 +10,6 @@ > > #ifndef __I915_OA_SKLGT3_H__ > > #define __I915_OA_SKLGT3_H__ > > -extern void i915_perf_load_test_config_sklgt3(struct drm_i915_private *dev_priv); > > +extern void i915_perf_load_test_config_sklgt3(struct drm_i915_private *i915); > > #endif > > diff --git a/drivers/gpu/drm/i915/i915_oa_sklgt4.c b/drivers/gpu/drm/i915/i915_oa_sklgt4.c > > index b291a6eb8a87..b846da6198f8 100644 > > --- a/drivers/gpu/drm/i915/i915_oa_sklgt4.c > > +++ b/drivers/gpu/drm/i915/i915_oa_sklgt4.c > > @@ -63,28 +63,28 @@ show_test_oa_id(struct device *kdev, struct device_attribute *attr, char *buf) > > } > > void > > -i915_perf_load_test_config_sklgt4(struct drm_i915_private *dev_priv) > > +i915_perf_load_test_config_sklgt4(struct drm_i915_private *i915) > > { > > - strlcpy(dev_priv->perf.oa.test_config.uuid, > > + strlcpy(i915->perf.oa.test_config.uuid, > > "882fa433-1f4a-4a67-a962-c741888fe5f5", > > - sizeof(dev_priv->perf.oa.test_config.uuid)); > > - dev_priv->perf.oa.test_config.id = 1; > > + sizeof(i915->perf.oa.test_config.uuid)); > > + i915->perf.oa.test_config.id = 1; > > - dev_priv->perf.oa.test_config.mux_regs = mux_config_test_oa; > > - dev_priv->perf.oa.test_config.mux_regs_len = ARRAY_SIZE(mux_config_test_oa); > > + i915->perf.oa.test_config.mux_regs = mux_config_test_oa; > > + i915->perf.oa.test_config.mux_regs_len = ARRAY_SIZE(mux_config_test_oa); > > - dev_priv->perf.oa.test_config.b_counter_regs = b_counter_config_test_oa; > > - dev_priv->perf.oa.test_config.b_counter_regs_len = ARRAY_SIZE(b_counter_config_test_oa); > > + i915->perf.oa.test_config.b_counter_regs = b_counter_config_test_oa; > > + i915->perf.oa.test_config.b_counter_regs_len = ARRAY_SIZE(b_counter_config_test_oa); > > - dev_priv->perf.oa.test_config.flex_regs = flex_eu_config_test_oa; > > - dev_priv->perf.oa.test_config.flex_regs_len = ARRAY_SIZE(flex_eu_config_test_oa); > > + i915->perf.oa.test_config.flex_regs = flex_eu_config_test_oa; > > + i915->perf.oa.test_config.flex_regs_len = ARRAY_SIZE(flex_eu_config_test_oa); > > - dev_priv->perf.oa.test_config.sysfs_metric.name = "882fa433-1f4a-4a67-a962-c741888fe5f5"; > > - dev_priv->perf.oa.test_config.sysfs_metric.attrs = dev_priv->perf.oa.test_config.attrs; > > + i915->perf.oa.test_config.sysfs_metric.name = "882fa433-1f4a-4a67-a962-c741888fe5f5"; > > + i915->perf.oa.test_config.sysfs_metric.attrs = i915->perf.oa.test_config.attrs; > > - dev_priv->perf.oa.test_config.attrs[0] = &dev_priv->perf.oa.test_config.sysfs_metric_id.attr; > > + i915->perf.oa.test_config.attrs[0] = &i915->perf.oa.test_config.sysfs_metric_id.attr; > > - dev_priv->perf.oa.test_config.sysfs_metric_id.attr.name = "id"; > > - dev_priv->perf.oa.test_config.sysfs_metric_id.attr.mode = 0444; > > - dev_priv->perf.oa.test_config.sysfs_metric_id.show = show_test_oa_id; > > + i915->perf.oa.test_config.sysfs_metric_id.attr.name = "id"; > > + i915->perf.oa.test_config.sysfs_metric_id.attr.mode = 0444; > > + i915->perf.oa.test_config.sysfs_metric_id.show = show_test_oa_id; > > } > > diff --git a/drivers/gpu/drm/i915/i915_oa_sklgt4.h b/drivers/gpu/drm/i915/i915_oa_sklgt4.h > > index 8dcf849d131e..f3f6d7bf28e0 100644 > > --- a/drivers/gpu/drm/i915/i915_oa_sklgt4.h > > +++ b/drivers/gpu/drm/i915/i915_oa_sklgt4.h > > @@ -10,6 +10,6 @@ > > #ifndef __I915_OA_SKLGT4_H__ > > #define __I915_OA_SKLGT4_H__ > > -extern void i915_perf_load_test_config_sklgt4(struct drm_i915_private *dev_priv); > > +extern void i915_perf_load_test_config_sklgt4(struct drm_i915_private *i915); > > #endif > > diff --git a/drivers/gpu/drm/i915/i915_perf.c b/drivers/gpu/drm/i915/i915_perf.c > > index d92ddfada262..e0ef77ec4da3 100644 > > --- a/drivers/gpu/drm/i915/i915_perf.c > > +++ b/drivers/gpu/drm/i915/i915_perf.c > > @@ -366,7 +366,7 @@ struct perf_open_properties { > > int oa_period_exponent; > > }; > > -static void free_oa_config(struct drm_i915_private *dev_priv, > > +static void free_oa_config(struct drm_i915_private *i915, > > struct i915_oa_config *oa_config) > > { > > if (!PTR_ERR(oa_config->flex_regs)) > > @@ -378,48 +378,48 @@ static void free_oa_config(struct drm_i915_private *dev_priv, > > kfree(oa_config); > > } > > -static void put_oa_config(struct drm_i915_private *dev_priv, > > +static void put_oa_config(struct drm_i915_private *i915, > > struct i915_oa_config *oa_config) > > { > > if (!atomic_dec_and_test(&oa_config->ref_count)) > > return; > > - free_oa_config(dev_priv, oa_config); > > + free_oa_config(i915, oa_config); > > } > > -static int get_oa_config(struct drm_i915_private *dev_priv, > > +static int get_oa_config(struct drm_i915_private *i915, > > int metrics_set, > > struct i915_oa_config **out_config) > > { > > int ret; > > if (metrics_set == 1) { > > - *out_config = &dev_priv->perf.oa.test_config; > > - atomic_inc(&dev_priv->perf.oa.test_config.ref_count); > > + *out_config = &i915->perf.oa.test_config; > > + atomic_inc(&i915->perf.oa.test_config.ref_count); > > return 0; > > } > > - ret = mutex_lock_interruptible(&dev_priv->perf.metrics_lock); > > + ret = mutex_lock_interruptible(&i915->perf.metrics_lock); > > if (ret) > > return ret; > > - *out_config = idr_find(&dev_priv->perf.metrics_idr, metrics_set); > > + *out_config = idr_find(&i915->perf.metrics_idr, metrics_set); > > if (!*out_config) > > ret = -EINVAL; > > else > > atomic_inc(&(*out_config)->ref_count); > > - mutex_unlock(&dev_priv->perf.metrics_lock); > > + mutex_unlock(&i915->perf.metrics_lock); > > return ret; > > } > > -static u32 gen8_oa_hw_tail_read(struct drm_i915_private *dev_priv) > > +static u32 gen8_oa_hw_tail_read(struct drm_i915_private *i915) > > { > > return I915_READ(GEN8_OATAILPTR) & GEN8_OATAILPTR_MASK; > > } > > -static u32 gen7_oa_hw_tail_read(struct drm_i915_private *dev_priv) > > +static u32 gen7_oa_hw_tail_read(struct drm_i915_private *i915) > > { > > u32 oastatus1 = I915_READ(GEN7_OASTATUS1); > > @@ -428,7 +428,7 @@ static u32 gen7_oa_hw_tail_read(struct drm_i915_private *dev_priv) > > /** > > * oa_buffer_check_unlocked - check for data and update tail ptr state > > - * @dev_priv: i915 device instance > > + * @i915: i915 device instance > > * > > * This is either called via fops (for blocking reads in user ctx) or the poll > > * check hrtimer (atomic ctx) to check the OA buffer tail pointer and check > > @@ -450,9 +450,9 @@ static u32 gen7_oa_hw_tail_read(struct drm_i915_private *dev_priv) > > * > > * Returns: %true if the OA buffer contains data, else %false > > */ > > -static bool oa_buffer_check_unlocked(struct drm_i915_private *dev_priv) > > +static bool oa_buffer_check_unlocked(struct drm_i915_private *i915) > > { > > - int report_size = dev_priv->perf.oa.oa_buffer.format_size; > > + int report_size = i915->perf.oa.oa_buffer.format_size; > > unsigned long flags; > > unsigned int aged_idx; > > u32 head, hw_tail, aged_tail, aging_tail; > > @@ -462,19 +462,19 @@ static bool oa_buffer_check_unlocked(struct drm_i915_private *dev_priv) > > * could result in an OA buffer reset which might reset the head, > > * tails[] and aged_tail state. > > */ > > - spin_lock_irqsave(&dev_priv->perf.oa.oa_buffer.ptr_lock, flags); > > + spin_lock_irqsave(&i915->perf.oa.oa_buffer.ptr_lock, flags); > > /* NB: The head we observe here might effectively be a little out of > > * date (between head and tails[aged_idx].offset if there is currently > > * a read() in progress. > > */ > > - head = dev_priv->perf.oa.oa_buffer.head; > > + head = i915->perf.oa.oa_buffer.head; > > - aged_idx = dev_priv->perf.oa.oa_buffer.aged_tail_idx; > > - aged_tail = dev_priv->perf.oa.oa_buffer.tails[aged_idx].offset; > > - aging_tail = dev_priv->perf.oa.oa_buffer.tails[!aged_idx].offset; > > + aged_idx = i915->perf.oa.oa_buffer.aged_tail_idx; > > + aged_tail = i915->perf.oa.oa_buffer.tails[aged_idx].offset; > > + aging_tail = i915->perf.oa.oa_buffer.tails[!aged_idx].offset; > > - hw_tail = dev_priv->perf.oa.ops.oa_hw_tail_read(dev_priv); > > + hw_tail = i915->perf.oa.ops.oa_hw_tail_read(i915); > > /* The tail pointer increases in 64 byte increments, > > * not in report_size steps... > > @@ -494,16 +494,16 @@ static bool oa_buffer_check_unlocked(struct drm_i915_private *dev_priv) > > * available) without needing to wait for a later hrtimer callback. > > */ > > if (aging_tail != INVALID_TAIL_PTR && > > - ((now - dev_priv->perf.oa.oa_buffer.aging_timestamp) > > > + ((now - i915->perf.oa.oa_buffer.aging_timestamp) > > > OA_TAIL_MARGIN_NSEC)) { > > aged_idx ^= 1; > > - dev_priv->perf.oa.oa_buffer.aged_tail_idx = aged_idx; > > + i915->perf.oa.oa_buffer.aged_tail_idx = aged_idx; > > aged_tail = aging_tail; > > /* Mark that we need a new pointer to start aging... */ > > - dev_priv->perf.oa.oa_buffer.tails[!aged_idx].offset = INVALID_TAIL_PTR; > > + i915->perf.oa.oa_buffer.tails[!aged_idx].offset = INVALID_TAIL_PTR; > > aging_tail = INVALID_TAIL_PTR; > > } > > @@ -518,7 +518,7 @@ static bool oa_buffer_check_unlocked(struct drm_i915_private *dev_priv) > > if (aging_tail == INVALID_TAIL_PTR && > > (aged_tail == INVALID_TAIL_PTR || > > OA_TAKEN(hw_tail, aged_tail) >= report_size)) { > > - struct i915_vma *vma = dev_priv->perf.oa.oa_buffer.vma; > > + struct i915_vma *vma = i915->perf.oa.oa_buffer.vma; > > u32 gtt_offset = i915_ggtt_offset(vma); > > /* Be paranoid and do a bounds check on the pointer read back > > @@ -527,16 +527,16 @@ static bool oa_buffer_check_unlocked(struct drm_i915_private *dev_priv) > > */ > > if (hw_tail >= gtt_offset && > > hw_tail < (gtt_offset + OA_BUFFER_SIZE)) { > > - dev_priv->perf.oa.oa_buffer.tails[!aged_idx].offset = > > + i915->perf.oa.oa_buffer.tails[!aged_idx].offset = > > aging_tail = hw_tail; > > - dev_priv->perf.oa.oa_buffer.aging_timestamp = now; > > + i915->perf.oa.oa_buffer.aging_timestamp = now; > > } else { > > DRM_ERROR("Ignoring spurious out of range OA buffer tail pointer = %u\n", > > hw_tail); > > } > > } > > - spin_unlock_irqrestore(&dev_priv->perf.oa.oa_buffer.ptr_lock, flags); > > + spin_unlock_irqrestore(&i915->perf.oa.oa_buffer.ptr_lock, flags); > > return aged_tail == INVALID_TAIL_PTR ? > > false : OA_TAKEN(aged_tail, head) >= report_size; > > @@ -599,8 +599,8 @@ static int append_oa_sample(struct i915_perf_stream *stream, > > size_t *offset, > > const u8 *report) > > { > > - struct drm_i915_private *dev_priv = stream->dev_priv; > > - int report_size = dev_priv->perf.oa.oa_buffer.format_size; > > + struct drm_i915_private *i915 = stream->i915; > > + int report_size = i915->perf.oa.oa_buffer.format_size; > > struct drm_i915_perf_record_header header; > > u32 sample_flags = stream->sample_flags; > > @@ -651,10 +651,10 @@ static int gen8_append_oa_reports(struct i915_perf_stream *stream, > > size_t count, > > size_t *offset) > > { > > - struct drm_i915_private *dev_priv = stream->dev_priv; > > - int report_size = dev_priv->perf.oa.oa_buffer.format_size; > > - u8 *oa_buf_base = dev_priv->perf.oa.oa_buffer.vaddr; > > - u32 gtt_offset = i915_ggtt_offset(dev_priv->perf.oa.oa_buffer.vma); > > + struct drm_i915_private *i915 = stream->i915; > > + int report_size = i915->perf.oa.oa_buffer.format_size; > > + u8 *oa_buf_base = i915->perf.oa.oa_buffer.vaddr; > > + u32 gtt_offset = i915_ggtt_offset(i915->perf.oa.oa_buffer.vma); > > u32 mask = (OA_BUFFER_SIZE - 1); > > size_t start_offset = *offset; > > unsigned long flags; > > @@ -666,13 +666,13 @@ static int gen8_append_oa_reports(struct i915_perf_stream *stream, > > if (WARN_ON(!stream->enabled)) > > return -EIO; > > - spin_lock_irqsave(&dev_priv->perf.oa.oa_buffer.ptr_lock, flags); > > + spin_lock_irqsave(&i915->perf.oa.oa_buffer.ptr_lock, flags); > > - head = dev_priv->perf.oa.oa_buffer.head; > > - aged_tail_idx = dev_priv->perf.oa.oa_buffer.aged_tail_idx; > > - tail = dev_priv->perf.oa.oa_buffer.tails[aged_tail_idx].offset; > > + head = i915->perf.oa.oa_buffer.head; > > + aged_tail_idx = i915->perf.oa.oa_buffer.aged_tail_idx; > > + tail = i915->perf.oa.oa_buffer.tails[aged_tail_idx].offset; > > - spin_unlock_irqrestore(&dev_priv->perf.oa.oa_buffer.ptr_lock, flags); > > + spin_unlock_irqrestore(&i915->perf.oa.oa_buffer.ptr_lock, flags); > > /* > > * An invalid tail pointer here means we're still waiting for the poll > > @@ -736,12 +736,12 @@ static int gen8_append_oa_reports(struct i915_perf_stream *stream, > > reason = ((report32[0] >> OAREPORT_REASON_SHIFT) & > > OAREPORT_REASON_MASK); > > if (reason == 0) { > > - if (__ratelimit(&dev_priv->perf.oa.spurious_report_rs)) > > + if (__ratelimit(&i915->perf.oa.spurious_report_rs)) > > DRM_NOTE("Skipping spurious, invalid OA report\n"); > > continue; > > } > > - ctx_id = report32[2] & dev_priv->perf.oa.specific_ctx_id_mask; > > + ctx_id = report32[2] & i915->perf.oa.specific_ctx_id_mask; > > /* > > * Squash whatever is in the CTX_ID field if it's marked as > > @@ -751,7 +751,7 @@ static int gen8_append_oa_reports(struct i915_perf_stream *stream, > > * Note: that we don't clear the valid_ctx_bit so userspace can > > * understand that the ID has been squashed by the kernel. > > */ > > - if (!(report32[0] & dev_priv->perf.oa.gen8_valid_ctx_bit)) > > + if (!(report32[0] & i915->perf.oa.gen8_valid_ctx_bit)) > > ctx_id = report32[2] = INVALID_CTX_ID; > > /* > > @@ -785,18 +785,18 @@ static int gen8_append_oa_reports(struct i915_perf_stream *stream, > > * switches since it's not-uncommon for periodic samples to > > * identify a switch before any 'context switch' report. > > */ > > - if (!dev_priv->perf.oa.exclusive_stream->ctx || > > - dev_priv->perf.oa.specific_ctx_id == ctx_id || > > - (dev_priv->perf.oa.oa_buffer.last_ctx_id == > > - dev_priv->perf.oa.specific_ctx_id) || > > + if (!i915->perf.oa.exclusive_stream->ctx || > > + i915->perf.oa.specific_ctx_id == ctx_id || > > + (i915->perf.oa.oa_buffer.last_ctx_id == > > + i915->perf.oa.specific_ctx_id) || > > reason & OAREPORT_REASON_CTX_SWITCH) { > > /* > > * While filtering for a single context we avoid > > * leaking the IDs of other contexts. > > */ > > - if (dev_priv->perf.oa.exclusive_stream->ctx && > > - dev_priv->perf.oa.specific_ctx_id != ctx_id) { > > + if (i915->perf.oa.exclusive_stream->ctx && > > + i915->perf.oa.specific_ctx_id != ctx_id) { > > report32[2] = INVALID_CTX_ID; > > } > > @@ -805,7 +805,7 @@ static int gen8_append_oa_reports(struct i915_perf_stream *stream, > > if (ret) > > break; > > - dev_priv->perf.oa.oa_buffer.last_ctx_id = ctx_id; > > + i915->perf.oa.oa_buffer.last_ctx_id = ctx_id; > > } > > /* > > @@ -819,7 +819,7 @@ static int gen8_append_oa_reports(struct i915_perf_stream *stream, > > } > > if (start_offset != *offset) { > > - spin_lock_irqsave(&dev_priv->perf.oa.oa_buffer.ptr_lock, flags); > > + spin_lock_irqsave(&i915->perf.oa.oa_buffer.ptr_lock, flags); > > /* > > * We removed the gtt_offset for the copy loop above, indexing > > @@ -828,9 +828,9 @@ static int gen8_append_oa_reports(struct i915_perf_stream *stream, > > head += gtt_offset; > > I915_WRITE(GEN8_OAHEADPTR, head & GEN8_OAHEADPTR_MASK); > > - dev_priv->perf.oa.oa_buffer.head = head; > > + i915->perf.oa.oa_buffer.head = head; > > - spin_unlock_irqrestore(&dev_priv->perf.oa.oa_buffer.ptr_lock, flags); > > + spin_unlock_irqrestore(&i915->perf.oa.oa_buffer.ptr_lock, flags); > > } > > return ret; > > @@ -861,11 +861,11 @@ static int gen8_oa_read(struct i915_perf_stream *stream, > > size_t count, > > size_t *offset) > > { > > - struct drm_i915_private *dev_priv = stream->dev_priv; > > + struct drm_i915_private *i915 = stream->i915; > > u32 oastatus; > > int ret; > > - if (WARN_ON(!dev_priv->perf.oa.oa_buffer.vaddr)) > > + if (WARN_ON(!i915->perf.oa.oa_buffer.vaddr)) > > return -EIO; > > oastatus = I915_READ(GEN8_OASTATUS); > > @@ -891,10 +891,10 @@ static int gen8_oa_read(struct i915_perf_stream *stream, > > return ret; > > DRM_DEBUG("OA buffer overflow (exponent = %d): force restart\n", > > - dev_priv->perf.oa.period_exponent); > > + i915->perf.oa.period_exponent); > > - dev_priv->perf.oa.ops.oa_disable(stream); > > - dev_priv->perf.oa.ops.oa_enable(stream); > > + i915->perf.oa.ops.oa_disable(stream); > > + i915->perf.oa.ops.oa_enable(stream); > > /* > > * Note: .oa_enable() is expected to re-init the oabuffer and > > @@ -940,10 +940,10 @@ static int gen7_append_oa_reports(struct i915_perf_stream *stream, > > size_t count, > > size_t *offset) > > { > > - struct drm_i915_private *dev_priv = stream->dev_priv; > > - int report_size = dev_priv->perf.oa.oa_buffer.format_size; > > - u8 *oa_buf_base = dev_priv->perf.oa.oa_buffer.vaddr; > > - u32 gtt_offset = i915_ggtt_offset(dev_priv->perf.oa.oa_buffer.vma); > > + struct drm_i915_private *i915 = stream->i915; > > + int report_size = i915->perf.oa.oa_buffer.format_size; > > + u8 *oa_buf_base = i915->perf.oa.oa_buffer.vaddr; > > + u32 gtt_offset = i915_ggtt_offset(i915->perf.oa.oa_buffer.vma); > > u32 mask = (OA_BUFFER_SIZE - 1); > > size_t start_offset = *offset; > > unsigned long flags; > > @@ -955,13 +955,13 @@ static int gen7_append_oa_reports(struct i915_perf_stream *stream, > > if (WARN_ON(!stream->enabled)) > > return -EIO; > > - spin_lock_irqsave(&dev_priv->perf.oa.oa_buffer.ptr_lock, flags); > > + spin_lock_irqsave(&i915->perf.oa.oa_buffer.ptr_lock, flags); > > - head = dev_priv->perf.oa.oa_buffer.head; > > - aged_tail_idx = dev_priv->perf.oa.oa_buffer.aged_tail_idx; > > - tail = dev_priv->perf.oa.oa_buffer.tails[aged_tail_idx].offset; > > + head = i915->perf.oa.oa_buffer.head; > > + aged_tail_idx = i915->perf.oa.oa_buffer.aged_tail_idx; > > + tail = i915->perf.oa.oa_buffer.tails[aged_tail_idx].offset; > > - spin_unlock_irqrestore(&dev_priv->perf.oa.oa_buffer.ptr_lock, flags); > > + spin_unlock_irqrestore(&i915->perf.oa.oa_buffer.ptr_lock, flags); > > /* An invalid tail pointer here means we're still waiting for the poll > > * hrtimer callback to give us a pointer > > @@ -1014,7 +1014,7 @@ static int gen7_append_oa_reports(struct i915_perf_stream *stream, > > * copying it to userspace... > > */ > > if (report32[0] == 0) { > > - if (__ratelimit(&dev_priv->perf.oa.spurious_report_rs)) > > + if (__ratelimit(&i915->perf.oa.spurious_report_rs)) > > DRM_NOTE("Skipping spurious, invalid OA report\n"); > > continue; > > } > > @@ -1033,7 +1033,7 @@ static int gen7_append_oa_reports(struct i915_perf_stream *stream, > > } > > if (start_offset != *offset) { > > - spin_lock_irqsave(&dev_priv->perf.oa.oa_buffer.ptr_lock, flags); > > + spin_lock_irqsave(&i915->perf.oa.oa_buffer.ptr_lock, flags); > > /* We removed the gtt_offset for the copy loop above, indexing > > * relative to oa_buf_base so put back here... > > @@ -1043,9 +1043,9 @@ static int gen7_append_oa_reports(struct i915_perf_stream *stream, > > I915_WRITE(GEN7_OASTATUS2, > > ((head & GEN7_OASTATUS2_HEAD_MASK) | > > GEN7_OASTATUS2_MEM_SELECT_GGTT)); > > - dev_priv->perf.oa.oa_buffer.head = head; > > + i915->perf.oa.oa_buffer.head = head; > > - spin_unlock_irqrestore(&dev_priv->perf.oa.oa_buffer.ptr_lock, flags); > > + spin_unlock_irqrestore(&i915->perf.oa.oa_buffer.ptr_lock, flags); > > } > > return ret; > > @@ -1072,11 +1072,11 @@ static int gen7_oa_read(struct i915_perf_stream *stream, > > size_t count, > > size_t *offset) > > { > > - struct drm_i915_private *dev_priv = stream->dev_priv; > > + struct drm_i915_private *i915 = stream->i915; > > u32 oastatus1; > > int ret; > > - if (WARN_ON(!dev_priv->perf.oa.oa_buffer.vaddr)) > > + if (WARN_ON(!i915->perf.oa.oa_buffer.vaddr)) > > return -EIO; > > oastatus1 = I915_READ(GEN7_OASTATUS1); > > @@ -1086,7 +1086,7 @@ static int gen7_oa_read(struct i915_perf_stream *stream, > > * may be updated asynchronously) so we ignore status bits > > * that have already been reported to userspace. > > */ > > - oastatus1 &= ~dev_priv->perf.oa.gen7_latched_oastatus1; > > + oastatus1 &= ~i915->perf.oa.gen7_latched_oastatus1; > > /* We treat OABUFFER_OVERFLOW as a significant error: > > * > > @@ -1115,10 +1115,10 @@ static int gen7_oa_read(struct i915_perf_stream *stream, > > return ret; > > DRM_DEBUG("OA buffer overflow (exponent = %d): force restart\n", > > - dev_priv->perf.oa.period_exponent); > > + i915->perf.oa.period_exponent); > > - dev_priv->perf.oa.ops.oa_disable(stream); > > - dev_priv->perf.oa.ops.oa_enable(stream); > > + i915->perf.oa.ops.oa_disable(stream); > > + i915->perf.oa.ops.oa_enable(stream); > > oastatus1 = I915_READ(GEN7_OASTATUS1); > > } > > @@ -1128,7 +1128,7 @@ static int gen7_oa_read(struct i915_perf_stream *stream, > > DRM_I915_PERF_RECORD_OA_REPORT_LOST); > > if (ret) > > return ret; > > - dev_priv->perf.oa.gen7_latched_oastatus1 |= > > + i915->perf.oa.gen7_latched_oastatus1 |= > > GEN7_OASTATUS1_REPORT_LOST; > > } > > @@ -1151,14 +1151,14 @@ static int gen7_oa_read(struct i915_perf_stream *stream, > > */ > > static int i915_oa_wait_unlocked(struct i915_perf_stream *stream) > > { > > - struct drm_i915_private *dev_priv = stream->dev_priv; > > + struct drm_i915_private *i915 = stream->i915; > > /* We would wait indefinitely if periodic sampling is not enabled */ > > - if (!dev_priv->perf.oa.periodic) > > + if (!i915->perf.oa.periodic) > > return -EIO; > > - return wait_event_interruptible(dev_priv->perf.oa.poll_wq, > > - oa_buffer_check_unlocked(dev_priv)); > > + return wait_event_interruptible(i915->perf.oa.poll_wq, > > + oa_buffer_check_unlocked(i915)); > > } > > /** > > @@ -1175,9 +1175,9 @@ static void i915_oa_poll_wait(struct i915_perf_stream *stream, > > struct file *file, > > poll_table *wait) > > { > > - struct drm_i915_private *dev_priv = stream->dev_priv; > > + struct drm_i915_private *i915 = stream->i915; > > - poll_wait(file, &dev_priv->perf.oa.poll_wq, wait); > > + poll_wait(file, &i915->perf.oa.poll_wq, wait); > > } > > /** > > @@ -1197,9 +1197,9 @@ static int i915_oa_read(struct i915_perf_stream *stream, > > size_t count, > > size_t *offset) > > { > > - struct drm_i915_private *dev_priv = stream->dev_priv; > > + struct drm_i915_private *i915 = stream->i915; > > - return dev_priv->perf.oa.ops.read(stream, buf, count, offset); > > + return i915->perf.oa.ops.read(stream, buf, count, offset); > > } > > static struct intel_context *oa_pin_context(struct drm_i915_private *i915, > > @@ -1248,7 +1248,7 @@ static struct intel_context *oa_pin_context(struct drm_i915_private *i915, > > */ > > static int oa_get_render_ctx_id(struct i915_perf_stream *stream) > > { > > - struct drm_i915_private *i915 = stream->dev_priv; > > + struct drm_i915_private *i915 = stream->i915; > > struct intel_context *ce; > > ce = oa_pin_context(i915, stream->ctx); > > @@ -1330,17 +1330,17 @@ static int oa_get_render_ctx_id(struct i915_perf_stream *stream) > > */ > > static void oa_put_render_ctx_id(struct i915_perf_stream *stream) > > { > > - struct drm_i915_private *dev_priv = stream->dev_priv; > > + struct drm_i915_private *i915 = stream->i915; > > struct intel_context *ce; > > - dev_priv->perf.oa.specific_ctx_id = INVALID_CTX_ID; > > - dev_priv->perf.oa.specific_ctx_id_mask = 0; > > + i915->perf.oa.specific_ctx_id = INVALID_CTX_ID; > > + i915->perf.oa.specific_ctx_id_mask = 0; > > - ce = fetch_and_zero(&dev_priv->perf.oa.pinned_ctx); > > + ce = fetch_and_zero(&i915->perf.oa.pinned_ctx); > > if (ce) { > > - mutex_lock(&dev_priv->drm.struct_mutex); > > + mutex_lock(&i915->drm.struct_mutex); > > intel_context_unpin(ce); > > - mutex_unlock(&dev_priv->drm.struct_mutex); > > + mutex_unlock(&i915->drm.struct_mutex); > > } > > } > > @@ -1359,64 +1359,64 @@ free_oa_buffer(struct drm_i915_private *i915) > > static void i915_oa_stream_destroy(struct i915_perf_stream *stream) > > { > > - struct drm_i915_private *dev_priv = stream->dev_priv; > > + struct drm_i915_private *i915 = stream->i915; > > - BUG_ON(stream != dev_priv->perf.oa.exclusive_stream); > > + BUG_ON(stream != i915->perf.oa.exclusive_stream); > > /* > > * Unset exclusive_stream first, it will be checked while disabling > > * the metric set on gen8+. > > */ > > - mutex_lock(&dev_priv->drm.struct_mutex); > > - dev_priv->perf.oa.exclusive_stream = NULL; > > - dev_priv->perf.oa.ops.disable_metric_set(dev_priv); > > - mutex_unlock(&dev_priv->drm.struct_mutex); > > + mutex_lock(&i915->drm.struct_mutex); > > + i915->perf.oa.exclusive_stream = NULL; > > + i915->perf.oa.ops.disable_metric_set(i915); > > + mutex_unlock(&i915->drm.struct_mutex); > > - free_oa_buffer(dev_priv); > > + free_oa_buffer(i915); > > - intel_uncore_forcewake_put(&dev_priv->uncore, FORCEWAKE_ALL); > > - intel_runtime_pm_put(dev_priv, stream->wakeref); > > + intel_uncore_forcewake_put(&i915->uncore, FORCEWAKE_ALL); > > + intel_runtime_pm_put(i915, stream->wakeref); > > if (stream->ctx) > > oa_put_render_ctx_id(stream); > > - put_oa_config(dev_priv, stream->oa_config); > > + put_oa_config(i915, stream->oa_config); > > - if (dev_priv->perf.oa.spurious_report_rs.missed) { > > + if (i915->perf.oa.spurious_report_rs.missed) { > > DRM_NOTE("%d spurious OA report notices suppressed due to ratelimiting\n", > > - dev_priv->perf.oa.spurious_report_rs.missed); > > + i915->perf.oa.spurious_report_rs.missed); > > } > > } > > -static void gen7_init_oa_buffer(struct drm_i915_private *dev_priv) > > +static void gen7_init_oa_buffer(struct drm_i915_private *i915) > > { > > - u32 gtt_offset = i915_ggtt_offset(dev_priv->perf.oa.oa_buffer.vma); > > + u32 gtt_offset = i915_ggtt_offset(i915->perf.oa.oa_buffer.vma); > > unsigned long flags; > > - spin_lock_irqsave(&dev_priv->perf.oa.oa_buffer.ptr_lock, flags); > > + spin_lock_irqsave(&i915->perf.oa.oa_buffer.ptr_lock, flags); > > /* Pre-DevBDW: OABUFFER must be set with counters off, > > * before OASTATUS1, but after OASTATUS2 > > */ > > I915_WRITE(GEN7_OASTATUS2, > > gtt_offset | GEN7_OASTATUS2_MEM_SELECT_GGTT); /* head */ > > - dev_priv->perf.oa.oa_buffer.head = gtt_offset; > > + i915->perf.oa.oa_buffer.head = gtt_offset; > > I915_WRITE(GEN7_OABUFFER, gtt_offset); > > I915_WRITE(GEN7_OASTATUS1, gtt_offset | OABUFFER_SIZE_16M); /* tail */ > > /* Mark that we need updated tail pointers to read from... */ > > - dev_priv->perf.oa.oa_buffer.tails[0].offset = INVALID_TAIL_PTR; > > - dev_priv->perf.oa.oa_buffer.tails[1].offset = INVALID_TAIL_PTR; > > + i915->perf.oa.oa_buffer.tails[0].offset = INVALID_TAIL_PTR; > > + i915->perf.oa.oa_buffer.tails[1].offset = INVALID_TAIL_PTR; > > - spin_unlock_irqrestore(&dev_priv->perf.oa.oa_buffer.ptr_lock, flags); > > + spin_unlock_irqrestore(&i915->perf.oa.oa_buffer.ptr_lock, flags); > > /* On Haswell we have to track which OASTATUS1 flags we've > > * already seen since they can't be cleared while periodic > > * sampling is enabled. > > */ > > - dev_priv->perf.oa.gen7_latched_oastatus1 = 0; > > + i915->perf.oa.gen7_latched_oastatus1 = 0; > > /* NB: although the OA buffer will initially be allocated > > * zeroed via shmfs (and so this memset is redundant when > > @@ -1429,24 +1429,24 @@ static void gen7_init_oa_buffer(struct drm_i915_private *dev_priv) > > * the assumption that new reports are being written to zeroed > > * memory... > > */ > > - memset(dev_priv->perf.oa.oa_buffer.vaddr, 0, OA_BUFFER_SIZE); > > + memset(i915->perf.oa.oa_buffer.vaddr, 0, OA_BUFFER_SIZE); > > /* Maybe make ->pollin per-stream state if we support multiple > > * concurrent streams in the future. > > */ > > - dev_priv->perf.oa.pollin = false; > > + i915->perf.oa.pollin = false; > > } > > -static void gen8_init_oa_buffer(struct drm_i915_private *dev_priv) > > +static void gen8_init_oa_buffer(struct drm_i915_private *i915) > > { > > - u32 gtt_offset = i915_ggtt_offset(dev_priv->perf.oa.oa_buffer.vma); > > + u32 gtt_offset = i915_ggtt_offset(i915->perf.oa.oa_buffer.vma); > > unsigned long flags; > > - spin_lock_irqsave(&dev_priv->perf.oa.oa_buffer.ptr_lock, flags); > > + spin_lock_irqsave(&i915->perf.oa.oa_buffer.ptr_lock, flags); > > I915_WRITE(GEN8_OASTATUS, 0); > > I915_WRITE(GEN8_OAHEADPTR, gtt_offset); > > - dev_priv->perf.oa.oa_buffer.head = gtt_offset; > > + i915->perf.oa.oa_buffer.head = gtt_offset; > > I915_WRITE(GEN8_OABUFFER_UDW, 0); > > @@ -1463,17 +1463,17 @@ static void gen8_init_oa_buffer(struct drm_i915_private *dev_priv) > > I915_WRITE(GEN8_OATAILPTR, gtt_offset & GEN8_OATAILPTR_MASK); > > /* Mark that we need updated tail pointers to read from... */ > > - dev_priv->perf.oa.oa_buffer.tails[0].offset = INVALID_TAIL_PTR; > > - dev_priv->perf.oa.oa_buffer.tails[1].offset = INVALID_TAIL_PTR; > > + i915->perf.oa.oa_buffer.tails[0].offset = INVALID_TAIL_PTR; > > + i915->perf.oa.oa_buffer.tails[1].offset = INVALID_TAIL_PTR; > > /* > > * Reset state used to recognise context switches, affecting which > > * reports we will forward to userspace while filtering for a single > > * context. > > */ > > - dev_priv->perf.oa.oa_buffer.last_ctx_id = INVALID_CTX_ID; > > + i915->perf.oa.oa_buffer.last_ctx_id = INVALID_CTX_ID; > > - spin_unlock_irqrestore(&dev_priv->perf.oa.oa_buffer.ptr_lock, flags); > > + spin_unlock_irqrestore(&i915->perf.oa.oa_buffer.ptr_lock, flags); > > /* > > * NB: although the OA buffer will initially be allocated > > @@ -1487,32 +1487,32 @@ static void gen8_init_oa_buffer(struct drm_i915_private *dev_priv) > > * the assumption that new reports are being written to zeroed > > * memory... > > */ > > - memset(dev_priv->perf.oa.oa_buffer.vaddr, 0, OA_BUFFER_SIZE); > > + memset(i915->perf.oa.oa_buffer.vaddr, 0, OA_BUFFER_SIZE); > > /* > > * Maybe make ->pollin per-stream state if we support multiple > > * concurrent streams in the future. > > */ > > - dev_priv->perf.oa.pollin = false; > > + i915->perf.oa.pollin = false; > > } > > -static int alloc_oa_buffer(struct drm_i915_private *dev_priv) > > +static int alloc_oa_buffer(struct drm_i915_private *i915) > > { > > struct drm_i915_gem_object *bo; > > struct i915_vma *vma; > > int ret; > > - if (WARN_ON(dev_priv->perf.oa.oa_buffer.vma)) > > + if (WARN_ON(i915->perf.oa.oa_buffer.vma)) > > return -ENODEV; > > - ret = i915_mutex_lock_interruptible(&dev_priv->drm); > > + ret = i915_mutex_lock_interruptible(&i915->drm); > > if (ret) > > return ret; > > BUILD_BUG_ON_NOT_POWER_OF_2(OA_BUFFER_SIZE); > > BUILD_BUG_ON(OA_BUFFER_SIZE < SZ_128K || OA_BUFFER_SIZE > SZ_16M); > > - bo = i915_gem_object_create_shmem(dev_priv, OA_BUFFER_SIZE); > > + bo = i915_gem_object_create_shmem(i915, OA_BUFFER_SIZE); > > if (IS_ERR(bo)) { > > DRM_ERROR("Failed to allocate OA buffer\n"); > > ret = PTR_ERR(bo); > > @@ -1527,18 +1527,18 @@ static int alloc_oa_buffer(struct drm_i915_private *dev_priv) > > ret = PTR_ERR(vma); > > goto err_unref; > > } > > - dev_priv->perf.oa.oa_buffer.vma = vma; > > + i915->perf.oa.oa_buffer.vma = vma; > > - dev_priv->perf.oa.oa_buffer.vaddr = > > + i915->perf.oa.oa_buffer.vaddr = > > i915_gem_object_pin_map(bo, I915_MAP_WB); > > - if (IS_ERR(dev_priv->perf.oa.oa_buffer.vaddr)) { > > - ret = PTR_ERR(dev_priv->perf.oa.oa_buffer.vaddr); > > + if (IS_ERR(i915->perf.oa.oa_buffer.vaddr)) { > > + ret = PTR_ERR(i915->perf.oa.oa_buffer.vaddr); > > goto err_unpin; > > } > > DRM_DEBUG_DRIVER("OA Buffer initialized, gtt offset = 0x%x, vaddr = %p\n", > > - i915_ggtt_offset(dev_priv->perf.oa.oa_buffer.vma), > > - dev_priv->perf.oa.oa_buffer.vaddr); > > + i915_ggtt_offset(i915->perf.oa.oa_buffer.vma), > > + i915->perf.oa.oa_buffer.vaddr); > > goto unlock; > > @@ -1548,15 +1548,15 @@ static int alloc_oa_buffer(struct drm_i915_private *dev_priv) > > err_unref: > > i915_gem_object_put(bo); > > - dev_priv->perf.oa.oa_buffer.vaddr = NULL; > > - dev_priv->perf.oa.oa_buffer.vma = NULL; > > + i915->perf.oa.oa_buffer.vaddr = NULL; > > + i915->perf.oa.oa_buffer.vma = NULL; > > unlock: > > - mutex_unlock(&dev_priv->drm.struct_mutex); > > + mutex_unlock(&i915->drm.struct_mutex); > > return ret; > > } > > -static void config_oa_regs(struct drm_i915_private *dev_priv, > > +static void config_oa_regs(struct drm_i915_private *i915, > > const struct i915_oa_reg *regs, > > u32 n_regs) > > { > > @@ -1571,7 +1571,7 @@ static void config_oa_regs(struct drm_i915_private *dev_priv, > > static int hsw_enable_metric_set(struct i915_perf_stream *stream) > > { > > - struct drm_i915_private *dev_priv = stream->dev_priv; > > + struct drm_i915_private *i915 = stream->i915; > > const struct i915_oa_config *oa_config = stream->oa_config; > > /* PRM: > > @@ -1588,7 +1588,7 @@ static int hsw_enable_metric_set(struct i915_perf_stream *stream) > > I915_WRITE(GEN6_UCGCTL1, (I915_READ(GEN6_UCGCTL1) | > > GEN6_CSUNIT_CLOCK_GATE_DISABLE)); > > - config_oa_regs(dev_priv, oa_config->mux_regs, oa_config->mux_regs_len); > > + config_oa_regs(i915, oa_config->mux_regs, oa_config->mux_regs_len); > > /* It apparently takes a fairly long time for a new MUX > > * configuration to be be applied after these register writes. > > @@ -1613,13 +1613,13 @@ static int hsw_enable_metric_set(struct i915_perf_stream *stream) > > */ > > usleep_range(15000, 20000); > > - config_oa_regs(dev_priv, oa_config->b_counter_regs, > > + config_oa_regs(i915, oa_config->b_counter_regs, > > oa_config->b_counter_regs_len); > > return 0; > > } > > -static void hsw_disable_metric_set(struct drm_i915_private *dev_priv) > > +static void hsw_disable_metric_set(struct drm_i915_private *i915) > > { > > I915_WRITE(GEN6_UCGCTL1, (I915_READ(GEN6_UCGCTL1) & > > ~GEN6_CSUNIT_CLOCK_GATE_DISABLE)); > > @@ -1716,15 +1716,15 @@ gen8_update_reg_state_unlocked(struct intel_context *ce, > > * > > * Note: it's only the RCS/Render context that has any OA state. > > */ > > -static int gen8_configure_all_contexts(struct drm_i915_private *dev_priv, > > +static int gen8_configure_all_contexts(struct drm_i915_private *i915, > > const struct i915_oa_config *oa_config) > > { > > - unsigned int map_type = i915_coherent_map_type(dev_priv); > > + unsigned int map_type = i915_coherent_map_type(i915); > > struct i915_gem_context *ctx; > > struct i915_request *rq; > > int ret; > > - lockdep_assert_held(&dev_priv->drm.struct_mutex); > > + lockdep_assert_held(&i915->drm.struct_mutex); > > /* > > * The OA register config is setup through the context image. This image > > @@ -1739,14 +1739,14 @@ static int gen8_configure_all_contexts(struct drm_i915_private *dev_priv, > > * So far the best way to work around this issue seems to be draining > > * the GPU from any submitted work. > > */ > > - ret = i915_gem_wait_for_idle(dev_priv, > > + ret = i915_gem_wait_for_idle(i915, > > I915_WAIT_LOCKED, > > MAX_SCHEDULE_TIMEOUT); > > if (ret) > > return ret; > > /* Update all contexts now that we've stalled the submission. */ > > - list_for_each_entry(ctx, &dev_priv->contexts.list, link) { > > + list_for_each_entry(ctx, &i915->contexts.list, link) { > > struct i915_gem_engines_iter it; > > struct intel_context *ce; > > @@ -1783,7 +1783,7 @@ static int gen8_configure_all_contexts(struct drm_i915_private *dev_priv, > > * Apply the configuration by doing one context restore of the edited > > * context image. > > */ > > - rq = i915_request_create(dev_priv->engine[RCS0]->kernel_context); > > + rq = i915_request_create(i915->engine[RCS0]->kernel_context); > > if (IS_ERR(rq)) > > return PTR_ERR(rq); > > @@ -1794,7 +1794,7 @@ static int gen8_configure_all_contexts(struct drm_i915_private *dev_priv, > > static int gen8_enable_metric_set(struct i915_perf_stream *stream) > > { > > - struct drm_i915_private *dev_priv = stream->dev_priv; > > + struct drm_i915_private *i915 = stream->i915; > > const struct i915_oa_config *oa_config = stream->oa_config; > > int ret; > > @@ -1821,7 +1821,7 @@ static int gen8_enable_metric_set(struct i915_perf_stream *stream) > > * be read back from automatically triggered reports, as part of the > > * RPT_ID field. > > */ > > - if (IS_GEN_RANGE(dev_priv, 9, 11)) { > > + if (IS_GEN_RANGE(i915, 9, 11)) { > > I915_WRITE(GEN8_OA_DEBUG, > > _MASKED_BIT_ENABLE(GEN9_OA_DEBUG_DISABLE_CLK_RATIO_REPORTS | > > GEN9_OA_DEBUG_INCLUDE_CLK_RATIO)); > > @@ -1832,31 +1832,31 @@ static int gen8_enable_metric_set(struct i915_perf_stream *stream) > > * to make sure all slices/subslices are ON before writing to NOA > > * registers. > > */ > > - ret = gen8_configure_all_contexts(dev_priv, oa_config); > > + ret = gen8_configure_all_contexts(i915, oa_config); > > if (ret) > > return ret; > > - config_oa_regs(dev_priv, oa_config->mux_regs, oa_config->mux_regs_len); > > + config_oa_regs(i915, oa_config->mux_regs, oa_config->mux_regs_len); > > - config_oa_regs(dev_priv, oa_config->b_counter_regs, > > + config_oa_regs(i915, oa_config->b_counter_regs, > > oa_config->b_counter_regs_len); > > return 0; > > } > > -static void gen8_disable_metric_set(struct drm_i915_private *dev_priv) > > +static void gen8_disable_metric_set(struct drm_i915_private *i915) > > { > > /* Reset all contexts' slices/subslices configurations. */ > > - gen8_configure_all_contexts(dev_priv, NULL); > > + gen8_configure_all_contexts(i915, NULL); > > I915_WRITE(GDT_CHICKEN_BITS, (I915_READ(GDT_CHICKEN_BITS) & > > ~GT_NOA_ENABLE)); > > } > > -static void gen10_disable_metric_set(struct drm_i915_private *dev_priv) > > +static void gen10_disable_metric_set(struct drm_i915_private *i915) > > { > > /* Reset all contexts' slices/subslices configurations. */ > > - gen8_configure_all_contexts(dev_priv, NULL); > > + gen8_configure_all_contexts(i915, NULL); > > /* Make sure we disable noa to save power. */ > > I915_WRITE(RPM_CONFIG1, > > @@ -1865,12 +1865,12 @@ static void gen10_disable_metric_set(struct drm_i915_private *dev_priv) > > static void gen7_oa_enable(struct i915_perf_stream *stream) > > { > > - struct drm_i915_private *dev_priv = stream->dev_priv; > > + struct drm_i915_private *i915 = stream->i915; > > struct i915_gem_context *ctx = stream->ctx; > > - u32 ctx_id = dev_priv->perf.oa.specific_ctx_id; > > - bool periodic = dev_priv->perf.oa.periodic; > > - u32 period_exponent = dev_priv->perf.oa.period_exponent; > > - u32 report_format = dev_priv->perf.oa.oa_buffer.format; > > + u32 ctx_id = i915->perf.oa.specific_ctx_id; > > + bool periodic = i915->perf.oa.periodic; > > + u32 period_exponent = i915->perf.oa.period_exponent; > > + u32 report_format = i915->perf.oa.oa_buffer.format; > > /* > > * Reset buf pointers so we don't forward reports from before now. > > @@ -1881,7 +1881,7 @@ static void gen7_oa_enable(struct i915_perf_stream *stream) > > * on the assumption that certain fields are written to zeroed > > * memory which this helps maintains. > > */ > > - gen7_init_oa_buffer(dev_priv); > > + gen7_init_oa_buffer(i915); > > I915_WRITE(GEN7_OACONTROL, > > (ctx_id & GEN7_OACONTROL_CTX_MASK) | > > @@ -1895,8 +1895,8 @@ static void gen7_oa_enable(struct i915_perf_stream *stream) > > static void gen8_oa_enable(struct i915_perf_stream *stream) > > { > > - struct drm_i915_private *dev_priv = stream->dev_priv; > > - u32 report_format = dev_priv->perf.oa.oa_buffer.format; > > + struct drm_i915_private *i915 = stream->i915; > > + u32 report_format = i915->perf.oa.oa_buffer.format; > > /* > > * Reset buf pointers so we don't forward reports from before now. > > @@ -1907,7 +1907,7 @@ static void gen8_oa_enable(struct i915_perf_stream *stream) > > * on the assumption that certain fields are written to zeroed > > * memory which this helps maintains. > > */ > > - gen8_init_oa_buffer(dev_priv); > > + gen8_init_oa_buffer(i915); > > /* > > * Note: we don't rely on the hardware to perform single context > > @@ -1930,19 +1930,19 @@ static void gen8_oa_enable(struct i915_perf_stream *stream) > > */ > > static void i915_oa_stream_enable(struct i915_perf_stream *stream) > > { > > - struct drm_i915_private *dev_priv = stream->dev_priv; > > + struct drm_i915_private *i915 = stream->i915; > > - dev_priv->perf.oa.ops.oa_enable(stream); > > + i915->perf.oa.ops.oa_enable(stream); > > - if (dev_priv->perf.oa.periodic) > > - hrtimer_start(&dev_priv->perf.oa.poll_check_timer, > > + if (i915->perf.oa.periodic) > > + hrtimer_start(&i915->perf.oa.poll_check_timer, > > ns_to_ktime(POLL_PERIOD), > > HRTIMER_MODE_REL_PINNED); > > } > > static void gen7_oa_disable(struct i915_perf_stream *stream) > > { > > - struct intel_uncore *uncore = &stream->dev_priv->uncore; > > + struct intel_uncore *uncore = &stream->i915->uncore; > > intel_uncore_write(uncore, GEN7_OACONTROL, 0); > > if (intel_wait_for_register(uncore, > > @@ -1953,7 +1953,7 @@ static void gen7_oa_disable(struct i915_perf_stream *stream) > > static void gen8_oa_disable(struct i915_perf_stream *stream) > > { > > - struct intel_uncore *uncore = &stream->dev_priv->uncore; > > + struct intel_uncore *uncore = &stream->i915->uncore; > > intel_uncore_write(uncore, GEN8_OACONTROL, 0); > > if (intel_wait_for_register(uncore, > > @@ -1972,12 +1972,12 @@ static void gen8_oa_disable(struct i915_perf_stream *stream) > > */ > > static void i915_oa_stream_disable(struct i915_perf_stream *stream) > > { > > - struct drm_i915_private *dev_priv = stream->dev_priv; > > + struct drm_i915_private *i915 = stream->i915; > > - dev_priv->perf.oa.ops.oa_disable(stream); > > + i915->perf.oa.ops.oa_disable(stream); > > - if (dev_priv->perf.oa.periodic) > > - hrtimer_cancel(&dev_priv->perf.oa.poll_check_timer); > > + if (i915->perf.oa.periodic) > > + hrtimer_cancel(&i915->perf.oa.poll_check_timer); > > } > > static const struct i915_perf_stream_ops i915_oa_stream_ops = { > > @@ -2011,7 +2011,7 @@ static int i915_oa_stream_init(struct i915_perf_stream *stream, > > struct drm_i915_perf_open_param *param, > > struct perf_open_properties *props) > > { > > - struct drm_i915_private *dev_priv = stream->dev_priv; > > + struct drm_i915_private *i915 = stream->i915; > > int format_size; > > int ret; > > @@ -2019,7 +2019,7 @@ static int i915_oa_stream_init(struct i915_perf_stream *stream, > > * reason then don't let userspace try their luck with config > > * IDs > > */ > > - if (!dev_priv->perf.metrics_kobj) { > > + if (!i915->perf.metrics_kobj) { > > DRM_DEBUG("OA metrics weren't advertised via sysfs\n"); > > return -EINVAL; > > } > > @@ -2029,7 +2029,7 @@ static int i915_oa_stream_init(struct i915_perf_stream *stream, > > return -EINVAL; > > } > > - if (!dev_priv->perf.oa.ops.enable_metric_set) { > > + if (!i915->perf.oa.ops.enable_metric_set) { > > DRM_DEBUG("OA unit not supported\n"); > > return -ENODEV; > > } > > @@ -2038,7 +2038,7 @@ static int i915_oa_stream_init(struct i915_perf_stream *stream, > > * counter reports and marshal to the appropriate client > > * we currently only allow exclusive access > > */ > > - if (dev_priv->perf.oa.exclusive_stream) { > > + if (i915->perf.oa.exclusive_stream) { > > DRM_DEBUG("OA unit already in use\n"); > > return -EBUSY; > > } > > @@ -2059,32 +2059,32 @@ static int i915_oa_stream_init(struct i915_perf_stream *stream, > > * > > * Using the same limiting factors as printk_ratelimit() > > */ > > - ratelimit_state_init(&dev_priv->perf.oa.spurious_report_rs, > > + ratelimit_state_init(&i915->perf.oa.spurious_report_rs, > > 5 * HZ, 10); > > /* Since we use a DRM_NOTE for spurious reports it would be > > * inconsistent to let __ratelimit() automatically print a warning for > > * throttling. > > */ > > - ratelimit_set_flags(&dev_priv->perf.oa.spurious_report_rs, > > + ratelimit_set_flags(&i915->perf.oa.spurious_report_rs, > > RATELIMIT_MSG_ON_RELEASE); > > stream->sample_size = sizeof(struct drm_i915_perf_record_header); > > - format_size = dev_priv->perf.oa.oa_formats[props->oa_format].size; > > + format_size = i915->perf.oa.oa_formats[props->oa_format].size; > > stream->sample_flags |= SAMPLE_OA_REPORT; > > stream->sample_size += format_size; > > - dev_priv->perf.oa.oa_buffer.format_size = format_size; > > - if (WARN_ON(dev_priv->perf.oa.oa_buffer.format_size == 0)) > > + i915->perf.oa.oa_buffer.format_size = format_size; > > + if (WARN_ON(i915->perf.oa.oa_buffer.format_size == 0)) > > return -EINVAL; > > - dev_priv->perf.oa.oa_buffer.format = > > - dev_priv->perf.oa.oa_formats[props->oa_format].format; > > + i915->perf.oa.oa_buffer.format = > > + i915->perf.oa.oa_formats[props->oa_format].format; > > - dev_priv->perf.oa.periodic = props->oa_periodic; > > - if (dev_priv->perf.oa.periodic) > > - dev_priv->perf.oa.period_exponent = props->oa_period_exponent; > > + i915->perf.oa.periodic = props->oa_periodic; > > + if (i915->perf.oa.periodic) > > + i915->perf.oa.period_exponent = props->oa_period_exponent; > > if (stream->ctx) { > > ret = oa_get_render_ctx_id(stream); > > @@ -2094,7 +2094,7 @@ static int i915_oa_stream_init(struct i915_perf_stream *stream, > > } > > } > > - ret = get_oa_config(dev_priv, props->metrics_set, &stream->oa_config); > > + ret = get_oa_config(i915, props->metrics_set, &stream->oa_config); > > if (ret) { > > DRM_DEBUG("Invalid OA config id=%i\n", props->metrics_set); > > goto err_config; > > @@ -2112,43 +2112,43 @@ static int i915_oa_stream_init(struct i915_perf_stream *stream, > > * In our case we are expecting that taking pm + FORCEWAKE > > * references will effectively disable RC6. > > */ > > - stream->wakeref = intel_runtime_pm_get(dev_priv); > > - intel_uncore_forcewake_get(&dev_priv->uncore, FORCEWAKE_ALL); > > + stream->wakeref = intel_runtime_pm_get(i915); > > + intel_uncore_forcewake_get(&i915->uncore, FORCEWAKE_ALL); > > - ret = alloc_oa_buffer(dev_priv); > > + ret = alloc_oa_buffer(i915); > > if (ret) > > goto err_oa_buf_alloc; > > - ret = i915_mutex_lock_interruptible(&dev_priv->drm); > > + ret = i915_mutex_lock_interruptible(&i915->drm); > > if (ret) > > goto err_lock; > > stream->ops = &i915_oa_stream_ops; > > - dev_priv->perf.oa.exclusive_stream = stream; > > + i915->perf.oa.exclusive_stream = stream; > > - ret = dev_priv->perf.oa.ops.enable_metric_set(stream); > > + ret = i915->perf.oa.ops.enable_metric_set(stream); > > if (ret) { > > DRM_DEBUG("Unable to enable metric set\n"); > > goto err_enable; > > } > > - mutex_unlock(&dev_priv->drm.struct_mutex); > > + mutex_unlock(&i915->drm.struct_mutex); > > return 0; > > err_enable: > > - dev_priv->perf.oa.exclusive_stream = NULL; > > - dev_priv->perf.oa.ops.disable_metric_set(dev_priv); > > - mutex_unlock(&dev_priv->drm.struct_mutex); > > + i915->perf.oa.exclusive_stream = NULL; > > + i915->perf.oa.ops.disable_metric_set(i915); > > + mutex_unlock(&i915->drm.struct_mutex); > > err_lock: > > - free_oa_buffer(dev_priv); > > + free_oa_buffer(i915); > > err_oa_buf_alloc: > > - put_oa_config(dev_priv, stream->oa_config); > > + put_oa_config(i915, stream->oa_config); > > - intel_uncore_forcewake_put(&dev_priv->uncore, FORCEWAKE_ALL); > > - intel_runtime_pm_put(dev_priv, stream->wakeref); > > + intel_uncore_forcewake_put(&i915->uncore, FORCEWAKE_ALL); > > + intel_runtime_pm_put(i915, stream->wakeref); > > err_config: > > if (stream->ctx) > > @@ -2238,7 +2238,7 @@ static ssize_t i915_perf_read(struct file *file, > > loff_t *ppos) > > { > > struct i915_perf_stream *stream = file->private_data; > > - struct drm_i915_private *dev_priv = stream->dev_priv; > > + struct drm_i915_private *i915 = stream->i915; > > ssize_t ret; > > /* To ensure it's handled consistently we simply treat all reads of a > > @@ -2261,15 +2261,15 @@ static ssize_t i915_perf_read(struct file *file, > > if (ret) > > return ret; > > - mutex_lock(&dev_priv->perf.lock); > > + mutex_lock(&i915->perf.lock); > > ret = i915_perf_read_locked(stream, file, > > buf, count, ppos); > > - mutex_unlock(&dev_priv->perf.lock); > > + mutex_unlock(&i915->perf.lock); > > } while (ret == -EAGAIN); > > } else { > > - mutex_lock(&dev_priv->perf.lock); > > + mutex_lock(&i915->perf.lock); > > ret = i915_perf_read_locked(stream, file, buf, count, ppos); > > - mutex_unlock(&dev_priv->perf.lock); > > + mutex_unlock(&i915->perf.lock); > > } > > /* We allow the poll checking to sometimes report false positive EPOLLIN > > @@ -2284,7 +2284,7 @@ static ssize_t i915_perf_read(struct file *file, > > /* Maybe make ->pollin per-stream state if we support multiple > > * concurrent streams in the future. > > */ > > - dev_priv->perf.oa.pollin = false; > > + i915->perf.oa.pollin = false; > > } > > return ret; > > @@ -2292,13 +2292,13 @@ static ssize_t i915_perf_read(struct file *file, > > static enum hrtimer_restart oa_poll_check_timer_cb(struct hrtimer *hrtimer) > > { > > - struct drm_i915_private *dev_priv = > > - container_of(hrtimer, typeof(*dev_priv), > > + struct drm_i915_private *i915 = > > + container_of(hrtimer, typeof(*i915), > > perf.oa.poll_check_timer); > > - if (oa_buffer_check_unlocked(dev_priv)) { > > - dev_priv->perf.oa.pollin = true; > > - wake_up(&dev_priv->perf.oa.poll_wq); > > + if (oa_buffer_check_unlocked(i915)) { > > + i915->perf.oa.pollin = true; > > + wake_up(&i915->perf.oa.poll_wq); > > } > > hrtimer_forward_now(hrtimer, ns_to_ktime(POLL_PERIOD)); > > @@ -2308,7 +2308,7 @@ static enum hrtimer_restart oa_poll_check_timer_cb(struct hrtimer *hrtimer) > > /** > > * i915_perf_poll_locked - poll_wait() with a suitable wait queue for stream > > - * @dev_priv: i915 device instance > > + * @i915: i915 device instance > > * @stream: An i915 perf stream > > * @file: An i915 perf stream file > > * @wait: poll() state table > > @@ -2322,7 +2322,7 @@ static enum hrtimer_restart oa_poll_check_timer_cb(struct hrtimer *hrtimer) > > * > > * Returns: any poll events that are ready without sleeping > > */ > > -static __poll_t i915_perf_poll_locked(struct drm_i915_private *dev_priv, > > +static __poll_t i915_perf_poll_locked(struct drm_i915_private *i915, > > struct i915_perf_stream *stream, > > struct file *file, > > poll_table *wait) > > @@ -2337,7 +2337,7 @@ static __poll_t i915_perf_poll_locked(struct drm_i915_private *dev_priv, > > * the hrtimer/oa_poll_check_timer_cb to notify us when there are > > * samples to read. > > */ > > - if (dev_priv->perf.oa.pollin) > > + if (i915->perf.oa.pollin) > > events |= EPOLLIN; > > return events; > > @@ -2359,12 +2359,12 @@ static __poll_t i915_perf_poll_locked(struct drm_i915_private *dev_priv, > > static __poll_t i915_perf_poll(struct file *file, poll_table *wait) > > { > > struct i915_perf_stream *stream = file->private_data; > > - struct drm_i915_private *dev_priv = stream->dev_priv; > > + struct drm_i915_private *i915 = stream->i915; > > __poll_t ret; > > - mutex_lock(&dev_priv->perf.lock); > > - ret = i915_perf_poll_locked(dev_priv, stream, file, wait); > > - mutex_unlock(&dev_priv->perf.lock); > > + mutex_lock(&i915->perf.lock); > > + ret = i915_perf_poll_locked(i915, stream, file, wait); > > + mutex_unlock(&i915->perf.lock); > > return ret; > > } > > @@ -2461,12 +2461,12 @@ static long i915_perf_ioctl(struct file *file, > > unsigned long arg) > > { > > struct i915_perf_stream *stream = file->private_data; > > - struct drm_i915_private *dev_priv = stream->dev_priv; > > + struct drm_i915_private *i915 = stream->i915; > > long ret; > > - mutex_lock(&dev_priv->perf.lock); > > + mutex_lock(&i915->perf.lock); > > ret = i915_perf_ioctl_locked(stream, cmd, arg); > > - mutex_unlock(&dev_priv->perf.lock); > > + mutex_unlock(&i915->perf.lock); > > return ret; > > } > > @@ -2511,11 +2511,11 @@ static void i915_perf_destroy_locked(struct i915_perf_stream *stream) > > static int i915_perf_release(struct inode *inode, struct file *file) > > { > > struct i915_perf_stream *stream = file->private_data; > > - struct drm_i915_private *dev_priv = stream->dev_priv; > > + struct drm_i915_private *i915 = stream->i915; > > - mutex_lock(&dev_priv->perf.lock); > > + mutex_lock(&i915->perf.lock); > > i915_perf_destroy_locked(stream); > > - mutex_unlock(&dev_priv->perf.lock); > > + mutex_unlock(&i915->perf.lock); > > return 0; > > } > > @@ -2537,7 +2537,7 @@ static const struct file_operations fops = { > > /** > > * i915_perf_open_ioctl_locked - DRM ioctl() for userspace to open a stream FD > > - * @dev_priv: i915 device instance > > + * @i915: i915 device instance > > * @param: The open parameters passed to 'DRM_I915_PERF_OPEN` > > * @props: individually validated u64 property value pairs > > * @file: drm file > > @@ -2560,7 +2560,7 @@ static const struct file_operations fops = { > > * Returns: zero on success or a negative error code. > > */ > > static int > > -i915_perf_open_ioctl_locked(struct drm_i915_private *dev_priv, > > +i915_perf_open_ioctl_locked(struct drm_i915_private *i915, > > struct drm_i915_perf_open_param *param, > > struct perf_open_properties *props, > > struct drm_file *file) > > @@ -2599,7 +2599,7 @@ i915_perf_open_ioctl_locked(struct drm_i915_private *dev_priv, > > * MI_REPORT_PERF_COUNT commands and so consider it a privileged op to > > * enable the OA unit by default. > > */ > > - if (IS_HASWELL(dev_priv) && specific_ctx) > > + if (IS_HASWELL(i915) && specific_ctx) > > privileged_op = false; > > /* Similar to perf's kernel.perf_paranoid_cpu sysctl option > > @@ -2620,7 +2620,7 @@ i915_perf_open_ioctl_locked(struct drm_i915_private *dev_priv, > > goto err_ctx; > > } > > - stream->dev_priv = dev_priv; > > + stream->i915 = i915; > > stream->ctx = specific_ctx; > > ret = i915_oa_stream_init(stream, param, props); > > @@ -2636,7 +2636,7 @@ i915_perf_open_ioctl_locked(struct drm_i915_private *dev_priv, > > goto err_flags; > > } > > - list_add(&stream->link, &dev_priv->perf.streams); > > + list_add(&stream->link, &i915->perf.streams); > > if (param->flags & I915_PERF_FLAG_FD_CLOEXEC) > > f_flags |= O_CLOEXEC; > > @@ -2668,15 +2668,15 @@ i915_perf_open_ioctl_locked(struct drm_i915_private *dev_priv, > > return ret; > > } > > -static u64 oa_exponent_to_ns(struct drm_i915_private *dev_priv, int exponent) > > +static u64 oa_exponent_to_ns(struct drm_i915_private *i915, int exponent) > > { > > return div64_u64(1000000000ULL * (2ULL << exponent), > > - 1000ULL * RUNTIME_INFO(dev_priv)->cs_timestamp_frequency_khz); > > + 1000ULL * RUNTIME_INFO(i915)->cs_timestamp_frequency_khz); > > } > > /** > > * read_properties_unlocked - validate + copy userspace stream open properties > > - * @dev_priv: i915 device instance > > + * @i915: i915 device instance > > * @uprops: The array of u64 key value pairs given by userspace > > * @n_props: The number of key value pairs expected in @uprops > > * @props: The stream configuration built up while validating properties > > @@ -2689,7 +2689,7 @@ static u64 oa_exponent_to_ns(struct drm_i915_private *dev_priv, int exponent) > > * we shouldn't validate or assume anything about ordering here. This doesn't > > * rule out defining new properties with ordering requirements in the future. > > */ > > -static int read_properties_unlocked(struct drm_i915_private *dev_priv, > > +static int read_properties_unlocked(struct drm_i915_private *i915, > > u64 __user *uprops, > > u32 n_props, > > struct perf_open_properties *props) > > @@ -2755,7 +2755,7 @@ static int read_properties_unlocked(struct drm_i915_private *dev_priv, > > value); > > return -EINVAL; > > } > > - if (!dev_priv->perf.oa.oa_formats[value].size) { > > + if (!i915->perf.oa.oa_formats[value].size) { > > DRM_DEBUG("Unsupported OA report format %llu\n", > > value); > > return -EINVAL; > > @@ -2776,7 +2776,7 @@ static int read_properties_unlocked(struct drm_i915_private *dev_priv, > > */ > > BUILD_BUG_ON(sizeof(oa_period) != 8); > > - oa_period = oa_exponent_to_ns(dev_priv, value); > > + oa_period = oa_exponent_to_ns(i915, value); > > /* This check is primarily to ensure that oa_period <= > > * UINT32_MAX (before passing to do_div which only > > @@ -2839,13 +2839,13 @@ static int read_properties_unlocked(struct drm_i915_private *dev_priv, > > int i915_perf_open_ioctl(struct drm_device *dev, void *data, > > struct drm_file *file) > > { > > - struct drm_i915_private *dev_priv = dev->dev_private; > > + struct drm_i915_private *i915 = dev->dev_private; > > struct drm_i915_perf_open_param *param = data; > > struct perf_open_properties props; > > u32 known_open_flags; > > int ret; > > - if (!dev_priv->perf.initialized) { > > + if (!i915->perf.initialized) { > > DRM_DEBUG("i915 perf interface not available for this system\n"); > > return -ENOTSUPP; > > } > > @@ -2858,124 +2858,124 @@ int i915_perf_open_ioctl(struct drm_device *dev, void *data, > > return -EINVAL; > > } > > - ret = read_properties_unlocked(dev_priv, > > + ret = read_properties_unlocked(i915, > > u64_to_user_ptr(param->properties_ptr), > > param->num_properties, > > &props); > > if (ret) > > return ret; > > - mutex_lock(&dev_priv->perf.lock); > > - ret = i915_perf_open_ioctl_locked(dev_priv, param, &props, file); > > - mutex_unlock(&dev_priv->perf.lock); > > + mutex_lock(&i915->perf.lock); > > + ret = i915_perf_open_ioctl_locked(i915, param, &props, file); > > + mutex_unlock(&i915->perf.lock); > > return ret; > > } > > /** > > * i915_perf_register - exposes i915-perf to userspace > > - * @dev_priv: i915 device instance > > + * @i915: i915 device instance > > * > > * In particular OA metric sets are advertised under a sysfs metrics/ > > * directory allowing userspace to enumerate valid IDs that can be > > * used to open an i915-perf stream. > > */ > > -void i915_perf_register(struct drm_i915_private *dev_priv) > > +void i915_perf_register(struct drm_i915_private *i915) > > { > > int ret; > > - if (!dev_priv->perf.initialized) > > + if (!i915->perf.initialized) > > return; > > /* To be sure we're synchronized with an attempted > > * i915_perf_open_ioctl(); considering that we register after > > * being exposed to userspace. > > */ > > - mutex_lock(&dev_priv->perf.lock); > > + mutex_lock(&i915->perf.lock); > > - dev_priv->perf.metrics_kobj = > > + i915->perf.metrics_kobj = > > kobject_create_and_add("metrics", > > - &dev_priv->drm.primary->kdev->kobj); > > - if (!dev_priv->perf.metrics_kobj) > > + &i915->drm.primary->kdev->kobj); > > + if (!i915->perf.metrics_kobj) > > goto exit; > > - sysfs_attr_init(&dev_priv->perf.oa.test_config.sysfs_metric_id.attr); > > - > > - if (INTEL_GEN(dev_priv) >= 11) { > > - i915_perf_load_test_config_icl(dev_priv); > > - } else if (IS_CANNONLAKE(dev_priv)) { > > - i915_perf_load_test_config_cnl(dev_priv); > > - } else if (IS_COFFEELAKE(dev_priv)) { > > - if (IS_CFL_GT2(dev_priv)) > > - i915_perf_load_test_config_cflgt2(dev_priv); > > - if (IS_CFL_GT3(dev_priv)) > > - i915_perf_load_test_config_cflgt3(dev_priv); > > - } else if (IS_GEMINILAKE(dev_priv)) { > > - i915_perf_load_test_config_glk(dev_priv); > > - } else if (IS_KABYLAKE(dev_priv)) { > > - if (IS_KBL_GT2(dev_priv)) > > - i915_perf_load_test_config_kblgt2(dev_priv); > > - else if (IS_KBL_GT3(dev_priv)) > > - i915_perf_load_test_config_kblgt3(dev_priv); > > - } else if (IS_BROXTON(dev_priv)) { > > - i915_perf_load_test_config_bxt(dev_priv); > > - } else if (IS_SKYLAKE(dev_priv)) { > > - if (IS_SKL_GT2(dev_priv)) > > - i915_perf_load_test_config_sklgt2(dev_priv); > > - else if (IS_SKL_GT3(dev_priv)) > > - i915_perf_load_test_config_sklgt3(dev_priv); > > - else if (IS_SKL_GT4(dev_priv)) > > - i915_perf_load_test_config_sklgt4(dev_priv); > > - } else if (IS_CHERRYVIEW(dev_priv)) { > > - i915_perf_load_test_config_chv(dev_priv); > > - } else if (IS_BROADWELL(dev_priv)) { > > - i915_perf_load_test_config_bdw(dev_priv); > > - } else if (IS_HASWELL(dev_priv)) { > > - i915_perf_load_test_config_hsw(dev_priv); > > -} > > - > > - if (dev_priv->perf.oa.test_config.id == 0) > > + sysfs_attr_init(&i915->perf.oa.test_config.sysfs_metric_id.attr); > > + > > + if (INTEL_GEN(i915) >= 11) { > > + i915_perf_load_test_config_icl(i915); > > + } else if (IS_CANNONLAKE(i915)) { > > + i915_perf_load_test_config_cnl(i915); > > + } else if (IS_COFFEELAKE(i915)) { > > + if (IS_CFL_GT2(i915)) > > + i915_perf_load_test_config_cflgt2(i915); > > + if (IS_CFL_GT3(i915)) > > + i915_perf_load_test_config_cflgt3(i915); > > + } else if (IS_GEMINILAKE(i915)) { > > + i915_perf_load_test_config_glk(i915); > > + } else if (IS_KABYLAKE(i915)) { > > + if (IS_KBL_GT2(i915)) > > + i915_perf_load_test_config_kblgt2(i915); > > + else if (IS_KBL_GT3(i915)) > > + i915_perf_load_test_config_kblgt3(i915); > > + } else if (IS_BROXTON(i915)) { > > + i915_perf_load_test_config_bxt(i915); > > + } else if (IS_SKYLAKE(i915)) { > > + if (IS_SKL_GT2(i915)) > > + i915_perf_load_test_config_sklgt2(i915); > > + else if (IS_SKL_GT3(i915)) > > + i915_perf_load_test_config_sklgt3(i915); > > + else if (IS_SKL_GT4(i915)) > > + i915_perf_load_test_config_sklgt4(i915); > > + } else if (IS_CHERRYVIEW(i915)) { > > + i915_perf_load_test_config_chv(i915); > > + } else if (IS_BROADWELL(i915)) { > > + i915_perf_load_test_config_bdw(i915); > > + } else if (IS_HASWELL(i915)) { > > + i915_perf_load_test_config_hsw(i915); > > +} > > + > > + if (i915->perf.oa.test_config.id == 0) > > goto sysfs_error; > > - ret = sysfs_create_group(dev_priv->perf.metrics_kobj, > > - &dev_priv->perf.oa.test_config.sysfs_metric); > > + ret = sysfs_create_group(i915->perf.metrics_kobj, > > + &i915->perf.oa.test_config.sysfs_metric); > > if (ret) > > goto sysfs_error; > > - atomic_set(&dev_priv->perf.oa.test_config.ref_count, 1); > > + atomic_set(&i915->perf.oa.test_config.ref_count, 1); > > goto exit; > > sysfs_error: > > - kobject_put(dev_priv->perf.metrics_kobj); > > - dev_priv->perf.metrics_kobj = NULL; > > + kobject_put(i915->perf.metrics_kobj); > > + i915->perf.metrics_kobj = NULL; > > exit: > > - mutex_unlock(&dev_priv->perf.lock); > > + mutex_unlock(&i915->perf.lock); > > } > > /** > > * i915_perf_unregister - hide i915-perf from userspace > > - * @dev_priv: i915 device instance > > + * @i915: i915 device instance > > * > > * i915-perf state cleanup is split up into an 'unregister' and > > * 'deinit' phase where the interface is first hidden from > > * userspace by i915_perf_unregister() before cleaning up > > * remaining state in i915_perf_fini(). > > */ > > -void i915_perf_unregister(struct drm_i915_private *dev_priv) > > +void i915_perf_unregister(struct drm_i915_private *i915) > > { > > - if (!dev_priv->perf.metrics_kobj) > > + if (!i915->perf.metrics_kobj) > > return; > > - sysfs_remove_group(dev_priv->perf.metrics_kobj, > > - &dev_priv->perf.oa.test_config.sysfs_metric); > > + sysfs_remove_group(i915->perf.metrics_kobj, > > + &i915->perf.oa.test_config.sysfs_metric); > > - kobject_put(dev_priv->perf.metrics_kobj); > > - dev_priv->perf.metrics_kobj = NULL; > > + kobject_put(i915->perf.metrics_kobj); > > + i915->perf.metrics_kobj = NULL; > > } > > -static bool gen8_is_valid_flex_addr(struct drm_i915_private *dev_priv, u32 addr) > > +static bool gen8_is_valid_flex_addr(struct drm_i915_private *i915, u32 addr) > > { > > static const i915_reg_t flex_eu_regs[] = { > > EU_PERF_CNTL0, > > @@ -2995,7 +2995,7 @@ static bool gen8_is_valid_flex_addr(struct drm_i915_private *dev_priv, u32 addr) > > return false; > > } > > -static bool gen7_is_valid_b_counter_addr(struct drm_i915_private *dev_priv, u32 addr) > > +static bool gen7_is_valid_b_counter_addr(struct drm_i915_private *i915, u32 addr) > > { > > return (addr >= i915_mmio_reg_offset(OASTARTTRIG1) && > > addr <= i915_mmio_reg_offset(OASTARTTRIG8)) || > > @@ -3005,7 +3005,7 @@ static bool gen7_is_valid_b_counter_addr(struct drm_i915_private *dev_priv, u32 > > addr <= i915_mmio_reg_offset(OACEC7_1)); > > } > > -static bool gen7_is_valid_mux_addr(struct drm_i915_private *dev_priv, u32 addr) > > +static bool gen7_is_valid_mux_addr(struct drm_i915_private *i915, u32 addr) > > { > > return addr == i915_mmio_reg_offset(HALF_SLICE_CHICKEN2) || > > (addr >= i915_mmio_reg_offset(MICRO_BP0_0) && > > @@ -3016,34 +3016,34 @@ static bool gen7_is_valid_mux_addr(struct drm_i915_private *dev_priv, u32 addr) > > addr <= i915_mmio_reg_offset(OA_PERFMATRIX_HI)); > > } > > -static bool gen8_is_valid_mux_addr(struct drm_i915_private *dev_priv, u32 addr) > > +static bool gen8_is_valid_mux_addr(struct drm_i915_private *i915, u32 addr) > > { > > - return gen7_is_valid_mux_addr(dev_priv, addr) || > > + return gen7_is_valid_mux_addr(i915, addr) || > > addr == i915_mmio_reg_offset(WAIT_FOR_RC6_EXIT) || > > (addr >= i915_mmio_reg_offset(RPM_CONFIG0) && > > addr <= i915_mmio_reg_offset(NOA_CONFIG(8))); > > } > > -static bool gen10_is_valid_mux_addr(struct drm_i915_private *dev_priv, u32 addr) > > +static bool gen10_is_valid_mux_addr(struct drm_i915_private *i915, u32 addr) > > { > > - return gen8_is_valid_mux_addr(dev_priv, addr) || > > + return gen8_is_valid_mux_addr(i915, addr) || > > addr == i915_mmio_reg_offset(GEN10_NOA_WRITE_HIGH) || > > (addr >= i915_mmio_reg_offset(OA_PERFCNT3_LO) && > > addr <= i915_mmio_reg_offset(OA_PERFCNT4_HI)); > > } > > -static bool hsw_is_valid_mux_addr(struct drm_i915_private *dev_priv, u32 addr) > > +static bool hsw_is_valid_mux_addr(struct drm_i915_private *i915, u32 addr) > > { > > - return gen7_is_valid_mux_addr(dev_priv, addr) || > > + return gen7_is_valid_mux_addr(i915, addr) || > > (addr >= 0x25100 && addr <= 0x2FF90) || > > (addr >= i915_mmio_reg_offset(HSW_MBVID2_NOA0) && > > addr <= i915_mmio_reg_offset(HSW_MBVID2_NOA9)) || > > addr == i915_mmio_reg_offset(HSW_MBVID2_MISR0); > > } > > -static bool chv_is_valid_mux_addr(struct drm_i915_private *dev_priv, u32 addr) > > +static bool chv_is_valid_mux_addr(struct drm_i915_private *i915, u32 addr) > > { > > - return gen7_is_valid_mux_addr(dev_priv, addr) || > > + return gen7_is_valid_mux_addr(i915, addr) || > > (addr >= 0x182300 && addr <= 0x1823A4); > > } > > @@ -3066,8 +3066,8 @@ static u32 mask_reg_value(u32 reg, u32 val) > > return val; > > } > > -static struct i915_oa_reg *alloc_oa_regs(struct drm_i915_private *dev_priv, > > - bool (*is_valid)(struct drm_i915_private *dev_priv, u32 addr), > > +static struct i915_oa_reg *alloc_oa_regs(struct drm_i915_private *i915, > > + bool (*is_valid)(struct drm_i915_private *i915, u32 addr), > > u32 __user *regs, > > u32 n_regs) > > { > > @@ -3097,7 +3097,7 @@ static struct i915_oa_reg *alloc_oa_regs(struct drm_i915_private *dev_priv, > > if (err) > > goto addr_err; > > - if (!is_valid(dev_priv, addr)) { > > + if (!is_valid(i915, addr)) { > > DRM_DEBUG("Invalid oa_reg address: %X\n", addr); > > err = -EINVAL; > > goto addr_err; > > @@ -3130,7 +3130,7 @@ static ssize_t show_dynamic_id(struct device *dev, > > return sprintf(buf, "%d\n", oa_config->id); > > } > > -static int create_dynamic_oa_sysfs_entry(struct drm_i915_private *dev_priv, > > +static int create_dynamic_oa_sysfs_entry(struct drm_i915_private *i915, > > struct i915_oa_config *oa_config) > > { > > sysfs_attr_init(&oa_config->sysfs_metric_id.attr); > > @@ -3145,7 +3145,7 @@ static int create_dynamic_oa_sysfs_entry(struct drm_i915_private *dev_priv, > > oa_config->sysfs_metric.name = oa_config->uuid; > > oa_config->sysfs_metric.attrs = oa_config->attrs; > > - return sysfs_create_group(dev_priv->perf.metrics_kobj, > > + return sysfs_create_group(i915->perf.metrics_kobj, > > &oa_config->sysfs_metric); > > } > > @@ -3165,17 +3165,17 @@ static int create_dynamic_oa_sysfs_entry(struct drm_i915_private *dev_priv, > > int i915_perf_add_config_ioctl(struct drm_device *dev, void *data, > > struct drm_file *file) > > { > > - struct drm_i915_private *dev_priv = dev->dev_private; > > + struct drm_i915_private *i915 = dev->dev_private; > > struct drm_i915_perf_oa_config *args = data; > > struct i915_oa_config *oa_config, *tmp; > > int err, id; > > - if (!dev_priv->perf.initialized) { > > + if (!i915->perf.initialized) { > > DRM_DEBUG("i915 perf interface not available for this system\n"); > > return -ENOTSUPP; > > } > > - if (!dev_priv->perf.metrics_kobj) { > > + if (!i915->perf.metrics_kobj) { > > DRM_DEBUG("OA metrics weren't advertised via sysfs\n"); > > return -EINVAL; > > } > > @@ -3213,8 +3213,8 @@ int i915_perf_add_config_ioctl(struct drm_device *dev, void *data, > > oa_config->mux_regs_len = args->n_mux_regs; > > oa_config->mux_regs = > > - alloc_oa_regs(dev_priv, > > - dev_priv->perf.oa.ops.is_valid_mux_reg, > > + alloc_oa_regs(i915, > > + i915->perf.oa.ops.is_valid_mux_reg, > > u64_to_user_ptr(args->mux_regs_ptr), > > args->n_mux_regs); > > @@ -3226,8 +3226,8 @@ int i915_perf_add_config_ioctl(struct drm_device *dev, void *data, > > oa_config->b_counter_regs_len = args->n_boolean_regs; > > oa_config->b_counter_regs = > > - alloc_oa_regs(dev_priv, > > - dev_priv->perf.oa.ops.is_valid_b_counter_reg, > > + alloc_oa_regs(i915, > > + i915->perf.oa.ops.is_valid_b_counter_reg, > > u64_to_user_ptr(args->boolean_regs_ptr), > > args->n_boolean_regs); > > @@ -3237,7 +3237,7 @@ int i915_perf_add_config_ioctl(struct drm_device *dev, void *data, > > goto reg_err; > > } > > - if (INTEL_GEN(dev_priv) < 8) { > > + if (INTEL_GEN(i915) < 8) { > > if (args->n_flex_regs != 0) { > > err = -EINVAL; > > goto reg_err; > > @@ -3245,8 +3245,8 @@ int i915_perf_add_config_ioctl(struct drm_device *dev, void *data, > > } else { > > oa_config->flex_regs_len = args->n_flex_regs; > > oa_config->flex_regs = > > - alloc_oa_regs(dev_priv, > > - dev_priv->perf.oa.ops.is_valid_flex_reg, > > + alloc_oa_regs(i915, > > + i915->perf.oa.ops.is_valid_flex_reg, > > u64_to_user_ptr(args->flex_regs_ptr), > > args->n_flex_regs); > > @@ -3257,14 +3257,14 @@ int i915_perf_add_config_ioctl(struct drm_device *dev, void *data, > > } > > } > > - err = mutex_lock_interruptible(&dev_priv->perf.metrics_lock); > > + err = mutex_lock_interruptible(&i915->perf.metrics_lock); > > if (err) > > goto reg_err; > > /* We shouldn't have too many configs, so this iteration shouldn't be > > * too costly. > > */ > > - idr_for_each_entry(&dev_priv->perf.metrics_idr, tmp, id) { > > + idr_for_each_entry(&i915->perf.metrics_idr, tmp, id) { > > if (!strcmp(tmp->uuid, oa_config->uuid)) { > > DRM_DEBUG("OA config already exists with this uuid\n"); > > err = -EADDRINUSE; > > @@ -3272,14 +3272,14 @@ int i915_perf_add_config_ioctl(struct drm_device *dev, void *data, > > } > > } > > - err = create_dynamic_oa_sysfs_entry(dev_priv, oa_config); > > + err = create_dynamic_oa_sysfs_entry(i915, oa_config); > > if (err) { > > DRM_DEBUG("Failed to create sysfs entry for OA config\n"); > > goto sysfs_err; > > } > > /* Config id 0 is invalid, id 1 for kernel stored test config. */ > > - oa_config->id = idr_alloc(&dev_priv->perf.metrics_idr, > > + oa_config->id = idr_alloc(&i915->perf.metrics_idr, > > oa_config, 2, > > 0, GFP_KERNEL); > > if (oa_config->id < 0) { > > @@ -3288,16 +3288,16 @@ int i915_perf_add_config_ioctl(struct drm_device *dev, void *data, > > goto sysfs_err; > > } > > - mutex_unlock(&dev_priv->perf.metrics_lock); > > + mutex_unlock(&i915->perf.metrics_lock); > > DRM_DEBUG("Added config %s id=%i\n", oa_config->uuid, oa_config->id); > > return oa_config->id; > > sysfs_err: > > - mutex_unlock(&dev_priv->perf.metrics_lock); > > + mutex_unlock(&i915->perf.metrics_lock); > > reg_err: > > - put_oa_config(dev_priv, oa_config); > > + put_oa_config(i915, oa_config); > > DRM_DEBUG("Failed to add new OA config\n"); > > return err; > > } > > @@ -3316,12 +3316,12 @@ int i915_perf_add_config_ioctl(struct drm_device *dev, void *data, > > int i915_perf_remove_config_ioctl(struct drm_device *dev, void *data, > > struct drm_file *file) > > { > > - struct drm_i915_private *dev_priv = dev->dev_private; > > + struct drm_i915_private *i915 = dev->dev_private; > > u64 *arg = data; > > struct i915_oa_config *oa_config; > > int ret; > > - if (!dev_priv->perf.initialized) { > > + if (!i915->perf.initialized) { > > DRM_DEBUG("i915 perf interface not available for this system\n"); > > return -ENOTSUPP; > > } > > @@ -3331,11 +3331,11 @@ int i915_perf_remove_config_ioctl(struct drm_device *dev, void *data, > > return -EACCES; > > } > > - ret = mutex_lock_interruptible(&dev_priv->perf.metrics_lock); > > + ret = mutex_lock_interruptible(&i915->perf.metrics_lock); > > if (ret) > > goto lock_err; > > - oa_config = idr_find(&dev_priv->perf.metrics_idr, *arg); > > + oa_config = idr_find(&i915->perf.metrics_idr, *arg); > > if (!oa_config) { > > DRM_DEBUG("Failed to remove unknown OA config\n"); > > ret = -ENOENT; > > @@ -3344,17 +3344,17 @@ int i915_perf_remove_config_ioctl(struct drm_device *dev, void *data, > > GEM_BUG_ON(*arg != oa_config->id); > > - sysfs_remove_group(dev_priv->perf.metrics_kobj, > > + sysfs_remove_group(i915->perf.metrics_kobj, > > &oa_config->sysfs_metric); > > - idr_remove(&dev_priv->perf.metrics_idr, *arg); > > + idr_remove(&i915->perf.metrics_idr, *arg); > > DRM_DEBUG("Removed config %s id=%i\n", oa_config->uuid, oa_config->id); > > - put_oa_config(dev_priv, oa_config); > > + put_oa_config(i915, oa_config); > > config_err: > > - mutex_unlock(&dev_priv->perf.metrics_lock); > > + mutex_unlock(&i915->perf.metrics_lock); > > lock_err: > > return ret; > > } > > @@ -3403,135 +3403,135 @@ static struct ctl_table dev_root[] = { > > /** > > * i915_perf_init - initialize i915-perf state on module load > > - * @dev_priv: i915 device instance > > + * @i915: i915 device instance > > * > > * Initializes i915-perf state without exposing anything to userspace. > > * > > * Note: i915-perf initialization is split into an 'init' and 'register' > > * phase with the i915_perf_register() exposing state to userspace. > > */ > > -void i915_perf_init(struct drm_i915_private *dev_priv) > > +void i915_perf_init(struct drm_i915_private *i915) > > { > > - if (IS_HASWELL(dev_priv)) { > > - dev_priv->perf.oa.ops.is_valid_b_counter_reg = > > + if (IS_HASWELL(i915)) { > > + i915->perf.oa.ops.is_valid_b_counter_reg = > > gen7_is_valid_b_counter_addr; > > - dev_priv->perf.oa.ops.is_valid_mux_reg = > > + i915->perf.oa.ops.is_valid_mux_reg = > > hsw_is_valid_mux_addr; > > - dev_priv->perf.oa.ops.is_valid_flex_reg = NULL; > > - dev_priv->perf.oa.ops.enable_metric_set = hsw_enable_metric_set; > > - dev_priv->perf.oa.ops.disable_metric_set = hsw_disable_metric_set; > > - dev_priv->perf.oa.ops.oa_enable = gen7_oa_enable; > > - dev_priv->perf.oa.ops.oa_disable = gen7_oa_disable; > > - dev_priv->perf.oa.ops.read = gen7_oa_read; > > - dev_priv->perf.oa.ops.oa_hw_tail_read = > > + i915->perf.oa.ops.is_valid_flex_reg = NULL; > > + i915->perf.oa.ops.enable_metric_set = hsw_enable_metric_set; > > + i915->perf.oa.ops.disable_metric_set = hsw_disable_metric_set; > > + i915->perf.oa.ops.oa_enable = gen7_oa_enable; > > + i915->perf.oa.ops.oa_disable = gen7_oa_disable; > > + i915->perf.oa.ops.read = gen7_oa_read; > > + i915->perf.oa.ops.oa_hw_tail_read = > > gen7_oa_hw_tail_read; > > - dev_priv->perf.oa.oa_formats = hsw_oa_formats; > > - } else if (HAS_LOGICAL_RING_CONTEXTS(dev_priv)) { > > + i915->perf.oa.oa_formats = hsw_oa_formats; > > + } else if (HAS_LOGICAL_RING_CONTEXTS(i915)) { > > /* Note: that although we could theoretically also support the > > * legacy ringbuffer mode on BDW (and earlier iterations of > > * this driver, before upstreaming did this) it didn't seem > > * worth the complexity to maintain now that BDW+ enable > > * execlist mode by default. > > */ > > - dev_priv->perf.oa.oa_formats = gen8_plus_oa_formats; > > + i915->perf.oa.oa_formats = gen8_plus_oa_formats; > > - dev_priv->perf.oa.ops.oa_enable = gen8_oa_enable; > > - dev_priv->perf.oa.ops.oa_disable = gen8_oa_disable; > > - dev_priv->perf.oa.ops.read = gen8_oa_read; > > - dev_priv->perf.oa.ops.oa_hw_tail_read = gen8_oa_hw_tail_read; > > + i915->perf.oa.ops.oa_enable = gen8_oa_enable; > > + i915->perf.oa.ops.oa_disable = gen8_oa_disable; > > + i915->perf.oa.ops.read = gen8_oa_read; > > + i915->perf.oa.ops.oa_hw_tail_read = gen8_oa_hw_tail_read; > > - if (IS_GEN_RANGE(dev_priv, 8, 9)) { > > - dev_priv->perf.oa.ops.is_valid_b_counter_reg = > > + if (IS_GEN_RANGE(i915, 8, 9)) { > > + i915->perf.oa.ops.is_valid_b_counter_reg = > > gen7_is_valid_b_counter_addr; > > - dev_priv->perf.oa.ops.is_valid_mux_reg = > > + i915->perf.oa.ops.is_valid_mux_reg = > > gen8_is_valid_mux_addr; > > - dev_priv->perf.oa.ops.is_valid_flex_reg = > > + i915->perf.oa.ops.is_valid_flex_reg = > > gen8_is_valid_flex_addr; > > - if (IS_CHERRYVIEW(dev_priv)) { > > - dev_priv->perf.oa.ops.is_valid_mux_reg = > > + if (IS_CHERRYVIEW(i915)) { > > + i915->perf.oa.ops.is_valid_mux_reg = > > chv_is_valid_mux_addr; > > } > > - dev_priv->perf.oa.ops.enable_metric_set = gen8_enable_metric_set; > > - dev_priv->perf.oa.ops.disable_metric_set = gen8_disable_metric_set; > > + i915->perf.oa.ops.enable_metric_set = gen8_enable_metric_set; > > + i915->perf.oa.ops.disable_metric_set = gen8_disable_metric_set; > > - if (IS_GEN(dev_priv, 8)) { > > - dev_priv->perf.oa.ctx_oactxctrl_offset = 0x120; > > - dev_priv->perf.oa.ctx_flexeu0_offset = 0x2ce; > > + if (IS_GEN(i915, 8)) { > > + i915->perf.oa.ctx_oactxctrl_offset = 0x120; > > + i915->perf.oa.ctx_flexeu0_offset = 0x2ce; > > - dev_priv->perf.oa.gen8_valid_ctx_bit = (1<<25); > > + i915->perf.oa.gen8_valid_ctx_bit = (1<<25); > > } else { > > - dev_priv->perf.oa.ctx_oactxctrl_offset = 0x128; > > - dev_priv->perf.oa.ctx_flexeu0_offset = 0x3de; > > + i915->perf.oa.ctx_oactxctrl_offset = 0x128; > > + i915->perf.oa.ctx_flexeu0_offset = 0x3de; > > - dev_priv->perf.oa.gen8_valid_ctx_bit = (1<<16); > > + i915->perf.oa.gen8_valid_ctx_bit = (1<<16); > > } > > - } else if (IS_GEN_RANGE(dev_priv, 10, 11)) { > > - dev_priv->perf.oa.ops.is_valid_b_counter_reg = > > + } else if (IS_GEN_RANGE(i915, 10, 11)) { > > + i915->perf.oa.ops.is_valid_b_counter_reg = > > gen7_is_valid_b_counter_addr; > > - dev_priv->perf.oa.ops.is_valid_mux_reg = > > + i915->perf.oa.ops.is_valid_mux_reg = > > gen10_is_valid_mux_addr; > > - dev_priv->perf.oa.ops.is_valid_flex_reg = > > + i915->perf.oa.ops.is_valid_flex_reg = > > gen8_is_valid_flex_addr; > > - dev_priv->perf.oa.ops.enable_metric_set = gen8_enable_metric_set; > > - dev_priv->perf.oa.ops.disable_metric_set = gen10_disable_metric_set; > > + i915->perf.oa.ops.enable_metric_set = gen8_enable_metric_set; > > + i915->perf.oa.ops.disable_metric_set = gen10_disable_metric_set; > > - dev_priv->perf.oa.ctx_oactxctrl_offset = 0x128; > > - dev_priv->perf.oa.ctx_flexeu0_offset = 0x3de; > > + i915->perf.oa.ctx_oactxctrl_offset = 0x128; > > + i915->perf.oa.ctx_flexeu0_offset = 0x3de; > > - dev_priv->perf.oa.gen8_valid_ctx_bit = (1<<16); > > + i915->perf.oa.gen8_valid_ctx_bit = (1<<16); > > } > > } > > - if (dev_priv->perf.oa.ops.enable_metric_set) { > > - hrtimer_init(&dev_priv->perf.oa.poll_check_timer, > > + if (i915->perf.oa.ops.enable_metric_set) { > > + hrtimer_init(&i915->perf.oa.poll_check_timer, > > CLOCK_MONOTONIC, HRTIMER_MODE_REL); > > - dev_priv->perf.oa.poll_check_timer.function = oa_poll_check_timer_cb; > > - init_waitqueue_head(&dev_priv->perf.oa.poll_wq); > > + i915->perf.oa.poll_check_timer.function = oa_poll_check_timer_cb; > > + init_waitqueue_head(&i915->perf.oa.poll_wq); > > - INIT_LIST_HEAD(&dev_priv->perf.streams); > > - mutex_init(&dev_priv->perf.lock); > > - spin_lock_init(&dev_priv->perf.oa.oa_buffer.ptr_lock); > > + INIT_LIST_HEAD(&i915->perf.streams); > > + mutex_init(&i915->perf.lock); > > + spin_lock_init(&i915->perf.oa.oa_buffer.ptr_lock); > > oa_sample_rate_hard_limit = 1000 * > > - (RUNTIME_INFO(dev_priv)->cs_timestamp_frequency_khz / 2); > > - dev_priv->perf.sysctl_header = register_sysctl_table(dev_root); > > + (RUNTIME_INFO(i915)->cs_timestamp_frequency_khz / 2); > > + i915->perf.sysctl_header = register_sysctl_table(dev_root); > > - mutex_init(&dev_priv->perf.metrics_lock); > > - idr_init(&dev_priv->perf.metrics_idr); > > + mutex_init(&i915->perf.metrics_lock); > > + idr_init(&i915->perf.metrics_idr); > > - dev_priv->perf.initialized = true; > > + i915->perf.initialized = true; > > } > > } > > static int destroy_config(int id, void *p, void *data) > > { > > - struct drm_i915_private *dev_priv = data; > > + struct drm_i915_private *i915 = data; > > struct i915_oa_config *oa_config = p; > > - put_oa_config(dev_priv, oa_config); > > + put_oa_config(i915, oa_config); > > return 0; > > } > > /** > > * i915_perf_fini - Counter part to i915_perf_init() > > - * @dev_priv: i915 device instance > > + * @i915: i915 device instance > > */ > > -void i915_perf_fini(struct drm_i915_private *dev_priv) > > +void i915_perf_fini(struct drm_i915_private *i915) > > { > > - if (!dev_priv->perf.initialized) > > + if (!i915->perf.initialized) > > return; > > - idr_for_each(&dev_priv->perf.metrics_idr, destroy_config, dev_priv); > > - idr_destroy(&dev_priv->perf.metrics_idr); > > + idr_for_each(&i915->perf.metrics_idr, destroy_config, i915); > > + idr_destroy(&i915->perf.metrics_idr); > > - unregister_sysctl_table(dev_priv->perf.sysctl_header); > > + unregister_sysctl_table(i915->perf.sysctl_header); > > - memset(&dev_priv->perf.oa.ops, 0, sizeof(dev_priv->perf.oa.ops)); > > + memset(&i915->perf.oa.ops, 0, sizeof(i915->perf.oa.ops)); > > - dev_priv->perf.initialized = false; > > + i915->perf.initialized = false; > > } > > diff --git a/drivers/gpu/drm/i915/i915_pmu.c b/drivers/gpu/drm/i915/i915_pmu.c > > index eb9c0e0e545c..c4861c12100c 100644 > > --- a/drivers/gpu/drm/i915/i915_pmu.c > > +++ b/drivers/gpu/drm/i915/i915_pmu.c > > @@ -159,24 +159,24 @@ add_sample(struct i915_pmu_sample *sample, u32 val) > > } > > static void > > -engines_sample(struct drm_i915_private *dev_priv, unsigned int period_ns) > > +engines_sample(struct drm_i915_private *i915, unsigned int period_ns) > > { > > struct intel_engine_cs *engine; > > enum intel_engine_id id; > > intel_wakeref_t wakeref; > > unsigned long flags; > > - if ((dev_priv->pmu.enable & ENGINE_SAMPLE_MASK) == 0) > > + if ((i915->pmu.enable & ENGINE_SAMPLE_MASK) == 0) > > return; > > wakeref = 0; > > - if (READ_ONCE(dev_priv->gt.awake)) > > - wakeref = intel_runtime_pm_get_if_in_use(dev_priv); > > + if (READ_ONCE(i915->gt.awake)) > > + wakeref = intel_runtime_pm_get_if_in_use(i915); > > if (!wakeref) > > return; > > - spin_lock_irqsave(&dev_priv->uncore.lock, flags); > > - for_each_engine(engine, dev_priv, id) { > > + spin_lock_irqsave(&i915->uncore.lock, flags); > > + for_each_engine(engine, i915, id) { > > struct intel_engine_pmu *pmu = &engine->pmu; > > bool busy; > > u32 val; > > @@ -205,9 +205,9 @@ engines_sample(struct drm_i915_private *dev_priv, unsigned int period_ns) > > if (busy) > > add_sample(&pmu->sample[I915_SAMPLE_BUSY], period_ns); > > } > > - spin_unlock_irqrestore(&dev_priv->uncore.lock, flags); > > + spin_unlock_irqrestore(&i915->uncore.lock, flags); > > - intel_runtime_pm_put(dev_priv, wakeref); > > + intel_runtime_pm_put(i915, wakeref); > > } > > static void > > @@ -217,33 +217,33 @@ add_sample_mult(struct i915_pmu_sample *sample, u32 val, u32 mul) > > } > > static void > > -frequency_sample(struct drm_i915_private *dev_priv, unsigned int period_ns) > > +frequency_sample(struct drm_i915_private *i915, unsigned int period_ns) > > { > > - if (dev_priv->pmu.enable & > > + if (i915->pmu.enable & > > config_enabled_mask(I915_PMU_ACTUAL_FREQUENCY)) { > > u32 val; > > - val = dev_priv->gt_pm.rps.cur_freq; > > - if (dev_priv->gt.awake) { > > + val = i915->gt_pm.rps.cur_freq; > > + if (i915->gt.awake) { > > intel_wakeref_t wakeref; > > - with_intel_runtime_pm_if_in_use(dev_priv, wakeref) { > > - val = intel_uncore_read_notrace(&dev_priv->uncore, > > + with_intel_runtime_pm_if_in_use(i915, wakeref) { > > + val = intel_uncore_read_notrace(&i915->uncore, > > GEN6_RPSTAT1); > > - val = intel_get_cagf(dev_priv, val); > > + val = intel_get_cagf(i915, val); > > } > > } > > - add_sample_mult(&dev_priv->pmu.sample[__I915_SAMPLE_FREQ_ACT], > > - intel_gpu_freq(dev_priv, val), > > + add_sample_mult(&i915->pmu.sample[__I915_SAMPLE_FREQ_ACT], > > + intel_gpu_freq(i915, val), > > period_ns / 1000); > > } > > - if (dev_priv->pmu.enable & > > + if (i915->pmu.enable & > > config_enabled_mask(I915_PMU_REQUESTED_FREQUENCY)) { > > - add_sample_mult(&dev_priv->pmu.sample[__I915_SAMPLE_FREQ_REQ], > > - intel_gpu_freq(dev_priv, > > - dev_priv->gt_pm.rps.cur_freq), > > + add_sample_mult(&i915->pmu.sample[__I915_SAMPLE_FREQ_REQ], > > + intel_gpu_freq(i915, > > + i915->gt_pm.rps.cur_freq), > > period_ns / 1000); > > } > > } > > diff --git a/drivers/gpu/drm/i915/i915_query.c b/drivers/gpu/drm/i915/i915_query.c > > index 7b7016171057..87606362b56f 100644 > > --- a/drivers/gpu/drm/i915/i915_query.c > > +++ b/drivers/gpu/drm/i915/i915_query.c > > @@ -31,10 +31,10 @@ static int copy_query_item(void *query_hdr, size_t query_sz, > > return 0; > > } > > -static int query_topology_info(struct drm_i915_private *dev_priv, > > +static int query_topology_info(struct drm_i915_private *i915, > > struct drm_i915_query_item *query_item) > > { > > - const struct sseu_dev_info *sseu = &RUNTIME_INFO(dev_priv)->sseu; > > + const struct sseu_dev_info *sseu = &RUNTIME_INFO(i915)->sseu; > > struct drm_i915_query_topology_info topo; > > u32 slice_length, subslice_length, eu_length, total_length; > > u8 subslice_stride = GEN_SSEU_STRIDE(sseu->max_subslices); > > @@ -143,7 +143,7 @@ query_engine_info(struct drm_i915_private *i915, > > return len; > > } > > -static int (* const i915_query_funcs[])(struct drm_i915_private *dev_priv, > > +static int (* const i915_query_funcs[])(struct drm_i915_private *i915, > > struct drm_i915_query_item *query_item) = { > > query_topology_info, > > query_engine_info, > > @@ -151,7 +151,7 @@ static int (* const i915_query_funcs[])(struct drm_i915_private *dev_priv, > > int i915_query_ioctl(struct drm_device *dev, void *data, struct drm_file *file) > > { > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > struct drm_i915_query *args = data; > > struct drm_i915_query_item __user *user_item_ptr = > > u64_to_user_ptr(args->items_ptr); > > @@ -180,7 +180,7 @@ int i915_query_ioctl(struct drm_device *dev, void *data, struct drm_file *file) > > if (func_idx < ARRAY_SIZE(i915_query_funcs)) { > > func_idx = array_index_nospec(func_idx, > > ARRAY_SIZE(i915_query_funcs)); > > - ret = i915_query_funcs[func_idx](dev_priv, &item); > > + ret = i915_query_funcs[func_idx](i915, &item); > > } > > /* Only write the length back to userspace if they differ. */ > > diff --git a/drivers/gpu/drm/i915/i915_reg.h b/drivers/gpu/drm/i915/i915_reg.h > > index edf9f93934a1..b522970bdf74 100644 > > --- a/drivers/gpu/drm/i915/i915_reg.h > > +++ b/drivers/gpu/drm/i915/i915_reg.h > > @@ -205,7 +205,7 @@ static inline bool i915_mmio_reg_valid(i915_reg_t reg) > > #define VLV_MIPI_BASE VLV_DISPLAY_BASE > > #define BXT_MIPI_BASE 0x60000 > > -#define DISPLAY_MMIO_BASE(dev_priv) (INTEL_INFO(dev_priv)->display_mmio_offset) > > +#define DISPLAY_MMIO_BASE(i915) (INTEL_INFO(i915)->display_mmio_offset) > > /* > > * Given the first two numbers __a and __b of arbitrarily many evenly spaced > > @@ -247,15 +247,15 @@ static inline bool i915_mmio_reg_valid(i915_reg_t reg) > > * Device info offset array based helpers for groups of registers with unevenly > > * spaced base offsets. > > */ > > -#define _MMIO_PIPE2(pipe, reg) _MMIO(INTEL_INFO(dev_priv)->pipe_offsets[pipe] - \ > > - INTEL_INFO(dev_priv)->pipe_offsets[PIPE_A] + (reg) + \ > > - DISPLAY_MMIO_BASE(dev_priv)) > > -#define _MMIO_TRANS2(pipe, reg) _MMIO(INTEL_INFO(dev_priv)->trans_offsets[(pipe)] - \ > > - INTEL_INFO(dev_priv)->trans_offsets[TRANSCODER_A] + (reg) + \ > > - DISPLAY_MMIO_BASE(dev_priv)) > > -#define _CURSOR2(pipe, reg) _MMIO(INTEL_INFO(dev_priv)->cursor_offsets[(pipe)] - \ > > - INTEL_INFO(dev_priv)->cursor_offsets[PIPE_A] + (reg) + \ > > - DISPLAY_MMIO_BASE(dev_priv)) > > +#define _MMIO_PIPE2(pipe, reg) _MMIO(INTEL_INFO(i915)->pipe_offsets[pipe] - \ > > + INTEL_INFO(i915)->pipe_offsets[PIPE_A] + (reg) + \ > > + DISPLAY_MMIO_BASE(i915)) > > +#define _MMIO_TRANS2(pipe, reg) _MMIO(INTEL_INFO(i915)->trans_offsets[(pipe)] - \ > > + INTEL_INFO(i915)->trans_offsets[TRANSCODER_A] + (reg) + \ > > + DISPLAY_MMIO_BASE(i915)) > > +#define _CURSOR2(pipe, reg) _MMIO(INTEL_INFO(i915)->cursor_offsets[(pipe)] - \ > > + INTEL_INFO(i915)->cursor_offsets[PIPE_A] + (reg) + \ > > + DISPLAY_MMIO_BASE(i915)) > > #define __MASKED_FIELD(mask, value) ((mask) << 16 | (value)) > > #define _MASKED_FIELD(mask, value) ({ \ > > @@ -1307,7 +1307,7 @@ enum i915_power_well_id { > > #define DPIO_CMNRST (1 << 0) > > #define DPIO_PHY(pipe) ((pipe) >> 1) > > -#define DPIO_PHY_IOSF_PORT(phy) (dev_priv->dpio_phy_iosf_port[phy]) > > +#define DPIO_PHY_IOSF_PORT(phy) (i915->dpio_phy_iosf_port[phy]) > > /* > > * Per pipe/PLL DPIO regs > > @@ -3004,9 +3004,9 @@ enum i915_power_well_id { > > #define PM_VEBOX_CS_ERROR_INTERRUPT (1 << 12) /* hsw+ */ > > #define PM_VEBOX_USER_INTERRUPT (1 << 10) /* hsw+ */ > > -#define GT_PARITY_ERROR(dev_priv) \ > > +#define GT_PARITY_ERROR(i915) \ > > (GT_RENDER_L3_PARITY_ERROR_INTERRUPT | \ > > - (IS_HASWELL(dev_priv) ? GT_RENDER_L3_PARITY_ERROR_INTERRUPT_S1 : 0)) > > + (IS_HASWELL(i915) ? GT_RENDER_L3_PARITY_ERROR_INTERRUPT_S1 : 0)) > > /* These are all the "old" interrupts */ > > #define ILK_BSD_USER_INTERRUPT (1 << 5) > > @@ -3195,7 +3195,7 @@ enum i915_power_well_id { > > /* > > * GPIO regs > > */ > > -#define GPIO(gpio) _MMIO(dev_priv->gpio_mmio_base + 0x5010 + \ > > +#define GPIO(gpio) _MMIO(i915->gpio_mmio_base + 0x5010 + \ > > 4 * (gpio)) > > # define GPIO_CLOCK_DIR_MASK (1 << 0) > > @@ -3213,7 +3213,7 @@ enum i915_power_well_id { > > # define GPIO_DATA_VAL_IN (1 << 12) > > # define GPIO_DATA_PULLUP_DISABLE (1 << 13) > > -#define GMBUS0 _MMIO(dev_priv->gpio_mmio_base + 0x5100) /* clock/port select */ > > +#define GMBUS0 _MMIO(i915->gpio_mmio_base + 0x5100) /* clock/port select */ > > #define GMBUS_AKSV_SELECT (1 << 11) > > #define GMBUS_RATE_100KHZ (0 << 8) > > #define GMBUS_RATE_50KHZ (1 << 8) > > @@ -3240,7 +3240,7 @@ enum i915_power_well_id { > > #define GMBUS_PIN_12_TC4_ICP 12 > > #define GMBUS_NUM_PINS 13 /* including 0 */ > > -#define GMBUS1 _MMIO(dev_priv->gpio_mmio_base + 0x5104) /* command/status */ > > +#define GMBUS1 _MMIO(i915->gpio_mmio_base + 0x5104) /* command/status */ > > #define GMBUS_SW_CLR_INT (1 << 31) > > #define GMBUS_SW_RDY (1 << 30) > > #define GMBUS_ENT (1 << 29) /* enable timeout */ > > @@ -3255,7 +3255,7 @@ enum i915_power_well_id { > > #define GMBUS_SLAVE_ADDR_SHIFT 1 > > #define GMBUS_SLAVE_READ (1 << 0) > > #define GMBUS_SLAVE_WRITE (0 << 0) > > -#define GMBUS2 _MMIO(dev_priv->gpio_mmio_base + 0x5108) /* status */ > > +#define GMBUS2 _MMIO(i915->gpio_mmio_base + 0x5108) /* status */ > > #define GMBUS_INUSE (1 << 15) > > #define GMBUS_HW_WAIT_PHASE (1 << 14) > > #define GMBUS_STALL_TIMEOUT (1 << 13) > > @@ -3263,22 +3263,22 @@ enum i915_power_well_id { > > #define GMBUS_HW_RDY (1 << 11) > > #define GMBUS_SATOER (1 << 10) > > #define GMBUS_ACTIVE (1 << 9) > > -#define GMBUS3 _MMIO(dev_priv->gpio_mmio_base + 0x510c) /* data buffer bytes 3-0 */ > > -#define GMBUS4 _MMIO(dev_priv->gpio_mmio_base + 0x5110) /* interrupt mask (Pineview+) */ > > +#define GMBUS3 _MMIO(i915->gpio_mmio_base + 0x510c) /* data buffer bytes 3-0 */ > > +#define GMBUS4 _MMIO(i915->gpio_mmio_base + 0x5110) /* interrupt mask (Pineview+) */ > > #define GMBUS_SLAVE_TIMEOUT_EN (1 << 4) > > #define GMBUS_NAK_EN (1 << 3) > > #define GMBUS_IDLE_EN (1 << 2) > > #define GMBUS_HW_WAIT_EN (1 << 1) > > #define GMBUS_HW_RDY_EN (1 << 0) > > -#define GMBUS5 _MMIO(dev_priv->gpio_mmio_base + 0x5120) /* byte index */ > > +#define GMBUS5 _MMIO(i915->gpio_mmio_base + 0x5120) /* byte index */ > > #define GMBUS_2BYTE_INDEX_EN (1 << 31) > > /* > > * Clock control & power management > > */ > > -#define _DPLL_A (DISPLAY_MMIO_BASE(dev_priv) + 0x6014) > > -#define _DPLL_B (DISPLAY_MMIO_BASE(dev_priv) + 0x6018) > > -#define _CHV_DPLL_C (DISPLAY_MMIO_BASE(dev_priv) + 0x6030) > > +#define _DPLL_A (DISPLAY_MMIO_BASE(i915) + 0x6014) > > +#define _DPLL_B (DISPLAY_MMIO_BASE(i915) + 0x6018) > > +#define _CHV_DPLL_C (DISPLAY_MMIO_BASE(i915) + 0x6030) > > #define DPLL(pipe) _MMIO_PIPE3((pipe), _DPLL_A, _DPLL_B, _CHV_DPLL_C) > > #define VGA0 _MMIO(0x6000) > > @@ -3375,9 +3375,9 @@ enum i915_power_well_id { > > #define SDVO_MULTIPLIER_SHIFT_HIRES 4 > > #define SDVO_MULTIPLIER_SHIFT_VGA 0 > > -#define _DPLL_A_MD (DISPLAY_MMIO_BASE(dev_priv) + 0x601c) > > -#define _DPLL_B_MD (DISPLAY_MMIO_BASE(dev_priv) + 0x6020) > > -#define _CHV_DPLL_C_MD (DISPLAY_MMIO_BASE(dev_priv) + 0x603c) > > +#define _DPLL_A_MD (DISPLAY_MMIO_BASE(i915) + 0x601c) > > +#define _DPLL_B_MD (DISPLAY_MMIO_BASE(i915) + 0x6020) > > +#define _CHV_DPLL_C_MD (DISPLAY_MMIO_BASE(i915) + 0x603c) > > #define DPLL_MD(pipe) _MMIO_PIPE3((pipe), _DPLL_A_MD, _DPLL_B_MD, _CHV_DPLL_C_MD) > > /* > > @@ -3449,7 +3449,7 @@ enum i915_power_well_id { > > #define DSTATE_PLL_D3_OFF (1 << 3) > > #define DSTATE_GFX_CLOCK_GATING (1 << 1) > > #define DSTATE_DOT_CLOCK_GATING (1 << 0) > > -#define DSPCLK_GATE_D _MMIO(DISPLAY_MMIO_BASE(dev_priv) + 0x6200) > > +#define DSPCLK_GATE_D _MMIO(DISPLAY_MMIO_BASE(i915) + 0x6200) > > # define DPUNIT_B_CLOCK_GATE_DISABLE (1 << 30) /* 965 */ > > # define VSUNIT_CLOCK_GATE_DISABLE (1 << 29) /* 965 */ > > # define VRHUNIT_CLOCK_GATE_DISABLE (1 << 28) /* 965 */ > > @@ -3589,7 +3589,7 @@ enum i915_power_well_id { > > #define _PALETTE_A 0xa000 > > #define _PALETTE_B 0xa800 > > #define _CHV_PALETTE_C 0xc000 > > -#define PALETTE(pipe, i) _MMIO(DISPLAY_MMIO_BASE(dev_priv) + \ > > +#define PALETTE(pipe, i) _MMIO(DISPLAY_MMIO_BASE(i915) + \ > > _PICK((pipe), _PALETTE_A, \ > > _PALETTE_B, _CHV_PALETTE_C) + \ > > (i) * 4) > > @@ -3936,8 +3936,8 @@ enum i915_power_well_id { > > #define INTERVAL_1_28_US(us) roundup(((us) * 100) >> 7, 25) > > #define INTERVAL_1_33_US(us) (((us) * 3) >> 2) > > #define INTERVAL_0_833_US(us) (((us) * 6) / 5) > > -#define GT_INTERVAL_FROM_US(dev_priv, us) (INTEL_GEN(dev_priv) >= 9 ? \ > > - (IS_GEN9_LP(dev_priv) ? \ > > +#define GT_INTERVAL_FROM_US(i915, us) (INTEL_GEN(i915) >= 9 ? \ > > + (IS_GEN9_LP(i915) ? \ > > INTERVAL_0_833_US(us) : \ > > INTERVAL_1_33_US(us)) : \ > > INTERVAL_1_28_US(us)) > > @@ -3945,8 +3945,8 @@ enum i915_power_well_id { > > #define INTERVAL_1_28_TO_US(interval) (((interval) << 7) / 100) > > #define INTERVAL_1_33_TO_US(interval) (((interval) << 2) / 3) > > #define INTERVAL_0_833_TO_US(interval) (((interval) * 5) / 6) > > -#define GT_PM_INTERVAL_TO_US(dev_priv, interval) (INTEL_GEN(dev_priv) >= 9 ? \ > > - (IS_GEN9_LP(dev_priv) ? \ > > +#define GT_PM_INTERVAL_TO_US(i915, interval) (INTEL_GEN(i915) >= 9 ? \ > > + (IS_GEN9_LP(i915) ? \ > > INTERVAL_0_833_TO_US(interval) : \ > > INTERVAL_1_33_TO_US(interval)) : \ > > INTERVAL_1_28_TO_US(interval)) > > @@ -4219,7 +4219,7 @@ enum { > > /* HSW+ eDP PSR registers */ > > #define HSW_EDP_PSR_BASE 0x64800 > > #define BDW_EDP_PSR_BASE 0x6f800 > > -#define EDP_PSR_CTL _MMIO(dev_priv->psr_mmio_base + 0) > > +#define EDP_PSR_CTL _MMIO(i915->psr_mmio_base + 0) > > #define EDP_PSR_ENABLE (1 << 31) > > #define BDW_PSR_SINGLE_FRAME (1 << 30) > > #define EDP_PSR_RESTORE_PSR_ACTIVE_CTX_MASK (1 << 29) /* SW can't modify */ > > @@ -4256,16 +4256,16 @@ enum { > > #define EDP_PSR_TRANSCODER_A_SHIFT 8 > > #define EDP_PSR_TRANSCODER_EDP_SHIFT 0 > > -#define EDP_PSR_AUX_CTL _MMIO(dev_priv->psr_mmio_base + 0x10) > > +#define EDP_PSR_AUX_CTL _MMIO(i915->psr_mmio_base + 0x10) > > #define EDP_PSR_AUX_CTL_TIME_OUT_MASK (3 << 26) > > #define EDP_PSR_AUX_CTL_MESSAGE_SIZE_MASK (0x1f << 20) > > #define EDP_PSR_AUX_CTL_PRECHARGE_2US_MASK (0xf << 16) > > #define EDP_PSR_AUX_CTL_ERROR_INTERRUPT (1 << 11) > > #define EDP_PSR_AUX_CTL_BIT_CLOCK_2X_MASK (0x7ff) > > -#define EDP_PSR_AUX_DATA(i) _MMIO(dev_priv->psr_mmio_base + 0x14 + (i) * 4) /* 5 registers */ > > +#define EDP_PSR_AUX_DATA(i) _MMIO(i915->psr_mmio_base + 0x14 + (i) * 4) /* 5 registers */ > > -#define EDP_PSR_STATUS _MMIO(dev_priv->psr_mmio_base + 0x40) > > +#define EDP_PSR_STATUS _MMIO(i915->psr_mmio_base + 0x40) > > #define EDP_PSR_STATUS_STATE_MASK (7 << 29) > > #define EDP_PSR_STATUS_STATE_SHIFT 29 > > #define EDP_PSR_STATUS_STATE_IDLE (0 << 29) > > @@ -4290,10 +4290,10 @@ enum { > > #define EDP_PSR_STATUS_SENDING_TP1 (1 << 4) > > #define EDP_PSR_STATUS_IDLE_MASK 0xf > > -#define EDP_PSR_PERF_CNT _MMIO(dev_priv->psr_mmio_base + 0x44) > > +#define EDP_PSR_PERF_CNT _MMIO(i915->psr_mmio_base + 0x44) > > #define EDP_PSR_PERF_CNT_MASK 0xffffff > > -#define EDP_PSR_DEBUG _MMIO(dev_priv->psr_mmio_base + 0x60) /* PSR_MASK on SKL+ */ > > +#define EDP_PSR_DEBUG _MMIO(i915->psr_mmio_base + 0x60) /* PSR_MASK on SKL+ */ > > #define EDP_PSR_DEBUG_MASK_MAX_SLEEP (1 << 28) > > #define EDP_PSR_DEBUG_MASK_LPSP (1 << 27) > > #define EDP_PSR_DEBUG_MASK_MEMUP (1 << 26) > > @@ -4405,7 +4405,7 @@ enum { > > /* Hotplug control (945+ only) */ > > -#define PORT_HOTPLUG_EN _MMIO(DISPLAY_MMIO_BASE(dev_priv) + 0x61110) > > +#define PORT_HOTPLUG_EN _MMIO(DISPLAY_MMIO_BASE(i915) + 0x61110) > > #define PORTB_HOTPLUG_INT_EN (1 << 29) > > #define PORTC_HOTPLUG_INT_EN (1 << 28) > > #define PORTD_HOTPLUG_INT_EN (1 << 27) > > @@ -4435,7 +4435,7 @@ enum { > > #define CRT_HOTPLUG_DETECT_VOLTAGE_325MV (0 << 2) > > #define CRT_HOTPLUG_DETECT_VOLTAGE_475MV (1 << 2) > > -#define PORT_HOTPLUG_STAT _MMIO(DISPLAY_MMIO_BASE(dev_priv) + 0x61114) > > +#define PORT_HOTPLUG_STAT _MMIO(DISPLAY_MMIO_BASE(i915) + 0x61114) > > /* > > * HDMI/DP bits are g4x+ > > * > > @@ -4517,7 +4517,7 @@ enum { > > #define PORT_DFT_I9XX _MMIO(0x61150) > > #define DC_BALANCE_RESET (1 << 25) > > -#define PORT_DFT2_G4X _MMIO(DISPLAY_MMIO_BASE(dev_priv) + 0x61154) > > +#define PORT_DFT2_G4X _MMIO(DISPLAY_MMIO_BASE(i915) + 0x61154) > > #define DC_BALANCE_RESET_VLV (1 << 31) > > #define PIPE_SCRAMBLE_RESET_MASK ((1 << 14) | (0x3 << 0)) > > #define PIPE_C_SCRAMBLE_RESET (1 << 14) /* chv */ > > @@ -4719,7 +4719,7 @@ enum { > > #define VLV_PPS_BASE (VLV_DISPLAY_BASE + PPS_BASE) > > #define PCH_PPS_BASE 0xC7200 > > -#define _MMIO_PPS(pps_idx, reg) _MMIO(dev_priv->pps_mmio_base - \ > > +#define _MMIO_PPS(pps_idx, reg) _MMIO(i915->pps_mmio_base - \ > > PPS_BASE + (reg) + \ > > (pps_idx) * 0x100) > > @@ -4792,7 +4792,7 @@ enum { > > #define PANEL_POWER_CYCLE_DELAY_MASK REG_GENMASK(4, 0) > > /* Panel fitting */ > > -#define PFIT_CONTROL _MMIO(DISPLAY_MMIO_BASE(dev_priv) + 0x61230) > > +#define PFIT_CONTROL _MMIO(DISPLAY_MMIO_BASE(i915) + 0x61230) > > #define PFIT_ENABLE (1 << 31) > > #define PFIT_PIPE_MASK (3 << 29) > > #define PFIT_PIPE_SHIFT 29 > > @@ -4810,7 +4810,7 @@ enum { > > #define PFIT_SCALING_PROGRAMMED (1 << 26) > > #define PFIT_SCALING_PILLAR (2 << 26) > > #define PFIT_SCALING_LETTER (3 << 26) > > -#define PFIT_PGM_RATIOS _MMIO(DISPLAY_MMIO_BASE(dev_priv) + 0x61234) > > +#define PFIT_PGM_RATIOS _MMIO(DISPLAY_MMIO_BASE(i915) + 0x61234) > > /* Pre-965 */ > > #define PFIT_VERT_SCALE_SHIFT 20 > > #define PFIT_VERT_SCALE_MASK 0xfff00000 > > @@ -4822,25 +4822,25 @@ enum { > > #define PFIT_HORIZ_SCALE_SHIFT_965 0 > > #define PFIT_HORIZ_SCALE_MASK_965 0x00001fff > > -#define PFIT_AUTO_RATIOS _MMIO(DISPLAY_MMIO_BASE(dev_priv) + 0x61238) > > +#define PFIT_AUTO_RATIOS _MMIO(DISPLAY_MMIO_BASE(i915) + 0x61238) > > -#define _VLV_BLC_PWM_CTL2_A (DISPLAY_MMIO_BASE(dev_priv) + 0x61250) > > -#define _VLV_BLC_PWM_CTL2_B (DISPLAY_MMIO_BASE(dev_priv) + 0x61350) > > +#define _VLV_BLC_PWM_CTL2_A (DISPLAY_MMIO_BASE(i915) + 0x61250) > > +#define _VLV_BLC_PWM_CTL2_B (DISPLAY_MMIO_BASE(i915) + 0x61350) > > #define VLV_BLC_PWM_CTL2(pipe) _MMIO_PIPE(pipe, _VLV_BLC_PWM_CTL2_A, \ > > _VLV_BLC_PWM_CTL2_B) > > -#define _VLV_BLC_PWM_CTL_A (DISPLAY_MMIO_BASE(dev_priv) + 0x61254) > > -#define _VLV_BLC_PWM_CTL_B (DISPLAY_MMIO_BASE(dev_priv) + 0x61354) > > +#define _VLV_BLC_PWM_CTL_A (DISPLAY_MMIO_BASE(i915) + 0x61254) > > +#define _VLV_BLC_PWM_CTL_B (DISPLAY_MMIO_BASE(i915) + 0x61354) > > #define VLV_BLC_PWM_CTL(pipe) _MMIO_PIPE(pipe, _VLV_BLC_PWM_CTL_A, \ > > _VLV_BLC_PWM_CTL_B) > > -#define _VLV_BLC_HIST_CTL_A (DISPLAY_MMIO_BASE(dev_priv) + 0x61260) > > -#define _VLV_BLC_HIST_CTL_B (DISPLAY_MMIO_BASE(dev_priv) + 0x61360) > > +#define _VLV_BLC_HIST_CTL_A (DISPLAY_MMIO_BASE(i915) + 0x61260) > > +#define _VLV_BLC_HIST_CTL_B (DISPLAY_MMIO_BASE(i915) + 0x61360) > > #define VLV_BLC_HIST_CTL(pipe) _MMIO_PIPE(pipe, _VLV_BLC_HIST_CTL_A, \ > > _VLV_BLC_HIST_CTL_B) > > /* Backlight control */ > > -#define BLC_PWM_CTL2 _MMIO(DISPLAY_MMIO_BASE(dev_priv) + 0x61250) /* 965+ only */ > > +#define BLC_PWM_CTL2 _MMIO(DISPLAY_MMIO_BASE(i915) + 0x61250) /* 965+ only */ > > #define BLM_PWM_ENABLE (1 << 31) > > #define BLM_COMBINATION_MODE (1 << 30) /* gen4 only */ > > #define BLM_PIPE_SELECT (1 << 29) > > @@ -4863,7 +4863,7 @@ enum { > > #define BLM_PHASE_IN_COUNT_MASK (0xff << 8) > > #define BLM_PHASE_IN_INCR_SHIFT (0) > > #define BLM_PHASE_IN_INCR_MASK (0xff << 0) > > -#define BLC_PWM_CTL _MMIO(DISPLAY_MMIO_BASE(dev_priv) + 0x61254) > > +#define BLC_PWM_CTL _MMIO(DISPLAY_MMIO_BASE(i915) + 0x61254) > > /* > > * This is the most significant 15 bits of the number of backlight cycles in a > > * complete cycle of the modulated backlight control. > > @@ -4885,7 +4885,7 @@ enum { > > #define BACKLIGHT_DUTY_CYCLE_MASK_PNV (0xfffe) > > #define BLM_POLARITY_PNV (1 << 0) /* pnv only */ > > -#define BLC_HIST_CTL _MMIO(DISPLAY_MMIO_BASE(dev_priv) + 0x61260) > > +#define BLC_HIST_CTL _MMIO(DISPLAY_MMIO_BASE(i915) + 0x61260) > > #define BLM_HISTOGRAM_ENABLE (1 << 31) > > /* New registers for PCH-split platforms. Safe where new bits show up, the > > @@ -5510,47 +5510,47 @@ enum { > > * is 20 bytes in each direction, hence the 5 fixed > > * data registers > > */ > > -#define _DPA_AUX_CH_CTL (DISPLAY_MMIO_BASE(dev_priv) + 0x64010) > > -#define _DPA_AUX_CH_DATA1 (DISPLAY_MMIO_BASE(dev_priv) + 0x64014) > > -#define _DPA_AUX_CH_DATA2 (DISPLAY_MMIO_BASE(dev_priv) + 0x64018) > > -#define _DPA_AUX_CH_DATA3 (DISPLAY_MMIO_BASE(dev_priv) + 0x6401c) > > -#define _DPA_AUX_CH_DATA4 (DISPLAY_MMIO_BASE(dev_priv) + 0x64020) > > -#define _DPA_AUX_CH_DATA5 (DISPLAY_MMIO_BASE(dev_priv) + 0x64024) > > - > > -#define _DPB_AUX_CH_CTL (DISPLAY_MMIO_BASE(dev_priv) + 0x64110) > > -#define _DPB_AUX_CH_DATA1 (DISPLAY_MMIO_BASE(dev_priv) + 0x64114) > > -#define _DPB_AUX_CH_DATA2 (DISPLAY_MMIO_BASE(dev_priv) + 0x64118) > > -#define _DPB_AUX_CH_DATA3 (DISPLAY_MMIO_BASE(dev_priv) + 0x6411c) > > -#define _DPB_AUX_CH_DATA4 (DISPLAY_MMIO_BASE(dev_priv) + 0x64120) > > -#define _DPB_AUX_CH_DATA5 (DISPLAY_MMIO_BASE(dev_priv) + 0x64124) > > - > > -#define _DPC_AUX_CH_CTL (DISPLAY_MMIO_BASE(dev_priv) + 0x64210) > > -#define _DPC_AUX_CH_DATA1 (DISPLAY_MMIO_BASE(dev_priv) + 0x64214) > > -#define _DPC_AUX_CH_DATA2 (DISPLAY_MMIO_BASE(dev_priv) + 0x64218) > > -#define _DPC_AUX_CH_DATA3 (DISPLAY_MMIO_BASE(dev_priv) + 0x6421c) > > -#define _DPC_AUX_CH_DATA4 (DISPLAY_MMIO_BASE(dev_priv) + 0x64220) > > -#define _DPC_AUX_CH_DATA5 (DISPLAY_MMIO_BASE(dev_priv) + 0x64224) > > - > > -#define _DPD_AUX_CH_CTL (DISPLAY_MMIO_BASE(dev_priv) + 0x64310) > > -#define _DPD_AUX_CH_DATA1 (DISPLAY_MMIO_BASE(dev_priv) + 0x64314) > > -#define _DPD_AUX_CH_DATA2 (DISPLAY_MMIO_BASE(dev_priv) + 0x64318) > > -#define _DPD_AUX_CH_DATA3 (DISPLAY_MMIO_BASE(dev_priv) + 0x6431c) > > -#define _DPD_AUX_CH_DATA4 (DISPLAY_MMIO_BASE(dev_priv) + 0x64320) > > -#define _DPD_AUX_CH_DATA5 (DISPLAY_MMIO_BASE(dev_priv) + 0x64324) > > - > > -#define _DPE_AUX_CH_CTL (DISPLAY_MMIO_BASE(dev_priv) + 0x64410) > > -#define _DPE_AUX_CH_DATA1 (DISPLAY_MMIO_BASE(dev_priv) + 0x64414) > > -#define _DPE_AUX_CH_DATA2 (DISPLAY_MMIO_BASE(dev_priv) + 0x64418) > > -#define _DPE_AUX_CH_DATA3 (DISPLAY_MMIO_BASE(dev_priv) + 0x6441c) > > -#define _DPE_AUX_CH_DATA4 (DISPLAY_MMIO_BASE(dev_priv) + 0x64420) > > -#define _DPE_AUX_CH_DATA5 (DISPLAY_MMIO_BASE(dev_priv) + 0x64424) > > - > > -#define _DPF_AUX_CH_CTL (DISPLAY_MMIO_BASE(dev_priv) + 0x64510) > > -#define _DPF_AUX_CH_DATA1 (DISPLAY_MMIO_BASE(dev_priv) + 0x64514) > > -#define _DPF_AUX_CH_DATA2 (DISPLAY_MMIO_BASE(dev_priv) + 0x64518) > > -#define _DPF_AUX_CH_DATA3 (DISPLAY_MMIO_BASE(dev_priv) + 0x6451c) > > -#define _DPF_AUX_CH_DATA4 (DISPLAY_MMIO_BASE(dev_priv) + 0x64520) > > -#define _DPF_AUX_CH_DATA5 (DISPLAY_MMIO_BASE(dev_priv) + 0x64524) > > +#define _DPA_AUX_CH_CTL (DISPLAY_MMIO_BASE(i915) + 0x64010) > > +#define _DPA_AUX_CH_DATA1 (DISPLAY_MMIO_BASE(i915) + 0x64014) > > +#define _DPA_AUX_CH_DATA2 (DISPLAY_MMIO_BASE(i915) + 0x64018) > > +#define _DPA_AUX_CH_DATA3 (DISPLAY_MMIO_BASE(i915) + 0x6401c) > > +#define _DPA_AUX_CH_DATA4 (DISPLAY_MMIO_BASE(i915) + 0x64020) > > +#define _DPA_AUX_CH_DATA5 (DISPLAY_MMIO_BASE(i915) + 0x64024) > > + > > +#define _DPB_AUX_CH_CTL (DISPLAY_MMIO_BASE(i915) + 0x64110) > > +#define _DPB_AUX_CH_DATA1 (DISPLAY_MMIO_BASE(i915) + 0x64114) > > +#define _DPB_AUX_CH_DATA2 (DISPLAY_MMIO_BASE(i915) + 0x64118) > > +#define _DPB_AUX_CH_DATA3 (DISPLAY_MMIO_BASE(i915) + 0x6411c) > > +#define _DPB_AUX_CH_DATA4 (DISPLAY_MMIO_BASE(i915) + 0x64120) > > +#define _DPB_AUX_CH_DATA5 (DISPLAY_MMIO_BASE(i915) + 0x64124) > > + > > +#define _DPC_AUX_CH_CTL (DISPLAY_MMIO_BASE(i915) + 0x64210) > > +#define _DPC_AUX_CH_DATA1 (DISPLAY_MMIO_BASE(i915) + 0x64214) > > +#define _DPC_AUX_CH_DATA2 (DISPLAY_MMIO_BASE(i915) + 0x64218) > > +#define _DPC_AUX_CH_DATA3 (DISPLAY_MMIO_BASE(i915) + 0x6421c) > > +#define _DPC_AUX_CH_DATA4 (DISPLAY_MMIO_BASE(i915) + 0x64220) > > +#define _DPC_AUX_CH_DATA5 (DISPLAY_MMIO_BASE(i915) + 0x64224) > > + > > +#define _DPD_AUX_CH_CTL (DISPLAY_MMIO_BASE(i915) + 0x64310) > > +#define _DPD_AUX_CH_DATA1 (DISPLAY_MMIO_BASE(i915) + 0x64314) > > +#define _DPD_AUX_CH_DATA2 (DISPLAY_MMIO_BASE(i915) + 0x64318) > > +#define _DPD_AUX_CH_DATA3 (DISPLAY_MMIO_BASE(i915) + 0x6431c) > > +#define _DPD_AUX_CH_DATA4 (DISPLAY_MMIO_BASE(i915) + 0x64320) > > +#define _DPD_AUX_CH_DATA5 (DISPLAY_MMIO_BASE(i915) + 0x64324) > > + > > +#define _DPE_AUX_CH_CTL (DISPLAY_MMIO_BASE(i915) + 0x64410) > > +#define _DPE_AUX_CH_DATA1 (DISPLAY_MMIO_BASE(i915) + 0x64414) > > +#define _DPE_AUX_CH_DATA2 (DISPLAY_MMIO_BASE(i915) + 0x64418) > > +#define _DPE_AUX_CH_DATA3 (DISPLAY_MMIO_BASE(i915) + 0x6441c) > > +#define _DPE_AUX_CH_DATA4 (DISPLAY_MMIO_BASE(i915) + 0x64420) > > +#define _DPE_AUX_CH_DATA5 (DISPLAY_MMIO_BASE(i915) + 0x64424) > > + > > +#define _DPF_AUX_CH_CTL (DISPLAY_MMIO_BASE(i915) + 0x64510) > > +#define _DPF_AUX_CH_DATA1 (DISPLAY_MMIO_BASE(i915) + 0x64514) > > +#define _DPF_AUX_CH_DATA2 (DISPLAY_MMIO_BASE(i915) + 0x64518) > > +#define _DPF_AUX_CH_DATA3 (DISPLAY_MMIO_BASE(i915) + 0x6451c) > > +#define _DPF_AUX_CH_DATA4 (DISPLAY_MMIO_BASE(i915) + 0x64520) > > +#define _DPF_AUX_CH_DATA5 (DISPLAY_MMIO_BASE(i915) + 0x64524) > > #define DP_AUX_CH_CTL(aux_ch) _MMIO_PORT(aux_ch, _DPA_AUX_CH_CTL, _DPB_AUX_CH_CTL) > > #define DP_AUX_CH_DATA(aux_ch, i) _MMIO(_PORT(aux_ch, _DPA_AUX_CH_DATA1, _DPB_AUX_CH_DATA1) + (i) * 4) /* 5 registers */ > > @@ -5843,7 +5843,7 @@ enum { > > #define DPINVGTT_STATUS_MASK 0xff > > #define DPINVGTT_STATUS_MASK_CHV 0xfff > > -#define DSPARB _MMIO(DISPLAY_MMIO_BASE(dev_priv) + 0x70030) > > +#define DSPARB _MMIO(DISPLAY_MMIO_BASE(i915) + 0x70030) > > #define DSPARB_CSTART_MASK (0x7f << 7) > > #define DSPARB_CSTART_SHIFT 7 > > #define DSPARB_BSTART_MASK (0x7f) > > @@ -5878,7 +5878,7 @@ enum { > > #define DSPARB_SPRITEF_MASK_VLV (0xff << 8) > > /* pnv/gen4/g4x/vlv/chv */ > > -#define DSPFW1 _MMIO(DISPLAY_MMIO_BASE(dev_priv) + 0x70034) > > +#define DSPFW1 _MMIO(DISPLAY_MMIO_BASE(i915) + 0x70034) > > #define DSPFW_SR_SHIFT 23 > > #define DSPFW_SR_MASK (0x1ff << 23) > > #define DSPFW_CURSORB_SHIFT 16 > > @@ -5889,7 +5889,7 @@ enum { > > #define DSPFW_PLANEA_SHIFT 0 > > #define DSPFW_PLANEA_MASK (0x7f << 0) > > #define DSPFW_PLANEA_MASK_VLV (0xff << 0) /* vlv/chv */ > > -#define DSPFW2 _MMIO(DISPLAY_MMIO_BASE(dev_priv) + 0x70038) > > +#define DSPFW2 _MMIO(DISPLAY_MMIO_BASE(i915) + 0x70038) > > #define DSPFW_FBC_SR_EN (1 << 31) /* g4x */ > > #define DSPFW_FBC_SR_SHIFT 28 > > #define DSPFW_FBC_SR_MASK (0x7 << 28) /* g4x */ > > @@ -5905,7 +5905,7 @@ enum { > > #define DSPFW_SPRITEA_SHIFT 0 > > #define DSPFW_SPRITEA_MASK (0x7f << 0) /* g4x */ > > #define DSPFW_SPRITEA_MASK_VLV (0xff << 0) /* vlv/chv */ > > -#define DSPFW3 _MMIO(DISPLAY_MMIO_BASE(dev_priv) + 0x7003c) > > +#define DSPFW3 _MMIO(DISPLAY_MMIO_BASE(i915) + 0x7003c) > > #define DSPFW_HPLL_SR_EN (1 << 31) > > #define PINEVIEW_SELF_REFRESH_EN (1 << 30) > > #define DSPFW_CURSOR_SR_SHIFT 24 > > @@ -6322,35 +6322,35 @@ enum { > > * [10:1f] all > > * [30:32] all > > */ > > -#define SWF0(i) _MMIO(DISPLAY_MMIO_BASE(dev_priv) + 0x70410 + (i) * 4) > > -#define SWF1(i) _MMIO(DISPLAY_MMIO_BASE(dev_priv) + 0x71410 + (i) * 4) > > -#define SWF3(i) _MMIO(DISPLAY_MMIO_BASE(dev_priv) + 0x72414 + (i) * 4) > > +#define SWF0(i) _MMIO(DISPLAY_MMIO_BASE(i915) + 0x70410 + (i) * 4) > > +#define SWF1(i) _MMIO(DISPLAY_MMIO_BASE(i915) + 0x71410 + (i) * 4) > > +#define SWF3(i) _MMIO(DISPLAY_MMIO_BASE(i915) + 0x72414 + (i) * 4) > > #define SWF_ILK(i) _MMIO(0x4F000 + (i) * 4) > > /* Pipe B */ > > -#define _PIPEBDSL (DISPLAY_MMIO_BASE(dev_priv) + 0x71000) > > -#define _PIPEBCONF (DISPLAY_MMIO_BASE(dev_priv) + 0x71008) > > -#define _PIPEBSTAT (DISPLAY_MMIO_BASE(dev_priv) + 0x71024) > > +#define _PIPEBDSL (DISPLAY_MMIO_BASE(i915) + 0x71000) > > +#define _PIPEBCONF (DISPLAY_MMIO_BASE(i915) + 0x71008) > > +#define _PIPEBSTAT (DISPLAY_MMIO_BASE(i915) + 0x71024) > > #define _PIPEBFRAMEHIGH 0x71040 > > #define _PIPEBFRAMEPIXEL 0x71044 > > -#define _PIPEB_FRMCOUNT_G4X (DISPLAY_MMIO_BASE(dev_priv) + 0x71040) > > -#define _PIPEB_FLIPCOUNT_G4X (DISPLAY_MMIO_BASE(dev_priv) + 0x71044) > > +#define _PIPEB_FRMCOUNT_G4X (DISPLAY_MMIO_BASE(i915) + 0x71040) > > +#define _PIPEB_FLIPCOUNT_G4X (DISPLAY_MMIO_BASE(i915) + 0x71044) > > /* Display B control */ > > -#define _DSPBCNTR (DISPLAY_MMIO_BASE(dev_priv) + 0x71180) > > +#define _DSPBCNTR (DISPLAY_MMIO_BASE(i915) + 0x71180) > > #define DISPPLANE_ALPHA_TRANS_ENABLE (1 << 15) > > #define DISPPLANE_ALPHA_TRANS_DISABLE 0 > > #define DISPPLANE_SPRITE_ABOVE_DISPLAY 0 > > #define DISPPLANE_SPRITE_ABOVE_OVERLAY (1) > > -#define _DSPBADDR (DISPLAY_MMIO_BASE(dev_priv) + 0x71184) > > -#define _DSPBSTRIDE (DISPLAY_MMIO_BASE(dev_priv) + 0x71188) > > -#define _DSPBPOS (DISPLAY_MMIO_BASE(dev_priv) + 0x7118C) > > -#define _DSPBSIZE (DISPLAY_MMIO_BASE(dev_priv) + 0x71190) > > -#define _DSPBSURF (DISPLAY_MMIO_BASE(dev_priv) + 0x7119C) > > -#define _DSPBTILEOFF (DISPLAY_MMIO_BASE(dev_priv) + 0x711A4) > > -#define _DSPBOFFSET (DISPLAY_MMIO_BASE(dev_priv) + 0x711A4) > > -#define _DSPBSURFLIVE (DISPLAY_MMIO_BASE(dev_priv) + 0x711AC) > > +#define _DSPBADDR (DISPLAY_MMIO_BASE(i915) + 0x71184) > > +#define _DSPBSTRIDE (DISPLAY_MMIO_BASE(i915) + 0x71188) > > +#define _DSPBPOS (DISPLAY_MMIO_BASE(i915) + 0x7118C) > > +#define _DSPBSIZE (DISPLAY_MMIO_BASE(i915) + 0x71190) > > +#define _DSPBSURF (DISPLAY_MMIO_BASE(i915) + 0x7119C) > > +#define _DSPBTILEOFF (DISPLAY_MMIO_BASE(i915) + 0x711A4) > > +#define _DSPBOFFSET (DISPLAY_MMIO_BASE(i915) + 0x711A4) > > +#define _DSPBSURFLIVE (DISPLAY_MMIO_BASE(i915) + 0x711AC) > > /* ICL DSI 0 and 1 */ > > #define _PIPEDSI0CONF 0x7b008 > > @@ -8946,7 +8946,7 @@ enum { > > #define GEN9_ENABLE_GPGPU_PREEMPTION (1 << 2) > > /* Audio */ > > -#define G4X_AUD_VID_DID _MMIO(DISPLAY_MMIO_BASE(dev_priv) + 0x62020) > > +#define G4X_AUD_VID_DID _MMIO(DISPLAY_MMIO_BASE(i915) + 0x62020) > > #define INTEL_AUDIO_DEVCL 0x808629FB > > #define INTEL_AUDIO_DEVBLC 0x80862801 > > #define INTEL_AUDIO_DEVCTG 0x80862802 > > @@ -10469,8 +10469,8 @@ enum skl_power_gate { > > /* MIPI DSI Controller and D-PHY registers */ > > -#define _MIPIA_DEVICE_READY (dev_priv->mipi_mmio_base + 0xb000) > > -#define _MIPIC_DEVICE_READY (dev_priv->mipi_mmio_base + 0xb800) > > +#define _MIPIA_DEVICE_READY (i915->mipi_mmio_base + 0xb000) > > +#define _MIPIC_DEVICE_READY (i915->mipi_mmio_base + 0xb800) > > #define MIPI_DEVICE_READY(port) _MMIO_MIPI(port, _MIPIA_DEVICE_READY, _MIPIC_DEVICE_READY) > > #define BUS_POSSESSION (1 << 3) /* set to give bus to receiver */ > > #define ULPS_STATE_MASK (3 << 1) > > @@ -10479,11 +10479,11 @@ enum skl_power_gate { > > #define ULPS_STATE_NORMAL_OPERATION (0 << 1) > > #define DEVICE_READY (1 << 0) > > -#define _MIPIA_INTR_STAT (dev_priv->mipi_mmio_base + 0xb004) > > -#define _MIPIC_INTR_STAT (dev_priv->mipi_mmio_base + 0xb804) > > +#define _MIPIA_INTR_STAT (i915->mipi_mmio_base + 0xb004) > > +#define _MIPIC_INTR_STAT (i915->mipi_mmio_base + 0xb804) > > #define MIPI_INTR_STAT(port) _MMIO_MIPI(port, _MIPIA_INTR_STAT, _MIPIC_INTR_STAT) > > -#define _MIPIA_INTR_EN (dev_priv->mipi_mmio_base + 0xb008) > > -#define _MIPIC_INTR_EN (dev_priv->mipi_mmio_base + 0xb808) > > +#define _MIPIA_INTR_EN (i915->mipi_mmio_base + 0xb008) > > +#define _MIPIC_INTR_EN (i915->mipi_mmio_base + 0xb808) > > #define MIPI_INTR_EN(port) _MMIO_MIPI(port, _MIPIA_INTR_EN, _MIPIC_INTR_EN) > > #define TEARING_EFFECT (1 << 31) > > #define SPL_PKT_SENT_INTERRUPT (1 << 30) > > @@ -10518,8 +10518,8 @@ enum skl_power_gate { > > #define RXSOT_SYNC_ERROR (1 << 1) > > #define RXSOT_ERROR (1 << 0) > > -#define _MIPIA_DSI_FUNC_PRG (dev_priv->mipi_mmio_base + 0xb00c) > > -#define _MIPIC_DSI_FUNC_PRG (dev_priv->mipi_mmio_base + 0xb80c) > > +#define _MIPIA_DSI_FUNC_PRG (i915->mipi_mmio_base + 0xb00c) > > +#define _MIPIC_DSI_FUNC_PRG (i915->mipi_mmio_base + 0xb80c) > > #define MIPI_DSI_FUNC_PRG(port) _MMIO_MIPI(port, _MIPIA_DSI_FUNC_PRG, _MIPIC_DSI_FUNC_PRG) > > #define CMD_MODE_DATA_WIDTH_MASK (7 << 13) > > #define CMD_MODE_NOT_SUPPORTED (0 << 13) > > @@ -10541,78 +10541,78 @@ enum skl_power_gate { > > #define DATA_LANES_PRG_REG_SHIFT 0 > > #define DATA_LANES_PRG_REG_MASK (7 << 0) > > -#define _MIPIA_HS_TX_TIMEOUT (dev_priv->mipi_mmio_base + 0xb010) > > -#define _MIPIC_HS_TX_TIMEOUT (dev_priv->mipi_mmio_base + 0xb810) > > +#define _MIPIA_HS_TX_TIMEOUT (i915->mipi_mmio_base + 0xb010) > > +#define _MIPIC_HS_TX_TIMEOUT (i915->mipi_mmio_base + 0xb810) > > #define MIPI_HS_TX_TIMEOUT(port) _MMIO_MIPI(port, _MIPIA_HS_TX_TIMEOUT, _MIPIC_HS_TX_TIMEOUT) > > #define HIGH_SPEED_TX_TIMEOUT_COUNTER_MASK 0xffffff > > -#define _MIPIA_LP_RX_TIMEOUT (dev_priv->mipi_mmio_base + 0xb014) > > -#define _MIPIC_LP_RX_TIMEOUT (dev_priv->mipi_mmio_base + 0xb814) > > +#define _MIPIA_LP_RX_TIMEOUT (i915->mipi_mmio_base + 0xb014) > > +#define _MIPIC_LP_RX_TIMEOUT (i915->mipi_mmio_base + 0xb814) > > #define MIPI_LP_RX_TIMEOUT(port) _MMIO_MIPI(port, _MIPIA_LP_RX_TIMEOUT, _MIPIC_LP_RX_TIMEOUT) > > #define LOW_POWER_RX_TIMEOUT_COUNTER_MASK 0xffffff > > -#define _MIPIA_TURN_AROUND_TIMEOUT (dev_priv->mipi_mmio_base + 0xb018) > > -#define _MIPIC_TURN_AROUND_TIMEOUT (dev_priv->mipi_mmio_base + 0xb818) > > +#define _MIPIA_TURN_AROUND_TIMEOUT (i915->mipi_mmio_base + 0xb018) > > +#define _MIPIC_TURN_AROUND_TIMEOUT (i915->mipi_mmio_base + 0xb818) > > #define MIPI_TURN_AROUND_TIMEOUT(port) _MMIO_MIPI(port, _MIPIA_TURN_AROUND_TIMEOUT, _MIPIC_TURN_AROUND_TIMEOUT) > > #define TURN_AROUND_TIMEOUT_MASK 0x3f > > -#define _MIPIA_DEVICE_RESET_TIMER (dev_priv->mipi_mmio_base + 0xb01c) > > -#define _MIPIC_DEVICE_RESET_TIMER (dev_priv->mipi_mmio_base + 0xb81c) > > +#define _MIPIA_DEVICE_RESET_TIMER (i915->mipi_mmio_base + 0xb01c) > > +#define _MIPIC_DEVICE_RESET_TIMER (i915->mipi_mmio_base + 0xb81c) > > #define MIPI_DEVICE_RESET_TIMER(port) _MMIO_MIPI(port, _MIPIA_DEVICE_RESET_TIMER, _MIPIC_DEVICE_RESET_TIMER) > > #define DEVICE_RESET_TIMER_MASK 0xffff > > -#define _MIPIA_DPI_RESOLUTION (dev_priv->mipi_mmio_base + 0xb020) > > -#define _MIPIC_DPI_RESOLUTION (dev_priv->mipi_mmio_base + 0xb820) > > +#define _MIPIA_DPI_RESOLUTION (i915->mipi_mmio_base + 0xb020) > > +#define _MIPIC_DPI_RESOLUTION (i915->mipi_mmio_base + 0xb820) > > #define MIPI_DPI_RESOLUTION(port) _MMIO_MIPI(port, _MIPIA_DPI_RESOLUTION, _MIPIC_DPI_RESOLUTION) > > #define VERTICAL_ADDRESS_SHIFT 16 > > #define VERTICAL_ADDRESS_MASK (0xffff << 16) > > #define HORIZONTAL_ADDRESS_SHIFT 0 > > #define HORIZONTAL_ADDRESS_MASK 0xffff > > -#define _MIPIA_DBI_FIFO_THROTTLE (dev_priv->mipi_mmio_base + 0xb024) > > -#define _MIPIC_DBI_FIFO_THROTTLE (dev_priv->mipi_mmio_base + 0xb824) > > +#define _MIPIA_DBI_FIFO_THROTTLE (i915->mipi_mmio_base + 0xb024) > > +#define _MIPIC_DBI_FIFO_THROTTLE (i915->mipi_mmio_base + 0xb824) > > #define MIPI_DBI_FIFO_THROTTLE(port) _MMIO_MIPI(port, _MIPIA_DBI_FIFO_THROTTLE, _MIPIC_DBI_FIFO_THROTTLE) > > #define DBI_FIFO_EMPTY_HALF (0 << 0) > > #define DBI_FIFO_EMPTY_QUARTER (1 << 0) > > #define DBI_FIFO_EMPTY_7_LOCATIONS (2 << 0) > > /* regs below are bits 15:0 */ > > -#define _MIPIA_HSYNC_PADDING_COUNT (dev_priv->mipi_mmio_base + 0xb028) > > -#define _MIPIC_HSYNC_PADDING_COUNT (dev_priv->mipi_mmio_base + 0xb828) > > +#define _MIPIA_HSYNC_PADDING_COUNT (i915->mipi_mmio_base + 0xb028) > > +#define _MIPIC_HSYNC_PADDING_COUNT (i915->mipi_mmio_base + 0xb828) > > #define MIPI_HSYNC_PADDING_COUNT(port) _MMIO_MIPI(port, _MIPIA_HSYNC_PADDING_COUNT, _MIPIC_HSYNC_PADDING_COUNT) > > -#define _MIPIA_HBP_COUNT (dev_priv->mipi_mmio_base + 0xb02c) > > -#define _MIPIC_HBP_COUNT (dev_priv->mipi_mmio_base + 0xb82c) > > +#define _MIPIA_HBP_COUNT (i915->mipi_mmio_base + 0xb02c) > > +#define _MIPIC_HBP_COUNT (i915->mipi_mmio_base + 0xb82c) > > #define MIPI_HBP_COUNT(port) _MMIO_MIPI(port, _MIPIA_HBP_COUNT, _MIPIC_HBP_COUNT) > > -#define _MIPIA_HFP_COUNT (dev_priv->mipi_mmio_base + 0xb030) > > -#define _MIPIC_HFP_COUNT (dev_priv->mipi_mmio_base + 0xb830) > > +#define _MIPIA_HFP_COUNT (i915->mipi_mmio_base + 0xb030) > > +#define _MIPIC_HFP_COUNT (i915->mipi_mmio_base + 0xb830) > > #define MIPI_HFP_COUNT(port) _MMIO_MIPI(port, _MIPIA_HFP_COUNT, _MIPIC_HFP_COUNT) > > -#define _MIPIA_HACTIVE_AREA_COUNT (dev_priv->mipi_mmio_base + 0xb034) > > -#define _MIPIC_HACTIVE_AREA_COUNT (dev_priv->mipi_mmio_base + 0xb834) > > +#define _MIPIA_HACTIVE_AREA_COUNT (i915->mipi_mmio_base + 0xb034) > > +#define _MIPIC_HACTIVE_AREA_COUNT (i915->mipi_mmio_base + 0xb834) > > #define MIPI_HACTIVE_AREA_COUNT(port) _MMIO_MIPI(port, _MIPIA_HACTIVE_AREA_COUNT, _MIPIC_HACTIVE_AREA_COUNT) > > -#define _MIPIA_VSYNC_PADDING_COUNT (dev_priv->mipi_mmio_base + 0xb038) > > -#define _MIPIC_VSYNC_PADDING_COUNT (dev_priv->mipi_mmio_base + 0xb838) > > +#define _MIPIA_VSYNC_PADDING_COUNT (i915->mipi_mmio_base + 0xb038) > > +#define _MIPIC_VSYNC_PADDING_COUNT (i915->mipi_mmio_base + 0xb838) > > #define MIPI_VSYNC_PADDING_COUNT(port) _MMIO_MIPI(port, _MIPIA_VSYNC_PADDING_COUNT, _MIPIC_VSYNC_PADDING_COUNT) > > -#define _MIPIA_VBP_COUNT (dev_priv->mipi_mmio_base + 0xb03c) > > -#define _MIPIC_VBP_COUNT (dev_priv->mipi_mmio_base + 0xb83c) > > +#define _MIPIA_VBP_COUNT (i915->mipi_mmio_base + 0xb03c) > > +#define _MIPIC_VBP_COUNT (i915->mipi_mmio_base + 0xb83c) > > #define MIPI_VBP_COUNT(port) _MMIO_MIPI(port, _MIPIA_VBP_COUNT, _MIPIC_VBP_COUNT) > > -#define _MIPIA_VFP_COUNT (dev_priv->mipi_mmio_base + 0xb040) > > -#define _MIPIC_VFP_COUNT (dev_priv->mipi_mmio_base + 0xb840) > > +#define _MIPIA_VFP_COUNT (i915->mipi_mmio_base + 0xb040) > > +#define _MIPIC_VFP_COUNT (i915->mipi_mmio_base + 0xb840) > > #define MIPI_VFP_COUNT(port) _MMIO_MIPI(port, _MIPIA_VFP_COUNT, _MIPIC_VFP_COUNT) > > -#define _MIPIA_HIGH_LOW_SWITCH_COUNT (dev_priv->mipi_mmio_base + 0xb044) > > -#define _MIPIC_HIGH_LOW_SWITCH_COUNT (dev_priv->mipi_mmio_base + 0xb844) > > +#define _MIPIA_HIGH_LOW_SWITCH_COUNT (i915->mipi_mmio_base + 0xb044) > > +#define _MIPIC_HIGH_LOW_SWITCH_COUNT (i915->mipi_mmio_base + 0xb844) > > #define MIPI_HIGH_LOW_SWITCH_COUNT(port) _MMIO_MIPI(port, _MIPIA_HIGH_LOW_SWITCH_COUNT, _MIPIC_HIGH_LOW_SWITCH_COUNT) > > /* regs above are bits 15:0 */ > > -#define _MIPIA_DPI_CONTROL (dev_priv->mipi_mmio_base + 0xb048) > > -#define _MIPIC_DPI_CONTROL (dev_priv->mipi_mmio_base + 0xb848) > > +#define _MIPIA_DPI_CONTROL (i915->mipi_mmio_base + 0xb048) > > +#define _MIPIC_DPI_CONTROL (i915->mipi_mmio_base + 0xb848) > > #define MIPI_DPI_CONTROL(port) _MMIO_MIPI(port, _MIPIA_DPI_CONTROL, _MIPIC_DPI_CONTROL) > > #define DPI_LP_MODE (1 << 6) > > #define BACKLIGHT_OFF (1 << 5) > > @@ -10622,27 +10622,27 @@ enum skl_power_gate { > > #define TURN_ON (1 << 1) > > #define SHUTDOWN (1 << 0) > > -#define _MIPIA_DPI_DATA (dev_priv->mipi_mmio_base + 0xb04c) > > -#define _MIPIC_DPI_DATA (dev_priv->mipi_mmio_base + 0xb84c) > > +#define _MIPIA_DPI_DATA (i915->mipi_mmio_base + 0xb04c) > > +#define _MIPIC_DPI_DATA (i915->mipi_mmio_base + 0xb84c) > > #define MIPI_DPI_DATA(port) _MMIO_MIPI(port, _MIPIA_DPI_DATA, _MIPIC_DPI_DATA) > > #define COMMAND_BYTE_SHIFT 0 > > #define COMMAND_BYTE_MASK (0x3f << 0) > > -#define _MIPIA_INIT_COUNT (dev_priv->mipi_mmio_base + 0xb050) > > -#define _MIPIC_INIT_COUNT (dev_priv->mipi_mmio_base + 0xb850) > > +#define _MIPIA_INIT_COUNT (i915->mipi_mmio_base + 0xb050) > > +#define _MIPIC_INIT_COUNT (i915->mipi_mmio_base + 0xb850) > > #define MIPI_INIT_COUNT(port) _MMIO_MIPI(port, _MIPIA_INIT_COUNT, _MIPIC_INIT_COUNT) > > #define MASTER_INIT_TIMER_SHIFT 0 > > #define MASTER_INIT_TIMER_MASK (0xffff << 0) > > -#define _MIPIA_MAX_RETURN_PKT_SIZE (dev_priv->mipi_mmio_base + 0xb054) > > -#define _MIPIC_MAX_RETURN_PKT_SIZE (dev_priv->mipi_mmio_base + 0xb854) > > +#define _MIPIA_MAX_RETURN_PKT_SIZE (i915->mipi_mmio_base + 0xb054) > > +#define _MIPIC_MAX_RETURN_PKT_SIZE (i915->mipi_mmio_base + 0xb854) > > #define MIPI_MAX_RETURN_PKT_SIZE(port) _MMIO_MIPI(port, \ > > _MIPIA_MAX_RETURN_PKT_SIZE, _MIPIC_MAX_RETURN_PKT_SIZE) > > #define MAX_RETURN_PKT_SIZE_SHIFT 0 > > #define MAX_RETURN_PKT_SIZE_MASK (0x3ff << 0) > > -#define _MIPIA_VIDEO_MODE_FORMAT (dev_priv->mipi_mmio_base + 0xb058) > > -#define _MIPIC_VIDEO_MODE_FORMAT (dev_priv->mipi_mmio_base + 0xb858) > > +#define _MIPIA_VIDEO_MODE_FORMAT (i915->mipi_mmio_base + 0xb058) > > +#define _MIPIC_VIDEO_MODE_FORMAT (i915->mipi_mmio_base + 0xb858) > > #define MIPI_VIDEO_MODE_FORMAT(port) _MMIO_MIPI(port, _MIPIA_VIDEO_MODE_FORMAT, _MIPIC_VIDEO_MODE_FORMAT) > > #define RANDOM_DPI_DISPLAY_RESOLUTION (1 << 4) > > #define DISABLE_VIDEO_BTA (1 << 3) > > @@ -10651,8 +10651,8 @@ enum skl_power_gate { > > #define VIDEO_MODE_NON_BURST_WITH_SYNC_EVENTS (2 << 0) > > #define VIDEO_MODE_BURST (3 << 0) > > -#define _MIPIA_EOT_DISABLE (dev_priv->mipi_mmio_base + 0xb05c) > > -#define _MIPIC_EOT_DISABLE (dev_priv->mipi_mmio_base + 0xb85c) > > +#define _MIPIA_EOT_DISABLE (i915->mipi_mmio_base + 0xb05c) > > +#define _MIPIC_EOT_DISABLE (i915->mipi_mmio_base + 0xb85c) > > #define MIPI_EOT_DISABLE(port) _MMIO_MIPI(port, _MIPIA_EOT_DISABLE, _MIPIC_EOT_DISABLE) > > #define BXT_DEFEATURE_DPI_FIFO_CTR (1 << 9) > > #define BXT_DPHY_DEFEATURE_EN (1 << 8) > > @@ -10665,35 +10665,35 @@ enum skl_power_gate { > > #define CLOCKSTOP (1 << 1) > > #define EOT_DISABLE (1 << 0) > > -#define _MIPIA_LP_BYTECLK (dev_priv->mipi_mmio_base + 0xb060) > > -#define _MIPIC_LP_BYTECLK (dev_priv->mipi_mmio_base + 0xb860) > > +#define _MIPIA_LP_BYTECLK (i915->mipi_mmio_base + 0xb060) > > +#define _MIPIC_LP_BYTECLK (i915->mipi_mmio_base + 0xb860) > > #define MIPI_LP_BYTECLK(port) _MMIO_MIPI(port, _MIPIA_LP_BYTECLK, _MIPIC_LP_BYTECLK) > > #define LP_BYTECLK_SHIFT 0 > > #define LP_BYTECLK_MASK (0xffff << 0) > > -#define _MIPIA_TLPX_TIME_COUNT (dev_priv->mipi_mmio_base + 0xb0a4) > > -#define _MIPIC_TLPX_TIME_COUNT (dev_priv->mipi_mmio_base + 0xb8a4) > > +#define _MIPIA_TLPX_TIME_COUNT (i915->mipi_mmio_base + 0xb0a4) > > +#define _MIPIC_TLPX_TIME_COUNT (i915->mipi_mmio_base + 0xb8a4) > > #define MIPI_TLPX_TIME_COUNT(port) _MMIO_MIPI(port, _MIPIA_TLPX_TIME_COUNT, _MIPIC_TLPX_TIME_COUNT) > > -#define _MIPIA_CLK_LANE_TIMING (dev_priv->mipi_mmio_base + 0xb098) > > -#define _MIPIC_CLK_LANE_TIMING (dev_priv->mipi_mmio_base + 0xb898) > > +#define _MIPIA_CLK_LANE_TIMING (i915->mipi_mmio_base + 0xb098) > > +#define _MIPIC_CLK_LANE_TIMING (i915->mipi_mmio_base + 0xb898) > > #define MIPI_CLK_LANE_TIMING(port) _MMIO_MIPI(port, _MIPIA_CLK_LANE_TIMING, _MIPIC_CLK_LANE_TIMING) > > /* bits 31:0 */ > > -#define _MIPIA_LP_GEN_DATA (dev_priv->mipi_mmio_base + 0xb064) > > -#define _MIPIC_LP_GEN_DATA (dev_priv->mipi_mmio_base + 0xb864) > > +#define _MIPIA_LP_GEN_DATA (i915->mipi_mmio_base + 0xb064) > > +#define _MIPIC_LP_GEN_DATA (i915->mipi_mmio_base + 0xb864) > > #define MIPI_LP_GEN_DATA(port) _MMIO_MIPI(port, _MIPIA_LP_GEN_DATA, _MIPIC_LP_GEN_DATA) > > /* bits 31:0 */ > > -#define _MIPIA_HS_GEN_DATA (dev_priv->mipi_mmio_base + 0xb068) > > -#define _MIPIC_HS_GEN_DATA (dev_priv->mipi_mmio_base + 0xb868) > > +#define _MIPIA_HS_GEN_DATA (i915->mipi_mmio_base + 0xb068) > > +#define _MIPIC_HS_GEN_DATA (i915->mipi_mmio_base + 0xb868) > > #define MIPI_HS_GEN_DATA(port) _MMIO_MIPI(port, _MIPIA_HS_GEN_DATA, _MIPIC_HS_GEN_DATA) > > -#define _MIPIA_LP_GEN_CTRL (dev_priv->mipi_mmio_base + 0xb06c) > > -#define _MIPIC_LP_GEN_CTRL (dev_priv->mipi_mmio_base + 0xb86c) > > +#define _MIPIA_LP_GEN_CTRL (i915->mipi_mmio_base + 0xb06c) > > +#define _MIPIC_LP_GEN_CTRL (i915->mipi_mmio_base + 0xb86c) > > #define MIPI_LP_GEN_CTRL(port) _MMIO_MIPI(port, _MIPIA_LP_GEN_CTRL, _MIPIC_LP_GEN_CTRL) > > -#define _MIPIA_HS_GEN_CTRL (dev_priv->mipi_mmio_base + 0xb070) > > -#define _MIPIC_HS_GEN_CTRL (dev_priv->mipi_mmio_base + 0xb870) > > +#define _MIPIA_HS_GEN_CTRL (i915->mipi_mmio_base + 0xb070) > > +#define _MIPIC_HS_GEN_CTRL (i915->mipi_mmio_base + 0xb870) > > #define MIPI_HS_GEN_CTRL(port) _MMIO_MIPI(port, _MIPIA_HS_GEN_CTRL, _MIPIC_HS_GEN_CTRL) > > #define LONG_PACKET_WORD_COUNT_SHIFT 8 > > #define LONG_PACKET_WORD_COUNT_MASK (0xffff << 8) > > @@ -10705,8 +10705,8 @@ enum skl_power_gate { > > #define DATA_TYPE_MASK (0x3f << 0) > > /* data type values, see include/video/mipi_display.h */ > > -#define _MIPIA_GEN_FIFO_STAT (dev_priv->mipi_mmio_base + 0xb074) > > -#define _MIPIC_GEN_FIFO_STAT (dev_priv->mipi_mmio_base + 0xb874) > > +#define _MIPIA_GEN_FIFO_STAT (i915->mipi_mmio_base + 0xb074) > > +#define _MIPIC_GEN_FIFO_STAT (i915->mipi_mmio_base + 0xb874) > > #define MIPI_GEN_FIFO_STAT(port) _MMIO_MIPI(port, _MIPIA_GEN_FIFO_STAT, _MIPIC_GEN_FIFO_STAT) > > #define DPI_FIFO_EMPTY (1 << 28) > > #define DBI_FIFO_EMPTY (1 << 27) > > @@ -10723,15 +10723,15 @@ enum skl_power_gate { > > #define HS_DATA_FIFO_HALF_EMPTY (1 << 1) > > #define HS_DATA_FIFO_FULL (1 << 0) > > -#define _MIPIA_HS_LS_DBI_ENABLE (dev_priv->mipi_mmio_base + 0xb078) > > -#define _MIPIC_HS_LS_DBI_ENABLE (dev_priv->mipi_mmio_base + 0xb878) > > +#define _MIPIA_HS_LS_DBI_ENABLE (i915->mipi_mmio_base + 0xb078) > > +#define _MIPIC_HS_LS_DBI_ENABLE (i915->mipi_mmio_base + 0xb878) > > #define MIPI_HS_LP_DBI_ENABLE(port) _MMIO_MIPI(port, _MIPIA_HS_LS_DBI_ENABLE, _MIPIC_HS_LS_DBI_ENABLE) > > #define DBI_HS_LP_MODE_MASK (1 << 0) > > #define DBI_LP_MODE (1 << 0) > > #define DBI_HS_MODE (0 << 0) > > -#define _MIPIA_DPHY_PARAM (dev_priv->mipi_mmio_base + 0xb080) > > -#define _MIPIC_DPHY_PARAM (dev_priv->mipi_mmio_base + 0xb880) > > +#define _MIPIA_DPHY_PARAM (i915->mipi_mmio_base + 0xb080) > > +#define _MIPIC_DPHY_PARAM (i915->mipi_mmio_base + 0xb880) > > #define MIPI_DPHY_PARAM(port) _MMIO_MIPI(port, _MIPIA_DPHY_PARAM, _MIPIC_DPHY_PARAM) > > #define EXIT_ZERO_COUNT_SHIFT 24 > > #define EXIT_ZERO_COUNT_MASK (0x3f << 24) > > @@ -10978,34 +10978,34 @@ enum skl_power_gate { > > #define TA_TIMEOUT_VALUE(x) ((x) << 0) > > /* bits 31:0 */ > > -#define _MIPIA_DBI_BW_CTRL (dev_priv->mipi_mmio_base + 0xb084) > > -#define _MIPIC_DBI_BW_CTRL (dev_priv->mipi_mmio_base + 0xb884) > > +#define _MIPIA_DBI_BW_CTRL (i915->mipi_mmio_base + 0xb084) > > +#define _MIPIC_DBI_BW_CTRL (i915->mipi_mmio_base + 0xb884) > > #define MIPI_DBI_BW_CTRL(port) _MMIO_MIPI(port, _MIPIA_DBI_BW_CTRL, _MIPIC_DBI_BW_CTRL) > > -#define _MIPIA_CLK_LANE_SWITCH_TIME_CNT (dev_priv->mipi_mmio_base + 0xb088) > > -#define _MIPIC_CLK_LANE_SWITCH_TIME_CNT (dev_priv->mipi_mmio_base + 0xb888) > > +#define _MIPIA_CLK_LANE_SWITCH_TIME_CNT (i915->mipi_mmio_base + 0xb088) > > +#define _MIPIC_CLK_LANE_SWITCH_TIME_CNT (i915->mipi_mmio_base + 0xb888) > > #define MIPI_CLK_LANE_SWITCH_TIME_CNT(port) _MMIO_MIPI(port, _MIPIA_CLK_LANE_SWITCH_TIME_CNT, _MIPIC_CLK_LANE_SWITCH_TIME_CNT) > > #define LP_HS_SSW_CNT_SHIFT 16 > > #define LP_HS_SSW_CNT_MASK (0xffff << 16) > > #define HS_LP_PWR_SW_CNT_SHIFT 0 > > #define HS_LP_PWR_SW_CNT_MASK (0xffff << 0) > > -#define _MIPIA_STOP_STATE_STALL (dev_priv->mipi_mmio_base + 0xb08c) > > -#define _MIPIC_STOP_STATE_STALL (dev_priv->mipi_mmio_base + 0xb88c) > > +#define _MIPIA_STOP_STATE_STALL (i915->mipi_mmio_base + 0xb08c) > > +#define _MIPIC_STOP_STATE_STALL (i915->mipi_mmio_base + 0xb88c) > > #define MIPI_STOP_STATE_STALL(port) _MMIO_MIPI(port, _MIPIA_STOP_STATE_STALL, _MIPIC_STOP_STATE_STALL) > > #define STOP_STATE_STALL_COUNTER_SHIFT 0 > > #define STOP_STATE_STALL_COUNTER_MASK (0xff << 0) > > -#define _MIPIA_INTR_STAT_REG_1 (dev_priv->mipi_mmio_base + 0xb090) > > -#define _MIPIC_INTR_STAT_REG_1 (dev_priv->mipi_mmio_base + 0xb890) > > +#define _MIPIA_INTR_STAT_REG_1 (i915->mipi_mmio_base + 0xb090) > > +#define _MIPIC_INTR_STAT_REG_1 (i915->mipi_mmio_base + 0xb890) > > #define MIPI_INTR_STAT_REG_1(port) _MMIO_MIPI(port, _MIPIA_INTR_STAT_REG_1, _MIPIC_INTR_STAT_REG_1) > > -#define _MIPIA_INTR_EN_REG_1 (dev_priv->mipi_mmio_base + 0xb094) > > -#define _MIPIC_INTR_EN_REG_1 (dev_priv->mipi_mmio_base + 0xb894) > > +#define _MIPIA_INTR_EN_REG_1 (i915->mipi_mmio_base + 0xb094) > > +#define _MIPIC_INTR_EN_REG_1 (i915->mipi_mmio_base + 0xb894) > > #define MIPI_INTR_EN_REG_1(port) _MMIO_MIPI(port, _MIPIA_INTR_EN_REG_1, _MIPIC_INTR_EN_REG_1) > > #define RX_CONTENTION_DETECTED (1 << 0) > > /* XXX: only pipe A ?!? */ > > -#define MIPIA_DBI_TYPEC_CTRL (dev_priv->mipi_mmio_base + 0xb100) > > +#define MIPIA_DBI_TYPEC_CTRL (i915->mipi_mmio_base + 0xb100) > > #define DBI_TYPEC_ENABLE (1 << 31) > > #define DBI_TYPEC_WIP (1 << 30) > > #define DBI_TYPEC_OPTION_SHIFT 28 > > @@ -11019,8 +11019,8 @@ enum skl_power_gate { > > /* MIPI adapter registers */ > > -#define _MIPIA_CTRL (dev_priv->mipi_mmio_base + 0xb104) > > -#define _MIPIC_CTRL (dev_priv->mipi_mmio_base + 0xb904) > > +#define _MIPIA_CTRL (i915->mipi_mmio_base + 0xb104) > > +#define _MIPIC_CTRL (i915->mipi_mmio_base + 0xb904) > > #define MIPI_CTRL(port) _MMIO_MIPI(port, _MIPIA_CTRL, _MIPIC_CTRL) > > #define ESCAPE_CLOCK_DIVIDER_SHIFT 5 /* A only */ > > #define ESCAPE_CLOCK_DIVIDER_MASK (3 << 5) > > @@ -11052,21 +11052,21 @@ enum skl_power_gate { > > #define GLK_MIPIIO_PORT_POWERED (1 << 1) /* RO */ > > #define GLK_MIPIIO_ENABLE (1 << 0) > > -#define _MIPIA_DATA_ADDRESS (dev_priv->mipi_mmio_base + 0xb108) > > -#define _MIPIC_DATA_ADDRESS (dev_priv->mipi_mmio_base + 0xb908) > > +#define _MIPIA_DATA_ADDRESS (i915->mipi_mmio_base + 0xb108) > > +#define _MIPIC_DATA_ADDRESS (i915->mipi_mmio_base + 0xb908) > > #define MIPI_DATA_ADDRESS(port) _MMIO_MIPI(port, _MIPIA_DATA_ADDRESS, _MIPIC_DATA_ADDRESS) > > #define DATA_MEM_ADDRESS_SHIFT 5 > > #define DATA_MEM_ADDRESS_MASK (0x7ffffff << 5) > > #define DATA_VALID (1 << 0) > > -#define _MIPIA_DATA_LENGTH (dev_priv->mipi_mmio_base + 0xb10c) > > -#define _MIPIC_DATA_LENGTH (dev_priv->mipi_mmio_base + 0xb90c) > > +#define _MIPIA_DATA_LENGTH (i915->mipi_mmio_base + 0xb10c) > > +#define _MIPIC_DATA_LENGTH (i915->mipi_mmio_base + 0xb90c) > > #define MIPI_DATA_LENGTH(port) _MMIO_MIPI(port, _MIPIA_DATA_LENGTH, _MIPIC_DATA_LENGTH) > > #define DATA_LENGTH_SHIFT 0 > > #define DATA_LENGTH_MASK (0xfffff << 0) > > -#define _MIPIA_COMMAND_ADDRESS (dev_priv->mipi_mmio_base + 0xb110) > > -#define _MIPIC_COMMAND_ADDRESS (dev_priv->mipi_mmio_base + 0xb910) > > +#define _MIPIA_COMMAND_ADDRESS (i915->mipi_mmio_base + 0xb110) > > +#define _MIPIC_COMMAND_ADDRESS (i915->mipi_mmio_base + 0xb910) > > #define MIPI_COMMAND_ADDRESS(port) _MMIO_MIPI(port, _MIPIA_COMMAND_ADDRESS, _MIPIC_COMMAND_ADDRESS) > > #define COMMAND_MEM_ADDRESS_SHIFT 5 > > #define COMMAND_MEM_ADDRESS_MASK (0x7ffffff << 5) > > @@ -11074,18 +11074,18 @@ enum skl_power_gate { > > #define MEMORY_WRITE_DATA_FROM_PIPE_RENDERING (1 << 1) > > #define COMMAND_VALID (1 << 0) > > -#define _MIPIA_COMMAND_LENGTH (dev_priv->mipi_mmio_base + 0xb114) > > -#define _MIPIC_COMMAND_LENGTH (dev_priv->mipi_mmio_base + 0xb914) > > +#define _MIPIA_COMMAND_LENGTH (i915->mipi_mmio_base + 0xb114) > > +#define _MIPIC_COMMAND_LENGTH (i915->mipi_mmio_base + 0xb914) > > #define MIPI_COMMAND_LENGTH(port) _MMIO_MIPI(port, _MIPIA_COMMAND_LENGTH, _MIPIC_COMMAND_LENGTH) > > #define COMMAND_LENGTH_SHIFT(n) (8 * (n)) /* n: 0...3 */ > > #define COMMAND_LENGTH_MASK(n) (0xff << (8 * (n))) > > -#define _MIPIA_READ_DATA_RETURN0 (dev_priv->mipi_mmio_base + 0xb118) > > -#define _MIPIC_READ_DATA_RETURN0 (dev_priv->mipi_mmio_base + 0xb918) > > +#define _MIPIA_READ_DATA_RETURN0 (i915->mipi_mmio_base + 0xb118) > > +#define _MIPIC_READ_DATA_RETURN0 (i915->mipi_mmio_base + 0xb918) > > #define MIPI_READ_DATA_RETURN(port, n) _MMIO(_MIPI(port, _MIPIA_READ_DATA_RETURN0, _MIPIC_READ_DATA_RETURN0) + 4 * (n)) /* n: 0...7 */ > > -#define _MIPIA_READ_DATA_VALID (dev_priv->mipi_mmio_base + 0xb138) > > -#define _MIPIC_READ_DATA_VALID (dev_priv->mipi_mmio_base + 0xb938) > > +#define _MIPIA_READ_DATA_VALID (i915->mipi_mmio_base + 0xb138) > > +#define _MIPIC_READ_DATA_VALID (i915->mipi_mmio_base + 0xb938) > > #define MIPI_READ_DATA_VALID(port) _MMIO_MIPI(port, _MIPIA_READ_DATA_VALID, _MIPIC_READ_DATA_VALID) > > #define READ_DATA_VALID(n) (1 << (n)) > > diff --git a/drivers/gpu/drm/i915/i915_suspend.c b/drivers/gpu/drm/i915/i915_suspend.c > > index 581201bcb81a..5b3ad7c9240e 100644 > > --- a/drivers/gpu/drm/i915/i915_suspend.c > > +++ b/drivers/gpu/drm/i915/i915_suspend.c > > @@ -31,121 +31,121 @@ > > #include "intel_fbc.h" > > #include "intel_gmbus.h" > > -static void i915_save_display(struct drm_i915_private *dev_priv) > > +static void i915_save_display(struct drm_i915_private *i915) > > { > > /* Display arbitration control */ > > - if (INTEL_GEN(dev_priv) <= 4) > > - dev_priv->regfile.saveDSPARB = I915_READ(DSPARB); > > + if (INTEL_GEN(i915) <= 4) > > + i915->regfile.saveDSPARB = I915_READ(DSPARB); > > /* save FBC interval */ > > - if (HAS_FBC(dev_priv) && INTEL_GEN(dev_priv) <= 4 && !IS_G4X(dev_priv)) > > - dev_priv->regfile.saveFBC_CONTROL = I915_READ(FBC_CONTROL); > > + if (HAS_FBC(i915) && INTEL_GEN(i915) <= 4 && !IS_G4X(i915)) > > + i915->regfile.saveFBC_CONTROL = I915_READ(FBC_CONTROL); > > } > > -static void i915_restore_display(struct drm_i915_private *dev_priv) > > +static void i915_restore_display(struct drm_i915_private *i915) > > { > > /* Display arbitration */ > > - if (INTEL_GEN(dev_priv) <= 4) > > - I915_WRITE(DSPARB, dev_priv->regfile.saveDSPARB); > > + if (INTEL_GEN(i915) <= 4) > > + I915_WRITE(DSPARB, i915->regfile.saveDSPARB); > > /* only restore FBC info on the platform that supports FBC*/ > > - intel_fbc_global_disable(dev_priv); > > + intel_fbc_global_disable(i915); > > /* restore FBC interval */ > > - if (HAS_FBC(dev_priv) && INTEL_GEN(dev_priv) <= 4 && !IS_G4X(dev_priv)) > > - I915_WRITE(FBC_CONTROL, dev_priv->regfile.saveFBC_CONTROL); > > + if (HAS_FBC(i915) && INTEL_GEN(i915) <= 4 && !IS_G4X(i915)) > > + I915_WRITE(FBC_CONTROL, i915->regfile.saveFBC_CONTROL); > > - i915_redisable_vga(dev_priv); > > + i915_redisable_vga(i915); > > } > > -int i915_save_state(struct drm_i915_private *dev_priv) > > +int i915_save_state(struct drm_i915_private *i915) > > { > > - struct pci_dev *pdev = dev_priv->drm.pdev; > > + struct pci_dev *pdev = i915->drm.pdev; > > int i; > > - mutex_lock(&dev_priv->drm.struct_mutex); > > + mutex_lock(&i915->drm.struct_mutex); > > - i915_save_display(dev_priv); > > + i915_save_display(i915); > > - if (IS_GEN(dev_priv, 4)) > > + if (IS_GEN(i915, 4)) > > pci_read_config_word(pdev, GCDGMBUS, > > - &dev_priv->regfile.saveGCDGMBUS); > > + &i915->regfile.saveGCDGMBUS); > > /* Cache mode state */ > > - if (INTEL_GEN(dev_priv) < 7) > > - dev_priv->regfile.saveCACHE_MODE_0 = I915_READ(CACHE_MODE_0); > > + if (INTEL_GEN(i915) < 7) > > + i915->regfile.saveCACHE_MODE_0 = I915_READ(CACHE_MODE_0); > > /* Memory Arbitration state */ > > - dev_priv->regfile.saveMI_ARB_STATE = I915_READ(MI_ARB_STATE); > > + i915->regfile.saveMI_ARB_STATE = I915_READ(MI_ARB_STATE); > > /* Scratch space */ > > - if (IS_GEN(dev_priv, 2) && IS_MOBILE(dev_priv)) { > > + if (IS_GEN(i915, 2) && IS_MOBILE(i915)) { > > for (i = 0; i < 7; i++) { > > - dev_priv->regfile.saveSWF0[i] = I915_READ(SWF0(i)); > > - dev_priv->regfile.saveSWF1[i] = I915_READ(SWF1(i)); > > + i915->regfile.saveSWF0[i] = I915_READ(SWF0(i)); > > + i915->regfile.saveSWF1[i] = I915_READ(SWF1(i)); > > } > > for (i = 0; i < 3; i++) > > - dev_priv->regfile.saveSWF3[i] = I915_READ(SWF3(i)); > > - } else if (IS_GEN(dev_priv, 2)) { > > + i915->regfile.saveSWF3[i] = I915_READ(SWF3(i)); > > + } else if (IS_GEN(i915, 2)) { > > for (i = 0; i < 7; i++) > > - dev_priv->regfile.saveSWF1[i] = I915_READ(SWF1(i)); > > - } else if (HAS_GMCH(dev_priv)) { > > + i915->regfile.saveSWF1[i] = I915_READ(SWF1(i)); > > + } else if (HAS_GMCH(i915)) { > > for (i = 0; i < 16; i++) { > > - dev_priv->regfile.saveSWF0[i] = I915_READ(SWF0(i)); > > - dev_priv->regfile.saveSWF1[i] = I915_READ(SWF1(i)); > > + i915->regfile.saveSWF0[i] = I915_READ(SWF0(i)); > > + i915->regfile.saveSWF1[i] = I915_READ(SWF1(i)); > > } > > for (i = 0; i < 3; i++) > > - dev_priv->regfile.saveSWF3[i] = I915_READ(SWF3(i)); > > + i915->regfile.saveSWF3[i] = I915_READ(SWF3(i)); > > } > > - mutex_unlock(&dev_priv->drm.struct_mutex); > > + mutex_unlock(&i915->drm.struct_mutex); > > return 0; > > } > > -int i915_restore_state(struct drm_i915_private *dev_priv) > > +int i915_restore_state(struct drm_i915_private *i915) > > { > > - struct pci_dev *pdev = dev_priv->drm.pdev; > > + struct pci_dev *pdev = i915->drm.pdev; > > int i; > > - mutex_lock(&dev_priv->drm.struct_mutex); > > + mutex_lock(&i915->drm.struct_mutex); > > - if (IS_GEN(dev_priv, 4)) > > + if (IS_GEN(i915, 4)) > > pci_write_config_word(pdev, GCDGMBUS, > > - dev_priv->regfile.saveGCDGMBUS); > > - i915_restore_display(dev_priv); > > + i915->regfile.saveGCDGMBUS); > > + i915_restore_display(i915); > > /* Cache mode state */ > > - if (INTEL_GEN(dev_priv) < 7) > > - I915_WRITE(CACHE_MODE_0, dev_priv->regfile.saveCACHE_MODE_0 | > > + if (INTEL_GEN(i915) < 7) > > + I915_WRITE(CACHE_MODE_0, i915->regfile.saveCACHE_MODE_0 | > > 0xffff0000); > > /* Memory arbitration state */ > > - I915_WRITE(MI_ARB_STATE, dev_priv->regfile.saveMI_ARB_STATE | 0xffff0000); > > + I915_WRITE(MI_ARB_STATE, i915->regfile.saveMI_ARB_STATE | 0xffff0000); > > /* Scratch space */ > > - if (IS_GEN(dev_priv, 2) && IS_MOBILE(dev_priv)) { > > + if (IS_GEN(i915, 2) && IS_MOBILE(i915)) { > > for (i = 0; i < 7; i++) { > > - I915_WRITE(SWF0(i), dev_priv->regfile.saveSWF0[i]); > > - I915_WRITE(SWF1(i), dev_priv->regfile.saveSWF1[i]); > > + I915_WRITE(SWF0(i), i915->regfile.saveSWF0[i]); > > + I915_WRITE(SWF1(i), i915->regfile.saveSWF1[i]); > > } > > for (i = 0; i < 3; i++) > > - I915_WRITE(SWF3(i), dev_priv->regfile.saveSWF3[i]); > > - } else if (IS_GEN(dev_priv, 2)) { > > + I915_WRITE(SWF3(i), i915->regfile.saveSWF3[i]); > > + } else if (IS_GEN(i915, 2)) { > > for (i = 0; i < 7; i++) > > - I915_WRITE(SWF1(i), dev_priv->regfile.saveSWF1[i]); > > - } else if (HAS_GMCH(dev_priv)) { > > + I915_WRITE(SWF1(i), i915->regfile.saveSWF1[i]); > > + } else if (HAS_GMCH(i915)) { > > for (i = 0; i < 16; i++) { > > - I915_WRITE(SWF0(i), dev_priv->regfile.saveSWF0[i]); > > - I915_WRITE(SWF1(i), dev_priv->regfile.saveSWF1[i]); > > + I915_WRITE(SWF0(i), i915->regfile.saveSWF0[i]); > > + I915_WRITE(SWF1(i), i915->regfile.saveSWF1[i]); > > } > > for (i = 0; i < 3; i++) > > - I915_WRITE(SWF3(i), dev_priv->regfile.saveSWF3[i]); > > + I915_WRITE(SWF3(i), i915->regfile.saveSWF3[i]); > > } > > - mutex_unlock(&dev_priv->drm.struct_mutex); > > + mutex_unlock(&i915->drm.struct_mutex); > > - intel_gmbus_reset(dev_priv); > > + intel_gmbus_reset(i915); > > return 0; > > } > > diff --git a/drivers/gpu/drm/i915/i915_sysfs.c b/drivers/gpu/drm/i915/i915_sysfs.c > > index 3ef07b987d40..9c89be3ec4c1 100644 > > --- a/drivers/gpu/drm/i915/i915_sysfs.c > > +++ b/drivers/gpu/drm/i915/i915_sysfs.c > > @@ -42,14 +42,14 @@ static inline struct drm_i915_private *kdev_minor_to_i915(struct device *kdev) > > } > > #ifdef CONFIG_PM > > -static u32 calc_residency(struct drm_i915_private *dev_priv, > > +static u32 calc_residency(struct drm_i915_private *i915, > > i915_reg_t reg) > > { > > intel_wakeref_t wakeref; > > u64 res = 0; > > - with_intel_runtime_pm(dev_priv, wakeref) > > - res = intel_rc6_residency_us(dev_priv, reg); > > + with_intel_runtime_pm(i915, wakeref) > > + res = intel_rc6_residency_us(i915, reg); > > return DIV_ROUND_CLOSEST_ULL(res, 1000); > > } > > @@ -57,15 +57,15 @@ static u32 calc_residency(struct drm_i915_private *dev_priv, > > static ssize_t > > show_rc6_mask(struct device *kdev, struct device_attribute *attr, char *buf) > > { > > - struct drm_i915_private *dev_priv = kdev_minor_to_i915(kdev); > > + struct drm_i915_private *i915 = kdev_minor_to_i915(kdev); > > unsigned int mask; > > mask = 0; > > - if (HAS_RC6(dev_priv)) > > + if (HAS_RC6(i915)) > > mask |= BIT(0); > > - if (HAS_RC6p(dev_priv)) > > + if (HAS_RC6p(i915)) > > mask |= BIT(1); > > - if (HAS_RC6pp(dev_priv)) > > + if (HAS_RC6pp(i915)) > > mask |= BIT(2); > > return snprintf(buf, PAGE_SIZE, "%x\n", mask); > > @@ -74,32 +74,32 @@ show_rc6_mask(struct device *kdev, struct device_attribute *attr, char *buf) > > static ssize_t > > show_rc6_ms(struct device *kdev, struct device_attribute *attr, char *buf) > > { > > - struct drm_i915_private *dev_priv = kdev_minor_to_i915(kdev); > > - u32 rc6_residency = calc_residency(dev_priv, GEN6_GT_GFX_RC6); > > + struct drm_i915_private *i915 = kdev_minor_to_i915(kdev); > > + u32 rc6_residency = calc_residency(i915, GEN6_GT_GFX_RC6); > > return snprintf(buf, PAGE_SIZE, "%u\n", rc6_residency); > > } > > static ssize_t > > show_rc6p_ms(struct device *kdev, struct device_attribute *attr, char *buf) > > { > > - struct drm_i915_private *dev_priv = kdev_minor_to_i915(kdev); > > - u32 rc6p_residency = calc_residency(dev_priv, GEN6_GT_GFX_RC6p); > > + struct drm_i915_private *i915 = kdev_minor_to_i915(kdev); > > + u32 rc6p_residency = calc_residency(i915, GEN6_GT_GFX_RC6p); > > return snprintf(buf, PAGE_SIZE, "%u\n", rc6p_residency); > > } > > static ssize_t > > show_rc6pp_ms(struct device *kdev, struct device_attribute *attr, char *buf) > > { > > - struct drm_i915_private *dev_priv = kdev_minor_to_i915(kdev); > > - u32 rc6pp_residency = calc_residency(dev_priv, GEN6_GT_GFX_RC6pp); > > + struct drm_i915_private *i915 = kdev_minor_to_i915(kdev); > > + u32 rc6pp_residency = calc_residency(i915, GEN6_GT_GFX_RC6pp); > > return snprintf(buf, PAGE_SIZE, "%u\n", rc6pp_residency); > > } > > static ssize_t > > show_media_rc6_ms(struct device *kdev, struct device_attribute *attr, char *buf) > > { > > - struct drm_i915_private *dev_priv = kdev_minor_to_i915(kdev); > > - u32 rc6_residency = calc_residency(dev_priv, VLV_GT_MEDIA_RC6); > > + struct drm_i915_private *i915 = kdev_minor_to_i915(kdev); > > + u32 rc6_residency = calc_residency(i915, VLV_GT_MEDIA_RC6); > > return snprintf(buf, PAGE_SIZE, "%u\n", rc6_residency); > > } > > @@ -142,9 +142,9 @@ static const struct attribute_group media_rc6_attr_group = { > > }; > > #endif > > -static int l3_access_valid(struct drm_i915_private *dev_priv, loff_t offset) > > +static int l3_access_valid(struct drm_i915_private *i915, loff_t offset) > > { > > - if (!HAS_L3_DPF(dev_priv)) > > + if (!HAS_L3_DPF(i915)) > > return -EPERM; > > if (offset % 4 != 0) > > @@ -162,14 +162,14 @@ i915_l3_read(struct file *filp, struct kobject *kobj, > > loff_t offset, size_t count) > > { > > struct device *kdev = kobj_to_dev(kobj); > > - struct drm_i915_private *dev_priv = kdev_minor_to_i915(kdev); > > - struct drm_device *dev = &dev_priv->drm; > > + struct drm_i915_private *i915 = kdev_minor_to_i915(kdev); > > + struct drm_device *dev = &i915->drm; > > int slice = (int)(uintptr_t)attr->private; > > int ret; > > count = round_down(count, 4); > > - ret = l3_access_valid(dev_priv, offset); > > + ret = l3_access_valid(i915, offset); > > if (ret) > > return ret; > > @@ -179,9 +179,9 @@ i915_l3_read(struct file *filp, struct kobject *kobj, > > if (ret) > > return ret; > > - if (dev_priv->l3_parity.remap_info[slice]) > > + if (i915->l3_parity.remap_info[slice]) > > memcpy(buf, > > - dev_priv->l3_parity.remap_info[slice] + (offset/4), > > + i915->l3_parity.remap_info[slice] + (offset/4), > > count); > > else > > memset(buf, 0, count); > > @@ -197,14 +197,14 @@ i915_l3_write(struct file *filp, struct kobject *kobj, > > loff_t offset, size_t count) > > { > > struct device *kdev = kobj_to_dev(kobj); > > - struct drm_i915_private *dev_priv = kdev_minor_to_i915(kdev); > > - struct drm_device *dev = &dev_priv->drm; > > + struct drm_i915_private *i915 = kdev_minor_to_i915(kdev); > > + struct drm_device *dev = &i915->drm; > > struct i915_gem_context *ctx; > > int slice = (int)(uintptr_t)attr->private; > > u32 **remap_info; > > int ret; > > - ret = l3_access_valid(dev_priv, offset); > > + ret = l3_access_valid(i915, offset); > > if (ret) > > return ret; > > @@ -212,7 +212,7 @@ i915_l3_write(struct file *filp, struct kobject *kobj, > > if (ret) > > return ret; > > - remap_info = &dev_priv->l3_parity.remap_info[slice]; > > + remap_info = &i915->l3_parity.remap_info[slice]; > > if (!*remap_info) { > > *remap_info = kzalloc(GEN7_L3LOG_SIZE, GFP_KERNEL); > > if (!*remap_info) { > > @@ -228,7 +228,7 @@ i915_l3_write(struct file *filp, struct kobject *kobj, > > memcpy(*remap_info + (offset/4), buf, count); > > /* NB: We defer the remapping until we switch to the context */ > > - list_for_each_entry(ctx, &dev_priv->contexts.list, link) > > + list_for_each_entry(ctx, &i915->contexts.list, link) > > ctx->remap_slice |= (1<<slice); > > ret = count; > > @@ -260,52 +260,52 @@ static const struct bin_attribute dpf_attrs_1 = { > > static ssize_t gt_act_freq_mhz_show(struct device *kdev, > > struct device_attribute *attr, char *buf) > > { > > - struct drm_i915_private *dev_priv = kdev_minor_to_i915(kdev); > > + struct drm_i915_private *i915 = kdev_minor_to_i915(kdev); > > intel_wakeref_t wakeref; > > u32 freq; > > - wakeref = intel_runtime_pm_get(dev_priv); > > + wakeref = intel_runtime_pm_get(i915); > > - if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) { > > - vlv_punit_get(dev_priv); > > - freq = vlv_punit_read(dev_priv, PUNIT_REG_GPU_FREQ_STS); > > - vlv_punit_put(dev_priv); > > + if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) { > > + vlv_punit_get(i915); > > + freq = vlv_punit_read(i915, PUNIT_REG_GPU_FREQ_STS); > > + vlv_punit_put(i915); > > freq = (freq >> 8) & 0xff; > > } else { > > - freq = intel_get_cagf(dev_priv, I915_READ(GEN6_RPSTAT1)); > > + freq = intel_get_cagf(i915, I915_READ(GEN6_RPSTAT1)); > > } > > - intel_runtime_pm_put(dev_priv, wakeref); > > + intel_runtime_pm_put(i915, wakeref); > > - return snprintf(buf, PAGE_SIZE, "%d\n", intel_gpu_freq(dev_priv, freq)); > > + return snprintf(buf, PAGE_SIZE, "%d\n", intel_gpu_freq(i915, freq)); > > } > > static ssize_t gt_cur_freq_mhz_show(struct device *kdev, > > struct device_attribute *attr, char *buf) > > { > > - struct drm_i915_private *dev_priv = kdev_minor_to_i915(kdev); > > + struct drm_i915_private *i915 = kdev_minor_to_i915(kdev); > > return snprintf(buf, PAGE_SIZE, "%d\n", > > - intel_gpu_freq(dev_priv, > > - dev_priv->gt_pm.rps.cur_freq)); > > + intel_gpu_freq(i915, > > + i915->gt_pm.rps.cur_freq)); > > } > > static ssize_t gt_boost_freq_mhz_show(struct device *kdev, struct device_attribute *attr, char *buf) > > { > > - struct drm_i915_private *dev_priv = kdev_minor_to_i915(kdev); > > + struct drm_i915_private *i915 = kdev_minor_to_i915(kdev); > > return snprintf(buf, PAGE_SIZE, "%d\n", > > - intel_gpu_freq(dev_priv, > > - dev_priv->gt_pm.rps.boost_freq)); > > + intel_gpu_freq(i915, > > + i915->gt_pm.rps.boost_freq)); > > } > > static ssize_t gt_boost_freq_mhz_store(struct device *kdev, > > struct device_attribute *attr, > > const char *buf, size_t count) > > { > > - struct drm_i915_private *dev_priv = kdev_minor_to_i915(kdev); > > - struct intel_rps *rps = &dev_priv->gt_pm.rps; > > + struct drm_i915_private *i915 = kdev_minor_to_i915(kdev); > > + struct intel_rps *rps = &i915->gt_pm.rps; > > bool boost = false; > > ssize_t ret; > > u32 val; > > @@ -315,7 +315,7 @@ static ssize_t gt_boost_freq_mhz_store(struct device *kdev, > > return ret; > > /* Validate against (static) hardware limits */ > > - val = intel_freq_opcode(dev_priv, val); > > + val = intel_freq_opcode(i915, val); > > if (val < rps->min_freq || val > rps->max_freq) > > return -EINVAL; > > @@ -334,28 +334,28 @@ static ssize_t gt_boost_freq_mhz_store(struct device *kdev, > > static ssize_t vlv_rpe_freq_mhz_show(struct device *kdev, > > struct device_attribute *attr, char *buf) > > { > > - struct drm_i915_private *dev_priv = kdev_minor_to_i915(kdev); > > + struct drm_i915_private *i915 = kdev_minor_to_i915(kdev); > > return snprintf(buf, PAGE_SIZE, "%d\n", > > - intel_gpu_freq(dev_priv, > > - dev_priv->gt_pm.rps.efficient_freq)); > > + intel_gpu_freq(i915, > > + i915->gt_pm.rps.efficient_freq)); > > } > > static ssize_t gt_max_freq_mhz_show(struct device *kdev, struct device_attribute *attr, char *buf) > > { > > - struct drm_i915_private *dev_priv = kdev_minor_to_i915(kdev); > > + struct drm_i915_private *i915 = kdev_minor_to_i915(kdev); > > return snprintf(buf, PAGE_SIZE, "%d\n", > > - intel_gpu_freq(dev_priv, > > - dev_priv->gt_pm.rps.max_freq_softlimit)); > > + intel_gpu_freq(i915, > > + i915->gt_pm.rps.max_freq_softlimit)); > > } > > static ssize_t gt_max_freq_mhz_store(struct device *kdev, > > struct device_attribute *attr, > > const char *buf, size_t count) > > { > > - struct drm_i915_private *dev_priv = kdev_minor_to_i915(kdev); > > - struct intel_rps *rps = &dev_priv->gt_pm.rps; > > + struct drm_i915_private *i915 = kdev_minor_to_i915(kdev); > > + struct intel_rps *rps = &i915->gt_pm.rps; > > intel_wakeref_t wakeref; > > u32 val; > > ssize_t ret; > > @@ -364,10 +364,10 @@ static ssize_t gt_max_freq_mhz_store(struct device *kdev, > > if (ret) > > return ret; > > - wakeref = intel_runtime_pm_get(dev_priv); > > + wakeref = intel_runtime_pm_get(i915); > > mutex_lock(&rps->lock); > > - val = intel_freq_opcode(dev_priv, val); > > + val = intel_freq_opcode(i915, val); > > if (val < rps->min_freq || > > val > rps->max_freq || > > val < rps->min_freq_softlimit) { > > @@ -377,7 +377,7 @@ static ssize_t gt_max_freq_mhz_store(struct device *kdev, > > if (val > rps->rp0_freq) > > DRM_DEBUG("User requested overclocking to %d\n", > > - intel_gpu_freq(dev_priv, val)); > > + intel_gpu_freq(i915, val)); > > rps->max_freq_softlimit = val; > > @@ -388,30 +388,30 @@ static ssize_t gt_max_freq_mhz_store(struct device *kdev, > > /* We still need *_set_rps to process the new max_delay and > > * update the interrupt limits and PMINTRMSK even though > > * frequency request may be unchanged. */ > > - ret = intel_set_rps(dev_priv, val); > > + ret = intel_set_rps(i915, val); > > unlock: > > mutex_unlock(&rps->lock); > > - intel_runtime_pm_put(dev_priv, wakeref); > > + intel_runtime_pm_put(i915, wakeref); > > return ret ?: count; > > } > > static ssize_t gt_min_freq_mhz_show(struct device *kdev, struct device_attribute *attr, char *buf) > > { > > - struct drm_i915_private *dev_priv = kdev_minor_to_i915(kdev); > > + struct drm_i915_private *i915 = kdev_minor_to_i915(kdev); > > return snprintf(buf, PAGE_SIZE, "%d\n", > > - intel_gpu_freq(dev_priv, > > - dev_priv->gt_pm.rps.min_freq_softlimit)); > > + intel_gpu_freq(i915, > > + i915->gt_pm.rps.min_freq_softlimit)); > > } > > static ssize_t gt_min_freq_mhz_store(struct device *kdev, > > struct device_attribute *attr, > > const char *buf, size_t count) > > { > > - struct drm_i915_private *dev_priv = kdev_minor_to_i915(kdev); > > - struct intel_rps *rps = &dev_priv->gt_pm.rps; > > + struct drm_i915_private *i915 = kdev_minor_to_i915(kdev); > > + struct intel_rps *rps = &i915->gt_pm.rps; > > intel_wakeref_t wakeref; > > u32 val; > > ssize_t ret; > > @@ -420,10 +420,10 @@ static ssize_t gt_min_freq_mhz_store(struct device *kdev, > > if (ret) > > return ret; > > - wakeref = intel_runtime_pm_get(dev_priv); > > + wakeref = intel_runtime_pm_get(i915); > > mutex_lock(&rps->lock); > > - val = intel_freq_opcode(dev_priv, val); > > + val = intel_freq_opcode(i915, val); > > if (val < rps->min_freq || > > val > rps->max_freq || > > val > rps->max_freq_softlimit) { > > @@ -440,11 +440,11 @@ static ssize_t gt_min_freq_mhz_store(struct device *kdev, > > /* We still need *_set_rps to process the new min_delay and > > * update the interrupt limits and PMINTRMSK even though > > * frequency request may be unchanged. */ > > - ret = intel_set_rps(dev_priv, val); > > + ret = intel_set_rps(i915, val); > > unlock: > > mutex_unlock(&rps->lock); > > - intel_runtime_pm_put(dev_priv, wakeref); > > + intel_runtime_pm_put(i915, wakeref); > > return ret ?: count; > > } > > @@ -465,16 +465,16 @@ static DEVICE_ATTR(gt_RPn_freq_mhz, S_IRUGO, gt_rp_mhz_show, NULL); > > /* For now we have a static number of RP states */ > > static ssize_t gt_rp_mhz_show(struct device *kdev, struct device_attribute *attr, char *buf) > > { > > - struct drm_i915_private *dev_priv = kdev_minor_to_i915(kdev); > > - struct intel_rps *rps = &dev_priv->gt_pm.rps; > > + struct drm_i915_private *i915 = kdev_minor_to_i915(kdev); > > + struct intel_rps *rps = &i915->gt_pm.rps; > > u32 val; > > if (attr == &dev_attr_gt_RP0_freq_mhz) > > - val = intel_gpu_freq(dev_priv, rps->rp0_freq); > > + val = intel_gpu_freq(i915, rps->rp0_freq); > > else if (attr == &dev_attr_gt_RP1_freq_mhz) > > - val = intel_gpu_freq(dev_priv, rps->rp1_freq); > > + val = intel_gpu_freq(i915, rps->rp1_freq); > > else if (attr == &dev_attr_gt_RPn_freq_mhz) > > - val = intel_gpu_freq(dev_priv, rps->min_freq); > > + val = intel_gpu_freq(i915, rps->min_freq); > > else > > BUG(); > > @@ -540,10 +540,10 @@ static ssize_t error_state_write(struct file *file, struct kobject *kobj, > > loff_t off, size_t count) > > { > > struct device *kdev = kobj_to_dev(kobj); > > - struct drm_i915_private *dev_priv = kdev_minor_to_i915(kdev); > > + struct drm_i915_private *i915 = kdev_minor_to_i915(kdev); > > DRM_DEBUG_DRIVER("Resetting error state\n"); > > - i915_reset_error_state(dev_priv); > > + i915_reset_error_state(i915); > > return count; > > } > > @@ -571,37 +571,37 @@ static void i915_setup_error_capture(struct device *kdev) {} > > static void i915_teardown_error_capture(struct device *kdev) {} > > #endif > > -void i915_setup_sysfs(struct drm_i915_private *dev_priv) > > +void i915_setup_sysfs(struct drm_i915_private *i915) > > { > > - struct device *kdev = dev_priv->drm.primary->kdev; > > + struct device *kdev = i915->drm.primary->kdev; > > int ret; > > #ifdef CONFIG_PM > > - if (HAS_RC6(dev_priv)) { > > + if (HAS_RC6(i915)) { > > ret = sysfs_merge_group(&kdev->kobj, > > &rc6_attr_group); > > if (ret) > > DRM_ERROR("RC6 residency sysfs setup failed\n"); > > } > > - if (HAS_RC6p(dev_priv)) { > > + if (HAS_RC6p(i915)) { > > ret = sysfs_merge_group(&kdev->kobj, > > &rc6p_attr_group); > > if (ret) > > DRM_ERROR("RC6p residency sysfs setup failed\n"); > > } > > - if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) { > > + if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) { > > ret = sysfs_merge_group(&kdev->kobj, > > &media_rc6_attr_group); > > if (ret) > > DRM_ERROR("Media RC6 residency sysfs setup failed\n"); > > } > > #endif > > - if (HAS_L3_DPF(dev_priv)) { > > + if (HAS_L3_DPF(i915)) { > > ret = device_create_bin_file(kdev, &dpf_attrs); > > if (ret) > > DRM_ERROR("l3 parity sysfs setup failed\n"); > > - if (NUM_L3_SLICES(dev_priv) > 1) { > > + if (NUM_L3_SLICES(i915) > 1) { > > ret = device_create_bin_file(kdev, > > &dpf_attrs_1); > > if (ret) > > @@ -610,9 +610,9 @@ void i915_setup_sysfs(struct drm_i915_private *dev_priv) > > } > > ret = 0; > > - if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) > > + if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) > > ret = sysfs_create_files(&kdev->kobj, vlv_attrs); > > - else if (INTEL_GEN(dev_priv) >= 6) > > + else if (INTEL_GEN(i915) >= 6) > > ret = sysfs_create_files(&kdev->kobj, gen6_attrs); > > if (ret) > > DRM_ERROR("RPS sysfs setup failed\n"); > > @@ -620,13 +620,13 @@ void i915_setup_sysfs(struct drm_i915_private *dev_priv) > > i915_setup_error_capture(kdev); > > } > > -void i915_teardown_sysfs(struct drm_i915_private *dev_priv) > > +void i915_teardown_sysfs(struct drm_i915_private *i915) > > { > > - struct device *kdev = dev_priv->drm.primary->kdev; > > + struct device *kdev = i915->drm.primary->kdev; > > i915_teardown_error_capture(kdev); > > - if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) > > + if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) > > sysfs_remove_files(&kdev->kobj, vlv_attrs); > > else > > sysfs_remove_files(&kdev->kobj, gen6_attrs); > > diff --git a/drivers/gpu/drm/i915/i915_trace.h b/drivers/gpu/drm/i915/i915_trace.h > > index 5c8cfaa70d72..9548b9fe4199 100644 > > --- a/drivers/gpu/drm/i915/i915_trace.h > > +++ b/drivers/gpu/drm/i915/i915_trace.h > > @@ -21,8 +21,8 @@ > > /* watermark/fifo updates */ > > TRACE_EVENT(intel_pipe_enable, > > - TP_PROTO(struct drm_i915_private *dev_priv, enum pipe pipe), > > - TP_ARGS(dev_priv, pipe), > > + TP_PROTO(struct drm_i915_private *i915, enum pipe pipe), > > + TP_ARGS(i915, pipe), > > TP_STRUCT__entry( > > __array(u32, frame, 3) > > @@ -32,11 +32,11 @@ TRACE_EVENT(intel_pipe_enable, > > TP_fast_assign( > > enum pipe _pipe; > > - for_each_pipe(dev_priv, _pipe) { > > + for_each_pipe(i915, _pipe) { > > __entry->frame[_pipe] = > > - dev_priv->drm.driver->get_vblank_counter(&dev_priv->drm, _pipe); > > + i915->drm.driver->get_vblank_counter(&i915->drm, _pipe); > > __entry->scanline[_pipe] = > > - intel_get_crtc_scanline(intel_get_crtc_for_pipe(dev_priv, _pipe)); > > + intel_get_crtc_scanline(intel_get_crtc_for_pipe(i915, _pipe)); > > } > > __entry->pipe = pipe; > > ), > > @@ -49,8 +49,8 @@ TRACE_EVENT(intel_pipe_enable, > > ); > > TRACE_EVENT(intel_pipe_disable, > > - TP_PROTO(struct drm_i915_private *dev_priv, enum pipe pipe), > > - TP_ARGS(dev_priv, pipe), > > + TP_PROTO(struct drm_i915_private *i915, enum pipe pipe), > > + TP_ARGS(i915, pipe), > > TP_STRUCT__entry( > > __array(u32, frame, 3) > > @@ -60,11 +60,11 @@ TRACE_EVENT(intel_pipe_disable, > > TP_fast_assign( > > enum pipe _pipe; > > - for_each_pipe(dev_priv, _pipe) { > > + for_each_pipe(i915, _pipe) { > > __entry->frame[_pipe] = > > - dev_priv->drm.driver->get_vblank_counter(&dev_priv->drm, _pipe); > > + i915->drm.driver->get_vblank_counter(&i915->drm, _pipe); > > __entry->scanline[_pipe] = > > - intel_get_crtc_scanline(intel_get_crtc_for_pipe(dev_priv, _pipe)); > > + intel_get_crtc_scanline(intel_get_crtc_for_pipe(i915, _pipe)); > > } > > __entry->pipe = pipe; > > ), > > @@ -102,8 +102,8 @@ TRACE_EVENT(intel_pipe_crc, > > ); > > TRACE_EVENT(intel_cpu_fifo_underrun, > > - TP_PROTO(struct drm_i915_private *dev_priv, enum pipe pipe), > > - TP_ARGS(dev_priv, pipe), > > + TP_PROTO(struct drm_i915_private *i915, enum pipe pipe), > > + TP_ARGS(i915, pipe), > > TP_STRUCT__entry( > > __field(enum pipe, pipe) > > @@ -113,8 +113,8 @@ TRACE_EVENT(intel_cpu_fifo_underrun, > > TP_fast_assign( > > __entry->pipe = pipe; > > - __entry->frame = dev_priv->drm.driver->get_vblank_counter(&dev_priv->drm, pipe); > > - __entry->scanline = intel_get_crtc_scanline(intel_get_crtc_for_pipe(dev_priv, pipe)); > > + __entry->frame = i915->drm.driver->get_vblank_counter(&i915->drm, pipe); > > + __entry->scanline = intel_get_crtc_scanline(intel_get_crtc_for_pipe(i915, pipe)); > > ), > > TP_printk("pipe %c, frame=%u, scanline=%u", > > @@ -123,8 +123,8 @@ TRACE_EVENT(intel_cpu_fifo_underrun, > > ); > > TRACE_EVENT(intel_pch_fifo_underrun, > > - TP_PROTO(struct drm_i915_private *dev_priv, enum pipe pch_transcoder), > > - TP_ARGS(dev_priv, pch_transcoder), > > + TP_PROTO(struct drm_i915_private *i915, enum pipe pch_transcoder), > > + TP_ARGS(i915, pch_transcoder), > > TP_STRUCT__entry( > > __field(enum pipe, pipe) > > @@ -135,8 +135,8 @@ TRACE_EVENT(intel_pch_fifo_underrun, > > TP_fast_assign( > > enum pipe pipe = pch_transcoder; > > __entry->pipe = pipe; > > - __entry->frame = dev_priv->drm.driver->get_vblank_counter(&dev_priv->drm, pipe); > > - __entry->scanline = intel_get_crtc_scanline(intel_get_crtc_for_pipe(dev_priv, pipe)); > > + __entry->frame = i915->drm.driver->get_vblank_counter(&i915->drm, pipe); > > + __entry->scanline = intel_get_crtc_scanline(intel_get_crtc_for_pipe(i915, pipe)); > > ), > > TP_printk("pch transcoder %c, frame=%u, scanline=%u", > > @@ -145,8 +145,8 @@ TRACE_EVENT(intel_pch_fifo_underrun, > > ); > > TRACE_EVENT(intel_memory_cxsr, > > - TP_PROTO(struct drm_i915_private *dev_priv, bool old, bool new), > > - TP_ARGS(dev_priv, old, new), > > + TP_PROTO(struct drm_i915_private *i915, bool old, bool new), > > + TP_ARGS(i915, old, new), > > TP_STRUCT__entry( > > __array(u32, frame, 3) > > @@ -157,11 +157,11 @@ TRACE_EVENT(intel_memory_cxsr, > > TP_fast_assign( > > enum pipe pipe; > > - for_each_pipe(dev_priv, pipe) { > > + for_each_pipe(i915, pipe) { > > __entry->frame[pipe] = > > - dev_priv->drm.driver->get_vblank_counter(&dev_priv->drm, pipe); > > + i915->drm.driver->get_vblank_counter(&i915->drm, pipe); > > __entry->scanline[pipe] = > > - intel_get_crtc_scanline(intel_get_crtc_for_pipe(dev_priv, pipe)); > > + intel_get_crtc_scanline(intel_get_crtc_for_pipe(i915, pipe)); > > } > > __entry->old = old; > > __entry->new = new; > > diff --git a/drivers/gpu/drm/i915/i915_vgpu.c b/drivers/gpu/drm/i915/i915_vgpu.c > > index 94d3992b599d..96874a4819a9 100644 > > --- a/drivers/gpu/drm/i915/i915_vgpu.c > > +++ b/drivers/gpu/drm/i915/i915_vgpu.c > > @@ -53,14 +53,14 @@ > > /** > > * i915_check_vgpu - detect virtual GPU > > - * @dev_priv: i915 device private > > + * @i915: i915 device private > > * > > * This function is called at the initialization stage, to detect whether > > * running on a vGPU. > > */ > > -void i915_check_vgpu(struct drm_i915_private *dev_priv) > > +void i915_check_vgpu(struct drm_i915_private *i915) > > { > > - struct intel_uncore *uncore = &dev_priv->uncore; > > + struct intel_uncore *uncore = &i915->uncore; > > u64 magic; > > u16 version_major; > > @@ -76,15 +76,15 @@ void i915_check_vgpu(struct drm_i915_private *dev_priv) > > return; > > } > > - dev_priv->vgpu.caps = __raw_uncore_read32(uncore, vgtif_reg(vgt_caps)); > > + i915->vgpu.caps = __raw_uncore_read32(uncore, vgtif_reg(vgt_caps)); > > - dev_priv->vgpu.active = true; > > + i915->vgpu.active = true; > > DRM_INFO("Virtual GPU for Intel GVT-g detected.\n"); > > } > > -bool intel_vgpu_has_full_ppgtt(struct drm_i915_private *dev_priv) > > +bool intel_vgpu_has_full_ppgtt(struct drm_i915_private *i915) > > { > > - return dev_priv->vgpu.caps & VGT_CAPS_FULL_PPGTT; > > + return i915->vgpu.caps & VGT_CAPS_FULL_PPGTT; > > } > > struct _balloon_info_ { > > @@ -112,22 +112,22 @@ static void vgt_deballoon_space(struct i915_ggtt *ggtt, > > /** > > * intel_vgt_deballoon - deballoon reserved graphics address trunks > > - * @dev_priv: i915 device private data > > + * @i915: i915 device private data > > * > > * This function is called to deallocate the ballooned-out graphic memory, when > > * driver is unloaded or when ballooning fails. > > */ > > -void intel_vgt_deballoon(struct drm_i915_private *dev_priv) > > +void intel_vgt_deballoon(struct drm_i915_private *i915) > > { > > int i; > > - if (!intel_vgpu_active(dev_priv)) > > + if (!intel_vgpu_active(i915)) > > return; > > DRM_DEBUG("VGT deballoon.\n"); > > for (i = 0; i < 4; i++) > > - vgt_deballoon_space(&dev_priv->ggtt, &bl_info.space[i]); > > + vgt_deballoon_space(&i915->ggtt, &bl_info.space[i]); > > } > > static int vgt_balloon_space(struct i915_ggtt *ggtt, > > @@ -153,7 +153,7 @@ static int vgt_balloon_space(struct i915_ggtt *ggtt, > > /** > > * intel_vgt_balloon - balloon out reserved graphics address trunks > > - * @dev_priv: i915 device private data > > + * @i915: i915 device private data > > * > > * This function is called at the initialization stage, to balloon out the > > * graphic address space allocated to other vGPUs, by marking these spaces as > > @@ -195,16 +195,16 @@ static int vgt_balloon_space(struct i915_ggtt *ggtt, > > * Returns: > > * zero on success, non-zero if configuration invalid or ballooning failed > > */ > > -int intel_vgt_balloon(struct drm_i915_private *dev_priv) > > +int intel_vgt_balloon(struct drm_i915_private *i915) > > { > > - struct i915_ggtt *ggtt = &dev_priv->ggtt; > > + struct i915_ggtt *ggtt = &i915->ggtt; > > unsigned long ggtt_end = ggtt->vm.total; > > unsigned long mappable_base, mappable_size, mappable_end; > > unsigned long unmappable_base, unmappable_size, unmappable_end; > > int ret; > > - if (!intel_vgpu_active(dev_priv)) > > + if (!intel_vgpu_active(i915)) > > return 0; > > mappable_base = I915_READ(vgtif_reg(avail_rs.mappable_gmadr.base)); > > diff --git a/drivers/gpu/drm/i915/i915_vgpu.h b/drivers/gpu/drm/i915/i915_vgpu.h > > index ebe1b7bced98..0fa0bfb88bed 100644 > > --- a/drivers/gpu/drm/i915/i915_vgpu.h > > +++ b/drivers/gpu/drm/i915/i915_vgpu.h > > @@ -26,23 +26,23 @@ > > #include "i915_pvinfo.h" > > -void i915_check_vgpu(struct drm_i915_private *dev_priv); > > +void i915_check_vgpu(struct drm_i915_private *i915); > > -bool intel_vgpu_has_full_ppgtt(struct drm_i915_private *dev_priv); > > +bool intel_vgpu_has_full_ppgtt(struct drm_i915_private *i915); > > static inline bool > > -intel_vgpu_has_hwsp_emulation(struct drm_i915_private *dev_priv) > > +intel_vgpu_has_hwsp_emulation(struct drm_i915_private *i915) > > { > > - return dev_priv->vgpu.caps & VGT_CAPS_HWSP_EMULATION; > > + return i915->vgpu.caps & VGT_CAPS_HWSP_EMULATION; > > } > > static inline bool > > -intel_vgpu_has_huge_gtt(struct drm_i915_private *dev_priv) > > +intel_vgpu_has_huge_gtt(struct drm_i915_private *i915) > > { > > - return dev_priv->vgpu.caps & VGT_CAPS_HUGE_GTT; > > + return i915->vgpu.caps & VGT_CAPS_HUGE_GTT; > > } > > -int intel_vgt_balloon(struct drm_i915_private *dev_priv); > > -void intel_vgt_deballoon(struct drm_i915_private *dev_priv); > > +int intel_vgt_balloon(struct drm_i915_private *i915); > > +void intel_vgt_deballoon(struct drm_i915_private *i915); > > #endif /* _I915_VGPU_H_ */ > > diff --git a/drivers/gpu/drm/i915/i915_vma.c b/drivers/gpu/drm/i915/i915_vma.c > > index cb341e4acf99..e3f4d7006a1b 100644 > > --- a/drivers/gpu/drm/i915/i915_vma.c > > +++ b/drivers/gpu/drm/i915/i915_vma.c > > @@ -555,7 +555,7 @@ static void assert_bind_count(const struct drm_i915_gem_object *obj) > > static int > > i915_vma_insert(struct i915_vma *vma, u64 size, u64 alignment, u64 flags) > > { > > - struct drm_i915_private *dev_priv = vma->vm->i915; > > + struct drm_i915_private *i915 = vma->vm->i915; > > unsigned int cache_level; > > u64 start, end; > > int ret; > > @@ -581,7 +581,7 @@ i915_vma_insert(struct i915_vma *vma, u64 size, u64 alignment, u64 flags) > > end = vma->vm->total; > > if (flags & PIN_MAPPABLE) > > - end = min_t(u64, end, dev_priv->ggtt.mappable_end); > > + end = min_t(u64, end, i915->ggtt.mappable_end); > > if (flags & PIN_ZONE_4G) > > end = min_t(u64, end, (1ULL << 32) - I915_GTT_PAGE_SIZE); > > GEM_BUG_ON(!IS_ALIGNED(end, I915_GTT_PAGE_SIZE)); > > diff --git a/drivers/gpu/drm/i915/icl_dsi.c b/drivers/gpu/drm/i915/icl_dsi.c > > index 74448e6bf749..d0192e093f6f 100644 > > --- a/drivers/gpu/drm/i915/icl_dsi.c > > +++ b/drivers/gpu/drm/i915/icl_dsi.c > > @@ -35,32 +35,32 @@ > > #include "intel_dsi.h" > > #include "intel_panel.h" > > -static inline int header_credits_available(struct drm_i915_private *dev_priv, > > +static inline int header_credits_available(struct drm_i915_private *i915, > > enum transcoder dsi_trans) > > { > > return (I915_READ(DSI_CMD_TXCTL(dsi_trans)) & FREE_HEADER_CREDIT_MASK) > > >> FREE_HEADER_CREDIT_SHIFT; > > } > > -static inline int payload_credits_available(struct drm_i915_private *dev_priv, > > +static inline int payload_credits_available(struct drm_i915_private *i915, > > enum transcoder dsi_trans) > > { > > return (I915_READ(DSI_CMD_TXCTL(dsi_trans)) & FREE_PLOAD_CREDIT_MASK) > > >> FREE_PLOAD_CREDIT_SHIFT; > > } > > -static void wait_for_header_credits(struct drm_i915_private *dev_priv, > > +static void wait_for_header_credits(struct drm_i915_private *i915, > > enum transcoder dsi_trans) > > { > > - if (wait_for_us(header_credits_available(dev_priv, dsi_trans) >= > > + if (wait_for_us(header_credits_available(i915, dsi_trans) >= > > MAX_HEADER_CREDIT, 100)) > > DRM_ERROR("DSI header credits not released\n"); > > } > > -static void wait_for_payload_credits(struct drm_i915_private *dev_priv, > > +static void wait_for_payload_credits(struct drm_i915_private *i915, > > enum transcoder dsi_trans) > > { > > - if (wait_for_us(payload_credits_available(dev_priv, dsi_trans) >= > > + if (wait_for_us(payload_credits_available(i915, dsi_trans) >= > > MAX_PLOAD_CREDIT, 100)) > > DRM_ERROR("DSI payload credits not released\n"); > > } > > @@ -75,7 +75,7 @@ static enum transcoder dsi_port_to_transcoder(enum port port) > > static void wait_for_cmds_dispatched_to_panel(struct intel_encoder *encoder) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base); > > struct mipi_dsi_device *dsi; > > enum port port; > > @@ -85,8 +85,8 @@ static void wait_for_cmds_dispatched_to_panel(struct intel_encoder *encoder) > > /* wait for header/payload credits to be released */ > > for_each_dsi_port(port, intel_dsi->ports) { > > dsi_trans = dsi_port_to_transcoder(port); > > - wait_for_header_credits(dev_priv, dsi_trans); > > - wait_for_payload_credits(dev_priv, dsi_trans); > > + wait_for_header_credits(i915, dsi_trans); > > + wait_for_payload_credits(i915, dsi_trans); > > } > > /* send nop DCS command */ > > @@ -102,7 +102,7 @@ static void wait_for_cmds_dispatched_to_panel(struct intel_encoder *encoder) > > /* wait for header credits to be released */ > > for_each_dsi_port(port, intel_dsi->ports) { > > dsi_trans = dsi_port_to_transcoder(port); > > - wait_for_header_credits(dev_priv, dsi_trans); > > + wait_for_header_credits(i915, dsi_trans); > > } > > /* wait for LP TX in progress bit to be cleared */ > > @@ -118,7 +118,7 @@ static bool add_payld_to_queue(struct intel_dsi_host *host, const u8 *data, > > u32 len) > > { > > struct intel_dsi *intel_dsi = host->intel_dsi; > > - struct drm_i915_private *dev_priv = to_i915(intel_dsi->base.base.dev); > > + struct drm_i915_private *i915 = to_i915(intel_dsi->base.base.dev); > > enum transcoder dsi_trans = dsi_port_to_transcoder(host->port); > > int free_credits; > > int i, j; > > @@ -126,7 +126,7 @@ static bool add_payld_to_queue(struct intel_dsi_host *host, const u8 *data, > > for (i = 0; i < len; i += 4) { > > u32 tmp = 0; > > - free_credits = payload_credits_available(dev_priv, dsi_trans); > > + free_credits = payload_credits_available(i915, dsi_trans); > > if (free_credits < 1) { > > DRM_ERROR("Payload credit not available\n"); > > return false; > > @@ -145,13 +145,13 @@ static int dsi_send_pkt_hdr(struct intel_dsi_host *host, > > struct mipi_dsi_packet pkt, bool enable_lpdt) > > { > > struct intel_dsi *intel_dsi = host->intel_dsi; > > - struct drm_i915_private *dev_priv = to_i915(intel_dsi->base.base.dev); > > + struct drm_i915_private *i915 = to_i915(intel_dsi->base.base.dev); > > enum transcoder dsi_trans = dsi_port_to_transcoder(host->port); > > u32 tmp; > > int free_credits; > > /* check if header credit available */ > > - free_credits = header_credits_available(dev_priv, dsi_trans); > > + free_credits = header_credits_available(i915, dsi_trans); > > if (free_credits < 1) { > > DRM_ERROR("send pkt header failed, not enough hdr credits\n"); > > return -1; > > @@ -200,7 +200,7 @@ static int dsi_send_pkt_payld(struct intel_dsi_host *host, > > static void dsi_program_swing_and_deemphasis(struct intel_encoder *encoder) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base); > > enum port port; > > u32 tmp; > > @@ -266,7 +266,7 @@ static void dsi_program_swing_and_deemphasis(struct intel_encoder *encoder) > > static void configure_dual_link_mode(struct intel_encoder *encoder, > > const struct intel_crtc_state *pipe_config) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base); > > u32 dss_ctl1; > > @@ -304,7 +304,7 @@ static void configure_dual_link_mode(struct intel_encoder *encoder, > > static void gen11_dsi_program_esc_clk_div(struct intel_encoder *encoder) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base); > > enum port port; > > u32 bpp = mipi_dsi_pixel_format_to_bpp(intel_dsi->pixel_format); > > @@ -329,7 +329,7 @@ static void gen11_dsi_program_esc_clk_div(struct intel_encoder *encoder) > > } > > } > > -static void get_dsi_io_power_domains(struct drm_i915_private *dev_priv, > > +static void get_dsi_io_power_domains(struct drm_i915_private *i915, > > struct intel_dsi *intel_dsi) > > { > > enum port port; > > @@ -337,7 +337,7 @@ static void get_dsi_io_power_domains(struct drm_i915_private *dev_priv, > > for_each_dsi_port(port, intel_dsi->ports) { > > WARN_ON(intel_dsi->io_wakeref[port]); > > intel_dsi->io_wakeref[port] = > > - intel_display_power_get(dev_priv, > > + intel_display_power_get(i915, > > port == PORT_A ? > > POWER_DOMAIN_PORT_DDI_A_IO : > > POWER_DOMAIN_PORT_DDI_B_IO); > > @@ -346,7 +346,7 @@ static void get_dsi_io_power_domains(struct drm_i915_private *dev_priv, > > static void gen11_dsi_enable_io_power(struct intel_encoder *encoder) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base); > > enum port port; > > u32 tmp; > > @@ -357,23 +357,23 @@ static void gen11_dsi_enable_io_power(struct intel_encoder *encoder) > > I915_WRITE(ICL_DSI_IO_MODECTL(port), tmp); > > } > > - get_dsi_io_power_domains(dev_priv, intel_dsi); > > + get_dsi_io_power_domains(i915, intel_dsi); > > } > > static void gen11_dsi_power_up_lanes(struct intel_encoder *encoder) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base); > > enum port port; > > for_each_dsi_port(port, intel_dsi->ports) > > - intel_combo_phy_power_up_lanes(dev_priv, port, true, > > + intel_combo_phy_power_up_lanes(i915, port, true, > > intel_dsi->lane_count, false); > > } > > static void gen11_dsi_config_phy_lanes_sequence(struct intel_encoder *encoder) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base); > > enum port port; > > u32 tmp; > > @@ -409,7 +409,7 @@ static void gen11_dsi_config_phy_lanes_sequence(struct intel_encoder *encoder) > > static void gen11_dsi_voltage_swing_program_seq(struct intel_encoder *encoder) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base); > > u32 tmp; > > enum port port; > > @@ -461,7 +461,7 @@ static void gen11_dsi_voltage_swing_program_seq(struct intel_encoder *encoder) > > static void gen11_dsi_enable_ddi_buffer(struct intel_encoder *encoder) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base); > > u32 tmp; > > enum port port; > > @@ -480,7 +480,7 @@ static void gen11_dsi_enable_ddi_buffer(struct intel_encoder *encoder) > > static void gen11_dsi_setup_dphy_timings(struct intel_encoder *encoder) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base); > > u32 tmp; > > enum port port; > > @@ -535,48 +535,48 @@ static void gen11_dsi_setup_dphy_timings(struct intel_encoder *encoder) > > static void gen11_dsi_gate_clocks(struct intel_encoder *encoder) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base); > > u32 tmp; > > enum port port; > > - mutex_lock(&dev_priv->dpll_lock); > > + mutex_lock(&i915->dpll_lock); > > tmp = I915_READ(DPCLKA_CFGCR0_ICL); > > for_each_dsi_port(port, intel_dsi->ports) { > > tmp |= DPCLKA_CFGCR0_DDI_CLK_OFF(port); > > } > > I915_WRITE(DPCLKA_CFGCR0_ICL, tmp); > > - mutex_unlock(&dev_priv->dpll_lock); > > + mutex_unlock(&i915->dpll_lock); > > } > > static void gen11_dsi_ungate_clocks(struct intel_encoder *encoder) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base); > > u32 tmp; > > enum port port; > > - mutex_lock(&dev_priv->dpll_lock); > > + mutex_lock(&i915->dpll_lock); > > tmp = I915_READ(DPCLKA_CFGCR0_ICL); > > for_each_dsi_port(port, intel_dsi->ports) { > > tmp &= ~DPCLKA_CFGCR0_DDI_CLK_OFF(port); > > } > > I915_WRITE(DPCLKA_CFGCR0_ICL, tmp); > > - mutex_unlock(&dev_priv->dpll_lock); > > + mutex_unlock(&i915->dpll_lock); > > } > > static void gen11_dsi_map_pll(struct intel_encoder *encoder, > > const struct intel_crtc_state *crtc_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base); > > struct intel_shared_dpll *pll = crtc_state->shared_dpll; > > enum port port; > > u32 val; > > - mutex_lock(&dev_priv->dpll_lock); > > + mutex_lock(&i915->dpll_lock); > > val = I915_READ(DPCLKA_CFGCR0_ICL); > > for_each_dsi_port(port, intel_dsi->ports) { > > @@ -592,14 +592,14 @@ static void gen11_dsi_map_pll(struct intel_encoder *encoder, > > POSTING_READ(DPCLKA_CFGCR0_ICL); > > - mutex_unlock(&dev_priv->dpll_lock); > > + mutex_unlock(&i915->dpll_lock); > > } > > static void > > gen11_dsi_configure_transcoder(struct intel_encoder *encoder, > > const struct intel_crtc_state *pipe_config) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base); > > struct intel_crtc *intel_crtc = to_intel_crtc(pipe_config->base.crtc); > > enum pipe pipe = intel_crtc->pipe; > > @@ -736,7 +736,7 @@ static void > > gen11_dsi_set_transcoder_timings(struct intel_encoder *encoder, > > const struct intel_crtc_state *pipe_config) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base); > > const struct drm_display_mode *adjusted_mode = > > &pipe_config->base.adjusted_mode; > > @@ -849,7 +849,7 @@ gen11_dsi_set_transcoder_timings(struct intel_encoder *encoder, > > static void gen11_dsi_enable_transcoder(struct intel_encoder *encoder) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base); > > enum port port; > > enum transcoder dsi_trans; > > @@ -862,7 +862,7 @@ static void gen11_dsi_enable_transcoder(struct intel_encoder *encoder) > > I915_WRITE(PIPECONF(dsi_trans), tmp); > > /* wait for transcoder to be enabled */ > > - if (intel_wait_for_register(&dev_priv->uncore, > > + if (intel_wait_for_register(&i915->uncore, > > PIPECONF(dsi_trans), > > I965_PIPECONF_ACTIVE, > > I965_PIPECONF_ACTIVE, 10)) > > @@ -872,7 +872,7 @@ static void gen11_dsi_enable_transcoder(struct intel_encoder *encoder) > > static void gen11_dsi_setup_timeouts(struct intel_encoder *encoder) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base); > > enum port port; > > enum transcoder dsi_trans; > > @@ -950,7 +950,7 @@ gen11_dsi_enable_port_and_phy(struct intel_encoder *encoder, > > static void gen11_dsi_powerup_panel(struct intel_encoder *encoder) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base); > > struct mipi_dsi_device *dsi; > > enum port port; > > @@ -1026,7 +1026,7 @@ static void gen11_dsi_pre_enable(struct intel_encoder *encoder, > > static void gen11_dsi_disable_transcoder(struct intel_encoder *encoder) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base); > > enum port port; > > enum transcoder dsi_trans; > > @@ -1041,7 +1041,7 @@ static void gen11_dsi_disable_transcoder(struct intel_encoder *encoder) > > I915_WRITE(PIPECONF(dsi_trans), tmp); > > /* wait for transcoder to be disabled */ > > - if (intel_wait_for_register(&dev_priv->uncore, > > + if (intel_wait_for_register(&i915->uncore, > > PIPECONF(dsi_trans), > > I965_PIPECONF_ACTIVE, 0, 50)) > > DRM_ERROR("DSI trancoder not disabled\n"); > > @@ -1062,7 +1062,7 @@ static void gen11_dsi_powerdown_panel(struct intel_encoder *encoder) > > static void gen11_dsi_deconfigure_trancoder(struct intel_encoder *encoder) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base); > > enum port port; > > enum transcoder dsi_trans; > > @@ -1103,7 +1103,7 @@ static void gen11_dsi_deconfigure_trancoder(struct intel_encoder *encoder) > > static void gen11_dsi_disable_port(struct intel_encoder *encoder) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base); > > u32 tmp; > > enum port port; > > @@ -1125,7 +1125,7 @@ static void gen11_dsi_disable_port(struct intel_encoder *encoder) > > static void gen11_dsi_disable_io_power(struct intel_encoder *encoder) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base); > > enum port port; > > u32 tmp; > > @@ -1134,7 +1134,7 @@ static void gen11_dsi_disable_io_power(struct intel_encoder *encoder) > > intel_wakeref_t wakeref; > > wakeref = fetch_and_zero(&intel_dsi->io_wakeref[port]); > > - intel_display_power_put(dev_priv, > > + intel_display_power_put(i915, > > port == PORT_A ? > > POWER_DOMAIN_PORT_DDI_A_IO : > > POWER_DOMAIN_PORT_DDI_B_IO, > > @@ -1205,13 +1205,13 @@ static void gen11_dsi_get_timings(struct intel_encoder *encoder, > > static void gen11_dsi_get_config(struct intel_encoder *encoder, > > struct intel_crtc_state *pipe_config) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_crtc *crtc = to_intel_crtc(pipe_config->base.crtc); > > struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base); > > /* FIXME: adapt icl_ddi_clock_get() for DSI and use that? */ > > pipe_config->port_clock = > > - cnl_calc_wrpll_link(dev_priv, &pipe_config->dpll_hw_state); > > + cnl_calc_wrpll_link(i915, &pipe_config->dpll_hw_state); > > pipe_config->base.adjusted_mode.crtc_clock = intel_dsi->pclk; > > if (intel_dsi->dual_link) > > @@ -1263,7 +1263,7 @@ static void gen11_dsi_get_power_domains(struct intel_encoder *encoder, > > static bool gen11_dsi_get_hw_state(struct intel_encoder *encoder, > > enum pipe *pipe) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base); > > enum transcoder dsi_trans; > > intel_wakeref_t wakeref; > > @@ -1271,7 +1271,7 @@ static bool gen11_dsi_get_hw_state(struct intel_encoder *encoder, > > bool ret = false; > > u32 tmp; > > - wakeref = intel_display_power_get_if_enabled(dev_priv, > > + wakeref = intel_display_power_get_if_enabled(i915, > > encoder->power_domain); > > if (!wakeref) > > return false; > > @@ -1298,7 +1298,7 @@ static bool gen11_dsi_get_hw_state(struct intel_encoder *encoder, > > ret = tmp & PIPECONF_ENABLE; > > } > > out: > > - intel_display_power_put(dev_priv, encoder->power_domain, wakeref); > > + intel_display_power_put(i915, encoder->power_domain, wakeref); > > return ret; > > } > > @@ -1391,8 +1391,8 @@ static const struct mipi_dsi_host_ops gen11_dsi_host_ops = { > > static void icl_dphy_param_init(struct intel_dsi *intel_dsi) > > { > > struct drm_device *dev = intel_dsi->base.base.dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > - struct mipi_config *mipi_config = dev_priv->vbt.dsi.config; > > + struct drm_i915_private *i915 = to_i915(dev); > > + struct mipi_config *mipi_config = i915->vbt.dsi.config; > > u32 tlpx_ns; > > u32 prepare_cnt, exit_zero_cnt, clk_zero_cnt, trail_cnt; > > u32 ths_prepare_ns, tclk_trail_ns; > > @@ -1487,9 +1487,9 @@ static void icl_dphy_param_init(struct intel_dsi *intel_dsi) > > intel_dsi_log_params(intel_dsi); > > } > > -void icl_dsi_init(struct drm_i915_private *dev_priv) > > +void icl_dsi_init(struct drm_i915_private *i915) > > { > > - struct drm_device *dev = &dev_priv->drm; > > + struct drm_device *dev = &i915->drm; > > struct intel_dsi *intel_dsi; > > struct intel_encoder *encoder; > > struct intel_connector *intel_connector; > > @@ -1497,7 +1497,7 @@ void icl_dsi_init(struct drm_i915_private *dev_priv) > > struct drm_display_mode *fixed_mode; > > enum port port; > > - if (!intel_bios_is_dsi_present(dev_priv, &port)) > > + if (!intel_bios_is_dsi_present(i915, &port)) > > return; > > intel_dsi = kzalloc(sizeof(*intel_dsi), GFP_KERNEL); > > @@ -1556,13 +1556,13 @@ void icl_dsi_init(struct drm_i915_private *dev_priv) > > intel_panel_init(&intel_connector->panel, fixed_mode, NULL); > > intel_panel_setup_backlight(connector, INVALID_PIPE); > > - if (dev_priv->vbt.dsi.config->dual_link) > > + if (i915->vbt.dsi.config->dual_link) > > intel_dsi->ports = BIT(PORT_A) | BIT(PORT_B); > > else > > intel_dsi->ports = BIT(port); > > - intel_dsi->dcs_backlight_ports = dev_priv->vbt.dsi.bl_ports; > > - intel_dsi->dcs_cabc_ports = dev_priv->vbt.dsi.cabc_ports; > > + intel_dsi->dcs_backlight_ports = i915->vbt.dsi.bl_ports; > > + intel_dsi->dcs_cabc_ports = i915->vbt.dsi.cabc_ports; > > for_each_dsi_port(port, intel_dsi->ports) { > > struct intel_dsi_host *host; > > diff --git a/drivers/gpu/drm/i915/intel_atomic.c b/drivers/gpu/drm/i915/intel_atomic.c > > index 6b985e895a97..45782e6ff0f4 100644 > > --- a/drivers/gpu/drm/i915/intel_atomic.c > > +++ b/drivers/gpu/drm/i915/intel_atomic.c > > @@ -54,13 +54,13 @@ int intel_digital_connector_atomic_get_property(struct drm_connector *connector, > > u64 *val) > > { > > struct drm_device *dev = connector->dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > struct intel_digital_connector_state *intel_conn_state = > > to_intel_digital_connector_state(state); > > - if (property == dev_priv->force_audio_property) > > + if (property == i915->force_audio_property) > > *val = intel_conn_state->force_audio; > > - else if (property == dev_priv->broadcast_rgb_property) > > + else if (property == i915->broadcast_rgb_property) > > *val = intel_conn_state->broadcast_rgb; > > else { > > DRM_DEBUG_ATOMIC("Unknown property [PROP:%d:%s]\n", > > @@ -86,16 +86,16 @@ int intel_digital_connector_atomic_set_property(struct drm_connector *connector, > > u64 val) > > { > > struct drm_device *dev = connector->dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > struct intel_digital_connector_state *intel_conn_state = > > to_intel_digital_connector_state(state); > > - if (property == dev_priv->force_audio_property) { > > + if (property == i915->force_audio_property) { > > intel_conn_state->force_audio = val; > > return 0; > > } > > - if (property == dev_priv->broadcast_rgb_property) { > > + if (property == i915->broadcast_rgb_property) { > > intel_conn_state->broadcast_rgb = val; > > return 0; > > } > > @@ -227,7 +227,7 @@ static void intel_atomic_setup_scaler(struct intel_crtc_scaler_state *scaler_sta > > struct intel_plane_state *plane_state, > > int *scaler_id) > > { > > - struct drm_i915_private *dev_priv = to_i915(intel_crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(intel_crtc->base.dev); > > int j; > > u32 mode; > > @@ -251,10 +251,10 @@ static void intel_atomic_setup_scaler(struct intel_crtc_scaler_state *scaler_sta > > plane_state->base.fb->format->is_yuv && > > plane_state->base.fb->format->num_planes > 1) { > > struct intel_plane *plane = to_intel_plane(plane_state->base.plane); > > - if (IS_GEN(dev_priv, 9) && > > - !IS_GEMINILAKE(dev_priv)) { > > + if (IS_GEN(i915, 9) && > > + !IS_GEMINILAKE(i915)) { > > mode = SKL_PS_SCALER_MODE_NV12; > > - } else if (icl_is_hdr_plane(dev_priv, plane->id)) { > > + } else if (icl_is_hdr_plane(i915, plane->id)) { > > /* > > * On gen11+'s HDR planes we only use the scaler for > > * scaling. They have a dedicated chroma upsampler, so > > @@ -267,7 +267,7 @@ static void intel_atomic_setup_scaler(struct intel_crtc_scaler_state *scaler_sta > > if (plane_state->linked_plane) > > mode |= PS_PLANE_Y_SEL(plane_state->linked_plane->id); > > } > > - } else if (INTEL_GEN(dev_priv) > 9 || IS_GEMINILAKE(dev_priv)) { > > + } else if (INTEL_GEN(i915) > 9 || IS_GEMINILAKE(i915)) { > > mode = PS_SCALER_MODE_NORMAL; > > } else if (num_scalers_need == 1 && intel_crtc->num_scalers > 1) { > > /* > > @@ -290,7 +290,7 @@ static void intel_atomic_setup_scaler(struct intel_crtc_scaler_state *scaler_sta > > /** > > * intel_atomic_setup_scalers() - setup scalers for crtc per staged requests > > - * @dev_priv: i915 device > > + * @i915: i915 device > > * @intel_crtc: intel crtc > > * @crtc_state: incoming crtc_state to validate and setup scalers > > * > > @@ -305,7 +305,7 @@ static void intel_atomic_setup_scaler(struct intel_crtc_scaler_state *scaler_sta > > * 0 - scalers were setup succesfully > > * error code - otherwise > > */ > > -int intel_atomic_setup_scalers(struct drm_i915_private *dev_priv, > > +int intel_atomic_setup_scalers(struct drm_i915_private *i915, > > struct intel_crtc *intel_crtc, > > struct intel_crtc_state *crtc_state) > > { > > @@ -369,7 +369,7 @@ int intel_atomic_setup_scalers(struct drm_i915_private *dev_priv, > > */ > > if (!plane) { > > struct drm_plane_state *state; > > - plane = drm_plane_from_index(&dev_priv->drm, i); > > + plane = drm_plane_from_index(&i915->drm, i); > > state = drm_atomic_get_plane_state(drm_state, plane); > > if (IS_ERR(state)) { > > DRM_DEBUG_KMS("Failed to add [PLANE:%d] to drm_state\n", > > diff --git a/drivers/gpu/drm/i915/intel_atomic.h b/drivers/gpu/drm/i915/intel_atomic.h > > index 1c8507da1a69..5482fd04d639 100644 > > --- a/drivers/gpu/drm/i915/intel_atomic.h > > +++ b/drivers/gpu/drm/i915/intel_atomic.h > > @@ -42,7 +42,7 @@ struct intel_crtc_state * > > intel_atomic_get_crtc_state(struct drm_atomic_state *state, > > struct intel_crtc *crtc); > > -int intel_atomic_setup_scalers(struct drm_i915_private *dev_priv, > > +int intel_atomic_setup_scalers(struct drm_i915_private *i915, > > struct intel_crtc *intel_crtc, > > struct intel_crtc_state *crtc_state); > > diff --git a/drivers/gpu/drm/i915/intel_audio.c b/drivers/gpu/drm/i915/intel_audio.c > > index 840daff12246..f463d8452fe9 100644 > > --- a/drivers/gpu/drm/i915/intel_audio.c > > +++ b/drivers/gpu/drm/i915/intel_audio.c > > @@ -230,7 +230,7 @@ static bool intel_eld_uptodate(struct drm_connector *connector, > > i915_reg_t reg_elda, u32 bits_elda, > > i915_reg_t reg_edid) > > { > > - struct drm_i915_private *dev_priv = to_i915(connector->dev); > > + struct drm_i915_private *i915 = to_i915(connector->dev); > > const u8 *eld = connector->eld; > > u32 tmp; > > int i; > > @@ -256,7 +256,7 @@ static void g4x_audio_codec_disable(struct intel_encoder *encoder, > > const struct intel_crtc_state *old_crtc_state, > > const struct drm_connector_state *old_conn_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > u32 eldv, tmp; > > DRM_DEBUG_KMS("Disable audio codec\n"); > > @@ -277,7 +277,7 @@ static void g4x_audio_codec_enable(struct intel_encoder *encoder, > > const struct intel_crtc_state *crtc_state, > > const struct drm_connector_state *conn_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct drm_connector *connector = conn_state->connector; > > const u8 *eld = connector->eld; > > u32 eldv; > > @@ -317,8 +317,8 @@ static void > > hsw_dp_audio_config_update(struct intel_encoder *encoder, > > const struct intel_crtc_state *crtc_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > - struct i915_audio_component *acomp = dev_priv->audio_component; > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > + struct i915_audio_component *acomp = i915->audio_component; > > enum transcoder cpu_transcoder = crtc_state->cpu_transcoder; > > enum port port = encoder->port; > > const struct dp_aud_n_m *nm; > > @@ -364,8 +364,8 @@ static void > > hsw_hdmi_audio_config_update(struct intel_encoder *encoder, > > const struct intel_crtc_state *crtc_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > - struct i915_audio_component *acomp = dev_priv->audio_component; > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > + struct i915_audio_component *acomp = i915->audio_component; > > enum transcoder cpu_transcoder = crtc_state->cpu_transcoder; > > enum port port = encoder->port; > > int n, rate; > > @@ -416,14 +416,14 @@ static void hsw_audio_codec_disable(struct intel_encoder *encoder, > > const struct intel_crtc_state *old_crtc_state, > > const struct drm_connector_state *old_conn_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > enum transcoder cpu_transcoder = old_crtc_state->cpu_transcoder; > > u32 tmp; > > DRM_DEBUG_KMS("Disable audio codec on transcoder %s\n", > > transcoder_name(cpu_transcoder)); > > - mutex_lock(&dev_priv->av_mutex); > > + mutex_lock(&i915->av_mutex); > > /* Disable timestamps */ > > tmp = I915_READ(HSW_AUD_CFG(cpu_transcoder)); > > @@ -441,14 +441,14 @@ static void hsw_audio_codec_disable(struct intel_encoder *encoder, > > tmp &= ~AUDIO_OUTPUT_ENABLE(cpu_transcoder); > > I915_WRITE(HSW_AUD_PIN_ELD_CP_VLD, tmp); > > - mutex_unlock(&dev_priv->av_mutex); > > + mutex_unlock(&i915->av_mutex); > > } > > static void hsw_audio_codec_enable(struct intel_encoder *encoder, > > const struct intel_crtc_state *crtc_state, > > const struct drm_connector_state *conn_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct drm_connector *connector = conn_state->connector; > > enum transcoder cpu_transcoder = crtc_state->cpu_transcoder; > > const u8 *eld = connector->eld; > > @@ -458,7 +458,7 @@ static void hsw_audio_codec_enable(struct intel_encoder *encoder, > > DRM_DEBUG_KMS("Enable audio codec on transcoder %s, %u bytes ELD\n", > > transcoder_name(cpu_transcoder), drm_eld_size(eld)); > > - mutex_lock(&dev_priv->av_mutex); > > + mutex_lock(&i915->av_mutex); > > /* Enable audio presence detect, invalidate ELD */ > > tmp = I915_READ(HSW_AUD_PIN_ELD_CP_VLD); > > @@ -491,14 +491,14 @@ static void hsw_audio_codec_enable(struct intel_encoder *encoder, > > /* Enable timestamps */ > > hsw_audio_config_update(encoder, crtc_state); > > - mutex_unlock(&dev_priv->av_mutex); > > + mutex_unlock(&i915->av_mutex); > > } > > static void ilk_audio_codec_disable(struct intel_encoder *encoder, > > const struct intel_crtc_state *old_crtc_state, > > const struct drm_connector_state *old_conn_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_crtc *crtc = to_intel_crtc(old_crtc_state->base.crtc); > > enum pipe pipe = crtc->pipe; > > enum port port = encoder->port; > > @@ -511,10 +511,10 @@ static void ilk_audio_codec_disable(struct intel_encoder *encoder, > > if (WARN_ON(port == PORT_A)) > > return; > > - if (HAS_PCH_IBX(dev_priv)) { > > + if (HAS_PCH_IBX(i915)) { > > aud_config = IBX_AUD_CFG(pipe); > > aud_cntrl_st2 = IBX_AUD_CNTL_ST2; > > - } else if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) { > > + } else if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) { > > aud_config = VLV_AUD_CFG(pipe); > > aud_cntrl_st2 = VLV_AUD_CNTL_ST2; > > } else { > > @@ -544,7 +544,7 @@ static void ilk_audio_codec_enable(struct intel_encoder *encoder, > > const struct intel_crtc_state *crtc_state, > > const struct drm_connector_state *conn_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc); > > struct drm_connector *connector = conn_state->connector; > > enum pipe pipe = crtc->pipe; > > @@ -567,13 +567,13 @@ static void ilk_audio_codec_enable(struct intel_encoder *encoder, > > * infrastructure is not there yet. > > */ > > - if (HAS_PCH_IBX(dev_priv)) { > > + if (HAS_PCH_IBX(i915)) { > > hdmiw_hdmiedid = IBX_HDMIW_HDMIEDID(pipe); > > aud_config = IBX_AUD_CFG(pipe); > > aud_cntl_st = IBX_AUD_CNTL_ST(pipe); > > aud_cntrl_st2 = IBX_AUD_CNTL_ST2; > > - } else if (IS_VALLEYVIEW(dev_priv) || > > - IS_CHERRYVIEW(dev_priv)) { > > + } else if (IS_VALLEYVIEW(i915) || > > + IS_CHERRYVIEW(i915)) { > > hdmiw_hdmiedid = VLV_HDMIW_HDMIEDID(pipe); > > aud_config = VLV_AUD_CFG(pipe); > > aud_cntl_st = VLV_AUD_CNTL_ST(pipe); > > @@ -632,8 +632,8 @@ void intel_audio_codec_enable(struct intel_encoder *encoder, > > const struct intel_crtc_state *crtc_state, > > const struct drm_connector_state *conn_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > - struct i915_audio_component *acomp = dev_priv->audio_component; > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > + struct i915_audio_component *acomp = i915->audio_component; > > struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc); > > struct drm_connector *connector = conn_state->connector; > > const struct drm_display_mode *adjusted_mode = > > @@ -654,17 +654,17 @@ void intel_audio_codec_enable(struct intel_encoder *encoder, > > connector->eld[6] = drm_av_sync_delay(connector, adjusted_mode) / 2; > > - if (dev_priv->display.audio_codec_enable) > > - dev_priv->display.audio_codec_enable(encoder, > > + if (i915->display.audio_codec_enable) > > + i915->display.audio_codec_enable(encoder, > > crtc_state, > > conn_state); > > - mutex_lock(&dev_priv->av_mutex); > > + mutex_lock(&i915->av_mutex); > > encoder->audio_connector = connector; > > /* referred in audio callbacks */ > > - dev_priv->av_enc_map[pipe] = encoder; > > - mutex_unlock(&dev_priv->av_mutex); > > + i915->av_enc_map[pipe] = encoder; > > + mutex_unlock(&i915->av_mutex); > > if (acomp && acomp->base.audio_ops && > > acomp->base.audio_ops->pin_eld_notify) { > > @@ -675,7 +675,7 @@ void intel_audio_codec_enable(struct intel_encoder *encoder, > > (int) port, (int) pipe); > > } > > - intel_lpe_audio_notify(dev_priv, pipe, port, connector->eld, > > + intel_lpe_audio_notify(i915, pipe, port, connector->eld, > > crtc_state->port_clock, > > intel_crtc_has_dp_encoder(crtc_state)); > > } > > @@ -693,21 +693,21 @@ void intel_audio_codec_disable(struct intel_encoder *encoder, > > const struct intel_crtc_state *old_crtc_state, > > const struct drm_connector_state *old_conn_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > - struct i915_audio_component *acomp = dev_priv->audio_component; > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > + struct i915_audio_component *acomp = i915->audio_component; > > struct intel_crtc *crtc = to_intel_crtc(old_crtc_state->base.crtc); > > enum port port = encoder->port; > > enum pipe pipe = crtc->pipe; > > - if (dev_priv->display.audio_codec_disable) > > - dev_priv->display.audio_codec_disable(encoder, > > + if (i915->display.audio_codec_disable) > > + i915->display.audio_codec_disable(encoder, > > old_crtc_state, > > old_conn_state); > > - mutex_lock(&dev_priv->av_mutex); > > + mutex_lock(&i915->av_mutex); > > encoder->audio_connector = NULL; > > - dev_priv->av_enc_map[pipe] = NULL; > > - mutex_unlock(&dev_priv->av_mutex); > > + i915->av_enc_map[pipe] = NULL; > > + mutex_unlock(&i915->av_mutex); > > if (acomp && acomp->base.audio_ops && > > acomp->base.audio_ops->pin_eld_notify) { > > @@ -718,31 +718,31 @@ void intel_audio_codec_disable(struct intel_encoder *encoder, > > (int) port, (int) pipe); > > } > > - intel_lpe_audio_notify(dev_priv, pipe, port, NULL, 0, false); > > + intel_lpe_audio_notify(i915, pipe, port, NULL, 0, false); > > } > > /** > > * intel_init_audio_hooks - Set up chip specific audio hooks > > - * @dev_priv: device private > > + * @i915: device private > > */ > > -void intel_init_audio_hooks(struct drm_i915_private *dev_priv) > > +void intel_init_audio_hooks(struct drm_i915_private *i915) > > { > > - if (IS_G4X(dev_priv)) { > > - dev_priv->display.audio_codec_enable = g4x_audio_codec_enable; > > - dev_priv->display.audio_codec_disable = g4x_audio_codec_disable; > > - } else if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) { > > - dev_priv->display.audio_codec_enable = ilk_audio_codec_enable; > > - dev_priv->display.audio_codec_disable = ilk_audio_codec_disable; > > - } else if (IS_HASWELL(dev_priv) || INTEL_GEN(dev_priv) >= 8) { > > - dev_priv->display.audio_codec_enable = hsw_audio_codec_enable; > > - dev_priv->display.audio_codec_disable = hsw_audio_codec_disable; > > - } else if (HAS_PCH_SPLIT(dev_priv)) { > > - dev_priv->display.audio_codec_enable = ilk_audio_codec_enable; > > - dev_priv->display.audio_codec_disable = ilk_audio_codec_disable; > > + if (IS_G4X(i915)) { > > + i915->display.audio_codec_enable = g4x_audio_codec_enable; > > + i915->display.audio_codec_disable = g4x_audio_codec_disable; > > + } else if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) { > > + i915->display.audio_codec_enable = ilk_audio_codec_enable; > > + i915->display.audio_codec_disable = ilk_audio_codec_disable; > > + } else if (IS_HASWELL(i915) || INTEL_GEN(i915) >= 8) { > > + i915->display.audio_codec_enable = hsw_audio_codec_enable; > > + i915->display.audio_codec_disable = hsw_audio_codec_disable; > > + } else if (HAS_PCH_SPLIT(i915)) { > > + i915->display.audio_codec_enable = ilk_audio_codec_enable; > > + i915->display.audio_codec_disable = ilk_audio_codec_disable; > > } > > } > > -static void glk_force_audio_cdclk(struct drm_i915_private *dev_priv, > > +static void glk_force_audio_cdclk(struct drm_i915_private *i915, > > bool enable) > > { > > struct drm_modeset_acquire_ctx ctx; > > @@ -750,7 +750,7 @@ static void glk_force_audio_cdclk(struct drm_i915_private *dev_priv, > > int ret; > > drm_modeset_acquire_init(&ctx, 0); > > - state = drm_atomic_state_alloc(&dev_priv->drm); > > + state = drm_atomic_state_alloc(&i915->drm); > > if (WARN_ON(!state)) > > return; > > @@ -762,11 +762,11 @@ static void glk_force_audio_cdclk(struct drm_i915_private *dev_priv, > > enable ? 2 * 96000 : 0; > > /* > > - * Protects dev_priv->cdclk.force_min_cdclk > > + * Protects i915->cdclk.force_min_cdclk > > * Need to lock this here in case we have no active pipes > > * and thus wouldn't lock it during the commit otherwise. > > */ > > - ret = drm_modeset_lock(&dev_priv->drm.mode_config.connection_mutex, > > + ret = drm_modeset_lock(&i915->drm.mode_config.connection_mutex, > > &ctx); > > if (!ret) > > ret = drm_atomic_commit(state); > > @@ -787,18 +787,18 @@ static void glk_force_audio_cdclk(struct drm_i915_private *dev_priv, > > static unsigned long i915_audio_component_get_power(struct device *kdev) > > { > > - struct drm_i915_private *dev_priv = kdev_to_i915(kdev); > > + struct drm_i915_private *i915 = kdev_to_i915(kdev); > > intel_wakeref_t ret; > > /* Catch potential impedance mismatches before they occur! */ > > BUILD_BUG_ON(sizeof(intel_wakeref_t) > sizeof(unsigned long)); > > - ret = intel_display_power_get(dev_priv, POWER_DOMAIN_AUDIO); > > + ret = intel_display_power_get(i915, POWER_DOMAIN_AUDIO); > > /* Force CDCLK to 2*BCLK as long as we need audio to be powered. */ > > - if (dev_priv->audio_power_refcount++ == 0) > > - if (IS_CANNONLAKE(dev_priv) || IS_GEMINILAKE(dev_priv)) > > - glk_force_audio_cdclk(dev_priv, true); > > + if (i915->audio_power_refcount++ == 0) > > + if (IS_CANNONLAKE(i915) || IS_GEMINILAKE(i915)) > > + glk_force_audio_cdclk(i915, true); > > return ret; > > } > > @@ -806,24 +806,24 @@ static unsigned long i915_audio_component_get_power(struct device *kdev) > > static void i915_audio_component_put_power(struct device *kdev, > > unsigned long cookie) > > { > > - struct drm_i915_private *dev_priv = kdev_to_i915(kdev); > > + struct drm_i915_private *i915 = kdev_to_i915(kdev); > > /* Stop forcing CDCLK to 2*BCLK if no need for audio to be powered. */ > > - if (--dev_priv->audio_power_refcount == 0) > > - if (IS_CANNONLAKE(dev_priv) || IS_GEMINILAKE(dev_priv)) > > - glk_force_audio_cdclk(dev_priv, false); > > + if (--i915->audio_power_refcount == 0) > > + if (IS_CANNONLAKE(i915) || IS_GEMINILAKE(i915)) > > + glk_force_audio_cdclk(i915, false); > > - intel_display_power_put(dev_priv, POWER_DOMAIN_AUDIO, cookie); > > + intel_display_power_put(i915, POWER_DOMAIN_AUDIO, cookie); > > } > > static void i915_audio_component_codec_wake_override(struct device *kdev, > > bool enable) > > { > > - struct drm_i915_private *dev_priv = kdev_to_i915(kdev); > > + struct drm_i915_private *i915 = kdev_to_i915(kdev); > > unsigned long cookie; > > u32 tmp; > > - if (!IS_GEN(dev_priv, 9)) > > + if (!IS_GEN(i915, 9)) > > return; > > cookie = i915_audio_component_get_power(kdev); > > @@ -850,12 +850,12 @@ static void i915_audio_component_codec_wake_override(struct device *kdev, > > /* Get CDCLK in kHz */ > > static int i915_audio_component_get_cdclk_freq(struct device *kdev) > > { > > - struct drm_i915_private *dev_priv = kdev_to_i915(kdev); > > + struct drm_i915_private *i915 = kdev_to_i915(kdev); > > - if (WARN_ON_ONCE(!HAS_DDI(dev_priv))) > > + if (WARN_ON_ONCE(!HAS_DDI(i915))) > > return -ENODEV; > > - return dev_priv->cdclk.hw.cdclk; > > + return i915->cdclk.hw.cdclk; > > } > > /* > > @@ -868,17 +868,17 @@ static int i915_audio_component_get_cdclk_freq(struct device *kdev) > > * will get the right intel_encoder with port matched > > * Non-MST & (pipe < 0): get the right intel_encoder with port matched > > */ > > -static struct intel_encoder *get_saved_enc(struct drm_i915_private *dev_priv, > > +static struct intel_encoder *get_saved_enc(struct drm_i915_private *i915, > > int port, int pipe) > > { > > struct intel_encoder *encoder; > > /* MST */ > > if (pipe >= 0) { > > - if (WARN_ON(pipe >= ARRAY_SIZE(dev_priv->av_enc_map))) > > + if (WARN_ON(pipe >= ARRAY_SIZE(i915->av_enc_map))) > > return NULL; > > - encoder = dev_priv->av_enc_map[pipe]; > > + encoder = i915->av_enc_map[pipe]; > > /* > > * when bootup, audio driver may not know it is > > * MST or not. So it will poll all the port & pipe > > @@ -893,8 +893,8 @@ static struct intel_encoder *get_saved_enc(struct drm_i915_private *dev_priv, > > if (pipe > 0) > > return NULL; > > - for_each_pipe(dev_priv, pipe) { > > - encoder = dev_priv->av_enc_map[pipe]; > > + for_each_pipe(i915, pipe) { > > + encoder = i915->av_enc_map[pipe]; > > if (encoder == NULL) > > continue; > > @@ -911,21 +911,21 @@ static struct intel_encoder *get_saved_enc(struct drm_i915_private *dev_priv, > > static int i915_audio_component_sync_audio_rate(struct device *kdev, int port, > > int pipe, int rate) > > { > > - struct drm_i915_private *dev_priv = kdev_to_i915(kdev); > > - struct i915_audio_component *acomp = dev_priv->audio_component; > > + struct drm_i915_private *i915 = kdev_to_i915(kdev); > > + struct i915_audio_component *acomp = i915->audio_component; > > struct intel_encoder *encoder; > > struct intel_crtc *crtc; > > unsigned long cookie; > > int err = 0; > > - if (!HAS_DDI(dev_priv)) > > + if (!HAS_DDI(i915)) > > return 0; > > cookie = i915_audio_component_get_power(kdev); > > - mutex_lock(&dev_priv->av_mutex); > > + mutex_lock(&i915->av_mutex); > > /* 1. get the pipe */ > > - encoder = get_saved_enc(dev_priv, port, pipe); > > + encoder = get_saved_enc(i915, port, pipe); > > if (!encoder || !encoder->base.crtc) { > > DRM_DEBUG_KMS("Not valid for port %c\n", port_name(port)); > > err = -ENODEV; > > @@ -940,7 +940,7 @@ static int i915_audio_component_sync_audio_rate(struct device *kdev, int port, > > hsw_audio_config_update(encoder, crtc->config); > > unlock: > > - mutex_unlock(&dev_priv->av_mutex); > > + mutex_unlock(&i915->av_mutex); > > i915_audio_component_put_power(kdev, cookie); > > return err; > > } > > @@ -949,17 +949,17 @@ static int i915_audio_component_get_eld(struct device *kdev, int port, > > int pipe, bool *enabled, > > unsigned char *buf, int max_bytes) > > { > > - struct drm_i915_private *dev_priv = kdev_to_i915(kdev); > > + struct drm_i915_private *i915 = kdev_to_i915(kdev); > > struct intel_encoder *intel_encoder; > > const u8 *eld; > > int ret = -EINVAL; > > - mutex_lock(&dev_priv->av_mutex); > > + mutex_lock(&i915->av_mutex); > > - intel_encoder = get_saved_enc(dev_priv, port, pipe); > > + intel_encoder = get_saved_enc(i915, port, pipe); > > if (!intel_encoder) { > > DRM_DEBUG_KMS("Not valid for port %c\n", port_name(port)); > > - mutex_unlock(&dev_priv->av_mutex); > > + mutex_unlock(&i915->av_mutex); > > return ret; > > } > > @@ -971,7 +971,7 @@ static int i915_audio_component_get_eld(struct device *kdev, int port, > > memcpy(buf, eld, min(max_bytes, ret)); > > } > > - mutex_unlock(&dev_priv->av_mutex); > > + mutex_unlock(&i915->av_mutex); > > return ret; > > } > > @@ -989,7 +989,7 @@ static int i915_audio_component_bind(struct device *i915_kdev, > > struct device *hda_kdev, void *data) > > { > > struct i915_audio_component *acomp = data; > > - struct drm_i915_private *dev_priv = kdev_to_i915(i915_kdev); > > + struct drm_i915_private *i915 = kdev_to_i915(i915_kdev); > > int i; > > if (WARN_ON(acomp->base.ops || acomp->base.dev)) > > @@ -998,14 +998,14 @@ static int i915_audio_component_bind(struct device *i915_kdev, > > if (WARN_ON(!device_link_add(hda_kdev, i915_kdev, DL_FLAG_STATELESS))) > > return -ENOMEM; > > - drm_modeset_lock_all(&dev_priv->drm); > > + drm_modeset_lock_all(&i915->drm); > > acomp->base.ops = &i915_audio_component_ops; > > acomp->base.dev = i915_kdev; > > BUILD_BUG_ON(MAX_PORTS != I915_MAX_PORTS); > > for (i = 0; i < ARRAY_SIZE(acomp->aud_sample_rate); i++) > > acomp->aud_sample_rate[i] = 0; > > - dev_priv->audio_component = acomp; > > - drm_modeset_unlock_all(&dev_priv->drm); > > + i915->audio_component = acomp; > > + drm_modeset_unlock_all(&i915->drm); > > return 0; > > } > > @@ -1014,13 +1014,13 @@ static void i915_audio_component_unbind(struct device *i915_kdev, > > struct device *hda_kdev, void *data) > > { > > struct i915_audio_component *acomp = data; > > - struct drm_i915_private *dev_priv = kdev_to_i915(i915_kdev); > > + struct drm_i915_private *i915 = kdev_to_i915(i915_kdev); > > - drm_modeset_lock_all(&dev_priv->drm); > > + drm_modeset_lock_all(&i915->drm); > > acomp->base.ops = NULL; > > acomp->base.dev = NULL; > > - dev_priv->audio_component = NULL; > > - drm_modeset_unlock_all(&dev_priv->drm); > > + i915->audio_component = NULL; > > + drm_modeset_unlock_all(&i915->drm); > > device_link_remove(hda_kdev, i915_kdev); > > } > > @@ -1032,7 +1032,7 @@ static const struct component_ops i915_audio_component_bind_ops = { > > /** > > * i915_audio_component_init - initialize and register the audio component > > - * @dev_priv: i915 device instance > > + * @i915: i915 device instance > > * > > * This will register with the component framework a child component which > > * will bind dynamically to the snd_hda_intel driver's corresponding master > > @@ -1046,11 +1046,11 @@ static const struct component_ops i915_audio_component_bind_ops = { > > * We ignore any error during registration and continue with reduced > > * functionality (i.e. without HDMI audio). > > */ > > -static void i915_audio_component_init(struct drm_i915_private *dev_priv) > > +static void i915_audio_component_init(struct drm_i915_private *i915) > > { > > int ret; > > - ret = component_add_typed(dev_priv->drm.dev, > > + ret = component_add_typed(i915->drm.dev, > > &i915_audio_component_bind_ops, > > I915_COMPONENT_AUDIO); > > if (ret < 0) { > > @@ -1059,46 +1059,46 @@ static void i915_audio_component_init(struct drm_i915_private *dev_priv) > > return; > > } > > - dev_priv->audio_component_registered = true; > > + i915->audio_component_registered = true; > > } > > /** > > * i915_audio_component_cleanup - deregister the audio component > > - * @dev_priv: i915 device instance > > + * @i915: i915 device instance > > * > > * Deregisters the audio component, breaking any existing binding to the > > * corresponding snd_hda_intel driver's master component. > > */ > > -static void i915_audio_component_cleanup(struct drm_i915_private *dev_priv) > > +static void i915_audio_component_cleanup(struct drm_i915_private *i915) > > { > > - if (!dev_priv->audio_component_registered) > > + if (!i915->audio_component_registered) > > return; > > - component_del(dev_priv->drm.dev, &i915_audio_component_bind_ops); > > - dev_priv->audio_component_registered = false; > > + component_del(i915->drm.dev, &i915_audio_component_bind_ops); > > + i915->audio_component_registered = false; > > } > > /** > > * intel_audio_init() - Initialize the audio driver either using > > * component framework or using lpe audio bridge > > - * @dev_priv: the i915 drm device private data > > + * @i915: the i915 drm device private data > > * > > */ > > -void intel_audio_init(struct drm_i915_private *dev_priv) > > +void intel_audio_init(struct drm_i915_private *i915) > > { > > - if (intel_lpe_audio_init(dev_priv) < 0) > > - i915_audio_component_init(dev_priv); > > + if (intel_lpe_audio_init(i915) < 0) > > + i915_audio_component_init(i915); > > } > > /** > > * intel_audio_deinit() - deinitialize the audio driver > > - * @dev_priv: the i915 drm device private data > > + * @i915: the i915 drm device private data > > * > > */ > > -void intel_audio_deinit(struct drm_i915_private *dev_priv) > > +void intel_audio_deinit(struct drm_i915_private *i915) > > { > > - if ((dev_priv)->lpe_audio.platdev != NULL) > > - intel_lpe_audio_teardown(dev_priv); > > + if ((i915)->lpe_audio.platdev != NULL) > > + intel_lpe_audio_teardown(i915); > > else > > - i915_audio_component_cleanup(dev_priv); > > + i915_audio_component_cleanup(i915); > > } > > diff --git a/drivers/gpu/drm/i915/intel_audio.h b/drivers/gpu/drm/i915/intel_audio.h > > index a3657c7a7ba2..c8fde42a9fde 100644 > > --- a/drivers/gpu/drm/i915/intel_audio.h > > +++ b/drivers/gpu/drm/i915/intel_audio.h > > @@ -11,14 +11,14 @@ struct drm_i915_private; > > struct intel_crtc_state; > > struct intel_encoder; > > -void intel_init_audio_hooks(struct drm_i915_private *dev_priv); > > +void intel_init_audio_hooks(struct drm_i915_private *i915); > > void intel_audio_codec_enable(struct intel_encoder *encoder, > > const struct intel_crtc_state *crtc_state, > > const struct drm_connector_state *conn_state); > > void intel_audio_codec_disable(struct intel_encoder *encoder, > > const struct intel_crtc_state *old_crtc_state, > > const struct drm_connector_state *old_conn_state); > > -void intel_audio_init(struct drm_i915_private *dev_priv); > > -void intel_audio_deinit(struct drm_i915_private *dev_priv); > > +void intel_audio_init(struct drm_i915_private *i915); > > +void intel_audio_deinit(struct drm_i915_private *i915); > > #endif /* __INTEL_AUDIO_H__ */ > > diff --git a/drivers/gpu/drm/i915/intel_bios.c b/drivers/gpu/drm/i915/intel_bios.c > > index 1c037dfa83f5..7bdac55d08cc 100644 > > --- a/drivers/gpu/drm/i915/intel_bios.c > > +++ b/drivers/gpu/drm/i915/intel_bios.c > > @@ -202,7 +202,7 @@ get_lvds_fp_timing(const struct bdb_header *bdb, > > /* Try to find integrated panel data */ > > static void > > -parse_lfp_panel_data(struct drm_i915_private *dev_priv, > > +parse_lfp_panel_data(struct drm_i915_private *i915, > > const struct bdb_header *bdb) > > { > > const struct bdb_lvds_options *lvds_options; > > @@ -219,9 +219,9 @@ parse_lfp_panel_data(struct drm_i915_private *dev_priv, > > if (!lvds_options) > > return; > > - dev_priv->vbt.lvds_dither = lvds_options->pixel_dither; > > + i915->vbt.lvds_dither = lvds_options->pixel_dither; > > - ret = intel_opregion_get_panel_type(dev_priv); > > + ret = intel_opregion_get_panel_type(i915); > > if (ret >= 0) { > > WARN_ON(ret > 0xf); > > panel_type = ret; > > @@ -236,7 +236,7 @@ parse_lfp_panel_data(struct drm_i915_private *dev_priv, > > DRM_DEBUG_KMS("Panel type: %d (VBT)\n", panel_type); > > } > > - dev_priv->vbt.panel_type = panel_type; > > + i915->vbt.panel_type = panel_type; > > drrs_mode = (lvds_options->dps_panel_type_bits > > >> (panel_type * 2)) & MODE_MASK; > > @@ -247,15 +247,15 @@ parse_lfp_panel_data(struct drm_i915_private *dev_priv, > > */ > > switch (drrs_mode) { > > case 0: > > - dev_priv->vbt.drrs_type = STATIC_DRRS_SUPPORT; > > + i915->vbt.drrs_type = STATIC_DRRS_SUPPORT; > > DRM_DEBUG_KMS("DRRS supported mode is static\n"); > > break; > > case 2: > > - dev_priv->vbt.drrs_type = SEAMLESS_DRRS_SUPPORT; > > + i915->vbt.drrs_type = SEAMLESS_DRRS_SUPPORT; > > DRM_DEBUG_KMS("DRRS supported mode is seamless\n"); > > break; > > default: > > - dev_priv->vbt.drrs_type = DRRS_NOT_SUPPORTED; > > + i915->vbt.drrs_type = DRRS_NOT_SUPPORTED; > > DRM_DEBUG_KMS("DRRS not supported (VBT input)\n"); > > break; > > } > > @@ -278,7 +278,7 @@ parse_lfp_panel_data(struct drm_i915_private *dev_priv, > > fill_detail_timing_data(panel_fixed_mode, panel_dvo_timing); > > - dev_priv->vbt.lfp_lvds_vbt_mode = panel_fixed_mode; > > + i915->vbt.lfp_lvds_vbt_mode = panel_fixed_mode; > > DRM_DEBUG_KMS("Found panel mode in BIOS VBT tables:\n"); > > drm_mode_debug_printmodeline(panel_fixed_mode); > > @@ -290,20 +290,20 @@ parse_lfp_panel_data(struct drm_i915_private *dev_priv, > > /* check the resolution, just to be sure */ > > if (fp_timing->x_res == panel_fixed_mode->hdisplay && > > fp_timing->y_res == panel_fixed_mode->vdisplay) { > > - dev_priv->vbt.bios_lvds_val = fp_timing->lvds_reg_val; > > + i915->vbt.bios_lvds_val = fp_timing->lvds_reg_val; > > DRM_DEBUG_KMS("VBT initial LVDS value %x\n", > > - dev_priv->vbt.bios_lvds_val); > > + i915->vbt.bios_lvds_val); > > } > > } > > } > > static void > > -parse_lfp_backlight(struct drm_i915_private *dev_priv, > > +parse_lfp_backlight(struct drm_i915_private *i915, > > const struct bdb_header *bdb) > > { > > const struct bdb_lfp_backlight_data *backlight_data; > > const struct lfp_backlight_data_entry *entry; > > - int panel_type = dev_priv->vbt.panel_type; > > + int panel_type = i915->vbt.panel_type; > > backlight_data = find_section(bdb, BDB_LVDS_BACKLIGHT); > > if (!backlight_data) > > @@ -317,38 +317,38 @@ parse_lfp_backlight(struct drm_i915_private *dev_priv, > > entry = &backlight_data->data[panel_type]; > > - dev_priv->vbt.backlight.present = entry->type == BDB_BACKLIGHT_TYPE_PWM; > > - if (!dev_priv->vbt.backlight.present) { > > + i915->vbt.backlight.present = entry->type == BDB_BACKLIGHT_TYPE_PWM; > > + if (!i915->vbt.backlight.present) { > > DRM_DEBUG_KMS("PWM backlight not present in VBT (type %u)\n", > > entry->type); > > return; > > } > > - dev_priv->vbt.backlight.type = INTEL_BACKLIGHT_DISPLAY_DDI; > > + i915->vbt.backlight.type = INTEL_BACKLIGHT_DISPLAY_DDI; > > if (bdb->version >= 191 && > > get_blocksize(backlight_data) >= sizeof(*backlight_data)) { > > const struct lfp_backlight_control_method *method; > > method = &backlight_data->backlight_control[panel_type]; > > - dev_priv->vbt.backlight.type = method->type; > > - dev_priv->vbt.backlight.controller = method->controller; > > + i915->vbt.backlight.type = method->type; > > + i915->vbt.backlight.controller = method->controller; > > } > > - dev_priv->vbt.backlight.pwm_freq_hz = entry->pwm_freq_hz; > > - dev_priv->vbt.backlight.active_low_pwm = entry->active_low_pwm; > > - dev_priv->vbt.backlight.min_brightness = entry->min_brightness; > > + i915->vbt.backlight.pwm_freq_hz = entry->pwm_freq_hz; > > + i915->vbt.backlight.active_low_pwm = entry->active_low_pwm; > > + i915->vbt.backlight.min_brightness = entry->min_brightness; > > DRM_DEBUG_KMS("VBT backlight PWM modulation frequency %u Hz, " > > "active %s, min brightness %u, level %u, controller %u\n", > > - dev_priv->vbt.backlight.pwm_freq_hz, > > - dev_priv->vbt.backlight.active_low_pwm ? "low" : "high", > > - dev_priv->vbt.backlight.min_brightness, > > + i915->vbt.backlight.pwm_freq_hz, > > + i915->vbt.backlight.active_low_pwm ? "low" : "high", > > + i915->vbt.backlight.min_brightness, > > backlight_data->level[panel_type], > > - dev_priv->vbt.backlight.controller); > > + i915->vbt.backlight.controller); > > } > > /* Try to find sdvo panel data */ > > static void > > -parse_sdvo_panel_data(struct drm_i915_private *dev_priv, > > +parse_sdvo_panel_data(struct drm_i915_private *i915, > > const struct bdb_header *bdb) > > { > > const struct bdb_sdvo_panel_dtds *dtds; > > @@ -381,16 +381,16 @@ parse_sdvo_panel_data(struct drm_i915_private *dev_priv, > > fill_detail_timing_data(panel_fixed_mode, &dtds->dtds[index]); > > - dev_priv->vbt.sdvo_lvds_vbt_mode = panel_fixed_mode; > > + i915->vbt.sdvo_lvds_vbt_mode = panel_fixed_mode; > > DRM_DEBUG_KMS("Found SDVO panel mode in BIOS VBT tables:\n"); > > drm_mode_debug_printmodeline(panel_fixed_mode); > > } > > -static int intel_bios_ssc_frequency(struct drm_i915_private *dev_priv, > > +static int intel_bios_ssc_frequency(struct drm_i915_private *i915, > > bool alternate) > > { > > - switch (INTEL_GEN(dev_priv)) { > > + switch (INTEL_GEN(i915)) { > > case 2: > > return alternate ? 66667 : 48000; > > case 3: > > @@ -402,7 +402,7 @@ static int intel_bios_ssc_frequency(struct drm_i915_private *dev_priv, > > } > > static void > > -parse_general_features(struct drm_i915_private *dev_priv, > > +parse_general_features(struct drm_i915_private *i915, > > const struct bdb_header *bdb) > > { > > const struct bdb_general_features *general; > > @@ -411,30 +411,30 @@ parse_general_features(struct drm_i915_private *dev_priv, > > if (!general) > > return; > > - dev_priv->vbt.int_tv_support = general->int_tv_support; > > + i915->vbt.int_tv_support = general->int_tv_support; > > /* int_crt_support can't be trusted on earlier platforms */ > > if (bdb->version >= 155 && > > - (HAS_DDI(dev_priv) || IS_VALLEYVIEW(dev_priv))) > > - dev_priv->vbt.int_crt_support = general->int_crt_support; > > - dev_priv->vbt.lvds_use_ssc = general->enable_ssc; > > - dev_priv->vbt.lvds_ssc_freq = > > - intel_bios_ssc_frequency(dev_priv, general->ssc_freq); > > - dev_priv->vbt.display_clock_mode = general->display_clock_mode; > > - dev_priv->vbt.fdi_rx_polarity_inverted = general->fdi_rx_polarity_inverted; > > + (HAS_DDI(i915) || IS_VALLEYVIEW(i915))) > > + i915->vbt.int_crt_support = general->int_crt_support; > > + i915->vbt.lvds_use_ssc = general->enable_ssc; > > + i915->vbt.lvds_ssc_freq = > > + intel_bios_ssc_frequency(i915, general->ssc_freq); > > + i915->vbt.display_clock_mode = general->display_clock_mode; > > + i915->vbt.fdi_rx_polarity_inverted = general->fdi_rx_polarity_inverted; > > if (bdb->version >= 181) { > > - dev_priv->vbt.orientation = general->rotate_180 ? > > + i915->vbt.orientation = general->rotate_180 ? > > DRM_MODE_PANEL_ORIENTATION_BOTTOM_UP : > > DRM_MODE_PANEL_ORIENTATION_NORMAL; > > } else { > > - dev_priv->vbt.orientation = DRM_MODE_PANEL_ORIENTATION_UNKNOWN; > > + i915->vbt.orientation = DRM_MODE_PANEL_ORIENTATION_UNKNOWN; > > } > > DRM_DEBUG_KMS("BDB_GENERAL_FEATURES int_tv_support %d int_crt_support %d lvds_use_ssc %d lvds_ssc_freq %d display_clock_mode %d fdi_rx_polarity_inverted %d\n", > > - dev_priv->vbt.int_tv_support, > > - dev_priv->vbt.int_crt_support, > > - dev_priv->vbt.lvds_use_ssc, > > - dev_priv->vbt.lvds_ssc_freq, > > - dev_priv->vbt.display_clock_mode, > > - dev_priv->vbt.fdi_rx_polarity_inverted); > > + i915->vbt.int_tv_support, > > + i915->vbt.int_crt_support, > > + i915->vbt.lvds_use_ssc, > > + i915->vbt.lvds_ssc_freq, > > + i915->vbt.display_clock_mode, > > + i915->vbt.fdi_rx_polarity_inverted); > > } > > static const struct child_device_config * > > @@ -444,7 +444,7 @@ child_device_ptr(const struct bdb_general_definitions *defs, int i) > > } > > static void > > -parse_sdvo_device_mapping(struct drm_i915_private *dev_priv, u8 bdb_version) > > +parse_sdvo_device_mapping(struct drm_i915_private *i915, u8 bdb_version) > > { > > struct sdvo_device_mapping *mapping; > > const struct child_device_config *child; > > @@ -454,13 +454,13 @@ parse_sdvo_device_mapping(struct drm_i915_private *dev_priv, u8 bdb_version) > > * Only parse SDVO mappings on gens that could have SDVO. This isn't > > * accurate and doesn't have to be, as long as it's not too strict. > > */ > > - if (!IS_GEN_RANGE(dev_priv, 3, 7)) { > > + if (!IS_GEN_RANGE(i915, 3, 7)) { > > DRM_DEBUG_KMS("Skipping SDVO device mapping\n"); > > return; > > } > > - for (i = 0, count = 0; i < dev_priv->vbt.child_dev_num; i++) { > > - child = dev_priv->vbt.child_dev + i; > > + for (i = 0, count = 0; i < i915->vbt.child_dev_num; i++) { > > + child = i915->vbt.child_dev + i; > > if (child->slave_addr != SLAVE_ADDR1 && > > child->slave_addr != SLAVE_ADDR2) { > > @@ -481,7 +481,7 @@ parse_sdvo_device_mapping(struct drm_i915_private *dev_priv, u8 bdb_version) > > child->slave_addr, > > (child->dvo_port == DEVICE_PORT_DVOB) ? > > "SDVOB" : "SDVOC"); > > - mapping = &dev_priv->vbt.sdvo_mappings[child->dvo_port - 1]; > > + mapping = &i915->vbt.sdvo_mappings[child->dvo_port - 1]; > > if (!mapping->initialized) { > > mapping->dvo_port = child->dvo_port; > > mapping->slave_addr = child->slave_addr; > > @@ -515,7 +515,7 @@ parse_sdvo_device_mapping(struct drm_i915_private *dev_priv, u8 bdb_version) > > } > > static void > > -parse_driver_features(struct drm_i915_private *dev_priv, > > +parse_driver_features(struct drm_i915_private *i915, > > const struct bdb_header *bdb) > > { > > const struct bdb_driver_features *driver; > > @@ -524,14 +524,14 @@ parse_driver_features(struct drm_i915_private *dev_priv, > > if (!driver) > > return; > > - if (INTEL_GEN(dev_priv) >= 5) { > > + if (INTEL_GEN(i915) >= 5) { > > /* > > * Note that we consider BDB_DRIVER_FEATURE_INT_SDVO_LVDS > > * to mean "eDP". The VBT spec doesn't agree with that > > * interpretation, but real world VBTs seem to. > > */ > > if (driver->lvds_config != BDB_DRIVER_FEATURE_INT_LVDS) > > - dev_priv->vbt.int_lvds_support = 0; > > + i915->vbt.int_lvds_support = 0; > > } else { > > /* > > * FIXME it's not clear which BDB version has the LVDS config > > @@ -547,7 +547,7 @@ parse_driver_features(struct drm_i915_private *dev_priv, > > if (bdb->version >= 134 && > > driver->lvds_config != BDB_DRIVER_FEATURE_INT_LVDS && > > driver->lvds_config != BDB_DRIVER_FEATURE_INT_SDVO_LVDS) > > - dev_priv->vbt.int_lvds_support = 0; > > + i915->vbt.int_lvds_support = 0; > > } > > DRM_DEBUG_KMS("DRRS State Enabled:%d\n", driver->drrs_enabled); > > @@ -558,17 +558,17 @@ parse_driver_features(struct drm_i915_private *dev_priv, > > * driver->drrs_enabled=false > > */ > > if (!driver->drrs_enabled) > > - dev_priv->vbt.drrs_type = DRRS_NOT_SUPPORTED; > > - dev_priv->vbt.psr.enable = driver->psr_enabled; > > + i915->vbt.drrs_type = DRRS_NOT_SUPPORTED; > > + i915->vbt.psr.enable = driver->psr_enabled; > > } > > static void > > -parse_edp(struct drm_i915_private *dev_priv, const struct bdb_header *bdb) > > +parse_edp(struct drm_i915_private *i915, const struct bdb_header *bdb) > > { > > const struct bdb_edp *edp; > > const struct edp_power_seq *edp_pps; > > const struct edp_fast_link_params *edp_link_params; > > - int panel_type = dev_priv->vbt.panel_type; > > + int panel_type = i915->vbt.panel_type; > > edp = find_section(bdb, BDB_EDP); > > if (!edp) > > @@ -576,13 +576,13 @@ parse_edp(struct drm_i915_private *dev_priv, const struct bdb_header *bdb) > > switch ((edp->color_depth >> (panel_type * 2)) & 3) { > > case EDP_18BPP: > > - dev_priv->vbt.edp.bpp = 18; > > + i915->vbt.edp.bpp = 18; > > break; > > case EDP_24BPP: > > - dev_priv->vbt.edp.bpp = 24; > > + i915->vbt.edp.bpp = 24; > > break; > > case EDP_30BPP: > > - dev_priv->vbt.edp.bpp = 30; > > + i915->vbt.edp.bpp = 30; > > break; > > } > > @@ -590,14 +590,14 @@ parse_edp(struct drm_i915_private *dev_priv, const struct bdb_header *bdb) > > edp_pps = &edp->power_seqs[panel_type]; > > edp_link_params = &edp->fast_link_params[panel_type]; > > - dev_priv->vbt.edp.pps = *edp_pps; > > + i915->vbt.edp.pps = *edp_pps; > > switch (edp_link_params->rate) { > > case EDP_RATE_1_62: > > - dev_priv->vbt.edp.rate = DP_LINK_BW_1_62; > > + i915->vbt.edp.rate = DP_LINK_BW_1_62; > > break; > > case EDP_RATE_2_7: > > - dev_priv->vbt.edp.rate = DP_LINK_BW_2_7; > > + i915->vbt.edp.rate = DP_LINK_BW_2_7; > > break; > > default: > > DRM_DEBUG_KMS("VBT has unknown eDP link rate value %u\n", > > @@ -607,13 +607,13 @@ parse_edp(struct drm_i915_private *dev_priv, const struct bdb_header *bdb) > > switch (edp_link_params->lanes) { > > case EDP_LANE_1: > > - dev_priv->vbt.edp.lanes = 1; > > + i915->vbt.edp.lanes = 1; > > break; > > case EDP_LANE_2: > > - dev_priv->vbt.edp.lanes = 2; > > + i915->vbt.edp.lanes = 2; > > break; > > case EDP_LANE_4: > > - dev_priv->vbt.edp.lanes = 4; > > + i915->vbt.edp.lanes = 4; > > break; > > default: > > DRM_DEBUG_KMS("VBT has unknown eDP lane count value %u\n", > > @@ -623,16 +623,16 @@ parse_edp(struct drm_i915_private *dev_priv, const struct bdb_header *bdb) > > switch (edp_link_params->preemphasis) { > > case EDP_PREEMPHASIS_NONE: > > - dev_priv->vbt.edp.preemphasis = DP_TRAIN_PRE_EMPH_LEVEL_0; > > + i915->vbt.edp.preemphasis = DP_TRAIN_PRE_EMPH_LEVEL_0; > > break; > > case EDP_PREEMPHASIS_3_5dB: > > - dev_priv->vbt.edp.preemphasis = DP_TRAIN_PRE_EMPH_LEVEL_1; > > + i915->vbt.edp.preemphasis = DP_TRAIN_PRE_EMPH_LEVEL_1; > > break; > > case EDP_PREEMPHASIS_6dB: > > - dev_priv->vbt.edp.preemphasis = DP_TRAIN_PRE_EMPH_LEVEL_2; > > + i915->vbt.edp.preemphasis = DP_TRAIN_PRE_EMPH_LEVEL_2; > > break; > > case EDP_PREEMPHASIS_9_5dB: > > - dev_priv->vbt.edp.preemphasis = DP_TRAIN_PRE_EMPH_LEVEL_3; > > + i915->vbt.edp.preemphasis = DP_TRAIN_PRE_EMPH_LEVEL_3; > > break; > > default: > > DRM_DEBUG_KMS("VBT has unknown eDP pre-emphasis value %u\n", > > @@ -642,16 +642,16 @@ parse_edp(struct drm_i915_private *dev_priv, const struct bdb_header *bdb) > > switch (edp_link_params->vswing) { > > case EDP_VSWING_0_4V: > > - dev_priv->vbt.edp.vswing = DP_TRAIN_VOLTAGE_SWING_LEVEL_0; > > + i915->vbt.edp.vswing = DP_TRAIN_VOLTAGE_SWING_LEVEL_0; > > break; > > case EDP_VSWING_0_6V: > > - dev_priv->vbt.edp.vswing = DP_TRAIN_VOLTAGE_SWING_LEVEL_1; > > + i915->vbt.edp.vswing = DP_TRAIN_VOLTAGE_SWING_LEVEL_1; > > break; > > case EDP_VSWING_0_8V: > > - dev_priv->vbt.edp.vswing = DP_TRAIN_VOLTAGE_SWING_LEVEL_2; > > + i915->vbt.edp.vswing = DP_TRAIN_VOLTAGE_SWING_LEVEL_2; > > break; > > case EDP_VSWING_1_2V: > > - dev_priv->vbt.edp.vswing = DP_TRAIN_VOLTAGE_SWING_LEVEL_3; > > + i915->vbt.edp.vswing = DP_TRAIN_VOLTAGE_SWING_LEVEL_3; > > break; > > default: > > DRM_DEBUG_KMS("VBT has unknown eDP voltage swing value %u\n", > > @@ -664,21 +664,21 @@ parse_edp(struct drm_i915_private *dev_priv, const struct bdb_header *bdb) > > /* Don't read from VBT if module parameter has valid value*/ > > if (i915_modparams.edp_vswing) { > > - dev_priv->vbt.edp.low_vswing = > > + i915->vbt.edp.low_vswing = > > i915_modparams.edp_vswing == 1; > > } else { > > vswing = (edp->edp_vswing_preemph >> (panel_type * 4)) & 0xF; > > - dev_priv->vbt.edp.low_vswing = vswing == 0; > > + i915->vbt.edp.low_vswing = vswing == 0; > > } > > } > > } > > static void > > -parse_psr(struct drm_i915_private *dev_priv, const struct bdb_header *bdb) > > +parse_psr(struct drm_i915_private *i915, const struct bdb_header *bdb) > > { > > const struct bdb_psr *psr; > > const struct psr_table *psr_table; > > - int panel_type = dev_priv->vbt.panel_type; > > + int panel_type = i915->vbt.panel_type; > > psr = find_section(bdb, BDB_PSR); > > if (!psr) { > > @@ -688,25 +688,25 @@ parse_psr(struct drm_i915_private *dev_priv, const struct bdb_header *bdb) > > psr_table = &psr->psr_table[panel_type]; > > - dev_priv->vbt.psr.full_link = psr_table->full_link; > > - dev_priv->vbt.psr.require_aux_wakeup = psr_table->require_aux_to_wakeup; > > + i915->vbt.psr.full_link = psr_table->full_link; > > + i915->vbt.psr.require_aux_wakeup = psr_table->require_aux_to_wakeup; > > /* Allowed VBT values goes from 0 to 15 */ > > - dev_priv->vbt.psr.idle_frames = psr_table->idle_frames < 0 ? 0 : > > + i915->vbt.psr.idle_frames = psr_table->idle_frames < 0 ? 0 : > > psr_table->idle_frames > 15 ? 15 : psr_table->idle_frames; > > switch (psr_table->lines_to_wait) { > > case 0: > > - dev_priv->vbt.psr.lines_to_wait = PSR_0_LINES_TO_WAIT; > > + i915->vbt.psr.lines_to_wait = PSR_0_LINES_TO_WAIT; > > break; > > case 1: > > - dev_priv->vbt.psr.lines_to_wait = PSR_1_LINE_TO_WAIT; > > + i915->vbt.psr.lines_to_wait = PSR_1_LINE_TO_WAIT; > > break; > > case 2: > > - dev_priv->vbt.psr.lines_to_wait = PSR_4_LINES_TO_WAIT; > > + i915->vbt.psr.lines_to_wait = PSR_4_LINES_TO_WAIT; > > break; > > case 3: > > - dev_priv->vbt.psr.lines_to_wait = PSR_8_LINES_TO_WAIT; > > + i915->vbt.psr.lines_to_wait = PSR_8_LINES_TO_WAIT; > > break; > > default: > > DRM_DEBUG_KMS("VBT has unknown PSR lines to wait %u\n", > > @@ -719,48 +719,48 @@ parse_psr(struct drm_i915_private *dev_priv, const struct bdb_header *bdb) > > * Old decimal value is wake up time in multiples of 100 us. > > */ > > if (bdb->version >= 205 && > > - (IS_GEN9_BC(dev_priv) || IS_GEMINILAKE(dev_priv) || > > - INTEL_GEN(dev_priv) >= 10)) { > > + (IS_GEN9_BC(i915) || IS_GEMINILAKE(i915) || > > + INTEL_GEN(i915) >= 10)) { > > switch (psr_table->tp1_wakeup_time) { > > case 0: > > - dev_priv->vbt.psr.tp1_wakeup_time_us = 500; > > + i915->vbt.psr.tp1_wakeup_time_us = 500; > > break; > > case 1: > > - dev_priv->vbt.psr.tp1_wakeup_time_us = 100; > > + i915->vbt.psr.tp1_wakeup_time_us = 100; > > break; > > case 3: > > - dev_priv->vbt.psr.tp1_wakeup_time_us = 0; > > + i915->vbt.psr.tp1_wakeup_time_us = 0; > > break; > > default: > > DRM_DEBUG_KMS("VBT tp1 wakeup time value %d is outside range[0-3], defaulting to max value 2500us\n", > > psr_table->tp1_wakeup_time); > > /* fallthrough */ > > case 2: > > - dev_priv->vbt.psr.tp1_wakeup_time_us = 2500; > > + i915->vbt.psr.tp1_wakeup_time_us = 2500; > > break; > > } > > switch (psr_table->tp2_tp3_wakeup_time) { > > case 0: > > - dev_priv->vbt.psr.tp2_tp3_wakeup_time_us = 500; > > + i915->vbt.psr.tp2_tp3_wakeup_time_us = 500; > > break; > > case 1: > > - dev_priv->vbt.psr.tp2_tp3_wakeup_time_us = 100; > > + i915->vbt.psr.tp2_tp3_wakeup_time_us = 100; > > break; > > case 3: > > - dev_priv->vbt.psr.tp2_tp3_wakeup_time_us = 0; > > + i915->vbt.psr.tp2_tp3_wakeup_time_us = 0; > > break; > > default: > > DRM_DEBUG_KMS("VBT tp2_tp3 wakeup time value %d is outside range[0-3], defaulting to max value 2500us\n", > > psr_table->tp2_tp3_wakeup_time); > > /* fallthrough */ > > case 2: > > - dev_priv->vbt.psr.tp2_tp3_wakeup_time_us = 2500; > > + i915->vbt.psr.tp2_tp3_wakeup_time_us = 2500; > > break; > > } > > } else { > > - dev_priv->vbt.psr.tp1_wakeup_time_us = psr_table->tp1_wakeup_time * 100; > > - dev_priv->vbt.psr.tp2_tp3_wakeup_time_us = psr_table->tp2_tp3_wakeup_time * 100; > > + i915->vbt.psr.tp1_wakeup_time_us = psr_table->tp1_wakeup_time * 100; > > + i915->vbt.psr.tp2_tp3_wakeup_time_us = psr_table->tp2_tp3_wakeup_time * 100; > > } > > if (bdb->version >= 226) { > > @@ -782,74 +782,74 @@ parse_psr(struct drm_i915_private *dev_priv, const struct bdb_header *bdb) > > wakeup_time = 2500; > > break; > > } > > - dev_priv->vbt.psr.psr2_tp2_tp3_wakeup_time_us = wakeup_time; > > + i915->vbt.psr.psr2_tp2_tp3_wakeup_time_us = wakeup_time; > > } else { > > /* Reusing PSR1 wakeup time for PSR2 in older VBTs */ > > - dev_priv->vbt.psr.psr2_tp2_tp3_wakeup_time_us = dev_priv->vbt.psr.tp2_tp3_wakeup_time_us; > > + i915->vbt.psr.psr2_tp2_tp3_wakeup_time_us = i915->vbt.psr.tp2_tp3_wakeup_time_us; > > } > > } > > -static void parse_dsi_backlight_ports(struct drm_i915_private *dev_priv, > > +static void parse_dsi_backlight_ports(struct drm_i915_private *i915, > > u16 version, enum port port) > > { > > - if (!dev_priv->vbt.dsi.config->dual_link || version < 197) { > > - dev_priv->vbt.dsi.bl_ports = BIT(port); > > - if (dev_priv->vbt.dsi.config->cabc_supported) > > - dev_priv->vbt.dsi.cabc_ports = BIT(port); > > + if (!i915->vbt.dsi.config->dual_link || version < 197) { > > + i915->vbt.dsi.bl_ports = BIT(port); > > + if (i915->vbt.dsi.config->cabc_supported) > > + i915->vbt.dsi.cabc_ports = BIT(port); > > return; > > } > > - switch (dev_priv->vbt.dsi.config->dl_dcs_backlight_ports) { > > + switch (i915->vbt.dsi.config->dl_dcs_backlight_ports) { > > case DL_DCS_PORT_A: > > - dev_priv->vbt.dsi.bl_ports = BIT(PORT_A); > > + i915->vbt.dsi.bl_ports = BIT(PORT_A); > > break; > > case DL_DCS_PORT_C: > > - dev_priv->vbt.dsi.bl_ports = BIT(PORT_C); > > + i915->vbt.dsi.bl_ports = BIT(PORT_C); > > break; > > default: > > case DL_DCS_PORT_A_AND_C: > > - dev_priv->vbt.dsi.bl_ports = BIT(PORT_A) | BIT(PORT_C); > > + i915->vbt.dsi.bl_ports = BIT(PORT_A) | BIT(PORT_C); > > break; > > } > > - if (!dev_priv->vbt.dsi.config->cabc_supported) > > + if (!i915->vbt.dsi.config->cabc_supported) > > return; > > - switch (dev_priv->vbt.dsi.config->dl_dcs_cabc_ports) { > > + switch (i915->vbt.dsi.config->dl_dcs_cabc_ports) { > > case DL_DCS_PORT_A: > > - dev_priv->vbt.dsi.cabc_ports = BIT(PORT_A); > > + i915->vbt.dsi.cabc_ports = BIT(PORT_A); > > break; > > case DL_DCS_PORT_C: > > - dev_priv->vbt.dsi.cabc_ports = BIT(PORT_C); > > + i915->vbt.dsi.cabc_ports = BIT(PORT_C); > > break; > > default: > > case DL_DCS_PORT_A_AND_C: > > - dev_priv->vbt.dsi.cabc_ports = > > + i915->vbt.dsi.cabc_ports = > > BIT(PORT_A) | BIT(PORT_C); > > break; > > } > > } > > static void > > -parse_mipi_config(struct drm_i915_private *dev_priv, > > +parse_mipi_config(struct drm_i915_private *i915, > > const struct bdb_header *bdb) > > { > > const struct bdb_mipi_config *start; > > const struct mipi_config *config; > > const struct mipi_pps_data *pps; > > - int panel_type = dev_priv->vbt.panel_type; > > + int panel_type = i915->vbt.panel_type; > > enum port port; > > /* parse MIPI blocks only if LFP type is MIPI */ > > - if (!intel_bios_is_dsi_present(dev_priv, &port)) > > + if (!intel_bios_is_dsi_present(i915, &port)) > > return; > > /* Initialize this to undefined indicating no generic MIPI support */ > > - dev_priv->vbt.dsi.panel_id = MIPI_DSI_UNDEFINED_PANEL_ID; > > + i915->vbt.dsi.panel_id = MIPI_DSI_UNDEFINED_PANEL_ID; > > /* Block #40 is already parsed and panel_fixed_mode is > > - * stored in dev_priv->lfp_lvds_vbt_mode > > + * stored in i915->lfp_lvds_vbt_mode > > * resuse this when needed > > */ > > @@ -873,17 +873,17 @@ parse_mipi_config(struct drm_i915_private *dev_priv, > > pps = &start->pps[panel_type]; > > /* store as of now full data. Trim when we realise all is not needed */ > > - dev_priv->vbt.dsi.config = kmemdup(config, sizeof(struct mipi_config), GFP_KERNEL); > > - if (!dev_priv->vbt.dsi.config) > > + i915->vbt.dsi.config = kmemdup(config, sizeof(struct mipi_config), GFP_KERNEL); > > + if (!i915->vbt.dsi.config) > > return; > > - dev_priv->vbt.dsi.pps = kmemdup(pps, sizeof(struct mipi_pps_data), GFP_KERNEL); > > - if (!dev_priv->vbt.dsi.pps) { > > - kfree(dev_priv->vbt.dsi.config); > > + i915->vbt.dsi.pps = kmemdup(pps, sizeof(struct mipi_pps_data), GFP_KERNEL); > > + if (!i915->vbt.dsi.pps) { > > + kfree(i915->vbt.dsi.config); > > return; > > } > > - parse_dsi_backlight_ports(dev_priv, bdb->version, port); > > + parse_dsi_backlight_ports(i915, bdb->version, port); > > /* FIXME is the 90 vs. 270 correct? */ > > switch (config->rotation) { > > @@ -892,25 +892,25 @@ parse_mipi_config(struct drm_i915_private *dev_priv, > > * Most (all?) VBTs claim 0 degrees despite having > > * an upside down panel, thus we do not trust this. > > */ > > - dev_priv->vbt.dsi.orientation = > > + i915->vbt.dsi.orientation = > > DRM_MODE_PANEL_ORIENTATION_UNKNOWN; > > break; > > case ENABLE_ROTATION_90: > > - dev_priv->vbt.dsi.orientation = > > + i915->vbt.dsi.orientation = > > DRM_MODE_PANEL_ORIENTATION_RIGHT_UP; > > break; > > case ENABLE_ROTATION_180: > > - dev_priv->vbt.dsi.orientation = > > + i915->vbt.dsi.orientation = > > DRM_MODE_PANEL_ORIENTATION_BOTTOM_UP; > > break; > > case ENABLE_ROTATION_270: > > - dev_priv->vbt.dsi.orientation = > > + i915->vbt.dsi.orientation = > > DRM_MODE_PANEL_ORIENTATION_LEFT_UP; > > break; > > } > > /* We have mandatory mipi config blocks. Initialize as generic panel */ > > - dev_priv->vbt.dsi.panel_id = MIPI_DSI_GENERIC_PANEL_ID; > > + i915->vbt.dsi.panel_id = MIPI_DSI_GENERIC_PANEL_ID; > > } > > /* Find the sequence block and size for the given panel. */ > > @@ -1073,12 +1073,12 @@ static int goto_next_sequence_v3(const u8 *data, int index, int total) > > * Get len of pre-fixed deassert fragment from a v1 init OTP sequence, > > * skip all delay + gpio operands and stop at the first DSI packet op. > > */ > > -static int get_init_otp_deassert_fragment_len(struct drm_i915_private *dev_priv) > > +static int get_init_otp_deassert_fragment_len(struct drm_i915_private *i915) > > { > > - const u8 *data = dev_priv->vbt.dsi.sequence[MIPI_SEQ_INIT_OTP]; > > + const u8 *data = i915->vbt.dsi.sequence[MIPI_SEQ_INIT_OTP]; > > int index, len; > > - if (WARN_ON(!data || dev_priv->vbt.dsi.seq_version != 1)) > > + if (WARN_ON(!data || i915->vbt.dsi.seq_version != 1)) > > return 0; > > /* index = 1 to skip sequence byte */ > > @@ -1106,54 +1106,54 @@ static int get_init_otp_deassert_fragment_len(struct drm_i915_private *dev_priv) > > * these devices we split the init OTP sequence into a deassert sequence and > > * the actual init OTP part. > > */ > > -static void fixup_mipi_sequences(struct drm_i915_private *dev_priv) > > +static void fixup_mipi_sequences(struct drm_i915_private *i915) > > { > > u8 *init_otp; > > int len; > > /* Limit this to VLV for now. */ > > - if (!IS_VALLEYVIEW(dev_priv)) > > + if (!IS_VALLEYVIEW(i915)) > > return; > > /* Limit this to v1 vid-mode sequences */ > > - if (dev_priv->vbt.dsi.config->is_cmd_mode || > > - dev_priv->vbt.dsi.seq_version != 1) > > + if (i915->vbt.dsi.config->is_cmd_mode || > > + i915->vbt.dsi.seq_version != 1) > > return; > > /* Only do this if there are otp and assert seqs and no deassert seq */ > > - if (!dev_priv->vbt.dsi.sequence[MIPI_SEQ_INIT_OTP] || > > - !dev_priv->vbt.dsi.sequence[MIPI_SEQ_ASSERT_RESET] || > > - dev_priv->vbt.dsi.sequence[MIPI_SEQ_DEASSERT_RESET]) > > + if (!i915->vbt.dsi.sequence[MIPI_SEQ_INIT_OTP] || > > + !i915->vbt.dsi.sequence[MIPI_SEQ_ASSERT_RESET] || > > + i915->vbt.dsi.sequence[MIPI_SEQ_DEASSERT_RESET]) > > return; > > /* The deassert-sequence ends at the first DSI packet */ > > - len = get_init_otp_deassert_fragment_len(dev_priv); > > + len = get_init_otp_deassert_fragment_len(i915); > > if (!len) > > return; > > DRM_DEBUG_KMS("Using init OTP fragment to deassert reset\n"); > > /* Copy the fragment, update seq byte and terminate it */ > > - init_otp = (u8 *)dev_priv->vbt.dsi.sequence[MIPI_SEQ_INIT_OTP]; > > - dev_priv->vbt.dsi.deassert_seq = kmemdup(init_otp, len + 1, GFP_KERNEL); > > - if (!dev_priv->vbt.dsi.deassert_seq) > > + init_otp = (u8 *)i915->vbt.dsi.sequence[MIPI_SEQ_INIT_OTP]; > > + i915->vbt.dsi.deassert_seq = kmemdup(init_otp, len + 1, GFP_KERNEL); > > + if (!i915->vbt.dsi.deassert_seq) > > return; > > - dev_priv->vbt.dsi.deassert_seq[0] = MIPI_SEQ_DEASSERT_RESET; > > - dev_priv->vbt.dsi.deassert_seq[len] = MIPI_SEQ_ELEM_END; > > + i915->vbt.dsi.deassert_seq[0] = MIPI_SEQ_DEASSERT_RESET; > > + i915->vbt.dsi.deassert_seq[len] = MIPI_SEQ_ELEM_END; > > /* Use the copy for deassert */ > > - dev_priv->vbt.dsi.sequence[MIPI_SEQ_DEASSERT_RESET] = > > - dev_priv->vbt.dsi.deassert_seq; > > + i915->vbt.dsi.sequence[MIPI_SEQ_DEASSERT_RESET] = > > + i915->vbt.dsi.deassert_seq; > > /* Replace the last byte of the fragment with init OTP seq byte */ > > init_otp[len - 1] = MIPI_SEQ_INIT_OTP; > > /* And make MIPI_MIPI_SEQ_INIT_OTP point to it */ > > - dev_priv->vbt.dsi.sequence[MIPI_SEQ_INIT_OTP] = init_otp + len - 1; > > + i915->vbt.dsi.sequence[MIPI_SEQ_INIT_OTP] = init_otp + len - 1; > > } > > static void > > -parse_mipi_sequence(struct drm_i915_private *dev_priv, > > +parse_mipi_sequence(struct drm_i915_private *i915, > > const struct bdb_header *bdb) > > { > > - int panel_type = dev_priv->vbt.panel_type; > > + int panel_type = i915->vbt.panel_type; > > const struct bdb_mipi_sequence *sequence; > > const u8 *seq_data; > > u32 seq_size; > > @@ -1161,7 +1161,7 @@ parse_mipi_sequence(struct drm_i915_private *dev_priv, > > int index = 0; > > /* Only our generic panel driver uses the sequence block. */ > > - if (dev_priv->vbt.dsi.panel_id != MIPI_DSI_GENERIC_PANEL_ID) > > + if (i915->vbt.dsi.panel_id != MIPI_DSI_GENERIC_PANEL_ID) > > return; > > sequence = find_section(bdb, BDB_MIPI_SEQUENCE); > > @@ -1202,7 +1202,7 @@ parse_mipi_sequence(struct drm_i915_private *dev_priv, > > if (seq_id == MIPI_SEQ_TEAR_ON || seq_id == MIPI_SEQ_TEAR_OFF) > > DRM_DEBUG_KMS("Unsupported sequence %u\n", seq_id); > > - dev_priv->vbt.dsi.sequence[seq_id] = data + index; > > + i915->vbt.dsi.sequence[seq_id] = data + index; > > if (sequence->version >= 3) > > index = goto_next_sequence_v3(data, index, seq_size); > > @@ -1214,18 +1214,18 @@ parse_mipi_sequence(struct drm_i915_private *dev_priv, > > } > > } > > - dev_priv->vbt.dsi.data = data; > > - dev_priv->vbt.dsi.size = seq_size; > > - dev_priv->vbt.dsi.seq_version = sequence->version; > > + i915->vbt.dsi.data = data; > > + i915->vbt.dsi.size = seq_size; > > + i915->vbt.dsi.seq_version = sequence->version; > > - fixup_mipi_sequences(dev_priv); > > + fixup_mipi_sequences(i915); > > DRM_DEBUG_DRIVER("MIPI related VBT parsing complete\n"); > > return; > > err: > > kfree(data); > > - memset(dev_priv->vbt.dsi.sequence, 0, sizeof(dev_priv->vbt.dsi.sequence)); > > + memset(i915->vbt.dsi.sequence, 0, sizeof(i915->vbt.dsi.sequence)); > > } > > static u8 translate_iboost(u8 val) > > @@ -1254,16 +1254,16 @@ static enum port get_port_by_ddc_pin(struct drm_i915_private *i915, u8 ddc_pin) > > return PORT_NONE; > > } > > -static void sanitize_ddc_pin(struct drm_i915_private *dev_priv, > > +static void sanitize_ddc_pin(struct drm_i915_private *i915, > > enum port port) > > { > > - struct ddi_vbt_port_info *info = &dev_priv->vbt.ddi_port_info[port]; > > + struct ddi_vbt_port_info *info = &i915->vbt.ddi_port_info[port]; > > enum port p; > > if (!info->alternate_ddc_pin) > > return; > > - p = get_port_by_ddc_pin(dev_priv, info->alternate_ddc_pin); > > + p = get_port_by_ddc_pin(i915, info->alternate_ddc_pin); > > if (p != PORT_NONE) { > > DRM_DEBUG_KMS("port %c trying to use the same DDC pin (0x%x) as port %c, " > > "disabling port %c DVI/HDMI support\n", > > @@ -1300,16 +1300,16 @@ static enum port get_port_by_aux_ch(struct drm_i915_private *i915, u8 aux_ch) > > return PORT_NONE; > > } > > -static void sanitize_aux_ch(struct drm_i915_private *dev_priv, > > +static void sanitize_aux_ch(struct drm_i915_private *i915, > > enum port port) > > { > > - struct ddi_vbt_port_info *info = &dev_priv->vbt.ddi_port_info[port]; > > + struct ddi_vbt_port_info *info = &i915->vbt.ddi_port_info[port]; > > enum port p; > > if (!info->alternate_aux_channel) > > return; > > - p = get_port_by_aux_ch(dev_priv, info->alternate_aux_channel); > > + p = get_port_by_aux_ch(i915, info->alternate_aux_channel); > > if (p != PORT_NONE) { > > DRM_DEBUG_KMS("port %c trying to use the same AUX CH (0x%x) as port %c, " > > "disabling port %c DP support\n", > > @@ -1347,15 +1347,15 @@ static const u8 icp_ddc_pin_map[] = { > > [ICL_DDC_BUS_PORT_4] = GMBUS_PIN_12_TC4_ICP, > > }; > > -static u8 map_ddc_pin(struct drm_i915_private *dev_priv, u8 vbt_pin) > > +static u8 map_ddc_pin(struct drm_i915_private *i915, u8 vbt_pin) > > { > > const u8 *ddc_pin_map; > > int n_entries; > > - if (HAS_PCH_ICP(dev_priv)) { > > + if (HAS_PCH_ICP(i915)) { > > ddc_pin_map = icp_ddc_pin_map; > > n_entries = ARRAY_SIZE(icp_ddc_pin_map); > > - } else if (HAS_PCH_CNP(dev_priv)) { > > + } else if (HAS_PCH_CNP(i915)) { > > ddc_pin_map = cnp_ddc_pin_map; > > n_entries = ARRAY_SIZE(cnp_ddc_pin_map); > > } else { > > @@ -1401,7 +1401,7 @@ static enum port dvo_port_to_port(u8 dvo_port) > > return PORT_NONE; > > } > > -static void parse_ddi_port(struct drm_i915_private *dev_priv, > > +static void parse_ddi_port(struct drm_i915_private *i915, > > const struct child_device_config *child, > > u8 bdb_version) > > { > > @@ -1413,7 +1413,7 @@ static void parse_ddi_port(struct drm_i915_private *dev_priv, > > if (port == PORT_NONE) > > return; > > - info = &dev_priv->vbt.ddi_port_info[port]; > > + info = &i915->vbt.ddi_port_info[port]; > > if (info->child) { > > DRM_DEBUG_KMS("More than one child device for port %c in VBT, using the first.\n", > > @@ -1447,7 +1447,7 @@ static void parse_ddi_port(struct drm_i915_private *dev_priv, > > DRM_DEBUG_KMS("Port %c VBT info: CRT:%d DVI:%d HDMI:%d DP:%d eDP:%d LSPCON:%d USB-Type-C:%d TBT:%d\n", > > port_name(port), is_crt, is_dvi, is_hdmi, is_dp, is_edp, > > - HAS_LSPCON(dev_priv) && child->lspcon, > > + HAS_LSPCON(i915) && child->lspcon, > > info->supports_typec_usb, info->supports_tbt); > > if (is_edp && is_dvi) > > @@ -1469,10 +1469,10 @@ static void parse_ddi_port(struct drm_i915_private *dev_priv, > > if (is_dvi) { > > u8 ddc_pin; > > - ddc_pin = map_ddc_pin(dev_priv, child->ddc_pin); > > - if (intel_gmbus_is_valid_pin(dev_priv, ddc_pin)) { > > + ddc_pin = map_ddc_pin(i915, child->ddc_pin); > > + if (intel_gmbus_is_valid_pin(i915, ddc_pin)) { > > info->alternate_ddc_pin = ddc_pin; > > - sanitize_ddc_pin(dev_priv, port); > > + sanitize_ddc_pin(i915, port); > > } else { > > DRM_DEBUG_KMS("Port %c has invalid DDC pin %d, " > > "sticking to defaults\n", > > @@ -1483,7 +1483,7 @@ static void parse_ddi_port(struct drm_i915_private *dev_priv, > > if (is_dp) { > > info->alternate_aux_channel = child->aux_channel; > > - sanitize_aux_ch(dev_priv, port); > > + sanitize_aux_ch(i915, port); > > } > > if (bdb_version >= 158) { > > @@ -1553,26 +1553,26 @@ static void parse_ddi_port(struct drm_i915_private *dev_priv, > > info->child = child; > > } > > -static void parse_ddi_ports(struct drm_i915_private *dev_priv, u8 bdb_version) > > +static void parse_ddi_ports(struct drm_i915_private *i915, u8 bdb_version) > > { > > const struct child_device_config *child; > > int i; > > - if (!HAS_DDI(dev_priv) && !IS_CHERRYVIEW(dev_priv)) > > + if (!HAS_DDI(i915) && !IS_CHERRYVIEW(i915)) > > return; > > if (bdb_version < 155) > > return; > > - for (i = 0; i < dev_priv->vbt.child_dev_num; i++) { > > - child = dev_priv->vbt.child_dev + i; > > + for (i = 0; i < i915->vbt.child_dev_num; i++) { > > + child = i915->vbt.child_dev + i; > > - parse_ddi_port(dev_priv, child, bdb_version); > > + parse_ddi_port(i915, child, bdb_version); > > } > > } > > static void > > -parse_general_definitions(struct drm_i915_private *dev_priv, > > +parse_general_definitions(struct drm_i915_private *i915, > > const struct bdb_header *bdb) > > { > > const struct bdb_general_definitions *defs; > > @@ -1597,8 +1597,8 @@ parse_general_definitions(struct drm_i915_private *dev_priv, > > bus_pin = defs->crt_ddc_gmbus_pin; > > DRM_DEBUG_KMS("crt_ddc_bus_pin: %d\n", bus_pin); > > - if (intel_gmbus_is_valid_pin(dev_priv, bus_pin)) > > - dev_priv->vbt.crt_ddc_pin = bus_pin; > > + if (intel_gmbus_is_valid_pin(i915, bus_pin)) > > + i915->vbt.crt_ddc_pin = bus_pin; > > if (bdb->version < 106) { > > expected_size = 22; > > @@ -1645,13 +1645,13 @@ parse_general_definitions(struct drm_i915_private *dev_priv, > > DRM_DEBUG_KMS("no child dev is parsed from VBT\n"); > > return; > > } > > - dev_priv->vbt.child_dev = kcalloc(count, sizeof(*child), GFP_KERNEL); > > - if (!dev_priv->vbt.child_dev) { > > + i915->vbt.child_dev = kcalloc(count, sizeof(*child), GFP_KERNEL); > > + if (!i915->vbt.child_dev) { > > DRM_DEBUG_KMS("No memory space for child device\n"); > > return; > > } > > - dev_priv->vbt.child_dev_num = count; > > + i915->vbt.child_dev_num = count; > > count = 0; > > for (i = 0; i < child_device_num; i++) { > > child = child_device_ptr(defs, i); > > @@ -1663,7 +1663,7 @@ parse_general_definitions(struct drm_i915_private *dev_priv, > > * (child_dev_size) of the child device. Accessing the data must > > * depend on VBT version. > > */ > > - memcpy(dev_priv->vbt.child_dev + count, child, > > + memcpy(i915->vbt.child_dev + count, child, > > min_t(size_t, defs->child_dev_size, sizeof(*child))); > > count++; > > } > > @@ -1671,41 +1671,41 @@ parse_general_definitions(struct drm_i915_private *dev_priv, > > /* Common defaults which may be overridden by VBT. */ > > static void > > -init_vbt_defaults(struct drm_i915_private *dev_priv) > > +init_vbt_defaults(struct drm_i915_private *i915) > > { > > enum port port; > > - dev_priv->vbt.crt_ddc_pin = GMBUS_PIN_VGADDC; > > + i915->vbt.crt_ddc_pin = GMBUS_PIN_VGADDC; > > /* Default to having backlight */ > > - dev_priv->vbt.backlight.present = true; > > + i915->vbt.backlight.present = true; > > /* LFP panel data */ > > - dev_priv->vbt.lvds_dither = 1; > > + i915->vbt.lvds_dither = 1; > > /* SDVO panel data */ > > - dev_priv->vbt.sdvo_lvds_vbt_mode = NULL; > > + i915->vbt.sdvo_lvds_vbt_mode = NULL; > > /* general features */ > > - dev_priv->vbt.int_tv_support = 1; > > - dev_priv->vbt.int_crt_support = 1; > > + i915->vbt.int_tv_support = 1; > > + i915->vbt.int_crt_support = 1; > > /* driver features */ > > - dev_priv->vbt.int_lvds_support = 1; > > + i915->vbt.int_lvds_support = 1; > > /* Default to using SSC */ > > - dev_priv->vbt.lvds_use_ssc = 1; > > + i915->vbt.lvds_use_ssc = 1; > > /* > > * Core/SandyBridge/IvyBridge use alternative (120MHz) reference > > * clock for LVDS. > > */ > > - dev_priv->vbt.lvds_ssc_freq = intel_bios_ssc_frequency(dev_priv, > > - !HAS_PCH_SPLIT(dev_priv)); > > - DRM_DEBUG_KMS("Set default to SSC at %d kHz\n", dev_priv->vbt.lvds_ssc_freq); > > + i915->vbt.lvds_ssc_freq = intel_bios_ssc_frequency(i915, > > + !HAS_PCH_SPLIT(i915)); > > + DRM_DEBUG_KMS("Set default to SSC at %d kHz\n", i915->vbt.lvds_ssc_freq); > > for (port = PORT_A; port < I915_MAX_PORTS; port++) { > > struct ddi_vbt_port_info *info = > > - &dev_priv->vbt.ddi_port_info[port]; > > + &i915->vbt.ddi_port_info[port]; > > info->hdmi_level_shift = HDMI_LEVEL_SHIFT_UNKNOWN; > > } > > @@ -1713,19 +1713,19 @@ init_vbt_defaults(struct drm_i915_private *dev_priv) > > /* Defaults to initialize only if there is no VBT. */ > > static void > > -init_vbt_missing_defaults(struct drm_i915_private *dev_priv) > > +init_vbt_missing_defaults(struct drm_i915_private *i915) > > { > > enum port port; > > for (port = PORT_A; port < I915_MAX_PORTS; port++) { > > struct ddi_vbt_port_info *info = > > - &dev_priv->vbt.ddi_port_info[port]; > > + &i915->vbt.ddi_port_info[port]; > > /* > > * VBT has the TypeC mode (native,TBT/USB) and we don't want > > * to detect it. > > */ > > - if (intel_port_is_tc(dev_priv, port)) > > + if (intel_port_is_tc(i915, port)) > > continue; > > info->supports_dvi = (port != PORT_A && port != PORT_E); > > @@ -1811,25 +1811,25 @@ static const struct vbt_header *find_vbt(void __iomem *bios, size_t size) > > /** > > * intel_bios_init - find VBT and initialize settings from the BIOS > > - * @dev_priv: i915 device instance > > + * @i915: i915 device instance > > * > > * Parse and initialize settings from the Video BIOS Tables (VBT). If the VBT > > * was not found in ACPI OpRegion, try to find it in PCI ROM first. Also > > * initialize some defaults if the VBT is not present at all. > > */ > > -void intel_bios_init(struct drm_i915_private *dev_priv) > > +void intel_bios_init(struct drm_i915_private *i915) > > { > > - struct pci_dev *pdev = dev_priv->drm.pdev; > > - const struct vbt_header *vbt = dev_priv->opregion.vbt; > > + struct pci_dev *pdev = i915->drm.pdev; > > + const struct vbt_header *vbt = i915->opregion.vbt; > > const struct bdb_header *bdb; > > u8 __iomem *bios = NULL; > > - if (!HAS_DISPLAY(dev_priv)) { > > + if (!HAS_DISPLAY(i915)) { > > DRM_DEBUG_KMS("Skipping VBT init due to disabled display.\n"); > > return; > > } > > - init_vbt_defaults(dev_priv); > > + init_vbt_defaults(i915); > > /* If the OpRegion does not have VBT, look in PCI ROM. */ > > if (!vbt) { > > @@ -1852,25 +1852,25 @@ void intel_bios_init(struct drm_i915_private *dev_priv) > > (int)sizeof(vbt->signature), vbt->signature, bdb->version); > > /* Grab useful general definitions */ > > - parse_general_features(dev_priv, bdb); > > - parse_general_definitions(dev_priv, bdb); > > - parse_lfp_panel_data(dev_priv, bdb); > > - parse_lfp_backlight(dev_priv, bdb); > > - parse_sdvo_panel_data(dev_priv, bdb); > > - parse_driver_features(dev_priv, bdb); > > - parse_edp(dev_priv, bdb); > > - parse_psr(dev_priv, bdb); > > - parse_mipi_config(dev_priv, bdb); > > - parse_mipi_sequence(dev_priv, bdb); > > + parse_general_features(i915, bdb); > > + parse_general_definitions(i915, bdb); > > + parse_lfp_panel_data(i915, bdb); > > + parse_lfp_backlight(i915, bdb); > > + parse_sdvo_panel_data(i915, bdb); > > + parse_driver_features(i915, bdb); > > + parse_edp(i915, bdb); > > + parse_psr(i915, bdb); > > + parse_mipi_config(i915, bdb); > > + parse_mipi_sequence(i915, bdb); > > /* Further processing on pre-parsed data */ > > - parse_sdvo_device_mapping(dev_priv, bdb->version); > > - parse_ddi_ports(dev_priv, bdb->version); > > + parse_sdvo_device_mapping(i915, bdb->version); > > + parse_ddi_ports(i915, bdb->version); > > out: > > if (!vbt) { > > DRM_INFO("Failed to find VBIOS tables (VBT)\n"); > > - init_vbt_missing_defaults(dev_priv); > > + init_vbt_missing_defaults(i915); > > } > > if (bios) > > @@ -1879,47 +1879,47 @@ void intel_bios_init(struct drm_i915_private *dev_priv) > > /** > > * intel_bios_cleanup - Free any resources allocated by intel_bios_init() > > - * @dev_priv: i915 device instance > > + * @i915: i915 device instance > > */ > > -void intel_bios_cleanup(struct drm_i915_private *dev_priv) > > +void intel_bios_cleanup(struct drm_i915_private *i915) > > { > > - kfree(dev_priv->vbt.child_dev); > > - dev_priv->vbt.child_dev = NULL; > > - dev_priv->vbt.child_dev_num = 0; > > - kfree(dev_priv->vbt.sdvo_lvds_vbt_mode); > > - dev_priv->vbt.sdvo_lvds_vbt_mode = NULL; > > - kfree(dev_priv->vbt.lfp_lvds_vbt_mode); > > - dev_priv->vbt.lfp_lvds_vbt_mode = NULL; > > - kfree(dev_priv->vbt.dsi.data); > > - dev_priv->vbt.dsi.data = NULL; > > - kfree(dev_priv->vbt.dsi.pps); > > - dev_priv->vbt.dsi.pps = NULL; > > - kfree(dev_priv->vbt.dsi.config); > > - dev_priv->vbt.dsi.config = NULL; > > - kfree(dev_priv->vbt.dsi.deassert_seq); > > - dev_priv->vbt.dsi.deassert_seq = NULL; > > + kfree(i915->vbt.child_dev); > > + i915->vbt.child_dev = NULL; > > + i915->vbt.child_dev_num = 0; > > + kfree(i915->vbt.sdvo_lvds_vbt_mode); > > + i915->vbt.sdvo_lvds_vbt_mode = NULL; > > + kfree(i915->vbt.lfp_lvds_vbt_mode); > > + i915->vbt.lfp_lvds_vbt_mode = NULL; > > + kfree(i915->vbt.dsi.data); > > + i915->vbt.dsi.data = NULL; > > + kfree(i915->vbt.dsi.pps); > > + i915->vbt.dsi.pps = NULL; > > + kfree(i915->vbt.dsi.config); > > + i915->vbt.dsi.config = NULL; > > + kfree(i915->vbt.dsi.deassert_seq); > > + i915->vbt.dsi.deassert_seq = NULL; > > } > > /** > > * intel_bios_is_tv_present - is integrated TV present in VBT > > - * @dev_priv: i915 device instance > > + * @i915: i915 device instance > > * > > * Return true if TV is present. If no child devices were parsed from VBT, > > * assume TV is present. > > */ > > -bool intel_bios_is_tv_present(struct drm_i915_private *dev_priv) > > +bool intel_bios_is_tv_present(struct drm_i915_private *i915) > > { > > const struct child_device_config *child; > > int i; > > - if (!dev_priv->vbt.int_tv_support) > > + if (!i915->vbt.int_tv_support) > > return false; > > - if (!dev_priv->vbt.child_dev_num) > > + if (!i915->vbt.child_dev_num) > > return true; > > - for (i = 0; i < dev_priv->vbt.child_dev_num; i++) { > > - child = dev_priv->vbt.child_dev + i; > > + for (i = 0; i < i915->vbt.child_dev_num; i++) { > > + child = i915->vbt.child_dev + i; > > /* > > * If the device type is not TV, continue. > > */ > > @@ -1943,22 +1943,22 @@ bool intel_bios_is_tv_present(struct drm_i915_private *dev_priv) > > /** > > * intel_bios_is_lvds_present - is LVDS present in VBT > > - * @dev_priv: i915 device instance > > + * @i915: i915 device instance > > * @i2c_pin: i2c pin for LVDS if present > > * > > * Return true if LVDS is present. If no child devices were parsed from VBT, > > * assume LVDS is present. > > */ > > -bool intel_bios_is_lvds_present(struct drm_i915_private *dev_priv, u8 *i2c_pin) > > +bool intel_bios_is_lvds_present(struct drm_i915_private *i915, u8 *i2c_pin) > > { > > const struct child_device_config *child; > > int i; > > - if (!dev_priv->vbt.child_dev_num) > > + if (!i915->vbt.child_dev_num) > > return true; > > - for (i = 0; i < dev_priv->vbt.child_dev_num; i++) { > > - child = dev_priv->vbt.child_dev + i; > > + for (i = 0; i < i915->vbt.child_dev_num; i++) { > > + child = i915->vbt.child_dev + i; > > /* If the device type is not LFP, continue. > > * We have to check both the new identifiers as well as the > > @@ -1968,7 +1968,7 @@ bool intel_bios_is_lvds_present(struct drm_i915_private *dev_priv, u8 *i2c_pin) > > child->device_type != DEVICE_TYPE_LFP) > > continue; > > - if (intel_gmbus_is_valid_pin(dev_priv, child->i2c_pin)) > > + if (intel_gmbus_is_valid_pin(i915, child->i2c_pin)) > > *i2c_pin = child->i2c_pin; > > /* However, we cannot trust the BIOS writers to populate > > @@ -1984,7 +1984,7 @@ bool intel_bios_is_lvds_present(struct drm_i915_private *dev_priv, u8 *i2c_pin) > > * additional data. Trust that if the VBT was written into > > * the OpRegion then they have validated the LVDS's existence. > > */ > > - if (dev_priv->opregion.vbt) > > + if (i915->opregion.vbt) > > return true; > > } > > @@ -1993,12 +1993,12 @@ bool intel_bios_is_lvds_present(struct drm_i915_private *dev_priv, u8 *i2c_pin) > > /** > > * intel_bios_is_port_present - is the specified digital port present > > - * @dev_priv: i915 device instance > > + * @i915: i915 device instance > > * @port: port to check > > * > > * Return true if the device in %port is present. > > */ > > -bool intel_bios_is_port_present(struct drm_i915_private *dev_priv, enum port port) > > +bool intel_bios_is_port_present(struct drm_i915_private *i915, enum port port) > > { > > const struct child_device_config *child; > > static const struct { > > @@ -2012,9 +2012,9 @@ bool intel_bios_is_port_present(struct drm_i915_private *dev_priv, enum port por > > }; > > int i; > > - if (HAS_DDI(dev_priv)) { > > + if (HAS_DDI(i915)) { > > const struct ddi_vbt_port_info *port_info = > > - &dev_priv->vbt.ddi_port_info[port]; > > + &i915->vbt.ddi_port_info[port]; > > return port_info->supports_dp || > > port_info->supports_dvi || > > @@ -2025,11 +2025,11 @@ bool intel_bios_is_port_present(struct drm_i915_private *dev_priv, enum port por > > if (WARN_ON(port == PORT_A) || port >= ARRAY_SIZE(port_mapping)) > > return false; > > - if (!dev_priv->vbt.child_dev_num) > > + if (!i915->vbt.child_dev_num) > > return false; > > - for (i = 0; i < dev_priv->vbt.child_dev_num; i++) { > > - child = dev_priv->vbt.child_dev + i; > > + for (i = 0; i < i915->vbt.child_dev_num; i++) { > > + child = i915->vbt.child_dev + i; > > if ((child->dvo_port == port_mapping[port].dp || > > child->dvo_port == port_mapping[port].hdmi) && > > @@ -2043,12 +2043,12 @@ bool intel_bios_is_port_present(struct drm_i915_private *dev_priv, enum port por > > /** > > * intel_bios_is_port_edp - is the device in given port eDP > > - * @dev_priv: i915 device instance > > + * @i915: i915 device instance > > * @port: port to check > > * > > * Return true if the device in %port is eDP. > > */ > > -bool intel_bios_is_port_edp(struct drm_i915_private *dev_priv, enum port port) > > +bool intel_bios_is_port_edp(struct drm_i915_private *i915, enum port port) > > { > > const struct child_device_config *child; > > static const short port_mapping[] = { > > @@ -2060,14 +2060,14 @@ bool intel_bios_is_port_edp(struct drm_i915_private *dev_priv, enum port port) > > }; > > int i; > > - if (HAS_DDI(dev_priv)) > > - return dev_priv->vbt.ddi_port_info[port].supports_edp; > > + if (HAS_DDI(i915)) > > + return i915->vbt.ddi_port_info[port].supports_edp; > > - if (!dev_priv->vbt.child_dev_num) > > + if (!i915->vbt.child_dev_num) > > return false; > > - for (i = 0; i < dev_priv->vbt.child_dev_num; i++) { > > - child = dev_priv->vbt.child_dev + i; > > + for (i = 0; i < i915->vbt.child_dev_num; i++) { > > + child = i915->vbt.child_dev + i; > > if (child->dvo_port == port_mapping[port] && > > (child->device_type & DEVICE_TYPE_eDP_BITS) == > > @@ -2113,14 +2113,14 @@ static bool child_dev_is_dp_dual_mode(const struct child_device_config *child, > > return false; > > } > > -bool intel_bios_is_port_dp_dual_mode(struct drm_i915_private *dev_priv, > > +bool intel_bios_is_port_dp_dual_mode(struct drm_i915_private *i915, > > enum port port) > > { > > const struct child_device_config *child; > > int i; > > - for (i = 0; i < dev_priv->vbt.child_dev_num; i++) { > > - child = dev_priv->vbt.child_dev + i; > > + for (i = 0; i < i915->vbt.child_dev_num; i++) { > > + child = i915->vbt.child_dev + i; > > if (child_dev_is_dp_dual_mode(child, port)) > > return true; > > @@ -2131,20 +2131,20 @@ bool intel_bios_is_port_dp_dual_mode(struct drm_i915_private *dev_priv, > > /** > > * intel_bios_is_dsi_present - is DSI present in VBT > > - * @dev_priv: i915 device instance > > + * @i915: i915 device instance > > * @port: port for DSI if present > > * > > * Return true if DSI is present, and return the port in %port. > > */ > > -bool intel_bios_is_dsi_present(struct drm_i915_private *dev_priv, > > +bool intel_bios_is_dsi_present(struct drm_i915_private *i915, > > enum port *port) > > { > > const struct child_device_config *child; > > u8 dvo_port; > > int i; > > - for (i = 0; i < dev_priv->vbt.child_dev_num; i++) { > > - child = dev_priv->vbt.child_dev + i; > > + for (i = 0; i < i915->vbt.child_dev_num; i++) { > > + child = i915->vbt.child_dev + i; > > if (!(child->device_type & DEVICE_TYPE_MIPI_OUTPUT)) > > continue; > > @@ -2152,8 +2152,8 @@ bool intel_bios_is_dsi_present(struct drm_i915_private *dev_priv, > > dvo_port = child->dvo_port; > > if (dvo_port == DVO_PORT_MIPIA || > > - (dvo_port == DVO_PORT_MIPIB && INTEL_GEN(dev_priv) >= 11) || > > - (dvo_port == DVO_PORT_MIPIC && INTEL_GEN(dev_priv) < 11)) { > > + (dvo_port == DVO_PORT_MIPIB && INTEL_GEN(i915) >= 11) || > > + (dvo_port == DVO_PORT_MIPIC && INTEL_GEN(i915) < 11)) { > > if (port) > > *port = dvo_port - DVO_PORT_MIPIA; > > return true; > > @@ -2205,11 +2205,11 @@ intel_bios_is_lspcon_present(const struct drm_i915_private *i915, > > return HAS_LSPCON(i915) && child && child->lspcon; > > } > > -enum aux_ch intel_bios_port_aux_ch(struct drm_i915_private *dev_priv, > > +enum aux_ch intel_bios_port_aux_ch(struct drm_i915_private *i915, > > enum port port) > > { > > const struct ddi_vbt_port_info *info = > > - &dev_priv->vbt.ddi_port_info[port]; > > + &i915->vbt.ddi_port_info[port]; > > enum aux_ch aux_ch; > > if (!info->alternate_aux_channel) { > > diff --git a/drivers/gpu/drm/i915/intel_bios.h b/drivers/gpu/drm/i915/intel_bios.h > > index 4e42cfaf61a7..d292bd1790ae 100644 > > --- a/drivers/gpu/drm/i915/intel_bios.h > > +++ b/drivers/gpu/drm/i915/intel_bios.h > > @@ -226,19 +226,19 @@ struct mipi_pps_data { > > u16 panel_power_cycle_delay; > > } __packed; > > -void intel_bios_init(struct drm_i915_private *dev_priv); > > -void intel_bios_cleanup(struct drm_i915_private *dev_priv); > > +void intel_bios_init(struct drm_i915_private *i915); > > +void intel_bios_cleanup(struct drm_i915_private *i915); > > bool intel_bios_is_valid_vbt(const void *buf, size_t size); > > -bool intel_bios_is_tv_present(struct drm_i915_private *dev_priv); > > -bool intel_bios_is_lvds_present(struct drm_i915_private *dev_priv, u8 *i2c_pin); > > -bool intel_bios_is_port_present(struct drm_i915_private *dev_priv, enum port port); > > -bool intel_bios_is_port_edp(struct drm_i915_private *dev_priv, enum port port); > > -bool intel_bios_is_port_dp_dual_mode(struct drm_i915_private *dev_priv, enum port port); > > -bool intel_bios_is_dsi_present(struct drm_i915_private *dev_priv, enum port *port); > > +bool intel_bios_is_tv_present(struct drm_i915_private *i915); > > +bool intel_bios_is_lvds_present(struct drm_i915_private *i915, u8 *i2c_pin); > > +bool intel_bios_is_port_present(struct drm_i915_private *i915, enum port port); > > +bool intel_bios_is_port_edp(struct drm_i915_private *i915, enum port port); > > +bool intel_bios_is_port_dp_dual_mode(struct drm_i915_private *i915, enum port port); > > +bool intel_bios_is_dsi_present(struct drm_i915_private *i915, enum port *port); > > bool intel_bios_is_port_hpd_inverted(const struct drm_i915_private *i915, > > enum port port); > > bool intel_bios_is_lspcon_present(const struct drm_i915_private *i915, > > enum port port); > > -enum aux_ch intel_bios_port_aux_ch(struct drm_i915_private *dev_priv, enum port port); > > +enum aux_ch intel_bios_port_aux_ch(struct drm_i915_private *i915, enum port port); > > #endif /* _INTEL_BIOS_H_ */ > > diff --git a/drivers/gpu/drm/i915/intel_bw.c b/drivers/gpu/drm/i915/intel_bw.c > > index 753ac3165061..1aca44a258d3 100644 > > --- a/drivers/gpu/drm/i915/intel_bw.c > > +++ b/drivers/gpu/drm/i915/intel_bw.c > > @@ -22,13 +22,13 @@ struct intel_qgv_info { > > enum intel_dram_type dram_type; > > }; > > -static int icl_pcode_read_mem_global_info(struct drm_i915_private *dev_priv, > > +static int icl_pcode_read_mem_global_info(struct drm_i915_private *i915, > > struct intel_qgv_info *qi) > > { > > u32 val = 0; > > int ret; > > - ret = sandybridge_pcode_read(dev_priv, > > + ret = sandybridge_pcode_read(i915, > > ICL_PCODE_MEM_SUBSYSYSTEM_INFO | > > ICL_PCODE_MEM_SS_READ_GLOBAL_INFO, > > &val, NULL); > > @@ -61,14 +61,14 @@ static int icl_pcode_read_mem_global_info(struct drm_i915_private *dev_priv, > > return 0; > > } > > -static int icl_pcode_read_qgv_point_info(struct drm_i915_private *dev_priv, > > +static int icl_pcode_read_qgv_point_info(struct drm_i915_private *i915, > > struct intel_qgv_point *sp, > > int point) > > { > > u32 val = 0, val2; > > int ret; > > - ret = sandybridge_pcode_read(dev_priv, > > + ret = sandybridge_pcode_read(i915, > > ICL_PCODE_MEM_SUBSYSYSTEM_INFO | > > ICL_PCODE_MEM_SS_READ_QGV_POINT_INFO(point), > > &val, &val2); > > @@ -87,12 +87,12 @@ static int icl_pcode_read_qgv_point_info(struct drm_i915_private *dev_priv, > > return 0; > > } > > -static int icl_get_qgv_points(struct drm_i915_private *dev_priv, > > +static int icl_get_qgv_points(struct drm_i915_private *i915, > > struct intel_qgv_info *qi) > > { > > int i, ret; > > - ret = icl_pcode_read_mem_global_info(dev_priv, qi); > > + ret = icl_pcode_read_mem_global_info(i915, qi); > > if (ret) > > return ret; > > @@ -102,7 +102,7 @@ static int icl_get_qgv_points(struct drm_i915_private *dev_priv, > > for (i = 0; i < qi->num_points; i++) { > > struct intel_qgv_point *sp = &qi->points[i]; > > - ret = icl_pcode_read_qgv_point_info(dev_priv, sp, i); > > + ret = icl_pcode_read_qgv_point_info(i915, sp, i); > > if (ret) > > return ret; > > @@ -142,7 +142,7 @@ static const struct intel_sa_info icl_sa_info = { > > .displayrtids = 128, > > }; > > -static int icl_get_bw_info(struct drm_i915_private *dev_priv) > > +static int icl_get_bw_info(struct drm_i915_private *i915) > > { > > struct intel_qgv_info qi = {}; > > const struct intel_sa_info *sa = &icl_sa_info; > > @@ -154,7 +154,7 @@ static int icl_get_bw_info(struct drm_i915_private *dev_priv) > > int maxdebw; > > int i, ret; > > - ret = icl_get_qgv_points(dev_priv, &qi); > > + ret = icl_get_qgv_points(i915, &qi); > > if (ret) { > > DRM_DEBUG_KMS("Failed to get memory subsystem information, ignoring bandwidth limits"); > > return ret; > > @@ -170,8 +170,8 @@ static int icl_get_bw_info(struct drm_i915_private *dev_priv) > > icl_calc_bw(dclk_max, 16, 1) * 6 / 10); /* 60% */ > > ipqdepth = min(ipqdepthpch, sa->displayrtids / num_channels); > > - for (i = 0; i < ARRAY_SIZE(dev_priv->max_bw); i++) { > > - struct intel_bw_info *bi = &dev_priv->max_bw[i]; > > + for (i = 0; i < ARRAY_SIZE(i915->max_bw); i++) { > > + struct intel_bw_info *bi = &i915->max_bw[i]; > > int clpchgroup; > > int j; > > @@ -206,18 +206,18 @@ static int icl_get_bw_info(struct drm_i915_private *dev_priv) > > return 0; > > } > > -static unsigned int icl_max_bw(struct drm_i915_private *dev_priv, > > +static unsigned int icl_max_bw(struct drm_i915_private *i915, > > int num_planes, int qgv_point) > > { > > int i; > > /* Did we initialize the bw limits successfully? */ > > - if (dev_priv->max_bw[0].num_planes == 0) > > + if (i915->max_bw[0].num_planes == 0) > > return UINT_MAX; > > - for (i = 0; i < ARRAY_SIZE(dev_priv->max_bw); i++) { > > + for (i = 0; i < ARRAY_SIZE(i915->max_bw); i++) { > > const struct intel_bw_info *bi = > > - &dev_priv->max_bw[i]; > > + &i915->max_bw[i]; > > if (num_planes >= bi->num_planes) > > return bi->deratedbw[qgv_point]; > > @@ -226,24 +226,24 @@ static unsigned int icl_max_bw(struct drm_i915_private *dev_priv, > > return 0; > > } > > -void intel_bw_init_hw(struct drm_i915_private *dev_priv) > > +void intel_bw_init_hw(struct drm_i915_private *i915) > > { > > - if (IS_GEN(dev_priv, 11)) > > - icl_get_bw_info(dev_priv); > > + if (IS_GEN(i915, 11)) > > + icl_get_bw_info(i915); > > } > > -static unsigned int intel_max_data_rate(struct drm_i915_private *dev_priv, > > +static unsigned int intel_max_data_rate(struct drm_i915_private *i915, > > int num_planes) > > { > > - if (IS_GEN(dev_priv, 11)) > > + if (IS_GEN(i915, 11)) > > /* > > * FIXME with SAGV disabled maybe we can assume > > * point 1 will always be used? Seems to match > > * the behaviour observed in the wild. > > */ > > - return min3(icl_max_bw(dev_priv, num_planes, 0), > > - icl_max_bw(dev_priv, num_planes, 1), > > - icl_max_bw(dev_priv, num_planes, 2)); > > + return min3(icl_max_bw(i915, num_planes, 0), > > + icl_max_bw(i915, num_planes, 1), > > + icl_max_bw(i915, num_planes, 2)); > > else > > return UINT_MAX; > > } > > @@ -293,25 +293,25 @@ void intel_bw_crtc_update(struct intel_bw_state *bw_state, > > bw_state->num_active_planes[crtc->pipe]); > > } > > -static unsigned int intel_bw_num_active_planes(struct drm_i915_private *dev_priv, > > +static unsigned int intel_bw_num_active_planes(struct drm_i915_private *i915, > > const struct intel_bw_state *bw_state) > > { > > unsigned int num_active_planes = 0; > > enum pipe pipe; > > - for_each_pipe(dev_priv, pipe) > > + for_each_pipe(i915, pipe) > > num_active_planes += bw_state->num_active_planes[pipe]; > > return num_active_planes; > > } > > -static unsigned int intel_bw_data_rate(struct drm_i915_private *dev_priv, > > +static unsigned int intel_bw_data_rate(struct drm_i915_private *i915, > > const struct intel_bw_state *bw_state) > > { > > unsigned int data_rate = 0; > > enum pipe pipe; > > - for_each_pipe(dev_priv, pipe) > > + for_each_pipe(i915, pipe) > > data_rate += bw_state->data_rate[pipe]; > > return data_rate; > > @@ -319,7 +319,7 @@ static unsigned int intel_bw_data_rate(struct drm_i915_private *dev_priv, > > int intel_bw_atomic_check(struct intel_atomic_state *state) > > { > > - struct drm_i915_private *dev_priv = to_i915(state->base.dev); > > + struct drm_i915_private *i915 = to_i915(state->base.dev); > > struct intel_crtc_state *new_crtc_state, *old_crtc_state; > > struct intel_bw_state *bw_state = NULL; > > unsigned int data_rate, max_data_rate; > > @@ -328,7 +328,7 @@ int intel_bw_atomic_check(struct intel_atomic_state *state) > > int i; > > /* FIXME earlier gens need some checks too */ > > - if (INTEL_GEN(dev_priv) < 11) > > + if (INTEL_GEN(i915) < 11) > > return 0; > > for_each_oldnew_intel_crtc_in_state(state, crtc, old_crtc_state, > > @@ -366,10 +366,10 @@ int intel_bw_atomic_check(struct intel_atomic_state *state) > > if (!bw_state) > > return 0; > > - data_rate = intel_bw_data_rate(dev_priv, bw_state); > > - num_active_planes = intel_bw_num_active_planes(dev_priv, bw_state); > > + data_rate = intel_bw_data_rate(i915, bw_state); > > + num_active_planes = intel_bw_num_active_planes(i915, bw_state); > > - max_data_rate = intel_max_data_rate(dev_priv, num_active_planes); > > + max_data_rate = intel_max_data_rate(i915, num_active_planes); > > data_rate = DIV_ROUND_UP(data_rate, 1000); > > @@ -406,7 +406,7 @@ static const struct drm_private_state_funcs intel_bw_funcs = { > > .atomic_destroy_state = intel_bw_destroy_state, > > }; > > -int intel_bw_init(struct drm_i915_private *dev_priv) > > +int intel_bw_init(struct drm_i915_private *i915) > > { > > struct intel_bw_state *state; > > @@ -414,7 +414,7 @@ int intel_bw_init(struct drm_i915_private *dev_priv) > > if (!state) > > return -ENOMEM; > > - drm_atomic_private_obj_init(&dev_priv->drm, &dev_priv->bw_obj, > > + drm_atomic_private_obj_init(&i915->drm, &i915->bw_obj, > > &state->base, &intel_bw_funcs); > > return 0; > > diff --git a/drivers/gpu/drm/i915/intel_bw.h b/drivers/gpu/drm/i915/intel_bw.h > > index e9d9c6d63bc3..396e5029f9a4 100644 > > --- a/drivers/gpu/drm/i915/intel_bw.h > > +++ b/drivers/gpu/drm/i915/intel_bw.h > > @@ -27,19 +27,19 @@ struct intel_bw_state { > > static inline struct intel_bw_state * > > intel_atomic_get_bw_state(struct intel_atomic_state *state) > > { > > - struct drm_i915_private *dev_priv = to_i915(state->base.dev); > > + struct drm_i915_private *i915 = to_i915(state->base.dev); > > struct drm_private_state *bw_state; > > bw_state = drm_atomic_get_private_obj_state(&state->base, > > - &dev_priv->bw_obj); > > + &i915->bw_obj); > > if (IS_ERR(bw_state)) > > return ERR_CAST(bw_state); > > return to_intel_bw_state(bw_state); > > } > > -void intel_bw_init_hw(struct drm_i915_private *dev_priv); > > -int intel_bw_init(struct drm_i915_private *dev_priv); > > +void intel_bw_init_hw(struct drm_i915_private *i915); > > +int intel_bw_init(struct drm_i915_private *i915); > > int intel_bw_atomic_check(struct intel_atomic_state *state); > > void intel_bw_crtc_update(struct intel_bw_state *bw_state, > > const struct intel_crtc_state *crtc_state); > > diff --git a/drivers/gpu/drm/i915/intel_cdclk.c b/drivers/gpu/drm/i915/intel_cdclk.c > > index 8993ab283562..02c617180c4b 100644 > > --- a/drivers/gpu/drm/i915/intel_cdclk.c > > +++ b/drivers/gpu/drm/i915/intel_cdclk.c > > @@ -53,46 +53,46 @@ > > * dividers can be programmed correctly. > > */ > > -static void fixed_133mhz_get_cdclk(struct drm_i915_private *dev_priv, > > +static void fixed_133mhz_get_cdclk(struct drm_i915_private *i915, > > struct intel_cdclk_state *cdclk_state) > > { > > cdclk_state->cdclk = 133333; > > } > > -static void fixed_200mhz_get_cdclk(struct drm_i915_private *dev_priv, > > +static void fixed_200mhz_get_cdclk(struct drm_i915_private *i915, > > struct intel_cdclk_state *cdclk_state) > > { > > cdclk_state->cdclk = 200000; > > } > > -static void fixed_266mhz_get_cdclk(struct drm_i915_private *dev_priv, > > +static void fixed_266mhz_get_cdclk(struct drm_i915_private *i915, > > struct intel_cdclk_state *cdclk_state) > > { > > cdclk_state->cdclk = 266667; > > } > > -static void fixed_333mhz_get_cdclk(struct drm_i915_private *dev_priv, > > +static void fixed_333mhz_get_cdclk(struct drm_i915_private *i915, > > struct intel_cdclk_state *cdclk_state) > > { > > cdclk_state->cdclk = 333333; > > } > > -static void fixed_400mhz_get_cdclk(struct drm_i915_private *dev_priv, > > +static void fixed_400mhz_get_cdclk(struct drm_i915_private *i915, > > struct intel_cdclk_state *cdclk_state) > > { > > cdclk_state->cdclk = 400000; > > } > > -static void fixed_450mhz_get_cdclk(struct drm_i915_private *dev_priv, > > +static void fixed_450mhz_get_cdclk(struct drm_i915_private *i915, > > struct intel_cdclk_state *cdclk_state) > > { > > cdclk_state->cdclk = 450000; > > } > > -static void i85x_get_cdclk(struct drm_i915_private *dev_priv, > > +static void i85x_get_cdclk(struct drm_i915_private *i915, > > struct intel_cdclk_state *cdclk_state) > > { > > - struct pci_dev *pdev = dev_priv->drm.pdev; > > + struct pci_dev *pdev = i915->drm.pdev; > > u16 hpllcc = 0; > > /* > > @@ -131,10 +131,10 @@ static void i85x_get_cdclk(struct drm_i915_private *dev_priv, > > } > > } > > -static void i915gm_get_cdclk(struct drm_i915_private *dev_priv, > > +static void i915gm_get_cdclk(struct drm_i915_private *i915, > > struct intel_cdclk_state *cdclk_state) > > { > > - struct pci_dev *pdev = dev_priv->drm.pdev; > > + struct pci_dev *pdev = i915->drm.pdev; > > u16 gcfgc = 0; > > pci_read_config_word(pdev, GCFGC, &gcfgc); > > @@ -155,10 +155,10 @@ static void i915gm_get_cdclk(struct drm_i915_private *dev_priv, > > } > > } > > -static void i945gm_get_cdclk(struct drm_i915_private *dev_priv, > > +static void i945gm_get_cdclk(struct drm_i915_private *i915, > > struct intel_cdclk_state *cdclk_state) > > { > > - struct pci_dev *pdev = dev_priv->drm.pdev; > > + struct pci_dev *pdev = i915->drm.pdev; > > u16 gcfgc = 0; > > pci_read_config_word(pdev, GCFGC, &gcfgc); > > @@ -179,7 +179,7 @@ static void i945gm_get_cdclk(struct drm_i915_private *dev_priv, > > } > > } > > -static unsigned int intel_hpll_vco(struct drm_i915_private *dev_priv) > > +static unsigned int intel_hpll_vco(struct drm_i915_private *i915) > > { > > static const unsigned int blb_vco[8] = { > > [0] = 3200000, > > @@ -223,20 +223,20 @@ static unsigned int intel_hpll_vco(struct drm_i915_private *dev_priv) > > u8 tmp = 0; > > /* FIXME other chipsets? */ > > - if (IS_GM45(dev_priv)) > > + if (IS_GM45(i915)) > > vco_table = ctg_vco; > > - else if (IS_G45(dev_priv)) > > + else if (IS_G45(i915)) > > vco_table = elk_vco; > > - else if (IS_I965GM(dev_priv)) > > + else if (IS_I965GM(i915)) > > vco_table = cl_vco; > > - else if (IS_PINEVIEW(dev_priv)) > > + else if (IS_PINEVIEW(i915)) > > vco_table = pnv_vco; > > - else if (IS_G33(dev_priv)) > > + else if (IS_G33(i915)) > > vco_table = blb_vco; > > else > > return 0; > > - tmp = I915_READ(IS_PINEVIEW(dev_priv) || IS_MOBILE(dev_priv) ? > > + tmp = I915_READ(IS_PINEVIEW(i915) || IS_MOBILE(i915) ? > > HPLLVCO_MOBILE : HPLLVCO); > > vco = vco_table[tmp & 0x7]; > > @@ -248,10 +248,10 @@ static unsigned int intel_hpll_vco(struct drm_i915_private *dev_priv) > > return vco; > > } > > -static void g33_get_cdclk(struct drm_i915_private *dev_priv, > > +static void g33_get_cdclk(struct drm_i915_private *i915, > > struct intel_cdclk_state *cdclk_state) > > { > > - struct pci_dev *pdev = dev_priv->drm.pdev; > > + struct pci_dev *pdev = i915->drm.pdev; > > static const u8 div_3200[] = { 12, 10, 8, 7, 5, 16 }; > > static const u8 div_4000[] = { 14, 12, 10, 8, 6, 20 }; > > static const u8 div_4800[] = { 20, 14, 12, 10, 8, 24 }; > > @@ -260,7 +260,7 @@ static void g33_get_cdclk(struct drm_i915_private *dev_priv, > > unsigned int cdclk_sel; > > u16 tmp = 0; > > - cdclk_state->vco = intel_hpll_vco(dev_priv); > > + cdclk_state->vco = intel_hpll_vco(i915); > > pci_read_config_word(pdev, GCFGC, &tmp); > > @@ -296,10 +296,10 @@ static void g33_get_cdclk(struct drm_i915_private *dev_priv, > > cdclk_state->cdclk = 190476; > > } > > -static void pnv_get_cdclk(struct drm_i915_private *dev_priv, > > +static void pnv_get_cdclk(struct drm_i915_private *i915, > > struct intel_cdclk_state *cdclk_state) > > { > > - struct pci_dev *pdev = dev_priv->drm.pdev; > > + struct pci_dev *pdev = i915->drm.pdev; > > u16 gcfgc = 0; > > pci_read_config_word(pdev, GCFGC, &gcfgc); > > @@ -329,10 +329,10 @@ static void pnv_get_cdclk(struct drm_i915_private *dev_priv, > > } > > } > > -static void i965gm_get_cdclk(struct drm_i915_private *dev_priv, > > +static void i965gm_get_cdclk(struct drm_i915_private *i915, > > struct intel_cdclk_state *cdclk_state) > > { > > - struct pci_dev *pdev = dev_priv->drm.pdev; > > + struct pci_dev *pdev = i915->drm.pdev; > > static const u8 div_3200[] = { 16, 10, 8 }; > > static const u8 div_4000[] = { 20, 12, 10 }; > > static const u8 div_5333[] = { 24, 16, 14 }; > > @@ -340,7 +340,7 @@ static void i965gm_get_cdclk(struct drm_i915_private *dev_priv, > > unsigned int cdclk_sel; > > u16 tmp = 0; > > - cdclk_state->vco = intel_hpll_vco(dev_priv); > > + cdclk_state->vco = intel_hpll_vco(i915); > > pci_read_config_word(pdev, GCFGC, &tmp); > > @@ -373,14 +373,14 @@ static void i965gm_get_cdclk(struct drm_i915_private *dev_priv, > > cdclk_state->cdclk = 200000; > > } > > -static void gm45_get_cdclk(struct drm_i915_private *dev_priv, > > +static void gm45_get_cdclk(struct drm_i915_private *i915, > > struct intel_cdclk_state *cdclk_state) > > { > > - struct pci_dev *pdev = dev_priv->drm.pdev; > > + struct pci_dev *pdev = i915->drm.pdev; > > unsigned int cdclk_sel; > > u16 tmp = 0; > > - cdclk_state->vco = intel_hpll_vco(dev_priv); > > + cdclk_state->vco = intel_hpll_vco(i915); > > pci_read_config_word(pdev, GCFGC, &tmp); > > @@ -403,7 +403,7 @@ static void gm45_get_cdclk(struct drm_i915_private *dev_priv, > > } > > } > > -static void hsw_get_cdclk(struct drm_i915_private *dev_priv, > > +static void hsw_get_cdclk(struct drm_i915_private *i915, > > struct intel_cdclk_state *cdclk_state) > > { > > u32 lcpll = I915_READ(LCPLL_CTL); > > @@ -415,15 +415,15 @@ static void hsw_get_cdclk(struct drm_i915_private *dev_priv, > > cdclk_state->cdclk = 450000; > > else if (freq == LCPLL_CLK_FREQ_450) > > cdclk_state->cdclk = 450000; > > - else if (IS_HSW_ULT(dev_priv)) > > + else if (IS_HSW_ULT(i915)) > > cdclk_state->cdclk = 337500; > > else > > cdclk_state->cdclk = 540000; > > } > > -static int vlv_calc_cdclk(struct drm_i915_private *dev_priv, int min_cdclk) > > +static int vlv_calc_cdclk(struct drm_i915_private *i915, int min_cdclk) > > { > > - int freq_320 = (dev_priv->hpll_freq << 1) % 320000 != 0 ? > > + int freq_320 = (i915->hpll_freq << 1) % 320000 != 0 ? > > 333333 : 320000; > > /* > > @@ -431,7 +431,7 @@ static int vlv_calc_cdclk(struct drm_i915_private *dev_priv, int min_cdclk) > > * Not sure what's wrong. For now use 200MHz only when all pipes > > * are off. > > */ > > - if (IS_VALLEYVIEW(dev_priv) && min_cdclk > freq_320) > > + if (IS_VALLEYVIEW(i915) && min_cdclk > freq_320) > > return 400000; > > else if (min_cdclk > 266667) > > return freq_320; > > @@ -441,9 +441,9 @@ static int vlv_calc_cdclk(struct drm_i915_private *dev_priv, int min_cdclk) > > return 200000; > > } > > -static u8 vlv_calc_voltage_level(struct drm_i915_private *dev_priv, int cdclk) > > +static u8 vlv_calc_voltage_level(struct drm_i915_private *i915, int cdclk) > > { > > - if (IS_VALLEYVIEW(dev_priv)) { > > + if (IS_VALLEYVIEW(i915)) { > > if (cdclk >= 320000) /* jump to highest voltage for 400MHz too */ > > return 2; > > else if (cdclk >= 266667) > > @@ -456,29 +456,29 @@ static u8 vlv_calc_voltage_level(struct drm_i915_private *dev_priv, int cdclk) > > * hardware has shown that we just need to write the desired > > * CCK divider into the Punit register. > > */ > > - return DIV_ROUND_CLOSEST(dev_priv->hpll_freq << 1, cdclk) - 1; > > + return DIV_ROUND_CLOSEST(i915->hpll_freq << 1, cdclk) - 1; > > } > > } > > -static void vlv_get_cdclk(struct drm_i915_private *dev_priv, > > +static void vlv_get_cdclk(struct drm_i915_private *i915, > > struct intel_cdclk_state *cdclk_state) > > { > > u32 val; > > - vlv_iosf_sb_get(dev_priv, > > + vlv_iosf_sb_get(i915, > > BIT(VLV_IOSF_SB_CCK) | BIT(VLV_IOSF_SB_PUNIT)); > > - cdclk_state->vco = vlv_get_hpll_vco(dev_priv); > > - cdclk_state->cdclk = vlv_get_cck_clock(dev_priv, "cdclk", > > + cdclk_state->vco = vlv_get_hpll_vco(i915); > > + cdclk_state->cdclk = vlv_get_cck_clock(i915, "cdclk", > > CCK_DISPLAY_CLOCK_CONTROL, > > cdclk_state->vco); > > - val = vlv_punit_read(dev_priv, PUNIT_REG_DSPSSPM); > > + val = vlv_punit_read(i915, PUNIT_REG_DSPSSPM); > > - vlv_iosf_sb_put(dev_priv, > > + vlv_iosf_sb_put(i915, > > BIT(VLV_IOSF_SB_CCK) | BIT(VLV_IOSF_SB_PUNIT)); > > - if (IS_VALLEYVIEW(dev_priv)) > > + if (IS_VALLEYVIEW(i915)) > > cdclk_state->voltage_level = (val & DSPFREQGUAR_MASK) >> > > DSPFREQGUAR_SHIFT; > > else > > @@ -486,18 +486,18 @@ static void vlv_get_cdclk(struct drm_i915_private *dev_priv, > > DSPFREQGUAR_SHIFT_CHV; > > } > > -static void vlv_program_pfi_credits(struct drm_i915_private *dev_priv) > > +static void vlv_program_pfi_credits(struct drm_i915_private *i915) > > { > > unsigned int credits, default_credits; > > - if (IS_CHERRYVIEW(dev_priv)) > > + if (IS_CHERRYVIEW(i915)) > > default_credits = PFI_CREDIT(12); > > else > > default_credits = PFI_CREDIT(8); > > - if (dev_priv->cdclk.hw.cdclk >= dev_priv->czclk_freq) { > > + if (i915->cdclk.hw.cdclk >= i915->czclk_freq) { > > /* CHV suggested value is 31 or 63 */ > > - if (IS_CHERRYVIEW(dev_priv)) > > + if (IS_CHERRYVIEW(i915)) > > credits = PFI_CREDIT_63; > > else > > credits = PFI_CREDIT(15); > > @@ -522,7 +522,7 @@ static void vlv_program_pfi_credits(struct drm_i915_private *dev_priv) > > WARN_ON(I915_READ(GCI_CONTROL) & PFI_CREDIT_RESEND); > > } > > -static void vlv_set_cdclk(struct drm_i915_private *dev_priv, > > +static void vlv_set_cdclk(struct drm_i915_private *i915, > > const struct intel_cdclk_state *cdclk_state, > > enum pipe pipe) > > { > > @@ -548,18 +548,18 @@ static void vlv_set_cdclk(struct drm_i915_private *dev_priv, > > * a system suspend. So grab the PIPE-A domain, which covers > > * the HW blocks needed for the following programming. > > */ > > - wakeref = intel_display_power_get(dev_priv, POWER_DOMAIN_PIPE_A); > > + wakeref = intel_display_power_get(i915, POWER_DOMAIN_PIPE_A); > > - vlv_iosf_sb_get(dev_priv, > > + vlv_iosf_sb_get(i915, > > BIT(VLV_IOSF_SB_CCK) | > > BIT(VLV_IOSF_SB_BUNIT) | > > BIT(VLV_IOSF_SB_PUNIT)); > > - val = vlv_punit_read(dev_priv, PUNIT_REG_DSPSSPM); > > + val = vlv_punit_read(i915, PUNIT_REG_DSPSSPM); > > val &= ~DSPFREQGUAR_MASK; > > val |= (cmd << DSPFREQGUAR_SHIFT); > > - vlv_punit_write(dev_priv, PUNIT_REG_DSPSSPM, val); > > - if (wait_for((vlv_punit_read(dev_priv, PUNIT_REG_DSPSSPM) & > > + vlv_punit_write(i915, PUNIT_REG_DSPSSPM, val); > > + if (wait_for((vlv_punit_read(i915, PUNIT_REG_DSPSSPM) & > > DSPFREQSTAT_MASK) == (cmd << DSPFREQSTAT_SHIFT), > > 50)) { > > DRM_ERROR("timed out waiting for CDclk change\n"); > > @@ -568,23 +568,23 @@ static void vlv_set_cdclk(struct drm_i915_private *dev_priv, > > if (cdclk == 400000) { > > u32 divider; > > - divider = DIV_ROUND_CLOSEST(dev_priv->hpll_freq << 1, > > + divider = DIV_ROUND_CLOSEST(i915->hpll_freq << 1, > > cdclk) - 1; > > /* adjust cdclk divider */ > > - val = vlv_cck_read(dev_priv, CCK_DISPLAY_CLOCK_CONTROL); > > + val = vlv_cck_read(i915, CCK_DISPLAY_CLOCK_CONTROL); > > val &= ~CCK_FREQUENCY_VALUES; > > val |= divider; > > - vlv_cck_write(dev_priv, CCK_DISPLAY_CLOCK_CONTROL, val); > > + vlv_cck_write(i915, CCK_DISPLAY_CLOCK_CONTROL, val); > > - if (wait_for((vlv_cck_read(dev_priv, CCK_DISPLAY_CLOCK_CONTROL) & > > + if (wait_for((vlv_cck_read(i915, CCK_DISPLAY_CLOCK_CONTROL) & > > CCK_FREQUENCY_STATUS) == (divider << CCK_FREQUENCY_STATUS_SHIFT), > > 50)) > > DRM_ERROR("timed out waiting for CDclk change\n"); > > } > > /* adjust self-refresh exit latency value */ > > - val = vlv_bunit_read(dev_priv, BUNIT_REG_BISOC); > > + val = vlv_bunit_read(i915, BUNIT_REG_BISOC); > > val &= ~0x7f; > > /* > > @@ -595,21 +595,21 @@ static void vlv_set_cdclk(struct drm_i915_private *dev_priv, > > val |= 4500 / 250; /* 4.5 usec */ > > else > > val |= 3000 / 250; /* 3.0 usec */ > > - vlv_bunit_write(dev_priv, BUNIT_REG_BISOC, val); > > + vlv_bunit_write(i915, BUNIT_REG_BISOC, val); > > - vlv_iosf_sb_put(dev_priv, > > + vlv_iosf_sb_put(i915, > > BIT(VLV_IOSF_SB_CCK) | > > BIT(VLV_IOSF_SB_BUNIT) | > > BIT(VLV_IOSF_SB_PUNIT)); > > - intel_update_cdclk(dev_priv); > > + intel_update_cdclk(i915); > > - vlv_program_pfi_credits(dev_priv); > > + vlv_program_pfi_credits(i915); > > - intel_display_power_put(dev_priv, POWER_DOMAIN_PIPE_A, wakeref); > > + intel_display_power_put(i915, POWER_DOMAIN_PIPE_A, wakeref); > > } > > -static void chv_set_cdclk(struct drm_i915_private *dev_priv, > > +static void chv_set_cdclk(struct drm_i915_private *i915, > > const struct intel_cdclk_state *cdclk_state, > > enum pipe pipe) > > { > > @@ -634,26 +634,26 @@ static void chv_set_cdclk(struct drm_i915_private *dev_priv, > > * a system suspend. So grab the PIPE-A domain, which covers > > * the HW blocks needed for the following programming. > > */ > > - wakeref = intel_display_power_get(dev_priv, POWER_DOMAIN_PIPE_A); > > + wakeref = intel_display_power_get(i915, POWER_DOMAIN_PIPE_A); > > - vlv_punit_get(dev_priv); > > - val = vlv_punit_read(dev_priv, PUNIT_REG_DSPSSPM); > > + vlv_punit_get(i915); > > + val = vlv_punit_read(i915, PUNIT_REG_DSPSSPM); > > val &= ~DSPFREQGUAR_MASK_CHV; > > val |= (cmd << DSPFREQGUAR_SHIFT_CHV); > > - vlv_punit_write(dev_priv, PUNIT_REG_DSPSSPM, val); > > - if (wait_for((vlv_punit_read(dev_priv, PUNIT_REG_DSPSSPM) & > > + vlv_punit_write(i915, PUNIT_REG_DSPSSPM, val); > > + if (wait_for((vlv_punit_read(i915, PUNIT_REG_DSPSSPM) & > > DSPFREQSTAT_MASK_CHV) == (cmd << DSPFREQSTAT_SHIFT_CHV), > > 50)) { > > DRM_ERROR("timed out waiting for CDclk change\n"); > > } > > - vlv_punit_put(dev_priv); > > + vlv_punit_put(i915); > > - intel_update_cdclk(dev_priv); > > + intel_update_cdclk(i915); > > - vlv_program_pfi_credits(dev_priv); > > + vlv_program_pfi_credits(i915); > > - intel_display_power_put(dev_priv, POWER_DOMAIN_PIPE_A, wakeref); > > + intel_display_power_put(i915, POWER_DOMAIN_PIPE_A, wakeref); > > } > > static int bdw_calc_cdclk(int min_cdclk) > > @@ -683,7 +683,7 @@ static u8 bdw_calc_voltage_level(int cdclk) > > } > > } > > -static void bdw_get_cdclk(struct drm_i915_private *dev_priv, > > +static void bdw_get_cdclk(struct drm_i915_private *i915, > > struct intel_cdclk_state *cdclk_state) > > { > > u32 lcpll = I915_READ(LCPLL_CTL); > > @@ -710,7 +710,7 @@ static void bdw_get_cdclk(struct drm_i915_private *dev_priv, > > bdw_calc_voltage_level(cdclk_state->cdclk); > > } > > -static void bdw_set_cdclk(struct drm_i915_private *dev_priv, > > +static void bdw_set_cdclk(struct drm_i915_private *i915, > > const struct intel_cdclk_state *cdclk_state, > > enum pipe pipe) > > { > > @@ -726,7 +726,7 @@ static void bdw_set_cdclk(struct drm_i915_private *dev_priv, > > "trying to change cdclk frequency with cdclk not enabled\n")) > > return; > > - ret = sandybridge_pcode_write(dev_priv, > > + ret = sandybridge_pcode_write(i915, > > BDW_PCODE_DISPLAY_FREQ_CHANGE_REQ, 0x0); > > if (ret) { > > DRM_ERROR("failed to inform pcode about cdclk change\n"); > > @@ -776,12 +776,12 @@ static void bdw_set_cdclk(struct drm_i915_private *dev_priv, > > LCPLL_CD_SOURCE_FCLK_DONE) == 0, 1)) > > DRM_ERROR("Switching back to LCPLL failed\n"); > > - sandybridge_pcode_write(dev_priv, HSW_PCODE_DE_WRITE_FREQ_REQ, > > + sandybridge_pcode_write(i915, HSW_PCODE_DE_WRITE_FREQ_REQ, > > cdclk_state->voltage_level); > > I915_WRITE(CDCLK_FREQ, DIV_ROUND_CLOSEST(cdclk, 1000) - 1); > > - intel_update_cdclk(dev_priv); > > + intel_update_cdclk(i915); > > } > > static int skl_calc_cdclk(int min_cdclk, int vco) > > @@ -819,7 +819,7 @@ static u8 skl_calc_voltage_level(int cdclk) > > return 0; > > } > > -static void skl_dpll0_update(struct drm_i915_private *dev_priv, > > +static void skl_dpll0_update(struct drm_i915_private *i915, > > struct intel_cdclk_state *cdclk_state) > > { > > u32 val; > > @@ -859,12 +859,12 @@ static void skl_dpll0_update(struct drm_i915_private *dev_priv, > > } > > } > > -static void skl_get_cdclk(struct drm_i915_private *dev_priv, > > +static void skl_get_cdclk(struct drm_i915_private *i915, > > struct intel_cdclk_state *cdclk_state) > > { > > u32 cdctl; > > - skl_dpll0_update(dev_priv, cdclk_state); > > + skl_dpll0_update(i915, cdclk_state); > > cdclk_state->cdclk = cdclk_state->bypass = cdclk_state->ref; > > @@ -926,18 +926,18 @@ static int skl_cdclk_decimal(int cdclk) > > return DIV_ROUND_CLOSEST(cdclk - 1000, 500); > > } > > -static void skl_set_preferred_cdclk_vco(struct drm_i915_private *dev_priv, > > +static void skl_set_preferred_cdclk_vco(struct drm_i915_private *i915, > > int vco) > > { > > - bool changed = dev_priv->skl_preferred_vco_freq != vco; > > + bool changed = i915->skl_preferred_vco_freq != vco; > > - dev_priv->skl_preferred_vco_freq = vco; > > + i915->skl_preferred_vco_freq = vco; > > if (changed) > > - intel_update_max_cdclk(dev_priv); > > + intel_update_max_cdclk(i915); > > } > > -static void skl_dpll0_enable(struct drm_i915_private *dev_priv, int vco) > > +static void skl_dpll0_enable(struct drm_i915_private *i915, int vco) > > { > > u32 val; > > @@ -969,29 +969,29 @@ static void skl_dpll0_enable(struct drm_i915_private *dev_priv, int vco) > > I915_WRITE(LCPLL1_CTL, I915_READ(LCPLL1_CTL) | LCPLL_PLL_ENABLE); > > - if (intel_wait_for_register(&dev_priv->uncore, > > + if (intel_wait_for_register(&i915->uncore, > > LCPLL1_CTL, LCPLL_PLL_LOCK, LCPLL_PLL_LOCK, > > 5)) > > DRM_ERROR("DPLL0 not locked\n"); > > - dev_priv->cdclk.hw.vco = vco; > > + i915->cdclk.hw.vco = vco; > > /* We'll want to keep using the current vco from now on. */ > > - skl_set_preferred_cdclk_vco(dev_priv, vco); > > + skl_set_preferred_cdclk_vco(i915, vco); > > } > > -static void skl_dpll0_disable(struct drm_i915_private *dev_priv) > > +static void skl_dpll0_disable(struct drm_i915_private *i915) > > { > > I915_WRITE(LCPLL1_CTL, I915_READ(LCPLL1_CTL) & ~LCPLL_PLL_ENABLE); > > - if (intel_wait_for_register(&dev_priv->uncore, > > + if (intel_wait_for_register(&i915->uncore, > > LCPLL1_CTL, LCPLL_PLL_LOCK, 0, > > 1)) > > DRM_ERROR("Couldn't disable DPLL0\n"); > > - dev_priv->cdclk.hw.vco = 0; > > + i915->cdclk.hw.vco = 0; > > } > > -static void skl_set_cdclk(struct drm_i915_private *dev_priv, > > +static void skl_set_cdclk(struct drm_i915_private *i915, > > const struct intel_cdclk_state *cdclk_state, > > enum pipe pipe) > > { > > @@ -1008,9 +1008,9 @@ static void skl_set_cdclk(struct drm_i915_private *dev_priv, > > * use the corresponding VCO freq as that always leads to using the > > * minimum 308MHz CDCLK. > > */ > > - WARN_ON_ONCE(IS_SKYLAKE(dev_priv) && vco == 8640000); > > + WARN_ON_ONCE(IS_SKYLAKE(i915) && vco == 8640000); > > - ret = skl_pcode_request(dev_priv, SKL_PCODE_CDCLK_CONTROL, > > + ret = skl_pcode_request(i915, SKL_PCODE_CDCLK_CONTROL, > > SKL_CDCLK_PREPARE_FOR_CHANGE, > > SKL_CDCLK_READY_FOR_CHANGE, > > SKL_CDCLK_READY_FOR_CHANGE, 3); > > @@ -1023,7 +1023,7 @@ static void skl_set_cdclk(struct drm_i915_private *dev_priv, > > /* Choose frequency for this cdclk */ > > switch (cdclk) { > > default: > > - WARN_ON(cdclk != dev_priv->cdclk.hw.bypass); > > + WARN_ON(cdclk != i915->cdclk.hw.bypass); > > WARN_ON(vco != 0); > > /* fall through */ > > case 308571: > > @@ -1043,13 +1043,13 @@ static void skl_set_cdclk(struct drm_i915_private *dev_priv, > > break; > > } > > - if (dev_priv->cdclk.hw.vco != 0 && > > - dev_priv->cdclk.hw.vco != vco) > > - skl_dpll0_disable(dev_priv); > > + if (i915->cdclk.hw.vco != 0 && > > + i915->cdclk.hw.vco != vco) > > + skl_dpll0_disable(i915); > > cdclk_ctl = I915_READ(CDCLK_CTL); > > - if (dev_priv->cdclk.hw.vco != vco) { > > + if (i915->cdclk.hw.vco != vco) { > > /* Wa Display #1183: skl,kbl,cfl */ > > cdclk_ctl &= ~(CDCLK_FREQ_SEL_MASK | CDCLK_FREQ_DECIMAL_MASK); > > cdclk_ctl |= freq_select | skl_cdclk_decimal(cdclk); > > @@ -1061,8 +1061,8 @@ static void skl_set_cdclk(struct drm_i915_private *dev_priv, > > I915_WRITE(CDCLK_CTL, cdclk_ctl); > > POSTING_READ(CDCLK_CTL); > > - if (dev_priv->cdclk.hw.vco != vco) > > - skl_dpll0_enable(dev_priv, vco); > > + if (i915->cdclk.hw.vco != vco) > > + skl_dpll0_enable(i915, vco); > > /* Wa Display #1183: skl,kbl,cfl */ > > cdclk_ctl &= ~(CDCLK_FREQ_SEL_MASK | CDCLK_FREQ_DECIMAL_MASK); > > @@ -1077,13 +1077,13 @@ static void skl_set_cdclk(struct drm_i915_private *dev_priv, > > POSTING_READ(CDCLK_CTL); > > /* inform PCU of the change */ > > - sandybridge_pcode_write(dev_priv, SKL_PCODE_CDCLK_CONTROL, > > + sandybridge_pcode_write(i915, SKL_PCODE_CDCLK_CONTROL, > > cdclk_state->voltage_level); > > - intel_update_cdclk(dev_priv); > > + intel_update_cdclk(i915); > > } > > -static void skl_sanitize_cdclk(struct drm_i915_private *dev_priv) > > +static void skl_sanitize_cdclk(struct drm_i915_private *i915) > > { > > u32 cdctl, expected; > > @@ -1095,12 +1095,12 @@ static void skl_sanitize_cdclk(struct drm_i915_private *dev_priv) > > if ((I915_READ(SWF_ILK(0x18)) & 0x00FFFFFF) == 0) > > goto sanitize; > > - intel_update_cdclk(dev_priv); > > - intel_dump_cdclk_state(&dev_priv->cdclk.hw, "Current CDCLK"); > > + intel_update_cdclk(i915); > > + intel_dump_cdclk_state(&i915->cdclk.hw, "Current CDCLK"); > > /* Is PLL enabled and locked ? */ > > - if (dev_priv->cdclk.hw.vco == 0 || > > - dev_priv->cdclk.hw.cdclk == dev_priv->cdclk.hw.bypass) > > + if (i915->cdclk.hw.vco == 0 || > > + i915->cdclk.hw.cdclk == i915->cdclk.hw.bypass) > > goto sanitize; > > /* DPLL okay; verify the cdclock > > @@ -1111,7 +1111,7 @@ static void skl_sanitize_cdclk(struct drm_i915_private *dev_priv) > > */ > > cdctl = I915_READ(CDCLK_CTL); > > expected = (cdctl & CDCLK_FREQ_SEL_MASK) | > > - skl_cdclk_decimal(dev_priv->cdclk.hw.cdclk); > > + skl_cdclk_decimal(i915->cdclk.hw.cdclk); > > if (cdctl == expected) > > /* All well; nothing to sanitize */ > > return; > > @@ -1120,49 +1120,49 @@ static void skl_sanitize_cdclk(struct drm_i915_private *dev_priv) > > DRM_DEBUG_KMS("Sanitizing cdclk programmed by pre-os\n"); > > /* force cdclk programming */ > > - dev_priv->cdclk.hw.cdclk = 0; > > + i915->cdclk.hw.cdclk = 0; > > /* force full PLL disable + enable */ > > - dev_priv->cdclk.hw.vco = -1; > > + i915->cdclk.hw.vco = -1; > > } > > -static void skl_init_cdclk(struct drm_i915_private *dev_priv) > > +static void skl_init_cdclk(struct drm_i915_private *i915) > > { > > struct intel_cdclk_state cdclk_state; > > - skl_sanitize_cdclk(dev_priv); > > + skl_sanitize_cdclk(i915); > > - if (dev_priv->cdclk.hw.cdclk != 0 && > > - dev_priv->cdclk.hw.vco != 0) { > > + if (i915->cdclk.hw.cdclk != 0 && > > + i915->cdclk.hw.vco != 0) { > > /* > > * Use the current vco as our initial > > * guess as to what the preferred vco is. > > */ > > - if (dev_priv->skl_preferred_vco_freq == 0) > > - skl_set_preferred_cdclk_vco(dev_priv, > > - dev_priv->cdclk.hw.vco); > > + if (i915->skl_preferred_vco_freq == 0) > > + skl_set_preferred_cdclk_vco(i915, > > + i915->cdclk.hw.vco); > > return; > > } > > - cdclk_state = dev_priv->cdclk.hw; > > + cdclk_state = i915->cdclk.hw; > > - cdclk_state.vco = dev_priv->skl_preferred_vco_freq; > > + cdclk_state.vco = i915->skl_preferred_vco_freq; > > if (cdclk_state.vco == 0) > > cdclk_state.vco = 8100000; > > cdclk_state.cdclk = skl_calc_cdclk(0, cdclk_state.vco); > > cdclk_state.voltage_level = skl_calc_voltage_level(cdclk_state.cdclk); > > - skl_set_cdclk(dev_priv, &cdclk_state, INVALID_PIPE); > > + skl_set_cdclk(i915, &cdclk_state, INVALID_PIPE); > > } > > -static void skl_uninit_cdclk(struct drm_i915_private *dev_priv) > > +static void skl_uninit_cdclk(struct drm_i915_private *i915) > > { > > - struct intel_cdclk_state cdclk_state = dev_priv->cdclk.hw; > > + struct intel_cdclk_state cdclk_state = i915->cdclk.hw; > > cdclk_state.cdclk = cdclk_state.bypass; > > cdclk_state.vco = 0; > > cdclk_state.voltage_level = skl_calc_voltage_level(cdclk_state.cdclk); > > - skl_set_cdclk(dev_priv, &cdclk_state, INVALID_PIPE); > > + skl_set_cdclk(i915, &cdclk_state, INVALID_PIPE); > > } > > static int bxt_calc_cdclk(int min_cdclk) > > @@ -1194,11 +1194,11 @@ static u8 bxt_calc_voltage_level(int cdclk) > > return DIV_ROUND_UP(cdclk, 25000); > > } > > -static int bxt_de_pll_vco(struct drm_i915_private *dev_priv, int cdclk) > > +static int bxt_de_pll_vco(struct drm_i915_private *i915, int cdclk) > > { > > int ratio; > > - if (cdclk == dev_priv->cdclk.hw.bypass) > > + if (cdclk == i915->cdclk.hw.bypass) > > return 0; > > switch (cdclk) { > > @@ -1216,14 +1216,14 @@ static int bxt_de_pll_vco(struct drm_i915_private *dev_priv, int cdclk) > > break; > > } > > - return dev_priv->cdclk.hw.ref * ratio; > > + return i915->cdclk.hw.ref * ratio; > > } > > -static int glk_de_pll_vco(struct drm_i915_private *dev_priv, int cdclk) > > +static int glk_de_pll_vco(struct drm_i915_private *i915, int cdclk) > > { > > int ratio; > > - if (cdclk == dev_priv->cdclk.hw.bypass) > > + if (cdclk == i915->cdclk.hw.bypass) > > return 0; > > switch (cdclk) { > > @@ -1237,10 +1237,10 @@ static int glk_de_pll_vco(struct drm_i915_private *dev_priv, int cdclk) > > break; > > } > > - return dev_priv->cdclk.hw.ref * ratio; > > + return i915->cdclk.hw.ref * ratio; > > } > > -static void bxt_de_pll_update(struct drm_i915_private *dev_priv, > > +static void bxt_de_pll_update(struct drm_i915_private *i915, > > struct intel_cdclk_state *cdclk_state) > > { > > u32 val; > > @@ -1259,13 +1259,13 @@ static void bxt_de_pll_update(struct drm_i915_private *dev_priv, > > cdclk_state->vco = (val & BXT_DE_PLL_RATIO_MASK) * cdclk_state->ref; > > } > > -static void bxt_get_cdclk(struct drm_i915_private *dev_priv, > > +static void bxt_get_cdclk(struct drm_i915_private *i915, > > struct intel_cdclk_state *cdclk_state) > > { > > u32 divider; > > int div; > > - bxt_de_pll_update(dev_priv, cdclk_state); > > + bxt_de_pll_update(i915, cdclk_state); > > cdclk_state->cdclk = cdclk_state->bypass = cdclk_state->ref; > > @@ -1279,7 +1279,7 @@ static void bxt_get_cdclk(struct drm_i915_private *dev_priv, > > div = 2; > > break; > > case BXT_CDCLK_CD2X_DIV_SEL_1_5: > > - WARN(IS_GEMINILAKE(dev_priv), "Unsupported divider\n"); > > + WARN(IS_GEMINILAKE(i915), "Unsupported divider\n"); > > div = 3; > > break; > > case BXT_CDCLK_CD2X_DIV_SEL_2: > > @@ -1304,22 +1304,22 @@ static void bxt_get_cdclk(struct drm_i915_private *dev_priv, > > bxt_calc_voltage_level(cdclk_state->cdclk); > > } > > -static void bxt_de_pll_disable(struct drm_i915_private *dev_priv) > > +static void bxt_de_pll_disable(struct drm_i915_private *i915) > > { > > I915_WRITE(BXT_DE_PLL_ENABLE, 0); > > /* Timeout 200us */ > > - if (intel_wait_for_register(&dev_priv->uncore, > > + if (intel_wait_for_register(&i915->uncore, > > BXT_DE_PLL_ENABLE, BXT_DE_PLL_LOCK, 0, > > 1)) > > DRM_ERROR("timeout waiting for DE PLL unlock\n"); > > - dev_priv->cdclk.hw.vco = 0; > > + i915->cdclk.hw.vco = 0; > > } > > -static void bxt_de_pll_enable(struct drm_i915_private *dev_priv, int vco) > > +static void bxt_de_pll_enable(struct drm_i915_private *i915, int vco) > > { > > - int ratio = DIV_ROUND_CLOSEST(vco, dev_priv->cdclk.hw.ref); > > + int ratio = DIV_ROUND_CLOSEST(vco, i915->cdclk.hw.ref); > > u32 val; > > val = I915_READ(BXT_DE_PLL_CTL); > > @@ -1330,17 +1330,17 @@ static void bxt_de_pll_enable(struct drm_i915_private *dev_priv, int vco) > > I915_WRITE(BXT_DE_PLL_ENABLE, BXT_DE_PLL_PLL_ENABLE); > > /* Timeout 200us */ > > - if (intel_wait_for_register(&dev_priv->uncore, > > + if (intel_wait_for_register(&i915->uncore, > > BXT_DE_PLL_ENABLE, > > BXT_DE_PLL_LOCK, > > BXT_DE_PLL_LOCK, > > 1)) > > DRM_ERROR("timeout waiting for DE PLL lock\n"); > > - dev_priv->cdclk.hw.vco = vco; > > + i915->cdclk.hw.vco = vco; > > } > > -static void bxt_set_cdclk(struct drm_i915_private *dev_priv, > > +static void bxt_set_cdclk(struct drm_i915_private *i915, > > const struct intel_cdclk_state *cdclk_state, > > enum pipe pipe) > > { > > @@ -1352,14 +1352,14 @@ static void bxt_set_cdclk(struct drm_i915_private *dev_priv, > > /* cdclk = vco / 2 / div{1,1.5,2,4} */ > > switch (DIV_ROUND_CLOSEST(vco, cdclk)) { > > default: > > - WARN_ON(cdclk != dev_priv->cdclk.hw.bypass); > > + WARN_ON(cdclk != i915->cdclk.hw.bypass); > > WARN_ON(vco != 0); > > /* fall through */ > > case 2: > > divider = BXT_CDCLK_CD2X_DIV_SEL_1; > > break; > > case 3: > > - WARN(IS_GEMINILAKE(dev_priv), "Unsupported divider\n"); > > + WARN(IS_GEMINILAKE(i915), "Unsupported divider\n"); > > divider = BXT_CDCLK_CD2X_DIV_SEL_1_5; > > break; > > case 4: > > @@ -1375,7 +1375,7 @@ static void bxt_set_cdclk(struct drm_i915_private *dev_priv, > > * requires us to wait up to 150usec, but that leads to timeouts; > > * the 2ms used here is based on experiment. > > */ > > - ret = sandybridge_pcode_write_timeout(dev_priv, > > + ret = sandybridge_pcode_write_timeout(i915, > > HSW_PCODE_DE_WRITE_FREQ_REQ, > > 0x80000000, 150, 2); > > if (ret) { > > @@ -1384,12 +1384,12 @@ static void bxt_set_cdclk(struct drm_i915_private *dev_priv, > > return; > > } > > - if (dev_priv->cdclk.hw.vco != 0 && > > - dev_priv->cdclk.hw.vco != vco) > > - bxt_de_pll_disable(dev_priv); > > + if (i915->cdclk.hw.vco != 0 && > > + i915->cdclk.hw.vco != vco) > > + bxt_de_pll_disable(i915); > > - if (dev_priv->cdclk.hw.vco != vco) > > - bxt_de_pll_enable(dev_priv, vco); > > + if (i915->cdclk.hw.vco != vco) > > + bxt_de_pll_enable(i915, vco); > > val = divider | skl_cdclk_decimal(cdclk); > > if (pipe == INVALID_PIPE) > > @@ -1405,7 +1405,7 @@ static void bxt_set_cdclk(struct drm_i915_private *dev_priv, > > I915_WRITE(CDCLK_CTL, val); > > if (pipe != INVALID_PIPE) > > - intel_wait_for_vblank(dev_priv, pipe); > > + intel_wait_for_vblank(i915, pipe); > > /* > > * The timeout isn't specified, the 2ms used here is based on > > @@ -1413,7 +1413,7 @@ static void bxt_set_cdclk(struct drm_i915_private *dev_priv, > > * FIXME: Waiting for the request completion could be delayed until > > * the next PCODE request based on BSpec. > > */ > > - ret = sandybridge_pcode_write_timeout(dev_priv, > > + ret = sandybridge_pcode_write_timeout(i915, > > HSW_PCODE_DE_WRITE_FREQ_REQ, > > cdclk_state->voltage_level, 150, 2); > > if (ret) { > > @@ -1422,18 +1422,18 @@ static void bxt_set_cdclk(struct drm_i915_private *dev_priv, > > return; > > } > > - intel_update_cdclk(dev_priv); > > + intel_update_cdclk(i915); > > } > > -static void bxt_sanitize_cdclk(struct drm_i915_private *dev_priv) > > +static void bxt_sanitize_cdclk(struct drm_i915_private *i915) > > { > > u32 cdctl, expected; > > - intel_update_cdclk(dev_priv); > > - intel_dump_cdclk_state(&dev_priv->cdclk.hw, "Current CDCLK"); > > + intel_update_cdclk(i915); > > + intel_dump_cdclk_state(&i915->cdclk.hw, "Current CDCLK"); > > - if (dev_priv->cdclk.hw.vco == 0 || > > - dev_priv->cdclk.hw.cdclk == dev_priv->cdclk.hw.bypass) > > + if (i915->cdclk.hw.vco == 0 || > > + i915->cdclk.hw.cdclk == i915->cdclk.hw.bypass) > > goto sanitize; > > /* DPLL okay; verify the cdclock > > @@ -1451,12 +1451,12 @@ static void bxt_sanitize_cdclk(struct drm_i915_private *dev_priv) > > cdctl &= ~BXT_CDCLK_CD2X_PIPE_NONE; > > expected = (cdctl & BXT_CDCLK_CD2X_DIV_SEL_MASK) | > > - skl_cdclk_decimal(dev_priv->cdclk.hw.cdclk); > > + skl_cdclk_decimal(i915->cdclk.hw.cdclk); > > /* > > * Disable SSA Precharge when CD clock frequency < 500 MHz, > > * enable otherwise. > > */ > > - if (dev_priv->cdclk.hw.cdclk >= 500000) > > + if (i915->cdclk.hw.cdclk >= 500000) > > expected |= BXT_CDCLK_SSA_PRECHARGE_ENABLE; > > if (cdctl == expected) > > @@ -1467,50 +1467,50 @@ static void bxt_sanitize_cdclk(struct drm_i915_private *dev_priv) > > DRM_DEBUG_KMS("Sanitizing cdclk programmed by pre-os\n"); > > /* force cdclk programming */ > > - dev_priv->cdclk.hw.cdclk = 0; > > + i915->cdclk.hw.cdclk = 0; > > /* force full PLL disable + enable */ > > - dev_priv->cdclk.hw.vco = -1; > > + i915->cdclk.hw.vco = -1; > > } > > -static void bxt_init_cdclk(struct drm_i915_private *dev_priv) > > +static void bxt_init_cdclk(struct drm_i915_private *i915) > > { > > struct intel_cdclk_state cdclk_state; > > - bxt_sanitize_cdclk(dev_priv); > > + bxt_sanitize_cdclk(i915); > > - if (dev_priv->cdclk.hw.cdclk != 0 && > > - dev_priv->cdclk.hw.vco != 0) > > + if (i915->cdclk.hw.cdclk != 0 && > > + i915->cdclk.hw.vco != 0) > > return; > > - cdclk_state = dev_priv->cdclk.hw; > > + cdclk_state = i915->cdclk.hw; > > /* > > * FIXME: > > * - The initial CDCLK needs to be read from VBT. > > * Need to make this change after VBT has changes for BXT. > > */ > > - if (IS_GEMINILAKE(dev_priv)) { > > + if (IS_GEMINILAKE(i915)) { > > cdclk_state.cdclk = glk_calc_cdclk(0); > > - cdclk_state.vco = glk_de_pll_vco(dev_priv, cdclk_state.cdclk); > > + cdclk_state.vco = glk_de_pll_vco(i915, cdclk_state.cdclk); > > } else { > > cdclk_state.cdclk = bxt_calc_cdclk(0); > > - cdclk_state.vco = bxt_de_pll_vco(dev_priv, cdclk_state.cdclk); > > + cdclk_state.vco = bxt_de_pll_vco(i915, cdclk_state.cdclk); > > } > > cdclk_state.voltage_level = bxt_calc_voltage_level(cdclk_state.cdclk); > > - bxt_set_cdclk(dev_priv, &cdclk_state, INVALID_PIPE); > > + bxt_set_cdclk(i915, &cdclk_state, INVALID_PIPE); > > } > > -static void bxt_uninit_cdclk(struct drm_i915_private *dev_priv) > > +static void bxt_uninit_cdclk(struct drm_i915_private *i915) > > { > > - struct intel_cdclk_state cdclk_state = dev_priv->cdclk.hw; > > + struct intel_cdclk_state cdclk_state = i915->cdclk.hw; > > cdclk_state.cdclk = cdclk_state.bypass; > > cdclk_state.vco = 0; > > cdclk_state.voltage_level = bxt_calc_voltage_level(cdclk_state.cdclk); > > - bxt_set_cdclk(dev_priv, &cdclk_state, INVALID_PIPE); > > + bxt_set_cdclk(i915, &cdclk_state, INVALID_PIPE); > > } > > static int cnl_calc_cdclk(int min_cdclk) > > @@ -1533,7 +1533,7 @@ static u8 cnl_calc_voltage_level(int cdclk) > > return 0; > > } > > -static void cnl_cdclk_pll_update(struct drm_i915_private *dev_priv, > > +static void cnl_cdclk_pll_update(struct drm_i915_private *i915, > > struct intel_cdclk_state *cdclk_state) > > { > > u32 val; > > @@ -1555,13 +1555,13 @@ static void cnl_cdclk_pll_update(struct drm_i915_private *dev_priv, > > cdclk_state->vco = (val & CNL_CDCLK_PLL_RATIO_MASK) * cdclk_state->ref; > > } > > -static void cnl_get_cdclk(struct drm_i915_private *dev_priv, > > +static void cnl_get_cdclk(struct drm_i915_private *i915, > > struct intel_cdclk_state *cdclk_state) > > { > > u32 divider; > > int div; > > - cnl_cdclk_pll_update(dev_priv, cdclk_state); > > + cnl_cdclk_pll_update(i915, cdclk_state); > > cdclk_state->cdclk = cdclk_state->bypass = cdclk_state->ref; > > @@ -1593,7 +1593,7 @@ static void cnl_get_cdclk(struct drm_i915_private *dev_priv, > > cnl_calc_voltage_level(cdclk_state->cdclk); > > } > > -static void cnl_cdclk_pll_disable(struct drm_i915_private *dev_priv) > > +static void cnl_cdclk_pll_disable(struct drm_i915_private *i915) > > { > > u32 val; > > @@ -1605,12 +1605,12 @@ static void cnl_cdclk_pll_disable(struct drm_i915_private *dev_priv) > > if (wait_for((I915_READ(BXT_DE_PLL_ENABLE) & BXT_DE_PLL_LOCK) == 0, 1)) > > DRM_ERROR("timeout waiting for CDCLK PLL unlock\n"); > > - dev_priv->cdclk.hw.vco = 0; > > + i915->cdclk.hw.vco = 0; > > } > > -static void cnl_cdclk_pll_enable(struct drm_i915_private *dev_priv, int vco) > > +static void cnl_cdclk_pll_enable(struct drm_i915_private *i915, int vco) > > { > > - int ratio = DIV_ROUND_CLOSEST(vco, dev_priv->cdclk.hw.ref); > > + int ratio = DIV_ROUND_CLOSEST(vco, i915->cdclk.hw.ref); > > u32 val; > > val = CNL_CDCLK_PLL_RATIO(ratio); > > @@ -1623,10 +1623,10 @@ static void cnl_cdclk_pll_enable(struct drm_i915_private *dev_priv, int vco) > > if (wait_for((I915_READ(BXT_DE_PLL_ENABLE) & BXT_DE_PLL_LOCK) != 0, 1)) > > DRM_ERROR("timeout waiting for CDCLK PLL lock\n"); > > - dev_priv->cdclk.hw.vco = vco; > > + i915->cdclk.hw.vco = vco; > > } > > -static void cnl_set_cdclk(struct drm_i915_private *dev_priv, > > +static void cnl_set_cdclk(struct drm_i915_private *i915, > > const struct intel_cdclk_state *cdclk_state, > > enum pipe pipe) > > { > > @@ -1635,7 +1635,7 @@ static void cnl_set_cdclk(struct drm_i915_private *dev_priv, > > u32 val, divider; > > int ret; > > - ret = skl_pcode_request(dev_priv, SKL_PCODE_CDCLK_CONTROL, > > + ret = skl_pcode_request(i915, SKL_PCODE_CDCLK_CONTROL, > > SKL_CDCLK_PREPARE_FOR_CHANGE, > > SKL_CDCLK_READY_FOR_CHANGE, > > SKL_CDCLK_READY_FOR_CHANGE, 3); > > @@ -1648,7 +1648,7 @@ static void cnl_set_cdclk(struct drm_i915_private *dev_priv, > > /* cdclk = vco / 2 / div{1,2} */ > > switch (DIV_ROUND_CLOSEST(vco, cdclk)) { > > default: > > - WARN_ON(cdclk != dev_priv->cdclk.hw.bypass); > > + WARN_ON(cdclk != i915->cdclk.hw.bypass); > > WARN_ON(vco != 0); > > /* fall through */ > > case 2: > > @@ -1659,12 +1659,12 @@ static void cnl_set_cdclk(struct drm_i915_private *dev_priv, > > break; > > } > > - if (dev_priv->cdclk.hw.vco != 0 && > > - dev_priv->cdclk.hw.vco != vco) > > - cnl_cdclk_pll_disable(dev_priv); > > + if (i915->cdclk.hw.vco != 0 && > > + i915->cdclk.hw.vco != vco) > > + cnl_cdclk_pll_disable(i915); > > - if (dev_priv->cdclk.hw.vco != vco) > > - cnl_cdclk_pll_enable(dev_priv, vco); > > + if (i915->cdclk.hw.vco != vco) > > + cnl_cdclk_pll_enable(i915, vco); > > val = divider | skl_cdclk_decimal(cdclk); > > if (pipe == INVALID_PIPE) > > @@ -1674,26 +1674,26 @@ static void cnl_set_cdclk(struct drm_i915_private *dev_priv, > > I915_WRITE(CDCLK_CTL, val); > > if (pipe != INVALID_PIPE) > > - intel_wait_for_vblank(dev_priv, pipe); > > + intel_wait_for_vblank(i915, pipe); > > /* inform PCU of the change */ > > - sandybridge_pcode_write(dev_priv, SKL_PCODE_CDCLK_CONTROL, > > + sandybridge_pcode_write(i915, SKL_PCODE_CDCLK_CONTROL, > > cdclk_state->voltage_level); > > - intel_update_cdclk(dev_priv); > > + intel_update_cdclk(i915); > > /* > > * Can't read out the voltage level :( > > * Let's just assume everything is as expected. > > */ > > - dev_priv->cdclk.hw.voltage_level = cdclk_state->voltage_level; > > + i915->cdclk.hw.voltage_level = cdclk_state->voltage_level; > > } > > -static int cnl_cdclk_pll_vco(struct drm_i915_private *dev_priv, int cdclk) > > +static int cnl_cdclk_pll_vco(struct drm_i915_private *i915, int cdclk) > > { > > int ratio; > > - if (cdclk == dev_priv->cdclk.hw.bypass) > > + if (cdclk == i915->cdclk.hw.bypass) > > return 0; > > switch (cdclk) { > > @@ -1702,25 +1702,25 @@ static int cnl_cdclk_pll_vco(struct drm_i915_private *dev_priv, int cdclk) > > /* fall through */ > > case 168000: > > case 336000: > > - ratio = dev_priv->cdclk.hw.ref == 19200 ? 35 : 28; > > + ratio = i915->cdclk.hw.ref == 19200 ? 35 : 28; > > break; > > case 528000: > > - ratio = dev_priv->cdclk.hw.ref == 19200 ? 55 : 44; > > + ratio = i915->cdclk.hw.ref == 19200 ? 55 : 44; > > break; > > } > > - return dev_priv->cdclk.hw.ref * ratio; > > + return i915->cdclk.hw.ref * ratio; > > } > > -static void cnl_sanitize_cdclk(struct drm_i915_private *dev_priv) > > +static void cnl_sanitize_cdclk(struct drm_i915_private *i915) > > { > > u32 cdctl, expected; > > - intel_update_cdclk(dev_priv); > > - intel_dump_cdclk_state(&dev_priv->cdclk.hw, "Current CDCLK"); > > + intel_update_cdclk(i915); > > + intel_dump_cdclk_state(&i915->cdclk.hw, "Current CDCLK"); > > - if (dev_priv->cdclk.hw.vco == 0 || > > - dev_priv->cdclk.hw.cdclk == dev_priv->cdclk.hw.bypass) > > + if (i915->cdclk.hw.vco == 0 || > > + i915->cdclk.hw.cdclk == i915->cdclk.hw.bypass) > > goto sanitize; > > /* DPLL okay; verify the cdclock > > @@ -1738,7 +1738,7 @@ static void cnl_sanitize_cdclk(struct drm_i915_private *dev_priv) > > cdctl &= ~BXT_CDCLK_CD2X_PIPE_NONE; > > expected = (cdctl & BXT_CDCLK_CD2X_DIV_SEL_MASK) | > > - skl_cdclk_decimal(dev_priv->cdclk.hw.cdclk); > > + skl_cdclk_decimal(i915->cdclk.hw.cdclk); > > if (cdctl == expected) > > /* All well; nothing to sanitize */ > > @@ -1748,10 +1748,10 @@ static void cnl_sanitize_cdclk(struct drm_i915_private *dev_priv) > > DRM_DEBUG_KMS("Sanitizing cdclk programmed by pre-os\n"); > > /* force cdclk programming */ > > - dev_priv->cdclk.hw.cdclk = 0; > > + i915->cdclk.hw.cdclk = 0; > > /* force full PLL disable + enable */ > > - dev_priv->cdclk.hw.vco = -1; > > + i915->cdclk.hw.vco = -1; > > } > > static int icl_calc_cdclk(int min_cdclk, unsigned int ref) > > @@ -1781,11 +1781,11 @@ static int icl_calc_cdclk(int min_cdclk, unsigned int ref) > > return ranges[0]; > > } > > -static int icl_calc_cdclk_pll_vco(struct drm_i915_private *dev_priv, int cdclk) > > +static int icl_calc_cdclk_pll_vco(struct drm_i915_private *i915, int cdclk) > > { > > int ratio; > > - if (cdclk == dev_priv->cdclk.hw.bypass) > > + if (cdclk == i915->cdclk.hw.bypass) > > return 0; > > switch (cdclk) { > > @@ -1795,21 +1795,21 @@ static int icl_calc_cdclk_pll_vco(struct drm_i915_private *dev_priv, int cdclk) > > case 307200: > > case 556800: > > case 652800: > > - WARN_ON(dev_priv->cdclk.hw.ref != 19200 && > > - dev_priv->cdclk.hw.ref != 38400); > > + WARN_ON(i915->cdclk.hw.ref != 19200 && > > + i915->cdclk.hw.ref != 38400); > > break; > > case 312000: > > case 552000: > > case 648000: > > - WARN_ON(dev_priv->cdclk.hw.ref != 24000); > > + WARN_ON(i915->cdclk.hw.ref != 24000); > > } > > - ratio = cdclk / (dev_priv->cdclk.hw.ref / 2); > > + ratio = cdclk / (i915->cdclk.hw.ref / 2); > > - return dev_priv->cdclk.hw.ref * ratio; > > + return i915->cdclk.hw.ref * ratio; > > } > > -static void icl_set_cdclk(struct drm_i915_private *dev_priv, > > +static void icl_set_cdclk(struct drm_i915_private *i915, > > const struct intel_cdclk_state *cdclk_state, > > enum pipe pipe) > > { > > @@ -1817,7 +1817,7 @@ static void icl_set_cdclk(struct drm_i915_private *dev_priv, > > unsigned int vco = cdclk_state->vco; > > int ret; > > - ret = skl_pcode_request(dev_priv, SKL_PCODE_CDCLK_CONTROL, > > + ret = skl_pcode_request(i915, SKL_PCODE_CDCLK_CONTROL, > > SKL_CDCLK_PREPARE_FOR_CHANGE, > > SKL_CDCLK_READY_FOR_CHANGE, > > SKL_CDCLK_READY_FOR_CHANGE, 3); > > @@ -1827,12 +1827,12 @@ static void icl_set_cdclk(struct drm_i915_private *dev_priv, > > return; > > } > > - if (dev_priv->cdclk.hw.vco != 0 && > > - dev_priv->cdclk.hw.vco != vco) > > - cnl_cdclk_pll_disable(dev_priv); > > + if (i915->cdclk.hw.vco != 0 && > > + i915->cdclk.hw.vco != vco) > > + cnl_cdclk_pll_disable(i915); > > - if (dev_priv->cdclk.hw.vco != vco) > > - cnl_cdclk_pll_enable(dev_priv, vco); > > + if (i915->cdclk.hw.vco != vco) > > + cnl_cdclk_pll_enable(i915, vco); > > /* > > * On ICL CD2X_DIV can only be 1, so we'll never end up changing the > > @@ -1842,16 +1842,16 @@ static void icl_set_cdclk(struct drm_i915_private *dev_priv, > > I915_WRITE(CDCLK_CTL, ICL_CDCLK_CD2X_PIPE_NONE | > > skl_cdclk_decimal(cdclk)); > > - sandybridge_pcode_write(dev_priv, SKL_PCODE_CDCLK_CONTROL, > > + sandybridge_pcode_write(i915, SKL_PCODE_CDCLK_CONTROL, > > cdclk_state->voltage_level); > > - intel_update_cdclk(dev_priv); > > + intel_update_cdclk(i915); > > /* > > * Can't read out the voltage level :( > > * Let's just assume everything is as expected. > > */ > > - dev_priv->cdclk.hw.voltage_level = cdclk_state->voltage_level; > > + i915->cdclk.hw.voltage_level = cdclk_state->voltage_level; > > } > > static u8 icl_calc_voltage_level(int cdclk) > > @@ -1864,7 +1864,7 @@ static u8 icl_calc_voltage_level(int cdclk) > > return 0; > > } > > -static void icl_get_cdclk(struct drm_i915_private *dev_priv, > > +static void icl_get_cdclk(struct drm_i915_private *i915, > > struct intel_cdclk_state *cdclk_state) > > { > > u32 val; > > @@ -1915,17 +1915,17 @@ static void icl_get_cdclk(struct drm_i915_private *dev_priv, > > icl_calc_voltage_level(cdclk_state->cdclk); > > } > > -static void icl_init_cdclk(struct drm_i915_private *dev_priv) > > +static void icl_init_cdclk(struct drm_i915_private *i915) > > { > > struct intel_cdclk_state sanitized_state; > > u32 val; > > - /* This sets dev_priv->cdclk.hw. */ > > - intel_update_cdclk(dev_priv); > > - intel_dump_cdclk_state(&dev_priv->cdclk.hw, "Current CDCLK"); > > + /* This sets i915->cdclk.hw. */ > > + intel_update_cdclk(i915); > > + intel_dump_cdclk_state(&i915->cdclk.hw, "Current CDCLK"); > > /* This means CDCLK disabled. */ > > - if (dev_priv->cdclk.hw.cdclk == dev_priv->cdclk.hw.bypass) > > + if (i915->cdclk.hw.cdclk == i915->cdclk.hw.bypass) > > goto sanitize; > > val = I915_READ(CDCLK_CTL); > > @@ -1934,7 +1934,7 @@ static void icl_init_cdclk(struct drm_i915_private *dev_priv) > > goto sanitize; > > if ((val & CDCLK_FREQ_DECIMAL_MASK) != > > - skl_cdclk_decimal(dev_priv->cdclk.hw.cdclk)) > > + skl_cdclk_decimal(i915->cdclk.hw.cdclk)) > > goto sanitize; > > return; > > @@ -1942,62 +1942,62 @@ static void icl_init_cdclk(struct drm_i915_private *dev_priv) > > sanitize: > > DRM_DEBUG_KMS("Sanitizing cdclk programmed by pre-os\n"); > > - sanitized_state.ref = dev_priv->cdclk.hw.ref; > > + sanitized_state.ref = i915->cdclk.hw.ref; > > sanitized_state.cdclk = icl_calc_cdclk(0, sanitized_state.ref); > > - sanitized_state.vco = icl_calc_cdclk_pll_vco(dev_priv, > > + sanitized_state.vco = icl_calc_cdclk_pll_vco(i915, > > sanitized_state.cdclk); > > sanitized_state.voltage_level = > > icl_calc_voltage_level(sanitized_state.cdclk); > > - icl_set_cdclk(dev_priv, &sanitized_state, INVALID_PIPE); > > + icl_set_cdclk(i915, &sanitized_state, INVALID_PIPE); > > } > > -static void icl_uninit_cdclk(struct drm_i915_private *dev_priv) > > +static void icl_uninit_cdclk(struct drm_i915_private *i915) > > { > > - struct intel_cdclk_state cdclk_state = dev_priv->cdclk.hw; > > + struct intel_cdclk_state cdclk_state = i915->cdclk.hw; > > cdclk_state.cdclk = cdclk_state.bypass; > > cdclk_state.vco = 0; > > cdclk_state.voltage_level = icl_calc_voltage_level(cdclk_state.cdclk); > > - icl_set_cdclk(dev_priv, &cdclk_state, INVALID_PIPE); > > + icl_set_cdclk(i915, &cdclk_state, INVALID_PIPE); > > } > > -static void cnl_init_cdclk(struct drm_i915_private *dev_priv) > > +static void cnl_init_cdclk(struct drm_i915_private *i915) > > { > > struct intel_cdclk_state cdclk_state; > > - cnl_sanitize_cdclk(dev_priv); > > + cnl_sanitize_cdclk(i915); > > - if (dev_priv->cdclk.hw.cdclk != 0 && > > - dev_priv->cdclk.hw.vco != 0) > > + if (i915->cdclk.hw.cdclk != 0 && > > + i915->cdclk.hw.vco != 0) > > return; > > - cdclk_state = dev_priv->cdclk.hw; > > + cdclk_state = i915->cdclk.hw; > > cdclk_state.cdclk = cnl_calc_cdclk(0); > > - cdclk_state.vco = cnl_cdclk_pll_vco(dev_priv, cdclk_state.cdclk); > > + cdclk_state.vco = cnl_cdclk_pll_vco(i915, cdclk_state.cdclk); > > cdclk_state.voltage_level = cnl_calc_voltage_level(cdclk_state.cdclk); > > - cnl_set_cdclk(dev_priv, &cdclk_state, INVALID_PIPE); > > + cnl_set_cdclk(i915, &cdclk_state, INVALID_PIPE); > > } > > -static void cnl_uninit_cdclk(struct drm_i915_private *dev_priv) > > +static void cnl_uninit_cdclk(struct drm_i915_private *i915) > > { > > - struct intel_cdclk_state cdclk_state = dev_priv->cdclk.hw; > > + struct intel_cdclk_state cdclk_state = i915->cdclk.hw; > > cdclk_state.cdclk = cdclk_state.bypass; > > cdclk_state.vco = 0; > > cdclk_state.voltage_level = cnl_calc_voltage_level(cdclk_state.cdclk); > > - cnl_set_cdclk(dev_priv, &cdclk_state, INVALID_PIPE); > > + cnl_set_cdclk(i915, &cdclk_state, INVALID_PIPE); > > } > > /** > > * intel_cdclk_init - Initialize CDCLK > > * @i915: i915 device > > * > > - * Initialize CDCLK. This consists mainly of initializing dev_priv->cdclk.hw and > > + * Initialize CDCLK. This consists mainly of initializing i915->cdclk.hw and > > * sanitizing the state of the hardware if needed. This is generally done only > > * during the display core initialization sequence, after which the DMC will > > * take care of turning CDCLK off/on as needed. > > @@ -2051,19 +2051,19 @@ bool intel_cdclk_needs_modeset(const struct intel_cdclk_state *a, > > /** > > * intel_cdclk_needs_cd2x_update - Determine if two CDCLK states require a cd2x divider update > > - * @dev_priv: Not a CDCLK state, it's the drm_i915_private! > > + * @i915: Not a CDCLK state, it's the drm_i915_private! > > * @a: first CDCLK state > > * @b: second CDCLK state > > * > > * Returns: > > * True if the CDCLK states require just a cd2x divider update, false if not. > > */ > > -bool intel_cdclk_needs_cd2x_update(struct drm_i915_private *dev_priv, > > +bool intel_cdclk_needs_cd2x_update(struct drm_i915_private *i915, > > const struct intel_cdclk_state *a, > > const struct intel_cdclk_state *b) > > { > > /* Older hw doesn't have the capability */ > > - if (INTEL_GEN(dev_priv) < 10 && !IS_GEN9_LP(dev_priv)) > > + if (INTEL_GEN(i915) < 10 && !IS_GEN9_LP(i915)) > > return false; > > return a->cdclk != b->cdclk && > > @@ -2100,10 +2100,10 @@ bool intel_cdclk_changed(const struct intel_cdclk_state *a, > > */ > > void intel_cdclk_swap_state(struct intel_atomic_state *state) > > { > > - struct drm_i915_private *dev_priv = to_i915(state->base.dev); > > + struct drm_i915_private *i915 = to_i915(state->base.dev); > > - swap(state->cdclk.logical, dev_priv->cdclk.logical); > > - swap(state->cdclk.actual, dev_priv->cdclk.actual); > > + swap(state->cdclk.logical, i915->cdclk.logical); > > + swap(state->cdclk.actual, i915->cdclk.actual); > > } > > void intel_dump_cdclk_state(const struct intel_cdclk_state *cdclk_state, > > @@ -2117,37 +2117,37 @@ void intel_dump_cdclk_state(const struct intel_cdclk_state *cdclk_state, > > /** > > * intel_set_cdclk - Push the CDCLK state to the hardware > > - * @dev_priv: i915 device > > + * @i915: i915 device > > * @cdclk_state: new CDCLK state > > * @pipe: pipe with which to synchronize the update > > * > > * Program the hardware based on the passed in CDCLK state, > > * if necessary. > > */ > > -static void intel_set_cdclk(struct drm_i915_private *dev_priv, > > +static void intel_set_cdclk(struct drm_i915_private *i915, > > const struct intel_cdclk_state *cdclk_state, > > enum pipe pipe) > > { > > - if (!intel_cdclk_changed(&dev_priv->cdclk.hw, cdclk_state)) > > + if (!intel_cdclk_changed(&i915->cdclk.hw, cdclk_state)) > > return; > > - if (WARN_ON_ONCE(!dev_priv->display.set_cdclk)) > > + if (WARN_ON_ONCE(!i915->display.set_cdclk)) > > return; > > intel_dump_cdclk_state(cdclk_state, "Changing CDCLK to"); > > - dev_priv->display.set_cdclk(dev_priv, cdclk_state, pipe); > > + i915->display.set_cdclk(i915, cdclk_state, pipe); > > - if (WARN(intel_cdclk_changed(&dev_priv->cdclk.hw, cdclk_state), > > + if (WARN(intel_cdclk_changed(&i915->cdclk.hw, cdclk_state), > > "cdclk state doesn't match!\n")) { > > - intel_dump_cdclk_state(&dev_priv->cdclk.hw, "[hw state]"); > > + intel_dump_cdclk_state(&i915->cdclk.hw, "[hw state]"); > > intel_dump_cdclk_state(cdclk_state, "[sw state]"); > > } > > } > > /** > > * intel_set_cdclk_pre_plane_update - Push the CDCLK state to the hardware > > - * @dev_priv: i915 device > > + * @i915: i915 device > > * @old_state: old CDCLK state > > * @new_state: new CDCLK state > > * @pipe: pipe with which to synchronize the update > > @@ -2156,18 +2156,18 @@ static void intel_set_cdclk(struct drm_i915_private *dev_priv, > > * in CDCLK state, if necessary. > > */ > > void > > -intel_set_cdclk_pre_plane_update(struct drm_i915_private *dev_priv, > > +intel_set_cdclk_pre_plane_update(struct drm_i915_private *i915, > > const struct intel_cdclk_state *old_state, > > const struct intel_cdclk_state *new_state, > > enum pipe pipe) > > { > > if (pipe == INVALID_PIPE || old_state->cdclk <= new_state->cdclk) > > - intel_set_cdclk(dev_priv, new_state, pipe); > > + intel_set_cdclk(i915, new_state, pipe); > > } > > /** > > * intel_set_cdclk_post_plane_update - Push the CDCLK state to the hardware > > - * @dev_priv: i915 device > > + * @i915: i915 device > > * @old_state: old CDCLK state > > * @new_state: new CDCLK state > > * @pipe: pipe with which to synchronize the update > > @@ -2176,24 +2176,24 @@ intel_set_cdclk_pre_plane_update(struct drm_i915_private *dev_priv, > > * in CDCLK state, if necessary. > > */ > > void > > -intel_set_cdclk_post_plane_update(struct drm_i915_private *dev_priv, > > +intel_set_cdclk_post_plane_update(struct drm_i915_private *i915, > > const struct intel_cdclk_state *old_state, > > const struct intel_cdclk_state *new_state, > > enum pipe pipe) > > { > > if (pipe != INVALID_PIPE && old_state->cdclk > new_state->cdclk) > > - intel_set_cdclk(dev_priv, new_state, pipe); > > + intel_set_cdclk(i915, new_state, pipe); > > } > > -static int intel_pixel_rate_to_cdclk(struct drm_i915_private *dev_priv, > > +static int intel_pixel_rate_to_cdclk(struct drm_i915_private *i915, > > int pixel_rate) > > { > > - if (INTEL_GEN(dev_priv) >= 10 || IS_GEMINILAKE(dev_priv)) > > + if (INTEL_GEN(i915) >= 10 || IS_GEMINILAKE(i915)) > > return DIV_ROUND_UP(pixel_rate, 2); > > - else if (IS_GEN(dev_priv, 9) || > > - IS_BROADWELL(dev_priv) || IS_HASWELL(dev_priv)) > > + else if (IS_GEN(i915, 9) || > > + IS_BROADWELL(i915) || IS_HASWELL(i915)) > > return pixel_rate; > > - else if (IS_CHERRYVIEW(dev_priv)) > > + else if (IS_CHERRYVIEW(i915)) > > return DIV_ROUND_UP(pixel_rate * 100, 95); > > else > > return DIV_ROUND_UP(pixel_rate * 100, 90); > > @@ -2201,17 +2201,17 @@ static int intel_pixel_rate_to_cdclk(struct drm_i915_private *dev_priv, > > int intel_crtc_compute_min_cdclk(const struct intel_crtc_state *crtc_state) > > { > > - struct drm_i915_private *dev_priv = > > + struct drm_i915_private *i915 = > > to_i915(crtc_state->base.crtc->dev); > > int min_cdclk; > > if (!crtc_state->base.enable) > > return 0; > > - min_cdclk = intel_pixel_rate_to_cdclk(dev_priv, crtc_state->pixel_rate); > > + min_cdclk = intel_pixel_rate_to_cdclk(i915, crtc_state->pixel_rate); > > /* pixel rate mustn't exceed 95% of cdclk with IPS on BDW */ > > - if (IS_BROADWELL(dev_priv) && hsw_crtc_state_ips_capable(crtc_state)) > > + if (IS_BROADWELL(i915) && hsw_crtc_state_ips_capable(crtc_state)) > > min_cdclk = DIV_ROUND_UP(min_cdclk * 100, 95); > > /* BSpec says "Do not use DisplayPort with CDCLK less than 432 MHz, > > @@ -2223,10 +2223,10 @@ int intel_crtc_compute_min_cdclk(const struct intel_crtc_state *crtc_state) > > crtc_state->has_audio && > > crtc_state->port_clock >= 540000 && > > crtc_state->lane_count == 4) { > > - if (IS_CANNONLAKE(dev_priv) || IS_GEMINILAKE(dev_priv)) { > > + if (IS_CANNONLAKE(i915) || IS_GEMINILAKE(i915)) { > > /* Display WA #1145: glk,cnl */ > > min_cdclk = max(316800, min_cdclk); > > - } else if (IS_GEN(dev_priv, 9) || IS_BROADWELL(dev_priv)) { > > + } else if (IS_GEN(i915, 9) || IS_BROADWELL(i915)) { > > /* Display WA #1144: skl,bxt */ > > min_cdclk = max(432000, min_cdclk); > > } > > @@ -2236,7 +2236,7 @@ int intel_crtc_compute_min_cdclk(const struct intel_crtc_state *crtc_state) > > * According to BSpec, "The CD clock frequency must be at least twice > > * the frequency of the Azalia BCLK." and BCLK is 96 MHz by default. > > */ > > - if (crtc_state->has_audio && INTEL_GEN(dev_priv) >= 9) > > + if (crtc_state->has_audio && INTEL_GEN(i915) >= 9) > > min_cdclk = max(2 * 96000, min_cdclk); > > /* > > @@ -2244,7 +2244,7 @@ int intel_crtc_compute_min_cdclk(const struct intel_crtc_state *crtc_state) > > * than 320000KHz. > > */ > > if (intel_crtc_has_type(crtc_state, INTEL_OUTPUT_DSI) && > > - IS_VALLEYVIEW(dev_priv)) > > + IS_VALLEYVIEW(i915)) > > min_cdclk = max(320000, min_cdclk); > > /* > > @@ -2253,12 +2253,12 @@ int intel_crtc_compute_min_cdclk(const struct intel_crtc_state *crtc_state) > > * correct for DSI PLL and DE PLL. > > */ > > if (intel_crtc_has_type(crtc_state, INTEL_OUTPUT_DSI) && > > - IS_GEMINILAKE(dev_priv)) > > + IS_GEMINILAKE(i915)) > > min_cdclk = max(158400, min_cdclk); > > - if (min_cdclk > dev_priv->max_cdclk_freq) { > > + if (min_cdclk > i915->max_cdclk_freq) { > > DRM_DEBUG_KMS("required cdclk (%d kHz) exceeds max (%d kHz)\n", > > - min_cdclk, dev_priv->max_cdclk_freq); > > + min_cdclk, i915->max_cdclk_freq); > > return -EINVAL; > > } > > @@ -2267,13 +2267,13 @@ int intel_crtc_compute_min_cdclk(const struct intel_crtc_state *crtc_state) > > static int intel_compute_min_cdclk(struct intel_atomic_state *state) > > { > > - struct drm_i915_private *dev_priv = to_i915(state->base.dev); > > + struct drm_i915_private *i915 = to_i915(state->base.dev); > > struct intel_crtc *crtc; > > struct intel_crtc_state *crtc_state; > > int min_cdclk, i; > > enum pipe pipe; > > - memcpy(state->min_cdclk, dev_priv->min_cdclk, > > + memcpy(state->min_cdclk, i915->min_cdclk, > > sizeof(state->min_cdclk)); > > for_each_new_intel_crtc_in_state(state, crtc, crtc_state, i) { > > @@ -2285,7 +2285,7 @@ static int intel_compute_min_cdclk(struct intel_atomic_state *state) > > } > > min_cdclk = state->cdclk.force_min_cdclk; > > - for_each_pipe(dev_priv, pipe) > > + for_each_pipe(i915, pipe) > > min_cdclk = max(state->min_cdclk[pipe], min_cdclk); > > return min_cdclk; > > @@ -2302,14 +2302,14 @@ static int intel_compute_min_cdclk(struct intel_atomic_state *state) > > */ > > static u8 cnl_compute_min_voltage_level(struct intel_atomic_state *state) > > { > > - struct drm_i915_private *dev_priv = to_i915(state->base.dev); > > + struct drm_i915_private *i915 = to_i915(state->base.dev); > > struct intel_crtc *crtc; > > struct intel_crtc_state *crtc_state; > > u8 min_voltage_level; > > int i; > > enum pipe pipe; > > - memcpy(state->min_voltage_level, dev_priv->min_voltage_level, > > + memcpy(state->min_voltage_level, i915->min_voltage_level, > > sizeof(state->min_voltage_level)); > > for_each_new_intel_crtc_in_state(state, crtc, crtc_state, i) { > > @@ -2321,7 +2321,7 @@ static u8 cnl_compute_min_voltage_level(struct intel_atomic_state *state) > > } > > min_voltage_level = 0; > > - for_each_pipe(dev_priv, pipe) > > + for_each_pipe(i915, pipe) > > min_voltage_level = max(state->min_voltage_level[pipe], > > min_voltage_level); > > @@ -2330,25 +2330,25 @@ static u8 cnl_compute_min_voltage_level(struct intel_atomic_state *state) > > static int vlv_modeset_calc_cdclk(struct intel_atomic_state *state) > > { > > - struct drm_i915_private *dev_priv = to_i915(state->base.dev); > > + struct drm_i915_private *i915 = to_i915(state->base.dev); > > int min_cdclk, cdclk; > > min_cdclk = intel_compute_min_cdclk(state); > > if (min_cdclk < 0) > > return min_cdclk; > > - cdclk = vlv_calc_cdclk(dev_priv, min_cdclk); > > + cdclk = vlv_calc_cdclk(i915, min_cdclk); > > state->cdclk.logical.cdclk = cdclk; > > state->cdclk.logical.voltage_level = > > - vlv_calc_voltage_level(dev_priv, cdclk); > > + vlv_calc_voltage_level(i915, cdclk); > > if (!state->active_crtcs) { > > - cdclk = vlv_calc_cdclk(dev_priv, state->cdclk.force_min_cdclk); > > + cdclk = vlv_calc_cdclk(i915, state->cdclk.force_min_cdclk); > > state->cdclk.actual.cdclk = cdclk; > > state->cdclk.actual.voltage_level = > > - vlv_calc_voltage_level(dev_priv, cdclk); > > + vlv_calc_voltage_level(i915, cdclk); > > } else { > > state->cdclk.actual = state->cdclk.logical; > > } > > @@ -2389,14 +2389,14 @@ static int bdw_modeset_calc_cdclk(struct intel_atomic_state *state) > > static int skl_dpll0_vco(struct intel_atomic_state *state) > > { > > - struct drm_i915_private *dev_priv = to_i915(state->base.dev); > > + struct drm_i915_private *i915 = to_i915(state->base.dev); > > struct intel_crtc *crtc; > > struct intel_crtc_state *crtc_state; > > int vco, i; > > vco = state->cdclk.logical.vco; > > if (!vco) > > - vco = dev_priv->skl_preferred_vco_freq; > > + vco = i915->skl_preferred_vco_freq; > > for_each_new_intel_crtc_in_state(state, crtc, crtc_state, i) { > > if (!crtc_state->base.enable) > > @@ -2460,19 +2460,19 @@ static int skl_modeset_calc_cdclk(struct intel_atomic_state *state) > > static int bxt_modeset_calc_cdclk(struct intel_atomic_state *state) > > { > > - struct drm_i915_private *dev_priv = to_i915(state->base.dev); > > + struct drm_i915_private *i915 = to_i915(state->base.dev); > > int min_cdclk, cdclk, vco; > > min_cdclk = intel_compute_min_cdclk(state); > > if (min_cdclk < 0) > > return min_cdclk; > > - if (IS_GEMINILAKE(dev_priv)) { > > + if (IS_GEMINILAKE(i915)) { > > cdclk = glk_calc_cdclk(min_cdclk); > > - vco = glk_de_pll_vco(dev_priv, cdclk); > > + vco = glk_de_pll_vco(i915, cdclk); > > } else { > > cdclk = bxt_calc_cdclk(min_cdclk); > > - vco = bxt_de_pll_vco(dev_priv, cdclk); > > + vco = bxt_de_pll_vco(i915, cdclk); > > } > > state->cdclk.logical.vco = vco; > > @@ -2481,12 +2481,12 @@ static int bxt_modeset_calc_cdclk(struct intel_atomic_state *state) > > bxt_calc_voltage_level(cdclk); > > if (!state->active_crtcs) { > > - if (IS_GEMINILAKE(dev_priv)) { > > + if (IS_GEMINILAKE(i915)) { > > cdclk = glk_calc_cdclk(state->cdclk.force_min_cdclk); > > - vco = glk_de_pll_vco(dev_priv, cdclk); > > + vco = glk_de_pll_vco(i915, cdclk); > > } else { > > cdclk = bxt_calc_cdclk(state->cdclk.force_min_cdclk); > > - vco = bxt_de_pll_vco(dev_priv, cdclk); > > + vco = bxt_de_pll_vco(i915, cdclk); > > } > > state->cdclk.actual.vco = vco; > > @@ -2502,7 +2502,7 @@ static int bxt_modeset_calc_cdclk(struct intel_atomic_state *state) > > static int cnl_modeset_calc_cdclk(struct intel_atomic_state *state) > > { > > - struct drm_i915_private *dev_priv = to_i915(state->base.dev); > > + struct drm_i915_private *i915 = to_i915(state->base.dev); > > int min_cdclk, cdclk, vco; > > min_cdclk = intel_compute_min_cdclk(state); > > @@ -2510,7 +2510,7 @@ static int cnl_modeset_calc_cdclk(struct intel_atomic_state *state) > > return min_cdclk; > > cdclk = cnl_calc_cdclk(min_cdclk); > > - vco = cnl_cdclk_pll_vco(dev_priv, cdclk); > > + vco = cnl_cdclk_pll_vco(i915, cdclk); > > state->cdclk.logical.vco = vco; > > state->cdclk.logical.cdclk = cdclk; > > @@ -2520,7 +2520,7 @@ static int cnl_modeset_calc_cdclk(struct intel_atomic_state *state) > > if (!state->active_crtcs) { > > cdclk = cnl_calc_cdclk(state->cdclk.force_min_cdclk); > > - vco = cnl_cdclk_pll_vco(dev_priv, cdclk); > > + vco = cnl_cdclk_pll_vco(i915, cdclk); > > state->cdclk.actual.vco = vco; > > state->cdclk.actual.cdclk = cdclk; > > @@ -2535,7 +2535,7 @@ static int cnl_modeset_calc_cdclk(struct intel_atomic_state *state) > > static int icl_modeset_calc_cdclk(struct intel_atomic_state *state) > > { > > - struct drm_i915_private *dev_priv = to_i915(state->base.dev); > > + struct drm_i915_private *i915 = to_i915(state->base.dev); > > unsigned int ref = state->cdclk.logical.ref; > > int min_cdclk, cdclk, vco; > > @@ -2544,7 +2544,7 @@ static int icl_modeset_calc_cdclk(struct intel_atomic_state *state) > > return min_cdclk; > > cdclk = icl_calc_cdclk(min_cdclk, ref); > > - vco = icl_calc_cdclk_pll_vco(dev_priv, cdclk); > > + vco = icl_calc_cdclk_pll_vco(i915, cdclk); > > state->cdclk.logical.vco = vco; > > state->cdclk.logical.cdclk = cdclk; > > @@ -2554,7 +2554,7 @@ static int icl_modeset_calc_cdclk(struct intel_atomic_state *state) > > if (!state->active_crtcs) { > > cdclk = icl_calc_cdclk(state->cdclk.force_min_cdclk, ref); > > - vco = icl_calc_cdclk_pll_vco(dev_priv, cdclk); > > + vco = icl_calc_cdclk_pll_vco(i915, cdclk); > > state->cdclk.actual.vco = vco; > > state->cdclk.actual.cdclk = cdclk; > > @@ -2567,18 +2567,18 @@ static int icl_modeset_calc_cdclk(struct intel_atomic_state *state) > > return 0; > > } > > -static int intel_compute_max_dotclk(struct drm_i915_private *dev_priv) > > +static int intel_compute_max_dotclk(struct drm_i915_private *i915) > > { > > - int max_cdclk_freq = dev_priv->max_cdclk_freq; > > + int max_cdclk_freq = i915->max_cdclk_freq; > > - if (INTEL_GEN(dev_priv) >= 10 || IS_GEMINILAKE(dev_priv)) > > + if (INTEL_GEN(i915) >= 10 || IS_GEMINILAKE(i915)) > > return 2 * max_cdclk_freq; > > - else if (IS_GEN(dev_priv, 9) || > > - IS_BROADWELL(dev_priv) || IS_HASWELL(dev_priv)) > > + else if (IS_GEN(i915, 9) || > > + IS_BROADWELL(i915) || IS_HASWELL(i915)) > > return max_cdclk_freq; > > - else if (IS_CHERRYVIEW(dev_priv)) > > + else if (IS_CHERRYVIEW(i915)) > > return max_cdclk_freq*95/100; > > - else if (INTEL_GEN(dev_priv) < 4) > > + else if (INTEL_GEN(i915) < 4) > > return 2*max_cdclk_freq*90/100; > > else > > return max_cdclk_freq*90/100; > > @@ -2586,26 +2586,26 @@ static int intel_compute_max_dotclk(struct drm_i915_private *dev_priv) > > /** > > * intel_update_max_cdclk - Determine the maximum support CDCLK frequency > > - * @dev_priv: i915 device > > + * @i915: i915 device > > * > > * Determine the maximum CDCLK frequency the platform supports, and also > > * derive the maximum dot clock frequency the maximum CDCLK frequency > > * allows. > > */ > > -void intel_update_max_cdclk(struct drm_i915_private *dev_priv) > > +void intel_update_max_cdclk(struct drm_i915_private *i915) > > { > > - if (INTEL_GEN(dev_priv) >= 11) { > > - if (dev_priv->cdclk.hw.ref == 24000) > > - dev_priv->max_cdclk_freq = 648000; > > + if (INTEL_GEN(i915) >= 11) { > > + if (i915->cdclk.hw.ref == 24000) > > + i915->max_cdclk_freq = 648000; > > else > > - dev_priv->max_cdclk_freq = 652800; > > - } else if (IS_CANNONLAKE(dev_priv)) { > > - dev_priv->max_cdclk_freq = 528000; > > - } else if (IS_GEN9_BC(dev_priv)) { > > + i915->max_cdclk_freq = 652800; > > + } else if (IS_CANNONLAKE(i915)) { > > + i915->max_cdclk_freq = 528000; > > + } else if (IS_GEN9_BC(i915)) { > > u32 limit = I915_READ(SKL_DFSM) & SKL_DFSM_CDCLK_LIMIT_MASK; > > int max_cdclk, vco; > > - vco = dev_priv->skl_preferred_vco_freq; > > + vco = i915->skl_preferred_vco_freq; > > WARN_ON(vco != 8100000 && vco != 8640000); > > /* > > @@ -2622,12 +2622,12 @@ void intel_update_max_cdclk(struct drm_i915_private *dev_priv) > > else > > max_cdclk = 308571; > > - dev_priv->max_cdclk_freq = skl_calc_cdclk(max_cdclk, vco); > > - } else if (IS_GEMINILAKE(dev_priv)) { > > - dev_priv->max_cdclk_freq = 316800; > > - } else if (IS_BROXTON(dev_priv)) { > > - dev_priv->max_cdclk_freq = 624000; > > - } else if (IS_BROADWELL(dev_priv)) { > > + i915->max_cdclk_freq = skl_calc_cdclk(max_cdclk, vco); > > + } else if (IS_GEMINILAKE(i915)) { > > + i915->max_cdclk_freq = 316800; > > + } else if (IS_BROXTON(i915)) { > > + i915->max_cdclk_freq = 624000; > > + } else if (IS_BROADWELL(i915)) { > > /* > > * FIXME with extra cooling we can allow > > * 540 MHz for ULX and 675 Mhz for ULT. > > @@ -2635,40 +2635,40 @@ void intel_update_max_cdclk(struct drm_i915_private *dev_priv) > > * available? PCI ID, VTB, something else? > > */ > > if (I915_READ(FUSE_STRAP) & HSW_CDCLK_LIMIT) > > - dev_priv->max_cdclk_freq = 450000; > > - else if (IS_BDW_ULX(dev_priv)) > > - dev_priv->max_cdclk_freq = 450000; > > - else if (IS_BDW_ULT(dev_priv)) > > - dev_priv->max_cdclk_freq = 540000; > > + i915->max_cdclk_freq = 450000; > > + else if (IS_BDW_ULX(i915)) > > + i915->max_cdclk_freq = 450000; > > + else if (IS_BDW_ULT(i915)) > > + i915->max_cdclk_freq = 540000; > > else > > - dev_priv->max_cdclk_freq = 675000; > > - } else if (IS_CHERRYVIEW(dev_priv)) { > > - dev_priv->max_cdclk_freq = 320000; > > - } else if (IS_VALLEYVIEW(dev_priv)) { > > - dev_priv->max_cdclk_freq = 400000; > > + i915->max_cdclk_freq = 675000; > > + } else if (IS_CHERRYVIEW(i915)) { > > + i915->max_cdclk_freq = 320000; > > + } else if (IS_VALLEYVIEW(i915)) { > > + i915->max_cdclk_freq = 400000; > > } else { > > /* otherwise assume cdclk is fixed */ > > - dev_priv->max_cdclk_freq = dev_priv->cdclk.hw.cdclk; > > + i915->max_cdclk_freq = i915->cdclk.hw.cdclk; > > } > > - dev_priv->max_dotclk_freq = intel_compute_max_dotclk(dev_priv); > > + i915->max_dotclk_freq = intel_compute_max_dotclk(i915); > > DRM_DEBUG_DRIVER("Max CD clock rate: %d kHz\n", > > - dev_priv->max_cdclk_freq); > > + i915->max_cdclk_freq); > > DRM_DEBUG_DRIVER("Max dotclock rate: %d kHz\n", > > - dev_priv->max_dotclk_freq); > > + i915->max_dotclk_freq); > > } > > /** > > * intel_update_cdclk - Determine the current CDCLK frequency > > - * @dev_priv: i915 device > > + * @i915: i915 device > > * > > * Determine the current CDCLK frequency. > > */ > > -void intel_update_cdclk(struct drm_i915_private *dev_priv) > > +void intel_update_cdclk(struct drm_i915_private *i915) > > { > > - dev_priv->display.get_cdclk(dev_priv, &dev_priv->cdclk.hw); > > + i915->display.get_cdclk(i915, &i915->cdclk.hw); > > /* > > * 9:0 CMBUS [sic] CDCLK frequency (cdfreq): > > @@ -2676,12 +2676,12 @@ void intel_update_cdclk(struct drm_i915_private *dev_priv) > > * of cdclk that generates 4MHz reference clock freq which is used to > > * generate GMBus clock. This will vary with the cdclk freq. > > */ > > - if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) > > + if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) > > I915_WRITE(GMBUSFREQ_VLV, > > - DIV_ROUND_UP(dev_priv->cdclk.hw.cdclk, 1000)); > > + DIV_ROUND_UP(i915->cdclk.hw.cdclk, 1000)); > > } > > -static int cnp_rawclk(struct drm_i915_private *dev_priv) > > +static int cnp_rawclk(struct drm_i915_private *i915) > > { > > u32 rawclk; > > int divider, fraction; > > @@ -2702,7 +2702,7 @@ static int cnp_rawclk(struct drm_i915_private *dev_priv) > > rawclk |= CNP_RAWCLK_DEN(DIV_ROUND_CLOSEST(numerator * 1000, > > fraction) - 1); > > - if (INTEL_PCH_TYPE(dev_priv) >= PCH_ICP) > > + if (INTEL_PCH_TYPE(i915) >= PCH_ICP) > > rawclk |= ICP_RAWCLK_NUM(numerator); > > } > > @@ -2710,19 +2710,19 @@ static int cnp_rawclk(struct drm_i915_private *dev_priv) > > return divider + fraction; > > } > > -static int pch_rawclk(struct drm_i915_private *dev_priv) > > +static int pch_rawclk(struct drm_i915_private *i915) > > { > > return (I915_READ(PCH_RAWCLK_FREQ) & RAWCLK_FREQ_MASK) * 1000; > > } > > -static int vlv_hrawclk(struct drm_i915_private *dev_priv) > > +static int vlv_hrawclk(struct drm_i915_private *i915) > > { > > /* RAWCLK_FREQ_VLV register updated from power well code */ > > - return vlv_get_cck_clock_hpll(dev_priv, "hrawclk", > > + return vlv_get_cck_clock_hpll(i915, "hrawclk", > > CCK_DISPLAY_REF_CLOCK_CONTROL); > > } > > -static int g4x_hrawclk(struct drm_i915_private *dev_priv) > > +static int g4x_hrawclk(struct drm_i915_private *i915) > > { > > u32 clkcfg; > > @@ -2750,104 +2750,104 @@ static int g4x_hrawclk(struct drm_i915_private *dev_priv) > > /** > > * intel_update_rawclk - Determine the current RAWCLK frequency > > - * @dev_priv: i915 device > > + * @i915: i915 device > > * > > * Determine the current RAWCLK frequency. RAWCLK is a fixed > > * frequency clock so this needs to done only once. > > */ > > -void intel_update_rawclk(struct drm_i915_private *dev_priv) > > -{ > > - if (INTEL_PCH_TYPE(dev_priv) >= PCH_CNP) > > - dev_priv->rawclk_freq = cnp_rawclk(dev_priv); > > - else if (HAS_PCH_SPLIT(dev_priv)) > > - dev_priv->rawclk_freq = pch_rawclk(dev_priv); > > - else if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) > > - dev_priv->rawclk_freq = vlv_hrawclk(dev_priv); > > - else if (IS_G4X(dev_priv) || IS_PINEVIEW(dev_priv)) > > - dev_priv->rawclk_freq = g4x_hrawclk(dev_priv); > > +void intel_update_rawclk(struct drm_i915_private *i915) > > +{ > > + if (INTEL_PCH_TYPE(i915) >= PCH_CNP) > > + i915->rawclk_freq = cnp_rawclk(i915); > > + else if (HAS_PCH_SPLIT(i915)) > > + i915->rawclk_freq = pch_rawclk(i915); > > + else if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) > > + i915->rawclk_freq = vlv_hrawclk(i915); > > + else if (IS_G4X(i915) || IS_PINEVIEW(i915)) > > + i915->rawclk_freq = g4x_hrawclk(i915); > > else > > /* no rawclk on other platforms, or no need to know it */ > > return; > > - DRM_DEBUG_DRIVER("rawclk rate: %d kHz\n", dev_priv->rawclk_freq); > > + DRM_DEBUG_DRIVER("rawclk rate: %d kHz\n", i915->rawclk_freq); > > } > > /** > > * intel_init_cdclk_hooks - Initialize CDCLK related modesetting hooks > > - * @dev_priv: i915 device > > + * @i915: i915 device > > */ > > -void intel_init_cdclk_hooks(struct drm_i915_private *dev_priv) > > -{ > > - if (INTEL_GEN(dev_priv) >= 11) { > > - dev_priv->display.set_cdclk = icl_set_cdclk; > > - dev_priv->display.modeset_calc_cdclk = icl_modeset_calc_cdclk; > > - } else if (IS_CANNONLAKE(dev_priv)) { > > - dev_priv->display.set_cdclk = cnl_set_cdclk; > > - dev_priv->display.modeset_calc_cdclk = cnl_modeset_calc_cdclk; > > - } else if (IS_GEN9_LP(dev_priv)) { > > - dev_priv->display.set_cdclk = bxt_set_cdclk; > > - dev_priv->display.modeset_calc_cdclk = bxt_modeset_calc_cdclk; > > - } else if (IS_GEN9_BC(dev_priv)) { > > - dev_priv->display.set_cdclk = skl_set_cdclk; > > - dev_priv->display.modeset_calc_cdclk = skl_modeset_calc_cdclk; > > - } else if (IS_BROADWELL(dev_priv)) { > > - dev_priv->display.set_cdclk = bdw_set_cdclk; > > - dev_priv->display.modeset_calc_cdclk = bdw_modeset_calc_cdclk; > > - } else if (IS_CHERRYVIEW(dev_priv)) { > > - dev_priv->display.set_cdclk = chv_set_cdclk; > > - dev_priv->display.modeset_calc_cdclk = vlv_modeset_calc_cdclk; > > - } else if (IS_VALLEYVIEW(dev_priv)) { > > - dev_priv->display.set_cdclk = vlv_set_cdclk; > > - dev_priv->display.modeset_calc_cdclk = vlv_modeset_calc_cdclk; > > +void intel_init_cdclk_hooks(struct drm_i915_private *i915) > > +{ > > + if (INTEL_GEN(i915) >= 11) { > > + i915->display.set_cdclk = icl_set_cdclk; > > + i915->display.modeset_calc_cdclk = icl_modeset_calc_cdclk; > > + } else if (IS_CANNONLAKE(i915)) { > > + i915->display.set_cdclk = cnl_set_cdclk; > > + i915->display.modeset_calc_cdclk = cnl_modeset_calc_cdclk; > > + } else if (IS_GEN9_LP(i915)) { > > + i915->display.set_cdclk = bxt_set_cdclk; > > + i915->display.modeset_calc_cdclk = bxt_modeset_calc_cdclk; > > + } else if (IS_GEN9_BC(i915)) { > > + i915->display.set_cdclk = skl_set_cdclk; > > + i915->display.modeset_calc_cdclk = skl_modeset_calc_cdclk; > > + } else if (IS_BROADWELL(i915)) { > > + i915->display.set_cdclk = bdw_set_cdclk; > > + i915->display.modeset_calc_cdclk = bdw_modeset_calc_cdclk; > > + } else if (IS_CHERRYVIEW(i915)) { > > + i915->display.set_cdclk = chv_set_cdclk; > > + i915->display.modeset_calc_cdclk = vlv_modeset_calc_cdclk; > > + } else if (IS_VALLEYVIEW(i915)) { > > + i915->display.set_cdclk = vlv_set_cdclk; > > + i915->display.modeset_calc_cdclk = vlv_modeset_calc_cdclk; > > } > > - if (INTEL_GEN(dev_priv) >= 11) > > - dev_priv->display.get_cdclk = icl_get_cdclk; > > - else if (IS_CANNONLAKE(dev_priv)) > > - dev_priv->display.get_cdclk = cnl_get_cdclk; > > - else if (IS_GEN9_LP(dev_priv)) > > - dev_priv->display.get_cdclk = bxt_get_cdclk; > > - else if (IS_GEN9_BC(dev_priv)) > > - dev_priv->display.get_cdclk = skl_get_cdclk; > > - else if (IS_BROADWELL(dev_priv)) > > - dev_priv->display.get_cdclk = bdw_get_cdclk; > > - else if (IS_HASWELL(dev_priv)) > > - dev_priv->display.get_cdclk = hsw_get_cdclk; > > - else if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) > > - dev_priv->display.get_cdclk = vlv_get_cdclk; > > - else if (IS_GEN(dev_priv, 6) || IS_IVYBRIDGE(dev_priv)) > > - dev_priv->display.get_cdclk = fixed_400mhz_get_cdclk; > > - else if (IS_GEN(dev_priv, 5)) > > - dev_priv->display.get_cdclk = fixed_450mhz_get_cdclk; > > - else if (IS_GM45(dev_priv)) > > - dev_priv->display.get_cdclk = gm45_get_cdclk; > > - else if (IS_G45(dev_priv)) > > - dev_priv->display.get_cdclk = g33_get_cdclk; > > - else if (IS_I965GM(dev_priv)) > > - dev_priv->display.get_cdclk = i965gm_get_cdclk; > > - else if (IS_I965G(dev_priv)) > > - dev_priv->display.get_cdclk = fixed_400mhz_get_cdclk; > > - else if (IS_PINEVIEW(dev_priv)) > > - dev_priv->display.get_cdclk = pnv_get_cdclk; > > - else if (IS_G33(dev_priv)) > > - dev_priv->display.get_cdclk = g33_get_cdclk; > > - else if (IS_I945GM(dev_priv)) > > - dev_priv->display.get_cdclk = i945gm_get_cdclk; > > - else if (IS_I945G(dev_priv)) > > - dev_priv->display.get_cdclk = fixed_400mhz_get_cdclk; > > - else if (IS_I915GM(dev_priv)) > > - dev_priv->display.get_cdclk = i915gm_get_cdclk; > > - else if (IS_I915G(dev_priv)) > > - dev_priv->display.get_cdclk = fixed_333mhz_get_cdclk; > > - else if (IS_I865G(dev_priv)) > > - dev_priv->display.get_cdclk = fixed_266mhz_get_cdclk; > > - else if (IS_I85X(dev_priv)) > > - dev_priv->display.get_cdclk = i85x_get_cdclk; > > - else if (IS_I845G(dev_priv)) > > - dev_priv->display.get_cdclk = fixed_200mhz_get_cdclk; > > + if (INTEL_GEN(i915) >= 11) > > + i915->display.get_cdclk = icl_get_cdclk; > > + else if (IS_CANNONLAKE(i915)) > > + i915->display.get_cdclk = cnl_get_cdclk; > > + else if (IS_GEN9_LP(i915)) > > + i915->display.get_cdclk = bxt_get_cdclk; > > + else if (IS_GEN9_BC(i915)) > > + i915->display.get_cdclk = skl_get_cdclk; > > + else if (IS_BROADWELL(i915)) > > + i915->display.get_cdclk = bdw_get_cdclk; > > + else if (IS_HASWELL(i915)) > > + i915->display.get_cdclk = hsw_get_cdclk; > > + else if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) > > + i915->display.get_cdclk = vlv_get_cdclk; > > + else if (IS_GEN(i915, 6) || IS_IVYBRIDGE(i915)) > > + i915->display.get_cdclk = fixed_400mhz_get_cdclk; > > + else if (IS_GEN(i915, 5)) > > + i915->display.get_cdclk = fixed_450mhz_get_cdclk; > > + else if (IS_GM45(i915)) > > + i915->display.get_cdclk = gm45_get_cdclk; > > + else if (IS_G45(i915)) > > + i915->display.get_cdclk = g33_get_cdclk; > > + else if (IS_I965GM(i915)) > > + i915->display.get_cdclk = i965gm_get_cdclk; > > + else if (IS_I965G(i915)) > > + i915->display.get_cdclk = fixed_400mhz_get_cdclk; > > + else if (IS_PINEVIEW(i915)) > > + i915->display.get_cdclk = pnv_get_cdclk; > > + else if (IS_G33(i915)) > > + i915->display.get_cdclk = g33_get_cdclk; > > + else if (IS_I945GM(i915)) > > + i915->display.get_cdclk = i945gm_get_cdclk; > > + else if (IS_I945G(i915)) > > + i915->display.get_cdclk = fixed_400mhz_get_cdclk; > > + else if (IS_I915GM(i915)) > > + i915->display.get_cdclk = i915gm_get_cdclk; > > + else if (IS_I915G(i915)) > > + i915->display.get_cdclk = fixed_333mhz_get_cdclk; > > + else if (IS_I865G(i915)) > > + i915->display.get_cdclk = fixed_266mhz_get_cdclk; > > + else if (IS_I85X(i915)) > > + i915->display.get_cdclk = i85x_get_cdclk; > > + else if (IS_I845G(i915)) > > + i915->display.get_cdclk = fixed_200mhz_get_cdclk; > > else { /* 830 */ > > - WARN(!IS_I830(dev_priv), > > + WARN(!IS_I830(i915), > > "Unknown platform. Assuming 133 MHz CDCLK\n"); > > - dev_priv->display.get_cdclk = fixed_133mhz_get_cdclk; > > + i915->display.get_cdclk = fixed_133mhz_get_cdclk; > > } > > } > > diff --git a/drivers/gpu/drm/i915/intel_cdclk.h b/drivers/gpu/drm/i915/intel_cdclk.h > > index 4d6f7f5f8930..4e17102af66c 100644 > > --- a/drivers/gpu/drm/i915/intel_cdclk.h > > +++ b/drivers/gpu/drm/i915/intel_cdclk.h > > @@ -18,11 +18,11 @@ struct intel_crtc_state; > > int intel_crtc_compute_min_cdclk(const struct intel_crtc_state *crtc_state); > > void intel_cdclk_init(struct drm_i915_private *i915); > > void intel_cdclk_uninit(struct drm_i915_private *i915); > > -void intel_init_cdclk_hooks(struct drm_i915_private *dev_priv); > > -void intel_update_max_cdclk(struct drm_i915_private *dev_priv); > > -void intel_update_cdclk(struct drm_i915_private *dev_priv); > > -void intel_update_rawclk(struct drm_i915_private *dev_priv); > > -bool intel_cdclk_needs_cd2x_update(struct drm_i915_private *dev_priv, > > +void intel_init_cdclk_hooks(struct drm_i915_private *i915); > > +void intel_update_max_cdclk(struct drm_i915_private *i915); > > +void intel_update_cdclk(struct drm_i915_private *i915); > > +void intel_update_rawclk(struct drm_i915_private *i915); > > +bool intel_cdclk_needs_cd2x_update(struct drm_i915_private *i915, > > const struct intel_cdclk_state *a, > > const struct intel_cdclk_state *b); > > bool intel_cdclk_needs_modeset(const struct intel_cdclk_state *a, > > @@ -31,12 +31,12 @@ bool intel_cdclk_changed(const struct intel_cdclk_state *a, > > const struct intel_cdclk_state *b); > > void intel_cdclk_swap_state(struct intel_atomic_state *state); > > void > > -intel_set_cdclk_pre_plane_update(struct drm_i915_private *dev_priv, > > +intel_set_cdclk_pre_plane_update(struct drm_i915_private *i915, > > const struct intel_cdclk_state *old_state, > > const struct intel_cdclk_state *new_state, > > enum pipe pipe); > > void > > -intel_set_cdclk_post_plane_update(struct drm_i915_private *dev_priv, > > +intel_set_cdclk_post_plane_update(struct drm_i915_private *i915, > > const struct intel_cdclk_state *old_state, > > const struct intel_cdclk_state *new_state, > > enum pipe pipe); > > diff --git a/drivers/gpu/drm/i915/intel_color.c b/drivers/gpu/drm/i915/intel_color.c > > index 0b8cf3e8c963..efc67e4a0ba5 100644 > > --- a/drivers/gpu/drm/i915/intel_color.c > > +++ b/drivers/gpu/drm/i915/intel_color.c > > @@ -137,7 +137,7 @@ static void ilk_update_pipe_csc(struct intel_crtc *crtc, > > const u16 coeff[9], > > const u16 postoff[3]) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > enum pipe pipe = crtc->pipe; > > I915_WRITE(PIPE_CSC_PREOFF_HI(pipe), preoff[0]); > > @@ -153,7 +153,7 @@ static void ilk_update_pipe_csc(struct intel_crtc *crtc, > > I915_WRITE(PIPE_CSC_COEFF_RV_GV(pipe), coeff[6] << 16 | coeff[7]); > > I915_WRITE(PIPE_CSC_COEFF_BV(pipe), coeff[8] << 16); > > - if (INTEL_GEN(dev_priv) >= 7) { > > + if (INTEL_GEN(i915) >= 7) { > > I915_WRITE(PIPE_CSC_POSTOFF_HI(pipe), postoff[0]); > > I915_WRITE(PIPE_CSC_POSTOFF_ME(pipe), postoff[1]); > > I915_WRITE(PIPE_CSC_POSTOFF_LO(pipe), postoff[2]); > > @@ -165,7 +165,7 @@ static void icl_update_output_csc(struct intel_crtc *crtc, > > const u16 coeff[9], > > const u16 postoff[3]) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > enum pipe pipe = crtc->pipe; > > I915_WRITE(PIPE_CSC_OUTPUT_PREOFF_HI(pipe), preoff[0]); > > @@ -188,15 +188,15 @@ static void icl_update_output_csc(struct intel_crtc *crtc, > > static bool ilk_csc_limited_range(const struct intel_crtc_state *crtc_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc_state->base.crtc->dev); > > + struct drm_i915_private *i915 = to_i915(crtc_state->base.crtc->dev); > > /* > > * FIXME if there's a gamma LUT after the CSC, we should > > * do the range compression using the gamma LUT instead. > > */ > > return crtc_state->limited_color_range && > > - (IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv) || > > - IS_GEN_RANGE(dev_priv, 9, 10)); > > + (IS_HASWELL(i915) || IS_BROADWELL(i915) || > > + IS_GEN_RANGE(i915, 9, 10)); > > } > > static void ilk_csc_convert_ctm(const struct intel_crtc_state *crtc_state, > > @@ -254,7 +254,7 @@ static void ilk_csc_convert_ctm(const struct intel_crtc_state *crtc_state, > > static void ilk_load_csc_matrix(const struct intel_crtc_state *crtc_state) > > { > > struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc); > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > bool limited_color_range = ilk_csc_limited_range(crtc_state); > > if (crtc_state->base.ctm) { > > @@ -280,7 +280,7 @@ static void ilk_load_csc_matrix(const struct intel_crtc_state *crtc_state) > > * LUT is needed but CSC is not we need to load an > > * identity matrix. > > */ > > - WARN_ON(!IS_CANNONLAKE(dev_priv) && !IS_GEMINILAKE(dev_priv)); > > + WARN_ON(!IS_CANNONLAKE(i915) && !IS_GEMINILAKE(i915)); > > ilk_update_pipe_csc(crtc, ilk_csc_off_zero, > > ilk_csc_coeff_identity, > > @@ -293,7 +293,7 @@ static void ilk_load_csc_matrix(const struct intel_crtc_state *crtc_state) > > static void icl_load_csc_matrix(const struct intel_crtc_state *crtc_state) > > { > > struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc); > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > if (crtc_state->base.ctm) { > > u16 coeff[9]; > > @@ -322,7 +322,7 @@ static void icl_load_csc_matrix(const struct intel_crtc_state *crtc_state) > > static void cherryview_load_csc_matrix(const struct intel_crtc_state *crtc_state) > > { > > struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc); > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > enum pipe pipe = crtc->pipe; > > if (crtc_state->base.ctm) { > > @@ -388,15 +388,15 @@ static void i9xx_load_luts_internal(const struct intel_crtc_state *crtc_state, > > const struct drm_property_blob *blob) > > { > > struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc); > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > enum pipe pipe = crtc->pipe; > > int i; > > - if (HAS_GMCH(dev_priv)) { > > + if (HAS_GMCH(i915)) { > > if (intel_crtc_has_type(crtc_state, INTEL_OUTPUT_DSI)) > > - assert_dsi_pll_enabled(dev_priv); > > + assert_dsi_pll_enabled(i915); > > else > > - assert_pll_enabled(dev_priv, pipe); > > + assert_pll_enabled(i915, pipe); > > } > > if (blob) { > > @@ -408,7 +408,7 @@ static void i9xx_load_luts_internal(const struct intel_crtc_state *crtc_state, > > (drm_color_lut_extract(lut[i].green, 8) << 8) | > > drm_color_lut_extract(lut[i].blue, 8); > > - if (HAS_GMCH(dev_priv)) > > + if (HAS_GMCH(i915)) > > I915_WRITE(PALETTE(pipe, i), word); > > else > > I915_WRITE(LGC_PALETTE(pipe, i), word); > > @@ -424,7 +424,7 @@ static void i9xx_load_luts(const struct intel_crtc_state *crtc_state) > > static void i9xx_color_commit(const struct intel_crtc_state *crtc_state) > > { > > struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc); > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > enum pipe pipe = crtc->pipe; > > u32 val; > > @@ -437,7 +437,7 @@ static void i9xx_color_commit(const struct intel_crtc_state *crtc_state) > > static void ilk_color_commit(const struct intel_crtc_state *crtc_state) > > { > > struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc); > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > enum pipe pipe = crtc->pipe; > > u32 val; > > @@ -452,7 +452,7 @@ static void ilk_color_commit(const struct intel_crtc_state *crtc_state) > > static void hsw_color_commit(const struct intel_crtc_state *crtc_state) > > { > > struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc); > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > I915_WRITE(GAMMA_MODE(crtc->pipe), crtc_state->gamma_mode); > > @@ -462,7 +462,7 @@ static void hsw_color_commit(const struct intel_crtc_state *crtc_state) > > static void skl_color_commit(const struct intel_crtc_state *crtc_state) > > { > > struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc); > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > enum pipe pipe = crtc->pipe; > > u32 val = 0; > > @@ -479,7 +479,7 @@ static void skl_color_commit(const struct intel_crtc_state *crtc_state) > > I915_WRITE(GAMMA_MODE(crtc->pipe), crtc_state->gamma_mode); > > - if (INTEL_GEN(dev_priv) >= 11) > > + if (INTEL_GEN(i915) >= 11) > > icl_load_csc_matrix(crtc_state); > > else > > ilk_load_csc_matrix(crtc_state); > > @@ -488,7 +488,7 @@ static void skl_color_commit(const struct intel_crtc_state *crtc_state) > > static void i965_load_lut_10p6(struct intel_crtc *crtc, > > const struct drm_property_blob *blob) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > const struct drm_color_lut *lut = blob->data; > > int i, lut_size = drm_color_lut_size(blob); > > enum pipe pipe = crtc->pipe; > > @@ -519,7 +519,7 @@ static void i965_load_luts(const struct intel_crtc_state *crtc_state) > > static void ilk_load_lut_10(struct intel_crtc *crtc, > > const struct drm_property_blob *blob) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > const struct drm_color_lut *lut = blob->data; > > int i, lut_size = drm_color_lut_size(blob); > > enum pipe pipe = crtc->pipe; > > @@ -556,7 +556,7 @@ static void ivb_load_lut_10(struct intel_crtc *crtc, > > const struct drm_property_blob *blob, > > u32 prec_index) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > int hw_lut_size = ivb_lut_10_size(prec_index); > > const struct drm_color_lut *lut = blob->data; > > int i, lut_size = drm_color_lut_size(blob); > > @@ -583,7 +583,7 @@ static void bdw_load_lut_10(struct intel_crtc *crtc, > > const struct drm_property_blob *blob, > > u32 prec_index) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > int hw_lut_size = ivb_lut_10_size(prec_index); > > const struct drm_color_lut *lut = blob->data; > > int i, lut_size = drm_color_lut_size(blob); > > @@ -609,7 +609,7 @@ static void bdw_load_lut_10(struct intel_crtc *crtc, > > static void ivb_load_lut_10_max(struct intel_crtc *crtc) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > enum pipe pipe = crtc->pipe; > > /* Program the max register to clamp values > 1.0. */ > > @@ -622,7 +622,7 @@ static void ivb_load_lut_10_max(struct intel_crtc *crtc) > > * ToDo: Extend the ABI to be able to program values > > * from 3.0 to 7.0 > > */ > > - if (INTEL_GEN(dev_priv) >= 10 || IS_GEMINILAKE(dev_priv)) { > > + if (INTEL_GEN(i915) >= 10 || IS_GEMINILAKE(i915)) { > > I915_WRITE(PREC_PAL_EXT2_GC_MAX(pipe, 0), 1 << 16); > > I915_WRITE(PREC_PAL_EXT2_GC_MAX(pipe, 1), 1 << 16); > > I915_WRITE(PREC_PAL_EXT2_GC_MAX(pipe, 2), 1 << 16); > > @@ -678,9 +678,9 @@ static void bdw_load_luts(const struct intel_crtc_state *crtc_state) > > static void glk_load_degamma_lut(const struct intel_crtc_state *crtc_state) > > { > > struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc); > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > enum pipe pipe = crtc->pipe; > > - const u32 lut_size = INTEL_INFO(dev_priv)->color.degamma_lut_size; > > + const u32 lut_size = INTEL_INFO(i915)->color.degamma_lut_size; > > const struct drm_color_lut *lut = crtc_state->base.degamma_lut->data; > > u32 i; > > @@ -717,9 +717,9 @@ static void glk_load_degamma_lut(const struct intel_crtc_state *crtc_state) > > static void glk_load_degamma_lut_linear(const struct intel_crtc_state *crtc_state) > > { > > struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc); > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > enum pipe pipe = crtc->pipe; > > - const u32 lut_size = INTEL_INFO(dev_priv)->color.degamma_lut_size; > > + const u32 lut_size = INTEL_INFO(i915)->color.degamma_lut_size; > > u32 i; > > /* > > @@ -798,7 +798,7 @@ static u32 chv_cgm_degamma_udw(const struct drm_color_lut *color) > > static void chv_load_cgm_degamma(struct intel_crtc *crtc, > > const struct drm_property_blob *blob) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > const struct drm_color_lut *lut = blob->data; > > int i, lut_size = drm_color_lut_size(blob); > > enum pipe pipe = crtc->pipe; > > @@ -825,7 +825,7 @@ static u32 chv_cgm_gamma_udw(const struct drm_color_lut *color) > > static void chv_load_cgm_gamma(struct intel_crtc *crtc, > > const struct drm_property_blob *blob) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > const struct drm_color_lut *lut = blob->data; > > int i, lut_size = drm_color_lut_size(blob); > > enum pipe pipe = crtc->pipe; > > @@ -860,37 +860,37 @@ static void chv_load_luts(const struct intel_crtc_state *crtc_state) > > void intel_color_load_luts(const struct intel_crtc_state *crtc_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc_state->base.crtc->dev); > > + struct drm_i915_private *i915 = to_i915(crtc_state->base.crtc->dev); > > - dev_priv->display.load_luts(crtc_state); > > + i915->display.load_luts(crtc_state); > > } > > void intel_color_commit(const struct intel_crtc_state *crtc_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc_state->base.crtc->dev); > > + struct drm_i915_private *i915 = to_i915(crtc_state->base.crtc->dev); > > - dev_priv->display.color_commit(crtc_state); > > + i915->display.color_commit(crtc_state); > > } > > int intel_color_check(struct intel_crtc_state *crtc_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc_state->base.crtc->dev); > > + struct drm_i915_private *i915 = to_i915(crtc_state->base.crtc->dev); > > - return dev_priv->display.color_check(crtc_state); > > + return i915->display.color_check(crtc_state); > > } > > void intel_color_get_config(struct intel_crtc_state *crtc_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc_state->base.crtc->dev); > > + struct drm_i915_private *i915 = to_i915(crtc_state->base.crtc->dev); > > - if (dev_priv->display.read_luts) > > - dev_priv->display.read_luts(crtc_state); > > + if (i915->display.read_luts) > > + i915->display.read_luts(crtc_state); > > } > > static bool need_plane_update(struct intel_plane *plane, > > const struct intel_crtc_state *crtc_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(plane->base.dev); > > + struct drm_i915_private *i915 = to_i915(plane->base.dev); > > /* > > * On pre-SKL the pipe gamma enable and pipe csc enable for > > @@ -898,7 +898,7 @@ static bool need_plane_update(struct intel_plane *plane, > > * We have to reconfigure that even if the plane is inactive. > > */ > > return crtc_state->active_planes & BIT(plane->id) || > > - (INTEL_GEN(dev_priv) < 9 && > > + (INTEL_GEN(i915) < 9 && > > plane->id == PLANE_PRIMARY); > > } > > @@ -906,7 +906,7 @@ static int > > intel_color_add_affected_planes(struct intel_crtc_state *new_crtc_state) > > { > > struct intel_crtc *crtc = to_intel_crtc(new_crtc_state->base.crtc); > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > struct intel_atomic_state *state = > > to_intel_atomic_state(new_crtc_state->base.state); > > const struct intel_crtc_state *old_crtc_state = > > @@ -921,7 +921,7 @@ intel_color_add_affected_planes(struct intel_crtc_state *new_crtc_state) > > new_crtc_state->csc_enable == old_crtc_state->csc_enable) > > return 0; > > - for_each_intel_plane_on_crtc(&dev_priv->drm, crtc, plane) { > > + for_each_intel_plane_on_crtc(&i915->drm, crtc, plane) { > > struct intel_plane_state *plane_state; > > if (!need_plane_update(plane, new_crtc_state)) > > @@ -956,7 +956,7 @@ static int check_lut_size(const struct drm_property_blob *lut, int expected) > > static int check_luts(const struct intel_crtc_state *crtc_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc_state->base.crtc->dev); > > + struct drm_i915_private *i915 = to_i915(crtc_state->base.crtc->dev); > > const struct drm_property_blob *gamma_lut = crtc_state->base.gamma_lut; > > const struct drm_property_blob *degamma_lut = crtc_state->base.degamma_lut; > > int gamma_length, degamma_length; > > @@ -972,10 +972,10 @@ static int check_luts(const struct intel_crtc_state *crtc_state) > > return -EINVAL; > > } > > - degamma_length = INTEL_INFO(dev_priv)->color.degamma_lut_size; > > - gamma_length = INTEL_INFO(dev_priv)->color.gamma_lut_size; > > - degamma_tests = INTEL_INFO(dev_priv)->color.degamma_lut_tests; > > - gamma_tests = INTEL_INFO(dev_priv)->color.gamma_lut_tests; > > + degamma_length = INTEL_INFO(i915)->color.degamma_lut_size; > > + gamma_length = INTEL_INFO(i915)->color.gamma_lut_size; > > + degamma_tests = INTEL_INFO(i915)->color.degamma_lut_tests; > > + gamma_tests = INTEL_INFO(i915)->color.gamma_lut_tests; > > if (check_lut_size(degamma_lut, degamma_length) || > > check_lut_size(gamma_lut, gamma_length)) > > @@ -1255,56 +1255,56 @@ static int icl_color_check(struct intel_crtc_state *crtc_state) > > void intel_color_init(struct intel_crtc *crtc) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > - bool has_ctm = INTEL_INFO(dev_priv)->color.degamma_lut_size != 0; > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > + bool has_ctm = INTEL_INFO(i915)->color.degamma_lut_size != 0; > > drm_mode_crtc_set_gamma_size(&crtc->base, 256); > > - if (HAS_GMCH(dev_priv)) { > > - if (IS_CHERRYVIEW(dev_priv)) { > > - dev_priv->display.color_check = chv_color_check; > > - dev_priv->display.color_commit = i9xx_color_commit; > > - dev_priv->display.load_luts = chv_load_luts; > > - } else if (INTEL_GEN(dev_priv) >= 4) { > > - dev_priv->display.color_check = i9xx_color_check; > > - dev_priv->display.color_commit = i9xx_color_commit; > > - dev_priv->display.load_luts = i965_load_luts; > > + if (HAS_GMCH(i915)) { > > + if (IS_CHERRYVIEW(i915)) { > > + i915->display.color_check = chv_color_check; > > + i915->display.color_commit = i9xx_color_commit; > > + i915->display.load_luts = chv_load_luts; > > + } else if (INTEL_GEN(i915) >= 4) { > > + i915->display.color_check = i9xx_color_check; > > + i915->display.color_commit = i9xx_color_commit; > > + i915->display.load_luts = i965_load_luts; > > } else { > > - dev_priv->display.color_check = i9xx_color_check; > > - dev_priv->display.color_commit = i9xx_color_commit; > > - dev_priv->display.load_luts = i9xx_load_luts; > > + i915->display.color_check = i9xx_color_check; > > + i915->display.color_commit = i9xx_color_commit; > > + i915->display.load_luts = i9xx_load_luts; > > } > > } else { > > - if (INTEL_GEN(dev_priv) >= 11) > > - dev_priv->display.color_check = icl_color_check; > > - else if (INTEL_GEN(dev_priv) >= 10 || IS_GEMINILAKE(dev_priv)) > > - dev_priv->display.color_check = glk_color_check; > > - else if (INTEL_GEN(dev_priv) >= 7) > > - dev_priv->display.color_check = ivb_color_check; > > + if (INTEL_GEN(i915) >= 11) > > + i915->display.color_check = icl_color_check; > > + else if (INTEL_GEN(i915) >= 10 || IS_GEMINILAKE(i915)) > > + i915->display.color_check = glk_color_check; > > + else if (INTEL_GEN(i915) >= 7) > > + i915->display.color_check = ivb_color_check; > > else > > - dev_priv->display.color_check = ilk_color_check; > > + i915->display.color_check = ilk_color_check; > > - if (INTEL_GEN(dev_priv) >= 9) > > - dev_priv->display.color_commit = skl_color_commit; > > - else if (IS_BROADWELL(dev_priv) || IS_HASWELL(dev_priv)) > > - dev_priv->display.color_commit = hsw_color_commit; > > + if (INTEL_GEN(i915) >= 9) > > + i915->display.color_commit = skl_color_commit; > > + else if (IS_BROADWELL(i915) || IS_HASWELL(i915)) > > + i915->display.color_commit = hsw_color_commit; > > else > > - dev_priv->display.color_commit = ilk_color_commit; > > - > > - if (INTEL_GEN(dev_priv) >= 11) > > - dev_priv->display.load_luts = icl_load_luts; > > - else if (IS_CANNONLAKE(dev_priv) || IS_GEMINILAKE(dev_priv)) > > - dev_priv->display.load_luts = glk_load_luts; > > - else if (INTEL_GEN(dev_priv) >= 8) > > - dev_priv->display.load_luts = bdw_load_luts; > > - else if (INTEL_GEN(dev_priv) >= 7) > > - dev_priv->display.load_luts = ivb_load_luts; > > + i915->display.color_commit = ilk_color_commit; > > + > > + if (INTEL_GEN(i915) >= 11) > > + i915->display.load_luts = icl_load_luts; > > + else if (IS_CANNONLAKE(i915) || IS_GEMINILAKE(i915)) > > + i915->display.load_luts = glk_load_luts; > > + else if (INTEL_GEN(i915) >= 8) > > + i915->display.load_luts = bdw_load_luts; > > + else if (INTEL_GEN(i915) >= 7) > > + i915->display.load_luts = ivb_load_luts; > > else > > - dev_priv->display.load_luts = ilk_load_luts; > > + i915->display.load_luts = ilk_load_luts; > > } > > drm_crtc_enable_color_mgmt(&crtc->base, > > - INTEL_INFO(dev_priv)->color.degamma_lut_size, > > + INTEL_INFO(i915)->color.degamma_lut_size, > > has_ctm, > > - INTEL_INFO(dev_priv)->color.gamma_lut_size); > > + INTEL_INFO(i915)->color.gamma_lut_size); > > } > > diff --git a/drivers/gpu/drm/i915/intel_combo_phy.c b/drivers/gpu/drm/i915/intel_combo_phy.c > > index 841708da5a56..0e7c18d783ca 100644 > > --- a/drivers/gpu/drm/i915/intel_combo_phy.c > > +++ b/drivers/gpu/drm/i915/intel_combo_phy.c > > @@ -44,7 +44,7 @@ static const struct cnl_procmon { > > * on its name. > > */ > > static const struct cnl_procmon * > > -cnl_get_procmon_ref_values(struct drm_i915_private *dev_priv, enum port port) > > +cnl_get_procmon_ref_values(struct drm_i915_private *i915, enum port port) > > { > > const struct cnl_procmon *procmon; > > u32 val; > > @@ -74,13 +74,13 @@ cnl_get_procmon_ref_values(struct drm_i915_private *dev_priv, enum port port) > > return procmon; > > } > > -static void cnl_set_procmon_ref_values(struct drm_i915_private *dev_priv, > > +static void cnl_set_procmon_ref_values(struct drm_i915_private *i915, > > enum port port) > > { > > const struct cnl_procmon *procmon; > > u32 val; > > - procmon = cnl_get_procmon_ref_values(dev_priv, port); > > + procmon = cnl_get_procmon_ref_values(i915, port); > > val = I915_READ(ICL_PORT_COMP_DW1(port)); > > val &= ~((0xff << 16) | 0xff); > > @@ -91,7 +91,7 @@ static void cnl_set_procmon_ref_values(struct drm_i915_private *dev_priv, > > I915_WRITE(ICL_PORT_COMP_DW10(port), procmon->dw10); > > } > > -static bool check_phy_reg(struct drm_i915_private *dev_priv, > > +static bool check_phy_reg(struct drm_i915_private *i915, > > enum port port, i915_reg_t reg, u32 mask, > > u32 expected_val) > > { > > @@ -108,47 +108,47 @@ static bool check_phy_reg(struct drm_i915_private *dev_priv, > > return true; > > } > > -static bool cnl_verify_procmon_ref_values(struct drm_i915_private *dev_priv, > > +static bool cnl_verify_procmon_ref_values(struct drm_i915_private *i915, > > enum port port) > > { > > const struct cnl_procmon *procmon; > > bool ret; > > - procmon = cnl_get_procmon_ref_values(dev_priv, port); > > + procmon = cnl_get_procmon_ref_values(i915, port); > > - ret = check_phy_reg(dev_priv, port, ICL_PORT_COMP_DW1(port), > > + ret = check_phy_reg(i915, port, ICL_PORT_COMP_DW1(port), > > (0xff << 16) | 0xff, procmon->dw1); > > - ret &= check_phy_reg(dev_priv, port, ICL_PORT_COMP_DW9(port), > > + ret &= check_phy_reg(i915, port, ICL_PORT_COMP_DW9(port), > > -1U, procmon->dw9); > > - ret &= check_phy_reg(dev_priv, port, ICL_PORT_COMP_DW10(port), > > + ret &= check_phy_reg(i915, port, ICL_PORT_COMP_DW10(port), > > -1U, procmon->dw10); > > return ret; > > } > > -static bool cnl_combo_phy_enabled(struct drm_i915_private *dev_priv) > > +static bool cnl_combo_phy_enabled(struct drm_i915_private *i915) > > { > > return !(I915_READ(CHICKEN_MISC_2) & CNL_COMP_PWR_DOWN) && > > (I915_READ(CNL_PORT_COMP_DW0) & COMP_INIT); > > } > > -static bool cnl_combo_phy_verify_state(struct drm_i915_private *dev_priv) > > +static bool cnl_combo_phy_verify_state(struct drm_i915_private *i915) > > { > > enum port port = PORT_A; > > bool ret; > > - if (!cnl_combo_phy_enabled(dev_priv)) > > + if (!cnl_combo_phy_enabled(i915)) > > return false; > > - ret = cnl_verify_procmon_ref_values(dev_priv, port); > > + ret = cnl_verify_procmon_ref_values(i915, port); > > - ret &= check_phy_reg(dev_priv, port, CNL_PORT_CL1CM_DW5, > > + ret &= check_phy_reg(i915, port, CNL_PORT_CL1CM_DW5, > > CL_POWER_DOWN_ENABLE, CL_POWER_DOWN_ENABLE); > > return ret; > > } > > -static void cnl_combo_phys_init(struct drm_i915_private *dev_priv) > > +static void cnl_combo_phys_init(struct drm_i915_private *i915) > > { > > u32 val; > > @@ -157,7 +157,7 @@ static void cnl_combo_phys_init(struct drm_i915_private *dev_priv) > > I915_WRITE(CHICKEN_MISC_2, val); > > /* Dummy PORT_A to get the correct CNL register from the ICL macro */ > > - cnl_set_procmon_ref_values(dev_priv, PORT_A); > > + cnl_set_procmon_ref_values(i915, PORT_A); > > val = I915_READ(CNL_PORT_COMP_DW0); > > val |= COMP_INIT; > > @@ -168,11 +168,11 @@ static void cnl_combo_phys_init(struct drm_i915_private *dev_priv) > > I915_WRITE(CNL_PORT_CL1CM_DW5, val); > > } > > -static void cnl_combo_phys_uninit(struct drm_i915_private *dev_priv) > > +static void cnl_combo_phys_uninit(struct drm_i915_private *i915) > > { > > u32 val; > > - if (!cnl_combo_phy_verify_state(dev_priv)) > > + if (!cnl_combo_phy_verify_state(i915)) > > DRM_WARN("Combo PHY HW state changed unexpectedly.\n"); > > val = I915_READ(CHICKEN_MISC_2); > > @@ -180,7 +180,7 @@ static void cnl_combo_phys_uninit(struct drm_i915_private *dev_priv) > > I915_WRITE(CHICKEN_MISC_2, val); > > } > > -static bool icl_combo_phy_enabled(struct drm_i915_private *dev_priv, > > +static bool icl_combo_phy_enabled(struct drm_i915_private *i915, > > enum port port) > > { > > return !(I915_READ(ICL_PHY_MISC(port)) & > > @@ -188,27 +188,27 @@ static bool icl_combo_phy_enabled(struct drm_i915_private *dev_priv, > > (I915_READ(ICL_PORT_COMP_DW0(port)) & COMP_INIT); > > } > > -static bool icl_combo_phy_verify_state(struct drm_i915_private *dev_priv, > > +static bool icl_combo_phy_verify_state(struct drm_i915_private *i915, > > enum port port) > > { > > bool ret; > > - if (!icl_combo_phy_enabled(dev_priv, port)) > > + if (!icl_combo_phy_enabled(i915, port)) > > return false; > > - ret = cnl_verify_procmon_ref_values(dev_priv, port); > > + ret = cnl_verify_procmon_ref_values(i915, port); > > if (port == PORT_A) > > - ret &= check_phy_reg(dev_priv, port, ICL_PORT_COMP_DW8(port), > > + ret &= check_phy_reg(i915, port, ICL_PORT_COMP_DW8(port), > > IREFGEN, IREFGEN); > > - ret &= check_phy_reg(dev_priv, port, ICL_PORT_CL_DW5(port), > > + ret &= check_phy_reg(i915, port, ICL_PORT_CL_DW5(port), > > CL_POWER_DOWN_ENABLE, CL_POWER_DOWN_ENABLE); > > return ret; > > } > > -void intel_combo_phy_power_up_lanes(struct drm_i915_private *dev_priv, > > +void intel_combo_phy_power_up_lanes(struct drm_i915_private *i915, > > enum port port, bool is_dsi, > > int lane_count, bool lane_reversal) > > { > > @@ -260,14 +260,14 @@ void intel_combo_phy_power_up_lanes(struct drm_i915_private *dev_priv, > > I915_WRITE(ICL_PORT_CL_DW10(port), val); > > } > > -static void icl_combo_phys_init(struct drm_i915_private *dev_priv) > > +static void icl_combo_phys_init(struct drm_i915_private *i915) > > { > > enum port port; > > - for_each_combo_port(dev_priv, port) { > > + for_each_combo_port(i915, port) { > > u32 val; > > - if (icl_combo_phy_verify_state(dev_priv, port)) { > > + if (icl_combo_phy_verify_state(i915, port)) { > > DRM_DEBUG_DRIVER("Port %c combo PHY already enabled, won't reprogram it.\n", > > port_name(port)); > > continue; > > @@ -277,7 +277,7 @@ static void icl_combo_phys_init(struct drm_i915_private *dev_priv) > > val &= ~ICL_PHY_MISC_DE_IO_COMP_PWR_DOWN; > > I915_WRITE(ICL_PHY_MISC(port), val); > > - cnl_set_procmon_ref_values(dev_priv, port); > > + cnl_set_procmon_ref_values(i915, port); > > if (port == PORT_A) { > > val = I915_READ(ICL_PORT_COMP_DW8(port)); > > @@ -295,15 +295,15 @@ static void icl_combo_phys_init(struct drm_i915_private *dev_priv) > > } > > } > > -static void icl_combo_phys_uninit(struct drm_i915_private *dev_priv) > > +static void icl_combo_phys_uninit(struct drm_i915_private *i915) > > { > > enum port port; > > - for_each_combo_port_reverse(dev_priv, port) { > > + for_each_combo_port_reverse(i915, port) { > > u32 val; > > if (port == PORT_A && > > - !icl_combo_phy_verify_state(dev_priv, port)) > > + !icl_combo_phy_verify_state(i915, port)) > > DRM_WARN("Port %c combo PHY HW state changed unexpectedly\n", > > port_name(port)); > > diff --git a/drivers/gpu/drm/i915/intel_combo_phy.h b/drivers/gpu/drm/i915/intel_combo_phy.h > > index e6e195a83b19..a08286cfcecf 100644 > > --- a/drivers/gpu/drm/i915/intel_combo_phy.h > > +++ b/drivers/gpu/drm/i915/intel_combo_phy.h > > @@ -11,9 +11,9 @@ > > struct drm_i915_private; > > -void intel_combo_phy_init(struct drm_i915_private *dev_priv); > > -void intel_combo_phy_uninit(struct drm_i915_private *dev_priv); > > -void intel_combo_phy_power_up_lanes(struct drm_i915_private *dev_priv, > > +void intel_combo_phy_init(struct drm_i915_private *i915); > > +void intel_combo_phy_uninit(struct drm_i915_private *i915); > > +void intel_combo_phy_power_up_lanes(struct drm_i915_private *i915, > > enum port port, bool is_dsi, > > int lane_count, bool lane_reversal); > > diff --git a/drivers/gpu/drm/i915/intel_connector.c b/drivers/gpu/drm/i915/intel_connector.c > > index 073b6c3ab7cc..2912804d5a0d 100644 > > --- a/drivers/gpu/drm/i915/intel_connector.c > > +++ b/drivers/gpu/drm/i915/intel_connector.c > > @@ -219,10 +219,10 @@ void > > intel_attach_force_audio_property(struct drm_connector *connector) > > { > > struct drm_device *dev = connector->dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > struct drm_property *prop; > > - prop = dev_priv->force_audio_property; > > + prop = i915->force_audio_property; > > if (prop == NULL) { > > prop = drm_property_create_enum(dev, 0, > > "audio", > > @@ -231,7 +231,7 @@ intel_attach_force_audio_property(struct drm_connector *connector) > > if (prop == NULL) > > return; > > - dev_priv->force_audio_property = prop; > > + i915->force_audio_property = prop; > > } > > drm_object_attach_property(&connector->base, prop, 0); > > } > > @@ -246,10 +246,10 @@ void > > intel_attach_broadcast_rgb_property(struct drm_connector *connector) > > { > > struct drm_device *dev = connector->dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > struct drm_property *prop; > > - prop = dev_priv->broadcast_rgb_property; > > + prop = i915->broadcast_rgb_property; > > if (prop == NULL) { > > prop = drm_property_create_enum(dev, DRM_MODE_PROP_ENUM, > > "Broadcast RGB", > > @@ -258,7 +258,7 @@ intel_attach_broadcast_rgb_property(struct drm_connector *connector) > > if (prop == NULL) > > return; > > - dev_priv->broadcast_rgb_property = prop; > > + i915->broadcast_rgb_property = prop; > > } > > drm_object_attach_property(&connector->base, prop, 0); > > diff --git a/drivers/gpu/drm/i915/intel_crt.c b/drivers/gpu/drm/i915/intel_crt.c > > index 3fcf2f84bcce..2a794e0bf457 100644 > > --- a/drivers/gpu/drm/i915/intel_crt.c > > +++ b/drivers/gpu/drm/i915/intel_crt.c > > @@ -70,7 +70,7 @@ static struct intel_crt *intel_attached_crt(struct drm_connector *connector) > > return intel_encoder_to_crt(intel_attached_encoder(connector)); > > } > > -bool intel_crt_port_enabled(struct drm_i915_private *dev_priv, > > +bool intel_crt_port_enabled(struct drm_i915_private *i915, > > i915_reg_t adpa_reg, enum pipe *pipe) > > { > > u32 val; > > @@ -78,7 +78,7 @@ bool intel_crt_port_enabled(struct drm_i915_private *dev_priv, > > val = I915_READ(adpa_reg); > > /* asserts want to know the pipe even if the port is disabled */ > > - if (HAS_PCH_CPT(dev_priv)) > > + if (HAS_PCH_CPT(i915)) > > *pipe = (val & ADPA_PIPE_SEL_MASK_CPT) >> ADPA_PIPE_SEL_SHIFT_CPT; > > else > > *pipe = (val & ADPA_PIPE_SEL_MASK) >> ADPA_PIPE_SEL_SHIFT; > > @@ -89,26 +89,26 @@ bool intel_crt_port_enabled(struct drm_i915_private *dev_priv, > > static bool intel_crt_get_hw_state(struct intel_encoder *encoder, > > enum pipe *pipe) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_crt *crt = intel_encoder_to_crt(encoder); > > intel_wakeref_t wakeref; > > bool ret; > > - wakeref = intel_display_power_get_if_enabled(dev_priv, > > + wakeref = intel_display_power_get_if_enabled(i915, > > encoder->power_domain); > > if (!wakeref) > > return false; > > - ret = intel_crt_port_enabled(dev_priv, crt->adpa_reg, pipe); > > + ret = intel_crt_port_enabled(i915, crt->adpa_reg, pipe); > > - intel_display_power_put(dev_priv, encoder->power_domain, wakeref); > > + intel_display_power_put(i915, encoder->power_domain, wakeref); > > return ret; > > } > > static unsigned int intel_crt_get_flags(struct intel_encoder *encoder) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_crt *crt = intel_encoder_to_crt(encoder); > > u32 tmp, flags = 0; > > @@ -140,7 +140,7 @@ static void intel_crt_get_config(struct intel_encoder *encoder, > > static void hsw_crt_get_config(struct intel_encoder *encoder, > > struct intel_crtc_state *pipe_config) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > intel_ddi_get_config(encoder, pipe_config); > > @@ -150,7 +150,7 @@ static void hsw_crt_get_config(struct intel_encoder *encoder, > > DRM_MODE_FLAG_NVSYNC); > > pipe_config->base.adjusted_mode.flags |= intel_crt_get_flags(encoder); > > - pipe_config->base.adjusted_mode.crtc_clock = lpt_get_iclkip(dev_priv); > > + pipe_config->base.adjusted_mode.crtc_clock = lpt_get_iclkip(i915); > > } > > /* Note: The caller is required to filter out dpms modes not supported by the > > @@ -159,13 +159,13 @@ static void intel_crt_set_dpms(struct intel_encoder *encoder, > > const struct intel_crtc_state *crtc_state, > > int mode) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_crt *crt = intel_encoder_to_crt(encoder); > > struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc); > > const struct drm_display_mode *adjusted_mode = &crtc_state->base.adjusted_mode; > > u32 adpa; > > - if (INTEL_GEN(dev_priv) >= 5) > > + if (INTEL_GEN(i915) >= 5) > > adpa = ADPA_HOTPLUG_BITS; > > else > > adpa = 0; > > @@ -176,14 +176,14 @@ static void intel_crt_set_dpms(struct intel_encoder *encoder, > > adpa |= ADPA_VSYNC_ACTIVE_HIGH; > > /* For CPT allow 3 pipe config, for others just use A or B */ > > - if (HAS_PCH_LPT(dev_priv)) > > + if (HAS_PCH_LPT(i915)) > > ; /* Those bits don't exist here */ > > - else if (HAS_PCH_CPT(dev_priv)) > > + else if (HAS_PCH_CPT(i915)) > > adpa |= ADPA_PIPE_SEL_CPT(crtc->pipe); > > else > > adpa |= ADPA_PIPE_SEL(crtc->pipe); > > - if (!HAS_PCH_SPLIT(dev_priv)) > > + if (!HAS_PCH_SPLIT(i915)) > > I915_WRITE(BCLRPAT(crtc->pipe), 0); > > switch (mode) { > > @@ -228,57 +228,57 @@ static void hsw_disable_crt(struct intel_encoder *encoder, > > const struct intel_crtc_state *old_crtc_state, > > const struct drm_connector_state *old_conn_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > WARN_ON(!old_crtc_state->has_pch_encoder); > > - intel_set_pch_fifo_underrun_reporting(dev_priv, PIPE_A, false); > > + intel_set_pch_fifo_underrun_reporting(i915, PIPE_A, false); > > } > > static void hsw_post_disable_crt(struct intel_encoder *encoder, > > const struct intel_crtc_state *old_crtc_state, > > const struct drm_connector_state *old_conn_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > intel_ddi_disable_pipe_clock(old_crtc_state); > > pch_post_disable_crt(encoder, old_crtc_state, old_conn_state); > > - lpt_disable_pch_transcoder(dev_priv); > > - lpt_disable_iclkip(dev_priv); > > + lpt_disable_pch_transcoder(i915); > > + lpt_disable_iclkip(i915); > > intel_ddi_fdi_post_disable(encoder, old_crtc_state, old_conn_state); > > WARN_ON(!old_crtc_state->has_pch_encoder); > > - intel_set_pch_fifo_underrun_reporting(dev_priv, PIPE_A, true); > > + intel_set_pch_fifo_underrun_reporting(i915, PIPE_A, true); > > } > > static void hsw_pre_pll_enable_crt(struct intel_encoder *encoder, > > const struct intel_crtc_state *crtc_state, > > const struct drm_connector_state *conn_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > WARN_ON(!crtc_state->has_pch_encoder); > > - intel_set_pch_fifo_underrun_reporting(dev_priv, PIPE_A, false); > > + intel_set_pch_fifo_underrun_reporting(i915, PIPE_A, false); > > } > > static void hsw_pre_enable_crt(struct intel_encoder *encoder, > > const struct intel_crtc_state *crtc_state, > > const struct drm_connector_state *conn_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc); > > enum pipe pipe = crtc->pipe; > > WARN_ON(!crtc_state->has_pch_encoder); > > - intel_set_cpu_fifo_underrun_reporting(dev_priv, pipe, false); > > + intel_set_cpu_fifo_underrun_reporting(i915, pipe, false); > > - dev_priv->display.fdi_link_train(crtc, crtc_state); > > + i915->display.fdi_link_train(crtc, crtc_state); > > intel_ddi_enable_pipe_clock(crtc_state); > > } > > @@ -287,7 +287,7 @@ static void hsw_enable_crt(struct intel_encoder *encoder, > > const struct intel_crtc_state *crtc_state, > > const struct drm_connector_state *conn_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc); > > enum pipe pipe = crtc->pipe; > > @@ -295,10 +295,10 @@ static void hsw_enable_crt(struct intel_encoder *encoder, > > intel_crt_set_dpms(encoder, crtc_state, DRM_MODE_DPMS_ON); > > - intel_wait_for_vblank(dev_priv, pipe); > > - intel_wait_for_vblank(dev_priv, pipe); > > - intel_set_cpu_fifo_underrun_reporting(dev_priv, pipe, true); > > - intel_set_pch_fifo_underrun_reporting(dev_priv, PIPE_A, true); > > + intel_wait_for_vblank(i915, pipe); > > + intel_wait_for_vblank(i915, pipe); > > + intel_set_cpu_fifo_underrun_reporting(i915, pipe, true); > > + intel_set_pch_fifo_underrun_reporting(i915, PIPE_A, true); > > } > > static void intel_enable_crt(struct intel_encoder *encoder, > > @@ -313,8 +313,8 @@ intel_crt_mode_valid(struct drm_connector *connector, > > struct drm_display_mode *mode) > > { > > struct drm_device *dev = connector->dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > - int max_dotclk = dev_priv->max_dotclk_freq; > > + struct drm_i915_private *i915 = to_i915(dev); > > + int max_dotclk = i915->max_dotclk_freq; > > int max_clock; > > if (mode->flags & DRM_MODE_FLAG_DBLSCAN) > > @@ -323,15 +323,15 @@ intel_crt_mode_valid(struct drm_connector *connector, > > if (mode->clock < 25000) > > return MODE_CLOCK_LOW; > > - if (HAS_PCH_LPT(dev_priv)) > > + if (HAS_PCH_LPT(i915)) > > max_clock = 180000; > > - else if (IS_VALLEYVIEW(dev_priv)) > > + else if (IS_VALLEYVIEW(i915)) > > /* > > * 270 MHz due to current DPLL limits, > > * DAC limit supposedly 355 MHz. > > */ > > max_clock = 270000; > > - else if (IS_GEN_RANGE(dev_priv, 3, 4)) > > + else if (IS_GEN_RANGE(i915, 3, 4)) > > max_clock = 400000; > > else > > max_clock = 350000; > > @@ -342,7 +342,7 @@ intel_crt_mode_valid(struct drm_connector *connector, > > return MODE_CLOCK_HIGH; > > /* The FDI receiver on LPT only supports 8bpc and only has 2 lanes. */ > > - if (HAS_PCH_LPT(dev_priv) && > > + if (HAS_PCH_LPT(i915) && > > (ironlake_get_lanes_required(mode->clock, 270000, 24) > 2)) > > return MODE_CLOCK_HIGH; > > @@ -388,7 +388,7 @@ static int hsw_crt_compute_config(struct intel_encoder *encoder, > > struct intel_crtc_state *pipe_config, > > struct drm_connector_state *conn_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct drm_display_mode *adjusted_mode = > > &pipe_config->base.adjusted_mode; > > @@ -404,7 +404,7 @@ static int hsw_crt_compute_config(struct intel_encoder *encoder, > > pipe_config->output_format = INTEL_OUTPUT_FORMAT_RGB; > > /* LPT FDI RX only supports 8bpc. */ > > - if (HAS_PCH_LPT(dev_priv)) { > > + if (HAS_PCH_LPT(i915)) { > > if (pipe_config->bw_constrained && pipe_config->pipe_bpp < 24) { > > DRM_DEBUG_KMS("LPT only supports 24bpp\n"); > > return -EINVAL; > > @@ -423,13 +423,13 @@ static bool intel_ironlake_crt_detect_hotplug(struct drm_connector *connector) > > { > > struct drm_device *dev = connector->dev; > > struct intel_crt *crt = intel_attached_crt(connector); > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > u32 adpa; > > bool ret; > > /* The first time through, trigger an explicit detection cycle */ > > if (crt->force_hotplug_required) { > > - bool turn_off_dac = HAS_PCH_SPLIT(dev_priv); > > + bool turn_off_dac = HAS_PCH_SPLIT(i915); > > u32 save_adpa; > > crt->force_hotplug_required = 0; > > @@ -443,7 +443,7 @@ static bool intel_ironlake_crt_detect_hotplug(struct drm_connector *connector) > > I915_WRITE(crt->adpa_reg, adpa); > > - if (intel_wait_for_register(&dev_priv->uncore, > > + if (intel_wait_for_register(&i915->uncore, > > crt->adpa_reg, > > ADPA_CRT_HOTPLUG_FORCE_TRIGGER, 0, > > 1000)) > > @@ -470,7 +470,7 @@ static bool valleyview_crt_detect_hotplug(struct drm_connector *connector) > > { > > struct drm_device *dev = connector->dev; > > struct intel_crt *crt = intel_attached_crt(connector); > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > bool reenable_hpd; > > u32 adpa; > > bool ret; > > @@ -488,7 +488,7 @@ static bool valleyview_crt_detect_hotplug(struct drm_connector *connector) > > * > > * Just disable HPD interrupts here to prevent this > > */ > > - reenable_hpd = intel_hpd_disable(dev_priv, crt->base.hpd_pin); > > + reenable_hpd = intel_hpd_disable(i915, crt->base.hpd_pin); > > save_adpa = adpa = I915_READ(crt->adpa_reg); > > DRM_DEBUG_KMS("trigger hotplug detect cycle: adpa=0x%x\n", adpa); > > @@ -497,7 +497,7 @@ static bool valleyview_crt_detect_hotplug(struct drm_connector *connector) > > I915_WRITE(crt->adpa_reg, adpa); > > - if (intel_wait_for_register(&dev_priv->uncore, > > + if (intel_wait_for_register(&i915->uncore, > > crt->adpa_reg, > > ADPA_CRT_HOTPLUG_FORCE_TRIGGER, 0, > > 1000)) { > > @@ -515,7 +515,7 @@ static bool valleyview_crt_detect_hotplug(struct drm_connector *connector) > > DRM_DEBUG_KMS("valleyview hotplug adpa=0x%x, result %d\n", adpa, ret); > > if (reenable_hpd) > > - intel_hpd_enable(dev_priv, crt->base.hpd_pin); > > + intel_hpd_enable(i915, crt->base.hpd_pin); > > return ret; > > } > > @@ -523,15 +523,15 @@ static bool valleyview_crt_detect_hotplug(struct drm_connector *connector) > > static bool intel_crt_detect_hotplug(struct drm_connector *connector) > > { > > struct drm_device *dev = connector->dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > u32 stat; > > bool ret = false; > > int i, tries = 0; > > - if (HAS_PCH_SPLIT(dev_priv)) > > + if (HAS_PCH_SPLIT(i915)) > > return intel_ironlake_crt_detect_hotplug(connector); > > - if (IS_VALLEYVIEW(dev_priv)) > > + if (IS_VALLEYVIEW(i915)) > > return valleyview_crt_detect_hotplug(connector); > > /* > > @@ -539,18 +539,18 @@ static bool intel_crt_detect_hotplug(struct drm_connector *connector) > > * to get a reliable result. > > */ > > - if (IS_G45(dev_priv)) > > + if (IS_G45(i915)) > > tries = 2; > > else > > tries = 1; > > for (i = 0; i < tries ; i++) { > > /* turn on the FORCE_DETECT */ > > - i915_hotplug_interrupt_update(dev_priv, > > + i915_hotplug_interrupt_update(i915, > > CRT_HOTPLUG_FORCE_DETECT, > > CRT_HOTPLUG_FORCE_DETECT); > > /* wait for FORCE_DETECT to go off */ > > - if (intel_wait_for_register(&dev_priv->uncore, PORT_HOTPLUG_EN, > > + if (intel_wait_for_register(&i915->uncore, PORT_HOTPLUG_EN, > > CRT_HOTPLUG_FORCE_DETECT, 0, > > 1000)) > > DRM_DEBUG_KMS("timed out waiting for FORCE_DETECT to go off"); > > @@ -563,7 +563,7 @@ static bool intel_crt_detect_hotplug(struct drm_connector *connector) > > /* clear the interrupt we just generated, if any */ > > I915_WRITE(PORT_HOTPLUG_STAT, CRT_HOTPLUG_INT_STATUS); > > - i915_hotplug_interrupt_update(dev_priv, CRT_HOTPLUG_FORCE_DETECT, 0); > > + i915_hotplug_interrupt_update(i915, CRT_HOTPLUG_FORCE_DETECT, 0); > > return ret; > > } > > @@ -605,14 +605,14 @@ static int intel_crt_ddc_get_modes(struct drm_connector *connector, > > static bool intel_crt_detect_ddc(struct drm_connector *connector) > > { > > struct intel_crt *crt = intel_attached_crt(connector); > > - struct drm_i915_private *dev_priv = to_i915(crt->base.base.dev); > > + struct drm_i915_private *i915 = to_i915(crt->base.base.dev); > > struct edid *edid; > > struct i2c_adapter *i2c; > > bool ret = false; > > BUG_ON(crt->base.type != INTEL_OUTPUT_ANALOG); > > - i2c = intel_gmbus_get_adapter(dev_priv, dev_priv->vbt.crt_ddc_pin); > > + i2c = intel_gmbus_get_adapter(i915, i915->vbt.crt_ddc_pin); > > edid = intel_crt_get_edid(connector, i2c); > > if (edid) { > > @@ -642,8 +642,8 @@ static enum drm_connector_status > > intel_crt_load_detect(struct intel_crt *crt, u32 pipe) > > { > > struct drm_device *dev = crt->base.base.dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > - struct intel_uncore *uncore = &dev_priv->uncore; > > + struct drm_i915_private *i915 = to_i915(dev); > > + struct intel_uncore *uncore = &i915->uncore; > > u32 save_bclrpat; > > u32 save_vtotal; > > u32 vtotal, vactive; > > @@ -677,7 +677,7 @@ intel_crt_load_detect(struct intel_crt *crt, u32 pipe) > > /* Set the border color to purple. */ > > intel_uncore_write(uncore, bclrpat_reg, 0x500050); > > - if (!IS_GEN(dev_priv, 2)) { > > + if (!IS_GEN(i915, 2)) { > > u32 pipeconf = intel_uncore_read(uncore, pipeconf_reg); > > intel_uncore_write(uncore, > > pipeconf_reg, > > @@ -685,7 +685,7 @@ intel_crt_load_detect(struct intel_crt *crt, u32 pipe) > > intel_uncore_posting_read(uncore, pipeconf_reg); > > /* Wait for next Vblank to substitue > > * border color for Color info */ > > - intel_wait_for_vblank(dev_priv, pipe); > > + intel_wait_for_vblank(i915, pipe); > > st00 = intel_uncore_read8(uncore, _VGA_MSR_WRITE); > > status = ((st00 & (1 << 4)) != 0) ? > > connector_status_connected : > > @@ -789,7 +789,7 @@ intel_crt_detect(struct drm_connector *connector, > > struct drm_modeset_acquire_ctx *ctx, > > bool force) > > { > > - struct drm_i915_private *dev_priv = to_i915(connector->dev); > > + struct drm_i915_private *i915 = to_i915(connector->dev); > > struct intel_crt *crt = intel_attached_crt(connector); > > struct intel_encoder *intel_encoder = &crt->base; > > intel_wakeref_t wakeref; > > @@ -801,7 +801,7 @@ intel_crt_detect(struct drm_connector *connector, > > force); > > if (i915_modparams.load_detect_test) { > > - wakeref = intel_display_power_get(dev_priv, > > + wakeref = intel_display_power_get(i915, > > intel_encoder->power_domain); > > goto load_detect; > > } > > @@ -810,10 +810,10 @@ intel_crt_detect(struct drm_connector *connector, > > if (dmi_check_system(intel_spurious_crt_detect)) > > return connector_status_disconnected; > > - wakeref = intel_display_power_get(dev_priv, > > + wakeref = intel_display_power_get(i915, > > intel_encoder->power_domain); > > - if (I915_HAS_HOTPLUG(dev_priv)) { > > + if (I915_HAS_HOTPLUG(i915)) { > > /* We can not rely on the HPD pin always being correctly wired > > * up, for example many KVM do not pass it through, and so > > * only trust an assertion that the monitor is connected. > > @@ -835,7 +835,7 @@ intel_crt_detect(struct drm_connector *connector, > > * broken monitor (without edid) to work behind a broken kvm (that fails > > * to have the right resistors for HP detection) needs to fix this up. > > * For now just bail out. */ > > - if (I915_HAS_HOTPLUG(dev_priv)) { > > + if (I915_HAS_HOTPLUG(i915)) { > > status = connector_status_disconnected; > > goto out; > > } > > @@ -851,7 +851,7 @@ intel_crt_detect(struct drm_connector *connector, > > if (ret > 0) { > > if (intel_crt_detect_ddc(connector)) > > status = connector_status_connected; > > - else if (INTEL_GEN(dev_priv) < 4) > > + else if (INTEL_GEN(i915) < 4) > > status = intel_crt_load_detect(crt, > > to_intel_crtc(connector->state->crtc)->pipe); > > else if (i915_modparams.load_detect_test) > > @@ -866,44 +866,44 @@ intel_crt_detect(struct drm_connector *connector, > > } > > out: > > - intel_display_power_put(dev_priv, intel_encoder->power_domain, wakeref); > > + intel_display_power_put(i915, intel_encoder->power_domain, wakeref); > > return status; > > } > > static int intel_crt_get_modes(struct drm_connector *connector) > > { > > struct drm_device *dev = connector->dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > struct intel_crt *crt = intel_attached_crt(connector); > > struct intel_encoder *intel_encoder = &crt->base; > > intel_wakeref_t wakeref; > > struct i2c_adapter *i2c; > > int ret; > > - wakeref = intel_display_power_get(dev_priv, > > + wakeref = intel_display_power_get(i915, > > intel_encoder->power_domain); > > - i2c = intel_gmbus_get_adapter(dev_priv, dev_priv->vbt.crt_ddc_pin); > > + i2c = intel_gmbus_get_adapter(i915, i915->vbt.crt_ddc_pin); > > ret = intel_crt_ddc_get_modes(connector, i2c); > > - if (ret || !IS_G4X(dev_priv)) > > + if (ret || !IS_G4X(i915)) > > goto out; > > /* Try to probe digital port for output in DVI-I -> VGA mode. */ > > - i2c = intel_gmbus_get_adapter(dev_priv, GMBUS_PIN_DPB); > > + i2c = intel_gmbus_get_adapter(i915, GMBUS_PIN_DPB); > > ret = intel_crt_ddc_get_modes(connector, i2c); > > out: > > - intel_display_power_put(dev_priv, intel_encoder->power_domain, wakeref); > > + intel_display_power_put(i915, intel_encoder->power_domain, wakeref); > > return ret; > > } > > void intel_crt_reset(struct drm_encoder *encoder) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->dev); > > + struct drm_i915_private *i915 = to_i915(encoder->dev); > > struct intel_crt *crt = intel_encoder_to_crt(to_intel_encoder(encoder)); > > - if (INTEL_GEN(dev_priv) >= 5) { > > + if (INTEL_GEN(i915) >= 5) { > > u32 adpa; > > adpa = I915_READ(crt->adpa_reg); > > @@ -942,7 +942,7 @@ static const struct drm_encoder_funcs intel_crt_enc_funcs = { > > .destroy = intel_encoder_destroy, > > }; > > -void intel_crt_init(struct drm_i915_private *dev_priv) > > +void intel_crt_init(struct drm_i915_private *i915) > > { > > struct drm_connector *connector; > > struct intel_crt *crt; > > @@ -950,9 +950,9 @@ void intel_crt_init(struct drm_i915_private *dev_priv) > > i915_reg_t adpa_reg; > > u32 adpa; > > - if (HAS_PCH_SPLIT(dev_priv)) > > + if (HAS_PCH_SPLIT(i915)) > > adpa_reg = PCH_ADPA; > > - else if (IS_VALLEYVIEW(dev_priv)) > > + else if (IS_VALLEYVIEW(i915)) > > adpa_reg = VLV_ADPA; > > else > > adpa_reg = ADPA; > > @@ -986,22 +986,22 @@ void intel_crt_init(struct drm_i915_private *dev_priv) > > connector = &intel_connector->base; > > crt->connector = intel_connector; > > - drm_connector_init(&dev_priv->drm, &intel_connector->base, > > + drm_connector_init(&i915->drm, &intel_connector->base, > > &intel_crt_connector_funcs, DRM_MODE_CONNECTOR_VGA); > > - drm_encoder_init(&dev_priv->drm, &crt->base.base, &intel_crt_enc_funcs, > > + drm_encoder_init(&i915->drm, &crt->base.base, &intel_crt_enc_funcs, > > DRM_MODE_ENCODER_DAC, "CRT"); > > intel_connector_attach_encoder(intel_connector, &crt->base); > > crt->base.type = INTEL_OUTPUT_ANALOG; > > crt->base.cloneable = (1 << INTEL_OUTPUT_DVO) | (1 << INTEL_OUTPUT_HDMI); > > - if (IS_I830(dev_priv)) > > + if (IS_I830(i915)) > > crt->base.crtc_mask = (1 << 0); > > else > > crt->base.crtc_mask = (1 << 0) | (1 << 1) | (1 << 2); > > - if (IS_GEN(dev_priv, 2)) > > + if (IS_GEN(i915, 2)) > > connector->interlace_allowed = 0; > > else > > connector->interlace_allowed = 1; > > @@ -1011,13 +1011,13 @@ void intel_crt_init(struct drm_i915_private *dev_priv) > > crt->base.power_domain = POWER_DOMAIN_PORT_CRT; > > - if (I915_HAS_HOTPLUG(dev_priv) && > > + if (I915_HAS_HOTPLUG(i915) && > > !dmi_check_system(intel_spurious_crt_detect)) { > > crt->base.hpd_pin = HPD_CRT; > > crt->base.hotplug = intel_encoder_hotplug; > > } > > - if (HAS_DDI(dev_priv)) { > > + if (HAS_DDI(i915)) { > > crt->base.port = PORT_E; > > crt->base.get_config = hsw_crt_get_config; > > crt->base.get_hw_state = intel_ddi_get_hw_state; > > @@ -1028,7 +1028,7 @@ void intel_crt_init(struct drm_i915_private *dev_priv) > > crt->base.disable = hsw_disable_crt; > > crt->base.post_disable = hsw_post_disable_crt; > > } else { > > - if (HAS_PCH_SPLIT(dev_priv)) { > > + if (HAS_PCH_SPLIT(i915)) { > > crt->base.compute_config = pch_crt_compute_config; > > crt->base.disable = pch_disable_crt; > > crt->base.post_disable = pch_post_disable_crt; > > @@ -1045,7 +1045,7 @@ void intel_crt_init(struct drm_i915_private *dev_priv) > > drm_connector_helper_add(connector, &intel_crt_connector_helper_funcs); > > - if (!I915_HAS_HOTPLUG(dev_priv)) > > + if (!I915_HAS_HOTPLUG(i915)) > > intel_connector->polled = DRM_CONNECTOR_POLL_CONNECT; > > /* > > @@ -1058,11 +1058,11 @@ void intel_crt_init(struct drm_i915_private *dev_priv) > > * polarity and link reversal bits or not, instead of relying on the > > * BIOS. > > */ > > - if (HAS_PCH_LPT(dev_priv)) { > > + if (HAS_PCH_LPT(i915)) { > > u32 fdi_config = FDI_RX_POLARITY_REVERSED_LPT | > > FDI_RX_LINK_REVERSAL_OVERRIDE; > > - dev_priv->fdi_rx_config = I915_READ(FDI_RX_CTL(PIPE_A)) & fdi_config; > > + i915->fdi_rx_config = I915_READ(FDI_RX_CTL(PIPE_A)) & fdi_config; > > } > > intel_crt_reset(&crt->base.base); > > diff --git a/drivers/gpu/drm/i915/intel_crt.h b/drivers/gpu/drm/i915/intel_crt.h > > index 1b3fba359efc..f771fbcf8b27 100644 > > --- a/drivers/gpu/drm/i915/intel_crt.h > > +++ b/drivers/gpu/drm/i915/intel_crt.h > > @@ -13,9 +13,9 @@ struct drm_encoder; > > struct drm_i915_private; > > struct drm_i915_private; > > -bool intel_crt_port_enabled(struct drm_i915_private *dev_priv, > > +bool intel_crt_port_enabled(struct drm_i915_private *i915, > > i915_reg_t adpa_reg, enum pipe *pipe); > > -void intel_crt_init(struct drm_i915_private *dev_priv); > > +void intel_crt_init(struct drm_i915_private *i915); > > void intel_crt_reset(struct drm_encoder *encoder); > > #endif /* __INTEL_CRT_H__ */ > > diff --git a/drivers/gpu/drm/i915/intel_csr.c b/drivers/gpu/drm/i915/intel_csr.c > > index bf0eebd385b9..29ea3cb6bbaf 100644 > > --- a/drivers/gpu/drm/i915/intel_csr.c > > +++ b/drivers/gpu/drm/i915/intel_csr.c > > @@ -206,18 +206,18 @@ static const struct stepping_info icl_stepping_info[] = { > > static const struct stepping_info no_stepping_info = { '*', '*' }; > > static const struct stepping_info * > > -intel_get_stepping_info(struct drm_i915_private *dev_priv) > > +intel_get_stepping_info(struct drm_i915_private *i915) > > { > > const struct stepping_info *si; > > unsigned int size; > > - if (IS_ICELAKE(dev_priv)) { > > + if (IS_ICELAKE(i915)) { > > size = ARRAY_SIZE(icl_stepping_info); > > si = icl_stepping_info; > > - } else if (IS_SKYLAKE(dev_priv)) { > > + } else if (IS_SKYLAKE(i915)) { > > size = ARRAY_SIZE(skl_stepping_info); > > si = skl_stepping_info; > > - } else if (IS_BROXTON(dev_priv)) { > > + } else if (IS_BROXTON(i915)) { > > size = ARRAY_SIZE(bxt_stepping_info); > > si = bxt_stepping_info; > > } else { > > @@ -225,19 +225,19 @@ intel_get_stepping_info(struct drm_i915_private *dev_priv) > > si = NULL; > > } > > - if (INTEL_REVID(dev_priv) < size) > > - return si + INTEL_REVID(dev_priv); > > + if (INTEL_REVID(i915) < size) > > + return si + INTEL_REVID(i915); > > return &no_stepping_info; > > } > > -static void gen9_set_dc_state_debugmask(struct drm_i915_private *dev_priv) > > +static void gen9_set_dc_state_debugmask(struct drm_i915_private *i915) > > { > > u32 val, mask; > > mask = DC_STATE_DEBUG_MASK_MEMORY_UP; > > - if (IS_GEN9_LP(dev_priv)) > > + if (IS_GEN9_LP(i915)) > > mask |= DC_STATE_DEBUG_MASK_CORES; > > /* The below bit doesn't need to be cleared ever afterwards */ > > @@ -251,29 +251,29 @@ static void gen9_set_dc_state_debugmask(struct drm_i915_private *dev_priv) > > /** > > * intel_csr_load_program() - write the firmware from memory to register. > > - * @dev_priv: i915 drm device. > > + * @i915: i915 drm device. > > * > > * CSR firmware is read from a .bin file and kept in internal memory one time. > > * Everytime display comes back from low power state this function is called to > > * copy the firmware from internal memory to registers. > > */ > > -void intel_csr_load_program(struct drm_i915_private *dev_priv) > > +void intel_csr_load_program(struct drm_i915_private *i915) > > { > > - u32 *payload = dev_priv->csr.dmc_payload; > > + u32 *payload = i915->csr.dmc_payload; > > u32 i, fw_size; > > - if (!HAS_CSR(dev_priv)) { > > + if (!HAS_CSR(i915)) { > > DRM_ERROR("No CSR support available for this platform\n"); > > return; > > } > > - if (!dev_priv->csr.dmc_payload) { > > + if (!i915->csr.dmc_payload) { > > DRM_ERROR("Tried to program CSR with empty payload\n"); > > return; > > } > > - fw_size = dev_priv->csr.dmc_fw_size; > > - assert_rpm_wakelock_held(dev_priv); > > + fw_size = i915->csr.dmc_fw_size; > > + assert_rpm_wakelock_held(i915); > > preempt_disable(); > > @@ -282,24 +282,24 @@ void intel_csr_load_program(struct drm_i915_private *dev_priv) > > preempt_enable(); > > - for (i = 0; i < dev_priv->csr.mmio_count; i++) { > > - I915_WRITE(dev_priv->csr.mmioaddr[i], > > - dev_priv->csr.mmiodata[i]); > > + for (i = 0; i < i915->csr.mmio_count; i++) { > > + I915_WRITE(i915->csr.mmioaddr[i], > > + i915->csr.mmiodata[i]); > > } > > - dev_priv->csr.dc_state = 0; > > + i915->csr.dc_state = 0; > > - gen9_set_dc_state_debugmask(dev_priv); > > + gen9_set_dc_state_debugmask(i915); > > } > > -static u32 *parse_csr_fw(struct drm_i915_private *dev_priv, > > +static u32 *parse_csr_fw(struct drm_i915_private *i915, > > const struct firmware *fw) > > { > > struct intel_css_header *css_header; > > struct intel_package_header *package_header; > > struct intel_dmc_header *dmc_header; > > - struct intel_csr *csr = &dev_priv->csr; > > - const struct stepping_info *si = intel_get_stepping_info(dev_priv); > > + struct intel_csr *csr = &i915->csr; > > + const struct stepping_info *si = intel_get_stepping_info(i915); > > u32 dmc_offset = CSR_DEFAULT_FW_OFFSET, readcount = 0, nbytes; > > u32 i; > > u32 *dmc_payload; > > @@ -430,48 +430,48 @@ static u32 *parse_csr_fw(struct drm_i915_private *dev_priv, > > return NULL; > > } > > -static void intel_csr_runtime_pm_get(struct drm_i915_private *dev_priv) > > +static void intel_csr_runtime_pm_get(struct drm_i915_private *i915) > > { > > - WARN_ON(dev_priv->csr.wakeref); > > - dev_priv->csr.wakeref = > > - intel_display_power_get(dev_priv, POWER_DOMAIN_INIT); > > + WARN_ON(i915->csr.wakeref); > > + i915->csr.wakeref = > > + intel_display_power_get(i915, POWER_DOMAIN_INIT); > > } > > -static void intel_csr_runtime_pm_put(struct drm_i915_private *dev_priv) > > +static void intel_csr_runtime_pm_put(struct drm_i915_private *i915) > > { > > intel_wakeref_t wakeref __maybe_unused = > > - fetch_and_zero(&dev_priv->csr.wakeref); > > + fetch_and_zero(&i915->csr.wakeref); > > - intel_display_power_put(dev_priv, POWER_DOMAIN_INIT, wakeref); > > + intel_display_power_put(i915, POWER_DOMAIN_INIT, wakeref); > > } > > static void csr_load_work_fn(struct work_struct *work) > > { > > - struct drm_i915_private *dev_priv; > > + struct drm_i915_private *i915; > > struct intel_csr *csr; > > const struct firmware *fw = NULL; > > - dev_priv = container_of(work, typeof(*dev_priv), csr.work); > > - csr = &dev_priv->csr; > > + i915 = container_of(work, typeof(*i915), csr.work); > > + csr = &i915->csr; > > - request_firmware(&fw, dev_priv->csr.fw_path, &dev_priv->drm.pdev->dev); > > + request_firmware(&fw, i915->csr.fw_path, &i915->drm.pdev->dev); > > if (fw) > > - dev_priv->csr.dmc_payload = parse_csr_fw(dev_priv, fw); > > + i915->csr.dmc_payload = parse_csr_fw(i915, fw); > > - if (dev_priv->csr.dmc_payload) { > > - intel_csr_load_program(dev_priv); > > - intel_csr_runtime_pm_put(dev_priv); > > + if (i915->csr.dmc_payload) { > > + intel_csr_load_program(i915); > > + intel_csr_runtime_pm_put(i915); > > DRM_INFO("Finished loading DMC firmware %s (v%u.%u)\n", > > - dev_priv->csr.fw_path, > > + i915->csr.fw_path, > > CSR_VERSION_MAJOR(csr->version), > > CSR_VERSION_MINOR(csr->version)); > > } else { > > - dev_notice(dev_priv->drm.dev, > > + dev_notice(i915->drm.dev, > > "Failed to load DMC firmware %s." > > " Disabling runtime power management.\n", > > csr->fw_path); > > - dev_notice(dev_priv->drm.dev, "DMC firmware homepage: %s", > > + dev_notice(i915->drm.dev, "DMC firmware homepage: %s", > > INTEL_UC_FIRMWARE_URL); > > } > > @@ -480,18 +480,18 @@ static void csr_load_work_fn(struct work_struct *work) > > /** > > * intel_csr_ucode_init() - initialize the firmware loading. > > - * @dev_priv: i915 drm device. > > + * @i915: i915 drm device. > > * > > * This function is called at the time of loading the display driver to read > > * firmware from a .bin file and copied into a internal memory. > > */ > > -void intel_csr_ucode_init(struct drm_i915_private *dev_priv) > > +void intel_csr_ucode_init(struct drm_i915_private *i915) > > { > > - struct intel_csr *csr = &dev_priv->csr; > > + struct intel_csr *csr = &i915->csr; > > - INIT_WORK(&dev_priv->csr.work, csr_load_work_fn); > > + INIT_WORK(&i915->csr.work, csr_load_work_fn); > > - if (!HAS_CSR(dev_priv)) > > + if (!HAS_CSR(i915)) > > return; > > /* > > @@ -502,32 +502,32 @@ void intel_csr_ucode_init(struct drm_i915_private *dev_priv) > > * suspend as runtime suspend *requires* a working CSR for whatever > > * reason. > > */ > > - intel_csr_runtime_pm_get(dev_priv); > > + intel_csr_runtime_pm_get(i915); > > - if (INTEL_GEN(dev_priv) >= 12) { > > + if (INTEL_GEN(i915) >= 12) { > > /* Allow to load fw via parameter using the last known size */ > > csr->max_fw_size = GEN12_CSR_MAX_FW_SIZE; > > - } else if (IS_GEN(dev_priv, 11)) { > > + } else if (IS_GEN(i915, 11)) { > > csr->fw_path = ICL_CSR_PATH; > > csr->required_version = ICL_CSR_VERSION_REQUIRED; > > csr->max_fw_size = ICL_CSR_MAX_FW_SIZE; > > - } else if (IS_CANNONLAKE(dev_priv)) { > > + } else if (IS_CANNONLAKE(i915)) { > > csr->fw_path = CNL_CSR_PATH; > > csr->required_version = CNL_CSR_VERSION_REQUIRED; > > csr->max_fw_size = CNL_CSR_MAX_FW_SIZE; > > - } else if (IS_GEMINILAKE(dev_priv)) { > > + } else if (IS_GEMINILAKE(i915)) { > > csr->fw_path = GLK_CSR_PATH; > > csr->required_version = GLK_CSR_VERSION_REQUIRED; > > csr->max_fw_size = GLK_CSR_MAX_FW_SIZE; > > - } else if (IS_KABYLAKE(dev_priv) || IS_COFFEELAKE(dev_priv)) { > > + } else if (IS_KABYLAKE(i915) || IS_COFFEELAKE(i915)) { > > csr->fw_path = KBL_CSR_PATH; > > csr->required_version = KBL_CSR_VERSION_REQUIRED; > > csr->max_fw_size = KBL_CSR_MAX_FW_SIZE; > > - } else if (IS_SKYLAKE(dev_priv)) { > > + } else if (IS_SKYLAKE(i915)) { > > csr->fw_path = SKL_CSR_PATH; > > csr->required_version = SKL_CSR_VERSION_REQUIRED; > > csr->max_fw_size = SKL_CSR_MAX_FW_SIZE; > > - } else if (IS_BROXTON(dev_priv)) { > > + } else if (IS_BROXTON(i915)) { > > csr->fw_path = BXT_CSR_PATH; > > csr->required_version = BXT_CSR_VERSION_REQUIRED; > > csr->max_fw_size = BXT_CSR_MAX_FW_SIZE; > > @@ -551,63 +551,63 @@ void intel_csr_ucode_init(struct drm_i915_private *dev_priv) > > } > > DRM_DEBUG_KMS("Loading %s\n", csr->fw_path); > > - schedule_work(&dev_priv->csr.work); > > + schedule_work(&i915->csr.work); > > } > > /** > > * intel_csr_ucode_suspend() - prepare CSR firmware before system suspend > > - * @dev_priv: i915 drm device > > + * @i915: i915 drm device > > * > > * Prepare the DMC firmware before entering system suspend. This includes > > * flushing pending work items and releasing any resources acquired during > > * init. > > */ > > -void intel_csr_ucode_suspend(struct drm_i915_private *dev_priv) > > +void intel_csr_ucode_suspend(struct drm_i915_private *i915) > > { > > - if (!HAS_CSR(dev_priv)) > > + if (!HAS_CSR(i915)) > > return; > > - flush_work(&dev_priv->csr.work); > > + flush_work(&i915->csr.work); > > /* Drop the reference held in case DMC isn't loaded. */ > > - if (!dev_priv->csr.dmc_payload) > > - intel_csr_runtime_pm_put(dev_priv); > > + if (!i915->csr.dmc_payload) > > + intel_csr_runtime_pm_put(i915); > > } > > /** > > * intel_csr_ucode_resume() - init CSR firmware during system resume > > - * @dev_priv: i915 drm device > > + * @i915: i915 drm device > > * > > * Reinitialize the DMC firmware during system resume, reacquiring any > > * resources released in intel_csr_ucode_suspend(). > > */ > > -void intel_csr_ucode_resume(struct drm_i915_private *dev_priv) > > +void intel_csr_ucode_resume(struct drm_i915_private *i915) > > { > > - if (!HAS_CSR(dev_priv)) > > + if (!HAS_CSR(i915)) > > return; > > /* > > * Reacquire the reference to keep RPM disabled in case DMC isn't > > * loaded. > > */ > > - if (!dev_priv->csr.dmc_payload) > > - intel_csr_runtime_pm_get(dev_priv); > > + if (!i915->csr.dmc_payload) > > + intel_csr_runtime_pm_get(i915); > > } > > /** > > * intel_csr_ucode_fini() - unload the CSR firmware. > > - * @dev_priv: i915 drm device. > > + * @i915: i915 drm device. > > * > > * Firmmware unloading includes freeing the internal memory and reset the > > * firmware loading status. > > */ > > -void intel_csr_ucode_fini(struct drm_i915_private *dev_priv) > > +void intel_csr_ucode_fini(struct drm_i915_private *i915) > > { > > - if (!HAS_CSR(dev_priv)) > > + if (!HAS_CSR(i915)) > > return; > > - intel_csr_ucode_suspend(dev_priv); > > - WARN_ON(dev_priv->csr.wakeref); > > + intel_csr_ucode_suspend(i915); > > + WARN_ON(i915->csr.wakeref); > > - kfree(dev_priv->csr.dmc_payload); > > + kfree(i915->csr.dmc_payload); > > } > > diff --git a/drivers/gpu/drm/i915/intel_ddi.c b/drivers/gpu/drm/i915/intel_ddi.c > > index 7925a176f900..eb61731f1abc 100644 > > --- a/drivers/gpu/drm/i915/intel_ddi.c > > +++ b/drivers/gpu/drm/i915/intel_ddi.c > > @@ -586,9 +586,9 @@ static const struct icl_mg_phy_ddi_buf_trans icl_mg_phy_ddi_translations[] = { > > }; > > static const struct ddi_buf_trans * > > -bdw_get_buf_trans_edp(struct drm_i915_private *dev_priv, int *n_entries) > > +bdw_get_buf_trans_edp(struct drm_i915_private *i915, int *n_entries) > > { > > - if (dev_priv->vbt.edp.low_vswing) { > > + if (i915->vbt.edp.low_vswing) { > > *n_entries = ARRAY_SIZE(bdw_ddi_translations_edp); > > return bdw_ddi_translations_edp; > > } else { > > @@ -598,12 +598,12 @@ bdw_get_buf_trans_edp(struct drm_i915_private *dev_priv, int *n_entries) > > } > > static const struct ddi_buf_trans * > > -skl_get_buf_trans_dp(struct drm_i915_private *dev_priv, int *n_entries) > > +skl_get_buf_trans_dp(struct drm_i915_private *i915, int *n_entries) > > { > > - if (IS_SKL_ULX(dev_priv)) { > > + if (IS_SKL_ULX(i915)) { > > *n_entries = ARRAY_SIZE(skl_y_ddi_translations_dp); > > return skl_y_ddi_translations_dp; > > - } else if (IS_SKL_ULT(dev_priv)) { > > + } else if (IS_SKL_ULT(i915)) { > > *n_entries = ARRAY_SIZE(skl_u_ddi_translations_dp); > > return skl_u_ddi_translations_dp; > > } else { > > @@ -613,12 +613,12 @@ skl_get_buf_trans_dp(struct drm_i915_private *dev_priv, int *n_entries) > > } > > static const struct ddi_buf_trans * > > -kbl_get_buf_trans_dp(struct drm_i915_private *dev_priv, int *n_entries) > > +kbl_get_buf_trans_dp(struct drm_i915_private *i915, int *n_entries) > > { > > - if (IS_KBL_ULX(dev_priv) || IS_CFL_ULX(dev_priv)) { > > + if (IS_KBL_ULX(i915) || IS_CFL_ULX(i915)) { > > *n_entries = ARRAY_SIZE(kbl_y_ddi_translations_dp); > > return kbl_y_ddi_translations_dp; > > - } else if (IS_KBL_ULT(dev_priv) || IS_CFL_ULT(dev_priv)) { > > + } else if (IS_KBL_ULT(i915) || IS_CFL_ULT(i915)) { > > *n_entries = ARRAY_SIZE(kbl_u_ddi_translations_dp); > > return kbl_u_ddi_translations_dp; > > } else { > > @@ -628,15 +628,15 @@ kbl_get_buf_trans_dp(struct drm_i915_private *dev_priv, int *n_entries) > > } > > static const struct ddi_buf_trans * > > -skl_get_buf_trans_edp(struct drm_i915_private *dev_priv, int *n_entries) > > +skl_get_buf_trans_edp(struct drm_i915_private *i915, int *n_entries) > > { > > - if (dev_priv->vbt.edp.low_vswing) { > > - if (IS_SKL_ULX(dev_priv) || IS_KBL_ULX(dev_priv) || > > - IS_CFL_ULX(dev_priv)) { > > + if (i915->vbt.edp.low_vswing) { > > + if (IS_SKL_ULX(i915) || IS_KBL_ULX(i915) || > > + IS_CFL_ULX(i915)) { > > *n_entries = ARRAY_SIZE(skl_y_ddi_translations_edp); > > return skl_y_ddi_translations_edp; > > - } else if (IS_SKL_ULT(dev_priv) || IS_KBL_ULT(dev_priv) || > > - IS_CFL_ULT(dev_priv)) { > > + } else if (IS_SKL_ULT(i915) || IS_KBL_ULT(i915) || > > + IS_CFL_ULT(i915)) { > > *n_entries = ARRAY_SIZE(skl_u_ddi_translations_edp); > > return skl_u_ddi_translations_edp; > > } else { > > @@ -645,17 +645,17 @@ skl_get_buf_trans_edp(struct drm_i915_private *dev_priv, int *n_entries) > > } > > } > > - if (IS_KABYLAKE(dev_priv) || IS_COFFEELAKE(dev_priv)) > > - return kbl_get_buf_trans_dp(dev_priv, n_entries); > > + if (IS_KABYLAKE(i915) || IS_COFFEELAKE(i915)) > > + return kbl_get_buf_trans_dp(i915, n_entries); > > else > > - return skl_get_buf_trans_dp(dev_priv, n_entries); > > + return skl_get_buf_trans_dp(i915, n_entries); > > } > > static const struct ddi_buf_trans * > > -skl_get_buf_trans_hdmi(struct drm_i915_private *dev_priv, int *n_entries) > > +skl_get_buf_trans_hdmi(struct drm_i915_private *i915, int *n_entries) > > { > > - if (IS_SKL_ULX(dev_priv) || IS_KBL_ULX(dev_priv) || > > - IS_CFL_ULX(dev_priv)) { > > + if (IS_SKL_ULX(i915) || IS_KBL_ULX(i915) || > > + IS_CFL_ULX(i915)) { > > *n_entries = ARRAY_SIZE(skl_y_ddi_translations_hdmi); > > return skl_y_ddi_translations_hdmi; > > } else { > > @@ -674,23 +674,23 @@ static int skl_buf_trans_num_entries(enum port port, int n_entries) > > } > > static const struct ddi_buf_trans * > > -intel_ddi_get_buf_trans_dp(struct drm_i915_private *dev_priv, > > +intel_ddi_get_buf_trans_dp(struct drm_i915_private *i915, > > enum port port, int *n_entries) > > { > > - if (IS_KABYLAKE(dev_priv) || IS_COFFEELAKE(dev_priv)) { > > + if (IS_KABYLAKE(i915) || IS_COFFEELAKE(i915)) { > > const struct ddi_buf_trans *ddi_translations = > > - kbl_get_buf_trans_dp(dev_priv, n_entries); > > + kbl_get_buf_trans_dp(i915, n_entries); > > *n_entries = skl_buf_trans_num_entries(port, *n_entries); > > return ddi_translations; > > - } else if (IS_SKYLAKE(dev_priv)) { > > + } else if (IS_SKYLAKE(i915)) { > > const struct ddi_buf_trans *ddi_translations = > > - skl_get_buf_trans_dp(dev_priv, n_entries); > > + skl_get_buf_trans_dp(i915, n_entries); > > *n_entries = skl_buf_trans_num_entries(port, *n_entries); > > return ddi_translations; > > - } else if (IS_BROADWELL(dev_priv)) { > > + } else if (IS_BROADWELL(i915)) { > > *n_entries = ARRAY_SIZE(bdw_ddi_translations_dp); > > return bdw_ddi_translations_dp; > > - } else if (IS_HASWELL(dev_priv)) { > > + } else if (IS_HASWELL(i915)) { > > *n_entries = ARRAY_SIZE(hsw_ddi_translations_dp); > > return hsw_ddi_translations_dp; > > } > > @@ -700,17 +700,17 @@ intel_ddi_get_buf_trans_dp(struct drm_i915_private *dev_priv, > > } > > static const struct ddi_buf_trans * > > -intel_ddi_get_buf_trans_edp(struct drm_i915_private *dev_priv, > > +intel_ddi_get_buf_trans_edp(struct drm_i915_private *i915, > > enum port port, int *n_entries) > > { > > - if (IS_GEN9_BC(dev_priv)) { > > + if (IS_GEN9_BC(i915)) { > > const struct ddi_buf_trans *ddi_translations = > > - skl_get_buf_trans_edp(dev_priv, n_entries); > > + skl_get_buf_trans_edp(i915, n_entries); > > *n_entries = skl_buf_trans_num_entries(port, *n_entries); > > return ddi_translations; > > - } else if (IS_BROADWELL(dev_priv)) { > > - return bdw_get_buf_trans_edp(dev_priv, n_entries); > > - } else if (IS_HASWELL(dev_priv)) { > > + } else if (IS_BROADWELL(i915)) { > > + return bdw_get_buf_trans_edp(i915, n_entries); > > + } else if (IS_HASWELL(i915)) { > > *n_entries = ARRAY_SIZE(hsw_ddi_translations_dp); > > return hsw_ddi_translations_dp; > > } > > @@ -720,13 +720,13 @@ intel_ddi_get_buf_trans_edp(struct drm_i915_private *dev_priv, > > } > > static const struct ddi_buf_trans * > > -intel_ddi_get_buf_trans_fdi(struct drm_i915_private *dev_priv, > > +intel_ddi_get_buf_trans_fdi(struct drm_i915_private *i915, > > int *n_entries) > > { > > - if (IS_BROADWELL(dev_priv)) { > > + if (IS_BROADWELL(i915)) { > > *n_entries = ARRAY_SIZE(bdw_ddi_translations_fdi); > > return bdw_ddi_translations_fdi; > > - } else if (IS_HASWELL(dev_priv)) { > > + } else if (IS_HASWELL(i915)) { > > *n_entries = ARRAY_SIZE(hsw_ddi_translations_fdi); > > return hsw_ddi_translations_fdi; > > } > > @@ -736,15 +736,15 @@ intel_ddi_get_buf_trans_fdi(struct drm_i915_private *dev_priv, > > } > > static const struct ddi_buf_trans * > > -intel_ddi_get_buf_trans_hdmi(struct drm_i915_private *dev_priv, > > +intel_ddi_get_buf_trans_hdmi(struct drm_i915_private *i915, > > int *n_entries) > > { > > - if (IS_GEN9_BC(dev_priv)) { > > - return skl_get_buf_trans_hdmi(dev_priv, n_entries); > > - } else if (IS_BROADWELL(dev_priv)) { > > + if (IS_GEN9_BC(i915)) { > > + return skl_get_buf_trans_hdmi(i915, n_entries); > > + } else if (IS_BROADWELL(i915)) { > > *n_entries = ARRAY_SIZE(bdw_ddi_translations_hdmi); > > return bdw_ddi_translations_hdmi; > > - } else if (IS_HASWELL(dev_priv)) { > > + } else if (IS_HASWELL(i915)) { > > *n_entries = ARRAY_SIZE(hsw_ddi_translations_hdmi); > > return hsw_ddi_translations_hdmi; > > } > > @@ -754,32 +754,32 @@ intel_ddi_get_buf_trans_hdmi(struct drm_i915_private *dev_priv, > > } > > static const struct bxt_ddi_buf_trans * > > -bxt_get_buf_trans_dp(struct drm_i915_private *dev_priv, int *n_entries) > > +bxt_get_buf_trans_dp(struct drm_i915_private *i915, int *n_entries) > > { > > *n_entries = ARRAY_SIZE(bxt_ddi_translations_dp); > > return bxt_ddi_translations_dp; > > } > > static const struct bxt_ddi_buf_trans * > > -bxt_get_buf_trans_edp(struct drm_i915_private *dev_priv, int *n_entries) > > +bxt_get_buf_trans_edp(struct drm_i915_private *i915, int *n_entries) > > { > > - if (dev_priv->vbt.edp.low_vswing) { > > + if (i915->vbt.edp.low_vswing) { > > *n_entries = ARRAY_SIZE(bxt_ddi_translations_edp); > > return bxt_ddi_translations_edp; > > } > > - return bxt_get_buf_trans_dp(dev_priv, n_entries); > > + return bxt_get_buf_trans_dp(i915, n_entries); > > } > > static const struct bxt_ddi_buf_trans * > > -bxt_get_buf_trans_hdmi(struct drm_i915_private *dev_priv, int *n_entries) > > +bxt_get_buf_trans_hdmi(struct drm_i915_private *i915, int *n_entries) > > { > > *n_entries = ARRAY_SIZE(bxt_ddi_translations_hdmi); > > return bxt_ddi_translations_hdmi; > > } > > static const struct cnl_ddi_buf_trans * > > -cnl_get_buf_trans_hdmi(struct drm_i915_private *dev_priv, int *n_entries) > > +cnl_get_buf_trans_hdmi(struct drm_i915_private *i915, int *n_entries) > > { > > u32 voltage = I915_READ(CNL_PORT_COMP_DW3) & VOLTAGE_INFO_MASK; > > @@ -800,7 +800,7 @@ cnl_get_buf_trans_hdmi(struct drm_i915_private *dev_priv, int *n_entries) > > } > > static const struct cnl_ddi_buf_trans * > > -cnl_get_buf_trans_dp(struct drm_i915_private *dev_priv, int *n_entries) > > +cnl_get_buf_trans_dp(struct drm_i915_private *i915, int *n_entries) > > { > > u32 voltage = I915_READ(CNL_PORT_COMP_DW3) & VOLTAGE_INFO_MASK; > > @@ -821,11 +821,11 @@ cnl_get_buf_trans_dp(struct drm_i915_private *dev_priv, int *n_entries) > > } > > static const struct cnl_ddi_buf_trans * > > -cnl_get_buf_trans_edp(struct drm_i915_private *dev_priv, int *n_entries) > > +cnl_get_buf_trans_edp(struct drm_i915_private *i915, int *n_entries) > > { > > u32 voltage = I915_READ(CNL_PORT_COMP_DW3) & VOLTAGE_INFO_MASK; > > - if (dev_priv->vbt.edp.low_vswing) { > > + if (i915->vbt.edp.low_vswing) { > > if (voltage == VOLTAGE_INFO_0_85V) { > > *n_entries = ARRAY_SIZE(cnl_ddi_translations_edp_0_85V); > > return cnl_ddi_translations_edp_0_85V; > > @@ -841,12 +841,12 @@ cnl_get_buf_trans_edp(struct drm_i915_private *dev_priv, int *n_entries) > > } > > return NULL; > > } else { > > - return cnl_get_buf_trans_dp(dev_priv, n_entries); > > + return cnl_get_buf_trans_dp(i915, n_entries); > > } > > } > > static const struct cnl_ddi_buf_trans * > > -icl_get_combo_buf_trans(struct drm_i915_private *dev_priv, enum port port, > > +icl_get_combo_buf_trans(struct drm_i915_private *i915, enum port port, > > int type, int rate, int *n_entries) > > { > > if (type == INTEL_OUTPUT_HDMI) { > > @@ -855,7 +855,7 @@ icl_get_combo_buf_trans(struct drm_i915_private *dev_priv, enum port port, > > } else if (rate > 540000 && type == INTEL_OUTPUT_EDP) { > > *n_entries = ARRAY_SIZE(icl_combo_phy_ddi_translations_edp_hbr3); > > return icl_combo_phy_ddi_translations_edp_hbr3; > > - } else if (type == INTEL_OUTPUT_EDP && dev_priv->vbt.edp.low_vswing) { > > + } else if (type == INTEL_OUTPUT_EDP && i915->vbt.edp.low_vswing) { > > *n_entries = ARRAY_SIZE(icl_combo_phy_ddi_translations_edp_hbr2); > > return icl_combo_phy_ddi_translations_edp_hbr2; > > } > > @@ -864,33 +864,33 @@ icl_get_combo_buf_trans(struct drm_i915_private *dev_priv, enum port port, > > return icl_combo_phy_ddi_translations_dp_hbr2; > > } > > -static int intel_ddi_hdmi_level(struct drm_i915_private *dev_priv, enum port port) > > +static int intel_ddi_hdmi_level(struct drm_i915_private *i915, enum port port) > > { > > int n_entries, level, default_entry; > > - level = dev_priv->vbt.ddi_port_info[port].hdmi_level_shift; > > + level = i915->vbt.ddi_port_info[port].hdmi_level_shift; > > - if (INTEL_GEN(dev_priv) >= 11) { > > - if (intel_port_is_combophy(dev_priv, port)) > > - icl_get_combo_buf_trans(dev_priv, port, INTEL_OUTPUT_HDMI, > > + if (INTEL_GEN(i915) >= 11) { > > + if (intel_port_is_combophy(i915, port)) > > + icl_get_combo_buf_trans(i915, port, INTEL_OUTPUT_HDMI, > > 0, &n_entries); > > else > > n_entries = ARRAY_SIZE(icl_mg_phy_ddi_translations); > > default_entry = n_entries - 1; > > - } else if (IS_CANNONLAKE(dev_priv)) { > > - cnl_get_buf_trans_hdmi(dev_priv, &n_entries); > > + } else if (IS_CANNONLAKE(i915)) { > > + cnl_get_buf_trans_hdmi(i915, &n_entries); > > default_entry = n_entries - 1; > > - } else if (IS_GEN9_LP(dev_priv)) { > > - bxt_get_buf_trans_hdmi(dev_priv, &n_entries); > > + } else if (IS_GEN9_LP(i915)) { > > + bxt_get_buf_trans_hdmi(i915, &n_entries); > > default_entry = n_entries - 1; > > - } else if (IS_GEN9_BC(dev_priv)) { > > - intel_ddi_get_buf_trans_hdmi(dev_priv, &n_entries); > > + } else if (IS_GEN9_BC(i915)) { > > + intel_ddi_get_buf_trans_hdmi(i915, &n_entries); > > default_entry = 8; > > - } else if (IS_BROADWELL(dev_priv)) { > > - intel_ddi_get_buf_trans_hdmi(dev_priv, &n_entries); > > + } else if (IS_BROADWELL(i915)) { > > + intel_ddi_get_buf_trans_hdmi(i915, &n_entries); > > default_entry = 7; > > - } else if (IS_HASWELL(dev_priv)) { > > - intel_ddi_get_buf_trans_hdmi(dev_priv, &n_entries); > > + } else if (IS_HASWELL(i915)) { > > + intel_ddi_get_buf_trans_hdmi(i915, &n_entries); > > default_entry = 6; > > } else { > > WARN(1, "ddi translation table missing\n"); > > @@ -917,25 +917,25 @@ static int intel_ddi_hdmi_level(struct drm_i915_private *dev_priv, enum port por > > static void intel_prepare_dp_ddi_buffers(struct intel_encoder *encoder, > > const struct intel_crtc_state *crtc_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > u32 iboost_bit = 0; > > int i, n_entries; > > enum port port = encoder->port; > > const struct ddi_buf_trans *ddi_translations; > > if (intel_crtc_has_type(crtc_state, INTEL_OUTPUT_ANALOG)) > > - ddi_translations = intel_ddi_get_buf_trans_fdi(dev_priv, > > + ddi_translations = intel_ddi_get_buf_trans_fdi(i915, > > &n_entries); > > else if (intel_crtc_has_type(crtc_state, INTEL_OUTPUT_EDP)) > > - ddi_translations = intel_ddi_get_buf_trans_edp(dev_priv, port, > > + ddi_translations = intel_ddi_get_buf_trans_edp(i915, port, > > &n_entries); > > else > > - ddi_translations = intel_ddi_get_buf_trans_dp(dev_priv, port, > > + ddi_translations = intel_ddi_get_buf_trans_dp(i915, port, > > &n_entries); > > /* If we're boosting the current, set bit 31 of trans1 */ > > - if (IS_GEN9_BC(dev_priv) && > > - dev_priv->vbt.ddi_port_info[port].dp_boost_level) > > + if (IS_GEN9_BC(i915) && > > + i915->vbt.ddi_port_info[port].dp_boost_level) > > iboost_bit = DDI_BUF_BALANCE_LEG_ENABLE; > > for (i = 0; i < n_entries; i++) { > > @@ -954,13 +954,13 @@ static void intel_prepare_dp_ddi_buffers(struct intel_encoder *encoder, > > static void intel_prepare_hdmi_ddi_buffers(struct intel_encoder *encoder, > > int level) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > u32 iboost_bit = 0; > > int n_entries; > > enum port port = encoder->port; > > const struct ddi_buf_trans *ddi_translations; > > - ddi_translations = intel_ddi_get_buf_trans_hdmi(dev_priv, &n_entries); > > + ddi_translations = intel_ddi_get_buf_trans_hdmi(i915, &n_entries); > > if (WARN_ON_ONCE(!ddi_translations)) > > return; > > @@ -968,8 +968,8 @@ static void intel_prepare_hdmi_ddi_buffers(struct intel_encoder *encoder, > > level = n_entries - 1; > > /* If we're boosting the current, set bit 31 of trans1 */ > > - if (IS_GEN9_BC(dev_priv) && > > - dev_priv->vbt.ddi_port_info[port].hdmi_boost_level) > > + if (IS_GEN9_BC(i915) && > > + i915->vbt.ddi_port_info[port].hdmi_boost_level) > > iboost_bit = DDI_BUF_BALANCE_LEG_ENABLE; > > /* Entry 9 is for HDMI: */ > > @@ -979,7 +979,7 @@ static void intel_prepare_hdmi_ddi_buffers(struct intel_encoder *encoder, > > ddi_translations[level].trans2); > > } > > -static void intel_wait_ddi_buf_idle(struct drm_i915_private *dev_priv, > > +static void intel_wait_ddi_buf_idle(struct drm_i915_private *i915, > > enum port port) > > { > > i915_reg_t reg = DDI_BUF_CTL(port); > > @@ -1064,7 +1064,7 @@ void hsw_fdi_link_train(struct intel_crtc *crtc, > > const struct intel_crtc_state *crtc_state) > > { > > struct drm_device *dev = crtc->base.dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > struct intel_encoder *encoder; > > u32 temp, i, rx_ctl_val, ddi_pll_sel; > > @@ -1085,7 +1085,7 @@ void hsw_fdi_link_train(struct intel_crtc *crtc, > > FDI_RX_TP1_TO_TP2_48 | FDI_RX_FDI_DELAY_90); > > /* Enable the PCH Receiver FDI PLL */ > > - rx_ctl_val = dev_priv->fdi_rx_config | FDI_RX_ENHANCE_FRAME_ENABLE | > > + rx_ctl_val = i915->fdi_rx_config | FDI_RX_ENHANCE_FRAME_ENABLE | > > FDI_RX_PLL_ENABLE | > > FDI_DP_PORT_WIDTH(crtc_state->fdi_lanes); > > I915_WRITE(FDI_RX_CTL(PIPE_A), rx_ctl_val); > > @@ -1174,7 +1174,7 @@ void hsw_fdi_link_train(struct intel_crtc *crtc, > > I915_WRITE(DP_TP_CTL(PORT_E), temp); > > POSTING_READ(DP_TP_CTL(PORT_E)); > > - intel_wait_ddi_buf_idle(dev_priv, PORT_E); > > + intel_wait_ddi_buf_idle(i915, PORT_E); > > /* Reset FDI_RX_MISC pwrdn lanes */ > > temp = I915_READ(FDI_RX_MISC(PIPE_A)); > > @@ -1223,7 +1223,7 @@ intel_ddi_get_crtc_encoder(struct intel_crtc *crtc) > > return ret; > > } > > -static int hsw_ddi_calc_wrpll_link(struct drm_i915_private *dev_priv, > > +static int hsw_ddi_calc_wrpll_link(struct drm_i915_private *i915, > > i915_reg_t reg) > > { > > int refclk; > > @@ -1238,7 +1238,7 @@ static int hsw_ddi_calc_wrpll_link(struct drm_i915_private *dev_priv, > > * non-SSC for non-ULT HSW. Check FUSE_STRAP3 > > * for the non-SSC reference frequency. > > */ > > - if (IS_HASWELL(dev_priv) && !IS_HSW_ULT(dev_priv)) { > > + if (IS_HASWELL(i915) && !IS_HSW_ULT(i915)) { > > if (I915_READ(FUSE_STRAP3) & HSW_REF_CLK_SELECT) > > refclk = 24; > > else > > @@ -1325,7 +1325,7 @@ static int skl_calc_wrpll_link(const struct intel_dpll_hw_state *pll_state) > > return dco_freq / (p0 * p1 * p2 * 5); > > } > > -int cnl_calc_wrpll_link(struct drm_i915_private *dev_priv, > > +int cnl_calc_wrpll_link(struct drm_i915_private *i915, > > struct intel_dpll_hw_state *pll_state) > > { > > u32 p0, p1, p2, dco_freq, ref_clock; > > @@ -1367,7 +1367,7 @@ int cnl_calc_wrpll_link(struct drm_i915_private *dev_priv, > > break; > > } > > - ref_clock = cnl_hdmi_pll_ref_clock(dev_priv); > > + ref_clock = cnl_hdmi_pll_ref_clock(i915); > > dco_freq = (pll_state->cfgcr0 & DPLL_CFGCR0_DCO_INTEGER_MASK) > > * ref_clock; > > @@ -1381,7 +1381,7 @@ int cnl_calc_wrpll_link(struct drm_i915_private *dev_priv, > > return dco_freq / (p0 * p1 * p2 * 5); > > } > > -static int icl_calc_tbt_pll_link(struct drm_i915_private *dev_priv, > > +static int icl_calc_tbt_pll_link(struct drm_i915_private *i915, > > enum port port) > > { > > u32 val = I915_READ(DDI_CLK_SEL(port)) & DDI_CLK_SEL_MASK; > > @@ -1403,13 +1403,13 @@ static int icl_calc_tbt_pll_link(struct drm_i915_private *dev_priv, > > } > > } > > -static int icl_calc_mg_pll_link(struct drm_i915_private *dev_priv, > > +static int icl_calc_mg_pll_link(struct drm_i915_private *i915, > > const struct intel_dpll_hw_state *pll_state) > > { > > u32 m1, m2_int, m2_frac, div1, div2, ref_clock; > > u64 tmp; > > - ref_clock = dev_priv->cdclk.hw.ref; > > + ref_clock = i915->cdclk.hw.ref; > > m1 = pll_state->mg_pll_div1 & MG_PLL_DIV1_FBPREDIV_MASK; > > m2_int = pll_state->mg_pll_div0 & MG_PLL_DIV0_FBDIV_INT_MASK; > > @@ -1483,21 +1483,21 @@ static void ddi_dotclock_get(struct intel_crtc_state *pipe_config) > > static void icl_ddi_clock_get(struct intel_encoder *encoder, > > struct intel_crtc_state *pipe_config) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_dpll_hw_state *pll_state = &pipe_config->dpll_hw_state; > > enum port port = encoder->port; > > int link_clock; > > - if (intel_port_is_combophy(dev_priv, port)) { > > - link_clock = cnl_calc_wrpll_link(dev_priv, pll_state); > > + if (intel_port_is_combophy(i915, port)) { > > + link_clock = cnl_calc_wrpll_link(i915, pll_state); > > } else { > > - enum intel_dpll_id pll_id = intel_get_shared_dpll_id(dev_priv, > > + enum intel_dpll_id pll_id = intel_get_shared_dpll_id(i915, > > pipe_config->shared_dpll); > > if (pll_id == DPLL_ID_ICL_TBTPLL) > > - link_clock = icl_calc_tbt_pll_link(dev_priv, port); > > + link_clock = icl_calc_tbt_pll_link(i915, port); > > else > > - link_clock = icl_calc_mg_pll_link(dev_priv, pll_state); > > + link_clock = icl_calc_mg_pll_link(i915, pll_state); > > } > > pipe_config->port_clock = link_clock; > > @@ -1508,12 +1508,12 @@ static void icl_ddi_clock_get(struct intel_encoder *encoder, > > static void cnl_ddi_clock_get(struct intel_encoder *encoder, > > struct intel_crtc_state *pipe_config) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_dpll_hw_state *pll_state = &pipe_config->dpll_hw_state; > > int link_clock; > > if (pll_state->cfgcr0 & DPLL_CFGCR0_HDMI_MODE) { > > - link_clock = cnl_calc_wrpll_link(dev_priv, pll_state); > > + link_clock = cnl_calc_wrpll_link(i915, pll_state); > > } else { > > link_clock = pll_state->cfgcr0 & DPLL_CFGCR0_LINK_RATE_MASK; > > @@ -1604,7 +1604,7 @@ static void skl_ddi_clock_get(struct intel_encoder *encoder, > > static void hsw_ddi_clock_get(struct intel_encoder *encoder, > > struct intel_crtc_state *pipe_config) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > int link_clock = 0; > > u32 val, pll; > > @@ -1620,10 +1620,10 @@ static void hsw_ddi_clock_get(struct intel_encoder *encoder, > > link_clock = 270000; > > break; > > case PORT_CLK_SEL_WRPLL1: > > - link_clock = hsw_ddi_calc_wrpll_link(dev_priv, WRPLL_CTL(0)); > > + link_clock = hsw_ddi_calc_wrpll_link(i915, WRPLL_CTL(0)); > > break; > > case PORT_CLK_SEL_WRPLL2: > > - link_clock = hsw_ddi_calc_wrpll_link(dev_priv, WRPLL_CTL(1)); > > + link_clock = hsw_ddi_calc_wrpll_link(i915, WRPLL_CTL(1)); > > break; > > case PORT_CLK_SEL_SPLL: > > pll = I915_READ(SPLL_CTL) & SPLL_FREQ_MASK; > > @@ -1675,24 +1675,24 @@ static void bxt_ddi_clock_get(struct intel_encoder *encoder, > > static void intel_ddi_clock_get(struct intel_encoder *encoder, > > struct intel_crtc_state *pipe_config) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > - if (INTEL_GEN(dev_priv) >= 11) > > + if (INTEL_GEN(i915) >= 11) > > icl_ddi_clock_get(encoder, pipe_config); > > - else if (IS_CANNONLAKE(dev_priv)) > > + else if (IS_CANNONLAKE(i915)) > > cnl_ddi_clock_get(encoder, pipe_config); > > - else if (IS_GEN9_LP(dev_priv)) > > + else if (IS_GEN9_LP(i915)) > > bxt_ddi_clock_get(encoder, pipe_config); > > - else if (IS_GEN9_BC(dev_priv)) > > + else if (IS_GEN9_BC(i915)) > > skl_ddi_clock_get(encoder, pipe_config); > > - else if (INTEL_GEN(dev_priv) <= 8) > > + else if (INTEL_GEN(i915) <= 8) > > hsw_ddi_clock_get(encoder, pipe_config); > > } > > void intel_ddi_set_pipe_settings(const struct intel_crtc_state *crtc_state) > > { > > struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc); > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > enum transcoder cpu_transcoder = crtc_state->cpu_transcoder; > > u32 temp; > > @@ -1746,7 +1746,7 @@ void intel_ddi_set_vc_payload_alloc(const struct intel_crtc_state *crtc_state, > > bool state) > > { > > struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc); > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > enum transcoder cpu_transcoder = crtc_state->cpu_transcoder; > > u32 temp; > > @@ -1762,7 +1762,7 @@ void intel_ddi_enable_transcoder_func(const struct intel_crtc_state *crtc_state) > > { > > struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc); > > struct intel_encoder *encoder = intel_ddi_get_crtc_encoder(crtc); > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > enum pipe pipe = crtc->pipe; > > enum transcoder cpu_transcoder = crtc_state->cpu_transcoder; > > enum port port = encoder->port; > > @@ -1845,7 +1845,7 @@ void intel_ddi_enable_transcoder_func(const struct intel_crtc_state *crtc_state) > > void intel_ddi_disable_transcoder_func(const struct intel_crtc_state *crtc_state) > > { > > struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc); > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > enum transcoder cpu_transcoder = crtc_state->cpu_transcoder; > > i915_reg_t reg = TRANS_DDI_FUNC_CTL(cpu_transcoder); > > u32 val = I915_READ(reg); > > @@ -1854,7 +1854,7 @@ void intel_ddi_disable_transcoder_func(const struct intel_crtc_state *crtc_state > > val |= TRANS_DDI_PORT_NONE; > > I915_WRITE(reg, val); > > - if (dev_priv->quirks & QUIRK_INCREASE_DDI_DISABLED_TIME && > > + if (i915->quirks & QUIRK_INCREASE_DDI_DISABLED_TIME && > > intel_crtc_has_type(crtc_state, INTEL_OUTPUT_HDMI)) { > > DRM_DEBUG_KMS("Quirk Increase DDI disabled time\n"); > > /* Quirk time at 100ms for reliable operation */ > > @@ -1866,13 +1866,13 @@ int intel_ddi_toggle_hdcp_signalling(struct intel_encoder *intel_encoder, > > bool enable) > > { > > struct drm_device *dev = intel_encoder->base.dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > intel_wakeref_t wakeref; > > enum pipe pipe = 0; > > int ret = 0; > > u32 tmp; > > - wakeref = intel_display_power_get_if_enabled(dev_priv, > > + wakeref = intel_display_power_get_if_enabled(i915, > > intel_encoder->power_domain); > > if (WARN_ON(!wakeref)) > > return -ENXIO; > > @@ -1889,14 +1889,14 @@ int intel_ddi_toggle_hdcp_signalling(struct intel_encoder *intel_encoder, > > tmp &= ~TRANS_DDI_HDCP_SIGNALLING; > > I915_WRITE(TRANS_DDI_FUNC_CTL(pipe), tmp); > > out: > > - intel_display_power_put(dev_priv, intel_encoder->power_domain, wakeref); > > + intel_display_power_put(i915, intel_encoder->power_domain, wakeref); > > return ret; > > } > > bool intel_ddi_connector_get_hw_state(struct intel_connector *intel_connector) > > { > > struct drm_device *dev = intel_connector->base.dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > struct intel_encoder *encoder = intel_connector->encoder; > > int type = intel_connector->base.connector_type; > > enum port port = encoder->port; > > @@ -1906,7 +1906,7 @@ bool intel_ddi_connector_get_hw_state(struct intel_connector *intel_connector) > > u32 tmp; > > bool ret; > > - wakeref = intel_display_power_get_if_enabled(dev_priv, > > + wakeref = intel_display_power_get_if_enabled(i915, > > encoder->power_domain); > > if (!wakeref) > > return false; > > @@ -1916,7 +1916,7 @@ bool intel_ddi_connector_get_hw_state(struct intel_connector *intel_connector) > > goto out; > > } > > - if (HAS_TRANSCODER_EDP(dev_priv) && port == PORT_A) > > + if (HAS_TRANSCODER_EDP(i915) && port == PORT_A) > > cpu_transcoder = TRANSCODER_EDP; > > else > > cpu_transcoder = (enum transcoder) pipe; > > @@ -1950,7 +1950,7 @@ bool intel_ddi_connector_get_hw_state(struct intel_connector *intel_connector) > > } > > out: > > - intel_display_power_put(dev_priv, encoder->power_domain, wakeref); > > + intel_display_power_put(i915, encoder->power_domain, wakeref); > > return ret; > > } > > @@ -1959,7 +1959,7 @@ static void intel_ddi_get_encoder_pipes(struct intel_encoder *encoder, > > u8 *pipe_mask, bool *is_dp_mst) > > { > > struct drm_device *dev = encoder->base.dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > enum port port = encoder->port; > > intel_wakeref_t wakeref; > > enum pipe p; > > @@ -1969,7 +1969,7 @@ static void intel_ddi_get_encoder_pipes(struct intel_encoder *encoder, > > *pipe_mask = 0; > > *is_dp_mst = false; > > - wakeref = intel_display_power_get_if_enabled(dev_priv, > > + wakeref = intel_display_power_get_if_enabled(i915, > > encoder->power_domain); > > if (!wakeref) > > return; > > @@ -1978,7 +1978,7 @@ static void intel_ddi_get_encoder_pipes(struct intel_encoder *encoder, > > if (!(tmp & DDI_BUF_CTL_ENABLE)) > > goto out; > > - if (HAS_TRANSCODER_EDP(dev_priv) && port == PORT_A) { > > + if (HAS_TRANSCODER_EDP(i915) && port == PORT_A) { > > tmp = I915_READ(TRANS_DDI_FUNC_CTL(TRANSCODER_EDP)); > > switch (tmp & TRANS_DDI_EDP_INPUT_MASK) { > > @@ -2001,7 +2001,7 @@ static void intel_ddi_get_encoder_pipes(struct intel_encoder *encoder, > > } > > mst_pipe_mask = 0; > > - for_each_pipe(dev_priv, p) { > > + for_each_pipe(i915, p) { > > enum transcoder cpu_transcoder = (enum transcoder)p; > > tmp = I915_READ(TRANS_DDI_FUNC_CTL(cpu_transcoder)); > > @@ -2033,7 +2033,7 @@ static void intel_ddi_get_encoder_pipes(struct intel_encoder *encoder, > > *is_dp_mst = mst_pipe_mask; > > out: > > - if (*pipe_mask && IS_GEN9_LP(dev_priv)) { > > + if (*pipe_mask && IS_GEN9_LP(i915)) { > > tmp = I915_READ(BXT_PHY_CTL(port)); > > if ((tmp & (BXT_PHY_CMNLANE_POWERDOWN_ACK | > > BXT_PHY_LANE_POWERDOWN_ACK | > > @@ -2042,7 +2042,7 @@ static void intel_ddi_get_encoder_pipes(struct intel_encoder *encoder, > > "(PHY_CTL %08x)\n", port_name(port), tmp); > > } > > - intel_display_power_put(dev_priv, encoder->power_domain, wakeref); > > + intel_display_power_put(i915, encoder->power_domain, wakeref); > > } > > bool intel_ddi_get_hw_state(struct intel_encoder *encoder, > > @@ -2083,7 +2083,7 @@ intel_ddi_main_link_aux_domain(struct intel_digital_port *dig_port) > > static void intel_ddi_get_power_domains(struct intel_encoder *encoder, > > struct intel_crtc_state *crtc_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_digital_port *dig_port; > > /* > > @@ -2095,29 +2095,29 @@ static void intel_ddi_get_power_domains(struct intel_encoder *encoder, > > return; > > dig_port = enc_to_dig_port(&encoder->base); > > - intel_display_power_get(dev_priv, dig_port->ddi_io_power_domain); > > + intel_display_power_get(i915, dig_port->ddi_io_power_domain); > > /* > > * AUX power is only needed for (e)DP mode, and for HDMI mode on TC > > * ports. > > */ > > if (intel_crtc_has_dp_encoder(crtc_state) || > > - intel_port_is_tc(dev_priv, encoder->port)) > > - intel_display_power_get(dev_priv, > > + intel_port_is_tc(i915, encoder->port)) > > + intel_display_power_get(i915, > > intel_ddi_main_link_aux_domain(dig_port)); > > /* > > * VDSC power is needed when DSC is enabled > > */ > > if (crtc_state->dsc_params.compression_enable) > > - intel_display_power_get(dev_priv, > > + intel_display_power_get(i915, > > intel_dsc_power_domain(crtc_state)); > > } > > void intel_ddi_enable_pipe_clock(const struct intel_crtc_state *crtc_state) > > { > > struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc); > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > struct intel_encoder *encoder = intel_ddi_get_crtc_encoder(crtc); > > enum port port = encoder->port; > > enum transcoder cpu_transcoder = crtc_state->cpu_transcoder; > > @@ -2129,7 +2129,7 @@ void intel_ddi_enable_pipe_clock(const struct intel_crtc_state *crtc_state) > > void intel_ddi_disable_pipe_clock(const struct intel_crtc_state *crtc_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc_state->base.crtc->dev); > > + struct drm_i915_private *i915 = to_i915(crtc_state->base.crtc->dev); > > enum transcoder cpu_transcoder = crtc_state->cpu_transcoder; > > if (cpu_transcoder != TRANSCODER_EDP) > > @@ -2137,7 +2137,7 @@ void intel_ddi_disable_pipe_clock(const struct intel_crtc_state *crtc_state) > > TRANS_CLK_SEL_DISABLED); > > } > > -static void _skl_ddi_set_iboost(struct drm_i915_private *dev_priv, > > +static void _skl_ddi_set_iboost(struct drm_i915_private *i915, > > enum port port, u8 iboost) > > { > > u32 tmp; > > @@ -2155,25 +2155,25 @@ static void skl_ddi_set_iboost(struct intel_encoder *encoder, > > int level, enum intel_output_type type) > > { > > struct intel_digital_port *intel_dig_port = enc_to_dig_port(&encoder->base); > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > enum port port = encoder->port; > > u8 iboost; > > if (type == INTEL_OUTPUT_HDMI) > > - iboost = dev_priv->vbt.ddi_port_info[port].hdmi_boost_level; > > + iboost = i915->vbt.ddi_port_info[port].hdmi_boost_level; > > else > > - iboost = dev_priv->vbt.ddi_port_info[port].dp_boost_level; > > + iboost = i915->vbt.ddi_port_info[port].dp_boost_level; > > if (iboost == 0) { > > const struct ddi_buf_trans *ddi_translations; > > int n_entries; > > if (type == INTEL_OUTPUT_HDMI) > > - ddi_translations = intel_ddi_get_buf_trans_hdmi(dev_priv, &n_entries); > > + ddi_translations = intel_ddi_get_buf_trans_hdmi(i915, &n_entries); > > else if (type == INTEL_OUTPUT_EDP) > > - ddi_translations = intel_ddi_get_buf_trans_edp(dev_priv, port, &n_entries); > > + ddi_translations = intel_ddi_get_buf_trans_edp(i915, port, &n_entries); > > else > > - ddi_translations = intel_ddi_get_buf_trans_dp(dev_priv, port, &n_entries); > > + ddi_translations = intel_ddi_get_buf_trans_dp(i915, port, &n_entries); > > if (WARN_ON_ONCE(!ddi_translations)) > > return; > > @@ -2189,33 +2189,33 @@ static void skl_ddi_set_iboost(struct intel_encoder *encoder, > > return; > > } > > - _skl_ddi_set_iboost(dev_priv, port, iboost); > > + _skl_ddi_set_iboost(i915, port, iboost); > > if (port == PORT_A && intel_dig_port->max_lanes == 4) > > - _skl_ddi_set_iboost(dev_priv, PORT_E, iboost); > > + _skl_ddi_set_iboost(i915, PORT_E, iboost); > > } > > static void bxt_ddi_vswing_sequence(struct intel_encoder *encoder, > > int level, enum intel_output_type type) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > const struct bxt_ddi_buf_trans *ddi_translations; > > enum port port = encoder->port; > > int n_entries; > > if (type == INTEL_OUTPUT_HDMI) > > - ddi_translations = bxt_get_buf_trans_hdmi(dev_priv, &n_entries); > > + ddi_translations = bxt_get_buf_trans_hdmi(i915, &n_entries); > > else if (type == INTEL_OUTPUT_EDP) > > - ddi_translations = bxt_get_buf_trans_edp(dev_priv, &n_entries); > > + ddi_translations = bxt_get_buf_trans_edp(i915, &n_entries); > > else > > - ddi_translations = bxt_get_buf_trans_dp(dev_priv, &n_entries); > > + ddi_translations = bxt_get_buf_trans_dp(i915, &n_entries); > > if (WARN_ON_ONCE(!ddi_translations)) > > return; > > if (WARN_ON_ONCE(level >= n_entries)) > > level = n_entries - 1; > > - bxt_ddi_phy_set_signal_level(dev_priv, port, > > + bxt_ddi_phy_set_signal_level(i915, port, > > ddi_translations[level].margin, > > ddi_translations[level].scale, > > ddi_translations[level].enable, > > @@ -2224,32 +2224,32 @@ static void bxt_ddi_vswing_sequence(struct intel_encoder *encoder, > > u8 intel_ddi_dp_voltage_max(struct intel_encoder *encoder) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_dp *intel_dp = enc_to_intel_dp(&encoder->base); > > enum port port = encoder->port; > > int n_entries; > > - if (INTEL_GEN(dev_priv) >= 11) { > > - if (intel_port_is_combophy(dev_priv, port)) > > - icl_get_combo_buf_trans(dev_priv, port, encoder->type, > > + if (INTEL_GEN(i915) >= 11) { > > + if (intel_port_is_combophy(i915, port)) > > + icl_get_combo_buf_trans(i915, port, encoder->type, > > intel_dp->link_rate, &n_entries); > > else > > n_entries = ARRAY_SIZE(icl_mg_phy_ddi_translations); > > - } else if (IS_CANNONLAKE(dev_priv)) { > > + } else if (IS_CANNONLAKE(i915)) { > > if (encoder->type == INTEL_OUTPUT_EDP) > > - cnl_get_buf_trans_edp(dev_priv, &n_entries); > > + cnl_get_buf_trans_edp(i915, &n_entries); > > else > > - cnl_get_buf_trans_dp(dev_priv, &n_entries); > > - } else if (IS_GEN9_LP(dev_priv)) { > > + cnl_get_buf_trans_dp(i915, &n_entries); > > + } else if (IS_GEN9_LP(i915)) { > > if (encoder->type == INTEL_OUTPUT_EDP) > > - bxt_get_buf_trans_edp(dev_priv, &n_entries); > > + bxt_get_buf_trans_edp(i915, &n_entries); > > else > > - bxt_get_buf_trans_dp(dev_priv, &n_entries); > > + bxt_get_buf_trans_dp(i915, &n_entries); > > } else { > > if (encoder->type == INTEL_OUTPUT_EDP) > > - intel_ddi_get_buf_trans_edp(dev_priv, port, &n_entries); > > + intel_ddi_get_buf_trans_edp(i915, port, &n_entries); > > else > > - intel_ddi_get_buf_trans_dp(dev_priv, port, &n_entries); > > + intel_ddi_get_buf_trans_dp(i915, port, &n_entries); > > } > > if (WARN_ON(n_entries < 1)) > > @@ -2284,18 +2284,18 @@ u8 intel_ddi_dp_pre_emphasis_max(struct intel_encoder *encoder, u8 voltage_swing > > static void cnl_ddi_vswing_program(struct intel_encoder *encoder, > > int level, enum intel_output_type type) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > const struct cnl_ddi_buf_trans *ddi_translations; > > enum port port = encoder->port; > > int n_entries, ln; > > u32 val; > > if (type == INTEL_OUTPUT_HDMI) > > - ddi_translations = cnl_get_buf_trans_hdmi(dev_priv, &n_entries); > > + ddi_translations = cnl_get_buf_trans_hdmi(i915, &n_entries); > > else if (type == INTEL_OUTPUT_EDP) > > - ddi_translations = cnl_get_buf_trans_edp(dev_priv, &n_entries); > > + ddi_translations = cnl_get_buf_trans_edp(i915, &n_entries); > > else > > - ddi_translations = cnl_get_buf_trans_dp(dev_priv, &n_entries); > > + ddi_translations = cnl_get_buf_trans_dp(i915, &n_entries); > > if (WARN_ON_ONCE(!ddi_translations)) > > return; > > @@ -2348,7 +2348,7 @@ static void cnl_ddi_vswing_program(struct intel_encoder *encoder, > > static void cnl_ddi_vswing_sequence(struct intel_encoder *encoder, > > int level, enum intel_output_type type) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > enum port port = encoder->port; > > int width, rate, ln; > > u32 val; > > @@ -2412,7 +2412,7 @@ static void cnl_ddi_vswing_sequence(struct intel_encoder *encoder, > > I915_WRITE(CNL_PORT_TX_DW5_GRP(port), val); > > } > > -static void icl_ddi_combo_vswing_program(struct drm_i915_private *dev_priv, > > +static void icl_ddi_combo_vswing_program(struct drm_i915_private *i915, > > u32 level, enum port port, int type, > > int rate) > > { > > @@ -2420,7 +2420,7 @@ static void icl_ddi_combo_vswing_program(struct drm_i915_private *dev_priv, > > u32 n_entries, val; > > int ln; > > - ddi_translations = icl_get_combo_buf_trans(dev_priv, port, type, > > + ddi_translations = icl_get_combo_buf_trans(i915, port, type, > > rate, &n_entries); > > if (!ddi_translations) > > return; > > @@ -2472,7 +2472,7 @@ static void icl_combo_phy_ddi_vswing_sequence(struct intel_encoder *encoder, > > u32 level, > > enum intel_output_type type) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > enum port port = encoder->port; > > int width = 0; > > int rate = 0; > > @@ -2530,7 +2530,7 @@ static void icl_combo_phy_ddi_vswing_sequence(struct intel_encoder *encoder, > > I915_WRITE(ICL_PORT_TX_DW5_GRP(port), val); > > /* 5. Program swing and de-emphasis */ > > - icl_ddi_combo_vswing_program(dev_priv, level, port, type, rate); > > + icl_ddi_combo_vswing_program(i915, level, port, type, rate); > > /* 6. Set training enable to trigger update */ > > val = I915_READ(ICL_PORT_TX_DW5_LN0(port)); > > @@ -2542,7 +2542,7 @@ static void icl_mg_phy_ddi_vswing_sequence(struct intel_encoder *encoder, > > int link_clock, > > u32 level) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > enum port port = encoder->port; > > const struct icl_mg_phy_ddi_buf_trans *ddi_translations; > > u32 n_entries, val; > > @@ -2662,10 +2662,10 @@ static void icl_ddi_vswing_sequence(struct intel_encoder *encoder, > > u32 level, > > enum intel_output_type type) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > enum port port = encoder->port; > > - if (intel_port_is_combophy(dev_priv, port)) > > + if (intel_port_is_combophy(i915, port)) > > icl_combo_phy_ddi_vswing_sequence(encoder, level, type); > > else > > icl_mg_phy_ddi_vswing_sequence(encoder, link_clock, level); > > @@ -2698,14 +2698,14 @@ static u32 intel_ddi_dp_level(struct intel_dp *intel_dp) > > u32 bxt_signal_levels(struct intel_dp *intel_dp) > > { > > struct intel_digital_port *dport = dp_to_dig_port(intel_dp); > > - struct drm_i915_private *dev_priv = to_i915(dport->base.base.dev); > > + struct drm_i915_private *i915 = to_i915(dport->base.base.dev); > > struct intel_encoder *encoder = &dport->base; > > int level = intel_ddi_dp_level(intel_dp); > > - if (INTEL_GEN(dev_priv) >= 11) > > + if (INTEL_GEN(i915) >= 11) > > icl_ddi_vswing_sequence(encoder, intel_dp->link_rate, > > level, encoder->type); > > - else if (IS_CANNONLAKE(dev_priv)) > > + else if (IS_CANNONLAKE(i915)) > > cnl_ddi_vswing_sequence(encoder, level, encoder->type); > > else > > bxt_ddi_vswing_sequence(encoder, level, encoder->type); > > @@ -2716,24 +2716,24 @@ u32 bxt_signal_levels(struct intel_dp *intel_dp) > > u32 ddi_signal_levels(struct intel_dp *intel_dp) > > { > > struct intel_digital_port *dport = dp_to_dig_port(intel_dp); > > - struct drm_i915_private *dev_priv = to_i915(dport->base.base.dev); > > + struct drm_i915_private *i915 = to_i915(dport->base.base.dev); > > struct intel_encoder *encoder = &dport->base; > > int level = intel_ddi_dp_level(intel_dp); > > - if (IS_GEN9_BC(dev_priv)) > > + if (IS_GEN9_BC(i915)) > > skl_ddi_set_iboost(encoder, level, encoder->type); > > return DDI_BUF_TRANS_SELECT(level); > > } > > static inline > > -u32 icl_dpclka_cfgcr0_clk_off(struct drm_i915_private *dev_priv, > > +u32 icl_dpclka_cfgcr0_clk_off(struct drm_i915_private *i915, > > enum port port) > > { > > - if (intel_port_is_combophy(dev_priv, port)) { > > + if (intel_port_is_combophy(i915, port)) { > > return ICL_DPCLKA_CFGCR0_DDI_CLK_OFF(port); > > - } else if (intel_port_is_tc(dev_priv, port)) { > > - enum tc_port tc_port = intel_port_to_tc(dev_priv, port); > > + } else if (intel_port_is_tc(i915, port)) { > > + enum tc_port tc_port = intel_port_to_tc(i915, port); > > return ICL_DPCLKA_CFGCR0_TC_CLK_OFF(tc_port); > > } > > @@ -2744,47 +2744,47 @@ u32 icl_dpclka_cfgcr0_clk_off(struct drm_i915_private *dev_priv, > > static void icl_map_plls_to_ports(struct intel_encoder *encoder, > > const struct intel_crtc_state *crtc_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_shared_dpll *pll = crtc_state->shared_dpll; > > enum port port = encoder->port; > > u32 val; > > - mutex_lock(&dev_priv->dpll_lock); > > + mutex_lock(&i915->dpll_lock); > > val = I915_READ(DPCLKA_CFGCR0_ICL); > > - WARN_ON((val & icl_dpclka_cfgcr0_clk_off(dev_priv, port)) == 0); > > + WARN_ON((val & icl_dpclka_cfgcr0_clk_off(i915, port)) == 0); > > - if (intel_port_is_combophy(dev_priv, port)) { > > + if (intel_port_is_combophy(i915, port)) { > > val &= ~DPCLKA_CFGCR0_DDI_CLK_SEL_MASK(port); > > val |= DPCLKA_CFGCR0_DDI_CLK_SEL(pll->info->id, port); > > I915_WRITE(DPCLKA_CFGCR0_ICL, val); > > POSTING_READ(DPCLKA_CFGCR0_ICL); > > } > > - val &= ~icl_dpclka_cfgcr0_clk_off(dev_priv, port); > > + val &= ~icl_dpclka_cfgcr0_clk_off(i915, port); > > I915_WRITE(DPCLKA_CFGCR0_ICL, val); > > - mutex_unlock(&dev_priv->dpll_lock); > > + mutex_unlock(&i915->dpll_lock); > > } > > static void icl_unmap_plls_to_ports(struct intel_encoder *encoder) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > enum port port = encoder->port; > > u32 val; > > - mutex_lock(&dev_priv->dpll_lock); > > + mutex_lock(&i915->dpll_lock); > > val = I915_READ(DPCLKA_CFGCR0_ICL); > > - val |= icl_dpclka_cfgcr0_clk_off(dev_priv, port); > > + val |= icl_dpclka_cfgcr0_clk_off(i915, port); > > I915_WRITE(DPCLKA_CFGCR0_ICL, val); > > - mutex_unlock(&dev_priv->dpll_lock); > > + mutex_unlock(&i915->dpll_lock); > > } > > void icl_sanitize_encoder_pll_mapping(struct intel_encoder *encoder) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > u32 val; > > enum port port; > > u32 port_mask; > > @@ -2821,7 +2821,7 @@ void icl_sanitize_encoder_pll_mapping(struct intel_encoder *encoder) > > * Sanity check that we haven't incorrectly registered another > > * encoder using any of the ports of this DSI encoder. > > */ > > - for_each_intel_encoder(&dev_priv->drm, other_encoder) { > > + for_each_intel_encoder(&i915->drm, other_encoder) { > > if (other_encoder == encoder) > > continue; > > @@ -2838,7 +2838,7 @@ void icl_sanitize_encoder_pll_mapping(struct intel_encoder *encoder) > > val = I915_READ(DPCLKA_CFGCR0_ICL); > > for_each_port_masked(port, port_mask) { > > bool ddi_clk_ungated = !(val & > > - icl_dpclka_cfgcr0_clk_off(dev_priv, > > + icl_dpclka_cfgcr0_clk_off(i915, > > port)); > > if (ddi_clk_needed == ddi_clk_ungated) > > @@ -2853,7 +2853,7 @@ void icl_sanitize_encoder_pll_mapping(struct intel_encoder *encoder) > > DRM_NOTE("Port %c is disabled/in DSI mode with an ungated DDI clock, gate it\n", > > port_name(port)); > > - val |= icl_dpclka_cfgcr0_clk_off(dev_priv, port); > > + val |= icl_dpclka_cfgcr0_clk_off(i915, port); > > I915_WRITE(DPCLKA_CFGCR0_ICL, val); > > } > > } > > @@ -2861,7 +2861,7 @@ void icl_sanitize_encoder_pll_mapping(struct intel_encoder *encoder) > > static void intel_ddi_clk_select(struct intel_encoder *encoder, > > const struct intel_crtc_state *crtc_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > enum port port = encoder->port; > > u32 val; > > const struct intel_shared_dpll *pll = crtc_state->shared_dpll; > > @@ -2869,13 +2869,13 @@ static void intel_ddi_clk_select(struct intel_encoder *encoder, > > if (WARN_ON(!pll)) > > return; > > - mutex_lock(&dev_priv->dpll_lock); > > + mutex_lock(&i915->dpll_lock); > > - if (INTEL_GEN(dev_priv) >= 11) { > > - if (!intel_port_is_combophy(dev_priv, port)) > > + if (INTEL_GEN(i915) >= 11) { > > + if (!intel_port_is_combophy(i915, port)) > > I915_WRITE(DDI_CLK_SEL(port), > > icl_pll_to_ddi_clk_sel(encoder, crtc_state)); > > - } else if (IS_CANNONLAKE(dev_priv)) { > > + } else if (IS_CANNONLAKE(i915)) { > > /* Configure DPCLKA_CFGCR0 to map the DPLL to the DDI. */ > > val = I915_READ(DPCLKA_CFGCR0); > > val &= ~DPCLKA_CFGCR0_DDI_CLK_SEL_MASK(port); > > @@ -2890,7 +2890,7 @@ static void intel_ddi_clk_select(struct intel_encoder *encoder, > > val = I915_READ(DPCLKA_CFGCR0); > > val &= ~DPCLKA_CFGCR0_DDI_CLK_OFF(port); > > I915_WRITE(DPCLKA_CFGCR0, val); > > - } else if (IS_GEN9_BC(dev_priv)) { > > + } else if (IS_GEN9_BC(i915)) { > > /* DDI -> PLL mapping */ > > val = I915_READ(DPLL_CTRL2); > > @@ -2901,37 +2901,37 @@ static void intel_ddi_clk_select(struct intel_encoder *encoder, > > I915_WRITE(DPLL_CTRL2, val); > > - } else if (INTEL_GEN(dev_priv) < 9) { > > + } else if (INTEL_GEN(i915) < 9) { > > I915_WRITE(PORT_CLK_SEL(port), hsw_pll_to_ddi_pll_sel(pll)); > > } > > - mutex_unlock(&dev_priv->dpll_lock); > > + mutex_unlock(&i915->dpll_lock); > > } > > static void intel_ddi_clk_disable(struct intel_encoder *encoder) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > enum port port = encoder->port; > > - if (INTEL_GEN(dev_priv) >= 11) { > > - if (!intel_port_is_combophy(dev_priv, port)) > > + if (INTEL_GEN(i915) >= 11) { > > + if (!intel_port_is_combophy(i915, port)) > > I915_WRITE(DDI_CLK_SEL(port), DDI_CLK_SEL_NONE); > > - } else if (IS_CANNONLAKE(dev_priv)) { > > + } else if (IS_CANNONLAKE(i915)) { > > I915_WRITE(DPCLKA_CFGCR0, I915_READ(DPCLKA_CFGCR0) | > > DPCLKA_CFGCR0_DDI_CLK_OFF(port)); > > - } else if (IS_GEN9_BC(dev_priv)) { > > + } else if (IS_GEN9_BC(i915)) { > > I915_WRITE(DPLL_CTRL2, I915_READ(DPLL_CTRL2) | > > DPLL_CTRL2_DDI_CLK_OFF(port)); > > - } else if (INTEL_GEN(dev_priv) < 9) { > > + } else if (INTEL_GEN(i915) < 9) { > > I915_WRITE(PORT_CLK_SEL(port), PORT_CLK_SEL_NONE); > > } > > } > > static void icl_enable_phy_clock_gating(struct intel_digital_port *dig_port) > > { > > - struct drm_i915_private *dev_priv = to_i915(dig_port->base.base.dev); > > + struct drm_i915_private *i915 = to_i915(dig_port->base.base.dev); > > enum port port = dig_port->base.port; > > - enum tc_port tc_port = intel_port_to_tc(dev_priv, port); > > + enum tc_port tc_port = intel_port_to_tc(i915, port); > > u32 val; > > int ln; > > @@ -2961,9 +2961,9 @@ static void icl_enable_phy_clock_gating(struct intel_digital_port *dig_port) > > static void icl_disable_phy_clock_gating(struct intel_digital_port *dig_port) > > { > > - struct drm_i915_private *dev_priv = to_i915(dig_port->base.base.dev); > > + struct drm_i915_private *i915 = to_i915(dig_port->base.base.dev); > > enum port port = dig_port->base.port; > > - enum tc_port tc_port = intel_port_to_tc(dev_priv, port); > > + enum tc_port tc_port = intel_port_to_tc(i915, port); > > u32 val; > > int ln; > > @@ -2993,9 +2993,9 @@ static void icl_disable_phy_clock_gating(struct intel_digital_port *dig_port) > > static void icl_program_mg_dp_mode(struct intel_digital_port *intel_dig_port) > > { > > - struct drm_i915_private *dev_priv = to_i915(intel_dig_port->base.base.dev); > > + struct drm_i915_private *i915 = to_i915(intel_dig_port->base.base.dev); > > enum port port = intel_dig_port->base.port; > > - enum tc_port tc_port = intel_port_to_tc(dev_priv, port); > > + enum tc_port tc_port = intel_port_to_tc(i915, port); > > u32 ln0, ln1, lane_info; > > if (tc_port == PORT_TC_NONE || intel_dig_port->tc_type == TC_PORT_TBT) > > @@ -3069,7 +3069,7 @@ static void intel_dp_sink_set_fec_ready(struct intel_dp *intel_dp, > > static void intel_ddi_enable_fec(struct intel_encoder *encoder, > > const struct intel_crtc_state *crtc_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > enum port port = encoder->port; > > u32 val; > > @@ -3080,7 +3080,7 @@ static void intel_ddi_enable_fec(struct intel_encoder *encoder, > > val |= DP_TP_CTL_FEC_ENABLE; > > I915_WRITE(DP_TP_CTL(port), val); > > - if (intel_wait_for_register(&dev_priv->uncore, DP_TP_STATUS(port), > > + if (intel_wait_for_register(&i915->uncore, DP_TP_STATUS(port), > > DP_TP_STATUS_FEC_ENABLE_LIVE, > > DP_TP_STATUS_FEC_ENABLE_LIVE, > > 1)) > > @@ -3090,7 +3090,7 @@ static void intel_ddi_enable_fec(struct intel_encoder *encoder, > > static void intel_ddi_disable_fec_state(struct intel_encoder *encoder, > > const struct intel_crtc_state *crtc_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > enum port port = encoder->port; > > u32 val; > > @@ -3108,7 +3108,7 @@ static void intel_ddi_pre_enable_dp(struct intel_encoder *encoder, > > const struct drm_connector_state *conn_state) > > { > > struct intel_dp *intel_dp = enc_to_intel_dp(&encoder->base); > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > enum port port = encoder->port; > > struct intel_digital_port *dig_port = enc_to_dig_port(&encoder->base); > > bool is_mst = intel_crtc_has_type(crtc_state, INTEL_OUTPUT_DP_MST); > > @@ -3123,26 +3123,26 @@ static void intel_ddi_pre_enable_dp(struct intel_encoder *encoder, > > intel_ddi_clk_select(encoder, crtc_state); > > - intel_display_power_get(dev_priv, dig_port->ddi_io_power_domain); > > + intel_display_power_get(i915, dig_port->ddi_io_power_domain); > > icl_program_mg_dp_mode(dig_port); > > icl_disable_phy_clock_gating(dig_port); > > - if (INTEL_GEN(dev_priv) >= 11) > > + if (INTEL_GEN(i915) >= 11) > > icl_ddi_vswing_sequence(encoder, crtc_state->port_clock, > > level, encoder->type); > > - else if (IS_CANNONLAKE(dev_priv)) > > + else if (IS_CANNONLAKE(i915)) > > cnl_ddi_vswing_sequence(encoder, level, encoder->type); > > - else if (IS_GEN9_LP(dev_priv)) > > + else if (IS_GEN9_LP(i915)) > > bxt_ddi_vswing_sequence(encoder, level, encoder->type); > > else > > intel_prepare_dp_ddi_buffers(encoder, crtc_state); > > - if (intel_port_is_combophy(dev_priv, port)) { > > + if (intel_port_is_combophy(i915, port)) { > > bool lane_reversal = > > dig_port->saved_port_bits & DDI_BUF_PORT_REVERSAL; > > - intel_combo_phy_power_up_lanes(dev_priv, port, false, > > + intel_combo_phy_power_up_lanes(i915, port, false, > > crtc_state->lane_count, > > lane_reversal); > > } > > @@ -3154,7 +3154,7 @@ static void intel_ddi_pre_enable_dp(struct intel_encoder *encoder, > > true); > > intel_dp_sink_set_fec_ready(intel_dp, crtc_state); > > intel_dp_start_link_train(intel_dp); > > - if (port != PORT_A || INTEL_GEN(dev_priv) >= 9) > > + if (port != PORT_A || INTEL_GEN(i915) >= 9) > > intel_dp_stop_link_train(intel_dp); > > intel_ddi_enable_fec(encoder, crtc_state); > > @@ -3173,32 +3173,32 @@ static void intel_ddi_pre_enable_hdmi(struct intel_encoder *encoder, > > { > > struct intel_digital_port *intel_dig_port = enc_to_dig_port(&encoder->base); > > struct intel_hdmi *intel_hdmi = &intel_dig_port->hdmi; > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > enum port port = encoder->port; > > - int level = intel_ddi_hdmi_level(dev_priv, port); > > + int level = intel_ddi_hdmi_level(i915, port); > > struct intel_digital_port *dig_port = enc_to_dig_port(&encoder->base); > > intel_dp_dual_mode_set_tmds_output(intel_hdmi, true); > > intel_ddi_clk_select(encoder, crtc_state); > > - intel_display_power_get(dev_priv, dig_port->ddi_io_power_domain); > > + intel_display_power_get(i915, dig_port->ddi_io_power_domain); > > icl_program_mg_dp_mode(dig_port); > > icl_disable_phy_clock_gating(dig_port); > > - if (INTEL_GEN(dev_priv) >= 11) > > + if (INTEL_GEN(i915) >= 11) > > icl_ddi_vswing_sequence(encoder, crtc_state->port_clock, > > level, INTEL_OUTPUT_HDMI); > > - else if (IS_CANNONLAKE(dev_priv)) > > + else if (IS_CANNONLAKE(i915)) > > cnl_ddi_vswing_sequence(encoder, level, INTEL_OUTPUT_HDMI); > > - else if (IS_GEN9_LP(dev_priv)) > > + else if (IS_GEN9_LP(i915)) > > bxt_ddi_vswing_sequence(encoder, level, INTEL_OUTPUT_HDMI); > > else > > intel_prepare_hdmi_ddi_buffers(encoder, level); > > icl_enable_phy_clock_gating(dig_port); > > - if (IS_GEN9_BC(dev_priv)) > > + if (IS_GEN9_BC(i915)) > > skl_ddi_set_iboost(encoder, level, INTEL_OUTPUT_HDMI); > > intel_ddi_enable_pipe_clock(crtc_state); > > @@ -3213,7 +3213,7 @@ static void intel_ddi_pre_enable(struct intel_encoder *encoder, > > const struct drm_connector_state *conn_state) > > { > > struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc); > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > enum pipe pipe = crtc->pipe; > > /* > > @@ -3231,10 +3231,10 @@ static void intel_ddi_pre_enable(struct intel_encoder *encoder, > > WARN_ON(crtc_state->has_pch_encoder); > > - if (INTEL_GEN(dev_priv) >= 11) > > + if (INTEL_GEN(i915) >= 11) > > icl_map_plls_to_ports(encoder, crtc_state); > > - intel_set_cpu_fifo_underrun_reporting(dev_priv, pipe, true); > > + intel_set_cpu_fifo_underrun_reporting(i915, pipe, true); > > if (intel_crtc_has_type(crtc_state, INTEL_OUTPUT_HDMI)) { > > intel_ddi_pre_enable_hdmi(encoder, crtc_state, conn_state); > > @@ -3257,7 +3257,7 @@ static void intel_ddi_pre_enable(struct intel_encoder *encoder, > > static void intel_disable_ddi_buf(struct intel_encoder *encoder, > > const struct intel_crtc_state *crtc_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > enum port port = encoder->port; > > bool wait = false; > > u32 val; > > @@ -3278,14 +3278,14 @@ static void intel_disable_ddi_buf(struct intel_encoder *encoder, > > intel_ddi_disable_fec_state(encoder, crtc_state); > > if (wait) > > - intel_wait_ddi_buf_idle(dev_priv, port); > > + intel_wait_ddi_buf_idle(i915, port); > > } > > static void intel_ddi_post_disable_dp(struct intel_encoder *encoder, > > const struct intel_crtc_state *old_crtc_state, > > const struct drm_connector_state *old_conn_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_digital_port *dig_port = enc_to_dig_port(&encoder->base); > > struct intel_dp *intel_dp = &dig_port->dp; > > bool is_mst = intel_crtc_has_type(old_crtc_state, > > @@ -3305,7 +3305,7 @@ static void intel_ddi_post_disable_dp(struct intel_encoder *encoder, > > intel_edp_panel_vdd_on(intel_dp); > > intel_edp_panel_off(intel_dp); > > - intel_display_power_put_unchecked(dev_priv, > > + intel_display_power_put_unchecked(i915, > > dig_port->ddi_io_power_domain); > > intel_ddi_clk_disable(encoder); > > @@ -3315,7 +3315,7 @@ static void intel_ddi_post_disable_hdmi(struct intel_encoder *encoder, > > const struct intel_crtc_state *old_crtc_state, > > const struct drm_connector_state *old_conn_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_digital_port *dig_port = enc_to_dig_port(&encoder->base); > > struct intel_hdmi *intel_hdmi = &dig_port->hdmi; > > @@ -3326,7 +3326,7 @@ static void intel_ddi_post_disable_hdmi(struct intel_encoder *encoder, > > intel_disable_ddi_buf(encoder, old_crtc_state); > > - intel_display_power_put_unchecked(dev_priv, > > + intel_display_power_put_unchecked(i915, > > dig_port->ddi_io_power_domain); > > intel_ddi_clk_disable(encoder); > > @@ -3338,7 +3338,7 @@ static void intel_ddi_post_disable(struct intel_encoder *encoder, > > const struct intel_crtc_state *old_crtc_state, > > const struct drm_connector_state *old_conn_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > /* > > * When called from DP MST code: > > @@ -3360,7 +3360,7 @@ static void intel_ddi_post_disable(struct intel_encoder *encoder, > > intel_ddi_post_disable_dp(encoder, > > old_crtc_state, old_conn_state); > > - if (INTEL_GEN(dev_priv) >= 11) > > + if (INTEL_GEN(i915) >= 11) > > icl_unmap_plls_to_ports(encoder); > > } > > @@ -3368,7 +3368,7 @@ void intel_ddi_fdi_post_disable(struct intel_encoder *encoder, > > const struct intel_crtc_state *old_crtc_state, > > const struct drm_connector_state *old_conn_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > u32 val; > > /* > > @@ -3402,11 +3402,11 @@ static void intel_enable_ddi_dp(struct intel_encoder *encoder, > > const struct intel_crtc_state *crtc_state, > > const struct drm_connector_state *conn_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_dp *intel_dp = enc_to_intel_dp(&encoder->base); > > enum port port = encoder->port; > > - if (port == PORT_A && INTEL_GEN(dev_priv) < 9) > > + if (port == PORT_A && INTEL_GEN(i915) < 9) > > intel_dp_stop_link_train(intel_dp); > > intel_edp_backlight_on(crtc_state, conn_state); > > @@ -3419,7 +3419,7 @@ static void intel_enable_ddi_dp(struct intel_encoder *encoder, > > } > > static i915_reg_t > > -gen9_chicken_trans_reg_by_port(struct drm_i915_private *dev_priv, > > +gen9_chicken_trans_reg_by_port(struct drm_i915_private *i915, > > enum port port) > > { > > static const i915_reg_t regs[] = { > > @@ -3430,7 +3430,7 @@ gen9_chicken_trans_reg_by_port(struct drm_i915_private *dev_priv, > > [PORT_E] = CHICKEN_TRANS_A, > > }; > > - WARN_ON(INTEL_GEN(dev_priv) < 9); > > + WARN_ON(INTEL_GEN(i915) < 9); > > if (WARN_ON(port < PORT_A || port > PORT_E)) > > port = PORT_A; > > @@ -3442,7 +3442,7 @@ static void intel_enable_ddi_hdmi(struct intel_encoder *encoder, > > const struct intel_crtc_state *crtc_state, > > const struct drm_connector_state *conn_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_digital_port *dig_port = enc_to_dig_port(&encoder->base); > > struct drm_connector *connector = conn_state->connector; > > enum port port = encoder->port; > > @@ -3454,14 +3454,14 @@ static void intel_enable_ddi_hdmi(struct intel_encoder *encoder, > > connector->base.id, connector->name); > > /* Display WA #1143: skl,kbl,cfl */ > > - if (IS_GEN9_BC(dev_priv)) { > > + if (IS_GEN9_BC(i915)) { > > /* > > * For some reason these chicken bits have been > > * stuffed into a transcoder register, event though > > * the bits affect a specific DDI port rather than > > * a specific transcoder. > > */ > > - i915_reg_t reg = gen9_chicken_trans_reg_by_port(dev_priv, port); > > + i915_reg_t reg = gen9_chicken_trans_reg_by_port(i915, port); > > u32 val; > > val = I915_READ(reg); > > @@ -3595,9 +3595,9 @@ static void intel_ddi_set_fia_lane_count(struct intel_encoder *encoder, > > const struct intel_crtc_state *pipe_config, > > enum port port) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_digital_port *dig_port = enc_to_dig_port(&encoder->base); > > - enum tc_port tc_port = intel_port_to_tc(dev_priv, port); > > + enum tc_port tc_port = intel_port_to_tc(i915, port); > > u32 val = I915_READ(PORT_TX_DFLEXDPMLE1); > > bool lane_reversal = dig_port->saved_port_bits & DDI_BUF_PORT_REVERSAL; > > @@ -3625,16 +3625,16 @@ intel_ddi_pre_pll_enable(struct intel_encoder *encoder, > > const struct intel_crtc_state *crtc_state, > > const struct drm_connector_state *conn_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_digital_port *dig_port = enc_to_dig_port(&encoder->base); > > enum port port = encoder->port; > > if (intel_crtc_has_dp_encoder(crtc_state) || > > - intel_port_is_tc(dev_priv, encoder->port)) > > - intel_display_power_get(dev_priv, > > + intel_port_is_tc(i915, encoder->port)) > > + intel_display_power_get(i915, > > intel_ddi_main_link_aux_domain(dig_port)); > > - if (IS_GEN9_LP(dev_priv)) > > + if (IS_GEN9_LP(i915)) > > bxt_ddi_phy_set_lane_optim_mask(encoder, > > crtc_state->lane_lat_optim_mask); > > @@ -3654,19 +3654,19 @@ intel_ddi_post_pll_disable(struct intel_encoder *encoder, > > const struct intel_crtc_state *crtc_state, > > const struct drm_connector_state *conn_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_digital_port *dig_port = enc_to_dig_port(&encoder->base); > > if (intel_crtc_has_dp_encoder(crtc_state) || > > - intel_port_is_tc(dev_priv, encoder->port)) > > - intel_display_power_put_unchecked(dev_priv, > > + intel_port_is_tc(i915, encoder->port)) > > + intel_display_power_put_unchecked(i915, > > intel_ddi_main_link_aux_domain(dig_port)); > > } > > static void intel_ddi_prepare_link_retrain(struct intel_dp *intel_dp) > > { > > struct intel_digital_port *intel_dig_port = dp_to_dig_port(intel_dp); > > - struct drm_i915_private *dev_priv = > > + struct drm_i915_private *i915 = > > to_i915(intel_dig_port->base.base.dev); > > enum port port = intel_dig_port->base.port; > > u32 val; > > @@ -3687,7 +3687,7 @@ static void intel_ddi_prepare_link_retrain(struct intel_dp *intel_dp) > > POSTING_READ(DP_TP_CTL(port)); > > if (wait) > > - intel_wait_ddi_buf_idle(dev_priv, port); > > + intel_wait_ddi_buf_idle(i915, port); > > } > > val = DP_TP_CTL_ENABLE | > > @@ -3709,32 +3709,32 @@ static void intel_ddi_prepare_link_retrain(struct intel_dp *intel_dp) > > udelay(600); > > } > > -static bool intel_ddi_is_audio_enabled(struct drm_i915_private *dev_priv, > > +static bool intel_ddi_is_audio_enabled(struct drm_i915_private *i915, > > enum transcoder cpu_transcoder) > > { > > if (cpu_transcoder == TRANSCODER_EDP) > > return false; > > - if (!intel_display_power_is_enabled(dev_priv, POWER_DOMAIN_AUDIO)) > > + if (!intel_display_power_is_enabled(i915, POWER_DOMAIN_AUDIO)) > > return false; > > return I915_READ(HSW_AUD_PIN_ELD_CP_VLD) & > > AUDIO_OUTPUT_ENABLE(cpu_transcoder); > > } > > -void intel_ddi_compute_min_voltage_level(struct drm_i915_private *dev_priv, > > +void intel_ddi_compute_min_voltage_level(struct drm_i915_private *i915, > > struct intel_crtc_state *crtc_state) > > { > > - if (INTEL_GEN(dev_priv) >= 11 && crtc_state->port_clock > 594000) > > + if (INTEL_GEN(i915) >= 11 && crtc_state->port_clock > 594000) > > crtc_state->min_voltage_level = 1; > > - else if (IS_CANNONLAKE(dev_priv) && crtc_state->port_clock > 594000) > > + else if (IS_CANNONLAKE(i915) && crtc_state->port_clock > 594000) > > crtc_state->min_voltage_level = 2; > > } > > void intel_ddi_get_config(struct intel_encoder *encoder, > > struct intel_crtc_state *pipe_config) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_crtc *intel_crtc = to_intel_crtc(pipe_config->base.crtc); > > enum transcoder cpu_transcoder = pipe_config->cpu_transcoder; > > struct intel_digital_port *intel_dig_port; > > @@ -3816,10 +3816,10 @@ void intel_ddi_get_config(struct intel_encoder *encoder, > > } > > pipe_config->has_audio = > > - intel_ddi_is_audio_enabled(dev_priv, cpu_transcoder); > > + intel_ddi_is_audio_enabled(i915, cpu_transcoder); > > - if (encoder->type == INTEL_OUTPUT_EDP && dev_priv->vbt.edp.bpp && > > - pipe_config->pipe_bpp > dev_priv->vbt.edp.bpp) { > > + if (encoder->type == INTEL_OUTPUT_EDP && i915->vbt.edp.bpp && > > + pipe_config->pipe_bpp > i915->vbt.edp.bpp) { > > /* > > * This is a big fat ugly hack. > > * > > @@ -3834,17 +3834,17 @@ void intel_ddi_get_config(struct intel_encoder *encoder, > > * load. > > */ > > DRM_DEBUG_KMS("pipe has %d bpp for eDP panel, overriding BIOS-provided max %d bpp\n", > > - pipe_config->pipe_bpp, dev_priv->vbt.edp.bpp); > > - dev_priv->vbt.edp.bpp = pipe_config->pipe_bpp; > > + pipe_config->pipe_bpp, i915->vbt.edp.bpp); > > + i915->vbt.edp.bpp = pipe_config->pipe_bpp; > > } > > intel_ddi_clock_get(encoder, pipe_config); > > - if (IS_GEN9_LP(dev_priv)) > > + if (IS_GEN9_LP(i915)) > > pipe_config->lane_lat_optim_mask = > > bxt_ddi_phy_get_lane_lat_optim_mask(encoder); > > - intel_ddi_compute_min_voltage_level(dev_priv, pipe_config); > > + intel_ddi_compute_min_voltage_level(i915, pipe_config); > > intel_hdmi_read_gcp_infoframe(encoder, pipe_config); > > @@ -3885,11 +3885,11 @@ static int intel_ddi_compute_config(struct intel_encoder *encoder, > > struct drm_connector_state *conn_state) > > { > > struct intel_crtc *crtc = to_intel_crtc(pipe_config->base.crtc); > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > enum port port = encoder->port; > > int ret; > > - if (HAS_TRANSCODER_EDP(dev_priv) && port == PORT_A) > > + if (HAS_TRANSCODER_EDP(i915) && port == PORT_A) > > pipe_config->cpu_transcoder = TRANSCODER_EDP; > > if (intel_crtc_has_type(pipe_config, INTEL_OUTPUT_HDMI)) > > @@ -3899,17 +3899,17 @@ static int intel_ddi_compute_config(struct intel_encoder *encoder, > > if (ret) > > return ret; > > - if (IS_HASWELL(dev_priv) && crtc->pipe == PIPE_A && > > + if (IS_HASWELL(i915) && crtc->pipe == PIPE_A && > > pipe_config->cpu_transcoder == TRANSCODER_EDP) > > pipe_config->pch_pfit.force_thru = > > pipe_config->pch_pfit.enabled || > > pipe_config->crc_enabled; > > - if (IS_GEN9_LP(dev_priv)) > > + if (IS_GEN9_LP(i915)) > > pipe_config->lane_lat_optim_mask = > > bxt_ddi_phy_calc_lane_lat_optim_mask(pipe_config->lane_count); > > - intel_ddi_compute_min_voltage_level(dev_priv, pipe_config); > > + intel_ddi_compute_min_voltage_level(i915, pipe_config); > > return 0; > > } > > @@ -4013,11 +4013,11 @@ static int modeset_pipe(struct drm_crtc *crtc, > > static int intel_hdmi_reset_link(struct intel_encoder *encoder, > > struct drm_modeset_acquire_ctx *ctx) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_hdmi *hdmi = enc_to_intel_hdmi(&encoder->base); > > struct intel_connector *connector = hdmi->attached_connector; > > struct i2c_adapter *adapter = > > - intel_gmbus_get_adapter(dev_priv, hdmi->ddc_bus); > > + intel_gmbus_get_adapter(i915, hdmi->ddc_bus); > > struct drm_connector_state *conn_state; > > struct intel_crtc_state *crtc_state; > > struct intel_crtc *crtc; > > @@ -4027,7 +4027,7 @@ static int intel_hdmi_reset_link(struct intel_encoder *encoder, > > if (!connector || connector->base.status != connector_status_connected) > > return 0; > > - ret = drm_modeset_lock(&dev_priv->drm.mode_config.connection_mutex, > > + ret = drm_modeset_lock(&i915->drm.mode_config.connection_mutex, > > ctx); > > if (ret) > > return ret; > > @@ -4131,7 +4131,7 @@ intel_ddi_init_hdmi_connector(struct intel_digital_port *intel_dig_port) > > static bool intel_ddi_a_force_4_lanes(struct intel_digital_port *dport) > > { > > - struct drm_i915_private *dev_priv = to_i915(dport->base.base.dev); > > + struct drm_i915_private *i915 = to_i915(dport->base.base.dev); > > if (dport->base.port != PORT_A) > > return false; > > @@ -4142,7 +4142,7 @@ static bool intel_ddi_a_force_4_lanes(struct intel_digital_port *dport) > > /* Broxton/Geminilake: Bspec says that DDI_A_4_LANES is the only > > * supported configuration > > */ > > - if (IS_GEN9_LP(dev_priv)) > > + if (IS_GEN9_LP(i915)) > > return true; > > /* Cannonlake: Most of SKUs don't support DDI_E, and the only > > @@ -4150,8 +4150,8 @@ static bool intel_ddi_a_force_4_lanes(struct intel_digital_port *dport) > > * DDI_F what makes DDI_E useless. However for this > > * case let's trust VBT info. > > */ > > - if (IS_CANNONLAKE(dev_priv) && > > - !intel_bios_is_port_present(dev_priv, PORT_E)) > > + if (IS_CANNONLAKE(i915) && > > + !intel_bios_is_port_present(i915, PORT_E)) > > return true; > > return false; > > @@ -4160,11 +4160,11 @@ static bool intel_ddi_a_force_4_lanes(struct intel_digital_port *dport) > > static int > > intel_ddi_max_lanes(struct intel_digital_port *intel_dport) > > { > > - struct drm_i915_private *dev_priv = to_i915(intel_dport->base.base.dev); > > + struct drm_i915_private *i915 = to_i915(intel_dport->base.base.dev); > > enum port port = intel_dport->base.port; > > int max_lanes = 4; > > - if (INTEL_GEN(dev_priv) >= 11) > > + if (INTEL_GEN(i915) >= 11) > > return max_lanes; > > if (port == PORT_A || port == PORT_E) { > > @@ -4189,10 +4189,10 @@ intel_ddi_max_lanes(struct intel_digital_port *intel_dport) > > return max_lanes; > > } > > -void intel_ddi_init(struct drm_i915_private *dev_priv, enum port port) > > +void intel_ddi_init(struct drm_i915_private *i915, enum port port) > > { > > struct ddi_vbt_port_info *port_info = > > - &dev_priv->vbt.ddi_port_info[port]; > > + &i915->vbt.ddi_port_info[port]; > > struct intel_digital_port *intel_dig_port; > > struct intel_encoder *intel_encoder; > > struct drm_encoder *encoder; > > @@ -4202,7 +4202,7 @@ void intel_ddi_init(struct drm_i915_private *dev_priv, enum port port) > > init_hdmi = port_info->supports_dvi || port_info->supports_hdmi; > > init_dp = port_info->supports_dp; > > - if (intel_bios_is_lspcon_present(dev_priv, port)) { > > + if (intel_bios_is_lspcon_present(i915, port)) { > > /* > > * Lspcon device needs to be driven with DP connector > > * with special detection sequence. So make sure DP > > @@ -4227,7 +4227,7 @@ void intel_ddi_init(struct drm_i915_private *dev_priv, enum port port) > > intel_encoder = &intel_dig_port->base; > > encoder = &intel_encoder->base; > > - drm_encoder_init(&dev_priv->drm, encoder, &intel_ddi_funcs, > > + drm_encoder_init(&i915->drm, encoder, &intel_ddi_funcs, > > DRM_MODE_ENCODER_TMDS, "DDI %c", port_name(port)); > > intel_encoder->hotplug = intel_ddi_hotplug; > > @@ -4248,10 +4248,10 @@ void intel_ddi_init(struct drm_i915_private *dev_priv, enum port port) > > intel_encoder->power_domain = intel_port_to_power_domain(port); > > intel_encoder->port = port; > > intel_encoder->cloneable = 0; > > - for_each_pipe(dev_priv, pipe) > > + for_each_pipe(i915, pipe) > > intel_encoder->crtc_mask |= BIT(pipe); > > - if (INTEL_GEN(dev_priv) >= 11) > > + if (INTEL_GEN(i915) >= 11) > > intel_dig_port->saved_port_bits = I915_READ(DDI_BUF_CTL(port)) & > > DDI_BUF_PORT_REVERSAL; > > else > > @@ -4259,9 +4259,9 @@ void intel_ddi_init(struct drm_i915_private *dev_priv, enum port port) > > (DDI_BUF_PORT_REVERSAL | DDI_A_4_LANES); > > intel_dig_port->dp.output_reg = INVALID_MMIO_REG; > > intel_dig_port->max_lanes = intel_ddi_max_lanes(intel_dig_port); > > - intel_dig_port->aux_ch = intel_bios_port_aux_ch(dev_priv, port); > > + intel_dig_port->aux_ch = intel_bios_port_aux_ch(i915, port); > > - intel_dig_port->tc_legacy_port = intel_port_is_tc(dev_priv, port) && > > + intel_dig_port->tc_legacy_port = intel_port_is_tc(i915, port) && > > !port_info->supports_typec_usb && > > !port_info->supports_tbt; > > @@ -4324,7 +4324,7 @@ void intel_ddi_init(struct drm_i915_private *dev_priv, enum port port) > > intel_infoframe_init(intel_dig_port); > > - if (intel_port_is_tc(dev_priv, port)) > > + if (intel_port_is_tc(i915, port)) > > intel_digital_port_connected(intel_encoder); > > return; > > diff --git a/drivers/gpu/drm/i915/intel_ddi.h b/drivers/gpu/drm/i915/intel_ddi.h > > index a08365da2643..50dd6fac4e3b 100644 > > --- a/drivers/gpu/drm/i915/intel_ddi.h > > +++ b/drivers/gpu/drm/i915/intel_ddi.h > > @@ -24,7 +24,7 @@ void intel_ddi_fdi_post_disable(struct intel_encoder *intel_encoder, > > const struct drm_connector_state *old_conn_state); > > void hsw_fdi_link_train(struct intel_crtc *crtc, > > const struct intel_crtc_state *crtc_state); > > -void intel_ddi_init(struct drm_i915_private *dev_priv, enum port port); > > +void intel_ddi_init(struct drm_i915_private *i915, enum port port); > > bool intel_ddi_get_hw_state(struct intel_encoder *encoder, enum pipe *pipe); > > void intel_ddi_enable_transcoder_func(const struct intel_crtc_state *crtc_state); > > void intel_ddi_disable_transcoder_func(const struct intel_crtc_state *crtc_state); > > @@ -36,7 +36,7 @@ void intel_ddi_get_config(struct intel_encoder *encoder, > > struct intel_crtc_state *pipe_config); > > void intel_ddi_set_vc_payload_alloc(const struct intel_crtc_state *crtc_state, > > bool state); > > -void intel_ddi_compute_min_voltage_level(struct drm_i915_private *dev_priv, > > +void intel_ddi_compute_min_voltage_level(struct drm_i915_private *i915, > > struct intel_crtc_state *crtc_state); > > u32 bxt_signal_levels(struct intel_dp *intel_dp); > > u32 ddi_signal_levels(struct intel_dp *intel_dp); > > @@ -46,7 +46,7 @@ u8 intel_ddi_dp_pre_emphasis_max(struct intel_encoder *encoder, > > int intel_ddi_toggle_hdcp_signalling(struct intel_encoder *intel_encoder, > > bool enable); > > void icl_sanitize_encoder_pll_mapping(struct intel_encoder *encoder); > > -int cnl_calc_wrpll_link(struct drm_i915_private *dev_priv, > > +int cnl_calc_wrpll_link(struct drm_i915_private *i915, > > struct intel_dpll_hw_state *state); > > #endif /* __INTEL_DDI_H__ */ > > diff --git a/drivers/gpu/drm/i915/intel_device_info.c b/drivers/gpu/drm/i915/intel_device_info.c > > index 7135d8dc32a7..a670881b3f1e 100644 > > --- a/drivers/gpu/drm/i915/intel_device_info.c > > +++ b/drivers/gpu/drm/i915/intel_device_info.c > > @@ -182,15 +182,15 @@ static u16 compute_eu_total(const struct sseu_dev_info *sseu) > > return total; > > } > > -static void gen11_sseu_info_init(struct drm_i915_private *dev_priv) > > +static void gen11_sseu_info_init(struct drm_i915_private *i915) > > { > > - struct sseu_dev_info *sseu = &RUNTIME_INFO(dev_priv)->sseu; > > + struct sseu_dev_info *sseu = &RUNTIME_INFO(i915)->sseu; > > u8 s_en; > > u32 ss_en, ss_en_mask; > > u8 eu_en; > > int s; > > - if (IS_ELKHARTLAKE(dev_priv)) { > > + if (IS_ELKHARTLAKE(i915)) { > > sseu->max_slices = 1; > > sseu->max_subslices = 4; > > sseu->max_eus_per_subslice = 8; > > @@ -227,9 +227,9 @@ static void gen11_sseu_info_init(struct drm_i915_private *dev_priv) > > sseu->has_eu_pg = 1; > > } > > -static void gen10_sseu_info_init(struct drm_i915_private *dev_priv) > > +static void gen10_sseu_info_init(struct drm_i915_private *i915) > > { > > - struct sseu_dev_info *sseu = &RUNTIME_INFO(dev_priv)->sseu; > > + struct sseu_dev_info *sseu = &RUNTIME_INFO(i915)->sseu; > > const u32 fuse2 = I915_READ(GEN8_FUSE2); > > int s, ss; > > const int eu_mask = 0xff; > > @@ -305,9 +305,9 @@ static void gen10_sseu_info_init(struct drm_i915_private *dev_priv) > > sseu->has_eu_pg = 1; > > } > > -static void cherryview_sseu_info_init(struct drm_i915_private *dev_priv) > > +static void cherryview_sseu_info_init(struct drm_i915_private *i915) > > { > > - struct sseu_dev_info *sseu = &RUNTIME_INFO(dev_priv)->sseu; > > + struct sseu_dev_info *sseu = &RUNTIME_INFO(i915)->sseu; > > u32 fuse; > > fuse = I915_READ(CHV_FUSE_GT); > > @@ -359,10 +359,10 @@ static void cherryview_sseu_info_init(struct drm_i915_private *dev_priv) > > sseu->has_eu_pg = (sseu->eu_per_subslice > 2); > > } > > -static void gen9_sseu_info_init(struct drm_i915_private *dev_priv) > > +static void gen9_sseu_info_init(struct drm_i915_private *i915) > > { > > - struct intel_device_info *info = mkwrite_device_info(dev_priv); > > - struct sseu_dev_info *sseu = &RUNTIME_INFO(dev_priv)->sseu; > > + struct intel_device_info *info = mkwrite_device_info(i915); > > + struct sseu_dev_info *sseu = &RUNTIME_INFO(i915)->sseu; > > int s, ss; > > u32 fuse2, eu_disable, subslice_mask; > > const u8 eu_mask = 0xff; > > @@ -371,8 +371,8 @@ static void gen9_sseu_info_init(struct drm_i915_private *dev_priv) > > sseu->slice_mask = (fuse2 & GEN8_F2_S_ENA_MASK) >> GEN8_F2_S_ENA_SHIFT; > > /* BXT has a single slice and at most 3 subslices. */ > > - sseu->max_slices = IS_GEN9_LP(dev_priv) ? 1 : 3; > > - sseu->max_subslices = IS_GEN9_LP(dev_priv) ? 3 : 4; > > + sseu->max_slices = IS_GEN9_LP(i915) ? 1 : 3; > > + sseu->max_subslices = IS_GEN9_LP(i915) ? 3 : 4; > > sseu->max_eus_per_subslice = 8; > > /* > > @@ -442,12 +442,12 @@ static void gen9_sseu_info_init(struct drm_i915_private *dev_priv) > > * pair per subslice. > > */ > > sseu->has_slice_pg = > > - !IS_GEN9_LP(dev_priv) && hweight8(sseu->slice_mask) > 1; > > + !IS_GEN9_LP(i915) && hweight8(sseu->slice_mask) > 1; > > sseu->has_subslice_pg = > > - IS_GEN9_LP(dev_priv) && intel_sseu_subslice_total(sseu) > 1; > > + IS_GEN9_LP(i915) && intel_sseu_subslice_total(sseu) > 1; > > sseu->has_eu_pg = sseu->eu_per_subslice > 2; > > - if (IS_GEN9_LP(dev_priv)) { > > + if (IS_GEN9_LP(i915)) { > > #define IS_SS_DISABLED(ss) (!(sseu->subslice_mask[0] & BIT(ss))) > > info->has_pooled_eu = hweight8(sseu->subslice_mask[0]) == 3; > > @@ -464,9 +464,9 @@ static void gen9_sseu_info_init(struct drm_i915_private *dev_priv) > > } > > } > > -static void broadwell_sseu_info_init(struct drm_i915_private *dev_priv) > > +static void broadwell_sseu_info_init(struct drm_i915_private *i915) > > { > > - struct sseu_dev_info *sseu = &RUNTIME_INFO(dev_priv)->sseu; > > + struct sseu_dev_info *sseu = &RUNTIME_INFO(i915)->sseu; > > int s, ss; > > u32 fuse2, subslice_mask, eu_disable[3]; /* s_max */ > > @@ -547,9 +547,9 @@ static void broadwell_sseu_info_init(struct drm_i915_private *dev_priv) > > sseu->has_eu_pg = 0; > > } > > -static void haswell_sseu_info_init(struct drm_i915_private *dev_priv) > > +static void haswell_sseu_info_init(struct drm_i915_private *i915) > > { > > - struct sseu_dev_info *sseu = &RUNTIME_INFO(dev_priv)->sseu; > > + struct sseu_dev_info *sseu = &RUNTIME_INFO(i915)->sseu; > > u32 fuse1; > > int s, ss; > > @@ -557,9 +557,9 @@ static void haswell_sseu_info_init(struct drm_i915_private *dev_priv) > > * There isn't a register to tell us how many slices/subslices. We > > * work off the PCI-ids here. > > */ > > - switch (INTEL_INFO(dev_priv)->gt) { > > + switch (INTEL_INFO(i915)->gt) { > > default: > > - MISSING_CASE(INTEL_INFO(dev_priv)->gt); > > + MISSING_CASE(INTEL_INFO(i915)->gt); > > /* fall through */ > > case 1: > > sseu->slice_mask = BIT(0); > > @@ -612,7 +612,7 @@ static void haswell_sseu_info_init(struct drm_i915_private *dev_priv) > > sseu->has_eu_pg = 0; > > } > > -static u32 read_reference_ts_freq(struct drm_i915_private *dev_priv) > > +static u32 read_reference_ts_freq(struct drm_i915_private *i915) > > { > > u32 ts_override = I915_READ(GEN9_TIMESTAMP_OVERRIDE); > > u32 base_freq, frac_freq; > > @@ -629,7 +629,7 @@ static u32 read_reference_ts_freq(struct drm_i915_private *dev_priv) > > return base_freq + frac_freq; > > } > > -static u32 gen10_get_crystal_clock_freq(struct drm_i915_private *dev_priv, > > +static u32 gen10_get_crystal_clock_freq(struct drm_i915_private *i915, > > u32 rpm_config_reg) > > { > > u32 f19_2_mhz = 19200; > > @@ -649,7 +649,7 @@ static u32 gen10_get_crystal_clock_freq(struct drm_i915_private *dev_priv, > > } > > } > > -static u32 gen11_get_crystal_clock_freq(struct drm_i915_private *dev_priv, > > +static u32 gen11_get_crystal_clock_freq(struct drm_i915_private *i915, > > u32 rpm_config_reg) > > { > > u32 f19_2_mhz = 19200; > > @@ -675,21 +675,21 @@ static u32 gen11_get_crystal_clock_freq(struct drm_i915_private *dev_priv, > > } > > } > > -static u32 read_timestamp_frequency(struct drm_i915_private *dev_priv) > > +static u32 read_timestamp_frequency(struct drm_i915_private *i915) > > { > > u32 f12_5_mhz = 12500; > > u32 f19_2_mhz = 19200; > > u32 f24_mhz = 24000; > > - if (INTEL_GEN(dev_priv) <= 4) { > > + if (INTEL_GEN(i915) <= 4) { > > /* PRMs say: > > * > > * "The value in this register increments once every 16 > > * hclks." (through the “Clocking Configuration” > > * (“CLKCFG”) MCHBAR register) > > */ > > - return dev_priv->rawclk_freq / 16; > > - } else if (INTEL_GEN(dev_priv) <= 8) { > > + return i915->rawclk_freq / 16; > > + } else if (INTEL_GEN(i915) <= 8) { > > /* PRMs say: > > * > > * "The PCU TSC counts 10ns increments; this timestamp > > @@ -697,14 +697,14 @@ static u32 read_timestamp_frequency(struct drm_i915_private *dev_priv) > > * rolling over every 1.5 hours). > > */ > > return f12_5_mhz; > > - } else if (INTEL_GEN(dev_priv) <= 9) { > > + } else if (INTEL_GEN(i915) <= 9) { > > u32 ctc_reg = I915_READ(CTC_MODE); > > u32 freq = 0; > > if ((ctc_reg & CTC_SOURCE_PARAMETER_MASK) == CTC_SOURCE_DIVIDE_LOGIC) { > > - freq = read_reference_ts_freq(dev_priv); > > + freq = read_reference_ts_freq(i915); > > } else { > > - freq = IS_GEN9_LP(dev_priv) ? f19_2_mhz : f24_mhz; > > + freq = IS_GEN9_LP(i915) ? f19_2_mhz : f24_mhz; > > /* Now figure out how the command stream's timestamp > > * register increments from this frequency (it might > > @@ -715,7 +715,7 @@ static u32 read_timestamp_frequency(struct drm_i915_private *dev_priv) > > } > > return freq; > > - } else if (INTEL_GEN(dev_priv) <= 11) { > > + } else if (INTEL_GEN(i915) <= 11) { > > u32 ctc_reg = I915_READ(CTC_MODE); > > u32 freq = 0; > > @@ -725,15 +725,15 @@ static u32 read_timestamp_frequency(struct drm_i915_private *dev_priv) > > * tells us which one we should use. > > */ > > if ((ctc_reg & CTC_SOURCE_PARAMETER_MASK) == CTC_SOURCE_DIVIDE_LOGIC) { > > - freq = read_reference_ts_freq(dev_priv); > > + freq = read_reference_ts_freq(i915); > > } else { > > u32 rpm_config_reg = I915_READ(RPM_CONFIG0); > > - if (INTEL_GEN(dev_priv) <= 10) > > - freq = gen10_get_crystal_clock_freq(dev_priv, > > + if (INTEL_GEN(i915) <= 10) > > + freq = gen10_get_crystal_clock_freq(i915, > > rpm_config_reg); > > else > > - freq = gen11_get_crystal_clock_freq(dev_priv, > > + freq = gen11_get_crystal_clock_freq(i915, > > rpm_config_reg); > > /* Now figure out how the command stream's timestamp > > @@ -841,7 +841,7 @@ void intel_device_info_subplatform_init(struct drm_i915_private *i915) > > /** > > * intel_device_info_runtime_init - initialize runtime info > > - * @dev_priv: the i915 device > > + * @i915: the i915 device > > * > > * Determine various intel_device_info fields at runtime. > > * > > @@ -855,16 +855,16 @@ void intel_device_info_subplatform_init(struct drm_i915_private *i915) > > * - after the PCH has been detected, > > * - before the first usage of the fields it can tweak. > > */ > > -void intel_device_info_runtime_init(struct drm_i915_private *dev_priv) > > +void intel_device_info_runtime_init(struct drm_i915_private *i915) > > { > > - struct intel_device_info *info = mkwrite_device_info(dev_priv); > > - struct intel_runtime_info *runtime = RUNTIME_INFO(dev_priv); > > + struct intel_device_info *info = mkwrite_device_info(i915); > > + struct intel_runtime_info *runtime = RUNTIME_INFO(i915); > > enum pipe pipe; > > - if (INTEL_GEN(dev_priv) >= 10) { > > - for_each_pipe(dev_priv, pipe) > > + if (INTEL_GEN(i915) >= 10) { > > + for_each_pipe(i915, pipe) > > runtime->num_scalers[pipe] = 2; > > - } else if (IS_GEN(dev_priv, 9)) { > > + } else if (IS_GEN(i915, 9)) { > > runtime->num_scalers[PIPE_A] = 2; > > runtime->num_scalers[PIPE_B] = 2; > > runtime->num_scalers[PIPE_C] = 1; > > @@ -872,13 +872,13 @@ void intel_device_info_runtime_init(struct drm_i915_private *dev_priv) > > BUILD_BUG_ON(BITS_PER_TYPE(intel_engine_mask_t) < I915_NUM_ENGINES); > > - if (INTEL_GEN(dev_priv) >= 11) > > - for_each_pipe(dev_priv, pipe) > > + if (INTEL_GEN(i915) >= 11) > > + for_each_pipe(i915, pipe) > > runtime->num_sprites[pipe] = 6; > > - else if (IS_GEN(dev_priv, 10) || IS_GEMINILAKE(dev_priv)) > > - for_each_pipe(dev_priv, pipe) > > + else if (IS_GEN(i915, 10) || IS_GEMINILAKE(i915)) > > + for_each_pipe(i915, pipe) > > runtime->num_sprites[pipe] = 3; > > - else if (IS_BROXTON(dev_priv)) { > > + else if (IS_BROXTON(i915)) { > > /* > > * Skylake and Broxton currently don't expose the topmost plane as its > > * use is exclusive with the legacy cursor and we only want to expose > > @@ -891,20 +891,20 @@ void intel_device_info_runtime_init(struct drm_i915_private *dev_priv) > > runtime->num_sprites[PIPE_A] = 2; > > runtime->num_sprites[PIPE_B] = 2; > > runtime->num_sprites[PIPE_C] = 1; > > - } else if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) { > > - for_each_pipe(dev_priv, pipe) > > + } else if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) { > > + for_each_pipe(i915, pipe) > > runtime->num_sprites[pipe] = 2; > > - } else if (INTEL_GEN(dev_priv) >= 5 || IS_G4X(dev_priv)) { > > - for_each_pipe(dev_priv, pipe) > > + } else if (INTEL_GEN(i915) >= 5 || IS_G4X(i915)) { > > + for_each_pipe(i915, pipe) > > runtime->num_sprites[pipe] = 1; > > } > > if (i915_modparams.disable_display) { > > DRM_INFO("Display disabled (module parameter)\n"); > > info->num_pipes = 0; > > - } else if (HAS_DISPLAY(dev_priv) && > > - (IS_GEN_RANGE(dev_priv, 7, 8)) && > > - HAS_PCH_SPLIT(dev_priv)) { > > + } else if (HAS_DISPLAY(i915) && > > + (IS_GEN_RANGE(i915, 7, 8)) && > > + HAS_PCH_SPLIT(i915)) { > > u32 fuse_strap = I915_READ(FUSE_STRAP); > > u32 sfuse_strap = I915_READ(SFUSE_STRAP); > > @@ -919,7 +919,7 @@ void intel_device_info_runtime_init(struct drm_i915_private *dev_priv) > > */ > > if (fuse_strap & ILK_INTERNAL_DISPLAY_DISABLE || > > sfuse_strap & SFUSE_STRAP_DISPLAY_DISABLED || > > - (HAS_PCH_CPT(dev_priv) && > > + (HAS_PCH_CPT(i915) && > > !(sfuse_strap & SFUSE_STRAP_FUSE_LOCK))) { > > DRM_INFO("Display fused off, disabling\n"); > > info->num_pipes = 0; > > @@ -927,7 +927,7 @@ void intel_device_info_runtime_init(struct drm_i915_private *dev_priv) > > DRM_INFO("PipeC fused off\n"); > > info->num_pipes -= 1; > > } > > - } else if (HAS_DISPLAY(dev_priv) && INTEL_GEN(dev_priv) >= 9) { > > + } else if (HAS_DISPLAY(i915) && INTEL_GEN(i915) >= 9) { > > u32 dfsm = I915_READ(SKL_DFSM); > > u8 disabled_mask = 0; > > bool invalid; > > @@ -961,26 +961,26 @@ void intel_device_info_runtime_init(struct drm_i915_private *dev_priv) > > } > > /* Initialize slice/subslice/EU info */ > > - if (IS_HASWELL(dev_priv)) > > - haswell_sseu_info_init(dev_priv); > > - else if (IS_CHERRYVIEW(dev_priv)) > > - cherryview_sseu_info_init(dev_priv); > > - else if (IS_BROADWELL(dev_priv)) > > - broadwell_sseu_info_init(dev_priv); > > - else if (IS_GEN(dev_priv, 9)) > > - gen9_sseu_info_init(dev_priv); > > - else if (IS_GEN(dev_priv, 10)) > > - gen10_sseu_info_init(dev_priv); > > - else if (INTEL_GEN(dev_priv) >= 11) > > - gen11_sseu_info_init(dev_priv); > > - > > - if (IS_GEN(dev_priv, 6) && intel_vtd_active()) { > > + if (IS_HASWELL(i915)) > > + haswell_sseu_info_init(i915); > > + else if (IS_CHERRYVIEW(i915)) > > + cherryview_sseu_info_init(i915); > > + else if (IS_BROADWELL(i915)) > > + broadwell_sseu_info_init(i915); > > + else if (IS_GEN(i915, 9)) > > + gen9_sseu_info_init(i915); > > + else if (IS_GEN(i915, 10)) > > + gen10_sseu_info_init(i915); > > + else if (INTEL_GEN(i915) >= 11) > > + gen11_sseu_info_init(i915); > > + > > + if (IS_GEN(i915, 6) && intel_vtd_active()) { > > DRM_INFO("Disabling ppGTT for VT-d support\n"); > > info->ppgtt_type = INTEL_PPGTT_NONE; > > } > > /* Initialize command stream timestamp frequency */ > > - runtime->cs_timestamp_frequency_khz = read_timestamp_frequency(dev_priv); > > + runtime->cs_timestamp_frequency_khz = read_timestamp_frequency(i915); > > } > > void intel_driver_caps_print(const struct intel_driver_caps *caps, > > @@ -997,16 +997,16 @@ void intel_driver_caps_print(const struct intel_driver_caps *caps, > > * this point (but later we need to prune the forcewake domains for engines that > > * are indeed fused off). > > */ > > -void intel_device_info_init_mmio(struct drm_i915_private *dev_priv) > > +void intel_device_info_init_mmio(struct drm_i915_private *i915) > > { > > - struct intel_device_info *info = mkwrite_device_info(dev_priv); > > + struct intel_device_info *info = mkwrite_device_info(i915); > > unsigned int logical_vdbox = 0; > > unsigned int i; > > u32 media_fuse; > > u16 vdbox_mask; > > u16 vebox_mask; > > - if (INTEL_GEN(dev_priv) < 11) > > + if (INTEL_GEN(i915) < 11) > > return; > > media_fuse = ~I915_READ(GEN11_GT_VEBOX_VDBOX_DISABLE); > > @@ -1016,7 +1016,7 @@ void intel_device_info_init_mmio(struct drm_i915_private *dev_priv) > > GEN11_GT_VEBOX_DISABLE_SHIFT; > > for (i = 0; i < I915_MAX_VCS; i++) { > > - if (!HAS_ENGINE(dev_priv, _VCS(i))) > > + if (!HAS_ENGINE(i915, _VCS(i))) > > continue; > > if (!(BIT(i) & vdbox_mask)) { > > @@ -1030,14 +1030,14 @@ void intel_device_info_init_mmio(struct drm_i915_private *dev_priv) > > * hooked up to an SFC (Scaler & Format Converter) unit. > > */ > > if (logical_vdbox++ % 2 == 0) > > - RUNTIME_INFO(dev_priv)->vdbox_sfc_access |= BIT(i); > > + RUNTIME_INFO(i915)->vdbox_sfc_access |= BIT(i); > > } > > DRM_DEBUG_DRIVER("vdbox enable: %04x, instances: %04lx\n", > > - vdbox_mask, VDBOX_MASK(dev_priv)); > > - GEM_BUG_ON(vdbox_mask != VDBOX_MASK(dev_priv)); > > + vdbox_mask, VDBOX_MASK(i915)); > > + GEM_BUG_ON(vdbox_mask != VDBOX_MASK(i915)); > > for (i = 0; i < I915_MAX_VECS; i++) { > > - if (!HAS_ENGINE(dev_priv, _VECS(i))) > > + if (!HAS_ENGINE(i915, _VECS(i))) > > continue; > > if (!(BIT(i) & vebox_mask)) { > > @@ -1046,6 +1046,6 @@ void intel_device_info_init_mmio(struct drm_i915_private *dev_priv) > > } > > } > > DRM_DEBUG_DRIVER("vebox enable: %04x, instances: %04lx\n", > > - vebox_mask, VEBOX_MASK(dev_priv)); > > - GEM_BUG_ON(vebox_mask != VEBOX_MASK(dev_priv)); > > + vebox_mask, VEBOX_MASK(i915)); > > + GEM_BUG_ON(vebox_mask != VEBOX_MASK(i915)); > > } > > diff --git a/drivers/gpu/drm/i915/intel_device_info.h b/drivers/gpu/drm/i915/intel_device_info.h > > index 3ea953a230b3..93e5cc95e5c0 100644 > > --- a/drivers/gpu/drm/i915/intel_device_info.h > > +++ b/drivers/gpu/drm/i915/intel_device_info.h > > @@ -218,8 +218,8 @@ struct intel_driver_caps { > > const char *intel_platform_name(enum intel_platform platform); > > -void intel_device_info_subplatform_init(struct drm_i915_private *dev_priv); > > -void intel_device_info_runtime_init(struct drm_i915_private *dev_priv); > > +void intel_device_info_subplatform_init(struct drm_i915_private *i915); > > +void intel_device_info_runtime_init(struct drm_i915_private *i915); > > void intel_device_info_dump_flags(const struct intel_device_info *info, > > struct drm_printer *p); > > void intel_device_info_dump_runtime(const struct intel_runtime_info *info, > > @@ -227,7 +227,7 @@ void intel_device_info_dump_runtime(const struct intel_runtime_info *info, > > void intel_device_info_dump_topology(const struct sseu_dev_info *sseu, > > struct drm_printer *p); > > -void intel_device_info_init_mmio(struct drm_i915_private *dev_priv); > > +void intel_device_info_init_mmio(struct drm_i915_private *i915); > > void intel_driver_caps_print(const struct intel_driver_caps *caps, > > struct drm_printer *p); > > diff --git a/drivers/gpu/drm/i915/intel_display.c b/drivers/gpu/drm/i915/intel_display.c > > index 1b1ddb48ca7a..be26ad9fc194 100644 > > --- a/drivers/gpu/drm/i915/intel_display.c > > +++ b/drivers/gpu/drm/i915/intel_display.c > > @@ -156,24 +156,24 @@ struct intel_limit { > > }; > > /* returns HPLL frequency in kHz */ > > -int vlv_get_hpll_vco(struct drm_i915_private *dev_priv) > > +int vlv_get_hpll_vco(struct drm_i915_private *i915) > > { > > int hpll_freq, vco_freq[] = { 800, 1600, 2000, 2400 }; > > /* Obtain SKU information */ > > - hpll_freq = vlv_cck_read(dev_priv, CCK_FUSE_REG) & > > + hpll_freq = vlv_cck_read(i915, CCK_FUSE_REG) & > > CCK_FUSE_HPLL_FREQ_MASK; > > return vco_freq[hpll_freq] * 1000; > > } > > -int vlv_get_cck_clock(struct drm_i915_private *dev_priv, > > +int vlv_get_cck_clock(struct drm_i915_private *i915, > > const char *name, u32 reg, int ref_freq) > > { > > u32 val; > > int divider; > > - val = vlv_cck_read(dev_priv, reg); > > + val = vlv_cck_read(i915, reg); > > divider = val & CCK_FREQUENCY_VALUES; > > WARN((val & CCK_FREQUENCY_STATUS) != > > @@ -183,42 +183,42 @@ int vlv_get_cck_clock(struct drm_i915_private *dev_priv, > > return DIV_ROUND_CLOSEST(ref_freq << 1, divider + 1); > > } > > -int vlv_get_cck_clock_hpll(struct drm_i915_private *dev_priv, > > +int vlv_get_cck_clock_hpll(struct drm_i915_private *i915, > > const char *name, u32 reg) > > { > > int hpll; > > - vlv_cck_get(dev_priv); > > + vlv_cck_get(i915); > > - if (dev_priv->hpll_freq == 0) > > - dev_priv->hpll_freq = vlv_get_hpll_vco(dev_priv); > > + if (i915->hpll_freq == 0) > > + i915->hpll_freq = vlv_get_hpll_vco(i915); > > - hpll = vlv_get_cck_clock(dev_priv, name, reg, dev_priv->hpll_freq); > > + hpll = vlv_get_cck_clock(i915, name, reg, i915->hpll_freq); > > - vlv_cck_put(dev_priv); > > + vlv_cck_put(i915); > > return hpll; > > } > > -static void intel_update_czclk(struct drm_i915_private *dev_priv) > > +static void intel_update_czclk(struct drm_i915_private *i915) > > { > > - if (!(IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv))) > > + if (!(IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915))) > > return; > > - dev_priv->czclk_freq = vlv_get_cck_clock_hpll(dev_priv, "czclk", > > + i915->czclk_freq = vlv_get_cck_clock_hpll(i915, "czclk", > > CCK_CZ_CLOCK_CONTROL); > > - DRM_DEBUG_DRIVER("CZ clock rate: %d kHz\n", dev_priv->czclk_freq); > > + DRM_DEBUG_DRIVER("CZ clock rate: %d kHz\n", i915->czclk_freq); > > } > > static inline u32 /* units of 100MHz */ > > -intel_fdi_link_freq(struct drm_i915_private *dev_priv, > > +intel_fdi_link_freq(struct drm_i915_private *i915, > > const struct intel_crtc_state *pipe_config) > > { > > - if (HAS_DDI(dev_priv)) > > + if (HAS_DDI(i915)) > > return pipe_config->port_clock; /* SPLL */ > > else > > - return dev_priv->fdi_pll_freq; > > + return i915->fdi_pll_freq; > > } > > static const struct intel_limit intel_limits_i8xx_dac = { > > @@ -488,7 +488,7 @@ static const struct intel_limit intel_limits_bxt = { > > /* WA Display #0827: Gen9:all */ > > static void > > -skl_wa_827(struct drm_i915_private *dev_priv, int pipe, bool enable) > > +skl_wa_827(struct drm_i915_private *i915, int pipe, bool enable) > > { > > if (enable) > > I915_WRITE(CLKGATE_DIS_PSL(pipe), > > @@ -502,7 +502,7 @@ skl_wa_827(struct drm_i915_private *dev_priv, int pipe, bool enable) > > /* Wa_2006604312:icl */ > > static void > > -icl_wa_scalerclkgating(struct drm_i915_private *dev_priv, enum pipe pipe, > > +icl_wa_scalerclkgating(struct drm_i915_private *i915, enum pipe pipe, > > bool enable) > > { > > if (enable) > > @@ -588,7 +588,7 @@ int chv_calc_dpll_params(int refclk, struct dpll *clock) > > * Returns whether the given set of divisors are valid for a given refclk with > > * the given connectors. > > */ > > -static bool intel_PLL_is_valid(struct drm_i915_private *dev_priv, > > +static bool intel_PLL_is_valid(struct drm_i915_private *i915, > > const struct intel_limit *limit, > > const struct dpll *clock) > > { > > @@ -601,13 +601,13 @@ static bool intel_PLL_is_valid(struct drm_i915_private *dev_priv, > > if (clock->m1 < limit->m1.min || limit->m1.max < clock->m1) > > INTELPllInvalid("m1 out of range\n"); > > - if (!IS_PINEVIEW(dev_priv) && !IS_VALLEYVIEW(dev_priv) && > > - !IS_CHERRYVIEW(dev_priv) && !IS_GEN9_LP(dev_priv)) > > + if (!IS_PINEVIEW(i915) && !IS_VALLEYVIEW(i915) && > > + !IS_CHERRYVIEW(i915) && !IS_GEN9_LP(i915)) > > if (clock->m1 <= clock->m2) > > INTELPllInvalid("m1 <= m2\n"); > > - if (!IS_VALLEYVIEW(dev_priv) && !IS_CHERRYVIEW(dev_priv) && > > - !IS_GEN9_LP(dev_priv)) { > > + if (!IS_VALLEYVIEW(i915) && !IS_CHERRYVIEW(i915) && > > + !IS_GEN9_LP(i915)) { > > if (clock->p < limit->p.min || limit->p.max < clock->p) > > INTELPllInvalid("p out of range\n"); > > if (clock->m < limit->m.min || limit->m.max < clock->m) > > @@ -630,7 +630,7 @@ i9xx_select_p2_div(const struct intel_limit *limit, > > const struct intel_crtc_state *crtc_state, > > int target) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc_state->base.crtc->dev); > > + struct drm_i915_private *i915 = to_i915(crtc_state->base.crtc->dev); > > if (intel_crtc_has_type(crtc_state, INTEL_OUTPUT_LVDS)) { > > /* > > @@ -638,7 +638,7 @@ i9xx_select_p2_div(const struct intel_limit *limit, > > * We haven't figured out how to reliably set up different > > * single/dual channel state, if we even can. > > */ > > - if (intel_is_dual_link_lvds(dev_priv)) > > + if (intel_is_dual_link_lvds(i915)) > > return limit->p2.p2_fast; > > else > > return limit->p2.p2_slow; > > @@ -1016,22 +1016,22 @@ bool intel_crtc_active(struct intel_crtc *crtc) > > crtc->config->base.adjusted_mode.crtc_clock; > > } > > -enum transcoder intel_pipe_to_cpu_transcoder(struct drm_i915_private *dev_priv, > > +enum transcoder intel_pipe_to_cpu_transcoder(struct drm_i915_private *i915, > > enum pipe pipe) > > { > > - struct intel_crtc *crtc = intel_get_crtc_for_pipe(dev_priv, pipe); > > + struct intel_crtc *crtc = intel_get_crtc_for_pipe(i915, pipe); > > return crtc->config->cpu_transcoder; > > } > > -static bool pipe_scanline_is_moving(struct drm_i915_private *dev_priv, > > +static bool pipe_scanline_is_moving(struct drm_i915_private *i915, > > enum pipe pipe) > > { > > i915_reg_t reg = PIPEDSL(pipe); > > u32 line1, line2; > > u32 line_mask; > > - if (IS_GEN(dev_priv, 2)) > > + if (IS_GEN(i915, 2)) > > line_mask = DSL_LINEMASK_GEN2; > > else > > line_mask = DSL_LINEMASK_GEN3; > > @@ -1045,11 +1045,11 @@ static bool pipe_scanline_is_moving(struct drm_i915_private *dev_priv, > > static void wait_for_pipe_scanline_moving(struct intel_crtc *crtc, bool state) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > enum pipe pipe = crtc->pipe; > > /* Wait for the display line to settle/start moving */ > > - if (wait_for(pipe_scanline_is_moving(dev_priv, pipe) == state, 100)) > > + if (wait_for(pipe_scanline_is_moving(i915, pipe) == state, 100)) > > DRM_ERROR("pipe %c scanline %s wait timed out\n", > > pipe_name(pipe), onoff(state)); > > } > > @@ -1068,14 +1068,14 @@ static void > > intel_wait_for_pipe_off(const struct intel_crtc_state *old_crtc_state) > > { > > struct intel_crtc *crtc = to_intel_crtc(old_crtc_state->base.crtc); > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > - if (INTEL_GEN(dev_priv) >= 4) { > > + if (INTEL_GEN(i915) >= 4) { > > enum transcoder cpu_transcoder = old_crtc_state->cpu_transcoder; > > i915_reg_t reg = PIPECONF(cpu_transcoder); > > /* Wait for the Pipe State to go off */ > > - if (intel_wait_for_register(&dev_priv->uncore, > > + if (intel_wait_for_register(&i915->uncore, > > reg, I965_PIPECONF_ACTIVE, 0, > > 100)) > > WARN(1, "pipe_off wait timed out\n"); > > @@ -1085,7 +1085,7 @@ intel_wait_for_pipe_off(const struct intel_crtc_state *old_crtc_state) > > } > > /* Only for pre-ILK configs */ > > -void assert_pll(struct drm_i915_private *dev_priv, > > +void assert_pll(struct drm_i915_private *i915, > > enum pipe pipe, bool state) > > { > > u32 val; > > @@ -1099,14 +1099,14 @@ void assert_pll(struct drm_i915_private *dev_priv, > > } > > /* XXX: the dsi pll is shared between MIPI DSI ports */ > > -void assert_dsi_pll(struct drm_i915_private *dev_priv, bool state) > > +void assert_dsi_pll(struct drm_i915_private *i915, bool state) > > { > > u32 val; > > bool cur_state; > > - vlv_cck_get(dev_priv); > > - val = vlv_cck_read(dev_priv, CCK_REG_DSI_PLL_CONTROL); > > - vlv_cck_put(dev_priv); > > + vlv_cck_get(i915); > > + val = vlv_cck_read(i915, CCK_REG_DSI_PLL_CONTROL); > > + vlv_cck_put(i915); > > cur_state = val & DSI_PLL_VCO_EN; > > I915_STATE_WARN(cur_state != state, > > @@ -1114,14 +1114,14 @@ void assert_dsi_pll(struct drm_i915_private *dev_priv, bool state) > > onoff(state), onoff(cur_state)); > > } > > -static void assert_fdi_tx(struct drm_i915_private *dev_priv, > > +static void assert_fdi_tx(struct drm_i915_private *i915, > > enum pipe pipe, bool state) > > { > > bool cur_state; > > - enum transcoder cpu_transcoder = intel_pipe_to_cpu_transcoder(dev_priv, > > + enum transcoder cpu_transcoder = intel_pipe_to_cpu_transcoder(i915, > > pipe); > > - if (HAS_DDI(dev_priv)) { > > + if (HAS_DDI(i915)) { > > /* DDI does not have a specific FDI_TX register */ > > u32 val = I915_READ(TRANS_DDI_FUNC_CTL(cpu_transcoder)); > > cur_state = !!(val & TRANS_DDI_FUNC_ENABLE); > > @@ -1136,7 +1136,7 @@ static void assert_fdi_tx(struct drm_i915_private *dev_priv, > > #define assert_fdi_tx_enabled(d, p) assert_fdi_tx(d, p, true) > > #define assert_fdi_tx_disabled(d, p) assert_fdi_tx(d, p, false) > > -static void assert_fdi_rx(struct drm_i915_private *dev_priv, > > +static void assert_fdi_rx(struct drm_i915_private *i915, > > enum pipe pipe, bool state) > > { > > u32 val; > > @@ -1151,24 +1151,24 @@ static void assert_fdi_rx(struct drm_i915_private *dev_priv, > > #define assert_fdi_rx_enabled(d, p) assert_fdi_rx(d, p, true) > > #define assert_fdi_rx_disabled(d, p) assert_fdi_rx(d, p, false) > > -static void assert_fdi_tx_pll_enabled(struct drm_i915_private *dev_priv, > > +static void assert_fdi_tx_pll_enabled(struct drm_i915_private *i915, > > enum pipe pipe) > > { > > u32 val; > > /* ILK FDI PLL is always enabled */ > > - if (IS_GEN(dev_priv, 5)) > > + if (IS_GEN(i915, 5)) > > return; > > /* On Haswell, DDI ports are responsible for the FDI PLL setup */ > > - if (HAS_DDI(dev_priv)) > > + if (HAS_DDI(i915)) > > return; > > val = I915_READ(FDI_TX_CTL(pipe)); > > I915_STATE_WARN(!(val & FDI_TX_PLL_ENABLE), "FDI TX PLL assertion failure, should be active but is disabled\n"); > > } > > -void assert_fdi_rx_pll(struct drm_i915_private *dev_priv, > > +void assert_fdi_rx_pll(struct drm_i915_private *i915, > > enum pipe pipe, bool state) > > { > > u32 val; > > @@ -1181,17 +1181,17 @@ void assert_fdi_rx_pll(struct drm_i915_private *dev_priv, > > onoff(state), onoff(cur_state)); > > } > > -void assert_panel_unlocked(struct drm_i915_private *dev_priv, enum pipe pipe) > > +void assert_panel_unlocked(struct drm_i915_private *i915, enum pipe pipe) > > { > > i915_reg_t pp_reg; > > u32 val; > > enum pipe panel_pipe = INVALID_PIPE; > > bool locked = true; > > - if (WARN_ON(HAS_DDI(dev_priv))) > > + if (WARN_ON(HAS_DDI(i915))) > > return; > > - if (HAS_PCH_SPLIT(dev_priv)) { > > + if (HAS_PCH_SPLIT(i915)) { > > u32 port_sel; > > pp_reg = PP_CONTROL(0); > > @@ -1199,22 +1199,22 @@ void assert_panel_unlocked(struct drm_i915_private *dev_priv, enum pipe pipe) > > switch (port_sel) { > > case PANEL_PORT_SELECT_LVDS: > > - intel_lvds_port_enabled(dev_priv, PCH_LVDS, &panel_pipe); > > + intel_lvds_port_enabled(i915, PCH_LVDS, &panel_pipe); > > break; > > case PANEL_PORT_SELECT_DPA: > > - intel_dp_port_enabled(dev_priv, DP_A, PORT_A, &panel_pipe); > > + intel_dp_port_enabled(i915, DP_A, PORT_A, &panel_pipe); > > break; > > case PANEL_PORT_SELECT_DPC: > > - intel_dp_port_enabled(dev_priv, PCH_DP_C, PORT_C, &panel_pipe); > > + intel_dp_port_enabled(i915, PCH_DP_C, PORT_C, &panel_pipe); > > break; > > case PANEL_PORT_SELECT_DPD: > > - intel_dp_port_enabled(dev_priv, PCH_DP_D, PORT_D, &panel_pipe); > > + intel_dp_port_enabled(i915, PCH_DP_D, PORT_D, &panel_pipe); > > break; > > default: > > MISSING_CASE(port_sel); > > break; > > } > > - } else if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) { > > + } else if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) { > > /* presumably write lock depends on pipe, not port select */ > > pp_reg = PP_CONTROL(pipe); > > panel_pipe = pipe; > > @@ -1225,7 +1225,7 @@ void assert_panel_unlocked(struct drm_i915_private *dev_priv, enum pipe pipe) > > port_sel = I915_READ(PP_ON_DELAYS(0)) & PANEL_PORT_SELECT_MASK; > > WARN_ON(port_sel != PANEL_PORT_SELECT_LVDS); > > - intel_lvds_port_enabled(dev_priv, LVDS, &panel_pipe); > > + intel_lvds_port_enabled(i915, LVDS, &panel_pipe); > > } > > val = I915_READ(pp_reg); > > @@ -1238,26 +1238,26 @@ void assert_panel_unlocked(struct drm_i915_private *dev_priv, enum pipe pipe) > > pipe_name(pipe)); > > } > > -void assert_pipe(struct drm_i915_private *dev_priv, > > +void assert_pipe(struct drm_i915_private *i915, > > enum pipe pipe, bool state) > > { > > bool cur_state; > > - enum transcoder cpu_transcoder = intel_pipe_to_cpu_transcoder(dev_priv, > > + enum transcoder cpu_transcoder = intel_pipe_to_cpu_transcoder(i915, > > pipe); > > enum intel_display_power_domain power_domain; > > intel_wakeref_t wakeref; > > /* we keep both pipes enabled on 830 */ > > - if (IS_I830(dev_priv)) > > + if (IS_I830(i915)) > > state = true; > > power_domain = POWER_DOMAIN_TRANSCODER(cpu_transcoder); > > - wakeref = intel_display_power_get_if_enabled(dev_priv, power_domain); > > + wakeref = intel_display_power_get_if_enabled(i915, power_domain); > > if (wakeref) { > > u32 val = I915_READ(PIPECONF(cpu_transcoder)); > > cur_state = !!(val & PIPECONF_ENABLE); > > - intel_display_power_put(dev_priv, power_domain, wakeref); > > + intel_display_power_put(i915, power_domain, wakeref); > > } else { > > cur_state = false; > > } > > @@ -1284,10 +1284,10 @@ static void assert_plane(struct intel_plane *plane, bool state) > > static void assert_planes_disabled(struct intel_crtc *crtc) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > struct intel_plane *plane; > > - for_each_intel_plane_on_crtc(&dev_priv->drm, crtc, plane) > > + for_each_intel_plane_on_crtc(&i915->drm, crtc, plane) > > assert_plane_disabled(plane); > > } > > @@ -1297,7 +1297,7 @@ static void assert_vblank_disabled(struct drm_crtc *crtc) > > drm_crtc_vblank_put(crtc); > > } > > -void assert_pch_transcoder_disabled(struct drm_i915_private *dev_priv, > > +void assert_pch_transcoder_disabled(struct drm_i915_private *i915, > > enum pipe pipe) > > { > > u32 val; > > @@ -1310,78 +1310,78 @@ void assert_pch_transcoder_disabled(struct drm_i915_private *dev_priv, > > pipe_name(pipe)); > > } > > -static void assert_pch_dp_disabled(struct drm_i915_private *dev_priv, > > +static void assert_pch_dp_disabled(struct drm_i915_private *i915, > > enum pipe pipe, enum port port, > > i915_reg_t dp_reg) > > { > > enum pipe port_pipe; > > bool state; > > - state = intel_dp_port_enabled(dev_priv, dp_reg, port, &port_pipe); > > + state = intel_dp_port_enabled(i915, dp_reg, port, &port_pipe); > > I915_STATE_WARN(state && port_pipe == pipe, > > "PCH DP %c enabled on transcoder %c, should be disabled\n", > > port_name(port), pipe_name(pipe)); > > - I915_STATE_WARN(HAS_PCH_IBX(dev_priv) && !state && port_pipe == PIPE_B, > > + I915_STATE_WARN(HAS_PCH_IBX(i915) && !state && port_pipe == PIPE_B, > > "IBX PCH DP %c still using transcoder B\n", > > port_name(port)); > > } > > -static void assert_pch_hdmi_disabled(struct drm_i915_private *dev_priv, > > +static void assert_pch_hdmi_disabled(struct drm_i915_private *i915, > > enum pipe pipe, enum port port, > > i915_reg_t hdmi_reg) > > { > > enum pipe port_pipe; > > bool state; > > - state = intel_sdvo_port_enabled(dev_priv, hdmi_reg, &port_pipe); > > + state = intel_sdvo_port_enabled(i915, hdmi_reg, &port_pipe); > > I915_STATE_WARN(state && port_pipe == pipe, > > "PCH HDMI %c enabled on transcoder %c, should be disabled\n", > > port_name(port), pipe_name(pipe)); > > - I915_STATE_WARN(HAS_PCH_IBX(dev_priv) && !state && port_pipe == PIPE_B, > > + I915_STATE_WARN(HAS_PCH_IBX(i915) && !state && port_pipe == PIPE_B, > > "IBX PCH HDMI %c still using transcoder B\n", > > port_name(port)); > > } > > -static void assert_pch_ports_disabled(struct drm_i915_private *dev_priv, > > +static void assert_pch_ports_disabled(struct drm_i915_private *i915, > > enum pipe pipe) > > { > > enum pipe port_pipe; > > - assert_pch_dp_disabled(dev_priv, pipe, PORT_B, PCH_DP_B); > > - assert_pch_dp_disabled(dev_priv, pipe, PORT_C, PCH_DP_C); > > - assert_pch_dp_disabled(dev_priv, pipe, PORT_D, PCH_DP_D); > > + assert_pch_dp_disabled(i915, pipe, PORT_B, PCH_DP_B); > > + assert_pch_dp_disabled(i915, pipe, PORT_C, PCH_DP_C); > > + assert_pch_dp_disabled(i915, pipe, PORT_D, PCH_DP_D); > > - I915_STATE_WARN(intel_crt_port_enabled(dev_priv, PCH_ADPA, &port_pipe) && > > + I915_STATE_WARN(intel_crt_port_enabled(i915, PCH_ADPA, &port_pipe) && > > port_pipe == pipe, > > "PCH VGA enabled on transcoder %c, should be disabled\n", > > pipe_name(pipe)); > > - I915_STATE_WARN(intel_lvds_port_enabled(dev_priv, PCH_LVDS, &port_pipe) && > > + I915_STATE_WARN(intel_lvds_port_enabled(i915, PCH_LVDS, &port_pipe) && > > port_pipe == pipe, > > "PCH LVDS enabled on transcoder %c, should be disabled\n", > > pipe_name(pipe)); > > /* PCH SDVOB multiplex with HDMIB */ > > - assert_pch_hdmi_disabled(dev_priv, pipe, PORT_B, PCH_HDMIB); > > - assert_pch_hdmi_disabled(dev_priv, pipe, PORT_C, PCH_HDMIC); > > - assert_pch_hdmi_disabled(dev_priv, pipe, PORT_D, PCH_HDMID); > > + assert_pch_hdmi_disabled(i915, pipe, PORT_B, PCH_HDMIB); > > + assert_pch_hdmi_disabled(i915, pipe, PORT_C, PCH_HDMIC); > > + assert_pch_hdmi_disabled(i915, pipe, PORT_D, PCH_HDMID); > > } > > static void _vlv_enable_pll(struct intel_crtc *crtc, > > const struct intel_crtc_state *pipe_config) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > enum pipe pipe = crtc->pipe; > > I915_WRITE(DPLL(pipe), pipe_config->dpll_hw_state.dpll); > > POSTING_READ(DPLL(pipe)); > > udelay(150); > > - if (intel_wait_for_register(&dev_priv->uncore, > > + if (intel_wait_for_register(&i915->uncore, > > DPLL(pipe), > > DPLL_LOCK_VLV, > > DPLL_LOCK_VLV, > > @@ -1392,13 +1392,13 @@ static void _vlv_enable_pll(struct intel_crtc *crtc, > > static void vlv_enable_pll(struct intel_crtc *crtc, > > const struct intel_crtc_state *pipe_config) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > enum pipe pipe = crtc->pipe; > > - assert_pipe_disabled(dev_priv, pipe); > > + assert_pipe_disabled(i915, pipe); > > /* PLL is protected by panel, make sure we can write it */ > > - assert_panel_unlocked(dev_priv, pipe); > > + assert_panel_unlocked(i915, pipe); > > if (pipe_config->dpll_hw_state.dpll & DPLL_VCO_ENABLE) > > _vlv_enable_pll(crtc, pipe_config); > > @@ -1411,19 +1411,19 @@ static void vlv_enable_pll(struct intel_crtc *crtc, > > static void _chv_enable_pll(struct intel_crtc *crtc, > > const struct intel_crtc_state *pipe_config) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > enum pipe pipe = crtc->pipe; > > enum dpio_channel port = vlv_pipe_to_channel(pipe); > > u32 tmp; > > - vlv_dpio_get(dev_priv); > > + vlv_dpio_get(i915); > > /* Enable back the 10bit clock to display controller */ > > - tmp = vlv_dpio_read(dev_priv, pipe, CHV_CMN_DW14(port)); > > + tmp = vlv_dpio_read(i915, pipe, CHV_CMN_DW14(port)); > > tmp |= DPIO_DCLKP_EN; > > - vlv_dpio_write(dev_priv, pipe, CHV_CMN_DW14(port), tmp); > > + vlv_dpio_write(i915, pipe, CHV_CMN_DW14(port), tmp); > > - vlv_dpio_put(dev_priv); > > + vlv_dpio_put(i915); > > /* > > * Need to wait > 100ns between dclkp clock enable bit and PLL enable. > > @@ -1434,7 +1434,7 @@ static void _chv_enable_pll(struct intel_crtc *crtc, > > I915_WRITE(DPLL(pipe), pipe_config->dpll_hw_state.dpll); > > /* Check PLL is locked */ > > - if (intel_wait_for_register(&dev_priv->uncore, > > + if (intel_wait_for_register(&i915->uncore, > > DPLL(pipe), DPLL_LOCK_VLV, DPLL_LOCK_VLV, > > 1)) > > DRM_ERROR("PLL %d failed to lock\n", pipe); > > @@ -1443,13 +1443,13 @@ static void _chv_enable_pll(struct intel_crtc *crtc, > > static void chv_enable_pll(struct intel_crtc *crtc, > > const struct intel_crtc_state *pipe_config) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > enum pipe pipe = crtc->pipe; > > - assert_pipe_disabled(dev_priv, pipe); > > + assert_pipe_disabled(i915, pipe); > > /* PLL is protected by panel, make sure we can write it */ > > - assert_panel_unlocked(dev_priv, pipe); > > + assert_panel_unlocked(i915, pipe); > > if (pipe_config->dpll_hw_state.dpll & DPLL_VCO_ENABLE) > > _chv_enable_pll(crtc, pipe_config); > > @@ -1464,7 +1464,7 @@ static void chv_enable_pll(struct intel_crtc *crtc, > > I915_WRITE(CBR4_VLV, CBR_DPLLBMD_PIPE(pipe)); > > I915_WRITE(DPLL_MD(PIPE_B), pipe_config->dpll_hw_state.dpll_md); > > I915_WRITE(CBR4_VLV, 0); > > - dev_priv->chv_dpll_md[pipe] = pipe_config->dpll_hw_state.dpll_md; > > + i915->chv_dpll_md[pipe] = pipe_config->dpll_hw_state.dpll_md; > > /* > > * DPLLB VGA mode also seems to cause problems. > > @@ -1477,27 +1477,27 @@ static void chv_enable_pll(struct intel_crtc *crtc, > > } > > } > > -static bool i9xx_has_pps(struct drm_i915_private *dev_priv) > > +static bool i9xx_has_pps(struct drm_i915_private *i915) > > { > > - if (IS_I830(dev_priv)) > > + if (IS_I830(i915)) > > return false; > > - return IS_PINEVIEW(dev_priv) || IS_MOBILE(dev_priv); > > + return IS_PINEVIEW(i915) || IS_MOBILE(i915); > > } > > static void i9xx_enable_pll(struct intel_crtc *crtc, > > const struct intel_crtc_state *crtc_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > i915_reg_t reg = DPLL(crtc->pipe); > > u32 dpll = crtc_state->dpll_hw_state.dpll; > > int i; > > - assert_pipe_disabled(dev_priv, crtc->pipe); > > + assert_pipe_disabled(i915, crtc->pipe); > > /* PLL is protected by panel, make sure we can write it */ > > - if (i9xx_has_pps(dev_priv)) > > - assert_panel_unlocked(dev_priv, crtc->pipe); > > + if (i9xx_has_pps(i915)) > > + assert_panel_unlocked(i915, crtc->pipe); > > /* > > * Apparently we need to have VGA mode enabled prior to changing > > @@ -1511,7 +1511,7 @@ static void i9xx_enable_pll(struct intel_crtc *crtc, > > POSTING_READ(reg); > > udelay(150); > > - if (INTEL_GEN(dev_priv) >= 4) { > > + if (INTEL_GEN(i915) >= 4) { > > I915_WRITE(DPLL_MD(crtc->pipe), > > crtc_state->dpll_hw_state.dpll_md); > > } else { > > @@ -1534,26 +1534,26 @@ static void i9xx_enable_pll(struct intel_crtc *crtc, > > static void i9xx_disable_pll(const struct intel_crtc_state *crtc_state) > > { > > struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc); > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > enum pipe pipe = crtc->pipe; > > /* Don't disable pipe or pipe PLLs if needed */ > > - if (IS_I830(dev_priv)) > > + if (IS_I830(i915)) > > return; > > /* Make sure the pipe isn't still relying on us */ > > - assert_pipe_disabled(dev_priv, pipe); > > + assert_pipe_disabled(i915, pipe); > > I915_WRITE(DPLL(pipe), DPLL_VGA_MODE_DIS); > > POSTING_READ(DPLL(pipe)); > > } > > -static void vlv_disable_pll(struct drm_i915_private *dev_priv, enum pipe pipe) > > +static void vlv_disable_pll(struct drm_i915_private *i915, enum pipe pipe) > > { > > u32 val; > > /* Make sure the pipe isn't still relying on us */ > > - assert_pipe_disabled(dev_priv, pipe); > > + assert_pipe_disabled(i915, pipe); > > val = DPLL_INTEGRATED_REF_CLK_VLV | > > DPLL_REF_CLK_ENABLE_VLV | DPLL_VGA_MODE_DIS; > > @@ -1564,13 +1564,13 @@ static void vlv_disable_pll(struct drm_i915_private *dev_priv, enum pipe pipe) > > POSTING_READ(DPLL(pipe)); > > } > > -static void chv_disable_pll(struct drm_i915_private *dev_priv, enum pipe pipe) > > +static void chv_disable_pll(struct drm_i915_private *i915, enum pipe pipe) > > { > > enum dpio_channel port = vlv_pipe_to_channel(pipe); > > u32 val; > > /* Make sure the pipe isn't still relying on us */ > > - assert_pipe_disabled(dev_priv, pipe); > > + assert_pipe_disabled(i915, pipe); > > val = DPLL_SSC_REF_CLK_CHV | > > DPLL_REF_CLK_ENABLE_VLV | DPLL_VGA_MODE_DIS; > > @@ -1580,17 +1580,17 @@ static void chv_disable_pll(struct drm_i915_private *dev_priv, enum pipe pipe) > > I915_WRITE(DPLL(pipe), val); > > POSTING_READ(DPLL(pipe)); > > - vlv_dpio_get(dev_priv); > > + vlv_dpio_get(i915); > > /* Disable 10bit clock to display controller */ > > - val = vlv_dpio_read(dev_priv, pipe, CHV_CMN_DW14(port)); > > + val = vlv_dpio_read(i915, pipe, CHV_CMN_DW14(port)); > > val &= ~DPIO_DCLKP_EN; > > - vlv_dpio_write(dev_priv, pipe, CHV_CMN_DW14(port), val); > > + vlv_dpio_write(i915, pipe, CHV_CMN_DW14(port), val); > > - vlv_dpio_put(dev_priv); > > + vlv_dpio_put(i915); > > } > > -void vlv_wait_port_ready(struct drm_i915_private *dev_priv, > > +void vlv_wait_port_ready(struct drm_i915_private *i915, > > struct intel_digital_port *dport, > > unsigned int expected_mask) > > { > > @@ -1615,7 +1615,7 @@ void vlv_wait_port_ready(struct drm_i915_private *dev_priv, > > BUG(); > > } > > - if (intel_wait_for_register(&dev_priv->uncore, > > + if (intel_wait_for_register(&i915->uncore, > > dpll_reg, port_mask, expected_mask, > > 1000)) > > WARN(1, "timed out waiting for port %c ready: got 0x%x, expected 0x%x\n", > > @@ -1626,19 +1626,19 @@ void vlv_wait_port_ready(struct drm_i915_private *dev_priv, > > static void ironlake_enable_pch_transcoder(const struct intel_crtc_state *crtc_state) > > { > > struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc); > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > enum pipe pipe = crtc->pipe; > > i915_reg_t reg; > > u32 val, pipeconf_val; > > /* Make sure PCH DPLL is enabled */ > > - assert_shared_dpll_enabled(dev_priv, crtc_state->shared_dpll); > > + assert_shared_dpll_enabled(i915, crtc_state->shared_dpll); > > /* FDI must be feeding us bits for PCH ports */ > > - assert_fdi_tx_enabled(dev_priv, pipe); > > - assert_fdi_rx_enabled(dev_priv, pipe); > > + assert_fdi_tx_enabled(i915, pipe); > > + assert_fdi_rx_enabled(i915, pipe); > > - if (HAS_PCH_CPT(dev_priv)) { > > + if (HAS_PCH_CPT(i915)) { > > /* Workaround: Set the timing override bit before enabling the > > * pch transcoder. */ > > reg = TRANS_CHICKEN2(pipe); > > @@ -1651,7 +1651,7 @@ static void ironlake_enable_pch_transcoder(const struct intel_crtc_state *crtc_s > > val = I915_READ(reg); > > pipeconf_val = I915_READ(PIPECONF(pipe)); > > - if (HAS_PCH_IBX(dev_priv)) { > > + if (HAS_PCH_IBX(i915)) { > > /* > > * Make the BPC in transcoder be consistent with > > * that in pipeconf reg. For HDMI we must use 8bpc > > @@ -1666,7 +1666,7 @@ static void ironlake_enable_pch_transcoder(const struct intel_crtc_state *crtc_s > > val &= ~TRANS_INTERLACE_MASK; > > if ((pipeconf_val & PIPECONF_INTERLACE_MASK) == PIPECONF_INTERLACED_ILK) { > > - if (HAS_PCH_IBX(dev_priv) && > > + if (HAS_PCH_IBX(i915) && > > intel_crtc_has_type(crtc_state, INTEL_OUTPUT_SDVO)) > > val |= TRANS_LEGACY_INTERLACED_ILK; > > else > > @@ -1676,20 +1676,20 @@ static void ironlake_enable_pch_transcoder(const struct intel_crtc_state *crtc_s > > } > > I915_WRITE(reg, val | TRANS_ENABLE); > > - if (intel_wait_for_register(&dev_priv->uncore, > > + if (intel_wait_for_register(&i915->uncore, > > reg, TRANS_STATE_ENABLE, TRANS_STATE_ENABLE, > > 100)) > > DRM_ERROR("failed to enable transcoder %c\n", pipe_name(pipe)); > > } > > -static void lpt_enable_pch_transcoder(struct drm_i915_private *dev_priv, > > +static void lpt_enable_pch_transcoder(struct drm_i915_private *i915, > > enum transcoder cpu_transcoder) > > { > > u32 val, pipeconf_val; > > /* FDI must be feeding us bits for PCH ports */ > > - assert_fdi_tx_enabled(dev_priv, (enum pipe) cpu_transcoder); > > - assert_fdi_rx_enabled(dev_priv, PIPE_A); > > + assert_fdi_tx_enabled(i915, (enum pipe) cpu_transcoder); > > + assert_fdi_rx_enabled(i915, PIPE_A); > > /* Workaround: set timing override bit. */ > > val = I915_READ(TRANS_CHICKEN2(PIPE_A)); > > @@ -1706,7 +1706,7 @@ static void lpt_enable_pch_transcoder(struct drm_i915_private *dev_priv, > > val |= TRANS_PROGRESSIVE; > > I915_WRITE(LPT_TRANSCONF, val); > > - if (intel_wait_for_register(&dev_priv->uncore, > > + if (intel_wait_for_register(&i915->uncore, > > LPT_TRANSCONF, > > TRANS_STATE_ENABLE, > > TRANS_STATE_ENABLE, > > @@ -1714,30 +1714,30 @@ static void lpt_enable_pch_transcoder(struct drm_i915_private *dev_priv, > > DRM_ERROR("Failed to enable PCH transcoder\n"); > > } > > -static void ironlake_disable_pch_transcoder(struct drm_i915_private *dev_priv, > > +static void ironlake_disable_pch_transcoder(struct drm_i915_private *i915, > > enum pipe pipe) > > { > > i915_reg_t reg; > > u32 val; > > /* FDI relies on the transcoder */ > > - assert_fdi_tx_disabled(dev_priv, pipe); > > - assert_fdi_rx_disabled(dev_priv, pipe); > > + assert_fdi_tx_disabled(i915, pipe); > > + assert_fdi_rx_disabled(i915, pipe); > > /* Ports must be off as well */ > > - assert_pch_ports_disabled(dev_priv, pipe); > > + assert_pch_ports_disabled(i915, pipe); > > reg = PCH_TRANSCONF(pipe); > > val = I915_READ(reg); > > val &= ~TRANS_ENABLE; > > I915_WRITE(reg, val); > > /* wait for PCH transcoder off, transcoder state */ > > - if (intel_wait_for_register(&dev_priv->uncore, > > + if (intel_wait_for_register(&i915->uncore, > > reg, TRANS_STATE_ENABLE, 0, > > 50)) > > DRM_ERROR("failed to disable transcoder %c\n", pipe_name(pipe)); > > - if (HAS_PCH_CPT(dev_priv)) { > > + if (HAS_PCH_CPT(i915)) { > > /* Workaround: Clear the timing override chicken bit again. */ > > reg = TRANS_CHICKEN2(pipe); > > val = I915_READ(reg); > > @@ -1746,7 +1746,7 @@ static void ironlake_disable_pch_transcoder(struct drm_i915_private *dev_priv, > > } > > } > > -void lpt_disable_pch_transcoder(struct drm_i915_private *dev_priv) > > +void lpt_disable_pch_transcoder(struct drm_i915_private *i915) > > { > > u32 val; > > @@ -1754,7 +1754,7 @@ void lpt_disable_pch_transcoder(struct drm_i915_private *dev_priv) > > val &= ~TRANS_ENABLE; > > I915_WRITE(LPT_TRANSCONF, val); > > /* wait for PCH transcoder off, transcoder state */ > > - if (intel_wait_for_register(&dev_priv->uncore, > > + if (intel_wait_for_register(&i915->uncore, > > LPT_TRANSCONF, TRANS_STATE_ENABLE, 0, > > 50)) > > DRM_ERROR("Failed to disable PCH transcoder\n"); > > @@ -1767,9 +1767,9 @@ void lpt_disable_pch_transcoder(struct drm_i915_private *dev_priv) > > enum pipe intel_crtc_pch_transcoder(struct intel_crtc *crtc) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > - if (HAS_PCH_LPT(dev_priv)) > > + if (HAS_PCH_LPT(i915)) > > return PIPE_A; > > else > > return crtc->pipe; > > @@ -1777,19 +1777,19 @@ enum pipe intel_crtc_pch_transcoder(struct intel_crtc *crtc) > > static u32 intel_crtc_max_vblank_count(const struct intel_crtc_state *crtc_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc_state->base.crtc->dev); > > + struct drm_i915_private *i915 = to_i915(crtc_state->base.crtc->dev); > > /* > > * On i965gm the hardware frame counter reads > > * zero when the TV encoder is enabled :( > > */ > > - if (IS_I965GM(dev_priv) && > > + if (IS_I965GM(i915) && > > (crtc_state->output_types & BIT(INTEL_OUTPUT_TVOUT))) > > return 0; > > - if (INTEL_GEN(dev_priv) >= 5 || IS_G4X(dev_priv)) > > + if (INTEL_GEN(i915) >= 5 || IS_G4X(i915)) > > return 0xffffffff; /* full 32 bit counter */ > > - else if (INTEL_GEN(dev_priv) >= 3) > > + else if (INTEL_GEN(i915) >= 3) > > return 0xffffff; /* only 24 bits of frame count */ > > else > > return 0; /* Gen2 doesn't have a hardware frame counter */ > > @@ -1807,7 +1807,7 @@ static void intel_crtc_vblank_on(const struct intel_crtc_state *crtc_state) > > static void intel_enable_pipe(const struct intel_crtc_state *new_crtc_state) > > { > > struct intel_crtc *crtc = to_intel_crtc(new_crtc_state->base.crtc); > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > enum transcoder cpu_transcoder = new_crtc_state->cpu_transcoder; > > enum pipe pipe = crtc->pipe; > > i915_reg_t reg; > > @@ -1822,29 +1822,29 @@ static void intel_enable_pipe(const struct intel_crtc_state *new_crtc_state) > > * a plane. On ILK+ the pipe PLLs are integrated, so we don't > > * need the check. > > */ > > - if (HAS_GMCH(dev_priv)) { > > + if (HAS_GMCH(i915)) { > > if (intel_crtc_has_type(new_crtc_state, INTEL_OUTPUT_DSI)) > > - assert_dsi_pll_enabled(dev_priv); > > + assert_dsi_pll_enabled(i915); > > else > > - assert_pll_enabled(dev_priv, pipe); > > + assert_pll_enabled(i915, pipe); > > } else { > > if (new_crtc_state->has_pch_encoder) { > > /* if driving the PCH, we need FDI enabled */ > > - assert_fdi_rx_pll_enabled(dev_priv, > > + assert_fdi_rx_pll_enabled(i915, > > intel_crtc_pch_transcoder(crtc)); > > - assert_fdi_tx_pll_enabled(dev_priv, > > + assert_fdi_tx_pll_enabled(i915, > > (enum pipe) cpu_transcoder); > > } > > /* FIXME: assert CPU port conditions for SNB+ */ > > } > > - trace_intel_pipe_enable(dev_priv, pipe); > > + trace_intel_pipe_enable(i915, pipe); > > reg = PIPECONF(cpu_transcoder); > > val = I915_READ(reg); > > if (val & PIPECONF_ENABLE) { > > /* we keep both pipes enabled on 830 */ > > - WARN_ON(!IS_I830(dev_priv)); > > + WARN_ON(!IS_I830(i915)); > > return; > > } > > @@ -1865,7 +1865,7 @@ static void intel_enable_pipe(const struct intel_crtc_state *new_crtc_state) > > static void intel_disable_pipe(const struct intel_crtc_state *old_crtc_state) > > { > > struct intel_crtc *crtc = to_intel_crtc(old_crtc_state->base.crtc); > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > enum transcoder cpu_transcoder = old_crtc_state->cpu_transcoder; > > enum pipe pipe = crtc->pipe; > > i915_reg_t reg; > > @@ -1879,7 +1879,7 @@ static void intel_disable_pipe(const struct intel_crtc_state *old_crtc_state) > > */ > > assert_planes_disabled(crtc); > > - trace_intel_pipe_disable(dev_priv, pipe); > > + trace_intel_pipe_disable(i915, pipe); > > reg = PIPECONF(cpu_transcoder); > > val = I915_READ(reg); > > @@ -1894,7 +1894,7 @@ static void intel_disable_pipe(const struct intel_crtc_state *old_crtc_state) > > val &= ~PIPECONF_DOUBLE_WIDE; > > /* Don't disable pipe or pipe PLLs if needed */ > > - if (!IS_I830(dev_priv)) > > + if (!IS_I830(i915)) > > val &= ~PIPECONF_ENABLE; > > I915_WRITE(reg, val); > > @@ -1902,22 +1902,22 @@ static void intel_disable_pipe(const struct intel_crtc_state *old_crtc_state) > > intel_wait_for_pipe_off(old_crtc_state); > > } > > -static unsigned int intel_tile_size(const struct drm_i915_private *dev_priv) > > +static unsigned int intel_tile_size(const struct drm_i915_private *i915) > > { > > - return IS_GEN(dev_priv, 2) ? 2048 : 4096; > > + return IS_GEN(i915, 2) ? 2048 : 4096; > > } > > static unsigned int > > intel_tile_width_bytes(const struct drm_framebuffer *fb, int color_plane) > > { > > - struct drm_i915_private *dev_priv = to_i915(fb->dev); > > + struct drm_i915_private *i915 = to_i915(fb->dev); > > unsigned int cpp = fb->format->cpp[color_plane]; > > switch (fb->modifier) { > > case DRM_FORMAT_MOD_LINEAR: > > - return intel_tile_size(dev_priv); > > + return intel_tile_size(i915); > > case I915_FORMAT_MOD_X_TILED: > > - if (IS_GEN(dev_priv, 2)) > > + if (IS_GEN(i915, 2)) > > return 128; > > else > > return 512; > > @@ -1926,7 +1926,7 @@ intel_tile_width_bytes(const struct drm_framebuffer *fb, int color_plane) > > return 128; > > /* fall through */ > > case I915_FORMAT_MOD_Y_TILED: > > - if (IS_GEN(dev_priv, 2) || HAS_128_BYTE_Y_TILING(dev_priv)) > > + if (IS_GEN(i915, 2) || HAS_128_BYTE_Y_TILING(i915)) > > return 128; > > else > > return 512; > > @@ -2017,26 +2017,26 @@ intel_fill_fb_ggtt_view(struct i915_ggtt_view *view, > > } > > } > > -static unsigned int intel_cursor_alignment(const struct drm_i915_private *dev_priv) > > +static unsigned int intel_cursor_alignment(const struct drm_i915_private *i915) > > { > > - if (IS_I830(dev_priv)) > > + if (IS_I830(i915)) > > return 16 * 1024; > > - else if (IS_I85X(dev_priv)) > > + else if (IS_I85X(i915)) > > return 256; > > - else if (IS_I845G(dev_priv) || IS_I865G(dev_priv)) > > + else if (IS_I845G(i915) || IS_I865G(i915)) > > return 32; > > else > > return 4 * 1024; > > } > > -static unsigned int intel_linear_alignment(const struct drm_i915_private *dev_priv) > > +static unsigned int intel_linear_alignment(const struct drm_i915_private *i915) > > { > > - if (INTEL_GEN(dev_priv) >= 9) > > + if (INTEL_GEN(i915) >= 9) > > return 256 * 1024; > > - else if (IS_I965G(dev_priv) || IS_I965GM(dev_priv) || > > - IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) > > + else if (IS_I965G(i915) || IS_I965GM(i915) || > > + IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) > > return 128 * 1024; > > - else if (INTEL_GEN(dev_priv) >= 4) > > + else if (INTEL_GEN(i915) >= 4) > > return 4 * 1024; > > else > > return 0; > > @@ -2045,7 +2045,7 @@ static unsigned int intel_linear_alignment(const struct drm_i915_private *dev_pr > > static unsigned int intel_surf_alignment(const struct drm_framebuffer *fb, > > int color_plane) > > { > > - struct drm_i915_private *dev_priv = to_i915(fb->dev); > > + struct drm_i915_private *i915 = to_i915(fb->dev); > > /* AUX_DIST needs only 4K alignment */ > > if (color_plane == 1) > > @@ -2053,9 +2053,9 @@ static unsigned int intel_surf_alignment(const struct drm_framebuffer *fb, > > switch (fb->modifier) { > > case DRM_FORMAT_MOD_LINEAR: > > - return intel_linear_alignment(dev_priv); > > + return intel_linear_alignment(i915); > > case I915_FORMAT_MOD_X_TILED: > > - if (INTEL_GEN(dev_priv) >= 9) > > + if (INTEL_GEN(i915) >= 9) > > return 256 * 1024; > > return 0; > > case I915_FORMAT_MOD_Y_TILED_CCS: > > @@ -2072,9 +2072,9 @@ static unsigned int intel_surf_alignment(const struct drm_framebuffer *fb, > > static bool intel_plane_uses_fence(const struct intel_plane_state *plane_state) > > { > > struct intel_plane *plane = to_intel_plane(plane_state->base.plane); > > - struct drm_i915_private *dev_priv = to_i915(plane->base.dev); > > + struct drm_i915_private *i915 = to_i915(plane->base.dev); > > - return INTEL_GEN(dev_priv) < 4 || > > + return INTEL_GEN(i915) < 4 || > > (plane->has_fbc && > > plane_state->view.type == I915_GGTT_VIEW_NORMAL); > > } > > @@ -2086,7 +2086,7 @@ intel_pin_and_fence_fb_obj(struct drm_framebuffer *fb, > > unsigned long *out_flags) > > { > > struct drm_device *dev = fb->dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > struct drm_i915_gem_object *obj = intel_fb_obj(fb); > > intel_wakeref_t wakeref; > > struct i915_vma *vma; > > @@ -2102,7 +2102,7 @@ intel_pin_and_fence_fb_obj(struct drm_framebuffer *fb, > > * we should always have valid PTE following the scanout preventing > > * the VT-d warning. > > */ > > - if (intel_scanout_needs_vtd_wa(dev_priv) && alignment < 256 * 1024) > > + if (intel_scanout_needs_vtd_wa(i915) && alignment < 256 * 1024) > > alignment = 256 * 1024; > > /* > > @@ -2112,10 +2112,10 @@ intel_pin_and_fence_fb_obj(struct drm_framebuffer *fb, > > * intel_runtime_pm_put(), so it is correct to wrap only the > > * pin/unpin/fence and not more. > > */ > > - wakeref = intel_runtime_pm_get(dev_priv); > > + wakeref = intel_runtime_pm_get(i915); > > i915_gem_object_lock(obj); > > - atomic_inc(&dev_priv->gpu_error.pending_fb_pin); > > + atomic_inc(&i915->gpu_error.pending_fb_pin); > > pinctl = 0; > > @@ -2126,7 +2126,7 @@ intel_pin_and_fence_fb_obj(struct drm_framebuffer *fb, > > * complicated than this. For example, Cherryview appears quite > > * happy to scanout from anywhere within its global aperture. > > */ > > - if (HAS_GMCH(dev_priv)) > > + if (HAS_GMCH(i915)) > > pinctl |= PIN_MAPPABLE; > > vma = i915_gem_object_pin_to_display_plane(obj, > > @@ -2154,7 +2154,7 @@ intel_pin_and_fence_fb_obj(struct drm_framebuffer *fb, > > * mode that matches the user configuration. > > */ > > ret = i915_vma_pin_fence(vma); > > - if (ret != 0 && INTEL_GEN(dev_priv) < 4) { > > + if (ret != 0 && INTEL_GEN(i915) < 4) { > > i915_gem_object_unpin_from_display_plane(vma); > > vma = ERR_PTR(ret); > > goto err; > > @@ -2166,10 +2166,10 @@ intel_pin_and_fence_fb_obj(struct drm_framebuffer *fb, > > i915_vma_get(vma); > > err: > > - atomic_dec(&dev_priv->gpu_error.pending_fb_pin); > > + atomic_dec(&i915->gpu_error.pending_fb_pin); > > i915_gem_object_unlock(obj); > > - intel_runtime_pm_put(dev_priv, wakeref); > > + intel_runtime_pm_put(i915, wakeref); > > return vma; > > } > > @@ -2265,7 +2265,7 @@ static u32 intel_adjust_aligned_offset(int *x, int *y, > > unsigned int pitch, > > u32 old_offset, u32 new_offset) > > { > > - struct drm_i915_private *dev_priv = to_i915(fb->dev); > > + struct drm_i915_private *i915 = to_i915(fb->dev); > > unsigned int cpp = fb->format->cpp[color_plane]; > > WARN_ON(new_offset > old_offset); > > @@ -2274,7 +2274,7 @@ static u32 intel_adjust_aligned_offset(int *x, int *y, > > unsigned int tile_size, tile_width, tile_height; > > unsigned int pitch_tiles; > > - tile_size = intel_tile_size(dev_priv); > > + tile_size = intel_tile_size(i915); > > intel_tile_dims(fb, color_plane, &tile_width, &tile_height); > > if (drm_rotation_90_or_270(rotation)) { > > @@ -2326,7 +2326,7 @@ static u32 intel_plane_adjust_aligned_offset(int *x, int *y, > > * used. This is why the user has to pass in the pitch since it > > * is specified in the rotated orientation. > > */ > > -static u32 intel_compute_aligned_offset(struct drm_i915_private *dev_priv, > > +static u32 intel_compute_aligned_offset(struct drm_i915_private *i915, > > int *x, int *y, > > const struct drm_framebuffer *fb, > > int color_plane, > > @@ -2344,7 +2344,7 @@ static u32 intel_compute_aligned_offset(struct drm_i915_private *dev_priv, > > unsigned int tile_size, tile_width, tile_height; > > unsigned int tile_rows, tiles, pitch_tiles; > > - tile_size = intel_tile_size(dev_priv); > > + tile_size = intel_tile_size(i915); > > intel_tile_dims(fb, color_plane, &tile_width, &tile_height); > > if (drm_rotation_90_or_270(rotation)) { > > @@ -2382,18 +2382,18 @@ static u32 intel_plane_compute_aligned_offset(int *x, int *y, > > int color_plane) > > { > > struct intel_plane *intel_plane = to_intel_plane(state->base.plane); > > - struct drm_i915_private *dev_priv = to_i915(intel_plane->base.dev); > > + struct drm_i915_private *i915 = to_i915(intel_plane->base.dev); > > const struct drm_framebuffer *fb = state->base.fb; > > unsigned int rotation = state->base.rotation; > > int pitch = state->color_plane[color_plane].stride; > > u32 alignment; > > if (intel_plane->id == PLANE_CURSOR) > > - alignment = intel_cursor_alignment(dev_priv); > > + alignment = intel_cursor_alignment(i915); > > else > > alignment = intel_surf_alignment(fb, color_plane); > > - return intel_compute_aligned_offset(dev_priv, x, y, fb, color_plane, > > + return intel_compute_aligned_offset(i915, x, y, fb, color_plane, > > pitch, rotation, alignment); > > } > > @@ -2402,11 +2402,11 @@ static int intel_fb_offset_to_xy(int *x, int *y, > > const struct drm_framebuffer *fb, > > int color_plane) > > { > > - struct drm_i915_private *dev_priv = to_i915(fb->dev); > > + struct drm_i915_private *i915 = to_i915(fb->dev); > > unsigned int height; > > if (fb->modifier != DRM_FORMAT_MOD_LINEAR && > > - fb->offsets[color_plane] % intel_tile_size(dev_priv)) { > > + fb->offsets[color_plane] % intel_tile_size(i915)) { > > DRM_DEBUG_KMS("Misaligned offset 0x%08x for color plane %d\n", > > fb->offsets[color_plane], color_plane); > > return -EINVAL; > > @@ -2507,7 +2507,7 @@ bool is_ccs_modifier(u64 modifier) > > modifier == I915_FORMAT_MOD_Yf_TILED_CCS; > > } > > -u32 intel_plane_fb_max_stride(struct drm_i915_private *dev_priv, > > +u32 intel_plane_fb_max_stride(struct drm_i915_private *i915, > > u32 pixel_format, u64 modifier) > > { > > struct intel_crtc *crtc; > > @@ -2517,7 +2517,7 @@ u32 intel_plane_fb_max_stride(struct drm_i915_private *dev_priv, > > * We assume the primary plane for pipe A has > > * the highest stride limits of them all. > > */ > > - crtc = intel_get_crtc_for_pipe(dev_priv, PIPE_A); > > + crtc = intel_get_crtc_for_pipe(i915, PIPE_A); > > plane = to_intel_plane(crtc->base.primary); > > return plane->max_stride(plane, pixel_format, modifier, > > @@ -2525,7 +2525,7 @@ u32 intel_plane_fb_max_stride(struct drm_i915_private *dev_priv, > > } > > static > > -u32 intel_fb_max_stride(struct drm_i915_private *dev_priv, > > +u32 intel_fb_max_stride(struct drm_i915_private *i915, > > u32 pixel_format, u64 modifier) > > { > > /* > > @@ -2535,22 +2535,22 @@ u32 intel_fb_max_stride(struct drm_i915_private *dev_priv, > > * The new CCS hash mode makes remapping impossible > > */ > > if (!is_ccs_modifier(modifier)) { > > - if (INTEL_GEN(dev_priv) >= 7) > > + if (INTEL_GEN(i915) >= 7) > > return 256*1024; > > - else if (INTEL_GEN(dev_priv) >= 4) > > + else if (INTEL_GEN(i915) >= 4) > > return 128*1024; > > } > > - return intel_plane_fb_max_stride(dev_priv, pixel_format, modifier); > > + return intel_plane_fb_max_stride(i915, pixel_format, modifier); > > } > > static u32 > > intel_fb_stride_alignment(const struct drm_framebuffer *fb, int color_plane) > > { > > - struct drm_i915_private *dev_priv = to_i915(fb->dev); > > + struct drm_i915_private *i915 = to_i915(fb->dev); > > if (fb->modifier == DRM_FORMAT_MOD_LINEAR) { > > - u32 max_stride = intel_plane_fb_max_stride(dev_priv, > > + u32 max_stride = intel_plane_fb_max_stride(i915, > > fb->format->format, > > fb->modifier); > > @@ -2559,7 +2559,7 @@ intel_fb_stride_alignment(const struct drm_framebuffer *fb, int color_plane) > > * we need the stride to be page aligned. > > */ > > if (fb->pitches[color_plane] > max_stride) > > - return intel_tile_size(dev_priv); > > + return intel_tile_size(i915); > > else > > return 64; > > } else { > > @@ -2570,7 +2570,7 @@ intel_fb_stride_alignment(const struct drm_framebuffer *fb, int color_plane) > > bool intel_plane_can_remap(const struct intel_plane_state *plane_state) > > { > > struct intel_plane *plane = to_intel_plane(plane_state->base.plane); > > - struct drm_i915_private *dev_priv = to_i915(plane->base.dev); > > + struct drm_i915_private *i915 = to_i915(plane->base.dev); > > const struct drm_framebuffer *fb = plane_state->base.fb; > > int i; > > @@ -2584,7 +2584,7 @@ bool intel_plane_can_remap(const struct intel_plane_state *plane_state) > > * Would also need to deal with the fence POT alignment > > * and gen2 2KiB GTT tile size. > > */ > > - if (INTEL_GEN(dev_priv) < 4) > > + if (INTEL_GEN(i915) < 4) > > return false; > > /* > > @@ -2596,7 +2596,7 @@ bool intel_plane_can_remap(const struct intel_plane_state *plane_state) > > /* Linear needs a page aligned stride for remapping */ > > if (fb->modifier == DRM_FORMAT_MOD_LINEAR) { > > - unsigned int alignment = intel_tile_size(dev_priv) - 1; > > + unsigned int alignment = intel_tile_size(i915) - 1; > > for (i = 0; i < fb->format->num_planes; i++) { > > if (fb->pitches[i] & alignment) > > @@ -2636,7 +2636,7 @@ static bool intel_plane_needs_remap(const struct intel_plane_state *plane_state) > > } > > static int > > -intel_fill_fb_info(struct drm_i915_private *dev_priv, > > +intel_fill_fb_info(struct drm_i915_private *i915, > > struct drm_framebuffer *fb) > > { > > struct intel_framebuffer *intel_fb = to_intel_framebuffer(fb); > > @@ -2645,7 +2645,7 @@ intel_fill_fb_info(struct drm_i915_private *dev_priv, > > u32 gtt_offset_rotated = 0; > > unsigned int max_size = 0; > > int i, num_planes = fb->format->num_planes; > > - unsigned int tile_size = intel_tile_size(dev_priv); > > + unsigned int tile_size = intel_tile_size(i915); > > for (i = 0; i < num_planes; i++) { > > unsigned int width, height; > > @@ -2719,7 +2719,7 @@ intel_fill_fb_info(struct drm_i915_private *dev_priv, > > intel_fb->normal[i].x = x; > > intel_fb->normal[i].y = y; > > - offset = intel_compute_aligned_offset(dev_priv, &x, &y, fb, i, > > + offset = intel_compute_aligned_offset(i915, &x, &y, fb, i, > > fb->pitches[i], > > DRM_MODE_ROTATE_0, > > tile_size); > > @@ -2803,14 +2803,14 @@ intel_fill_fb_info(struct drm_i915_private *dev_priv, > > static void > > intel_plane_remap_gtt(struct intel_plane_state *plane_state) > > { > > - struct drm_i915_private *dev_priv = > > + struct drm_i915_private *i915 = > > to_i915(plane_state->base.plane->dev); > > struct drm_framebuffer *fb = plane_state->base.fb; > > struct intel_framebuffer *intel_fb = to_intel_framebuffer(fb); > > struct intel_rotation_info *info = &plane_state->view.rotated; > > unsigned int rotation = plane_state->base.rotation; > > int i, num_planes = fb->format->num_planes; > > - unsigned int tile_size = intel_tile_size(dev_priv); > > + unsigned int tile_size = intel_tile_size(i915); > > unsigned int src_x, src_y; > > unsigned int src_w, src_h; > > u32 gtt_offset = 0; > > @@ -2860,7 +2860,7 @@ intel_plane_remap_gtt(struct intel_plane_state *plane_state) > > x += intel_fb->normal[i].x; > > y += intel_fb->normal[i].y; > > - offset = intel_compute_aligned_offset(dev_priv, &x, &y, > > + offset = intel_compute_aligned_offset(i915, &x, &y, > > fb, i, fb->pitches[i], > > DRM_MODE_ROTATE_0, tile_size); > > offset /= tile_size; > > @@ -3046,7 +3046,7 @@ intel_alloc_initial_plane_obj(struct intel_crtc *crtc, > > struct intel_initial_plane_config *plane_config) > > { > > struct drm_device *dev = crtc->base.dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > struct drm_i915_gem_object *obj = NULL; > > struct drm_mode_fb_cmd2 mode_cmd = { 0 }; > > struct drm_framebuffer *fb = &plane_config->fb->base; > > @@ -3062,7 +3062,7 @@ intel_alloc_initial_plane_obj(struct intel_crtc *crtc, > > /* If the FB is too big, just don't use it since fbdev is not very > > * important and we should probably use that space with FBC or other > > * features. */ > > - if (size_aligned * 2 > dev_priv->stolen_usable_size) > > + if (size_aligned * 2 > i915->stolen_usable_size) > > return false; > > switch (fb->modifier) { > > @@ -3077,7 +3077,7 @@ intel_alloc_initial_plane_obj(struct intel_crtc *crtc, > > } > > mutex_lock(&dev->struct_mutex); > > - obj = i915_gem_object_create_stolen_for_preallocated(dev_priv, > > + obj = i915_gem_object_create_stolen_for_preallocated(i915, > > base_aligned, > > base_aligned, > > size_aligned); > > @@ -3135,7 +3135,7 @@ intel_set_plane_visible(struct intel_crtc_state *crtc_state, > > static void fixup_active_planes(struct intel_crtc_state *crtc_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc_state->base.crtc->dev); > > + struct drm_i915_private *i915 = to_i915(crtc_state->base.crtc->dev); > > struct drm_plane *plane; > > /* > > @@ -3145,7 +3145,7 @@ static void fixup_active_planes(struct intel_crtc_state *crtc_state) > > */ > > crtc_state->active_planes = 0; > > - drm_for_each_plane_mask(plane, &dev_priv->drm, > > + drm_for_each_plane_mask(plane, &i915->drm, > > crtc_state->base.plane_mask) > > crtc_state->active_planes |= BIT(to_intel_plane(plane)->id); > > } > > @@ -3177,7 +3177,7 @@ intel_find_initial_plane_obj(struct intel_crtc *intel_crtc, > > struct intel_initial_plane_config *plane_config) > > { > > struct drm_device *dev = intel_crtc->base.dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > struct drm_crtc *c; > > struct drm_i915_gem_object *obj; > > struct drm_plane *primary = intel_crtc->base.primary; > > @@ -3272,7 +3272,7 @@ intel_find_initial_plane_obj(struct intel_crtc *intel_crtc, > > intel_state->base.dst = drm_plane_state_dest(plane_state); > > if (i915_gem_object_is_tiled(obj)) > > - dev_priv->preserve_bios_swizzle = true; > > + i915->preserve_bios_swizzle = true; > > plane_state->fb = fb; > > plane_state->crtc = &intel_crtc->base; > > @@ -3381,7 +3381,7 @@ static bool skl_check_main_ccs_coordinates(struct intel_plane_state *plane_state > > static int skl_check_main_surface(struct intel_plane_state *plane_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(plane_state->base.plane->dev); > > + struct drm_i915_private *i915 = to_i915(plane_state->base.plane->dev); > > const struct drm_framebuffer *fb = plane_state->base.fb; > > unsigned int rotation = plane_state->base.rotation; > > int x = plane_state->base.src.x1 >> 16; > > @@ -3392,9 +3392,9 @@ static int skl_check_main_surface(struct intel_plane_state *plane_state) > > int max_height = 4096; > > u32 alignment, offset, aux_offset = plane_state->color_plane[1].offset; > > - if (INTEL_GEN(dev_priv) >= 11) > > + if (INTEL_GEN(i915) >= 11) > > max_width = icl_max_plane_width(fb, 0, rotation); > > - else if (INTEL_GEN(dev_priv) >= 10 || IS_GEMINILAKE(dev_priv)) > > + else if (INTEL_GEN(i915) >= 10 || IS_GEMINILAKE(i915)) > > max_width = glk_max_plane_width(fb, 0, rotation); > > else > > max_width = skl_max_plane_width(fb, 0, rotation); > > @@ -3564,16 +3564,16 @@ i9xx_plane_max_stride(struct intel_plane *plane, > > u32 pixel_format, u64 modifier, > > unsigned int rotation) > > { > > - struct drm_i915_private *dev_priv = to_i915(plane->base.dev); > > + struct drm_i915_private *i915 = to_i915(plane->base.dev); > > - if (!HAS_GMCH(dev_priv)) { > > + if (!HAS_GMCH(i915)) { > > return 32*1024; > > - } else if (INTEL_GEN(dev_priv) >= 4) { > > + } else if (INTEL_GEN(i915) >= 4) { > > if (modifier == I915_FORMAT_MOD_X_TILED) > > return 16*1024; > > else > > return 32*1024; > > - } else if (INTEL_GEN(dev_priv) >= 3) { > > + } else if (INTEL_GEN(i915) >= 3) { > > if (modifier == I915_FORMAT_MOD_X_TILED) > > return 8*1024; > > else > > @@ -3589,7 +3589,7 @@ i9xx_plane_max_stride(struct intel_plane *plane, > > static u32 i9xx_plane_ctl_crtc(const struct intel_crtc_state *crtc_state) > > { > > struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc); > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > u32 dspcntr = 0; > > if (crtc_state->gamma_enable) > > @@ -3598,7 +3598,7 @@ static u32 i9xx_plane_ctl_crtc(const struct intel_crtc_state *crtc_state) > > if (crtc_state->csc_enable) > > dspcntr |= DISPPLANE_PIPE_CSC_ENABLE; > > - if (INTEL_GEN(dev_priv) < 5) > > + if (INTEL_GEN(i915) < 5) > > dspcntr |= DISPPLANE_SEL_PIPE(crtc->pipe); > > return dspcntr; > > @@ -3607,7 +3607,7 @@ static u32 i9xx_plane_ctl_crtc(const struct intel_crtc_state *crtc_state) > > static u32 i9xx_plane_ctl(const struct intel_crtc_state *crtc_state, > > const struct intel_plane_state *plane_state) > > { > > - struct drm_i915_private *dev_priv = > > + struct drm_i915_private *i915 = > > to_i915(plane_state->base.plane->dev); > > const struct drm_framebuffer *fb = plane_state->base.fb; > > unsigned int rotation = plane_state->base.rotation; > > @@ -3615,8 +3615,8 @@ static u32 i9xx_plane_ctl(const struct intel_crtc_state *crtc_state, > > dspcntr = DISPLAY_PLANE_ENABLE; > > - if (IS_G4X(dev_priv) || IS_GEN(dev_priv, 5) || > > - IS_GEN(dev_priv, 6) || IS_IVYBRIDGE(dev_priv)) > > + if (IS_G4X(i915) || IS_GEN(i915, 5) || > > + IS_GEN(i915, 6) || IS_IVYBRIDGE(i915)) > > dspcntr |= DISPPLANE_TRICKLE_FEED_DISABLE; > > switch (fb->format->format) { > > @@ -3646,7 +3646,7 @@ static u32 i9xx_plane_ctl(const struct intel_crtc_state *crtc_state, > > return 0; > > } > > - if (INTEL_GEN(dev_priv) >= 4 && > > + if (INTEL_GEN(i915) >= 4 && > > fb->modifier == I915_FORMAT_MOD_X_TILED) > > dspcntr |= DISPPLANE_TILED; > > @@ -3661,7 +3661,7 @@ static u32 i9xx_plane_ctl(const struct intel_crtc_state *crtc_state, > > int i9xx_check_plane_surface(struct intel_plane_state *plane_state) > > { > > - struct drm_i915_private *dev_priv = > > + struct drm_i915_private *i915 = > > to_i915(plane_state->base.plane->dev); > > int src_x, src_y; > > u32 offset; > > @@ -3679,7 +3679,7 @@ int i9xx_check_plane_surface(struct intel_plane_state *plane_state) > > intel_add_fb_offsets(&src_x, &src_y, plane_state, 0); > > - if (INTEL_GEN(dev_priv) >= 4) > > + if (INTEL_GEN(i915) >= 4) > > offset = intel_plane_compute_aligned_offset(&src_x, &src_y, > > plane_state, 0); > > else > > @@ -3694,7 +3694,7 @@ int i9xx_check_plane_surface(struct intel_plane_state *plane_state) > > (src_y << 16) - plane_state->base.src.y1); > > /* HSW/BDW do this automagically in hardware */ > > - if (!IS_HASWELL(dev_priv) && !IS_BROADWELL(dev_priv)) { > > + if (!IS_HASWELL(i915) && !IS_BROADWELL(i915)) { > > unsigned int rotation = plane_state->base.rotation; > > int src_w = drm_rect_width(&plane_state->base.src) >> 16; > > int src_h = drm_rect_height(&plane_state->base.src) >> 16; > > @@ -3752,7 +3752,7 @@ static void i9xx_update_plane(struct intel_plane *plane, > > const struct intel_crtc_state *crtc_state, > > const struct intel_plane_state *plane_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(plane->base.dev); > > + struct drm_i915_private *i915 = to_i915(plane->base.dev); > > enum i9xx_plane_id i9xx_plane = plane->i9xx_plane; > > u32 linear_offset; > > int x = plane_state->color_plane[0].x; > > @@ -3765,16 +3765,16 @@ static void i9xx_update_plane(struct intel_plane *plane, > > linear_offset = intel_fb_xy_to_linear(x, y, plane_state, 0); > > - if (INTEL_GEN(dev_priv) >= 4) > > + if (INTEL_GEN(i915) >= 4) > > dspaddr_offset = plane_state->color_plane[0].offset; > > else > > dspaddr_offset = linear_offset; > > - spin_lock_irqsave(&dev_priv->uncore.lock, irqflags); > > + spin_lock_irqsave(&i915->uncore.lock, irqflags); > > I915_WRITE_FW(DSPSTRIDE(i9xx_plane), plane_state->color_plane[0].stride); > > - if (INTEL_GEN(dev_priv) < 4) { > > + if (INTEL_GEN(i915) < 4) { > > /* pipesrc and dspsize control the size that is scaled from, > > * which should always be the user's requested size. > > */ > > @@ -3782,7 +3782,7 @@ static void i9xx_update_plane(struct intel_plane *plane, > > I915_WRITE_FW(DSPSIZE(i9xx_plane), > > ((crtc_state->pipe_src_h - 1) << 16) | > > (crtc_state->pipe_src_w - 1)); > > - } else if (IS_CHERRYVIEW(dev_priv) && i9xx_plane == PLANE_B) { > > + } else if (IS_CHERRYVIEW(i915) && i9xx_plane == PLANE_B) { > > I915_WRITE_FW(PRIMPOS(i9xx_plane), 0); > > I915_WRITE_FW(PRIMSIZE(i9xx_plane), > > ((crtc_state->pipe_src_h - 1) << 16) | > > @@ -3790,9 +3790,9 @@ static void i9xx_update_plane(struct intel_plane *plane, > > I915_WRITE_FW(PRIMCNSTALPHA(i9xx_plane), 0); > > } > > - if (IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv)) { > > + if (IS_HASWELL(i915) || IS_BROADWELL(i915)) { > > I915_WRITE_FW(DSPOFFSET(i9xx_plane), (y << 16) | x); > > - } else if (INTEL_GEN(dev_priv) >= 4) { > > + } else if (INTEL_GEN(i915) >= 4) { > > I915_WRITE_FW(DSPLINOFF(i9xx_plane), linear_offset); > > I915_WRITE_FW(DSPTILEOFF(i9xx_plane), (y << 16) | x); > > } > > @@ -3803,7 +3803,7 @@ static void i9xx_update_plane(struct intel_plane *plane, > > * the control register just before the surface register. > > */ > > I915_WRITE_FW(DSPCNTR(i9xx_plane), dspcntr); > > - if (INTEL_GEN(dev_priv) >= 4) > > + if (INTEL_GEN(i915) >= 4) > > I915_WRITE_FW(DSPSURF(i9xx_plane), > > intel_plane_ggtt_offset(plane_state) + > > dspaddr_offset); > > @@ -3812,13 +3812,13 @@ static void i9xx_update_plane(struct intel_plane *plane, > > intel_plane_ggtt_offset(plane_state) + > > dspaddr_offset); > > - spin_unlock_irqrestore(&dev_priv->uncore.lock, irqflags); > > + spin_unlock_irqrestore(&i915->uncore.lock, irqflags); > > } > > static void i9xx_disable_plane(struct intel_plane *plane, > > const struct intel_crtc_state *crtc_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(plane->base.dev); > > + struct drm_i915_private *i915 = to_i915(plane->base.dev); > > enum i9xx_plane_id i9xx_plane = plane->i9xx_plane; > > unsigned long irqflags; > > u32 dspcntr; > > @@ -3835,21 +3835,21 @@ static void i9xx_disable_plane(struct intel_plane *plane, > > */ > > dspcntr = i9xx_plane_ctl_crtc(crtc_state); > > - spin_lock_irqsave(&dev_priv->uncore.lock, irqflags); > > + spin_lock_irqsave(&i915->uncore.lock, irqflags); > > I915_WRITE_FW(DSPCNTR(i9xx_plane), dspcntr); > > - if (INTEL_GEN(dev_priv) >= 4) > > + if (INTEL_GEN(i915) >= 4) > > I915_WRITE_FW(DSPSURF(i9xx_plane), 0); > > else > > I915_WRITE_FW(DSPADDR(i9xx_plane), 0); > > - spin_unlock_irqrestore(&dev_priv->uncore.lock, irqflags); > > + spin_unlock_irqrestore(&i915->uncore.lock, irqflags); > > } > > static bool i9xx_plane_get_hw_state(struct intel_plane *plane, > > enum pipe *pipe) > > { > > - struct drm_i915_private *dev_priv = to_i915(plane->base.dev); > > + struct drm_i915_private *i915 = to_i915(plane->base.dev); > > enum intel_display_power_domain power_domain; > > enum i9xx_plane_id i9xx_plane = plane->i9xx_plane; > > intel_wakeref_t wakeref; > > @@ -3862,7 +3862,7 @@ static bool i9xx_plane_get_hw_state(struct intel_plane *plane, > > * display power wells. > > */ > > power_domain = POWER_DOMAIN_PIPE(plane->pipe); > > - wakeref = intel_display_power_get_if_enabled(dev_priv, power_domain); > > + wakeref = intel_display_power_get_if_enabled(i915, power_domain); > > if (!wakeref) > > return false; > > @@ -3870,13 +3870,13 @@ static bool i9xx_plane_get_hw_state(struct intel_plane *plane, > > ret = val & DISPLAY_PLANE_ENABLE; > > - if (INTEL_GEN(dev_priv) >= 5) > > + if (INTEL_GEN(i915) >= 5) > > *pipe = plane->pipe; > > else > > *pipe = (val & DISPPLANE_SEL_PIPE_MASK) >> > > DISPPLANE_SEL_PIPE_SHIFT; > > - intel_display_power_put(dev_priv, power_domain, wakeref); > > + intel_display_power_put(i915, power_domain, wakeref); > > return ret; > > } > > @@ -3884,7 +3884,7 @@ static bool i9xx_plane_get_hw_state(struct intel_plane *plane, > > static void skl_detach_scaler(struct intel_crtc *intel_crtc, int id) > > { > > struct drm_device *dev = intel_crtc->base.dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > I915_WRITE(SKL_PS_CTRL(intel_crtc->pipe, id), 0); > > I915_WRITE(SKL_PS_WIN_POS(intel_crtc->pipe, id), 0); > > @@ -4091,10 +4091,10 @@ static u32 cnl_plane_ctl_flip(unsigned int reflect) > > u32 skl_plane_ctl_crtc(const struct intel_crtc_state *crtc_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc_state->base.crtc->dev); > > + struct drm_i915_private *i915 = to_i915(crtc_state->base.crtc->dev); > > u32 plane_ctl = 0; > > - if (INTEL_GEN(dev_priv) >= 10 || IS_GEMINILAKE(dev_priv)) > > + if (INTEL_GEN(i915) >= 10 || IS_GEMINILAKE(i915)) > > return plane_ctl; > > if (crtc_state->gamma_enable) > > @@ -4109,7 +4109,7 @@ u32 skl_plane_ctl_crtc(const struct intel_crtc_state *crtc_state) > > u32 skl_plane_ctl(const struct intel_crtc_state *crtc_state, > > const struct intel_plane_state *plane_state) > > { > > - struct drm_i915_private *dev_priv = > > + struct drm_i915_private *i915 = > > to_i915(plane_state->base.plane->dev); > > const struct drm_framebuffer *fb = plane_state->base.fb; > > unsigned int rotation = plane_state->base.rotation; > > @@ -4118,7 +4118,7 @@ u32 skl_plane_ctl(const struct intel_crtc_state *crtc_state, > > plane_ctl = PLANE_CTL_ENABLE; > > - if (INTEL_GEN(dev_priv) < 10 && !IS_GEMINILAKE(dev_priv)) { > > + if (INTEL_GEN(i915) < 10 && !IS_GEMINILAKE(i915)) { > > plane_ctl |= skl_plane_ctl_alpha(plane_state); > > plane_ctl |= PLANE_CTL_PLANE_GAMMA_DISABLE; > > @@ -4133,7 +4133,7 @@ u32 skl_plane_ctl(const struct intel_crtc_state *crtc_state, > > plane_ctl |= skl_plane_ctl_tiling(fb->modifier); > > plane_ctl |= skl_plane_ctl_rotate(rotation & DRM_MODE_ROTATE_MASK); > > - if (INTEL_GEN(dev_priv) >= 10) > > + if (INTEL_GEN(i915) >= 10) > > plane_ctl |= cnl_plane_ctl_flip(rotation & > > DRM_MODE_REFLECT_MASK); > > @@ -4147,10 +4147,10 @@ u32 skl_plane_ctl(const struct intel_crtc_state *crtc_state, > > u32 glk_plane_color_ctl_crtc(const struct intel_crtc_state *crtc_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc_state->base.crtc->dev); > > + struct drm_i915_private *i915 = to_i915(crtc_state->base.crtc->dev); > > u32 plane_color_ctl = 0; > > - if (INTEL_GEN(dev_priv) >= 11) > > + if (INTEL_GEN(i915) >= 11) > > return plane_color_ctl; > > if (crtc_state->gamma_enable) > > @@ -4165,7 +4165,7 @@ u32 glk_plane_color_ctl_crtc(const struct intel_crtc_state *crtc_state) > > u32 glk_plane_color_ctl(const struct intel_crtc_state *crtc_state, > > const struct intel_plane_state *plane_state) > > { > > - struct drm_i915_private *dev_priv = > > + struct drm_i915_private *i915 = > > to_i915(plane_state->base.plane->dev); > > const struct drm_framebuffer *fb = plane_state->base.fb; > > struct intel_plane *plane = to_intel_plane(plane_state->base.plane); > > @@ -4174,7 +4174,7 @@ u32 glk_plane_color_ctl(const struct intel_crtc_state *crtc_state, > > plane_color_ctl |= PLANE_COLOR_PLANE_GAMMA_DISABLE; > > plane_color_ctl |= glk_plane_color_ctl_alpha(plane_state); > > - if (fb->format->is_yuv && !icl_is_hdr_plane(dev_priv, plane->id)) { > > + if (fb->format->is_yuv && !icl_is_hdr_plane(i915, plane->id)) { > > if (plane_state->base.color_encoding == DRM_COLOR_YCBCR_BT709) > > plane_color_ctl |= PLANE_COLOR_CSC_MODE_YUV709_TO_RGB709; > > else > > @@ -4228,31 +4228,31 @@ __intel_display_resume(struct drm_device *dev, > > return ret; > > } > > -static bool gpu_reset_clobbers_display(struct drm_i915_private *dev_priv) > > +static bool gpu_reset_clobbers_display(struct drm_i915_private *i915) > > { > > - return (INTEL_INFO(dev_priv)->gpu_reset_clobbers_display && > > - intel_has_gpu_reset(dev_priv)); > > + return (INTEL_INFO(i915)->gpu_reset_clobbers_display && > > + intel_has_gpu_reset(i915)); > > } > > -void intel_prepare_reset(struct drm_i915_private *dev_priv) > > +void intel_prepare_reset(struct drm_i915_private *i915) > > { > > - struct drm_device *dev = &dev_priv->drm; > > - struct drm_modeset_acquire_ctx *ctx = &dev_priv->reset_ctx; > > + struct drm_device *dev = &i915->drm; > > + struct drm_modeset_acquire_ctx *ctx = &i915->reset_ctx; > > struct drm_atomic_state *state; > > int ret; > > /* reset doesn't touch the display */ > > if (!i915_modparams.force_reset_modeset_test && > > - !gpu_reset_clobbers_display(dev_priv)) > > + !gpu_reset_clobbers_display(i915)) > > return; > > /* We have a modeset vs reset deadlock, defensively unbreak it. */ > > - set_bit(I915_RESET_MODESET, &dev_priv->gpu_error.flags); > > - wake_up_all(&dev_priv->gpu_error.wait_queue); > > + set_bit(I915_RESET_MODESET, &i915->gpu_error.flags); > > + wake_up_all(&i915->gpu_error.wait_queue); > > - if (atomic_read(&dev_priv->gpu_error.pending_fb_pin)) { > > + if (atomic_read(&i915->gpu_error.pending_fb_pin)) { > > DRM_DEBUG_KMS("Modeset potentially stuck, unbreaking through wedging\n"); > > - i915_gem_set_wedged(dev_priv); > > + i915_gem_set_wedged(i915); > > } > > /* > > @@ -4286,27 +4286,27 @@ void intel_prepare_reset(struct drm_i915_private *dev_priv) > > return; > > } > > - dev_priv->modeset_restore_state = state; > > + i915->modeset_restore_state = state; > > state->acquire_ctx = ctx; > > } > > -void intel_finish_reset(struct drm_i915_private *dev_priv) > > +void intel_finish_reset(struct drm_i915_private *i915) > > { > > - struct drm_device *dev = &dev_priv->drm; > > - struct drm_modeset_acquire_ctx *ctx = &dev_priv->reset_ctx; > > + struct drm_device *dev = &i915->drm; > > + struct drm_modeset_acquire_ctx *ctx = &i915->reset_ctx; > > struct drm_atomic_state *state; > > int ret; > > /* reset doesn't touch the display */ > > - if (!test_bit(I915_RESET_MODESET, &dev_priv->gpu_error.flags)) > > + if (!test_bit(I915_RESET_MODESET, &i915->gpu_error.flags)) > > return; > > - state = fetch_and_zero(&dev_priv->modeset_restore_state); > > + state = fetch_and_zero(&i915->modeset_restore_state); > > if (!state) > > goto unlock; > > /* reset doesn't touch the display */ > > - if (!gpu_reset_clobbers_display(dev_priv)) { > > + if (!gpu_reset_clobbers_display(i915)) { > > /* for testing only restore the display */ > > ret = __intel_display_resume(dev, state, ctx); > > if (ret) > > @@ -4316,20 +4316,20 @@ void intel_finish_reset(struct drm_i915_private *dev_priv) > > * The display has been reset as well, > > * so need a full re-initialization. > > */ > > - intel_pps_unlock_regs_wa(dev_priv); > > + intel_pps_unlock_regs_wa(i915); > > intel_modeset_init_hw(dev); > > - intel_init_clock_gating(dev_priv); > > + intel_init_clock_gating(i915); > > - spin_lock_irq(&dev_priv->irq_lock); > > - if (dev_priv->display.hpd_irq_setup) > > - dev_priv->display.hpd_irq_setup(dev_priv); > > - spin_unlock_irq(&dev_priv->irq_lock); > > + spin_lock_irq(&i915->irq_lock); > > + if (i915->display.hpd_irq_setup) > > + i915->display.hpd_irq_setup(i915); > > + spin_unlock_irq(&i915->irq_lock); > > ret = __intel_display_resume(dev, state, ctx); > > if (ret) > > DRM_ERROR("Restoring old state failed with %i\n", ret); > > - intel_hpd_init(dev_priv); > > + intel_hpd_init(i915); > > } > > drm_atomic_state_put(state); > > @@ -4338,12 +4338,12 @@ void intel_finish_reset(struct drm_i915_private *dev_priv) > > drm_modeset_acquire_fini(ctx); > > mutex_unlock(&dev->mode_config.mutex); > > - clear_bit(I915_RESET_MODESET, &dev_priv->gpu_error.flags); > > + clear_bit(I915_RESET_MODESET, &i915->gpu_error.flags); > > } > > static void icl_set_pipe_chicken(struct intel_crtc *crtc) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > enum pipe pipe = crtc->pipe; > > u32 tmp; > > @@ -4369,7 +4369,7 @@ static void intel_update_pipe_config(const struct intel_crtc_state *old_crtc_sta > > const struct intel_crtc_state *new_crtc_state) > > { > > struct intel_crtc *crtc = to_intel_crtc(new_crtc_state->base.crtc); > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > /* drm_atomic_helper_update_legacy_modeset_state might not be called. */ > > crtc->base.mode = new_crtc_state->base.mode; > > @@ -4388,26 +4388,26 @@ static void intel_update_pipe_config(const struct intel_crtc_state *old_crtc_sta > > (new_crtc_state->pipe_src_h - 1)); > > /* on skylake this is done by detaching scalers */ > > - if (INTEL_GEN(dev_priv) >= 9) { > > + if (INTEL_GEN(i915) >= 9) { > > skl_detach_scalers(new_crtc_state); > > if (new_crtc_state->pch_pfit.enabled) > > skylake_pfit_enable(new_crtc_state); > > - } else if (HAS_PCH_SPLIT(dev_priv)) { > > + } else if (HAS_PCH_SPLIT(i915)) { > > if (new_crtc_state->pch_pfit.enabled) > > ironlake_pfit_enable(new_crtc_state); > > else if (old_crtc_state->pch_pfit.enabled) > > ironlake_pfit_disable(old_crtc_state); > > } > > - if (INTEL_GEN(dev_priv) >= 11) > > + if (INTEL_GEN(i915) >= 11) > > icl_set_pipe_chicken(crtc); > > } > > static void intel_fdi_normal_train(struct intel_crtc *crtc) > > { > > struct drm_device *dev = crtc->base.dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > int pipe = crtc->pipe; > > i915_reg_t reg; > > u32 temp; > > @@ -4415,7 +4415,7 @@ static void intel_fdi_normal_train(struct intel_crtc *crtc) > > /* enable normal train */ > > reg = FDI_TX_CTL(pipe); > > temp = I915_READ(reg); > > - if (IS_IVYBRIDGE(dev_priv)) { > > + if (IS_IVYBRIDGE(i915)) { > > temp &= ~FDI_LINK_TRAIN_NONE_IVB; > > temp |= FDI_LINK_TRAIN_NONE_IVB | FDI_TX_ENHANCE_FRAME_ENABLE; > > } else { > > @@ -4426,7 +4426,7 @@ static void intel_fdi_normal_train(struct intel_crtc *crtc) > > reg = FDI_RX_CTL(pipe); > > temp = I915_READ(reg); > > - if (HAS_PCH_CPT(dev_priv)) { > > + if (HAS_PCH_CPT(i915)) { > > temp &= ~FDI_LINK_TRAIN_PATTERN_MASK_CPT; > > temp |= FDI_LINK_TRAIN_NORMAL_CPT; > > } else { > > @@ -4440,7 +4440,7 @@ static void intel_fdi_normal_train(struct intel_crtc *crtc) > > udelay(1000); > > /* IVB wants error correction enabled */ > > - if (IS_IVYBRIDGE(dev_priv)) > > + if (IS_IVYBRIDGE(i915)) > > I915_WRITE(reg, I915_READ(reg) | FDI_FS_ERRC_ENABLE | > > FDI_FE_ERRC_ENABLE); > > } > > @@ -4450,13 +4450,13 @@ static void ironlake_fdi_link_train(struct intel_crtc *crtc, > > const struct intel_crtc_state *crtc_state) > > { > > struct drm_device *dev = crtc->base.dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > int pipe = crtc->pipe; > > i915_reg_t reg; > > u32 temp, tries; > > /* FDI needs bits from pipe first */ > > - assert_pipe_enabled(dev_priv, pipe); > > + assert_pipe_enabled(i915, pipe); > > /* Train 1: umask FDI RX Interrupt symbol_lock and bit_lock bit > > for train result */ > > @@ -4551,7 +4551,7 @@ static void gen6_fdi_link_train(struct intel_crtc *crtc, > > const struct intel_crtc_state *crtc_state) > > { > > struct drm_device *dev = crtc->base.dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > int pipe = crtc->pipe; > > i915_reg_t reg; > > u32 temp, i, retry; > > @@ -4584,7 +4584,7 @@ static void gen6_fdi_link_train(struct intel_crtc *crtc, > > reg = FDI_RX_CTL(pipe); > > temp = I915_READ(reg); > > - if (HAS_PCH_CPT(dev_priv)) { > > + if (HAS_PCH_CPT(i915)) { > > temp &= ~FDI_LINK_TRAIN_PATTERN_MASK_CPT; > > temp |= FDI_LINK_TRAIN_PATTERN_1_CPT; > > } else { > > @@ -4628,7 +4628,7 @@ static void gen6_fdi_link_train(struct intel_crtc *crtc, > > temp = I915_READ(reg); > > temp &= ~FDI_LINK_TRAIN_NONE; > > temp |= FDI_LINK_TRAIN_PATTERN_2; > > - if (IS_GEN(dev_priv, 6)) { > > + if (IS_GEN(i915, 6)) { > > temp &= ~FDI_LINK_TRAIN_VOL_EMP_MASK; > > /* SNB-B */ > > temp |= FDI_LINK_TRAIN_400MV_0DB_SNB_B; > > @@ -4637,7 +4637,7 @@ static void gen6_fdi_link_train(struct intel_crtc *crtc, > > reg = FDI_RX_CTL(pipe); > > temp = I915_READ(reg); > > - if (HAS_PCH_CPT(dev_priv)) { > > + if (HAS_PCH_CPT(i915)) { > > temp &= ~FDI_LINK_TRAIN_PATTERN_MASK_CPT; > > temp |= FDI_LINK_TRAIN_PATTERN_2_CPT; > > } else { > > @@ -4684,7 +4684,7 @@ static void ivb_manual_fdi_link_train(struct intel_crtc *crtc, > > const struct intel_crtc_state *crtc_state) > > { > > struct drm_device *dev = crtc->base.dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > int pipe = crtc->pipe; > > i915_reg_t reg; > > u32 temp, i, j; > > @@ -4802,7 +4802,7 @@ static void ivb_manual_fdi_link_train(struct intel_crtc *crtc, > > static void ironlake_fdi_pll_enable(const struct intel_crtc_state *crtc_state) > > { > > struct intel_crtc *intel_crtc = to_intel_crtc(crtc_state->base.crtc); > > - struct drm_i915_private *dev_priv = to_i915(intel_crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(intel_crtc->base.dev); > > int pipe = intel_crtc->pipe; > > i915_reg_t reg; > > u32 temp; > > @@ -4839,7 +4839,7 @@ static void ironlake_fdi_pll_enable(const struct intel_crtc_state *crtc_state) > > static void ironlake_fdi_pll_disable(struct intel_crtc *intel_crtc) > > { > > struct drm_device *dev = intel_crtc->base.dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > int pipe = intel_crtc->pipe; > > i915_reg_t reg; > > u32 temp; > > @@ -4869,7 +4869,7 @@ static void ironlake_fdi_pll_disable(struct intel_crtc *intel_crtc) > > static void ironlake_fdi_disable(struct drm_crtc *crtc) > > { > > struct drm_device *dev = crtc->dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > struct intel_crtc *intel_crtc = to_intel_crtc(crtc); > > int pipe = intel_crtc->pipe; > > i915_reg_t reg; > > @@ -4891,7 +4891,7 @@ static void ironlake_fdi_disable(struct drm_crtc *crtc) > > udelay(100); > > /* Ironlake workaround, disable clock pointer after downing FDI */ > > - if (HAS_PCH_IBX(dev_priv)) > > + if (HAS_PCH_IBX(i915)) > > I915_WRITE(FDI_RX_CHICKEN(pipe), FDI_RX_PHASE_SYNC_POINTER_OVR); > > /* still set train pattern 1 */ > > @@ -4903,7 +4903,7 @@ static void ironlake_fdi_disable(struct drm_crtc *crtc) > > reg = FDI_RX_CTL(pipe); > > temp = I915_READ(reg); > > - if (HAS_PCH_CPT(dev_priv)) { > > + if (HAS_PCH_CPT(i915)) { > > temp &= ~FDI_LINK_TRAIN_PATTERN_MASK_CPT; > > temp |= FDI_LINK_TRAIN_PATTERN_1_CPT; > > } else { > > @@ -4919,12 +4919,12 @@ static void ironlake_fdi_disable(struct drm_crtc *crtc) > > udelay(100); > > } > > -bool intel_has_pending_fb_unpin(struct drm_i915_private *dev_priv) > > +bool intel_has_pending_fb_unpin(struct drm_i915_private *i915) > > { > > struct drm_crtc *crtc; > > bool cleanup_done; > > - drm_for_each_crtc(crtc, &dev_priv->drm) { > > + drm_for_each_crtc(crtc, &i915->drm) { > > struct drm_crtc_commit *commit; > > spin_lock(&crtc->commit_lock); > > commit = list_first_entry_or_null(&crtc->commit_list, > > @@ -4944,31 +4944,31 @@ bool intel_has_pending_fb_unpin(struct drm_i915_private *dev_priv) > > return false; > > } > > -void lpt_disable_iclkip(struct drm_i915_private *dev_priv) > > +void lpt_disable_iclkip(struct drm_i915_private *i915) > > { > > u32 temp; > > I915_WRITE(PIXCLK_GATE, PIXCLK_GATE_GATE); > > - mutex_lock(&dev_priv->sb_lock); > > + mutex_lock(&i915->sb_lock); > > - temp = intel_sbi_read(dev_priv, SBI_SSCCTL6, SBI_ICLK); > > + temp = intel_sbi_read(i915, SBI_SSCCTL6, SBI_ICLK); > > temp |= SBI_SSCCTL_DISABLE; > > - intel_sbi_write(dev_priv, SBI_SSCCTL6, temp, SBI_ICLK); > > + intel_sbi_write(i915, SBI_SSCCTL6, temp, SBI_ICLK); > > - mutex_unlock(&dev_priv->sb_lock); > > + mutex_unlock(&i915->sb_lock); > > } > > /* Program iCLKIP clock to the desired frequency */ > > static void lpt_program_iclkip(const struct intel_crtc_state *crtc_state) > > { > > struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc); > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > int clock = crtc_state->base.adjusted_mode.crtc_clock; > > u32 divsel, phaseinc, auxdiv, phasedir = 0; > > u32 temp; > > - lpt_disable_iclkip(dev_priv); > > + lpt_disable_iclkip(i915); > > /* The iCLK virtual clock root frequency is in MHz, > > * but the adjusted_mode->crtc_clock in in KHz. To get the > > @@ -5007,30 +5007,30 @@ static void lpt_program_iclkip(const struct intel_crtc_state *crtc_state) > > phasedir, > > phaseinc); > > - mutex_lock(&dev_priv->sb_lock); > > + mutex_lock(&i915->sb_lock); > > /* Program SSCDIVINTPHASE6 */ > > - temp = intel_sbi_read(dev_priv, SBI_SSCDIVINTPHASE6, SBI_ICLK); > > + temp = intel_sbi_read(i915, SBI_SSCDIVINTPHASE6, SBI_ICLK); > > temp &= ~SBI_SSCDIVINTPHASE_DIVSEL_MASK; > > temp |= SBI_SSCDIVINTPHASE_DIVSEL(divsel); > > temp &= ~SBI_SSCDIVINTPHASE_INCVAL_MASK; > > temp |= SBI_SSCDIVINTPHASE_INCVAL(phaseinc); > > temp |= SBI_SSCDIVINTPHASE_DIR(phasedir); > > temp |= SBI_SSCDIVINTPHASE_PROPAGATE; > > - intel_sbi_write(dev_priv, SBI_SSCDIVINTPHASE6, temp, SBI_ICLK); > > + intel_sbi_write(i915, SBI_SSCDIVINTPHASE6, temp, SBI_ICLK); > > /* Program SSCAUXDIV */ > > - temp = intel_sbi_read(dev_priv, SBI_SSCAUXDIV6, SBI_ICLK); > > + temp = intel_sbi_read(i915, SBI_SSCAUXDIV6, SBI_ICLK); > > temp &= ~SBI_SSCAUXDIV_FINALDIV2SEL(1); > > temp |= SBI_SSCAUXDIV_FINALDIV2SEL(auxdiv); > > - intel_sbi_write(dev_priv, SBI_SSCAUXDIV6, temp, SBI_ICLK); > > + intel_sbi_write(i915, SBI_SSCAUXDIV6, temp, SBI_ICLK); > > /* Enable modulator and associated divider */ > > - temp = intel_sbi_read(dev_priv, SBI_SSCCTL6, SBI_ICLK); > > + temp = intel_sbi_read(i915, SBI_SSCCTL6, SBI_ICLK); > > temp &= ~SBI_SSCCTL_DISABLE; > > - intel_sbi_write(dev_priv, SBI_SSCCTL6, temp, SBI_ICLK); > > + intel_sbi_write(i915, SBI_SSCCTL6, temp, SBI_ICLK); > > - mutex_unlock(&dev_priv->sb_lock); > > + mutex_unlock(&i915->sb_lock); > > /* Wait for initialization time */ > > udelay(24); > > @@ -5038,7 +5038,7 @@ static void lpt_program_iclkip(const struct intel_crtc_state *crtc_state) > > I915_WRITE(PIXCLK_GATE, PIXCLK_GATE_UNGATE); > > } > > -int lpt_get_iclkip(struct drm_i915_private *dev_priv) > > +int lpt_get_iclkip(struct drm_i915_private *i915) > > { > > u32 divsel, phaseinc, auxdiv; > > u32 iclk_virtual_root_freq = 172800 * 1000; > > @@ -5049,25 +5049,25 @@ int lpt_get_iclkip(struct drm_i915_private *dev_priv) > > if ((I915_READ(PIXCLK_GATE) & PIXCLK_GATE_UNGATE) == 0) > > return 0; > > - mutex_lock(&dev_priv->sb_lock); > > + mutex_lock(&i915->sb_lock); > > - temp = intel_sbi_read(dev_priv, SBI_SSCCTL6, SBI_ICLK); > > + temp = intel_sbi_read(i915, SBI_SSCCTL6, SBI_ICLK); > > if (temp & SBI_SSCCTL_DISABLE) { > > - mutex_unlock(&dev_priv->sb_lock); > > + mutex_unlock(&i915->sb_lock); > > return 0; > > } > > - temp = intel_sbi_read(dev_priv, SBI_SSCDIVINTPHASE6, SBI_ICLK); > > + temp = intel_sbi_read(i915, SBI_SSCDIVINTPHASE6, SBI_ICLK); > > divsel = (temp & SBI_SSCDIVINTPHASE_DIVSEL_MASK) >> > > SBI_SSCDIVINTPHASE_DIVSEL_SHIFT; > > phaseinc = (temp & SBI_SSCDIVINTPHASE_INCVAL_MASK) >> > > SBI_SSCDIVINTPHASE_INCVAL_SHIFT; > > - temp = intel_sbi_read(dev_priv, SBI_SSCAUXDIV6, SBI_ICLK); > > + temp = intel_sbi_read(i915, SBI_SSCAUXDIV6, SBI_ICLK); > > auxdiv = (temp & SBI_SSCAUXDIV_FINALDIV2SEL_MASK) >> > > SBI_SSCAUXDIV_FINALDIV2SEL_SHIFT; > > - mutex_unlock(&dev_priv->sb_lock); > > + mutex_unlock(&i915->sb_lock); > > desired_divisor = (divsel + 2) * iclk_pi_range + phaseinc; > > @@ -5079,7 +5079,7 @@ static void ironlake_pch_transcoder_set_timings(const struct intel_crtc_state *c > > enum pipe pch_transcoder) > > { > > struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc); > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > enum transcoder cpu_transcoder = crtc_state->cpu_transcoder; > > I915_WRITE(PCH_TRANS_HTOTAL(pch_transcoder), > > @@ -5099,7 +5099,7 @@ static void ironlake_pch_transcoder_set_timings(const struct intel_crtc_state *c > > I915_READ(VSYNCSHIFT(cpu_transcoder))); > > } > > -static void cpt_set_fdi_bc_bifurcation(struct drm_i915_private *dev_priv, bool enable) > > +static void cpt_set_fdi_bc_bifurcation(struct drm_i915_private *i915, bool enable) > > { > > u32 temp; > > @@ -5122,20 +5122,20 @@ static void cpt_set_fdi_bc_bifurcation(struct drm_i915_private *dev_priv, bool e > > static void ivybridge_update_fdi_bc_bifurcation(const struct intel_crtc_state *crtc_state) > > { > > struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc); > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > switch (crtc->pipe) { > > case PIPE_A: > > break; > > case PIPE_B: > > if (crtc_state->fdi_lanes > 2) > > - cpt_set_fdi_bc_bifurcation(dev_priv, false); > > + cpt_set_fdi_bc_bifurcation(i915, false); > > else > > - cpt_set_fdi_bc_bifurcation(dev_priv, true); > > + cpt_set_fdi_bc_bifurcation(i915, true); > > break; > > case PIPE_C: > > - cpt_set_fdi_bc_bifurcation(dev_priv, true); > > + cpt_set_fdi_bc_bifurcation(i915, true); > > break; > > default: > > @@ -5185,13 +5185,13 @@ static void ironlake_pch_enable(const struct intel_atomic_state *state, > > { > > struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc); > > struct drm_device *dev = crtc->base.dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > int pipe = crtc->pipe; > > u32 temp; > > - assert_pch_transcoder_disabled(dev_priv, pipe); > > + assert_pch_transcoder_disabled(i915, pipe); > > - if (IS_IVYBRIDGE(dev_priv)) > > + if (IS_IVYBRIDGE(i915)) > > ivybridge_update_fdi_bc_bifurcation(crtc_state); > > /* Write the TU size bits before fdi link training, so that error > > @@ -5200,18 +5200,18 @@ static void ironlake_pch_enable(const struct intel_atomic_state *state, > > I915_READ(PIPE_DATA_M1(pipe)) & TU_SIZE_MASK); > > /* For PCH output, training FDI link */ > > - dev_priv->display.fdi_link_train(crtc, crtc_state); > > + i915->display.fdi_link_train(crtc, crtc_state); > > /* We need to program the right clock selection before writing the pixel > > * mutliplier into the DPLL. */ > > - if (HAS_PCH_CPT(dev_priv)) { > > + if (HAS_PCH_CPT(i915)) { > > u32 sel; > > temp = I915_READ(PCH_DPLL_SEL); > > temp |= TRANS_DPLL_ENABLE(pipe); > > sel = TRANS_DPLLB_SEL(pipe); > > if (crtc_state->shared_dpll == > > - intel_get_shared_dpll_by_id(dev_priv, DPLL_ID_PCH_PLL_B)) > > + intel_get_shared_dpll_by_id(i915, DPLL_ID_PCH_PLL_B)) > > temp |= sel; > > else > > temp &= ~sel; > > @@ -5228,13 +5228,13 @@ static void ironlake_pch_enable(const struct intel_atomic_state *state, > > intel_enable_shared_dpll(crtc_state); > > /* set transcoder timing, panel must allow it */ > > - assert_panel_unlocked(dev_priv, pipe); > > + assert_panel_unlocked(i915, pipe); > > ironlake_pch_transcoder_set_timings(crtc_state, pipe); > > intel_fdi_normal_train(crtc); > > /* For PCH DP, enable TRANS_DP_CTL */ > > - if (HAS_PCH_CPT(dev_priv) && > > + if (HAS_PCH_CPT(i915) && > > intel_crtc_has_dp_encoder(crtc_state)) { > > const struct drm_display_mode *adjusted_mode = > > &crtc_state->base.adjusted_mode; > > @@ -5268,22 +5268,22 @@ static void lpt_pch_enable(const struct intel_atomic_state *state, > > const struct intel_crtc_state *crtc_state) > > { > > struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc); > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > enum transcoder cpu_transcoder = crtc_state->cpu_transcoder; > > - assert_pch_transcoder_disabled(dev_priv, PIPE_A); > > + assert_pch_transcoder_disabled(i915, PIPE_A); > > lpt_program_iclkip(crtc_state); > > /* Set transcoder timing. */ > > ironlake_pch_transcoder_set_timings(crtc_state, PIPE_A); > > - lpt_enable_pch_transcoder(dev_priv, cpu_transcoder); > > + lpt_enable_pch_transcoder(i915, cpu_transcoder); > > } > > static void cpt_verify_modeset(struct drm_device *dev, int pipe) > > { > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > i915_reg_t dslreg = PIPEDSL(pipe); > > u32 temp; > > @@ -5385,7 +5385,7 @@ skl_update_scaler(struct intel_crtc_state *crtc_state, bool force_detach, > > &crtc_state->scaler_state; > > struct intel_crtc *intel_crtc = > > to_intel_crtc(crtc_state->base.crtc); > > - struct drm_i915_private *dev_priv = to_i915(intel_crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(intel_crtc->base.dev); > > const struct drm_display_mode *adjusted_mode = > > &crtc_state->base.adjusted_mode; > > @@ -5403,7 +5403,7 @@ skl_update_scaler(struct intel_crtc_state *crtc_state, bool force_detach, > > * Once NV12 is enabled, handle it here while allocating scaler > > * for NV12. > > */ > > - if (INTEL_GEN(dev_priv) >= 9 && crtc_state->base.enable && > > + if (INTEL_GEN(i915) >= 9 && crtc_state->base.enable && > > need_scaler && adjusted_mode->flags & DRM_MODE_FLAG_INTERLACE) { > > DRM_DEBUG_KMS("Pipe/Plane scaling not supported with IF-ID mode\n"); > > return -EINVAL; > > @@ -5442,10 +5442,10 @@ skl_update_scaler(struct intel_crtc_state *crtc_state, bool force_detach, > > /* range checks */ > > if (src_w < SKL_MIN_SRC_W || src_h < SKL_MIN_SRC_H || > > dst_w < SKL_MIN_DST_W || dst_h < SKL_MIN_DST_H || > > - (INTEL_GEN(dev_priv) >= 11 && > > + (INTEL_GEN(i915) >= 11 && > > (src_w > ICL_MAX_SRC_W || src_h > ICL_MAX_SRC_H || > > dst_w > ICL_MAX_DST_W || dst_h > ICL_MAX_DST_H)) || > > - (INTEL_GEN(dev_priv) < 11 && > > + (INTEL_GEN(i915) < 11 && > > (src_w > SKL_MAX_SRC_W || src_h > SKL_MAX_SRC_H || > > dst_w > SKL_MAX_DST_W || dst_h > SKL_MAX_DST_H))) { > > DRM_DEBUG_KMS("scaler_user index %u.%u: src %ux%u dst %ux%u " > > @@ -5502,14 +5502,14 @@ static int skl_update_scaler_plane(struct intel_crtc_state *crtc_state, > > { > > struct intel_plane *intel_plane = > > to_intel_plane(plane_state->base.plane); > > - struct drm_i915_private *dev_priv = to_i915(intel_plane->base.dev); > > + struct drm_i915_private *i915 = to_i915(intel_plane->base.dev); > > struct drm_framebuffer *fb = plane_state->base.fb; > > int ret; > > bool force_detach = !fb || !plane_state->base.visible; > > bool need_scaler = false; > > /* Pre-gen11 and SDR planes always need a scaler for planar formats. */ > > - if (!icl_is_hdr_plane(dev_priv, intel_plane->id) && > > + if (!icl_is_hdr_plane(i915, intel_plane->id) && > > fb && is_planar_yuv_format(fb->format->format)) > > need_scaler = true; > > @@ -5582,7 +5582,7 @@ static void skylake_scaler_disable(struct intel_crtc *crtc) > > static void skylake_pfit_enable(const struct intel_crtc_state *crtc_state) > > { > > struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc); > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > enum pipe pipe = crtc->pipe; > > const struct intel_crtc_scaler_state *scaler_state = > > &crtc_state->scaler_state; > > @@ -5619,7 +5619,7 @@ static void skylake_pfit_enable(const struct intel_crtc_state *crtc_state) > > static void ironlake_pfit_enable(const struct intel_crtc_state *crtc_state) > > { > > struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc); > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > int pipe = crtc->pipe; > > if (crtc_state->pch_pfit.enabled) { > > @@ -5627,7 +5627,7 @@ static void ironlake_pfit_enable(const struct intel_crtc_state *crtc_state) > > * as some pre-programmed values are broken, > > * e.g. x201. > > */ > > - if (IS_IVYBRIDGE(dev_priv) || IS_HASWELL(dev_priv)) > > + if (IS_IVYBRIDGE(i915) || IS_HASWELL(i915)) > > I915_WRITE(PF_CTL(pipe), PF_ENABLE | PF_FILTER_MED_3x3 | > > PF_PIPE_SEL_IVB(pipe)); > > else > > @@ -5641,7 +5641,7 @@ void hsw_enable_ips(const struct intel_crtc_state *crtc_state) > > { > > struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc); > > struct drm_device *dev = crtc->base.dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > if (!crtc_state->ips_enabled) > > return; > > @@ -5653,8 +5653,8 @@ void hsw_enable_ips(const struct intel_crtc_state *crtc_state) > > */ > > WARN_ON(!(crtc_state->active_planes & ~BIT(PLANE_CURSOR))); > > - if (IS_BROADWELL(dev_priv)) { > > - WARN_ON(sandybridge_pcode_write(dev_priv, DISPLAY_IPS_CONTROL, > > + if (IS_BROADWELL(i915)) { > > + WARN_ON(sandybridge_pcode_write(i915, DISPLAY_IPS_CONTROL, > > IPS_ENABLE | IPS_PCODE_CONTROL)); > > /* Quoting Art Runyan: "its not safe to expect any particular > > * value in IPS_CTL bit 31 after enabling IPS through the > > @@ -5668,7 +5668,7 @@ void hsw_enable_ips(const struct intel_crtc_state *crtc_state) > > * and don't wait for vblanks until the end of crtc_enable, then > > * the HW state readout code will complain that the expected > > * IPS_CTL value is not the one we read. */ > > - if (intel_wait_for_register(&dev_priv->uncore, > > + if (intel_wait_for_register(&i915->uncore, > > IPS_CTL, IPS_ENABLE, IPS_ENABLE, > > 50)) > > DRM_ERROR("Timed out waiting for IPS enable\n"); > > @@ -5679,19 +5679,19 @@ void hsw_disable_ips(const struct intel_crtc_state *crtc_state) > > { > > struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc); > > struct drm_device *dev = crtc->base.dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > if (!crtc_state->ips_enabled) > > return; > > - if (IS_BROADWELL(dev_priv)) { > > - WARN_ON(sandybridge_pcode_write(dev_priv, DISPLAY_IPS_CONTROL, 0)); > > + if (IS_BROADWELL(i915)) { > > + WARN_ON(sandybridge_pcode_write(i915, DISPLAY_IPS_CONTROL, 0)); > > /* > > * Wait for PCODE to finish disabling IPS. The BSpec specified > > * 42ms timeout value leads to occasional timeouts so use 100ms > > * instead. > > */ > > - if (intel_wait_for_register(&dev_priv->uncore, > > + if (intel_wait_for_register(&i915->uncore, > > IPS_CTL, IPS_ENABLE, 0, > > 100)) > > DRM_ERROR("Timed out waiting for IPS disable\n"); > > @@ -5701,7 +5701,7 @@ void hsw_disable_ips(const struct intel_crtc_state *crtc_state) > > } > > /* We need to wait for a vblank before we can disable the plane. */ > > - intel_wait_for_vblank(dev_priv, crtc->pipe); > > + intel_wait_for_vblank(i915, crtc->pipe); > > } > > static void intel_crtc_dpms_overlay_disable(struct intel_crtc *intel_crtc) > > @@ -5735,7 +5735,7 @@ intel_post_enable_primary(struct drm_crtc *crtc, > > const struct intel_crtc_state *new_crtc_state) > > { > > struct drm_device *dev = crtc->dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > struct intel_crtc *intel_crtc = to_intel_crtc(crtc); > > int pipe = intel_crtc->pipe; > > @@ -5746,12 +5746,12 @@ intel_post_enable_primary(struct drm_crtc *crtc, > > * FIXME: Need to fix the logic to work when we turn off all planes > > * but leave the pipe running. > > */ > > - if (IS_GEN(dev_priv, 2)) > > - intel_set_cpu_fifo_underrun_reporting(dev_priv, pipe, true); > > + if (IS_GEN(i915, 2)) > > + intel_set_cpu_fifo_underrun_reporting(i915, pipe, true); > > /* Underruns don't always raise interrupts, so check manually. */ > > - intel_check_cpu_fifo_underruns(dev_priv); > > - intel_check_pch_fifo_underruns(dev_priv); > > + intel_check_cpu_fifo_underruns(i915); > > + intel_check_pch_fifo_underruns(i915); > > } > > /* FIXME get rid of this and use pre_plane_update */ > > @@ -5759,7 +5759,7 @@ static void > > intel_pre_disable_primary_noatomic(struct drm_crtc *crtc) > > { > > struct drm_device *dev = crtc->dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > struct intel_crtc *intel_crtc = to_intel_crtc(crtc); > > int pipe = intel_crtc->pipe; > > @@ -5767,8 +5767,8 @@ intel_pre_disable_primary_noatomic(struct drm_crtc *crtc) > > * Gen2 reports pipe underruns whenever all planes are disabled. > > * So disable underrun reporting before all the planes get disabled. > > */ > > - if (IS_GEN(dev_priv, 2)) > > - intel_set_cpu_fifo_underrun_reporting(dev_priv, pipe, false); > > + if (IS_GEN(i915, 2)) > > + intel_set_cpu_fifo_underrun_reporting(i915, pipe, false); > > hsw_disable_ips(to_intel_crtc_state(crtc->state)); > > @@ -5781,16 +5781,16 @@ intel_pre_disable_primary_noatomic(struct drm_crtc *crtc) > > * event which is after the vblank start event, so we need to have a > > * wait-for-vblank between disabling the plane and the pipe. > > */ > > - if (HAS_GMCH(dev_priv) && > > - intel_set_memory_cxsr(dev_priv, false)) > > - intel_wait_for_vblank(dev_priv, pipe); > > + if (HAS_GMCH(i915) && > > + intel_set_memory_cxsr(i915, false)) > > + intel_wait_for_vblank(i915, pipe); > > } > > static bool hsw_pre_update_disable_ips(const struct intel_crtc_state *old_crtc_state, > > const struct intel_crtc_state *new_crtc_state) > > { > > struct intel_crtc *crtc = to_intel_crtc(new_crtc_state->base.crtc); > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > if (!old_crtc_state->ips_enabled) > > return false; > > @@ -5804,7 +5804,7 @@ static bool hsw_pre_update_disable_ips(const struct intel_crtc_state *old_crtc_s > > * > > * Disable IPS before we program the LUT. > > */ > > - if (IS_HASWELL(dev_priv) && > > + if (IS_HASWELL(i915) && > > (new_crtc_state->base.color_mgmt_changed || > > new_crtc_state->update_pipe) && > > new_crtc_state->gamma_mode == GAMMA_MODE_MODE_SPLIT) > > @@ -5817,7 +5817,7 @@ static bool hsw_post_update_enable_ips(const struct intel_crtc_state *old_crtc_s > > const struct intel_crtc_state *new_crtc_state) > > { > > struct intel_crtc *crtc = to_intel_crtc(new_crtc_state->base.crtc); > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > if (!new_crtc_state->ips_enabled) > > return false; > > @@ -5831,7 +5831,7 @@ static bool hsw_post_update_enable_ips(const struct intel_crtc_state *old_crtc_s > > * > > * Re-enable IPS after the LUT has been programmed. > > */ > > - if (IS_HASWELL(dev_priv) && > > + if (IS_HASWELL(i915) && > > (new_crtc_state->base.color_mgmt_changed || > > new_crtc_state->update_pipe) && > > new_crtc_state->gamma_mode == GAMMA_MODE_MODE_SPLIT) > > @@ -5848,24 +5848,24 @@ static bool hsw_post_update_enable_ips(const struct intel_crtc_state *old_crtc_s > > return !old_crtc_state->ips_enabled; > > } > > -static bool needs_nv12_wa(struct drm_i915_private *dev_priv, > > +static bool needs_nv12_wa(struct drm_i915_private *i915, > > const struct intel_crtc_state *crtc_state) > > { > > if (!crtc_state->nv12_planes) > > return false; > > /* WA Display #0827: Gen9:all */ > > - if (IS_GEN(dev_priv, 9) && !IS_GEMINILAKE(dev_priv)) > > + if (IS_GEN(i915, 9) && !IS_GEMINILAKE(i915)) > > return true; > > return false; > > } > > -static bool needs_scalerclk_wa(struct drm_i915_private *dev_priv, > > +static bool needs_scalerclk_wa(struct drm_i915_private *i915, > > const struct intel_crtc_state *crtc_state) > > { > > /* Wa_2006604312:icl */ > > - if (crtc_state->scaler_state.scaler_users > 0 && IS_ICELAKE(dev_priv)) > > + if (crtc_state->scaler_state.scaler_users > 0 && IS_ICELAKE(i915)) > > return true; > > return false; > > @@ -5875,7 +5875,7 @@ static void intel_post_plane_update(struct intel_crtc_state *old_crtc_state) > > { > > struct intel_crtc *crtc = to_intel_crtc(old_crtc_state->base.crtc); > > struct drm_device *dev = crtc->base.dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > struct drm_atomic_state *old_state = old_crtc_state->base.state; > > struct intel_crtc_state *pipe_config = > > intel_atomic_get_new_crtc_state(to_intel_atomic_state(old_state), > > @@ -5904,13 +5904,13 @@ static void intel_post_plane_update(struct intel_crtc_state *old_crtc_state) > > intel_post_enable_primary(&crtc->base, pipe_config); > > } > > - if (needs_nv12_wa(dev_priv, old_crtc_state) && > > - !needs_nv12_wa(dev_priv, pipe_config)) > > - skl_wa_827(dev_priv, crtc->pipe, false); > > + if (needs_nv12_wa(i915, old_crtc_state) && > > + !needs_nv12_wa(i915, pipe_config)) > > + skl_wa_827(i915, crtc->pipe, false); > > - if (needs_scalerclk_wa(dev_priv, old_crtc_state) && > > - !needs_scalerclk_wa(dev_priv, pipe_config)) > > - icl_wa_scalerclkgating(dev_priv, crtc->pipe, false); > > + if (needs_scalerclk_wa(i915, old_crtc_state) && > > + !needs_scalerclk_wa(i915, pipe_config)) > > + icl_wa_scalerclkgating(i915, crtc->pipe, false); > > } > > static void intel_pre_plane_update(struct intel_crtc_state *old_crtc_state, > > @@ -5918,7 +5918,7 @@ static void intel_pre_plane_update(struct intel_crtc_state *old_crtc_state, > > { > > struct intel_crtc *crtc = to_intel_crtc(old_crtc_state->base.crtc); > > struct drm_device *dev = crtc->base.dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > struct drm_atomic_state *old_state = old_crtc_state->base.state; > > struct drm_plane *primary = crtc->base.primary; > > struct drm_plane_state *old_primary_state = > > @@ -5940,20 +5940,20 @@ static void intel_pre_plane_update(struct intel_crtc_state *old_crtc_state, > > * Gen2 reports pipe underruns whenever all planes are disabled. > > * So disable underrun reporting before all the planes get disabled. > > */ > > - if (IS_GEN(dev_priv, 2) && old_primary_state->visible && > > + if (IS_GEN(i915, 2) && old_primary_state->visible && > > (modeset || !new_primary_state->base.visible)) > > - intel_set_cpu_fifo_underrun_reporting(dev_priv, crtc->pipe, false); > > + intel_set_cpu_fifo_underrun_reporting(i915, crtc->pipe, false); > > } > > /* Display WA 827 */ > > - if (!needs_nv12_wa(dev_priv, old_crtc_state) && > > - needs_nv12_wa(dev_priv, pipe_config)) > > - skl_wa_827(dev_priv, crtc->pipe, true); > > + if (!needs_nv12_wa(i915, old_crtc_state) && > > + needs_nv12_wa(i915, pipe_config)) > > + skl_wa_827(i915, crtc->pipe, true); > > /* Wa_2006604312:icl */ > > - if (!needs_scalerclk_wa(dev_priv, old_crtc_state) && > > - needs_scalerclk_wa(dev_priv, pipe_config)) > > - icl_wa_scalerclkgating(dev_priv, crtc->pipe, true); > > + if (!needs_scalerclk_wa(i915, old_crtc_state) && > > + needs_scalerclk_wa(i915, pipe_config)) > > + icl_wa_scalerclkgating(i915, crtc->pipe, true); > > /* > > * Vblank time updates from the shadow to live plane control register > > @@ -5964,9 +5964,9 @@ static void intel_pre_plane_update(struct intel_crtc_state *old_crtc_state, > > * event which is after the vblank start event, so we need to have a > > * wait-for-vblank between disabling the plane and the pipe. > > */ > > - if (HAS_GMCH(dev_priv) && old_crtc_state->base.active && > > - pipe_config->disable_cxsr && intel_set_memory_cxsr(dev_priv, false)) > > - intel_wait_for_vblank(dev_priv, crtc->pipe); > > + if (HAS_GMCH(i915) && old_crtc_state->base.active && > > + pipe_config->disable_cxsr && intel_set_memory_cxsr(i915, false)) > > + intel_wait_for_vblank(i915, crtc->pipe); > > /* > > * IVB workaround: must disable low power watermarks for at least > > @@ -5977,7 +5977,7 @@ static void intel_pre_plane_update(struct intel_crtc_state *old_crtc_state, > > */ > > if (pipe_config->disable_lp_wm && ilk_disable_lp_wm(dev) && > > old_crtc_state->base.active) > > - intel_wait_for_vblank(dev_priv, crtc->pipe); > > + intel_wait_for_vblank(i915, crtc->pipe); > > /* > > * If we're doing a modeset, we're done. No need to do any pre-vblank > > @@ -6000,8 +6000,8 @@ static void intel_pre_plane_update(struct intel_crtc_state *old_crtc_state, > > * we'll continue to update watermarks the old way, if flags tell > > * us to. > > */ > > - if (dev_priv->display.initial_watermarks != NULL) > > - dev_priv->display.initial_watermarks(old_intel_state, > > + if (i915->display.initial_watermarks != NULL) > > + i915->display.initial_watermarks(old_intel_state, > > pipe_config); > > else if (pipe_config->update_wm_pre) > > intel_update_watermarks(crtc); > > @@ -6010,7 +6010,7 @@ static void intel_pre_plane_update(struct intel_crtc_state *old_crtc_state, > > static void intel_crtc_disable_planes(struct intel_atomic_state *state, > > struct intel_crtc *crtc) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > const struct intel_crtc_state *new_crtc_state = > > intel_atomic_get_new_crtc_state(state, crtc); > > unsigned int update_mask = new_crtc_state->update_planes; > > @@ -6032,7 +6032,7 @@ static void intel_crtc_disable_planes(struct intel_atomic_state *state, > > fb_bits |= plane->frontbuffer_bit; > > } > > - intel_frontbuffer_flip(dev_priv, fb_bits); > > + intel_frontbuffer_flip(i915, fb_bits); > > } > > static void intel_encoders_pre_pll_enable(struct drm_crtc *crtc, > > @@ -6190,7 +6190,7 @@ static void ironlake_crtc_enable(struct intel_crtc_state *pipe_config, > > { > > struct drm_crtc *crtc = pipe_config->base.crtc; > > struct drm_device *dev = crtc->dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > struct intel_crtc *intel_crtc = to_intel_crtc(crtc); > > int pipe = intel_crtc->pipe; > > struct intel_atomic_state *old_intel_state = > > @@ -6209,8 +6209,8 @@ static void ironlake_crtc_enable(struct intel_crtc_state *pipe_config, > > * > > * Spurious PCH underruns also occur during PCH enabling. > > */ > > - intel_set_cpu_fifo_underrun_reporting(dev_priv, pipe, false); > > - intel_set_pch_fifo_underrun_reporting(dev_priv, pipe, false); > > + intel_set_cpu_fifo_underrun_reporting(i915, pipe, false); > > + intel_set_pch_fifo_underrun_reporting(i915, pipe, false); > > if (pipe_config->has_pch_encoder) > > intel_prepare_shared_dpll(pipe_config); > > @@ -6238,8 +6238,8 @@ static void ironlake_crtc_enable(struct intel_crtc_state *pipe_config, > > * enabling. */ > > ironlake_fdi_pll_enable(pipe_config); > > } else { > > - assert_fdi_tx_disabled(dev_priv, pipe); > > - assert_fdi_rx_disabled(dev_priv, pipe); > > + assert_fdi_tx_disabled(i915, pipe); > > + assert_fdi_rx_disabled(i915, pipe); > > } > > ironlake_pfit_enable(pipe_config); > > @@ -6253,8 +6253,8 @@ static void ironlake_crtc_enable(struct intel_crtc_state *pipe_config, > > /* update DSPCNTR to configure gamma for pipe bottom color */ > > intel_disable_primary_plane(pipe_config); > > - if (dev_priv->display.initial_watermarks != NULL) > > - dev_priv->display.initial_watermarks(old_intel_state, pipe_config); > > + if (i915->display.initial_watermarks != NULL) > > + i915->display.initial_watermarks(old_intel_state, pipe_config); > > intel_enable_pipe(pipe_config); > > if (pipe_config->has_pch_encoder) > > @@ -6265,7 +6265,7 @@ static void ironlake_crtc_enable(struct intel_crtc_state *pipe_config, > > intel_encoders_enable(crtc, pipe_config, old_state); > > - if (HAS_PCH_CPT(dev_priv)) > > + if (HAS_PCH_CPT(i915)) > > cpt_verify_modeset(dev, intel_crtc->pipe); > > /* > > @@ -6275,11 +6275,11 @@ static void ironlake_crtc_enable(struct intel_crtc_state *pipe_config, > > * in case there are more corner cases we don't know about. > > */ > > if (pipe_config->has_pch_encoder) { > > - intel_wait_for_vblank(dev_priv, pipe); > > - intel_wait_for_vblank(dev_priv, pipe); > > + intel_wait_for_vblank(i915, pipe); > > + intel_wait_for_vblank(i915, pipe); > > } > > - intel_set_cpu_fifo_underrun_reporting(dev_priv, pipe, true); > > - intel_set_pch_fifo_underrun_reporting(dev_priv, pipe, true); > > + intel_set_cpu_fifo_underrun_reporting(i915, pipe, true); > > + intel_set_pch_fifo_underrun_reporting(i915, pipe, true); > > } > > /* IPS only exists on ULT machines and is tied to pipe A. */ > > @@ -6288,7 +6288,7 @@ static bool hsw_crtc_supports_ips(struct intel_crtc *crtc) > > return HAS_IPS(to_i915(crtc->base.dev)) && crtc->pipe == PIPE_A; > > } > > -static void glk_pipe_scaler_clock_gating_wa(struct drm_i915_private *dev_priv, > > +static void glk_pipe_scaler_clock_gating_wa(struct drm_i915_private *i915, > > enum pipe pipe, bool apply) > > { > > u32 val = I915_READ(CLKGATE_DIS_PSL(pipe)); > > @@ -6304,7 +6304,7 @@ static void glk_pipe_scaler_clock_gating_wa(struct drm_i915_private *dev_priv, > > static void icl_pipe_mbus_enable(struct intel_crtc *crtc) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > enum pipe pipe = crtc->pipe; > > u32 val; > > @@ -6319,7 +6319,7 @@ static void haswell_crtc_enable(struct intel_crtc_state *pipe_config, > > struct drm_atomic_state *old_state) > > { > > struct drm_crtc *crtc = pipe_config->base.crtc; > > - struct drm_i915_private *dev_priv = to_i915(crtc->dev); > > + struct drm_i915_private *i915 = to_i915(crtc->dev); > > struct intel_crtc *intel_crtc = to_intel_crtc(crtc); > > int pipe = intel_crtc->pipe, hsw_workaround_pipe; > > enum transcoder cpu_transcoder = pipe_config->cpu_transcoder; > > @@ -6359,18 +6359,18 @@ static void haswell_crtc_enable(struct intel_crtc_state *pipe_config, > > if (!transcoder_is_dsi(cpu_transcoder)) > > haswell_set_pipeconf(pipe_config); > > - if (INTEL_GEN(dev_priv) >= 9 || IS_BROADWELL(dev_priv)) > > + if (INTEL_GEN(i915) >= 9 || IS_BROADWELL(i915)) > > bdw_set_pipemisc(pipe_config); > > intel_crtc->active = true; > > /* Display WA #1180: WaDisableScalarClockGating: glk, cnl */ > > - psl_clkgate_wa = (IS_GEMINILAKE(dev_priv) || IS_CANNONLAKE(dev_priv)) && > > + psl_clkgate_wa = (IS_GEMINILAKE(i915) || IS_CANNONLAKE(i915)) && > > pipe_config->pch_pfit.enabled; > > if (psl_clkgate_wa) > > - glk_pipe_scaler_clock_gating_wa(dev_priv, pipe, true); > > + glk_pipe_scaler_clock_gating_wa(i915, pipe, true); > > - if (INTEL_GEN(dev_priv) >= 9) > > + if (INTEL_GEN(i915) >= 9) > > skylake_pfit_enable(pipe_config); > > else > > ironlake_pfit_enable(pipe_config); > > @@ -6382,20 +6382,20 @@ static void haswell_crtc_enable(struct intel_crtc_state *pipe_config, > > intel_color_load_luts(pipe_config); > > intel_color_commit(pipe_config); > > /* update DSPCNTR to configure gamma/csc for pipe bottom color */ > > - if (INTEL_GEN(dev_priv) < 9) > > + if (INTEL_GEN(i915) < 9) > > intel_disable_primary_plane(pipe_config); > > - if (INTEL_GEN(dev_priv) >= 11) > > + if (INTEL_GEN(i915) >= 11) > > icl_set_pipe_chicken(intel_crtc); > > intel_ddi_set_pipe_settings(pipe_config); > > if (!transcoder_is_dsi(cpu_transcoder)) > > intel_ddi_enable_transcoder_func(pipe_config); > > - if (dev_priv->display.initial_watermarks != NULL) > > - dev_priv->display.initial_watermarks(old_intel_state, pipe_config); > > + if (i915->display.initial_watermarks != NULL) > > + i915->display.initial_watermarks(old_intel_state, pipe_config); > > - if (INTEL_GEN(dev_priv) >= 11) > > + if (INTEL_GEN(i915) >= 11) > > icl_pipe_mbus_enable(intel_crtc); > > /* XXX: Do the pipe assertions at the right place for BXT DSI. */ > > @@ -6414,23 +6414,23 @@ static void haswell_crtc_enable(struct intel_crtc_state *pipe_config, > > intel_encoders_enable(crtc, pipe_config, old_state); > > if (psl_clkgate_wa) { > > - intel_wait_for_vblank(dev_priv, pipe); > > - glk_pipe_scaler_clock_gating_wa(dev_priv, pipe, false); > > + intel_wait_for_vblank(i915, pipe); > > + glk_pipe_scaler_clock_gating_wa(i915, pipe, false); > > } > > /* If we change the relative order between pipe/planes enabling, we need > > * to change the workaround. */ > > hsw_workaround_pipe = pipe_config->hsw_workaround_pipe; > > - if (IS_HASWELL(dev_priv) && hsw_workaround_pipe != INVALID_PIPE) { > > - intel_wait_for_vblank(dev_priv, hsw_workaround_pipe); > > - intel_wait_for_vblank(dev_priv, hsw_workaround_pipe); > > + if (IS_HASWELL(i915) && hsw_workaround_pipe != INVALID_PIPE) { > > + intel_wait_for_vblank(i915, hsw_workaround_pipe); > > + intel_wait_for_vblank(i915, hsw_workaround_pipe); > > } > > } > > static void ironlake_pfit_disable(const struct intel_crtc_state *old_crtc_state) > > { > > struct intel_crtc *crtc = to_intel_crtc(old_crtc_state->base.crtc); > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > enum pipe pipe = crtc->pipe; > > /* To avoid upsetting the power well on haswell only disable the pfit if > > @@ -6447,7 +6447,7 @@ static void ironlake_crtc_disable(struct intel_crtc_state *old_crtc_state, > > { > > struct drm_crtc *crtc = old_crtc_state->base.crtc; > > struct drm_device *dev = crtc->dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > struct intel_crtc *intel_crtc = to_intel_crtc(crtc); > > int pipe = intel_crtc->pipe; > > @@ -6456,8 +6456,8 @@ static void ironlake_crtc_disable(struct intel_crtc_state *old_crtc_state, > > * pipe is already disabled, but FDI RX/TX is still enabled. > > * Happens at least with VGA+HDMI cloning. Suppress them. > > */ > > - intel_set_cpu_fifo_underrun_reporting(dev_priv, pipe, false); > > - intel_set_pch_fifo_underrun_reporting(dev_priv, pipe, false); > > + intel_set_cpu_fifo_underrun_reporting(i915, pipe, false); > > + intel_set_pch_fifo_underrun_reporting(i915, pipe, false); > > intel_encoders_disable(crtc, old_crtc_state, old_state); > > @@ -6474,9 +6474,9 @@ static void ironlake_crtc_disable(struct intel_crtc_state *old_crtc_state, > > intel_encoders_post_disable(crtc, old_crtc_state, old_state); > > if (old_crtc_state->has_pch_encoder) { > > - ironlake_disable_pch_transcoder(dev_priv, pipe); > > + ironlake_disable_pch_transcoder(i915, pipe); > > - if (HAS_PCH_CPT(dev_priv)) { > > + if (HAS_PCH_CPT(i915)) { > > i915_reg_t reg; > > u32 temp; > > @@ -6497,15 +6497,15 @@ static void ironlake_crtc_disable(struct intel_crtc_state *old_crtc_state, > > ironlake_fdi_pll_disable(intel_crtc); > > } > > - intel_set_cpu_fifo_underrun_reporting(dev_priv, pipe, true); > > - intel_set_pch_fifo_underrun_reporting(dev_priv, pipe, true); > > + intel_set_cpu_fifo_underrun_reporting(i915, pipe, true); > > + intel_set_pch_fifo_underrun_reporting(i915, pipe, true); > > } > > static void haswell_crtc_disable(struct intel_crtc_state *old_crtc_state, > > struct drm_atomic_state *old_state) > > { > > struct drm_crtc *crtc = old_crtc_state->base.crtc; > > - struct drm_i915_private *dev_priv = to_i915(crtc->dev); > > + struct drm_i915_private *i915 = to_i915(crtc->dev); > > struct intel_crtc *intel_crtc = to_intel_crtc(crtc); > > enum transcoder cpu_transcoder = old_crtc_state->cpu_transcoder; > > @@ -6526,7 +6526,7 @@ static void haswell_crtc_disable(struct intel_crtc_state *old_crtc_state, > > intel_dsc_disable(old_crtc_state); > > - if (INTEL_GEN(dev_priv) >= 9) > > + if (INTEL_GEN(i915) >= 9) > > skylake_scaler_disable(intel_crtc); > > else > > ironlake_pfit_disable(old_crtc_state); > > @@ -6539,7 +6539,7 @@ static void haswell_crtc_disable(struct intel_crtc_state *old_crtc_state, > > static void i9xx_pfit_enable(const struct intel_crtc_state *crtc_state) > > { > > struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc); > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > if (!crtc_state->gmch_pfit.control) > > return; > > @@ -6549,7 +6549,7 @@ static void i9xx_pfit_enable(const struct intel_crtc_state *crtc_state) > > * according to register description and PRM. > > */ > > WARN_ON(I915_READ(PFIT_CONTROL) & PFIT_ENABLE); > > - assert_pipe_disabled(dev_priv, crtc->pipe); > > + assert_pipe_disabled(i915, crtc->pipe); > > I915_WRITE(PFIT_PGM_RATIOS, crtc_state->gmch_pfit.pgm_ratios); > > I915_WRITE(PFIT_CONTROL, crtc_state->gmch_pfit.control); > > @@ -6559,31 +6559,31 @@ static void i9xx_pfit_enable(const struct intel_crtc_state *crtc_state) > > I915_WRITE(BCLRPAT(crtc->pipe), 0); > > } > > -bool intel_port_is_combophy(struct drm_i915_private *dev_priv, enum port port) > > +bool intel_port_is_combophy(struct drm_i915_private *i915, enum port port) > > { > > if (port == PORT_NONE) > > return false; > > - if (IS_ELKHARTLAKE(dev_priv)) > > + if (IS_ELKHARTLAKE(i915)) > > return port <= PORT_C; > > - if (INTEL_GEN(dev_priv) >= 11) > > + if (INTEL_GEN(i915) >= 11) > > return port <= PORT_B; > > return false; > > } > > -bool intel_port_is_tc(struct drm_i915_private *dev_priv, enum port port) > > +bool intel_port_is_tc(struct drm_i915_private *i915, enum port port) > > { > > - if (INTEL_GEN(dev_priv) >= 11 && !IS_ELKHARTLAKE(dev_priv)) > > + if (INTEL_GEN(i915) >= 11 && !IS_ELKHARTLAKE(i915)) > > return port >= PORT_C && port <= PORT_F; > > return false; > > } > > -enum tc_port intel_port_to_tc(struct drm_i915_private *dev_priv, enum port port) > > +enum tc_port intel_port_to_tc(struct drm_i915_private *i915, enum port port) > > { > > - if (!intel_port_is_tc(dev_priv, port)) > > + if (!intel_port_is_tc(i915, port)) > > return PORT_TC_NONE; > > return port - PORT_C; > > @@ -6636,7 +6636,7 @@ static u64 get_crtc_power_domains(struct drm_crtc *crtc, > > struct intel_crtc_state *crtc_state) > > { > > struct drm_device *dev = crtc->dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > struct drm_encoder *encoder; > > struct intel_crtc *intel_crtc = to_intel_crtc(crtc); > > enum pipe pipe = intel_crtc->pipe; > > @@ -6658,7 +6658,7 @@ static u64 get_crtc_power_domains(struct drm_crtc *crtc, > > mask |= BIT_ULL(intel_encoder->power_domain); > > } > > - if (HAS_DDI(dev_priv) && crtc_state->has_audio) > > + if (HAS_DDI(i915) && crtc_state->has_audio) > > mask |= BIT_ULL(POWER_DOMAIN_AUDIO); > > if (crtc_state->shared_dpll) > > @@ -6671,7 +6671,7 @@ static u64 > > modeset_get_crtc_power_domains(struct drm_crtc *crtc, > > struct intel_crtc_state *crtc_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc->dev); > > + struct drm_i915_private *i915 = to_i915(crtc->dev); > > struct intel_crtc *intel_crtc = to_intel_crtc(crtc); > > enum intel_display_power_domain domain; > > u64 domains, new_domains, old_domains; > > @@ -6683,18 +6683,18 @@ modeset_get_crtc_power_domains(struct drm_crtc *crtc, > > domains = new_domains & ~old_domains; > > for_each_power_domain(domain, domains) > > - intel_display_power_get(dev_priv, domain); > > + intel_display_power_get(i915, domain); > > return old_domains & ~new_domains; > > } > > -static void modeset_put_power_domains(struct drm_i915_private *dev_priv, > > +static void modeset_put_power_domains(struct drm_i915_private *i915, > > u64 domains) > > { > > enum intel_display_power_domain domain; > > for_each_power_domain(domain, domains) > > - intel_display_power_put_unchecked(dev_priv, domain); > > + intel_display_power_put_unchecked(i915, domain); > > } > > static void valleyview_crtc_enable(struct intel_crtc_state *pipe_config, > > @@ -6704,7 +6704,7 @@ static void valleyview_crtc_enable(struct intel_crtc_state *pipe_config, > > to_intel_atomic_state(old_state); > > struct drm_crtc *crtc = pipe_config->base.crtc; > > struct drm_device *dev = crtc->dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > struct intel_crtc *intel_crtc = to_intel_crtc(crtc); > > int pipe = intel_crtc->pipe; > > @@ -6717,7 +6717,7 @@ static void valleyview_crtc_enable(struct intel_crtc_state *pipe_config, > > intel_set_pipe_timings(pipe_config); > > intel_set_pipe_src_size(pipe_config); > > - if (IS_CHERRYVIEW(dev_priv) && pipe == PIPE_B) { > > + if (IS_CHERRYVIEW(i915) && pipe == PIPE_B) { > > I915_WRITE(CHV_BLEND(pipe), CHV_BLEND_LEGACY); > > I915_WRITE(CHV_CANVAS(pipe), 0); > > } > > @@ -6726,11 +6726,11 @@ static void valleyview_crtc_enable(struct intel_crtc_state *pipe_config, > > intel_crtc->active = true; > > - intel_set_cpu_fifo_underrun_reporting(dev_priv, pipe, true); > > + intel_set_cpu_fifo_underrun_reporting(i915, pipe, true); > > intel_encoders_pre_pll_enable(crtc, pipe_config, old_state); > > - if (IS_CHERRYVIEW(dev_priv)) { > > + if (IS_CHERRYVIEW(i915)) { > > chv_prepare_pll(intel_crtc, pipe_config); > > chv_enable_pll(intel_crtc, pipe_config); > > } else { > > @@ -6747,7 +6747,7 @@ static void valleyview_crtc_enable(struct intel_crtc_state *pipe_config, > > /* update DSPCNTR to configure gamma for pipe bottom color */ > > intel_disable_primary_plane(pipe_config); > > - dev_priv->display.initial_watermarks(old_intel_state, > > + i915->display.initial_watermarks(old_intel_state, > > pipe_config); > > intel_enable_pipe(pipe_config); > > @@ -6760,7 +6760,7 @@ static void valleyview_crtc_enable(struct intel_crtc_state *pipe_config, > > static void i9xx_set_pll_dividers(const struct intel_crtc_state *crtc_state) > > { > > struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc); > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > I915_WRITE(FP0(crtc->pipe), crtc_state->dpll_hw_state.fp0); > > I915_WRITE(FP1(crtc->pipe), crtc_state->dpll_hw_state.fp1); > > @@ -6773,7 +6773,7 @@ static void i9xx_crtc_enable(struct intel_crtc_state *pipe_config, > > to_intel_atomic_state(old_state); > > struct drm_crtc *crtc = pipe_config->base.crtc; > > struct drm_device *dev = crtc->dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > struct intel_crtc *intel_crtc = to_intel_crtc(crtc); > > enum pipe pipe = intel_crtc->pipe; > > @@ -6792,8 +6792,8 @@ static void i9xx_crtc_enable(struct intel_crtc_state *pipe_config, > > intel_crtc->active = true; > > - if (!IS_GEN(dev_priv, 2)) > > - intel_set_cpu_fifo_underrun_reporting(dev_priv, pipe, true); > > + if (!IS_GEN(i915, 2)) > > + intel_set_cpu_fifo_underrun_reporting(i915, pipe, true); > > intel_encoders_pre_enable(crtc, pipe_config, old_state); > > @@ -6806,8 +6806,8 @@ static void i9xx_crtc_enable(struct intel_crtc_state *pipe_config, > > /* update DSPCNTR to configure gamma for pipe bottom color */ > > intel_disable_primary_plane(pipe_config); > > - if (dev_priv->display.initial_watermarks != NULL) > > - dev_priv->display.initial_watermarks(old_intel_state, > > + if (i915->display.initial_watermarks != NULL) > > + i915->display.initial_watermarks(old_intel_state, > > pipe_config); > > else > > intel_update_watermarks(intel_crtc); > > @@ -6822,12 +6822,12 @@ static void i9xx_crtc_enable(struct intel_crtc_state *pipe_config, > > static void i9xx_pfit_disable(const struct intel_crtc_state *old_crtc_state) > > { > > struct intel_crtc *crtc = to_intel_crtc(old_crtc_state->base.crtc); > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > if (!old_crtc_state->gmch_pfit.control) > > return; > > - assert_pipe_disabled(dev_priv, crtc->pipe); > > + assert_pipe_disabled(i915, crtc->pipe); > > DRM_DEBUG_KMS("disabling pfit, current: 0x%08x\n", > > I915_READ(PFIT_CONTROL)); > > @@ -6839,7 +6839,7 @@ static void i9xx_crtc_disable(struct intel_crtc_state *old_crtc_state, > > { > > struct drm_crtc *crtc = old_crtc_state->base.crtc; > > struct drm_device *dev = crtc->dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > struct intel_crtc *intel_crtc = to_intel_crtc(crtc); > > int pipe = intel_crtc->pipe; > > @@ -6847,8 +6847,8 @@ static void i9xx_crtc_disable(struct intel_crtc_state *old_crtc_state, > > * On gen2 planes are double buffered but the pipe isn't, so we must > > * wait for planes to fully turn off before disabling the pipe. > > */ > > - if (IS_GEN(dev_priv, 2)) > > - intel_wait_for_vblank(dev_priv, pipe); > > + if (IS_GEN(i915, 2)) > > + intel_wait_for_vblank(i915, pipe); > > intel_encoders_disable(crtc, old_crtc_state, old_state); > > @@ -6862,25 +6862,25 @@ static void i9xx_crtc_disable(struct intel_crtc_state *old_crtc_state, > > intel_encoders_post_disable(crtc, old_crtc_state, old_state); > > if (!intel_crtc_has_type(old_crtc_state, INTEL_OUTPUT_DSI)) { > > - if (IS_CHERRYVIEW(dev_priv)) > > - chv_disable_pll(dev_priv, pipe); > > - else if (IS_VALLEYVIEW(dev_priv)) > > - vlv_disable_pll(dev_priv, pipe); > > + if (IS_CHERRYVIEW(i915)) > > + chv_disable_pll(i915, pipe); > > + else if (IS_VALLEYVIEW(i915)) > > + vlv_disable_pll(i915, pipe); > > else > > i9xx_disable_pll(old_crtc_state); > > } > > intel_encoders_post_pll_disable(crtc, old_crtc_state, old_state); > > - if (!IS_GEN(dev_priv, 2)) > > - intel_set_cpu_fifo_underrun_reporting(dev_priv, pipe, false); > > + if (!IS_GEN(i915, 2)) > > + intel_set_cpu_fifo_underrun_reporting(i915, pipe, false); > > - if (!dev_priv->display.initial_watermarks) > > + if (!i915->display.initial_watermarks) > > intel_update_watermarks(intel_crtc); > > /* clock the pipe down to 640x480@60 to potentially save power */ > > - if (IS_I830(dev_priv)) > > - i830_enable_pipe(dev_priv, pipe); > > + if (IS_I830(i915)) > > + i830_enable_pipe(i915, pipe); > > } > > static void intel_crtc_disable_noatomic(struct drm_crtc *crtc, > > @@ -6888,9 +6888,9 @@ static void intel_crtc_disable_noatomic(struct drm_crtc *crtc, > > { > > struct intel_encoder *encoder; > > struct intel_crtc *intel_crtc = to_intel_crtc(crtc); > > - struct drm_i915_private *dev_priv = to_i915(crtc->dev); > > + struct drm_i915_private *i915 = to_i915(crtc->dev); > > struct intel_bw_state *bw_state = > > - to_intel_bw_state(dev_priv->bw_obj.state); > > + to_intel_bw_state(i915->bw_obj.state); > > enum intel_display_power_domain domain; > > struct intel_plane *plane; > > u64 domains; > > @@ -6901,7 +6901,7 @@ static void intel_crtc_disable_noatomic(struct drm_crtc *crtc, > > if (!intel_crtc->active) > > return; > > - for_each_intel_plane_on_crtc(&dev_priv->drm, intel_crtc, plane) { > > + for_each_intel_plane_on_crtc(&i915->drm, intel_crtc, plane) { > > const struct intel_plane_state *plane_state = > > to_intel_plane_state(plane->base.state); > > @@ -6924,7 +6924,7 @@ static void intel_crtc_disable_noatomic(struct drm_crtc *crtc, > > WARN_ON(IS_ERR(crtc_state) || ret); > > - dev_priv->display.crtc_disable(crtc_state, state); > > + i915->display.crtc_disable(crtc_state, state); > > drm_atomic_state_put(state); > > @@ -6947,12 +6947,12 @@ static void intel_crtc_disable_noatomic(struct drm_crtc *crtc, > > domains = intel_crtc->enabled_power_domains; > > for_each_power_domain(domain, domains) > > - intel_display_power_put_unchecked(dev_priv, domain); > > + intel_display_power_put_unchecked(i915, domain); > > intel_crtc->enabled_power_domains = 0; > > - dev_priv->active_crtcs &= ~(1 << intel_crtc->pipe); > > - dev_priv->min_cdclk[intel_crtc->pipe] = 0; > > - dev_priv->min_voltage_level[intel_crtc->pipe] = 0; > > + i915->active_crtcs &= ~(1 << intel_crtc->pipe); > > + i915->min_cdclk[intel_crtc->pipe] = 0; > > + i915->min_voltage_level[intel_crtc->pipe] = 0; > > bw_state->data_rate[intel_crtc->pipe] = 0; > > bw_state->num_active_planes[intel_crtc->pipe] = 0; > > @@ -6964,7 +6964,7 @@ static void intel_crtc_disable_noatomic(struct drm_crtc *crtc, > > */ > > int intel_display_suspend(struct drm_device *dev) > > { > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > struct drm_atomic_state *state; > > int ret; > > @@ -6973,7 +6973,7 @@ int intel_display_suspend(struct drm_device *dev) > > if (ret) > > DRM_ERROR("Suspending crtc's failed with %i\n", ret); > > else > > - dev_priv->modeset_restore_state = state; > > + i915->modeset_restore_state = state; > > return ret; > > } > > @@ -7035,7 +7035,7 @@ static int pipe_required_fdi_lanes(struct intel_crtc_state *crtc_state) > > static int ironlake_check_fdi_lanes(struct drm_device *dev, enum pipe pipe, > > struct intel_crtc_state *pipe_config) > > { > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > struct drm_atomic_state *state = pipe_config->base.state; > > struct intel_crtc *other_crtc; > > struct intel_crtc_state *other_crtc_state; > > @@ -7048,7 +7048,7 @@ static int ironlake_check_fdi_lanes(struct drm_device *dev, enum pipe pipe, > > return -EINVAL; > > } > > - if (IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv)) { > > + if (IS_HASWELL(i915) || IS_BROADWELL(i915)) { > > if (pipe_config->fdi_lanes > 2) { > > DRM_DEBUG_KMS("only 2 lanes on haswell, required: %i lanes\n", > > pipe_config->fdi_lanes); > > @@ -7058,7 +7058,7 @@ static int ironlake_check_fdi_lanes(struct drm_device *dev, enum pipe pipe, > > } > > } > > - if (INTEL_INFO(dev_priv)->num_pipes == 2) > > + if (INTEL_INFO(i915)->num_pipes == 2) > > return 0; > > /* Ivybridge 3 pipe is really complicated */ > > @@ -7069,7 +7069,7 @@ static int ironlake_check_fdi_lanes(struct drm_device *dev, enum pipe pipe, > > if (pipe_config->fdi_lanes <= 2) > > return 0; > > - other_crtc = intel_get_crtc_for_pipe(dev_priv, PIPE_C); > > + other_crtc = intel_get_crtc_for_pipe(i915, PIPE_C); > > other_crtc_state = > > intel_atomic_get_crtc_state(state, other_crtc); > > if (IS_ERR(other_crtc_state)) > > @@ -7088,7 +7088,7 @@ static int ironlake_check_fdi_lanes(struct drm_device *dev, enum pipe pipe, > > return -EINVAL; > > } > > - other_crtc = intel_get_crtc_for_pipe(dev_priv, PIPE_B); > > + other_crtc = intel_get_crtc_for_pipe(i915, PIPE_B); > > other_crtc_state = > > intel_atomic_get_crtc_state(state, other_crtc); > > if (IS_ERR(other_crtc_state)) > > @@ -7156,7 +7156,7 @@ static int ironlake_fdi_compute_config(struct intel_crtc *intel_crtc, > > bool hsw_crtc_state_ips_capable(const struct intel_crtc_state *crtc_state) > > { > > struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc); > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > /* IPS only exists on ULT machines and is tied to pipe A. */ > > if (!hsw_crtc_supports_ips(crtc)) > > @@ -7175,8 +7175,8 @@ bool hsw_crtc_state_ips_capable(const struct intel_crtc_state *crtc_state) > > * > > * Should measure whether using a lower cdclk w/o IPS > > */ > > - if (IS_BROADWELL(dev_priv) && > > - crtc_state->pixel_rate > dev_priv->max_cdclk_freq * 95 / 100) > > + if (IS_BROADWELL(i915) && > > + crtc_state->pixel_rate > i915->max_cdclk_freq * 95 / 100) > > return false; > > return true; > > @@ -7184,7 +7184,7 @@ bool hsw_crtc_state_ips_capable(const struct intel_crtc_state *crtc_state) > > static bool hsw_compute_ips_config(struct intel_crtc_state *crtc_state) > > { > > - struct drm_i915_private *dev_priv = > > + struct drm_i915_private *i915 = > > to_i915(crtc_state->base.crtc->dev); > > struct intel_atomic_state *intel_state = > > to_intel_atomic_state(crtc_state->base.state); > > @@ -7206,7 +7206,7 @@ static bool hsw_compute_ips_config(struct intel_crtc_state *crtc_state) > > return false; > > /* pixel rate mustn't exceed 95% of cdclk with IPS on BDW */ > > - if (IS_BROADWELL(dev_priv) && > > + if (IS_BROADWELL(i915) && > > crtc_state->pixel_rate > intel_state->cdclk.logical.cdclk * 95 / 100) > > return false; > > @@ -7215,11 +7215,11 @@ static bool hsw_compute_ips_config(struct intel_crtc_state *crtc_state) > > static bool intel_crtc_supports_double_wide(const struct intel_crtc *crtc) > > { > > - const struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + const struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > /* GDG double wide on either pipe, otherwise pipe A only */ > > - return INTEL_GEN(dev_priv) < 4 && > > - (crtc->pipe == PIPE_A || IS_I915G(dev_priv)); > > + return INTEL_GEN(i915) < 4 && > > + (crtc->pipe == PIPE_A || IS_I915G(i915)); > > } > > static u32 ilk_pipe_pixel_rate(const struct intel_crtc_state *pipe_config) > > @@ -7259,9 +7259,9 @@ static u32 ilk_pipe_pixel_rate(const struct intel_crtc_state *pipe_config) > > static void intel_crtc_compute_pixel_rate(struct intel_crtc_state *crtc_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc_state->base.crtc->dev); > > + struct drm_i915_private *i915 = to_i915(crtc_state->base.crtc->dev); > > - if (HAS_GMCH(dev_priv)) > > + if (HAS_GMCH(i915)) > > /* FIXME calculate proper pipe pixel rate for GMCH pfit */ > > crtc_state->pixel_rate = > > crtc_state->base.adjusted_mode.crtc_clock; > > @@ -7273,12 +7273,12 @@ static void intel_crtc_compute_pixel_rate(struct intel_crtc_state *crtc_state) > > static int intel_crtc_compute_config(struct intel_crtc *crtc, > > struct intel_crtc_state *pipe_config) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > const struct drm_display_mode *adjusted_mode = &pipe_config->base.adjusted_mode; > > - int clock_limit = dev_priv->max_dotclk_freq; > > + int clock_limit = i915->max_dotclk_freq; > > - if (INTEL_GEN(dev_priv) < 4) { > > - clock_limit = dev_priv->max_cdclk_freq * 9 / 10; > > + if (INTEL_GEN(i915) < 4) { > > + clock_limit = i915->max_cdclk_freq * 9 / 10; > > /* > > * Enable double wide mode when the dot clock > > @@ -7286,7 +7286,7 @@ static int intel_crtc_compute_config(struct intel_crtc *crtc, > > */ > > if (intel_crtc_supports_double_wide(crtc) && > > adjusted_mode->crtc_clock > clock_limit) { > > - clock_limit = dev_priv->max_dotclk_freq; > > + clock_limit = i915->max_dotclk_freq; > > pipe_config->double_wide = true; > > } > > } > > @@ -7323,7 +7323,7 @@ static int intel_crtc_compute_config(struct intel_crtc *crtc, > > } > > if (intel_crtc_has_type(pipe_config, INTEL_OUTPUT_LVDS) && > > - intel_is_dual_link_lvds(dev_priv)) { > > + intel_is_dual_link_lvds(i915)) { > > DRM_DEBUG_KMS("Odd pipe source width not supported with dual link LVDS\n"); > > return -EINVAL; > > } > > @@ -7332,7 +7332,7 @@ static int intel_crtc_compute_config(struct intel_crtc *crtc, > > /* Cantiga+ cannot handle modes with a hsync front porch of 0. > > * WaPruneModeWithIncorrectHsyncOffset:ctg,elk,ilk,snb,ivb,vlv,hsw. > > */ > > - if ((INTEL_GEN(dev_priv) > 4 || IS_G4X(dev_priv)) && > > + if ((INTEL_GEN(i915) > 4 || IS_G4X(i915)) && > > adjusted_mode->crtc_hsync_start == adjusted_mode->crtc_hdisplay) > > return -EINVAL; > > @@ -7392,12 +7392,12 @@ intel_link_compute_m_n(u16 bits_per_pixel, int nlanes, > > constant_n); > > } > > -static inline bool intel_panel_use_ssc(struct drm_i915_private *dev_priv) > > +static inline bool intel_panel_use_ssc(struct drm_i915_private *i915) > > { > > if (i915_modparams.panel_use_ssc >= 0) > > return i915_modparams.panel_use_ssc != 0; > > - return dev_priv->vbt.lvds_use_ssc > > - && !(dev_priv->quirks & QUIRK_LVDS_SSC_DISABLE); > > + return i915->vbt.lvds_use_ssc > > + && !(i915->quirks & QUIRK_LVDS_SSC_DISABLE); > > } > > static u32 pnv_dpll_compute_fp(struct dpll *dpll) > > @@ -7414,10 +7414,10 @@ static void i9xx_update_pll_dividers(struct intel_crtc *crtc, > > struct intel_crtc_state *crtc_state, > > struct dpll *reduced_clock) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > u32 fp, fp2 = 0; > > - if (IS_PINEVIEW(dev_priv)) { > > + if (IS_PINEVIEW(i915)) { > > fp = pnv_dpll_compute_fp(&crtc_state->dpll); > > if (reduced_clock) > > fp2 = pnv_dpll_compute_fp(reduced_clock); > > @@ -7437,7 +7437,7 @@ static void i9xx_update_pll_dividers(struct intel_crtc *crtc, > > } > > } > > -static void vlv_pllb_recal_opamp(struct drm_i915_private *dev_priv, enum pipe > > +static void vlv_pllb_recal_opamp(struct drm_i915_private *i915, enum pipe > > pipe) > > { > > u32 reg_val; > > @@ -7446,31 +7446,31 @@ static void vlv_pllb_recal_opamp(struct drm_i915_private *dev_priv, enum pipe > > * PLLB opamp always calibrates to max value of 0x3f, force enable it > > * and set it to a reasonable value instead. > > */ > > - reg_val = vlv_dpio_read(dev_priv, pipe, VLV_PLL_DW9(1)); > > + reg_val = vlv_dpio_read(i915, pipe, VLV_PLL_DW9(1)); > > reg_val &= 0xffffff00; > > reg_val |= 0x00000030; > > - vlv_dpio_write(dev_priv, pipe, VLV_PLL_DW9(1), reg_val); > > + vlv_dpio_write(i915, pipe, VLV_PLL_DW9(1), reg_val); > > - reg_val = vlv_dpio_read(dev_priv, pipe, VLV_REF_DW13); > > + reg_val = vlv_dpio_read(i915, pipe, VLV_REF_DW13); > > reg_val &= 0x00ffffff; > > reg_val |= 0x8c000000; > > - vlv_dpio_write(dev_priv, pipe, VLV_REF_DW13, reg_val); > > + vlv_dpio_write(i915, pipe, VLV_REF_DW13, reg_val); > > - reg_val = vlv_dpio_read(dev_priv, pipe, VLV_PLL_DW9(1)); > > + reg_val = vlv_dpio_read(i915, pipe, VLV_PLL_DW9(1)); > > reg_val &= 0xffffff00; > > - vlv_dpio_write(dev_priv, pipe, VLV_PLL_DW9(1), reg_val); > > + vlv_dpio_write(i915, pipe, VLV_PLL_DW9(1), reg_val); > > - reg_val = vlv_dpio_read(dev_priv, pipe, VLV_REF_DW13); > > + reg_val = vlv_dpio_read(i915, pipe, VLV_REF_DW13); > > reg_val &= 0x00ffffff; > > reg_val |= 0xb0000000; > > - vlv_dpio_write(dev_priv, pipe, VLV_REF_DW13, reg_val); > > + vlv_dpio_write(i915, pipe, VLV_REF_DW13, reg_val); > > } > > static void intel_pch_transcoder_set_m_n(const struct intel_crtc_state *crtc_state, > > const struct intel_link_m_n *m_n) > > { > > struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc); > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > enum pipe pipe = crtc->pipe; > > I915_WRITE(PCH_TRANS_DATA_M1(pipe), TU_SIZE(m_n->tu) | m_n->gmch_m); > > @@ -7479,17 +7479,17 @@ static void intel_pch_transcoder_set_m_n(const struct intel_crtc_state *crtc_sta > > I915_WRITE(PCH_TRANS_LINK_N1(pipe), m_n->link_n); > > } > > -static bool transcoder_has_m2_n2(struct drm_i915_private *dev_priv, > > +static bool transcoder_has_m2_n2(struct drm_i915_private *i915, > > enum transcoder transcoder) > > { > > - if (IS_HASWELL(dev_priv)) > > + if (IS_HASWELL(i915)) > > return transcoder == TRANSCODER_EDP; > > /* > > * Strictly speaking some registers are available before > > * gen7, but we only support DRRS on gen7+ > > */ > > - return IS_GEN(dev_priv, 7) || IS_CHERRYVIEW(dev_priv); > > + return IS_GEN(i915, 7) || IS_CHERRYVIEW(i915); > > } > > static void intel_cpu_transcoder_set_m_n(const struct intel_crtc_state *crtc_state, > > @@ -7497,11 +7497,11 @@ static void intel_cpu_transcoder_set_m_n(const struct intel_crtc_state *crtc_sta > > const struct intel_link_m_n *m2_n2) > > { > > struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc); > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > enum pipe pipe = crtc->pipe; > > enum transcoder transcoder = crtc_state->cpu_transcoder; > > - if (INTEL_GEN(dev_priv) >= 5) { > > + if (INTEL_GEN(i915) >= 5) { > > I915_WRITE(PIPE_DATA_M1(transcoder), TU_SIZE(m_n->tu) | m_n->gmch_m); > > I915_WRITE(PIPE_DATA_N1(transcoder), m_n->gmch_n); > > I915_WRITE(PIPE_LINK_M1(transcoder), m_n->link_m); > > @@ -7511,7 +7511,7 @@ static void intel_cpu_transcoder_set_m_n(const struct intel_crtc_state *crtc_sta > > * (to make sure the registers are not unnecessarily accessed). > > */ > > if (m2_n2 && crtc_state->has_drrs && > > - transcoder_has_m2_n2(dev_priv, transcoder)) { > > + transcoder_has_m2_n2(i915, transcoder)) { > > I915_WRITE(PIPE_DATA_M2(transcoder), > > TU_SIZE(m2_n2->tu) | m2_n2->gmch_m); > > I915_WRITE(PIPE_DATA_N2(transcoder), m2_n2->gmch_n); > > @@ -7588,7 +7588,7 @@ static void vlv_prepare_pll(struct intel_crtc *crtc, > > const struct intel_crtc_state *pipe_config) > > { > > struct drm_device *dev = crtc->base.dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > enum pipe pipe = crtc->pipe; > > u32 mdiv; > > u32 bestn, bestm1, bestm2, bestp1, bestp2; > > @@ -7603,7 +7603,7 @@ static void vlv_prepare_pll(struct intel_crtc *crtc, > > if ((pipe_config->dpll_hw_state.dpll & DPLL_VCO_ENABLE) == 0) > > return; > > - vlv_dpio_get(dev_priv); > > + vlv_dpio_get(i915); > > bestn = pipe_config->dpll.n; > > bestm1 = pipe_config->dpll.m1; > > @@ -7615,18 +7615,18 @@ static void vlv_prepare_pll(struct intel_crtc *crtc, > > /* PLL B needs special handling */ > > if (pipe == PIPE_B) > > - vlv_pllb_recal_opamp(dev_priv, pipe); > > + vlv_pllb_recal_opamp(i915, pipe); > > /* Set up Tx target for periodic Rcomp update */ > > - vlv_dpio_write(dev_priv, pipe, VLV_PLL_DW9_BCAST, 0x0100000f); > > + vlv_dpio_write(i915, pipe, VLV_PLL_DW9_BCAST, 0x0100000f); > > /* Disable target IRef on PLL */ > > - reg_val = vlv_dpio_read(dev_priv, pipe, VLV_PLL_DW8(pipe)); > > + reg_val = vlv_dpio_read(i915, pipe, VLV_PLL_DW8(pipe)); > > reg_val &= 0x00ffffff; > > - vlv_dpio_write(dev_priv, pipe, VLV_PLL_DW8(pipe), reg_val); > > + vlv_dpio_write(i915, pipe, VLV_PLL_DW8(pipe), reg_val); > > /* Disable fast lock */ > > - vlv_dpio_write(dev_priv, pipe, VLV_CMN_DW0, 0x610); > > + vlv_dpio_write(i915, pipe, VLV_CMN_DW0, 0x610); > > /* Set idtafcrecal before PLL is enabled */ > > mdiv = ((bestm1 << DPIO_M1DIV_SHIFT) | (bestm2 & DPIO_M2DIV_MASK)); > > @@ -7640,55 +7640,55 @@ static void vlv_prepare_pll(struct intel_crtc *crtc, > > * Note: don't use the DAC post divider as it seems unstable. > > */ > > mdiv |= (DPIO_POST_DIV_HDMIDP << DPIO_POST_DIV_SHIFT); > > - vlv_dpio_write(dev_priv, pipe, VLV_PLL_DW3(pipe), mdiv); > > + vlv_dpio_write(i915, pipe, VLV_PLL_DW3(pipe), mdiv); > > mdiv |= DPIO_ENABLE_CALIBRATION; > > - vlv_dpio_write(dev_priv, pipe, VLV_PLL_DW3(pipe), mdiv); > > + vlv_dpio_write(i915, pipe, VLV_PLL_DW3(pipe), mdiv); > > /* Set HBR and RBR LPF coefficients */ > > if (pipe_config->port_clock == 162000 || > > intel_crtc_has_type(pipe_config, INTEL_OUTPUT_ANALOG) || > > intel_crtc_has_type(pipe_config, INTEL_OUTPUT_HDMI)) > > - vlv_dpio_write(dev_priv, pipe, VLV_PLL_DW10(pipe), > > + vlv_dpio_write(i915, pipe, VLV_PLL_DW10(pipe), > > 0x009f0003); > > else > > - vlv_dpio_write(dev_priv, pipe, VLV_PLL_DW10(pipe), > > + vlv_dpio_write(i915, pipe, VLV_PLL_DW10(pipe), > > 0x00d0000f); > > if (intel_crtc_has_dp_encoder(pipe_config)) { > > /* Use SSC source */ > > if (pipe == PIPE_A) > > - vlv_dpio_write(dev_priv, pipe, VLV_PLL_DW5(pipe), > > + vlv_dpio_write(i915, pipe, VLV_PLL_DW5(pipe), > > 0x0df40000); > > else > > - vlv_dpio_write(dev_priv, pipe, VLV_PLL_DW5(pipe), > > + vlv_dpio_write(i915, pipe, VLV_PLL_DW5(pipe), > > 0x0df70000); > > } else { /* HDMI or VGA */ > > /* Use bend source */ > > if (pipe == PIPE_A) > > - vlv_dpio_write(dev_priv, pipe, VLV_PLL_DW5(pipe), > > + vlv_dpio_write(i915, pipe, VLV_PLL_DW5(pipe), > > 0x0df70000); > > else > > - vlv_dpio_write(dev_priv, pipe, VLV_PLL_DW5(pipe), > > + vlv_dpio_write(i915, pipe, VLV_PLL_DW5(pipe), > > 0x0df40000); > > } > > - coreclk = vlv_dpio_read(dev_priv, pipe, VLV_PLL_DW7(pipe)); > > + coreclk = vlv_dpio_read(i915, pipe, VLV_PLL_DW7(pipe)); > > coreclk = (coreclk & 0x0000ff00) | 0x01c00000; > > if (intel_crtc_has_dp_encoder(pipe_config)) > > coreclk |= 0x01000000; > > - vlv_dpio_write(dev_priv, pipe, VLV_PLL_DW7(pipe), coreclk); > > + vlv_dpio_write(i915, pipe, VLV_PLL_DW7(pipe), coreclk); > > - vlv_dpio_write(dev_priv, pipe, VLV_PLL_DW11(pipe), 0x87871000); > > + vlv_dpio_write(i915, pipe, VLV_PLL_DW11(pipe), 0x87871000); > > - vlv_dpio_put(dev_priv); > > + vlv_dpio_put(i915); > > } > > static void chv_prepare_pll(struct intel_crtc *crtc, > > const struct intel_crtc_state *pipe_config) > > { > > struct drm_device *dev = crtc->base.dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > enum pipe pipe = crtc->pipe; > > enum dpio_channel port = vlv_pipe_to_channel(pipe); > > u32 loopfilter, tribuf_calcntr; > > @@ -7714,42 +7714,42 @@ static void chv_prepare_pll(struct intel_crtc *crtc, > > dpio_val = 0; > > loopfilter = 0; > > - vlv_dpio_get(dev_priv); > > + vlv_dpio_get(i915); > > /* p1 and p2 divider */ > > - vlv_dpio_write(dev_priv, pipe, CHV_CMN_DW13(port), > > + vlv_dpio_write(i915, pipe, CHV_CMN_DW13(port), > > 5 << DPIO_CHV_S1_DIV_SHIFT | > > bestp1 << DPIO_CHV_P1_DIV_SHIFT | > > bestp2 << DPIO_CHV_P2_DIV_SHIFT | > > 1 << DPIO_CHV_K_DIV_SHIFT); > > /* Feedback post-divider - m2 */ > > - vlv_dpio_write(dev_priv, pipe, CHV_PLL_DW0(port), bestm2); > > + vlv_dpio_write(i915, pipe, CHV_PLL_DW0(port), bestm2); > > /* Feedback refclk divider - n and m1 */ > > - vlv_dpio_write(dev_priv, pipe, CHV_PLL_DW1(port), > > + vlv_dpio_write(i915, pipe, CHV_PLL_DW1(port), > > DPIO_CHV_M1_DIV_BY_2 | > > 1 << DPIO_CHV_N_DIV_SHIFT); > > /* M2 fraction division */ > > - vlv_dpio_write(dev_priv, pipe, CHV_PLL_DW2(port), bestm2_frac); > > + vlv_dpio_write(i915, pipe, CHV_PLL_DW2(port), bestm2_frac); > > /* M2 fraction division enable */ > > - dpio_val = vlv_dpio_read(dev_priv, pipe, CHV_PLL_DW3(port)); > > + dpio_val = vlv_dpio_read(i915, pipe, CHV_PLL_DW3(port)); > > dpio_val &= ~(DPIO_CHV_FEEDFWD_GAIN_MASK | DPIO_CHV_FRAC_DIV_EN); > > dpio_val |= (2 << DPIO_CHV_FEEDFWD_GAIN_SHIFT); > > if (bestm2_frac) > > dpio_val |= DPIO_CHV_FRAC_DIV_EN; > > - vlv_dpio_write(dev_priv, pipe, CHV_PLL_DW3(port), dpio_val); > > + vlv_dpio_write(i915, pipe, CHV_PLL_DW3(port), dpio_val); > > /* Program digital lock detect threshold */ > > - dpio_val = vlv_dpio_read(dev_priv, pipe, CHV_PLL_DW9(port)); > > + dpio_val = vlv_dpio_read(i915, pipe, CHV_PLL_DW9(port)); > > dpio_val &= ~(DPIO_CHV_INT_LOCK_THRESHOLD_MASK | > > DPIO_CHV_INT_LOCK_THRESHOLD_SEL_COARSE); > > dpio_val |= (0x5 << DPIO_CHV_INT_LOCK_THRESHOLD_SHIFT); > > if (!bestm2_frac) > > dpio_val |= DPIO_CHV_INT_LOCK_THRESHOLD_SEL_COARSE; > > - vlv_dpio_write(dev_priv, pipe, CHV_PLL_DW9(port), dpio_val); > > + vlv_dpio_write(i915, pipe, CHV_PLL_DW9(port), dpio_val); > > /* Loop filter */ > > if (vco == 5400000) { > > @@ -7774,24 +7774,24 @@ static void chv_prepare_pll(struct intel_crtc *crtc, > > loopfilter |= (0x3 << DPIO_CHV_GAIN_CTRL_SHIFT); > > tribuf_calcntr = 0; > > } > > - vlv_dpio_write(dev_priv, pipe, CHV_PLL_DW6(port), loopfilter); > > + vlv_dpio_write(i915, pipe, CHV_PLL_DW6(port), loopfilter); > > - dpio_val = vlv_dpio_read(dev_priv, pipe, CHV_PLL_DW8(port)); > > + dpio_val = vlv_dpio_read(i915, pipe, CHV_PLL_DW8(port)); > > dpio_val &= ~DPIO_CHV_TDC_TARGET_CNT_MASK; > > dpio_val |= (tribuf_calcntr << DPIO_CHV_TDC_TARGET_CNT_SHIFT); > > - vlv_dpio_write(dev_priv, pipe, CHV_PLL_DW8(port), dpio_val); > > + vlv_dpio_write(i915, pipe, CHV_PLL_DW8(port), dpio_val); > > /* AFC Recal */ > > - vlv_dpio_write(dev_priv, pipe, CHV_CMN_DW14(port), > > - vlv_dpio_read(dev_priv, pipe, CHV_CMN_DW14(port)) | > > + vlv_dpio_write(i915, pipe, CHV_CMN_DW14(port), > > + vlv_dpio_read(i915, pipe, CHV_CMN_DW14(port)) | > > DPIO_AFC_RECAL); > > - vlv_dpio_put(dev_priv); > > + vlv_dpio_put(i915); > > } > > /** > > * vlv_force_pll_on - forcibly enable just the PLL > > - * @dev_priv: i915 private structure > > + * @i915: i915 private structure > > * @pipe: pipe PLL to enable > > * @dpll: PLL configuration > > * > > @@ -7799,10 +7799,10 @@ static void chv_prepare_pll(struct intel_crtc *crtc, > > * in cases where we need the PLL enabled even when @pipe is not going to > > * be enabled. > > */ > > -int vlv_force_pll_on(struct drm_i915_private *dev_priv, enum pipe pipe, > > +int vlv_force_pll_on(struct drm_i915_private *i915, enum pipe pipe, > > const struct dpll *dpll) > > { > > - struct intel_crtc *crtc = intel_get_crtc_for_pipe(dev_priv, pipe); > > + struct intel_crtc *crtc = intel_get_crtc_for_pipe(i915, pipe); > > struct intel_crtc_state *pipe_config; > > pipe_config = kzalloc(sizeof(*pipe_config), GFP_KERNEL); > > @@ -7813,7 +7813,7 @@ int vlv_force_pll_on(struct drm_i915_private *dev_priv, enum pipe pipe, > > pipe_config->pixel_multiplier = 1; > > pipe_config->dpll = *dpll; > > - if (IS_CHERRYVIEW(dev_priv)) { > > + if (IS_CHERRYVIEW(i915)) { > > chv_compute_dpll(crtc, pipe_config); > > chv_prepare_pll(crtc, pipe_config); > > chv_enable_pll(crtc, pipe_config); > > @@ -7830,25 +7830,25 @@ int vlv_force_pll_on(struct drm_i915_private *dev_priv, enum pipe pipe, > > /** > > * vlv_force_pll_off - forcibly disable just the PLL > > - * @dev_priv: i915 private structure > > + * @i915: i915 private structure > > * @pipe: pipe PLL to disable > > * > > * Disable the PLL for @pipe. To be used in cases where we need > > * the PLL enabled even when @pipe is not going to be enabled. > > */ > > -void vlv_force_pll_off(struct drm_i915_private *dev_priv, enum pipe pipe) > > +void vlv_force_pll_off(struct drm_i915_private *i915, enum pipe pipe) > > { > > - if (IS_CHERRYVIEW(dev_priv)) > > - chv_disable_pll(dev_priv, pipe); > > + if (IS_CHERRYVIEW(i915)) > > + chv_disable_pll(i915, pipe); > > else > > - vlv_disable_pll(dev_priv, pipe); > > + vlv_disable_pll(i915, pipe); > > } > > static void i9xx_compute_dpll(struct intel_crtc *crtc, > > struct intel_crtc_state *crtc_state, > > struct dpll *reduced_clock) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > u32 dpll; > > struct dpll *clock = &crtc_state->dpll; > > @@ -7861,8 +7861,8 @@ static void i9xx_compute_dpll(struct intel_crtc *crtc, > > else > > dpll |= DPLLB_MODE_DAC_SERIAL; > > - if (IS_I945G(dev_priv) || IS_I945GM(dev_priv) || > > - IS_G33(dev_priv) || IS_PINEVIEW(dev_priv)) { > > + if (IS_I945G(i915) || IS_I945GM(i915) || > > + IS_G33(i915) || IS_PINEVIEW(i915)) { > > dpll |= (crtc_state->pixel_multiplier - 1) > > << SDVO_MULTIPLIER_SHIFT_HIRES; > > } > > @@ -7875,11 +7875,11 @@ static void i9xx_compute_dpll(struct intel_crtc *crtc, > > dpll |= DPLL_SDVO_HIGH_SPEED; > > /* compute bitmask from p1 value */ > > - if (IS_PINEVIEW(dev_priv)) > > + if (IS_PINEVIEW(i915)) > > dpll |= (1 << (clock->p1 - 1)) << DPLL_FPA01_P1_POST_DIV_SHIFT_PINEVIEW; > > else { > > dpll |= (1 << (clock->p1 - 1)) << DPLL_FPA01_P1_POST_DIV_SHIFT; > > - if (IS_G4X(dev_priv) && reduced_clock) > > + if (IS_G4X(i915) && reduced_clock) > > dpll |= (1 << (reduced_clock->p1 - 1)) << DPLL_FPA1_P1_POST_DIV_SHIFT; > > } > > switch (clock->p2) { > > @@ -7896,13 +7896,13 @@ static void i9xx_compute_dpll(struct intel_crtc *crtc, > > dpll |= DPLLB_LVDS_P2_CLOCK_DIV_14; > > break; > > } > > - if (INTEL_GEN(dev_priv) >= 4) > > + if (INTEL_GEN(i915) >= 4) > > dpll |= (6 << PLL_LOAD_PULSE_PHASE_SHIFT); > > if (crtc_state->sdvo_tv_clock) > > dpll |= PLL_REF_INPUT_TVCLKINBC; > > else if (intel_crtc_has_type(crtc_state, INTEL_OUTPUT_LVDS) && > > - intel_panel_use_ssc(dev_priv)) > > + intel_panel_use_ssc(i915)) > > dpll |= PLLB_REF_INPUT_SPREADSPECTRUMIN; > > else > > dpll |= PLL_REF_INPUT_DREFCLK; > > @@ -7910,7 +7910,7 @@ static void i9xx_compute_dpll(struct intel_crtc *crtc, > > dpll |= DPLL_VCO_ENABLE; > > crtc_state->dpll_hw_state.dpll = dpll; > > - if (INTEL_GEN(dev_priv) >= 4) { > > + if (INTEL_GEN(i915) >= 4) { > > u32 dpll_md = (crtc_state->pixel_multiplier - 1) > > << DPLL_MD_UDI_MULTIPLIER_SHIFT; > > crtc_state->dpll_hw_state.dpll_md = dpll_md; > > @@ -7922,7 +7922,7 @@ static void i8xx_compute_dpll(struct intel_crtc *crtc, > > struct dpll *reduced_clock) > > { > > struct drm_device *dev = crtc->base.dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > u32 dpll; > > struct dpll *clock = &crtc_state->dpll; > > @@ -7953,12 +7953,12 @@ static void i8xx_compute_dpll(struct intel_crtc *crtc, > > * both DPLLS. The spec says we should disable the DVO 2X clock > > * when not needed, but this seems to work fine in practice. > > */ > > - if (IS_I830(dev_priv) || > > + if (IS_I830(i915) || > > intel_crtc_has_type(crtc_state, INTEL_OUTPUT_DVO)) > > dpll |= DPLL_DVO_2X_MODE; > > if (intel_crtc_has_type(crtc_state, INTEL_OUTPUT_LVDS) && > > - intel_panel_use_ssc(dev_priv)) > > + intel_panel_use_ssc(i915)) > > dpll |= PLLB_REF_INPUT_SPREADSPECTRUMIN; > > else > > dpll |= PLL_REF_INPUT_DREFCLK; > > @@ -7970,7 +7970,7 @@ static void i8xx_compute_dpll(struct intel_crtc *crtc, > > static void intel_set_pipe_timings(const struct intel_crtc_state *crtc_state) > > { > > struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc); > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > enum pipe pipe = crtc->pipe; > > enum transcoder cpu_transcoder = crtc_state->cpu_transcoder; > > const struct drm_display_mode *adjusted_mode = &crtc_state->base.adjusted_mode; > > @@ -7996,7 +7996,7 @@ static void intel_set_pipe_timings(const struct intel_crtc_state *crtc_state) > > vsyncshift += adjusted_mode->crtc_htotal; > > } > > - if (INTEL_GEN(dev_priv) > 3) > > + if (INTEL_GEN(i915) > 3) > > I915_WRITE(VSYNCSHIFT(cpu_transcoder), vsyncshift); > > I915_WRITE(HTOTAL(cpu_transcoder), > > @@ -8023,7 +8023,7 @@ static void intel_set_pipe_timings(const struct intel_crtc_state *crtc_state) > > * programmed with the VTOTAL_EDP value. Same for VTOTAL_C. This is > > * documented on the DDI_FUNC_CTL register description, EDP Input Select > > * bits. */ > > - if (IS_HASWELL(dev_priv) && cpu_transcoder == TRANSCODER_EDP && > > + if (IS_HASWELL(i915) && cpu_transcoder == TRANSCODER_EDP && > > (pipe == PIPE_B || pipe == PIPE_C)) > > I915_WRITE(VTOTAL(pipe), I915_READ(VTOTAL(cpu_transcoder))); > > @@ -8032,7 +8032,7 @@ static void intel_set_pipe_timings(const struct intel_crtc_state *crtc_state) > > static void intel_set_pipe_src_size(const struct intel_crtc_state *crtc_state) > > { > > struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc); > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > enum pipe pipe = crtc->pipe; > > /* pipesrc controls the size that is scaled from, which should > > @@ -8047,7 +8047,7 @@ static void intel_get_pipe_timings(struct intel_crtc *crtc, > > struct intel_crtc_state *pipe_config) > > { > > struct drm_device *dev = crtc->base.dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > enum transcoder cpu_transcoder = pipe_config->cpu_transcoder; > > u32 tmp; > > @@ -8092,7 +8092,7 @@ static void intel_get_pipe_src_size(struct intel_crtc *crtc, > > struct intel_crtc_state *pipe_config) > > { > > struct drm_device *dev = crtc->base.dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > u32 tmp; > > tmp = I915_READ(PIPESRC(crtc->pipe)); > > @@ -8129,21 +8129,21 @@ void intel_mode_from_pipe_config(struct drm_display_mode *mode, > > static void i9xx_set_pipeconf(const struct intel_crtc_state *crtc_state) > > { > > struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc); > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > u32 pipeconf; > > pipeconf = 0; > > /* we keep both pipes enabled on 830 */ > > - if (IS_I830(dev_priv)) > > + if (IS_I830(i915)) > > pipeconf |= I915_READ(PIPECONF(crtc->pipe)) & PIPECONF_ENABLE; > > if (crtc_state->double_wide) > > pipeconf |= PIPECONF_DOUBLE_WIDE; > > /* only g4x and later have fancy bpc/dither controls */ > > - if (IS_G4X(dev_priv) || IS_VALLEYVIEW(dev_priv) || > > - IS_CHERRYVIEW(dev_priv)) { > > + if (IS_G4X(i915) || IS_VALLEYVIEW(i915) || > > + IS_CHERRYVIEW(i915)) { > > /* Bspec claims that we can't use dithering for 30bpp pipes. */ > > if (crtc_state->dither && crtc_state->pipe_bpp != 30) > > pipeconf |= PIPECONF_DITHER_EN | > > @@ -8166,7 +8166,7 @@ static void i9xx_set_pipeconf(const struct intel_crtc_state *crtc_state) > > } > > if (crtc_state->base.adjusted_mode.flags & DRM_MODE_FLAG_INTERLACE) { > > - if (INTEL_GEN(dev_priv) < 4 || > > + if (INTEL_GEN(i915) < 4 || > > intel_crtc_has_type(crtc_state, INTEL_OUTPUT_SDVO)) > > pipeconf |= PIPECONF_INTERLACE_W_FIELD_INDICATION; > > else > > @@ -8175,7 +8175,7 @@ static void i9xx_set_pipeconf(const struct intel_crtc_state *crtc_state) > > pipeconf |= PIPECONF_PROGRESSIVE; > > } > > - if ((IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) && > > + if ((IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) && > > crtc_state->limited_color_range) > > pipeconf |= PIPECONF_COLOR_RANGE_SELECT; > > @@ -8189,7 +8189,7 @@ static int i8xx_crtc_compute_clock(struct intel_crtc *crtc, > > struct intel_crtc_state *crtc_state) > > { > > struct drm_device *dev = crtc->base.dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > const struct intel_limit *limit; > > int refclk = 48000; > > @@ -8197,8 +8197,8 @@ static int i8xx_crtc_compute_clock(struct intel_crtc *crtc, > > sizeof(crtc_state->dpll_hw_state)); > > if (intel_crtc_has_type(crtc_state, INTEL_OUTPUT_LVDS)) { > > - if (intel_panel_use_ssc(dev_priv)) { > > - refclk = dev_priv->vbt.lvds_ssc_freq; > > + if (intel_panel_use_ssc(i915)) { > > + refclk = i915->vbt.lvds_ssc_freq; > > DRM_DEBUG_KMS("using SSC reference clock of %d kHz\n", refclk); > > } > > @@ -8224,7 +8224,7 @@ static int i8xx_crtc_compute_clock(struct intel_crtc *crtc, > > static int g4x_crtc_compute_clock(struct intel_crtc *crtc, > > struct intel_crtc_state *crtc_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > const struct intel_limit *limit; > > int refclk = 96000; > > @@ -8232,12 +8232,12 @@ static int g4x_crtc_compute_clock(struct intel_crtc *crtc, > > sizeof(crtc_state->dpll_hw_state)); > > if (intel_crtc_has_type(crtc_state, INTEL_OUTPUT_LVDS)) { > > - if (intel_panel_use_ssc(dev_priv)) { > > - refclk = dev_priv->vbt.lvds_ssc_freq; > > + if (intel_panel_use_ssc(i915)) { > > + refclk = i915->vbt.lvds_ssc_freq; > > DRM_DEBUG_KMS("using SSC reference clock of %d kHz\n", refclk); > > } > > - if (intel_is_dual_link_lvds(dev_priv)) > > + if (intel_is_dual_link_lvds(i915)) > > limit = &intel_limits_g4x_dual_channel_lvds; > > else > > limit = &intel_limits_g4x_single_channel_lvds; > > @@ -8267,7 +8267,7 @@ static int pnv_crtc_compute_clock(struct intel_crtc *crtc, > > struct intel_crtc_state *crtc_state) > > { > > struct drm_device *dev = crtc->base.dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > const struct intel_limit *limit; > > int refclk = 96000; > > @@ -8275,8 +8275,8 @@ static int pnv_crtc_compute_clock(struct intel_crtc *crtc, > > sizeof(crtc_state->dpll_hw_state)); > > if (intel_crtc_has_type(crtc_state, INTEL_OUTPUT_LVDS)) { > > - if (intel_panel_use_ssc(dev_priv)) { > > - refclk = dev_priv->vbt.lvds_ssc_freq; > > + if (intel_panel_use_ssc(i915)) { > > + refclk = i915->vbt.lvds_ssc_freq; > > DRM_DEBUG_KMS("using SSC reference clock of %d kHz\n", refclk); > > } > > @@ -8301,7 +8301,7 @@ static int i9xx_crtc_compute_clock(struct intel_crtc *crtc, > > struct intel_crtc_state *crtc_state) > > { > > struct drm_device *dev = crtc->base.dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > const struct intel_limit *limit; > > int refclk = 96000; > > @@ -8309,8 +8309,8 @@ static int i9xx_crtc_compute_clock(struct intel_crtc *crtc, > > sizeof(crtc_state->dpll_hw_state)); > > if (intel_crtc_has_type(crtc_state, INTEL_OUTPUT_LVDS)) { > > - if (intel_panel_use_ssc(dev_priv)) { > > - refclk = dev_priv->vbt.lvds_ssc_freq; > > + if (intel_panel_use_ssc(i915)) { > > + refclk = i915->vbt.lvds_ssc_freq; > > DRM_DEBUG_KMS("using SSC reference clock of %d kHz\n", refclk); > > } > > @@ -8373,22 +8373,22 @@ static int vlv_crtc_compute_clock(struct intel_crtc *crtc, > > return 0; > > } > > -static bool i9xx_has_pfit(struct drm_i915_private *dev_priv) > > +static bool i9xx_has_pfit(struct drm_i915_private *i915) > > { > > - if (IS_I830(dev_priv)) > > + if (IS_I830(i915)) > > return false; > > - return INTEL_GEN(dev_priv) >= 4 || > > - IS_PINEVIEW(dev_priv) || IS_MOBILE(dev_priv); > > + return INTEL_GEN(i915) >= 4 || > > + IS_PINEVIEW(i915) || IS_MOBILE(i915); > > } > > static void i9xx_get_pfit_config(struct intel_crtc *crtc, > > struct intel_crtc_state *pipe_config) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > u32 tmp; > > - if (!i9xx_has_pfit(dev_priv)) > > + if (!i9xx_has_pfit(i915)) > > return; > > tmp = I915_READ(PFIT_CONTROL); > > @@ -8396,7 +8396,7 @@ static void i9xx_get_pfit_config(struct intel_crtc *crtc, > > return; > > /* Check whether the pfit is attached to our pipe. */ > > - if (INTEL_GEN(dev_priv) < 4) { > > + if (INTEL_GEN(i915) < 4) { > > if (crtc->pipe != PIPE_B) > > return; > > } else { > > @@ -8412,7 +8412,7 @@ static void vlv_crtc_clock_get(struct intel_crtc *crtc, > > struct intel_crtc_state *pipe_config) > > { > > struct drm_device *dev = crtc->base.dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > int pipe = pipe_config->cpu_transcoder; > > struct dpll clock; > > u32 mdiv; > > @@ -8422,9 +8422,9 @@ static void vlv_crtc_clock_get(struct intel_crtc *crtc, > > if ((pipe_config->dpll_hw_state.dpll & DPLL_VCO_ENABLE) == 0) > > return; > > - vlv_dpio_get(dev_priv); > > - mdiv = vlv_dpio_read(dev_priv, pipe, VLV_PLL_DW3(pipe)); > > - vlv_dpio_put(dev_priv); > > + vlv_dpio_get(i915); > > + mdiv = vlv_dpio_read(i915, pipe, VLV_PLL_DW3(pipe)); > > + vlv_dpio_put(i915); > > clock.m1 = (mdiv >> DPIO_M1DIV_SHIFT) & 7; > > clock.m2 = mdiv & DPIO_M2DIV_MASK; > > @@ -8440,7 +8440,7 @@ i9xx_get_initial_plane_config(struct intel_crtc *crtc, > > struct intel_initial_plane_config *plane_config) > > { > > struct drm_device *dev = crtc->base.dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > struct intel_plane *plane = to_intel_plane(crtc->base.primary); > > enum i9xx_plane_id i9xx_plane = plane->i9xx_plane; > > enum pipe pipe; > > @@ -8467,7 +8467,7 @@ i9xx_get_initial_plane_config(struct intel_crtc *crtc, > > val = I915_READ(DSPCNTR(i9xx_plane)); > > - if (INTEL_GEN(dev_priv) >= 4) { > > + if (INTEL_GEN(i915) >= 4) { > > if (val & DISPPLANE_TILED) { > > plane_config->tiling = I915_TILING_X; > > fb->modifier = I915_FORMAT_MOD_X_TILED; > > @@ -8477,7 +8477,7 @@ i9xx_get_initial_plane_config(struct intel_crtc *crtc, > > plane_config->rotation = DRM_MODE_ROTATE_180; > > } > > - if (IS_CHERRYVIEW(dev_priv) && pipe == PIPE_B && > > + if (IS_CHERRYVIEW(i915) && pipe == PIPE_B && > > val & DISPPLANE_MIRROR) > > plane_config->rotation |= DRM_MODE_REFLECT_X; > > @@ -8485,10 +8485,10 @@ i9xx_get_initial_plane_config(struct intel_crtc *crtc, > > fourcc = i9xx_format_to_fourcc(pixel_format); > > fb->format = drm_format_info(fourcc); > > - if (IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv)) { > > + if (IS_HASWELL(i915) || IS_BROADWELL(i915)) { > > offset = I915_READ(DSPOFFSET(i9xx_plane)); > > base = I915_READ(DSPSURF(i9xx_plane)) & 0xfffff000; > > - } else if (INTEL_GEN(dev_priv) >= 4) { > > + } else if (INTEL_GEN(i915) >= 4) { > > if (plane_config->tiling) > > offset = I915_READ(DSPTILEOFF(i9xx_plane)); > > else > > @@ -8522,7 +8522,7 @@ static void chv_crtc_clock_get(struct intel_crtc *crtc, > > struct intel_crtc_state *pipe_config) > > { > > struct drm_device *dev = crtc->base.dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > int pipe = pipe_config->cpu_transcoder; > > enum dpio_channel port = vlv_pipe_to_channel(pipe); > > struct dpll clock; > > @@ -8533,13 +8533,13 @@ static void chv_crtc_clock_get(struct intel_crtc *crtc, > > if ((pipe_config->dpll_hw_state.dpll & DPLL_VCO_ENABLE) == 0) > > return; > > - vlv_dpio_get(dev_priv); > > - cmn_dw13 = vlv_dpio_read(dev_priv, pipe, CHV_CMN_DW13(port)); > > - pll_dw0 = vlv_dpio_read(dev_priv, pipe, CHV_PLL_DW0(port)); > > - pll_dw1 = vlv_dpio_read(dev_priv, pipe, CHV_PLL_DW1(port)); > > - pll_dw2 = vlv_dpio_read(dev_priv, pipe, CHV_PLL_DW2(port)); > > - pll_dw3 = vlv_dpio_read(dev_priv, pipe, CHV_PLL_DW3(port)); > > - vlv_dpio_put(dev_priv); > > + vlv_dpio_get(i915); > > + cmn_dw13 = vlv_dpio_read(i915, pipe, CHV_CMN_DW13(port)); > > + pll_dw0 = vlv_dpio_read(i915, pipe, CHV_PLL_DW0(port)); > > + pll_dw1 = vlv_dpio_read(i915, pipe, CHV_PLL_DW1(port)); > > + pll_dw2 = vlv_dpio_read(i915, pipe, CHV_PLL_DW2(port)); > > + pll_dw3 = vlv_dpio_read(i915, pipe, CHV_PLL_DW3(port)); > > + vlv_dpio_put(i915); > > clock.m1 = (pll_dw1 & 0x7) == DPIO_CHV_M1_DIV_BY_2 ? 2 : 0; > > clock.m2 = (pll_dw0 & 0xff) << 22; > > @@ -8555,12 +8555,12 @@ static void chv_crtc_clock_get(struct intel_crtc *crtc, > > static void intel_get_crtc_ycbcr_config(struct intel_crtc *crtc, > > struct intel_crtc_state *pipe_config) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > enum intel_output_format output = INTEL_OUTPUT_FORMAT_RGB; > > pipe_config->lspcon_downsampling = false; > > - if (IS_BROADWELL(dev_priv) || INTEL_GEN(dev_priv) >= 9) { > > + if (IS_BROADWELL(i915) || INTEL_GEN(i915) >= 9) { > > u32 tmp = I915_READ(PIPEMISC(crtc->pipe)); > > if (tmp & PIPEMISC_OUTPUT_COLORSPACE_YUV) { > > @@ -8571,8 +8571,8 @@ static void intel_get_crtc_ycbcr_config(struct intel_crtc *crtc, > > /* We support 4:2:0 in full blend mode only */ > > if (!blend) > > output = INTEL_OUTPUT_FORMAT_INVALID; > > - else if (!(IS_GEMINILAKE(dev_priv) || > > - INTEL_GEN(dev_priv) >= 10)) > > + else if (!(IS_GEMINILAKE(i915) || > > + INTEL_GEN(i915) >= 10)) > > output = INTEL_OUTPUT_FORMAT_INVALID; > > else > > output = INTEL_OUTPUT_FORMAT_YCBCR420; > > @@ -8599,7 +8599,7 @@ static void i9xx_get_pipe_color_config(struct intel_crtc_state *crtc_state) > > { > > struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc); > > struct intel_plane *plane = to_intel_plane(crtc->base.primary); > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > enum i9xx_plane_id i9xx_plane = plane->i9xx_plane; > > u32 tmp; > > @@ -8608,7 +8608,7 @@ static void i9xx_get_pipe_color_config(struct intel_crtc_state *crtc_state) > > if (tmp & DISPPLANE_GAMMA_ENABLE) > > crtc_state->gamma_enable = true; > > - if (!HAS_GMCH(dev_priv) && > > + if (!HAS_GMCH(i915) && > > tmp & DISPPLANE_PIPE_CSC_ENABLE) > > crtc_state->csc_enable = true; > > } > > @@ -8616,14 +8616,14 @@ static void i9xx_get_pipe_color_config(struct intel_crtc_state *crtc_state) > > static bool i9xx_get_pipe_config(struct intel_crtc *crtc, > > struct intel_crtc_state *pipe_config) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > enum intel_display_power_domain power_domain; > > intel_wakeref_t wakeref; > > u32 tmp; > > bool ret; > > power_domain = POWER_DOMAIN_PIPE(crtc->pipe); > > - wakeref = intel_display_power_get_if_enabled(dev_priv, power_domain); > > + wakeref = intel_display_power_get_if_enabled(i915, power_domain); > > if (!wakeref) > > return false; > > @@ -8637,8 +8637,8 @@ static bool i9xx_get_pipe_config(struct intel_crtc *crtc, > > if (!(tmp & PIPECONF_ENABLE)) > > goto out; > > - if (IS_G4X(dev_priv) || IS_VALLEYVIEW(dev_priv) || > > - IS_CHERRYVIEW(dev_priv)) { > > + if (IS_G4X(i915) || IS_VALLEYVIEW(i915) || > > + IS_CHERRYVIEW(i915)) { > > switch (tmp & PIPECONF_BPC_MASK) { > > case PIPECONF_6BPC: > > pipe_config->pipe_bpp = 18; > > @@ -8654,20 +8654,20 @@ static bool i9xx_get_pipe_config(struct intel_crtc *crtc, > > } > > } > > - if ((IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) && > > + if ((IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) && > > (tmp & PIPECONF_COLOR_RANGE_SELECT)) > > pipe_config->limited_color_range = true; > > pipe_config->gamma_mode = (tmp & PIPECONF_GAMMA_MODE_MASK_I9XX) >> > > PIPECONF_GAMMA_MODE_SHIFT; > > - if (IS_CHERRYVIEW(dev_priv)) > > + if (IS_CHERRYVIEW(i915)) > > pipe_config->cgm_mode = I915_READ(CGM_PIPE_MODE(crtc->pipe)); > > i9xx_get_pipe_color_config(pipe_config); > > intel_color_get_config(pipe_config); > > - if (INTEL_GEN(dev_priv) < 4) > > + if (INTEL_GEN(i915) < 4) > > pipe_config->double_wide = tmp & PIPECONF_DOUBLE_WIDE; > > intel_get_pipe_timings(crtc, pipe_config); > > @@ -8675,18 +8675,18 @@ static bool i9xx_get_pipe_config(struct intel_crtc *crtc, > > i9xx_get_pfit_config(crtc, pipe_config); > > - if (INTEL_GEN(dev_priv) >= 4) { > > + if (INTEL_GEN(i915) >= 4) { > > /* No way to read it out on pipes B and C */ > > - if (IS_CHERRYVIEW(dev_priv) && crtc->pipe != PIPE_A) > > - tmp = dev_priv->chv_dpll_md[crtc->pipe]; > > + if (IS_CHERRYVIEW(i915) && crtc->pipe != PIPE_A) > > + tmp = i915->chv_dpll_md[crtc->pipe]; > > else > > tmp = I915_READ(DPLL_MD(crtc->pipe)); > > pipe_config->pixel_multiplier = > > ((tmp & DPLL_MD_UDI_MULTIPLIER_MASK) > > >> DPLL_MD_UDI_MULTIPLIER_SHIFT) + 1; > > pipe_config->dpll_hw_state.dpll_md = tmp; > > - } else if (IS_I945G(dev_priv) || IS_I945GM(dev_priv) || > > - IS_G33(dev_priv) || IS_PINEVIEW(dev_priv)) { > > + } else if (IS_I945G(i915) || IS_I945GM(i915) || > > + IS_G33(i915) || IS_PINEVIEW(i915)) { > > tmp = I915_READ(DPLL(crtc->pipe)); > > pipe_config->pixel_multiplier = > > ((tmp & SDVO_MULTIPLIER_MASK) > > @@ -8698,7 +8698,7 @@ static bool i9xx_get_pipe_config(struct intel_crtc *crtc, > > pipe_config->pixel_multiplier = 1; > > } > > pipe_config->dpll_hw_state.dpll = I915_READ(DPLL(crtc->pipe)); > > - if (!IS_VALLEYVIEW(dev_priv) && !IS_CHERRYVIEW(dev_priv)) { > > + if (!IS_VALLEYVIEW(i915) && !IS_CHERRYVIEW(i915)) { > > pipe_config->dpll_hw_state.fp0 = I915_READ(FP0(crtc->pipe)); > > pipe_config->dpll_hw_state.fp1 = I915_READ(FP1(crtc->pipe)); > > } else { > > @@ -8708,9 +8708,9 @@ static bool i9xx_get_pipe_config(struct intel_crtc *crtc, > > DPLL_PORTB_READY_MASK); > > } > > - if (IS_CHERRYVIEW(dev_priv)) > > + if (IS_CHERRYVIEW(i915)) > > chv_crtc_clock_get(crtc, pipe_config); > > - else if (IS_VALLEYVIEW(dev_priv)) > > + else if (IS_VALLEYVIEW(i915)) > > vlv_crtc_clock_get(crtc, pipe_config); > > else > > i9xx_crtc_clock_get(crtc, pipe_config); > > @@ -8726,12 +8726,12 @@ static bool i9xx_get_pipe_config(struct intel_crtc *crtc, > > ret = true; > > out: > > - intel_display_power_put(dev_priv, power_domain, wakeref); > > + intel_display_power_put(i915, power_domain, wakeref); > > return ret; > > } > > -static void ironlake_init_pch_refclk(struct drm_i915_private *dev_priv) > > +static void ironlake_init_pch_refclk(struct drm_i915_private *i915) > > { > > struct intel_encoder *encoder; > > int i; > > @@ -8744,7 +8744,7 @@ static void ironlake_init_pch_refclk(struct drm_i915_private *dev_priv) > > bool using_ssc_source = false; > > /* We need to take the global config into account */ > > - for_each_intel_encoder(&dev_priv->drm, encoder) { > > + for_each_intel_encoder(&i915->drm, encoder) { > > switch (encoder->type) { > > case INTEL_OUTPUT_LVDS: > > has_panel = true; > > @@ -8760,8 +8760,8 @@ static void ironlake_init_pch_refclk(struct drm_i915_private *dev_priv) > > } > > } > > - if (HAS_PCH_IBX(dev_priv)) { > > - has_ck505 = dev_priv->vbt.display_clock_mode; > > + if (HAS_PCH_IBX(i915)) { > > + has_ck505 = i915->vbt.display_clock_mode; > > can_ssc = has_ck505; > > } else { > > has_ck505 = false; > > @@ -8769,7 +8769,7 @@ static void ironlake_init_pch_refclk(struct drm_i915_private *dev_priv) > > } > > /* Check if any DPLLs are using the SSC source */ > > - for (i = 0; i < dev_priv->num_shared_dpll; i++) { > > + for (i = 0; i < i915->num_shared_dpll; i++) { > > u32 temp = I915_READ(PCH_DPLL(i)); > > if (!(temp & DPLL_VCO_ENABLE)) > > @@ -8810,11 +8810,11 @@ static void ironlake_init_pch_refclk(struct drm_i915_private *dev_priv) > > if (has_panel) { > > final |= DREF_SSC_SOURCE_ENABLE; > > - if (intel_panel_use_ssc(dev_priv) && can_ssc) > > + if (intel_panel_use_ssc(i915) && can_ssc) > > final |= DREF_SSC1_ENABLE; > > if (has_cpu_edp) { > > - if (intel_panel_use_ssc(dev_priv) && can_ssc) > > + if (intel_panel_use_ssc(i915) && can_ssc) > > final |= DREF_CPU_SOURCE_OUTPUT_DOWNSPREAD; > > else > > final |= DREF_CPU_SOURCE_OUTPUT_NONSPREAD; > > @@ -8841,7 +8841,7 @@ static void ironlake_init_pch_refclk(struct drm_i915_private *dev_priv) > > val |= DREF_SSC_SOURCE_ENABLE; > > /* SSC must be turned on before enabling the CPU output */ > > - if (intel_panel_use_ssc(dev_priv) && can_ssc) { > > + if (intel_panel_use_ssc(i915) && can_ssc) { > > DRM_DEBUG_KMS("Using SSC on panel\n"); > > val |= DREF_SSC1_ENABLE; > > } else > > @@ -8856,7 +8856,7 @@ static void ironlake_init_pch_refclk(struct drm_i915_private *dev_priv) > > /* Enable CPU source on CPU attached eDP */ > > if (has_cpu_edp) { > > - if (intel_panel_use_ssc(dev_priv) && can_ssc) { > > + if (intel_panel_use_ssc(i915) && can_ssc) { > > DRM_DEBUG_KMS("Using SSC on eDP\n"); > > val |= DREF_CPU_SOURCE_OUTPUT_DOWNSPREAD; > > } else > > @@ -8898,7 +8898,7 @@ static void ironlake_init_pch_refclk(struct drm_i915_private *dev_priv) > > BUG_ON(val != final); > > } > > -static void lpt_reset_fdi_mphy(struct drm_i915_private *dev_priv) > > +static void lpt_reset_fdi_mphy(struct drm_i915_private *i915) > > { > > u32 tmp; > > @@ -8920,78 +8920,78 @@ static void lpt_reset_fdi_mphy(struct drm_i915_private *dev_priv) > > } > > /* WaMPhyProgramming:hsw */ > > -static void lpt_program_fdi_mphy(struct drm_i915_private *dev_priv) > > +static void lpt_program_fdi_mphy(struct drm_i915_private *i915) > > { > > u32 tmp; > > - tmp = intel_sbi_read(dev_priv, 0x8008, SBI_MPHY); > > + tmp = intel_sbi_read(i915, 0x8008, SBI_MPHY); > > tmp &= ~(0xFF << 24); > > tmp |= (0x12 << 24); > > - intel_sbi_write(dev_priv, 0x8008, tmp, SBI_MPHY); > > + intel_sbi_write(i915, 0x8008, tmp, SBI_MPHY); > > - tmp = intel_sbi_read(dev_priv, 0x2008, SBI_MPHY); > > + tmp = intel_sbi_read(i915, 0x2008, SBI_MPHY); > > tmp |= (1 << 11); > > - intel_sbi_write(dev_priv, 0x2008, tmp, SBI_MPHY); > > + intel_sbi_write(i915, 0x2008, tmp, SBI_MPHY); > > - tmp = intel_sbi_read(dev_priv, 0x2108, SBI_MPHY); > > + tmp = intel_sbi_read(i915, 0x2108, SBI_MPHY); > > tmp |= (1 << 11); > > - intel_sbi_write(dev_priv, 0x2108, tmp, SBI_MPHY); > > + intel_sbi_write(i915, 0x2108, tmp, SBI_MPHY); > > - tmp = intel_sbi_read(dev_priv, 0x206C, SBI_MPHY); > > + tmp = intel_sbi_read(i915, 0x206C, SBI_MPHY); > > tmp |= (1 << 24) | (1 << 21) | (1 << 18); > > - intel_sbi_write(dev_priv, 0x206C, tmp, SBI_MPHY); > > + intel_sbi_write(i915, 0x206C, tmp, SBI_MPHY); > > - tmp = intel_sbi_read(dev_priv, 0x216C, SBI_MPHY); > > + tmp = intel_sbi_read(i915, 0x216C, SBI_MPHY); > > tmp |= (1 << 24) | (1 << 21) | (1 << 18); > > - intel_sbi_write(dev_priv, 0x216C, tmp, SBI_MPHY); > > + intel_sbi_write(i915, 0x216C, tmp, SBI_MPHY); > > - tmp = intel_sbi_read(dev_priv, 0x2080, SBI_MPHY); > > + tmp = intel_sbi_read(i915, 0x2080, SBI_MPHY); > > tmp &= ~(7 << 13); > > tmp |= (5 << 13); > > - intel_sbi_write(dev_priv, 0x2080, tmp, SBI_MPHY); > > + intel_sbi_write(i915, 0x2080, tmp, SBI_MPHY); > > - tmp = intel_sbi_read(dev_priv, 0x2180, SBI_MPHY); > > + tmp = intel_sbi_read(i915, 0x2180, SBI_MPHY); > > tmp &= ~(7 << 13); > > tmp |= (5 << 13); > > - intel_sbi_write(dev_priv, 0x2180, tmp, SBI_MPHY); > > + intel_sbi_write(i915, 0x2180, tmp, SBI_MPHY); > > - tmp = intel_sbi_read(dev_priv, 0x208C, SBI_MPHY); > > + tmp = intel_sbi_read(i915, 0x208C, SBI_MPHY); > > tmp &= ~0xFF; > > tmp |= 0x1C; > > - intel_sbi_write(dev_priv, 0x208C, tmp, SBI_MPHY); > > + intel_sbi_write(i915, 0x208C, tmp, SBI_MPHY); > > - tmp = intel_sbi_read(dev_priv, 0x218C, SBI_MPHY); > > + tmp = intel_sbi_read(i915, 0x218C, SBI_MPHY); > > tmp &= ~0xFF; > > tmp |= 0x1C; > > - intel_sbi_write(dev_priv, 0x218C, tmp, SBI_MPHY); > > + intel_sbi_write(i915, 0x218C, tmp, SBI_MPHY); > > - tmp = intel_sbi_read(dev_priv, 0x2098, SBI_MPHY); > > + tmp = intel_sbi_read(i915, 0x2098, SBI_MPHY); > > tmp &= ~(0xFF << 16); > > tmp |= (0x1C << 16); > > - intel_sbi_write(dev_priv, 0x2098, tmp, SBI_MPHY); > > + intel_sbi_write(i915, 0x2098, tmp, SBI_MPHY); > > - tmp = intel_sbi_read(dev_priv, 0x2198, SBI_MPHY); > > + tmp = intel_sbi_read(i915, 0x2198, SBI_MPHY); > > tmp &= ~(0xFF << 16); > > tmp |= (0x1C << 16); > > - intel_sbi_write(dev_priv, 0x2198, tmp, SBI_MPHY); > > + intel_sbi_write(i915, 0x2198, tmp, SBI_MPHY); > > - tmp = intel_sbi_read(dev_priv, 0x20C4, SBI_MPHY); > > + tmp = intel_sbi_read(i915, 0x20C4, SBI_MPHY); > > tmp |= (1 << 27); > > - intel_sbi_write(dev_priv, 0x20C4, tmp, SBI_MPHY); > > + intel_sbi_write(i915, 0x20C4, tmp, SBI_MPHY); > > - tmp = intel_sbi_read(dev_priv, 0x21C4, SBI_MPHY); > > + tmp = intel_sbi_read(i915, 0x21C4, SBI_MPHY); > > tmp |= (1 << 27); > > - intel_sbi_write(dev_priv, 0x21C4, tmp, SBI_MPHY); > > + intel_sbi_write(i915, 0x21C4, tmp, SBI_MPHY); > > - tmp = intel_sbi_read(dev_priv, 0x20EC, SBI_MPHY); > > + tmp = intel_sbi_read(i915, 0x20EC, SBI_MPHY); > > tmp &= ~(0xF << 28); > > tmp |= (4 << 28); > > - intel_sbi_write(dev_priv, 0x20EC, tmp, SBI_MPHY); > > + intel_sbi_write(i915, 0x20EC, tmp, SBI_MPHY); > > - tmp = intel_sbi_read(dev_priv, 0x21EC, SBI_MPHY); > > + tmp = intel_sbi_read(i915, 0x21EC, SBI_MPHY); > > tmp &= ~(0xF << 28); > > tmp |= (4 << 28); > > - intel_sbi_write(dev_priv, 0x21EC, tmp, SBI_MPHY); > > + intel_sbi_write(i915, 0x21EC, tmp, SBI_MPHY); > > } > > /* Implements 3 different sequences from BSpec chapter "Display iCLK > > @@ -9000,69 +9000,69 @@ static void lpt_program_fdi_mphy(struct drm_i915_private *dev_priv) > > * - Sequence to enable CLKOUT_DP without spread > > * - Sequence to enable CLKOUT_DP for FDI usage and configure PCH FDI I/O > > */ > > -static void lpt_enable_clkout_dp(struct drm_i915_private *dev_priv, > > +static void lpt_enable_clkout_dp(struct drm_i915_private *i915, > > bool with_spread, bool with_fdi) > > { > > u32 reg, tmp; > > if (WARN(with_fdi && !with_spread, "FDI requires downspread\n")) > > with_spread = true; > > - if (WARN(HAS_PCH_LPT_LP(dev_priv) && > > + if (WARN(HAS_PCH_LPT_LP(i915) && > > with_fdi, "LP PCH doesn't have FDI\n")) > > with_fdi = false; > > - mutex_lock(&dev_priv->sb_lock); > > + mutex_lock(&i915->sb_lock); > > - tmp = intel_sbi_read(dev_priv, SBI_SSCCTL, SBI_ICLK); > > + tmp = intel_sbi_read(i915, SBI_SSCCTL, SBI_ICLK); > > tmp &= ~SBI_SSCCTL_DISABLE; > > tmp |= SBI_SSCCTL_PATHALT; > > - intel_sbi_write(dev_priv, SBI_SSCCTL, tmp, SBI_ICLK); > > + intel_sbi_write(i915, SBI_SSCCTL, tmp, SBI_ICLK); > > udelay(24); > > if (with_spread) { > > - tmp = intel_sbi_read(dev_priv, SBI_SSCCTL, SBI_ICLK); > > + tmp = intel_sbi_read(i915, SBI_SSCCTL, SBI_ICLK); > > tmp &= ~SBI_SSCCTL_PATHALT; > > - intel_sbi_write(dev_priv, SBI_SSCCTL, tmp, SBI_ICLK); > > + intel_sbi_write(i915, SBI_SSCCTL, tmp, SBI_ICLK); > > if (with_fdi) { > > - lpt_reset_fdi_mphy(dev_priv); > > - lpt_program_fdi_mphy(dev_priv); > > + lpt_reset_fdi_mphy(i915); > > + lpt_program_fdi_mphy(i915); > > } > > } > > - reg = HAS_PCH_LPT_LP(dev_priv) ? SBI_GEN0 : SBI_DBUFF0; > > - tmp = intel_sbi_read(dev_priv, reg, SBI_ICLK); > > + reg = HAS_PCH_LPT_LP(i915) ? SBI_GEN0 : SBI_DBUFF0; > > + tmp = intel_sbi_read(i915, reg, SBI_ICLK); > > tmp |= SBI_GEN0_CFG_BUFFENABLE_DISABLE; > > - intel_sbi_write(dev_priv, reg, tmp, SBI_ICLK); > > + intel_sbi_write(i915, reg, tmp, SBI_ICLK); > > - mutex_unlock(&dev_priv->sb_lock); > > + mutex_unlock(&i915->sb_lock); > > } > > /* Sequence to disable CLKOUT_DP */ > > -void lpt_disable_clkout_dp(struct drm_i915_private *dev_priv) > > +void lpt_disable_clkout_dp(struct drm_i915_private *i915) > > { > > u32 reg, tmp; > > - mutex_lock(&dev_priv->sb_lock); > > + mutex_lock(&i915->sb_lock); > > - reg = HAS_PCH_LPT_LP(dev_priv) ? SBI_GEN0 : SBI_DBUFF0; > > - tmp = intel_sbi_read(dev_priv, reg, SBI_ICLK); > > + reg = HAS_PCH_LPT_LP(i915) ? SBI_GEN0 : SBI_DBUFF0; > > + tmp = intel_sbi_read(i915, reg, SBI_ICLK); > > tmp &= ~SBI_GEN0_CFG_BUFFENABLE_DISABLE; > > - intel_sbi_write(dev_priv, reg, tmp, SBI_ICLK); > > + intel_sbi_write(i915, reg, tmp, SBI_ICLK); > > - tmp = intel_sbi_read(dev_priv, SBI_SSCCTL, SBI_ICLK); > > + tmp = intel_sbi_read(i915, SBI_SSCCTL, SBI_ICLK); > > if (!(tmp & SBI_SSCCTL_DISABLE)) { > > if (!(tmp & SBI_SSCCTL_PATHALT)) { > > tmp |= SBI_SSCCTL_PATHALT; > > - intel_sbi_write(dev_priv, SBI_SSCCTL, tmp, SBI_ICLK); > > + intel_sbi_write(i915, SBI_SSCCTL, tmp, SBI_ICLK); > > udelay(32); > > } > > tmp |= SBI_SSCCTL_DISABLE; > > - intel_sbi_write(dev_priv, SBI_SSCCTL, tmp, SBI_ICLK); > > + intel_sbi_write(i915, SBI_SSCCTL, tmp, SBI_ICLK); > > } > > - mutex_unlock(&dev_priv->sb_lock); > > + mutex_unlock(&i915->sb_lock); > > } > > #define BEND_IDX(steps) ((50 + (steps)) / 5) > > @@ -9097,7 +9097,7 @@ static const u16 sscdivintphase[] = { > > * < 0 slow down the clock, > 0 speed up the clock, 0 == no bend (135MHz) > > * change in clock period = -(steps / 10) * 5.787 ps > > */ > > -static void lpt_bend_clkout_dp(struct drm_i915_private *dev_priv, int steps) > > +static void lpt_bend_clkout_dp(struct drm_i915_private *i915, int steps) > > { > > u32 tmp; > > int idx = BEND_IDX(steps); > > @@ -9108,25 +9108,25 @@ static void lpt_bend_clkout_dp(struct drm_i915_private *dev_priv, int steps) > > if (WARN_ON(idx >= ARRAY_SIZE(sscdivintphase))) > > return; > > - mutex_lock(&dev_priv->sb_lock); > > + mutex_lock(&i915->sb_lock); > > if (steps % 10 != 0) > > tmp = 0xAAAAAAAB; > > else > > tmp = 0x00000000; > > - intel_sbi_write(dev_priv, SBI_SSCDITHPHASE, tmp, SBI_ICLK); > > + intel_sbi_write(i915, SBI_SSCDITHPHASE, tmp, SBI_ICLK); > > - tmp = intel_sbi_read(dev_priv, SBI_SSCDIVINTPHASE, SBI_ICLK); > > + tmp = intel_sbi_read(i915, SBI_SSCDIVINTPHASE, SBI_ICLK); > > tmp &= 0xffff0000; > > tmp |= sscdivintphase[idx]; > > - intel_sbi_write(dev_priv, SBI_SSCDIVINTPHASE, tmp, SBI_ICLK); > > + intel_sbi_write(i915, SBI_SSCDIVINTPHASE, tmp, SBI_ICLK); > > - mutex_unlock(&dev_priv->sb_lock); > > + mutex_unlock(&i915->sb_lock); > > } > > #undef BEND_IDX > > -static bool spll_uses_pch_ssc(struct drm_i915_private *dev_priv) > > +static bool spll_uses_pch_ssc(struct drm_i915_private *i915) > > { > > u32 fuse_strap = I915_READ(FUSE_STRAP); > > u32 ctl = I915_READ(SPLL_CTL); > > @@ -9138,14 +9138,14 @@ static bool spll_uses_pch_ssc(struct drm_i915_private *dev_priv) > > (fuse_strap & HSW_CPU_SSC_ENABLE) == 0) > > return true; > > - if (IS_BROADWELL(dev_priv) && > > + if (IS_BROADWELL(i915) && > > (ctl & SPLL_REF_MASK) == SPLL_REF_PCH_SSC_BDW) > > return true; > > return false; > > } > > -static bool wrpll_uses_pch_ssc(struct drm_i915_private *dev_priv, > > +static bool wrpll_uses_pch_ssc(struct drm_i915_private *i915, > > enum intel_dpll_id id) > > { > > u32 fuse_strap = I915_READ(FUSE_STRAP); > > @@ -9157,7 +9157,7 @@ static bool wrpll_uses_pch_ssc(struct drm_i915_private *dev_priv, > > if ((ctl & WRPLL_REF_MASK) == WRPLL_REF_PCH_SSC) > > return true; > > - if ((IS_BROADWELL(dev_priv) || IS_HSW_ULT(dev_priv)) && > > + if ((IS_BROADWELL(i915) || IS_HSW_ULT(i915)) && > > (ctl & WRPLL_REF_MASK) == WRPLL_REF_MUXED_SSC_BDW && > > (fuse_strap & HSW_CPU_SSC_ENABLE) == 0) > > return true; > > @@ -9165,13 +9165,13 @@ static bool wrpll_uses_pch_ssc(struct drm_i915_private *dev_priv, > > return false; > > } > > -static void lpt_init_pch_refclk(struct drm_i915_private *dev_priv) > > +static void lpt_init_pch_refclk(struct drm_i915_private *i915) > > { > > struct intel_encoder *encoder; > > bool pch_ssc_in_use = false; > > bool has_fdi = false; > > - for_each_intel_encoder(&dev_priv->drm, encoder) { > > + for_each_intel_encoder(&i915->drm, encoder) { > > switch (encoder->type) { > > case INTEL_OUTPUT_ANALOG: > > has_fdi = true; > > @@ -9196,17 +9196,17 @@ static void lpt_init_pch_refclk(struct drm_i915_private *dev_priv) > > * clock hierarchy. That would also allow us to do > > * clock bending finally. > > */ > > - if (spll_uses_pch_ssc(dev_priv)) { > > + if (spll_uses_pch_ssc(i915)) { > > DRM_DEBUG_KMS("SPLL using PCH SSC\n"); > > pch_ssc_in_use = true; > > } > > - if (wrpll_uses_pch_ssc(dev_priv, DPLL_ID_WRPLL1)) { > > + if (wrpll_uses_pch_ssc(i915, DPLL_ID_WRPLL1)) { > > DRM_DEBUG_KMS("WRPLL1 using PCH SSC\n"); > > pch_ssc_in_use = true; > > } > > - if (wrpll_uses_pch_ssc(dev_priv, DPLL_ID_WRPLL2)) { > > + if (wrpll_uses_pch_ssc(i915, DPLL_ID_WRPLL2)) { > > DRM_DEBUG_KMS("WRPLL2 using PCH SSC\n"); > > pch_ssc_in_use = true; > > } > > @@ -9215,28 +9215,28 @@ static void lpt_init_pch_refclk(struct drm_i915_private *dev_priv) > > return; > > if (has_fdi) { > > - lpt_bend_clkout_dp(dev_priv, 0); > > - lpt_enable_clkout_dp(dev_priv, true, true); > > + lpt_bend_clkout_dp(i915, 0); > > + lpt_enable_clkout_dp(i915, true, true); > > } else { > > - lpt_disable_clkout_dp(dev_priv); > > + lpt_disable_clkout_dp(i915); > > } > > } > > /* > > * Initialize reference clocks when the driver loads > > */ > > -void intel_init_pch_refclk(struct drm_i915_private *dev_priv) > > +void intel_init_pch_refclk(struct drm_i915_private *i915) > > { > > - if (HAS_PCH_IBX(dev_priv) || HAS_PCH_CPT(dev_priv)) > > - ironlake_init_pch_refclk(dev_priv); > > - else if (HAS_PCH_LPT(dev_priv)) > > - lpt_init_pch_refclk(dev_priv); > > + if (HAS_PCH_IBX(i915) || HAS_PCH_CPT(i915)) > > + ironlake_init_pch_refclk(i915); > > + else if (HAS_PCH_LPT(i915)) > > + lpt_init_pch_refclk(i915); > > } > > static void ironlake_set_pipeconf(const struct intel_crtc_state *crtc_state) > > { > > struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc); > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > enum pipe pipe = crtc->pipe; > > u32 val; > > @@ -9280,11 +9280,11 @@ static void ironlake_set_pipeconf(const struct intel_crtc_state *crtc_state) > > static void haswell_set_pipeconf(const struct intel_crtc_state *crtc_state) > > { > > struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc); > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > enum transcoder cpu_transcoder = crtc_state->cpu_transcoder; > > u32 val = 0; > > - if (IS_HASWELL(dev_priv) && crtc_state->dither) > > + if (IS_HASWELL(i915) && crtc_state->dither) > > val |= (PIPECONF_DITHER_EN | PIPECONF_DITHER_TYPE_SP); > > if (crtc_state->base.adjusted_mode.flags & DRM_MODE_FLAG_INTERLACE) > > @@ -9299,7 +9299,7 @@ static void haswell_set_pipeconf(const struct intel_crtc_state *crtc_state) > > static void bdw_set_pipemisc(const struct intel_crtc_state *crtc_state) > > { > > struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc); > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > u32 val = 0; > > switch (crtc_state->pipe_bpp) { > > @@ -9331,7 +9331,7 @@ static void bdw_set_pipemisc(const struct intel_crtc_state *crtc_state) > > val |= PIPEMISC_YUV420_ENABLE | > > PIPEMISC_YUV420_MODE_FULL_BLEND; > > - if (INTEL_GEN(dev_priv) >= 11 && > > + if (INTEL_GEN(i915) >= 11 && > > (crtc_state->active_planes & ~(icl_hdr_plane_mask() | > > BIT(PLANE_CURSOR))) == 0) > > val |= PIPEMISC_HDR_MODE_PRECISION; > > @@ -9341,7 +9341,7 @@ static void bdw_set_pipemisc(const struct intel_crtc_state *crtc_state) > > int bdw_get_pipemisc_bpp(struct intel_crtc *crtc) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > u32 tmp; > > tmp = I915_READ(PIPEMISC(crtc->pipe)); > > @@ -9381,17 +9381,17 @@ static void ironlake_compute_dpll(struct intel_crtc *crtc, > > struct intel_crtc_state *crtc_state, > > struct dpll *reduced_clock) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > u32 dpll, fp, fp2; > > int factor; > > /* Enable autotuning of the PLL clock (if permissible) */ > > factor = 21; > > if (intel_crtc_has_type(crtc_state, INTEL_OUTPUT_LVDS)) { > > - if ((intel_panel_use_ssc(dev_priv) && > > - dev_priv->vbt.lvds_ssc_freq == 100000) || > > - (HAS_PCH_IBX(dev_priv) && > > - intel_is_dual_link_lvds(dev_priv))) > > + if ((intel_panel_use_ssc(i915) && > > + i915->vbt.lvds_ssc_freq == 100000) || > > + (HAS_PCH_IBX(i915) && > > + intel_is_dual_link_lvds(i915))) > > factor = 25; > > } else if (crtc_state->sdvo_tv_clock) { > > factor = 20; > > @@ -9442,7 +9442,7 @@ static void ironlake_compute_dpll(struct intel_crtc *crtc, > > * clear if it''s a win or loss power wise. No point in doing > > * this on ILK at all since it has a fixed DPLL<->pipe mapping. > > */ > > - if (INTEL_INFO(dev_priv)->num_pipes == 3 && > > + if (INTEL_INFO(i915)->num_pipes == 3 && > > intel_crtc_has_type(crtc_state, INTEL_OUTPUT_ANALOG)) > > dpll |= DPLL_SDVO_HIGH_SPEED; > > @@ -9467,7 +9467,7 @@ static void ironlake_compute_dpll(struct intel_crtc *crtc, > > } > > if (intel_crtc_has_type(crtc_state, INTEL_OUTPUT_LVDS) && > > - intel_panel_use_ssc(dev_priv)) > > + intel_panel_use_ssc(i915)) > > dpll |= PLLB_REF_INPUT_SPREADSPECTRUMIN; > > else > > dpll |= PLL_REF_INPUT_DREFCLK; > > @@ -9482,7 +9482,7 @@ static void ironlake_compute_dpll(struct intel_crtc *crtc, > > static int ironlake_crtc_compute_clock(struct intel_crtc *crtc, > > struct intel_crtc_state *crtc_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > const struct intel_limit *limit; > > int refclk = 120000; > > @@ -9494,13 +9494,13 @@ static int ironlake_crtc_compute_clock(struct intel_crtc *crtc, > > return 0; > > if (intel_crtc_has_type(crtc_state, INTEL_OUTPUT_LVDS)) { > > - if (intel_panel_use_ssc(dev_priv)) { > > + if (intel_panel_use_ssc(i915)) { > > DRM_DEBUG_KMS("using SSC reference clock of %d kHz\n", > > - dev_priv->vbt.lvds_ssc_freq); > > - refclk = dev_priv->vbt.lvds_ssc_freq; > > + i915->vbt.lvds_ssc_freq); > > + refclk = i915->vbt.lvds_ssc_freq; > > } > > - if (intel_is_dual_link_lvds(dev_priv)) { > > + if (intel_is_dual_link_lvds(i915)) { > > if (refclk == 100000) > > limit = &intel_limits_ironlake_dual_lvds_100m; > > else > > @@ -9537,7 +9537,7 @@ static void intel_pch_transcoder_get_m_n(struct intel_crtc *crtc, > > struct intel_link_m_n *m_n) > > { > > struct drm_device *dev = crtc->base.dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > enum pipe pipe = crtc->pipe; > > m_n->link_m = I915_READ(PCH_TRANS_LINK_M1(pipe)); > > @@ -9554,10 +9554,10 @@ static void intel_cpu_transcoder_get_m_n(struct intel_crtc *crtc, > > struct intel_link_m_n *m_n, > > struct intel_link_m_n *m2_n2) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > enum pipe pipe = crtc->pipe; > > - if (INTEL_GEN(dev_priv) >= 5) { > > + if (INTEL_GEN(i915) >= 5) { > > m_n->link_m = I915_READ(PIPE_LINK_M1(transcoder)); > > m_n->link_n = I915_READ(PIPE_LINK_N1(transcoder)); > > m_n->gmch_m = I915_READ(PIPE_DATA_M1(transcoder)) > > @@ -9566,7 +9566,7 @@ static void intel_cpu_transcoder_get_m_n(struct intel_crtc *crtc, > > m_n->tu = ((I915_READ(PIPE_DATA_M1(transcoder)) > > & TU_SIZE_MASK) >> TU_SIZE_SHIFT) + 1; > > - if (m2_n2 && transcoder_has_m2_n2(dev_priv, transcoder)) { > > + if (m2_n2 && transcoder_has_m2_n2(i915, transcoder)) { > > m2_n2->link_m = I915_READ(PIPE_LINK_M2(transcoder)); > > m2_n2->link_n = I915_READ(PIPE_LINK_N2(transcoder)); > > m2_n2->gmch_m = I915_READ(PIPE_DATA_M2(transcoder)) > > @@ -9608,7 +9608,7 @@ static void skylake_get_pfit_config(struct intel_crtc *crtc, > > struct intel_crtc_state *pipe_config) > > { > > struct drm_device *dev = crtc->base.dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > struct intel_crtc_scaler_state *scaler_state = &pipe_config->scaler_state; > > u32 ps_ctrl = 0; > > int id = -1; > > @@ -9640,7 +9640,7 @@ skylake_get_initial_plane_config(struct intel_crtc *crtc, > > struct intel_initial_plane_config *plane_config) > > { > > struct drm_device *dev = crtc->base.dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > struct intel_plane *plane = to_intel_plane(crtc->base.primary); > > enum plane_id plane_id = plane->id; > > enum pipe pipe; > > @@ -9667,12 +9667,12 @@ skylake_get_initial_plane_config(struct intel_crtc *crtc, > > val = I915_READ(PLANE_CTL(pipe, plane_id)); > > - if (INTEL_GEN(dev_priv) >= 11) > > + if (INTEL_GEN(i915) >= 11) > > pixel_format = val & ICL_PLANE_CTL_FORMAT_MASK; > > else > > pixel_format = val & PLANE_CTL_FORMAT_MASK; > > - if (INTEL_GEN(dev_priv) >= 10 || IS_GEMINILAKE(dev_priv)) { > > + if (INTEL_GEN(i915) >= 10 || IS_GEMINILAKE(i915)) { > > alpha = I915_READ(PLANE_COLOR_CTL(pipe, plane_id)); > > alpha &= PLANE_COLOR_ALPHA_MASK; > > } else { > > @@ -9729,7 +9729,7 @@ skylake_get_initial_plane_config(struct intel_crtc *crtc, > > break; > > } > > - if (INTEL_GEN(dev_priv) >= 10 && > > + if (INTEL_GEN(i915) >= 10 && > > val & PLANE_CTL_FLIP_HORIZONTAL) > > plane_config->rotation |= DRM_MODE_REFLECT_X; > > @@ -9766,7 +9766,7 @@ static void ironlake_get_pfit_config(struct intel_crtc *crtc, > > struct intel_crtc_state *pipe_config) > > { > > struct drm_device *dev = crtc->base.dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > u32 tmp; > > tmp = I915_READ(PF_CTL(crtc->pipe)); > > @@ -9779,7 +9779,7 @@ static void ironlake_get_pfit_config(struct intel_crtc *crtc, > > /* We currently do not free assignements of panel fitters on > > * ivb/hsw (since we don't use the higher upscaling modes which > > * differentiates them) so just WARN about this case for now. */ > > - if (IS_GEN(dev_priv, 7)) { > > + if (IS_GEN(i915, 7)) { > > WARN_ON((tmp & PF_PIPE_SEL_MASK_IVB) != > > PF_PIPE_SEL_IVB(crtc->pipe)); > > } > > @@ -9790,14 +9790,14 @@ static bool ironlake_get_pipe_config(struct intel_crtc *crtc, > > struct intel_crtc_state *pipe_config) > > { > > struct drm_device *dev = crtc->base.dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > enum intel_display_power_domain power_domain; > > intel_wakeref_t wakeref; > > u32 tmp; > > bool ret; > > power_domain = POWER_DOMAIN_PIPE(crtc->pipe); > > - wakeref = intel_display_power_get_if_enabled(dev_priv, power_domain); > > + wakeref = intel_display_power_get_if_enabled(i915, power_domain); > > if (!wakeref) > > return false; > > @@ -9850,7 +9850,7 @@ static bool ironlake_get_pipe_config(struct intel_crtc *crtc, > > ironlake_get_fdi_m_n_config(crtc, pipe_config); > > - if (HAS_PCH_IBX(dev_priv)) { > > + if (HAS_PCH_IBX(i915)) { > > /* > > * The pipe->pch transcoder and pch transcoder->pll > > * mapping is fixed. > > @@ -9865,10 +9865,10 @@ static bool ironlake_get_pipe_config(struct intel_crtc *crtc, > > } > > pipe_config->shared_dpll = > > - intel_get_shared_dpll_by_id(dev_priv, pll_id); > > + intel_get_shared_dpll_by_id(i915, pll_id); > > pll = pipe_config->shared_dpll; > > - WARN_ON(!pll->info->funcs->get_hw_state(dev_priv, pll, > > + WARN_ON(!pll->info->funcs->get_hw_state(i915, pll, > > &pipe_config->dpll_hw_state)); > > tmp = pipe_config->dpll_hw_state.dpll; > > @@ -9889,19 +9889,19 @@ static bool ironlake_get_pipe_config(struct intel_crtc *crtc, > > ret = true; > > out: > > - intel_display_power_put(dev_priv, power_domain, wakeref); > > + intel_display_power_put(i915, power_domain, wakeref); > > return ret; > > } > > static int haswell_crtc_compute_clock(struct intel_crtc *crtc, > > struct intel_crtc_state *crtc_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > struct intel_atomic_state *state = > > to_intel_atomic_state(crtc_state->base.state); > > if (!intel_crtc_has_type(crtc_state, INTEL_OUTPUT_DSI) || > > - INTEL_GEN(dev_priv) >= 11) { > > + INTEL_GEN(i915) >= 11) { > > struct intel_encoder *encoder = > > intel_get_crtc_new_encoder(state, crtc_state); > > @@ -9915,7 +9915,7 @@ static int haswell_crtc_compute_clock(struct intel_crtc *crtc, > > return 0; > > } > > -static void cannonlake_get_ddi_pll(struct drm_i915_private *dev_priv, > > +static void cannonlake_get_ddi_pll(struct drm_i915_private *i915, > > enum port port, > > struct intel_crtc_state *pipe_config) > > { > > @@ -9928,10 +9928,10 @@ static void cannonlake_get_ddi_pll(struct drm_i915_private *dev_priv, > > if (WARN_ON(id < SKL_DPLL0 || id > SKL_DPLL2)) > > return; > > - pipe_config->shared_dpll = intel_get_shared_dpll_by_id(dev_priv, id); > > + pipe_config->shared_dpll = intel_get_shared_dpll_by_id(i915, id); > > } > > -static void icelake_get_ddi_pll(struct drm_i915_private *dev_priv, > > +static void icelake_get_ddi_pll(struct drm_i915_private *i915, > > enum port port, > > struct intel_crtc_state *pipe_config) > > { > > @@ -9939,21 +9939,21 @@ static void icelake_get_ddi_pll(struct drm_i915_private *dev_priv, > > u32 temp; > > /* TODO: TBT pll not implemented. */ > > - if (intel_port_is_combophy(dev_priv, port)) { > > + if (intel_port_is_combophy(i915, port)) { > > temp = I915_READ(DPCLKA_CFGCR0_ICL) & > > DPCLKA_CFGCR0_DDI_CLK_SEL_MASK(port); > > id = temp >> DPCLKA_CFGCR0_DDI_CLK_SEL_SHIFT(port); > > - } else if (intel_port_is_tc(dev_priv, port)) { > > - id = icl_tc_port_to_pll_id(intel_port_to_tc(dev_priv, port)); > > + } else if (intel_port_is_tc(i915, port)) { > > + id = icl_tc_port_to_pll_id(intel_port_to_tc(i915, port)); > > } else { > > WARN(1, "Invalid port %x\n", port); > > return; > > } > > - pipe_config->shared_dpll = intel_get_shared_dpll_by_id(dev_priv, id); > > + pipe_config->shared_dpll = intel_get_shared_dpll_by_id(i915, id); > > } > > -static void bxt_get_ddi_pll(struct drm_i915_private *dev_priv, > > +static void bxt_get_ddi_pll(struct drm_i915_private *i915, > > enum port port, > > struct intel_crtc_state *pipe_config) > > { > > @@ -9974,10 +9974,10 @@ static void bxt_get_ddi_pll(struct drm_i915_private *dev_priv, > > return; > > } > > - pipe_config->shared_dpll = intel_get_shared_dpll_by_id(dev_priv, id); > > + pipe_config->shared_dpll = intel_get_shared_dpll_by_id(i915, id); > > } > > -static void skylake_get_ddi_pll(struct drm_i915_private *dev_priv, > > +static void skylake_get_ddi_pll(struct drm_i915_private *i915, > > enum port port, > > struct intel_crtc_state *pipe_config) > > { > > @@ -9990,10 +9990,10 @@ static void skylake_get_ddi_pll(struct drm_i915_private *dev_priv, > > if (WARN_ON(id < SKL_DPLL0 || id > SKL_DPLL3)) > > return; > > - pipe_config->shared_dpll = intel_get_shared_dpll_by_id(dev_priv, id); > > + pipe_config->shared_dpll = intel_get_shared_dpll_by_id(i915, id); > > } > > -static void haswell_get_ddi_pll(struct drm_i915_private *dev_priv, > > +static void haswell_get_ddi_pll(struct drm_i915_private *i915, > > enum port port, > > struct intel_crtc_state *pipe_config) > > { > > @@ -10026,7 +10026,7 @@ static void haswell_get_ddi_pll(struct drm_i915_private *dev_priv, > > return; > > } > > - pipe_config->shared_dpll = intel_get_shared_dpll_by_id(dev_priv, id); > > + pipe_config->shared_dpll = intel_get_shared_dpll_by_id(i915, id); > > } > > static bool hsw_get_transcoder_state(struct intel_crtc *crtc, > > @@ -10035,7 +10035,7 @@ static bool hsw_get_transcoder_state(struct intel_crtc *crtc, > > intel_wakeref_t *wakerefs) > > { > > struct drm_device *dev = crtc->base.dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > enum intel_display_power_domain power_domain; > > unsigned long panel_transcoder_mask = 0; > > unsigned long enabled_panel_transcoders = 0; > > @@ -10043,11 +10043,11 @@ static bool hsw_get_transcoder_state(struct intel_crtc *crtc, > > intel_wakeref_t wf; > > u32 tmp; > > - if (INTEL_GEN(dev_priv) >= 11) > > + if (INTEL_GEN(i915) >= 11) > > panel_transcoder_mask |= > > BIT(TRANSCODER_DSI_0) | BIT(TRANSCODER_DSI_1); > > - if (HAS_TRANSCODER_EDP(dev_priv)) > > + if (HAS_TRANSCODER_EDP(i915)) > > panel_transcoder_mask |= BIT(TRANSCODER_EDP); > > /* > > @@ -10062,7 +10062,7 @@ static bool hsw_get_transcoder_state(struct intel_crtc *crtc, > > */ > > for_each_set_bit(panel_transcoder, > > &panel_transcoder_mask, > > - ARRAY_SIZE(INTEL_INFO(dev_priv)->trans_offsets)) { > > + ARRAY_SIZE(INTEL_INFO(i915)->trans_offsets)) { > > bool force_thru = false; > > enum pipe trans_pipe; > > @@ -10113,7 +10113,7 @@ static bool hsw_get_transcoder_state(struct intel_crtc *crtc, > > power_domain = POWER_DOMAIN_TRANSCODER(pipe_config->cpu_transcoder); > > WARN_ON(*power_domain_mask & BIT_ULL(power_domain)); > > - wf = intel_display_power_get_if_enabled(dev_priv, power_domain); > > + wf = intel_display_power_get_if_enabled(i915, power_domain); > > if (!wf) > > return false; > > @@ -10131,7 +10131,7 @@ static bool bxt_get_dsi_transcoder_state(struct intel_crtc *crtc, > > intel_wakeref_t *wakerefs) > > { > > struct drm_device *dev = crtc->base.dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > enum intel_display_power_domain power_domain; > > enum transcoder cpu_transcoder; > > intel_wakeref_t wf; > > @@ -10147,7 +10147,7 @@ static bool bxt_get_dsi_transcoder_state(struct intel_crtc *crtc, > > power_domain = POWER_DOMAIN_TRANSCODER(cpu_transcoder); > > WARN_ON(*power_domain_mask & BIT_ULL(power_domain)); > > - wf = intel_display_power_get_if_enabled(dev_priv, power_domain); > > + wf = intel_display_power_get_if_enabled(i915, power_domain); > > if (!wf) > > continue; > > @@ -10161,7 +10161,7 @@ static bool bxt_get_dsi_transcoder_state(struct intel_crtc *crtc, > > * registers/MIPI[BXT]. We can break out here early, since we > > * need the same DSI PLL to be enabled for both DSI ports. > > */ > > - if (!bxt_dsi_pll_is_enabled(dev_priv)) > > + if (!bxt_dsi_pll_is_enabled(i915)) > > break; > > /* XXX: this works for video mode only */ > > @@ -10183,7 +10183,7 @@ static bool bxt_get_dsi_transcoder_state(struct intel_crtc *crtc, > > static void haswell_get_ddi_port_state(struct intel_crtc *crtc, > > struct intel_crtc_state *pipe_config) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > struct intel_shared_dpll *pll; > > enum port port; > > u32 tmp; > > @@ -10192,20 +10192,20 @@ static void haswell_get_ddi_port_state(struct intel_crtc *crtc, > > port = (tmp & TRANS_DDI_PORT_MASK) >> TRANS_DDI_PORT_SHIFT; > > - if (INTEL_GEN(dev_priv) >= 11) > > - icelake_get_ddi_pll(dev_priv, port, pipe_config); > > - else if (IS_CANNONLAKE(dev_priv)) > > - cannonlake_get_ddi_pll(dev_priv, port, pipe_config); > > - else if (IS_GEN9_BC(dev_priv)) > > - skylake_get_ddi_pll(dev_priv, port, pipe_config); > > - else if (IS_GEN9_LP(dev_priv)) > > - bxt_get_ddi_pll(dev_priv, port, pipe_config); > > + if (INTEL_GEN(i915) >= 11) > > + icelake_get_ddi_pll(i915, port, pipe_config); > > + else if (IS_CANNONLAKE(i915)) > > + cannonlake_get_ddi_pll(i915, port, pipe_config); > > + else if (IS_GEN9_BC(i915)) > > + skylake_get_ddi_pll(i915, port, pipe_config); > > + else if (IS_GEN9_LP(i915)) > > + bxt_get_ddi_pll(i915, port, pipe_config); > > else > > - haswell_get_ddi_pll(dev_priv, port, pipe_config); > > + haswell_get_ddi_pll(i915, port, pipe_config); > > pll = pipe_config->shared_dpll; > > if (pll) { > > - WARN_ON(!pll->info->funcs->get_hw_state(dev_priv, pll, > > + WARN_ON(!pll->info->funcs->get_hw_state(i915, pll, > > &pipe_config->dpll_hw_state)); > > } > > @@ -10214,7 +10214,7 @@ static void haswell_get_ddi_port_state(struct intel_crtc *crtc, > > * DDI E. So just check whether this pipe is wired to DDI E and whether > > * the PCH transcoder is on. > > */ > > - if (INTEL_GEN(dev_priv) < 9 && > > + if (INTEL_GEN(i915) < 9 && > > (port == PORT_E) && I915_READ(LPT_TRANSCONF) & TRANS_ENABLE) { > > pipe_config->has_pch_encoder = true; > > @@ -10229,7 +10229,7 @@ static void haswell_get_ddi_port_state(struct intel_crtc *crtc, > > static bool haswell_get_pipe_config(struct intel_crtc *crtc, > > struct intel_crtc_state *pipe_config) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > intel_wakeref_t wakerefs[POWER_DOMAIN_NUM], wf; > > enum intel_display_power_domain power_domain; > > u64 power_domain_mask; > > @@ -10238,7 +10238,7 @@ static bool haswell_get_pipe_config(struct intel_crtc *crtc, > > intel_crtc_init_scalers(crtc, pipe_config); > > power_domain = POWER_DOMAIN_PIPE(crtc->pipe); > > - wf = intel_display_power_get_if_enabled(dev_priv, power_domain); > > + wf = intel_display_power_get_if_enabled(i915, power_domain); > > if (!wf) > > return false; > > @@ -10250,7 +10250,7 @@ static bool haswell_get_pipe_config(struct intel_crtc *crtc, > > active = hsw_get_transcoder_state(crtc, pipe_config, > > &power_domain_mask, wakerefs); > > - if (IS_GEN9_LP(dev_priv) && > > + if (IS_GEN9_LP(i915) && > > bxt_get_dsi_transcoder_state(crtc, pipe_config, > > &power_domain_mask, wakerefs)) { > > WARN_ON(active); > > @@ -10261,7 +10261,7 @@ static bool haswell_get_pipe_config(struct intel_crtc *crtc, > > goto out; > > if (!transcoder_is_dsi(pipe_config->cpu_transcoder) || > > - INTEL_GEN(dev_priv) >= 11) { > > + INTEL_GEN(i915) >= 11) { > > haswell_get_ddi_port_state(crtc, pipe_config); > > intel_get_pipe_timings(crtc, pipe_config); > > } > > @@ -10273,7 +10273,7 @@ static bool haswell_get_pipe_config(struct intel_crtc *crtc, > > pipe_config->csc_mode = I915_READ(PIPE_CSC_MODE(crtc->pipe)); > > - if (INTEL_GEN(dev_priv) >= 9) { > > + if (INTEL_GEN(i915) >= 9) { > > u32 tmp = I915_READ(SKL_BOTTOM_COLOR(crtc->pipe)); > > if (tmp & SKL_BOTTOM_COLOR_GAMMA_ENABLE) > > @@ -10290,19 +10290,19 @@ static bool haswell_get_pipe_config(struct intel_crtc *crtc, > > power_domain = POWER_DOMAIN_PIPE_PANEL_FITTER(crtc->pipe); > > WARN_ON(power_domain_mask & BIT_ULL(power_domain)); > > - wf = intel_display_power_get_if_enabled(dev_priv, power_domain); > > + wf = intel_display_power_get_if_enabled(i915, power_domain); > > if (wf) { > > wakerefs[power_domain] = wf; > > power_domain_mask |= BIT_ULL(power_domain); > > - if (INTEL_GEN(dev_priv) >= 9) > > + if (INTEL_GEN(i915) >= 9) > > skylake_get_pfit_config(crtc, pipe_config); > > else > > ironlake_get_pfit_config(crtc, pipe_config); > > } > > if (hsw_crtc_supports_ips(crtc)) { > > - if (IS_HASWELL(dev_priv)) > > + if (IS_HASWELL(i915)) > > pipe_config->ips_enabled = I915_READ(IPS_CTL) & IPS_ENABLE; > > else { > > /* > > @@ -10324,7 +10324,7 @@ static bool haswell_get_pipe_config(struct intel_crtc *crtc, > > out: > > for_each_power_domain(power_domain, power_domain_mask) > > - intel_display_power_put(dev_priv, > > + intel_display_power_put(i915, > > power_domain, wakerefs[power_domain]); > > return active; > > @@ -10332,13 +10332,13 @@ static bool haswell_get_pipe_config(struct intel_crtc *crtc, > > static u32 intel_cursor_base(const struct intel_plane_state *plane_state) > > { > > - struct drm_i915_private *dev_priv = > > + struct drm_i915_private *i915 = > > to_i915(plane_state->base.plane->dev); > > const struct drm_framebuffer *fb = plane_state->base.fb; > > const struct drm_i915_gem_object *obj = intel_fb_obj(fb); > > u32 base; > > - if (INTEL_INFO(dev_priv)->display.cursor_needs_physical) > > + if (INTEL_INFO(i915)->display.cursor_needs_physical) > > base = obj->phys_handle->busaddr; > > else > > base = intel_plane_ggtt_offset(plane_state); > > @@ -10346,7 +10346,7 @@ static u32 intel_cursor_base(const struct intel_plane_state *plane_state) > > base += plane_state->color_plane[0].offset; > > /* ILK+ do this automagically */ > > - if (HAS_GMCH(dev_priv) && > > + if (HAS_GMCH(i915) && > > plane_state->base.rotation & DRM_MODE_ROTATE_180) > > base += (plane_state->base.crtc_h * > > plane_state->base.crtc_w - 1) * fb->format->cpp[0]; > > @@ -10532,7 +10532,7 @@ static void i845_update_cursor(struct intel_plane *plane, > > const struct intel_crtc_state *crtc_state, > > const struct intel_plane_state *plane_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(plane->base.dev); > > + struct drm_i915_private *i915 = to_i915(plane->base.dev); > > u32 cntl = 0, base = 0, pos = 0, size = 0; > > unsigned long irqflags; > > @@ -10549,7 +10549,7 @@ static void i845_update_cursor(struct intel_plane *plane, > > pos = intel_cursor_position(plane_state); > > } > > - spin_lock_irqsave(&dev_priv->uncore.lock, irqflags); > > + spin_lock_irqsave(&i915->uncore.lock, irqflags); > > /* On these chipsets we can only modify the base/size/stride > > * whilst the cursor is disabled. > > @@ -10570,7 +10570,7 @@ static void i845_update_cursor(struct intel_plane *plane, > > I915_WRITE_FW(CURPOS(PIPE_A), pos); > > } > > - spin_unlock_irqrestore(&dev_priv->uncore.lock, irqflags); > > + spin_unlock_irqrestore(&i915->uncore.lock, irqflags); > > } > > static void i845_disable_cursor(struct intel_plane *plane, > > @@ -10582,13 +10582,13 @@ static void i845_disable_cursor(struct intel_plane *plane, > > static bool i845_cursor_get_hw_state(struct intel_plane *plane, > > enum pipe *pipe) > > { > > - struct drm_i915_private *dev_priv = to_i915(plane->base.dev); > > + struct drm_i915_private *i915 = to_i915(plane->base.dev); > > enum intel_display_power_domain power_domain; > > intel_wakeref_t wakeref; > > bool ret; > > power_domain = POWER_DOMAIN_PIPE(PIPE_A); > > - wakeref = intel_display_power_get_if_enabled(dev_priv, power_domain); > > + wakeref = intel_display_power_get_if_enabled(i915, power_domain); > > if (!wakeref) > > return false; > > @@ -10596,7 +10596,7 @@ static bool i845_cursor_get_hw_state(struct intel_plane *plane, > > *pipe = PIPE_A; > > - intel_display_power_put(dev_priv, power_domain, wakeref); > > + intel_display_power_put(i915, power_domain, wakeref); > > return ret; > > } > > @@ -10612,10 +10612,10 @@ i9xx_cursor_max_stride(struct intel_plane *plane, > > static u32 i9xx_cursor_ctl_crtc(const struct intel_crtc_state *crtc_state) > > { > > struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc); > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > u32 cntl = 0; > > - if (INTEL_GEN(dev_priv) >= 11) > > + if (INTEL_GEN(i915) >= 11) > > return cntl; > > if (crtc_state->gamma_enable) > > @@ -10624,7 +10624,7 @@ static u32 i9xx_cursor_ctl_crtc(const struct intel_crtc_state *crtc_state) > > if (crtc_state->csc_enable) > > cntl |= MCURSOR_PIPE_CSC_ENABLE; > > - if (INTEL_GEN(dev_priv) < 5 && !IS_G4X(dev_priv)) > > + if (INTEL_GEN(i915) < 5 && !IS_G4X(i915)) > > cntl |= MCURSOR_PIPE_SELECT(crtc->pipe); > > return cntl; > > @@ -10633,11 +10633,11 @@ static u32 i9xx_cursor_ctl_crtc(const struct intel_crtc_state *crtc_state) > > static u32 i9xx_cursor_ctl(const struct intel_crtc_state *crtc_state, > > const struct intel_plane_state *plane_state) > > { > > - struct drm_i915_private *dev_priv = > > + struct drm_i915_private *i915 = > > to_i915(plane_state->base.plane->dev); > > u32 cntl = 0; > > - if (IS_GEN(dev_priv, 6) || IS_IVYBRIDGE(dev_priv)) > > + if (IS_GEN(i915, 6) || IS_IVYBRIDGE(i915)) > > cntl |= MCURSOR_TRICKLE_FEED_DISABLE; > > switch (plane_state->base.crtc_w) { > > @@ -10663,7 +10663,7 @@ static u32 i9xx_cursor_ctl(const struct intel_crtc_state *crtc_state, > > static bool i9xx_cursor_size_ok(const struct intel_plane_state *plane_state) > > { > > - struct drm_i915_private *dev_priv = > > + struct drm_i915_private *i915 = > > to_i915(plane_state->base.plane->dev); > > int width = plane_state->base.crtc_w; > > int height = plane_state->base.crtc_h; > > @@ -10687,7 +10687,7 @@ static bool i9xx_cursor_size_ok(const struct intel_plane_state *plane_state) > > * cursor is not rotated. Everything else requires square > > * cursors. > > */ > > - if (HAS_CUR_FBC(dev_priv) && > > + if (HAS_CUR_FBC(i915) && > > plane_state->base.rotation & DRM_MODE_ROTATE_0) { > > if (height < 8 || height > width) > > return false; > > @@ -10703,7 +10703,7 @@ static int i9xx_check_cursor(struct intel_crtc_state *crtc_state, > > struct intel_plane_state *plane_state) > > { > > struct intel_plane *plane = to_intel_plane(plane_state->base.plane); > > - struct drm_i915_private *dev_priv = to_i915(plane->base.dev); > > + struct drm_i915_private *i915 = to_i915(plane->base.dev); > > const struct drm_framebuffer *fb = plane_state->base.fb; > > enum pipe pipe = plane->pipe; > > int ret; > > @@ -10743,7 +10743,7 @@ static int i9xx_check_cursor(struct intel_crtc_state *crtc_state, > > * display power well must be turned off and on again. > > * Refuse the put the cursor into that compromised position. > > */ > > - if (IS_CHERRYVIEW(dev_priv) && pipe == PIPE_C && > > + if (IS_CHERRYVIEW(i915) && pipe == PIPE_C && > > plane_state->base.visible && plane_state->base.crtc_x < 0) { > > DRM_DEBUG_KMS("CHV cursor C not allowed to straddle the left screen edge\n"); > > return -EINVAL; > > @@ -10758,7 +10758,7 @@ static void i9xx_update_cursor(struct intel_plane *plane, > > const struct intel_crtc_state *crtc_state, > > const struct intel_plane_state *plane_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(plane->base.dev); > > + struct drm_i915_private *i915 = to_i915(plane->base.dev); > > enum pipe pipe = plane->pipe; > > u32 cntl = 0, base = 0, pos = 0, fbc_ctl = 0; > > unsigned long irqflags; > > @@ -10774,7 +10774,7 @@ static void i9xx_update_cursor(struct intel_plane *plane, > > pos = intel_cursor_position(plane_state); > > } > > - spin_lock_irqsave(&dev_priv->uncore.lock, irqflags); > > + spin_lock_irqsave(&i915->uncore.lock, irqflags); > > /* > > * On some platforms writing CURCNTR first will also > > @@ -10796,13 +10796,13 @@ static void i9xx_update_cursor(struct intel_plane *plane, > > * the CURCNTR write arms the update. > > */ > > - if (INTEL_GEN(dev_priv) >= 9) > > + if (INTEL_GEN(i915) >= 9) > > skl_write_cursor_wm(plane, crtc_state); > > if (plane->cursor.base != base || > > plane->cursor.size != fbc_ctl || > > plane->cursor.cntl != cntl) { > > - if (HAS_CUR_FBC(dev_priv)) > > + if (HAS_CUR_FBC(i915)) > > I915_WRITE_FW(CUR_FBC_CTL(pipe), fbc_ctl); > > I915_WRITE_FW(CURCNTR(pipe), cntl); > > I915_WRITE_FW(CURPOS(pipe), pos); > > @@ -10816,7 +10816,7 @@ static void i9xx_update_cursor(struct intel_plane *plane, > > I915_WRITE_FW(CURBASE(pipe), base); > > } > > - spin_unlock_irqrestore(&dev_priv->uncore.lock, irqflags); > > + spin_unlock_irqrestore(&i915->uncore.lock, irqflags); > > } > > static void i9xx_disable_cursor(struct intel_plane *plane, > > @@ -10828,7 +10828,7 @@ static void i9xx_disable_cursor(struct intel_plane *plane, > > static bool i9xx_cursor_get_hw_state(struct intel_plane *plane, > > enum pipe *pipe) > > { > > - struct drm_i915_private *dev_priv = to_i915(plane->base.dev); > > + struct drm_i915_private *i915 = to_i915(plane->base.dev); > > enum intel_display_power_domain power_domain; > > intel_wakeref_t wakeref; > > bool ret; > > @@ -10840,7 +10840,7 @@ static bool i9xx_cursor_get_hw_state(struct intel_plane *plane, > > * display power wells. > > */ > > power_domain = POWER_DOMAIN_PIPE(plane->pipe); > > - wakeref = intel_display_power_get_if_enabled(dev_priv, power_domain); > > + wakeref = intel_display_power_get_if_enabled(i915, power_domain); > > if (!wakeref) > > return false; > > @@ -10848,13 +10848,13 @@ static bool i9xx_cursor_get_hw_state(struct intel_plane *plane, > > ret = val & MCURSOR_MODE; > > - if (INTEL_GEN(dev_priv) >= 5 || IS_G4X(dev_priv)) > > + if (INTEL_GEN(i915) >= 5 || IS_G4X(i915)) > > *pipe = plane->pipe; > > else > > *pipe = (val & MCURSOR_PIPE_SELECT_MASK) >> > > MCURSOR_PIPE_SELECT_SHIFT; > > - intel_display_power_put(dev_priv, power_domain, wakeref); > > + intel_display_power_put(i915, power_domain, wakeref); > > return ret; > > } > > @@ -10924,7 +10924,7 @@ int intel_get_load_detect_pipe(struct drm_connector *connector, > > struct drm_encoder *encoder = &intel_encoder->base; > > struct drm_crtc *crtc = NULL; > > struct drm_device *dev = encoder->dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > struct drm_mode_config *config = &dev->mode_config; > > struct drm_atomic_state *state = NULL, *restore_state = NULL; > > struct drm_connector_state *connector_state; > > @@ -11051,7 +11051,7 @@ int intel_get_load_detect_pipe(struct drm_connector *connector, > > drm_atomic_state_put(state); > > /* let the connector get through one full cycle before testing */ > > - intel_wait_for_vblank(dev_priv, intel_crtc->pipe); > > + intel_wait_for_vblank(i915, intel_crtc->pipe); > > return true; > > fail: > > @@ -11096,14 +11096,14 @@ void intel_release_load_detect_pipe(struct drm_connector *connector, > > static int i9xx_pll_refclk(struct drm_device *dev, > > const struct intel_crtc_state *pipe_config) > > { > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > u32 dpll = pipe_config->dpll_hw_state.dpll; > > if ((dpll & PLL_REF_INPUT_MASK) == PLLB_REF_INPUT_SPREADSPECTRUMIN) > > - return dev_priv->vbt.lvds_ssc_freq; > > - else if (HAS_PCH_SPLIT(dev_priv)) > > + return i915->vbt.lvds_ssc_freq; > > + else if (HAS_PCH_SPLIT(i915)) > > return 120000; > > - else if (!IS_GEN(dev_priv, 2)) > > + else if (!IS_GEN(i915, 2)) > > return 96000; > > else > > return 48000; > > @@ -11114,7 +11114,7 @@ static void i9xx_crtc_clock_get(struct intel_crtc *crtc, > > struct intel_crtc_state *pipe_config) > > { > > struct drm_device *dev = crtc->base.dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > int pipe = pipe_config->cpu_transcoder; > > u32 dpll = pipe_config->dpll_hw_state.dpll; > > u32 fp; > > @@ -11128,7 +11128,7 @@ static void i9xx_crtc_clock_get(struct intel_crtc *crtc, > > fp = pipe_config->dpll_hw_state.fp1; > > clock.m1 = (fp & FP_M1_DIV_MASK) >> FP_M1_DIV_SHIFT; > > - if (IS_PINEVIEW(dev_priv)) { > > + if (IS_PINEVIEW(i915)) { > > clock.n = ffs((fp & FP_N_PINEVIEW_DIV_MASK) >> FP_N_DIV_SHIFT) - 1; > > clock.m2 = (fp & FP_M2_PINEVIEW_DIV_MASK) >> FP_M2_DIV_SHIFT; > > } else { > > @@ -11136,8 +11136,8 @@ static void i9xx_crtc_clock_get(struct intel_crtc *crtc, > > clock.m2 = (fp & FP_M2_DIV_MASK) >> FP_M2_DIV_SHIFT; > > } > > - if (!IS_GEN(dev_priv, 2)) { > > - if (IS_PINEVIEW(dev_priv)) > > + if (!IS_GEN(i915, 2)) { > > + if (IS_PINEVIEW(i915)) > > clock.p1 = ffs((dpll & DPLL_FPA01_P1_POST_DIV_MASK_PINEVIEW) >> > > DPLL_FPA01_P1_POST_DIV_SHIFT_PINEVIEW); > > else > > @@ -11159,12 +11159,12 @@ static void i9xx_crtc_clock_get(struct intel_crtc *crtc, > > return; > > } > > - if (IS_PINEVIEW(dev_priv)) > > + if (IS_PINEVIEW(i915)) > > port_clock = pnv_calc_dpll_params(refclk, &clock); > > else > > port_clock = i9xx_calc_dpll_params(refclk, &clock); > > } else { > > - u32 lvds = IS_I830(dev_priv) ? 0 : I915_READ(LVDS); > > + u32 lvds = IS_I830(i915) ? 0 : I915_READ(LVDS); > > bool is_lvds = (pipe == 1) && (lvds & LVDS_PORT_EN); > > if (is_lvds) { > > @@ -11221,7 +11221,7 @@ int intel_dotclock_calculate(int link_freq, > > static void ironlake_pch_clock_get(struct intel_crtc *crtc, > > struct intel_crtc_state *pipe_config) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > /* read out port_clock from the DPLL */ > > i9xx_crtc_clock_get(crtc, pipe_config); > > @@ -11232,7 +11232,7 @@ static void ironlake_pch_clock_get(struct intel_crtc *crtc, > > * Calculate one based on the FDI configuration. > > */ > > pipe_config->base.adjusted_mode.crtc_clock = > > - intel_dotclock_calculate(intel_fdi_link_freq(dev_priv, pipe_config), > > + intel_dotclock_calculate(intel_fdi_link_freq(i915, pipe_config), > > &pipe_config->fdi_m_n); > > } > > @@ -11240,7 +11240,7 @@ static void ironlake_pch_clock_get(struct intel_crtc *crtc, > > struct drm_display_mode * > > intel_encoder_current_mode(struct intel_encoder *encoder) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_crtc_state *crtc_state; > > struct drm_display_mode *mode; > > struct intel_crtc *crtc; > > @@ -11249,7 +11249,7 @@ intel_encoder_current_mode(struct intel_encoder *encoder) > > if (!encoder->get_hw_state(encoder, &pipe)) > > return NULL; > > - crtc = intel_get_crtc_for_pipe(dev_priv, pipe); > > + crtc = intel_get_crtc_for_pipe(i915, pipe); > > mode = kzalloc(sizeof(*mode), GFP_KERNEL); > > if (!mode) > > @@ -11263,7 +11263,7 @@ intel_encoder_current_mode(struct intel_encoder *encoder) > > crtc_state->base.crtc = &crtc->base; > > - if (!dev_priv->display.get_pipe_config(crtc, crtc_state)) { > > + if (!i915->display.get_pipe_config(crtc, crtc_state)) { > > kfree(crtc_state); > > kfree(mode); > > return NULL; > > @@ -11337,7 +11337,7 @@ int intel_plane_atomic_calc_changes(const struct intel_crtc_state *old_crtc_stat > > struct intel_crtc *intel_crtc = to_intel_crtc(crtc); > > struct intel_plane *plane = to_intel_plane(plane_state->plane); > > struct drm_device *dev = crtc->dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > bool mode_changed = needs_modeset(crtc_state); > > bool was_crtc_enabled = old_crtc_state->base.active; > > bool is_crtc_enabled = crtc_state->active; > > @@ -11345,7 +11345,7 @@ int intel_plane_atomic_calc_changes(const struct intel_crtc_state *old_crtc_stat > > struct drm_framebuffer *fb = plane_state->fb; > > int ret; > > - if (INTEL_GEN(dev_priv) >= 9 && plane->id != PLANE_CURSOR) { > > + if (INTEL_GEN(i915) >= 9 && plane->id != PLANE_CURSOR) { > > ret = skl_update_scaler_plane( > > to_intel_crtc_state(crtc_state), > > to_intel_plane_state(plane_state)); > > @@ -11395,14 +11395,14 @@ int intel_plane_atomic_calc_changes(const struct intel_crtc_state *old_crtc_stat > > turn_off, turn_on, mode_changed); > > if (turn_on) { > > - if (INTEL_GEN(dev_priv) < 5 && !IS_G4X(dev_priv)) > > + if (INTEL_GEN(i915) < 5 && !IS_G4X(i915)) > > pipe_config->update_wm_pre = true; > > /* must disable cxsr around plane enable/disable */ > > if (plane->id != PLANE_CURSOR) > > pipe_config->disable_cxsr = true; > > } else if (turn_off) { > > - if (INTEL_GEN(dev_priv) < 5 && !IS_G4X(dev_priv)) > > + if (INTEL_GEN(i915) < 5 && !IS_G4X(i915)) > > pipe_config->update_wm_post = true; > > /* must disable cxsr around plane enable/disable */ > > @@ -11410,7 +11410,7 @@ int intel_plane_atomic_calc_changes(const struct intel_crtc_state *old_crtc_stat > > pipe_config->disable_cxsr = true; > > } else if (intel_wm_need_update(to_intel_plane_state(plane->base.state), > > to_intel_plane_state(plane_state))) { > > - if (INTEL_GEN(dev_priv) < 5 && !IS_G4X(dev_priv)) { > > + if (INTEL_GEN(i915) < 5 && !IS_G4X(i915)) { > > /* FIXME bollocks */ > > pipe_config->update_wm_pre = true; > > pipe_config->update_wm_post = true; > > @@ -11454,8 +11454,8 @@ int intel_plane_atomic_calc_changes(const struct intel_crtc_state *old_crtc_stat > > * plane, not only sprite plane. > > */ > > if (plane->id != PLANE_CURSOR && > > - (IS_GEN_RANGE(dev_priv, 5, 6) || > > - IS_IVYBRIDGE(dev_priv)) && > > + (IS_GEN_RANGE(i915, 5, 6) || > > + IS_IVYBRIDGE(i915)) && > > (turn_on || (!needs_scaling(old_plane_state) && > > needs_scaling(to_intel_plane_state(plane_state))))) > > pipe_config->disable_lp_wm = true; > > @@ -11519,13 +11519,13 @@ static int icl_add_linked_planes(struct intel_atomic_state *state) > > static int icl_check_nv12_planes(struct intel_crtc_state *crtc_state) > > { > > struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc); > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > struct intel_atomic_state *state = to_intel_atomic_state(crtc_state->base.state); > > struct intel_plane *plane, *linked; > > struct intel_plane_state *plane_state; > > int i; > > - if (INTEL_GEN(dev_priv) < 11) > > + if (INTEL_GEN(i915) < 11) > > return 0; > > /* > > @@ -11555,7 +11555,7 @@ static int icl_check_nv12_planes(struct intel_crtc_state *crtc_state) > > !(crtc_state->nv12_planes & BIT(plane->id))) > > continue; > > - for_each_intel_plane_on_crtc(&dev_priv->drm, crtc, linked) { > > + for_each_intel_plane_on_crtc(&i915->drm, crtc, linked) { > > if (!icl_is_nv12_y_plane(linked->id)) > > continue; > > @@ -11602,21 +11602,21 @@ static bool c8_planes_changed(const struct intel_crtc_state *new_crtc_state) > > static int intel_crtc_atomic_check(struct drm_crtc *crtc, > > struct drm_crtc_state *crtc_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc->dev); > > + struct drm_i915_private *i915 = to_i915(crtc->dev); > > struct intel_crtc *intel_crtc = to_intel_crtc(crtc); > > struct intel_crtc_state *pipe_config = > > to_intel_crtc_state(crtc_state); > > int ret; > > bool mode_changed = needs_modeset(crtc_state); > > - if (INTEL_GEN(dev_priv) < 5 && !IS_G4X(dev_priv) && > > + if (INTEL_GEN(i915) < 5 && !IS_G4X(i915) && > > mode_changed && !crtc_state->active) > > pipe_config->update_wm_post = true; > > if (mode_changed && crtc_state->enable && > > - dev_priv->display.crtc_compute_clock && > > + i915->display.crtc_compute_clock && > > !WARN_ON(pipe_config->shared_dpll)) { > > - ret = dev_priv->display.crtc_compute_clock(intel_crtc, > > + ret = i915->display.crtc_compute_clock(intel_crtc, > > pipe_config); > > if (ret) > > return ret; > > @@ -11637,16 +11637,16 @@ static int intel_crtc_atomic_check(struct drm_crtc *crtc, > > } > > ret = 0; > > - if (dev_priv->display.compute_pipe_wm) { > > - ret = dev_priv->display.compute_pipe_wm(pipe_config); > > + if (i915->display.compute_pipe_wm) { > > + ret = i915->display.compute_pipe_wm(pipe_config); > > if (ret) { > > DRM_DEBUG_KMS("Target pipe watermarks are invalid\n"); > > return ret; > > } > > } > > - if (dev_priv->display.compute_intermediate_wm) { > > - if (WARN_ON(!dev_priv->display.compute_pipe_wm)) > > + if (i915->display.compute_intermediate_wm) { > > + if (WARN_ON(!i915->display.compute_pipe_wm)) > > return 0; > > /* > > @@ -11654,14 +11654,14 @@ static int intel_crtc_atomic_check(struct drm_crtc *crtc, > > * old state and the new state. We can program these > > * immediately. > > */ > > - ret = dev_priv->display.compute_intermediate_wm(pipe_config); > > + ret = i915->display.compute_intermediate_wm(pipe_config); > > if (ret) { > > DRM_DEBUG_KMS("No valid intermediate pipe watermarks are possible\n"); > > return ret; > > } > > } > > - if (INTEL_GEN(dev_priv) >= 9) { > > + if (INTEL_GEN(i915) >= 9) { > > if (mode_changed || pipe_config->update_pipe) > > ret = skl_update_scaler_crtc(pipe_config); > > @@ -11671,11 +11671,11 @@ static int intel_crtc_atomic_check(struct drm_crtc *crtc, > > ret = skl_check_pipe_max_pixel_rate(intel_crtc, > > pipe_config); > > if (!ret) > > - ret = intel_atomic_setup_scalers(dev_priv, intel_crtc, > > + ret = intel_atomic_setup_scalers(i915, intel_crtc, > > pipe_config); > > } > > - if (HAS_IPS(dev_priv)) > > + if (HAS_IPS(i915)) > > pipe_config->ips_enabled = hsw_compute_ips_config(pipe_config); > > return ret; > > @@ -11752,16 +11752,16 @@ static int > > compute_baseline_pipe_bpp(struct intel_crtc *crtc, > > struct intel_crtc_state *pipe_config) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > struct drm_atomic_state *state = pipe_config->base.state; > > struct drm_connector *connector; > > struct drm_connector_state *connector_state; > > int bpp, i; > > - if ((IS_G4X(dev_priv) || IS_VALLEYVIEW(dev_priv) || > > - IS_CHERRYVIEW(dev_priv))) > > + if ((IS_G4X(i915) || IS_VALLEYVIEW(i915) || > > + IS_CHERRYVIEW(i915))) > > bpp = 10*3; > > - else if (INTEL_GEN(dev_priv) >= 5) > > + else if (INTEL_GEN(i915) >= 5) > > bpp = 12*3; > > else > > bpp = 8*3; > > @@ -11807,13 +11807,13 @@ intel_dump_m_n_config(const struct intel_crtc_state *pipe_config, > > } > > static void > > -intel_dump_infoframe(struct drm_i915_private *dev_priv, > > +intel_dump_infoframe(struct drm_i915_private *i915, > > const union hdmi_infoframe *frame) > > { > > if ((drm_debug & DRM_UT_KMS) == 0) > > return; > > - hdmi_infoframe_log(KERN_DEBUG, dev_priv->drm.dev, frame); > > + hdmi_infoframe_log(KERN_DEBUG, i915->drm.dev, frame); > > } > > #define OUTPUT_TYPE(x) [INTEL_OUTPUT_ ## x] = #x > > @@ -11907,7 +11907,7 @@ static void intel_dump_pipe_config(const struct intel_crtc_state *pipe_config, > > const char *context) > > { > > struct intel_crtc *crtc = to_intel_crtc(pipe_config->base.crtc); > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > const struct intel_plane_state *plane_state; > > struct intel_plane *plane; > > char buf[64]; > > @@ -11953,13 +11953,13 @@ static void intel_dump_pipe_config(const struct intel_crtc_state *pipe_config, > > DRM_DEBUG_KMS("GCP: 0x%x\n", pipe_config->infoframes.gcp); > > if (pipe_config->infoframes.enable & > > intel_hdmi_infoframe_enable(HDMI_INFOFRAME_TYPE_AVI)) > > - intel_dump_infoframe(dev_priv, &pipe_config->infoframes.avi); > > + intel_dump_infoframe(i915, &pipe_config->infoframes.avi); > > if (pipe_config->infoframes.enable & > > intel_hdmi_infoframe_enable(HDMI_INFOFRAME_TYPE_SPD)) > > - intel_dump_infoframe(dev_priv, &pipe_config->infoframes.spd); > > + intel_dump_infoframe(i915, &pipe_config->infoframes.spd); > > if (pipe_config->infoframes.enable & > > intel_hdmi_infoframe_enable(HDMI_INFOFRAME_TYPE_VENDOR)) > > - intel_dump_infoframe(dev_priv, &pipe_config->infoframes.hdmi); > > + intel_dump_infoframe(i915, &pipe_config->infoframes.hdmi); > > DRM_DEBUG_KMS("requested mode:\n"); > > drm_mode_debug_printmodeline(&pipe_config->base.mode); > > @@ -11971,13 +11971,13 @@ static void intel_dump_pipe_config(const struct intel_crtc_state *pipe_config, > > pipe_config->pipe_src_w, pipe_config->pipe_src_h, > > pipe_config->pixel_rate); > > - if (INTEL_GEN(dev_priv) >= 9) > > + if (INTEL_GEN(i915) >= 9) > > DRM_DEBUG_KMS("num_scalers: %d, scaler_users: 0x%x, scaler_id: %d\n", > > crtc->num_scalers, > > pipe_config->scaler_state.scaler_users, > > pipe_config->scaler_state.scaler_id); > > - if (HAS_GMCH(dev_priv)) > > + if (HAS_GMCH(i915)) > > DRM_DEBUG_KMS("gmch pfit: control: 0x%08x, ratios: 0x%08x, lvds border: 0x%08x\n", > > pipe_config->gmch_pfit.control, > > pipe_config->gmch_pfit.pgm_ratios, > > @@ -11992,7 +11992,7 @@ static void intel_dump_pipe_config(const struct intel_crtc_state *pipe_config, > > DRM_DEBUG_KMS("ips: %i, double wide: %i\n", > > pipe_config->ips_enabled, pipe_config->double_wide); > > - intel_dpll_dump_hw_state(dev_priv, &pipe_config->dpll_hw_state); > > + intel_dpll_dump_hw_state(i915, &pipe_config->dpll_hw_state); > > dump_planes: > > if (!state) > > @@ -12073,7 +12073,7 @@ static bool check_digital_port_conflicts(struct intel_atomic_state *state) > > static int > > clear_intel_crtc_state(struct intel_crtc_state *crtc_state) > > { > > - struct drm_i915_private *dev_priv = > > + struct drm_i915_private *i915 = > > to_i915(crtc_state->base.crtc->dev); > > struct intel_crtc_state *saved_state; > > @@ -12090,8 +12090,8 @@ clear_intel_crtc_state(struct intel_crtc_state *crtc_state) > > saved_state->shared_dpll = crtc_state->shared_dpll; > > saved_state->dpll_hw_state = crtc_state->dpll_hw_state; > > saved_state->crc_enabled = crtc_state->crc_enabled; > > - if (IS_G4X(dev_priv) || > > - IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) > > + if (IS_G4X(i915) || > > + IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) > > saved_state->wm = crtc_state->wm; > > /* Keep base drm_crtc_state intact, only clear our extended struct */ > > @@ -12316,7 +12316,7 @@ intel_compare_infoframe(const union hdmi_infoframe *a, > > } > > static void > > -pipe_config_infoframe_err(struct drm_i915_private *dev_priv, > > +pipe_config_infoframe_err(struct drm_i915_private *i915, > > bool adjust, const char *name, > > const union hdmi_infoframe *a, > > const union hdmi_infoframe *b) > > @@ -12327,15 +12327,15 @@ pipe_config_infoframe_err(struct drm_i915_private *dev_priv, > > drm_dbg(DRM_UT_KMS, "mismatch in %s infoframe", name); > > drm_dbg(DRM_UT_KMS, "expected:"); > > - hdmi_infoframe_log(KERN_DEBUG, dev_priv->drm.dev, a); > > + hdmi_infoframe_log(KERN_DEBUG, i915->drm.dev, a); > > drm_dbg(DRM_UT_KMS, "found"); > > - hdmi_infoframe_log(KERN_DEBUG, dev_priv->drm.dev, b); > > + hdmi_infoframe_log(KERN_DEBUG, i915->drm.dev, b); > > } else { > > drm_err("mismatch in %s infoframe", name); > > drm_err("expected:"); > > - hdmi_infoframe_log(KERN_ERR, dev_priv->drm.dev, a); > > + hdmi_infoframe_log(KERN_ERR, i915->drm.dev, a); > > drm_err("found"); > > - hdmi_infoframe_log(KERN_ERR, dev_priv->drm.dev, b); > > + hdmi_infoframe_log(KERN_ERR, i915->drm.dev, b); > > } > > } > > @@ -12357,17 +12357,17 @@ pipe_config_err(bool adjust, const char *name, const char *format, ...) > > va_end(args); > > } > > -static bool fastboot_enabled(struct drm_i915_private *dev_priv) > > +static bool fastboot_enabled(struct drm_i915_private *i915) > > { > > if (i915_modparams.fastboot != -1) > > return i915_modparams.fastboot; > > /* Enable fastboot by default on Skylake and newer */ > > - if (INTEL_GEN(dev_priv) >= 9) > > + if (INTEL_GEN(i915) >= 9) > > return true; > > /* Enable fastboot by default on VLV and CHV */ > > - if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) > > + if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) > > return true; > > /* Disabled by default on all others */ > > @@ -12375,7 +12375,7 @@ static bool fastboot_enabled(struct drm_i915_private *dev_priv) > > } > > static bool > > -intel_pipe_config_compare(struct drm_i915_private *dev_priv, > > +intel_pipe_config_compare(struct drm_i915_private *i915, > > struct intel_crtc_state *current_config, > > struct intel_crtc_state *pipe_config, > > bool adjust) > > @@ -12385,7 +12385,7 @@ intel_pipe_config_compare(struct drm_i915_private *dev_priv, > > (current_config->base.mode.private_flags & I915_MODE_FLAG_INHERITED) && > > !(pipe_config->base.mode.private_flags & I915_MODE_FLAG_INHERITED); > > - if (fixup_inherited && !fastboot_enabled(dev_priv)) { > > + if (fixup_inherited && !fastboot_enabled(i915)) { > > DRM_DEBUG_KMS("initial modeset and fastboot not set\n"); > > ret = false; > > } > > @@ -12525,7 +12525,7 @@ intel_pipe_config_compare(struct drm_i915_private *dev_priv, > > #define PIPE_CONF_CHECK_INFOFRAME(name) do { \ > > if (!intel_compare_infoframe(¤t_config->infoframes.name, \ > > &pipe_config->infoframes.name)) { \ > > - pipe_config_infoframe_err(dev_priv, adjust, __stringify(name), \ > > + pipe_config_infoframe_err(i915, adjust, __stringify(name), \ > > ¤t_config->infoframes.name, \ > > &pipe_config->infoframes.name); \ > > ret = false; \ > > @@ -12544,7 +12544,7 @@ intel_pipe_config_compare(struct drm_i915_private *dev_priv, > > PIPE_CONF_CHECK_I(lane_count); > > PIPE_CONF_CHECK_X(lane_lat_optim_mask); > > - if (INTEL_GEN(dev_priv) < 8) { > > + if (INTEL_GEN(i915) < 8) { > > PIPE_CONF_CHECK_M_N(dp_m_n); > > if (current_config->has_drrs) > > @@ -12571,8 +12571,8 @@ intel_pipe_config_compare(struct drm_i915_private *dev_priv, > > PIPE_CONF_CHECK_I(pixel_multiplier); > > PIPE_CONF_CHECK_I(output_format); > > PIPE_CONF_CHECK_BOOL(has_hdmi_sink); > > - if ((INTEL_GEN(dev_priv) < 8 && !IS_HASWELL(dev_priv)) || > > - IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) > > + if ((INTEL_GEN(i915) < 8 && !IS_HASWELL(i915)) || > > + IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) > > PIPE_CONF_CHECK_BOOL(limited_color_range); > > PIPE_CONF_CHECK_BOOL(hdmi_scrambling); > > @@ -12597,7 +12597,7 @@ intel_pipe_config_compare(struct drm_i915_private *dev_priv, > > PIPE_CONF_CHECK_X(gmch_pfit.control); > > /* pfit ratios are autocomputed by the hw on gen4+ */ > > - if (INTEL_GEN(dev_priv) < 4) > > + if (INTEL_GEN(i915) < 4) > > PIPE_CONF_CHECK_X(gmch_pfit.pgm_ratios); > > PIPE_CONF_CHECK_X(gmch_pfit.lvds_border_bits); > > @@ -12621,7 +12621,7 @@ intel_pipe_config_compare(struct drm_i915_private *dev_priv, > > PIPE_CONF_CHECK_CLOCK_FUZZY(pixel_rate); > > PIPE_CONF_CHECK_X(gamma_mode); > > - if (IS_CHERRYVIEW(dev_priv)) > > + if (IS_CHERRYVIEW(i915)) > > PIPE_CONF_CHECK_X(cgm_mode); > > else > > PIPE_CONF_CHECK_X(csc_mode); > > @@ -12667,7 +12667,7 @@ intel_pipe_config_compare(struct drm_i915_private *dev_priv, > > PIPE_CONF_CHECK_X(dsi_pll.ctrl); > > PIPE_CONF_CHECK_X(dsi_pll.div); > > - if (IS_G4X(dev_priv) || INTEL_GEN(dev_priv) >= 5) > > + if (IS_G4X(i915) || INTEL_GEN(i915) >= 5) > > PIPE_CONF_CHECK_I(pipe_bpp); > > PIPE_CONF_CHECK_CLOCK_FUZZY(base.adjusted_mode.crtc_clock); > > @@ -12694,11 +12694,11 @@ intel_pipe_config_compare(struct drm_i915_private *dev_priv, > > return ret; > > } > > -static void intel_pipe_config_sanity_check(struct drm_i915_private *dev_priv, > > +static void intel_pipe_config_sanity_check(struct drm_i915_private *i915, > > const struct intel_crtc_state *pipe_config) > > { > > if (pipe_config->has_pch_encoder) { > > - int fdi_dotclock = intel_dotclock_calculate(intel_fdi_link_freq(dev_priv, pipe_config), > > + int fdi_dotclock = intel_dotclock_calculate(intel_fdi_link_freq(i915, pipe_config), > > &pipe_config->fdi_m_n); > > int dotclock = pipe_config->base.adjusted_mode.crtc_clock; > > @@ -12715,7 +12715,7 @@ static void intel_pipe_config_sanity_check(struct drm_i915_private *dev_priv, > > static void verify_wm_state(struct drm_crtc *crtc, > > struct drm_crtc_state *new_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc->dev); > > + struct drm_i915_private *i915 = to_i915(crtc->dev); > > struct skl_hw_state { > > struct skl_ddb_entry ddb_y[I915_MAX_PLANES]; > > struct skl_ddb_entry ddb_uv[I915_MAX_PLANES]; > > @@ -12727,9 +12727,9 @@ static void verify_wm_state(struct drm_crtc *crtc, > > struct skl_ddb_entry *hw_ddb_entry, *sw_ddb_entry; > > struct intel_crtc *intel_crtc = to_intel_crtc(crtc); > > const enum pipe pipe = intel_crtc->pipe; > > - int plane, level, max_level = ilk_wm_max_level(dev_priv); > > + int plane, level, max_level = ilk_wm_max_level(i915); > > - if (INTEL_GEN(dev_priv) < 9 || !new_state->active) > > + if (INTEL_GEN(i915) < 9 || !new_state->active) > > return; > > hw = kzalloc(sizeof(*hw), GFP_KERNEL); > > @@ -12741,17 +12741,17 @@ static void verify_wm_state(struct drm_crtc *crtc, > > skl_pipe_ddb_get_hw_state(intel_crtc, hw->ddb_y, hw->ddb_uv); > > - skl_ddb_get_hw_state(dev_priv, &hw->ddb); > > - sw_ddb = &dev_priv->wm.skl_hw.ddb; > > + skl_ddb_get_hw_state(i915, &hw->ddb); > > + sw_ddb = &i915->wm.skl_hw.ddb; > > - if (INTEL_GEN(dev_priv) >= 11 && > > + if (INTEL_GEN(i915) >= 11 && > > hw->ddb.enabled_slices != sw_ddb->enabled_slices) > > DRM_ERROR("mismatch in DBUF Slices (expected %u, got %u)\n", > > sw_ddb->enabled_slices, > > hw->ddb.enabled_slices); > > /* planes */ > > - for_each_universal_plane(dev_priv, pipe, plane) { > > + for_each_universal_plane(i915, pipe, plane) { > > struct skl_plane_wm *hw_plane_wm, *sw_plane_wm; > > hw_plane_wm = &hw->wm.planes[plane]; > > @@ -12933,7 +12933,7 @@ verify_crtc_state(struct drm_crtc *crtc, > > struct drm_crtc_state *new_crtc_state) > > { > > struct drm_device *dev = crtc->dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > struct intel_encoder *encoder; > > struct intel_crtc *intel_crtc = to_intel_crtc(crtc); > > struct intel_crtc_state *pipe_config, *sw_config; > > @@ -12949,10 +12949,10 @@ verify_crtc_state(struct drm_crtc *crtc, > > DRM_DEBUG_KMS("[CRTC:%d:%s]\n", crtc->base.id, crtc->name); > > - active = dev_priv->display.get_pipe_config(intel_crtc, pipe_config); > > + active = i915->display.get_pipe_config(intel_crtc, pipe_config); > > /* we keep both pipes enabled on 830 */ > > - if (IS_I830(dev_priv)) > > + if (IS_I830(i915)) > > active = new_crtc_state->active; > > I915_STATE_WARN(new_crtc_state->active != active, > > @@ -12984,10 +12984,10 @@ verify_crtc_state(struct drm_crtc *crtc, > > if (!new_crtc_state->active) > > return; > > - intel_pipe_config_sanity_check(dev_priv, pipe_config); > > + intel_pipe_config_sanity_check(i915, pipe_config); > > sw_config = to_intel_crtc_state(new_crtc_state); > > - if (!intel_pipe_config_compare(dev_priv, sw_config, > > + if (!intel_pipe_config_compare(i915, sw_config, > > pipe_config, false)) { > > I915_STATE_WARN(1, "pipe state doesn't match!\n"); > > intel_dump_pipe_config(pipe_config, NULL, "[hw state]"); > > @@ -13009,7 +13009,7 @@ intel_verify_planes(struct intel_atomic_state *state) > > } > > static void > > -verify_single_dpll_state(struct drm_i915_private *dev_priv, > > +verify_single_dpll_state(struct drm_i915_private *i915, > > struct intel_shared_dpll *pll, > > struct drm_crtc *crtc, > > struct drm_crtc_state *new_state) > > @@ -13022,7 +13022,7 @@ verify_single_dpll_state(struct drm_i915_private *dev_priv, > > DRM_DEBUG_KMS("%s\n", pll->info->name); > > - active = pll->info->funcs->get_hw_state(dev_priv, pll, &dpll_hw_state); > > + active = pll->info->funcs->get_hw_state(i915, pll, &dpll_hw_state); > > if (!(pll->info->flags & INTEL_DPLL_ALWAYS_ON)) { > > I915_STATE_WARN(!pll->on && pll->active_mask, > > @@ -13068,12 +13068,12 @@ verify_shared_dpll_state(struct drm_device *dev, struct drm_crtc *crtc, > > struct drm_crtc_state *old_crtc_state, > > struct drm_crtc_state *new_crtc_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > struct intel_crtc_state *old_state = to_intel_crtc_state(old_crtc_state); > > struct intel_crtc_state *new_state = to_intel_crtc_state(new_crtc_state); > > if (new_state->shared_dpll) > > - verify_single_dpll_state(dev_priv, new_state->shared_dpll, crtc, new_crtc_state); > > + verify_single_dpll_state(i915, new_state->shared_dpll, crtc, new_crtc_state); > > if (old_state->shared_dpll && > > old_state->shared_dpll != new_state->shared_dpll) { > > @@ -13108,11 +13108,11 @@ intel_modeset_verify_crtc(struct drm_crtc *crtc, > > static void > > verify_disabled_dpll_state(struct drm_device *dev) > > { > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > int i; > > - for (i = 0; i < dev_priv->num_shared_dpll; i++) > > - verify_single_dpll_state(dev_priv, &dev_priv->shared_dplls[i], NULL, NULL); > > + for (i = 0; i < i915->num_shared_dpll; i++) > > + verify_single_dpll_state(i915, &i915->shared_dplls[i], NULL, NULL); > > } > > static void > > @@ -13127,7 +13127,7 @@ intel_modeset_verify_disabled(struct drm_device *dev, > > static void update_scanline_offset(const struct intel_crtc_state *crtc_state) > > { > > struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc); > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > /* > > * The scanline counter increments at the leading edge of hsync. > > @@ -13156,7 +13156,7 @@ static void update_scanline_offset(const struct intel_crtc_state *crtc_state) > > * However if queried just before the start of vblank we'll get an > > * answer that's slightly in the future. > > */ > > - if (IS_GEN(dev_priv, 2)) { > > + if (IS_GEN(i915, 2)) { > > const struct drm_display_mode *adjusted_mode = &crtc_state->base.adjusted_mode; > > int vtotal; > > @@ -13165,7 +13165,7 @@ static void update_scanline_offset(const struct intel_crtc_state *crtc_state) > > vtotal /= 2; > > crtc->scanline_offset = vtotal - 1; > > - } else if (HAS_DDI(dev_priv) && > > + } else if (HAS_DDI(i915) && > > intel_crtc_has_type(crtc_state, INTEL_OUTPUT_HDMI)) { > > crtc->scanline_offset = 2; > > } else > > @@ -13174,12 +13174,12 @@ static void update_scanline_offset(const struct intel_crtc_state *crtc_state) > > static void intel_modeset_clear_plls(struct intel_atomic_state *state) > > { > > - struct drm_i915_private *dev_priv = to_i915(state->base.dev); > > + struct drm_i915_private *i915 = to_i915(state->base.dev); > > struct intel_crtc_state *old_crtc_state, *new_crtc_state; > > struct intel_crtc *crtc; > > int i; > > - if (!dev_priv->display.crtc_compute_clock) > > + if (!i915->display.crtc_compute_clock) > > return; > > for_each_oldnew_intel_crtc_in_state(state, crtc, old_crtc_state, > > @@ -13311,7 +13311,7 @@ static int intel_modeset_all_pipes(struct drm_atomic_state *state) > > static int intel_modeset_checks(struct intel_atomic_state *state) > > { > > - struct drm_i915_private *dev_priv = to_i915(state->base.dev); > > + struct drm_i915_private *i915 = to_i915(state->base.dev); > > struct intel_crtc_state *old_crtc_state, *new_crtc_state; > > struct intel_crtc *crtc; > > int ret = 0, i; > > @@ -13323,12 +13323,12 @@ static int intel_modeset_checks(struct intel_atomic_state *state) > > /* keep the current setting */ > > if (!state->cdclk.force_min_cdclk_changed) > > - state->cdclk.force_min_cdclk = dev_priv->cdclk.force_min_cdclk; > > + state->cdclk.force_min_cdclk = i915->cdclk.force_min_cdclk; > > state->modeset = true; > > - state->active_crtcs = dev_priv->active_crtcs; > > - state->cdclk.logical = dev_priv->cdclk.logical; > > - state->cdclk.actual = dev_priv->cdclk.actual; > > + state->active_crtcs = i915->active_crtcs; > > + state->cdclk.logical = i915->cdclk.logical; > > + state->cdclk.actual = i915->cdclk.actual; > > state->cdclk.pipe = INVALID_PIPE; > > for_each_oldnew_intel_crtc_in_state(state, crtc, old_crtc_state, > > @@ -13349,19 +13349,19 @@ static int intel_modeset_checks(struct intel_atomic_state *state) > > * mode set on this crtc. For other crtcs we need to use the > > * adjusted_mode bits in the crtc directly. > > */ > > - if (dev_priv->display.modeset_calc_cdclk) { > > + if (i915->display.modeset_calc_cdclk) { > > enum pipe pipe; > > - ret = dev_priv->display.modeset_calc_cdclk(state); > > + ret = i915->display.modeset_calc_cdclk(state); > > if (ret < 0) > > return ret; > > /* > > - * Writes to dev_priv->cdclk.logical must protected by > > + * Writes to i915->cdclk.logical must protected by > > * holding all the crtc locks, even if we don't end up > > * touching the hardware > > */ > > - if (intel_cdclk_changed(&dev_priv->cdclk.logical, > > + if (intel_cdclk_changed(&i915->cdclk.logical, > > &state->cdclk.logical)) { > > ret = intel_lock_all_pipes(&state->base); > > if (ret < 0) > > @@ -13373,7 +13373,7 @@ static int intel_modeset_checks(struct intel_atomic_state *state) > > struct drm_crtc_state *crtc_state; > > pipe = ilog2(state->active_crtcs); > > - crtc = &intel_get_crtc_for_pipe(dev_priv, pipe)->base; > > + crtc = &intel_get_crtc_for_pipe(i915, pipe)->base; > > crtc_state = drm_atomic_get_new_crtc_state(&state->base, crtc); > > if (crtc_state && needs_modeset(crtc_state)) > > pipe = INVALID_PIPE; > > @@ -13383,15 +13383,15 @@ static int intel_modeset_checks(struct intel_atomic_state *state) > > /* All pipes must be switched off while we change the cdclk. */ > > if (pipe != INVALID_PIPE && > > - intel_cdclk_needs_cd2x_update(dev_priv, > > - &dev_priv->cdclk.actual, > > + intel_cdclk_needs_cd2x_update(i915, > > + &i915->cdclk.actual, > > &state->cdclk.actual)) { > > ret = intel_lock_all_pipes(&state->base); > > if (ret < 0) > > return ret; > > state->cdclk.pipe = pipe; > > - } else if (intel_cdclk_needs_modeset(&dev_priv->cdclk.actual, > > + } else if (intel_cdclk_needs_modeset(&i915->cdclk.actual, > > &state->cdclk.actual)) { > > ret = intel_modeset_all_pipes(&state->base); > > if (ret < 0) > > @@ -13410,7 +13410,7 @@ static int intel_modeset_checks(struct intel_atomic_state *state) > > intel_modeset_clear_plls(state); > > - if (IS_HASWELL(dev_priv)) > > + if (IS_HASWELL(i915)) > > return haswell_mode_set_planes_workaround(state); > > return 0; > > @@ -13424,11 +13424,11 @@ static int intel_modeset_checks(struct intel_atomic_state *state) > > static int calc_watermark_data(struct intel_atomic_state *state) > > { > > struct drm_device *dev = state->base.dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > /* Is there platform-specific watermark information to calculate? */ > > - if (dev_priv->display.compute_global_watermarks) > > - return dev_priv->display.compute_global_watermarks(state); > > + if (i915->display.compute_global_watermarks) > > + return i915->display.compute_global_watermarks(state); > > return 0; > > } > > @@ -13441,7 +13441,7 @@ static int calc_watermark_data(struct intel_atomic_state *state) > > static int intel_atomic_check(struct drm_device *dev, > > struct drm_atomic_state *_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > struct intel_atomic_state *state = to_intel_atomic_state(_state); > > struct intel_crtc_state *old_crtc_state, *new_crtc_state; > > struct intel_crtc *crtc; > > @@ -13474,7 +13474,7 @@ static int intel_atomic_check(struct drm_device *dev, > > if (ret) > > goto fail; > > - if (intel_pipe_config_compare(dev_priv, old_crtc_state, > > + if (intel_pipe_config_compare(i915, old_crtc_state, > > new_crtc_state, true)) { > > new_crtc_state->base.mode_changed = false; > > new_crtc_state->update_pipe = true; > > @@ -13493,7 +13493,7 @@ static int intel_atomic_check(struct drm_device *dev, > > if (ret) > > goto fail; > > } else { > > - state->cdclk.logical = dev_priv->cdclk.logical; > > + state->cdclk.logical = i915->cdclk.logical; > > } > > ret = icl_add_linked_planes(state); > > @@ -13504,7 +13504,7 @@ static int intel_atomic_check(struct drm_device *dev, > > if (ret) > > goto fail; > > - intel_fbc_choose_crtc(dev_priv, state); > > + intel_fbc_choose_crtc(i915, state); > > ret = calc_watermark_data(state); > > if (ret) > > goto fail; > > @@ -13564,7 +13564,7 @@ static void intel_update_crtc(struct drm_crtc *crtc, > > struct drm_crtc_state *new_crtc_state) > > { > > struct drm_device *dev = crtc->dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > struct intel_crtc *intel_crtc = to_intel_crtc(crtc); > > struct intel_crtc_state *pipe_config = to_intel_crtc_state(new_crtc_state); > > bool modeset = needs_modeset(new_crtc_state); > > @@ -13574,7 +13574,7 @@ static void intel_update_crtc(struct drm_crtc *crtc, > > if (modeset) { > > update_scanline_offset(pipe_config); > > - dev_priv->display.crtc_enable(pipe_config, state); > > + i915->display.crtc_enable(pipe_config, state); > > /* vblanks work again, re-enable pipe CRC. */ > > intel_crtc_enable_pipe_crc(intel_crtc); > > @@ -13593,7 +13593,7 @@ static void intel_update_crtc(struct drm_crtc *crtc, > > intel_begin_crtc_commit(to_intel_atomic_state(state), intel_crtc); > > - if (INTEL_GEN(dev_priv) >= 9) > > + if (INTEL_GEN(i915) >= 9) > > skl_update_planes_on_crtc(to_intel_atomic_state(state), intel_crtc); > > else > > i9xx_update_planes_on_crtc(to_intel_atomic_state(state), intel_crtc); > > @@ -13618,7 +13618,7 @@ static void intel_update_crtcs(struct drm_atomic_state *state) > > static void skl_update_crtcs(struct drm_atomic_state *state) > > { > > - struct drm_i915_private *dev_priv = to_i915(state->dev); > > + struct drm_i915_private *i915 = to_i915(state->dev); > > struct intel_atomic_state *intel_state = to_intel_atomic_state(state); > > struct drm_crtc *crtc; > > struct intel_crtc *intel_crtc; > > @@ -13628,7 +13628,7 @@ static void skl_update_crtcs(struct drm_atomic_state *state) > > bool progress; > > enum pipe pipe; > > int i; > > - u8 hw_enabled_slices = dev_priv->wm.skl_hw.ddb.enabled_slices; > > + u8 hw_enabled_slices = i915->wm.skl_hw.ddb.enabled_slices; > > u8 required_slices = intel_state->wm_results.ddb.enabled_slices; > > struct skl_ddb_entry entries[I915_MAX_PIPES] = {}; > > @@ -13638,8 +13638,8 @@ static void skl_update_crtcs(struct drm_atomic_state *state) > > entries[i] = to_intel_crtc_state(old_crtc_state)->wm.skl.ddb; > > /* If 2nd DBuf slice required, enable it here */ > > - if (INTEL_GEN(dev_priv) >= 11 && required_slices > hw_enabled_slices) > > - icl_dbuf_slices_update(dev_priv, required_slices); > > + if (INTEL_GEN(i915) >= 11 && required_slices > hw_enabled_slices) > > + icl_dbuf_slices_update(i915, required_slices); > > /* > > * Whenever the number of active pipes changes, we need to make sure we > > @@ -13663,7 +13663,7 @@ static void skl_update_crtcs(struct drm_atomic_state *state) > > if (skl_ddb_allocation_overlaps(&cstate->wm.skl.ddb, > > entries, > > - INTEL_INFO(dev_priv)->num_pipes, i)) > > + INTEL_INFO(i915)->num_pipes, i)) > > continue; > > updated |= cmask; > > @@ -13685,57 +13685,57 @@ static void skl_update_crtcs(struct drm_atomic_state *state) > > new_crtc_state); > > if (vbl_wait) > > - intel_wait_for_vblank(dev_priv, pipe); > > + intel_wait_for_vblank(i915, pipe); > > progress = true; > > } > > } while (progress); > > /* If 2nd DBuf slice is no more required disable it */ > > - if (INTEL_GEN(dev_priv) >= 11 && required_slices < hw_enabled_slices) > > - icl_dbuf_slices_update(dev_priv, required_slices); > > + if (INTEL_GEN(i915) >= 11 && required_slices < hw_enabled_slices) > > + icl_dbuf_slices_update(i915, required_slices); > > } > > -static void intel_atomic_helper_free_state(struct drm_i915_private *dev_priv) > > +static void intel_atomic_helper_free_state(struct drm_i915_private *i915) > > { > > struct intel_atomic_state *state, *next; > > struct llist_node *freed; > > - freed = llist_del_all(&dev_priv->atomic_helper.free_list); > > + freed = llist_del_all(&i915->atomic_helper.free_list); > > llist_for_each_entry_safe(state, next, freed, freed) > > drm_atomic_state_put(&state->base); > > } > > static void intel_atomic_helper_free_state_worker(struct work_struct *work) > > { > > - struct drm_i915_private *dev_priv = > > - container_of(work, typeof(*dev_priv), atomic_helper.free_work); > > + struct drm_i915_private *i915 = > > + container_of(work, typeof(*i915), atomic_helper.free_work); > > - intel_atomic_helper_free_state(dev_priv); > > + intel_atomic_helper_free_state(i915); > > } > > static void intel_atomic_commit_fence_wait(struct intel_atomic_state *intel_state) > > { > > struct wait_queue_entry wait_fence, wait_reset; > > - struct drm_i915_private *dev_priv = to_i915(intel_state->base.dev); > > + struct drm_i915_private *i915 = to_i915(intel_state->base.dev); > > init_wait_entry(&wait_fence, 0); > > init_wait_entry(&wait_reset, 0); > > for (;;) { > > prepare_to_wait(&intel_state->commit_ready.wait, > > &wait_fence, TASK_UNINTERRUPTIBLE); > > - prepare_to_wait(&dev_priv->gpu_error.wait_queue, > > + prepare_to_wait(&i915->gpu_error.wait_queue, > > &wait_reset, TASK_UNINTERRUPTIBLE); > > if (i915_sw_fence_done(&intel_state->commit_ready) > > - || test_bit(I915_RESET_MODESET, &dev_priv->gpu_error.flags)) > > + || test_bit(I915_RESET_MODESET, &i915->gpu_error.flags)) > > break; > > schedule(); > > } > > finish_wait(&intel_state->commit_ready.wait, &wait_fence); > > - finish_wait(&dev_priv->gpu_error.wait_queue, &wait_reset); > > + finish_wait(&i915->gpu_error.wait_queue, &wait_reset); > > } > > static void intel_atomic_cleanup_work(struct work_struct *work) > > @@ -13755,7 +13755,7 @@ static void intel_atomic_commit_tail(struct drm_atomic_state *state) > > { > > struct drm_device *dev = state->dev; > > struct intel_atomic_state *intel_state = to_intel_atomic_state(state); > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > struct drm_crtc_state *old_crtc_state, *new_crtc_state; > > struct intel_crtc_state *new_intel_crtc_state, *old_intel_crtc_state; > > struct drm_crtc *crtc; > > @@ -13769,7 +13769,7 @@ static void intel_atomic_commit_tail(struct drm_atomic_state *state) > > drm_atomic_helper_wait_for_dependencies(state); > > if (intel_state->modeset) > > - wakeref = intel_display_power_get(dev_priv, POWER_DOMAIN_MODESET); > > + wakeref = intel_display_power_get(i915, POWER_DOMAIN_MODESET); > > for_each_oldnew_crtc_in_state(state, crtc, old_crtc_state, new_crtc_state, i) { > > old_intel_crtc_state = to_intel_crtc_state(old_crtc_state); > > @@ -13798,7 +13798,7 @@ static void intel_atomic_commit_tail(struct drm_atomic_state *state) > > */ > > intel_crtc_disable_pipe_crc(intel_crtc); > > - dev_priv->display.crtc_disable(old_intel_crtc_state, state); > > + i915->display.crtc_disable(old_intel_crtc_state, state); > > intel_crtc->active = false; > > intel_fbc_disable(intel_crtc); > > intel_disable_shared_dpll(old_intel_crtc_state); > > @@ -13807,14 +13807,14 @@ static void intel_atomic_commit_tail(struct drm_atomic_state *state) > > * Underruns don't always raise > > * interrupts, so check manually. > > */ > > - intel_check_cpu_fifo_underruns(dev_priv); > > - intel_check_pch_fifo_underruns(dev_priv); > > + intel_check_cpu_fifo_underruns(i915); > > + intel_check_pch_fifo_underruns(i915); > > /* FIXME unify this for all platforms */ > > if (!new_crtc_state->active && > > - !HAS_GMCH(dev_priv) && > > - dev_priv->display.initial_watermarks) > > - dev_priv->display.initial_watermarks(intel_state, > > + !HAS_GMCH(i915) && > > + i915->display.initial_watermarks) > > + i915->display.initial_watermarks(intel_state, > > new_intel_crtc_state); > > } > > } > > @@ -13826,9 +13826,9 @@ static void intel_atomic_commit_tail(struct drm_atomic_state *state) > > if (intel_state->modeset) { > > drm_atomic_helper_update_legacy_modeset_state(state->dev, state); > > - intel_set_cdclk_pre_plane_update(dev_priv, > > + intel_set_cdclk_pre_plane_update(i915, > > &intel_state->cdclk.actual, > > - &dev_priv->cdclk.actual, > > + &i915->cdclk.actual, > > intel_state->cdclk.pipe); > > /* > > @@ -13836,7 +13836,7 @@ static void intel_atomic_commit_tail(struct drm_atomic_state *state) > > * have more then one pipe enabled > > */ > > if (!intel_can_enable_sagv(state)) > > - intel_disable_sagv(dev_priv); > > + intel_disable_sagv(i915); > > intel_modeset_verify_disabled(dev, state); > > } > > @@ -13856,12 +13856,12 @@ static void intel_atomic_commit_tail(struct drm_atomic_state *state) > > } > > /* Now enable the clocks, plane, pipe, and connectors that we set up. */ > > - dev_priv->display.update_crtcs(state); > > + i915->display.update_crtcs(state); > > if (intel_state->modeset) > > - intel_set_cdclk_post_plane_update(dev_priv, > > + intel_set_cdclk_post_plane_update(i915, > > &intel_state->cdclk.actual, > > - &dev_priv->cdclk.actual, > > + &i915->cdclk.actual, > > intel_state->cdclk.pipe); > > /* FIXME: We should call drm_atomic_helper_commit_hw_done() here > > @@ -13895,8 +13895,8 @@ static void intel_atomic_commit_tail(struct drm_atomic_state *state) > > for_each_new_crtc_in_state(state, crtc, new_crtc_state, i) { > > new_intel_crtc_state = to_intel_crtc_state(new_crtc_state); > > - if (dev_priv->display.optimize_watermarks) > > - dev_priv->display.optimize_watermarks(intel_state, > > + if (i915->display.optimize_watermarks) > > + i915->display.optimize_watermarks(intel_state, > > new_intel_crtc_state); > > } > > @@ -13904,7 +13904,7 @@ static void intel_atomic_commit_tail(struct drm_atomic_state *state) > > intel_post_plane_update(to_intel_crtc_state(old_crtc_state)); > > if (put_domains[i]) > > - modeset_put_power_domains(dev_priv, put_domains[i]); > > + modeset_put_power_domains(i915, put_domains[i]); > > intel_modeset_verify_crtc(crtc, state, old_crtc_state, new_crtc_state); > > } > > @@ -13913,7 +13913,7 @@ static void intel_atomic_commit_tail(struct drm_atomic_state *state) > > intel_verify_planes(intel_state); > > if (intel_state->modeset && intel_can_enable_sagv(state)) > > - intel_enable_sagv(dev_priv); > > + intel_enable_sagv(i915); > > drm_atomic_helper_commit_hw_done(state); > > @@ -13924,10 +13924,10 @@ static void intel_atomic_commit_tail(struct drm_atomic_state *state) > > * so enable debugging for the next modeset - and hope we catch > > * the culprit. > > */ > > - intel_uncore_arm_unclaimed_mmio_detection(&dev_priv->uncore); > > - intel_display_power_put(dev_priv, POWER_DOMAIN_MODESET, wakeref); > > + intel_uncore_arm_unclaimed_mmio_detection(&i915->uncore); > > + intel_display_power_put(i915, POWER_DOMAIN_MODESET, wakeref); > > } > > - intel_runtime_pm_put(dev_priv, intel_state->wakeref); > > + intel_runtime_pm_put(i915, intel_state->wakeref); > > /* > > * Defer the cleanup of the old state to a separate worker to not > > @@ -14003,10 +14003,10 @@ static int intel_atomic_commit(struct drm_device *dev, > > bool nonblock) > > { > > struct intel_atomic_state *intel_state = to_intel_atomic_state(state); > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > int ret = 0; > > - intel_state->wakeref = intel_runtime_pm_get(dev_priv); > > + intel_state->wakeref = intel_runtime_pm_get(i915); > > drm_atomic_state_get(state); > > i915_sw_fence_init(&intel_state->commit_ready, > > @@ -14029,7 +14029,7 @@ static int intel_atomic_commit(struct drm_device *dev, > > * FIXME doing watermarks and fb cleanup from a vblank worker > > * (assuming we had any) would solve these problems. > > */ > > - if (INTEL_GEN(dev_priv) < 9 && state->legacy_cursor_update) { > > + if (INTEL_GEN(i915) < 9 && state->legacy_cursor_update) { > > struct intel_crtc_state *new_crtc_state; > > struct intel_crtc *crtc; > > int i; > > @@ -14044,7 +14044,7 @@ static int intel_atomic_commit(struct drm_device *dev, > > if (ret) { > > DRM_DEBUG_ATOMIC("Preparing state failed with %i\n", ret); > > i915_sw_fence_commit(&intel_state->commit_ready); > > - intel_runtime_pm_put(dev_priv, intel_state->wakeref); > > + intel_runtime_pm_put(i915, intel_state->wakeref); > > return ret; > > } > > @@ -14056,21 +14056,21 @@ static int intel_atomic_commit(struct drm_device *dev, > > i915_sw_fence_commit(&intel_state->commit_ready); > > drm_atomic_helper_cleanup_planes(dev, state); > > - intel_runtime_pm_put(dev_priv, intel_state->wakeref); > > + intel_runtime_pm_put(i915, intel_state->wakeref); > > return ret; > > } > > - dev_priv->wm.distrust_bios_wm = false; > > + i915->wm.distrust_bios_wm = false; > > intel_shared_dpll_swap_state(state); > > intel_atomic_track_fbs(state); > > if (intel_state->modeset) { > > - memcpy(dev_priv->min_cdclk, intel_state->min_cdclk, > > + memcpy(i915->min_cdclk, intel_state->min_cdclk, > > sizeof(intel_state->min_cdclk)); > > - memcpy(dev_priv->min_voltage_level, > > + memcpy(i915->min_voltage_level, > > intel_state->min_voltage_level, > > sizeof(intel_state->min_voltage_level)); > > - dev_priv->active_crtcs = intel_state->active_crtcs; > > - dev_priv->cdclk.force_min_cdclk = > > + i915->active_crtcs = intel_state->active_crtcs; > > + i915->cdclk.force_min_cdclk = > > intel_state->cdclk.force_min_cdclk; > > intel_cdclk_swap_state(intel_state); > > @@ -14081,12 +14081,12 @@ static int intel_atomic_commit(struct drm_device *dev, > > i915_sw_fence_commit(&intel_state->commit_ready); > > if (nonblock && intel_state->modeset) { > > - queue_work(dev_priv->modeset_wq, &state->commit_work); > > + queue_work(i915->modeset_wq, &state->commit_work); > > } else if (nonblock) { > > queue_work(system_unbound_wq, &state->commit_work); > > } else { > > if (intel_state->modeset) > > - flush_workqueue(dev_priv->modeset_wq); > > + flush_workqueue(i915->modeset_wq); > > intel_atomic_commit_tail(state); > > } > > @@ -14166,14 +14166,14 @@ static void add_rps_boost_after_vblank(struct drm_crtc *crtc, > > static int intel_plane_pin_fb(struct intel_plane_state *plane_state) > > { > > struct intel_plane *plane = to_intel_plane(plane_state->base.plane); > > - struct drm_i915_private *dev_priv = to_i915(plane->base.dev); > > + struct drm_i915_private *i915 = to_i915(plane->base.dev); > > struct drm_framebuffer *fb = plane_state->base.fb; > > struct i915_vma *vma; > > if (plane->id == PLANE_CURSOR && > > - INTEL_INFO(dev_priv)->display.cursor_needs_physical) { > > + INTEL_INFO(i915)->display.cursor_needs_physical) { > > struct drm_i915_gem_object *obj = intel_fb_obj(fb); > > - const int align = intel_cursor_alignment(dev_priv); > > + const int align = intel_cursor_alignment(i915); > > int err; > > err = i915_gem_object_attach_phys(obj, align); > > @@ -14231,7 +14231,7 @@ intel_prepare_plane_fb(struct drm_plane *plane, > > { > > struct intel_atomic_state *intel_state = > > to_intel_atomic_state(new_state->state); > > - struct drm_i915_private *dev_priv = to_i915(plane->dev); > > + struct drm_i915_private *i915 = to_i915(plane->dev); > > struct drm_framebuffer *fb = new_state->fb; > > struct drm_i915_gem_object *obj = intel_fb_obj(fb); > > struct drm_i915_gem_object *old_obj = intel_fb_obj(plane->state->fb); > > @@ -14279,7 +14279,7 @@ intel_prepare_plane_fb(struct drm_plane *plane, > > if (ret) > > return ret; > > - ret = mutex_lock_interruptible(&dev_priv->drm.struct_mutex); > > + ret = mutex_lock_interruptible(&i915->drm.struct_mutex); > > if (ret) { > > i915_gem_object_unpin_pages(obj); > > return ret; > > @@ -14287,7 +14287,7 @@ intel_prepare_plane_fb(struct drm_plane *plane, > > ret = intel_plane_pin_fb(to_intel_plane_state(new_state)); > > - mutex_unlock(&dev_priv->drm.struct_mutex); > > + mutex_unlock(&i915->drm.struct_mutex); > > i915_gem_object_unpin_pages(obj); > > if (ret) > > return ret; > > @@ -14323,7 +14323,7 @@ intel_prepare_plane_fb(struct drm_plane *plane, > > * maximum clocks following a vblank miss (see do_rps_boost()). > > */ > > if (!intel_state->rps_interactive) { > > - intel_rps_mark_interactive(dev_priv, true); > > + intel_rps_mark_interactive(i915, true); > > intel_state->rps_interactive = true; > > } > > @@ -14345,17 +14345,17 @@ intel_cleanup_plane_fb(struct drm_plane *plane, > > { > > struct intel_atomic_state *intel_state = > > to_intel_atomic_state(old_state->state); > > - struct drm_i915_private *dev_priv = to_i915(plane->dev); > > + struct drm_i915_private *i915 = to_i915(plane->dev); > > if (intel_state->rps_interactive) { > > - intel_rps_mark_interactive(dev_priv, false); > > + intel_rps_mark_interactive(i915, false); > > intel_state->rps_interactive = false; > > } > > /* Should only be called after a successful intel_prepare_plane_fb()! */ > > - mutex_lock(&dev_priv->drm.struct_mutex); > > + mutex_lock(&i915->drm.struct_mutex); > > intel_plane_unpin_fb(to_intel_plane_state(old_state)); > > - mutex_unlock(&dev_priv->drm.struct_mutex); > > + mutex_unlock(&i915->drm.struct_mutex); > > } > > int > > @@ -14363,7 +14363,7 @@ skl_max_scale(const struct intel_crtc_state *crtc_state, > > u32 pixel_format) > > { > > struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc); > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > int max_scale, mult; > > int crtc_clock, max_dotclk, tmpclk1, tmpclk2; > > @@ -14373,7 +14373,7 @@ skl_max_scale(const struct intel_crtc_state *crtc_state, > > crtc_clock = crtc_state->base.adjusted_mode.crtc_clock; > > max_dotclk = to_intel_atomic_state(crtc_state->base.state)->cdclk.logical.cdclk; > > - if (IS_GEMINILAKE(dev_priv) || INTEL_GEN(dev_priv) >= 10) > > + if (IS_GEMINILAKE(i915) || INTEL_GEN(i915) >= 10) > > max_dotclk *= 2; > > if (WARN_ON_ONCE(!crtc_clock || max_dotclk < crtc_clock)) > > @@ -14396,7 +14396,7 @@ skl_max_scale(const struct intel_crtc_state *crtc_state, > > static void intel_begin_crtc_commit(struct intel_atomic_state *state, > > struct intel_crtc *crtc) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > struct intel_crtc_state *old_crtc_state = > > intel_atomic_get_old_crtc_state(state, crtc); > > struct intel_crtc_state *new_crtc_state = > > @@ -14415,31 +14415,31 @@ static void intel_begin_crtc_commit(struct intel_atomic_state *state, > > if (new_crtc_state->update_pipe) > > intel_update_pipe_config(old_crtc_state, new_crtc_state); > > - else if (INTEL_GEN(dev_priv) >= 9) > > + else if (INTEL_GEN(i915) >= 9) > > skl_detach_scalers(new_crtc_state); > > - if (INTEL_GEN(dev_priv) >= 9 || IS_BROADWELL(dev_priv)) > > + if (INTEL_GEN(i915) >= 9 || IS_BROADWELL(i915)) > > bdw_set_pipemisc(new_crtc_state); > > out: > > - if (dev_priv->display.atomic_update_watermarks) > > - dev_priv->display.atomic_update_watermarks(state, > > + if (i915->display.atomic_update_watermarks) > > + i915->display.atomic_update_watermarks(state, > > new_crtc_state); > > } > > void intel_crtc_arm_fifo_underrun(struct intel_crtc *crtc, > > struct intel_crtc_state *crtc_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > - if (!IS_GEN(dev_priv, 2)) > > - intel_set_cpu_fifo_underrun_reporting(dev_priv, crtc->pipe, true); > > + if (!IS_GEN(i915, 2)) > > + intel_set_cpu_fifo_underrun_reporting(i915, crtc->pipe, true); > > if (crtc_state->has_pch_encoder) { > > enum pipe pch_transcoder = > > intel_crtc_pch_transcoder(crtc); > > - intel_set_pch_fifo_underrun_reporting(dev_priv, pch_transcoder, true); > > + intel_set_pch_fifo_underrun_reporting(i915, pch_transcoder, true); > > } > > } > > @@ -14559,7 +14559,7 @@ intel_legacy_cursor_update(struct drm_plane *plane, > > u32 src_w, u32 src_h, > > struct drm_modeset_acquire_ctx *ctx) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc->dev); > > + struct drm_i915_private *i915 = to_i915(crtc->dev); > > int ret; > > struct drm_plane_state *old_plane_state, *new_plane_state; > > struct intel_plane *intel_plane = to_intel_plane(plane); > > @@ -14626,7 +14626,7 @@ intel_legacy_cursor_update(struct drm_plane *plane, > > if (ret) > > goto out_free; > > - ret = mutex_lock_interruptible(&dev_priv->drm.struct_mutex); > > + ret = mutex_lock_interruptible(&i915->drm.struct_mutex); > > if (ret) > > goto out_free; > > @@ -14664,7 +14664,7 @@ intel_legacy_cursor_update(struct drm_plane *plane, > > intel_plane_unpin_fb(to_intel_plane_state(old_plane_state)); > > out_unlock: > > - mutex_unlock(&dev_priv->drm.struct_mutex); > > + mutex_unlock(&i915->drm.struct_mutex); > > out_free: > > if (new_crtc_state) > > intel_crtc_destroy_state(crtc, &new_crtc_state->base); > > @@ -14691,25 +14691,25 @@ static const struct drm_plane_funcs intel_cursor_plane_funcs = { > > .format_mod_supported = intel_cursor_format_mod_supported, > > }; > > -static bool i9xx_plane_has_fbc(struct drm_i915_private *dev_priv, > > +static bool i9xx_plane_has_fbc(struct drm_i915_private *i915, > > enum i9xx_plane_id i9xx_plane) > > { > > - if (!HAS_FBC(dev_priv)) > > + if (!HAS_FBC(i915)) > > return false; > > - if (IS_BROADWELL(dev_priv) || IS_HASWELL(dev_priv)) > > + if (IS_BROADWELL(i915) || IS_HASWELL(i915)) > > return i9xx_plane == PLANE_A; /* tied to pipe A */ > > - else if (IS_IVYBRIDGE(dev_priv)) > > + else if (IS_IVYBRIDGE(i915)) > > return i9xx_plane == PLANE_A || i9xx_plane == PLANE_B || > > i9xx_plane == PLANE_C; > > - else if (INTEL_GEN(dev_priv) >= 4) > > + else if (INTEL_GEN(i915) >= 4) > > return i9xx_plane == PLANE_A || i9xx_plane == PLANE_B; > > else > > return i9xx_plane == PLANE_A; > > } > > static struct intel_plane * > > -intel_primary_plane_create(struct drm_i915_private *dev_priv, enum pipe pipe) > > +intel_primary_plane_create(struct drm_i915_private *i915, enum pipe pipe) > > { > > struct intel_plane *plane; > > const struct drm_plane_funcs *plane_funcs; > > @@ -14720,8 +14720,8 @@ intel_primary_plane_create(struct drm_i915_private *dev_priv, enum pipe pipe) > > int num_formats; > > int ret; > > - if (INTEL_GEN(dev_priv) >= 9) > > - return skl_universal_plane_create(dev_priv, pipe, > > + if (INTEL_GEN(i915) >= 9) > > + return skl_universal_plane_create(i915, pipe, > > PLANE_PRIMARY); > > plane = intel_plane_alloc(); > > @@ -14733,21 +14733,21 @@ intel_primary_plane_create(struct drm_i915_private *dev_priv, enum pipe pipe) > > * On gen2/3 only plane A can do FBC, but the panel fitter and LVDS > > * port is hooked to pipe B. Hence we want plane A feeding pipe B. > > */ > > - if (HAS_FBC(dev_priv) && INTEL_GEN(dev_priv) < 4) > > + if (HAS_FBC(i915) && INTEL_GEN(i915) < 4) > > plane->i9xx_plane = (enum i9xx_plane_id) !pipe; > > else > > plane->i9xx_plane = (enum i9xx_plane_id) pipe; > > plane->id = PLANE_PRIMARY; > > plane->frontbuffer_bit = INTEL_FRONTBUFFER(pipe, plane->id); > > - plane->has_fbc = i9xx_plane_has_fbc(dev_priv, plane->i9xx_plane); > > + plane->has_fbc = i9xx_plane_has_fbc(i915, plane->i9xx_plane); > > if (plane->has_fbc) { > > - struct intel_fbc *fbc = &dev_priv->fbc; > > + struct intel_fbc *fbc = &i915->fbc; > > fbc->possible_framebuffer_bits |= plane->frontbuffer_bit; > > } > > - if (INTEL_GEN(dev_priv) >= 4) { > > + if (INTEL_GEN(i915) >= 4) { > > formats = i965_primary_formats; > > num_formats = ARRAY_SIZE(i965_primary_formats); > > modifiers = i9xx_format_modifiers; > > @@ -14775,14 +14775,14 @@ intel_primary_plane_create(struct drm_i915_private *dev_priv, enum pipe pipe) > > possible_crtcs = BIT(pipe); > > - if (INTEL_GEN(dev_priv) >= 5 || IS_G4X(dev_priv)) > > - ret = drm_universal_plane_init(&dev_priv->drm, &plane->base, > > + if (INTEL_GEN(i915) >= 5 || IS_G4X(i915)) > > + ret = drm_universal_plane_init(&i915->drm, &plane->base, > > possible_crtcs, plane_funcs, > > formats, num_formats, modifiers, > > DRM_PLANE_TYPE_PRIMARY, > > "primary %c", pipe_name(pipe)); > > else > > - ret = drm_universal_plane_init(&dev_priv->drm, &plane->base, > > + ret = drm_universal_plane_init(&i915->drm, &plane->base, > > possible_crtcs, plane_funcs, > > formats, num_formats, modifiers, > > DRM_PLANE_TYPE_PRIMARY, > > @@ -14791,18 +14791,18 @@ intel_primary_plane_create(struct drm_i915_private *dev_priv, enum pipe pipe) > > if (ret) > > goto fail; > > - if (IS_CHERRYVIEW(dev_priv) && pipe == PIPE_B) { > > + if (IS_CHERRYVIEW(i915) && pipe == PIPE_B) { > > supported_rotations = > > DRM_MODE_ROTATE_0 | DRM_MODE_ROTATE_180 | > > DRM_MODE_REFLECT_X; > > - } else if (INTEL_GEN(dev_priv) >= 4) { > > + } else if (INTEL_GEN(i915) >= 4) { > > supported_rotations = > > DRM_MODE_ROTATE_0 | DRM_MODE_ROTATE_180; > > } else { > > supported_rotations = DRM_MODE_ROTATE_0; > > } > > - if (INTEL_GEN(dev_priv) >= 4) > > + if (INTEL_GEN(i915) >= 4) > > drm_plane_create_rotation_property(&plane->base, > > DRM_MODE_ROTATE_0, > > supported_rotations); > > @@ -14818,7 +14818,7 @@ intel_primary_plane_create(struct drm_i915_private *dev_priv, enum pipe pipe) > > } > > static struct intel_plane * > > -intel_cursor_plane_create(struct drm_i915_private *dev_priv, > > +intel_cursor_plane_create(struct drm_i915_private *i915, > > enum pipe pipe) > > { > > unsigned int possible_crtcs; > > @@ -14834,7 +14834,7 @@ intel_cursor_plane_create(struct drm_i915_private *dev_priv, > > cursor->id = PLANE_CURSOR; > > cursor->frontbuffer_bit = INTEL_FRONTBUFFER(pipe, cursor->id); > > - if (IS_I845G(dev_priv) || IS_I865G(dev_priv)) { > > + if (IS_I845G(i915) || IS_I865G(i915)) { > > cursor->max_stride = i845_cursor_max_stride; > > cursor->update_plane = i845_update_cursor; > > cursor->disable_plane = i845_disable_cursor; > > @@ -14851,12 +14851,12 @@ intel_cursor_plane_create(struct drm_i915_private *dev_priv, > > cursor->cursor.base = ~0; > > cursor->cursor.cntl = ~0; > > - if (IS_I845G(dev_priv) || IS_I865G(dev_priv) || HAS_CUR_FBC(dev_priv)) > > + if (IS_I845G(i915) || IS_I865G(i915) || HAS_CUR_FBC(i915)) > > cursor->cursor.size = ~0; > > possible_crtcs = BIT(pipe); > > - ret = drm_universal_plane_init(&dev_priv->drm, &cursor->base, > > + ret = drm_universal_plane_init(&i915->drm, &cursor->base, > > possible_crtcs, &intel_cursor_plane_funcs, > > intel_cursor_formats, > > ARRAY_SIZE(intel_cursor_formats), > > @@ -14866,7 +14866,7 @@ intel_cursor_plane_create(struct drm_i915_private *dev_priv, > > if (ret) > > goto fail; > > - if (INTEL_GEN(dev_priv) >= 4) > > + if (INTEL_GEN(i915) >= 4) > > drm_plane_create_rotation_property(&cursor->base, > > DRM_MODE_ROTATE_0, > > DRM_MODE_ROTATE_0 | > > @@ -14887,10 +14887,10 @@ static void intel_crtc_init_scalers(struct intel_crtc *crtc, > > { > > struct intel_crtc_scaler_state *scaler_state = > > &crtc_state->scaler_state; > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > int i; > > - crtc->num_scalers = RUNTIME_INFO(dev_priv)->num_scalers[crtc->pipe]; > > + crtc->num_scalers = RUNTIME_INFO(i915)->num_scalers[crtc->pipe]; > > if (!crtc->num_scalers) > > return; > > @@ -14904,7 +14904,7 @@ static void intel_crtc_init_scalers(struct intel_crtc *crtc, > > scaler_state->scaler_id = -1; > > } > > -static int intel_crtc_init(struct drm_i915_private *dev_priv, enum pipe pipe) > > +static int intel_crtc_init(struct drm_i915_private *i915, enum pipe pipe) > > { > > struct intel_crtc *intel_crtc; > > struct intel_crtc_state *crtc_state = NULL; > > @@ -14924,17 +14924,17 @@ static int intel_crtc_init(struct drm_i915_private *dev_priv, enum pipe pipe) > > __drm_atomic_helper_crtc_reset(&intel_crtc->base, &crtc_state->base); > > intel_crtc->config = crtc_state; > > - primary = intel_primary_plane_create(dev_priv, pipe); > > + primary = intel_primary_plane_create(i915, pipe); > > if (IS_ERR(primary)) { > > ret = PTR_ERR(primary); > > goto fail; > > } > > intel_crtc->plane_ids_mask |= BIT(primary->id); > > - for_each_sprite(dev_priv, pipe, sprite) { > > + for_each_sprite(i915, pipe, sprite) { > > struct intel_plane *plane; > > - plane = intel_sprite_plane_create(dev_priv, pipe, sprite); > > + plane = intel_sprite_plane_create(i915, pipe, sprite); > > if (IS_ERR(plane)) { > > ret = PTR_ERR(plane); > > goto fail; > > @@ -14942,14 +14942,14 @@ static int intel_crtc_init(struct drm_i915_private *dev_priv, enum pipe pipe) > > intel_crtc->plane_ids_mask |= BIT(plane->id); > > } > > - cursor = intel_cursor_plane_create(dev_priv, pipe); > > + cursor = intel_cursor_plane_create(i915, pipe); > > if (IS_ERR(cursor)) { > > ret = PTR_ERR(cursor); > > goto fail; > > } > > intel_crtc->plane_ids_mask |= BIT(cursor->id); > > - ret = drm_crtc_init_with_planes(&dev_priv->drm, &intel_crtc->base, > > + ret = drm_crtc_init_with_planes(&i915->drm, &intel_crtc->base, > > &primary->base, &cursor->base, > > &intel_crtc_funcs, > > "pipe %c", pipe_name(pipe)); > > @@ -14961,16 +14961,16 @@ static int intel_crtc_init(struct drm_i915_private *dev_priv, enum pipe pipe) > > /* initialize shared scalers */ > > intel_crtc_init_scalers(intel_crtc, crtc_state); > > - BUG_ON(pipe >= ARRAY_SIZE(dev_priv->pipe_to_crtc_mapping) || > > - dev_priv->pipe_to_crtc_mapping[pipe] != NULL); > > - dev_priv->pipe_to_crtc_mapping[pipe] = intel_crtc; > > + BUG_ON(pipe >= ARRAY_SIZE(i915->pipe_to_crtc_mapping) || > > + i915->pipe_to_crtc_mapping[pipe] != NULL); > > + i915->pipe_to_crtc_mapping[pipe] = intel_crtc; > > - if (INTEL_GEN(dev_priv) < 9) { > > + if (INTEL_GEN(i915) < 9) { > > enum i9xx_plane_id i9xx_plane = primary->i9xx_plane; > > - BUG_ON(i9xx_plane >= ARRAY_SIZE(dev_priv->plane_to_crtc_mapping) || > > - dev_priv->plane_to_crtc_mapping[i9xx_plane] != NULL); > > - dev_priv->plane_to_crtc_mapping[i9xx_plane] = intel_crtc; > > + BUG_ON(i9xx_plane >= ARRAY_SIZE(i915->plane_to_crtc_mapping) || > > + i915->plane_to_crtc_mapping[i9xx_plane] != NULL); > > + i915->plane_to_crtc_mapping[i9xx_plane] = intel_crtc; > > } > > drm_crtc_helper_add(&intel_crtc->base, &intel_helper_funcs); > > @@ -15026,29 +15026,29 @@ static int intel_encoder_clones(struct intel_encoder *encoder) > > return index_mask; > > } > > -static bool ilk_has_edp_a(struct drm_i915_private *dev_priv) > > +static bool ilk_has_edp_a(struct drm_i915_private *i915) > > { > > - if (!IS_MOBILE(dev_priv)) > > + if (!IS_MOBILE(i915)) > > return false; > > if ((I915_READ(DP_A) & DP_DETECTED) == 0) > > return false; > > - if (IS_GEN(dev_priv, 5) && (I915_READ(FUSE_STRAP) & ILK_eDP_A_DISABLE)) > > + if (IS_GEN(i915, 5) && (I915_READ(FUSE_STRAP) & ILK_eDP_A_DISABLE)) > > return false; > > return true; > > } > > -static bool intel_ddi_crt_present(struct drm_i915_private *dev_priv) > > +static bool intel_ddi_crt_present(struct drm_i915_private *i915) > > { > > - if (INTEL_GEN(dev_priv) >= 9) > > + if (INTEL_GEN(i915) >= 9) > > return false; > > - if (IS_HSW_ULT(dev_priv) || IS_BDW_ULT(dev_priv)) > > + if (IS_HSW_ULT(i915) || IS_BDW_ULT(i915)) > > return false; > > - if (HAS_PCH_LPT_H(dev_priv) && > > + if (HAS_PCH_LPT_H(i915) && > > I915_READ(SFUSE_STRAP) & SFUSE_STRAP_CRT_DISABLED) > > return false; > > @@ -15056,24 +15056,24 @@ static bool intel_ddi_crt_present(struct drm_i915_private *dev_priv) > > if (I915_READ(DDI_BUF_CTL(PORT_A)) & DDI_A_4_LANES) > > return false; > > - if (!dev_priv->vbt.int_crt_support) > > + if (!i915->vbt.int_crt_support) > > return false; > > return true; > > } > > -void intel_pps_unlock_regs_wa(struct drm_i915_private *dev_priv) > > +void intel_pps_unlock_regs_wa(struct drm_i915_private *i915) > > { > > int pps_num; > > int pps_idx; > > - if (HAS_DDI(dev_priv)) > > + if (HAS_DDI(i915)) > > return; > > /* > > * This w/a is needed at least on CPT/PPT, but to be sure apply it > > * everywhere where registers can be write protected. > > */ > > - if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) > > + if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) > > pps_num = 2; > > else > > pps_num = 1; > > @@ -15086,65 +15086,65 @@ void intel_pps_unlock_regs_wa(struct drm_i915_private *dev_priv) > > } > > } > > -static void intel_pps_init(struct drm_i915_private *dev_priv) > > +static void intel_pps_init(struct drm_i915_private *i915) > > { > > - if (HAS_PCH_SPLIT(dev_priv) || IS_GEN9_LP(dev_priv)) > > - dev_priv->pps_mmio_base = PCH_PPS_BASE; > > - else if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) > > - dev_priv->pps_mmio_base = VLV_PPS_BASE; > > + if (HAS_PCH_SPLIT(i915) || IS_GEN9_LP(i915)) > > + i915->pps_mmio_base = PCH_PPS_BASE; > > + else if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) > > + i915->pps_mmio_base = VLV_PPS_BASE; > > else > > - dev_priv->pps_mmio_base = PPS_BASE; > > + i915->pps_mmio_base = PPS_BASE; > > - intel_pps_unlock_regs_wa(dev_priv); > > + intel_pps_unlock_regs_wa(i915); > > } > > -static void intel_setup_outputs(struct drm_i915_private *dev_priv) > > +static void intel_setup_outputs(struct drm_i915_private *i915) > > { > > struct intel_encoder *encoder; > > bool dpd_is_edp = false; > > - intel_pps_init(dev_priv); > > + intel_pps_init(i915); > > - if (!HAS_DISPLAY(dev_priv)) > > + if (!HAS_DISPLAY(i915)) > > return; > > - if (IS_ELKHARTLAKE(dev_priv)) { > > - intel_ddi_init(dev_priv, PORT_A); > > - intel_ddi_init(dev_priv, PORT_B); > > - intel_ddi_init(dev_priv, PORT_C); > > - icl_dsi_init(dev_priv); > > - } else if (INTEL_GEN(dev_priv) >= 11) { > > - intel_ddi_init(dev_priv, PORT_A); > > - intel_ddi_init(dev_priv, PORT_B); > > - intel_ddi_init(dev_priv, PORT_C); > > - intel_ddi_init(dev_priv, PORT_D); > > - intel_ddi_init(dev_priv, PORT_E); > > + if (IS_ELKHARTLAKE(i915)) { > > + intel_ddi_init(i915, PORT_A); > > + intel_ddi_init(i915, PORT_B); > > + intel_ddi_init(i915, PORT_C); > > + icl_dsi_init(i915); > > + } else if (INTEL_GEN(i915) >= 11) { > > + intel_ddi_init(i915, PORT_A); > > + intel_ddi_init(i915, PORT_B); > > + intel_ddi_init(i915, PORT_C); > > + intel_ddi_init(i915, PORT_D); > > + intel_ddi_init(i915, PORT_E); > > /* > > * On some ICL SKUs port F is not present. No strap bits for > > * this, so rely on VBT. > > * Work around broken VBTs on SKUs known to have no port F. > > */ > > - if (IS_ICL_WITH_PORT_F(dev_priv) && > > - intel_bios_is_port_present(dev_priv, PORT_F)) > > - intel_ddi_init(dev_priv, PORT_F); > > + if (IS_ICL_WITH_PORT_F(i915) && > > + intel_bios_is_port_present(i915, PORT_F)) > > + intel_ddi_init(i915, PORT_F); > > - icl_dsi_init(dev_priv); > > - } else if (IS_GEN9_LP(dev_priv)) { > > + icl_dsi_init(i915); > > + } else if (IS_GEN9_LP(i915)) { > > /* > > * FIXME: Broxton doesn't support port detection via the > > * DDI_BUF_CTL_A or SFUSE_STRAP registers, find another way to > > * detect the ports. > > */ > > - intel_ddi_init(dev_priv, PORT_A); > > - intel_ddi_init(dev_priv, PORT_B); > > - intel_ddi_init(dev_priv, PORT_C); > > + intel_ddi_init(i915, PORT_A); > > + intel_ddi_init(i915, PORT_B); > > + intel_ddi_init(i915, PORT_C); > > - vlv_dsi_init(dev_priv); > > - } else if (HAS_DDI(dev_priv)) { > > + vlv_dsi_init(i915); > > + } else if (HAS_DDI(i915)) { > > int found; > > - if (intel_ddi_crt_present(dev_priv)) > > - intel_crt_init(dev_priv); > > + if (intel_ddi_crt_present(i915)) > > + intel_crt_init(i915); > > /* > > * Haswell uses DDI functions to detect digital outputs. > > @@ -15153,29 +15153,29 @@ static void intel_setup_outputs(struct drm_i915_private *dev_priv) > > */ > > found = I915_READ(DDI_BUF_CTL(PORT_A)) & DDI_INIT_DISPLAY_DETECTED; > > /* WaIgnoreDDIAStrap: skl */ > > - if (found || IS_GEN9_BC(dev_priv)) > > - intel_ddi_init(dev_priv, PORT_A); > > + if (found || IS_GEN9_BC(i915)) > > + intel_ddi_init(i915, PORT_A); > > /* DDI B, C, D, and F detection is indicated by the SFUSE_STRAP > > * register */ > > found = I915_READ(SFUSE_STRAP); > > if (found & SFUSE_STRAP_DDIB_DETECTED) > > - intel_ddi_init(dev_priv, PORT_B); > > + intel_ddi_init(i915, PORT_B); > > if (found & SFUSE_STRAP_DDIC_DETECTED) > > - intel_ddi_init(dev_priv, PORT_C); > > + intel_ddi_init(i915, PORT_C); > > if (found & SFUSE_STRAP_DDID_DETECTED) > > - intel_ddi_init(dev_priv, PORT_D); > > + intel_ddi_init(i915, PORT_D); > > if (found & SFUSE_STRAP_DDIF_DETECTED) > > - intel_ddi_init(dev_priv, PORT_F); > > + intel_ddi_init(i915, PORT_F); > > /* > > * On SKL we don't have a way to detect DDI-E so we rely on VBT. > > */ > > - if (IS_GEN9_BC(dev_priv) && > > - intel_bios_is_port_present(dev_priv, PORT_E)) > > - intel_ddi_init(dev_priv, PORT_E); > > + if (IS_GEN9_BC(i915) && > > + intel_bios_is_port_present(i915, PORT_E)) > > + intel_ddi_init(i915, PORT_E); > > - } else if (HAS_PCH_SPLIT(dev_priv)) { > > + } else if (HAS_PCH_SPLIT(i915)) { > > int found; > > /* > > @@ -15183,39 +15183,39 @@ static void intel_setup_outputs(struct drm_i915_private *dev_priv) > > * to prevent the registration of both eDP and LVDS and the > > * incorrect sharing of the PPS. > > */ > > - intel_lvds_init(dev_priv); > > - intel_crt_init(dev_priv); > > + intel_lvds_init(i915); > > + intel_crt_init(i915); > > - dpd_is_edp = intel_dp_is_port_edp(dev_priv, PORT_D); > > + dpd_is_edp = intel_dp_is_port_edp(i915, PORT_D); > > - if (ilk_has_edp_a(dev_priv)) > > - intel_dp_init(dev_priv, DP_A, PORT_A); > > + if (ilk_has_edp_a(i915)) > > + intel_dp_init(i915, DP_A, PORT_A); > > if (I915_READ(PCH_HDMIB) & SDVO_DETECTED) { > > /* PCH SDVOB multiplex with HDMIB */ > > - found = intel_sdvo_init(dev_priv, PCH_SDVOB, PORT_B); > > + found = intel_sdvo_init(i915, PCH_SDVOB, PORT_B); > > if (!found) > > - intel_hdmi_init(dev_priv, PCH_HDMIB, PORT_B); > > + intel_hdmi_init(i915, PCH_HDMIB, PORT_B); > > if (!found && (I915_READ(PCH_DP_B) & DP_DETECTED)) > > - intel_dp_init(dev_priv, PCH_DP_B, PORT_B); > > + intel_dp_init(i915, PCH_DP_B, PORT_B); > > } > > if (I915_READ(PCH_HDMIC) & SDVO_DETECTED) > > - intel_hdmi_init(dev_priv, PCH_HDMIC, PORT_C); > > + intel_hdmi_init(i915, PCH_HDMIC, PORT_C); > > if (!dpd_is_edp && I915_READ(PCH_HDMID) & SDVO_DETECTED) > > - intel_hdmi_init(dev_priv, PCH_HDMID, PORT_D); > > + intel_hdmi_init(i915, PCH_HDMID, PORT_D); > > if (I915_READ(PCH_DP_C) & DP_DETECTED) > > - intel_dp_init(dev_priv, PCH_DP_C, PORT_C); > > + intel_dp_init(i915, PCH_DP_C, PORT_C); > > if (I915_READ(PCH_DP_D) & DP_DETECTED) > > - intel_dp_init(dev_priv, PCH_DP_D, PORT_D); > > - } else if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) { > > + intel_dp_init(i915, PCH_DP_D, PORT_D); > > + } else if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) { > > bool has_edp, has_port; > > - if (IS_VALLEYVIEW(dev_priv) && dev_priv->vbt.int_crt_support) > > - intel_crt_init(dev_priv); > > + if (IS_VALLEYVIEW(i915) && i915->vbt.int_crt_support) > > + intel_crt_init(i915); > > /* > > * The DP_DETECTED bit is the latched state of the DDC > > @@ -15232,97 +15232,97 @@ static void intel_setup_outputs(struct drm_i915_private *dev_priv) > > * trust the port type the VBT declares as we've seen at least > > * HDMI ports that the VBT claim are DP or eDP. > > */ > > - has_edp = intel_dp_is_port_edp(dev_priv, PORT_B); > > - has_port = intel_bios_is_port_present(dev_priv, PORT_B); > > + has_edp = intel_dp_is_port_edp(i915, PORT_B); > > + has_port = intel_bios_is_port_present(i915, PORT_B); > > if (I915_READ(VLV_DP_B) & DP_DETECTED || has_port) > > - has_edp &= intel_dp_init(dev_priv, VLV_DP_B, PORT_B); > > + has_edp &= intel_dp_init(i915, VLV_DP_B, PORT_B); > > if ((I915_READ(VLV_HDMIB) & SDVO_DETECTED || has_port) && !has_edp) > > - intel_hdmi_init(dev_priv, VLV_HDMIB, PORT_B); > > + intel_hdmi_init(i915, VLV_HDMIB, PORT_B); > > - has_edp = intel_dp_is_port_edp(dev_priv, PORT_C); > > - has_port = intel_bios_is_port_present(dev_priv, PORT_C); > > + has_edp = intel_dp_is_port_edp(i915, PORT_C); > > + has_port = intel_bios_is_port_present(i915, PORT_C); > > if (I915_READ(VLV_DP_C) & DP_DETECTED || has_port) > > - has_edp &= intel_dp_init(dev_priv, VLV_DP_C, PORT_C); > > + has_edp &= intel_dp_init(i915, VLV_DP_C, PORT_C); > > if ((I915_READ(VLV_HDMIC) & SDVO_DETECTED || has_port) && !has_edp) > > - intel_hdmi_init(dev_priv, VLV_HDMIC, PORT_C); > > + intel_hdmi_init(i915, VLV_HDMIC, PORT_C); > > - if (IS_CHERRYVIEW(dev_priv)) { > > + if (IS_CHERRYVIEW(i915)) { > > /* > > * eDP not supported on port D, > > * so no need to worry about it > > */ > > - has_port = intel_bios_is_port_present(dev_priv, PORT_D); > > + has_port = intel_bios_is_port_present(i915, PORT_D); > > if (I915_READ(CHV_DP_D) & DP_DETECTED || has_port) > > - intel_dp_init(dev_priv, CHV_DP_D, PORT_D); > > + intel_dp_init(i915, CHV_DP_D, PORT_D); > > if (I915_READ(CHV_HDMID) & SDVO_DETECTED || has_port) > > - intel_hdmi_init(dev_priv, CHV_HDMID, PORT_D); > > + intel_hdmi_init(i915, CHV_HDMID, PORT_D); > > } > > - vlv_dsi_init(dev_priv); > > - } else if (IS_PINEVIEW(dev_priv)) { > > - intel_lvds_init(dev_priv); > > - intel_crt_init(dev_priv); > > - } else if (IS_GEN_RANGE(dev_priv, 3, 4)) { > > + vlv_dsi_init(i915); > > + } else if (IS_PINEVIEW(i915)) { > > + intel_lvds_init(i915); > > + intel_crt_init(i915); > > + } else if (IS_GEN_RANGE(i915, 3, 4)) { > > bool found = false; > > - if (IS_MOBILE(dev_priv)) > > - intel_lvds_init(dev_priv); > > + if (IS_MOBILE(i915)) > > + intel_lvds_init(i915); > > - intel_crt_init(dev_priv); > > + intel_crt_init(i915); > > if (I915_READ(GEN3_SDVOB) & SDVO_DETECTED) { > > DRM_DEBUG_KMS("probing SDVOB\n"); > > - found = intel_sdvo_init(dev_priv, GEN3_SDVOB, PORT_B); > > - if (!found && IS_G4X(dev_priv)) { > > + found = intel_sdvo_init(i915, GEN3_SDVOB, PORT_B); > > + if (!found && IS_G4X(i915)) { > > DRM_DEBUG_KMS("probing HDMI on SDVOB\n"); > > - intel_hdmi_init(dev_priv, GEN4_HDMIB, PORT_B); > > + intel_hdmi_init(i915, GEN4_HDMIB, PORT_B); > > } > > - if (!found && IS_G4X(dev_priv)) > > - intel_dp_init(dev_priv, DP_B, PORT_B); > > + if (!found && IS_G4X(i915)) > > + intel_dp_init(i915, DP_B, PORT_B); > > } > > /* Before G4X SDVOC doesn't have its own detect register */ > > if (I915_READ(GEN3_SDVOB) & SDVO_DETECTED) { > > DRM_DEBUG_KMS("probing SDVOC\n"); > > - found = intel_sdvo_init(dev_priv, GEN3_SDVOC, PORT_C); > > + found = intel_sdvo_init(i915, GEN3_SDVOC, PORT_C); > > } > > if (!found && (I915_READ(GEN3_SDVOC) & SDVO_DETECTED)) { > > - if (IS_G4X(dev_priv)) { > > + if (IS_G4X(i915)) { > > DRM_DEBUG_KMS("probing HDMI on SDVOC\n"); > > - intel_hdmi_init(dev_priv, GEN4_HDMIC, PORT_C); > > + intel_hdmi_init(i915, GEN4_HDMIC, PORT_C); > > } > > - if (IS_G4X(dev_priv)) > > - intel_dp_init(dev_priv, DP_C, PORT_C); > > + if (IS_G4X(i915)) > > + intel_dp_init(i915, DP_C, PORT_C); > > } > > - if (IS_G4X(dev_priv) && (I915_READ(DP_D) & DP_DETECTED)) > > - intel_dp_init(dev_priv, DP_D, PORT_D); > > + if (IS_G4X(i915) && (I915_READ(DP_D) & DP_DETECTED)) > > + intel_dp_init(i915, DP_D, PORT_D); > > - if (SUPPORTS_TV(dev_priv)) > > - intel_tv_init(dev_priv); > > - } else if (IS_GEN(dev_priv, 2)) { > > - if (IS_I85X(dev_priv)) > > - intel_lvds_init(dev_priv); > > + if (SUPPORTS_TV(i915)) > > + intel_tv_init(i915); > > + } else if (IS_GEN(i915, 2)) { > > + if (IS_I85X(i915)) > > + intel_lvds_init(i915); > > - intel_crt_init(dev_priv); > > - intel_dvo_init(dev_priv); > > + intel_crt_init(i915); > > + intel_dvo_init(i915); > > } > > - intel_psr_init(dev_priv); > > + intel_psr_init(i915); > > - for_each_intel_encoder(&dev_priv->drm, encoder) { > > + for_each_intel_encoder(&i915->drm, encoder) { > > encoder->base.possible_crtcs = encoder->crtc_mask; > > encoder->base.possible_clones = > > intel_encoder_clones(encoder); > > } > > - intel_init_pch_refclk(dev_priv); > > + intel_init_pch_refclk(i915); > > - drm_helper_move_panel_connectors_to_head(&dev_priv->drm); > > + drm_helper_move_panel_connectors_to_head(&i915->drm); > > } > > static void intel_user_framebuffer_destroy(struct drm_framebuffer *fb) > > @@ -15379,7 +15379,7 @@ static int intel_framebuffer_init(struct intel_framebuffer *intel_fb, > > struct drm_i915_gem_object *obj, > > struct drm_mode_fb_cmd2 *mode_cmd) > > { > > - struct drm_i915_private *dev_priv = to_i915(obj->base.dev); > > + struct drm_i915_private *i915 = to_i915(obj->base.dev); > > struct drm_framebuffer *fb = &intel_fb->base; > > u32 max_stride; > > unsigned int tiling, stride; > > @@ -15411,7 +15411,7 @@ static int intel_framebuffer_init(struct intel_framebuffer *intel_fb, > > } > > } > > - if (!drm_any_plane_has_format(&dev_priv->drm, > > + if (!drm_any_plane_has_format(&i915->drm, > > mode_cmd->pixel_format, > > mode_cmd->modifier[0])) { > > struct drm_format_name_buf format_name; > > @@ -15427,13 +15427,13 @@ static int intel_framebuffer_init(struct intel_framebuffer *intel_fb, > > * gen2/3 display engine uses the fence if present, > > * so the tiling mode must match the fb modifier exactly. > > */ > > - if (INTEL_GEN(dev_priv) < 4 && > > + if (INTEL_GEN(i915) < 4 && > > tiling != intel_fb_modifier_to_tiling(mode_cmd->modifier[0])) { > > DRM_DEBUG_KMS("tiling_mode must match fb modifier exactly on gen2/3\n"); > > goto err; > > } > > - max_stride = intel_fb_max_stride(dev_priv, mode_cmd->pixel_format, > > + max_stride = intel_fb_max_stride(i915, mode_cmd->pixel_format, > > mode_cmd->modifier[0]); > > if (mode_cmd->pitches[0] > max_stride) { > > DRM_DEBUG_KMS("%s pitch (%u) must be at most %d\n", > > @@ -15457,7 +15457,7 @@ static int intel_framebuffer_init(struct intel_framebuffer *intel_fb, > > if (mode_cmd->offsets[0] != 0) > > goto err; > > - drm_helper_mode_fill_fb_struct(&dev_priv->drm, fb, mode_cmd); > > + drm_helper_mode_fill_fb_struct(&i915->drm, fb, mode_cmd); > > for (i = 0; i < fb->format->num_planes; i++) { > > u32 stride_alignment; > > @@ -15478,7 +15478,7 @@ static int intel_framebuffer_init(struct intel_framebuffer *intel_fb, > > * require the entire fb to accommodate that to avoid > > * potential runtime errors at plane configuration time. > > */ > > - if (IS_GEN(dev_priv, 9) && i == 0 && fb->width > 3840 && > > + if (IS_GEN(i915, 9) && i == 0 && fb->width > 3840 && > > is_ccs_modifier(fb->modifier)) > > stride_alignment *= 4; > > @@ -15491,11 +15491,11 @@ static int intel_framebuffer_init(struct intel_framebuffer *intel_fb, > > fb->obj[i] = &obj->base; > > } > > - ret = intel_fill_fb_info(dev_priv, fb); > > + ret = intel_fill_fb_info(i915, fb); > > if (ret) > > goto err; > > - ret = drm_framebuffer_init(&dev_priv->drm, fb, &intel_fb_funcs); > > + ret = drm_framebuffer_init(&i915->drm, fb, &intel_fb_funcs); > > if (ret) { > > DRM_ERROR("framebuffer init failed %d\n", ret); > > goto err; > > @@ -15545,7 +15545,7 @@ static enum drm_mode_status > > intel_mode_valid(struct drm_device *dev, > > const struct drm_display_mode *mode) > > { > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > int hdisplay_max, htotal_max; > > int vdisplay_max, vtotal_max; > > @@ -15578,13 +15578,13 @@ intel_mode_valid(struct drm_device *dev, > > DRM_MODE_FLAG_CLKDIV2)) > > return MODE_BAD; > > - if (INTEL_GEN(dev_priv) >= 9 || > > - IS_BROADWELL(dev_priv) || IS_HASWELL(dev_priv)) { > > + if (INTEL_GEN(i915) >= 9 || > > + IS_BROADWELL(i915) || IS_HASWELL(i915)) { > > hdisplay_max = 8192; /* FDI max 4096 handled elsewhere */ > > vdisplay_max = 4096; > > htotal_max = 8192; > > vtotal_max = 8192; > > - } else if (INTEL_GEN(dev_priv) >= 3) { > > + } else if (INTEL_GEN(i915) >= 3) { > > hdisplay_max = 4096; > > vdisplay_max = 4096; > > htotal_max = 8192; > > @@ -15625,113 +15625,113 @@ static const struct drm_mode_config_funcs intel_mode_funcs = { > > /** > > * intel_init_display_hooks - initialize the display modesetting hooks > > - * @dev_priv: device private > > + * @i915: device private > > */ > > -void intel_init_display_hooks(struct drm_i915_private *dev_priv) > > +void intel_init_display_hooks(struct drm_i915_private *i915) > > { > > - intel_init_cdclk_hooks(dev_priv); > > + intel_init_cdclk_hooks(i915); > > - if (INTEL_GEN(dev_priv) >= 9) { > > - dev_priv->display.get_pipe_config = haswell_get_pipe_config; > > - dev_priv->display.get_initial_plane_config = > > + if (INTEL_GEN(i915) >= 9) { > > + i915->display.get_pipe_config = haswell_get_pipe_config; > > + i915->display.get_initial_plane_config = > > skylake_get_initial_plane_config; > > - dev_priv->display.crtc_compute_clock = > > + i915->display.crtc_compute_clock = > > haswell_crtc_compute_clock; > > - dev_priv->display.crtc_enable = haswell_crtc_enable; > > - dev_priv->display.crtc_disable = haswell_crtc_disable; > > - } else if (HAS_DDI(dev_priv)) { > > - dev_priv->display.get_pipe_config = haswell_get_pipe_config; > > - dev_priv->display.get_initial_plane_config = > > + i915->display.crtc_enable = haswell_crtc_enable; > > + i915->display.crtc_disable = haswell_crtc_disable; > > + } else if (HAS_DDI(i915)) { > > + i915->display.get_pipe_config = haswell_get_pipe_config; > > + i915->display.get_initial_plane_config = > > i9xx_get_initial_plane_config; > > - dev_priv->display.crtc_compute_clock = > > + i915->display.crtc_compute_clock = > > haswell_crtc_compute_clock; > > - dev_priv->display.crtc_enable = haswell_crtc_enable; > > - dev_priv->display.crtc_disable = haswell_crtc_disable; > > - } else if (HAS_PCH_SPLIT(dev_priv)) { > > - dev_priv->display.get_pipe_config = ironlake_get_pipe_config; > > - dev_priv->display.get_initial_plane_config = > > + i915->display.crtc_enable = haswell_crtc_enable; > > + i915->display.crtc_disable = haswell_crtc_disable; > > + } else if (HAS_PCH_SPLIT(i915)) { > > + i915->display.get_pipe_config = ironlake_get_pipe_config; > > + i915->display.get_initial_plane_config = > > i9xx_get_initial_plane_config; > > - dev_priv->display.crtc_compute_clock = > > + i915->display.crtc_compute_clock = > > ironlake_crtc_compute_clock; > > - dev_priv->display.crtc_enable = ironlake_crtc_enable; > > - dev_priv->display.crtc_disable = ironlake_crtc_disable; > > - } else if (IS_CHERRYVIEW(dev_priv)) { > > - dev_priv->display.get_pipe_config = i9xx_get_pipe_config; > > - dev_priv->display.get_initial_plane_config = > > + i915->display.crtc_enable = ironlake_crtc_enable; > > + i915->display.crtc_disable = ironlake_crtc_disable; > > + } else if (IS_CHERRYVIEW(i915)) { > > + i915->display.get_pipe_config = i9xx_get_pipe_config; > > + i915->display.get_initial_plane_config = > > i9xx_get_initial_plane_config; > > - dev_priv->display.crtc_compute_clock = chv_crtc_compute_clock; > > - dev_priv->display.crtc_enable = valleyview_crtc_enable; > > - dev_priv->display.crtc_disable = i9xx_crtc_disable; > > - } else if (IS_VALLEYVIEW(dev_priv)) { > > - dev_priv->display.get_pipe_config = i9xx_get_pipe_config; > > - dev_priv->display.get_initial_plane_config = > > + i915->display.crtc_compute_clock = chv_crtc_compute_clock; > > + i915->display.crtc_enable = valleyview_crtc_enable; > > + i915->display.crtc_disable = i9xx_crtc_disable; > > + } else if (IS_VALLEYVIEW(i915)) { > > + i915->display.get_pipe_config = i9xx_get_pipe_config; > > + i915->display.get_initial_plane_config = > > i9xx_get_initial_plane_config; > > - dev_priv->display.crtc_compute_clock = vlv_crtc_compute_clock; > > - dev_priv->display.crtc_enable = valleyview_crtc_enable; > > - dev_priv->display.crtc_disable = i9xx_crtc_disable; > > - } else if (IS_G4X(dev_priv)) { > > - dev_priv->display.get_pipe_config = i9xx_get_pipe_config; > > - dev_priv->display.get_initial_plane_config = > > + i915->display.crtc_compute_clock = vlv_crtc_compute_clock; > > + i915->display.crtc_enable = valleyview_crtc_enable; > > + i915->display.crtc_disable = i9xx_crtc_disable; > > + } else if (IS_G4X(i915)) { > > + i915->display.get_pipe_config = i9xx_get_pipe_config; > > + i915->display.get_initial_plane_config = > > i9xx_get_initial_plane_config; > > - dev_priv->display.crtc_compute_clock = g4x_crtc_compute_clock; > > - dev_priv->display.crtc_enable = i9xx_crtc_enable; > > - dev_priv->display.crtc_disable = i9xx_crtc_disable; > > - } else if (IS_PINEVIEW(dev_priv)) { > > - dev_priv->display.get_pipe_config = i9xx_get_pipe_config; > > - dev_priv->display.get_initial_plane_config = > > + i915->display.crtc_compute_clock = g4x_crtc_compute_clock; > > + i915->display.crtc_enable = i9xx_crtc_enable; > > + i915->display.crtc_disable = i9xx_crtc_disable; > > + } else if (IS_PINEVIEW(i915)) { > > + i915->display.get_pipe_config = i9xx_get_pipe_config; > > + i915->display.get_initial_plane_config = > > i9xx_get_initial_plane_config; > > - dev_priv->display.crtc_compute_clock = pnv_crtc_compute_clock; > > - dev_priv->display.crtc_enable = i9xx_crtc_enable; > > - dev_priv->display.crtc_disable = i9xx_crtc_disable; > > - } else if (!IS_GEN(dev_priv, 2)) { > > - dev_priv->display.get_pipe_config = i9xx_get_pipe_config; > > - dev_priv->display.get_initial_plane_config = > > + i915->display.crtc_compute_clock = pnv_crtc_compute_clock; > > + i915->display.crtc_enable = i9xx_crtc_enable; > > + i915->display.crtc_disable = i9xx_crtc_disable; > > + } else if (!IS_GEN(i915, 2)) { > > + i915->display.get_pipe_config = i9xx_get_pipe_config; > > + i915->display.get_initial_plane_config = > > i9xx_get_initial_plane_config; > > - dev_priv->display.crtc_compute_clock = i9xx_crtc_compute_clock; > > - dev_priv->display.crtc_enable = i9xx_crtc_enable; > > - dev_priv->display.crtc_disable = i9xx_crtc_disable; > > + i915->display.crtc_compute_clock = i9xx_crtc_compute_clock; > > + i915->display.crtc_enable = i9xx_crtc_enable; > > + i915->display.crtc_disable = i9xx_crtc_disable; > > } else { > > - dev_priv->display.get_pipe_config = i9xx_get_pipe_config; > > - dev_priv->display.get_initial_plane_config = > > + i915->display.get_pipe_config = i9xx_get_pipe_config; > > + i915->display.get_initial_plane_config = > > i9xx_get_initial_plane_config; > > - dev_priv->display.crtc_compute_clock = i8xx_crtc_compute_clock; > > - dev_priv->display.crtc_enable = i9xx_crtc_enable; > > - dev_priv->display.crtc_disable = i9xx_crtc_disable; > > + i915->display.crtc_compute_clock = i8xx_crtc_compute_clock; > > + i915->display.crtc_enable = i9xx_crtc_enable; > > + i915->display.crtc_disable = i9xx_crtc_disable; > > } > > - if (IS_GEN(dev_priv, 5)) { > > - dev_priv->display.fdi_link_train = ironlake_fdi_link_train; > > - } else if (IS_GEN(dev_priv, 6)) { > > - dev_priv->display.fdi_link_train = gen6_fdi_link_train; > > - } else if (IS_IVYBRIDGE(dev_priv)) { > > + if (IS_GEN(i915, 5)) { > > + i915->display.fdi_link_train = ironlake_fdi_link_train; > > + } else if (IS_GEN(i915, 6)) { > > + i915->display.fdi_link_train = gen6_fdi_link_train; > > + } else if (IS_IVYBRIDGE(i915)) { > > /* FIXME: detect B0+ stepping and use auto training */ > > - dev_priv->display.fdi_link_train = ivb_manual_fdi_link_train; > > - } else if (IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv)) { > > - dev_priv->display.fdi_link_train = hsw_fdi_link_train; > > + i915->display.fdi_link_train = ivb_manual_fdi_link_train; > > + } else if (IS_HASWELL(i915) || IS_BROADWELL(i915)) { > > + i915->display.fdi_link_train = hsw_fdi_link_train; > > } > > - if (INTEL_GEN(dev_priv) >= 9) > > - dev_priv->display.update_crtcs = skl_update_crtcs; > > + if (INTEL_GEN(i915) >= 9) > > + i915->display.update_crtcs = skl_update_crtcs; > > else > > - dev_priv->display.update_crtcs = intel_update_crtcs; > > + i915->display.update_crtcs = intel_update_crtcs; > > } > > -static i915_reg_t i915_vgacntrl_reg(struct drm_i915_private *dev_priv) > > +static i915_reg_t i915_vgacntrl_reg(struct drm_i915_private *i915) > > { > > - if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) > > + if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) > > return VLV_VGACNTRL; > > - else if (INTEL_GEN(dev_priv) >= 5) > > + else if (INTEL_GEN(i915) >= 5) > > return CPU_VGACNTRL; > > else > > return VGACNTRL; > > } > > /* Disable the VGA plane that we never use */ > > -static void i915_disable_vga(struct drm_i915_private *dev_priv) > > +static void i915_disable_vga(struct drm_i915_private *i915) > > { > > - struct pci_dev *pdev = dev_priv->drm.pdev; > > + struct pci_dev *pdev = i915->drm.pdev; > > u8 sr1; > > - i915_reg_t vga_reg = i915_vgacntrl_reg(dev_priv); > > + i915_reg_t vga_reg = i915_vgacntrl_reg(i915); > > /* WaEnableVGAAccessThroughIOPort:ctg,elk,ilk,snb,ivb,vlv,hsw */ > > vga_get_uninterruptible(pdev, VGA_RSRC_LEGACY_IO); > > @@ -15747,11 +15747,11 @@ static void i915_disable_vga(struct drm_i915_private *dev_priv) > > void intel_modeset_init_hw(struct drm_device *dev) > > { > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > - intel_update_cdclk(dev_priv); > > - intel_dump_cdclk_state(&dev_priv->cdclk.hw, "Current CDCLK"); > > - dev_priv->cdclk.logical = dev_priv->cdclk.actual = dev_priv->cdclk.hw; > > + intel_update_cdclk(i915); > > + intel_dump_cdclk_state(&i915->cdclk.hw, "Current CDCLK"); > > + i915->cdclk.logical = i915->cdclk.actual = i915->cdclk.hw; > > } > > /* > > @@ -15766,7 +15766,7 @@ void intel_modeset_init_hw(struct drm_device *dev) > > */ > > static void sanitize_watermarks(struct drm_device *dev) > > { > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > struct drm_atomic_state *state; > > struct intel_atomic_state *intel_state; > > struct drm_crtc *crtc; > > @@ -15776,7 +15776,7 @@ static void sanitize_watermarks(struct drm_device *dev) > > int i; > > /* Only supported on platforms that use atomic watermark design */ > > - if (!dev_priv->display.optimize_watermarks) > > + if (!i915->display.optimize_watermarks) > > return; > > /* > > @@ -15804,7 +15804,7 @@ static void sanitize_watermarks(struct drm_device *dev) > > * intermediate watermarks (since we don't trust the current > > * watermarks). > > */ > > - if (!HAS_GMCH(dev_priv)) > > + if (!HAS_GMCH(i915)) > > intel_state->skip_intermediate_wm = true; > > ret = intel_atomic_check(dev, state); > > @@ -15829,7 +15829,7 @@ static void sanitize_watermarks(struct drm_device *dev) > > struct intel_crtc_state *cs = to_intel_crtc_state(cstate); > > cs->wm.need_postvbl_update = true; > > - dev_priv->display.optimize_watermarks(intel_state, cs); > > + i915->display.optimize_watermarks(intel_state, cs); > > to_intel_crtc_state(crtc->state)->wm = cs->wm; > > } > > @@ -15841,20 +15841,20 @@ static void sanitize_watermarks(struct drm_device *dev) > > drm_modeset_acquire_fini(&ctx); > > } > > -static void intel_update_fdi_pll_freq(struct drm_i915_private *dev_priv) > > +static void intel_update_fdi_pll_freq(struct drm_i915_private *i915) > > { > > - if (IS_GEN(dev_priv, 5)) { > > + if (IS_GEN(i915, 5)) { > > u32 fdi_pll_clk = > > I915_READ(FDI_PLL_BIOS_0) & FDI_PLL_FB_CLOCK_MASK; > > - dev_priv->fdi_pll_freq = (fdi_pll_clk + 2) * 10000; > > - } else if (IS_GEN(dev_priv, 6) || IS_IVYBRIDGE(dev_priv)) { > > - dev_priv->fdi_pll_freq = 270000; > > + i915->fdi_pll_freq = (fdi_pll_clk + 2) * 10000; > > + } else if (IS_GEN(i915, 6) || IS_IVYBRIDGE(i915)) { > > + i915->fdi_pll_freq = 270000; > > } else { > > return; > > } > > - DRM_DEBUG_DRIVER("FDI PLL freq=%d\n", dev_priv->fdi_pll_freq); > > + DRM_DEBUG_DRIVER("FDI PLL freq=%d\n", i915->fdi_pll_freq); > > } > > static int intel_initial_commit(struct drm_device *dev) > > @@ -15915,17 +15915,17 @@ static int intel_initial_commit(struct drm_device *dev) > > int intel_modeset_init(struct drm_device *dev) > > { > > - struct drm_i915_private *dev_priv = to_i915(dev); > > - struct i915_ggtt *ggtt = &dev_priv->ggtt; > > + struct drm_i915_private *i915 = to_i915(dev); > > + struct i915_ggtt *ggtt = &i915->ggtt; > > enum pipe pipe; > > struct intel_crtc *crtc; > > int ret; > > - dev_priv->modeset_wq = alloc_ordered_workqueue("i915_modeset", 0); > > + i915->modeset_wq = alloc_ordered_workqueue("i915_modeset", 0); > > drm_mode_config_init(dev); > > - ret = intel_bw_init(dev_priv); > > + ret = intel_bw_init(i915); > > if (ret) > > return ret; > > @@ -15939,15 +15939,15 @@ int intel_modeset_init(struct drm_device *dev) > > dev->mode_config.funcs = &intel_mode_funcs; > > - init_llist_head(&dev_priv->atomic_helper.free_list); > > - INIT_WORK(&dev_priv->atomic_helper.free_work, > > + init_llist_head(&i915->atomic_helper.free_list); > > + INIT_WORK(&i915->atomic_helper.free_work, > > intel_atomic_helper_free_state_worker); > > - intel_init_quirks(dev_priv); > > + intel_init_quirks(i915); > > - intel_fbc_init(dev_priv); > > + intel_fbc_init(i915); > > - intel_init_pm(dev_priv); > > + intel_init_pm(i915); > > /* > > * There may be no VBT; and if the BIOS enabled SSC we can > > @@ -15955,15 +15955,15 @@ int intel_modeset_init(struct drm_device *dev) > > * BIOS isn't using it, don't assume it will work even if the VBT > > * indicates as much. > > */ > > - if (HAS_PCH_IBX(dev_priv) || HAS_PCH_CPT(dev_priv)) { > > + if (HAS_PCH_IBX(i915) || HAS_PCH_CPT(i915)) { > > bool bios_lvds_use_ssc = !!(I915_READ(PCH_DREF_CONTROL) & > > DREF_SSC1_ENABLE); > > - if (dev_priv->vbt.lvds_use_ssc != bios_lvds_use_ssc) { > > + if (i915->vbt.lvds_use_ssc != bios_lvds_use_ssc) { > > DRM_DEBUG_KMS("SSC %sabled by BIOS, overriding VBT which says %sabled\n", > > bios_lvds_use_ssc ? "en" : "dis", > > - dev_priv->vbt.lvds_use_ssc ? "en" : "dis"); > > - dev_priv->vbt.lvds_use_ssc = bios_lvds_use_ssc; > > + i915->vbt.lvds_use_ssc ? "en" : "dis"); > > + i915->vbt.lvds_use_ssc = bios_lvds_use_ssc; > > } > > } > > @@ -15971,13 +15971,13 @@ int intel_modeset_init(struct drm_device *dev) > > * Maximum framebuffer dimensions, chosen to match > > * the maximum render engine surface size on gen4+. > > */ > > - if (INTEL_GEN(dev_priv) >= 7) { > > + if (INTEL_GEN(i915) >= 7) { > > dev->mode_config.max_width = 16384; > > dev->mode_config.max_height = 16384; > > - } else if (INTEL_GEN(dev_priv) >= 4) { > > + } else if (INTEL_GEN(i915) >= 4) { > > dev->mode_config.max_width = 8192; > > dev->mode_config.max_height = 8192; > > - } else if (IS_GEN(dev_priv, 3)) { > > + } else if (IS_GEN(i915, 3)) { > > dev->mode_config.max_width = 4096; > > dev->mode_config.max_height = 4096; > > } else { > > @@ -15985,10 +15985,10 @@ int intel_modeset_init(struct drm_device *dev) > > dev->mode_config.max_height = 2048; > > } > > - if (IS_I845G(dev_priv) || IS_I865G(dev_priv)) { > > - dev->mode_config.cursor_width = IS_I845G(dev_priv) ? 64 : 512; > > + if (IS_I845G(i915) || IS_I865G(i915)) { > > + dev->mode_config.cursor_width = IS_I845G(i915) ? 64 : 512; > > dev->mode_config.cursor_height = 1023; > > - } else if (IS_GEN(dev_priv, 2)) { > > + } else if (IS_GEN(i915, 2)) { > > dev->mode_config.cursor_width = 64; > > dev->mode_config.cursor_height = 64; > > } else { > > @@ -15999,11 +15999,11 @@ int intel_modeset_init(struct drm_device *dev) > > dev->mode_config.fb_base = ggtt->gmadr.start; > > DRM_DEBUG_KMS("%d display pipe%s available.\n", > > - INTEL_INFO(dev_priv)->num_pipes, > > - INTEL_INFO(dev_priv)->num_pipes > 1 ? "s" : ""); > > + INTEL_INFO(i915)->num_pipes, > > + INTEL_INFO(i915)->num_pipes > 1 ? "s" : ""); > > - for_each_pipe(dev_priv, pipe) { > > - ret = intel_crtc_init(dev_priv, pipe); > > + for_each_pipe(i915, pipe) { > > + ret = intel_crtc_init(i915, pipe); > > if (ret) { > > drm_mode_config_cleanup(dev); > > return ret; > > @@ -16011,19 +16011,19 @@ int intel_modeset_init(struct drm_device *dev) > > } > > intel_shared_dpll_init(dev); > > - intel_update_fdi_pll_freq(dev_priv); > > + intel_update_fdi_pll_freq(i915); > > - intel_update_czclk(dev_priv); > > + intel_update_czclk(i915); > > intel_modeset_init_hw(dev); > > - intel_hdcp_component_init(dev_priv); > > + intel_hdcp_component_init(i915); > > - if (dev_priv->max_cdclk_freq == 0) > > - intel_update_max_cdclk(dev_priv); > > + if (i915->max_cdclk_freq == 0) > > + intel_update_max_cdclk(i915); > > /* Just disable it once at startup */ > > - i915_disable_vga(dev_priv); > > - intel_setup_outputs(dev_priv); > > + i915_disable_vga(i915); > > + intel_setup_outputs(i915); > > drm_modeset_lock_all(dev); > > intel_modeset_setup_hw_state(dev, dev->mode_config.acquire_ctx); > > @@ -16042,7 +16042,7 @@ int intel_modeset_init(struct drm_device *dev) > > * can even allow for smooth boot transitions if the BIOS > > * fb is large enough for the active pipe configuration. > > */ > > - dev_priv->display.get_initial_plane_config(crtc, > > + i915->display.get_initial_plane_config(crtc, > > &plane_config); > > /* > > @@ -16057,7 +16057,7 @@ int intel_modeset_init(struct drm_device *dev) > > * Note that we need to do this after reconstructing the BIOS fb's > > * since the watermark calculation done here will use pstate->fb. > > */ > > - if (!HAS_GMCH(dev_priv)) > > + if (!HAS_GMCH(i915)) > > sanitize_watermarks(dev); > > /* > > @@ -16073,9 +16073,9 @@ int intel_modeset_init(struct drm_device *dev) > > return 0; > > } > > -void i830_enable_pipe(struct drm_i915_private *dev_priv, enum pipe pipe) > > +void i830_enable_pipe(struct drm_i915_private *i915, enum pipe pipe) > > { > > - struct intel_crtc *crtc = intel_get_crtc_for_pipe(dev_priv, pipe); > > + struct intel_crtc *crtc = intel_get_crtc_for_pipe(i915, pipe); > > /* 640x480@60Hz, ~25175 kHz */ > > struct dpll clock = { > > .m1 = 18, > > @@ -16143,9 +16143,9 @@ void i830_enable_pipe(struct drm_i915_private *dev_priv, enum pipe pipe) > > intel_wait_for_pipe_scanline_moving(crtc); > > } > > -void i830_disable_pipe(struct drm_i915_private *dev_priv, enum pipe pipe) > > +void i830_disable_pipe(struct drm_i915_private *i915, enum pipe pipe) > > { > > - struct intel_crtc *crtc = intel_get_crtc_for_pipe(dev_priv, pipe); > > + struct intel_crtc *crtc = intel_get_crtc_for_pipe(i915, pipe); > > DRM_DEBUG_KMS("disabling pipe %c due to force quirk\n", > > pipe_name(pipe)); > > @@ -16166,14 +16166,14 @@ void i830_disable_pipe(struct drm_i915_private *dev_priv, enum pipe pipe) > > } > > static void > > -intel_sanitize_plane_mapping(struct drm_i915_private *dev_priv) > > +intel_sanitize_plane_mapping(struct drm_i915_private *i915) > > { > > struct intel_crtc *crtc; > > - if (INTEL_GEN(dev_priv) >= 4) > > + if (INTEL_GEN(i915) >= 4) > > return; > > - for_each_intel_crtc(&dev_priv->drm, crtc) { > > + for_each_intel_crtc(&i915->drm, crtc) { > > struct intel_plane *plane = > > to_intel_plane(crtc->base.primary); > > struct intel_crtc *plane_crtc; > > @@ -16188,7 +16188,7 @@ intel_sanitize_plane_mapping(struct drm_i915_private *dev_priv) > > DRM_DEBUG_KMS("[PLANE:%d:%s] attached to the wrong pipe, disabling plane\n", > > plane->base.base.id, plane->base.name); > > - plane_crtc = intel_get_crtc_for_pipe(dev_priv, pipe); > > + plane_crtc = intel_get_crtc_for_pipe(i915, pipe); > > intel_plane_disable_noatomic(plane_crtc, plane); > > } > > } > > @@ -16215,18 +16215,18 @@ static struct intel_connector *intel_encoder_find_connector(struct intel_encoder > > return NULL; > > } > > -static bool has_pch_trancoder(struct drm_i915_private *dev_priv, > > +static bool has_pch_trancoder(struct drm_i915_private *i915, > > enum pipe pch_transcoder) > > { > > - return HAS_PCH_IBX(dev_priv) || HAS_PCH_CPT(dev_priv) || > > - (HAS_PCH_LPT_H(dev_priv) && pch_transcoder == PIPE_A); > > + return HAS_PCH_IBX(i915) || HAS_PCH_CPT(i915) || > > + (HAS_PCH_LPT_H(i915) && pch_transcoder == PIPE_A); > > } > > static void intel_sanitize_crtc(struct intel_crtc *crtc, > > struct drm_modeset_acquire_ctx *ctx) > > { > > struct drm_device *dev = crtc->base.dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > struct intel_crtc_state *crtc_state = to_intel_crtc_state(crtc->base.state); > > enum transcoder cpu_transcoder = crtc_state->cpu_transcoder; > > @@ -16255,7 +16255,7 @@ static void intel_sanitize_crtc(struct intel_crtc *crtc, > > * Disable any background color set by the BIOS, but enable the > > * gamma and CSC to match how we program our planes. > > */ > > - if (INTEL_GEN(dev_priv) >= 9) > > + if (INTEL_GEN(i915) >= 9) > > I915_WRITE(SKL_BOTTOM_COLOR(crtc->pipe), > > SKL_BOTTOM_COLOR_GAMMA_ENABLE | > > SKL_BOTTOM_COLOR_CSC_ENABLE); > > @@ -16266,7 +16266,7 @@ static void intel_sanitize_crtc(struct intel_crtc *crtc, > > if (crtc_state->base.active && !intel_crtc_has_encoders(crtc)) > > intel_crtc_disable_noatomic(&crtc->base, ctx); > > - if (crtc_state->base.active || HAS_GMCH(dev_priv)) { > > + if (crtc_state->base.active || HAS_GMCH(i915)) { > > /* > > * We start out with underrun reporting disabled to avoid races. > > * For correct bookkeeping mark this on active crtcs. > > @@ -16290,14 +16290,14 @@ static void intel_sanitize_crtc(struct intel_crtc *crtc, > > * PCH transcoders B and C would prevent enabling the south > > * error interrupt (see cpt_can_enable_serr_int()). > > */ > > - if (has_pch_trancoder(dev_priv, crtc->pipe)) > > + if (has_pch_trancoder(i915, crtc->pipe)) > > crtc->pch_fifo_underrun_disabled = true; > > } > > } > > static bool has_bogus_dpll_config(const struct intel_crtc_state *crtc_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc_state->base.crtc->dev); > > + struct drm_i915_private *i915 = to_i915(crtc_state->base.crtc->dev); > > /* > > * Some SNB BIOSen (eg. ASUS K53SV) are known to misprogram > > @@ -16309,7 +16309,7 @@ static bool has_bogus_dpll_config(const struct intel_crtc_state *crtc_state) > > * without several WARNs, but for now let's take the easy > > * road. > > */ > > - return IS_GEN(dev_priv, 6) && > > + return IS_GEN(i915, 6) && > > crtc_state->base.active && > > crtc_state->shared_dpll && > > crtc_state->port_clock == 0; > > @@ -16317,7 +16317,7 @@ static bool has_bogus_dpll_config(const struct intel_crtc_state *crtc_state) > > static void intel_sanitize_encoder(struct intel_encoder *encoder) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_connector *connector; > > struct intel_crtc *crtc = to_intel_crtc(encoder->base.crtc); > > struct intel_crtc_state *crtc_state = crtc ? > > @@ -16378,21 +16378,21 @@ static void intel_sanitize_encoder(struct intel_encoder *encoder) > > /* notify opregion of the sanitized encoder state */ > > intel_opregion_notify_encoder(encoder, connector && has_active_crtc); > > - if (INTEL_GEN(dev_priv) >= 11) > > + if (INTEL_GEN(i915) >= 11) > > icl_sanitize_encoder_pll_mapping(encoder); > > } > > -void i915_redisable_vga_power_on(struct drm_i915_private *dev_priv) > > +void i915_redisable_vga_power_on(struct drm_i915_private *i915) > > { > > - i915_reg_t vga_reg = i915_vgacntrl_reg(dev_priv); > > + i915_reg_t vga_reg = i915_vgacntrl_reg(i915); > > if (!(I915_READ(vga_reg) & VGA_DISP_DISABLE)) { > > DRM_DEBUG_KMS("Something enabled VGA plane, disabling it\n"); > > - i915_disable_vga(dev_priv); > > + i915_disable_vga(i915); > > } > > } > > -void i915_redisable_vga(struct drm_i915_private *dev_priv) > > +void i915_redisable_vga(struct drm_i915_private *i915) > > { > > intel_wakeref_t wakeref; > > @@ -16405,23 +16405,23 @@ void i915_redisable_vga(struct drm_i915_private *dev_priv) > > * follow the "don't touch the power well if we don't need it" policy > > * the rest of the driver uses. > > */ > > - wakeref = intel_display_power_get_if_enabled(dev_priv, > > + wakeref = intel_display_power_get_if_enabled(i915, > > POWER_DOMAIN_VGA); > > if (!wakeref) > > return; > > - i915_redisable_vga_power_on(dev_priv); > > + i915_redisable_vga_power_on(i915); > > - intel_display_power_put(dev_priv, POWER_DOMAIN_VGA, wakeref); > > + intel_display_power_put(i915, POWER_DOMAIN_VGA, wakeref); > > } > > /* FIXME read out full plane state for all planes */ > > -static void readout_plane_state(struct drm_i915_private *dev_priv) > > +static void readout_plane_state(struct drm_i915_private *i915) > > { > > struct intel_plane *plane; > > struct intel_crtc *crtc; > > - for_each_intel_plane(&dev_priv->drm, plane) { > > + for_each_intel_plane(&i915->drm, plane) { > > struct intel_plane_state *plane_state = > > to_intel_plane_state(plane->base.state); > > struct intel_crtc_state *crtc_state; > > @@ -16430,7 +16430,7 @@ static void readout_plane_state(struct drm_i915_private *dev_priv) > > visible = plane->get_hw_state(plane, &pipe); > > - crtc = intel_get_crtc_for_pipe(dev_priv, pipe); > > + crtc = intel_get_crtc_for_pipe(i915, pipe); > > crtc_state = to_intel_crtc_state(crtc->base.state); > > intel_set_plane_visible(crtc_state, plane_state, visible); > > @@ -16440,7 +16440,7 @@ static void readout_plane_state(struct drm_i915_private *dev_priv) > > enableddisabled(visible), pipe_name(pipe)); > > } > > - for_each_intel_crtc(&dev_priv->drm, crtc) { > > + for_each_intel_crtc(&i915->drm, crtc) { > > struct intel_crtc_state *crtc_state = > > to_intel_crtc_state(crtc->base.state); > > @@ -16450,7 +16450,7 @@ static void readout_plane_state(struct drm_i915_private *dev_priv) > > static void intel_modeset_readout_hw_state(struct drm_device *dev) > > { > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > enum pipe pipe; > > struct intel_crtc *crtc; > > struct intel_encoder *encoder; > > @@ -16458,7 +16458,7 @@ static void intel_modeset_readout_hw_state(struct drm_device *dev) > > struct drm_connector_list_iter conn_iter; > > int i; > > - dev_priv->active_crtcs = 0; > > + i915->active_crtcs = 0; > > for_each_intel_crtc(dev, crtc) { > > struct intel_crtc_state *crtc_state = > > @@ -16469,25 +16469,25 @@ static void intel_modeset_readout_hw_state(struct drm_device *dev) > > __drm_atomic_helper_crtc_reset(&crtc->base, &crtc_state->base); > > crtc_state->base.active = crtc_state->base.enable = > > - dev_priv->display.get_pipe_config(crtc, crtc_state); > > + i915->display.get_pipe_config(crtc, crtc_state); > > crtc->base.enabled = crtc_state->base.enable; > > crtc->active = crtc_state->base.active; > > if (crtc_state->base.active) > > - dev_priv->active_crtcs |= 1 << crtc->pipe; > > + i915->active_crtcs |= 1 << crtc->pipe; > > DRM_DEBUG_KMS("[CRTC:%d:%s] hw state readout: %s\n", > > crtc->base.base.id, crtc->base.name, > > enableddisabled(crtc_state->base.active)); > > } > > - readout_plane_state(dev_priv); > > + readout_plane_state(i915); > > - for (i = 0; i < dev_priv->num_shared_dpll; i++) { > > - struct intel_shared_dpll *pll = &dev_priv->shared_dplls[i]; > > + for (i = 0; i < i915->num_shared_dpll; i++) { > > + struct intel_shared_dpll *pll = &i915->shared_dplls[i]; > > - pll->on = pll->info->funcs->get_hw_state(dev_priv, pll, > > + pll->on = pll->info->funcs->get_hw_state(i915, pll, > > &pll->state.hw_state); > > pll->state.crtc_mask = 0; > > for_each_intel_crtc(dev, crtc) { > > @@ -16510,7 +16510,7 @@ static void intel_modeset_readout_hw_state(struct drm_device *dev) > > if (encoder->get_hw_state(encoder, &pipe)) { > > struct intel_crtc_state *crtc_state; > > - crtc = intel_get_crtc_for_pipe(dev_priv, pipe); > > + crtc = intel_get_crtc_for_pipe(i915, pipe); > > crtc_state = to_intel_crtc_state(crtc->base.state); > > encoder->base.crtc = &crtc->base; > > @@ -16558,7 +16558,7 @@ static void intel_modeset_readout_hw_state(struct drm_device *dev) > > for_each_intel_crtc(dev, crtc) { > > struct intel_bw_state *bw_state = > > - to_intel_bw_state(dev_priv->bw_obj.state); > > + to_intel_bw_state(i915->bw_obj.state); > > struct intel_crtc_state *crtc_state = > > to_intel_crtc_state(crtc->base.state); > > struct intel_plane *plane; > > @@ -16585,7 +16585,7 @@ static void intel_modeset_readout_hw_state(struct drm_device *dev) > > intel_crtc_compute_pixel_rate(crtc_state); > > - if (dev_priv->display.modeset_calc_cdclk) { > > + if (i915->display.modeset_calc_cdclk) { > > min_cdclk = intel_crtc_compute_min_cdclk(crtc_state); > > if (WARN_ON(min_cdclk < 0)) > > min_cdclk = 0; > > @@ -16596,11 +16596,11 @@ static void intel_modeset_readout_hw_state(struct drm_device *dev) > > update_scanline_offset(crtc_state); > > } > > - dev_priv->min_cdclk[crtc->pipe] = min_cdclk; > > - dev_priv->min_voltage_level[crtc->pipe] = > > + i915->min_cdclk[crtc->pipe] = min_cdclk; > > + i915->min_voltage_level[crtc->pipe] = > > crtc_state->min_voltage_level; > > - for_each_intel_plane_on_crtc(&dev_priv->drm, crtc, plane) { > > + for_each_intel_plane_on_crtc(&i915->drm, crtc, plane) { > > const struct intel_plane_state *plane_state = > > to_intel_plane_state(plane->base.state); > > @@ -16615,16 +16615,16 @@ static void intel_modeset_readout_hw_state(struct drm_device *dev) > > intel_bw_crtc_update(bw_state, crtc_state); > > - intel_pipe_config_sanity_check(dev_priv, crtc_state); > > + intel_pipe_config_sanity_check(i915, crtc_state); > > } > > } > > static void > > -get_encoder_power_domains(struct drm_i915_private *dev_priv) > > +get_encoder_power_domains(struct drm_i915_private *i915) > > { > > struct intel_encoder *encoder; > > - for_each_intel_encoder(&dev_priv->drm, encoder) { > > + for_each_intel_encoder(&i915->drm, encoder) { > > struct intel_crtc_state *crtc_state; > > if (!encoder->get_power_domains) > > @@ -16642,14 +16642,14 @@ get_encoder_power_domains(struct drm_i915_private *dev_priv) > > } > > } > > -static void intel_early_display_was(struct drm_i915_private *dev_priv) > > +static void intel_early_display_was(struct drm_i915_private *i915) > > { > > /* Display WA #1185 WaDisableDARBFClkGating:cnl,glk */ > > - if (IS_CANNONLAKE(dev_priv) || IS_GEMINILAKE(dev_priv)) > > + if (IS_CANNONLAKE(i915) || IS_GEMINILAKE(i915)) > > I915_WRITE(GEN9_CLKGATE_DIS_0, I915_READ(GEN9_CLKGATE_DIS_0) | > > DARBF_GATING_DIS); > > - if (IS_HASWELL(dev_priv)) { > > + if (IS_HASWELL(i915)) { > > /* > > * WaRsPkgCStateDisplayPMReq:hsw > > * System hang if this isn't done before disabling all planes! > > @@ -16659,7 +16659,7 @@ static void intel_early_display_was(struct drm_i915_private *dev_priv) > > } > > } > > -static void ibx_sanitize_pch_hdmi_port(struct drm_i915_private *dev_priv, > > +static void ibx_sanitize_pch_hdmi_port(struct drm_i915_private *i915, > > enum port port, i915_reg_t hdmi_reg) > > { > > u32 val = I915_READ(hdmi_reg); > > @@ -16677,7 +16677,7 @@ static void ibx_sanitize_pch_hdmi_port(struct drm_i915_private *dev_priv, > > I915_WRITE(hdmi_reg, val); > > } > > -static void ibx_sanitize_pch_dp_port(struct drm_i915_private *dev_priv, > > +static void ibx_sanitize_pch_dp_port(struct drm_i915_private *i915, > > enum port port, i915_reg_t dp_reg) > > { > > u32 val = I915_READ(dp_reg); > > @@ -16695,7 +16695,7 @@ static void ibx_sanitize_pch_dp_port(struct drm_i915_private *dev_priv, > > I915_WRITE(dp_reg, val); > > } > > -static void ibx_sanitize_pch_ports(struct drm_i915_private *dev_priv) > > +static void ibx_sanitize_pch_ports(struct drm_i915_private *i915) > > { > > /* > > * The BIOS may select transcoder B on some of the PCH > > @@ -16708,14 +16708,14 @@ static void ibx_sanitize_pch_ports(struct drm_i915_private *dev_priv) > > * (see. intel_dp_link_down(), intel_disable_hdmi(), > > * intel_disable_sdvo()). > > */ > > - ibx_sanitize_pch_dp_port(dev_priv, PORT_B, PCH_DP_B); > > - ibx_sanitize_pch_dp_port(dev_priv, PORT_C, PCH_DP_C); > > - ibx_sanitize_pch_dp_port(dev_priv, PORT_D, PCH_DP_D); > > + ibx_sanitize_pch_dp_port(i915, PORT_B, PCH_DP_B); > > + ibx_sanitize_pch_dp_port(i915, PORT_C, PCH_DP_C); > > + ibx_sanitize_pch_dp_port(i915, PORT_D, PCH_DP_D); > > /* PCH SDVOB multiplex with HDMIB */ > > - ibx_sanitize_pch_hdmi_port(dev_priv, PORT_B, PCH_HDMIB); > > - ibx_sanitize_pch_hdmi_port(dev_priv, PORT_C, PCH_HDMIC); > > - ibx_sanitize_pch_hdmi_port(dev_priv, PORT_D, PCH_HDMID); > > + ibx_sanitize_pch_hdmi_port(i915, PORT_B, PCH_HDMIB); > > + ibx_sanitize_pch_hdmi_port(i915, PORT_C, PCH_HDMIC); > > + ibx_sanitize_pch_hdmi_port(i915, PORT_D, PCH_HDMID); > > } > > /* Scan out the current hw modeset state, > > @@ -16725,29 +16725,29 @@ static void > > intel_modeset_setup_hw_state(struct drm_device *dev, > > struct drm_modeset_acquire_ctx *ctx) > > { > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > struct intel_crtc_state *crtc_state; > > struct intel_encoder *encoder; > > struct intel_crtc *crtc; > > intel_wakeref_t wakeref; > > int i; > > - wakeref = intel_display_power_get(dev_priv, POWER_DOMAIN_INIT); > > + wakeref = intel_display_power_get(i915, POWER_DOMAIN_INIT); > > - intel_early_display_was(dev_priv); > > + intel_early_display_was(i915); > > intel_modeset_readout_hw_state(dev); > > /* HW state is read out, now we need to sanitize this mess. */ > > - get_encoder_power_domains(dev_priv); > > + get_encoder_power_domains(i915); > > - if (HAS_PCH_IBX(dev_priv)) > > - ibx_sanitize_pch_ports(dev_priv); > > + if (HAS_PCH_IBX(i915)) > > + ibx_sanitize_pch_ports(i915); > > /* > > * intel_sanitize_plane_mapping() may need to do vblank > > * waits, so we need vblank interrupts restored beforehand. > > */ > > - for_each_intel_crtc(&dev_priv->drm, crtc) { > > + for_each_intel_crtc(&i915->drm, crtc) { > > crtc_state = to_intel_crtc_state(crtc->base.state); > > drm_crtc_vblank_reset(&crtc->base); > > @@ -16756,12 +16756,12 @@ intel_modeset_setup_hw_state(struct drm_device *dev, > > intel_crtc_vblank_on(crtc_state); > > } > > - intel_sanitize_plane_mapping(dev_priv); > > + intel_sanitize_plane_mapping(i915); > > for_each_intel_encoder(dev, encoder) > > intel_sanitize_encoder(encoder); > > - for_each_intel_crtc(&dev_priv->drm, crtc) { > > + for_each_intel_crtc(&i915->drm, crtc) { > > crtc_state = to_intel_crtc_state(crtc->base.state); > > intel_sanitize_crtc(crtc, ctx); > > intel_dump_pipe_config(crtc_state, NULL, "[setup_hw_state]"); > > @@ -16769,8 +16769,8 @@ intel_modeset_setup_hw_state(struct drm_device *dev, > > intel_modeset_update_connector_atomic_state(dev); > > - for (i = 0; i < dev_priv->num_shared_dpll; i++) { > > - struct intel_shared_dpll *pll = &dev_priv->shared_dplls[i]; > > + for (i = 0; i < i915->num_shared_dpll; i++) { > > + struct intel_shared_dpll *pll = &i915->shared_dplls[i]; > > if (!pll->on || pll->active_mask) > > continue; > > @@ -16778,20 +16778,20 @@ intel_modeset_setup_hw_state(struct drm_device *dev, > > DRM_DEBUG_KMS("%s enabled but not in use, disabling\n", > > pll->info->name); > > - pll->info->funcs->disable(dev_priv, pll); > > + pll->info->funcs->disable(i915, pll); > > pll->on = false; > > } > > - if (IS_G4X(dev_priv)) { > > - g4x_wm_get_hw_state(dev_priv); > > - g4x_wm_sanitize(dev_priv); > > - } else if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) { > > - vlv_wm_get_hw_state(dev_priv); > > - vlv_wm_sanitize(dev_priv); > > - } else if (INTEL_GEN(dev_priv) >= 9) { > > - skl_wm_get_hw_state(dev_priv); > > - } else if (HAS_PCH_SPLIT(dev_priv)) { > > - ilk_wm_get_hw_state(dev_priv); > > + if (IS_G4X(i915)) { > > + g4x_wm_get_hw_state(i915); > > + g4x_wm_sanitize(i915); > > + } else if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) { > > + vlv_wm_get_hw_state(i915); > > + vlv_wm_sanitize(i915); > > + } else if (INTEL_GEN(i915) >= 9) { > > + skl_wm_get_hw_state(i915); > > + } else if (HAS_PCH_SPLIT(i915)) { > > + ilk_wm_get_hw_state(i915); > > } > > for_each_intel_crtc(dev, crtc) { > > @@ -16800,22 +16800,22 @@ intel_modeset_setup_hw_state(struct drm_device *dev, > > crtc_state = to_intel_crtc_state(crtc->base.state); > > put_domains = modeset_get_crtc_power_domains(&crtc->base, crtc_state); > > if (WARN_ON(put_domains)) > > - modeset_put_power_domains(dev_priv, put_domains); > > + modeset_put_power_domains(i915, put_domains); > > } > > - intel_display_power_put(dev_priv, POWER_DOMAIN_INIT, wakeref); > > + intel_display_power_put(i915, POWER_DOMAIN_INIT, wakeref); > > - intel_fbc_init_pipe_state(dev_priv); > > + intel_fbc_init_pipe_state(i915); > > } > > void intel_display_resume(struct drm_device *dev) > > { > > - struct drm_i915_private *dev_priv = to_i915(dev); > > - struct drm_atomic_state *state = dev_priv->modeset_restore_state; > > + struct drm_i915_private *i915 = to_i915(dev); > > + struct drm_atomic_state *state = i915->modeset_restore_state; > > struct drm_modeset_acquire_ctx ctx; > > int ret; > > - dev_priv->modeset_restore_state = NULL; > > + i915->modeset_restore_state = NULL; > > if (state) > > state->acquire_ctx = &ctx; > > @@ -16832,7 +16832,7 @@ void intel_display_resume(struct drm_device *dev) > > if (!ret) > > ret = __intel_display_resume(dev, state, &ctx); > > - intel_enable_ipc(dev_priv); > > + intel_enable_ipc(i915); > > drm_modeset_drop_locks(&ctx); > > drm_modeset_acquire_fini(&ctx); > > @@ -16862,19 +16862,19 @@ static void intel_hpd_poll_fini(struct drm_device *dev) > > void intel_modeset_cleanup(struct drm_device *dev) > > { > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > - flush_workqueue(dev_priv->modeset_wq); > > + flush_workqueue(i915->modeset_wq); > > - flush_work(&dev_priv->atomic_helper.free_work); > > - WARN_ON(!llist_empty(&dev_priv->atomic_helper.free_list)); > > + flush_work(&i915->atomic_helper.free_work); > > + WARN_ON(!llist_empty(&i915->atomic_helper.free_list)); > > /* > > * Interrupts and polling as the first thing to avoid creating havoc. > > * Too much stuff here (turning of connectors, ...) would > > * experience fancy races otherwise. > > */ > > - intel_irq_uninstall(dev_priv); > > + intel_irq_uninstall(i915); > > /* > > * Due to the hpd irq storm handling the hotplug work can re-arm the > > @@ -16883,37 +16883,37 @@ void intel_modeset_cleanup(struct drm_device *dev) > > intel_hpd_poll_fini(dev); > > /* poll work can call into fbdev, hence clean that up afterwards */ > > - intel_fbdev_fini(dev_priv); > > + intel_fbdev_fini(i915); > > intel_unregister_dsm_handler(); > > - intel_fbc_global_disable(dev_priv); > > + intel_fbc_global_disable(i915); > > /* flush any delayed tasks or pending work */ > > flush_scheduled_work(); > > - intel_hdcp_component_fini(dev_priv); > > + intel_hdcp_component_fini(i915); > > drm_mode_config_cleanup(dev); > > - intel_overlay_cleanup(dev_priv); > > + intel_overlay_cleanup(i915); > > - intel_gmbus_teardown(dev_priv); > > + intel_gmbus_teardown(i915); > > - destroy_workqueue(dev_priv->modeset_wq); > > + destroy_workqueue(i915->modeset_wq); > > - intel_fbc_cleanup_cfb(dev_priv); > > + intel_fbc_cleanup_cfb(i915); > > } > > /* > > * set vga decode state - true == enable VGA decode > > */ > > -int intel_modeset_vga_set_state(struct drm_i915_private *dev_priv, bool state) > > +int intel_modeset_vga_set_state(struct drm_i915_private *i915, bool state) > > { > > - unsigned reg = INTEL_GEN(dev_priv) >= 6 ? SNB_GMCH_CTRL : INTEL_GMCH_CTRL; > > + unsigned reg = INTEL_GEN(i915) >= 6 ? SNB_GMCH_CTRL : INTEL_GMCH_CTRL; > > u16 gmch_ctrl; > > - if (pci_read_config_word(dev_priv->bridge_dev, reg, &gmch_ctrl)) { > > + if (pci_read_config_word(i915->bridge_dev, reg, &gmch_ctrl)) { > > DRM_ERROR("failed to read control word\n"); > > return -EIO; > > } > > @@ -16926,7 +16926,7 @@ int intel_modeset_vga_set_state(struct drm_i915_private *dev_priv, bool state) > > else > > gmch_ctrl |= INTEL_GMCH_VGA_DISABLE; > > - if (pci_write_config_word(dev_priv->bridge_dev, reg, gmch_ctrl)) { > > + if (pci_write_config_word(i915->bridge_dev, reg, gmch_ctrl)) { > > DRM_ERROR("failed to write control word\n"); > > return -EIO; > > } > > @@ -16980,7 +16980,7 @@ struct intel_display_error_state { > > }; > > struct intel_display_error_state * > > -intel_display_capture_error_state(struct drm_i915_private *dev_priv) > > +intel_display_capture_error_state(struct drm_i915_private *i915) > > { > > struct intel_display_error_state *error; > > int transcoders[] = { > > @@ -16993,19 +16993,19 @@ intel_display_capture_error_state(struct drm_i915_private *dev_priv) > > BUILD_BUG_ON(ARRAY_SIZE(transcoders) != ARRAY_SIZE(error->transcoder)); > > - if (!HAS_DISPLAY(dev_priv)) > > + if (!HAS_DISPLAY(i915)) > > return NULL; > > error = kzalloc(sizeof(*error), GFP_ATOMIC); > > if (error == NULL) > > return NULL; > > - if (IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv)) > > + if (IS_HASWELL(i915) || IS_BROADWELL(i915)) > > error->power_well_driver = I915_READ(HSW_PWR_WELL_CTL2); > > - for_each_pipe(dev_priv, i) { > > + for_each_pipe(i915, i) { > > error->pipe[i].power_domain_on = > > - __intel_display_power_is_enabled(dev_priv, > > + __intel_display_power_is_enabled(i915, > > POWER_DOMAIN_PIPE(i)); > > if (!error->pipe[i].power_domain_on) > > continue; > > @@ -17016,32 +17016,32 @@ intel_display_capture_error_state(struct drm_i915_private *dev_priv) > > error->plane[i].control = I915_READ(DSPCNTR(i)); > > error->plane[i].stride = I915_READ(DSPSTRIDE(i)); > > - if (INTEL_GEN(dev_priv) <= 3) { > > + if (INTEL_GEN(i915) <= 3) { > > error->plane[i].size = I915_READ(DSPSIZE(i)); > > error->plane[i].pos = I915_READ(DSPPOS(i)); > > } > > - if (INTEL_GEN(dev_priv) <= 7 && !IS_HASWELL(dev_priv)) > > + if (INTEL_GEN(i915) <= 7 && !IS_HASWELL(i915)) > > error->plane[i].addr = I915_READ(DSPADDR(i)); > > - if (INTEL_GEN(dev_priv) >= 4) { > > + if (INTEL_GEN(i915) >= 4) { > > error->plane[i].surface = I915_READ(DSPSURF(i)); > > error->plane[i].tile_offset = I915_READ(DSPTILEOFF(i)); > > } > > error->pipe[i].source = I915_READ(PIPESRC(i)); > > - if (HAS_GMCH(dev_priv)) > > + if (HAS_GMCH(i915)) > > error->pipe[i].stat = I915_READ(PIPESTAT(i)); > > } > > for (i = 0; i < ARRAY_SIZE(error->transcoder); i++) { > > enum transcoder cpu_transcoder = transcoders[i]; > > - if (!INTEL_INFO(dev_priv)->trans_offsets[cpu_transcoder]) > > + if (!INTEL_INFO(i915)->trans_offsets[cpu_transcoder]) > > continue; > > error->transcoder[i].available = true; > > error->transcoder[i].power_domain_on = > > - __intel_display_power_is_enabled(dev_priv, > > + __intel_display_power_is_enabled(i915, > > POWER_DOMAIN_TRANSCODER(cpu_transcoder)); > > if (!error->transcoder[i].power_domain_on) > > continue; > > @@ -17066,17 +17066,17 @@ void > > intel_display_print_error_state(struct drm_i915_error_state_buf *m, > > struct intel_display_error_state *error) > > { > > - struct drm_i915_private *dev_priv = m->i915; > > + struct drm_i915_private *i915 = m->i915; > > int i; > > if (!error) > > return; > > - err_printf(m, "Num Pipes: %d\n", INTEL_INFO(dev_priv)->num_pipes); > > - if (IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv)) > > + err_printf(m, "Num Pipes: %d\n", INTEL_INFO(i915)->num_pipes); > > + if (IS_HASWELL(i915) || IS_BROADWELL(i915)) > > err_printf(m, "PWR_WELL_CTL2: %08x\n", > > error->power_well_driver); > > - for_each_pipe(dev_priv, i) { > > + for_each_pipe(i915, i) { > > err_printf(m, "Pipe [%d]:\n", i); > > err_printf(m, " Power: %s\n", > > onoff(error->pipe[i].power_domain_on)); > > @@ -17086,13 +17086,13 @@ intel_display_print_error_state(struct drm_i915_error_state_buf *m, > > err_printf(m, "Plane [%d]:\n", i); > > err_printf(m, " CNTR: %08x\n", error->plane[i].control); > > err_printf(m, " STRIDE: %08x\n", error->plane[i].stride); > > - if (INTEL_GEN(dev_priv) <= 3) { > > + if (INTEL_GEN(i915) <= 3) { > > err_printf(m, " SIZE: %08x\n", error->plane[i].size); > > err_printf(m, " POS: %08x\n", error->plane[i].pos); > > } > > - if (INTEL_GEN(dev_priv) <= 7 && !IS_HASWELL(dev_priv)) > > + if (INTEL_GEN(i915) <= 7 && !IS_HASWELL(i915)) > > err_printf(m, " ADDR: %08x\n", error->plane[i].addr); > > - if (INTEL_GEN(dev_priv) >= 4) { > > + if (INTEL_GEN(i915) >= 4) { > > err_printf(m, " SURF: %08x\n", error->plane[i].surface); > > err_printf(m, " TILEOFF: %08x\n", error->plane[i].tile_offset); > > } > > diff --git a/drivers/gpu/drm/i915/intel_display.h b/drivers/gpu/drm/i915/intel_display.h > > index ee6b8194a459..7e4b4b1c0ac6 100644 > > --- a/drivers/gpu/drm/i915/intel_display.h > > +++ b/drivers/gpu/drm/i915/intel_display.h > > @@ -125,7 +125,7 @@ enum i9xx_plane_id { > > }; > > #define plane_name(p) ((p) + 'A') > > -#define sprite_name(p, s) ((p) * RUNTIME_INFO(dev_priv)->num_sprites[(p)] + (s) + 'A') > > +#define sprite_name(p, s) ((p) * RUNTIME_INFO(i915)->num_sprites[(p)] + (s) + 'A') > > /* > > * Per-pipe plane identifier. > > @@ -353,8 +353,8 @@ void intel_link_compute_m_n(u16 bpp, int nlanes, > > struct intel_link_m_n *m_n, > > bool constant_n); > > bool is_ccs_modifier(u64 modifier); > > -void lpt_disable_clkout_dp(struct drm_i915_private *dev_priv); > > -u32 intel_plane_fb_max_stride(struct drm_i915_private *dev_priv, > > +void lpt_disable_clkout_dp(struct drm_i915_private *i915); > > +u32 intel_plane_fb_max_stride(struct drm_i915_private *i915, > > u32 pixel_format, u64 modifier); > > bool intel_plane_can_remap(const struct intel_plane_state *plane_state); > > diff --git a/drivers/gpu/drm/i915/intel_display_power.c b/drivers/gpu/drm/i915/intel_display_power.c > > index bb9ef1cea5db..9acfad5c83ef 100644 > > --- a/drivers/gpu/drm/i915/intel_display_power.c > > +++ b/drivers/gpu/drm/i915/intel_display_power.c > > @@ -17,7 +17,7 @@ > > #include "intel_hotplug.h" > > #include "intel_sideband.h" > > -bool intel_display_power_well_is_enabled(struct drm_i915_private *dev_priv, > > +bool intel_display_power_well_is_enabled(struct drm_i915_private *i915, > > enum i915_power_well_id power_well_id); > > const char * > > @@ -122,42 +122,42 @@ intel_display_power_domain_str(enum intel_display_power_domain domain) > > } > > } > > -static void intel_power_well_enable(struct drm_i915_private *dev_priv, > > +static void intel_power_well_enable(struct drm_i915_private *i915, > > struct i915_power_well *power_well) > > { > > DRM_DEBUG_KMS("enabling %s\n", power_well->desc->name); > > - power_well->desc->ops->enable(dev_priv, power_well); > > + power_well->desc->ops->enable(i915, power_well); > > power_well->hw_enabled = true; > > } > > -static void intel_power_well_disable(struct drm_i915_private *dev_priv, > > +static void intel_power_well_disable(struct drm_i915_private *i915, > > struct i915_power_well *power_well) > > { > > DRM_DEBUG_KMS("disabling %s\n", power_well->desc->name); > > power_well->hw_enabled = false; > > - power_well->desc->ops->disable(dev_priv, power_well); > > + power_well->desc->ops->disable(i915, power_well); > > } > > -static void intel_power_well_get(struct drm_i915_private *dev_priv, > > +static void intel_power_well_get(struct drm_i915_private *i915, > > struct i915_power_well *power_well) > > { > > if (!power_well->count++) > > - intel_power_well_enable(dev_priv, power_well); > > + intel_power_well_enable(i915, power_well); > > } > > -static void intel_power_well_put(struct drm_i915_private *dev_priv, > > +static void intel_power_well_put(struct drm_i915_private *i915, > > struct i915_power_well *power_well) > > { > > WARN(!power_well->count, "Use count on power well %s is already zero", > > power_well->desc->name); > > if (!--power_well->count) > > - intel_power_well_disable(dev_priv, power_well); > > + intel_power_well_disable(i915, power_well); > > } > > /** > > * __intel_display_power_is_enabled - unlocked check for a power domain > > - * @dev_priv: i915 device instance > > + * @i915: i915 device instance > > * @domain: power domain to check > > * > > * This is the unlocked version of intel_display_power_is_enabled() and should > > @@ -167,18 +167,18 @@ static void intel_power_well_put(struct drm_i915_private *dev_priv, > > * Returns: > > * True when the power domain is enabled, false otherwise. > > */ > > -bool __intel_display_power_is_enabled(struct drm_i915_private *dev_priv, > > +bool __intel_display_power_is_enabled(struct drm_i915_private *i915, > > enum intel_display_power_domain domain) > > { > > struct i915_power_well *power_well; > > bool is_enabled; > > - if (dev_priv->runtime_pm.suspended) > > + if (i915->runtime_pm.suspended) > > return false; > > is_enabled = true; > > - for_each_power_domain_well_reverse(dev_priv, power_well, BIT_ULL(domain)) { > > + for_each_power_domain_well_reverse(i915, power_well, BIT_ULL(domain)) { > > if (power_well->desc->always_on) > > continue; > > @@ -193,7 +193,7 @@ bool __intel_display_power_is_enabled(struct drm_i915_private *dev_priv, > > /** > > * intel_display_power_is_enabled - check for a power domain > > - * @dev_priv: i915 device instance > > + * @i915: i915 device instance > > * @domain: power domain to check > > * > > * This function can be used to check the hw power domain state. It is mostly > > @@ -208,16 +208,16 @@ bool __intel_display_power_is_enabled(struct drm_i915_private *dev_priv, > > * Returns: > > * True when the power domain is enabled, false otherwise. > > */ > > -bool intel_display_power_is_enabled(struct drm_i915_private *dev_priv, > > +bool intel_display_power_is_enabled(struct drm_i915_private *i915, > > enum intel_display_power_domain domain) > > { > > struct i915_power_domains *power_domains; > > bool ret; > > - power_domains = &dev_priv->power_domains; > > + power_domains = &i915->power_domains; > > mutex_lock(&power_domains->lock); > > - ret = __intel_display_power_is_enabled(dev_priv, domain); > > + ret = __intel_display_power_is_enabled(i915, domain); > > mutex_unlock(&power_domains->lock); > > return ret; > > @@ -229,10 +229,10 @@ bool intel_display_power_is_enabled(struct drm_i915_private *dev_priv, > > * to be enabled, and it will only be disabled if none of the registers is > > * requesting it to be enabled. > > */ > > -static void hsw_power_well_post_enable(struct drm_i915_private *dev_priv, > > +static void hsw_power_well_post_enable(struct drm_i915_private *i915, > > u8 irq_pipe_mask, bool has_vga) > > { > > - struct pci_dev *pdev = dev_priv->drm.pdev; > > + struct pci_dev *pdev = i915->drm.pdev; > > /* > > * After we re-enable the power well, if we touch VGA register 0x3d5 > > @@ -251,31 +251,31 @@ static void hsw_power_well_post_enable(struct drm_i915_private *dev_priv, > > } > > if (irq_pipe_mask) > > - gen8_irq_power_well_post_enable(dev_priv, irq_pipe_mask); > > + gen8_irq_power_well_post_enable(i915, irq_pipe_mask); > > } > > -static void hsw_power_well_pre_disable(struct drm_i915_private *dev_priv, > > +static void hsw_power_well_pre_disable(struct drm_i915_private *i915, > > u8 irq_pipe_mask) > > { > > if (irq_pipe_mask) > > - gen8_irq_power_well_pre_disable(dev_priv, irq_pipe_mask); > > + gen8_irq_power_well_pre_disable(i915, irq_pipe_mask); > > } > > -static void hsw_wait_for_power_well_enable(struct drm_i915_private *dev_priv, > > +static void hsw_wait_for_power_well_enable(struct drm_i915_private *i915, > > struct i915_power_well *power_well) > > { > > const struct i915_power_well_regs *regs = power_well->desc->hsw.regs; > > int pw_idx = power_well->desc->hsw.idx; > > /* Timeout for PW1:10 us, AUX:not specified, other PWs:20 us. */ > > - WARN_ON(intel_wait_for_register(&dev_priv->uncore, > > + WARN_ON(intel_wait_for_register(&i915->uncore, > > regs->driver, > > HSW_PWR_WELL_CTL_STATE(pw_idx), > > HSW_PWR_WELL_CTL_STATE(pw_idx), > > 1)); > > } > > -static u32 hsw_power_well_requesters(struct drm_i915_private *dev_priv, > > +static u32 hsw_power_well_requesters(struct drm_i915_private *i915, > > const struct i915_power_well_regs *regs, > > int pw_idx) > > { > > @@ -291,7 +291,7 @@ static u32 hsw_power_well_requesters(struct drm_i915_private *dev_priv, > > return ret; > > } > > -static void hsw_wait_for_power_well_disable(struct drm_i915_private *dev_priv, > > +static void hsw_wait_for_power_well_disable(struct drm_i915_private *i915, > > struct i915_power_well *power_well) > > { > > const struct i915_power_well_regs *regs = power_well->desc->hsw.regs; > > @@ -310,7 +310,7 @@ static void hsw_wait_for_power_well_disable(struct drm_i915_private *dev_priv, > > */ > > wait_for((disabled = !(I915_READ(regs->driver) & > > HSW_PWR_WELL_CTL_STATE(pw_idx))) || > > - (reqs = hsw_power_well_requesters(dev_priv, regs, pw_idx)), 1); > > + (reqs = hsw_power_well_requesters(i915, regs, pw_idx)), 1); > > if (disabled) > > return; > > @@ -319,16 +319,16 @@ static void hsw_wait_for_power_well_disable(struct drm_i915_private *dev_priv, > > !!(reqs & 1), !!(reqs & 2), !!(reqs & 4), !!(reqs & 8)); > > } > > -static void gen9_wait_for_power_well_fuses(struct drm_i915_private *dev_priv, > > +static void gen9_wait_for_power_well_fuses(struct drm_i915_private *i915, > > enum skl_power_gate pg) > > { > > /* Timeout 5us for PG#0, for other PGs 1us */ > > - WARN_ON(intel_wait_for_register(&dev_priv->uncore, SKL_FUSE_STATUS, > > + WARN_ON(intel_wait_for_register(&i915->uncore, SKL_FUSE_STATUS, > > SKL_FUSE_PG_DIST_STATUS(pg), > > SKL_FUSE_PG_DIST_STATUS(pg), 1)); > > } > > -static void hsw_power_well_enable(struct drm_i915_private *dev_priv, > > +static void hsw_power_well_enable(struct drm_i915_private *i915, > > struct i915_power_well *power_well) > > { > > const struct i915_power_well_regs *regs = power_well->desc->hsw.regs; > > @@ -338,7 +338,7 @@ static void hsw_power_well_enable(struct drm_i915_private *dev_priv, > > u32 val; > > if (wait_fuses) { > > - pg = INTEL_GEN(dev_priv) >= 11 ? ICL_PW_CTL_IDX_TO_PG(pw_idx) : > > + pg = INTEL_GEN(i915) >= 11 ? ICL_PW_CTL_IDX_TO_PG(pw_idx) : > > SKL_PW_CTL_IDX_TO_PG(pw_idx); > > /* > > * For PW1 we have to wait both for the PW0/PG0 fuse state > > @@ -348,15 +348,15 @@ static void hsw_power_well_enable(struct drm_i915_private *dev_priv, > > * after the enabling. > > */ > > if (pg == SKL_PG1) > > - gen9_wait_for_power_well_fuses(dev_priv, SKL_PG0); > > + gen9_wait_for_power_well_fuses(i915, SKL_PG0); > > } > > val = I915_READ(regs->driver); > > I915_WRITE(regs->driver, val | HSW_PWR_WELL_CTL_REQ(pw_idx)); > > - hsw_wait_for_power_well_enable(dev_priv, power_well); > > + hsw_wait_for_power_well_enable(i915, power_well); > > /* Display WA #1178: cnl */ > > - if (IS_CANNONLAKE(dev_priv) && > > + if (IS_CANNONLAKE(i915) && > > pw_idx >= GLK_PW_CTL_IDX_AUX_B && > > pw_idx <= CNL_PW_CTL_IDX_AUX_F) { > > val = I915_READ(CNL_AUX_ANAOVRD1(pw_idx)); > > @@ -365,32 +365,32 @@ static void hsw_power_well_enable(struct drm_i915_private *dev_priv, > > } > > if (wait_fuses) > > - gen9_wait_for_power_well_fuses(dev_priv, pg); > > + gen9_wait_for_power_well_fuses(i915, pg); > > - hsw_power_well_post_enable(dev_priv, > > + hsw_power_well_post_enable(i915, > > power_well->desc->hsw.irq_pipe_mask, > > power_well->desc->hsw.has_vga); > > } > > -static void hsw_power_well_disable(struct drm_i915_private *dev_priv, > > +static void hsw_power_well_disable(struct drm_i915_private *i915, > > struct i915_power_well *power_well) > > { > > const struct i915_power_well_regs *regs = power_well->desc->hsw.regs; > > int pw_idx = power_well->desc->hsw.idx; > > u32 val; > > - hsw_power_well_pre_disable(dev_priv, > > + hsw_power_well_pre_disable(i915, > > power_well->desc->hsw.irq_pipe_mask); > > val = I915_READ(regs->driver); > > I915_WRITE(regs->driver, val & ~HSW_PWR_WELL_CTL_REQ(pw_idx)); > > - hsw_wait_for_power_well_disable(dev_priv, power_well); > > + hsw_wait_for_power_well_disable(i915, power_well); > > } > > #define ICL_AUX_PW_TO_PORT(pw_idx) ((pw_idx) - ICL_PW_CTL_IDX_AUX_A) > > static void > > -icl_combo_phy_aux_power_well_enable(struct drm_i915_private *dev_priv, > > +icl_combo_phy_aux_power_well_enable(struct drm_i915_private *i915, > > struct i915_power_well *power_well) > > { > > const struct i915_power_well_regs *regs = power_well->desc->hsw.regs; > > @@ -404,12 +404,12 @@ icl_combo_phy_aux_power_well_enable(struct drm_i915_private *dev_priv, > > val = I915_READ(ICL_PORT_CL_DW12(port)); > > I915_WRITE(ICL_PORT_CL_DW12(port), val | ICL_LANE_ENABLE_AUX); > > - hsw_wait_for_power_well_enable(dev_priv, power_well); > > + hsw_wait_for_power_well_enable(i915, power_well); > > /* Display WA #1178: icl */ > > - if (IS_ICELAKE(dev_priv) && > > + if (IS_ICELAKE(i915) && > > pw_idx >= ICL_PW_CTL_IDX_AUX_A && pw_idx <= ICL_PW_CTL_IDX_AUX_B && > > - !intel_bios_is_port_edp(dev_priv, port)) { > > + !intel_bios_is_port_edp(i915, port)) { > > val = I915_READ(ICL_AUX_ANAOVRD1(pw_idx)); > > val |= ICL_AUX_ANAOVRD1_ENABLE | ICL_AUX_ANAOVRD1_LDO_BYPASS; > > I915_WRITE(ICL_AUX_ANAOVRD1(pw_idx), val); > > @@ -417,7 +417,7 @@ icl_combo_phy_aux_power_well_enable(struct drm_i915_private *dev_priv, > > } > > static void > > -icl_combo_phy_aux_power_well_disable(struct drm_i915_private *dev_priv, > > +icl_combo_phy_aux_power_well_disable(struct drm_i915_private *i915, > > struct i915_power_well *power_well) > > { > > const struct i915_power_well_regs *regs = power_well->desc->hsw.regs; > > @@ -431,14 +431,14 @@ icl_combo_phy_aux_power_well_disable(struct drm_i915_private *dev_priv, > > val = I915_READ(regs->driver); > > I915_WRITE(regs->driver, val & ~HSW_PWR_WELL_CTL_REQ(pw_idx)); > > - hsw_wait_for_power_well_disable(dev_priv, power_well); > > + hsw_wait_for_power_well_disable(i915, power_well); > > } > > #define ICL_AUX_PW_TO_CH(pw_idx) \ > > ((pw_idx) - ICL_PW_CTL_IDX_AUX_A + AUX_CH_A) > > static void > > -icl_tc_phy_aux_power_well_enable(struct drm_i915_private *dev_priv, > > +icl_tc_phy_aux_power_well_enable(struct drm_i915_private *i915, > > struct i915_power_well *power_well) > > { > > enum aux_ch aux_ch = ICL_AUX_PW_TO_CH(power_well->desc->hsw.idx); > > @@ -450,7 +450,7 @@ icl_tc_phy_aux_power_well_enable(struct drm_i915_private *dev_priv, > > val |= DP_AUX_CH_CTL_TBT_IO; > > I915_WRITE(DP_AUX_CH_CTL(aux_ch), val); > > - hsw_power_well_enable(dev_priv, power_well); > > + hsw_power_well_enable(i915, power_well); > > } > > /* > > @@ -458,7 +458,7 @@ icl_tc_phy_aux_power_well_enable(struct drm_i915_private *dev_priv, > > * enable it, so check if it's enabled and also check if we've requested it to > > * be enabled. > > */ > > -static bool hsw_power_well_enabled(struct drm_i915_private *dev_priv, > > +static bool hsw_power_well_enabled(struct drm_i915_private *i915, > > struct i915_power_well *power_well) > > { > > const struct i915_power_well_regs *regs = power_well->desc->hsw.regs; > > @@ -476,14 +476,14 @@ static bool hsw_power_well_enabled(struct drm_i915_private *dev_priv, > > * BIOS's own request bits, which are forced-on for these power wells > > * when exiting DC5/6. > > */ > > - if (IS_GEN(dev_priv, 9) && !IS_GEN9_LP(dev_priv) && > > + if (IS_GEN(i915, 9) && !IS_GEN9_LP(i915) && > > (id == SKL_DISP_PW_1 || id == SKL_DISP_PW_MISC_IO)) > > val |= I915_READ(regs->bios); > > return (val & mask) == mask; > > } > > -static void assert_can_enable_dc9(struct drm_i915_private *dev_priv) > > +static void assert_can_enable_dc9(struct drm_i915_private *i915) > > { > > WARN_ONCE((I915_READ(DC_STATE_EN) & DC_STATE_EN_DC9), > > "DC9 already programmed to be enabled.\n"); > > @@ -492,7 +492,7 @@ static void assert_can_enable_dc9(struct drm_i915_private *dev_priv) > > WARN_ONCE(I915_READ(HSW_PWR_WELL_CTL2) & > > HSW_PWR_WELL_CTL_REQ(SKL_PW_CTL_IDX_PW_2), > > "Power well 2 on.\n"); > > - WARN_ONCE(intel_irqs_enabled(dev_priv), > > + WARN_ONCE(intel_irqs_enabled(i915), > > "Interrupts not disabled yet.\n"); > > /* > > @@ -504,9 +504,9 @@ static void assert_can_enable_dc9(struct drm_i915_private *dev_priv) > > */ > > } > > -static void assert_can_disable_dc9(struct drm_i915_private *dev_priv) > > +static void assert_can_disable_dc9(struct drm_i915_private *i915) > > { > > - WARN_ONCE(intel_irqs_enabled(dev_priv), > > + WARN_ONCE(intel_irqs_enabled(i915), > > "Interrupts not disabled yet.\n"); > > WARN_ONCE(I915_READ(DC_STATE_EN) & DC_STATE_EN_UPTO_DC5, > > "DC5 still not disabled.\n"); > > @@ -520,7 +520,7 @@ static void assert_can_disable_dc9(struct drm_i915_private *dev_priv) > > */ > > } > > -static void gen9_write_dc_state(struct drm_i915_private *dev_priv, > > +static void gen9_write_dc_state(struct drm_i915_private *i915, > > u32 state) > > { > > int rewrites = 0; > > @@ -557,14 +557,14 @@ static void gen9_write_dc_state(struct drm_i915_private *dev_priv, > > state, rewrites); > > } > > -static u32 gen9_dc_mask(struct drm_i915_private *dev_priv) > > +static u32 gen9_dc_mask(struct drm_i915_private *i915) > > { > > u32 mask; > > mask = DC_STATE_EN_UPTO_DC5; > > - if (INTEL_GEN(dev_priv) >= 11) > > + if (INTEL_GEN(i915) >= 11) > > mask |= DC_STATE_EN_UPTO_DC6 | DC_STATE_EN_DC9; > > - else if (IS_GEN9_LP(dev_priv)) > > + else if (IS_GEN9_LP(i915)) > > mask |= DC_STATE_EN_DC9; > > else > > mask |= DC_STATE_EN_UPTO_DC6; > > @@ -572,20 +572,20 @@ static u32 gen9_dc_mask(struct drm_i915_private *dev_priv) > > return mask; > > } > > -void gen9_sanitize_dc_state(struct drm_i915_private *dev_priv) > > +void gen9_sanitize_dc_state(struct drm_i915_private *i915) > > { > > u32 val; > > - val = I915_READ(DC_STATE_EN) & gen9_dc_mask(dev_priv); > > + val = I915_READ(DC_STATE_EN) & gen9_dc_mask(i915); > > DRM_DEBUG_KMS("Resetting DC state tracking from %02x to %02x\n", > > - dev_priv->csr.dc_state, val); > > - dev_priv->csr.dc_state = val; > > + i915->csr.dc_state, val); > > + i915->csr.dc_state = val; > > } > > /** > > * gen9_set_dc_state - set target display C power state > > - * @dev_priv: i915 device instance > > + * @i915: i915 device instance > > * @state: target DC power state > > * - DC_STATE_DISABLE > > * - DC_STATE_EN_UPTO_DC5 > > @@ -606,35 +606,35 @@ void gen9_sanitize_dc_state(struct drm_i915_private *dev_priv) > > * back on and register state is restored. This is guaranteed by the MMIO write > > * to DC_STATE_EN blocking until the state is restored. > > */ > > -static void gen9_set_dc_state(struct drm_i915_private *dev_priv, u32 state) > > +static void gen9_set_dc_state(struct drm_i915_private *i915, u32 state) > > { > > u32 val; > > u32 mask; > > - if (WARN_ON_ONCE(state & ~dev_priv->csr.allowed_dc_mask)) > > - state &= dev_priv->csr.allowed_dc_mask; > > + if (WARN_ON_ONCE(state & ~i915->csr.allowed_dc_mask)) > > + state &= i915->csr.allowed_dc_mask; > > val = I915_READ(DC_STATE_EN); > > - mask = gen9_dc_mask(dev_priv); > > + mask = gen9_dc_mask(i915); > > DRM_DEBUG_KMS("Setting DC state from %02x to %02x\n", > > val & mask, state); > > /* Check if DMC is ignoring our DC state requests */ > > - if ((val & mask) != dev_priv->csr.dc_state) > > + if ((val & mask) != i915->csr.dc_state) > > DRM_ERROR("DC state mismatch (0x%x -> 0x%x)\n", > > - dev_priv->csr.dc_state, val & mask); > > + i915->csr.dc_state, val & mask); > > val &= ~mask; > > val |= state; > > - gen9_write_dc_state(dev_priv, val); > > + gen9_write_dc_state(i915, val); > > - dev_priv->csr.dc_state = val & mask; > > + i915->csr.dc_state = val & mask; > > } > > -void bxt_enable_dc9(struct drm_i915_private *dev_priv) > > +void bxt_enable_dc9(struct drm_i915_private *i915) > > { > > - assert_can_enable_dc9(dev_priv); > > + assert_can_enable_dc9(i915); > > DRM_DEBUG_KMS("Enabling DC9\n"); > > /* > > @@ -642,23 +642,23 @@ void bxt_enable_dc9(struct drm_i915_private *dev_priv) > > * platforms with South Display Engine on PCH, > > * because PPS registers are always on. > > */ > > - if (!HAS_PCH_SPLIT(dev_priv)) > > - intel_power_sequencer_reset(dev_priv); > > - gen9_set_dc_state(dev_priv, DC_STATE_EN_DC9); > > + if (!HAS_PCH_SPLIT(i915)) > > + intel_power_sequencer_reset(i915); > > + gen9_set_dc_state(i915, DC_STATE_EN_DC9); > > } > > -void bxt_disable_dc9(struct drm_i915_private *dev_priv) > > +void bxt_disable_dc9(struct drm_i915_private *i915) > > { > > - assert_can_disable_dc9(dev_priv); > > + assert_can_disable_dc9(i915); > > DRM_DEBUG_KMS("Disabling DC9\n"); > > - gen9_set_dc_state(dev_priv, DC_STATE_DISABLE); > > + gen9_set_dc_state(i915, DC_STATE_DISABLE); > > - intel_pps_unlock_regs_wa(dev_priv); > > + intel_pps_unlock_regs_wa(i915); > > } > > -static void assert_csr_loaded(struct drm_i915_private *dev_priv) > > +static void assert_csr_loaded(struct drm_i915_private *i915) > > { > > WARN_ONCE(!I915_READ(CSR_PROGRAM(0)), > > "CSR program storage start is NULL\n"); > > @@ -667,12 +667,12 @@ static void assert_csr_loaded(struct drm_i915_private *dev_priv) > > } > > static struct i915_power_well * > > -lookup_power_well(struct drm_i915_private *dev_priv, > > +lookup_power_well(struct drm_i915_private *i915, > > enum i915_power_well_id power_well_id) > > { > > struct i915_power_well *power_well; > > - for_each_power_well(dev_priv, power_well) > > + for_each_power_well(i915, power_well) > > if (power_well->desc->id == power_well_id) > > return power_well; > > @@ -684,62 +684,62 @@ lookup_power_well(struct drm_i915_private *dev_priv, > > * our driver. > > */ > > WARN(1, "Power well %d not defined for this platform\n", power_well_id); > > - return &dev_priv->power_domains.power_wells[0]; > > + return &i915->power_domains.power_wells[0]; > > } > > -static void assert_can_enable_dc5(struct drm_i915_private *dev_priv) > > +static void assert_can_enable_dc5(struct drm_i915_private *i915) > > { > > - bool pg2_enabled = intel_display_power_well_is_enabled(dev_priv, > > + bool pg2_enabled = intel_display_power_well_is_enabled(i915, > > SKL_DISP_PW_2); > > WARN_ONCE(pg2_enabled, "PG2 not disabled to enable DC5.\n"); > > WARN_ONCE((I915_READ(DC_STATE_EN) & DC_STATE_EN_UPTO_DC5), > > "DC5 already programmed to be enabled.\n"); > > - assert_rpm_wakelock_held(dev_priv); > > + assert_rpm_wakelock_held(i915); > > - assert_csr_loaded(dev_priv); > > + assert_csr_loaded(i915); > > } > > -void gen9_enable_dc5(struct drm_i915_private *dev_priv) > > +void gen9_enable_dc5(struct drm_i915_private *i915) > > { > > - assert_can_enable_dc5(dev_priv); > > + assert_can_enable_dc5(i915); > > DRM_DEBUG_KMS("Enabling DC5\n"); > > /* Wa Display #1183: skl,kbl,cfl */ > > - if (IS_GEN9_BC(dev_priv)) > > + if (IS_GEN9_BC(i915)) > > I915_WRITE(GEN8_CHICKEN_DCPR_1, I915_READ(GEN8_CHICKEN_DCPR_1) | > > SKL_SELECT_ALTERNATE_DC_EXIT); > > - gen9_set_dc_state(dev_priv, DC_STATE_EN_UPTO_DC5); > > + gen9_set_dc_state(i915, DC_STATE_EN_UPTO_DC5); > > } > > -static void assert_can_enable_dc6(struct drm_i915_private *dev_priv) > > +static void assert_can_enable_dc6(struct drm_i915_private *i915) > > { > > WARN_ONCE(I915_READ(UTIL_PIN_CTL) & UTIL_PIN_ENABLE, > > "Backlight is not disabled.\n"); > > WARN_ONCE((I915_READ(DC_STATE_EN) & DC_STATE_EN_UPTO_DC6), > > "DC6 already programmed to be enabled.\n"); > > - assert_csr_loaded(dev_priv); > > + assert_csr_loaded(i915); > > } > > -void skl_enable_dc6(struct drm_i915_private *dev_priv) > > +void skl_enable_dc6(struct drm_i915_private *i915) > > { > > - assert_can_enable_dc6(dev_priv); > > + assert_can_enable_dc6(i915); > > DRM_DEBUG_KMS("Enabling DC6\n"); > > /* Wa Display #1183: skl,kbl,cfl */ > > - if (IS_GEN9_BC(dev_priv)) > > + if (IS_GEN9_BC(i915)) > > I915_WRITE(GEN8_CHICKEN_DCPR_1, I915_READ(GEN8_CHICKEN_DCPR_1) | > > SKL_SELECT_ALTERNATE_DC_EXIT); > > - gen9_set_dc_state(dev_priv, DC_STATE_EN_UPTO_DC6); > > + gen9_set_dc_state(i915, DC_STATE_EN_UPTO_DC6); > > } > > -static void hsw_power_well_sync_hw(struct drm_i915_private *dev_priv, > > +static void hsw_power_well_sync_hw(struct drm_i915_private *i915, > > struct i915_power_well *power_well) > > { > > const struct i915_power_well_regs *regs = power_well->desc->hsw.regs; > > @@ -757,52 +757,52 @@ static void hsw_power_well_sync_hw(struct drm_i915_private *dev_priv, > > } > > } > > -static void bxt_dpio_cmn_power_well_enable(struct drm_i915_private *dev_priv, > > +static void bxt_dpio_cmn_power_well_enable(struct drm_i915_private *i915, > > struct i915_power_well *power_well) > > { > > - bxt_ddi_phy_init(dev_priv, power_well->desc->bxt.phy); > > + bxt_ddi_phy_init(i915, power_well->desc->bxt.phy); > > } > > -static void bxt_dpio_cmn_power_well_disable(struct drm_i915_private *dev_priv, > > +static void bxt_dpio_cmn_power_well_disable(struct drm_i915_private *i915, > > struct i915_power_well *power_well) > > { > > - bxt_ddi_phy_uninit(dev_priv, power_well->desc->bxt.phy); > > + bxt_ddi_phy_uninit(i915, power_well->desc->bxt.phy); > > } > > -static bool bxt_dpio_cmn_power_well_enabled(struct drm_i915_private *dev_priv, > > +static bool bxt_dpio_cmn_power_well_enabled(struct drm_i915_private *i915, > > struct i915_power_well *power_well) > > { > > - return bxt_ddi_phy_is_enabled(dev_priv, power_well->desc->bxt.phy); > > + return bxt_ddi_phy_is_enabled(i915, power_well->desc->bxt.phy); > > } > > -static void bxt_verify_ddi_phy_power_wells(struct drm_i915_private *dev_priv) > > +static void bxt_verify_ddi_phy_power_wells(struct drm_i915_private *i915) > > { > > struct i915_power_well *power_well; > > - power_well = lookup_power_well(dev_priv, BXT_DISP_PW_DPIO_CMN_A); > > + power_well = lookup_power_well(i915, BXT_DISP_PW_DPIO_CMN_A); > > if (power_well->count > 0) > > - bxt_ddi_phy_verify_state(dev_priv, power_well->desc->bxt.phy); > > + bxt_ddi_phy_verify_state(i915, power_well->desc->bxt.phy); > > - power_well = lookup_power_well(dev_priv, VLV_DISP_PW_DPIO_CMN_BC); > > + power_well = lookup_power_well(i915, VLV_DISP_PW_DPIO_CMN_BC); > > if (power_well->count > 0) > > - bxt_ddi_phy_verify_state(dev_priv, power_well->desc->bxt.phy); > > + bxt_ddi_phy_verify_state(i915, power_well->desc->bxt.phy); > > - if (IS_GEMINILAKE(dev_priv)) { > > - power_well = lookup_power_well(dev_priv, > > + if (IS_GEMINILAKE(i915)) { > > + power_well = lookup_power_well(i915, > > GLK_DISP_PW_DPIO_CMN_C); > > if (power_well->count > 0) > > - bxt_ddi_phy_verify_state(dev_priv, > > + bxt_ddi_phy_verify_state(i915, > > power_well->desc->bxt.phy); > > } > > } > > -static bool gen9_dc_off_power_well_enabled(struct drm_i915_private *dev_priv, > > +static bool gen9_dc_off_power_well_enabled(struct drm_i915_private *i915, > > struct i915_power_well *power_well) > > { > > return (I915_READ(DC_STATE_EN) & DC_STATE_EN_UPTO_DC5_DC6_MASK) == 0; > > } > > -static void gen9_assert_dbuf_enabled(struct drm_i915_private *dev_priv) > > +static void gen9_assert_dbuf_enabled(struct drm_i915_private *i915) > > { > > u32 tmp = I915_READ(DBUF_CTL); > > @@ -811,92 +811,92 @@ static void gen9_assert_dbuf_enabled(struct drm_i915_private *dev_priv) > > "Unexpected DBuf power power state (0x%08x)\n", tmp); > > } > > -static void gen9_dc_off_power_well_enable(struct drm_i915_private *dev_priv, > > +static void gen9_dc_off_power_well_enable(struct drm_i915_private *i915, > > struct i915_power_well *power_well) > > { > > struct intel_cdclk_state cdclk_state = {}; > > - gen9_set_dc_state(dev_priv, DC_STATE_DISABLE); > > + gen9_set_dc_state(i915, DC_STATE_DISABLE); > > - dev_priv->display.get_cdclk(dev_priv, &cdclk_state); > > + i915->display.get_cdclk(i915, &cdclk_state); > > /* Can't read out voltage_level so can't use intel_cdclk_changed() */ > > - WARN_ON(intel_cdclk_needs_modeset(&dev_priv->cdclk.hw, &cdclk_state)); > > + WARN_ON(intel_cdclk_needs_modeset(&i915->cdclk.hw, &cdclk_state)); > > - gen9_assert_dbuf_enabled(dev_priv); > > + gen9_assert_dbuf_enabled(i915); > > - if (IS_GEN9_LP(dev_priv)) > > - bxt_verify_ddi_phy_power_wells(dev_priv); > > + if (IS_GEN9_LP(i915)) > > + bxt_verify_ddi_phy_power_wells(i915); > > - if (INTEL_GEN(dev_priv) >= 11) > > + if (INTEL_GEN(i915) >= 11) > > /* > > * DMC retains HW context only for port A, the other combo > > * PHY's HW context for port B is lost after DC transitions, > > * so we need to restore it manually. > > */ > > - intel_combo_phy_init(dev_priv); > > + intel_combo_phy_init(i915); > > } > > -static void gen9_dc_off_power_well_disable(struct drm_i915_private *dev_priv, > > +static void gen9_dc_off_power_well_disable(struct drm_i915_private *i915, > > struct i915_power_well *power_well) > > { > > - if (!dev_priv->csr.dmc_payload) > > + if (!i915->csr.dmc_payload) > > return; > > - if (dev_priv->csr.allowed_dc_mask & DC_STATE_EN_UPTO_DC6) > > - skl_enable_dc6(dev_priv); > > - else if (dev_priv->csr.allowed_dc_mask & DC_STATE_EN_UPTO_DC5) > > - gen9_enable_dc5(dev_priv); > > + if (i915->csr.allowed_dc_mask & DC_STATE_EN_UPTO_DC6) > > + skl_enable_dc6(i915); > > + else if (i915->csr.allowed_dc_mask & DC_STATE_EN_UPTO_DC5) > > + gen9_enable_dc5(i915); > > } > > -static void i9xx_power_well_sync_hw_noop(struct drm_i915_private *dev_priv, > > +static void i9xx_power_well_sync_hw_noop(struct drm_i915_private *i915, > > struct i915_power_well *power_well) > > { > > } > > -static void i9xx_always_on_power_well_noop(struct drm_i915_private *dev_priv, > > +static void i9xx_always_on_power_well_noop(struct drm_i915_private *i915, > > struct i915_power_well *power_well) > > { > > } > > -static bool i9xx_always_on_power_well_enabled(struct drm_i915_private *dev_priv, > > +static bool i9xx_always_on_power_well_enabled(struct drm_i915_private *i915, > > struct i915_power_well *power_well) > > { > > return true; > > } > > -static void i830_pipes_power_well_enable(struct drm_i915_private *dev_priv, > > +static void i830_pipes_power_well_enable(struct drm_i915_private *i915, > > struct i915_power_well *power_well) > > { > > if ((I915_READ(PIPECONF(PIPE_A)) & PIPECONF_ENABLE) == 0) > > - i830_enable_pipe(dev_priv, PIPE_A); > > + i830_enable_pipe(i915, PIPE_A); > > if ((I915_READ(PIPECONF(PIPE_B)) & PIPECONF_ENABLE) == 0) > > - i830_enable_pipe(dev_priv, PIPE_B); > > + i830_enable_pipe(i915, PIPE_B); > > } > > -static void i830_pipes_power_well_disable(struct drm_i915_private *dev_priv, > > +static void i830_pipes_power_well_disable(struct drm_i915_private *i915, > > struct i915_power_well *power_well) > > { > > - i830_disable_pipe(dev_priv, PIPE_B); > > - i830_disable_pipe(dev_priv, PIPE_A); > > + i830_disable_pipe(i915, PIPE_B); > > + i830_disable_pipe(i915, PIPE_A); > > } > > -static bool i830_pipes_power_well_enabled(struct drm_i915_private *dev_priv, > > +static bool i830_pipes_power_well_enabled(struct drm_i915_private *i915, > > struct i915_power_well *power_well) > > { > > return I915_READ(PIPECONF(PIPE_A)) & PIPECONF_ENABLE && > > I915_READ(PIPECONF(PIPE_B)) & PIPECONF_ENABLE; > > } > > -static void i830_pipes_power_well_sync_hw(struct drm_i915_private *dev_priv, > > +static void i830_pipes_power_well_sync_hw(struct drm_i915_private *i915, > > struct i915_power_well *power_well) > > { > > if (power_well->count > 0) > > - i830_pipes_power_well_enable(dev_priv, power_well); > > + i830_pipes_power_well_enable(i915, power_well); > > else > > - i830_pipes_power_well_disable(dev_priv, power_well); > > + i830_pipes_power_well_disable(i915, power_well); > > } > > -static void vlv_set_power_well(struct drm_i915_private *dev_priv, > > +static void vlv_set_power_well(struct drm_i915_private *i915, > > struct i915_power_well *power_well, bool enable) > > { > > int pw_idx = power_well->desc->vlv.idx; > > @@ -908,43 +908,43 @@ static void vlv_set_power_well(struct drm_i915_private *dev_priv, > > state = enable ? PUNIT_PWRGT_PWR_ON(pw_idx) : > > PUNIT_PWRGT_PWR_GATE(pw_idx); > > - vlv_punit_get(dev_priv); > > + vlv_punit_get(i915); > > #define COND \ > > - ((vlv_punit_read(dev_priv, PUNIT_REG_PWRGT_STATUS) & mask) == state) > > + ((vlv_punit_read(i915, PUNIT_REG_PWRGT_STATUS) & mask) == state) > > if (COND) > > goto out; > > - ctrl = vlv_punit_read(dev_priv, PUNIT_REG_PWRGT_CTRL); > > + ctrl = vlv_punit_read(i915, PUNIT_REG_PWRGT_CTRL); > > ctrl &= ~mask; > > ctrl |= state; > > - vlv_punit_write(dev_priv, PUNIT_REG_PWRGT_CTRL, ctrl); > > + vlv_punit_write(i915, PUNIT_REG_PWRGT_CTRL, ctrl); > > if (wait_for(COND, 100)) > > DRM_ERROR("timeout setting power well state %08x (%08x)\n", > > state, > > - vlv_punit_read(dev_priv, PUNIT_REG_PWRGT_CTRL)); > > + vlv_punit_read(i915, PUNIT_REG_PWRGT_CTRL)); > > #undef COND > > out: > > - vlv_punit_put(dev_priv); > > + vlv_punit_put(i915); > > } > > -static void vlv_power_well_enable(struct drm_i915_private *dev_priv, > > +static void vlv_power_well_enable(struct drm_i915_private *i915, > > struct i915_power_well *power_well) > > { > > - vlv_set_power_well(dev_priv, power_well, true); > > + vlv_set_power_well(i915, power_well, true); > > } > > -static void vlv_power_well_disable(struct drm_i915_private *dev_priv, > > +static void vlv_power_well_disable(struct drm_i915_private *i915, > > struct i915_power_well *power_well) > > { > > - vlv_set_power_well(dev_priv, power_well, false); > > + vlv_set_power_well(i915, power_well, false); > > } > > -static bool vlv_power_well_enabled(struct drm_i915_private *dev_priv, > > +static bool vlv_power_well_enabled(struct drm_i915_private *i915, > > struct i915_power_well *power_well) > > { > > int pw_idx = power_well->desc->vlv.idx; > > @@ -956,9 +956,9 @@ static bool vlv_power_well_enabled(struct drm_i915_private *dev_priv, > > mask = PUNIT_PWRGT_MASK(pw_idx); > > ctrl = PUNIT_PWRGT_PWR_ON(pw_idx); > > - vlv_punit_get(dev_priv); > > + vlv_punit_get(i915); > > - state = vlv_punit_read(dev_priv, PUNIT_REG_PWRGT_STATUS) & mask; > > + state = vlv_punit_read(i915, PUNIT_REG_PWRGT_STATUS) & mask; > > /* > > * We only ever set the power-on and power-gate states, anything > > * else is unexpected. > > @@ -972,15 +972,15 @@ static bool vlv_power_well_enabled(struct drm_i915_private *dev_priv, > > * A transient state at this point would mean some unexpected party > > * is poking at the power controls too. > > */ > > - ctrl = vlv_punit_read(dev_priv, PUNIT_REG_PWRGT_CTRL) & mask; > > + ctrl = vlv_punit_read(i915, PUNIT_REG_PWRGT_CTRL) & mask; > > WARN_ON(ctrl != state); > > - vlv_punit_put(dev_priv); > > + vlv_punit_put(i915); > > return enabled; > > } > > -static void vlv_init_display_clock_gating(struct drm_i915_private *dev_priv) > > +static void vlv_init_display_clock_gating(struct drm_i915_private *i915) > > { > > u32 val; > > @@ -1001,13 +1001,13 @@ static void vlv_init_display_clock_gating(struct drm_i915_private *dev_priv) > > I915_WRITE(MI_ARB_VLV, MI_ARB_DISPLAY_TRICKLE_FEED_DISABLE); > > I915_WRITE(CBR1_VLV, 0); > > - WARN_ON(dev_priv->rawclk_freq == 0); > > + WARN_ON(i915->rawclk_freq == 0); > > I915_WRITE(RAWCLK_FREQ_VLV, > > - DIV_ROUND_CLOSEST(dev_priv->rawclk_freq, 1000)); > > + DIV_ROUND_CLOSEST(i915->rawclk_freq, 1000)); > > } > > -static void vlv_display_power_well_init(struct drm_i915_private *dev_priv) > > +static void vlv_display_power_well_init(struct drm_i915_private *i915) > > { > > struct intel_encoder *encoder; > > enum pipe pipe; > > @@ -1020,7 +1020,7 @@ static void vlv_display_power_well_init(struct drm_i915_private *dev_priv) > > * > > * CHV DPLL B/C have some issues if VGA mode is enabled. > > */ > > - for_each_pipe(dev_priv, pipe) { > > + for_each_pipe(i915, pipe) { > > u32 val = I915_READ(DPLL(pipe)); > > val |= DPLL_REF_CLK_ENABLE_VLV | DPLL_VGA_MODE_DIS; > > @@ -1030,71 +1030,71 @@ static void vlv_display_power_well_init(struct drm_i915_private *dev_priv) > > I915_WRITE(DPLL(pipe), val); > > } > > - vlv_init_display_clock_gating(dev_priv); > > + vlv_init_display_clock_gating(i915); > > - spin_lock_irq(&dev_priv->irq_lock); > > - valleyview_enable_display_irqs(dev_priv); > > - spin_unlock_irq(&dev_priv->irq_lock); > > + spin_lock_irq(&i915->irq_lock); > > + valleyview_enable_display_irqs(i915); > > + spin_unlock_irq(&i915->irq_lock); > > /* > > * During driver initialization/resume we can avoid restoring the > > * part of the HW/SW state that will be inited anyway explicitly. > > */ > > - if (dev_priv->power_domains.initializing) > > + if (i915->power_domains.initializing) > > return; > > - intel_hpd_init(dev_priv); > > + intel_hpd_init(i915); > > /* Re-enable the ADPA, if we have one */ > > - for_each_intel_encoder(&dev_priv->drm, encoder) { > > + for_each_intel_encoder(&i915->drm, encoder) { > > if (encoder->type == INTEL_OUTPUT_ANALOG) > > intel_crt_reset(&encoder->base); > > } > > - i915_redisable_vga_power_on(dev_priv); > > + i915_redisable_vga_power_on(i915); > > - intel_pps_unlock_regs_wa(dev_priv); > > + intel_pps_unlock_regs_wa(i915); > > } > > -static void vlv_display_power_well_deinit(struct drm_i915_private *dev_priv) > > +static void vlv_display_power_well_deinit(struct drm_i915_private *i915) > > { > > - spin_lock_irq(&dev_priv->irq_lock); > > - valleyview_disable_display_irqs(dev_priv); > > - spin_unlock_irq(&dev_priv->irq_lock); > > + spin_lock_irq(&i915->irq_lock); > > + valleyview_disable_display_irqs(i915); > > + spin_unlock_irq(&i915->irq_lock); > > /* make sure we're done processing display irqs */ > > - synchronize_irq(dev_priv->drm.irq); > > + synchronize_irq(i915->drm.irq); > > - intel_power_sequencer_reset(dev_priv); > > + intel_power_sequencer_reset(i915); > > /* Prevent us from re-enabling polling on accident in late suspend */ > > - if (!dev_priv->drm.dev->power.is_suspended) > > - intel_hpd_poll_init(dev_priv); > > + if (!i915->drm.dev->power.is_suspended) > > + intel_hpd_poll_init(i915); > > } > > -static void vlv_display_power_well_enable(struct drm_i915_private *dev_priv, > > +static void vlv_display_power_well_enable(struct drm_i915_private *i915, > > struct i915_power_well *power_well) > > { > > - vlv_set_power_well(dev_priv, power_well, true); > > + vlv_set_power_well(i915, power_well, true); > > - vlv_display_power_well_init(dev_priv); > > + vlv_display_power_well_init(i915); > > } > > -static void vlv_display_power_well_disable(struct drm_i915_private *dev_priv, > > +static void vlv_display_power_well_disable(struct drm_i915_private *i915, > > struct i915_power_well *power_well) > > { > > - vlv_display_power_well_deinit(dev_priv); > > + vlv_display_power_well_deinit(i915); > > - vlv_set_power_well(dev_priv, power_well, false); > > + vlv_set_power_well(i915, power_well, false); > > } > > -static void vlv_dpio_cmn_power_well_enable(struct drm_i915_private *dev_priv, > > +static void vlv_dpio_cmn_power_well_enable(struct drm_i915_private *i915, > > struct i915_power_well *power_well) > > { > > /* since ref/cri clock was enabled */ > > udelay(1); /* >10ns for cmnreset, >0ns for sidereset */ > > - vlv_set_power_well(dev_priv, power_well, true); > > + vlv_set_power_well(i915, power_well, true); > > /* > > * From VLV2A0_DP_eDP_DPIO_driver_vbios_notes_10.docx - > > @@ -1110,31 +1110,31 @@ static void vlv_dpio_cmn_power_well_enable(struct drm_i915_private *dev_priv, > > I915_WRITE(DPIO_CTL, I915_READ(DPIO_CTL) | DPIO_CMNRST); > > } > > -static void vlv_dpio_cmn_power_well_disable(struct drm_i915_private *dev_priv, > > +static void vlv_dpio_cmn_power_well_disable(struct drm_i915_private *i915, > > struct i915_power_well *power_well) > > { > > enum pipe pipe; > > - for_each_pipe(dev_priv, pipe) > > - assert_pll_disabled(dev_priv, pipe); > > + for_each_pipe(i915, pipe) > > + assert_pll_disabled(i915, pipe); > > /* Assert common reset */ > > I915_WRITE(DPIO_CTL, I915_READ(DPIO_CTL) & ~DPIO_CMNRST); > > - vlv_set_power_well(dev_priv, power_well, false); > > + vlv_set_power_well(i915, power_well, false); > > } > > #define POWER_DOMAIN_MASK (GENMASK_ULL(POWER_DOMAIN_NUM - 1, 0)) > > #define BITS_SET(val, bits) (((val) & (bits)) == (bits)) > > -static void assert_chv_phy_status(struct drm_i915_private *dev_priv) > > +static void assert_chv_phy_status(struct drm_i915_private *i915) > > { > > struct i915_power_well *cmn_bc = > > - lookup_power_well(dev_priv, VLV_DISP_PW_DPIO_CMN_BC); > > + lookup_power_well(i915, VLV_DISP_PW_DPIO_CMN_BC); > > struct i915_power_well *cmn_d = > > - lookup_power_well(dev_priv, CHV_DISP_PW_DPIO_CMN_D); > > - u32 phy_control = dev_priv->chv_phy_control; > > + lookup_power_well(i915, CHV_DISP_PW_DPIO_CMN_D); > > + u32 phy_control = i915->chv_phy_control; > > u32 phy_status = 0; > > u32 phy_status_mask = 0xffffffff; > > @@ -1145,7 +1145,7 @@ static void assert_chv_phy_status(struct drm_i915_private *dev_priv) > > * reset (ie. the power well has been disabled at > > * least once). > > */ > > - if (!dev_priv->chv_phy_assert[DPIO_PHY0]) > > + if (!i915->chv_phy_assert[DPIO_PHY0]) > > phy_status_mask &= ~(PHY_STATUS_CMN_LDO(DPIO_PHY0, DPIO_CH0) | > > PHY_STATUS_SPLINE_LDO(DPIO_PHY0, DPIO_CH0, 0) | > > PHY_STATUS_SPLINE_LDO(DPIO_PHY0, DPIO_CH0, 1) | > > @@ -1153,12 +1153,12 @@ static void assert_chv_phy_status(struct drm_i915_private *dev_priv) > > PHY_STATUS_SPLINE_LDO(DPIO_PHY0, DPIO_CH1, 0) | > > PHY_STATUS_SPLINE_LDO(DPIO_PHY0, DPIO_CH1, 1)); > > - if (!dev_priv->chv_phy_assert[DPIO_PHY1]) > > + if (!i915->chv_phy_assert[DPIO_PHY1]) > > phy_status_mask &= ~(PHY_STATUS_CMN_LDO(DPIO_PHY1, DPIO_CH0) | > > PHY_STATUS_SPLINE_LDO(DPIO_PHY1, DPIO_CH0, 0) | > > PHY_STATUS_SPLINE_LDO(DPIO_PHY1, DPIO_CH0, 1)); > > - if (cmn_bc->desc->ops->is_enabled(dev_priv, cmn_bc)) { > > + if (cmn_bc->desc->ops->is_enabled(i915, cmn_bc)) { > > phy_status |= PHY_POWERGOOD(DPIO_PHY0); > > /* this assumes override is only used to enable lanes */ > > @@ -1199,7 +1199,7 @@ static void assert_chv_phy_status(struct drm_i915_private *dev_priv) > > phy_status |= PHY_STATUS_SPLINE_LDO(DPIO_PHY0, DPIO_CH1, 1); > > } > > - if (cmn_d->desc->ops->is_enabled(dev_priv, cmn_d)) { > > + if (cmn_d->desc->ops->is_enabled(i915, cmn_d)) { > > phy_status |= PHY_POWERGOOD(DPIO_PHY1); > > /* this assumes override is only used to enable lanes */ > > @@ -1224,19 +1224,19 @@ static void assert_chv_phy_status(struct drm_i915_private *dev_priv) > > * The PHY may be busy with some initial calibration and whatnot, > > * so the power state can take a while to actually change. > > */ > > - if (intel_wait_for_register(&dev_priv->uncore, > > + if (intel_wait_for_register(&i915->uncore, > > DISPLAY_PHY_STATUS, > > phy_status_mask, > > phy_status, > > 10)) > > DRM_ERROR("Unexpected PHY_STATUS 0x%08x, expected 0x%08x (PHY_CONTROL=0x%08x)\n", > > I915_READ(DISPLAY_PHY_STATUS) & phy_status_mask, > > - phy_status, dev_priv->chv_phy_control); > > + phy_status, i915->chv_phy_control); > > } > > #undef BITS_SET > > -static void chv_dpio_cmn_power_well_enable(struct drm_i915_private *dev_priv, > > +static void chv_dpio_cmn_power_well_enable(struct drm_i915_private *i915, > > struct i915_power_well *power_well) > > { > > enum dpio_phy phy; > > @@ -1256,51 +1256,51 @@ static void chv_dpio_cmn_power_well_enable(struct drm_i915_private *dev_priv, > > /* since ref/cri clock was enabled */ > > udelay(1); /* >10ns for cmnreset, >0ns for sidereset */ > > - vlv_set_power_well(dev_priv, power_well, true); > > + vlv_set_power_well(i915, power_well, true); > > /* Poll for phypwrgood signal */ > > - if (intel_wait_for_register(&dev_priv->uncore, > > + if (intel_wait_for_register(&i915->uncore, > > DISPLAY_PHY_STATUS, > > PHY_POWERGOOD(phy), > > PHY_POWERGOOD(phy), > > 1)) > > DRM_ERROR("Display PHY %d is not power up\n", phy); > > - vlv_dpio_get(dev_priv); > > + vlv_dpio_get(i915); > > /* Enable dynamic power down */ > > - tmp = vlv_dpio_read(dev_priv, pipe, CHV_CMN_DW28); > > + tmp = vlv_dpio_read(i915, pipe, CHV_CMN_DW28); > > tmp |= DPIO_DYNPWRDOWNEN_CH0 | DPIO_CL1POWERDOWNEN | > > DPIO_SUS_CLK_CONFIG_GATE_CLKREQ; > > - vlv_dpio_write(dev_priv, pipe, CHV_CMN_DW28, tmp); > > + vlv_dpio_write(i915, pipe, CHV_CMN_DW28, tmp); > > if (power_well->desc->id == VLV_DISP_PW_DPIO_CMN_BC) { > > - tmp = vlv_dpio_read(dev_priv, pipe, _CHV_CMN_DW6_CH1); > > + tmp = vlv_dpio_read(i915, pipe, _CHV_CMN_DW6_CH1); > > tmp |= DPIO_DYNPWRDOWNEN_CH1; > > - vlv_dpio_write(dev_priv, pipe, _CHV_CMN_DW6_CH1, tmp); > > + vlv_dpio_write(i915, pipe, _CHV_CMN_DW6_CH1, tmp); > > } else { > > /* > > * Force the non-existing CL2 off. BXT does this > > * too, so maybe it saves some power even though > > * CL2 doesn't exist? > > */ > > - tmp = vlv_dpio_read(dev_priv, pipe, CHV_CMN_DW30); > > + tmp = vlv_dpio_read(i915, pipe, CHV_CMN_DW30); > > tmp |= DPIO_CL2_LDOFUSE_PWRENB; > > - vlv_dpio_write(dev_priv, pipe, CHV_CMN_DW30, tmp); > > + vlv_dpio_write(i915, pipe, CHV_CMN_DW30, tmp); > > } > > - vlv_dpio_put(dev_priv); > > + vlv_dpio_put(i915); > > - dev_priv->chv_phy_control |= PHY_COM_LANE_RESET_DEASSERT(phy); > > - I915_WRITE(DISPLAY_PHY_CONTROL, dev_priv->chv_phy_control); > > + i915->chv_phy_control |= PHY_COM_LANE_RESET_DEASSERT(phy); > > + I915_WRITE(DISPLAY_PHY_CONTROL, i915->chv_phy_control); > > DRM_DEBUG_KMS("Enabled DPIO PHY%d (PHY_CONTROL=0x%08x)\n", > > - phy, dev_priv->chv_phy_control); > > + phy, i915->chv_phy_control); > > - assert_chv_phy_status(dev_priv); > > + assert_chv_phy_status(i915); > > } > > -static void chv_dpio_cmn_power_well_disable(struct drm_i915_private *dev_priv, > > +static void chv_dpio_cmn_power_well_disable(struct drm_i915_private *i915, > > struct i915_power_well *power_well) > > { > > enum dpio_phy phy; > > @@ -1310,28 +1310,28 @@ static void chv_dpio_cmn_power_well_disable(struct drm_i915_private *dev_priv, > > if (power_well->desc->id == VLV_DISP_PW_DPIO_CMN_BC) { > > phy = DPIO_PHY0; > > - assert_pll_disabled(dev_priv, PIPE_A); > > - assert_pll_disabled(dev_priv, PIPE_B); > > + assert_pll_disabled(i915, PIPE_A); > > + assert_pll_disabled(i915, PIPE_B); > > } else { > > phy = DPIO_PHY1; > > - assert_pll_disabled(dev_priv, PIPE_C); > > + assert_pll_disabled(i915, PIPE_C); > > } > > - dev_priv->chv_phy_control &= ~PHY_COM_LANE_RESET_DEASSERT(phy); > > - I915_WRITE(DISPLAY_PHY_CONTROL, dev_priv->chv_phy_control); > > + i915->chv_phy_control &= ~PHY_COM_LANE_RESET_DEASSERT(phy); > > + I915_WRITE(DISPLAY_PHY_CONTROL, i915->chv_phy_control); > > - vlv_set_power_well(dev_priv, power_well, false); > > + vlv_set_power_well(i915, power_well, false); > > DRM_DEBUG_KMS("Disabled DPIO PHY%d (PHY_CONTROL=0x%08x)\n", > > - phy, dev_priv->chv_phy_control); > > + phy, i915->chv_phy_control); > > /* PHY is fully reset now, so we can enable the PHY state asserts */ > > - dev_priv->chv_phy_assert[phy] = true; > > + i915->chv_phy_assert[phy] = true; > > - assert_chv_phy_status(dev_priv); > > + assert_chv_phy_status(i915); > > } > > -static void assert_chv_phy_powergate(struct drm_i915_private *dev_priv, enum dpio_phy phy, > > +static void assert_chv_phy_powergate(struct drm_i915_private *i915, enum dpio_phy phy, > > enum dpio_channel ch, bool override, unsigned int mask) > > { > > enum pipe pipe = phy == DPIO_PHY0 ? PIPE_A : PIPE_C; > > @@ -1344,7 +1344,7 @@ static void assert_chv_phy_powergate(struct drm_i915_private *dev_priv, enum dpi > > * reset (ie. the power well has been disabled at > > * least once). > > */ > > - if (!dev_priv->chv_phy_assert[phy]) > > + if (!i915->chv_phy_assert[phy]) > > return; > > if (ch == DPIO_CH0) > > @@ -1352,9 +1352,9 @@ static void assert_chv_phy_powergate(struct drm_i915_private *dev_priv, enum dpi > > else > > reg = _CHV_CMN_DW6_CH1; > > - vlv_dpio_get(dev_priv); > > - val = vlv_dpio_read(dev_priv, pipe, reg); > > - vlv_dpio_put(dev_priv); > > + vlv_dpio_get(i915); > > + val = vlv_dpio_read(i915, pipe, reg); > > + vlv_dpio_put(i915); > > /* > > * This assumes !override is only used when the port is disabled. > > @@ -1393,30 +1393,30 @@ static void assert_chv_phy_powergate(struct drm_i915_private *dev_priv, enum dpi > > reg, val); > > } > > -bool chv_phy_powergate_ch(struct drm_i915_private *dev_priv, enum dpio_phy phy, > > +bool chv_phy_powergate_ch(struct drm_i915_private *i915, enum dpio_phy phy, > > enum dpio_channel ch, bool override) > > { > > - struct i915_power_domains *power_domains = &dev_priv->power_domains; > > + struct i915_power_domains *power_domains = &i915->power_domains; > > bool was_override; > > mutex_lock(&power_domains->lock); > > - was_override = dev_priv->chv_phy_control & PHY_CH_POWER_DOWN_OVRD_EN(phy, ch); > > + was_override = i915->chv_phy_control & PHY_CH_POWER_DOWN_OVRD_EN(phy, ch); > > if (override == was_override) > > goto out; > > if (override) > > - dev_priv->chv_phy_control |= PHY_CH_POWER_DOWN_OVRD_EN(phy, ch); > > + i915->chv_phy_control |= PHY_CH_POWER_DOWN_OVRD_EN(phy, ch); > > else > > - dev_priv->chv_phy_control &= ~PHY_CH_POWER_DOWN_OVRD_EN(phy, ch); > > + i915->chv_phy_control &= ~PHY_CH_POWER_DOWN_OVRD_EN(phy, ch); > > - I915_WRITE(DISPLAY_PHY_CONTROL, dev_priv->chv_phy_control); > > + I915_WRITE(DISPLAY_PHY_CONTROL, i915->chv_phy_control); > > DRM_DEBUG_KMS("Power gating DPIO PHY%d CH%d (DPIO_PHY_CONTROL=0x%08x)\n", > > - phy, ch, dev_priv->chv_phy_control); > > + phy, ch, i915->chv_phy_control); > > - assert_chv_phy_status(dev_priv); > > + assert_chv_phy_status(i915); > > out: > > mutex_unlock(&power_domains->lock); > > @@ -1427,43 +1427,43 @@ bool chv_phy_powergate_ch(struct drm_i915_private *dev_priv, enum dpio_phy phy, > > void chv_phy_powergate_lanes(struct intel_encoder *encoder, > > bool override, unsigned int mask) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > - struct i915_power_domains *power_domains = &dev_priv->power_domains; > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > + struct i915_power_domains *power_domains = &i915->power_domains; > > enum dpio_phy phy = vlv_dport_to_phy(enc_to_dig_port(&encoder->base)); > > enum dpio_channel ch = vlv_dport_to_channel(enc_to_dig_port(&encoder->base)); > > mutex_lock(&power_domains->lock); > > - dev_priv->chv_phy_control &= ~PHY_CH_POWER_DOWN_OVRD(0xf, phy, ch); > > - dev_priv->chv_phy_control |= PHY_CH_POWER_DOWN_OVRD(mask, phy, ch); > > + i915->chv_phy_control &= ~PHY_CH_POWER_DOWN_OVRD(0xf, phy, ch); > > + i915->chv_phy_control |= PHY_CH_POWER_DOWN_OVRD(mask, phy, ch); > > if (override) > > - dev_priv->chv_phy_control |= PHY_CH_POWER_DOWN_OVRD_EN(phy, ch); > > + i915->chv_phy_control |= PHY_CH_POWER_DOWN_OVRD_EN(phy, ch); > > else > > - dev_priv->chv_phy_control &= ~PHY_CH_POWER_DOWN_OVRD_EN(phy, ch); > > + i915->chv_phy_control &= ~PHY_CH_POWER_DOWN_OVRD_EN(phy, ch); > > - I915_WRITE(DISPLAY_PHY_CONTROL, dev_priv->chv_phy_control); > > + I915_WRITE(DISPLAY_PHY_CONTROL, i915->chv_phy_control); > > DRM_DEBUG_KMS("Power gating DPIO PHY%d CH%d lanes 0x%x (PHY_CONTROL=0x%08x)\n", > > - phy, ch, mask, dev_priv->chv_phy_control); > > + phy, ch, mask, i915->chv_phy_control); > > - assert_chv_phy_status(dev_priv); > > + assert_chv_phy_status(i915); > > - assert_chv_phy_powergate(dev_priv, phy, ch, override, mask); > > + assert_chv_phy_powergate(i915, phy, ch, override, mask); > > mutex_unlock(&power_domains->lock); > > } > > -static bool chv_pipe_power_well_enabled(struct drm_i915_private *dev_priv, > > +static bool chv_pipe_power_well_enabled(struct drm_i915_private *i915, > > struct i915_power_well *power_well) > > { > > enum pipe pipe = PIPE_A; > > bool enabled; > > u32 state, ctrl; > > - vlv_punit_get(dev_priv); > > + vlv_punit_get(i915); > > - state = vlv_punit_read(dev_priv, PUNIT_REG_DSPSSPM) & DP_SSS_MASK(pipe); > > + state = vlv_punit_read(i915, PUNIT_REG_DSPSSPM) & DP_SSS_MASK(pipe); > > /* > > * We only ever set the power-on and power-gate states, anything > > * else is unexpected. > > @@ -1475,15 +1475,15 @@ static bool chv_pipe_power_well_enabled(struct drm_i915_private *dev_priv, > > * A transient state at this point would mean some unexpected party > > * is poking at the power controls too. > > */ > > - ctrl = vlv_punit_read(dev_priv, PUNIT_REG_DSPSSPM) & DP_SSC_MASK(pipe); > > + ctrl = vlv_punit_read(i915, PUNIT_REG_DSPSSPM) & DP_SSC_MASK(pipe); > > WARN_ON(ctrl << 16 != state); > > - vlv_punit_put(dev_priv); > > + vlv_punit_put(i915); > > return enabled; > > } > > -static void chv_set_pipe_power_well(struct drm_i915_private *dev_priv, > > +static void chv_set_pipe_power_well(struct drm_i915_private *i915, > > struct i915_power_well *power_well, > > bool enable) > > { > > @@ -1493,44 +1493,44 @@ static void chv_set_pipe_power_well(struct drm_i915_private *dev_priv, > > state = enable ? DP_SSS_PWR_ON(pipe) : DP_SSS_PWR_GATE(pipe); > > - vlv_punit_get(dev_priv); > > + vlv_punit_get(i915); > > #define COND \ > > - ((vlv_punit_read(dev_priv, PUNIT_REG_DSPSSPM) & DP_SSS_MASK(pipe)) == state) > > + ((vlv_punit_read(i915, PUNIT_REG_DSPSSPM) & DP_SSS_MASK(pipe)) == state) > > if (COND) > > goto out; > > - ctrl = vlv_punit_read(dev_priv, PUNIT_REG_DSPSSPM); > > + ctrl = vlv_punit_read(i915, PUNIT_REG_DSPSSPM); > > ctrl &= ~DP_SSC_MASK(pipe); > > ctrl |= enable ? DP_SSC_PWR_ON(pipe) : DP_SSC_PWR_GATE(pipe); > > - vlv_punit_write(dev_priv, PUNIT_REG_DSPSSPM, ctrl); > > + vlv_punit_write(i915, PUNIT_REG_DSPSSPM, ctrl); > > if (wait_for(COND, 100)) > > DRM_ERROR("timeout setting power well state %08x (%08x)\n", > > state, > > - vlv_punit_read(dev_priv, PUNIT_REG_DSPSSPM)); > > + vlv_punit_read(i915, PUNIT_REG_DSPSSPM)); > > #undef COND > > out: > > - vlv_punit_put(dev_priv); > > + vlv_punit_put(i915); > > } > > -static void chv_pipe_power_well_enable(struct drm_i915_private *dev_priv, > > +static void chv_pipe_power_well_enable(struct drm_i915_private *i915, > > struct i915_power_well *power_well) > > { > > - chv_set_pipe_power_well(dev_priv, power_well, true); > > + chv_set_pipe_power_well(i915, power_well, true); > > - vlv_display_power_well_init(dev_priv); > > + vlv_display_power_well_init(i915); > > } > > -static void chv_pipe_power_well_disable(struct drm_i915_private *dev_priv, > > +static void chv_pipe_power_well_disable(struct drm_i915_private *i915, > > struct i915_power_well *power_well) > > { > > - vlv_display_power_well_deinit(dev_priv); > > + vlv_display_power_well_deinit(i915); > > - chv_set_pipe_power_well(dev_priv, power_well, false); > > + chv_set_pipe_power_well(i915, power_well, false); > > } > > static u64 __async_put_domains_mask(struct i915_power_domains *power_domains) > > @@ -1627,10 +1627,10 @@ async_put_domains_clear_domain(struct i915_power_domains *power_domains, > > } > > static bool > > -intel_display_power_grab_async_put_ref(struct drm_i915_private *dev_priv, > > +intel_display_power_grab_async_put_ref(struct drm_i915_private *i915, > > enum intel_display_power_domain domain) > > { > > - struct i915_power_domains *power_domains = &dev_priv->power_domains; > > + struct i915_power_domains *power_domains = &i915->power_domains; > > bool ret = false; > > if (!(async_put_domains_mask(power_domains) & BIT_ULL(domain))) > > @@ -1644,7 +1644,7 @@ intel_display_power_grab_async_put_ref(struct drm_i915_private *dev_priv, > > goto out_verify; > > cancel_delayed_work(&power_domains->async_put_work); > > - intel_runtime_pm_put_raw(dev_priv, > > + intel_runtime_pm_put_raw(i915, > > fetch_and_zero(&power_domains->async_put_wakeref)); > > out_verify: > > verify_async_put_domains_state(power_domains); > > @@ -1653,24 +1653,24 @@ intel_display_power_grab_async_put_ref(struct drm_i915_private *dev_priv, > > } > > static void > > -__intel_display_power_get_domain(struct drm_i915_private *dev_priv, > > +__intel_display_power_get_domain(struct drm_i915_private *i915, > > enum intel_display_power_domain domain) > > { > > - struct i915_power_domains *power_domains = &dev_priv->power_domains; > > + struct i915_power_domains *power_domains = &i915->power_domains; > > struct i915_power_well *power_well; > > - if (intel_display_power_grab_async_put_ref(dev_priv, domain)) > > + if (intel_display_power_grab_async_put_ref(i915, domain)) > > return; > > - for_each_power_domain_well(dev_priv, power_well, BIT_ULL(domain)) > > - intel_power_well_get(dev_priv, power_well); > > + for_each_power_domain_well(i915, power_well, BIT_ULL(domain)) > > + intel_power_well_get(i915, power_well); > > power_domains->domain_use_count[domain]++; > > } > > /** > > * intel_display_power_get - grab a power domain reference > > - * @dev_priv: i915 device instance > > + * @i915: i915 device instance > > * @domain: power domain to reference > > * > > * This function grabs a power domain reference for @domain and ensures that the > > @@ -1680,14 +1680,14 @@ __intel_display_power_get_domain(struct drm_i915_private *dev_priv, > > * Any power domain reference obtained by this function must have a symmetric > > * call to intel_display_power_put() to release the reference again. > > */ > > -intel_wakeref_t intel_display_power_get(struct drm_i915_private *dev_priv, > > +intel_wakeref_t intel_display_power_get(struct drm_i915_private *i915, > > enum intel_display_power_domain domain) > > { > > - struct i915_power_domains *power_domains = &dev_priv->power_domains; > > - intel_wakeref_t wakeref = intel_runtime_pm_get(dev_priv); > > + struct i915_power_domains *power_domains = &i915->power_domains; > > + intel_wakeref_t wakeref = intel_runtime_pm_get(i915); > > mutex_lock(&power_domains->lock); > > - __intel_display_power_get_domain(dev_priv, domain); > > + __intel_display_power_get_domain(i915, domain); > > mutex_unlock(&power_domains->lock); > > return wakeref; > > @@ -1695,7 +1695,7 @@ intel_wakeref_t intel_display_power_get(struct drm_i915_private *dev_priv, > > /** > > * intel_display_power_get_if_enabled - grab a reference for an enabled display power domain > > - * @dev_priv: i915 device instance > > + * @i915: i915 device instance > > * @domain: power domain to reference > > * > > * This function grabs a power domain reference for @domain and ensures that the > > @@ -1706,21 +1706,21 @@ intel_wakeref_t intel_display_power_get(struct drm_i915_private *dev_priv, > > * call to intel_display_power_put() to release the reference again. > > */ > > intel_wakeref_t > > -intel_display_power_get_if_enabled(struct drm_i915_private *dev_priv, > > +intel_display_power_get_if_enabled(struct drm_i915_private *i915, > > enum intel_display_power_domain domain) > > { > > - struct i915_power_domains *power_domains = &dev_priv->power_domains; > > + struct i915_power_domains *power_domains = &i915->power_domains; > > intel_wakeref_t wakeref; > > bool is_enabled; > > - wakeref = intel_runtime_pm_get_if_in_use(dev_priv); > > + wakeref = intel_runtime_pm_get_if_in_use(i915); > > if (!wakeref) > > return false; > > mutex_lock(&power_domains->lock); > > - if (__intel_display_power_is_enabled(dev_priv, domain)) { > > - __intel_display_power_get_domain(dev_priv, domain); > > + if (__intel_display_power_is_enabled(i915, domain)) { > > + __intel_display_power_get_domain(i915, domain); > > is_enabled = true; > > } else { > > is_enabled = false; > > @@ -1729,7 +1729,7 @@ intel_display_power_get_if_enabled(struct drm_i915_private *dev_priv, > > mutex_unlock(&power_domains->lock); > > if (!is_enabled) { > > - intel_runtime_pm_put(dev_priv, wakeref); > > + intel_runtime_pm_put(i915, wakeref); > > wakeref = 0; > > } > > @@ -1737,14 +1737,14 @@ intel_display_power_get_if_enabled(struct drm_i915_private *dev_priv, > > } > > static void > > -__intel_display_power_put_domain(struct drm_i915_private *dev_priv, > > +__intel_display_power_put_domain(struct drm_i915_private *i915, > > enum intel_display_power_domain domain) > > { > > struct i915_power_domains *power_domains; > > struct i915_power_well *power_well; > > const char *name = intel_display_power_domain_str(domain); > > - power_domains = &dev_priv->power_domains; > > + power_domains = &i915->power_domains; > > WARN(!power_domains->domain_use_count[domain], > > "Use count on domain %s is already zero\n", > > @@ -1755,23 +1755,23 @@ __intel_display_power_put_domain(struct drm_i915_private *dev_priv, > > power_domains->domain_use_count[domain]--; > > - for_each_power_domain_well_reverse(dev_priv, power_well, BIT_ULL(domain)) > > - intel_power_well_put(dev_priv, power_well); > > + for_each_power_domain_well_reverse(i915, power_well, BIT_ULL(domain)) > > + intel_power_well_put(i915, power_well); > > } > > -static void __intel_display_power_put(struct drm_i915_private *dev_priv, > > +static void __intel_display_power_put(struct drm_i915_private *i915, > > enum intel_display_power_domain domain) > > { > > - struct i915_power_domains *power_domains = &dev_priv->power_domains; > > + struct i915_power_domains *power_domains = &i915->power_domains; > > mutex_lock(&power_domains->lock); > > - __intel_display_power_put_domain(dev_priv, domain); > > + __intel_display_power_put_domain(i915, domain); > > mutex_unlock(&power_domains->lock); > > } > > /** > > * intel_display_power_put_unchecked - release an unchecked power domain reference > > - * @dev_priv: i915 device instance > > + * @i915: i915 device instance > > * @domain: power domain to reference > > * > > * This function drops the power domain reference obtained by > > @@ -1782,11 +1782,11 @@ static void __intel_display_power_put(struct drm_i915_private *dev_priv, > > * new code, as the correctness of its use cannot be checked. Always use > > * intel_display_power_put() instead. > > */ > > -void intel_display_power_put_unchecked(struct drm_i915_private *dev_priv, > > +void intel_display_power_put_unchecked(struct drm_i915_private *i915, > > enum intel_display_power_domain domain) > > { > > - __intel_display_power_put(dev_priv, domain); > > - intel_runtime_pm_put_unchecked(dev_priv); > > + __intel_display_power_put(i915, domain); > > + intel_runtime_pm_put_unchecked(i915); > > } > > static void > > @@ -1803,7 +1803,7 @@ queue_async_put_domains_work(struct i915_power_domains *power_domains, > > static void > > release_async_put_domains(struct i915_power_domains *power_domains, u64 mask) > > { > > - struct drm_i915_private *dev_priv = > > + struct drm_i915_private *i915 = > > container_of(power_domains, struct drm_i915_private, > > power_domains); > > enum intel_display_power_domain domain; > > @@ -1814,26 +1814,26 @@ release_async_put_domains(struct i915_power_domains *power_domains, u64 mask) > > * wakeref to make the state checker happy about the HW access during > > * power well disabling. > > */ > > - assert_rpm_raw_wakeref_held(dev_priv); > > - wakeref = intel_runtime_pm_get(dev_priv); > > + assert_rpm_raw_wakeref_held(i915); > > + wakeref = intel_runtime_pm_get(i915); > > for_each_power_domain(domain, mask) { > > /* Clear before put, so put's sanity check is happy. */ > > async_put_domains_clear_domain(power_domains, domain); > > - __intel_display_power_put_domain(dev_priv, domain); > > + __intel_display_power_put_domain(i915, domain); > > } > > - intel_runtime_pm_put(dev_priv, wakeref); > > + intel_runtime_pm_put(i915, wakeref); > > } > > static void > > intel_display_power_put_async_work(struct work_struct *work) > > { > > - struct drm_i915_private *dev_priv = > > + struct drm_i915_private *i915 = > > container_of(work, struct drm_i915_private, > > power_domains.async_put_work.work); > > - struct i915_power_domains *power_domains = &dev_priv->power_domains; > > - intel_wakeref_t new_work_wakeref = intel_runtime_pm_get_raw(dev_priv); > > + struct i915_power_domains *power_domains = &i915->power_domains; > > + intel_wakeref_t new_work_wakeref = intel_runtime_pm_get_raw(i915); > > intel_wakeref_t old_work_wakeref = 0; > > mutex_lock(&power_domains->lock); > > @@ -1863,9 +1863,9 @@ intel_display_power_put_async_work(struct work_struct *work) > > mutex_unlock(&power_domains->lock); > > if (old_work_wakeref) > > - intel_runtime_pm_put_raw(dev_priv, old_work_wakeref); > > + intel_runtime_pm_put_raw(i915, old_work_wakeref); > > if (new_work_wakeref) > > - intel_runtime_pm_put_raw(dev_priv, new_work_wakeref); > > + intel_runtime_pm_put_raw(i915, new_work_wakeref); > > } > > /** > > @@ -1974,7 +1974,7 @@ intel_display_power_flush_work_sync(struct drm_i915_private *i915) > > #if IS_ENABLED(CONFIG_DRM_I915_DEBUG_RUNTIME_PM) > > /** > > * intel_display_power_put - release a power domain reference > > - * @dev_priv: i915 device instance > > + * @i915: i915 device instance > > * @domain: power domain to reference > > * @wakeref: wakeref acquired for the reference that is being released > > * > > @@ -1982,12 +1982,12 @@ intel_display_power_flush_work_sync(struct drm_i915_private *i915) > > * intel_display_power_get() and might power down the corresponding hardware > > * block right away if this is the last reference. > > */ > > -void intel_display_power_put(struct drm_i915_private *dev_priv, > > +void intel_display_power_put(struct drm_i915_private *i915, > > enum intel_display_power_domain domain, > > intel_wakeref_t wakeref) > > { > > - __intel_display_power_put(dev_priv, domain); > > - intel_runtime_pm_put(dev_priv, wakeref); > > + __intel_display_power_put(i915, domain); > > + intel_runtime_pm_put(i915, wakeref); > > } > > #endif > > @@ -2655,14 +2655,14 @@ static const struct i915_power_well_desc chv_power_wells[] = { > > }, > > }; > > -bool intel_display_power_well_is_enabled(struct drm_i915_private *dev_priv, > > +bool intel_display_power_well_is_enabled(struct drm_i915_private *i915, > > enum i915_power_well_id power_well_id) > > { > > struct i915_power_well *power_well; > > bool ret; > > - power_well = lookup_power_well(dev_priv, power_well_id); > > - ret = power_well->desc->ops->is_enabled(dev_priv, power_well); > > + power_well = lookup_power_well(i915, power_well_id); > > + ret = power_well->desc->ops->is_enabled(i915, power_well); > > return ret; > > } > > @@ -3352,7 +3352,7 @@ static const struct i915_power_well_desc icl_power_wells[] = { > > }; > > static int > > -sanitize_disable_power_well_option(const struct drm_i915_private *dev_priv, > > +sanitize_disable_power_well_option(const struct drm_i915_private *i915, > > int disable_power_well) > > { > > if (disable_power_well >= 0) > > @@ -3361,14 +3361,14 @@ sanitize_disable_power_well_option(const struct drm_i915_private *dev_priv, > > return 1; > > } > > -static u32 get_allowed_dc_mask(const struct drm_i915_private *dev_priv, > > +static u32 get_allowed_dc_mask(const struct drm_i915_private *i915, > > int enable_dc) > > { > > u32 mask; > > int requested_dc; > > int max_dc; > > - if (INTEL_GEN(dev_priv) >= 11) { > > + if (INTEL_GEN(i915) >= 11) { > > max_dc = 2; > > /* > > * DC9 has a separate HW flow from the rest of the DC states, > > @@ -3376,10 +3376,10 @@ static u32 get_allowed_dc_mask(const struct drm_i915_private *dev_priv, > > * suspend/resume, so allow it unconditionally. > > */ > > mask = DC_STATE_EN_DC9; > > - } else if (IS_GEN(dev_priv, 10) || IS_GEN9_BC(dev_priv)) { > > + } else if (IS_GEN(i915, 10) || IS_GEN9_BC(i915)) { > > max_dc = 2; > > mask = 0; > > - } else if (IS_GEN9_LP(dev_priv)) { > > + } else if (IS_GEN9_LP(i915)) { > > max_dc = 1; > > mask = DC_STATE_EN_DC9; > > } else { > > @@ -3451,21 +3451,21 @@ __set_power_wells(struct i915_power_domains *power_domains, > > /** > > * intel_power_domains_init - initializes the power domain structures > > - * @dev_priv: i915 device instance > > + * @i915: i915 device instance > > * > > - * Initializes the power domain structures for @dev_priv depending upon the > > + * Initializes the power domain structures for @i915 depending upon the > > * supported platform. > > */ > > -int intel_power_domains_init(struct drm_i915_private *dev_priv) > > +int intel_power_domains_init(struct drm_i915_private *i915) > > { > > - struct i915_power_domains *power_domains = &dev_priv->power_domains; > > + struct i915_power_domains *power_domains = &i915->power_domains; > > int err; > > i915_modparams.disable_power_well = > > - sanitize_disable_power_well_option(dev_priv, > > + sanitize_disable_power_well_option(i915, > > i915_modparams.disable_power_well); > > - dev_priv->csr.allowed_dc_mask = > > - get_allowed_dc_mask(dev_priv, i915_modparams.enable_dc); > > + i915->csr.allowed_dc_mask = > > + get_allowed_dc_mask(i915, i915_modparams.enable_dc); > > BUILD_BUG_ON(POWER_DOMAIN_NUM > 64); > > @@ -3478,9 +3478,9 @@ int intel_power_domains_init(struct drm_i915_private *dev_priv) > > * The enabling order will be from lower to higher indexed wells, > > * the disabling order is reversed. > > */ > > - if (IS_GEN(dev_priv, 11)) { > > + if (IS_GEN(i915, 11)) { > > err = set_power_wells(power_domains, icl_power_wells); > > - } else if (IS_CANNONLAKE(dev_priv)) { > > + } else if (IS_CANNONLAKE(i915)) { > > err = set_power_wells(power_domains, cnl_power_wells); > > /* > > @@ -3489,23 +3489,23 @@ int intel_power_domains_init(struct drm_i915_private *dev_priv) > > * timeouts, lets remove them from the list > > * for the SKUs without port F. > > */ > > - if (!IS_CNL_WITH_PORT_F(dev_priv)) > > + if (!IS_CNL_WITH_PORT_F(i915)) > > power_domains->power_well_count -= 2; > > - } else if (IS_GEMINILAKE(dev_priv)) { > > + } else if (IS_GEMINILAKE(i915)) { > > err = set_power_wells(power_domains, glk_power_wells); > > - } else if (IS_BROXTON(dev_priv)) { > > + } else if (IS_BROXTON(i915)) { > > err = set_power_wells(power_domains, bxt_power_wells); > > - } else if (IS_GEN9_BC(dev_priv)) { > > + } else if (IS_GEN9_BC(i915)) { > > err = set_power_wells(power_domains, skl_power_wells); > > - } else if (IS_CHERRYVIEW(dev_priv)) { > > + } else if (IS_CHERRYVIEW(i915)) { > > err = set_power_wells(power_domains, chv_power_wells); > > - } else if (IS_BROADWELL(dev_priv)) { > > + } else if (IS_BROADWELL(i915)) { > > err = set_power_wells(power_domains, bdw_power_wells); > > - } else if (IS_HASWELL(dev_priv)) { > > + } else if (IS_HASWELL(i915)) { > > err = set_power_wells(power_domains, hsw_power_wells); > > - } else if (IS_VALLEYVIEW(dev_priv)) { > > + } else if (IS_VALLEYVIEW(i915)) { > > err = set_power_wells(power_domains, vlv_power_wells); > > - } else if (IS_I830(dev_priv)) { > > + } else if (IS_I830(i915)) { > > err = set_power_wells(power_domains, i830_power_wells); > > } else { > > err = set_power_wells(power_domains, i9xx_always_on_power_well); > > @@ -3516,31 +3516,31 @@ int intel_power_domains_init(struct drm_i915_private *dev_priv) > > /** > > * intel_power_domains_cleanup - clean up power domains resources > > - * @dev_priv: i915 device instance > > + * @i915: i915 device instance > > * > > * Release any resources acquired by intel_power_domains_init() > > */ > > -void intel_power_domains_cleanup(struct drm_i915_private *dev_priv) > > +void intel_power_domains_cleanup(struct drm_i915_private *i915) > > { > > - kfree(dev_priv->power_domains.power_wells); > > + kfree(i915->power_domains.power_wells); > > } > > -static void intel_power_domains_sync_hw(struct drm_i915_private *dev_priv) > > +static void intel_power_domains_sync_hw(struct drm_i915_private *i915) > > { > > - struct i915_power_domains *power_domains = &dev_priv->power_domains; > > + struct i915_power_domains *power_domains = &i915->power_domains; > > struct i915_power_well *power_well; > > mutex_lock(&power_domains->lock); > > - for_each_power_well(dev_priv, power_well) { > > - power_well->desc->ops->sync_hw(dev_priv, power_well); > > + for_each_power_well(i915, power_well) { > > + power_well->desc->ops->sync_hw(i915, power_well); > > power_well->hw_enabled = > > - power_well->desc->ops->is_enabled(dev_priv, power_well); > > + power_well->desc->ops->is_enabled(i915, power_well); > > } > > mutex_unlock(&power_domains->lock); > > } > > static inline > > -bool intel_dbuf_slice_set(struct drm_i915_private *dev_priv, > > +bool intel_dbuf_slice_set(struct drm_i915_private *i915, > > i915_reg_t reg, bool enable) > > { > > u32 val, status; > > @@ -3560,30 +3560,30 @@ bool intel_dbuf_slice_set(struct drm_i915_private *dev_priv, > > return true; > > } > > -static void gen9_dbuf_enable(struct drm_i915_private *dev_priv) > > +static void gen9_dbuf_enable(struct drm_i915_private *i915) > > { > > - intel_dbuf_slice_set(dev_priv, DBUF_CTL, true); > > + intel_dbuf_slice_set(i915, DBUF_CTL, true); > > } > > -static void gen9_dbuf_disable(struct drm_i915_private *dev_priv) > > +static void gen9_dbuf_disable(struct drm_i915_private *i915) > > { > > - intel_dbuf_slice_set(dev_priv, DBUF_CTL, false); > > + intel_dbuf_slice_set(i915, DBUF_CTL, false); > > } > > -static u8 intel_dbuf_max_slices(struct drm_i915_private *dev_priv) > > +static u8 intel_dbuf_max_slices(struct drm_i915_private *i915) > > { > > - if (INTEL_GEN(dev_priv) < 11) > > + if (INTEL_GEN(i915) < 11) > > return 1; > > return 2; > > } > > -void icl_dbuf_slices_update(struct drm_i915_private *dev_priv, > > +void icl_dbuf_slices_update(struct drm_i915_private *i915, > > u8 req_slices) > > { > > - const u8 hw_enabled_slices = dev_priv->wm.skl_hw.ddb.enabled_slices; > > + const u8 hw_enabled_slices = i915->wm.skl_hw.ddb.enabled_slices; > > bool ret; > > - if (req_slices > intel_dbuf_max_slices(dev_priv)) { > > + if (req_slices > intel_dbuf_max_slices(i915)) { > > DRM_ERROR("Invalid number of dbuf slices requested\n"); > > return; > > } > > @@ -3592,15 +3592,15 @@ void icl_dbuf_slices_update(struct drm_i915_private *dev_priv, > > return; > > if (req_slices > hw_enabled_slices) > > - ret = intel_dbuf_slice_set(dev_priv, DBUF_CTL_S2, true); > > + ret = intel_dbuf_slice_set(i915, DBUF_CTL_S2, true); > > else > > - ret = intel_dbuf_slice_set(dev_priv, DBUF_CTL_S2, false); > > + ret = intel_dbuf_slice_set(i915, DBUF_CTL_S2, false); > > if (ret) > > - dev_priv->wm.skl_hw.ddb.enabled_slices = req_slices; > > + i915->wm.skl_hw.ddb.enabled_slices = req_slices; > > } > > -static void icl_dbuf_enable(struct drm_i915_private *dev_priv) > > +static void icl_dbuf_enable(struct drm_i915_private *i915) > > { > > I915_WRITE(DBUF_CTL_S1, I915_READ(DBUF_CTL_S1) | DBUF_POWER_REQUEST); > > I915_WRITE(DBUF_CTL_S2, I915_READ(DBUF_CTL_S2) | DBUF_POWER_REQUEST); > > @@ -3616,10 +3616,10 @@ static void icl_dbuf_enable(struct drm_i915_private *dev_priv) > > * FIXME: for now pretend that we only have 1 slice, see > > * intel_enabled_dbuf_slices_num(). > > */ > > - dev_priv->wm.skl_hw.ddb.enabled_slices = 1; > > + i915->wm.skl_hw.ddb.enabled_slices = 1; > > } > > -static void icl_dbuf_disable(struct drm_i915_private *dev_priv) > > +static void icl_dbuf_disable(struct drm_i915_private *i915) > > { > > I915_WRITE(DBUF_CTL_S1, I915_READ(DBUF_CTL_S1) & ~DBUF_POWER_REQUEST); > > I915_WRITE(DBUF_CTL_S2, I915_READ(DBUF_CTL_S2) & ~DBUF_POWER_REQUEST); > > @@ -3635,10 +3635,10 @@ static void icl_dbuf_disable(struct drm_i915_private *dev_priv) > > * FIXME: for now pretend that the first slice is always > > * enabled, see intel_enabled_dbuf_slices_num(). > > */ > > - dev_priv->wm.skl_hw.ddb.enabled_slices = 1; > > + i915->wm.skl_hw.ddb.enabled_slices = 1; > > } > > -static void icl_mbus_init(struct drm_i915_private *dev_priv) > > +static void icl_mbus_init(struct drm_i915_private *i915) > > { > > u32 val; > > @@ -3650,7 +3650,7 @@ static void icl_mbus_init(struct drm_i915_private *dev_priv) > > I915_WRITE(MBUS_ABOX_CTL, val); > > } > > -static void hsw_assert_cdclk(struct drm_i915_private *dev_priv) > > +static void hsw_assert_cdclk(struct drm_i915_private *i915) > > { > > u32 val = I915_READ(LCPLL_CTL); > > @@ -3670,9 +3670,9 @@ static void hsw_assert_cdclk(struct drm_i915_private *dev_priv) > > DRM_ERROR("LCPLL not using non-SSC reference\n"); > > } > > -static void assert_can_disable_lcpll(struct drm_i915_private *dev_priv) > > +static void assert_can_disable_lcpll(struct drm_i915_private *i915) > > { > > - struct drm_device *dev = &dev_priv->drm; > > + struct drm_device *dev = &i915->drm; > > struct intel_crtc *crtc; > > for_each_intel_crtc(dev, crtc) > > @@ -3691,7 +3691,7 @@ static void assert_can_disable_lcpll(struct drm_i915_private *dev_priv) > > "Panel power on\n"); > > I915_STATE_WARN(I915_READ(BLC_PWM_CPU_CTL2) & BLM_PWM_ENABLE, > > "CPU PWM1 enabled\n"); > > - if (IS_HASWELL(dev_priv)) > > + if (IS_HASWELL(i915)) > > I915_STATE_WARN(I915_READ(HSW_BLC_PWM2_CTL) & BLM_PWM_ENABLE, > > "CPU PWM2 enabled\n"); > > I915_STATE_WARN(I915_READ(BLC_PWM_PCH_CTL1) & BLM_PCH_PWM_ENABLE, > > @@ -3707,21 +3707,21 @@ static void assert_can_disable_lcpll(struct drm_i915_private *dev_priv) > > * gen-specific and since we only disable LCPLL after we fully disable > > * the interrupts, the check below should be enough. > > */ > > - I915_STATE_WARN(intel_irqs_enabled(dev_priv), "IRQs enabled\n"); > > + I915_STATE_WARN(intel_irqs_enabled(i915), "IRQs enabled\n"); > > } > > -static u32 hsw_read_dcomp(struct drm_i915_private *dev_priv) > > +static u32 hsw_read_dcomp(struct drm_i915_private *i915) > > { > > - if (IS_HASWELL(dev_priv)) > > + if (IS_HASWELL(i915)) > > return I915_READ(D_COMP_HSW); > > else > > return I915_READ(D_COMP_BDW); > > } > > -static void hsw_write_dcomp(struct drm_i915_private *dev_priv, u32 val) > > +static void hsw_write_dcomp(struct drm_i915_private *i915, u32 val) > > { > > - if (IS_HASWELL(dev_priv)) { > > - if (sandybridge_pcode_write(dev_priv, > > + if (IS_HASWELL(i915)) { > > + if (sandybridge_pcode_write(i915, > > GEN6_PCODE_WRITE_D_COMP, val)) > > DRM_DEBUG_KMS("Failed to write to D_COMP\n"); > > } else { > > @@ -3738,12 +3738,12 @@ static void hsw_write_dcomp(struct drm_i915_private *dev_priv, u32 val) > > * register. Callers should take care of disabling all the display engine > > * functions, doing the mode unset, fixing interrupts, etc. > > */ > > -static void hsw_disable_lcpll(struct drm_i915_private *dev_priv, > > +static void hsw_disable_lcpll(struct drm_i915_private *i915, > > bool switch_to_fclk, bool allow_power_down) > > { > > u32 val; > > - assert_can_disable_lcpll(dev_priv); > > + assert_can_disable_lcpll(i915); > > val = I915_READ(LCPLL_CTL); > > @@ -3762,16 +3762,16 @@ static void hsw_disable_lcpll(struct drm_i915_private *dev_priv, > > I915_WRITE(LCPLL_CTL, val); > > POSTING_READ(LCPLL_CTL); > > - if (intel_wait_for_register(&dev_priv->uncore, LCPLL_CTL, > > + if (intel_wait_for_register(&i915->uncore, LCPLL_CTL, > > LCPLL_PLL_LOCK, 0, 1)) > > DRM_ERROR("LCPLL still locked\n"); > > - val = hsw_read_dcomp(dev_priv); > > + val = hsw_read_dcomp(i915); > > val |= D_COMP_COMP_DISABLE; > > - hsw_write_dcomp(dev_priv, val); > > + hsw_write_dcomp(i915, val); > > ndelay(100); > > - if (wait_for((hsw_read_dcomp(dev_priv) & > > + if (wait_for((hsw_read_dcomp(i915) & > > D_COMP_RCOMP_IN_PROGRESS) == 0, 1)) > > DRM_ERROR("D_COMP RCOMP still in progress\n"); > > @@ -3787,7 +3787,7 @@ static void hsw_disable_lcpll(struct drm_i915_private *dev_priv, > > * Fully restores LCPLL, disallowing power down and switching back to LCPLL > > * source. > > */ > > -static void hsw_restore_lcpll(struct drm_i915_private *dev_priv) > > +static void hsw_restore_lcpll(struct drm_i915_private *i915) > > { > > u32 val; > > @@ -3801,7 +3801,7 @@ static void hsw_restore_lcpll(struct drm_i915_private *dev_priv) > > * Make sure we're not on PC8 state before disabling PC8, otherwise > > * we'll hang the machine. To prevent PC8 state, just enable force_wake. > > */ > > - intel_uncore_forcewake_get(&dev_priv->uncore, FORCEWAKE_ALL); > > + intel_uncore_forcewake_get(&i915->uncore, FORCEWAKE_ALL); > > if (val & LCPLL_POWER_DOWN_ALLOW) { > > val &= ~LCPLL_POWER_DOWN_ALLOW; > > @@ -3809,16 +3809,16 @@ static void hsw_restore_lcpll(struct drm_i915_private *dev_priv) > > POSTING_READ(LCPLL_CTL); > > } > > - val = hsw_read_dcomp(dev_priv); > > + val = hsw_read_dcomp(i915); > > val |= D_COMP_COMP_FORCE; > > val &= ~D_COMP_COMP_DISABLE; > > - hsw_write_dcomp(dev_priv, val); > > + hsw_write_dcomp(i915, val); > > val = I915_READ(LCPLL_CTL); > > val &= ~LCPLL_PLL_DISABLE; > > I915_WRITE(LCPLL_CTL, val); > > - if (intel_wait_for_register(&dev_priv->uncore, LCPLL_CTL, > > + if (intel_wait_for_register(&i915->uncore, LCPLL_CTL, > > LCPLL_PLL_LOCK, LCPLL_PLL_LOCK, 5)) > > DRM_ERROR("LCPLL not locked yet\n"); > > @@ -3832,10 +3832,10 @@ static void hsw_restore_lcpll(struct drm_i915_private *dev_priv) > > DRM_ERROR("Switching back to LCPLL failed\n"); > > } > > - intel_uncore_forcewake_put(&dev_priv->uncore, FORCEWAKE_ALL); > > + intel_uncore_forcewake_put(&i915->uncore, FORCEWAKE_ALL); > > - intel_update_cdclk(dev_priv); > > - intel_dump_cdclk_state(&dev_priv->cdclk.hw, "Current CDCLK"); > > + intel_update_cdclk(i915); > > + intel_dump_cdclk_state(&i915->cdclk.hw, "Current CDCLK"); > > } > > /* > > @@ -3861,45 +3861,45 @@ static void hsw_restore_lcpll(struct drm_i915_private *dev_priv) > > * For more, read "Display Sequences for Package C8" on the hardware > > * documentation. > > */ > > -void hsw_enable_pc8(struct drm_i915_private *dev_priv) > > +void hsw_enable_pc8(struct drm_i915_private *i915) > > { > > u32 val; > > DRM_DEBUG_KMS("Enabling package C8+\n"); > > - if (HAS_PCH_LPT_LP(dev_priv)) { > > + if (HAS_PCH_LPT_LP(i915)) { > > val = I915_READ(SOUTH_DSPCLK_GATE_D); > > val &= ~PCH_LP_PARTITION_LEVEL_DISABLE; > > I915_WRITE(SOUTH_DSPCLK_GATE_D, val); > > } > > - lpt_disable_clkout_dp(dev_priv); > > - hsw_disable_lcpll(dev_priv, true, true); > > + lpt_disable_clkout_dp(i915); > > + hsw_disable_lcpll(i915, true, true); > > } > > -void hsw_disable_pc8(struct drm_i915_private *dev_priv) > > +void hsw_disable_pc8(struct drm_i915_private *i915) > > { > > u32 val; > > DRM_DEBUG_KMS("Disabling package C8+\n"); > > - hsw_restore_lcpll(dev_priv); > > - intel_init_pch_refclk(dev_priv); > > + hsw_restore_lcpll(i915); > > + intel_init_pch_refclk(i915); > > - if (HAS_PCH_LPT_LP(dev_priv)) { > > + if (HAS_PCH_LPT_LP(i915)) { > > val = I915_READ(SOUTH_DSPCLK_GATE_D); > > val |= PCH_LP_PARTITION_LEVEL_DISABLE; > > I915_WRITE(SOUTH_DSPCLK_GATE_D, val); > > } > > } > > -static void intel_pch_reset_handshake(struct drm_i915_private *dev_priv, > > +static void intel_pch_reset_handshake(struct drm_i915_private *i915, > > bool enable) > > { > > i915_reg_t reg; > > u32 reset_bits, val; > > - if (IS_IVYBRIDGE(dev_priv)) { > > + if (IS_IVYBRIDGE(i915)) { > > reg = GEN7_MSG_CTL; > > reset_bits = WAIT_FOR_PCH_FLR_ACK | WAIT_FOR_PCH_RESET_ACK; > > } else { > > @@ -3917,46 +3917,46 @@ static void intel_pch_reset_handshake(struct drm_i915_private *dev_priv, > > I915_WRITE(reg, val); > > } > > -static void skl_display_core_init(struct drm_i915_private *dev_priv, > > +static void skl_display_core_init(struct drm_i915_private *i915, > > bool resume) > > { > > - struct i915_power_domains *power_domains = &dev_priv->power_domains; > > + struct i915_power_domains *power_domains = &i915->power_domains; > > struct i915_power_well *well; > > - gen9_set_dc_state(dev_priv, DC_STATE_DISABLE); > > + gen9_set_dc_state(i915, DC_STATE_DISABLE); > > /* enable PCH reset handshake */ > > - intel_pch_reset_handshake(dev_priv, !HAS_PCH_NOP(dev_priv)); > > + intel_pch_reset_handshake(i915, !HAS_PCH_NOP(i915)); > > /* enable PG1 and Misc I/O */ > > mutex_lock(&power_domains->lock); > > - well = lookup_power_well(dev_priv, SKL_DISP_PW_1); > > - intel_power_well_enable(dev_priv, well); > > + well = lookup_power_well(i915, SKL_DISP_PW_1); > > + intel_power_well_enable(i915, well); > > - well = lookup_power_well(dev_priv, SKL_DISP_PW_MISC_IO); > > - intel_power_well_enable(dev_priv, well); > > + well = lookup_power_well(i915, SKL_DISP_PW_MISC_IO); > > + intel_power_well_enable(i915, well); > > mutex_unlock(&power_domains->lock); > > - intel_cdclk_init(dev_priv); > > + intel_cdclk_init(i915); > > - gen9_dbuf_enable(dev_priv); > > + gen9_dbuf_enable(i915); > > - if (resume && dev_priv->csr.dmc_payload) > > - intel_csr_load_program(dev_priv); > > + if (resume && i915->csr.dmc_payload) > > + intel_csr_load_program(i915); > > } > > -static void skl_display_core_uninit(struct drm_i915_private *dev_priv) > > +static void skl_display_core_uninit(struct drm_i915_private *i915) > > { > > - struct i915_power_domains *power_domains = &dev_priv->power_domains; > > + struct i915_power_domains *power_domains = &i915->power_domains; > > struct i915_power_well *well; > > - gen9_set_dc_state(dev_priv, DC_STATE_DISABLE); > > + gen9_set_dc_state(i915, DC_STATE_DISABLE); > > - gen9_dbuf_disable(dev_priv); > > + gen9_dbuf_disable(i915); > > - intel_cdclk_uninit(dev_priv); > > + intel_cdclk_uninit(i915); > > /* The spec doesn't call for removing the reset handshake flag */ > > /* disable PG1 and Misc I/O */ > > @@ -3969,21 +3969,21 @@ static void skl_display_core_uninit(struct drm_i915_private *dev_priv) > > * Note that even though the driver's request is removed power well 1 > > * may stay enabled after this due to DMC's own request on it. > > */ > > - well = lookup_power_well(dev_priv, SKL_DISP_PW_1); > > - intel_power_well_disable(dev_priv, well); > > + well = lookup_power_well(i915, SKL_DISP_PW_1); > > + intel_power_well_disable(i915, well); > > mutex_unlock(&power_domains->lock); > > usleep_range(10, 30); /* 10 us delay per Bspec */ > > } > > -void bxt_display_core_init(struct drm_i915_private *dev_priv, > > +void bxt_display_core_init(struct drm_i915_private *i915, > > bool resume) > > { > > - struct i915_power_domains *power_domains = &dev_priv->power_domains; > > + struct i915_power_domains *power_domains = &i915->power_domains; > > struct i915_power_well *well; > > - gen9_set_dc_state(dev_priv, DC_STATE_DISABLE); > > + gen9_set_dc_state(i915, DC_STATE_DISABLE); > > /* > > * NDE_RSTWRN_OPT RST PCH Handshake En must always be 0b on BXT > > @@ -3991,34 +3991,34 @@ void bxt_display_core_init(struct drm_i915_private *dev_priv, > > * Move the handshake programming to initialization sequence. > > * Previously was left up to BIOS. > > */ > > - intel_pch_reset_handshake(dev_priv, false); > > + intel_pch_reset_handshake(i915, false); > > /* Enable PG1 */ > > mutex_lock(&power_domains->lock); > > - well = lookup_power_well(dev_priv, SKL_DISP_PW_1); > > - intel_power_well_enable(dev_priv, well); > > + well = lookup_power_well(i915, SKL_DISP_PW_1); > > + intel_power_well_enable(i915, well); > > mutex_unlock(&power_domains->lock); > > - intel_cdclk_init(dev_priv); > > + intel_cdclk_init(i915); > > - gen9_dbuf_enable(dev_priv); > > + gen9_dbuf_enable(i915); > > - if (resume && dev_priv->csr.dmc_payload) > > - intel_csr_load_program(dev_priv); > > + if (resume && i915->csr.dmc_payload) > > + intel_csr_load_program(i915); > > } > > -void bxt_display_core_uninit(struct drm_i915_private *dev_priv) > > +void bxt_display_core_uninit(struct drm_i915_private *i915) > > { > > - struct i915_power_domains *power_domains = &dev_priv->power_domains; > > + struct i915_power_domains *power_domains = &i915->power_domains; > > struct i915_power_well *well; > > - gen9_set_dc_state(dev_priv, DC_STATE_DISABLE); > > + gen9_set_dc_state(i915, DC_STATE_DISABLE); > > - gen9_dbuf_disable(dev_priv); > > + gen9_dbuf_disable(i915); > > - intel_cdclk_uninit(dev_priv); > > + intel_cdclk_uninit(i915); > > /* The spec doesn't call for removing the reset handshake flag */ > > @@ -4029,60 +4029,60 @@ void bxt_display_core_uninit(struct drm_i915_private *dev_priv) > > */ > > mutex_lock(&power_domains->lock); > > - well = lookup_power_well(dev_priv, SKL_DISP_PW_1); > > - intel_power_well_disable(dev_priv, well); > > + well = lookup_power_well(i915, SKL_DISP_PW_1); > > + intel_power_well_disable(i915, well); > > mutex_unlock(&power_domains->lock); > > usleep_range(10, 30); /* 10 us delay per Bspec */ > > } > > -static void cnl_display_core_init(struct drm_i915_private *dev_priv, bool resume) > > +static void cnl_display_core_init(struct drm_i915_private *i915, bool resume) > > { > > - struct i915_power_domains *power_domains = &dev_priv->power_domains; > > + struct i915_power_domains *power_domains = &i915->power_domains; > > struct i915_power_well *well; > > - gen9_set_dc_state(dev_priv, DC_STATE_DISABLE); > > + gen9_set_dc_state(i915, DC_STATE_DISABLE); > > /* 1. Enable PCH Reset Handshake */ > > - intel_pch_reset_handshake(dev_priv, !HAS_PCH_NOP(dev_priv)); > > + intel_pch_reset_handshake(i915, !HAS_PCH_NOP(i915)); > > /* 2-3. */ > > - intel_combo_phy_init(dev_priv); > > + intel_combo_phy_init(i915); > > /* > > * 4. Enable Power Well 1 (PG1). > > * The AUX IO power wells will be enabled on demand. > > */ > > mutex_lock(&power_domains->lock); > > - well = lookup_power_well(dev_priv, SKL_DISP_PW_1); > > - intel_power_well_enable(dev_priv, well); > > + well = lookup_power_well(i915, SKL_DISP_PW_1); > > + intel_power_well_enable(i915, well); > > mutex_unlock(&power_domains->lock); > > /* 5. Enable CD clock */ > > - intel_cdclk_init(dev_priv); > > + intel_cdclk_init(i915); > > /* 6. Enable DBUF */ > > - gen9_dbuf_enable(dev_priv); > > + gen9_dbuf_enable(i915); > > - if (resume && dev_priv->csr.dmc_payload) > > - intel_csr_load_program(dev_priv); > > + if (resume && i915->csr.dmc_payload) > > + intel_csr_load_program(i915); > > } > > -static void cnl_display_core_uninit(struct drm_i915_private *dev_priv) > > +static void cnl_display_core_uninit(struct drm_i915_private *i915) > > { > > - struct i915_power_domains *power_domains = &dev_priv->power_domains; > > + struct i915_power_domains *power_domains = &i915->power_domains; > > struct i915_power_well *well; > > - gen9_set_dc_state(dev_priv, DC_STATE_DISABLE); > > + gen9_set_dc_state(i915, DC_STATE_DISABLE); > > /* 1. Disable all display engine functions -> aready done */ > > /* 2. Disable DBUF */ > > - gen9_dbuf_disable(dev_priv); > > + gen9_dbuf_disable(i915); > > /* 3. Disable CD clock */ > > - intel_cdclk_uninit(dev_priv); > > + intel_cdclk_uninit(i915); > > /* > > * 4. Disable Power Well 1 (PG1). > > @@ -4090,66 +4090,66 @@ static void cnl_display_core_uninit(struct drm_i915_private *dev_priv) > > * disabled at this point. > > */ > > mutex_lock(&power_domains->lock); > > - well = lookup_power_well(dev_priv, SKL_DISP_PW_1); > > - intel_power_well_disable(dev_priv, well); > > + well = lookup_power_well(i915, SKL_DISP_PW_1); > > + intel_power_well_disable(i915, well); > > mutex_unlock(&power_domains->lock); > > usleep_range(10, 30); /* 10 us delay per Bspec */ > > /* 5. */ > > - intel_combo_phy_uninit(dev_priv); > > + intel_combo_phy_uninit(i915); > > } > > -void icl_display_core_init(struct drm_i915_private *dev_priv, > > +void icl_display_core_init(struct drm_i915_private *i915, > > bool resume) > > { > > - struct i915_power_domains *power_domains = &dev_priv->power_domains; > > + struct i915_power_domains *power_domains = &i915->power_domains; > > struct i915_power_well *well; > > - gen9_set_dc_state(dev_priv, DC_STATE_DISABLE); > > + gen9_set_dc_state(i915, DC_STATE_DISABLE); > > /* 1. Enable PCH reset handshake. */ > > - intel_pch_reset_handshake(dev_priv, !HAS_PCH_NOP(dev_priv)); > > + intel_pch_reset_handshake(i915, !HAS_PCH_NOP(i915)); > > /* 2. Initialize all combo phys */ > > - intel_combo_phy_init(dev_priv); > > + intel_combo_phy_init(i915); > > /* > > * 3. Enable Power Well 1 (PG1). > > * The AUX IO power wells will be enabled on demand. > > */ > > mutex_lock(&power_domains->lock); > > - well = lookup_power_well(dev_priv, SKL_DISP_PW_1); > > - intel_power_well_enable(dev_priv, well); > > + well = lookup_power_well(i915, SKL_DISP_PW_1); > > + intel_power_well_enable(i915, well); > > mutex_unlock(&power_domains->lock); > > /* 4. Enable CDCLK. */ > > - intel_cdclk_init(dev_priv); > > + intel_cdclk_init(i915); > > /* 5. Enable DBUF. */ > > - icl_dbuf_enable(dev_priv); > > + icl_dbuf_enable(i915); > > /* 6. Setup MBUS. */ > > - icl_mbus_init(dev_priv); > > + icl_mbus_init(i915); > > - if (resume && dev_priv->csr.dmc_payload) > > - intel_csr_load_program(dev_priv); > > + if (resume && i915->csr.dmc_payload) > > + intel_csr_load_program(i915); > > } > > -void icl_display_core_uninit(struct drm_i915_private *dev_priv) > > +void icl_display_core_uninit(struct drm_i915_private *i915) > > { > > - struct i915_power_domains *power_domains = &dev_priv->power_domains; > > + struct i915_power_domains *power_domains = &i915->power_domains; > > struct i915_power_well *well; > > - gen9_set_dc_state(dev_priv, DC_STATE_DISABLE); > > + gen9_set_dc_state(i915, DC_STATE_DISABLE); > > /* 1. Disable all display engine functions -> aready done */ > > /* 2. Disable DBUF */ > > - icl_dbuf_disable(dev_priv); > > + icl_dbuf_disable(i915); > > /* 3. Disable CD clock */ > > - intel_cdclk_uninit(dev_priv); > > + intel_cdclk_uninit(i915); > > /* > > * 4. Disable Power Well 1 (PG1). > > @@ -4157,20 +4157,20 @@ void icl_display_core_uninit(struct drm_i915_private *dev_priv) > > * disabled at this point. > > */ > > mutex_lock(&power_domains->lock); > > - well = lookup_power_well(dev_priv, SKL_DISP_PW_1); > > - intel_power_well_disable(dev_priv, well); > > + well = lookup_power_well(i915, SKL_DISP_PW_1); > > + intel_power_well_disable(i915, well); > > mutex_unlock(&power_domains->lock); > > /* 5. */ > > - intel_combo_phy_uninit(dev_priv); > > + intel_combo_phy_uninit(i915); > > } > > -static void chv_phy_control_init(struct drm_i915_private *dev_priv) > > +static void chv_phy_control_init(struct drm_i915_private *i915) > > { > > struct i915_power_well *cmn_bc = > > - lookup_power_well(dev_priv, VLV_DISP_PW_DPIO_CMN_BC); > > + lookup_power_well(i915, VLV_DISP_PW_DPIO_CMN_BC); > > struct i915_power_well *cmn_d = > > - lookup_power_well(dev_priv, CHV_DISP_PW_DPIO_CMN_D); > > + lookup_power_well(i915, CHV_DISP_PW_DPIO_CMN_D); > > /* > > * DISPLAY_PHY_CONTROL can get corrupted if read. As a > > @@ -4179,7 +4179,7 @@ static void chv_phy_control_init(struct drm_i915_private *dev_priv) > > * power well state and lane status to reconstruct the > > * expected initial value. > > */ > > - dev_priv->chv_phy_control = > > + i915->chv_phy_control = > > PHY_LDO_SEQ_DELAY(PHY_LDO_DELAY_600NS, DPIO_PHY0) | > > PHY_LDO_SEQ_DELAY(PHY_LDO_DELAY_600NS, DPIO_PHY1) | > > PHY_CH_POWER_MODE(PHY_CH_DEEP_PSR, DPIO_PHY0, DPIO_CH0) | > > @@ -4193,7 +4193,7 @@ static void chv_phy_control_init(struct drm_i915_private *dev_priv) > > * override and set the lane powerdown bits accding to the > > * current lane status. > > */ > > - if (cmn_bc->desc->ops->is_enabled(dev_priv, cmn_bc)) { > > + if (cmn_bc->desc->ops->is_enabled(i915, cmn_bc)) { > > u32 status = I915_READ(DPLL(PIPE_A)); > > unsigned int mask; > > @@ -4201,30 +4201,30 @@ static void chv_phy_control_init(struct drm_i915_private *dev_priv) > > if (mask == 0xf) > > mask = 0x0; > > else > > - dev_priv->chv_phy_control |= > > + i915->chv_phy_control |= > > PHY_CH_POWER_DOWN_OVRD_EN(DPIO_PHY0, DPIO_CH0); > > - dev_priv->chv_phy_control |= > > + i915->chv_phy_control |= > > PHY_CH_POWER_DOWN_OVRD(mask, DPIO_PHY0, DPIO_CH0); > > mask = (status & DPLL_PORTC_READY_MASK) >> 4; > > if (mask == 0xf) > > mask = 0x0; > > else > > - dev_priv->chv_phy_control |= > > + i915->chv_phy_control |= > > PHY_CH_POWER_DOWN_OVRD_EN(DPIO_PHY0, DPIO_CH1); > > - dev_priv->chv_phy_control |= > > + i915->chv_phy_control |= > > PHY_CH_POWER_DOWN_OVRD(mask, DPIO_PHY0, DPIO_CH1); > > - dev_priv->chv_phy_control |= PHY_COM_LANE_RESET_DEASSERT(DPIO_PHY0); > > + i915->chv_phy_control |= PHY_COM_LANE_RESET_DEASSERT(DPIO_PHY0); > > - dev_priv->chv_phy_assert[DPIO_PHY0] = false; > > + i915->chv_phy_assert[DPIO_PHY0] = false; > > } else { > > - dev_priv->chv_phy_assert[DPIO_PHY0] = true; > > + i915->chv_phy_assert[DPIO_PHY0] = true; > > } > > - if (cmn_d->desc->ops->is_enabled(dev_priv, cmn_d)) { > > + if (cmn_d->desc->ops->is_enabled(i915, cmn_d)) { > > u32 status = I915_READ(DPIO_PHY_STATUS); > > unsigned int mask; > > @@ -4233,42 +4233,42 @@ static void chv_phy_control_init(struct drm_i915_private *dev_priv) > > if (mask == 0xf) > > mask = 0x0; > > else > > - dev_priv->chv_phy_control |= > > + i915->chv_phy_control |= > > PHY_CH_POWER_DOWN_OVRD_EN(DPIO_PHY1, DPIO_CH0); > > - dev_priv->chv_phy_control |= > > + i915->chv_phy_control |= > > PHY_CH_POWER_DOWN_OVRD(mask, DPIO_PHY1, DPIO_CH0); > > - dev_priv->chv_phy_control |= PHY_COM_LANE_RESET_DEASSERT(DPIO_PHY1); > > + i915->chv_phy_control |= PHY_COM_LANE_RESET_DEASSERT(DPIO_PHY1); > > - dev_priv->chv_phy_assert[DPIO_PHY1] = false; > > + i915->chv_phy_assert[DPIO_PHY1] = false; > > } else { > > - dev_priv->chv_phy_assert[DPIO_PHY1] = true; > > + i915->chv_phy_assert[DPIO_PHY1] = true; > > } > > - I915_WRITE(DISPLAY_PHY_CONTROL, dev_priv->chv_phy_control); > > + I915_WRITE(DISPLAY_PHY_CONTROL, i915->chv_phy_control); > > DRM_DEBUG_KMS("Initial PHY_CONTROL=0x%08x\n", > > - dev_priv->chv_phy_control); > > + i915->chv_phy_control); > > } > > -static void vlv_cmnlane_wa(struct drm_i915_private *dev_priv) > > +static void vlv_cmnlane_wa(struct drm_i915_private *i915) > > { > > struct i915_power_well *cmn = > > - lookup_power_well(dev_priv, VLV_DISP_PW_DPIO_CMN_BC); > > + lookup_power_well(i915, VLV_DISP_PW_DPIO_CMN_BC); > > struct i915_power_well *disp2d = > > - lookup_power_well(dev_priv, VLV_DISP_PW_DISP2D); > > + lookup_power_well(i915, VLV_DISP_PW_DISP2D); > > /* If the display might be already active skip this */ > > - if (cmn->desc->ops->is_enabled(dev_priv, cmn) && > > - disp2d->desc->ops->is_enabled(dev_priv, disp2d) && > > + if (cmn->desc->ops->is_enabled(i915, cmn) && > > + disp2d->desc->ops->is_enabled(i915, disp2d) && > > I915_READ(DPIO_CTL) & DPIO_CMNRST) > > return; > > DRM_DEBUG_KMS("toggling display PHY side reset\n"); > > /* cmnlane needs DPLL registers */ > > - disp2d->desc->ops->enable(dev_priv, disp2d); > > + disp2d->desc->ops->enable(i915, disp2d); > > /* > > * From VLV2A0_DP_eDP_HDMI_DPIO_driver_vbios_notes_11.docx: > > @@ -4277,27 +4277,27 @@ static void vlv_cmnlane_wa(struct drm_i915_private *dev_priv) > > * Simply ungating isn't enough to reset the PHY enough to get > > * ports and lanes running. > > */ > > - cmn->desc->ops->disable(dev_priv, cmn); > > + cmn->desc->ops->disable(i915, cmn); > > } > > -static bool vlv_punit_is_power_gated(struct drm_i915_private *dev_priv, u32 reg0) > > +static bool vlv_punit_is_power_gated(struct drm_i915_private *i915, u32 reg0) > > { > > bool ret; > > - vlv_punit_get(dev_priv); > > - ret = (vlv_punit_read(dev_priv, reg0) & SSPM0_SSC_MASK) == SSPM0_SSC_PWR_GATE; > > - vlv_punit_put(dev_priv); > > + vlv_punit_get(i915); > > + ret = (vlv_punit_read(i915, reg0) & SSPM0_SSC_MASK) == SSPM0_SSC_PWR_GATE; > > + vlv_punit_put(i915); > > return ret; > > } > > -static void assert_ved_power_gated(struct drm_i915_private *dev_priv) > > +static void assert_ved_power_gated(struct drm_i915_private *i915) > > { > > - WARN(!vlv_punit_is_power_gated(dev_priv, PUNIT_REG_VEDSSPM0), > > + WARN(!vlv_punit_is_power_gated(i915, PUNIT_REG_VEDSSPM0), > > "VED not power gated\n"); > > } > > -static void assert_isp_power_gated(struct drm_i915_private *dev_priv) > > +static void assert_isp_power_gated(struct drm_i915_private *i915) > > { > > static const struct pci_device_id isp_ids[] = { > > {PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x0f38)}, > > @@ -4306,11 +4306,11 @@ static void assert_isp_power_gated(struct drm_i915_private *dev_priv) > > }; > > WARN(!pci_dev_present(isp_ids) && > > - !vlv_punit_is_power_gated(dev_priv, PUNIT_REG_ISPSSPM0), > > + !vlv_punit_is_power_gated(i915, PUNIT_REG_ISPSSPM0), > > "ISP not power gated\n"); > > } > > -static void intel_power_domains_verify_state(struct drm_i915_private *dev_priv); > > +static void intel_power_domains_verify_state(struct drm_i915_private *i915); > > /** > > * intel_power_domains_init_hw - initialize hardware power domain state > > diff --git a/drivers/gpu/drm/i915/intel_display_power.h b/drivers/gpu/drm/i915/intel_display_power.h > > index ff57b0a7fe59..0310fe114ecc 100644 > > --- a/drivers/gpu/drm/i915/intel_display_power.h > > +++ b/drivers/gpu/drm/i915/intel_display_power.h > > @@ -80,23 +80,23 @@ struct i915_power_well_ops { > > * during driver init and resume time, possibly after first calling > > * the enable/disable handlers. > > */ > > - void (*sync_hw)(struct drm_i915_private *dev_priv, > > + void (*sync_hw)(struct drm_i915_private *i915, > > struct i915_power_well *power_well); > > /* > > * Enable the well and resources that depend on it (for example > > * interrupts located on the well). Called after the 0->1 refcount > > * transition. > > */ > > - void (*enable)(struct drm_i915_private *dev_priv, > > + void (*enable)(struct drm_i915_private *i915, > > struct i915_power_well *power_well); > > /* > > * Disable the well and resources that depend on it. Called after > > * the 1->0 refcount transition. > > */ > > - void (*disable)(struct drm_i915_private *dev_priv, > > + void (*disable)(struct drm_i915_private *i915, > > struct i915_power_well *power_well); > > /* Returns the hw enabled state. */ > > - bool (*is_enabled)(struct drm_i915_private *dev_priv, > > + bool (*is_enabled)(struct drm_i915_private *i915, > > struct i915_power_well *power_well); > > }; > > @@ -204,48 +204,48 @@ struct i915_power_domains { > > for_each_power_well_reverse(__dev_priv, __power_well) \ > > for_each_if((__power_well)->desc->domains & (__domain_mask)) > > -void skl_enable_dc6(struct drm_i915_private *dev_priv); > > -void gen9_sanitize_dc_state(struct drm_i915_private *dev_priv); > > -void bxt_enable_dc9(struct drm_i915_private *dev_priv); > > -void bxt_disable_dc9(struct drm_i915_private *dev_priv); > > -void gen9_enable_dc5(struct drm_i915_private *dev_priv); > > +void skl_enable_dc6(struct drm_i915_private *i915); > > +void gen9_sanitize_dc_state(struct drm_i915_private *i915); > > +void bxt_enable_dc9(struct drm_i915_private *i915); > > +void bxt_disable_dc9(struct drm_i915_private *i915); > > +void gen9_enable_dc5(struct drm_i915_private *i915); > > -int intel_power_domains_init(struct drm_i915_private *dev_priv); > > -void intel_power_domains_cleanup(struct drm_i915_private *dev_priv); > > -void intel_power_domains_init_hw(struct drm_i915_private *dev_priv, bool resume); > > -void intel_power_domains_fini_hw(struct drm_i915_private *dev_priv); > > -void icl_display_core_init(struct drm_i915_private *dev_priv, bool resume); > > -void icl_display_core_uninit(struct drm_i915_private *dev_priv); > > -void intel_power_domains_enable(struct drm_i915_private *dev_priv); > > -void intel_power_domains_disable(struct drm_i915_private *dev_priv); > > -void intel_power_domains_suspend(struct drm_i915_private *dev_priv, > > +int intel_power_domains_init(struct drm_i915_private *i915); > > +void intel_power_domains_cleanup(struct drm_i915_private *i915); > > +void intel_power_domains_init_hw(struct drm_i915_private *i915, bool resume); > > +void intel_power_domains_fini_hw(struct drm_i915_private *i915); > > +void icl_display_core_init(struct drm_i915_private *i915, bool resume); > > +void icl_display_core_uninit(struct drm_i915_private *i915); > > +void intel_power_domains_enable(struct drm_i915_private *i915); > > +void intel_power_domains_disable(struct drm_i915_private *i915); > > +void intel_power_domains_suspend(struct drm_i915_private *i915, > > enum i915_drm_suspend_mode); > > -void intel_power_domains_resume(struct drm_i915_private *dev_priv); > > -void hsw_enable_pc8(struct drm_i915_private *dev_priv); > > -void hsw_disable_pc8(struct drm_i915_private *dev_priv); > > -void bxt_display_core_init(struct drm_i915_private *dev_priv, bool resume); > > -void bxt_display_core_uninit(struct drm_i915_private *dev_priv); > > +void intel_power_domains_resume(struct drm_i915_private *i915); > > +void hsw_enable_pc8(struct drm_i915_private *i915); > > +void hsw_disable_pc8(struct drm_i915_private *i915); > > +void bxt_display_core_init(struct drm_i915_private *i915, bool resume); > > +void bxt_display_core_uninit(struct drm_i915_private *i915); > > const char * > > intel_display_power_domain_str(enum intel_display_power_domain domain); > > -bool intel_display_power_is_enabled(struct drm_i915_private *dev_priv, > > +bool intel_display_power_is_enabled(struct drm_i915_private *i915, > > enum intel_display_power_domain domain); > > -bool __intel_display_power_is_enabled(struct drm_i915_private *dev_priv, > > +bool __intel_display_power_is_enabled(struct drm_i915_private *i915, > > enum intel_display_power_domain domain); > > -intel_wakeref_t intel_display_power_get(struct drm_i915_private *dev_priv, > > +intel_wakeref_t intel_display_power_get(struct drm_i915_private *i915, > > enum intel_display_power_domain domain); > > intel_wakeref_t > > -intel_display_power_get_if_enabled(struct drm_i915_private *dev_priv, > > +intel_display_power_get_if_enabled(struct drm_i915_private *i915, > > enum intel_display_power_domain domain); > > -void intel_display_power_put_unchecked(struct drm_i915_private *dev_priv, > > +void intel_display_power_put_unchecked(struct drm_i915_private *i915, > > enum intel_display_power_domain domain); > > void __intel_display_power_put_async(struct drm_i915_private *i915, > > enum intel_display_power_domain domain, > > intel_wakeref_t wakeref); > > void intel_display_power_flush_work(struct drm_i915_private *i915); > > #if IS_ENABLED(CONFIG_DRM_I915_DEBUG_RUNTIME_PM) > > -void intel_display_power_put(struct drm_i915_private *dev_priv, > > +void intel_display_power_put(struct drm_i915_private *i915, > > enum intel_display_power_domain domain, > > intel_wakeref_t wakeref); > > static inline void > > @@ -277,12 +277,12 @@ intel_display_power_put_async(struct drm_i915_private *i915, > > for ((wf) = intel_display_power_get((i915), (domain)); (wf); \ > > intel_display_power_put_async((i915), (domain), (wf)), (wf) = 0) > > -void icl_dbuf_slices_update(struct drm_i915_private *dev_priv, > > +void icl_dbuf_slices_update(struct drm_i915_private *i915, > > u8 req_slices); > > void chv_phy_powergate_lanes(struct intel_encoder *encoder, > > bool override, unsigned int mask); > > -bool chv_phy_powergate_ch(struct drm_i915_private *dev_priv, enum dpio_phy phy, > > +bool chv_phy_powergate_ch(struct drm_i915_private *i915, enum dpio_phy phy, > > enum dpio_channel ch, bool override); > > #endif /* __INTEL_DISPLAY_POWER_H__ */ > > diff --git a/drivers/gpu/drm/i915/intel_dp.c b/drivers/gpu/drm/i915/intel_dp.c > > index 4336df46fe78..d4dc2b9506e5 100644 > > --- a/drivers/gpu/drm/i915/intel_dp.c > > +++ b/drivers/gpu/drm/i915/intel_dp.c > > @@ -160,7 +160,7 @@ static bool edp_panel_vdd_on(struct intel_dp *intel_dp); > > static void edp_panel_vdd_off(struct intel_dp *intel_dp, bool sync); > > static void vlv_init_panel_power_sequencer(struct intel_encoder *encoder, > > const struct intel_crtc_state *crtc_state); > > -static void vlv_steal_power_sequencer(struct drm_i915_private *dev_priv, > > +static void vlv_steal_power_sequencer(struct drm_i915_private *i915, > > enum pipe pipe); > > static void intel_dp_unset_edid(struct intel_dp *intel_dp); > > @@ -214,8 +214,8 @@ static int intel_dp_max_common_rate(struct intel_dp *intel_dp) > > static int intel_dp_get_fia_supported_lane_count(struct intel_dp *intel_dp) > > { > > struct intel_digital_port *dig_port = dp_to_dig_port(intel_dp); > > - struct drm_i915_private *dev_priv = to_i915(dig_port->base.base.dev); > > - enum tc_port tc_port = intel_port_to_tc(dev_priv, dig_port->base.port); > > + struct drm_i915_private *i915 = to_i915(dig_port->base.base.dev); > > + enum tc_port tc_port = intel_port_to_tc(i915, dig_port->base.port); > > intel_wakeref_t wakeref; > > u32 lane_info; > > @@ -223,7 +223,7 @@ static int intel_dp_get_fia_supported_lane_count(struct intel_dp *intel_dp) > > return 4; > > lane_info = 0; > > - with_intel_display_power(dev_priv, POWER_DOMAIN_DISPLAY_CORE, wakeref) > > + with_intel_display_power(i915, POWER_DOMAIN_DISPLAY_CORE, wakeref) > > lane_info = (I915_READ(PORT_TX_DFLEXDPSP) & > > DP_LANE_ASSIGNMENT_MASK(tc_port)) >> > > DP_LANE_ASSIGNMENT_SHIFT(tc_port); > > @@ -284,8 +284,8 @@ intel_dp_downstream_max_dotclock(struct intel_dp *intel_dp) > > { > > struct intel_digital_port *intel_dig_port = dp_to_dig_port(intel_dp); > > struct intel_encoder *encoder = &intel_dig_port->base; > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > - int max_dotclk = dev_priv->max_dotclk_freq; > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > + int max_dotclk = i915->max_dotclk_freq; > > int ds_max_dotclk; > > int type = intel_dp->downstream_ports[0] & DP_DS_PORT_TYPE_MASK; > > @@ -305,7 +305,7 @@ intel_dp_downstream_max_dotclock(struct intel_dp *intel_dp) > > static int cnl_max_source_rate(struct intel_dp *intel_dp) > > { > > struct intel_digital_port *dig_port = dp_to_dig_port(intel_dp); > > - struct drm_i915_private *dev_priv = to_i915(dig_port->base.base.dev); > > + struct drm_i915_private *i915 = to_i915(dig_port->base.base.dev); > > enum port port = dig_port->base.port; > > u32 voltage = I915_READ(CNL_PORT_COMP_DW3) & VOLTAGE_INFO_MASK; > > @@ -315,7 +315,7 @@ static int cnl_max_source_rate(struct intel_dp *intel_dp) > > return 540000; > > /* For this SKU 8.1G is supported in all ports */ > > - if (IS_CNL_WITH_PORT_F(dev_priv)) > > + if (IS_CNL_WITH_PORT_F(i915)) > > return 810000; > > /* For other SKUs, max rate on ports A and D is 5.4G */ > > @@ -328,11 +328,11 @@ static int cnl_max_source_rate(struct intel_dp *intel_dp) > > static int icl_max_source_rate(struct intel_dp *intel_dp) > > { > > struct intel_digital_port *dig_port = dp_to_dig_port(intel_dp); > > - struct drm_i915_private *dev_priv = to_i915(dig_port->base.base.dev); > > + struct drm_i915_private *i915 = to_i915(dig_port->base.base.dev); > > enum port port = dig_port->base.port; > > - if (intel_port_is_combophy(dev_priv, port) && > > - !IS_ELKHARTLAKE(dev_priv) && > > + if (intel_port_is_combophy(i915, port) && > > + !IS_ELKHARTLAKE(i915) && > > !intel_dp_is_edp(intel_dp)) > > return 540000; > > @@ -359,30 +359,30 @@ intel_dp_set_source_rates(struct intel_dp *intel_dp) > > 162000, 270000 > > }; > > struct intel_digital_port *dig_port = dp_to_dig_port(intel_dp); > > - struct drm_i915_private *dev_priv = to_i915(dig_port->base.base.dev); > > + struct drm_i915_private *i915 = to_i915(dig_port->base.base.dev); > > const struct ddi_vbt_port_info *info = > > - &dev_priv->vbt.ddi_port_info[dig_port->base.port]; > > + &i915->vbt.ddi_port_info[dig_port->base.port]; > > const int *source_rates; > > int size, max_rate = 0, vbt_max_rate = info->dp_max_link_rate; > > /* This should only be done once */ > > WARN_ON(intel_dp->source_rates || intel_dp->num_source_rates); > > - if (INTEL_GEN(dev_priv) >= 10) { > > + if (INTEL_GEN(i915) >= 10) { > > source_rates = cnl_rates; > > size = ARRAY_SIZE(cnl_rates); > > - if (IS_GEN(dev_priv, 10)) > > + if (IS_GEN(i915, 10)) > > max_rate = cnl_max_source_rate(intel_dp); > > else > > max_rate = icl_max_source_rate(intel_dp); > > - } else if (IS_GEN9_LP(dev_priv)) { > > + } else if (IS_GEN9_LP(i915)) { > > source_rates = bxt_rates; > > size = ARRAY_SIZE(bxt_rates); > > - } else if (IS_GEN9_BC(dev_priv)) { > > + } else if (IS_GEN9_BC(i915)) { > > source_rates = skl_rates; > > size = ARRAY_SIZE(skl_rates); > > - } else if ((IS_HASWELL(dev_priv) && !IS_HSW_ULX(dev_priv)) || > > - IS_BROADWELL(dev_priv)) { > > + } else if ((IS_HASWELL(i915) && !IS_HSW_ULX(i915)) || > > + IS_BROADWELL(i915)) { > > source_rates = hsw_rates; > > size = ARRAY_SIZE(hsw_rates); > > } else { > > @@ -532,7 +532,7 @@ intel_dp_mode_valid(struct drm_connector *connector, > > struct intel_dp *intel_dp = intel_attached_dp(connector); > > struct intel_connector *intel_connector = to_intel_connector(connector); > > struct drm_display_mode *fixed_mode = intel_connector->panel.fixed_mode; > > - struct drm_i915_private *dev_priv = to_i915(connector->dev); > > + struct drm_i915_private *i915 = to_i915(connector->dev); > > int target_clock = mode->clock; > > int max_rate, mode_rate, max_lanes, max_link_clock; > > int max_dotclk; > > @@ -564,7 +564,7 @@ intel_dp_mode_valid(struct drm_connector *connector, > > * Output bpp is stored in 6.4 format so right shift by 4 to get the > > * integer value since we support only integer values of bpp. > > */ > > - if ((INTEL_GEN(dev_priv) >= 10 || IS_GEMINILAKE(dev_priv)) && > > + if ((INTEL_GEN(i915) >= 10 || IS_GEMINILAKE(i915)) && > > drm_dp_sink_supports_dsc(intel_dp->dsc_dpcd)) { > > if (intel_dp_is_edp(intel_dp)) { > > dsc_max_output_bpp = > > @@ -630,17 +630,17 @@ intel_dp_pps_init(struct intel_dp *intel_dp); > > static intel_wakeref_t > > pps_lock(struct intel_dp *intel_dp) > > { > > - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp); > > + struct drm_i915_private *i915 = dp_to_i915(intel_dp); > > intel_wakeref_t wakeref; > > /* > > * See intel_power_sequencer_reset() why we need > > * a power domain reference here. > > */ > > - wakeref = intel_display_power_get(dev_priv, > > + wakeref = intel_display_power_get(i915, > > intel_aux_power_domain(dp_to_dig_port(intel_dp))); > > - mutex_lock(&dev_priv->pps_mutex); > > + mutex_lock(&i915->pps_mutex); > > return wakeref; > > } > > @@ -648,10 +648,10 @@ pps_lock(struct intel_dp *intel_dp) > > static intel_wakeref_t > > pps_unlock(struct intel_dp *intel_dp, intel_wakeref_t wakeref) > > { > > - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp); > > + struct drm_i915_private *i915 = dp_to_i915(intel_dp); > > - mutex_unlock(&dev_priv->pps_mutex); > > - intel_display_power_put(dev_priv, > > + mutex_unlock(&i915->pps_mutex); > > + intel_display_power_put(i915, > > intel_aux_power_domain(dp_to_dig_port(intel_dp)), > > wakeref); > > return 0; > > @@ -663,7 +663,7 @@ pps_unlock(struct intel_dp *intel_dp, intel_wakeref_t wakeref) > > static void > > vlv_power_sequencer_kick(struct intel_dp *intel_dp) > > { > > - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp); > > + struct drm_i915_private *i915 = dp_to_i915(intel_dp); > > struct intel_digital_port *intel_dig_port = dp_to_dig_port(intel_dp); > > enum pipe pipe = intel_dp->pps_pipe; > > bool pll_enabled, release_cl_override = false; > > @@ -687,7 +687,7 @@ vlv_power_sequencer_kick(struct intel_dp *intel_dp) > > DP |= DP_PORT_WIDTH(1); > > DP |= DP_LINK_TRAIN_PAT_1; > > - if (IS_CHERRYVIEW(dev_priv)) > > + if (IS_CHERRYVIEW(i915)) > > DP |= DP_PIPE_SEL_CHV(pipe); > > else > > DP |= DP_PIPE_SEL(pipe); > > @@ -699,10 +699,10 @@ vlv_power_sequencer_kick(struct intel_dp *intel_dp) > > * So enable temporarily it if it's not already enabled. > > */ > > if (!pll_enabled) { > > - release_cl_override = IS_CHERRYVIEW(dev_priv) && > > - !chv_phy_powergate_ch(dev_priv, phy, ch, true); > > + release_cl_override = IS_CHERRYVIEW(i915) && > > + !chv_phy_powergate_ch(i915, phy, ch, true); > > - if (vlv_force_pll_on(dev_priv, pipe, IS_CHERRYVIEW(dev_priv) ? > > + if (vlv_force_pll_on(i915, pipe, IS_CHERRYVIEW(i915) ? > > &chv_dpll[0].dpll : &vlv_dpll[0].dpll)) { > > DRM_ERROR("Failed to force on pll for pipe %c!\n", > > pipe_name(pipe)); > > @@ -726,14 +726,14 @@ vlv_power_sequencer_kick(struct intel_dp *intel_dp) > > POSTING_READ(intel_dp->output_reg); > > if (!pll_enabled) { > > - vlv_force_pll_off(dev_priv, pipe); > > + vlv_force_pll_off(i915, pipe); > > if (release_cl_override) > > - chv_phy_powergate_ch(dev_priv, phy, ch, false); > > + chv_phy_powergate_ch(i915, phy, ch, false); > > } > > } > > -static enum pipe vlv_find_free_pps(struct drm_i915_private *dev_priv) > > +static enum pipe vlv_find_free_pps(struct drm_i915_private *i915) > > { > > struct intel_encoder *encoder; > > unsigned int pipes = (1 << PIPE_A) | (1 << PIPE_B); > > @@ -742,7 +742,7 @@ static enum pipe vlv_find_free_pps(struct drm_i915_private *dev_priv) > > * We don't have power sequencer currently. > > * Pick one that's not used by other ports. > > */ > > - for_each_intel_dp(&dev_priv->drm, encoder) { > > + for_each_intel_dp(&i915->drm, encoder) { > > struct intel_dp *intel_dp = enc_to_intel_dp(&encoder->base); > > if (encoder->type == INTEL_OUTPUT_EDP) { > > @@ -768,11 +768,11 @@ static enum pipe vlv_find_free_pps(struct drm_i915_private *dev_priv) > > static enum pipe > > vlv_power_sequencer_pipe(struct intel_dp *intel_dp) > > { > > - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp); > > + struct drm_i915_private *i915 = dp_to_i915(intel_dp); > > struct intel_digital_port *intel_dig_port = dp_to_dig_port(intel_dp); > > enum pipe pipe; > > - lockdep_assert_held(&dev_priv->pps_mutex); > > + lockdep_assert_held(&i915->pps_mutex); > > /* We should never land here with regular DP ports */ > > WARN_ON(!intel_dp_is_edp(intel_dp)); > > @@ -783,7 +783,7 @@ vlv_power_sequencer_pipe(struct intel_dp *intel_dp) > > if (intel_dp->pps_pipe != INVALID_PIPE) > > return intel_dp->pps_pipe; > > - pipe = vlv_find_free_pps(dev_priv); > > + pipe = vlv_find_free_pps(i915); > > /* > > * Didn't find one. This should not happen since there > > @@ -792,7 +792,7 @@ vlv_power_sequencer_pipe(struct intel_dp *intel_dp) > > if (WARN_ON(pipe == INVALID_PIPE)) > > pipe = PIPE_A; > > - vlv_steal_power_sequencer(dev_priv, pipe); > > + vlv_steal_power_sequencer(i915, pipe); > > intel_dp->pps_pipe = pipe; > > DRM_DEBUG_KMS("picked pipe %c power sequencer for port %c\n", > > @@ -815,10 +815,10 @@ vlv_power_sequencer_pipe(struct intel_dp *intel_dp) > > static int > > bxt_power_sequencer_idx(struct intel_dp *intel_dp) > > { > > - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp); > > - int backlight_controller = dev_priv->vbt.backlight.controller; > > + struct drm_i915_private *i915 = dp_to_i915(intel_dp); > > + int backlight_controller = i915->vbt.backlight.controller; > > - lockdep_assert_held(&dev_priv->pps_mutex); > > + lockdep_assert_held(&i915->pps_mutex); > > /* We should never land here with regular DP ports */ > > WARN_ON(!intel_dp_is_edp(intel_dp)); > > @@ -837,29 +837,29 @@ bxt_power_sequencer_idx(struct intel_dp *intel_dp) > > return backlight_controller; > > } > > -typedef bool (*vlv_pipe_check)(struct drm_i915_private *dev_priv, > > +typedef bool (*vlv_pipe_check)(struct drm_i915_private *i915, > > enum pipe pipe); > > -static bool vlv_pipe_has_pp_on(struct drm_i915_private *dev_priv, > > +static bool vlv_pipe_has_pp_on(struct drm_i915_private *i915, > > enum pipe pipe) > > { > > return I915_READ(PP_STATUS(pipe)) & PP_ON; > > } > > -static bool vlv_pipe_has_vdd_on(struct drm_i915_private *dev_priv, > > +static bool vlv_pipe_has_vdd_on(struct drm_i915_private *i915, > > enum pipe pipe) > > { > > return I915_READ(PP_CONTROL(pipe)) & EDP_FORCE_VDD; > > } > > -static bool vlv_pipe_any(struct drm_i915_private *dev_priv, > > +static bool vlv_pipe_any(struct drm_i915_private *i915, > > enum pipe pipe) > > { > > return true; > > } > > static enum pipe > > -vlv_initial_pps_pipe(struct drm_i915_private *dev_priv, > > +vlv_initial_pps_pipe(struct drm_i915_private *i915, > > enum port port, > > vlv_pipe_check pipe_check) > > { > > @@ -872,7 +872,7 @@ vlv_initial_pps_pipe(struct drm_i915_private *dev_priv, > > if (port_sel != PANEL_PORT_SELECT_VLV(port)) > > continue; > > - if (!pipe_check(dev_priv, pipe)) > > + if (!pipe_check(i915, pipe)) > > continue; > > return pipe; > > @@ -884,23 +884,23 @@ vlv_initial_pps_pipe(struct drm_i915_private *dev_priv, > > static void > > vlv_initial_power_sequencer_setup(struct intel_dp *intel_dp) > > { > > - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp); > > + struct drm_i915_private *i915 = dp_to_i915(intel_dp); > > struct intel_digital_port *intel_dig_port = dp_to_dig_port(intel_dp); > > enum port port = intel_dig_port->base.port; > > - lockdep_assert_held(&dev_priv->pps_mutex); > > + lockdep_assert_held(&i915->pps_mutex); > > /* try to find a pipe with this port selected */ > > /* first pick one where the panel is on */ > > - intel_dp->pps_pipe = vlv_initial_pps_pipe(dev_priv, port, > > + intel_dp->pps_pipe = vlv_initial_pps_pipe(i915, port, > > vlv_pipe_has_pp_on); > > /* didn't find one? pick one where vdd is on */ > > if (intel_dp->pps_pipe == INVALID_PIPE) > > - intel_dp->pps_pipe = vlv_initial_pps_pipe(dev_priv, port, > > + intel_dp->pps_pipe = vlv_initial_pps_pipe(i915, port, > > vlv_pipe_has_vdd_on); > > /* didn't find one? pick one with just the correct port */ > > if (intel_dp->pps_pipe == INVALID_PIPE) > > - intel_dp->pps_pipe = vlv_initial_pps_pipe(dev_priv, port, > > + intel_dp->pps_pipe = vlv_initial_pps_pipe(i915, port, > > vlv_pipe_any); > > /* didn't find one? just let vlv_power_sequencer_pipe() pick one when needed */ > > @@ -917,12 +917,12 @@ vlv_initial_power_sequencer_setup(struct intel_dp *intel_dp) > > intel_dp_init_panel_power_sequencer_registers(intel_dp, false); > > } > > -void intel_power_sequencer_reset(struct drm_i915_private *dev_priv) > > +void intel_power_sequencer_reset(struct drm_i915_private *i915) > > { > > struct intel_encoder *encoder; > > - if (WARN_ON(!IS_VALLEYVIEW(dev_priv) && !IS_CHERRYVIEW(dev_priv) && > > - !IS_GEN9_LP(dev_priv))) > > + if (WARN_ON(!IS_VALLEYVIEW(i915) && !IS_CHERRYVIEW(i915) && > > + !IS_GEN9_LP(i915))) > > return; > > /* > > @@ -935,7 +935,7 @@ void intel_power_sequencer_reset(struct drm_i915_private *dev_priv) > > * should use them always. > > */ > > - for_each_intel_dp(&dev_priv->drm, encoder) { > > + for_each_intel_dp(&i915->drm, encoder) { > > struct intel_dp *intel_dp = enc_to_intel_dp(&encoder->base); > > WARN_ON(intel_dp->active_pipe != INVALID_PIPE); > > @@ -943,7 +943,7 @@ void intel_power_sequencer_reset(struct drm_i915_private *dev_priv) > > if (encoder->type != INTEL_OUTPUT_EDP) > > continue; > > - if (IS_GEN9_LP(dev_priv)) > > + if (IS_GEN9_LP(i915)) > > intel_dp->pps_reset = true; > > else > > intel_dp->pps_pipe = INVALID_PIPE; > > @@ -961,14 +961,14 @@ struct pps_registers { > > static void intel_pps_get_registers(struct intel_dp *intel_dp, > > struct pps_registers *regs) > > { > > - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp); > > + struct drm_i915_private *i915 = dp_to_i915(intel_dp); > > int pps_idx = 0; > > memset(regs, 0, sizeof(*regs)); > > - if (IS_GEN9_LP(dev_priv)) > > + if (IS_GEN9_LP(i915)) > > pps_idx = bxt_power_sequencer_idx(intel_dp); > > - else if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) > > + else if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) > > pps_idx = vlv_power_sequencer_pipe(intel_dp); > > regs->pp_ctrl = PP_CONTROL(pps_idx); > > @@ -977,7 +977,7 @@ static void intel_pps_get_registers(struct intel_dp *intel_dp, > > regs->pp_off = PP_OFF_DELAYS(pps_idx); > > /* Cycle delay moved from PP_DIVISOR to PP_CONTROL */ > > - if (IS_GEN9_LP(dev_priv) || INTEL_PCH_TYPE(dev_priv) >= PCH_CNP) > > + if (IS_GEN9_LP(i915) || INTEL_PCH_TYPE(i915) >= PCH_CNP) > > regs->pp_div = INVALID_MMIO_REG; > > else > > regs->pp_div = PP_DIVISOR(pps_idx); > > @@ -1010,14 +1010,14 @@ static int edp_notify_handler(struct notifier_block *this, unsigned long code, > > { > > struct intel_dp *intel_dp = container_of(this, typeof(* intel_dp), > > edp_notifier); > > - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp); > > + struct drm_i915_private *i915 = dp_to_i915(intel_dp); > > intel_wakeref_t wakeref; > > if (!intel_dp_is_edp(intel_dp) || code != SYS_RESTART) > > return 0; > > with_pps_lock(intel_dp, wakeref) { > > - if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) { > > + if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) { > > enum pipe pipe = vlv_power_sequencer_pipe(intel_dp); > > i915_reg_t pp_ctrl_reg, pp_div_reg; > > u32 pp_div; > > @@ -1039,11 +1039,11 @@ static int edp_notify_handler(struct notifier_block *this, unsigned long code, > > static bool edp_have_panel_power(struct intel_dp *intel_dp) > > { > > - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp); > > + struct drm_i915_private *i915 = dp_to_i915(intel_dp); > > - lockdep_assert_held(&dev_priv->pps_mutex); > > + lockdep_assert_held(&i915->pps_mutex); > > - if ((IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) && > > + if ((IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) && > > intel_dp->pps_pipe == INVALID_PIPE) > > return false; > > @@ -1052,11 +1052,11 @@ static bool edp_have_panel_power(struct intel_dp *intel_dp) > > static bool edp_have_panel_vdd(struct intel_dp *intel_dp) > > { > > - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp); > > + struct drm_i915_private *i915 = dp_to_i915(intel_dp); > > - lockdep_assert_held(&dev_priv->pps_mutex); > > + lockdep_assert_held(&i915->pps_mutex); > > - if ((IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) && > > + if ((IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) && > > intel_dp->pps_pipe == INVALID_PIPE) > > return false; > > @@ -1066,7 +1066,7 @@ static bool edp_have_panel_vdd(struct intel_dp *intel_dp) > > static void > > intel_dp_check_edp(struct intel_dp *intel_dp) > > { > > - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp); > > + struct drm_i915_private *i915 = dp_to_i915(intel_dp); > > if (!intel_dp_is_edp(intel_dp)) > > return; > > @@ -1103,7 +1103,7 @@ intel_dp_aux_wait_done(struct intel_dp *intel_dp) > > static u32 g4x_get_aux_clock_divider(struct intel_dp *intel_dp, int index) > > { > > - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp); > > + struct drm_i915_private *i915 = dp_to_i915(intel_dp); > > if (index) > > return 0; > > @@ -1112,12 +1112,12 @@ static u32 g4x_get_aux_clock_divider(struct intel_dp *intel_dp, int index) > > * The clock divider is based off the hrawclk, and would like to run at > > * 2MHz. So, take the hrawclk value and divide by 2000 and use that > > */ > > - return DIV_ROUND_CLOSEST(dev_priv->rawclk_freq, 2000); > > + return DIV_ROUND_CLOSEST(i915->rawclk_freq, 2000); > > } > > static u32 ilk_get_aux_clock_divider(struct intel_dp *intel_dp, int index) > > { > > - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp); > > + struct drm_i915_private *i915 = dp_to_i915(intel_dp); > > struct intel_digital_port *dig_port = dp_to_dig_port(intel_dp); > > if (index) > > @@ -1129,17 +1129,17 @@ static u32 ilk_get_aux_clock_divider(struct intel_dp *intel_dp, int index) > > * divide by 2000 and use that > > */ > > if (dig_port->aux_ch == AUX_CH_A) > > - return DIV_ROUND_CLOSEST(dev_priv->cdclk.hw.cdclk, 2000); > > + return DIV_ROUND_CLOSEST(i915->cdclk.hw.cdclk, 2000); > > else > > - return DIV_ROUND_CLOSEST(dev_priv->rawclk_freq, 2000); > > + return DIV_ROUND_CLOSEST(i915->rawclk_freq, 2000); > > } > > static u32 hsw_get_aux_clock_divider(struct intel_dp *intel_dp, int index) > > { > > - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp); > > + struct drm_i915_private *i915 = dp_to_i915(intel_dp); > > struct intel_digital_port *dig_port = dp_to_dig_port(intel_dp); > > - if (dig_port->aux_ch != AUX_CH_A && HAS_PCH_LPT_H(dev_priv)) { > > + if (dig_port->aux_ch != AUX_CH_A && HAS_PCH_LPT_H(i915)) { > > /* Workaround for non-ULT HSW */ > > switch (index) { > > case 0: return 63; > > @@ -1166,16 +1166,16 @@ static u32 g4x_get_aux_send_ctl(struct intel_dp *intel_dp, > > u32 aux_clock_divider) > > { > > struct intel_digital_port *intel_dig_port = dp_to_dig_port(intel_dp); > > - struct drm_i915_private *dev_priv = > > + struct drm_i915_private *i915 = > > to_i915(intel_dig_port->base.base.dev); > > u32 precharge, timeout; > > - if (IS_GEN(dev_priv, 6)) > > + if (IS_GEN(i915, 6)) > > precharge = 3; > > else > > precharge = 5; > > - if (IS_BROADWELL(dev_priv)) > > + if (IS_BROADWELL(i915)) > > timeout = DP_AUX_CH_CTL_TIME_OUT_600us; > > else > > timeout = DP_AUX_CH_CTL_TIME_OUT_400us; > > @@ -1481,7 +1481,7 @@ intel_dp_aux_transfer(struct drm_dp_aux *aux, struct drm_dp_aux_msg *msg) > > static i915_reg_t g4x_aux_ctl_reg(struct intel_dp *intel_dp) > > { > > - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp); > > + struct drm_i915_private *i915 = dp_to_i915(intel_dp); > > struct intel_digital_port *dig_port = dp_to_dig_port(intel_dp); > > enum aux_ch aux_ch = dig_port->aux_ch; > > @@ -1498,7 +1498,7 @@ static i915_reg_t g4x_aux_ctl_reg(struct intel_dp *intel_dp) > > static i915_reg_t g4x_aux_data_reg(struct intel_dp *intel_dp, int index) > > { > > - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp); > > + struct drm_i915_private *i915 = dp_to_i915(intel_dp); > > struct intel_digital_port *dig_port = dp_to_dig_port(intel_dp); > > enum aux_ch aux_ch = dig_port->aux_ch; > > @@ -1515,7 +1515,7 @@ static i915_reg_t g4x_aux_data_reg(struct intel_dp *intel_dp, int index) > > static i915_reg_t ilk_aux_ctl_reg(struct intel_dp *intel_dp) > > { > > - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp); > > + struct drm_i915_private *i915 = dp_to_i915(intel_dp); > > struct intel_digital_port *dig_port = dp_to_dig_port(intel_dp); > > enum aux_ch aux_ch = dig_port->aux_ch; > > @@ -1534,7 +1534,7 @@ static i915_reg_t ilk_aux_ctl_reg(struct intel_dp *intel_dp) > > static i915_reg_t ilk_aux_data_reg(struct intel_dp *intel_dp, int index) > > { > > - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp); > > + struct drm_i915_private *i915 = dp_to_i915(intel_dp); > > struct intel_digital_port *dig_port = dp_to_dig_port(intel_dp); > > enum aux_ch aux_ch = dig_port->aux_ch; > > @@ -1553,7 +1553,7 @@ static i915_reg_t ilk_aux_data_reg(struct intel_dp *intel_dp, int index) > > static i915_reg_t skl_aux_ctl_reg(struct intel_dp *intel_dp) > > { > > - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp); > > + struct drm_i915_private *i915 = dp_to_i915(intel_dp); > > struct intel_digital_port *dig_port = dp_to_dig_port(intel_dp); > > enum aux_ch aux_ch = dig_port->aux_ch; > > @@ -1573,7 +1573,7 @@ static i915_reg_t skl_aux_ctl_reg(struct intel_dp *intel_dp) > > static i915_reg_t skl_aux_data_reg(struct intel_dp *intel_dp, int index) > > { > > - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp); > > + struct drm_i915_private *i915 = dp_to_i915(intel_dp); > > struct intel_digital_port *dig_port = dp_to_dig_port(intel_dp); > > enum aux_ch aux_ch = dig_port->aux_ch; > > @@ -1600,14 +1600,14 @@ intel_dp_aux_fini(struct intel_dp *intel_dp) > > static void > > intel_dp_aux_init(struct intel_dp *intel_dp) > > { > > - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp); > > + struct drm_i915_private *i915 = dp_to_i915(intel_dp); > > struct intel_digital_port *dig_port = dp_to_dig_port(intel_dp); > > struct intel_encoder *encoder = &dig_port->base; > > - if (INTEL_GEN(dev_priv) >= 9) { > > + if (INTEL_GEN(i915) >= 9) { > > intel_dp->aux_ch_ctl_reg = skl_aux_ctl_reg; > > intel_dp->aux_ch_data_reg = skl_aux_data_reg; > > - } else if (HAS_PCH_SPLIT(dev_priv)) { > > + } else if (HAS_PCH_SPLIT(i915)) { > > intel_dp->aux_ch_ctl_reg = ilk_aux_ctl_reg; > > intel_dp->aux_ch_data_reg = ilk_aux_data_reg; > > } else { > > @@ -1615,16 +1615,16 @@ intel_dp_aux_init(struct intel_dp *intel_dp) > > intel_dp->aux_ch_data_reg = g4x_aux_data_reg; > > } > > - if (INTEL_GEN(dev_priv) >= 9) > > + if (INTEL_GEN(i915) >= 9) > > intel_dp->get_aux_clock_divider = skl_get_aux_clock_divider; > > - else if (IS_BROADWELL(dev_priv) || IS_HASWELL(dev_priv)) > > + else if (IS_BROADWELL(i915) || IS_HASWELL(i915)) > > intel_dp->get_aux_clock_divider = hsw_get_aux_clock_divider; > > - else if (HAS_PCH_SPLIT(dev_priv)) > > + else if (HAS_PCH_SPLIT(i915)) > > intel_dp->get_aux_clock_divider = ilk_get_aux_clock_divider; > > else > > intel_dp->get_aux_clock_divider = g4x_get_aux_clock_divider; > > - if (INTEL_GEN(dev_priv) >= 9) > > + if (INTEL_GEN(i915) >= 9) > > intel_dp->get_aux_send_ctl = skl_get_aux_send_ctl; > > else > > intel_dp->get_aux_send_ctl = g4x_get_aux_send_ctl; > > @@ -1655,20 +1655,20 @@ static void > > intel_dp_set_clock(struct intel_encoder *encoder, > > struct intel_crtc_state *pipe_config) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > const struct dp_link_dpll *divisor = NULL; > > int i, count = 0; > > - if (IS_G4X(dev_priv)) { > > + if (IS_G4X(i915)) { > > divisor = g4x_dpll; > > count = ARRAY_SIZE(g4x_dpll); > > - } else if (HAS_PCH_SPLIT(dev_priv)) { > > + } else if (HAS_PCH_SPLIT(i915)) { > > divisor = pch_dpll; > > count = ARRAY_SIZE(pch_dpll); > > - } else if (IS_CHERRYVIEW(dev_priv)) { > > + } else if (IS_CHERRYVIEW(i915)) { > > divisor = chv_dpll; > > count = ARRAY_SIZE(chv_dpll); > > - } else if (IS_VALLEYVIEW(dev_priv)) { > > + } else if (IS_VALLEYVIEW(i915)) { > > divisor = vlv_dpll; > > count = ARRAY_SIZE(vlv_dpll); > > } > > @@ -1760,9 +1760,9 @@ void intel_dp_compute_rate(struct intel_dp *intel_dp, int port_clock, > > static bool intel_dp_source_supports_fec(struct intel_dp *intel_dp, > > const struct intel_crtc_state *pipe_config) > > { > > - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp); > > + struct drm_i915_private *i915 = dp_to_i915(intel_dp); > > - return INTEL_GEN(dev_priv) >= 11 && > > + return INTEL_GEN(i915) >= 11 && > > pipe_config->cpu_transcoder != TRANSCODER_A; > > } > > @@ -1776,9 +1776,9 @@ static bool intel_dp_supports_fec(struct intel_dp *intel_dp, > > static bool intel_dp_source_supports_dsc(struct intel_dp *intel_dp, > > const struct intel_crtc_state *pipe_config) > > { > > - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp); > > + struct drm_i915_private *i915 = dp_to_i915(intel_dp); > > - return INTEL_GEN(dev_priv) >= 10 && > > + return INTEL_GEN(i915) >= 10 && > > pipe_config->cpu_transcoder != TRANSCODER_A; > > } > > @@ -1795,7 +1795,7 @@ static bool intel_dp_supports_dsc(struct intel_dp *intel_dp, > > static int intel_dp_compute_bpp(struct intel_dp *intel_dp, > > struct intel_crtc_state *pipe_config) > > { > > - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp); > > + struct drm_i915_private *i915 = dp_to_i915(intel_dp); > > struct intel_connector *intel_connector = intel_dp->attached_connector; > > int bpp, bpc; > > @@ -1808,10 +1808,10 @@ static int intel_dp_compute_bpp(struct intel_dp *intel_dp, > > if (intel_dp_is_edp(intel_dp)) { > > /* Get bpp from vbt only for panels that dont have bpp in edid */ > > if (intel_connector->base.display_info.bpc == 0 && > > - dev_priv->vbt.edp.bpp && dev_priv->vbt.edp.bpp < bpp) { > > + i915->vbt.edp.bpp && i915->vbt.edp.bpp < bpp) { > > DRM_DEBUG_KMS("clamping bpp for eDP panel to BIOS-provided %i\n", > > - dev_priv->vbt.edp.bpp); > > - bpp = dev_priv->vbt.edp.bpp; > > + i915->vbt.edp.bpp); > > + bpp = i915->vbt.edp.bpp; > > } > > } > > @@ -1924,7 +1924,7 @@ static int intel_dp_dsc_compute_config(struct intel_dp *intel_dp, > > struct link_config_limits *limits) > > { > > struct intel_digital_port *dig_port = dp_to_dig_port(intel_dp); > > - struct drm_i915_private *dev_priv = to_i915(dig_port->base.base.dev); > > + struct drm_i915_private *i915 = to_i915(dig_port->base.base.dev); > > struct drm_display_mode *adjusted_mode = &pipe_config->base.adjusted_mode; > > u8 dsc_max_bpc; > > int pipe_bpp; > > @@ -1988,7 +1988,7 @@ static int intel_dp_dsc_compute_config(struct intel_dp *intel_dp, > > * is greater than the maximum Cdclock and if slice count is even > > * then we need to use 2 VDSC instances. > > */ > > - if (adjusted_mode->crtc_clock > dev_priv->max_cdclk_freq) { > > + if (adjusted_mode->crtc_clock > i915->max_cdclk_freq) { > > if (pipe_config->dsc_params.slice_count > 1) { > > pipe_config->dsc_params.dsc_split = true; > > } else { > > @@ -2168,7 +2168,7 @@ intel_dp_compute_config(struct intel_encoder *encoder, > > struct intel_crtc_state *pipe_config, > > struct drm_connector_state *conn_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct drm_display_mode *adjusted_mode = &pipe_config->base.adjusted_mode; > > struct intel_dp *intel_dp = enc_to_intel_dp(&encoder->base); > > struct intel_lspcon *lspcon = enc_to_intel_lspcon(&encoder->base); > > @@ -2181,7 +2181,7 @@ intel_dp_compute_config(struct intel_encoder *encoder, > > DP_DPCD_QUIRK_CONSTANT_N); > > int ret = 0, output_bpp; > > - if (HAS_PCH_SPLIT(dev_priv) && !HAS_DDI(dev_priv) && port != PORT_A) > > + if (HAS_PCH_SPLIT(i915) && !HAS_DDI(i915) && port != PORT_A) > > pipe_config->has_pch_encoder = true; > > pipe_config->output_format = INTEL_OUTPUT_FORMAT_RGB; > > @@ -2195,7 +2195,7 @@ intel_dp_compute_config(struct intel_encoder *encoder, > > return ret; > > pipe_config->has_drrs = false; > > - if (IS_G4X(dev_priv) || port == PORT_A) > > + if (IS_G4X(i915) || port == PORT_A) > > pipe_config->has_audio = false; > > else if (intel_conn_state->force_audio == HDMI_AUDIO_AUTO) > > pipe_config->has_audio = intel_dp->has_audio; > > @@ -2206,13 +2206,13 @@ intel_dp_compute_config(struct intel_encoder *encoder, > > intel_fixed_panel_mode(intel_connector->panel.fixed_mode, > > adjusted_mode); > > - if (INTEL_GEN(dev_priv) >= 9) { > > + if (INTEL_GEN(i915) >= 9) { > > ret = skl_update_scaler_crtc(pipe_config); > > if (ret) > > return ret; > > } > > - if (HAS_GMCH(dev_priv)) > > + if (HAS_GMCH(i915)) > > intel_gmch_panel_fitting(intel_crtc, pipe_config, > > conn_state->scaling_mode); > > else > > @@ -2223,7 +2223,7 @@ intel_dp_compute_config(struct intel_encoder *encoder, > > if (adjusted_mode->flags & DRM_MODE_FLAG_DBLSCAN) > > return -EINVAL; > > - if (HAS_GMCH(dev_priv) && > > + if (HAS_GMCH(i915) && > > adjusted_mode->flags & DRM_MODE_FLAG_INTERLACE) > > return -EINVAL; > > @@ -2250,7 +2250,7 @@ intel_dp_compute_config(struct intel_encoder *encoder, > > constant_n); > > if (intel_connector->panel.downclock_mode != NULL && > > - dev_priv->drrs.type == SEAMLESS_DRRS_SUPPORT) { > > + i915->drrs.type == SEAMLESS_DRRS_SUPPORT) { > > pipe_config->has_drrs = true; > > intel_link_compute_m_n(output_bpp, > > pipe_config->lane_count, > > @@ -2260,7 +2260,7 @@ intel_dp_compute_config(struct intel_encoder *encoder, > > constant_n); > > } > > - if (!HAS_DDI(dev_priv)) > > + if (!HAS_DDI(i915)) > > intel_dp_set_clock(encoder, pipe_config); > > intel_psr_compute_config(intel_dp, pipe_config); > > @@ -2281,7 +2281,7 @@ void intel_dp_set_link_params(struct intel_dp *intel_dp, > > static void intel_dp_prepare(struct intel_encoder *encoder, > > const struct intel_crtc_state *pipe_config) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_dp *intel_dp = enc_to_intel_dp(&encoder->base); > > enum port port = encoder->port; > > struct intel_crtc *crtc = to_intel_crtc(pipe_config->base.crtc); > > @@ -2320,7 +2320,7 @@ static void intel_dp_prepare(struct intel_encoder *encoder, > > /* Split out the IBX/CPU vs CPT settings */ > > - if (IS_IVYBRIDGE(dev_priv) && port == PORT_A) { > > + if (IS_IVYBRIDGE(i915) && port == PORT_A) { > > if (adjusted_mode->flags & DRM_MODE_FLAG_PHSYNC) > > intel_dp->DP |= DP_SYNC_HS_HIGH; > > if (adjusted_mode->flags & DRM_MODE_FLAG_PVSYNC) > > @@ -2331,7 +2331,7 @@ static void intel_dp_prepare(struct intel_encoder *encoder, > > intel_dp->DP |= DP_ENHANCED_FRAMING; > > intel_dp->DP |= DP_PIPE_SEL_IVB(crtc->pipe); > > - } else if (HAS_PCH_CPT(dev_priv) && port != PORT_A) { > > + } else if (HAS_PCH_CPT(i915) && port != PORT_A) { > > u32 trans_dp; > > intel_dp->DP |= DP_LINK_TRAIN_OFF_CPT; > > @@ -2343,7 +2343,7 @@ static void intel_dp_prepare(struct intel_encoder *encoder, > > trans_dp &= ~TRANS_DP_ENH_FRAMING; > > I915_WRITE(TRANS_DP_CTL(crtc->pipe), trans_dp); > > } else { > > - if (IS_G4X(dev_priv) && pipe_config->limited_color_range) > > + if (IS_G4X(i915) && pipe_config->limited_color_range) > > intel_dp->DP |= DP_COLOR_RANGE_16_235; > > if (adjusted_mode->flags & DRM_MODE_FLAG_PHSYNC) > > @@ -2355,7 +2355,7 @@ static void intel_dp_prepare(struct intel_encoder *encoder, > > if (drm_dp_enhanced_frame_cap(intel_dp->dpcd)) > > intel_dp->DP |= DP_ENHANCED_FRAMING; > > - if (IS_CHERRYVIEW(dev_priv)) > > + if (IS_CHERRYVIEW(i915)) > > intel_dp->DP |= DP_PIPE_SEL_CHV(crtc->pipe); > > else > > intel_dp->DP |= DP_PIPE_SEL(crtc->pipe); > > @@ -2377,10 +2377,10 @@ static void wait_panel_status(struct intel_dp *intel_dp, > > u32 mask, > > u32 value) > > { > > - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp); > > + struct drm_i915_private *i915 = dp_to_i915(intel_dp); > > i915_reg_t pp_stat_reg, pp_ctrl_reg; > > - lockdep_assert_held(&dev_priv->pps_mutex); > > + lockdep_assert_held(&i915->pps_mutex); > > intel_pps_verify_state(intel_dp); > > @@ -2392,7 +2392,7 @@ static void wait_panel_status(struct intel_dp *intel_dp, > > I915_READ(pp_stat_reg), > > I915_READ(pp_ctrl_reg)); > > - if (intel_wait_for_register(&dev_priv->uncore, > > + if (intel_wait_for_register(&i915->uncore, > > pp_stat_reg, mask, value, > > 5000)) > > DRM_ERROR("Panel status timeout: status %08x control %08x\n", > > @@ -2453,13 +2453,13 @@ static void edp_wait_backlight_off(struct intel_dp *intel_dp) > > static u32 ironlake_get_pp_control(struct intel_dp *intel_dp) > > { > > - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp); > > + struct drm_i915_private *i915 = dp_to_i915(intel_dp); > > u32 control; > > - lockdep_assert_held(&dev_priv->pps_mutex); > > + lockdep_assert_held(&i915->pps_mutex); > > control = I915_READ(_pp_ctrl_reg(intel_dp)); > > - if (WARN_ON(!HAS_DDI(dev_priv) && > > + if (WARN_ON(!HAS_DDI(i915) && > > (control & PANEL_UNLOCK_MASK) != PANEL_UNLOCK_REGS)) { > > control &= ~PANEL_UNLOCK_MASK; > > control |= PANEL_UNLOCK_REGS; > > @@ -2474,13 +2474,13 @@ static u32 ironlake_get_pp_control(struct intel_dp *intel_dp) > > */ > > static bool edp_panel_vdd_on(struct intel_dp *intel_dp) > > { > > - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp); > > + struct drm_i915_private *i915 = dp_to_i915(intel_dp); > > struct intel_digital_port *intel_dig_port = dp_to_dig_port(intel_dp); > > u32 pp; > > i915_reg_t pp_stat_reg, pp_ctrl_reg; > > bool need_to_disable = !intel_dp->want_panel_vdd; > > - lockdep_assert_held(&dev_priv->pps_mutex); > > + lockdep_assert_held(&i915->pps_mutex); > > if (!intel_dp_is_edp(intel_dp)) > > return false; > > @@ -2491,7 +2491,7 @@ static bool edp_panel_vdd_on(struct intel_dp *intel_dp) > > if (edp_have_panel_vdd(intel_dp)) > > return need_to_disable; > > - intel_display_power_get(dev_priv, > > + intel_display_power_get(i915, > > intel_aux_power_domain(intel_dig_port)); > > DRM_DEBUG_KMS("Turning eDP port %c VDD on\n", > > @@ -2546,13 +2546,13 @@ void intel_edp_panel_vdd_on(struct intel_dp *intel_dp) > > static void edp_panel_vdd_off_sync(struct intel_dp *intel_dp) > > { > > - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp); > > + struct drm_i915_private *i915 = dp_to_i915(intel_dp); > > struct intel_digital_port *intel_dig_port = > > dp_to_dig_port(intel_dp); > > u32 pp; > > i915_reg_t pp_stat_reg, pp_ctrl_reg; > > - lockdep_assert_held(&dev_priv->pps_mutex); > > + lockdep_assert_held(&i915->pps_mutex); > > WARN_ON(intel_dp->want_panel_vdd); > > @@ -2578,7 +2578,7 @@ static void edp_panel_vdd_off_sync(struct intel_dp *intel_dp) > > if ((pp & PANEL_POWER_ON) == 0) > > intel_dp->panel_power_off_time = ktime_get_boottime(); > > - intel_display_power_put_unchecked(dev_priv, > > + intel_display_power_put_unchecked(i915, > > intel_aux_power_domain(intel_dig_port)); > > } > > @@ -2615,9 +2615,9 @@ static void edp_panel_vdd_schedule_off(struct intel_dp *intel_dp) > > */ > > static void edp_panel_vdd_off(struct intel_dp *intel_dp, bool sync) > > { > > - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp); > > + struct drm_i915_private *i915 = dp_to_i915(intel_dp); > > - lockdep_assert_held(&dev_priv->pps_mutex); > > + lockdep_assert_held(&i915->pps_mutex); > > if (!intel_dp_is_edp(intel_dp)) > > return; > > @@ -2635,11 +2635,11 @@ static void edp_panel_vdd_off(struct intel_dp *intel_dp, bool sync) > > static void edp_panel_on(struct intel_dp *intel_dp) > > { > > - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp); > > + struct drm_i915_private *i915 = dp_to_i915(intel_dp); > > u32 pp; > > i915_reg_t pp_ctrl_reg; > > - lockdep_assert_held(&dev_priv->pps_mutex); > > + lockdep_assert_held(&i915->pps_mutex); > > if (!intel_dp_is_edp(intel_dp)) > > return; > > @@ -2656,7 +2656,7 @@ static void edp_panel_on(struct intel_dp *intel_dp) > > pp_ctrl_reg = _pp_ctrl_reg(intel_dp); > > pp = ironlake_get_pp_control(intel_dp); > > - if (IS_GEN(dev_priv, 5)) { > > + if (IS_GEN(i915, 5)) { > > /* ILK workaround: disable reset around power sequence */ > > pp &= ~PANEL_POWER_RESET; > > I915_WRITE(pp_ctrl_reg, pp); > > @@ -2664,7 +2664,7 @@ static void edp_panel_on(struct intel_dp *intel_dp) > > } > > pp |= PANEL_POWER_ON; > > - if (!IS_GEN(dev_priv, 5)) > > + if (!IS_GEN(i915, 5)) > > pp |= PANEL_POWER_RESET; > > I915_WRITE(pp_ctrl_reg, pp); > > @@ -2673,7 +2673,7 @@ static void edp_panel_on(struct intel_dp *intel_dp) > > wait_panel_on(intel_dp); > > intel_dp->last_power_on = jiffies; > > - if (IS_GEN(dev_priv, 5)) { > > + if (IS_GEN(i915, 5)) { > > pp |= PANEL_POWER_RESET; /* restore panel reset bit */ > > I915_WRITE(pp_ctrl_reg, pp); > > POSTING_READ(pp_ctrl_reg); > > @@ -2694,12 +2694,12 @@ void intel_edp_panel_on(struct intel_dp *intel_dp) > > static void edp_panel_off(struct intel_dp *intel_dp) > > { > > - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp); > > + struct drm_i915_private *i915 = dp_to_i915(intel_dp); > > struct intel_digital_port *dig_port = dp_to_dig_port(intel_dp); > > u32 pp; > > i915_reg_t pp_ctrl_reg; > > - lockdep_assert_held(&dev_priv->pps_mutex); > > + lockdep_assert_held(&i915->pps_mutex); > > if (!intel_dp_is_edp(intel_dp)) > > return; > > @@ -2727,7 +2727,7 @@ static void edp_panel_off(struct intel_dp *intel_dp) > > intel_dp->panel_power_off_time = ktime_get_boottime(); > > /* We got a reference when we enabled the VDD. */ > > - intel_display_power_put_unchecked(dev_priv, intel_aux_power_domain(dig_port)); > > + intel_display_power_put_unchecked(i915, intel_aux_power_domain(dig_port)); > > } > > void intel_edp_panel_off(struct intel_dp *intel_dp) > > @@ -2744,7 +2744,7 @@ void intel_edp_panel_off(struct intel_dp *intel_dp) > > /* Enable backlight in the panel power control. */ > > static void _intel_edp_backlight_on(struct intel_dp *intel_dp) > > { > > - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp); > > + struct drm_i915_private *i915 = dp_to_i915(intel_dp); > > intel_wakeref_t wakeref; > > /* > > @@ -2785,7 +2785,7 @@ void intel_edp_backlight_on(const struct intel_crtc_state *crtc_state, > > /* Disable backlight in the panel power control. */ > > static void _intel_edp_backlight_off(struct intel_dp *intel_dp) > > { > > - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp); > > + struct drm_i915_private *i915 = dp_to_i915(intel_dp); > > intel_wakeref_t wakeref; > > if (!intel_dp_is_edp(intel_dp)) > > @@ -2849,7 +2849,7 @@ static void intel_edp_backlight_power(struct intel_connector *connector, > > static void assert_dp_port(struct intel_dp *intel_dp, bool state) > > { > > struct intel_digital_port *dig_port = dp_to_dig_port(intel_dp); > > - struct drm_i915_private *dev_priv = to_i915(dig_port->base.base.dev); > > + struct drm_i915_private *i915 = to_i915(dig_port->base.base.dev); > > bool cur_state = I915_READ(intel_dp->output_reg) & DP_PORT_EN; > > I915_STATE_WARN(cur_state != state, > > @@ -2859,7 +2859,7 @@ static void assert_dp_port(struct intel_dp *intel_dp, bool state) > > } > > #define assert_dp_port_disabled(d) assert_dp_port((d), false) > > -static void assert_edp_pll(struct drm_i915_private *dev_priv, bool state) > > +static void assert_edp_pll(struct drm_i915_private *i915, bool state) > > { > > bool cur_state = I915_READ(DP_A) & DP_PLL_ENABLE; > > @@ -2874,11 +2874,11 @@ static void ironlake_edp_pll_on(struct intel_dp *intel_dp, > > const struct intel_crtc_state *pipe_config) > > { > > struct intel_crtc *crtc = to_intel_crtc(pipe_config->base.crtc); > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > - assert_pipe_disabled(dev_priv, crtc->pipe); > > + assert_pipe_disabled(i915, crtc->pipe); > > assert_dp_port_disabled(intel_dp); > > - assert_edp_pll_disabled(dev_priv); > > + assert_edp_pll_disabled(i915); > > DRM_DEBUG_KMS("enabling eDP PLL for clock %d\n", > > pipe_config->port_clock); > > @@ -2900,8 +2900,8 @@ static void ironlake_edp_pll_on(struct intel_dp *intel_dp, > > * 1. Wait for the start of vertical blank on the enabled pipe going to FDI > > * 2. Program DP PLL enable > > */ > > - if (IS_GEN(dev_priv, 5)) > > - intel_wait_for_vblank_if_active(dev_priv, !crtc->pipe); > > + if (IS_GEN(i915, 5)) > > + intel_wait_for_vblank_if_active(i915, !crtc->pipe); > > intel_dp->DP |= DP_PLL_ENABLE; > > @@ -2914,11 +2914,11 @@ static void ironlake_edp_pll_off(struct intel_dp *intel_dp, > > const struct intel_crtc_state *old_crtc_state) > > { > > struct intel_crtc *crtc = to_intel_crtc(old_crtc_state->base.crtc); > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > - assert_pipe_disabled(dev_priv, crtc->pipe); > > + assert_pipe_disabled(i915, crtc->pipe); > > assert_dp_port_disabled(intel_dp); > > - assert_edp_pll_enabled(dev_priv); > > + assert_edp_pll_enabled(i915); > > DRM_DEBUG_KMS("disabling eDP PLL\n"); > > @@ -2999,12 +2999,12 @@ void intel_dp_sink_dpms(struct intel_dp *intel_dp, int mode) > > mode == DRM_MODE_DPMS_ON ? "enable" : "disable"); > > } > > -static bool cpt_dp_port_selected(struct drm_i915_private *dev_priv, > > +static bool cpt_dp_port_selected(struct drm_i915_private *i915, > > enum port port, enum pipe *pipe) > > { > > enum pipe p; > > - for_each_pipe(dev_priv, p) { > > + for_each_pipe(i915, p) { > > u32 val = I915_READ(TRANS_DP_CTL(p)); > > if ((val & TRANS_DP_PORT_SEL_MASK) == TRANS_DP_PORT_SEL(port)) { > > @@ -3021,7 +3021,7 @@ static bool cpt_dp_port_selected(struct drm_i915_private *dev_priv, > > return false; > > } > > -bool intel_dp_port_enabled(struct drm_i915_private *dev_priv, > > +bool intel_dp_port_enabled(struct drm_i915_private *i915, > > i915_reg_t dp_reg, enum port port, > > enum pipe *pipe) > > { > > @@ -3033,11 +3033,11 @@ bool intel_dp_port_enabled(struct drm_i915_private *dev_priv, > > ret = val & DP_PORT_EN; > > /* asserts want to know the pipe even if the port is disabled */ > > - if (IS_IVYBRIDGE(dev_priv) && port == PORT_A) > > + if (IS_IVYBRIDGE(i915) && port == PORT_A) > > *pipe = (val & DP_PIPE_SEL_MASK_IVB) >> DP_PIPE_SEL_SHIFT_IVB; > > - else if (HAS_PCH_CPT(dev_priv) && port != PORT_A) > > - ret &= cpt_dp_port_selected(dev_priv, port, pipe); > > - else if (IS_CHERRYVIEW(dev_priv)) > > + else if (HAS_PCH_CPT(i915) && port != PORT_A) > > + ret &= cpt_dp_port_selected(i915, port, pipe); > > + else if (IS_CHERRYVIEW(i915)) > > *pipe = (val & DP_PIPE_SEL_MASK_CHV) >> DP_PIPE_SEL_SHIFT_CHV; > > else > > *pipe = (val & DP_PIPE_SEL_MASK) >> DP_PIPE_SEL_SHIFT; > > @@ -3048,20 +3048,20 @@ bool intel_dp_port_enabled(struct drm_i915_private *dev_priv, > > static bool intel_dp_get_hw_state(struct intel_encoder *encoder, > > enum pipe *pipe) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_dp *intel_dp = enc_to_intel_dp(&encoder->base); > > intel_wakeref_t wakeref; > > bool ret; > > - wakeref = intel_display_power_get_if_enabled(dev_priv, > > + wakeref = intel_display_power_get_if_enabled(i915, > > encoder->power_domain); > > if (!wakeref) > > return false; > > - ret = intel_dp_port_enabled(dev_priv, intel_dp->output_reg, > > + ret = intel_dp_port_enabled(i915, intel_dp->output_reg, > > encoder->port, pipe); > > - intel_display_power_put(dev_priv, encoder->power_domain, wakeref); > > + intel_display_power_put(i915, encoder->power_domain, wakeref); > > return ret; > > } > > @@ -3069,7 +3069,7 @@ static bool intel_dp_get_hw_state(struct intel_encoder *encoder, > > static void intel_dp_get_config(struct intel_encoder *encoder, > > struct intel_crtc_state *pipe_config) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_dp *intel_dp = enc_to_intel_dp(&encoder->base); > > u32 tmp, flags = 0; > > enum port port = encoder->port; > > @@ -3084,7 +3084,7 @@ static void intel_dp_get_config(struct intel_encoder *encoder, > > pipe_config->has_audio = tmp & DP_AUDIO_OUTPUT_ENABLE && port != PORT_A; > > - if (HAS_PCH_CPT(dev_priv) && port != PORT_A) { > > + if (HAS_PCH_CPT(i915) && port != PORT_A) { > > u32 trans_dp = I915_READ(TRANS_DP_CTL(crtc->pipe)); > > if (trans_dp & TRANS_DP_HSYNC_ACTIVE_HIGH) > > @@ -3110,7 +3110,7 @@ static void intel_dp_get_config(struct intel_encoder *encoder, > > pipe_config->base.adjusted_mode.flags |= flags; > > - if (IS_G4X(dev_priv) && tmp & DP_COLOR_RANGE_16_235) > > + if (IS_G4X(i915) && tmp & DP_COLOR_RANGE_16_235) > > pipe_config->limited_color_range = true; > > pipe_config->lane_count = > > @@ -3129,8 +3129,8 @@ static void intel_dp_get_config(struct intel_encoder *encoder, > > intel_dotclock_calculate(pipe_config->port_clock, > > &pipe_config->dp_m_n); > > - if (intel_dp_is_edp(intel_dp) && dev_priv->vbt.edp.bpp && > > - pipe_config->pipe_bpp > dev_priv->vbt.edp.bpp) { > > + if (intel_dp_is_edp(intel_dp) && i915->vbt.edp.bpp && > > + pipe_config->pipe_bpp > i915->vbt.edp.bpp) { > > /* > > * This is a big fat ugly hack. > > * > > @@ -3145,8 +3145,8 @@ static void intel_dp_get_config(struct intel_encoder *encoder, > > * load. > > */ > > DRM_DEBUG_KMS("pipe has %d bpp for eDP panel, overriding BIOS-provided max %d bpp\n", > > - pipe_config->pipe_bpp, dev_priv->vbt.edp.bpp); > > - dev_priv->vbt.edp.bpp = pipe_config->pipe_bpp; > > + pipe_config->pipe_bpp, i915->vbt.edp.bpp); > > + i915->vbt.edp.bpp = pipe_config->pipe_bpp; > > } > > } > > @@ -3215,16 +3215,16 @@ static void chv_post_disable_dp(struct intel_encoder *encoder, > > const struct intel_crtc_state *old_crtc_state, > > const struct drm_connector_state *old_conn_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > intel_dp_link_down(encoder, old_crtc_state); > > - vlv_dpio_get(dev_priv); > > + vlv_dpio_get(i915); > > /* Assert data lane reset */ > > chv_data_lane_soft_reset(encoder, old_crtc_state, true); > > - vlv_dpio_put(dev_priv); > > + vlv_dpio_put(i915); > > } > > static void > > @@ -3232,7 +3232,7 @@ _intel_dp_set_link_train(struct intel_dp *intel_dp, > > u32 *DP, > > u8 dp_train_pat) > > { > > - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp); > > + struct drm_i915_private *i915 = dp_to_i915(intel_dp); > > struct intel_digital_port *intel_dig_port = dp_to_dig_port(intel_dp); > > enum port port = intel_dig_port->base.port; > > u8 train_pat_mask = drm_dp_training_pattern_mask(intel_dp->dpcd); > > @@ -3241,7 +3241,7 @@ _intel_dp_set_link_train(struct intel_dp *intel_dp, > > DRM_DEBUG_KMS("Using DP training pattern TPS%d\n", > > dp_train_pat & train_pat_mask); > > - if (HAS_DDI(dev_priv)) { > > + if (HAS_DDI(i915)) { > > u32 temp = I915_READ(DP_TP_CTL(port)); > > if (dp_train_pat & DP_LINK_SCRAMBLING_DISABLE) > > @@ -3270,8 +3270,8 @@ _intel_dp_set_link_train(struct intel_dp *intel_dp, > > } > > I915_WRITE(DP_TP_CTL(port), temp); > > - } else if ((IS_IVYBRIDGE(dev_priv) && port == PORT_A) || > > - (HAS_PCH_CPT(dev_priv) && port != PORT_A)) { > > + } else if ((IS_IVYBRIDGE(i915) && port == PORT_A) || > > + (HAS_PCH_CPT(i915) && port != PORT_A)) { > > *DP &= ~DP_LINK_TRAIN_MASK_CPT; > > switch (dp_train_pat & DP_TRAINING_PATTERN_MASK) { > > @@ -3314,7 +3314,7 @@ _intel_dp_set_link_train(struct intel_dp *intel_dp, > > static void intel_dp_enable_port(struct intel_dp *intel_dp, > > const struct intel_crtc_state *old_crtc_state) > > { > > - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp); > > + struct drm_i915_private *i915 = dp_to_i915(intel_dp); > > /* enable with pattern 1 (as per spec) */ > > @@ -3338,7 +3338,7 @@ static void intel_enable_dp(struct intel_encoder *encoder, > > const struct intel_crtc_state *pipe_config, > > const struct drm_connector_state *conn_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_dp *intel_dp = enc_to_intel_dp(&encoder->base); > > struct intel_crtc *crtc = to_intel_crtc(pipe_config->base.crtc); > > u32 dp_reg = I915_READ(intel_dp->output_reg); > > @@ -3349,7 +3349,7 @@ static void intel_enable_dp(struct intel_encoder *encoder, > > return; > > with_pps_lock(intel_dp, wakeref) { > > - if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) > > + if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) > > vlv_init_panel_power_sequencer(encoder, pipe_config); > > intel_dp_enable_port(intel_dp, pipe_config); > > @@ -3359,13 +3359,13 @@ static void intel_enable_dp(struct intel_encoder *encoder, > > edp_panel_vdd_off(intel_dp, true); > > } > > - if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) { > > + if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) { > > unsigned int lane_mask = 0x0; > > - if (IS_CHERRYVIEW(dev_priv)) > > + if (IS_CHERRYVIEW(i915)) > > lane_mask = intel_dp_unused_lane_mask(pipe_config->lane_count); > > - vlv_wait_port_ready(dev_priv, dp_to_dig_port(intel_dp), > > + vlv_wait_port_ready(i915, dp_to_dig_port(intel_dp), > > lane_mask); > > } > > @@ -3412,7 +3412,7 @@ static void g4x_pre_enable_dp(struct intel_encoder *encoder, > > static void vlv_detach_power_sequencer(struct intel_dp *intel_dp) > > { > > struct intel_digital_port *intel_dig_port = dp_to_dig_port(intel_dp); > > - struct drm_i915_private *dev_priv = to_i915(intel_dig_port->base.base.dev); > > + struct drm_i915_private *i915 = to_i915(intel_dig_port->base.base.dev); > > enum pipe pipe = intel_dp->pps_pipe; > > i915_reg_t pp_on_reg = PP_ON_DELAYS(pipe); > > @@ -3440,14 +3440,14 @@ static void vlv_detach_power_sequencer(struct intel_dp *intel_dp) > > intel_dp->pps_pipe = INVALID_PIPE; > > } > > -static void vlv_steal_power_sequencer(struct drm_i915_private *dev_priv, > > +static void vlv_steal_power_sequencer(struct drm_i915_private *i915, > > enum pipe pipe) > > { > > struct intel_encoder *encoder; > > - lockdep_assert_held(&dev_priv->pps_mutex); > > + lockdep_assert_held(&i915->pps_mutex); > > - for_each_intel_dp(&dev_priv->drm, encoder) { > > + for_each_intel_dp(&i915->drm, encoder) { > > struct intel_dp *intel_dp = enc_to_intel_dp(&encoder->base); > > enum port port = encoder->port; > > @@ -3469,11 +3469,11 @@ static void vlv_steal_power_sequencer(struct drm_i915_private *dev_priv, > > static void vlv_init_panel_power_sequencer(struct intel_encoder *encoder, > > const struct intel_crtc_state *crtc_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_dp *intel_dp = enc_to_intel_dp(&encoder->base); > > struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc); > > - lockdep_assert_held(&dev_priv->pps_mutex); > > + lockdep_assert_held(&i915->pps_mutex); > > WARN_ON(intel_dp->active_pipe != INVALID_PIPE); > > @@ -3491,7 +3491,7 @@ static void vlv_init_panel_power_sequencer(struct intel_encoder *encoder, > > * We may be stealing the power > > * sequencer from another port. > > */ > > - vlv_steal_power_sequencer(dev_priv, crtc->pipe); > > + vlv_steal_power_sequencer(i915, crtc->pipe); > > intel_dp->active_pipe = crtc->pipe; > > @@ -3570,17 +3570,17 @@ intel_dp_get_link_status(struct intel_dp *intel_dp, u8 link_status[DP_LINK_STATU > > u8 > > intel_dp_voltage_max(struct intel_dp *intel_dp) > > { > > - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp); > > + struct drm_i915_private *i915 = dp_to_i915(intel_dp); > > struct intel_encoder *encoder = &dp_to_dig_port(intel_dp)->base; > > enum port port = encoder->port; > > - if (HAS_DDI(dev_priv)) > > + if (HAS_DDI(i915)) > > return intel_ddi_dp_voltage_max(encoder); > > - else if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) > > + else if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) > > return DP_TRAIN_VOLTAGE_SWING_LEVEL_3; > > - else if (IS_IVYBRIDGE(dev_priv) && port == PORT_A) > > + else if (IS_IVYBRIDGE(i915) && port == PORT_A) > > return DP_TRAIN_VOLTAGE_SWING_LEVEL_2; > > - else if (HAS_PCH_CPT(dev_priv) && port != PORT_A) > > + else if (HAS_PCH_CPT(i915) && port != PORT_A) > > return DP_TRAIN_VOLTAGE_SWING_LEVEL_3; > > else > > return DP_TRAIN_VOLTAGE_SWING_LEVEL_2; > > @@ -3589,13 +3589,13 @@ intel_dp_voltage_max(struct intel_dp *intel_dp) > > u8 > > intel_dp_pre_emphasis_max(struct intel_dp *intel_dp, u8 voltage_swing) > > { > > - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp); > > + struct drm_i915_private *i915 = dp_to_i915(intel_dp); > > struct intel_encoder *encoder = &dp_to_dig_port(intel_dp)->base; > > enum port port = encoder->port; > > - if (HAS_DDI(dev_priv)) { > > + if (HAS_DDI(i915)) { > > return intel_ddi_dp_pre_emphasis_max(encoder, voltage_swing); > > - } else if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) { > > + } else if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) { > > switch (voltage_swing & DP_TRAIN_VOLTAGE_SWING_MASK) { > > case DP_TRAIN_VOLTAGE_SWING_LEVEL_0: > > return DP_TRAIN_PRE_EMPH_LEVEL_3; > > @@ -3607,7 +3607,7 @@ intel_dp_pre_emphasis_max(struct intel_dp *intel_dp, u8 voltage_swing) > > default: > > return DP_TRAIN_PRE_EMPH_LEVEL_0; > > } > > - } else if (IS_IVYBRIDGE(dev_priv) && port == PORT_A) { > > + } else if (IS_IVYBRIDGE(i915) && port == PORT_A) { > > switch (voltage_swing & DP_TRAIN_VOLTAGE_SWING_MASK) { > > case DP_TRAIN_VOLTAGE_SWING_LEVEL_0: > > return DP_TRAIN_PRE_EMPH_LEVEL_2; > > @@ -3901,25 +3901,25 @@ ivb_cpu_edp_signal_levels(u8 train_set) > > void > > intel_dp_set_signal_levels(struct intel_dp *intel_dp) > > { > > - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp); > > + struct drm_i915_private *i915 = dp_to_i915(intel_dp); > > struct intel_digital_port *intel_dig_port = dp_to_dig_port(intel_dp); > > enum port port = intel_dig_port->base.port; > > u32 signal_levels, mask = 0; > > u8 train_set = intel_dp->train_set[0]; > > - if (IS_GEN9_LP(dev_priv) || INTEL_GEN(dev_priv) >= 10) { > > + if (IS_GEN9_LP(i915) || INTEL_GEN(i915) >= 10) { > > signal_levels = bxt_signal_levels(intel_dp); > > - } else if (HAS_DDI(dev_priv)) { > > + } else if (HAS_DDI(i915)) { > > signal_levels = ddi_signal_levels(intel_dp); > > mask = DDI_BUF_EMP_MASK; > > - } else if (IS_CHERRYVIEW(dev_priv)) { > > + } else if (IS_CHERRYVIEW(i915)) { > > signal_levels = chv_signal_levels(intel_dp); > > - } else if (IS_VALLEYVIEW(dev_priv)) { > > + } else if (IS_VALLEYVIEW(i915)) { > > signal_levels = vlv_signal_levels(intel_dp); > > - } else if (IS_IVYBRIDGE(dev_priv) && port == PORT_A) { > > + } else if (IS_IVYBRIDGE(i915) && port == PORT_A) { > > signal_levels = ivb_cpu_edp_signal_levels(train_set); > > mask = EDP_LINK_TRAIN_VOL_EMP_MASK_IVB; > > - } else if (IS_GEN(dev_priv, 6) && port == PORT_A) { > > + } else if (IS_GEN(i915, 6) && port == PORT_A) { > > signal_levels = snb_cpu_edp_signal_levels(train_set); > > mask = EDP_LINK_TRAIN_VOL_EMP_MASK_SNB; > > } else { > > @@ -3947,7 +3947,7 @@ intel_dp_program_link_training_pattern(struct intel_dp *intel_dp, > > u8 dp_train_pat) > > { > > struct intel_digital_port *intel_dig_port = dp_to_dig_port(intel_dp); > > - struct drm_i915_private *dev_priv = > > + struct drm_i915_private *i915 = > > to_i915(intel_dig_port->base.base.dev); > > _intel_dp_set_link_train(intel_dp, &intel_dp->DP, dp_train_pat); > > @@ -3958,12 +3958,12 @@ intel_dp_program_link_training_pattern(struct intel_dp *intel_dp, > > void intel_dp_set_idle_link_train(struct intel_dp *intel_dp) > > { > > - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp); > > + struct drm_i915_private *i915 = dp_to_i915(intel_dp); > > struct intel_digital_port *intel_dig_port = dp_to_dig_port(intel_dp); > > enum port port = intel_dig_port->base.port; > > u32 val; > > - if (!HAS_DDI(dev_priv)) > > + if (!HAS_DDI(i915)) > > return; > > val = I915_READ(DP_TP_CTL(port)); > > @@ -3981,7 +3981,7 @@ void intel_dp_set_idle_link_train(struct intel_dp *intel_dp) > > if (port == PORT_A) > > return; > > - if (intel_wait_for_register(&dev_priv->uncore, DP_TP_STATUS(port), > > + if (intel_wait_for_register(&i915->uncore, DP_TP_STATUS(port), > > DP_TP_STATUS_IDLE_DONE, > > DP_TP_STATUS_IDLE_DONE, > > 1)) > > @@ -3992,7 +3992,7 @@ static void > > intel_dp_link_down(struct intel_encoder *encoder, > > const struct intel_crtc_state *old_crtc_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_dp *intel_dp = enc_to_intel_dp(&encoder->base); > > struct intel_crtc *crtc = to_intel_crtc(old_crtc_state->base.crtc); > > enum port port = encoder->port; > > @@ -4003,8 +4003,8 @@ intel_dp_link_down(struct intel_encoder *encoder, > > DRM_DEBUG_KMS("\n"); > > - if ((IS_IVYBRIDGE(dev_priv) && port == PORT_A) || > > - (HAS_PCH_CPT(dev_priv) && port != PORT_A)) { > > + if ((IS_IVYBRIDGE(i915) && port == PORT_A) || > > + (HAS_PCH_CPT(i915) && port != PORT_A)) { > > DP &= ~DP_LINK_TRAIN_MASK_CPT; > > DP |= DP_LINK_TRAIN_PAT_IDLE_CPT; > > } else { > > @@ -4023,13 +4023,13 @@ intel_dp_link_down(struct intel_encoder *encoder, > > * to transcoder A after disabling it to allow the > > * matching HDMI port to be enabled on transcoder A. > > */ > > - if (HAS_PCH_IBX(dev_priv) && crtc->pipe == PIPE_B && port != PORT_A) { > > + if (HAS_PCH_IBX(i915) && crtc->pipe == PIPE_B && port != PORT_A) { > > /* > > * We get CPU/PCH FIFO underruns on the other pipe when > > * doing the workaround. Sweep them under the rug. > > */ > > - intel_set_cpu_fifo_underrun_reporting(dev_priv, PIPE_A, false); > > - intel_set_pch_fifo_underrun_reporting(dev_priv, PIPE_A, false); > > + intel_set_cpu_fifo_underrun_reporting(i915, PIPE_A, false); > > + intel_set_pch_fifo_underrun_reporting(i915, PIPE_A, false); > > /* always enable with pattern 1 (as per spec) */ > > DP &= ~(DP_PIPE_SEL_MASK | DP_LINK_TRAIN_MASK); > > @@ -4042,16 +4042,16 @@ intel_dp_link_down(struct intel_encoder *encoder, > > I915_WRITE(intel_dp->output_reg, DP); > > POSTING_READ(intel_dp->output_reg); > > - intel_wait_for_vblank_if_active(dev_priv, PIPE_A); > > - intel_set_cpu_fifo_underrun_reporting(dev_priv, PIPE_A, true); > > - intel_set_pch_fifo_underrun_reporting(dev_priv, PIPE_A, true); > > + intel_wait_for_vblank_if_active(i915, PIPE_A); > > + intel_set_cpu_fifo_underrun_reporting(i915, PIPE_A, true); > > + intel_set_pch_fifo_underrun_reporting(i915, PIPE_A, true); > > } > > msleep(intel_dp->panel_power_down_delay); > > intel_dp->DP = DP; > > - if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) { > > + if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) { > > intel_wakeref_t wakeref; > > with_pps_lock(intel_dp, wakeref) > > @@ -4156,7 +4156,7 @@ static void intel_dp_get_dsc_sink_cap(struct intel_dp *intel_dp) > > static bool > > intel_edp_init_dpcd(struct intel_dp *intel_dp) > > { > > - struct drm_i915_private *dev_priv = > > + struct drm_i915_private *i915 = > > to_i915(dp_to_dig_port(intel_dp)->base.base.dev); > > /* this function is meant to be called only once */ > > @@ -4169,7 +4169,7 @@ intel_edp_init_dpcd(struct intel_dp *intel_dp) > > drm_dp_is_branch(intel_dp->dpcd)); > > if (intel_dp->dpcd[DP_DPCD_REV] >= 0x11) > > - dev_priv->no_aux_handshake = intel_dp->dpcd[DP_MAX_DOWNSPREAD] & > > + i915->no_aux_handshake = intel_dp->dpcd[DP_MAX_DOWNSPREAD] & > > DP_NO_AUX_HANDSHAKE_LINK_TRAINING; > > /* > > @@ -4230,7 +4230,7 @@ intel_edp_init_dpcd(struct intel_dp *intel_dp) > > intel_dp_set_common_rates(intel_dp); > > /* Read the eDP DSC DPCD registers */ > > - if (INTEL_GEN(dev_priv) >= 10 || IS_GEMINILAKE(dev_priv)) > > + if (INTEL_GEN(i915) >= 10 || IS_GEMINILAKE(i915)) > > intel_dp_get_dsc_sink_cap(intel_dp); > > return true; > > @@ -4804,7 +4804,7 @@ intel_dp_needs_link_retrain(struct intel_dp *intel_dp) > > int intel_dp_retrain_link(struct intel_encoder *encoder, > > struct drm_modeset_acquire_ctx *ctx) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_dp *intel_dp = enc_to_intel_dp(&encoder->base); > > struct intel_connector *connector = intel_dp->attached_connector; > > struct drm_connector_state *conn_state; > > @@ -4817,7 +4817,7 @@ int intel_dp_retrain_link(struct intel_encoder *encoder, > > if (!connector || connector->base.status != connector_status_connected) > > return 0; > > - ret = drm_modeset_lock(&dev_priv->drm.mode_config.connection_mutex, > > + ret = drm_modeset_lock(&i915->drm.mode_config.connection_mutex, > > ctx); > > if (ret) > > return ret; > > @@ -4847,20 +4847,20 @@ int intel_dp_retrain_link(struct intel_encoder *encoder, > > return 0; > > /* Suppress underruns caused by re-training */ > > - intel_set_cpu_fifo_underrun_reporting(dev_priv, crtc->pipe, false); > > + intel_set_cpu_fifo_underrun_reporting(i915, crtc->pipe, false); > > if (crtc_state->has_pch_encoder) > > - intel_set_pch_fifo_underrun_reporting(dev_priv, > > + intel_set_pch_fifo_underrun_reporting(i915, > > intel_crtc_pch_transcoder(crtc), false); > > intel_dp_start_link_train(intel_dp); > > intel_dp_stop_link_train(intel_dp); > > /* Keep underrun reporting disabled until things are stable */ > > - intel_wait_for_vblank(dev_priv, crtc->pipe); > > + intel_wait_for_vblank(i915, crtc->pipe); > > - intel_set_cpu_fifo_underrun_reporting(dev_priv, crtc->pipe, true); > > + intel_set_cpu_fifo_underrun_reporting(i915, crtc->pipe, true); > > if (crtc_state->has_pch_encoder) > > - intel_set_pch_fifo_underrun_reporting(dev_priv, > > + intel_set_pch_fifo_underrun_reporting(i915, > > intel_crtc_pch_transcoder(crtc), true); > > return 0; > > @@ -4946,7 +4946,7 @@ static void intel_dp_check_service_irq(struct intel_dp *intel_dp) > > static bool > > intel_dp_short_pulse(struct intel_dp *intel_dp) > > { > > - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp); > > + struct drm_i915_private *i915 = dp_to_i915(intel_dp); > > u8 old_sink_count = intel_dp->sink_count; > > bool ret; > > @@ -4983,7 +4983,7 @@ intel_dp_short_pulse(struct intel_dp *intel_dp) > > if (intel_dp->compliance.test_type == DP_TEST_LINK_TRAINING) { > > DRM_DEBUG_KMS("Link Training Compliance Test requested\n"); > > /* Send a Hotplug Uevent to userspace to start modeset */ > > - drm_kms_helper_hotplug_event(&dev_priv->drm); > > + drm_kms_helper_hotplug_event(&i915->drm); > > } > > return true; > > @@ -5052,7 +5052,7 @@ edp_detect(struct intel_dp *intel_dp) > > static bool ibx_digital_port_connected(struct intel_encoder *encoder) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > u32 bit; > > switch (encoder->hpd_pin) { > > @@ -5075,7 +5075,7 @@ static bool ibx_digital_port_connected(struct intel_encoder *encoder) > > static bool cpt_digital_port_connected(struct intel_encoder *encoder) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > u32 bit; > > switch (encoder->hpd_pin) { > > @@ -5098,7 +5098,7 @@ static bool cpt_digital_port_connected(struct intel_encoder *encoder) > > static bool spt_digital_port_connected(struct intel_encoder *encoder) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > u32 bit; > > switch (encoder->hpd_pin) { > > @@ -5117,7 +5117,7 @@ static bool spt_digital_port_connected(struct intel_encoder *encoder) > > static bool g4x_digital_port_connected(struct intel_encoder *encoder) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > u32 bit; > > switch (encoder->hpd_pin) { > > @@ -5140,7 +5140,7 @@ static bool g4x_digital_port_connected(struct intel_encoder *encoder) > > static bool gm45_digital_port_connected(struct intel_encoder *encoder) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > u32 bit; > > switch (encoder->hpd_pin) { > > @@ -5163,7 +5163,7 @@ static bool gm45_digital_port_connected(struct intel_encoder *encoder) > > static bool ilk_digital_port_connected(struct intel_encoder *encoder) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > if (encoder->hpd_pin == HPD_PORT_A) > > return I915_READ(DEISR) & DE_DP_A_HOTPLUG; > > @@ -5173,7 +5173,7 @@ static bool ilk_digital_port_connected(struct intel_encoder *encoder) > > static bool snb_digital_port_connected(struct intel_encoder *encoder) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > if (encoder->hpd_pin == HPD_PORT_A) > > return I915_READ(DEISR) & DE_DP_A_HOTPLUG; > > @@ -5183,7 +5183,7 @@ static bool snb_digital_port_connected(struct intel_encoder *encoder) > > static bool ivb_digital_port_connected(struct intel_encoder *encoder) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > if (encoder->hpd_pin == HPD_PORT_A) > > return I915_READ(DEISR) & DE_DP_A_HOTPLUG_IVB; > > @@ -5193,7 +5193,7 @@ static bool ivb_digital_port_connected(struct intel_encoder *encoder) > > static bool bdw_digital_port_connected(struct intel_encoder *encoder) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > if (encoder->hpd_pin == HPD_PORT_A) > > return I915_READ(GEN8_DE_PORT_ISR) & GEN8_PORT_DP_A_HOTPLUG; > > @@ -5203,7 +5203,7 @@ static bool bdw_digital_port_connected(struct intel_encoder *encoder) > > static bool bxt_digital_port_connected(struct intel_encoder *encoder) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > u32 bit; > > switch (encoder->hpd_pin) { > > @@ -5224,7 +5224,7 @@ static bool bxt_digital_port_connected(struct intel_encoder *encoder) > > return I915_READ(GEN8_DE_PORT_ISR) & bit; > > } > > -static bool icl_combo_port_connected(struct drm_i915_private *dev_priv, > > +static bool icl_combo_port_connected(struct drm_i915_private *i915, > > struct intel_digital_port *intel_dig_port) > > { > > enum port port = intel_dig_port->base.port; > > @@ -5247,7 +5247,7 @@ static const char *tc_type_name(enum tc_port_type type) > > return names[type]; > > } > > -static void icl_update_tc_port_type(struct drm_i915_private *dev_priv, > > +static void icl_update_tc_port_type(struct drm_i915_private *i915, > > struct intel_digital_port *intel_dig_port, > > bool is_legacy, bool is_typec, bool is_tbt) > > { > > @@ -5295,10 +5295,10 @@ static void icl_update_tc_port_type(struct drm_i915_private *dev_priv, > > * will require a lot of coordination with user space and thorough testing for > > * the extra possible cases. > > */ > > -static bool icl_tc_phy_connect(struct drm_i915_private *dev_priv, > > +static bool icl_tc_phy_connect(struct drm_i915_private *i915, > > struct intel_digital_port *dig_port) > > { > > - enum tc_port tc_port = intel_port_to_tc(dev_priv, dig_port->base.port); > > + enum tc_port tc_port = intel_port_to_tc(i915, dig_port->base.port); > > u32 val; > > if (dig_port->tc_type != TC_PORT_LEGACY && > > @@ -5329,7 +5329,7 @@ static bool icl_tc_phy_connect(struct drm_i915_private *dev_priv, > > if (dig_port->tc_type == TC_PORT_TYPEC && > > !(I915_READ(PORT_TX_DFLEXDPSP) & TC_LIVE_STATE_TC(tc_port))) { > > DRM_DEBUG_KMS("TC PHY %d sudden disconnect.\n", tc_port); > > - icl_tc_phy_disconnect(dev_priv, dig_port); > > + icl_tc_phy_disconnect(i915, dig_port); > > return false; > > } > > @@ -5340,10 +5340,10 @@ static bool icl_tc_phy_connect(struct drm_i915_private *dev_priv, > > * See the comment at the connect function. This implements the Disconnect > > * Flow. > > */ > > -void icl_tc_phy_disconnect(struct drm_i915_private *dev_priv, > > +void icl_tc_phy_disconnect(struct drm_i915_private *i915, > > struct intel_digital_port *dig_port) > > { > > - enum tc_port tc_port = intel_port_to_tc(dev_priv, dig_port->base.port); > > + enum tc_port tc_port = intel_port_to_tc(i915, dig_port->base.port); > > if (dig_port->tc_type == TC_PORT_UNKNOWN) > > return; > > @@ -5378,11 +5378,11 @@ void icl_tc_phy_disconnect(struct drm_i915_private *dev_priv, > > * connected ports are usable, and avoids exposing to the users objects they > > * can't really use. > > */ > > -static bool icl_tc_port_connected(struct drm_i915_private *dev_priv, > > +static bool icl_tc_port_connected(struct drm_i915_private *i915, > > struct intel_digital_port *intel_dig_port) > > { > > enum port port = intel_dig_port->base.port; > > - enum tc_port tc_port = intel_port_to_tc(dev_priv, port); > > + enum tc_port tc_port = intel_port_to_tc(i915, port); > > bool is_legacy, is_typec, is_tbt; > > u32 dpsp; > > @@ -5407,15 +5407,15 @@ static bool icl_tc_port_connected(struct drm_i915_private *dev_priv, > > is_tbt = dpsp & TC_LIVE_STATE_TBT(tc_port); > > if (!is_legacy && !is_typec && !is_tbt) { > > - icl_tc_phy_disconnect(dev_priv, intel_dig_port); > > + icl_tc_phy_disconnect(i915, intel_dig_port); > > return false; > > } > > - icl_update_tc_port_type(dev_priv, intel_dig_port, is_legacy, is_typec, > > + icl_update_tc_port_type(i915, intel_dig_port, is_legacy, is_typec, > > is_tbt); > > - if (!icl_tc_phy_connect(dev_priv, intel_dig_port)) > > + if (!icl_tc_phy_connect(i915, intel_dig_port)) > > return false; > > return true; > > @@ -5423,13 +5423,13 @@ static bool icl_tc_port_connected(struct drm_i915_private *dev_priv, > > static bool icl_digital_port_connected(struct intel_encoder *encoder) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_digital_port *dig_port = enc_to_dig_port(&encoder->base); > > - if (intel_port_is_combophy(dev_priv, encoder->port)) > > - return icl_combo_port_connected(dev_priv, dig_port); > > - else if (intel_port_is_tc(dev_priv, encoder->port)) > > - return icl_tc_port_connected(dev_priv, dig_port); > > + if (intel_port_is_combophy(i915, encoder->port)) > > + return icl_combo_port_connected(i915, dig_port); > > + else if (intel_port_is_tc(i915, encoder->port)) > > + return icl_tc_port_connected(i915, dig_port); > > else > > MISSING_CASE(encoder->hpd_pin); > > @@ -5449,41 +5449,41 @@ static bool icl_digital_port_connected(struct intel_encoder *encoder) > > */ > > static bool __intel_digital_port_connected(struct intel_encoder *encoder) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > - if (HAS_GMCH(dev_priv)) { > > - if (IS_GM45(dev_priv)) > > + if (HAS_GMCH(i915)) { > > + if (IS_GM45(i915)) > > return gm45_digital_port_connected(encoder); > > else > > return g4x_digital_port_connected(encoder); > > } > > - if (INTEL_GEN(dev_priv) >= 11) > > + if (INTEL_GEN(i915) >= 11) > > return icl_digital_port_connected(encoder); > > - else if (IS_GEN(dev_priv, 10) || IS_GEN9_BC(dev_priv)) > > + else if (IS_GEN(i915, 10) || IS_GEN9_BC(i915)) > > return spt_digital_port_connected(encoder); > > - else if (IS_GEN9_LP(dev_priv)) > > + else if (IS_GEN9_LP(i915)) > > return bxt_digital_port_connected(encoder); > > - else if (IS_GEN(dev_priv, 8)) > > + else if (IS_GEN(i915, 8)) > > return bdw_digital_port_connected(encoder); > > - else if (IS_GEN(dev_priv, 7)) > > + else if (IS_GEN(i915, 7)) > > return ivb_digital_port_connected(encoder); > > - else if (IS_GEN(dev_priv, 6)) > > + else if (IS_GEN(i915, 6)) > > return snb_digital_port_connected(encoder); > > - else if (IS_GEN(dev_priv, 5)) > > + else if (IS_GEN(i915, 5)) > > return ilk_digital_port_connected(encoder); > > - MISSING_CASE(INTEL_GEN(dev_priv)); > > + MISSING_CASE(INTEL_GEN(i915)); > > return false; > > } > > bool intel_digital_port_connected(struct intel_encoder *encoder) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > bool is_connected = false; > > intel_wakeref_t wakeref; > > - with_intel_display_power(dev_priv, POWER_DOMAIN_DISPLAY_CORE, wakeref) > > + with_intel_display_power(i915, POWER_DOMAIN_DISPLAY_CORE, wakeref) > > is_connected = __intel_digital_port_connected(encoder); > > return is_connected; > > @@ -5537,7 +5537,7 @@ intel_dp_detect(struct drm_connector *connector, > > struct drm_modeset_acquire_ctx *ctx, > > bool force) > > { > > - struct drm_i915_private *dev_priv = to_i915(connector->dev); > > + struct drm_i915_private *i915 = to_i915(connector->dev); > > struct intel_dp *intel_dp = intel_attached_dp(connector); > > struct intel_digital_port *dig_port = dp_to_dig_port(intel_dp); > > struct intel_encoder *encoder = &dig_port->base; > > @@ -5545,7 +5545,7 @@ intel_dp_detect(struct drm_connector *connector, > > DRM_DEBUG_KMS("[CONNECTOR:%d:%s]\n", > > connector->base.id, connector->name); > > - WARN_ON(!drm_modeset_is_locked(&dev_priv->drm.mode_config.connection_mutex)); > > + WARN_ON(!drm_modeset_is_locked(&i915->drm.mode_config.connection_mutex)); > > /* Can't disconnect eDP */ > > if (intel_dp_is_edp(intel_dp)) > > @@ -5584,7 +5584,7 @@ intel_dp_detect(struct drm_connector *connector, > > intel_dp_print_rates(intel_dp); > > /* Read DP Sink DSC Cap DPCD regs for DP v1.4 */ > > - if (INTEL_GEN(dev_priv) >= 11) > > + if (INTEL_GEN(i915) >= 11) > > intel_dp_get_dsc_sink_cap(intel_dp); > > drm_dp_read_desc(&intel_dp->aux, &intel_dp->desc, > > @@ -5642,7 +5642,7 @@ intel_dp_force(struct drm_connector *connector) > > struct intel_dp *intel_dp = intel_attached_dp(connector); > > struct intel_digital_port *dig_port = dp_to_dig_port(intel_dp); > > struct intel_encoder *intel_encoder = &dig_port->base; > > - struct drm_i915_private *dev_priv = to_i915(intel_encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(intel_encoder->base.dev); > > enum intel_display_power_domain aux_domain = > > intel_aux_power_domain(dig_port); > > intel_wakeref_t wakeref; > > @@ -5654,11 +5654,11 @@ intel_dp_force(struct drm_connector *connector) > > if (connector->status != connector_status_connected) > > return; > > - wakeref = intel_display_power_get(dev_priv, aux_domain); > > + wakeref = intel_display_power_get(i915, aux_domain); > > intel_dp_set_edid(intel_dp); > > - intel_display_power_put(dev_priv, aux_domain, wakeref); > > + intel_display_power_put(i915, aux_domain, wakeref); > > } > > static int intel_dp_get_modes(struct drm_connector *connector) > > @@ -6362,10 +6362,10 @@ static const struct intel_hdcp_shim intel_dp_hdcp_shim = { > > static void intel_edp_panel_vdd_sanitize(struct intel_dp *intel_dp) > > { > > - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp); > > + struct drm_i915_private *i915 = dp_to_i915(intel_dp); > > struct intel_digital_port *dig_port = dp_to_dig_port(intel_dp); > > - lockdep_assert_held(&dev_priv->pps_mutex); > > + lockdep_assert_held(&i915->pps_mutex); > > if (!edp_have_panel_vdd(intel_dp)) > > return; > > @@ -6377,18 +6377,18 @@ static void intel_edp_panel_vdd_sanitize(struct intel_dp *intel_dp) > > * indefinitely. > > */ > > DRM_DEBUG_KMS("VDD left on by BIOS, adjusting state tracking\n"); > > - intel_display_power_get(dev_priv, intel_aux_power_domain(dig_port)); > > + intel_display_power_get(i915, intel_aux_power_domain(dig_port)); > > edp_panel_vdd_schedule_off(intel_dp); > > } > > static enum pipe vlv_active_pipe(struct intel_dp *intel_dp) > > { > > - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp); > > + struct drm_i915_private *i915 = dp_to_i915(intel_dp); > > struct intel_encoder *encoder = &dp_to_dig_port(intel_dp)->base; > > enum pipe pipe; > > - if (intel_dp_port_enabled(dev_priv, intel_dp->output_reg, > > + if (intel_dp_port_enabled(i915, intel_dp->output_reg, > > encoder->port, &pipe)) > > return pipe; > > @@ -6397,12 +6397,12 @@ static enum pipe vlv_active_pipe(struct intel_dp *intel_dp) > > void intel_dp_encoder_reset(struct drm_encoder *encoder) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->dev); > > + struct drm_i915_private *i915 = to_i915(encoder->dev); > > struct intel_dp *intel_dp = enc_to_intel_dp(encoder); > > struct intel_lspcon *lspcon = dp_to_lspcon(intel_dp); > > intel_wakeref_t wakeref; > > - if (!HAS_DDI(dev_priv)) > > + if (!HAS_DDI(i915)) > > intel_dp->DP = I915_READ(intel_dp->output_reg); > > if (lspcon->active) > > @@ -6410,12 +6410,12 @@ void intel_dp_encoder_reset(struct drm_encoder *encoder) > > intel_dp->reset_link_params = true; > > - if (!IS_VALLEYVIEW(dev_priv) && !IS_CHERRYVIEW(dev_priv) && > > + if (!IS_VALLEYVIEW(i915) && !IS_CHERRYVIEW(i915) && > > !intel_dp_is_edp(intel_dp)) > > return; > > with_pps_lock(intel_dp, wakeref) { > > - if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) > > + if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) > > intel_dp->active_pipe = vlv_active_pipe(intel_dp); > > if (intel_dp_is_edp(intel_dp)) { > > @@ -6508,41 +6508,41 @@ intel_dp_hpd_pulse(struct intel_digital_port *intel_dig_port, bool long_hpd) > > } > > /* check the VBT to see whether the eDP is on another port */ > > -bool intel_dp_is_port_edp(struct drm_i915_private *dev_priv, enum port port) > > +bool intel_dp_is_port_edp(struct drm_i915_private *i915, enum port port) > > { > > /* > > * eDP not supported on g4x. so bail out early just > > * for a bit extra safety in case the VBT is bonkers. > > */ > > - if (INTEL_GEN(dev_priv) < 5) > > + if (INTEL_GEN(i915) < 5) > > return false; > > - if (INTEL_GEN(dev_priv) < 9 && port == PORT_A) > > + if (INTEL_GEN(i915) < 9 && port == PORT_A) > > return true; > > - return intel_bios_is_port_edp(dev_priv, port); > > + return intel_bios_is_port_edp(i915, port); > > } > > static void > > intel_dp_add_properties(struct intel_dp *intel_dp, struct drm_connector *connector) > > { > > - struct drm_i915_private *dev_priv = to_i915(connector->dev); > > + struct drm_i915_private *i915 = to_i915(connector->dev); > > enum port port = dp_to_dig_port(intel_dp)->base.port; > > - if (!IS_G4X(dev_priv) && port != PORT_A) > > + if (!IS_G4X(i915) && port != PORT_A) > > intel_attach_force_audio_property(connector); > > intel_attach_broadcast_rgb_property(connector); > > - if (HAS_GMCH(dev_priv)) > > + if (HAS_GMCH(i915)) > > drm_connector_attach_max_bpc_property(connector, 6, 10); > > - else if (INTEL_GEN(dev_priv) >= 5) > > + else if (INTEL_GEN(i915) >= 5) > > drm_connector_attach_max_bpc_property(connector, 6, 12); > > if (intel_dp_is_edp(intel_dp)) { > > u32 allowed_scalers; > > allowed_scalers = BIT(DRM_MODE_SCALE_ASPECT) | BIT(DRM_MODE_SCALE_FULLSCREEN); > > - if (!HAS_GMCH(dev_priv)) > > + if (!HAS_GMCH(i915)) > > allowed_scalers |= BIT(DRM_MODE_SCALE_CENTER); > > drm_connector_attach_scaling_mode_property(connector, allowed_scalers); > > @@ -6562,7 +6562,7 @@ static void intel_dp_init_panel_power_timestamps(struct intel_dp *intel_dp) > > static void > > intel_pps_readout_hw_state(struct intel_dp *intel_dp, struct edp_power_seq *seq) > > { > > - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp); > > + struct drm_i915_private *i915 = dp_to_i915(intel_dp); > > u32 pp_on, pp_off, pp_ctl; > > struct pps_registers regs; > > @@ -6571,7 +6571,7 @@ intel_pps_readout_hw_state(struct intel_dp *intel_dp, struct edp_power_seq *seq) > > pp_ctl = ironlake_get_pp_control(intel_dp); > > /* Ensure PPS is unlocked */ > > - if (!HAS_DDI(dev_priv)) > > + if (!HAS_DDI(i915)) > > I915_WRITE(regs.pp_ctrl, pp_ctl); > > pp_on = I915_READ(regs.pp_on); > > @@ -6621,11 +6621,11 @@ intel_pps_verify_state(struct intel_dp *intel_dp) > > static void > > intel_dp_init_panel_power_sequencer(struct intel_dp *intel_dp) > > { > > - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp); > > + struct drm_i915_private *i915 = dp_to_i915(intel_dp); > > struct edp_power_seq cur, vbt, spec, > > *final = &intel_dp->pps_delays; > > - lockdep_assert_held(&dev_priv->pps_mutex); > > + lockdep_assert_held(&i915->pps_mutex); > > /* already initialized? */ > > if (final->t11_t12 != 0) > > @@ -6635,13 +6635,13 @@ intel_dp_init_panel_power_sequencer(struct intel_dp *intel_dp) > > intel_pps_dump_state("cur", &cur); > > - vbt = dev_priv->vbt.edp.pps; > > + vbt = i915->vbt.edp.pps; > > /* On Toshiba Satellite P50-C-18C system the VBT T12 delay > > * of 500ms appears to be too short. Ocassionally the panel > > * just fails to power back on. Increasing the delay to 800ms > > * seems sufficient to avoid this problem. > > */ > > - if (dev_priv->quirks & QUIRK_INCREASE_T12_DELAY) { > > + if (i915->quirks & QUIRK_INCREASE_T12_DELAY) { > > vbt.t11_t12 = max_t(u16, vbt.t11_t12, 1300 * 10); > > DRM_DEBUG_KMS("Increasing T12 panel delay as per the quirk to %d\n", > > vbt.t11_t12); > > @@ -6714,14 +6714,14 @@ static void > > intel_dp_init_panel_power_sequencer_registers(struct intel_dp *intel_dp, > > bool force_disable_vdd) > > { > > - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp); > > + struct drm_i915_private *i915 = dp_to_i915(intel_dp); > > u32 pp_on, pp_off, port_sel = 0; > > - int div = dev_priv->rawclk_freq / 1000; > > + int div = i915->rawclk_freq / 1000; > > struct pps_registers regs; > > enum port port = dp_to_dig_port(intel_dp)->base.port; > > const struct edp_power_seq *seq = &intel_dp->pps_delays; > > - lockdep_assert_held(&dev_priv->pps_mutex); > > + lockdep_assert_held(&i915->pps_mutex); > > intel_pps_get_registers(intel_dp, ®s); > > @@ -6757,9 +6757,9 @@ intel_dp_init_panel_power_sequencer_registers(struct intel_dp *intel_dp, > > /* Haswell doesn't have any port selection bits for the panel > > * power sequencer any more. */ > > - if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) { > > + if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) { > > port_sel = PANEL_PORT_SELECT_VLV(port); > > - } else if (HAS_PCH_IBX(dev_priv) || HAS_PCH_CPT(dev_priv)) { > > + } else if (HAS_PCH_IBX(i915) || HAS_PCH_CPT(i915)) { > > switch (port) { > > case PORT_A: > > port_sel = PANEL_PORT_SELECT_DPA; > > @@ -6807,9 +6807,9 @@ intel_dp_init_panel_power_sequencer_registers(struct intel_dp *intel_dp, > > static void intel_dp_pps_init(struct intel_dp *intel_dp) > > { > > - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp); > > + struct drm_i915_private *i915 = dp_to_i915(intel_dp); > > - if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) { > > + if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) { > > vlv_initial_power_sequencer_setup(intel_dp); > > } else { > > intel_dp_init_panel_power_sequencer(intel_dp); > > @@ -6819,7 +6819,7 @@ static void intel_dp_pps_init(struct intel_dp *intel_dp) > > /** > > * intel_dp_set_drrs_state - program registers for RR switch to take effect > > - * @dev_priv: i915 device > > + * @i915: i915 device > > * @crtc_state: a pointer to the active intel_crtc_state > > * @refresh_rate: RR to be programmed > > * > > @@ -6828,15 +6828,15 @@ static void intel_dp_pps_init(struct intel_dp *intel_dp) > > * supported by the panel or to any other RR based on media playback (in > > * this case, RR value needs to be passed from user space). > > * > > - * The caller of this function needs to take a lock on dev_priv->drrs. > > + * The caller of this function needs to take a lock on i915->drrs. > > */ > > -static void intel_dp_set_drrs_state(struct drm_i915_private *dev_priv, > > +static void intel_dp_set_drrs_state(struct drm_i915_private *i915, > > const struct intel_crtc_state *crtc_state, > > int refresh_rate) > > { > > struct intel_encoder *encoder; > > struct intel_digital_port *dig_port = NULL; > > - struct intel_dp *intel_dp = dev_priv->drrs.dp; > > + struct intel_dp *intel_dp = i915->drrs.dp; > > struct intel_crtc *intel_crtc = to_intel_crtc(crtc_state->base.crtc); > > enum drrs_refresh_rate_type index = DRRS_HIGH_RR; > > @@ -6858,7 +6858,7 @@ static void intel_dp_set_drrs_state(struct drm_i915_private *dev_priv, > > return; > > } > > - if (dev_priv->drrs.type < SEAMLESS_DRRS_SUPPORT) { > > + if (i915->drrs.type < SEAMLESS_DRRS_SUPPORT) { > > DRM_DEBUG_KMS("Only Seamless DRRS supported.\n"); > > return; > > } > > @@ -6867,7 +6867,7 @@ static void intel_dp_set_drrs_state(struct drm_i915_private *dev_priv, > > refresh_rate) > > index = DRRS_LOW_RR; > > - if (index == dev_priv->drrs.refresh_rate_type) { > > + if (index == i915->drrs.refresh_rate_type) { > > DRM_DEBUG_KMS( > > "DRRS requested for previously set RR...ignoring\n"); > > return; > > @@ -6878,7 +6878,7 @@ static void intel_dp_set_drrs_state(struct drm_i915_private *dev_priv, > > return; > > } > > - if (INTEL_GEN(dev_priv) >= 8 && !IS_CHERRYVIEW(dev_priv)) { > > + if (INTEL_GEN(i915) >= 8 && !IS_CHERRYVIEW(i915)) { > > switch (index) { > > case DRRS_HIGH_RR: > > intel_dp_set_m_n(crtc_state, M1_N1); > > @@ -6890,18 +6890,18 @@ static void intel_dp_set_drrs_state(struct drm_i915_private *dev_priv, > > default: > > DRM_ERROR("Unsupported refreshrate type\n"); > > } > > - } else if (INTEL_GEN(dev_priv) > 6) { > > + } else if (INTEL_GEN(i915) > 6) { > > i915_reg_t reg = PIPECONF(crtc_state->cpu_transcoder); > > u32 val; > > val = I915_READ(reg); > > if (index > DRRS_HIGH_RR) { > > - if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) > > + if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) > > val |= PIPECONF_EDP_RR_MODE_SWITCH_VLV; > > else > > val |= PIPECONF_EDP_RR_MODE_SWITCH; > > } else { > > - if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) > > + if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) > > val &= ~PIPECONF_EDP_RR_MODE_SWITCH_VLV; > > else > > val &= ~PIPECONF_EDP_RR_MODE_SWITCH; > > @@ -6909,7 +6909,7 @@ static void intel_dp_set_drrs_state(struct drm_i915_private *dev_priv, > > I915_WRITE(reg, val); > > } > > - dev_priv->drrs.refresh_rate_type = index; > > + i915->drrs.refresh_rate_type = index; > > DRM_DEBUG_KMS("eDP Refresh Rate set to : %dHz\n", refresh_rate); > > } > > @@ -6924,30 +6924,30 @@ static void intel_dp_set_drrs_state(struct drm_i915_private *dev_priv, > > void intel_edp_drrs_enable(struct intel_dp *intel_dp, > > const struct intel_crtc_state *crtc_state) > > { > > - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp); > > + struct drm_i915_private *i915 = dp_to_i915(intel_dp); > > if (!crtc_state->has_drrs) { > > DRM_DEBUG_KMS("Panel doesn't support DRRS\n"); > > return; > > } > > - if (dev_priv->psr.enabled) { > > + if (i915->psr.enabled) { > > DRM_DEBUG_KMS("PSR enabled. Not enabling DRRS.\n"); > > return; > > } > > - mutex_lock(&dev_priv->drrs.mutex); > > - if (dev_priv->drrs.dp) { > > + mutex_lock(&i915->drrs.mutex); > > + if (i915->drrs.dp) { > > DRM_DEBUG_KMS("DRRS already enabled\n"); > > goto unlock; > > } > > - dev_priv->drrs.busy_frontbuffer_bits = 0; > > + i915->drrs.busy_frontbuffer_bits = 0; > > - dev_priv->drrs.dp = intel_dp; > > + i915->drrs.dp = intel_dp; > > unlock: > > - mutex_unlock(&dev_priv->drrs.mutex); > > + mutex_unlock(&i915->drrs.mutex); > > } > > /** > > @@ -6959,36 +6959,36 @@ void intel_edp_drrs_enable(struct intel_dp *intel_dp, > > void intel_edp_drrs_disable(struct intel_dp *intel_dp, > > const struct intel_crtc_state *old_crtc_state) > > { > > - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp); > > + struct drm_i915_private *i915 = dp_to_i915(intel_dp); > > if (!old_crtc_state->has_drrs) > > return; > > - mutex_lock(&dev_priv->drrs.mutex); > > - if (!dev_priv->drrs.dp) { > > - mutex_unlock(&dev_priv->drrs.mutex); > > + mutex_lock(&i915->drrs.mutex); > > + if (!i915->drrs.dp) { > > + mutex_unlock(&i915->drrs.mutex); > > return; > > } > > - if (dev_priv->drrs.refresh_rate_type == DRRS_LOW_RR) > > - intel_dp_set_drrs_state(dev_priv, old_crtc_state, > > + if (i915->drrs.refresh_rate_type == DRRS_LOW_RR) > > + intel_dp_set_drrs_state(i915, old_crtc_state, > > intel_dp->attached_connector->panel.fixed_mode->vrefresh); > > - dev_priv->drrs.dp = NULL; > > - mutex_unlock(&dev_priv->drrs.mutex); > > + i915->drrs.dp = NULL; > > + mutex_unlock(&i915->drrs.mutex); > > - cancel_delayed_work_sync(&dev_priv->drrs.work); > > + cancel_delayed_work_sync(&i915->drrs.work); > > } > > static void intel_edp_drrs_downclock_work(struct work_struct *work) > > { > > - struct drm_i915_private *dev_priv = > > - container_of(work, typeof(*dev_priv), drrs.work.work); > > + struct drm_i915_private *i915 = > > + container_of(work, typeof(*i915), drrs.work.work); > > struct intel_dp *intel_dp; > > - mutex_lock(&dev_priv->drrs.mutex); > > + mutex_lock(&i915->drrs.mutex); > > - intel_dp = dev_priv->drrs.dp; > > + intel_dp = i915->drrs.dp; > > if (!intel_dp) > > goto unlock; > > @@ -6998,23 +6998,23 @@ static void intel_edp_drrs_downclock_work(struct work_struct *work) > > * recheck. > > */ > > - if (dev_priv->drrs.busy_frontbuffer_bits) > > + if (i915->drrs.busy_frontbuffer_bits) > > goto unlock; > > - if (dev_priv->drrs.refresh_rate_type != DRRS_LOW_RR) { > > + if (i915->drrs.refresh_rate_type != DRRS_LOW_RR) { > > struct drm_crtc *crtc = dp_to_dig_port(intel_dp)->base.base.crtc; > > - intel_dp_set_drrs_state(dev_priv, to_intel_crtc(crtc)->config, > > + intel_dp_set_drrs_state(i915, to_intel_crtc(crtc)->config, > > intel_dp->attached_connector->panel.downclock_mode->vrefresh); > > } > > unlock: > > - mutex_unlock(&dev_priv->drrs.mutex); > > + mutex_unlock(&i915->drrs.mutex); > > } > > /** > > * intel_edp_drrs_invalidate - Disable Idleness DRRS > > - * @dev_priv: i915 device > > + * @i915: i915 device > > * @frontbuffer_bits: frontbuffer plane tracking bits > > * > > * This function gets called everytime rendering on the given planes start. > > @@ -7022,40 +7022,40 @@ static void intel_edp_drrs_downclock_work(struct work_struct *work) > > * > > * Dirty frontbuffers relevant to DRRS are tracked in busy_frontbuffer_bits. > > */ > > -void intel_edp_drrs_invalidate(struct drm_i915_private *dev_priv, > > +void intel_edp_drrs_invalidate(struct drm_i915_private *i915, > > unsigned int frontbuffer_bits) > > { > > struct drm_crtc *crtc; > > enum pipe pipe; > > - if (dev_priv->drrs.type == DRRS_NOT_SUPPORTED) > > + if (i915->drrs.type == DRRS_NOT_SUPPORTED) > > return; > > - cancel_delayed_work(&dev_priv->drrs.work); > > + cancel_delayed_work(&i915->drrs.work); > > - mutex_lock(&dev_priv->drrs.mutex); > > - if (!dev_priv->drrs.dp) { > > - mutex_unlock(&dev_priv->drrs.mutex); > > + mutex_lock(&i915->drrs.mutex); > > + if (!i915->drrs.dp) { > > + mutex_unlock(&i915->drrs.mutex); > > return; > > } > > - crtc = dp_to_dig_port(dev_priv->drrs.dp)->base.base.crtc; > > + crtc = dp_to_dig_port(i915->drrs.dp)->base.base.crtc; > > pipe = to_intel_crtc(crtc)->pipe; > > frontbuffer_bits &= INTEL_FRONTBUFFER_ALL_MASK(pipe); > > - dev_priv->drrs.busy_frontbuffer_bits |= frontbuffer_bits; > > + i915->drrs.busy_frontbuffer_bits |= frontbuffer_bits; > > /* invalidate means busy screen hence upclock */ > > - if (frontbuffer_bits && dev_priv->drrs.refresh_rate_type == DRRS_LOW_RR) > > - intel_dp_set_drrs_state(dev_priv, to_intel_crtc(crtc)->config, > > - dev_priv->drrs.dp->attached_connector->panel.fixed_mode->vrefresh); > > + if (frontbuffer_bits && i915->drrs.refresh_rate_type == DRRS_LOW_RR) > > + intel_dp_set_drrs_state(i915, to_intel_crtc(crtc)->config, > > + i915->drrs.dp->attached_connector->panel.fixed_mode->vrefresh); > > - mutex_unlock(&dev_priv->drrs.mutex); > > + mutex_unlock(&i915->drrs.mutex); > > } > > /** > > * intel_edp_drrs_flush - Restart Idleness DRRS > > - * @dev_priv: i915 device > > + * @i915: i915 device > > * @frontbuffer_bits: frontbuffer plane tracking bits > > * > > * This function gets called every time rendering on the given planes has > > @@ -7065,42 +7065,42 @@ void intel_edp_drrs_invalidate(struct drm_i915_private *dev_priv, > > * > > * Dirty frontbuffers relevant to DRRS are tracked in busy_frontbuffer_bits. > > */ > > -void intel_edp_drrs_flush(struct drm_i915_private *dev_priv, > > +void intel_edp_drrs_flush(struct drm_i915_private *i915, > > unsigned int frontbuffer_bits) > > { > > struct drm_crtc *crtc; > > enum pipe pipe; > > - if (dev_priv->drrs.type == DRRS_NOT_SUPPORTED) > > + if (i915->drrs.type == DRRS_NOT_SUPPORTED) > > return; > > - cancel_delayed_work(&dev_priv->drrs.work); > > + cancel_delayed_work(&i915->drrs.work); > > - mutex_lock(&dev_priv->drrs.mutex); > > - if (!dev_priv->drrs.dp) { > > - mutex_unlock(&dev_priv->drrs.mutex); > > + mutex_lock(&i915->drrs.mutex); > > + if (!i915->drrs.dp) { > > + mutex_unlock(&i915->drrs.mutex); > > return; > > } > > - crtc = dp_to_dig_port(dev_priv->drrs.dp)->base.base.crtc; > > + crtc = dp_to_dig_port(i915->drrs.dp)->base.base.crtc; > > pipe = to_intel_crtc(crtc)->pipe; > > frontbuffer_bits &= INTEL_FRONTBUFFER_ALL_MASK(pipe); > > - dev_priv->drrs.busy_frontbuffer_bits &= ~frontbuffer_bits; > > + i915->drrs.busy_frontbuffer_bits &= ~frontbuffer_bits; > > /* flush means busy screen hence upclock */ > > - if (frontbuffer_bits && dev_priv->drrs.refresh_rate_type == DRRS_LOW_RR) > > - intel_dp_set_drrs_state(dev_priv, to_intel_crtc(crtc)->config, > > - dev_priv->drrs.dp->attached_connector->panel.fixed_mode->vrefresh); > > + if (frontbuffer_bits && i915->drrs.refresh_rate_type == DRRS_LOW_RR) > > + intel_dp_set_drrs_state(i915, to_intel_crtc(crtc)->config, > > + i915->drrs.dp->attached_connector->panel.fixed_mode->vrefresh); > > /* > > * flush also means no more activity hence schedule downclock, if all > > * other fbs are quiescent too > > */ > > - if (!dev_priv->drrs.busy_frontbuffer_bits) > > - schedule_delayed_work(&dev_priv->drrs.work, > > + if (!i915->drrs.busy_frontbuffer_bits) > > + schedule_delayed_work(&i915->drrs.work, > > msecs_to_jiffies(1000)); > > - mutex_unlock(&dev_priv->drrs.mutex); > > + mutex_unlock(&i915->drrs.mutex); > > } > > /** > > @@ -7157,18 +7157,18 @@ static struct drm_display_mode * > > intel_dp_drrs_init(struct intel_connector *connector, > > struct drm_display_mode *fixed_mode) > > { > > - struct drm_i915_private *dev_priv = to_i915(connector->base.dev); > > + struct drm_i915_private *i915 = to_i915(connector->base.dev); > > struct drm_display_mode *downclock_mode = NULL; > > - INIT_DELAYED_WORK(&dev_priv->drrs.work, intel_edp_drrs_downclock_work); > > - mutex_init(&dev_priv->drrs.mutex); > > + INIT_DELAYED_WORK(&i915->drrs.work, intel_edp_drrs_downclock_work); > > + mutex_init(&i915->drrs.mutex); > > - if (INTEL_GEN(dev_priv) <= 6) { > > + if (INTEL_GEN(i915) <= 6) { > > DRM_DEBUG_KMS("DRRS supported for Gen7 and above\n"); > > return NULL; > > } > > - if (dev_priv->vbt.drrs_type != SEAMLESS_DRRS_SUPPORT) { > > + if (i915->vbt.drrs_type != SEAMLESS_DRRS_SUPPORT) { > > DRM_DEBUG_KMS("VBT doesn't support DRRS\n"); > > return NULL; > > } > > @@ -7179,9 +7179,9 @@ intel_dp_drrs_init(struct intel_connector *connector, > > return NULL; > > } > > - dev_priv->drrs.type = dev_priv->vbt.drrs_type; > > + i915->drrs.type = i915->vbt.drrs_type; > > - dev_priv->drrs.refresh_rate_type = DRRS_HIGH_RR; > > + i915->drrs.refresh_rate_type = DRRS_HIGH_RR; > > DRM_DEBUG_KMS("seamless DRRS supported for eDP panel.\n"); > > return downclock_mode; > > } > > @@ -7189,8 +7189,8 @@ intel_dp_drrs_init(struct intel_connector *connector, > > static bool intel_edp_init_connector(struct intel_dp *intel_dp, > > struct intel_connector *intel_connector) > > { > > - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp); > > - struct drm_device *dev = &dev_priv->drm; > > + struct drm_i915_private *i915 = dp_to_i915(intel_dp); > > + struct drm_device *dev = &i915->drm; > > struct drm_connector *connector = &intel_connector->base; > > struct drm_display_mode *fixed_mode = NULL; > > struct drm_display_mode *downclock_mode = NULL; > > @@ -7210,8 +7210,8 @@ static bool intel_edp_init_connector(struct intel_dp *intel_dp, > > * eDP and LVDS bail out early in this case to prevent interfering > > * with an already powered-on LVDS power sequencer. > > */ > > - if (intel_get_lvds_encoder(dev_priv)) { > > - WARN_ON(!(HAS_PCH_IBX(dev_priv) || HAS_PCH_CPT(dev_priv))); > > + if (intel_get_lvds_encoder(i915)) { > > + WARN_ON(!(HAS_PCH_IBX(i915) || HAS_PCH_CPT(i915))); > > DRM_INFO("LVDS was detected, not registering eDP\n"); > > return false; > > @@ -7256,7 +7256,7 @@ static bool intel_edp_init_connector(struct intel_dp *intel_dp, > > fixed_mode = intel_panel_vbt_fixed_mode(intel_connector); > > mutex_unlock(&dev->mode_config.mutex); > > - if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) { > > + if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) { > > intel_dp->edp_notifier.notifier_call = edp_notify_handler; > > register_reboot_notifier(&intel_dp->edp_notifier); > > @@ -7330,7 +7330,7 @@ intel_dp_init_connector(struct intel_digital_port *intel_dig_port, > > struct intel_dp *intel_dp = &intel_dig_port->dp; > > struct intel_encoder *intel_encoder = &intel_dig_port->base; > > struct drm_device *dev = intel_encoder->base.dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > enum port port = intel_encoder->port; > > int type; > > @@ -7353,18 +7353,18 @@ intel_dp_init_connector(struct intel_digital_port *intel_dig_port, > > intel_dp->DP = I915_READ(intel_dp->output_reg); > > intel_dp->attached_connector = intel_connector; > > - if (intel_dp_is_port_edp(dev_priv, port)) { > > + if (intel_dp_is_port_edp(i915, port)) { > > /* > > * Currently we don't support eDP on TypeC ports, although in > > * theory it could work on TypeC legacy ports. > > */ > > - WARN_ON(intel_port_is_tc(dev_priv, port)); > > + WARN_ON(intel_port_is_tc(i915, port)); > > type = DRM_MODE_CONNECTOR_eDP; > > } else { > > type = DRM_MODE_CONNECTOR_DisplayPort; > > } > > - if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) > > + if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) > > intel_dp->active_pipe = vlv_active_pipe(intel_dp); > > /* > > @@ -7376,7 +7376,7 @@ intel_dp_init_connector(struct intel_digital_port *intel_dig_port, > > intel_encoder->type = INTEL_OUTPUT_EDP; > > /* eDP only on port B and/or C on vlv/chv */ > > - if (WARN_ON((IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) && > > + if (WARN_ON((IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) && > > intel_dp_is_edp(intel_dp) && > > port != PORT_B && port != PORT_C)) > > return false; > > @@ -7388,26 +7388,26 @@ intel_dp_init_connector(struct intel_digital_port *intel_dig_port, > > drm_connector_init(dev, connector, &intel_dp_connector_funcs, type); > > drm_connector_helper_add(connector, &intel_dp_connector_helper_funcs); > > - if (!HAS_GMCH(dev_priv)) > > + if (!HAS_GMCH(i915)) > > connector->interlace_allowed = true; > > connector->doublescan_allowed = 0; > > - if (INTEL_GEN(dev_priv) >= 11) > > + if (INTEL_GEN(i915) >= 11) > > connector->ycbcr_420_allowed = true; > > - intel_encoder->hpd_pin = intel_hpd_pin_default(dev_priv, port); > > + intel_encoder->hpd_pin = intel_hpd_pin_default(i915, port); > > intel_dp_aux_init(intel_dp); > > intel_connector_attach_encoder(intel_connector, intel_encoder); > > - if (HAS_DDI(dev_priv)) > > + if (HAS_DDI(i915)) > > intel_connector->get_hw_state = intel_ddi_connector_get_hw_state; > > else > > intel_connector->get_hw_state = intel_connector_get_hw_state; > > /* init MST on ports that can support it */ > > - if (HAS_DP_MST(dev_priv) && !intel_dp_is_edp(intel_dp) && > > + if (HAS_DP_MST(i915) && !intel_dp_is_edp(intel_dp) && > > (port == PORT_B || port == PORT_C || > > port == PORT_D || port == PORT_F)) > > intel_dp_mst_encoder_init(intel_dig_port, > > @@ -7421,7 +7421,7 @@ intel_dp_init_connector(struct intel_digital_port *intel_dig_port, > > intel_dp_add_properties(intel_dp, connector); > > - if (is_hdcp_supported(dev_priv, port) && !intel_dp_is_edp(intel_dp)) { > > + if (is_hdcp_supported(i915, port) && !intel_dp_is_edp(intel_dp)) { > > int ret = intel_hdcp_init(intel_connector, &intel_dp_hdcp_shim); > > if (ret) > > DRM_DEBUG_KMS("HDCP init failed, skipping.\n"); > > @@ -7431,7 +7431,7 @@ intel_dp_init_connector(struct intel_digital_port *intel_dig_port, > > * 0xd. Failure to do so will result in spurious interrupts being > > * generated on the port when a cable is not attached. > > */ > > - if (IS_G45(dev_priv)) { > > + if (IS_G45(i915)) { > > u32 temp = I915_READ(PEG_BAND_GAP_DATA); > > I915_WRITE(PEG_BAND_GAP_DATA, (temp & ~0xf) | 0xd); > > } > > @@ -7444,7 +7444,7 @@ intel_dp_init_connector(struct intel_digital_port *intel_dig_port, > > return false; > > } > > -bool intel_dp_init(struct drm_i915_private *dev_priv, > > +bool intel_dp_init(struct drm_i915_private *i915, > > i915_reg_t output_reg, > > enum port port) > > { > > @@ -7464,7 +7464,7 @@ bool intel_dp_init(struct drm_i915_private *dev_priv, > > intel_encoder = &intel_dig_port->base; > > encoder = &intel_encoder->base; > > - if (drm_encoder_init(&dev_priv->drm, &intel_encoder->base, > > + if (drm_encoder_init(&i915->drm, &intel_encoder->base, > > &intel_dp_enc_funcs, DRM_MODE_ENCODER_TMDS, > > "DP %c", port_name(port))) > > goto err_encoder_init; > > @@ -7475,14 +7475,14 @@ bool intel_dp_init(struct drm_i915_private *dev_priv, > > intel_encoder->get_config = intel_dp_get_config; > > intel_encoder->update_pipe = intel_panel_update_backlight; > > intel_encoder->suspend = intel_dp_encoder_suspend; > > - if (IS_CHERRYVIEW(dev_priv)) { > > + if (IS_CHERRYVIEW(i915)) { > > intel_encoder->pre_pll_enable = chv_dp_pre_pll_enable; > > intel_encoder->pre_enable = chv_pre_enable_dp; > > intel_encoder->enable = vlv_enable_dp; > > intel_encoder->disable = vlv_disable_dp; > > intel_encoder->post_disable = chv_post_disable_dp; > > intel_encoder->post_pll_disable = chv_dp_post_pll_disable; > > - } else if (IS_VALLEYVIEW(dev_priv)) { > > + } else if (IS_VALLEYVIEW(i915)) { > > intel_encoder->pre_pll_enable = vlv_dp_pre_pll_enable; > > intel_encoder->pre_enable = vlv_pre_enable_dp; > > intel_encoder->enable = vlv_enable_dp; > > @@ -7500,7 +7500,7 @@ bool intel_dp_init(struct drm_i915_private *dev_priv, > > intel_encoder->type = INTEL_OUTPUT_DP; > > intel_encoder->power_domain = intel_port_to_power_domain(port); > > - if (IS_CHERRYVIEW(dev_priv)) { > > + if (IS_CHERRYVIEW(i915)) { > > if (port == PORT_D) > > intel_encoder->crtc_mask = 1 << 2; > > else > > @@ -7516,7 +7516,7 @@ bool intel_dp_init(struct drm_i915_private *dev_priv, > > if (port != PORT_A) > > intel_infoframe_init(intel_dig_port); > > - intel_dig_port->aux_ch = intel_bios_port_aux_ch(dev_priv, port); > > + intel_dig_port->aux_ch = intel_bios_port_aux_ch(i915, port); > > if (!intel_dp_init_connector(intel_dig_port, intel_connector)) > > goto err_init_connector; > > @@ -7531,11 +7531,11 @@ bool intel_dp_init(struct drm_i915_private *dev_priv, > > return false; > > } > > -void intel_dp_mst_suspend(struct drm_i915_private *dev_priv) > > +void intel_dp_mst_suspend(struct drm_i915_private *i915) > > { > > struct intel_encoder *encoder; > > - for_each_intel_encoder(&dev_priv->drm, encoder) { > > + for_each_intel_encoder(&i915->drm, encoder) { > > struct intel_dp *intel_dp; > > if (encoder->type != INTEL_OUTPUT_DDI) > > @@ -7551,11 +7551,11 @@ void intel_dp_mst_suspend(struct drm_i915_private *dev_priv) > > } > > } > > -void intel_dp_mst_resume(struct drm_i915_private *dev_priv) > > +void intel_dp_mst_resume(struct drm_i915_private *i915) > > { > > struct intel_encoder *encoder; > > - for_each_intel_encoder(&dev_priv->drm, encoder) { > > + for_each_intel_encoder(&i915->drm, encoder) { > > struct intel_dp *intel_dp; > > int ret; > > diff --git a/drivers/gpu/drm/i915/intel_dp.h b/drivers/gpu/drm/i915/intel_dp.h > > index da70b1a41c83..b6fecb185b23 100644 > > --- a/drivers/gpu/drm/i915/intel_dp.h > > +++ b/drivers/gpu/drm/i915/intel_dp.h > > @@ -35,10 +35,10 @@ void intel_dp_adjust_compliance_config(struct intel_dp *intel_dp, > > bool intel_dp_limited_color_range(const struct intel_crtc_state *crtc_state, > > const struct drm_connector_state *conn_state); > > int intel_dp_min_bpp(const struct intel_crtc_state *crtc_state); > > -bool intel_dp_port_enabled(struct drm_i915_private *dev_priv, > > +bool intel_dp_port_enabled(struct drm_i915_private *i915, > > i915_reg_t dp_reg, enum port port, > > enum pipe *pipe); > > -bool intel_dp_init(struct drm_i915_private *dev_priv, i915_reg_t output_reg, > > +bool intel_dp_init(struct drm_i915_private *i915, i915_reg_t output_reg, > > enum port port); > > bool intel_dp_init_connector(struct intel_digital_port *intel_dig_port, > > struct intel_connector *intel_connector); > > @@ -60,7 +60,7 @@ int intel_dp_compute_config(struct intel_encoder *encoder, > > struct intel_crtc_state *pipe_config, > > struct drm_connector_state *conn_state); > > bool intel_dp_is_edp(struct intel_dp *intel_dp); > > -bool intel_dp_is_port_edp(struct drm_i915_private *dev_priv, enum port port); > > +bool intel_dp_is_port_edp(struct drm_i915_private *i915, enum port port); > > enum irqreturn intel_dp_hpd_pulse(struct intel_digital_port *intel_dig_port, > > bool long_hpd); > > void intel_edp_backlight_on(const struct intel_crtc_state *crtc_state, > > @@ -69,21 +69,21 @@ void intel_edp_backlight_off(const struct drm_connector_state *conn_state); > > void intel_edp_panel_vdd_on(struct intel_dp *intel_dp); > > void intel_edp_panel_on(struct intel_dp *intel_dp); > > void intel_edp_panel_off(struct intel_dp *intel_dp); > > -void intel_dp_mst_suspend(struct drm_i915_private *dev_priv); > > -void intel_dp_mst_resume(struct drm_i915_private *dev_priv); > > +void intel_dp_mst_suspend(struct drm_i915_private *i915); > > +void intel_dp_mst_resume(struct drm_i915_private *i915); > > int intel_dp_max_link_rate(struct intel_dp *intel_dp); > > int intel_dp_max_lane_count(struct intel_dp *intel_dp); > > int intel_dp_rate_select(struct intel_dp *intel_dp, int rate); > > -void intel_power_sequencer_reset(struct drm_i915_private *dev_priv); > > +void intel_power_sequencer_reset(struct drm_i915_private *i915); > > u32 intel_dp_pack_aux(const u8 *src, int src_bytes); > > void intel_edp_drrs_enable(struct intel_dp *intel_dp, > > const struct intel_crtc_state *crtc_state); > > void intel_edp_drrs_disable(struct intel_dp *intel_dp, > > const struct intel_crtc_state *crtc_state); > > -void intel_edp_drrs_invalidate(struct drm_i915_private *dev_priv, > > +void intel_edp_drrs_invalidate(struct drm_i915_private *i915, > > unsigned int frontbuffer_bits); > > -void intel_edp_drrs_flush(struct drm_i915_private *dev_priv, > > +void intel_edp_drrs_flush(struct drm_i915_private *i915, > > unsigned int frontbuffer_bits); > > void > > @@ -112,7 +112,7 @@ bool intel_dp_get_colorimetry_status(struct intel_dp *intel_dp); > > int intel_dp_link_required(int pixel_clock, int bpp); > > int intel_dp_max_data_rate(int max_link_clock, int max_lanes); > > bool intel_digital_port_connected(struct intel_encoder *encoder); > > -void icl_tc_phy_disconnect(struct drm_i915_private *dev_priv, > > +void icl_tc_phy_disconnect(struct drm_i915_private *i915, > > struct intel_digital_port *dig_port); > > static inline unsigned int intel_dp_unused_lane_mask(int lane_count) > > diff --git a/drivers/gpu/drm/i915/intel_dp_aux_backlight.c b/drivers/gpu/drm/i915/intel_dp_aux_backlight.c > > index 7ded95a334db..7a3bd29dcf0b 100644 > > --- a/drivers/gpu/drm/i915/intel_dp_aux_backlight.c > > +++ b/drivers/gpu/drm/i915/intel_dp_aux_backlight.c > > @@ -109,7 +109,7 @@ intel_dp_aux_set_backlight(const struct drm_connector_state *conn_state, u32 lev > > */ > > static bool intel_dp_aux_set_pwm_freq(struct intel_connector *connector) > > { > > - struct drm_i915_private *dev_priv = to_i915(connector->base.dev); > > + struct drm_i915_private *i915 = to_i915(connector->base.dev); > > struct intel_dp *intel_dp = enc_to_intel_dp(&connector->encoder->base); > > int freq, fxp, fxp_min, fxp_max, fxp_actual, f = 1; > > u8 pn, pn_min, pn_max; > > @@ -118,7 +118,7 @@ static bool intel_dp_aux_set_pwm_freq(struct intel_connector *connector) > > * Note that, if F x P is out of supported range, the maximum value or > > * minimum value will applied automatically. So no need to check that. > > */ > > - freq = dev_priv->vbt.backlight.pwm_freq_hz; > > + freq = i915->vbt.backlight.pwm_freq_hz; > > DRM_DEBUG_KMS("VBT defined backlight frequency %u Hz\n", freq); > > if (!freq) { > > DRM_DEBUG_KMS("Use panel default backlight frequency\n"); > > diff --git a/drivers/gpu/drm/i915/intel_dp_mst.c b/drivers/gpu/drm/i915/intel_dp_mst.c > > index 0caf645fbbb8..ab21dc690c9f 100644 > > --- a/drivers/gpu/drm/i915/intel_dp_mst.c > > +++ b/drivers/gpu/drm/i915/intel_dp_mst.c > > @@ -91,7 +91,7 @@ static int intel_dp_mst_compute_config(struct intel_encoder *encoder, > > struct intel_crtc_state *pipe_config, > > struct drm_connector_state *conn_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_dp_mst_encoder *intel_mst = enc_to_mst(&encoder->base); > > struct intel_dp *intel_dp = &intel_mst->primary->dp; > > struct intel_connector *connector = > > @@ -140,11 +140,11 @@ static int intel_dp_mst_compute_config(struct intel_encoder *encoder, > > pipe_config->limited_color_range = > > intel_dp_limited_color_range(pipe_config, conn_state); > > - if (IS_GEN9_LP(dev_priv)) > > + if (IS_GEN9_LP(i915)) > > pipe_config->lane_lat_optim_mask = > > bxt_ddi_phy_calc_lane_lat_optim_mask(pipe_config->lane_count); > > - intel_ddi_compute_min_voltage_level(dev_priv, pipe_config); > > + intel_ddi_compute_min_voltage_level(i915, pipe_config); > > return 0; > > } > > @@ -285,7 +285,7 @@ static void intel_mst_pre_enable_dp(struct intel_encoder *encoder, > > struct intel_dp_mst_encoder *intel_mst = enc_to_mst(&encoder->base); > > struct intel_digital_port *intel_dig_port = intel_mst->primary; > > struct intel_dp *intel_dp = &intel_dig_port->dp; > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > enum port port = intel_dig_port->base.port; > > struct intel_connector *connector = > > to_intel_connector(conn_state->connector); > > @@ -332,12 +332,12 @@ static void intel_mst_enable_dp(struct intel_encoder *encoder, > > struct intel_dp_mst_encoder *intel_mst = enc_to_mst(&encoder->base); > > struct intel_digital_port *intel_dig_port = intel_mst->primary; > > struct intel_dp *intel_dp = &intel_dig_port->dp; > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > enum port port = intel_dig_port->base.port; > > DRM_DEBUG_KMS("active links %d\n", intel_dp->active_mst_links); > > - if (intel_wait_for_register(&dev_priv->uncore, > > + if (intel_wait_for_register(&i915->uncore, > > DP_TP_STATUS(port), > > DP_TP_STATUS_ACT_SENT, > > DP_TP_STATUS_ACT_SENT, > > @@ -495,7 +495,7 @@ static struct drm_connector *intel_dp_add_mst_connector(struct drm_dp_mst_topolo > > struct intel_dp *intel_dp = container_of(mgr, struct intel_dp, mst_mgr); > > struct intel_digital_port *intel_dig_port = dp_to_dig_port(intel_dp); > > struct drm_device *dev = intel_dig_port->base.base.dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > struct intel_connector *intel_connector; > > struct drm_connector *connector; > > enum pipe pipe; > > @@ -520,7 +520,7 @@ static struct drm_connector *intel_dp_add_mst_connector(struct drm_dp_mst_topolo > > drm_connector_helper_add(connector, &intel_dp_mst_connector_helper_funcs); > > - for_each_pipe(dev_priv, pipe) { > > + for_each_pipe(i915, pipe) { > > struct drm_encoder *enc = > > &intel_dp->mst_encoders[pipe]->base.base; > > @@ -549,10 +549,10 @@ static struct drm_connector *intel_dp_add_mst_connector(struct drm_dp_mst_topolo > > static void intel_dp_register_mst_connector(struct drm_connector *connector) > > { > > - struct drm_i915_private *dev_priv = to_i915(connector->dev); > > + struct drm_i915_private *i915 = to_i915(connector->dev); > > - if (dev_priv->fbdev) > > - drm_fb_helper_add_one_connector(&dev_priv->fbdev->helper, > > + if (i915->fbdev) > > + drm_fb_helper_add_one_connector(&i915->fbdev->helper, > > connector); > > drm_connector_register(connector); > > @@ -561,13 +561,13 @@ static void intel_dp_register_mst_connector(struct drm_connector *connector) > > static void intel_dp_destroy_mst_connector(struct drm_dp_mst_topology_mgr *mgr, > > struct drm_connector *connector) > > { > > - struct drm_i915_private *dev_priv = to_i915(connector->dev); > > + struct drm_i915_private *i915 = to_i915(connector->dev); > > DRM_DEBUG_KMS("[CONNECTOR:%d:%s]\n", connector->base.id, connector->name); > > drm_connector_unregister(connector); > > - if (dev_priv->fbdev) > > - drm_fb_helper_remove_one_connector(&dev_priv->fbdev->helper, > > + if (i915->fbdev) > > + drm_fb_helper_remove_one_connector(&i915->fbdev->helper, > > connector); > > drm_connector_put(connector); > > @@ -622,10 +622,10 @@ static bool > > intel_dp_create_fake_mst_encoders(struct intel_digital_port *intel_dig_port) > > { > > struct intel_dp *intel_dp = &intel_dig_port->dp; > > - struct drm_i915_private *dev_priv = to_i915(intel_dig_port->base.base.dev); > > + struct drm_i915_private *i915 = to_i915(intel_dig_port->base.base.dev); > > enum pipe pipe; > > - for_each_pipe(dev_priv, pipe) > > + for_each_pipe(i915, pipe) > > intel_dp->mst_encoders[pipe] = intel_dp_create_fake_mst_encoder(intel_dig_port, pipe); > > return true; > > } > > diff --git a/drivers/gpu/drm/i915/intel_dpio_phy.c b/drivers/gpu/drm/i915/intel_dpio_phy.c > > index bdbe41759827..c85672bb1a7d 100644 > > --- a/drivers/gpu/drm/i915/intel_dpio_phy.c > > +++ b/drivers/gpu/drm/i915/intel_dpio_phy.c > > @@ -212,9 +212,9 @@ static const struct bxt_ddi_phy_info glk_ddi_phy_info[] = { > > }; > > static const struct bxt_ddi_phy_info * > > -bxt_get_phy_list(struct drm_i915_private *dev_priv, int *count) > > +bxt_get_phy_list(struct drm_i915_private *i915, int *count) > > { > > - if (IS_GEMINILAKE(dev_priv)) { > > + if (IS_GEMINILAKE(i915)) { > > *count = ARRAY_SIZE(glk_ddi_phy_info); > > return glk_ddi_phy_info; > > } else { > > @@ -224,22 +224,22 @@ bxt_get_phy_list(struct drm_i915_private *dev_priv, int *count) > > } > > static const struct bxt_ddi_phy_info * > > -bxt_get_phy_info(struct drm_i915_private *dev_priv, enum dpio_phy phy) > > +bxt_get_phy_info(struct drm_i915_private *i915, enum dpio_phy phy) > > { > > int count; > > const struct bxt_ddi_phy_info *phy_list = > > - bxt_get_phy_list(dev_priv, &count); > > + bxt_get_phy_list(i915, &count); > > return &phy_list[phy]; > > } > > -void bxt_port_to_phy_channel(struct drm_i915_private *dev_priv, enum port port, > > +void bxt_port_to_phy_channel(struct drm_i915_private *i915, enum port port, > > enum dpio_phy *phy, enum dpio_channel *ch) > > { > > const struct bxt_ddi_phy_info *phy_info, *phys; > > int i, count; > > - phys = bxt_get_phy_list(dev_priv, &count); > > + phys = bxt_get_phy_list(i915, &count); > > for (i = 0; i < count; i++) { > > phy_info = &phys[i]; > > @@ -263,7 +263,7 @@ void bxt_port_to_phy_channel(struct drm_i915_private *dev_priv, enum port port, > > *ch = DPIO_CH0; > > } > > -void bxt_ddi_phy_set_signal_level(struct drm_i915_private *dev_priv, > > +void bxt_ddi_phy_set_signal_level(struct drm_i915_private *i915, > > enum port port, u32 margin, u32 scale, > > u32 enable, u32 deemphasis) > > { > > @@ -271,7 +271,7 @@ void bxt_ddi_phy_set_signal_level(struct drm_i915_private *dev_priv, > > enum dpio_phy phy; > > enum dpio_channel ch; > > - bxt_port_to_phy_channel(dev_priv, port, &phy, &ch); > > + bxt_port_to_phy_channel(i915, port, &phy, &ch); > > /* > > * While we write to the group register to program all lanes at once we > > @@ -306,12 +306,12 @@ void bxt_ddi_phy_set_signal_level(struct drm_i915_private *dev_priv, > > I915_WRITE(BXT_PORT_PCS_DW10_GRP(phy, ch), val); > > } > > -bool bxt_ddi_phy_is_enabled(struct drm_i915_private *dev_priv, > > +bool bxt_ddi_phy_is_enabled(struct drm_i915_private *i915, > > enum dpio_phy phy) > > { > > const struct bxt_ddi_phy_info *phy_info; > > - phy_info = bxt_get_phy_info(dev_priv, phy); > > + phy_info = bxt_get_phy_info(i915, phy); > > if (!(I915_READ(BXT_P_CR_GT_DISP_PWRON) & phy_info->pwron_mask)) > > return false; > > @@ -334,37 +334,37 @@ bool bxt_ddi_phy_is_enabled(struct drm_i915_private *dev_priv, > > return true; > > } > > -static u32 bxt_get_grc(struct drm_i915_private *dev_priv, enum dpio_phy phy) > > +static u32 bxt_get_grc(struct drm_i915_private *i915, enum dpio_phy phy) > > { > > u32 val = I915_READ(BXT_PORT_REF_DW6(phy)); > > return (val & GRC_CODE_MASK) >> GRC_CODE_SHIFT; > > } > > -static void bxt_phy_wait_grc_done(struct drm_i915_private *dev_priv, > > +static void bxt_phy_wait_grc_done(struct drm_i915_private *i915, > > enum dpio_phy phy) > > { > > - if (intel_wait_for_register(&dev_priv->uncore, > > + if (intel_wait_for_register(&i915->uncore, > > BXT_PORT_REF_DW3(phy), > > GRC_DONE, GRC_DONE, > > 10)) > > DRM_ERROR("timeout waiting for PHY%d GRC\n", phy); > > } > > -static void _bxt_ddi_phy_init(struct drm_i915_private *dev_priv, > > +static void _bxt_ddi_phy_init(struct drm_i915_private *i915, > > enum dpio_phy phy) > > { > > const struct bxt_ddi_phy_info *phy_info; > > u32 val; > > - phy_info = bxt_get_phy_info(dev_priv, phy); > > + phy_info = bxt_get_phy_info(i915, phy); > > - if (bxt_ddi_phy_is_enabled(dev_priv, phy)) { > > + if (bxt_ddi_phy_is_enabled(i915, phy)) { > > /* Still read out the GRC value for state verification */ > > if (phy_info->rcomp_phy != -1) > > - dev_priv->bxt_phy_grc = bxt_get_grc(dev_priv, phy); > > + i915->bxt_phy_grc = bxt_get_grc(i915, phy); > > - if (bxt_ddi_phy_verify_state(dev_priv, phy)) { > > + if (bxt_ddi_phy_verify_state(i915, phy)) { > > DRM_DEBUG_DRIVER("DDI PHY %d already enabled, " > > "won't reprogram it\n", phy); > > return; > > @@ -386,7 +386,7 @@ static void _bxt_ddi_phy_init(struct drm_i915_private *dev_priv, > > * The flag should get set in 100us according to the HW team, but > > * use 1ms due to occasional timeouts observed with that. > > */ > > - if (intel_wait_for_register_fw(&dev_priv->uncore, > > + if (intel_wait_for_register_fw(&i915->uncore, > > BXT_PORT_CL1CM_DW0(phy), > > PHY_RESERVED | PHY_POWER_GOOD, > > PHY_POWER_GOOD, > > @@ -419,14 +419,14 @@ static void _bxt_ddi_phy_init(struct drm_i915_private *dev_priv, > > if (phy_info->rcomp_phy != -1) { > > u32 grc_code; > > - bxt_phy_wait_grc_done(dev_priv, phy_info->rcomp_phy); > > + bxt_phy_wait_grc_done(i915, phy_info->rcomp_phy); > > /* > > * PHY0 isn't connected to an RCOMP resistor so copy over > > * the corresponding calibrated value from PHY1, and disable > > * the automatic calibration on PHY0. > > */ > > - val = dev_priv->bxt_phy_grc = bxt_get_grc(dev_priv, > > + val = i915->bxt_phy_grc = bxt_get_grc(i915, > > phy_info->rcomp_phy); > > grc_code = val << GRC_CODE_FAST_SHIFT | > > val << GRC_CODE_SLOW_SHIFT | > > @@ -446,12 +446,12 @@ static void _bxt_ddi_phy_init(struct drm_i915_private *dev_priv, > > I915_WRITE(BXT_PHY_CTL_FAMILY(phy), val); > > } > > -void bxt_ddi_phy_uninit(struct drm_i915_private *dev_priv, enum dpio_phy phy) > > +void bxt_ddi_phy_uninit(struct drm_i915_private *i915, enum dpio_phy phy) > > { > > const struct bxt_ddi_phy_info *phy_info; > > u32 val; > > - phy_info = bxt_get_phy_info(dev_priv, phy); > > + phy_info = bxt_get_phy_info(i915, phy); > > val = I915_READ(BXT_PHY_CTL_FAMILY(phy)); > > val &= ~COMMON_RESET_DIS; > > @@ -462,34 +462,34 @@ void bxt_ddi_phy_uninit(struct drm_i915_private *dev_priv, enum dpio_phy phy) > > I915_WRITE(BXT_P_CR_GT_DISP_PWRON, val); > > } > > -void bxt_ddi_phy_init(struct drm_i915_private *dev_priv, enum dpio_phy phy) > > +void bxt_ddi_phy_init(struct drm_i915_private *i915, enum dpio_phy phy) > > { > > const struct bxt_ddi_phy_info *phy_info = > > - bxt_get_phy_info(dev_priv, phy); > > + bxt_get_phy_info(i915, phy); > > enum dpio_phy rcomp_phy = phy_info->rcomp_phy; > > bool was_enabled; > > - lockdep_assert_held(&dev_priv->power_domains.lock); > > + lockdep_assert_held(&i915->power_domains.lock); > > was_enabled = true; > > if (rcomp_phy != -1) > > - was_enabled = bxt_ddi_phy_is_enabled(dev_priv, rcomp_phy); > > + was_enabled = bxt_ddi_phy_is_enabled(i915, rcomp_phy); > > /* > > * We need to copy the GRC calibration value from rcomp_phy, > > * so make sure it's powered up. > > */ > > if (!was_enabled) > > - _bxt_ddi_phy_init(dev_priv, rcomp_phy); > > + _bxt_ddi_phy_init(i915, rcomp_phy); > > - _bxt_ddi_phy_init(dev_priv, phy); > > + _bxt_ddi_phy_init(i915, phy); > > if (!was_enabled) > > - bxt_ddi_phy_uninit(dev_priv, rcomp_phy); > > + bxt_ddi_phy_uninit(i915, rcomp_phy); > > } > > static bool __printf(6, 7) > > -__phy_reg_verify_state(struct drm_i915_private *dev_priv, enum dpio_phy phy, > > +__phy_reg_verify_state(struct drm_i915_private *i915, enum dpio_phy phy, > > i915_reg_t reg, u32 mask, u32 expected, > > const char *reg_fmt, ...) > > { > > @@ -515,20 +515,20 @@ __phy_reg_verify_state(struct drm_i915_private *dev_priv, enum dpio_phy phy, > > return false; > > } > > -bool bxt_ddi_phy_verify_state(struct drm_i915_private *dev_priv, > > +bool bxt_ddi_phy_verify_state(struct drm_i915_private *i915, > > enum dpio_phy phy) > > { > > const struct bxt_ddi_phy_info *phy_info; > > u32 mask; > > bool ok; > > - phy_info = bxt_get_phy_info(dev_priv, phy); > > + phy_info = bxt_get_phy_info(i915, phy); > > #define _CHK(reg, mask, exp, fmt, ...) \ > > - __phy_reg_verify_state(dev_priv, phy, reg, mask, exp, fmt, \ > > + __phy_reg_verify_state(i915, phy, reg, mask, exp, fmt, \ > > ## __VA_ARGS__) > > - if (!bxt_ddi_phy_is_enabled(dev_priv, phy)) > > + if (!bxt_ddi_phy_is_enabled(i915, phy)) > > return false; > > ok = true; > > @@ -552,7 +552,7 @@ bool bxt_ddi_phy_verify_state(struct drm_i915_private *dev_priv, > > "BXT_PORT_CL2CM_DW6(%d)", phy); > > if (phy_info->rcomp_phy != -1) { > > - u32 grc_code = dev_priv->bxt_phy_grc; > > + u32 grc_code = i915->bxt_phy_grc; > > grc_code = grc_code << GRC_CODE_FAST_SHIFT | > > grc_code << GRC_CODE_SLOW_SHIFT | > > @@ -591,13 +591,13 @@ bxt_ddi_phy_calc_lane_lat_optim_mask(u8 lane_count) > > void bxt_ddi_phy_set_lane_optim_mask(struct intel_encoder *encoder, > > u8 lane_lat_optim_mask) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > enum port port = encoder->port; > > enum dpio_phy phy; > > enum dpio_channel ch; > > int lane; > > - bxt_port_to_phy_channel(dev_priv, port, &phy, &ch); > > + bxt_port_to_phy_channel(i915, port, &phy, &ch); > > for (lane = 0; lane < 4; lane++) { > > u32 val = I915_READ(BXT_PORT_TX_DW14_LN(phy, ch, lane)); > > @@ -617,14 +617,14 @@ void bxt_ddi_phy_set_lane_optim_mask(struct intel_encoder *encoder, > > u8 > > bxt_ddi_phy_get_lane_lat_optim_mask(struct intel_encoder *encoder) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > enum port port = encoder->port; > > enum dpio_phy phy; > > enum dpio_channel ch; > > int lane; > > u8 mask; > > - bxt_port_to_phy_channel(dev_priv, port, &phy, &ch); > > + bxt_port_to_phy_channel(i915, port, &phy, &ch); > > mask = 0; > > for (lane = 0; lane < 4; lane++) { > > @@ -642,7 +642,7 @@ void chv_set_phy_signal_level(struct intel_encoder *encoder, > > u32 deemph_reg_value, u32 margin_reg_value, > > bool uniq_trans_scale) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_digital_port *dport = enc_to_dig_port(&encoder->base); > > struct intel_crtc *intel_crtc = to_intel_crtc(encoder->base.crtc); > > enum dpio_channel ch = vlv_dport_to_channel(dport); > > @@ -650,46 +650,46 @@ void chv_set_phy_signal_level(struct intel_encoder *encoder, > > u32 val; > > int i; > > - vlv_dpio_get(dev_priv); > > + vlv_dpio_get(i915); > > /* Clear calc init */ > > - val = vlv_dpio_read(dev_priv, pipe, VLV_PCS01_DW10(ch)); > > + val = vlv_dpio_read(i915, pipe, VLV_PCS01_DW10(ch)); > > val &= ~(DPIO_PCS_SWING_CALC_TX0_TX2 | DPIO_PCS_SWING_CALC_TX1_TX3); > > val &= ~(DPIO_PCS_TX1DEEMP_MASK | DPIO_PCS_TX2DEEMP_MASK); > > val |= DPIO_PCS_TX1DEEMP_9P5 | DPIO_PCS_TX2DEEMP_9P5; > > - vlv_dpio_write(dev_priv, pipe, VLV_PCS01_DW10(ch), val); > > + vlv_dpio_write(i915, pipe, VLV_PCS01_DW10(ch), val); > > if (intel_crtc->config->lane_count > 2) { > > - val = vlv_dpio_read(dev_priv, pipe, VLV_PCS23_DW10(ch)); > > + val = vlv_dpio_read(i915, pipe, VLV_PCS23_DW10(ch)); > > val &= ~(DPIO_PCS_SWING_CALC_TX0_TX2 | DPIO_PCS_SWING_CALC_TX1_TX3); > > val &= ~(DPIO_PCS_TX1DEEMP_MASK | DPIO_PCS_TX2DEEMP_MASK); > > val |= DPIO_PCS_TX1DEEMP_9P5 | DPIO_PCS_TX2DEEMP_9P5; > > - vlv_dpio_write(dev_priv, pipe, VLV_PCS23_DW10(ch), val); > > + vlv_dpio_write(i915, pipe, VLV_PCS23_DW10(ch), val); > > } > > - val = vlv_dpio_read(dev_priv, pipe, VLV_PCS01_DW9(ch)); > > + val = vlv_dpio_read(i915, pipe, VLV_PCS01_DW9(ch)); > > val &= ~(DPIO_PCS_TX1MARGIN_MASK | DPIO_PCS_TX2MARGIN_MASK); > > val |= DPIO_PCS_TX1MARGIN_000 | DPIO_PCS_TX2MARGIN_000; > > - vlv_dpio_write(dev_priv, pipe, VLV_PCS01_DW9(ch), val); > > + vlv_dpio_write(i915, pipe, VLV_PCS01_DW9(ch), val); > > if (intel_crtc->config->lane_count > 2) { > > - val = vlv_dpio_read(dev_priv, pipe, VLV_PCS23_DW9(ch)); > > + val = vlv_dpio_read(i915, pipe, VLV_PCS23_DW9(ch)); > > val &= ~(DPIO_PCS_TX1MARGIN_MASK | DPIO_PCS_TX2MARGIN_MASK); > > val |= DPIO_PCS_TX1MARGIN_000 | DPIO_PCS_TX2MARGIN_000; > > - vlv_dpio_write(dev_priv, pipe, VLV_PCS23_DW9(ch), val); > > + vlv_dpio_write(i915, pipe, VLV_PCS23_DW9(ch), val); > > } > > /* Program swing deemph */ > > for (i = 0; i < intel_crtc->config->lane_count; i++) { > > - val = vlv_dpio_read(dev_priv, pipe, CHV_TX_DW4(ch, i)); > > + val = vlv_dpio_read(i915, pipe, CHV_TX_DW4(ch, i)); > > val &= ~DPIO_SWING_DEEMPH9P5_MASK; > > val |= deemph_reg_value << DPIO_SWING_DEEMPH9P5_SHIFT; > > - vlv_dpio_write(dev_priv, pipe, CHV_TX_DW4(ch, i), val); > > + vlv_dpio_write(i915, pipe, CHV_TX_DW4(ch, i), val); > > } > > /* Program swing margin */ > > for (i = 0; i < intel_crtc->config->lane_count; i++) { > > - val = vlv_dpio_read(dev_priv, pipe, CHV_TX_DW2(ch, i)); > > + val = vlv_dpio_read(i915, pipe, CHV_TX_DW2(ch, i)); > > val &= ~DPIO_SWING_MARGIN000_MASK; > > val |= margin_reg_value << DPIO_SWING_MARGIN000_SHIFT; > > @@ -702,7 +702,7 @@ void chv_set_phy_signal_level(struct intel_encoder *encoder, > > val &= ~(0xff << DPIO_UNIQ_TRANS_SCALE_SHIFT); > > val |= 0x9a << DPIO_UNIQ_TRANS_SCALE_SHIFT; > > - vlv_dpio_write(dev_priv, pipe, CHV_TX_DW2(ch, i), val); > > + vlv_dpio_write(i915, pipe, CHV_TX_DW2(ch, i), val); > > } > > /* > > @@ -712,70 +712,70 @@ void chv_set_phy_signal_level(struct intel_encoder *encoder, > > * 27 for ch0 and ch1. > > */ > > for (i = 0; i < intel_crtc->config->lane_count; i++) { > > - val = vlv_dpio_read(dev_priv, pipe, CHV_TX_DW3(ch, i)); > > + val = vlv_dpio_read(i915, pipe, CHV_TX_DW3(ch, i)); > > if (uniq_trans_scale) > > val |= DPIO_TX_UNIQ_TRANS_SCALE_EN; > > else > > val &= ~DPIO_TX_UNIQ_TRANS_SCALE_EN; > > - vlv_dpio_write(dev_priv, pipe, CHV_TX_DW3(ch, i), val); > > + vlv_dpio_write(i915, pipe, CHV_TX_DW3(ch, i), val); > > } > > /* Start swing calculation */ > > - val = vlv_dpio_read(dev_priv, pipe, VLV_PCS01_DW10(ch)); > > + val = vlv_dpio_read(i915, pipe, VLV_PCS01_DW10(ch)); > > val |= DPIO_PCS_SWING_CALC_TX0_TX2 | DPIO_PCS_SWING_CALC_TX1_TX3; > > - vlv_dpio_write(dev_priv, pipe, VLV_PCS01_DW10(ch), val); > > + vlv_dpio_write(i915, pipe, VLV_PCS01_DW10(ch), val); > > if (intel_crtc->config->lane_count > 2) { > > - val = vlv_dpio_read(dev_priv, pipe, VLV_PCS23_DW10(ch)); > > + val = vlv_dpio_read(i915, pipe, VLV_PCS23_DW10(ch)); > > val |= DPIO_PCS_SWING_CALC_TX0_TX2 | DPIO_PCS_SWING_CALC_TX1_TX3; > > - vlv_dpio_write(dev_priv, pipe, VLV_PCS23_DW10(ch), val); > > + vlv_dpio_write(i915, pipe, VLV_PCS23_DW10(ch), val); > > } > > - vlv_dpio_put(dev_priv); > > + vlv_dpio_put(i915); > > } > > void chv_data_lane_soft_reset(struct intel_encoder *encoder, > > const struct intel_crtc_state *crtc_state, > > bool reset) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > enum dpio_channel ch = vlv_dport_to_channel(enc_to_dig_port(&encoder->base)); > > struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc); > > enum pipe pipe = crtc->pipe; > > u32 val; > > - val = vlv_dpio_read(dev_priv, pipe, VLV_PCS01_DW0(ch)); > > + val = vlv_dpio_read(i915, pipe, VLV_PCS01_DW0(ch)); > > if (reset) > > val &= ~(DPIO_PCS_TX_LANE2_RESET | DPIO_PCS_TX_LANE1_RESET); > > else > > val |= DPIO_PCS_TX_LANE2_RESET | DPIO_PCS_TX_LANE1_RESET; > > - vlv_dpio_write(dev_priv, pipe, VLV_PCS01_DW0(ch), val); > > + vlv_dpio_write(i915, pipe, VLV_PCS01_DW0(ch), val); > > if (crtc_state->lane_count > 2) { > > - val = vlv_dpio_read(dev_priv, pipe, VLV_PCS23_DW0(ch)); > > + val = vlv_dpio_read(i915, pipe, VLV_PCS23_DW0(ch)); > > if (reset) > > val &= ~(DPIO_PCS_TX_LANE2_RESET | DPIO_PCS_TX_LANE1_RESET); > > else > > val |= DPIO_PCS_TX_LANE2_RESET | DPIO_PCS_TX_LANE1_RESET; > > - vlv_dpio_write(dev_priv, pipe, VLV_PCS23_DW0(ch), val); > > + vlv_dpio_write(i915, pipe, VLV_PCS23_DW0(ch), val); > > } > > - val = vlv_dpio_read(dev_priv, pipe, VLV_PCS01_DW1(ch)); > > + val = vlv_dpio_read(i915, pipe, VLV_PCS01_DW1(ch)); > > val |= CHV_PCS_REQ_SOFTRESET_EN; > > if (reset) > > val &= ~DPIO_PCS_CLK_SOFT_RESET; > > else > > val |= DPIO_PCS_CLK_SOFT_RESET; > > - vlv_dpio_write(dev_priv, pipe, VLV_PCS01_DW1(ch), val); > > + vlv_dpio_write(i915, pipe, VLV_PCS01_DW1(ch), val); > > if (crtc_state->lane_count > 2) { > > - val = vlv_dpio_read(dev_priv, pipe, VLV_PCS23_DW1(ch)); > > + val = vlv_dpio_read(i915, pipe, VLV_PCS23_DW1(ch)); > > val |= CHV_PCS_REQ_SOFTRESET_EN; > > if (reset) > > val &= ~DPIO_PCS_CLK_SOFT_RESET; > > else > > val |= DPIO_PCS_CLK_SOFT_RESET; > > - vlv_dpio_write(dev_priv, pipe, VLV_PCS23_DW1(ch), val); > > + vlv_dpio_write(i915, pipe, VLV_PCS23_DW1(ch), val); > > } > > } > > @@ -783,7 +783,7 @@ void chv_phy_pre_pll_enable(struct intel_encoder *encoder, > > const struct intel_crtc_state *crtc_state) > > { > > struct intel_digital_port *dport = enc_to_dig_port(&encoder->base); > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc); > > enum dpio_channel ch = vlv_dport_to_channel(dport); > > enum pipe pipe = crtc->pipe; > > @@ -797,51 +797,51 @@ void chv_phy_pre_pll_enable(struct intel_encoder *encoder, > > */ > > if (ch == DPIO_CH0 && pipe == PIPE_B) > > dport->release_cl2_override = > > - !chv_phy_powergate_ch(dev_priv, DPIO_PHY0, DPIO_CH1, true); > > + !chv_phy_powergate_ch(i915, DPIO_PHY0, DPIO_CH1, true); > > chv_phy_powergate_lanes(encoder, true, lane_mask); > > - vlv_dpio_get(dev_priv); > > + vlv_dpio_get(i915); > > /* Assert data lane reset */ > > chv_data_lane_soft_reset(encoder, crtc_state, true); > > /* program left/right clock distribution */ > > if (pipe != PIPE_B) { > > - val = vlv_dpio_read(dev_priv, pipe, _CHV_CMN_DW5_CH0); > > + val = vlv_dpio_read(i915, pipe, _CHV_CMN_DW5_CH0); > > val &= ~(CHV_BUFLEFTENA1_MASK | CHV_BUFRIGHTENA1_MASK); > > if (ch == DPIO_CH0) > > val |= CHV_BUFLEFTENA1_FORCE; > > if (ch == DPIO_CH1) > > val |= CHV_BUFRIGHTENA1_FORCE; > > - vlv_dpio_write(dev_priv, pipe, _CHV_CMN_DW5_CH0, val); > > + vlv_dpio_write(i915, pipe, _CHV_CMN_DW5_CH0, val); > > } else { > > - val = vlv_dpio_read(dev_priv, pipe, _CHV_CMN_DW1_CH1); > > + val = vlv_dpio_read(i915, pipe, _CHV_CMN_DW1_CH1); > > val &= ~(CHV_BUFLEFTENA2_MASK | CHV_BUFRIGHTENA2_MASK); > > if (ch == DPIO_CH0) > > val |= CHV_BUFLEFTENA2_FORCE; > > if (ch == DPIO_CH1) > > val |= CHV_BUFRIGHTENA2_FORCE; > > - vlv_dpio_write(dev_priv, pipe, _CHV_CMN_DW1_CH1, val); > > + vlv_dpio_write(i915, pipe, _CHV_CMN_DW1_CH1, val); > > } > > /* program clock channel usage */ > > - val = vlv_dpio_read(dev_priv, pipe, VLV_PCS01_DW8(ch)); > > + val = vlv_dpio_read(i915, pipe, VLV_PCS01_DW8(ch)); > > val |= CHV_PCS_USEDCLKCHANNEL_OVRRIDE; > > if (pipe != PIPE_B) > > val &= ~CHV_PCS_USEDCLKCHANNEL; > > else > > val |= CHV_PCS_USEDCLKCHANNEL; > > - vlv_dpio_write(dev_priv, pipe, VLV_PCS01_DW8(ch), val); > > + vlv_dpio_write(i915, pipe, VLV_PCS01_DW8(ch), val); > > if (crtc_state->lane_count > 2) { > > - val = vlv_dpio_read(dev_priv, pipe, VLV_PCS23_DW8(ch)); > > + val = vlv_dpio_read(i915, pipe, VLV_PCS23_DW8(ch)); > > val |= CHV_PCS_USEDCLKCHANNEL_OVRRIDE; > > if (pipe != PIPE_B) > > val &= ~CHV_PCS_USEDCLKCHANNEL; > > else > > val |= CHV_PCS_USEDCLKCHANNEL; > > - vlv_dpio_write(dev_priv, pipe, VLV_PCS23_DW8(ch), val); > > + vlv_dpio_write(i915, pipe, VLV_PCS23_DW8(ch), val); > > } > > /* > > @@ -849,14 +849,14 @@ void chv_phy_pre_pll_enable(struct intel_encoder *encoder, > > * matches the pipe, but here we need to > > * pick the CL based on the port. > > */ > > - val = vlv_dpio_read(dev_priv, pipe, CHV_CMN_DW19(ch)); > > + val = vlv_dpio_read(i915, pipe, CHV_CMN_DW19(ch)); > > if (pipe != PIPE_B) > > val &= ~CHV_CMN_USEDCLKCHANNEL; > > else > > val |= CHV_CMN_USEDCLKCHANNEL; > > - vlv_dpio_write(dev_priv, pipe, CHV_CMN_DW19(ch), val); > > + vlv_dpio_write(i915, pipe, CHV_CMN_DW19(ch), val); > > - vlv_dpio_put(dev_priv); > > + vlv_dpio_put(i915); > > } > > void chv_phy_pre_encoder_enable(struct intel_encoder *encoder, > > @@ -864,24 +864,24 @@ void chv_phy_pre_encoder_enable(struct intel_encoder *encoder, > > { > > struct intel_dp *intel_dp = enc_to_intel_dp(&encoder->base); > > struct intel_digital_port *dport = dp_to_dig_port(intel_dp); > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc); > > enum dpio_channel ch = vlv_dport_to_channel(dport); > > enum pipe pipe = crtc->pipe; > > int data, i, stagger; > > u32 val; > > - vlv_dpio_get(dev_priv); > > + vlv_dpio_get(i915); > > /* allow hardware to manage TX FIFO reset source */ > > - val = vlv_dpio_read(dev_priv, pipe, VLV_PCS01_DW11(ch)); > > + val = vlv_dpio_read(i915, pipe, VLV_PCS01_DW11(ch)); > > val &= ~DPIO_LANEDESKEW_STRAP_OVRD; > > - vlv_dpio_write(dev_priv, pipe, VLV_PCS01_DW11(ch), val); > > + vlv_dpio_write(i915, pipe, VLV_PCS01_DW11(ch), val); > > if (crtc_state->lane_count > 2) { > > - val = vlv_dpio_read(dev_priv, pipe, VLV_PCS23_DW11(ch)); > > + val = vlv_dpio_read(i915, pipe, VLV_PCS23_DW11(ch)); > > val &= ~DPIO_LANEDESKEW_STRAP_OVRD; > > - vlv_dpio_write(dev_priv, pipe, VLV_PCS23_DW11(ch), val); > > + vlv_dpio_write(i915, pipe, VLV_PCS23_DW11(ch), val); > > } > > /* Program Tx lane latency optimal setting*/ > > @@ -891,7 +891,7 @@ void chv_phy_pre_encoder_enable(struct intel_encoder *encoder, > > data = 0x0; > > else > > data = (i == 1) ? 0x0 : 0x1; > > - vlv_dpio_write(dev_priv, pipe, CHV_TX_DW14(ch, i), > > + vlv_dpio_write(i915, pipe, CHV_TX_DW14(ch, i), > > data << DPIO_UPAR_SHIFT); > > } > > @@ -907,17 +907,17 @@ void chv_phy_pre_encoder_enable(struct intel_encoder *encoder, > > else > > stagger = 0x2; > > - val = vlv_dpio_read(dev_priv, pipe, VLV_PCS01_DW11(ch)); > > + val = vlv_dpio_read(i915, pipe, VLV_PCS01_DW11(ch)); > > val |= DPIO_TX2_STAGGER_MASK(0x1f); > > - vlv_dpio_write(dev_priv, pipe, VLV_PCS01_DW11(ch), val); > > + vlv_dpio_write(i915, pipe, VLV_PCS01_DW11(ch), val); > > if (crtc_state->lane_count > 2) { > > - val = vlv_dpio_read(dev_priv, pipe, VLV_PCS23_DW11(ch)); > > + val = vlv_dpio_read(i915, pipe, VLV_PCS23_DW11(ch)); > > val |= DPIO_TX2_STAGGER_MASK(0x1f); > > - vlv_dpio_write(dev_priv, pipe, VLV_PCS23_DW11(ch), val); > > + vlv_dpio_write(i915, pipe, VLV_PCS23_DW11(ch), val); > > } > > - vlv_dpio_write(dev_priv, pipe, VLV_PCS01_DW12(ch), > > + vlv_dpio_write(i915, pipe, VLV_PCS01_DW12(ch), > > DPIO_LANESTAGGER_STRAP(stagger) | > > DPIO_LANESTAGGER_STRAP_OVRD | > > DPIO_TX1_STAGGER_MASK(0x1f) | > > @@ -925,7 +925,7 @@ void chv_phy_pre_encoder_enable(struct intel_encoder *encoder, > > DPIO_TX2_STAGGER_MULT(0)); > > if (crtc_state->lane_count > 2) { > > - vlv_dpio_write(dev_priv, pipe, VLV_PCS23_DW12(ch), > > + vlv_dpio_write(i915, pipe, VLV_PCS23_DW12(ch), > > DPIO_LANESTAGGER_STRAP(stagger) | > > DPIO_LANESTAGGER_STRAP_OVRD | > > DPIO_TX1_STAGGER_MASK(0x1f) | > > @@ -936,16 +936,16 @@ void chv_phy_pre_encoder_enable(struct intel_encoder *encoder, > > /* Deassert data lane reset */ > > chv_data_lane_soft_reset(encoder, crtc_state, false); > > - vlv_dpio_put(dev_priv); > > + vlv_dpio_put(i915); > > } > > void chv_phy_release_cl2_override(struct intel_encoder *encoder) > > { > > struct intel_digital_port *dport = enc_to_dig_port(&encoder->base); > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > if (dport->release_cl2_override) { > > - chv_phy_powergate_ch(dev_priv, DPIO_PHY0, DPIO_CH1, false); > > + chv_phy_powergate_ch(i915, DPIO_PHY0, DPIO_CH1, false); > > dport->release_cl2_override = false; > > } > > } > > @@ -953,24 +953,24 @@ void chv_phy_release_cl2_override(struct intel_encoder *encoder) > > void chv_phy_post_pll_disable(struct intel_encoder *encoder, > > const struct intel_crtc_state *old_crtc_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > enum pipe pipe = to_intel_crtc(old_crtc_state->base.crtc)->pipe; > > u32 val; > > - vlv_dpio_get(dev_priv); > > + vlv_dpio_get(i915); > > /* disable left/right clock distribution */ > > if (pipe != PIPE_B) { > > - val = vlv_dpio_read(dev_priv, pipe, _CHV_CMN_DW5_CH0); > > + val = vlv_dpio_read(i915, pipe, _CHV_CMN_DW5_CH0); > > val &= ~(CHV_BUFLEFTENA1_MASK | CHV_BUFRIGHTENA1_MASK); > > - vlv_dpio_write(dev_priv, pipe, _CHV_CMN_DW5_CH0, val); > > + vlv_dpio_write(i915, pipe, _CHV_CMN_DW5_CH0, val); > > } else { > > - val = vlv_dpio_read(dev_priv, pipe, _CHV_CMN_DW1_CH1); > > + val = vlv_dpio_read(i915, pipe, _CHV_CMN_DW1_CH1); > > val &= ~(CHV_BUFLEFTENA2_MASK | CHV_BUFRIGHTENA2_MASK); > > - vlv_dpio_write(dev_priv, pipe, _CHV_CMN_DW1_CH1, val); > > + vlv_dpio_write(i915, pipe, _CHV_CMN_DW1_CH1, val); > > } > > - vlv_dpio_put(dev_priv); > > + vlv_dpio_put(i915); > > /* > > * Leave the power down bit cleared for at least one > > @@ -988,57 +988,57 @@ void vlv_set_phy_signal_level(struct intel_encoder *encoder, > > u32 demph_reg_value, u32 preemph_reg_value, > > u32 uniqtranscale_reg_value, u32 tx3_demph) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_crtc *intel_crtc = to_intel_crtc(encoder->base.crtc); > > struct intel_digital_port *dport = enc_to_dig_port(&encoder->base); > > enum dpio_channel port = vlv_dport_to_channel(dport); > > enum pipe pipe = intel_crtc->pipe; > > - vlv_dpio_get(dev_priv); > > + vlv_dpio_get(i915); > > - vlv_dpio_write(dev_priv, pipe, VLV_TX_DW5(port), 0x00000000); > > - vlv_dpio_write(dev_priv, pipe, VLV_TX_DW4(port), demph_reg_value); > > - vlv_dpio_write(dev_priv, pipe, VLV_TX_DW2(port), > > + vlv_dpio_write(i915, pipe, VLV_TX_DW5(port), 0x00000000); > > + vlv_dpio_write(i915, pipe, VLV_TX_DW4(port), demph_reg_value); > > + vlv_dpio_write(i915, pipe, VLV_TX_DW2(port), > > uniqtranscale_reg_value); > > - vlv_dpio_write(dev_priv, pipe, VLV_TX_DW3(port), 0x0C782040); > > + vlv_dpio_write(i915, pipe, VLV_TX_DW3(port), 0x0C782040); > > if (tx3_demph) > > - vlv_dpio_write(dev_priv, pipe, VLV_TX3_DW4(port), tx3_demph); > > + vlv_dpio_write(i915, pipe, VLV_TX3_DW4(port), tx3_demph); > > - vlv_dpio_write(dev_priv, pipe, VLV_PCS_DW11(port), 0x00030000); > > - vlv_dpio_write(dev_priv, pipe, VLV_PCS_DW9(port), preemph_reg_value); > > - vlv_dpio_write(dev_priv, pipe, VLV_TX_DW5(port), DPIO_TX_OCALINIT_EN); > > + vlv_dpio_write(i915, pipe, VLV_PCS_DW11(port), 0x00030000); > > + vlv_dpio_write(i915, pipe, VLV_PCS_DW9(port), preemph_reg_value); > > + vlv_dpio_write(i915, pipe, VLV_TX_DW5(port), DPIO_TX_OCALINIT_EN); > > - vlv_dpio_put(dev_priv); > > + vlv_dpio_put(i915); > > } > > void vlv_phy_pre_pll_enable(struct intel_encoder *encoder, > > const struct intel_crtc_state *crtc_state) > > { > > struct intel_digital_port *dport = enc_to_dig_port(&encoder->base); > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc); > > enum dpio_channel port = vlv_dport_to_channel(dport); > > enum pipe pipe = crtc->pipe; > > /* Program Tx lane resets to default */ > > - vlv_dpio_get(dev_priv); > > + vlv_dpio_get(i915); > > - vlv_dpio_write(dev_priv, pipe, VLV_PCS_DW0(port), > > + vlv_dpio_write(i915, pipe, VLV_PCS_DW0(port), > > DPIO_PCS_TX_LANE2_RESET | > > DPIO_PCS_TX_LANE1_RESET); > > - vlv_dpio_write(dev_priv, pipe, VLV_PCS_DW1(port), > > + vlv_dpio_write(i915, pipe, VLV_PCS_DW1(port), > > DPIO_PCS_CLK_CRI_RXEB_EIOS_EN | > > DPIO_PCS_CLK_CRI_RXDIGFILTSG_EN | > > (1<<DPIO_PCS_CLK_DATAWIDTH_SHIFT) | > > DPIO_PCS_CLK_SOFT_RESET); > > /* Fix up inter-pair skew failure */ > > - vlv_dpio_write(dev_priv, pipe, VLV_PCS_DW12(port), 0x00750f00); > > - vlv_dpio_write(dev_priv, pipe, VLV_TX_DW11(port), 0x00001500); > > - vlv_dpio_write(dev_priv, pipe, VLV_TX_DW14(port), 0x40400000); > > + vlv_dpio_write(i915, pipe, VLV_PCS_DW12(port), 0x00750f00); > > + vlv_dpio_write(i915, pipe, VLV_TX_DW11(port), 0x00001500); > > + vlv_dpio_write(i915, pipe, VLV_TX_DW14(port), 0x40400000); > > - vlv_dpio_put(dev_priv); > > + vlv_dpio_put(i915); > > } > > void vlv_phy_pre_encoder_enable(struct intel_encoder *encoder, > > @@ -1046,42 +1046,42 @@ void vlv_phy_pre_encoder_enable(struct intel_encoder *encoder, > > { > > struct intel_dp *intel_dp = enc_to_intel_dp(&encoder->base); > > struct intel_digital_port *dport = dp_to_dig_port(intel_dp); > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc); > > enum dpio_channel port = vlv_dport_to_channel(dport); > > enum pipe pipe = crtc->pipe; > > u32 val; > > - vlv_dpio_get(dev_priv); > > + vlv_dpio_get(i915); > > /* Enable clock channels for this port */ > > - val = vlv_dpio_read(dev_priv, pipe, VLV_PCS01_DW8(port)); > > + val = vlv_dpio_read(i915, pipe, VLV_PCS01_DW8(port)); > > val = 0; > > if (pipe) > > val |= (1<<21); > > else > > val &= ~(1<<21); > > val |= 0x001000c4; > > - vlv_dpio_write(dev_priv, pipe, VLV_PCS_DW8(port), val); > > + vlv_dpio_write(i915, pipe, VLV_PCS_DW8(port), val); > > /* Program lane clock */ > > - vlv_dpio_write(dev_priv, pipe, VLV_PCS_DW14(port), 0x00760018); > > - vlv_dpio_write(dev_priv, pipe, VLV_PCS_DW23(port), 0x00400888); > > + vlv_dpio_write(i915, pipe, VLV_PCS_DW14(port), 0x00760018); > > + vlv_dpio_write(i915, pipe, VLV_PCS_DW23(port), 0x00400888); > > - vlv_dpio_put(dev_priv); > > + vlv_dpio_put(i915); > > } > > void vlv_phy_reset_lanes(struct intel_encoder *encoder, > > const struct intel_crtc_state *old_crtc_state) > > { > > struct intel_digital_port *dport = enc_to_dig_port(&encoder->base); > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_crtc *crtc = to_intel_crtc(old_crtc_state->base.crtc); > > enum dpio_channel port = vlv_dport_to_channel(dport); > > enum pipe pipe = crtc->pipe; > > - vlv_dpio_get(dev_priv); > > - vlv_dpio_write(dev_priv, pipe, VLV_PCS_DW0(port), 0x00000000); > > - vlv_dpio_write(dev_priv, pipe, VLV_PCS_DW1(port), 0x00e00060); > > - vlv_dpio_put(dev_priv); > > + vlv_dpio_get(i915); > > + vlv_dpio_write(i915, pipe, VLV_PCS_DW0(port), 0x00000000); > > + vlv_dpio_write(i915, pipe, VLV_PCS_DW1(port), 0x00e00060); > > + vlv_dpio_put(i915); > > } > > diff --git a/drivers/gpu/drm/i915/intel_dpio_phy.h b/drivers/gpu/drm/i915/intel_dpio_phy.h > > index f418aab90b7e..d4c9393bebce 100644 > > --- a/drivers/gpu/drm/i915/intel_dpio_phy.h > > +++ b/drivers/gpu/drm/i915/intel_dpio_phy.h > > @@ -15,16 +15,16 @@ struct drm_i915_private; > > struct intel_crtc_state; > > struct intel_encoder; > > -void bxt_port_to_phy_channel(struct drm_i915_private *dev_priv, enum port port, > > +void bxt_port_to_phy_channel(struct drm_i915_private *i915, enum port port, > > enum dpio_phy *phy, enum dpio_channel *ch); > > -void bxt_ddi_phy_set_signal_level(struct drm_i915_private *dev_priv, > > +void bxt_ddi_phy_set_signal_level(struct drm_i915_private *i915, > > enum port port, u32 margin, u32 scale, > > u32 enable, u32 deemphasis); > > -void bxt_ddi_phy_init(struct drm_i915_private *dev_priv, enum dpio_phy phy); > > -void bxt_ddi_phy_uninit(struct drm_i915_private *dev_priv, enum dpio_phy phy); > > -bool bxt_ddi_phy_is_enabled(struct drm_i915_private *dev_priv, > > +void bxt_ddi_phy_init(struct drm_i915_private *i915, enum dpio_phy phy); > > +void bxt_ddi_phy_uninit(struct drm_i915_private *i915, enum dpio_phy phy); > > +bool bxt_ddi_phy_is_enabled(struct drm_i915_private *i915, > > enum dpio_phy phy); > > -bool bxt_ddi_phy_verify_state(struct drm_i915_private *dev_priv, > > +bool bxt_ddi_phy_verify_state(struct drm_i915_private *i915, > > enum dpio_phy phy); > > u8 bxt_ddi_phy_calc_lane_lat_optim_mask(u8 lane_count); > > void bxt_ddi_phy_set_lane_optim_mask(struct intel_encoder *encoder, > > diff --git a/drivers/gpu/drm/i915/intel_dpll_mgr.c b/drivers/gpu/drm/i915/intel_dpll_mgr.c > > index 2d4e7b9a7b9d..2dabcdf27df4 100644 > > --- a/drivers/gpu/drm/i915/intel_dpll_mgr.c > > +++ b/drivers/gpu/drm/i915/intel_dpll_mgr.c > > @@ -45,14 +45,14 @@ > > */ > > static void > > -intel_atomic_duplicate_dpll_state(struct drm_i915_private *dev_priv, > > +intel_atomic_duplicate_dpll_state(struct drm_i915_private *i915, > > struct intel_shared_dpll_state *shared_dpll) > > { > > enum intel_dpll_id i; > > /* Copy shared dpll state */ > > - for (i = 0; i < dev_priv->num_shared_dpll; i++) { > > - struct intel_shared_dpll *pll = &dev_priv->shared_dplls[i]; > > + for (i = 0; i < i915->num_shared_dpll; i++) { > > + struct intel_shared_dpll *pll = &i915->shared_dplls[i]; > > shared_dpll[i] = pll->state; > > } > > @@ -77,40 +77,40 @@ intel_atomic_get_shared_dpll_state(struct drm_atomic_state *s) > > /** > > * intel_get_shared_dpll_by_id - get a DPLL given its id > > - * @dev_priv: i915 device instance > > + * @i915: i915 device instance > > * @id: pll id > > * > > * Returns: > > * A pointer to the DPLL with @id > > */ > > struct intel_shared_dpll * > > -intel_get_shared_dpll_by_id(struct drm_i915_private *dev_priv, > > +intel_get_shared_dpll_by_id(struct drm_i915_private *i915, > > enum intel_dpll_id id) > > { > > - return &dev_priv->shared_dplls[id]; > > + return &i915->shared_dplls[id]; > > } > > /** > > * intel_get_shared_dpll_id - get the id of a DPLL > > - * @dev_priv: i915 device instance > > + * @i915: i915 device instance > > * @pll: the DPLL > > * > > * Returns: > > * The id of @pll > > */ > > enum intel_dpll_id > > -intel_get_shared_dpll_id(struct drm_i915_private *dev_priv, > > +intel_get_shared_dpll_id(struct drm_i915_private *i915, > > struct intel_shared_dpll *pll) > > { > > - if (WARN_ON(pll < dev_priv->shared_dplls|| > > - pll > &dev_priv->shared_dplls[dev_priv->num_shared_dpll])) > > + if (WARN_ON(pll < i915->shared_dplls|| > > + pll > &i915->shared_dplls[i915->num_shared_dpll])) > > return -1; > > - return (enum intel_dpll_id) (pll - dev_priv->shared_dplls); > > + return (enum intel_dpll_id) (pll - i915->shared_dplls); > > } > > /* For ILK+ */ > > -void assert_shared_dpll(struct drm_i915_private *dev_priv, > > +void assert_shared_dpll(struct drm_i915_private *i915, > > struct intel_shared_dpll *pll, > > bool state) > > { > > @@ -120,7 +120,7 @@ void assert_shared_dpll(struct drm_i915_private *dev_priv, > > if (WARN(!pll, "asserting DPLL %s with no DPLL\n", onoff(state))) > > return; > > - cur_state = pll->info->funcs->get_hw_state(dev_priv, pll, &hw_state); > > + cur_state = pll->info->funcs->get_hw_state(i915, pll, &hw_state); > > I915_STATE_WARN(cur_state != state, > > "%s assertion failure (expected %s, current %s)\n", > > pll->info->name, onoff(state), onoff(cur_state)); > > @@ -136,22 +136,22 @@ void assert_shared_dpll(struct drm_i915_private *dev_priv, > > void intel_prepare_shared_dpll(const struct intel_crtc_state *crtc_state) > > { > > struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc); > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > struct intel_shared_dpll *pll = crtc_state->shared_dpll; > > if (WARN_ON(pll == NULL)) > > return; > > - mutex_lock(&dev_priv->dpll_lock); > > + mutex_lock(&i915->dpll_lock); > > WARN_ON(!pll->state.crtc_mask); > > if (!pll->active_mask) { > > DRM_DEBUG_DRIVER("setting up %s\n", pll->info->name); > > WARN_ON(pll->on); > > - assert_shared_dpll_disabled(dev_priv, pll); > > + assert_shared_dpll_disabled(i915, pll); > > - pll->info->funcs->prepare(dev_priv, pll); > > + pll->info->funcs->prepare(i915, pll); > > } > > - mutex_unlock(&dev_priv->dpll_lock); > > + mutex_unlock(&i915->dpll_lock); > > } > > /** > > @@ -163,7 +163,7 @@ void intel_prepare_shared_dpll(const struct intel_crtc_state *crtc_state) > > void intel_enable_shared_dpll(const struct intel_crtc_state *crtc_state) > > { > > struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc); > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > struct intel_shared_dpll *pll = crtc_state->shared_dpll; > > unsigned int crtc_mask = drm_crtc_mask(&crtc->base); > > unsigned int old_mask; > > @@ -171,7 +171,7 @@ void intel_enable_shared_dpll(const struct intel_crtc_state *crtc_state) > > if (WARN_ON(pll == NULL)) > > return; > > - mutex_lock(&dev_priv->dpll_lock); > > + mutex_lock(&i915->dpll_lock); > > old_mask = pll->active_mask; > > if (WARN_ON(!(pll->state.crtc_mask & crtc_mask)) || > > @@ -186,17 +186,17 @@ void intel_enable_shared_dpll(const struct intel_crtc_state *crtc_state) > > if (old_mask) { > > WARN_ON(!pll->on); > > - assert_shared_dpll_enabled(dev_priv, pll); > > + assert_shared_dpll_enabled(i915, pll); > > goto out; > > } > > WARN_ON(pll->on); > > DRM_DEBUG_KMS("enabling %s\n", pll->info->name); > > - pll->info->funcs->enable(dev_priv, pll); > > + pll->info->funcs->enable(i915, pll); > > pll->on = true; > > out: > > - mutex_unlock(&dev_priv->dpll_lock); > > + mutex_unlock(&i915->dpll_lock); > > } > > /** > > @@ -208,18 +208,18 @@ void intel_enable_shared_dpll(const struct intel_crtc_state *crtc_state) > > void intel_disable_shared_dpll(const struct intel_crtc_state *crtc_state) > > { > > struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc); > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > struct intel_shared_dpll *pll = crtc_state->shared_dpll; > > unsigned int crtc_mask = drm_crtc_mask(&crtc->base); > > /* PCH only available on ILK+ */ > > - if (INTEL_GEN(dev_priv) < 5) > > + if (INTEL_GEN(i915) < 5) > > return; > > if (pll == NULL) > > return; > > - mutex_lock(&dev_priv->dpll_lock); > > + mutex_lock(&i915->dpll_lock); > > if (WARN_ON(!(pll->active_mask & crtc_mask))) > > goto out; > > @@ -227,7 +227,7 @@ void intel_disable_shared_dpll(const struct intel_crtc_state *crtc_state) > > pll->info->name, pll->active_mask, pll->on, > > crtc->base.base.id); > > - assert_shared_dpll_enabled(dev_priv, pll); > > + assert_shared_dpll_enabled(i915, pll); > > WARN_ON(!pll->on); > > pll->active_mask &= ~crtc_mask; > > @@ -235,11 +235,11 @@ void intel_disable_shared_dpll(const struct intel_crtc_state *crtc_state) > > goto out; > > DRM_DEBUG_KMS("disabling %s\n", pll->info->name); > > - pll->info->funcs->disable(dev_priv, pll); > > + pll->info->funcs->disable(i915, pll); > > pll->on = false; > > out: > > - mutex_unlock(&dev_priv->dpll_lock); > > + mutex_unlock(&i915->dpll_lock); > > } > > static struct intel_shared_dpll * > > @@ -248,7 +248,7 @@ intel_find_shared_dpll(struct intel_crtc_state *crtc_state, > > enum intel_dpll_id range_max) > > { > > struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc); > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > struct intel_shared_dpll *pll, *unused_pll = NULL; > > struct intel_shared_dpll_state *shared_dpll; > > enum intel_dpll_id i; > > @@ -256,7 +256,7 @@ intel_find_shared_dpll(struct intel_crtc_state *crtc_state, > > shared_dpll = intel_atomic_get_shared_dpll_state(crtc_state->base.state); > > for (i = range_min; i <= range_max; i++) { > > - pll = &dev_priv->shared_dplls[i]; > > + pll = &i915->shared_dplls[i]; > > /* Only want to check enabled timings first */ > > if (shared_dpll[i].crtc_mask == 0) { > > @@ -322,7 +322,7 @@ intel_reference_shared_dpll(struct intel_shared_dpll *pll, > > */ > > void intel_shared_dpll_swap_state(struct drm_atomic_state *state) > > { > > - struct drm_i915_private *dev_priv = to_i915(state->dev); > > + struct drm_i915_private *i915 = to_i915(state->dev); > > struct intel_shared_dpll_state *shared_dpll; > > struct intel_shared_dpll *pll; > > enum intel_dpll_id i; > > @@ -331,10 +331,10 @@ void intel_shared_dpll_swap_state(struct drm_atomic_state *state) > > return; > > shared_dpll = to_intel_atomic_state(state)->shared_dpll; > > - for (i = 0; i < dev_priv->num_shared_dpll; i++) { > > + for (i = 0; i < i915->num_shared_dpll; i++) { > > struct intel_shared_dpll_state tmp; > > - pll = &dev_priv->shared_dplls[i]; > > + pll = &i915->shared_dplls[i]; > > tmp = pll->state; > > pll->state = shared_dpll[i]; > > @@ -342,7 +342,7 @@ void intel_shared_dpll_swap_state(struct drm_atomic_state *state) > > } > > } > > -static bool ibx_pch_dpll_get_hw_state(struct drm_i915_private *dev_priv, > > +static bool ibx_pch_dpll_get_hw_state(struct drm_i915_private *i915, > > struct intel_shared_dpll *pll, > > struct intel_dpll_hw_state *hw_state) > > { > > @@ -350,7 +350,7 @@ static bool ibx_pch_dpll_get_hw_state(struct drm_i915_private *dev_priv, > > intel_wakeref_t wakeref; > > u32 val; > > - wakeref = intel_display_power_get_if_enabled(dev_priv, > > + wakeref = intel_display_power_get_if_enabled(i915, > > POWER_DOMAIN_DISPLAY_CORE); > > if (!wakeref) > > return false; > > @@ -360,12 +360,12 @@ static bool ibx_pch_dpll_get_hw_state(struct drm_i915_private *dev_priv, > > hw_state->fp0 = I915_READ(PCH_FP0(id)); > > hw_state->fp1 = I915_READ(PCH_FP1(id)); > > - intel_display_power_put(dev_priv, POWER_DOMAIN_DISPLAY_CORE, wakeref); > > + intel_display_power_put(i915, POWER_DOMAIN_DISPLAY_CORE, wakeref); > > return val & DPLL_VCO_ENABLE; > > } > > -static void ibx_pch_dpll_prepare(struct drm_i915_private *dev_priv, > > +static void ibx_pch_dpll_prepare(struct drm_i915_private *i915, > > struct intel_shared_dpll *pll) > > { > > const enum intel_dpll_id id = pll->info->id; > > @@ -374,12 +374,12 @@ static void ibx_pch_dpll_prepare(struct drm_i915_private *dev_priv, > > I915_WRITE(PCH_FP1(id), pll->state.hw_state.fp1); > > } > > -static void ibx_assert_pch_refclk_enabled(struct drm_i915_private *dev_priv) > > +static void ibx_assert_pch_refclk_enabled(struct drm_i915_private *i915) > > { > > u32 val; > > bool enabled; > > - I915_STATE_WARN_ON(!(HAS_PCH_IBX(dev_priv) || HAS_PCH_CPT(dev_priv))); > > + I915_STATE_WARN_ON(!(HAS_PCH_IBX(i915) || HAS_PCH_CPT(i915))); > > val = I915_READ(PCH_DREF_CONTROL); > > enabled = !!(val & (DREF_SSC_SOURCE_MASK | DREF_NONSPREAD_SOURCE_MASK | > > @@ -387,13 +387,13 @@ static void ibx_assert_pch_refclk_enabled(struct drm_i915_private *dev_priv) > > I915_STATE_WARN(!enabled, "PCH refclk assertion failure, should be active but is disabled\n"); > > } > > -static void ibx_pch_dpll_enable(struct drm_i915_private *dev_priv, > > +static void ibx_pch_dpll_enable(struct drm_i915_private *i915, > > struct intel_shared_dpll *pll) > > { > > const enum intel_dpll_id id = pll->info->id; > > /* PCH refclock must be enabled first */ > > - ibx_assert_pch_refclk_enabled(dev_priv); > > + ibx_assert_pch_refclk_enabled(i915); > > I915_WRITE(PCH_DPLL(id), pll->state.hw_state.dpll); > > @@ -411,7 +411,7 @@ static void ibx_pch_dpll_enable(struct drm_i915_private *dev_priv, > > udelay(200); > > } > > -static void ibx_pch_dpll_disable(struct drm_i915_private *dev_priv, > > +static void ibx_pch_dpll_disable(struct drm_i915_private *i915, > > struct intel_shared_dpll *pll) > > { > > const enum intel_dpll_id id = pll->info->id; > > @@ -426,14 +426,14 @@ ibx_get_dpll(struct intel_crtc_state *crtc_state, > > struct intel_encoder *encoder) > > { > > struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc); > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > struct intel_shared_dpll *pll; > > enum intel_dpll_id i; > > - if (HAS_PCH_IBX(dev_priv)) { > > + if (HAS_PCH_IBX(i915)) { > > /* Ironlake PCH has a fixed PLL->PCH pipe mapping. */ > > i = (enum intel_dpll_id) crtc->pipe; > > - pll = &dev_priv->shared_dplls[i]; > > + pll = &i915->shared_dplls[i]; > > DRM_DEBUG_KMS("[CRTC:%d:%s] using pre-allocated %s\n", > > crtc->base.base.id, crtc->base.name, > > @@ -453,7 +453,7 @@ ibx_get_dpll(struct intel_crtc_state *crtc_state, > > return pll; > > } > > -static void ibx_dump_hw_state(struct drm_i915_private *dev_priv, > > +static void ibx_dump_hw_state(struct drm_i915_private *i915, > > const struct intel_dpll_hw_state *hw_state) > > { > > DRM_DEBUG_KMS("dpll_hw_state: dpll: 0x%x, dpll_md: 0x%x, " > > @@ -471,7 +471,7 @@ static const struct intel_shared_dpll_funcs ibx_pch_dpll_funcs = { > > .get_hw_state = ibx_pch_dpll_get_hw_state, > > }; > > -static void hsw_ddi_wrpll_enable(struct drm_i915_private *dev_priv, > > +static void hsw_ddi_wrpll_enable(struct drm_i915_private *i915, > > struct intel_shared_dpll *pll) > > { > > const enum intel_dpll_id id = pll->info->id; > > @@ -481,7 +481,7 @@ static void hsw_ddi_wrpll_enable(struct drm_i915_private *dev_priv, > > udelay(20); > > } > > -static void hsw_ddi_spll_enable(struct drm_i915_private *dev_priv, > > +static void hsw_ddi_spll_enable(struct drm_i915_private *i915, > > struct intel_shared_dpll *pll) > > { > > I915_WRITE(SPLL_CTL, pll->state.hw_state.spll); > > @@ -489,7 +489,7 @@ static void hsw_ddi_spll_enable(struct drm_i915_private *dev_priv, > > udelay(20); > > } > > -static void hsw_ddi_wrpll_disable(struct drm_i915_private *dev_priv, > > +static void hsw_ddi_wrpll_disable(struct drm_i915_private *i915, > > struct intel_shared_dpll *pll) > > { > > const enum intel_dpll_id id = pll->info->id; > > @@ -500,7 +500,7 @@ static void hsw_ddi_wrpll_disable(struct drm_i915_private *dev_priv, > > POSTING_READ(WRPLL_CTL(id)); > > } > > -static void hsw_ddi_spll_disable(struct drm_i915_private *dev_priv, > > +static void hsw_ddi_spll_disable(struct drm_i915_private *i915, > > struct intel_shared_dpll *pll) > > { > > u32 val; > > @@ -510,7 +510,7 @@ static void hsw_ddi_spll_disable(struct drm_i915_private *dev_priv, > > POSTING_READ(SPLL_CTL); > > } > > -static bool hsw_ddi_wrpll_get_hw_state(struct drm_i915_private *dev_priv, > > +static bool hsw_ddi_wrpll_get_hw_state(struct drm_i915_private *i915, > > struct intel_shared_dpll *pll, > > struct intel_dpll_hw_state *hw_state) > > { > > @@ -518,7 +518,7 @@ static bool hsw_ddi_wrpll_get_hw_state(struct drm_i915_private *dev_priv, > > intel_wakeref_t wakeref; > > u32 val; > > - wakeref = intel_display_power_get_if_enabled(dev_priv, > > + wakeref = intel_display_power_get_if_enabled(i915, > > POWER_DOMAIN_DISPLAY_CORE); > > if (!wakeref) > > return false; > > @@ -526,19 +526,19 @@ static bool hsw_ddi_wrpll_get_hw_state(struct drm_i915_private *dev_priv, > > val = I915_READ(WRPLL_CTL(id)); > > hw_state->wrpll = val; > > - intel_display_power_put(dev_priv, POWER_DOMAIN_DISPLAY_CORE, wakeref); > > + intel_display_power_put(i915, POWER_DOMAIN_DISPLAY_CORE, wakeref); > > return val & WRPLL_PLL_ENABLE; > > } > > -static bool hsw_ddi_spll_get_hw_state(struct drm_i915_private *dev_priv, > > +static bool hsw_ddi_spll_get_hw_state(struct drm_i915_private *i915, > > struct intel_shared_dpll *pll, > > struct intel_dpll_hw_state *hw_state) > > { > > intel_wakeref_t wakeref; > > u32 val; > > - wakeref = intel_display_power_get_if_enabled(dev_priv, > > + wakeref = intel_display_power_get_if_enabled(i915, > > POWER_DOMAIN_DISPLAY_CORE); > > if (!wakeref) > > return false; > > @@ -546,7 +546,7 @@ static bool hsw_ddi_spll_get_hw_state(struct drm_i915_private *dev_priv, > > val = I915_READ(SPLL_CTL); > > hw_state->spll = val; > > - intel_display_power_put(dev_priv, POWER_DOMAIN_DISPLAY_CORE, wakeref); > > + intel_display_power_put(i915, POWER_DOMAIN_DISPLAY_CORE, wakeref); > > return val & SPLL_PLL_ENABLE; > > } > > @@ -793,7 +793,7 @@ static struct intel_shared_dpll *hsw_ddi_hdmi_get_dpll(struct intel_crtc_state * > > static struct intel_shared_dpll * > > hsw_ddi_dp_get_dpll(struct intel_crtc_state *crtc_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc_state->base.crtc->dev); > > + struct drm_i915_private *i915 = to_i915(crtc_state->base.crtc->dev); > > struct intel_shared_dpll *pll; > > enum intel_dpll_id pll_id; > > int clock = crtc_state->port_clock; > > @@ -813,7 +813,7 @@ hsw_ddi_dp_get_dpll(struct intel_crtc_state *crtc_state) > > return NULL; > > } > > - pll = intel_get_shared_dpll_by_id(dev_priv, pll_id); > > + pll = intel_get_shared_dpll_by_id(i915, pll_id); > > if (!pll) > > return NULL; > > @@ -855,7 +855,7 @@ hsw_get_dpll(struct intel_crtc_state *crtc_state, > > return pll; > > } > > -static void hsw_dump_hw_state(struct drm_i915_private *dev_priv, > > +static void hsw_dump_hw_state(struct drm_i915_private *i915, > > const struct intel_dpll_hw_state *hw_state) > > { > > DRM_DEBUG_KMS("dpll_hw_state: wrpll: 0x%x spll: 0x%x\n", > > @@ -874,17 +874,17 @@ static const struct intel_shared_dpll_funcs hsw_ddi_spll_funcs = { > > .get_hw_state = hsw_ddi_spll_get_hw_state, > > }; > > -static void hsw_ddi_lcpll_enable(struct drm_i915_private *dev_priv, > > +static void hsw_ddi_lcpll_enable(struct drm_i915_private *i915, > > struct intel_shared_dpll *pll) > > { > > } > > -static void hsw_ddi_lcpll_disable(struct drm_i915_private *dev_priv, > > +static void hsw_ddi_lcpll_disable(struct drm_i915_private *i915, > > struct intel_shared_dpll *pll) > > { > > } > > -static bool hsw_ddi_lcpll_get_hw_state(struct drm_i915_private *dev_priv, > > +static bool hsw_ddi_lcpll_get_hw_state(struct drm_i915_private *i915, > > struct intel_shared_dpll *pll, > > struct intel_dpll_hw_state *hw_state) > > { > > @@ -928,7 +928,7 @@ static const struct skl_dpll_regs skl_dpll_regs[4] = { > > }, > > }; > > -static void skl_ddi_pll_write_ctrl1(struct drm_i915_private *dev_priv, > > +static void skl_ddi_pll_write_ctrl1(struct drm_i915_private *i915, > > struct intel_shared_dpll *pll) > > { > > const enum intel_dpll_id id = pll->info->id; > > @@ -945,13 +945,13 @@ static void skl_ddi_pll_write_ctrl1(struct drm_i915_private *dev_priv, > > POSTING_READ(DPLL_CTRL1); > > } > > -static void skl_ddi_pll_enable(struct drm_i915_private *dev_priv, > > +static void skl_ddi_pll_enable(struct drm_i915_private *i915, > > struct intel_shared_dpll *pll) > > { > > const struct skl_dpll_regs *regs = skl_dpll_regs; > > const enum intel_dpll_id id = pll->info->id; > > - skl_ddi_pll_write_ctrl1(dev_priv, pll); > > + skl_ddi_pll_write_ctrl1(i915, pll); > > I915_WRITE(regs[id].cfgcr1, pll->state.hw_state.cfgcr1); > > I915_WRITE(regs[id].cfgcr2, pll->state.hw_state.cfgcr2); > > @@ -962,7 +962,7 @@ static void skl_ddi_pll_enable(struct drm_i915_private *dev_priv, > > I915_WRITE(regs[id].ctl, > > I915_READ(regs[id].ctl) | LCPLL_PLL_ENABLE); > > - if (intel_wait_for_register(&dev_priv->uncore, > > + if (intel_wait_for_register(&i915->uncore, > > DPLL_STATUS, > > DPLL_LOCK(id), > > DPLL_LOCK(id), > > @@ -970,13 +970,13 @@ static void skl_ddi_pll_enable(struct drm_i915_private *dev_priv, > > DRM_ERROR("DPLL %d not locked\n", id); > > } > > -static void skl_ddi_dpll0_enable(struct drm_i915_private *dev_priv, > > +static void skl_ddi_dpll0_enable(struct drm_i915_private *i915, > > struct intel_shared_dpll *pll) > > { > > - skl_ddi_pll_write_ctrl1(dev_priv, pll); > > + skl_ddi_pll_write_ctrl1(i915, pll); > > } > > -static void skl_ddi_pll_disable(struct drm_i915_private *dev_priv, > > +static void skl_ddi_pll_disable(struct drm_i915_private *i915, > > struct intel_shared_dpll *pll) > > { > > const struct skl_dpll_regs *regs = skl_dpll_regs; > > @@ -988,12 +988,12 @@ static void skl_ddi_pll_disable(struct drm_i915_private *dev_priv, > > POSTING_READ(regs[id].ctl); > > } > > -static void skl_ddi_dpll0_disable(struct drm_i915_private *dev_priv, > > +static void skl_ddi_dpll0_disable(struct drm_i915_private *i915, > > struct intel_shared_dpll *pll) > > { > > } > > -static bool skl_ddi_pll_get_hw_state(struct drm_i915_private *dev_priv, > > +static bool skl_ddi_pll_get_hw_state(struct drm_i915_private *i915, > > struct intel_shared_dpll *pll, > > struct intel_dpll_hw_state *hw_state) > > { > > @@ -1003,7 +1003,7 @@ static bool skl_ddi_pll_get_hw_state(struct drm_i915_private *dev_priv, > > intel_wakeref_t wakeref; > > bool ret; > > - wakeref = intel_display_power_get_if_enabled(dev_priv, > > + wakeref = intel_display_power_get_if_enabled(i915, > > POWER_DOMAIN_DISPLAY_CORE); > > if (!wakeref) > > return false; > > @@ -1025,12 +1025,12 @@ static bool skl_ddi_pll_get_hw_state(struct drm_i915_private *dev_priv, > > ret = true; > > out: > > - intel_display_power_put(dev_priv, POWER_DOMAIN_DISPLAY_CORE, wakeref); > > + intel_display_power_put(i915, POWER_DOMAIN_DISPLAY_CORE, wakeref); > > return ret; > > } > > -static bool skl_ddi_dpll0_get_hw_state(struct drm_i915_private *dev_priv, > > +static bool skl_ddi_dpll0_get_hw_state(struct drm_i915_private *i915, > > struct intel_shared_dpll *pll, > > struct intel_dpll_hw_state *hw_state) > > { > > @@ -1040,7 +1040,7 @@ static bool skl_ddi_dpll0_get_hw_state(struct drm_i915_private *dev_priv, > > u32 val; > > bool ret; > > - wakeref = intel_display_power_get_if_enabled(dev_priv, > > + wakeref = intel_display_power_get_if_enabled(i915, > > POWER_DOMAIN_DISPLAY_CORE); > > if (!wakeref) > > return false; > > @@ -1058,7 +1058,7 @@ static bool skl_ddi_dpll0_get_hw_state(struct drm_i915_private *dev_priv, > > ret = true; > > out: > > - intel_display_power_put(dev_priv, POWER_DOMAIN_DISPLAY_CORE, wakeref); > > + intel_display_power_put(i915, POWER_DOMAIN_DISPLAY_CORE, wakeref); > > return ret; > > } > > @@ -1424,7 +1424,7 @@ skl_get_dpll(struct intel_crtc_state *crtc_state, > > return pll; > > } > > -static void skl_dump_hw_state(struct drm_i915_private *dev_priv, > > +static void skl_dump_hw_state(struct drm_i915_private *i915, > > const struct intel_dpll_hw_state *hw_state) > > { > > DRM_DEBUG_KMS("dpll_hw_state: " > > @@ -1446,7 +1446,7 @@ static const struct intel_shared_dpll_funcs skl_ddi_dpll0_funcs = { > > .get_hw_state = skl_ddi_dpll0_get_hw_state, > > }; > > -static void bxt_ddi_pll_enable(struct drm_i915_private *dev_priv, > > +static void bxt_ddi_pll_enable(struct drm_i915_private *i915, > > struct intel_shared_dpll *pll) > > { > > u32 temp; > > @@ -1454,14 +1454,14 @@ static void bxt_ddi_pll_enable(struct drm_i915_private *dev_priv, > > enum dpio_phy phy; > > enum dpio_channel ch; > > - bxt_port_to_phy_channel(dev_priv, port, &phy, &ch); > > + bxt_port_to_phy_channel(i915, port, &phy, &ch); > > /* Non-SSC reference */ > > temp = I915_READ(BXT_PORT_PLL_ENABLE(port)); > > temp |= PORT_PLL_REF_SEL; > > I915_WRITE(BXT_PORT_PLL_ENABLE(port), temp); > > - if (IS_GEMINILAKE(dev_priv)) { > > + if (IS_GEMINILAKE(i915)) { > > temp = I915_READ(BXT_PORT_PLL_ENABLE(port)); > > temp |= PORT_PLL_POWER_ENABLE; > > I915_WRITE(BXT_PORT_PLL_ENABLE(port), temp); > > @@ -1549,7 +1549,7 @@ static void bxt_ddi_pll_enable(struct drm_i915_private *dev_priv, > > 200)) > > DRM_ERROR("PLL %d not locked\n", port); > > - if (IS_GEMINILAKE(dev_priv)) { > > + if (IS_GEMINILAKE(i915)) { > > temp = I915_READ(BXT_PORT_TX_DW5_LN0(phy, ch)); > > temp |= DCC_DELAY_RANGE_2; > > I915_WRITE(BXT_PORT_TX_DW5_GRP(phy, ch), temp); > > @@ -1566,7 +1566,7 @@ static void bxt_ddi_pll_enable(struct drm_i915_private *dev_priv, > > I915_WRITE(BXT_PORT_PCS_DW12_GRP(phy, ch), temp); > > } > > -static void bxt_ddi_pll_disable(struct drm_i915_private *dev_priv, > > +static void bxt_ddi_pll_disable(struct drm_i915_private *i915, > > struct intel_shared_dpll *pll) > > { > > enum port port = (enum port)pll->info->id; /* 1:1 port->PLL mapping */ > > @@ -1577,7 +1577,7 @@ static void bxt_ddi_pll_disable(struct drm_i915_private *dev_priv, > > I915_WRITE(BXT_PORT_PLL_ENABLE(port), temp); > > POSTING_READ(BXT_PORT_PLL_ENABLE(port)); > > - if (IS_GEMINILAKE(dev_priv)) { > > + if (IS_GEMINILAKE(i915)) { > > temp = I915_READ(BXT_PORT_PLL_ENABLE(port)); > > temp &= ~PORT_PLL_POWER_ENABLE; > > I915_WRITE(BXT_PORT_PLL_ENABLE(port), temp); > > @@ -1588,7 +1588,7 @@ static void bxt_ddi_pll_disable(struct drm_i915_private *dev_priv, > > } > > } > > -static bool bxt_ddi_pll_get_hw_state(struct drm_i915_private *dev_priv, > > +static bool bxt_ddi_pll_get_hw_state(struct drm_i915_private *i915, > > struct intel_shared_dpll *pll, > > struct intel_dpll_hw_state *hw_state) > > { > > @@ -1599,9 +1599,9 @@ static bool bxt_ddi_pll_get_hw_state(struct drm_i915_private *dev_priv, > > u32 val; > > bool ret; > > - bxt_port_to_phy_channel(dev_priv, port, &phy, &ch); > > + bxt_port_to_phy_channel(i915, port, &phy, &ch); > > - wakeref = intel_display_power_get_if_enabled(dev_priv, > > + wakeref = intel_display_power_get_if_enabled(i915, > > POWER_DOMAIN_DISPLAY_CORE); > > if (!wakeref) > > return false; > > @@ -1660,7 +1660,7 @@ static bool bxt_ddi_pll_get_hw_state(struct drm_i915_private *dev_priv, > > ret = true; > > out: > > - intel_display_power_put(dev_priv, POWER_DOMAIN_DISPLAY_CORE, wakeref); > > + intel_display_power_put(i915, POWER_DOMAIN_DISPLAY_CORE, wakeref); > > return ret; > > } > > @@ -1832,7 +1832,7 @@ bxt_get_dpll(struct intel_crtc_state *crtc_state, > > struct intel_encoder *encoder) > > { > > struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc); > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > struct intel_shared_dpll *pll; > > enum intel_dpll_id id; > > @@ -1846,7 +1846,7 @@ bxt_get_dpll(struct intel_crtc_state *crtc_state, > > /* 1:1 mapping between ports and PLLs */ > > id = (enum intel_dpll_id) encoder->port; > > - pll = intel_get_shared_dpll_by_id(dev_priv, id); > > + pll = intel_get_shared_dpll_by_id(i915, id); > > DRM_DEBUG_KMS("[CRTC:%d:%s] using pre-allocated %s\n", > > crtc->base.base.id, crtc->base.name, pll->info->name); > > @@ -1856,7 +1856,7 @@ bxt_get_dpll(struct intel_crtc_state *crtc_state, > > return pll; > > } > > -static void bxt_dump_hw_state(struct drm_i915_private *dev_priv, > > +static void bxt_dump_hw_state(struct drm_i915_private *i915, > > const struct intel_dpll_hw_state *hw_state) > > { > > DRM_DEBUG_KMS("dpll_hw_state: ebb0: 0x%x, ebb4: 0x%x," > > @@ -1887,7 +1887,7 @@ struct intel_dpll_mgr { > > struct intel_shared_dpll *(*get_dpll)(struct intel_crtc_state *crtc_state, > > struct intel_encoder *encoder); > > - void (*dump_hw_state)(struct drm_i915_private *dev_priv, > > + void (*dump_hw_state)(struct drm_i915_private *i915, > > const struct intel_dpll_hw_state *hw_state); > > }; > > @@ -1946,7 +1946,7 @@ static const struct intel_dpll_mgr bxt_pll_mgr = { > > .dump_hw_state = bxt_dump_hw_state, > > }; > > -static void cnl_ddi_pll_enable(struct drm_i915_private *dev_priv, > > +static void cnl_ddi_pll_enable(struct drm_i915_private *i915, > > struct intel_shared_dpll *pll) > > { > > const enum intel_dpll_id id = pll->info->id; > > @@ -1958,7 +1958,7 @@ static void cnl_ddi_pll_enable(struct drm_i915_private *dev_priv, > > I915_WRITE(CNL_DPLL_ENABLE(id), val); > > /* 2. Wait for DPLL power state enabled in DPLL_ENABLE. */ > > - if (intel_wait_for_register(&dev_priv->uncore, > > + if (intel_wait_for_register(&i915->uncore, > > CNL_DPLL_ENABLE(id), > > PLL_POWER_STATE, > > PLL_POWER_STATE, > > @@ -1999,7 +1999,7 @@ static void cnl_ddi_pll_enable(struct drm_i915_private *dev_priv, > > I915_WRITE(CNL_DPLL_ENABLE(id), val); > > /* 7. Wait for PLL lock status in DPLL_ENABLE. */ > > - if (intel_wait_for_register(&dev_priv->uncore, > > + if (intel_wait_for_register(&i915->uncore, > > CNL_DPLL_ENABLE(id), > > PLL_LOCK, > > PLL_LOCK, > > @@ -2021,7 +2021,7 @@ static void cnl_ddi_pll_enable(struct drm_i915_private *dev_priv, > > */ > > } > > -static void cnl_ddi_pll_disable(struct drm_i915_private *dev_priv, > > +static void cnl_ddi_pll_disable(struct drm_i915_private *i915, > > struct intel_shared_dpll *pll) > > { > > const enum intel_dpll_id id = pll->info->id; > > @@ -2047,7 +2047,7 @@ static void cnl_ddi_pll_disable(struct drm_i915_private *dev_priv, > > I915_WRITE(CNL_DPLL_ENABLE(id), val); > > /* 4. Wait for PLL not locked status in DPLL_ENABLE. */ > > - if (intel_wait_for_register(&dev_priv->uncore, > > + if (intel_wait_for_register(&i915->uncore, > > CNL_DPLL_ENABLE(id), > > PLL_LOCK, > > 0, > > @@ -2069,7 +2069,7 @@ static void cnl_ddi_pll_disable(struct drm_i915_private *dev_priv, > > I915_WRITE(CNL_DPLL_ENABLE(id), val); > > /* 7. Wait for DPLL power state disabled in DPLL_ENABLE. */ > > - if (intel_wait_for_register(&dev_priv->uncore, > > + if (intel_wait_for_register(&i915->uncore, > > CNL_DPLL_ENABLE(id), > > PLL_POWER_STATE, > > 0, > > @@ -2077,7 +2077,7 @@ static void cnl_ddi_pll_disable(struct drm_i915_private *dev_priv, > > DRM_ERROR("PLL %d Power not disabled\n", id); > > } > > -static bool cnl_ddi_pll_get_hw_state(struct drm_i915_private *dev_priv, > > +static bool cnl_ddi_pll_get_hw_state(struct drm_i915_private *i915, > > struct intel_shared_dpll *pll, > > struct intel_dpll_hw_state *hw_state) > > { > > @@ -2086,7 +2086,7 @@ static bool cnl_ddi_pll_get_hw_state(struct drm_i915_private *dev_priv, > > u32 val; > > bool ret; > > - wakeref = intel_display_power_get_if_enabled(dev_priv, > > + wakeref = intel_display_power_get_if_enabled(i915, > > POWER_DOMAIN_DISPLAY_CORE); > > if (!wakeref) > > return false; > > @@ -2107,7 +2107,7 @@ static bool cnl_ddi_pll_get_hw_state(struct drm_i915_private *dev_priv, > > ret = true; > > out: > > - intel_display_power_put(dev_priv, POWER_DOMAIN_DISPLAY_CORE, wakeref); > > + intel_display_power_put(i915, POWER_DOMAIN_DISPLAY_CORE, wakeref); > > return ret; > > } > > @@ -2199,15 +2199,15 @@ static void cnl_wrpll_params_populate(struct skl_wrpll_params *params, > > params->dco_fraction = dco & 0x7fff; > > } > > -int cnl_hdmi_pll_ref_clock(struct drm_i915_private *dev_priv) > > +int cnl_hdmi_pll_ref_clock(struct drm_i915_private *i915) > > { > > - int ref_clock = dev_priv->cdclk.hw.ref; > > + int ref_clock = i915->cdclk.hw.ref; > > /* > > * For ICL+, the spec states: if reference frequency is 38.4, > > * use 19.2 because the DPLL automatically divides that by 2. > > */ > > - if (INTEL_GEN(dev_priv) >= 11 && ref_clock == 38400) > > + if (INTEL_GEN(i915) >= 11 && ref_clock == 38400) > > ref_clock = 19200; > > return ref_clock; > > @@ -2217,7 +2217,7 @@ static bool > > cnl_ddi_calculate_wrpll(struct intel_crtc_state *crtc_state, > > struct skl_wrpll_params *wrpll_params) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc_state->base.crtc->dev); > > + struct drm_i915_private *i915 = to_i915(crtc_state->base.crtc->dev); > > u32 afe_clock = crtc_state->port_clock * 5; > > u32 ref_clock; > > u32 dco_min = 7998000; > > @@ -2252,7 +2252,7 @@ cnl_ddi_calculate_wrpll(struct intel_crtc_state *crtc_state, > > cnl_wrpll_get_multipliers(best_div, &pdiv, &qdiv, &kdiv); > > - ref_clock = cnl_hdmi_pll_ref_clock(dev_priv); > > + ref_clock = cnl_hdmi_pll_ref_clock(i915); > > cnl_wrpll_params_populate(wrpll_params, best_dco, ref_clock, > > pdiv, qdiv, kdiv); > > @@ -2370,7 +2370,7 @@ cnl_get_dpll(struct intel_crtc_state *crtc_state, > > return pll; > > } > > -static void cnl_dump_hw_state(struct drm_i915_private *dev_priv, > > +static void cnl_dump_hw_state(struct drm_i915_private *i915, > > const struct intel_dpll_hw_state *hw_state) > > { > > DRM_DEBUG_KMS("dpll_hw_state: " > > @@ -2476,9 +2476,9 @@ static const struct skl_wrpll_params icl_tbt_pll_19_2MHz_values = { > > static bool icl_calc_dp_combo_pll(struct intel_crtc_state *crtc_state, > > struct skl_wrpll_params *pll_params) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc_state->base.crtc->dev); > > + struct drm_i915_private *i915 = to_i915(crtc_state->base.crtc->dev); > > const struct icl_combo_pll_params *params = > > - dev_priv->cdclk.hw.ref == 24000 ? > > + i915->cdclk.hw.ref == 24000 ? > > icl_dp_combo_pll_24MHz_values : > > icl_dp_combo_pll_19_2MHz_values; > > int clock = crtc_state->port_clock; > > @@ -2498,9 +2498,9 @@ static bool icl_calc_dp_combo_pll(struct intel_crtc_state *crtc_state, > > static bool icl_calc_tbt_pll(struct intel_crtc_state *crtc_state, > > struct skl_wrpll_params *pll_params) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc_state->base.crtc->dev); > > + struct drm_i915_private *i915 = to_i915(crtc_state->base.crtc->dev); > > - *pll_params = dev_priv->cdclk.hw.ref == 24000 ? > > + *pll_params = i915->cdclk.hw.ref == 24000 ? > > icl_tbt_pll_24MHz_values : icl_tbt_pll_19_2MHz_values; > > return true; > > } > > @@ -2508,12 +2508,12 @@ static bool icl_calc_tbt_pll(struct intel_crtc_state *crtc_state, > > static bool icl_calc_dpll_state(struct intel_crtc_state *crtc_state, > > struct intel_encoder *encoder) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc_state->base.crtc->dev); > > + struct drm_i915_private *i915 = to_i915(crtc_state->base.crtc->dev); > > u32 cfgcr0, cfgcr1; > > struct skl_wrpll_params pll_params = { 0 }; > > bool ret; > > - if (intel_port_is_tc(dev_priv, encoder->port)) > > + if (intel_port_is_tc(i915, encoder->port)) > > ret = icl_calc_tbt_pll(crtc_state, &pll_params); > > else if (intel_crtc_has_type(crtc_state, INTEL_OUTPUT_HDMI) || > > intel_crtc_has_type(crtc_state, INTEL_OUTPUT_DSI)) > > @@ -2629,9 +2629,9 @@ static bool icl_mg_pll_find_divisors(int clock_khz, bool is_dp, bool use_ssc, > > */ > > static bool icl_calc_mg_pll_state(struct intel_crtc_state *crtc_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc_state->base.crtc->dev); > > + struct drm_i915_private *i915 = to_i915(crtc_state->base.crtc->dev); > > struct intel_dpll_hw_state *pll_state = &crtc_state->dpll_hw_state; > > - int refclk_khz = dev_priv->cdclk.hw.ref; > > + int refclk_khz = i915->cdclk.hw.ref; > > int clock = crtc_state->port_clock; > > u32 dco_khz, m1div, m2div_int, m2div_rem, m2div_frac; > > u32 iref_ndiv, iref_trim, iref_pulse_w; > > @@ -2796,18 +2796,18 @@ static struct intel_shared_dpll * > > icl_get_dpll(struct intel_crtc_state *crtc_state, > > struct intel_encoder *encoder) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc_state->base.crtc->dev); > > + struct drm_i915_private *i915 = to_i915(crtc_state->base.crtc->dev); > > struct intel_digital_port *intel_dig_port; > > struct intel_shared_dpll *pll; > > enum port port = encoder->port; > > enum intel_dpll_id min, max; > > bool ret; > > - if (intel_port_is_combophy(dev_priv, port)) { > > + if (intel_port_is_combophy(i915, port)) { > > min = DPLL_ID_ICL_DPLL0; > > max = DPLL_ID_ICL_DPLL1; > > ret = icl_calc_dpll_state(crtc_state, encoder); > > - } else if (intel_port_is_tc(dev_priv, port)) { > > + } else if (intel_port_is_tc(i915, port)) { > > if (encoder->type == INTEL_OUTPUT_DP_MST) { > > struct intel_dp_mst_encoder *mst_encoder; > > @@ -2824,7 +2824,7 @@ icl_get_dpll(struct intel_crtc_state *crtc_state, > > } else { > > enum tc_port tc_port; > > - tc_port = intel_port_to_tc(dev_priv, port); > > + tc_port = intel_port_to_tc(i915, port); > > min = icl_tc_port_to_pll_id(tc_port); > > max = min; > > ret = icl_calc_mg_pll_state(crtc_state); > > @@ -2851,7 +2851,7 @@ icl_get_dpll(struct intel_crtc_state *crtc_state, > > return pll; > > } > > -static bool mg_pll_get_hw_state(struct drm_i915_private *dev_priv, > > +static bool mg_pll_get_hw_state(struct drm_i915_private *i915, > > struct intel_shared_dpll *pll, > > struct intel_dpll_hw_state *hw_state) > > { > > @@ -2861,7 +2861,7 @@ static bool mg_pll_get_hw_state(struct drm_i915_private *dev_priv, > > bool ret = false; > > u32 val; > > - wakeref = intel_display_power_get_if_enabled(dev_priv, > > + wakeref = intel_display_power_get_if_enabled(i915, > > POWER_DOMAIN_DISPLAY_CORE); > > if (!wakeref) > > return false; > > @@ -2896,7 +2896,7 @@ static bool mg_pll_get_hw_state(struct drm_i915_private *dev_priv, > > hw_state->mg_pll_tdc_coldst_bias = > > I915_READ(MG_PLL_TDC_COLDST_BIAS(tc_port)); > > - if (dev_priv->cdclk.hw.ref == 38400) { > > + if (i915->cdclk.hw.ref == 38400) { > > hw_state->mg_pll_tdc_coldst_bias_mask = MG_PLL_TDC_COLDST_COLDSTART; > > hw_state->mg_pll_bias_mask = 0; > > } else { > > @@ -2909,11 +2909,11 @@ static bool mg_pll_get_hw_state(struct drm_i915_private *dev_priv, > > ret = true; > > out: > > - intel_display_power_put(dev_priv, POWER_DOMAIN_DISPLAY_CORE, wakeref); > > + intel_display_power_put(i915, POWER_DOMAIN_DISPLAY_CORE, wakeref); > > return ret; > > } > > -static bool icl_pll_get_hw_state(struct drm_i915_private *dev_priv, > > +static bool icl_pll_get_hw_state(struct drm_i915_private *i915, > > struct intel_shared_dpll *pll, > > struct intel_dpll_hw_state *hw_state, > > i915_reg_t enable_reg) > > @@ -2923,7 +2923,7 @@ static bool icl_pll_get_hw_state(struct drm_i915_private *dev_priv, > > bool ret = false; > > u32 val; > > - wakeref = intel_display_power_get_if_enabled(dev_priv, > > + wakeref = intel_display_power_get_if_enabled(i915, > > POWER_DOMAIN_DISPLAY_CORE); > > if (!wakeref) > > return false; > > @@ -2937,26 +2937,26 @@ static bool icl_pll_get_hw_state(struct drm_i915_private *dev_priv, > > ret = true; > > out: > > - intel_display_power_put(dev_priv, POWER_DOMAIN_DISPLAY_CORE, wakeref); > > + intel_display_power_put(i915, POWER_DOMAIN_DISPLAY_CORE, wakeref); > > return ret; > > } > > -static bool combo_pll_get_hw_state(struct drm_i915_private *dev_priv, > > +static bool combo_pll_get_hw_state(struct drm_i915_private *i915, > > struct intel_shared_dpll *pll, > > struct intel_dpll_hw_state *hw_state) > > { > > - return icl_pll_get_hw_state(dev_priv, pll, hw_state, > > + return icl_pll_get_hw_state(i915, pll, hw_state, > > CNL_DPLL_ENABLE(pll->info->id)); > > } > > -static bool tbt_pll_get_hw_state(struct drm_i915_private *dev_priv, > > +static bool tbt_pll_get_hw_state(struct drm_i915_private *i915, > > struct intel_shared_dpll *pll, > > struct intel_dpll_hw_state *hw_state) > > { > > - return icl_pll_get_hw_state(dev_priv, pll, hw_state, TBT_PLL_ENABLE); > > + return icl_pll_get_hw_state(i915, pll, hw_state, TBT_PLL_ENABLE); > > } > > -static void icl_dpll_write(struct drm_i915_private *dev_priv, > > +static void icl_dpll_write(struct drm_i915_private *i915, > > struct intel_shared_dpll *pll) > > { > > struct intel_dpll_hw_state *hw_state = &pll->state.hw_state; > > @@ -2967,7 +2967,7 @@ static void icl_dpll_write(struct drm_i915_private *dev_priv, > > POSTING_READ(ICL_DPLL_CFGCR1(id)); > > } > > -static void icl_mg_pll_write(struct drm_i915_private *dev_priv, > > +static void icl_mg_pll_write(struct drm_i915_private *i915, > > struct intel_shared_dpll *pll) > > { > > struct intel_dpll_hw_state *hw_state = &pll->state.hw_state; > > @@ -3017,7 +3017,7 @@ static void icl_mg_pll_write(struct drm_i915_private *dev_priv, > > POSTING_READ(MG_PLL_TDC_COLDST_BIAS(tc_port)); > > } > > -static void icl_pll_power_enable(struct drm_i915_private *dev_priv, > > +static void icl_pll_power_enable(struct drm_i915_private *i915, > > struct intel_shared_dpll *pll, > > i915_reg_t enable_reg) > > { > > @@ -3031,12 +3031,12 @@ static void icl_pll_power_enable(struct drm_i915_private *dev_priv, > > * The spec says we need to "wait" but it also says it should be > > * immediate. > > */ > > - if (intel_wait_for_register(&dev_priv->uncore, enable_reg, > > + if (intel_wait_for_register(&i915->uncore, enable_reg, > > PLL_POWER_STATE, PLL_POWER_STATE, 1)) > > DRM_ERROR("PLL %d Power not enabled\n", pll->info->id); > > } > > -static void icl_pll_enable(struct drm_i915_private *dev_priv, > > +static void icl_pll_enable(struct drm_i915_private *i915, > > struct intel_shared_dpll *pll, > > i915_reg_t enable_reg) > > { > > @@ -3047,19 +3047,19 @@ static void icl_pll_enable(struct drm_i915_private *dev_priv, > > I915_WRITE(enable_reg, val); > > /* Timeout is actually 600us. */ > > - if (intel_wait_for_register(&dev_priv->uncore, enable_reg, > > + if (intel_wait_for_register(&i915->uncore, enable_reg, > > PLL_LOCK, PLL_LOCK, 1)) > > DRM_ERROR("PLL %d not locked\n", pll->info->id); > > } > > -static void combo_pll_enable(struct drm_i915_private *dev_priv, > > +static void combo_pll_enable(struct drm_i915_private *i915, > > struct intel_shared_dpll *pll) > > { > > i915_reg_t enable_reg = CNL_DPLL_ENABLE(pll->info->id); > > - icl_pll_power_enable(dev_priv, pll, enable_reg); > > + icl_pll_power_enable(i915, pll, enable_reg); > > - icl_dpll_write(dev_priv, pll); > > + icl_dpll_write(i915, pll); > > /* > > * DVFS pre sequence would be here, but in our driver the cdclk code > > @@ -3067,17 +3067,17 @@ static void combo_pll_enable(struct drm_i915_private *dev_priv, > > * nothing here. > > */ > > - icl_pll_enable(dev_priv, pll, enable_reg); > > + icl_pll_enable(i915, pll, enable_reg); > > /* DVFS post sequence would be here. See the comment above. */ > > } > > -static void tbt_pll_enable(struct drm_i915_private *dev_priv, > > +static void tbt_pll_enable(struct drm_i915_private *i915, > > struct intel_shared_dpll *pll) > > { > > - icl_pll_power_enable(dev_priv, pll, TBT_PLL_ENABLE); > > + icl_pll_power_enable(i915, pll, TBT_PLL_ENABLE); > > - icl_dpll_write(dev_priv, pll); > > + icl_dpll_write(i915, pll); > > /* > > * DVFS pre sequence would be here, but in our driver the cdclk code > > @@ -3085,20 +3085,20 @@ static void tbt_pll_enable(struct drm_i915_private *dev_priv, > > * nothing here. > > */ > > - icl_pll_enable(dev_priv, pll, TBT_PLL_ENABLE); > > + icl_pll_enable(i915, pll, TBT_PLL_ENABLE); > > /* DVFS post sequence would be here. See the comment above. */ > > } > > -static void mg_pll_enable(struct drm_i915_private *dev_priv, > > +static void mg_pll_enable(struct drm_i915_private *i915, > > struct intel_shared_dpll *pll) > > { > > i915_reg_t enable_reg = > > MG_PLL_ENABLE(icl_pll_id_to_tc_port(pll->info->id)); > > - icl_pll_power_enable(dev_priv, pll, enable_reg); > > + icl_pll_power_enable(i915, pll, enable_reg); > > - icl_mg_pll_write(dev_priv, pll); > > + icl_mg_pll_write(i915, pll); > > /* > > * DVFS pre sequence would be here, but in our driver the cdclk code > > @@ -3106,12 +3106,12 @@ static void mg_pll_enable(struct drm_i915_private *dev_priv, > > * nothing here. > > */ > > - icl_pll_enable(dev_priv, pll, enable_reg); > > + icl_pll_enable(i915, pll, enable_reg); > > /* DVFS post sequence would be here. See the comment above. */ > > } > > -static void icl_pll_disable(struct drm_i915_private *dev_priv, > > +static void icl_pll_disable(struct drm_i915_private *i915, > > struct intel_shared_dpll *pll, > > i915_reg_t enable_reg) > > { > > @@ -3130,7 +3130,7 @@ static void icl_pll_disable(struct drm_i915_private *dev_priv, > > I915_WRITE(enable_reg, val); > > /* Timeout is actually 1us. */ > > - if (intel_wait_for_register(&dev_priv->uncore, > > + if (intel_wait_for_register(&i915->uncore, > > enable_reg, PLL_LOCK, 0, 1)) > > DRM_ERROR("PLL %d locked\n", pll->info->id); > > @@ -3144,33 +3144,33 @@ static void icl_pll_disable(struct drm_i915_private *dev_priv, > > * The spec says we need to "wait" but it also says it should be > > * immediate. > > */ > > - if (intel_wait_for_register(&dev_priv->uncore, > > + if (intel_wait_for_register(&i915->uncore, > > enable_reg, PLL_POWER_STATE, 0, 1)) > > DRM_ERROR("PLL %d Power not disabled\n", pll->info->id); > > } > > -static void combo_pll_disable(struct drm_i915_private *dev_priv, > > +static void combo_pll_disable(struct drm_i915_private *i915, > > struct intel_shared_dpll *pll) > > { > > - icl_pll_disable(dev_priv, pll, CNL_DPLL_ENABLE(pll->info->id)); > > + icl_pll_disable(i915, pll, CNL_DPLL_ENABLE(pll->info->id)); > > } > > -static void tbt_pll_disable(struct drm_i915_private *dev_priv, > > +static void tbt_pll_disable(struct drm_i915_private *i915, > > struct intel_shared_dpll *pll) > > { > > - icl_pll_disable(dev_priv, pll, TBT_PLL_ENABLE); > > + icl_pll_disable(i915, pll, TBT_PLL_ENABLE); > > } > > -static void mg_pll_disable(struct drm_i915_private *dev_priv, > > +static void mg_pll_disable(struct drm_i915_private *i915, > > struct intel_shared_dpll *pll) > > { > > i915_reg_t enable_reg = > > MG_PLL_ENABLE(icl_pll_id_to_tc_port(pll->info->id)); > > - icl_pll_disable(dev_priv, pll, enable_reg); > > + icl_pll_disable(i915, pll, enable_reg); > > } > > -static void icl_dump_hw_state(struct drm_i915_private *dev_priv, > > +static void icl_dump_hw_state(struct drm_i915_private *i915, > > const struct intel_dpll_hw_state *hw_state) > > { > > DRM_DEBUG_KMS("dpll_hw_state: cfgcr0: 0x%x, cfgcr1: 0x%x, " > > @@ -3247,28 +3247,28 @@ static const struct intel_dpll_mgr ehl_pll_mgr = { > > */ > > void intel_shared_dpll_init(struct drm_device *dev) > > { > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > const struct intel_dpll_mgr *dpll_mgr = NULL; > > const struct dpll_info *dpll_info; > > int i; > > - if (IS_ELKHARTLAKE(dev_priv)) > > + if (IS_ELKHARTLAKE(i915)) > > dpll_mgr = &ehl_pll_mgr; > > - else if (INTEL_GEN(dev_priv) >= 11) > > + else if (INTEL_GEN(i915) >= 11) > > dpll_mgr = &icl_pll_mgr; > > - else if (IS_CANNONLAKE(dev_priv)) > > + else if (IS_CANNONLAKE(i915)) > > dpll_mgr = &cnl_pll_mgr; > > - else if (IS_GEN9_BC(dev_priv)) > > + else if (IS_GEN9_BC(i915)) > > dpll_mgr = &skl_pll_mgr; > > - else if (IS_GEN9_LP(dev_priv)) > > + else if (IS_GEN9_LP(i915)) > > dpll_mgr = &bxt_pll_mgr; > > - else if (HAS_DDI(dev_priv)) > > + else if (HAS_DDI(i915)) > > dpll_mgr = &hsw_pll_mgr; > > - else if (HAS_PCH_IBX(dev_priv) || HAS_PCH_CPT(dev_priv)) > > + else if (HAS_PCH_IBX(i915) || HAS_PCH_CPT(i915)) > > dpll_mgr = &pch_pll_mgr; > > if (!dpll_mgr) { > > - dev_priv->num_shared_dpll = 0; > > + i915->num_shared_dpll = 0; > > return; > > } > > @@ -3276,14 +3276,14 @@ void intel_shared_dpll_init(struct drm_device *dev) > > for (i = 0; dpll_info[i].name; i++) { > > WARN_ON(i != dpll_info[i].id); > > - dev_priv->shared_dplls[i].info = &dpll_info[i]; > > + i915->shared_dplls[i].info = &dpll_info[i]; > > } > > - dev_priv->dpll_mgr = dpll_mgr; > > - dev_priv->num_shared_dpll = i; > > - mutex_init(&dev_priv->dpll_lock); > > + i915->dpll_mgr = dpll_mgr; > > + i915->num_shared_dpll = i; > > + mutex_init(&i915->dpll_lock); > > - BUG_ON(dev_priv->num_shared_dpll > I915_NUM_PLLS); > > + BUG_ON(i915->num_shared_dpll > I915_NUM_PLLS); > > } > > /** > > @@ -3304,8 +3304,8 @@ struct intel_shared_dpll * > > intel_get_shared_dpll(struct intel_crtc_state *crtc_state, > > struct intel_encoder *encoder) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc_state->base.crtc->dev); > > - const struct intel_dpll_mgr *dpll_mgr = dev_priv->dpll_mgr; > > + struct drm_i915_private *i915 = to_i915(crtc_state->base.crtc->dev); > > + const struct intel_dpll_mgr *dpll_mgr = i915->dpll_mgr; > > if (WARN_ON(!dpll_mgr)) > > return NULL; > > @@ -3335,16 +3335,16 @@ void intel_release_shared_dpll(struct intel_shared_dpll *dpll, > > /** > > * intel_shared_dpll_dump_hw_state - write hw_state to dmesg > > - * @dev_priv: i915 drm device > > + * @i915: i915 drm device > > * @hw_state: hw state to be written to the log > > * > > * Write the relevant values in @hw_state to dmesg using DRM_DEBUG_KMS. > > */ > > -void intel_dpll_dump_hw_state(struct drm_i915_private *dev_priv, > > +void intel_dpll_dump_hw_state(struct drm_i915_private *i915, > > const struct intel_dpll_hw_state *hw_state) > > { > > - if (dev_priv->dpll_mgr) { > > - dev_priv->dpll_mgr->dump_hw_state(dev_priv, hw_state); > > + if (i915->dpll_mgr) { > > + i915->dpll_mgr->dump_hw_state(i915, hw_state); > > } else { > > /* fallback for platforms that don't use the shared dpll > > * infrastructure > > diff --git a/drivers/gpu/drm/i915/intel_dpll_mgr.h b/drivers/gpu/drm/i915/intel_dpll_mgr.h > > index d0570414f3d1..b068c8441328 100644 > > --- a/drivers/gpu/drm/i915/intel_dpll_mgr.h > > +++ b/drivers/gpu/drm/i915/intel_dpll_mgr.h > > @@ -221,7 +221,7 @@ struct intel_shared_dpll_funcs { > > * Called from intel_prepare_shared_dpll() function unless the PLL > > * is already enabled. > > */ > > - void (*prepare)(struct drm_i915_private *dev_priv, > > + void (*prepare)(struct drm_i915_private *i915, > > struct intel_shared_dpll *pll); > > /** > > @@ -230,7 +230,7 @@ struct intel_shared_dpll_funcs { > > * Hook for enabling the pll, called from intel_enable_shared_dpll() > > * if the pll is not already enabled. > > */ > > - void (*enable)(struct drm_i915_private *dev_priv, > > + void (*enable)(struct drm_i915_private *i915, > > struct intel_shared_dpll *pll); > > /** > > @@ -240,7 +240,7 @@ struct intel_shared_dpll_funcs { > > * only when it is safe to disable the pll, i.e., there are no more > > * tracked users for it. > > */ > > - void (*disable)(struct drm_i915_private *dev_priv, > > + void (*disable)(struct drm_i915_private *i915, > > struct intel_shared_dpll *pll); > > /** > > @@ -250,7 +250,7 @@ struct intel_shared_dpll_funcs { > > * registers. This is used for initial hw state readout and state > > * verification after a mode set. > > */ > > - bool (*get_hw_state)(struct drm_i915_private *dev_priv, > > + bool (*get_hw_state)(struct drm_i915_private *i915, > > struct intel_shared_dpll *pll, > > struct intel_dpll_hw_state *hw_state); > > }; > > @@ -271,7 +271,7 @@ struct dpll_info { > > /** > > * @id: unique indentifier for this DPLL; should match the index in the > > - * dev_priv->shared_dplls array > > + * i915->shared_dplls array > > */ > > enum intel_dpll_id id; > > @@ -321,12 +321,12 @@ struct intel_shared_dpll { > > /* shared dpll functions */ > > struct intel_shared_dpll * > > -intel_get_shared_dpll_by_id(struct drm_i915_private *dev_priv, > > +intel_get_shared_dpll_by_id(struct drm_i915_private *i915, > > enum intel_dpll_id id); > > enum intel_dpll_id > > -intel_get_shared_dpll_id(struct drm_i915_private *dev_priv, > > +intel_get_shared_dpll_id(struct drm_i915_private *i915, > > struct intel_shared_dpll *pll); > > -void assert_shared_dpll(struct drm_i915_private *dev_priv, > > +void assert_shared_dpll(struct drm_i915_private *i915, > > struct intel_shared_dpll *pll, > > bool state); > > #define assert_shared_dpll_enabled(d, p) assert_shared_dpll(d, p, true) > > @@ -342,9 +342,9 @@ void intel_disable_shared_dpll(const struct intel_crtc_state *crtc_state); > > void intel_shared_dpll_swap_state(struct drm_atomic_state *state); > > void intel_shared_dpll_init(struct drm_device *dev); > > -void intel_dpll_dump_hw_state(struct drm_i915_private *dev_priv, > > +void intel_dpll_dump_hw_state(struct drm_i915_private *i915, > > const struct intel_dpll_hw_state *hw_state); > > -int cnl_hdmi_pll_ref_clock(struct drm_i915_private *dev_priv); > > +int cnl_hdmi_pll_ref_clock(struct drm_i915_private *i915); > > enum intel_dpll_id icl_tc_port_to_pll_id(enum tc_port tc_port); > > bool intel_dpll_is_combophy(enum intel_dpll_id id); > > diff --git a/drivers/gpu/drm/i915/intel_drv.h b/drivers/gpu/drm/i915/intel_drv.h > > index 3e337317f77e..23065dda82d9 100644 > > --- a/drivers/gpu/drm/i915/intel_drv.h > > +++ b/drivers/gpu/drm/i915/intel_drv.h > > @@ -963,7 +963,7 @@ struct intel_crtc { > > struct intel_crtc_state *config; > > - /* Access to these should be protected by dev_priv->irq_lock. */ > > + /* Access to these should be protected by i915->irq_lock. */ > > bool cpu_fifo_underrun_disabled; > > bool pch_fifo_underrun_disabled; > > @@ -1293,15 +1293,15 @@ vlv_pipe_to_channel(enum pipe pipe) > > } > > static inline struct intel_crtc * > > -intel_get_crtc_for_pipe(struct drm_i915_private *dev_priv, enum pipe pipe) > > +intel_get_crtc_for_pipe(struct drm_i915_private *i915, enum pipe pipe) > > { > > - return dev_priv->pipe_to_crtc_mapping[pipe]; > > + return i915->pipe_to_crtc_mapping[pipe]; > > } > > static inline struct intel_crtc * > > -intel_get_crtc_for_plane(struct drm_i915_private *dev_priv, enum i9xx_plane_id plane) > > +intel_get_crtc_for_plane(struct drm_i915_private *i915, enum i9xx_plane_id plane) > > { > > - return dev_priv->plane_to_crtc_mapping[plane]; > > + return i915->plane_to_crtc_mapping[plane]; > > } > > struct intel_load_detect_pipe { > > @@ -1447,17 +1447,17 @@ intel_atomic_get_new_crtc_state(struct intel_atomic_state *state, > > /* intel_display.c */ > > void intel_plane_destroy(struct drm_plane *plane); > > -void i830_enable_pipe(struct drm_i915_private *dev_priv, enum pipe pipe); > > -void i830_disable_pipe(struct drm_i915_private *dev_priv, enum pipe pipe); > > +void i830_enable_pipe(struct drm_i915_private *i915, enum pipe pipe); > > +void i830_disable_pipe(struct drm_i915_private *i915, enum pipe pipe); > > enum pipe intel_crtc_pch_transcoder(struct intel_crtc *crtc); > > -int vlv_get_hpll_vco(struct drm_i915_private *dev_priv); > > -int vlv_get_cck_clock(struct drm_i915_private *dev_priv, > > +int vlv_get_hpll_vco(struct drm_i915_private *i915); > > +int vlv_get_cck_clock(struct drm_i915_private *i915, > > const char *name, u32 reg, int ref_freq); > > -int vlv_get_cck_clock_hpll(struct drm_i915_private *dev_priv, > > +int vlv_get_cck_clock_hpll(struct drm_i915_private *i915, > > const char *name, u32 reg); > > -void lpt_disable_pch_transcoder(struct drm_i915_private *dev_priv); > > -void lpt_disable_iclkip(struct drm_i915_private *dev_priv); > > -void intel_init_display_hooks(struct drm_i915_private *dev_priv); > > +void lpt_disable_pch_transcoder(struct drm_i915_private *i915); > > +void lpt_disable_iclkip(struct drm_i915_private *i915); > > +void intel_init_display_hooks(struct drm_i915_private *i915); > > unsigned int intel_fb_xy_to_linear(int x, int y, > > const struct intel_plane_state *state, > > int plane); > > @@ -1467,19 +1467,19 @@ void intel_add_fb_offsets(int *x, int *y, > > const struct intel_plane_state *state, int plane); > > unsigned int intel_rotation_info_size(const struct intel_rotation_info *rot_info); > > unsigned int intel_remapped_info_size(const struct intel_remapped_info *rem_info); > > -bool intel_has_pending_fb_unpin(struct drm_i915_private *dev_priv); > > +bool intel_has_pending_fb_unpin(struct drm_i915_private *i915); > > int intel_display_suspend(struct drm_device *dev); > > -void intel_pps_unlock_regs_wa(struct drm_i915_private *dev_priv); > > +void intel_pps_unlock_regs_wa(struct drm_i915_private *i915); > > void intel_encoder_destroy(struct drm_encoder *encoder); > > struct drm_display_mode * > > intel_encoder_current_mode(struct intel_encoder *encoder); > > -bool intel_port_is_combophy(struct drm_i915_private *dev_priv, enum port port); > > -bool intel_port_is_tc(struct drm_i915_private *dev_priv, enum port port); > > -enum tc_port intel_port_to_tc(struct drm_i915_private *dev_priv, > > +bool intel_port_is_combophy(struct drm_i915_private *i915, enum port port); > > +bool intel_port_is_tc(struct drm_i915_private *i915, enum port port); > > +enum tc_port intel_port_to_tc(struct drm_i915_private *i915, > > enum port port); > > int intel_get_pipe_from_crtc_id_ioctl(struct drm_device *dev, void *data, > > struct drm_file *file_priv); > > -enum transcoder intel_pipe_to_cpu_transcoder(struct drm_i915_private *dev_priv, > > +enum transcoder intel_pipe_to_cpu_transcoder(struct drm_i915_private *i915, > > enum pipe pipe); > > static inline bool > > intel_crtc_has_type(const struct intel_crtc_state *crtc_state, > > @@ -1496,23 +1496,23 @@ intel_crtc_has_dp_encoder(const struct intel_crtc_state *crtc_state) > > (1 << INTEL_OUTPUT_EDP)); > > } > > static inline void > > -intel_wait_for_vblank(struct drm_i915_private *dev_priv, enum pipe pipe) > > +intel_wait_for_vblank(struct drm_i915_private *i915, enum pipe pipe) > > { > > - drm_wait_one_vblank(&dev_priv->drm, pipe); > > + drm_wait_one_vblank(&i915->drm, pipe); > > } > > static inline void > > -intel_wait_for_vblank_if_active(struct drm_i915_private *dev_priv, int pipe) > > +intel_wait_for_vblank_if_active(struct drm_i915_private *i915, int pipe) > > { > > - const struct intel_crtc *crtc = intel_get_crtc_for_pipe(dev_priv, pipe); > > + const struct intel_crtc *crtc = intel_get_crtc_for_pipe(i915, pipe); > > if (crtc->active) > > - intel_wait_for_vblank(dev_priv, pipe); > > + intel_wait_for_vblank(i915, pipe); > > } > > u32 intel_crtc_get_vblank_counter(struct intel_crtc *crtc); > > int ironlake_get_lanes_required(int target_clock, int link_bw, int bpp); > > -void vlv_wait_port_ready(struct drm_i915_private *dev_priv, > > +void vlv_wait_port_ready(struct drm_i915_private *i915, > > struct intel_digital_port *dport, > > unsigned int expected_mask); > > int intel_get_load_detect_pipe(struct drm_connector *connector, > > @@ -1536,34 +1536,34 @@ int intel_prepare_plane_fb(struct drm_plane *plane, > > void intel_cleanup_plane_fb(struct drm_plane *plane, > > struct drm_plane_state *old_state); > > -void assert_pch_transcoder_disabled(struct drm_i915_private *dev_priv, > > +void assert_pch_transcoder_disabled(struct drm_i915_private *i915, > > enum pipe pipe); > > -int vlv_force_pll_on(struct drm_i915_private *dev_priv, enum pipe pipe, > > +int vlv_force_pll_on(struct drm_i915_private *i915, enum pipe pipe, > > const struct dpll *dpll); > > -void vlv_force_pll_off(struct drm_i915_private *dev_priv, enum pipe pipe); > > -int lpt_get_iclkip(struct drm_i915_private *dev_priv); > > +void vlv_force_pll_off(struct drm_i915_private *i915, enum pipe pipe); > > +int lpt_get_iclkip(struct drm_i915_private *i915); > > bool intel_fuzzy_clock_check(int clock1, int clock2); > > /* modesetting asserts */ > > -void assert_panel_unlocked(struct drm_i915_private *dev_priv, > > +void assert_panel_unlocked(struct drm_i915_private *i915, > > enum pipe pipe); > > -void assert_pll(struct drm_i915_private *dev_priv, > > +void assert_pll(struct drm_i915_private *i915, > > enum pipe pipe, bool state); > > #define assert_pll_enabled(d, p) assert_pll(d, p, true) > > #define assert_pll_disabled(d, p) assert_pll(d, p, false) > > -void assert_dsi_pll(struct drm_i915_private *dev_priv, bool state); > > +void assert_dsi_pll(struct drm_i915_private *i915, bool state); > > #define assert_dsi_pll_enabled(d) assert_dsi_pll(d, true) > > #define assert_dsi_pll_disabled(d) assert_dsi_pll(d, false) > > -void assert_fdi_rx_pll(struct drm_i915_private *dev_priv, > > +void assert_fdi_rx_pll(struct drm_i915_private *i915, > > enum pipe pipe, bool state); > > #define assert_fdi_rx_pll_enabled(d, p) assert_fdi_rx_pll(d, p, true) > > #define assert_fdi_rx_pll_disabled(d, p) assert_fdi_rx_pll(d, p, false) > > -void assert_pipe(struct drm_i915_private *dev_priv, enum pipe pipe, bool state); > > +void assert_pipe(struct drm_i915_private *i915, enum pipe pipe, bool state); > > #define assert_pipe_enabled(d, p) assert_pipe(d, p, true) > > #define assert_pipe_disabled(d, p) assert_pipe(d, p, false) > > -void intel_prepare_reset(struct drm_i915_private *dev_priv); > > -void intel_finish_reset(struct drm_i915_private *dev_priv); > > +void intel_prepare_reset(struct drm_i915_private *i915); > > +void intel_finish_reset(struct drm_i915_private *i915); > > void intel_dp_get_m_n(struct intel_crtc *crtc, > > struct intel_crtc_state *pipe_config); > > void intel_dp_set_m_n(const struct intel_crtc_state *crtc_state, > > diff --git a/drivers/gpu/drm/i915/intel_dsi.c b/drivers/gpu/drm/i915/intel_dsi.c > > index 5fec02aceaed..09414da3d818 100644 > > --- a/drivers/gpu/drm/i915/intel_dsi.c > > +++ b/drivers/gpu/drm/i915/intel_dsi.c > > @@ -113,14 +113,14 @@ struct intel_dsi_host *intel_dsi_host_init(struct intel_dsi *intel_dsi, > > enum drm_panel_orientation > > intel_dsi_get_panel_orientation(struct intel_connector *connector) > > { > > - struct drm_i915_private *dev_priv = to_i915(connector->base.dev); > > + struct drm_i915_private *i915 = to_i915(connector->base.dev); > > enum drm_panel_orientation orientation; > > - orientation = dev_priv->vbt.dsi.orientation; > > + orientation = i915->vbt.dsi.orientation; > > if (orientation != DRM_MODE_PANEL_ORIENTATION_UNKNOWN) > > return orientation; > > - orientation = dev_priv->vbt.orientation; > > + orientation = i915->vbt.orientation; > > if (orientation != DRM_MODE_PANEL_ORIENTATION_UNKNOWN) > > return orientation; > > diff --git a/drivers/gpu/drm/i915/intel_dsi.h b/drivers/gpu/drm/i915/intel_dsi.h > > index 6d20434636cd..2f477263d1f6 100644 > > --- a/drivers/gpu/drm/i915/intel_dsi.h > > +++ b/drivers/gpu/drm/i915/intel_dsi.h > > @@ -155,7 +155,7 @@ static inline u16 intel_dsi_encoder_ports(struct intel_encoder *encoder) > > } > > /* icl_dsi.c */ > > -void icl_dsi_init(struct drm_i915_private *dev_priv); > > +void icl_dsi_init(struct drm_i915_private *i915); > > /* intel_dsi.c */ > > int intel_dsi_bitrate(const struct intel_dsi *intel_dsi); > > @@ -172,7 +172,7 @@ enum drm_mode_status intel_dsi_mode_valid(struct drm_connector *connector, > > struct intel_dsi_host *intel_dsi_host_init(struct intel_dsi *intel_dsi, > > const struct mipi_dsi_host_ops *funcs, > > enum port port); > > -void vlv_dsi_init(struct drm_i915_private *dev_priv); > > +void vlv_dsi_init(struct drm_i915_private *i915); > > /* vlv_dsi_pll.c */ > > int vlv_dsi_pll_compute(struct intel_encoder *encoder, > > @@ -184,7 +184,7 @@ u32 vlv_dsi_get_pclk(struct intel_encoder *encoder, > > struct intel_crtc_state *config); > > void vlv_dsi_reset_clocks(struct intel_encoder *encoder, enum port port); > > -bool bxt_dsi_pll_is_enabled(struct drm_i915_private *dev_priv); > > +bool bxt_dsi_pll_is_enabled(struct drm_i915_private *i915); > > int bxt_dsi_pll_compute(struct intel_encoder *encoder, > > struct intel_crtc_state *config); > > void bxt_dsi_pll_enable(struct intel_encoder *encoder, > > diff --git a/drivers/gpu/drm/i915/intel_dsi_dcs_backlight.c b/drivers/gpu/drm/i915/intel_dsi_dcs_backlight.c > > index 8c33262cb0b2..e5dd4962705c 100644 > > --- a/drivers/gpu/drm/i915/intel_dsi_dcs_backlight.c > > +++ b/drivers/gpu/drm/i915/intel_dsi_dcs_backlight.c > > @@ -159,11 +159,11 @@ static int dcs_setup_backlight(struct intel_connector *connector, > > int intel_dsi_dcs_init_backlight_funcs(struct intel_connector *intel_connector) > > { > > struct drm_device *dev = intel_connector->base.dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > struct intel_encoder *encoder = intel_connector->encoder; > > struct intel_panel *panel = &intel_connector->panel; > > - if (dev_priv->vbt.backlight.type != INTEL_BACKLIGHT_DSI_DCS) > > + if (i915->vbt.backlight.type != INTEL_BACKLIGHT_DSI_DCS) > > return -ENODEV; > > if (WARN_ON(encoder->type != INTEL_OUTPUT_DSI)) > > diff --git a/drivers/gpu/drm/i915/intel_dsi_vbt.c b/drivers/gpu/drm/i915/intel_dsi_vbt.c > > index e5b178660408..d44ebdfd8760 100644 > > --- a/drivers/gpu/drm/i915/intel_dsi_vbt.c > > +++ b/drivers/gpu/drm/i915/intel_dsi_vbt.c > > @@ -121,7 +121,7 @@ static inline enum port intel_dsi_seq_port_to_port(u8 port) > > static const u8 *mipi_exec_send_packet(struct intel_dsi *intel_dsi, > > const u8 *data) > > { > > - struct drm_i915_private *dev_priv = to_i915(intel_dsi->base.base.dev); > > + struct drm_i915_private *i915 = to_i915(intel_dsi->base.base.dev); > > struct mipi_dsi_device *dsi_device; > > u8 type, flags, seq_port; > > u16 len; > > @@ -192,7 +192,7 @@ static const u8 *mipi_exec_send_packet(struct intel_dsi *intel_dsi, > > break; > > } > > - if (INTEL_GEN(dev_priv) < 11) > > + if (INTEL_GEN(i915) < 11) > > vlv_dsi_wait_for_fifo_empty(intel_dsi, port); > > out: > > @@ -213,7 +213,7 @@ static const u8 *mipi_exec_delay(struct intel_dsi *intel_dsi, const u8 *data) > > return data; > > } > > -static void vlv_exec_gpio(struct drm_i915_private *dev_priv, > > +static void vlv_exec_gpio(struct drm_i915_private *i915, > > u8 gpio_source, u8 gpio_index, bool value) > > { > > struct gpio_map *map; > > @@ -228,7 +228,7 @@ static void vlv_exec_gpio(struct drm_i915_private *dev_priv, > > map = &vlv_gpio_table[gpio_index]; > > - if (dev_priv->vbt.dsi.seq_version >= 3) { > > + if (i915->vbt.dsi.seq_version >= 3) { > > /* XXX: this assumes vlv_gpio_table only has NC GPIOs. */ > > port = IOSF_PORT_GPIO_NC; > > } else { > > @@ -246,26 +246,26 @@ static void vlv_exec_gpio(struct drm_i915_private *dev_priv, > > pconf0 = VLV_GPIO_PCONF0(map->base_offset); > > padval = VLV_GPIO_PAD_VAL(map->base_offset); > > - vlv_iosf_sb_get(dev_priv, BIT(VLV_IOSF_SB_GPIO)); > > + vlv_iosf_sb_get(i915, BIT(VLV_IOSF_SB_GPIO)); > > if (!map->init) { > > /* FIXME: remove constant below */ > > - vlv_iosf_sb_write(dev_priv, port, pconf0, 0x2000CC00); > > + vlv_iosf_sb_write(i915, port, pconf0, 0x2000CC00); > > map->init = true; > > } > > tmp = 0x4 | value; > > - vlv_iosf_sb_write(dev_priv, port, padval, tmp); > > - vlv_iosf_sb_put(dev_priv, BIT(VLV_IOSF_SB_GPIO)); > > + vlv_iosf_sb_write(i915, port, padval, tmp); > > + vlv_iosf_sb_put(i915, BIT(VLV_IOSF_SB_GPIO)); > > } > > -static void chv_exec_gpio(struct drm_i915_private *dev_priv, > > +static void chv_exec_gpio(struct drm_i915_private *i915, > > u8 gpio_source, u8 gpio_index, bool value) > > { > > u16 cfg0, cfg1; > > u16 family_num; > > u8 port; > > - if (dev_priv->vbt.dsi.seq_version >= 3) { > > + if (i915->vbt.dsi.seq_version >= 3) { > > if (gpio_index >= CHV_GPIO_IDX_START_SE) { > > /* XXX: it's unclear whether 255->57 is part of SE. */ > > gpio_index -= CHV_GPIO_IDX_START_SE; > > @@ -301,15 +301,15 @@ static void chv_exec_gpio(struct drm_i915_private *dev_priv, > > cfg0 = CHV_GPIO_PAD_CFG0(family_num, gpio_index); > > cfg1 = CHV_GPIO_PAD_CFG1(family_num, gpio_index); > > - vlv_iosf_sb_get(dev_priv, BIT(VLV_IOSF_SB_GPIO)); > > - vlv_iosf_sb_write(dev_priv, port, cfg1, 0); > > - vlv_iosf_sb_write(dev_priv, port, cfg0, > > + vlv_iosf_sb_get(i915, BIT(VLV_IOSF_SB_GPIO)); > > + vlv_iosf_sb_write(i915, port, cfg1, 0); > > + vlv_iosf_sb_write(i915, port, cfg0, > > CHV_GPIO_GPIOEN | CHV_GPIO_GPIOCFG_GPO | > > CHV_GPIO_GPIOTXSTATE(value)); > > - vlv_iosf_sb_put(dev_priv, BIT(VLV_IOSF_SB_GPIO)); > > + vlv_iosf_sb_put(i915, BIT(VLV_IOSF_SB_GPIO)); > > } > > -static void bxt_exec_gpio(struct drm_i915_private *dev_priv, > > +static void bxt_exec_gpio(struct drm_i915_private *i915, > > u8 gpio_source, u8 gpio_index, bool value) > > { > > /* XXX: this table is a quick ugly hack. */ > > @@ -317,7 +317,7 @@ static void bxt_exec_gpio(struct drm_i915_private *dev_priv, > > struct gpio_desc *gpio_desc = bxt_gpio_table[gpio_index]; > > if (!gpio_desc) { > > - gpio_desc = devm_gpiod_get_index(dev_priv->drm.dev, > > + gpio_desc = devm_gpiod_get_index(i915->drm.dev, > > NULL, gpio_index, > > value ? GPIOD_OUT_LOW : > > GPIOD_OUT_HIGH); > > @@ -334,7 +334,7 @@ static void bxt_exec_gpio(struct drm_i915_private *dev_priv, > > gpiod_set_value(gpio_desc, value); > > } > > -static void icl_exec_gpio(struct drm_i915_private *dev_priv, > > +static void icl_exec_gpio(struct drm_i915_private *i915, > > u8 gpio_source, u8 gpio_index, bool value) > > { > > DRM_DEBUG_KMS("Skipping ICL GPIO element execution\n"); > > @@ -343,19 +343,19 @@ static void icl_exec_gpio(struct drm_i915_private *dev_priv, > > static const u8 *mipi_exec_gpio(struct intel_dsi *intel_dsi, const u8 *data) > > { > > struct drm_device *dev = intel_dsi->base.base.dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > u8 gpio_source, gpio_index = 0, gpio_number; > > bool value; > > DRM_DEBUG_KMS("\n"); > > - if (dev_priv->vbt.dsi.seq_version >= 3) > > + if (i915->vbt.dsi.seq_version >= 3) > > gpio_index = *data++; > > gpio_number = *data++; > > /* gpio source in sequence v2 only */ > > - if (dev_priv->vbt.dsi.seq_version == 2) > > + if (i915->vbt.dsi.seq_version == 2) > > gpio_source = (*data >> 1) & 3; > > else > > gpio_source = 0; > > @@ -363,14 +363,14 @@ static const u8 *mipi_exec_gpio(struct intel_dsi *intel_dsi, const u8 *data) > > /* pull up/down */ > > value = *data++ & 1; > > - if (INTEL_GEN(dev_priv) >= 11) > > - icl_exec_gpio(dev_priv, gpio_source, gpio_index, value); > > - else if (IS_VALLEYVIEW(dev_priv)) > > - vlv_exec_gpio(dev_priv, gpio_source, gpio_number, value); > > - else if (IS_CHERRYVIEW(dev_priv)) > > - chv_exec_gpio(dev_priv, gpio_source, gpio_number, value); > > + if (INTEL_GEN(i915) >= 11) > > + icl_exec_gpio(i915, gpio_source, gpio_index, value); > > + else if (IS_VALLEYVIEW(i915)) > > + vlv_exec_gpio(i915, gpio_source, gpio_number, value); > > + else if (IS_CHERRYVIEW(i915)) > > + chv_exec_gpio(i915, gpio_source, gpio_number, value); > > else > > - bxt_exec_gpio(dev_priv, gpio_source, gpio_index, value); > > + bxt_exec_gpio(i915, gpio_source, gpio_index, value); > > return data; > > } > > @@ -456,14 +456,14 @@ static const char *sequence_name(enum mipi_seq seq_id) > > void intel_dsi_vbt_exec_sequence(struct intel_dsi *intel_dsi, > > enum mipi_seq seq_id) > > { > > - struct drm_i915_private *dev_priv = to_i915(intel_dsi->base.base.dev); > > + struct drm_i915_private *i915 = to_i915(intel_dsi->base.base.dev); > > const u8 *data; > > fn_mipi_elem_exec mipi_elem_exec; > > - if (WARN_ON(seq_id >= ARRAY_SIZE(dev_priv->vbt.dsi.sequence))) > > + if (WARN_ON(seq_id >= ARRAY_SIZE(i915->vbt.dsi.sequence))) > > return; > > - data = dev_priv->vbt.dsi.sequence[seq_id]; > > + data = i915->vbt.dsi.sequence[seq_id]; > > if (!data) > > return; > > @@ -476,7 +476,7 @@ void intel_dsi_vbt_exec_sequence(struct intel_dsi *intel_dsi, > > data++; > > /* Skip Size of Sequence. */ > > - if (dev_priv->vbt.dsi.seq_version >= 3) > > + if (i915->vbt.dsi.seq_version >= 3) > > data += 4; > > while (1) { > > @@ -492,7 +492,7 @@ void intel_dsi_vbt_exec_sequence(struct intel_dsi *intel_dsi, > > mipi_elem_exec = NULL; > > /* Size of Operation. */ > > - if (dev_priv->vbt.dsi.seq_version >= 3) > > + if (i915->vbt.dsi.seq_version >= 3) > > operation_size = *data++; > > if (mipi_elem_exec) { > > @@ -521,10 +521,10 @@ void intel_dsi_vbt_exec_sequence(struct intel_dsi *intel_dsi, > > void intel_dsi_msleep(struct intel_dsi *intel_dsi, int msec) > > { > > - struct drm_i915_private *dev_priv = to_i915(intel_dsi->base.base.dev); > > + struct drm_i915_private *i915 = to_i915(intel_dsi->base.base.dev); > > /* For v3 VBTs in vid-mode the delays are part of the VBT sequences */ > > - if (is_vid_mode(intel_dsi) && dev_priv->vbt.dsi.seq_version >= 3) > > + if (is_vid_mode(intel_dsi) && i915->vbt.dsi.seq_version >= 3) > > return; > > msleep(msec); > > @@ -571,10 +571,10 @@ void intel_dsi_log_params(struct intel_dsi *intel_dsi) > > bool intel_dsi_vbt_init(struct intel_dsi *intel_dsi, u16 panel_id) > > { > > struct drm_device *dev = intel_dsi->base.base.dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > - struct mipi_config *mipi_config = dev_priv->vbt.dsi.config; > > - struct mipi_pps_data *pps = dev_priv->vbt.dsi.pps; > > - struct drm_display_mode *mode = dev_priv->vbt.lfp_lvds_vbt_mode; > > + struct drm_i915_private *i915 = to_i915(dev); > > + struct mipi_config *mipi_config = i915->vbt.dsi.config; > > + struct mipi_pps_data *pps = i915->vbt.dsi.pps; > > + struct drm_display_mode *mode = i915->vbt.lfp_lvds_vbt_mode; > > u16 burst_mode_ratio; > > enum port port; > > diff --git a/drivers/gpu/drm/i915/intel_dvo.c b/drivers/gpu/drm/i915/intel_dvo.c > > index 22666d28f4aa..6ced95b73538 100644 > > --- a/drivers/gpu/drm/i915/intel_dvo.c > > +++ b/drivers/gpu/drm/i915/intel_dvo.c > > @@ -133,7 +133,7 @@ static struct intel_dvo *intel_attached_dvo(struct drm_connector *connector) > > static bool intel_dvo_connector_get_hw_state(struct intel_connector *connector) > > { > > struct drm_device *dev = connector->base.dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > struct intel_dvo *intel_dvo = intel_attached_dvo(&connector->base); > > u32 tmp; > > @@ -148,7 +148,7 @@ static bool intel_dvo_connector_get_hw_state(struct intel_connector *connector) > > static bool intel_dvo_get_hw_state(struct intel_encoder *encoder, > > enum pipe *pipe) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_dvo *intel_dvo = enc_to_dvo(encoder); > > u32 tmp; > > @@ -162,7 +162,7 @@ static bool intel_dvo_get_hw_state(struct intel_encoder *encoder, > > static void intel_dvo_get_config(struct intel_encoder *encoder, > > struct intel_crtc_state *pipe_config) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_dvo *intel_dvo = enc_to_dvo(encoder); > > u32 tmp, flags = 0; > > @@ -187,7 +187,7 @@ static void intel_disable_dvo(struct intel_encoder *encoder, > > const struct intel_crtc_state *old_crtc_state, > > const struct drm_connector_state *old_conn_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_dvo *intel_dvo = enc_to_dvo(encoder); > > i915_reg_t dvo_reg = intel_dvo->dev.dvo_reg; > > u32 temp = I915_READ(dvo_reg); > > @@ -201,7 +201,7 @@ static void intel_enable_dvo(struct intel_encoder *encoder, > > const struct intel_crtc_state *pipe_config, > > const struct drm_connector_state *conn_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_dvo *intel_dvo = enc_to_dvo(encoder); > > i915_reg_t dvo_reg = intel_dvo->dev.dvo_reg; > > u32 temp = I915_READ(dvo_reg); > > @@ -276,7 +276,7 @@ static void intel_dvo_pre_enable(struct intel_encoder *encoder, > > const struct intel_crtc_state *pipe_config, > > const struct drm_connector_state *conn_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_crtc *crtc = to_intel_crtc(pipe_config->base.crtc); > > const struct drm_display_mode *adjusted_mode = &pipe_config->base.adjusted_mode; > > struct intel_dvo *intel_dvo = enc_to_dvo(encoder); > > @@ -319,7 +319,7 @@ intel_dvo_detect(struct drm_connector *connector, bool force) > > static int intel_dvo_get_modes(struct drm_connector *connector) > > { > > - struct drm_i915_private *dev_priv = to_i915(connector->dev); > > + struct drm_i915_private *i915 = to_i915(connector->dev); > > const struct drm_display_mode *fixed_mode = > > to_intel_connector(connector)->panel.fixed_mode; > > @@ -330,7 +330,7 @@ static int intel_dvo_get_modes(struct drm_connector *connector) > > * that's not the case. > > */ > > intel_ddc_get_modes(connector, > > - intel_gmbus_get_adapter(dev_priv, GMBUS_PIN_DPC)); > > + intel_gmbus_get_adapter(i915, GMBUS_PIN_DPC)); > > if (!list_empty(&connector->probed_modes)) > > return 1; > > @@ -406,7 +406,7 @@ static enum port intel_dvo_port(i915_reg_t dvo_reg) > > return PORT_C; > > } > > -void intel_dvo_init(struct drm_i915_private *dev_priv) > > +void intel_dvo_init(struct drm_i915_private *i915) > > { > > struct intel_encoder *intel_encoder; > > struct intel_dvo *intel_dvo; > > @@ -452,7 +452,7 @@ void intel_dvo_init(struct drm_i915_private *dev_priv) > > * special cases, but otherwise default to what's defined > > * in the spec. > > */ > > - if (intel_gmbus_is_valid_pin(dev_priv, dvo->gpio)) > > + if (intel_gmbus_is_valid_pin(i915, dvo->gpio)) > > gpio = dvo->gpio; > > else if (dvo->type == INTEL_DVO_CHIP_LVDS) > > gpio = GMBUS_PIN_SSC; > > @@ -464,7 +464,7 @@ void intel_dvo_init(struct drm_i915_private *dev_priv) > > * It appears that everything is on GPIOE except for panels > > * on i830 laptops, which are on GPIOB (DVOA). > > */ > > - i2c = intel_gmbus_get_adapter(dev_priv, gpio); > > + i2c = intel_gmbus_get_adapter(i915, gpio); > > intel_dvo->dev = *dvo; > > @@ -480,7 +480,7 @@ void intel_dvo_init(struct drm_i915_private *dev_priv) > > * have the clock enabled before we attempt to > > * initialize the device. > > */ > > - for_each_pipe(dev_priv, pipe) { > > + for_each_pipe(i915, pipe) { > > dpll[pipe] = I915_READ(DPLL(pipe)); > > I915_WRITE(DPLL(pipe), dpll[pipe] | DPLL_DVO_2X_MODE); > > } > > @@ -488,7 +488,7 @@ void intel_dvo_init(struct drm_i915_private *dev_priv) > > dvoinit = dvo->dev_ops->init(&intel_dvo->dev, i2c); > > /* restore the DVO 2x clock state to original */ > > - for_each_pipe(dev_priv, pipe) { > > + for_each_pipe(i915, pipe) { > > I915_WRITE(DPLL(pipe), dpll[pipe]); > > } > > @@ -498,7 +498,7 @@ void intel_dvo_init(struct drm_i915_private *dev_priv) > > continue; > > port = intel_dvo_port(dvo->dvo_reg); > > - drm_encoder_init(&dev_priv->drm, &intel_encoder->base, > > + drm_encoder_init(&i915->drm, &intel_encoder->base, > > &intel_dvo_enc_funcs, encoder_type, > > "DVO %c", port_name(port)); > > @@ -511,14 +511,14 @@ void intel_dvo_init(struct drm_i915_private *dev_priv) > > case INTEL_DVO_CHIP_TMDS: > > intel_encoder->cloneable = (1 << INTEL_OUTPUT_ANALOG) | > > (1 << INTEL_OUTPUT_DVO); > > - drm_connector_init(&dev_priv->drm, connector, > > + drm_connector_init(&i915->drm, connector, > > &intel_dvo_connector_funcs, > > DRM_MODE_CONNECTOR_DVII); > > encoder_type = DRM_MODE_ENCODER_TMDS; > > break; > > case INTEL_DVO_CHIP_LVDS: > > intel_encoder->cloneable = 0; > > - drm_connector_init(&dev_priv->drm, connector, > > + drm_connector_init(&i915->drm, connector, > > &intel_dvo_connector_funcs, > > DRM_MODE_CONNECTOR_LVDS); > > encoder_type = DRM_MODE_ENCODER_LVDS; > > diff --git a/drivers/gpu/drm/i915/intel_dvo.h b/drivers/gpu/drm/i915/intel_dvo.h > > index 3ed0fdf8efff..d865cd439bb7 100644 > > --- a/drivers/gpu/drm/i915/intel_dvo.h > > +++ b/drivers/gpu/drm/i915/intel_dvo.h > > @@ -8,6 +8,6 @@ > > struct drm_i915_private; > > -void intel_dvo_init(struct drm_i915_private *dev_priv); > > +void intel_dvo_init(struct drm_i915_private *i915); > > #endif /* __INTEL_DVO_H__ */ > > diff --git a/drivers/gpu/drm/i915/intel_dvo_dev.h b/drivers/gpu/drm/i915/intel_dvo_dev.h > > index 94a6ae1e0292..1d184dc7922b 100644 > > --- a/drivers/gpu/drm/i915/intel_dvo_dev.h > > +++ b/drivers/gpu/drm/i915/intel_dvo_dev.h > > @@ -40,7 +40,7 @@ struct intel_dvo_device { > > int slave_addr; > > const struct intel_dvo_dev_ops *dev_ops; > > - void *dev_priv; > > + void *i915; > > struct i2c_adapter *i2c_bus; > > }; > > diff --git a/drivers/gpu/drm/i915/intel_fbc.c b/drivers/gpu/drm/i915/intel_fbc.c > > index 5679f2fffb7c..627fed2a55d2 100644 > > --- a/drivers/gpu/drm/i915/intel_fbc.c > > +++ b/drivers/gpu/drm/i915/intel_fbc.c > > @@ -45,14 +45,14 @@ > > #include "intel_fbc.h" > > #include "intel_frontbuffer.h" > > -static inline bool fbc_supported(struct drm_i915_private *dev_priv) > > +static inline bool fbc_supported(struct drm_i915_private *i915) > > { > > - return HAS_FBC(dev_priv); > > + return HAS_FBC(i915); > > } > > -static inline bool no_fbc_on_multiple_pipes(struct drm_i915_private *dev_priv) > > +static inline bool no_fbc_on_multiple_pipes(struct drm_i915_private *i915) > > { > > - return INTEL_GEN(dev_priv) <= 3; > > + return INTEL_GEN(i915) <= 3; > > } > > /* > > @@ -82,22 +82,22 @@ static void intel_fbc_get_plane_source_size(struct intel_fbc_state_cache *cache, > > *height = cache->plane.src_h; > > } > > -static int intel_fbc_calculate_cfb_size(struct drm_i915_private *dev_priv, > > +static int intel_fbc_calculate_cfb_size(struct drm_i915_private *i915, > > struct intel_fbc_state_cache *cache) > > { > > int lines; > > intel_fbc_get_plane_source_size(cache, NULL, &lines); > > - if (IS_GEN(dev_priv, 7)) > > + if (IS_GEN(i915, 7)) > > lines = min(lines, 2048); > > - else if (INTEL_GEN(dev_priv) >= 8) > > + else if (INTEL_GEN(i915) >= 8) > > lines = min(lines, 2560); > > /* Hardware needs the full buffer stride, not just the active area. */ > > return lines * cache->fb.stride; > > } > > -static void i8xx_fbc_deactivate(struct drm_i915_private *dev_priv) > > +static void i8xx_fbc_deactivate(struct drm_i915_private *i915) > > { > > u32 fbc_ctl; > > @@ -110,7 +110,7 @@ static void i8xx_fbc_deactivate(struct drm_i915_private *dev_priv) > > I915_WRITE(FBC_CONTROL, fbc_ctl); > > /* Wait for compressing bit to clear */ > > - if (intel_wait_for_register(&dev_priv->uncore, > > + if (intel_wait_for_register(&i915->uncore, > > FBC_STATUS, FBC_STAT_COMPRESSING, 0, > > 10)) { > > DRM_DEBUG_KMS("FBC idle timed out\n"); > > @@ -118,9 +118,9 @@ static void i8xx_fbc_deactivate(struct drm_i915_private *dev_priv) > > } > > } > > -static void i8xx_fbc_activate(struct drm_i915_private *dev_priv) > > +static void i8xx_fbc_activate(struct drm_i915_private *i915) > > { > > - struct intel_fbc_reg_params *params = &dev_priv->fbc.params; > > + struct intel_fbc_reg_params *params = &i915->fbc.params; > > int cfb_pitch; > > int i; > > u32 fbc_ctl; > > @@ -131,7 +131,7 @@ static void i8xx_fbc_activate(struct drm_i915_private *dev_priv) > > cfb_pitch = params->fb.stride; > > /* FBC_CTL wants 32B or 64B units */ > > - if (IS_GEN(dev_priv, 2)) > > + if (IS_GEN(i915, 2)) > > cfb_pitch = (cfb_pitch / 32) - 1; > > else > > cfb_pitch = (cfb_pitch / 64) - 1; > > @@ -140,7 +140,7 @@ static void i8xx_fbc_activate(struct drm_i915_private *dev_priv) > > for (i = 0; i < (FBC_LL_SIZE / 32) + 1; i++) > > I915_WRITE(FBC_TAG(i), 0); > > - if (IS_GEN(dev_priv, 4)) { > > + if (IS_GEN(i915, 4)) { > > u32 fbc_ctl2; > > /* Set it up... */ > > @@ -154,21 +154,21 @@ static void i8xx_fbc_activate(struct drm_i915_private *dev_priv) > > fbc_ctl = I915_READ(FBC_CONTROL); > > fbc_ctl &= 0x3fff << FBC_CTL_INTERVAL_SHIFT; > > fbc_ctl |= FBC_CTL_EN | FBC_CTL_PERIODIC; > > - if (IS_I945GM(dev_priv)) > > + if (IS_I945GM(i915)) > > fbc_ctl |= FBC_CTL_C3_IDLE; /* 945 needs special SR handling */ > > fbc_ctl |= (cfb_pitch & 0xff) << FBC_CTL_STRIDE_SHIFT; > > fbc_ctl |= params->vma->fence->id; > > I915_WRITE(FBC_CONTROL, fbc_ctl); > > } > > -static bool i8xx_fbc_is_active(struct drm_i915_private *dev_priv) > > +static bool i8xx_fbc_is_active(struct drm_i915_private *i915) > > { > > return I915_READ(FBC_CONTROL) & FBC_CTL_EN; > > } > > -static void g4x_fbc_activate(struct drm_i915_private *dev_priv) > > +static void g4x_fbc_activate(struct drm_i915_private *i915) > > { > > - struct intel_fbc_reg_params *params = &dev_priv->fbc.params; > > + struct intel_fbc_reg_params *params = &i915->fbc.params; > > u32 dpfc_ctl; > > dpfc_ctl = DPFC_CTL_PLANE(params->crtc.i9xx_plane) | DPFC_SR_EN; > > @@ -188,7 +188,7 @@ static void g4x_fbc_activate(struct drm_i915_private *dev_priv) > > I915_WRITE(DPFC_CONTROL, dpfc_ctl | DPFC_CTL_EN); > > } > > -static void g4x_fbc_deactivate(struct drm_i915_private *dev_priv) > > +static void g4x_fbc_deactivate(struct drm_i915_private *i915) > > { > > u32 dpfc_ctl; > > @@ -200,23 +200,23 @@ static void g4x_fbc_deactivate(struct drm_i915_private *dev_priv) > > } > > } > > -static bool g4x_fbc_is_active(struct drm_i915_private *dev_priv) > > +static bool g4x_fbc_is_active(struct drm_i915_private *i915) > > { > > return I915_READ(DPFC_CONTROL) & DPFC_CTL_EN; > > } > > /* This function forces a CFB recompression through the nuke operation. */ > > -static void intel_fbc_recompress(struct drm_i915_private *dev_priv) > > +static void intel_fbc_recompress(struct drm_i915_private *i915) > > { > > I915_WRITE(MSG_FBC_REND_STATE, FBC_REND_NUKE); > > POSTING_READ(MSG_FBC_REND_STATE); > > } > > -static void ilk_fbc_activate(struct drm_i915_private *dev_priv) > > +static void ilk_fbc_activate(struct drm_i915_private *i915) > > { > > - struct intel_fbc_reg_params *params = &dev_priv->fbc.params; > > + struct intel_fbc_reg_params *params = &i915->fbc.params; > > u32 dpfc_ctl; > > - int threshold = dev_priv->fbc.threshold; > > + int threshold = i915->fbc.threshold; > > dpfc_ctl = DPFC_CTL_PLANE(params->crtc.i9xx_plane); > > if (params->fb.format->cpp[0] == 2) > > @@ -237,9 +237,9 @@ static void ilk_fbc_activate(struct drm_i915_private *dev_priv) > > if (params->flags & PLANE_HAS_FENCE) { > > dpfc_ctl |= DPFC_CTL_FENCE_EN; > > - if (IS_GEN(dev_priv, 5)) > > + if (IS_GEN(i915, 5)) > > dpfc_ctl |= params->vma->fence->id; > > - if (IS_GEN(dev_priv, 6)) { > > + if (IS_GEN(i915, 6)) { > > I915_WRITE(SNB_DPFC_CTL_SA, > > SNB_CPU_FENCE_ENABLE | > > params->vma->fence->id); > > @@ -247,7 +247,7 @@ static void ilk_fbc_activate(struct drm_i915_private *dev_priv) > > params->crtc.fence_y_offset); > > } > > } else { > > - if (IS_GEN(dev_priv, 6)) { > > + if (IS_GEN(i915, 6)) { > > I915_WRITE(SNB_DPFC_CTL_SA, 0); > > I915_WRITE(DPFC_CPU_FENCE_OFFSET, 0); > > } > > @@ -259,10 +259,10 @@ static void ilk_fbc_activate(struct drm_i915_private *dev_priv) > > /* enable it... */ > > I915_WRITE(ILK_DPFC_CONTROL, dpfc_ctl | DPFC_CTL_EN); > > - intel_fbc_recompress(dev_priv); > > + intel_fbc_recompress(i915); > > } > > -static void ilk_fbc_deactivate(struct drm_i915_private *dev_priv) > > +static void ilk_fbc_deactivate(struct drm_i915_private *i915) > > { > > u32 dpfc_ctl; > > @@ -274,19 +274,19 @@ static void ilk_fbc_deactivate(struct drm_i915_private *dev_priv) > > } > > } > > -static bool ilk_fbc_is_active(struct drm_i915_private *dev_priv) > > +static bool ilk_fbc_is_active(struct drm_i915_private *i915) > > { > > return I915_READ(ILK_DPFC_CONTROL) & DPFC_CTL_EN; > > } > > -static void gen7_fbc_activate(struct drm_i915_private *dev_priv) > > +static void gen7_fbc_activate(struct drm_i915_private *i915) > > { > > - struct intel_fbc_reg_params *params = &dev_priv->fbc.params; > > + struct intel_fbc_reg_params *params = &i915->fbc.params; > > u32 dpfc_ctl; > > - int threshold = dev_priv->fbc.threshold; > > + int threshold = i915->fbc.threshold; > > /* Display WA #0529: skl, kbl, bxt. */ > > - if (IS_GEN(dev_priv, 9) && !IS_GEMINILAKE(dev_priv)) { > > + if (IS_GEN(i915, 9) && !IS_GEMINILAKE(i915)) { > > u32 val = I915_READ(CHICKEN_MISC_4); > > val &= ~(FBC_STRIDE_OVERRIDE | FBC_STRIDE_MASK); > > @@ -299,7 +299,7 @@ static void gen7_fbc_activate(struct drm_i915_private *dev_priv) > > } > > dpfc_ctl = 0; > > - if (IS_IVYBRIDGE(dev_priv)) > > + if (IS_IVYBRIDGE(i915)) > > dpfc_ctl |= IVB_DPFC_CTL_PLANE(params->crtc.i9xx_plane); > > if (params->fb.format->cpp[0] == 2) > > @@ -329,15 +329,15 @@ static void gen7_fbc_activate(struct drm_i915_private *dev_priv) > > I915_WRITE(DPFC_CPU_FENCE_OFFSET, 0); > > } > > - if (dev_priv->fbc.false_color) > > + if (i915->fbc.false_color) > > dpfc_ctl |= FBC_CTL_FALSE_COLOR; > > - if (IS_IVYBRIDGE(dev_priv)) { > > + if (IS_IVYBRIDGE(i915)) { > > /* WaFbcAsynchFlipDisableFbcQueue:ivb */ > > I915_WRITE(ILK_DISPLAY_CHICKEN1, > > I915_READ(ILK_DISPLAY_CHICKEN1) | > > ILK_FBCQ_DIS); > > - } else if (IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv)) { > > + } else if (IS_HASWELL(i915) || IS_BROADWELL(i915)) { > > /* WaFbcAsynchFlipDisableFbcQueue:hsw,bdw */ > > I915_WRITE(CHICKEN_PIPESL_1(params->crtc.pipe), > > I915_READ(CHICKEN_PIPESL_1(params->crtc.pipe)) | > > @@ -346,72 +346,72 @@ static void gen7_fbc_activate(struct drm_i915_private *dev_priv) > > I915_WRITE(ILK_DPFC_CONTROL, dpfc_ctl | DPFC_CTL_EN); > > - intel_fbc_recompress(dev_priv); > > + intel_fbc_recompress(i915); > > } > > -static bool intel_fbc_hw_is_active(struct drm_i915_private *dev_priv) > > +static bool intel_fbc_hw_is_active(struct drm_i915_private *i915) > > { > > - if (INTEL_GEN(dev_priv) >= 5) > > - return ilk_fbc_is_active(dev_priv); > > - else if (IS_GM45(dev_priv)) > > - return g4x_fbc_is_active(dev_priv); > > + if (INTEL_GEN(i915) >= 5) > > + return ilk_fbc_is_active(i915); > > + else if (IS_GM45(i915)) > > + return g4x_fbc_is_active(i915); > > else > > - return i8xx_fbc_is_active(dev_priv); > > + return i8xx_fbc_is_active(i915); > > } > > -static void intel_fbc_hw_activate(struct drm_i915_private *dev_priv) > > +static void intel_fbc_hw_activate(struct drm_i915_private *i915) > > { > > - struct intel_fbc *fbc = &dev_priv->fbc; > > + struct intel_fbc *fbc = &i915->fbc; > > fbc->active = true; > > - if (INTEL_GEN(dev_priv) >= 7) > > - gen7_fbc_activate(dev_priv); > > - else if (INTEL_GEN(dev_priv) >= 5) > > - ilk_fbc_activate(dev_priv); > > - else if (IS_GM45(dev_priv)) > > - g4x_fbc_activate(dev_priv); > > + if (INTEL_GEN(i915) >= 7) > > + gen7_fbc_activate(i915); > > + else if (INTEL_GEN(i915) >= 5) > > + ilk_fbc_activate(i915); > > + else if (IS_GM45(i915)) > > + g4x_fbc_activate(i915); > > else > > - i8xx_fbc_activate(dev_priv); > > + i8xx_fbc_activate(i915); > > } > > -static void intel_fbc_hw_deactivate(struct drm_i915_private *dev_priv) > > +static void intel_fbc_hw_deactivate(struct drm_i915_private *i915) > > { > > - struct intel_fbc *fbc = &dev_priv->fbc; > > + struct intel_fbc *fbc = &i915->fbc; > > fbc->active = false; > > - if (INTEL_GEN(dev_priv) >= 5) > > - ilk_fbc_deactivate(dev_priv); > > - else if (IS_GM45(dev_priv)) > > - g4x_fbc_deactivate(dev_priv); > > + if (INTEL_GEN(i915) >= 5) > > + ilk_fbc_deactivate(i915); > > + else if (IS_GM45(i915)) > > + g4x_fbc_deactivate(i915); > > else > > - i8xx_fbc_deactivate(dev_priv); > > + i8xx_fbc_deactivate(i915); > > } > > /** > > * intel_fbc_is_active - Is FBC active? > > - * @dev_priv: i915 device instance > > + * @i915: i915 device instance > > * > > * This function is used to verify the current state of FBC. > > * > > * FIXME: This should be tracked in the plane config eventually > > * instead of queried at runtime for most callers. > > */ > > -bool intel_fbc_is_active(struct drm_i915_private *dev_priv) > > +bool intel_fbc_is_active(struct drm_i915_private *i915) > > { > > - return dev_priv->fbc.active; > > + return i915->fbc.active; > > } > > -static void intel_fbc_deactivate(struct drm_i915_private *dev_priv, > > +static void intel_fbc_deactivate(struct drm_i915_private *i915, > > const char *reason) > > { > > - struct intel_fbc *fbc = &dev_priv->fbc; > > + struct intel_fbc *fbc = &i915->fbc; > > WARN_ON(!mutex_is_locked(&fbc->lock)); > > if (fbc->active) > > - intel_fbc_hw_deactivate(dev_priv); > > + intel_fbc_hw_deactivate(i915); > > fbc->no_fbc_reason = reason; > > } > > @@ -419,12 +419,12 @@ static void intel_fbc_deactivate(struct drm_i915_private *dev_priv, > > static bool multiple_pipes_ok(struct intel_crtc *crtc, > > struct intel_plane_state *plane_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > - struct intel_fbc *fbc = &dev_priv->fbc; > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > + struct intel_fbc *fbc = &i915->fbc; > > enum pipe pipe = crtc->pipe; > > /* Don't even bother tracking anything we don't need. */ > > - if (!no_fbc_on_multiple_pipes(dev_priv)) > > + if (!no_fbc_on_multiple_pipes(i915)) > > return true; > > if (plane_state->base.visible) > > @@ -435,7 +435,7 @@ static bool multiple_pipes_ok(struct intel_crtc *crtc, > > return (fbc->visible_pipes_mask & ~(1 << pipe)) != 0; > > } > > -static int find_compression_threshold(struct drm_i915_private *dev_priv, > > +static int find_compression_threshold(struct drm_i915_private *i915, > > struct drm_mm_node *node, > > int size, > > int fb_cpp) > > @@ -448,8 +448,8 @@ static int find_compression_threshold(struct drm_i915_private *dev_priv, > > * reserved range size, so it always assumes the maximum (8mb) is used. > > * If we enable FBC using a CFB on that memory range we'll get FIFO > > * underruns, even if that range is not reserved by the BIOS. */ > > - if (IS_BROADWELL(dev_priv) || IS_GEN9_BC(dev_priv)) > > - end = resource_size(&dev_priv->dsm) - 8 * 1024 * 1024; > > + if (IS_BROADWELL(i915) || IS_GEN9_BC(i915)) > > + end = resource_size(&i915->dsm) - 8 * 1024 * 1024; > > else > > end = U64_MAX; > > @@ -461,7 +461,7 @@ static int find_compression_threshold(struct drm_i915_private *dev_priv, > > */ > > /* Try to over-allocate to reduce reallocations and fragmentation. */ > > - ret = i915_gem_stolen_insert_node_in_range(dev_priv, node, size <<= 1, > > + ret = i915_gem_stolen_insert_node_in_range(i915, node, size <<= 1, > > 4096, 0, end); > > if (ret == 0) > > return compression_threshold; > > @@ -472,9 +472,9 @@ static int find_compression_threshold(struct drm_i915_private *dev_priv, > > (fb_cpp == 2 && compression_threshold == 2)) > > return 0; > > - ret = i915_gem_stolen_insert_node_in_range(dev_priv, node, size >>= 1, > > + ret = i915_gem_stolen_insert_node_in_range(i915, node, size >>= 1, > > 4096, 0, end); > > - if (ret && INTEL_GEN(dev_priv) <= 4) { > > + if (ret && INTEL_GEN(i915) <= 4) { > > return 0; > > } else if (ret) { > > compression_threshold <<= 1; > > @@ -486,17 +486,17 @@ static int find_compression_threshold(struct drm_i915_private *dev_priv, > > static int intel_fbc_alloc_cfb(struct intel_crtc *crtc) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > - struct intel_fbc *fbc = &dev_priv->fbc; > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > + struct intel_fbc *fbc = &i915->fbc; > > struct drm_mm_node *uninitialized_var(compressed_llb); > > int size, fb_cpp, ret; > > WARN_ON(drm_mm_node_allocated(&fbc->compressed_fb)); > > - size = intel_fbc_calculate_cfb_size(dev_priv, &fbc->state_cache); > > + size = intel_fbc_calculate_cfb_size(i915, &fbc->state_cache); > > fb_cpp = fbc->state_cache.fb.format->cpp[0]; > > - ret = find_compression_threshold(dev_priv, &fbc->compressed_fb, > > + ret = find_compression_threshold(i915, &fbc->compressed_fb, > > size, fb_cpp); > > if (!ret) > > goto err_llb; > > @@ -507,32 +507,32 @@ static int intel_fbc_alloc_cfb(struct intel_crtc *crtc) > > fbc->threshold = ret; > > - if (INTEL_GEN(dev_priv) >= 5) > > + if (INTEL_GEN(i915) >= 5) > > I915_WRITE(ILK_DPFC_CB_BASE, fbc->compressed_fb.start); > > - else if (IS_GM45(dev_priv)) { > > + else if (IS_GM45(i915)) { > > I915_WRITE(DPFC_CB_BASE, fbc->compressed_fb.start); > > } else { > > compressed_llb = kzalloc(sizeof(*compressed_llb), GFP_KERNEL); > > if (!compressed_llb) > > goto err_fb; > > - ret = i915_gem_stolen_insert_node(dev_priv, compressed_llb, > > + ret = i915_gem_stolen_insert_node(i915, compressed_llb, > > 4096, 4096); > > if (ret) > > goto err_fb; > > fbc->compressed_llb = compressed_llb; > > - GEM_BUG_ON(range_overflows_t(u64, dev_priv->dsm.start, > > + GEM_BUG_ON(range_overflows_t(u64, i915->dsm.start, > > fbc->compressed_fb.start, > > U32_MAX)); > > - GEM_BUG_ON(range_overflows_t(u64, dev_priv->dsm.start, > > + GEM_BUG_ON(range_overflows_t(u64, i915->dsm.start, > > fbc->compressed_llb->start, > > U32_MAX)); > > I915_WRITE(FBC_CFB_BASE, > > - dev_priv->dsm.start + fbc->compressed_fb.start); > > + i915->dsm.start + fbc->compressed_fb.start); > > I915_WRITE(FBC_LL_BASE, > > - dev_priv->dsm.start + compressed_llb->start); > > + i915->dsm.start + compressed_llb->start); > > } > > DRM_DEBUG_KMS("reserved %llu bytes of contiguous stolen space for FBC, threshold: %d\n", > > @@ -542,39 +542,39 @@ static int intel_fbc_alloc_cfb(struct intel_crtc *crtc) > > err_fb: > > kfree(compressed_llb); > > - i915_gem_stolen_remove_node(dev_priv, &fbc->compressed_fb); > > + i915_gem_stolen_remove_node(i915, &fbc->compressed_fb); > > err_llb: > > - if (drm_mm_initialized(&dev_priv->mm.stolen)) > > + if (drm_mm_initialized(&i915->mm.stolen)) > > pr_info_once("drm: not enough stolen space for compressed buffer (need %d more bytes), disabling. Hint: you may be able to increase stolen memory size in the BIOS to avoid this.\n", size); > > return -ENOSPC; > > } > > -static void __intel_fbc_cleanup_cfb(struct drm_i915_private *dev_priv) > > +static void __intel_fbc_cleanup_cfb(struct drm_i915_private *i915) > > { > > - struct intel_fbc *fbc = &dev_priv->fbc; > > + struct intel_fbc *fbc = &i915->fbc; > > if (drm_mm_node_allocated(&fbc->compressed_fb)) > > - i915_gem_stolen_remove_node(dev_priv, &fbc->compressed_fb); > > + i915_gem_stolen_remove_node(i915, &fbc->compressed_fb); > > if (fbc->compressed_llb) { > > - i915_gem_stolen_remove_node(dev_priv, fbc->compressed_llb); > > + i915_gem_stolen_remove_node(i915, fbc->compressed_llb); > > kfree(fbc->compressed_llb); > > } > > } > > -void intel_fbc_cleanup_cfb(struct drm_i915_private *dev_priv) > > +void intel_fbc_cleanup_cfb(struct drm_i915_private *i915) > > { > > - struct intel_fbc *fbc = &dev_priv->fbc; > > + struct intel_fbc *fbc = &i915->fbc; > > - if (!fbc_supported(dev_priv)) > > + if (!fbc_supported(i915)) > > return; > > mutex_lock(&fbc->lock); > > - __intel_fbc_cleanup_cfb(dev_priv); > > + __intel_fbc_cleanup_cfb(i915); > > mutex_unlock(&fbc->lock); > > } > > -static bool stride_is_valid(struct drm_i915_private *dev_priv, > > +static bool stride_is_valid(struct drm_i915_private *i915, > > unsigned int stride) > > { > > /* This should have been caught earlier. */ > > @@ -585,10 +585,10 @@ static bool stride_is_valid(struct drm_i915_private *dev_priv, > > if (stride < 512) > > return false; > > - if (IS_GEN(dev_priv, 2) || IS_GEN(dev_priv, 3)) > > + if (IS_GEN(i915, 2) || IS_GEN(i915, 3)) > > return stride == 4096 || stride == 8192; > > - if (IS_GEN(dev_priv, 4) && !IS_G4X(dev_priv) && stride < 2048) > > + if (IS_GEN(i915, 4) && !IS_G4X(i915) && stride < 2048) > > return false; > > if (stride > 16384) > > @@ -597,7 +597,7 @@ static bool stride_is_valid(struct drm_i915_private *dev_priv, > > return true; > > } > > -static bool pixel_format_is_valid(struct drm_i915_private *dev_priv, > > +static bool pixel_format_is_valid(struct drm_i915_private *i915, > > u32 pixel_format) > > { > > switch (pixel_format) { > > @@ -607,10 +607,10 @@ static bool pixel_format_is_valid(struct drm_i915_private *dev_priv, > > case DRM_FORMAT_XRGB1555: > > case DRM_FORMAT_RGB565: > > /* 16bpp not supported on gen2 */ > > - if (IS_GEN(dev_priv, 2)) > > + if (IS_GEN(i915, 2)) > > return false; > > /* WaFbcOnly1to1Ratio:ctg */ > > - if (IS_G4X(dev_priv)) > > + if (IS_G4X(i915)) > > return false; > > return true; > > default: > > @@ -626,17 +626,17 @@ static bool pixel_format_is_valid(struct drm_i915_private *dev_priv, > > */ > > static bool intel_fbc_hw_tracking_covers_screen(struct intel_crtc *crtc) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > - struct intel_fbc *fbc = &dev_priv->fbc; > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > + struct intel_fbc *fbc = &i915->fbc; > > unsigned int effective_w, effective_h, max_w, max_h; > > - if (INTEL_GEN(dev_priv) >= 10 || IS_GEMINILAKE(dev_priv)) { > > + if (INTEL_GEN(i915) >= 10 || IS_GEMINILAKE(i915)) { > > max_w = 5120; > > max_h = 4096; > > - } else if (INTEL_GEN(dev_priv) >= 8 || IS_HASWELL(dev_priv)) { > > + } else if (INTEL_GEN(i915) >= 8 || IS_HASWELL(i915)) { > > max_w = 4096; > > max_h = 4096; > > - } else if (IS_G4X(dev_priv) || INTEL_GEN(dev_priv) >= 5) { > > + } else if (IS_G4X(i915) || INTEL_GEN(i915) >= 5) { > > max_w = 4096; > > max_h = 2048; > > } else { > > @@ -656,8 +656,8 @@ static void intel_fbc_update_state_cache(struct intel_crtc *crtc, > > struct intel_crtc_state *crtc_state, > > struct intel_plane_state *plane_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > - struct intel_fbc *fbc = &dev_priv->fbc; > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > + struct intel_fbc *fbc = &i915->fbc; > > struct intel_fbc_state_cache *cache = &fbc->state_cache; > > struct drm_framebuffer *fb = plane_state->base.fb; > > @@ -665,7 +665,7 @@ static void intel_fbc_update_state_cache(struct intel_crtc *crtc, > > cache->flags = 0; > > cache->crtc.mode_flags = crtc_state->base.adjusted_mode.flags; > > - if (IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv)) > > + if (IS_HASWELL(i915) || IS_BROADWELL(i915)) > > cache->crtc.hsw_bdw_pixel_rate = crtc_state->pixel_rate; > > cache->plane.rotation = plane_state->base.rotation; > > @@ -697,8 +697,8 @@ static void intel_fbc_update_state_cache(struct intel_crtc *crtc, > > static bool intel_fbc_can_activate(struct intel_crtc *crtc) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > - struct intel_fbc *fbc = &dev_priv->fbc; > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > + struct intel_fbc *fbc = &i915->fbc; > > struct intel_fbc_state_cache *cache = &fbc->state_cache; > > /* We don't need to use a state cache here since this information is > > @@ -741,18 +741,18 @@ static bool intel_fbc_can_activate(struct intel_crtc *crtc) > > fbc->no_fbc_reason = "framebuffer not tiled or fenced"; > > return false; > > } > > - if (INTEL_GEN(dev_priv) <= 4 && !IS_G4X(dev_priv) && > > + if (INTEL_GEN(i915) <= 4 && !IS_G4X(i915) && > > cache->plane.rotation != DRM_MODE_ROTATE_0) { > > fbc->no_fbc_reason = "rotation unsupported"; > > return false; > > } > > - if (!stride_is_valid(dev_priv, cache->fb.stride)) { > > + if (!stride_is_valid(i915, cache->fb.stride)) { > > fbc->no_fbc_reason = "framebuffer stride not supported"; > > return false; > > } > > - if (!pixel_format_is_valid(dev_priv, cache->fb.format->format)) { > > + if (!pixel_format_is_valid(i915, cache->fb.format->format)) { > > fbc->no_fbc_reason = "pixel format is invalid"; > > return false; > > } > > @@ -764,8 +764,8 @@ static bool intel_fbc_can_activate(struct intel_crtc *crtc) > > } > > /* WaFbcExceedCdClockThreshold:hsw,bdw */ > > - if ((IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv)) && > > - cache->crtc.hsw_bdw_pixel_rate >= dev_priv->cdclk.hw.cdclk * 95 / 100) { > > + if ((IS_HASWELL(i915) || IS_BROADWELL(i915)) && > > + cache->crtc.hsw_bdw_pixel_rate >= i915->cdclk.hw.cdclk * 95 / 100) { > > fbc->no_fbc_reason = "pixel rate is too big"; > > return false; > > } > > @@ -780,7 +780,7 @@ static bool intel_fbc_can_activate(struct intel_crtc *crtc) > > * we didn't get any invalidate/deactivate calls, but this would require > > * a lot of tracking just for a specific case. If we conclude it's an > > * important case, we can implement it later. */ > > - if (intel_fbc_calculate_cfb_size(dev_priv, &fbc->state_cache) > > > + if (intel_fbc_calculate_cfb_size(i915, &fbc->state_cache) > > > fbc->compressed_fb.size * fbc->threshold) { > > fbc->no_fbc_reason = "CFB requirements changed"; > > return false; > > @@ -791,7 +791,7 @@ static bool intel_fbc_can_activate(struct intel_crtc *crtc) > > * having a Y offset that isn't divisible by 4 causes FIFO underrun > > * and screen flicker. > > */ > > - if (IS_GEN_RANGE(dev_priv, 9, 10) && > > + if (IS_GEN_RANGE(i915, 9, 10) && > > (fbc->state_cache.plane.adjusted_y & 3)) { > > fbc->no_fbc_reason = "plane Y offset is misaligned"; > > return false; > > @@ -800,11 +800,11 @@ static bool intel_fbc_can_activate(struct intel_crtc *crtc) > > return true; > > } > > -static bool intel_fbc_can_enable(struct drm_i915_private *dev_priv) > > +static bool intel_fbc_can_enable(struct drm_i915_private *i915) > > { > > - struct intel_fbc *fbc = &dev_priv->fbc; > > + struct intel_fbc *fbc = &i915->fbc; > > - if (intel_vgpu_active(dev_priv)) { > > + if (intel_vgpu_active(i915)) { > > fbc->no_fbc_reason = "VGPU is active"; > > return false; > > } > > @@ -825,8 +825,8 @@ static bool intel_fbc_can_enable(struct drm_i915_private *dev_priv) > > static void intel_fbc_get_reg_params(struct intel_crtc *crtc, > > struct intel_fbc_reg_params *params) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > - struct intel_fbc *fbc = &dev_priv->fbc; > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > + struct intel_fbc *fbc = &i915->fbc; > > struct intel_fbc_state_cache *cache = &fbc->state_cache; > > /* Since all our fields are integer types, use memset here so the > > @@ -844,9 +844,9 @@ static void intel_fbc_get_reg_params(struct intel_crtc *crtc, > > params->fb.format = cache->fb.format; > > params->fb.stride = cache->fb.stride; > > - params->cfb_size = intel_fbc_calculate_cfb_size(dev_priv, cache); > > + params->cfb_size = intel_fbc_calculate_cfb_size(i915, cache); > > - if (IS_GEN(dev_priv, 9) && !IS_GEMINILAKE(dev_priv)) > > + if (IS_GEN(i915, 9) && !IS_GEMINILAKE(i915)) > > params->gen9_wa_cfb_stride = DIV_ROUND_UP(cache->plane.src_w, > > 32 * fbc->threshold) * 8; > > } > > @@ -855,11 +855,11 @@ void intel_fbc_pre_update(struct intel_crtc *crtc, > > struct intel_crtc_state *crtc_state, > > struct intel_plane_state *plane_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > - struct intel_fbc *fbc = &dev_priv->fbc; > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > + struct intel_fbc *fbc = &i915->fbc; > > const char *reason = "update pending"; > > - if (!fbc_supported(dev_priv)) > > + if (!fbc_supported(i915)) > > return; > > mutex_lock(&fbc->lock); > > @@ -876,21 +876,21 @@ void intel_fbc_pre_update(struct intel_crtc *crtc, > > fbc->flip_pending = true; > > deactivate: > > - intel_fbc_deactivate(dev_priv, reason); > > + intel_fbc_deactivate(i915, reason); > > unlock: > > mutex_unlock(&fbc->lock); > > } > > /** > > * __intel_fbc_disable - disable FBC > > - * @dev_priv: i915 device instance > > + * @i915: i915 device instance > > * > > * This is the low level function that actually disables FBC. Callers should > > * grab the FBC lock. > > */ > > -static void __intel_fbc_disable(struct drm_i915_private *dev_priv) > > +static void __intel_fbc_disable(struct drm_i915_private *i915) > > { > > - struct intel_fbc *fbc = &dev_priv->fbc; > > + struct intel_fbc *fbc = &i915->fbc; > > struct intel_crtc *crtc = fbc->crtc; > > WARN_ON(!mutex_is_locked(&fbc->lock)); > > @@ -899,7 +899,7 @@ static void __intel_fbc_disable(struct drm_i915_private *dev_priv) > > DRM_DEBUG_KMS("Disabling FBC on pipe %c\n", pipe_name(crtc->pipe)); > > - __intel_fbc_cleanup_cfb(dev_priv); > > + __intel_fbc_cleanup_cfb(i915); > > fbc->enabled = false; > > fbc->crtc = NULL; > > @@ -907,8 +907,8 @@ static void __intel_fbc_disable(struct drm_i915_private *dev_priv) > > static void __intel_fbc_post_update(struct intel_crtc *crtc) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > - struct intel_fbc *fbc = &dev_priv->fbc; > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > + struct intel_fbc *fbc = &i915->fbc; > > WARN_ON(!mutex_is_locked(&fbc->lock)); > > @@ -919,8 +919,8 @@ static void __intel_fbc_post_update(struct intel_crtc *crtc) > > WARN_ON(fbc->active); > > if (!i915_modparams.enable_fbc) { > > - intel_fbc_deactivate(dev_priv, "disabled at runtime per module param"); > > - __intel_fbc_disable(dev_priv); > > + intel_fbc_deactivate(i915, "disabled at runtime per module param"); > > + __intel_fbc_disable(i915); > > return; > > } > > @@ -931,18 +931,18 @@ static void __intel_fbc_post_update(struct intel_crtc *crtc) > > return; > > if (!fbc->busy_bits) { > > - intel_fbc_deactivate(dev_priv, "FBC enabled (active or scheduled)"); > > - intel_fbc_hw_activate(dev_priv); > > + intel_fbc_deactivate(i915, "FBC enabled (active or scheduled)"); > > + intel_fbc_hw_activate(i915); > > } else > > - intel_fbc_deactivate(dev_priv, "frontbuffer write"); > > + intel_fbc_deactivate(i915, "frontbuffer write"); > > } > > void intel_fbc_post_update(struct intel_crtc *crtc) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > - struct intel_fbc *fbc = &dev_priv->fbc; > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > + struct intel_fbc *fbc = &i915->fbc; > > - if (!fbc_supported(dev_priv)) > > + if (!fbc_supported(i915)) > > return; > > mutex_lock(&fbc->lock); > > @@ -958,13 +958,13 @@ static unsigned int intel_fbc_get_frontbuffer_bit(struct intel_fbc *fbc) > > return fbc->possible_framebuffer_bits; > > } > > -void intel_fbc_invalidate(struct drm_i915_private *dev_priv, > > +void intel_fbc_invalidate(struct drm_i915_private *i915, > > unsigned int frontbuffer_bits, > > enum fb_op_origin origin) > > { > > - struct intel_fbc *fbc = &dev_priv->fbc; > > + struct intel_fbc *fbc = &i915->fbc; > > - if (!fbc_supported(dev_priv)) > > + if (!fbc_supported(i915)) > > return; > > if (origin == ORIGIN_GTT || origin == ORIGIN_FLIP) > > @@ -975,17 +975,17 @@ void intel_fbc_invalidate(struct drm_i915_private *dev_priv, > > fbc->busy_bits |= intel_fbc_get_frontbuffer_bit(fbc) & frontbuffer_bits; > > if (fbc->enabled && fbc->busy_bits) > > - intel_fbc_deactivate(dev_priv, "frontbuffer write"); > > + intel_fbc_deactivate(i915, "frontbuffer write"); > > mutex_unlock(&fbc->lock); > > } > > -void intel_fbc_flush(struct drm_i915_private *dev_priv, > > +void intel_fbc_flush(struct drm_i915_private *i915, > > unsigned int frontbuffer_bits, enum fb_op_origin origin) > > { > > - struct intel_fbc *fbc = &dev_priv->fbc; > > + struct intel_fbc *fbc = &i915->fbc; > > - if (!fbc_supported(dev_priv)) > > + if (!fbc_supported(i915)) > > return; > > mutex_lock(&fbc->lock); > > @@ -998,7 +998,7 @@ void intel_fbc_flush(struct drm_i915_private *dev_priv, > > if (!fbc->busy_bits && fbc->enabled && > > (frontbuffer_bits & intel_fbc_get_frontbuffer_bit(fbc))) { > > if (fbc->active) > > - intel_fbc_recompress(dev_priv); > > + intel_fbc_recompress(i915); > > else if (!fbc->flip_pending) > > __intel_fbc_post_update(fbc->crtc); > > } > > @@ -1009,7 +1009,7 @@ void intel_fbc_flush(struct drm_i915_private *dev_priv, > > /** > > * intel_fbc_choose_crtc - select a CRTC to enable FBC on > > - * @dev_priv: i915 device instance > > + * @i915: i915 device instance > > * @state: the atomic state structure > > * > > * This function looks at the proposed state for CRTCs and planes, then chooses > > @@ -1017,12 +1017,12 @@ void intel_fbc_flush(struct drm_i915_private *dev_priv, > > * true. > > * > > * Later, intel_fbc_enable is going to look for state->enable_fbc and then maybe > > - * enable FBC for the chosen CRTC. If it does, it will set dev_priv->fbc.crtc. > > + * enable FBC for the chosen CRTC. If it does, it will set i915->fbc.crtc. > > */ > > -void intel_fbc_choose_crtc(struct drm_i915_private *dev_priv, > > +void intel_fbc_choose_crtc(struct drm_i915_private *i915, > > struct intel_atomic_state *state) > > { > > - struct intel_fbc *fbc = &dev_priv->fbc; > > + struct intel_fbc *fbc = &i915->fbc; > > struct intel_plane *plane; > > struct intel_plane_state *plane_state; > > bool crtc_chosen = false; > > @@ -1035,7 +1035,7 @@ void intel_fbc_choose_crtc(struct drm_i915_private *dev_priv, > > !intel_atomic_get_new_crtc_state(state, fbc->crtc)) > > goto out; > > - if (!intel_fbc_can_enable(dev_priv)) > > + if (!intel_fbc_can_enable(i915)) > > goto out; > > /* Simply choose the first CRTC that is compatible and has a visible > > @@ -1081,10 +1081,10 @@ void intel_fbc_enable(struct intel_crtc *crtc, > > struct intel_crtc_state *crtc_state, > > struct intel_plane_state *plane_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > - struct intel_fbc *fbc = &dev_priv->fbc; > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > + struct intel_fbc *fbc = &i915->fbc; > > - if (!fbc_supported(dev_priv)) > > + if (!fbc_supported(i915)) > > return; > > mutex_lock(&fbc->lock); > > @@ -1127,44 +1127,44 @@ void intel_fbc_enable(struct intel_crtc *crtc, > > */ > > void intel_fbc_disable(struct intel_crtc *crtc) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > - struct intel_fbc *fbc = &dev_priv->fbc; > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > + struct intel_fbc *fbc = &i915->fbc; > > - if (!fbc_supported(dev_priv)) > > + if (!fbc_supported(i915)) > > return; > > mutex_lock(&fbc->lock); > > if (fbc->crtc == crtc) > > - __intel_fbc_disable(dev_priv); > > + __intel_fbc_disable(i915); > > mutex_unlock(&fbc->lock); > > } > > /** > > * intel_fbc_global_disable - globally disable FBC > > - * @dev_priv: i915 device instance > > + * @i915: i915 device instance > > * > > * This function disables FBC regardless of which CRTC is associated with it. > > */ > > -void intel_fbc_global_disable(struct drm_i915_private *dev_priv) > > +void intel_fbc_global_disable(struct drm_i915_private *i915) > > { > > - struct intel_fbc *fbc = &dev_priv->fbc; > > + struct intel_fbc *fbc = &i915->fbc; > > - if (!fbc_supported(dev_priv)) > > + if (!fbc_supported(i915)) > > return; > > mutex_lock(&fbc->lock); > > if (fbc->enabled) { > > WARN_ON(fbc->crtc->active); > > - __intel_fbc_disable(dev_priv); > > + __intel_fbc_disable(i915); > > } > > mutex_unlock(&fbc->lock); > > } > > static void intel_fbc_underrun_work_fn(struct work_struct *work) > > { > > - struct drm_i915_private *dev_priv = > > + struct drm_i915_private *i915 = > > container_of(work, struct drm_i915_private, fbc.underrun_work); > > - struct intel_fbc *fbc = &dev_priv->fbc; > > + struct intel_fbc *fbc = &i915->fbc; > > mutex_lock(&fbc->lock); > > @@ -1175,42 +1175,42 @@ static void intel_fbc_underrun_work_fn(struct work_struct *work) > > DRM_DEBUG_KMS("Disabling FBC due to FIFO underrun.\n"); > > fbc->underrun_detected = true; > > - intel_fbc_deactivate(dev_priv, "FIFO underrun"); > > + intel_fbc_deactivate(i915, "FIFO underrun"); > > out: > > mutex_unlock(&fbc->lock); > > } > > /* > > * intel_fbc_reset_underrun - reset FBC fifo underrun status. > > - * @dev_priv: i915 device instance > > + * @i915: i915 device instance > > * > > * See intel_fbc_handle_fifo_underrun_irq(). For automated testing we > > * want to re-enable FBC after an underrun to increase test coverage. > > */ > > -int intel_fbc_reset_underrun(struct drm_i915_private *dev_priv) > > +int intel_fbc_reset_underrun(struct drm_i915_private *i915) > > { > > int ret; > > - cancel_work_sync(&dev_priv->fbc.underrun_work); > > + cancel_work_sync(&i915->fbc.underrun_work); > > - ret = mutex_lock_interruptible(&dev_priv->fbc.lock); > > + ret = mutex_lock_interruptible(&i915->fbc.lock); > > if (ret) > > return ret; > > - if (dev_priv->fbc.underrun_detected) { > > + if (i915->fbc.underrun_detected) { > > DRM_DEBUG_KMS("Re-allowing FBC after fifo underrun\n"); > > - dev_priv->fbc.no_fbc_reason = "FIFO underrun cleared"; > > + i915->fbc.no_fbc_reason = "FIFO underrun cleared"; > > } > > - dev_priv->fbc.underrun_detected = false; > > - mutex_unlock(&dev_priv->fbc.lock); > > + i915->fbc.underrun_detected = false; > > + mutex_unlock(&i915->fbc.lock); > > return 0; > > } > > /** > > * intel_fbc_handle_fifo_underrun_irq - disable FBC when we get a FIFO underrun > > - * @dev_priv: i915 device instance > > + * @i915: i915 device instance > > * > > * Without FBC, most underruns are harmless and don't really cause too many > > * problems, except for an annoying message on dmesg. With FBC, underruns can > > @@ -1222,11 +1222,11 @@ int intel_fbc_reset_underrun(struct drm_i915_private *dev_priv) > > * > > * This function is called from the IRQ handler. > > */ > > -void intel_fbc_handle_fifo_underrun_irq(struct drm_i915_private *dev_priv) > > +void intel_fbc_handle_fifo_underrun_irq(struct drm_i915_private *i915) > > { > > - struct intel_fbc *fbc = &dev_priv->fbc; > > + struct intel_fbc *fbc = &i915->fbc; > > - if (!fbc_supported(dev_priv)) > > + if (!fbc_supported(i915)) > > return; > > /* There's no guarantee that underrun_detected won't be set to true > > @@ -1243,24 +1243,24 @@ void intel_fbc_handle_fifo_underrun_irq(struct drm_i915_private *dev_priv) > > /** > > * intel_fbc_init_pipe_state - initialize FBC's CRTC visibility tracking > > - * @dev_priv: i915 device instance > > + * @i915: i915 device instance > > * > > * The FBC code needs to track CRTC visibility since the older platforms can't > > * have FBC enabled while multiple pipes are used. This function does the > > * initial setup at driver load to make sure FBC is matching the real hardware. > > */ > > -void intel_fbc_init_pipe_state(struct drm_i915_private *dev_priv) > > +void intel_fbc_init_pipe_state(struct drm_i915_private *i915) > > { > > struct intel_crtc *crtc; > > /* Don't even bother tracking anything if we don't need. */ > > - if (!no_fbc_on_multiple_pipes(dev_priv)) > > + if (!no_fbc_on_multiple_pipes(i915)) > > return; > > - for_each_intel_crtc(&dev_priv->drm, crtc) > > + for_each_intel_crtc(&i915->drm, crtc) > > if (intel_crtc_active(crtc) && > > crtc->base.primary->state->visible) > > - dev_priv->fbc.visible_pipes_mask |= (1 << crtc->pipe); > > + i915->fbc.visible_pipes_mask |= (1 << crtc->pipe); > > } > > /* > > @@ -1272,29 +1272,29 @@ void intel_fbc_init_pipe_state(struct drm_i915_private *dev_priv) > > * space to change the value during runtime without sanitizing it again. IGT > > * relies on being able to change i915.enable_fbc at runtime. > > */ > > -static int intel_sanitize_fbc_option(struct drm_i915_private *dev_priv) > > +static int intel_sanitize_fbc_option(struct drm_i915_private *i915) > > { > > if (i915_modparams.enable_fbc >= 0) > > return !!i915_modparams.enable_fbc; > > - if (!HAS_FBC(dev_priv)) > > + if (!HAS_FBC(i915)) > > return 0; > > /* https://bugs.freedesktop.org/show_bug.cgi?id=108085 */ > > - if (IS_GEMINILAKE(dev_priv)) > > + if (IS_GEMINILAKE(i915)) > > return 0; > > - if (IS_BROADWELL(dev_priv) || INTEL_GEN(dev_priv) >= 9) > > + if (IS_BROADWELL(i915) || INTEL_GEN(i915) >= 9) > > return 1; > > return 0; > > } > > -static bool need_fbc_vtd_wa(struct drm_i915_private *dev_priv) > > +static bool need_fbc_vtd_wa(struct drm_i915_private *i915) > > { > > /* WaFbcTurnOffFbcWhenHyperVisorIsUsed:skl,bxt */ > > if (intel_vtd_active() && > > - (IS_SKYLAKE(dev_priv) || IS_BROXTON(dev_priv))) { > > + (IS_SKYLAKE(i915) || IS_BROXTON(i915))) { > > DRM_INFO("Disabling framebuffer compression (FBC) to prevent screen flicker with VT-d enabled\n"); > > return true; > > } > > @@ -1304,38 +1304,38 @@ static bool need_fbc_vtd_wa(struct drm_i915_private *dev_priv) > > /** > > * intel_fbc_init - Initialize FBC > > - * @dev_priv: the i915 device > > + * @i915: the i915 device > > * > > * This function might be called during PM init process. > > */ > > -void intel_fbc_init(struct drm_i915_private *dev_priv) > > +void intel_fbc_init(struct drm_i915_private *i915) > > { > > - struct intel_fbc *fbc = &dev_priv->fbc; > > + struct intel_fbc *fbc = &i915->fbc; > > INIT_WORK(&fbc->underrun_work, intel_fbc_underrun_work_fn); > > mutex_init(&fbc->lock); > > fbc->enabled = false; > > fbc->active = false; > > - if (need_fbc_vtd_wa(dev_priv)) > > - mkwrite_device_info(dev_priv)->display.has_fbc = false; > > + if (need_fbc_vtd_wa(i915)) > > + mkwrite_device_info(i915)->display.has_fbc = false; > > - i915_modparams.enable_fbc = intel_sanitize_fbc_option(dev_priv); > > + i915_modparams.enable_fbc = intel_sanitize_fbc_option(i915); > > DRM_DEBUG_KMS("Sanitized enable_fbc value: %d\n", > > i915_modparams.enable_fbc); > > - if (!HAS_FBC(dev_priv)) { > > + if (!HAS_FBC(i915)) { > > fbc->no_fbc_reason = "unsupported by this chipset"; > > return; > > } > > /* This value was pulled out of someone's hat */ > > - if (INTEL_GEN(dev_priv) <= 4 && !IS_GM45(dev_priv)) > > + if (INTEL_GEN(i915) <= 4 && !IS_GM45(i915)) > > I915_WRITE(FBC_CONTROL, 500 << FBC_CTL_INTERVAL_SHIFT); > > /* We still don't have any sort of hardware state readout for FBC, so > > * deactivate it in case the BIOS activated it to make sure software > > * matches the hardware state. */ > > - if (intel_fbc_hw_is_active(dev_priv)) > > - intel_fbc_hw_deactivate(dev_priv); > > + if (intel_fbc_hw_is_active(i915)) > > + intel_fbc_hw_deactivate(i915); > > } > > diff --git a/drivers/gpu/drm/i915/intel_fbc.h b/drivers/gpu/drm/i915/intel_fbc.h > > index 50272eda8d43..b7dc6cb43497 100644 > > --- a/drivers/gpu/drm/i915/intel_fbc.h > > +++ b/drivers/gpu/drm/i915/intel_fbc.h > > @@ -16,27 +16,27 @@ struct intel_crtc; > > struct intel_crtc_state; > > struct intel_plane_state; > > -void intel_fbc_choose_crtc(struct drm_i915_private *dev_priv, > > +void intel_fbc_choose_crtc(struct drm_i915_private *i915, > > struct intel_atomic_state *state); > > -bool intel_fbc_is_active(struct drm_i915_private *dev_priv); > > +bool intel_fbc_is_active(struct drm_i915_private *i915); > > void intel_fbc_pre_update(struct intel_crtc *crtc, > > struct intel_crtc_state *crtc_state, > > struct intel_plane_state *plane_state); > > void intel_fbc_post_update(struct intel_crtc *crtc); > > -void intel_fbc_init(struct drm_i915_private *dev_priv); > > -void intel_fbc_init_pipe_state(struct drm_i915_private *dev_priv); > > +void intel_fbc_init(struct drm_i915_private *i915); > > +void intel_fbc_init_pipe_state(struct drm_i915_private *i915); > > void intel_fbc_enable(struct intel_crtc *crtc, > > struct intel_crtc_state *crtc_state, > > struct intel_plane_state *plane_state); > > void intel_fbc_disable(struct intel_crtc *crtc); > > -void intel_fbc_global_disable(struct drm_i915_private *dev_priv); > > -void intel_fbc_invalidate(struct drm_i915_private *dev_priv, > > +void intel_fbc_global_disable(struct drm_i915_private *i915); > > +void intel_fbc_invalidate(struct drm_i915_private *i915, > > unsigned int frontbuffer_bits, > > enum fb_op_origin origin); > > -void intel_fbc_flush(struct drm_i915_private *dev_priv, > > +void intel_fbc_flush(struct drm_i915_private *i915, > > unsigned int frontbuffer_bits, enum fb_op_origin origin); > > -void intel_fbc_cleanup_cfb(struct drm_i915_private *dev_priv); > > -void intel_fbc_handle_fifo_underrun_irq(struct drm_i915_private *dev_priv); > > -int intel_fbc_reset_underrun(struct drm_i915_private *dev_priv); > > +void intel_fbc_cleanup_cfb(struct drm_i915_private *i915); > > +void intel_fbc_handle_fifo_underrun_irq(struct drm_i915_private *i915); > > +int intel_fbc_reset_underrun(struct drm_i915_private *i915); > > #endif /* __INTEL_FBC_H__ */ > > diff --git a/drivers/gpu/drm/i915/intel_fbdev.c b/drivers/gpu/drm/i915/intel_fbdev.c > > index 0d3a6fa674e6..93dee9ff5a58 100644 > > --- a/drivers/gpu/drm/i915/intel_fbdev.c > > +++ b/drivers/gpu/drm/i915/intel_fbdev.c > > @@ -117,7 +117,7 @@ static int intelfb_alloc(struct drm_fb_helper *helper, > > container_of(helper, struct intel_fbdev, helper); > > struct drm_framebuffer *fb; > > struct drm_device *dev = helper->dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > struct drm_mode_fb_cmd2 mode_cmd = {}; > > struct drm_i915_gem_object *obj; > > int size, ret; > > @@ -141,10 +141,10 @@ static int intelfb_alloc(struct drm_fb_helper *helper, > > * important and we should probably use that space with FBC or other > > * features. */ > > obj = NULL; > > - if (size * 2 < dev_priv->stolen_usable_size) > > - obj = i915_gem_object_create_stolen(dev_priv, size); > > + if (size * 2 < i915->stolen_usable_size) > > + obj = i915_gem_object_create_stolen(i915, size); > > if (obj == NULL) > > - obj = i915_gem_object_create_shmem(dev_priv, size); > > + obj = i915_gem_object_create_shmem(i915, size); > > if (IS_ERR(obj)) { > > DRM_ERROR("failed to allocate framebuffer\n"); > > ret = PTR_ERR(obj); > > @@ -174,9 +174,9 @@ static int intelfb_create(struct drm_fb_helper *helper, > > container_of(helper, struct intel_fbdev, helper); > > struct intel_framebuffer *intel_fb = ifbdev->fb; > > struct drm_device *dev = helper->dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > - struct pci_dev *pdev = dev_priv->drm.pdev; > > - struct i915_ggtt *ggtt = &dev_priv->ggtt; > > + struct drm_i915_private *i915 = to_i915(dev); > > + struct pci_dev *pdev = i915->drm.pdev; > > + struct i915_ggtt *ggtt = &i915->ggtt; > > const struct i915_ggtt_view view = { > > .type = I915_GGTT_VIEW_NORMAL, > > }; > > @@ -213,7 +213,7 @@ static int intelfb_create(struct drm_fb_helper *helper, > > } > > mutex_lock(&dev->struct_mutex); > > - wakeref = intel_runtime_pm_get(dev_priv); > > + wakeref = intel_runtime_pm_get(i915); > > /* Pin the GGTT vma for our access via info->screen_base. > > * This also validates that any existing fb inherited from the > > @@ -272,7 +272,7 @@ static int intelfb_create(struct drm_fb_helper *helper, > > ifbdev->vma = vma; > > ifbdev->vma_flags = flags; > > - intel_runtime_pm_put(dev_priv, wakeref); > > + intel_runtime_pm_put(i915, wakeref); > > mutex_unlock(&dev->struct_mutex); > > vga_switcheroo_client_fb_set(pdev, info); > > return 0; > > @@ -280,7 +280,7 @@ static int intelfb_create(struct drm_fb_helper *helper, > > out_unpin: > > intel_unpin_fb_vma(vma, flags); > > out_unlock: > > - intel_runtime_pm_put(dev_priv, wakeref); > > + intel_runtime_pm_put(i915, wakeref); > > mutex_unlock(&dev->struct_mutex); > > return ret; > > } > > @@ -447,11 +447,11 @@ static void intel_fbdev_suspend_worker(struct work_struct *work) > > int intel_fbdev_init(struct drm_device *dev) > > { > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > struct intel_fbdev *ifbdev; > > int ret; > > - if (WARN_ON(!HAS_DISPLAY(dev_priv))) > > + if (WARN_ON(!HAS_DISPLAY(i915))) > > return -ENODEV; > > ifbdev = kzalloc(sizeof(struct intel_fbdev), GFP_KERNEL); > > @@ -470,8 +470,8 @@ int intel_fbdev_init(struct drm_device *dev) > > return ret; > > } > > - dev_priv->fbdev = ifbdev; > > - INIT_WORK(&dev_priv->fbdev_suspend_work, intel_fbdev_suspend_worker); > > + i915->fbdev = ifbdev; > > + INIT_WORK(&i915->fbdev_suspend_work, intel_fbdev_suspend_worker); > > drm_fb_helper_single_add_all_connectors(&ifbdev->helper); > > @@ -508,23 +508,23 @@ static void intel_fbdev_sync(struct intel_fbdev *ifbdev) > > ifbdev->cookie = 0; > > } > > -void intel_fbdev_unregister(struct drm_i915_private *dev_priv) > > +void intel_fbdev_unregister(struct drm_i915_private *i915) > > { > > - struct intel_fbdev *ifbdev = dev_priv->fbdev; > > + struct intel_fbdev *ifbdev = i915->fbdev; > > if (!ifbdev) > > return; > > - cancel_work_sync(&dev_priv->fbdev_suspend_work); > > + cancel_work_sync(&i915->fbdev_suspend_work); > > if (!current_is_async()) > > intel_fbdev_sync(ifbdev); > > drm_fb_helper_unregister_fbi(&ifbdev->helper); > > } > > -void intel_fbdev_fini(struct drm_i915_private *dev_priv) > > +void intel_fbdev_fini(struct drm_i915_private *i915) > > { > > - struct intel_fbdev *ifbdev = fetch_and_zero(&dev_priv->fbdev); > > + struct intel_fbdev *ifbdev = fetch_and_zero(&i915->fbdev); > > if (!ifbdev) > > return; > > @@ -554,8 +554,8 @@ static void intel_fbdev_hpd_set_suspend(struct intel_fbdev *ifbdev, int state) > > void intel_fbdev_set_suspend(struct drm_device *dev, int state, bool synchronous) > > { > > - struct drm_i915_private *dev_priv = to_i915(dev); > > - struct intel_fbdev *ifbdev = dev_priv->fbdev; > > + struct drm_i915_private *i915 = to_i915(dev); > > + struct intel_fbdev *ifbdev = i915->fbdev; > > struct fb_info *info; > > if (!ifbdev || !ifbdev->vma) > > @@ -572,7 +572,7 @@ void intel_fbdev_set_suspend(struct drm_device *dev, int state, bool synchronous > > * ourselves, so only flush outstanding work upon suspend! > > */ > > if (state != FBINFO_STATE_RUNNING) > > - flush_work(&dev_priv->fbdev_suspend_work); > > + flush_work(&i915->fbdev_suspend_work); > > console_lock(); > > } else { > > @@ -586,7 +586,7 @@ void intel_fbdev_set_suspend(struct drm_device *dev, int state, bool synchronous > > /* Don't block our own workqueue as this can > > * be run in parallel with other i915.ko tasks. > > */ > > - schedule_work(&dev_priv->fbdev_suspend_work); > > + schedule_work(&i915->fbdev_suspend_work); > > return; > > } > > } > > diff --git a/drivers/gpu/drm/i915/intel_fbdev.h b/drivers/gpu/drm/i915/intel_fbdev.h > > index de7c84250eb5..4d0c8c851a5e 100644 > > --- a/drivers/gpu/drm/i915/intel_fbdev.h > > +++ b/drivers/gpu/drm/i915/intel_fbdev.h > > @@ -14,8 +14,8 @@ struct drm_i915_private; > > #ifdef CONFIG_DRM_FBDEV_EMULATION > > int intel_fbdev_init(struct drm_device *dev); > > void intel_fbdev_initial_config_async(struct drm_device *dev); > > -void intel_fbdev_unregister(struct drm_i915_private *dev_priv); > > -void intel_fbdev_fini(struct drm_i915_private *dev_priv); > > +void intel_fbdev_unregister(struct drm_i915_private *i915); > > +void intel_fbdev_fini(struct drm_i915_private *i915); > > void intel_fbdev_set_suspend(struct drm_device *dev, int state, bool synchronous); > > void intel_fbdev_output_poll_changed(struct drm_device *dev); > > void intel_fbdev_restore_mode(struct drm_device *dev); > > @@ -29,11 +29,11 @@ static inline void intel_fbdev_initial_config_async(struct drm_device *dev) > > { > > } > > -static inline void intel_fbdev_unregister(struct drm_i915_private *dev_priv) > > +static inline void intel_fbdev_unregister(struct drm_i915_private *i915) > > { > > } > > -static inline void intel_fbdev_fini(struct drm_i915_private *dev_priv) > > +static inline void intel_fbdev_fini(struct drm_i915_private *i915) > > { > > } > > diff --git a/drivers/gpu/drm/i915/intel_fifo_underrun.c b/drivers/gpu/drm/i915/intel_fifo_underrun.c > > index 8545ad32bb50..08f092addae2 100644 > > --- a/drivers/gpu/drm/i915/intel_fifo_underrun.c > > +++ b/drivers/gpu/drm/i915/intel_fifo_underrun.c > > @@ -52,14 +52,14 @@ > > static bool ivb_can_enable_err_int(struct drm_device *dev) > > { > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > struct intel_crtc *crtc; > > enum pipe pipe; > > - lockdep_assert_held(&dev_priv->irq_lock); > > + lockdep_assert_held(&i915->irq_lock); > > - for_each_pipe(dev_priv, pipe) { > > - crtc = intel_get_crtc_for_pipe(dev_priv, pipe); > > + for_each_pipe(i915, pipe) { > > + crtc = intel_get_crtc_for_pipe(i915, pipe); > > if (crtc->cpu_fifo_underrun_disabled) > > return false; > > @@ -70,14 +70,14 @@ static bool ivb_can_enable_err_int(struct drm_device *dev) > > static bool cpt_can_enable_serr_int(struct drm_device *dev) > > { > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > enum pipe pipe; > > struct intel_crtc *crtc; > > - lockdep_assert_held(&dev_priv->irq_lock); > > + lockdep_assert_held(&i915->irq_lock); > > - for_each_pipe(dev_priv, pipe) { > > - crtc = intel_get_crtc_for_pipe(dev_priv, pipe); > > + for_each_pipe(i915, pipe) { > > + crtc = intel_get_crtc_for_pipe(i915, pipe); > > if (crtc->pch_fifo_underrun_disabled) > > return false; > > @@ -88,20 +88,20 @@ static bool cpt_can_enable_serr_int(struct drm_device *dev) > > static void i9xx_check_fifo_underruns(struct intel_crtc *crtc) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > i915_reg_t reg = PIPESTAT(crtc->pipe); > > u32 enable_mask; > > - lockdep_assert_held(&dev_priv->irq_lock); > > + lockdep_assert_held(&i915->irq_lock); > > if ((I915_READ(reg) & PIPE_FIFO_UNDERRUN_STATUS) == 0) > > return; > > - enable_mask = i915_pipestat_enable_mask(dev_priv, crtc->pipe); > > + enable_mask = i915_pipestat_enable_mask(i915, crtc->pipe); > > I915_WRITE(reg, enable_mask | PIPE_FIFO_UNDERRUN_STATUS); > > POSTING_READ(reg); > > - trace_intel_cpu_fifo_underrun(dev_priv, crtc->pipe); > > + trace_intel_cpu_fifo_underrun(i915, crtc->pipe); > > DRM_ERROR("pipe %c underrun\n", pipe_name(crtc->pipe)); > > } > > @@ -109,13 +109,13 @@ static void i9xx_set_fifo_underrun_reporting(struct drm_device *dev, > > enum pipe pipe, > > bool enable, bool old) > > { > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > i915_reg_t reg = PIPESTAT(pipe); > > - lockdep_assert_held(&dev_priv->irq_lock); > > + lockdep_assert_held(&i915->irq_lock); > > if (enable) { > > - u32 enable_mask = i915_pipestat_enable_mask(dev_priv, pipe); > > + u32 enable_mask = i915_pipestat_enable_mask(i915, pipe); > > I915_WRITE(reg, enable_mask | PIPE_FIFO_UNDERRUN_STATUS); > > POSTING_READ(reg); > > @@ -128,23 +128,23 @@ static void i9xx_set_fifo_underrun_reporting(struct drm_device *dev, > > static void ironlake_set_fifo_underrun_reporting(struct drm_device *dev, > > enum pipe pipe, bool enable) > > { > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > u32 bit = (pipe == PIPE_A) ? > > DE_PIPEA_FIFO_UNDERRUN : DE_PIPEB_FIFO_UNDERRUN; > > if (enable) > > - ilk_enable_display_irq(dev_priv, bit); > > + ilk_enable_display_irq(i915, bit); > > else > > - ilk_disable_display_irq(dev_priv, bit); > > + ilk_disable_display_irq(i915, bit); > > } > > static void ivybridge_check_fifo_underruns(struct intel_crtc *crtc) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > enum pipe pipe = crtc->pipe; > > u32 err_int = I915_READ(GEN7_ERR_INT); > > - lockdep_assert_held(&dev_priv->irq_lock); > > + lockdep_assert_held(&i915->irq_lock); > > if ((err_int & ERR_INT_FIFO_UNDERRUN(pipe)) == 0) > > return; > > @@ -152,7 +152,7 @@ static void ivybridge_check_fifo_underruns(struct intel_crtc *crtc) > > I915_WRITE(GEN7_ERR_INT, ERR_INT_FIFO_UNDERRUN(pipe)); > > POSTING_READ(GEN7_ERR_INT); > > - trace_intel_cpu_fifo_underrun(dev_priv, pipe); > > + trace_intel_cpu_fifo_underrun(i915, pipe); > > DRM_ERROR("fifo underrun on pipe %c\n", pipe_name(pipe)); > > } > > @@ -160,16 +160,16 @@ static void ivybridge_set_fifo_underrun_reporting(struct drm_device *dev, > > enum pipe pipe, > > bool enable, bool old) > > { > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > if (enable) { > > I915_WRITE(GEN7_ERR_INT, ERR_INT_FIFO_UNDERRUN(pipe)); > > if (!ivb_can_enable_err_int(dev)) > > return; > > - ilk_enable_display_irq(dev_priv, DE_ERR_INT_IVB); > > + ilk_enable_display_irq(i915, DE_ERR_INT_IVB); > > } else { > > - ilk_disable_display_irq(dev_priv, DE_ERR_INT_IVB); > > + ilk_disable_display_irq(i915, DE_ERR_INT_IVB); > > if (old && > > I915_READ(GEN7_ERR_INT) & ERR_INT_FIFO_UNDERRUN(pipe)) { > > @@ -182,35 +182,35 @@ static void ivybridge_set_fifo_underrun_reporting(struct drm_device *dev, > > static void broadwell_set_fifo_underrun_reporting(struct drm_device *dev, > > enum pipe pipe, bool enable) > > { > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > if (enable) > > - bdw_enable_pipe_irq(dev_priv, pipe, GEN8_PIPE_FIFO_UNDERRUN); > > + bdw_enable_pipe_irq(i915, pipe, GEN8_PIPE_FIFO_UNDERRUN); > > else > > - bdw_disable_pipe_irq(dev_priv, pipe, GEN8_PIPE_FIFO_UNDERRUN); > > + bdw_disable_pipe_irq(i915, pipe, GEN8_PIPE_FIFO_UNDERRUN); > > } > > static void ibx_set_fifo_underrun_reporting(struct drm_device *dev, > > enum pipe pch_transcoder, > > bool enable) > > { > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > u32 bit = (pch_transcoder == PIPE_A) ? > > SDE_TRANSA_FIFO_UNDER : SDE_TRANSB_FIFO_UNDER; > > if (enable) > > - ibx_enable_display_interrupt(dev_priv, bit); > > + ibx_enable_display_interrupt(i915, bit); > > else > > - ibx_disable_display_interrupt(dev_priv, bit); > > + ibx_disable_display_interrupt(i915, bit); > > } > > static void cpt_check_pch_fifo_underruns(struct intel_crtc *crtc) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > enum pipe pch_transcoder = crtc->pipe; > > u32 serr_int = I915_READ(SERR_INT); > > - lockdep_assert_held(&dev_priv->irq_lock); > > + lockdep_assert_held(&i915->irq_lock); > > if ((serr_int & SERR_INT_TRANS_FIFO_UNDERRUN(pch_transcoder)) == 0) > > return; > > @@ -218,7 +218,7 @@ static void cpt_check_pch_fifo_underruns(struct intel_crtc *crtc) > > I915_WRITE(SERR_INT, SERR_INT_TRANS_FIFO_UNDERRUN(pch_transcoder)); > > POSTING_READ(SERR_INT); > > - trace_intel_pch_fifo_underrun(dev_priv, pch_transcoder); > > + trace_intel_pch_fifo_underrun(i915, pch_transcoder); > > DRM_ERROR("pch fifo underrun on pch transcoder %c\n", > > pipe_name(pch_transcoder)); > > } > > @@ -227,7 +227,7 @@ static void cpt_set_fifo_underrun_reporting(struct drm_device *dev, > > enum pipe pch_transcoder, > > bool enable, bool old) > > { > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > if (enable) { > > I915_WRITE(SERR_INT, > > @@ -236,9 +236,9 @@ static void cpt_set_fifo_underrun_reporting(struct drm_device *dev, > > if (!cpt_can_enable_serr_int(dev)) > > return; > > - ibx_enable_display_interrupt(dev_priv, SDE_ERROR_CPT); > > + ibx_enable_display_interrupt(i915, SDE_ERROR_CPT); > > } else { > > - ibx_disable_display_interrupt(dev_priv, SDE_ERROR_CPT); > > + ibx_disable_display_interrupt(i915, SDE_ERROR_CPT); > > if (old && I915_READ(SERR_INT) & > > SERR_INT_TRANS_FIFO_UNDERRUN(pch_transcoder)) { > > @@ -251,22 +251,22 @@ static void cpt_set_fifo_underrun_reporting(struct drm_device *dev, > > static bool __intel_set_cpu_fifo_underrun_reporting(struct drm_device *dev, > > enum pipe pipe, bool enable) > > { > > - struct drm_i915_private *dev_priv = to_i915(dev); > > - struct intel_crtc *crtc = intel_get_crtc_for_pipe(dev_priv, pipe); > > + struct drm_i915_private *i915 = to_i915(dev); > > + struct intel_crtc *crtc = intel_get_crtc_for_pipe(i915, pipe); > > bool old; > > - lockdep_assert_held(&dev_priv->irq_lock); > > + lockdep_assert_held(&i915->irq_lock); > > old = !crtc->cpu_fifo_underrun_disabled; > > crtc->cpu_fifo_underrun_disabled = !enable; > > - if (HAS_GMCH(dev_priv)) > > + if (HAS_GMCH(i915)) > > i9xx_set_fifo_underrun_reporting(dev, pipe, enable, old); > > - else if (IS_GEN_RANGE(dev_priv, 5, 6)) > > + else if (IS_GEN_RANGE(i915, 5, 6)) > > ironlake_set_fifo_underrun_reporting(dev, pipe, enable); > > - else if (IS_GEN(dev_priv, 7)) > > + else if (IS_GEN(i915, 7)) > > ivybridge_set_fifo_underrun_reporting(dev, pipe, enable, old); > > - else if (INTEL_GEN(dev_priv) >= 8) > > + else if (INTEL_GEN(i915) >= 8) > > broadwell_set_fifo_underrun_reporting(dev, pipe, enable); > > return old; > > @@ -274,7 +274,7 @@ static bool __intel_set_cpu_fifo_underrun_reporting(struct drm_device *dev, > > /** > > * intel_set_cpu_fifo_underrun_reporting - set cpu fifo underrrun reporting state > > - * @dev_priv: i915 device instance > > + * @i915: i915 device instance > > * @pipe: (CPU) pipe to set state for > > * @enable: whether underruns should be reported or not > > * > > @@ -288,23 +288,23 @@ static bool __intel_set_cpu_fifo_underrun_reporting(struct drm_device *dev, > > * > > * Returns the previous state of underrun reporting. > > */ > > -bool intel_set_cpu_fifo_underrun_reporting(struct drm_i915_private *dev_priv, > > +bool intel_set_cpu_fifo_underrun_reporting(struct drm_i915_private *i915, > > enum pipe pipe, bool enable) > > { > > unsigned long flags; > > bool ret; > > - spin_lock_irqsave(&dev_priv->irq_lock, flags); > > - ret = __intel_set_cpu_fifo_underrun_reporting(&dev_priv->drm, pipe, > > + spin_lock_irqsave(&i915->irq_lock, flags); > > + ret = __intel_set_cpu_fifo_underrun_reporting(&i915->drm, pipe, > > enable); > > - spin_unlock_irqrestore(&dev_priv->irq_lock, flags); > > + spin_unlock_irqrestore(&i915->irq_lock, flags); > > return ret; > > } > > /** > > * intel_set_pch_fifo_underrun_reporting - set PCH fifo underrun reporting state > > - * @dev_priv: i915 device instance > > + * @i915: i915 device instance > > * @pch_transcoder: the PCH transcoder (same as pipe on IVB and older) > > * @enable: whether underruns should be reported or not > > * > > @@ -316,12 +316,12 @@ bool intel_set_cpu_fifo_underrun_reporting(struct drm_i915_private *dev_priv, > > * > > * Returns the previous state of underrun reporting. > > */ > > -bool intel_set_pch_fifo_underrun_reporting(struct drm_i915_private *dev_priv, > > +bool intel_set_pch_fifo_underrun_reporting(struct drm_i915_private *i915, > > enum pipe pch_transcoder, > > bool enable) > > { > > struct intel_crtc *crtc = > > - intel_get_crtc_for_pipe(dev_priv, pch_transcoder); > > + intel_get_crtc_for_pipe(i915, pch_transcoder); > > unsigned long flags; > > bool old; > > @@ -334,71 +334,71 @@ bool intel_set_pch_fifo_underrun_reporting(struct drm_i915_private *dev_priv, > > * crtc on LPT won't cause issues. > > */ > > - spin_lock_irqsave(&dev_priv->irq_lock, flags); > > + spin_lock_irqsave(&i915->irq_lock, flags); > > old = !crtc->pch_fifo_underrun_disabled; > > crtc->pch_fifo_underrun_disabled = !enable; > > - if (HAS_PCH_IBX(dev_priv)) > > - ibx_set_fifo_underrun_reporting(&dev_priv->drm, > > + if (HAS_PCH_IBX(i915)) > > + ibx_set_fifo_underrun_reporting(&i915->drm, > > pch_transcoder, > > enable); > > else > > - cpt_set_fifo_underrun_reporting(&dev_priv->drm, > > + cpt_set_fifo_underrun_reporting(&i915->drm, > > pch_transcoder, > > enable, old); > > - spin_unlock_irqrestore(&dev_priv->irq_lock, flags); > > + spin_unlock_irqrestore(&i915->irq_lock, flags); > > return old; > > } > > /** > > * intel_cpu_fifo_underrun_irq_handler - handle CPU fifo underrun interrupt > > - * @dev_priv: i915 device instance > > + * @i915: i915 device instance > > * @pipe: (CPU) pipe to set state for > > * > > * This handles a CPU fifo underrun interrupt, generating an underrun warning > > * into dmesg if underrun reporting is enabled and then disables the underrun > > * interrupt to avoid an irq storm. > > */ > > -void intel_cpu_fifo_underrun_irq_handler(struct drm_i915_private *dev_priv, > > +void intel_cpu_fifo_underrun_irq_handler(struct drm_i915_private *i915, > > enum pipe pipe) > > { > > - struct intel_crtc *crtc = intel_get_crtc_for_pipe(dev_priv, pipe); > > + struct intel_crtc *crtc = intel_get_crtc_for_pipe(i915, pipe); > > /* We may be called too early in init, thanks BIOS! */ > > if (crtc == NULL) > > return; > > /* GMCH can't disable fifo underruns, filter them. */ > > - if (HAS_GMCH(dev_priv) && > > + if (HAS_GMCH(i915) && > > crtc->cpu_fifo_underrun_disabled) > > return; > > - if (intel_set_cpu_fifo_underrun_reporting(dev_priv, pipe, false)) { > > - trace_intel_cpu_fifo_underrun(dev_priv, pipe); > > + if (intel_set_cpu_fifo_underrun_reporting(i915, pipe, false)) { > > + trace_intel_cpu_fifo_underrun(i915, pipe); > > DRM_ERROR("CPU pipe %c FIFO underrun\n", > > pipe_name(pipe)); > > } > > - intel_fbc_handle_fifo_underrun_irq(dev_priv); > > + intel_fbc_handle_fifo_underrun_irq(i915); > > } > > /** > > * intel_pch_fifo_underrun_irq_handler - handle PCH fifo underrun interrupt > > - * @dev_priv: i915 device instance > > + * @i915: i915 device instance > > * @pch_transcoder: the PCH transcoder (same as pipe on IVB and older) > > * > > * This handles a PCH fifo underrun interrupt, generating an underrun warning > > * into dmesg if underrun reporting is enabled and then disables the underrun > > * interrupt to avoid an irq storm. > > */ > > -void intel_pch_fifo_underrun_irq_handler(struct drm_i915_private *dev_priv, > > +void intel_pch_fifo_underrun_irq_handler(struct drm_i915_private *i915, > > enum pipe pch_transcoder) > > { > > - if (intel_set_pch_fifo_underrun_reporting(dev_priv, pch_transcoder, > > + if (intel_set_pch_fifo_underrun_reporting(i915, pch_transcoder, > > false)) { > > - trace_intel_pch_fifo_underrun(dev_priv, pch_transcoder); > > + trace_intel_pch_fifo_underrun(i915, pch_transcoder); > > DRM_ERROR("PCH transcoder %c FIFO underrun\n", > > pipe_name(pch_transcoder)); > > } > > @@ -406,53 +406,53 @@ void intel_pch_fifo_underrun_irq_handler(struct drm_i915_private *dev_priv, > > /** > > * intel_check_cpu_fifo_underruns - check for CPU fifo underruns immediately > > - * @dev_priv: i915 device instance > > + * @i915: i915 device instance > > * > > * Check for CPU fifo underruns immediately. Useful on IVB/HSW where the shared > > * error interrupt may have been disabled, and so CPU fifo underruns won't > > * necessarily raise an interrupt, and on GMCH platforms where underruns never > > * raise an interrupt. > > */ > > -void intel_check_cpu_fifo_underruns(struct drm_i915_private *dev_priv) > > +void intel_check_cpu_fifo_underruns(struct drm_i915_private *i915) > > { > > struct intel_crtc *crtc; > > - spin_lock_irq(&dev_priv->irq_lock); > > + spin_lock_irq(&i915->irq_lock); > > - for_each_intel_crtc(&dev_priv->drm, crtc) { > > + for_each_intel_crtc(&i915->drm, crtc) { > > if (crtc->cpu_fifo_underrun_disabled) > > continue; > > - if (HAS_GMCH(dev_priv)) > > + if (HAS_GMCH(i915)) > > i9xx_check_fifo_underruns(crtc); > > - else if (IS_GEN(dev_priv, 7)) > > + else if (IS_GEN(i915, 7)) > > ivybridge_check_fifo_underruns(crtc); > > } > > - spin_unlock_irq(&dev_priv->irq_lock); > > + spin_unlock_irq(&i915->irq_lock); > > } > > /** > > * intel_check_pch_fifo_underruns - check for PCH fifo underruns immediately > > - * @dev_priv: i915 device instance > > + * @i915: i915 device instance > > * > > * Check for PCH fifo underruns immediately. Useful on CPT/PPT where the shared > > * error interrupt may have been disabled, and so PCH fifo underruns won't > > * necessarily raise an interrupt. > > */ > > -void intel_check_pch_fifo_underruns(struct drm_i915_private *dev_priv) > > +void intel_check_pch_fifo_underruns(struct drm_i915_private *i915) > > { > > struct intel_crtc *crtc; > > - spin_lock_irq(&dev_priv->irq_lock); > > + spin_lock_irq(&i915->irq_lock); > > - for_each_intel_crtc(&dev_priv->drm, crtc) { > > + for_each_intel_crtc(&i915->drm, crtc) { > > if (crtc->pch_fifo_underrun_disabled) > > continue; > > - if (HAS_PCH_CPT(dev_priv)) > > + if (HAS_PCH_CPT(i915)) > > cpt_check_pch_fifo_underruns(crtc); > > } > > - spin_unlock_irq(&dev_priv->irq_lock); > > + spin_unlock_irq(&i915->irq_lock); > > } > > diff --git a/drivers/gpu/drm/i915/intel_fifo_underrun.h b/drivers/gpu/drm/i915/intel_fifo_underrun.h > > index e04f22ac1f49..290909a67de8 100644 > > --- a/drivers/gpu/drm/i915/intel_fifo_underrun.h > > +++ b/drivers/gpu/drm/i915/intel_fifo_underrun.h > > @@ -12,16 +12,16 @@ > > struct drm_i915_private; > > -bool intel_set_cpu_fifo_underrun_reporting(struct drm_i915_private *dev_priv, > > +bool intel_set_cpu_fifo_underrun_reporting(struct drm_i915_private *i915, > > enum pipe pipe, bool enable); > > -bool intel_set_pch_fifo_underrun_reporting(struct drm_i915_private *dev_priv, > > +bool intel_set_pch_fifo_underrun_reporting(struct drm_i915_private *i915, > > enum pipe pch_transcoder, > > bool enable); > > -void intel_cpu_fifo_underrun_irq_handler(struct drm_i915_private *dev_priv, > > +void intel_cpu_fifo_underrun_irq_handler(struct drm_i915_private *i915, > > enum pipe pipe); > > -void intel_pch_fifo_underrun_irq_handler(struct drm_i915_private *dev_priv, > > +void intel_pch_fifo_underrun_irq_handler(struct drm_i915_private *i915, > > enum pipe pch_transcoder); > > -void intel_check_cpu_fifo_underruns(struct drm_i915_private *dev_priv); > > -void intel_check_pch_fifo_underruns(struct drm_i915_private *dev_priv); > > +void intel_check_cpu_fifo_underruns(struct drm_i915_private *i915); > > +void intel_check_pch_fifo_underruns(struct drm_i915_private *i915); > > #endif /* __INTEL_FIFO_UNDERRUN_H__ */ > > diff --git a/drivers/gpu/drm/i915/intel_frontbuffer.c b/drivers/gpu/drm/i915/intel_frontbuffer.c > > index d6036b9ad16a..161139de3f28 100644 > > --- a/drivers/gpu/drm/i915/intel_frontbuffer.c > > +++ b/drivers/gpu/drm/i915/intel_frontbuffer.c > > @@ -66,24 +66,24 @@ void __intel_fb_obj_invalidate(struct drm_i915_gem_object *obj, > > enum fb_op_origin origin, > > unsigned int frontbuffer_bits) > > { > > - struct drm_i915_private *dev_priv = to_i915(obj->base.dev); > > + struct drm_i915_private *i915 = to_i915(obj->base.dev); > > if (origin == ORIGIN_CS) { > > - spin_lock(&dev_priv->fb_tracking.lock); > > - dev_priv->fb_tracking.busy_bits |= frontbuffer_bits; > > - dev_priv->fb_tracking.flip_bits &= ~frontbuffer_bits; > > - spin_unlock(&dev_priv->fb_tracking.lock); > > + spin_lock(&i915->fb_tracking.lock); > > + i915->fb_tracking.busy_bits |= frontbuffer_bits; > > + i915->fb_tracking.flip_bits &= ~frontbuffer_bits; > > + spin_unlock(&i915->fb_tracking.lock); > > } > > might_sleep(); > > - intel_psr_invalidate(dev_priv, frontbuffer_bits, origin); > > - intel_edp_drrs_invalidate(dev_priv, frontbuffer_bits); > > - intel_fbc_invalidate(dev_priv, frontbuffer_bits, origin); > > + intel_psr_invalidate(i915, frontbuffer_bits, origin); > > + intel_edp_drrs_invalidate(i915, frontbuffer_bits); > > + intel_fbc_invalidate(i915, frontbuffer_bits, origin); > > } > > /** > > * intel_frontbuffer_flush - flush frontbuffer > > - * @dev_priv: i915 device > > + * @i915: i915 device > > * @frontbuffer_bits: frontbuffer plane tracking bits > > * @origin: which operation caused the flush > > * > > @@ -93,45 +93,45 @@ void __intel_fb_obj_invalidate(struct drm_i915_gem_object *obj, > > * > > * Can be called without any locks held. > > */ > > -static void intel_frontbuffer_flush(struct drm_i915_private *dev_priv, > > +static void intel_frontbuffer_flush(struct drm_i915_private *i915, > > unsigned frontbuffer_bits, > > enum fb_op_origin origin) > > { > > /* Delay flushing when rings are still busy.*/ > > - spin_lock(&dev_priv->fb_tracking.lock); > > - frontbuffer_bits &= ~dev_priv->fb_tracking.busy_bits; > > - spin_unlock(&dev_priv->fb_tracking.lock); > > + spin_lock(&i915->fb_tracking.lock); > > + frontbuffer_bits &= ~i915->fb_tracking.busy_bits; > > + spin_unlock(&i915->fb_tracking.lock); > > if (!frontbuffer_bits) > > return; > > might_sleep(); > > - intel_edp_drrs_flush(dev_priv, frontbuffer_bits); > > - intel_psr_flush(dev_priv, frontbuffer_bits, origin); > > - intel_fbc_flush(dev_priv, frontbuffer_bits, origin); > > + intel_edp_drrs_flush(i915, frontbuffer_bits); > > + intel_psr_flush(i915, frontbuffer_bits, origin); > > + intel_fbc_flush(i915, frontbuffer_bits, origin); > > } > > void __intel_fb_obj_flush(struct drm_i915_gem_object *obj, > > enum fb_op_origin origin, > > unsigned int frontbuffer_bits) > > { > > - struct drm_i915_private *dev_priv = to_i915(obj->base.dev); > > + struct drm_i915_private *i915 = to_i915(obj->base.dev); > > if (origin == ORIGIN_CS) { > > - spin_lock(&dev_priv->fb_tracking.lock); > > + spin_lock(&i915->fb_tracking.lock); > > /* Filter out new bits since rendering started. */ > > - frontbuffer_bits &= dev_priv->fb_tracking.busy_bits; > > - dev_priv->fb_tracking.busy_bits &= ~frontbuffer_bits; > > - spin_unlock(&dev_priv->fb_tracking.lock); > > + frontbuffer_bits &= i915->fb_tracking.busy_bits; > > + i915->fb_tracking.busy_bits &= ~frontbuffer_bits; > > + spin_unlock(&i915->fb_tracking.lock); > > } > > if (frontbuffer_bits) > > - intel_frontbuffer_flush(dev_priv, frontbuffer_bits, origin); > > + intel_frontbuffer_flush(i915, frontbuffer_bits, origin); > > } > > /** > > * intel_frontbuffer_flip_prepare - prepare asynchronous frontbuffer flip > > - * @dev_priv: i915 device > > + * @i915: i915 device > > * @frontbuffer_bits: frontbuffer plane tracking bits > > * > > * This function gets called after scheduling a flip on @obj. The actual > > @@ -141,19 +141,19 @@ void __intel_fb_obj_flush(struct drm_i915_gem_object *obj, > > * > > * Can be called without any locks held. > > */ > > -void intel_frontbuffer_flip_prepare(struct drm_i915_private *dev_priv, > > +void intel_frontbuffer_flip_prepare(struct drm_i915_private *i915, > > unsigned frontbuffer_bits) > > { > > - spin_lock(&dev_priv->fb_tracking.lock); > > - dev_priv->fb_tracking.flip_bits |= frontbuffer_bits; > > + spin_lock(&i915->fb_tracking.lock); > > + i915->fb_tracking.flip_bits |= frontbuffer_bits; > > /* Remove stale busy bits due to the old buffer. */ > > - dev_priv->fb_tracking.busy_bits &= ~frontbuffer_bits; > > - spin_unlock(&dev_priv->fb_tracking.lock); > > + i915->fb_tracking.busy_bits &= ~frontbuffer_bits; > > + spin_unlock(&i915->fb_tracking.lock); > > } > > /** > > * intel_frontbuffer_flip_complete - complete asynchronous frontbuffer flip > > - * @dev_priv: i915 device > > + * @i915: i915 device > > * @frontbuffer_bits: frontbuffer plane tracking bits > > * > > * This function gets called after the flip has been latched and will complete > > @@ -161,23 +161,23 @@ void intel_frontbuffer_flip_prepare(struct drm_i915_private *dev_priv, > > * > > * Can be called without any locks held. > > */ > > -void intel_frontbuffer_flip_complete(struct drm_i915_private *dev_priv, > > +void intel_frontbuffer_flip_complete(struct drm_i915_private *i915, > > unsigned frontbuffer_bits) > > { > > - spin_lock(&dev_priv->fb_tracking.lock); > > + spin_lock(&i915->fb_tracking.lock); > > /* Mask any cancelled flips. */ > > - frontbuffer_bits &= dev_priv->fb_tracking.flip_bits; > > - dev_priv->fb_tracking.flip_bits &= ~frontbuffer_bits; > > - spin_unlock(&dev_priv->fb_tracking.lock); > > + frontbuffer_bits &= i915->fb_tracking.flip_bits; > > + i915->fb_tracking.flip_bits &= ~frontbuffer_bits; > > + spin_unlock(&i915->fb_tracking.lock); > > if (frontbuffer_bits) > > - intel_frontbuffer_flush(dev_priv, > > + intel_frontbuffer_flush(i915, > > frontbuffer_bits, ORIGIN_FLIP); > > } > > /** > > * intel_frontbuffer_flip - synchronous frontbuffer flip > > - * @dev_priv: i915 device > > + * @i915: i915 device > > * @frontbuffer_bits: frontbuffer plane tracking bits > > * > > * This function gets called after scheduling a flip on @obj. This is for > > @@ -186,13 +186,13 @@ void intel_frontbuffer_flip_complete(struct drm_i915_private *dev_priv, > > * > > * Can be called without any locks held. > > */ > > -void intel_frontbuffer_flip(struct drm_i915_private *dev_priv, > > +void intel_frontbuffer_flip(struct drm_i915_private *i915, > > unsigned frontbuffer_bits) > > { > > - spin_lock(&dev_priv->fb_tracking.lock); > > + spin_lock(&i915->fb_tracking.lock); > > /* Remove stale busy bits due to the old buffer. */ > > - dev_priv->fb_tracking.busy_bits &= ~frontbuffer_bits; > > - spin_unlock(&dev_priv->fb_tracking.lock); > > + i915->fb_tracking.busy_bits &= ~frontbuffer_bits; > > + spin_unlock(&i915->fb_tracking.lock); > > - intel_frontbuffer_flush(dev_priv, frontbuffer_bits, ORIGIN_FLIP); > > + intel_frontbuffer_flush(i915, frontbuffer_bits, ORIGIN_FLIP); > > } > > diff --git a/drivers/gpu/drm/i915/intel_frontbuffer.h b/drivers/gpu/drm/i915/intel_frontbuffer.h > > index 5727320c8084..55c42efee7ce 100644 > > --- a/drivers/gpu/drm/i915/intel_frontbuffer.h > > +++ b/drivers/gpu/drm/i915/intel_frontbuffer.h > > @@ -37,11 +37,11 @@ enum fb_op_origin { > > ORIGIN_DIRTYFB, > > }; > > -void intel_frontbuffer_flip_prepare(struct drm_i915_private *dev_priv, > > +void intel_frontbuffer_flip_prepare(struct drm_i915_private *i915, > > unsigned frontbuffer_bits); > > -void intel_frontbuffer_flip_complete(struct drm_i915_private *dev_priv, > > +void intel_frontbuffer_flip_complete(struct drm_i915_private *i915, > > unsigned frontbuffer_bits); > > -void intel_frontbuffer_flip(struct drm_i915_private *dev_priv, > > +void intel_frontbuffer_flip(struct drm_i915_private *i915, > > unsigned frontbuffer_bits); > > void __intel_fb_obj_invalidate(struct drm_i915_gem_object *obj, > > diff --git a/drivers/gpu/drm/i915/intel_gmbus.c b/drivers/gpu/drm/i915/intel_gmbus.c > > index aa88e6e7cc65..f1b9844b591e 100644 > > --- a/drivers/gpu/drm/i915/intel_gmbus.c > > +++ b/drivers/gpu/drm/i915/intel_gmbus.c > > @@ -89,42 +89,42 @@ static const struct gmbus_pin gmbus_pins_icp[] = { > > }; > > /* pin is expected to be valid */ > > -static const struct gmbus_pin *get_gmbus_pin(struct drm_i915_private *dev_priv, > > +static const struct gmbus_pin *get_gmbus_pin(struct drm_i915_private *i915, > > unsigned int pin) > > { > > - if (HAS_PCH_ICP(dev_priv)) > > + if (HAS_PCH_ICP(i915)) > > return &gmbus_pins_icp[pin]; > > - else if (HAS_PCH_CNP(dev_priv)) > > + else if (HAS_PCH_CNP(i915)) > > return &gmbus_pins_cnp[pin]; > > - else if (IS_GEN9_LP(dev_priv)) > > + else if (IS_GEN9_LP(i915)) > > return &gmbus_pins_bxt[pin]; > > - else if (IS_GEN9_BC(dev_priv)) > > + else if (IS_GEN9_BC(i915)) > > return &gmbus_pins_skl[pin]; > > - else if (IS_BROADWELL(dev_priv)) > > + else if (IS_BROADWELL(i915)) > > return &gmbus_pins_bdw[pin]; > > else > > return &gmbus_pins[pin]; > > } > > -bool intel_gmbus_is_valid_pin(struct drm_i915_private *dev_priv, > > +bool intel_gmbus_is_valid_pin(struct drm_i915_private *i915, > > unsigned int pin) > > { > > unsigned int size; > > - if (HAS_PCH_ICP(dev_priv)) > > + if (HAS_PCH_ICP(i915)) > > size = ARRAY_SIZE(gmbus_pins_icp); > > - else if (HAS_PCH_CNP(dev_priv)) > > + else if (HAS_PCH_CNP(i915)) > > size = ARRAY_SIZE(gmbus_pins_cnp); > > - else if (IS_GEN9_LP(dev_priv)) > > + else if (IS_GEN9_LP(i915)) > > size = ARRAY_SIZE(gmbus_pins_bxt); > > - else if (IS_GEN9_BC(dev_priv)) > > + else if (IS_GEN9_BC(i915)) > > size = ARRAY_SIZE(gmbus_pins_skl); > > - else if (IS_BROADWELL(dev_priv)) > > + else if (IS_BROADWELL(i915)) > > size = ARRAY_SIZE(gmbus_pins_bdw); > > else > > size = ARRAY_SIZE(gmbus_pins); > > - return pin < size && get_gmbus_pin(dev_priv, pin)->name; > > + return pin < size && get_gmbus_pin(i915, pin)->name; > > } > > /* Intel GPIO access functions */ > > @@ -138,13 +138,13 @@ to_intel_gmbus(struct i2c_adapter *i2c) > > } > > void > > -intel_gmbus_reset(struct drm_i915_private *dev_priv) > > +intel_gmbus_reset(struct drm_i915_private *i915) > > { > > I915_WRITE(GMBUS0, 0); > > I915_WRITE(GMBUS4, 0); > > } > > -static void pnv_gmbus_clock_gating(struct drm_i915_private *dev_priv, > > +static void pnv_gmbus_clock_gating(struct drm_i915_private *i915, > > bool enable) > > { > > u32 val; > > @@ -158,7 +158,7 @@ static void pnv_gmbus_clock_gating(struct drm_i915_private *dev_priv, > > I915_WRITE(DSPCLK_GATE_D, val); > > } > > -static void pch_gmbus_clock_gating(struct drm_i915_private *dev_priv, > > +static void pch_gmbus_clock_gating(struct drm_i915_private *i915, > > bool enable) > > { > > u32 val; > > @@ -171,7 +171,7 @@ static void pch_gmbus_clock_gating(struct drm_i915_private *dev_priv, > > I915_WRITE(SOUTH_DSPCLK_GATE_D, val); > > } > > -static void bxt_gmbus_clock_gating(struct drm_i915_private *dev_priv, > > +static void bxt_gmbus_clock_gating(struct drm_i915_private *i915, > > bool enable) > > { > > u32 val; > > @@ -186,7 +186,7 @@ static void bxt_gmbus_clock_gating(struct drm_i915_private *dev_priv, > > static u32 get_reserved(struct intel_gmbus *bus) > > { > > - struct drm_i915_private *i915 = bus->dev_priv; > > + struct drm_i915_private *i915 = bus->i915; > > struct intel_uncore *uncore = &i915->uncore; > > u32 reserved = 0; > > @@ -202,7 +202,7 @@ static u32 get_reserved(struct intel_gmbus *bus) > > static int get_clock(void *data) > > { > > struct intel_gmbus *bus = data; > > - struct intel_uncore *uncore = &bus->dev_priv->uncore; > > + struct intel_uncore *uncore = &bus->i915->uncore; > > u32 reserved = get_reserved(bus); > > intel_uncore_write_notrace(uncore, > > @@ -217,7 +217,7 @@ static int get_clock(void *data) > > static int get_data(void *data) > > { > > struct intel_gmbus *bus = data; > > - struct intel_uncore *uncore = &bus->dev_priv->uncore; > > + struct intel_uncore *uncore = &bus->i915->uncore; > > u32 reserved = get_reserved(bus); > > intel_uncore_write_notrace(uncore, > > @@ -232,7 +232,7 @@ static int get_data(void *data) > > static void set_clock(void *data, int state_high) > > { > > struct intel_gmbus *bus = data; > > - struct intel_uncore *uncore = &bus->dev_priv->uncore; > > + struct intel_uncore *uncore = &bus->i915->uncore; > > u32 reserved = get_reserved(bus); > > u32 clock_bits; > > @@ -251,7 +251,7 @@ static void set_clock(void *data, int state_high) > > static void set_data(void *data, int state_high) > > { > > struct intel_gmbus *bus = data; > > - struct intel_uncore *uncore = &bus->dev_priv->uncore; > > + struct intel_uncore *uncore = &bus->i915->uncore; > > u32 reserved = get_reserved(bus); > > u32 data_bits; > > @@ -271,12 +271,12 @@ intel_gpio_pre_xfer(struct i2c_adapter *adapter) > > struct intel_gmbus *bus = container_of(adapter, > > struct intel_gmbus, > > adapter); > > - struct drm_i915_private *dev_priv = bus->dev_priv; > > + struct drm_i915_private *i915 = bus->i915; > > - intel_gmbus_reset(dev_priv); > > + intel_gmbus_reset(i915); > > - if (IS_PINEVIEW(dev_priv)) > > - pnv_gmbus_clock_gating(dev_priv, false); > > + if (IS_PINEVIEW(i915)) > > + pnv_gmbus_clock_gating(i915, false); > > set_data(bus, 1); > > set_clock(bus, 1); > > @@ -290,24 +290,24 @@ intel_gpio_post_xfer(struct i2c_adapter *adapter) > > struct intel_gmbus *bus = container_of(adapter, > > struct intel_gmbus, > > adapter); > > - struct drm_i915_private *dev_priv = bus->dev_priv; > > + struct drm_i915_private *i915 = bus->i915; > > set_data(bus, 1); > > set_clock(bus, 1); > > - if (IS_PINEVIEW(dev_priv)) > > - pnv_gmbus_clock_gating(dev_priv, true); > > + if (IS_PINEVIEW(i915)) > > + pnv_gmbus_clock_gating(i915, true); > > } > > static void > > intel_gpio_setup(struct intel_gmbus *bus, unsigned int pin) > > { > > - struct drm_i915_private *dev_priv = bus->dev_priv; > > + struct drm_i915_private *i915 = bus->i915; > > struct i2c_algo_bit_data *algo; > > algo = &bus->bit_algo; > > - bus->gpio_reg = GPIO(get_gmbus_pin(dev_priv, pin)->gpio); > > + bus->gpio_reg = GPIO(get_gmbus_pin(i915, pin)->gpio); > > bus->adapter.algo_data = algo; > > algo->setsda = set_data; > > algo->setscl = set_clock; > > @@ -320,7 +320,7 @@ intel_gpio_setup(struct intel_gmbus *bus, unsigned int pin) > > algo->data = bus; > > } > > -static int gmbus_wait(struct drm_i915_private *dev_priv, u32 status, u32 irq_en) > > +static int gmbus_wait(struct drm_i915_private *i915, u32 status, u32 irq_en) > > { > > DEFINE_WAIT(wait); > > u32 gmbus2; > > @@ -330,10 +330,10 @@ static int gmbus_wait(struct drm_i915_private *dev_priv, u32 status, u32 irq_en) > > * we also need to check for NAKs besides the hw ready/idle signal, we > > * need to wake up periodically and check that ourselves. > > */ > > - if (!HAS_GMBUS_IRQ(dev_priv)) > > + if (!HAS_GMBUS_IRQ(i915)) > > irq_en = 0; > > - add_wait_queue(&dev_priv->gmbus_wait_queue, &wait); > > + add_wait_queue(&i915->gmbus_wait_queue, &wait); > > I915_WRITE_FW(GMBUS4, irq_en); > > status |= GMBUS_SATOER; > > @@ -342,7 +342,7 @@ static int gmbus_wait(struct drm_i915_private *dev_priv, u32 status, u32 irq_en) > > ret = wait_for((gmbus2 = I915_READ_FW(GMBUS2)) & status, 50); > > I915_WRITE_FW(GMBUS4, 0); > > - remove_wait_queue(&dev_priv->gmbus_wait_queue, &wait); > > + remove_wait_queue(&i915->gmbus_wait_queue, &wait); > > if (gmbus2 & GMBUS_SATOER) > > return -ENXIO; > > @@ -351,7 +351,7 @@ static int gmbus_wait(struct drm_i915_private *dev_priv, u32 status, u32 irq_en) > > } > > static int > > -gmbus_wait_idle(struct drm_i915_private *dev_priv) > > +gmbus_wait_idle(struct drm_i915_private *i915) > > { > > DEFINE_WAIT(wait); > > u32 irq_enable; > > @@ -359,36 +359,36 @@ gmbus_wait_idle(struct drm_i915_private *dev_priv) > > /* Important: The hw handles only the first bit, so set only one! */ > > irq_enable = 0; > > - if (HAS_GMBUS_IRQ(dev_priv)) > > + if (HAS_GMBUS_IRQ(i915)) > > irq_enable = GMBUS_IDLE_EN; > > - add_wait_queue(&dev_priv->gmbus_wait_queue, &wait); > > + add_wait_queue(&i915->gmbus_wait_queue, &wait); > > I915_WRITE_FW(GMBUS4, irq_enable); > > - ret = intel_wait_for_register_fw(&dev_priv->uncore, > > + ret = intel_wait_for_register_fw(&i915->uncore, > > GMBUS2, GMBUS_ACTIVE, 0, > > 10); > > I915_WRITE_FW(GMBUS4, 0); > > - remove_wait_queue(&dev_priv->gmbus_wait_queue, &wait); > > + remove_wait_queue(&i915->gmbus_wait_queue, &wait); > > return ret; > > } > > static inline > > -unsigned int gmbus_max_xfer_size(struct drm_i915_private *dev_priv) > > +unsigned int gmbus_max_xfer_size(struct drm_i915_private *i915) > > { > > - return INTEL_GEN(dev_priv) >= 9 ? GEN9_GMBUS_BYTE_COUNT_MAX : > > + return INTEL_GEN(i915) >= 9 ? GEN9_GMBUS_BYTE_COUNT_MAX : > > GMBUS_BYTE_COUNT_MAX; > > } > > static int > > -gmbus_xfer_read_chunk(struct drm_i915_private *dev_priv, > > +gmbus_xfer_read_chunk(struct drm_i915_private *i915, > > unsigned short addr, u8 *buf, unsigned int len, > > u32 gmbus0_reg, u32 gmbus1_index) > > { > > unsigned int size = len; > > - bool burst_read = len > gmbus_max_xfer_size(dev_priv); > > + bool burst_read = len > gmbus_max_xfer_size(i915); > > bool extra_byte_added = false; > > if (burst_read) { > > @@ -414,7 +414,7 @@ gmbus_xfer_read_chunk(struct drm_i915_private *dev_priv, > > int ret; > > u32 val, loop = 0; > > - ret = gmbus_wait(dev_priv, GMBUS_HW_RDY, GMBUS_HW_RDY_EN); > > + ret = gmbus_wait(i915, GMBUS_HW_RDY, GMBUS_HW_RDY_EN); > > if (ret) > > return ret; > > @@ -446,7 +446,7 @@ gmbus_xfer_read_chunk(struct drm_i915_private *dev_priv, > > #define INTEL_GMBUS_BURST_READ_MAX_LEN 767U > > static int > > -gmbus_xfer_read(struct drm_i915_private *dev_priv, struct i2c_msg *msg, > > +gmbus_xfer_read(struct drm_i915_private *i915, struct i2c_msg *msg, > > u32 gmbus0_reg, u32 gmbus1_index) > > { > > u8 *buf = msg->buf; > > @@ -455,12 +455,12 @@ gmbus_xfer_read(struct drm_i915_private *dev_priv, struct i2c_msg *msg, > > int ret; > > do { > > - if (HAS_GMBUS_BURST_READ(dev_priv)) > > + if (HAS_GMBUS_BURST_READ(i915)) > > len = min(rx_size, INTEL_GMBUS_BURST_READ_MAX_LEN); > > else > > - len = min(rx_size, gmbus_max_xfer_size(dev_priv)); > > + len = min(rx_size, gmbus_max_xfer_size(i915)); > > - ret = gmbus_xfer_read_chunk(dev_priv, msg->addr, buf, len, > > + ret = gmbus_xfer_read_chunk(i915, msg->addr, buf, len, > > gmbus0_reg, gmbus1_index); > > if (ret) > > return ret; > > @@ -473,7 +473,7 @@ gmbus_xfer_read(struct drm_i915_private *dev_priv, struct i2c_msg *msg, > > } > > static int > > -gmbus_xfer_write_chunk(struct drm_i915_private *dev_priv, > > +gmbus_xfer_write_chunk(struct drm_i915_private *i915, > > unsigned short addr, u8 *buf, unsigned int len, > > u32 gmbus1_index) > > { > > @@ -502,7 +502,7 @@ gmbus_xfer_write_chunk(struct drm_i915_private *dev_priv, > > I915_WRITE_FW(GMBUS3, val); > > - ret = gmbus_wait(dev_priv, GMBUS_HW_RDY, GMBUS_HW_RDY_EN); > > + ret = gmbus_wait(i915, GMBUS_HW_RDY, GMBUS_HW_RDY_EN); > > if (ret) > > return ret; > > } > > @@ -511,7 +511,7 @@ gmbus_xfer_write_chunk(struct drm_i915_private *dev_priv, > > } > > static int > > -gmbus_xfer_write(struct drm_i915_private *dev_priv, struct i2c_msg *msg, > > +gmbus_xfer_write(struct drm_i915_private *i915, struct i2c_msg *msg, > > u32 gmbus1_index) > > { > > u8 *buf = msg->buf; > > @@ -520,9 +520,9 @@ gmbus_xfer_write(struct drm_i915_private *dev_priv, struct i2c_msg *msg, > > int ret; > > do { > > - len = min(tx_size, gmbus_max_xfer_size(dev_priv)); > > + len = min(tx_size, gmbus_max_xfer_size(i915)); > > - ret = gmbus_xfer_write_chunk(dev_priv, msg->addr, buf, len, > > + ret = gmbus_xfer_write_chunk(i915, msg->addr, buf, len, > > gmbus1_index); > > if (ret) > > return ret; > > @@ -549,7 +549,7 @@ gmbus_is_index_xfer(struct i2c_msg *msgs, int i, int num) > > } > > static int > > -gmbus_index_xfer(struct drm_i915_private *dev_priv, struct i2c_msg *msgs, > > +gmbus_index_xfer(struct drm_i915_private *i915, struct i2c_msg *msgs, > > u32 gmbus0_reg) > > { > > u32 gmbus1_index = 0; > > @@ -568,10 +568,10 @@ gmbus_index_xfer(struct drm_i915_private *dev_priv, struct i2c_msg *msgs, > > I915_WRITE_FW(GMBUS5, gmbus5); > > if (msgs[1].flags & I2C_M_RD) > > - ret = gmbus_xfer_read(dev_priv, &msgs[1], gmbus0_reg, > > + ret = gmbus_xfer_read(i915, &msgs[1], gmbus0_reg, > > gmbus1_index); > > else > > - ret = gmbus_xfer_write(dev_priv, &msgs[1], gmbus1_index); > > + ret = gmbus_xfer_write(i915, &msgs[1], gmbus1_index); > > /* Clear GMBUS5 after each index transfer */ > > if (gmbus5) > > @@ -587,15 +587,15 @@ do_gmbus_xfer(struct i2c_adapter *adapter, struct i2c_msg *msgs, int num, > > struct intel_gmbus *bus = container_of(adapter, > > struct intel_gmbus, > > adapter); > > - struct drm_i915_private *dev_priv = bus->dev_priv; > > + struct drm_i915_private *i915 = bus->i915; > > int i = 0, inc, try = 0; > > int ret = 0; > > /* Display WA #0868: skl,bxt,kbl,cfl,glk,cnl */ > > - if (IS_GEN9_LP(dev_priv)) > > - bxt_gmbus_clock_gating(dev_priv, false); > > - else if (HAS_PCH_SPT(dev_priv) || HAS_PCH_CNP(dev_priv)) > > - pch_gmbus_clock_gating(dev_priv, false); > > + if (IS_GEN9_LP(i915)) > > + bxt_gmbus_clock_gating(i915, false); > > + else if (HAS_PCH_SPT(i915) || HAS_PCH_CNP(i915)) > > + pch_gmbus_clock_gating(i915, false); > > retry: > > I915_WRITE_FW(GMBUS0, gmbus0_source | bus->reg0); > > @@ -603,18 +603,18 @@ do_gmbus_xfer(struct i2c_adapter *adapter, struct i2c_msg *msgs, int num, > > for (; i < num; i += inc) { > > inc = 1; > > if (gmbus_is_index_xfer(msgs, i, num)) { > > - ret = gmbus_index_xfer(dev_priv, &msgs[i], > > + ret = gmbus_index_xfer(i915, &msgs[i], > > gmbus0_source | bus->reg0); > > inc = 2; /* an index transmission is two msgs */ > > } else if (msgs[i].flags & I2C_M_RD) { > > - ret = gmbus_xfer_read(dev_priv, &msgs[i], > > + ret = gmbus_xfer_read(i915, &msgs[i], > > gmbus0_source | bus->reg0, 0); > > } else { > > - ret = gmbus_xfer_write(dev_priv, &msgs[i], 0); > > + ret = gmbus_xfer_write(i915, &msgs[i], 0); > > } > > if (!ret) > > - ret = gmbus_wait(dev_priv, > > + ret = gmbus_wait(i915, > > GMBUS_HW_WAIT_PHASE, GMBUS_HW_WAIT_EN); > > if (ret == -ETIMEDOUT) > > goto timeout; > > @@ -632,7 +632,7 @@ do_gmbus_xfer(struct i2c_adapter *adapter, struct i2c_msg *msgs, int num, > > * We will re-enable it at the start of the next xfer, > > * till then let it sleep. > > */ > > - if (gmbus_wait_idle(dev_priv)) { > > + if (gmbus_wait_idle(i915)) { > > DRM_DEBUG_KMS("GMBUS [%s] timed out waiting for idle\n", > > adapter->name); > > ret = -ETIMEDOUT; > > @@ -656,7 +656,7 @@ do_gmbus_xfer(struct i2c_adapter *adapter, struct i2c_msg *msgs, int num, > > * it's slow responding and only answers on the 2nd retry. > > */ > > ret = -ENXIO; > > - if (gmbus_wait_idle(dev_priv)) { > > + if (gmbus_wait_idle(i915)) { > > DRM_DEBUG_KMS("GMBUS [%s] timed out after NAK\n", > > adapter->name); > > ret = -ETIMEDOUT; > > @@ -701,10 +701,10 @@ do_gmbus_xfer(struct i2c_adapter *adapter, struct i2c_msg *msgs, int num, > > out: > > /* Display WA #0868: skl,bxt,kbl,cfl,glk,cnl */ > > - if (IS_GEN9_LP(dev_priv)) > > - bxt_gmbus_clock_gating(dev_priv, true); > > - else if (HAS_PCH_SPT(dev_priv) || HAS_PCH_CNP(dev_priv)) > > - pch_gmbus_clock_gating(dev_priv, true); > > + if (IS_GEN9_LP(i915)) > > + bxt_gmbus_clock_gating(i915, true); > > + else if (HAS_PCH_SPT(i915) || HAS_PCH_CNP(i915)) > > + pch_gmbus_clock_gating(i915, true); > > return ret; > > } > > @@ -714,11 +714,11 @@ gmbus_xfer(struct i2c_adapter *adapter, struct i2c_msg *msgs, int num) > > { > > struct intel_gmbus *bus = > > container_of(adapter, struct intel_gmbus, adapter); > > - struct drm_i915_private *dev_priv = bus->dev_priv; > > + struct drm_i915_private *i915 = bus->i915; > > intel_wakeref_t wakeref; > > int ret; > > - wakeref = intel_display_power_get(dev_priv, POWER_DOMAIN_GMBUS); > > + wakeref = intel_display_power_get(i915, POWER_DOMAIN_GMBUS); > > if (bus->force_bit) { > > ret = i2c_bit_algo.master_xfer(adapter, msgs, num); > > @@ -730,7 +730,7 @@ gmbus_xfer(struct i2c_adapter *adapter, struct i2c_msg *msgs, int num) > > bus->force_bit |= GMBUS_FORCE_BIT_RETRY; > > } > > - intel_display_power_put(dev_priv, POWER_DOMAIN_GMBUS, wakeref); > > + intel_display_power_put(i915, POWER_DOMAIN_GMBUS, wakeref); > > return ret; > > } > > @@ -739,7 +739,7 @@ int intel_gmbus_output_aksv(struct i2c_adapter *adapter) > > { > > struct intel_gmbus *bus = > > container_of(adapter, struct intel_gmbus, adapter); > > - struct drm_i915_private *dev_priv = bus->dev_priv; > > + struct drm_i915_private *i915 = bus->i915; > > u8 cmd = DRM_HDCP_DDC_AKSV; > > u8 buf[DRM_HDCP_KSV_LEN] = { 0 }; > > struct i2c_msg msgs[] = { > > @@ -759,8 +759,8 @@ int intel_gmbus_output_aksv(struct i2c_adapter *adapter) > > intel_wakeref_t wakeref; > > int ret; > > - wakeref = intel_display_power_get(dev_priv, POWER_DOMAIN_GMBUS); > > - mutex_lock(&dev_priv->gmbus_mutex); > > + wakeref = intel_display_power_get(i915, POWER_DOMAIN_GMBUS); > > + mutex_lock(&i915->gmbus_mutex); > > /* > > * In order to output Aksv to the receiver, use an indexed write to > > @@ -769,8 +769,8 @@ int intel_gmbus_output_aksv(struct i2c_adapter *adapter) > > */ > > ret = do_gmbus_xfer(adapter, msgs, ARRAY_SIZE(msgs), GMBUS_AKSV_SELECT); > > - mutex_unlock(&dev_priv->gmbus_mutex); > > - intel_display_power_put(dev_priv, POWER_DOMAIN_GMBUS, wakeref); > > + mutex_unlock(&i915->gmbus_mutex); > > + intel_display_power_put(i915, POWER_DOMAIN_GMBUS, wakeref); > > return ret; > > } > > @@ -793,27 +793,27 @@ static void gmbus_lock_bus(struct i2c_adapter *adapter, > > unsigned int flags) > > { > > struct intel_gmbus *bus = to_intel_gmbus(adapter); > > - struct drm_i915_private *dev_priv = bus->dev_priv; > > + struct drm_i915_private *i915 = bus->i915; > > - mutex_lock(&dev_priv->gmbus_mutex); > > + mutex_lock(&i915->gmbus_mutex); > > } > > static int gmbus_trylock_bus(struct i2c_adapter *adapter, > > unsigned int flags) > > { > > struct intel_gmbus *bus = to_intel_gmbus(adapter); > > - struct drm_i915_private *dev_priv = bus->dev_priv; > > + struct drm_i915_private *i915 = bus->i915; > > - return mutex_trylock(&dev_priv->gmbus_mutex); > > + return mutex_trylock(&i915->gmbus_mutex); > > } > > static void gmbus_unlock_bus(struct i2c_adapter *adapter, > > unsigned int flags) > > { > > struct intel_gmbus *bus = to_intel_gmbus(adapter); > > - struct drm_i915_private *dev_priv = bus->dev_priv; > > + struct drm_i915_private *i915 = bus->i915; > > - mutex_unlock(&dev_priv->gmbus_mutex); > > + mutex_unlock(&i915->gmbus_mutex); > > } > > static const struct i2c_lock_operations gmbus_lock_ops = { > > @@ -824,45 +824,45 @@ static const struct i2c_lock_operations gmbus_lock_ops = { > > /** > > * intel_gmbus_setup - instantiate all Intel i2c GMBuses > > - * @dev_priv: i915 device private > > + * @i915: i915 device private > > */ > > -int intel_gmbus_setup(struct drm_i915_private *dev_priv) > > +int intel_gmbus_setup(struct drm_i915_private *i915) > > { > > - struct pci_dev *pdev = dev_priv->drm.pdev; > > + struct pci_dev *pdev = i915->drm.pdev; > > struct intel_gmbus *bus; > > unsigned int pin; > > int ret; > > - if (!HAS_DISPLAY(dev_priv)) > > + if (!HAS_DISPLAY(i915)) > > return 0; > > - if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) > > - dev_priv->gpio_mmio_base = VLV_DISPLAY_BASE; > > - else if (!HAS_GMCH(dev_priv)) > > + if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) > > + i915->gpio_mmio_base = VLV_DISPLAY_BASE; > > + else if (!HAS_GMCH(i915)) > > /* > > * Broxton uses the same PCH offsets for South Display Engine, > > * even though it doesn't have a PCH. > > */ > > - dev_priv->gpio_mmio_base = PCH_DISPLAY_BASE; > > + i915->gpio_mmio_base = PCH_DISPLAY_BASE; > > - mutex_init(&dev_priv->gmbus_mutex); > > - init_waitqueue_head(&dev_priv->gmbus_wait_queue); > > + mutex_init(&i915->gmbus_mutex); > > + init_waitqueue_head(&i915->gmbus_wait_queue); > > - for (pin = 0; pin < ARRAY_SIZE(dev_priv->gmbus); pin++) { > > - if (!intel_gmbus_is_valid_pin(dev_priv, pin)) > > + for (pin = 0; pin < ARRAY_SIZE(i915->gmbus); pin++) { > > + if (!intel_gmbus_is_valid_pin(i915, pin)) > > continue; > > - bus = &dev_priv->gmbus[pin]; > > + bus = &i915->gmbus[pin]; > > bus->adapter.owner = THIS_MODULE; > > bus->adapter.class = I2C_CLASS_DDC; > > snprintf(bus->adapter.name, > > sizeof(bus->adapter.name), > > "i915 gmbus %s", > > - get_gmbus_pin(dev_priv, pin)->name); > > + get_gmbus_pin(i915, pin)->name); > > bus->adapter.dev.parent = &pdev->dev; > > - bus->dev_priv = dev_priv; > > + bus->i915 = i915; > > bus->adapter.algo = &gmbus_algorithm; > > bus->adapter.lock_ops = &gmbus_lock_ops; > > @@ -877,7 +877,7 @@ int intel_gmbus_setup(struct drm_i915_private *dev_priv) > > bus->reg0 = pin | GMBUS_RATE_100KHZ; > > /* gmbus seems to be broken on i830 */ > > - if (IS_I830(dev_priv)) > > + if (IS_I830(i915)) > > bus->force_bit = 1; > > intel_gpio_setup(bus, pin); > > @@ -887,28 +887,28 @@ int intel_gmbus_setup(struct drm_i915_private *dev_priv) > > goto err; > > } > > - intel_gmbus_reset(dev_priv); > > + intel_gmbus_reset(i915); > > return 0; > > err: > > while (pin--) { > > - if (!intel_gmbus_is_valid_pin(dev_priv, pin)) > > + if (!intel_gmbus_is_valid_pin(i915, pin)) > > continue; > > - bus = &dev_priv->gmbus[pin]; > > + bus = &i915->gmbus[pin]; > > i2c_del_adapter(&bus->adapter); > > } > > return ret; > > } > > -struct i2c_adapter *intel_gmbus_get_adapter(struct drm_i915_private *dev_priv, > > +struct i2c_adapter *intel_gmbus_get_adapter(struct drm_i915_private *i915, > > unsigned int pin) > > { > > - if (WARN_ON(!intel_gmbus_is_valid_pin(dev_priv, pin))) > > + if (WARN_ON(!intel_gmbus_is_valid_pin(i915, pin))) > > return NULL; > > - return &dev_priv->gmbus[pin].adapter; > > + return &i915->gmbus[pin].adapter; > > } > > void intel_gmbus_set_speed(struct i2c_adapter *adapter, int speed) > > @@ -921,16 +921,16 @@ void intel_gmbus_set_speed(struct i2c_adapter *adapter, int speed) > > void intel_gmbus_force_bit(struct i2c_adapter *adapter, bool force_bit) > > { > > struct intel_gmbus *bus = to_intel_gmbus(adapter); > > - struct drm_i915_private *dev_priv = bus->dev_priv; > > + struct drm_i915_private *i915 = bus->i915; > > - mutex_lock(&dev_priv->gmbus_mutex); > > + mutex_lock(&i915->gmbus_mutex); > > bus->force_bit += force_bit ? 1 : -1; > > DRM_DEBUG_KMS("%sabling bit-banging on %s. force bit now %d\n", > > force_bit ? "en" : "dis", adapter->name, > > bus->force_bit); > > - mutex_unlock(&dev_priv->gmbus_mutex); > > + mutex_unlock(&i915->gmbus_mutex); > > } > > bool intel_gmbus_is_forced_bit(struct i2c_adapter *adapter) > > @@ -940,16 +940,16 @@ bool intel_gmbus_is_forced_bit(struct i2c_adapter *adapter) > > return bus->force_bit; > > } > > -void intel_gmbus_teardown(struct drm_i915_private *dev_priv) > > +void intel_gmbus_teardown(struct drm_i915_private *i915) > > { > > struct intel_gmbus *bus; > > unsigned int pin; > > - for (pin = 0; pin < ARRAY_SIZE(dev_priv->gmbus); pin++) { > > - if (!intel_gmbus_is_valid_pin(dev_priv, pin)) > > + for (pin = 0; pin < ARRAY_SIZE(i915->gmbus); pin++) { > > + if (!intel_gmbus_is_valid_pin(i915, pin)) > > continue; > > - bus = &dev_priv->gmbus[pin]; > > + bus = &i915->gmbus[pin]; > > i2c_del_adapter(&bus->adapter); > > } > > } > > diff --git a/drivers/gpu/drm/i915/intel_gmbus.h b/drivers/gpu/drm/i915/intel_gmbus.h > > index d989085b8d22..4a4196ca8343 100644 > > --- a/drivers/gpu/drm/i915/intel_gmbus.h > > +++ b/drivers/gpu/drm/i915/intel_gmbus.h > > @@ -11,17 +11,17 @@ > > struct drm_i915_private; > > struct i2c_adapter; > > -int intel_gmbus_setup(struct drm_i915_private *dev_priv); > > -void intel_gmbus_teardown(struct drm_i915_private *dev_priv); > > -bool intel_gmbus_is_valid_pin(struct drm_i915_private *dev_priv, > > +int intel_gmbus_setup(struct drm_i915_private *i915); > > +void intel_gmbus_teardown(struct drm_i915_private *i915); > > +bool intel_gmbus_is_valid_pin(struct drm_i915_private *i915, > > unsigned int pin); > > int intel_gmbus_output_aksv(struct i2c_adapter *adapter); > > struct i2c_adapter * > > -intel_gmbus_get_adapter(struct drm_i915_private *dev_priv, unsigned int pin); > > +intel_gmbus_get_adapter(struct drm_i915_private *i915, unsigned int pin); > > void intel_gmbus_set_speed(struct i2c_adapter *adapter, int speed); > > void intel_gmbus_force_bit(struct i2c_adapter *adapter, bool force_bit); > > bool intel_gmbus_is_forced_bit(struct i2c_adapter *adapter); > > -void intel_gmbus_reset(struct drm_i915_private *dev_priv); > > +void intel_gmbus_reset(struct drm_i915_private *i915); > > #endif /* __INTEL_GMBUS_H__ */ > > diff --git a/drivers/gpu/drm/i915/intel_guc.c b/drivers/gpu/drm/i915/intel_guc.c > > index c40a6efdd33a..f08ac29c2a4f 100644 > > --- a/drivers/gpu/drm/i915/intel_guc.c > > +++ b/drivers/gpu/drm/i915/intel_guc.c > > @@ -29,14 +29,14 @@ > > static void gen8_guc_raise_irq(struct intel_guc *guc) > > { > > - struct drm_i915_private *dev_priv = guc_to_i915(guc); > > + struct drm_i915_private *i915 = guc_to_i915(guc); > > I915_WRITE(GUC_SEND_INTERRUPT, GUC_SEND_TRIGGER); > > } > > static void gen11_guc_raise_irq(struct intel_guc *guc) > > { > > - struct drm_i915_private *dev_priv = guc_to_i915(guc); > > + struct drm_i915_private *i915 = guc_to_i915(guc); > > I915_WRITE(GEN11_GUC_HOST_INTERRUPT, 0); > > } > > @@ -52,11 +52,11 @@ static inline i915_reg_t guc_send_reg(struct intel_guc *guc, u32 i) > > void intel_guc_init_send_regs(struct intel_guc *guc) > > { > > - struct drm_i915_private *dev_priv = guc_to_i915(guc); > > + struct drm_i915_private *i915 = guc_to_i915(guc); > > enum forcewake_domains fw_domains = 0; > > unsigned int i; > > - if (INTEL_GEN(dev_priv) >= 11) { > > + if (INTEL_GEN(i915) >= 11) { > > guc->send_regs.base = > > i915_mmio_reg_offset(GEN11_SOFT_SCRATCH(0)); > > guc->send_regs.count = GEN11_SOFT_SCRATCH_COUNT; > > @@ -67,7 +67,7 @@ void intel_guc_init_send_regs(struct intel_guc *guc) > > } > > for (i = 0; i < guc->send_regs.count; i++) { > > - fw_domains |= intel_uncore_forcewake_for_reg(&dev_priv->uncore, > > + fw_domains |= intel_uncore_forcewake_for_reg(&i915->uncore, > > guc_send_reg(guc, i), > > FW_REG_READ | FW_REG_WRITE); > > } > > @@ -101,7 +101,7 @@ void intel_guc_init_early(struct intel_guc *guc) > > static int guc_init_wq(struct intel_guc *guc) > > { > > - struct drm_i915_private *dev_priv = guc_to_i915(guc); > > + struct drm_i915_private *i915 = guc_to_i915(guc); > > /* > > * GuC log buffer flush work item has to do register access to > > @@ -135,8 +135,8 @@ static int guc_init_wq(struct intel_guc *guc) > > * to make sure we're always sending a single preemption request with a > > * single workitem. > > */ > > - if (HAS_LOGICAL_RING_PREEMPTION(dev_priv) && > > - USES_GUC_SUBMISSION(dev_priv)) { > > + if (HAS_LOGICAL_RING_PREEMPTION(i915) && > > + USES_GUC_SUBMISSION(i915)) { > > guc->preempt_wq = alloc_ordered_workqueue("i915-guc_preempt", > > WQ_HIGHPRI); > > if (!guc->preempt_wq) { > > @@ -211,7 +211,7 @@ static void guc_shared_data_destroy(struct intel_guc *guc) > > int intel_guc_init(struct intel_guc *guc) > > { > > - struct drm_i915_private *dev_priv = guc_to_i915(guc); > > + struct drm_i915_private *i915 = guc_to_i915(guc); > > int ret; > > ret = intel_uc_fw_init(&guc->fw); > > @@ -237,7 +237,7 @@ int intel_guc_init(struct intel_guc *guc) > > goto err_ads; > > /* We need to notify the guc whenever we change the GGTT */ > > - i915_ggtt_enable_guc(dev_priv); > > + i915_ggtt_enable_guc(i915); > > return 0; > > @@ -256,9 +256,9 @@ int intel_guc_init(struct intel_guc *guc) > > void intel_guc_fini(struct intel_guc *guc) > > { > > - struct drm_i915_private *dev_priv = guc_to_i915(guc); > > + struct drm_i915_private *i915 = guc_to_i915(guc); > > - i915_ggtt_disable_guc(dev_priv); > > + i915_ggtt_disable_guc(i915); > > intel_guc_ct_fini(&guc->ct); > > @@ -366,7 +366,7 @@ static u32 guc_ctl_ads_flags(struct intel_guc *guc) > > */ > > void intel_guc_init_params(struct intel_guc *guc) > > { > > - struct drm_i915_private *dev_priv = guc_to_i915(guc); > > + struct drm_i915_private *i915 = guc_to_i915(guc); > > u32 params[GUC_CTL_MAX_DWORDS]; > > int i; > > @@ -386,14 +386,14 @@ void intel_guc_init_params(struct intel_guc *guc) > > * they are power context saved so it's ok to release forcewake > > * when we are done here and take it again at xfer time. > > */ > > - intel_uncore_forcewake_get(&dev_priv->uncore, FORCEWAKE_BLITTER); > > + intel_uncore_forcewake_get(&i915->uncore, FORCEWAKE_BLITTER); > > I915_WRITE(SOFT_SCRATCH(0), 0); > > for (i = 0; i < GUC_CTL_MAX_DWORDS; i++) > > I915_WRITE(SOFT_SCRATCH(1 + i), params[i]); > > - intel_uncore_forcewake_put(&dev_priv->uncore, FORCEWAKE_BLITTER); > > + intel_uncore_forcewake_put(&i915->uncore, FORCEWAKE_BLITTER); > > } > > int intel_guc_send_nop(struct intel_guc *guc, const u32 *action, u32 len, > > @@ -414,8 +414,8 @@ void intel_guc_to_host_event_handler_nop(struct intel_guc *guc) > > int intel_guc_send_mmio(struct intel_guc *guc, const u32 *action, u32 len, > > u32 *response_buf, u32 response_buf_size) > > { > > - struct drm_i915_private *dev_priv = guc_to_i915(guc); > > - struct intel_uncore *uncore = &dev_priv->uncore; > > + struct drm_i915_private *i915 = guc_to_i915(guc); > > + struct intel_uncore *uncore = &i915->uncore; > > u32 status; > > int i; > > int ret; > > @@ -497,12 +497,12 @@ int intel_guc_to_host_process_recv_msg(struct intel_guc *guc, > > int intel_guc_sample_forcewake(struct intel_guc *guc) > > { > > - struct drm_i915_private *dev_priv = guc_to_i915(guc); > > + struct drm_i915_private *i915 = guc_to_i915(guc); > > u32 action[2]; > > action[0] = INTEL_GUC_ACTION_SAMPLE_FORCEWAKE; > > /* WaRsDisableCoarsePowerGating:skl,cnl */ > > - if (!HAS_RC6(dev_priv) || NEEDS_WaRsDisableCoarsePowerGating(dev_priv)) > > + if (!HAS_RC6(i915) || NEEDS_WaRsDisableCoarsePowerGating(i915)) > > action[1] = 0; > > else > > /* bit 0 and 1 are for Render and Media domain separately */ > > @@ -538,7 +538,7 @@ int intel_guc_auth_huc(struct intel_guc *guc, u32 rsa_offset) > > */ > > int intel_guc_suspend(struct intel_guc *guc) > > { > > - struct drm_i915_private *dev_priv = guc_to_i915(guc); > > + struct drm_i915_private *i915 = guc_to_i915(guc); > > int ret; > > u32 status; > > u32 action[] = { > > @@ -562,7 +562,7 @@ int intel_guc_suspend(struct intel_guc *guc) > > if (ret) > > return ret; > > - ret = __intel_wait_for_register(&dev_priv->uncore, SOFT_SCRATCH(14), > > + ret = __intel_wait_for_register(&i915->uncore, SOFT_SCRATCH(14), > > INTEL_GUC_SLEEP_STATE_INVALID_MASK, > > 0, 0, 10, &status); > > if (ret) > > @@ -658,17 +658,17 @@ int intel_guc_resume(struct intel_guc *guc) > > */ > > struct i915_vma *intel_guc_allocate_vma(struct intel_guc *guc, u32 size) > > { > > - struct drm_i915_private *dev_priv = guc_to_i915(guc); > > + struct drm_i915_private *i915 = guc_to_i915(guc); > > struct drm_i915_gem_object *obj; > > struct i915_vma *vma; > > u64 flags; > > int ret; > > - obj = i915_gem_object_create_shmem(dev_priv, size); > > + obj = i915_gem_object_create_shmem(i915, size); > > if (IS_ERR(obj)) > > return ERR_CAST(obj); > > - vma = i915_vma_instance(obj, &dev_priv->ggtt.vm, NULL); > > + vma = i915_vma_instance(obj, &i915->ggtt.vm, NULL); > > if (IS_ERR(vma)) > > goto err; > > diff --git a/drivers/gpu/drm/i915/intel_guc_ads.c b/drivers/gpu/drm/i915/intel_guc_ads.c > > index ecb69fc94218..4ac9ed2dd467 100644 > > --- a/drivers/gpu/drm/i915/intel_guc_ads.c > > +++ b/drivers/gpu/drm/i915/intel_guc_ads.c > > @@ -85,7 +85,7 @@ struct __guc_ads_blob { > > static int __guc_ads_init(struct intel_guc *guc) > > { > > - struct drm_i915_private *dev_priv = guc_to_i915(guc); > > + struct drm_i915_private *i915 = guc_to_i915(guc); > > struct __guc_ads_blob *blob; > > const u32 skipped_size = LRC_PPHWSP_SZ * PAGE_SIZE + LR_HW_CONTEXT_SIZE; > > u32 base; > > @@ -115,18 +115,18 @@ static int __guc_ads_init(struct intel_guc *guc) > > */ > > blob->ads.golden_context_lrca[engine_class] = 0; > > blob->ads.eng_state_size[engine_class] = > > - intel_engine_context_size(dev_priv, engine_class) - > > + intel_engine_context_size(i915, engine_class) - > > skipped_size; > > } > > /* System info */ > > - blob->system_info.slice_enabled = hweight8(RUNTIME_INFO(dev_priv)->sseu.slice_mask); > > + blob->system_info.slice_enabled = hweight8(RUNTIME_INFO(i915)->sseu.slice_mask); > > blob->system_info.rcs_enabled = 1; > > blob->system_info.bcs_enabled = 1; > > - blob->system_info.vdbox_enable_mask = VDBOX_MASK(dev_priv); > > - blob->system_info.vebox_enable_mask = VEBOX_MASK(dev_priv); > > - blob->system_info.vdbox_sfc_support_mask = RUNTIME_INFO(dev_priv)->vdbox_sfc_access; > > + blob->system_info.vdbox_enable_mask = VDBOX_MASK(i915); > > + blob->system_info.vebox_enable_mask = VEBOX_MASK(i915); > > + blob->system_info.vdbox_sfc_support_mask = RUNTIME_INFO(i915)->vdbox_sfc_access; > > base = intel_guc_ggtt_offset(guc, guc->ads_vma); > > diff --git a/drivers/gpu/drm/i915/intel_guc_fw.c b/drivers/gpu/drm/i915/intel_guc_fw.c > > index 72cdafd9636a..2beb90c62a40 100644 > > --- a/drivers/gpu/drm/i915/intel_guc_fw.c > > +++ b/drivers/gpu/drm/i915/intel_guc_fw.c > > @@ -125,7 +125,7 @@ void intel_guc_fw_init_early(struct intel_guc *guc) > > static void guc_prepare_xfer(struct intel_guc *guc) > > { > > - struct drm_i915_private *dev_priv = guc_to_i915(guc); > > + struct drm_i915_private *i915 = guc_to_i915(guc); > > /* Must program this register before loading the ucode with DMA */ > > I915_WRITE(GUC_SHIM_CONTROL, GUC_DISABLE_SRAM_INIT_TO_ZEROES | > > @@ -135,12 +135,12 @@ static void guc_prepare_xfer(struct intel_guc *guc) > > GUC_ENABLE_READ_CACHE_FOR_WOPCM_DATA | > > GUC_ENABLE_MIA_CLOCK_GATING); > > - if (IS_GEN9_LP(dev_priv)) > > + if (IS_GEN9_LP(i915)) > > I915_WRITE(GEN9LP_GT_PM_CONFIG, GT_DOORBELL_ENABLE); > > else > > I915_WRITE(GEN9_GT_PM_CONFIG, GT_DOORBELL_ENABLE); > > - if (IS_GEN(dev_priv, 9)) { > > + if (IS_GEN(i915, 9)) { > > /* DOP Clock Gating Enable for GuC clocks */ > > I915_WRITE(GEN7_MISCCPCTL, (GEN8_DOP_CLOCK_GATE_GUC_ENABLE | > > I915_READ(GEN7_MISCCPCTL))); > > @@ -153,7 +153,7 @@ static void guc_prepare_xfer(struct intel_guc *guc) > > /* Copy RSA signature from the fw image to HW for verification */ > > static void guc_xfer_rsa(struct intel_guc *guc) > > { > > - struct drm_i915_private *dev_priv = guc_to_i915(guc); > > + struct drm_i915_private *i915 = guc_to_i915(guc); > > struct intel_uc_fw *fw = &guc->fw; > > struct sg_table *pages = fw->obj->mm.pages; > > u32 rsa[UOS_RSA_SCRATCH_COUNT]; > > @@ -168,7 +168,7 @@ static void guc_xfer_rsa(struct intel_guc *guc) > > static bool guc_xfer_completed(struct intel_guc *guc, u32 *status) > > { > > - struct drm_i915_private *dev_priv = guc_to_i915(guc); > > + struct drm_i915_private *i915 = guc_to_i915(guc); > > /* Did we complete the xfer? */ > > *status = I915_READ(DMA_CTRL); > > @@ -186,7 +186,7 @@ static bool guc_xfer_completed(struct intel_guc *guc, u32 *status) > > */ > > static inline bool guc_ready(struct intel_guc *guc, u32 *status) > > { > > - struct drm_i915_private *dev_priv = guc_to_i915(guc); > > + struct drm_i915_private *i915 = guc_to_i915(guc); > > u32 val = I915_READ(GUC_STATUS); > > u32 uk_val = val & GS_UKERNEL_MASK; > > @@ -234,7 +234,7 @@ static int guc_wait_ucode(struct intel_guc *guc) > > */ > > static int guc_xfer_ucode(struct intel_guc *guc) > > { > > - struct drm_i915_private *dev_priv = guc_to_i915(guc); > > + struct drm_i915_private *i915 = guc_to_i915(guc); > > struct intel_uc_fw *guc_fw = &guc->fw; > > unsigned long offset; > > @@ -267,12 +267,12 @@ static int guc_xfer_ucode(struct intel_guc *guc) > > static int guc_fw_xfer(struct intel_uc_fw *guc_fw) > > { > > struct intel_guc *guc = container_of(guc_fw, struct intel_guc, fw); > > - struct drm_i915_private *dev_priv = guc_to_i915(guc); > > + struct drm_i915_private *i915 = guc_to_i915(guc); > > int ret; > > GEM_BUG_ON(guc_fw->type != INTEL_UC_FW_TYPE_GUC); > > - intel_uncore_forcewake_get(&dev_priv->uncore, FORCEWAKE_ALL); > > + intel_uncore_forcewake_get(&i915->uncore, FORCEWAKE_ALL); > > guc_prepare_xfer(guc); > > @@ -285,7 +285,7 @@ static int guc_fw_xfer(struct intel_uc_fw *guc_fw) > > ret = guc_xfer_ucode(guc); > > - intel_uncore_forcewake_put(&dev_priv->uncore, FORCEWAKE_ALL); > > + intel_uncore_forcewake_put(&i915->uncore, FORCEWAKE_ALL); > > return ret; > > } > > diff --git a/drivers/gpu/drm/i915/intel_guc_log.c b/drivers/gpu/drm/i915/intel_guc_log.c > > index 67eadc82c396..300ed940f0b7 100644 > > --- a/drivers/gpu/drm/i915/intel_guc_log.c > > +++ b/drivers/gpu/drm/i915/intel_guc_log.c > > @@ -390,7 +390,7 @@ void intel_guc_log_init_early(struct intel_guc_log *log) > > static int guc_log_relay_create(struct intel_guc_log *log) > > { > > struct intel_guc *guc = log_to_guc(log); > > - struct drm_i915_private *dev_priv = guc_to_i915(guc); > > + struct drm_i915_private *i915 = guc_to_i915(guc); > > struct rchan *guc_log_relay_chan; > > size_t n_subbufs, subbuf_size; > > int ret; > > @@ -409,9 +409,9 @@ static int guc_log_relay_create(struct intel_guc_log *log) > > n_subbufs = 8; > > guc_log_relay_chan = relay_open("guc_log", > > - dev_priv->drm.primary->debugfs_root, > > + i915->drm.primary->debugfs_root, > > subbuf_size, n_subbufs, > > - &relay_callbacks, dev_priv); > > + &relay_callbacks, i915); > > if (!guc_log_relay_chan) { > > DRM_ERROR("Couldn't create relay chan for GuC logging\n"); > > @@ -436,7 +436,7 @@ static void guc_log_relay_destroy(struct intel_guc_log *log) > > static void guc_log_capture_logs(struct intel_guc_log *log) > > { > > struct intel_guc *guc = log_to_guc(log); > > - struct drm_i915_private *dev_priv = guc_to_i915(guc); > > + struct drm_i915_private *i915 = guc_to_i915(guc); > > intel_wakeref_t wakeref; > > guc_read_update_log_buffer(log); > > @@ -445,7 +445,7 @@ static void guc_log_capture_logs(struct intel_guc_log *log) > > * Generally device is expected to be active only at this > > * time, so get/put should be really quick. > > */ > > - with_intel_runtime_pm(dev_priv, wakeref) > > + with_intel_runtime_pm(i915, wakeref) > > guc_action_flush_log_complete(guc); > > } > > @@ -505,7 +505,7 @@ void intel_guc_log_destroy(struct intel_guc_log *log) > > int intel_guc_log_set_level(struct intel_guc_log *log, u32 level) > > { > > struct intel_guc *guc = log_to_guc(log); > > - struct drm_i915_private *dev_priv = guc_to_i915(guc); > > + struct drm_i915_private *i915 = guc_to_i915(guc); > > intel_wakeref_t wakeref; > > int ret = 0; > > @@ -519,12 +519,12 @@ int intel_guc_log_set_level(struct intel_guc_log *log, u32 level) > > if (level < GUC_LOG_LEVEL_DISABLED || level > GUC_LOG_LEVEL_MAX) > > return -EINVAL; > > - mutex_lock(&dev_priv->drm.struct_mutex); > > + mutex_lock(&i915->drm.struct_mutex); > > if (log->level == level) > > goto out_unlock; > > - with_intel_runtime_pm(dev_priv, wakeref) > > + with_intel_runtime_pm(i915, wakeref) > > ret = guc_action_control_log(guc, > > GUC_LOG_LEVEL_IS_VERBOSE(level), > > GUC_LOG_LEVEL_IS_ENABLED(level), > > @@ -537,7 +537,7 @@ int intel_guc_log_set_level(struct intel_guc_log *log, u32 level) > > log->level = level; > > out_unlock: > > - mutex_unlock(&dev_priv->drm.struct_mutex); > > + mutex_unlock(&i915->drm.struct_mutex); > > return ret; > > } > > diff --git a/drivers/gpu/drm/i915/intel_guc_submission.c b/drivers/gpu/drm/i915/intel_guc_submission.c > > index 97f6970d8da8..013d85278270 100644 > > --- a/drivers/gpu/drm/i915/intel_guc_submission.c > > +++ b/drivers/gpu/drm/i915/intel_guc_submission.c > > @@ -203,7 +203,7 @@ static struct guc_doorbell_info *__get_doorbell(struct intel_guc_client *client) > > static bool __doorbell_valid(struct intel_guc *guc, u16 db_id) > > { > > - struct drm_i915_private *dev_priv = guc_to_i915(guc); > > + struct drm_i915_private *i915 = guc_to_i915(guc); > > GEM_BUG_ON(db_id >= GUC_NUM_DOORBELLS); > > return I915_READ(GEN8_DRBREGL(db_id)) & GEN8_DRB_VALID; > > @@ -1014,7 +1014,7 @@ static bool guc_verify_doorbells(struct intel_guc *guc) > > /** > > * guc_client_alloc() - Allocate an intel_guc_client > > - * @dev_priv: driver private data structure > > + * @i915: driver private data structure > > * @engines: The set of engines to enable for this client > > * @priority: four levels priority _CRITICAL, _HIGH, _NORMAL and _LOW > > * The kernel client to replace ExecList submission is created with > > @@ -1026,13 +1026,13 @@ static bool guc_verify_doorbells(struct intel_guc *guc) > > * Return: An intel_guc_client object if success, else NULL. > > */ > > static struct intel_guc_client * > > -guc_client_alloc(struct drm_i915_private *dev_priv, > > +guc_client_alloc(struct drm_i915_private *i915, > > u32 engines, > > u32 priority, > > struct i915_gem_context *ctx) > > { > > struct intel_guc_client *client; > > - struct intel_guc *guc = &dev_priv->guc; > > + struct intel_guc *guc = &i915->guc; > > struct i915_vma *vma; > > void *vaddr; > > int ret; > > @@ -1129,27 +1129,27 @@ static inline bool ctx_save_restore_disabled(struct intel_context *ce) > > static int guc_clients_create(struct intel_guc *guc) > > { > > - struct drm_i915_private *dev_priv = guc_to_i915(guc); > > + struct drm_i915_private *i915 = guc_to_i915(guc); > > struct intel_guc_client *client; > > GEM_BUG_ON(guc->execbuf_client); > > GEM_BUG_ON(guc->preempt_client); > > - client = guc_client_alloc(dev_priv, > > - INTEL_INFO(dev_priv)->engine_mask, > > + client = guc_client_alloc(i915, > > + INTEL_INFO(i915)->engine_mask, > > GUC_CLIENT_PRIORITY_KMD_NORMAL, > > - dev_priv->kernel_context); > > + i915->kernel_context); > > if (IS_ERR(client)) { > > DRM_ERROR("Failed to create GuC client for submission!\n"); > > return PTR_ERR(client); > > } > > guc->execbuf_client = client; > > - if (dev_priv->preempt_context) { > > - client = guc_client_alloc(dev_priv, > > - INTEL_INFO(dev_priv)->engine_mask, > > + if (i915->preempt_context) { > > + client = guc_client_alloc(i915, > > + INTEL_INFO(i915)->engine_mask, > > GUC_CLIENT_PRIORITY_KMD_HIGH, > > - dev_priv->preempt_context); > > + i915->preempt_context); > > if (IS_ERR(client)) { > > DRM_ERROR("Failed to create GuC client for preemption!\n"); > > guc_client_free(guc->execbuf_client); > > @@ -1244,7 +1244,7 @@ static void guc_clients_disable(struct intel_guc *guc) > > */ > > int intel_guc_submission_init(struct intel_guc *guc) > > { > > - struct drm_i915_private *dev_priv = guc_to_i915(guc); > > + struct drm_i915_private *i915 = guc_to_i915(guc); > > struct intel_engine_cs *engine; > > enum intel_engine_id id; > > int ret; > > @@ -1266,7 +1266,7 @@ int intel_guc_submission_init(struct intel_guc *guc) > > if (ret) > > goto err_pool; > > - for_each_engine(engine, dev_priv, id) { > > + for_each_engine(engine, i915, id) { > > guc->preempt_work[id].engine = engine; > > INIT_WORK(&guc->preempt_work[id].work, inject_preempt_context); > > } > > @@ -1280,11 +1280,11 @@ int intel_guc_submission_init(struct intel_guc *guc) > > void intel_guc_submission_fini(struct intel_guc *guc) > > { > > - struct drm_i915_private *dev_priv = guc_to_i915(guc); > > + struct drm_i915_private *i915 = guc_to_i915(guc); > > struct intel_engine_cs *engine; > > enum intel_engine_id id; > > - for_each_engine(engine, dev_priv, id) > > + for_each_engine(engine, i915, id) > > cancel_work_sync(&guc->preempt_work[id].work); > > guc_clients_destroy(guc); > > @@ -1294,9 +1294,9 @@ void intel_guc_submission_fini(struct intel_guc *guc) > > guc_stage_desc_pool_destroy(guc); > > } > > -static void guc_interrupts_capture(struct drm_i915_private *dev_priv) > > +static void guc_interrupts_capture(struct drm_i915_private *i915) > > { > > - struct intel_rps *rps = &dev_priv->gt_pm.rps; > > + struct intel_rps *rps = &i915->gt_pm.rps; > > struct intel_engine_cs *engine; > > enum intel_engine_id id; > > int irqs; > > @@ -1305,7 +1305,7 @@ static void guc_interrupts_capture(struct drm_i915_private *dev_priv) > > * to GuC > > */ > > irqs = _MASKED_BIT_ENABLE(GFX_INTERRUPT_STEERING); > > - for_each_engine(engine, dev_priv, id) > > + for_each_engine(engine, i915, id) > > ENGINE_WRITE(engine, RING_MODE_GEN7, irqs); > > /* route USER_INTERRUPT to Host, all others are sent to GuC. */ > > @@ -1339,9 +1339,9 @@ static void guc_interrupts_capture(struct drm_i915_private *dev_priv) > > rps->pm_intrmsk_mbz &= ~GEN8_PMINTR_DISABLE_REDIRECT_TO_GUC; > > } > > -static void guc_interrupts_release(struct drm_i915_private *dev_priv) > > +static void guc_interrupts_release(struct drm_i915_private *i915) > > { > > - struct intel_rps *rps = &dev_priv->gt_pm.rps; > > + struct intel_rps *rps = &i915->gt_pm.rps; > > struct intel_engine_cs *engine; > > enum intel_engine_id id; > > int irqs; > > @@ -1352,7 +1352,7 @@ static void guc_interrupts_release(struct drm_i915_private *dev_priv) > > */ > > irqs = _MASKED_FIELD(GFX_FORWARD_VBLANK_MASK, GFX_FORWARD_VBLANK_NEVER); > > irqs |= _MASKED_BIT_DISABLE(GFX_INTERRUPT_STEERING); > > - for_each_engine(engine, dev_priv, id) > > + for_each_engine(engine, i915, id) > > ENGINE_WRITE(engine, RING_MODE_GEN7, irqs); > > /* route all GT interrupts to the host */ > > @@ -1408,7 +1408,7 @@ static void guc_set_default_submission(struct intel_engine_cs *engine) > > int intel_guc_submission_enable(struct intel_guc *guc) > > { > > - struct drm_i915_private *dev_priv = guc_to_i915(guc); > > + struct drm_i915_private *i915 = guc_to_i915(guc); > > struct intel_engine_cs *engine; > > enum intel_engine_id id; > > int err; > > @@ -1433,9 +1433,9 @@ int intel_guc_submission_enable(struct intel_guc *guc) > > return err; > > /* Take over from manual control of ELSP (execlists) */ > > - guc_interrupts_capture(dev_priv); > > + guc_interrupts_capture(i915); > > - for_each_engine(engine, dev_priv, id) { > > + for_each_engine(engine, i915, id) { > > engine->set_default_submission = guc_set_default_submission; > > engine->set_default_submission(engine); > > } > > @@ -1445,11 +1445,11 @@ int intel_guc_submission_enable(struct intel_guc *guc) > > void intel_guc_submission_disable(struct intel_guc *guc) > > { > > - struct drm_i915_private *dev_priv = guc_to_i915(guc); > > + struct drm_i915_private *i915 = guc_to_i915(guc); > > - GEM_BUG_ON(dev_priv->gt.awake); /* GT should be parked first */ > > + GEM_BUG_ON(i915->gt.awake); /* GT should be parked first */ > > - guc_interrupts_release(dev_priv); > > + guc_interrupts_release(i915); > > guc_clients_disable(guc); > > } > > diff --git a/drivers/gpu/drm/i915/intel_gvt.c b/drivers/gpu/drm/i915/intel_gvt.c > > index 1d7d26e4cf14..b50d31a2e095 100644 > > --- a/drivers/gpu/drm/i915/intel_gvt.c > > +++ b/drivers/gpu/drm/i915/intel_gvt.c > > @@ -39,17 +39,17 @@ > > * doc is available on https://01.org/group/2230/documentation-list. > > */ > > -static bool is_supported_device(struct drm_i915_private *dev_priv) > > +static bool is_supported_device(struct drm_i915_private *i915) > > { > > - if (IS_BROADWELL(dev_priv)) > > + if (IS_BROADWELL(i915)) > > return true; > > - if (IS_SKYLAKE(dev_priv)) > > + if (IS_SKYLAKE(i915)) > > return true; > > - if (IS_KABYLAKE(dev_priv)) > > + if (IS_KABYLAKE(i915)) > > return true; > > - if (IS_BROXTON(dev_priv)) > > + if (IS_BROXTON(i915)) > > return true; > > - if (IS_COFFEELAKE(dev_priv)) > > + if (IS_COFFEELAKE(i915)) > > return true; > > return false; > > @@ -57,21 +57,21 @@ static bool is_supported_device(struct drm_i915_private *dev_priv) > > /** > > * intel_gvt_sanitize_options - sanitize GVT related options > > - * @dev_priv: drm i915 private data > > + * @i915: drm i915 private data > > * > > * This function is called at the i915 options sanitize stage. > > */ > > -void intel_gvt_sanitize_options(struct drm_i915_private *dev_priv) > > +void intel_gvt_sanitize_options(struct drm_i915_private *i915) > > { > > if (!i915_modparams.enable_gvt) > > return; > > - if (intel_vgpu_active(dev_priv)) { > > + if (intel_vgpu_active(i915)) { > > DRM_INFO("GVT-g is disabled for guest\n"); > > goto bail; > > } > > - if (!is_supported_device(dev_priv)) { > > + if (!is_supported_device(i915)) { > > DRM_INFO("Unsupported device. GVT-g is disabled\n"); > > goto bail; > > } > > @@ -83,7 +83,7 @@ void intel_gvt_sanitize_options(struct drm_i915_private *dev_priv) > > /** > > * intel_gvt_init - initialize GVT components > > - * @dev_priv: drm i915 private data > > + * @i915: drm i915 private data > > * > > * This function is called at the initialization stage to create a GVT device. > > * > > @@ -91,7 +91,7 @@ void intel_gvt_sanitize_options(struct drm_i915_private *dev_priv) > > * Zero on success, negative error code if failed. > > * > > */ > > -int intel_gvt_init(struct drm_i915_private *dev_priv) > > +int intel_gvt_init(struct drm_i915_private *i915) > > { > > int ret; > > @@ -103,12 +103,12 @@ int intel_gvt_init(struct drm_i915_private *dev_priv) > > return 0; > > } > > - if (USES_GUC_SUBMISSION(dev_priv)) { > > + if (USES_GUC_SUBMISSION(i915)) { > > DRM_ERROR("i915 GVT-g loading failed due to Graphics virtualization is not yet supported with GuC submission\n"); > > return -EIO; > > } > > - ret = intel_gvt_init_device(dev_priv); > > + ret = intel_gvt_init_device(i915); > > if (ret) { > > DRM_DEBUG_DRIVER("Fail to init GVT device\n"); > > goto bail; > > @@ -123,15 +123,15 @@ int intel_gvt_init(struct drm_i915_private *dev_priv) > > /** > > * intel_gvt_cleanup - cleanup GVT components when i915 driver is unloading > > - * @dev_priv: drm i915 private * > > + * @i915: drm i915 private * > > * > > * This function is called at the i915 driver unloading stage, to shutdown > > * GVT components and release the related resources. > > */ > > -void intel_gvt_cleanup(struct drm_i915_private *dev_priv) > > +void intel_gvt_cleanup(struct drm_i915_private *i915) > > { > > - if (!intel_gvt_active(dev_priv)) > > + if (!intel_gvt_active(i915)) > > return; > > - intel_gvt_clean_device(dev_priv); > > + intel_gvt_clean_device(i915); > > } > > diff --git a/drivers/gpu/drm/i915/intel_gvt.h b/drivers/gpu/drm/i915/intel_gvt.h > > index 61b246470282..6f81e26cb9ad 100644 > > --- a/drivers/gpu/drm/i915/intel_gvt.h > > +++ b/drivers/gpu/drm/i915/intel_gvt.h > > @@ -27,22 +27,22 @@ > > struct intel_gvt; > > #ifdef CONFIG_DRM_I915_GVT > > -int intel_gvt_init(struct drm_i915_private *dev_priv); > > -void intel_gvt_cleanup(struct drm_i915_private *dev_priv); > > -int intel_gvt_init_device(struct drm_i915_private *dev_priv); > > -void intel_gvt_clean_device(struct drm_i915_private *dev_priv); > > +int intel_gvt_init(struct drm_i915_private *i915); > > +void intel_gvt_cleanup(struct drm_i915_private *i915); > > +int intel_gvt_init_device(struct drm_i915_private *i915); > > +void intel_gvt_clean_device(struct drm_i915_private *i915); > > int intel_gvt_init_host(void); > > -void intel_gvt_sanitize_options(struct drm_i915_private *dev_priv); > > +void intel_gvt_sanitize_options(struct drm_i915_private *i915); > > #else > > -static inline int intel_gvt_init(struct drm_i915_private *dev_priv) > > +static inline int intel_gvt_init(struct drm_i915_private *i915) > > { > > return 0; > > } > > -static inline void intel_gvt_cleanup(struct drm_i915_private *dev_priv) > > +static inline void intel_gvt_cleanup(struct drm_i915_private *i915) > > { > > } > > -static inline void intel_gvt_sanitize_options(struct drm_i915_private *dev_priv) > > +static inline void intel_gvt_sanitize_options(struct drm_i915_private *i915) > > { > > } > > #endif > > diff --git a/drivers/gpu/drm/i915/intel_hdcp.c b/drivers/gpu/drm/i915/intel_hdcp.c > > index bc3a94d491c4..7df5f19b52d5 100644 > > --- a/drivers/gpu/drm/i915/intel_hdcp.c > > +++ b/drivers/gpu/drm/i915/intel_hdcp.c > > @@ -81,7 +81,7 @@ bool intel_hdcp_capable(struct intel_connector *connector) > > /* Is HDCP2.2 capable on Platform and Sink */ > > bool intel_hdcp2_capable(struct intel_connector *connector) > > { > > - struct drm_i915_private *dev_priv = to_i915(connector->base.dev); > > + struct drm_i915_private *i915 = to_i915(connector->base.dev); > > struct intel_digital_port *intel_dig_port = conn_to_dig_port(connector); > > struct intel_hdcp *hdcp = &connector->hdcp; > > bool capable = false; > > @@ -91,12 +91,12 @@ bool intel_hdcp2_capable(struct intel_connector *connector) > > return false; > > /* MEI interface is solid */ > > - mutex_lock(&dev_priv->hdcp_comp_mutex); > > - if (!dev_priv->hdcp_comp_added || !dev_priv->hdcp_master) { > > - mutex_unlock(&dev_priv->hdcp_comp_mutex); > > + mutex_lock(&i915->hdcp_comp_mutex); > > + if (!i915->hdcp_comp_added || !i915->hdcp_master) { > > + mutex_unlock(&i915->hdcp_comp_mutex); > > return false; > > } > > - mutex_unlock(&dev_priv->hdcp_comp_mutex); > > + mutex_unlock(&i915->hdcp_comp_mutex); > > /* Sink's capability for HDCP2.2 */ > > hdcp->shim->hdcp_2_2_capable(intel_dig_port, &capable); > > @@ -106,7 +106,7 @@ bool intel_hdcp2_capable(struct intel_connector *connector) > > static inline bool intel_hdcp_in_use(struct intel_connector *connector) > > { > > - struct drm_i915_private *dev_priv = to_i915(connector->base.dev); > > + struct drm_i915_private *i915 = to_i915(connector->base.dev); > > enum port port = connector->encoder->port; > > u32 reg; > > @@ -116,7 +116,7 @@ static inline bool intel_hdcp_in_use(struct intel_connector *connector) > > static inline bool intel_hdcp2_in_use(struct intel_connector *connector) > > { > > - struct drm_i915_private *dev_priv = to_i915(connector->base.dev); > > + struct drm_i915_private *i915 = to_i915(connector->base.dev); > > enum port port = connector->encoder->port; > > u32 reg; > > @@ -145,9 +145,9 @@ static int intel_hdcp_poll_ksv_fifo(struct intel_digital_port *intel_dig_port, > > return 0; > > } > > -static bool hdcp_key_loadable(struct drm_i915_private *dev_priv) > > +static bool hdcp_key_loadable(struct drm_i915_private *i915) > > { > > - struct i915_power_domains *power_domains = &dev_priv->power_domains; > > + struct i915_power_domains *power_domains = &i915->power_domains; > > struct i915_power_well *power_well; > > enum i915_power_well_id id; > > bool enabled = false; > > @@ -156,7 +156,7 @@ static bool hdcp_key_loadable(struct drm_i915_private *dev_priv) > > * On HSW and BDW, Display HW loads the Key as soon as Display resumes. > > * On all BXT+, SW can load the keys only when the PW#1 is turned on. > > */ > > - if (IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv)) > > + if (IS_HASWELL(i915) || IS_BROADWELL(i915)) > > id = HSW_DISP_PW_GLOBAL; > > else > > id = SKL_DISP_PW_1; > > @@ -164,9 +164,9 @@ static bool hdcp_key_loadable(struct drm_i915_private *dev_priv) > > mutex_lock(&power_domains->lock); > > /* PG1 (power well #1) needs to be enabled */ > > - for_each_power_well(dev_priv, power_well) { > > + for_each_power_well(i915, power_well) { > > if (power_well->desc->id == id) { > > - enabled = power_well->desc->ops->is_enabled(dev_priv, > > + enabled = power_well->desc->ops->is_enabled(i915, > > power_well); > > break; > > } > > @@ -182,14 +182,14 @@ static bool hdcp_key_loadable(struct drm_i915_private *dev_priv) > > return enabled; > > } > > -static void intel_hdcp_clear_keys(struct drm_i915_private *dev_priv) > > +static void intel_hdcp_clear_keys(struct drm_i915_private *i915) > > { > > I915_WRITE(HDCP_KEY_CONF, HDCP_CLEAR_KEYS_TRIGGER); > > I915_WRITE(HDCP_KEY_STATUS, HDCP_KEY_LOAD_DONE | HDCP_KEY_LOAD_STATUS | > > HDCP_FUSE_IN_PROGRESS | HDCP_FUSE_ERROR | HDCP_FUSE_DONE); > > } > > -static int intel_hdcp_load_keys(struct drm_i915_private *dev_priv) > > +static int intel_hdcp_load_keys(struct drm_i915_private *i915) > > { > > int ret; > > u32 val; > > @@ -202,7 +202,7 @@ static int intel_hdcp_load_keys(struct drm_i915_private *dev_priv) > > * On HSW and BDW HW loads the HDCP1.4 Key when Display comes > > * out of reset. So if Key is not already loaded, its an error state. > > */ > > - if (IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv)) > > + if (IS_HASWELL(i915) || IS_BROADWELL(i915)) > > if (!(I915_READ(HDCP_KEY_STATUS) & HDCP_KEY_LOAD_DONE)) > > return -ENXIO; > > @@ -213,8 +213,8 @@ static int intel_hdcp_load_keys(struct drm_i915_private *dev_priv) > > * platforms except BXT and GLK, differ in the key load trigger process > > * from other platforms. So GEN9_BC uses the GT Driver Mailbox i/f. > > */ > > - if (IS_GEN9_BC(dev_priv)) { > > - ret = sandybridge_pcode_write(dev_priv, > > + if (IS_GEN9_BC(i915)) { > > + ret = sandybridge_pcode_write(i915, > > SKL_PCODE_LOAD_HDCP_KEYS, 1); > > if (ret) { > > DRM_ERROR("Failed to initiate HDCP key load (%d)\n", > > @@ -226,7 +226,7 @@ static int intel_hdcp_load_keys(struct drm_i915_private *dev_priv) > > } > > /* Wait for the keys to load (500us) */ > > - ret = __intel_wait_for_register(&dev_priv->uncore, HDCP_KEY_STATUS, > > + ret = __intel_wait_for_register(&i915->uncore, HDCP_KEY_STATUS, > > HDCP_KEY_LOAD_DONE, HDCP_KEY_LOAD_DONE, > > 10, 1, &val); > > if (ret) > > @@ -241,10 +241,10 @@ static int intel_hdcp_load_keys(struct drm_i915_private *dev_priv) > > } > > /* Returns updated SHA-1 index */ > > -static int intel_write_sha_text(struct drm_i915_private *dev_priv, u32 sha_text) > > +static int intel_write_sha_text(struct drm_i915_private *i915, u32 sha_text) > > { > > I915_WRITE(HDCP_SHA_TEXT, sha_text); > > - if (intel_wait_for_register(&dev_priv->uncore, HDCP_REP_CTL, > > + if (intel_wait_for_register(&i915->uncore, HDCP_REP_CTL, > > HDCP_SHA1_READY, HDCP_SHA1_READY, 1)) { > > DRM_ERROR("Timed out waiting for SHA1 ready\n"); > > return -ETIMEDOUT; > > @@ -279,11 +279,11 @@ int intel_hdcp_validate_v_prime(struct intel_digital_port *intel_dig_port, > > const struct intel_hdcp_shim *shim, > > u8 *ksv_fifo, u8 num_downstream, u8 *bstatus) > > { > > - struct drm_i915_private *dev_priv; > > + struct drm_i915_private *i915; > > u32 vprime, sha_text, sha_leftovers, rep_ctl; > > int ret, i, j, sha_idx; > > - dev_priv = intel_dig_port->base.base.dev->dev_private; > > + i915 = intel_dig_port->base.base.dev->dev_private; > > /* Process V' values from the receiver */ > > for (i = 0; i < DRM_HDCP_V_PRIME_NUM_PARTS; i++) { > > @@ -317,7 +317,7 @@ int intel_hdcp_validate_v_prime(struct intel_digital_port *intel_dig_port, > > for (j = 0; j < sha_empty; j++) > > sha_text |= ksv[j] << ((sizeof(sha_text) - j - 1) * 8); > > - ret = intel_write_sha_text(dev_priv, sha_text); > > + ret = intel_write_sha_text(i915, sha_text); > > if (ret < 0) > > return ret; > > @@ -340,7 +340,7 @@ int intel_hdcp_validate_v_prime(struct intel_digital_port *intel_dig_port, > > if (sizeof(sha_text) > sha_leftovers) > > continue; > > - ret = intel_write_sha_text(dev_priv, sha_text); > > + ret = intel_write_sha_text(i915, sha_text); > > if (ret < 0) > > return ret; > > sha_leftovers = 0; > > @@ -357,7 +357,7 @@ int intel_hdcp_validate_v_prime(struct intel_digital_port *intel_dig_port, > > if (sha_leftovers == 0) { > > /* Write 16 bits of text, 16 bits of M0 */ > > I915_WRITE(HDCP_REP_CTL, rep_ctl | HDCP_SHA1_TEXT_16); > > - ret = intel_write_sha_text(dev_priv, > > + ret = intel_write_sha_text(i915, > > bstatus[0] << 8 | bstatus[1]); > > if (ret < 0) > > return ret; > > @@ -365,14 +365,14 @@ int intel_hdcp_validate_v_prime(struct intel_digital_port *intel_dig_port, > > /* Write 32 bits of M0 */ > > I915_WRITE(HDCP_REP_CTL, rep_ctl | HDCP_SHA1_TEXT_0); > > - ret = intel_write_sha_text(dev_priv, 0); > > + ret = intel_write_sha_text(i915, 0); > > if (ret < 0) > > return ret; > > sha_idx += sizeof(sha_text); > > /* Write 16 bits of M0 */ > > I915_WRITE(HDCP_REP_CTL, rep_ctl | HDCP_SHA1_TEXT_16); > > - ret = intel_write_sha_text(dev_priv, 0); > > + ret = intel_write_sha_text(i915, 0); > > if (ret < 0) > > return ret; > > sha_idx += sizeof(sha_text); > > @@ -383,21 +383,21 @@ int intel_hdcp_validate_v_prime(struct intel_digital_port *intel_dig_port, > > sha_text |= bstatus[0] << 16 | bstatus[1] << 8; > > /* Only 24-bits of data, must be in the LSB */ > > sha_text = (sha_text & 0xffffff00) >> 8; > > - ret = intel_write_sha_text(dev_priv, sha_text); > > + ret = intel_write_sha_text(i915, sha_text); > > if (ret < 0) > > return ret; > > sha_idx += sizeof(sha_text); > > /* Write 32 bits of M0 */ > > I915_WRITE(HDCP_REP_CTL, rep_ctl | HDCP_SHA1_TEXT_0); > > - ret = intel_write_sha_text(dev_priv, 0); > > + ret = intel_write_sha_text(i915, 0); > > if (ret < 0) > > return ret; > > sha_idx += sizeof(sha_text); > > /* Write 24 bits of M0 */ > > I915_WRITE(HDCP_REP_CTL, rep_ctl | HDCP_SHA1_TEXT_8); > > - ret = intel_write_sha_text(dev_priv, 0); > > + ret = intel_write_sha_text(i915, 0); > > if (ret < 0) > > return ret; > > sha_idx += sizeof(sha_text); > > @@ -406,7 +406,7 @@ int intel_hdcp_validate_v_prime(struct intel_digital_port *intel_dig_port, > > /* Write 32 bits of text */ > > I915_WRITE(HDCP_REP_CTL, rep_ctl | HDCP_SHA1_TEXT_32); > > sha_text |= bstatus[0] << 24 | bstatus[1] << 16; > > - ret = intel_write_sha_text(dev_priv, sha_text); > > + ret = intel_write_sha_text(i915, sha_text); > > if (ret < 0) > > return ret; > > sha_idx += sizeof(sha_text); > > @@ -414,7 +414,7 @@ int intel_hdcp_validate_v_prime(struct intel_digital_port *intel_dig_port, > > /* Write 64 bits of M0 */ > > I915_WRITE(HDCP_REP_CTL, rep_ctl | HDCP_SHA1_TEXT_0); > > for (i = 0; i < 2; i++) { > > - ret = intel_write_sha_text(dev_priv, 0); > > + ret = intel_write_sha_text(i915, 0); > > if (ret < 0) > > return ret; > > sha_idx += sizeof(sha_text); > > @@ -423,28 +423,28 @@ int intel_hdcp_validate_v_prime(struct intel_digital_port *intel_dig_port, > > /* Write 32 bits of text */ > > I915_WRITE(HDCP_REP_CTL, rep_ctl | HDCP_SHA1_TEXT_32); > > sha_text |= bstatus[0] << 24; > > - ret = intel_write_sha_text(dev_priv, sha_text); > > + ret = intel_write_sha_text(i915, sha_text); > > if (ret < 0) > > return ret; > > sha_idx += sizeof(sha_text); > > /* Write 8 bits of text, 24 bits of M0 */ > > I915_WRITE(HDCP_REP_CTL, rep_ctl | HDCP_SHA1_TEXT_8); > > - ret = intel_write_sha_text(dev_priv, bstatus[1]); > > + ret = intel_write_sha_text(i915, bstatus[1]); > > if (ret < 0) > > return ret; > > sha_idx += sizeof(sha_text); > > /* Write 32 bits of M0 */ > > I915_WRITE(HDCP_REP_CTL, rep_ctl | HDCP_SHA1_TEXT_0); > > - ret = intel_write_sha_text(dev_priv, 0); > > + ret = intel_write_sha_text(i915, 0); > > if (ret < 0) > > return ret; > > sha_idx += sizeof(sha_text); > > /* Write 8 bits of M0 */ > > I915_WRITE(HDCP_REP_CTL, rep_ctl | HDCP_SHA1_TEXT_24); > > - ret = intel_write_sha_text(dev_priv, 0); > > + ret = intel_write_sha_text(i915, 0); > > if (ret < 0) > > return ret; > > sha_idx += sizeof(sha_text); > > @@ -457,7 +457,7 @@ int intel_hdcp_validate_v_prime(struct intel_digital_port *intel_dig_port, > > I915_WRITE(HDCP_REP_CTL, rep_ctl | HDCP_SHA1_TEXT_32); > > /* Fill up to 64-4 bytes with zeros (leave the last write for length) */ > > while ((sha_idx % 64) < (64 - sizeof(sha_text))) { > > - ret = intel_write_sha_text(dev_priv, 0); > > + ret = intel_write_sha_text(i915, 0); > > if (ret < 0) > > return ret; > > sha_idx += sizeof(sha_text); > > @@ -469,13 +469,13 @@ int intel_hdcp_validate_v_prime(struct intel_digital_port *intel_dig_port, > > * - 10 bytes for BINFO/BSTATUS(2), M0(8) > > */ > > sha_text = (num_downstream * 5 + 10) * 8; > > - ret = intel_write_sha_text(dev_priv, sha_text); > > + ret = intel_write_sha_text(i915, sha_text); > > if (ret < 0) > > return ret; > > /* Tell the HW we're done with the hash and wait for it to ACK */ > > I915_WRITE(HDCP_REP_CTL, rep_ctl | HDCP_SHA1_COMPLETE_HASH); > > - if (intel_wait_for_register(&dev_priv->uncore, HDCP_REP_CTL, > > + if (intel_wait_for_register(&i915->uncore, HDCP_REP_CTL, > > HDCP_SHA1_COMPLETE, > > HDCP_SHA1_COMPLETE, 1)) { > > DRM_ERROR("Timed out waiting for SHA1 complete\n"); > > @@ -571,7 +571,7 @@ static int intel_hdcp_auth(struct intel_connector *connector) > > struct intel_hdcp *hdcp = &connector->hdcp; > > struct drm_device *dev = connector->base.dev; > > const struct intel_hdcp_shim *shim = hdcp->shim; > > - struct drm_i915_private *dev_priv; > > + struct drm_i915_private *i915; > > enum port port; > > unsigned long r0_prime_gen_start; > > int ret, i, tries = 2; > > @@ -589,7 +589,7 @@ static int intel_hdcp_auth(struct intel_connector *connector) > > } ri; > > bool repeater_present, hdcp_capable; > > - dev_priv = intel_dig_port->base.base.dev->dev_private; > > + i915 = intel_dig_port->base.base.dev->dev_private; > > port = intel_dig_port->base.port; > > @@ -615,7 +615,7 @@ static int intel_hdcp_auth(struct intel_connector *connector) > > I915_WRITE(PORT_HDCP_CONF(port), HDCP_CONF_CAPTURE_AN); > > /* Wait for An to be acquired */ > > - if (intel_wait_for_register(&dev_priv->uncore, PORT_HDCP_STATUS(port), > > + if (intel_wait_for_register(&i915->uncore, PORT_HDCP_STATUS(port), > > HDCP_STATUS_AN_READY, > > HDCP_STATUS_AN_READY, 1)) { > > DRM_ERROR("Timed out waiting for An\n"); > > @@ -701,7 +701,7 @@ static int intel_hdcp_auth(struct intel_connector *connector) > > } > > /* Wait for encryption confirmation */ > > - if (intel_wait_for_register(&dev_priv->uncore, PORT_HDCP_STATUS(port), > > + if (intel_wait_for_register(&i915->uncore, PORT_HDCP_STATUS(port), > > HDCP_STATUS_ENC, HDCP_STATUS_ENC, > > ENCRYPT_STATUS_CHANGE_TIMEOUT_MS)) { > > DRM_ERROR("Timed out waiting for encryption\n"); > > @@ -723,7 +723,7 @@ static int intel_hdcp_auth(struct intel_connector *connector) > > static int _intel_hdcp_disable(struct intel_connector *connector) > > { > > struct intel_hdcp *hdcp = &connector->hdcp; > > - struct drm_i915_private *dev_priv = connector->base.dev->dev_private; > > + struct drm_i915_private *i915 = connector->base.dev->dev_private; > > struct intel_digital_port *intel_dig_port = conn_to_dig_port(connector); > > enum port port = intel_dig_port->base.port; > > int ret; > > @@ -733,7 +733,7 @@ static int _intel_hdcp_disable(struct intel_connector *connector) > > hdcp->hdcp_encrypted = false; > > I915_WRITE(PORT_HDCP_CONF(port), 0); > > - if (intel_wait_for_register(&dev_priv->uncore, > > + if (intel_wait_for_register(&i915->uncore, > > PORT_HDCP_STATUS(port), ~0, 0, > > ENCRYPT_STATUS_CHANGE_TIMEOUT_MS)) { > > DRM_ERROR("Failed to disable HDCP, timeout clearing status\n"); > > @@ -753,22 +753,22 @@ static int _intel_hdcp_disable(struct intel_connector *connector) > > static int _intel_hdcp_enable(struct intel_connector *connector) > > { > > struct intel_hdcp *hdcp = &connector->hdcp; > > - struct drm_i915_private *dev_priv = connector->base.dev->dev_private; > > + struct drm_i915_private *i915 = connector->base.dev->dev_private; > > int i, ret, tries = 3; > > DRM_DEBUG_KMS("[%s:%d] HDCP is being enabled...\n", > > connector->base.name, connector->base.base.id); > > - if (!hdcp_key_loadable(dev_priv)) { > > + if (!hdcp_key_loadable(i915)) { > > DRM_ERROR("HDCP key Load is not possible\n"); > > return -ENXIO; > > } > > for (i = 0; i < KEY_LOAD_TRIES; i++) { > > - ret = intel_hdcp_load_keys(dev_priv); > > + ret = intel_hdcp_load_keys(i915); > > if (!ret) > > break; > > - intel_hdcp_clear_keys(dev_priv); > > + intel_hdcp_clear_keys(i915); > > } > > if (ret) { > > DRM_ERROR("Could not load HDCP keys, (%d)\n", ret); > > @@ -803,7 +803,7 @@ struct intel_connector *intel_hdcp_to_connector(struct intel_hdcp *hdcp) > > static int intel_hdcp_check_link(struct intel_connector *connector) > > { > > struct intel_hdcp *hdcp = &connector->hdcp; > > - struct drm_i915_private *dev_priv = connector->base.dev->dev_private; > > + struct drm_i915_private *i915 = connector->base.dev->dev_private; > > struct intel_digital_port *intel_dig_port = conn_to_dig_port(connector); > > enum port port = intel_dig_port->base.port; > > int ret = 0; > > @@ -884,10 +884,10 @@ static void intel_hdcp_prop_work(struct work_struct *work) > > drm_modeset_unlock(&dev->mode_config.connection_mutex); > > } > > -bool is_hdcp_supported(struct drm_i915_private *dev_priv, enum port port) > > +bool is_hdcp_supported(struct drm_i915_private *i915, enum port port) > > { > > /* PORT E doesn't have HDCP, and PORT F is disabled */ > > - return INTEL_GEN(dev_priv) >= 9 && port < PORT_E; > > + return INTEL_GEN(i915) >= 9 && port < PORT_E; > > } > > static int > > @@ -895,22 +895,22 @@ hdcp2_prepare_ake_init(struct intel_connector *connector, > > struct hdcp2_ake_init *ake_data) > > { > > struct hdcp_port_data *data = &connector->hdcp.port_data; > > - struct drm_i915_private *dev_priv = to_i915(connector->base.dev); > > + struct drm_i915_private *i915 = to_i915(connector->base.dev); > > struct i915_hdcp_comp_master *comp; > > int ret; > > - mutex_lock(&dev_priv->hdcp_comp_mutex); > > - comp = dev_priv->hdcp_master; > > + mutex_lock(&i915->hdcp_comp_mutex); > > + comp = i915->hdcp_master; > > if (!comp || !comp->ops) { > > - mutex_unlock(&dev_priv->hdcp_comp_mutex); > > + mutex_unlock(&i915->hdcp_comp_mutex); > > return -EINVAL; > > } > > ret = comp->ops->initiate_hdcp2_session(comp->mei_dev, data, ake_data); > > if (ret) > > DRM_DEBUG_KMS("Prepare_ake_init failed. %d\n", ret); > > - mutex_unlock(&dev_priv->hdcp_comp_mutex); > > + mutex_unlock(&i915->hdcp_comp_mutex); > > return ret; > > } > > @@ -923,15 +923,15 @@ hdcp2_verify_rx_cert_prepare_km(struct intel_connector *connector, > > size_t *msg_sz) > > { > > struct hdcp_port_data *data = &connector->hdcp.port_data; > > - struct drm_i915_private *dev_priv = to_i915(connector->base.dev); > > + struct drm_i915_private *i915 = to_i915(connector->base.dev); > > struct i915_hdcp_comp_master *comp; > > int ret; > > - mutex_lock(&dev_priv->hdcp_comp_mutex); > > - comp = dev_priv->hdcp_master; > > + mutex_lock(&i915->hdcp_comp_mutex); > > + comp = i915->hdcp_master; > > if (!comp || !comp->ops) { > > - mutex_unlock(&dev_priv->hdcp_comp_mutex); > > + mutex_unlock(&i915->hdcp_comp_mutex); > > return -EINVAL; > > } > > @@ -940,7 +940,7 @@ hdcp2_verify_rx_cert_prepare_km(struct intel_connector *connector, > > ek_pub_km, msg_sz); > > if (ret < 0) > > DRM_DEBUG_KMS("Verify rx_cert failed. %d\n", ret); > > - mutex_unlock(&dev_priv->hdcp_comp_mutex); > > + mutex_unlock(&i915->hdcp_comp_mutex); > > return ret; > > } > > @@ -949,22 +949,22 @@ static int hdcp2_verify_hprime(struct intel_connector *connector, > > struct hdcp2_ake_send_hprime *rx_hprime) > > { > > struct hdcp_port_data *data = &connector->hdcp.port_data; > > - struct drm_i915_private *dev_priv = to_i915(connector->base.dev); > > + struct drm_i915_private *i915 = to_i915(connector->base.dev); > > struct i915_hdcp_comp_master *comp; > > int ret; > > - mutex_lock(&dev_priv->hdcp_comp_mutex); > > - comp = dev_priv->hdcp_master; > > + mutex_lock(&i915->hdcp_comp_mutex); > > + comp = i915->hdcp_master; > > if (!comp || !comp->ops) { > > - mutex_unlock(&dev_priv->hdcp_comp_mutex); > > + mutex_unlock(&i915->hdcp_comp_mutex); > > return -EINVAL; > > } > > ret = comp->ops->verify_hprime(comp->mei_dev, data, rx_hprime); > > if (ret < 0) > > DRM_DEBUG_KMS("Verify hprime failed. %d\n", ret); > > - mutex_unlock(&dev_priv->hdcp_comp_mutex); > > + mutex_unlock(&i915->hdcp_comp_mutex); > > return ret; > > } > > @@ -974,22 +974,22 @@ hdcp2_store_pairing_info(struct intel_connector *connector, > > struct hdcp2_ake_send_pairing_info *pairing_info) > > { > > struct hdcp_port_data *data = &connector->hdcp.port_data; > > - struct drm_i915_private *dev_priv = to_i915(connector->base.dev); > > + struct drm_i915_private *i915 = to_i915(connector->base.dev); > > struct i915_hdcp_comp_master *comp; > > int ret; > > - mutex_lock(&dev_priv->hdcp_comp_mutex); > > - comp = dev_priv->hdcp_master; > > + mutex_lock(&i915->hdcp_comp_mutex); > > + comp = i915->hdcp_master; > > if (!comp || !comp->ops) { > > - mutex_unlock(&dev_priv->hdcp_comp_mutex); > > + mutex_unlock(&i915->hdcp_comp_mutex); > > return -EINVAL; > > } > > ret = comp->ops->store_pairing_info(comp->mei_dev, data, pairing_info); > > if (ret < 0) > > DRM_DEBUG_KMS("Store pairing info failed. %d\n", ret); > > - mutex_unlock(&dev_priv->hdcp_comp_mutex); > > + mutex_unlock(&i915->hdcp_comp_mutex); > > return ret; > > } > > @@ -999,22 +999,22 @@ hdcp2_prepare_lc_init(struct intel_connector *connector, > > struct hdcp2_lc_init *lc_init) > > { > > struct hdcp_port_data *data = &connector->hdcp.port_data; > > - struct drm_i915_private *dev_priv = to_i915(connector->base.dev); > > + struct drm_i915_private *i915 = to_i915(connector->base.dev); > > struct i915_hdcp_comp_master *comp; > > int ret; > > - mutex_lock(&dev_priv->hdcp_comp_mutex); > > - comp = dev_priv->hdcp_master; > > + mutex_lock(&i915->hdcp_comp_mutex); > > + comp = i915->hdcp_master; > > if (!comp || !comp->ops) { > > - mutex_unlock(&dev_priv->hdcp_comp_mutex); > > + mutex_unlock(&i915->hdcp_comp_mutex); > > return -EINVAL; > > } > > ret = comp->ops->initiate_locality_check(comp->mei_dev, data, lc_init); > > if (ret < 0) > > DRM_DEBUG_KMS("Prepare lc_init failed. %d\n", ret); > > - mutex_unlock(&dev_priv->hdcp_comp_mutex); > > + mutex_unlock(&i915->hdcp_comp_mutex); > > return ret; > > } > > @@ -1024,22 +1024,22 @@ hdcp2_verify_lprime(struct intel_connector *connector, > > struct hdcp2_lc_send_lprime *rx_lprime) > > { > > struct hdcp_port_data *data = &connector->hdcp.port_data; > > - struct drm_i915_private *dev_priv = to_i915(connector->base.dev); > > + struct drm_i915_private *i915 = to_i915(connector->base.dev); > > struct i915_hdcp_comp_master *comp; > > int ret; > > - mutex_lock(&dev_priv->hdcp_comp_mutex); > > - comp = dev_priv->hdcp_master; > > + mutex_lock(&i915->hdcp_comp_mutex); > > + comp = i915->hdcp_master; > > if (!comp || !comp->ops) { > > - mutex_unlock(&dev_priv->hdcp_comp_mutex); > > + mutex_unlock(&i915->hdcp_comp_mutex); > > return -EINVAL; > > } > > ret = comp->ops->verify_lprime(comp->mei_dev, data, rx_lprime); > > if (ret < 0) > > DRM_DEBUG_KMS("Verify L_Prime failed. %d\n", ret); > > - mutex_unlock(&dev_priv->hdcp_comp_mutex); > > + mutex_unlock(&i915->hdcp_comp_mutex); > > return ret; > > } > > @@ -1048,22 +1048,22 @@ static int hdcp2_prepare_skey(struct intel_connector *connector, > > struct hdcp2_ske_send_eks *ske_data) > > { > > struct hdcp_port_data *data = &connector->hdcp.port_data; > > - struct drm_i915_private *dev_priv = to_i915(connector->base.dev); > > + struct drm_i915_private *i915 = to_i915(connector->base.dev); > > struct i915_hdcp_comp_master *comp; > > int ret; > > - mutex_lock(&dev_priv->hdcp_comp_mutex); > > - comp = dev_priv->hdcp_master; > > + mutex_lock(&i915->hdcp_comp_mutex); > > + comp = i915->hdcp_master; > > if (!comp || !comp->ops) { > > - mutex_unlock(&dev_priv->hdcp_comp_mutex); > > + mutex_unlock(&i915->hdcp_comp_mutex); > > return -EINVAL; > > } > > ret = comp->ops->get_session_key(comp->mei_dev, data, ske_data); > > if (ret < 0) > > DRM_DEBUG_KMS("Get session key failed. %d\n", ret); > > - mutex_unlock(&dev_priv->hdcp_comp_mutex); > > + mutex_unlock(&i915->hdcp_comp_mutex); > > return ret; > > } > > @@ -1075,15 +1075,15 @@ hdcp2_verify_rep_topology_prepare_ack(struct intel_connector *connector, > > struct hdcp2_rep_send_ack *rep_send_ack) > > { > > struct hdcp_port_data *data = &connector->hdcp.port_data; > > - struct drm_i915_private *dev_priv = to_i915(connector->base.dev); > > + struct drm_i915_private *i915 = to_i915(connector->base.dev); > > struct i915_hdcp_comp_master *comp; > > int ret; > > - mutex_lock(&dev_priv->hdcp_comp_mutex); > > - comp = dev_priv->hdcp_master; > > + mutex_lock(&i915->hdcp_comp_mutex); > > + comp = i915->hdcp_master; > > if (!comp || !comp->ops) { > > - mutex_unlock(&dev_priv->hdcp_comp_mutex); > > + mutex_unlock(&i915->hdcp_comp_mutex); > > return -EINVAL; > > } > > @@ -1092,7 +1092,7 @@ hdcp2_verify_rep_topology_prepare_ack(struct intel_connector *connector, > > rep_send_ack); > > if (ret < 0) > > DRM_DEBUG_KMS("Verify rep topology failed. %d\n", ret); > > - mutex_unlock(&dev_priv->hdcp_comp_mutex); > > + mutex_unlock(&i915->hdcp_comp_mutex); > > return ret; > > } > > @@ -1102,22 +1102,22 @@ hdcp2_verify_mprime(struct intel_connector *connector, > > struct hdcp2_rep_stream_ready *stream_ready) > > { > > struct hdcp_port_data *data = &connector->hdcp.port_data; > > - struct drm_i915_private *dev_priv = to_i915(connector->base.dev); > > + struct drm_i915_private *i915 = to_i915(connector->base.dev); > > struct i915_hdcp_comp_master *comp; > > int ret; > > - mutex_lock(&dev_priv->hdcp_comp_mutex); > > - comp = dev_priv->hdcp_master; > > + mutex_lock(&i915->hdcp_comp_mutex); > > + comp = i915->hdcp_master; > > if (!comp || !comp->ops) { > > - mutex_unlock(&dev_priv->hdcp_comp_mutex); > > + mutex_unlock(&i915->hdcp_comp_mutex); > > return -EINVAL; > > } > > ret = comp->ops->verify_mprime(comp->mei_dev, data, stream_ready); > > if (ret < 0) > > DRM_DEBUG_KMS("Verify mprime failed. %d\n", ret); > > - mutex_unlock(&dev_priv->hdcp_comp_mutex); > > + mutex_unlock(&i915->hdcp_comp_mutex); > > return ret; > > } > > @@ -1125,43 +1125,43 @@ hdcp2_verify_mprime(struct intel_connector *connector, > > static int hdcp2_authenticate_port(struct intel_connector *connector) > > { > > struct hdcp_port_data *data = &connector->hdcp.port_data; > > - struct drm_i915_private *dev_priv = to_i915(connector->base.dev); > > + struct drm_i915_private *i915 = to_i915(connector->base.dev); > > struct i915_hdcp_comp_master *comp; > > int ret; > > - mutex_lock(&dev_priv->hdcp_comp_mutex); > > - comp = dev_priv->hdcp_master; > > + mutex_lock(&i915->hdcp_comp_mutex); > > + comp = i915->hdcp_master; > > if (!comp || !comp->ops) { > > - mutex_unlock(&dev_priv->hdcp_comp_mutex); > > + mutex_unlock(&i915->hdcp_comp_mutex); > > return -EINVAL; > > } > > ret = comp->ops->enable_hdcp_authentication(comp->mei_dev, data); > > if (ret < 0) > > DRM_DEBUG_KMS("Enable hdcp auth failed. %d\n", ret); > > - mutex_unlock(&dev_priv->hdcp_comp_mutex); > > + mutex_unlock(&i915->hdcp_comp_mutex); > > return ret; > > } > > static int hdcp2_close_mei_session(struct intel_connector *connector) > > { > > - struct drm_i915_private *dev_priv = to_i915(connector->base.dev); > > + struct drm_i915_private *i915 = to_i915(connector->base.dev); > > struct i915_hdcp_comp_master *comp; > > int ret; > > - mutex_lock(&dev_priv->hdcp_comp_mutex); > > - comp = dev_priv->hdcp_master; > > + mutex_lock(&i915->hdcp_comp_mutex); > > + comp = i915->hdcp_master; > > if (!comp || !comp->ops) { > > - mutex_unlock(&dev_priv->hdcp_comp_mutex); > > + mutex_unlock(&i915->hdcp_comp_mutex); > > return -EINVAL; > > } > > ret = comp->ops->close_hdcp_session(comp->mei_dev, > > &connector->hdcp.port_data); > > - mutex_unlock(&dev_priv->hdcp_comp_mutex); > > + mutex_unlock(&i915->hdcp_comp_mutex); > > return ret; > > } > > @@ -1488,7 +1488,7 @@ static int hdcp2_authenticate_sink(struct intel_connector *connector) > > static int hdcp2_enable_encryption(struct intel_connector *connector) > > { > > struct intel_digital_port *intel_dig_port = conn_to_dig_port(connector); > > - struct drm_i915_private *dev_priv = to_i915(connector->base.dev); > > + struct drm_i915_private *i915 = to_i915(connector->base.dev); > > struct intel_hdcp *hdcp = &connector->hdcp; > > enum port port = connector->encoder->port; > > int ret; > > @@ -1511,7 +1511,7 @@ static int hdcp2_enable_encryption(struct intel_connector *connector) > > CTL_LINK_ENCRYPTION_REQ); > > } > > - ret = intel_wait_for_register(&dev_priv->uncore, HDCP2_STATUS_DDI(port), > > + ret = intel_wait_for_register(&i915->uncore, HDCP2_STATUS_DDI(port), > > LINK_ENCRYPTION_STATUS, > > LINK_ENCRYPTION_STATUS, > > ENCRYPT_STATUS_CHANGE_TIMEOUT_MS); > > @@ -1522,7 +1522,7 @@ static int hdcp2_enable_encryption(struct intel_connector *connector) > > static int hdcp2_disable_encryption(struct intel_connector *connector) > > { > > struct intel_digital_port *intel_dig_port = conn_to_dig_port(connector); > > - struct drm_i915_private *dev_priv = to_i915(connector->base.dev); > > + struct drm_i915_private *i915 = to_i915(connector->base.dev); > > struct intel_hdcp *hdcp = &connector->hdcp; > > enum port port = connector->encoder->port; > > int ret; > > @@ -1532,7 +1532,7 @@ static int hdcp2_disable_encryption(struct intel_connector *connector) > > I915_WRITE(HDCP2_CTL_DDI(port), > > I915_READ(HDCP2_CTL_DDI(port)) & ~CTL_LINK_ENCRYPTION_REQ); > > - ret = intel_wait_for_register(&dev_priv->uncore, HDCP2_STATUS_DDI(port), > > + ret = intel_wait_for_register(&i915->uncore, HDCP2_STATUS_DDI(port), > > LINK_ENCRYPTION_STATUS, 0x0, > > ENCRYPT_STATUS_CHANGE_TIMEOUT_MS); > > if (ret == -ETIMEDOUT) > > @@ -1628,7 +1628,7 @@ static int _intel_hdcp2_disable(struct intel_connector *connector) > > static int intel_hdcp2_check_link(struct intel_connector *connector) > > { > > struct intel_digital_port *intel_dig_port = conn_to_dig_port(connector); > > - struct drm_i915_private *dev_priv = to_i915(connector->base.dev); > > + struct drm_i915_private *i915 = to_i915(connector->base.dev); > > struct intel_hdcp *hdcp = &connector->hdcp; > > enum port port = connector->encoder->port; > > int ret = 0; > > @@ -1721,13 +1721,13 @@ static void intel_hdcp_check_work(struct work_struct *work) > > static int i915_hdcp_component_bind(struct device *i915_kdev, > > struct device *mei_kdev, void *data) > > { > > - struct drm_i915_private *dev_priv = kdev_to_i915(i915_kdev); > > + struct drm_i915_private *i915 = kdev_to_i915(i915_kdev); > > DRM_DEBUG("I915 HDCP comp bind\n"); > > - mutex_lock(&dev_priv->hdcp_comp_mutex); > > - dev_priv->hdcp_master = (struct i915_hdcp_comp_master *)data; > > - dev_priv->hdcp_master->mei_dev = mei_kdev; > > - mutex_unlock(&dev_priv->hdcp_comp_mutex); > > + mutex_lock(&i915->hdcp_comp_mutex); > > + i915->hdcp_master = (struct i915_hdcp_comp_master *)data; > > + i915->hdcp_master->mei_dev = mei_kdev; > > + mutex_unlock(&i915->hdcp_comp_mutex); > > return 0; > > } > > @@ -1735,12 +1735,12 @@ static int i915_hdcp_component_bind(struct device *i915_kdev, > > static void i915_hdcp_component_unbind(struct device *i915_kdev, > > struct device *mei_kdev, void *data) > > { > > - struct drm_i915_private *dev_priv = kdev_to_i915(i915_kdev); > > + struct drm_i915_private *i915 = kdev_to_i915(i915_kdev); > > DRM_DEBUG("I915 HDCP comp unbind\n"); > > - mutex_lock(&dev_priv->hdcp_comp_mutex); > > - dev_priv->hdcp_master = NULL; > > - mutex_unlock(&dev_priv->hdcp_comp_mutex); > > + mutex_lock(&i915->hdcp_comp_mutex); > > + i915->hdcp_master = NULL; > > + mutex_unlock(&i915->hdcp_comp_mutex); > > } > > static const struct component_ops i915_hdcp_component_ops = { > > @@ -1773,34 +1773,34 @@ static inline int initialize_hdcp_port_data(struct intel_connector *connector) > > return 0; > > } > > -static bool is_hdcp2_supported(struct drm_i915_private *dev_priv) > > +static bool is_hdcp2_supported(struct drm_i915_private *i915) > > { > > if (!IS_ENABLED(CONFIG_INTEL_MEI_HDCP)) > > return false; > > - return (INTEL_GEN(dev_priv) >= 10 || IS_GEMINILAKE(dev_priv) || > > - IS_KABYLAKE(dev_priv)); > > + return (INTEL_GEN(i915) >= 10 || IS_GEMINILAKE(i915) || > > + IS_KABYLAKE(i915)); > > } > > -void intel_hdcp_component_init(struct drm_i915_private *dev_priv) > > +void intel_hdcp_component_init(struct drm_i915_private *i915) > > { > > int ret; > > - if (!is_hdcp2_supported(dev_priv)) > > + if (!is_hdcp2_supported(i915)) > > return; > > - mutex_lock(&dev_priv->hdcp_comp_mutex); > > - WARN_ON(dev_priv->hdcp_comp_added); > > + mutex_lock(&i915->hdcp_comp_mutex); > > + WARN_ON(i915->hdcp_comp_added); > > - dev_priv->hdcp_comp_added = true; > > - mutex_unlock(&dev_priv->hdcp_comp_mutex); > > - ret = component_add_typed(dev_priv->drm.dev, &i915_hdcp_component_ops, > > + i915->hdcp_comp_added = true; > > + mutex_unlock(&i915->hdcp_comp_mutex); > > + ret = component_add_typed(i915->drm.dev, &i915_hdcp_component_ops, > > I915_COMPONENT_HDCP); > > if (ret < 0) { > > DRM_DEBUG_KMS("Failed at component add(%d)\n", ret); > > - mutex_lock(&dev_priv->hdcp_comp_mutex); > > - dev_priv->hdcp_comp_added = false; > > - mutex_unlock(&dev_priv->hdcp_comp_mutex); > > + mutex_lock(&i915->hdcp_comp_mutex); > > + i915->hdcp_comp_added = false; > > + mutex_unlock(&i915->hdcp_comp_mutex); > > return; > > } > > } > > @@ -1822,7 +1822,7 @@ static void intel_hdcp2_init(struct intel_connector *connector) > > int intel_hdcp_init(struct intel_connector *connector, > > const struct intel_hdcp_shim *shim) > > { > > - struct drm_i915_private *dev_priv = to_i915(connector->base.dev); > > + struct drm_i915_private *i915 = to_i915(connector->base.dev); > > struct intel_hdcp *hdcp = &connector->hdcp; > > int ret; > > @@ -1838,7 +1838,7 @@ int intel_hdcp_init(struct intel_connector *connector, > > INIT_DELAYED_WORK(&hdcp->check_work, intel_hdcp_check_work); > > INIT_WORK(&hdcp->prop_work, intel_hdcp_prop_work); > > - if (is_hdcp2_supported(dev_priv)) > > + if (is_hdcp2_supported(i915)) > > intel_hdcp2_init(connector); > > init_waitqueue_head(&hdcp->cp_irq_queue); > > @@ -1905,18 +1905,18 @@ int intel_hdcp_disable(struct intel_connector *connector) > > return ret; > > } > > -void intel_hdcp_component_fini(struct drm_i915_private *dev_priv) > > +void intel_hdcp_component_fini(struct drm_i915_private *i915) > > { > > - mutex_lock(&dev_priv->hdcp_comp_mutex); > > - if (!dev_priv->hdcp_comp_added) { > > - mutex_unlock(&dev_priv->hdcp_comp_mutex); > > + mutex_lock(&i915->hdcp_comp_mutex); > > + if (!i915->hdcp_comp_added) { > > + mutex_unlock(&i915->hdcp_comp_mutex); > > return; > > } > > - dev_priv->hdcp_comp_added = false; > > - mutex_unlock(&dev_priv->hdcp_comp_mutex); > > + i915->hdcp_comp_added = false; > > + mutex_unlock(&i915->hdcp_comp_mutex); > > - component_del(dev_priv->drm.dev, &i915_hdcp_component_ops); > > + component_del(i915->drm.dev, &i915_hdcp_component_ops); > > } > > void intel_hdcp_cleanup(struct intel_connector *connector) > > diff --git a/drivers/gpu/drm/i915/intel_hdcp.h b/drivers/gpu/drm/i915/intel_hdcp.h > > index be8da85c866a..24079dda1ac4 100644 > > --- a/drivers/gpu/drm/i915/intel_hdcp.h > > +++ b/drivers/gpu/drm/i915/intel_hdcp.h > > @@ -23,11 +23,11 @@ int intel_hdcp_init(struct intel_connector *connector, > > const struct intel_hdcp_shim *hdcp_shim); > > int intel_hdcp_enable(struct intel_connector *connector); > > int intel_hdcp_disable(struct intel_connector *connector); > > -bool is_hdcp_supported(struct drm_i915_private *dev_priv, enum port port); > > +bool is_hdcp_supported(struct drm_i915_private *i915, enum port port); > > bool intel_hdcp_capable(struct intel_connector *connector); > > bool intel_hdcp2_capable(struct intel_connector *connector); > > -void intel_hdcp_component_init(struct drm_i915_private *dev_priv); > > -void intel_hdcp_component_fini(struct drm_i915_private *dev_priv); > > +void intel_hdcp_component_init(struct drm_i915_private *i915); > > +void intel_hdcp_component_fini(struct drm_i915_private *i915); > > void intel_hdcp_cleanup(struct intel_connector *connector); > > void intel_hdcp_handle_cp_irq(struct intel_connector *connector); > > diff --git a/drivers/gpu/drm/i915/intel_hdmi.c b/drivers/gpu/drm/i915/intel_hdmi.c > > index 187a2b828b97..c5be41ba36da 100644 > > --- a/drivers/gpu/drm/i915/intel_hdmi.c > > +++ b/drivers/gpu/drm/i915/intel_hdmi.c > > @@ -67,17 +67,17 @@ static void > > assert_hdmi_port_disabled(struct intel_hdmi *intel_hdmi) > > { > > struct drm_device *dev = intel_hdmi_to_dev(intel_hdmi); > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > u32 enabled_bits; > > - enabled_bits = HAS_DDI(dev_priv) ? DDI_BUF_CTL_ENABLE : SDVO_ENABLE; > > + enabled_bits = HAS_DDI(i915) ? DDI_BUF_CTL_ENABLE : SDVO_ENABLE; > > WARN(I915_READ(intel_hdmi->hdmi_reg) & enabled_bits, > > "HDMI port enabled, expecting disabled\n"); > > } > > static void > > -assert_hdmi_transcoder_func_disabled(struct drm_i915_private *dev_priv, > > +assert_hdmi_transcoder_func_disabled(struct drm_i915_private *i915, > > enum transcoder cpu_transcoder) > > { > > WARN(I915_READ(TRANS_DDI_FUNC_CTL(cpu_transcoder)) & > > @@ -163,7 +163,7 @@ static u32 hsw_infoframe_enable(unsigned int type) > > } > > static i915_reg_t > > -hsw_dip_data_reg(struct drm_i915_private *dev_priv, > > +hsw_dip_data_reg(struct drm_i915_private *i915, > > enum transcoder cpu_transcoder, > > unsigned int type, > > int i) > > @@ -207,7 +207,7 @@ static void g4x_write_infoframe(struct intel_encoder *encoder, > > const void *frame, ssize_t len) > > { > > const u32 *data = frame; > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > u32 val = I915_READ(VIDEO_DIP_CTL); > > int i; > > @@ -241,7 +241,7 @@ static void g4x_read_infoframe(struct intel_encoder *encoder, > > unsigned int type, > > void *frame, ssize_t len) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > u32 val, *data = frame; > > int i; > > @@ -259,7 +259,7 @@ static void g4x_read_infoframe(struct intel_encoder *encoder, > > static u32 g4x_infoframes_enabled(struct intel_encoder *encoder, > > const struct intel_crtc_state *pipe_config) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > u32 val = I915_READ(VIDEO_DIP_CTL); > > if ((val & VIDEO_DIP_ENABLE) == 0) > > @@ -278,7 +278,7 @@ static void ibx_write_infoframe(struct intel_encoder *encoder, > > const void *frame, ssize_t len) > > { > > const u32 *data = frame; > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_crtc *intel_crtc = to_intel_crtc(crtc_state->base.crtc); > > i915_reg_t reg = TVIDEO_DIP_CTL(intel_crtc->pipe); > > u32 val = I915_READ(reg); > > @@ -314,7 +314,7 @@ static void ibx_read_infoframe(struct intel_encoder *encoder, > > unsigned int type, > > void *frame, ssize_t len) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc); > > u32 val, *data = frame; > > int i; > > @@ -333,7 +333,7 @@ static void ibx_read_infoframe(struct intel_encoder *encoder, > > static u32 ibx_infoframes_enabled(struct intel_encoder *encoder, > > const struct intel_crtc_state *pipe_config) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > enum pipe pipe = to_intel_crtc(pipe_config->base.crtc)->pipe; > > i915_reg_t reg = TVIDEO_DIP_CTL(pipe); > > u32 val = I915_READ(reg); > > @@ -355,7 +355,7 @@ static void cpt_write_infoframe(struct intel_encoder *encoder, > > const void *frame, ssize_t len) > > { > > const u32 *data = frame; > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_crtc *intel_crtc = to_intel_crtc(crtc_state->base.crtc); > > i915_reg_t reg = TVIDEO_DIP_CTL(intel_crtc->pipe); > > u32 val = I915_READ(reg); > > @@ -394,7 +394,7 @@ static void cpt_read_infoframe(struct intel_encoder *encoder, > > unsigned int type, > > void *frame, ssize_t len) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc); > > u32 val, *data = frame; > > int i; > > @@ -413,7 +413,7 @@ static void cpt_read_infoframe(struct intel_encoder *encoder, > > static u32 cpt_infoframes_enabled(struct intel_encoder *encoder, > > const struct intel_crtc_state *pipe_config) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > enum pipe pipe = to_intel_crtc(pipe_config->base.crtc)->pipe; > > u32 val = I915_READ(TVIDEO_DIP_CTL(pipe)); > > @@ -431,7 +431,7 @@ static void vlv_write_infoframe(struct intel_encoder *encoder, > > const void *frame, ssize_t len) > > { > > const u32 *data = frame; > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_crtc *intel_crtc = to_intel_crtc(crtc_state->base.crtc); > > i915_reg_t reg = VLV_TVIDEO_DIP_CTL(intel_crtc->pipe); > > u32 val = I915_READ(reg); > > @@ -467,7 +467,7 @@ static void vlv_read_infoframe(struct intel_encoder *encoder, > > unsigned int type, > > void *frame, ssize_t len) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc); > > u32 val, *data = frame; > > int i; > > @@ -486,7 +486,7 @@ static void vlv_read_infoframe(struct intel_encoder *encoder, > > static u32 vlv_infoframes_enabled(struct intel_encoder *encoder, > > const struct intel_crtc_state *pipe_config) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > enum pipe pipe = to_intel_crtc(pipe_config->base.crtc)->pipe; > > u32 val = I915_READ(VLV_TVIDEO_DIP_CTL(pipe)); > > @@ -507,7 +507,7 @@ static void hsw_write_infoframe(struct intel_encoder *encoder, > > const void *frame, ssize_t len) > > { > > const u32 *data = frame; > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > enum transcoder cpu_transcoder = crtc_state->cpu_transcoder; > > i915_reg_t ctl_reg = HSW_TVIDEO_DIP_CTL(cpu_transcoder); > > int data_size; > > @@ -520,13 +520,13 @@ static void hsw_write_infoframe(struct intel_encoder *encoder, > > I915_WRITE(ctl_reg, val); > > for (i = 0; i < len; i += 4) { > > - I915_WRITE(hsw_dip_data_reg(dev_priv, cpu_transcoder, > > + I915_WRITE(hsw_dip_data_reg(i915, cpu_transcoder, > > type, i >> 2), *data); > > data++; > > } > > /* Write every possible data byte to force correct ECC calculation. */ > > for (; i < data_size; i += 4) > > - I915_WRITE(hsw_dip_data_reg(dev_priv, cpu_transcoder, > > + I915_WRITE(hsw_dip_data_reg(i915, cpu_transcoder, > > type, i >> 2), 0); > > val |= hsw_infoframe_enable(type); > > @@ -539,7 +539,7 @@ static void hsw_read_infoframe(struct intel_encoder *encoder, > > unsigned int type, > > void *frame, ssize_t len) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > enum transcoder cpu_transcoder = crtc_state->cpu_transcoder; > > u32 val, *data = frame; > > int i; > > @@ -547,14 +547,14 @@ static void hsw_read_infoframe(struct intel_encoder *encoder, > > val = I915_READ(HSW_TVIDEO_DIP_CTL(cpu_transcoder)); > > for (i = 0; i < len; i += 4) > > - *data++ = I915_READ(hsw_dip_data_reg(dev_priv, cpu_transcoder, > > + *data++ = I915_READ(hsw_dip_data_reg(i915, cpu_transcoder, > > type, i >> 2)); > > } > > static u32 hsw_infoframes_enabled(struct intel_encoder *encoder, > > const struct intel_crtc_state *pipe_config) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > u32 val = I915_READ(HSW_TVIDEO_DIP_CTL(pipe_config->cpu_transcoder)); > > u32 mask; > > @@ -562,7 +562,7 @@ static u32 hsw_infoframes_enabled(struct intel_encoder *encoder, > > VIDEO_DIP_ENABLE_GCP_HSW | VIDEO_DIP_ENABLE_VS_HSW | > > VIDEO_DIP_ENABLE_GMP_HSW | VIDEO_DIP_ENABLE_SPD_HSW); > > - if (INTEL_GEN(dev_priv) >= 10 || IS_GEMINILAKE(dev_priv)) > > + if (INTEL_GEN(i915) >= 10 || IS_GEMINILAKE(i915)) > > mask |= VIDEO_DIP_ENABLE_DRM_GLK; > > return val & mask; > > @@ -593,7 +593,7 @@ u32 intel_hdmi_infoframe_enable(unsigned int type) > > u32 intel_hdmi_infoframes_enabled(struct intel_encoder *encoder, > > const struct intel_crtc_state *crtc_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_digital_port *dig_port = enc_to_dig_port(&encoder->base); > > u32 val, ret = 0; > > int i; > > @@ -604,7 +604,7 @@ u32 intel_hdmi_infoframes_enabled(struct intel_encoder *encoder, > > for (i = 0; i < ARRAY_SIZE(infoframe_type_to_idx); i++) { > > unsigned int type = infoframe_type_to_idx[i]; > > - if (HAS_DDI(dev_priv)) { > > + if (HAS_DDI(i915)) { > > if (val & hsw_infoframe_enable(type)) > > ret |= BIT(i); > > } else { > > @@ -804,10 +804,10 @@ intel_hdmi_compute_drm_infoframe(struct intel_encoder *encoder, > > struct drm_connector_state *conn_state) > > { > > struct hdmi_drm_infoframe *frame = &crtc_state->infoframes.drm.drm; > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > int ret; > > - if (!(INTEL_GEN(dev_priv) >= 10 || IS_GEMINILAKE(dev_priv))) > > + if (!(INTEL_GEN(i915) >= 10 || IS_GEMINILAKE(i915))) > > return true; > > if (!crtc_state->has_infoframe) > > @@ -837,7 +837,7 @@ static void g4x_set_infoframes(struct intel_encoder *encoder, > > const struct intel_crtc_state *crtc_state, > > const struct drm_connector_state *conn_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_digital_port *intel_dig_port = enc_to_dig_port(&encoder->base); > > struct intel_hdmi *intel_hdmi = &intel_dig_port->hdmi; > > i915_reg_t reg = VIDEO_DIP_CTL; > > @@ -947,7 +947,7 @@ static bool intel_hdmi_set_gcp_infoframe(struct intel_encoder *encoder, > > const struct intel_crtc_state *crtc_state, > > const struct drm_connector_state *conn_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc); > > i915_reg_t reg; > > @@ -955,11 +955,11 @@ static bool intel_hdmi_set_gcp_infoframe(struct intel_encoder *encoder, > > intel_hdmi_infoframe_enable(HDMI_PACKET_TYPE_GENERAL_CONTROL)) == 0) > > return false; > > - if (HAS_DDI(dev_priv)) > > + if (HAS_DDI(i915)) > > reg = HSW_TVIDEO_DIP_GCP(crtc_state->cpu_transcoder); > > - else if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) > > + else if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) > > reg = VLV_TVIDEO_DIP_GCP(crtc->pipe); > > - else if (HAS_PCH_SPLIT(dev_priv)) > > + else if (HAS_PCH_SPLIT(i915)) > > reg = TVIDEO_DIP_GCP(crtc->pipe); > > else > > return false; > > @@ -972,7 +972,7 @@ static bool intel_hdmi_set_gcp_infoframe(struct intel_encoder *encoder, > > void intel_hdmi_read_gcp_infoframe(struct intel_encoder *encoder, > > struct intel_crtc_state *crtc_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc); > > i915_reg_t reg; > > @@ -980,11 +980,11 @@ void intel_hdmi_read_gcp_infoframe(struct intel_encoder *encoder, > > intel_hdmi_infoframe_enable(HDMI_PACKET_TYPE_GENERAL_CONTROL)) == 0) > > return; > > - if (HAS_DDI(dev_priv)) > > + if (HAS_DDI(i915)) > > reg = HSW_TVIDEO_DIP_GCP(crtc_state->cpu_transcoder); > > - else if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) > > + else if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) > > reg = VLV_TVIDEO_DIP_GCP(crtc->pipe); > > - else if (HAS_PCH_SPLIT(dev_priv)) > > + else if (HAS_PCH_SPLIT(i915)) > > reg = TVIDEO_DIP_GCP(crtc->pipe); > > else > > return; > > @@ -996,9 +996,9 @@ static void intel_hdmi_compute_gcp_infoframe(struct intel_encoder *encoder, > > struct intel_crtc_state *crtc_state, > > struct drm_connector_state *conn_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > - if (IS_G4X(dev_priv) || !crtc_state->has_infoframe) > > + if (IS_G4X(i915) || !crtc_state->has_infoframe) > > return; > > crtc_state->infoframes.enable |= > > @@ -1019,7 +1019,7 @@ static void ibx_set_infoframes(struct intel_encoder *encoder, > > const struct intel_crtc_state *crtc_state, > > const struct drm_connector_state *conn_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_crtc *intel_crtc = to_intel_crtc(crtc_state->base.crtc); > > struct intel_digital_port *intel_dig_port = enc_to_dig_port(&encoder->base); > > struct intel_hdmi *intel_hdmi = &intel_dig_port->hdmi; > > @@ -1078,7 +1078,7 @@ static void cpt_set_infoframes(struct intel_encoder *encoder, > > const struct intel_crtc_state *crtc_state, > > const struct drm_connector_state *conn_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_crtc *intel_crtc = to_intel_crtc(crtc_state->base.crtc); > > struct intel_hdmi *intel_hdmi = enc_to_intel_hdmi(&encoder->base); > > i915_reg_t reg = TVIDEO_DIP_CTL(intel_crtc->pipe); > > @@ -1127,7 +1127,7 @@ static void vlv_set_infoframes(struct intel_encoder *encoder, > > const struct intel_crtc_state *crtc_state, > > const struct drm_connector_state *conn_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_crtc *intel_crtc = to_intel_crtc(crtc_state->base.crtc); > > struct intel_hdmi *intel_hdmi = enc_to_intel_hdmi(&encoder->base); > > i915_reg_t reg = VLV_TVIDEO_DIP_CTL(intel_crtc->pipe); > > @@ -1185,11 +1185,11 @@ static void hsw_set_infoframes(struct intel_encoder *encoder, > > const struct intel_crtc_state *crtc_state, > > const struct drm_connector_state *conn_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > i915_reg_t reg = HSW_TVIDEO_DIP_CTL(crtc_state->cpu_transcoder); > > u32 val = I915_READ(reg); > > - assert_hdmi_transcoder_func_disabled(dev_priv, > > + assert_hdmi_transcoder_func_disabled(i915, > > crtc_state->cpu_transcoder); > > val &= ~(VIDEO_DIP_ENABLE_VSC_HSW | VIDEO_DIP_ENABLE_AVI_HSW | > > @@ -1225,9 +1225,9 @@ static void hsw_set_infoframes(struct intel_encoder *encoder, > > void intel_dp_dual_mode_set_tmds_output(struct intel_hdmi *hdmi, bool enable) > > { > > - struct drm_i915_private *dev_priv = to_i915(intel_hdmi_to_dev(hdmi)); > > + struct drm_i915_private *i915 = to_i915(intel_hdmi_to_dev(hdmi)); > > struct i2c_adapter *adapter = > > - intel_gmbus_get_adapter(dev_priv, hdmi->ddc_bus); > > + intel_gmbus_get_adapter(i915, hdmi->ddc_bus); > > if (hdmi->dp_dual_mode.type < DRM_DP_DUAL_MODE_TYPE2_DVI) > > return; > > @@ -1243,9 +1243,9 @@ static int intel_hdmi_hdcp_read(struct intel_digital_port *intel_dig_port, > > unsigned int offset, void *buffer, size_t size) > > { > > struct intel_hdmi *hdmi = &intel_dig_port->hdmi; > > - struct drm_i915_private *dev_priv = > > + struct drm_i915_private *i915 = > > intel_dig_port->base.base.dev->dev_private; > > - struct i2c_adapter *adapter = intel_gmbus_get_adapter(dev_priv, > > + struct i2c_adapter *adapter = intel_gmbus_get_adapter(i915, > > hdmi->ddc_bus); > > int ret; > > u8 start = offset & 0xff; > > @@ -1273,9 +1273,9 @@ static int intel_hdmi_hdcp_write(struct intel_digital_port *intel_dig_port, > > unsigned int offset, void *buffer, size_t size) > > { > > struct intel_hdmi *hdmi = &intel_dig_port->hdmi; > > - struct drm_i915_private *dev_priv = > > + struct drm_i915_private *i915 = > > intel_dig_port->base.base.dev->dev_private; > > - struct i2c_adapter *adapter = intel_gmbus_get_adapter(dev_priv, > > + struct i2c_adapter *adapter = intel_gmbus_get_adapter(i915, > > hdmi->ddc_bus); > > int ret; > > u8 *write_buf; > > @@ -1308,9 +1308,9 @@ int intel_hdmi_hdcp_write_an_aksv(struct intel_digital_port *intel_dig_port, > > u8 *an) > > { > > struct intel_hdmi *hdmi = &intel_dig_port->hdmi; > > - struct drm_i915_private *dev_priv = > > + struct drm_i915_private *i915 = > > intel_dig_port->base.base.dev->dev_private; > > - struct i2c_adapter *adapter = intel_gmbus_get_adapter(dev_priv, > > + struct i2c_adapter *adapter = intel_gmbus_get_adapter(i915, > > hdmi->ddc_bus); > > int ret; > > @@ -1428,7 +1428,7 @@ int intel_hdmi_hdcp_read_v_prime_part(struct intel_digital_port *intel_dig_port, > > static int kbl_repositioning_enc_en_signal(struct intel_connector *connector) > > { > > - struct drm_i915_private *dev_priv = to_i915(connector->base.dev); > > + struct drm_i915_private *i915 = to_i915(connector->base.dev); > > struct intel_digital_port *intel_dig_port = conn_to_dig_port(connector); > > struct drm_crtc *crtc = connector->base.state->crtc; > > struct intel_crtc *intel_crtc = container_of(crtc, > > @@ -1463,7 +1463,7 @@ int intel_hdmi_hdcp_toggle_signalling(struct intel_digital_port *intel_dig_port, > > { > > struct intel_hdmi *hdmi = &intel_dig_port->hdmi; > > struct intel_connector *connector = hdmi->attached_connector; > > - struct drm_i915_private *dev_priv = to_i915(connector->base.dev); > > + struct drm_i915_private *i915 = to_i915(connector->base.dev); > > int ret; > > if (!enable) > > @@ -1480,7 +1480,7 @@ int intel_hdmi_hdcp_toggle_signalling(struct intel_digital_port *intel_dig_port, > > * WA: To fix incorrect positioning of the window of > > * opportunity and enc_en signalling in KABYLAKE. > > */ > > - if (IS_KABYLAKE(dev_priv) && enable) > > + if (IS_KABYLAKE(i915) && enable) > > return kbl_repositioning_enc_en_signal(connector); > > return 0; > > @@ -1489,7 +1489,7 @@ int intel_hdmi_hdcp_toggle_signalling(struct intel_digital_port *intel_dig_port, > > static > > bool intel_hdmi_hdcp_check_link(struct intel_digital_port *intel_dig_port) > > { > > - struct drm_i915_private *dev_priv = > > + struct drm_i915_private *i915 = > > intel_dig_port->base.base.dev->dev_private; > > enum port port = intel_dig_port->base.port; > > int ret; > > @@ -1720,7 +1720,7 @@ static void intel_hdmi_prepare(struct intel_encoder *encoder, > > const struct intel_crtc_state *crtc_state) > > { > > struct drm_device *dev = encoder->base.dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc); > > struct intel_hdmi *intel_hdmi = enc_to_intel_hdmi(&encoder->base); > > const struct drm_display_mode *adjusted_mode = &crtc_state->base.adjusted_mode; > > @@ -1729,7 +1729,7 @@ static void intel_hdmi_prepare(struct intel_encoder *encoder, > > intel_dp_dual_mode_set_tmds_output(intel_hdmi, true); > > hdmi_val = SDVO_ENCODING_HDMI; > > - if (!HAS_PCH_SPLIT(dev_priv) && crtc_state->limited_color_range) > > + if (!HAS_PCH_SPLIT(i915) && crtc_state->limited_color_range) > > hdmi_val |= HDMI_COLOR_RANGE_16_235; > > if (adjusted_mode->flags & DRM_MODE_FLAG_PVSYNC) > > hdmi_val |= SDVO_VSYNC_ACTIVE_HIGH; > > @@ -1744,9 +1744,9 @@ static void intel_hdmi_prepare(struct intel_encoder *encoder, > > if (crtc_state->has_hdmi_sink) > > hdmi_val |= HDMI_MODE_SELECT_HDMI; > > - if (HAS_PCH_CPT(dev_priv)) > > + if (HAS_PCH_CPT(i915)) > > hdmi_val |= SDVO_PIPE_SEL_CPT(crtc->pipe); > > - else if (IS_CHERRYVIEW(dev_priv)) > > + else if (IS_CHERRYVIEW(i915)) > > hdmi_val |= SDVO_PIPE_SEL_CHV(crtc->pipe); > > else > > hdmi_val |= SDVO_PIPE_SEL(crtc->pipe); > > @@ -1758,19 +1758,19 @@ static void intel_hdmi_prepare(struct intel_encoder *encoder, > > static bool intel_hdmi_get_hw_state(struct intel_encoder *encoder, > > enum pipe *pipe) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_hdmi *intel_hdmi = enc_to_intel_hdmi(&encoder->base); > > intel_wakeref_t wakeref; > > bool ret; > > - wakeref = intel_display_power_get_if_enabled(dev_priv, > > + wakeref = intel_display_power_get_if_enabled(i915, > > encoder->power_domain); > > if (!wakeref) > > return false; > > - ret = intel_sdvo_port_enabled(dev_priv, intel_hdmi->hdmi_reg, pipe); > > + ret = intel_sdvo_port_enabled(i915, intel_hdmi->hdmi_reg, pipe); > > - intel_display_power_put(dev_priv, encoder->power_domain, wakeref); > > + intel_display_power_put(i915, encoder->power_domain, wakeref); > > return ret; > > } > > @@ -1780,7 +1780,7 @@ static void intel_hdmi_get_config(struct intel_encoder *encoder, > > { > > struct intel_hdmi *intel_hdmi = enc_to_intel_hdmi(&encoder->base); > > struct drm_device *dev = encoder->base.dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > u32 tmp, flags = 0; > > int dotclock; > > @@ -1810,7 +1810,7 @@ static void intel_hdmi_get_config(struct intel_encoder *encoder, > > if (tmp & HDMI_AUDIO_ENABLE) > > pipe_config->has_audio = true; > > - if (!HAS_PCH_SPLIT(dev_priv) && > > + if (!HAS_PCH_SPLIT(i915) && > > tmp & HDMI_COLOR_RANGE_16_235) > > pipe_config->limited_color_range = true; > > @@ -1858,7 +1858,7 @@ static void g4x_enable_hdmi(struct intel_encoder *encoder, > > const struct drm_connector_state *conn_state) > > { > > struct drm_device *dev = encoder->base.dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > struct intel_hdmi *intel_hdmi = enc_to_intel_hdmi(&encoder->base); > > u32 temp; > > @@ -1880,7 +1880,7 @@ static void ibx_enable_hdmi(struct intel_encoder *encoder, > > const struct drm_connector_state *conn_state) > > { > > struct drm_device *dev = encoder->base.dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > struct intel_hdmi *intel_hdmi = enc_to_intel_hdmi(&encoder->base); > > u32 temp; > > @@ -1930,7 +1930,7 @@ static void cpt_enable_hdmi(struct intel_encoder *encoder, > > const struct drm_connector_state *conn_state) > > { > > struct drm_device *dev = encoder->base.dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > struct intel_crtc *crtc = to_intel_crtc(pipe_config->base.crtc); > > struct intel_hdmi *intel_hdmi = enc_to_intel_hdmi(&encoder->base); > > enum pipe pipe = crtc->pipe; > > @@ -1991,7 +1991,7 @@ static void intel_disable_hdmi(struct intel_encoder *encoder, > > const struct drm_connector_state *old_conn_state) > > { > > struct drm_device *dev = encoder->base.dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > struct intel_hdmi *intel_hdmi = enc_to_intel_hdmi(&encoder->base); > > struct intel_digital_port *intel_dig_port = > > hdmi_to_dig_port(intel_hdmi); > > @@ -2009,13 +2009,13 @@ static void intel_disable_hdmi(struct intel_encoder *encoder, > > * to transcoder A after disabling it to allow the > > * matching DP port to be enabled on transcoder A. > > */ > > - if (HAS_PCH_IBX(dev_priv) && crtc->pipe == PIPE_B) { > > + if (HAS_PCH_IBX(i915) && crtc->pipe == PIPE_B) { > > /* > > * We get CPU/PCH FIFO underruns on the other pipe when > > * doing the workaround. Sweep them under the rug. > > */ > > - intel_set_cpu_fifo_underrun_reporting(dev_priv, PIPE_A, false); > > - intel_set_pch_fifo_underrun_reporting(dev_priv, PIPE_A, false); > > + intel_set_cpu_fifo_underrun_reporting(i915, PIPE_A, false); > > + intel_set_pch_fifo_underrun_reporting(i915, PIPE_A, false); > > temp &= ~SDVO_PIPE_SEL_MASK; > > temp |= SDVO_ENABLE | SDVO_PIPE_SEL(PIPE_A); > > @@ -2032,9 +2032,9 @@ static void intel_disable_hdmi(struct intel_encoder *encoder, > > I915_WRITE(intel_hdmi->hdmi_reg, temp); > > POSTING_READ(intel_hdmi->hdmi_reg); > > - intel_wait_for_vblank_if_active(dev_priv, PIPE_A); > > - intel_set_cpu_fifo_underrun_reporting(dev_priv, PIPE_A, true); > > - intel_set_pch_fifo_underrun_reporting(dev_priv, PIPE_A, true); > > + intel_wait_for_vblank_if_active(i915, PIPE_A); > > + intel_set_cpu_fifo_underrun_reporting(i915, PIPE_A, true); > > + intel_set_pch_fifo_underrun_reporting(i915, PIPE_A, true); > > } > > intel_dig_port->set_infoframes(encoder, > > @@ -2073,16 +2073,16 @@ static void pch_post_disable_hdmi(struct intel_encoder *encoder, > > static int intel_hdmi_source_max_tmds_clock(struct intel_encoder *encoder) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > const struct ddi_vbt_port_info *info = > > - &dev_priv->vbt.ddi_port_info[encoder->port]; > > + &i915->vbt.ddi_port_info[encoder->port]; > > int max_tmds_clock; > > - if (INTEL_GEN(dev_priv) >= 10 || IS_GEMINILAKE(dev_priv)) > > + if (INTEL_GEN(i915) >= 10 || IS_GEMINILAKE(i915)) > > max_tmds_clock = 594000; > > - else if (INTEL_GEN(dev_priv) >= 8 || IS_HASWELL(dev_priv)) > > + else if (INTEL_GEN(i915) >= 8 || IS_HASWELL(i915)) > > max_tmds_clock = 300000; > > - else if (INTEL_GEN(dev_priv) >= 5) > > + else if (INTEL_GEN(i915) >= 5) > > max_tmds_clock = 225000; > > else > > max_tmds_clock = 165000; > > @@ -2123,7 +2123,7 @@ hdmi_port_clock_valid(struct intel_hdmi *hdmi, > > int clock, bool respect_downstream_limits, > > bool force_dvi) > > { > > - struct drm_i915_private *dev_priv = to_i915(intel_hdmi_to_dev(hdmi)); > > + struct drm_i915_private *i915 = to_i915(intel_hdmi_to_dev(hdmi)); > > if (clock < 25000) > > return MODE_CLOCK_LOW; > > @@ -2131,11 +2131,11 @@ hdmi_port_clock_valid(struct intel_hdmi *hdmi, > > return MODE_CLOCK_HIGH; > > /* BXT DPLL can't generate 223-240 MHz */ > > - if (IS_GEN9_LP(dev_priv) && clock > 223333 && clock < 240000) > > + if (IS_GEN9_LP(i915) && clock > 223333 && clock < 240000) > > return MODE_CLOCK_RANGE; > > /* CHV DPLL can't generate 216-240 MHz */ > > - if (IS_CHERRYVIEW(dev_priv) && clock > 216000 && clock < 240000) > > + if (IS_CHERRYVIEW(i915) && clock > 216000 && clock < 240000) > > return MODE_CLOCK_RANGE; > > return MODE_OK; > > @@ -2147,7 +2147,7 @@ intel_hdmi_mode_valid(struct drm_connector *connector, > > { > > struct intel_hdmi *hdmi = intel_attached_hdmi(connector); > > struct drm_device *dev = intel_hdmi_to_dev(hdmi); > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > enum drm_mode_status status; > > int clock; > > int max_dotclk = to_i915(connector->dev)->max_dotclk_freq; > > @@ -2176,12 +2176,12 @@ intel_hdmi_mode_valid(struct drm_connector *connector, > > if (hdmi->has_hdmi_sink && !force_dvi) { > > /* if we can't do 8bpc we may still be able to do 12bpc */ > > - if (status != MODE_OK && !HAS_GMCH(dev_priv)) > > + if (status != MODE_OK && !HAS_GMCH(i915)) > > status = hdmi_port_clock_valid(hdmi, clock * 3 / 2, > > true, force_dvi); > > /* if we can't do 8,12bpc we may still be able to do 10bpc */ > > - if (status != MODE_OK && INTEL_GEN(dev_priv) >= 11) > > + if (status != MODE_OK && INTEL_GEN(i915) >= 11) > > status = hdmi_port_clock_valid(hdmi, clock * 5 / 4, > > true, force_dvi); > > } > > @@ -2192,7 +2192,7 @@ intel_hdmi_mode_valid(struct drm_connector *connector, > > static bool hdmi_deep_color_possible(const struct intel_crtc_state *crtc_state, > > int bpc) > > { > > - struct drm_i915_private *dev_priv = > > + struct drm_i915_private *i915 = > > to_i915(crtc_state->base.crtc->dev); > > struct drm_atomic_state *state = crtc_state->base.state; > > struct drm_connector_state *connector_state; > > @@ -2201,10 +2201,10 @@ static bool hdmi_deep_color_possible(const struct intel_crtc_state *crtc_state, > > &crtc_state->base.adjusted_mode; > > int i; > > - if (HAS_GMCH(dev_priv)) > > + if (HAS_GMCH(i915)) > > return false; > > - if (bpc == 10 && INTEL_GEN(dev_priv) < 11) > > + if (bpc == 10 && INTEL_GEN(i915) < 11) > > return false; > > if (crtc_state->pipe_bpp < bpc * 3) > > @@ -2246,13 +2246,13 @@ static bool hdmi_deep_color_possible(const struct intel_crtc_state *crtc_state, > > } > > /* Display WA #1139: glk */ > > - if (bpc == 12 && IS_GLK_REVID(dev_priv, 0, GLK_REVID_A1) && > > + if (bpc == 12 && IS_GLK_REVID(i915, 0, GLK_REVID_A1) && > > adjusted_mode->htotal > 5460) > > return false; > > /* Display Wa_1405510057:icl */ > > if (crtc_state->output_format == INTEL_OUTPUT_FORMAT_YCBCR420 && > > - bpc == 10 && INTEL_GEN(dev_priv) >= 11 && > > + bpc == 10 && INTEL_GEN(i915) >= 11 && > > (adjusted_mode->crtc_hblank_end - > > adjusted_mode->crtc_hblank_start) % 8 == 2) > > return false; > > @@ -2297,7 +2297,7 @@ int intel_hdmi_compute_config(struct intel_encoder *encoder, > > struct drm_connector_state *conn_state) > > { > > struct intel_hdmi *intel_hdmi = enc_to_intel_hdmi(&encoder->base); > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct drm_display_mode *adjusted_mode = &pipe_config->base.adjusted_mode; > > struct drm_connector *connector = conn_state->connector; > > struct drm_scdc *scdc = &connector->display_info.hdmi.scdc; > > @@ -2345,7 +2345,7 @@ int intel_hdmi_compute_config(struct intel_encoder *encoder, > > } > > } > > - if (HAS_PCH_SPLIT(dev_priv) && !HAS_DDI(dev_priv)) > > + if (HAS_PCH_SPLIT(i915) && !HAS_DDI(i915)) > > pipe_config->has_pch_encoder = true; > > if (pipe_config->has_hdmi_sink) { > > @@ -2399,8 +2399,8 @@ int intel_hdmi_compute_config(struct intel_encoder *encoder, > > pipe_config->lane_count = 4; > > - if (scdc->scrambling.supported && (INTEL_GEN(dev_priv) >= 10 || > > - IS_GEMINILAKE(dev_priv))) { > > + if (scdc->scrambling.supported && (INTEL_GEN(i915) >= 10 || > > + IS_GEMINILAKE(i915))) { > > if (scdc->scrambling.low_rates) > > pipe_config->hdmi_scrambling = true; > > @@ -2453,11 +2453,11 @@ intel_hdmi_unset_edid(struct drm_connector *connector) > > static void > > intel_hdmi_dp_dual_mode_detect(struct drm_connector *connector, bool has_edid) > > { > > - struct drm_i915_private *dev_priv = to_i915(connector->dev); > > + struct drm_i915_private *i915 = to_i915(connector->dev); > > struct intel_hdmi *hdmi = intel_attached_hdmi(connector); > > enum port port = hdmi_to_dig_port(hdmi)->base.port; > > struct i2c_adapter *adapter = > > - intel_gmbus_get_adapter(dev_priv, hdmi->ddc_bus); > > + intel_gmbus_get_adapter(i915, hdmi->ddc_bus); > > enum drm_dp_dual_mode_type type = drm_dp_dual_mode_detect(adapter); > > /* > > @@ -2477,7 +2477,7 @@ intel_hdmi_dp_dual_mode_detect(struct drm_connector *connector, bool has_edid) > > * Make sure not to set limits for that port. > > */ > > if (has_edid && !connector->override_edid && > > - intel_bios_is_port_dp_dual_mode(dev_priv, port)) { > > + intel_bios_is_port_dp_dual_mode(i915, port)) { > > DRM_DEBUG_KMS("Assuming DP dual mode adaptor presence based on VBT\n"); > > type = DRM_DP_DUAL_MODE_TYPE1_DVI; > > } else { > > @@ -2500,16 +2500,16 @@ intel_hdmi_dp_dual_mode_detect(struct drm_connector *connector, bool has_edid) > > static bool > > intel_hdmi_set_edid(struct drm_connector *connector) > > { > > - struct drm_i915_private *dev_priv = to_i915(connector->dev); > > + struct drm_i915_private *i915 = to_i915(connector->dev); > > struct intel_hdmi *intel_hdmi = intel_attached_hdmi(connector); > > intel_wakeref_t wakeref; > > struct edid *edid; > > bool connected = false; > > struct i2c_adapter *i2c; > > - wakeref = intel_display_power_get(dev_priv, POWER_DOMAIN_GMBUS); > > + wakeref = intel_display_power_get(i915, POWER_DOMAIN_GMBUS); > > - i2c = intel_gmbus_get_adapter(dev_priv, intel_hdmi->ddc_bus); > > + i2c = intel_gmbus_get_adapter(i915, intel_hdmi->ddc_bus); > > edid = drm_get_edid(connector, i2c); > > @@ -2522,7 +2522,7 @@ intel_hdmi_set_edid(struct drm_connector *connector) > > intel_hdmi_dp_dual_mode_detect(connector, edid != NULL); > > - intel_display_power_put(dev_priv, POWER_DOMAIN_GMBUS, wakeref); > > + intel_display_power_put(i915, POWER_DOMAIN_GMBUS, wakeref); > > to_intel_connector(connector)->detect_edid = edid; > > if (edid && edid->input & DRM_EDID_INPUT_DIGITAL) { > > @@ -2541,7 +2541,7 @@ static enum drm_connector_status > > intel_hdmi_detect(struct drm_connector *connector, bool force) > > { > > enum drm_connector_status status = connector_status_disconnected; > > - struct drm_i915_private *dev_priv = to_i915(connector->dev); > > + struct drm_i915_private *i915 = to_i915(connector->dev); > > struct intel_hdmi *intel_hdmi = intel_attached_hdmi(connector); > > struct intel_encoder *encoder = &hdmi_to_dig_port(intel_hdmi)->base; > > intel_wakeref_t wakeref; > > @@ -2549,9 +2549,9 @@ intel_hdmi_detect(struct drm_connector *connector, bool force) > > DRM_DEBUG_KMS("[CONNECTOR:%d:%s]\n", > > connector->base.id, connector->name); > > - wakeref = intel_display_power_get(dev_priv, POWER_DOMAIN_GMBUS); > > + wakeref = intel_display_power_get(i915, POWER_DOMAIN_GMBUS); > > - if (INTEL_GEN(dev_priv) >= 11 && > > + if (INTEL_GEN(i915) >= 11 && > > !intel_digital_port_connected(encoder)) > > goto out; > > @@ -2561,7 +2561,7 @@ intel_hdmi_detect(struct drm_connector *connector, bool force) > > status = connector_status_connected; > > out: > > - intel_display_power_put(dev_priv, POWER_DOMAIN_GMBUS, wakeref); > > + intel_display_power_put(i915, POWER_DOMAIN_GMBUS, wakeref); > > if (status != connector_status_connected) > > cec_notifier_phys_addr_invalidate(intel_hdmi->cec_notifier); > > @@ -2613,7 +2613,7 @@ static void vlv_hdmi_pre_enable(struct intel_encoder *encoder, > > const struct drm_connector_state *conn_state) > > { > > struct intel_digital_port *dport = enc_to_dig_port(&encoder->base); > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > vlv_phy_pre_encoder_enable(encoder, pipe_config); > > @@ -2627,7 +2627,7 @@ static void vlv_hdmi_pre_enable(struct intel_encoder *encoder, > > g4x_enable_hdmi(encoder, pipe_config, conn_state); > > - vlv_wait_port_ready(dev_priv, dport, 0x0); > > + vlv_wait_port_ready(i915, dport, 0x0); > > } > > static void vlv_hdmi_pre_pll_enable(struct intel_encoder *encoder, > > @@ -2668,14 +2668,14 @@ static void chv_hdmi_post_disable(struct intel_encoder *encoder, > > const struct drm_connector_state *old_conn_state) > > { > > struct drm_device *dev = encoder->base.dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > - vlv_dpio_get(dev_priv); > > + vlv_dpio_get(i915); > > /* Assert data lane reset */ > > chv_data_lane_soft_reset(encoder, old_crtc_state, true); > > - vlv_dpio_put(dev_priv); > > + vlv_dpio_put(i915); > > } > > static void chv_hdmi_pre_enable(struct intel_encoder *encoder, > > @@ -2684,7 +2684,7 @@ static void chv_hdmi_pre_enable(struct intel_encoder *encoder, > > { > > struct intel_digital_port *dport = enc_to_dig_port(&encoder->base); > > struct drm_device *dev = encoder->base.dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > chv_phy_pre_encoder_enable(encoder, pipe_config); > > @@ -2698,7 +2698,7 @@ static void chv_hdmi_pre_enable(struct intel_encoder *encoder, > > g4x_enable_hdmi(encoder, pipe_config, conn_state); > > - vlv_wait_port_ready(dev_priv, dport, 0x0); > > + vlv_wait_port_ready(i915, dport, 0x0); > > /* Second common lane will stay alive on its own now */ > > chv_phy_release_cl2_override(encoder); > > @@ -2707,10 +2707,10 @@ static void chv_hdmi_pre_enable(struct intel_encoder *encoder, > > static struct i2c_adapter * > > intel_hdmi_get_i2c_adapter(struct drm_connector *connector) > > { > > - struct drm_i915_private *dev_priv = to_i915(connector->dev); > > + struct drm_i915_private *i915 = to_i915(connector->dev); > > struct intel_hdmi *intel_hdmi = intel_attached_hdmi(connector); > > - return intel_gmbus_get_adapter(dev_priv, intel_hdmi->ddc_bus); > > + return intel_gmbus_get_adapter(i915, intel_hdmi->ddc_bus); > > } > > static void intel_hdmi_create_i2c_symlink(struct drm_connector *connector) > > @@ -2791,7 +2791,7 @@ static const struct drm_encoder_funcs intel_hdmi_enc_funcs = { > > static void > > intel_hdmi_add_properties(struct intel_hdmi *intel_hdmi, struct drm_connector *connector) > > { > > - struct drm_i915_private *dev_priv = to_i915(connector->dev); > > + struct drm_i915_private *i915 = to_i915(connector->dev); > > struct intel_digital_port *intel_dig_port = > > hdmi_to_dig_port(intel_hdmi); > > @@ -2810,11 +2810,11 @@ intel_hdmi_add_properties(struct intel_hdmi *intel_hdmi, struct drm_connector *c > > drm_connector_attach_content_type_property(connector); > > connector->state->picture_aspect_ratio = HDMI_PICTURE_ASPECT_NONE; > > - if (INTEL_GEN(dev_priv) >= 10 || IS_GEMINILAKE(dev_priv)) > > + if (INTEL_GEN(i915) >= 10 || IS_GEMINILAKE(i915)) > > drm_object_attach_property(&connector->base, > > connector->dev->mode_config.hdr_output_metadata_property, 0); > > - if (!HAS_GMCH(dev_priv)) > > + if (!HAS_GMCH(i915)) > > drm_connector_attach_max_bpc_property(connector, 8, 12); > > } > > @@ -2841,12 +2841,12 @@ bool intel_hdmi_handle_sink_scrambling(struct intel_encoder *encoder, > > bool high_tmds_clock_ratio, > > bool scrambling) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_hdmi *intel_hdmi = enc_to_intel_hdmi(&encoder->base); > > struct drm_scrambling *sink_scrambling = > > &connector->display_info.hdmi.scdc.scrambling; > > struct i2c_adapter *adapter = > > - intel_gmbus_get_adapter(dev_priv, intel_hdmi->ddc_bus); > > + intel_gmbus_get_adapter(i915, intel_hdmi->ddc_bus); > > if (!sink_scrambling->supported) > > return true; > > @@ -2861,7 +2861,7 @@ bool intel_hdmi_handle_sink_scrambling(struct intel_encoder *encoder, > > drm_scdc_set_scrambling(adapter, scrambling); > > } > > -static u8 chv_port_to_ddc_pin(struct drm_i915_private *dev_priv, enum port port) > > +static u8 chv_port_to_ddc_pin(struct drm_i915_private *i915, enum port port) > > { > > u8 ddc_pin; > > @@ -2883,7 +2883,7 @@ static u8 chv_port_to_ddc_pin(struct drm_i915_private *dev_priv, enum port port) > > return ddc_pin; > > } > > -static u8 bxt_port_to_ddc_pin(struct drm_i915_private *dev_priv, enum port port) > > +static u8 bxt_port_to_ddc_pin(struct drm_i915_private *i915, enum port port) > > { > > u8 ddc_pin; > > @@ -2902,7 +2902,7 @@ static u8 bxt_port_to_ddc_pin(struct drm_i915_private *dev_priv, enum port port) > > return ddc_pin; > > } > > -static u8 cnp_port_to_ddc_pin(struct drm_i915_private *dev_priv, > > +static u8 cnp_port_to_ddc_pin(struct drm_i915_private *i915, > > enum port port) > > { > > u8 ddc_pin; > > @@ -2928,7 +2928,7 @@ static u8 cnp_port_to_ddc_pin(struct drm_i915_private *dev_priv, > > return ddc_pin; > > } > > -static u8 icl_port_to_ddc_pin(struct drm_i915_private *dev_priv, enum port port) > > +static u8 icl_port_to_ddc_pin(struct drm_i915_private *i915, enum port port) > > { > > u8 ddc_pin; > > @@ -2959,7 +2959,7 @@ static u8 icl_port_to_ddc_pin(struct drm_i915_private *dev_priv, enum port port) > > return ddc_pin; > > } > > -static u8 g4x_port_to_ddc_pin(struct drm_i915_private *dev_priv, > > +static u8 g4x_port_to_ddc_pin(struct drm_i915_private *i915, > > enum port port) > > { > > u8 ddc_pin; > > @@ -2982,11 +2982,11 @@ static u8 g4x_port_to_ddc_pin(struct drm_i915_private *dev_priv, > > return ddc_pin; > > } > > -static u8 intel_hdmi_ddc_pin(struct drm_i915_private *dev_priv, > > +static u8 intel_hdmi_ddc_pin(struct drm_i915_private *i915, > > enum port port) > > { > > const struct ddi_vbt_port_info *info = > > - &dev_priv->vbt.ddi_port_info[port]; > > + &i915->vbt.ddi_port_info[port]; > > u8 ddc_pin; > > if (info->alternate_ddc_pin) { > > @@ -2995,16 +2995,16 @@ static u8 intel_hdmi_ddc_pin(struct drm_i915_private *dev_priv, > > return info->alternate_ddc_pin; > > } > > - if (HAS_PCH_ICP(dev_priv)) > > - ddc_pin = icl_port_to_ddc_pin(dev_priv, port); > > - else if (HAS_PCH_CNP(dev_priv)) > > - ddc_pin = cnp_port_to_ddc_pin(dev_priv, port); > > - else if (IS_GEN9_LP(dev_priv)) > > - ddc_pin = bxt_port_to_ddc_pin(dev_priv, port); > > - else if (IS_CHERRYVIEW(dev_priv)) > > - ddc_pin = chv_port_to_ddc_pin(dev_priv, port); > > + if (HAS_PCH_ICP(i915)) > > + ddc_pin = icl_port_to_ddc_pin(i915, port); > > + else if (HAS_PCH_CNP(i915)) > > + ddc_pin = cnp_port_to_ddc_pin(i915, port); > > + else if (IS_GEN9_LP(i915)) > > + ddc_pin = bxt_port_to_ddc_pin(i915, port); > > + else if (IS_CHERRYVIEW(i915)) > > + ddc_pin = chv_port_to_ddc_pin(i915, port); > > else > > - ddc_pin = g4x_port_to_ddc_pin(dev_priv, port); > > + ddc_pin = g4x_port_to_ddc_pin(i915, port); > > DRM_DEBUG_KMS("Using DDC pin 0x%x for port %c (platform default)\n", > > ddc_pin, port_name(port)); > > @@ -3014,20 +3014,20 @@ static u8 intel_hdmi_ddc_pin(struct drm_i915_private *dev_priv, > > void intel_infoframe_init(struct intel_digital_port *intel_dig_port) > > { > > - struct drm_i915_private *dev_priv = > > + struct drm_i915_private *i915 = > > to_i915(intel_dig_port->base.base.dev); > > - if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) { > > + if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) { > > intel_dig_port->write_infoframe = vlv_write_infoframe; > > intel_dig_port->read_infoframe = vlv_read_infoframe; > > intel_dig_port->set_infoframes = vlv_set_infoframes; > > intel_dig_port->infoframes_enabled = vlv_infoframes_enabled; > > - } else if (IS_G4X(dev_priv)) { > > + } else if (IS_G4X(i915)) { > > intel_dig_port->write_infoframe = g4x_write_infoframe; > > intel_dig_port->read_infoframe = g4x_read_infoframe; > > intel_dig_port->set_infoframes = g4x_set_infoframes; > > intel_dig_port->infoframes_enabled = g4x_infoframes_enabled; > > - } else if (HAS_DDI(dev_priv)) { > > + } else if (HAS_DDI(i915)) { > > if (intel_dig_port->lspcon.active) { > > intel_dig_port->write_infoframe = lspcon_write_infoframe; > > intel_dig_port->read_infoframe = lspcon_read_infoframe; > > @@ -3039,7 +3039,7 @@ void intel_infoframe_init(struct intel_digital_port *intel_dig_port) > > intel_dig_port->set_infoframes = hsw_set_infoframes; > > intel_dig_port->infoframes_enabled = hsw_infoframes_enabled; > > } > > - } else if (HAS_PCH_IBX(dev_priv)) { > > + } else if (HAS_PCH_IBX(i915)) { > > intel_dig_port->write_infoframe = ibx_write_infoframe; > > intel_dig_port->read_infoframe = ibx_read_infoframe; > > intel_dig_port->set_infoframes = ibx_set_infoframes; > > @@ -3059,7 +3059,7 @@ void intel_hdmi_init_connector(struct intel_digital_port *intel_dig_port, > > struct intel_hdmi *intel_hdmi = &intel_dig_port->hdmi; > > struct intel_encoder *intel_encoder = &intel_dig_port->base; > > struct drm_device *dev = intel_encoder->base.dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > enum port port = intel_encoder->port; > > DRM_DEBUG_KMS("Adding HDMI connector on port %c\n", > > @@ -3078,16 +3078,16 @@ void intel_hdmi_init_connector(struct intel_digital_port *intel_dig_port, > > connector->doublescan_allowed = 0; > > connector->stereo_allowed = 1; > > - if (INTEL_GEN(dev_priv) >= 10 || IS_GEMINILAKE(dev_priv)) > > + if (INTEL_GEN(i915) >= 10 || IS_GEMINILAKE(i915)) > > connector->ycbcr_420_allowed = true; > > - intel_hdmi->ddc_bus = intel_hdmi_ddc_pin(dev_priv, port); > > + intel_hdmi->ddc_bus = intel_hdmi_ddc_pin(i915, port); > > if (WARN_ON(port == PORT_A)) > > return; > > - intel_encoder->hpd_pin = intel_hpd_pin_default(dev_priv, port); > > + intel_encoder->hpd_pin = intel_hpd_pin_default(i915, port); > > - if (HAS_DDI(dev_priv)) > > + if (HAS_DDI(i915)) > > intel_connector->get_hw_state = intel_ddi_connector_get_hw_state; > > else > > intel_connector->get_hw_state = intel_connector_get_hw_state; > > @@ -3097,7 +3097,7 @@ void intel_hdmi_init_connector(struct intel_digital_port *intel_dig_port, > > intel_connector_attach_encoder(intel_connector, intel_encoder); > > intel_hdmi->attached_connector = intel_connector; > > - if (is_hdcp_supported(dev_priv, port)) { > > + if (is_hdcp_supported(i915, port)) { > > int ret = intel_hdcp_init(intel_connector, > > &intel_hdmi_hdcp_shim); > > if (ret) > > @@ -3108,7 +3108,7 @@ void intel_hdmi_init_connector(struct intel_digital_port *intel_dig_port, > > * 0xd. Failure to do so will result in spurious interrupts being > > * generated on the port when a cable is not attached. > > */ > > - if (IS_G45(dev_priv)) { > > + if (IS_G45(i915)) { > > u32 temp = I915_READ(PEG_BAND_GAP_DATA); > > I915_WRITE(PEG_BAND_GAP_DATA, (temp & ~0xf) | 0xd); > > } > > @@ -3119,7 +3119,7 @@ void intel_hdmi_init_connector(struct intel_digital_port *intel_dig_port, > > DRM_DEBUG_KMS("CEC notifier get failed\n"); > > } > > -void intel_hdmi_init(struct drm_i915_private *dev_priv, > > +void intel_hdmi_init(struct drm_i915_private *i915, > > i915_reg_t hdmi_reg, enum port port) > > { > > struct intel_digital_port *intel_dig_port; > > @@ -3138,13 +3138,13 @@ void intel_hdmi_init(struct drm_i915_private *dev_priv, > > intel_encoder = &intel_dig_port->base; > > - drm_encoder_init(&dev_priv->drm, &intel_encoder->base, > > + drm_encoder_init(&i915->drm, &intel_encoder->base, > > &intel_hdmi_enc_funcs, DRM_MODE_ENCODER_TMDS, > > "HDMI %c", port_name(port)); > > intel_encoder->hotplug = intel_encoder_hotplug; > > intel_encoder->compute_config = intel_hdmi_compute_config; > > - if (HAS_PCH_SPLIT(dev_priv)) { > > + if (HAS_PCH_SPLIT(i915)) { > > intel_encoder->disable = pch_disable_hdmi; > > intel_encoder->post_disable = pch_post_disable_hdmi; > > } else { > > @@ -3152,22 +3152,22 @@ void intel_hdmi_init(struct drm_i915_private *dev_priv, > > } > > intel_encoder->get_hw_state = intel_hdmi_get_hw_state; > > intel_encoder->get_config = intel_hdmi_get_config; > > - if (IS_CHERRYVIEW(dev_priv)) { > > + if (IS_CHERRYVIEW(i915)) { > > intel_encoder->pre_pll_enable = chv_hdmi_pre_pll_enable; > > intel_encoder->pre_enable = chv_hdmi_pre_enable; > > intel_encoder->enable = vlv_enable_hdmi; > > intel_encoder->post_disable = chv_hdmi_post_disable; > > intel_encoder->post_pll_disable = chv_hdmi_post_pll_disable; > > - } else if (IS_VALLEYVIEW(dev_priv)) { > > + } else if (IS_VALLEYVIEW(i915)) { > > intel_encoder->pre_pll_enable = vlv_hdmi_pre_pll_enable; > > intel_encoder->pre_enable = vlv_hdmi_pre_enable; > > intel_encoder->enable = vlv_enable_hdmi; > > intel_encoder->post_disable = vlv_hdmi_post_disable; > > } else { > > intel_encoder->pre_enable = intel_hdmi_pre_enable; > > - if (HAS_PCH_CPT(dev_priv)) > > + if (HAS_PCH_CPT(i915)) > > intel_encoder->enable = cpt_enable_hdmi; > > - else if (HAS_PCH_IBX(dev_priv)) > > + else if (HAS_PCH_IBX(i915)) > > intel_encoder->enable = ibx_enable_hdmi; > > else > > intel_encoder->enable = g4x_enable_hdmi; > > @@ -3176,7 +3176,7 @@ void intel_hdmi_init(struct drm_i915_private *dev_priv, > > intel_encoder->type = INTEL_OUTPUT_HDMI; > > intel_encoder->power_domain = intel_port_to_power_domain(port); > > intel_encoder->port = port; > > - if (IS_CHERRYVIEW(dev_priv)) { > > + if (IS_CHERRYVIEW(i915)) { > > if (port == PORT_D) > > intel_encoder->crtc_mask = 1 << 2; > > else > > @@ -3190,7 +3190,7 @@ void intel_hdmi_init(struct drm_i915_private *dev_priv, > > * to work on real hardware. And since g4x can send infoframes to > > * only one port anyway, nothing is lost by allowing it. > > */ > > - if (IS_G4X(dev_priv)) > > + if (IS_G4X(i915)) > > intel_encoder->cloneable |= 1 << INTEL_OUTPUT_HDMI; > > intel_dig_port->hdmi.hdmi_reg = hdmi_reg; > > @@ -3199,6 +3199,6 @@ void intel_hdmi_init(struct drm_i915_private *dev_priv, > > intel_infoframe_init(intel_dig_port); > > - intel_dig_port->aux_ch = intel_bios_port_aux_ch(dev_priv, port); > > + intel_dig_port->aux_ch = intel_bios_port_aux_ch(i915, port); > > intel_hdmi_init_connector(intel_dig_port, intel_connector); > > } > > diff --git a/drivers/gpu/drm/i915/intel_hdmi.h b/drivers/gpu/drm/i915/intel_hdmi.h > > index 106c2e0bc3c9..46f03356b54b 100644 > > --- a/drivers/gpu/drm/i915/intel_hdmi.h > > +++ b/drivers/gpu/drm/i915/intel_hdmi.h > > @@ -24,7 +24,7 @@ struct intel_hdmi; > > struct drm_connector_state; > > union hdmi_infoframe; > > -void intel_hdmi_init(struct drm_i915_private *dev_priv, i915_reg_t hdmi_reg, > > +void intel_hdmi_init(struct drm_i915_private *i915, i915_reg_t hdmi_reg, > > enum port port); > > void intel_hdmi_init_connector(struct intel_digital_port *intel_dig_port, > > struct intel_connector *intel_connector); > > diff --git a/drivers/gpu/drm/i915/intel_hotplug.c b/drivers/gpu/drm/i915/intel_hotplug.c > > index ff9eb3c855d3..a551b9ab2e17 100644 > > --- a/drivers/gpu/drm/i915/intel_hotplug.c > > +++ b/drivers/gpu/drm/i915/intel_hotplug.c > > @@ -78,7 +78,7 @@ > > /** > > * intel_hpd_pin_default - return default pin associated with certain port. > > - * @dev_priv: private driver data pointer > > + * @i915: private driver data pointer > > * @port: the hpd port to get associated pin > > * > > * It is only valid and used by digital port encoder. > > @@ -86,7 +86,7 @@ > > * Return pin that is associatade with @port and HDP_NONE if no pin is > > * hard associated with that @port. > > */ > > -enum hpd_pin intel_hpd_pin_default(struct drm_i915_private *dev_priv, > > +enum hpd_pin intel_hpd_pin_default(struct drm_i915_private *i915, > > enum port port) > > { > > switch (port) { > > @@ -101,7 +101,7 @@ enum hpd_pin intel_hpd_pin_default(struct drm_i915_private *dev_priv, > > case PORT_E: > > return HPD_PORT_E; > > case PORT_F: > > - if (IS_CNL_WITH_PORT_F(dev_priv)) > > + if (IS_CNL_WITH_PORT_F(i915)) > > return HPD_PORT_E; > > return HPD_PORT_F; > > default: > > @@ -115,7 +115,7 @@ enum hpd_pin intel_hpd_pin_default(struct drm_i915_private *dev_priv, > > /** > > * intel_hpd_irq_storm_detect - gather stats and detect HPD IRQ storm on a pin > > - * @dev_priv: private driver data pointer > > + * @i915: private driver data pointer > > * @pin: the pin to gather stats on > > * @long_hpd: whether the HPD IRQ was long or short > > * > > @@ -124,13 +124,13 @@ enum hpd_pin intel_hpd_pin_default(struct drm_i915_private *dev_priv, > > * responsible for further action. > > * > > * The number of IRQs that are allowed within @HPD_STORM_DETECT_PERIOD is > > - * stored in @dev_priv->hotplug.hpd_storm_threshold which defaults to > > + * stored in @i915->hotplug.hpd_storm_threshold which defaults to > > * @HPD_STORM_DEFAULT_THRESHOLD. Long IRQs count as +10 to this threshold, and > > * short IRQs count as +1. If this threshold is exceeded, it's considered an > > * IRQ storm and the IRQ state is set to @HPD_MARK_DISABLED. > > * > > * By default, most systems will only count long IRQs towards > > - * &dev_priv->hotplug.hpd_storm_threshold. However, some older systems also > > + * &i915->hotplug.hpd_storm_threshold. However, some older systems also > > * suffer from short IRQ storms and must also track these. Because short IRQ > > * storms are naturally caused by sideband interactions with DP MST devices, > > * short IRQ detection is only enabled for systems without DP MST support. > > @@ -142,10 +142,10 @@ enum hpd_pin intel_hpd_pin_default(struct drm_i915_private *dev_priv, > > * > > * Return true if an IRQ storm was detected on @pin. > > */ > > -static bool intel_hpd_irq_storm_detect(struct drm_i915_private *dev_priv, > > +static bool intel_hpd_irq_storm_detect(struct drm_i915_private *i915, > > enum hpd_pin pin, bool long_hpd) > > { > > - struct i915_hotplug *hpd = &dev_priv->hotplug; > > + struct i915_hotplug *hpd = &i915->hotplug; > > unsigned long start = hpd->stats[pin].last_jiffies; > > unsigned long end = start + msecs_to_jiffies(HPD_STORM_DETECT_PERIOD); > > const int increment = long_hpd ? 10 : 1; > > @@ -153,7 +153,7 @@ static bool intel_hpd_irq_storm_detect(struct drm_i915_private *dev_priv, > > bool storm = false; > > if (!threshold || > > - (!long_hpd && !dev_priv->hotplug.hpd_short_storm_enabled)) > > + (!long_hpd && !i915->hotplug.hpd_short_storm_enabled)) > > return false; > > if (!time_in_range(jiffies, start, end)) { > > @@ -175,9 +175,9 @@ static bool intel_hpd_irq_storm_detect(struct drm_i915_private *dev_priv, > > } > > static void > > -intel_hpd_irq_storm_switch_to_polling(struct drm_i915_private *dev_priv) > > +intel_hpd_irq_storm_switch_to_polling(struct drm_i915_private *i915) > > { > > - struct drm_device *dev = &dev_priv->drm; > > + struct drm_device *dev = &i915->drm; > > struct intel_connector *intel_connector; > > struct intel_encoder *intel_encoder; > > struct drm_connector *connector; > > @@ -185,7 +185,7 @@ intel_hpd_irq_storm_switch_to_polling(struct drm_i915_private *dev_priv) > > enum hpd_pin pin; > > bool hpd_disabled = false; > > - lockdep_assert_held(&dev_priv->irq_lock); > > + lockdep_assert_held(&i915->irq_lock); > > drm_connector_list_iter_begin(dev, &conn_iter); > > drm_for_each_connector_iter(connector, &conn_iter) { > > @@ -199,14 +199,14 @@ intel_hpd_irq_storm_switch_to_polling(struct drm_i915_private *dev_priv) > > pin = intel_encoder->hpd_pin; > > if (pin == HPD_NONE || > > - dev_priv->hotplug.stats[pin].state != HPD_MARK_DISABLED) > > + i915->hotplug.stats[pin].state != HPD_MARK_DISABLED) > > continue; > > DRM_INFO("HPD interrupt storm detected on connector %s: " > > "switching from hotplug detection to polling\n", > > connector->name); > > - dev_priv->hotplug.stats[pin].state = HPD_DISABLED; > > + i915->hotplug.stats[pin].state = HPD_DISABLED; > > connector->polled = DRM_CONNECTOR_POLL_CONNECT > > | DRM_CONNECTOR_POLL_DISCONNECT; > > hpd_disabled = true; > > @@ -216,31 +216,31 @@ intel_hpd_irq_storm_switch_to_polling(struct drm_i915_private *dev_priv) > > /* Enable polling and queue hotplug re-enabling. */ > > if (hpd_disabled) { > > drm_kms_helper_poll_enable(dev); > > - mod_delayed_work(system_wq, &dev_priv->hotplug.reenable_work, > > + mod_delayed_work(system_wq, &i915->hotplug.reenable_work, > > msecs_to_jiffies(HPD_STORM_REENABLE_DELAY)); > > } > > } > > static void intel_hpd_irq_storm_reenable_work(struct work_struct *work) > > { > > - struct drm_i915_private *dev_priv = > > - container_of(work, typeof(*dev_priv), > > + struct drm_i915_private *i915 = > > + container_of(work, typeof(*i915), > > hotplug.reenable_work.work); > > - struct drm_device *dev = &dev_priv->drm; > > + struct drm_device *dev = &i915->drm; > > intel_wakeref_t wakeref; > > enum hpd_pin pin; > > - wakeref = intel_runtime_pm_get(dev_priv); > > + wakeref = intel_runtime_pm_get(i915); > > - spin_lock_irq(&dev_priv->irq_lock); > > + spin_lock_irq(&i915->irq_lock); > > for_each_hpd_pin(pin) { > > struct drm_connector *connector; > > struct drm_connector_list_iter conn_iter; > > - if (dev_priv->hotplug.stats[pin].state != HPD_DISABLED) > > + if (i915->hotplug.stats[pin].state != HPD_DISABLED) > > continue; > > - dev_priv->hotplug.stats[pin].state = HPD_ENABLED; > > + i915->hotplug.stats[pin].state = HPD_ENABLED; > > drm_connector_list_iter_begin(dev, &conn_iter); > > drm_for_each_connector_iter(connector, &conn_iter) { > > @@ -259,11 +259,11 @@ static void intel_hpd_irq_storm_reenable_work(struct work_struct *work) > > } > > drm_connector_list_iter_end(&conn_iter); > > } > > - if (dev_priv->display_irqs_enabled && dev_priv->display.hpd_irq_setup) > > - dev_priv->display.hpd_irq_setup(dev_priv); > > - spin_unlock_irq(&dev_priv->irq_lock); > > + if (i915->display_irqs_enabled && i915->display.hpd_irq_setup) > > + i915->display.hpd_irq_setup(i915); > > + spin_unlock_irq(&i915->irq_lock); > > - intel_runtime_pm_put(dev_priv, wakeref); > > + intel_runtime_pm_put(i915, wakeref); > > } > > bool intel_encoder_hotplug(struct intel_encoder *encoder, > > @@ -298,20 +298,20 @@ static bool intel_encoder_has_hpd_pulse(struct intel_encoder *encoder) > > static void i915_digport_work_func(struct work_struct *work) > > { > > - struct drm_i915_private *dev_priv = > > + struct drm_i915_private *i915 = > > container_of(work, struct drm_i915_private, hotplug.dig_port_work); > > u32 long_port_mask, short_port_mask; > > struct intel_encoder *encoder; > > u32 old_bits = 0; > > - spin_lock_irq(&dev_priv->irq_lock); > > - long_port_mask = dev_priv->hotplug.long_port_mask; > > - dev_priv->hotplug.long_port_mask = 0; > > - short_port_mask = dev_priv->hotplug.short_port_mask; > > - dev_priv->hotplug.short_port_mask = 0; > > - spin_unlock_irq(&dev_priv->irq_lock); > > + spin_lock_irq(&i915->irq_lock); > > + long_port_mask = i915->hotplug.long_port_mask; > > + i915->hotplug.long_port_mask = 0; > > + short_port_mask = i915->hotplug.short_port_mask; > > + i915->hotplug.short_port_mask = 0; > > + spin_unlock_irq(&i915->irq_lock); > > - for_each_intel_encoder(&dev_priv->drm, encoder) { > > + for_each_intel_encoder(&i915->drm, encoder) { > > struct intel_digital_port *dig_port; > > enum port port = encoder->port; > > bool long_hpd, short_hpd; > > @@ -336,10 +336,10 @@ static void i915_digport_work_func(struct work_struct *work) > > } > > if (old_bits) { > > - spin_lock_irq(&dev_priv->irq_lock); > > - dev_priv->hotplug.event_bits |= old_bits; > > - spin_unlock_irq(&dev_priv->irq_lock); > > - schedule_work(&dev_priv->hotplug.hotplug_work); > > + spin_lock_irq(&i915->irq_lock); > > + i915->hotplug.event_bits |= old_bits; > > + spin_unlock_irq(&i915->irq_lock); > > + schedule_work(&i915->hotplug.hotplug_work); > > } > > } > > @@ -348,9 +348,9 @@ static void i915_digport_work_func(struct work_struct *work) > > */ > > static void i915_hotplug_work_func(struct work_struct *work) > > { > > - struct drm_i915_private *dev_priv = > > + struct drm_i915_private *i915 = > > container_of(work, struct drm_i915_private, hotplug.hotplug_work); > > - struct drm_device *dev = &dev_priv->drm; > > + struct drm_device *dev = &i915->drm; > > struct intel_connector *intel_connector; > > struct intel_encoder *intel_encoder; > > struct drm_connector *connector; > > @@ -361,15 +361,15 @@ static void i915_hotplug_work_func(struct work_struct *work) > > mutex_lock(&dev->mode_config.mutex); > > DRM_DEBUG_KMS("running encoder hotplug functions\n"); > > - spin_lock_irq(&dev_priv->irq_lock); > > + spin_lock_irq(&i915->irq_lock); > > - hpd_event_bits = dev_priv->hotplug.event_bits; > > - dev_priv->hotplug.event_bits = 0; > > + hpd_event_bits = i915->hotplug.event_bits; > > + i915->hotplug.event_bits = 0; > > /* Enable polling for connectors which had HPD IRQ storms */ > > - intel_hpd_irq_storm_switch_to_polling(dev_priv); > > + intel_hpd_irq_storm_switch_to_polling(i915); > > - spin_unlock_irq(&dev_priv->irq_lock); > > + spin_unlock_irq(&i915->irq_lock); > > drm_connector_list_iter_begin(dev, &conn_iter); > > drm_for_each_connector_iter(connector, &conn_iter) { > > @@ -395,7 +395,7 @@ static void i915_hotplug_work_func(struct work_struct *work) > > /** > > * intel_hpd_irq_handler - main hotplug irq handler > > - * @dev_priv: drm_i915_private > > + * @i915: drm_i915_private > > * @pin_mask: a mask of hpd pins that have triggered the irq > > * @long_mask: a mask of hpd pins that may be long hpd pulses > > * > > @@ -409,7 +409,7 @@ static void i915_hotplug_work_func(struct work_struct *work) > > * Here, we do hotplug irq storm detection and mitigation, and pass further > > * processing to appropriate bottom halves. > > */ > > -void intel_hpd_irq_handler(struct drm_i915_private *dev_priv, > > +void intel_hpd_irq_handler(struct drm_i915_private *i915, > > u32 pin_mask, u32 long_mask) > > { > > struct intel_encoder *encoder; > > @@ -422,7 +422,7 @@ void intel_hpd_irq_handler(struct drm_i915_private *dev_priv, > > if (!pin_mask) > > return; > > - spin_lock(&dev_priv->irq_lock); > > + spin_lock(&i915->irq_lock); > > /* > > * Determine whether ->hpd_pulse() exists for each pin, and > > @@ -430,7 +430,7 @@ void intel_hpd_irq_handler(struct drm_i915_private *dev_priv, > > * as each pin may have up to two encoders (HDMI and DP) and > > * only the one of them (DP) will have ->hpd_pulse(). > > */ > > - for_each_intel_encoder(&dev_priv->drm, encoder) { > > + for_each_intel_encoder(&i915->drm, encoder) { > > bool has_hpd_pulse = intel_encoder_has_hpd_pulse(encoder); > > enum port port = encoder->port; > > bool long_hpd; > > @@ -450,10 +450,10 @@ void intel_hpd_irq_handler(struct drm_i915_private *dev_priv, > > if (long_hpd) { > > long_hpd_pulse_mask |= BIT(pin); > > - dev_priv->hotplug.long_port_mask |= BIT(port); > > + i915->hotplug.long_port_mask |= BIT(port); > > } else { > > short_hpd_pulse_mask |= BIT(pin); > > - dev_priv->hotplug.short_port_mask |= BIT(port); > > + i915->hotplug.short_port_mask |= BIT(port); > > } > > } > > @@ -464,19 +464,19 @@ void intel_hpd_irq_handler(struct drm_i915_private *dev_priv, > > if (!(BIT(pin) & pin_mask)) > > continue; > > - if (dev_priv->hotplug.stats[pin].state == HPD_DISABLED) { > > + if (i915->hotplug.stats[pin].state == HPD_DISABLED) { > > /* > > * On GMCH platforms the interrupt mask bits only > > * prevent irq generation, not the setting of the > > * hotplug bits itself. So only WARN about unexpected > > * interrupts on saner platforms. > > */ > > - WARN_ONCE(!HAS_GMCH(dev_priv), > > + WARN_ONCE(!HAS_GMCH(i915), > > "Received HPD interrupt on pin %d although disabled\n", pin); > > continue; > > } > > - if (dev_priv->hotplug.stats[pin].state != HPD_ENABLED) > > + if (i915->hotplug.stats[pin].state != HPD_ENABLED) > > continue; > > /* > > @@ -487,13 +487,13 @@ void intel_hpd_irq_handler(struct drm_i915_private *dev_priv, > > if (((short_hpd_pulse_mask | long_hpd_pulse_mask) & BIT(pin))) { > > long_hpd = long_hpd_pulse_mask & BIT(pin); > > } else { > > - dev_priv->hotplug.event_bits |= BIT(pin); > > + i915->hotplug.event_bits |= BIT(pin); > > long_hpd = true; > > queue_hp = true; > > } > > - if (intel_hpd_irq_storm_detect(dev_priv, pin, long_hpd)) { > > - dev_priv->hotplug.event_bits &= ~BIT(pin); > > + if (intel_hpd_irq_storm_detect(i915, pin, long_hpd)) { > > + i915->hotplug.event_bits &= ~BIT(pin); > > storm_detected = true; > > queue_hp = true; > > } > > @@ -503,9 +503,9 @@ void intel_hpd_irq_handler(struct drm_i915_private *dev_priv, > > * Disable any IRQs that storms were detected on. Polling enablement > > * happens later in our hotplug work. > > */ > > - if (storm_detected && dev_priv->display_irqs_enabled) > > - dev_priv->display.hpd_irq_setup(dev_priv); > > - spin_unlock(&dev_priv->irq_lock); > > + if (storm_detected && i915->display_irqs_enabled) > > + i915->display.hpd_irq_setup(i915); > > + spin_unlock(&i915->irq_lock); > > /* > > * Our hotplug handler can grab modeset locks (by calling down into the > > @@ -514,14 +514,14 @@ void intel_hpd_irq_handler(struct drm_i915_private *dev_priv, > > * deadlock. > > */ > > if (queue_dig) > > - queue_work(dev_priv->hotplug.dp_wq, &dev_priv->hotplug.dig_port_work); > > + queue_work(i915->hotplug.dp_wq, &i915->hotplug.dig_port_work); > > if (queue_hp) > > - schedule_work(&dev_priv->hotplug.hotplug_work); > > + schedule_work(&i915->hotplug.hotplug_work); > > } > > /** > > * intel_hpd_init - initializes and enables hpd support > > - * @dev_priv: i915 device instance > > + * @i915: i915 device instance > > * > > * This function enables the hotplug support. It requires that interrupts have > > * already been enabled with intel_irq_init_hw(). From this point on hotplug and > > @@ -533,43 +533,43 @@ void intel_hpd_irq_handler(struct drm_i915_private *dev_priv, > > * > > * Also see: intel_hpd_poll_init(), which enables connector polling > > */ > > -void intel_hpd_init(struct drm_i915_private *dev_priv) > > +void intel_hpd_init(struct drm_i915_private *i915) > > { > > int i; > > for_each_hpd_pin(i) { > > - dev_priv->hotplug.stats[i].count = 0; > > - dev_priv->hotplug.stats[i].state = HPD_ENABLED; > > + i915->hotplug.stats[i].count = 0; > > + i915->hotplug.stats[i].state = HPD_ENABLED; > > } > > - WRITE_ONCE(dev_priv->hotplug.poll_enabled, false); > > - schedule_work(&dev_priv->hotplug.poll_init_work); > > + WRITE_ONCE(i915->hotplug.poll_enabled, false); > > + schedule_work(&i915->hotplug.poll_init_work); > > /* > > * Interrupt setup is already guaranteed to be single-threaded, this is > > * just to make the assert_spin_locked checks happy. > > */ > > - if (dev_priv->display_irqs_enabled && dev_priv->display.hpd_irq_setup) { > > - spin_lock_irq(&dev_priv->irq_lock); > > - if (dev_priv->display_irqs_enabled) > > - dev_priv->display.hpd_irq_setup(dev_priv); > > - spin_unlock_irq(&dev_priv->irq_lock); > > + if (i915->display_irqs_enabled && i915->display.hpd_irq_setup) { > > + spin_lock_irq(&i915->irq_lock); > > + if (i915->display_irqs_enabled) > > + i915->display.hpd_irq_setup(i915); > > + spin_unlock_irq(&i915->irq_lock); > > } > > } > > static void i915_hpd_poll_init_work(struct work_struct *work) > > { > > - struct drm_i915_private *dev_priv = > > + struct drm_i915_private *i915 = > > container_of(work, struct drm_i915_private, > > hotplug.poll_init_work); > > - struct drm_device *dev = &dev_priv->drm; > > + struct drm_device *dev = &i915->drm; > > struct drm_connector *connector; > > struct drm_connector_list_iter conn_iter; > > bool enabled; > > mutex_lock(&dev->mode_config.mutex); > > - enabled = READ_ONCE(dev_priv->hotplug.poll_enabled); > > + enabled = READ_ONCE(i915->hotplug.poll_enabled); > > drm_connector_list_iter_begin(dev, &conn_iter); > > drm_for_each_connector_iter(connector, &conn_iter) { > > @@ -582,7 +582,7 @@ static void i915_hpd_poll_init_work(struct work_struct *work) > > if (intel_connector->mst_port) > > continue; > > - if (!connector->polled && I915_HAS_HOTPLUG(dev_priv) && > > + if (!connector->polled && I915_HAS_HOTPLUG(i915) && > > intel_connector->encoder->hpd_pin > HPD_NONE) { > > connector->polled = enabled ? > > DRM_CONNECTOR_POLL_CONNECT | > > @@ -607,7 +607,7 @@ static void i915_hpd_poll_init_work(struct work_struct *work) > > /** > > * intel_hpd_poll_init - enables/disables polling for connectors with hpd > > - * @dev_priv: i915 device instance > > + * @i915: i915 device instance > > * > > * This function enables polling for all connectors, regardless of whether or > > * not they support hotplug detection. Under certain conditions HPD may not be > > @@ -621,9 +621,9 @@ static void i915_hpd_poll_init_work(struct work_struct *work) > > * > > * Also see: intel_hpd_init(), which restores hpd handling. > > */ > > -void intel_hpd_poll_init(struct drm_i915_private *dev_priv) > > +void intel_hpd_poll_init(struct drm_i915_private *i915) > > { > > - WRITE_ONCE(dev_priv->hotplug.poll_enabled, true); > > + WRITE_ONCE(i915->hotplug.poll_enabled, true); > > /* > > * We might already be holding dev->mode_config.mutex, so do this in a > > @@ -631,57 +631,57 @@ void intel_hpd_poll_init(struct drm_i915_private *dev_priv) > > * As well, there's no issue if we race here since we always reschedule > > * this worker anyway > > */ > > - schedule_work(&dev_priv->hotplug.poll_init_work); > > + schedule_work(&i915->hotplug.poll_init_work); > > } > > -void intel_hpd_init_work(struct drm_i915_private *dev_priv) > > +void intel_hpd_init_work(struct drm_i915_private *i915) > > { > > - INIT_WORK(&dev_priv->hotplug.hotplug_work, i915_hotplug_work_func); > > - INIT_WORK(&dev_priv->hotplug.dig_port_work, i915_digport_work_func); > > - INIT_WORK(&dev_priv->hotplug.poll_init_work, i915_hpd_poll_init_work); > > - INIT_DELAYED_WORK(&dev_priv->hotplug.reenable_work, > > + INIT_WORK(&i915->hotplug.hotplug_work, i915_hotplug_work_func); > > + INIT_WORK(&i915->hotplug.dig_port_work, i915_digport_work_func); > > + INIT_WORK(&i915->hotplug.poll_init_work, i915_hpd_poll_init_work); > > + INIT_DELAYED_WORK(&i915->hotplug.reenable_work, > > intel_hpd_irq_storm_reenable_work); > > } > > -void intel_hpd_cancel_work(struct drm_i915_private *dev_priv) > > +void intel_hpd_cancel_work(struct drm_i915_private *i915) > > { > > - spin_lock_irq(&dev_priv->irq_lock); > > + spin_lock_irq(&i915->irq_lock); > > - dev_priv->hotplug.long_port_mask = 0; > > - dev_priv->hotplug.short_port_mask = 0; > > - dev_priv->hotplug.event_bits = 0; > > + i915->hotplug.long_port_mask = 0; > > + i915->hotplug.short_port_mask = 0; > > + i915->hotplug.event_bits = 0; > > - spin_unlock_irq(&dev_priv->irq_lock); > > + spin_unlock_irq(&i915->irq_lock); > > - cancel_work_sync(&dev_priv->hotplug.dig_port_work); > > - cancel_work_sync(&dev_priv->hotplug.hotplug_work); > > - cancel_work_sync(&dev_priv->hotplug.poll_init_work); > > - cancel_delayed_work_sync(&dev_priv->hotplug.reenable_work); > > + cancel_work_sync(&i915->hotplug.dig_port_work); > > + cancel_work_sync(&i915->hotplug.hotplug_work); > > + cancel_work_sync(&i915->hotplug.poll_init_work); > > + cancel_delayed_work_sync(&i915->hotplug.reenable_work); > > } > > -bool intel_hpd_disable(struct drm_i915_private *dev_priv, enum hpd_pin pin) > > +bool intel_hpd_disable(struct drm_i915_private *i915, enum hpd_pin pin) > > { > > bool ret = false; > > if (pin == HPD_NONE) > > return false; > > - spin_lock_irq(&dev_priv->irq_lock); > > - if (dev_priv->hotplug.stats[pin].state == HPD_ENABLED) { > > - dev_priv->hotplug.stats[pin].state = HPD_DISABLED; > > + spin_lock_irq(&i915->irq_lock); > > + if (i915->hotplug.stats[pin].state == HPD_ENABLED) { > > + i915->hotplug.stats[pin].state = HPD_DISABLED; > > ret = true; > > } > > - spin_unlock_irq(&dev_priv->irq_lock); > > + spin_unlock_irq(&i915->irq_lock); > > return ret; > > } > > -void intel_hpd_enable(struct drm_i915_private *dev_priv, enum hpd_pin pin) > > +void intel_hpd_enable(struct drm_i915_private *i915, enum hpd_pin pin) > > { > > if (pin == HPD_NONE) > > return; > > - spin_lock_irq(&dev_priv->irq_lock); > > - dev_priv->hotplug.stats[pin].state = HPD_ENABLED; > > - spin_unlock_irq(&dev_priv->irq_lock); > > + spin_lock_irq(&i915->irq_lock); > > + i915->hotplug.stats[pin].state = HPD_ENABLED; > > + spin_unlock_irq(&i915->irq_lock); > > } > > diff --git a/drivers/gpu/drm/i915/intel_hotplug.h b/drivers/gpu/drm/i915/intel_hotplug.h > > index 805f897dbb7a..2fcc3148b369 100644 > > --- a/drivers/gpu/drm/i915/intel_hotplug.h > > +++ b/drivers/gpu/drm/i915/intel_hotplug.h > > @@ -14,17 +14,17 @@ struct drm_i915_private; > > struct intel_connector; > > struct intel_encoder; > > -void intel_hpd_poll_init(struct drm_i915_private *dev_priv); > > +void intel_hpd_poll_init(struct drm_i915_private *i915); > > bool intel_encoder_hotplug(struct intel_encoder *encoder, > > struct intel_connector *connector); > > -void intel_hpd_irq_handler(struct drm_i915_private *dev_priv, > > +void intel_hpd_irq_handler(struct drm_i915_private *i915, > > u32 pin_mask, u32 long_mask); > > -void intel_hpd_init(struct drm_i915_private *dev_priv); > > -void intel_hpd_init_work(struct drm_i915_private *dev_priv); > > -void intel_hpd_cancel_work(struct drm_i915_private *dev_priv); > > -enum hpd_pin intel_hpd_pin_default(struct drm_i915_private *dev_priv, > > +void intel_hpd_init(struct drm_i915_private *i915); > > +void intel_hpd_init_work(struct drm_i915_private *i915); > > +void intel_hpd_cancel_work(struct drm_i915_private *i915); > > +enum hpd_pin intel_hpd_pin_default(struct drm_i915_private *i915, > > enum port port); > > -bool intel_hpd_disable(struct drm_i915_private *dev_priv, enum hpd_pin pin); > > -void intel_hpd_enable(struct drm_i915_private *dev_priv, enum hpd_pin pin); > > +bool intel_hpd_disable(struct drm_i915_private *i915, enum hpd_pin pin); > > +void intel_hpd_enable(struct drm_i915_private *i915, enum hpd_pin pin); > > #endif /* __INTEL_HOTPLUG_H__ */ > > diff --git a/drivers/gpu/drm/i915/intel_huc.c b/drivers/gpu/drm/i915/intel_huc.c > > index 8572a0588efc..2b23662698a9 100644 > > --- a/drivers/gpu/drm/i915/intel_huc.c > > +++ b/drivers/gpu/drm/i915/intel_huc.c > > @@ -167,14 +167,14 @@ int intel_huc_auth(struct intel_huc *huc) > > */ > > int intel_huc_check_status(struct intel_huc *huc) > > { > > - struct drm_i915_private *dev_priv = huc_to_i915(huc); > > + struct drm_i915_private *i915 = huc_to_i915(huc); > > intel_wakeref_t wakeref; > > bool status = false; > > - if (!HAS_HUC(dev_priv)) > > + if (!HAS_HUC(i915)) > > return -ENODEV; > > - with_intel_runtime_pm(dev_priv, wakeref) > > + with_intel_runtime_pm(i915, wakeref) > > status = (I915_READ(huc->status.reg) & huc->status.mask) == > > huc->status.value; > > diff --git a/drivers/gpu/drm/i915/intel_huc_fw.c b/drivers/gpu/drm/i915/intel_huc_fw.c > > index 05cbf8338f53..03aed0664726 100644 > > --- a/drivers/gpu/drm/i915/intel_huc_fw.c > > +++ b/drivers/gpu/drm/i915/intel_huc_fw.c > > @@ -69,34 +69,34 @@ MODULE_FIRMWARE(I915_ICL_HUC_UCODE); > > static void huc_fw_select(struct intel_uc_fw *huc_fw) > > { > > struct intel_huc *huc = container_of(huc_fw, struct intel_huc, fw); > > - struct drm_i915_private *dev_priv = huc_to_i915(huc); > > + struct drm_i915_private *i915 = huc_to_i915(huc); > > GEM_BUG_ON(huc_fw->type != INTEL_UC_FW_TYPE_HUC); > > - if (!HAS_HUC(dev_priv)) > > + if (!HAS_HUC(i915)) > > return; > > if (i915_modparams.huc_firmware_path) { > > huc_fw->path = i915_modparams.huc_firmware_path; > > huc_fw->major_ver_wanted = 0; > > huc_fw->minor_ver_wanted = 0; > > - } else if (IS_SKYLAKE(dev_priv)) { > > + } else if (IS_SKYLAKE(i915)) { > > huc_fw->path = I915_SKL_HUC_UCODE; > > huc_fw->major_ver_wanted = SKL_HUC_FW_MAJOR; > > huc_fw->minor_ver_wanted = SKL_HUC_FW_MINOR; > > - } else if (IS_BROXTON(dev_priv)) { > > + } else if (IS_BROXTON(i915)) { > > huc_fw->path = I915_BXT_HUC_UCODE; > > huc_fw->major_ver_wanted = BXT_HUC_FW_MAJOR; > > huc_fw->minor_ver_wanted = BXT_HUC_FW_MINOR; > > - } else if (IS_KABYLAKE(dev_priv) || IS_COFFEELAKE(dev_priv)) { > > + } else if (IS_KABYLAKE(i915) || IS_COFFEELAKE(i915)) { > > huc_fw->path = I915_KBL_HUC_UCODE; > > huc_fw->major_ver_wanted = KBL_HUC_FW_MAJOR; > > huc_fw->minor_ver_wanted = KBL_HUC_FW_MINOR; > > - } else if (IS_GEMINILAKE(dev_priv)) { > > + } else if (IS_GEMINILAKE(i915)) { > > huc_fw->path = I915_GLK_HUC_UCODE; > > huc_fw->major_ver_wanted = GLK_HUC_FW_MAJOR; > > huc_fw->minor_ver_wanted = GLK_HUC_FW_MINOR; > > - } else if (IS_ICELAKE(dev_priv)) { > > + } else if (IS_ICELAKE(i915)) { > > huc_fw->path = I915_ICL_HUC_UCODE; > > huc_fw->major_ver_wanted = ICL_HUC_FW_MAJOR; > > huc_fw->minor_ver_wanted = ICL_HUC_FW_MINOR; > > @@ -135,8 +135,8 @@ static void huc_xfer_rsa(struct intel_huc *huc) > > static int huc_xfer_ucode(struct intel_huc *huc) > > { > > struct intel_uc_fw *huc_fw = &huc->fw; > > - struct drm_i915_private *dev_priv = huc_to_i915(huc); > > - struct intel_uncore *uncore = &dev_priv->uncore; > > + struct drm_i915_private *i915 = huc_to_i915(huc); > > + struct intel_uncore *uncore = &i915->uncore; > > unsigned long offset = 0; > > u32 size; > > int ret; > > diff --git a/drivers/gpu/drm/i915/intel_lpe_audio.c b/drivers/gpu/drm/i915/intel_lpe_audio.c > > index b19800b58442..d95ae05ccebc 100644 > > --- a/drivers/gpu/drm/i915/intel_lpe_audio.c > > +++ b/drivers/gpu/drm/i915/intel_lpe_audio.c > > @@ -73,12 +73,12 @@ > > #include "i915_drv.h" > > #include "intel_lpe_audio.h" > > -#define HAS_LPE_AUDIO(dev_priv) ((dev_priv)->lpe_audio.platdev != NULL) > > +#define HAS_LPE_AUDIO(i915) ((i915)->lpe_audio.platdev != NULL) > > static struct platform_device * > > -lpe_audio_platdev_create(struct drm_i915_private *dev_priv) > > +lpe_audio_platdev_create(struct drm_i915_private *i915) > > { > > - struct drm_device *dev = &dev_priv->drm; > > + struct drm_device *dev = &i915->drm; > > struct platform_device_info pinfo = {}; > > struct resource *rsc; > > struct platform_device *platdev; > > @@ -94,7 +94,7 @@ lpe_audio_platdev_create(struct drm_i915_private *dev_priv) > > return ERR_PTR(-ENOMEM); > > } > > - rsc[0].start = rsc[0].end = dev_priv->lpe_audio.irq; > > + rsc[0].start = rsc[0].end = i915->lpe_audio.irq; > > rsc[0].flags = IORESOURCE_IRQ; > > rsc[0].name = "hdmi-lpe-audio-irq"; > > @@ -114,8 +114,8 @@ lpe_audio_platdev_create(struct drm_i915_private *dev_priv) > > pinfo.size_data = sizeof(*pdata); > > pinfo.dma_mask = DMA_BIT_MASK(32); > > - pdata->num_pipes = INTEL_INFO(dev_priv)->num_pipes; > > - pdata->num_ports = IS_CHERRYVIEW(dev_priv) ? 3 : 2; /* B,C,D or B,C */ > > + pdata->num_pipes = INTEL_INFO(i915)->num_pipes; > > + pdata->num_ports = IS_CHERRYVIEW(i915) ? 3 : 2; /* B,C,D or B,C */ > > pdata->port[0].pipe = -1; > > pdata->port[1].pipe = -1; > > pdata->port[2].pipe = -1; > > @@ -135,7 +135,7 @@ lpe_audio_platdev_create(struct drm_i915_private *dev_priv) > > return platdev; > > } > > -static void lpe_audio_platdev_destroy(struct drm_i915_private *dev_priv) > > +static void lpe_audio_platdev_destroy(struct drm_i915_private *i915) > > { > > /* XXX Note that platform_device_register_full() allocates a dma_mask > > * and never frees it. We can't free it here as we cannot guarantee > > @@ -145,7 +145,7 @@ static void lpe_audio_platdev_destroy(struct drm_i915_private *dev_priv) > > * than us fiddle with its internals. > > */ > > - platform_device_unregister(dev_priv->lpe_audio.platdev); > > + platform_device_unregister(i915->lpe_audio.platdev); > > } > > static void lpe_audio_irq_unmask(struct irq_data *d) > > @@ -162,24 +162,24 @@ static struct irq_chip lpe_audio_irqchip = { > > .irq_unmask = lpe_audio_irq_unmask, > > }; > > -static int lpe_audio_irq_init(struct drm_i915_private *dev_priv) > > +static int lpe_audio_irq_init(struct drm_i915_private *i915) > > { > > - int irq = dev_priv->lpe_audio.irq; > > + int irq = i915->lpe_audio.irq; > > - WARN_ON(!intel_irqs_enabled(dev_priv)); > > + WARN_ON(!intel_irqs_enabled(i915)); > > irq_set_chip_and_handler_name(irq, > > &lpe_audio_irqchip, > > handle_simple_irq, > > "hdmi_lpe_audio_irq_handler"); > > - return irq_set_chip_data(irq, dev_priv); > > + return irq_set_chip_data(irq, i915); > > } > > -static bool lpe_audio_detect(struct drm_i915_private *dev_priv) > > +static bool lpe_audio_detect(struct drm_i915_private *i915) > > { > > int lpe_present = false; > > - if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) { > > + if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) { > > static const struct pci_device_id atom_hdaudio_ids[] = { > > /* Baytrail */ > > {PCI_DEVICE(PCI_VENDOR_ID_INTEL, 0x0f04)}, > > @@ -196,21 +196,21 @@ static bool lpe_audio_detect(struct drm_i915_private *dev_priv) > > return lpe_present; > > } > > -static int lpe_audio_setup(struct drm_i915_private *dev_priv) > > +static int lpe_audio_setup(struct drm_i915_private *i915) > > { > > int ret; > > - dev_priv->lpe_audio.irq = irq_alloc_desc(0); > > - if (dev_priv->lpe_audio.irq < 0) { > > + i915->lpe_audio.irq = irq_alloc_desc(0); > > + if (i915->lpe_audio.irq < 0) { > > DRM_ERROR("Failed to allocate IRQ desc: %d\n", > > - dev_priv->lpe_audio.irq); > > - ret = dev_priv->lpe_audio.irq; > > + i915->lpe_audio.irq); > > + ret = i915->lpe_audio.irq; > > goto err; > > } > > - DRM_DEBUG("irq = %d\n", dev_priv->lpe_audio.irq); > > + DRM_DEBUG("irq = %d\n", i915->lpe_audio.irq); > > - ret = lpe_audio_irq_init(dev_priv); > > + ret = lpe_audio_irq_init(i915); > > if (ret) { > > DRM_ERROR("Failed to initialize irqchip for lpe audio: %d\n", > > @@ -218,10 +218,10 @@ static int lpe_audio_setup(struct drm_i915_private *dev_priv) > > goto err_free_irq; > > } > > - dev_priv->lpe_audio.platdev = lpe_audio_platdev_create(dev_priv); > > + i915->lpe_audio.platdev = lpe_audio_platdev_create(i915); > > - if (IS_ERR(dev_priv->lpe_audio.platdev)) { > > - ret = PTR_ERR(dev_priv->lpe_audio.platdev); > > + if (IS_ERR(i915->lpe_audio.platdev)) { > > + ret = PTR_ERR(i915->lpe_audio.platdev); > > DRM_ERROR("Failed to create lpe audio platform device: %d\n", > > ret); > > goto err_free_irq; > > @@ -234,28 +234,28 @@ static int lpe_audio_setup(struct drm_i915_private *dev_priv) > > return 0; > > err_free_irq: > > - irq_free_desc(dev_priv->lpe_audio.irq); > > + irq_free_desc(i915->lpe_audio.irq); > > err: > > - dev_priv->lpe_audio.irq = -1; > > - dev_priv->lpe_audio.platdev = NULL; > > + i915->lpe_audio.irq = -1; > > + i915->lpe_audio.platdev = NULL; > > return ret; > > } > > /** > > * intel_lpe_audio_irq_handler() - forwards the LPE audio irq > > - * @dev_priv: the i915 drm device private data > > + * @i915: the i915 drm device private data > > * > > * the LPE Audio irq is forwarded to the irq handler registered by LPE audio > > * driver. > > */ > > -void intel_lpe_audio_irq_handler(struct drm_i915_private *dev_priv) > > +void intel_lpe_audio_irq_handler(struct drm_i915_private *i915) > > { > > int ret; > > - if (!HAS_LPE_AUDIO(dev_priv)) > > + if (!HAS_LPE_AUDIO(i915)) > > return; > > - ret = generic_handle_irq(dev_priv->lpe_audio.irq); > > + ret = generic_handle_irq(i915->lpe_audio.irq); > > if (ret) > > DRM_ERROR_RATELIMITED("error handling LPE audio irq: %d\n", > > ret); > > @@ -264,17 +264,17 @@ void intel_lpe_audio_irq_handler(struct drm_i915_private *dev_priv) > > /** > > * intel_lpe_audio_init() - detect and setup the bridge between HDMI LPE Audio > > * driver and i915 > > - * @dev_priv: the i915 drm device private data > > + * @i915: the i915 drm device private data > > * > > * Return: 0 if successful. non-zero if detection or > > * llocation/initialization fails > > */ > > -int intel_lpe_audio_init(struct drm_i915_private *dev_priv) > > +int intel_lpe_audio_init(struct drm_i915_private *i915) > > { > > int ret = -ENODEV; > > - if (lpe_audio_detect(dev_priv)) { > > - ret = lpe_audio_setup(dev_priv); > > + if (lpe_audio_detect(i915)) { > > + ret = lpe_audio_setup(i915); > > if (ret < 0) > > DRM_ERROR("failed to setup LPE Audio bridge\n"); > > } > > @@ -284,31 +284,31 @@ int intel_lpe_audio_init(struct drm_i915_private *dev_priv) > > /** > > * intel_lpe_audio_teardown() - destroy the bridge between HDMI LPE > > * audio driver and i915 > > - * @dev_priv: the i915 drm device private data > > + * @i915: the i915 drm device private data > > * > > * release all the resources for LPE audio <-> i915 bridge. > > */ > > -void intel_lpe_audio_teardown(struct drm_i915_private *dev_priv) > > +void intel_lpe_audio_teardown(struct drm_i915_private *i915) > > { > > struct irq_desc *desc; > > - if (!HAS_LPE_AUDIO(dev_priv)) > > + if (!HAS_LPE_AUDIO(i915)) > > return; > > - desc = irq_to_desc(dev_priv->lpe_audio.irq); > > + desc = irq_to_desc(i915->lpe_audio.irq); > > - lpe_audio_platdev_destroy(dev_priv); > > + lpe_audio_platdev_destroy(i915); > > - irq_free_desc(dev_priv->lpe_audio.irq); > > + irq_free_desc(i915->lpe_audio.irq); > > - dev_priv->lpe_audio.irq = -1; > > - dev_priv->lpe_audio.platdev = NULL; > > + i915->lpe_audio.irq = -1; > > + i915->lpe_audio.platdev = NULL; > > } > > /** > > * intel_lpe_audio_notify() - notify lpe audio event > > * audio driver and i915 > > - * @dev_priv: the i915 drm device private data > > + * @i915: the i915 drm device private data > > * @pipe: pipe > > * @port: port > > * @eld : ELD data > > @@ -317,7 +317,7 @@ void intel_lpe_audio_teardown(struct drm_i915_private *dev_priv) > > * > > * Notify lpe audio driver of eld change. > > */ > > -void intel_lpe_audio_notify(struct drm_i915_private *dev_priv, > > +void intel_lpe_audio_notify(struct drm_i915_private *i915, > > enum pipe pipe, enum port port, > > const void *eld, int ls_clock, bool dp_output) > > { > > @@ -326,10 +326,10 @@ void intel_lpe_audio_notify(struct drm_i915_private *dev_priv, > > struct intel_hdmi_lpe_audio_port_pdata *ppdata; > > u32 audio_enable; > > - if (!HAS_LPE_AUDIO(dev_priv)) > > + if (!HAS_LPE_AUDIO(i915)) > > return; > > - pdata = dev_get_platdata(&dev_priv->lpe_audio.platdev->dev); > > + pdata = dev_get_platdata(&i915->lpe_audio.platdev->dev); > > ppdata = &pdata->port[port - PORT_B]; > > spin_lock_irqsave(&pdata->lpe_audio_slock, irqflags); > > @@ -357,7 +357,7 @@ void intel_lpe_audio_notify(struct drm_i915_private *dev_priv, > > } > > if (pdata->notify_audio_lpe) > > - pdata->notify_audio_lpe(dev_priv->lpe_audio.platdev, port - PORT_B); > > + pdata->notify_audio_lpe(i915->lpe_audio.platdev, port - PORT_B); > > spin_unlock_irqrestore(&pdata->lpe_audio_slock, irqflags); > > } > > diff --git a/drivers/gpu/drm/i915/intel_lpe_audio.h b/drivers/gpu/drm/i915/intel_lpe_audio.h > > index f848c5038714..df43abce3c95 100644 > > --- a/drivers/gpu/drm/i915/intel_lpe_audio.h > > +++ b/drivers/gpu/drm/i915/intel_lpe_audio.h > > @@ -12,10 +12,10 @@ enum pipe; > > enum port; > > struct drm_i915_private; > > -int intel_lpe_audio_init(struct drm_i915_private *dev_priv); > > -void intel_lpe_audio_teardown(struct drm_i915_private *dev_priv); > > -void intel_lpe_audio_irq_handler(struct drm_i915_private *dev_priv); > > -void intel_lpe_audio_notify(struct drm_i915_private *dev_priv, > > +int intel_lpe_audio_init(struct drm_i915_private *i915); > > +void intel_lpe_audio_teardown(struct drm_i915_private *i915); > > +void intel_lpe_audio_irq_handler(struct drm_i915_private *i915); > > +void intel_lpe_audio_notify(struct drm_i915_private *i915, > > enum pipe pipe, enum port port, > > const void *eld, int ls_clock, bool dp_output); > > diff --git a/drivers/gpu/drm/i915/intel_lspcon.c b/drivers/gpu/drm/i915/intel_lspcon.c > > index 7028d0cf3bb1..695465f97195 100644 > > --- a/drivers/gpu/drm/i915/intel_lspcon.c > > +++ b/drivers/gpu/drm/i915/intel_lspcon.c > > @@ -555,10 +555,10 @@ bool lspcon_init(struct intel_digital_port *intel_dig_port) > > struct intel_dp *dp = &intel_dig_port->dp; > > struct intel_lspcon *lspcon = &intel_dig_port->lspcon; > > struct drm_device *dev = intel_dig_port->base.base.dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > struct drm_connector *connector = &dp->attached_connector->base; > > - if (!HAS_LSPCON(dev_priv)) { > > + if (!HAS_LSPCON(i915)) { > > DRM_ERROR("LSPCON is not supported on this platform\n"); > > return false; > > } > > diff --git a/drivers/gpu/drm/i915/intel_lvds.c b/drivers/gpu/drm/i915/intel_lvds.c > > index efefed62a7f8..0405f9834827 100644 > > --- a/drivers/gpu/drm/i915/intel_lvds.c > > +++ b/drivers/gpu/drm/i915/intel_lvds.c > > @@ -80,7 +80,7 @@ static struct intel_lvds_encoder *to_lvds_encoder(struct drm_encoder *encoder) > > return container_of(encoder, struct intel_lvds_encoder, base.base); > > } > > -bool intel_lvds_port_enabled(struct drm_i915_private *dev_priv, > > +bool intel_lvds_port_enabled(struct drm_i915_private *i915, > > i915_reg_t lvds_reg, enum pipe *pipe) > > { > > u32 val; > > @@ -88,7 +88,7 @@ bool intel_lvds_port_enabled(struct drm_i915_private *dev_priv, > > val = I915_READ(lvds_reg); > > /* asserts want to know the pipe even if the port is disabled */ > > - if (HAS_PCH_CPT(dev_priv)) > > + if (HAS_PCH_CPT(i915)) > > *pipe = (val & LVDS_PIPE_SEL_MASK_CPT) >> LVDS_PIPE_SEL_SHIFT_CPT; > > else > > *pipe = (val & LVDS_PIPE_SEL_MASK) >> LVDS_PIPE_SEL_SHIFT; > > @@ -99,19 +99,19 @@ bool intel_lvds_port_enabled(struct drm_i915_private *dev_priv, > > static bool intel_lvds_get_hw_state(struct intel_encoder *encoder, > > enum pipe *pipe) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_lvds_encoder *lvds_encoder = to_lvds_encoder(&encoder->base); > > intel_wakeref_t wakeref; > > bool ret; > > - wakeref = intel_display_power_get_if_enabled(dev_priv, > > + wakeref = intel_display_power_get_if_enabled(i915, > > encoder->power_domain); > > if (!wakeref) > > return false; > > - ret = intel_lvds_port_enabled(dev_priv, lvds_encoder->reg, pipe); > > + ret = intel_lvds_port_enabled(i915, lvds_encoder->reg, pipe); > > - intel_display_power_put(dev_priv, encoder->power_domain, wakeref); > > + intel_display_power_put(i915, encoder->power_domain, wakeref); > > return ret; > > } > > @@ -119,7 +119,7 @@ static bool intel_lvds_get_hw_state(struct intel_encoder *encoder, > > static void intel_lvds_get_config(struct intel_encoder *encoder, > > struct intel_crtc_state *pipe_config) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_lvds_encoder *lvds_encoder = to_lvds_encoder(&encoder->base); > > u32 tmp, flags = 0; > > @@ -137,12 +137,12 @@ static void intel_lvds_get_config(struct intel_encoder *encoder, > > pipe_config->base.adjusted_mode.flags |= flags; > > - if (INTEL_GEN(dev_priv) < 5) > > + if (INTEL_GEN(i915) < 5) > > pipe_config->gmch_pfit.lvds_border_bits = > > tmp & LVDS_BORDER_ENABLE; > > /* gen2/3 store dither state in pfit control, needs to match */ > > - if (INTEL_GEN(dev_priv) < 4) { > > + if (INTEL_GEN(i915) < 4) { > > tmp = I915_READ(PFIT_CONTROL); > > pipe_config->gmch_pfit.control |= tmp & PANEL_8TO6_DITHER_ENABLE; > > @@ -151,7 +151,7 @@ static void intel_lvds_get_config(struct intel_encoder *encoder, > > pipe_config->base.adjusted_mode.crtc_clock = pipe_config->port_clock; > > } > > -static void intel_lvds_pps_get_hw_state(struct drm_i915_private *dev_priv, > > +static void intel_lvds_pps_get_hw_state(struct drm_i915_private *i915, > > struct intel_lvds_pps *pps) > > { > > u32 val; > > @@ -180,7 +180,7 @@ static void intel_lvds_pps_get_hw_state(struct drm_i915_private *dev_priv, > > /* Convert from 100ms to 100us units */ > > pps->t4 = val * 1000; > > - if (INTEL_GEN(dev_priv) <= 4 && > > + if (INTEL_GEN(i915) <= 4 && > > pps->t1_t2 == 0 && pps->t5 == 0 && pps->t3 == 0 && pps->tx == 0) { > > DRM_DEBUG_KMS("Panel power timings uninitialized, " > > "setting defaults\n"); > > @@ -198,7 +198,7 @@ static void intel_lvds_pps_get_hw_state(struct drm_i915_private *dev_priv, > > pps->divider, pps->port, pps->powerdown_on_reset); > > } > > -static void intel_lvds_pps_init_hw(struct drm_i915_private *dev_priv, > > +static void intel_lvds_pps_init_hw(struct drm_i915_private *i915, > > struct intel_lvds_pps *pps) > > { > > u32 val; > > @@ -229,26 +229,26 @@ static void intel_pre_enable_lvds(struct intel_encoder *encoder, > > const struct drm_connector_state *conn_state) > > { > > struct intel_lvds_encoder *lvds_encoder = to_lvds_encoder(&encoder->base); > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_crtc *crtc = to_intel_crtc(pipe_config->base.crtc); > > const struct drm_display_mode *adjusted_mode = &pipe_config->base.adjusted_mode; > > int pipe = crtc->pipe; > > u32 temp; > > - if (HAS_PCH_SPLIT(dev_priv)) { > > - assert_fdi_rx_pll_disabled(dev_priv, pipe); > > - assert_shared_dpll_disabled(dev_priv, > > + if (HAS_PCH_SPLIT(i915)) { > > + assert_fdi_rx_pll_disabled(i915, pipe); > > + assert_shared_dpll_disabled(i915, > > pipe_config->shared_dpll); > > } else { > > - assert_pll_disabled(dev_priv, pipe); > > + assert_pll_disabled(i915, pipe); > > } > > - intel_lvds_pps_init_hw(dev_priv, &lvds_encoder->init_pps); > > + intel_lvds_pps_init_hw(i915, &lvds_encoder->init_pps); > > temp = lvds_encoder->init_lvds_val; > > temp |= LVDS_PORT_EN | LVDS_A0A2_CLKA_POWER_UP; > > - if (HAS_PCH_CPT(dev_priv)) { > > + if (HAS_PCH_CPT(i915)) { > > temp &= ~LVDS_PIPE_SEL_MASK_CPT; > > temp |= LVDS_PIPE_SEL_CPT(pipe); > > } else { > > @@ -283,7 +283,7 @@ static void intel_pre_enable_lvds(struct intel_encoder *encoder, > > * special lvds dither control bit on pch-split platforms, dithering is > > * only controlled through the PIPECONF reg. > > */ > > - if (IS_GEN(dev_priv, 4)) { > > + if (IS_GEN(i915, 4)) { > > /* > > * Bspec wording suggests that LVDS port dithering only exists > > * for 18bpp panels. > > @@ -311,14 +311,14 @@ static void intel_enable_lvds(struct intel_encoder *encoder, > > { > > struct drm_device *dev = encoder->base.dev; > > struct intel_lvds_encoder *lvds_encoder = to_lvds_encoder(&encoder->base); > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > I915_WRITE(lvds_encoder->reg, I915_READ(lvds_encoder->reg) | LVDS_PORT_EN); > > I915_WRITE(PP_CONTROL(0), I915_READ(PP_CONTROL(0)) | PANEL_POWER_ON); > > POSTING_READ(lvds_encoder->reg); > > - if (intel_wait_for_register(&dev_priv->uncore, > > + if (intel_wait_for_register(&i915->uncore, > > PP_STATUS(0), PP_ON, PP_ON, 5000)) > > DRM_ERROR("timed out waiting for panel to power on\n"); > > @@ -330,10 +330,10 @@ static void intel_disable_lvds(struct intel_encoder *encoder, > > const struct drm_connector_state *old_conn_state) > > { > > struct intel_lvds_encoder *lvds_encoder = to_lvds_encoder(&encoder->base); > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > I915_WRITE(PP_CONTROL(0), I915_READ(PP_CONTROL(0)) & ~PANEL_POWER_ON); > > - if (intel_wait_for_register(&dev_priv->uncore, > > + if (intel_wait_for_register(&i915->uncore, > > PP_STATUS(0), PP_ON, 0, 1000)) > > DRM_ERROR("timed out waiting for panel to power off\n"); > > @@ -389,7 +389,7 @@ static int intel_lvds_compute_config(struct intel_encoder *intel_encoder, > > struct intel_crtc_state *pipe_config, > > struct drm_connector_state *conn_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(intel_encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(intel_encoder->base.dev); > > struct intel_lvds_encoder *lvds_encoder = > > to_lvds_encoder(&intel_encoder->base); > > struct intel_connector *intel_connector = > > @@ -399,7 +399,7 @@ static int intel_lvds_compute_config(struct intel_encoder *intel_encoder, > > unsigned int lvds_bpp; > > /* Should never happen!! */ > > - if (INTEL_GEN(dev_priv) < 4 && intel_crtc->pipe == 0) { > > + if (INTEL_GEN(i915) < 4 && intel_crtc->pipe == 0) { > > DRM_ERROR("Can't support LVDS on pipe A\n"); > > return -EINVAL; > > } > > @@ -429,7 +429,7 @@ static int intel_lvds_compute_config(struct intel_encoder *intel_encoder, > > if (adjusted_mode->flags & DRM_MODE_FLAG_DBLSCAN) > > return -EINVAL; > > - if (HAS_PCH_SPLIT(dev_priv)) { > > + if (HAS_PCH_SPLIT(i915)) { > > pipe_config->has_pch_encoder = true; > > intel_pch_panel_fitting(intel_crtc, pipe_config, > > @@ -751,11 +751,11 @@ static const struct dmi_system_id intel_dual_link_lvds[] = { > > { } /* terminating entry */ > > }; > > -struct intel_encoder *intel_get_lvds_encoder(struct drm_i915_private *dev_priv) > > +struct intel_encoder *intel_get_lvds_encoder(struct drm_i915_private *i915) > > { > > struct intel_encoder *encoder; > > - for_each_intel_encoder(&dev_priv->drm, encoder) { > > + for_each_intel_encoder(&i915->drm, encoder) { > > if (encoder->type == INTEL_OUTPUT_LVDS) > > return encoder; > > } > > @@ -763,9 +763,9 @@ struct intel_encoder *intel_get_lvds_encoder(struct drm_i915_private *dev_priv) > > return NULL; > > } > > -bool intel_is_dual_link_lvds(struct drm_i915_private *dev_priv) > > +bool intel_is_dual_link_lvds(struct drm_i915_private *i915) > > { > > - struct intel_encoder *encoder = intel_get_lvds_encoder(dev_priv); > > + struct intel_encoder *encoder = intel_get_lvds_encoder(i915); > > return encoder && to_lvds_encoder(&encoder->base)->is_dual_link; > > } > > @@ -774,7 +774,7 @@ static bool compute_is_dual_link_lvds(struct intel_lvds_encoder *lvds_encoder) > > { > > struct drm_device *dev = lvds_encoder->base.base.dev; > > unsigned int val; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > /* use the module option value if specified */ > > if (i915_modparams.lvds_channel_mode > 0) > > @@ -794,26 +794,26 @@ static bool compute_is_dual_link_lvds(struct intel_lvds_encoder *lvds_encoder) > > * register is uninitialized. > > */ > > val = I915_READ(lvds_encoder->reg); > > - if (HAS_PCH_CPT(dev_priv)) > > + if (HAS_PCH_CPT(i915)) > > val &= ~(LVDS_DETECTED | LVDS_PIPE_SEL_MASK_CPT); > > else > > val &= ~(LVDS_DETECTED | LVDS_PIPE_SEL_MASK); > > if (val == 0) > > - val = dev_priv->vbt.bios_lvds_val; > > + val = i915->vbt.bios_lvds_val; > > return (val & LVDS_CLKB_POWER_MASK) == LVDS_CLKB_POWER_UP; > > } > > /** > > * intel_lvds_init - setup LVDS connectors on this device > > - * @dev_priv: i915 device > > + * @i915: i915 device > > * > > * Create the connector, register the LVDS DDC bus, and try to figure out what > > * modes we can display on the LVDS panel (if present). > > */ > > -void intel_lvds_init(struct drm_i915_private *dev_priv) > > +void intel_lvds_init(struct drm_i915_private *i915) > > { > > - struct drm_device *dev = &dev_priv->drm; > > + struct drm_device *dev = &i915->drm; > > struct intel_lvds_encoder *lvds_encoder; > > struct intel_encoder *intel_encoder; > > struct intel_connector *intel_connector; > > @@ -829,30 +829,30 @@ void intel_lvds_init(struct drm_i915_private *dev_priv) > > /* Skip init on machines we know falsely report LVDS */ > > if (dmi_check_system(intel_no_lvds)) { > > - WARN(!dev_priv->vbt.int_lvds_support, > > + WARN(!i915->vbt.int_lvds_support, > > "Useless DMI match. Internal LVDS support disabled by VBT\n"); > > return; > > } > > - if (!dev_priv->vbt.int_lvds_support) { > > + if (!i915->vbt.int_lvds_support) { > > DRM_DEBUG_KMS("Internal LVDS support disabled by VBT\n"); > > return; > > } > > - if (HAS_PCH_SPLIT(dev_priv)) > > + if (HAS_PCH_SPLIT(i915)) > > lvds_reg = PCH_LVDS; > > else > > lvds_reg = LVDS; > > lvds = I915_READ(lvds_reg); > > - if (HAS_PCH_SPLIT(dev_priv)) { > > + if (HAS_PCH_SPLIT(i915)) { > > if ((lvds & LVDS_DETECTED) == 0) > > return; > > } > > pin = GMBUS_PIN_PANEL; > > - if (!intel_bios_is_lvds_present(dev_priv, &pin)) { > > + if (!intel_bios_is_lvds_present(i915, &pin)) { > > if ((lvds & LVDS_PORT_EN) == 0) { > > DRM_DEBUG_KMS("LVDS is not present in VBT\n"); > > return; > > @@ -884,7 +884,7 @@ void intel_lvds_init(struct drm_i915_private *dev_priv) > > intel_encoder->enable = intel_enable_lvds; > > intel_encoder->pre_enable = intel_pre_enable_lvds; > > intel_encoder->compute_config = intel_lvds_compute_config; > > - if (HAS_PCH_SPLIT(dev_priv)) { > > + if (HAS_PCH_SPLIT(i915)) { > > intel_encoder->disable = pch_disable_lvds; > > intel_encoder->post_disable = pch_post_disable_lvds; > > } else { > > @@ -901,9 +901,9 @@ void intel_lvds_init(struct drm_i915_private *dev_priv) > > intel_encoder->power_domain = POWER_DOMAIN_PORT_OTHER; > > intel_encoder->port = PORT_NONE; > > intel_encoder->cloneable = 0; > > - if (HAS_PCH_SPLIT(dev_priv)) > > + if (HAS_PCH_SPLIT(i915)) > > intel_encoder->crtc_mask = (1 << 0) | (1 << 1) | (1 << 2); > > - else if (IS_GEN(dev_priv, 4)) > > + else if (IS_GEN(i915, 4)) > > intel_encoder->crtc_mask = (1 << 0) | (1 << 1); > > else > > intel_encoder->crtc_mask = (1 << 1); > > @@ -922,7 +922,7 @@ void intel_lvds_init(struct drm_i915_private *dev_priv) > > drm_connector_attach_scaling_mode_property(connector, allowed_scalers); > > connector->state->scaling_mode = DRM_MODE_SCALE_ASPECT; > > - intel_lvds_pps_get_hw_state(dev_priv, &lvds_encoder->init_pps); > > + intel_lvds_pps_get_hw_state(i915, &lvds_encoder->init_pps); > > lvds_encoder->init_lvds_val = lvds; > > /* > > @@ -940,10 +940,10 @@ void intel_lvds_init(struct drm_i915_private *dev_priv) > > mutex_lock(&dev->mode_config.mutex); > > if (vga_switcheroo_handler_flags() & VGA_SWITCHEROO_CAN_SWITCH_DDC) > > edid = drm_get_edid_switcheroo(connector, > > - intel_gmbus_get_adapter(dev_priv, pin)); > > + intel_gmbus_get_adapter(i915, pin)); > > else > > edid = drm_get_edid(connector, > > - intel_gmbus_get_adapter(dev_priv, pin)); > > + intel_gmbus_get_adapter(i915, pin)); > > if (edid) { > > if (drm_add_edid_modes(connector, edid)) { > > drm_connector_update_edid_property(connector, > > diff --git a/drivers/gpu/drm/i915/intel_lvds.h b/drivers/gpu/drm/i915/intel_lvds.h > > index bc9c8b84ba2f..4afdcb4d6352 100644 > > --- a/drivers/gpu/drm/i915/intel_lvds.h > > +++ b/drivers/gpu/drm/i915/intel_lvds.h > > @@ -13,10 +13,10 @@ > > enum pipe; > > struct drm_i915_private; > > -bool intel_lvds_port_enabled(struct drm_i915_private *dev_priv, > > +bool intel_lvds_port_enabled(struct drm_i915_private *i915, > > i915_reg_t lvds_reg, enum pipe *pipe); > > -void intel_lvds_init(struct drm_i915_private *dev_priv); > > -struct intel_encoder *intel_get_lvds_encoder(struct drm_i915_private *dev_priv); > > -bool intel_is_dual_link_lvds(struct drm_i915_private *dev_priv); > > +void intel_lvds_init(struct drm_i915_private *i915); > > +struct intel_encoder *intel_get_lvds_encoder(struct drm_i915_private *i915); > > +bool intel_is_dual_link_lvds(struct drm_i915_private *i915); > > #endif /* __INTEL_LVDS_H__ */ > > diff --git a/drivers/gpu/drm/i915/intel_opregion.c b/drivers/gpu/drm/i915/intel_opregion.c > > index 8fa1159d097f..6cbe775d7537 100644 > > --- a/drivers/gpu/drm/i915/intel_opregion.c > > +++ b/drivers/gpu/drm/i915/intel_opregion.c > > @@ -266,11 +266,11 @@ struct opregion_asle_ext { > > #define MAX_DSLP 1500 > > -static int swsci(struct drm_i915_private *dev_priv, > > +static int swsci(struct drm_i915_private *i915, > > u32 function, u32 parm, u32 *parm_out) > > { > > - struct opregion_swsci *swsci = dev_priv->opregion.swsci; > > - struct pci_dev *pdev = dev_priv->drm.pdev; > > + struct opregion_swsci *swsci = i915->opregion.swsci; > > + struct pci_dev *pdev = i915->drm.pdev; > > u32 main_function, sub_function, scic; > > u16 swsci_val; > > u32 dslp; > > @@ -285,11 +285,11 @@ static int swsci(struct drm_i915_private *dev_priv, > > /* Check if we can call the function. See swsci_setup for details. */ > > if (main_function == SWSCI_SBCB) { > > - if ((dev_priv->opregion.swsci_sbcb_sub_functions & > > + if ((i915->opregion.swsci_sbcb_sub_functions & > > (1 << sub_function)) == 0) > > return -EINVAL; > > } else if (main_function == SWSCI_GBDA) { > > - if ((dev_priv->opregion.swsci_gbda_sub_functions & > > + if ((i915->opregion.swsci_gbda_sub_functions & > > (1 << sub_function)) == 0) > > return -EINVAL; > > } > > @@ -363,13 +363,13 @@ static int swsci(struct drm_i915_private *dev_priv, > > int intel_opregion_notify_encoder(struct intel_encoder *intel_encoder, > > bool enable) > > { > > - struct drm_i915_private *dev_priv = to_i915(intel_encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(intel_encoder->base.dev); > > u32 parm = 0; > > u32 type = 0; > > u32 port; > > /* don't care about old stuff for now */ > > - if (!HAS_DDI(dev_priv)) > > + if (!HAS_DDI(i915)) > > return 0; > > if (intel_encoder->type == INTEL_OUTPUT_DSI) > > @@ -409,7 +409,7 @@ int intel_opregion_notify_encoder(struct intel_encoder *intel_encoder, > > parm |= type << (16 + port * 3); > > - return swsci(dev_priv, SWSCI_SBCB_DISPLAY_POWER_STATE, parm, NULL); > > + return swsci(i915, SWSCI_SBCB_DISPLAY_POWER_STATE, parm, NULL); > > } > > static const struct { > > @@ -423,29 +423,29 @@ static const struct { > > { PCI_D3cold, 0x04 }, > > }; > > -int intel_opregion_notify_adapter(struct drm_i915_private *dev_priv, > > +int intel_opregion_notify_adapter(struct drm_i915_private *i915, > > pci_power_t state) > > { > > int i; > > - if (!HAS_DDI(dev_priv)) > > + if (!HAS_DDI(i915)) > > return 0; > > for (i = 0; i < ARRAY_SIZE(power_state_map); i++) { > > if (state == power_state_map[i].pci_power_state) > > - return swsci(dev_priv, SWSCI_SBCB_ADAPTER_POWER_STATE, > > + return swsci(i915, SWSCI_SBCB_ADAPTER_POWER_STATE, > > power_state_map[i].parm, NULL); > > } > > return -EINVAL; > > } > > -static u32 asle_set_backlight(struct drm_i915_private *dev_priv, u32 bclp) > > +static u32 asle_set_backlight(struct drm_i915_private *i915, u32 bclp) > > { > > struct intel_connector *connector; > > struct drm_connector_list_iter conn_iter; > > - struct opregion_asle *asle = dev_priv->opregion.asle; > > - struct drm_device *dev = &dev_priv->drm; > > + struct opregion_asle *asle = i915->opregion.asle; > > + struct drm_device *dev = &i915->drm; > > DRM_DEBUG_DRIVER("bclp = 0x%08x\n", bclp); > > @@ -480,7 +480,7 @@ static u32 asle_set_backlight(struct drm_i915_private *dev_priv, u32 bclp) > > return 0; > > } > > -static u32 asle_set_als_illum(struct drm_i915_private *dev_priv, u32 alsi) > > +static u32 asle_set_als_illum(struct drm_i915_private *i915, u32 alsi) > > { > > /* alsi is the current ALS reading in lux. 0 indicates below sensor > > range, 0xffff indicates above sensor range. 1-0xfffe are valid */ > > @@ -488,13 +488,13 @@ static u32 asle_set_als_illum(struct drm_i915_private *dev_priv, u32 alsi) > > return ASLC_ALS_ILLUM_FAILED; > > } > > -static u32 asle_set_pwm_freq(struct drm_i915_private *dev_priv, u32 pfmb) > > +static u32 asle_set_pwm_freq(struct drm_i915_private *i915, u32 pfmb) > > { > > DRM_DEBUG_DRIVER("PWM freq is not supported\n"); > > return ASLC_PWM_FREQ_FAILED; > > } > > -static u32 asle_set_pfit(struct drm_i915_private *dev_priv, u32 pfit) > > +static u32 asle_set_pfit(struct drm_i915_private *i915, u32 pfit) > > { > > /* Panel fitting is currently controlled by the X code, so this is a > > noop until modesetting support works fully */ > > @@ -502,13 +502,13 @@ static u32 asle_set_pfit(struct drm_i915_private *dev_priv, u32 pfit) > > return ASLC_PFIT_FAILED; > > } > > -static u32 asle_set_supported_rotation_angles(struct drm_i915_private *dev_priv, u32 srot) > > +static u32 asle_set_supported_rotation_angles(struct drm_i915_private *i915, u32 srot) > > { > > DRM_DEBUG_DRIVER("SROT is not supported\n"); > > return ASLC_ROTATION_ANGLES_FAILED; > > } > > -static u32 asle_set_button_array(struct drm_i915_private *dev_priv, u32 iuer) > > +static u32 asle_set_button_array(struct drm_i915_private *i915, u32 iuer) > > { > > if (!iuer) > > DRM_DEBUG_DRIVER("Button array event is not supported (nothing)\n"); > > @@ -526,7 +526,7 @@ static u32 asle_set_button_array(struct drm_i915_private *dev_priv, u32 iuer) > > return ASLC_BUTTON_ARRAY_FAILED; > > } > > -static u32 asle_set_convertible(struct drm_i915_private *dev_priv, u32 iuer) > > +static u32 asle_set_convertible(struct drm_i915_private *i915, u32 iuer) > > { > > if (iuer & ASLE_IUER_CONVERTIBLE) > > DRM_DEBUG_DRIVER("Convertible is not supported (clamshell)\n"); > > @@ -536,7 +536,7 @@ static u32 asle_set_convertible(struct drm_i915_private *dev_priv, u32 iuer) > > return ASLC_CONVERTIBLE_FAILED; > > } > > -static u32 asle_set_docking(struct drm_i915_private *dev_priv, u32 iuer) > > +static u32 asle_set_docking(struct drm_i915_private *i915, u32 iuer) > > { > > if (iuer & ASLE_IUER_DOCKING) > > DRM_DEBUG_DRIVER("Docking is not supported (docked)\n"); > > @@ -546,7 +546,7 @@ static u32 asle_set_docking(struct drm_i915_private *dev_priv, u32 iuer) > > return ASLC_DOCKING_FAILED; > > } > > -static u32 asle_isct_state(struct drm_i915_private *dev_priv) > > +static u32 asle_isct_state(struct drm_i915_private *i915) > > { > > DRM_DEBUG_DRIVER("ISCT is not supported\n"); > > return ASLC_ISCT_STATE_FAILED; > > @@ -556,9 +556,9 @@ static void asle_work(struct work_struct *work) > > { > > struct intel_opregion *opregion = > > container_of(work, struct intel_opregion, asle_work); > > - struct drm_i915_private *dev_priv = > > + struct drm_i915_private *i915 = > > container_of(opregion, struct drm_i915_private, opregion); > > - struct opregion_asle *asle = dev_priv->opregion.asle; > > + struct opregion_asle *asle = i915->opregion.asle; > > u32 aslc_stat = 0; > > u32 aslc_req; > > @@ -574,40 +574,40 @@ static void asle_work(struct work_struct *work) > > } > > if (aslc_req & ASLC_SET_ALS_ILLUM) > > - aslc_stat |= asle_set_als_illum(dev_priv, asle->alsi); > > + aslc_stat |= asle_set_als_illum(i915, asle->alsi); > > if (aslc_req & ASLC_SET_BACKLIGHT) > > - aslc_stat |= asle_set_backlight(dev_priv, asle->bclp); > > + aslc_stat |= asle_set_backlight(i915, asle->bclp); > > if (aslc_req & ASLC_SET_PFIT) > > - aslc_stat |= asle_set_pfit(dev_priv, asle->pfit); > > + aslc_stat |= asle_set_pfit(i915, asle->pfit); > > if (aslc_req & ASLC_SET_PWM_FREQ) > > - aslc_stat |= asle_set_pwm_freq(dev_priv, asle->pfmb); > > + aslc_stat |= asle_set_pwm_freq(i915, asle->pfmb); > > if (aslc_req & ASLC_SUPPORTED_ROTATION_ANGLES) > > - aslc_stat |= asle_set_supported_rotation_angles(dev_priv, > > + aslc_stat |= asle_set_supported_rotation_angles(i915, > > asle->srot); > > if (aslc_req & ASLC_BUTTON_ARRAY) > > - aslc_stat |= asle_set_button_array(dev_priv, asle->iuer); > > + aslc_stat |= asle_set_button_array(i915, asle->iuer); > > if (aslc_req & ASLC_CONVERTIBLE_INDICATOR) > > - aslc_stat |= asle_set_convertible(dev_priv, asle->iuer); > > + aslc_stat |= asle_set_convertible(i915, asle->iuer); > > if (aslc_req & ASLC_DOCKING_INDICATOR) > > - aslc_stat |= asle_set_docking(dev_priv, asle->iuer); > > + aslc_stat |= asle_set_docking(i915, asle->iuer); > > if (aslc_req & ASLC_ISCT_STATE_CHANGE) > > - aslc_stat |= asle_isct_state(dev_priv); > > + aslc_stat |= asle_isct_state(i915); > > asle->aslc = aslc_stat; > > } > > -void intel_opregion_asle_intr(struct drm_i915_private *dev_priv) > > +void intel_opregion_asle_intr(struct drm_i915_private *i915) > > { > > - if (dev_priv->opregion.asle) > > - schedule_work(&dev_priv->opregion.asle_work); > > + if (i915->opregion.asle) > > + schedule_work(&i915->opregion.asle_work); > > } > > #define ACPI_EV_DISPLAY_SWITCH (1<<0) > > @@ -702,9 +702,9 @@ static u32 acpi_display_type(struct intel_connector *connector) > > return display_type; > > } > > -static void intel_didl_outputs(struct drm_i915_private *dev_priv) > > +static void intel_didl_outputs(struct drm_i915_private *i915) > > { > > - struct intel_opregion *opregion = &dev_priv->opregion; > > + struct intel_opregion *opregion = &i915->opregion; > > struct intel_connector *connector; > > struct drm_connector_list_iter conn_iter; > > int i = 0, max_outputs; > > @@ -720,7 +720,7 @@ static void intel_didl_outputs(struct drm_i915_private *dev_priv) > > max_outputs = ARRAY_SIZE(opregion->acpi->didl) + > > ARRAY_SIZE(opregion->acpi->did2); > > - drm_connector_list_iter_begin(&dev_priv->drm, &conn_iter); > > + drm_connector_list_iter_begin(&i915->drm, &conn_iter); > > for_each_intel_connector_iter(connector, &conn_iter) { > > u32 device_id, type; > > @@ -749,9 +749,9 @@ static void intel_didl_outputs(struct drm_i915_private *dev_priv) > > set_did(opregion, i, 0); > > } > > -static void intel_setup_cadls(struct drm_i915_private *dev_priv) > > +static void intel_setup_cadls(struct drm_i915_private *i915) > > { > > - struct intel_opregion *opregion = &dev_priv->opregion; > > + struct intel_opregion *opregion = &i915->opregion; > > struct intel_connector *connector; > > struct drm_connector_list_iter conn_iter; > > int i = 0; > > @@ -766,7 +766,7 @@ static void intel_setup_cadls(struct drm_i915_private *dev_priv) > > * Note that internal panels should be at the front of the connector > > * list already, ensuring they're not left out. > > */ > > - drm_connector_list_iter_begin(&dev_priv->drm, &conn_iter); > > + drm_connector_list_iter_begin(&i915->drm, &conn_iter); > > for_each_intel_connector_iter(connector, &conn_iter) { > > if (i >= ARRAY_SIZE(opregion->acpi->cadl)) > > break; > > @@ -779,9 +779,9 @@ static void intel_setup_cadls(struct drm_i915_private *dev_priv) > > opregion->acpi->cadl[i] = 0; > > } > > -static void swsci_setup(struct drm_i915_private *dev_priv) > > +static void swsci_setup(struct drm_i915_private *i915) > > { > > - struct intel_opregion *opregion = &dev_priv->opregion; > > + struct intel_opregion *opregion = &i915->opregion; > > bool requested_callbacks = false; > > u32 tmp; > > @@ -790,7 +790,7 @@ static void swsci_setup(struct drm_i915_private *dev_priv) > > opregion->swsci_sbcb_sub_functions = 1; > > /* We use GBDA to ask for supported GBDA calls. */ > > - if (swsci(dev_priv, SWSCI_GBDA_SUPPORTED_CALLS, 0, &tmp) == 0) { > > + if (swsci(i915, SWSCI_GBDA_SUPPORTED_CALLS, 0, &tmp) == 0) { > > /* make the bits match the sub-function codes */ > > tmp <<= 1; > > opregion->swsci_gbda_sub_functions |= tmp; > > @@ -801,7 +801,7 @@ static void swsci_setup(struct drm_i915_private *dev_priv) > > * must not call interfaces that are not specifically requested by the > > * bios. > > */ > > - if (swsci(dev_priv, SWSCI_GBDA_REQUESTED_CALLBACKS, 0, &tmp) == 0) { > > + if (swsci(i915, SWSCI_GBDA_REQUESTED_CALLBACKS, 0, &tmp) == 0) { > > /* here, the bits already match sub-function codes */ > > opregion->swsci_sbcb_sub_functions |= tmp; > > requested_callbacks = true; > > @@ -812,7 +812,7 @@ static void swsci_setup(struct drm_i915_private *dev_priv) > > * the callback is _requested_. But we still can't call interfaces that > > * are not requested. > > */ > > - if (swsci(dev_priv, SWSCI_SBCB_SUPPORTED_CALLBACKS, 0, &tmp) == 0) { > > + if (swsci(i915, SWSCI_SBCB_SUPPORTED_CALLBACKS, 0, &tmp) == 0) { > > /* make the bits match the sub-function codes */ > > u32 low = tmp & 0x7ff; > > u32 high = tmp & ~0xfff; /* bit 11 is reserved */ > > @@ -854,9 +854,9 @@ static const struct dmi_system_id intel_no_opregion_vbt[] = { > > { } > > }; > > -static int intel_load_vbt_firmware(struct drm_i915_private *dev_priv) > > +static int intel_load_vbt_firmware(struct drm_i915_private *i915) > > { > > - struct intel_opregion *opregion = &dev_priv->opregion; > > + struct intel_opregion *opregion = &i915->opregion; > > const struct firmware *fw = NULL; > > const char *name = i915_modparams.vbt_firmware; > > int ret; > > @@ -864,7 +864,7 @@ static int intel_load_vbt_firmware(struct drm_i915_private *dev_priv) > > if (!name || !*name) > > return -ENOENT; > > - ret = request_firmware(&fw, name, &dev_priv->drm.pdev->dev); > > + ret = request_firmware(&fw, name, &i915->drm.pdev->dev); > > if (ret) { > > DRM_ERROR("Requesting VBT firmware \"%s\" failed (%d)\n", > > name, ret); > > @@ -891,10 +891,10 @@ static int intel_load_vbt_firmware(struct drm_i915_private *dev_priv) > > return ret; > > } > > -int intel_opregion_setup(struct drm_i915_private *dev_priv) > > +int intel_opregion_setup(struct drm_i915_private *i915) > > { > > - struct intel_opregion *opregion = &dev_priv->opregion; > > - struct pci_dev *pdev = dev_priv->drm.pdev; > > + struct intel_opregion *opregion = &i915->opregion; > > + struct pci_dev *pdev = i915->drm.pdev; > > u32 asls, mboxes; > > char buf[sizeof(OPREGION_SIGNATURE)]; > > int err = 0; > > @@ -945,7 +945,7 @@ int intel_opregion_setup(struct drm_i915_private *dev_priv) > > if (mboxes & MBOX_SWSCI) { > > DRM_DEBUG_DRIVER("SWSCI supported\n"); > > opregion->swsci = base + OPREGION_SWSCI_OFFSET; > > - swsci_setup(dev_priv); > > + swsci_setup(i915); > > } > > if (mboxes & MBOX_ASLE) { > > @@ -958,7 +958,7 @@ int intel_opregion_setup(struct drm_i915_private *dev_priv) > > if (mboxes & MBOX_ASLE_EXT) > > DRM_DEBUG_DRIVER("ASLE extension supported\n"); > > - if (intel_load_vbt_firmware(dev_priv) == 0) > > + if (intel_load_vbt_firmware(i915) == 0) > > goto out; > > if (dmi_check_system(intel_no_opregion_vbt)) > > @@ -1043,12 +1043,12 @@ static const struct dmi_system_id intel_use_opregion_panel_type[] = { > > }; > > int > > -intel_opregion_get_panel_type(struct drm_i915_private *dev_priv) > > +intel_opregion_get_panel_type(struct drm_i915_private *i915) > > { > > u32 panel_details; > > int ret; > > - ret = swsci(dev_priv, SWSCI_GBDA_PANEL_DETAILS, 0x0, &panel_details); > > + ret = swsci(i915, SWSCI_GBDA_PANEL_DETAILS, 0x0, &panel_details); > > if (ret) { > > DRM_DEBUG_KMS("Failed to get panel details from OpRegion (%d)\n", > > ret); > > diff --git a/drivers/gpu/drm/i915/intel_opregion.h b/drivers/gpu/drm/i915/intel_opregion.h > > index 4aa68ffbd30e..8e0b7bdecbe4 100644 > > --- a/drivers/gpu/drm/i915/intel_opregion.h > > +++ b/drivers/gpu/drm/i915/intel_opregion.h > > @@ -56,47 +56,47 @@ struct intel_opregion { > > #ifdef CONFIG_ACPI > > -int intel_opregion_setup(struct drm_i915_private *dev_priv); > > +int intel_opregion_setup(struct drm_i915_private *i915); > > -void intel_opregion_register(struct drm_i915_private *dev_priv); > > -void intel_opregion_unregister(struct drm_i915_private *dev_priv); > > +void intel_opregion_register(struct drm_i915_private *i915); > > +void intel_opregion_unregister(struct drm_i915_private *i915); > > -void intel_opregion_resume(struct drm_i915_private *dev_priv); > > -void intel_opregion_suspend(struct drm_i915_private *dev_priv, > > +void intel_opregion_resume(struct drm_i915_private *i915); > > +void intel_opregion_suspend(struct drm_i915_private *i915, > > pci_power_t state); > > -void intel_opregion_asle_intr(struct drm_i915_private *dev_priv); > > +void intel_opregion_asle_intr(struct drm_i915_private *i915); > > int intel_opregion_notify_encoder(struct intel_encoder *intel_encoder, > > bool enable); > > -int intel_opregion_notify_adapter(struct drm_i915_private *dev_priv, > > +int intel_opregion_notify_adapter(struct drm_i915_private *i915, > > pci_power_t state); > > -int intel_opregion_get_panel_type(struct drm_i915_private *dev_priv); > > +int intel_opregion_get_panel_type(struct drm_i915_private *i915); > > #else /* CONFIG_ACPI*/ > > -static inline int intel_opregion_setup(struct drm_i915_private *dev_priv) > > +static inline int intel_opregion_setup(struct drm_i915_private *i915) > > { > > return 0; > > } > > -static inline void intel_opregion_register(struct drm_i915_private *dev_priv) > > +static inline void intel_opregion_register(struct drm_i915_private *i915) > > { > > } > > -static inline void intel_opregion_unregister(struct drm_i915_private *dev_priv) > > +static inline void intel_opregion_unregister(struct drm_i915_private *i915) > > { > > } > > -static inline void intel_opregion_resume(struct drm_i915_private *dev_priv) > > +static inline void intel_opregion_resume(struct drm_i915_private *i915) > > { > > } > > -static inline void intel_opregion_suspend(struct drm_i915_private *dev_priv, > > +static inline void intel_opregion_suspend(struct drm_i915_private *i915, > > pci_power_t state) > > { > > } > > -static inline void intel_opregion_asle_intr(struct drm_i915_private *dev_priv) > > +static inline void intel_opregion_asle_intr(struct drm_i915_private *i915) > > { > > } > > diff --git a/drivers/gpu/drm/i915/intel_overlay.c b/drivers/gpu/drm/i915/intel_overlay.c > > index a2ac06a08715..802bc4c72ecd 100644 > > --- a/drivers/gpu/drm/i915/intel_overlay.c > > +++ b/drivers/gpu/drm/i915/intel_overlay.c > > @@ -193,10 +193,10 @@ struct intel_overlay { > > struct i915_active_request last_flip; > > }; > > -static void i830_overlay_clock_gating(struct drm_i915_private *dev_priv, > > +static void i830_overlay_clock_gating(struct drm_i915_private *i915, > > bool enable) > > { > > - struct pci_dev *pdev = dev_priv->drm.pdev; > > + struct pci_dev *pdev = i915->drm.pdev; > > u8 val; > > /* WA_OVERLAY_CLKGATE:alm */ > > @@ -247,7 +247,7 @@ static struct i915_request *alloc_request(struct intel_overlay *overlay) > > /* overlay needs to be disable in OCMD reg */ > > static int intel_overlay_on(struct intel_overlay *overlay) > > { > > - struct drm_i915_private *dev_priv = overlay->i915; > > + struct drm_i915_private *i915 = overlay->i915; > > struct i915_request *rq; > > u32 *cs; > > @@ -265,8 +265,8 @@ static int intel_overlay_on(struct intel_overlay *overlay) > > overlay->active = true; > > - if (IS_I830(dev_priv)) > > - i830_overlay_clock_gating(dev_priv, false); > > + if (IS_I830(i915)) > > + i830_overlay_clock_gating(i915, false); > > *cs++ = MI_OVERLAY_FLIP | MI_OVERLAY_ON; > > *cs++ = overlay->flip_addr | OFC_UPDATE; > > @@ -303,7 +303,7 @@ static int intel_overlay_continue(struct intel_overlay *overlay, > > struct i915_vma *vma, > > bool load_polyphase_filter) > > { > > - struct drm_i915_private *dev_priv = overlay->i915; > > + struct drm_i915_private *i915 = overlay->i915; > > struct i915_request *rq; > > u32 flip_addr = overlay->flip_addr; > > u32 tmp, *cs; > > @@ -369,7 +369,7 @@ static void intel_overlay_off_tail(struct i915_active_request *active, > > { > > struct intel_overlay *overlay = > > container_of(active, typeof(*overlay), last_flip); > > - struct drm_i915_private *dev_priv = overlay->i915; > > + struct drm_i915_private *i915 = overlay->i915; > > intel_overlay_release_old_vma(overlay); > > @@ -377,8 +377,8 @@ static void intel_overlay_off_tail(struct i915_active_request *active, > > overlay->crtc = NULL; > > overlay->active = false; > > - if (IS_I830(dev_priv)) > > - i830_overlay_clock_gating(dev_priv, true); > > + if (IS_I830(i915)) > > + i830_overlay_clock_gating(i915, true); > > } > > /* overlay needs to be disabled in OCMD reg */ > > @@ -437,11 +437,11 @@ static int intel_overlay_recover_from_interrupt(struct intel_overlay *overlay) > > */ > > static int intel_overlay_release_old_vid(struct intel_overlay *overlay) > > { > > - struct drm_i915_private *dev_priv = overlay->i915; > > + struct drm_i915_private *i915 = overlay->i915; > > u32 *cs; > > int ret; > > - lockdep_assert_held(&dev_priv->drm.struct_mutex); > > + lockdep_assert_held(&i915->drm.struct_mutex); > > /* Only wait if there is actually an old frame to release to > > * guarantee forward progress. > > @@ -477,9 +477,9 @@ static int intel_overlay_release_old_vid(struct intel_overlay *overlay) > > return 0; > > } > > -void intel_overlay_reset(struct drm_i915_private *dev_priv) > > +void intel_overlay_reset(struct drm_i915_private *i915) > > { > > - struct intel_overlay *overlay = dev_priv->overlay; > > + struct intel_overlay *overlay = i915->overlay; > > if (!overlay) > > return; > > @@ -540,11 +540,11 @@ static int uv_vsubsampling(u32 format) > > } > > } > > -static u32 calc_swidthsw(struct drm_i915_private *dev_priv, u32 offset, u32 width) > > +static u32 calc_swidthsw(struct drm_i915_private *i915, u32 offset, u32 width) > > { > > u32 sw; > > - if (IS_GEN(dev_priv, 2)) > > + if (IS_GEN(i915, 2)) > > sw = ALIGN((offset & 31) + width, 32); > > else > > sw = ALIGN((offset & 63) + width, 64); > > @@ -749,21 +749,21 @@ static int intel_overlay_do_put_image(struct intel_overlay *overlay, > > struct drm_intel_overlay_put_image *params) > > { > > struct overlay_registers __iomem *regs = overlay->regs; > > - struct drm_i915_private *dev_priv = overlay->i915; > > + struct drm_i915_private *i915 = overlay->i915; > > u32 swidth, swidthsw, sheight, ostride; > > enum pipe pipe = overlay->crtc->pipe; > > bool scale_changed = false; > > struct i915_vma *vma; > > int ret, tmp_width; > > - lockdep_assert_held(&dev_priv->drm.struct_mutex); > > - WARN_ON(!drm_modeset_is_locked(&dev_priv->drm.mode_config.connection_mutex)); > > + lockdep_assert_held(&i915->drm.struct_mutex); > > + WARN_ON(!drm_modeset_is_locked(&i915->drm.mode_config.connection_mutex)); > > ret = intel_overlay_release_old_vid(overlay); > > if (ret != 0) > > return ret; > > - atomic_inc(&dev_priv->gpu_error.pending_fb_pin); > > + atomic_inc(&i915->gpu_error.pending_fb_pin); > > i915_gem_object_lock(new_bo); > > vma = i915_gem_object_pin_to_display_plane(new_bo, > > @@ -783,7 +783,7 @@ static int intel_overlay_do_put_image(struct intel_overlay *overlay, > > u32 oconfig; > > oconfig = OCONF_CC_OUT_8BIT; > > - if (IS_GEN(dev_priv, 4)) > > + if (IS_GEN(i915, 4)) > > oconfig |= OCONF_CSC_MODE_BT709; > > oconfig |= pipe == 0 ? > > OCONF_PIPE_A : OCONF_PIPE_B; > > @@ -804,7 +804,7 @@ static int intel_overlay_do_put_image(struct intel_overlay *overlay, > > tmp_width = params->src_width; > > swidth = params->src_width; > > - swidthsw = calc_swidthsw(dev_priv, params->offset_Y, tmp_width); > > + swidthsw = calc_swidthsw(i915, params->offset_Y, tmp_width); > > sheight = params->src_height; > > iowrite32(i915_ggtt_offset(vma) + params->offset_Y, ®s->OBUF_0Y); > > ostride = params->stride_Y; > > @@ -817,9 +817,9 @@ static int intel_overlay_do_put_image(struct intel_overlay *overlay, > > swidth |= (params->src_width / uv_hscale) << 16; > > sheight |= (params->src_height / uv_vscale) << 16; > > - tmp_U = calc_swidthsw(dev_priv, params->offset_U, > > + tmp_U = calc_swidthsw(i915, params->offset_U, > > params->src_width / uv_hscale); > > - tmp_V = calc_swidthsw(dev_priv, params->offset_V, > > + tmp_V = calc_swidthsw(i915, params->offset_V, > > params->src_width / uv_hscale); > > swidthsw |= max(tmp_U, tmp_V) << 16; > > @@ -851,18 +851,18 @@ static int intel_overlay_do_put_image(struct intel_overlay *overlay, > > out_unpin: > > i915_gem_object_unpin_from_display_plane(vma); > > out_pin_section: > > - atomic_dec(&dev_priv->gpu_error.pending_fb_pin); > > + atomic_dec(&i915->gpu_error.pending_fb_pin); > > return ret; > > } > > int intel_overlay_switch_off(struct intel_overlay *overlay) > > { > > - struct drm_i915_private *dev_priv = overlay->i915; > > + struct drm_i915_private *i915 = overlay->i915; > > int ret; > > - lockdep_assert_held(&dev_priv->drm.struct_mutex); > > - WARN_ON(!drm_modeset_is_locked(&dev_priv->drm.mode_config.connection_mutex)); > > + lockdep_assert_held(&i915->drm.struct_mutex); > > + WARN_ON(!drm_modeset_is_locked(&i915->drm.mode_config.connection_mutex)); > > ret = intel_overlay_recover_from_interrupt(overlay); > > if (ret != 0) > > @@ -895,14 +895,14 @@ static int check_overlay_possible_on_crtc(struct intel_overlay *overlay, > > static void update_pfit_vscale_ratio(struct intel_overlay *overlay) > > { > > - struct drm_i915_private *dev_priv = overlay->i915; > > + struct drm_i915_private *i915 = overlay->i915; > > u32 pfit_control = I915_READ(PFIT_CONTROL); > > u32 ratio; > > /* XXX: This is not the same logic as in the xorg driver, but more in > > * line with the intel documentation for the i965 > > */ > > - if (INTEL_GEN(dev_priv) >= 4) { > > + if (INTEL_GEN(i915) >= 4) { > > /* on i965 use the PGM reg to read out the autoscaler values */ > > ratio = I915_READ(PFIT_PGM_RATIOS) >> PFIT_VERT_SCALE_SHIFT_965; > > } else { > > @@ -947,7 +947,7 @@ static int check_overlay_scaling(struct drm_intel_overlay_put_image *rec) > > return 0; > > } > > -static int check_overlay_src(struct drm_i915_private *dev_priv, > > +static int check_overlay_src(struct drm_i915_private *i915, > > struct drm_intel_overlay_put_image *rec, > > struct drm_i915_gem_object *new_bo) > > { > > @@ -958,7 +958,7 @@ static int check_overlay_src(struct drm_i915_private *dev_priv, > > u32 tmp; > > /* check src dimensions */ > > - if (IS_I845G(dev_priv) || IS_I830(dev_priv)) { > > + if (IS_I845G(i915) || IS_I830(i915)) { > > if (rec->src_height > IMAGE_MAX_HEIGHT_LEGACY || > > rec->src_width > IMAGE_MAX_WIDTH_LEGACY) > > return -EINVAL; > > @@ -1010,14 +1010,14 @@ static int check_overlay_src(struct drm_i915_private *dev_priv, > > return -EINVAL; > > /* stride checking */ > > - if (IS_I830(dev_priv) || IS_I845G(dev_priv)) > > + if (IS_I830(i915) || IS_I845G(i915)) > > stride_mask = 255; > > else > > stride_mask = 63; > > if (rec->stride_Y & stride_mask || rec->stride_UV & stride_mask) > > return -EINVAL; > > - if (IS_GEN(dev_priv, 4) && rec->stride_Y < 512) > > + if (IS_GEN(i915, 4) && rec->stride_Y < 512) > > return -EINVAL; > > tmp = (rec->flags & I915_OVERLAY_TYPE_MASK) == I915_OVERLAY_YUV_PLANAR ? > > @@ -1062,14 +1062,14 @@ int intel_overlay_put_image_ioctl(struct drm_device *dev, void *data, > > struct drm_file *file_priv) > > { > > struct drm_intel_overlay_put_image *params = data; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > struct intel_overlay *overlay; > > struct drm_crtc *drmmode_crtc; > > struct intel_crtc *crtc; > > struct drm_i915_gem_object *new_bo; > > int ret; > > - overlay = dev_priv->overlay; > > + overlay = i915->overlay; > > if (!overlay) { > > DRM_DEBUG("userspace bug: no overlay\n"); > > return -ENODEV; > > @@ -1148,7 +1148,7 @@ int intel_overlay_put_image_ioctl(struct drm_device *dev, void *data, > > goto out_unlock; > > } > > - ret = check_overlay_src(dev_priv, params, new_bo); > > + ret = check_overlay_src(i915, params, new_bo); > > if (ret != 0) > > goto out_unlock; > > @@ -1231,11 +1231,11 @@ int intel_overlay_attrs_ioctl(struct drm_device *dev, void *data, > > struct drm_file *file_priv) > > { > > struct drm_intel_overlay_attrs *attrs = data; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > struct intel_overlay *overlay; > > int ret; > > - overlay = dev_priv->overlay; > > + overlay = i915->overlay; > > if (!overlay) { > > DRM_DEBUG("userspace bug: no overlay\n"); > > return -ENODEV; > > @@ -1251,7 +1251,7 @@ int intel_overlay_attrs_ioctl(struct drm_device *dev, void *data, > > attrs->contrast = overlay->contrast; > > attrs->saturation = overlay->saturation; > > - if (!IS_GEN(dev_priv, 2)) { > > + if (!IS_GEN(i915, 2)) { > > attrs->gamma0 = I915_READ(OGAMC0); > > attrs->gamma1 = I915_READ(OGAMC1); > > attrs->gamma2 = I915_READ(OGAMC2); > > @@ -1275,7 +1275,7 @@ int intel_overlay_attrs_ioctl(struct drm_device *dev, void *data, > > update_reg_attrs(overlay, overlay->regs); > > if (attrs->flags & I915_OVERLAY_UPDATE_GAMMA) { > > - if (IS_GEN(dev_priv, 2)) > > + if (IS_GEN(i915, 2)) > > goto out_unlock; > > if (overlay->active) { > > @@ -1351,19 +1351,19 @@ static int get_registers(struct intel_overlay *overlay, bool use_phys) > > return err; > > } > > -void intel_overlay_setup(struct drm_i915_private *dev_priv) > > +void intel_overlay_setup(struct drm_i915_private *i915) > > { > > struct intel_overlay *overlay; > > int ret; > > - if (!HAS_OVERLAY(dev_priv)) > > + if (!HAS_OVERLAY(i915)) > > return; > > overlay = kzalloc(sizeof(*overlay), GFP_KERNEL); > > if (!overlay) > > return; > > - overlay->i915 = dev_priv; > > + overlay->i915 = i915; > > overlay->color_key = 0x0101fe; > > overlay->color_key_enabled = true; > > @@ -1373,7 +1373,7 @@ void intel_overlay_setup(struct drm_i915_private *dev_priv) > > INIT_ACTIVE_REQUEST(&overlay->last_flip); > > - ret = get_registers(overlay, OVERLAY_NEEDS_PHYSICAL(dev_priv)); > > + ret = get_registers(overlay, OVERLAY_NEEDS_PHYSICAL(i915)); > > if (ret) > > goto out_free; > > @@ -1387,7 +1387,7 @@ void intel_overlay_setup(struct drm_i915_private *dev_priv) > > update_polyphase_filter(overlay->regs); > > update_reg_attrs(overlay, overlay->regs); > > - dev_priv->overlay = overlay; > > + i915->overlay = overlay; > > DRM_INFO("Initialized overlay support.\n"); > > return; > > @@ -1397,11 +1397,11 @@ void intel_overlay_setup(struct drm_i915_private *dev_priv) > > kfree(overlay); > > } > > -void intel_overlay_cleanup(struct drm_i915_private *dev_priv) > > +void intel_overlay_cleanup(struct drm_i915_private *i915) > > { > > struct intel_overlay *overlay; > > - overlay = fetch_and_zero(&dev_priv->overlay); > > + overlay = fetch_and_zero(&i915->overlay); > > if (!overlay) > > return; > > @@ -1427,9 +1427,9 @@ struct intel_overlay_error_state { > > }; > > struct intel_overlay_error_state * > > -intel_overlay_capture_error_state(struct drm_i915_private *dev_priv) > > +intel_overlay_capture_error_state(struct drm_i915_private *i915) > > { > > - struct intel_overlay *overlay = dev_priv->overlay; > > + struct intel_overlay *overlay = i915->overlay; > > struct intel_overlay_error_state *error; > > if (!overlay || !overlay->active) > > diff --git a/drivers/gpu/drm/i915/intel_overlay.h b/drivers/gpu/drm/i915/intel_overlay.h > > index a167c28acd27..2e8771cffee9 100644 > > --- a/drivers/gpu/drm/i915/intel_overlay.h > > +++ b/drivers/gpu/drm/i915/intel_overlay.h > > @@ -13,16 +13,16 @@ struct drm_i915_private; > > struct intel_overlay; > > struct intel_overlay_error_state; > > -void intel_overlay_setup(struct drm_i915_private *dev_priv); > > -void intel_overlay_cleanup(struct drm_i915_private *dev_priv); > > +void intel_overlay_setup(struct drm_i915_private *i915); > > +void intel_overlay_cleanup(struct drm_i915_private *i915); > > int intel_overlay_switch_off(struct intel_overlay *overlay); > > int intel_overlay_put_image_ioctl(struct drm_device *dev, void *data, > > struct drm_file *file_priv); > > int intel_overlay_attrs_ioctl(struct drm_device *dev, void *data, > > struct drm_file *file_priv); > > -void intel_overlay_reset(struct drm_i915_private *dev_priv); > > +void intel_overlay_reset(struct drm_i915_private *i915); > > struct intel_overlay_error_state * > > -intel_overlay_capture_error_state(struct drm_i915_private *dev_priv); > > +intel_overlay_capture_error_state(struct drm_i915_private *i915); > > void intel_overlay_print_error_state(struct drm_i915_error_state_buf *e, > > struct intel_overlay_error_state *error); > > diff --git a/drivers/gpu/drm/i915/intel_panel.c b/drivers/gpu/drm/i915/intel_panel.c > > index 9cd4e37e3934..d80115089445 100644 > > --- a/drivers/gpu/drm/i915/intel_panel.c > > +++ b/drivers/gpu/drm/i915/intel_panel.c > > @@ -65,7 +65,7 @@ struct drm_display_mode * > > intel_panel_edid_downclock_mode(struct intel_connector *connector, > > const struct drm_display_mode *fixed_mode) > > { > > - struct drm_i915_private *dev_priv = to_i915(connector->base.dev); > > + struct drm_i915_private *i915 = to_i915(connector->base.dev); > > const struct drm_display_mode *scan, *best_mode = NULL; > > struct drm_display_mode *downclock_mode; > > int best_clock = fixed_mode->clock; > > @@ -92,7 +92,7 @@ intel_panel_edid_downclock_mode(struct intel_connector *connector, > > if (!best_mode) > > return NULL; > > - downclock_mode = drm_mode_duplicate(&dev_priv->drm, best_mode); > > + downclock_mode = drm_mode_duplicate(&i915->drm, best_mode); > > if (!downclock_mode) > > return NULL; > > @@ -106,7 +106,7 @@ intel_panel_edid_downclock_mode(struct intel_connector *connector, > > struct drm_display_mode * > > intel_panel_edid_fixed_mode(struct intel_connector *connector) > > { > > - struct drm_i915_private *dev_priv = to_i915(connector->base.dev); > > + struct drm_i915_private *i915 = to_i915(connector->base.dev); > > const struct drm_display_mode *scan; > > struct drm_display_mode *fixed_mode; > > @@ -118,7 +118,7 @@ intel_panel_edid_fixed_mode(struct intel_connector *connector) > > if ((scan->type & DRM_MODE_TYPE_PREFERRED) == 0) > > continue; > > - fixed_mode = drm_mode_duplicate(&dev_priv->drm, scan); > > + fixed_mode = drm_mode_duplicate(&i915->drm, scan); > > if (!fixed_mode) > > return NULL; > > @@ -132,7 +132,7 @@ intel_panel_edid_fixed_mode(struct intel_connector *connector) > > scan = list_first_entry(&connector->base.probed_modes, > > typeof(*scan), head); > > - fixed_mode = drm_mode_duplicate(&dev_priv->drm, scan); > > + fixed_mode = drm_mode_duplicate(&i915->drm, scan); > > if (!fixed_mode) > > return NULL; > > @@ -148,15 +148,15 @@ intel_panel_edid_fixed_mode(struct intel_connector *connector) > > struct drm_display_mode * > > intel_panel_vbt_fixed_mode(struct intel_connector *connector) > > { > > - struct drm_i915_private *dev_priv = to_i915(connector->base.dev); > > + struct drm_i915_private *i915 = to_i915(connector->base.dev); > > struct drm_display_info *info = &connector->base.display_info; > > struct drm_display_mode *fixed_mode; > > - if (!dev_priv->vbt.lfp_lvds_vbt_mode) > > + if (!i915->vbt.lfp_lvds_vbt_mode) > > return NULL; > > - fixed_mode = drm_mode_duplicate(&dev_priv->drm, > > - dev_priv->vbt.lfp_lvds_vbt_mode); > > + fixed_mode = drm_mode_duplicate(&i915->drm, > > + i915->vbt.lfp_lvds_vbt_mode); > > if (!fixed_mode) > > return NULL; > > @@ -378,7 +378,7 @@ void intel_gmch_panel_fitting(struct intel_crtc *intel_crtc, > > struct intel_crtc_state *pipe_config, > > int fitting_mode) > > { > > - struct drm_i915_private *dev_priv = to_i915(intel_crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(intel_crtc->base.dev); > > u32 pfit_control = 0, pfit_pgm_ratios = 0, border = 0; > > struct drm_display_mode *adjusted_mode = &pipe_config->base.adjusted_mode; > > @@ -399,7 +399,7 @@ void intel_gmch_panel_fitting(struct intel_crtc *intel_crtc, > > break; > > case DRM_MODE_SCALE_ASPECT: > > /* Scale but preserve the aspect ratio */ > > - if (INTEL_GEN(dev_priv) >= 4) > > + if (INTEL_GEN(i915) >= 4) > > i965_scale_aspect(pipe_config, &pfit_control); > > else > > i9xx_scale_aspect(pipe_config, &pfit_control, > > @@ -413,7 +413,7 @@ void intel_gmch_panel_fitting(struct intel_crtc *intel_crtc, > > if (pipe_config->pipe_src_h != adjusted_mode->crtc_vdisplay || > > pipe_config->pipe_src_w != adjusted_mode->crtc_hdisplay) { > > pfit_control |= PFIT_ENABLE; > > - if (INTEL_GEN(dev_priv) >= 4) > > + if (INTEL_GEN(i915) >= 4) > > pfit_control |= PFIT_SCALING_AUTO; > > else > > pfit_control |= (VERT_AUTO_SCALE | > > @@ -429,7 +429,7 @@ void intel_gmch_panel_fitting(struct intel_crtc *intel_crtc, > > /* 965+ wants fuzzy fitting */ > > /* FIXME: handle multiple panels by failing gracefully */ > > - if (INTEL_GEN(dev_priv) >= 4) > > + if (INTEL_GEN(i915) >= 4) > > pfit_control |= ((intel_crtc->pipe << PFIT_PIPE_SHIFT) | > > PFIT_FILTER_FUZZY); > > @@ -440,7 +440,7 @@ void intel_gmch_panel_fitting(struct intel_crtc *intel_crtc, > > } > > /* Make sure pre-965 set dither correctly for 18bpp panels. */ > > - if (INTEL_GEN(dev_priv) < 4 && pipe_config->pipe_bpp == 18) > > + if (INTEL_GEN(i915) < 4 && pipe_config->pipe_bpp == 18) > > pfit_control |= PANEL_8TO6_DITHER_ENABLE; > > pipe_config->gmch_pfit.control = pfit_control; > > @@ -517,7 +517,7 @@ static inline u32 scale_hw_to_user(struct intel_connector *connector, > > static u32 intel_panel_compute_brightness(struct intel_connector *connector, > > u32 val) > > { > > - struct drm_i915_private *dev_priv = to_i915(connector->base.dev); > > + struct drm_i915_private *i915 = to_i915(connector->base.dev); > > struct intel_panel *panel = &connector->panel; > > WARN_ON(panel->backlight.max == 0); > > @@ -526,7 +526,7 @@ static u32 intel_panel_compute_brightness(struct intel_connector *connector, > > return val; > > if (i915_modparams.invert_brightness > 0 || > > - dev_priv->quirks & QUIRK_INVERT_BRIGHTNESS) { > > + i915->quirks & QUIRK_INVERT_BRIGHTNESS) { > > return panel->backlight.max - val + panel->backlight.min; > > } > > @@ -535,39 +535,39 @@ static u32 intel_panel_compute_brightness(struct intel_connector *connector, > > static u32 lpt_get_backlight(struct intel_connector *connector) > > { > > - struct drm_i915_private *dev_priv = to_i915(connector->base.dev); > > + struct drm_i915_private *i915 = to_i915(connector->base.dev); > > return I915_READ(BLC_PWM_PCH_CTL2) & BACKLIGHT_DUTY_CYCLE_MASK; > > } > > static u32 pch_get_backlight(struct intel_connector *connector) > > { > > - struct drm_i915_private *dev_priv = to_i915(connector->base.dev); > > + struct drm_i915_private *i915 = to_i915(connector->base.dev); > > return I915_READ(BLC_PWM_CPU_CTL) & BACKLIGHT_DUTY_CYCLE_MASK; > > } > > static u32 i9xx_get_backlight(struct intel_connector *connector) > > { > > - struct drm_i915_private *dev_priv = to_i915(connector->base.dev); > > + struct drm_i915_private *i915 = to_i915(connector->base.dev); > > struct intel_panel *panel = &connector->panel; > > u32 val; > > val = I915_READ(BLC_PWM_CTL) & BACKLIGHT_DUTY_CYCLE_MASK; > > - if (INTEL_GEN(dev_priv) < 4) > > + if (INTEL_GEN(i915) < 4) > > val >>= 1; > > if (panel->backlight.combination_mode) { > > u8 lbpc; > > - pci_read_config_byte(dev_priv->drm.pdev, LBPC, &lbpc); > > + pci_read_config_byte(i915->drm.pdev, LBPC, &lbpc); > > val *= lbpc; > > } > > return val; > > } > > -static u32 _vlv_get_backlight(struct drm_i915_private *dev_priv, enum pipe pipe) > > +static u32 _vlv_get_backlight(struct drm_i915_private *i915, enum pipe pipe) > > { > > if (WARN_ON(pipe != PIPE_A && pipe != PIPE_B)) > > return 0; > > @@ -577,15 +577,15 @@ static u32 _vlv_get_backlight(struct drm_i915_private *dev_priv, enum pipe pipe) > > static u32 vlv_get_backlight(struct intel_connector *connector) > > { > > - struct drm_i915_private *dev_priv = to_i915(connector->base.dev); > > + struct drm_i915_private *i915 = to_i915(connector->base.dev); > > enum pipe pipe = intel_connector_get_pipe(connector); > > - return _vlv_get_backlight(dev_priv, pipe); > > + return _vlv_get_backlight(i915, pipe); > > } > > static u32 bxt_get_backlight(struct intel_connector *connector) > > { > > - struct drm_i915_private *dev_priv = to_i915(connector->base.dev); > > + struct drm_i915_private *i915 = to_i915(connector->base.dev); > > struct intel_panel *panel = &connector->panel; > > return I915_READ(BXT_BLC_PWM_DUTY(panel->backlight.controller)); > > @@ -603,7 +603,7 @@ static u32 pwm_get_backlight(struct intel_connector *connector) > > static void lpt_set_backlight(const struct drm_connector_state *conn_state, u32 level) > > { > > struct intel_connector *connector = to_intel_connector(conn_state->connector); > > - struct drm_i915_private *dev_priv = to_i915(connector->base.dev); > > + struct drm_i915_private *i915 = to_i915(connector->base.dev); > > u32 val = I915_READ(BLC_PWM_PCH_CTL2) & ~BACKLIGHT_DUTY_CYCLE_MASK; > > I915_WRITE(BLC_PWM_PCH_CTL2, val | level); > > @@ -612,7 +612,7 @@ static void lpt_set_backlight(const struct drm_connector_state *conn_state, u32 > > static void pch_set_backlight(const struct drm_connector_state *conn_state, u32 level) > > { > > struct intel_connector *connector = to_intel_connector(conn_state->connector); > > - struct drm_i915_private *dev_priv = to_i915(connector->base.dev); > > + struct drm_i915_private *i915 = to_i915(connector->base.dev); > > u32 tmp; > > tmp = I915_READ(BLC_PWM_CPU_CTL) & ~BACKLIGHT_DUTY_CYCLE_MASK; > > @@ -622,7 +622,7 @@ static void pch_set_backlight(const struct drm_connector_state *conn_state, u32 > > static void i9xx_set_backlight(const struct drm_connector_state *conn_state, u32 level) > > { > > struct intel_connector *connector = to_intel_connector(conn_state->connector); > > - struct drm_i915_private *dev_priv = to_i915(connector->base.dev); > > + struct drm_i915_private *i915 = to_i915(connector->base.dev); > > struct intel_panel *panel = &connector->panel; > > u32 tmp, mask; > > @@ -633,10 +633,10 @@ static void i9xx_set_backlight(const struct drm_connector_state *conn_state, u32 > > lbpc = level * 0xfe / panel->backlight.max + 1; > > level /= lbpc; > > - pci_write_config_byte(dev_priv->drm.pdev, LBPC, lbpc); > > + pci_write_config_byte(i915->drm.pdev, LBPC, lbpc); > > } > > - if (IS_GEN(dev_priv, 4)) { > > + if (IS_GEN(i915, 4)) { > > mask = BACKLIGHT_DUTY_CYCLE_MASK; > > } else { > > level <<= 1; > > @@ -650,7 +650,7 @@ static void i9xx_set_backlight(const struct drm_connector_state *conn_state, u32 > > static void vlv_set_backlight(const struct drm_connector_state *conn_state, u32 level) > > { > > struct intel_connector *connector = to_intel_connector(conn_state->connector); > > - struct drm_i915_private *dev_priv = to_i915(connector->base.dev); > > + struct drm_i915_private *i915 = to_i915(connector->base.dev); > > enum pipe pipe = to_intel_crtc(conn_state->crtc)->pipe; > > u32 tmp; > > @@ -661,7 +661,7 @@ static void vlv_set_backlight(const struct drm_connector_state *conn_state, u32 > > static void bxt_set_backlight(const struct drm_connector_state *conn_state, u32 level) > > { > > struct intel_connector *connector = to_intel_connector(conn_state->connector); > > - struct drm_i915_private *dev_priv = to_i915(connector->base.dev); > > + struct drm_i915_private *i915 = to_i915(connector->base.dev); > > struct intel_panel *panel = &connector->panel; > > I915_WRITE(BXT_BLC_PWM_DUTY(panel->backlight.controller), level); > > @@ -694,7 +694,7 @@ void intel_panel_set_backlight_acpi(const struct drm_connector_state *conn_state > > u32 user_level, u32 user_max) > > { > > struct intel_connector *connector = to_intel_connector(conn_state->connector); > > - struct drm_i915_private *dev_priv = to_i915(connector->base.dev); > > + struct drm_i915_private *i915 = to_i915(connector->base.dev); > > struct intel_panel *panel = &connector->panel; > > u32 hw_level; > > @@ -707,7 +707,7 @@ void intel_panel_set_backlight_acpi(const struct drm_connector_state *conn_state > > if (!panel->backlight.present || !conn_state->crtc) > > return; > > - mutex_lock(&dev_priv->backlight_lock); > > + mutex_lock(&i915->backlight_lock); > > WARN_ON(panel->backlight.max == 0); > > @@ -723,13 +723,13 @@ void intel_panel_set_backlight_acpi(const struct drm_connector_state *conn_state > > if (panel->backlight.enabled) > > intel_panel_actually_set_backlight(conn_state, hw_level); > > - mutex_unlock(&dev_priv->backlight_lock); > > + mutex_unlock(&i915->backlight_lock); > > } > > static void lpt_disable_backlight(const struct drm_connector_state *old_conn_state) > > { > > struct intel_connector *connector = to_intel_connector(old_conn_state->connector); > > - struct drm_i915_private *dev_priv = to_i915(connector->base.dev); > > + struct drm_i915_private *i915 = to_i915(connector->base.dev); > > u32 tmp; > > intel_panel_actually_set_backlight(old_conn_state, 0); > > @@ -755,7 +755,7 @@ static void lpt_disable_backlight(const struct drm_connector_state *old_conn_sta > > static void pch_disable_backlight(const struct drm_connector_state *old_conn_state) > > { > > struct intel_connector *connector = to_intel_connector(old_conn_state->connector); > > - struct drm_i915_private *dev_priv = to_i915(connector->base.dev); > > + struct drm_i915_private *i915 = to_i915(connector->base.dev); > > u32 tmp; > > intel_panel_actually_set_backlight(old_conn_state, 0); > > @@ -774,7 +774,7 @@ static void i9xx_disable_backlight(const struct drm_connector_state *old_conn_st > > static void i965_disable_backlight(const struct drm_connector_state *old_conn_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(old_conn_state->connector->dev); > > + struct drm_i915_private *i915 = to_i915(old_conn_state->connector->dev); > > u32 tmp; > > intel_panel_actually_set_backlight(old_conn_state, 0); > > @@ -786,7 +786,7 @@ static void i965_disable_backlight(const struct drm_connector_state *old_conn_st > > static void vlv_disable_backlight(const struct drm_connector_state *old_conn_state) > > { > > struct intel_connector *connector = to_intel_connector(old_conn_state->connector); > > - struct drm_i915_private *dev_priv = to_i915(connector->base.dev); > > + struct drm_i915_private *i915 = to_i915(connector->base.dev); > > enum pipe pipe = to_intel_crtc(old_conn_state->crtc)->pipe; > > u32 tmp; > > @@ -799,7 +799,7 @@ static void vlv_disable_backlight(const struct drm_connector_state *old_conn_sta > > static void bxt_disable_backlight(const struct drm_connector_state *old_conn_state) > > { > > struct intel_connector *connector = to_intel_connector(old_conn_state->connector); > > - struct drm_i915_private *dev_priv = to_i915(connector->base.dev); > > + struct drm_i915_private *i915 = to_i915(connector->base.dev); > > struct intel_panel *panel = &connector->panel; > > u32 tmp, val; > > @@ -819,7 +819,7 @@ static void bxt_disable_backlight(const struct drm_connector_state *old_conn_sta > > static void cnp_disable_backlight(const struct drm_connector_state *old_conn_state) > > { > > struct intel_connector *connector = to_intel_connector(old_conn_state->connector); > > - struct drm_i915_private *dev_priv = to_i915(connector->base.dev); > > + struct drm_i915_private *i915 = to_i915(connector->base.dev); > > struct intel_panel *panel = &connector->panel; > > u32 tmp; > > @@ -844,7 +844,7 @@ static void pwm_disable_backlight(const struct drm_connector_state *old_conn_sta > > void intel_panel_disable_backlight(const struct drm_connector_state *old_conn_state) > > { > > struct intel_connector *connector = to_intel_connector(old_conn_state->connector); > > - struct drm_i915_private *dev_priv = to_i915(connector->base.dev); > > + struct drm_i915_private *i915 = to_i915(connector->base.dev); > > struct intel_panel *panel = &connector->panel; > > if (!panel->backlight.present) > > @@ -856,26 +856,26 @@ void intel_panel_disable_backlight(const struct drm_connector_state *old_conn_st > > * backlight. This will leave the backlight on unnecessarily when > > * another client is not activated. > > */ > > - if (dev_priv->drm.switch_power_state == DRM_SWITCH_POWER_CHANGING) { > > + if (i915->drm.switch_power_state == DRM_SWITCH_POWER_CHANGING) { > > DRM_DEBUG_DRIVER("Skipping backlight disable on vga switch\n"); > > return; > > } > > - mutex_lock(&dev_priv->backlight_lock); > > + mutex_lock(&i915->backlight_lock); > > if (panel->backlight.device) > > panel->backlight.device->props.power = FB_BLANK_POWERDOWN; > > panel->backlight.enabled = false; > > panel->backlight.disable(old_conn_state); > > - mutex_unlock(&dev_priv->backlight_lock); > > + mutex_unlock(&i915->backlight_lock); > > } > > static void lpt_enable_backlight(const struct intel_crtc_state *crtc_state, > > const struct drm_connector_state *conn_state) > > { > > struct intel_connector *connector = to_intel_connector(conn_state->connector); > > - struct drm_i915_private *dev_priv = to_i915(connector->base.dev); > > + struct drm_i915_private *i915 = to_i915(connector->base.dev); > > struct intel_panel *panel = &connector->panel; > > u32 pch_ctl1, pch_ctl2, schicken; > > @@ -886,7 +886,7 @@ static void lpt_enable_backlight(const struct intel_crtc_state *crtc_state, > > I915_WRITE(BLC_PWM_PCH_CTL1, pch_ctl1); > > } > > - if (HAS_PCH_LPT(dev_priv)) { > > + if (HAS_PCH_LPT(i915)) { > > schicken = I915_READ(SOUTH_CHICKEN2); > > if (panel->backlight.alternate_pwm_increment) > > schicken |= LPT_PWM_GRANULARITY; > > @@ -910,7 +910,7 @@ static void lpt_enable_backlight(const struct intel_crtc_state *crtc_state, > > pch_ctl1 |= BLM_PCH_POLARITY; > > /* After LPT, override is the default. */ > > - if (HAS_PCH_LPT(dev_priv)) > > + if (HAS_PCH_LPT(i915)) > > pch_ctl1 |= BLM_PCH_OVERRIDE_ENABLE; > > I915_WRITE(BLC_PWM_PCH_CTL1, pch_ctl1); > > @@ -925,7 +925,7 @@ static void pch_enable_backlight(const struct intel_crtc_state *crtc_state, > > const struct drm_connector_state *conn_state) > > { > > struct intel_connector *connector = to_intel_connector(conn_state->connector); > > - struct drm_i915_private *dev_priv = to_i915(connector->base.dev); > > + struct drm_i915_private *i915 = to_i915(connector->base.dev); > > struct intel_panel *panel = &connector->panel; > > enum transcoder cpu_transcoder = crtc_state->cpu_transcoder; > > u32 cpu_ctl2, pch_ctl1, pch_ctl2; > > @@ -971,7 +971,7 @@ static void i9xx_enable_backlight(const struct intel_crtc_state *crtc_state, > > const struct drm_connector_state *conn_state) > > { > > struct intel_connector *connector = to_intel_connector(conn_state->connector); > > - struct drm_i915_private *dev_priv = to_i915(connector->base.dev); > > + struct drm_i915_private *i915 = to_i915(connector->base.dev); > > struct intel_panel *panel = &connector->panel; > > u32 ctl, freq; > > @@ -988,7 +988,7 @@ static void i9xx_enable_backlight(const struct intel_crtc_state *crtc_state, > > ctl = freq << 17; > > if (panel->backlight.combination_mode) > > ctl |= BLM_LEGACY_MODE; > > - if (IS_PINEVIEW(dev_priv) && panel->backlight.active_low_pwm) > > + if (IS_PINEVIEW(i915) && panel->backlight.active_low_pwm) > > ctl |= BLM_POLARITY_PNV; > > I915_WRITE(BLC_PWM_CTL, ctl); > > @@ -1002,7 +1002,7 @@ static void i9xx_enable_backlight(const struct intel_crtc_state *crtc_state, > > * 855gm only, but checking for gen2 is safe, as 855gm is the only gen2 > > * that has backlight. > > */ > > - if (IS_GEN(dev_priv, 2)) > > + if (IS_GEN(i915, 2)) > > I915_WRITE(BLC_HIST_CTL, BLM_HISTOGRAM_ENABLE); > > } > > @@ -1010,7 +1010,7 @@ static void i965_enable_backlight(const struct intel_crtc_state *crtc_state, > > const struct drm_connector_state *conn_state) > > { > > struct intel_connector *connector = to_intel_connector(conn_state->connector); > > - struct drm_i915_private *dev_priv = to_i915(connector->base.dev); > > + struct drm_i915_private *i915 = to_i915(connector->base.dev); > > struct intel_panel *panel = &connector->panel; > > enum pipe pipe = to_intel_crtc(conn_state->crtc)->pipe; > > u32 ctl, ctl2, freq; > > @@ -1045,7 +1045,7 @@ static void vlv_enable_backlight(const struct intel_crtc_state *crtc_state, > > const struct drm_connector_state *conn_state) > > { > > struct intel_connector *connector = to_intel_connector(conn_state->connector); > > - struct drm_i915_private *dev_priv = to_i915(connector->base.dev); > > + struct drm_i915_private *i915 = to_i915(connector->base.dev); > > struct intel_panel *panel = &connector->panel; > > enum pipe pipe = to_intel_crtc(crtc_state->base.crtc)->pipe; > > u32 ctl, ctl2; > > @@ -1075,7 +1075,7 @@ static void bxt_enable_backlight(const struct intel_crtc_state *crtc_state, > > const struct drm_connector_state *conn_state) > > { > > struct intel_connector *connector = to_intel_connector(conn_state->connector); > > - struct drm_i915_private *dev_priv = to_i915(connector->base.dev); > > + struct drm_i915_private *i915 = to_i915(connector->base.dev); > > struct intel_panel *panel = &connector->panel; > > enum pipe pipe = to_intel_crtc(crtc_state->base.crtc)->pipe; > > u32 pwm_ctl, val; > > @@ -1123,7 +1123,7 @@ static void cnp_enable_backlight(const struct intel_crtc_state *crtc_state, > > const struct drm_connector_state *conn_state) > > { > > struct intel_connector *connector = to_intel_connector(conn_state->connector); > > - struct drm_i915_private *dev_priv = to_i915(connector->base.dev); > > + struct drm_i915_private *i915 = to_i915(connector->base.dev); > > struct intel_panel *panel = &connector->panel; > > u32 pwm_ctl; > > @@ -1187,7 +1187,7 @@ void intel_panel_enable_backlight(const struct intel_crtc_state *crtc_state, > > const struct drm_connector_state *conn_state) > > { > > struct intel_connector *connector = to_intel_connector(conn_state->connector); > > - struct drm_i915_private *dev_priv = to_i915(connector->base.dev); > > + struct drm_i915_private *i915 = to_i915(connector->base.dev); > > struct intel_panel *panel = &connector->panel; > > enum pipe pipe = to_intel_crtc(crtc_state->base.crtc)->pipe; > > @@ -1196,28 +1196,28 @@ void intel_panel_enable_backlight(const struct intel_crtc_state *crtc_state, > > DRM_DEBUG_KMS("pipe %c\n", pipe_name(pipe)); > > - mutex_lock(&dev_priv->backlight_lock); > > + mutex_lock(&i915->backlight_lock); > > __intel_panel_enable_backlight(crtc_state, conn_state); > > - mutex_unlock(&dev_priv->backlight_lock); > > + mutex_unlock(&i915->backlight_lock); > > } > > #if IS_ENABLED(CONFIG_BACKLIGHT_CLASS_DEVICE) > > static u32 intel_panel_get_backlight(struct intel_connector *connector) > > { > > - struct drm_i915_private *dev_priv = to_i915(connector->base.dev); > > + struct drm_i915_private *i915 = to_i915(connector->base.dev); > > struct intel_panel *panel = &connector->panel; > > u32 val = 0; > > - mutex_lock(&dev_priv->backlight_lock); > > + mutex_lock(&i915->backlight_lock); > > if (panel->backlight.enabled) { > > val = panel->backlight.get(connector); > > val = intel_panel_compute_brightness(connector, val); > > } > > - mutex_unlock(&dev_priv->backlight_lock); > > + mutex_unlock(&i915->backlight_lock); > > DRM_DEBUG_DRIVER("get backlight PWM = %d\n", val); > > return val; > > @@ -1228,14 +1228,14 @@ static void intel_panel_set_backlight(const struct drm_connector_state *conn_sta > > u32 user_level, u32 user_max) > > { > > struct intel_connector *connector = to_intel_connector(conn_state->connector); > > - struct drm_i915_private *dev_priv = to_i915(connector->base.dev); > > + struct drm_i915_private *i915 = to_i915(connector->base.dev); > > struct intel_panel *panel = &connector->panel; > > u32 hw_level; > > if (!panel->backlight.present) > > return; > > - mutex_lock(&dev_priv->backlight_lock); > > + mutex_lock(&i915->backlight_lock); > > WARN_ON(panel->backlight.max == 0); > > @@ -1245,7 +1245,7 @@ static void intel_panel_set_backlight(const struct drm_connector_state *conn_sta > > if (panel->backlight.enabled) > > intel_panel_actually_set_backlight(conn_state, hw_level); > > - mutex_unlock(&dev_priv->backlight_lock); > > + mutex_unlock(&i915->backlight_lock); > > } > > static int intel_backlight_device_update_status(struct backlight_device *bd) > > @@ -1284,11 +1284,11 @@ static int intel_backlight_device_get_brightness(struct backlight_device *bd) > > { > > struct intel_connector *connector = bl_get_data(bd); > > struct drm_device *dev = connector->base.dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > intel_wakeref_t wakeref; > > int ret = 0; > > - with_intel_runtime_pm(dev_priv, wakeref) { > > + with_intel_runtime_pm(i915, wakeref) { > > u32 hw_level; > > drm_modeset_lock(&dev->mode_config.connection_mutex, NULL); > > @@ -1378,9 +1378,9 @@ void intel_backlight_device_unregister(struct intel_connector *connector) > > */ > > static u32 cnp_hz_to_pwm(struct intel_connector *connector, u32 pwm_freq_hz) > > { > > - struct drm_i915_private *dev_priv = to_i915(connector->base.dev); > > + struct drm_i915_private *i915 = to_i915(connector->base.dev); > > - return DIV_ROUND_CLOSEST(KHz(dev_priv->rawclk_freq), pwm_freq_hz); > > + return DIV_ROUND_CLOSEST(KHz(i915->rawclk_freq), pwm_freq_hz); > > } > > /* > > @@ -1416,7 +1416,7 @@ static u32 spt_hz_to_pwm(struct intel_connector *connector, u32 pwm_freq_hz) > > */ > > static u32 lpt_hz_to_pwm(struct intel_connector *connector, u32 pwm_freq_hz) > > { > > - struct drm_i915_private *dev_priv = to_i915(connector->base.dev); > > + struct drm_i915_private *i915 = to_i915(connector->base.dev); > > struct intel_panel *panel = &connector->panel; > > u32 mul, clock; > > @@ -1425,7 +1425,7 @@ static u32 lpt_hz_to_pwm(struct intel_connector *connector, u32 pwm_freq_hz) > > else > > mul = 128; > > - if (HAS_PCH_LPT_H(dev_priv)) > > + if (HAS_PCH_LPT_H(i915)) > > clock = MHz(135); /* LPT:H */ > > else > > clock = MHz(24); /* LPT:LP */ > > @@ -1439,9 +1439,9 @@ static u32 lpt_hz_to_pwm(struct intel_connector *connector, u32 pwm_freq_hz) > > */ > > static u32 pch_hz_to_pwm(struct intel_connector *connector, u32 pwm_freq_hz) > > { > > - struct drm_i915_private *dev_priv = to_i915(connector->base.dev); > > + struct drm_i915_private *i915 = to_i915(connector->base.dev); > > - return DIV_ROUND_CLOSEST(KHz(dev_priv->rawclk_freq), pwm_freq_hz * 128); > > + return DIV_ROUND_CLOSEST(KHz(i915->rawclk_freq), pwm_freq_hz * 128); > > } > > /* > > @@ -1454,13 +1454,13 @@ static u32 pch_hz_to_pwm(struct intel_connector *connector, u32 pwm_freq_hz) > > */ > > static u32 i9xx_hz_to_pwm(struct intel_connector *connector, u32 pwm_freq_hz) > > { > > - struct drm_i915_private *dev_priv = to_i915(connector->base.dev); > > + struct drm_i915_private *i915 = to_i915(connector->base.dev); > > int clock; > > - if (IS_PINEVIEW(dev_priv)) > > - clock = KHz(dev_priv->rawclk_freq); > > + if (IS_PINEVIEW(i915)) > > + clock = KHz(i915->rawclk_freq); > > else > > - clock = KHz(dev_priv->cdclk.hw.cdclk); > > + clock = KHz(i915->cdclk.hw.cdclk); > > return DIV_ROUND_CLOSEST(clock, pwm_freq_hz * 32); > > } > > @@ -1472,13 +1472,13 @@ static u32 i9xx_hz_to_pwm(struct intel_connector *connector, u32 pwm_freq_hz) > > */ > > static u32 i965_hz_to_pwm(struct intel_connector *connector, u32 pwm_freq_hz) > > { > > - struct drm_i915_private *dev_priv = to_i915(connector->base.dev); > > + struct drm_i915_private *i915 = to_i915(connector->base.dev); > > int clock; > > - if (IS_G4X(dev_priv)) > > - clock = KHz(dev_priv->rawclk_freq); > > + if (IS_G4X(i915)) > > + clock = KHz(i915->rawclk_freq); > > else > > - clock = KHz(dev_priv->cdclk.hw.cdclk); > > + clock = KHz(i915->cdclk.hw.cdclk); > > return DIV_ROUND_CLOSEST(clock, pwm_freq_hz * 128); > > } > > @@ -1490,17 +1490,17 @@ static u32 i965_hz_to_pwm(struct intel_connector *connector, u32 pwm_freq_hz) > > */ > > static u32 vlv_hz_to_pwm(struct intel_connector *connector, u32 pwm_freq_hz) > > { > > - struct drm_i915_private *dev_priv = to_i915(connector->base.dev); > > + struct drm_i915_private *i915 = to_i915(connector->base.dev); > > int mul, clock; > > if ((I915_READ(CBR1_VLV) & CBR_PWM_CLOCK_MUX_SELECT) == 0) { > > - if (IS_CHERRYVIEW(dev_priv)) > > + if (IS_CHERRYVIEW(i915)) > > clock = KHz(19200); > > else > > clock = MHz(25); > > mul = 16; > > } else { > > - clock = KHz(dev_priv->rawclk_freq); > > + clock = KHz(i915->rawclk_freq); > > mul = 128; > > } > > @@ -1509,9 +1509,9 @@ static u32 vlv_hz_to_pwm(struct intel_connector *connector, u32 pwm_freq_hz) > > static u32 get_backlight_max_vbt(struct intel_connector *connector) > > { > > - struct drm_i915_private *dev_priv = to_i915(connector->base.dev); > > + struct drm_i915_private *i915 = to_i915(connector->base.dev); > > struct intel_panel *panel = &connector->panel; > > - u16 pwm_freq_hz = dev_priv->vbt.backlight.pwm_freq_hz; > > + u16 pwm_freq_hz = i915->vbt.backlight.pwm_freq_hz; > > u32 pwm; > > if (!panel->backlight.hz_to_pwm) { > > @@ -1542,7 +1542,7 @@ static u32 get_backlight_max_vbt(struct intel_connector *connector) > > */ > > static u32 get_backlight_min_vbt(struct intel_connector *connector) > > { > > - struct drm_i915_private *dev_priv = to_i915(connector->base.dev); > > + struct drm_i915_private *i915 = to_i915(connector->base.dev); > > struct intel_panel *panel = &connector->panel; > > int min; > > @@ -1555,10 +1555,10 @@ static u32 get_backlight_min_vbt(struct intel_connector *connector) > > * against this by letting the minimum be at most (arbitrarily chosen) > > * 25% of the max. > > */ > > - min = clamp_t(int, dev_priv->vbt.backlight.min_brightness, 0, 64); > > - if (min != dev_priv->vbt.backlight.min_brightness) { > > + min = clamp_t(int, i915->vbt.backlight.min_brightness, 0, 64); > > + if (min != i915->vbt.backlight.min_brightness) { > > DRM_DEBUG_KMS("clamping VBT min backlight %d/255 to %d/255\n", > > - dev_priv->vbt.backlight.min_brightness, min); > > + i915->vbt.backlight.min_brightness, min); > > } > > /* vbt value is a coefficient in range [0..255] */ > > @@ -1567,12 +1567,12 @@ static u32 get_backlight_min_vbt(struct intel_connector *connector) > > static int lpt_setup_backlight(struct intel_connector *connector, enum pipe unused) > > { > > - struct drm_i915_private *dev_priv = to_i915(connector->base.dev); > > + struct drm_i915_private *i915 = to_i915(connector->base.dev); > > struct intel_panel *panel = &connector->panel; > > u32 cpu_ctl2, pch_ctl1, pch_ctl2, val; > > bool alt, cpu_mode; > > - if (HAS_PCH_LPT(dev_priv)) > > + if (HAS_PCH_LPT(i915)) > > alt = I915_READ(SOUTH_CHICKEN2) & LPT_PWM_GRANULARITY; > > else > > alt = I915_READ(SOUTH_CHICKEN1) & SPT_PWM_GRANULARITY; > > @@ -1596,7 +1596,7 @@ static int lpt_setup_backlight(struct intel_connector *connector, enum pipe unus > > panel->backlight.enabled = pch_ctl1 & BLM_PCH_PWM_ENABLE; > > - cpu_mode = panel->backlight.enabled && HAS_PCH_LPT(dev_priv) && > > + cpu_mode = panel->backlight.enabled && HAS_PCH_LPT(i915) && > > !(pch_ctl1 & BLM_PCH_OVERRIDE_ENABLE) && > > (cpu_ctl2 & BLM_PWM_ENABLE); > > if (cpu_mode) > > @@ -1622,7 +1622,7 @@ static int lpt_setup_backlight(struct intel_connector *connector, enum pipe unus > > static int pch_setup_backlight(struct intel_connector *connector, enum pipe unused) > > { > > - struct drm_i915_private *dev_priv = to_i915(connector->base.dev); > > + struct drm_i915_private *i915 = to_i915(connector->base.dev); > > struct intel_panel *panel = &connector->panel; > > u32 cpu_ctl2, pch_ctl1, pch_ctl2, val; > > @@ -1654,16 +1654,16 @@ static int pch_setup_backlight(struct intel_connector *connector, enum pipe unus > > static int i9xx_setup_backlight(struct intel_connector *connector, enum pipe unused) > > { > > - struct drm_i915_private *dev_priv = to_i915(connector->base.dev); > > + struct drm_i915_private *i915 = to_i915(connector->base.dev); > > struct intel_panel *panel = &connector->panel; > > u32 ctl, val; > > ctl = I915_READ(BLC_PWM_CTL); > > - if (IS_GEN(dev_priv, 2) || IS_I915GM(dev_priv) || IS_I945GM(dev_priv)) > > + if (IS_GEN(i915, 2) || IS_I915GM(i915) || IS_I945GM(i915)) > > panel->backlight.combination_mode = ctl & BLM_LEGACY_MODE; > > - if (IS_PINEVIEW(dev_priv)) > > + if (IS_PINEVIEW(i915)) > > panel->backlight.active_low_pwm = ctl & BLM_POLARITY_PNV; > > panel->backlight.max = ctl >> 17; > > @@ -1693,7 +1693,7 @@ static int i9xx_setup_backlight(struct intel_connector *connector, enum pipe unu > > static int i965_setup_backlight(struct intel_connector *connector, enum pipe unused) > > { > > - struct drm_i915_private *dev_priv = to_i915(connector->base.dev); > > + struct drm_i915_private *i915 = to_i915(connector->base.dev); > > struct intel_panel *panel = &connector->panel; > > u32 ctl, ctl2, val; > > @@ -1727,7 +1727,7 @@ static int i965_setup_backlight(struct intel_connector *connector, enum pipe unu > > static int vlv_setup_backlight(struct intel_connector *connector, enum pipe pipe) > > { > > - struct drm_i915_private *dev_priv = to_i915(connector->base.dev); > > + struct drm_i915_private *i915 = to_i915(connector->base.dev); > > struct intel_panel *panel = &connector->panel; > > u32 ctl, ctl2, val; > > @@ -1748,7 +1748,7 @@ static int vlv_setup_backlight(struct intel_connector *connector, enum pipe pipe > > panel->backlight.min = get_backlight_min_vbt(connector); > > - val = _vlv_get_backlight(dev_priv, pipe); > > + val = _vlv_get_backlight(i915, pipe); > > val = intel_panel_compute_brightness(connector, val); > > panel->backlight.level = clamp(val, panel->backlight.min, > > panel->backlight.max); > > @@ -1761,11 +1761,11 @@ static int vlv_setup_backlight(struct intel_connector *connector, enum pipe pipe > > static int > > bxt_setup_backlight(struct intel_connector *connector, enum pipe unused) > > { > > - struct drm_i915_private *dev_priv = to_i915(connector->base.dev); > > + struct drm_i915_private *i915 = to_i915(connector->base.dev); > > struct intel_panel *panel = &connector->panel; > > u32 pwm_ctl, val; > > - panel->backlight.controller = dev_priv->vbt.backlight.controller; > > + panel->backlight.controller = i915->vbt.backlight.controller; > > pwm_ctl = I915_READ(BXT_BLC_PWM_CTL(panel->backlight.controller)); > > @@ -1801,7 +1801,7 @@ bxt_setup_backlight(struct intel_connector *connector, enum pipe unused) > > static int > > cnp_setup_backlight(struct intel_connector *connector, enum pipe unused) > > { > > - struct drm_i915_private *dev_priv = to_i915(connector->base.dev); > > + struct drm_i915_private *i915 = to_i915(connector->base.dev); > > struct intel_panel *panel = &connector->panel; > > u32 pwm_ctl, val; > > @@ -1881,28 +1881,28 @@ void intel_panel_update_backlight(struct intel_encoder *encoder, > > const struct drm_connector_state *conn_state) > > { > > struct intel_connector *connector = to_intel_connector(conn_state->connector); > > - struct drm_i915_private *dev_priv = to_i915(connector->base.dev); > > + struct drm_i915_private *i915 = to_i915(connector->base.dev); > > struct intel_panel *panel = &connector->panel; > > if (!panel->backlight.present) > > return; > > - mutex_lock(&dev_priv->backlight_lock); > > + mutex_lock(&i915->backlight_lock); > > if (!panel->backlight.enabled) > > __intel_panel_enable_backlight(crtc_state, conn_state); > > - mutex_unlock(&dev_priv->backlight_lock); > > + mutex_unlock(&i915->backlight_lock); > > } > > int intel_panel_setup_backlight(struct drm_connector *connector, enum pipe pipe) > > { > > - struct drm_i915_private *dev_priv = to_i915(connector->dev); > > + struct drm_i915_private *i915 = to_i915(connector->dev); > > struct intel_connector *intel_connector = to_intel_connector(connector); > > struct intel_panel *panel = &intel_connector->panel; > > int ret; > > - if (!dev_priv->vbt.backlight.present) { > > - if (dev_priv->quirks & QUIRK_BACKLIGHT_PRESENT) { > > + if (!i915->vbt.backlight.present) { > > + if (i915->quirks & QUIRK_BACKLIGHT_PRESENT) { > > DRM_DEBUG_KMS("no backlight present per VBT, but present per quirk\n"); > > } else { > > DRM_DEBUG_KMS("no backlight present per VBT\n"); > > @@ -1915,9 +1915,9 @@ int intel_panel_setup_backlight(struct drm_connector *connector, enum pipe pipe) > > return -ENODEV; > > /* set level and max in panel struct */ > > - mutex_lock(&dev_priv->backlight_lock); > > + mutex_lock(&i915->backlight_lock); > > ret = panel->backlight.setup(intel_connector, pipe); > > - mutex_unlock(&dev_priv->backlight_lock); > > + mutex_unlock(&i915->backlight_lock); > > if (ret) { > > DRM_DEBUG_KMS("failed to setup backlight for connector %s\n", > > @@ -1950,7 +1950,7 @@ intel_panel_init_backlight_funcs(struct intel_panel *panel) > > { > > struct intel_connector *connector = > > container_of(panel, struct intel_connector, panel); > > - struct drm_i915_private *dev_priv = to_i915(connector->base.dev); > > + struct drm_i915_private *i915 = to_i915(connector->base.dev); > > if (connector->base.connector_type == DRM_MODE_CONNECTOR_eDP && > > intel_dp_aux_init_backlight_funcs(connector) == 0) > > @@ -1960,38 +1960,38 @@ intel_panel_init_backlight_funcs(struct intel_panel *panel) > > intel_dsi_dcs_init_backlight_funcs(connector) == 0) > > return; > > - if (IS_GEN9_LP(dev_priv)) { > > + if (IS_GEN9_LP(i915)) { > > panel->backlight.setup = bxt_setup_backlight; > > panel->backlight.enable = bxt_enable_backlight; > > panel->backlight.disable = bxt_disable_backlight; > > panel->backlight.set = bxt_set_backlight; > > panel->backlight.get = bxt_get_backlight; > > panel->backlight.hz_to_pwm = bxt_hz_to_pwm; > > - } else if (INTEL_PCH_TYPE(dev_priv) >= PCH_CNP) { > > + } else if (INTEL_PCH_TYPE(i915) >= PCH_CNP) { > > panel->backlight.setup = cnp_setup_backlight; > > panel->backlight.enable = cnp_enable_backlight; > > panel->backlight.disable = cnp_disable_backlight; > > panel->backlight.set = bxt_set_backlight; > > panel->backlight.get = bxt_get_backlight; > > panel->backlight.hz_to_pwm = cnp_hz_to_pwm; > > - } else if (INTEL_PCH_TYPE(dev_priv) >= PCH_LPT) { > > + } else if (INTEL_PCH_TYPE(i915) >= PCH_LPT) { > > panel->backlight.setup = lpt_setup_backlight; > > panel->backlight.enable = lpt_enable_backlight; > > panel->backlight.disable = lpt_disable_backlight; > > panel->backlight.set = lpt_set_backlight; > > panel->backlight.get = lpt_get_backlight; > > - if (HAS_PCH_LPT(dev_priv)) > > + if (HAS_PCH_LPT(i915)) > > panel->backlight.hz_to_pwm = lpt_hz_to_pwm; > > else > > panel->backlight.hz_to_pwm = spt_hz_to_pwm; > > - } else if (HAS_PCH_SPLIT(dev_priv)) { > > + } else if (HAS_PCH_SPLIT(i915)) { > > panel->backlight.setup = pch_setup_backlight; > > panel->backlight.enable = pch_enable_backlight; > > panel->backlight.disable = pch_disable_backlight; > > panel->backlight.set = pch_set_backlight; > > panel->backlight.get = pch_get_backlight; > > panel->backlight.hz_to_pwm = pch_hz_to_pwm; > > - } else if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) { > > + } else if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) { > > if (connector->base.connector_type == DRM_MODE_CONNECTOR_DSI) { > > panel->backlight.setup = pwm_setup_backlight; > > panel->backlight.enable = pwm_enable_backlight; > > @@ -2006,7 +2006,7 @@ intel_panel_init_backlight_funcs(struct intel_panel *panel) > > panel->backlight.get = vlv_get_backlight; > > panel->backlight.hz_to_pwm = vlv_hz_to_pwm; > > } > > - } else if (IS_GEN(dev_priv, 4)) { > > + } else if (IS_GEN(i915, 4)) { > > panel->backlight.setup = i965_setup_backlight; > > panel->backlight.enable = i965_enable_backlight; > > panel->backlight.disable = i965_disable_backlight; > > diff --git a/drivers/gpu/drm/i915/intel_pipe_crc.c b/drivers/gpu/drm/i915/intel_pipe_crc.c > > index 1e2c4307d05a..4a4409f4076f 100644 > > --- a/drivers/gpu/drm/i915/intel_pipe_crc.c > > +++ b/drivers/gpu/drm/i915/intel_pipe_crc.c > > @@ -70,11 +70,11 @@ static int i8xx_pipe_crc_ctl_reg(enum intel_pipe_crc_source *source, > > return 0; > > } > > -static int i9xx_pipe_crc_auto_source(struct drm_i915_private *dev_priv, > > +static int i9xx_pipe_crc_auto_source(struct drm_i915_private *i915, > > enum pipe pipe, > > enum intel_pipe_crc_source *source) > > { > > - struct drm_device *dev = &dev_priv->drm; > > + struct drm_device *dev = &i915->drm; > > struct intel_encoder *encoder; > > struct intel_crtc *crtc; > > struct intel_digital_port *dig_port; > > @@ -124,7 +124,7 @@ static int i9xx_pipe_crc_auto_source(struct drm_i915_private *dev_priv, > > return ret; > > } > > -static int vlv_pipe_crc_ctl_reg(struct drm_i915_private *dev_priv, > > +static int vlv_pipe_crc_ctl_reg(struct drm_i915_private *i915, > > enum pipe pipe, > > enum intel_pipe_crc_source *source, > > u32 *val) > > @@ -132,7 +132,7 @@ static int vlv_pipe_crc_ctl_reg(struct drm_i915_private *dev_priv, > > bool need_stable_symbols = false; > > if (*source == INTEL_PIPE_CRC_SOURCE_AUTO) { > > - int ret = i9xx_pipe_crc_auto_source(dev_priv, pipe, source); > > + int ret = i9xx_pipe_crc_auto_source(i915, pipe, source); > > if (ret) > > return ret; > > } > > @@ -150,7 +150,7 @@ static int vlv_pipe_crc_ctl_reg(struct drm_i915_private *dev_priv, > > need_stable_symbols = true; > > break; > > case INTEL_PIPE_CRC_SOURCE_DP_D: > > - if (!IS_CHERRYVIEW(dev_priv)) > > + if (!IS_CHERRYVIEW(i915)) > > return -EINVAL; > > *val = PIPE_CRC_ENABLE | PIPE_CRC_SOURCE_DP_D_VLV; > > need_stable_symbols = true; > > @@ -194,13 +194,13 @@ static int vlv_pipe_crc_ctl_reg(struct drm_i915_private *dev_priv, > > return 0; > > } > > -static int i9xx_pipe_crc_ctl_reg(struct drm_i915_private *dev_priv, > > +static int i9xx_pipe_crc_ctl_reg(struct drm_i915_private *i915, > > enum pipe pipe, > > enum intel_pipe_crc_source *source, > > u32 *val) > > { > > if (*source == INTEL_PIPE_CRC_SOURCE_AUTO) { > > - int ret = i9xx_pipe_crc_auto_source(dev_priv, pipe, source); > > + int ret = i9xx_pipe_crc_auto_source(i915, pipe, source); > > if (ret) > > return ret; > > } > > @@ -210,7 +210,7 @@ static int i9xx_pipe_crc_ctl_reg(struct drm_i915_private *dev_priv, > > *val = PIPE_CRC_ENABLE | PIPE_CRC_SOURCE_PIPE_I9XX; > > break; > > case INTEL_PIPE_CRC_SOURCE_TV: > > - if (!SUPPORTS_TV(dev_priv)) > > + if (!SUPPORTS_TV(i915)) > > return -EINVAL; > > *val = PIPE_CRC_ENABLE | PIPE_CRC_SOURCE_TV_PRE; > > break; > > @@ -234,7 +234,7 @@ static int i9xx_pipe_crc_ctl_reg(struct drm_i915_private *dev_priv, > > return 0; > > } > > -static void vlv_undo_pipe_scramble_reset(struct drm_i915_private *dev_priv, > > +static void vlv_undo_pipe_scramble_reset(struct drm_i915_private *i915, > > enum pipe pipe) > > { > > u32 tmp = I915_READ(PORT_DFT2_G4X); > > @@ -286,7 +286,7 @@ static int ilk_pipe_crc_ctl_reg(enum intel_pipe_crc_source *source, > > static void > > intel_crtc_crc_setup_workarounds(struct intel_crtc *crtc, bool enable) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > struct intel_crtc_state *pipe_config; > > struct drm_atomic_state *state; > > struct drm_modeset_acquire_ctx ctx; > > @@ -294,7 +294,7 @@ intel_crtc_crc_setup_workarounds(struct intel_crtc *crtc, bool enable) > > drm_modeset_acquire_init(&ctx, 0); > > - state = drm_atomic_state_alloc(&dev_priv->drm); > > + state = drm_atomic_state_alloc(&i915->drm); > > if (!state) { > > ret = -ENOMEM; > > goto unlock; > > @@ -312,7 +312,7 @@ intel_crtc_crc_setup_workarounds(struct intel_crtc *crtc, bool enable) > > pipe_config->base.mode_changed = pipe_config->has_psr; > > pipe_config->crc_enabled = enable; > > - if (IS_HASWELL(dev_priv) && > > + if (IS_HASWELL(i915) && > > pipe_config->base.active && crtc->pipe == PIPE_A && > > pipe_config->cpu_transcoder == TRANSCODER_EDP) > > pipe_config->base.mode_changed = true; > > @@ -333,7 +333,7 @@ intel_crtc_crc_setup_workarounds(struct intel_crtc *crtc, bool enable) > > drm_modeset_acquire_fini(&ctx); > > } > > -static int ivb_pipe_crc_ctl_reg(struct drm_i915_private *dev_priv, > > +static int ivb_pipe_crc_ctl_reg(struct drm_i915_private *i915, > > enum pipe pipe, > > enum intel_pipe_crc_source *source, > > u32 *val) > > @@ -361,7 +361,7 @@ static int ivb_pipe_crc_ctl_reg(struct drm_i915_private *dev_priv, > > return 0; > > } > > -static int skl_pipe_crc_ctl_reg(struct drm_i915_private *dev_priv, > > +static int skl_pipe_crc_ctl_reg(struct drm_i915_private *i915, > > enum pipe pipe, > > enum intel_pipe_crc_source *source, > > u32 *val) > > @@ -404,22 +404,22 @@ static int skl_pipe_crc_ctl_reg(struct drm_i915_private *dev_priv, > > return 0; > > } > > -static int get_new_crc_ctl_reg(struct drm_i915_private *dev_priv, > > +static int get_new_crc_ctl_reg(struct drm_i915_private *i915, > > enum pipe pipe, > > enum intel_pipe_crc_source *source, u32 *val) > > { > > - if (IS_GEN(dev_priv, 2)) > > + if (IS_GEN(i915, 2)) > > return i8xx_pipe_crc_ctl_reg(source, val); > > - else if (INTEL_GEN(dev_priv) < 5) > > - return i9xx_pipe_crc_ctl_reg(dev_priv, pipe, source, val); > > - else if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) > > - return vlv_pipe_crc_ctl_reg(dev_priv, pipe, source, val); > > - else if (IS_GEN_RANGE(dev_priv, 5, 6)) > > + else if (INTEL_GEN(i915) < 5) > > + return i9xx_pipe_crc_ctl_reg(i915, pipe, source, val); > > + else if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) > > + return vlv_pipe_crc_ctl_reg(i915, pipe, source, val); > > + else if (IS_GEN_RANGE(i915, 5, 6)) > > return ilk_pipe_crc_ctl_reg(source, val); > > - else if (INTEL_GEN(dev_priv) < 9) > > - return ivb_pipe_crc_ctl_reg(dev_priv, pipe, source, val); > > + else if (INTEL_GEN(i915) < 9) > > + return ivb_pipe_crc_ctl_reg(i915, pipe, source, val); > > else > > - return skl_pipe_crc_ctl_reg(dev_priv, pipe, source, val); > > + return skl_pipe_crc_ctl_reg(i915, pipe, source, val); > > } > > static int > > @@ -440,18 +440,18 @@ display_crc_ctl_parse_source(const char *buf, enum intel_pipe_crc_source *s) > > return 0; > > } > > -void intel_display_crc_init(struct drm_i915_private *dev_priv) > > +void intel_display_crc_init(struct drm_i915_private *i915) > > { > > enum pipe pipe; > > - for_each_pipe(dev_priv, pipe) { > > - struct intel_pipe_crc *pipe_crc = &dev_priv->pipe_crc[pipe]; > > + for_each_pipe(i915, pipe) { > > + struct intel_pipe_crc *pipe_crc = &i915->pipe_crc[pipe]; > > spin_lock_init(&pipe_crc->lock); > > } > > } > > -static int i8xx_crc_source_valid(struct drm_i915_private *dev_priv, > > +static int i8xx_crc_source_valid(struct drm_i915_private *i915, > > const enum intel_pipe_crc_source source) > > { > > switch (source) { > > @@ -463,7 +463,7 @@ static int i8xx_crc_source_valid(struct drm_i915_private *dev_priv, > > } > > } > > -static int i9xx_crc_source_valid(struct drm_i915_private *dev_priv, > > +static int i9xx_crc_source_valid(struct drm_i915_private *i915, > > const enum intel_pipe_crc_source source) > > { > > switch (source) { > > @@ -476,7 +476,7 @@ static int i9xx_crc_source_valid(struct drm_i915_private *dev_priv, > > } > > } > > -static int vlv_crc_source_valid(struct drm_i915_private *dev_priv, > > +static int vlv_crc_source_valid(struct drm_i915_private *i915, > > const enum intel_pipe_crc_source source) > > { > > switch (source) { > > @@ -491,7 +491,7 @@ static int vlv_crc_source_valid(struct drm_i915_private *dev_priv, > > } > > } > > -static int ilk_crc_source_valid(struct drm_i915_private *dev_priv, > > +static int ilk_crc_source_valid(struct drm_i915_private *i915, > > const enum intel_pipe_crc_source source) > > { > > switch (source) { > > @@ -505,7 +505,7 @@ static int ilk_crc_source_valid(struct drm_i915_private *dev_priv, > > } > > } > > -static int ivb_crc_source_valid(struct drm_i915_private *dev_priv, > > +static int ivb_crc_source_valid(struct drm_i915_private *i915, > > const enum intel_pipe_crc_source source) > > { > > switch (source) { > > @@ -519,7 +519,7 @@ static int ivb_crc_source_valid(struct drm_i915_private *dev_priv, > > } > > } > > -static int skl_crc_source_valid(struct drm_i915_private *dev_priv, > > +static int skl_crc_source_valid(struct drm_i915_private *i915, > > const enum intel_pipe_crc_source source) > > { > > switch (source) { > > @@ -539,21 +539,21 @@ static int skl_crc_source_valid(struct drm_i915_private *dev_priv, > > } > > static int > > -intel_is_valid_crc_source(struct drm_i915_private *dev_priv, > > +intel_is_valid_crc_source(struct drm_i915_private *i915, > > const enum intel_pipe_crc_source source) > > { > > - if (IS_GEN(dev_priv, 2)) > > - return i8xx_crc_source_valid(dev_priv, source); > > - else if (INTEL_GEN(dev_priv) < 5) > > - return i9xx_crc_source_valid(dev_priv, source); > > - else if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) > > - return vlv_crc_source_valid(dev_priv, source); > > - else if (IS_GEN_RANGE(dev_priv, 5, 6)) > > - return ilk_crc_source_valid(dev_priv, source); > > - else if (INTEL_GEN(dev_priv) < 9) > > - return ivb_crc_source_valid(dev_priv, source); > > + if (IS_GEN(i915, 2)) > > + return i8xx_crc_source_valid(i915, source); > > + else if (INTEL_GEN(i915) < 5) > > + return i9xx_crc_source_valid(i915, source); > > + else if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) > > + return vlv_crc_source_valid(i915, source); > > + else if (IS_GEN_RANGE(i915, 5, 6)) > > + return ilk_crc_source_valid(i915, source); > > + else if (INTEL_GEN(i915) < 9) > > + return ivb_crc_source_valid(i915, source); > > else > > - return skl_crc_source_valid(dev_priv, source); > > + return skl_crc_source_valid(i915, source); > > } > > const char *const *intel_crtc_get_crc_sources(struct drm_crtc *crtc, > > @@ -566,7 +566,7 @@ const char *const *intel_crtc_get_crc_sources(struct drm_crtc *crtc, > > int intel_crtc_verify_crc_source(struct drm_crtc *crtc, const char *source_name, > > size_t *values_cnt) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc->dev); > > + struct drm_i915_private *i915 = to_i915(crtc->dev); > > enum intel_pipe_crc_source source; > > if (display_crc_ctl_parse_source(source_name, &source) < 0) { > > @@ -575,7 +575,7 @@ int intel_crtc_verify_crc_source(struct drm_crtc *crtc, const char *source_name, > > } > > if (source == INTEL_PIPE_CRC_SOURCE_AUTO || > > - intel_is_valid_crc_source(dev_priv, source) == 0) { > > + intel_is_valid_crc_source(i915, source) == 0) { > > *values_cnt = 5; > > return 0; > > } > > @@ -585,8 +585,8 @@ int intel_crtc_verify_crc_source(struct drm_crtc *crtc, const char *source_name, > > int intel_crtc_set_crc_source(struct drm_crtc *crtc, const char *source_name) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc->dev); > > - struct intel_pipe_crc *pipe_crc = &dev_priv->pipe_crc[crtc->index]; > > + struct drm_i915_private *i915 = to_i915(crtc->dev); > > + struct intel_pipe_crc *pipe_crc = &i915->pipe_crc[crtc->index]; > > enum intel_display_power_domain power_domain; > > enum intel_pipe_crc_source source; > > intel_wakeref_t wakeref; > > @@ -600,7 +600,7 @@ int intel_crtc_set_crc_source(struct drm_crtc *crtc, const char *source_name) > > } > > power_domain = POWER_DOMAIN_PIPE(crtc->index); > > - wakeref = intel_display_power_get_if_enabled(dev_priv, power_domain); > > + wakeref = intel_display_power_get_if_enabled(i915, power_domain); > > if (!wakeref) { > > DRM_DEBUG_KMS("Trying to capture CRC while pipe is off\n"); > > return -EIO; > > @@ -610,7 +610,7 @@ int intel_crtc_set_crc_source(struct drm_crtc *crtc, const char *source_name) > > if (enable) > > intel_crtc_crc_setup_workarounds(to_intel_crtc(crtc), true); > > - ret = get_new_crc_ctl_reg(dev_priv, crtc->index, &source, &val); > > + ret = get_new_crc_ctl_reg(i915, crtc->index, &source, &val); > > if (ret != 0) > > goto out; > > @@ -619,8 +619,8 @@ int intel_crtc_set_crc_source(struct drm_crtc *crtc, const char *source_name) > > POSTING_READ(PIPE_CRC_CTL(crtc->index)); > > if (!source) { > > - if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) > > - vlv_undo_pipe_scramble_reset(dev_priv, crtc->index); > > + if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) > > + vlv_undo_pipe_scramble_reset(i915, crtc->index); > > } > > pipe_crc->skipped = 0; > > @@ -629,7 +629,7 @@ int intel_crtc_set_crc_source(struct drm_crtc *crtc, const char *source_name) > > if (!enable) > > intel_crtc_crc_setup_workarounds(to_intel_crtc(crtc), false); > > - intel_display_power_put(dev_priv, power_domain, wakeref); > > + intel_display_power_put(i915, power_domain, wakeref); > > return ret; > > } > > @@ -637,14 +637,14 @@ int intel_crtc_set_crc_source(struct drm_crtc *crtc, const char *source_name) > > void intel_crtc_enable_pipe_crc(struct intel_crtc *intel_crtc) > > { > > struct drm_crtc *crtc = &intel_crtc->base; > > - struct drm_i915_private *dev_priv = to_i915(crtc->dev); > > - struct intel_pipe_crc *pipe_crc = &dev_priv->pipe_crc[crtc->index]; > > + struct drm_i915_private *i915 = to_i915(crtc->dev); > > + struct intel_pipe_crc *pipe_crc = &i915->pipe_crc[crtc->index]; > > u32 val = 0; > > if (!crtc->crc.opened) > > return; > > - if (get_new_crc_ctl_reg(dev_priv, crtc->index, &pipe_crc->source, &val) < 0) > > + if (get_new_crc_ctl_reg(i915, crtc->index, &pipe_crc->source, &val) < 0) > > return; > > /* Don't need pipe_crc->lock here, IRQs are not generated. */ > > @@ -657,8 +657,8 @@ void intel_crtc_enable_pipe_crc(struct intel_crtc *intel_crtc) > > void intel_crtc_disable_pipe_crc(struct intel_crtc *intel_crtc) > > { > > struct drm_crtc *crtc = &intel_crtc->base; > > - struct drm_i915_private *dev_priv = to_i915(crtc->dev); > > - struct intel_pipe_crc *pipe_crc = &dev_priv->pipe_crc[crtc->index]; > > + struct drm_i915_private *i915 = to_i915(crtc->dev); > > + struct intel_pipe_crc *pipe_crc = &i915->pipe_crc[crtc->index]; > > /* Swallow crc's until we stop generating them. */ > > spin_lock_irq(&pipe_crc->lock); > > @@ -667,5 +667,5 @@ void intel_crtc_disable_pipe_crc(struct intel_crtc *intel_crtc) > > I915_WRITE(PIPE_CRC_CTL(crtc->index), 0); > > POSTING_READ(PIPE_CRC_CTL(crtc->index)); > > - synchronize_irq(dev_priv->drm.irq); > > + synchronize_irq(i915->drm.irq); > > } > > diff --git a/drivers/gpu/drm/i915/intel_pipe_crc.h b/drivers/gpu/drm/i915/intel_pipe_crc.h > > index db258a756fc6..42048ab1391a 100644 > > --- a/drivers/gpu/drm/i915/intel_pipe_crc.h > > +++ b/drivers/gpu/drm/i915/intel_pipe_crc.h > > @@ -13,7 +13,7 @@ struct drm_i915_private; > > struct intel_crtc; > > #ifdef CONFIG_DEBUG_FS > > -void intel_display_crc_init(struct drm_i915_private *dev_priv); > > +void intel_display_crc_init(struct drm_i915_private *i915); > > int intel_crtc_set_crc_source(struct drm_crtc *crtc, const char *source_name); > > int intel_crtc_verify_crc_source(struct drm_crtc *crtc, > > const char *source_name, size_t *values_cnt); > > @@ -22,7 +22,7 @@ const char *const *intel_crtc_get_crc_sources(struct drm_crtc *crtc, > > void intel_crtc_disable_pipe_crc(struct intel_crtc *crtc); > > void intel_crtc_enable_pipe_crc(struct intel_crtc *crtc); > > #else > > -static inline void intel_display_crc_init(struct drm_i915_private *dev_priv) {} > > +static inline void intel_display_crc_init(struct drm_i915_private *i915) {} > > #define intel_crtc_set_crc_source NULL > > #define intel_crtc_verify_crc_source NULL > > #define intel_crtc_get_crc_sources NULL > > diff --git a/drivers/gpu/drm/i915/intel_pm.c b/drivers/gpu/drm/i915/intel_pm.c > > index 2c7f3ebc0117..4b939160cd95 100644 > > --- a/drivers/gpu/drm/i915/intel_pm.c > > +++ b/drivers/gpu/drm/i915/intel_pm.c > > @@ -63,9 +63,9 @@ > > * require higher latency to switch to and wake up. > > */ > > -static void gen9_init_clock_gating(struct drm_i915_private *dev_priv) > > +static void gen9_init_clock_gating(struct drm_i915_private *i915) > > { > > - if (HAS_LLC(dev_priv)) { > > + if (HAS_LLC(i915)) { > > /* > > * WaCompressedResourceDisplayNewHashMode:skl,kbl > > * Display WA #0390: skl,kbl > > @@ -96,16 +96,16 @@ static void gen9_init_clock_gating(struct drm_i915_private *dev_priv) > > I915_WRITE(ILK_DPFC_CHICKEN, I915_READ(ILK_DPFC_CHICKEN) | > > ILK_DPFC_DISABLE_DUMMY0); > > - if (IS_SKYLAKE(dev_priv)) { > > + if (IS_SKYLAKE(i915)) { > > /* WaDisableDopClockGating */ > > I915_WRITE(GEN7_MISCCPCTL, I915_READ(GEN7_MISCCPCTL) > > & ~GEN7_DOP_CLOCK_GATE_ENABLE); > > } > > } > > -static void bxt_init_clock_gating(struct drm_i915_private *dev_priv) > > +static void bxt_init_clock_gating(struct drm_i915_private *i915) > > { > > - gen9_init_clock_gating(dev_priv); > > + gen9_init_clock_gating(i915); > > /* WaDisableSDEUnitClockGating:bxt */ > > I915_WRITE(GEN8_UCGCTL6, I915_READ(GEN8_UCGCTL6) | > > @@ -126,9 +126,9 @@ static void bxt_init_clock_gating(struct drm_i915_private *dev_priv) > > PWM1_GATING_DIS | PWM2_GATING_DIS); > > } > > -static void glk_init_clock_gating(struct drm_i915_private *dev_priv) > > +static void glk_init_clock_gating(struct drm_i915_private *i915) > > { > > - gen9_init_clock_gating(dev_priv); > > + gen9_init_clock_gating(i915); > > /* > > * WaDisablePWMClockGating:glk > > @@ -139,7 +139,7 @@ static void glk_init_clock_gating(struct drm_i915_private *dev_priv) > > PWM1_GATING_DIS | PWM2_GATING_DIS); > > /* WaDDIIOTimeout:glk */ > > - if (IS_GLK_REVID(dev_priv, 0, GLK_REVID_A1)) { > > + if (IS_GLK_REVID(i915, 0, GLK_REVID_A1)) { > > u32 val = I915_READ(CHICKEN_MISC_2); > > val &= ~(GLK_CL0_PWR_DOWN | > > GLK_CL1_PWR_DOWN | > > @@ -149,7 +149,7 @@ static void glk_init_clock_gating(struct drm_i915_private *dev_priv) > > } > > -static void i915_pineview_get_mem_freq(struct drm_i915_private *dev_priv) > > +static void i915_pineview_get_mem_freq(struct drm_i915_private *i915) > > { > > u32 tmp; > > @@ -157,100 +157,100 @@ static void i915_pineview_get_mem_freq(struct drm_i915_private *dev_priv) > > switch (tmp & CLKCFG_FSB_MASK) { > > case CLKCFG_FSB_533: > > - dev_priv->fsb_freq = 533; /* 133*4 */ > > + i915->fsb_freq = 533; /* 133*4 */ > > break; > > case CLKCFG_FSB_800: > > - dev_priv->fsb_freq = 800; /* 200*4 */ > > + i915->fsb_freq = 800; /* 200*4 */ > > break; > > case CLKCFG_FSB_667: > > - dev_priv->fsb_freq = 667; /* 167*4 */ > > + i915->fsb_freq = 667; /* 167*4 */ > > break; > > case CLKCFG_FSB_400: > > - dev_priv->fsb_freq = 400; /* 100*4 */ > > + i915->fsb_freq = 400; /* 100*4 */ > > break; > > } > > switch (tmp & CLKCFG_MEM_MASK) { > > case CLKCFG_MEM_533: > > - dev_priv->mem_freq = 533; > > + i915->mem_freq = 533; > > break; > > case CLKCFG_MEM_667: > > - dev_priv->mem_freq = 667; > > + i915->mem_freq = 667; > > break; > > case CLKCFG_MEM_800: > > - dev_priv->mem_freq = 800; > > + i915->mem_freq = 800; > > break; > > } > > /* detect pineview DDR3 setting */ > > tmp = I915_READ(CSHRDDR3CTL); > > - dev_priv->is_ddr3 = (tmp & CSHRDDR3CTL_DDR3) ? 1 : 0; > > + i915->is_ddr3 = (tmp & CSHRDDR3CTL_DDR3) ? 1 : 0; > > } > > -static void i915_ironlake_get_mem_freq(struct drm_i915_private *dev_priv) > > +static void i915_ironlake_get_mem_freq(struct drm_i915_private *i915) > > { > > u16 ddrpll, csipll; > > - ddrpll = intel_uncore_read16(&dev_priv->uncore, DDRMPLL1); > > - csipll = intel_uncore_read16(&dev_priv->uncore, CSIPLL0); > > + ddrpll = intel_uncore_read16(&i915->uncore, DDRMPLL1); > > + csipll = intel_uncore_read16(&i915->uncore, CSIPLL0); > > switch (ddrpll & 0xff) { > > case 0xc: > > - dev_priv->mem_freq = 800; > > + i915->mem_freq = 800; > > break; > > case 0x10: > > - dev_priv->mem_freq = 1066; > > + i915->mem_freq = 1066; > > break; > > case 0x14: > > - dev_priv->mem_freq = 1333; > > + i915->mem_freq = 1333; > > break; > > case 0x18: > > - dev_priv->mem_freq = 1600; > > + i915->mem_freq = 1600; > > break; > > default: > > DRM_DEBUG_DRIVER("unknown memory frequency 0x%02x\n", > > ddrpll & 0xff); > > - dev_priv->mem_freq = 0; > > + i915->mem_freq = 0; > > break; > > } > > - dev_priv->ips.r_t = dev_priv->mem_freq; > > + i915->ips.r_t = i915->mem_freq; > > switch (csipll & 0x3ff) { > > case 0x00c: > > - dev_priv->fsb_freq = 3200; > > + i915->fsb_freq = 3200; > > break; > > case 0x00e: > > - dev_priv->fsb_freq = 3733; > > + i915->fsb_freq = 3733; > > break; > > case 0x010: > > - dev_priv->fsb_freq = 4266; > > + i915->fsb_freq = 4266; > > break; > > case 0x012: > > - dev_priv->fsb_freq = 4800; > > + i915->fsb_freq = 4800; > > break; > > case 0x014: > > - dev_priv->fsb_freq = 5333; > > + i915->fsb_freq = 5333; > > break; > > case 0x016: > > - dev_priv->fsb_freq = 5866; > > + i915->fsb_freq = 5866; > > break; > > case 0x018: > > - dev_priv->fsb_freq = 6400; > > + i915->fsb_freq = 6400; > > break; > > default: > > DRM_DEBUG_DRIVER("unknown fsb frequency 0x%04x\n", > > csipll & 0x3ff); > > - dev_priv->fsb_freq = 0; > > + i915->fsb_freq = 0; > > break; > > } > > - if (dev_priv->fsb_freq == 3200) { > > - dev_priv->ips.c_m = 0; > > - } else if (dev_priv->fsb_freq > 3200 && dev_priv->fsb_freq <= 4800) { > > - dev_priv->ips.c_m = 1; > > + if (i915->fsb_freq == 3200) { > > + i915->ips.c_m = 0; > > + } else if (i915->fsb_freq > 3200 && i915->fsb_freq <= 4800) { > > + i915->ips.c_m = 1; > > } else { > > - dev_priv->ips.c_m = 2; > > + i915->ips.c_m = 2; > > } > > } > > @@ -316,61 +316,61 @@ static const struct cxsr_latency *intel_get_cxsr_latency(bool is_desktop, > > return NULL; > > } > > -static void chv_set_memory_dvfs(struct drm_i915_private *dev_priv, bool enable) > > +static void chv_set_memory_dvfs(struct drm_i915_private *i915, bool enable) > > { > > u32 val; > > - vlv_punit_get(dev_priv); > > + vlv_punit_get(i915); > > - val = vlv_punit_read(dev_priv, PUNIT_REG_DDR_SETUP2); > > + val = vlv_punit_read(i915, PUNIT_REG_DDR_SETUP2); > > if (enable) > > val &= ~FORCE_DDR_HIGH_FREQ; > > else > > val |= FORCE_DDR_HIGH_FREQ; > > val &= ~FORCE_DDR_LOW_FREQ; > > val |= FORCE_DDR_FREQ_REQ_ACK; > > - vlv_punit_write(dev_priv, PUNIT_REG_DDR_SETUP2, val); > > + vlv_punit_write(i915, PUNIT_REG_DDR_SETUP2, val); > > - if (wait_for((vlv_punit_read(dev_priv, PUNIT_REG_DDR_SETUP2) & > > + if (wait_for((vlv_punit_read(i915, PUNIT_REG_DDR_SETUP2) & > > FORCE_DDR_FREQ_REQ_ACK) == 0, 3)) > > DRM_ERROR("timed out waiting for Punit DDR DVFS request\n"); > > - vlv_punit_put(dev_priv); > > + vlv_punit_put(i915); > > } > > -static void chv_set_memory_pm5(struct drm_i915_private *dev_priv, bool enable) > > +static void chv_set_memory_pm5(struct drm_i915_private *i915, bool enable) > > { > > u32 val; > > - vlv_punit_get(dev_priv); > > + vlv_punit_get(i915); > > - val = vlv_punit_read(dev_priv, PUNIT_REG_DSPSSPM); > > + val = vlv_punit_read(i915, PUNIT_REG_DSPSSPM); > > if (enable) > > val |= DSP_MAXFIFO_PM5_ENABLE; > > else > > val &= ~DSP_MAXFIFO_PM5_ENABLE; > > - vlv_punit_write(dev_priv, PUNIT_REG_DSPSSPM, val); > > + vlv_punit_write(i915, PUNIT_REG_DSPSSPM, val); > > - vlv_punit_put(dev_priv); > > + vlv_punit_put(i915); > > } > > #define FW_WM(value, plane) \ > > (((value) << DSPFW_ ## plane ## _SHIFT) & DSPFW_ ## plane ## _MASK) > > -static bool _intel_set_memory_cxsr(struct drm_i915_private *dev_priv, bool enable) > > +static bool _intel_set_memory_cxsr(struct drm_i915_private *i915, bool enable) > > { > > bool was_enabled; > > u32 val; > > - if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) { > > + if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) { > > was_enabled = I915_READ(FW_BLC_SELF_VLV) & FW_CSPWRDWNEN; > > I915_WRITE(FW_BLC_SELF_VLV, enable ? FW_CSPWRDWNEN : 0); > > POSTING_READ(FW_BLC_SELF_VLV); > > - } else if (IS_G4X(dev_priv) || IS_I965GM(dev_priv)) { > > + } else if (IS_G4X(i915) || IS_I965GM(i915)) { > > was_enabled = I915_READ(FW_BLC_SELF) & FW_BLC_SELF_EN; > > I915_WRITE(FW_BLC_SELF, enable ? FW_BLC_SELF_EN : 0); > > POSTING_READ(FW_BLC_SELF); > > - } else if (IS_PINEVIEW(dev_priv)) { > > + } else if (IS_PINEVIEW(i915)) { > > val = I915_READ(DSPFW3); > > was_enabled = val & PINEVIEW_SELF_REFRESH_EN; > > if (enable) > > @@ -379,13 +379,13 @@ static bool _intel_set_memory_cxsr(struct drm_i915_private *dev_priv, bool enabl > > val &= ~PINEVIEW_SELF_REFRESH_EN; > > I915_WRITE(DSPFW3, val); > > POSTING_READ(DSPFW3); > > - } else if (IS_I945G(dev_priv) || IS_I945GM(dev_priv)) { > > + } else if (IS_I945G(i915) || IS_I945GM(i915)) { > > was_enabled = I915_READ(FW_BLC_SELF) & FW_BLC_SELF_EN; > > val = enable ? _MASKED_BIT_ENABLE(FW_BLC_SELF_EN) : > > _MASKED_BIT_DISABLE(FW_BLC_SELF_EN); > > I915_WRITE(FW_BLC_SELF, val); > > POSTING_READ(FW_BLC_SELF); > > - } else if (IS_I915GM(dev_priv)) { > > + } else if (IS_I915GM(i915)) { > > /* > > * FIXME can't find a bit like this for 915G, and > > * and yet it does have the related watermark in > > @@ -400,7 +400,7 @@ static bool _intel_set_memory_cxsr(struct drm_i915_private *dev_priv, bool enabl > > return false; > > } > > - trace_intel_memory_cxsr(dev_priv, was_enabled, enable); > > + trace_intel_memory_cxsr(i915, was_enabled, enable); > > DRM_DEBUG_KMS("memory self-refresh is %s (was %s)\n", > > enableddisabled(enable), > > @@ -411,7 +411,7 @@ static bool _intel_set_memory_cxsr(struct drm_i915_private *dev_priv, bool enabl > > /** > > * intel_set_memory_cxsr - Configure CxSR state > > - * @dev_priv: i915 device > > + * @i915: i915 device > > * @enable: Allow vs. disallow CxSR > > * > > * Allow or disallow the system to enter a special CxSR > > @@ -446,17 +446,17 @@ static bool _intel_set_memory_cxsr(struct drm_i915_private *dev_priv, bool enabl > > * the hardware w.r.t. HPLL SR when writing to plane registers. > > * Disallowing just CxSR is sufficient. > > */ > > -bool intel_set_memory_cxsr(struct drm_i915_private *dev_priv, bool enable) > > +bool intel_set_memory_cxsr(struct drm_i915_private *i915, bool enable) > > { > > bool ret; > > - mutex_lock(&dev_priv->wm.wm_mutex); > > - ret = _intel_set_memory_cxsr(dev_priv, enable); > > - if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) > > - dev_priv->wm.vlv.cxsr = enable; > > - else if (IS_G4X(dev_priv)) > > - dev_priv->wm.g4x.cxsr = enable; > > - mutex_unlock(&dev_priv->wm.wm_mutex); > > + mutex_lock(&i915->wm.wm_mutex); > > + ret = _intel_set_memory_cxsr(i915, enable); > > + if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) > > + i915->wm.vlv.cxsr = enable; > > + else if (IS_G4X(i915)) > > + i915->wm.g4x.cxsr = enable; > > + mutex_unlock(&i915->wm.wm_mutex); > > return ret; > > } > > @@ -483,7 +483,7 @@ static const int pessimal_latency_ns = 5000; > > static void vlv_get_fifo_size(struct intel_crtc_state *crtc_state) > > { > > struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc); > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > struct vlv_fifo_state *fifo_state = &crtc_state->wm.vlv.fifo_state; > > enum pipe pipe = crtc->pipe; > > int sprite0_start, sprite1_start; > > @@ -519,7 +519,7 @@ static void vlv_get_fifo_size(struct intel_crtc_state *crtc_state) > > fifo_state->plane[PLANE_CURSOR] = 63; > > } > > -static int i9xx_get_fifo_size(struct drm_i915_private *dev_priv, > > +static int i9xx_get_fifo_size(struct drm_i915_private *i915, > > enum i9xx_plane_id i9xx_plane) > > { > > u32 dsparb = I915_READ(DSPARB); > > @@ -535,7 +535,7 @@ static int i9xx_get_fifo_size(struct drm_i915_private *dev_priv, > > return size; > > } > > -static int i830_get_fifo_size(struct drm_i915_private *dev_priv, > > +static int i830_get_fifo_size(struct drm_i915_private *i915, > > enum i9xx_plane_id i9xx_plane) > > { > > u32 dsparb = I915_READ(DSPARB); > > @@ -552,7 +552,7 @@ static int i830_get_fifo_size(struct drm_i915_private *dev_priv, > > return size; > > } > > -static int i845_get_fifo_size(struct drm_i915_private *dev_priv, > > +static int i845_get_fifo_size(struct drm_i915_private *i915, > > enum i9xx_plane_id i9xx_plane) > > { > > u32 dsparb = I915_READ(DSPARB); > > @@ -805,9 +805,9 @@ static bool is_enabling(int old, int new, int threshold) > > return old < threshold && new >= threshold; > > } > > -static int intel_wm_num_levels(struct drm_i915_private *dev_priv) > > +static int intel_wm_num_levels(struct drm_i915_private *i915) > > { > > - return dev_priv->wm.max_level + 1; > > + return i915->wm.max_level + 1; > > } > > static bool intel_wm_plane_visible(const struct intel_crtc_state *crtc_state, > > @@ -833,11 +833,11 @@ static bool intel_wm_plane_visible(const struct intel_crtc_state *crtc_state, > > return plane_state->base.visible; > > } > > -static struct intel_crtc *single_enabled_crtc(struct drm_i915_private *dev_priv) > > +static struct intel_crtc *single_enabled_crtc(struct drm_i915_private *i915) > > { > > struct intel_crtc *crtc, *enabled = NULL; > > - for_each_intel_crtc(&dev_priv->drm, crtc) { > > + for_each_intel_crtc(&i915->drm, crtc) { > > if (intel_crtc_active(crtc)) { > > if (enabled) > > return NULL; > > @@ -850,23 +850,23 @@ static struct intel_crtc *single_enabled_crtc(struct drm_i915_private *dev_priv) > > static void pineview_update_wm(struct intel_crtc *unused_crtc) > > { > > - struct drm_i915_private *dev_priv = to_i915(unused_crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(unused_crtc->base.dev); > > struct intel_crtc *crtc; > > const struct cxsr_latency *latency; > > u32 reg; > > unsigned int wm; > > - latency = intel_get_cxsr_latency(!IS_MOBILE(dev_priv), > > - dev_priv->is_ddr3, > > - dev_priv->fsb_freq, > > - dev_priv->mem_freq); > > + latency = intel_get_cxsr_latency(!IS_MOBILE(i915), > > + i915->is_ddr3, > > + i915->fsb_freq, > > + i915->mem_freq); > > if (!latency) { > > DRM_DEBUG_KMS("Unknown FSB/MEM found, disable CxSR\n"); > > - intel_set_memory_cxsr(dev_priv, false); > > + intel_set_memory_cxsr(i915, false); > > return; > > } > > - crtc = single_enabled_crtc(dev_priv); > > + crtc = single_enabled_crtc(i915); > > if (crtc) { > > const struct drm_display_mode *adjusted_mode = > > &crtc->config->base.adjusted_mode; > > @@ -913,9 +913,9 @@ static void pineview_update_wm(struct intel_crtc *unused_crtc) > > I915_WRITE(DSPFW3, reg); > > DRM_DEBUG_KMS("DSPFW3 register is %x\n", reg); > > - intel_set_memory_cxsr(dev_priv, true); > > + intel_set_memory_cxsr(i915, true); > > } else { > > - intel_set_memory_cxsr(dev_priv, false); > > + intel_set_memory_cxsr(i915, false); > > } > > } > > @@ -936,13 +936,13 @@ static unsigned int g4x_tlb_miss_wa(int fifo_size, int width, int cpp) > > return max(0, tlb_miss); > > } > > -static void g4x_write_wm_values(struct drm_i915_private *dev_priv, > > +static void g4x_write_wm_values(struct drm_i915_private *i915, > > const struct g4x_wm_values *wm) > > { > > enum pipe pipe; > > - for_each_pipe(dev_priv, pipe) > > - trace_g4x_wm(intel_get_crtc_for_pipe(dev_priv, pipe), wm); > > + for_each_pipe(i915, pipe) > > + trace_g4x_wm(intel_get_crtc_for_pipe(i915, pipe), wm); > > I915_WRITE(DSPFW1, > > FW_WM(wm->sr.plane, SR) | > > @@ -968,13 +968,13 @@ static void g4x_write_wm_values(struct drm_i915_private *dev_priv, > > #define FW_WM_VLV(value, plane) \ > > (((value) << DSPFW_ ## plane ## _SHIFT) & DSPFW_ ## plane ## _MASK_VLV) > > -static void vlv_write_wm_values(struct drm_i915_private *dev_priv, > > +static void vlv_write_wm_values(struct drm_i915_private *i915, > > const struct vlv_wm_values *wm) > > { > > enum pipe pipe; > > - for_each_pipe(dev_priv, pipe) { > > - trace_vlv_wm(intel_get_crtc_for_pipe(dev_priv, pipe), wm); > > + for_each_pipe(i915, pipe) { > > + trace_vlv_wm(intel_get_crtc_for_pipe(i915, pipe), wm); > > I915_WRITE(VLV_DDL(pipe), > > (wm->ddl[pipe].plane[PLANE_CURSOR] << DDL_CURSOR_SHIFT) | > > @@ -1006,7 +1006,7 @@ static void vlv_write_wm_values(struct drm_i915_private *dev_priv, > > I915_WRITE(DSPFW3, > > FW_WM(wm->sr.cursor, CURSOR_SR)); > > - if (IS_CHERRYVIEW(dev_priv)) { > > + if (IS_CHERRYVIEW(i915)) { > > I915_WRITE(DSPFW7_CHV, > > FW_WM_VLV(wm->pipe[PIPE_B].plane[PLANE_SPRITE1], SPRITED) | > > FW_WM_VLV(wm->pipe[PIPE_B].plane[PLANE_SPRITE0], SPRITEC)); > > @@ -1046,14 +1046,14 @@ static void vlv_write_wm_values(struct drm_i915_private *dev_priv, > > #undef FW_WM_VLV > > -static void g4x_setup_wm_latency(struct drm_i915_private *dev_priv) > > +static void g4x_setup_wm_latency(struct drm_i915_private *i915) > > { > > /* all latencies in usec */ > > - dev_priv->wm.pri_latency[G4X_WM_LEVEL_NORMAL] = 5; > > - dev_priv->wm.pri_latency[G4X_WM_LEVEL_SR] = 12; > > - dev_priv->wm.pri_latency[G4X_WM_LEVEL_HPLL] = 35; > > + i915->wm.pri_latency[G4X_WM_LEVEL_NORMAL] = 5; > > + i915->wm.pri_latency[G4X_WM_LEVEL_SR] = 12; > > + i915->wm.pri_latency[G4X_WM_LEVEL_HPLL] = 35; > > - dev_priv->wm.max_level = G4X_WM_LEVEL_HPLL; > > + i915->wm.max_level = G4X_WM_LEVEL_HPLL; > > } > > static int g4x_plane_fifo_size(enum plane_id plane_id, int level) > > @@ -1103,10 +1103,10 @@ static u16 g4x_compute_wm(const struct intel_crtc_state *crtc_state, > > int level) > > { > > struct intel_plane *plane = to_intel_plane(plane_state->base.plane); > > - struct drm_i915_private *dev_priv = to_i915(plane->base.dev); > > + struct drm_i915_private *i915 = to_i915(plane->base.dev); > > const struct drm_display_mode *adjusted_mode = > > &crtc_state->base.adjusted_mode; > > - unsigned int latency = dev_priv->wm.pri_latency[level] * 10; > > + unsigned int latency = i915->wm.pri_latency[level] * 10; > > unsigned int clock, htotal, cpp, width, wm; > > if (latency == 0) > > @@ -1126,7 +1126,7 @@ static u16 g4x_compute_wm(const struct intel_crtc_state *crtc_state, > > * the HPLL watermark, which seems a little strange. > > * Let's use 32bpp for the HPLL watermark as well. > > */ > > - if (IS_GM45(dev_priv) && plane->id == PLANE_PRIMARY && > > + if (IS_GM45(i915) && plane->id == PLANE_PRIMARY && > > level != G4X_WM_LEVEL_NORMAL) > > cpp = 4; > > else > > @@ -1165,10 +1165,10 @@ static u16 g4x_compute_wm(const struct intel_crtc_state *crtc_state, > > static bool g4x_raw_plane_wm_set(struct intel_crtc_state *crtc_state, > > int level, enum plane_id plane_id, u16 value) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc_state->base.crtc->dev); > > + struct drm_i915_private *i915 = to_i915(crtc_state->base.crtc->dev); > > bool dirty = false; > > - for (; level < intel_wm_num_levels(dev_priv); level++) { > > + for (; level < intel_wm_num_levels(i915); level++) { > > struct g4x_pipe_wm *raw = &crtc_state->wm.g4x.raw[level]; > > dirty |= raw->plane[plane_id] != value; > > @@ -1181,13 +1181,13 @@ static bool g4x_raw_plane_wm_set(struct intel_crtc_state *crtc_state, > > static bool g4x_raw_fbc_wm_set(struct intel_crtc_state *crtc_state, > > int level, u16 value) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc_state->base.crtc->dev); > > + struct drm_i915_private *i915 = to_i915(crtc_state->base.crtc->dev); > > bool dirty = false; > > /* NORMAL level doesn't have an FBC watermark */ > > level = max(level, G4X_WM_LEVEL_SR); > > - for (; level < intel_wm_num_levels(dev_priv); level++) { > > + for (; level < intel_wm_num_levels(i915); level++) { > > struct g4x_pipe_wm *raw = &crtc_state->wm.g4x.raw[level]; > > dirty |= raw->fbc != value; > > @@ -1283,9 +1283,9 @@ static bool g4x_raw_plane_wm_is_valid(const struct intel_crtc_state *crtc_state, > > static bool g4x_raw_crtc_wm_is_valid(const struct intel_crtc_state *crtc_state, > > int level) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc_state->base.crtc->dev); > > + struct drm_i915_private *i915 = to_i915(crtc_state->base.crtc->dev); > > - if (level > dev_priv->wm.max_level) > > + if (level > i915->wm.max_level) > > return false; > > return g4x_raw_plane_wm_is_valid(crtc_state, PLANE_PRIMARY, level) && > > @@ -1484,7 +1484,7 @@ static int g4x_compute_intermediate_wm(struct intel_crtc_state *new_crtc_state) > > return 0; > > } > > -static void g4x_merge_wm(struct drm_i915_private *dev_priv, > > +static void g4x_merge_wm(struct drm_i915_private *i915, > > struct g4x_wm_values *wm) > > { > > struct intel_crtc *crtc; > > @@ -1494,7 +1494,7 @@ static void g4x_merge_wm(struct drm_i915_private *dev_priv, > > wm->hpll_en = true; > > wm->fbc_en = true; > > - for_each_intel_crtc(&dev_priv->drm, crtc) { > > + for_each_intel_crtc(&i915->drm, crtc) { > > const struct g4x_wm_state *wm_state = &crtc->wm.active.g4x; > > if (!crtc->active) > > @@ -1516,7 +1516,7 @@ static void g4x_merge_wm(struct drm_i915_private *dev_priv, > > wm->fbc_en = false; > > } > > - for_each_intel_crtc(&dev_priv->drm, crtc) { > > + for_each_intel_crtc(&i915->drm, crtc) { > > const struct g4x_wm_state *wm_state = &crtc->wm.active.g4x; > > enum pipe pipe = crtc->pipe; > > @@ -1528,23 +1528,23 @@ static void g4x_merge_wm(struct drm_i915_private *dev_priv, > > } > > } > > -static void g4x_program_watermarks(struct drm_i915_private *dev_priv) > > +static void g4x_program_watermarks(struct drm_i915_private *i915) > > { > > - struct g4x_wm_values *old_wm = &dev_priv->wm.g4x; > > + struct g4x_wm_values *old_wm = &i915->wm.g4x; > > struct g4x_wm_values new_wm = {}; > > - g4x_merge_wm(dev_priv, &new_wm); > > + g4x_merge_wm(i915, &new_wm); > > if (memcmp(old_wm, &new_wm, sizeof(new_wm)) == 0) > > return; > > if (is_disabling(old_wm->cxsr, new_wm.cxsr, true)) > > - _intel_set_memory_cxsr(dev_priv, false); > > + _intel_set_memory_cxsr(i915, false); > > - g4x_write_wm_values(dev_priv, &new_wm); > > + g4x_write_wm_values(i915, &new_wm); > > if (is_enabling(old_wm->cxsr, new_wm.cxsr, true)) > > - _intel_set_memory_cxsr(dev_priv, true); > > + _intel_set_memory_cxsr(i915, true); > > *old_wm = new_wm; > > } > > @@ -1552,28 +1552,28 @@ static void g4x_program_watermarks(struct drm_i915_private *dev_priv) > > static void g4x_initial_watermarks(struct intel_atomic_state *state, > > struct intel_crtc_state *crtc_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc_state->base.crtc->dev); > > + struct drm_i915_private *i915 = to_i915(crtc_state->base.crtc->dev); > > struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc); > > - mutex_lock(&dev_priv->wm.wm_mutex); > > + mutex_lock(&i915->wm.wm_mutex); > > crtc->wm.active.g4x = crtc_state->wm.g4x.intermediate; > > - g4x_program_watermarks(dev_priv); > > - mutex_unlock(&dev_priv->wm.wm_mutex); > > + g4x_program_watermarks(i915); > > + mutex_unlock(&i915->wm.wm_mutex); > > } > > static void g4x_optimize_watermarks(struct intel_atomic_state *state, > > struct intel_crtc_state *crtc_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc_state->base.crtc->dev); > > + struct drm_i915_private *i915 = to_i915(crtc_state->base.crtc->dev); > > struct intel_crtc *intel_crtc = to_intel_crtc(crtc_state->base.crtc); > > if (!crtc_state->wm.need_postvbl_update) > > return; > > - mutex_lock(&dev_priv->wm.wm_mutex); > > + mutex_lock(&i915->wm.wm_mutex); > > intel_crtc->wm.active.g4x = crtc_state->wm.g4x.optimal; > > - g4x_program_watermarks(dev_priv); > > - mutex_unlock(&dev_priv->wm.wm_mutex); > > + g4x_program_watermarks(i915); > > + mutex_unlock(&i915->wm.wm_mutex); > > } > > /* latency must be in 0.1us units. */ > > @@ -1592,18 +1592,18 @@ static unsigned int vlv_wm_method2(unsigned int pixel_rate, > > return ret; > > } > > -static void vlv_setup_wm_latency(struct drm_i915_private *dev_priv) > > +static void vlv_setup_wm_latency(struct drm_i915_private *i915) > > { > > /* all latencies in usec */ > > - dev_priv->wm.pri_latency[VLV_WM_LEVEL_PM2] = 3; > > + i915->wm.pri_latency[VLV_WM_LEVEL_PM2] = 3; > > - dev_priv->wm.max_level = VLV_WM_LEVEL_PM2; > > + i915->wm.max_level = VLV_WM_LEVEL_PM2; > > - if (IS_CHERRYVIEW(dev_priv)) { > > - dev_priv->wm.pri_latency[VLV_WM_LEVEL_PM5] = 12; > > - dev_priv->wm.pri_latency[VLV_WM_LEVEL_DDR_DVFS] = 33; > > + if (IS_CHERRYVIEW(i915)) { > > + i915->wm.pri_latency[VLV_WM_LEVEL_PM5] = 12; > > + i915->wm.pri_latency[VLV_WM_LEVEL_DDR_DVFS] = 33; > > - dev_priv->wm.max_level = VLV_WM_LEVEL_DDR_DVFS; > > + i915->wm.max_level = VLV_WM_LEVEL_DDR_DVFS; > > } > > } > > @@ -1612,12 +1612,12 @@ static u16 vlv_compute_wm_level(const struct intel_crtc_state *crtc_state, > > int level) > > { > > struct intel_plane *plane = to_intel_plane(plane_state->base.plane); > > - struct drm_i915_private *dev_priv = to_i915(plane->base.dev); > > + struct drm_i915_private *i915 = to_i915(plane->base.dev); > > const struct drm_display_mode *adjusted_mode = > > &crtc_state->base.adjusted_mode; > > unsigned int clock, htotal, cpp, width, wm; > > - if (dev_priv->wm.pri_latency[level] == 0) > > + if (i915->wm.pri_latency[level] == 0) > > return USHRT_MAX; > > if (!intel_wm_plane_visible(crtc_state, plane_state)) > > @@ -1638,7 +1638,7 @@ static u16 vlv_compute_wm_level(const struct intel_crtc_state *crtc_state, > > wm = 63; > > } else { > > wm = vlv_wm_method2(clock, htotal, width, cpp, > > - dev_priv->wm.pri_latency[level] * 10); > > + i915->wm.pri_latency[level] * 10); > > } > > return min_t(unsigned int, wm, USHRT_MAX); > > @@ -1736,9 +1736,9 @@ static int vlv_compute_fifo(struct intel_crtc_state *crtc_state) > > static void vlv_invalidate_wms(struct intel_crtc *crtc, > > struct vlv_wm_state *wm_state, int level) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > - for (; level < intel_wm_num_levels(dev_priv); level++) { > > + for (; level < intel_wm_num_levels(i915); level++) { > > enum plane_id plane_id; > > for_each_plane_id_on_crtc(crtc, plane_id) > > @@ -1764,8 +1764,8 @@ static u16 vlv_invert_wm_value(u16 wm, u16 fifo_size) > > static bool vlv_raw_plane_wm_set(struct intel_crtc_state *crtc_state, > > int level, enum plane_id plane_id, u16 value) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc_state->base.crtc->dev); > > - int num_levels = intel_wm_num_levels(dev_priv); > > + struct drm_i915_private *i915 = to_i915(crtc_state->base.crtc->dev); > > + int num_levels = intel_wm_num_levels(i915); > > bool dirty = false; > > for (; level < num_levels; level++) { > > @@ -1840,7 +1840,7 @@ static bool vlv_raw_crtc_wm_is_valid(const struct intel_crtc_state *crtc_state, > > static int vlv_compute_pipe_wm(struct intel_crtc_state *crtc_state) > > { > > struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc); > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > struct intel_atomic_state *state = > > to_intel_atomic_state(crtc_state->base.state); > > struct vlv_wm_state *wm_state = &crtc_state->wm.vlv.optimal; > > @@ -1897,7 +1897,7 @@ static int vlv_compute_pipe_wm(struct intel_crtc_state *crtc_state) > > } > > /* initially allow all levels */ > > - wm_state->num_levels = intel_wm_num_levels(dev_priv); > > + wm_state->num_levels = intel_wm_num_levels(i915); > > /* > > * Note that enabling cxsr with no primary/sprite planes > > * enabled can wedge the pipe. Hence we only allow cxsr > > @@ -1907,7 +1907,7 @@ static int vlv_compute_pipe_wm(struct intel_crtc_state *crtc_state) > > for (level = 0; level < wm_state->num_levels; level++) { > > const struct g4x_pipe_wm *raw = &crtc_state->wm.vlv.raw[level]; > > - const int sr_fifo_size = INTEL_INFO(dev_priv)->num_pipes * 512 - 1; > > + const int sr_fifo_size = INTEL_INFO(i915)->num_pipes * 512 - 1; > > if (!vlv_raw_crtc_wm_is_valid(crtc_state, level)) > > break; > > @@ -1948,8 +1948,8 @@ static void vlv_atomic_update_fifo(struct intel_atomic_state *state, > > struct intel_crtc_state *crtc_state) > > { > > struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc); > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > - struct intel_uncore *uncore = &dev_priv->uncore; > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > + struct intel_uncore *uncore = &i915->uncore; > > const struct vlv_fifo_state *fifo_state = > > &crtc_state->wm.vlv.fifo_state; > > int sprite0_start, sprite1_start, fifo_size; > > @@ -2092,16 +2092,16 @@ static int vlv_compute_intermediate_wm(struct intel_crtc_state *new_crtc_state) > > return 0; > > } > > -static void vlv_merge_wm(struct drm_i915_private *dev_priv, > > +static void vlv_merge_wm(struct drm_i915_private *i915, > > struct vlv_wm_values *wm) > > { > > struct intel_crtc *crtc; > > int num_active_crtcs = 0; > > - wm->level = dev_priv->wm.max_level; > > + wm->level = i915->wm.max_level; > > wm->cxsr = true; > > - for_each_intel_crtc(&dev_priv->drm, crtc) { > > + for_each_intel_crtc(&i915->drm, crtc) { > > const struct vlv_wm_state *wm_state = &crtc->wm.active.vlv; > > if (!crtc->active) > > @@ -2120,7 +2120,7 @@ static void vlv_merge_wm(struct drm_i915_private *dev_priv, > > if (num_active_crtcs > 1) > > wm->level = VLV_WM_LEVEL_PM2; > > - for_each_intel_crtc(&dev_priv->drm, crtc) { > > + for_each_intel_crtc(&i915->drm, crtc) { > > const struct vlv_wm_state *wm_state = &crtc->wm.active.vlv; > > enum pipe pipe = crtc->pipe; > > @@ -2135,35 +2135,35 @@ static void vlv_merge_wm(struct drm_i915_private *dev_priv, > > } > > } > > -static void vlv_program_watermarks(struct drm_i915_private *dev_priv) > > +static void vlv_program_watermarks(struct drm_i915_private *i915) > > { > > - struct vlv_wm_values *old_wm = &dev_priv->wm.vlv; > > + struct vlv_wm_values *old_wm = &i915->wm.vlv; > > struct vlv_wm_values new_wm = {}; > > - vlv_merge_wm(dev_priv, &new_wm); > > + vlv_merge_wm(i915, &new_wm); > > if (memcmp(old_wm, &new_wm, sizeof(new_wm)) == 0) > > return; > > if (is_disabling(old_wm->level, new_wm.level, VLV_WM_LEVEL_DDR_DVFS)) > > - chv_set_memory_dvfs(dev_priv, false); > > + chv_set_memory_dvfs(i915, false); > > if (is_disabling(old_wm->level, new_wm.level, VLV_WM_LEVEL_PM5)) > > - chv_set_memory_pm5(dev_priv, false); > > + chv_set_memory_pm5(i915, false); > > if (is_disabling(old_wm->cxsr, new_wm.cxsr, true)) > > - _intel_set_memory_cxsr(dev_priv, false); > > + _intel_set_memory_cxsr(i915, false); > > - vlv_write_wm_values(dev_priv, &new_wm); > > + vlv_write_wm_values(i915, &new_wm); > > if (is_enabling(old_wm->cxsr, new_wm.cxsr, true)) > > - _intel_set_memory_cxsr(dev_priv, true); > > + _intel_set_memory_cxsr(i915, true); > > if (is_enabling(old_wm->level, new_wm.level, VLV_WM_LEVEL_PM5)) > > - chv_set_memory_pm5(dev_priv, true); > > + chv_set_memory_pm5(i915, true); > > if (is_enabling(old_wm->level, new_wm.level, VLV_WM_LEVEL_DDR_DVFS)) > > - chv_set_memory_dvfs(dev_priv, true); > > + chv_set_memory_dvfs(i915, true); > > *old_wm = new_wm; > > } > > @@ -2171,40 +2171,40 @@ static void vlv_program_watermarks(struct drm_i915_private *dev_priv) > > static void vlv_initial_watermarks(struct intel_atomic_state *state, > > struct intel_crtc_state *crtc_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc_state->base.crtc->dev); > > + struct drm_i915_private *i915 = to_i915(crtc_state->base.crtc->dev); > > struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc); > > - mutex_lock(&dev_priv->wm.wm_mutex); > > + mutex_lock(&i915->wm.wm_mutex); > > crtc->wm.active.vlv = crtc_state->wm.vlv.intermediate; > > - vlv_program_watermarks(dev_priv); > > - mutex_unlock(&dev_priv->wm.wm_mutex); > > + vlv_program_watermarks(i915); > > + mutex_unlock(&i915->wm.wm_mutex); > > } > > static void vlv_optimize_watermarks(struct intel_atomic_state *state, > > struct intel_crtc_state *crtc_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc_state->base.crtc->dev); > > + struct drm_i915_private *i915 = to_i915(crtc_state->base.crtc->dev); > > struct intel_crtc *intel_crtc = to_intel_crtc(crtc_state->base.crtc); > > if (!crtc_state->wm.need_postvbl_update) > > return; > > - mutex_lock(&dev_priv->wm.wm_mutex); > > + mutex_lock(&i915->wm.wm_mutex); > > intel_crtc->wm.active.vlv = crtc_state->wm.vlv.optimal; > > - vlv_program_watermarks(dev_priv); > > - mutex_unlock(&dev_priv->wm.wm_mutex); > > + vlv_program_watermarks(i915); > > + mutex_unlock(&i915->wm.wm_mutex); > > } > > static void i965_update_wm(struct intel_crtc *unused_crtc) > > { > > - struct drm_i915_private *dev_priv = to_i915(unused_crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(unused_crtc->base.dev); > > struct intel_crtc *crtc; > > int srwm = 1; > > int cursor_sr = 16; > > bool cxsr_enabled; > > /* Calc sr entries for one plane configs */ > > - crtc = single_enabled_crtc(dev_priv); > > + crtc = single_enabled_crtc(i915); > > if (crtc) { > > /* self-refresh has much higher latency */ > > static const int sr_latency_ns = 12000; > > @@ -2246,7 +2246,7 @@ static void i965_update_wm(struct intel_crtc *unused_crtc) > > } else { > > cxsr_enabled = false; > > /* Turn off self refresh if both pipes are enabled */ > > - intel_set_memory_cxsr(dev_priv, false); > > + intel_set_memory_cxsr(i915, false); > > } > > DRM_DEBUG_KMS("Setting FIFO watermarks - A: 8, B: 8, C: 8, SR %d\n", > > @@ -2263,14 +2263,14 @@ static void i965_update_wm(struct intel_crtc *unused_crtc) > > I915_WRITE(DSPFW3, FW_WM(cursor_sr, CURSOR_SR)); > > if (cxsr_enabled) > > - intel_set_memory_cxsr(dev_priv, true); > > + intel_set_memory_cxsr(i915, true); > > } > > #undef FW_WM > > static void i9xx_update_wm(struct intel_crtc *unused_crtc) > > { > > - struct drm_i915_private *dev_priv = to_i915(unused_crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(unused_crtc->base.dev); > > const struct intel_watermark_params *wm_info; > > u32 fwater_lo; > > u32 fwater_hi; > > @@ -2279,15 +2279,15 @@ static void i9xx_update_wm(struct intel_crtc *unused_crtc) > > int planea_wm, planeb_wm; > > struct intel_crtc *crtc, *enabled = NULL; > > - if (IS_I945GM(dev_priv)) > > + if (IS_I945GM(i915)) > > wm_info = &i945_wm_info; > > - else if (!IS_GEN(dev_priv, 2)) > > + else if (!IS_GEN(i915, 2)) > > wm_info = &i915_wm_info; > > else > > wm_info = &i830_a_wm_info; > > - fifo_size = dev_priv->display.get_fifo_size(dev_priv, PLANE_A); > > - crtc = intel_get_crtc_for_plane(dev_priv, PLANE_A); > > + fifo_size = i915->display.get_fifo_size(i915, PLANE_A); > > + crtc = intel_get_crtc_for_plane(i915, PLANE_A); > > if (intel_crtc_active(crtc)) { > > const struct drm_display_mode *adjusted_mode = > > &crtc->config->base.adjusted_mode; > > @@ -2295,7 +2295,7 @@ static void i9xx_update_wm(struct intel_crtc *unused_crtc) > > crtc->base.primary->state->fb; > > int cpp; > > - if (IS_GEN(dev_priv, 2)) > > + if (IS_GEN(i915, 2)) > > cpp = 4; > > else > > cpp = fb->format->cpp[0]; > > @@ -2310,11 +2310,11 @@ static void i9xx_update_wm(struct intel_crtc *unused_crtc) > > planea_wm = wm_info->max_wm; > > } > > - if (IS_GEN(dev_priv, 2)) > > + if (IS_GEN(i915, 2)) > > wm_info = &i830_bc_wm_info; > > - fifo_size = dev_priv->display.get_fifo_size(dev_priv, PLANE_B); > > - crtc = intel_get_crtc_for_plane(dev_priv, PLANE_B); > > + fifo_size = i915->display.get_fifo_size(i915, PLANE_B); > > + crtc = intel_get_crtc_for_plane(i915, PLANE_B); > > if (intel_crtc_active(crtc)) { > > const struct drm_display_mode *adjusted_mode = > > &crtc->config->base.adjusted_mode; > > @@ -2322,7 +2322,7 @@ static void i9xx_update_wm(struct intel_crtc *unused_crtc) > > crtc->base.primary->state->fb; > > int cpp; > > - if (IS_GEN(dev_priv, 2)) > > + if (IS_GEN(i915, 2)) > > cpp = 4; > > else > > cpp = fb->format->cpp[0]; > > @@ -2342,7 +2342,7 @@ static void i9xx_update_wm(struct intel_crtc *unused_crtc) > > DRM_DEBUG_KMS("FIFO watermarks - A: %d, B: %d\n", planea_wm, planeb_wm); > > - if (IS_I915GM(dev_priv) && enabled) { > > + if (IS_I915GM(i915) && enabled) { > > struct drm_i915_gem_object *obj; > > obj = intel_fb_obj(enabled->base.primary->state->fb); > > @@ -2358,10 +2358,10 @@ static void i9xx_update_wm(struct intel_crtc *unused_crtc) > > cwm = 2; > > /* Play safe and disable self-refresh before adjusting watermarks. */ > > - intel_set_memory_cxsr(dev_priv, false); > > + intel_set_memory_cxsr(i915, false); > > /* Calc sr entries for one plane configs */ > > - if (HAS_FW_BLC(dev_priv) && enabled) { > > + if (HAS_FW_BLC(i915) && enabled) { > > /* self-refresh has much higher latency */ > > static const int sr_latency_ns = 6000; > > const struct drm_display_mode *adjusted_mode = > > @@ -2374,7 +2374,7 @@ static void i9xx_update_wm(struct intel_crtc *unused_crtc) > > int cpp; > > int entries; > > - if (IS_I915GM(dev_priv) || IS_I945GM(dev_priv)) > > + if (IS_I915GM(i915) || IS_I945GM(i915)) > > cpp = 4; > > else > > cpp = fb->format->cpp[0]; > > @@ -2387,7 +2387,7 @@ static void i9xx_update_wm(struct intel_crtc *unused_crtc) > > if (srwm < 0) > > srwm = 1; > > - if (IS_I945G(dev_priv) || IS_I945GM(dev_priv)) > > + if (IS_I945G(i915) || IS_I945GM(i915)) > > I915_WRITE(FW_BLC_SELF, > > FW_BLC_SELF_FIFO_MASK | (srwm & 0xff)); > > else > > @@ -2408,25 +2408,25 @@ static void i9xx_update_wm(struct intel_crtc *unused_crtc) > > I915_WRITE(FW_BLC2, fwater_hi); > > if (enabled) > > - intel_set_memory_cxsr(dev_priv, true); > > + intel_set_memory_cxsr(i915, true); > > } > > static void i845_update_wm(struct intel_crtc *unused_crtc) > > { > > - struct drm_i915_private *dev_priv = to_i915(unused_crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(unused_crtc->base.dev); > > struct intel_crtc *crtc; > > const struct drm_display_mode *adjusted_mode; > > u32 fwater_lo; > > int planea_wm; > > - crtc = single_enabled_crtc(dev_priv); > > + crtc = single_enabled_crtc(i915); > > if (crtc == NULL) > > return; > > adjusted_mode = &crtc->config->base.adjusted_mode; > > planea_wm = intel_calculate_wm(adjusted_mode->crtc_clock, > > &i845_wm_info, > > - dev_priv->display.get_fifo_size(dev_priv, PLANE_A), > > + i915->display.get_fifo_size(i915, PLANE_A), > > 4, pessimal_latency_ns); > > fwater_lo = I915_READ(FW_BLC) & ~0xfff; > > fwater_lo |= (3<<8) | planea_wm; > > @@ -2586,24 +2586,24 @@ static u32 ilk_compute_fbc_wm(const struct intel_crtc_state *cstate, > > } > > static unsigned int > > -ilk_display_fifo_size(const struct drm_i915_private *dev_priv) > > +ilk_display_fifo_size(const struct drm_i915_private *i915) > > { > > - if (INTEL_GEN(dev_priv) >= 8) > > + if (INTEL_GEN(i915) >= 8) > > return 3072; > > - else if (INTEL_GEN(dev_priv) >= 7) > > + else if (INTEL_GEN(i915) >= 7) > > return 768; > > else > > return 512; > > } > > static unsigned int > > -ilk_plane_wm_reg_max(const struct drm_i915_private *dev_priv, > > +ilk_plane_wm_reg_max(const struct drm_i915_private *i915, > > int level, bool is_sprite) > > { > > - if (INTEL_GEN(dev_priv) >= 8) > > + if (INTEL_GEN(i915) >= 8) > > /* BDW primary/sprite plane watermarks */ > > return level == 0 ? 255 : 2047; > > - else if (INTEL_GEN(dev_priv) >= 7) > > + else if (INTEL_GEN(i915) >= 7) > > /* IVB/HSW primary/sprite plane watermarks */ > > return level == 0 ? 127 : 1023; > > else if (!is_sprite) > > @@ -2615,30 +2615,30 @@ ilk_plane_wm_reg_max(const struct drm_i915_private *dev_priv, > > } > > static unsigned int > > -ilk_cursor_wm_reg_max(const struct drm_i915_private *dev_priv, int level) > > +ilk_cursor_wm_reg_max(const struct drm_i915_private *i915, int level) > > { > > - if (INTEL_GEN(dev_priv) >= 7) > > + if (INTEL_GEN(i915) >= 7) > > return level == 0 ? 63 : 255; > > else > > return level == 0 ? 31 : 63; > > } > > -static unsigned int ilk_fbc_wm_reg_max(const struct drm_i915_private *dev_priv) > > +static unsigned int ilk_fbc_wm_reg_max(const struct drm_i915_private *i915) > > { > > - if (INTEL_GEN(dev_priv) >= 8) > > + if (INTEL_GEN(i915) >= 8) > > return 31; > > else > > return 15; > > } > > /* Calculate the maximum primary/sprite plane watermark */ > > -static unsigned int ilk_plane_wm_max(const struct drm_i915_private *dev_priv, > > +static unsigned int ilk_plane_wm_max(const struct drm_i915_private *i915, > > int level, > > const struct intel_wm_config *config, > > enum intel_ddb_partitioning ddb_partitioning, > > bool is_sprite) > > { > > - unsigned int fifo_size = ilk_display_fifo_size(dev_priv); > > + unsigned int fifo_size = ilk_display_fifo_size(i915); > > /* if sprites aren't enabled, sprites get nothing */ > > if (is_sprite && !config->sprites_enabled) > > @@ -2646,14 +2646,14 @@ static unsigned int ilk_plane_wm_max(const struct drm_i915_private *dev_priv, > > /* HSW allows LP1+ watermarks even with multiple pipes */ > > if (level == 0 || config->num_pipes_active > 1) { > > - fifo_size /= INTEL_INFO(dev_priv)->num_pipes; > > + fifo_size /= INTEL_INFO(i915)->num_pipes; > > /* > > * For some reason the non self refresh > > * FIFO size is only half of the self > > * refresh FIFO size on ILK/SNB. > > */ > > - if (INTEL_GEN(dev_priv) <= 6) > > + if (INTEL_GEN(i915) <= 6) > > fifo_size /= 2; > > } > > @@ -2669,11 +2669,11 @@ static unsigned int ilk_plane_wm_max(const struct drm_i915_private *dev_priv, > > } > > /* clamp to max that the registers can hold */ > > - return min(fifo_size, ilk_plane_wm_reg_max(dev_priv, level, is_sprite)); > > + return min(fifo_size, ilk_plane_wm_reg_max(i915, level, is_sprite)); > > } > > /* Calculate the maximum cursor plane watermark */ > > -static unsigned int ilk_cursor_wm_max(const struct drm_i915_private *dev_priv, > > +static unsigned int ilk_cursor_wm_max(const struct drm_i915_private *i915, > > int level, > > const struct intel_wm_config *config) > > { > > @@ -2682,29 +2682,29 @@ static unsigned int ilk_cursor_wm_max(const struct drm_i915_private *dev_priv, > > return 64; > > /* otherwise just report max that registers can hold */ > > - return ilk_cursor_wm_reg_max(dev_priv, level); > > + return ilk_cursor_wm_reg_max(i915, level); > > } > > -static void ilk_compute_wm_maximums(const struct drm_i915_private *dev_priv, > > +static void ilk_compute_wm_maximums(const struct drm_i915_private *i915, > > int level, > > const struct intel_wm_config *config, > > enum intel_ddb_partitioning ddb_partitioning, > > struct ilk_wm_maximums *max) > > { > > - max->pri = ilk_plane_wm_max(dev_priv, level, config, ddb_partitioning, false); > > - max->spr = ilk_plane_wm_max(dev_priv, level, config, ddb_partitioning, true); > > - max->cur = ilk_cursor_wm_max(dev_priv, level, config); > > - max->fbc = ilk_fbc_wm_reg_max(dev_priv); > > + max->pri = ilk_plane_wm_max(i915, level, config, ddb_partitioning, false); > > + max->spr = ilk_plane_wm_max(i915, level, config, ddb_partitioning, true); > > + max->cur = ilk_cursor_wm_max(i915, level, config); > > + max->fbc = ilk_fbc_wm_reg_max(i915); > > } > > -static void ilk_compute_wm_reg_maximums(const struct drm_i915_private *dev_priv, > > +static void ilk_compute_wm_reg_maximums(const struct drm_i915_private *i915, > > int level, > > struct ilk_wm_maximums *max) > > { > > - max->pri = ilk_plane_wm_reg_max(dev_priv, level, false); > > - max->spr = ilk_plane_wm_reg_max(dev_priv, level, true); > > - max->cur = ilk_cursor_wm_reg_max(dev_priv, level); > > - max->fbc = ilk_fbc_wm_reg_max(dev_priv); > > + max->pri = ilk_plane_wm_reg_max(i915, level, false); > > + max->spr = ilk_plane_wm_reg_max(i915, level, true); > > + max->cur = ilk_cursor_wm_reg_max(i915, level); > > + max->fbc = ilk_fbc_wm_reg_max(i915); > > } > > static bool ilk_validate_wm_level(int level, > > @@ -2748,7 +2748,7 @@ static bool ilk_validate_wm_level(int level, > > return ret; > > } > > -static void ilk_compute_wm_level(const struct drm_i915_private *dev_priv, > > +static void ilk_compute_wm_level(const struct drm_i915_private *i915, > > const struct intel_crtc *intel_crtc, > > int level, > > struct intel_crtc_state *cstate, > > @@ -2757,9 +2757,9 @@ static void ilk_compute_wm_level(const struct drm_i915_private *dev_priv, > > const struct intel_plane_state *curstate, > > struct intel_wm_level *result) > > { > > - u16 pri_latency = dev_priv->wm.pri_latency[level]; > > - u16 spr_latency = dev_priv->wm.spr_latency[level]; > > - u16 cur_latency = dev_priv->wm.cur_latency[level]; > > + u16 pri_latency = i915->wm.pri_latency[level]; > > + u16 spr_latency = i915->wm.spr_latency[level]; > > + u16 cur_latency = i915->wm.cur_latency[level]; > > /* WM1+ latency values stored in 0.5us units */ > > if (level > 0) { > > @@ -2811,19 +2811,19 @@ hsw_compute_linetime_wm(const struct intel_crtc_state *cstate) > > PIPE_WM_LINETIME_TIME(linetime); > > } > > -static void intel_read_wm_latency(struct drm_i915_private *dev_priv, > > +static void intel_read_wm_latency(struct drm_i915_private *i915, > > u16 wm[8]) > > { > > - struct intel_uncore *uncore = &dev_priv->uncore; > > + struct intel_uncore *uncore = &i915->uncore; > > - if (INTEL_GEN(dev_priv) >= 9) { > > + if (INTEL_GEN(i915) >= 9) { > > u32 val; > > int ret, i; > > - int level, max_level = ilk_wm_max_level(dev_priv); > > + int level, max_level = ilk_wm_max_level(i915); > > /* read the first set of memory latencies[0:3] */ > > val = 0; /* data0 to be programmed to 0 for first set */ > > - ret = sandybridge_pcode_read(dev_priv, > > + ret = sandybridge_pcode_read(i915, > > GEN9_PCODE_READ_MEM_LATENCY, > > &val, NULL); > > @@ -2842,7 +2842,7 @@ static void intel_read_wm_latency(struct drm_i915_private *dev_priv, > > /* read the second set of memory latencies[4:7] */ > > val = 1; /* data0 to be programmed to 1 for second set */ > > - ret = sandybridge_pcode_read(dev_priv, > > + ret = sandybridge_pcode_read(i915, > > GEN9_PCODE_READ_MEM_LATENCY, > > &val, NULL); > > if (ret) { > > @@ -2893,10 +2893,10 @@ static void intel_read_wm_latency(struct drm_i915_private *dev_priv, > > * any underrun. If not able to get Dimm info assume 16GB dimm > > * to avoid any underrun. > > */ > > - if (dev_priv->dram_info.is_16gb_dimm) > > + if (i915->dram_info.is_16gb_dimm) > > wm[0] += 1; > > - } else if (IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv)) { > > + } else if (IS_HASWELL(i915) || IS_BROADWELL(i915)) { > > u64 sskpd = intel_uncore_read64(uncore, MCH_SSKPD); > > wm[0] = (sskpd >> 56) & 0xFF; > > @@ -2906,14 +2906,14 @@ static void intel_read_wm_latency(struct drm_i915_private *dev_priv, > > wm[2] = (sskpd >> 12) & 0xFF; > > wm[3] = (sskpd >> 20) & 0x1FF; > > wm[4] = (sskpd >> 32) & 0x1FF; > > - } else if (INTEL_GEN(dev_priv) >= 6) { > > + } else if (INTEL_GEN(i915) >= 6) { > > u32 sskpd = intel_uncore_read(uncore, MCH_SSKPD); > > wm[0] = (sskpd >> SSKPD_WM0_SHIFT) & SSKPD_WM_MASK; > > wm[1] = (sskpd >> SSKPD_WM1_SHIFT) & SSKPD_WM_MASK; > > wm[2] = (sskpd >> SSKPD_WM2_SHIFT) & SSKPD_WM_MASK; > > wm[3] = (sskpd >> SSKPD_WM3_SHIFT) & SSKPD_WM_MASK; > > - } else if (INTEL_GEN(dev_priv) >= 5) { > > + } else if (INTEL_GEN(i915) >= 5) { > > u32 mltr = intel_uncore_read(uncore, MLTR_ILK); > > /* ILK primary LP0 latency is 700 ns */ > > @@ -2921,44 +2921,44 @@ static void intel_read_wm_latency(struct drm_i915_private *dev_priv, > > wm[1] = (mltr >> MLTR_WM1_SHIFT) & ILK_SRLT_MASK; > > wm[2] = (mltr >> MLTR_WM2_SHIFT) & ILK_SRLT_MASK; > > } else { > > - MISSING_CASE(INTEL_DEVID(dev_priv)); > > + MISSING_CASE(INTEL_DEVID(i915)); > > } > > } > > -static void intel_fixup_spr_wm_latency(struct drm_i915_private *dev_priv, > > +static void intel_fixup_spr_wm_latency(struct drm_i915_private *i915, > > u16 wm[5]) > > { > > /* ILK sprite LP0 latency is 1300 ns */ > > - if (IS_GEN(dev_priv, 5)) > > + if (IS_GEN(i915, 5)) > > wm[0] = 13; > > } > > -static void intel_fixup_cur_wm_latency(struct drm_i915_private *dev_priv, > > +static void intel_fixup_cur_wm_latency(struct drm_i915_private *i915, > > u16 wm[5]) > > { > > /* ILK cursor LP0 latency is 1300 ns */ > > - if (IS_GEN(dev_priv, 5)) > > + if (IS_GEN(i915, 5)) > > wm[0] = 13; > > } > > -int ilk_wm_max_level(const struct drm_i915_private *dev_priv) > > +int ilk_wm_max_level(const struct drm_i915_private *i915) > > { > > /* how many WM levels are we expecting */ > > - if (INTEL_GEN(dev_priv) >= 9) > > + if (INTEL_GEN(i915) >= 9) > > return 7; > > - else if (IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv)) > > + else if (IS_HASWELL(i915) || IS_BROADWELL(i915)) > > return 4; > > - else if (INTEL_GEN(dev_priv) >= 6) > > + else if (INTEL_GEN(i915) >= 6) > > return 3; > > else > > return 2; > > } > > -static void intel_print_wm_latency(struct drm_i915_private *dev_priv, > > +static void intel_print_wm_latency(struct drm_i915_private *i915, > > const char *name, > > const u16 wm[8]) > > { > > - int level, max_level = ilk_wm_max_level(dev_priv); > > + int level, max_level = ilk_wm_max_level(i915); > > for (level = 0; level <= max_level; level++) { > > unsigned int latency = wm[level]; > > @@ -2973,7 +2973,7 @@ static void intel_print_wm_latency(struct drm_i915_private *dev_priv, > > * - latencies are in us on gen9. > > * - before then, WM1+ latency values are in 0.5us units > > */ > > - if (INTEL_GEN(dev_priv) >= 9) > > + if (INTEL_GEN(i915) >= 9) > > latency *= 10; > > else if (level > 0) > > latency *= 5; > > @@ -2984,10 +2984,10 @@ static void intel_print_wm_latency(struct drm_i915_private *dev_priv, > > } > > } > > -static bool ilk_increase_wm_latency(struct drm_i915_private *dev_priv, > > +static bool ilk_increase_wm_latency(struct drm_i915_private *i915, > > u16 wm[5], u16 min) > > { > > - int level, max_level = ilk_wm_max_level(dev_priv); > > + int level, max_level = ilk_wm_max_level(i915); > > if (wm[0] >= min) > > return false; > > @@ -2999,7 +2999,7 @@ static bool ilk_increase_wm_latency(struct drm_i915_private *dev_priv, > > return true; > > } > > -static void snb_wm_latency_quirk(struct drm_i915_private *dev_priv) > > +static void snb_wm_latency_quirk(struct drm_i915_private *i915) > > { > > bool changed; > > @@ -3007,20 +3007,20 @@ static void snb_wm_latency_quirk(struct drm_i915_private *dev_priv) > > * The BIOS provided WM memory latency values are often > > * inadequate for high resolution displays. Adjust them. > > */ > > - changed = ilk_increase_wm_latency(dev_priv, dev_priv->wm.pri_latency, 12) | > > - ilk_increase_wm_latency(dev_priv, dev_priv->wm.spr_latency, 12) | > > - ilk_increase_wm_latency(dev_priv, dev_priv->wm.cur_latency, 12); > > + changed = ilk_increase_wm_latency(i915, i915->wm.pri_latency, 12) | > > + ilk_increase_wm_latency(i915, i915->wm.spr_latency, 12) | > > + ilk_increase_wm_latency(i915, i915->wm.cur_latency, 12); > > if (!changed) > > return; > > DRM_DEBUG_KMS("WM latency values increased to avoid potential underruns\n"); > > - intel_print_wm_latency(dev_priv, "Primary", dev_priv->wm.pri_latency); > > - intel_print_wm_latency(dev_priv, "Sprite", dev_priv->wm.spr_latency); > > - intel_print_wm_latency(dev_priv, "Cursor", dev_priv->wm.cur_latency); > > + intel_print_wm_latency(i915, "Primary", i915->wm.pri_latency); > > + intel_print_wm_latency(i915, "Sprite", i915->wm.spr_latency); > > + intel_print_wm_latency(i915, "Cursor", i915->wm.cur_latency); > > } > > -static void snb_wm_lp3_irq_quirk(struct drm_i915_private *dev_priv) > > +static void snb_wm_lp3_irq_quirk(struct drm_i915_private *i915) > > { > > /* > > * On some SNB machines (Thinkpad X220 Tablet at least) > > @@ -3033,50 +3033,50 @@ static void snb_wm_lp3_irq_quirk(struct drm_i915_private *dev_priv) > > * interrupts only. To play it safe we disable LP3 > > * watermarks entirely. > > */ > > - if (dev_priv->wm.pri_latency[3] == 0 && > > - dev_priv->wm.spr_latency[3] == 0 && > > - dev_priv->wm.cur_latency[3] == 0) > > + if (i915->wm.pri_latency[3] == 0 && > > + i915->wm.spr_latency[3] == 0 && > > + i915->wm.cur_latency[3] == 0) > > return; > > - dev_priv->wm.pri_latency[3] = 0; > > - dev_priv->wm.spr_latency[3] = 0; > > - dev_priv->wm.cur_latency[3] = 0; > > + i915->wm.pri_latency[3] = 0; > > + i915->wm.spr_latency[3] = 0; > > + i915->wm.cur_latency[3] = 0; > > DRM_DEBUG_KMS("LP3 watermarks disabled due to potential for lost interrupts\n"); > > - intel_print_wm_latency(dev_priv, "Primary", dev_priv->wm.pri_latency); > > - intel_print_wm_latency(dev_priv, "Sprite", dev_priv->wm.spr_latency); > > - intel_print_wm_latency(dev_priv, "Cursor", dev_priv->wm.cur_latency); > > + intel_print_wm_latency(i915, "Primary", i915->wm.pri_latency); > > + intel_print_wm_latency(i915, "Sprite", i915->wm.spr_latency); > > + intel_print_wm_latency(i915, "Cursor", i915->wm.cur_latency); > > } > > -static void ilk_setup_wm_latency(struct drm_i915_private *dev_priv) > > +static void ilk_setup_wm_latency(struct drm_i915_private *i915) > > { > > - intel_read_wm_latency(dev_priv, dev_priv->wm.pri_latency); > > + intel_read_wm_latency(i915, i915->wm.pri_latency); > > - memcpy(dev_priv->wm.spr_latency, dev_priv->wm.pri_latency, > > - sizeof(dev_priv->wm.pri_latency)); > > - memcpy(dev_priv->wm.cur_latency, dev_priv->wm.pri_latency, > > - sizeof(dev_priv->wm.pri_latency)); > > + memcpy(i915->wm.spr_latency, i915->wm.pri_latency, > > + sizeof(i915->wm.pri_latency)); > > + memcpy(i915->wm.cur_latency, i915->wm.pri_latency, > > + sizeof(i915->wm.pri_latency)); > > - intel_fixup_spr_wm_latency(dev_priv, dev_priv->wm.spr_latency); > > - intel_fixup_cur_wm_latency(dev_priv, dev_priv->wm.cur_latency); > > + intel_fixup_spr_wm_latency(i915, i915->wm.spr_latency); > > + intel_fixup_cur_wm_latency(i915, i915->wm.cur_latency); > > - intel_print_wm_latency(dev_priv, "Primary", dev_priv->wm.pri_latency); > > - intel_print_wm_latency(dev_priv, "Sprite", dev_priv->wm.spr_latency); > > - intel_print_wm_latency(dev_priv, "Cursor", dev_priv->wm.cur_latency); > > + intel_print_wm_latency(i915, "Primary", i915->wm.pri_latency); > > + intel_print_wm_latency(i915, "Sprite", i915->wm.spr_latency); > > + intel_print_wm_latency(i915, "Cursor", i915->wm.cur_latency); > > - if (IS_GEN(dev_priv, 6)) { > > - snb_wm_latency_quirk(dev_priv); > > - snb_wm_lp3_irq_quirk(dev_priv); > > + if (IS_GEN(i915, 6)) { > > + snb_wm_latency_quirk(i915); > > + snb_wm_lp3_irq_quirk(i915); > > } > > } > > -static void skl_setup_wm_latency(struct drm_i915_private *dev_priv) > > +static void skl_setup_wm_latency(struct drm_i915_private *i915) > > { > > - intel_read_wm_latency(dev_priv, dev_priv->wm.skl_latency); > > - intel_print_wm_latency(dev_priv, "Gen9 Plane", dev_priv->wm.skl_latency); > > + intel_read_wm_latency(i915, i915->wm.skl_latency); > > + intel_print_wm_latency(i915, "Gen9 Plane", i915->wm.skl_latency); > > } > > -static bool ilk_validate_pipe_wm(const struct drm_i915_private *dev_priv, > > +static bool ilk_validate_pipe_wm(const struct drm_i915_private *i915, > > struct intel_pipe_wm *pipe_wm) > > { > > /* LP0 watermark maximums depend on this pipe alone */ > > @@ -3088,7 +3088,7 @@ static bool ilk_validate_pipe_wm(const struct drm_i915_private *dev_priv, > > struct ilk_wm_maximums max; > > /* LP0 watermarks always use 1/2 DDB partitioning */ > > - ilk_compute_wm_maximums(dev_priv, 0, &config, INTEL_DDB_PART_1_2, &max); > > + ilk_compute_wm_maximums(i915, 0, &config, INTEL_DDB_PART_1_2, &max); > > /* At least LP0 must be valid */ > > if (!ilk_validate_wm_level(0, &max, &pipe_wm->wm[0])) { > > @@ -3106,13 +3106,13 @@ static int ilk_compute_pipe_wm(struct intel_crtc_state *cstate) > > struct intel_crtc *intel_crtc = to_intel_crtc(cstate->base.crtc); > > struct intel_pipe_wm *pipe_wm; > > struct drm_device *dev = state->dev; > > - const struct drm_i915_private *dev_priv = to_i915(dev); > > + const struct drm_i915_private *i915 = to_i915(dev); > > struct drm_plane *plane; > > const struct drm_plane_state *plane_state; > > const struct intel_plane_state *pristate = NULL; > > const struct intel_plane_state *sprstate = NULL; > > const struct intel_plane_state *curstate = NULL; > > - int level, max_level = ilk_wm_max_level(dev_priv), usable_level; > > + int level, max_level = ilk_wm_max_level(i915), usable_level; > > struct ilk_wm_maximums max; > > pipe_wm = &cstate->wm.ilk.optimal; > > @@ -3139,7 +3139,7 @@ static int ilk_compute_pipe_wm(struct intel_crtc_state *cstate) > > usable_level = max_level; > > /* ILK/SNB: LP2+ watermarks only w/o sprites */ > > - if (INTEL_GEN(dev_priv) <= 6 && pipe_wm->sprites_enabled) > > + if (INTEL_GEN(i915) <= 6 && pipe_wm->sprites_enabled) > > usable_level = 1; > > /* ILK/SNB/IVB: LP1+ watermarks only w/o scaling */ > > @@ -3147,21 +3147,21 @@ static int ilk_compute_pipe_wm(struct intel_crtc_state *cstate) > > usable_level = 0; > > memset(&pipe_wm->wm, 0, sizeof(pipe_wm->wm)); > > - ilk_compute_wm_level(dev_priv, intel_crtc, 0, cstate, > > + ilk_compute_wm_level(i915, intel_crtc, 0, cstate, > > pristate, sprstate, curstate, &pipe_wm->wm[0]); > > - if (IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv)) > > + if (IS_HASWELL(i915) || IS_BROADWELL(i915)) > > pipe_wm->linetime = hsw_compute_linetime_wm(cstate); > > - if (!ilk_validate_pipe_wm(dev_priv, pipe_wm)) > > + if (!ilk_validate_pipe_wm(i915, pipe_wm)) > > return -EINVAL; > > - ilk_compute_wm_reg_maximums(dev_priv, 1, &max); > > + ilk_compute_wm_reg_maximums(i915, 1, &max); > > for (level = 1; level <= usable_level; level++) { > > struct intel_wm_level *wm = &pipe_wm->wm[level]; > > - ilk_compute_wm_level(dev_priv, intel_crtc, level, cstate, > > + ilk_compute_wm_level(i915, intel_crtc, level, cstate, > > pristate, sprstate, curstate, wm); > > /* > > @@ -3186,14 +3186,14 @@ static int ilk_compute_pipe_wm(struct intel_crtc_state *cstate) > > static int ilk_compute_intermediate_wm(struct intel_crtc_state *newstate) > > { > > struct intel_crtc *intel_crtc = to_intel_crtc(newstate->base.crtc); > > - struct drm_i915_private *dev_priv = to_i915(intel_crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(intel_crtc->base.dev); > > struct intel_pipe_wm *a = &newstate->wm.ilk.intermediate; > > struct intel_atomic_state *intel_state = > > to_intel_atomic_state(newstate->base.state); > > const struct intel_crtc_state *oldstate = > > intel_atomic_get_old_crtc_state(intel_state, intel_crtc); > > const struct intel_pipe_wm *b = &oldstate->wm.ilk.optimal; > > - int level, max_level = ilk_wm_max_level(dev_priv); > > + int level, max_level = ilk_wm_max_level(i915); > > /* > > * Start with the final, target watermarks, then combine with the > > @@ -3226,7 +3226,7 @@ static int ilk_compute_intermediate_wm(struct intel_crtc_state *newstate) > > * there's no safe way to transition from the old state to > > * the new state, so we need to fail the atomic transaction. > > */ > > - if (!ilk_validate_pipe_wm(dev_priv, a)) > > + if (!ilk_validate_pipe_wm(i915, a)) > > return -EINVAL; > > /* > > @@ -3242,7 +3242,7 @@ static int ilk_compute_intermediate_wm(struct intel_crtc_state *newstate) > > /* > > * Merge the watermarks from all active pipes for a specific level. > > */ > > -static void ilk_merge_wm_level(struct drm_i915_private *dev_priv, > > +static void ilk_merge_wm_level(struct drm_i915_private *i915, > > int level, > > struct intel_wm_level *ret_wm) > > { > > @@ -3250,7 +3250,7 @@ static void ilk_merge_wm_level(struct drm_i915_private *dev_priv, > > ret_wm->enable = true; > > - for_each_intel_crtc(&dev_priv->drm, intel_crtc) { > > + for_each_intel_crtc(&i915->drm, intel_crtc) { > > const struct intel_pipe_wm *active = &intel_crtc->wm.active.ilk; > > const struct intel_wm_level *wm = &active->wm[level]; > > @@ -3275,27 +3275,27 @@ static void ilk_merge_wm_level(struct drm_i915_private *dev_priv, > > /* > > * Merge all low power watermarks for all active pipes. > > */ > > -static void ilk_wm_merge(struct drm_i915_private *dev_priv, > > +static void ilk_wm_merge(struct drm_i915_private *i915, > > const struct intel_wm_config *config, > > const struct ilk_wm_maximums *max, > > struct intel_pipe_wm *merged) > > { > > - int level, max_level = ilk_wm_max_level(dev_priv); > > + int level, max_level = ilk_wm_max_level(i915); > > int last_enabled_level = max_level; > > /* ILK/SNB/IVB: LP1+ watermarks only w/ single pipe */ > > - if ((INTEL_GEN(dev_priv) <= 6 || IS_IVYBRIDGE(dev_priv)) && > > + if ((INTEL_GEN(i915) <= 6 || IS_IVYBRIDGE(i915)) && > > config->num_pipes_active > 1) > > last_enabled_level = 0; > > /* ILK: FBC WM must be disabled always */ > > - merged->fbc_wm_enabled = INTEL_GEN(dev_priv) >= 6; > > + merged->fbc_wm_enabled = INTEL_GEN(i915) >= 6; > > /* merge each WM1+ level */ > > for (level = 1; level <= max_level; level++) { > > struct intel_wm_level *wm = &merged->wm[level]; > > - ilk_merge_wm_level(dev_priv, level, wm); > > + ilk_merge_wm_level(i915, level, wm); > > if (level > last_enabled_level) > > wm->enable = false; > > @@ -3320,8 +3320,8 @@ static void ilk_wm_merge(struct drm_i915_private *dev_priv, > > * What we should check here is whether FBC can be > > * enabled sometime later. > > */ > > - if (IS_GEN(dev_priv, 5) && !merged->fbc_wm_enabled && > > - intel_fbc_is_active(dev_priv)) { > > + if (IS_GEN(i915, 5) && !merged->fbc_wm_enabled && > > + intel_fbc_is_active(i915)) { > > for (level = 2; level <= max_level; level++) { > > struct intel_wm_level *wm = &merged->wm[level]; > > @@ -3337,16 +3337,16 @@ static int ilk_wm_lp_to_level(int wm_lp, const struct intel_pipe_wm *pipe_wm) > > } > > /* The value we need to program into the WM_LPx latency field */ > > -static unsigned int ilk_wm_lp_latency(struct drm_i915_private *dev_priv, > > +static unsigned int ilk_wm_lp_latency(struct drm_i915_private *i915, > > int level) > > { > > - if (IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv)) > > + if (IS_HASWELL(i915) || IS_BROADWELL(i915)) > > return 2 * level; > > else > > - return dev_priv->wm.pri_latency[level]; > > + return i915->wm.pri_latency[level]; > > } > > -static void ilk_compute_wm_results(struct drm_i915_private *dev_priv, > > +static void ilk_compute_wm_results(struct drm_i915_private *i915, > > const struct intel_pipe_wm *merged, > > enum intel_ddb_partitioning partitioning, > > struct ilk_wm_values *results) > > @@ -3370,14 +3370,14 @@ static void ilk_compute_wm_results(struct drm_i915_private *dev_priv, > > * disabled. Doing otherwise could cause underruns. > > */ > > results->wm_lp[wm_lp - 1] = > > - (ilk_wm_lp_latency(dev_priv, level) << WM1_LP_LATENCY_SHIFT) | > > + (ilk_wm_lp_latency(i915, level) << WM1_LP_LATENCY_SHIFT) | > > (r->pri_val << WM1_LP_SR_SHIFT) | > > r->cur_val; > > if (r->enable) > > results->wm_lp[wm_lp - 1] |= WM1_LP_SR_EN; > > - if (INTEL_GEN(dev_priv) >= 8) > > + if (INTEL_GEN(i915) >= 8) > > results->wm_lp[wm_lp - 1] |= > > r->fbc_val << WM1_LP_FBC_SHIFT_BDW; > > else > > @@ -3388,7 +3388,7 @@ static void ilk_compute_wm_results(struct drm_i915_private *dev_priv, > > * Always set WM1S_LP_EN when spr_val != 0, even if the > > * level is disabled. Doing otherwise could cause underruns. > > */ > > - if (INTEL_GEN(dev_priv) <= 6 && r->spr_val) { > > + if (INTEL_GEN(i915) <= 6 && r->spr_val) { > > WARN_ON(wm_lp != 1); > > results->wm_lp_spr[wm_lp - 1] = WM1S_LP_EN | r->spr_val; > > } else > > @@ -3396,7 +3396,7 @@ static void ilk_compute_wm_results(struct drm_i915_private *dev_priv, > > } > > /* LP0 register values */ > > - for_each_intel_crtc(&dev_priv->drm, intel_crtc) { > > + for_each_intel_crtc(&i915->drm, intel_crtc) { > > enum pipe pipe = intel_crtc->pipe; > > const struct intel_wm_level *r = > > &intel_crtc->wm.active.ilk.wm[0]; > > @@ -3416,11 +3416,11 @@ static void ilk_compute_wm_results(struct drm_i915_private *dev_priv, > > /* Find the result with the highest level enabled. Check for enable_fbc_wm in > > * case both are at the same level. Prefer r1 in case they're the same. */ > > static struct intel_pipe_wm * > > -ilk_find_best_result(struct drm_i915_private *dev_priv, > > +ilk_find_best_result(struct drm_i915_private *i915, > > struct intel_pipe_wm *r1, > > struct intel_pipe_wm *r2) > > { > > - int level, max_level = ilk_wm_max_level(dev_priv); > > + int level, max_level = ilk_wm_max_level(i915); > > int level1 = 0, level2 = 0; > > for (level = 1; level <= max_level; level++) { > > @@ -3450,7 +3450,7 @@ ilk_find_best_result(struct drm_i915_private *dev_priv, > > #define WM_DIRTY_FBC (1 << 24) > > #define WM_DIRTY_DDB (1 << 25) > > -static unsigned int ilk_compute_wm_dirty(struct drm_i915_private *dev_priv, > > +static unsigned int ilk_compute_wm_dirty(struct drm_i915_private *i915, > > const struct ilk_wm_values *old, > > const struct ilk_wm_values *new) > > { > > @@ -3458,7 +3458,7 @@ static unsigned int ilk_compute_wm_dirty(struct drm_i915_private *dev_priv, > > enum pipe pipe; > > int wm_lp; > > - for_each_pipe(dev_priv, pipe) { > > + for_each_pipe(i915, pipe) { > > if (old->wm_linetime[pipe] != new->wm_linetime[pipe]) { > > dirty |= WM_DIRTY_LINETIME(pipe); > > /* Must disable LP1+ watermarks too */ > > @@ -3502,10 +3502,10 @@ static unsigned int ilk_compute_wm_dirty(struct drm_i915_private *dev_priv, > > return dirty; > > } > > -static bool _ilk_disable_lp_wm(struct drm_i915_private *dev_priv, > > +static bool _ilk_disable_lp_wm(struct drm_i915_private *i915, > > unsigned int dirty) > > { > > - struct ilk_wm_values *previous = &dev_priv->wm.hw; > > + struct ilk_wm_values *previous = &i915->wm.hw; > > bool changed = false; > > if (dirty & WM_DIRTY_LP(3) && previous->wm_lp[2] & WM1_LP_SR_EN) { > > @@ -3536,18 +3536,18 @@ static bool _ilk_disable_lp_wm(struct drm_i915_private *dev_priv, > > * The spec says we shouldn't write when we don't need, because every write > > * causes WMs to be re-evaluated, expending some power. > > */ > > -static void ilk_write_wm_values(struct drm_i915_private *dev_priv, > > +static void ilk_write_wm_values(struct drm_i915_private *i915, > > struct ilk_wm_values *results) > > { > > - struct ilk_wm_values *previous = &dev_priv->wm.hw; > > + struct ilk_wm_values *previous = &i915->wm.hw; > > unsigned int dirty; > > u32 val; > > - dirty = ilk_compute_wm_dirty(dev_priv, previous, results); > > + dirty = ilk_compute_wm_dirty(i915, previous, results); > > if (!dirty) > > return; > > - _ilk_disable_lp_wm(dev_priv, dirty); > > + _ilk_disable_lp_wm(i915, dirty); > > if (dirty & WM_DIRTY_PIPE(PIPE_A)) > > I915_WRITE(WM0_PIPEA_ILK, results->wm_pipe[0]); > > @@ -3564,7 +3564,7 @@ static void ilk_write_wm_values(struct drm_i915_private *dev_priv, > > I915_WRITE(PIPE_WM_LINETIME(PIPE_C), results->wm_linetime[2]); > > if (dirty & WM_DIRTY_DDB) { > > - if (IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv)) { > > + if (IS_HASWELL(i915) || IS_BROADWELL(i915)) { > > val = I915_READ(WM_MISC); > > if (results->partitioning == INTEL_DDB_PART_1_2) > > val &= ~WM_MISC_DATA_PARTITION_5_6; > > @@ -3594,7 +3594,7 @@ static void ilk_write_wm_values(struct drm_i915_private *dev_priv, > > previous->wm_lp_spr[0] != results->wm_lp_spr[0]) > > I915_WRITE(WM1S_LP_ILK, results->wm_lp_spr[0]); > > - if (INTEL_GEN(dev_priv) >= 7) { > > + if (INTEL_GEN(i915) >= 7) { > > if (dirty & WM_DIRTY_LP(2) && previous->wm_lp_spr[1] != results->wm_lp_spr[1]) > > I915_WRITE(WM2S_LP_IVB, results->wm_lp_spr[1]); > > if (dirty & WM_DIRTY_LP(3) && previous->wm_lp_spr[2] != results->wm_lp_spr[2]) > > @@ -3608,17 +3608,17 @@ static void ilk_write_wm_values(struct drm_i915_private *dev_priv, > > if (dirty & WM_DIRTY_LP(3) && previous->wm_lp[2] != results->wm_lp[2]) > > I915_WRITE(WM3_LP_ILK, results->wm_lp[2]); > > - dev_priv->wm.hw = *results; > > + i915->wm.hw = *results; > > } > > bool ilk_disable_lp_wm(struct drm_device *dev) > > { > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > - return _ilk_disable_lp_wm(dev_priv, WM_DIRTY_LP_ALL); > > + return _ilk_disable_lp_wm(i915, WM_DIRTY_LP_ALL); > > } > > -static u8 intel_enabled_dbuf_slices_num(struct drm_i915_private *dev_priv) > > +static u8 intel_enabled_dbuf_slices_num(struct drm_i915_private *i915) > > { > > u8 enabled_slices; > > @@ -3626,7 +3626,7 @@ static u8 intel_enabled_dbuf_slices_num(struct drm_i915_private *dev_priv) > > enabled_slices = 1; > > /* Gen prior to GEN11 have only one DBuf slice */ > > - if (INTEL_GEN(dev_priv) < 11) > > + if (INTEL_GEN(i915) < 11) > > return enabled_slices; > > /* > > @@ -3644,16 +3644,16 @@ static u8 intel_enabled_dbuf_slices_num(struct drm_i915_private *dev_priv) > > * FIXME: We still don't have the proper code detect if we need to apply the WA, > > * so assume we'll always need it in order to avoid underruns. > > */ > > -static bool skl_needs_memory_bw_wa(struct drm_i915_private *dev_priv) > > +static bool skl_needs_memory_bw_wa(struct drm_i915_private *i915) > > { > > - return IS_GEN9_BC(dev_priv) || IS_BROXTON(dev_priv); > > + return IS_GEN9_BC(i915) || IS_BROXTON(i915); > > } > > static bool > > -intel_has_sagv(struct drm_i915_private *dev_priv) > > +intel_has_sagv(struct drm_i915_private *i915) > > { > > - return (IS_GEN9_BC(dev_priv) || INTEL_GEN(dev_priv) >= 10) && > > - dev_priv->sagv_status != I915_SAGV_NOT_CONTROLLED; > > + return (IS_GEN9_BC(i915) || INTEL_GEN(i915) >= 10) && > > + i915->sagv_status != I915_SAGV_NOT_CONTROLLED; > > } > > /* > > @@ -3668,18 +3668,18 @@ intel_has_sagv(struct drm_i915_private *dev_priv) > > * - We're not using an interlaced display configuration > > */ > > int > > -intel_enable_sagv(struct drm_i915_private *dev_priv) > > +intel_enable_sagv(struct drm_i915_private *i915) > > { > > int ret; > > - if (!intel_has_sagv(dev_priv)) > > + if (!intel_has_sagv(i915)) > > return 0; > > - if (dev_priv->sagv_status == I915_SAGV_ENABLED) > > + if (i915->sagv_status == I915_SAGV_ENABLED) > > return 0; > > DRM_DEBUG_KMS("Enabling SAGV\n"); > > - ret = sandybridge_pcode_write(dev_priv, GEN9_PCODE_SAGV_CONTROL, > > + ret = sandybridge_pcode_write(i915, GEN9_PCODE_SAGV_CONTROL, > > GEN9_SAGV_ENABLE); > > /* We don't need to wait for SAGV when enabling */ > > @@ -3688,33 +3688,33 @@ intel_enable_sagv(struct drm_i915_private *dev_priv) > > * Some skl systems, pre-release machines in particular, > > * don't actually have SAGV. > > */ > > - if (IS_SKYLAKE(dev_priv) && ret == -ENXIO) { > > + if (IS_SKYLAKE(i915) && ret == -ENXIO) { > > DRM_DEBUG_DRIVER("No SAGV found on system, ignoring\n"); > > - dev_priv->sagv_status = I915_SAGV_NOT_CONTROLLED; > > + i915->sagv_status = I915_SAGV_NOT_CONTROLLED; > > return 0; > > } else if (ret < 0) { > > DRM_ERROR("Failed to enable SAGV\n"); > > return ret; > > } > > - dev_priv->sagv_status = I915_SAGV_ENABLED; > > + i915->sagv_status = I915_SAGV_ENABLED; > > return 0; > > } > > int > > -intel_disable_sagv(struct drm_i915_private *dev_priv) > > +intel_disable_sagv(struct drm_i915_private *i915) > > { > > int ret; > > - if (!intel_has_sagv(dev_priv)) > > + if (!intel_has_sagv(i915)) > > return 0; > > - if (dev_priv->sagv_status == I915_SAGV_DISABLED) > > + if (i915->sagv_status == I915_SAGV_DISABLED) > > return 0; > > DRM_DEBUG_KMS("Disabling SAGV\n"); > > /* bspec says to keep retrying for at least 1 ms */ > > - ret = skl_pcode_request(dev_priv, GEN9_PCODE_SAGV_CONTROL, > > + ret = skl_pcode_request(i915, GEN9_PCODE_SAGV_CONTROL, > > GEN9_SAGV_DISABLE, > > GEN9_SAGV_IS_DISABLED, GEN9_SAGV_IS_DISABLED, > > 1); > > @@ -3722,23 +3722,23 @@ intel_disable_sagv(struct drm_i915_private *dev_priv) > > * Some skl systems, pre-release machines in particular, > > * don't actually have SAGV. > > */ > > - if (IS_SKYLAKE(dev_priv) && ret == -ENXIO) { > > + if (IS_SKYLAKE(i915) && ret == -ENXIO) { > > DRM_DEBUG_DRIVER("No SAGV found on system, ignoring\n"); > > - dev_priv->sagv_status = I915_SAGV_NOT_CONTROLLED; > > + i915->sagv_status = I915_SAGV_NOT_CONTROLLED; > > return 0; > > } else if (ret < 0) { > > DRM_ERROR("Failed to disable SAGV (%d)\n", ret); > > return ret; > > } > > - dev_priv->sagv_status = I915_SAGV_DISABLED; > > + i915->sagv_status = I915_SAGV_DISABLED; > > return 0; > > } > > bool intel_can_enable_sagv(struct drm_atomic_state *state) > > { > > struct drm_device *dev = state->dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > struct intel_atomic_state *intel_state = to_intel_atomic_state(state); > > struct intel_crtc *crtc; > > struct intel_plane *plane; > > @@ -3747,12 +3747,12 @@ bool intel_can_enable_sagv(struct drm_atomic_state *state) > > int level, latency; > > int sagv_block_time_us; > > - if (!intel_has_sagv(dev_priv)) > > + if (!intel_has_sagv(i915)) > > return false; > > - if (IS_GEN(dev_priv, 9)) > > + if (IS_GEN(i915, 9)) > > sagv_block_time_us = 30; > > - else if (IS_GEN(dev_priv, 10)) > > + else if (IS_GEN(i915, 10)) > > sagv_block_time_us = 20; > > else > > sagv_block_time_us = 10; > > @@ -3772,7 +3772,7 @@ bool intel_can_enable_sagv(struct drm_atomic_state *state) > > /* Since we're now guaranteed to only have one active CRTC... */ > > pipe = ffs(intel_state->active_crtcs) - 1; > > - crtc = intel_get_crtc_for_pipe(dev_priv, pipe); > > + crtc = intel_get_crtc_for_pipe(i915, pipe); > > cstate = to_intel_crtc_state(crtc->base.state); > > if (crtc->base.state->adjusted_mode.flags & DRM_MODE_FLAG_INTERLACE) > > @@ -3787,13 +3787,13 @@ bool intel_can_enable_sagv(struct drm_atomic_state *state) > > continue; > > /* Find the highest enabled wm level for this plane */ > > - for (level = ilk_wm_max_level(dev_priv); > > + for (level = ilk_wm_max_level(i915); > > !wm->wm[level].plane_en; --level) > > { } > > - latency = dev_priv->wm.skl_latency[level]; > > + latency = i915->wm.skl_latency[level]; > > - if (skl_needs_memory_bw_wa(dev_priv) && > > + if (skl_needs_memory_bw_wa(i915) && > > plane->base.state->fb->modifier == > > I915_FORMAT_MOD_X_TILED) > > latency += 15; > > @@ -3810,7 +3810,7 @@ bool intel_can_enable_sagv(struct drm_atomic_state *state) > > return true; > > } > > -static u16 intel_get_ddb_size(struct drm_i915_private *dev_priv, > > +static u16 intel_get_ddb_size(struct drm_i915_private *i915, > > const struct intel_crtc_state *cstate, > > const u64 total_data_rate, > > const int num_active, > > @@ -3818,11 +3818,11 @@ static u16 intel_get_ddb_size(struct drm_i915_private *dev_priv, > > { > > const struct drm_display_mode *adjusted_mode; > > u64 total_data_bw; > > - u16 ddb_size = INTEL_INFO(dev_priv)->ddb_size; > > + u16 ddb_size = INTEL_INFO(i915)->ddb_size; > > WARN_ON(ddb_size == 0); > > - if (INTEL_GEN(dev_priv) < 11) > > + if (INTEL_GEN(i915) < 11) > > return ddb_size - 4; /* 4 blocks for bypass path allocation */ > > adjusted_mode = &cstate->base.adjusted_mode; > > @@ -3847,7 +3847,7 @@ static u16 intel_get_ddb_size(struct drm_i915_private *dev_priv, > > } > > static void > > -skl_ddb_get_pipe_allocation_limits(struct drm_i915_private *dev_priv, > > +skl_ddb_get_pipe_allocation_limits(struct drm_i915_private *i915, > > const struct intel_crtc_state *cstate, > > const u64 total_data_rate, > > struct skl_ddb_allocation *ddb, > > @@ -3867,16 +3867,16 @@ skl_ddb_get_pipe_allocation_limits(struct drm_i915_private *dev_priv, > > if (WARN_ON(!state) || !cstate->base.active) { > > alloc->start = 0; > > alloc->end = 0; > > - *num_active = hweight32(dev_priv->active_crtcs); > > + *num_active = hweight32(i915->active_crtcs); > > return; > > } > > if (intel_state->active_pipe_changes) > > *num_active = hweight32(intel_state->active_crtcs); > > else > > - *num_active = hweight32(dev_priv->active_crtcs); > > + *num_active = hweight32(i915->active_crtcs); > > - ddb_size = intel_get_ddb_size(dev_priv, cstate, total_data_rate, > > + ddb_size = intel_get_ddb_size(i915, cstate, total_data_rate, > > *num_active, ddb); > > /* > > @@ -3939,8 +3939,8 @@ static unsigned int > > skl_cursor_allocation(const struct intel_crtc_state *crtc_state, > > int num_active) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc_state->base.crtc->dev); > > - int level, max_level = ilk_wm_max_level(dev_priv); > > + struct drm_i915_private *i915 = to_i915(crtc_state->base.crtc->dev); > > + int level, max_level = ilk_wm_max_level(i915); > > struct skl_wm_level wm = {}; > > int ret, min_ddb_alloc = 0; > > struct skl_wm_params wp; > > @@ -3963,7 +3963,7 @@ skl_cursor_allocation(const struct intel_crtc_state *crtc_state, > > return max(num_active == 1 ? 32 : 8, min_ddb_alloc); > > } > > -static void skl_ddb_entry_init_from_hw(struct drm_i915_private *dev_priv, > > +static void skl_ddb_entry_init_from_hw(struct drm_i915_private *i915, > > struct skl_ddb_entry *entry, u32 reg) > > { > > @@ -3975,7 +3975,7 @@ static void skl_ddb_entry_init_from_hw(struct drm_i915_private *dev_priv, > > } > > static void > > -skl_ddb_get_hw_plane_state(struct drm_i915_private *dev_priv, > > +skl_ddb_get_hw_plane_state(struct drm_i915_private *i915, > > const enum pipe pipe, > > const enum plane_id plane_id, > > struct skl_ddb_entry *ddb_y, > > @@ -3987,7 +3987,7 @@ skl_ddb_get_hw_plane_state(struct drm_i915_private *dev_priv, > > /* Cursor doesn't support NV12/planar, so no extra calculation needed */ > > if (plane_id == PLANE_CURSOR) { > > val = I915_READ(CUR_BUF_CFG(pipe)); > > - skl_ddb_entry_init_from_hw(dev_priv, ddb_y, val); > > + skl_ddb_entry_init_from_hw(i915, ddb_y, val); > > return; > > } > > @@ -3999,9 +3999,9 @@ skl_ddb_get_hw_plane_state(struct drm_i915_private *dev_priv, > > val & PLANE_CTL_ORDER_RGBX, > > val & PLANE_CTL_ALPHA_MASK); > > - if (INTEL_GEN(dev_priv) >= 11) { > > + if (INTEL_GEN(i915) >= 11) { > > val = I915_READ(PLANE_BUF_CFG(pipe, plane_id)); > > - skl_ddb_entry_init_from_hw(dev_priv, ddb_y, val); > > + skl_ddb_entry_init_from_hw(i915, ddb_y, val); > > } else { > > val = I915_READ(PLANE_BUF_CFG(pipe, plane_id)); > > val2 = I915_READ(PLANE_NV12_BUF_CFG(pipe, plane_id)); > > @@ -4009,8 +4009,8 @@ skl_ddb_get_hw_plane_state(struct drm_i915_private *dev_priv, > > if (is_planar_yuv_format(fourcc)) > > swap(val, val2); > > - skl_ddb_entry_init_from_hw(dev_priv, ddb_y, val); > > - skl_ddb_entry_init_from_hw(dev_priv, ddb_uv, val2); > > + skl_ddb_entry_init_from_hw(i915, ddb_y, val); > > + skl_ddb_entry_init_from_hw(i915, ddb_uv, val2); > > } > > } > > @@ -4018,30 +4018,30 @@ void skl_pipe_ddb_get_hw_state(struct intel_crtc *crtc, > > struct skl_ddb_entry *ddb_y, > > struct skl_ddb_entry *ddb_uv) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > enum intel_display_power_domain power_domain; > > enum pipe pipe = crtc->pipe; > > intel_wakeref_t wakeref; > > enum plane_id plane_id; > > power_domain = POWER_DOMAIN_PIPE(pipe); > > - wakeref = intel_display_power_get_if_enabled(dev_priv, power_domain); > > + wakeref = intel_display_power_get_if_enabled(i915, power_domain); > > if (!wakeref) > > return; > > for_each_plane_id_on_crtc(crtc, plane_id) > > - skl_ddb_get_hw_plane_state(dev_priv, pipe, > > + skl_ddb_get_hw_plane_state(i915, pipe, > > plane_id, > > &ddb_y[plane_id], > > &ddb_uv[plane_id]); > > - intel_display_power_put(dev_priv, power_domain, wakeref); > > + intel_display_power_put(i915, power_domain, wakeref); > > } > > -void skl_ddb_get_hw_state(struct drm_i915_private *dev_priv, > > +void skl_ddb_get_hw_state(struct drm_i915_private *i915, > > struct skl_ddb_allocation *ddb /* out */) > > { > > - ddb->enabled_slices = intel_enabled_dbuf_slices_num(dev_priv); > > + ddb->enabled_slices = intel_enabled_dbuf_slices_num(i915); > > } > > /* > > @@ -4138,7 +4138,7 @@ skl_pipe_downscale_amount(const struct intel_crtc_state *crtc_state) > > int skl_check_pipe_max_pixel_rate(struct intel_crtc *intel_crtc, > > struct intel_crtc_state *cstate) > > { > > - struct drm_i915_private *dev_priv = to_i915(intel_crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(intel_crtc->base.dev); > > struct drm_crtc_state *crtc_state = &cstate->base; > > struct drm_atomic_state *state = crtc_state->state; > > struct drm_plane *plane; > > @@ -4181,7 +4181,7 @@ int skl_check_pipe_max_pixel_rate(struct intel_crtc *intel_crtc, > > crtc_clock = crtc_state->adjusted_mode.crtc_clock; > > dotclk = to_intel_atomic_state(state)->cdclk.logical.cdclk; > > - if (IS_GEMINILAKE(dev_priv) || INTEL_GEN(dev_priv) >= 10) > > + if (IS_GEMINILAKE(i915) || INTEL_GEN(i915) >= 10) > > dotclk *= 2; > > pipe_max_pixel_rate = div_round_up_u32_fixed16(dotclk, pipe_downscale); > > @@ -4341,7 +4341,7 @@ skl_allocate_pipe_ddb(struct intel_crtc_state *cstate, > > { > > struct drm_atomic_state *state = cstate->base.state; > > struct drm_crtc *crtc = cstate->base.crtc; > > - struct drm_i915_private *dev_priv = to_i915(crtc->dev); > > + struct drm_i915_private *i915 = to_i915(crtc->dev); > > struct intel_crtc *intel_crtc = to_intel_crtc(crtc); > > struct skl_ddb_entry *alloc = &cstate->wm.skl.ddb; > > u16 alloc_size, start = 0; > > @@ -4367,7 +4367,7 @@ skl_allocate_pipe_ddb(struct intel_crtc_state *cstate, > > return 0; > > } > > - if (INTEL_GEN(dev_priv) >= 11) > > + if (INTEL_GEN(i915) >= 11) > > total_data_rate = > > icl_get_total_relative_data_rate(cstate, > > plane_data_rate); > > @@ -4378,7 +4378,7 @@ skl_allocate_pipe_ddb(struct intel_crtc_state *cstate, > > uv_plane_data_rate); > > - skl_ddb_get_pipe_allocation_limits(dev_priv, cstate, total_data_rate, > > + skl_ddb_get_pipe_allocation_limits(i915, cstate, total_data_rate, > > ddb, alloc, &num_active); > > alloc_size = skl_ddb_entry_size(alloc); > > if (alloc_size == 0) > > @@ -4398,7 +4398,7 @@ skl_allocate_pipe_ddb(struct intel_crtc_state *cstate, > > * Find the highest watermark level for which we can satisfy the block > > * requirement of active planes. > > */ > > - for (level = ilk_wm_max_level(dev_priv); level >= 0; level--) { > > + for (level = ilk_wm_max_level(i915); level >= 0; level--) { > > blocks = 0; > > for_each_plane_id_on_crtc(intel_crtc, plane_id) { > > const struct skl_plane_wm *wm = > > @@ -4484,7 +4484,7 @@ skl_allocate_pipe_ddb(struct intel_crtc_state *cstate, > > continue; > > /* Gen11+ uses a separate plane for UV watermarks */ > > - WARN_ON(INTEL_GEN(dev_priv) >= 11 && uv_total[plane_id]); > > + WARN_ON(INTEL_GEN(i915) >= 11 && uv_total[plane_id]); > > /* Leave disabled planes at (0,0) */ > > if (total[plane_id]) { > > @@ -4506,7 +4506,7 @@ skl_allocate_pipe_ddb(struct intel_crtc_state *cstate, > > * all levels as "enabled." Go back now and disable the ones > > * that aren't actually possible. > > */ > > - for (level++; level <= ilk_wm_max_level(dev_priv); level++) { > > + for (level++; level <= ilk_wm_max_level(i915); level++) { > > for_each_plane_id_on_crtc(intel_crtc, plane_id) { > > struct skl_plane_wm *wm = > > &cstate->wm.skl.optimal.planes[plane_id]; > > @@ -4531,7 +4531,7 @@ skl_allocate_pipe_ddb(struct intel_crtc_state *cstate, > > * Wa_1408961008:icl, ehl > > * Underruns with WM1+ disabled > > */ > > - if (IS_GEN(dev_priv, 11) && > > + if (IS_GEN(i915, 11) && > > level == 1 && wm->wm[0].plane_en) { > > wm->wm[level].plane_res_b = wm->wm[0].plane_res_b; > > wm->wm[level].plane_res_l = wm->wm[0].plane_res_l; > > @@ -4562,7 +4562,7 @@ skl_allocate_pipe_ddb(struct intel_crtc_state *cstate, > > * 2xcdclk is 1350 MHz and the pixel rate should never exceed that. > > */ > > static uint_fixed_16_16_t > > -skl_wm_method1(const struct drm_i915_private *dev_priv, u32 pixel_rate, > > +skl_wm_method1(const struct drm_i915_private *i915, u32 pixel_rate, > > u8 cpp, u32 latency, u32 dbuf_block_size) > > { > > u32 wm_intermediate_val; > > @@ -4574,7 +4574,7 @@ skl_wm_method1(const struct drm_i915_private *dev_priv, u32 pixel_rate, > > wm_intermediate_val = latency * pixel_rate * cpp; > > ret = div_fixed16(wm_intermediate_val, 1000 * dbuf_block_size); > > - if (INTEL_GEN(dev_priv) >= 10) > > + if (INTEL_GEN(i915) >= 10) > > ret = add_fixed16_u32(ret, 1); > > return ret; > > @@ -4648,7 +4648,7 @@ skl_compute_wm_params(const struct intel_crtc_state *crtc_state, > > int color_plane) > > { > > struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc); > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > u32 interm_pbpl; > > /* only planar format has two planes */ > > @@ -4673,7 +4673,7 @@ skl_compute_wm_params(const struct intel_crtc_state *crtc_state, > > wp->cpp = format->cpp[color_plane]; > > wp->plane_pixel_rate = plane_pixel_rate; > > - if (INTEL_GEN(dev_priv) >= 11 && > > + if (INTEL_GEN(i915) >= 11 && > > modifier == I915_FORMAT_MOD_Yf_TILED && wp->cpp == 1) > > wp->dbuf_block_size = 256; > > else > > @@ -4698,7 +4698,7 @@ skl_compute_wm_params(const struct intel_crtc_state *crtc_state, > > wp->y_min_scanlines = 4; > > } > > - if (skl_needs_memory_bw_wa(dev_priv)) > > + if (skl_needs_memory_bw_wa(i915)) > > wp->y_min_scanlines *= 2; > > wp->plane_bytes_per_line = wp->width * wp->cpp; > > @@ -4707,12 +4707,12 @@ skl_compute_wm_params(const struct intel_crtc_state *crtc_state, > > wp->y_min_scanlines, > > wp->dbuf_block_size); > > - if (INTEL_GEN(dev_priv) >= 10) > > + if (INTEL_GEN(i915) >= 10) > > interm_pbpl++; > > wp->plane_blocks_per_line = div_fixed16(interm_pbpl, > > wp->y_min_scanlines); > > - } else if (wp->x_tiled && IS_GEN(dev_priv, 9)) { > > + } else if (wp->x_tiled && IS_GEN(i915, 9)) { > > interm_pbpl = DIV_ROUND_UP(wp->plane_bytes_per_line, > > wp->dbuf_block_size); > > wp->plane_blocks_per_line = u32_to_fixed16(interm_pbpl); > > @@ -4758,9 +4758,9 @@ skl_compute_plane_wm_params(const struct intel_crtc_state *crtc_state, > > wp, color_plane); > > } > > -static bool skl_wm_has_lines(struct drm_i915_private *dev_priv, int level) > > +static bool skl_wm_has_lines(struct drm_i915_private *i915, int level) > > { > > - if (INTEL_GEN(dev_priv) >= 10 || IS_GEMINILAKE(dev_priv)) > > + if (INTEL_GEN(i915) >= 10 || IS_GEMINILAKE(i915)) > > return true; > > /* The number of lines are ignored for the level 0 watermark. */ > > @@ -4773,8 +4773,8 @@ static void skl_compute_plane_wm(const struct intel_crtc_state *cstate, > > const struct skl_wm_level *result_prev, > > struct skl_wm_level *result /* out */) > > { > > - struct drm_i915_private *dev_priv = to_i915(cstate->base.crtc->dev); > > - u32 latency = dev_priv->wm.skl_latency[level]; > > + struct drm_i915_private *i915 = to_i915(cstate->base.crtc->dev); > > + u32 latency = i915->wm.skl_latency[level]; > > uint_fixed_16_16_t method1, method2; > > uint_fixed_16_16_t selected_result; > > u32 res_blocks, res_lines, min_ddb_alloc = 0; > > @@ -4789,14 +4789,14 @@ static void skl_compute_plane_wm(const struct intel_crtc_state *cstate, > > * WaIncreaseLatencyIPCEnabled: kbl,cfl > > * Display WA #1141: kbl,cfl > > */ > > - if ((IS_KABYLAKE(dev_priv) || IS_COFFEELAKE(dev_priv)) || > > - dev_priv->ipc_enabled) > > + if ((IS_KABYLAKE(i915) || IS_COFFEELAKE(i915)) || > > + i915->ipc_enabled) > > latency += 4; > > - if (skl_needs_memory_bw_wa(dev_priv) && wp->x_tiled) > > + if (skl_needs_memory_bw_wa(i915) && wp->x_tiled) > > latency += 15; > > - method1 = skl_wm_method1(dev_priv, wp->plane_pixel_rate, > > + method1 = skl_wm_method1(i915, wp->plane_pixel_rate, > > wp->cpp, latency, wp->dbuf_block_size); > > method2 = skl_wm_method2(wp->plane_pixel_rate, > > cstate->base.adjusted_mode.crtc_htotal, > > @@ -4811,8 +4811,8 @@ static void skl_compute_plane_wm(const struct intel_crtc_state *cstate, > > (wp->plane_bytes_per_line / wp->dbuf_block_size < 1)) { > > selected_result = method2; > > } else if (latency >= wp->linetime_us) { > > - if (IS_GEN(dev_priv, 9) && > > - !IS_GEMINILAKE(dev_priv)) > > + if (IS_GEN(i915, 9) && > > + !IS_GEMINILAKE(i915)) > > selected_result = min_fixed16(method1, method2); > > else > > selected_result = method2; > > @@ -4825,7 +4825,7 @@ static void skl_compute_plane_wm(const struct intel_crtc_state *cstate, > > res_lines = div_round_up_fixed16(selected_result, > > wp->plane_blocks_per_line); > > - if (IS_GEN9_BC(dev_priv) || IS_BROXTON(dev_priv)) { > > + if (IS_GEN9_BC(i915) || IS_BROXTON(i915)) { > > /* Display WA #1125: skl,bxt,kbl */ > > if (level == 0 && wp->rc_surface) > > res_blocks += > > @@ -4852,7 +4852,7 @@ static void skl_compute_plane_wm(const struct intel_crtc_state *cstate, > > } > > } > > - if (INTEL_GEN(dev_priv) >= 11) { > > + if (INTEL_GEN(i915) >= 11) { > > if (wp->y_tiled) { > > int extra_lines; > > @@ -4870,7 +4870,7 @@ static void skl_compute_plane_wm(const struct intel_crtc_state *cstate, > > } > > } > > - if (!skl_wm_has_lines(dev_priv, level)) > > + if (!skl_wm_has_lines(i915, level)) > > res_lines = 0; > > if (res_lines > 31) { > > @@ -4897,8 +4897,8 @@ skl_compute_wm_levels(const struct intel_crtc_state *cstate, > > const struct skl_wm_params *wm_params, > > struct skl_wm_level *levels) > > { > > - struct drm_i915_private *dev_priv = to_i915(cstate->base.crtc->dev); > > - int level, max_level = ilk_wm_max_level(dev_priv); > > + struct drm_i915_private *i915 = to_i915(cstate->base.crtc->dev); > > + int level, max_level = ilk_wm_max_level(i915); > > struct skl_wm_level *result_prev = &levels[0]; > > for (level = 0; level <= max_level; level++) { > > @@ -4915,7 +4915,7 @@ static u32 > > skl_compute_linetime_wm(const struct intel_crtc_state *cstate) > > { > > struct drm_atomic_state *state = cstate->base.state; > > - struct drm_i915_private *dev_priv = to_i915(state->dev); > > + struct drm_i915_private *i915 = to_i915(state->dev); > > uint_fixed_16_16_t linetime_us; > > u32 linetime_wm; > > @@ -4923,7 +4923,7 @@ skl_compute_linetime_wm(const struct intel_crtc_state *cstate) > > linetime_wm = fixed16_to_u32_round_up(mul_u32_fixed16(8, linetime_us)); > > /* Display WA #1135: BXT:ALL GLK:ALL */ > > - if (IS_GEN9_LP(dev_priv) && dev_priv->ipc_enabled) > > + if (IS_GEN9_LP(i915) && i915->ipc_enabled) > > linetime_wm /= 2; > > return linetime_wm; > > @@ -4934,21 +4934,21 @@ static void skl_compute_transition_wm(const struct intel_crtc_state *cstate, > > struct skl_plane_wm *wm) > > { > > struct drm_device *dev = cstate->base.crtc->dev; > > - const struct drm_i915_private *dev_priv = to_i915(dev); > > + const struct drm_i915_private *i915 = to_i915(dev); > > u16 trans_min, trans_y_tile_min; > > const u16 trans_amount = 10; /* This is configurable amount */ > > u16 wm0_sel_res_b, trans_offset_b, res_blocks; > > /* Transition WM are not recommended by HW team for GEN9 */ > > - if (INTEL_GEN(dev_priv) <= 9) > > + if (INTEL_GEN(i915) <= 9) > > return; > > /* Transition WM don't make any sense if ipc is disabled */ > > - if (!dev_priv->ipc_enabled) > > + if (!i915->ipc_enabled) > > return; > > trans_min = 14; > > - if (INTEL_GEN(dev_priv) >= 11) > > + if (INTEL_GEN(i915) >= 11) > > trans_min = 4; > > trans_offset_b = trans_min + trans_amount; > > @@ -4974,7 +4974,7 @@ static void skl_compute_transition_wm(const struct intel_crtc_state *cstate, > > res_blocks = wm0_sel_res_b + trans_offset_b; > > /* WA BUG:1938466 add one block for non y-tile planes */ > > - if (IS_CNL_REVID(dev_priv, CNL_REVID_A0, CNL_REVID_A0)) > > + if (IS_CNL_REVID(i915, CNL_REVID_A0, CNL_REVID_A0)) > > res_blocks += 1; > > } > > @@ -5093,7 +5093,7 @@ static int icl_build_plane_wm(struct intel_crtc_state *crtc_state, > > static int skl_build_pipe_wm(struct intel_crtc_state *cstate) > > { > > - struct drm_i915_private *dev_priv = to_i915(cstate->base.crtc->dev); > > + struct drm_i915_private *i915 = to_i915(cstate->base.crtc->dev); > > struct skl_pipe_wm *pipe_wm = &cstate->wm.skl.optimal; > > struct drm_crtc_state *crtc_state = &cstate->base; > > struct drm_plane *plane; > > @@ -5110,7 +5110,7 @@ static int skl_build_pipe_wm(struct intel_crtc_state *cstate) > > const struct intel_plane_state *intel_pstate = > > to_intel_plane_state(pstate); > > - if (INTEL_GEN(dev_priv) >= 11) > > + if (INTEL_GEN(i915) >= 11) > > ret = icl_build_plane_wm(cstate, intel_pstate); > > else > > ret = skl_build_plane_wm(cstate, intel_pstate); > > @@ -5123,7 +5123,7 @@ static int skl_build_pipe_wm(struct intel_crtc_state *cstate) > > return 0; > > } > > -static void skl_ddb_entry_write(struct drm_i915_private *dev_priv, > > +static void skl_ddb_entry_write(struct drm_i915_private *i915, > > i915_reg_t reg, > > const struct skl_ddb_entry *entry) > > { > > @@ -5133,7 +5133,7 @@ static void skl_ddb_entry_write(struct drm_i915_private *dev_priv, > > I915_WRITE_FW(reg, 0); > > } > > -static void skl_write_wm_level(struct drm_i915_private *dev_priv, > > +static void skl_write_wm_level(struct drm_i915_private *i915, > > i915_reg_t reg, > > const struct skl_wm_level *level) > > { > > @@ -5152,8 +5152,8 @@ static void skl_write_wm_level(struct drm_i915_private *dev_priv, > > void skl_write_plane_wm(struct intel_plane *plane, > > const struct intel_crtc_state *crtc_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(plane->base.dev); > > - int level, max_level = ilk_wm_max_level(dev_priv); > > + struct drm_i915_private *i915 = to_i915(plane->base.dev); > > + int level, max_level = ilk_wm_max_level(i915); > > enum plane_id plane_id = plane->id; > > enum pipe pipe = plane->pipe; > > const struct skl_plane_wm *wm = > > @@ -5164,14 +5164,14 @@ void skl_write_plane_wm(struct intel_plane *plane, > > &crtc_state->wm.skl.plane_ddb_uv[plane_id]; > > for (level = 0; level <= max_level; level++) { > > - skl_write_wm_level(dev_priv, PLANE_WM(pipe, plane_id, level), > > + skl_write_wm_level(i915, PLANE_WM(pipe, plane_id, level), > > &wm->wm[level]); > > } > > - skl_write_wm_level(dev_priv, PLANE_WM_TRANS(pipe, plane_id), > > + skl_write_wm_level(i915, PLANE_WM_TRANS(pipe, plane_id), > > &wm->trans_wm); > > - if (INTEL_GEN(dev_priv) >= 11) { > > - skl_ddb_entry_write(dev_priv, > > + if (INTEL_GEN(i915) >= 11) { > > + skl_ddb_entry_write(i915, > > PLANE_BUF_CFG(pipe, plane_id), ddb_y); > > return; > > } > > @@ -5179,17 +5179,17 @@ void skl_write_plane_wm(struct intel_plane *plane, > > if (wm->is_planar) > > swap(ddb_y, ddb_uv); > > - skl_ddb_entry_write(dev_priv, > > + skl_ddb_entry_write(i915, > > PLANE_BUF_CFG(pipe, plane_id), ddb_y); > > - skl_ddb_entry_write(dev_priv, > > + skl_ddb_entry_write(i915, > > PLANE_NV12_BUF_CFG(pipe, plane_id), ddb_uv); > > } > > void skl_write_cursor_wm(struct intel_plane *plane, > > const struct intel_crtc_state *crtc_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(plane->base.dev); > > - int level, max_level = ilk_wm_max_level(dev_priv); > > + struct drm_i915_private *i915 = to_i915(plane->base.dev); > > + int level, max_level = ilk_wm_max_level(i915); > > enum plane_id plane_id = plane->id; > > enum pipe pipe = plane->pipe; > > const struct skl_plane_wm *wm = > > @@ -5198,12 +5198,12 @@ void skl_write_cursor_wm(struct intel_plane *plane, > > &crtc_state->wm.skl.plane_ddb_y[plane_id]; > > for (level = 0; level <= max_level; level++) { > > - skl_write_wm_level(dev_priv, CUR_WM(pipe, level), > > + skl_write_wm_level(i915, CUR_WM(pipe, level), > > &wm->wm[level]); > > } > > - skl_write_wm_level(dev_priv, CUR_WM_TRANS(pipe), &wm->trans_wm); > > + skl_write_wm_level(i915, CUR_WM_TRANS(pipe), &wm->trans_wm); > > - skl_ddb_entry_write(dev_priv, CUR_BUF_CFG(pipe), ddb); > > + skl_ddb_entry_write(i915, CUR_BUF_CFG(pipe), ddb); > > } > > bool skl_wm_level_equals(const struct skl_wm_level *l1, > > @@ -5215,11 +5215,11 @@ bool skl_wm_level_equals(const struct skl_wm_level *l1, > > l1->plane_res_b == l2->plane_res_b; > > } > > -static bool skl_plane_wm_equals(struct drm_i915_private *dev_priv, > > +static bool skl_plane_wm_equals(struct drm_i915_private *i915, > > const struct skl_plane_wm *wm1, > > const struct skl_plane_wm *wm2) > > { > > - int level, max_level = ilk_wm_max_level(dev_priv); > > + int level, max_level = ilk_wm_max_level(i915); > > for (level = 0; level <= max_level; level++) { > > if (!skl_wm_level_equals(&wm1->wm[level], &wm2->wm[level]) || > > @@ -5234,11 +5234,11 @@ static bool skl_pipe_wm_equals(struct intel_crtc *crtc, > > const struct skl_pipe_wm *wm1, > > const struct skl_pipe_wm *wm2) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > enum plane_id plane_id; > > for_each_plane_id_on_crtc(crtc, plane_id) { > > - if (!skl_plane_wm_equals(dev_priv, > > + if (!skl_plane_wm_equals(i915, > > &wm1->planes[plane_id], > > &wm2->planes[plane_id])) > > return false; > > @@ -5287,10 +5287,10 @@ skl_ddb_add_affected_planes(const struct intel_crtc_state *old_crtc_state, > > { > > struct intel_atomic_state *state = to_intel_atomic_state(new_crtc_state->base.state); > > struct intel_crtc *crtc = to_intel_crtc(new_crtc_state->base.crtc); > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > struct intel_plane *plane; > > - for_each_intel_plane_on_crtc(&dev_priv->drm, crtc, plane) { > > + for_each_intel_plane_on_crtc(&i915->drm, crtc, plane) { > > struct intel_plane_state *plane_state; > > enum plane_id plane_id = plane->id; > > @@ -5313,14 +5313,14 @@ skl_ddb_add_affected_planes(const struct intel_crtc_state *old_crtc_state, > > static int > > skl_compute_ddb(struct intel_atomic_state *state) > > { > > - const struct drm_i915_private *dev_priv = to_i915(state->base.dev); > > + const struct drm_i915_private *i915 = to_i915(state->base.dev); > > struct skl_ddb_allocation *ddb = &state->wm_results.ddb; > > struct intel_crtc_state *old_crtc_state; > > struct intel_crtc_state *new_crtc_state; > > struct intel_crtc *crtc; > > int ret, i; > > - memcpy(ddb, &dev_priv->wm.skl_hw.ddb, sizeof(*ddb)); > > + memcpy(ddb, &i915->wm.skl_hw.ddb, sizeof(*ddb)); > > for_each_oldnew_intel_crtc_in_state(state, crtc, old_crtc_state, > > new_crtc_state, i) { > > @@ -5345,7 +5345,7 @@ static char enast(bool enable) > > static void > > skl_print_wm_changes(struct intel_atomic_state *state) > > { > > - struct drm_i915_private *dev_priv = to_i915(state->base.dev); > > + struct drm_i915_private *i915 = to_i915(state->base.dev); > > const struct intel_crtc_state *old_crtc_state; > > const struct intel_crtc_state *new_crtc_state; > > struct intel_plane *plane; > > @@ -5362,7 +5362,7 @@ skl_print_wm_changes(struct intel_atomic_state *state) > > old_pipe_wm = &old_crtc_state->wm.skl.optimal; > > new_pipe_wm = &new_crtc_state->wm.skl.optimal; > > - for_each_intel_plane_on_crtc(&dev_priv->drm, crtc, plane) { > > + for_each_intel_plane_on_crtc(&i915->drm, crtc, plane) { > > enum plane_id plane_id = plane->id; > > const struct skl_ddb_entry *old, *new; > > @@ -5378,14 +5378,14 @@ skl_print_wm_changes(struct intel_atomic_state *state) > > skl_ddb_entry_size(old), skl_ddb_entry_size(new)); > > } > > - for_each_intel_plane_on_crtc(&dev_priv->drm, crtc, plane) { > > + for_each_intel_plane_on_crtc(&i915->drm, crtc, plane) { > > enum plane_id plane_id = plane->id; > > const struct skl_plane_wm *old_wm, *new_wm; > > old_wm = &old_pipe_wm->planes[plane_id]; > > new_wm = &new_pipe_wm->planes[plane_id]; > > - if (skl_plane_wm_equals(dev_priv, old_wm, new_wm)) > > + if (skl_plane_wm_equals(i915, old_wm, new_wm)) > > continue; > > DRM_DEBUG_KMS("[PLANE:%d:%s] level %cwm0,%cwm1,%cwm2,%cwm3,%cwm4,%cwm5,%cwm6,%cwm7,%ctwm" > > @@ -5460,7 +5460,7 @@ static int > > skl_ddb_add_affected_pipes(struct intel_atomic_state *state, bool *changed) > > { > > struct drm_device *dev = state->base.dev; > > - const struct drm_i915_private *dev_priv = to_i915(dev); > > + const struct drm_i915_private *i915 = to_i915(dev); > > struct intel_crtc *crtc; > > struct intel_crtc_state *crtc_state; > > u32 realloc_pipes = pipes_modified(state); > > @@ -5470,14 +5470,14 @@ skl_ddb_add_affected_pipes(struct intel_atomic_state *state, bool *changed) > > * When we distrust bios wm we always need to recompute to set the > > * expected DDB allocations for each CRTC. > > */ > > - if (dev_priv->wm.distrust_bios_wm) > > + if (i915->wm.distrust_bios_wm) > > (*changed) = true; > > /* > > * If this transaction isn't actually touching any CRTC's, don't > > * bother with watermark calculation. Note that if we pass this > > * test, we're guaranteed to hold at least one CRTC state mutex, > > - * which means we can safely use values like dev_priv->active_crtcs > > + * which means we can safely use values like i915->active_crtcs > > * since any racing commits that want to update them would need to > > * hold _all_ CRTC state mutexes. > > */ > > @@ -5493,7 +5493,7 @@ skl_ddb_add_affected_pipes(struct intel_atomic_state *state, bool *changed) > > * pretend that all pipes switched active status so that we'll > > * ensure a full DDB recompute. > > */ > > - if (dev_priv->wm.distrust_bios_wm) { > > + if (i915->wm.distrust_bios_wm) { > > ret = drm_modeset_lock(&dev->mode_config.connection_mutex, > > state->base.acquire_ctx); > > if (ret) > > @@ -5508,7 +5508,7 @@ skl_ddb_add_affected_pipes(struct intel_atomic_state *state, bool *changed) > > * on the first commit too. > > */ > > if (!state->modeset) > > - state->active_crtcs = dev_priv->active_crtcs; > > + state->active_crtcs = i915->active_crtcs; > > } > > /* > > @@ -5567,14 +5567,14 @@ skl_ddb_add_affected_pipes(struct intel_atomic_state *state, bool *changed) > > static int skl_wm_add_affected_planes(struct intel_atomic_state *state, > > struct intel_crtc *crtc) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > const struct intel_crtc_state *old_crtc_state = > > intel_atomic_get_old_crtc_state(state, crtc); > > struct intel_crtc_state *new_crtc_state = > > intel_atomic_get_new_crtc_state(state, crtc); > > struct intel_plane *plane; > > - for_each_intel_plane_on_crtc(&dev_priv->drm, crtc, plane) { > > + for_each_intel_plane_on_crtc(&i915->drm, crtc, plane) { > > struct intel_plane_state *plane_state; > > enum plane_id plane_id = plane->id; > > @@ -5587,7 +5587,7 @@ static int skl_wm_add_affected_planes(struct intel_atomic_state *state, > > * with the software state. > > */ > > if (!drm_atomic_crtc_needs_modeset(&new_crtc_state->base) && > > - skl_plane_wm_equals(dev_priv, > > + skl_plane_wm_equals(i915, > > &old_crtc_state->wm.skl.optimal.planes[plane_id], > > &new_crtc_state->wm.skl.optimal.planes[plane_id])) > > continue; > > @@ -5654,7 +5654,7 @@ static void skl_atomic_update_crtc_wm(struct intel_atomic_state *state, > > struct intel_crtc_state *cstate) > > { > > struct intel_crtc *crtc = to_intel_crtc(cstate->base.crtc); > > - struct drm_i915_private *dev_priv = to_i915(state->base.dev); > > + struct drm_i915_private *i915 = to_i915(state->base.dev); > > struct skl_pipe_wm *pipe_wm = &cstate->wm.skl.optimal; > > enum pipe pipe = crtc->pipe; > > @@ -5669,27 +5669,27 @@ static void skl_initial_wm(struct intel_atomic_state *state, > > { > > struct intel_crtc *intel_crtc = to_intel_crtc(cstate->base.crtc); > > struct drm_device *dev = intel_crtc->base.dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > struct skl_ddb_values *results = &state->wm_results; > > if ((results->dirty_pipes & drm_crtc_mask(&intel_crtc->base)) == 0) > > return; > > - mutex_lock(&dev_priv->wm.wm_mutex); > > + mutex_lock(&i915->wm.wm_mutex); > > if (cstate->base.active_changed) > > skl_atomic_update_crtc_wm(state, cstate); > > - mutex_unlock(&dev_priv->wm.wm_mutex); > > + mutex_unlock(&i915->wm.wm_mutex); > > } > > -static void ilk_compute_wm_config(struct drm_i915_private *dev_priv, > > +static void ilk_compute_wm_config(struct drm_i915_private *i915, > > struct intel_wm_config *config) > > { > > struct intel_crtc *crtc; > > /* Compute the currently _active_ config */ > > - for_each_intel_crtc(&dev_priv->drm, crtc) { > > + for_each_intel_crtc(&i915->drm, crtc) { > > const struct intel_pipe_wm *wm = &crtc->wm.active.ilk; > > if (!wm->pipe_enabled) > > @@ -5701,7 +5701,7 @@ static void ilk_compute_wm_config(struct drm_i915_private *dev_priv, > > } > > } > > -static void ilk_program_watermarks(struct drm_i915_private *dev_priv) > > +static void ilk_program_watermarks(struct drm_i915_private *i915) > > { > > struct intel_pipe_wm lp_wm_1_2 = {}, lp_wm_5_6 = {}, *best_lp_wm; > > struct ilk_wm_maximums max; > > @@ -5709,18 +5709,18 @@ static void ilk_program_watermarks(struct drm_i915_private *dev_priv) > > struct ilk_wm_values results = {}; > > enum intel_ddb_partitioning partitioning; > > - ilk_compute_wm_config(dev_priv, &config); > > + ilk_compute_wm_config(i915, &config); > > - ilk_compute_wm_maximums(dev_priv, 1, &config, INTEL_DDB_PART_1_2, &max); > > - ilk_wm_merge(dev_priv, &config, &max, &lp_wm_1_2); > > + ilk_compute_wm_maximums(i915, 1, &config, INTEL_DDB_PART_1_2, &max); > > + ilk_wm_merge(i915, &config, &max, &lp_wm_1_2); > > /* 5/6 split only in single pipe config on IVB+ */ > > - if (INTEL_GEN(dev_priv) >= 7 && > > + if (INTEL_GEN(i915) >= 7 && > > config.num_pipes_active == 1 && config.sprites_enabled) { > > - ilk_compute_wm_maximums(dev_priv, 1, &config, INTEL_DDB_PART_5_6, &max); > > - ilk_wm_merge(dev_priv, &config, &max, &lp_wm_5_6); > > + ilk_compute_wm_maximums(i915, 1, &config, INTEL_DDB_PART_5_6, &max); > > + ilk_wm_merge(i915, &config, &max, &lp_wm_5_6); > > - best_lp_wm = ilk_find_best_result(dev_priv, &lp_wm_1_2, &lp_wm_5_6); > > + best_lp_wm = ilk_find_best_result(i915, &lp_wm_1_2, &lp_wm_5_6); > > } else { > > best_lp_wm = &lp_wm_1_2; > > } > > @@ -5728,35 +5728,35 @@ static void ilk_program_watermarks(struct drm_i915_private *dev_priv) > > partitioning = (best_lp_wm == &lp_wm_1_2) ? > > INTEL_DDB_PART_1_2 : INTEL_DDB_PART_5_6; > > - ilk_compute_wm_results(dev_priv, best_lp_wm, partitioning, &results); > > + ilk_compute_wm_results(i915, best_lp_wm, partitioning, &results); > > - ilk_write_wm_values(dev_priv, &results); > > + ilk_write_wm_values(i915, &results); > > } > > static void ilk_initial_watermarks(struct intel_atomic_state *state, > > struct intel_crtc_state *cstate) > > { > > - struct drm_i915_private *dev_priv = to_i915(cstate->base.crtc->dev); > > + struct drm_i915_private *i915 = to_i915(cstate->base.crtc->dev); > > struct intel_crtc *intel_crtc = to_intel_crtc(cstate->base.crtc); > > - mutex_lock(&dev_priv->wm.wm_mutex); > > + mutex_lock(&i915->wm.wm_mutex); > > intel_crtc->wm.active.ilk = cstate->wm.ilk.intermediate; > > - ilk_program_watermarks(dev_priv); > > - mutex_unlock(&dev_priv->wm.wm_mutex); > > + ilk_program_watermarks(i915); > > + mutex_unlock(&i915->wm.wm_mutex); > > } > > static void ilk_optimize_watermarks(struct intel_atomic_state *state, > > struct intel_crtc_state *cstate) > > { > > - struct drm_i915_private *dev_priv = to_i915(cstate->base.crtc->dev); > > + struct drm_i915_private *i915 = to_i915(cstate->base.crtc->dev); > > struct intel_crtc *intel_crtc = to_intel_crtc(cstate->base.crtc); > > - mutex_lock(&dev_priv->wm.wm_mutex); > > + mutex_lock(&i915->wm.wm_mutex); > > if (cstate->wm.need_postvbl_update) { > > intel_crtc->wm.active.ilk = cstate->wm.ilk.optimal; > > - ilk_program_watermarks(dev_priv); > > + ilk_program_watermarks(i915); > > } > > - mutex_unlock(&dev_priv->wm.wm_mutex); > > + mutex_unlock(&i915->wm.wm_mutex); > > } > > static inline void skl_wm_level_from_reg_val(u32 val, > > @@ -5772,13 +5772,13 @@ static inline void skl_wm_level_from_reg_val(u32 val, > > void skl_pipe_wm_get_hw_state(struct intel_crtc *crtc, > > struct skl_pipe_wm *out) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > enum pipe pipe = crtc->pipe; > > int level, max_level; > > enum plane_id plane_id; > > u32 val; > > - max_level = ilk_wm_max_level(dev_priv); > > + max_level = ilk_wm_max_level(i915); > > for_each_plane_id_on_crtc(crtc, plane_id) { > > struct skl_plane_wm *wm = &out->planes[plane_id]; > > @@ -5806,15 +5806,15 @@ void skl_pipe_wm_get_hw_state(struct intel_crtc *crtc, > > out->linetime = I915_READ(PIPE_WM_LINETIME(pipe)); > > } > > -void skl_wm_get_hw_state(struct drm_i915_private *dev_priv) > > +void skl_wm_get_hw_state(struct drm_i915_private *i915) > > { > > - struct skl_ddb_values *hw = &dev_priv->wm.skl_hw; > > - struct skl_ddb_allocation *ddb = &dev_priv->wm.skl_hw.ddb; > > + struct skl_ddb_values *hw = &i915->wm.skl_hw; > > + struct skl_ddb_allocation *ddb = &i915->wm.skl_hw.ddb; > > struct intel_crtc *crtc; > > struct intel_crtc_state *cstate; > > - skl_ddb_get_hw_state(dev_priv, ddb); > > - for_each_intel_crtc(&dev_priv->drm, crtc) { > > + skl_ddb_get_hw_state(i915, ddb); > > + for_each_intel_crtc(&i915->drm, crtc) { > > cstate = to_intel_crtc_state(crtc->base.state); > > skl_pipe_wm_get_hw_state(crtc, &cstate->wm.skl.optimal); > > @@ -5823,17 +5823,17 @@ void skl_wm_get_hw_state(struct drm_i915_private *dev_priv) > > hw->dirty_pipes |= drm_crtc_mask(&crtc->base); > > } > > - if (dev_priv->active_crtcs) { > > + if (i915->active_crtcs) { > > /* Fully recompute DDB on first atomic commit */ > > - dev_priv->wm.distrust_bios_wm = true; > > + i915->wm.distrust_bios_wm = true; > > } > > } > > static void ilk_pipe_wm_get_hw_state(struct intel_crtc *crtc) > > { > > struct drm_device *dev = crtc->base.dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > - struct ilk_wm_values *hw = &dev_priv->wm.hw; > > + struct drm_i915_private *i915 = to_i915(dev); > > + struct ilk_wm_values *hw = &i915->wm.hw; > > struct intel_crtc_state *cstate = to_intel_crtc_state(crtc->base.state); > > struct intel_pipe_wm *active = &cstate->wm.ilk.optimal; > > enum pipe pipe = crtc->pipe; > > @@ -5844,7 +5844,7 @@ static void ilk_pipe_wm_get_hw_state(struct intel_crtc *crtc) > > }; > > hw->wm_pipe[pipe] = I915_READ(wm0_pipe_reg[pipe]); > > - if (IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv)) > > + if (IS_HASWELL(i915) || IS_BROADWELL(i915)) > > hw->wm_linetime[pipe] = I915_READ(PIPE_WM_LINETIME(pipe)); > > memset(active, 0, sizeof(*active)); > > @@ -5866,7 +5866,7 @@ static void ilk_pipe_wm_get_hw_state(struct intel_crtc *crtc) > > active->wm[0].cur_val = tmp & WM0_PIPE_CURSOR_MASK; > > active->linetime = hw->wm_linetime[pipe]; > > } else { > > - int level, max_level = ilk_wm_max_level(dev_priv); > > + int level, max_level = ilk_wm_max_level(i915); > > /* > > * For inactive pipes, all watermark levels > > @@ -5885,7 +5885,7 @@ static void ilk_pipe_wm_get_hw_state(struct intel_crtc *crtc) > > #define _FW_WM_VLV(value, plane) \ > > (((value) & DSPFW_ ## plane ## _MASK_VLV) >> DSPFW_ ## plane ## _SHIFT) > > -static void g4x_read_wm_values(struct drm_i915_private *dev_priv, > > +static void g4x_read_wm_values(struct drm_i915_private *i915, > > struct g4x_wm_values *wm) > > { > > u32 tmp; > > @@ -5911,13 +5911,13 @@ static void g4x_read_wm_values(struct drm_i915_private *dev_priv, > > wm->hpll.plane = _FW_WM(tmp, HPLL_SR); > > } > > -static void vlv_read_wm_values(struct drm_i915_private *dev_priv, > > +static void vlv_read_wm_values(struct drm_i915_private *i915, > > struct vlv_wm_values *wm) > > { > > enum pipe pipe; > > u32 tmp; > > - for_each_pipe(dev_priv, pipe) { > > + for_each_pipe(i915, pipe) { > > tmp = I915_READ(VLV_DDL(pipe)); > > wm->ddl[pipe].plane[PLANE_PRIMARY] = > > @@ -5944,7 +5944,7 @@ static void vlv_read_wm_values(struct drm_i915_private *dev_priv, > > tmp = I915_READ(DSPFW3); > > wm->sr.cursor = _FW_WM(tmp, CURSOR_SR); > > - if (IS_CHERRYVIEW(dev_priv)) { > > + if (IS_CHERRYVIEW(i915)) { > > tmp = I915_READ(DSPFW7_CHV); > > wm->pipe[PIPE_B].plane[PLANE_SPRITE1] = _FW_WM_VLV(tmp, SPRITED); > > wm->pipe[PIPE_B].plane[PLANE_SPRITE0] = _FW_WM_VLV(tmp, SPRITEC); > > @@ -5987,16 +5987,16 @@ static void vlv_read_wm_values(struct drm_i915_private *dev_priv, > > #undef _FW_WM > > #undef _FW_WM_VLV > > -void g4x_wm_get_hw_state(struct drm_i915_private *dev_priv) > > +void g4x_wm_get_hw_state(struct drm_i915_private *i915) > > { > > - struct g4x_wm_values *wm = &dev_priv->wm.g4x; > > + struct g4x_wm_values *wm = &i915->wm.g4x; > > struct intel_crtc *crtc; > > - g4x_read_wm_values(dev_priv, wm); > > + g4x_read_wm_values(i915, wm); > > wm->cxsr = I915_READ(FW_BLC_SELF) & FW_BLC_SELF_EN; > > - for_each_intel_crtc(&dev_priv->drm, crtc) { > > + for_each_intel_crtc(&i915->drm, crtc) { > > struct intel_crtc_state *crtc_state = > > to_intel_crtc_state(crtc->base.state); > > struct g4x_wm_state *active = &crtc->wm.active.g4x; > > @@ -6071,16 +6071,16 @@ void g4x_wm_get_hw_state(struct drm_i915_private *dev_priv) > > yesno(wm->cxsr), yesno(wm->hpll_en), yesno(wm->fbc_en)); > > } > > -void g4x_wm_sanitize(struct drm_i915_private *dev_priv) > > +void g4x_wm_sanitize(struct drm_i915_private *i915) > > { > > struct intel_plane *plane; > > struct intel_crtc *crtc; > > - mutex_lock(&dev_priv->wm.wm_mutex); > > + mutex_lock(&i915->wm.wm_mutex); > > - for_each_intel_plane(&dev_priv->drm, plane) { > > + for_each_intel_plane(&i915->drm, plane) { > > struct intel_crtc *crtc = > > - intel_get_crtc_for_pipe(dev_priv, plane->pipe); > > + intel_get_crtc_for_pipe(i915, plane->pipe); > > struct intel_crtc_state *crtc_state = > > to_intel_crtc_state(crtc->base.state); > > struct intel_plane_state *plane_state = > > @@ -6113,7 +6113,7 @@ void g4x_wm_sanitize(struct drm_i915_private *dev_priv) > > } > > } > > - for_each_intel_crtc(&dev_priv->drm, crtc) { > > + for_each_intel_crtc(&i915->drm, crtc) { > > struct intel_crtc_state *crtc_state = > > to_intel_crtc_state(crtc->base.state); > > @@ -6122,26 +6122,26 @@ void g4x_wm_sanitize(struct drm_i915_private *dev_priv) > > crtc->wm.active.g4x = crtc_state->wm.g4x.optimal; > > } > > - g4x_program_watermarks(dev_priv); > > + g4x_program_watermarks(i915); > > - mutex_unlock(&dev_priv->wm.wm_mutex); > > + mutex_unlock(&i915->wm.wm_mutex); > > } > > -void vlv_wm_get_hw_state(struct drm_i915_private *dev_priv) > > +void vlv_wm_get_hw_state(struct drm_i915_private *i915) > > { > > - struct vlv_wm_values *wm = &dev_priv->wm.vlv; > > + struct vlv_wm_values *wm = &i915->wm.vlv; > > struct intel_crtc *crtc; > > u32 val; > > - vlv_read_wm_values(dev_priv, wm); > > + vlv_read_wm_values(i915, wm); > > wm->cxsr = I915_READ(FW_BLC_SELF_VLV) & FW_CSPWRDWNEN; > > wm->level = VLV_WM_LEVEL_PM2; > > - if (IS_CHERRYVIEW(dev_priv)) { > > - vlv_punit_get(dev_priv); > > + if (IS_CHERRYVIEW(i915)) { > > + vlv_punit_get(i915); > > - val = vlv_punit_read(dev_priv, PUNIT_REG_DSPSSPM); > > + val = vlv_punit_read(i915, PUNIT_REG_DSPSSPM); > > if (val & DSP_MAXFIFO_PM5_ENABLE) > > wm->level = VLV_WM_LEVEL_PM5; > > @@ -6154,25 +6154,25 @@ void vlv_wm_get_hw_state(struct drm_i915_private *dev_priv) > > * HIGH/LOW bits so that we don't actually change > > * the current state. > > */ > > - val = vlv_punit_read(dev_priv, PUNIT_REG_DDR_SETUP2); > > + val = vlv_punit_read(i915, PUNIT_REG_DDR_SETUP2); > > val |= FORCE_DDR_FREQ_REQ_ACK; > > - vlv_punit_write(dev_priv, PUNIT_REG_DDR_SETUP2, val); > > + vlv_punit_write(i915, PUNIT_REG_DDR_SETUP2, val); > > - if (wait_for((vlv_punit_read(dev_priv, PUNIT_REG_DDR_SETUP2) & > > + if (wait_for((vlv_punit_read(i915, PUNIT_REG_DDR_SETUP2) & > > FORCE_DDR_FREQ_REQ_ACK) == 0, 3)) { > > DRM_DEBUG_KMS("Punit not acking DDR DVFS request, " > > "assuming DDR DVFS is disabled\n"); > > - dev_priv->wm.max_level = VLV_WM_LEVEL_PM5; > > + i915->wm.max_level = VLV_WM_LEVEL_PM5; > > } else { > > - val = vlv_punit_read(dev_priv, PUNIT_REG_DDR_SETUP2); > > + val = vlv_punit_read(i915, PUNIT_REG_DDR_SETUP2); > > if ((val & FORCE_DDR_HIGH_FREQ) == 0) > > wm->level = VLV_WM_LEVEL_DDR_DVFS; > > } > > - vlv_punit_put(dev_priv); > > + vlv_punit_put(i915); > > } > > - for_each_intel_crtc(&dev_priv->drm, crtc) { > > + for_each_intel_crtc(&i915->drm, crtc) { > > struct intel_crtc_state *crtc_state = > > to_intel_crtc_state(crtc->base.state); > > struct vlv_wm_state *active = &crtc->wm.active.vlv; > > @@ -6224,16 +6224,16 @@ void vlv_wm_get_hw_state(struct drm_i915_private *dev_priv) > > wm->sr.plane, wm->sr.cursor, wm->level, wm->cxsr); > > } > > -void vlv_wm_sanitize(struct drm_i915_private *dev_priv) > > +void vlv_wm_sanitize(struct drm_i915_private *i915) > > { > > struct intel_plane *plane; > > struct intel_crtc *crtc; > > - mutex_lock(&dev_priv->wm.wm_mutex); > > + mutex_lock(&i915->wm.wm_mutex); > > - for_each_intel_plane(&dev_priv->drm, plane) { > > + for_each_intel_plane(&i915->drm, plane) { > > struct intel_crtc *crtc = > > - intel_get_crtc_for_pipe(dev_priv, plane->pipe); > > + intel_get_crtc_for_pipe(i915, plane->pipe); > > struct intel_crtc_state *crtc_state = > > to_intel_crtc_state(crtc->base.state); > > struct intel_plane_state *plane_state = > > @@ -6259,7 +6259,7 @@ void vlv_wm_sanitize(struct drm_i915_private *dev_priv) > > } > > } > > - for_each_intel_crtc(&dev_priv->drm, crtc) { > > + for_each_intel_crtc(&i915->drm, crtc) { > > struct intel_crtc_state *crtc_state = > > to_intel_crtc_state(crtc->base.state); > > @@ -6268,16 +6268,16 @@ void vlv_wm_sanitize(struct drm_i915_private *dev_priv) > > crtc->wm.active.vlv = crtc_state->wm.vlv.optimal; > > } > > - vlv_program_watermarks(dev_priv); > > + vlv_program_watermarks(i915); > > - mutex_unlock(&dev_priv->wm.wm_mutex); > > + mutex_unlock(&i915->wm.wm_mutex); > > } > > /* > > * FIXME should probably kill this and improve > > * the real watermark readout/sanitation instead > > */ > > -static void ilk_init_lp_watermarks(struct drm_i915_private *dev_priv) > > +static void ilk_init_lp_watermarks(struct drm_i915_private *i915) > > { > > I915_WRITE(WM3_LP_ILK, I915_READ(WM3_LP_ILK) & ~WM1_LP_SR_EN); > > I915_WRITE(WM2_LP_ILK, I915_READ(WM2_LP_ILK) & ~WM1_LP_SR_EN); > > @@ -6289,14 +6289,14 @@ static void ilk_init_lp_watermarks(struct drm_i915_private *dev_priv) > > */ > > } > > -void ilk_wm_get_hw_state(struct drm_i915_private *dev_priv) > > +void ilk_wm_get_hw_state(struct drm_i915_private *i915) > > { > > - struct ilk_wm_values *hw = &dev_priv->wm.hw; > > + struct ilk_wm_values *hw = &i915->wm.hw; > > struct intel_crtc *crtc; > > - ilk_init_lp_watermarks(dev_priv); > > + ilk_init_lp_watermarks(i915); > > - for_each_intel_crtc(&dev_priv->drm, crtc) > > + for_each_intel_crtc(&i915->drm, crtc) > > ilk_pipe_wm_get_hw_state(crtc); > > hw->wm_lp[0] = I915_READ(WM1_LP_ILK); > > @@ -6304,15 +6304,15 @@ void ilk_wm_get_hw_state(struct drm_i915_private *dev_priv) > > hw->wm_lp[2] = I915_READ(WM3_LP_ILK); > > hw->wm_lp_spr[0] = I915_READ(WM1S_LP_ILK); > > - if (INTEL_GEN(dev_priv) >= 7) { > > + if (INTEL_GEN(i915) >= 7) { > > hw->wm_lp_spr[1] = I915_READ(WM2S_LP_IVB); > > hw->wm_lp_spr[2] = I915_READ(WM3S_LP_IVB); > > } > > - if (IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv)) > > + if (IS_HASWELL(i915) || IS_BROADWELL(i915)) > > hw->partitioning = (I915_READ(WM_MISC) & WM_MISC_DATA_PARTITION_5_6) ? > > INTEL_DDB_PART_5_6 : INTEL_DDB_PART_1_2; > > - else if (IS_IVYBRIDGE(dev_priv)) > > + else if (IS_IVYBRIDGE(i915)) > > hw->partitioning = (I915_READ(DISP_ARB_CTL2) & DISP_DATA_PARTITION_5_6) ? > > INTEL_DDB_PART_5_6 : INTEL_DDB_PART_1_2; > > @@ -6355,22 +6355,22 @@ void ilk_wm_get_hw_state(struct drm_i915_private *dev_priv) > > */ > > void intel_update_watermarks(struct intel_crtc *crtc) > > { > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > - if (dev_priv->display.update_wm) > > - dev_priv->display.update_wm(crtc); > > + if (i915->display.update_wm) > > + i915->display.update_wm(crtc); > > } > > -void intel_enable_ipc(struct drm_i915_private *dev_priv) > > +void intel_enable_ipc(struct drm_i915_private *i915) > > { > > u32 val; > > - if (!HAS_IPC(dev_priv)) > > + if (!HAS_IPC(i915)) > > return; > > val = I915_READ(DISP_ARB_CTL2); > > - if (dev_priv->ipc_enabled) > > + if (i915->ipc_enabled) > > val |= DISP_IPC_ENABLE; > > else > > val &= ~DISP_IPC_ENABLE; > > @@ -6378,27 +6378,27 @@ void intel_enable_ipc(struct drm_i915_private *dev_priv) > > I915_WRITE(DISP_ARB_CTL2, val); > > } > > -static bool intel_can_enable_ipc(struct drm_i915_private *dev_priv) > > +static bool intel_can_enable_ipc(struct drm_i915_private *i915) > > { > > /* Display WA #0477 WaDisableIPC: skl */ > > - if (IS_SKYLAKE(dev_priv)) > > + if (IS_SKYLAKE(i915)) > > return false; > > /* Display WA #1141: SKL:all KBL:all CFL */ > > - if (IS_KABYLAKE(dev_priv) || IS_COFFEELAKE(dev_priv)) > > - return dev_priv->dram_info.symmetric_memory; > > + if (IS_KABYLAKE(i915) || IS_COFFEELAKE(i915)) > > + return i915->dram_info.symmetric_memory; > > return true; > > } > > -void intel_init_ipc(struct drm_i915_private *dev_priv) > > +void intel_init_ipc(struct drm_i915_private *i915) > > { > > - if (!HAS_IPC(dev_priv)) > > + if (!HAS_IPC(i915)) > > return; > > - dev_priv->ipc_enabled = intel_can_enable_ipc(dev_priv); > > + i915->ipc_enabled = intel_can_enable_ipc(i915); > > - intel_enable_ipc(dev_priv); > > + intel_enable_ipc(i915); > > } > > /* > > @@ -6430,9 +6430,9 @@ bool ironlake_set_drps(struct drm_i915_private *i915, u8 val) > > return true; > > } > > -static void ironlake_enable_drps(struct drm_i915_private *dev_priv) > > +static void ironlake_enable_drps(struct drm_i915_private *i915) > > { > > - struct intel_uncore *uncore = &dev_priv->uncore; > > + struct intel_uncore *uncore = &i915->uncore; > > u32 rgvmodectl; > > u8 fmax, fmin, fstart, vstart; > > @@ -6463,12 +6463,12 @@ static void ironlake_enable_drps(struct drm_i915_private *dev_priv) > > vstart = (intel_uncore_read(uncore, PXVFREQ(fstart)) & > > PXVFREQ_PX_MASK) >> PXVFREQ_PX_SHIFT; > > - dev_priv->ips.fmax = fmax; /* IPS callback will increase this */ > > - dev_priv->ips.fstart = fstart; > > + i915->ips.fmax = fmax; /* IPS callback will increase this */ > > + i915->ips.fstart = fstart; > > - dev_priv->ips.max_delay = fstart; > > - dev_priv->ips.min_delay = fmin; > > - dev_priv->ips.cur_delay = fstart; > > + i915->ips.max_delay = fstart; > > + i915->ips.min_delay = fmin; > > + i915->ips.cur_delay = fstart; > > DRM_DEBUG_DRIVER("fmax: %d, fmin: %d, fstart: %d\n", > > fmax, fmin, fstart); > > @@ -6492,15 +6492,15 @@ static void ironlake_enable_drps(struct drm_i915_private *dev_priv) > > DRM_ERROR("stuck trying to change perf mode\n"); > > mdelay(1); > > - ironlake_set_drps(dev_priv, fstart); > > + ironlake_set_drps(i915, fstart); > > - dev_priv->ips.last_count1 = > > + i915->ips.last_count1 = > > intel_uncore_read(uncore, DMIEC) + > > intel_uncore_read(uncore, DDREC) + > > intel_uncore_read(uncore, CSIEC); > > - dev_priv->ips.last_time1 = jiffies_to_msecs(jiffies); > > - dev_priv->ips.last_count2 = intel_uncore_read(uncore, GFXEC); > > - dev_priv->ips.last_time2 = ktime_get_raw_ns(); > > + i915->ips.last_time1 = jiffies_to_msecs(jiffies); > > + i915->ips.last_count2 = intel_uncore_read(uncore, GFXEC); > > + i915->ips.last_time2 = ktime_get_raw_ns(); > > spin_unlock_irq(&mchdev_lock); > > } > > @@ -6543,9 +6543,9 @@ static void ironlake_disable_drps(struct drm_i915_private *i915) > > * ourselves, instead of doing a rmw cycle (which might result in us clearing > > * all limits and the gpu stuck at whatever frequency it is at atm). > > */ > > -static u32 intel_rps_limits(struct drm_i915_private *dev_priv, u8 val) > > +static u32 intel_rps_limits(struct drm_i915_private *i915, u8 val) > > { > > - struct intel_rps *rps = &dev_priv->gt_pm.rps; > > + struct intel_rps *rps = &i915->gt_pm.rps; > > u32 limits; > > /* Only set the down limit when we've reached the lowest level to avoid > > @@ -6554,7 +6554,7 @@ static u32 intel_rps_limits(struct drm_i915_private *dev_priv, u8 val) > > * the hw runs at the minimal clock before selecting the desired > > * frequency, if the down threshold expires in that window we will not > > * receive a down interrupt. */ > > - if (INTEL_GEN(dev_priv) >= 9) { > > + if (INTEL_GEN(i915) >= 9) { > > limits = (rps->max_freq_softlimit) << 23; > > if (val <= rps->min_freq_softlimit) > > limits |= (rps->min_freq_softlimit) << 14; > > @@ -6567,9 +6567,9 @@ static u32 intel_rps_limits(struct drm_i915_private *dev_priv, u8 val) > > return limits; > > } > > -static void rps_set_power(struct drm_i915_private *dev_priv, int new_power) > > +static void rps_set_power(struct drm_i915_private *i915, int new_power) > > { > > - struct intel_rps *rps = &dev_priv->gt_pm.rps; > > + struct intel_rps *rps = &i915->gt_pm.rps; > > u32 threshold_up = 0, threshold_down = 0; /* in % */ > > u32 ei_up = 0, ei_down = 0; > > @@ -6614,23 +6614,23 @@ static void rps_set_power(struct drm_i915_private *dev_priv, int new_power) > > /* When byt can survive without system hang with dynamic > > * sw freq adjustments, this restriction can be lifted. > > */ > > - if (IS_VALLEYVIEW(dev_priv)) > > + if (IS_VALLEYVIEW(i915)) > > goto skip_hw_write; > > I915_WRITE(GEN6_RP_UP_EI, > > - GT_INTERVAL_FROM_US(dev_priv, ei_up)); > > + GT_INTERVAL_FROM_US(i915, ei_up)); > > I915_WRITE(GEN6_RP_UP_THRESHOLD, > > - GT_INTERVAL_FROM_US(dev_priv, > > + GT_INTERVAL_FROM_US(i915, > > ei_up * threshold_up / 100)); > > I915_WRITE(GEN6_RP_DOWN_EI, > > - GT_INTERVAL_FROM_US(dev_priv, ei_down)); > > + GT_INTERVAL_FROM_US(i915, ei_down)); > > I915_WRITE(GEN6_RP_DOWN_THRESHOLD, > > - GT_INTERVAL_FROM_US(dev_priv, > > + GT_INTERVAL_FROM_US(i915, > > ei_down * threshold_down / 100)); > > I915_WRITE(GEN6_RP_CONTROL, > > - (INTEL_GEN(dev_priv) > 9 ? 0 : GEN6_RP_MEDIA_TURBO) | > > + (INTEL_GEN(i915) > 9 ? 0 : GEN6_RP_MEDIA_TURBO) | > > GEN6_RP_MEDIA_HW_NORMAL_MODE | > > GEN6_RP_MEDIA_IS_GFX | > > GEN6_RP_ENABLE | > > @@ -6643,9 +6643,9 @@ static void rps_set_power(struct drm_i915_private *dev_priv, int new_power) > > rps->power.down_threshold = threshold_down; > > } > > -static void gen6_set_rps_thresholds(struct drm_i915_private *dev_priv, u8 val) > > +static void gen6_set_rps_thresholds(struct drm_i915_private *i915, u8 val) > > { > > - struct intel_rps *rps = &dev_priv->gt_pm.rps; > > + struct intel_rps *rps = &i915->gt_pm.rps; > > int new_power; > > new_power = rps->power.mode; > > @@ -6680,7 +6680,7 @@ static void gen6_set_rps_thresholds(struct drm_i915_private *dev_priv, u8 val) > > mutex_lock(&rps->power.mutex); > > if (rps->power.interactive) > > new_power = HIGH_POWER; > > - rps_set_power(dev_priv, new_power); > > + rps_set_power(i915, new_power); > > mutex_unlock(&rps->power.mutex); > > } > > @@ -6702,9 +6702,9 @@ void intel_rps_mark_interactive(struct drm_i915_private *i915, bool interactive) > > mutex_unlock(&rps->power.mutex); > > } > > -static u32 gen6_rps_pm_mask(struct drm_i915_private *dev_priv, u8 val) > > +static u32 gen6_rps_pm_mask(struct drm_i915_private *i915, u8 val) > > { > > - struct intel_rps *rps = &dev_priv->gt_pm.rps; > > + struct intel_rps *rps = &i915->gt_pm.rps; > > u32 mask = 0; > > /* We use UP_EI_EXPIRED interupts for both up/down in manual mode */ > > @@ -6713,28 +6713,28 @@ static u32 gen6_rps_pm_mask(struct drm_i915_private *dev_priv, u8 val) > > if (val < rps->max_freq_softlimit) > > mask |= GEN6_PM_RP_UP_EI_EXPIRED | GEN6_PM_RP_UP_THRESHOLD; > > - mask &= dev_priv->pm_rps_events; > > + mask &= i915->pm_rps_events; > > - return gen6_sanitize_rps_pm_mask(dev_priv, ~mask); > > + return gen6_sanitize_rps_pm_mask(i915, ~mask); > > } > > /* gen6_set_rps is called to update the frequency request, but should also be > > * called when the range (min_delay and max_delay) is modified so that we can > > * update the GEN6_RP_INTERRUPT_LIMITS register accordingly. */ > > -static int gen6_set_rps(struct drm_i915_private *dev_priv, u8 val) > > +static int gen6_set_rps(struct drm_i915_private *i915, u8 val) > > { > > - struct intel_rps *rps = &dev_priv->gt_pm.rps; > > + struct intel_rps *rps = &i915->gt_pm.rps; > > /* min/max delay may still have been modified so be sure to > > * write the limits value. > > */ > > if (val != rps->cur_freq) { > > - gen6_set_rps_thresholds(dev_priv, val); > > + gen6_set_rps_thresholds(i915, val); > > - if (INTEL_GEN(dev_priv) >= 9) > > + if (INTEL_GEN(i915) >= 9) > > I915_WRITE(GEN6_RPNSWREQ, > > GEN9_FREQUENCY(val)); > > - else if (IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv)) > > + else if (IS_HASWELL(i915) || IS_BROADWELL(i915)) > > I915_WRITE(GEN6_RPNSWREQ, > > HSW_FREQUENCY(val)); > > else > > @@ -6747,37 +6747,37 @@ static int gen6_set_rps(struct drm_i915_private *dev_priv, u8 val) > > /* Make sure we continue to get interrupts > > * until we hit the minimum or maximum frequencies. > > */ > > - I915_WRITE(GEN6_RP_INTERRUPT_LIMITS, intel_rps_limits(dev_priv, val)); > > - I915_WRITE(GEN6_PMINTRMSK, gen6_rps_pm_mask(dev_priv, val)); > > + I915_WRITE(GEN6_RP_INTERRUPT_LIMITS, intel_rps_limits(i915, val)); > > + I915_WRITE(GEN6_PMINTRMSK, gen6_rps_pm_mask(i915, val)); > > rps->cur_freq = val; > > - trace_intel_gpu_freq_change(intel_gpu_freq(dev_priv, val)); > > + trace_intel_gpu_freq_change(intel_gpu_freq(i915, val)); > > return 0; > > } > > -static int valleyview_set_rps(struct drm_i915_private *dev_priv, u8 val) > > +static int valleyview_set_rps(struct drm_i915_private *i915, u8 val) > > { > > int err; > > - if (WARN_ONCE(IS_CHERRYVIEW(dev_priv) && (val & 1), > > + if (WARN_ONCE(IS_CHERRYVIEW(i915) && (val & 1), > > "Odd GPU freq value\n")) > > val &= ~1; > > - I915_WRITE(GEN6_PMINTRMSK, gen6_rps_pm_mask(dev_priv, val)); > > + I915_WRITE(GEN6_PMINTRMSK, gen6_rps_pm_mask(i915, val)); > > - if (val != dev_priv->gt_pm.rps.cur_freq) { > > - vlv_punit_get(dev_priv); > > - err = vlv_punit_write(dev_priv, PUNIT_REG_GPU_FREQ_REQ, val); > > - vlv_punit_put(dev_priv); > > + if (val != i915->gt_pm.rps.cur_freq) { > > + vlv_punit_get(i915); > > + err = vlv_punit_write(i915, PUNIT_REG_GPU_FREQ_REQ, val); > > + vlv_punit_put(i915); > > if (err) > > return err; > > - gen6_set_rps_thresholds(dev_priv, val); > > + gen6_set_rps_thresholds(i915, val); > > } > > - dev_priv->gt_pm.rps.cur_freq = val; > > - trace_intel_gpu_freq_change(intel_gpu_freq(dev_priv, val)); > > + i915->gt_pm.rps.cur_freq = val; > > + trace_intel_gpu_freq_change(intel_gpu_freq(i915, val)); > > return 0; > > } > > @@ -6789,9 +6789,9 @@ static int valleyview_set_rps(struct drm_i915_private *dev_priv, u8 val) > > * 2. Request idle freq. > > * 3. Release Forcewake of Media well. > > */ > > -static void vlv_set_rps_idle(struct drm_i915_private *dev_priv) > > +static void vlv_set_rps_idle(struct drm_i915_private *i915) > > { > > - struct intel_rps *rps = &dev_priv->gt_pm.rps; > > + struct intel_rps *rps = &i915->gt_pm.rps; > > u32 val = rps->idle_freq; > > int err; > > @@ -6810,28 +6810,28 @@ static void vlv_set_rps_idle(struct drm_i915_private *dev_priv) > > * punit into committing the voltage change) as that takes a lot less > > * power than the render powerwell. > > */ > > - intel_uncore_forcewake_get(&dev_priv->uncore, FORCEWAKE_MEDIA); > > - err = valleyview_set_rps(dev_priv, val); > > - intel_uncore_forcewake_put(&dev_priv->uncore, FORCEWAKE_MEDIA); > > + intel_uncore_forcewake_get(&i915->uncore, FORCEWAKE_MEDIA); > > + err = valleyview_set_rps(i915, val); > > + intel_uncore_forcewake_put(&i915->uncore, FORCEWAKE_MEDIA); > > if (err) > > DRM_ERROR("Failed to set RPS for idle\n"); > > } > > -void gen6_rps_busy(struct drm_i915_private *dev_priv) > > +void gen6_rps_busy(struct drm_i915_private *i915) > > { > > - struct intel_rps *rps = &dev_priv->gt_pm.rps; > > + struct intel_rps *rps = &i915->gt_pm.rps; > > mutex_lock(&rps->lock); > > if (rps->enabled) { > > u8 freq; > > - if (dev_priv->pm_rps_events & GEN6_PM_RP_UP_EI_EXPIRED) > > - gen6_rps_reset_ei(dev_priv); > > + if (i915->pm_rps_events & GEN6_PM_RP_UP_EI_EXPIRED) > > + gen6_rps_reset_ei(i915); > > I915_WRITE(GEN6_PMINTRMSK, > > - gen6_rps_pm_mask(dev_priv, rps->cur_freq)); > > + gen6_rps_pm_mask(i915, rps->cur_freq)); > > - gen6_enable_rps_interrupts(dev_priv); > > + gen6_enable_rps_interrupts(i915); > > /* Use the user's desired frequency as a guide, but for better > > * performance, jump directly to RPe as our starting frequency. > > @@ -6839,7 +6839,7 @@ void gen6_rps_busy(struct drm_i915_private *dev_priv) > > freq = max(rps->cur_freq, > > rps->efficient_freq); > > - if (intel_set_rps(dev_priv, > > + if (intel_set_rps(i915, > > clamp(freq, > > rps->min_freq_softlimit, > > rps->max_freq_softlimit))) > > @@ -6848,26 +6848,26 @@ void gen6_rps_busy(struct drm_i915_private *dev_priv) > > mutex_unlock(&rps->lock); > > } > > -void gen6_rps_idle(struct drm_i915_private *dev_priv) > > +void gen6_rps_idle(struct drm_i915_private *i915) > > { > > - struct intel_rps *rps = &dev_priv->gt_pm.rps; > > + struct intel_rps *rps = &i915->gt_pm.rps; > > /* Flush our bottom-half so that it does not race with us > > * setting the idle frequency and so that it is bounded by > > * our rpm wakeref. And then disable the interrupts to stop any > > * futher RPS reclocking whilst we are asleep. > > */ > > - gen6_disable_rps_interrupts(dev_priv); > > + gen6_disable_rps_interrupts(i915); > > mutex_lock(&rps->lock); > > if (rps->enabled) { > > - if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) > > - vlv_set_rps_idle(dev_priv); > > + if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) > > + vlv_set_rps_idle(i915); > > else > > - gen6_set_rps(dev_priv, rps->idle_freq); > > + gen6_set_rps(i915, rps->idle_freq); > > rps->last_adj = 0; > > I915_WRITE(GEN6_PMINTRMSK, > > - gen6_sanitize_rps_pm_mask(dev_priv, ~0)); > > + gen6_sanitize_rps_pm_mask(i915, ~0)); > > } > > mutex_unlock(&rps->lock); > > } > > @@ -6904,9 +6904,9 @@ void gen6_rps_boost(struct i915_request *rq) > > atomic_inc(&rps->boosts); > > } > > -int intel_set_rps(struct drm_i915_private *dev_priv, u8 val) > > +int intel_set_rps(struct drm_i915_private *i915, u8 val) > > { > > - struct intel_rps *rps = &dev_priv->gt_pm.rps; > > + struct intel_rps *rps = &i915->gt_pm.rps; > > int err; > > lockdep_assert_held(&rps->lock); > > @@ -6918,63 +6918,63 @@ int intel_set_rps(struct drm_i915_private *dev_priv, u8 val) > > return 0; > > } > > - if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) > > - err = valleyview_set_rps(dev_priv, val); > > + if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) > > + err = valleyview_set_rps(i915, val); > > else > > - err = gen6_set_rps(dev_priv, val); > > + err = gen6_set_rps(i915, val); > > return err; > > } > > -static void gen9_disable_rc6(struct drm_i915_private *dev_priv) > > +static void gen9_disable_rc6(struct drm_i915_private *i915) > > { > > I915_WRITE(GEN6_RC_CONTROL, 0); > > I915_WRITE(GEN9_PG_ENABLE, 0); > > } > > -static void gen9_disable_rps(struct drm_i915_private *dev_priv) > > +static void gen9_disable_rps(struct drm_i915_private *i915) > > { > > I915_WRITE(GEN6_RP_CONTROL, 0); > > } > > -static void gen6_disable_rc6(struct drm_i915_private *dev_priv) > > +static void gen6_disable_rc6(struct drm_i915_private *i915) > > { > > I915_WRITE(GEN6_RC_CONTROL, 0); > > } > > -static void gen6_disable_rps(struct drm_i915_private *dev_priv) > > +static void gen6_disable_rps(struct drm_i915_private *i915) > > { > > I915_WRITE(GEN6_RPNSWREQ, 1 << 31); > > I915_WRITE(GEN6_RP_CONTROL, 0); > > } > > -static void cherryview_disable_rc6(struct drm_i915_private *dev_priv) > > +static void cherryview_disable_rc6(struct drm_i915_private *i915) > > { > > I915_WRITE(GEN6_RC_CONTROL, 0); > > } > > -static void cherryview_disable_rps(struct drm_i915_private *dev_priv) > > +static void cherryview_disable_rps(struct drm_i915_private *i915) > > { > > I915_WRITE(GEN6_RP_CONTROL, 0); > > } > > -static void valleyview_disable_rc6(struct drm_i915_private *dev_priv) > > +static void valleyview_disable_rc6(struct drm_i915_private *i915) > > { > > /* We're doing forcewake before Disabling RC6, > > * This what the BIOS expects when going into suspend */ > > - intel_uncore_forcewake_get(&dev_priv->uncore, FORCEWAKE_ALL); > > + intel_uncore_forcewake_get(&i915->uncore, FORCEWAKE_ALL); > > I915_WRITE(GEN6_RC_CONTROL, 0); > > - intel_uncore_forcewake_put(&dev_priv->uncore, FORCEWAKE_ALL); > > + intel_uncore_forcewake_put(&i915->uncore, FORCEWAKE_ALL); > > } > > -static void valleyview_disable_rps(struct drm_i915_private *dev_priv) > > +static void valleyview_disable_rps(struct drm_i915_private *i915) > > { > > I915_WRITE(GEN6_RP_CONTROL, 0); > > } > > -static bool bxt_check_bios_rc6_setup(struct drm_i915_private *dev_priv) > > +static bool bxt_check_bios_rc6_setup(struct drm_i915_private *i915) > > { > > bool enable_rc6 = true; > > unsigned long rc6_ctx_base; > > @@ -7000,8 +7000,8 @@ static bool bxt_check_bios_rc6_setup(struct drm_i915_private *dev_priv) > > * for this check. > > */ > > rc6_ctx_base = I915_READ(RC6_CTX_BASE) & RC6_CTX_BASE_MASK; > > - if (!((rc6_ctx_base >= dev_priv->dsm_reserved.start) && > > - (rc6_ctx_base + PAGE_SIZE < dev_priv->dsm_reserved.end))) { > > + if (!((rc6_ctx_base >= i915->dsm_reserved.start) && > > + (rc6_ctx_base + PAGE_SIZE < i915->dsm_reserved.end))) { > > DRM_DEBUG_DRIVER("RC6 Base address not as expected.\n"); > > enable_rc6 = false; > > } > > @@ -7062,14 +7062,14 @@ static bool sanitize_rc6(struct drm_i915_private *i915) > > return info->has_rc6; > > } > > -static void gen6_init_rps_frequencies(struct drm_i915_private *dev_priv) > > +static void gen6_init_rps_frequencies(struct drm_i915_private *i915) > > { > > - struct intel_rps *rps = &dev_priv->gt_pm.rps; > > + struct intel_rps *rps = &i915->gt_pm.rps; > > /* All of these values are in units of 50MHz */ > > /* static values from HW: RP0 > RP1 > RPn (min_freq) */ > > - if (IS_GEN9_LP(dev_priv)) { > > + if (IS_GEN9_LP(i915)) { > > u32 rp_state_cap = I915_READ(BXT_RP_STATE_CAP); > > rps->rp0_freq = (rp_state_cap >> 16) & 0xff; > > rps->rp1_freq = (rp_state_cap >> 8) & 0xff; > > @@ -7084,11 +7084,11 @@ static void gen6_init_rps_frequencies(struct drm_i915_private *dev_priv) > > rps->max_freq = rps->rp0_freq; > > rps->efficient_freq = rps->rp1_freq; > > - if (IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv) || > > - IS_GEN9_BC(dev_priv) || INTEL_GEN(dev_priv) >= 10) { > > + if (IS_HASWELL(i915) || IS_BROADWELL(i915) || > > + IS_GEN9_BC(i915) || INTEL_GEN(i915) >= 10) { > > u32 ddcc_status = 0; > > - if (sandybridge_pcode_read(dev_priv, > > + if (sandybridge_pcode_read(i915, > > HSW_PCODE_DYNAMIC_DUTY_CYCLE_CONTROL, > > &ddcc_status, NULL) == 0) > > rps->efficient_freq = > > @@ -7098,7 +7098,7 @@ static void gen6_init_rps_frequencies(struct drm_i915_private *dev_priv) > > rps->max_freq); > > } > > - if (IS_GEN9_BC(dev_priv) || INTEL_GEN(dev_priv) >= 10) { > > + if (IS_GEN9_BC(i915) || INTEL_GEN(i915) >= 10) { > > /* Store the frequency values in 16.66 MHZ units, which is > > * the natural hardware unit for SKL > > */ > > @@ -7110,45 +7110,45 @@ static void gen6_init_rps_frequencies(struct drm_i915_private *dev_priv) > > } > > } > > -static void reset_rps(struct drm_i915_private *dev_priv, > > +static void reset_rps(struct drm_i915_private *i915, > > int (*set)(struct drm_i915_private *, u8)) > > { > > - struct intel_rps *rps = &dev_priv->gt_pm.rps; > > + struct intel_rps *rps = &i915->gt_pm.rps; > > u8 freq = rps->cur_freq; > > /* force a reset */ > > rps->power.mode = -1; > > rps->cur_freq = -1; > > - if (set(dev_priv, freq)) > > + if (set(i915, freq)) > > DRM_ERROR("Failed to reset RPS to initial values\n"); > > } > > /* See the Gen9_GT_PM_Programming_Guide doc for the below */ > > -static void gen9_enable_rps(struct drm_i915_private *dev_priv) > > +static void gen9_enable_rps(struct drm_i915_private *i915) > > { > > - intel_uncore_forcewake_get(&dev_priv->uncore, FORCEWAKE_ALL); > > + intel_uncore_forcewake_get(&i915->uncore, FORCEWAKE_ALL); > > /* Program defaults and thresholds for RPS */ > > - if (IS_GEN(dev_priv, 9)) > > + if (IS_GEN(i915, 9)) > > I915_WRITE(GEN6_RC_VIDEO_FREQ, > > - GEN9_FREQUENCY(dev_priv->gt_pm.rps.rp1_freq)); > > + GEN9_FREQUENCY(i915->gt_pm.rps.rp1_freq)); > > /* 1 second timeout*/ > > I915_WRITE(GEN6_RP_DOWN_TIMEOUT, > > - GT_INTERVAL_FROM_US(dev_priv, 1000000)); > > + GT_INTERVAL_FROM_US(i915, 1000000)); > > I915_WRITE(GEN6_RP_IDLE_HYSTERSIS, 0xa); > > /* Leaning on the below call to gen6_set_rps to program/setup the > > * Up/Down EI & threshold registers, as well as the RP_CONTROL, > > * RP_INTERRUPT_LIMITS & RPNSWREQ registers */ > > - reset_rps(dev_priv, gen6_set_rps); > > + reset_rps(i915, gen6_set_rps); > > - intel_uncore_forcewake_put(&dev_priv->uncore, FORCEWAKE_ALL); > > + intel_uncore_forcewake_put(&i915->uncore, FORCEWAKE_ALL); > > } > > -static void gen11_enable_rc6(struct drm_i915_private *dev_priv) > > +static void gen11_enable_rc6(struct drm_i915_private *i915) > > { > > struct intel_engine_cs *engine; > > enum intel_engine_id id; > > @@ -7160,7 +7160,7 @@ static void gen11_enable_rc6(struct drm_i915_private *dev_priv) > > * 1b: Get forcewake during program sequence. Although the driver > > * hasn't enabled a state yet where we need forcewake, BIOS may have. > > */ > > - intel_uncore_forcewake_get(&dev_priv->uncore, FORCEWAKE_ALL); > > + intel_uncore_forcewake_get(&i915->uncore, FORCEWAKE_ALL); > > /* 2a: Disable RC states. */ > > I915_WRITE(GEN6_RC_CONTROL, 0); > > @@ -7171,10 +7171,10 @@ static void gen11_enable_rc6(struct drm_i915_private *dev_priv) > > I915_WRITE(GEN6_RC_EVALUATION_INTERVAL, 125000); /* 12500 * 1280ns */ > > I915_WRITE(GEN6_RC_IDLE_HYSTERSIS, 25); /* 25 * 1280ns */ > > - for_each_engine(engine, dev_priv, id) > > + for_each_engine(engine, i915, id) > > I915_WRITE(RING_MAX_IDLE(engine->mmio_base), 10); > > - if (HAS_GUC(dev_priv)) > > + if (HAS_GUC(i915)) > > I915_WRITE(GUC_MAX_IDLE_COUNT, 0xA); > > I915_WRITE(GEN6_RC_SLEEP, 0); > > @@ -7217,10 +7217,10 @@ static void gen11_enable_rc6(struct drm_i915_private *dev_priv) > > GEN9_MEDIA_PG_ENABLE | > > GEN11_MEDIA_SAMPLER_PG_ENABLE); > > - intel_uncore_forcewake_put(&dev_priv->uncore, FORCEWAKE_ALL); > > + intel_uncore_forcewake_put(&i915->uncore, FORCEWAKE_ALL); > > } > > -static void gen9_enable_rc6(struct drm_i915_private *dev_priv) > > +static void gen9_enable_rc6(struct drm_i915_private *i915) > > { > > struct intel_engine_cs *engine; > > enum intel_engine_id id; > > @@ -7231,16 +7231,16 @@ static void gen9_enable_rc6(struct drm_i915_private *dev_priv) > > /* 1b: Get forcewake during program sequence. Although the driver > > * hasn't enabled a state yet where we need forcewake, BIOS may have.*/ > > - intel_uncore_forcewake_get(&dev_priv->uncore, FORCEWAKE_ALL); > > + intel_uncore_forcewake_get(&i915->uncore, FORCEWAKE_ALL); > > /* 2a: Disable RC states. */ > > I915_WRITE(GEN6_RC_CONTROL, 0); > > /* 2b: Program RC6 thresholds.*/ > > - if (INTEL_GEN(dev_priv) >= 10) { > > + if (INTEL_GEN(i915) >= 10) { > > I915_WRITE(GEN6_RC6_WAKE_RATE_LIMIT, 54 << 16 | 85); > > I915_WRITE(GEN10_MEDIA_WAKE_RATE_LIMIT, 150); > > - } else if (IS_SKYLAKE(dev_priv)) { > > + } else if (IS_SKYLAKE(i915)) { > > /* > > * WaRsDoubleRc6WrlWithCoarsePowerGating:skl Doubling WRL only > > * when CPG is enabled > > @@ -7252,10 +7252,10 @@ static void gen9_enable_rc6(struct drm_i915_private *dev_priv) > > I915_WRITE(GEN6_RC_EVALUATION_INTERVAL, 125000); /* 12500 * 1280ns */ > > I915_WRITE(GEN6_RC_IDLE_HYSTERSIS, 25); /* 25 * 1280ns */ > > - for_each_engine(engine, dev_priv, id) > > + for_each_engine(engine, i915, id) > > I915_WRITE(RING_MAX_IDLE(engine->mmio_base), 10); > > - if (HAS_GUC(dev_priv)) > > + if (HAS_GUC(i915)) > > I915_WRITE(GUC_MAX_IDLE_COUNT, 0xA); > > I915_WRITE(GEN6_RC_SLEEP, 0); > > @@ -7288,7 +7288,7 @@ static void gen9_enable_rc6(struct drm_i915_private *dev_priv) > > I915_WRITE(GEN6_RC6_THRESHOLD, 37500); /* 37.5/125ms per EI */ > > /* WaRsUseTimeoutMode:cnl (pre-prod) */ > > - if (IS_CNL_REVID(dev_priv, CNL_REVID_A0, CNL_REVID_C0)) > > + if (IS_CNL_REVID(i915, CNL_REVID_A0, CNL_REVID_C0)) > > rc6_mode = GEN7_RC_CTL_TO_MODE; > > else > > rc6_mode = GEN6_RC_CTL_EI_MODE(1); > > @@ -7302,16 +7302,16 @@ static void gen9_enable_rc6(struct drm_i915_private *dev_priv) > > * 3b: Enable Coarse Power Gating only when RC6 is enabled. > > * WaRsDisableCoarsePowerGating:skl,cnl - Render/Media PG need to be disabled with RC6. > > */ > > - if (NEEDS_WaRsDisableCoarsePowerGating(dev_priv)) > > + if (NEEDS_WaRsDisableCoarsePowerGating(i915)) > > I915_WRITE(GEN9_PG_ENABLE, 0); > > else > > I915_WRITE(GEN9_PG_ENABLE, > > GEN9_RENDER_PG_ENABLE | GEN9_MEDIA_PG_ENABLE); > > - intel_uncore_forcewake_put(&dev_priv->uncore, FORCEWAKE_ALL); > > + intel_uncore_forcewake_put(&i915->uncore, FORCEWAKE_ALL); > > } > > -static void gen8_enable_rc6(struct drm_i915_private *dev_priv) > > +static void gen8_enable_rc6(struct drm_i915_private *i915) > > { > > struct intel_engine_cs *engine; > > enum intel_engine_id id; > > @@ -7321,7 +7321,7 @@ static void gen8_enable_rc6(struct drm_i915_private *dev_priv) > > /* 1b: Get forcewake during program sequence. Although the driver > > * hasn't enabled a state yet where we need forcewake, BIOS may have.*/ > > - intel_uncore_forcewake_get(&dev_priv->uncore, FORCEWAKE_ALL); > > + intel_uncore_forcewake_get(&i915->uncore, FORCEWAKE_ALL); > > /* 2a: Disable RC states. */ > > I915_WRITE(GEN6_RC_CONTROL, 0); > > @@ -7330,7 +7330,7 @@ static void gen8_enable_rc6(struct drm_i915_private *dev_priv) > > I915_WRITE(GEN6_RC6_WAKE_RATE_LIMIT, 40 << 16); > > I915_WRITE(GEN6_RC_EVALUATION_INTERVAL, 125000); /* 12500 * 1280ns */ > > I915_WRITE(GEN6_RC_IDLE_HYSTERSIS, 25); /* 25 * 1280ns */ > > - for_each_engine(engine, dev_priv, id) > > + for_each_engine(engine, i915, id) > > I915_WRITE(RING_MAX_IDLE(engine->mmio_base), 10); > > I915_WRITE(GEN6_RC_SLEEP, 0); > > I915_WRITE(GEN6_RC6_THRESHOLD, 625); /* 800us/1.28 for TO */ > > @@ -7342,14 +7342,14 @@ static void gen8_enable_rc6(struct drm_i915_private *dev_priv) > > GEN7_RC_CTL_TO_MODE | > > GEN6_RC_CTL_RC6_ENABLE); > > - intel_uncore_forcewake_put(&dev_priv->uncore, FORCEWAKE_ALL); > > + intel_uncore_forcewake_put(&i915->uncore, FORCEWAKE_ALL); > > } > > -static void gen8_enable_rps(struct drm_i915_private *dev_priv) > > +static void gen8_enable_rps(struct drm_i915_private *i915) > > { > > - struct intel_rps *rps = &dev_priv->gt_pm.rps; > > + struct intel_rps *rps = &i915->gt_pm.rps; > > - intel_uncore_forcewake_get(&dev_priv->uncore, FORCEWAKE_ALL); > > + intel_uncore_forcewake_get(&i915->uncore, FORCEWAKE_ALL); > > /* 1 Program defaults and thresholds for RPS*/ > > I915_WRITE(GEN6_RPNSWREQ, > > @@ -7380,12 +7380,12 @@ static void gen8_enable_rps(struct drm_i915_private *dev_priv) > > GEN6_RP_UP_BUSY_AVG | > > GEN6_RP_DOWN_IDLE_AVG); > > - reset_rps(dev_priv, gen6_set_rps); > > + reset_rps(i915, gen6_set_rps); > > - intel_uncore_forcewake_put(&dev_priv->uncore, FORCEWAKE_ALL); > > + intel_uncore_forcewake_put(&i915->uncore, FORCEWAKE_ALL); > > } > > -static void gen6_enable_rc6(struct drm_i915_private *dev_priv) > > +static void gen6_enable_rc6(struct drm_i915_private *i915) > > { > > struct intel_engine_cs *engine; > > enum intel_engine_id id; > > @@ -7402,7 +7402,7 @@ static void gen6_enable_rc6(struct drm_i915_private *dev_priv) > > I915_WRITE(GTFIFODBG, gtfifodbg); > > } > > - intel_uncore_forcewake_get(&dev_priv->uncore, FORCEWAKE_ALL); > > + intel_uncore_forcewake_get(&i915->uncore, FORCEWAKE_ALL); > > /* disable the counters and set deterministic thresholds */ > > I915_WRITE(GEN6_RC_CONTROL, 0); > > @@ -7413,12 +7413,12 @@ static void gen6_enable_rc6(struct drm_i915_private *dev_priv) > > I915_WRITE(GEN6_RC_EVALUATION_INTERVAL, 125000); > > I915_WRITE(GEN6_RC_IDLE_HYSTERSIS, 25); > > - for_each_engine(engine, dev_priv, id) > > + for_each_engine(engine, i915, id) > > I915_WRITE(RING_MAX_IDLE(engine->mmio_base), 10); > > I915_WRITE(GEN6_RC_SLEEP, 0); > > I915_WRITE(GEN6_RC1e_THRESHOLD, 1000); > > - if (IS_IVYBRIDGE(dev_priv)) > > + if (IS_IVYBRIDGE(i915)) > > I915_WRITE(GEN6_RC6_THRESHOLD, 125000); > > else > > I915_WRITE(GEN6_RC6_THRESHOLD, 50000); > > @@ -7427,9 +7427,9 @@ static void gen6_enable_rc6(struct drm_i915_private *dev_priv) > > /* We don't use those on Haswell */ > > rc6_mask = GEN6_RC_CTL_RC6_ENABLE; > > - if (HAS_RC6p(dev_priv)) > > + if (HAS_RC6p(i915)) > > rc6_mask |= GEN6_RC_CTL_RC6p_ENABLE; > > - if (HAS_RC6pp(dev_priv)) > > + if (HAS_RC6pp(i915)) > > rc6_mask |= GEN6_RC_CTL_RC6pp_ENABLE; > > I915_WRITE(GEN6_RC_CONTROL, > > rc6_mask | > > @@ -7437,24 +7437,24 @@ static void gen6_enable_rc6(struct drm_i915_private *dev_priv) > > GEN6_RC_CTL_HW_ENABLE); > > rc6vids = 0; > > - ret = sandybridge_pcode_read(dev_priv, GEN6_PCODE_READ_RC6VIDS, > > + ret = sandybridge_pcode_read(i915, GEN6_PCODE_READ_RC6VIDS, > > &rc6vids, NULL); > > - if (IS_GEN(dev_priv, 6) && ret) { > > + if (IS_GEN(i915, 6) && ret) { > > DRM_DEBUG_DRIVER("Couldn't check for BIOS workaround\n"); > > - } else if (IS_GEN(dev_priv, 6) && (GEN6_DECODE_RC6_VID(rc6vids & 0xff) < 450)) { > > + } else if (IS_GEN(i915, 6) && (GEN6_DECODE_RC6_VID(rc6vids & 0xff) < 450)) { > > DRM_DEBUG_DRIVER("You should update your BIOS. Correcting minimum rc6 voltage (%dmV->%dmV)\n", > > GEN6_DECODE_RC6_VID(rc6vids & 0xff), 450); > > rc6vids &= 0xffff00; > > rc6vids |= GEN6_ENCODE_RC6_VID(450); > > - ret = sandybridge_pcode_write(dev_priv, GEN6_PCODE_WRITE_RC6VIDS, rc6vids); > > + ret = sandybridge_pcode_write(i915, GEN6_PCODE_WRITE_RC6VIDS, rc6vids); > > if (ret) > > DRM_ERROR("Couldn't fix incorrect rc6 voltage\n"); > > } > > - intel_uncore_forcewake_put(&dev_priv->uncore, FORCEWAKE_ALL); > > + intel_uncore_forcewake_put(&i915->uncore, FORCEWAKE_ALL); > > } > > -static void gen6_enable_rps(struct drm_i915_private *dev_priv) > > +static void gen6_enable_rps(struct drm_i915_private *i915) > > { > > /* Here begins a magic sequence of register writes to enable > > * auto-downclocking. > > @@ -7462,20 +7462,20 @@ static void gen6_enable_rps(struct drm_i915_private *dev_priv) > > * Perhaps there might be some value in exposing these to > > * userspace... > > */ > > - intel_uncore_forcewake_get(&dev_priv->uncore, FORCEWAKE_ALL); > > + intel_uncore_forcewake_get(&i915->uncore, FORCEWAKE_ALL); > > /* Power down if completely idle for over 50ms */ > > I915_WRITE(GEN6_RP_DOWN_TIMEOUT, 50000); > > I915_WRITE(GEN6_RP_IDLE_HYSTERSIS, 10); > > - reset_rps(dev_priv, gen6_set_rps); > > + reset_rps(i915, gen6_set_rps); > > - intel_uncore_forcewake_put(&dev_priv->uncore, FORCEWAKE_ALL); > > + intel_uncore_forcewake_put(&i915->uncore, FORCEWAKE_ALL); > > } > > -static void gen6_update_ring_freq(struct drm_i915_private *dev_priv) > > +static void gen6_update_ring_freq(struct drm_i915_private *i915) > > { > > - struct intel_rps *rps = &dev_priv->gt_pm.rps; > > + struct intel_rps *rps = &i915->gt_pm.rps; > > const int min_freq = 15; > > const int scaling_factor = 180; > > unsigned int gpu_freq; > > @@ -7509,7 +7509,7 @@ static void gen6_update_ring_freq(struct drm_i915_private *dev_priv) > > min_gpu_freq = rps->min_freq; > > max_gpu_freq = rps->max_freq; > > - if (IS_GEN9_BC(dev_priv) || INTEL_GEN(dev_priv) >= 10) { > > + if (IS_GEN9_BC(i915) || INTEL_GEN(i915) >= 10) { > > /* Convert GT frequency to 50 HZ units */ > > min_gpu_freq /= GEN9_FREQ_SCALER; > > max_gpu_freq /= GEN9_FREQ_SCALER; > > @@ -7524,16 +7524,16 @@ static void gen6_update_ring_freq(struct drm_i915_private *dev_priv) > > const int diff = max_gpu_freq - gpu_freq; > > unsigned int ia_freq = 0, ring_freq = 0; > > - if (IS_GEN9_BC(dev_priv) || INTEL_GEN(dev_priv) >= 10) { > > + if (IS_GEN9_BC(i915) || INTEL_GEN(i915) >= 10) { > > /* > > * ring_freq = 2 * GT. ring_freq is in 100MHz units > > * No floor required for ring frequency on SKL. > > */ > > ring_freq = gpu_freq; > > - } else if (INTEL_GEN(dev_priv) >= 8) { > > + } else if (INTEL_GEN(i915) >= 8) { > > /* max(2 * GT, DDR). NB: GT is 50MHz units */ > > ring_freq = max(min_ring_freq, gpu_freq); > > - } else if (IS_HASWELL(dev_priv)) { > > + } else if (IS_HASWELL(i915)) { > > ring_freq = mult_frac(gpu_freq, 5, 4); > > ring_freq = max(min_ring_freq, ring_freq); > > /* leave ia_freq as the default, chosen by cpufreq */ > > @@ -7552,7 +7552,7 @@ static void gen6_update_ring_freq(struct drm_i915_private *dev_priv) > > ia_freq = DIV_ROUND_CLOSEST(ia_freq, 100); > > } > > - sandybridge_pcode_write(dev_priv, > > + sandybridge_pcode_write(i915, > > GEN6_PCODE_WRITE_MIN_FREQ_TABLE, > > ia_freq << GEN6_PCODE_FREQ_IA_RATIO_SHIFT | > > ring_freq << GEN6_PCODE_FREQ_RING_RATIO_SHIFT | > > @@ -7560,13 +7560,13 @@ static void gen6_update_ring_freq(struct drm_i915_private *dev_priv) > > } > > } > > -static int cherryview_rps_max_freq(struct drm_i915_private *dev_priv) > > +static int cherryview_rps_max_freq(struct drm_i915_private *i915) > > { > > u32 val, rp0; > > - val = vlv_punit_read(dev_priv, FB_GFX_FMAX_AT_VMAX_FUSE); > > + val = vlv_punit_read(i915, FB_GFX_FMAX_AT_VMAX_FUSE); > > - switch (RUNTIME_INFO(dev_priv)->sseu.eu_total) { > > + switch (RUNTIME_INFO(i915)->sseu.eu_total) { > > case 8: > > /* (2 * 4) config */ > > rp0 = (val >> FB_GFX_FMAX_AT_VMAX_2SS4EU_FUSE_SHIFT); > > @@ -7588,53 +7588,53 @@ static int cherryview_rps_max_freq(struct drm_i915_private *dev_priv) > > return rp0; > > } > > -static int cherryview_rps_rpe_freq(struct drm_i915_private *dev_priv) > > +static int cherryview_rps_rpe_freq(struct drm_i915_private *i915) > > { > > u32 val, rpe; > > - val = vlv_punit_read(dev_priv, PUNIT_GPU_DUTYCYCLE_REG); > > + val = vlv_punit_read(i915, PUNIT_GPU_DUTYCYCLE_REG); > > rpe = (val >> PUNIT_GPU_DUTYCYCLE_RPE_FREQ_SHIFT) & PUNIT_GPU_DUTYCYCLE_RPE_FREQ_MASK; > > return rpe; > > } > > -static int cherryview_rps_guar_freq(struct drm_i915_private *dev_priv) > > +static int cherryview_rps_guar_freq(struct drm_i915_private *i915) > > { > > u32 val, rp1; > > - val = vlv_punit_read(dev_priv, FB_GFX_FMAX_AT_VMAX_FUSE); > > + val = vlv_punit_read(i915, FB_GFX_FMAX_AT_VMAX_FUSE); > > rp1 = (val & FB_GFX_FREQ_FUSE_MASK); > > return rp1; > > } > > -static u32 cherryview_rps_min_freq(struct drm_i915_private *dev_priv) > > +static u32 cherryview_rps_min_freq(struct drm_i915_private *i915) > > { > > u32 val, rpn; > > - val = vlv_punit_read(dev_priv, FB_GFX_FMIN_AT_VMIN_FUSE); > > + val = vlv_punit_read(i915, FB_GFX_FMIN_AT_VMIN_FUSE); > > rpn = ((val >> FB_GFX_FMIN_AT_VMIN_FUSE_SHIFT) & > > FB_GFX_FREQ_FUSE_MASK); > > return rpn; > > } > > -static int valleyview_rps_guar_freq(struct drm_i915_private *dev_priv) > > +static int valleyview_rps_guar_freq(struct drm_i915_private *i915) > > { > > u32 val, rp1; > > - val = vlv_nc_read(dev_priv, IOSF_NC_FB_GFX_FREQ_FUSE); > > + val = vlv_nc_read(i915, IOSF_NC_FB_GFX_FREQ_FUSE); > > rp1 = (val & FB_GFX_FGUARANTEED_FREQ_FUSE_MASK) >> FB_GFX_FGUARANTEED_FREQ_FUSE_SHIFT; > > return rp1; > > } > > -static int valleyview_rps_max_freq(struct drm_i915_private *dev_priv) > > +static int valleyview_rps_max_freq(struct drm_i915_private *i915) > > { > > u32 val, rp0; > > - val = vlv_nc_read(dev_priv, IOSF_NC_FB_GFX_FREQ_FUSE); > > + val = vlv_nc_read(i915, IOSF_NC_FB_GFX_FREQ_FUSE); > > rp0 = (val & FB_GFX_MAX_FREQ_FUSE_MASK) >> FB_GFX_MAX_FREQ_FUSE_SHIFT; > > /* Clamp to max */ > > @@ -7643,23 +7643,23 @@ static int valleyview_rps_max_freq(struct drm_i915_private *dev_priv) > > return rp0; > > } > > -static int valleyview_rps_rpe_freq(struct drm_i915_private *dev_priv) > > +static int valleyview_rps_rpe_freq(struct drm_i915_private *i915) > > { > > u32 val, rpe; > > - val = vlv_nc_read(dev_priv, IOSF_NC_FB_GFX_FMAX_FUSE_LO); > > + val = vlv_nc_read(i915, IOSF_NC_FB_GFX_FMAX_FUSE_LO); > > rpe = (val & FB_FMAX_VMIN_FREQ_LO_MASK) >> FB_FMAX_VMIN_FREQ_LO_SHIFT; > > - val = vlv_nc_read(dev_priv, IOSF_NC_FB_GFX_FMAX_FUSE_HI); > > + val = vlv_nc_read(i915, IOSF_NC_FB_GFX_FMAX_FUSE_HI); > > rpe |= (val & FB_FMAX_VMIN_FREQ_HI_MASK) << 5; > > return rpe; > > } > > -static int valleyview_rps_min_freq(struct drm_i915_private *dev_priv) > > +static int valleyview_rps_min_freq(struct drm_i915_private *i915) > > { > > u32 val; > > - val = vlv_punit_read(dev_priv, PUNIT_REG_GPU_LFM) & 0xff; > > + val = vlv_punit_read(i915, PUNIT_REG_GPU_LFM) & 0xff; > > /* > > * According to the BYT Punit GPU turbo HAS 1.1.6.3 the minimum value > > * for the minimum frequency in GPLL mode is 0xc1. Contrary to this on > > @@ -7671,24 +7671,24 @@ static int valleyview_rps_min_freq(struct drm_i915_private *dev_priv) > > } > > /* Check that the pctx buffer wasn't move under us. */ > > -static void valleyview_check_pctx(struct drm_i915_private *dev_priv) > > +static void valleyview_check_pctx(struct drm_i915_private *i915) > > { > > unsigned long pctx_addr = I915_READ(VLV_PCBR) & ~4095; > > - WARN_ON(pctx_addr != dev_priv->dsm.start + > > - dev_priv->vlv_pctx->stolen->start); > > + WARN_ON(pctx_addr != i915->dsm.start + > > + i915->vlv_pctx->stolen->start); > > } > > /* Check that the pcbr address is not empty. */ > > -static void cherryview_check_pctx(struct drm_i915_private *dev_priv) > > +static void cherryview_check_pctx(struct drm_i915_private *i915) > > { > > unsigned long pctx_addr = I915_READ(VLV_PCBR) & ~4095; > > WARN_ON((pctx_addr >> VLV_PCBR_ADDR_SHIFT) == 0); > > } > > -static void cherryview_setup_pctx(struct drm_i915_private *dev_priv) > > +static void cherryview_setup_pctx(struct drm_i915_private *i915) > > { > > resource_size_t pctx_paddr, paddr; > > resource_size_t pctx_size = 32*1024; > > @@ -7697,7 +7697,7 @@ static void cherryview_setup_pctx(struct drm_i915_private *dev_priv) > > pcbr = I915_READ(VLV_PCBR); > > if ((pcbr >> VLV_PCBR_ADDR_SHIFT) == 0) { > > DRM_DEBUG_DRIVER("BIOS didn't set up PCBR, fixing up\n"); > > - paddr = dev_priv->dsm.end + 1 - pctx_size; > > + paddr = i915->dsm.end + 1 - pctx_size; > > GEM_BUG_ON(paddr > U32_MAX); > > pctx_paddr = (paddr & (~4095)); > > @@ -7707,7 +7707,7 @@ static void cherryview_setup_pctx(struct drm_i915_private *dev_priv) > > DRM_DEBUG_DRIVER("PCBR: 0x%08x\n", I915_READ(VLV_PCBR)); > > } > > -static void valleyview_setup_pctx(struct drm_i915_private *dev_priv) > > +static void valleyview_setup_pctx(struct drm_i915_private *i915) > > { > > struct drm_i915_gem_object *pctx; > > resource_size_t pctx_paddr; > > @@ -7719,8 +7719,8 @@ static void valleyview_setup_pctx(struct drm_i915_private *dev_priv) > > /* BIOS set it up already, grab the pre-alloc'd space */ > > resource_size_t pcbr_offset; > > - pcbr_offset = (pcbr & (~4095)) - dev_priv->dsm.start; > > - pctx = i915_gem_object_create_stolen_for_preallocated(dev_priv, > > + pcbr_offset = (pcbr & (~4095)) - i915->dsm.start; > > + pctx = i915_gem_object_create_stolen_for_preallocated(i915, > > pcbr_offset, > > I915_GTT_OFFSET_NONE, > > pctx_size); > > @@ -7737,148 +7737,148 @@ static void valleyview_setup_pctx(struct drm_i915_private *dev_priv) > > * overlap with other ranges, such as the frame buffer, protected > > * memory, or any other relevant ranges. > > */ > > - pctx = i915_gem_object_create_stolen(dev_priv, pctx_size); > > + pctx = i915_gem_object_create_stolen(i915, pctx_size); > > if (!pctx) { > > DRM_DEBUG("not enough stolen space for PCTX, disabling\n"); > > goto out; > > } > > GEM_BUG_ON(range_overflows_t(u64, > > - dev_priv->dsm.start, > > + i915->dsm.start, > > pctx->stolen->start, > > U32_MAX)); > > - pctx_paddr = dev_priv->dsm.start + pctx->stolen->start; > > + pctx_paddr = i915->dsm.start + pctx->stolen->start; > > I915_WRITE(VLV_PCBR, pctx_paddr); > > out: > > DRM_DEBUG_DRIVER("PCBR: 0x%08x\n", I915_READ(VLV_PCBR)); > > - dev_priv->vlv_pctx = pctx; > > + i915->vlv_pctx = pctx; > > } > > -static void valleyview_cleanup_pctx(struct drm_i915_private *dev_priv) > > +static void valleyview_cleanup_pctx(struct drm_i915_private *i915) > > { > > struct drm_i915_gem_object *pctx; > > - pctx = fetch_and_zero(&dev_priv->vlv_pctx); > > + pctx = fetch_and_zero(&i915->vlv_pctx); > > if (pctx) > > i915_gem_object_put(pctx); > > } > > -static void vlv_init_gpll_ref_freq(struct drm_i915_private *dev_priv) > > +static void vlv_init_gpll_ref_freq(struct drm_i915_private *i915) > > { > > - dev_priv->gt_pm.rps.gpll_ref_freq = > > - vlv_get_cck_clock(dev_priv, "GPLL ref", > > + i915->gt_pm.rps.gpll_ref_freq = > > + vlv_get_cck_clock(i915, "GPLL ref", > > CCK_GPLL_CLOCK_CONTROL, > > - dev_priv->czclk_freq); > > + i915->czclk_freq); > > DRM_DEBUG_DRIVER("GPLL reference freq: %d kHz\n", > > - dev_priv->gt_pm.rps.gpll_ref_freq); > > + i915->gt_pm.rps.gpll_ref_freq); > > } > > -static void valleyview_init_gt_powersave(struct drm_i915_private *dev_priv) > > +static void valleyview_init_gt_powersave(struct drm_i915_private *i915) > > { > > - struct intel_rps *rps = &dev_priv->gt_pm.rps; > > + struct intel_rps *rps = &i915->gt_pm.rps; > > u32 val; > > - valleyview_setup_pctx(dev_priv); > > + valleyview_setup_pctx(i915); > > - vlv_iosf_sb_get(dev_priv, > > + vlv_iosf_sb_get(i915, > > BIT(VLV_IOSF_SB_PUNIT) | > > BIT(VLV_IOSF_SB_NC) | > > BIT(VLV_IOSF_SB_CCK)); > > - vlv_init_gpll_ref_freq(dev_priv); > > + vlv_init_gpll_ref_freq(i915); > > - val = vlv_punit_read(dev_priv, PUNIT_REG_GPU_FREQ_STS); > > + val = vlv_punit_read(i915, PUNIT_REG_GPU_FREQ_STS); > > switch ((val >> 6) & 3) { > > case 0: > > case 1: > > - dev_priv->mem_freq = 800; > > + i915->mem_freq = 800; > > break; > > case 2: > > - dev_priv->mem_freq = 1066; > > + i915->mem_freq = 1066; > > break; > > case 3: > > - dev_priv->mem_freq = 1333; > > + i915->mem_freq = 1333; > > break; > > } > > - DRM_DEBUG_DRIVER("DDR speed: %d MHz\n", dev_priv->mem_freq); > > + DRM_DEBUG_DRIVER("DDR speed: %d MHz\n", i915->mem_freq); > > - rps->max_freq = valleyview_rps_max_freq(dev_priv); > > + rps->max_freq = valleyview_rps_max_freq(i915); > > rps->rp0_freq = rps->max_freq; > > DRM_DEBUG_DRIVER("max GPU freq: %d MHz (%u)\n", > > - intel_gpu_freq(dev_priv, rps->max_freq), > > + intel_gpu_freq(i915, rps->max_freq), > > rps->max_freq); > > - rps->efficient_freq = valleyview_rps_rpe_freq(dev_priv); > > + rps->efficient_freq = valleyview_rps_rpe_freq(i915); > > DRM_DEBUG_DRIVER("RPe GPU freq: %d MHz (%u)\n", > > - intel_gpu_freq(dev_priv, rps->efficient_freq), > > + intel_gpu_freq(i915, rps->efficient_freq), > > rps->efficient_freq); > > - rps->rp1_freq = valleyview_rps_guar_freq(dev_priv); > > + rps->rp1_freq = valleyview_rps_guar_freq(i915); > > DRM_DEBUG_DRIVER("RP1(Guar Freq) GPU freq: %d MHz (%u)\n", > > - intel_gpu_freq(dev_priv, rps->rp1_freq), > > + intel_gpu_freq(i915, rps->rp1_freq), > > rps->rp1_freq); > > - rps->min_freq = valleyview_rps_min_freq(dev_priv); > > + rps->min_freq = valleyview_rps_min_freq(i915); > > DRM_DEBUG_DRIVER("min GPU freq: %d MHz (%u)\n", > > - intel_gpu_freq(dev_priv, rps->min_freq), > > + intel_gpu_freq(i915, rps->min_freq), > > rps->min_freq); > > - vlv_iosf_sb_put(dev_priv, > > + vlv_iosf_sb_put(i915, > > BIT(VLV_IOSF_SB_PUNIT) | > > BIT(VLV_IOSF_SB_NC) | > > BIT(VLV_IOSF_SB_CCK)); > > } > > -static void cherryview_init_gt_powersave(struct drm_i915_private *dev_priv) > > +static void cherryview_init_gt_powersave(struct drm_i915_private *i915) > > { > > - struct intel_rps *rps = &dev_priv->gt_pm.rps; > > + struct intel_rps *rps = &i915->gt_pm.rps; > > u32 val; > > - cherryview_setup_pctx(dev_priv); > > + cherryview_setup_pctx(i915); > > - vlv_iosf_sb_get(dev_priv, > > + vlv_iosf_sb_get(i915, > > BIT(VLV_IOSF_SB_PUNIT) | > > BIT(VLV_IOSF_SB_NC) | > > BIT(VLV_IOSF_SB_CCK)); > > - vlv_init_gpll_ref_freq(dev_priv); > > + vlv_init_gpll_ref_freq(i915); > > - val = vlv_cck_read(dev_priv, CCK_FUSE_REG); > > + val = vlv_cck_read(i915, CCK_FUSE_REG); > > switch ((val >> 2) & 0x7) { > > case 3: > > - dev_priv->mem_freq = 2000; > > + i915->mem_freq = 2000; > > break; > > default: > > - dev_priv->mem_freq = 1600; > > + i915->mem_freq = 1600; > > break; > > } > > - DRM_DEBUG_DRIVER("DDR speed: %d MHz\n", dev_priv->mem_freq); > > + DRM_DEBUG_DRIVER("DDR speed: %d MHz\n", i915->mem_freq); > > - rps->max_freq = cherryview_rps_max_freq(dev_priv); > > + rps->max_freq = cherryview_rps_max_freq(i915); > > rps->rp0_freq = rps->max_freq; > > DRM_DEBUG_DRIVER("max GPU freq: %d MHz (%u)\n", > > - intel_gpu_freq(dev_priv, rps->max_freq), > > + intel_gpu_freq(i915, rps->max_freq), > > rps->max_freq); > > - rps->efficient_freq = cherryview_rps_rpe_freq(dev_priv); > > + rps->efficient_freq = cherryview_rps_rpe_freq(i915); > > DRM_DEBUG_DRIVER("RPe GPU freq: %d MHz (%u)\n", > > - intel_gpu_freq(dev_priv, rps->efficient_freq), > > + intel_gpu_freq(i915, rps->efficient_freq), > > rps->efficient_freq); > > - rps->rp1_freq = cherryview_rps_guar_freq(dev_priv); > > + rps->rp1_freq = cherryview_rps_guar_freq(i915); > > DRM_DEBUG_DRIVER("RP1(Guar) GPU freq: %d MHz (%u)\n", > > - intel_gpu_freq(dev_priv, rps->rp1_freq), > > + intel_gpu_freq(i915, rps->rp1_freq), > > rps->rp1_freq); > > - rps->min_freq = cherryview_rps_min_freq(dev_priv); > > + rps->min_freq = cherryview_rps_min_freq(i915); > > DRM_DEBUG_DRIVER("min GPU freq: %d MHz (%u)\n", > > - intel_gpu_freq(dev_priv, rps->min_freq), > > + intel_gpu_freq(i915, rps->min_freq), > > rps->min_freq); > > - vlv_iosf_sb_put(dev_priv, > > + vlv_iosf_sb_put(i915, > > BIT(VLV_IOSF_SB_PUNIT) | > > BIT(VLV_IOSF_SB_NC) | > > BIT(VLV_IOSF_SB_CCK)); > > @@ -7888,12 +7888,12 @@ static void cherryview_init_gt_powersave(struct drm_i915_private *dev_priv) > > "Odd GPU freq values\n"); > > } > > -static void valleyview_cleanup_gt_powersave(struct drm_i915_private *dev_priv) > > +static void valleyview_cleanup_gt_powersave(struct drm_i915_private *i915) > > { > > - valleyview_cleanup_pctx(dev_priv); > > + valleyview_cleanup_pctx(i915); > > } > > -static void cherryview_enable_rc6(struct drm_i915_private *dev_priv) > > +static void cherryview_enable_rc6(struct drm_i915_private *i915) > > { > > struct intel_engine_cs *engine; > > enum intel_engine_id id; > > @@ -7907,11 +7907,11 @@ static void cherryview_enable_rc6(struct drm_i915_private *dev_priv) > > I915_WRITE(GTFIFODBG, gtfifodbg); > > } > > - cherryview_check_pctx(dev_priv); > > + cherryview_check_pctx(i915); > > /* 1a & 1b: Get forcewake during program sequence. Although the driver > > * hasn't enabled a state yet where we need forcewake, BIOS may have.*/ > > - intel_uncore_forcewake_get(&dev_priv->uncore, FORCEWAKE_ALL); > > + intel_uncore_forcewake_get(&i915->uncore, FORCEWAKE_ALL); > > /* Disable RC states. */ > > I915_WRITE(GEN6_RC_CONTROL, 0); > > @@ -7921,7 +7921,7 @@ static void cherryview_enable_rc6(struct drm_i915_private *dev_priv) > > I915_WRITE(GEN6_RC_EVALUATION_INTERVAL, 125000); /* 12500 * 1280ns */ > > I915_WRITE(GEN6_RC_IDLE_HYSTERSIS, 25); /* 25 * 1280ns */ > > - for_each_engine(engine, dev_priv, id) > > + for_each_engine(engine, i915, id) > > I915_WRITE(RING_MAX_IDLE(engine->mmio_base), 10); > > I915_WRITE(GEN6_RC_SLEEP, 0); > > @@ -7943,14 +7943,14 @@ static void cherryview_enable_rc6(struct drm_i915_private *dev_priv) > > rc6_mode = GEN7_RC_CTL_TO_MODE; > > I915_WRITE(GEN6_RC_CONTROL, rc6_mode); > > - intel_uncore_forcewake_put(&dev_priv->uncore, FORCEWAKE_ALL); > > + intel_uncore_forcewake_put(&i915->uncore, FORCEWAKE_ALL); > > } > > -static void cherryview_enable_rps(struct drm_i915_private *dev_priv) > > +static void cherryview_enable_rps(struct drm_i915_private *i915) > > { > > u32 val; > > - intel_uncore_forcewake_get(&dev_priv->uncore, FORCEWAKE_ALL); > > + intel_uncore_forcewake_get(&i915->uncore, FORCEWAKE_ALL); > > /* 1: Program defaults and thresholds for RPS*/ > > I915_WRITE(GEN6_RP_DOWN_TIMEOUT, 1000000); > > @@ -7970,14 +7970,14 @@ static void cherryview_enable_rps(struct drm_i915_private *dev_priv) > > GEN6_RP_DOWN_IDLE_AVG); > > /* Setting Fixed Bias */ > > - vlv_punit_get(dev_priv); > > + vlv_punit_get(i915); > > val = VLV_OVERRIDE_EN | VLV_SOC_TDP_EN | CHV_BIAS_CPU_50_SOC_50; > > - vlv_punit_write(dev_priv, VLV_TURBO_SOC_OVERRIDE, val); > > + vlv_punit_write(i915, VLV_TURBO_SOC_OVERRIDE, val); > > - val = vlv_punit_read(dev_priv, PUNIT_REG_GPU_FREQ_STS); > > + val = vlv_punit_read(i915, PUNIT_REG_GPU_FREQ_STS); > > - vlv_punit_put(dev_priv); > > + vlv_punit_put(i915); > > /* RPS code assumes GPLL is used */ > > WARN_ONCE((val & GPLLENABLE) == 0, "GPLL not enabled\n"); > > @@ -7985,18 +7985,18 @@ static void cherryview_enable_rps(struct drm_i915_private *dev_priv) > > DRM_DEBUG_DRIVER("GPLL enabled? %s\n", yesno(val & GPLLENABLE)); > > DRM_DEBUG_DRIVER("GPU status: 0x%08x\n", val); > > - reset_rps(dev_priv, valleyview_set_rps); > > + reset_rps(i915, valleyview_set_rps); > > - intel_uncore_forcewake_put(&dev_priv->uncore, FORCEWAKE_ALL); > > + intel_uncore_forcewake_put(&i915->uncore, FORCEWAKE_ALL); > > } > > -static void valleyview_enable_rc6(struct drm_i915_private *dev_priv) > > +static void valleyview_enable_rc6(struct drm_i915_private *i915) > > { > > struct intel_engine_cs *engine; > > enum intel_engine_id id; > > u32 gtfifodbg; > > - valleyview_check_pctx(dev_priv); > > + valleyview_check_pctx(i915); > > gtfifodbg = I915_READ(GTFIFODBG); > > if (gtfifodbg) { > > @@ -8005,7 +8005,7 @@ static void valleyview_enable_rc6(struct drm_i915_private *dev_priv) > > I915_WRITE(GTFIFODBG, gtfifodbg); > > } > > - intel_uncore_forcewake_get(&dev_priv->uncore, FORCEWAKE_ALL); > > + intel_uncore_forcewake_get(&i915->uncore, FORCEWAKE_ALL); > > /* Disable RC states. */ > > I915_WRITE(GEN6_RC_CONTROL, 0); > > @@ -8014,7 +8014,7 @@ static void valleyview_enable_rc6(struct drm_i915_private *dev_priv) > > I915_WRITE(GEN6_RC_EVALUATION_INTERVAL, 125000); > > I915_WRITE(GEN6_RC_IDLE_HYSTERSIS, 25); > > - for_each_engine(engine, dev_priv, id) > > + for_each_engine(engine, i915, id) > > I915_WRITE(RING_MAX_IDLE(engine->mmio_base), 10); > > I915_WRITE(GEN6_RC6_THRESHOLD, 0x557); > > @@ -8030,14 +8030,14 @@ static void valleyview_enable_rc6(struct drm_i915_private *dev_priv) > > I915_WRITE(GEN6_RC_CONTROL, > > GEN7_RC_CTL_TO_MODE | VLV_RC_CTL_CTX_RST_PARALLEL); > > - intel_uncore_forcewake_put(&dev_priv->uncore, FORCEWAKE_ALL); > > + intel_uncore_forcewake_put(&i915->uncore, FORCEWAKE_ALL); > > } > > -static void valleyview_enable_rps(struct drm_i915_private *dev_priv) > > +static void valleyview_enable_rps(struct drm_i915_private *i915) > > { > > u32 val; > > - intel_uncore_forcewake_get(&dev_priv->uncore, FORCEWAKE_ALL); > > + intel_uncore_forcewake_get(&i915->uncore, FORCEWAKE_ALL); > > I915_WRITE(GEN6_RP_DOWN_TIMEOUT, 1000000); > > I915_WRITE(GEN6_RP_UP_THRESHOLD, 59400); > > @@ -8055,15 +8055,15 @@ static void valleyview_enable_rps(struct drm_i915_private *dev_priv) > > GEN6_RP_UP_BUSY_AVG | > > GEN6_RP_DOWN_IDLE_CONT); > > - vlv_punit_get(dev_priv); > > + vlv_punit_get(i915); > > /* Setting Fixed Bias */ > > val = VLV_OVERRIDE_EN | VLV_SOC_TDP_EN | VLV_BIAS_CPU_125_SOC_875; > > - vlv_punit_write(dev_priv, VLV_TURBO_SOC_OVERRIDE, val); > > + vlv_punit_write(i915, VLV_TURBO_SOC_OVERRIDE, val); > > - val = vlv_punit_read(dev_priv, PUNIT_REG_GPU_FREQ_STS); > > + val = vlv_punit_read(i915, PUNIT_REG_GPU_FREQ_STS); > > - vlv_punit_put(dev_priv); > > + vlv_punit_put(i915); > > /* RPS code assumes GPLL is used */ > > WARN_ONCE((val & GPLLENABLE) == 0, "GPLL not enabled\n"); > > @@ -8071,9 +8071,9 @@ static void valleyview_enable_rps(struct drm_i915_private *dev_priv) > > DRM_DEBUG_DRIVER("GPLL enabled? %s\n", yesno(val & GPLLENABLE)); > > DRM_DEBUG_DRIVER("GPU status: 0x%08x\n", val); > > - reset_rps(dev_priv, valleyview_set_rps); > > + reset_rps(i915, valleyview_set_rps); > > - intel_uncore_forcewake_put(&dev_priv->uncore, FORCEWAKE_ALL); > > + intel_uncore_forcewake_put(&i915->uncore, FORCEWAKE_ALL); > > } > > static unsigned long intel_pxfreq(u32 vidfreq) > > @@ -8105,7 +8105,7 @@ static const struct cparams { > > { 0, 800, 231, 23784 }, > > }; > > -static unsigned long __i915_chipset_val(struct drm_i915_private *dev_priv) > > +static unsigned long __i915_chipset_val(struct drm_i915_private *i915) > > { > > u64 total_count, diff, ret; > > u32 count1, count2, count3, m = 0, c = 0; > > @@ -8114,7 +8114,7 @@ static unsigned long __i915_chipset_val(struct drm_i915_private *dev_priv) > > lockdep_assert_held(&mchdev_lock); > > - diff1 = now - dev_priv->ips.last_time1; > > + diff1 = now - i915->ips.last_time1; > > /* Prevent division-by-zero if we are asking too fast. > > * Also, we don't get interesting results if we are polling > > @@ -8122,7 +8122,7 @@ static unsigned long __i915_chipset_val(struct drm_i915_private *dev_priv) > > * in such cases. > > */ > > if (diff1 <= 10) > > - return dev_priv->ips.chipset_power; > > + return i915->ips.chipset_power; > > count1 = I915_READ(DMIEC); > > count2 = I915_READ(DDREC); > > @@ -8131,16 +8131,16 @@ static unsigned long __i915_chipset_val(struct drm_i915_private *dev_priv) > > total_count = count1 + count2 + count3; > > /* FIXME: handle per-counter overflow */ > > - if (total_count < dev_priv->ips.last_count1) { > > - diff = ~0UL - dev_priv->ips.last_count1; > > + if (total_count < i915->ips.last_count1) { > > + diff = ~0UL - i915->ips.last_count1; > > diff += total_count; > > } else { > > - diff = total_count - dev_priv->ips.last_count1; > > + diff = total_count - i915->ips.last_count1; > > } > > for (i = 0; i < ARRAY_SIZE(cparams); i++) { > > - if (cparams[i].i == dev_priv->ips.c_m && > > - cparams[i].t == dev_priv->ips.r_t) { > > + if (cparams[i].i == i915->ips.c_m && > > + cparams[i].t == i915->ips.r_t) { > > m = cparams[i].m; > > c = cparams[i].c; > > break; > > @@ -8151,25 +8151,25 @@ static unsigned long __i915_chipset_val(struct drm_i915_private *dev_priv) > > ret = ((m * diff) + c); > > ret = div_u64(ret, 10); > > - dev_priv->ips.last_count1 = total_count; > > - dev_priv->ips.last_time1 = now; > > + i915->ips.last_count1 = total_count; > > + i915->ips.last_time1 = now; > > - dev_priv->ips.chipset_power = ret; > > + i915->ips.chipset_power = ret; > > return ret; > > } > > -unsigned long i915_chipset_val(struct drm_i915_private *dev_priv) > > +unsigned long i915_chipset_val(struct drm_i915_private *i915) > > { > > intel_wakeref_t wakeref; > > unsigned long val = 0; > > - if (!IS_GEN(dev_priv, 5)) > > + if (!IS_GEN(i915, 5)) > > return 0; > > - with_intel_runtime_pm(dev_priv, wakeref) { > > + with_intel_runtime_pm(i915, wakeref) { > > spin_lock_irq(&mchdev_lock); > > - val = __i915_chipset_val(dev_priv); > > + val = __i915_chipset_val(i915); > > spin_unlock_irq(&mchdev_lock); > > } > > @@ -8202,18 +8202,18 @@ static int _pxvid_to_vd(u8 pxvid) > > return (pxvid + 2) * 125; > > } > > -static u32 pvid_to_extvid(struct drm_i915_private *dev_priv, u8 pxvid) > > +static u32 pvid_to_extvid(struct drm_i915_private *i915, u8 pxvid) > > { > > const int vd = _pxvid_to_vd(pxvid); > > const int vm = vd - 1125; > > - if (INTEL_INFO(dev_priv)->is_mobile) > > + if (INTEL_INFO(i915)->is_mobile) > > return vm > 0 ? vm : 0; > > return vd; > > } > > -static void __i915_update_gfx_val(struct drm_i915_private *dev_priv) > > +static void __i915_update_gfx_val(struct drm_i915_private *i915) > > { > > u64 now, diff, diffms; > > u32 count; > > @@ -8221,7 +8221,7 @@ static void __i915_update_gfx_val(struct drm_i915_private *dev_priv) > > lockdep_assert_held(&mchdev_lock); > > now = ktime_get_raw_ns(); > > - diffms = now - dev_priv->ips.last_time2; > > + diffms = now - i915->ips.last_time2; > > do_div(diffms, NSEC_PER_MSEC); > > /* Don't divide by 0 */ > > @@ -8230,50 +8230,50 @@ static void __i915_update_gfx_val(struct drm_i915_private *dev_priv) > > count = I915_READ(GFXEC); > > - if (count < dev_priv->ips.last_count2) { > > - diff = ~0UL - dev_priv->ips.last_count2; > > + if (count < i915->ips.last_count2) { > > + diff = ~0UL - i915->ips.last_count2; > > diff += count; > > } else { > > - diff = count - dev_priv->ips.last_count2; > > + diff = count - i915->ips.last_count2; > > } > > - dev_priv->ips.last_count2 = count; > > - dev_priv->ips.last_time2 = now; > > + i915->ips.last_count2 = count; > > + i915->ips.last_time2 = now; > > /* More magic constants... */ > > diff = diff * 1181; > > diff = div_u64(diff, diffms * 10); > > - dev_priv->ips.gfx_power = diff; > > + i915->ips.gfx_power = diff; > > } > > -void i915_update_gfx_val(struct drm_i915_private *dev_priv) > > +void i915_update_gfx_val(struct drm_i915_private *i915) > > { > > intel_wakeref_t wakeref; > > - if (!IS_GEN(dev_priv, 5)) > > + if (!IS_GEN(i915, 5)) > > return; > > - with_intel_runtime_pm(dev_priv, wakeref) { > > + with_intel_runtime_pm(i915, wakeref) { > > spin_lock_irq(&mchdev_lock); > > - __i915_update_gfx_val(dev_priv); > > + __i915_update_gfx_val(i915); > > spin_unlock_irq(&mchdev_lock); > > } > > } > > -static unsigned long __i915_gfx_val(struct drm_i915_private *dev_priv) > > +static unsigned long __i915_gfx_val(struct drm_i915_private *i915) > > { > > unsigned long t, corr, state1, corr2, state2; > > u32 pxvid, ext_v; > > lockdep_assert_held(&mchdev_lock); > > - pxvid = I915_READ(PXVFREQ(dev_priv->gt_pm.rps.cur_freq)); > > + pxvid = I915_READ(PXVFREQ(i915->gt_pm.rps.cur_freq)); > > pxvid = (pxvid >> 24) & 0x7f; > > - ext_v = pvid_to_extvid(dev_priv, pxvid); > > + ext_v = pvid_to_extvid(i915, pxvid); > > state1 = ext_v; > > - t = i915_mch_val(dev_priv); > > + t = i915_mch_val(i915); > > /* Revel in the empirically derived constants */ > > @@ -8287,27 +8287,27 @@ static unsigned long __i915_gfx_val(struct drm_i915_private *dev_priv) > > corr = corr * ((150142 * state1) / 10000 - 78642); > > corr /= 100000; > > - corr2 = (corr * dev_priv->ips.corr); > > + corr2 = (corr * i915->ips.corr); > > state2 = (corr2 * state1) / 10000; > > state2 /= 100; /* convert to mW */ > > - __i915_update_gfx_val(dev_priv); > > + __i915_update_gfx_val(i915); > > - return dev_priv->ips.gfx_power + state2; > > + return i915->ips.gfx_power + state2; > > } > > -unsigned long i915_gfx_val(struct drm_i915_private *dev_priv) > > +unsigned long i915_gfx_val(struct drm_i915_private *i915) > > { > > intel_wakeref_t wakeref; > > unsigned long val = 0; > > - if (!IS_GEN(dev_priv, 5)) > > + if (!IS_GEN(i915, 5)) > > return 0; > > - with_intel_runtime_pm(dev_priv, wakeref) { > > + with_intel_runtime_pm(i915, wakeref) { > > spin_lock_irq(&mchdev_lock); > > - val = __i915_gfx_val(dev_priv); > > + val = __i915_gfx_val(i915); > > spin_unlock_irq(&mchdev_lock); > > } > > @@ -8471,11 +8471,11 @@ ips_ping_for_i915_load(void) > > } > > } > > -void intel_gpu_ips_init(struct drm_i915_private *dev_priv) > > +void intel_gpu_ips_init(struct drm_i915_private *i915) > > { > > /* We only register the i915 ips part with intel-ips once everything is > > * set up, to avoid intel-ips sneaking in and reading bogus values. */ > > - rcu_assign_pointer(i915_mch_dev, dev_priv); > > + rcu_assign_pointer(i915_mch_dev, i915); > > ips_ping_for_i915_load(); > > } > > @@ -8485,7 +8485,7 @@ void intel_gpu_ips_teardown(void) > > rcu_assign_pointer(i915_mch_dev, NULL); > > } > > -static void intel_init_emon(struct drm_i915_private *dev_priv) > > +static void intel_init_emon(struct drm_i915_private *i915) > > { > > u32 lcfuse; > > u8 pxw[16]; > > @@ -8552,40 +8552,40 @@ static void intel_init_emon(struct drm_i915_private *dev_priv) > > lcfuse = I915_READ(LCFUSE02); > > - dev_priv->ips.corr = (lcfuse & LCFUSE_HIV_MASK); > > + i915->ips.corr = (lcfuse & LCFUSE_HIV_MASK); > > } > > -void intel_init_gt_powersave(struct drm_i915_private *dev_priv) > > +void intel_init_gt_powersave(struct drm_i915_private *i915) > > { > > - struct intel_rps *rps = &dev_priv->gt_pm.rps; > > + struct intel_rps *rps = &i915->gt_pm.rps; > > /* > > * RPM depends on RC6 to save restore the GT HW context, so make RC6 a > > * requirement. > > */ > > - if (!sanitize_rc6(dev_priv)) { > > + if (!sanitize_rc6(i915)) { > > DRM_INFO("RC6 disabled, disabling runtime PM support\n"); > > - pm_runtime_get(&dev_priv->drm.pdev->dev); > > + pm_runtime_get(&i915->drm.pdev->dev); > > } > > /* Initialize RPS limits (for userspace) */ > > - if (IS_CHERRYVIEW(dev_priv)) > > - cherryview_init_gt_powersave(dev_priv); > > - else if (IS_VALLEYVIEW(dev_priv)) > > - valleyview_init_gt_powersave(dev_priv); > > - else if (INTEL_GEN(dev_priv) >= 6) > > - gen6_init_rps_frequencies(dev_priv); > > + if (IS_CHERRYVIEW(i915)) > > + cherryview_init_gt_powersave(i915); > > + else if (IS_VALLEYVIEW(i915)) > > + valleyview_init_gt_powersave(i915); > > + else if (INTEL_GEN(i915) >= 6) > > + gen6_init_rps_frequencies(i915); > > /* Derive initial user preferences/limits from the hardware limits */ > > rps->max_freq_softlimit = rps->max_freq; > > rps->min_freq_softlimit = rps->min_freq; > > /* After setting max-softlimit, find the overclock max freq */ > > - if (IS_GEN(dev_priv, 6) || > > - IS_IVYBRIDGE(dev_priv) || IS_HASWELL(dev_priv)) { > > + if (IS_GEN(i915, 6) || > > + IS_IVYBRIDGE(i915) || IS_HASWELL(i915)) { > > u32 params = 0; > > - sandybridge_pcode_read(dev_priv, GEN6_READ_OC_PARAMS, > > + sandybridge_pcode_read(i915, GEN6_READ_OC_PARAMS, > > ¶ms, NULL); > > if (params & BIT(31)) { /* OC supported */ > > DRM_DEBUG_DRIVER("Overclocking supported, max: %dMHz, overclock: %dMHz\n", > > @@ -8601,25 +8601,25 @@ void intel_init_gt_powersave(struct drm_i915_private *dev_priv) > > rps->cur_freq = rps->idle_freq; > > } > > -void intel_cleanup_gt_powersave(struct drm_i915_private *dev_priv) > > +void intel_cleanup_gt_powersave(struct drm_i915_private *i915) > > { > > - if (IS_VALLEYVIEW(dev_priv)) > > - valleyview_cleanup_gt_powersave(dev_priv); > > + if (IS_VALLEYVIEW(i915)) > > + valleyview_cleanup_gt_powersave(i915); > > - if (!HAS_RC6(dev_priv)) > > - pm_runtime_put(&dev_priv->drm.pdev->dev); > > + if (!HAS_RC6(i915)) > > + pm_runtime_put(&i915->drm.pdev->dev); > > } > > -void intel_sanitize_gt_powersave(struct drm_i915_private *dev_priv) > > +void intel_sanitize_gt_powersave(struct drm_i915_private *i915) > > { > > - dev_priv->gt_pm.rps.enabled = true; /* force RPS disabling */ > > - dev_priv->gt_pm.rc6.enabled = true; /* force RC6 disabling */ > > - intel_disable_gt_powersave(dev_priv); > > + i915->gt_pm.rps.enabled = true; /* force RPS disabling */ > > + i915->gt_pm.rc6.enabled = true; /* force RC6 disabling */ > > + intel_disable_gt_powersave(i915); > > - if (INTEL_GEN(dev_priv) >= 11) > > - gen11_reset_rps_interrupts(dev_priv); > > - else if (INTEL_GEN(dev_priv) >= 6) > > - gen6_reset_rps_interrupts(dev_priv); > > + if (INTEL_GEN(i915) >= 11) > > + gen11_reset_rps_interrupts(i915); > > + else if (INTEL_GEN(i915) >= 6) > > + gen6_reset_rps_interrupts(i915); > > } > > static inline void intel_disable_llc_pstate(struct drm_i915_private *i915) > > @@ -8634,56 +8634,56 @@ static inline void intel_disable_llc_pstate(struct drm_i915_private *i915) > > i915->gt_pm.llc_pstate.enabled = false; > > } > > -static void intel_disable_rc6(struct drm_i915_private *dev_priv) > > +static void intel_disable_rc6(struct drm_i915_private *i915) > > { > > - lockdep_assert_held(&dev_priv->gt_pm.rps.lock); > > + lockdep_assert_held(&i915->gt_pm.rps.lock); > > - if (!dev_priv->gt_pm.rc6.enabled) > > + if (!i915->gt_pm.rc6.enabled) > > return; > > - if (INTEL_GEN(dev_priv) >= 9) > > - gen9_disable_rc6(dev_priv); > > - else if (IS_CHERRYVIEW(dev_priv)) > > - cherryview_disable_rc6(dev_priv); > > - else if (IS_VALLEYVIEW(dev_priv)) > > - valleyview_disable_rc6(dev_priv); > > - else if (INTEL_GEN(dev_priv) >= 6) > > - gen6_disable_rc6(dev_priv); > > + if (INTEL_GEN(i915) >= 9) > > + gen9_disable_rc6(i915); > > + else if (IS_CHERRYVIEW(i915)) > > + cherryview_disable_rc6(i915); > > + else if (IS_VALLEYVIEW(i915)) > > + valleyview_disable_rc6(i915); > > + else if (INTEL_GEN(i915) >= 6) > > + gen6_disable_rc6(i915); > > - dev_priv->gt_pm.rc6.enabled = false; > > + i915->gt_pm.rc6.enabled = false; > > } > > -static void intel_disable_rps(struct drm_i915_private *dev_priv) > > +static void intel_disable_rps(struct drm_i915_private *i915) > > { > > - lockdep_assert_held(&dev_priv->gt_pm.rps.lock); > > + lockdep_assert_held(&i915->gt_pm.rps.lock); > > - if (!dev_priv->gt_pm.rps.enabled) > > + if (!i915->gt_pm.rps.enabled) > > return; > > - if (INTEL_GEN(dev_priv) >= 9) > > - gen9_disable_rps(dev_priv); > > - else if (IS_CHERRYVIEW(dev_priv)) > > - cherryview_disable_rps(dev_priv); > > - else if (IS_VALLEYVIEW(dev_priv)) > > - valleyview_disable_rps(dev_priv); > > - else if (INTEL_GEN(dev_priv) >= 6) > > - gen6_disable_rps(dev_priv); > > - else if (IS_IRONLAKE_M(dev_priv)) > > - ironlake_disable_drps(dev_priv); > > + if (INTEL_GEN(i915) >= 9) > > + gen9_disable_rps(i915); > > + else if (IS_CHERRYVIEW(i915)) > > + cherryview_disable_rps(i915); > > + else if (IS_VALLEYVIEW(i915)) > > + valleyview_disable_rps(i915); > > + else if (INTEL_GEN(i915) >= 6) > > + gen6_disable_rps(i915); > > + else if (IS_IRONLAKE_M(i915)) > > + ironlake_disable_drps(i915); > > - dev_priv->gt_pm.rps.enabled = false; > > + i915->gt_pm.rps.enabled = false; > > } > > -void intel_disable_gt_powersave(struct drm_i915_private *dev_priv) > > +void intel_disable_gt_powersave(struct drm_i915_private *i915) > > { > > - mutex_lock(&dev_priv->gt_pm.rps.lock); > > + mutex_lock(&i915->gt_pm.rps.lock); > > - intel_disable_rc6(dev_priv); > > - intel_disable_rps(dev_priv); > > - if (HAS_LLC(dev_priv)) > > - intel_disable_llc_pstate(dev_priv); > > + intel_disable_rc6(i915); > > + intel_disable_rps(i915); > > + if (HAS_LLC(i915)) > > + intel_disable_llc_pstate(i915); > > - mutex_unlock(&dev_priv->gt_pm.rps.lock); > > + mutex_unlock(&i915->gt_pm.rps.lock); > > } > > static inline void intel_enable_llc_pstate(struct drm_i915_private *i915) > > @@ -8698,51 +8698,51 @@ static inline void intel_enable_llc_pstate(struct drm_i915_private *i915) > > i915->gt_pm.llc_pstate.enabled = true; > > } > > -static void intel_enable_rc6(struct drm_i915_private *dev_priv) > > +static void intel_enable_rc6(struct drm_i915_private *i915) > > { > > - lockdep_assert_held(&dev_priv->gt_pm.rps.lock); > > + lockdep_assert_held(&i915->gt_pm.rps.lock); > > - if (dev_priv->gt_pm.rc6.enabled) > > + if (i915->gt_pm.rc6.enabled) > > return; > > - if (IS_CHERRYVIEW(dev_priv)) > > - cherryview_enable_rc6(dev_priv); > > - else if (IS_VALLEYVIEW(dev_priv)) > > - valleyview_enable_rc6(dev_priv); > > - else if (INTEL_GEN(dev_priv) >= 11) > > - gen11_enable_rc6(dev_priv); > > - else if (INTEL_GEN(dev_priv) >= 9) > > - gen9_enable_rc6(dev_priv); > > - else if (IS_BROADWELL(dev_priv)) > > - gen8_enable_rc6(dev_priv); > > - else if (INTEL_GEN(dev_priv) >= 6) > > - gen6_enable_rc6(dev_priv); > > + if (IS_CHERRYVIEW(i915)) > > + cherryview_enable_rc6(i915); > > + else if (IS_VALLEYVIEW(i915)) > > + valleyview_enable_rc6(i915); > > + else if (INTEL_GEN(i915) >= 11) > > + gen11_enable_rc6(i915); > > + else if (INTEL_GEN(i915) >= 9) > > + gen9_enable_rc6(i915); > > + else if (IS_BROADWELL(i915)) > > + gen8_enable_rc6(i915); > > + else if (INTEL_GEN(i915) >= 6) > > + gen6_enable_rc6(i915); > > - dev_priv->gt_pm.rc6.enabled = true; > > + i915->gt_pm.rc6.enabled = true; > > } > > -static void intel_enable_rps(struct drm_i915_private *dev_priv) > > +static void intel_enable_rps(struct drm_i915_private *i915) > > { > > - struct intel_rps *rps = &dev_priv->gt_pm.rps; > > + struct intel_rps *rps = &i915->gt_pm.rps; > > lockdep_assert_held(&rps->lock); > > if (rps->enabled) > > return; > > - if (IS_CHERRYVIEW(dev_priv)) { > > - cherryview_enable_rps(dev_priv); > > - } else if (IS_VALLEYVIEW(dev_priv)) { > > - valleyview_enable_rps(dev_priv); > > - } else if (INTEL_GEN(dev_priv) >= 9) { > > - gen9_enable_rps(dev_priv); > > - } else if (IS_BROADWELL(dev_priv)) { > > - gen8_enable_rps(dev_priv); > > - } else if (INTEL_GEN(dev_priv) >= 6) { > > - gen6_enable_rps(dev_priv); > > - } else if (IS_IRONLAKE_M(dev_priv)) { > > - ironlake_enable_drps(dev_priv); > > - intel_init_emon(dev_priv); > > + if (IS_CHERRYVIEW(i915)) { > > + cherryview_enable_rps(i915); > > + } else if (IS_VALLEYVIEW(i915)) { > > + valleyview_enable_rps(i915); > > + } else if (INTEL_GEN(i915) >= 9) { > > + gen9_enable_rps(i915); > > + } else if (IS_BROADWELL(i915)) { > > + gen8_enable_rps(i915); > > + } else if (INTEL_GEN(i915) >= 6) { > > + gen6_enable_rps(i915); > > + } else if (IS_IRONLAKE_M(i915)) { > > + ironlake_enable_drps(i915); > > + intel_init_emon(i915); > > } > > WARN_ON(rps->max_freq < rps->min_freq); > > @@ -8754,25 +8754,25 @@ static void intel_enable_rps(struct drm_i915_private *dev_priv) > > rps->enabled = true; > > } > > -void intel_enable_gt_powersave(struct drm_i915_private *dev_priv) > > +void intel_enable_gt_powersave(struct drm_i915_private *i915) > > { > > /* Powersaving is controlled by the host when inside a VM */ > > - if (intel_vgpu_active(dev_priv)) > > + if (intel_vgpu_active(i915)) > > return; > > - mutex_lock(&dev_priv->gt_pm.rps.lock); > > + mutex_lock(&i915->gt_pm.rps.lock); > > - if (HAS_RC6(dev_priv)) > > - intel_enable_rc6(dev_priv); > > - if (HAS_RPS(dev_priv)) > > - intel_enable_rps(dev_priv); > > - if (HAS_LLC(dev_priv)) > > - intel_enable_llc_pstate(dev_priv); > > + if (HAS_RC6(i915)) > > + intel_enable_rc6(i915); > > + if (HAS_RPS(i915)) > > + intel_enable_rps(i915); > > + if (HAS_LLC(i915)) > > + intel_enable_llc_pstate(i915); > > - mutex_unlock(&dev_priv->gt_pm.rps.lock); > > + mutex_unlock(&i915->gt_pm.rps.lock); > > } > > -static void ibx_init_clock_gating(struct drm_i915_private *dev_priv) > > +static void ibx_init_clock_gating(struct drm_i915_private *i915) > > { > > /* > > * On Ibex Peak and Cougar Point, we need to disable clock > > @@ -8782,11 +8782,11 @@ static void ibx_init_clock_gating(struct drm_i915_private *dev_priv) > > I915_WRITE(SOUTH_DSPCLK_GATE_D, PCH_DPLSUNIT_CLOCK_GATE_DISABLE); > > } > > -static void g4x_disable_trickle_feed(struct drm_i915_private *dev_priv) > > +static void g4x_disable_trickle_feed(struct drm_i915_private *i915) > > { > > enum pipe pipe; > > - for_each_pipe(dev_priv, pipe) { > > + for_each_pipe(i915, pipe) { > > I915_WRITE(DSPCNTR(pipe), > > I915_READ(DSPCNTR(pipe)) | > > DISPPLANE_TRICKLE_FEED_DISABLE); > > @@ -8796,7 +8796,7 @@ static void g4x_disable_trickle_feed(struct drm_i915_private *dev_priv) > > } > > } > > -static void ilk_init_clock_gating(struct drm_i915_private *dev_priv) > > +static void ilk_init_clock_gating(struct drm_i915_private *i915) > > { > > u32 dspclk_gate = ILK_VRHUNIT_CLOCK_GATE_DISABLE; > > @@ -8836,7 +8836,7 @@ static void ilk_init_clock_gating(struct drm_i915_private *dev_priv) > > * The bit 22 of 0x42004 > > * The bit 7,8,9 of 0x42020. > > */ > > - if (IS_IRONLAKE_M(dev_priv)) { > > + if (IS_IRONLAKE_M(i915)) { > > /* WaFbcAsynchFlipDisableFbcQueue:ilk */ > > I915_WRITE(ILK_DISPLAY_CHICKEN1, > > I915_READ(ILK_DISPLAY_CHICKEN1) | > > @@ -8862,12 +8862,12 @@ static void ilk_init_clock_gating(struct drm_i915_private *dev_priv) > > /* WaDisable_RenderCache_OperationalFlush:ilk */ > > I915_WRITE(CACHE_MODE_0, _MASKED_BIT_DISABLE(RC_OP_FLUSH_ENABLE)); > > - g4x_disable_trickle_feed(dev_priv); > > + g4x_disable_trickle_feed(i915); > > - ibx_init_clock_gating(dev_priv); > > + ibx_init_clock_gating(i915); > > } > > -static void cpt_init_clock_gating(struct drm_i915_private *dev_priv) > > +static void cpt_init_clock_gating(struct drm_i915_private *i915) > > { > > int pipe; > > u32 val; > > @@ -8885,11 +8885,11 @@ static void cpt_init_clock_gating(struct drm_i915_private *dev_priv) > > /* The below fixes the weird display corruption, a few pixels shifted > > * downward, on (only) LVDS of some HP laptops with IVY. > > */ > > - for_each_pipe(dev_priv, pipe) { > > + for_each_pipe(i915, pipe) { > > val = I915_READ(TRANS_CHICKEN2(pipe)); > > val |= TRANS_CHICKEN2_TIMING_OVERRIDE; > > val &= ~TRANS_CHICKEN2_FDI_POLARITY_REVERSED; > > - if (dev_priv->vbt.fdi_rx_polarity_inverted) > > + if (i915->vbt.fdi_rx_polarity_inverted) > > val |= TRANS_CHICKEN2_FDI_POLARITY_REVERSED; > > val &= ~TRANS_CHICKEN2_FRAME_START_DELAY_MASK; > > val &= ~TRANS_CHICKEN2_DISABLE_DEEP_COLOR_COUNTER; > > @@ -8897,13 +8897,13 @@ static void cpt_init_clock_gating(struct drm_i915_private *dev_priv) > > I915_WRITE(TRANS_CHICKEN2(pipe), val); > > } > > /* WADP0ClockGatingDisable */ > > - for_each_pipe(dev_priv, pipe) { > > + for_each_pipe(i915, pipe) { > > I915_WRITE(TRANS_CHICKEN1(pipe), > > TRANS_CHICKEN1_DP0UNIT_GC_DISABLE); > > } > > } > > -static void gen6_check_mch_setup(struct drm_i915_private *dev_priv) > > +static void gen6_check_mch_setup(struct drm_i915_private *i915) > > { > > u32 tmp; > > @@ -8913,7 +8913,7 @@ static void gen6_check_mch_setup(struct drm_i915_private *dev_priv) > > tmp); > > } > > -static void gen6_init_clock_gating(struct drm_i915_private *dev_priv) > > +static void gen6_init_clock_gating(struct drm_i915_private *i915) > > { > > u32 dspclk_gate = ILK_VRHUNIT_CLOCK_GATE_DISABLE; > > @@ -9000,14 +9000,14 @@ static void gen6_init_clock_gating(struct drm_i915_private *dev_priv) > > ILK_DPARBUNIT_CLOCK_GATE_ENABLE | > > ILK_DPFDUNIT_CLOCK_GATE_ENABLE); > > - g4x_disable_trickle_feed(dev_priv); > > + g4x_disable_trickle_feed(i915); > > - cpt_init_clock_gating(dev_priv); > > + cpt_init_clock_gating(i915); > > - gen6_check_mch_setup(dev_priv); > > + gen6_check_mch_setup(i915); > > } > > -static void gen7_setup_fixed_func_scheduler(struct drm_i915_private *dev_priv) > > +static void gen7_setup_fixed_func_scheduler(struct drm_i915_private *i915) > > { > > u32 reg = I915_READ(GEN7_FF_THREAD_MODE); > > @@ -9025,13 +9025,13 @@ static void gen7_setup_fixed_func_scheduler(struct drm_i915_private *dev_priv) > > I915_WRITE(GEN7_FF_THREAD_MODE, reg); > > } > > -static void lpt_init_clock_gating(struct drm_i915_private *dev_priv) > > +static void lpt_init_clock_gating(struct drm_i915_private *i915) > > { > > /* > > * TODO: this bit should only be enabled when really needed, then > > * disabled when not needed anymore in order to save power. > > */ > > - if (HAS_PCH_LPT_LP(dev_priv)) > > + if (HAS_PCH_LPT_LP(i915)) > > I915_WRITE(SOUTH_DSPCLK_GATE_D, > > I915_READ(SOUTH_DSPCLK_GATE_D) | > > PCH_LP_PARTITION_LEVEL_DISABLE); > > @@ -9042,9 +9042,9 @@ static void lpt_init_clock_gating(struct drm_i915_private *dev_priv) > > TRANS_CHICKEN1_DP0UNIT_GC_DISABLE); > > } > > -static void lpt_suspend_hw(struct drm_i915_private *dev_priv) > > +static void lpt_suspend_hw(struct drm_i915_private *i915) > > { > > - if (HAS_PCH_LPT_LP(dev_priv)) { > > + if (HAS_PCH_LPT_LP(i915)) { > > u32 val = I915_READ(SOUTH_DSPCLK_GATE_D); > > val &= ~PCH_LP_PARTITION_LEVEL_DISABLE; > > @@ -9052,7 +9052,7 @@ static void lpt_suspend_hw(struct drm_i915_private *dev_priv) > > } > > } > > -static void gen8_set_l3sqc_credits(struct drm_i915_private *dev_priv, > > +static void gen8_set_l3sqc_credits(struct drm_i915_private *i915, > > int general_prio_credits, > > int high_prio_credits) > > { > > @@ -9078,7 +9078,7 @@ static void gen8_set_l3sqc_credits(struct drm_i915_private *dev_priv, > > I915_WRITE(GEN7_MISCCPCTL, misccpctl); > > } > > -static void icl_init_clock_gating(struct drm_i915_private *dev_priv) > > +static void icl_init_clock_gating(struct drm_i915_private *i915) > > { > > /* This is not an Wa. Enable to reduce Sampler power */ > > I915_WRITE(GEN10_DFR_RATIO_EN_AND_CHICKEN, > > @@ -9089,9 +9089,9 @@ static void icl_init_clock_gating(struct drm_i915_private *dev_priv) > > _MASKED_BIT_ENABLE(GEN11_ENABLE_32_PLANE_MODE)); > > } > > -static void cnp_init_clock_gating(struct drm_i915_private *dev_priv) > > +static void cnp_init_clock_gating(struct drm_i915_private *i915) > > { > > - if (!HAS_PCH_CNP(dev_priv)) > > + if (!HAS_PCH_CNP(i915)) > > return; > > /* Display WA #1181 WaSouthDisplayDisablePWMCGEGating: cnp */ > > @@ -9099,10 +9099,10 @@ static void cnp_init_clock_gating(struct drm_i915_private *dev_priv) > > CNP_PWM_CGE_GATING_DISABLE); > > } > > -static void cnl_init_clock_gating(struct drm_i915_private *dev_priv) > > +static void cnl_init_clock_gating(struct drm_i915_private *i915) > > { > > u32 val; > > - cnp_init_clock_gating(dev_priv); > > + cnp_init_clock_gating(i915); > > /* This is not an Wa. Enable for better image quality */ > > I915_WRITE(_3D_CHICKEN3, > > @@ -9120,7 +9120,7 @@ static void cnl_init_clock_gating(struct drm_i915_private *dev_priv) > > /* ReadHitWriteOnlyDisable:cnl */ > > val |= RCCUNIT_CLKGATE_DIS; > > /* WaSarbUnitClockGatingDisable:cnl (pre-prod) */ > > - if (IS_CNL_REVID(dev_priv, CNL_REVID_A0, CNL_REVID_B0)) > > + if (IS_CNL_REVID(i915, CNL_REVID_A0, CNL_REVID_B0)) > > val |= SARBUNIT_CLKGATE_DIS; > > I915_WRITE(SLICE_UNIT_LEVEL_CLKGATE, val); > > @@ -9136,27 +9136,27 @@ static void cnl_init_clock_gating(struct drm_i915_private *dev_priv) > > I915_WRITE(UNSLICE_UNIT_LEVEL_CLKGATE, val); > > } > > -static void cfl_init_clock_gating(struct drm_i915_private *dev_priv) > > +static void cfl_init_clock_gating(struct drm_i915_private *i915) > > { > > - cnp_init_clock_gating(dev_priv); > > - gen9_init_clock_gating(dev_priv); > > + cnp_init_clock_gating(i915); > > + gen9_init_clock_gating(i915); > > /* WaFbcNukeOnHostModify:cfl */ > > I915_WRITE(ILK_DPFC_CHICKEN, I915_READ(ILK_DPFC_CHICKEN) | > > ILK_DPFC_NUKE_ON_ANY_MODIFICATION); > > } > > -static void kbl_init_clock_gating(struct drm_i915_private *dev_priv) > > +static void kbl_init_clock_gating(struct drm_i915_private *i915) > > { > > - gen9_init_clock_gating(dev_priv); > > + gen9_init_clock_gating(i915); > > /* WaDisableSDEUnitClockGating:kbl */ > > - if (IS_KBL_REVID(dev_priv, 0, KBL_REVID_B0)) > > + if (IS_KBL_REVID(i915, 0, KBL_REVID_B0)) > > I915_WRITE(GEN8_UCGCTL6, I915_READ(GEN8_UCGCTL6) | > > GEN8_SDEUNIT_CLOCK_GATE_DISABLE); > > /* WaDisableGamClockGating:kbl */ > > - if (IS_KBL_REVID(dev_priv, 0, KBL_REVID_B0)) > > + if (IS_KBL_REVID(i915, 0, KBL_REVID_B0)) > > I915_WRITE(GEN6_UCGCTL1, I915_READ(GEN6_UCGCTL1) | > > GEN6_GAMUNIT_CLOCK_GATE_DISABLE); > > @@ -9165,9 +9165,9 @@ static void kbl_init_clock_gating(struct drm_i915_private *dev_priv) > > ILK_DPFC_NUKE_ON_ANY_MODIFICATION); > > } > > -static void skl_init_clock_gating(struct drm_i915_private *dev_priv) > > +static void skl_init_clock_gating(struct drm_i915_private *i915) > > { > > - gen9_init_clock_gating(dev_priv); > > + gen9_init_clock_gating(i915); > > /* WAC6entrylatency:skl */ > > I915_WRITE(FBC_LLC_READ_CTRL, I915_READ(FBC_LLC_READ_CTRL) | > > @@ -9178,10 +9178,10 @@ static void skl_init_clock_gating(struct drm_i915_private *dev_priv) > > ILK_DPFC_NUKE_ON_ANY_MODIFICATION); > > } > > -static void bdw_init_clock_gating(struct drm_i915_private *dev_priv) > > +static void bdw_init_clock_gating(struct drm_i915_private *i915) > > { > > /* The GTT cache must be disabled if the system is using 2M pages. */ > > - bool can_use_gtt_cache = !HAS_PAGE_SIZES(dev_priv, > > + bool can_use_gtt_cache = !HAS_PAGE_SIZES(i915, > > I915_GTT_PAGE_SIZE_2M); > > enum pipe pipe; > > @@ -9193,7 +9193,7 @@ static void bdw_init_clock_gating(struct drm_i915_private *dev_priv) > > I915_READ(CHICKEN_PAR1_1) | DPA_MASK_VBLANK_SRD); > > /* WaPsrDPRSUnmaskVBlankInSRD:bdw */ > > - for_each_pipe(dev_priv, pipe) { > > + for_each_pipe(i915, pipe) { > > I915_WRITE(CHICKEN_PIPESL_1(pipe), > > I915_READ(CHICKEN_PIPESL_1(pipe)) | > > BDW_DPRS_MASK_VBLANK_SRD); > > @@ -9213,7 +9213,7 @@ static void bdw_init_clock_gating(struct drm_i915_private *dev_priv) > > GEN8_SDEUNIT_CLOCK_GATE_DISABLE); > > /* WaProgramL3SqcReg1Default:bdw */ > > - gen8_set_l3sqc_credits(dev_priv, 30, 2); > > + gen8_set_l3sqc_credits(i915, 30, 2); > > /* WaGttCachingOffByDefault:bdw */ > > I915_WRITE(HSW_GTT_CACHE_EN, can_use_gtt_cache ? GTT_CACHE_EN_ALL : 0); > > @@ -9222,7 +9222,7 @@ static void bdw_init_clock_gating(struct drm_i915_private *dev_priv) > > I915_WRITE(CHICKEN_PAR2_1, I915_READ(CHICKEN_PAR2_1) > > | KVM_CONFIG_CHANGE_NOTIFICATION_SELECT); > > - lpt_init_clock_gating(dev_priv); > > + lpt_init_clock_gating(i915); > > /* WaDisableDopClockGating:bdw > > * > > @@ -9233,7 +9233,7 @@ static void bdw_init_clock_gating(struct drm_i915_private *dev_priv) > > I915_READ(GEN6_UCGCTL1) | GEN6_EU_TCUNIT_CLOCK_GATE_DISABLE); > > } > > -static void hsw_init_clock_gating(struct drm_i915_private *dev_priv) > > +static void hsw_init_clock_gating(struct drm_i915_private *i915) > > { > > /* L3 caching of data atomics doesn't work -- disable it. */ > > I915_WRITE(HSW_SCRATCH1, HSW_SCRATCH1_L3_DATA_ATOMICS_DISABLE); > > @@ -9278,10 +9278,10 @@ static void hsw_init_clock_gating(struct drm_i915_private *dev_priv) > > /* WaSwitchSolVfFArbitrationPriority:hsw */ > > I915_WRITE(GAM_ECOCHK, I915_READ(GAM_ECOCHK) | HSW_ECOCHK_ARB_PRIO_SOL); > > - lpt_init_clock_gating(dev_priv); > > + lpt_init_clock_gating(i915); > > } > > -static void ivb_init_clock_gating(struct drm_i915_private *dev_priv) > > +static void ivb_init_clock_gating(struct drm_i915_private *i915) > > { > > u32 snpcr; > > @@ -9297,7 +9297,7 @@ static void ivb_init_clock_gating(struct drm_i915_private *dev_priv) > > CHICKEN3_DGMG_DONE_FIX_DISABLE); > > /* WaDisablePSDDualDispatchEnable:ivb */ > > - if (IS_IVB_GT1(dev_priv)) > > + if (IS_IVB_GT1(i915)) > > I915_WRITE(GEN7_HALF_SLICE_CHICKEN1, > > _MASKED_BIT_ENABLE(GEN7_PSD_SINGLE_PORT_DISPATCH_ENABLE)); > > @@ -9313,7 +9313,7 @@ static void ivb_init_clock_gating(struct drm_i915_private *dev_priv) > > GEN7_WA_FOR_GEN7_L3_CONTROL); > > I915_WRITE(GEN7_L3_CHICKEN_MODE_REGISTER, > > GEN7_WA_L3_CHICKEN_MODE); > > - if (IS_IVB_GT1(dev_priv)) > > + if (IS_IVB_GT1(i915)) > > I915_WRITE(GEN7_ROW_CHICKEN2, > > _MASKED_BIT_ENABLE(DOP_CLOCK_GATING_DISABLE)); > > else { > > @@ -9340,9 +9340,9 @@ static void ivb_init_clock_gating(struct drm_i915_private *dev_priv) > > I915_READ(GEN7_SQ_CHICKEN_MBCUNIT_CONFIG) | > > GEN7_SQ_CHICKEN_MBCUNIT_SQINTMOB); > > - g4x_disable_trickle_feed(dev_priv); > > + g4x_disable_trickle_feed(i915); > > - gen7_setup_fixed_func_scheduler(dev_priv); > > + gen7_setup_fixed_func_scheduler(i915); > > if (0) { /* causes HiZ corruption on ivb:gt1 */ > > /* enable HiZ Raw Stall Optimization */ > > @@ -9370,13 +9370,13 @@ static void ivb_init_clock_gating(struct drm_i915_private *dev_priv) > > snpcr |= GEN6_MBC_SNPCR_MED; > > I915_WRITE(GEN6_MBCUNIT_SNPCR, snpcr); > > - if (!HAS_PCH_NOP(dev_priv)) > > - cpt_init_clock_gating(dev_priv); > > + if (!HAS_PCH_NOP(i915)) > > + cpt_init_clock_gating(i915); > > - gen6_check_mch_setup(dev_priv); > > + gen6_check_mch_setup(i915); > > } > > -static void vlv_init_clock_gating(struct drm_i915_private *dev_priv) > > +static void vlv_init_clock_gating(struct drm_i915_private *i915) > > { > > /* WaDisableEarlyCull:vlv */ > > I915_WRITE(_3D_CHICKEN3, > > @@ -9409,7 +9409,7 @@ static void vlv_init_clock_gating(struct drm_i915_private *dev_priv) > > I915_READ(GEN7_SQ_CHICKEN_MBCUNIT_CONFIG) | > > GEN7_SQ_CHICKEN_MBCUNIT_SQINTMOB); > > - gen7_setup_fixed_func_scheduler(dev_priv); > > + gen7_setup_fixed_func_scheduler(i915); > > /* > > * According to the spec, bit 13 (RCZUNIT) must be set on IVB. > > @@ -9456,7 +9456,7 @@ static void vlv_init_clock_gating(struct drm_i915_private *dev_priv) > > I915_WRITE(VLV_GUNIT_CLOCK_GATE, GCFG_DIS); > > } > > -static void chv_init_clock_gating(struct drm_i915_private *dev_priv) > > +static void chv_init_clock_gating(struct drm_i915_private *i915) > > { > > /* WaVSRefCountFullforceMissDisable:chv */ > > /* WaDSRefCountFullforceMissDisable:chv */ > > @@ -9481,7 +9481,7 @@ static void chv_init_clock_gating(struct drm_i915_private *dev_priv) > > * See gfxspecs/Related Documents/Performance Guide/ > > * LSQC Setting Recommendations. > > */ > > - gen8_set_l3sqc_credits(dev_priv, 38, 2); > > + gen8_set_l3sqc_credits(i915, 38, 2); > > /* > > * GTT cache may not work with big pages, so if those > > @@ -9490,7 +9490,7 @@ static void chv_init_clock_gating(struct drm_i915_private *dev_priv) > > I915_WRITE(HSW_GTT_CACHE_EN, GTT_CACHE_EN_ALL); > > } > > -static void g4x_init_clock_gating(struct drm_i915_private *dev_priv) > > +static void g4x_init_clock_gating(struct drm_i915_private *i915) > > { > > u32 dspclk_gate; > > @@ -9502,7 +9502,7 @@ static void g4x_init_clock_gating(struct drm_i915_private *dev_priv) > > dspclk_gate = VRHUNIT_CLOCK_GATE_DISABLE | > > OVRUNIT_CLOCK_GATE_DISABLE | > > OVCUNIT_CLOCK_GATE_DISABLE; > > - if (IS_GM45(dev_priv)) > > + if (IS_GM45(i915)) > > dspclk_gate |= DSSUNIT_CLOCK_GATE_DISABLE; > > I915_WRITE(DSPCLK_GATE_D, dspclk_gate); > > @@ -9513,12 +9513,12 @@ static void g4x_init_clock_gating(struct drm_i915_private *dev_priv) > > /* WaDisable_RenderCache_OperationalFlush:g4x */ > > I915_WRITE(CACHE_MODE_0, _MASKED_BIT_DISABLE(RC_OP_FLUSH_ENABLE)); > > - g4x_disable_trickle_feed(dev_priv); > > + g4x_disable_trickle_feed(i915); > > } > > -static void i965gm_init_clock_gating(struct drm_i915_private *dev_priv) > > +static void i965gm_init_clock_gating(struct drm_i915_private *i915) > > { > > - struct intel_uncore *uncore = &dev_priv->uncore; > > + struct intel_uncore *uncore = &i915->uncore; > > intel_uncore_write(uncore, RENCLK_GATE_D1, I965_RCC_CLOCK_GATE_DISABLE); > > intel_uncore_write(uncore, RENCLK_GATE_D2, 0); > > @@ -9535,7 +9535,7 @@ static void i965gm_init_clock_gating(struct drm_i915_private *dev_priv) > > _MASKED_BIT_DISABLE(RC_OP_FLUSH_ENABLE)); > > } > > -static void i965g_init_clock_gating(struct drm_i915_private *dev_priv) > > +static void i965g_init_clock_gating(struct drm_i915_private *i915) > > { > > I915_WRITE(RENCLK_GATE_D1, I965_RCZ_CLOCK_GATE_DISABLE | > > I965_RCC_CLOCK_GATE_DISABLE | > > @@ -9550,7 +9550,7 @@ static void i965g_init_clock_gating(struct drm_i915_private *dev_priv) > > I915_WRITE(CACHE_MODE_0, _MASKED_BIT_DISABLE(RC_OP_FLUSH_ENABLE)); > > } > > -static void gen3_init_clock_gating(struct drm_i915_private *dev_priv) > > +static void gen3_init_clock_gating(struct drm_i915_private *i915) > > { > > u32 dstate = I915_READ(D_STATE); > > @@ -9558,7 +9558,7 @@ static void gen3_init_clock_gating(struct drm_i915_private *dev_priv) > > DSTATE_DOT_CLOCK_GATING; > > I915_WRITE(D_STATE, dstate); > > - if (IS_PINEVIEW(dev_priv)) > > + if (IS_PINEVIEW(i915)) > > I915_WRITE(ECOSKPD, _MASKED_BIT_ENABLE(ECO_GATING_CX_ONLY)); > > /* IIR "flip pending" means done if this bit is set */ > > @@ -9574,7 +9574,7 @@ static void gen3_init_clock_gating(struct drm_i915_private *dev_priv) > > _MASKED_BIT_ENABLE(MI_ARB_DISPLAY_TRICKLE_FEED_DISABLE)); > > } > > -static void i85x_init_clock_gating(struct drm_i915_private *dev_priv) > > +static void i85x_init_clock_gating(struct drm_i915_private *i915) > > { > > I915_WRITE(RENCLK_GATE_D1, SV_CLOCK_GATE_DISABLE); > > @@ -9586,168 +9586,168 @@ static void i85x_init_clock_gating(struct drm_i915_private *dev_priv) > > _MASKED_BIT_ENABLE(MEM_DISPLAY_TRICKLE_FEED_DISABLE)); > > } > > -static void i830_init_clock_gating(struct drm_i915_private *dev_priv) > > +static void i830_init_clock_gating(struct drm_i915_private *i915) > > { > > I915_WRITE(MEM_MODE, > > _MASKED_BIT_ENABLE(MEM_DISPLAY_A_TRICKLE_FEED_DISABLE) | > > _MASKED_BIT_ENABLE(MEM_DISPLAY_B_TRICKLE_FEED_DISABLE)); > > } > > -void intel_init_clock_gating(struct drm_i915_private *dev_priv) > > +void intel_init_clock_gating(struct drm_i915_private *i915) > > { > > - dev_priv->display.init_clock_gating(dev_priv); > > + i915->display.init_clock_gating(i915); > > } > > -void intel_suspend_hw(struct drm_i915_private *dev_priv) > > +void intel_suspend_hw(struct drm_i915_private *i915) > > { > > - if (HAS_PCH_LPT(dev_priv)) > > - lpt_suspend_hw(dev_priv); > > + if (HAS_PCH_LPT(i915)) > > + lpt_suspend_hw(i915); > > } > > -static void nop_init_clock_gating(struct drm_i915_private *dev_priv) > > +static void nop_init_clock_gating(struct drm_i915_private *i915) > > { > > DRM_DEBUG_KMS("No clock gating settings or workarounds applied.\n"); > > } > > /** > > * intel_init_clock_gating_hooks - setup the clock gating hooks > > - * @dev_priv: device private > > + * @i915: device private > > * > > * Setup the hooks that configure which clocks of a given platform can be > > * gated and also apply various GT and display specific workarounds for these > > * platforms. Note that some GT specific workarounds are applied separately > > * when GPU contexts or batchbuffers start their execution. > > */ > > -void intel_init_clock_gating_hooks(struct drm_i915_private *dev_priv) > > -{ > > - if (IS_GEN(dev_priv, 11)) > > - dev_priv->display.init_clock_gating = icl_init_clock_gating; > > - else if (IS_CANNONLAKE(dev_priv)) > > - dev_priv->display.init_clock_gating = cnl_init_clock_gating; > > - else if (IS_COFFEELAKE(dev_priv)) > > - dev_priv->display.init_clock_gating = cfl_init_clock_gating; > > - else if (IS_SKYLAKE(dev_priv)) > > - dev_priv->display.init_clock_gating = skl_init_clock_gating; > > - else if (IS_KABYLAKE(dev_priv)) > > - dev_priv->display.init_clock_gating = kbl_init_clock_gating; > > - else if (IS_BROXTON(dev_priv)) > > - dev_priv->display.init_clock_gating = bxt_init_clock_gating; > > - else if (IS_GEMINILAKE(dev_priv)) > > - dev_priv->display.init_clock_gating = glk_init_clock_gating; > > - else if (IS_BROADWELL(dev_priv)) > > - dev_priv->display.init_clock_gating = bdw_init_clock_gating; > > - else if (IS_CHERRYVIEW(dev_priv)) > > - dev_priv->display.init_clock_gating = chv_init_clock_gating; > > - else if (IS_HASWELL(dev_priv)) > > - dev_priv->display.init_clock_gating = hsw_init_clock_gating; > > - else if (IS_IVYBRIDGE(dev_priv)) > > - dev_priv->display.init_clock_gating = ivb_init_clock_gating; > > - else if (IS_VALLEYVIEW(dev_priv)) > > - dev_priv->display.init_clock_gating = vlv_init_clock_gating; > > - else if (IS_GEN(dev_priv, 6)) > > - dev_priv->display.init_clock_gating = gen6_init_clock_gating; > > - else if (IS_GEN(dev_priv, 5)) > > - dev_priv->display.init_clock_gating = ilk_init_clock_gating; > > - else if (IS_G4X(dev_priv)) > > - dev_priv->display.init_clock_gating = g4x_init_clock_gating; > > - else if (IS_I965GM(dev_priv)) > > - dev_priv->display.init_clock_gating = i965gm_init_clock_gating; > > - else if (IS_I965G(dev_priv)) > > - dev_priv->display.init_clock_gating = i965g_init_clock_gating; > > - else if (IS_GEN(dev_priv, 3)) > > - dev_priv->display.init_clock_gating = gen3_init_clock_gating; > > - else if (IS_I85X(dev_priv) || IS_I865G(dev_priv)) > > - dev_priv->display.init_clock_gating = i85x_init_clock_gating; > > - else if (IS_GEN(dev_priv, 2)) > > - dev_priv->display.init_clock_gating = i830_init_clock_gating; > > +void intel_init_clock_gating_hooks(struct drm_i915_private *i915) > > +{ > > + if (IS_GEN(i915, 11)) > > + i915->display.init_clock_gating = icl_init_clock_gating; > > + else if (IS_CANNONLAKE(i915)) > > + i915->display.init_clock_gating = cnl_init_clock_gating; > > + else if (IS_COFFEELAKE(i915)) > > + i915->display.init_clock_gating = cfl_init_clock_gating; > > + else if (IS_SKYLAKE(i915)) > > + i915->display.init_clock_gating = skl_init_clock_gating; > > + else if (IS_KABYLAKE(i915)) > > + i915->display.init_clock_gating = kbl_init_clock_gating; > > + else if (IS_BROXTON(i915)) > > + i915->display.init_clock_gating = bxt_init_clock_gating; > > + else if (IS_GEMINILAKE(i915)) > > + i915->display.init_clock_gating = glk_init_clock_gating; > > + else if (IS_BROADWELL(i915)) > > + i915->display.init_clock_gating = bdw_init_clock_gating; > > + else if (IS_CHERRYVIEW(i915)) > > + i915->display.init_clock_gating = chv_init_clock_gating; > > + else if (IS_HASWELL(i915)) > > + i915->display.init_clock_gating = hsw_init_clock_gating; > > + else if (IS_IVYBRIDGE(i915)) > > + i915->display.init_clock_gating = ivb_init_clock_gating; > > + else if (IS_VALLEYVIEW(i915)) > > + i915->display.init_clock_gating = vlv_init_clock_gating; > > + else if (IS_GEN(i915, 6)) > > + i915->display.init_clock_gating = gen6_init_clock_gating; > > + else if (IS_GEN(i915, 5)) > > + i915->display.init_clock_gating = ilk_init_clock_gating; > > + else if (IS_G4X(i915)) > > + i915->display.init_clock_gating = g4x_init_clock_gating; > > + else if (IS_I965GM(i915)) > > + i915->display.init_clock_gating = i965gm_init_clock_gating; > > + else if (IS_I965G(i915)) > > + i915->display.init_clock_gating = i965g_init_clock_gating; > > + else if (IS_GEN(i915, 3)) > > + i915->display.init_clock_gating = gen3_init_clock_gating; > > + else if (IS_I85X(i915) || IS_I865G(i915)) > > + i915->display.init_clock_gating = i85x_init_clock_gating; > > + else if (IS_GEN(i915, 2)) > > + i915->display.init_clock_gating = i830_init_clock_gating; > > else { > > - MISSING_CASE(INTEL_DEVID(dev_priv)); > > - dev_priv->display.init_clock_gating = nop_init_clock_gating; > > + MISSING_CASE(INTEL_DEVID(i915)); > > + i915->display.init_clock_gating = nop_init_clock_gating; > > } > > } > > /* Set up chip specific power management-related functions */ > > -void intel_init_pm(struct drm_i915_private *dev_priv) > > +void intel_init_pm(struct drm_i915_private *i915) > > { > > /* For cxsr */ > > - if (IS_PINEVIEW(dev_priv)) > > - i915_pineview_get_mem_freq(dev_priv); > > - else if (IS_GEN(dev_priv, 5)) > > - i915_ironlake_get_mem_freq(dev_priv); > > + if (IS_PINEVIEW(i915)) > > + i915_pineview_get_mem_freq(i915); > > + else if (IS_GEN(i915, 5)) > > + i915_ironlake_get_mem_freq(i915); > > /* For FIFO watermark updates */ > > - if (INTEL_GEN(dev_priv) >= 9) { > > - skl_setup_wm_latency(dev_priv); > > - dev_priv->display.initial_watermarks = skl_initial_wm; > > - dev_priv->display.atomic_update_watermarks = skl_atomic_update_crtc_wm; > > - dev_priv->display.compute_global_watermarks = skl_compute_wm; > > - } else if (HAS_PCH_SPLIT(dev_priv)) { > > - ilk_setup_wm_latency(dev_priv); > > - > > - if ((IS_GEN(dev_priv, 5) && dev_priv->wm.pri_latency[1] && > > - dev_priv->wm.spr_latency[1] && dev_priv->wm.cur_latency[1]) || > > - (!IS_GEN(dev_priv, 5) && dev_priv->wm.pri_latency[0] && > > - dev_priv->wm.spr_latency[0] && dev_priv->wm.cur_latency[0])) { > > - dev_priv->display.compute_pipe_wm = ilk_compute_pipe_wm; > > - dev_priv->display.compute_intermediate_wm = > > + if (INTEL_GEN(i915) >= 9) { > > + skl_setup_wm_latency(i915); > > + i915->display.initial_watermarks = skl_initial_wm; > > + i915->display.atomic_update_watermarks = skl_atomic_update_crtc_wm; > > + i915->display.compute_global_watermarks = skl_compute_wm; > > + } else if (HAS_PCH_SPLIT(i915)) { > > + ilk_setup_wm_latency(i915); > > + > > + if ((IS_GEN(i915, 5) && i915->wm.pri_latency[1] && > > + i915->wm.spr_latency[1] && i915->wm.cur_latency[1]) || > > + (!IS_GEN(i915, 5) && i915->wm.pri_latency[0] && > > + i915->wm.spr_latency[0] && i915->wm.cur_latency[0])) { > > + i915->display.compute_pipe_wm = ilk_compute_pipe_wm; > > + i915->display.compute_intermediate_wm = > > ilk_compute_intermediate_wm; > > - dev_priv->display.initial_watermarks = > > + i915->display.initial_watermarks = > > ilk_initial_watermarks; > > - dev_priv->display.optimize_watermarks = > > + i915->display.optimize_watermarks = > > ilk_optimize_watermarks; > > } else { > > DRM_DEBUG_KMS("Failed to read display plane latency. " > > "Disable CxSR\n"); > > } > > - } else if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) { > > - vlv_setup_wm_latency(dev_priv); > > - dev_priv->display.compute_pipe_wm = vlv_compute_pipe_wm; > > - dev_priv->display.compute_intermediate_wm = vlv_compute_intermediate_wm; > > - dev_priv->display.initial_watermarks = vlv_initial_watermarks; > > - dev_priv->display.optimize_watermarks = vlv_optimize_watermarks; > > - dev_priv->display.atomic_update_watermarks = vlv_atomic_update_fifo; > > - } else if (IS_G4X(dev_priv)) { > > - g4x_setup_wm_latency(dev_priv); > > - dev_priv->display.compute_pipe_wm = g4x_compute_pipe_wm; > > - dev_priv->display.compute_intermediate_wm = g4x_compute_intermediate_wm; > > - dev_priv->display.initial_watermarks = g4x_initial_watermarks; > > - dev_priv->display.optimize_watermarks = g4x_optimize_watermarks; > > - } else if (IS_PINEVIEW(dev_priv)) { > > - if (!intel_get_cxsr_latency(!IS_MOBILE(dev_priv), > > - dev_priv->is_ddr3, > > - dev_priv->fsb_freq, > > - dev_priv->mem_freq)) { > > + } else if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) { > > + vlv_setup_wm_latency(i915); > > + i915->display.compute_pipe_wm = vlv_compute_pipe_wm; > > + i915->display.compute_intermediate_wm = vlv_compute_intermediate_wm; > > + i915->display.initial_watermarks = vlv_initial_watermarks; > > + i915->display.optimize_watermarks = vlv_optimize_watermarks; > > + i915->display.atomic_update_watermarks = vlv_atomic_update_fifo; > > + } else if (IS_G4X(i915)) { > > + g4x_setup_wm_latency(i915); > > + i915->display.compute_pipe_wm = g4x_compute_pipe_wm; > > + i915->display.compute_intermediate_wm = g4x_compute_intermediate_wm; > > + i915->display.initial_watermarks = g4x_initial_watermarks; > > + i915->display.optimize_watermarks = g4x_optimize_watermarks; > > + } else if (IS_PINEVIEW(i915)) { > > + if (!intel_get_cxsr_latency(!IS_MOBILE(i915), > > + i915->is_ddr3, > > + i915->fsb_freq, > > + i915->mem_freq)) { > > DRM_INFO("failed to find known CxSR latency " > > "(found ddr%s fsb freq %d, mem freq %d), " > > "disabling CxSR\n", > > - (dev_priv->is_ddr3 == 1) ? "3" : "2", > > - dev_priv->fsb_freq, dev_priv->mem_freq); > > + (i915->is_ddr3 == 1) ? "3" : "2", > > + i915->fsb_freq, i915->mem_freq); > > /* Disable CxSR and never update its watermark again */ > > - intel_set_memory_cxsr(dev_priv, false); > > - dev_priv->display.update_wm = NULL; > > + intel_set_memory_cxsr(i915, false); > > + i915->display.update_wm = NULL; > > } else > > - dev_priv->display.update_wm = pineview_update_wm; > > - } else if (IS_GEN(dev_priv, 4)) { > > - dev_priv->display.update_wm = i965_update_wm; > > - } else if (IS_GEN(dev_priv, 3)) { > > - dev_priv->display.update_wm = i9xx_update_wm; > > - dev_priv->display.get_fifo_size = i9xx_get_fifo_size; > > - } else if (IS_GEN(dev_priv, 2)) { > > - if (INTEL_INFO(dev_priv)->num_pipes == 1) { > > - dev_priv->display.update_wm = i845_update_wm; > > - dev_priv->display.get_fifo_size = i845_get_fifo_size; > > + i915->display.update_wm = pineview_update_wm; > > + } else if (IS_GEN(i915, 4)) { > > + i915->display.update_wm = i965_update_wm; > > + } else if (IS_GEN(i915, 3)) { > > + i915->display.update_wm = i9xx_update_wm; > > + i915->display.get_fifo_size = i9xx_get_fifo_size; > > + } else if (IS_GEN(i915, 2)) { > > + if (INTEL_INFO(i915)->num_pipes == 1) { > > + i915->display.update_wm = i845_update_wm; > > + i915->display.get_fifo_size = i845_get_fifo_size; > > } else { > > - dev_priv->display.update_wm = i9xx_update_wm; > > - dev_priv->display.get_fifo_size = i830_get_fifo_size; > > + i915->display.update_wm = i9xx_update_wm; > > + i915->display.get_fifo_size = i830_get_fifo_size; > > } > > } else { > > DRM_ERROR("unexpected fall-through in intel_init_pm\n"); > > } > > } > > -static int byt_gpu_freq(struct drm_i915_private *dev_priv, int val) > > +static int byt_gpu_freq(struct drm_i915_private *i915, int val) > > { > > - struct intel_rps *rps = &dev_priv->gt_pm.rps; > > + struct intel_rps *rps = &i915->gt_pm.rps; > > /* > > * N = val - 0xb7 > > @@ -9756,16 +9756,16 @@ static int byt_gpu_freq(struct drm_i915_private *dev_priv, int val) > > return DIV_ROUND_CLOSEST(rps->gpll_ref_freq * (val - 0xb7), 1000); > > } > > -static int byt_freq_opcode(struct drm_i915_private *dev_priv, int val) > > +static int byt_freq_opcode(struct drm_i915_private *i915, int val) > > { > > - struct intel_rps *rps = &dev_priv->gt_pm.rps; > > + struct intel_rps *rps = &i915->gt_pm.rps; > > return DIV_ROUND_CLOSEST(1000 * val, rps->gpll_ref_freq) + 0xb7; > > } > > -static int chv_gpu_freq(struct drm_i915_private *dev_priv, int val) > > +static int chv_gpu_freq(struct drm_i915_private *i915, int val) > > { > > - struct intel_rps *rps = &dev_priv->gt_pm.rps; > > + struct intel_rps *rps = &i915->gt_pm.rps; > > /* > > * N = val / 2 > > @@ -9774,52 +9774,52 @@ static int chv_gpu_freq(struct drm_i915_private *dev_priv, int val) > > return DIV_ROUND_CLOSEST(rps->gpll_ref_freq * val, 2 * 2 * 1000); > > } > > -static int chv_freq_opcode(struct drm_i915_private *dev_priv, int val) > > +static int chv_freq_opcode(struct drm_i915_private *i915, int val) > > { > > - struct intel_rps *rps = &dev_priv->gt_pm.rps; > > + struct intel_rps *rps = &i915->gt_pm.rps; > > /* CHV needs even values */ > > return DIV_ROUND_CLOSEST(2 * 1000 * val, rps->gpll_ref_freq) * 2; > > } > > -int intel_gpu_freq(struct drm_i915_private *dev_priv, int val) > > +int intel_gpu_freq(struct drm_i915_private *i915, int val) > > { > > - if (INTEL_GEN(dev_priv) >= 9) > > + if (INTEL_GEN(i915) >= 9) > > return DIV_ROUND_CLOSEST(val * GT_FREQUENCY_MULTIPLIER, > > GEN9_FREQ_SCALER); > > - else if (IS_CHERRYVIEW(dev_priv)) > > - return chv_gpu_freq(dev_priv, val); > > - else if (IS_VALLEYVIEW(dev_priv)) > > - return byt_gpu_freq(dev_priv, val); > > + else if (IS_CHERRYVIEW(i915)) > > + return chv_gpu_freq(i915, val); > > + else if (IS_VALLEYVIEW(i915)) > > + return byt_gpu_freq(i915, val); > > else > > return val * GT_FREQUENCY_MULTIPLIER; > > } > > -int intel_freq_opcode(struct drm_i915_private *dev_priv, int val) > > +int intel_freq_opcode(struct drm_i915_private *i915, int val) > > { > > - if (INTEL_GEN(dev_priv) >= 9) > > + if (INTEL_GEN(i915) >= 9) > > return DIV_ROUND_CLOSEST(val * GEN9_FREQ_SCALER, > > GT_FREQUENCY_MULTIPLIER); > > - else if (IS_CHERRYVIEW(dev_priv)) > > - return chv_freq_opcode(dev_priv, val); > > - else if (IS_VALLEYVIEW(dev_priv)) > > - return byt_freq_opcode(dev_priv, val); > > + else if (IS_CHERRYVIEW(i915)) > > + return chv_freq_opcode(i915, val); > > + else if (IS_VALLEYVIEW(i915)) > > + return byt_freq_opcode(i915, val); > > else > > return DIV_ROUND_CLOSEST(val, GT_FREQUENCY_MULTIPLIER); > > } > > -void intel_pm_setup(struct drm_i915_private *dev_priv) > > +void intel_pm_setup(struct drm_i915_private *i915) > > { > > - mutex_init(&dev_priv->gt_pm.rps.lock); > > - mutex_init(&dev_priv->gt_pm.rps.power.mutex); > > + mutex_init(&i915->gt_pm.rps.lock); > > + mutex_init(&i915->gt_pm.rps.power.mutex); > > - atomic_set(&dev_priv->gt_pm.rps.num_waiters, 0); > > + atomic_set(&i915->gt_pm.rps.num_waiters, 0); > > - dev_priv->runtime_pm.suspended = false; > > - atomic_set(&dev_priv->runtime_pm.wakeref_count, 0); > > + i915->runtime_pm.suspended = false; > > + atomic_set(&i915->runtime_pm.wakeref_count, 0); > > } > > -static u64 vlv_residency_raw(struct drm_i915_private *dev_priv, > > +static u64 vlv_residency_raw(struct drm_i915_private *i915, > > const i915_reg_t reg) > > { > > u32 lower, upper, tmp; > > @@ -9829,7 +9829,7 @@ static u64 vlv_residency_raw(struct drm_i915_private *dev_priv, > > * The register accessed do not need forcewake. We borrow > > * uncore lock to prevent concurrent access to range reg. > > */ > > - lockdep_assert_held(&dev_priv->uncore.lock); > > + lockdep_assert_held(&i915->uncore.lock); > > /* > > * vlv and chv residency counters are 40 bits in width. > > @@ -9865,17 +9865,17 @@ static u64 vlv_residency_raw(struct drm_i915_private *dev_priv, > > return lower | (u64)upper << 8; > > } > > -u64 intel_rc6_residency_ns(struct drm_i915_private *dev_priv, > > +u64 intel_rc6_residency_ns(struct drm_i915_private *i915, > > const i915_reg_t reg) > > { > > - struct intel_uncore *uncore = &dev_priv->uncore; > > + struct intel_uncore *uncore = &i915->uncore; > > u64 time_hw, prev_hw, overflow_hw; > > unsigned int fw_domains; > > unsigned long flags; > > unsigned int i; > > u32 mul, div; > > - if (!HAS_RC6(dev_priv)) > > + if (!HAS_RC6(i915)) > > return 0; > > /* > > @@ -9887,7 +9887,7 @@ u64 intel_rc6_residency_ns(struct drm_i915_private *dev_priv, > > */ > > i = (i915_mmio_reg_offset(reg) - > > i915_mmio_reg_offset(GEN6_GT_GFX_RC6_LOCKED)) / sizeof(u32); > > - if (WARN_ON_ONCE(i >= ARRAY_SIZE(dev_priv->gt_pm.rc6.cur_residency))) > > + if (WARN_ON_ONCE(i >= ARRAY_SIZE(i915->gt_pm.rc6.cur_residency))) > > return 0; > > fw_domains = intel_uncore_forcewake_for_reg(uncore, reg, FW_REG_READ); > > @@ -9896,14 +9896,14 @@ u64 intel_rc6_residency_ns(struct drm_i915_private *dev_priv, > > intel_uncore_forcewake_get__locked(uncore, fw_domains); > > /* On VLV and CHV, residency time is in CZ units rather than 1.28us */ > > - if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) { > > + if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) { > > mul = 1000000; > > - div = dev_priv->czclk_freq; > > + div = i915->czclk_freq; > > overflow_hw = BIT_ULL(40); > > - time_hw = vlv_residency_raw(dev_priv, reg); > > + time_hw = vlv_residency_raw(i915, reg); > > } else { > > /* 833.33ns units on Gen9LP, 1.28us elsewhere. */ > > - if (IS_GEN9_LP(dev_priv)) { > > + if (IS_GEN9_LP(i915)) { > > mul = 10000; > > div = 12; > > } else { > > @@ -9921,8 +9921,8 @@ u64 intel_rc6_residency_ns(struct drm_i915_private *dev_priv, > > * But relying on a sufficient frequency of queries otherwise counters > > * can still wrap. > > */ > > - prev_hw = dev_priv->gt_pm.rc6.prev_hw_residency[i]; > > - dev_priv->gt_pm.rc6.prev_hw_residency[i] = time_hw; > > + prev_hw = i915->gt_pm.rc6.prev_hw_residency[i]; > > + i915->gt_pm.rc6.prev_hw_residency[i] = time_hw; > > /* RC6 delta from last sample. */ > > if (time_hw >= prev_hw) > > @@ -9931,8 +9931,8 @@ u64 intel_rc6_residency_ns(struct drm_i915_private *dev_priv, > > time_hw += overflow_hw - prev_hw; > > /* Add delta to RC6 extended raw driver copy. */ > > - time_hw += dev_priv->gt_pm.rc6.cur_residency[i]; > > - dev_priv->gt_pm.rc6.cur_residency[i] = time_hw; > > + time_hw += i915->gt_pm.rc6.cur_residency[i]; > > + i915->gt_pm.rc6.cur_residency[i] = time_hw; > > intel_uncore_forcewake_put__locked(uncore, fw_domains); > > spin_unlock_irqrestore(&uncore->lock, flags); > > @@ -9940,19 +9940,19 @@ u64 intel_rc6_residency_ns(struct drm_i915_private *dev_priv, > > return mul_u64_u32_div(time_hw, mul, div); > > } > > -u64 intel_rc6_residency_us(struct drm_i915_private *dev_priv, > > +u64 intel_rc6_residency_us(struct drm_i915_private *i915, > > i915_reg_t reg) > > { > > - return DIV_ROUND_UP_ULL(intel_rc6_residency_ns(dev_priv, reg), 1000); > > + return DIV_ROUND_UP_ULL(intel_rc6_residency_ns(i915, reg), 1000); > > } > > -u32 intel_get_cagf(struct drm_i915_private *dev_priv, u32 rpstat) > > +u32 intel_get_cagf(struct drm_i915_private *i915, u32 rpstat) > > { > > u32 cagf; > > - if (INTEL_GEN(dev_priv) >= 9) > > + if (INTEL_GEN(i915) >= 9) > > cagf = (rpstat & GEN9_CAGF_MASK) >> GEN9_CAGF_SHIFT; > > - else if (IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv)) > > + else if (IS_HASWELL(i915) || IS_BROADWELL(i915)) > > cagf = (rpstat & HSW_CAGF_MASK) >> HSW_CAGF_SHIFT; > > else > > cagf = (rpstat & GEN6_CAGF_MASK) >> GEN6_CAGF_SHIFT; > > diff --git a/drivers/gpu/drm/i915/intel_pm.h b/drivers/gpu/drm/i915/intel_pm.h > > index 1b489fa399e1..e2c4edf263f7 100644 > > --- a/drivers/gpu/drm/i915/intel_pm.h > > +++ b/drivers/gpu/drm/i915/intel_pm.h > > @@ -22,39 +22,39 @@ struct skl_ddb_entry; > > struct skl_pipe_wm; > > struct skl_wm_level; > > -void intel_init_clock_gating(struct drm_i915_private *dev_priv); > > -void intel_suspend_hw(struct drm_i915_private *dev_priv); > > -int ilk_wm_max_level(const struct drm_i915_private *dev_priv); > > +void intel_init_clock_gating(struct drm_i915_private *i915); > > +void intel_suspend_hw(struct drm_i915_private *i915); > > +int ilk_wm_max_level(const struct drm_i915_private *i915); > > void intel_update_watermarks(struct intel_crtc *crtc); > > -void intel_init_pm(struct drm_i915_private *dev_priv); > > -void intel_init_clock_gating_hooks(struct drm_i915_private *dev_priv); > > -void intel_pm_setup(struct drm_i915_private *dev_priv); > > -void intel_gpu_ips_init(struct drm_i915_private *dev_priv); > > +void intel_init_pm(struct drm_i915_private *i915); > > +void intel_init_clock_gating_hooks(struct drm_i915_private *i915); > > +void intel_pm_setup(struct drm_i915_private *i915); > > +void intel_gpu_ips_init(struct drm_i915_private *i915); > > void intel_gpu_ips_teardown(void); > > -void intel_init_gt_powersave(struct drm_i915_private *dev_priv); > > -void intel_cleanup_gt_powersave(struct drm_i915_private *dev_priv); > > -void intel_sanitize_gt_powersave(struct drm_i915_private *dev_priv); > > -void intel_enable_gt_powersave(struct drm_i915_private *dev_priv); > > -void intel_disable_gt_powersave(struct drm_i915_private *dev_priv); > > -void gen6_rps_busy(struct drm_i915_private *dev_priv); > > -void gen6_rps_idle(struct drm_i915_private *dev_priv); > > +void intel_init_gt_powersave(struct drm_i915_private *i915); > > +void intel_cleanup_gt_powersave(struct drm_i915_private *i915); > > +void intel_sanitize_gt_powersave(struct drm_i915_private *i915); > > +void intel_enable_gt_powersave(struct drm_i915_private *i915); > > +void intel_disable_gt_powersave(struct drm_i915_private *i915); > > +void gen6_rps_busy(struct drm_i915_private *i915); > > +void gen6_rps_idle(struct drm_i915_private *i915); > > void gen6_rps_boost(struct i915_request *rq); > > -void g4x_wm_get_hw_state(struct drm_i915_private *dev_priv); > > -void vlv_wm_get_hw_state(struct drm_i915_private *dev_priv); > > -void ilk_wm_get_hw_state(struct drm_i915_private *dev_priv); > > -void skl_wm_get_hw_state(struct drm_i915_private *dev_priv); > > +void g4x_wm_get_hw_state(struct drm_i915_private *i915); > > +void vlv_wm_get_hw_state(struct drm_i915_private *i915); > > +void ilk_wm_get_hw_state(struct drm_i915_private *i915); > > +void skl_wm_get_hw_state(struct drm_i915_private *i915); > > void skl_pipe_ddb_get_hw_state(struct intel_crtc *crtc, > > struct skl_ddb_entry *ddb_y, > > struct skl_ddb_entry *ddb_uv); > > -void skl_ddb_get_hw_state(struct drm_i915_private *dev_priv, > > +void skl_ddb_get_hw_state(struct drm_i915_private *i915, > > struct skl_ddb_allocation *ddb /* out */); > > void skl_pipe_wm_get_hw_state(struct intel_crtc *crtc, > > struct skl_pipe_wm *out); > > -void g4x_wm_sanitize(struct drm_i915_private *dev_priv); > > -void vlv_wm_sanitize(struct drm_i915_private *dev_priv); > > +void g4x_wm_sanitize(struct drm_i915_private *i915); > > +void vlv_wm_sanitize(struct drm_i915_private *i915); > > bool intel_can_enable_sagv(struct drm_atomic_state *state); > > -int intel_enable_sagv(struct drm_i915_private *dev_priv); > > -int intel_disable_sagv(struct drm_i915_private *dev_priv); > > +int intel_enable_sagv(struct drm_i915_private *i915); > > +int intel_disable_sagv(struct drm_i915_private *i915); > > bool skl_wm_level_equals(const struct skl_wm_level *l1, > > const struct skl_wm_level *l2); > > bool skl_ddb_allocation_overlaps(const struct skl_ddb_entry *ddb, > > @@ -67,24 +67,24 @@ void skl_write_cursor_wm(struct intel_plane *plane, > > bool ilk_disable_lp_wm(struct drm_device *dev); > > int skl_check_pipe_max_pixel_rate(struct intel_crtc *intel_crtc, > > struct intel_crtc_state *cstate); > > -void intel_init_ipc(struct drm_i915_private *dev_priv); > > -void intel_enable_ipc(struct drm_i915_private *dev_priv); > > +void intel_init_ipc(struct drm_i915_private *i915); > > +void intel_enable_ipc(struct drm_i915_private *i915); > > -int intel_gpu_freq(struct drm_i915_private *dev_priv, int val); > > -int intel_freq_opcode(struct drm_i915_private *dev_priv, int val); > > -u64 intel_rc6_residency_ns(struct drm_i915_private *dev_priv, i915_reg_t reg); > > -u64 intel_rc6_residency_us(struct drm_i915_private *dev_priv, i915_reg_t reg); > > +int intel_gpu_freq(struct drm_i915_private *i915, int val); > > +int intel_freq_opcode(struct drm_i915_private *i915, int val); > > +u64 intel_rc6_residency_ns(struct drm_i915_private *i915, i915_reg_t reg); > > +u64 intel_rc6_residency_us(struct drm_i915_private *i915, i915_reg_t reg); > > -u32 intel_get_cagf(struct drm_i915_private *dev_priv, u32 rpstat1); > > +u32 intel_get_cagf(struct drm_i915_private *i915, u32 rpstat1); > > -unsigned long i915_chipset_val(struct drm_i915_private *dev_priv); > > -unsigned long i915_mch_val(struct drm_i915_private *dev_priv); > > -unsigned long i915_gfx_val(struct drm_i915_private *dev_priv); > > -void i915_update_gfx_val(struct drm_i915_private *dev_priv); > > +unsigned long i915_chipset_val(struct drm_i915_private *i915); > > +unsigned long i915_mch_val(struct drm_i915_private *i915); > > +unsigned long i915_gfx_val(struct drm_i915_private *i915); > > +void i915_update_gfx_val(struct drm_i915_private *i915); > > -bool ironlake_set_drps(struct drm_i915_private *dev_priv, u8 val); > > -int intel_set_rps(struct drm_i915_private *dev_priv, u8 val); > > +bool ironlake_set_drps(struct drm_i915_private *i915, u8 val); > > +int intel_set_rps(struct drm_i915_private *i915, u8 val); > > void intel_rps_mark_interactive(struct drm_i915_private *i915, bool interactive); > > -bool intel_set_memory_cxsr(struct drm_i915_private *dev_priv, bool enable); > > +bool intel_set_memory_cxsr(struct drm_i915_private *i915, bool enable); > > #endif /* __INTEL_PM_H__ */ > > diff --git a/drivers/gpu/drm/i915/intel_psr.c b/drivers/gpu/drm/i915/intel_psr.c > > index 01ca502099df..072eea00b17a 100644 > > --- a/drivers/gpu/drm/i915/intel_psr.c > > +++ b/drivers/gpu/drm/i915/intel_psr.c > > @@ -71,14 +71,14 @@ static bool psr_global_enabled(u32 debug) > > } > > } > > -static bool intel_psr2_enabled(struct drm_i915_private *dev_priv, > > +static bool intel_psr2_enabled(struct drm_i915_private *i915, > > const struct intel_crtc_state *crtc_state) > > { > > /* Cannot enable DSC and PSR2 simultaneously */ > > WARN_ON(crtc_state->dsc_params.compression_enable && > > crtc_state->has_psr2); > > - switch (dev_priv->psr.debug & I915_PSR_DEBUG_MODE_MASK) { > > + switch (i915->psr.debug & I915_PSR_DEBUG_MODE_MASK) { > > case I915_PSR_DEBUG_DISABLE: > > case I915_PSR_DEBUG_FORCE_PSR1: > > return false; > > @@ -104,20 +104,20 @@ static int edp_psr_shift(enum transcoder cpu_transcoder) > > } > > } > > -void intel_psr_irq_control(struct drm_i915_private *dev_priv, u32 debug) > > +void intel_psr_irq_control(struct drm_i915_private *i915, u32 debug) > > { > > u32 debug_mask, mask; > > enum transcoder cpu_transcoder; > > u32 transcoders = BIT(TRANSCODER_EDP); > > - if (INTEL_GEN(dev_priv) >= 8) > > + if (INTEL_GEN(i915) >= 8) > > transcoders |= BIT(TRANSCODER_A) | > > BIT(TRANSCODER_B) | > > BIT(TRANSCODER_C); > > debug_mask = 0; > > mask = 0; > > - for_each_cpu_transcoder_masked(dev_priv, cpu_transcoder, transcoders) { > > + for_each_cpu_transcoder_masked(i915, cpu_transcoder, transcoders) { > > int shift = edp_psr_shift(cpu_transcoder); > > mask |= EDP_PSR_ERROR(shift); > > @@ -168,26 +168,26 @@ static void psr_event_print(u32 val, bool psr2_enabled) > > DRM_DEBUG_KMS("\tPSR disabled\n"); > > } > > -void intel_psr_irq_handler(struct drm_i915_private *dev_priv, u32 psr_iir) > > +void intel_psr_irq_handler(struct drm_i915_private *i915, u32 psr_iir) > > { > > u32 transcoders = BIT(TRANSCODER_EDP); > > enum transcoder cpu_transcoder; > > ktime_t time_ns = ktime_get(); > > u32 mask = 0; > > - if (INTEL_GEN(dev_priv) >= 8) > > + if (INTEL_GEN(i915) >= 8) > > transcoders |= BIT(TRANSCODER_A) | > > BIT(TRANSCODER_B) | > > BIT(TRANSCODER_C); > > - for_each_cpu_transcoder_masked(dev_priv, cpu_transcoder, transcoders) { > > + for_each_cpu_transcoder_masked(i915, cpu_transcoder, transcoders) { > > int shift = edp_psr_shift(cpu_transcoder); > > if (psr_iir & EDP_PSR_ERROR(shift)) { > > DRM_WARN("[transcoder %s] PSR aux error\n", > > transcoder_name(cpu_transcoder)); > > - dev_priv->psr.irq_aux_error = true; > > + i915->psr.irq_aux_error = true; > > /* > > * If this interruption is not masked it will keep > > @@ -201,19 +201,19 @@ void intel_psr_irq_handler(struct drm_i915_private *dev_priv, u32 psr_iir) > > } > > if (psr_iir & EDP_PSR_PRE_ENTRY(shift)) { > > - dev_priv->psr.last_entry_attempt = time_ns; > > + i915->psr.last_entry_attempt = time_ns; > > DRM_DEBUG_KMS("[transcoder %s] PSR entry attempt in 2 vblanks\n", > > transcoder_name(cpu_transcoder)); > > } > > if (psr_iir & EDP_PSR_POST_EXIT(shift)) { > > - dev_priv->psr.last_exit = time_ns; > > + i915->psr.last_exit = time_ns; > > DRM_DEBUG_KMS("[transcoder %s] PSR exit completed\n", > > transcoder_name(cpu_transcoder)); > > - if (INTEL_GEN(dev_priv) >= 9) { > > + if (INTEL_GEN(i915) >= 9) { > > u32 val = I915_READ(PSR_EVENT(cpu_transcoder)); > > - bool psr2_enabled = dev_priv->psr.psr2_enabled; > > + bool psr2_enabled = i915->psr.psr2_enabled; > > I915_WRITE(PSR_EVENT(cpu_transcoder), val); > > psr_event_print(val, psr2_enabled); > > @@ -225,7 +225,7 @@ void intel_psr_irq_handler(struct drm_i915_private *dev_priv, u32 psr_iir) > > mask |= I915_READ(EDP_PSR_IMR); > > I915_WRITE(EDP_PSR_IMR, mask); > > - schedule_work(&dev_priv->psr.work); > > + schedule_work(&i915->psr.work); > > } > > } > > @@ -279,7 +279,7 @@ static u16 intel_dp_get_su_x_granulartiy(struct intel_dp *intel_dp) > > void intel_psr_init_dpcd(struct intel_dp *intel_dp) > > { > > - struct drm_i915_private *dev_priv = > > + struct drm_i915_private *i915 = > > to_i915(dp_to_dig_port(intel_dp)->base.base.dev); > > drm_dp_dpcd_read(&intel_dp->aux, DP_PSR_SUPPORT, intel_dp->psr_dpcd, > > @@ -300,14 +300,14 @@ void intel_psr_init_dpcd(struct intel_dp *intel_dp) > > return; > > } > > - dev_priv->psr.sink_support = true; > > - dev_priv->psr.sink_sync_latency = > > + i915->psr.sink_support = true; > > + i915->psr.sink_sync_latency = > > intel_dp_get_sink_sync_latency(intel_dp); > > - WARN_ON(dev_priv->psr.dp); > > - dev_priv->psr.dp = intel_dp; > > + WARN_ON(i915->psr.dp); > > + i915->psr.dp = intel_dp; > > - if (INTEL_GEN(dev_priv) >= 9 && > > + if (INTEL_GEN(i915) >= 9 && > > (intel_dp->psr_dpcd[0] == DP_PSR2_WITH_Y_COORD_IS_SUPPORTED)) { > > bool y_req = intel_dp->psr_dpcd[1] & > > DP_PSR2_SU_Y_COORDINATE_REQUIRED; > > @@ -324,14 +324,14 @@ void intel_psr_init_dpcd(struct intel_dp *intel_dp) > > * Y-coordinate requirement panels we would need to enable > > * GTC first. > > */ > > - dev_priv->psr.sink_psr2_support = y_req && alpm; > > + i915->psr.sink_psr2_support = y_req && alpm; > > DRM_DEBUG_KMS("PSR2 %ssupported\n", > > - dev_priv->psr.sink_psr2_support ? "" : "not "); > > + i915->psr.sink_psr2_support ? "" : "not "); > > - if (dev_priv->psr.sink_psr2_support) { > > - dev_priv->psr.colorimetry_support = > > + if (i915->psr.sink_psr2_support) { > > + i915->psr.colorimetry_support = > > intel_dp_get_colorimetry_status(intel_dp); > > - dev_priv->psr.su_x_granularity = > > + i915->psr.su_x_granularity = > > intel_dp_get_su_x_granulartiy(intel_dp); > > } > > } > > @@ -341,15 +341,15 @@ static void intel_psr_setup_vsc(struct intel_dp *intel_dp, > > const struct intel_crtc_state *crtc_state) > > { > > struct intel_digital_port *intel_dig_port = dp_to_dig_port(intel_dp); > > - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp); > > + struct drm_i915_private *i915 = dp_to_i915(intel_dp); > > struct dp_sdp psr_vsc; > > - if (dev_priv->psr.psr2_enabled) { > > + if (i915->psr.psr2_enabled) { > > /* Prepare VSC Header for SU as per EDP 1.4 spec, Table 6.11 */ > > memset(&psr_vsc, 0, sizeof(psr_vsc)); > > psr_vsc.sdp_header.HB0 = 0; > > psr_vsc.sdp_header.HB1 = 0x7; > > - if (dev_priv->psr.colorimetry_support) { > > + if (i915->psr.colorimetry_support) { > > psr_vsc.sdp_header.HB2 = 0x5; > > psr_vsc.sdp_header.HB3 = 0x13; > > } else { > > @@ -372,7 +372,7 @@ static void intel_psr_setup_vsc(struct intel_dp *intel_dp, > > static void hsw_psr_setup_aux(struct intel_dp *intel_dp) > > { > > - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp); > > + struct drm_i915_private *i915 = dp_to_i915(intel_dp); > > u32 aux_clock_divider, aux_ctl; > > int i; > > static const u8 aux_msg[] = { > > @@ -405,19 +405,19 @@ static void hsw_psr_setup_aux(struct intel_dp *intel_dp) > > static void intel_psr_enable_sink(struct intel_dp *intel_dp) > > { > > - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp); > > + struct drm_i915_private *i915 = dp_to_i915(intel_dp); > > u8 dpcd_val = DP_PSR_ENABLE; > > /* Enable ALPM at sink for psr2 */ > > - if (dev_priv->psr.psr2_enabled) { > > + if (i915->psr.psr2_enabled) { > > drm_dp_dpcd_writeb(&intel_dp->aux, DP_RECEIVER_ALPM_CONFIG, > > DP_ALPM_ENABLE); > > dpcd_val |= DP_PSR_ENABLE_PSR2 | DP_PSR_IRQ_HPD_WITH_CRC_ERRORS; > > } else { > > - if (dev_priv->psr.link_standby) > > + if (i915->psr.link_standby) > > dpcd_val |= DP_PSR_MAIN_LINK_ACTIVE; > > - if (INTEL_GEN(dev_priv) >= 8) > > + if (INTEL_GEN(i915) >= 8) > > dpcd_val |= DP_PSR_CRC_VERIFICATION; > > } > > @@ -428,26 +428,26 @@ static void intel_psr_enable_sink(struct intel_dp *intel_dp) > > static u32 intel_psr1_get_tp_time(struct intel_dp *intel_dp) > > { > > - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp); > > + struct drm_i915_private *i915 = dp_to_i915(intel_dp); > > u32 val = 0; > > - if (INTEL_GEN(dev_priv) >= 11) > > + if (INTEL_GEN(i915) >= 11) > > val |= EDP_PSR_TP4_TIME_0US; > > - if (dev_priv->vbt.psr.tp1_wakeup_time_us == 0) > > + if (i915->vbt.psr.tp1_wakeup_time_us == 0) > > val |= EDP_PSR_TP1_TIME_0us; > > - else if (dev_priv->vbt.psr.tp1_wakeup_time_us <= 100) > > + else if (i915->vbt.psr.tp1_wakeup_time_us <= 100) > > val |= EDP_PSR_TP1_TIME_100us; > > - else if (dev_priv->vbt.psr.tp1_wakeup_time_us <= 500) > > + else if (i915->vbt.psr.tp1_wakeup_time_us <= 500) > > val |= EDP_PSR_TP1_TIME_500us; > > else > > val |= EDP_PSR_TP1_TIME_2500us; > > - if (dev_priv->vbt.psr.tp2_tp3_wakeup_time_us == 0) > > + if (i915->vbt.psr.tp2_tp3_wakeup_time_us == 0) > > val |= EDP_PSR_TP2_TP3_TIME_0us; > > - else if (dev_priv->vbt.psr.tp2_tp3_wakeup_time_us <= 100) > > + else if (i915->vbt.psr.tp2_tp3_wakeup_time_us <= 100) > > val |= EDP_PSR_TP2_TP3_TIME_100us; > > - else if (dev_priv->vbt.psr.tp2_tp3_wakeup_time_us <= 500) > > + else if (i915->vbt.psr.tp2_tp3_wakeup_time_us <= 500) > > val |= EDP_PSR_TP2_TP3_TIME_500us; > > else > > val |= EDP_PSR_TP2_TP3_TIME_2500us; > > @@ -463,31 +463,31 @@ static u32 intel_psr1_get_tp_time(struct intel_dp *intel_dp) > > static void hsw_activate_psr1(struct intel_dp *intel_dp) > > { > > - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp); > > + struct drm_i915_private *i915 = dp_to_i915(intel_dp); > > u32 max_sleep_time = 0x1f; > > u32 val = EDP_PSR_ENABLE; > > /* Let's use 6 as the minimum to cover all known cases including the > > * off-by-one issue that HW has in some cases. > > */ > > - int idle_frames = max(6, dev_priv->vbt.psr.idle_frames); > > + int idle_frames = max(6, i915->vbt.psr.idle_frames); > > /* sink_sync_latency of 8 means source has to wait for more than 8 > > * frames, we'll go with 9 frames for now > > */ > > - idle_frames = max(idle_frames, dev_priv->psr.sink_sync_latency + 1); > > + idle_frames = max(idle_frames, i915->psr.sink_sync_latency + 1); > > val |= idle_frames << EDP_PSR_IDLE_FRAME_SHIFT; > > val |= max_sleep_time << EDP_PSR_MAX_SLEEP_TIME_SHIFT; > > - if (IS_HASWELL(dev_priv)) > > + if (IS_HASWELL(i915)) > > val |= EDP_PSR_MIN_LINK_ENTRY_TIME_8_LINES; > > - if (dev_priv->psr.link_standby) > > + if (i915->psr.link_standby) > > val |= EDP_PSR_LINK_STANDBY; > > val |= intel_psr1_get_tp_time(intel_dp); > > - if (INTEL_GEN(dev_priv) >= 8) > > + if (INTEL_GEN(i915) >= 8) > > val |= EDP_PSR_CRC_ENABLE; > > val |= I915_READ(EDP_PSR_CTL) & EDP_PSR_RESTORE_PSR_ACTIVE_CTX_MASK; > > @@ -496,29 +496,29 @@ static void hsw_activate_psr1(struct intel_dp *intel_dp) > > static void hsw_activate_psr2(struct intel_dp *intel_dp) > > { > > - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp); > > + struct drm_i915_private *i915 = dp_to_i915(intel_dp); > > u32 val; > > /* Let's use 6 as the minimum to cover all known cases including the > > * off-by-one issue that HW has in some cases. > > */ > > - int idle_frames = max(6, dev_priv->vbt.psr.idle_frames); > > + int idle_frames = max(6, i915->vbt.psr.idle_frames); > > - idle_frames = max(idle_frames, dev_priv->psr.sink_sync_latency + 1); > > + idle_frames = max(idle_frames, i915->psr.sink_sync_latency + 1); > > val = idle_frames << EDP_PSR2_IDLE_FRAME_SHIFT; > > val |= EDP_PSR2_ENABLE | EDP_SU_TRACK_ENABLE; > > - if (INTEL_GEN(dev_priv) >= 10 || IS_GEMINILAKE(dev_priv)) > > + if (INTEL_GEN(i915) >= 10 || IS_GEMINILAKE(i915)) > > val |= EDP_Y_COORDINATE_ENABLE; > > - val |= EDP_PSR2_FRAME_BEFORE_SU(dev_priv->psr.sink_sync_latency + 1); > > + val |= EDP_PSR2_FRAME_BEFORE_SU(i915->psr.sink_sync_latency + 1); > > - if (dev_priv->vbt.psr.psr2_tp2_tp3_wakeup_time_us >= 0 && > > - dev_priv->vbt.psr.psr2_tp2_tp3_wakeup_time_us <= 50) > > + if (i915->vbt.psr.psr2_tp2_tp3_wakeup_time_us >= 0 && > > + i915->vbt.psr.psr2_tp2_tp3_wakeup_time_us <= 50) > > val |= EDP_PSR2_TP2_TIME_50us; > > - else if (dev_priv->vbt.psr.psr2_tp2_tp3_wakeup_time_us <= 100) > > + else if (i915->vbt.psr.psr2_tp2_tp3_wakeup_time_us <= 100) > > val |= EDP_PSR2_TP2_TIME_100us; > > - else if (dev_priv->vbt.psr.psr2_tp2_tp3_wakeup_time_us <= 500) > > + else if (i915->vbt.psr.psr2_tp2_tp3_wakeup_time_us <= 500) > > val |= EDP_PSR2_TP2_TIME_500us; > > else > > val |= EDP_PSR2_TP2_TIME_2500us; > > @@ -535,12 +535,12 @@ static void hsw_activate_psr2(struct intel_dp *intel_dp) > > static bool intel_psr2_config_valid(struct intel_dp *intel_dp, > > struct intel_crtc_state *crtc_state) > > { > > - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp); > > + struct drm_i915_private *i915 = dp_to_i915(intel_dp); > > int crtc_hdisplay = crtc_state->base.adjusted_mode.crtc_hdisplay; > > int crtc_vdisplay = crtc_state->base.adjusted_mode.crtc_vdisplay; > > int psr_max_h = 0, psr_max_v = 0; > > - if (!dev_priv->psr.sink_psr2_support) > > + if (!i915->psr.sink_psr2_support) > > return false; > > /* > > @@ -553,10 +553,10 @@ static bool intel_psr2_config_valid(struct intel_dp *intel_dp, > > return false; > > } > > - if (INTEL_GEN(dev_priv) >= 10 || IS_GEMINILAKE(dev_priv)) { > > + if (INTEL_GEN(i915) >= 10 || IS_GEMINILAKE(i915)) { > > psr_max_h = 4096; > > psr_max_v = 2304; > > - } else if (IS_GEN(dev_priv, 9)) { > > + } else if (IS_GEN(i915, 9)) { > > psr_max_h = 3640; > > psr_max_v = 2304; > > } > > @@ -574,9 +574,9 @@ static bool intel_psr2_config_valid(struct intel_dp *intel_dp, > > * only need to validate the SU block width is a multiple of > > * x granularity. > > */ > > - if (crtc_hdisplay % dev_priv->psr.su_x_granularity) { > > + if (crtc_hdisplay % i915->psr.su_x_granularity) { > > DRM_DEBUG_KMS("PSR2 not enabled, hdisplay(%d) not multiple of %d\n", > > - crtc_hdisplay, dev_priv->psr.su_x_granularity); > > + crtc_hdisplay, i915->psr.su_x_granularity); > > return false; > > } > > @@ -592,15 +592,15 @@ void intel_psr_compute_config(struct intel_dp *intel_dp, > > struct intel_crtc_state *crtc_state) > > { > > struct intel_digital_port *dig_port = dp_to_dig_port(intel_dp); > > - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp); > > + struct drm_i915_private *i915 = dp_to_i915(intel_dp); > > const struct drm_display_mode *adjusted_mode = > > &crtc_state->base.adjusted_mode; > > int psr_setup_time; > > - if (!CAN_PSR(dev_priv)) > > + if (!CAN_PSR(i915)) > > return; > > - if (intel_dp != dev_priv->psr.dp) > > + if (intel_dp != i915->psr.dp) > > return; > > /* > > @@ -615,7 +615,7 @@ void intel_psr_compute_config(struct intel_dp *intel_dp, > > return; > > } > > - if (dev_priv->psr.sink_not_reliable) { > > + if (i915->psr.sink_not_reliable) { > > DRM_DEBUG_KMS("PSR sink implementation is not reliable\n"); > > return; > > } > > @@ -645,24 +645,24 @@ void intel_psr_compute_config(struct intel_dp *intel_dp, > > static void intel_psr_activate(struct intel_dp *intel_dp) > > { > > - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp); > > + struct drm_i915_private *i915 = dp_to_i915(intel_dp); > > - if (INTEL_GEN(dev_priv) >= 9) > > + if (INTEL_GEN(i915) >= 9) > > WARN_ON(I915_READ(EDP_PSR2_CTL) & EDP_PSR2_ENABLE); > > WARN_ON(I915_READ(EDP_PSR_CTL) & EDP_PSR_ENABLE); > > - WARN_ON(dev_priv->psr.active); > > - lockdep_assert_held(&dev_priv->psr.lock); > > + WARN_ON(i915->psr.active); > > + lockdep_assert_held(&i915->psr.lock); > > /* psr1 and psr2 are mutually exclusive.*/ > > - if (dev_priv->psr.psr2_enabled) > > + if (i915->psr.psr2_enabled) > > hsw_activate_psr2(intel_dp); > > else > > hsw_activate_psr1(intel_dp); > > - dev_priv->psr.active = true; > > + i915->psr.active = true; > > } > > -static i915_reg_t gen9_chicken_trans_reg(struct drm_i915_private *dev_priv, > > +static i915_reg_t gen9_chicken_trans_reg(struct drm_i915_private *i915, > > enum transcoder cpu_transcoder) > > { > > static const i915_reg_t regs[] = { > > @@ -672,7 +672,7 @@ static i915_reg_t gen9_chicken_trans_reg(struct drm_i915_private *dev_priv, > > [TRANSCODER_EDP] = CHICKEN_TRANS_EDP, > > }; > > - WARN_ON(INTEL_GEN(dev_priv) < 9); > > + WARN_ON(INTEL_GEN(i915) < 9); > > if (WARN_ON(cpu_transcoder >= ARRAY_SIZE(regs) || > > !regs[cpu_transcoder].reg)) > > @@ -684,19 +684,19 @@ static i915_reg_t gen9_chicken_trans_reg(struct drm_i915_private *dev_priv, > > static void intel_psr_enable_source(struct intel_dp *intel_dp, > > const struct intel_crtc_state *crtc_state) > > { > > - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp); > > + struct drm_i915_private *i915 = dp_to_i915(intel_dp); > > enum transcoder cpu_transcoder = crtc_state->cpu_transcoder; > > u32 mask; > > /* Only HSW and BDW have PSR AUX registers that need to be setup. SKL+ > > * use hardcoded values PSR AUX transactions > > */ > > - if (IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv)) > > + if (IS_HASWELL(i915) || IS_BROADWELL(i915)) > > hsw_psr_setup_aux(intel_dp); > > - if (dev_priv->psr.psr2_enabled && (IS_GEN(dev_priv, 9) && > > - !IS_GEMINILAKE(dev_priv))) { > > - i915_reg_t reg = gen9_chicken_trans_reg(dev_priv, > > + if (i915->psr.psr2_enabled && (IS_GEN(i915, 9) && > > + !IS_GEMINILAKE(i915))) { > > + i915_reg_t reg = gen9_chicken_trans_reg(i915, > > cpu_transcoder); > > u32 chicken = I915_READ(reg); > > @@ -716,29 +716,29 @@ static void intel_psr_enable_source(struct intel_dp *intel_dp, > > EDP_PSR_DEBUG_MASK_LPSP | > > EDP_PSR_DEBUG_MASK_MAX_SLEEP; > > - if (INTEL_GEN(dev_priv) < 11) > > + if (INTEL_GEN(i915) < 11) > > mask |= EDP_PSR_DEBUG_MASK_DISP_REG_WRITE; > > I915_WRITE(EDP_PSR_DEBUG, mask); > > } > > -static void intel_psr_enable_locked(struct drm_i915_private *dev_priv, > > +static void intel_psr_enable_locked(struct drm_i915_private *i915, > > const struct intel_crtc_state *crtc_state) > > { > > - struct intel_dp *intel_dp = dev_priv->psr.dp; > > + struct intel_dp *intel_dp = i915->psr.dp; > > - WARN_ON(dev_priv->psr.enabled); > > + WARN_ON(i915->psr.enabled); > > - dev_priv->psr.psr2_enabled = intel_psr2_enabled(dev_priv, crtc_state); > > - dev_priv->psr.busy_frontbuffer_bits = 0; > > - dev_priv->psr.pipe = to_intel_crtc(crtc_state->base.crtc)->pipe; > > + i915->psr.psr2_enabled = intel_psr2_enabled(i915, crtc_state); > > + i915->psr.busy_frontbuffer_bits = 0; > > + i915->psr.pipe = to_intel_crtc(crtc_state->base.crtc)->pipe; > > DRM_DEBUG_KMS("Enabling PSR%s\n", > > - dev_priv->psr.psr2_enabled ? "2" : "1"); > > + i915->psr.psr2_enabled ? "2" : "1"); > > intel_psr_setup_vsc(intel_dp, crtc_state); > > intel_psr_enable_sink(intel_dp); > > intel_psr_enable_source(intel_dp, crtc_state); > > - dev_priv->psr.enabled = true; > > + i915->psr.enabled = true; > > intel_psr_activate(intel_dp); > > } > > @@ -753,41 +753,41 @@ static void intel_psr_enable_locked(struct drm_i915_private *dev_priv, > > void intel_psr_enable(struct intel_dp *intel_dp, > > const struct intel_crtc_state *crtc_state) > > { > > - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp); > > + struct drm_i915_private *i915 = dp_to_i915(intel_dp); > > if (!crtc_state->has_psr) > > return; > > - if (WARN_ON(!CAN_PSR(dev_priv))) > > + if (WARN_ON(!CAN_PSR(i915))) > > return; > > - WARN_ON(dev_priv->drrs.dp); > > + WARN_ON(i915->drrs.dp); > > - mutex_lock(&dev_priv->psr.lock); > > + mutex_lock(&i915->psr.lock); > > - if (!psr_global_enabled(dev_priv->psr.debug)) { > > + if (!psr_global_enabled(i915->psr.debug)) { > > DRM_DEBUG_KMS("PSR disabled by flag\n"); > > goto unlock; > > } > > - intel_psr_enable_locked(dev_priv, crtc_state); > > + intel_psr_enable_locked(i915, crtc_state); > > unlock: > > - mutex_unlock(&dev_priv->psr.lock); > > + mutex_unlock(&i915->psr.lock); > > } > > -static void intel_psr_exit(struct drm_i915_private *dev_priv) > > +static void intel_psr_exit(struct drm_i915_private *i915) > > { > > u32 val; > > - if (!dev_priv->psr.active) { > > - if (INTEL_GEN(dev_priv) >= 9) > > + if (!i915->psr.active) { > > + if (INTEL_GEN(i915) >= 9) > > WARN_ON(I915_READ(EDP_PSR2_CTL) & EDP_PSR2_ENABLE); > > WARN_ON(I915_READ(EDP_PSR_CTL) & EDP_PSR_ENABLE); > > return; > > } > > - if (dev_priv->psr.psr2_enabled) { > > + if (i915->psr.psr2_enabled) { > > val = I915_READ(EDP_PSR2_CTL); > > WARN_ON(!(val & EDP_PSR2_ENABLE)); > > I915_WRITE(EDP_PSR2_CTL, val & ~EDP_PSR2_ENABLE); > > @@ -796,26 +796,26 @@ static void intel_psr_exit(struct drm_i915_private *dev_priv) > > WARN_ON(!(val & EDP_PSR_ENABLE)); > > I915_WRITE(EDP_PSR_CTL, val & ~EDP_PSR_ENABLE); > > } > > - dev_priv->psr.active = false; > > + i915->psr.active = false; > > } > > static void intel_psr_disable_locked(struct intel_dp *intel_dp) > > { > > - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp); > > + struct drm_i915_private *i915 = dp_to_i915(intel_dp); > > i915_reg_t psr_status; > > u32 psr_status_mask; > > - lockdep_assert_held(&dev_priv->psr.lock); > > + lockdep_assert_held(&i915->psr.lock); > > - if (!dev_priv->psr.enabled) > > + if (!i915->psr.enabled) > > return; > > DRM_DEBUG_KMS("Disabling PSR%s\n", > > - dev_priv->psr.psr2_enabled ? "2" : "1"); > > + i915->psr.psr2_enabled ? "2" : "1"); > > - intel_psr_exit(dev_priv); > > + intel_psr_exit(i915); > > - if (dev_priv->psr.psr2_enabled) { > > + if (i915->psr.psr2_enabled) { > > psr_status = EDP_PSR2_STATUS; > > psr_status_mask = EDP_PSR2_STATUS_STATE_MASK; > > } else { > > @@ -824,14 +824,14 @@ static void intel_psr_disable_locked(struct intel_dp *intel_dp) > > } > > /* Wait till PSR is idle */ > > - if (intel_wait_for_register(&dev_priv->uncore, > > + if (intel_wait_for_register(&i915->uncore, > > psr_status, psr_status_mask, 0, 2000)) > > DRM_ERROR("Timed out waiting PSR idle state\n"); > > /* Disable PSR on Sink */ > > drm_dp_dpcd_writeb(&intel_dp->aux, DP_PSR_EN_CFG, 0); > > - dev_priv->psr.enabled = false; > > + i915->psr.enabled = false; > > } > > /** > > @@ -844,23 +844,23 @@ static void intel_psr_disable_locked(struct intel_dp *intel_dp) > > void intel_psr_disable(struct intel_dp *intel_dp, > > const struct intel_crtc_state *old_crtc_state) > > { > > - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp); > > + struct drm_i915_private *i915 = dp_to_i915(intel_dp); > > if (!old_crtc_state->has_psr) > > return; > > - if (WARN_ON(!CAN_PSR(dev_priv))) > > + if (WARN_ON(!CAN_PSR(i915))) > > return; > > - mutex_lock(&dev_priv->psr.lock); > > + mutex_lock(&i915->psr.lock); > > intel_psr_disable_locked(intel_dp); > > - mutex_unlock(&dev_priv->psr.lock); > > - cancel_work_sync(&dev_priv->psr.work); > > + mutex_unlock(&i915->psr.lock); > > + cancel_work_sync(&i915->psr.work); > > } > > -static void psr_force_hw_tracking_exit(struct drm_i915_private *dev_priv) > > +static void psr_force_hw_tracking_exit(struct drm_i915_private *i915) > > { > > /* > > * Display WA #0884: all > > @@ -871,7 +871,7 @@ static void psr_force_hw_tracking_exit(struct drm_i915_private *dev_priv) > > * but it makes more sense write to the current active > > * pipe. > > */ > > - I915_WRITE(CURSURFLIVE(dev_priv->psr.pipe), 0); > > + I915_WRITE(CURSURFLIVE(i915->psr.pipe), 0); > > } > > /** > > @@ -886,22 +886,22 @@ static void psr_force_hw_tracking_exit(struct drm_i915_private *dev_priv) > > void intel_psr_update(struct intel_dp *intel_dp, > > const struct intel_crtc_state *crtc_state) > > { > > - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp); > > - struct i915_psr *psr = &dev_priv->psr; > > + struct drm_i915_private *i915 = dp_to_i915(intel_dp); > > + struct i915_psr *psr = &i915->psr; > > bool enable, psr2_enable; > > - if (!CAN_PSR(dev_priv) || READ_ONCE(psr->dp) != intel_dp) > > + if (!CAN_PSR(i915) || READ_ONCE(psr->dp) != intel_dp) > > return; > > - mutex_lock(&dev_priv->psr.lock); > > + mutex_lock(&i915->psr.lock); > > enable = crtc_state->has_psr && psr_global_enabled(psr->debug); > > - psr2_enable = intel_psr2_enabled(dev_priv, crtc_state); > > + psr2_enable = intel_psr2_enabled(i915, crtc_state); > > if (enable == psr->enabled && psr2_enable == psr->psr2_enabled) { > > /* Force a PSR exit when enabling CRC to avoid CRC timeouts */ > > if (crtc_state->crc_enabled && psr->enabled) > > - psr_force_hw_tracking_exit(dev_priv); > > + psr_force_hw_tracking_exit(i915); > > goto unlock; > > } > > @@ -910,10 +910,10 @@ void intel_psr_update(struct intel_dp *intel_dp, > > intel_psr_disable_locked(intel_dp); > > if (enable) > > - intel_psr_enable_locked(dev_priv, crtc_state); > > + intel_psr_enable_locked(i915, crtc_state); > > unlock: > > - mutex_unlock(&dev_priv->psr.lock); > > + mutex_unlock(&i915->psr.lock); > > } > > /** > > @@ -930,13 +930,13 @@ int intel_psr_wait_for_idle(const struct intel_crtc_state *new_crtc_state, > > u32 *out_value) > > { > > struct intel_crtc *crtc = to_intel_crtc(new_crtc_state->base.crtc); > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > - if (!dev_priv->psr.enabled || !new_crtc_state->has_psr) > > + if (!i915->psr.enabled || !new_crtc_state->has_psr) > > return 0; > > /* FIXME: Update this for PSR2 if we need to wait for idle */ > > - if (READ_ONCE(dev_priv->psr.psr2_enabled)) > > + if (READ_ONCE(i915->psr.psr2_enabled)) > > return 0; > > /* > > @@ -946,22 +946,22 @@ int intel_psr_wait_for_idle(const struct intel_crtc_state *new_crtc_state, > > * defensive enough to cover everything. > > */ > > - return __intel_wait_for_register(&dev_priv->uncore, EDP_PSR_STATUS, > > + return __intel_wait_for_register(&i915->uncore, EDP_PSR_STATUS, > > EDP_PSR_STATUS_STATE_MASK, > > EDP_PSR_STATUS_STATE_IDLE, 2, 50, > > out_value); > > } > > -static bool __psr_wait_for_idle_locked(struct drm_i915_private *dev_priv) > > +static bool __psr_wait_for_idle_locked(struct drm_i915_private *i915) > > { > > i915_reg_t reg; > > u32 mask; > > int err; > > - if (!dev_priv->psr.enabled) > > + if (!i915->psr.enabled) > > return false; > > - if (dev_priv->psr.psr2_enabled) { > > + if (i915->psr.psr2_enabled) { > > reg = EDP_PSR2_STATUS; > > mask = EDP_PSR2_STATUS_STATE_MASK; > > } else { > > @@ -969,20 +969,20 @@ static bool __psr_wait_for_idle_locked(struct drm_i915_private *dev_priv) > > mask = EDP_PSR_STATUS_STATE_MASK; > > } > > - mutex_unlock(&dev_priv->psr.lock); > > + mutex_unlock(&i915->psr.lock); > > - err = intel_wait_for_register(&dev_priv->uncore, reg, mask, 0, 50); > > + err = intel_wait_for_register(&i915->uncore, reg, mask, 0, 50); > > if (err) > > DRM_ERROR("Timed out waiting for PSR Idle for re-enable\n"); > > /* After the unlocked wait, verify that PSR is still wanted! */ > > - mutex_lock(&dev_priv->psr.lock); > > - return err == 0 && dev_priv->psr.enabled; > > + mutex_lock(&i915->psr.lock); > > + return err == 0 && i915->psr.enabled; > > } > > -static int intel_psr_fastset_force(struct drm_i915_private *dev_priv) > > +static int intel_psr_fastset_force(struct drm_i915_private *i915) > > { > > - struct drm_device *dev = &dev_priv->drm; > > + struct drm_device *dev = &i915->drm; > > struct drm_modeset_acquire_ctx ctx; > > struct drm_atomic_state *state; > > struct drm_crtc *crtc; > > @@ -1032,7 +1032,7 @@ static int intel_psr_fastset_force(struct drm_i915_private *dev_priv) > > return err; > > } > > -int intel_psr_debug_set(struct drm_i915_private *dev_priv, u64 val) > > +int intel_psr_debug_set(struct drm_i915_private *i915, u64 val) > > { > > const u32 mode = val & I915_PSR_DEBUG_MODE_MASK; > > u32 old_mode; > > @@ -1044,25 +1044,25 @@ int intel_psr_debug_set(struct drm_i915_private *dev_priv, u64 val) > > return -EINVAL; > > } > > - ret = mutex_lock_interruptible(&dev_priv->psr.lock); > > + ret = mutex_lock_interruptible(&i915->psr.lock); > > if (ret) > > return ret; > > - old_mode = dev_priv->psr.debug & I915_PSR_DEBUG_MODE_MASK; > > - dev_priv->psr.debug = val; > > - intel_psr_irq_control(dev_priv, dev_priv->psr.debug); > > + old_mode = i915->psr.debug & I915_PSR_DEBUG_MODE_MASK; > > + i915->psr.debug = val; > > + intel_psr_irq_control(i915, i915->psr.debug); > > - mutex_unlock(&dev_priv->psr.lock); > > + mutex_unlock(&i915->psr.lock); > > if (old_mode != mode) > > - ret = intel_psr_fastset_force(dev_priv); > > + ret = intel_psr_fastset_force(i915); > > return ret; > > } > > -static void intel_psr_handle_irq(struct drm_i915_private *dev_priv) > > +static void intel_psr_handle_irq(struct drm_i915_private *i915) > > { > > - struct i915_psr *psr = &dev_priv->psr; > > + struct i915_psr *psr = &i915->psr; > > intel_psr_disable_locked(psr->dp); > > psr->sink_not_reliable = true; > > @@ -1072,16 +1072,16 @@ static void intel_psr_handle_irq(struct drm_i915_private *dev_priv) > > static void intel_psr_work(struct work_struct *work) > > { > > - struct drm_i915_private *dev_priv = > > - container_of(work, typeof(*dev_priv), psr.work); > > + struct drm_i915_private *i915 = > > + container_of(work, typeof(*i915), psr.work); > > - mutex_lock(&dev_priv->psr.lock); > > + mutex_lock(&i915->psr.lock); > > - if (!dev_priv->psr.enabled) > > + if (!i915->psr.enabled) > > goto unlock; > > - if (READ_ONCE(dev_priv->psr.irq_aux_error)) > > - intel_psr_handle_irq(dev_priv); > > + if (READ_ONCE(i915->psr.irq_aux_error)) > > + intel_psr_handle_irq(i915); > > /* > > * We have to make sure PSR is ready for re-enable > > @@ -1089,7 +1089,7 @@ static void intel_psr_work(struct work_struct *work) > > * PSR might take some time to get fully disabled > > * and be ready for re-enable. > > */ > > - if (!__psr_wait_for_idle_locked(dev_priv)) > > + if (!__psr_wait_for_idle_locked(i915)) > > goto unlock; > > /* > > @@ -1097,17 +1097,17 @@ static void intel_psr_work(struct work_struct *work) > > * recheck. Since psr_flush first clears this and then reschedules we > > * won't ever miss a flush when bailing out here. > > */ > > - if (dev_priv->psr.busy_frontbuffer_bits || dev_priv->psr.active) > > + if (i915->psr.busy_frontbuffer_bits || i915->psr.active) > > goto unlock; > > - intel_psr_activate(dev_priv->psr.dp); > > + intel_psr_activate(i915->psr.dp); > > unlock: > > - mutex_unlock(&dev_priv->psr.lock); > > + mutex_unlock(&i915->psr.lock); > > } > > /** > > * intel_psr_invalidate - Invalidade PSR > > - * @dev_priv: i915 device > > + * @i915: i915 device > > * @frontbuffer_bits: frontbuffer plane tracking bits > > * @origin: which operation caused the invalidate > > * > > @@ -1118,33 +1118,33 @@ static void intel_psr_work(struct work_struct *work) > > * > > * Dirty frontbuffers relevant to PSR are tracked in busy_frontbuffer_bits." > > */ > > -void intel_psr_invalidate(struct drm_i915_private *dev_priv, > > +void intel_psr_invalidate(struct drm_i915_private *i915, > > unsigned frontbuffer_bits, enum fb_op_origin origin) > > { > > - if (!CAN_PSR(dev_priv)) > > + if (!CAN_PSR(i915)) > > return; > > if (origin == ORIGIN_FLIP) > > return; > > - mutex_lock(&dev_priv->psr.lock); > > - if (!dev_priv->psr.enabled) { > > - mutex_unlock(&dev_priv->psr.lock); > > + mutex_lock(&i915->psr.lock); > > + if (!i915->psr.enabled) { > > + mutex_unlock(&i915->psr.lock); > > return; > > } > > - frontbuffer_bits &= INTEL_FRONTBUFFER_ALL_MASK(dev_priv->psr.pipe); > > - dev_priv->psr.busy_frontbuffer_bits |= frontbuffer_bits; > > + frontbuffer_bits &= INTEL_FRONTBUFFER_ALL_MASK(i915->psr.pipe); > > + i915->psr.busy_frontbuffer_bits |= frontbuffer_bits; > > if (frontbuffer_bits) > > - intel_psr_exit(dev_priv); > > + intel_psr_exit(i915); > > - mutex_unlock(&dev_priv->psr.lock); > > + mutex_unlock(&i915->psr.lock); > > } > > /** > > * intel_psr_flush - Flush PSR > > - * @dev_priv: i915 device > > + * @i915: i915 device > > * @frontbuffer_bits: frontbuffer plane tracking bits > > * @origin: which operation caused the flush > > * > > @@ -1155,55 +1155,55 @@ void intel_psr_invalidate(struct drm_i915_private *dev_priv, > > * > > * Dirty frontbuffers relevant to PSR are tracked in busy_frontbuffer_bits. > > */ > > -void intel_psr_flush(struct drm_i915_private *dev_priv, > > +void intel_psr_flush(struct drm_i915_private *i915, > > unsigned frontbuffer_bits, enum fb_op_origin origin) > > { > > - if (!CAN_PSR(dev_priv)) > > + if (!CAN_PSR(i915)) > > return; > > if (origin == ORIGIN_FLIP) > > return; > > - mutex_lock(&dev_priv->psr.lock); > > - if (!dev_priv->psr.enabled) { > > - mutex_unlock(&dev_priv->psr.lock); > > + mutex_lock(&i915->psr.lock); > > + if (!i915->psr.enabled) { > > + mutex_unlock(&i915->psr.lock); > > return; > > } > > - frontbuffer_bits &= INTEL_FRONTBUFFER_ALL_MASK(dev_priv->psr.pipe); > > - dev_priv->psr.busy_frontbuffer_bits &= ~frontbuffer_bits; > > + frontbuffer_bits &= INTEL_FRONTBUFFER_ALL_MASK(i915->psr.pipe); > > + i915->psr.busy_frontbuffer_bits &= ~frontbuffer_bits; > > /* By definition flush = invalidate + flush */ > > if (frontbuffer_bits) > > - psr_force_hw_tracking_exit(dev_priv); > > + psr_force_hw_tracking_exit(i915); > > - if (!dev_priv->psr.active && !dev_priv->psr.busy_frontbuffer_bits) > > - schedule_work(&dev_priv->psr.work); > > - mutex_unlock(&dev_priv->psr.lock); > > + if (!i915->psr.active && !i915->psr.busy_frontbuffer_bits) > > + schedule_work(&i915->psr.work); > > + mutex_unlock(&i915->psr.lock); > > } > > /** > > * intel_psr_init - Init basic PSR work and mutex. > > - * @dev_priv: i915 device private > > + * @i915: i915 device private > > * > > * This function is called only once at driver load to initialize basic > > * PSR stuff. > > */ > > -void intel_psr_init(struct drm_i915_private *dev_priv) > > +void intel_psr_init(struct drm_i915_private *i915) > > { > > u32 val; > > - if (!HAS_PSR(dev_priv)) > > + if (!HAS_PSR(i915)) > > return; > > - dev_priv->psr_mmio_base = IS_HASWELL(dev_priv) ? > > + i915->psr_mmio_base = IS_HASWELL(i915) ? > > HSW_EDP_PSR_BASE : BDW_EDP_PSR_BASE; > > - if (!dev_priv->psr.sink_support) > > + if (!i915->psr.sink_support) > > return; > > if (i915_modparams.enable_psr == -1) > > - if (INTEL_GEN(dev_priv) < 9 || !dev_priv->vbt.psr.enable) > > + if (INTEL_GEN(i915) < 9 || !i915->vbt.psr.enable) > > i915_modparams.enable_psr = 0; > > /* > > @@ -1218,31 +1218,31 @@ void intel_psr_init(struct drm_i915_private *dev_priv) > > val &= EDP_PSR_ERROR(edp_psr_shift(TRANSCODER_EDP)); > > if (val) { > > DRM_DEBUG_KMS("PSR interruption error set\n"); > > - dev_priv->psr.sink_not_reliable = true; > > + i915->psr.sink_not_reliable = true; > > } > > /* Set link_standby x link_off defaults */ > > - if (IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv)) > > + if (IS_HASWELL(i915) || IS_BROADWELL(i915)) > > /* HSW and BDW require workarounds that we don't implement. */ > > - dev_priv->psr.link_standby = false; > > + i915->psr.link_standby = false; > > else > > /* For new platforms let's respect VBT back again */ > > - dev_priv->psr.link_standby = dev_priv->vbt.psr.full_link; > > + i915->psr.link_standby = i915->vbt.psr.full_link; > > - INIT_WORK(&dev_priv->psr.work, intel_psr_work); > > - mutex_init(&dev_priv->psr.lock); > > + INIT_WORK(&i915->psr.work, intel_psr_work); > > + mutex_init(&i915->psr.lock); > > } > > void intel_psr_short_pulse(struct intel_dp *intel_dp) > > { > > - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp); > > - struct i915_psr *psr = &dev_priv->psr; > > + struct drm_i915_private *i915 = dp_to_i915(intel_dp); > > + struct i915_psr *psr = &i915->psr; > > u8 val; > > const u8 errors = DP_PSR_RFB_STORAGE_ERROR | > > DP_PSR_VSC_SDP_UNCORRECTABLE_ERROR | > > DP_PSR_LINK_CRC_ERROR; > > - if (!CAN_PSR(dev_priv) || !intel_dp_is_edp(intel_dp)) > > + if (!CAN_PSR(i915) || !intel_dp_is_edp(intel_dp)) > > return; > > mutex_lock(&psr->lock); > > @@ -1288,15 +1288,15 @@ void intel_psr_short_pulse(struct intel_dp *intel_dp) > > bool intel_psr_enabled(struct intel_dp *intel_dp) > > { > > - struct drm_i915_private *dev_priv = dp_to_i915(intel_dp); > > + struct drm_i915_private *i915 = dp_to_i915(intel_dp); > > bool ret; > > - if (!CAN_PSR(dev_priv) || !intel_dp_is_edp(intel_dp)) > > + if (!CAN_PSR(i915) || !intel_dp_is_edp(intel_dp)) > > return false; > > - mutex_lock(&dev_priv->psr.lock); > > - ret = (dev_priv->psr.dp == intel_dp && dev_priv->psr.enabled); > > - mutex_unlock(&dev_priv->psr.lock); > > + mutex_lock(&i915->psr.lock); > > + ret = (i915->psr.dp == intel_dp && i915->psr.enabled); > > + mutex_unlock(&i915->psr.lock); > > return ret; > > } > > diff --git a/drivers/gpu/drm/i915/intel_psr.h b/drivers/gpu/drm/i915/intel_psr.h > > index dc818826f36d..8a3ddc4ae44d 100644 > > --- a/drivers/gpu/drm/i915/intel_psr.h > > +++ b/drivers/gpu/drm/i915/intel_psr.h > > @@ -12,7 +12,7 @@ struct drm_i915_private; > > struct intel_crtc_state; > > struct intel_dp; > > -#define CAN_PSR(dev_priv) (HAS_PSR(dev_priv) && dev_priv->psr.sink_support) > > +#define CAN_PSR(i915) (HAS_PSR(i915) && i915->psr.sink_support) > > void intel_psr_init_dpcd(struct intel_dp *intel_dp); > > void intel_psr_enable(struct intel_dp *intel_dp, > > const struct intel_crtc_state *crtc_state); > > @@ -20,18 +20,18 @@ void intel_psr_disable(struct intel_dp *intel_dp, > > const struct intel_crtc_state *old_crtc_state); > > void intel_psr_update(struct intel_dp *intel_dp, > > const struct intel_crtc_state *crtc_state); > > -int intel_psr_debug_set(struct drm_i915_private *dev_priv, u64 value); > > -void intel_psr_invalidate(struct drm_i915_private *dev_priv, > > +int intel_psr_debug_set(struct drm_i915_private *i915, u64 value); > > +void intel_psr_invalidate(struct drm_i915_private *i915, > > unsigned frontbuffer_bits, > > enum fb_op_origin origin); > > -void intel_psr_flush(struct drm_i915_private *dev_priv, > > +void intel_psr_flush(struct drm_i915_private *i915, > > unsigned frontbuffer_bits, > > enum fb_op_origin origin); > > -void intel_psr_init(struct drm_i915_private *dev_priv); > > +void intel_psr_init(struct drm_i915_private *i915); > > void intel_psr_compute_config(struct intel_dp *intel_dp, > > struct intel_crtc_state *crtc_state); > > -void intel_psr_irq_control(struct drm_i915_private *dev_priv, u32 debug); > > -void intel_psr_irq_handler(struct drm_i915_private *dev_priv, u32 psr_iir); > > +void intel_psr_irq_control(struct drm_i915_private *i915, u32 debug); > > +void intel_psr_irq_handler(struct drm_i915_private *i915, u32 psr_iir); > > void intel_psr_short_pulse(struct intel_dp *intel_dp); > > int intel_psr_wait_for_idle(const struct intel_crtc_state *new_crtc_state, > > u32 *out_value); > > diff --git a/drivers/gpu/drm/i915/intel_quirks.h b/drivers/gpu/drm/i915/intel_quirks.h > > index b0fcff142a56..b617ca047695 100644 > > --- a/drivers/gpu/drm/i915/intel_quirks.h > > +++ b/drivers/gpu/drm/i915/intel_quirks.h > > @@ -8,6 +8,6 @@ > > struct drm_i915_private; > > -void intel_init_quirks(struct drm_i915_private *dev_priv); > > +void intel_init_quirks(struct drm_i915_private *i915); > > #endif /* __INTEL_QUIRKS_H__ */ > > diff --git a/drivers/gpu/drm/i915/intel_runtime_pm.h b/drivers/gpu/drm/i915/intel_runtime_pm.h > > index a7acceb13473..1042f1c0b125 100644 > > --- a/drivers/gpu/drm/i915/intel_runtime_pm.h > > +++ b/drivers/gpu/drm/i915/intel_runtime_pm.h > > @@ -20,10 +20,10 @@ enum i915_drm_suspend_mode { > > I915_DRM_SUSPEND_HIBERNATE, > > }; > > -void intel_runtime_pm_init_early(struct drm_i915_private *dev_priv); > > -void intel_runtime_pm_enable(struct drm_i915_private *dev_priv); > > -void intel_runtime_pm_disable(struct drm_i915_private *dev_priv); > > -void intel_runtime_pm_cleanup(struct drm_i915_private *dev_priv); > > +void intel_runtime_pm_init_early(struct drm_i915_private *i915); > > +void intel_runtime_pm_enable(struct drm_i915_private *i915); > > +void intel_runtime_pm_disable(struct drm_i915_private *i915); > > +void intel_runtime_pm_cleanup(struct drm_i915_private *i915); > > intel_wakeref_t intel_runtime_pm_get(struct drm_i915_private *i915); > > intel_wakeref_t intel_runtime_pm_get_if_in_use(struct drm_i915_private *i915); > > diff --git a/drivers/gpu/drm/i915/intel_sdvo.c b/drivers/gpu/drm/i915/intel_sdvo.c > > index 0860ae36bb87..b916c79d6b31 100644 > > --- a/drivers/gpu/drm/i915/intel_sdvo.c > > +++ b/drivers/gpu/drm/i915/intel_sdvo.c > > @@ -212,18 +212,18 @@ intel_sdvo_create_enhance_property(struct intel_sdvo *intel_sdvo, > > static void intel_sdvo_write_sdvox(struct intel_sdvo *intel_sdvo, u32 val) > > { > > struct drm_device *dev = intel_sdvo->base.base.dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > u32 bval = val, cval = val; > > int i; > > - if (HAS_PCH_SPLIT(dev_priv)) { > > + if (HAS_PCH_SPLIT(i915)) { > > I915_WRITE(intel_sdvo->sdvo_reg, val); > > POSTING_READ(intel_sdvo->sdvo_reg); > > /* > > * HW workaround, need to write this twice for issue > > * that may result in first write getting masked. > > */ > > - if (HAS_PCH_IBX(dev_priv)) { > > + if (HAS_PCH_IBX(i915)) { > > I915_WRITE(intel_sdvo->sdvo_reg, val); > > POSTING_READ(intel_sdvo->sdvo_reg); > > } > > @@ -1400,7 +1400,7 @@ static void intel_sdvo_pre_enable(struct intel_encoder *intel_encoder, > > const struct intel_crtc_state *crtc_state, > > const struct drm_connector_state *conn_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(intel_encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(intel_encoder->base.dev); > > struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc); > > const struct drm_display_mode *adjusted_mode = &crtc_state->base.adjusted_mode; > > const struct intel_sdvo_connector_state *sdvo_state = > > @@ -1481,13 +1481,13 @@ static void intel_sdvo_pre_enable(struct intel_encoder *intel_encoder, > > return; > > /* Set the SDVO control regs. */ > > - if (INTEL_GEN(dev_priv) >= 4) { > > + if (INTEL_GEN(i915) >= 4) { > > /* The real mode polarity is set by the SDVO commands, using > > * struct intel_sdvo_dtd. */ > > sdvox = SDVO_VSYNC_ACTIVE_HIGH | SDVO_HSYNC_ACTIVE_HIGH; > > - if (!HAS_PCH_SPLIT(dev_priv) && crtc_state->limited_color_range) > > + if (!HAS_PCH_SPLIT(i915) && crtc_state->limited_color_range) > > sdvox |= HDMI_COLOR_RANGE_16_235; > > - if (INTEL_GEN(dev_priv) < 5) > > + if (INTEL_GEN(i915) < 5) > > sdvox |= SDVO_BORDER_ENABLE; > > } else { > > sdvox = I915_READ(intel_sdvo->sdvo_reg); > > @@ -1498,15 +1498,15 @@ static void intel_sdvo_pre_enable(struct intel_encoder *intel_encoder, > > sdvox |= (9 << 19) | SDVO_BORDER_ENABLE; > > } > > - if (HAS_PCH_CPT(dev_priv)) > > + if (HAS_PCH_CPT(i915)) > > sdvox |= SDVO_PIPE_SEL_CPT(crtc->pipe); > > else > > sdvox |= SDVO_PIPE_SEL(crtc->pipe); > > - if (INTEL_GEN(dev_priv) >= 4) { > > + if (INTEL_GEN(i915) >= 4) { > > /* done in crtc_mode_set as the dpll_md reg must be written early */ > > - } else if (IS_I945G(dev_priv) || IS_I945GM(dev_priv) || > > - IS_G33(dev_priv) || IS_PINEVIEW(dev_priv)) { > > + } else if (IS_I945G(i915) || IS_I945GM(i915) || > > + IS_G33(i915) || IS_PINEVIEW(i915)) { > > /* done in crtc_mode_set as it lives inside the dpll register */ > > } else { > > sdvox |= (crtc_state->pixel_multiplier - 1) > > @@ -1514,7 +1514,7 @@ static void intel_sdvo_pre_enable(struct intel_encoder *intel_encoder, > > } > > if (input_dtd.part2.sdvo_flags & SDVO_NEED_TO_STALL && > > - INTEL_GEN(dev_priv) < 5) > > + INTEL_GEN(i915) < 5) > > sdvox |= SDVO_STALL_SELECT; > > intel_sdvo_write_sdvox(intel_sdvo, sdvox); > > } > > @@ -1531,7 +1531,7 @@ static bool intel_sdvo_connector_get_hw_state(struct intel_connector *connector) > > return active_outputs & intel_sdvo_connector->output_flag; > > } > > -bool intel_sdvo_port_enabled(struct drm_i915_private *dev_priv, > > +bool intel_sdvo_port_enabled(struct drm_i915_private *i915, > > i915_reg_t sdvo_reg, enum pipe *pipe) > > { > > u32 val; > > @@ -1539,9 +1539,9 @@ bool intel_sdvo_port_enabled(struct drm_i915_private *dev_priv, > > val = I915_READ(sdvo_reg); > > /* asserts want to know the pipe even if the port is disabled */ > > - if (HAS_PCH_CPT(dev_priv)) > > + if (HAS_PCH_CPT(i915)) > > *pipe = (val & SDVO_PIPE_SEL_MASK_CPT) >> SDVO_PIPE_SEL_SHIFT_CPT; > > - else if (IS_CHERRYVIEW(dev_priv)) > > + else if (IS_CHERRYVIEW(i915)) > > *pipe = (val & SDVO_PIPE_SEL_MASK_CHV) >> SDVO_PIPE_SEL_SHIFT_CHV; > > else > > *pipe = (val & SDVO_PIPE_SEL_MASK) >> SDVO_PIPE_SEL_SHIFT; > > @@ -1552,14 +1552,14 @@ bool intel_sdvo_port_enabled(struct drm_i915_private *dev_priv, > > static bool intel_sdvo_get_hw_state(struct intel_encoder *encoder, > > enum pipe *pipe) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_sdvo *intel_sdvo = to_sdvo(encoder); > > u16 active_outputs = 0; > > bool ret; > > intel_sdvo_get_active_outputs(intel_sdvo, &active_outputs); > > - ret = intel_sdvo_port_enabled(dev_priv, intel_sdvo->sdvo_reg, pipe); > > + ret = intel_sdvo_port_enabled(i915, intel_sdvo->sdvo_reg, pipe); > > return ret || active_outputs; > > } > > @@ -1568,7 +1568,7 @@ static void intel_sdvo_get_config(struct intel_encoder *encoder, > > struct intel_crtc_state *pipe_config) > > { > > struct drm_device *dev = encoder->base.dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > struct intel_sdvo *intel_sdvo = to_sdvo(encoder); > > struct intel_sdvo_dtd dtd; > > int encoder_pixel_multiplier = 0; > > @@ -1610,7 +1610,7 @@ static void intel_sdvo_get_config(struct intel_encoder *encoder, > > * encoder->get_config we so already have a valid pixel multplier on all > > * other platfroms. > > */ > > - if (IS_I915G(dev_priv) || IS_I915GM(dev_priv)) { > > + if (IS_I915G(i915) || IS_I915GM(i915)) { > > pipe_config->pixel_multiplier = > > ((sdvox & SDVO_PORT_MULTIPLY_MASK) > > >> SDVO_PORT_MULTIPLY_SHIFT) + 1; > > @@ -1693,7 +1693,7 @@ static void intel_disable_sdvo(struct intel_encoder *encoder, > > const struct intel_crtc_state *old_crtc_state, > > const struct drm_connector_state *conn_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_sdvo *intel_sdvo = to_sdvo(encoder); > > struct intel_crtc *crtc = to_intel_crtc(old_crtc_state->base.crtc); > > u32 temp; > > @@ -1716,13 +1716,13 @@ static void intel_disable_sdvo(struct intel_encoder *encoder, > > * to transcoder A after disabling it to allow the > > * matching DP port to be enabled on transcoder A. > > */ > > - if (HAS_PCH_IBX(dev_priv) && crtc->pipe == PIPE_B) { > > + if (HAS_PCH_IBX(i915) && crtc->pipe == PIPE_B) { > > /* > > * We get CPU/PCH FIFO underruns on the other pipe when > > * doing the workaround. Sweep them under the rug. > > */ > > - intel_set_cpu_fifo_underrun_reporting(dev_priv, PIPE_A, false); > > - intel_set_pch_fifo_underrun_reporting(dev_priv, PIPE_A, false); > > + intel_set_cpu_fifo_underrun_reporting(i915, PIPE_A, false); > > + intel_set_pch_fifo_underrun_reporting(i915, PIPE_A, false); > > temp &= ~SDVO_PIPE_SEL_MASK; > > temp |= SDVO_ENABLE | SDVO_PIPE_SEL(PIPE_A); > > @@ -1731,9 +1731,9 @@ static void intel_disable_sdvo(struct intel_encoder *encoder, > > temp &= ~SDVO_ENABLE; > > intel_sdvo_write_sdvox(intel_sdvo, temp); > > - intel_wait_for_vblank_if_active(dev_priv, PIPE_A); > > - intel_set_cpu_fifo_underrun_reporting(dev_priv, PIPE_A, true); > > - intel_set_pch_fifo_underrun_reporting(dev_priv, PIPE_A, true); > > + intel_wait_for_vblank_if_active(i915, PIPE_A); > > + intel_set_cpu_fifo_underrun_reporting(i915, PIPE_A, true); > > + intel_set_pch_fifo_underrun_reporting(i915, PIPE_A, true); > > } > > } > > @@ -1755,7 +1755,7 @@ static void intel_enable_sdvo(struct intel_encoder *encoder, > > const struct drm_connector_state *conn_state) > > { > > struct drm_device *dev = encoder->base.dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > struct intel_sdvo *intel_sdvo = to_sdvo(encoder); > > struct intel_crtc *intel_crtc = to_intel_crtc(pipe_config->base.crtc); > > u32 temp; > > @@ -1768,7 +1768,7 @@ static void intel_enable_sdvo(struct intel_encoder *encoder, > > intel_sdvo_write_sdvox(intel_sdvo, temp); > > for (i = 0; i < 2; i++) > > - intel_wait_for_vblank(dev_priv, intel_crtc->pipe); > > + intel_wait_for_vblank(i915, intel_crtc->pipe); > > success = intel_sdvo_get_trained_inputs(intel_sdvo, &input1, &input2); > > /* > > @@ -1865,17 +1865,17 @@ static bool intel_sdvo_get_capabilities(struct intel_sdvo *intel_sdvo, struct in > > static u16 intel_sdvo_get_hotplug_support(struct intel_sdvo *intel_sdvo) > > { > > - struct drm_i915_private *dev_priv = to_i915(intel_sdvo->base.base.dev); > > + struct drm_i915_private *i915 = to_i915(intel_sdvo->base.base.dev); > > u16 hotplug; > > - if (!I915_HAS_HOTPLUG(dev_priv)) > > + if (!I915_HAS_HOTPLUG(i915)) > > return 0; > > /* > > * HW Erratum: SDVO Hotplug is broken on all i945G chips, there's noise > > * on the line. > > */ > > - if (IS_I945G(dev_priv) || IS_I945GM(dev_priv)) > > + if (IS_I945G(i915) || IS_I945GM(i915)) > > return 0; > > if (!intel_sdvo_get_value(intel_sdvo, SDVO_CMD_GET_HOT_PLUG_SUPPORT, > > @@ -1919,11 +1919,11 @@ intel_sdvo_get_edid(struct drm_connector *connector) > > static struct edid * > > intel_sdvo_get_analog_edid(struct drm_connector *connector) > > { > > - struct drm_i915_private *dev_priv = to_i915(connector->dev); > > + struct drm_i915_private *i915 = to_i915(connector->dev); > > return drm_get_edid(connector, > > - intel_gmbus_get_adapter(dev_priv, > > - dev_priv->vbt.crt_ddc_pin)); > > + intel_gmbus_get_adapter(i915, > > + i915->vbt.crt_ddc_pin)); > > } > > static enum drm_connector_status > > @@ -2186,7 +2186,7 @@ static void intel_sdvo_get_tv_modes(struct drm_connector *connector) > > static void intel_sdvo_get_lvds_modes(struct drm_connector *connector) > > { > > struct intel_sdvo *intel_sdvo = intel_attached_sdvo(connector); > > - struct drm_i915_private *dev_priv = to_i915(connector->dev); > > + struct drm_i915_private *i915 = to_i915(connector->dev); > > struct drm_display_mode *newmode; > > DRM_DEBUG_KMS("[CONNECTOR:%d:%s]\n", > > @@ -2196,9 +2196,9 @@ static void intel_sdvo_get_lvds_modes(struct drm_connector *connector) > > * Fetch modes from VBT. For SDVO prefer the VBT mode since some > > * SDVO->LVDS transcoders can't cope with the EDID mode. > > */ > > - if (dev_priv->vbt.sdvo_lvds_vbt_mode != NULL) { > > + if (i915->vbt.sdvo_lvds_vbt_mode != NULL) { > > newmode = drm_mode_duplicate(connector->dev, > > - dev_priv->vbt.sdvo_lvds_vbt_mode); > > + i915->vbt.sdvo_lvds_vbt_mode); > > if (newmode != NULL) { > > /* Guarantee the mode is preferred */ > > newmode->type = (DRM_MODE_TYPE_PREFERRED | > > @@ -2488,15 +2488,15 @@ intel_sdvo_guess_ddc_bus(struct intel_sdvo *sdvo) > > * outputs, then LVDS outputs. > > */ > > static void > > -intel_sdvo_select_ddc_bus(struct drm_i915_private *dev_priv, > > +intel_sdvo_select_ddc_bus(struct drm_i915_private *i915, > > struct intel_sdvo *sdvo) > > { > > struct sdvo_device_mapping *mapping; > > if (sdvo->port == PORT_B) > > - mapping = &dev_priv->vbt.sdvo_mappings[0]; > > + mapping = &i915->vbt.sdvo_mappings[0]; > > else > > - mapping = &dev_priv->vbt.sdvo_mappings[1]; > > + mapping = &i915->vbt.sdvo_mappings[1]; > > if (mapping->initialized) > > sdvo->ddc_bus = 1 << ((mapping->ddc_pin & 0xf0) >> 4); > > @@ -2505,24 +2505,24 @@ intel_sdvo_select_ddc_bus(struct drm_i915_private *dev_priv, > > } > > static void > > -intel_sdvo_select_i2c_bus(struct drm_i915_private *dev_priv, > > +intel_sdvo_select_i2c_bus(struct drm_i915_private *i915, > > struct intel_sdvo *sdvo) > > { > > struct sdvo_device_mapping *mapping; > > u8 pin; > > if (sdvo->port == PORT_B) > > - mapping = &dev_priv->vbt.sdvo_mappings[0]; > > + mapping = &i915->vbt.sdvo_mappings[0]; > > else > > - mapping = &dev_priv->vbt.sdvo_mappings[1]; > > + mapping = &i915->vbt.sdvo_mappings[1]; > > if (mapping->initialized && > > - intel_gmbus_is_valid_pin(dev_priv, mapping->i2c_pin)) > > + intel_gmbus_is_valid_pin(i915, mapping->i2c_pin)) > > pin = mapping->i2c_pin; > > else > > pin = GMBUS_PIN_DPB; > > - sdvo->i2c = intel_gmbus_get_adapter(dev_priv, pin); > > + sdvo->i2c = intel_gmbus_get_adapter(i915, pin); > > /* > > * With gmbus we should be able to drive sdvo i2c at 2MHz, but somehow > > @@ -2546,17 +2546,17 @@ intel_sdvo_is_hdmi_connector(struct intel_sdvo *intel_sdvo, int device) > > } > > static u8 > > -intel_sdvo_get_slave_addr(struct drm_i915_private *dev_priv, > > +intel_sdvo_get_slave_addr(struct drm_i915_private *i915, > > struct intel_sdvo *sdvo) > > { > > struct sdvo_device_mapping *my_mapping, *other_mapping; > > if (sdvo->port == PORT_B) { > > - my_mapping = &dev_priv->vbt.sdvo_mappings[0]; > > - other_mapping = &dev_priv->vbt.sdvo_mappings[1]; > > + my_mapping = &i915->vbt.sdvo_mappings[0]; > > + other_mapping = &i915->vbt.sdvo_mappings[1]; > > } else { > > - my_mapping = &dev_priv->vbt.sdvo_mappings[1]; > > - other_mapping = &dev_priv->vbt.sdvo_mappings[0]; > > + my_mapping = &i915->vbt.sdvo_mappings[1]; > > + other_mapping = &i915->vbt.sdvo_mappings[0]; > > } > > /* If the BIOS described our SDVO device, take advantage of it. */ > > @@ -2616,10 +2616,10 @@ static void > > intel_sdvo_add_hdmi_properties(struct intel_sdvo *intel_sdvo, > > struct intel_sdvo_connector *connector) > > { > > - struct drm_i915_private *dev_priv = to_i915(connector->base.base.dev); > > + struct drm_i915_private *i915 = to_i915(connector->base.base.dev); > > intel_attach_force_audio_property(&connector->base.base); > > - if (INTEL_GEN(dev_priv) >= 4 && IS_MOBILE(dev_priv)) { > > + if (INTEL_GEN(i915) >= 4 && IS_MOBILE(i915)) { > > intel_attach_broadcast_rgb_property(&connector->base.base); > > } > > intel_attach_aspect_ratio_property(&connector->base.base); > > @@ -3180,9 +3180,9 @@ static const struct i2c_lock_operations proxy_lock_ops = { > > static bool > > intel_sdvo_init_ddc_proxy(struct intel_sdvo *sdvo, > > - struct drm_i915_private *dev_priv) > > + struct drm_i915_private *i915) > > { > > - struct pci_dev *pdev = dev_priv->drm.pdev; > > + struct pci_dev *pdev = i915->drm.pdev; > > sdvo->ddc.owner = THIS_MODULE; > > sdvo->ddc.class = I2C_CLASS_DDC; > > @@ -3195,23 +3195,23 @@ intel_sdvo_init_ddc_proxy(struct intel_sdvo *sdvo, > > return i2c_add_adapter(&sdvo->ddc) == 0; > > } > > -static void assert_sdvo_port_valid(const struct drm_i915_private *dev_priv, > > +static void assert_sdvo_port_valid(const struct drm_i915_private *i915, > > enum port port) > > { > > - if (HAS_PCH_SPLIT(dev_priv)) > > + if (HAS_PCH_SPLIT(i915)) > > WARN_ON(port != PORT_B); > > else > > WARN_ON(port != PORT_B && port != PORT_C); > > } > > -bool intel_sdvo_init(struct drm_i915_private *dev_priv, > > +bool intel_sdvo_init(struct drm_i915_private *i915, > > i915_reg_t sdvo_reg, enum port port) > > { > > struct intel_encoder *intel_encoder; > > struct intel_sdvo *intel_sdvo; > > int i; > > - assert_sdvo_port_valid(dev_priv, port); > > + assert_sdvo_port_valid(i915, port); > > intel_sdvo = kzalloc(sizeof(*intel_sdvo), GFP_KERNEL); > > if (!intel_sdvo) > > @@ -3220,9 +3220,9 @@ bool intel_sdvo_init(struct drm_i915_private *dev_priv, > > intel_sdvo->sdvo_reg = sdvo_reg; > > intel_sdvo->port = port; > > intel_sdvo->slave_addr = > > - intel_sdvo_get_slave_addr(dev_priv, intel_sdvo) >> 1; > > - intel_sdvo_select_i2c_bus(dev_priv, intel_sdvo); > > - if (!intel_sdvo_init_ddc_proxy(intel_sdvo, dev_priv)) > > + intel_sdvo_get_slave_addr(i915, intel_sdvo) >> 1; > > + intel_sdvo_select_i2c_bus(i915, intel_sdvo); > > + if (!intel_sdvo_init_ddc_proxy(intel_sdvo, i915)) > > goto err_i2c_bus; > > /* encoder type will be decided later */ > > @@ -3230,7 +3230,7 @@ bool intel_sdvo_init(struct drm_i915_private *dev_priv, > > intel_encoder->type = INTEL_OUTPUT_SDVO; > > intel_encoder->power_domain = POWER_DOMAIN_PORT_OTHER; > > intel_encoder->port = port; > > - drm_encoder_init(&dev_priv->drm, &intel_encoder->base, > > + drm_encoder_init(&i915->drm, &intel_encoder->base, > > &intel_sdvo_enc_funcs, 0, > > "SDVO %c", port_name(port)); > > @@ -3246,7 +3246,7 @@ bool intel_sdvo_init(struct drm_i915_private *dev_priv, > > } > > intel_encoder->compute_config = intel_sdvo_compute_config; > > - if (HAS_PCH_SPLIT(dev_priv)) { > > + if (HAS_PCH_SPLIT(i915)) { > > intel_encoder->disable = pch_disable_sdvo; > > intel_encoder->post_disable = pch_post_disable_sdvo; > > } else { > > @@ -3290,7 +3290,7 @@ bool intel_sdvo_init(struct drm_i915_private *dev_priv, > > */ > > intel_sdvo->base.cloneable = 0; > > - intel_sdvo_select_ddc_bus(dev_priv, intel_sdvo); > > + intel_sdvo_select_ddc_bus(i915, intel_sdvo); > > /* Set the input timing to the screen. Assume always input 0. */ > > if (!intel_sdvo_set_target_input(intel_sdvo)) > > diff --git a/drivers/gpu/drm/i915/intel_sdvo.h b/drivers/gpu/drm/i915/intel_sdvo.h > > index c9e05bcdd141..1ec6766c5ae1 100644 > > --- a/drivers/gpu/drm/i915/intel_sdvo.h > > +++ b/drivers/gpu/drm/i915/intel_sdvo.h > > @@ -15,9 +15,9 @@ > > struct drm_i915_private; > > enum pipe; > > -bool intel_sdvo_port_enabled(struct drm_i915_private *dev_priv, > > +bool intel_sdvo_port_enabled(struct drm_i915_private *i915, > > i915_reg_t sdvo_reg, enum pipe *pipe); > > -bool intel_sdvo_init(struct drm_i915_private *dev_priv, > > +bool intel_sdvo_init(struct drm_i915_private *i915, > > i915_reg_t reg, enum port port); > > #endif /* __INTEL_SDVO_H__ */ > > diff --git a/drivers/gpu/drm/i915/intel_sprite.c b/drivers/gpu/drm/i915/intel_sprite.c > > index c180815faabd..73871e5e8e19 100644 > > --- a/drivers/gpu/drm/i915/intel_sprite.c > > +++ b/drivers/gpu/drm/i915/intel_sprite.c > > @@ -94,12 +94,12 @@ int intel_usecs_to_scanlines(const struct drm_display_mode *adjusted_mode, > > void intel_pipe_update_start(const struct intel_crtc_state *new_crtc_state) > > { > > struct intel_crtc *crtc = to_intel_crtc(new_crtc_state->base.crtc); > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > const struct drm_display_mode *adjusted_mode = &new_crtc_state->base.adjusted_mode; > > long timeout = msecs_to_jiffies_timeout(1); > > int scanline, min, max, vblank_start; > > wait_queue_head_t *wq = drm_crtc_vblank_waitqueue(&crtc->base); > > - bool need_vlv_dsi_wa = (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) && > > + bool need_vlv_dsi_wa = (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) && > > intel_crtc_has_type(new_crtc_state, INTEL_OUTPUT_DSI); > > DEFINE_WAIT(wait); > > u32 psr_status; > > @@ -207,7 +207,7 @@ void intel_pipe_update_end(struct intel_crtc_state *new_crtc_state) > > int scanline_end = intel_get_crtc_scanline(crtc); > > u32 end_vbl_count = intel_crtc_get_vblank_counter(crtc); > > ktime_t end_vbl_time = ktime_get(); > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > trace_i915_pipe_update_end(crtc, end_vbl_count, scanline_end); > > @@ -227,7 +227,7 @@ void intel_pipe_update_end(struct intel_crtc_state *new_crtc_state) > > local_irq_enable(); > > - if (intel_vgpu_active(dev_priv)) > > + if (intel_vgpu_active(i915)) > > return; > > if (crtc->debug.start_vbl_count && > > @@ -353,7 +353,7 @@ skl_program_scaler(struct intel_plane *plane, > > const struct intel_crtc_state *crtc_state, > > const struct intel_plane_state *plane_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(plane->base.dev); > > + struct drm_i915_private *i915 = to_i915(plane->base.dev); > > enum pipe pipe = plane->pipe; > > int scaler_id = plane_state->scaler_id; > > const struct intel_scaler *scaler = > > @@ -375,7 +375,7 @@ skl_program_scaler(struct intel_plane *plane, > > /* TODO: handle sub-pixel coordinates */ > > if (is_planar_yuv_format(plane_state->base.fb->format->format) && > > - !icl_is_hdr_plane(dev_priv, plane->id)) { > > + !icl_is_hdr_plane(i915, plane->id)) { > > y_hphase = skl_scaler_calc_phase(1, hscale, false); > > y_vphase = skl_scaler_calc_phase(1, vscale, false); > > @@ -415,7 +415,7 @@ icl_program_input_csc(struct intel_plane *plane, > > const struct intel_crtc_state *crtc_state, > > const struct intel_plane_state *plane_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(plane->base.dev); > > + struct drm_i915_private *i915 = to_i915(plane->base.dev); > > enum pipe pipe = plane->pipe; > > enum plane_id plane_id = plane->id; > > @@ -507,7 +507,7 @@ skl_program_plane(struct intel_plane *plane, > > const struct intel_plane_state *plane_state, > > int color_plane, bool slave, u32 plane_ctl) > > { > > - struct drm_i915_private *dev_priv = to_i915(plane->base.dev); > > + struct drm_i915_private *i915 = to_i915(plane->base.dev); > > enum plane_id plane_id = plane->id; > > enum pipe pipe = plane->pipe; > > const struct drm_intel_sprite_colorkey *key = &plane_state->ckey; > > @@ -529,7 +529,7 @@ skl_program_plane(struct intel_plane *plane, > > plane_ctl |= skl_plane_ctl_crtc(crtc_state); > > - if (INTEL_GEN(dev_priv) >= 10 || IS_GEMINILAKE(dev_priv)) > > + if (INTEL_GEN(i915) >= 10 || IS_GEMINILAKE(i915)) > > plane_color_ctl = plane_state->color_ctl | > > glk_plane_color_ctl_crtc(crtc_state); > > @@ -549,7 +549,7 @@ skl_program_plane(struct intel_plane *plane, > > crtc_y = 0; > > } > > - spin_lock_irqsave(&dev_priv->uncore.lock, irqflags); > > + spin_lock_irqsave(&i915->uncore.lock, irqflags); > > I915_WRITE_FW(PLANE_STRIDE(pipe, plane_id), stride); > > I915_WRITE_FW(PLANE_POS(pipe, plane_id), (crtc_y << 16) | crtc_x); > > @@ -557,7 +557,7 @@ skl_program_plane(struct intel_plane *plane, > > I915_WRITE_FW(PLANE_AUX_DIST(pipe, plane_id), > > (plane_state->color_plane[1].offset - surf_addr) | aux_stride); > > - if (icl_is_hdr_plane(dev_priv, plane_id)) { > > + if (icl_is_hdr_plane(i915, plane_id)) { > > u32 cus_ctl = 0; > > if (linked) { > > @@ -578,10 +578,10 @@ skl_program_plane(struct intel_plane *plane, > > I915_WRITE_FW(PLANE_CUS_CTL(pipe, plane_id), cus_ctl); > > } > > - if (INTEL_GEN(dev_priv) >= 10 || IS_GEMINILAKE(dev_priv)) > > + if (INTEL_GEN(i915) >= 10 || IS_GEMINILAKE(i915)) > > I915_WRITE_FW(PLANE_COLOR_CTL(pipe, plane_id), plane_color_ctl); > > - if (fb->format->is_yuv && icl_is_hdr_plane(dev_priv, plane_id)) > > + if (fb->format->is_yuv && icl_is_hdr_plane(i915, plane_id)) > > icl_program_input_csc(plane, crtc_state, plane_state); > > skl_write_plane_wm(plane, crtc_state); > > @@ -592,7 +592,7 @@ skl_program_plane(struct intel_plane *plane, > > I915_WRITE_FW(PLANE_OFFSET(pipe, plane_id), (y << 16) | x); > > - if (INTEL_GEN(dev_priv) < 11) > > + if (INTEL_GEN(i915) < 11) > > I915_WRITE_FW(PLANE_AUX_OFFSET(pipe, plane_id), > > (plane_state->color_plane[1].y << 16) | > > plane_state->color_plane[1].x); > > @@ -609,7 +609,7 @@ skl_program_plane(struct intel_plane *plane, > > if (!slave && plane_state->scaler_id >= 0) > > skl_program_scaler(plane, crtc_state, plane_state); > > - spin_unlock_irqrestore(&dev_priv->uncore.lock, irqflags); > > + spin_unlock_irqrestore(&i915->uncore.lock, irqflags); > > } > > static void > > @@ -641,14 +641,14 @@ static void > > skl_disable_plane(struct intel_plane *plane, > > const struct intel_crtc_state *crtc_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(plane->base.dev); > > + struct drm_i915_private *i915 = to_i915(plane->base.dev); > > enum plane_id plane_id = plane->id; > > enum pipe pipe = plane->pipe; > > unsigned long irqflags; > > - spin_lock_irqsave(&dev_priv->uncore.lock, irqflags); > > + spin_lock_irqsave(&i915->uncore.lock, irqflags); > > - if (icl_is_hdr_plane(dev_priv, plane_id)) > > + if (icl_is_hdr_plane(i915, plane_id)) > > I915_WRITE_FW(PLANE_CUS_CTL(pipe, plane_id), 0); > > skl_write_plane_wm(plane, crtc_state); > > @@ -656,21 +656,21 @@ skl_disable_plane(struct intel_plane *plane, > > I915_WRITE_FW(PLANE_CTL(pipe, plane_id), 0); > > I915_WRITE_FW(PLANE_SURF(pipe, plane_id), 0); > > - spin_unlock_irqrestore(&dev_priv->uncore.lock, irqflags); > > + spin_unlock_irqrestore(&i915->uncore.lock, irqflags); > > } > > static bool > > skl_plane_get_hw_state(struct intel_plane *plane, > > enum pipe *pipe) > > { > > - struct drm_i915_private *dev_priv = to_i915(plane->base.dev); > > + struct drm_i915_private *i915 = to_i915(plane->base.dev); > > enum intel_display_power_domain power_domain; > > enum plane_id plane_id = plane->id; > > intel_wakeref_t wakeref; > > bool ret; > > power_domain = POWER_DOMAIN_PIPE(plane->pipe); > > - wakeref = intel_display_power_get_if_enabled(dev_priv, power_domain); > > + wakeref = intel_display_power_get_if_enabled(i915, power_domain); > > if (!wakeref) > > return false; > > @@ -678,7 +678,7 @@ skl_plane_get_hw_state(struct intel_plane *plane, > > *pipe = plane->pipe; > > - intel_display_power_put(dev_priv, power_domain, wakeref); > > + intel_display_power_put(i915, power_domain, wakeref); > > return ret; > > } > > @@ -687,7 +687,7 @@ static void > > chv_update_csc(const struct intel_plane_state *plane_state) > > { > > struct intel_plane *plane = to_intel_plane(plane_state->base.plane); > > - struct drm_i915_private *dev_priv = to_i915(plane->base.dev); > > + struct drm_i915_private *i915 = to_i915(plane->base.dev); > > const struct drm_framebuffer *fb = plane_state->base.fb; > > enum plane_id plane_id = plane->id; > > /* > > @@ -746,7 +746,7 @@ static void > > vlv_update_clrc(const struct intel_plane_state *plane_state) > > { > > struct intel_plane *plane = to_intel_plane(plane_state->base.plane); > > - struct drm_i915_private *dev_priv = to_i915(plane->base.dev); > > + struct drm_i915_private *i915 = to_i915(plane->base.dev); > > const struct drm_framebuffer *fb = plane_state->base.fb; > > enum pipe pipe = plane->pipe; > > enum plane_id plane_id = plane->id; > > @@ -863,7 +863,7 @@ vlv_update_plane(struct intel_plane *plane, > > const struct intel_crtc_state *crtc_state, > > const struct intel_plane_state *plane_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(plane->base.dev); > > + struct drm_i915_private *i915 = to_i915(plane->base.dev); > > enum pipe pipe = plane->pipe; > > enum plane_id plane_id = plane->id; > > u32 sprsurf_offset = plane_state->color_plane[0].offset; > > @@ -886,7 +886,7 @@ vlv_update_plane(struct intel_plane *plane, > > linear_offset = intel_fb_xy_to_linear(x, y, plane_state, 0); > > - spin_lock_irqsave(&dev_priv->uncore.lock, irqflags); > > + spin_lock_irqsave(&i915->uncore.lock, irqflags); > > I915_WRITE_FW(SPSTRIDE(pipe, plane_id), > > plane_state->color_plane[0].stride); > > @@ -894,7 +894,7 @@ vlv_update_plane(struct intel_plane *plane, > > I915_WRITE_FW(SPSIZE(pipe, plane_id), (crtc_h << 16) | crtc_w); > > I915_WRITE_FW(SPCONSTALPHA(pipe, plane_id), 0); > > - if (IS_CHERRYVIEW(dev_priv) && pipe == PIPE_B) > > + if (IS_CHERRYVIEW(i915) && pipe == PIPE_B) > > chv_update_csc(plane_state); > > if (key->flags) { > > @@ -917,38 +917,38 @@ vlv_update_plane(struct intel_plane *plane, > > vlv_update_clrc(plane_state); > > - spin_unlock_irqrestore(&dev_priv->uncore.lock, irqflags); > > + spin_unlock_irqrestore(&i915->uncore.lock, irqflags); > > } > > static void > > vlv_disable_plane(struct intel_plane *plane, > > const struct intel_crtc_state *crtc_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(plane->base.dev); > > + struct drm_i915_private *i915 = to_i915(plane->base.dev); > > enum pipe pipe = plane->pipe; > > enum plane_id plane_id = plane->id; > > unsigned long irqflags; > > - spin_lock_irqsave(&dev_priv->uncore.lock, irqflags); > > + spin_lock_irqsave(&i915->uncore.lock, irqflags); > > I915_WRITE_FW(SPCNTR(pipe, plane_id), 0); > > I915_WRITE_FW(SPSURF(pipe, plane_id), 0); > > - spin_unlock_irqrestore(&dev_priv->uncore.lock, irqflags); > > + spin_unlock_irqrestore(&i915->uncore.lock, irqflags); > > } > > static bool > > vlv_plane_get_hw_state(struct intel_plane *plane, > > enum pipe *pipe) > > { > > - struct drm_i915_private *dev_priv = to_i915(plane->base.dev); > > + struct drm_i915_private *i915 = to_i915(plane->base.dev); > > enum intel_display_power_domain power_domain; > > enum plane_id plane_id = plane->id; > > intel_wakeref_t wakeref; > > bool ret; > > power_domain = POWER_DOMAIN_PIPE(plane->pipe); > > - wakeref = intel_display_power_get_if_enabled(dev_priv, power_domain); > > + wakeref = intel_display_power_get_if_enabled(i915, power_domain); > > if (!wakeref) > > return false; > > @@ -956,7 +956,7 @@ vlv_plane_get_hw_state(struct intel_plane *plane, > > *pipe = plane->pipe; > > - intel_display_power_put(dev_priv, power_domain, wakeref); > > + intel_display_power_put(i915, power_domain, wakeref); > > return ret; > > } > > @@ -977,7 +977,7 @@ static u32 ivb_sprite_ctl_crtc(const struct intel_crtc_state *crtc_state) > > static u32 ivb_sprite_ctl(const struct intel_crtc_state *crtc_state, > > const struct intel_plane_state *plane_state) > > { > > - struct drm_i915_private *dev_priv = > > + struct drm_i915_private *i915 = > > to_i915(plane_state->base.plane->dev); > > const struct drm_framebuffer *fb = plane_state->base.fb; > > unsigned int rotation = plane_state->base.rotation; > > @@ -986,7 +986,7 @@ static u32 ivb_sprite_ctl(const struct intel_crtc_state *crtc_state, > > sprctl = SPRITE_ENABLE; > > - if (IS_IVYBRIDGE(dev_priv)) > > + if (IS_IVYBRIDGE(i915)) > > sprctl |= SPRITE_TRICKLE_FEED_DISABLE; > > switch (fb->format->format) { > > @@ -1038,7 +1038,7 @@ ivb_update_plane(struct intel_plane *plane, > > const struct intel_crtc_state *crtc_state, > > const struct intel_plane_state *plane_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(plane->base.dev); > > + struct drm_i915_private *i915 = to_i915(plane->base.dev); > > enum pipe pipe = plane->pipe; > > u32 sprsurf_offset = plane_state->color_plane[0].offset; > > u32 linear_offset; > > @@ -1067,12 +1067,12 @@ ivb_update_plane(struct intel_plane *plane, > > linear_offset = intel_fb_xy_to_linear(x, y, plane_state, 0); > > - spin_lock_irqsave(&dev_priv->uncore.lock, irqflags); > > + spin_lock_irqsave(&i915->uncore.lock, irqflags); > > I915_WRITE_FW(SPRSTRIDE(pipe), plane_state->color_plane[0].stride); > > I915_WRITE_FW(SPRPOS(pipe), (crtc_y << 16) | crtc_x); > > I915_WRITE_FW(SPRSIZE(pipe), (crtc_h << 16) | crtc_w); > > - if (IS_IVYBRIDGE(dev_priv)) > > + if (IS_IVYBRIDGE(i915)) > > I915_WRITE_FW(SPRSCALE(pipe), sprscale); > > if (key->flags) { > > @@ -1083,7 +1083,7 @@ ivb_update_plane(struct intel_plane *plane, > > /* HSW consolidates SPRTILEOFF and SPRLINOFF into a single SPROFFSET > > * register */ > > - if (IS_HASWELL(dev_priv) || IS_BROADWELL(dev_priv)) { > > + if (IS_HASWELL(i915) || IS_BROADWELL(i915)) { > > I915_WRITE_FW(SPROFFSET(pipe), (y << 16) | x); > > } else { > > I915_WRITE_FW(SPRLINOFF(pipe), linear_offset); > > @@ -1099,39 +1099,39 @@ ivb_update_plane(struct intel_plane *plane, > > I915_WRITE_FW(SPRSURF(pipe), > > intel_plane_ggtt_offset(plane_state) + sprsurf_offset); > > - spin_unlock_irqrestore(&dev_priv->uncore.lock, irqflags); > > + spin_unlock_irqrestore(&i915->uncore.lock, irqflags); > > } > > static void > > ivb_disable_plane(struct intel_plane *plane, > > const struct intel_crtc_state *crtc_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(plane->base.dev); > > + struct drm_i915_private *i915 = to_i915(plane->base.dev); > > enum pipe pipe = plane->pipe; > > unsigned long irqflags; > > - spin_lock_irqsave(&dev_priv->uncore.lock, irqflags); > > + spin_lock_irqsave(&i915->uncore.lock, irqflags); > > I915_WRITE_FW(SPRCTL(pipe), 0); > > /* Disable the scaler */ > > - if (IS_IVYBRIDGE(dev_priv)) > > + if (IS_IVYBRIDGE(i915)) > > I915_WRITE_FW(SPRSCALE(pipe), 0); > > I915_WRITE_FW(SPRSURF(pipe), 0); > > - spin_unlock_irqrestore(&dev_priv->uncore.lock, irqflags); > > + spin_unlock_irqrestore(&i915->uncore.lock, irqflags); > > } > > static bool > > ivb_plane_get_hw_state(struct intel_plane *plane, > > enum pipe *pipe) > > { > > - struct drm_i915_private *dev_priv = to_i915(plane->base.dev); > > + struct drm_i915_private *i915 = to_i915(plane->base.dev); > > enum intel_display_power_domain power_domain; > > intel_wakeref_t wakeref; > > bool ret; > > power_domain = POWER_DOMAIN_PIPE(plane->pipe); > > - wakeref = intel_display_power_get_if_enabled(dev_priv, power_domain); > > + wakeref = intel_display_power_get_if_enabled(i915, power_domain); > > if (!wakeref) > > return false; > > @@ -1139,7 +1139,7 @@ ivb_plane_get_hw_state(struct intel_plane *plane, > > *pipe = plane->pipe; > > - intel_display_power_put(dev_priv, power_domain, wakeref); > > + intel_display_power_put(i915, power_domain, wakeref); > > return ret; > > } > > @@ -1168,7 +1168,7 @@ static u32 g4x_sprite_ctl_crtc(const struct intel_crtc_state *crtc_state) > > static u32 g4x_sprite_ctl(const struct intel_crtc_state *crtc_state, > > const struct intel_plane_state *plane_state) > > { > > - struct drm_i915_private *dev_priv = > > + struct drm_i915_private *i915 = > > to_i915(plane_state->base.plane->dev); > > const struct drm_framebuffer *fb = plane_state->base.fb; > > unsigned int rotation = plane_state->base.rotation; > > @@ -1177,7 +1177,7 @@ static u32 g4x_sprite_ctl(const struct intel_crtc_state *crtc_state, > > dvscntr = DVS_ENABLE; > > - if (IS_GEN(dev_priv, 6)) > > + if (IS_GEN(i915, 6)) > > dvscntr |= DVS_TRICKLE_FEED_DISABLE; > > switch (fb->format->format) { > > @@ -1229,7 +1229,7 @@ g4x_update_plane(struct intel_plane *plane, > > const struct intel_crtc_state *crtc_state, > > const struct intel_plane_state *plane_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(plane->base.dev); > > + struct drm_i915_private *i915 = to_i915(plane->base.dev); > > enum pipe pipe = plane->pipe; > > u32 dvssurf_offset = plane_state->color_plane[0].offset; > > u32 linear_offset; > > @@ -1258,7 +1258,7 @@ g4x_update_plane(struct intel_plane *plane, > > linear_offset = intel_fb_xy_to_linear(x, y, plane_state, 0); > > - spin_lock_irqsave(&dev_priv->uncore.lock, irqflags); > > + spin_lock_irqsave(&i915->uncore.lock, irqflags); > > I915_WRITE_FW(DVSSTRIDE(pipe), plane_state->color_plane[0].stride); > > I915_WRITE_FW(DVSPOS(pipe), (crtc_y << 16) | crtc_x); > > @@ -1283,38 +1283,38 @@ g4x_update_plane(struct intel_plane *plane, > > I915_WRITE_FW(DVSSURF(pipe), > > intel_plane_ggtt_offset(plane_state) + dvssurf_offset); > > - spin_unlock_irqrestore(&dev_priv->uncore.lock, irqflags); > > + spin_unlock_irqrestore(&i915->uncore.lock, irqflags); > > } > > static void > > g4x_disable_plane(struct intel_plane *plane, > > const struct intel_crtc_state *crtc_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(plane->base.dev); > > + struct drm_i915_private *i915 = to_i915(plane->base.dev); > > enum pipe pipe = plane->pipe; > > unsigned long irqflags; > > - spin_lock_irqsave(&dev_priv->uncore.lock, irqflags); > > + spin_lock_irqsave(&i915->uncore.lock, irqflags); > > I915_WRITE_FW(DVSCNTR(pipe), 0); > > /* Disable the scaler */ > > I915_WRITE_FW(DVSSCALE(pipe), 0); > > I915_WRITE_FW(DVSSURF(pipe), 0); > > - spin_unlock_irqrestore(&dev_priv->uncore.lock, irqflags); > > + spin_unlock_irqrestore(&i915->uncore.lock, irqflags); > > } > > static bool > > g4x_plane_get_hw_state(struct intel_plane *plane, > > enum pipe *pipe) > > { > > - struct drm_i915_private *dev_priv = to_i915(plane->base.dev); > > + struct drm_i915_private *i915 = to_i915(plane->base.dev); > > enum intel_display_power_domain power_domain; > > intel_wakeref_t wakeref; > > bool ret; > > power_domain = POWER_DOMAIN_PIPE(plane->pipe); > > - wakeref = intel_display_power_get_if_enabled(dev_priv, power_domain); > > + wakeref = intel_display_power_get_if_enabled(i915, power_domain); > > if (!wakeref) > > return false; > > @@ -1322,7 +1322,7 @@ g4x_plane_get_hw_state(struct intel_plane *plane, > > *pipe = plane->pipe; > > - intel_display_power_put(dev_priv, power_domain, wakeref); > > + intel_display_power_put(i915, power_domain, wakeref); > > return ret; > > } > > @@ -1406,16 +1406,16 @@ g4x_sprite_check(struct intel_crtc_state *crtc_state, > > struct intel_plane_state *plane_state) > > { > > struct intel_plane *plane = to_intel_plane(plane_state->base.plane); > > - struct drm_i915_private *dev_priv = to_i915(plane->base.dev); > > + struct drm_i915_private *i915 = to_i915(plane->base.dev); > > int min_scale = DRM_PLANE_HELPER_NO_SCALING; > > int max_scale = DRM_PLANE_HELPER_NO_SCALING; > > int ret; > > if (intel_fb_scalable(plane_state->base.fb)) { > > - if (INTEL_GEN(dev_priv) < 7) { > > + if (INTEL_GEN(i915) < 7) { > > min_scale = 1; > > max_scale = 16 << 16; > > - } else if (IS_IVYBRIDGE(dev_priv)) { > > + } else if (IS_IVYBRIDGE(i915)) { > > min_scale = 1; > > max_scale = 2 << 16; > > } > > @@ -1443,7 +1443,7 @@ g4x_sprite_check(struct intel_crtc_state *crtc_state, > > if (ret) > > return ret; > > - if (INTEL_GEN(dev_priv) >= 7) > > + if (INTEL_GEN(i915) >= 7) > > plane_state->ctl = ivb_sprite_ctl(crtc_state, plane_state); > > else > > plane_state->ctl = g4x_sprite_ctl(crtc_state, plane_state); > > @@ -1454,11 +1454,11 @@ g4x_sprite_check(struct intel_crtc_state *crtc_state, > > int chv_plane_check_rotation(const struct intel_plane_state *plane_state) > > { > > struct intel_plane *plane = to_intel_plane(plane_state->base.plane); > > - struct drm_i915_private *dev_priv = to_i915(plane->base.dev); > > + struct drm_i915_private *i915 = to_i915(plane->base.dev); > > unsigned int rotation = plane_state->base.rotation; > > /* CHV ignores the mirror bit when the rotate bit is set :( */ > > - if (IS_CHERRYVIEW(dev_priv) && > > + if (IS_CHERRYVIEW(i915) && > > rotation & DRM_MODE_ROTATE_180 && > > rotation & DRM_MODE_REFLECT_X) { > > DRM_DEBUG_KMS("Cannot rotate and reflect at the same time\n"); > > @@ -1506,7 +1506,7 @@ static int skl_plane_check_fb(const struct intel_crtc_state *crtc_state, > > const struct intel_plane_state *plane_state) > > { > > struct intel_plane *plane = to_intel_plane(plane_state->base.plane); > > - struct drm_i915_private *dev_priv = to_i915(plane->base.dev); > > + struct drm_i915_private *i915 = to_i915(plane->base.dev); > > const struct drm_framebuffer *fb = plane_state->base.fb; > > unsigned int rotation = plane_state->base.rotation; > > struct drm_format_name_buf format_name; > > @@ -1540,7 +1540,7 @@ static int skl_plane_check_fb(const struct intel_crtc_state *crtc_state, > > */ > > switch (fb->format->format) { > > case DRM_FORMAT_RGB565: > > - if (INTEL_GEN(dev_priv) >= 11) > > + if (INTEL_GEN(i915) >= 11) > > break; > > /* fall through */ > > case DRM_FORMAT_C8: > > @@ -1579,7 +1579,7 @@ static int skl_plane_check_fb(const struct intel_crtc_state *crtc_state, > > static int skl_plane_check_dst_coordinates(const struct intel_crtc_state *crtc_state, > > const struct intel_plane_state *plane_state) > > { > > - struct drm_i915_private *dev_priv = > > + struct drm_i915_private *i915 = > > to_i915(plane_state->base.plane->dev); > > int crtc_x = plane_state->base.dst.x1; > > int crtc_w = drm_rect_width(&plane_state->base.dst); > > @@ -1594,7 +1594,7 @@ static int skl_plane_check_dst_coordinates(const struct intel_crtc_state *crtc_s > > * than the cursor ending less than 4 pixels from the left edge of the > > * screen may cause FIFO underflow and display corruption. > > */ > > - if ((IS_GEMINILAKE(dev_priv) || IS_CANNONLAKE(dev_priv)) && > > + if ((IS_GEMINILAKE(i915) || IS_CANNONLAKE(i915)) && > > (crtc_x + crtc_w < 4 || crtc_x > pipe_src_w - 4)) { > > DRM_DEBUG_KMS("requested plane X %s position %d invalid (valid range %d-%d)\n", > > crtc_x + crtc_w < 4 ? "end" : "start", > > @@ -1627,7 +1627,7 @@ static int skl_plane_check(struct intel_crtc_state *crtc_state, > > struct intel_plane_state *plane_state) > > { > > struct intel_plane *plane = to_intel_plane(plane_state->base.plane); > > - struct drm_i915_private *dev_priv = to_i915(plane->base.dev); > > + struct drm_i915_private *i915 = to_i915(plane->base.dev); > > const struct drm_framebuffer *fb = plane_state->base.fb; > > int min_scale = DRM_PLANE_HELPER_NO_SCALING; > > int max_scale = DRM_PLANE_HELPER_NO_SCALING; > > @@ -1675,23 +1675,23 @@ static int skl_plane_check(struct intel_crtc_state *crtc_state, > > plane_state->ctl = skl_plane_ctl(crtc_state, plane_state); > > - if (INTEL_GEN(dev_priv) >= 10 || IS_GEMINILAKE(dev_priv)) > > + if (INTEL_GEN(i915) >= 10 || IS_GEMINILAKE(i915)) > > plane_state->color_ctl = glk_plane_color_ctl(crtc_state, > > plane_state); > > return 0; > > } > > -static bool has_dst_key_in_primary_plane(struct drm_i915_private *dev_priv) > > +static bool has_dst_key_in_primary_plane(struct drm_i915_private *i915) > > { > > - return INTEL_GEN(dev_priv) >= 9; > > + return INTEL_GEN(i915) >= 9; > > } > > static void intel_plane_set_ckey(struct intel_plane_state *plane_state, > > const struct drm_intel_sprite_colorkey *set) > > { > > struct intel_plane *plane = to_intel_plane(plane_state->base.plane); > > - struct drm_i915_private *dev_priv = to_i915(plane->base.dev); > > + struct drm_i915_private *i915 = to_i915(plane->base.dev); > > struct drm_intel_sprite_colorkey *key = &plane_state->ckey; > > *key = *set; > > @@ -1708,7 +1708,7 @@ static void intel_plane_set_ckey(struct intel_plane_state *plane_state, > > * On SKL+ we want dst key enabled on > > * the primary and not on the sprite. > > */ > > - if (INTEL_GEN(dev_priv) >= 9 && plane->id != PLANE_PRIMARY && > > + if (INTEL_GEN(i915) >= 9 && plane->id != PLANE_PRIMARY && > > set->flags & I915_SET_COLORKEY_DESTINATION) > > key->flags = 0; > > } > > @@ -1716,7 +1716,7 @@ static void intel_plane_set_ckey(struct intel_plane_state *plane_state, > > int intel_sprite_set_colorkey_ioctl(struct drm_device *dev, void *data, > > struct drm_file *file_priv) > > { > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > struct drm_intel_sprite_colorkey *set = data; > > struct drm_plane *plane; > > struct drm_plane_state *plane_state; > > @@ -1734,7 +1734,7 @@ int intel_sprite_set_colorkey_ioctl(struct drm_device *dev, void *data, > > if ((set->flags & (I915_SET_COLORKEY_DESTINATION | I915_SET_COLORKEY_SOURCE)) == (I915_SET_COLORKEY_DESTINATION | I915_SET_COLORKEY_SOURCE)) > > return -EINVAL; > > - if ((IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) && > > + if ((IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) && > > set->flags & I915_SET_COLORKEY_DESTINATION) > > return -EINVAL; > > @@ -1747,7 +1747,7 @@ int intel_sprite_set_colorkey_ioctl(struct drm_device *dev, void *data, > > * Also multiple planes can't do destination keying on the same > > * pipe simultaneously. > > */ > > - if (INTEL_GEN(dev_priv) >= 9 && > > + if (INTEL_GEN(i915) >= 9 && > > to_intel_plane(plane)->id >= PLANE_SPRITE1 && > > set->flags & I915_SET_COLORKEY_DESTINATION) > > return -EINVAL; > > @@ -1771,9 +1771,9 @@ int intel_sprite_set_colorkey_ioctl(struct drm_device *dev, void *data, > > * On some platforms we have to configure > > * the dst colorkey on the primary plane. > > */ > > - if (!ret && has_dst_key_in_primary_plane(dev_priv)) { > > + if (!ret && has_dst_key_in_primary_plane(i915)) { > > struct intel_crtc *crtc = > > - intel_get_crtc_for_pipe(dev_priv, > > + intel_get_crtc_for_pipe(i915, > > to_intel_plane(plane)->pipe); > > plane_state = drm_atomic_get_plane_state(state, > > @@ -2197,26 +2197,26 @@ static const struct drm_plane_funcs skl_plane_funcs = { > > .format_mod_supported = skl_plane_format_mod_supported, > > }; > > -static bool skl_plane_has_fbc(struct drm_i915_private *dev_priv, > > +static bool skl_plane_has_fbc(struct drm_i915_private *i915, > > enum pipe pipe, enum plane_id plane_id) > > { > > - if (!HAS_FBC(dev_priv)) > > + if (!HAS_FBC(i915)) > > return false; > > return pipe == PIPE_A && plane_id == PLANE_PRIMARY; > > } > > -static bool skl_plane_has_planar(struct drm_i915_private *dev_priv, > > +static bool skl_plane_has_planar(struct drm_i915_private *i915, > > enum pipe pipe, enum plane_id plane_id) > > { > > - if (INTEL_GEN(dev_priv) >= 11) > > + if (INTEL_GEN(i915) >= 11) > > return plane_id <= PLANE_SPRITE3; > > /* Display WA #0870: skl, bxt */ > > - if (IS_SKYLAKE(dev_priv) || IS_BROXTON(dev_priv)) > > + if (IS_SKYLAKE(i915) || IS_BROXTON(i915)) > > return false; > > - if (IS_GEN(dev_priv, 9) && !IS_GEMINILAKE(dev_priv) && pipe == PIPE_C) > > + if (IS_GEN(i915, 9) && !IS_GEMINILAKE(i915) && pipe == PIPE_C) > > return false; > > if (plane_id != PLANE_PRIMARY && plane_id != PLANE_SPRITE0) > > @@ -2225,16 +2225,16 @@ static bool skl_plane_has_planar(struct drm_i915_private *dev_priv, > > return true; > > } > > -static bool skl_plane_has_ccs(struct drm_i915_private *dev_priv, > > +static bool skl_plane_has_ccs(struct drm_i915_private *i915, > > enum pipe pipe, enum plane_id plane_id) > > { > > if (plane_id == PLANE_CURSOR) > > return false; > > - if (INTEL_GEN(dev_priv) >= 10) > > + if (INTEL_GEN(i915) >= 10) > > return true; > > - if (IS_GEMINILAKE(dev_priv)) > > + if (IS_GEMINILAKE(i915)) > > return pipe != PIPE_C; > > return pipe != PIPE_C && > > @@ -2243,7 +2243,7 @@ static bool skl_plane_has_ccs(struct drm_i915_private *dev_priv, > > } > > struct intel_plane * > > -skl_universal_plane_create(struct drm_i915_private *dev_priv, > > +skl_universal_plane_create(struct drm_i915_private *i915, > > enum pipe pipe, enum plane_id plane_id) > > { > > struct intel_plane *plane; > > @@ -2263,9 +2263,9 @@ skl_universal_plane_create(struct drm_i915_private *dev_priv, > > plane->id = plane_id; > > plane->frontbuffer_bit = INTEL_FRONTBUFFER(pipe, plane_id); > > - plane->has_fbc = skl_plane_has_fbc(dev_priv, pipe, plane_id); > > + plane->has_fbc = skl_plane_has_fbc(i915, pipe, plane_id); > > if (plane->has_fbc) { > > - struct intel_fbc *fbc = &dev_priv->fbc; > > + struct intel_fbc *fbc = &i915->fbc; > > fbc->possible_framebuffer_bits |= plane->frontbuffer_bit; > > } > > @@ -2278,24 +2278,24 @@ skl_universal_plane_create(struct drm_i915_private *dev_priv, > > if (icl_is_nv12_y_plane(plane_id)) > > plane->update_slave = icl_update_slave; > > - if (skl_plane_has_planar(dev_priv, pipe, plane_id)) { > > - if (icl_is_hdr_plane(dev_priv, plane_id)) { > > + if (skl_plane_has_planar(i915, pipe, plane_id)) { > > + if (icl_is_hdr_plane(i915, plane_id)) { > > formats = icl_hdr_planar_formats; > > num_formats = ARRAY_SIZE(icl_hdr_planar_formats); > > - } else if (INTEL_GEN(dev_priv) >= 11) { > > + } else if (INTEL_GEN(i915) >= 11) { > > formats = icl_planar_formats; > > num_formats = ARRAY_SIZE(icl_planar_formats); > > - } else if (INTEL_GEN(dev_priv) == 10 || IS_GEMINILAKE(dev_priv)) { > > + } else if (INTEL_GEN(i915) == 10 || IS_GEMINILAKE(i915)) { > > formats = glk_planar_formats; > > num_formats = ARRAY_SIZE(glk_planar_formats); > > } else { > > formats = skl_planar_formats; > > num_formats = ARRAY_SIZE(skl_planar_formats); > > } > > - } else if (icl_is_hdr_plane(dev_priv, plane_id)) { > > + } else if (icl_is_hdr_plane(i915, plane_id)) { > > formats = icl_hdr_plane_formats; > > num_formats = ARRAY_SIZE(icl_hdr_plane_formats); > > - } else if (INTEL_GEN(dev_priv) >= 11) { > > + } else if (INTEL_GEN(i915) >= 11) { > > formats = icl_plane_formats; > > num_formats = ARRAY_SIZE(icl_plane_formats); > > } else { > > @@ -2303,7 +2303,7 @@ skl_universal_plane_create(struct drm_i915_private *dev_priv, > > num_formats = ARRAY_SIZE(skl_plane_formats); > > } > > - plane->has_ccs = skl_plane_has_ccs(dev_priv, pipe, plane_id); > > + plane->has_ccs = skl_plane_has_ccs(i915, pipe, plane_id); > > if (plane->has_ccs) > > modifiers = skl_plane_format_modifiers_ccs; > > else > > @@ -2316,7 +2316,7 @@ skl_universal_plane_create(struct drm_i915_private *dev_priv, > > possible_crtcs = BIT(pipe); > > - ret = drm_universal_plane_init(&dev_priv->drm, &plane->base, > > + ret = drm_universal_plane_init(&i915->drm, &plane->base, > > possible_crtcs, &skl_plane_funcs, > > formats, num_formats, modifiers, > > plane_type, > > @@ -2329,7 +2329,7 @@ skl_universal_plane_create(struct drm_i915_private *dev_priv, > > DRM_MODE_ROTATE_0 | DRM_MODE_ROTATE_90 | > > DRM_MODE_ROTATE_180 | DRM_MODE_ROTATE_270; > > - if (INTEL_GEN(dev_priv) >= 10) > > + if (INTEL_GEN(i915) >= 10) > > supported_rotations |= DRM_MODE_REFLECT_X; > > drm_plane_create_rotation_property(&plane->base, > > @@ -2361,7 +2361,7 @@ skl_universal_plane_create(struct drm_i915_private *dev_priv, > > } > > struct intel_plane * > > -intel_sprite_plane_create(struct drm_i915_private *dev_priv, > > +intel_sprite_plane_create(struct drm_i915_private *i915, > > enum pipe pipe, int sprite) > > { > > struct intel_plane *plane; > > @@ -2373,15 +2373,15 @@ intel_sprite_plane_create(struct drm_i915_private *dev_priv, > > int num_formats; > > int ret; > > - if (INTEL_GEN(dev_priv) >= 9) > > - return skl_universal_plane_create(dev_priv, pipe, > > + if (INTEL_GEN(i915) >= 9) > > + return skl_universal_plane_create(i915, pipe, > > PLANE_SPRITE0 + sprite); > > plane = intel_plane_alloc(); > > if (IS_ERR(plane)) > > return plane; > > - if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) { > > + if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) { > > plane->max_stride = i9xx_plane_max_stride; > > plane->update_plane = vlv_update_plane; > > plane->disable_plane = vlv_disable_plane; > > @@ -2393,7 +2393,7 @@ intel_sprite_plane_create(struct drm_i915_private *dev_priv, > > modifiers = i9xx_plane_format_modifiers; > > plane_funcs = &vlv_sprite_funcs; > > - } else if (INTEL_GEN(dev_priv) >= 7) { > > + } else if (INTEL_GEN(i915) >= 7) { > > plane->max_stride = g4x_sprite_max_stride; > > plane->update_plane = ivb_update_plane; > > plane->disable_plane = ivb_disable_plane; > > @@ -2413,7 +2413,7 @@ intel_sprite_plane_create(struct drm_i915_private *dev_priv, > > plane->check_plane = g4x_sprite_check; > > modifiers = i9xx_plane_format_modifiers; > > - if (IS_GEN(dev_priv, 6)) { > > + if (IS_GEN(i915, 6)) { > > formats = snb_plane_formats; > > num_formats = ARRAY_SIZE(snb_plane_formats); > > @@ -2426,7 +2426,7 @@ intel_sprite_plane_create(struct drm_i915_private *dev_priv, > > } > > } > > - if (IS_CHERRYVIEW(dev_priv) && pipe == PIPE_B) { > > + if (IS_CHERRYVIEW(i915) && pipe == PIPE_B) { > > supported_rotations = > > DRM_MODE_ROTATE_0 | DRM_MODE_ROTATE_180 | > > DRM_MODE_REFLECT_X; > > @@ -2441,7 +2441,7 @@ intel_sprite_plane_create(struct drm_i915_private *dev_priv, > > possible_crtcs = BIT(pipe); > > - ret = drm_universal_plane_init(&dev_priv->drm, &plane->base, > > + ret = drm_universal_plane_init(&i915->drm, &plane->base, > > possible_crtcs, plane_funcs, > > formats, num_formats, modifiers, > > DRM_PLANE_TYPE_OVERLAY, > > diff --git a/drivers/gpu/drm/i915/intel_sprite.h b/drivers/gpu/drm/i915/intel_sprite.h > > index 500f6bffb139..401295e05116 100644 > > --- a/drivers/gpu/drm/i915/intel_sprite.h > > +++ b/drivers/gpu/drm/i915/intel_sprite.h > > @@ -21,7 +21,7 @@ struct intel_plane_state; > > bool is_planar_yuv_format(u32 pixelformat); > > int intel_usecs_to_scanlines(const struct drm_display_mode *adjusted_mode, > > int usecs); > > -struct intel_plane *intel_sprite_plane_create(struct drm_i915_private *dev_priv, > > +struct intel_plane *intel_sprite_plane_create(struct drm_i915_private *i915, > > enum pipe pipe, int plane); > > int intel_sprite_set_colorkey_ioctl(struct drm_device *dev, void *data, > > struct drm_file *file_priv); > > @@ -31,7 +31,7 @@ int intel_plane_check_stride(const struct intel_plane_state *plane_state); > > int intel_plane_check_src_coordinates(struct intel_plane_state *plane_state); > > int chv_plane_check_rotation(const struct intel_plane_state *plane_state); > > struct intel_plane * > > -skl_universal_plane_create(struct drm_i915_private *dev_priv, > > +skl_universal_plane_create(struct drm_i915_private *i915, > > enum pipe pipe, enum plane_id plane_id); > > static inline bool icl_is_nv12_y_plane(enum plane_id id) > > @@ -49,10 +49,10 @@ static inline u8 icl_hdr_plane_mask(void) > > BIT(PLANE_SPRITE0) | BIT(PLANE_SPRITE1); > > } > > -static inline bool icl_is_hdr_plane(struct drm_i915_private *dev_priv, > > +static inline bool icl_is_hdr_plane(struct drm_i915_private *i915, > > enum plane_id plane_id) > > { > > - return INTEL_GEN(dev_priv) >= 11 && > > + return INTEL_GEN(i915) >= 11 && > > icl_hdr_plane_mask() & BIT(plane_id); > > } > > diff --git a/drivers/gpu/drm/i915/intel_tv.c b/drivers/gpu/drm/i915/intel_tv.c > > index 5dc594eafaf2..4e66261bf5af 100644 > > --- a/drivers/gpu/drm/i915/intel_tv.c > > +++ b/drivers/gpu/drm/i915/intel_tv.c > > @@ -906,7 +906,7 @@ static struct intel_tv *intel_attached_tv(struct drm_connector *connector) > > static bool > > intel_tv_get_hw_state(struct intel_encoder *encoder, enum pipe *pipe) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > u32 tmp = I915_READ(TV_CTL); > > *pipe = (tmp & TV_ENC_PIPE_SEL_MASK) >> TV_ENC_PIPE_SEL_SHIFT; > > @@ -920,10 +920,10 @@ intel_enable_tv(struct intel_encoder *encoder, > > const struct drm_connector_state *conn_state) > > { > > struct drm_device *dev = encoder->base.dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > /* Prevents vblank waits from timing out in intel_tv_detect_type() */ > > - intel_wait_for_vblank(dev_priv, > > + intel_wait_for_vblank(i915, > > to_intel_crtc(pipe_config->base.crtc)->pipe); > > I915_WRITE(TV_CTL, I915_READ(TV_CTL) | TV_ENC_ENABLE); > > @@ -935,7 +935,7 @@ intel_disable_tv(struct intel_encoder *encoder, > > const struct drm_connector_state *old_conn_state) > > { > > struct drm_device *dev = encoder->base.dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > I915_WRITE(TV_CTL, I915_READ(TV_CTL) & ~TV_ENC_ENABLE); > > } > > @@ -1084,7 +1084,7 @@ static void > > intel_tv_get_config(struct intel_encoder *encoder, > > struct intel_crtc_state *pipe_config) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct drm_display_mode *adjusted_mode = > > &pipe_config->base.adjusted_mode; > > struct drm_display_mode mode = {}; > > @@ -1158,15 +1158,15 @@ intel_tv_get_config(struct intel_encoder *encoder, > > adjusted_mode->crtc_clock /= 2; > > /* pixel counter doesn't work on i965gm TV output */ > > - if (IS_I965GM(dev_priv)) > > + if (IS_I965GM(i915)) > > adjusted_mode->private_flags |= > > I915_MODE_FLAG_USE_SCANLINE_COUNTER; > > } > > -static bool intel_tv_source_too_wide(struct drm_i915_private *dev_priv, > > +static bool intel_tv_source_too_wide(struct drm_i915_private *i915, > > int hdisplay) > > { > > - return IS_GEN(dev_priv, 3) && hdisplay > 1024; > > + return IS_GEN(i915, 3) && hdisplay > 1024; > > } > > static bool intel_tv_vert_scaling(const struct drm_display_mode *tv_mode, > > @@ -1184,7 +1184,7 @@ intel_tv_compute_config(struct intel_encoder *encoder, > > struct intel_crtc_state *pipe_config, > > struct drm_connector_state *conn_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_tv_connector_state *tv_conn_state = > > to_intel_tv_connector_state(conn_state); > > const struct tv_mode *tv_mode = intel_tv_mode_find(conn_state); > > @@ -1209,7 +1209,7 @@ intel_tv_compute_config(struct intel_encoder *encoder, > > intel_tv_mode_to_mode(adjusted_mode, tv_mode); > > drm_mode_set_crtcinfo(adjusted_mode, 0); > > - if (intel_tv_source_too_wide(dev_priv, hdisplay) || > > + if (intel_tv_source_too_wide(i915, hdisplay) || > > !intel_tv_vert_scaling(adjusted_mode, conn_state, vdisplay)) { > > int extra, top, bottom; > > @@ -1306,7 +1306,7 @@ intel_tv_compute_config(struct intel_encoder *encoder, > > * the active portion. Hence following this formula seems > > * more trouble that it's worth. > > * > > - * if (IS_GEN(dev_priv, 4)) { > > + * if (IS_GEN(i915, 4)) { > > * num = cdclk * (tv_mode->oversample >> !tv_mode->progressive); > > * den = tv_mode->clock; > > * } else { > > @@ -1327,7 +1327,7 @@ intel_tv_compute_config(struct intel_encoder *encoder, > > adjusted_mode->name[0] = '\0'; > > /* pixel counter doesn't work on i965gm TV output */ > > - if (IS_I965GM(dev_priv)) > > + if (IS_I965GM(i915)) > > adjusted_mode->private_flags |= > > I915_MODE_FLAG_USE_SCANLINE_COUNTER; > > @@ -1335,7 +1335,7 @@ intel_tv_compute_config(struct intel_encoder *encoder, > > } > > static void > > -set_tv_mode_timings(struct drm_i915_private *dev_priv, > > +set_tv_mode_timings(struct drm_i915_private *i915, > > const struct tv_mode *tv_mode, > > bool burst_ena) > > { > > @@ -1393,7 +1393,7 @@ set_tv_mode_timings(struct drm_i915_private *dev_priv, > > I915_WRITE(TV_V_CTL_7, vctl7); > > } > > -static void set_color_conversion(struct drm_i915_private *dev_priv, > > +static void set_color_conversion(struct drm_i915_private *i915, > > const struct color_conversion *color_conversion) > > { > > if (!color_conversion) > > @@ -1417,7 +1417,7 @@ static void intel_tv_pre_enable(struct intel_encoder *encoder, > > const struct intel_crtc_state *pipe_config, > > const struct drm_connector_state *conn_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_crtc *intel_crtc = to_intel_crtc(pipe_config->base.crtc); > > struct intel_tv *intel_tv = enc_to_tv(encoder); > > const struct intel_tv_connector_state *tv_conn_state = > > @@ -1507,18 +1507,18 @@ static void intel_tv_pre_enable(struct intel_encoder *encoder, > > tv_mode->dda3_inc << TV_SCDDA3_INC_SHIFT; > > /* Enable two fixes for the chips that need them. */ > > - if (IS_I915GM(dev_priv)) > > + if (IS_I915GM(i915)) > > tv_ctl |= TV_ENC_C0_FIX | TV_ENC_SDP_FIX; > > - set_tv_mode_timings(dev_priv, tv_mode, burst_ena); > > + set_tv_mode_timings(i915, tv_mode, burst_ena); > > I915_WRITE(TV_SC_CTL_1, scctl1); > > I915_WRITE(TV_SC_CTL_2, scctl2); > > I915_WRITE(TV_SC_CTL_3, scctl3); > > - set_color_conversion(dev_priv, color_conversion); > > + set_color_conversion(i915, color_conversion); > > - if (INTEL_GEN(dev_priv) >= 4) > > + if (INTEL_GEN(i915) >= 4) > > I915_WRITE(TV_CLR_KNOBS, 0x00404000); > > else > > I915_WRITE(TV_CLR_KNOBS, 0x00606000); > > @@ -1528,7 +1528,7 @@ static void intel_tv_pre_enable(struct intel_encoder *encoder, > > ((video_levels->black << TV_BLACK_LEVEL_SHIFT) | > > (video_levels->blank << TV_BLANK_LEVEL_SHIFT))); > > - assert_pipe_disabled(dev_priv, intel_crtc->pipe); > > + assert_pipe_disabled(i915, intel_crtc->pipe); > > /* Filter ctl must be set before TV_WIN_SIZE */ > > tv_filter_ctl = TV_AUTO_SCALE; > > @@ -1568,18 +1568,18 @@ intel_tv_detect_type(struct intel_tv *intel_tv, > > struct drm_crtc *crtc = connector->state->crtc; > > struct intel_crtc *intel_crtc = to_intel_crtc(crtc); > > struct drm_device *dev = connector->dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > u32 tv_ctl, save_tv_ctl; > > u32 tv_dac, save_tv_dac; > > int type; > > /* Disable TV interrupts around load detect or we'll recurse */ > > if (connector->polled & DRM_CONNECTOR_POLL_HPD) { > > - spin_lock_irq(&dev_priv->irq_lock); > > - i915_disable_pipestat(dev_priv, 0, > > + spin_lock_irq(&i915->irq_lock); > > + i915_disable_pipestat(i915, 0, > > PIPE_HOTPLUG_INTERRUPT_STATUS | > > PIPE_HOTPLUG_TV_INTERRUPT_STATUS); > > - spin_unlock_irq(&dev_priv->irq_lock); > > + spin_unlock_irq(&i915->irq_lock); > > } > > save_tv_dac = tv_dac = I915_READ(TV_DAC); > > @@ -1605,7 +1605,7 @@ intel_tv_detect_type(struct intel_tv *intel_tv, > > * The TV sense state should be cleared to zero on cantiga platform. Otherwise > > * the TV is misdetected. This is hardware requirement. > > */ > > - if (IS_GM45(dev_priv)) > > + if (IS_GM45(i915)) > > tv_dac &= ~(TVDAC_STATE_CHG_EN | TVDAC_A_SENSE_CTL | > > TVDAC_B_SENSE_CTL | TVDAC_C_SENSE_CTL); > > @@ -1613,7 +1613,7 @@ intel_tv_detect_type(struct intel_tv *intel_tv, > > I915_WRITE(TV_DAC, tv_dac); > > POSTING_READ(TV_DAC); > > - intel_wait_for_vblank(dev_priv, intel_crtc->pipe); > > + intel_wait_for_vblank(i915, intel_crtc->pipe); > > type = -1; > > tv_dac = I915_READ(TV_DAC); > > @@ -1643,15 +1643,15 @@ intel_tv_detect_type(struct intel_tv *intel_tv, > > POSTING_READ(TV_CTL); > > /* For unknown reasons the hw barfs if we don't do this vblank wait. */ > > - intel_wait_for_vblank(dev_priv, intel_crtc->pipe); > > + intel_wait_for_vblank(i915, intel_crtc->pipe); > > /* Restore interrupt config */ > > if (connector->polled & DRM_CONNECTOR_POLL_HPD) { > > - spin_lock_irq(&dev_priv->irq_lock); > > - i915_enable_pipestat(dev_priv, 0, > > + spin_lock_irq(&i915->irq_lock); > > + i915_enable_pipestat(i915, 0, > > PIPE_HOTPLUG_INTERRUPT_STATUS | > > PIPE_HOTPLUG_TV_INTERRUPT_STATUS); > > - spin_unlock_irq(&dev_priv->irq_lock); > > + spin_unlock_irq(&i915->irq_lock); > > } > > return type; > > @@ -1764,7 +1764,7 @@ intel_tv_set_mode_type(struct drm_display_mode *mode, > > static int > > intel_tv_get_modes(struct drm_connector *connector) > > { > > - struct drm_i915_private *dev_priv = to_i915(connector->dev); > > + struct drm_i915_private *i915 = to_i915(connector->dev); > > const struct tv_mode *tv_mode = intel_tv_mode_find(connector->state); > > int i, count = 0; > > @@ -1778,7 +1778,7 @@ intel_tv_get_modes(struct drm_connector *connector) > > continue; > > /* no vertical scaling with wide sources on gen3 */ > > - if (IS_GEN(dev_priv, 3) && input->w > 1024 && > > + if (IS_GEN(i915, 3) && input->w > 1024 && > > input->h > intel_tv_mode_vdisplay(tv_mode)) > > continue; > > @@ -1857,9 +1857,9 @@ static const struct drm_encoder_funcs intel_tv_enc_funcs = { > > }; > > void > > -intel_tv_init(struct drm_i915_private *dev_priv) > > +intel_tv_init(struct drm_i915_private *i915) > > { > > - struct drm_device *dev = &dev_priv->drm; > > + struct drm_device *dev = &i915->drm; > > struct drm_connector *connector; > > struct intel_tv *intel_tv; > > struct intel_encoder *intel_encoder; > > @@ -1872,7 +1872,7 @@ intel_tv_init(struct drm_i915_private *dev_priv) > > if ((I915_READ(TV_CTL) & TV_FUSE_STATE_MASK) == TV_FUSE_STATE_DISABLED) > > return; > > - if (!intel_bios_is_tv_present(dev_priv)) { > > + if (!intel_bios_is_tv_present(i915)) { > > DRM_DEBUG_KMS("Integrated TV is not present.\n"); > > return; > > } > > @@ -1966,7 +1966,7 @@ intel_tv_init(struct drm_i915_private *dev_priv) > > /* Create TV properties then attach current values */ > > for (i = 0; i < ARRAY_SIZE(tv_modes); i++) { > > /* 1080p50/1080p60 not supported on gen3 */ > > - if (IS_GEN(dev_priv, 3) && > > + if (IS_GEN(i915, 3) && > > tv_modes[i].oversample == 1) > > break; > > diff --git a/drivers/gpu/drm/i915/intel_tv.h b/drivers/gpu/drm/i915/intel_tv.h > > index 44518575ec5c..fc0b89563b4f 100644 > > --- a/drivers/gpu/drm/i915/intel_tv.h > > +++ b/drivers/gpu/drm/i915/intel_tv.h > > @@ -8,6 +8,6 @@ > > struct drm_i915_private; > > -void intel_tv_init(struct drm_i915_private *dev_priv); > > +void intel_tv_init(struct drm_i915_private *i915); > > #endif /* __INTEL_TV_H__ */ > > diff --git a/drivers/gpu/drm/i915/intel_uc.c b/drivers/gpu/drm/i915/intel_uc.c > > index a8e7f0ba7c3b..11e2dcf6b917 100644 > > --- a/drivers/gpu/drm/i915/intel_uc.c > > +++ b/drivers/gpu/drm/i915/intel_uc.c > > @@ -33,12 +33,12 @@ static void guc_free_load_err_log(struct intel_guc *guc); > > /* Reset GuC providing us with fresh state for both GuC and HuC. > > */ > > -static int __intel_uc_reset_hw(struct drm_i915_private *dev_priv) > > +static int __intel_uc_reset_hw(struct drm_i915_private *i915) > > { > > int ret; > > u32 guc_status; > > - ret = intel_reset_guc(dev_priv); > > + ret = intel_reset_guc(i915); > > if (ret) { > > DRM_ERROR("Failed to reset GuC, ret = %d\n", ret); > > return ret; > > diff --git a/drivers/gpu/drm/i915/intel_uc.h b/drivers/gpu/drm/i915/intel_uc.h > > index 3ea06c87dfcd..4ddd50d990b3 100644 > > --- a/drivers/gpu/drm/i915/intel_uc.h > > +++ b/drivers/gpu/drm/i915/intel_uc.h > > @@ -28,20 +28,20 @@ > > #include "intel_huc.h" > > #include "i915_params.h" > > -void intel_uc_init_early(struct drm_i915_private *dev_priv); > > -void intel_uc_cleanup_early(struct drm_i915_private *dev_priv); > > -void intel_uc_init_mmio(struct drm_i915_private *dev_priv); > > -int intel_uc_init_misc(struct drm_i915_private *dev_priv); > > -void intel_uc_fini_misc(struct drm_i915_private *dev_priv); > > -void intel_uc_sanitize(struct drm_i915_private *dev_priv); > > -int intel_uc_init_hw(struct drm_i915_private *dev_priv); > > -void intel_uc_fini_hw(struct drm_i915_private *dev_priv); > > -int intel_uc_init(struct drm_i915_private *dev_priv); > > -void intel_uc_fini(struct drm_i915_private *dev_priv); > > +void intel_uc_init_early(struct drm_i915_private *i915); > > +void intel_uc_cleanup_early(struct drm_i915_private *i915); > > +void intel_uc_init_mmio(struct drm_i915_private *i915); > > +int intel_uc_init_misc(struct drm_i915_private *i915); > > +void intel_uc_fini_misc(struct drm_i915_private *i915); > > +void intel_uc_sanitize(struct drm_i915_private *i915); > > +int intel_uc_init_hw(struct drm_i915_private *i915); > > +void intel_uc_fini_hw(struct drm_i915_private *i915); > > +int intel_uc_init(struct drm_i915_private *i915); > > +void intel_uc_fini(struct drm_i915_private *i915); > > void intel_uc_reset_prepare(struct drm_i915_private *i915); > > void intel_uc_suspend(struct drm_i915_private *i915); > > void intel_uc_runtime_suspend(struct drm_i915_private *i915); > > -int intel_uc_resume(struct drm_i915_private *dev_priv); > > +int intel_uc_resume(struct drm_i915_private *i915); > > static inline bool intel_uc_is_using_guc(struct drm_i915_private *i915) > > { > > diff --git a/drivers/gpu/drm/i915/intel_uc_fw.c b/drivers/gpu/drm/i915/intel_uc_fw.c > > index f342ddd47df8..d156911d1560 100644 > > --- a/drivers/gpu/drm/i915/intel_uc_fw.c > > +++ b/drivers/gpu/drm/i915/intel_uc_fw.c > > @@ -32,15 +32,15 @@ > > /** > > * intel_uc_fw_fetch - fetch uC firmware > > * > > - * @dev_priv: device private > > + * @i915: device private > > * @uc_fw: uC firmware > > * > > * Fetch uC firmware into GEM obj. > > */ > > -void intel_uc_fw_fetch(struct drm_i915_private *dev_priv, > > +void intel_uc_fw_fetch(struct drm_i915_private *i915, > > struct intel_uc_fw *uc_fw) > > { > > - struct pci_dev *pdev = dev_priv->drm.pdev; > > + struct pci_dev *pdev = i915->drm.pdev; > > struct drm_i915_gem_object *obj; > > const struct firmware *fw = NULL; > > struct uc_css_header *css; > > @@ -48,10 +48,10 @@ void intel_uc_fw_fetch(struct drm_i915_private *dev_priv, > > int err; > > if (!uc_fw->path) { > > - dev_info(dev_priv->drm.dev, > > + dev_info(i915->drm.dev, > > "%s: No firmware was defined for %s!\n", > > intel_uc_fw_type_repr(uc_fw->type), > > - intel_platform_name(INTEL_INFO(dev_priv)->platform)); > > + intel_platform_name(INTEL_INFO(i915)->platform)); > > return; > > } > > @@ -159,7 +159,7 @@ void intel_uc_fw_fetch(struct drm_i915_private *dev_priv, > > goto fail; > > } > > - obj = i915_gem_object_create_shmem_from_data(dev_priv, > > + obj = i915_gem_object_create_shmem_from_data(i915, > > fw->data, fw->size); > > if (IS_ERR(obj)) { > > err = PTR_ERR(obj); > > diff --git a/drivers/gpu/drm/i915/intel_uc_fw.h b/drivers/gpu/drm/i915/intel_uc_fw.h > > index ff98f8661d72..881c3aef074c 100644 > > --- a/drivers/gpu/drm/i915/intel_uc_fw.h > > +++ b/drivers/gpu/drm/i915/intel_uc_fw.h > > @@ -142,7 +142,7 @@ static inline u32 intel_uc_fw_get_upload_size(struct intel_uc_fw *uc_fw) > > return uc_fw->header_size + uc_fw->ucode_size; > > } > > -void intel_uc_fw_fetch(struct drm_i915_private *dev_priv, > > +void intel_uc_fw_fetch(struct drm_i915_private *i915, > > struct intel_uc_fw *uc_fw); > > void intel_uc_fw_cleanup_fetch(struct intel_uc_fw *uc_fw); > > int intel_uc_fw_upload(struct intel_uc_fw *uc_fw, > > diff --git a/drivers/gpu/drm/i915/intel_uncore.c b/drivers/gpu/drm/i915/intel_uncore.c > > index 85171a8b866a..87dec005165d 100644 > > --- a/drivers/gpu/drm/i915/intel_uncore.c > > +++ b/drivers/gpu/drm/i915/intel_uncore.c > > @@ -537,10 +537,10 @@ void intel_uncore_runtime_resume(struct intel_uncore *uncore) > > iosf_mbi_register_pmic_bus_access_notifier(&uncore->pmic_bus_access_nb); > > } > > -void intel_uncore_sanitize(struct drm_i915_private *dev_priv) > > +void intel_uncore_sanitize(struct drm_i915_private *i915) > > { > > /* BIOS often leaves RC6 enabled, but disable it for hw init */ > > - intel_sanitize_gt_powersave(dev_priv); > > + intel_sanitize_gt_powersave(i915); > > } > > static void __intel_uncore_forcewake_get(struct intel_uncore *uncore, > > @@ -647,7 +647,7 @@ void intel_uncore_forcewake_user_put(struct intel_uncore *uncore) > > * @fw_domains: forcewake domains to get reference on > > * > > * See intel_uncore_forcewake_get(). This variant places the onus > > - * on the caller to explicitly handle the dev_priv->uncore.lock spinlock. > > + * on the caller to explicitly handle the i915->uncore.lock spinlock. > > */ > > void intel_uncore_forcewake_get__locked(struct intel_uncore *uncore, > > enum forcewake_domains fw_domains) > > @@ -708,7 +708,7 @@ void intel_uncore_forcewake_put(struct intel_uncore *uncore, > > * @fw_domains: forcewake domains to get reference on > > * > > * See intel_uncore_forcewake_put(). This variant places the onus > > - * on the caller to explicitly handle the dev_priv->uncore.lock spinlock. > > + * on the caller to explicitly handle the i915->uncore.lock spinlock. > > */ > > void intel_uncore_forcewake_put__locked(struct intel_uncore *uncore, > > enum forcewake_domains fw_domains) > > @@ -821,10 +821,10 @@ find_fw_domain(struct intel_uncore *uncore, u32 offset) > > #define GEN_FW_RANGE(s, e, d) \ > > { .start = (s), .end = (e), .domains = (d) } > > -#define HAS_FWTABLE(dev_priv) \ > > - (INTEL_GEN(dev_priv) >= 9 || \ > > - IS_CHERRYVIEW(dev_priv) || \ > > - IS_VALLEYVIEW(dev_priv)) > > +#define HAS_FWTABLE(i915) \ > > + (INTEL_GEN(i915) >= 9 || \ > > + IS_CHERRYVIEW(i915) || \ > > + IS_VALLEYVIEW(i915)) > > /* *Must* be sorted by offset ranges! See intel_fw_table_check(). */ > > static const struct intel_forcewake_range __vlv_fw_ranges[] = { > > @@ -1461,7 +1461,7 @@ static void intel_uncore_fw_domains_init(struct intel_uncore *uncore) > > static int i915_pmic_bus_access_notifier(struct notifier_block *nb, > > unsigned long action, void *data) > > { > > - struct drm_i915_private *dev_priv = container_of(nb, > > + struct drm_i915_private *i915 = container_of(nb, > > struct drm_i915_private, uncore.pmic_bus_access_nb); > > switch (action) { > > @@ -1479,12 +1479,12 @@ static int i915_pmic_bus_access_notifier(struct notifier_block *nb, > > * wake reference -> disable wakeref asserts for the time of > > * the access. > > */ > > - disable_rpm_wakeref_asserts(dev_priv); > > - intel_uncore_forcewake_get(&dev_priv->uncore, FORCEWAKE_ALL); > > - enable_rpm_wakeref_asserts(dev_priv); > > + disable_rpm_wakeref_asserts(i915); > > + intel_uncore_forcewake_get(&i915->uncore, FORCEWAKE_ALL); > > + enable_rpm_wakeref_asserts(i915); > > break; > > case MBI_PMIC_BUS_ACCESS_END: > > - intel_uncore_forcewake_put(&dev_priv->uncore, FORCEWAKE_ALL); > > + intel_uncore_forcewake_put(&i915->uncore, FORCEWAKE_ALL); > > break; > > } > > diff --git a/drivers/gpu/drm/i915/intel_uncore.h b/drivers/gpu/drm/i915/intel_uncore.h > > index d6af3de70121..d30bda2c3444 100644 > > --- a/drivers/gpu/drm/i915/intel_uncore.h > > +++ b/drivers/gpu/drm/i915/intel_uncore.h > > @@ -177,7 +177,7 @@ intel_uncore_has_fifo(const struct intel_uncore *uncore) > > return uncore->flags & UNCORE_HAS_FIFO; > > } > > -void intel_uncore_sanitize(struct drm_i915_private *dev_priv); > > +void intel_uncore_sanitize(struct drm_i915_private *i915); > > void intel_uncore_init_early(struct intel_uncore *uncore); > > int intel_uncore_init_mmio(struct intel_uncore *uncore); > > void intel_uncore_prune_mmio_domains(struct intel_uncore *uncore); > > @@ -361,7 +361,7 @@ intel_uncore_read64_2x32(struct intel_uncore *uncore, > > * > > * Certain architectures will die if the same cacheline is concurrently accessed > > * by different clients (e.g. on Ivybridge). Access to registers should > > - * therefore generally be serialised, by either the dev_priv->uncore.lock or > > + * therefore generally be serialised, by either the i915->uncore.lock or > > * a more localised lock guarding all access to that bank of registers. > > */ > > #define intel_uncore_read_fw(...) __raw_uncore_read32(__VA_ARGS__) > > diff --git a/drivers/gpu/drm/i915/intel_vbt_defs.h b/drivers/gpu/drm/i915/intel_vbt_defs.h > > index 89ef14cafb6b..e2e9883d6b28 100644 > > --- a/drivers/gpu/drm/i915/intel_vbt_defs.h > > +++ b/drivers/gpu/drm/i915/intel_vbt_defs.h > > @@ -337,7 +337,7 @@ enum vbt_gmbus_ddi { > > * basically any of the fields to ensure the correct interpretation for the BDB > > * version in question. > > * > > - * When we copy the child device configs to dev_priv->vbt.child_dev, we reserve > > + * When we copy the child device configs to i915->vbt.child_dev, we reserve > > * space for the full structure below, and initialize the tail not actually > > * present in VBT to zeros. Accessing those fields is fine, as long as the > > * default zero is taken into account, again according to the BDB version. > > diff --git a/drivers/gpu/drm/i915/intel_vdsc.c b/drivers/gpu/drm/i915/intel_vdsc.c > > index ffec807b8960..164fd4b56112 100644 > > --- a/drivers/gpu/drm/i915/intel_vdsc.c > > +++ b/drivers/gpu/drm/i915/intel_vdsc.c > > @@ -478,7 +478,7 @@ static void intel_configure_pps_for_dsc_encoder(struct intel_encoder *encoder, > > const struct intel_crtc_state *crtc_state) > > { > > struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc); > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > const struct drm_dsc_config *vdsc_cfg = &crtc_state->dp_dsc_cfg; > > enum pipe pipe = crtc->pipe; > > enum transcoder cpu_transcoder = crtc_state->cpu_transcoder; > > @@ -897,7 +897,7 @@ void intel_dsc_enable(struct intel_encoder *encoder, > > const struct intel_crtc_state *crtc_state) > > { > > struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc); > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > enum pipe pipe = crtc->pipe; > > i915_reg_t dss_ctl1_reg, dss_ctl2_reg; > > u32 dss_ctl1_val = 0; > > @@ -907,7 +907,7 @@ void intel_dsc_enable(struct intel_encoder *encoder, > > return; > > /* Enable Power wells for VDSC/joining */ > > - intel_display_power_get(dev_priv, > > + intel_display_power_get(i915, > > intel_dsc_power_domain(crtc_state)); > > intel_configure_pps_for_dsc_encoder(encoder, crtc_state); > > @@ -933,7 +933,7 @@ void intel_dsc_enable(struct intel_encoder *encoder, > > void intel_dsc_disable(const struct intel_crtc_state *old_crtc_state) > > { > > struct intel_crtc *crtc = to_intel_crtc(old_crtc_state->base.crtc); > > - struct drm_i915_private *dev_priv = to_i915(crtc->base.dev); > > + struct drm_i915_private *i915 = to_i915(crtc->base.dev); > > enum pipe pipe = crtc->pipe; > > i915_reg_t dss_ctl1_reg, dss_ctl2_reg; > > u32 dss_ctl1_val = 0, dss_ctl2_val = 0; > > @@ -961,6 +961,6 @@ void intel_dsc_disable(const struct intel_crtc_state *old_crtc_state) > > I915_WRITE(dss_ctl2_reg, dss_ctl2_val); > > /* Disable Power wells for VDSC/joining */ > > - intel_display_power_put_unchecked(dev_priv, > > + intel_display_power_put_unchecked(i915, > > intel_dsc_power_domain(old_crtc_state)); > > } > > diff --git a/drivers/gpu/drm/i915/intel_wopcm.c b/drivers/gpu/drm/i915/intel_wopcm.c > > index 7b4ba84b9fb8..646540b97267 100644 > > --- a/drivers/gpu/drm/i915/intel_wopcm.c > > +++ b/drivers/gpu/drm/i915/intel_wopcm.c > > @@ -225,7 +225,7 @@ int intel_wopcm_init(struct intel_wopcm *wopcm) > > return 0; > > } > > -static inline int write_and_verify(struct drm_i915_private *dev_priv, > > +static inline int write_and_verify(struct drm_i915_private *i915, > > i915_reg_t reg, u32 val, u32 mask, > > u32 locked_bit) > > { > > @@ -252,27 +252,27 @@ static inline int write_and_verify(struct drm_i915_private *dev_priv, > > */ > > int intel_wopcm_init_hw(struct intel_wopcm *wopcm) > > { > > - struct drm_i915_private *dev_priv = wopcm_to_i915(wopcm); > > + struct drm_i915_private *i915 = wopcm_to_i915(wopcm); > > u32 huc_agent; > > u32 mask; > > int err; > > - if (!USES_GUC(dev_priv)) > > + if (!USES_GUC(i915)) > > return 0; > > - GEM_BUG_ON(!HAS_GUC(dev_priv)); > > + GEM_BUG_ON(!HAS_GUC(i915)); > > GEM_BUG_ON(!wopcm->guc.size); > > GEM_BUG_ON(!wopcm->guc.base); > > - err = write_and_verify(dev_priv, GUC_WOPCM_SIZE, wopcm->guc.size, > > + err = write_and_verify(i915, GUC_WOPCM_SIZE, wopcm->guc.size, > > GUC_WOPCM_SIZE_MASK | GUC_WOPCM_SIZE_LOCKED, > > GUC_WOPCM_SIZE_LOCKED); > > if (err) > > goto err_out; > > - huc_agent = USES_HUC(dev_priv) ? HUC_LOADING_AGENT_GUC : 0; > > + huc_agent = USES_HUC(i915) ? HUC_LOADING_AGENT_GUC : 0; > > mask = GUC_WOPCM_OFFSET_MASK | GUC_WOPCM_OFFSET_VALID | huc_agent; > > - err = write_and_verify(dev_priv, DMA_GUC_WOPCM_OFFSET, > > + err = write_and_verify(i915, DMA_GUC_WOPCM_OFFSET, > > wopcm->guc.base | huc_agent, mask, > > GUC_WOPCM_OFFSET_VALID); > > if (err) > > diff --git a/drivers/gpu/drm/i915/selftests/i915_gem_gtt.c b/drivers/gpu/drm/i915/selftests/i915_gem_gtt.c > > index 2093d08a7569..2514842236f6 100644 > > --- a/drivers/gpu/drm/i915/selftests/i915_gem_gtt.c > > +++ b/drivers/gpu/drm/i915/selftests/i915_gem_gtt.c > > @@ -147,17 +147,17 @@ fake_dma_object(struct drm_i915_private *i915, u64 size) > > static int igt_ppgtt_alloc(void *arg) > > { > > - struct drm_i915_private *dev_priv = arg; > > + struct drm_i915_private *i915 = arg; > > struct i915_ppgtt *ppgtt; > > u64 size, last, limit; > > int err = 0; > > /* Allocate a ppggt and try to fill the entire range */ > > - if (!HAS_PPGTT(dev_priv)) > > + if (!HAS_PPGTT(i915)) > > return 0; > > - ppgtt = __ppgtt_create(dev_priv); > > + ppgtt = __ppgtt_create(i915); > > if (IS_ERR(ppgtt)) > > return PTR_ERR(ppgtt); > > @@ -208,9 +208,9 @@ static int igt_ppgtt_alloc(void *arg) > > } > > err_ppgtt_cleanup: > > - mutex_lock(&dev_priv->drm.struct_mutex); > > + mutex_lock(&i915->drm.struct_mutex); > > i915_vm_put(&ppgtt->vm); > > - mutex_unlock(&dev_priv->drm.struct_mutex); > > + mutex_unlock(&i915->drm.struct_mutex); > > return err; > > } > > @@ -992,7 +992,7 @@ static int shrink_boom(struct drm_i915_private *i915, > > return err; > > } > > -static int exercise_ppgtt(struct drm_i915_private *dev_priv, > > +static int exercise_ppgtt(struct drm_i915_private *i915, > > int (*func)(struct drm_i915_private *i915, > > struct i915_address_space *vm, > > u64 hole_start, u64 hole_end, > > @@ -1003,15 +1003,15 @@ static int exercise_ppgtt(struct drm_i915_private *dev_priv, > > IGT_TIMEOUT(end_time); > > int err; > > - if (!HAS_FULL_PPGTT(dev_priv)) > > + if (!HAS_FULL_PPGTT(i915)) > > return 0; > > - file = mock_file(dev_priv); > > + file = mock_file(i915); > > if (IS_ERR(file)) > > return PTR_ERR(file); > > - mutex_lock(&dev_priv->drm.struct_mutex); > > - ppgtt = i915_ppgtt_create(dev_priv); > > + mutex_lock(&i915->drm.struct_mutex); > > + ppgtt = i915_ppgtt_create(i915); > > if (IS_ERR(ppgtt)) { > > err = PTR_ERR(ppgtt); > > goto out_unlock; > > @@ -1019,13 +1019,13 @@ static int exercise_ppgtt(struct drm_i915_private *dev_priv, > > GEM_BUG_ON(offset_in_page(ppgtt->vm.total)); > > GEM_BUG_ON(ppgtt->vm.closed); > > - err = func(dev_priv, &ppgtt->vm, 0, ppgtt->vm.total, end_time); > > + err = func(i915, &ppgtt->vm, 0, ppgtt->vm.total, end_time); > > i915_vm_put(&ppgtt->vm); > > out_unlock: > > - mutex_unlock(&dev_priv->drm.struct_mutex); > > + mutex_unlock(&i915->drm.struct_mutex); > > - mock_file_free(dev_priv, file); > > + mock_file_free(i915, file); > > return err; > > } > > diff --git a/drivers/gpu/drm/i915/selftests/intel_guc.c b/drivers/gpu/drm/i915/selftests/intel_guc.c > > index 7fd0321e0947..ab3e71c957df 100644 > > --- a/drivers/gpu/drm/i915/selftests/intel_guc.c > > +++ b/drivers/gpu/drm/i915/selftests/intel_guc.c > > @@ -107,12 +107,12 @@ static int validate_client(struct intel_guc_client *client, > > int client_priority, > > bool is_preempt_client) > > { > > - struct drm_i915_private *dev_priv = guc_to_i915(client->guc); > > + struct drm_i915_private *i915 = guc_to_i915(client->guc); > > struct i915_gem_context *ctx_owner = is_preempt_client ? > > - dev_priv->preempt_context : dev_priv->kernel_context; > > + i915->preempt_context : i915->kernel_context; > > if (client->owner != ctx_owner || > > - client->engines != INTEL_INFO(dev_priv)->engine_mask || > > + client->engines != INTEL_INFO(i915)->engine_mask || > > client->priority != client_priority || > > client->doorbell_id == GUC_DOORBELL_INVALID) > > return -EINVAL; > > @@ -137,16 +137,16 @@ static bool client_doorbell_in_sync(struct intel_guc_client *client) > > */ > > static int igt_guc_clients(void *args) > > { > > - struct drm_i915_private *dev_priv = args; > > + struct drm_i915_private *i915 = args; > > intel_wakeref_t wakeref; > > struct intel_guc *guc; > > int err = 0; > > - GEM_BUG_ON(!HAS_GUC(dev_priv)); > > - mutex_lock(&dev_priv->drm.struct_mutex); > > - wakeref = intel_runtime_pm_get(dev_priv); > > + GEM_BUG_ON(!HAS_GUC(i915)); > > + mutex_lock(&i915->drm.struct_mutex); > > + wakeref = intel_runtime_pm_get(i915); > > - guc = &dev_priv->guc; > > + guc = &i915->guc; > > if (!guc) { > > pr_err("No guc object!\n"); > > err = -EINVAL; > > @@ -227,8 +227,8 @@ static int igt_guc_clients(void *args) > > guc_clients_create(guc); > > guc_clients_enable(guc); > > unlock: > > - intel_runtime_pm_put(dev_priv, wakeref); > > - mutex_unlock(&dev_priv->drm.struct_mutex); > > + intel_runtime_pm_put(i915, wakeref); > > + mutex_unlock(&i915->drm.struct_mutex); > > return err; > > } > > @@ -239,17 +239,17 @@ static int igt_guc_clients(void *args) > > */ > > static int igt_guc_doorbells(void *arg) > > { > > - struct drm_i915_private *dev_priv = arg; > > + struct drm_i915_private *i915 = arg; > > intel_wakeref_t wakeref; > > struct intel_guc *guc; > > int i, err = 0; > > u16 db_id; > > - GEM_BUG_ON(!HAS_GUC(dev_priv)); > > - mutex_lock(&dev_priv->drm.struct_mutex); > > - wakeref = intel_runtime_pm_get(dev_priv); > > + GEM_BUG_ON(!HAS_GUC(i915)); > > + mutex_lock(&i915->drm.struct_mutex); > > + wakeref = intel_runtime_pm_get(i915); > > - guc = &dev_priv->guc; > > + guc = &i915->guc; > > if (!guc) { > > pr_err("No guc object!\n"); > > err = -EINVAL; > > @@ -261,10 +261,10 @@ static int igt_guc_doorbells(void *arg) > > goto unlock; > > for (i = 0; i < ATTEMPTS; i++) { > > - clients[i] = guc_client_alloc(dev_priv, > > - INTEL_INFO(dev_priv)->engine_mask, > > + clients[i] = guc_client_alloc(i915, > > + INTEL_INFO(i915)->engine_mask, > > i % GUC_CLIENT_PRIORITY_NUM, > > - dev_priv->kernel_context); > > + i915->kernel_context); > > if (!clients[i]) { > > pr_err("[%d] No guc client\n", i); > > @@ -340,20 +340,20 @@ static int igt_guc_doorbells(void *arg) > > guc_client_free(clients[i]); > > } > > unlock: > > - intel_runtime_pm_put(dev_priv, wakeref); > > - mutex_unlock(&dev_priv->drm.struct_mutex); > > + intel_runtime_pm_put(i915, wakeref); > > + mutex_unlock(&i915->drm.struct_mutex); > > return err; > > } > > -int intel_guc_live_selftest(struct drm_i915_private *dev_priv) > > +int intel_guc_live_selftest(struct drm_i915_private *i915) > > { > > static const struct i915_subtest tests[] = { > > SUBTEST(igt_guc_clients), > > SUBTEST(igt_guc_doorbells), > > }; > > - if (!USES_GUC_SUBMISSION(dev_priv)) > > + if (!USES_GUC_SUBMISSION(i915)) > > return 0; > > - return i915_subtests(tests, dev_priv); > > + return i915_subtests(tests, i915); > > } > > diff --git a/drivers/gpu/drm/i915/selftests/intel_uncore.c b/drivers/gpu/drm/i915/selftests/intel_uncore.c > > index e0d7ebecb215..b5304cfeda21 100644 > > --- a/drivers/gpu/drm/i915/selftests/intel_uncore.c > > +++ b/drivers/gpu/drm/i915/selftests/intel_uncore.c > > @@ -254,15 +254,15 @@ static int live_forcewake_ops(void *arg) > > static int live_forcewake_domains(void *arg) > > { > > #define FW_RANGE 0x40000 > > - struct drm_i915_private *dev_priv = arg; > > - struct intel_uncore *uncore = &dev_priv->uncore; > > + struct drm_i915_private *i915 = arg; > > + struct intel_uncore *uncore = &i915->uncore; > > unsigned long *valid; > > u32 offset; > > int err; > > - if (!HAS_FPGA_DBG_UNCLAIMED(dev_priv) && > > - !IS_VALLEYVIEW(dev_priv) && > > - !IS_CHERRYVIEW(dev_priv)) > > + if (!HAS_FPGA_DBG_UNCLAIMED(i915) && > > + !IS_VALLEYVIEW(i915) && > > + !IS_CHERRYVIEW(i915)) > > return 0; > > /* > > diff --git a/drivers/gpu/drm/i915/vlv_dsi.c b/drivers/gpu/drm/i915/vlv_dsi.c > > index e272d826210a..0178bdf0c5a9 100644 > > --- a/drivers/gpu/drm/i915/vlv_dsi.c > > +++ b/drivers/gpu/drm/i915/vlv_dsi.c > > @@ -78,19 +78,19 @@ void vlv_dsi_wait_for_fifo_empty(struct intel_dsi *intel_dsi, enum port port) > > { > > struct drm_encoder *encoder = &intel_dsi->base.base; > > struct drm_device *dev = encoder->dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > u32 mask; > > mask = LP_CTRL_FIFO_EMPTY | HS_CTRL_FIFO_EMPTY | > > LP_DATA_FIFO_EMPTY | HS_DATA_FIFO_EMPTY; > > - if (intel_wait_for_register(&dev_priv->uncore, > > + if (intel_wait_for_register(&i915->uncore, > > MIPI_GEN_FIFO_STAT(port), mask, mask, > > 100)) > > DRM_ERROR("DPI FIFOs are not empty\n"); > > } > > -static void write_data(struct drm_i915_private *dev_priv, > > +static void write_data(struct drm_i915_private *i915, > > i915_reg_t reg, > > const u8 *data, u32 len) > > { > > @@ -106,7 +106,7 @@ static void write_data(struct drm_i915_private *dev_priv, > > } > > } > > -static void read_data(struct drm_i915_private *dev_priv, > > +static void read_data(struct drm_i915_private *i915, > > i915_reg_t reg, > > u8 *data, u32 len) > > { > > @@ -125,7 +125,7 @@ static ssize_t intel_dsi_host_transfer(struct mipi_dsi_host *host, > > { > > struct intel_dsi_host *intel_dsi_host = to_intel_dsi_host(host); > > struct drm_device *dev = intel_dsi_host->intel_dsi->base.base.dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > enum port port = intel_dsi_host->port; > > struct mipi_dsi_packet packet; > > ssize_t ret; > > @@ -154,13 +154,13 @@ static ssize_t intel_dsi_host_transfer(struct mipi_dsi_host *host, > > /* note: this is never true for reads */ > > if (packet.payload_length) { > > - if (intel_wait_for_register(&dev_priv->uncore, > > + if (intel_wait_for_register(&i915->uncore, > > MIPI_GEN_FIFO_STAT(port), > > data_mask, 0, > > 50)) > > DRM_ERROR("Timeout waiting for HS/LP DATA FIFO !full\n"); > > - write_data(dev_priv, data_reg, packet.payload, > > + write_data(i915, data_reg, packet.payload, > > packet.payload_length); > > } > > @@ -168,7 +168,7 @@ static ssize_t intel_dsi_host_transfer(struct mipi_dsi_host *host, > > I915_WRITE(MIPI_INTR_STAT(port), GEN_READ_DATA_AVAIL); > > } > > - if (intel_wait_for_register(&dev_priv->uncore, > > + if (intel_wait_for_register(&i915->uncore, > > MIPI_GEN_FIFO_STAT(port), > > ctrl_mask, 0, > > 50)) { > > @@ -180,13 +180,13 @@ static ssize_t intel_dsi_host_transfer(struct mipi_dsi_host *host, > > /* ->rx_len is set only for reads */ > > if (msg->rx_len) { > > data_mask = GEN_READ_DATA_AVAIL; > > - if (intel_wait_for_register(&dev_priv->uncore, > > + if (intel_wait_for_register(&i915->uncore, > > MIPI_INTR_STAT(port), > > data_mask, data_mask, > > 50)) > > DRM_ERROR("Timeout waiting for read data.\n"); > > - read_data(dev_priv, data_reg, msg->rx_buf, msg->rx_len); > > + read_data(i915, data_reg, msg->rx_buf, msg->rx_len); > > } > > /* XXX: fix for reads and writes */ > > @@ -221,7 +221,7 @@ static int dpi_send_cmd(struct intel_dsi *intel_dsi, u32 cmd, bool hs, > > { > > struct drm_encoder *encoder = &intel_dsi->base.base; > > struct drm_device *dev = encoder->dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > u32 mask; > > /* XXX: pipe, hs */ > > @@ -240,7 +240,7 @@ static int dpi_send_cmd(struct intel_dsi *intel_dsi, u32 cmd, bool hs, > > I915_WRITE(MIPI_DPI_CONTROL(port), cmd); > > mask = SPL_PKT_SENT_INTERRUPT; > > - if (intel_wait_for_register(&dev_priv->uncore, > > + if (intel_wait_for_register(&i915->uncore, > > MIPI_INTR_STAT(port), mask, mask, > > 100)) > > DRM_ERROR("Video mode command 0x%08x send failed.\n", cmd); > > @@ -248,25 +248,25 @@ static int dpi_send_cmd(struct intel_dsi *intel_dsi, u32 cmd, bool hs, > > return 0; > > } > > -static void band_gap_reset(struct drm_i915_private *dev_priv) > > +static void band_gap_reset(struct drm_i915_private *i915) > > { > > - vlv_flisdsi_get(dev_priv); > > + vlv_flisdsi_get(i915); > > - vlv_flisdsi_write(dev_priv, 0x08, 0x0001); > > - vlv_flisdsi_write(dev_priv, 0x0F, 0x0005); > > - vlv_flisdsi_write(dev_priv, 0x0F, 0x0025); > > + vlv_flisdsi_write(i915, 0x08, 0x0001); > > + vlv_flisdsi_write(i915, 0x0F, 0x0005); > > + vlv_flisdsi_write(i915, 0x0F, 0x0025); > > udelay(150); > > - vlv_flisdsi_write(dev_priv, 0x0F, 0x0000); > > - vlv_flisdsi_write(dev_priv, 0x08, 0x0000); > > + vlv_flisdsi_write(i915, 0x0F, 0x0000); > > + vlv_flisdsi_write(i915, 0x08, 0x0000); > > - vlv_flisdsi_put(dev_priv); > > + vlv_flisdsi_put(i915); > > } > > static int intel_dsi_compute_config(struct intel_encoder *encoder, > > struct intel_crtc_state *pipe_config, > > struct drm_connector_state *conn_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_dsi *intel_dsi = container_of(encoder, struct intel_dsi, > > base); > > struct intel_connector *intel_connector = intel_dsi->attached_connector; > > @@ -281,7 +281,7 @@ static int intel_dsi_compute_config(struct intel_encoder *encoder, > > if (fixed_mode) { > > intel_fixed_panel_mode(fixed_mode, adjusted_mode); > > - if (HAS_GMCH(dev_priv)) > > + if (HAS_GMCH(i915)) > > intel_gmch_panel_fitting(crtc, pipe_config, > > conn_state->scaling_mode); > > else > > @@ -300,7 +300,7 @@ static int intel_dsi_compute_config(struct intel_encoder *encoder, > > else > > pipe_config->pipe_bpp = 18; > > - if (IS_GEN9_LP(dev_priv)) { > > + if (IS_GEN9_LP(i915)) { > > /* Enable Frame time stamp based scanline reporting */ > > adjusted_mode->private_flags |= > > I915_MODE_FLAG_GET_SCANLINE_FROM_TIMESTAMP; > > @@ -327,7 +327,7 @@ static int intel_dsi_compute_config(struct intel_encoder *encoder, > > static bool glk_dsi_enable_io(struct intel_encoder *encoder) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base); > > enum port port; > > u32 tmp; > > @@ -359,7 +359,7 @@ static bool glk_dsi_enable_io(struct intel_encoder *encoder) > > /* Wait for Pwr ACK */ > > for_each_dsi_port(port, intel_dsi->ports) { > > - if (intel_wait_for_register(&dev_priv->uncore, > > + if (intel_wait_for_register(&i915->uncore, > > MIPI_CTRL(port), > > GLK_MIPIIO_PORT_POWERED, > > GLK_MIPIIO_PORT_POWERED, > > @@ -378,14 +378,14 @@ static bool glk_dsi_enable_io(struct intel_encoder *encoder) > > static void glk_dsi_device_ready(struct intel_encoder *encoder) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base); > > enum port port; > > u32 val; > > /* Wait for MIPI PHY status bit to set */ > > for_each_dsi_port(port, intel_dsi->ports) { > > - if (intel_wait_for_register(&dev_priv->uncore, > > + if (intel_wait_for_register(&i915->uncore, > > MIPI_CTRL(port), > > GLK_PHY_STATUS_PORT_READY, > > GLK_PHY_STATUS_PORT_READY, > > @@ -413,7 +413,7 @@ static void glk_dsi_device_ready(struct intel_encoder *encoder) > > I915_WRITE(MIPI_DEVICE_READY(port), val); > > /* Wait for ULPS active */ > > - if (intel_wait_for_register(&dev_priv->uncore, > > + if (intel_wait_for_register(&i915->uncore, > > MIPI_CTRL(port), > > GLK_ULPS_NOT_ACTIVE, > > 0, > > @@ -440,7 +440,7 @@ static void glk_dsi_device_ready(struct intel_encoder *encoder) > > /* Wait for Stop state */ > > for_each_dsi_port(port, intel_dsi->ports) { > > - if (intel_wait_for_register(&dev_priv->uncore, > > + if (intel_wait_for_register(&i915->uncore, > > MIPI_CTRL(port), > > GLK_DATA_LANE_STOP_STATE, > > GLK_DATA_LANE_STOP_STATE, > > @@ -450,7 +450,7 @@ static void glk_dsi_device_ready(struct intel_encoder *encoder) > > /* Wait for AFE LATCH */ > > for_each_dsi_port(port, intel_dsi->ports) { > > - if (intel_wait_for_register(&dev_priv->uncore, > > + if (intel_wait_for_register(&i915->uncore, > > BXT_MIPI_PORT_CTRL(port), > > AFE_LATCHOUT, > > AFE_LATCHOUT, > > @@ -461,7 +461,7 @@ static void glk_dsi_device_ready(struct intel_encoder *encoder) > > static void bxt_dsi_device_ready(struct intel_encoder *encoder) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base); > > enum port port; > > u32 val; > > @@ -488,21 +488,21 @@ static void bxt_dsi_device_ready(struct intel_encoder *encoder) > > static void vlv_dsi_device_ready(struct intel_encoder *encoder) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base); > > enum port port; > > u32 val; > > DRM_DEBUG_KMS("\n"); > > - vlv_flisdsi_get(dev_priv); > > + vlv_flisdsi_get(i915); > > /* program rcomp for compliance, reduce from 50 ohms to 45 ohms > > * needed everytime after power gate */ > > - vlv_flisdsi_write(dev_priv, 0x04, 0x0004); > > - vlv_flisdsi_put(dev_priv); > > + vlv_flisdsi_write(i915, 0x04, 0x0004); > > + vlv_flisdsi_put(i915); > > /* bandgap reset is needed after everytime we do power gate */ > > - band_gap_reset(dev_priv); > > + band_gap_reset(i915); > > for_each_dsi_port(port, intel_dsi->ports) { > > @@ -527,11 +527,11 @@ static void vlv_dsi_device_ready(struct intel_encoder *encoder) > > static void intel_dsi_device_ready(struct intel_encoder *encoder) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > - if (IS_GEMINILAKE(dev_priv)) > > + if (IS_GEMINILAKE(i915)) > > glk_dsi_device_ready(encoder); > > - else if (IS_GEN9_LP(dev_priv)) > > + else if (IS_GEN9_LP(i915)) > > bxt_dsi_device_ready(encoder); > > else > > vlv_dsi_device_ready(encoder); > > @@ -539,7 +539,7 @@ static void intel_dsi_device_ready(struct intel_encoder *encoder) > > static void glk_dsi_enter_low_power_mode(struct intel_encoder *encoder) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base); > > enum port port; > > u32 val; > > @@ -554,7 +554,7 @@ static void glk_dsi_enter_low_power_mode(struct intel_encoder *encoder) > > /* Wait for MIPI PHY status bit to unset */ > > for_each_dsi_port(port, intel_dsi->ports) { > > - if (intel_wait_for_register(&dev_priv->uncore, > > + if (intel_wait_for_register(&i915->uncore, > > MIPI_CTRL(port), > > GLK_PHY_STATUS_PORT_READY, 0, 20)) > > DRM_ERROR("PHY is not turning OFF\n"); > > @@ -562,7 +562,7 @@ static void glk_dsi_enter_low_power_mode(struct intel_encoder *encoder) > > /* Wait for Pwr ACK bit to unset */ > > for_each_dsi_port(port, intel_dsi->ports) { > > - if (intel_wait_for_register(&dev_priv->uncore, > > + if (intel_wait_for_register(&i915->uncore, > > MIPI_CTRL(port), > > GLK_MIPIIO_PORT_POWERED, 0, 20)) > > DRM_ERROR("MIPI IO Port is not powergated\n"); > > @@ -571,7 +571,7 @@ static void glk_dsi_enter_low_power_mode(struct intel_encoder *encoder) > > static void glk_dsi_disable_mipi_io(struct intel_encoder *encoder) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base); > > enum port port; > > u32 tmp; > > @@ -583,7 +583,7 @@ static void glk_dsi_disable_mipi_io(struct intel_encoder *encoder) > > /* Wait for MIPI PHY status bit to unset */ > > for_each_dsi_port(port, intel_dsi->ports) { > > - if (intel_wait_for_register(&dev_priv->uncore, > > + if (intel_wait_for_register(&i915->uncore, > > MIPI_CTRL(port), > > GLK_PHY_STATUS_PORT_READY, 0, 20)) > > DRM_ERROR("PHY is not turning OFF\n"); > > @@ -605,14 +605,14 @@ static void glk_dsi_clear_device_ready(struct intel_encoder *encoder) > > static void vlv_dsi_clear_device_ready(struct intel_encoder *encoder) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base); > > enum port port; > > DRM_DEBUG_KMS("\n"); > > for_each_dsi_port(port, intel_dsi->ports) { > > /* Common bit for both MIPI Port A & MIPI Port C on VLV/CHV */ > > - i915_reg_t port_ctrl = IS_GEN9_LP(dev_priv) ? > > + i915_reg_t port_ctrl = IS_GEN9_LP(i915) ? > > BXT_MIPI_PORT_CTRL(port) : MIPI_PORT_CTRL(PORT_A); > > u32 val; > > @@ -632,8 +632,8 @@ static void vlv_dsi_clear_device_ready(struct intel_encoder *encoder) > > * On VLV/CHV, wait till Clock lanes are in LP-00 state for MIPI > > * Port A only. MIPI Port C has no similar bit for checking. > > */ > > - if ((IS_GEN9_LP(dev_priv) || port == PORT_A) && > > - intel_wait_for_register(&dev_priv->uncore, > > + if ((IS_GEN9_LP(i915) || port == PORT_A) && > > + intel_wait_for_register(&i915->uncore, > > port_ctrl, AFE_LATCHOUT, 0, > > 30)) > > DRM_ERROR("DSI LP not going Low\n"); > > @@ -651,14 +651,14 @@ static void vlv_dsi_clear_device_ready(struct intel_encoder *encoder) > > static void intel_dsi_port_enable(struct intel_encoder *encoder, > > const struct intel_crtc_state *crtc_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_crtc *crtc = to_intel_crtc(crtc_state->base.crtc); > > struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base); > > enum port port; > > if (intel_dsi->dual_link == DSI_DUAL_LINK_FRONT_BACK) { > > u32 temp; > > - if (IS_GEN9_LP(dev_priv)) { > > + if (IS_GEN9_LP(i915)) { > > for_each_dsi_port(port, intel_dsi->ports) { > > temp = I915_READ(MIPI_CTRL(port)); > > temp &= ~BXT_PIXEL_OVERLAP_CNT_MASK | > > @@ -676,7 +676,7 @@ static void intel_dsi_port_enable(struct intel_encoder *encoder, > > } > > for_each_dsi_port(port, intel_dsi->ports) { > > - i915_reg_t port_ctrl = IS_GEN9_LP(dev_priv) ? > > + i915_reg_t port_ctrl = IS_GEN9_LP(i915) ? > > BXT_MIPI_PORT_CTRL(port) : MIPI_PORT_CTRL(port); > > u32 temp; > > @@ -688,7 +688,7 @@ static void intel_dsi_port_enable(struct intel_encoder *encoder, > > if (intel_dsi->ports == (BIT(PORT_A) | BIT(PORT_C))) { > > temp |= (intel_dsi->dual_link - 1) > > << DUAL_LINK_MODE_SHIFT; > > - if (IS_BROXTON(dev_priv)) > > + if (IS_BROXTON(i915)) > > temp |= LANE_CONFIGURATION_DUAL_LINK_A; > > else > > temp |= crtc->pipe ? > > @@ -708,12 +708,12 @@ static void intel_dsi_port_enable(struct intel_encoder *encoder, > > static void intel_dsi_port_disable(struct intel_encoder *encoder) > > { > > struct drm_device *dev = encoder->base.dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base); > > enum port port; > > for_each_dsi_port(port, intel_dsi->ports) { > > - i915_reg_t port_ctrl = IS_GEN9_LP(dev_priv) ? > > + i915_reg_t port_ctrl = IS_GEN9_LP(i915) ? > > BXT_MIPI_PORT_CTRL(port) : MIPI_PORT_CTRL(port); > > u32 temp; > > @@ -775,7 +775,7 @@ static void intel_dsi_pre_enable(struct intel_encoder *encoder, > > { > > struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base); > > struct drm_crtc *crtc = pipe_config->base.crtc; > > - struct drm_i915_private *dev_priv = to_i915(crtc->dev); > > + struct drm_i915_private *i915 = to_i915(crtc->dev); > > struct intel_crtc *intel_crtc = to_intel_crtc(crtc); > > int pipe = intel_crtc->pipe; > > enum port port; > > @@ -784,13 +784,13 @@ static void intel_dsi_pre_enable(struct intel_encoder *encoder, > > DRM_DEBUG_KMS("\n"); > > - intel_set_cpu_fifo_underrun_reporting(dev_priv, pipe, true); > > + intel_set_cpu_fifo_underrun_reporting(i915, pipe, true); > > /* > > * The BIOS may leave the PLL in a wonky state where it doesn't > > * lock. It needs to be fully powered down to fix it. > > */ > > - if (IS_GEN9_LP(dev_priv)) { > > + if (IS_GEN9_LP(i915)) { > > bxt_dsi_pll_disable(encoder); > > bxt_dsi_pll_enable(encoder, pipe_config); > > } else { > > @@ -798,7 +798,7 @@ static void intel_dsi_pre_enable(struct intel_encoder *encoder, > > vlv_dsi_pll_enable(encoder, pipe_config); > > } > > - if (IS_BROXTON(dev_priv)) { > > + if (IS_BROXTON(i915)) { > > /* Add MIPI IO reset programming for modeset */ > > val = I915_READ(BXT_P_CR_GT_DISP_PWRON); > > I915_WRITE(BXT_P_CR_GT_DISP_PWRON, > > @@ -809,7 +809,7 @@ static void intel_dsi_pre_enable(struct intel_encoder *encoder, > > I915_WRITE(BXT_P_DSI_REGULATOR_TX_CTRL, 0); > > } > > - if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) { > > + if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) { > > u32 val; > > /* Disable DPOunit clock gating, can stall pipe */ > > @@ -818,7 +818,7 @@ static void intel_dsi_pre_enable(struct intel_encoder *encoder, > > I915_WRITE(DSPCLK_GATE_D, val); > > } > > - if (!IS_GEMINILAKE(dev_priv)) > > + if (!IS_GEMINILAKE(i915)) > > intel_dsi_prepare(encoder, pipe_config); > > /* Power on, try both CRC pmic gpio and VBT */ > > @@ -830,7 +830,7 @@ static void intel_dsi_pre_enable(struct intel_encoder *encoder, > > /* Deassert reset */ > > intel_dsi_vbt_exec_sequence(intel_dsi, MIPI_SEQ_DEASSERT_RESET); > > - if (IS_GEMINILAKE(dev_priv)) { > > + if (IS_GEMINILAKE(i915)) { > > glk_cold_boot = glk_dsi_enable_io(encoder); > > /* Prepare port in cold boot(s3/s4) scenario */ > > @@ -842,7 +842,7 @@ static void intel_dsi_pre_enable(struct intel_encoder *encoder, > > intel_dsi_device_ready(encoder); > > /* Prepare port in normal boot scenario */ > > - if (IS_GEMINILAKE(dev_priv) && !glk_cold_boot) > > + if (IS_GEMINILAKE(i915) && !glk_cold_boot) > > intel_dsi_prepare(encoder, pipe_config); > > /* Send initialization commands in LP mode */ > > @@ -901,9 +901,9 @@ static void intel_dsi_disable(struct intel_encoder *encoder, > > static void intel_dsi_clear_device_ready(struct intel_encoder *encoder) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > - if (IS_GEMINILAKE(dev_priv)) > > + if (IS_GEMINILAKE(i915)) > > glk_dsi_clear_device_ready(encoder); > > else > > vlv_dsi_clear_device_ready(encoder); > > @@ -913,7 +913,7 @@ static void intel_dsi_post_disable(struct intel_encoder *encoder, > > const struct intel_crtc_state *pipe_config, > > const struct drm_connector_state *conn_state) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base); > > enum port port; > > u32 val; > > @@ -941,7 +941,7 @@ static void intel_dsi_post_disable(struct intel_encoder *encoder, > > /* Transition to LP-00 */ > > intel_dsi_clear_device_ready(encoder); > > - if (IS_BROXTON(dev_priv)) { > > + if (IS_BROXTON(i915)) { > > /* Power down DSI regulator to save power */ > > I915_WRITE(BXT_P_DSI_REGULATOR_CFG, STAP_SELECT); > > I915_WRITE(BXT_P_DSI_REGULATOR_TX_CTRL, HS_IO_CTRL_SELECT); > > @@ -952,7 +952,7 @@ static void intel_dsi_post_disable(struct intel_encoder *encoder, > > val & ~MIPIO_RST_CTRL); > > } > > - if (IS_GEN9_LP(dev_priv)) { > > + if (IS_GEN9_LP(i915)) { > > bxt_dsi_pll_disable(encoder); > > } else { > > u32 val; > > @@ -983,7 +983,7 @@ static void intel_dsi_post_disable(struct intel_encoder *encoder, > > static bool intel_dsi_get_hw_state(struct intel_encoder *encoder, > > enum pipe *pipe) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base); > > intel_wakeref_t wakeref; > > enum port port; > > @@ -991,7 +991,7 @@ static bool intel_dsi_get_hw_state(struct intel_encoder *encoder, > > DRM_DEBUG_KMS("\n"); > > - wakeref = intel_display_power_get_if_enabled(dev_priv, > > + wakeref = intel_display_power_get_if_enabled(i915, > > encoder->power_domain); > > if (!wakeref) > > return false; > > @@ -1001,12 +1001,12 @@ static bool intel_dsi_get_hw_state(struct intel_encoder *encoder, > > * configuration, otherwise accessing DSI registers will hang the > > * machine. See BSpec North Display Engine registers/MIPI[BXT]. > > */ > > - if (IS_GEN9_LP(dev_priv) && !bxt_dsi_pll_is_enabled(dev_priv)) > > + if (IS_GEN9_LP(i915) && !bxt_dsi_pll_is_enabled(i915)) > > goto out_put_power; > > /* XXX: this only works for one DSI output */ > > for_each_dsi_port(port, intel_dsi->ports) { > > - i915_reg_t ctrl_reg = IS_GEN9_LP(dev_priv) ? > > + i915_reg_t ctrl_reg = IS_GEN9_LP(i915) ? > > BXT_MIPI_PORT_CTRL(port) : MIPI_PORT_CTRL(port); > > bool enabled = I915_READ(ctrl_reg) & DPI_ENABLE; > > @@ -1015,7 +1015,7 @@ static bool intel_dsi_get_hw_state(struct intel_encoder *encoder, > > * bit in port C control register does not get set. As a > > * workaround, check pipe B conf instead. > > */ > > - if ((IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) && > > + if ((IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) && > > port == PORT_C) > > enabled = I915_READ(PIPECONF(PIPE_B)) & PIPECONF_ENABLE; > > @@ -1031,7 +1031,7 @@ static bool intel_dsi_get_hw_state(struct intel_encoder *encoder, > > if (!(I915_READ(MIPI_DEVICE_READY(port)) & DEVICE_READY)) > > continue; > > - if (IS_GEN9_LP(dev_priv)) { > > + if (IS_GEN9_LP(i915)) { > > u32 tmp = I915_READ(MIPI_CTRL(port)); > > tmp &= BXT_PIPE_SELECT_MASK; > > tmp >>= BXT_PIPE_SELECT_SHIFT; > > @@ -1049,7 +1049,7 @@ static bool intel_dsi_get_hw_state(struct intel_encoder *encoder, > > } > > out_put_power: > > - intel_display_power_put(dev_priv, encoder->power_domain, wakeref); > > + intel_display_power_put(i915, encoder->power_domain, wakeref); > > return active; > > } > > @@ -1058,7 +1058,7 @@ static void bxt_dsi_get_pipe_config(struct intel_encoder *encoder, > > struct intel_crtc_state *pipe_config) > > { > > struct drm_device *dev = encoder->base.dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > struct drm_display_mode *adjusted_mode = > > &pipe_config->base.adjusted_mode; > > struct drm_display_mode *adjusted_mode_sw; > > @@ -1218,13 +1218,13 @@ static void bxt_dsi_get_pipe_config(struct intel_encoder *encoder, > > static void intel_dsi_get_config(struct intel_encoder *encoder, > > struct intel_crtc_state *pipe_config) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > u32 pclk; > > DRM_DEBUG_KMS("\n"); > > pipe_config->output_types |= BIT(INTEL_OUTPUT_DSI); > > - if (IS_GEN9_LP(dev_priv)) { > > + if (IS_GEN9_LP(i915)) { > > bxt_dsi_get_pipe_config(encoder, pipe_config); > > pclk = bxt_dsi_get_pclk(encoder, pipe_config); > > } else { > > @@ -1255,7 +1255,7 @@ static void set_dsi_timings(struct drm_encoder *encoder, > > const struct drm_display_mode *adjusted_mode) > > { > > struct drm_device *dev = encoder->dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > struct intel_dsi *intel_dsi = enc_to_intel_dsi(encoder); > > enum port port; > > unsigned int bpp = mipi_dsi_pixel_format_to_bpp(intel_dsi->pixel_format); > > @@ -1290,7 +1290,7 @@ static void set_dsi_timings(struct drm_encoder *encoder, > > hbp = txbyteclkhs(hbp, bpp, lane_count, intel_dsi->burst_mode_ratio); > > for_each_dsi_port(port, intel_dsi->ports) { > > - if (IS_GEN9_LP(dev_priv)) { > > + if (IS_GEN9_LP(i915)) { > > /* > > * Program hdisplay and vdisplay on MIPI transcoder. > > * This is different from calculated hactive and > > @@ -1342,7 +1342,7 @@ static void intel_dsi_prepare(struct intel_encoder *intel_encoder, > > { > > struct drm_encoder *encoder = &intel_encoder->base; > > struct drm_device *dev = encoder->dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > struct intel_crtc *intel_crtc = to_intel_crtc(pipe_config->base.crtc); > > struct intel_dsi *intel_dsi = enc_to_intel_dsi(encoder); > > const struct drm_display_mode *adjusted_mode = &pipe_config->base.adjusted_mode; > > @@ -1362,7 +1362,7 @@ static void intel_dsi_prepare(struct intel_encoder *intel_encoder, > > } > > for_each_dsi_port(port, intel_dsi->ports) { > > - if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) { > > + if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) { > > /* > > * escape clock divider, 20MHz, shared for A and C. > > * device ready must be off when doing this! txclkesc? > > @@ -1377,7 +1377,7 @@ static void intel_dsi_prepare(struct intel_encoder *intel_encoder, > > tmp &= ~READ_REQUEST_PRIORITY_MASK; > > I915_WRITE(MIPI_CTRL(port), tmp | > > READ_REQUEST_PRIORITY_HIGH); > > - } else if (IS_GEN9_LP(dev_priv)) { > > + } else if (IS_GEN9_LP(i915)) { > > enum pipe pipe = intel_crtc->pipe; > > tmp = I915_READ(MIPI_CTRL(port)); > > @@ -1415,7 +1415,7 @@ static void intel_dsi_prepare(struct intel_encoder *intel_encoder, > > if (intel_dsi->clock_stop) > > tmp |= CLOCKSTOP; > > - if (IS_GEN9_LP(dev_priv)) { > > + if (IS_GEN9_LP(i915)) { > > tmp |= BXT_DPHY_DEFEATURE_EN; > > if (!is_cmd_mode(intel_dsi)) > > tmp |= BXT_DEFEATURE_DPI_FIFO_CTR; > > @@ -1466,7 +1466,7 @@ static void intel_dsi_prepare(struct intel_encoder *intel_encoder, > > I915_WRITE(MIPI_INIT_COUNT(port), > > txclkesc(intel_dsi->escape_clk_div, 100)); > > - if (IS_GEN9_LP(dev_priv) && (!intel_dsi->dual_link)) { > > + if (IS_GEN9_LP(i915) && (!intel_dsi->dual_link)) { > > /* > > * BXT spec says write MIPI_INIT_COUNT for > > * both the ports, even if only one is > > @@ -1500,7 +1500,7 @@ static void intel_dsi_prepare(struct intel_encoder *intel_encoder, > > */ > > I915_WRITE(MIPI_LP_BYTECLK(port), intel_dsi->lp_byte_clk); > > - if (IS_GEMINILAKE(dev_priv)) { > > + if (IS_GEMINILAKE(i915)) { > > I915_WRITE(MIPI_TLPX_TIME_COUNT(port), > > intel_dsi->lp_byte_clk); > > /* Shadow of DPHY reg */ > > @@ -1533,19 +1533,19 @@ static void intel_dsi_prepare(struct intel_encoder *intel_encoder, > > static void intel_dsi_unprepare(struct intel_encoder *encoder) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base); > > enum port port; > > u32 val; > > - if (IS_GEMINILAKE(dev_priv)) > > + if (IS_GEMINILAKE(i915)) > > return; > > for_each_dsi_port(port, intel_dsi->ports) { > > /* Panel commands can be sent when clock is in LP11 */ > > I915_WRITE(MIPI_DEVICE_READY(port), 0x0); > > - if (IS_GEN9_LP(dev_priv)) > > + if (IS_GEN9_LP(i915)) > > bxt_dsi_reset_clocks(encoder, port); > > else > > vlv_dsi_reset_clocks(encoder, port); > > @@ -1594,7 +1594,7 @@ static const struct drm_connector_funcs intel_dsi_connector_funcs = { > > static enum drm_panel_orientation > > vlv_dsi_get_hw_panel_orientation(struct intel_connector *connector) > > { > > - struct drm_i915_private *dev_priv = to_i915(connector->base.dev); > > + struct drm_i915_private *i915 = to_i915(connector->base.dev); > > struct intel_encoder *encoder = connector->encoder; > > enum intel_display_power_domain power_domain; > > enum drm_panel_orientation orientation; > > @@ -1607,11 +1607,11 @@ vlv_dsi_get_hw_panel_orientation(struct intel_connector *connector) > > if (!encoder->get_hw_state(encoder, &pipe)) > > return DRM_MODE_PANEL_ORIENTATION_UNKNOWN; > > - crtc = intel_get_crtc_for_pipe(dev_priv, pipe); > > + crtc = intel_get_crtc_for_pipe(i915, pipe); > > plane = to_intel_plane(crtc->base.primary); > > power_domain = POWER_DOMAIN_PIPE(pipe); > > - wakeref = intel_display_power_get_if_enabled(dev_priv, power_domain); > > + wakeref = intel_display_power_get_if_enabled(i915, power_domain); > > if (!wakeref) > > return DRM_MODE_PANEL_ORIENTATION_UNKNOWN; > > @@ -1624,7 +1624,7 @@ vlv_dsi_get_hw_panel_orientation(struct intel_connector *connector) > > else > > orientation = DRM_MODE_PANEL_ORIENTATION_NORMAL; > > - intel_display_power_put(dev_priv, power_domain, wakeref); > > + intel_display_power_put(i915, power_domain, wakeref); > > return orientation; > > } > > @@ -1632,10 +1632,10 @@ vlv_dsi_get_hw_panel_orientation(struct intel_connector *connector) > > static enum drm_panel_orientation > > vlv_dsi_get_panel_orientation(struct intel_connector *connector) > > { > > - struct drm_i915_private *dev_priv = to_i915(connector->base.dev); > > + struct drm_i915_private *i915 = to_i915(connector->base.dev); > > enum drm_panel_orientation orientation; > > - if (IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) { > > + if (IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) { > > orientation = vlv_dsi_get_hw_panel_orientation(connector); > > if (orientation != DRM_MODE_PANEL_ORIENTATION_UNKNOWN) > > return orientation; > > @@ -1646,13 +1646,13 @@ vlv_dsi_get_panel_orientation(struct intel_connector *connector) > > static void intel_dsi_add_properties(struct intel_connector *connector) > > { > > - struct drm_i915_private *dev_priv = to_i915(connector->base.dev); > > + struct drm_i915_private *i915 = to_i915(connector->base.dev); > > if (connector->panel.fixed_mode) { > > u32 allowed_scalers; > > allowed_scalers = BIT(DRM_MODE_SCALE_ASPECT) | BIT(DRM_MODE_SCALE_FULLSCREEN); > > - if (!HAS_GMCH(dev_priv)) > > + if (!HAS_GMCH(i915)) > > allowed_scalers |= BIT(DRM_MODE_SCALE_CENTER); > > drm_connector_attach_scaling_mode_property(&connector->base, > > @@ -1679,8 +1679,8 @@ static void intel_dsi_add_properties(struct intel_connector *connector) > > static void vlv_dphy_param_init(struct intel_dsi *intel_dsi) > > { > > struct drm_device *dev = intel_dsi->base.base.dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > - struct mipi_config *mipi_config = dev_priv->vbt.dsi.config; > > + struct drm_i915_private *i915 = to_i915(dev); > > + struct mipi_config *mipi_config = i915->vbt.dsi.config; > > u32 tlpx_ns, extra_byte_count, tlpx_ui; > > u32 ui_num, ui_den; > > u32 prepare_cnt, exit_zero_cnt, clk_zero_cnt, trail_cnt; > > @@ -1727,7 +1727,7 @@ static void vlv_dphy_param_init(struct intel_dsi *intel_dsi) > > * For GEMINILAKE dphy_param_reg will be programmed in terms of > > * HS byte clock count for other platform in HS ddr clock count > > */ > > - mul = IS_GEMINILAKE(dev_priv) ? 8 : 2; > > + mul = IS_GEMINILAKE(i915) ? 8 : 2; > > ths_prepare_ns = max(mipi_config->ths_prepare, > > mipi_config->tclk_prepare); > > @@ -1837,9 +1837,9 @@ static void vlv_dphy_param_init(struct intel_dsi *intel_dsi) > > intel_dsi_log_params(intel_dsi); > > } > > -void vlv_dsi_init(struct drm_i915_private *dev_priv) > > +void vlv_dsi_init(struct drm_i915_private *i915) > > { > > - struct drm_device *dev = &dev_priv->drm; > > + struct drm_device *dev = &i915->drm; > > struct intel_dsi *intel_dsi; > > struct intel_encoder *intel_encoder; > > struct drm_encoder *encoder; > > @@ -1851,13 +1851,13 @@ void vlv_dsi_init(struct drm_i915_private *dev_priv) > > DRM_DEBUG_KMS("\n"); > > /* There is no detection method for MIPI so rely on VBT */ > > - if (!intel_bios_is_dsi_present(dev_priv, &port)) > > + if (!intel_bios_is_dsi_present(i915, &port)) > > return; > > - if (IS_GEN9_LP(dev_priv)) > > - dev_priv->mipi_mmio_base = BXT_MIPI_BASE; > > + if (IS_GEN9_LP(i915)) > > + i915->mipi_mmio_base = BXT_MIPI_BASE; > > else > > - dev_priv->mipi_mmio_base = VLV_MIPI_BASE; > > + i915->mipi_mmio_base = VLV_MIPI_BASE; > > intel_dsi = kzalloc(sizeof(*intel_dsi), GFP_KERNEL); > > if (!intel_dsi) > > @@ -1897,20 +1897,20 @@ void vlv_dsi_init(struct drm_i915_private *dev_priv) > > * On BYT/CHV, pipe A maps to MIPI DSI port A, pipe B maps to MIPI DSI > > * port C. BXT isn't limited like this. > > */ > > - if (IS_GEN9_LP(dev_priv)) > > + if (IS_GEN9_LP(i915)) > > intel_encoder->crtc_mask = BIT(PIPE_A) | BIT(PIPE_B) | BIT(PIPE_C); > > else if (port == PORT_A) > > intel_encoder->crtc_mask = BIT(PIPE_A); > > else > > intel_encoder->crtc_mask = BIT(PIPE_B); > > - if (dev_priv->vbt.dsi.config->dual_link) > > + if (i915->vbt.dsi.config->dual_link) > > intel_dsi->ports = BIT(PORT_A) | BIT(PORT_C); > > else > > intel_dsi->ports = BIT(port); > > - intel_dsi->dcs_backlight_ports = dev_priv->vbt.dsi.bl_ports; > > - intel_dsi->dcs_cabc_ports = dev_priv->vbt.dsi.cabc_ports; > > + intel_dsi->dcs_backlight_ports = i915->vbt.dsi.bl_ports; > > + intel_dsi->dcs_cabc_ports = i915->vbt.dsi.cabc_ports; > > /* Create a DSI host (and a device) for each port. */ > > for_each_dsi_port(port, intel_dsi->ports) { > > @@ -1949,8 +1949,8 @@ void vlv_dsi_init(struct drm_i915_private *dev_priv) > > * In case of BYT with CRC PMIC, we need to use GPIO for > > * Panel control. > > */ > > - if ((IS_VALLEYVIEW(dev_priv) || IS_CHERRYVIEW(dev_priv)) && > > - (dev_priv->vbt.dsi.config->pwm_blc == PPS_BLC_PMIC)) { > > + if ((IS_VALLEYVIEW(i915) || IS_CHERRYVIEW(i915)) && > > + (i915->vbt.dsi.config->pwm_blc == PPS_BLC_PMIC)) { > > intel_dsi->gpio_panel = > > gpiod_get(dev->dev, "panel", GPIOD_OUT_HIGH); > > diff --git a/drivers/gpu/drm/i915/vlv_dsi_pll.c b/drivers/gpu/drm/i915/vlv_dsi_pll.c > > index 99cc3e2e9c2c..26655d5472b4 100644 > > --- a/drivers/gpu/drm/i915/vlv_dsi_pll.c > > +++ b/drivers/gpu/drm/i915/vlv_dsi_pll.c > > @@ -53,7 +53,7 @@ static u32 dsi_clk_from_pclk(u32 pclk, enum mipi_dsi_pixel_format fmt, > > return dsi_clk_khz; > > } > > -static int dsi_calc_mnp(struct drm_i915_private *dev_priv, > > +static int dsi_calc_mnp(struct drm_i915_private *i915, > > struct intel_crtc_state *config, > > int target_dsi_clk) > > { > > @@ -68,7 +68,7 @@ static int dsi_calc_mnp(struct drm_i915_private *dev_priv, > > return -ECHRNG; > > } > > - if (IS_CHERRYVIEW(dev_priv)) { > > + if (IS_CHERRYVIEW(i915)) { > > ref_clk = 100000; > > n = 4; > > m_min = 70; > > @@ -116,7 +116,7 @@ static int dsi_calc_mnp(struct drm_i915_private *dev_priv, > > int vlv_dsi_pll_compute(struct intel_encoder *encoder, > > struct intel_crtc_state *config) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base); > > int ret; > > u32 dsi_clk; > > @@ -124,7 +124,7 @@ int vlv_dsi_pll_compute(struct intel_encoder *encoder, > > dsi_clk = dsi_clk_from_pclk(intel_dsi->pclk, intel_dsi->pixel_format, > > intel_dsi->lane_count); > > - ret = dsi_calc_mnp(dev_priv, config, dsi_clk); > > + ret = dsi_calc_mnp(i915, config, dsi_clk); > > if (ret) { > > DRM_DEBUG_KMS("dsi_calc_mnp failed\n"); > > return ret; > > @@ -147,15 +147,15 @@ int vlv_dsi_pll_compute(struct intel_encoder *encoder, > > void vlv_dsi_pll_enable(struct intel_encoder *encoder, > > const struct intel_crtc_state *config) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > DRM_DEBUG_KMS("\n"); > > - vlv_cck_get(dev_priv); > > + vlv_cck_get(i915); > > - vlv_cck_write(dev_priv, CCK_REG_DSI_PLL_CONTROL, 0); > > - vlv_cck_write(dev_priv, CCK_REG_DSI_PLL_DIVIDER, config->dsi_pll.div); > > - vlv_cck_write(dev_priv, CCK_REG_DSI_PLL_CONTROL, > > + vlv_cck_write(i915, CCK_REG_DSI_PLL_CONTROL, 0); > > + vlv_cck_write(i915, CCK_REG_DSI_PLL_DIVIDER, config->dsi_pll.div); > > + vlv_cck_write(i915, CCK_REG_DSI_PLL_CONTROL, > > config->dsi_pll.ctrl & ~DSI_PLL_VCO_EN); > > /* wait at least 0.5 us after ungating before enabling VCO, > > @@ -163,38 +163,38 @@ void vlv_dsi_pll_enable(struct intel_encoder *encoder, > > */ > > usleep_range(10, 50); > > - vlv_cck_write(dev_priv, CCK_REG_DSI_PLL_CONTROL, config->dsi_pll.ctrl); > > + vlv_cck_write(i915, CCK_REG_DSI_PLL_CONTROL, config->dsi_pll.ctrl); > > - if (wait_for(vlv_cck_read(dev_priv, CCK_REG_DSI_PLL_CONTROL) & > > + if (wait_for(vlv_cck_read(i915, CCK_REG_DSI_PLL_CONTROL) & > > DSI_PLL_LOCK, 20)) { > > - vlv_cck_put(dev_priv); > > + vlv_cck_put(i915); > > DRM_ERROR("DSI PLL lock failed\n"); > > return; > > } > > - vlv_cck_put(dev_priv); > > + vlv_cck_put(i915); > > DRM_DEBUG_KMS("DSI PLL locked\n"); > > } > > void vlv_dsi_pll_disable(struct intel_encoder *encoder) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > u32 tmp; > > DRM_DEBUG_KMS("\n"); > > - vlv_cck_get(dev_priv); > > + vlv_cck_get(i915); > > - tmp = vlv_cck_read(dev_priv, CCK_REG_DSI_PLL_CONTROL); > > + tmp = vlv_cck_read(i915, CCK_REG_DSI_PLL_CONTROL); > > tmp &= ~DSI_PLL_VCO_EN; > > tmp |= DSI_PLL_LDO_GATE; > > - vlv_cck_write(dev_priv, CCK_REG_DSI_PLL_CONTROL, tmp); > > + vlv_cck_write(i915, CCK_REG_DSI_PLL_CONTROL, tmp); > > - vlv_cck_put(dev_priv); > > + vlv_cck_put(i915); > > } > > -bool bxt_dsi_pll_is_enabled(struct drm_i915_private *dev_priv) > > +bool bxt_dsi_pll_is_enabled(struct drm_i915_private *i915) > > { > > bool enabled; > > u32 val; > > @@ -216,7 +216,7 @@ bool bxt_dsi_pll_is_enabled(struct drm_i915_private *dev_priv) > > * causes a system hang. > > */ > > val = I915_READ(BXT_DSI_PLL_CTL); > > - if (IS_GEMINILAKE(dev_priv)) { > > + if (IS_GEMINILAKE(i915)) { > > if (!(val & BXT_DSIA_16X_MASK)) { > > DRM_DEBUG_DRIVER("Invalid PLL divider (%08x)\n", val); > > enabled = false; > > @@ -233,7 +233,7 @@ bool bxt_dsi_pll_is_enabled(struct drm_i915_private *dev_priv) > > void bxt_dsi_pll_disable(struct intel_encoder *encoder) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > u32 val; > > DRM_DEBUG_KMS("\n"); > > @@ -246,7 +246,7 @@ void bxt_dsi_pll_disable(struct intel_encoder *encoder) > > * PLL lock should deassert within 200us. > > * Wait up to 1ms before timing out. > > */ > > - if (intel_wait_for_register(&dev_priv->uncore, > > + if (intel_wait_for_register(&i915->uncore, > > BXT_DSI_PLL_ENABLE, > > BXT_DSI_PLL_LOCKED, > > 0, > > @@ -257,21 +257,21 @@ void bxt_dsi_pll_disable(struct intel_encoder *encoder) > > u32 vlv_dsi_get_pclk(struct intel_encoder *encoder, > > struct intel_crtc_state *config) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base); > > int bpp = mipi_dsi_pixel_format_to_bpp(intel_dsi->pixel_format); > > u32 dsi_clock, pclk; > > u32 pll_ctl, pll_div; > > u32 m = 0, p = 0, n; > > - int refclk = IS_CHERRYVIEW(dev_priv) ? 100000 : 25000; > > + int refclk = IS_CHERRYVIEW(i915) ? 100000 : 25000; > > int i; > > DRM_DEBUG_KMS("\n"); > > - vlv_cck_get(dev_priv); > > - pll_ctl = vlv_cck_read(dev_priv, CCK_REG_DSI_PLL_CONTROL); > > - pll_div = vlv_cck_read(dev_priv, CCK_REG_DSI_PLL_DIVIDER); > > - vlv_cck_put(dev_priv); > > + vlv_cck_get(i915); > > + pll_ctl = vlv_cck_read(i915, CCK_REG_DSI_PLL_CONTROL); > > + pll_div = vlv_cck_read(i915, CCK_REG_DSI_PLL_DIVIDER); > > + vlv_cck_put(i915); > > config->dsi_pll.ctrl = pll_ctl & ~DSI_PLL_LOCK; > > config->dsi_pll.div = pll_div; > > @@ -325,7 +325,7 @@ u32 bxt_dsi_get_pclk(struct intel_encoder *encoder, > > u32 dsi_clk; > > u32 dsi_ratio; > > struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base); > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > int bpp = mipi_dsi_pixel_format_to_bpp(intel_dsi->pixel_format); > > config->dsi_pll.ctrl = I915_READ(BXT_DSI_PLL_CTL); > > @@ -343,7 +343,7 @@ u32 bxt_dsi_get_pclk(struct intel_encoder *encoder, > > void vlv_dsi_reset_clocks(struct intel_encoder *encoder, enum port port) > > { > > u32 temp; > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base); > > temp = I915_READ(MIPI_CTRL(port)); > > @@ -356,7 +356,7 @@ void vlv_dsi_reset_clocks(struct intel_encoder *encoder, enum port port) > > static void glk_dsi_program_esc_clock(struct drm_device *dev, > > const struct intel_crtc_state *config) > > { > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > u32 dsi_rate = 0; > > u32 pll_ratio = 0; > > u32 ddr_clk = 0; > > @@ -404,7 +404,7 @@ static void glk_dsi_program_esc_clock(struct drm_device *dev, > > static void bxt_dsi_program_clocks(struct drm_device *dev, enum port port, > > const struct intel_crtc_state *config) > > { > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > u32 tmp; > > u32 dsi_rate = 0; > > u32 pll_ratio = 0; > > @@ -457,7 +457,7 @@ static void bxt_dsi_program_clocks(struct drm_device *dev, enum port port, > > int bxt_dsi_pll_compute(struct intel_encoder *encoder, > > struct intel_crtc_state *config) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base); > > u8 dsi_ratio, dsi_ratio_min, dsi_ratio_max; > > u32 dsi_clk; > > @@ -472,7 +472,7 @@ int bxt_dsi_pll_compute(struct intel_encoder *encoder, > > */ > > dsi_ratio = DIV_ROUND_UP(dsi_clk * 2, BXT_REF_CLOCK_KHZ); > > - if (IS_BROXTON(dev_priv)) { > > + if (IS_BROXTON(i915)) { > > dsi_ratio_min = BXT_DSI_PLL_RATIO_MIN; > > dsi_ratio_max = BXT_DSI_PLL_RATIO_MAX; > > } else { > > @@ -496,7 +496,7 @@ int bxt_dsi_pll_compute(struct intel_encoder *encoder, > > /* As per recommendation from hardware team, > > * Prog PVD ratio =1 if dsi ratio <= 50 > > */ > > - if (IS_BROXTON(dev_priv) && dsi_ratio <= 50) > > + if (IS_BROXTON(i915) && dsi_ratio <= 50) > > config->dsi_pll.ctrl |= BXT_DSI_PLL_PVD_RATIO_1; > > return 0; > > @@ -505,7 +505,7 @@ int bxt_dsi_pll_compute(struct intel_encoder *encoder, > > void bxt_dsi_pll_enable(struct intel_encoder *encoder, > > const struct intel_crtc_state *config) > > { > > - struct drm_i915_private *dev_priv = to_i915(encoder->base.dev); > > + struct drm_i915_private *i915 = to_i915(encoder->base.dev); > > struct intel_dsi *intel_dsi = enc_to_intel_dsi(&encoder->base); > > enum port port; > > u32 val; > > @@ -517,7 +517,7 @@ void bxt_dsi_pll_enable(struct intel_encoder *encoder, > > POSTING_READ(BXT_DSI_PLL_CTL); > > /* Program TX, RX, Dphy clocks */ > > - if (IS_BROXTON(dev_priv)) { > > + if (IS_BROXTON(i915)) { > > for_each_dsi_port(port, intel_dsi->ports) > > bxt_dsi_program_clocks(encoder->base.dev, port, config); > > } else { > > @@ -530,7 +530,7 @@ void bxt_dsi_pll_enable(struct intel_encoder *encoder, > > I915_WRITE(BXT_DSI_PLL_ENABLE, val); > > /* Timeout and fail if PLL not locked */ > > - if (intel_wait_for_register(&dev_priv->uncore, > > + if (intel_wait_for_register(&i915->uncore, > > BXT_DSI_PLL_ENABLE, > > BXT_DSI_PLL_LOCKED, > > BXT_DSI_PLL_LOCKED, > > @@ -546,10 +546,10 @@ void bxt_dsi_reset_clocks(struct intel_encoder *encoder, enum port port) > > { > > u32 tmp; > > struct drm_device *dev = encoder->base.dev; > > - struct drm_i915_private *dev_priv = to_i915(dev); > > + struct drm_i915_private *i915 = to_i915(dev); > > /* Clear old configurations */ > > - if (IS_BROXTON(dev_priv)) { > > + if (IS_BROXTON(i915)) { > > tmp = I915_READ(BXT_MIPI_CLOCK_CTL); > > tmp &= ~(BXT_MIPI_TX_ESCLK_FIXDIV_MASK(port)); > > tmp &= ~(BXT_MIPI_RX_ESCLK_UPPER_FIXDIV_MASK(port)); > > _______________________________________________ Intel-gfx mailing list Intel-gfx@xxxxxxxxxxxxxxxxxxxxx https://lists.freedesktop.org/mailman/listinfo/intel-gfx