> We will discuss it tomorrow, I'm not sure of your approach right now > > We can stop running driver right after IDH_FLR_NOTIFICATION received and > continue driver right after IDH_FLR_NOTIFICATION_CMPL I have verified the method on my platform and work fine. Let's discuss it tomorrow if you have any concern. > -----Original Message----- > From: Yu, Xiangliang > Sent: Wednesday, January 11, 2017 10:43 PM > To: Liu, Monk <Monk.Liu at amd.com>; amd-gfx at lists.freedesktop.org > Cc: Liu, Shaoyun <Shaoyun.Liu at amd.com> > Subject: RE: [V3 06/11] drm/amdgpu/virt: implement VI virt operation > interfaces > > > -----Original Message----- > > From: Liu, Monk > > Sent: Wednesday, January 11, 2017 10:28 PM > > To: Yu, Xiangliang <Xiangliang.Yu at amd.com>; amd- > > gfx at lists.freedesktop.org > > Cc: Yu, Xiangliang <Xiangliang.Yu at amd.com>; Liu, Shaoyun > > <Shaoyun.Liu at amd.com> > > Subject: RE: [V3 06/11] drm/amdgpu/virt: implement VI virt operation > > interfaces > > > > +static int xgpu_vi_mailbox_rcv_irq(struct amdgpu_device *adev, > > + struct amdgpu_irq_src *source, > > + struct amdgpu_iv_entry *entry) { > > + int r; > > + > > + adev->virt.caps &= ~AMDGPU_SRIOV_CAPS_RUNTIME; > > + r = xgpu_vi_mailbox_rcv_msg(adev, IDH_FLR_NOTIFICATION); > > + /* do nothing for other msg */ > > + if (r) > > + return 0; > > + > > + /* TODO: need to save gfx states */ > > + schedule_delayed_work(&adev->virt.flr_work, > > + msecs_to_jiffies(VI_MAILBOX_RESET_TIME)); > > + > > + return 0; > > +} > > > > [ML] we should be able to receive at least two type of IRQ from mailbox: > > One is FLR_NOTIFICATION and another one is > IDH_FLR_NOTIFICATION_CMPL, > > > > The previous one means GIM told us GIM will start do FLR The later one > > means GIM told us GIM already finished FLR and from that on Guest > > driver can do re-initialize . > > > > You only handle the first IRQ, and dispatch a delay_work to monitor > > event of IDH_FLR_NOTIFICATION_CMPL, > > > > I suspect that correct scheme may be you handle both of those two type > > of IRQ , please check with @Frank and me on the detail tomorrow . > > As I think we may suspend gpu after receive FLR interrupt, polling method is > safer. > > > > > BR Monk > > > > -----Original Message----- > > From: Xiangliang Yu [mailto:Xiangliang.Yu at amd.com] > > Sent: Wednesday, January 11, 2017 9:18 PM > > To: amd-gfx at lists.freedesktop.org > > Cc: Yu, Xiangliang <Xiangliang.Yu at amd.com>; Liu, Monk > > <Monk.Liu at amd.com>; Liu, Shaoyun <Shaoyun.Liu at amd.com> > > Subject: [V3 06/11] drm/amdgpu/virt: implement VI virt operation > > interfaces > > > > VI has asic specific virt support, which including mailbox and golden > > registers init. > > > > Signed-off-by: Xiangliang Yu <Xiangliang.Yu at amd.com> > > Signed-off-by: Monk Liu <Monk.Liu at amd.com> > > Signed-off-by: shaoyunl <Shaoyun.Liu at amd.com> > > --- > > drivers/gpu/drm/amd/amdgpu/Makefile | 2 +- > > drivers/gpu/drm/amd/amdgpu/amdgpu_virt.h | 9 +- > > drivers/gpu/drm/amd/amdgpu/mxgpu_vi.c | 592 > > +++++++++++++++++++++++++++++++ > > drivers/gpu/drm/amd/amdgpu/mxgpu_vi.h | 55 +++ > > 4 files changed, 654 insertions(+), 4 deletions(-) create mode > > 100644 drivers/gpu/drm/amd/amdgpu/mxgpu_vi.c > > create mode 100644 drivers/gpu/drm/amd/amdgpu/mxgpu_vi.h > > > > diff --git a/drivers/gpu/drm/amd/amdgpu/Makefile > > b/drivers/gpu/drm/amd/amdgpu/Makefile > > index 0b8e470..388e7b0 100644 > > --- a/drivers/gpu/drm/amd/amdgpu/Makefile > > +++ b/drivers/gpu/drm/amd/amdgpu/Makefile > > @@ -40,7 +40,7 @@ amdgpu-$(CONFIG_DRM_AMDGPU_CIK)+= cik.o > cik_ih.o > > kv_smc.o kv_dpm.o \ amdgpu-$(CONFIG_DRM_AMDGPU_SI)+= si.o > gmc_v6_0.o > > gfx_v6_0.o si_ih.o si_dma.o dce_v6_0.o si_dpm.o si_smc.o > > > > amdgpu-y += \ > > - vi.o > > + vi.o mxgpu_vi.o > > > > # add GMC block > > amdgpu-y += \ > > diff --git a/drivers/gpu/drm/amd/amdgpu/amdgpu_virt.h > > b/drivers/gpu/drm/amd/amdgpu/amdgpu_virt.h > > index 45771b0..a409eec 100644 > > --- a/drivers/gpu/drm/amd/amdgpu/amdgpu_virt.h > > +++ b/drivers/gpu/drm/amd/amdgpu/amdgpu_virt.h > > @@ -41,9 +41,12 @@ struct amdgpu_virt_ops { > > > > /* GPU virtualization */ > > struct amdgpu_virt { > > - uint32_t caps; > > - uint32_t reg_val_offs; > > - struct mutex lock; > > + uint32_t caps; > > + uint32_t reg_val_offs; > > + struct mutex lock; > > + struct amdgpu_irq_src ack_irq; > > + struct amdgpu_irq_src rcv_irq; > > + struct delayed_work flr_work; > > const struct amdgpu_virt_ops *ops; > > }; > > > > diff --git a/drivers/gpu/drm/amd/amdgpu/mxgpu_vi.c > > b/drivers/gpu/drm/amd/amdgpu/mxgpu_vi.c > > new file mode 100644 > > index 0000000..d2622b6 > > --- /dev/null > > +++ b/drivers/gpu/drm/amd/amdgpu/mxgpu_vi.c > > @@ -0,0 +1,592 @@ > > +/* > > + * Copyright 2017 Advanced Micro Devices, Inc. > > + * > > + * Permission is hereby granted, free of charge, to any person > > +obtaining a > > + * copy of this software and associated documentation files (the > > +"Software"), > > + * to deal in the Software without restriction, including without > > +limitation > > + * the rights to use, copy, modify, merge, publish, distribute, > > +sublicense, > > + * and/or sell copies of the Software, and to permit persons to whom > > +the > > + * Software is furnished to do so, subject to the following conditions: > > + * > > + * The above copyright notice and this permission notice shall be > > +included in > > + * all copies or substantial portions of the Software. > > + * > > + * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY > KIND, > > +EXPRESS OR > > + * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF > > +MERCHANTABILITY, > > + * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN > NO > > EVENT > > +SHALL > > + * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, > > +DAMAGES OR > > + * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR > > +OTHERWISE, > > + * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR > > THE USE > > +OR > > + * OTHER DEALINGS IN THE SOFTWARE. > > + * > > + * Authors: Xiangliang.Yu at amd.com > > + */ > > + > > +#include "amdgpu.h" > > +#include "vi.h" > > +#include "bif/bif_5_0_d.h" > > +#include "bif/bif_5_0_sh_mask.h" > > +#include "vid.h" > > +#include "gca/gfx_8_0_d.h" > > +#include "gca/gfx_8_0_sh_mask.h" > > +#include "gmc_v8_0.h" > > +#include "gfx_v8_0.h" > > +#include "sdma_v3_0.h" > > +#include "tonga_ih.h" > > +#include "gmc/gmc_8_2_d.h" > > +#include "gmc/gmc_8_2_sh_mask.h" > > +#include "oss/oss_3_0_d.h" > > +#include "oss/oss_3_0_sh_mask.h" > > +#include "gca/gfx_8_0_sh_mask.h" > > +#include "dce/dce_10_0_d.h" > > +#include "dce/dce_10_0_sh_mask.h" > > +#include "smu/smu_7_1_3_d.h" > > +#include "mxgpu_vi.h" > > + > > +/* VI golden setting */ > > +static const u32 xgpu_fiji_mgcg_cgcg_init[] = { > > + mmRLC_CGTT_MGCG_OVERRIDE, 0xffffffff, 0xffffffff, > > + mmGRBM_GFX_INDEX, 0xffffffff, 0xe0000000, > > + mmCB_CGTT_SCLK_CTRL, 0xffffffff, 0x00000100, > > + mmCGTT_BCI_CLK_CTRL, 0xffffffff, 0x00000100, > > + mmCGTT_CP_CLK_CTRL, 0xffffffff, 0x00000100, > > + mmCGTT_CPC_CLK_CTRL, 0xffffffff, 0x00000100, > > + mmCGTT_CPF_CLK_CTRL, 0xffffffff, 0x40000100, > > + mmCGTT_DRM_CLK_CTRL0, 0xffffffff, 0x00600100, > > + mmCGTT_GDS_CLK_CTRL, 0xffffffff, 0x00000100, > > + mmCGTT_IA_CLK_CTRL, 0xffffffff, 0x06000100, > > + mmCGTT_PA_CLK_CTRL, 0xffffffff, 0x00000100, > > + mmCGTT_WD_CLK_CTRL, 0xffffffff, 0x06000100, > > + mmCGTT_PC_CLK_CTRL, 0xffffffff, 0x00000100, > > + mmCGTT_RLC_CLK_CTRL, 0xffffffff, 0x00000100, > > + mmCGTT_SC_CLK_CTRL, 0xffffffff, 0x00000100, > > + mmCGTT_SPI_CLK_CTRL, 0xffffffff, 0x00000100, > > + mmCGTT_SQ_CLK_CTRL, 0xffffffff, 0x00000100, > > + mmCGTT_SQG_CLK_CTRL, 0xffffffff, 0x00000100, > > + mmCGTT_SX_CLK_CTRL0, 0xffffffff, 0x00000100, > > + mmCGTT_SX_CLK_CTRL1, 0xffffffff, 0x00000100, > > + mmCGTT_SX_CLK_CTRL2, 0xffffffff, 0x00000100, > > + mmCGTT_SX_CLK_CTRL3, 0xffffffff, 0x00000100, > > + mmCGTT_SX_CLK_CTRL4, 0xffffffff, 0x00000100, > > + mmCGTT_TCI_CLK_CTRL, 0xffffffff, 0x00000100, > > + mmCGTT_TCP_CLK_CTRL, 0xffffffff, 0x00000100, > > + mmCGTT_VGT_CLK_CTRL, 0xffffffff, 0x06000100, > > + mmDB_CGTT_CLK_CTRL_0, 0xffffffff, 0x00000100, > > + mmTA_CGTT_CTRL, 0xffffffff, 0x00000100, > > + mmTCA_CGTT_SCLK_CTRL, 0xffffffff, 0x00000100, > > + mmTCC_CGTT_SCLK_CTRL, 0xffffffff, 0x00000100, > > + mmTD_CGTT_CTRL, 0xffffffff, 0x00000100, > > + mmGRBM_GFX_INDEX, 0xffffffff, 0xe0000000, > > + mmCGTS_SM_CTRL_REG, 0xffffffff, 0x96e00200, > > + mmCP_RB_WPTR_POLL_CNTL, 0xffffffff, 0x00900100, > > + mmRLC_CGCG_CGLS_CTRL, 0xffffffff, 0x0020003c, > > + mmPCIE_INDEX, 0xffffffff, 0x0140001c, > > + mmPCIE_DATA, 0x000f0000, 0x00000000, > > + mmSMC_IND_INDEX_4, 0xffffffff, 0xC060000C, > > + mmSMC_IND_DATA_4, 0xc0000fff, 0x00000100, > > + mmXDMA_CLOCK_GATING_CNTL, 0xffffffff, 0x00000100, > > + mmXDMA_MEM_POWER_CNTL, 0x00000101, 0x00000000, > > + mmMC_MEM_POWER_LS, 0xffffffff, 0x00000104, > > + mmCGTT_DRM_CLK_CTRL0, 0xff000fff, 0x00000100, > > + mmHDP_XDP_CGTT_BLK_CTRL, 0xc0000fff, 0x00000104, > > + mmCP_MEM_SLP_CNTL, 0x00000001, 0x00000001, > > + mmSDMA0_CLK_CTRL, 0xff000ff0, 0x00000100, > > + mmSDMA1_CLK_CTRL, 0xff000ff0, 0x00000100, }; > > + > > +static const u32 xgpu_fiji_golden_settings_a10[] = { > > + mmCB_HW_CONTROL_3, 0x000001ff, 0x00000040, > > + mmDB_DEBUG2, 0xf00fffff, 0x00000400, > > + mmDCI_CLK_CNTL, 0x00000080, 0x00000000, > > + mmFBC_DEBUG_COMP, 0x000000f0, 0x00000070, > > + mmFBC_MISC, 0x1f311fff, 0x12300000, > > + mmHDMI_CONTROL, 0x31000111, 0x00000011, > > + mmPA_SC_ENHANCE, 0xffffffff, 0x20000001, > > + mmPA_SC_LINE_STIPPLE_STATE, 0x0000ff0f, 0x00000000, > > + mmSDMA0_CHICKEN_BITS, 0xfc910007, 0x00810007, > > + mmSDMA0_GFX_IB_CNTL, 0x800f0111, 0x00000100, > > + mmSDMA0_RLC0_IB_CNTL, 0x800f0111, 0x00000100, > > + mmSDMA0_RLC1_IB_CNTL, 0x800f0111, 0x00000100, > > + mmSDMA1_CHICKEN_BITS, 0xfc910007, 0x00810007, > > + mmSDMA1_GFX_IB_CNTL, 0x800f0111, 0x00000100, > > + mmSDMA1_RLC0_IB_CNTL, 0x800f0111, 0x00000100, > > + mmSDMA1_RLC1_IB_CNTL, 0x800f0111, 0x00000100, > > + mmSQ_RANDOM_WAVE_PRI, 0x001fffff, 0x000006fd, > > + mmTA_CNTL_AUX, 0x000f000f, 0x000b0000, > > + mmTCC_EXE_DISABLE, 0x00000002, 0x00000002, > > + mmTCP_ADDR_CONFIG, 0x000003ff, 0x000000ff, > > + mmVGT_RESET_DEBUG, 0x00000004, 0x00000004, > > + mmVM_PRT_APERTURE0_LOW_ADDR, 0x0fffffff, 0x0fffffff, > > + mmVM_PRT_APERTURE1_LOW_ADDR, 0x0fffffff, 0x0fffffff, > > + mmVM_PRT_APERTURE2_LOW_ADDR, 0x0fffffff, 0x0fffffff, > > + mmVM_PRT_APERTURE3_LOW_ADDR, 0x0fffffff, 0x0fffffff, }; > > + > > +static const u32 xgpu_fiji_golden_common_all[] = { > > + mmGRBM_GFX_INDEX, 0xffffffff, 0xe0000000, > > + mmPA_SC_RASTER_CONFIG, 0xffffffff, 0x3a00161a, > > + mmPA_SC_RASTER_CONFIG_1, 0xffffffff, 0x0000002e, > > + mmGB_ADDR_CONFIG, 0xffffffff, 0x22011003, > > + mmSPI_RESOURCE_RESERVE_CU_0, 0xffffffff, 0x00000800, > > + mmSPI_RESOURCE_RESERVE_CU_1, 0xffffffff, 0x00000800, > > + mmSPI_RESOURCE_RESERVE_EN_CU_0, 0xffffffff, 0x00007FBF, > > + mmSPI_RESOURCE_RESERVE_EN_CU_1, 0xffffffff, 0x00007FAF, > > + mmGRBM_GFX_INDEX, 0xffffffff, 0xe0000000, > > + mmSPI_CONFIG_CNTL_1, 0x0000000f, 0x00000009, }; > > + > > +static const u32 xgpu_tonga_mgcg_cgcg_init[] = { > > + mmRLC_CGTT_MGCG_OVERRIDE, 0xffffffff, 0xffffffff, > > + mmGRBM_GFX_INDEX, 0xffffffff, 0xe0000000, > > + mmCB_CGTT_SCLK_CTRL, 0xffffffff, 0x00000100, > > + mmCGTT_BCI_CLK_CTRL, 0xffffffff, 0x00000100, > > + mmCGTT_CP_CLK_CTRL, 0xffffffff, 0x00000100, > > + mmCGTT_CPC_CLK_CTRL, 0xffffffff, 0x00000100, > > + mmCGTT_CPF_CLK_CTRL, 0xffffffff, 0x40000100, > > + mmCGTT_DRM_CLK_CTRL0, 0xffffffff, 0x00600100, > > + mmCGTT_GDS_CLK_CTRL, 0xffffffff, 0x00000100, > > + mmCGTT_IA_CLK_CTRL, 0xffffffff, 0x06000100, > > + mmCGTT_PA_CLK_CTRL, 0xffffffff, 0x00000100, > > + mmCGTT_WD_CLK_CTRL, 0xffffffff, 0x06000100, > > + mmCGTT_PC_CLK_CTRL, 0xffffffff, 0x00000100, > > + mmCGTT_RLC_CLK_CTRL, 0xffffffff, 0x00000100, > > + mmCGTT_SC_CLK_CTRL, 0xffffffff, 0x00000100, > > + mmCGTT_SPI_CLK_CTRL, 0xffffffff, 0x00000100, > > + mmCGTT_SQ_CLK_CTRL, 0xffffffff, 0x00000100, > > + mmCGTT_SQG_CLK_CTRL, 0xffffffff, 0x00000100, > > + mmCGTT_SX_CLK_CTRL0, 0xffffffff, 0x00000100, > > + mmCGTT_SX_CLK_CTRL1, 0xffffffff, 0x00000100, > > + mmCGTT_SX_CLK_CTRL2, 0xffffffff, 0x00000100, > > + mmCGTT_SX_CLK_CTRL3, 0xffffffff, 0x00000100, > > + mmCGTT_SX_CLK_CTRL4, 0xffffffff, 0x00000100, > > + mmCGTT_TCI_CLK_CTRL, 0xffffffff, 0x00000100, > > + mmCGTT_TCP_CLK_CTRL, 0xffffffff, 0x00000100, > > + mmCGTT_VGT_CLK_CTRL, 0xffffffff, 0x06000100, > > + mmDB_CGTT_CLK_CTRL_0, 0xffffffff, 0x00000100, > > + mmTA_CGTT_CTRL, 0xffffffff, 0x00000100, > > + mmTCA_CGTT_SCLK_CTRL, 0xffffffff, 0x00000100, > > + mmTCC_CGTT_SCLK_CTRL, 0xffffffff, 0x00000100, > > + mmTD_CGTT_CTRL, 0xffffffff, 0x00000100, > > + mmGRBM_GFX_INDEX, 0xffffffff, 0xe0000000, > > + mmCGTS_CU0_SP0_CTRL_REG, 0xffffffff, 0x00010000, > > + mmCGTS_CU0_LDS_SQ_CTRL_REG, 0xffffffff, 0x00030002, > > + mmCGTS_CU0_TA_SQC_CTRL_REG, 0xffffffff, 0x00040007, > > + mmCGTS_CU0_SP1_CTRL_REG, 0xffffffff, 0x00060005, > > + mmCGTS_CU0_TD_TCP_CTRL_REG, 0xffffffff, 0x00090008, > > + mmCGTS_CU1_SP0_CTRL_REG, 0xffffffff, 0x00010000, > > + mmCGTS_CU1_LDS_SQ_CTRL_REG, 0xffffffff, 0x00030002, > > + mmCGTS_CU1_TA_CTRL_REG, 0xffffffff, 0x00040007, > > + mmCGTS_CU1_SP1_CTRL_REG, 0xffffffff, 0x00060005, > > + mmCGTS_CU1_TD_TCP_CTRL_REG, 0xffffffff, 0x00090008, > > + mmCGTS_CU2_SP0_CTRL_REG, 0xffffffff, 0x00010000, > > + mmCGTS_CU2_LDS_SQ_CTRL_REG, 0xffffffff, 0x00030002, > > + mmCGTS_CU2_TA_CTRL_REG, 0xffffffff, 0x00040007, > > + mmCGTS_CU2_SP1_CTRL_REG, 0xffffffff, 0x00060005, > > + mmCGTS_CU2_TD_TCP_CTRL_REG, 0xffffffff, 0x00090008, > > + mmCGTS_CU3_SP0_CTRL_REG, 0xffffffff, 0x00010000, > > + mmCGTS_CU3_LDS_SQ_CTRL_REG, 0xffffffff, 0x00030002, > > + mmCGTS_CU3_TA_CTRL_REG, 0xffffffff, 0x00040007, > > + mmCGTS_CU3_SP1_CTRL_REG, 0xffffffff, 0x00060005, > > + mmCGTS_CU3_TD_TCP_CTRL_REG, 0xffffffff, 0x00090008, > > + mmCGTS_CU4_SP0_CTRL_REG, 0xffffffff, 0x00010000, > > + mmCGTS_CU4_LDS_SQ_CTRL_REG, 0xffffffff, 0x00030002, > > + mmCGTS_CU4_TA_SQC_CTRL_REG, 0xffffffff, 0x00040007, > > + mmCGTS_CU4_SP1_CTRL_REG, 0xffffffff, 0x00060005, > > + mmCGTS_CU4_TD_TCP_CTRL_REG, 0xffffffff, 0x00090008, > > + mmCGTS_CU5_SP0_CTRL_REG, 0xffffffff, 0x00010000, > > + mmCGTS_CU5_LDS_SQ_CTRL_REG, 0xffffffff, 0x00030002, > > + mmCGTS_CU5_TA_CTRL_REG, 0xffffffff, 0x00040007, > > + mmCGTS_CU5_SP1_CTRL_REG, 0xffffffff, 0x00060005, > > + mmCGTS_CU5_TD_TCP_CTRL_REG, 0xffffffff, 0x00090008, > > + mmCGTS_CU6_SP0_CTRL_REG, 0xffffffff, 0x00010000, > > + mmCGTS_CU6_LDS_SQ_CTRL_REG, 0xffffffff, 0x00030002, > > + mmCGTS_CU6_TA_CTRL_REG, 0xffffffff, 0x00040007, > > + mmCGTS_CU6_SP1_CTRL_REG, 0xffffffff, 0x00060005, > > + mmCGTS_CU6_TD_TCP_CTRL_REG, 0xffffffff, 0x00090008, > > + mmCGTS_CU7_SP0_CTRL_REG, 0xffffffff, 0x00010000, > > + mmCGTS_CU7_LDS_SQ_CTRL_REG, 0xffffffff, 0x00030002, > > + mmCGTS_CU7_TA_CTRL_REG, 0xffffffff, 0x00040007, > > + mmCGTS_CU7_SP1_CTRL_REG, 0xffffffff, 0x00060005, > > + mmCGTS_CU7_TD_TCP_CTRL_REG, 0xffffffff, 0x00090008, > > + mmCGTS_SM_CTRL_REG, 0xffffffff, 0x96e00200, > > + mmCP_RB_WPTR_POLL_CNTL, 0xffffffff, 0x00900100, > > + mmRLC_CGCG_CGLS_CTRL, 0xffffffff, 0x0020003c, > > + mmPCIE_INDEX, 0xffffffff, 0x0140001c, > > + mmPCIE_DATA, 0x000f0000, 0x00000000, > > + mmSMC_IND_INDEX_4, 0xffffffff, 0xC060000C, > > + mmSMC_IND_DATA_4, 0xc0000fff, 0x00000100, > > + mmXDMA_CLOCK_GATING_CNTL, 0xffffffff, 0x00000100, > > + mmXDMA_MEM_POWER_CNTL, 0x00000101, 0x00000000, > > + mmMC_MEM_POWER_LS, 0xffffffff, 0x00000104, > > + mmCGTT_DRM_CLK_CTRL0, 0xff000fff, 0x00000100, > > + mmHDP_XDP_CGTT_BLK_CTRL, 0xc0000fff, 0x00000104, > > + mmCP_MEM_SLP_CNTL, 0x00000001, 0x00000001, > > + mmSDMA0_CLK_CTRL, 0xff000ff0, 0x00000100, > > + mmSDMA1_CLK_CTRL, 0xff000ff0, 0x00000100, > > +}; > > + > > +static const u32 xgpu_tonga_golden_settings_a11[] = { > > + mmCB_HW_CONTROL, 0xfffdf3cf, 0x00007208, > > + mmCB_HW_CONTROL_3, 0x00000040, 0x00000040, > > + mmDB_DEBUG2, 0xf00fffff, 0x00000400, > > + mmDCI_CLK_CNTL, 0x00000080, 0x00000000, > > + mmFBC_DEBUG_COMP, 0x000000f0, 0x00000070, > > + mmFBC_MISC, 0x1f311fff, 0x12300000, > > + mmGB_GPU_ID, 0x0000000f, 0x00000000, > > + mmHDMI_CONTROL, 0x31000111, 0x00000011, > > + mmMC_ARB_WTM_GRPWT_RD, 0x00000003, 0x00000000, > > + mmMC_HUB_RDREQ_DMIF_LIMIT, 0x0000007f, 0x00000028, > > + mmMC_HUB_WDP_UMC, 0x00007fb6, 0x00000991, > > + mmPA_SC_ENHANCE, 0xffffffff, 0x20000001, > > + mmPA_SC_FIFO_DEPTH_CNTL, 0x000003ff, 0x000000fc, > > + mmPA_SC_LINE_STIPPLE_STATE, 0x0000ff0f, 0x00000000, > > + mmRLC_CGCG_CGLS_CTRL, 0x00000003, 0x0000003c, > > + mmSDMA0_CHICKEN_BITS, 0xfc910007, 0x00810007, > > + mmSDMA0_CLK_CTRL, 0xff000fff, 0x00000000, > > + mmSDMA0_GFX_IB_CNTL, 0x800f0111, 0x00000100, > > + mmSDMA0_RLC0_IB_CNTL, 0x800f0111, 0x00000100, > > + mmSDMA0_RLC1_IB_CNTL, 0x800f0111, 0x00000100, > > + mmSDMA1_CHICKEN_BITS, 0xfc910007, 0x00810007, > > + mmSDMA1_CLK_CTRL, 0xff000fff, 0x00000000, > > + mmSDMA1_GFX_IB_CNTL, 0x800f0111, 0x00000100, > > + mmSDMA1_RLC0_IB_CNTL, 0x800f0111, 0x00000100, > > + mmSDMA1_RLC1_IB_CNTL, 0x800f0111, 0x00000100, > > + mmSQ_RANDOM_WAVE_PRI, 0x001fffff, 0x000006fd, > > + mmTA_CNTL_AUX, 0x000f000f, 0x000b0000, > > + mmTCC_CTRL, 0x00100000, 0xf31fff7f, > > + mmTCC_EXE_DISABLE, 0x00000002, 0x00000002, > > + mmTCP_ADDR_CONFIG, 0x000003ff, 0x000002fb, > > + mmTCP_CHAN_STEER_HI, 0xffffffff, 0x0000543b, > > + mmTCP_CHAN_STEER_LO, 0xffffffff, 0xa9210876, > > + mmVGT_RESET_DEBUG, 0x00000004, 0x00000004, > > + mmVM_PRT_APERTURE0_LOW_ADDR, 0x0fffffff, 0x0fffffff, > > + mmVM_PRT_APERTURE1_LOW_ADDR, 0x0fffffff, 0x0fffffff, > > + mmVM_PRT_APERTURE2_LOW_ADDR, 0x0fffffff, 0x0fffffff, > > + mmVM_PRT_APERTURE3_LOW_ADDR, 0x0fffffff, 0x0fffffff, }; > > + > > +static const u32 xgpu_tonga_golden_common_all[] = { > > + mmGRBM_GFX_INDEX, 0xffffffff, 0xe0000000, > > + mmPA_SC_RASTER_CONFIG, 0xffffffff, 0x16000012, > > + mmPA_SC_RASTER_CONFIG_1, 0xffffffff, 0x0000002A, > > + mmGB_ADDR_CONFIG, 0xffffffff, 0x22011002, > > + mmSPI_RESOURCE_RESERVE_CU_0, 0xffffffff, 0x00000800, > > + mmSPI_RESOURCE_RESERVE_CU_1, 0xffffffff, 0x00000800, > > + mmSPI_RESOURCE_RESERVE_EN_CU_0, 0xffffffff, 0x00007FBF, }; > > + > > +void xgpu_vi_init_golden_registers(struct amdgpu_device *adev) { > > + switch (adev->asic_type) { > > + case CHIP_FIJI: > > + amdgpu_program_register_sequence(adev, > > + xgpu_fiji_mgcg_cgcg_init, > > + (const u32)ARRAY_SIZE( > > + xgpu_fiji_mgcg_cgcg_init)); > > + amdgpu_program_register_sequence(adev, > > + > > xgpu_fiji_golden_settings_a10, > > + (const u32)ARRAY_SIZE( > > + > > xgpu_fiji_golden_settings_a10)); > > + amdgpu_program_register_sequence(adev, > > + > > xgpu_fiji_golden_common_all, > > + (const u32)ARRAY_SIZE( > > + > > xgpu_fiji_golden_common_all)); > > + break; > > + case CHIP_TONGA: > > + amdgpu_program_register_sequence(adev, > > + xgpu_tonga_mgcg_cgcg_init, > > + (const u32)ARRAY_SIZE( > > + > > xgpu_tonga_mgcg_cgcg_init)); > > + amdgpu_program_register_sequence(adev, > > + > > xgpu_tonga_golden_settings_a11, > > + (const u32)ARRAY_SIZE( > > + > > xgpu_tonga_golden_settings_a11)); > > + amdgpu_program_register_sequence(adev, > > + > > xgpu_tonga_golden_common_all, > > + (const u32)ARRAY_SIZE( > > + > > xgpu_tonga_golden_common_all)); > > + break; > > + default: > > + BUG_ON("Doesn't support chip type.\n"); > > + break; > > + } > > +} > > + > > +/* > > + * Mailbox communication between GPU hypervisor and VFs */ static > > +void xgpu_vi_mailbox_send_ack(struct amdgpu_device *adev) { > > + u32 reg; > > + > > + reg = RREG32(mmMAILBOX_CONTROL); > > + reg = REG_SET_FIELD(reg, MAILBOX_CONTROL, RCV_MSG_ACK, 1); > > + WREG32(mmMAILBOX_CONTROL, reg); > > +} > > + > > +static void xgpu_vi_mailbox_set_valid(struct amdgpu_device *adev, > > +bool > > +val) { > > + u32 reg; > > + > > + reg = RREG32(mmMAILBOX_CONTROL); > > + reg = REG_SET_FIELD(reg, MAILBOX_CONTROL, > > + TRN_MSG_VALID, val ? 1 : 0); > > + WREG32(mmMAILBOX_CONTROL, reg); > > +} > > + > > +static void xgpu_vi_mailbox_trans_msg(struct amdgpu_device *adev, > > + enum idh_event event) > > +{ > > + u32 reg; > > + > > + reg = RREG32(mmMAILBOX_MSGBUF_TRN_DW0); > > + reg = REG_SET_FIELD(reg, MAILBOX_MSGBUF_TRN_DW0, > > + MSGBUF_DATA, event); > > + WREG32(mmMAILBOX_MSGBUF_TRN_DW0, reg); > > + > > + xgpu_vi_mailbox_set_valid(adev, true); } > > + > > +static int xgpu_vi_mailbox_rcv_msg(struct amdgpu_device *adev, > > + enum idh_event event) > > +{ > > + u32 reg; > > + > > + reg = RREG32(mmMAILBOX_MSGBUF_RCV_DW0); > > + if (reg != event) > > + return -ENOENT; > > + > > + /* send ack to PF */ > > + xgpu_vi_mailbox_send_ack(adev); > > + > > + return 0; > > +} > > + > > +static int xgpu_vi_poll_ack(struct amdgpu_device *adev) { > > + int r = 0, timeout = VI_MAILBOX_TIMEDOUT; > > + u32 mask = REG_FIELD_MASK(MAILBOX_CONTROL, TRN_MSG_ACK); > > + u32 reg; > > + > > + reg = RREG32(mmMAILBOX_CONTROL); > > + while (!(reg & mask)) { > > + if (timeout <= 0) { > > + pr_err("Doesn't get ack from pf.\n"); > > + r = -ETIME; > > + break; > > + } > > + msleep(1); > > + timeout -= 1; > > + > > + reg = RREG32(mmMAILBOX_CONTROL); > > + } > > + > > + return r; > > +} > > + > > +static int xgpu_vi_poll_msg(struct amdgpu_device *adev, enum > > +idh_event > > +event) { > > + int r = 0, timeout = VI_MAILBOX_TIMEDOUT; > > + > > + r = xgpu_vi_mailbox_rcv_msg(adev, event); > > + while (r) { > > + if (timeout <= 0) { > > + pr_err("Doesn't get ack from pf.\n"); > > + r = -ETIME; > > + break; > > + } > > + msleep(1); > > + timeout -= 1; > > + > > + r = xgpu_vi_mailbox_rcv_msg(adev, event); > > + } > > + > > + return r; > > +} > > + > > +static int xgpu_vi_send_access_requests(struct amdgpu_device *adev, > > + enum idh_request request) > > +{ > > + int r; > > + > > + xgpu_vi_mailbox_trans_msg(adev, request); > > + > > + /* start to poll ack */ > > + r = xgpu_vi_poll_ack(adev); > > + if (r) > > + return r; > > + > > + xgpu_vi_mailbox_set_valid(adev, false); > > + > > + /* start to check msg if request is idh_req_gpu_init_access */ > > + if (request == IDH_REQ_GPU_INIT_ACCESS) { > > + r = xgpu_vi_poll_msg(adev, IDH_READY_TO_ACCESS_GPU); > > + if (r) > > + return r; > > + } > > + > > + return 0; > > +} > > + > > +static int xgpu_vi_request_reset(struct amdgpu_device *adev) { > > + return xgpu_vi_send_access_requests(adev, > > IDH_REQ_GPU_RESET_ACCESS); } > > + > > +static int xgpu_vi_request_full_gpu_access(struct amdgpu_device *adev, > > + bool init) > > +{ > > + enum idh_event event; > > + > > + event = init ? IDH_REQ_GPU_INIT_ACCESS : > > IDH_REQ_GPU_FINI_ACCESS; > > + return xgpu_vi_send_access_requests(adev, event); } > > + > > +static int xgpu_vi_release_full_gpu_access(struct amdgpu_device *adev, > > + bool init) > > +{ > > + enum idh_event event; > > + int r = 0; > > + > > + event = init ? IDH_REL_GPU_INIT_ACCESS : > > IDH_REL_GPU_FINI_ACCESS; > > + r = xgpu_vi_send_access_requests(adev, event); > > + > > + return r; > > +} > > + > > +/* add support mailbox interrupts */ > > +static int xgpu_vi_mailbox_ack_irq(struct amdgpu_device *adev, > > + struct amdgpu_irq_src *source, > > + struct amdgpu_iv_entry *entry) { > > + DRM_DEBUG("get ack intr and do nothing.\n"); > > + return 0; > > +} > > + > > +static int xgpu_vi_set_mailbox_ack_irq(struct amdgpu_device *adev, > > + struct amdgpu_irq_src *src, > > + unsigned type, > > + enum amdgpu_interrupt_state state) { > > + u32 tmp = RREG32(mmMAILBOX_INT_CNTL); > > + > > + tmp = REG_SET_FIELD(tmp, MAILBOX_INT_CNTL, ACK_INT_EN, > > + (state == AMDGPU_IRQ_STATE_ENABLE) ? 1 : 0); > > + WREG32(mmMAILBOX_INT_CNTL, tmp); > > + > > + return 0; > > +} > > + > > +static void xgpu_vi_mailbox_flr_work(struct work_struct *work) { > > + struct amdgpu_virt *virt = container_of(work, > > + struct amdgpu_virt, flr_work.work); > > + struct amdgpu_device *adev = container_of(virt, > > + struct amdgpu_device, virt); > > + int r = 0; > > + > > + r = xgpu_vi_poll_msg(adev, IDH_FLR_NOTIFICATION_CMPL); > > + if (r) > > + DRM_ERROR("failed to get flr cmpl msg from hypervior.\n"); > > + > > + /* TODO: need to restore gfx states */ } > > + > > +static int xgpu_vi_set_mailbox_rcv_irq(struct amdgpu_device *adev, > > + struct amdgpu_irq_src *src, > > + unsigned type, > > + enum amdgpu_interrupt_state state) { > > + u32 tmp = RREG32(mmMAILBOX_INT_CNTL); > > + > > + tmp = REG_SET_FIELD(tmp, MAILBOX_INT_CNTL, VALID_INT_EN, > > + (state == AMDGPU_IRQ_STATE_ENABLE) ? 1 : 0); > > + WREG32(mmMAILBOX_INT_CNTL, tmp); > > + > > + return 0; > > +} > > + > > +static int xgpu_vi_mailbox_rcv_irq(struct amdgpu_device *adev, > > + struct amdgpu_irq_src *source, > > + struct amdgpu_iv_entry *entry) { > > + int r; > > + > > + adev->virt.caps &= ~AMDGPU_SRIOV_CAPS_RUNTIME; > > + r = xgpu_vi_mailbox_rcv_msg(adev, IDH_FLR_NOTIFICATION); > > + /* do nothing for other msg */ > > + if (r) > > + return 0; > > + > > + /* TODO: need to save gfx states */ > > + schedule_delayed_work(&adev->virt.flr_work, > > + msecs_to_jiffies(VI_MAILBOX_RESET_TIME)); > > + > > + return 0; > > +} > > + > > +static const struct amdgpu_irq_src_funcs > > +xgpu_vi_mailbox_ack_irq_funcs = > > { > > + .set = xgpu_vi_set_mailbox_ack_irq, > > + .process = xgpu_vi_mailbox_ack_irq, > > +}; > > + > > +static const struct amdgpu_irq_src_funcs > > +xgpu_vi_mailbox_rcv_irq_funcs = > > { > > + .set = xgpu_vi_set_mailbox_rcv_irq, > > + .process = xgpu_vi_mailbox_rcv_irq, > > +}; > > + > > +void xgpu_vi_mailbox_set_irq_funcs(struct amdgpu_device *adev) { > > + adev->virt.ack_irq.num_types = 1; > > + adev->virt.ack_irq.funcs = &xgpu_vi_mailbox_ack_irq_funcs; > > + adev->virt.rcv_irq.num_types = 1; > > + adev->virt.rcv_irq.funcs = &xgpu_vi_mailbox_rcv_irq_funcs; } > > + > > +int xgpu_vi_mailbox_add_irq_id(struct amdgpu_device *adev) { > > + int r; > > + > > + r = amdgpu_irq_add_id(adev, 135, &adev->virt.rcv_irq); > > + if (r) > > + return r; > > + > > + r = amdgpu_irq_add_id(adev, 138, &adev->virt.ack_irq); > > + if (r) { > > + amdgpu_irq_put(adev, &adev->virt.rcv_irq, 0); > > + return r; > > + } > > + > > + return 0; > > +} > > + > > +int xgpu_vi_mailbox_get_irq(struct amdgpu_device *adev) { > > + int r; > > + > > + r = amdgpu_irq_get(adev, &adev->virt.rcv_irq, 0); > > + if (r) > > + return r; > > + r = amdgpu_irq_get(adev, &adev->virt.ack_irq, 0); > > + if (r) { > > + amdgpu_irq_put(adev, &adev->virt.rcv_irq, 0); > > + return r; > > + } > > + > > + INIT_DELAYED_WORK(&adev->virt.flr_work, > > xgpu_vi_mailbox_flr_work); > > + > > + return 0; > > +} > > + > > +void xgpu_vi_mailbox_put_irq(struct amdgpu_device *adev) { > > + cancel_delayed_work_sync(&adev->virt.flr_work); > > + amdgpu_irq_put(adev, &adev->virt.ack_irq, 0); > > + amdgpu_irq_put(adev, &adev->virt.rcv_irq, 0); } > > + > > +const struct amdgpu_virt_ops xgpu_vi_virt_ops = { > > + .req_full_gpu = xgpu_vi_request_full_gpu_access, > > + .rel_full_gpu = xgpu_vi_release_full_gpu_access, > > + .reset_gpu = xgpu_vi_request_reset, > > +}; > > diff --git a/drivers/gpu/drm/amd/amdgpu/mxgpu_vi.h > > b/drivers/gpu/drm/amd/amdgpu/mxgpu_vi.h > > new file mode 100644 > > index 0000000..fd6216e > > --- /dev/null > > +++ b/drivers/gpu/drm/amd/amdgpu/mxgpu_vi.h > > @@ -0,0 +1,55 @@ > > +/* > > + * Copyright 2017 Advanced Micro Devices, Inc. > > + * > > + * Permission is hereby granted, free of charge, to any person > > +obtaining a > > + * copy of this software and associated documentation files (the > > +"Software"), > > + * to deal in the Software without restriction, including without > > +limitation > > + * the rights to use, copy, modify, merge, publish, distribute, > > +sublicense, > > + * and/or sell copies of the Software, and to permit persons to whom > > +the > > + * Software is furnished to do so, subject to the following conditions: > > + * > > + * The above copyright notice and this permission notice shall be > > +included in > > + * all copies or substantial portions of the Software. > > + * > > + * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY > KIND, > > +EXPRESS OR > > + * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF > > +MERCHANTABILITY, > > + * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN > NO > > EVENT > > +SHALL > > + * THE COPYRIGHT HOLDER(S) OR AUTHOR(S) BE LIABLE FOR ANY CLAIM, > > +DAMAGES OR > > + * OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR > > +OTHERWISE, > > + * ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR > > THE USE > > +OR > > + * OTHER DEALINGS IN THE SOFTWARE. > > + */ > > + > > +#ifndef __MXGPU_VI_H__ > > +#define __MXGPU_VI_H__ > > + > > +#define VI_MAILBOX_TIMEDOUT 150 > > +#define VI_MAILBOX_RESET_TIME 12 > > + > > +/* VI mailbox messages request */ > > +enum idh_request { > > + IDH_REQ_GPU_INIT_ACCESS = 1, > > + IDH_REL_GPU_INIT_ACCESS, > > + IDH_REQ_GPU_FINI_ACCESS, > > + IDH_REL_GPU_FINI_ACCESS, > > + IDH_REQ_GPU_RESET_ACCESS > > +}; > > + > > +/* VI mailbox messages data */ > > +enum idh_event { > > + IDH_CLR_MSG_BUF = 0, > > + IDH_READY_TO_ACCESS_GPU, > > + IDH_FLR_NOTIFICATION, > > + IDH_FLR_NOTIFICATION_CMPL, > > + IDH_EVENT_MAX > > +}; > > + > > +extern const struct amdgpu_virt_ops xgpu_vi_virt_ops; > > + > > +void xgpu_vi_init_golden_registers(struct amdgpu_device *adev); void > > +xgpu_vi_mailbox_set_irq_funcs(struct amdgpu_device *adev); int > > +xgpu_vi_mailbox_add_irq_id(struct amdgpu_device *adev); int > > +xgpu_vi_mailbox_get_irq(struct amdgpu_device *adev); void > > +xgpu_vi_mailbox_put_irq(struct amdgpu_device *adev); > > + > > +#endif > > -- > > 2.7.4