> -----Original Message----- > From: amd-gfx [mailto:amd-gfx-bounces at lists.freedesktop.org] On Behalf > Of Monk Liu > Sent: Tuesday, February 07, 2017 1:11 AM > To: amd-gfx at lists.freedesktop.org > Cc: Liu, Monk > Subject: [PATCH 20/20] drm/amdgpu:fix kiq_resume routine > > use is_load_stage to fix compute ring test failure issue > which occured after FLR/gpu_reset. > > we need backup a clean status of MQD which was created in drv load > stage, and use it in resume stage, otherwise KCQ and KIQ all may > faild in ring/ib test. > > Change-Id: I41be940454a6638e9a8a05f096601eaa1fbebaab > Signed-off-by: Monk Liu <Monk.Liu at amd.com> Please see my comments on patch 17. > --- > drivers/gpu/drm/amd/amdgpu/gfx_v8_0.c | 44 > ++++++++++++++++++++++++++--------- > 1 file changed, 33 insertions(+), 11 deletions(-) > > diff --git a/drivers/gpu/drm/amd/amdgpu/gfx_v8_0.c > b/drivers/gpu/drm/amd/amdgpu/gfx_v8_0.c > index 0ce00ff..4a641d3 100644 > --- a/drivers/gpu/drm/amd/amdgpu/gfx_v8_0.c > +++ b/drivers/gpu/drm/amd/amdgpu/gfx_v8_0.c > @@ -4877,24 +4877,46 @@ static int gfx_v8_0_kiq_init_queue(struct > amdgpu_ring *ring, > struct amdgpu_kiq *kiq = &adev->gfx.kiq; > uint64_t eop_gpu_addr; > bool is_kiq = (ring->funcs->type == AMDGPU_RING_TYPE_KIQ); > + int mqd_idx = AMDGPU_MAX_COMPUTE_RINGS; > > if (is_kiq) { > eop_gpu_addr = kiq->eop_gpu_addr; > gfx_v8_0_kiq_setting(&kiq->ring); > - } else > + } else { > eop_gpu_addr = adev->gfx.mec.hpd_eop_gpu_addr + > ring->queue * MEC_HPD_SIZE; > + mqd_idx = ring - &adev->gfx.compute_ring[0]; > + } > > - mutex_lock(&adev->srbm_mutex); > - vi_srbm_select(adev, ring->me, ring->pipe, ring->queue, 0); > + if (adev->is_load_stage) { > + memset((void *)mqd, 0, sizeof(*mqd)); > + mutex_lock(&adev->srbm_mutex); > + vi_srbm_select(adev, ring->me, ring->pipe, ring->queue, 0); > + gfx_v8_0_mqd_init(adev, mqd, mqd_gpu_addr, > eop_gpu_addr, ring); > + if (is_kiq) > + gfx_v8_0_kiq_init_register(adev, mqd, ring); > + vi_srbm_select(adev, 0, 0, 0, 0); > + mutex_unlock(&adev->srbm_mutex); > > - gfx_v8_0_mqd_init(adev, mqd, mqd_gpu_addr, eop_gpu_addr, > ring); > + if (adev->gfx.mec.mqd_backup[mqd_idx]) > + memcpy(adev->gfx.mec.mqd_backup[mqd_idx], > mqd, sizeof(*mqd)); > + } else { /* for GPU_RESET case */ > + /* reset MQD to a clean status */ > + if (adev->gfx.mec.mqd_backup[mqd_idx]) > + memcpy(mqd, adev- > >gfx.mec.mqd_backup[mqd_idx], sizeof(*mqd)); > > - if (is_kiq) > - gfx_v8_0_kiq_init_register(adev, mqd, ring); > - > - vi_srbm_select(adev, 0, 0, 0, 0); > - mutex_unlock(&adev->srbm_mutex); > + /* reset ring buffer */ > + ring->wptr = 0; > + amdgpu_ring_clear_ring(ring); > + > + if (is_kiq) { > + mutex_lock(&adev->srbm_mutex); > + vi_srbm_select(adev, ring->me, ring->pipe, ring->queue, > 0); > + gfx_v8_0_kiq_init_register(adev, mqd, ring); > + vi_srbm_select(adev, 0, 0, 0, 0); > + mutex_unlock(&adev->srbm_mutex); > + } > + } > > if (is_kiq) > gfx_v8_0_kiq_enable(ring); > @@ -4913,9 +4935,9 @@ static int gfx_v8_0_kiq_resume(struct > amdgpu_device *adev) > > ring = &adev->gfx.kiq.ring; > if (!amdgpu_bo_kmap(ring->mqd_obj, (void **)&ring->mqd_ptr)) { > - memset((void *)ring->mqd_ptr, 0, sizeof(struct vi_mqd)); > r = gfx_v8_0_kiq_init_queue(ring, ring->mqd_ptr, ring- > >mqd_gpu_addr); > amdgpu_bo_kunmap(ring->mqd_obj); > + ring->mqd_ptr = NULL; > if (r) > return r; > } else { > @@ -4925,9 +4947,9 @@ static int gfx_v8_0_kiq_resume(struct > amdgpu_device *adev) > for (i = 0; i < adev->gfx.num_compute_rings; i++) { > ring = &adev->gfx.compute_ring[i]; > if (!amdgpu_bo_kmap(ring->mqd_obj, (void **)&ring- > >mqd_ptr)) { > - memset((void *)ring->mqd_ptr, 0, sizeof(struct > vi_mqd)); > r = gfx_v8_0_kiq_init_queue(ring, ring->mqd_ptr, > ring->mqd_gpu_addr); > amdgpu_bo_kunmap(ring->mqd_obj); > + ring->mqd_ptr = NULL; > if (r) > return r; > } else { > -- > 2.7.4 > > _______________________________________________ > amd-gfx mailing list > amd-gfx at lists.freedesktop.org > https://lists.freedesktop.org/mailman/listinfo/amd-gfx