On Tue, Mar 4, 2025 at 8:57 AM Connor Abbott <cwabbott0@xxxxxxxxx> wrote: > > When things go wrong, the GPU is capable of quickly generating millions > of faulting translation requests per second. When that happens, in the > stall-on-fault model each access will stall until it wins the race to > signal the fault and then the RESUME register is written. This slows > processing page faults to a crawl as the GPU can generate faults much > faster than the CPU can acknowledge them. It also means that all > available resources in the SMMU are saturated waiting for the stalled > transactions, so that other transactions such as transactions generated > by the GMU, which shares a context bank with the GPU, cannot proceed. Nit, the GMU does not actually share a cb.. looking on x1e80100.dtsi, the GMU has cb 5 and gpu has 0 and 1. (Currently we just use the first, but I guess the 2nd would be used if we supported protected content?) fwiw, you can read this from dtsi, ie. in the GMU node, "iommus = <&adreno_smmu 5 0x0>;" > This causes a GMU watchdog timeout, which leads to a failed reset > because GX cannot collapse when there is a transaction pending and a > permanently hung GPU. > > On older platforms with qcom,smmu-v2, it seems that when one transaction > is stalled subsequent faulting transactions are terminated, which avoids > this problem, but the MMU-500 follows the spec here. > > To work around these problem, disable stall-on-fault as soon as we get a > page fault until a cooldown period after pagefaults stop. This allows > the GMU some guaranteed time to continue working. We only use > stall-on-fault to halt the GPU while we collect a devcoredump and we > always terminate the transaction afterward, so it's fine to miss some > subsequent page faults. We also keep it disabled so long as the current > devcoredump hasn't been deleted, because in that case we likely won't > capture another one if there's a fault. > > After this commit HFI messages still occasionally time out, because the > crashdump handler doesn't run fast enough to let the GMU resume, but the > driver seems to recover from it. This will probably go away after the > HFI timeout is increased. > > Signed-off-by: Connor Abbott <cwabbott0@xxxxxxxxx> > --- > drivers/gpu/drm/msm/adreno/a5xx_gpu.c | 2 ++ > drivers/gpu/drm/msm/adreno/a6xx_gpu.c | 4 ++++ > drivers/gpu/drm/msm/adreno/adreno_gpu.c | 42 ++++++++++++++++++++++++++++++++- > drivers/gpu/drm/msm/adreno/adreno_gpu.h | 24 +++++++++++++++++++ > drivers/gpu/drm/msm/msm_iommu.c | 9 +++++++ > drivers/gpu/drm/msm/msm_mmu.h | 1 + > 6 files changed, 81 insertions(+), 1 deletion(-) > > diff --git a/drivers/gpu/drm/msm/adreno/a5xx_gpu.c b/drivers/gpu/drm/msm/adreno/a5xx_gpu.c > index 71dca78cd7a5324e9ff5b14f173e2209fa42e196..670141531112c9d29cef8ef1fd51b74759fdd6d2 100644 > --- a/drivers/gpu/drm/msm/adreno/a5xx_gpu.c > +++ b/drivers/gpu/drm/msm/adreno/a5xx_gpu.c > @@ -131,6 +131,8 @@ static void a5xx_submit(struct msm_gpu *gpu, struct msm_gem_submit *submit) > struct msm_ringbuffer *ring = submit->ring; > unsigned int i, ibs = 0; > > + adreno_check_and_reenable_stall(adreno_gpu); > + > if (IS_ENABLED(CONFIG_DRM_MSM_GPU_SUDO) && submit->in_rb) { > ring->cur_ctx_seqno = 0; > a5xx_submit_in_rb(gpu, submit); > diff --git a/drivers/gpu/drm/msm/adreno/a6xx_gpu.c b/drivers/gpu/drm/msm/adreno/a6xx_gpu.c > index 0ae29a7c8a4d3f74236a35cc919f69d5c0a384a0..5a34cd2109a2d74c92841448a61ccb0d4f34e264 100644 > --- a/drivers/gpu/drm/msm/adreno/a6xx_gpu.c > +++ b/drivers/gpu/drm/msm/adreno/a6xx_gpu.c > @@ -212,6 +212,8 @@ static void a6xx_submit(struct msm_gpu *gpu, struct msm_gem_submit *submit) > struct msm_ringbuffer *ring = submit->ring; > unsigned int i, ibs = 0; > > + adreno_check_and_reenable_stall(adreno_gpu); > + > a6xx_set_pagetable(a6xx_gpu, ring, submit); > > get_stats_counter(ring, REG_A6XX_RBBM_PERFCTR_CP(0), > @@ -335,6 +337,8 @@ static void a7xx_submit(struct msm_gpu *gpu, struct msm_gem_submit *submit) > struct msm_ringbuffer *ring = submit->ring; > unsigned int i, ibs = 0; > > + adreno_check_and_reenable_stall(adreno_gpu); > + > /* > * Toggle concurrent binning for pagetable switch and set the thread to > * BR since only it can execute the pagetable switch packets. > diff --git a/drivers/gpu/drm/msm/adreno/adreno_gpu.c b/drivers/gpu/drm/msm/adreno/adreno_gpu.c > index 1238f326597808eb28b4c6822cbd41a26e555eb9..bac586101dc0494f46b069a8440a45825dfe9b5e 100644 > --- a/drivers/gpu/drm/msm/adreno/adreno_gpu.c > +++ b/drivers/gpu/drm/msm/adreno/adreno_gpu.c > @@ -246,16 +246,53 @@ u64 adreno_private_address_space_size(struct msm_gpu *gpu) > return SZ_4G; > } > > +void adreno_check_and_reenable_stall(struct adreno_gpu *adreno_gpu) > +{ > + struct msm_gpu *gpu = &adreno_gpu->base; > + unsigned long flags; > + > + /* > + * Wait until the cooldown period has passed and we would actually > + * collect a crashdump to re-enable stall-on-fault. > + */ > + spin_lock_irqsave(&adreno_gpu->fault_stall_lock, flags); > + if (!adreno_gpu->stall_enabled && > + ktime_after(ktime_get(), adreno_gpu->stall_reenable_time) && > + !READ_ONCE(gpu->crashstate)) { > + adreno_gpu->stall_enabled = true; > + > + gpu->aspace->mmu->funcs->set_stall(gpu->aspace->mmu, true); > + } > + spin_unlock_irqrestore(&adreno_gpu->fault_stall_lock, flags); > +} > + > #define ARM_SMMU_FSR_TF BIT(1) > #define ARM_SMMU_FSR_PF BIT(3) > #define ARM_SMMU_FSR_EF BIT(4) > +#define ARM_SMMU_FSR_SS BIT(30) > > int adreno_fault_handler(struct msm_gpu *gpu, unsigned long iova, int flags, > struct adreno_smmu_fault_info *info, const char *block, > u32 scratch[4]) > { > + struct adreno_gpu *adreno_gpu = to_adreno_gpu(gpu); > const char *type = "UNKNOWN"; > - bool do_devcoredump = info && !READ_ONCE(gpu->crashstate); > + bool do_devcoredump = info && (info->fsr & ARM_SMMU_FSR_SS) && > + !READ_ONCE(gpu->crashstate); > + unsigned long irq_flags; > + > + /* > + * In case there is a subsequent storm of pagefaults, disable > + * stall-on-fault for at least half a second. > + */ > + spin_lock_irqsave(&adreno_gpu->fault_stall_lock, irq_flags); > + if (adreno_gpu->stall_enabled) { > + adreno_gpu->stall_enabled = false; > + > + gpu->aspace->mmu->funcs->set_stall(gpu->aspace->mmu, false); > + } > + adreno_gpu->stall_reenable_time = ktime_add_ms(ktime_get(), 500); > + spin_unlock_irqrestore(&adreno_gpu->fault_stall_lock, irq_flags); > > /* > * If we aren't going to be resuming later from fault_worker, then do > @@ -1143,6 +1180,9 @@ int adreno_gpu_init(struct drm_device *drm, struct platform_device *pdev, > adreno_gpu->info->inactive_period); > pm_runtime_use_autosuspend(dev); > > + spin_lock_init(&adreno_gpu->fault_stall_lock); > + adreno_gpu->stall_enabled = true; > + > return msm_gpu_init(drm, pdev, &adreno_gpu->base, &funcs->base, > gpu_name, &adreno_gpu_config); > } > diff --git a/drivers/gpu/drm/msm/adreno/adreno_gpu.h b/drivers/gpu/drm/msm/adreno/adreno_gpu.h > index dcf454629ce037b2a8274a6699674ad754ce1f07..a528036b46216bd898f6d48c5fb0555c4c4b053b 100644 > --- a/drivers/gpu/drm/msm/adreno/adreno_gpu.h > +++ b/drivers/gpu/drm/msm/adreno/adreno_gpu.h > @@ -205,6 +205,28 @@ struct adreno_gpu { > /* firmware: */ > const struct firmware *fw[ADRENO_FW_MAX]; > > + /** > + * fault_stall_lock: nit, @fault_stall_lock: And for fault_stall_reenable_time/stall_enabled, it wouldn't hurt to add something along the lines of "Protected by @fault_stall_lock". I've been slowly trying to improve the comment docs over time, I have some of that in my vmbind patchset. Anyways, with those nits addressed, r-b BR, -R > + * > + * Serialize changes to stall-on-fault state. > + */ > + spinlock_t fault_stall_lock; > + > + /** > + * fault_stall_reenable_time: > + * > + * if stall_enabled is false, when to reenable stall-on-fault. > + */ > + ktime_t stall_reenable_time; > + > + /** > + * stall_enabled: > + * > + * Whether stall-on-fault is currently enabled. > + */ > + bool stall_enabled; > + > + > struct { > /** > * @rgb565_predicator: Unknown, introduced with A650 family, > @@ -629,6 +651,8 @@ int adreno_fault_handler(struct msm_gpu *gpu, unsigned long iova, int flags, > struct adreno_smmu_fault_info *info, const char *block, > u32 scratch[4]); > > +void adreno_check_and_reenable_stall(struct adreno_gpu *gpu); > + > int adreno_read_speedbin(struct device *dev, u32 *speedbin); > > /* > diff --git a/drivers/gpu/drm/msm/msm_iommu.c b/drivers/gpu/drm/msm/msm_iommu.c > index 2a94e82316f95c5f9dcc37ef0a4664a29e3492b2..8d5380e6dcc217c7c209b51527bf15748b3ada71 100644 > --- a/drivers/gpu/drm/msm/msm_iommu.c > +++ b/drivers/gpu/drm/msm/msm_iommu.c > @@ -351,6 +351,14 @@ static void msm_iommu_resume_translation(struct msm_mmu *mmu) > adreno_smmu->resume_translation(adreno_smmu->cookie, true); > } > > +static void msm_iommu_set_stall(struct msm_mmu *mmu, bool enable) > +{ > + struct adreno_smmu_priv *adreno_smmu = dev_get_drvdata(mmu->dev); > + > + if (adreno_smmu->set_stall) > + adreno_smmu->set_stall(adreno_smmu->cookie, enable); > +} > + > static void msm_iommu_detach(struct msm_mmu *mmu) > { > struct msm_iommu *iommu = to_msm_iommu(mmu); > @@ -399,6 +407,7 @@ static const struct msm_mmu_funcs funcs = { > .unmap = msm_iommu_unmap, > .destroy = msm_iommu_destroy, > .resume_translation = msm_iommu_resume_translation, > + .set_stall = msm_iommu_set_stall, > }; > > struct msm_mmu *msm_iommu_new(struct device *dev, unsigned long quirks) > diff --git a/drivers/gpu/drm/msm/msm_mmu.h b/drivers/gpu/drm/msm/msm_mmu.h > index 88af4f490881f2a6789ae2d03e1c02d10046331a..2694a356a17904e7572b767b16ed0cee806406cf 100644 > --- a/drivers/gpu/drm/msm/msm_mmu.h > +++ b/drivers/gpu/drm/msm/msm_mmu.h > @@ -16,6 +16,7 @@ struct msm_mmu_funcs { > int (*unmap)(struct msm_mmu *mmu, uint64_t iova, size_t len); > void (*destroy)(struct msm_mmu *mmu); > void (*resume_translation)(struct msm_mmu *mmu); > + void (*set_stall)(struct msm_mmu *mmu, bool enable); > }; > > enum msm_mmu_type { > > -- > 2.47.1 >