On 1/8/2025 07:59, Jiang Liu wrote:Subject: [RFC PATCH 00/13] Enhance device state machine to better support suspend/resume
I'm not sure how this happened, but your subject didn't end up in the subject of the thread on patch 0 so the thread just looks like an unsubjected thread.
Maybe it’s caused by one extra blank line at the header. Recently we were testing suspend/resume functionality with AMD GPUs, we have encountered several resource tracking related bugs, such as double buffer free, use after free and unbalanced irq reference count.
Can you share more aobut how you were hitting these issues? Are they specific to S3 or to s2idle flows? dGPU or APU?Are they only with SRIOV?Is there anything to do with the host influencing the failures to happen, or are you contriving the failures to find the bugs?I know we've had some reports about resource tracking warnings on the reset flows, but I haven't heard much about suspend/resume.
We are investigating to develop some advanced product features based on amdgpu suspend/resume. So we started by tested the suspend/resume functionality of AMD 308x GPUs with the following simple script: ``` echo platform > /sys/power/pm_test i=0 while true; do echo mem > /sys/power/state let i=i+1 echo $i sleep 1 done ```
It succeeds with the first and second iteration but always fails on following iterations on a bare metal servers with eight MI308X GPUs.
During analyze and root-cause the failure, we have encountered several crashes, resource leakages and false alarms.
With sriov in single VF mode, resume always fails. Seems some contexts/vram buffers get lost during suspend and haven’t be restored on resume, so cause failure. We haven’t tested sriov in multiple VFs mode yet. We need more help from AMD side to make SR work for SRIOV:) We have tried to solve these issues case by case, but found that may not be the right way. Especially about the unbalanced irq reference count, there will be new issues appear once we fixed the current known issues. After analyzing related source code, we found that there may be some fundamental implementaion flaws behind these resource tracking
implementationissues. The amdgpu driver has two major state machines to driver the device management flow, one is for ip blocks, the other is for ras blocks. The hook points defined in struct amd_ip_funcs for device setup/teardown are symmetric, but the implementation is asymmetric, sometime even ambiguous. The most obvious two issues we noticed are: 1) amdgpu_irq_get() are called from .late_init() but amdgpu_irq_put() are called from .hw_fini() instead of .early_fini(). 2) the way to reset ip_bloc.status.valid/sw/hw/late_initialized doesn't match the way to set those flags. When taking device suspend/resume into account, in addition to device probe/remove, things get much more complex. Some issues arise because many suspend/resume implementations directly reuse .hw_init/.hw_fini/ .late_init hook points.
So we try to fix those issues by two enhancements/refinements to current device management state machines. The first change is to make the ip block state machine and associated status flags work in stack-like way as below: Callback Status Flags early_init: valid = true sw_init: sw = true hw_init: hw = true late_init: late_initialized = true early_fini: late_initialized = false hw_fini: hw = false sw_fini: sw = false late_fini: valid = false
At a high level this makes sense to me, but I'd just call 'late' or 'late_init'.Another idea if you make it stack like is to do it as a true enum for the state machine and store it all in one variable.
I will add a patch to convert those bool flags into an enum. Thanks, Gerry Also do the same thing for ras block state machine, though it's much more simpler. The second change is fine tune the overall device management work flow as below: 1. amdgpu_driver_load_kms() amdgpu_device_init() amdgpu_device_ip_early_init() ip_blocks[i].early_init() ip_blocks[i].status.valid = true amdgpu_device_ip_init() amdgpu_ras_init() ip_blocks[i].sw_init() ip_blocks[i].status.sw = true ip_blocks[i].hw_init() ip_blocks[i].status.hw = true amdgpu_device_ip_late_init() ip_blocks[i].late_init() ip_blocks[i].status.late_initialized = true amdgpu_ras_late_init() ras_blocks[i].ras_late_init() amdgpu_ras_feature_enable_on_boot() 2. amdgpu_pmops_suspend()/amdgpu_pmops_freeze()/amdgpu_pmops_poweroff() amdgpu_device_suspend() amdgpu_ras_early_fini() ras_blocks[i].ras_early_fini() amdgpu_ras_feature_disable() amdgpu_ras_suspend() amdgpu_ras_disable_all_features() +++ ip_blocks[i].early_fini() +++ ip_blocks[i].status.late_initialized = false ip_blocks[i].suspend() 3. amdgpu_pmops_resume()/amdgpu_pmops_thaw()/amdgpu_pmops_restore() amdgpu_device_resume() amdgpu_device_ip_resume() ip_blocks[i].resume() amdgpu_device_ip_late_init() ip_blocks[i].late_init() ip_blocks[i].status.late_initialized = true amdgpu_ras_late_init() ras_blocks[i].ras_late_init() amdgpu_ras_feature_enable_on_boot() amdgpu_ras_resume() amdgpu_ras_enable_all_features() 4. amdgpu_driver_unload_kms() amdgpu_device_fini_hw() amdgpu_ras_early_fini() ras_blocks[i].ras_early_fini() +++ ip_blocks[i].early_fini() +++ ip_blocks[i].status.late_initialized = false ip_blocks[i].hw_fini() ip_blocks[i].status.hw = false 5. amdgpu_driver_release_kms() amdgpu_device_fini_sw() amdgpu_device_ip_fini() ip_blocks[i].sw_fini() ip_blocks[i].status.sw = false --- ip_blocks[i].status.valid = false +++ amdgpu_ras_fini() ip_blocks[i].late_fini() +++ ip_blocks[i].status.valid = false --- ip_blocks[i].status.late_initialized = false --- amdgpu_ras_fini() The main changes include: 1) invoke ip_blocks[i].early_fini in amdgpu_pmops_suspend(). Currently there's only one ip block which provides `early_fini` callback. We have add a check of `in_s3` to keep current behavior in function amdgpu_dm_early_fini(). So there should be no functional changes. 2) set ip_blocks[i].status.late_initialized to false after calling callback `early_fini`. We have auditted all usages of the late_initialized flag and no functional changes found. 3) only set ip_blocks[i].status.valid = false after calling the `late_fini` callback. 4) call amdgpu_ras_fini() before invoking ip_blocks[i].late_fini. Then we try to refine each subsystem, such as nbio, asic, gfx, gmc, ras etc, to follow the new design. Currently we have only taken the nbio and asic as examples to show the proposed changes. Once we have confirmed that's the right way to go, we will handle the lefting subsystems. This is in early stage and requesting for comments, any comments and suggestions are welcomed! Jiang Liu (13): amdgpu: wrong array index to get ip block for PSP drm/admgpu: add helper functions to track status for ras manager drm/amdgpu: add a flag to track ras debugfs creation status drm/amdgpu: free all resources on error recovery path of amdgpu_ras_init() drm/amdgpu: introduce a flag to track refcount held for features drm/amdgpu: enhance amdgpu_ras_block_late_fini() drm/amdgpu: enhance amdgpu_ras_pre_fini() to better support SR drm/admgpu: rename amdgpu_ras_pre_fini() to amdgpu_ras_early_fini() drm/amdgpu: make IP block state machine works in stack like way drm/admgpu: make device state machine work in stack like way drm/amdgpu/sdma: improve the way to manage irq reference count drm/amdgpu/nbio: improve the way to manage irq reference count drm/amdgpu/asic: make ip block operations symmetric by .early_fini() drivers/gpu/drm/amd/amdgpu/amdgpu.h | 40 +++++ drivers/gpu/drm/amd/amdgpu/amdgpu_device.c | 37 ++++- drivers/gpu/drm/amd/amdgpu/amdgpu_gfx.c | 2 +- drivers/gpu/drm/amd/amdgpu/amdgpu_jpeg.c | 2 +- drivers/gpu/drm/amd/amdgpu/amdgpu_nbio.c | 16 +- drivers/gpu/drm/amd/amdgpu/amdgpu_nbio.h | 1 + drivers/gpu/drm/amd/amdgpu/amdgpu_psp.c | 8 +- drivers/gpu/drm/amd/amdgpu/amdgpu_ras.c | 144 +++++++++++++----- drivers/gpu/drm/amd/amdgpu/amdgpu_ras.h | 16 +- drivers/gpu/drm/amd/amdgpu/amdgpu_sdma.c | 26 +++- drivers/gpu/drm/amd/amdgpu/amdgpu_sdma.h | 2 + drivers/gpu/drm/amd/amdgpu/amdgpu_umc.c | 2 +- drivers/gpu/drm/amd/amdgpu/amdgpu_vcn.c | 2 +- drivers/gpu/drm/amd/amdgpu/amdgpu_xgmi.c | 2 +- drivers/gpu/drm/amd/amdgpu/gfx_v9_4_3.c | 2 +- drivers/gpu/drm/amd/amdgpu/mmhub_v1_8.c | 2 +- drivers/gpu/drm/amd/amdgpu/nbio_v7_4.c | 1 + drivers/gpu/drm/amd/amdgpu/nbio_v7_9.c | 1 + drivers/gpu/drm/amd/amdgpu/nv.c | 14 +- drivers/gpu/drm/amd/amdgpu/sdma_v4_0.c | 8 - drivers/gpu/drm/amd/amdgpu/sdma_v4_4_2.c | 23 +-- drivers/gpu/drm/amd/amdgpu/soc15.c | 38 ++--- drivers/gpu/drm/amd/amdgpu/soc21.c | 35 +++-- drivers/gpu/drm/amd/amdgpu/soc24.c | 17 ++- .../gpu/drm/amd/display/amdgpu_dm/amdgpu_dm.c | 3 + 25 files changed, 326 insertions(+), 118 deletions(-)
|