On 11/21/2024 7:58 PM, Alex Deucher wrote: > On Wed, Nov 20, 2024 at 11:57 PM Lazar, Lijo <lijo.lazar@xxxxxxx> wrote: >> >> >> >> On 11/20/2024 11:58 PM, Alex Deucher wrote: >>> smu->workload_mask is IP specific and should not be messed with in >>> the common code. The mask bits vary across SMU versions. >>> >>> Move all handling of smu->workload_mask in to the backends and >>> simplify the code. Store the user's preference in smu->power_profile_mode >>> which will be reflected in sysfs. For internal driver profile >>> switches for KFD or VCN, just update the workload mask so that the >>> user's preference is retained. Remove all of the extra now unused >>> workload related elements in the smu structure. >>> >>> v2: use refcounts for workload profiles >>> v3: rework based on feedback from Lijo >>> v4: fix the refcount on failure, drop backend mask >>> v5: rework custom handling >>> v6: handle failure cleanup with custom profile >>> >>> Signed-off-by: Alex Deucher <alexander.deucher@xxxxxxx> >>> Cc: Kenneth Feng <kenneth.feng@xxxxxxx> >>> Cc: Lijo Lazar <lijo.lazar@xxxxxxx> >>> --- >>> drivers/gpu/drm/amd/pm/swsmu/amdgpu_smu.c | 150 ++++++++++------ >>> drivers/gpu/drm/amd/pm/swsmu/inc/amdgpu_smu.h | 15 +- >>> .../gpu/drm/amd/pm/swsmu/smu11/arcturus_ppt.c | 166 +++++++++-------- >>> .../gpu/drm/amd/pm/swsmu/smu11/navi10_ppt.c | 167 ++++++++++------- >>> .../amd/pm/swsmu/smu11/sienna_cichlid_ppt.c | 168 +++++++++++------- >>> .../gpu/drm/amd/pm/swsmu/smu11/vangogh_ppt.c | 41 ++--- >>> .../gpu/drm/amd/pm/swsmu/smu12/renoir_ppt.c | 43 ++--- >>> .../drm/amd/pm/swsmu/smu13/smu_v13_0_0_ppt.c | 167 +++++++++-------- >>> .../drm/amd/pm/swsmu/smu13/smu_v13_0_7_ppt.c | 138 ++++++++------ >>> .../drm/amd/pm/swsmu/smu14/smu_v14_0_2_ppt.c | 168 +++++++++++------- >>> drivers/gpu/drm/amd/pm/swsmu/smu_cmn.c | 25 +++ >>> drivers/gpu/drm/amd/pm/swsmu/smu_cmn.h | 4 + >>> 12 files changed, 736 insertions(+), 516 deletions(-) >>> >>> diff --git a/drivers/gpu/drm/amd/pm/swsmu/amdgpu_smu.c b/drivers/gpu/drm/amd/pm/swsmu/amdgpu_smu.c >>> index f99fe2508852..acaa1530c25c 100644 >>> --- a/drivers/gpu/drm/amd/pm/swsmu/amdgpu_smu.c >>> +++ b/drivers/gpu/drm/amd/pm/swsmu/amdgpu_smu.c >>> @@ -72,6 +72,10 @@ static int smu_set_power_limit(void *handle, uint32_t limit); >>> static int smu_set_fan_speed_rpm(void *handle, uint32_t speed); >>> static int smu_set_gfx_cgpg(struct smu_context *smu, bool enabled); >>> static int smu_set_mp1_state(void *handle, enum pp_mp1_state mp1_state); >>> +static void smu_power_profile_mode_get(struct smu_context *smu, >>> + enum PP_SMC_POWER_PROFILE profile_mode); >>> +static void smu_power_profile_mode_put(struct smu_context *smu, >>> + enum PP_SMC_POWER_PROFILE profile_mode); >>> >>> static int smu_sys_get_pp_feature_mask(void *handle, >>> char *buf) >>> @@ -1268,8 +1272,6 @@ static int smu_sw_init(struct amdgpu_ip_block *ip_block) >>> INIT_WORK(&smu->interrupt_work, smu_interrupt_work_fn); >>> atomic64_set(&smu->throttle_int_counter, 0); >>> smu->watermarks_bitmap = 0; >>> - smu->power_profile_mode = PP_SMC_POWER_PROFILE_BOOTUP_DEFAULT; >>> - smu->default_power_profile_mode = PP_SMC_POWER_PROFILE_BOOTUP_DEFAULT; >>> >>> for (i = 0; i < adev->vcn.num_vcn_inst; i++) >>> atomic_set(&smu->smu_power.power_gate.vcn_gated[i], 1); >>> @@ -1277,27 +1279,13 @@ static int smu_sw_init(struct amdgpu_ip_block *ip_block) >>> atomic_set(&smu->smu_power.power_gate.vpe_gated, 1); >>> atomic_set(&smu->smu_power.power_gate.umsch_mm_gated, 1); >>> >>> - smu->workload_prority[PP_SMC_POWER_PROFILE_BOOTUP_DEFAULT] = 0; >>> - smu->workload_prority[PP_SMC_POWER_PROFILE_FULLSCREEN3D] = 1; >>> - smu->workload_prority[PP_SMC_POWER_PROFILE_POWERSAVING] = 2; >>> - smu->workload_prority[PP_SMC_POWER_PROFILE_VIDEO] = 3; >>> - smu->workload_prority[PP_SMC_POWER_PROFILE_VR] = 4; >>> - smu->workload_prority[PP_SMC_POWER_PROFILE_COMPUTE] = 5; >>> - smu->workload_prority[PP_SMC_POWER_PROFILE_CUSTOM] = 6; >>> - >>> if (smu->is_apu || >>> !smu_is_workload_profile_available(smu, PP_SMC_POWER_PROFILE_FULLSCREEN3D)) >>> - smu->workload_mask = 1 << smu->workload_prority[PP_SMC_POWER_PROFILE_BOOTUP_DEFAULT]; >>> + smu->power_profile_mode = PP_SMC_POWER_PROFILE_BOOTUP_DEFAULT; >>> else >>> - smu->workload_mask = 1 << smu->workload_prority[PP_SMC_POWER_PROFILE_FULLSCREEN3D]; >>> - >>> - smu->workload_setting[0] = PP_SMC_POWER_PROFILE_BOOTUP_DEFAULT; >>> - smu->workload_setting[1] = PP_SMC_POWER_PROFILE_FULLSCREEN3D; >>> - smu->workload_setting[2] = PP_SMC_POWER_PROFILE_POWERSAVING; >>> - smu->workload_setting[3] = PP_SMC_POWER_PROFILE_VIDEO; >>> - smu->workload_setting[4] = PP_SMC_POWER_PROFILE_VR; >>> - smu->workload_setting[5] = PP_SMC_POWER_PROFILE_COMPUTE; >>> - smu->workload_setting[6] = PP_SMC_POWER_PROFILE_CUSTOM; >>> + smu->power_profile_mode = PP_SMC_POWER_PROFILE_FULLSCREEN3D; >>> + smu_power_profile_mode_get(smu, smu->power_profile_mode); >>> + >>> smu->display_config = &adev->pm.pm_display_cfg; >>> >>> smu->smu_dpm.dpm_level = AMD_DPM_FORCED_LEVEL_AUTO; >>> @@ -1350,6 +1338,11 @@ static int smu_sw_fini(struct amdgpu_ip_block *ip_block) >>> return ret; >>> } >>> >>> + if (smu->custom_profile_params) { >>> + kfree(smu->custom_profile_params); >>> + smu->custom_profile_params = NULL; >>> + } >>> + >>> smu_fini_microcode(smu); >>> >>> return 0; >>> @@ -2137,6 +2130,9 @@ static int smu_suspend(struct amdgpu_ip_block *ip_block) >>> if (!ret) >>> adev->gfx.gfx_off_entrycount = count; >>> >>> + /* clear this on suspend so it will get reprogrammed on resume */ >>> + smu->workload_mask = 0; >>> + >>> return 0; >>> } >>> >>> @@ -2249,25 +2245,49 @@ static int smu_enable_umd_pstate(void *handle, >>> } >>> >>> static int smu_bump_power_profile_mode(struct smu_context *smu, >>> - long *param, >>> - uint32_t param_size) >>> + long *custom_params, >>> + u32 custom_params_max_idx) >>> { >>> - int ret = 0; >>> + u32 workload_mask = 0; >>> + int i, ret = 0; >>> + >>> + for (i = 0; i < PP_SMC_POWER_PROFILE_COUNT; i++) { >>> + if (smu->workload_refcount[i]) >>> + workload_mask |= 1 << i; >>> + } >>> + >>> + if (smu->workload_mask == workload_mask) >>> + return 0; >>> >>> if (smu->ppt_funcs->set_power_profile_mode) >>> - ret = smu->ppt_funcs->set_power_profile_mode(smu, param, param_size); >>> + ret = smu->ppt_funcs->set_power_profile_mode(smu, workload_mask, >>> + custom_params, >>> + custom_params_max_idx); >>> + >>> + if (!ret) >>> + smu->workload_mask = workload_mask; >>> >>> return ret; >>> } >>> >>> +static void smu_power_profile_mode_get(struct smu_context *smu, >>> + enum PP_SMC_POWER_PROFILE profile_mode) >>> +{ >>> + smu->workload_refcount[profile_mode]++; >>> +} >>> + >>> +static void smu_power_profile_mode_put(struct smu_context *smu, >>> + enum PP_SMC_POWER_PROFILE profile_mode) >>> +{ >>> + if (smu->workload_refcount[profile_mode]) >>> + smu->workload_refcount[profile_mode]--; >>> +} >>> + >>> static int smu_adjust_power_state_dynamic(struct smu_context *smu, >>> enum amd_dpm_forced_level level, >>> - bool skip_display_settings, >>> - bool init) >>> + bool skip_display_settings) >>> { >>> int ret = 0; >>> - int index = 0; >>> - long workload[1]; >>> struct smu_dpm_context *smu_dpm_ctx = &(smu->smu_dpm); >>> >>> if (!skip_display_settings) { >>> @@ -2304,14 +2324,8 @@ static int smu_adjust_power_state_dynamic(struct smu_context *smu, >>> } >>> >>> if (smu_dpm_ctx->dpm_level != AMD_DPM_FORCED_LEVEL_MANUAL && >>> - smu_dpm_ctx->dpm_level != AMD_DPM_FORCED_LEVEL_PERF_DETERMINISM) { >>> - index = fls(smu->workload_mask); >>> - index = index > 0 && index <= WORKLOAD_POLICY_MAX ? index - 1 : 0; >>> - workload[0] = smu->workload_setting[index]; >>> - >>> - if (init || smu->power_profile_mode != workload[0]) >>> - smu_bump_power_profile_mode(smu, workload, 0); >>> - } >>> + smu_dpm_ctx->dpm_level != AMD_DPM_FORCED_LEVEL_PERF_DETERMINISM) >>> + smu_bump_power_profile_mode(smu, NULL, 0); >>> >>> return ret; >>> } >>> @@ -2330,13 +2344,13 @@ static int smu_handle_task(struct smu_context *smu, >>> ret = smu_pre_display_config_changed(smu); >>> if (ret) >>> return ret; >>> - ret = smu_adjust_power_state_dynamic(smu, level, false, false); >>> + ret = smu_adjust_power_state_dynamic(smu, level, false); >>> break; >>> case AMD_PP_TASK_COMPLETE_INIT: >>> - ret = smu_adjust_power_state_dynamic(smu, level, true, true); >>> + ret = smu_adjust_power_state_dynamic(smu, level, true); >>> break; >>> case AMD_PP_TASK_READJUST_POWER_STATE: >>> - ret = smu_adjust_power_state_dynamic(smu, level, true, false); >>> + ret = smu_adjust_power_state_dynamic(smu, level, true); >>> break; >>> default: >>> break; >>> @@ -2358,12 +2372,11 @@ static int smu_handle_dpm_task(void *handle, >>> >>> static int smu_switch_power_profile(void *handle, >>> enum PP_SMC_POWER_PROFILE type, >>> - bool en) >>> + bool enable) >>> { >>> struct smu_context *smu = handle; >>> struct smu_dpm_context *smu_dpm_ctx = &(smu->smu_dpm); >>> - long workload[1]; >>> - uint32_t index; >>> + int ret; >>> >>> if (!smu->pm_enabled || !smu->adev->pm.dpm_enabled) >>> return -EOPNOTSUPP; >>> @@ -2371,21 +2384,21 @@ static int smu_switch_power_profile(void *handle, >>> if (!(type < PP_SMC_POWER_PROFILE_CUSTOM)) >>> return -EINVAL; >>> >>> - if (!en) { >>> - smu->workload_mask &= ~(1 << smu->workload_prority[type]); >>> - index = fls(smu->workload_mask); >>> - index = index > 0 && index <= WORKLOAD_POLICY_MAX ? index - 1 : 0; >>> - workload[0] = smu->workload_setting[index]; >>> - } else { >>> - smu->workload_mask |= (1 << smu->workload_prority[type]); >>> - index = fls(smu->workload_mask); >>> - index = index <= WORKLOAD_POLICY_MAX ? index - 1 : 0; >>> - workload[0] = smu->workload_setting[index]; >>> - } >>> - >>> if (smu_dpm_ctx->dpm_level != AMD_DPM_FORCED_LEVEL_MANUAL && >>> - smu_dpm_ctx->dpm_level != AMD_DPM_FORCED_LEVEL_PERF_DETERMINISM) >>> - smu_bump_power_profile_mode(smu, workload, 0); >>> + smu_dpm_ctx->dpm_level != AMD_DPM_FORCED_LEVEL_PERF_DETERMINISM) { >>> + if (enable) >>> + smu_power_profile_mode_get(smu, type); >>> + else >>> + smu_power_profile_mode_put(smu, type); >>> + ret = smu_bump_power_profile_mode(smu, NULL, 0); >>> + if (ret) { >>> + if (enable) >>> + smu_power_profile_mode_put(smu, type); >>> + else >>> + smu_power_profile_mode_get(smu, type); >>> + return ret; >>> + } >>> + } >>> >>> return 0; >>> } >>> @@ -3084,12 +3097,35 @@ static int smu_set_power_profile_mode(void *handle, >>> uint32_t param_size) >>> { >>> struct smu_context *smu = handle; >>> + bool custom = false; >>> + int ret = 0; >>> >>> if (!smu->pm_enabled || !smu->adev->pm.dpm_enabled || >>> !smu->ppt_funcs->set_power_profile_mode) >>> return -EOPNOTSUPP; >>> >>> - return smu_bump_power_profile_mode(smu, param, param_size); >>> + if (param[param_size] == PP_SMC_POWER_PROFILE_CUSTOM) { >>> + custom = true; >>> + /* clear frontend mask so custom changes propogate */ >>> + smu->workload_mask = 0; >>> + } >>> + >>> + if ((param[param_size] != smu->power_profile_mode) || custom) { >>> + /* clear the old user preference */ >>> + smu_power_profile_mode_put(smu, smu->power_profile_mode); >>> + /* set the new user preference */ >>> + smu_power_profile_mode_get(smu, param[param_size]); >>> + ret = smu_bump_power_profile_mode(smu, >>> + custom ? param : NULL, >>> + custom ? param_size : 0); >>> + if (ret) >>> + smu_power_profile_mode_put(smu, param[param_size]); >>> + else >>> + /* store the user's preference */ >>> + smu->power_profile_mode = param[param_size]; >>> + } >>> + >>> + return ret; >>> } >>> >>> static int smu_get_fan_control_mode(void *handle, u32 *fan_mode) >>> diff --git a/drivers/gpu/drm/amd/pm/swsmu/inc/amdgpu_smu.h b/drivers/gpu/drm/amd/pm/swsmu/inc/amdgpu_smu.h >>> index d407777d6711..3630593bce61 100644 >>> --- a/drivers/gpu/drm/amd/pm/swsmu/inc/amdgpu_smu.h >>> +++ b/drivers/gpu/drm/amd/pm/swsmu/inc/amdgpu_smu.h >>> @@ -556,11 +556,13 @@ struct smu_context { >>> uint32_t hard_min_uclk_req_from_dal; >>> bool disable_uclk_switch; >>> >>> + /* asic agnostic workload mask */ >>> uint32_t workload_mask; >>> - uint32_t workload_prority[WORKLOAD_POLICY_MAX]; >>> - uint32_t workload_setting[WORKLOAD_POLICY_MAX]; >>> + /* default/user workload preference */ >>> uint32_t power_profile_mode; >>> - uint32_t default_power_profile_mode; >>> + uint32_t workload_refcount[PP_SMC_POWER_PROFILE_COUNT]; >>> + /* backend specific custom workload settings */ >>> + long *custom_profile_params; >>> bool pm_enabled; >>> bool is_apu; >>> >>> @@ -731,9 +733,12 @@ struct pptable_funcs { >>> * @set_power_profile_mode: Set a power profile mode. Also used to >>> * create/set custom power profile modes. >>> * &input: Power profile mode parameters. >>> - * &size: Size of &input. >>> + * &workload_mask: mask of workloads to enable >>> + * &custom_params: custom profile parameters >>> + * &custom_params_max_idx: max valid idx into custom_params >>> */ >>> - int (*set_power_profile_mode)(struct smu_context *smu, long *input, uint32_t size); >>> + int (*set_power_profile_mode)(struct smu_context *smu, u32 workload_mask, >>> + long *custom_params, u32 custom_params_max_idx); >>> >>> /** >>> * @dpm_set_vcn_enable: Enable/disable VCN engine dynamic power >>> diff --git a/drivers/gpu/drm/amd/pm/swsmu/smu11/arcturus_ppt.c b/drivers/gpu/drm/amd/pm/swsmu/smu11/arcturus_ppt.c >>> index a15754b1989f..8aa61a9f7778 100644 >>> --- a/drivers/gpu/drm/amd/pm/swsmu/smu11/arcturus_ppt.c >>> +++ b/drivers/gpu/drm/amd/pm/swsmu/smu11/arcturus_ppt.c >>> @@ -1445,98 +1445,120 @@ static int arcturus_get_power_profile_mode(struct smu_context *smu, >>> return size; >>> } >>> >>> -static int arcturus_set_power_profile_mode(struct smu_context *smu, >>> - long *input, >>> - uint32_t size) >>> +#define ARCTURUS_CUSTOM_PARAMS_COUNT 10 >>> +#define ARCTURUS_CUSTOM_PARAMS_CLOCK_COUNT 2 >>> +#define ARCTURUS_CUSTOM_PARAMS_SIZE (ARCTURUS_CUSTOM_PARAMS_CLOCK_COUNT * ARCTURUS_CUSTOM_PARAMS_COUNT * sizeof(long)) >>> + >>> +static int arcturus_set_power_profile_mode_coeff(struct smu_context *smu, >>> + long *input) >>> { >>> DpmActivityMonitorCoeffInt_t activity_monitor; >>> - int workload_type = 0; >>> - uint32_t profile_mode = input[size]; >>> - int ret = 0; >>> + int ret, idx; >>> >>> - if (profile_mode > PP_SMC_POWER_PROFILE_CUSTOM) { >>> - dev_err(smu->adev->dev, "Invalid power profile mode %d\n", profile_mode); >>> - return -EINVAL; >>> + ret = smu_cmn_update_table(smu, >>> + SMU_TABLE_ACTIVITY_MONITOR_COEFF, >>> + WORKLOAD_PPLIB_CUSTOM_BIT, >>> + (void *)(&activity_monitor), >>> + false); >>> + if (ret) { >>> + dev_err(smu->adev->dev, "[%s] Failed to get activity monitor!", __func__); >>> + return ret; >>> } >>> >>> + idx = 0 * ARCTURUS_CUSTOM_PARAMS_COUNT; >>> + if (input[idx]) { >>> + /* Gfxclk */ >>> + activity_monitor.Gfx_FPS = input[idx + 1]; >>> + activity_monitor.Gfx_UseRlcBusy = input[idx + 2]; >>> + activity_monitor.Gfx_MinActiveFreqType = input[idx + 3]; >>> + activity_monitor.Gfx_MinActiveFreq = input[idx + 4]; >>> + activity_monitor.Gfx_BoosterFreqType = input[idx + 5]; >>> + activity_monitor.Gfx_BoosterFreq = input[idx + 6]; >>> + activity_monitor.Gfx_PD_Data_limit_c = input[idx + 7]; >>> + activity_monitor.Gfx_PD_Data_error_coeff = input[idx + 8]; >>> + activity_monitor.Gfx_PD_Data_error_rate_coeff = input[idx + 9]; >>> + } >>> + idx = 1 * ARCTURUS_CUSTOM_PARAMS_COUNT; >>> + if (input[idx]) { >>> + /* Uclk */ >>> + activity_monitor.Mem_FPS = input[idx + 1]; >>> + activity_monitor.Mem_UseRlcBusy = input[idx + 2]; >>> + activity_monitor.Mem_MinActiveFreqType = input[idx + 3]; >>> + activity_monitor.Mem_MinActiveFreq = input[idx + 4]; >>> + activity_monitor.Mem_BoosterFreqType = input[idx + 5]; >>> + activity_monitor.Mem_BoosterFreq = input[idx + 6]; >>> + activity_monitor.Mem_PD_Data_limit_c = input[idx + 7]; >>> + activity_monitor.Mem_PD_Data_error_coeff = input[idx + 8]; >>> + activity_monitor.Mem_PD_Data_error_rate_coeff = input[idx + 9]; >>> + } >>> >>> - if ((profile_mode == PP_SMC_POWER_PROFILE_CUSTOM) && >>> - (smu->smc_fw_version >= 0x360d00)) { >>> - if (size != 10) >>> - return -EINVAL; >>> + ret = smu_cmn_update_table(smu, >>> + SMU_TABLE_ACTIVITY_MONITOR_COEFF, >>> + WORKLOAD_PPLIB_CUSTOM_BIT, >>> + (void *)(&activity_monitor), >>> + true); >>> + if (ret) { >>> + dev_err(smu->adev->dev, "[%s] Failed to set activity monitor!", __func__); >>> + return ret; >>> + } >>> >>> - ret = smu_cmn_update_table(smu, >>> - SMU_TABLE_ACTIVITY_MONITOR_COEFF, >>> - WORKLOAD_PPLIB_CUSTOM_BIT, >>> - (void *)(&activity_monitor), >>> - false); >>> - if (ret) { >>> - dev_err(smu->adev->dev, "[%s] Failed to get activity monitor!", __func__); >>> - return ret; >>> - } >>> + return ret; >>> +} >>> >>> - switch (input[0]) { >>> - case 0: /* Gfxclk */ >>> - activity_monitor.Gfx_FPS = input[1]; >>> - activity_monitor.Gfx_UseRlcBusy = input[2]; >>> - activity_monitor.Gfx_MinActiveFreqType = input[3]; >>> - activity_monitor.Gfx_MinActiveFreq = input[4]; >>> - activity_monitor.Gfx_BoosterFreqType = input[5]; >>> - activity_monitor.Gfx_BoosterFreq = input[6]; >>> - activity_monitor.Gfx_PD_Data_limit_c = input[7]; >>> - activity_monitor.Gfx_PD_Data_error_coeff = input[8]; >>> - activity_monitor.Gfx_PD_Data_error_rate_coeff = input[9]; >>> - break; >>> - case 1: /* Uclk */ >>> - activity_monitor.Mem_FPS = input[1]; >>> - activity_monitor.Mem_UseRlcBusy = input[2]; >>> - activity_monitor.Mem_MinActiveFreqType = input[3]; >>> - activity_monitor.Mem_MinActiveFreq = input[4]; >>> - activity_monitor.Mem_BoosterFreqType = input[5]; >>> - activity_monitor.Mem_BoosterFreq = input[6]; >>> - activity_monitor.Mem_PD_Data_limit_c = input[7]; >>> - activity_monitor.Mem_PD_Data_error_coeff = input[8]; >>> - activity_monitor.Mem_PD_Data_error_rate_coeff = input[9]; >>> - break; >>> - default: >>> +static int arcturus_set_power_profile_mode(struct smu_context *smu, >>> + u32 workload_mask, >>> + long *custom_params, >>> + u32 custom_params_max_idx) >>> +{ >>> + u32 backend_workload_mask = 0; >>> + int ret, idx = -1, i; >>> + >>> + smu_cmn_get_backend_workload_mask(smu, workload_mask, >>> + &backend_workload_mask); >>> + >>> + if (workload_mask & (1 << PP_SMC_POWER_PROFILE_CUSTOM)) { >>> + if (smu->smc_fw_version < 0x360d00) >>> return -EINVAL; >>> + if (!smu->custom_profile_params) { >>> + smu->custom_profile_params = >>> + kzalloc(ARCTURUS_CUSTOM_PARAMS_SIZE, GFP_KERNEL); >>> + if (!smu->custom_profile_params) >>> + return -ENOMEM; >>> } >>> - >>> - ret = smu_cmn_update_table(smu, >>> - SMU_TABLE_ACTIVITY_MONITOR_COEFF, >>> - WORKLOAD_PPLIB_CUSTOM_BIT, >>> - (void *)(&activity_monitor), >>> - true); >>> + if (custom_params && custom_params_max_idx) { >>> + if (custom_params_max_idx != ARCTURUS_CUSTOM_PARAMS_COUNT) >>> + return -EINVAL; >>> + if (custom_params[0] >= ARCTURUS_CUSTOM_PARAMS_CLOCK_COUNT) >>> + return -EINVAL; >>> + idx = custom_params[0] * ARCTURUS_CUSTOM_PARAMS_COUNT; >>> + smu->custom_profile_params[idx] = 1; >>> + for (i = 1; i < custom_params_max_idx; i++) >>> + smu->custom_profile_params[idx + i] = custom_params[i]; >>> + } >>> + ret = arcturus_set_power_profile_mode_coeff(smu, >>> + smu->custom_profile_params); >>> if (ret) { >>> - dev_err(smu->adev->dev, "[%s] Failed to set activity monitor!", __func__); >>> + if (idx != -1) >>> + smu->custom_profile_params[idx] = 0; >> >> I don't know if this is the right thing to do. This means we are only >> having a partial revert eventhough custom mode settings as a whole failed. >> >> 1) Current Mode = 3D >> 2) Pass Custom + GfxCLK custom settings >> 3) Pass Custom + MemCLK custom settings >> >> When 3) fails, we revert from the custom mode (put operation) and it >> goes back to 3D as the new settings failed. At a later point if user >> passes MemCLK custom settings, this is going to pick the initial GFXCLK >> custom settings also. Is that needed? >> > > I guess that is an open question. Each step is discrete and 2 > succeeded so it seemed logical to me that it should be retained. > I.e., if you are trying custom settings, it seems logical that if an > operation fails, you'd only need to redo the operation that failed. > E.g., one of the memclk parameters was bad so retry 3 with new > parameters; no need to do step 2 again. That said, custom is kind of > weird because you can modify different aspects of it with each > discrete operation. > I guess then we need to document this behavior. With that, Reviewed-by: Lijo Lazar <lijo.lazar@xxxxxxx> Thanks, Lijo > Alex > > >> Thanks, >> Lijo >>> return ret; >>> } >>> - } >>> - >>> - /* >>> - * Conv PP_SMC_POWER_PROFILE* to WORKLOAD_PPLIB_*_BIT >>> - * Not all profile modes are supported on arcturus. >>> - */ >>> - workload_type = smu_cmn_to_asic_specific_index(smu, >>> - CMN2ASIC_MAPPING_WORKLOAD, >>> - profile_mode); >>> - if (workload_type < 0) { >>> - dev_dbg(smu->adev->dev, "Unsupported power profile mode %d on arcturus\n", profile_mode); >>> - return -EINVAL; >>> + } else if (smu->custom_profile_params) { >>> + memset(smu->custom_profile_params, 0, ARCTURUS_CUSTOM_PARAMS_SIZE); >>> } >>> >>> ret = smu_cmn_send_smc_msg_with_param(smu, >>> - SMU_MSG_SetWorkloadMask, >>> - 1 << workload_type, >>> - NULL); >>> + SMU_MSG_SetWorkloadMask, >>> + backend_workload_mask, >>> + NULL); >>> if (ret) { >>> - dev_err(smu->adev->dev, "Fail to set workload type %d\n", workload_type); >>> + dev_err(smu->adev->dev, "Failed to set workload mask 0x%08x\n", >>> + workload_mask); >>> + if (idx != -1) >>> + smu->custom_profile_params[idx] = 0; >>> return ret; >>> } >>> >>> - smu->power_profile_mode = profile_mode; >>> - >>> - return 0; >>> + return ret; >>> } >>> >>> static int arcturus_set_performance_level(struct smu_context *smu, >>> diff --git a/drivers/gpu/drm/amd/pm/swsmu/smu11/navi10_ppt.c b/drivers/gpu/drm/amd/pm/swsmu/smu11/navi10_ppt.c >>> index faa8e7d9c3c6..7fad5dfb39c4 100644 >>> --- a/drivers/gpu/drm/amd/pm/swsmu/smu11/navi10_ppt.c >>> +++ b/drivers/gpu/drm/amd/pm/swsmu/smu11/navi10_ppt.c >>> @@ -2006,87 +2006,122 @@ static int navi10_get_power_profile_mode(struct smu_context *smu, char *buf) >>> return size; >>> } >>> >>> -static int navi10_set_power_profile_mode(struct smu_context *smu, long *input, uint32_t size) >>> +#define NAVI10_CUSTOM_PARAMS_COUNT 10 >>> +#define NAVI10_CUSTOM_PARAMS_CLOCKS_COUNT 3 >>> +#define NAVI10_CUSTOM_PARAMS_SIZE (NAVI10_CUSTOM_PARAMS_CLOCKS_COUNT * NAVI10_CUSTOM_PARAMS_COUNT * sizeof(long)) >>> + >>> +static int navi10_set_power_profile_mode_coeff(struct smu_context *smu, >>> + long *input) >>> { >>> DpmActivityMonitorCoeffInt_t activity_monitor; >>> - int workload_type, ret = 0; >>> + int ret, idx; >>> >>> - smu->power_profile_mode = input[size]; >>> + ret = smu_cmn_update_table(smu, >>> + SMU_TABLE_ACTIVITY_MONITOR_COEFF, WORKLOAD_PPLIB_CUSTOM_BIT, >>> + (void *)(&activity_monitor), false); >>> + if (ret) { >>> + dev_err(smu->adev->dev, "[%s] Failed to get activity monitor!", __func__); >>> + return ret; >>> + } >>> >>> - if (smu->power_profile_mode > PP_SMC_POWER_PROFILE_CUSTOM) { >>> - dev_err(smu->adev->dev, "Invalid power profile mode %d\n", smu->power_profile_mode); >>> - return -EINVAL; >>> + idx = 0 * NAVI10_CUSTOM_PARAMS_COUNT; >>> + if (input[idx]) { >>> + /* Gfxclk */ >>> + activity_monitor.Gfx_FPS = input[idx + 1]; >>> + activity_monitor.Gfx_MinFreqStep = input[idx + 2]; >>> + activity_monitor.Gfx_MinActiveFreqType = input[idx + 3]; >>> + activity_monitor.Gfx_MinActiveFreq = input[idx + 4]; >>> + activity_monitor.Gfx_BoosterFreqType = input[idx + 5]; >>> + activity_monitor.Gfx_BoosterFreq = input[idx + 6]; >>> + activity_monitor.Gfx_PD_Data_limit_c = input[idx + 7]; >>> + activity_monitor.Gfx_PD_Data_error_coeff = input[idx + 8]; >>> + activity_monitor.Gfx_PD_Data_error_rate_coeff = input[idx + 9]; >>> + } >>> + idx = 1 * NAVI10_CUSTOM_PARAMS_COUNT; >>> + if (input[idx]) { >>> + /* Socclk */ >>> + activity_monitor.Soc_FPS = input[idx + 1]; >>> + activity_monitor.Soc_MinFreqStep = input[idx + 2]; >>> + activity_monitor.Soc_MinActiveFreqType = input[idx + 3]; >>> + activity_monitor.Soc_MinActiveFreq = input[idx + 4]; >>> + activity_monitor.Soc_BoosterFreqType = input[idx + 5]; >>> + activity_monitor.Soc_BoosterFreq = input[idx + 6]; >>> + activity_monitor.Soc_PD_Data_limit_c = input[idx + 7]; >>> + activity_monitor.Soc_PD_Data_error_coeff = input[idx + 8]; >>> + activity_monitor.Soc_PD_Data_error_rate_coeff = input[idx + 9]; >>> + } >>> + idx = 2 * NAVI10_CUSTOM_PARAMS_COUNT; >>> + if (input[idx]) { >>> + /* Memclk */ >>> + activity_monitor.Mem_FPS = input[idx + 1]; >>> + activity_monitor.Mem_MinFreqStep = input[idx + 2]; >>> + activity_monitor.Mem_MinActiveFreqType = input[idx + 3]; >>> + activity_monitor.Mem_MinActiveFreq = input[idx + 4]; >>> + activity_monitor.Mem_BoosterFreqType = input[idx + 5]; >>> + activity_monitor.Mem_BoosterFreq = input[idx + 6]; >>> + activity_monitor.Mem_PD_Data_limit_c = input[idx + 7]; >>> + activity_monitor.Mem_PD_Data_error_coeff = input[idx + 8]; >>> + activity_monitor.Mem_PD_Data_error_rate_coeff = input[idx + 9]; >>> + } >>> + >>> + ret = smu_cmn_update_table(smu, >>> + SMU_TABLE_ACTIVITY_MONITOR_COEFF, WORKLOAD_PPLIB_CUSTOM_BIT, >>> + (void *)(&activity_monitor), true); >>> + if (ret) { >>> + dev_err(smu->adev->dev, "[%s] Failed to set activity monitor!", __func__); >>> + return ret; >>> } >>> >>> - if (smu->power_profile_mode == PP_SMC_POWER_PROFILE_CUSTOM) { >>> - if (size != 10) >>> - return -EINVAL; >>> + return ret; >>> +} >>> >>> - ret = smu_cmn_update_table(smu, >>> - SMU_TABLE_ACTIVITY_MONITOR_COEFF, WORKLOAD_PPLIB_CUSTOM_BIT, >>> - (void *)(&activity_monitor), false); >>> - if (ret) { >>> - dev_err(smu->adev->dev, "[%s] Failed to get activity monitor!", __func__); >>> - return ret; >>> - } >>> +static int navi10_set_power_profile_mode(struct smu_context *smu, >>> + u32 workload_mask, >>> + long *custom_params, >>> + u32 custom_params_max_idx) >>> +{ >>> + u32 backend_workload_mask = 0; >>> + int ret, idx = -1, i; >>> >>> - switch (input[0]) { >>> - case 0: /* Gfxclk */ >>> - activity_monitor.Gfx_FPS = input[1]; >>> - activity_monitor.Gfx_MinFreqStep = input[2]; >>> - activity_monitor.Gfx_MinActiveFreqType = input[3]; >>> - activity_monitor.Gfx_MinActiveFreq = input[4]; >>> - activity_monitor.Gfx_BoosterFreqType = input[5]; >>> - activity_monitor.Gfx_BoosterFreq = input[6]; >>> - activity_monitor.Gfx_PD_Data_limit_c = input[7]; >>> - activity_monitor.Gfx_PD_Data_error_coeff = input[8]; >>> - activity_monitor.Gfx_PD_Data_error_rate_coeff = input[9]; >>> - break; >>> - case 1: /* Socclk */ >>> - activity_monitor.Soc_FPS = input[1]; >>> - activity_monitor.Soc_MinFreqStep = input[2]; >>> - activity_monitor.Soc_MinActiveFreqType = input[3]; >>> - activity_monitor.Soc_MinActiveFreq = input[4]; >>> - activity_monitor.Soc_BoosterFreqType = input[5]; >>> - activity_monitor.Soc_BoosterFreq = input[6]; >>> - activity_monitor.Soc_PD_Data_limit_c = input[7]; >>> - activity_monitor.Soc_PD_Data_error_coeff = input[8]; >>> - activity_monitor.Soc_PD_Data_error_rate_coeff = input[9]; >>> - break; >>> - case 2: /* Memclk */ >>> - activity_monitor.Mem_FPS = input[1]; >>> - activity_monitor.Mem_MinFreqStep = input[2]; >>> - activity_monitor.Mem_MinActiveFreqType = input[3]; >>> - activity_monitor.Mem_MinActiveFreq = input[4]; >>> - activity_monitor.Mem_BoosterFreqType = input[5]; >>> - activity_monitor.Mem_BoosterFreq = input[6]; >>> - activity_monitor.Mem_PD_Data_limit_c = input[7]; >>> - activity_monitor.Mem_PD_Data_error_coeff = input[8]; >>> - activity_monitor.Mem_PD_Data_error_rate_coeff = input[9]; >>> - break; >>> - default: >>> - return -EINVAL; >>> - } >>> + smu_cmn_get_backend_workload_mask(smu, workload_mask, >>> + &backend_workload_mask); >>> >>> - ret = smu_cmn_update_table(smu, >>> - SMU_TABLE_ACTIVITY_MONITOR_COEFF, WORKLOAD_PPLIB_CUSTOM_BIT, >>> - (void *)(&activity_monitor), true); >>> + if (workload_mask & (1 << PP_SMC_POWER_PROFILE_CUSTOM)) { >>> + if (!smu->custom_profile_params) { >>> + smu->custom_profile_params = kzalloc(NAVI10_CUSTOM_PARAMS_SIZE, GFP_KERNEL); >>> + if (!smu->custom_profile_params) >>> + return -ENOMEM; >>> + } >>> + if (custom_params && custom_params_max_idx) { >>> + if (custom_params_max_idx != NAVI10_CUSTOM_PARAMS_COUNT) >>> + return -EINVAL; >>> + if (custom_params[0] >= NAVI10_CUSTOM_PARAMS_CLOCKS_COUNT) >>> + return -EINVAL; >>> + idx = custom_params[0] * NAVI10_CUSTOM_PARAMS_COUNT; >>> + smu->custom_profile_params[idx] = 1; >>> + for (i = 1; i < custom_params_max_idx; i++) >>> + smu->custom_profile_params[idx + i] = custom_params[i]; >>> + } >>> + ret = navi10_set_power_profile_mode_coeff(smu, >>> + smu->custom_profile_params); >>> if (ret) { >>> - dev_err(smu->adev->dev, "[%s] Failed to set activity monitor!", __func__); >>> + if (idx != -1) >>> + smu->custom_profile_params[idx] = 0; >>> return ret; >>> } >>> + } else if (smu->custom_profile_params) { >>> + memset(smu->custom_profile_params, 0, NAVI10_CUSTOM_PARAMS_SIZE); >>> } >>> >>> - /* conv PP_SMC_POWER_PROFILE* to WORKLOAD_PPLIB_*_BIT */ >>> - workload_type = smu_cmn_to_asic_specific_index(smu, >>> - CMN2ASIC_MAPPING_WORKLOAD, >>> - smu->power_profile_mode); >>> - if (workload_type < 0) >>> - return -EINVAL; >>> ret = smu_cmn_send_smc_msg_with_param(smu, SMU_MSG_SetWorkloadMask, >>> - 1 << workload_type, NULL); >>> - if (ret) >>> - dev_err(smu->adev->dev, "[%s] Failed to set work load mask!", __func__); >>> + backend_workload_mask, NULL); >>> + if (ret) { >>> + dev_err(smu->adev->dev, "Failed to set workload mask 0x%08x\n", >>> + workload_mask); >>> + if (idx != -1) >>> + smu->custom_profile_params[idx] = 0; >>> + return ret; >>> + } >>> >>> return ret; >>> } >>> diff --git a/drivers/gpu/drm/amd/pm/swsmu/smu11/sienna_cichlid_ppt.c b/drivers/gpu/drm/amd/pm/swsmu/smu11/sienna_cichlid_ppt.c >>> index 30d050a6e953..19a25fdc2f5b 100644 >>> --- a/drivers/gpu/drm/amd/pm/swsmu/smu11/sienna_cichlid_ppt.c >>> +++ b/drivers/gpu/drm/amd/pm/swsmu/smu11/sienna_cichlid_ppt.c >>> @@ -1704,90 +1704,126 @@ static int sienna_cichlid_get_power_profile_mode(struct smu_context *smu, char * >>> return size; >>> } >>> >>> -static int sienna_cichlid_set_power_profile_mode(struct smu_context *smu, long *input, uint32_t size) >>> +#define SIENNA_CICHLID_CUSTOM_PARAMS_COUNT 10 >>> +#define SIENNA_CICHLID_CUSTOM_PARAMS_CLOCK_COUNT 3 >>> +#define SIENNA_CICHLID_CUSTOM_PARAMS_SIZE (SIENNA_CICHLID_CUSTOM_PARAMS_CLOCK_COUNT * SIENNA_CICHLID_CUSTOM_PARAMS_COUNT * sizeof(long)) >>> + >>> +static int sienna_cichlid_set_power_profile_mode_coeff(struct smu_context *smu, >>> + long *input) >>> { >>> >>> DpmActivityMonitorCoeffIntExternal_t activity_monitor_external; >>> DpmActivityMonitorCoeffInt_t *activity_monitor = >>> &(activity_monitor_external.DpmActivityMonitorCoeffInt); >>> - int workload_type, ret = 0; >>> + int ret, idx; >>> >>> - smu->power_profile_mode = input[size]; >>> + ret = smu_cmn_update_table(smu, >>> + SMU_TABLE_ACTIVITY_MONITOR_COEFF, WORKLOAD_PPLIB_CUSTOM_BIT, >>> + (void *)(&activity_monitor_external), false); >>> + if (ret) { >>> + dev_err(smu->adev->dev, "[%s] Failed to get activity monitor!", __func__); >>> + return ret; >>> + } >>> >>> - if (smu->power_profile_mode > PP_SMC_POWER_PROFILE_CUSTOM) { >>> - dev_err(smu->adev->dev, "Invalid power profile mode %d\n", smu->power_profile_mode); >>> - return -EINVAL; >>> + idx = 0 * SIENNA_CICHLID_CUSTOM_PARAMS_COUNT; >>> + if (input[idx]) { >>> + /* Gfxclk */ >>> + activity_monitor->Gfx_FPS = input[idx + 1]; >>> + activity_monitor->Gfx_MinFreqStep = input[idx + 2]; >>> + activity_monitor->Gfx_MinActiveFreqType = input[idx + 3]; >>> + activity_monitor->Gfx_MinActiveFreq = input[idx + 4]; >>> + activity_monitor->Gfx_BoosterFreqType = input[idx + 5]; >>> + activity_monitor->Gfx_BoosterFreq = input[idx + 6]; >>> + activity_monitor->Gfx_PD_Data_limit_c = input[idx + 7]; >>> + activity_monitor->Gfx_PD_Data_error_coeff = input[idx + 8]; >>> + activity_monitor->Gfx_PD_Data_error_rate_coeff = input[idx + 9]; >>> + } >>> + idx = 1 * SIENNA_CICHLID_CUSTOM_PARAMS_COUNT; >>> + if (input[idx]) { >>> + /* Socclk */ >>> + activity_monitor->Fclk_FPS = input[idx + 1]; >>> + activity_monitor->Fclk_MinFreqStep = input[idx + 2]; >>> + activity_monitor->Fclk_MinActiveFreqType = input[idx + 3]; >>> + activity_monitor->Fclk_MinActiveFreq = input[idx + 4]; >>> + activity_monitor->Fclk_BoosterFreqType = input[idx + 5]; >>> + activity_monitor->Fclk_BoosterFreq = input[idx + 6]; >>> + activity_monitor->Fclk_PD_Data_limit_c = input[idx + 7]; >>> + activity_monitor->Fclk_PD_Data_error_coeff = input[idx + 8]; >>> + activity_monitor->Fclk_PD_Data_error_rate_coeff = input[idx + 9]; >>> + } >>> + idx = 2 * SIENNA_CICHLID_CUSTOM_PARAMS_COUNT; >>> + if (input[idx]) { >>> + /* Memclk */ >>> + activity_monitor->Mem_FPS = input[idx + 1]; >>> + activity_monitor->Mem_MinFreqStep = input[idx + 2]; >>> + activity_monitor->Mem_MinActiveFreqType = input[idx + 3]; >>> + activity_monitor->Mem_MinActiveFreq = input[idx + 4]; >>> + activity_monitor->Mem_BoosterFreqType = input[idx + 5]; >>> + activity_monitor->Mem_BoosterFreq = input[idx + 6]; >>> + activity_monitor->Mem_PD_Data_limit_c = input[idx + 7]; >>> + activity_monitor->Mem_PD_Data_error_coeff = input[idx + 8]; >>> + activity_monitor->Mem_PD_Data_error_rate_coeff = input[idx + 9]; >>> } >>> >>> - if (smu->power_profile_mode == PP_SMC_POWER_PROFILE_CUSTOM) { >>> - if (size != 10) >>> - return -EINVAL; >>> + ret = smu_cmn_update_table(smu, >>> + SMU_TABLE_ACTIVITY_MONITOR_COEFF, WORKLOAD_PPLIB_CUSTOM_BIT, >>> + (void *)(&activity_monitor_external), true); >>> + if (ret) { >>> + dev_err(smu->adev->dev, "[%s] Failed to set activity monitor!", __func__); >>> + return ret; >>> + } >>> >>> - ret = smu_cmn_update_table(smu, >>> - SMU_TABLE_ACTIVITY_MONITOR_COEFF, WORKLOAD_PPLIB_CUSTOM_BIT, >>> - (void *)(&activity_monitor_external), false); >>> - if (ret) { >>> - dev_err(smu->adev->dev, "[%s] Failed to get activity monitor!", __func__); >>> - return ret; >>> - } >>> + return ret; >>> +} >>> >>> - switch (input[0]) { >>> - case 0: /* Gfxclk */ >>> - activity_monitor->Gfx_FPS = input[1]; >>> - activity_monitor->Gfx_MinFreqStep = input[2]; >>> - activity_monitor->Gfx_MinActiveFreqType = input[3]; >>> - activity_monitor->Gfx_MinActiveFreq = input[4]; >>> - activity_monitor->Gfx_BoosterFreqType = input[5]; >>> - activity_monitor->Gfx_BoosterFreq = input[6]; >>> - activity_monitor->Gfx_PD_Data_limit_c = input[7]; >>> - activity_monitor->Gfx_PD_Data_error_coeff = input[8]; >>> - activity_monitor->Gfx_PD_Data_error_rate_coeff = input[9]; >>> - break; >>> - case 1: /* Socclk */ >>> - activity_monitor->Fclk_FPS = input[1]; >>> - activity_monitor->Fclk_MinFreqStep = input[2]; >>> - activity_monitor->Fclk_MinActiveFreqType = input[3]; >>> - activity_monitor->Fclk_MinActiveFreq = input[4]; >>> - activity_monitor->Fclk_BoosterFreqType = input[5]; >>> - activity_monitor->Fclk_BoosterFreq = input[6]; >>> - activity_monitor->Fclk_PD_Data_limit_c = input[7]; >>> - activity_monitor->Fclk_PD_Data_error_coeff = input[8]; >>> - activity_monitor->Fclk_PD_Data_error_rate_coeff = input[9]; >>> - break; >>> - case 2: /* Memclk */ >>> - activity_monitor->Mem_FPS = input[1]; >>> - activity_monitor->Mem_MinFreqStep = input[2]; >>> - activity_monitor->Mem_MinActiveFreqType = input[3]; >>> - activity_monitor->Mem_MinActiveFreq = input[4]; >>> - activity_monitor->Mem_BoosterFreqType = input[5]; >>> - activity_monitor->Mem_BoosterFreq = input[6]; >>> - activity_monitor->Mem_PD_Data_limit_c = input[7]; >>> - activity_monitor->Mem_PD_Data_error_coeff = input[8]; >>> - activity_monitor->Mem_PD_Data_error_rate_coeff = input[9]; >>> - break; >>> - default: >>> - return -EINVAL; >>> - } >>> +static int sienna_cichlid_set_power_profile_mode(struct smu_context *smu, >>> + u32 workload_mask, >>> + long *custom_params, >>> + u32 custom_params_max_idx) >>> +{ >>> + u32 backend_workload_mask = 0; >>> + int ret, idx = -1, i; >>> >>> - ret = smu_cmn_update_table(smu, >>> - SMU_TABLE_ACTIVITY_MONITOR_COEFF, WORKLOAD_PPLIB_CUSTOM_BIT, >>> - (void *)(&activity_monitor_external), true); >>> + smu_cmn_get_backend_workload_mask(smu, workload_mask, >>> + &backend_workload_mask); >>> + >>> + if (workload_mask & (1 << PP_SMC_POWER_PROFILE_CUSTOM)) { >>> + if (!smu->custom_profile_params) { >>> + smu->custom_profile_params = >>> + kzalloc(SIENNA_CICHLID_CUSTOM_PARAMS_SIZE, GFP_KERNEL); >>> + if (!smu->custom_profile_params) >>> + return -ENOMEM; >>> + } >>> + if (custom_params && custom_params_max_idx) { >>> + if (custom_params_max_idx != SIENNA_CICHLID_CUSTOM_PARAMS_COUNT) >>> + return -EINVAL; >>> + if (custom_params[0] >= SIENNA_CICHLID_CUSTOM_PARAMS_CLOCK_COUNT) >>> + return -EINVAL; >>> + idx = custom_params[0] * SIENNA_CICHLID_CUSTOM_PARAMS_COUNT; >>> + smu->custom_profile_params[idx] = 1; >>> + for (i = 1; i < custom_params_max_idx; i++) >>> + smu->custom_profile_params[idx + i] = custom_params[i]; >>> + } >>> + ret = sienna_cichlid_set_power_profile_mode_coeff(smu, >>> + smu->custom_profile_params); >>> if (ret) { >>> - dev_err(smu->adev->dev, "[%s] Failed to set activity monitor!", __func__); >>> + if (idx != -1) >>> + smu->custom_profile_params[idx] = 0; >>> return ret; >>> } >>> + } else if (smu->custom_profile_params) { >>> + memset(smu->custom_profile_params, 0, SIENNA_CICHLID_CUSTOM_PARAMS_SIZE); >>> } >>> >>> - /* conv PP_SMC_POWER_PROFILE* to WORKLOAD_PPLIB_*_BIT */ >>> - workload_type = smu_cmn_to_asic_specific_index(smu, >>> - CMN2ASIC_MAPPING_WORKLOAD, >>> - smu->power_profile_mode); >>> - if (workload_type < 0) >>> - return -EINVAL; >>> ret = smu_cmn_send_smc_msg_with_param(smu, SMU_MSG_SetWorkloadMask, >>> - 1 << workload_type, NULL); >>> - if (ret) >>> - dev_err(smu->adev->dev, "[%s] Failed to set work load mask!", __func__); >>> + backend_workload_mask, NULL); >>> + if (ret) { >>> + dev_err(smu->adev->dev, "Failed to set workload mask 0x%08x\n", >>> + workload_mask); >>> + if (idx != -1) >>> + smu->custom_profile_params[idx] = 0; >>> + return ret; >>> + } >>> >>> return ret; >>> } >>> diff --git a/drivers/gpu/drm/amd/pm/swsmu/smu11/vangogh_ppt.c b/drivers/gpu/drm/amd/pm/swsmu/smu11/vangogh_ppt.c >>> index cd3e9ba3eff4..a55ea76d7399 100644 >>> --- a/drivers/gpu/drm/amd/pm/swsmu/smu11/vangogh_ppt.c >>> +++ b/drivers/gpu/drm/amd/pm/swsmu/smu11/vangogh_ppt.c >>> @@ -1056,42 +1056,27 @@ static int vangogh_get_power_profile_mode(struct smu_context *smu, >>> return size; >>> } >>> >>> -static int vangogh_set_power_profile_mode(struct smu_context *smu, long *input, uint32_t size) >>> +static int vangogh_set_power_profile_mode(struct smu_context *smu, >>> + u32 workload_mask, >>> + long *custom_params, >>> + u32 custom_params_max_idx) >>> { >>> - int workload_type, ret; >>> - uint32_t profile_mode = input[size]; >>> + u32 backend_workload_mask = 0; >>> + int ret; >>> >>> - if (profile_mode >= PP_SMC_POWER_PROFILE_COUNT) { >>> - dev_err(smu->adev->dev, "Invalid power profile mode %d\n", profile_mode); >>> - return -EINVAL; >>> - } >>> - >>> - if (profile_mode == PP_SMC_POWER_PROFILE_BOOTUP_DEFAULT || >>> - profile_mode == PP_SMC_POWER_PROFILE_POWERSAVING) >>> - return 0; >>> - >>> - /* conv PP_SMC_POWER_PROFILE* to WORKLOAD_PPLIB_*_BIT */ >>> - workload_type = smu_cmn_to_asic_specific_index(smu, >>> - CMN2ASIC_MAPPING_WORKLOAD, >>> - profile_mode); >>> - if (workload_type < 0) { >>> - dev_dbg(smu->adev->dev, "Unsupported power profile mode %d on VANGOGH\n", >>> - profile_mode); >>> - return -EINVAL; >>> - } >>> + smu_cmn_get_backend_workload_mask(smu, workload_mask, >>> + &backend_workload_mask); >>> >>> ret = smu_cmn_send_smc_msg_with_param(smu, SMU_MSG_ActiveProcessNotify, >>> - 1 << workload_type, >>> - NULL); >>> + backend_workload_mask, >>> + NULL); >>> if (ret) { >>> - dev_err_once(smu->adev->dev, "Fail to set workload type %d\n", >>> - workload_type); >>> + dev_err_once(smu->adev->dev, "Fail to set workload mask 0x%08x\n", >>> + workload_mask); >>> return ret; >>> } >>> >>> - smu->power_profile_mode = profile_mode; >>> - >>> - return 0; >>> + return ret; >>> } >>> >>> static int vangogh_set_soft_freq_limited_range(struct smu_context *smu, >>> diff --git a/drivers/gpu/drm/amd/pm/swsmu/smu12/renoir_ppt.c b/drivers/gpu/drm/amd/pm/swsmu/smu12/renoir_ppt.c >>> index a34797f3576b..37d82a71a2d7 100644 >>> --- a/drivers/gpu/drm/amd/pm/swsmu/smu12/renoir_ppt.c >>> +++ b/drivers/gpu/drm/amd/pm/swsmu/smu12/renoir_ppt.c >>> @@ -864,44 +864,27 @@ static int renoir_force_clk_levels(struct smu_context *smu, >>> return ret; >>> } >>> >>> -static int renoir_set_power_profile_mode(struct smu_context *smu, long *input, uint32_t size) >>> +static int renoir_set_power_profile_mode(struct smu_context *smu, >>> + u32 workload_mask, >>> + long *custom_params, >>> + u32 custom_params_max_idx) >>> { >>> - int workload_type, ret; >>> - uint32_t profile_mode = input[size]; >>> + int ret; >>> + u32 backend_workload_mask = 0; >>> >>> - if (profile_mode > PP_SMC_POWER_PROFILE_CUSTOM) { >>> - dev_err(smu->adev->dev, "Invalid power profile mode %d\n", profile_mode); >>> - return -EINVAL; >>> - } >>> - >>> - if (profile_mode == PP_SMC_POWER_PROFILE_BOOTUP_DEFAULT || >>> - profile_mode == PP_SMC_POWER_PROFILE_POWERSAVING) >>> - return 0; >>> - >>> - /* conv PP_SMC_POWER_PROFILE* to WORKLOAD_PPLIB_*_BIT */ >>> - workload_type = smu_cmn_to_asic_specific_index(smu, >>> - CMN2ASIC_MAPPING_WORKLOAD, >>> - profile_mode); >>> - if (workload_type < 0) { >>> - /* >>> - * TODO: If some case need switch to powersave/default power mode >>> - * then can consider enter WORKLOAD_COMPUTE/WORKLOAD_CUSTOM for power saving. >>> - */ >>> - dev_dbg(smu->adev->dev, "Unsupported power profile mode %d on RENOIR\n", profile_mode); >>> - return -EINVAL; >>> - } >>> + smu_cmn_get_backend_workload_mask(smu, workload_mask, >>> + &backend_workload_mask); >>> >>> ret = smu_cmn_send_smc_msg_with_param(smu, SMU_MSG_ActiveProcessNotify, >>> - 1 << workload_type, >>> - NULL); >>> + backend_workload_mask, >>> + NULL); >>> if (ret) { >>> - dev_err_once(smu->adev->dev, "Fail to set workload type %d\n", workload_type); >>> + dev_err_once(smu->adev->dev, "Failed to set workload mask 0x08%x\n", >>> + workload_mask); >>> return ret; >>> } >>> >>> - smu->power_profile_mode = profile_mode; >>> - >>> - return 0; >>> + return ret; >>> } >>> >>> static int renoir_set_peak_clock_by_device(struct smu_context *smu) >>> diff --git a/drivers/gpu/drm/amd/pm/swsmu/smu13/smu_v13_0_0_ppt.c b/drivers/gpu/drm/amd/pm/swsmu/smu13/smu_v13_0_0_ppt.c >>> index 199bdd9720d3..3aa705aae4c0 100644 >>> --- a/drivers/gpu/drm/amd/pm/swsmu/smu13/smu_v13_0_0_ppt.c >>> +++ b/drivers/gpu/drm/amd/pm/swsmu/smu13/smu_v13_0_0_ppt.c >>> @@ -2571,82 +2571,76 @@ static int smu_v13_0_0_get_power_profile_mode(struct smu_context *smu, >>> return size; >>> } >>> >>> -static int smu_v13_0_0_set_power_profile_mode(struct smu_context *smu, >>> - long *input, >>> - uint32_t size) >>> +#define SMU_13_0_0_CUSTOM_PARAMS_COUNT 9 >>> +#define SMU_13_0_0_CUSTOM_PARAMS_CLOCK_COUNT 2 >>> +#define SMU_13_0_0_CUSTOM_PARAMS_SIZE (SMU_13_0_0_CUSTOM_PARAMS_CLOCK_COUNT * SMU_13_0_0_CUSTOM_PARAMS_COUNT * sizeof(long)) >>> + >>> +static int smu_v13_0_0_set_power_profile_mode_coeff(struct smu_context *smu, >>> + long *input) >>> { >>> DpmActivityMonitorCoeffIntExternal_t activity_monitor_external; >>> DpmActivityMonitorCoeffInt_t *activity_monitor = >>> &(activity_monitor_external.DpmActivityMonitorCoeffInt); >>> - int workload_type, ret = 0; >>> - u32 workload_mask, selected_workload_mask; >>> - >>> - smu->power_profile_mode = input[size]; >>> + int ret, idx; >>> >>> - if (smu->power_profile_mode >= PP_SMC_POWER_PROFILE_COUNT) { >>> - dev_err(smu->adev->dev, "Invalid power profile mode %d\n", smu->power_profile_mode); >>> - return -EINVAL; >>> + ret = smu_cmn_update_table(smu, >>> + SMU_TABLE_ACTIVITY_MONITOR_COEFF, >>> + WORKLOAD_PPLIB_CUSTOM_BIT, >>> + (void *)(&activity_monitor_external), >>> + false); >>> + if (ret) { >>> + dev_err(smu->adev->dev, "[%s] Failed to get activity monitor!", __func__); >>> + return ret; >>> } >>> >>> - if (smu->power_profile_mode == PP_SMC_POWER_PROFILE_CUSTOM) { >>> - if (size != 9) >>> - return -EINVAL; >>> - >>> - ret = smu_cmn_update_table(smu, >>> - SMU_TABLE_ACTIVITY_MONITOR_COEFF, >>> - WORKLOAD_PPLIB_CUSTOM_BIT, >>> - (void *)(&activity_monitor_external), >>> - false); >>> - if (ret) { >>> - dev_err(smu->adev->dev, "[%s] Failed to get activity monitor!", __func__); >>> - return ret; >>> - } >>> - >>> - switch (input[0]) { >>> - case 0: /* Gfxclk */ >>> - activity_monitor->Gfx_FPS = input[1]; >>> - activity_monitor->Gfx_MinActiveFreqType = input[2]; >>> - activity_monitor->Gfx_MinActiveFreq = input[3]; >>> - activity_monitor->Gfx_BoosterFreqType = input[4]; >>> - activity_monitor->Gfx_BoosterFreq = input[5]; >>> - activity_monitor->Gfx_PD_Data_limit_c = input[6]; >>> - activity_monitor->Gfx_PD_Data_error_coeff = input[7]; >>> - activity_monitor->Gfx_PD_Data_error_rate_coeff = input[8]; >>> - break; >>> - case 1: /* Fclk */ >>> - activity_monitor->Fclk_FPS = input[1]; >>> - activity_monitor->Fclk_MinActiveFreqType = input[2]; >>> - activity_monitor->Fclk_MinActiveFreq = input[3]; >>> - activity_monitor->Fclk_BoosterFreqType = input[4]; >>> - activity_monitor->Fclk_BoosterFreq = input[5]; >>> - activity_monitor->Fclk_PD_Data_limit_c = input[6]; >>> - activity_monitor->Fclk_PD_Data_error_coeff = input[7]; >>> - activity_monitor->Fclk_PD_Data_error_rate_coeff = input[8]; >>> - break; >>> - default: >>> - return -EINVAL; >>> - } >>> + idx = 0 * SMU_13_0_0_CUSTOM_PARAMS_COUNT; >>> + if (input[idx]) { >>> + /* Gfxclk */ >>> + activity_monitor->Gfx_FPS = input[idx + 1]; >>> + activity_monitor->Gfx_MinActiveFreqType = input[idx + 2]; >>> + activity_monitor->Gfx_MinActiveFreq = input[idx + 3]; >>> + activity_monitor->Gfx_BoosterFreqType = input[idx + 4]; >>> + activity_monitor->Gfx_BoosterFreq = input[idx + 5]; >>> + activity_monitor->Gfx_PD_Data_limit_c = input[idx + 6]; >>> + activity_monitor->Gfx_PD_Data_error_coeff = input[idx + 7]; >>> + activity_monitor->Gfx_PD_Data_error_rate_coeff = input[idx + 8]; >>> + } >>> + idx = 1 * SMU_13_0_0_CUSTOM_PARAMS_COUNT; >>> + if (input[idx]) { >>> + /* Fclk */ >>> + activity_monitor->Fclk_FPS = input[idx + 1]; >>> + activity_monitor->Fclk_MinActiveFreqType = input[idx + 2]; >>> + activity_monitor->Fclk_MinActiveFreq = input[idx + 3]; >>> + activity_monitor->Fclk_BoosterFreqType = input[idx + 4]; >>> + activity_monitor->Fclk_BoosterFreq = input[idx + 5]; >>> + activity_monitor->Fclk_PD_Data_limit_c = input[idx + 6]; >>> + activity_monitor->Fclk_PD_Data_error_coeff = input[idx + 7]; >>> + activity_monitor->Fclk_PD_Data_error_rate_coeff = input[idx + 8]; >>> + } >>> >>> - ret = smu_cmn_update_table(smu, >>> - SMU_TABLE_ACTIVITY_MONITOR_COEFF, >>> - WORKLOAD_PPLIB_CUSTOM_BIT, >>> - (void *)(&activity_monitor_external), >>> - true); >>> - if (ret) { >>> - dev_err(smu->adev->dev, "[%s] Failed to set activity monitor!", __func__); >>> - return ret; >>> - } >>> + ret = smu_cmn_update_table(smu, >>> + SMU_TABLE_ACTIVITY_MONITOR_COEFF, >>> + WORKLOAD_PPLIB_CUSTOM_BIT, >>> + (void *)(&activity_monitor_external), >>> + true); >>> + if (ret) { >>> + dev_err(smu->adev->dev, "[%s] Failed to set activity monitor!", __func__); >>> + return ret; >>> } >>> >>> - /* conv PP_SMC_POWER_PROFILE* to WORKLOAD_PPLIB_*_BIT */ >>> - workload_type = smu_cmn_to_asic_specific_index(smu, >>> - CMN2ASIC_MAPPING_WORKLOAD, >>> - smu->power_profile_mode); >>> + return ret; >>> +} >>> >>> - if (workload_type < 0) >>> - return -EINVAL; >>> +static int smu_v13_0_0_set_power_profile_mode(struct smu_context *smu, >>> + u32 workload_mask, >>> + long *custom_params, >>> + u32 custom_params_max_idx) >>> +{ >>> + u32 backend_workload_mask = 0; >>> + int workload_type, ret, idx = -1, i; >>> >>> - selected_workload_mask = workload_mask = 1 << workload_type; >>> + smu_cmn_get_backend_workload_mask(smu, workload_mask, >>> + &backend_workload_mask); >>> >>> /* Add optimizations for SMU13.0.0/10. Reuse the power saving profile */ >>> if ((amdgpu_ip_version(smu->adev, MP1_HWIP, 0) == IP_VERSION(13, 0, 0) && >>> @@ -2658,15 +2652,48 @@ static int smu_v13_0_0_set_power_profile_mode(struct smu_context *smu, >>> CMN2ASIC_MAPPING_WORKLOAD, >>> PP_SMC_POWER_PROFILE_POWERSAVING); >>> if (workload_type >= 0) >>> - workload_mask |= 1 << workload_type; >>> + backend_workload_mask |= 1 << workload_type; >>> + } >>> + >>> + if (workload_mask & (1 << PP_SMC_POWER_PROFILE_CUSTOM)) { >>> + if (!smu->custom_profile_params) { >>> + smu->custom_profile_params = >>> + kzalloc(SMU_13_0_0_CUSTOM_PARAMS_SIZE, GFP_KERNEL); >>> + if (!smu->custom_profile_params) >>> + return -ENOMEM; >>> + } >>> + if (custom_params && custom_params_max_idx) { >>> + if (custom_params_max_idx != SMU_13_0_0_CUSTOM_PARAMS_COUNT) >>> + return -EINVAL; >>> + if (custom_params[0] >= SMU_13_0_0_CUSTOM_PARAMS_CLOCK_COUNT) >>> + return -EINVAL; >>> + idx = custom_params[0] * SMU_13_0_0_CUSTOM_PARAMS_COUNT; >>> + smu->custom_profile_params[idx] = 1; >>> + for (i = 1; i < custom_params_max_idx; i++) >>> + smu->custom_profile_params[idx + i] = custom_params[i]; >>> + } >>> + ret = smu_v13_0_0_set_power_profile_mode_coeff(smu, >>> + smu->custom_profile_params); >>> + if (ret) { >>> + if (idx != -1) >>> + smu->custom_profile_params[idx] = 0; >>> + return ret; >>> + } >>> + } else if (smu->custom_profile_params) { >>> + memset(smu->custom_profile_params, 0, SMU_13_0_0_CUSTOM_PARAMS_SIZE); >>> } >>> >>> ret = smu_cmn_send_smc_msg_with_param(smu, >>> - SMU_MSG_SetWorkloadMask, >>> - workload_mask, >>> - NULL); >>> - if (!ret) >>> - smu->workload_mask = selected_workload_mask; >>> + SMU_MSG_SetWorkloadMask, >>> + backend_workload_mask, >>> + NULL); >>> + if (ret) { >>> + dev_err(smu->adev->dev, "Failed to set workload mask 0x%08x\n", >>> + workload_mask); >>> + if (idx != -1) >>> + smu->custom_profile_params[idx] = 0; >>> + return ret; >>> + } >>> >>> return ret; >>> } >>> diff --git a/drivers/gpu/drm/amd/pm/swsmu/smu13/smu_v13_0_7_ppt.c b/drivers/gpu/drm/amd/pm/swsmu/smu13/smu_v13_0_7_ppt.c >>> index 34c1e0c7e1e4..f4ac403b8b36 100644 >>> --- a/drivers/gpu/drm/amd/pm/swsmu/smu13/smu_v13_0_7_ppt.c >>> +++ b/drivers/gpu/drm/amd/pm/swsmu/smu13/smu_v13_0_7_ppt.c >>> @@ -2530,78 +2530,110 @@ do { \ >>> return result; >>> } >>> >>> -static int smu_v13_0_7_set_power_profile_mode(struct smu_context *smu, long *input, uint32_t size) >>> +#define SMU_13_0_7_CUSTOM_PARAMS_COUNT 8 >>> +#define SMU_13_0_7_CUSTOM_PARAMS_CLOCK_COUNT 2 >>> +#define SMU_13_0_7_CUSTOM_PARAMS_SIZE (SMU_13_0_7_CUSTOM_PARAMS_CLOCK_COUNT * SMU_13_0_7_CUSTOM_PARAMS_COUNT * sizeof(long)) >>> + >>> +static int smu_v13_0_7_set_power_profile_mode_coeff(struct smu_context *smu, >>> + long *input) >>> { >>> >>> DpmActivityMonitorCoeffIntExternal_t activity_monitor_external; >>> DpmActivityMonitorCoeffInt_t *activity_monitor = >>> &(activity_monitor_external.DpmActivityMonitorCoeffInt); >>> - int workload_type, ret = 0; >>> + int ret, idx; >>> >>> - smu->power_profile_mode = input[size]; >>> + ret = smu_cmn_update_table(smu, >>> + SMU_TABLE_ACTIVITY_MONITOR_COEFF, WORKLOAD_PPLIB_CUSTOM_BIT, >>> + (void *)(&activity_monitor_external), false); >>> + if (ret) { >>> + dev_err(smu->adev->dev, "[%s] Failed to get activity monitor!", __func__); >>> + return ret; >>> + } >>> >>> - if (smu->power_profile_mode > PP_SMC_POWER_PROFILE_WINDOW3D) { >>> - dev_err(smu->adev->dev, "Invalid power profile mode %d\n", smu->power_profile_mode); >>> - return -EINVAL; >>> + idx = 0 * SMU_13_0_7_CUSTOM_PARAMS_COUNT; >>> + if (input[idx]) { >>> + /* Gfxclk */ >>> + activity_monitor->Gfx_ActiveHystLimit = input[idx + 1]; >>> + activity_monitor->Gfx_IdleHystLimit = input[idx + 2]; >>> + activity_monitor->Gfx_FPS = input[idx + 3]; >>> + activity_monitor->Gfx_MinActiveFreqType = input[idx + 4]; >>> + activity_monitor->Gfx_BoosterFreqType = input[idx + 5]; >>> + activity_monitor->Gfx_MinActiveFreq = input[idx + 6]; >>> + activity_monitor->Gfx_BoosterFreq = input[idx + 7]; >>> + } >>> + idx = 1 * SMU_13_0_7_CUSTOM_PARAMS_COUNT; >>> + if (input[idx]) { >>> + /* Fclk */ >>> + activity_monitor->Fclk_ActiveHystLimit = input[idx + 1]; >>> + activity_monitor->Fclk_IdleHystLimit = input[idx + 2]; >>> + activity_monitor->Fclk_FPS = input[idx + 3]; >>> + activity_monitor->Fclk_MinActiveFreqType = input[idx + 4]; >>> + activity_monitor->Fclk_BoosterFreqType = input[idx + 5]; >>> + activity_monitor->Fclk_MinActiveFreq = input[idx + 6]; >>> + activity_monitor->Fclk_BoosterFreq = input[idx + 7]; >>> } >>> >>> - if (smu->power_profile_mode == PP_SMC_POWER_PROFILE_CUSTOM) { >>> - if (size != 8) >>> - return -EINVAL; >>> + ret = smu_cmn_update_table(smu, >>> + SMU_TABLE_ACTIVITY_MONITOR_COEFF, WORKLOAD_PPLIB_CUSTOM_BIT, >>> + (void *)(&activity_monitor_external), true); >>> + if (ret) { >>> + dev_err(smu->adev->dev, "[%s] Failed to set activity monitor!", __func__); >>> + return ret; >>> + } >>> >>> - ret = smu_cmn_update_table(smu, >>> - SMU_TABLE_ACTIVITY_MONITOR_COEFF, WORKLOAD_PPLIB_CUSTOM_BIT, >>> - (void *)(&activity_monitor_external), false); >>> - if (ret) { >>> - dev_err(smu->adev->dev, "[%s] Failed to get activity monitor!", __func__); >>> - return ret; >>> - } >>> + return ret; >>> +} >>> >>> - switch (input[0]) { >>> - case 0: /* Gfxclk */ >>> - activity_monitor->Gfx_ActiveHystLimit = input[1]; >>> - activity_monitor->Gfx_IdleHystLimit = input[2]; >>> - activity_monitor->Gfx_FPS = input[3]; >>> - activity_monitor->Gfx_MinActiveFreqType = input[4]; >>> - activity_monitor->Gfx_BoosterFreqType = input[5]; >>> - activity_monitor->Gfx_MinActiveFreq = input[6]; >>> - activity_monitor->Gfx_BoosterFreq = input[7]; >>> - break; >>> - case 1: /* Fclk */ >>> - activity_monitor->Fclk_ActiveHystLimit = input[1]; >>> - activity_monitor->Fclk_IdleHystLimit = input[2]; >>> - activity_monitor->Fclk_FPS = input[3]; >>> - activity_monitor->Fclk_MinActiveFreqType = input[4]; >>> - activity_monitor->Fclk_BoosterFreqType = input[5]; >>> - activity_monitor->Fclk_MinActiveFreq = input[6]; >>> - activity_monitor->Fclk_BoosterFreq = input[7]; >>> - break; >>> - default: >>> - return -EINVAL; >>> +static int smu_v13_0_7_set_power_profile_mode(struct smu_context *smu, >>> + u32 workload_mask, >>> + long *custom_params, >>> + u32 custom_params_max_idx) >>> +{ >>> + u32 backend_workload_mask = 0; >>> + int ret, idx = -1, i; >>> + >>> + smu_cmn_get_backend_workload_mask(smu, workload_mask, >>> + &backend_workload_mask); >>> + >>> + if (workload_mask & (1 << PP_SMC_POWER_PROFILE_CUSTOM)) { >>> + if (!smu->custom_profile_params) { >>> + smu->custom_profile_params = >>> + kzalloc(SMU_13_0_7_CUSTOM_PARAMS_SIZE, GFP_KERNEL); >>> + if (!smu->custom_profile_params) >>> + return -ENOMEM; >>> } >>> - >>> - ret = smu_cmn_update_table(smu, >>> - SMU_TABLE_ACTIVITY_MONITOR_COEFF, WORKLOAD_PPLIB_CUSTOM_BIT, >>> - (void *)(&activity_monitor_external), true); >>> + if (custom_params && custom_params_max_idx) { >>> + if (custom_params_max_idx != SMU_13_0_7_CUSTOM_PARAMS_COUNT) >>> + return -EINVAL; >>> + if (custom_params[0] >= SMU_13_0_7_CUSTOM_PARAMS_CLOCK_COUNT) >>> + return -EINVAL; >>> + idx = custom_params[0] * SMU_13_0_7_CUSTOM_PARAMS_COUNT; >>> + smu->custom_profile_params[idx] = 1; >>> + for (i = 1; i < custom_params_max_idx; i++) >>> + smu->custom_profile_params[idx + i] = custom_params[i]; >>> + } >>> + ret = smu_v13_0_7_set_power_profile_mode_coeff(smu, >>> + smu->custom_profile_params); >>> if (ret) { >>> - dev_err(smu->adev->dev, "[%s] Failed to set activity monitor!", __func__); >>> + if (idx != -1) >>> + smu->custom_profile_params[idx] = 0; >>> return ret; >>> } >>> + } else if (smu->custom_profile_params) { >>> + memset(smu->custom_profile_params, 0, SMU_13_0_7_CUSTOM_PARAMS_SIZE); >>> } >>> >>> - /* conv PP_SMC_POWER_PROFILE* to WORKLOAD_PPLIB_*_BIT */ >>> - workload_type = smu_cmn_to_asic_specific_index(smu, >>> - CMN2ASIC_MAPPING_WORKLOAD, >>> - smu->power_profile_mode); >>> - if (workload_type < 0) >>> - return -EINVAL; >>> ret = smu_cmn_send_smc_msg_with_param(smu, SMU_MSG_SetWorkloadMask, >>> - 1 << workload_type, NULL); >>> + backend_workload_mask, NULL); >>> >>> - if (ret) >>> - dev_err(smu->adev->dev, "[%s] Failed to set work load mask!", __func__); >>> - else >>> - smu->workload_mask = (1 << workload_type); >>> + if (ret) { >>> + dev_err(smu->adev->dev, "Failed to set workload mask 0x%08x\n", >>> + workload_mask); >>> + if (idx != -1) >>> + smu->custom_profile_params[idx] = 0; >>> + return ret; >>> + } >>> >>> return ret; >>> } >>> diff --git a/drivers/gpu/drm/amd/pm/swsmu/smu14/smu_v14_0_2_ppt.c b/drivers/gpu/drm/amd/pm/swsmu/smu14/smu_v14_0_2_ppt.c >>> index 98e01a06add8..6a565ce74d5b 100644 >>> --- a/drivers/gpu/drm/amd/pm/swsmu/smu14/smu_v14_0_2_ppt.c >>> +++ b/drivers/gpu/drm/amd/pm/swsmu/smu14/smu_v14_0_2_ppt.c >>> @@ -1739,90 +1739,120 @@ static int smu_v14_0_2_get_power_profile_mode(struct smu_context *smu, >>> return size; >>> } >>> >>> -static int smu_v14_0_2_set_power_profile_mode(struct smu_context *smu, >>> - long *input, >>> - uint32_t size) >>> +#define SMU_14_0_2_CUSTOM_PARAMS_COUNT 9 >>> +#define SMU_14_0_2_CUSTOM_PARAMS_CLOCK_COUNT 2 >>> +#define SMU_14_0_2_CUSTOM_PARAMS_SIZE (SMU_14_0_2_CUSTOM_PARAMS_CLOCK_COUNT * SMU_14_0_2_CUSTOM_PARAMS_COUNT * sizeof(long)) >>> + >>> +static int smu_v14_0_2_set_power_profile_mode_coeff(struct smu_context *smu, >>> + long *input) >>> { >>> DpmActivityMonitorCoeffIntExternal_t activity_monitor_external; >>> DpmActivityMonitorCoeffInt_t *activity_monitor = >>> &(activity_monitor_external.DpmActivityMonitorCoeffInt); >>> - int workload_type, ret = 0; >>> - uint32_t current_profile_mode = smu->power_profile_mode; >>> - smu->power_profile_mode = input[size]; >>> + int ret, idx; >>> >>> - if (smu->power_profile_mode >= PP_SMC_POWER_PROFILE_COUNT) { >>> - dev_err(smu->adev->dev, "Invalid power profile mode %d\n", smu->power_profile_mode); >>> - return -EINVAL; >>> + ret = smu_cmn_update_table(smu, >>> + SMU_TABLE_ACTIVITY_MONITOR_COEFF, >>> + WORKLOAD_PPLIB_CUSTOM_BIT, >>> + (void *)(&activity_monitor_external), >>> + false); >>> + if (ret) { >>> + dev_err(smu->adev->dev, "[%s] Failed to get activity monitor!", __func__); >>> + return ret; >>> } >>> >>> - if (smu->power_profile_mode == PP_SMC_POWER_PROFILE_CUSTOM) { >>> - if (size != 9) >>> - return -EINVAL; >>> + idx = 0 * SMU_14_0_2_CUSTOM_PARAMS_COUNT; >>> + if (input[idx]) { >>> + /* Gfxclk */ >>> + activity_monitor->Gfx_FPS = input[idx + 1]; >>> + activity_monitor->Gfx_MinActiveFreqType = input[idx + 2]; >>> + activity_monitor->Gfx_MinActiveFreq = input[idx + 3]; >>> + activity_monitor->Gfx_BoosterFreqType = input[idx + 4]; >>> + activity_monitor->Gfx_BoosterFreq = input[idx + 5]; >>> + activity_monitor->Gfx_PD_Data_limit_c = input[idx + 6]; >>> + activity_monitor->Gfx_PD_Data_error_coeff = input[idx + 7]; >>> + activity_monitor->Gfx_PD_Data_error_rate_coeff = input[idx + 8]; >>> + } >>> + idx = 1 * SMU_14_0_2_CUSTOM_PARAMS_COUNT; >>> + if (input[idx]) { >>> + /* Fclk */ >>> + activity_monitor->Fclk_FPS = input[idx + 1]; >>> + activity_monitor->Fclk_MinActiveFreqType = input[idx + 2]; >>> + activity_monitor->Fclk_MinActiveFreq = input[idx + 3]; >>> + activity_monitor->Fclk_BoosterFreqType = input[idx + 4]; >>> + activity_monitor->Fclk_BoosterFreq = input[idx + 5]; >>> + activity_monitor->Fclk_PD_Data_limit_c = input[idx + 6]; >>> + activity_monitor->Fclk_PD_Data_error_coeff = input[idx + 7]; >>> + activity_monitor->Fclk_PD_Data_error_rate_coeff = input[idx + 8]; >>> + } >>> >>> - ret = smu_cmn_update_table(smu, >>> - SMU_TABLE_ACTIVITY_MONITOR_COEFF, >>> - WORKLOAD_PPLIB_CUSTOM_BIT, >>> - (void *)(&activity_monitor_external), >>> - false); >>> - if (ret) { >>> - dev_err(smu->adev->dev, "[%s] Failed to get activity monitor!", __func__); >>> - return ret; >>> - } >>> + ret = smu_cmn_update_table(smu, >>> + SMU_TABLE_ACTIVITY_MONITOR_COEFF, >>> + WORKLOAD_PPLIB_CUSTOM_BIT, >>> + (void *)(&activity_monitor_external), >>> + true); >>> + if (ret) { >>> + dev_err(smu->adev->dev, "[%s] Failed to set activity monitor!", __func__); >>> + return ret; >>> + } >>> >>> - switch (input[0]) { >>> - case 0: /* Gfxclk */ >>> - activity_monitor->Gfx_FPS = input[1]; >>> - activity_monitor->Gfx_MinActiveFreqType = input[2]; >>> - activity_monitor->Gfx_MinActiveFreq = input[3]; >>> - activity_monitor->Gfx_BoosterFreqType = input[4]; >>> - activity_monitor->Gfx_BoosterFreq = input[5]; >>> - activity_monitor->Gfx_PD_Data_limit_c = input[6]; >>> - activity_monitor->Gfx_PD_Data_error_coeff = input[7]; >>> - activity_monitor->Gfx_PD_Data_error_rate_coeff = input[8]; >>> - break; >>> - case 1: /* Fclk */ >>> - activity_monitor->Fclk_FPS = input[1]; >>> - activity_monitor->Fclk_MinActiveFreqType = input[2]; >>> - activity_monitor->Fclk_MinActiveFreq = input[3]; >>> - activity_monitor->Fclk_BoosterFreqType = input[4]; >>> - activity_monitor->Fclk_BoosterFreq = input[5]; >>> - activity_monitor->Fclk_PD_Data_limit_c = input[6]; >>> - activity_monitor->Fclk_PD_Data_error_coeff = input[7]; >>> - activity_monitor->Fclk_PD_Data_error_rate_coeff = input[8]; >>> - break; >>> - default: >>> - return -EINVAL; >>> - } >>> + return ret; >>> +} >>> >>> - ret = smu_cmn_update_table(smu, >>> - SMU_TABLE_ACTIVITY_MONITOR_COEFF, >>> - WORKLOAD_PPLIB_CUSTOM_BIT, >>> - (void *)(&activity_monitor_external), >>> - true); >>> - if (ret) { >>> - dev_err(smu->adev->dev, "[%s] Failed to set activity monitor!", __func__); >>> - return ret; >>> - } >>> - } >>> +static int smu_v14_0_2_set_power_profile_mode(struct smu_context *smu, >>> + u32 workload_mask, >>> + long *custom_params, >>> + u32 custom_params_max_idx) >>> +{ >>> + u32 backend_workload_mask = 0; >>> + int ret, idx = -1, i; >>> + >>> + smu_cmn_get_backend_workload_mask(smu, workload_mask, >>> + &backend_workload_mask); >>> >>> - if (smu->power_profile_mode == PP_SMC_POWER_PROFILE_COMPUTE) >>> + /* disable deep sleep if compute is enabled */ >>> + if (workload_mask & (1 << PP_SMC_POWER_PROFILE_COMPUTE)) >>> smu_v14_0_deep_sleep_control(smu, false); >>> - else if (current_profile_mode == PP_SMC_POWER_PROFILE_COMPUTE) >>> + else >>> smu_v14_0_deep_sleep_control(smu, true); >>> >>> - /* conv PP_SMC_POWER_PROFILE* to WORKLOAD_PPLIB_*_BIT */ >>> - workload_type = smu_cmn_to_asic_specific_index(smu, >>> - CMN2ASIC_MAPPING_WORKLOAD, >>> - smu->power_profile_mode); >>> - if (workload_type < 0) >>> - return -EINVAL; >>> + if (workload_mask & (1 << PP_SMC_POWER_PROFILE_CUSTOM)) { >>> + if (!smu->custom_profile_params) { >>> + smu->custom_profile_params = >>> + kzalloc(SMU_14_0_2_CUSTOM_PARAMS_SIZE, GFP_KERNEL); >>> + if (!smu->custom_profile_params) >>> + return -ENOMEM; >>> + } >>> + if (custom_params && custom_params_max_idx) { >>> + if (custom_params_max_idx != SMU_14_0_2_CUSTOM_PARAMS_COUNT) >>> + return -EINVAL; >>> + if (custom_params[0] >= SMU_14_0_2_CUSTOM_PARAMS_CLOCK_COUNT) >>> + return -EINVAL; >>> + idx = custom_params[0] * SMU_14_0_2_CUSTOM_PARAMS_COUNT; >>> + smu->custom_profile_params[idx] = 1; >>> + for (i = 1; i < custom_params_max_idx; i++) >>> + smu->custom_profile_params[idx + i] = custom_params[i]; >>> + } >>> + ret = smu_v14_0_2_set_power_profile_mode_coeff(smu, >>> + smu->custom_profile_params); >>> + if (ret) { >>> + if (idx != -1) >>> + smu->custom_profile_params[idx] = 0; >>> + return ret; >>> + } >>> + } else if (smu->custom_profile_params) { >>> + memset(smu->custom_profile_params, 0, SMU_14_0_2_CUSTOM_PARAMS_SIZE); >>> + } >>> >>> - ret = smu_cmn_send_smc_msg_with_param(smu, >>> - SMU_MSG_SetWorkloadMask, >>> - 1 << workload_type, >>> - NULL); >>> - if (!ret) >>> - smu->workload_mask = 1 << workload_type; >>> + ret = smu_cmn_send_smc_msg_with_param(smu, SMU_MSG_SetWorkloadMask, >>> + backend_workload_mask, NULL); >>> + if (ret) { >>> + dev_err(smu->adev->dev, "Failed to set workload mask 0x%08x\n", >>> + workload_mask); >>> + if (idx != -1) >>> + smu->custom_profile_params[idx] = 0; >>> + return ret; >>> + } >>> >>> return ret; >>> } >>> diff --git a/drivers/gpu/drm/amd/pm/swsmu/smu_cmn.c b/drivers/gpu/drm/amd/pm/swsmu/smu_cmn.c >>> index 007a81e108ec..8f92b2777726 100644 >>> --- a/drivers/gpu/drm/amd/pm/swsmu/smu_cmn.c >>> +++ b/drivers/gpu/drm/amd/pm/swsmu/smu_cmn.c >>> @@ -1221,3 +1221,28 @@ void smu_cmn_generic_plpd_policy_desc(struct smu_dpm_policy *policy) >>> { >>> policy->desc = &xgmi_plpd_policy_desc; >>> } >>> + >>> +void smu_cmn_get_backend_workload_mask(struct smu_context *smu, >>> + u32 workload_mask, >>> + u32 *backend_workload_mask) >>> +{ >>> + int workload_type; >>> + u32 profile_mode; >>> + >>> + *backend_workload_mask = 0; >>> + >>> + for (profile_mode = 0; profile_mode < PP_SMC_POWER_PROFILE_COUNT; profile_mode++) { >>> + if (!(workload_mask & (1 << profile_mode))) >>> + continue; >>> + >>> + /* conv PP_SMC_POWER_PROFILE* to WORKLOAD_PPLIB_*_BIT */ >>> + workload_type = smu_cmn_to_asic_specific_index(smu, >>> + CMN2ASIC_MAPPING_WORKLOAD, >>> + profile_mode); >>> + >>> + if (workload_type < 0) >>> + continue; >>> + >>> + *backend_workload_mask |= 1 << workload_type; >>> + } >>> +} >>> diff --git a/drivers/gpu/drm/amd/pm/swsmu/smu_cmn.h b/drivers/gpu/drm/amd/pm/swsmu/smu_cmn.h >>> index 1de685defe85..a020277dec3e 100644 >>> --- a/drivers/gpu/drm/amd/pm/swsmu/smu_cmn.h >>> +++ b/drivers/gpu/drm/amd/pm/swsmu/smu_cmn.h >>> @@ -147,5 +147,9 @@ bool smu_cmn_is_audio_func_enabled(struct amdgpu_device *adev); >>> void smu_cmn_generic_soc_policy_desc(struct smu_dpm_policy *policy); >>> void smu_cmn_generic_plpd_policy_desc(struct smu_dpm_policy *policy); >>> >>> +void smu_cmn_get_backend_workload_mask(struct smu_context *smu, >>> + u32 workload_mask, >>> + u32 *backend_workload_mask); >>> + >>> #endif >>> #endif