Re: [PATCH v4 4/5] sched: Rename arch_update_thermal_pressure into arch_update_hw_pressure

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On 01/09/24 17:46, Vincent Guittot wrote:
> Now that cpufreq provides a pressure value to the scheduler, rename
> arch_update_thermal_pressure into HW pressure to reflect that it returns
> a pressure applied by HW (i.e. with a high frequency change) and not
> always related to thermal mitigation but also generated by max current
> limitation as an example. Such high frequency signal needs filtering to be
> smoothed and provide a value that reflects the average available capacity
> into the scheduler time scale.
> 
> Signed-off-by: Vincent Guittot <vincent.guittot@xxxxxxxxxx>
> ---

Reviewed-by: Qais Yousef <qyousef@xxxxxxxxxxx>

>  arch/arm/include/asm/topology.h               |  6 ++---
>  arch/arm64/include/asm/topology.h             |  6 ++---
>  drivers/base/arch_topology.c                  | 26 +++++++++----------
>  drivers/cpufreq/qcom-cpufreq-hw.c             |  4 +--
>  include/linux/arch_topology.h                 |  8 +++---
>  include/linux/sched/topology.h                |  8 +++---
>  .../{thermal_pressure.h => hw_pressure.h}     | 14 +++++-----
>  include/trace/events/sched.h                  |  2 +-
>  init/Kconfig                                  | 12 ++++-----
>  kernel/sched/core.c                           |  8 +++---
>  kernel/sched/fair.c                           | 16 ++++++------
>  kernel/sched/pelt.c                           | 18 ++++++-------
>  kernel/sched/pelt.h                           | 16 ++++++------
>  kernel/sched/sched.h                          | 10 +++----
>  14 files changed, 77 insertions(+), 77 deletions(-)
>  rename include/trace/events/{thermal_pressure.h => hw_pressure.h} (55%)
> 
> diff --git a/arch/arm/include/asm/topology.h b/arch/arm/include/asm/topology.h
> index 853c4f81ba4a..ad36b6570067 100644
> --- a/arch/arm/include/asm/topology.h
> +++ b/arch/arm/include/asm/topology.h
> @@ -22,9 +22,9 @@
>  /* Enable topology flag updates */
>  #define arch_update_cpu_topology topology_update_cpu_topology
>  
> -/* Replace task scheduler's default thermal pressure API */
> -#define arch_scale_thermal_pressure topology_get_thermal_pressure
> -#define arch_update_thermal_pressure	topology_update_thermal_pressure
> +/* Replace task scheduler's default HW pressure API */
> +#define arch_scale_hw_pressure topology_get_hw_pressure
> +#define arch_update_hw_pressure	topology_update_hw_pressure
>  
>  #else
>  
> diff --git a/arch/arm64/include/asm/topology.h b/arch/arm64/include/asm/topology.h
> index a323b109b9c4..0f6ef432fb84 100644
> --- a/arch/arm64/include/asm/topology.h
> +++ b/arch/arm64/include/asm/topology.h
> @@ -35,9 +35,9 @@ void update_freq_counters_refs(void);
>  /* Enable topology flag updates */
>  #define arch_update_cpu_topology topology_update_cpu_topology
>  
> -/* Replace task scheduler's default thermal pressure API */
> -#define arch_scale_thermal_pressure topology_get_thermal_pressure
> -#define arch_update_thermal_pressure	topology_update_thermal_pressure
> +/* Replace task scheduler's default HW pressure API */
> +#define arch_scale_hw_pressure topology_get_hw_pressure
> +#define arch_update_hw_pressure	topology_update_hw_pressure
>  
>  #include <asm-generic/topology.h>
>  
> diff --git a/drivers/base/arch_topology.c b/drivers/base/arch_topology.c
> index 5aaa0865625d..331ffc468ceb 100644
> --- a/drivers/base/arch_topology.c
> +++ b/drivers/base/arch_topology.c
> @@ -22,7 +22,7 @@
>  #include <linux/units.h>
>  
>  #define CREATE_TRACE_POINTS
> -#include <trace/events/thermal_pressure.h>
> +#include <trace/events/hw_pressure.h>
>  
>  static DEFINE_PER_CPU(struct scale_freq_data __rcu *, sft_data);
>  static struct cpumask scale_freq_counters_mask;
> @@ -160,26 +160,26 @@ void topology_set_cpu_scale(unsigned int cpu, unsigned long capacity)
>  	per_cpu(cpu_scale, cpu) = capacity;
>  }
>  
> -DEFINE_PER_CPU(unsigned long, thermal_pressure);
> +DEFINE_PER_CPU(unsigned long, hw_pressure);
>  
>  /**
> - * topology_update_thermal_pressure() - Update thermal pressure for CPUs
> + * topology_update_hw_pressure() - Update HW pressure for CPUs
>   * @cpus        : The related CPUs for which capacity has been reduced
>   * @capped_freq : The maximum allowed frequency that CPUs can run at
>   *
> - * Update the value of thermal pressure for all @cpus in the mask. The
> + * Update the value of HW pressure for all @cpus in the mask. The
>   * cpumask should include all (online+offline) affected CPUs, to avoid
>   * operating on stale data when hot-plug is used for some CPUs. The
>   * @capped_freq reflects the currently allowed max CPUs frequency due to
> - * thermal capping. It might be also a boost frequency value, which is bigger
> + * HW capping. It might be also a boost frequency value, which is bigger
>   * than the internal 'capacity_freq_ref' max frequency. In such case the
>   * pressure value should simply be removed, since this is an indication that
> - * there is no thermal throttling. The @capped_freq must be provided in kHz.
> + * there is no HW throttling. The @capped_freq must be provided in kHz.
>   */
> -void topology_update_thermal_pressure(const struct cpumask *cpus,
> +void topology_update_hw_pressure(const struct cpumask *cpus,
>  				      unsigned long capped_freq)
>  {
> -	unsigned long max_capacity, capacity, th_pressure;
> +	unsigned long max_capacity, capacity, hw_pressure;
>  	u32 max_freq;
>  	int cpu;
>  
> @@ -189,21 +189,21 @@ void topology_update_thermal_pressure(const struct cpumask *cpus,
>  
>  	/*
>  	 * Handle properly the boost frequencies, which should simply clean
> -	 * the thermal pressure value.
> +	 * the HW pressure value.
>  	 */
>  	if (max_freq <= capped_freq)
>  		capacity = max_capacity;
>  	else
>  		capacity = mult_frac(max_capacity, capped_freq, max_freq);
>  
> -	th_pressure = max_capacity - capacity;
> +	hw_pressure = max_capacity - capacity;
>  
> -	trace_thermal_pressure_update(cpu, th_pressure);
> +	trace_hw_pressure_update(cpu, hw_pressure);
>  
>  	for_each_cpu(cpu, cpus)
> -		WRITE_ONCE(per_cpu(thermal_pressure, cpu), th_pressure);
> +		WRITE_ONCE(per_cpu(hw_pressure, cpu), hw_pressure);
>  }
> -EXPORT_SYMBOL_GPL(topology_update_thermal_pressure);
> +EXPORT_SYMBOL_GPL(topology_update_hw_pressure);
>  
>  static ssize_t cpu_capacity_show(struct device *dev,
>  				 struct device_attribute *attr,
> diff --git a/drivers/cpufreq/qcom-cpufreq-hw.c b/drivers/cpufreq/qcom-cpufreq-hw.c
> index 70b0f21968a0..ec8df5496a0c 100644
> --- a/drivers/cpufreq/qcom-cpufreq-hw.c
> +++ b/drivers/cpufreq/qcom-cpufreq-hw.c
> @@ -347,8 +347,8 @@ static void qcom_lmh_dcvs_notify(struct qcom_cpufreq_data *data)
>  
>  	throttled_freq = freq_hz / HZ_PER_KHZ;
>  
> -	/* Update thermal pressure (the boost frequencies are accepted) */
> -	arch_update_thermal_pressure(policy->related_cpus, throttled_freq);
> +	/* Update HW pressure (the boost frequencies are accepted) */
> +	arch_update_hw_pressure(policy->related_cpus, throttled_freq);
>  
>  	/*
>  	 * In the unlikely case policy is unregistered do not enable
> diff --git a/include/linux/arch_topology.h b/include/linux/arch_topology.h
> index a63d61ca55af..b721f360d759 100644
> --- a/include/linux/arch_topology.h
> +++ b/include/linux/arch_topology.h
> @@ -60,14 +60,14 @@ void topology_scale_freq_tick(void);
>  void topology_set_scale_freq_source(struct scale_freq_data *data, const struct cpumask *cpus);
>  void topology_clear_scale_freq_source(enum scale_freq_source source, const struct cpumask *cpus);
>  
> -DECLARE_PER_CPU(unsigned long, thermal_pressure);
> +DECLARE_PER_CPU(unsigned long, hw_pressure);
>  
> -static inline unsigned long topology_get_thermal_pressure(int cpu)
> +static inline unsigned long topology_get_hw_pressure(int cpu)
>  {
> -	return per_cpu(thermal_pressure, cpu);
> +	return per_cpu(hw_pressure, cpu);
>  }
>  
> -void topology_update_thermal_pressure(const struct cpumask *cpus,
> +void topology_update_hw_pressure(const struct cpumask *cpus,
>  				      unsigned long capped_freq);
>  
>  struct cpu_topology {
> diff --git a/include/linux/sched/topology.h b/include/linux/sched/topology.h
> index a6e04b4a21d7..e3b2cf7de018 100644
> --- a/include/linux/sched/topology.h
> +++ b/include/linux/sched/topology.h
> @@ -264,17 +264,17 @@ unsigned long arch_scale_cpu_capacity(int cpu)
>  }
>  #endif
>  
> -#ifndef arch_scale_thermal_pressure
> +#ifndef arch_scale_hw_pressure
>  static __always_inline
> -unsigned long arch_scale_thermal_pressure(int cpu)
> +unsigned long arch_scale_hw_pressure(int cpu)
>  {
>  	return 0;
>  }
>  #endif
>  
> -#ifndef arch_update_thermal_pressure
> +#ifndef arch_update_hw_pressure
>  static __always_inline
> -void arch_update_thermal_pressure(const struct cpumask *cpus,
> +void arch_update_hw_pressure(const struct cpumask *cpus,
>  				  unsigned long capped_frequency)
>  { }
>  #endif
> diff --git a/include/trace/events/thermal_pressure.h b/include/trace/events/hw_pressure.h
> similarity index 55%
> rename from include/trace/events/thermal_pressure.h
> rename to include/trace/events/hw_pressure.h
> index b68680201360..b9cd68854128 100644
> --- a/include/trace/events/thermal_pressure.h
> +++ b/include/trace/events/hw_pressure.h
> @@ -1,27 +1,27 @@
>  /* SPDX-License-Identifier: GPL-2.0 */
>  #undef TRACE_SYSTEM
> -#define TRACE_SYSTEM thermal_pressure
> +#define TRACE_SYSTEM hw_pressure
>  
>  #if !defined(_TRACE_THERMAL_PRESSURE_H) || defined(TRACE_HEADER_MULTI_READ)
>  #define _TRACE_THERMAL_PRESSURE_H
>  
>  #include <linux/tracepoint.h>
>  
> -TRACE_EVENT(thermal_pressure_update,
> -	TP_PROTO(int cpu, unsigned long thermal_pressure),
> -	TP_ARGS(cpu, thermal_pressure),
> +TRACE_EVENT(hw_pressure_update,
> +	TP_PROTO(int cpu, unsigned long hw_pressure),
> +	TP_ARGS(cpu, hw_pressure),
>  
>  	TP_STRUCT__entry(
> -		__field(unsigned long, thermal_pressure)
> +		__field(unsigned long, hw_pressure)
>  		__field(int, cpu)
>  	),
>  
>  	TP_fast_assign(
> -		__entry->thermal_pressure = thermal_pressure;
> +		__entry->hw_pressure = hw_pressure;
>  		__entry->cpu = cpu;
>  	),
>  
> -	TP_printk("cpu=%d thermal_pressure=%lu", __entry->cpu, __entry->thermal_pressure)
> +	TP_printk("cpu=%d hw_pressure=%lu", __entry->cpu, __entry->hw_pressure)
>  );
>  #endif /* _TRACE_THERMAL_PRESSURE_H */
>  
> diff --git a/include/trace/events/sched.h b/include/trace/events/sched.h
> index dbb01b4b7451..d115d64c4011 100644
> --- a/include/trace/events/sched.h
> +++ b/include/trace/events/sched.h
> @@ -752,7 +752,7 @@ DECLARE_TRACE(pelt_dl_tp,
>  	TP_PROTO(struct rq *rq),
>  	TP_ARGS(rq));
>  
> -DECLARE_TRACE(pelt_thermal_tp,
> +DECLARE_TRACE(pelt_hw_tp,
>  	TP_PROTO(struct rq *rq),
>  	TP_ARGS(rq));
>  
> diff --git a/init/Kconfig b/init/Kconfig
> index 9ffb103fc927..37ceeb67e01c 100644
> --- a/init/Kconfig
> +++ b/init/Kconfig
> @@ -538,24 +538,24 @@ config HAVE_SCHED_AVG_IRQ
>  	depends on IRQ_TIME_ACCOUNTING || PARAVIRT_TIME_ACCOUNTING
>  	depends on SMP
>  
> -config SCHED_THERMAL_PRESSURE
> +config SCHED_HW_PRESSURE
>  	bool
>  	default y if ARM && ARM_CPU_TOPOLOGY
>  	default y if ARM64
>  	depends on SMP
>  	depends on CPU_FREQ_THERMAL
>  	help
> -	  Select this option to enable thermal pressure accounting in the
> -	  scheduler. Thermal pressure is the value conveyed to the scheduler
> +	  Select this option to enable HW pressure accounting in the
> +	  scheduler. HW pressure is the value conveyed to the scheduler
>  	  that reflects the reduction in CPU compute capacity resulted from
> -	  thermal throttling. Thermal throttling occurs when the performance of
> -	  a CPU is capped due to high operating temperatures.
> +	  HW throttling. HW throttling occurs when the performance of
> +	  a CPU is capped due to high operating temperatures as an example.
>  
>  	  If selected, the scheduler will be able to balance tasks accordingly,
>  	  i.e. put less load on throttled CPUs than on non/less throttled ones.
>  
>  	  This requires the architecture to implement
> -	  arch_update_thermal_pressure() and arch_scale_thermal_pressure().
> +	  arch_update_hw_pressure() and arch_scale_thermal_pressure().
>  
>  config BSD_PROCESS_ACCT
>  	bool "BSD Process Accounting"
> diff --git a/kernel/sched/core.c b/kernel/sched/core.c
> index db4be4921e7f..a6f084bdf1c5 100644
> --- a/kernel/sched/core.c
> +++ b/kernel/sched/core.c
> @@ -107,7 +107,7 @@ EXPORT_TRACEPOINT_SYMBOL_GPL(pelt_rt_tp);
>  EXPORT_TRACEPOINT_SYMBOL_GPL(pelt_dl_tp);
>  EXPORT_TRACEPOINT_SYMBOL_GPL(pelt_irq_tp);
>  EXPORT_TRACEPOINT_SYMBOL_GPL(pelt_se_tp);
> -EXPORT_TRACEPOINT_SYMBOL_GPL(pelt_thermal_tp);
> +EXPORT_TRACEPOINT_SYMBOL_GPL(pelt_hw_tp);
>  EXPORT_TRACEPOINT_SYMBOL_GPL(sched_cpu_capacity_tp);
>  EXPORT_TRACEPOINT_SYMBOL_GPL(sched_overutilized_tp);
>  EXPORT_TRACEPOINT_SYMBOL_GPL(sched_util_est_cfs_tp);
> @@ -5658,7 +5658,7 @@ void scheduler_tick(void)
>  	struct rq *rq = cpu_rq(cpu);
>  	struct task_struct *curr = rq->curr;
>  	struct rq_flags rf;
> -	unsigned long thermal_pressure;
> +	unsigned long hw_pressure;
>  	u64 resched_latency;
>  
>  	if (housekeeping_cpu(cpu, HK_TYPE_TICK))
> @@ -5669,8 +5669,8 @@ void scheduler_tick(void)
>  	rq_lock(rq, &rf);
>  
>  	update_rq_clock(rq);
> -	thermal_pressure = arch_scale_thermal_pressure(cpu_of(rq));
> -	update_thermal_load_avg(rq_clock_thermal(rq), rq, thermal_pressure);
> +	hw_pressure = arch_scale_hw_pressure(cpu_of(rq));
> +	update_hw_load_avg(rq_clock_hw(rq), rq, hw_pressure);
>  	curr->sched_class->task_tick(rq, curr, 0);
>  	if (sched_feat(LATENCY_WARN))
>  		resched_latency = cpu_resched_latency(rq);
> diff --git a/kernel/sched/fair.c b/kernel/sched/fair.c
> index e54bbf8b4936..f1c3d600d6d6 100644
> --- a/kernel/sched/fair.c
> +++ b/kernel/sched/fair.c
> @@ -78,7 +78,7 @@ static unsigned int normalized_sysctl_sched_base_slice	= 750000ULL;
>  
>  const_debug unsigned int sysctl_sched_migration_cost	= 500000UL;
>  
> -int sched_thermal_decay_shift;
> +int sched_hw_decay_shift;
>  static int __init setup_sched_thermal_decay_shift(char *str)
>  {
>  	int _shift = 0;
> @@ -86,7 +86,7 @@ static int __init setup_sched_thermal_decay_shift(char *str)
>  	if (kstrtoint(str, 0, &_shift))
>  		pr_warn("Unable to set scheduler thermal pressure decay shift parameter\n");
>  
> -	sched_thermal_decay_shift = clamp(_shift, 0, 10);
> +	sched_hw_decay_shift = clamp(_shift, 0, 10);
>  	return 1;
>  }
>  __setup("sched_thermal_decay_shift=", setup_sched_thermal_decay_shift);
> @@ -4914,7 +4914,7 @@ static inline unsigned long get_actual_cpu_capacity(int cpu)
>  {
>  	unsigned long capacity = arch_scale_cpu_capacity(cpu);
>  
> -	capacity -= max(thermal_load_avg(cpu_rq(cpu)), cpufreq_get_pressure(cpu));
> +	capacity -= max(hw_load_avg(cpu_rq(cpu)), cpufreq_get_pressure(cpu));
>  
>  	return capacity;
>  }
> @@ -4947,7 +4947,7 @@ static inline int util_fits_cpu(unsigned long util,
>  	 * Similarly if a task is capped to arch_scale_cpu_capacity(little_cpu), it
>  	 * should fit a little cpu even if there's some pressure.
>  	 *
> -	 * Only exception is for thermal pressure since it has a direct impact
> +	 * Only exception is for HW or cpufreq pressure since it has a direct impact
>  	 * on available OPP of the system.
>  	 *
>  	 * We honour it for uclamp_min only as a drop in performance level
> @@ -9202,7 +9202,7 @@ static inline bool others_have_blocked(struct rq *rq)
>  	if (READ_ONCE(rq->avg_dl.util_avg))
>  		return true;
>  
> -	if (thermal_load_avg(rq))
> +	if (hw_load_avg(rq))
>  		return true;
>  
>  #ifdef CONFIG_HAVE_SCHED_AVG_IRQ
> @@ -9234,7 +9234,7 @@ static bool __update_blocked_others(struct rq *rq, bool *done)
>  {
>  	const struct sched_class *curr_class;
>  	u64 now = rq_clock_pelt(rq);
> -	unsigned long thermal_pressure;
> +	unsigned long hw_pressure;
>  	bool decayed;
>  
>  	/*
> @@ -9243,11 +9243,11 @@ static bool __update_blocked_others(struct rq *rq, bool *done)
>  	 */
>  	curr_class = rq->curr->sched_class;
>  
> -	thermal_pressure = arch_scale_thermal_pressure(cpu_of(rq));
> +	hw_pressure = arch_scale_hw_pressure(cpu_of(rq));
>  
>  	decayed = update_rt_rq_load_avg(now, rq, curr_class == &rt_sched_class) |
>  		  update_dl_rq_load_avg(now, rq, curr_class == &dl_sched_class) |
> -		  update_thermal_load_avg(rq_clock_thermal(rq), rq, thermal_pressure) |
> +		  update_hw_load_avg(rq_clock_hw(rq), rq, hw_pressure) |
>  		  update_irq_load_avg(rq, 0);
>  
>  	if (others_have_blocked(rq))
> diff --git a/kernel/sched/pelt.c b/kernel/sched/pelt.c
> index 63b6cf898220..f951c44f1d52 100644
> --- a/kernel/sched/pelt.c
> +++ b/kernel/sched/pelt.c
> @@ -384,30 +384,30 @@ int update_dl_rq_load_avg(u64 now, struct rq *rq, int running)
>  	return 0;
>  }
>  
> -#ifdef CONFIG_SCHED_THERMAL_PRESSURE
> +#ifdef CONFIG_SCHED_HW_PRESSURE
>  /*
> - * thermal:
> + * hardware:
>   *
>   *   load_sum = \Sum se->avg.load_sum but se->avg.load_sum is not tracked
>   *
>   *   util_avg and runnable_load_avg are not supported and meaningless.
>   *
>   * Unlike rt/dl utilization tracking that track time spent by a cpu
> - * running a rt/dl task through util_avg, the average thermal pressure is
> - * tracked through load_avg. This is because thermal pressure signal is
> + * running a rt/dl task through util_avg, the average HW pressure is
> + * tracked through load_avg. This is because HW pressure signal is
>   * time weighted "delta" capacity unlike util_avg which is binary.
>   * "delta capacity" =  actual capacity  -
> - *			capped capacity a cpu due to a thermal event.
> + *			capped capacity a cpu due to a HW event.
>   */
>  
> -int update_thermal_load_avg(u64 now, struct rq *rq, u64 capacity)
> +int update_hw_load_avg(u64 now, struct rq *rq, u64 capacity)
>  {
> -	if (___update_load_sum(now, &rq->avg_thermal,
> +	if (___update_load_sum(now, &rq->avg_hw,
>  			       capacity,
>  			       capacity,
>  			       capacity)) {
> -		___update_load_avg(&rq->avg_thermal, 1);
> -		trace_pelt_thermal_tp(rq);
> +		___update_load_avg(&rq->avg_hw, 1);
> +		trace_pelt_hw_tp(rq);
>  		return 1;
>  	}
>  
> diff --git a/kernel/sched/pelt.h b/kernel/sched/pelt.h
> index 9e1083465fbc..2150062949d4 100644
> --- a/kernel/sched/pelt.h
> +++ b/kernel/sched/pelt.h
> @@ -7,21 +7,21 @@ int __update_load_avg_cfs_rq(u64 now, struct cfs_rq *cfs_rq);
>  int update_rt_rq_load_avg(u64 now, struct rq *rq, int running);
>  int update_dl_rq_load_avg(u64 now, struct rq *rq, int running);
>  
> -#ifdef CONFIG_SCHED_THERMAL_PRESSURE
> -int update_thermal_load_avg(u64 now, struct rq *rq, u64 capacity);
> +#ifdef CONFIG_SCHED_HW_PRESSURE
> +int update_hw_load_avg(u64 now, struct rq *rq, u64 capacity);
>  
> -static inline u64 thermal_load_avg(struct rq *rq)
> +static inline u64 hw_load_avg(struct rq *rq)
>  {
> -	return READ_ONCE(rq->avg_thermal.load_avg);
> +	return READ_ONCE(rq->avg_hw.load_avg);
>  }
>  #else
>  static inline int
> -update_thermal_load_avg(u64 now, struct rq *rq, u64 capacity)
> +update_hw_load_avg(u64 now, struct rq *rq, u64 capacity)
>  {
>  	return 0;
>  }
>  
> -static inline u64 thermal_load_avg(struct rq *rq)
> +static inline u64 hw_load_avg(struct rq *rq)
>  {
>  	return 0;
>  }
> @@ -202,12 +202,12 @@ update_dl_rq_load_avg(u64 now, struct rq *rq, int running)
>  }
>  
>  static inline int
> -update_thermal_load_avg(u64 now, struct rq *rq, u64 capacity)
> +update_hw_load_avg(u64 now, struct rq *rq, u64 capacity)
>  {
>  	return 0;
>  }
>  
> -static inline u64 thermal_load_avg(struct rq *rq)
> +static inline u64 hw_load_avg(struct rq *rq)
>  {
>  	return 0;
>  }
> diff --git a/kernel/sched/sched.h b/kernel/sched/sched.h
> index e58a54bda77d..677d24202eec 100644
> --- a/kernel/sched/sched.h
> +++ b/kernel/sched/sched.h
> @@ -1078,8 +1078,8 @@ struct rq {
>  #ifdef CONFIG_HAVE_SCHED_AVG_IRQ
>  	struct sched_avg	avg_irq;
>  #endif
> -#ifdef CONFIG_SCHED_THERMAL_PRESSURE
> -	struct sched_avg	avg_thermal;
> +#ifdef CONFIG_SCHED_HW_PRESSURE
> +	struct sched_avg	avg_hw;
>  #endif
>  	u64			idle_stamp;
>  	u64			avg_idle;
> @@ -1531,11 +1531,11 @@ static inline u64 rq_clock_task(struct rq *rq)
>   *	3			256
>   *	4			512
>   */
> -extern int sched_thermal_decay_shift;
> +extern int sched_hw_decay_shift;
>  
> -static inline u64 rq_clock_thermal(struct rq *rq)
> +static inline u64 rq_clock_hw(struct rq *rq)
>  {
> -	return rq_clock_task(rq) >> sched_thermal_decay_shift;
> +	return rq_clock_task(rq) >> sched_hw_decay_shift;
>  }
>  
>  static inline void rq_clock_skip_update(struct rq *rq)
> -- 
> 2.34.1
> 




[Index of Archives]     [Kernel Newbies]     [Security]     [Netfilter]     [Bugtraq]     [Linux FS]     [Yosemite Forum]     [MIPS Linux]     [ARM Linux]     [Linux Security]     [Linux RAID]     [Samba]     [Video 4 Linux]     [Device Mapper]     [Linux Resources]

  Powered by Linux