Re: [PATCH v8 02/16] sched/core: Add bucket local max tracking

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

 



On 02-Apr 11:41, Patrick Bellasi wrote:
> Because of bucketization, different task-specific clamp values are
> tracked in the same bucket.  For example, with 20% bucket size and
> assuming to have:
>   Task1: util_min=25%
>   Task2: util_min=35%
> both tasks will be refcounted in the [20..39]% bucket and always boosted
> only up to 20% thus implementing a simple floor aggregation normally
> used in histograms.
> 
> In systems with only few and well-defined clamp values, it would be
> useful to track the exact clamp value required by a task whenever
> possible. For example, if a system requires only 23% and 47% boost
> values then it's possible to track the exact boost required by each
> task using only 3 buckets of ~33% size each.
> 
> Introduce a mechanism to max aggregate the requested clamp values of
> RUNNABLE tasks in the same bucket. Keep it simple by resetting the
> bucket value to its base value only when a bucket becomes inactive.
> Allow a limited and controlled overboosting margin for tasks recounted
> in the same bucket.
> 
> In systems where the boost values are not known in advance, it is still
> possible to control the maximum acceptable overboosting margin by tuning
> the number of clamp groups. For example, 20 groups ensure a 5% maximum
> overboost.
> 
> Remove the rq bucket initialization code since a correct bucket value
> is now computed when a task is refcounted into a CPU's rq.
> 
> Signed-off-by: Patrick Bellasi <patrick.bellasi@xxxxxxx>
> Cc: Ingo Molnar <mingo@xxxxxxxxxx>
> Cc: Peter Zijlstra <peterz@xxxxxxxxxxxxx>
> 
> --
> Changes in v8:
>  Message-ID: <20190313193916.GQ2482@xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx>
>  - split this code out from the previous patch
> ---
>  kernel/sched/core.c | 46 ++++++++++++++++++++++++++-------------------
>  1 file changed, 27 insertions(+), 19 deletions(-)
> 
> diff --git a/kernel/sched/core.c b/kernel/sched/core.c
> index 032211b72110..6e1beae5f348 100644
> --- a/kernel/sched/core.c
> +++ b/kernel/sched/core.c
> @@ -778,6 +778,11 @@ unsigned int uclamp_rq_max_value(struct rq *rq, unsigned int clamp_id)
>   * When a task is enqueued on a rq, the clamp bucket currently defined by the
>   * task's uclamp::bucket_id is refcounted on that rq. This also immediately
>   * updates the rq's clamp value if required.
> + *
> + * Tasks can have a task-specific value requested from user-space, track
> + * within each bucket the maximum value for tasks refcounted in it.
> + * This "local max aggregation" allows to track the exact "requested" value
> + * for each bucket when all its RUNNABLE tasks require the same clamp.
>   */
>  static inline void uclamp_rq_inc_id(struct task_struct *p, struct rq *rq,
>  				    unsigned int clamp_id)
> @@ -789,8 +794,15 @@ static inline void uclamp_rq_inc_id(struct task_struct *p, struct rq *rq,
>  	bucket = &uc_rq->bucket[uc_se->bucket_id];
>  	bucket->tasks++;
>  
> +	/*
> +	 * Local max aggregation: rq buckets always track the max
> +	 * "requested" clamp value of its RUNNABLE tasks.
> +	 */
> +	if (uc_se->value > bucket->value)
> +		bucket->value = uc_se->value;
> +
>  	if (uc_se->value > READ_ONCE(uc_rq->value))
> -		WRITE_ONCE(uc_rq->value, bucket->value);
> +		WRITE_ONCE(uc_rq->value, uc_se->value);
>  }
>  
>  /*
> @@ -815,6 +827,12 @@ static inline void uclamp_rq_dec_id(struct task_struct *p, struct rq *rq,
>  	if (likely(bucket->tasks))
>  		bucket->tasks--;
>  
> +	/*
> +	 * Keep "local max aggregation" simple and accept to (possibly)
> +	 * overboost some RUNNABLE tasks in the same bucket.
> +	 * The rq clamp bucket value is reset to its base value whenever
> +	 * there are no more RUNNABLE tasks refcounting it.
> +	 */
>  	if (likely(bucket->tasks))
>  		return;
>  
> @@ -824,8 +842,14 @@ static inline void uclamp_rq_dec_id(struct task_struct *p, struct rq *rq,
>  	 * e.g. due to future modification, warn and fixup the expected value.
>  	 */
>  	SCHED_WARN_ON(bucket->value > rq_clamp);
> -	if (bucket->value >= rq_clamp)
> +	if (bucket->value >= rq_clamp) {
> +		/*
> +		 * Reset clamp bucket value to its nominal value whenever
> +		 * there are anymore RUNNABLE tasks refcounting it.
> +		 */
> +		bucket->value = uclamp_bucket_base_value(bucket->value);

While running tests on Android, I found that the snipped above should
be better done in uclamp_rq_inc_id() for two main reasons:
 1. because of the early return in this function, we skip the reset in
    case the task is not the last one running on a CPU, thus
    triggering the above SCHED_WARN_ON
 2. since a non active bucket is already ignored, we don't care about
    resetting its "local max value"

Will move the "max local update" in uclamp_rq_inc_id() where something
like:


    /*
     * Local max aggregation: rq buckets always track the max
     * "requested" clamp value of its RUNNABLE tasks.
     */
    if (bucket->tasks == 1 || uc_se->value > bucket->value)
    	bucket->value = uc_se->value;


Should grant to always have an updated local max every time a bucket
is active.


>  		WRITE_ONCE(uc_rq->value, uclamp_rq_max_value(rq, clamp_id));
> +	}
>  }
>  
>  static inline void uclamp_rq_inc(struct rq *rq, struct task_struct *p)
> @@ -855,25 +879,9 @@ static void __init init_uclamp(void)
>  	unsigned int clamp_id;
>  	int cpu;
>  
> -	for_each_possible_cpu(cpu) {
> -		struct uclamp_bucket *bucket;
> -		struct uclamp_rq *uc_rq;
> -		unsigned int bucket_id;
> -
> +	for_each_possible_cpu(cpu)
>  		memset(&cpu_rq(cpu)->uclamp, 0, sizeof(struct uclamp_rq));
>  
> -		for (clamp_id = 0; clamp_id < UCLAMP_CNT; ++clamp_id) {
> -			uc_rq = &cpu_rq(cpu)->uclamp[clamp_id];
> -
> -			bucket_id = 1;
> -			while (bucket_id < UCLAMP_BUCKETS) {
> -				bucket = &uc_rq->bucket[bucket_id];
> -				bucket->value = bucket_id * UCLAMP_BUCKET_DELTA;
> -				++bucket_id;
> -			}
> -		}
> -	}
> -
>  	for (clamp_id = 0; clamp_id < UCLAMP_CNT; ++clamp_id) {
>  		struct uclamp_se *uc_se = &init_task.uclamp[clamp_id];
>  
> -- 
> 2.20.1
> 

-- 
#include <best/regards.h>

Patrick Bellasi



[Index of Archives]     [Linux USB Devel]     [Video for Linux]     [Linux Audio Users]     [Yosemite News]     [Linux Kernel]     [Linux SCSI]

  Powered by Linux