On Wed, Jun 02, 2021 at 09:09:54PM +0200, Dietmar Eggemann wrote: > On 27/05/2021 17:38, Beata Michalska wrote: > > [...] > > > +/* > > + * Verify whether there is any CPU capacity asymmetry in a given sched domain. > > + * Provides sd_flags reflecting the asymmetry scope. > > + */ > > +static inline int > > +asym_cpu_capacity_classify(struct sched_domain *sd, > > + const struct cpumask *cpu_map) > > +{ > > + struct asym_cap_data *entry; > > + int sd_asym_flags = 0; > > + int asym_cap_count = 0; > > + int asym_cap_miss = 0; > > + > > + /* > > + * Count how many unique CPU capacities this domain spans across > > + * (compare sched_domain CPUs mask with ones representing available > > + * CPUs capacities). Take into account CPUs that might be offline: > > + * skip those. > > + */ > > + list_for_each_entry(entry, &asym_cap_list, link) { > > + if (cpumask_intersects(sched_domain_span(sd), > > + cpu_capacity_span(entry))) > > + ++asym_cap_count; > > + else if (cpumask_intersects(cpu_capacity_span(entry), cpu_map)) > > nit: `sd span, entry span` but `entry span, cpu_map`. Why not `cpu_map, entry span`? > Cannot recall any reason for that. > > + ++asym_cap_miss; > > + } > > + /* No asymmetry detected */ > > + if (WARN_ON_ONCE(!asym_cap_count) || asym_cap_count == 1) > > + goto leave; > > + > > + sd_asym_flags |= SD_ASYM_CPUCAPACITY; > > + > > + /* > > + * All the available capacities have been found within given sched > > + * domain: no misses reported. > > + */ > > + if (!asym_cap_miss) > > + sd_asym_flags |= SD_ASYM_CPUCAPACITY_FULL; > > + > > +leave: > > + return sd_asym_flags; > > +} > > Everything looks good except that I like this more compact version better, proposed in: > > https://lkml.kernel.org/r/YK9ESqNEo+uacyMD@xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx > > And passing `const struct cpumask *sd_span` instead of `struct > sched_domain *sd` into the function. > I do understand the parameter argument, but honestly don't see much difference in naming and switching single return for asymmetric topologies vs two return statement, but if that is more preferred/readable version I do not mind changing that as well. Thanks for the review. --- BR B. > > diff --git a/kernel/sched/topology.c b/kernel/sched/topology.c > index 77b73abbb9a4..0de8eebded9f 100644 > --- a/kernel/sched/topology.c > +++ b/kernel/sched/topology.c > @@ -1290,13 +1290,11 @@ static LIST_HEAD(asym_cap_list); > * Provides sd_flags reflecting the asymmetry scope. > */ > static inline int > -asym_cpu_capacity_classify(struct sched_domain *sd, > +asym_cpu_capacity_classify(const struct cpumask *sd_span, > const struct cpumask *cpu_map) > { > struct asym_cap_data *entry; > - int sd_asym_flags = 0; > - int asym_cap_count = 0; > - int asym_cap_miss = 0; > + int count = 0, miss = 0; > > /* > * Count how many unique CPU capacities this domain spans across > @@ -1305,27 +1303,20 @@ asym_cpu_capacity_classify(struct sched_domain *sd, > * skip those. > */ > list_for_each_entry(entry, &asym_cap_list, link) { > - if (cpumask_intersects(sched_domain_span(sd), > - cpu_capacity_span(entry))) > - ++asym_cap_count; > - else if (cpumask_intersects(cpu_capacity_span(entry), cpu_map)) > - ++asym_cap_miss; > + if (cpumask_intersects(sd_span, cpu_capacity_span(entry))) > + ++count; > + else if (cpumask_intersects(cpu_map, cpu_capacity_span(entry))) > + ++miss; > } > - /* No asymmetry detected */ > - if (WARN_ON_ONCE(!asym_cap_count) || asym_cap_count == 1) > - goto leave; > > - sd_asym_flags |= SD_ASYM_CPUCAPACITY; > + if (WARN_ON_ONCE(!count) || count == 1) /* No asymmetry */ > + return 0; > > - /* > - * All the available capacities have been found within given sched > - * domain: no misses reported. > - */ > - if (!asym_cap_miss) > - sd_asym_flags |= SD_ASYM_CPUCAPACITY_FULL; > + if (miss) /* Partial asymmetry */ > + return SD_ASYM_CPUCAPACITY; > > -leave: > - return sd_asym_flags; > + /* Full asymmetry */ > + return SD_ASYM_CPUCAPACITY | SD_ASYM_CPUCAPACITY_FULL; > } > > static inline void asym_cpu_capacity_update_data(int cpu) > @@ -1510,6 +1501,7 @@ sd_init(struct sched_domain_topology_level *tl, > struct sd_data *sdd = &tl->data; > struct sched_domain *sd = *per_cpu_ptr(sdd->sd, cpu); > int sd_id, sd_weight, sd_flags = 0; > + struct cpumask *sd_span; > > #ifdef CONFIG_NUMA > /* > @@ -1557,10 +1549,11 @@ sd_init(struct sched_domain_topology_level *tl, > #endif > }; > > - cpumask_and(sched_domain_span(sd), cpu_map, tl->mask(cpu)); > - sd_id = cpumask_first(sched_domain_span(sd)); > + sd_span = sched_domain_span(sd); > + cpumask_and(sd_span, cpu_map, tl->mask(cpu)); > + sd_id = cpumask_first(sd_span); > > - sd->flags |= asym_cpu_capacity_classify(sd, cpu_map); > + sd->flags |= asym_cpu_capacity_classify(sd_span, cpu_map); > > WARN_ONCE((sd->flags & (SD_SHARE_CPUCAPACITY | SD_ASYM_CPUCAPACITY)) == > (SD_SHARE_CPUCAPACITY | SD_ASYM_CPUCAPACITY), > -- > 2.25.1