Re: [PATCH 3/6] sched/fair.c: disambiguate existing/remaining "capacity" usage

From: Vincent Guittot
Date: Thu May 15 2014 - 03:39:33 EST


On 14 May 2014 22:57, Nicolas Pitre <nicolas.pitre@xxxxxxxxxx> wrote:
> We have "power" (which should actually become "capacity") and "capacity"
> which is a scaled down "capacity factor" in terms of possible tasks.
> Let's use "capa_factor" to make room for proper usage of "capacity"
> later.
>
> Signed-off-by: Nicolas Pitre <nico@xxxxxxxxxx>
> ---
> kernel/sched/fair.c | 42 +++++++++++++++++++++---------------------
> 1 file changed, 21 insertions(+), 21 deletions(-)
>
> diff --git a/kernel/sched/fair.c b/kernel/sched/fair.c
> index 0eda4c527e..2633c42692 100644
> --- a/kernel/sched/fair.c
> +++ b/kernel/sched/fair.c
> @@ -5487,7 +5487,7 @@ struct sg_lb_stats {
> unsigned long load_per_task;
> unsigned long group_power;
> unsigned int sum_nr_running; /* Nr tasks running in the group */
> - unsigned int group_capacity;
> + unsigned int group_capa_factor;

As it is mainly compared to sum_nr_running, you might rename it to
group_nr_capacity instead of group_capa_factor

> unsigned int idle_cpus;
> unsigned int group_weight;
> int group_imb; /* Is there an imbalance in the group ? */
> @@ -5782,15 +5782,15 @@ static inline int sg_imbalanced(struct sched_group *group)
> }
>
> /*
> - * Compute the group capacity.
> + * Compute the group capacity factor.
> *
> * Avoid the issue where N*frac(smt_power) >= 1 creates 'phantom' cores by
> * first dividing out the smt factor and computing the actual number of cores
> * and limit power unit capacity with that.
> */
> -static inline int sg_capacity(struct lb_env *env, struct sched_group *group)
> +static inline int sg_capa_factor(struct lb_env *env, struct sched_group *group)
> {
> - unsigned int capacity, smt, cpus;
> + unsigned int capa_factor, smt, cpus;
> unsigned int power, power_orig;
>
> power = group->sgp->power;
> @@ -5799,13 +5799,13 @@ static inline int sg_capacity(struct lb_env *env, struct sched_group *group)
>
> /* smt := ceil(cpus / power), assumes: 1 < smt_power < 2 */
> smt = DIV_ROUND_UP(SCHED_POWER_SCALE * cpus, power_orig);
> - capacity = cpus / smt; /* cores */
> + capa_factor = cpus / smt; /* cores */
>
> - capacity = min_t(unsigned, capacity, DIV_ROUND_CLOSEST(power, SCHED_POWER_SCALE));
> - if (!capacity)
> - capacity = fix_small_capacity(env->sd, group);
> + capa_factor = min_t(unsigned, capa_factor, DIV_ROUND_CLOSEST(power, SCHED_POWER_SCALE));
> + if (!capa_factor)
> + capa_factor = fix_small_capacity(env->sd, group);
>
> - return capacity;
> + return capa_factor;
> }
>
> /**
> @@ -5855,9 +5855,9 @@ static inline void update_sg_lb_stats(struct lb_env *env,
> sgs->group_weight = group->group_weight;
>
> sgs->group_imb = sg_imbalanced(group);
> - sgs->group_capacity = sg_capacity(env, group);
> + sgs->group_capa_factor = sg_capa_factor(env, group);
>
> - if (sgs->group_capacity > sgs->sum_nr_running)
> + if (sgs->group_capa_factor > sgs->sum_nr_running)
> sgs->group_has_free_capacity = 1;
> }
>
> @@ -5882,7 +5882,7 @@ static bool update_sd_pick_busiest(struct lb_env *env,
> if (sgs->avg_load <= sds->busiest_stat.avg_load)
> return false;
>
> - if (sgs->sum_nr_running > sgs->group_capacity)
> + if (sgs->sum_nr_running > sgs->group_capa_factor)
> return true;
>
> if (sgs->group_imb)
> @@ -5973,17 +5973,17 @@ static inline void update_sd_lb_stats(struct lb_env *env, struct sd_lb_stats *sd
>
> /*
> * In case the child domain prefers tasks go to siblings
> - * first, lower the sg capacity to one so that we'll try
> + * first, lower the sg capacity factor to one so that we'll try
> * and move all the excess tasks away. We lower the capacity
> * of a group only if the local group has the capacity to fit
> - * these excess tasks, i.e. nr_running < group_capacity. The
> + * these excess tasks, i.e. nr_running < group_capa_factor. The
> * extra check prevents the case where you always pull from the
> * heaviest group when it is already under-utilized (possible
> * with a large weight task outweighs the tasks on the system).
> */
> if (prefer_sibling && sds->local &&
> sds->local_stat.group_has_free_capacity)
> - sgs->group_capacity = min(sgs->group_capacity, 1U);
> + sgs->group_capa_factor = min(sgs->group_capa_factor, 1U);
>
> if (update_sd_pick_busiest(env, sds, sg, sgs)) {
> sds->busiest = sg;
> @@ -6157,7 +6157,7 @@ static inline void calculate_imbalance(struct lb_env *env, struct sd_lb_stats *s
> * have to drop below capacity to reach cpu-load equilibrium.
> */
> load_above_capacity =
> - (busiest->sum_nr_running - busiest->group_capacity);
> + (busiest->sum_nr_running - busiest->group_capa_factor);
>
> load_above_capacity *= (SCHED_LOAD_SCALE * SCHED_POWER_SCALE);
> load_above_capacity /= busiest->group_power;
> @@ -6301,7 +6301,7 @@ static struct rq *find_busiest_queue(struct lb_env *env,
> int i;
>
> for_each_cpu_and(i, sched_group_cpus(group), env->cpus) {
> - unsigned long power, capacity, wl;
> + unsigned long power, capa_factor, wl;
> enum fbq_type rt;
>
> rq = cpu_rq(i);
> @@ -6330,9 +6330,9 @@ static struct rq *find_busiest_queue(struct lb_env *env,
> continue;
>
> power = power_of(i);
> - capacity = DIV_ROUND_CLOSEST(power, SCHED_POWER_SCALE);
> - if (!capacity)
> - capacity = fix_small_capacity(env->sd, group);
> + capa_factor = DIV_ROUND_CLOSEST(power, SCHED_POWER_SCALE);
> + if (!capa_factor)
> + capa_factor = fix_small_capacity(env->sd, group);
>
> wl = weighted_cpuload(i);
>
> @@ -6340,7 +6340,7 @@ static struct rq *find_busiest_queue(struct lb_env *env,
> * When comparing with imbalance, use weighted_cpuload()
> * which is not scaled with the cpu power.
> */
> - if (capacity && rq->nr_running == 1 && wl > env->imbalance)
> + if (capa_factor && rq->nr_running == 1 && wl > env->imbalance)
> continue;
>
> /*
> --
> 1.8.4.108.g55ea5f6
>
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/