[tip:sched/core] sched: Clean up the load_idx selection in select_task_rq_fair

From: tip-bot for Peter Zijlstra
Date: Wed Sep 16 2009 - 11:10:38 EST


Commit-ID: 5158f4e4428c6b8d52796b3b460e95796123a114
Gitweb: http://git.kernel.org/tip/5158f4e4428c6b8d52796b3b460e95796123a114
Author: Peter Zijlstra <a.p.zijlstra@xxxxxxxxx>
AuthorDate: Wed, 16 Sep 2009 13:46:59 +0200
Committer: Ingo Molnar <mingo@xxxxxxx>
CommitDate: Wed, 16 Sep 2009 16:44:32 +0200

sched: Clean up the load_idx selection in select_task_rq_fair

Clean up the code a little.

Signed-off-by: Peter Zijlstra <a.p.zijlstra@xxxxxxxxx>
LKML-Reference: <new-submission>
Signed-off-by: Ingo Molnar <mingo@xxxxxxx>


---
kernel/sched_fair.c | 27 ++++++++-------------------
1 files changed, 8 insertions(+), 19 deletions(-)

diff --git a/kernel/sched_fair.c b/kernel/sched_fair.c
index 722d392..aeff40e 100644
--- a/kernel/sched_fair.c
+++ b/kernel/sched_fair.c
@@ -1248,26 +1248,11 @@ static int wake_affine(struct sched_domain *sd, struct task_struct *p, int sync)
*/
static struct sched_group *
find_idlest_group(struct sched_domain *sd, struct task_struct *p,
- int this_cpu, int flag)
+ int this_cpu, int load_idx)
{
struct sched_group *idlest = NULL, *this = NULL, *group = sd->groups;
unsigned long min_load = ULONG_MAX, this_load = 0;
int imbalance = 100 + (sd->imbalance_pct-100)/2;
- int load_idx = 0;
-
- switch (flag) {
- case SD_BALANCE_FORK:
- case SD_BALANCE_EXEC:
- load_idx = sd->forkexec_idx;
- break;
-
- case SD_BALANCE_WAKE:
- load_idx = sd->wake_idx;
- break;
-
- default:
- break;
- }

do {
unsigned long load, avg_load;
@@ -1346,14 +1331,14 @@ find_idlest_cpu(struct sched_group *group, struct task_struct *p, int this_cpu)
*
* preempt must be disabled.
*/
-static int select_task_rq_fair(struct task_struct *p, int sd_flag, int flags)
+static int select_task_rq_fair(struct task_struct *p, int sd_flag, int wake_flags)
{
struct sched_domain *tmp, *shares = NULL, *sd = NULL;
int cpu = smp_processor_id();
int prev_cpu = task_cpu(p);
int new_cpu = cpu;
int want_affine = 0;
- int sync = flags & WF_SYNC;
+ int sync = wake_flags & WF_SYNC;

if (sd_flag & SD_BALANCE_WAKE) {
if (sched_feat(AFFINE_WAKEUPS))
@@ -1413,6 +1398,7 @@ static int select_task_rq_fair(struct task_struct *p, int sd_flag, int flags)
update_shares(sd);

while (sd) {
+ int load_idx = sd->forkexec_idx;
struct sched_group *group;
int weight;

@@ -1421,7 +1407,10 @@ static int select_task_rq_fair(struct task_struct *p, int sd_flag, int flags)
continue;
}

- group = find_idlest_group(sd, p, cpu, sd_flag);
+ if (sd_flag & SD_BALANCE_WAKE)
+ load_idx = sd->wake_idx;
+
+ group = find_idlest_group(sd, p, cpu, load_idx);
if (!group) {
sd = sd->child;
continue;
--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/