[PATCH 1/2] sched: Rework migrate_tasks()

From: Kirill Tkhai
Date: Wed Jun 11 2014 - 05:52:21 EST



Currently migrate_tasks() skips throttled tasks,
because they are not pickable by pick_next_task().

These tasks stay on dead cpu even after they
becomes unthrottled. They are not schedulable
till user manually changes their affinity or till
cpu becomes alive again.

But for user this looks completely untransparent:
task hangs, but it's not obvious what he has to do,
because kernel does not report any problem.

This situation may easily be triggered intentionally.
Playing with extremely small cpu.cfs_quota_us causes
it almost in 100% cases. In usual life it's very rare,
but still possible for some unhappy user.

This patch changes migrate_tasks() to iterate throw
all of the threads in the system under tasklist_lock
is locked. This may a little worsen the performance,
but:

1)it guarantees all of tasks will migrate;

2)migrate_tasks() is not performance critical,
it just must do what it has to do.

Signed-off-by: Kirill Tkhai <ktkhai@xxxxxxxxxxxxx>
CC: Peter Zijlstra <peterz@xxxxxxxxxxxxx>
CC: Ingo Molnar <mingo@xxxxxxxxxx>
---
kernel/sched/core.c | 75 ++++++++++++++-------------------------------------
1 file changed, 20 insertions(+), 55 deletions(-)

diff --git a/kernel/sched/core.c b/kernel/sched/core.c
index d0d1565..da7f4c5 100644
--- a/kernel/sched/core.c
+++ b/kernel/sched/core.c
@@ -4795,78 +4795,43 @@ static void calc_load_migrate(struct rq *rq)
atomic_long_add(delta, &calc_load_tasks);
}

-static void put_prev_task_fake(struct rq *rq, struct task_struct *prev)
-{
-}
-
-static const struct sched_class fake_sched_class = {
- .put_prev_task = put_prev_task_fake,
-};
-
-static struct task_struct fake_task = {
- /*
- * Avoid pull_{rt,dl}_task()
- */
- .prio = MAX_PRIO + 1,
- .sched_class = &fake_sched_class,
-};
-
/*
* Migrate all tasks from the rq, sleeping tasks will be migrated by
* try_to_wake_up()->select_task_rq().
*
- * Called with rq->lock held even though we'er in stop_machine() and
- * there's no concurrency possible, we hold the required locks anyway
- * because of lock validation efforts.
+ * Called from stop_machine(), and there's no concurrency possible, but
+ * we hold the required locks anyway because of lock validation efforts.
*/
static void migrate_tasks(unsigned int dead_cpu)
{
struct rq *rq = cpu_rq(dead_cpu);
- struct task_struct *next, *stop = rq->stop;
+ struct task_struct *g, *p;
int dest_cpu;

- /*
- * Fudge the rq selection such that the below task selection loop
- * doesn't get stuck on the currently eligible stop task.
- *
- * We're currently inside stop_machine() and the rq is either stuck
- * in the stop_machine_cpu_stop() loop, or we're executing this code,
- * either way we should never end up calling schedule() until we're
- * done here.
- */
- rq->stop = NULL;
-
- /*
- * put_prev_task() and pick_next_task() sched
- * class method both need to have an up-to-date
- * value of rq->clock[_task]
- */
- update_rq_clock(rq);
-
- for ( ; ; ) {
+ read_lock(&tasklist_lock);
+ do_each_thread(g, p) {
/*
- * There's this thread running, bail when that's the only
- * remaining thread.
+ * We're inside stop_machine() and nobody can manage tasks
+ * in parallel. So, this unlocked check is correct.
*/
- if (rq->nr_running == 1)
- break;
+ if (!p->on_rq || task_cpu(p) != dead_cpu)
+ continue;

- next = pick_next_task(rq, &fake_task);
- BUG_ON(!next);
- next->sched_class->put_prev_task(rq, next);
+ if (p == rq->stop)
+ continue;

/* Find suitable destination for @next, with force if needed. */
- dest_cpu = select_fallback_rq(dead_cpu, next);
+ raw_spin_lock(&rq->lock);
+ dest_cpu = select_fallback_rq(dead_cpu, p);
raw_spin_unlock(&rq->lock);

- __migrate_task(next, dead_cpu, dest_cpu);
+ WARN_ON(!__migrate_task(p, dead_cpu, dest_cpu));

- raw_spin_lock(&rq->lock);
- }
+ } while_each_thread(g, p);
+ read_unlock(&tasklist_lock);

- rq->stop = stop;
+ BUG_ON(rq->nr_running != 1); /* the migration thread */
}
-
#endif /* CONFIG_HOTPLUG_CPU */

#if defined(CONFIG_SCHED_DEBUG) && defined(CONFIG_SYSCTL)
@@ -5107,16 +5072,16 @@ migration_call(struct notifier_block *nfb, unsigned long action, void *hcpu)

#ifdef CONFIG_HOTPLUG_CPU
case CPU_DYING:
+ BUG_ON(current != rq->stop);
sched_ttwu_pending();
/* Update our root-domain */
- raw_spin_lock_irqsave(&rq->lock, flags);
+ raw_spin_lock(&rq->lock);
if (rq->rd) {
BUG_ON(!cpumask_test_cpu(cpu, rq->rd->span));
set_rq_offline(rq);
}
+ raw_spin_unlock(&rq->lock);
migrate_tasks(cpu);
- BUG_ON(rq->nr_running != 1); /* the migration thread */
- raw_spin_unlock_irqrestore(&rq->lock, flags);
break;

case CPU_DEAD:



--
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@xxxxxxxxxxxxxxx
More majordomo info at http://vger.kernel.org/majordomo-info.html
Please read the FAQ at http://www.tux.org/lkml/