Home Home > GIT Browse
summaryrefslogtreecommitdiff
diff options
context:
space:
mode:
authorRusty Russell <rusty@rustcorp.com.au>2003-06-13 21:57:53 -0700
committerLinus Torvalds <torvalds@home.transmeta.com>2003-06-13 21:57:53 -0700
commit035406972c5e9ae212deea49cd2c4d7e369210a8 (patch)
tree7e6fd5b1d339e8b8dbd3c31900f70e4e78a69881
parent07de07defc8c727f287bf20d200b0cfeedfbfa51 (diff)
[PATCH] sched.c neatening and fixes.
1) Fix the comments for the migration_thread. A while back Ingo agreed they were exactly wrong, IIRC. 8). 2) Changed spin_lock_irqsave to spin_lock_irq, since it's in a kernel thread. 3) Don't repeat if the task has moved off the original CPU, just finish. This is because we are simply trying to push the task off this CPU: if it's already moved, great. Currently we might theoretically move a task which is actually running on another CPU, which is v. bad. 4) Replace the __ffs(p->cpus_allowed) with any_online_cpu(), since that's what it's for, and __ffs() can give the wrong answer, eg. if there's no CPU 0. 5) Move the core functionality of migrate_task into a separate function, move_task_away, which I want for the hotplug CPU patch.
-rw-r--r--kernel/sched.c73
1 files changed, 36 insertions, 37 deletions
diff --git a/kernel/sched.c b/kernel/sched.c
index 79c8794c1769..dead0e9dcad5 100644
--- a/kernel/sched.c
+++ b/kernel/sched.c
@@ -2296,7 +2296,7 @@ void set_cpus_allowed(task_t *p, unsigned long new_mask)
* it is sufficient to simply update the task's cpu field.
*/
if (!p->array && !task_running(rq, p)) {
- set_task_cpu(p, __ffs(p->cpus_allowed));
+ set_task_cpu(p, any_online_cpu(p->cpus_allowed));
task_rq_unlock(rq, &flags);
return;
}
@@ -2310,9 +2310,35 @@ void set_cpus_allowed(task_t *p, unsigned long new_mask)
wait_for_completion(&req.done);
}
+/* Move (not current) task off this cpu, onto dest cpu. */
+static void move_task_away(struct task_struct *p, int dest_cpu)
+{
+ runqueue_t *rq_dest;
+ unsigned long flags;
+
+ rq_dest = cpu_rq(dest_cpu);
+
+ local_irq_save(flags);
+ double_rq_lock(this_rq(), rq_dest);
+ if (task_cpu(p) != smp_processor_id())
+ goto out; /* Already moved */
+
+ set_task_cpu(p, dest_cpu);
+ if (p->array) {
+ deactivate_task(p, this_rq());
+ activate_task(p, rq_dest);
+ if (p->prio < rq_dest->curr->prio)
+ resched_task(rq_dest->curr);
+ }
+ out:
+ double_rq_unlock(this_rq(), rq_dest);
+ local_irq_restore(flags);
+}
+
/*
* migration_thread - this is a highprio system thread that performs
- * thread migration by 'pulling' threads into the target runqueue.
+ * thread migration by bumping thread off CPU then 'pushing' onto
+ * another runqueue.
*/
static int migration_thread(void * data)
{
@@ -2326,8 +2352,9 @@ static int migration_thread(void * data)
set_fs(KERNEL_DS);
/*
- * Either we are running on the right CPU, or there's a
- * a migration thread on the target CPU, guaranteed.
+ * Either we are running on the right CPU, or there's a a
+ * migration thread on this CPU, guaranteed (we're started
+ * serially).
*/
set_cpus_allowed(current, 1UL << cpu);
@@ -2337,51 +2364,23 @@ static int migration_thread(void * data)
rq->migration_thread = current;
for (;;) {
- runqueue_t *rq_src, *rq_dest;
struct list_head *head;
- int cpu_src, cpu_dest;
migration_req_t *req;
- unsigned long flags;
- task_t *p;
- spin_lock_irqsave(&rq->lock, flags);
+ spin_lock_irq(&rq->lock);
head = &rq->migration_queue;
current->state = TASK_INTERRUPTIBLE;
if (list_empty(head)) {
- spin_unlock_irqrestore(&rq->lock, flags);
+ spin_unlock_irq(&rq->lock);
schedule();
continue;
}
req = list_entry(head->next, migration_req_t, list);
list_del_init(head->next);
- spin_unlock_irqrestore(&rq->lock, flags);
-
- p = req->task;
- cpu_dest = __ffs(p->cpus_allowed & cpu_online_map);
- rq_dest = cpu_rq(cpu_dest);
-repeat:
- cpu_src = task_cpu(p);
- rq_src = cpu_rq(cpu_src);
-
- local_irq_save(flags);
- double_rq_lock(rq_src, rq_dest);
- if (task_cpu(p) != cpu_src) {
- double_rq_unlock(rq_src, rq_dest);
- local_irq_restore(flags);
- goto repeat;
- }
- if (rq_src == rq) {
- set_task_cpu(p, cpu_dest);
- if (p->array) {
- deactivate_task(p, rq_src);
- __activate_task(p, rq_dest);
- if (p->prio < rq_dest->curr->prio)
- resched_task(rq_dest->curr);
- }
- }
- double_rq_unlock(rq_src, rq_dest);
- local_irq_restore(flags);
+ spin_unlock_irq(&rq->lock);
+ move_task_away(req->task,
+ any_online_cpu(req->task->cpus_allowed));
complete(&req->done);
}
}