summary refs log tree commit diff
diff options
context:
space:
mode:
authorPeter Zijlstra <a.p.zijlstra@chello.nl>2008-11-04 21:25:10 +0100
committerIngo Molnar <mingo@elte.hu>2008-11-05 10:30:15 +0100
commit02479099c286894644f8e96c6bbb535ab64662fd (patch)
treedf25bc51ddb31a44c91b672311b532de626fadcb
parent4793241be408b3926ee00c704d7da3b3faf3a05f (diff)
downloadlinux-02479099c286894644f8e96c6bbb535ab64662fd.tar.gz
sched: fix buddies for group scheduling
Impact: scheduling order fix for group scheduling

For each level in the hierarchy, set the buddy to point to the right entity.
Therefore, when we do the hierarchical schedule, we have a fair chance of
ending up where we meant to.

Signed-off-by: Peter Zijlstra <a.p.zijlstra@chello.nl>
Acked-by: Mike Galbraith <efault@gmx.de>
Signed-off-by: Ingo Molnar <mingo@elte.hu>
-rw-r--r--kernel/sched_fair.c16
1 files changed, 14 insertions, 2 deletions
diff --git a/kernel/sched_fair.c b/kernel/sched_fair.c
index a6b1db8a0bd8..51aa3e102acb 100644
--- a/kernel/sched_fair.c
+++ b/kernel/sched_fair.c
@@ -1319,6 +1319,18 @@ wakeup_preempt_entity(struct sched_entity *curr, struct sched_entity *se)
 	return 0;
 }
 
+static void set_last_buddy(struct sched_entity *se)
+{
+	for_each_sched_entity(se)
+		cfs_rq_of(se)->last = se;
+}
+
+static void set_next_buddy(struct sched_entity *se)
+{
+	for_each_sched_entity(se)
+		cfs_rq_of(se)->next = se;
+}
+
 /*
  * Preempt the current task with a newly woken task if needed:
  */
@@ -1352,8 +1364,8 @@ static void check_preempt_wakeup(struct rq *rq, struct task_struct *p, int sync)
 	 * obvious reasons its a bad idea to schedule back to the idle thread.
 	 */
 	if (sched_feat(LAST_BUDDY) && likely(se->on_rq && curr != rq->idle))
-		cfs_rq_of(se)->last = se;
-	cfs_rq_of(pse)->next = pse;
+		set_last_buddy(se);
+	set_next_buddy(pse);
 
 	/*
 	 * We can come here with TIF_NEED_RESCHED already set from new task