sched/numa: Revert "Use effective_load() to balance NUMA loads"
authorPeter Zijlstra <peterz@infradead.org>
Fri, 11 Jul 2014 14:01:53 +0000 (16:01 +0200)
committerIngo Molnar <mingo@kernel.org>
Wed, 16 Jul 2014 11:38:23 +0000 (13:38 +0200)
Due to divergent trees, Rik find that this patch is no longer
required.

Requested-by: Rik van Riel <riel@redhat.com>
Signed-off-by: Peter Zijlstra <peterz@infradead.org>
Cc: Linus Torvalds <torvalds@linux-foundation.org>
Link: http://lkml.kernel.org/n/tip-u6odkgkw8wz3m7orgsjfo5pi@git.kernel.org
Signed-off-by: Ingo Molnar <mingo@kernel.org>
kernel/sched/fair.c

index f5f0cc9..45943b2 100644 (file)
@@ -1151,7 +1151,6 @@ static void task_numa_compare(struct task_numa_env *env,
        struct rq *src_rq = cpu_rq(env->src_cpu);
        struct rq *dst_rq = cpu_rq(env->dst_cpu);
        struct task_struct *cur;
-       struct task_group *tg;
        long src_load, dst_load;
        long load;
        long imp = env->p->numa_group ? groupimp : taskimp;
@@ -1223,14 +1222,9 @@ static void task_numa_compare(struct task_numa_env *env,
         * In the overloaded case, try and keep the load balanced.
         */
 balance:
-       src_load = env->src_stats.load;
-       dst_load = env->dst_stats.load;
-
-       /* Calculate the effect of moving env->p from src to dst. */
-       load = env->p->se.load.weight;
-       tg = task_group(env->p);
-       src_load += effective_load(tg, env->src_cpu, -load, -load);
-       dst_load += effective_load(tg, env->dst_cpu, load, load);
+       load = task_h_load(env->p);
+       dst_load = env->dst_stats.load + load;
+       src_load = env->src_stats.load - load;
 
        if (moveimp > imp && moveimp > env->best_imp) {
                /*
@@ -1250,11 +1244,9 @@ balance:
                goto unlock;
 
        if (cur) {
-               /* Cur moves in the opposite direction. */
-               load = cur->se.load.weight;
-               tg = task_group(cur);
-               src_load += effective_load(tg, env->src_cpu, load, load);
-               dst_load += effective_load(tg, env->dst_cpu, -load, -load);
+               load = task_h_load(cur);
+               dst_load -= load;
+               src_load += load;
        }
 
        if (load_too_imbalanced(src_load, dst_load, env))