aboutsummaryrefslogtreecommitdiffstats
path: root/kernel/sched_features.h
diff options
context:
space:
mode:
authorPeter Zijlstra <a.p.zijlstra@chello.nl>2008-06-27 13:41:12 +0200
committerIngo Molnar <mingo@elte.hu>2008-06-27 14:31:28 +0200
commitc9c294a630e28eec5f2865f028ecfc58d45c0a5a (patch)
tree62fcfc395e98ff7a4ff6bee3364d0bb86024b171 /kernel/sched_features.h
parenta7be37ac8e1565e00880531f4e2aff421a21c803 (diff)
downloadkernel_samsung_smdk4412-c9c294a630e28eec5f2865f028ecfc58d45c0a5a.zip
kernel_samsung_smdk4412-c9c294a630e28eec5f2865f028ecfc58d45c0a5a.tar.gz
kernel_samsung_smdk4412-c9c294a630e28eec5f2865f028ecfc58d45c0a5a.tar.bz2
sched: fix calc_delta_asym()
calc_delta_asym() is supposed to do the same as calc_delta_fair() except linearly shrink the result for negative nice processes - this causes them to have a smaller preemption threshold so that they are more easily preempted. The problem is that for task groups se->load.weight is the per cpu share of the actual task group weight; take that into account. Also provide a debug switch to disable the asymmetry (which I still don't like - but it does greatly benefit some workloads) This would explain the interactivity issues reported against group scheduling. Signed-off-by: Peter Zijlstra <a.p.zijlstra@chello.nl> Cc: Srivatsa Vaddagiri <vatsa@linux.vnet.ibm.com> Cc: Mike Galbraith <efault@gmx.de> Signed-off-by: Ingo Molnar <mingo@elte.hu>
Diffstat (limited to 'kernel/sched_features.h')
-rw-r--r--kernel/sched_features.h1
1 files changed, 1 insertions, 0 deletions
diff --git a/kernel/sched_features.h b/kernel/sched_features.h
index afa5491..04123af 100644
--- a/kernel/sched_features.h
+++ b/kernel/sched_features.h
@@ -7,3 +7,4 @@ SCHED_FEAT(CACHE_HOT_BUDDY, 1)
SCHED_FEAT(SYNC_WAKEUPS, 1)
SCHED_FEAT(HRTICK, 1)
SCHED_FEAT(DOUBLE_TICK, 0)
+SCHED_FEAT(ASYM_GRAN, 1)