diff options
author | Peter Zijlstra <peterz@infradead.org> | 2015-09-07 15:05:42 +0200 |
---|---|---|
committer | Ingo Molnar <mingo@kernel.org> | 2015-09-13 09:52:59 +0200 |
commit | 54a21385facbdcd89a78e8c3e5025f04c5f2b59c (patch) | |
tree | c730d1646928df978e88485fc7ed7fd208b5c120 /kernel/sched | |
parent | 98d8fd8126676f7ba6e133e65b2ca4b17989d32c (diff) | |
download | linux-54a21385facbdcd89a78e8c3e5025f04c5f2b59c.tar.bz2 |
sched/fair: Rename scale() to cap_scale()
Rename scale() to cap_scale() to better reflect its purpose, it is
after all not a general purpose scale function, it has
SCHED_CAPACITY_SHIFT hardcoded in it.
Signed-off-by: Peter Zijlstra (Intel) <peterz@infradead.org>
Cc: Linus Torvalds <torvalds@linux-foundation.org>
Cc: Mike Galbraith <efault@gmx.de>
Cc: Peter Zijlstra <peterz@infradead.org>
Cc: Thomas Gleixner <tglx@linutronix.de>
Cc: linux-kernel@vger.kernel.org
Signed-off-by: Ingo Molnar <mingo@kernel.org>
Diffstat (limited to 'kernel/sched')
-rw-r--r-- | kernel/sched/fair.c | 14 |
1 files changed, 7 insertions, 7 deletions
diff --git a/kernel/sched/fair.c b/kernel/sched/fair.c index 047fd1c78a94..7109047731eb 100644 --- a/kernel/sched/fair.c +++ b/kernel/sched/fair.c @@ -2515,7 +2515,7 @@ static u32 __compute_runnable_contrib(u64 n) return contrib + runnable_avg_yN_sum[n]; } -#define scale(v, s) ((v)*(s) >> SCHED_CAPACITY_SHIFT) +#define cap_scale(v, s) ((v)*(s) >> SCHED_CAPACITY_SHIFT) /* * We can represent the historical contribution to runnable average as the @@ -2588,7 +2588,7 @@ __update_load_avg(u64 now, int cpu, struct sched_avg *sa, * period and accrue it. */ delta_w = 1024 - delta_w; - scaled_delta_w = scale(delta_w, scale_freq); + scaled_delta_w = cap_scale(delta_w, scale_freq); if (weight) { sa->load_sum += weight * scaled_delta_w; if (cfs_rq) { @@ -2597,7 +2597,7 @@ __update_load_avg(u64 now, int cpu, struct sched_avg *sa, } } if (running) - sa->util_sum += scale(scaled_delta_w, scale_cpu); + sa->util_sum += cap_scale(scaled_delta_w, scale_cpu); delta -= delta_w; @@ -2614,25 +2614,25 @@ __update_load_avg(u64 now, int cpu, struct sched_avg *sa, /* Efficiently calculate \sum (1..n_period) 1024*y^i */ contrib = __compute_runnable_contrib(periods); - contrib = scale(contrib, scale_freq); + contrib = cap_scale(contrib, scale_freq); if (weight) { sa->load_sum += weight * contrib; if (cfs_rq) cfs_rq->runnable_load_sum += weight * contrib; } if (running) - sa->util_sum += scale(contrib, scale_cpu); + sa->util_sum += cap_scale(contrib, scale_cpu); } /* Remainder of delta accrued against u_0` */ - scaled_delta = scale(delta, scale_freq); + scaled_delta = cap_scale(delta, scale_freq); if (weight) { sa->load_sum += weight * scaled_delta; if (cfs_rq) cfs_rq->runnable_load_sum += weight * scaled_delta; } if (running) - sa->util_sum += scale(scaled_delta, scale_cpu); + sa->util_sum += cap_scale(scaled_delta, scale_cpu); sa->period_contrib += delta; |