git.openpandora.org
/
pandora-kernel.git
/ blobdiff
commit
grep
author
committer
pickaxe
?
search:
re
summary
|
shortlog
|
log
|
commit
|
commitdiff
|
tree
raw
|
inline
| side by side
tracing: Use one prologue for the wakeup tracer function tracers
[pandora-kernel.git]
/
kernel
/
sched_fair.c
diff --git
a/kernel/sched_fair.c
b/kernel/sched_fair.c
index
ab661eb
..
db3f674
100644
(file)
--- a/
kernel/sched_fair.c
+++ b/
kernel/sched_fair.c
@@
-54,13
+54,13
@@
enum sched_tunable_scaling sysctl_sched_tunable_scaling
* Minimal preemption granularity for CPU-bound tasks:
* (default: 2 msec * (1 + ilog(ncpus)), units: nanoseconds)
*/
* Minimal preemption granularity for CPU-bound tasks:
* (default: 2 msec * (1 + ilog(ncpus)), units: nanoseconds)
*/
-unsigned int sysctl_sched_min_granularity =
200
0000ULL;
-unsigned int normalized_sysctl_sched_min_granularity =
200
0000ULL;
+unsigned int sysctl_sched_min_granularity =
75
0000ULL;
+unsigned int normalized_sysctl_sched_min_granularity =
75
0000ULL;
/*
* is kept at sysctl_sched_latency / sysctl_sched_min_granularity
*/
/*
* is kept at sysctl_sched_latency / sysctl_sched_min_granularity
*/
-static unsigned int sched_nr_latency =
3
;
+static unsigned int sched_nr_latency =
8
;
/*
* After fork, child runs first. If set to 0 (default) then
/*
* After fork, child runs first. If set to 0 (default) then
@@
-1313,7
+1313,7
@@
static struct sched_group *
find_idlest_group(struct sched_domain *sd, struct task_struct *p,
int this_cpu, int load_idx)
{
find_idlest_group(struct sched_domain *sd, struct task_struct *p,
int this_cpu, int load_idx)
{
- struct sched_group *idlest = NULL, *
this = NULL, *
group = sd->groups;
+ struct sched_group *idlest = NULL, *group = sd->groups;
unsigned long min_load = ULONG_MAX, this_load = 0;
int imbalance = 100 + (sd->imbalance_pct-100)/2;
unsigned long min_load = ULONG_MAX, this_load = 0;
int imbalance = 100 + (sd->imbalance_pct-100)/2;
@@
-1348,7
+1348,6
@@
find_idlest_group(struct sched_domain *sd, struct task_struct *p,
if (local_group) {
this_load = avg_load;
if (local_group) {
this_load = avg_load;
- this = group;
} else if (avg_load < min_load) {
min_load = avg_load;
idlest = group;
} else if (avg_load < min_load) {
min_load = avg_load;
idlest = group;
@@
-2268,8
+2267,6
@@
unsigned long scale_rt_power(int cpu)
struct rq *rq = cpu_rq(cpu);
u64 total, available;
struct rq *rq = cpu_rq(cpu);
u64 total, available;
- sched_avg_update(rq);
-
total = sched_avg_period() + (rq->clock - rq->age_stamp);
available = total - rq->rt_avg;
total = sched_avg_period() + (rq->clock - rq->age_stamp);
available = total - rq->rt_avg;
@@
-3633,7
+3630,7
@@
static inline int nohz_kick_needed(struct rq *rq, int cpu)
if (time_before(now, nohz.next_balance))
return 0;
if (time_before(now, nohz.next_balance))
return 0;
- if (
!rq->nr_running
)
+ if (
rq->idle_at_tick
)
return 0;
first_pick_cpu = atomic_read(&nohz.first_pick_cpu);
return 0;
first_pick_cpu = atomic_read(&nohz.first_pick_cpu);