summaryrefslogtreecommitdiff
path: root/kernel
diff options
context:
space:
mode:
authorSrivatsa Vaddagiri <vatsa@codeaurora.org>2014-07-23 14:27:18 +0530
committerDavid Keitel <dkeitel@codeaurora.org>2016-03-23 20:00:03 -0700
commit1ffae4dc94bded073868b760fafa443a44303f55 (patch)
tree907ea9e3c6144c020938b04e0ddeb80f08b90308 /kernel
parent0b210afc21c31a1bd28b721a70a54aced29eedb0 (diff)
sched: window-stats: Handle policy change properly
sched_window_stat_policy influences task demand and thus various statistics maintained per-cpu like curr_runnable_sum. Changing policy non-atomically would lead to improper accounting. For example, when task is enqueued on a cpu's runqueue, its demand that is added to rq->cumulative_runnable_avg could be based on AVG policy and when its dequeued its demand that is removed can be based on MAX, leading to erroneous accounting. This change causes policy change to be "atomic" i.e all cpu's rq->lock are held and all task's window-stats are reset before policy is changed. Change-Id: I6a3e4fb7bc299dfc5c367693b5717a1ef518c32d CRs-Fixed: 687409 Signed-off-by: Srivatsa Vaddagiri <vatsa@codeaurora.org> [joonwoop@codeaurora.org: fixed minor conflict in include/linux/sched/sysctl.h. Signed-off-by: Joonwoo Park <joonwoop@codeaurora.org>
Diffstat (limited to 'kernel')
-rw-r--r--kernel/sched/core.c21
-rw-r--r--kernel/sched/fair.c34
-rw-r--r--kernel/sched/sched.h2
-rw-r--r--kernel/sysctl.c2
4 files changed, 54 insertions, 5 deletions
diff --git a/kernel/sched/core.c b/kernel/sched/core.c
index 6c8d5076878a..2245a69d8480 100644
--- a/kernel/sched/core.c
+++ b/kernel/sched/core.c
@@ -1123,6 +1123,15 @@ __read_mostly unsigned int sched_ravg_window = 10000000;
__read_mostly unsigned int sysctl_sched_window_stats_policy =
WINDOW_STATS_USE_AVG;
+/*
+ * copy of sysctl_sched_window_stats_policy. Required for atomically
+ * changing policy (see sched_window_stats_policy_update_handler() for details).
+ *
+ * Initialize both to same value!!
+ */
+static __read_mostly unsigned int sched_window_stats_policy =
+ WINDOW_STATS_USE_AVG;
+
/* 1 -> use PELT based load stats, 0 -> use window-based load stats */
unsigned int __read_mostly sched_use_pelt;
@@ -1243,9 +1252,9 @@ update_history(struct rq *rq, struct task_struct *p, u32 runtime, int samples,
compute_demand:
avg = div64_u64(sum, RAVG_HIST_SIZE);
- if (sysctl_sched_window_stats_policy == WINDOW_STATS_USE_RECENT)
+ if (sched_window_stats_policy == WINDOW_STATS_USE_RECENT)
demand = runtime;
- else if (sysctl_sched_window_stats_policy == WINDOW_STATS_USE_MAX)
+ else if (sched_window_stats_policy == WINDOW_STATS_USE_MAX)
demand = max;
else
demand = max(avg, runtime);
@@ -1515,7 +1524,8 @@ unsigned long sched_get_busy(int cpu)
}
/* Called with IRQs disabled */
-void reset_all_window_stats(u64 window_start, unsigned int window_size)
+void reset_all_window_stats(u64 window_start, unsigned int window_size,
+ int policy)
{
int cpu;
u64 wallclock;
@@ -1557,6 +1567,9 @@ void reset_all_window_stats(u64 window_start, unsigned int window_size)
fixup_nr_big_small_task(cpu);
}
+ if (policy >= 0)
+ sched_window_stats_policy = policy;
+
for_each_online_cpu(cpu) {
struct rq *rq = cpu_rq(cpu);
raw_spin_unlock(&rq->lock);
@@ -1589,7 +1602,7 @@ int sched_set_window(u64 window_start, unsigned int window_size)
BUG_ON(sched_clock() < ws);
- reset_all_window_stats(ws, window_size);
+ reset_all_window_stats(ws, window_size, -1);
local_irq_restore(flags);
diff --git a/kernel/sched/fair.c b/kernel/sched/fair.c
index 3d1ad0f45b28..c4e57cf18443 100644
--- a/kernel/sched/fair.c
+++ b/kernel/sched/fair.c
@@ -3144,6 +3144,40 @@ void post_big_small_task_count_change(void)
local_irq_enable();
}
+static DEFINE_MUTEX(policy_mutex);
+
+int sched_window_stats_policy_update_handler(struct ctl_table *table, int write,
+ void __user *buffer, size_t *lenp,
+ loff_t *ppos)
+{
+ int ret;
+ unsigned int *data = (unsigned int *)table->data;
+ unsigned int old_val;
+ unsigned long flags;
+
+ if (!sched_enable_hmp)
+ return -EINVAL;
+
+ mutex_lock(&policy_mutex);
+
+ old_val = *data;
+
+ ret = proc_dointvec_minmax(table, write, buffer, lenp, ppos);
+ if (ret || !write || (write && old_val == *data))
+ goto done;
+
+ local_irq_save(flags);
+
+ reset_all_window_stats(0, 0, sysctl_sched_window_stats_policy);
+
+ local_irq_restore(flags);
+
+done:
+ mutex_unlock(&policy_mutex);
+
+ return ret;
+}
+
/*
* Convert percentage value into absolute form. This will avoid div() operation
* in fast path, to convert task load in percentage scale.
diff --git a/kernel/sched/sched.h b/kernel/sched/sched.h
index e2a22f661559..1083fefea541 100644
--- a/kernel/sched/sched.h
+++ b/kernel/sched/sched.h
@@ -1027,6 +1027,8 @@ extern void inc_nr_big_small_task(struct rq *rq, struct task_struct *p);
extern void dec_nr_big_small_task(struct rq *rq, struct task_struct *p);
extern void set_hmp_defaults(void);
extern unsigned int power_cost_at_freq(int cpu, unsigned int freq);
+extern void reset_all_window_stats(u64 window_start, unsigned int window_size,
+ int policy);
#else /* CONFIG_SCHED_HMP */
diff --git a/kernel/sysctl.c b/kernel/sysctl.c
index abe1ea74f977..2c464882e2da 100644
--- a/kernel/sysctl.c
+++ b/kernel/sysctl.c
@@ -314,7 +314,7 @@ static struct ctl_table kern_table[] = {
.data = &sysctl_sched_window_stats_policy,
.maxlen = sizeof(unsigned int),
.mode = 0644,
- .proc_handler = proc_dointvec,
+ .proc_handler = sched_window_stats_policy_update_handler,
},
{
.procname = "sched_wakeup_load_threshold",