From: Neeraj Upadhyay <neeraj.upadhyay@kernel.org>
Use RCU watching state of a CPU to check whether RCU-tasks GP
need to wait for idle injection task on that CPU. Idle injection
tasks which are in deep-idle states where RCU is not watching or
which have transitioned to/from deep-idle state do not block
RCU-tasks grace period.
Signed-off-by: Neeraj Upadhyay <neeraj.upadhyay@kernel.org>
---
kernel/rcu/tasks.h | 63 +++++++++++++++++++++++++++++++++++-----------
1 file changed, 48 insertions(+), 15 deletions(-)
diff --git a/kernel/rcu/tasks.h b/kernel/rcu/tasks.h
index d8506d2e6f54..1947f9b6346d 100644
--- a/kernel/rcu/tasks.h
+++ b/kernel/rcu/tasks.h
@@ -38,6 +38,8 @@ typedef void (*postgp_func_t)(struct rcu_tasks *rtp);
* @rtpp: Pointer to the rcu_tasks structure.
* @rcu_watching_snap: Per-GP RCU-watching snapshot for idle tasks.
* @rcu_watching_snap_rec: RCU-watching snapshot recorded for idle task.
+ * @rcu_watching_idle_inj_snap: Per-GP RCU-watching snapshot for idle inject task.
+ * @rcu_watching_idle_inj_rec: RCU-watching snapshot recorded for idle inject task.
*/
struct rcu_tasks_percpu {
struct rcu_segcblist cblist;
@@ -56,6 +58,8 @@ struct rcu_tasks_percpu {
struct rcu_tasks *rtpp;
int rcu_watching_snap;
bool rcu_watching_snap_rec;
+ int rcu_watching_idle_inj_snap;
+ bool rcu_watching_idle_inj_rec;
};
/**
@@ -989,10 +993,34 @@ static void rcu_tasks_pregp_step(struct list_head *hop)
struct rcu_tasks_percpu *rtpcp = per_cpu_ptr(rcu_tasks.rtpcpu, cpu);
rtpcp->rcu_watching_snap_rec = false;
+ rtpcp->rcu_watching_idle_inj_rec = false;
}
}
#ifdef CONFIG_SMP
+static bool rcu_idle_check_rcu_watching(int *rcu_watching_snap, bool *rcu_watching_rec, int cpu)
+{
+ if (!*rcu_watching_rec) {
+ /*
+ * Do plain access. Ordering between remote CPU's pre idle accesses
+ * and post rcu-tasks grace period is provided by synchronize_rcu()
+ * in rcu_tasks_postgp().
+ */
+ *rcu_watching_snap = ct_rcu_watching_cpu(cpu);
+ *rcu_watching_rec = true;
+ if (rcu_watching_snap_in_eqs(*rcu_watching_snap))
+ /*
+ * RCU-idle contexts are RCU-tasks quiescent state for idle
+ * (and idle injection) tasks.
+ */
+ return false;
+ } else if (rcu_watching_snap_stopped_since(cpu, *rcu_watching_snap)) {
+ return false;
+ }
+
+ return true;
+}
+
static bool rcu_idle_task_is_holdout(struct task_struct *t, int cpu)
{
struct rcu_tasks_percpu *rtpcp = per_cpu_ptr(rcu_tasks.rtpcpu, cpu);
@@ -1010,22 +1038,16 @@ static bool rcu_idle_task_is_holdout(struct task_struct *t, int cpu)
if (!t->on_cpu)
return false;
- if (!rtpcp->rcu_watching_snap_rec) {
- /*
- * Do plain access. Ordering between remote CPU's pre idle accesses
- * and post rcu-tasks grace period is provided by synchronize_rcu()
- * in rcu_tasks_postgp().
- */
- rtpcp->rcu_watching_snap = ct_rcu_watching_cpu(cpu);
- rtpcp->rcu_watching_snap_rec = true;
- /* RCU-idle contexts are RCU-tasks quiescent state for idle tasks. */
- if (rcu_watching_snap_in_eqs(rtpcp->rcu_watching_snap))
- return false;
- } else if (rcu_watching_snap_stopped_since(cpu, rtpcp->rcu_watching_snap)) {
- return false;
- }
+ return rcu_idle_check_rcu_watching(&rtpcp->rcu_watching_snap,
+ &rtpcp->rcu_watching_snap_rec, cpu);
+}
- return true;
+static bool rcu_idle_inj_is_holdout(struct task_struct *t, int cpu)
+{
+ struct rcu_tasks_percpu *rtpcp = per_cpu_ptr(rcu_tasks.rtpcpu, cpu);
+
+ return rcu_idle_check_rcu_watching(&rtpcp->rcu_watching_idle_inj_snap,
+ &rtpcp->rcu_watching_idle_inj_rec, cpu);
}
#else /* #ifdef CONFIG_SMP */
static inline bool rcu_idle_task_is_holdout(struct task_struct *t, int cpu)
@@ -1037,6 +1059,15 @@ static inline bool rcu_idle_task_is_holdout(struct task_struct *t, int cpu)
*/
return false;
}
+
+static inline bool rcu_idle_inj_is_holdout(struct task_struct *t, int cpu)
+{
+ /*
+ * Idle injection tasks are PF_IDLE within preempt disabled
+ * region. So, we should not enter this call for !SMP.
+ */
+ return false;
+}
#endif
/* Check for quiescent states since the pregp's synchronize_rcu() */
@@ -1060,6 +1091,8 @@ static bool rcu_tasks_is_holdout(struct task_struct *t)
if (t == idle_task(cpu))
return rcu_idle_task_is_holdout(t, cpu);
+ else if (is_idle_task(t))
+ return rcu_idle_inj_is_holdout(t, cpu);
return true;
}
--
2.40.1