kernel/sched/ext_idle.c | 17 ++++++++++++++--- 1 file changed, 14 insertions(+), 3 deletions(-)
The node-aware idle kfunc helpers validate per-node idle tracking, but they
don't check whether built-in idle tracking itself is enabled.
As a result, when ops.update_idle() disables built-in idle tracking, the
node helpers can still read per-node idle masks and attempt idle CPU
selection. This violates the documented behavior and can expose stale
idle state to BPF schedulers.
Fix this by checking check_builtin_idle_enabled() in the node mask getters
and in scx_bpf_pick_idle_cpu_node(), matching the behavior of the non-node
helpers.
scx_bpf_pick_any_cpu_node() is different by: when built-in idle
tracking is disabled, it should skip idle selection and fall back directly
to the any-CPU path. Make it do so and match scx_bpf_pick_any_cpu().
Fixes: 01059219b0cf ("sched_ext: idle: Introduce node-aware idle cpu kfunc helpers")
Cc: stable@vger.kernel.org # v6.15+
Assisted-by: Codex:GPT-5
Signed-off-by: Joseph Salisbury <joseph.salisbury@oracle.com>
---
kernel/sched/ext_idle.c | 17 ++++++++++++++---
1 file changed, 14 insertions(+), 3 deletions(-)
diff --git a/kernel/sched/ext_idle.c b/kernel/sched/ext_idle.c
index ba298ac3ce6c..948f6b4f8ab5 100644
--- a/kernel/sched/ext_idle.c
+++ b/kernel/sched/ext_idle.c
@@ -1082,6 +1082,9 @@ __bpf_kfunc const struct cpumask *scx_bpf_get_idle_cpumask_node(int node)
if (node < 0)
return cpu_none_mask;
+ if (!check_builtin_idle_enabled(sch))
+ return cpu_none_mask;
+
return idle_cpumask(node)->cpu;
}
@@ -1137,6 +1140,9 @@ __bpf_kfunc const struct cpumask *scx_bpf_get_idle_smtmask_node(int node)
if (node < 0)
return cpu_none_mask;
+ if (!check_builtin_idle_enabled(sch))
+ return cpu_none_mask;
+
if (sched_smt_active())
return idle_cpumask(node)->smt;
else
@@ -1253,6 +1259,9 @@ __bpf_kfunc s32 scx_bpf_pick_idle_cpu_node(const struct cpumask *cpus_allowed,
if (node < 0)
return node;
+ if (!check_builtin_idle_enabled(sch))
+ return -EBUSY;
+
return scx_pick_idle_cpu(cpus_allowed, node, flags);
}
@@ -1337,9 +1346,11 @@ __bpf_kfunc s32 scx_bpf_pick_any_cpu_node(const struct cpumask *cpus_allowed,
if (node < 0)
return node;
- cpu = scx_pick_idle_cpu(cpus_allowed, node, flags);
- if (cpu >= 0)
- return cpu;
+ if (static_branch_likely(&scx_builtin_idle_enabled)) {
+ cpu = scx_pick_idle_cpu(cpus_allowed, node, flags);
+ if (cpu >= 0)
+ return cpu;
+ }
if (flags & SCX_PICK_IDLE_IN_NODE)
cpu = cpumask_any_and_distribute(cpumask_of_node(node), cpus_allowed);
--
2.47.3
Hi Joe,
On Tue, Mar 24, 2026 at 03:42:35PM -0400, Joseph Salisbury wrote:
> The node-aware idle kfunc helpers validate per-node idle tracking, but they
> don't check whether built-in idle tracking itself is enabled.
>
> As a result, when ops.update_idle() disables built-in idle tracking, the
> node helpers can still read per-node idle masks and attempt idle CPU
> selection. This violates the documented behavior and can expose stale
> idle state to BPF schedulers.
>
> Fix this by checking check_builtin_idle_enabled() in the node mask getters
> and in scx_bpf_pick_idle_cpu_node(), matching the behavior of the non-node
> helpers.
>
> scx_bpf_pick_any_cpu_node() is different by: when built-in idle
> tracking is disabled, it should skip idle selection and fall back directly
> to the any-CPU path. Make it do so and match scx_bpf_pick_any_cpu().
>
> Fixes: 01059219b0cf ("sched_ext: idle: Introduce node-aware idle cpu kfunc helpers")
> Cc: stable@vger.kernel.org # v6.15+
> Assisted-by: Codex:GPT-5
> Signed-off-by: Joseph Salisbury <joseph.salisbury@oracle.com>
We are already validating this at load time, see validate_ops():
...
/*
* SCX_OPS_BUILTIN_IDLE_PER_NODE requires built-in CPU idle
* selection policy to be enabled.
*/
if ((ops->flags & SCX_OPS_BUILTIN_IDLE_PER_NODE) &&
(ops->update_idle && !(ops->flags & SCX_OPS_KEEP_BUILTIN_IDLE))) {
scx_error(sch, "SCX_OPS_BUILTIN_IDLE_PER_NODE requires CPU idle selection enabled");
return -EINVAL;
}
...
In practice you can't have SCX_OPS_BUILTIN_IDLE_PER_NODE set without
built-in idle enabled if a scheduler is running and we are checking for
SCX_OPS_BUILTIN_IDLE_PER_NODE in validate_node(). So I think these extra
checks are not needed.
Thanks,
-Andrea
> ---
> kernel/sched/ext_idle.c | 17 ++++++++++++++---
> 1 file changed, 14 insertions(+), 3 deletions(-)
>
> diff --git a/kernel/sched/ext_idle.c b/kernel/sched/ext_idle.c
> index ba298ac3ce6c..948f6b4f8ab5 100644
> --- a/kernel/sched/ext_idle.c
> +++ b/kernel/sched/ext_idle.c
> @@ -1082,6 +1082,9 @@ __bpf_kfunc const struct cpumask *scx_bpf_get_idle_cpumask_node(int node)
> if (node < 0)
> return cpu_none_mask;
>
> + if (!check_builtin_idle_enabled(sch))
> + return cpu_none_mask;
> +
> return idle_cpumask(node)->cpu;
> }
>
> @@ -1137,6 +1140,9 @@ __bpf_kfunc const struct cpumask *scx_bpf_get_idle_smtmask_node(int node)
> if (node < 0)
> return cpu_none_mask;
>
> + if (!check_builtin_idle_enabled(sch))
> + return cpu_none_mask;
> +
> if (sched_smt_active())
> return idle_cpumask(node)->smt;
> else
> @@ -1253,6 +1259,9 @@ __bpf_kfunc s32 scx_bpf_pick_idle_cpu_node(const struct cpumask *cpus_allowed,
> if (node < 0)
> return node;
>
> + if (!check_builtin_idle_enabled(sch))
> + return -EBUSY;
> +
> return scx_pick_idle_cpu(cpus_allowed, node, flags);
> }
>
> @@ -1337,9 +1346,11 @@ __bpf_kfunc s32 scx_bpf_pick_any_cpu_node(const struct cpumask *cpus_allowed,
> if (node < 0)
> return node;
>
> - cpu = scx_pick_idle_cpu(cpus_allowed, node, flags);
> - if (cpu >= 0)
> - return cpu;
> + if (static_branch_likely(&scx_builtin_idle_enabled)) {
> + cpu = scx_pick_idle_cpu(cpus_allowed, node, flags);
> + if (cpu >= 0)
> + return cpu;
> + }
>
> if (flags & SCX_PICK_IDLE_IN_NODE)
> cpu = cpumask_any_and_distribute(cpumask_of_node(node), cpus_allowed);
> --
> 2.47.3
>
On 3/25/26 6:22 PM, Andrea Righi wrote:
> Hi Joe,
>
> On Tue, Mar 24, 2026 at 03:42:35PM -0400, Joseph Salisbury wrote:
>> The node-aware idle kfunc helpers validate per-node idle tracking, but they
>> don't check whether built-in idle tracking itself is enabled.
>>
>> As a result, when ops.update_idle() disables built-in idle tracking, the
>> node helpers can still read per-node idle masks and attempt idle CPU
>> selection. This violates the documented behavior and can expose stale
>> idle state to BPF schedulers.
>>
>> Fix this by checking check_builtin_idle_enabled() in the node mask getters
>> and in scx_bpf_pick_idle_cpu_node(), matching the behavior of the non-node
>> helpers.
>>
>> scx_bpf_pick_any_cpu_node() is different by: when built-in idle
>> tracking is disabled, it should skip idle selection and fall back directly
>> to the any-CPU path. Make it do so and match scx_bpf_pick_any_cpu().
>>
>> Fixes: 01059219b0cf ("sched_ext: idle: Introduce node-aware idle cpu kfunc helpers")
>> Cc: stable@vger.kernel.org # v6.15+
>> Assisted-by: Codex:GPT-5
>> Signed-off-by: Joseph Salisbury <joseph.salisbury@oracle.com>
> We are already validating this at load time, see validate_ops():
> ...
> /*
> * SCX_OPS_BUILTIN_IDLE_PER_NODE requires built-in CPU idle
> * selection policy to be enabled.
> */
> if ((ops->flags & SCX_OPS_BUILTIN_IDLE_PER_NODE) &&
> (ops->update_idle && !(ops->flags & SCX_OPS_KEEP_BUILTIN_IDLE))) {
> scx_error(sch, "SCX_OPS_BUILTIN_IDLE_PER_NODE requires CPU idle selection enabled");
> return -EINVAL;
> }
> ...
>
> In practice you can't have SCX_OPS_BUILTIN_IDLE_PER_NODE set without
> built-in idle enabled if a scheduler is running and we are checking for
> SCX_OPS_BUILTIN_IDLE_PER_NODE in validate_node(). So I think these extra
> checks are not needed.
>
> Thanks,
> -Andrea
Hi Andrea,
Thanks for the review. I missed the validate_ops() check and focused on
the helper-side behavior.
SCX_OPS_BUILTIN_IDLE_PER_NODE is rejected when ops.update_idle()
disables built-in idle unless SCX_OPS_KEEP_BUILTIN_IDLE is set, so the
state I was trying to guard is not reachable for a running scheduler.
That makes the added checks unnecessary.
I thought scx_bpf_pick_any_cpu_node() should mirror
scx_bpf_pick_any_cpu() and fall back when built-in idle is disabled.
However, that only makes sense for the non-node helper, where built-in
idle disabled is a valid configuration. For the per-node case,
validate_ops() rejects that combination, so there is no runtime case to
handle.
While looking at this, I noticed the comment above
scx_bpf_pick_any_cpu_node() still describes that unreachable
built-in-idle-disabled fallback case (Line 1364 in ext_idle.c in
mainline). I can create a comment-only cleanup to align the comment
with the current behavior. Do you think that is worth sending? Maybe
something like this:
- * If ops.update_idle() is implemented and %SCX_OPS_KEEP_BUILTIN_IDLE
is not
- * set, this function can't tell which CPUs are idle and will always
pick any
- * CPU.
+ * %SCX_OPS_BUILTIN_IDLE_PER_NODE requires built-in idle tracking, so
+ * this helper always attempts node-aware idle selection before falling
+ * back to picking any CPU.
Thanks for the explanation, and sorry for the noise.
Thanks,
Joe
>
>> ---
>> kernel/sched/ext_idle.c | 17 ++++++++++++++---
>> 1 file changed, 14 insertions(+), 3 deletions(-)
>>
>> diff --git a/kernel/sched/ext_idle.c b/kernel/sched/ext_idle.c
>> index ba298ac3ce6c..948f6b4f8ab5 100644
>> --- a/kernel/sched/ext_idle.c
>> +++ b/kernel/sched/ext_idle.c
>> @@ -1082,6 +1082,9 @@ __bpf_kfunc const struct cpumask *scx_bpf_get_idle_cpumask_node(int node)
>> if (node < 0)
>> return cpu_none_mask;
>>
>> + if (!check_builtin_idle_enabled(sch))
>> + return cpu_none_mask;
>> +
>> return idle_cpumask(node)->cpu;
>> }
>>
>> @@ -1137,6 +1140,9 @@ __bpf_kfunc const struct cpumask *scx_bpf_get_idle_smtmask_node(int node)
>> if (node < 0)
>> return cpu_none_mask;
>>
>> + if (!check_builtin_idle_enabled(sch))
>> + return cpu_none_mask;
>> +
>> if (sched_smt_active())
>> return idle_cpumask(node)->smt;
>> else
>> @@ -1253,6 +1259,9 @@ __bpf_kfunc s32 scx_bpf_pick_idle_cpu_node(const struct cpumask *cpus_allowed,
>> if (node < 0)
>> return node;
>>
>> + if (!check_builtin_idle_enabled(sch))
>> + return -EBUSY;
>> +
>> return scx_pick_idle_cpu(cpus_allowed, node, flags);
>> }
>>
>> @@ -1337,9 +1346,11 @@ __bpf_kfunc s32 scx_bpf_pick_any_cpu_node(const struct cpumask *cpus_allowed,
>> if (node < 0)
>> return node;
>>
>> - cpu = scx_pick_idle_cpu(cpus_allowed, node, flags);
>> - if (cpu >= 0)
>> - return cpu;
>> + if (static_branch_likely(&scx_builtin_idle_enabled)) {
>> + cpu = scx_pick_idle_cpu(cpus_allowed, node, flags);
>> + if (cpu >= 0)
>> + return cpu;
>> + }
>>
>> if (flags & SCX_PICK_IDLE_IN_NODE)
>> cpu = cpumask_any_and_distribute(cpumask_of_node(node), cpus_allowed);
>> --
>> 2.47.3
>>
© 2016 - 2026 Red Hat, Inc.