Rename cpu_corgroup_{map,mask} to cpu_corgrp_{map,mask} to free up the
cpu_corgroup_* namespace. cpu_corgroup_mask() will be added back in the
subsequent commit for CONFIG_SCHED_MC enablement.
No functional changes intended.
Signed-off-by: K Prateek Nayak <kprateek.nayak@amd.com>
---
arch/powerpc/kernel/smp.c | 26 +++++++++++++-------------
1 file changed, 13 insertions(+), 13 deletions(-)
diff --git a/arch/powerpc/kernel/smp.c b/arch/powerpc/kernel/smp.c
index 862f50c09539..4f48262658cc 100644
--- a/arch/powerpc/kernel/smp.c
+++ b/arch/powerpc/kernel/smp.c
@@ -87,7 +87,7 @@ DEFINE_PER_CPU(cpumask_var_t, cpu_sibling_map);
DEFINE_PER_CPU(cpumask_var_t, cpu_smallcore_map);
DEFINE_PER_CPU(cpumask_var_t, cpu_l2_cache_map);
DEFINE_PER_CPU(cpumask_var_t, cpu_core_map);
-static DEFINE_PER_CPU(cpumask_var_t, cpu_coregroup_map);
+static DEFINE_PER_CPU(cpumask_var_t, cpu_corgrp_map);
EXPORT_PER_CPU_SYMBOL(cpu_sibling_map);
EXPORT_PER_CPU_SYMBOL(cpu_l2_cache_map);
@@ -1045,9 +1045,9 @@ static const struct cpumask *tl_smt_mask(struct sched_domain_topology_level *tl,
}
#endif
-static struct cpumask *cpu_coregroup_mask(int cpu)
+static struct cpumask *cpu_corgrp_mask(int cpu)
{
- return per_cpu(cpu_coregroup_map, cpu);
+ return per_cpu(cpu_corgrp_map, cpu);
}
static bool has_coregroup_support(void)
@@ -1061,7 +1061,7 @@ static bool has_coregroup_support(void)
static const struct cpumask *cpu_mc_mask(struct sched_domain_topology_level *tl, int cpu)
{
- return cpu_coregroup_mask(cpu);
+ return cpu_corgrp_mask(cpu);
}
static const struct cpumask *cpu_pkg_mask(struct sched_domain_topology_level *tl, int cpu)
@@ -1124,7 +1124,7 @@ void __init smp_prepare_cpus(unsigned int max_cpus)
zalloc_cpumask_var_node(&per_cpu(cpu_core_map, cpu),
GFP_KERNEL, cpu_to_node(cpu));
if (has_coregroup_support())
- zalloc_cpumask_var_node(&per_cpu(cpu_coregroup_map, cpu),
+ zalloc_cpumask_var_node(&per_cpu(cpu_corgrp_map, cpu),
GFP_KERNEL, cpu_to_node(cpu));
#ifdef CONFIG_NUMA
@@ -1145,7 +1145,7 @@ void __init smp_prepare_cpus(unsigned int max_cpus)
cpumask_set_cpu(boot_cpuid, cpu_core_mask(boot_cpuid));
if (has_coregroup_support())
- cpumask_set_cpu(boot_cpuid, cpu_coregroup_mask(boot_cpuid));
+ cpumask_set_cpu(boot_cpuid, cpu_corgrp_mask(boot_cpuid));
init_big_cores();
if (has_big_cores) {
@@ -1510,8 +1510,8 @@ static void remove_cpu_from_masks(int cpu)
set_cpus_unrelated(cpu, i, cpu_core_mask);
if (has_coregroup_support()) {
- for_each_cpu(i, cpu_coregroup_mask(cpu))
- set_cpus_unrelated(cpu, i, cpu_coregroup_mask);
+ for_each_cpu(i, cpu_corgrp_mask(cpu))
+ set_cpus_unrelated(cpu, i, cpu_corgrp_mask);
}
}
#endif
@@ -1543,7 +1543,7 @@ static void update_coregroup_mask(int cpu, cpumask_var_t *mask)
if (!*mask) {
/* Assume only siblings are part of this CPU's coregroup */
for_each_cpu(i, submask_fn(cpu))
- set_cpus_related(cpu, i, cpu_coregroup_mask);
+ set_cpus_related(cpu, i, cpu_corgrp_mask);
return;
}
@@ -1551,18 +1551,18 @@ static void update_coregroup_mask(int cpu, cpumask_var_t *mask)
cpumask_and(*mask, cpu_online_mask, cpu_node_mask(cpu));
/* Update coregroup mask with all the CPUs that are part of submask */
- or_cpumasks_related(cpu, cpu, submask_fn, cpu_coregroup_mask);
+ or_cpumasks_related(cpu, cpu, submask_fn, cpu_corgrp_mask);
/* Skip all CPUs already part of coregroup mask */
- cpumask_andnot(*mask, *mask, cpu_coregroup_mask(cpu));
+ cpumask_andnot(*mask, *mask, cpu_corgrp_mask(cpu));
for_each_cpu(i, *mask) {
/* Skip all CPUs not part of this coregroup */
if (coregroup_id == cpu_to_coregroup_id(i)) {
- or_cpumasks_related(cpu, i, submask_fn, cpu_coregroup_mask);
+ or_cpumasks_related(cpu, i, submask_fn, cpu_corgrp_mask);
cpumask_andnot(*mask, *mask, submask_fn(i));
} else {
- cpumask_andnot(*mask, *mask, cpu_coregroup_mask(i));
+ cpumask_andnot(*mask, *mask, cpu_corgrp_mask(i));
}
}
}
--
2.34.1
Le 26/08/2025 à 06:13, K Prateek Nayak a écrit : > Rename cpu_corgroup_{map,mask} to cpu_corgrp_{map,mask} to free up the > cpu_corgroup_* namespace. cpu_corgroup_mask() will be added back in the > subsequent commit for CONFIG_SCHED_MC enablement. This renaming seems odd and uncomplete. For instance update_coregroup_mask() should probably be renamed as well shoudln't it ? When you say cpu_corgroup_mask() will be added back, you mean the same function or a completely different function but with the same name ? What's really the difference between corgrp and coregroup ? Shouldn't also has_coregroup_support() now be renamed has_corgrp_support() ? Christophe > > No functional changes intended. > > Signed-off-by: K Prateek Nayak <kprateek.nayak@amd.com> > --- > arch/powerpc/kernel/smp.c | 26 +++++++++++++------------- > 1 file changed, 13 insertions(+), 13 deletions(-) > > diff --git a/arch/powerpc/kernel/smp.c b/arch/powerpc/kernel/smp.c > index 862f50c09539..4f48262658cc 100644 > --- a/arch/powerpc/kernel/smp.c > +++ b/arch/powerpc/kernel/smp.c > @@ -87,7 +87,7 @@ DEFINE_PER_CPU(cpumask_var_t, cpu_sibling_map); > DEFINE_PER_CPU(cpumask_var_t, cpu_smallcore_map); > DEFINE_PER_CPU(cpumask_var_t, cpu_l2_cache_map); > DEFINE_PER_CPU(cpumask_var_t, cpu_core_map); > -static DEFINE_PER_CPU(cpumask_var_t, cpu_coregroup_map); > +static DEFINE_PER_CPU(cpumask_var_t, cpu_corgrp_map); > > EXPORT_PER_CPU_SYMBOL(cpu_sibling_map); > EXPORT_PER_CPU_SYMBOL(cpu_l2_cache_map); > @@ -1045,9 +1045,9 @@ static const struct cpumask *tl_smt_mask(struct sched_domain_topology_level *tl, > } > #endif > > -static struct cpumask *cpu_coregroup_mask(int cpu) > +static struct cpumask *cpu_corgrp_mask(int cpu) > { > - return per_cpu(cpu_coregroup_map, cpu); > + return per_cpu(cpu_corgrp_map, cpu); > } > > static bool has_coregroup_support(void) > @@ -1061,7 +1061,7 @@ static bool has_coregroup_support(void) > > static const struct cpumask *cpu_mc_mask(struct sched_domain_topology_level *tl, int cpu) > { > - return cpu_coregroup_mask(cpu); > + return cpu_corgrp_mask(cpu); > } > > static const struct cpumask *cpu_pkg_mask(struct sched_domain_topology_level *tl, int cpu) > @@ -1124,7 +1124,7 @@ void __init smp_prepare_cpus(unsigned int max_cpus) > zalloc_cpumask_var_node(&per_cpu(cpu_core_map, cpu), > GFP_KERNEL, cpu_to_node(cpu)); > if (has_coregroup_support()) > - zalloc_cpumask_var_node(&per_cpu(cpu_coregroup_map, cpu), > + zalloc_cpumask_var_node(&per_cpu(cpu_corgrp_map, cpu), > GFP_KERNEL, cpu_to_node(cpu)); > > #ifdef CONFIG_NUMA > @@ -1145,7 +1145,7 @@ void __init smp_prepare_cpus(unsigned int max_cpus) > cpumask_set_cpu(boot_cpuid, cpu_core_mask(boot_cpuid)); > > if (has_coregroup_support()) > - cpumask_set_cpu(boot_cpuid, cpu_coregroup_mask(boot_cpuid)); > + cpumask_set_cpu(boot_cpuid, cpu_corgrp_mask(boot_cpuid)); > > init_big_cores(); > if (has_big_cores) { > @@ -1510,8 +1510,8 @@ static void remove_cpu_from_masks(int cpu) > set_cpus_unrelated(cpu, i, cpu_core_mask); > > if (has_coregroup_support()) { > - for_each_cpu(i, cpu_coregroup_mask(cpu)) > - set_cpus_unrelated(cpu, i, cpu_coregroup_mask); > + for_each_cpu(i, cpu_corgrp_mask(cpu)) > + set_cpus_unrelated(cpu, i, cpu_corgrp_mask); > } > } > #endif > @@ -1543,7 +1543,7 @@ static void update_coregroup_mask(int cpu, cpumask_var_t *mask) > if (!*mask) { > /* Assume only siblings are part of this CPU's coregroup */ > for_each_cpu(i, submask_fn(cpu)) > - set_cpus_related(cpu, i, cpu_coregroup_mask); > + set_cpus_related(cpu, i, cpu_corgrp_mask); > > return; > } > @@ -1551,18 +1551,18 @@ static void update_coregroup_mask(int cpu, cpumask_var_t *mask) > cpumask_and(*mask, cpu_online_mask, cpu_node_mask(cpu)); > > /* Update coregroup mask with all the CPUs that are part of submask */ > - or_cpumasks_related(cpu, cpu, submask_fn, cpu_coregroup_mask); > + or_cpumasks_related(cpu, cpu, submask_fn, cpu_corgrp_mask); > > /* Skip all CPUs already part of coregroup mask */ > - cpumask_andnot(*mask, *mask, cpu_coregroup_mask(cpu)); > + cpumask_andnot(*mask, *mask, cpu_corgrp_mask(cpu)); > > for_each_cpu(i, *mask) { > /* Skip all CPUs not part of this coregroup */ > if (coregroup_id == cpu_to_coregroup_id(i)) { > - or_cpumasks_related(cpu, i, submask_fn, cpu_coregroup_mask); > + or_cpumasks_related(cpu, i, submask_fn, cpu_corgrp_mask); > cpumask_andnot(*mask, *mask, submask_fn(i)); > } else { > - cpumask_andnot(*mask, *mask, cpu_coregroup_mask(i)); > + cpumask_andnot(*mask, *mask, cpu_corgrp_mask(i)); > } > } > }
Hello Christophe, On 8/26/2025 10:32 AM, Christophe Leroy wrote: > > > Le 26/08/2025 à 06:13, K Prateek Nayak a écrit : >> Rename cpu_corgroup_{map,mask} to cpu_corgrp_{map,mask} to free up the >> cpu_corgroup_* namespace. cpu_corgroup_mask() will be added back in the >> subsequent commit for CONFIG_SCHED_MC enablement. > > This renaming seems odd and uncomplete. For instance update_coregroup_mask() should probably be renamed as well shoudln't it ? So this was a bad copypasta on my part! It should have been s/cpu_coregroup_*/cpu_coregrp_*/ > > When you say cpu_corgroup_mask() will be added back, you mean the same function or a completely different function but with the same name ? > > What's really the difference between corgrp and coregroup ? > > Shouldn't also has_coregroup_support() now be renamed has_corgrp_support() ? The main intention was that kernel/sched/topology.c uses the cpu_coregroup_mask() as the default function for to derive the mask for MC domain in the default topology and PPC uses it internally for this file only. Peter just exposed the cpu_coregroup_mask() as is in https://git.kernel.org/pub/scm/linux/kernel/git/peterz/queue.git/commit/?h=sched/core&id=6e890353ce7e983a621d30413d4fc6d228ae1b4f which should be fine too since the PPC side overrides the default topology and can decide to add or ommit the MC bits. I was erring on the side of caution and allowing cpu_coregroup_mask() to return the node mask if has_coregroup_support() returns false but given the MC domain is never added when has_coregroup_support() returns false, we don't need all this. -- Thanks and Regards, Prateek
© 2016 - 2025 Red Hat, Inc.