sched/topology: Remove redundant cpumask_and() in init_overlap_sched_group()
mask is built in build_balance_mask() by for_each_cpu(i, sg_span), so it must be a subset of sched_group_span(sg). So the cpumask_and() call is redundant - remove it. [ mingo: Adjusted the changelog a bit. ] Signed-off-by: Barry Song <song.bao.hua@hisilicon.com> Signed-off-by: Ingo Molnar <mingo@kernel.org> Reviewed-by: Valentin Schneider <Valentin.Schneider@arm.com> Link: https://lore.kernel.org/r/20210325023140.23456-1-song.bao.hua@hisilicon.com
This commit is contained in:
parent
c4681f3f1c
commit
0a2b65c03e
|
@ -934,7 +934,7 @@ static void init_overlap_sched_group(struct sched_domain *sd,
|
|||
int cpu;
|
||||
|
||||
build_balance_mask(sd, sg, mask);
|
||||
cpu = cpumask_first_and(sched_group_span(sg), mask);
|
||||
cpu = cpumask_first(mask);
|
||||
|
||||
sg->sgc = *per_cpu_ptr(sdd->sgc, cpu);
|
||||
if (atomic_inc_return(&sg->sgc->ref) == 1)
|
||||
|
|
Loading…
Reference in New Issue