sg_span           842 kernel/sched/topology.c 	const struct cpumask *sg_span = sched_group_span(sg);
sg_span           849 kernel/sched/topology.c 	for_each_cpu(i, sg_span) {
sg_span           861 kernel/sched/topology.c 		if (!cpumask_equal(sg_span, sched_domain_span(sibling->child)))
sg_span           880 kernel/sched/topology.c 	struct cpumask *sg_span;
sg_span           888 kernel/sched/topology.c 	sg_span = sched_group_span(sg);
sg_span           890 kernel/sched/topology.c 		cpumask_copy(sg_span, sched_domain_span(sd->child));
sg_span           892 kernel/sched/topology.c 		cpumask_copy(sg_span, sched_domain_span(sd));
sg_span           903 kernel/sched/topology.c 	struct cpumask *sg_span;
sg_span           920 kernel/sched/topology.c 	sg_span = sched_group_span(sg);
sg_span           921 kernel/sched/topology.c 	sg->sgc->capacity = SCHED_CAPACITY_SCALE * cpumask_weight(sg_span);
sg_span           939 kernel/sched/topology.c 		struct cpumask *sg_span;
sg_span           963 kernel/sched/topology.c 		sg_span = sched_group_span(sg);
sg_span           964 kernel/sched/topology.c 		cpumask_or(covered, covered, sg_span);