sg_span 842 kernel/sched/topology.c const struct cpumask *sg_span = sched_group_span(sg); sg_span 849 kernel/sched/topology.c for_each_cpu(i, sg_span) { sg_span 861 kernel/sched/topology.c if (!cpumask_equal(sg_span, sched_domain_span(sibling->child))) sg_span 880 kernel/sched/topology.c struct cpumask *sg_span; sg_span 888 kernel/sched/topology.c sg_span = sched_group_span(sg); sg_span 890 kernel/sched/topology.c cpumask_copy(sg_span, sched_domain_span(sd->child)); sg_span 892 kernel/sched/topology.c cpumask_copy(sg_span, sched_domain_span(sd)); sg_span 903 kernel/sched/topology.c struct cpumask *sg_span; sg_span 920 kernel/sched/topology.c sg_span = sched_group_span(sg); sg_span 921 kernel/sched/topology.c sg->sgc->capacity = SCHED_CAPACITY_SCALE * cpumask_weight(sg_span); sg_span 939 kernel/sched/topology.c struct cpumask *sg_span; sg_span 963 kernel/sched/topology.c sg_span = sched_group_span(sg); sg_span 964 kernel/sched/topology.c cpumask_or(covered, covered, sg_span);