Lines Matching refs:span
2297 for_each_cpu_and(i, rd->span, cpu_active_mask) in dl_bw_cpus()
3934 cpumask_t *span = rq->rd->span; in __sched_setscheduler() local
3941 if (!cpumask_subset(span, &p->cpus_allowed) || in __sched_setscheduler()
4457 if (!cpumask_subset(task_rq(p)->rd->span, new_mask)) { in sched_setaffinity()
5087 if (dl_task(p) && !cpumask_intersects(task_rq(p)->rd->span, in task_can_attach()
5536 BUG_ON(!cpumask_test_cpu(cpu, rq->rd->span)); in migration_call()
5549 BUG_ON(!cpumask_test_cpu(cpu, rq->rd->span)); in migration_call()
5831 free_cpumask_var(rd->span); in free_rootdomain()
5848 cpumask_clear_cpu(rq->cpu, old_rd->span); in rq_attach_root()
5862 cpumask_set_cpu(rq->cpu, rd->span); in rq_attach_root()
5876 if (!zalloc_cpumask_var(&rd->span, GFP_KERNEL)) in init_rootdomain()
5900 free_cpumask_var(rd->span); in init_rootdomain()
6110 const struct cpumask *span = sched_domain_span(sd); in build_group_mask() local
6115 for_each_cpu(i, span) { in build_group_mask()
6137 const struct cpumask *span = sched_domain_span(sd); in build_overlap_sched_groups() local
6145 for_each_cpu(i, span) { in build_overlap_sched_groups()
6237 const struct cpumask *span = sched_domain_span(sd); in build_sched_groups() local
6244 if (cpu != cpumask_first(span)) in build_sched_groups()
6252 for_each_cpu(i, span) { in build_sched_groups()
6262 for_each_cpu(j, span) { in build_sched_groups()