Lines Matching refs:per_cpu_ptr

6116 		sibling = *per_cpu_ptr(sdd->sd, i);  in build_group_mask()
6151 sibling = *per_cpu_ptr(sdd->sd, i); in build_overlap_sched_groups()
6171 sg->sgc = *per_cpu_ptr(sdd->sgc, i); in build_overlap_sched_groups()
6210 struct sched_domain *sd = *per_cpu_ptr(sdd->sd, cpu); in get_group()
6217 *sg = *per_cpu_ptr(sdd->sg, cpu); in get_group()
6218 (*sg)->sgc = *per_cpu_ptr(sdd->sgc, cpu); in get_group()
6391 WARN_ON_ONCE(*per_cpu_ptr(sdd->sd, cpu) != sd); in claim_allocations()
6392 *per_cpu_ptr(sdd->sd, cpu) = NULL; in claim_allocations()
6394 if (atomic_read(&(*per_cpu_ptr(sdd->sg, cpu))->ref)) in claim_allocations()
6395 *per_cpu_ptr(sdd->sg, cpu) = NULL; in claim_allocations()
6397 if (atomic_read(&(*per_cpu_ptr(sdd->sgc, cpu))->ref)) in claim_allocations()
6398 *per_cpu_ptr(sdd->sgc, cpu) = NULL; in claim_allocations()
6431 struct sched_domain *sd = *per_cpu_ptr(tl->data.sd, cpu); in sd_init()
6877 *per_cpu_ptr(sdd->sd, j) = sd; in __sdt_alloc()
6886 *per_cpu_ptr(sdd->sg, j) = sg; in __sdt_alloc()
6893 *per_cpu_ptr(sdd->sgc, j) = sgc; in __sdt_alloc()
6912 sd = *per_cpu_ptr(sdd->sd, j); in __sdt_free()
6915 kfree(*per_cpu_ptr(sdd->sd, j)); in __sdt_free()
6919 kfree(*per_cpu_ptr(sdd->sg, j)); in __sdt_free()
6921 kfree(*per_cpu_ptr(sdd->sgc, j)); in __sdt_free()
6990 *per_cpu_ptr(d.sd, i) = sd; in build_sched_domains()
7000 for (sd = *per_cpu_ptr(d.sd, i); sd; sd = sd->parent) { in build_sched_domains()
7017 for (sd = *per_cpu_ptr(d.sd, i); sd; sd = sd->parent) { in build_sched_domains()
7026 sd = *per_cpu_ptr(d.sd, i); in build_sched_domains()