Lines Matching refs:sched_domain

622 	struct sched_domain *sd;  in get_nohz_timer_target()
1676 struct sched_domain *sd; in ttwu_stat()
5377 sd_alloc_ctl_domain_table(struct sched_domain *sd) in sd_alloc_ctl_domain_table()
5420 struct sched_domain *sd; in sd_alloc_ctl_cpu_table()
5664 static int sched_domain_debug_one(struct sched_domain *sd, int cpu, int level, in sched_domain_debug_one()
5737 static void sched_domain_debug(struct sched_domain *sd, int cpu) in sched_domain_debug()
5768 static int sd_degenerate(struct sched_domain *sd) in sd_degenerate()
5793 sd_parent_degenerate(struct sched_domain *sd, struct sched_domain *parent) in sd_parent_degenerate()
5955 struct sched_domain *sd = container_of(rcu, struct sched_domain, rcu); in free_sched_domain()
5970 static void destroy_sched_domain(struct sched_domain *sd, int cpu) in destroy_sched_domain()
5975 static void destroy_sched_domains(struct sched_domain *sd, int cpu) in destroy_sched_domains()
5990 DEFINE_PER_CPU(struct sched_domain *, sd_llc);
5993 DEFINE_PER_CPU(struct sched_domain *, sd_numa);
5994 DEFINE_PER_CPU(struct sched_domain *, sd_busy);
5995 DEFINE_PER_CPU(struct sched_domain *, sd_asym);
5999 struct sched_domain *sd; in update_top_cache_domain()
6000 struct sched_domain *busy_sd = NULL; in update_top_cache_domain()
6028 cpu_attach_domain(struct sched_domain *sd, struct root_domain *rd, int cpu) in cpu_attach_domain()
6031 struct sched_domain *tmp; in cpu_attach_domain()
6035 struct sched_domain *parent = tmp->parent; in cpu_attach_domain()
6084 struct sched_domain ** __percpu sd;
6108 static void build_group_mask(struct sched_domain *sd, struct sched_group *sg) in build_group_mask()
6112 struct sched_domain *sibling; in build_group_mask()
6134 build_overlap_sched_groups(struct sched_domain *sd, int cpu) in build_overlap_sched_groups()
6140 struct sched_domain *sibling; in build_overlap_sched_groups()
6210 struct sched_domain *sd = *per_cpu_ptr(sdd->sd, cpu); in get_group()
6211 struct sched_domain *child = sd->child; in get_group()
6233 build_sched_groups(struct sched_domain *sd, int cpu) in build_sched_groups()
6291 static void init_sched_groups_capacity(int cpu, struct sched_domain *sd) in init_sched_groups_capacity()
6326 static void set_domain_attribute(struct sched_domain *sd, in set_domain_attribute()
6373 d->sd = alloc_percpu(struct sched_domain *); in __visit_domain_allocation_hell()
6387 static void claim_allocations(int cpu, struct sched_domain *sd) in claim_allocations()
6428 static struct sched_domain *
6431 struct sched_domain *sd = *per_cpu_ptr(tl->data.sd, cpu); in sd_init()
6449 *sd = (struct sched_domain){ in sd_init()
6855 sdd->sd = alloc_percpu(struct sched_domain *); in __sdt_alloc()
6868 struct sched_domain *sd; in __sdt_alloc()
6872 sd = kzalloc_node(sizeof(struct sched_domain) + cpumask_size(), in __sdt_alloc()
6909 struct sched_domain *sd; in __sdt_free()
6932 struct sched_domain *build_sched_domain(struct sched_domain_topology_level *tl, in build_sched_domain()
6934 struct sched_domain *child, int cpu) in build_sched_domain()
6936 struct sched_domain *sd = sd_init(tl, cpu); in build_sched_domain()
6974 struct sched_domain *sd; in build_sched_domains()