sched_domains_numa_masks 1286 kernel/sched/topology.c static struct cpumask		***sched_domains_numa_masks;
sched_domains_numa_masks 1464 kernel/sched/topology.c 	return sched_domains_numa_masks[sched_domains_curr_level][cpu_to_node(cpu)];
sched_domains_numa_masks 1631 kernel/sched/topology.c 	sched_domains_numa_masks = kzalloc(sizeof(void *) * level, GFP_KERNEL);
sched_domains_numa_masks 1632 kernel/sched/topology.c 	if (!sched_domains_numa_masks)
sched_domains_numa_masks 1640 kernel/sched/topology.c 		sched_domains_numa_masks[i] =
sched_domains_numa_masks 1642 kernel/sched/topology.c 		if (!sched_domains_numa_masks[i])
sched_domains_numa_masks 1650 kernel/sched/topology.c 			sched_domains_numa_masks[i][j] = mask;
sched_domains_numa_masks 1713 kernel/sched/topology.c 				cpumask_set_cpu(cpu, sched_domains_numa_masks[i][j]);
sched_domains_numa_masks 1724 kernel/sched/topology.c 			cpumask_clear_cpu(cpu, sched_domains_numa_masks[i][j]);
sched_domains_numa_masks 1741 kernel/sched/topology.c 		cpu = cpumask_any_and(cpus, sched_domains_numa_masks[i][j]);