sched_domains_numa_masks 1286 kernel/sched/topology.c static struct cpumask ***sched_domains_numa_masks; sched_domains_numa_masks 1464 kernel/sched/topology.c return sched_domains_numa_masks[sched_domains_curr_level][cpu_to_node(cpu)]; sched_domains_numa_masks 1631 kernel/sched/topology.c sched_domains_numa_masks = kzalloc(sizeof(void *) * level, GFP_KERNEL); sched_domains_numa_masks 1632 kernel/sched/topology.c if (!sched_domains_numa_masks) sched_domains_numa_masks 1640 kernel/sched/topology.c sched_domains_numa_masks[i] = sched_domains_numa_masks 1642 kernel/sched/topology.c if (!sched_domains_numa_masks[i]) sched_domains_numa_masks 1650 kernel/sched/topology.c sched_domains_numa_masks[i][j] = mask; sched_domains_numa_masks 1713 kernel/sched/topology.c cpumask_set_cpu(cpu, sched_domains_numa_masks[i][j]); sched_domains_numa_masks 1724 kernel/sched/topology.c cpumask_clear_cpu(cpu, sched_domains_numa_masks[i][j]); sched_domains_numa_masks 1741 kernel/sched/topology.c cpu = cpumask_any_and(cpus, sched_domains_numa_masks[i][j]);