/linux-4.4.14/arch/x86/mm/ |
H A D | numa.c | 119 /* setup nr_node_ids if not done yet */ setup_node_to_cpumask_map() 120 if (nr_node_ids == MAX_NUMNODES) setup_node_to_cpumask_map() 124 for (node = 0; node < nr_node_ids; node++) setup_node_to_cpumask_map() 128 pr_debug("Node to cpumask map for %d nodes\n", nr_node_ids); setup_node_to_cpumask_map() 851 if (node >= nr_node_ids) { cpumask_of_node() 853 "cpumask_of_node(%d): node > nr_node_ids(%d)\n", cpumask_of_node() 854 node, nr_node_ids); cpumask_of_node()
|
/linux-4.4.14/arch/x86/kernel/ |
H A D | setup_percpu.c | 173 pr_info("NR_CPUS:%d nr_cpumask_bits:%d nr_cpu_ids:%d nr_node_ids:%d\n", setup_per_cpu_areas() 174 NR_CPUS, nr_cpumask_bits, nr_cpu_ids, nr_node_ids); setup_per_cpu_areas()
|
/linux-4.4.14/include/linux/ |
H A D | nodemask.h | 432 extern int nr_node_ids; 473 #define nr_node_ids 1 macro
|
/linux-4.4.14/mm/ |
H A D | list_lru.c | 538 size_t size = sizeof(*lru->node) * nr_node_ids; __list_lru_init()
|
H A D | slab.h | 364 for (__node = 0; __node < nr_node_ids; __node++) \
|
H A D | ksm.c | 2233 buf = kcalloc(nr_node_ids + nr_node_ids, sizeof(*buf), merge_across_nodes_store() 2240 root_unstable_tree = buf + nr_node_ids; merge_across_nodes_store() 2247 ksm_nr_node_ids = knob ? 1 : nr_node_ids; merge_across_nodes_store()
|
H A D | slub.c | 3904 nr_node_ids * sizeof(struct kmem_cache_node *), kmem_cache_init() 3927 nr_cpu_ids, nr_node_ids); kmem_cache_init() 4482 nodes = kzalloc(sizeof(unsigned long) * nr_node_ids, GFP_KERNEL); show_slab_objects() 4561 for (node = 0; node < nr_node_ids; node++)
|
H A D | vmalloc.c | 2605 memset(counters, 0, nr_node_ids * sizeof(unsigned int)); show_numa_info() 2673 nr_node_ids * sizeof(unsigned int)); vmalloc_open()
|
H A D | slab.c | 886 size_t memsize = sizeof(void *) * nr_node_ids; alloc_alien_cache() 1442 * struct kmem_cache size depends on nr_node_ids & nr_cpu_ids kmem_cache_init() 1446 nr_node_ids * sizeof(struct kmem_cache_node *), kmem_cache_init()
|
H A D | page_alloc.c | 261 int nr_node_ids __read_mostly = MAX_NUMNODES; 263 EXPORT_SYMBOL(nr_node_ids); variable 5377 nr_node_ids = highest + 1; setup_nr_node_ids()
|
H A D | compaction.c | 1728 if (nid >= 0 && nid < nr_node_ids && node_online(nid)) { sysfs_compact_node()
|
H A D | hugetlb.c | 2513 for (nid = 0; nid < nr_node_ids; nid++) { kobj_to_node_hstate() 2568 for (nid = 0; nid < nr_node_ids; nid++) hugetlb_unregister_all_nodes()
|
H A D | memcontrol.c | 735 VM_BUG_ON((unsigned)nid >= nr_node_ids); mem_cgroup_node_nr_lru_pages() 4168 size += nr_node_ids * sizeof(struct mem_cgroup_per_node *); mem_cgroup_alloc()
|
H A D | vmscan.c | 235 if (nr_node_ids == 1) register_shrinker() 239 size *= nr_node_ids; register_shrinker()
|
/linux-4.4.14/arch/powerpc/mm/ |
H A D | numa.c | 78 /* setup nr_node_ids if not done yet */ setup_node_to_cpumask_map() 79 if (nr_node_ids == MAX_NUMNODES) setup_node_to_cpumask_map() 87 dbg("Node to cpumask map for %d nodes\n", nr_node_ids); setup_node_to_cpumask_map()
|
/linux-4.4.14/drivers/hv/ |
H A D | hv.c | 412 hv_context.hv_numa_map = kzalloc(sizeof(struct cpumask) * nr_node_ids, hv_synic_alloc()
|
H A D | channel_mgmt.c | 438 if (next_node == nr_node_ids) init_vp_index()
|
/linux-4.4.14/net/openvswitch/ |
H A D | flow_table.c | 759 + (nr_node_ids ovs_flow_init()
|
/linux-4.4.14/kernel/sched/ |
H A D | core.c | 5813 if (nr_node_ids == 1) sd_parent_degenerate() 6570 for (i = 0; i < nr_node_ids; i++) { sched_numa_warn() 6572 for (j = 0; j < nr_node_ids; j++) sched_numa_warn() 6653 sched_domains_numa_distance = kzalloc(sizeof(int) * nr_node_ids, GFP_KERNEL); sched_init_numa() 6665 for (i = 0; i < nr_node_ids; i++) { sched_init_numa() 6666 for (j = 0; j < nr_node_ids; j++) { sched_init_numa() 6667 for (k = 0; k < nr_node_ids; k++) { sched_init_numa() 6732 kzalloc(nr_node_ids * sizeof(void *), GFP_KERNEL); sched_init_numa() 6736 for (j = 0; j < nr_node_ids; j++) { sched_init_numa() 6793 for (j = 0; j < nr_node_ids; j++) { sched_domains_numa_masks_set() 6804 for (j = 0; j < nr_node_ids; j++) sched_domains_numa_masks_clear()
|
H A D | fair.c | 918 return NR_NUMA_HINT_FAULT_TYPES * (s * nr_node_ids + nid) + priv; task_faults_idx() 1943 4*nr_node_ids*sizeof(unsigned long); task_numa_group() 1954 nr_node_ids; task_numa_group() 1958 for (i = 0; i < NR_NUMA_HINT_FAULT_STATS * nr_node_ids; i++) task_numa_group() 2016 for (i = 0; i < NR_NUMA_HINT_FAULT_STATS * nr_node_ids; i++) { task_numa_group() 2048 for (i = 0; i < NR_NUMA_HINT_FAULT_STATS * nr_node_ids; i++) task_numa_free() 2083 NR_NUMA_HINT_FAULT_BUCKETS * nr_node_ids; task_numa_fault()
|
/linux-4.4.14/drivers/net/ethernet/cavium/thunder/ |
H A D | nic_main.c | 896 if (nr_node_ids > 1) nic_num_sqs_en()
|
/linux-4.4.14/net/sunrpc/ |
H A D | svc.c | 195 unsigned int maxpools = nr_node_ids; svc_pool_map_init_pernode()
|
/linux-4.4.14/kernel/ |
H A D | workqueue.c | 3543 ctx = kzalloc(sizeof(*ctx) + nr_node_ids * sizeof(ctx->pwq_tbl[0]), apply_wqattrs_prepare() 3843 tbl_size = nr_node_ids * sizeof(wq->numa_pwq_tbl[0]); __alloc_workqueue_key() 5241 tbl = kzalloc(nr_node_ids * sizeof(tbl[0]), GFP_KERNEL); wq_numa_init()
|