/linux-4.4.14/include/linux/ |
H A D | numa.h | 11 #define MAX_NUMNODES (1 << NODES_SHIFT) macro
|
H A D | nodemask.h | 44 * int first_node(mask) Number lowest set bit, or MAX_NUMNODES 45 * int next_node(node, mask) Next node past 'node', or MAX_NUMNODES 47 * MAX_NUMNODES. 95 typedef struct { DECLARE_BITMAP(bits, MAX_NUMNODES); } nodemask_t; 104 #define nodemask_pr_args(maskp) MAX_NUMNODES, (maskp)->bits 127 #define nodes_setall(dst) __nodes_setall(&(dst), MAX_NUMNODES) __nodes_setall() 133 #define nodes_clear(dst) __nodes_clear(&(dst), MAX_NUMNODES) __nodes_clear() 150 __nodes_and(&(dst), &(src1), &(src2), MAX_NUMNODES) __nodes_and() 158 __nodes_or(&(dst), &(src1), &(src2), MAX_NUMNODES) __nodes_or() 166 __nodes_xor(&(dst), &(src1), &(src2), MAX_NUMNODES) __nodes_xor() 174 __nodes_andnot(&(dst), &(src1), &(src2), MAX_NUMNODES) __nodes_andnot() 182 __nodes_complement(&(dst), &(src), MAX_NUMNODES) __nodes_complement() 190 __nodes_equal(&(src1), &(src2), MAX_NUMNODES) __nodes_equal() 198 __nodes_intersects(&(src1), &(src2), MAX_NUMNODES) __nodes_intersects() 206 __nodes_subset(&(src1), &(src2), MAX_NUMNODES) __nodes_subset() 213 #define nodes_empty(src) __nodes_empty(&(src), MAX_NUMNODES) __nodes_empty() 219 #define nodes_full(nodemask) __nodes_full(&(nodemask), MAX_NUMNODES) __nodes_full() 225 #define nodes_weight(nodemask) __nodes_weight(&(nodemask), MAX_NUMNODES) __nodes_weight() 232 __nodes_shift_right(&(dst), &(src), (n), MAX_NUMNODES) __nodes_shift_right() 240 __nodes_shift_left(&(dst), &(src), (n), MAX_NUMNODES) __nodes_shift_left() 248 > MAX_NUMNODES, then the silly min_ts could be dropped. */ 253 return min_t(int, MAX_NUMNODES, find_first_bit(srcp->bits, MAX_NUMNODES)); __first_node() 259 return min_t(int,MAX_NUMNODES,find_next_bit(srcp->bits, MAX_NUMNODES, n+1)); __next_node() 282 return min_t(int,MAX_NUMNODES, __first_unset_node() 283 find_first_zero_bit(maskp->bits, MAX_NUMNODES)); __first_unset_node() 286 #define NODE_MASK_LAST_WORD BITMAP_LAST_WORD_MASK(MAX_NUMNODES) 288 #if MAX_NUMNODES <= BITS_PER_LONG 292 [BITS_TO_LONGS(MAX_NUMNODES)-1] = NODE_MASK_LAST_WORD \ 299 [0 ... BITS_TO_LONGS(MAX_NUMNODES)-2] = ~0UL, \ 300 [BITS_TO_LONGS(MAX_NUMNODES)-1] = NODE_MASK_LAST_WORD \ 307 [0 ... BITS_TO_LONGS(MAX_NUMNODES)-1] = 0UL \ 313 __nodemask_parse_user((ubuf), (ulen), &(dst), MAX_NUMNODES) __nodemask_parse_user() 320 #define nodelist_parse(buf, dst) __nodelist_parse((buf), &(dst), MAX_NUMNODES) __nodelist_parse() 327 __node_remap((oldbit), &(old), &(new), MAX_NUMNODES) __node_remap() 335 __nodes_remap(&(dst), &(src), &(old), &(new), MAX_NUMNODES) __nodes_remap() 343 __nodes_onto(&(dst), &(orig), &(relmap), MAX_NUMNODES) __nodes_onto() 351 __nodes_fold(&(dst), &(orig), sz, MAX_NUMNODES) __nodes_fold() 358 #if MAX_NUMNODES > 1 361 (node) < MAX_NUMNODES; \ 363 #else /* MAX_NUMNODES == 1 */ 367 #endif /* MAX_NUMNODES */ 397 #if MAX_NUMNODES > 1 node_state() 472 #define next_online_node(nid) (MAX_NUMNODES) 481 #if defined(CONFIG_NUMA) && (MAX_NUMNODES > 1)
|
H A D | slab_def.h | 76 struct kmem_cache_node *node[MAX_NUMNODES];
|
H A D | hugetlb.h | 314 struct list_head hugepage_freelists[MAX_NUMNODES]; 315 unsigned int nr_huge_pages_node[MAX_NUMNODES]; 316 unsigned int free_huge_pages_node[MAX_NUMNODES]; 317 unsigned int surplus_huge_pages_node[MAX_NUMNODES];
|
H A D | topology.h | 122 extern int _node_numa_mem_[MAX_NUMNODES];
|
H A D | slub_def.h | 101 struct kmem_cache_node *node[MAX_NUMNODES];
|
H A D | memblock.h | 196 * @nid: node selector, %MAX_NUMNODES for all nodes
|
H A D | gfp.h | 425 VM_BUG_ON(nid < 0 || nid >= MAX_NUMNODES); __alloc_pages_node()
|
H A D | memcontrol.h | 260 #if MAX_NUMNODES > 1
|
H A D | mmzone.h | 577 #define MAX_ZONES_PER_ZONELIST (MAX_NUMNODES * MAX_NR_ZONES)
|
H A D | mm.h | 2310 #if MAX_NUMNODES > 1 page_is_guard()
|
/linux-4.4.14/include/acpi/ |
H A D | acpi_numa.h | 8 #if MAX_NUMNODES > 256 9 #define MAX_PXM_DOMAINS MAX_NUMNODES
|
/linux-4.4.14/arch/mips/include/asm/mach-loongson64/ |
H A D | topology.h | 15 extern unsigned char __node_distances[MAX_NUMNODES][MAX_NUMNODES];
|
/linux-4.4.14/arch/ia64/include/asm/ |
H A D | mmzone.h | 32 # define NR_NODE_MEMBLKS (MAX_NUMNODES * 8) 35 # define NR_NODE_MEMBLKS (MAX_NUMNODES * 4) 39 # define NR_NODE_MEMBLKS (MAX_NUMNODES * 4)
|
H A D | numa.h | 26 extern cpumask_t node_to_cpu_mask[MAX_NUMNODES] __cacheline_aligned; 27 extern pg_data_t *pgdat_list[MAX_NUMNODES]; 61 extern u8 numa_slit[MAX_NUMNODES * MAX_NUMNODES]; 62 #define node_distance(from,to) (numa_slit[(from) * MAX_NUMNODES + (to)])
|
H A D | acpi.h | 100 #if MAX_NUMNODES > 256 101 #define MAX_PXM_DOMAINS MAX_NUMNODES 106 extern int __initdata nid_to_pxm_map[MAX_NUMNODES];
|
H A D | nodedata.h | 29 struct pglist_data *pg_data_ptrs[MAX_NUMNODES];
|
/linux-4.4.14/arch/s390/include/asm/ |
H A D | numa.h | 22 extern cpumask_t node_to_cpumask_map[MAX_NUMNODES];
|
/linux-4.4.14/arch/metag/include/asm/ |
H A D | mmzone.h | 14 for (nid = 0; nid < MAX_NUMNODES; nid++) pfn_to_nid()
|
/linux-4.4.14/arch/sh/mm/ |
H A D | numa.c | 18 struct pglist_data *node_data[MAX_NUMNODES] __read_mostly; 34 BUG_ON(nid >= MAX_NUMNODES || nid <= 0); setup_bootmem_node()
|
/linux-4.4.14/arch/s390/numa/ |
H A D | numa.c | 23 pg_data_t *node_data[MAX_NUMNODES]; 26 cpumask_t node_to_cpumask_map[MAX_NUMNODES]; 102 for (nid = 0; nid < MAX_NUMNODES; nid++) numa_setup_memory()
|
H A D | mode_emu.c | 64 int per_node[MAX_NUMNODES]; /* Number of cores pinned to node */ 385 for (i = 0; i < MAX_NUMNODES; i++) toptree_to_topology() 523 emu_nodes = min(count, MAX_NUMNODES); early_parse_emu_nodes()
|
/linux-4.4.14/arch/ia64/include/asm/sn/ |
H A D | arch.h | 32 * should be deleted and TIOs should be included in MAX_NUMNODES. 34 #define MAX_TIO_NODES MAX_NUMNODES 35 #define MAX_COMPACT_NODES (MAX_NUMNODES + MAX_TIO_NODES)
|
/linux-4.4.14/arch/m32r/mm/ |
H A D | discontig.c | 21 struct pglist_data *node_data[MAX_NUMNODES]; 24 pg_data_t m32r_node_data[MAX_NUMNODES]; 33 static mem_prof_t mem_prof[MAX_NUMNODES];
|
/linux-4.4.14/arch/sh/include/asm/ |
H A D | mmzone.h | 16 for (nid = 0; nid < MAX_NUMNODES; nid++) pfn_to_nid()
|
/linux-4.4.14/arch/m32r/include/asm/ |
H A D | mmzone.h | 40 for (node = 0 ; node < MAX_NUMNODES ; node++) pfn_to_nid()
|
/linux-4.4.14/arch/metag/mm/ |
H A D | numa.c | 19 struct pglist_data *node_data[MAX_NUMNODES] __read_mostly; 37 BUG_ON(nid >= MAX_NUMNODES || nid <= 0); setup_bootmem_node()
|
/linux-4.4.14/arch/ia64/kernel/ |
H A D | numa.c | 30 cpumask_t node_to_cpu_mask[MAX_NUMNODES] __cacheline_aligned; 73 for(node=0; node < MAX_NUMNODES; node++) build_cpu_to_node_map()
|
H A D | uncached.c | 43 struct uncached_pool uncached_pools[MAX_NUMNODES]; 191 if (unlikely(starting_nid >= MAX_NUMNODES)) uncached_alloc_page() 211 } while ((nid = (nid + 1) % MAX_NUMNODES) != starting_nid); uncached_alloc_page()
|
H A D | machine_kexec.c | 150 VMCOREINFO_LENGTH(pgdat_list, MAX_NUMNODES); arch_crash_save_vmcoreinfo()
|
H A D | acpi.c | 540 * mapping with sparse node numbering iff MAX_PXM_DOMAINS <= MAX_NUMNODES. acpi_numa_arch_fixup() 574 for (i = 0; i < MAX_NUMNODES; i++) 575 for (j = 0; j < MAX_NUMNODES; j++)
|
H A D | iosapic.c | 660 iosapic_lists[iosapic_index].node == MAX_NUMNODES) get_target_cpu() 1071 iosapic_lists[index].node = MAX_NUMNODES; iosapic_init()
|
/linux-4.4.14/arch/ia64/mm/ |
H A D | numa.c | 37 u8 numa_slit[MAX_NUMNODES * MAX_NUMNODES];
|
H A D | discontig.c | 47 static struct early_node_data mem_data[MAX_NUMNODES] __initdata; 50 pg_data_t *pgdat_list[MAX_NUMNODES]; 195 ai = pcpu_alloc_alloc_info(MAX_NUMNODES, nr_cpu_ids); setup_per_cpu_areas() 557 for (node = MAX_NUMNODES - 1; node >= 0; node--) { 704 sparse_memory_present_with_active_regions(MAX_NUMNODES); paging_init()
|
/linux-4.4.14/arch/mips/loongson64/loongson-3/ |
H A D | numa.c | 34 static struct node_data prealloc__node_data[MAX_NUMNODES]; 35 unsigned char __node_distances[MAX_NUMNODES][MAX_NUMNODES]; 37 struct node_data *__node_data[MAX_NUMNODES]; 91 for (row = 0; row < MAX_NUMNODES; row++) init_topology_matrix() 92 for (col = 0; col < MAX_NUMNODES; col++) init_topology_matrix()
|
/linux-4.4.14/arch/x86/mm/ |
H A D | numa_emulation.c | 13 static int emu_nid_to_phys[MAX_NUMNODES]; 92 if (nr_nodes > MAX_NUMNODES) { split_nodes_interleave() 94 nr_nodes, MAX_NUMNODES); split_nodes_interleave() 95 nr_nodes = MAX_NUMNODES; split_nodes_interleave() 217 * The limit on emulated nodes is MAX_NUMNODES, so the size per node is split_nodes_size_interleave() 222 min_size = (max_addr - addr - mem_hole_size(addr, max_addr)) / MAX_NUMNODES; split_nodes_size_interleave() 273 ret = emu_setup_memblk(ei, pi, nid++ % MAX_NUMNODES, for_each_node_mask() 326 for (i = 0; i < MAX_NUMNODES; i++) numa_emulation()
|
H A D | numa.c | 26 struct pglist_data *node_data[MAX_NUMNODES] __read_mostly; 72 cpumask_var_t node_to_cpumask_map[MAX_NUMNODES]; 120 if (nr_node_ids == MAX_NUMNODES) setup_node_to_cpumask_map() 139 if (start > end || nid < 0 || nid >= MAX_NUMNODES) { numa_add_memblk_to() 495 if (r->nid != MAX_NUMNODES) for_each_memblock() 600 if (rr == MAX_NUMNODES) numa_init_array() 618 MAX_NUMNODES)); numa_init() 620 MAX_NUMNODES)); numa_init()
|
H A D | init.c | 440 for_each_mem_pfn_range(i, MAX_NUMNODES, &start_pfn, &end_pfn, NULL) { init_range_memory_mapping()
|
H A D | init_32.c | 710 sparse_memory_present_with_active_regions(MAX_NUMNODES); paging_init()
|
H A D | init_64.c | 651 sparse_memory_present_with_active_regions(MAX_NUMNODES); paging_init()
|
/linux-4.4.14/arch/tile/kernel/ |
H A D | setup.c | 58 struct pglist_data node_data[MAX_NUMNODES] __read_mostly; ABS() 62 unsigned long node_start_pfn[MAX_NUMNODES]; ABS() 63 unsigned long node_end_pfn[MAX_NUMNODES]; ABS() 64 unsigned long __initdata node_memmap_pfn[MAX_NUMNODES]; ABS() 65 unsigned long __initdata node_percpu_pfn[MAX_NUMNODES]; ABS() 66 unsigned long __initdata node_free_pfn[MAX_NUMNODES]; ABS() 68 static unsigned long __initdata node_percpu[MAX_NUMNODES]; ABS() 88 unsigned long node_lowmem_end_pfn[MAX_NUMNODES]; ABS() 95 int node_controller[MAX_NUMNODES] = { [0 ... MAX_NUMNODES-1] = -1 }; 114 static unsigned int __initdata maxnodemem_pfn[MAX_NUMNODES] = { 115 [0 ... MAX_NUMNODES-1] = -1U 146 if (node >= MAX_NUMNODES || *endp != ':') setup_maxnodemem() 391 BUILD_BUG_ON(MAX_NUMNODES > 127); setup_memory() 458 if (i >= MAX_NUMNODES) { setup_memory() 575 for (i = 1; i < MAX_NUMNODES; ++i) { setup_memory() 582 for (i = 0; i < MAX_NUMNODES; ++i) { setup_memory() 680 for (i = 0; i < MAX_NUMNODES; ++i) setup_bootmem_allocator() 847 struct cpumask node_2_cpu_mask[MAX_NUMNODES] __write_once; 885 int distance[MAX_NUMNODES][NR_CPUS]; setup_numa_mapping() 966 if (node == MAX_NUMNODES) 1514 static size_t __initdata pfn_offset[MAX_NUMNODES] = { 0 };
|
/linux-4.4.14/arch/x86/include/asm/ |
H A D | e820.h | 6 #define E820_X_MAX (E820MAX + 3 * MAX_NUMNODES)
|
H A D | numa.h | 11 #define NR_NODE_MEMBLKS (MAX_NUMNODES*2)
|
H A D | topology.h | 73 extern cpumask_var_t node_to_cpumask_map[MAX_NUMNODES];
|
/linux-4.4.14/mm/ |
H A D | mmzone.c | 21 if (nid == MAX_NUMNODES) next_online_pgdat()
|
H A D | mempolicy.c | 124 static struct mempolicy preferred_node_policy[MAX_NUMNODES]; 351 if (current->il_next >= MAX_NUMNODES) mpol_rebind_nodemask() 353 if (current->il_next >= MAX_NUMNODES) mpol_rebind_nodemask() 1258 if (nlongs > BITS_TO_LONGS(MAX_NUMNODES)) { get_nodes() 1261 for (k = BITS_TO_LONGS(MAX_NUMNODES); k < nlongs; k++) { get_nodes() 1271 nlongs = BITS_TO_LONGS(MAX_NUMNODES); get_nodes() 1286 const int nbytes = BITS_TO_LONGS(MAX_NUMNODES) * sizeof(long); copy_nodes_to_user() 1443 if (nmask != NULL && maxnode < MAX_NUMNODES) SYSCALL_DEFINE5() 1470 DECLARE_BITMAP(bm, MAX_NUMNODES); COMPAT_SYSCALL_DEFINE5() 1472 nr_bits = min_t(unsigned long, maxnode-1, MAX_NUMNODES); COMPAT_SYSCALL_DEFINE5() 1498 DECLARE_BITMAP(bm, MAX_NUMNODES); COMPAT_SYSCALL_DEFINE3() 1500 nr_bits = min_t(unsigned long, maxnode-1, MAX_NUMNODES); COMPAT_SYSCALL_DEFINE3() 1524 nr_bits = min_t(unsigned long, maxnode-1, MAX_NUMNODES); COMPAT_SYSCALL_DEFINE6() 1680 if (next >= MAX_NUMNODES) interleave_nodes() 1682 if (next < MAX_NUMNODES) interleave_nodes() 1786 get_random_int() % w, MAX_NUMNODES); node_random()
|
H A D | memblock.c | 299 memblock_set_region_node(&type->regions[0], MAX_NUMNODES); memblock_remove_region() 630 return memblock_add_region(base, size, MAX_NUMNODES, 0); memblock_add() 758 return memblock_reserve_region(base, size, MAX_NUMNODES, 0); memblock_reserve() 893 if (WARN_ONCE(nid == MAX_NUMNODES, __next_mem_range() 894 "Usage of MAX_NUMNODES is deprecated. Use NUMA_NO_NODE instead\n")) __next_mem_range() 998 if (WARN_ONCE(nid == MAX_NUMNODES, "Usage of MAX_NUMNODES is deprecated. Use NUMA_NO_NODE instead\n")) __next_mem_range_rev() 1091 if (nid == MAX_NUMNODES || nid == r->nid) __next_mem_pfn_range() 1260 if (WARN_ONCE(nid == MAX_NUMNODES, "Usage of MAX_NUMNODES is deprecated. Use NUMA_NO_NODE instead\n")) memblock_virt_alloc_internal() 1628 if (memblock_get_region_node(rgn) != MAX_NUMNODES) memblock_dump()
|
H A D | page_alloc.c | 89 int _node_numa_mem_[MAX_NUMNODES]; 253 static unsigned long __meminitdata zone_movable_pfn[MAX_NUMNODES]; 260 #if MAX_NUMNODES > 1 261 int nr_node_ids __read_mostly = MAX_NUMNODES; 4008 static int node_load[MAX_NUMNODES]; 4055 val *= (MAX_NODE_LOAD*MAX_NUMNODES); for_each_node_state() 4109 static int node_order[MAX_NUMNODES]; 4264 for (node = local_node + 1; node < MAX_NUMNODES; node++) { build_zonelists() 4819 * @nid: The node to free memory on. If MAX_NUMNODES, all nodes are freed. 4844 * @nid: The node to call memory_present for. If MAX_NUMNODES, all nodes will be used. 4860 * @nid: The nid to return the range for. If MAX_NUMNODES, the min and max PFN are returned. 4980 * Return the number of holes in a range on a node. If nid is MAX_NUMNODES, 5009 return __absent_pages_in_range(MAX_NUMNODES, start_pfn, end_pfn); absent_pages_in_range() 5368 #if MAX_NUMNODES > 1 5376 highest = find_last_bit(node_possible_map.bits, MAX_NUMNODES); setup_nr_node_ids() 5407 for_each_mem_pfn_range(i, MAX_NUMNODES, &start, &end, &nid) { node_map_pfn_alignment() 5458 return find_min_pfn_for_node(MAX_NUMNODES); find_min_pfn_with_active_regions() 5472 for_each_mem_pfn_range(i, MAX_NUMNODES, &start_pfn, &end_pfn, &nid) { early_calculate_totalpages() 5645 for (nid = 0; nid < MAX_NUMNODES; nid++) 5733 for (i = 0; i < MAX_NUMNODES; i++) { free_area_init_nodes() 5741 for_each_mem_pfn_range(i, MAX_NUMNODES, &start_pfn, &end_pfn, &nid) free_area_init_nodes()
|
H A D | page_isolation.c | 287 if (node == MAX_NUMNODES) alloc_migrate_target()
|
H A D | migrate.c | 1211 while (pm->node != MAX_NUMNODES && pm->page != p) new_page_node() 1214 if (pm->node == MAX_NUMNODES) new_page_node() 1231 * The pm array ends with node = MAX_NUMNODES. 1246 for (pp = pm; pp->node != MAX_NUMNODES; pp++) { do_move_page_to_node_array() 1366 if (node < 0 || node >= MAX_NUMNODES) do_pages_move() 1380 pm[chunk_nr_pages].node = MAX_NUMNODES; do_pages_move()
|
H A D | memcontrol.c | 134 struct mem_cgroup_tree_per_node *rb_tree_per_node[MAX_NUMNODES]; 804 #if MAX_NUMNODES > 1 memcg_check_events() 811 #if MAX_NUMNODES > 1 memcg_check_events() 1417 #if MAX_NUMNODES > 1 1493 if (node == MAX_NUMNODES) mem_cgroup_select_victim_node() 1501 if (unlikely(node == MAX_NUMNODES)) mem_cgroup_select_victim_node() 4252 memcg->last_scanned_node = MAX_NUMNODES; mem_cgroup_css_alloc()
|
H A D | bootmem.c | 37 bootmem_data_t bootmem_node_data[MAX_NUMNODES] __initdata;
|
H A D | huge_memory.c | 2343 static int khugepaged_node_load[MAX_NUMNODES]; 2360 for (i = 0; i < MAX_NUMNODES; i++) { khugepaged_scan_abort() 2376 for (nid = 0; nid < MAX_NUMNODES; nid++) khugepaged_find_target_node() 2384 for (nid = last_khugepaged_target_node + 1; nid < MAX_NUMNODES; khugepaged_find_target_node()
|
H A D | sparse.c | 39 #if MAX_NUMNODES <= 256
|
H A D | slab.c | 240 #define NUM_INIT_LISTS (2 * MAX_NUMNODES) 243 #define SIZE_NODE (MAX_NUMNODES) 616 if (node == MAX_NUMNODES) init_reap_node() 627 if (unlikely(node >= MAX_NUMNODES)) next_reap_node() 3094 VM_BUG_ON(nodeid < 0 || nodeid >= MAX_NUMNODES); ____cache_alloc_node()
|
H A D | hugetlb.c | 942 if (nid == MAX_NUMNODES) next_node_allowed() 944 VM_BUG_ON(nid >= MAX_NUMNODES); next_node_allowed() 2489 static struct node_hstate node_hstates[MAX_NUMNODES]; 2712 for (i = 0; i < MAX_NUMNODES; ++i) hugetlb_add_hstate()
|
/linux-4.4.14/drivers/acpi/ |
H A D | numa.c | 42 static int node_to_pxm_map[MAX_NUMNODES] 43 = { [0 ... MAX_NUMNODES - 1] = PXM_INVAL }; 79 if (nodes_weight(nodes_found_map) >= MAX_NUMNODES) acpi_map_pxm_to_node()
|
/linux-4.4.14/arch/x86/include/uapi/asm/ |
H A D | e820.h | 13 * kernel was built: MAX_NUMNODES == (1 << CONFIG_NODES_SHIFT),
|
/linux-4.4.14/arch/powerpc/platforms/cell/spufs/ |
H A D | sched.c | 315 for (n = 0; n < MAX_NUMNODES; n++, node++) { aff_ref_location() 327 node = (node < MAX_NUMNODES) ? node : 0; aff_ref_location() 598 for (n = 0; n < MAX_NUMNODES; n++, node++) { spu_get_idle() 599 node = (node < MAX_NUMNODES) ? node : 0; spu_get_idle() 646 for (n = 0; n < MAX_NUMNODES; n++, node++) { find_victim() 647 node = (node < MAX_NUMNODES) ? node : 0; find_victim() 971 for (node = 0; node < MAX_NUMNODES; node++) count_active_contexts() 1014 for (node = 0; node < MAX_NUMNODES; node++) { spusched_thread() 1164 for (node = 0; node < MAX_NUMNODES; node++) { spu_sched_exit()
|
H A D | inode.c | 371 for (node = 0; node < MAX_NUMNODES; node++) { spufs_assert_affinity() 377 if (node == MAX_NUMNODES) { spufs_assert_affinity()
|
/linux-4.4.14/arch/alpha/mm/ |
H A D | numa.c | 22 pg_data_t node_data[MAX_NUMNODES]; 261 for (nid = 0; nid < MAX_NUMNODES; nid++) setup_memory()
|
/linux-4.4.14/arch/sparc/mm/ |
H A D | init_64.c | 96 u64 numa_latency[MAX_NUMNODES][MAX_NUMNODES]; 787 static struct node_mem_mask node_masks[MAX_NUMNODES]; 793 cpumask_t numa_cpumask_lookup_table[MAX_NUMNODES]; 904 struct pglist_data *node_data[MAX_NUMNODES]; 1152 if ((from >= MAX_NUMNODES) || (to >= MAX_NUMNODES)) { __node_distance() 1164 for (i = 0; i < MAX_NUMNODES; i++) { find_best_numa_node_for_mlgroup() 1186 if (tnode == MAX_NUMNODES) mdesc_for_each_arc() 1287 for (i = 0; i < MAX_NUMNODES; i++) { numa_parse_mdesc() 1290 for (j = 0; j < MAX_NUMNODES; j++) { numa_parse_mdesc() 1359 for (i = 0; i < MAX_NUMNODES; i++) { bootmem_init_numa() 1360 for (j = 0; j < MAX_NUMNODES; j++) bootmem_init_numa() 1417 sparse_memory_present_with_active_regions(MAX_NUMNODES); bootmem_init()
|
/linux-4.4.14/arch/powerpc/platforms/cell/ |
H A D | spu_manage.c | 315 if (spu->node >= MAX_NUMNODES) { of_create_spu() 405 for (node = 0; node < MAX_NUMNODES; node++) { init_affinity_qs20_harcoded() 529 for (cbe = 0; cbe < MAX_NUMNODES; cbe++) init_affinity_fw()
|
H A D | spu_base.c | 49 struct cbe_spu_info cbe_spu_info[MAX_NUMNODES]; 768 for (i = 0; i < MAX_NUMNODES; i++) { init_spu_base()
|
/linux-4.4.14/arch/s390/mm/ |
H A D | init.c | 104 sparse_memory_present_with_active_regions(MAX_NUMNODES); paging_init()
|
/linux-4.4.14/arch/sh/kernel/ |
H A D | machine_kexec.c | 148 VMCOREINFO_LENGTH(node_data, MAX_NUMNODES); arch_crash_save_vmcoreinfo()
|
H A D | setup.c | 100 static struct resource mem_resources[MAX_NUMNODES];
|
/linux-4.4.14/arch/m68k/mm/ |
H A D | init.c | 49 pg_data_t pg_data_map[MAX_NUMNODES];
|
/linux-4.4.14/arch/mips/sgi-ip27/ |
H A D | ip27-nmi.c | 216 if (node == MAX_NUMNODES) cont_nmi_dump()
|
/linux-4.4.14/arch/tile/mm/ |
H A D | init.c | 80 static pte_t *l2_ptes[MAX_NUMNODES]; 81 static int num_l2_ptes[MAX_NUMNODES]; 97 BUG_ON(node >= MAX_NUMNODES); get_prealloc_pte() 441 for (i = 0; i < MAX_NUMNODES; ++i) { kernel_physical_mapping_init() 828 for (i = MAX_NUMNODES-1; i >= 0; --i) { mem_init() 841 for (i = MAX_NUMNODES-1; i >= 0; --i) { mem_init()
|
/linux-4.4.14/arch/powerpc/mm/ |
H A D | numa.c | 52 cpumask_var_t node_to_cpumask_map[MAX_NUMNODES]; 53 struct pglist_data *node_data[MAX_NUMNODES]; 66 static int distance_lookup_table[MAX_NUMNODES][MAX_DISTANCE_REF_POINTS]; 79 if (nr_node_ids == MAX_NUMNODES) setup_node_to_cpumask_map() 233 /* Returns nid in the range [0..MAX_NUMNODES-1], or -1 if no useful numa 247 if (nid == 0xffff || nid >= MAX_NUMNODES) associativity_to_nid() 509 if (nid == 0xffff || nid >= MAX_NUMNODES) of_drconf_to_nid_single()
|
/linux-4.4.14/drivers/staging/lustre/lustre/libcfs/linux/ |
H A D | linux-cpu.c | 420 if (node < 0 || node >= MAX_NUMNODES) { cfs_cpt_set_node() 444 if (node < 0 || node >= MAX_NUMNODES) { cfs_cpt_unset_node() 852 high = node ? MAX_NUMNODES - 1 : nr_cpu_ids - 1; cfs_cpt_table_create_pattern()
|
/linux-4.4.14/arch/sparc/kernel/ |
H A D | sysfs.c | 290 for (i = 0; i < MAX_NUMNODES; i++) register_nodes()
|
/linux-4.4.14/arch/tile/include/asm/ |
H A D | pci.h | 159 int mem_maps[MAX_NUMNODES]; pci_iounmap()
|
/linux-4.4.14/arch/x86/kernel/ |
H A D | machine_kexec_32.c | 267 VMCOREINFO_LENGTH(node_data, MAX_NUMNODES); arch_crash_save_vmcoreinfo()
|
H A D | pci-calgary_64.c | 164 static struct scal_detail *scal_devs[MAX_NUMNODES] __initdata; 165 static struct rio_detail *rio_devs[MAX_NUMNODES * 4] __initdata; 1246 if (numnodes > MAX_NUMNODES){ build_detail_arrays() 1248 "Calgary: MAX_NUMNODES too low! Defined as %d, " build_detail_arrays() 1250 MAX_NUMNODES, numnodes); build_detail_arrays()
|
H A D | machine_kexec_64.c | 336 VMCOREINFO_LENGTH(node_data, MAX_NUMNODES); arch_crash_save_vmcoreinfo()
|
H A D | e820.c | 1140 for_each_mem_pfn_range(i, MAX_NUMNODES, &start_pfn, &end_pfn, NULL) { memblock_find_dma_reserve()
|
/linux-4.4.14/arch/powerpc/kernel/ |
H A D | machine_kexec.c | 74 VMCOREINFO_LENGTH(node_data, MAX_NUMNODES); arch_crash_save_vmcoreinfo()
|
H A D | sysfs.c | 987 for (i = 0; i < MAX_NUMNODES; i++) register_nodes()
|
/linux-4.4.14/drivers/base/ |
H A D | node.c | 193 BUILD_BUG_ON(MAX_NUMNODES * 4 > PAGE_SIZE); node_read_distance() 312 struct node *node_devices[MAX_NUMNODES];
|
/linux-4.4.14/arch/powerpc/oprofile/cell/ |
H A D | spu_task_sync.c | 41 struct spu_buffer spu_buff[MAX_NUMNODES * SPUS_PER_NODE]; 144 static struct cached_info *spu_info[MAX_NUMNODES * 8];
|
/linux-4.4.14/arch/ia64/sn/kernel/sn2/ |
H A D | sn_hwperf.c | 206 return node < MAX_NUMNODES && node_online(node) && nr_cpus_node(node); sn_hwperf_has_cpus() 211 return node < MAX_NUMNODES && node_online(node) && NODE_DATA(node)->node_present_pages; sn_hwperf_has_mem()
|
H A D | sn2_smp.c | 172 short nasids[MAX_NUMNODES], nix; sn2_global_tlb_purge()
|
/linux-4.4.14/drivers/char/ |
H A D | mspec.c | 105 static unsigned long scratch_page[MAX_NUMNODES];
|
/linux-4.4.14/drivers/misc/sgi-gru/ |
H A D | grutables.h | 160 #define GRU_MAX_BLADES MAX_NUMNODES
|
/linux-4.4.14/arch/parisc/mm/ |
H A D | init.c | 52 struct node_map_data node_data[MAX_NUMNODES] __read_mostly;
|
/linux-4.4.14/arch/ia64/sn/kernel/ |
H A D | setup.c | 622 for (i=0; i < MAX_NUMNODES; i++) { sn_cpu_init()
|
/linux-4.4.14/arch/powerpc/oprofile/ |
H A D | op_model_cell.c | 141 static unsigned long spu_pm_cnt[MAX_NUMNODES * NUM_SPUS_PER_NODE]; 752 for (i=0; i < MAX_NUMNODES * NUM_SPUS_PER_NODE; i++) cell_reg_setup_spu_events()
|
/linux-4.4.14/arch/alpha/kernel/ |
H A D | setup.c | 83 struct cpumask node_to_cpumask_map[MAX_NUMNODES] __read_mostly;
|
/linux-4.4.14/kernel/ |
H A D | compat.c | 1118 nr_bits = min_t(unsigned long, maxnode - 1, MAX_NUMNODES); COMPAT_SYSCALL_DEFINE4()
|
H A D | cpuset.c | 1840 .max_write_len = (100U + 6 * MAX_NUMNODES), 2593 if (node == MAX_NUMNODES) cpuset_spread_node()
|
/linux-4.4.14/drivers/pci/ |
H A D | pci-sysfs.c | 219 if ((node < 0 && node != NUMA_NO_NODE) || node >= MAX_NUMNODES) numa_node_store()
|
/linux-4.4.14/fs/proc/ |
H A D | task_mmu.c | 1380 unsigned long node[MAX_NUMNODES];
|
/linux-4.4.14/drivers/misc/sgi-xp/ |
H A D | xpc_sn2.c | 563 static u64 xpc_prot_vec_sn2[MAX_NUMNODES];
|
/linux-4.4.14/drivers/block/mtip32xx/ |
H A D | mtip32xx.c | 4307 if (next_node == MAX_NUMNODES) mtip_get_next_rr_node()
|