nr_nodes 195 arch/mips/include/asm/mach-loongson64/boot_param.h u32 nr_nodes; nr_nodes 147 arch/mips/loongson64/common/env.c loongson_sysconf.nr_nodes = (loongson_sysconf.nr_cpus + nr_nodes 58 arch/mips/loongson64/loongson-3/numa.c for (i = 0; i < loongson_sysconf.nr_nodes; i++) { nr_nodes 213 arch/mips/loongson64/loongson-3/numa.c for (node = 0; node < loongson_sysconf.nr_nodes; node++) { nr_nodes 84 arch/x86/mm/numa_emulation.c u64 addr, u64 max_addr, int nr_nodes) nr_nodes 92 arch/x86/mm/numa_emulation.c if (nr_nodes <= 0) nr_nodes 94 arch/x86/mm/numa_emulation.c if (nr_nodes > MAX_NUMNODES) { nr_nodes 96 arch/x86/mm/numa_emulation.c nr_nodes, MAX_NUMNODES); nr_nodes 97 arch/x86/mm/numa_emulation.c nr_nodes = MAX_NUMNODES; nr_nodes 105 arch/x86/mm/numa_emulation.c size = PFN_PHYS((unsigned long)(size >> PAGE_SHIFT) / nr_nodes); nr_nodes 111 arch/x86/mm/numa_emulation.c big = ((size & ~FAKE_NODE_MIN_HASH_MASK) * nr_nodes) / nr_nodes 172 arch/x86/mm/numa_emulation.c ret = emu_setup_memblk(ei, pi, nid++ % nr_nodes, nr_nodes 200 arch/x86/mm/numa_emulation.c static u64 uniform_size(u64 max_addr, u64 base, u64 hole, int nr_nodes) nr_nodes 206 arch/x86/mm/numa_emulation.c return PFN_PHYS((max_pfn - base_pfn - hole_pfns) / nr_nodes); nr_nodes 218 arch/x86/mm/numa_emulation.c int nr_nodes, struct numa_memblk *pblk, nr_nodes 225 arch/x86/mm/numa_emulation.c if ((!size && !nr_nodes) || (nr_nodes && !pblk)) nr_nodes 238 arch/x86/mm/numa_emulation.c if (!nr_nodes) nr_nodes 239 arch/x86/mm/numa_emulation.c nr_nodes = MAX_NUMNODES; nr_nodes 247 arch/x86/mm/numa_emulation.c min_size = uniform_size(max_addr, addr, 0, nr_nodes); nr_nodes 258 arch/x86/mm/numa_emulation.c mem_hole_size(addr, max_addr), nr_nodes); nr_nodes 110 fs/btrfs/relocation.c int nr_nodes; nr_nodes 235 fs/btrfs/relocation.c ASSERT(!cache->nr_nodes); nr_nodes 249 fs/btrfs/relocation.c cache->nr_nodes++; nr_nodes 258 fs/btrfs/relocation.c cache->nr_nodes--; nr_nodes 171 lib/rbtree_test.c static void check_postorder_foreach(int nr_nodes) nr_nodes 178 lib/rbtree_test.c WARN_ON_ONCE(count != nr_nodes); nr_nodes 181 lib/rbtree_test.c static void check_postorder(int nr_nodes) nr_nodes 188 lib/rbtree_test.c WARN_ON_ONCE(count != nr_nodes); nr_nodes 191 lib/rbtree_test.c static void check(int nr_nodes) nr_nodes 211 lib/rbtree_test.c WARN_ON_ONCE(count != nr_nodes); nr_nodes 214 lib/rbtree_test.c check_postorder(nr_nodes); nr_nodes 215 lib/rbtree_test.c check_postorder_foreach(nr_nodes); nr_nodes 218 lib/rbtree_test.c static void check_augmented(int nr_nodes) nr_nodes 222 lib/rbtree_test.c check(nr_nodes); nr_nodes 1037 mm/hugetlb.c #define for_each_node_mask_to_alloc(hs, nr_nodes, node, mask) \ nr_nodes 1038 mm/hugetlb.c for (nr_nodes = nodes_weight(*mask); \ nr_nodes 1039 mm/hugetlb.c nr_nodes > 0 && \ nr_nodes 1041 mm/hugetlb.c nr_nodes--) nr_nodes 1043 mm/hugetlb.c #define for_each_node_mask_to_free(hs, nr_nodes, node, mask) \ nr_nodes 1044 mm/hugetlb.c for (nr_nodes = nodes_weight(*mask); \ nr_nodes 1045 mm/hugetlb.c nr_nodes > 0 && \ nr_nodes 1047 mm/hugetlb.c nr_nodes--) nr_nodes 1535 mm/hugetlb.c int nr_nodes, node; nr_nodes 1538 mm/hugetlb.c for_each_node_mask_to_alloc(h, nr_nodes, node, nodes_allowed) { nr_nodes 1562 mm/hugetlb.c int nr_nodes, node; nr_nodes 1565 mm/hugetlb.c for_each_node_mask_to_free(h, nr_nodes, node, nodes_allowed) { nr_nodes 2222 mm/hugetlb.c int nr_nodes, node; nr_nodes 2224 mm/hugetlb.c for_each_node_mask_to_alloc(h, nr_nodes, node, &node_states[N_MEMORY]) { nr_nodes 2399 mm/hugetlb.c int nr_nodes, node; nr_nodes 2404 mm/hugetlb.c for_each_node_mask_to_alloc(h, nr_nodes, node, nodes_allowed) { nr_nodes 2409 mm/hugetlb.c for_each_node_mask_to_free(h, nr_nodes, node, nodes_allowed) { nr_nodes 5588 mm/page_alloc.c unsigned nr_nodes) nr_nodes 5595 mm/page_alloc.c for (i = 0; i < nr_nodes; i++) { nr_nodes 5632 mm/page_alloc.c int node, load, nr_nodes = 0; nr_nodes 5653 mm/page_alloc.c node_order[nr_nodes++] = node; nr_nodes 5658 mm/page_alloc.c build_zonelists_in_node_order(pgdat, node_order, nr_nodes); nr_nodes 126 tools/perf/bench/numa.c int nr_nodes; nr_nodes 227 tools/perf/bench/numa.c int i, nr_nodes = 0; nr_nodes 229 tools/perf/bench/numa.c for (i = 0; i < g->p.nr_nodes; i++) { nr_nodes 231 tools/perf/bench/numa.c nr_nodes++; nr_nodes 234 tools/perf/bench/numa.c return nr_nodes; nr_nodes 335 tools/perf/bench/numa.c ret = set_mempolicy(MPOL_DEFAULT, NULL, g->p.nr_nodes-1); nr_nodes 348 tools/perf/bench/numa.c BUG_ON(g->p.nr_nodes > (int)sizeof(nodemask)*8); nr_nodes 673 tools/perf/bench/numa.c BUG_ON(step <= 0 || step >= g->p.nr_nodes); nr_nodes 686 tools/perf/bench/numa.c if (bind_node_0 >= g->p.nr_nodes || bind_node_1 >= g->p.nr_nodes) { nr_nodes 687 tools/perf/bench/numa.c printf("\nTest not applicable, system has only %d nodes.\n", g->p.nr_nodes); nr_nodes 983 tools/perf/bench/numa.c for (node = 0; node < g->p.nr_nodes; node++) nr_nodes 1012 tools/perf/bench/numa.c for (node = 0; node < g->p.nr_nodes; node++) { nr_nodes 1034 tools/perf/bench/numa.c for (node = 0; node < g->p.nr_nodes; node++) { nr_nodes 1396 tools/perf/bench/numa.c g->p.nr_nodes = numa_max_node() + 1; nr_nodes 1399 tools/perf/bench/numa.c BUG_ON(g->p.nr_nodes > MAX_NR_NODES || g->p.nr_nodes < 0);