nr_nodes          195 arch/mips/include/asm/mach-loongson64/boot_param.h 	u32 nr_nodes;
nr_nodes          147 arch/mips/loongson64/common/env.c 	loongson_sysconf.nr_nodes = (loongson_sysconf.nr_cpus +
nr_nodes           58 arch/mips/loongson64/loongson-3/numa.c 	for (i = 0; i < loongson_sysconf.nr_nodes; i++) {
nr_nodes          213 arch/mips/loongson64/loongson-3/numa.c 	for (node = 0; node < loongson_sysconf.nr_nodes; node++) {
nr_nodes           84 arch/x86/mm/numa_emulation.c 					 u64 addr, u64 max_addr, int nr_nodes)
nr_nodes           92 arch/x86/mm/numa_emulation.c 	if (nr_nodes <= 0)
nr_nodes           94 arch/x86/mm/numa_emulation.c 	if (nr_nodes > MAX_NUMNODES) {
nr_nodes           96 arch/x86/mm/numa_emulation.c 			nr_nodes, MAX_NUMNODES);
nr_nodes           97 arch/x86/mm/numa_emulation.c 		nr_nodes = MAX_NUMNODES;
nr_nodes          105 arch/x86/mm/numa_emulation.c 	size = PFN_PHYS((unsigned long)(size >> PAGE_SHIFT) / nr_nodes);
nr_nodes          111 arch/x86/mm/numa_emulation.c 	big = ((size & ~FAKE_NODE_MIN_HASH_MASK) * nr_nodes) /
nr_nodes          172 arch/x86/mm/numa_emulation.c 			ret = emu_setup_memblk(ei, pi, nid++ % nr_nodes,
nr_nodes          200 arch/x86/mm/numa_emulation.c static u64 uniform_size(u64 max_addr, u64 base, u64 hole, int nr_nodes)
nr_nodes          206 arch/x86/mm/numa_emulation.c 	return PFN_PHYS((max_pfn - base_pfn - hole_pfns) / nr_nodes);
nr_nodes          218 arch/x86/mm/numa_emulation.c 					      int nr_nodes, struct numa_memblk *pblk,
nr_nodes          225 arch/x86/mm/numa_emulation.c 	if ((!size && !nr_nodes) || (nr_nodes && !pblk))
nr_nodes          238 arch/x86/mm/numa_emulation.c 	if (!nr_nodes)
nr_nodes          239 arch/x86/mm/numa_emulation.c 		nr_nodes = MAX_NUMNODES;
nr_nodes          247 arch/x86/mm/numa_emulation.c 		min_size = uniform_size(max_addr, addr, 0, nr_nodes);
nr_nodes          258 arch/x86/mm/numa_emulation.c 				mem_hole_size(addr, max_addr), nr_nodes);
nr_nodes          110 fs/btrfs/relocation.c 	int nr_nodes;
nr_nodes          235 fs/btrfs/relocation.c 	ASSERT(!cache->nr_nodes);
nr_nodes          249 fs/btrfs/relocation.c 		cache->nr_nodes++;
nr_nodes          258 fs/btrfs/relocation.c 		cache->nr_nodes--;
nr_nodes          171 lib/rbtree_test.c static void check_postorder_foreach(int nr_nodes)
nr_nodes          178 lib/rbtree_test.c 	WARN_ON_ONCE(count != nr_nodes);
nr_nodes          181 lib/rbtree_test.c static void check_postorder(int nr_nodes)
nr_nodes          188 lib/rbtree_test.c 	WARN_ON_ONCE(count != nr_nodes);
nr_nodes          191 lib/rbtree_test.c static void check(int nr_nodes)
nr_nodes          211 lib/rbtree_test.c 	WARN_ON_ONCE(count != nr_nodes);
nr_nodes          214 lib/rbtree_test.c 	check_postorder(nr_nodes);
nr_nodes          215 lib/rbtree_test.c 	check_postorder_foreach(nr_nodes);
nr_nodes          218 lib/rbtree_test.c static void check_augmented(int nr_nodes)
nr_nodes          222 lib/rbtree_test.c 	check(nr_nodes);
nr_nodes         1037 mm/hugetlb.c   #define for_each_node_mask_to_alloc(hs, nr_nodes, node, mask)		\
nr_nodes         1038 mm/hugetlb.c   	for (nr_nodes = nodes_weight(*mask);				\
nr_nodes         1039 mm/hugetlb.c   		nr_nodes > 0 &&						\
nr_nodes         1041 mm/hugetlb.c   		nr_nodes--)
nr_nodes         1043 mm/hugetlb.c   #define for_each_node_mask_to_free(hs, nr_nodes, node, mask)		\
nr_nodes         1044 mm/hugetlb.c   	for (nr_nodes = nodes_weight(*mask);				\
nr_nodes         1045 mm/hugetlb.c   		nr_nodes > 0 &&						\
nr_nodes         1047 mm/hugetlb.c   		nr_nodes--)
nr_nodes         1535 mm/hugetlb.c   	int nr_nodes, node;
nr_nodes         1538 mm/hugetlb.c   	for_each_node_mask_to_alloc(h, nr_nodes, node, nodes_allowed) {
nr_nodes         1562 mm/hugetlb.c   	int nr_nodes, node;
nr_nodes         1565 mm/hugetlb.c   	for_each_node_mask_to_free(h, nr_nodes, node, nodes_allowed) {
nr_nodes         2222 mm/hugetlb.c   	int nr_nodes, node;
nr_nodes         2224 mm/hugetlb.c   	for_each_node_mask_to_alloc(h, nr_nodes, node, &node_states[N_MEMORY]) {
nr_nodes         2399 mm/hugetlb.c   	int nr_nodes, node;
nr_nodes         2404 mm/hugetlb.c   		for_each_node_mask_to_alloc(h, nr_nodes, node, nodes_allowed) {
nr_nodes         2409 mm/hugetlb.c   		for_each_node_mask_to_free(h, nr_nodes, node, nodes_allowed) {
nr_nodes         5588 mm/page_alloc.c 		unsigned nr_nodes)
nr_nodes         5595 mm/page_alloc.c 	for (i = 0; i < nr_nodes; i++) {
nr_nodes         5632 mm/page_alloc.c 	int node, load, nr_nodes = 0;
nr_nodes         5653 mm/page_alloc.c 		node_order[nr_nodes++] = node;
nr_nodes         5658 mm/page_alloc.c 	build_zonelists_in_node_order(pgdat, node_order, nr_nodes);
nr_nodes          126 tools/perf/bench/numa.c 	int			nr_nodes;
nr_nodes          227 tools/perf/bench/numa.c 	int i, nr_nodes = 0;
nr_nodes          229 tools/perf/bench/numa.c 	for (i = 0; i < g->p.nr_nodes; i++) {
nr_nodes          231 tools/perf/bench/numa.c 			nr_nodes++;
nr_nodes          234 tools/perf/bench/numa.c 	return nr_nodes;
nr_nodes          335 tools/perf/bench/numa.c 	ret = set_mempolicy(MPOL_DEFAULT, NULL, g->p.nr_nodes-1);
nr_nodes          348 tools/perf/bench/numa.c 	BUG_ON(g->p.nr_nodes > (int)sizeof(nodemask)*8);
nr_nodes          673 tools/perf/bench/numa.c 			BUG_ON(step <= 0 || step >= g->p.nr_nodes);
nr_nodes          686 tools/perf/bench/numa.c 		if (bind_node_0 >= g->p.nr_nodes || bind_node_1 >= g->p.nr_nodes) {
nr_nodes          687 tools/perf/bench/numa.c 			printf("\nTest not applicable, system has only %d nodes.\n", g->p.nr_nodes);
nr_nodes          983 tools/perf/bench/numa.c 	for (node = 0; node < g->p.nr_nodes; node++)
nr_nodes         1012 tools/perf/bench/numa.c 	for (node = 0; node < g->p.nr_nodes; node++) {
nr_nodes         1034 tools/perf/bench/numa.c 	for (node = 0; node < g->p.nr_nodes; node++) {
nr_nodes         1396 tools/perf/bench/numa.c 	g->p.nr_nodes = numa_max_node() + 1;
nr_nodes         1399 tools/perf/bench/numa.c 	BUG_ON(g->p.nr_nodes > MAX_NR_NODES || g->p.nr_nodes < 0);