new_nid          1094 arch/powerpc/mm/numa.c 	int new_nid;
new_nid          1217 arch/powerpc/mm/numa.c 	int new_nid;
new_nid          1223 arch/powerpc/mm/numa.c 	new_nid = associativity_to_nid(associativity);
new_nid          1224 arch/powerpc/mm/numa.c 	if (new_nid < 0 || !node_possible(new_nid))
new_nid          1225 arch/powerpc/mm/numa.c 		new_nid = first_online_node;
new_nid          1227 arch/powerpc/mm/numa.c 	if (NODE_DATA(new_nid) == NULL) {
new_nid          1236 arch/powerpc/mm/numa.c 		if (!topology_inited || try_online_node(new_nid))
new_nid          1237 arch/powerpc/mm/numa.c 			new_nid = first_online_node;
new_nid          1244 arch/powerpc/mm/numa.c 		new_nid = first_online_node;
new_nid          1249 arch/powerpc/mm/numa.c 		cpu, new_nid);
new_nid          1250 arch/powerpc/mm/numa.c 	return new_nid;
new_nid          1269 arch/powerpc/mm/numa.c 		int new_nid = update->new_nid;
new_nid          1274 arch/powerpc/mm/numa.c 		map_cpu_to_node(cpu, new_nid);
new_nid          1275 arch/powerpc/mm/numa.c 		set_cpu_numa_node(cpu, new_nid);
new_nid          1276 arch/powerpc/mm/numa.c 		set_cpu_numa_mem(cpu, local_memory_node(new_nid));
new_nid          1299 arch/powerpc/mm/numa.c 		nid = update->new_nid;
new_nid          1322 arch/powerpc/mm/numa.c 	int weight, new_nid, i = 0;
new_nid          1354 arch/powerpc/mm/numa.c 		new_nid = find_and_online_cpu_nid(cpu);
new_nid          1356 arch/powerpc/mm/numa.c 		if (new_nid == numa_cpu_lookup_table[cpu]) {
new_nid          1361 arch/powerpc/mm/numa.c 					new_nid, cpu);
new_nid          1370 arch/powerpc/mm/numa.c 			ud->new_nid = new_nid;
new_nid          1389 arch/powerpc/mm/numa.c 					  ud->old_nid, ud->new_nid);
new_nid          1427 arch/powerpc/mm/numa.c 		register_cpu_under_node(ud->cpu, ud->new_nid);
new_nid           977 arch/sparc/mm/init_64.c 	int prev_nid, new_nid;
new_nid           981 arch/sparc/mm/init_64.c 		for (new_nid = 0; new_nid < num_node_masks; new_nid++) {
new_nid           982 arch/sparc/mm/init_64.c 			struct node_mem_mask *p = &node_masks[new_nid];
new_nid           986 arch/sparc/mm/init_64.c 					prev_nid = new_nid;
new_nid           991 arch/sparc/mm/init_64.c 		if (new_nid == num_node_masks) {
new_nid           998 arch/sparc/mm/init_64.c 		if (prev_nid != new_nid)
new_nid           421 fs/f2fs/xattr.c 	nid_t new_nid = 0;
new_nid           425 fs/f2fs/xattr.c 		if (!f2fs_alloc_nid(sbi, &new_nid))
new_nid           435 fs/f2fs/xattr.c 				f2fs_alloc_nid_failed(sbi, new_nid);
new_nid           446 fs/f2fs/xattr.c 			f2fs_alloc_nid_failed(sbi, new_nid);
new_nid           462 fs/f2fs/xattr.c 			f2fs_alloc_nid_failed(sbi, new_nid);
new_nid           465 fs/f2fs/xattr.c 		f2fs_bug_on(sbi, new_nid);
new_nid           469 fs/f2fs/xattr.c 		set_new_dnode(&dn, inode, NULL, NULL, new_nid);
new_nid           473 fs/f2fs/xattr.c 			f2fs_alloc_nid_failed(sbi, new_nid);
new_nid           476 fs/f2fs/xattr.c 		f2fs_alloc_nid_done(sbi, new_nid);
new_nid          5171 mm/hugetlb.c   		int new_nid = page_to_nid(newpage);
new_nid          5179 mm/hugetlb.c   			h->surplus_huge_pages_node[new_nid]++;