nodes_weight      157 arch/x86/mm/amdtopology.c 	if (!nodes_weight(numa_nodes_parsed))
nodes_weight      125 arch/x86/mm/numa_emulation.c 	while (nodes_weight(physnode_mask)) {
nodes_weight      272 arch/x86/mm/numa_emulation.c 	while (nodes_weight(physnode_mask)) {
nodes_weight       64 drivers/acpi/numa.c 		if (nodes_weight(nodes_found_map) >= MAX_NUMNODES)
nodes_weight      430 include/linux/nodemask.h 	return nodes_weight(node_states[state]);
nodes_weight       25 lib/nodemask.c 	w = nodes_weight(*maskp);
nodes_weight     1038 mm/hugetlb.c   	for (nr_nodes = nodes_weight(*mask);				\
nodes_weight     1044 mm/hugetlb.c   	for (nr_nodes = nodes_weight(*mask);				\
nodes_weight      163 mm/mempolicy.c 	nodes_fold(tmp, *orig, nodes_weight(*rel));
nodes_weight     1126 mm/mempolicy.c 			if ((nodes_weight(*from) != nodes_weight(*to)) &&
nodes_weight     1890 mm/mempolicy.c 	unsigned nnodes = nodes_weight(pol->v.nodes);
nodes_weight     7113 mm/page_alloc.c 	int usable_nodes = nodes_weight(node_states[N_MEMORY]);