nodes_weight 157 arch/x86/mm/amdtopology.c if (!nodes_weight(numa_nodes_parsed)) nodes_weight 125 arch/x86/mm/numa_emulation.c while (nodes_weight(physnode_mask)) { nodes_weight 272 arch/x86/mm/numa_emulation.c while (nodes_weight(physnode_mask)) { nodes_weight 64 drivers/acpi/numa.c if (nodes_weight(nodes_found_map) >= MAX_NUMNODES) nodes_weight 430 include/linux/nodemask.h return nodes_weight(node_states[state]); nodes_weight 25 lib/nodemask.c w = nodes_weight(*maskp); nodes_weight 1038 mm/hugetlb.c for (nr_nodes = nodes_weight(*mask); \ nodes_weight 1044 mm/hugetlb.c for (nr_nodes = nodes_weight(*mask); \ nodes_weight 163 mm/mempolicy.c nodes_fold(tmp, *orig, nodes_weight(*rel)); nodes_weight 1126 mm/mempolicy.c if ((nodes_weight(*from) != nodes_weight(*to)) && nodes_weight 1890 mm/mempolicy.c unsigned nnodes = nodes_weight(pol->v.nodes); nodes_weight 7113 mm/page_alloc.c int usable_nodes = nodes_weight(node_states[N_MEMORY]);