Lines Matching refs:nodes_allowed
939 static int next_node_allowed(int nid, nodemask_t *nodes_allowed) in next_node_allowed() argument
941 nid = next_node(nid, *nodes_allowed); in next_node_allowed()
943 nid = first_node(*nodes_allowed); in next_node_allowed()
949 static int get_valid_node_allowed(int nid, nodemask_t *nodes_allowed) in get_valid_node_allowed() argument
951 if (!node_isset(nid, *nodes_allowed)) in get_valid_node_allowed()
952 nid = next_node_allowed(nid, nodes_allowed); in get_valid_node_allowed()
963 nodemask_t *nodes_allowed) in hstate_next_node_to_alloc() argument
967 VM_BUG_ON(!nodes_allowed); in hstate_next_node_to_alloc()
969 nid = get_valid_node_allowed(h->next_nid_to_alloc, nodes_allowed); in hstate_next_node_to_alloc()
970 h->next_nid_to_alloc = next_node_allowed(nid, nodes_allowed); in hstate_next_node_to_alloc()
981 static int hstate_next_node_to_free(struct hstate *h, nodemask_t *nodes_allowed) in hstate_next_node_to_free() argument
985 VM_BUG_ON(!nodes_allowed); in hstate_next_node_to_free()
987 nid = get_valid_node_allowed(h->next_nid_to_free, nodes_allowed); in hstate_next_node_to_free()
988 h->next_nid_to_free = next_node_allowed(nid, nodes_allowed); in hstate_next_node_to_free()
1118 nodemask_t *nodes_allowed) in alloc_fresh_gigantic_page() argument
1123 for_each_node_mask_to_alloc(h, nr_nodes, node, nodes_allowed) { in alloc_fresh_gigantic_page()
1139 nodemask_t *nodes_allowed) { return 0; } in alloc_fresh_gigantic_page() argument
1351 static int alloc_fresh_huge_page(struct hstate *h, nodemask_t *nodes_allowed) in alloc_fresh_huge_page() argument
1357 for_each_node_mask_to_alloc(h, nr_nodes, node, nodes_allowed) { in alloc_fresh_huge_page()
1379 static int free_pool_huge_page(struct hstate *h, nodemask_t *nodes_allowed, in free_pool_huge_page() argument
1385 for_each_node_mask_to_free(h, nr_nodes, node, nodes_allowed) { in free_pool_huge_page()
2079 nodemask_t *nodes_allowed) in try_to_free_low() argument
2086 for_each_node_mask(i, *nodes_allowed) { in try_to_free_low()
2103 nodemask_t *nodes_allowed) in try_to_free_low() argument
2113 static int adjust_pool_surplus(struct hstate *h, nodemask_t *nodes_allowed, in adjust_pool_surplus() argument
2121 for_each_node_mask_to_alloc(h, nr_nodes, node, nodes_allowed) { in adjust_pool_surplus()
2126 for_each_node_mask_to_free(h, nr_nodes, node, nodes_allowed) { in adjust_pool_surplus()
2142 nodemask_t *nodes_allowed) in set_max_huge_pages() argument
2162 if (!adjust_pool_surplus(h, nodes_allowed, -1)) in set_max_huge_pages()
2174 ret = alloc_fresh_gigantic_page(h, nodes_allowed); in set_max_huge_pages()
2176 ret = alloc_fresh_huge_page(h, nodes_allowed); in set_max_huge_pages()
2203 try_to_free_low(h, min_count, nodes_allowed); in set_max_huge_pages()
2205 if (!free_pool_huge_page(h, nodes_allowed, 0)) in set_max_huge_pages()
2210 if (!adjust_pool_surplus(h, nodes_allowed, 1)) in set_max_huge_pages()
2266 NODEMASK_ALLOC(nodemask_t, nodes_allowed, GFP_KERNEL | __GFP_NORETRY); in __nr_hugepages_store_common()
2278 init_nodemask_of_mempolicy(nodes_allowed))) { in __nr_hugepages_store_common()
2279 NODEMASK_FREE(nodes_allowed); in __nr_hugepages_store_common()
2280 nodes_allowed = &node_states[N_MEMORY]; in __nr_hugepages_store_common()
2282 } else if (nodes_allowed) { in __nr_hugepages_store_common()
2288 init_nodemask_of_node(nodes_allowed, nid); in __nr_hugepages_store_common()
2290 nodes_allowed = &node_states[N_MEMORY]; in __nr_hugepages_store_common()
2292 h->max_huge_pages = set_max_huge_pages(h, count, nodes_allowed); in __nr_hugepages_store_common()
2294 if (nodes_allowed != &node_states[N_MEMORY]) in __nr_hugepages_store_common()
2295 NODEMASK_FREE(nodes_allowed); in __nr_hugepages_store_common()
2299 NODEMASK_FREE(nodes_allowed); in __nr_hugepages_store_common()