/linux-4.1.27/include/linux/ |
D | nodemask.h | 382 N_MEMORY, /* The node has memory(regular, high, movable) */ enumerator 384 N_MEMORY = N_HIGH_MEMORY, 422 #define first_memory_node first_node(node_states[N_MEMORY]) 429 return next_node(nid, node_states[N_MEMORY]); in next_memory_node()
|
D | cpuset.h | 161 #define cpuset_current_mems_allowed (node_states[N_MEMORY])
|
/linux-4.1.27/mm/ |
D | memory_hotplug.c | 903 if (N_MEMORY == N_NORMAL_MEMORY) in node_states_check_changes_online() 928 if (N_MEMORY == N_HIGH_MEMORY) in node_states_check_changes_online() 944 if (!node_state(nid, N_MEMORY)) in node_states_check_changes_online() 958 node_set_state(node, N_MEMORY); in node_states_set_node() 1594 if (N_MEMORY == N_NORMAL_MEMORY) in node_states_check_changes_offline() 1622 if (N_MEMORY == N_HIGH_MEMORY) in node_states_check_changes_offline() 1659 if ((N_MEMORY != N_NORMAL_MEMORY) && in node_states_clear_node() 1663 if ((N_MEMORY != N_HIGH_MEMORY) && in node_states_clear_node() 1665 node_clear_state(node, N_MEMORY); in node_states_clear_node()
|
D | hugetlb.c | 1418 if (!free_pool_huge_page(h, &node_states[N_MEMORY], 1)) in return_unused_surplus_pages() 1544 for_each_node_mask_to_alloc(h, nr_nodes, node, &node_states[N_MEMORY]) { in alloc_bootmem_huge_page() 1619 &node_states[N_MEMORY])) in hugetlb_hstate_alloc_pages() 1866 nodes_allowed = &node_states[N_MEMORY]; in __nr_hugepages_store_common() 1876 nodes_allowed = &node_states[N_MEMORY]; in __nr_hugepages_store_common() 1880 if (nodes_allowed != &node_states[N_MEMORY]) in __nr_hugepages_store_common() 2198 for_each_node_state(nid, N_MEMORY) { in hugetlb_register_all_nodes() 2301 h->next_nid_to_alloc = first_node(node_states[N_MEMORY]); in hugetlb_add_hstate() 2302 h->next_nid_to_free = first_node(node_states[N_MEMORY]); in hugetlb_add_hstate() 2479 for_each_node_state(nid, N_MEMORY) in hugetlb_show_meminfo()
|
D | page_ext.c | 363 for_each_node_state(nid, N_MEMORY) { in page_ext_init()
|
D | page_alloc.c | 103 [N_MEMORY] = { { [0] = 1UL } }, 1970 &node_states[N_MEMORY]; in zlc_setup() 3613 for_each_node_state(n, N_MEMORY) { in find_next_best_node() 5213 node_set_state(nid, N_MEMORY); in early_calculate_totalpages() 5230 nodemask_t saved_node_state = node_states[N_MEMORY]; in find_zone_movable_pfns_for_nodes() 5232 int usable_nodes = nodes_weight(node_states[N_MEMORY]); in find_zone_movable_pfns_for_nodes() 5290 for_each_node_state(nid, N_MEMORY) { in find_zone_movable_pfns_for_nodes() 5383 node_states[N_MEMORY] = saved_node_state; in find_zone_movable_pfns_for_nodes() 5391 if (N_MEMORY == N_NORMAL_MEMORY) in check_for_memory() 5488 node_set_state(nid, N_MEMORY); in free_area_init_nodes()
|
D | mempolicy.c | 224 cpuset_current_mems_allowed, node_states[N_MEMORY]); in mpol_set_nodemask() 1404 if (!nodes_subset(*new, node_states[N_MEMORY])) { in SYSCALL_DEFINE4() 2592 for_each_node_state(nid, N_MEMORY) { in numa_policy_init() 2665 if (!nodes_subset(nodes, node_states[N_MEMORY])) in mpol_parse_str() 2699 nodes = node_states[N_MEMORY]; in mpol_parse_str()
|
D | vmstat.c | 1132 if (!node_state(pgdat->node_id, N_MEMORY)) in pagetypeinfo_show() 1597 if (!node_state(pgdat->node_id, N_MEMORY)) in unusable_show()
|
D | oom_kill.c | 226 if (nodemask && !nodes_subset(node_states[N_MEMORY], *nodemask)) { in constrained_alloc()
|
D | memcontrol.c | 895 for_each_node_state(nid, N_MEMORY) in mem_cgroup_nr_lru_pages() 1639 memcg->scan_nodes = node_states[N_MEMORY]; in mem_cgroup_may_update_nodemask() 1641 for_each_node_mask(nid, node_states[N_MEMORY]) { in mem_cgroup_may_update_nodemask() 3423 for_each_node_state(nid, N_MEMORY) { in memcg_numa_stat_show() 3438 for_each_node_state(nid, N_MEMORY) { in memcg_numa_stat_show()
|
D | vmscan.c | 3519 for_each_node_state(nid, N_MEMORY) { in cpu_callback() 3575 for_each_node_state(nid, N_MEMORY) in kswapd_init()
|
D | migrate.c | 1344 if (!node_state(node, N_MEMORY)) in do_pages_move()
|
/linux-4.1.27/drivers/base/ |
D | node.c | 230 node_state(node->dev.id, N_MEMORY)) { in hugetlb_register_node() 646 [N_MEMORY] = _NODE_ATTR(has_memory, N_MEMORY), 659 &node_state_attr[N_MEMORY].attr.attr,
|
/linux-4.1.27/kernel/ |
D | cpuset.c | 354 while (!nodes_intersects(cs->effective_mems, node_states[N_MEMORY])) in guarantee_online_mems() 356 nodes_and(*pmask, cs->effective_mems, node_states[N_MEMORY]); in guarantee_online_mems() 2250 new_mems = node_states[N_MEMORY]; in cpuset_hotplug_workfn() 2342 top_cpuset.mems_allowed = node_states[N_MEMORY]; in cpuset_init_smp() 2346 top_cpuset.effective_mems = node_states[N_MEMORY]; in cpuset_init_smp()
|
D | kthread.c | 491 set_mems_allowed(node_states[N_MEMORY]); in kthreadd()
|
/linux-4.1.27/init/ |
D | main.c | 986 set_mems_allowed(node_states[N_MEMORY]); in kernel_init_freeable()
|
/linux-4.1.27/arch/x86/mm/ |
D | init_64.c | 660 node_clear_state(0, N_MEMORY); in paging_init() 661 if (N_MEMORY != N_NORMAL_MEMORY) in paging_init()
|
/linux-4.1.27/fs/proc/ |
D | kcore.c | 256 for_each_node_state(nid, N_MEMORY) { in kcore_update_ram()
|
D | task_mmu.c | 1406 if (!node_isset(nid, node_states[N_MEMORY])) in can_gather_numa_stats() 1560 for_each_node_state(nid, N_MEMORY) in show_numa_map()
|
/linux-4.1.27/Documentation/ |
D | memory-hotplug.txt | 422 status_change_nid is set node id when N_MEMORY of nodemask is (will be)
|
/linux-4.1.27/Documentation/cgroups/ |
D | cpusets.txt | 221 automatically tracks the value of node_states[N_MEMORY]--i.e.,
|