/linux-4.1.27/mm/ |
D | oom_kill.c | 122 struct mem_cgroup *memcg, const nodemask_t *nodemask) in oom_unkillable_task() argument 134 if (!has_intersects_mems_allowed(p, nodemask)) in oom_unkillable_task() 150 const nodemask_t *nodemask, unsigned long totalpages) in oom_badness() argument 155 if (oom_unkillable_task(p, memcg, nodemask)) in oom_badness() 199 gfp_t gfp_mask, nodemask_t *nodemask, in constrained_alloc() argument 226 if (nodemask && !nodes_subset(node_states[N_MEMORY], *nodemask)) { in constrained_alloc() 228 for_each_node_mask(nid, *nodemask) in constrained_alloc() 235 high_zoneidx, nodemask) in constrained_alloc() 249 gfp_t gfp_mask, nodemask_t *nodemask, in constrained_alloc() argument 258 unsigned long totalpages, const nodemask_t *nodemask, in oom_scan_process_thread() argument [all …]
|
D | vmscan.c | 75 nodemask_t *nodemask; member 2473 gfp_zone(sc->gfp_mask), sc->nodemask) { in shrink_zones() 2678 nodemask_t *nodemask) in throttle_direct_reclaim() argument 2716 gfp_zone(gfp_mask), nodemask) { in throttle_direct_reclaim() 2762 gfp_t gfp_mask, nodemask_t *nodemask) in try_to_free_pages() argument 2769 .nodemask = nodemask, in try_to_free_pages() 2781 if (throttle_direct_reclaim(gfp_mask, zonelist, nodemask)) in try_to_free_pages()
|
D | internal.h | 127 nodemask_t *nodemask; member
|
D | page_alloc.c | 2132 ac->nodemask) { in get_page_from_freelist() 2448 if (out_of_memory(ac->zonelist, gfp_mask, order, ac->nodemask, false) in __alloc_pages_may_oom() 2542 ac->nodemask); in __perform_reclaim() 2615 ac->high_zoneidx, ac->nodemask) in wake_all_kswapds() 2720 if (!(alloc_flags & ALLOC_CPUSET) && !ac->nodemask) { in __alloc_pages_slowpath() 2870 struct zonelist *zonelist, nodemask_t *nodemask) in __alloc_pages_nodemask() argument 2879 .nodemask = nodemask, in __alloc_pages_nodemask() 2910 ac.nodemask ? : &cpuset_current_mems_allowed, in __alloc_pages_nodemask()
|
D | mempolicy.c | 1811 nodemask_t **nodemask) in huge_zonelist() argument 1816 *nodemask = NULL; /* assume !MPOL_BIND */ in huge_zonelist() 1824 *nodemask = &(*mpol)->v.nodes; in huge_zonelist()
|
D | hugetlb.c | 633 nodemask_t *nodemask; in dequeue_huge_page_vma() local 655 htlb_alloc_mask(h), &mpol, &nodemask); in dequeue_huge_page_vma() 658 MAX_NR_ZONES - 1, nodemask) { in dequeue_huge_page_vma()
|
D | compaction.c | 1500 ac->nodemask) { in try_to_compact_pages()
|
/linux-4.1.27/include/linux/ |
D | cpuset.h | 49 int cpuset_nodemask_valid_mems_allowed(nodemask_t *nodemask); 121 static inline void set_mems_allowed(nodemask_t nodemask) in set_mems_allowed() argument 128 current->mems_allowed = nodemask; in set_mems_allowed() 164 static inline int cpuset_nodemask_valid_mems_allowed(nodemask_t *nodemask) in cpuset_nodemask_valid_mems_allowed() argument 226 static inline void set_mems_allowed(nodemask_t nodemask) in set_mems_allowed() argument
|
D | oom.h | 55 struct mem_cgroup *memcg, const nodemask_t *nodemask, 62 struct mem_cgroup *memcg, nodemask_t *nodemask, 69 int order, const nodemask_t *nodemask, 73 unsigned long totalpages, const nodemask_t *nodemask,
|
D | nodemask.h | 140 #define node_isset(node, nodemask) test_bit((node), (nodemask).bits) argument 142 #define node_test_and_set(node, nodemask) \ argument 143 __node_test_and_set((node), &(nodemask)) 219 #define nodes_full(nodemask) __nodes_full(&(nodemask), MAX_NUMNODES) argument 225 #define nodes_weight(nodemask) __nodes_weight(&(nodemask), MAX_NUMNODES) argument
|
D | mempolicy.h | 150 struct mempolicy **mpol, nodemask_t **nodemask); 259 struct mempolicy **mpol, nodemask_t **nodemask) in huge_zonelist() argument 262 *nodemask = NULL; in huge_zonelist()
|
D | mmzone.h | 1017 #define for_each_zone_zonelist_nodemask(zone, z, zlist, highidx, nodemask) \ argument 1018 for (z = first_zones_zonelist(zlist, highidx, nodemask, &zone); \ 1020 z = next_zones_zonelist(++z, highidx, nodemask), \
|
D | gfp.h | 294 struct zonelist *zonelist, nodemask_t *nodemask);
|
/linux-4.1.27/Documentation/vm/ |
D | numa_memory_policy.txt | 186 always preferred by passing an empty nodemask with this mode. 187 If an empty nodemask is passed, the policy cannot use the 217 MPOL_F_STATIC_NODES: This flag specifies that the nodemask passed by 223 nodemask (Bind, Interleave) is remapped to the new set of 236 3 is allowed from the user's nodemask, the "interleave" only 237 occurs over that node. If no nodes from the user's nodemask are 242 MPOL_PREFERRED policies that were created with an empty nodemask 245 MPOL_F_RELATIVE_NODES: This flag specifies that the nodemask passed 247 set of allowed nodes. The kernel stores the user-passed nodemask, 248 and if the allowed nodes changes, then that original nodemask will [all …]
|
/linux-4.1.27/kernel/ |
D | smp.c | 365 const struct cpumask *nodemask; in smp_call_function_any() local 374 nodemask = cpumask_of_node(cpu_to_node(cpu)); in smp_call_function_any() 375 for (cpu = cpumask_first_and(nodemask, mask); cpu < nr_cpu_ids; in smp_call_function_any() 376 cpu = cpumask_next_and(cpu, nodemask, mask)) { in smp_call_function_any()
|
D | cpuset.c | 2433 int cpuset_nodemask_valid_mems_allowed(nodemask_t *nodemask) in cpuset_nodemask_valid_mems_allowed() argument 2435 return nodes_intersects(*nodemask, current->mems_allowed); in cpuset_nodemask_valid_mems_allowed()
|
/linux-4.1.27/drivers/staging/lustre/lustre/libcfs/linux/ |
D | linux-cpu.c | 570 nodemask_t *nodemask; in cfs_cpt_bind() local 578 nodemask = cptab->ctb_nodemask; in cfs_cpt_bind() 581 nodemask = cptab->ctb_parts[cpt].cpt_nodemask; in cfs_cpt_bind() 595 set_mems_allowed(*nodemask); in cfs_cpt_bind()
|
/linux-4.1.27/Documentation/ |
D | memory-hotplug.txt | 418 status_change_nid_normal is set node id when N_NORMAL_MEMORY of nodemask 419 is (will be) set/clear, if this is -1, then nodemask status is not changed. 420 status_change_nid_high is set node id when N_HIGH_MEMORY of nodemask 421 is (will be) set/clear, if this is -1, then nodemask status is not changed. 422 status_change_nid is set node id when N_MEMORY of nodemask is (will be) 424 node loses all memory. If this is -1, then nodemask status is not changed.
|
D | printk-formats.txt | 276 bitmap and its derivatives such as cpumask and nodemask: 281 For printing bitmap and its derivatives such as cpumask and nodemask,
|
/linux-4.1.27/tools/perf/bench/ |
D | numa.c | 285 unsigned long nodemask; in bind_to_memnode() local 291 BUG_ON(g->p.nr_nodes > (int)sizeof(nodemask)); in bind_to_memnode() 292 nodemask = 1L << node; in bind_to_memnode() 294 ret = set_mempolicy(MPOL_BIND, &nodemask, sizeof(nodemask)*8); in bind_to_memnode() 295 dprintf("binding to node %d, mask: %016lx => %d\n", node, nodemask, ret); in bind_to_memnode()
|
/linux-4.1.27/arch/x86/mm/ |
D | numa.c | 319 static void __init numa_nodemask_from_meminfo(nodemask_t *nodemask, in numa_nodemask_from_meminfo() argument 327 node_set(mi->blk[i].nid, *nodemask); in numa_nodemask_from_meminfo()
|
/linux-4.1.27/kernel/irq/ |
D | manage.c | 355 const struct cpumask *nodemask = cpumask_of_node(node); in setup_affinity() local 358 if (cpumask_intersects(mask, nodemask)) in setup_affinity() 359 cpumask_and(mask, mask, nodemask); in setup_affinity()
|
/linux-4.1.27/kernel/sched/ |
D | core.c | 1308 const struct cpumask *nodemask = NULL; in select_fallback_rq() local 1318 nodemask = cpumask_of_node(nid); in select_fallback_rq() 1321 for_each_cpu(dest_cpu, nodemask) { in select_fallback_rq()
|