Lines Matching refs:numa_group
877 struct numa_group { struct
907 return p->numa_group ? p->numa_group->gid : 0; in task_numa_group_id() argument
932 if (!p->numa_group) in group_faults()
935 return p->numa_group->faults[task_faults_idx(NUMA_MEM, nid, 0)] + in group_faults()
936 p->numa_group->faults[task_faults_idx(NUMA_MEM, nid, 1)]; in group_faults()
939 static inline unsigned long group_faults_cpu(struct numa_group *group, int nid) in group_faults_cpu()
1040 if (!p->numa_group) in group_weight()
1043 total_faults = p->numa_group->total_faults; in group_weight()
1057 struct numa_group *ng = p->numa_group; in should_numa_migrate_memory()
1262 long imp = env->p->numa_group ? groupimp : taskimp; in task_numa_compare()
1304 if (cur->numa_group == env->p->numa_group) { in task_numa_compare()
1311 if (cur->numa_group) in task_numa_compare()
1319 if (cur->numa_group) in task_numa_compare()
1496 if (env.best_cpu == -1 || (p->numa_group && in task_numa_migrate()
1497 nodes_weight(p->numa_group->active_nodes) > 1)) { in task_numa_migrate()
1531 if (p->numa_group) { in task_numa_migrate()
1537 if (node_isset(nid, p->numa_group->active_nodes)) in task_numa_migrate()
1597 static void update_numa_active_node_mask(struct numa_group *numa_group) in update_numa_active_node_mask() argument
1603 faults = group_faults_cpu(numa_group, nid); in update_numa_active_node_mask()
1609 faults = group_faults_cpu(numa_group, nid); in update_numa_active_node_mask()
1610 if (!node_isset(nid, numa_group->active_nodes)) { in update_numa_active_node_mask()
1612 node_set(nid, numa_group->active_nodes); in update_numa_active_node_mask()
1614 node_clear(nid, numa_group->active_nodes); in update_numa_active_node_mask()
1835 if (p->numa_group) { in task_numa_placement()
1836 group_lock = &p->numa_group->lock; in task_numa_placement()
1877 if (p->numa_group) { in task_numa_placement()
1885 p->numa_group->faults[mem_idx] += diff; in task_numa_placement()
1886 p->numa_group->faults_cpu[mem_idx] += f_diff; in task_numa_placement()
1887 p->numa_group->total_faults += diff; in task_numa_placement()
1888 group_faults += p->numa_group->faults[mem_idx]; in task_numa_placement()
1905 if (p->numa_group) { in task_numa_placement()
1906 update_numa_active_node_mask(p->numa_group); in task_numa_placement()
1921 static inline int get_numa_group(struct numa_group *grp) in get_numa_group()
1926 static inline void put_numa_group(struct numa_group *grp) in put_numa_group()
1935 struct numa_group *grp, *my_grp; in task_numa_group()
1941 if (unlikely(!p->numa_group)) { in task_numa_group()
1942 unsigned int size = sizeof(struct numa_group) + in task_numa_group()
1964 rcu_assign_pointer(p->numa_group, grp); in task_numa_group()
1973 grp = rcu_dereference(tsk->numa_group); in task_numa_group()
1977 my_grp = p->numa_group; in task_numa_group()
2029 rcu_assign_pointer(p->numa_group, grp); in task_numa_group()
2041 struct numa_group *grp = p->numa_group; in task_numa_free()
2054 RCU_INIT_POINTER(p->numa_group, NULL); in task_numa_free()
2111 if (!priv && !local && p->numa_group && in task_numa_fault()
2112 node_isset(cpu_node, p->numa_group->active_nodes) && in task_numa_fault()
2113 node_isset(mem_node, p->numa_group->active_nodes)) in task_numa_fault()
5589 struct numa_group *numa_group = rcu_dereference(p->numa_group); in migrate_degrades_locality() local
5617 if (numa_group) { in migrate_degrades_locality()
8335 if (p->numa_group) { in show_numa_stats()
8336 gsf = p->numa_group->faults[task_faults_idx(NUMA_MEM, node, 0)], in show_numa_stats()
8337 gpf = p->numa_group->faults[task_faults_idx(NUMA_MEM, node, 1)]; in show_numa_stats()