Lines Matching refs:local
1620 unsigned long local = p->numa_faults_locality[1]; in update_task_scan_period() local
1629 if (local + shared == 0 || p->numa_faults_locality[2]) { in update_task_scan_period()
1646 ratio = (local * NUMA_PERIOD_SLOTS) / (local + remote); in update_task_scan_period()
2043 int local = !!(flags & TNF_FAULT_LOCAL); in task_numa_fault() local
2084 if (!priv && !local && p->numa_group && in task_numa_fault()
2087 local = 1; in task_numa_fault()
2105 p->numa_faults_locality[local] += pages; in task_numa_fault()
5962 struct sched_group *local; /* Local group in this sd */ member
5981 .local = NULL, in init_sd_lb_stats()
6412 sds->local = sg; in update_sd_lb_stats()
6436 if (prefer_sibling && sds->local && in update_sd_lb_stats()
6524 struct sg_lb_stats *local, *busiest; in fix_small_imbalance() local
6526 local = &sds->local_stat; in fix_small_imbalance()
6529 if (!local->sum_nr_running) in fix_small_imbalance()
6530 local->load_per_task = cpu_avg_load_per_task(env->dst_cpu); in fix_small_imbalance()
6531 else if (busiest->load_per_task > local->load_per_task) in fix_small_imbalance()
6539 local->avg_load + (scaled_busy_load_per_task * imbn)) { in fix_small_imbalance()
6552 capa_now += local->group_capacity * in fix_small_imbalance()
6553 min(local->load_per_task, local->avg_load); in fix_small_imbalance()
6567 local->group_capacity; in fix_small_imbalance()
6570 local->group_capacity; in fix_small_imbalance()
6572 capa_move += local->group_capacity * in fix_small_imbalance()
6573 min(local->load_per_task, local->avg_load + tmp); in fix_small_imbalance()
6590 struct sg_lb_stats *local, *busiest; in calculate_imbalance() local
6592 local = &sds->local_stat; in calculate_imbalance()
6610 local->avg_load >= sds->avg_load) { in calculate_imbalance()
6619 local->group_type == group_overloaded) { in calculate_imbalance()
6641 (sds->avg_load - local->avg_load) * local->group_capacity in calculate_imbalance()
6675 struct sg_lb_stats *local, *busiest; in find_busiest_group() local
6685 local = &sds.local_stat; in find_busiest_group()
6709 if (env->idle == CPU_NEWLY_IDLE && group_has_capacity(env, local) && in find_busiest_group()
6717 if (local->avg_load >= busiest->avg_load) in find_busiest_group()
6724 if (local->avg_load >= sds.avg_load) in find_busiest_group()
6736 (local->idle_cpus <= (busiest->idle_cpus + 1))) in find_busiest_group()
6744 env->sd->imbalance_pct * local->avg_load) in find_busiest_group()