Lines Matching refs:SCHED_CAPACITY_SCALE
1167 smt = DIV_ROUND_UP(SCHED_CAPACITY_SCALE * cpus, ns->compute_capacity); in update_numa_stats()
1171 DIV_ROUND_CLOSEST(ns->compute_capacity, SCHED_CAPACITY_SCALE)); in update_numa_stats()
4770 avg_load = (avg_load * SCHED_CAPACITY_SCALE) / group->sgc->capacity; in find_idlest_group()
6099 if (likely(used < SCHED_CAPACITY_SCALE)) in scale_rt_capacity()
6100 return SCHED_CAPACITY_SCALE - used; in scale_rt_capacity()
6336 sgs->avg_load = (sgs->group_load*SCHED_CAPACITY_SCALE) / sgs->group_capacity; in update_sg_lb_stats()
6544 SCHED_CAPACITY_SCALE); in check_asym_packing()
6573 (busiest->load_per_task * SCHED_CAPACITY_SCALE) / in fix_small_imbalance()
6592 capa_now /= SCHED_CAPACITY_SCALE; in fix_small_imbalance()
6603 busiest->load_per_task * SCHED_CAPACITY_SCALE) { in fix_small_imbalance()
6607 tmp = (busiest->load_per_task * SCHED_CAPACITY_SCALE) / in fix_small_imbalance()
6612 capa_move /= SCHED_CAPACITY_SCALE; in fix_small_imbalance()
6680 ) / SCHED_CAPACITY_SCALE; in calculate_imbalance()
6735 sds.avg_load = (SCHED_CAPACITY_SCALE * sds.total_load) in find_busiest_group()