Searched refs:SCHED_CAPACITY_SCALE (Results 1 – 5 of 5) sorted by relevance
310 set_capacity_scale(cpu, SCHED_CAPACITY_SCALE); in init_cpu_topology()
1161 smt = DIV_ROUND_UP(SCHED_CAPACITY_SCALE * cpus, ns->compute_capacity); in update_numa_stats()1165 DIV_ROUND_CLOSEST(ns->compute_capacity, SCHED_CAPACITY_SCALE)); in update_numa_stats()4680 avg_load = (avg_load * SCHED_CAPACITY_SCALE) / group->sgc->capacity; in find_idlest_group()6025 return SCHED_CAPACITY_SCALE; in default_scale_cpu_capacity()6054 if (likely(used < SCHED_CAPACITY_SCALE)) in scale_rt_capacity()6055 return SCHED_CAPACITY_SCALE - used; in scale_rt_capacity()6062 unsigned long capacity = SCHED_CAPACITY_SCALE; in update_cpu_capacity()6298 sgs->avg_load = (sgs->group_load*SCHED_CAPACITY_SCALE) / sgs->group_capacity; in update_sg_lb_stats()6506 SCHED_CAPACITY_SCALE); in check_asym_packing()6535 (busiest->load_per_task * SCHED_CAPACITY_SCALE) / in fix_small_imbalance()[all …]
1396 return SCHED_CAPACITY_SCALE; in arch_scale_freq_capacity()
5461 if (group->sgc->capacity != SCHED_CAPACITY_SCALE) { in sched_domain_debug_one()5923 sg->sgc->capacity = SCHED_CAPACITY_SCALE * cpumask_weight(sg_span); in build_overlap_sched_groups()7216 rq->cpu_capacity = rq->cpu_capacity_orig = SCHED_CAPACITY_SCALE; in sched_init()
901 #define SCHED_CAPACITY_SCALE (1L << SCHED_CAPACITY_SHIFT) macro