Lines Matching refs:avg
681 p->se.avg.runnable_avg_sum = p->se.avg.running_avg_sum = slice; in init_task_runnable_average()
682 p->se.avg.avg_period = slice; in init_task_runnable_average()
1690 delta = p->se.avg.runnable_avg_sum; in numa_get_avg_runtime()
1691 *period = p->se.avg.avg_period; in numa_get_avg_runtime()
2609 decays -= se->avg.decay_count; in __synchronize_entity_decay()
2610 se->avg.decay_count = 0; in __synchronize_entity_decay()
2614 se->avg.load_avg_contrib = decay_load(se->avg.load_avg_contrib, decays); in __synchronize_entity_decay()
2615 se->avg.utilization_avg_contrib = in __synchronize_entity_decay()
2616 decay_load(se->avg.utilization_avg_contrib, decays); in __synchronize_entity_decay()
2670 se->avg.load_avg_contrib = div_u64(contrib, in __update_group_entity_contrib()
2698 se->avg.load_avg_contrib *= runnable_avg; in __update_group_entity_contrib()
2699 se->avg.load_avg_contrib >>= NICE_0_SHIFT; in __update_group_entity_contrib()
2705 __update_entity_runnable_avg(rq_clock_task(rq), cpu_of(rq), &rq->avg, in update_rq_runnable_avg()
2707 __update_tg_runnable_avg(&rq->avg, &rq->cfs); in update_rq_runnable_avg()
2723 contrib = se->avg.runnable_avg_sum * scale_load_down(se->load.weight); in __update_task_entity_contrib()
2724 contrib /= (se->avg.avg_period + 1); in __update_task_entity_contrib()
2725 se->avg.load_avg_contrib = scale_load(contrib); in __update_task_entity_contrib()
2731 long old_contrib = se->avg.load_avg_contrib; in __update_entity_load_avg_contrib()
2736 __update_tg_runnable_avg(&se->avg, group_cfs_rq(se)); in __update_entity_load_avg_contrib()
2740 return se->avg.load_avg_contrib - old_contrib; in __update_entity_load_avg_contrib()
2749 contrib = se->avg.running_avg_sum * scale_load_down(SCHED_LOAD_SCALE); in __update_task_entity_utilization()
2750 contrib /= (se->avg.avg_period + 1); in __update_task_entity_utilization()
2751 se->avg.utilization_avg_contrib = scale_load(contrib); in __update_task_entity_utilization()
2756 long old_contrib = se->avg.utilization_avg_contrib; in __update_entity_utilization_avg_contrib()
2761 se->avg.utilization_avg_contrib = in __update_entity_utilization_avg_contrib()
2764 return se->avg.utilization_avg_contrib - old_contrib; in __update_entity_utilization_avg_contrib()
2796 if (!__update_entity_runnable_avg(now, cpu, &se->avg, se->on_rq, in update_entity_load_avg()
2857 if (unlikely(se->avg.decay_count <= 0)) { in enqueue_entity_load_avg()
2858 se->avg.last_runnable_update = rq_clock_task(rq_of(cfs_rq)); in enqueue_entity_load_avg()
2859 if (se->avg.decay_count) { in enqueue_entity_load_avg()
2868 se->avg.last_runnable_update -= (-se->avg.decay_count) in enqueue_entity_load_avg()
2872 se->avg.decay_count = 0; in enqueue_entity_load_avg()
2881 subtract_blocked_load_contrib(cfs_rq, se->avg.load_avg_contrib); in enqueue_entity_load_avg()
2885 cfs_rq->runnable_load_avg += se->avg.load_avg_contrib; in enqueue_entity_load_avg()
2886 cfs_rq->utilization_load_avg += se->avg.utilization_avg_contrib; in enqueue_entity_load_avg()
2904 cfs_rq->runnable_load_avg -= se->avg.load_avg_contrib; in dequeue_entity_load_avg()
2905 cfs_rq->utilization_load_avg -= se->avg.utilization_avg_contrib; in dequeue_entity_load_avg()
2907 cfs_rq->blocked_load_avg += se->avg.load_avg_contrib; in dequeue_entity_load_avg()
2908 se->avg.decay_count = atomic64_read(&cfs_rq->decay_counter); in dequeue_entity_load_avg()
4925 if (se->avg.decay_count) { in migrate_task_rq_fair()
4926 se->avg.decay_count = -__synchronize_entity_decay(se); in migrate_task_rq_fair()
4927 atomic_long_add(se->avg.load_avg_contrib, in migrate_task_rq_fair()
5838 if (!se->avg.runnable_avg_sum && !cfs_rq->nr_running) in __update_blocked_averages_cpu()
5900 load = div64_ul(load * se->avg.load_avg_contrib, in update_cfs_rq_h_load()
5913 return div64_ul(p->se.avg.load_avg_contrib * cfs_rq->h_load, in task_h_load()
5923 return p->se.avg.load_avg_contrib; in task_h_load()
6036 u64 total, used, age_stamp, avg; in scale_rt_capacity() local
6044 avg = ACCESS_ONCE(rq->rt_avg); in scale_rt_capacity()
6052 used = div_u64(avg, total); in scale_rt_capacity()
7931 if (se->avg.decay_count) { in switched_from_fair()
7933 subtract_blocked_load_contrib(cfs_rq, se->avg.load_avg_contrib); in switched_from_fair()
8043 se->avg.decay_count = atomic64_read(&cfs_rq->decay_counter); in task_move_group_fair()
8044 cfs_rq->blocked_load_avg += se->avg.load_avg_contrib; in task_move_group_fair()