Lines Matching refs:delta
435 s64 delta = (s64)(vruntime - max_vruntime); in max_vruntime() local
436 if (delta > 0) in max_vruntime()
444 s64 delta = (s64)(vruntime - min_vruntime); in min_vruntime() local
445 if (delta < 0) in min_vruntime()
596 static inline u64 calc_delta_fair(u64 delta, struct sched_entity *se) in calc_delta_fair() argument
599 delta = __calc_delta(delta, NICE_0_LOAD, &se->load); in calc_delta_fair()
601 return delta; in calc_delta_fair()
1703 u64 runtime, delta, now; in numa_get_avg_runtime() local
1709 delta = runtime - p->last_sum_exec_runtime; in numa_get_avg_runtime()
1712 delta = p->se.avg.load_sum / p->se.load.weight; in numa_get_avg_runtime()
1719 return delta; in numa_get_avg_runtime()
2563 u64 delta, scaled_delta, periods; in __update_load_avg() local
2568 delta = now - sa->last_update_time; in __update_load_avg()
2573 if ((s64)delta < 0) { in __update_load_avg()
2582 delta >>= 10; in __update_load_avg()
2583 if (!delta) in __update_load_avg()
2592 if (delta + delta_w >= 1024) { in __update_load_avg()
2615 delta -= delta_w; in __update_load_avg()
2618 periods = delta / 1024; in __update_load_avg()
2619 delta %= 1024; in __update_load_avg()
2641 scaled_delta = cap_scale(delta, scale_freq); in __update_load_avg()
2650 sa->period_contrib += delta; in __update_load_avg()
2671 long delta = cfs_rq->avg.load_avg - cfs_rq->tg_load_avg_contrib; in update_tg_load_avg() local
2673 if (force || abs(delta) > cfs_rq->tg_load_avg_contrib / 64) { in update_tg_load_avg()
2674 atomic_long_add(delta, &cfs_rq->tg->load_avg); in update_tg_load_avg()
2898 u64 delta = rq_clock(rq_of(cfs_rq)) - se->statistics.sleep_start; in enqueue_sleeper() local
2900 if ((s64)delta < 0) in enqueue_sleeper()
2901 delta = 0; in enqueue_sleeper()
2903 if (unlikely(delta > se->statistics.sleep_max)) in enqueue_sleeper()
2904 se->statistics.sleep_max = delta; in enqueue_sleeper()
2907 se->statistics.sum_sleep_runtime += delta; in enqueue_sleeper()
2910 account_scheduler_latency(tsk, delta >> 10, 1); in enqueue_sleeper()
2911 trace_sched_stat_sleep(tsk, delta); in enqueue_sleeper()
2915 u64 delta = rq_clock(rq_of(cfs_rq)) - se->statistics.block_start; in enqueue_sleeper() local
2917 if ((s64)delta < 0) in enqueue_sleeper()
2918 delta = 0; in enqueue_sleeper()
2920 if (unlikely(delta > se->statistics.block_max)) in enqueue_sleeper()
2921 se->statistics.block_max = delta; in enqueue_sleeper()
2924 se->statistics.sum_sleep_runtime += delta; in enqueue_sleeper()
2928 se->statistics.iowait_sum += delta; in enqueue_sleeper()
2930 trace_sched_stat_iowait(tsk, delta); in enqueue_sleeper()
2933 trace_sched_stat_blocked(tsk, delta); in enqueue_sleeper()
2943 delta >> 20); in enqueue_sleeper()
2945 account_scheduler_latency(tsk, delta >> 10, 0); in enqueue_sleeper()
3133 s64 delta; in check_preempt_tick() local
3156 delta = curr->vruntime - se->vruntime; in check_preempt_tick()
3158 if (delta < 0) in check_preempt_tick()
3161 if (delta > ideal_runtime) in check_preempt_tick()
4092 s64 delta = slice - ran; in hrtick_start_fair() local
4094 if (delta < 0) { in hrtick_start_fair()
4099 hrtick_start(rq, delta); in hrtick_start_fair()
5553 s64 delta; in task_hot() local
5576 delta = rq_clock_task(env->src_rq) - p->se.exec_start; in task_hot()
5578 return delta < (s64)sysctl_sched_migration_cost; in task_hot()
6082 s64 delta; in scale_rt_capacity() local
6090 delta = __rq_clock_broken(rq) - age_stamp; in scale_rt_capacity()
6092 if (unlikely(delta < 0)) in scale_rt_capacity()
6093 delta = 0; in scale_rt_capacity()
6095 total = sched_avg_period() + delta; in scale_rt_capacity()