Lines Matching refs:rq_of

249 static inline struct rq *rq_of(struct cfs_rq *cfs_rq)  in rq_of()  function
296 cfs_rq->tg->parent->cfs_rq[cpu_of(rq_of(cfs_rq))]->on_list) { in list_add_leaf_cfs_rq()
298 &rq_of(cfs_rq)->leaf_cfs_rq_list); in list_add_leaf_cfs_rq()
301 &rq_of(cfs_rq)->leaf_cfs_rq_list); in list_add_leaf_cfs_rq()
374 static inline struct rq *rq_of(struct cfs_rq *cfs_rq) in rq_of() function
704 u64 now = rq_clock_task(rq_of(cfs_rq)); in update_curr()
744 schedstat_set(se->statistics.wait_start, rq_clock(rq_of(cfs_rq))); in update_stats_wait_start()
764 rq_clock(rq_of(cfs_rq)) - se->statistics.wait_start)); in update_stats_wait_end()
767 rq_clock(rq_of(cfs_rq)) - se->statistics.wait_start); in update_stats_wait_end()
771 rq_clock(rq_of(cfs_rq)) - se->statistics.wait_start); in update_stats_wait_end()
797 se->exec_start = rq_clock_task(rq_of(cfs_rq)); in update_stats_curr_start()
2335 update_load_add(&rq_of(cfs_rq)->load, se->load.weight); in account_entity_enqueue()
2338 struct rq *rq = rq_of(cfs_rq); in account_entity_enqueue()
2352 update_load_sub(&rq_of(cfs_rq)->load, se->load.weight); in account_entity_dequeue()
2354 account_numa_dequeue(rq_of(cfs_rq), task_of(se)); in account_entity_dequeue()
2427 se = tg->se[cpu_of(rq_of(cfs_rq))]; in update_cfs_shares()
2704 decayed = __update_load_avg(now, cpu_of(rq_of(cfs_rq)), sa, in update_cfs_rq_load_avg()
2720 int cpu = cpu_of(rq_of(cfs_rq)); in update_load_avg()
2744 __update_load_avg(cfs_rq->avg.last_update_time, cpu_of(rq_of(cfs_rq)), in attach_entity_load_avg()
2763 __update_load_avg(cfs_rq->avg.last_update_time, cpu_of(rq_of(cfs_rq)), in detach_entity_load_avg()
2783 __update_load_avg(now, cpu_of(rq_of(cfs_rq)), sa, in enqueue_entity_load_avg()
2833 __update_load_avg(last_update_time, cpu_of(rq_of(cfs_rq)), &se->avg, 0, 0, NULL); in remove_entity_load_avg()
2898 u64 delta = rq_clock(rq_of(cfs_rq)) - se->statistics.sleep_start; in enqueue_sleeper()
2915 u64 delta = rq_clock(rq_of(cfs_rq)) - se->statistics.block_start; in enqueue_sleeper()
3096 se->statistics.sleep_start = rq_clock(rq_of(cfs_rq)); in dequeue_entity()
3098 se->statistics.block_start = rq_clock(rq_of(cfs_rq)); in dequeue_entity()
3138 resched_curr(rq_of(cfs_rq)); in check_preempt_tick()
3162 resched_curr(rq_of(cfs_rq)); in check_preempt_tick()
3188 if (rq_of(cfs_rq)->load.weight >= 2*se->load.weight) { in set_next_entity()
3302 resched_curr(rq_of(cfs_rq)); in entity_tick()
3309 hrtimer_active(&rq_of(cfs_rq)->hrtick_timer)) in entity_tick()
3395 return rq_clock_task(rq_of(cfs_rq)) - cfs_rq->throttled_clock_task_time; in cfs_rq_clock_task()
3444 if (likely((s64)(rq_clock(rq_of(cfs_rq)) - cfs_rq->runtime_expires) < 0)) in expire_cfs_rq_runtime()
3484 resched_curr(rq_of(cfs_rq)); in __account_cfs_rq_runtime()
3557 struct rq *rq = rq_of(cfs_rq); in throttle_cfs_rq()
3563 se = cfs_rq->tg->se[cpu_of(rq_of(cfs_rq))]; in throttle_cfs_rq()
3611 struct rq *rq = rq_of(cfs_rq); in unthrottle_cfs_rq()
3666 struct rq *rq = rq_of(cfs_rq); in distribute_cfs_runtime()
4037 return rq_clock_task(rq_of(cfs_rq)); in cfs_rq_clock_task()
5927 struct rq *rq = rq_of(cfs_rq); in update_cfs_rq_h_load()