Lines Matching defs:rq

317 #define for_each_leaf_cfs_rq(rq, cfs_rq) \  argument
392 struct rq *rq = task_rq(p); in cfs_rq_of() local
411 #define for_each_leaf_cfs_rq(rq, cfs_rq) \ argument
736 static void update_curr_fair(struct rq *rq) in update_curr_fair()
865 static void account_numa_enqueue(struct rq *rq, struct task_struct *p) in account_numa_enqueue()
871 static void account_numa_dequeue(struct rq *rq, struct task_struct *p) in account_numa_dequeue()
1146 struct rq *rq = cpu_rq(cpu); in update_numa_stats() local
2285 void task_tick_numa(struct rq *rq, struct task_struct *curr) in task_tick_numa()
2317 static void task_tick_numa(struct rq *rq, struct task_struct *curr) in task_tick_numa()
2321 static inline void account_numa_enqueue(struct rq *rq, struct task_struct *p) in account_numa_enqueue()
2325 static inline void account_numa_dequeue(struct rq *rq, struct task_struct *p) in account_numa_dequeue()
2338 struct rq *rq = rq_of(cfs_rq); in account_entity_enqueue() local
2882 static inline int idle_balance(struct rq *rq) in idle_balance()
3527 struct rq *rq = data; in tg_unthrottle_up() local
3544 struct rq *rq = data; in tg_throttle_down() local
3557 struct rq *rq = rq_of(cfs_rq); in throttle_cfs_rq() local
3611 struct rq *rq = rq_of(cfs_rq); in unthrottle_cfs_rq() local
3666 struct rq *rq = rq_of(cfs_rq); in distribute_cfs_runtime() local
3997 static void __maybe_unused update_runtime_enabled(struct rq *rq) in update_runtime_enabled()
4010 static void __maybe_unused unthrottle_offline_cfs_rqs(struct rq *rq) in unthrottle_offline_cfs_rqs()
4072 static inline void update_runtime_enabled(struct rq *rq) {} in update_runtime_enabled()
4073 static inline void unthrottle_offline_cfs_rqs(struct rq *rq) {} in unthrottle_offline_cfs_rqs()
4082 static void hrtick_start_fair(struct rq *rq, struct task_struct *p) in hrtick_start_fair()
4108 static void hrtick_update(struct rq *rq) in hrtick_update()
4120 hrtick_start_fair(struct rq *rq, struct task_struct *p) in hrtick_start_fair()
4124 static inline void hrtick_update(struct rq *rq) in hrtick_update()
4135 enqueue_task_fair(struct rq *rq, struct task_struct *p, int flags) in enqueue_task_fair()
4183 static void dequeue_task_fair(struct rq *rq, struct task_struct *p, int flags) in dequeue_task_fair()
4435 struct rq *rq = cpu_rq(cpu); in source_load() local
4450 struct rq *rq = cpu_rq(cpu); in target_load() local
4471 struct rq *rq = cpu_rq(cpu); in cpu_avg_load_per_task() local
4801 struct rq *rq = cpu_rq(i); in find_idlest_cpu() local
5114 static void check_preempt_wakeup(struct rq *rq, struct task_struct *p, int wake_flags) in check_preempt_wakeup()
5198 pick_next_task_fair(struct rq *rq, struct task_struct *prev) in pick_next_task_fair()
5331 static void put_prev_task_fair(struct rq *rq, struct task_struct *prev) in put_prev_task_fair()
5347 static void yield_task_fair(struct rq *rq) in yield_task_fair()
5378 static bool yield_to_task_fair(struct rq *rq, struct task_struct *p, bool preempt) in yield_to_task_fair()
5853 static void attach_task(struct rq *rq, struct task_struct *p) in attach_task()
5867 static void attach_one_task(struct rq *rq, struct task_struct *p) in attach_one_task()
5898 struct rq *rq = cpu_rq(cpu); in update_blocked_averages() local
5927 struct rq *rq = rq_of(cfs_rq); in update_cfs_rq_h_load() local
5969 struct rq *rq = cpu_rq(cpu); in update_blocked_averages() local
6080 struct rq *rq = cpu_rq(cpu); in scale_rt_capacity() local
6148 struct rq *rq = cpu_rq(cpu); in update_group_capacity() local
6191 check_cpu_capacity(struct rq *rq, struct sched_domain *sd) in check_cpu_capacity()
6310 struct rq *rq = cpu_rq(i); in update_sg_lb_stats() local
6406 static inline enum fbq_type fbq_classify_rq(struct rq *rq) in fbq_classify_rq()
6420 static inline enum fbq_type fbq_classify_rq(struct rq *rq) in fbq_classify_rq()
6802 struct rq *busiest = NULL, *rq; in find_busiest_queue() local
7413 static inline int on_null_domain(struct rq *rq) in on_null_domain()
7573 static void rebalance_domains(struct rq *rq, enum cpu_idle_type idle) in rebalance_domains()
7684 struct rq *rq; in nohz_idle_balance() local
7748 static inline bool nohz_kick_needed(struct rq *rq) in nohz_kick_needed()
7841 void trigger_load_balance(struct rq *rq) in trigger_load_balance()
7855 static void rq_online_fair(struct rq *rq) in rq_online_fair()
7862 static void rq_offline_fair(struct rq *rq) in rq_offline_fair()
7875 static void task_tick_fair(struct rq *rq, struct task_struct *curr, int queued) in task_tick_fair()
7899 struct rq *rq = this_rq(); in task_fork_fair() local
7944 prio_changed_fair(struct rq *rq, struct task_struct *p, int oldprio) in prio_changed_fair()
8026 static void switched_from_fair(struct rq *rq, struct task_struct *p) in switched_from_fair()
8031 static void switched_to_fair(struct rq *rq, struct task_struct *p) in switched_to_fair()
8053 static void set_curr_task_fair(struct rq *rq) in set_curr_task_fair()
8155 struct rq *rq = cpu_rq(cpu); in unregister_fair_sched_group() local
8174 struct rq *rq = cpu_rq(cpu); in init_tg_cfs_entry() local
8222 struct rq *rq = cpu_rq(i); in sched_group_set_shares() local
8254 static unsigned int get_rr_interval_fair(struct rq *rq, struct task_struct *task) in get_rr_interval_fair()