Searched refs:task_rq (Results 1 – 5 of 5) sorted by relevance
704 #define task_rq(p) cpu_rq(task_cpu(p)) macro1423 rq = task_rq(p); in __task_rq_lock()1425 if (likely(rq == task_rq(p) && !task_on_rq_migrating(p))) in __task_rq_lock()1445 rq = task_rq(p); in task_rq_lock()1463 if (likely(rq == task_rq(p) && !task_on_rq_migrating(p))) in task_rq_lock()
36 struct rq *rq = task_rq(p); in dl_rq_of_se()1253 best_cpu = cpudl_find(&task_rq(task)->rd->cpudl, in find_later_rq()1338 if (unlikely(task_rq(task) != rq || in find_lock_later_rq()1586 rq = task_rq(p); in set_cpus_allowed_dl()1728 push_dl_task(rq) && rq != task_rq(p)) in switched_to_dl()
239 return task_rq(p); in rq_of_rt_se()1562 if (!cpupri_find(&task_rq(task)->rd->cpupri, task, lowest_mask)) in find_lowest_rq()1655 if (unlikely(task_rq(task) != rq || in find_lock_lowest_rq()2082 rq = task_rq(p); in set_cpus_allowed_rt()2174 push_rt_task(rq) && rq != task_rq(p)) in switched_to_rt()
1042 lockdep_is_held(&task_rq(p)->lock))); in set_task_cpu()1063 src_rq = task_rq(p); in __migrate_swap_task()1197 rq = task_rq(p); in wait_task_inactive()1722 struct rq *rq = task_rq(p); in try_to_wake_up_local()4074 if (!cpumask_subset(task_rq(p)->rd->span, new_mask)) { in sched_setaffinity()4340 p_rq = task_rq(p); in yield_to()4351 if (task_rq(p) != p_rq) { in yield_to()4694 if (dl_task(p) && !cpumask_intersects(task_rq(p)->rd->span, in task_can_attach()4736 struct rq *rq = task_rq(p); in move_queued_task()
391 return &task_rq(p)->cfs; in task_cfs_rq()397 struct rq *rq = task_rq(p); in cfs_rq_of()4174 WARN_ON(task_rq(p) != rq); in hrtick_start_fair()5774 BUG_ON(task_rq(p) != rq); in attach_task()