src_rq 1758 kernel/sched/core.c struct rq *src_rq, *dst_rq; src_rq 1761 kernel/sched/core.c src_rq = task_rq(p); src_rq 1764 kernel/sched/core.c rq_pin_lock(src_rq, &srf); src_rq 1767 kernel/sched/core.c deactivate_task(src_rq, p, 0); src_rq 1773 kernel/sched/core.c rq_unpin_lock(src_rq, &srf); src_rq 1793 kernel/sched/core.c struct rq *src_rq, *dst_rq; src_rq 1799 kernel/sched/core.c src_rq = cpu_rq(arg->src_cpu); src_rq 1804 kernel/sched/core.c double_rq_lock(src_rq, dst_rq); src_rq 1824 kernel/sched/core.c double_rq_unlock(src_rq, dst_rq); src_rq 2139 kernel/sched/deadline.c struct rq *src_rq; src_rq 2155 kernel/sched/deadline.c src_rq = cpu_rq(cpu); src_rq 2163 kernel/sched/deadline.c src_rq->dl.earliest_dl.next)) src_rq 2167 kernel/sched/deadline.c double_lock_balance(this_rq, src_rq); src_rq 2173 kernel/sched/deadline.c if (src_rq->dl.dl_nr_running <= 1) src_rq 2176 kernel/sched/deadline.c p = pick_earliest_pushable_dl_task(src_rq, this_cpu); src_rq 2187 kernel/sched/deadline.c WARN_ON(p == src_rq->curr); src_rq 2195 kernel/sched/deadline.c src_rq->curr->dl.deadline)) src_rq 2200 kernel/sched/deadline.c deactivate_task(src_rq, p, 0); src_rq 2208 kernel/sched/deadline.c double_unlock_balance(this_rq, src_rq); src_rq 7125 kernel/sched/fair.c struct rq *src_rq; src_rq 7156 kernel/sched/fair.c lockdep_assert_held(&env->src_rq->lock); src_rq 7177 kernel/sched/fair.c delta = rq_clock_task(env->src_rq) - p->se.exec_start; src_rq 7208 kernel/sched/fair.c if (env->src_rq->nr_running > env->src_rq->nr_preferred_running) src_rq 7250 kernel/sched/fair.c lockdep_assert_held(&env->src_rq->lock); src_rq 7295 kernel/sched/fair.c if (task_running(env->src_rq, p)) { src_rq 7328 kernel/sched/fair.c lockdep_assert_held(&env->src_rq->lock); src_rq 7330 kernel/sched/fair.c deactivate_task(env->src_rq, p, DEQUEUE_NOCLOCK); src_rq 7344 kernel/sched/fair.c lockdep_assert_held(&env->src_rq->lock); src_rq 7347 kernel/sched/fair.c &env->src_rq->cfs_tasks, se.group_node) { src_rq 7375 kernel/sched/fair.c struct list_head *tasks = &env->src_rq->cfs_tasks; src_rq 7380 kernel/sched/fair.c lockdep_assert_held(&env->src_rq->lock); src_rq 7390 kernel/sched/fair.c if (env->idle != CPU_NOT_IDLE && env->src_rq->nr_running <= 1) src_rq 8763 kernel/sched/fair.c (env->src_rq->cfs.h_nr_running == 1)) { src_rq 8764 kernel/sched/fair.c if ((check_cpu_capacity(env->src_rq, sd)) && src_rq 8879 kernel/sched/fair.c env.src_rq = busiest; src_rq 9186 kernel/sched/fair.c .src_rq = busiest_rq, src_rq 2054 kernel/sched/rt.c struct rq *src_rq; src_rq 2082 kernel/sched/rt.c src_rq = cpu_rq(cpu); src_rq 2091 kernel/sched/rt.c if (src_rq->rt.highest_prio.next >= src_rq 2100 kernel/sched/rt.c double_lock_balance(this_rq, src_rq); src_rq 2106 kernel/sched/rt.c p = pick_highest_pushable_task(src_rq, this_cpu); src_rq 2113 kernel/sched/rt.c WARN_ON(p == src_rq->curr); src_rq 2124 kernel/sched/rt.c if (p->prio < src_rq->curr->prio) src_rq 2129 kernel/sched/rt.c deactivate_task(src_rq, p, 0); src_rq 2140 kernel/sched/rt.c double_unlock_balance(this_rq, src_rq);