later_rq          531 kernel/sched/deadline.c 	struct rq *later_rq = NULL;
later_rq          534 kernel/sched/deadline.c 	later_rq = find_lock_later_rq(p, rq);
later_rq          535 kernel/sched/deadline.c 	if (!later_rq) {
later_rq          557 kernel/sched/deadline.c 		later_rq = cpu_rq(cpu);
later_rq          558 kernel/sched/deadline.c 		double_lock_balance(rq, later_rq);
later_rq          571 kernel/sched/deadline.c 		add_rq_bw(&p->dl, &later_rq->dl);
later_rq          572 kernel/sched/deadline.c 		add_running_bw(&p->dl, &later_rq->dl);
later_rq          575 kernel/sched/deadline.c 		add_rq_bw(&p->dl, &later_rq->dl);
later_rq          588 kernel/sched/deadline.c 	dl_b = &later_rq->rd->dl_bw;
later_rq          590 kernel/sched/deadline.c 	__dl_add(dl_b, p->dl.dl_bw, cpumask_weight(later_rq->rd->span));
later_rq          593 kernel/sched/deadline.c 	set_task_cpu(p, later_rq->cpu);
later_rq          594 kernel/sched/deadline.c 	double_unlock_balance(later_rq, rq);
later_rq          596 kernel/sched/deadline.c 	return later_rq;
later_rq         1966 kernel/sched/deadline.c 	struct rq *later_rq = NULL;
later_rq         1976 kernel/sched/deadline.c 		later_rq = cpu_rq(cpu);
later_rq         1978 kernel/sched/deadline.c 		if (later_rq->dl.dl_nr_running &&
later_rq         1980 kernel/sched/deadline.c 					later_rq->dl.earliest_dl.curr)) {
later_rq         1986 kernel/sched/deadline.c 			later_rq = NULL;
later_rq         1991 kernel/sched/deadline.c 		if (double_lock_balance(rq, later_rq)) {
later_rq         1993 kernel/sched/deadline.c 				     !cpumask_test_cpu(later_rq->cpu, task->cpus_ptr) ||
later_rq         1997 kernel/sched/deadline.c 				double_unlock_balance(rq, later_rq);
later_rq         1998 kernel/sched/deadline.c 				later_rq = NULL;
later_rq         2008 kernel/sched/deadline.c 		if (!later_rq->dl.dl_nr_running ||
later_rq         2010 kernel/sched/deadline.c 				   later_rq->dl.earliest_dl.curr))
later_rq         2014 kernel/sched/deadline.c 		double_unlock_balance(rq, later_rq);
later_rq         2015 kernel/sched/deadline.c 		later_rq = NULL;
later_rq         2018 kernel/sched/deadline.c 	return later_rq;
later_rq         2049 kernel/sched/deadline.c 	struct rq *later_rq;
later_rq         2079 kernel/sched/deadline.c 	later_rq = find_lock_later_rq(next_task, rq);
later_rq         2080 kernel/sched/deadline.c 	if (!later_rq) {
later_rq         2107 kernel/sched/deadline.c 	set_task_cpu(next_task, later_rq->cpu);
later_rq         2113 kernel/sched/deadline.c 	update_rq_clock(later_rq);
later_rq         2114 kernel/sched/deadline.c 	activate_task(later_rq, next_task, ENQUEUE_NOCLOCK);
later_rq         2117 kernel/sched/deadline.c 	resched_curr(later_rq);
later_rq         2119 kernel/sched/deadline.c 	double_unlock_balance(rq, later_rq);