rq_clock         2231 kernel/sched/core.c 		u64 delta = rq_clock(rq) - rq->idle_stamp;
rq_clock          265 kernel/sched/deadline.c 	zerolag_time -= rq_clock(rq);
rq_clock          661 kernel/sched/deadline.c 	WARN_ON(dl_time_before(rq_clock(rq), dl_se->deadline));
rq_clock          676 kernel/sched/deadline.c 	dl_se->deadline = rq_clock(rq) + dl_se->dl_deadline;
rq_clock          711 kernel/sched/deadline.c 		dl_se->deadline = rq_clock(rq) + pi_se->dl_deadline;
rq_clock          738 kernel/sched/deadline.c 	if (dl_time_before(dl_se->deadline, rq_clock(rq))) {
rq_clock          740 kernel/sched/deadline.c 		dl_se->deadline = rq_clock(rq) + pi_se->dl_deadline;
rq_clock          825 kernel/sched/deadline.c 	u64 laxity = dl_se->deadline - rq_clock(rq);
rq_clock          833 kernel/sched/deadline.c 	WARN_ON(dl_time_before(dl_se->deadline, rq_clock(rq)));
rq_clock          890 kernel/sched/deadline.c 	if (dl_time_before(dl_se->deadline, rq_clock(rq)) ||
rq_clock          891 kernel/sched/deadline.c 	    dl_entity_overflow(dl_se, pi_se, rq_clock(rq))) {
rq_clock          894 kernel/sched/deadline.c 			     !dl_time_before(dl_se->deadline, rq_clock(rq)) &&
rq_clock          900 kernel/sched/deadline.c 		dl_se->deadline = rq_clock(rq) + pi_se->dl_deadline;
rq_clock          937 kernel/sched/deadline.c 	delta = ktime_to_ns(now) - rq_clock(rq);
rq_clock         1116 kernel/sched/deadline.c 	if (dl_time_before(dl_se->deadline, rq_clock(rq)) &&
rq_clock         1117 kernel/sched/deadline.c 	    dl_time_before(rq_clock(rq), dl_next_period(dl_se))) {
rq_clock         1460 kernel/sched/deadline.c 				 rq_clock(rq_of_dl_rq(dl_rq_of_se(dl_se))))) {
rq_clock          881 kernel/sched/fair.c 	wait_start = rq_clock(rq_of(cfs_rq));
rq_clock          900 kernel/sched/fair.c 	delta = rq_clock(rq_of(cfs_rq)) - schedstat_val(se->statistics.wait_start);
rq_clock          939 kernel/sched/fair.c 		u64 delta = rq_clock(rq_of(cfs_rq)) - sleep_start;
rq_clock          956 kernel/sched/fair.c 		u64 delta = rq_clock(rq_of(cfs_rq)) - block_start;
rq_clock         1030 kernel/sched/fair.c 				      rq_clock(rq_of(cfs_rq)));
rq_clock         1033 kernel/sched/fair.c 				      rq_clock(rq_of(cfs_rq)));
rq_clock         4534 kernel/sched/fair.c 	cfs_rq->throttled_clock = rq_clock(rq);
rq_clock         4573 kernel/sched/fair.c 	cfs_b->throttled_time += rq_clock(rq) - cfs_rq->throttled_clock;
rq_clock         9812 kernel/sched/fair.c 	this_rq->idle_stamp = rq_clock(this_rq);
rq_clock         2311 kernel/sched/sched.h 		data->func(data, rq_clock(rq), flags);
rq_clock          158 kernel/sched/stats.h 	unsigned long long now = rq_clock(rq), delta = 0;
rq_clock          177 kernel/sched/stats.h 	unsigned long long now = rq_clock(rq), delta = 0;
rq_clock          198 kernel/sched/stats.h 			t->sched_info.last_queued = rq_clock(rq);
rq_clock          212 kernel/sched/stats.h 	unsigned long long delta = rq_clock(rq) - t->sched_info.last_arrival;