Lines Matching refs:curr

78 	dl_rq->earliest_dl.curr = dl_rq->earliest_dl.next = 0;  in init_dl_rq()
648 if (dl_task(rq->curr)) in dl_task_timer()
716 struct task_struct *curr = rq->curr; in update_curr_dl() local
717 struct sched_dl_entity *dl_se = &curr->dl; in update_curr_dl()
720 if (!dl_task(curr) || !on_dl_rq(dl_se)) in update_curr_dl()
731 delta_exec = rq_clock_task(rq) - curr->se.exec_start; in update_curr_dl()
735 schedstat_set(curr->se.statistics.exec_max, in update_curr_dl()
736 max(curr->se.statistics.exec_max, delta_exec)); in update_curr_dl()
738 curr->se.sum_exec_runtime += delta_exec; in update_curr_dl()
739 account_group_exec_runtime(curr, delta_exec); in update_curr_dl()
741 curr->se.exec_start = rq_clock_task(rq); in update_curr_dl()
742 cpuacct_charge(curr, delta_exec); in update_curr_dl()
749 __dequeue_task_dl(rq, curr, 0); in update_curr_dl()
750 if (unlikely(dl_se->dl_boosted || !start_dl_timer(curr))) in update_curr_dl()
751 enqueue_task_dl(rq, curr, ENQUEUE_REPLENISH); in update_curr_dl()
753 if (!is_leftmost(curr, &rq->dl)) in update_curr_dl()
801 if (dl_rq->earliest_dl.curr == 0 || in inc_dl_deadline()
802 dl_time_before(deadline, dl_rq->earliest_dl.curr)) { in inc_dl_deadline()
809 dl_rq->earliest_dl.next = dl_rq->earliest_dl.curr; in inc_dl_deadline()
810 dl_rq->earliest_dl.curr = deadline; in inc_dl_deadline()
833 dl_rq->earliest_dl.curr = 0; in dec_dl_deadline()
841 dl_rq->earliest_dl.curr = entry->deadline; in dec_dl_deadline()
1020 struct task_struct *p = rq->curr; in yield_task_dl()
1029 rq->curr->dl.dl_yielded = 1; in yield_task_dl()
1049 struct task_struct *curr; in select_task_rq_dl() local
1058 curr = READ_ONCE(rq->curr); /* unlocked access */ in select_task_rq_dl()
1069 if (unlikely(dl_task(curr)) && in select_task_rq_dl()
1070 (curr->nr_cpus_allowed < 2 || in select_task_rq_dl()
1071 !dl_entity_preempt(&p->dl, &curr->dl)) && in select_task_rq_dl()
1077 cpu_rq(target)->dl.earliest_dl.curr) || in select_task_rq_dl()
1093 if (rq->curr->nr_cpus_allowed == 1 || in check_preempt_equal_dl()
1094 cpudl_find(&rq->rd->cpudl, rq->curr, NULL) == -1) in check_preempt_equal_dl()
1117 if (dl_entity_preempt(&p->dl, &rq->curr->dl)) { in check_preempt_curr_dl()
1127 if ((p->dl.deadline == rq->curr->dl.deadline) && in check_preempt_curr_dl()
1128 !test_tsk_need_resched(rq->curr)) in check_preempt_curr_dl()
1256 struct task_struct *p = rq->curr; in set_curr_task_dl()
1430 later_rq->dl.earliest_dl.curr)) { in find_lock_later_rq()
1460 later_rq->dl.earliest_dl.curr)) in find_lock_later_rq()
1510 if (unlikely(next_task == rq->curr)) { in push_dl_task()
1520 if (dl_task(rq->curr) && in push_dl_task()
1521 dl_time_before(next_task->dl.deadline, rq->curr->dl.deadline) && in push_dl_task()
1522 rq->curr->nr_cpus_allowed > 1) { in push_dl_task()
1608 dl_time_before(this_rq->dl.earliest_dl.curr, in pull_dl_task()
1632 this_rq->dl.earliest_dl.curr))) { in pull_dl_task()
1633 WARN_ON(p == src_rq->curr); in pull_dl_task()
1641 src_rq->curr->dl.deadline)) in pull_dl_task()
1668 !test_tsk_need_resched(rq->curr) && in task_woken_dl()
1670 dl_task(rq->curr) && in task_woken_dl()
1671 (rq->curr->nr_cpus_allowed < 2 || in task_woken_dl()
1672 !dl_entity_preempt(&p->dl, &rq->curr->dl))) { in task_woken_dl()
1718 cpudl_set(&rq->rd->cpudl, rq->cpu, rq->dl.earliest_dl.curr, 1); in rq_online_dl()
1770 if (task_on_rq_queued(p) && rq->curr != p) { in switched_to_dl()
1775 if (dl_task(rq->curr)) in switched_to_dl()
1790 if (task_on_rq_queued(p) || rq->curr == p) { in prio_changed_dl()
1806 if (dl_time_before(rq->dl.earliest_dl.curr, p->dl.deadline)) in prio_changed_dl()