Searched refs:on_rq (Results 1 - 8 of 8) sorted by relevance

/linux-4.1.27/kernel/sched/
H A Dsched.h20 /* task_struct::on_rq states: */
1034 return p->on_rq == TASK_ON_RQ_QUEUED; task_on_rq_queued()
1039 return p->on_rq == TASK_ON_RQ_MIGRATING; task_on_rq_migrating()
1221 void (*task_move_group) (struct task_struct *p, int on_rq);
1451 * [S] ->on_rq = MIGRATING [L] rq = task_rq()
1454 * [L] ->on_rq
H A Dfair.c638 u64 slice = __sched_period(cfs_rq->nr_running + !se->on_rq); sched_slice()
647 if (unlikely(!se->on_rq)) { for_each_sched_entity()
2364 if (se->on_rq) { reweight_entity()
2373 if (se->on_rq) reweight_entity()
2796 if (!__update_entity_runnable_avg(now, cpu, &se->avg, se->on_rq, update_entity_load_avg()
2806 if (se->on_rq) { update_entity_load_avg()
3091 se->on_rq = 1; enqueue_entity()
3173 se->on_rq = 0; dequeue_entity()
3235 if (se->on_rq) { set_next_entity()
3331 if (prev->on_rq) put_prev_entity()
3338 if (prev->on_rq) { put_prev_entity()
3342 /* in !on_rq case, update occurred at dequeue */ put_prev_entity()
3649 if (!se->on_rq) for_each_sched_entity()
3703 if (se->on_rq) for_each_sched_entity()
4228 if (se->on_rq) for_each_sched_entity()
5089 if (unlikely(!se->on_rq || curr == rq->idle)) check_preempt_wakeup()
5130 if (curr->on_rq) pick_next_task_fair()
5274 if (!se->on_rq || throttled_hierarchy(cfs_rq_of(se))) yield_to_task_fair()
5649 p->on_rq = TASK_ON_RQ_MIGRATING; detach_task()
5775 p->on_rq = TASK_ON_RQ_QUEUED; attach_task()
H A Dcore.c1028 !p->on_rq); set_task_cpu()
1451 p->on_rq = TASK_ON_RQ_QUEUED; ttwu_activate()
1502 * the task is still ->on_rq.
1675 if (p->on_rq && ttwu_remote(p, wake_flags)) try_to_wake_up()
1798 p->on_rq = 0; __sched_fork()
1800 p->se.on_rq = 0; __sched_fork()
2096 p->on_rq = TASK_ON_RQ_QUEUED; wake_up_new_task()
2468 * If we see ->on_cpu without ->on_rq, the task is leaving, and has task_sched_runtime()
2477 * Must be ->curr _and_ ->on_rq. If dequeued, we would task_sched_runtime()
2774 prev->on_rq = 0; __schedule()
4630 idle->on_rq = TASK_ON_RQ_QUEUED; init_idle()
4741 p->on_rq = TASK_ON_RQ_MIGRATING; move_queued_task()
4749 p->on_rq = TASK_ON_RQ_QUEUED; move_queued_task()
H A Ddeadline.c605 * prev->on_rq = 0; dl_task_timer()
/linux-4.1.27/kernel/rcu/
H A Dupdate.c555 !ACCESS_ONCE(t->on_rq) || check_holdout_task()
618 * Wait for all pre-existing t->on_rq and t->nvcsw rcu_tasks_kthread()
642 if (t != current && ACCESS_ONCE(t->on_rq) && for_each_process_thread()
690 * Because ->on_rq and ->nvcsw are not guaranteed
/linux-4.1.27/kernel/
H A Dcpu.c307 if (!p->on_rq) do_each_thread()
/linux-4.1.27/kernel/trace/
H A Dtrace_selftest.c1108 while (p->on_rq) { trace_selftest_startup_wakeup()
/linux-4.1.27/include/linux/
H A Dsched.h1184 unsigned int on_rq; member in struct:sched_entity
1309 int on_rq; member in struct:task_struct

Completed in 498 milliseconds