Searched refs:on_rq (Results 1 - 8 of 8) sorted by relevance
/linux-4.1.27/kernel/sched/ |
H A D | sched.h | 20 /* task_struct::on_rq states: */ 1034 return p->on_rq == TASK_ON_RQ_QUEUED; task_on_rq_queued() 1039 return p->on_rq == TASK_ON_RQ_MIGRATING; task_on_rq_migrating() 1221 void (*task_move_group) (struct task_struct *p, int on_rq); 1451 * [S] ->on_rq = MIGRATING [L] rq = task_rq() 1454 * [L] ->on_rq
|
H A D | fair.c | 638 u64 slice = __sched_period(cfs_rq->nr_running + !se->on_rq); sched_slice() 647 if (unlikely(!se->on_rq)) { for_each_sched_entity() 2364 if (se->on_rq) { reweight_entity() 2373 if (se->on_rq) reweight_entity() 2796 if (!__update_entity_runnable_avg(now, cpu, &se->avg, se->on_rq, update_entity_load_avg() 2806 if (se->on_rq) { update_entity_load_avg() 3091 se->on_rq = 1; enqueue_entity() 3173 se->on_rq = 0; dequeue_entity() 3235 if (se->on_rq) { set_next_entity() 3331 if (prev->on_rq) put_prev_entity() 3338 if (prev->on_rq) { put_prev_entity() 3342 /* in !on_rq case, update occurred at dequeue */ put_prev_entity() 3649 if (!se->on_rq) for_each_sched_entity() 3703 if (se->on_rq) for_each_sched_entity() 4228 if (se->on_rq) for_each_sched_entity() 5089 if (unlikely(!se->on_rq || curr == rq->idle)) check_preempt_wakeup() 5130 if (curr->on_rq) pick_next_task_fair() 5274 if (!se->on_rq || throttled_hierarchy(cfs_rq_of(se))) yield_to_task_fair() 5649 p->on_rq = TASK_ON_RQ_MIGRATING; detach_task() 5775 p->on_rq = TASK_ON_RQ_QUEUED; attach_task()
|
H A D | core.c | 1028 !p->on_rq); set_task_cpu() 1451 p->on_rq = TASK_ON_RQ_QUEUED; ttwu_activate() 1502 * the task is still ->on_rq. 1675 if (p->on_rq && ttwu_remote(p, wake_flags)) try_to_wake_up() 1798 p->on_rq = 0; __sched_fork() 1800 p->se.on_rq = 0; __sched_fork() 2096 p->on_rq = TASK_ON_RQ_QUEUED; wake_up_new_task() 2468 * If we see ->on_cpu without ->on_rq, the task is leaving, and has task_sched_runtime() 2477 * Must be ->curr _and_ ->on_rq. If dequeued, we would task_sched_runtime() 2774 prev->on_rq = 0; __schedule() 4630 idle->on_rq = TASK_ON_RQ_QUEUED; init_idle() 4741 p->on_rq = TASK_ON_RQ_MIGRATING; move_queued_task() 4749 p->on_rq = TASK_ON_RQ_QUEUED; move_queued_task()
|
H A D | deadline.c | 605 * prev->on_rq = 0; dl_task_timer()
|
/linux-4.1.27/kernel/rcu/ |
H A D | update.c | 555 !ACCESS_ONCE(t->on_rq) || check_holdout_task() 618 * Wait for all pre-existing t->on_rq and t->nvcsw rcu_tasks_kthread() 642 if (t != current && ACCESS_ONCE(t->on_rq) && for_each_process_thread() 690 * Because ->on_rq and ->nvcsw are not guaranteed
|
/linux-4.1.27/kernel/ |
H A D | cpu.c | 307 if (!p->on_rq) do_each_thread()
|
/linux-4.1.27/kernel/trace/ |
H A D | trace_selftest.c | 1108 while (p->on_rq) { trace_selftest_startup_wakeup()
|
/linux-4.1.27/include/linux/ |
H A D | sched.h | 1184 unsigned int on_rq; member in struct:sched_entity 1309 int on_rq; member in struct:task_struct
|
Completed in 498 milliseconds