Searched refs:this_rq (Results 1 – 8 of 8) sorted by relevance
/linux-4.4.14/kernel/sched/ |
D | loadavg.c | 81 long calc_load_fold_active(struct rq *this_rq) in calc_load_fold_active() argument 85 nr_active = this_rq->nr_running; in calc_load_fold_active() 86 nr_active += (long)this_rq->nr_uninterruptible; in calc_load_fold_active() 88 if (nr_active != this_rq->calc_load_active) { in calc_load_fold_active() 89 delta = nr_active - this_rq->calc_load_active; in calc_load_fold_active() 90 this_rq->calc_load_active = nr_active; in calc_load_fold_active() 184 struct rq *this_rq = this_rq(); in calc_load_enter_idle() local 191 delta = calc_load_fold_active(this_rq); in calc_load_enter_idle() 201 struct rq *this_rq = this_rq(); in calc_load_exit_idle() local 206 if (time_before(jiffies, this_rq->calc_load_update)) in calc_load_exit_idle() [all …]
|
D | sched.h | 29 extern void calc_global_load_tick(struct rq *this_rq); 30 extern long calc_load_fold_active(struct rq *this_rq); 33 extern void update_cpu_load_active(struct rq *this_rq); 35 static inline void update_cpu_load_active(struct rq *this_rq) { } in update_cpu_load_active() argument 705 #define this_rq() this_cpu_ptr(&runqueues) macro 1199 void (*task_woken) (struct rq *this_rq, struct task_struct *task); 1218 void (*switched_from) (struct rq *this_rq, struct task_struct *task); 1219 void (*switched_to) (struct rq *this_rq, struct task_struct *task); 1220 void (*prio_changed) (struct rq *this_rq, struct task_struct *task, 1255 extern void idle_enter_fair(struct rq *this_rq); [all …]
|
D | fair.c | 2843 void idle_enter_fair(struct rq *this_rq) in idle_enter_fair() argument 2852 void idle_exit_fair(struct rq *this_rq) in idle_exit_fair() argument 2866 static int idle_balance(struct rq *this_rq); 4314 static void __update_cpu_load(struct rq *this_rq, unsigned long this_load, in __update_cpu_load() argument 4319 this_rq->nr_load_updates++; in __update_cpu_load() 4322 this_rq->cpu_load[0] = this_load; /* Fasttrack for idx 0 */ in __update_cpu_load() 4328 old_load = this_rq->cpu_load[i]; in __update_cpu_load() 4339 this_rq->cpu_load[i] = (old_load * (scale - 1) + new_load) >> i; in __update_cpu_load() 4342 sched_avg_update(this_rq); in __update_cpu_load() 4369 static void update_idle_cpu_load(struct rq *this_rq) in update_idle_cpu_load() argument [all …]
|
D | deadline.c | 1580 static void pull_dl_task(struct rq *this_rq) in pull_dl_task() argument 1582 int this_cpu = this_rq->cpu, cpu; in pull_dl_task() 1588 if (likely(!dl_overloaded(this_rq))) in pull_dl_task() 1597 for_each_cpu(cpu, this_rq->rd->dlo_mask) { in pull_dl_task() 1607 if (this_rq->dl.dl_nr_running && in pull_dl_task() 1608 dl_time_before(this_rq->dl.earliest_dl.curr, in pull_dl_task() 1613 double_lock_balance(this_rq, src_rq); in pull_dl_task() 1630 (!this_rq->dl.dl_nr_running || in pull_dl_task() 1632 this_rq->dl.earliest_dl.curr))) { in pull_dl_task() 1648 activate_task(this_rq, p, 0); in pull_dl_task() [all …]
|
D | rt.c | 263 static void pull_rt_task(struct rq *this_rq); 425 static inline void pull_rt_task(struct rq *this_rq) in pull_rt_task() argument 542 return this_rq()->rd->span; in sched_rt_period_mask() 1962 static void pull_rt_task(struct rq *this_rq) in pull_rt_task() argument 1964 int this_cpu = this_rq->cpu, cpu; in pull_rt_task() 1969 if (likely(!rt_overloaded(this_rq))) in pull_rt_task() 1980 tell_cpu_to_push(this_rq); in pull_rt_task() 1985 for_each_cpu(cpu, this_rq->rd->rto_mask) { in pull_rt_task() 1999 this_rq->rt.highest_prio.curr) in pull_rt_task() 2007 double_lock_balance(this_rq, src_rq); in pull_rt_task() [all …]
|
D | cputime.c | 249 struct rq *rq = this_rq(); in account_idle_time() 265 steal -= this_rq()->prev_steal_time; in steal_account_process_tick() 273 this_rq()->prev_steal_time += jiffies_to_nsecs(steal_jiffies); in steal_account_process_tick() 376 struct rq *rq = this_rq(); in irqtime_account_idle_ticks() 467 struct rq *rq = this_rq(); in account_process_tick()
|
D | idle.c | 24 idle_set_state(this_rq(), idle_state); in sched_idle_set_state()
|
D | core.c | 299 rq = this_rq(); in this_rq_lock() 376 if (rq == this_rq()) { in hrtick_start() 742 if (this_rq()->nr_running > 1) in sched_can_stop_tick() 1127 struct rq *rq = this_rq(); in migration_cpu_stop() 1667 struct rq *rq = this_rq(); in ttwu_stat() 1788 struct rq *rq = this_rq(); in sched_ttwu_pending() 1818 if (llist_empty(&this_rq()->wake_list) && !got_nohz_idle_kick()) in scheduler_ipi() 1841 this_rq()->idle_balance = 1; in scheduler_ipi() 2019 if (WARN_ON_ONCE(rq != this_rq()) || in try_to_wake_up_local() 2541 struct rq *rq = this_rq(); in finish_task_switch() [all …]
|