Home
last modified time | relevance | path

Searched refs:cpu_rq (Results 1 – 9 of 9) sorted by relevance

/linux-4.4.14/kernel/sched/
Dcpuacct.c108 raw_spin_lock_irq(&cpu_rq(cpu)->lock); in cpuacct_cpuusage_read()
110 raw_spin_unlock_irq(&cpu_rq(cpu)->lock); in cpuacct_cpuusage_read()
126 raw_spin_lock_irq(&cpu_rq(cpu)->lock); in cpuacct_cpuusage_write()
128 raw_spin_unlock_irq(&cpu_rq(cpu)->lock); in cpuacct_cpuusage_write()
Dcore.c396 hrtick_clear(cpu_rq(cpu)); in hotplug_hrtick()
600 struct rq *rq = cpu_rq(cpu); in resched_cpu()
655 struct rq *rq = cpu_rq(cpu); in wake_up_idle_cpu()
919 struct task_struct *old_stop = cpu_rq(cpu)->stop; in sched_set_stop_task()
935 cpu_rq(cpu)->stop = stop; in sched_set_stop_task()
1079 rq = cpu_rq(new_cpu); in move_queued_task()
1311 dst_rq = cpu_rq(cpu); in __migrate_swap_task()
1341 src_rq = cpu_rq(arg->src_cpu); in migrate_swap_stop()
1342 dst_rq = cpu_rq(arg->dst_cpu); in migrate_swap_stop()
1849 struct rq *rq = cpu_rq(cpu); in ttwu_queue_remote()
[all …]
Ddeadline.c267 later_rq = cpu_rq(cpu); in dl_task_offline_migration()
1055 rq = cpu_rq(cpu); in select_task_rq_dl()
1077 cpu_rq(target)->dl.earliest_dl.curr) || in select_task_rq_dl()
1078 (cpu_rq(target)->dl.dl_nr_running == 0))) in select_task_rq_dl()
1426 later_rq = cpu_rq(cpu); in find_lock_later_rq()
1601 src_rq = cpu_rq(cpu); in pull_dl_task()
1856 print_dl_rq(m, cpu, &cpu_rq(cpu)->dl); in print_dl_stats()
Dfair.c1146 struct rq *rq = cpu_rq(cpu); in update_numa_stats()
1257 struct rq *src_rq = cpu_rq(env->src_cpu); in task_numa_compare()
1258 struct rq *dst_rq = cpu_rq(env->dst_cpu); in task_numa_compare()
1968 tsk = READ_ONCE(cpu_rq(cpu)->curr); in task_numa_group()
4348 return cfs_rq_runnable_load_avg(&cpu_rq(cpu)->cfs); in weighted_cpuload()
4435 struct rq *rq = cpu_rq(cpu); in source_load()
4450 struct rq *rq = cpu_rq(cpu); in target_load()
4461 return cpu_rq(cpu)->cpu_capacity; in capacity_of()
4466 return cpu_rq(cpu)->cpu_capacity_orig; in capacity_orig_of()
4471 struct rq *rq = cpu_rq(cpu); in cpu_avg_load_per_task()
[all …]
Drt.c162 struct rq *rq = cpu_rq(cpu); in init_tg_rt_entry()
616 return &cpu_rq(cpu)->rt; in sched_rt_period_rt_rq()
1326 rq = cpu_rq(cpu); in select_task_rq_rt()
1363 p->prio < cpu_rq(target)->rt.highest_prio.curr) in select_task_rq_rt()
1646 lowest_rq = cpu_rq(cpu); in find_lock_lowest_rq()
1855 next_rq = cpu_rq(cpu); in find_next_push_cpu()
1912 rq = cpu_rq(this_cpu); in try_to_push_tasks()
1989 src_rq = cpu_rq(cpu); in pull_rt_task()
2304 for_each_rt_rq(rt_rq, iter, cpu_rq(cpu)) in print_rt_stats()
Dstats.c29 rq = cpu_rq(cpu); in show_schedstat()
Dsched.h704 #define cpu_rq(cpu) (&per_cpu(runqueues, (cpu))) macro
706 #define task_rq(p) cpu_rq(task_cpu(p))
707 #define cpu_curr(cpu) (cpu_rq(cpu)->curr)
794 for (__sd = rcu_dereference_check_sched_domain(cpu_rq(cpu)->sd); \
1722 #define nohz_flags(cpu) (&cpu_rq(cpu)->nohz_flags)
Ddebug.c170 struct rq *rq = cpu_rq(cpu); in print_cfs_rq()
189 rq0_min_vruntime = cpu_rq(0)->cfs.min_vruntime; in print_cfs_rq()
269 struct rq *rq = cpu_rq(cpu); in print_cpu()
/linux-4.4.14/tools/perf/Documentation/
Dperf-probe.txt184 … be moved easily by modifying schedule(), but the same line matching 'rq=cpu_rq*' may still exist …