rt_bandwidth 4907 kernel/sched/core.c task_group(p)->rt_bandwidth.rt_runtime == 0 && rt_bandwidth 6606 kernel/sched/core.c init_rt_bandwidth(&root_task_group.rt_bandwidth, rt_bandwidth 13 kernel/sched/rt.c static int do_sched_rt_period_timer(struct rt_bandwidth *rt_b, int overrun); rt_bandwidth 15 kernel/sched/rt.c struct rt_bandwidth def_rt_bandwidth; rt_bandwidth 19 kernel/sched/rt.c struct rt_bandwidth *rt_b = rt_bandwidth 20 kernel/sched/rt.c container_of(timer, struct rt_bandwidth, rt_period_timer); rt_bandwidth 41 kernel/sched/rt.c void init_rt_bandwidth(struct rt_bandwidth *rt_b, u64 period, u64 runtime) rt_bandwidth 53 kernel/sched/rt.c static void start_rt_bandwidth(struct rt_bandwidth *rt_b) rt_bandwidth 106 kernel/sched/rt.c static void destroy_rt_bandwidth(struct rt_bandwidth *rt_b) rt_bandwidth 143 kernel/sched/rt.c destroy_rt_bandwidth(&tg->rt_bandwidth); rt_bandwidth 196 kernel/sched/rt.c init_rt_bandwidth(&tg->rt_bandwidth, rt_bandwidth 211 kernel/sched/rt.c rt_rq->rt_runtime = tg->rt_bandwidth.rt_runtime; rt_bandwidth 452 kernel/sched/rt.c return ktime_to_ns(rt_rq->tg->rt_bandwidth.rt_period); rt_bandwidth 553 kernel/sched/rt.c struct rt_rq *sched_rt_period_rt_rq(struct rt_bandwidth *rt_b, int cpu) rt_bandwidth 555 kernel/sched/rt.c return container_of(rt_b, struct task_group, rt_bandwidth)->rt_rq[cpu]; rt_bandwidth 558 kernel/sched/rt.c static inline struct rt_bandwidth *sched_rt_bandwidth(struct rt_rq *rt_rq) rt_bandwidth 560 kernel/sched/rt.c return &rt_rq->tg->rt_bandwidth; rt_bandwidth 615 kernel/sched/rt.c struct rt_rq *sched_rt_period_rt_rq(struct rt_bandwidth *rt_b, int cpu) rt_bandwidth 620 kernel/sched/rt.c static inline struct rt_bandwidth *sched_rt_bandwidth(struct rt_rq *rt_rq) rt_bandwidth 629 kernel/sched/rt.c struct rt_bandwidth *rt_b = sched_rt_bandwidth(rt_rq); rt_bandwidth 641 kernel/sched/rt.c struct rt_bandwidth *rt_b = sched_rt_bandwidth(rt_rq); rt_bandwidth 701 kernel/sched/rt.c struct rt_bandwidth *rt_b = sched_rt_bandwidth(rt_rq); rt_bandwidth 785 kernel/sched/rt.c struct rt_bandwidth *rt_b = sched_rt_bandwidth(rt_rq); rt_bandwidth 812 kernel/sched/rt.c static int do_sched_rt_period_timer(struct rt_bandwidth *rt_b, int overrun) rt_bandwidth 828 kernel/sched/rt.c if (rt_b == &root_task_group.rt_bandwidth) rt_bandwidth 924 kernel/sched/rt.c struct rt_bandwidth *rt_b = sched_rt_bandwidth(rt_rq); rt_bandwidth 1131 kernel/sched/rt.c start_rt_bandwidth(&rt_rq->tg->rt_bandwidth); rt_bandwidth 2434 kernel/sched/rt.c period = ktime_to_ns(tg->rt_bandwidth.rt_period); rt_bandwidth 2435 kernel/sched/rt.c runtime = tg->rt_bandwidth.rt_runtime; rt_bandwidth 2466 kernel/sched/rt.c period = ktime_to_ns(child->rt_bandwidth.rt_period); rt_bandwidth 2467 kernel/sched/rt.c runtime = child->rt_bandwidth.rt_runtime; rt_bandwidth 2522 kernel/sched/rt.c raw_spin_lock_irq(&tg->rt_bandwidth.rt_runtime_lock); rt_bandwidth 2523 kernel/sched/rt.c tg->rt_bandwidth.rt_period = ns_to_ktime(rt_period); rt_bandwidth 2524 kernel/sched/rt.c tg->rt_bandwidth.rt_runtime = rt_runtime; rt_bandwidth 2533 kernel/sched/rt.c raw_spin_unlock_irq(&tg->rt_bandwidth.rt_runtime_lock); rt_bandwidth 2545 kernel/sched/rt.c rt_period = ktime_to_ns(tg->rt_bandwidth.rt_period); rt_bandwidth 2559 kernel/sched/rt.c if (tg->rt_bandwidth.rt_runtime == RUNTIME_INF) rt_bandwidth 2562 kernel/sched/rt.c rt_runtime_us = tg->rt_bandwidth.rt_runtime; rt_bandwidth 2575 kernel/sched/rt.c rt_runtime = tg->rt_bandwidth.rt_runtime; rt_bandwidth 2584 kernel/sched/rt.c rt_period_us = ktime_to_ns(tg->rt_bandwidth.rt_period); rt_bandwidth 2605 kernel/sched/rt.c if (rt_task(tsk) && tg->rt_bandwidth.rt_runtime == 0) rt_bandwidth 385 kernel/sched/sched.h struct rt_bandwidth rt_bandwidth; rt_bandwidth 1880 kernel/sched/sched.h extern struct rt_bandwidth def_rt_bandwidth; rt_bandwidth 1881 kernel/sched/sched.h extern void init_rt_bandwidth(struct rt_bandwidth *rt_b, u64 period, u64 runtime);