rt_runtime 4907 kernel/sched/core.c task_group(p)->rt_bandwidth.rt_runtime == 0 && rt_runtime 6657 kernel/sched/core.c rq->rt.rt_runtime = def_rt_bandwidth.rt_runtime; rt_runtime 584 kernel/sched/debug.c PN(rt_runtime); rt_runtime 44 kernel/sched/rt.c rt_b->rt_runtime = runtime; rt_runtime 55 kernel/sched/rt.c if (!rt_bandwidth_enabled() || rt_b->rt_runtime == RUNTIME_INF) rt_runtime 101 kernel/sched/rt.c rt_rq->rt_runtime = 0; rt_runtime 211 kernel/sched/rt.c rt_rq->rt_runtime = tg->rt_bandwidth.rt_runtime; rt_runtime 447 kernel/sched/rt.c return rt_rq->rt_runtime; rt_runtime 567 kernel/sched/rt.c return rt_rq->rt_runtime; rt_runtime 632 kernel/sched/rt.c rt_rq->rt_time < rt_b->rt_runtime); rt_runtime 663 kernel/sched/rt.c if (iter->rt_runtime == RUNTIME_INF) rt_runtime 670 kernel/sched/rt.c diff = iter->rt_runtime - iter->rt_time; rt_runtime 673 kernel/sched/rt.c if (rt_rq->rt_runtime + diff > rt_period) rt_runtime 674 kernel/sched/rt.c diff = rt_period - rt_rq->rt_runtime; rt_runtime 675 kernel/sched/rt.c iter->rt_runtime -= diff; rt_runtime 676 kernel/sched/rt.c rt_rq->rt_runtime += diff; rt_runtime 677 kernel/sched/rt.c if (rt_rq->rt_runtime == rt_period) { rt_runtime 712 kernel/sched/rt.c if (rt_rq->rt_runtime == RUNTIME_INF || rt_runtime 713 kernel/sched/rt.c rt_rq->rt_runtime == rt_b->rt_runtime) rt_runtime 722 kernel/sched/rt.c want = rt_b->rt_runtime - rt_rq->rt_runtime; rt_runtime 734 kernel/sched/rt.c if (iter == rt_rq || iter->rt_runtime == RUNTIME_INF) rt_runtime 739 kernel/sched/rt.c diff = min_t(s64, iter->rt_runtime, want); rt_runtime 740 kernel/sched/rt.c iter->rt_runtime -= diff; rt_runtime 743 kernel/sched/rt.c iter->rt_runtime -= want; rt_runtime 763 kernel/sched/rt.c rt_rq->rt_runtime = RUNTIME_INF; rt_runtime 789 kernel/sched/rt.c rt_rq->rt_runtime = rt_b->rt_runtime; rt_runtime 802 kernel/sched/rt.c if (rt_rq->rt_time > rt_rq->rt_runtime) { rt_runtime 842 kernel/sched/rt.c if (!sched_feat(RT_RUNTIME_SHARE) && rt_rq->rt_runtime != RUNTIME_INF) rt_runtime 843 kernel/sched/rt.c rt_rq->rt_runtime = rt_b->rt_runtime; rt_runtime 858 kernel/sched/rt.c runtime = rt_rq->rt_runtime; rt_runtime 890 kernel/sched/rt.c if (!throttled && (!rt_bandwidth_enabled() || rt_b->rt_runtime == RUNTIME_INF)) rt_runtime 930 kernel/sched/rt.c if (likely(rt_b->rt_runtime)) { rt_runtime 2424 kernel/sched/rt.c u64 rt_runtime; rt_runtime 2435 kernel/sched/rt.c runtime = tg->rt_bandwidth.rt_runtime; rt_runtime 2439 kernel/sched/rt.c runtime = d->rt_runtime; rt_runtime 2467 kernel/sched/rt.c runtime = child->rt_bandwidth.rt_runtime; rt_runtime 2471 kernel/sched/rt.c runtime = d->rt_runtime; rt_runtime 2490 kernel/sched/rt.c .rt_runtime = runtime, rt_runtime 2501 kernel/sched/rt.c u64 rt_period, u64 rt_runtime) rt_runtime 2509 kernel/sched/rt.c if (tg == &root_task_group && rt_runtime == 0) rt_runtime 2518 kernel/sched/rt.c err = __rt_schedulable(tg, rt_period, rt_runtime); rt_runtime 2524 kernel/sched/rt.c tg->rt_bandwidth.rt_runtime = rt_runtime; rt_runtime 2530 kernel/sched/rt.c rt_rq->rt_runtime = rt_runtime; rt_runtime 2543 kernel/sched/rt.c u64 rt_runtime, rt_period; rt_runtime 2546 kernel/sched/rt.c rt_runtime = (u64)rt_runtime_us * NSEC_PER_USEC; rt_runtime 2548 kernel/sched/rt.c rt_runtime = RUNTIME_INF; rt_runtime 2552 kernel/sched/rt.c return tg_set_rt_bandwidth(tg, rt_period, rt_runtime); rt_runtime 2559 kernel/sched/rt.c if (tg->rt_bandwidth.rt_runtime == RUNTIME_INF) rt_runtime 2562 kernel/sched/rt.c rt_runtime_us = tg->rt_bandwidth.rt_runtime; rt_runtime 2569 kernel/sched/rt.c u64 rt_runtime, rt_period; rt_runtime 2575 kernel/sched/rt.c rt_runtime = tg->rt_bandwidth.rt_runtime; rt_runtime 2577 kernel/sched/rt.c return tg_set_rt_bandwidth(tg, rt_period, rt_runtime); rt_runtime 2605 kernel/sched/rt.c if (rt_task(tsk) && tg->rt_bandwidth.rt_runtime == 0) rt_runtime 2622 kernel/sched/rt.c rt_rq->rt_runtime = global_rt_runtime(); rt_runtime 2645 kernel/sched/rt.c def_rt_bandwidth.rt_runtime = global_rt_runtime(); rt_runtime 243 kernel/sched/sched.h u64 rt_runtime; rt_runtime 620 kernel/sched/sched.h u64 rt_runtime;