rt_nr_running 685 kernel/sched/core.c fifo_nr_running = rq->rt.rt_nr_running - rq->rt.rr_nr_running; rt_nr_running 578 kernel/sched/debug.c PU(rt_nr_running); rt_nr_running 496 kernel/sched/rt.c if (rt_rq->rt_nr_running) { rt_nr_running 592 kernel/sched/rt.c if (!rt_rq->rt_nr_running) rt_nr_running 844 kernel/sched/rt.c skip = !rt_rq->rt_time && !rt_rq->rt_nr_running; rt_nr_running 871 kernel/sched/rt.c if (rt_rq->rt_nr_running && rq->curr == rq->idle) rt_nr_running 874 kernel/sched/rt.c if (rt_rq->rt_time || rt_rq->rt_nr_running) rt_nr_running 877 kernel/sched/rt.c } else if (rt_rq->rt_nr_running) { rt_nr_running 1007 kernel/sched/rt.c sub_nr_running(rq, rt_rq->rt_nr_running); rt_nr_running 1025 kernel/sched/rt.c if (rt_rq->rt_nr_running) { rt_nr_running 1026 kernel/sched/rt.c add_nr_running(rq, rt_rq->rt_nr_running); rt_nr_running 1094 kernel/sched/rt.c if (rt_rq->rt_nr_running) { rt_nr_running 1140 kernel/sched/rt.c WARN_ON(!rt_rq->rt_nr_running && rt_rq->rt_nr_boosted); rt_nr_running 1162 kernel/sched/rt.c return group_rq->rt_nr_running; rt_nr_running 1187 kernel/sched/rt.c rt_rq->rt_nr_running += rt_se_nr_running(rt_se); rt_nr_running 1199 kernel/sched/rt.c WARN_ON(!rt_rq->rt_nr_running); rt_nr_running 1200 kernel/sched/rt.c rt_rq->rt_nr_running -= rt_se_nr_running(rt_se); rt_nr_running 1244 kernel/sched/rt.c if (group_rq && (rt_rq_throttled(group_rq) || !group_rq->rt_nr_running)) { rt_nr_running 1319 kernel/sched/rt.c if (rt_rq && rt_rq->rt_nr_running) rt_nr_running 2197 kernel/sched/rt.c if (!task_on_rq_queued(p) || rq->rt.rt_nr_running) rt_nr_running 599 kernel/sched/sched.h unsigned int rt_nr_running; rt_nr_running 634 kernel/sched/sched.h return rt_rq->rt_queued && rt_rq->rt_nr_running;