Lines Matching refs:period
93 void start_bandwidth_timer(struct hrtimer *period_timer, ktime_t period) in start_bandwidth_timer() argument
103 hrtimer_forward(period_timer, now, period); in start_bandwidth_timer()
726 s64 period = sched_avg_period(); in sched_avg_update() local
728 while ((s64)(rq_clock(rq) - rq->age_stamp) > period) { in sched_avg_update()
735 rq->age_stamp += period; in sched_avg_update()
1971 unsigned long to_ratio(u64 period, u64 runtime) in to_ratio() argument
1981 if (period == 0) in to_ratio()
1984 return div64_u64(runtime << 20, period); in to_ratio()
2035 u64 period = attr->sched_period ?: attr->sched_deadline; in dl_overflow() local
2037 u64 new_bw = dl_policy(policy) ? to_ratio(period, runtime) : 0; in dl_overflow()
7619 u64 period, runtime; in tg_rt_schedulable() local
7621 period = ktime_to_ns(tg->rt_bandwidth.rt_period); in tg_rt_schedulable()
7625 period = d->rt_period; in tg_rt_schedulable()
7632 if (runtime > period && runtime != RUNTIME_INF) in tg_rt_schedulable()
7641 total = to_ratio(period, runtime); in tg_rt_schedulable()
7653 period = ktime_to_ns(child->rt_bandwidth.rt_period); in tg_rt_schedulable()
7657 period = d->rt_period; in tg_rt_schedulable()
7661 sum += to_ratio(period, runtime); in tg_rt_schedulable()
7670 static int __rt_schedulable(struct task_group *tg, u64 period, u64 runtime) in __rt_schedulable() argument
7676 .rt_period = period, in __rt_schedulable()
7818 u64 period = global_rt_period(); in sched_dl_global_validate() local
7819 u64 new_bw = to_ratio(period, runtime); in sched_dl_global_validate()
8073 static int __cfs_schedulable(struct task_group *tg, u64 period, u64 runtime);
8075 static int tg_set_cfs_bandwidth(struct task_group *tg, u64 period, u64 quota) in tg_set_cfs_bandwidth() argument
8088 if (quota < min_cfs_quota_period || period < min_cfs_quota_period) in tg_set_cfs_bandwidth()
8096 if (period > max_cfs_quota_period) in tg_set_cfs_bandwidth()
8105 ret = __cfs_schedulable(tg, period, quota); in tg_set_cfs_bandwidth()
8118 cfs_b->period = ns_to_ktime(period); in tg_set_cfs_bandwidth()
8152 u64 quota, period; in tg_set_cfs_quota() local
8154 period = ktime_to_ns(tg->cfs_bandwidth.period); in tg_set_cfs_quota()
8160 return tg_set_cfs_bandwidth(tg, period, quota); in tg_set_cfs_quota()
8178 u64 quota, period; in tg_set_cfs_period() local
8180 period = (u64)cfs_period_us * NSEC_PER_USEC; in tg_set_cfs_period()
8183 return tg_set_cfs_bandwidth(tg, period, quota); in tg_set_cfs_period()
8190 cfs_period_us = ktime_to_ns(tg->cfs_bandwidth.period); in tg_get_cfs_period()
8222 u64 period, quota; member
8232 u64 quota, period; in normalize_cfs_quota() local
8235 period = d->period; in normalize_cfs_quota()
8238 period = tg_get_cfs_period(tg); in normalize_cfs_quota()
8246 return to_ratio(period, quota); in normalize_cfs_quota()
8277 static int __cfs_schedulable(struct task_group *tg, u64 period, u64 quota) in __cfs_schedulable() argument
8282 .period = period, in __cfs_schedulable()
8287 do_div(data.period, NSEC_PER_USEC); in __cfs_schedulable()