LOAD_AVG_MAX 2026 kernel/sched/fair.c *period = LOAD_AVG_MAX; LOAD_AVG_MAX 2893 kernel/sched/fair.c u32 divider = LOAD_AVG_MAX - 1024 + se->avg.period_contrib; LOAD_AVG_MAX 3302 kernel/sched/fair.c se->avg.util_sum = se->avg.util_avg * LOAD_AVG_MAX; LOAD_AVG_MAX 3306 kernel/sched/fair.c cfs_rq->avg.util_sum = cfs_rq->avg.util_avg * LOAD_AVG_MAX; LOAD_AVG_MAX 3328 kernel/sched/fair.c runnable_sum = min(runnable_sum, (long)LOAD_AVG_MAX); LOAD_AVG_MAX 3353 kernel/sched/fair.c load_avg = div_s64(load_sum, LOAD_AVG_MAX); LOAD_AVG_MAX 3364 kernel/sched/fair.c runnable_load_avg = div_s64(runnable_load_sum, LOAD_AVG_MAX); LOAD_AVG_MAX 3478 kernel/sched/fair.c u32 divider = LOAD_AVG_MAX - 1024 + sa->period_contrib; LOAD_AVG_MAX 3521 kernel/sched/fair.c u32 divider = LOAD_AVG_MAX - 1024 + cfs_rq->avg.period_contrib; LOAD_AVG_MAX 81 kernel/sched/pelt.c c2 = LOAD_AVG_MAX - decay_load(LOAD_AVG_MAX, periods) - 1024; LOAD_AVG_MAX 229 kernel/sched/pelt.c u32 divider = LOAD_AVG_MAX - 1024 + sa->period_contrib; LOAD_AVG_MAX 99 kernel/sched/pelt.h u32 divider = ((LOAD_AVG_MAX - 1024) << SCHED_CAPACITY_SHIFT) - LOAD_AVG_MAX;