LOAD_AVG_MAX     2026 kernel/sched/fair.c 		*period = LOAD_AVG_MAX;
LOAD_AVG_MAX     2893 kernel/sched/fair.c 		u32 divider = LOAD_AVG_MAX - 1024 + se->avg.period_contrib;
LOAD_AVG_MAX     3302 kernel/sched/fair.c 	se->avg.util_sum = se->avg.util_avg * LOAD_AVG_MAX;
LOAD_AVG_MAX     3306 kernel/sched/fair.c 	cfs_rq->avg.util_sum = cfs_rq->avg.util_avg * LOAD_AVG_MAX;
LOAD_AVG_MAX     3328 kernel/sched/fair.c 		runnable_sum = min(runnable_sum, (long)LOAD_AVG_MAX);
LOAD_AVG_MAX     3353 kernel/sched/fair.c 	load_avg = div_s64(load_sum, LOAD_AVG_MAX);
LOAD_AVG_MAX     3364 kernel/sched/fair.c 	runnable_load_avg = div_s64(runnable_load_sum, LOAD_AVG_MAX);
LOAD_AVG_MAX     3478 kernel/sched/fair.c 		u32 divider = LOAD_AVG_MAX - 1024 + sa->period_contrib;
LOAD_AVG_MAX     3521 kernel/sched/fair.c 	u32 divider = LOAD_AVG_MAX - 1024 + cfs_rq->avg.period_contrib;
LOAD_AVG_MAX       81 kernel/sched/pelt.c 	c2 = LOAD_AVG_MAX - decay_load(LOAD_AVG_MAX, periods) - 1024;
LOAD_AVG_MAX      229 kernel/sched/pelt.c 	u32 divider = LOAD_AVG_MAX - 1024 + sa->period_contrib;
LOAD_AVG_MAX       99 kernel/sched/pelt.h 	u32 divider = ((LOAD_AVG_MAX - 1024) << SCHED_CAPACITY_SHIFT) - LOAD_AVG_MAX;