Searched refs:busiest (Results 1 – 4 of 4) sorted by relevance
6019 struct sched_group *busiest; /* Busiest group in this sd */ member6038 .busiest = NULL, in init_sd_lb_stats()6365 struct sg_lb_stats *busiest = &sds->busiest_stat; in update_sd_pick_busiest() local6367 if (sgs->group_type > busiest->group_type) in update_sd_pick_busiest()6370 if (sgs->group_type < busiest->group_type) in update_sd_pick_busiest()6373 if (sgs->avg_load <= busiest->avg_load) in update_sd_pick_busiest()6386 if (!sds->busiest) in update_sd_pick_busiest()6389 if (group_first_cpu(sds->busiest) > group_first_cpu(sg)) in update_sd_pick_busiest()6482 sds->busiest = sg; in update_sd_lb_stats()6535 if (!sds->busiest) in check_asym_packing()[all …]
1523 static inline int _double_lock_balance(struct rq *this_rq, struct rq *busiest) in _double_lock_balance() argument1525 __acquires(busiest->lock) in _double_lock_balance()1529 double_rq_lock(this_rq, busiest); in _double_lock_balance()1542 static inline int _double_lock_balance(struct rq *this_rq, struct rq *busiest) in _double_lock_balance() argument1544 __acquires(busiest->lock) in _double_lock_balance()1549 if (unlikely(!raw_spin_trylock(&busiest->lock))) { in _double_lock_balance()1550 if (busiest < this_rq) { in _double_lock_balance()1552 raw_spin_lock(&busiest->lock); in _double_lock_balance()1557 raw_spin_lock_nested(&busiest->lock, in _double_lock_balance()1568 static inline int double_lock_balance(struct rq *this_rq, struct rq *busiest) in double_lock_balance() argument[all …]
42 Initially, load_balance() finds the busiest group in the current sched domain.43 If it succeeds, it looks for the busiest runqueue of all the CPUs' runqueues in45 CPU's runqueue and the newly found busiest one and starts moving tasks from it
38 API-update-sd-pick-busiest43 API-find-busiest-group