Searched refs:busiest (Results 1 – 3 of 3) sorted by relevance
8319 .busiest = NULL, in init_sd_lb_stats()9400 busiest->sum_nr_running, busiest->group_weight); in calculate_imbalance()9441 (busiest->avg_load - sds->avg_load) * busiest->group_capacity, in calculate_imbalance()9503 if (!sds.busiest) in find_busiest_group()9697 busiest = rq; in find_busiest_queue()9714 busiest = rq; in find_busiest_queue()9721 busiest = rq; in find_busiest_queue()9732 busiest = rq; in find_busiest_queue()9740 return busiest; in find_busiest_queue()9852 struct rq *busiest; in load_balance() local[all …]
2518 __acquires(busiest->lock) in _double_lock_balance()2522 double_rq_lock(this_rq, busiest); in _double_lock_balance()2537 __acquires(busiest->lock) in _double_lock_balance()2540 if (__rq_lockp(this_rq) == __rq_lockp(busiest)) in _double_lock_balance()2543 if (likely(raw_spin_rq_trylock(busiest))) in _double_lock_balance()2546 if (rq_order_less(this_rq, busiest)) { in _double_lock_balance()2552 double_rq_lock(this_rq, busiest); in _double_lock_balance()2566 return _double_lock_balance(this_rq, busiest); in double_lock_balance()2570 __releases(busiest->lock) in double_unlock_balance()2572 if (__rq_lockp(this_rq) != __rq_lockp(busiest)) in double_unlock_balance()[all …]
48 Initially, load_balance() finds the busiest group in the current sched domain.49 If it succeeds, it looks for the busiest runqueue of all the CPUs' runqueues in51 CPU's runqueue and the newly found busiest one and starts moving tasks from it
Completed in 31 milliseconds