Lines Matching refs:this_rq

99 extern void calc_global_load_tick(struct rq *this_rq);
100 extern long calc_load_fold_active(struct rq *this_rq, long adjust);
1366 #define this_rq() this_cpu_ptr(&runqueues) macro
1649 rq = this_rq(); in this_rq_lock_irq()
2121 void (*task_woken)(struct rq *this_rq, struct task_struct *task);
2142 void (*switched_from)(struct rq *this_rq, struct task_struct *task);
2143 void (*switched_to) (struct rq *this_rq, struct task_struct *task);
2144 void (*prio_changed) (struct rq *this_rq, struct task_struct *task,
2516 static inline int _double_lock_balance(struct rq *this_rq, struct rq *busiest) in _double_lock_balance() argument
2517 __releases(this_rq->lock) in _double_lock_balance()
2519 __acquires(this_rq->lock) in _double_lock_balance()
2521 raw_spin_rq_unlock(this_rq); in _double_lock_balance()
2522 double_rq_lock(this_rq, busiest); in _double_lock_balance()
2535 static inline int _double_lock_balance(struct rq *this_rq, struct rq *busiest) in _double_lock_balance() argument
2536 __releases(this_rq->lock) in _double_lock_balance()
2538 __acquires(this_rq->lock) in _double_lock_balance()
2540 if (__rq_lockp(this_rq) == __rq_lockp(busiest)) in _double_lock_balance()
2546 if (rq_order_less(this_rq, busiest)) { in _double_lock_balance()
2551 raw_spin_rq_unlock(this_rq); in _double_lock_balance()
2552 double_rq_lock(this_rq, busiest); in _double_lock_balance()
2562 static inline int double_lock_balance(struct rq *this_rq, struct rq *busiest) in double_lock_balance() argument
2566 return _double_lock_balance(this_rq, busiest); in double_lock_balance()
2569 static inline void double_unlock_balance(struct rq *this_rq, struct rq *busiest) in double_unlock_balance() argument
2572 if (__rq_lockp(this_rq) != __rq_lockp(busiest)) in double_unlock_balance()
2574 lock_set_subclass(&__rq_lockp(this_rq)->dep_map, 0, _RET_IP_); in double_unlock_balance()