Searched refs:this_rq (Results 1 – 9 of 9) sorted by relevance
/linux/kernel/sched/ |
A D | loadavg.c | 83 nr_active = this_rq->nr_running - adjust; in calc_load_fold_active() 84 nr_active += (int)this_rq->nr_uninterruptible; in calc_load_fold_active() 86 if (nr_active != this_rq->calc_load_active) { in calc_load_fold_active() 87 delta = nr_active - this_rq->calc_load_active; in calc_load_fold_active() 88 this_rq->calc_load_active = nr_active; in calc_load_fold_active() 252 calc_load_nohz_fold(this_rq()); in calc_load_nohz_start() 266 struct rq *this_rq = this_rq(); in calc_load_nohz_stop() local 281 this_rq->calc_load_update += LOAD_FREQ; in calc_load_nohz_stop() 386 void calc_global_load_tick(struct rq *this_rq) in calc_global_load_tick() argument 393 delta = calc_load_fold_active(this_rq, 0); in calc_global_load_tick() [all …]
|
A D | sched.h | 1366 #define this_rq() this_cpu_ptr(&runqueues) macro 1649 rq = this_rq(); in this_rq_lock_irq() 2517 __releases(this_rq->lock) in _double_lock_balance() 2519 __acquires(this_rq->lock) in _double_lock_balance() 2521 raw_spin_rq_unlock(this_rq); in _double_lock_balance() 2522 double_rq_lock(this_rq, busiest); in _double_lock_balance() 2536 __releases(this_rq->lock) in _double_lock_balance() 2538 __acquires(this_rq->lock) in _double_lock_balance() 2546 if (rq_order_less(this_rq, busiest)) { in _double_lock_balance() 2551 raw_spin_rq_unlock(this_rq); in _double_lock_balance() [all …]
|
A D | rt.c | 591 return this_rq()->rd->span; in sched_rt_period_mask() 2260 rq = this_rq(); in rto_push_irq_work_func() 2292 int this_cpu = this_rq->cpu, cpu; in pull_rt_task() 2309 cpumask_test_cpu(this_rq->cpu, this_rq->rd->rto_mask)) in pull_rt_task() 2314 tell_cpu_to_push(this_rq); in pull_rt_task() 2333 this_rq->rt.highest_prio.curr) in pull_rt_task() 2342 double_lock_balance(this_rq, src_rq); in pull_rt_task() 2374 activate_task(this_rq, p, 0); in pull_rt_task() 2388 raw_spin_rq_unlock(this_rq); in pull_rt_task() 2391 raw_spin_rq_lock(this_rq); in pull_rt_task() [all …]
|
A D | fair.c | 6242 struct rq *this_rq = this_rq(); in select_idle_cpu() local 6263 while (this_rq->wake_stamp < now && this_rq->wake_avg_idle) { in select_idle_cpu() 6264 this_rq->wake_stamp++; in select_idle_cpu() 6306 this_rq->wake_avg_idle -= min(this_rq->wake_avg_idle, time); in select_idle_cpu() 9859 .dst_rq = this_rq, in load_balance() 10872 if (this_rq->ttwu_pending) in newidle_balance() 10879 this_rq->idle_stamp = rq_clock(this_rq); in newidle_balance() 10948 raw_spin_rq_lock(this_rq); in newidle_balance() 10962 if (this_rq->nr_running != this_rq->cfs.h_nr_running) in newidle_balance() 10971 this_rq->idle_stamp = 0; in newidle_balance() [all …]
|
A D | deadline.c | 2322 int this_cpu = this_rq->cpu, cpu; in pull_dl_task() 2328 if (likely(!dl_overloaded(this_rq))) in pull_dl_task() 2347 if (this_rq->dl.dl_nr_running && in pull_dl_task() 2354 double_lock_balance(this_rq, src_rq); in pull_dl_task() 2371 (!this_rq->dl.dl_nr_running || in pull_dl_task() 2373 this_rq->dl.earliest_dl.curr))) { in pull_dl_task() 2390 activate_task(this_rq, p, 0); in pull_dl_task() 2398 double_unlock_balance(this_rq, src_rq); in pull_dl_task() 2401 raw_spin_rq_unlock(this_rq); in pull_dl_task() 2404 raw_spin_rq_lock(this_rq); in pull_dl_task() [all …]
|
A D | cputime.c | 222 struct rq *rq = this_rq(); in account_idle_time() 242 steal -= this_rq()->prev_steal_time; in steal_account_process_time() 245 this_rq()->prev_steal_time += steal; in steal_account_process_time() 385 } else if (p == this_rq()->idle) { in irqtime_account_process_tick() 491 else if ((p != this_rq()->idle) || (irq_count() != HARDIRQ_OFFSET)) in account_process_tick()
|
A D | idle.c | 22 idle_set_state(this_rq(), idle_state); in sched_idle_set_state()
|
A D | membarrier.c | 234 struct rq *rq = this_rq(); in membarrier_update_current_mm()
|
A D | core.c | 760 if (rq == this_rq()) in hrtick_start() 2161 this_rq()->nr_pinned++; in migrate_disable() 2190 this_rq()->nr_pinned--; in migrate_enable() 2318 struct rq *rq = this_rq(); in migration_cpu_stop() 3510 rq = this_rq(); in ttwu_stat() 3651 struct rq *rq = this_rq(); in sched_ttwu_pending() 4812 struct rq *rq = this_rq(); in finish_task_switch() 8282 rq = this_rq(); in yield_to() 8770 BUG_ON(current != this_rq()->idle); in idle_task_exit() 8783 struct rq *rq = this_rq(); in __balance_push_cpu_stop() [all …]
|
Completed in 61 milliseconds