Home
last modified time | relevance | path

Searched refs:cpu_rq (Results 1 – 13 of 13) sorted by relevance

/linux/kernel/sched/
A Dcpuacct.c115 raw_spin_rq_lock_irq(cpu_rq(cpu)); in cpuacct_cpuusage_read()
129 raw_spin_rq_unlock_irq(cpu_rq(cpu)); in cpuacct_cpuusage_read()
144 raw_spin_rq_lock_irq(cpu_rq(cpu)); in cpuacct_cpuusage_write()
151 raw_spin_rq_unlock_irq(cpu_rq(cpu)); in cpuacct_cpuusage_write()
256 raw_spin_rq_lock_irq(cpu_rq(cpu)); in cpuacct_all_seq_show()
262 raw_spin_rq_unlock_irq(cpu_rq(cpu)); in cpuacct_all_seq_show()
A Dmembarrier.c277 if (!(READ_ONCE(cpu_rq(cpu)->membarrier_state) & in membarrier_global_expedited()
285 p = rcu_dereference(cpu_rq(cpu)->curr); in membarrier_global_expedited()
357 p = rcu_dereference(cpu_rq(cpu_id)->curr); in membarrier_private_expedited()
370 p = rcu_dereference(cpu_rq(cpu)->curr); in membarrier_private_expedited()
465 struct rq *rq = cpu_rq(cpu); in sync_runqueues_membarrier_state()
A Dcore.c2258 rq = cpu_rq(new_cpu); in move_queued_task()
5754 rq_i = cpu_rq(i); in pick_next_task()
5771 rq_i = cpu_rq(i); in pick_next_task()
5811 rq_i = cpu_rq(i); in pick_next_task()
5859 struct rq *dst = cpu_rq(this), *src = cpu_rq(that); in try_steal_cookie()
5981 rq = cpu_rq(t); in sched_core_cpu_starting()
5993 rq = cpu_rq(t); in sched_core_cpu_starting()
6028 core_rq = cpu_rq(t); in sched_core_cpu_deactivate()
6044 rq = cpu_rq(t); in sched_core_cpu_deactivate()
6141 rq = cpu_rq(cpu); in __schedule()
[all …]
A Dfair.c1615 rq = cpu_rq(env->dst_cpu); in task_numa_assign()
1630 rq = cpu_rq(env->best_cpu); in task_numa_assign()
6000 struct rq *rq = cpu_rq(i); in find_idlest_group_cpu()
6502 cfs_rq = &cpu_rq(cpu)->cfs; in cpu_util()
6533 cfs_rq = &cpu_rq(cpu)->cfs; in cpu_util_without()
8681 struct rq *rq = cpu_rq(i); in update_sg_lb_stats()
8955 struct rq *rq = cpu_rq(i); in update_sg_wakeup_stats()
9621 rq = cpu_rq(i); in find_busiest_queue()
10728 rq = cpu_rq(balance_cpu); in _nohz_idle_balance()
11519 rq = cpu_rq(i); in online_fair_sched_group()
[all …]
A Ddeadline.c73 return &cpu_rq(i)->rd->dl_bw; in dl_bw_of()
78 struct root_domain *rd = cpu_rq(i)->rd; in dl_bw_cpus()
97 struct root_domain *rd = cpu_rq(i)->rd; in __dl_bw_capacity()
125 struct root_domain *rd = cpu_rq(cpu)->rd; in dl_bw_visited()
136 return &cpu_rq(i)->dl.dl_bw; in dl_bw_of()
622 later_rq = cpu_rq(cpu); in dl_task_offline_migration()
1764 rq = cpu_rq(cpu); in select_task_rq_dl()
2159 later_rq = cpu_rq(cpu); in find_lock_later_rq()
2341 src_rq = cpu_rq(cpu); in pull_dl_task()
2755 init_dl_rq_bw_ratio(&cpu_rq(cpu)->dl); in sched_dl_do_global()
[all …]
A Ddebug.c585 struct rq *rq = cpu_rq(cpu); in print_cfs_rq()
606 rq0_min_vruntime = cpu_rq(0)->cfs.min_vruntime; in print_cfs_rq()
706 dl_bw = &cpu_rq(cpu)->rd->dl_bw; in print_dl_rq()
718 struct rq *rq = cpu_rq(cpu); in print_cpu()
1085 cpu, latency, cpu_rq(cpu)->ticks_without_resched); in resched_latency_warn()
A Dsched.h1365 #define cpu_rq(cpu) (&per_cpu(runqueues, (cpu))) macro
1367 #define task_rq(p) cpu_rq(task_cpu(p))
1368 #define cpu_curr(cpu) (cpu_rq(cpu)->curr)
1726 for (__sd = rcu_dereference_check_sched_domain(cpu_rq(cpu)->sd); \
2709 #define nohz_flags(cpu) (&cpu_rq(cpu)->nohz_flags)
2732 struct rq *rq = cpu_rq(i); in __dl_update()
2909 return cpu_rq(cpu)->cpu_capacity_orig; in capacity_orig_of()
A Drt.c166 struct rq *rq = cpu_rq(cpu); in init_tg_rt_entry()
665 return &cpu_rq(cpu)->rt; in sched_rt_period_rt_rq()
1564 rq = cpu_rq(cpu); in select_task_rq_rt()
1614 p->prio < cpu_rq(target)->rt.highest_prio.curr) in select_task_rq_rt()
1937 lowest_rq = cpu_rq(cpu); in find_lock_lowest_rq()
2323 src_rq = cpu_rq(cpu); in pull_rt_task()
2886 struct rt_rq *rt_rq = &cpu_rq(i)->rt; in sched_rt_global_constraints()
2986 for_each_rt_rq(rt_rq, iter, cpu_rq(cpu)) in print_rt_stats()
A Dstats.c133 rq = cpu_rq(cpu); in show_schedstat()
A Dcpufreq_schedutil.c166 struct rq *rq = cpu_rq(sg_cpu->cpu); in sugov_get_util()
315 if (cpu_bw_dl(cpu_rq(sg_cpu->cpu)) > sg_cpu->bw_dl) in ignore_dl_rate_limit()
A Dtopology.c357 struct root_domain *rd = cpu_rq(cpu)->rd; in build_perf_domains()
689 struct rq *rq = cpu_rq(cpu); in cpu_attach_domain()
2259 rq = cpu_rq(i); in build_sched_domains()
2462 rd = cpu_rq(cpumask_any(doms_cur[i]))->rd; in partition_sched_domains_locked()
2499 cpu_rq(cpumask_first(doms_cur[j]))->rd->pd) { in partition_sched_domains_locked()
A Dcputime.c973 rq = cpu_rq(cpu); in kcpustat_field()
1060 rq = cpu_rq(cpu); in kcpustat_cpu_fetch()
/linux/tools/perf/Documentation/
A Dperf-probe.txt234 … be moved easily by modifying schedule(), but the same line matching 'rq=cpu_rq*' may still exist …

Completed in 70 milliseconds