/linux/drivers/xen/ |
A D | pcpu.c | 108 struct pcpu *pcpu = container_of(dev, struct pcpu, dev); in online_store() local 168 struct pcpu *pcpu) in pcpu_online_status() argument 185 struct pcpu *pcpu; in get_pcpu() local 197 struct pcpu *pcpu = container_of(dev, struct pcpu, dev); in pcpu_release() local 207 if (!pcpu) in unregister_and_remove_pcpu() 215 static int register_pcpu(struct pcpu *pcpu) in register_pcpu() argument 220 if (!pcpu) in register_pcpu() 240 struct pcpu *pcpu; in create_and_register_pcpu() local 246 pcpu = kzalloc(sizeof(struct pcpu), GFP_KERNEL); in create_and_register_pcpu() 272 struct pcpu *pcpu = NULL; in sync_pcpu() local [all …]
|
/linux/arch/s390/kernel/ |
A D | smp.c | 149 static inline int pcpu_stopped(struct pcpu *pcpu) in pcpu_stopped() argument 159 static inline int pcpu_running(struct pcpu *pcpu) in pcpu_running() argument 316 static void pcpu_delegate(struct pcpu *pcpu, in pcpu_delegate() argument 372 struct pcpu *pcpu; in smp_call_online_cpu() local 448 struct pcpu *pcpu = pcpu_devices + cpu; in smp_emergency_stop() local 614 struct pcpu *pcpu; in smp_store_status() local 773 struct pcpu *pcpu; in smp_add_core() local 922 struct pcpu *pcpu = pcpu_devices + cpu; in __cpu_up() local 977 struct pcpu *pcpu; in __cpu_die() local 1021 struct pcpu *pcpu = pcpu_devices; in smp_prepare_boot_cpu() local [all …]
|
/linux/arch/arm/common/ |
A D | mcpm_platsmp.c | 20 unsigned int *pcpu, unsigned int *pcluster) in cpu_to_pcpu() argument 25 *pcpu = MPIDR_AFFINITY_LEVEL(mpidr, 0); in cpu_to_pcpu() 31 unsigned int pcpu, pcluster, ret; in mcpm_boot_secondary() local 34 cpu_to_pcpu(cpu, &pcpu, &pcluster); in mcpm_boot_secondary() 37 __func__, cpu, pcpu, pcluster); in mcpm_boot_secondary() 39 mcpm_set_entry_vector(pcpu, pcluster, NULL); in mcpm_boot_secondary() 40 ret = mcpm_cpu_power_up(pcpu, pcluster); in mcpm_boot_secondary() 58 unsigned int pcpu, pcluster; in mcpm_cpu_kill() local 60 cpu_to_pcpu(cpu, &pcpu, &pcluster); in mcpm_cpu_kill() 73 unsigned int mpidr, pcpu, pcluster; in mcpm_cpu_die() local [all …]
|
/linux/drivers/gpu/drm/amd/display/amdgpu_dm/ |
A D | dc_fpu.c | 59 int *pcpu, depth = 0; in dc_assert_fp_enabled() local 61 pcpu = get_cpu_ptr(&fpu_recursion_depth); in dc_assert_fp_enabled() 62 depth = *pcpu; in dc_assert_fp_enabled() 83 int *pcpu; in dc_fpu_begin() local 85 pcpu = get_cpu_ptr(&fpu_recursion_depth); in dc_fpu_begin() 86 *pcpu += 1; in dc_fpu_begin() 88 if (*pcpu == 1) { in dc_fpu_begin() 121 int *pcpu; in dc_fpu_end() local 123 pcpu = get_cpu_ptr(&fpu_recursion_depth); in dc_fpu_end() 124 *pcpu -= 1; in dc_fpu_end() [all …]
|
/linux/arch/x86/events/amd/ |
A D | ibs.c | 405 struct cpu_perf_ibs *pcpu = this_cpu_ptr(perf_ibs->pcpu); in perf_ibs_start() local 436 struct cpu_perf_ibs *pcpu = this_cpu_ptr(perf_ibs->pcpu); in perf_ibs_stop() local 489 struct cpu_perf_ibs *pcpu = this_cpu_ptr(perf_ibs->pcpu); in perf_ibs_add() local 496 pcpu->event = event; in perf_ibs_add() 507 struct cpu_perf_ibs *pcpu = this_cpu_ptr(perf_ibs->pcpu); in perf_ibs_del() local 514 pcpu->event = NULL; in perf_ibs_del() 587 struct cpu_perf_ibs *pcpu = this_cpu_ptr(perf_ibs->pcpu); in perf_ibs_handle_irq() local 737 if (!pcpu) in perf_ibs_pmu_init() 740 perf_ibs->pcpu = pcpu; in perf_ibs_pmu_init() 755 perf_ibs->pcpu = NULL; in perf_ibs_pmu_init() [all …]
|
/linux/arch/powerpc/platforms/pseries/ |
A D | smp.c | 56 int smp_query_cpu_stopped(unsigned int pcpu) in smp_query_cpu_stopped() argument 67 status = rtas_call(qcss_tok, 1, 2, &cpu_status, pcpu); in smp_query_cpu_stopped() 93 unsigned int pcpu; in smp_startup_cpu() local 100 pcpu = get_hard_smp_processor_id(lcpu); in smp_startup_cpu() 103 if (smp_query_cpu_stopped(pcpu) == QCSS_NOT_STOPPED){ in smp_startup_cpu() 116 status = rtas_call(start_cpu, 3, 1, NULL, pcpu, start_here, pcpu); in smp_startup_cpu()
|
A D | hotplug-cpu.c | 117 unsigned int pcpu = get_hard_smp_processor_id(cpu); in pseries_cpu_die() local 121 cpu_status = smp_query_cpu_stopped(pcpu); in pseries_cpu_die() 128 cpu, pcpu); in pseries_cpu_die() 137 cpu, pcpu); in pseries_cpu_die()
|
/linux/kernel/ |
A D | padata.c | 54 target_cpu = cpumask_first(pd->cpumask.pcpu); in padata_index_to_cpu() 546 for_each_cpu(cpu, pd->cpumask.pcpu) { in padata_init_reorder_list() 577 cpumask_and(pd->cpumask.pcpu, pinst->cpumask.pcpu, cpu_online_mask); in padata_alloc_pd() 585 pd->cpu = cpumask_first(pd->cpumask.pcpu); in padata_alloc_pd() 591 free_cpumask_var(pd->cpumask.pcpu); in padata_alloc_pd() 604 free_cpumask_var(pd->cpumask.pcpu); in padata_free_pd() 732 parallel_mask = pinst->cpumask.pcpu; in padata_set_cpumask() 830 free_cpumask_var(pinst->cpumask.pcpu); in __padata_free() 865 cpumask = pinst->cpumask.pcpu; in show_cpumask() 992 free_cpumask_var(pinst->cpumask.pcpu); in padata_alloc() [all …]
|
A D | smp.c | 90 struct cfd_percpu __percpu *pcpu; member 113 cfd->pcpu = alloc_percpu(struct cfd_percpu); in smpcfd_prepare_cpu() 114 if (!cfd->pcpu) { in smpcfd_prepare_cpu() 129 free_percpu(cfd->pcpu); in smpcfd_dead_cpu() 290 struct cfd_percpu *pcpu = per_cpu_ptr(cfd->pcpu, cpu); in csd_lock_print_extended() local 419 struct cfd_percpu *pcpu = per_cpu_ptr(cfd->pcpu, cpu); in __smp_call_single_queue_debug() local 421 cfd_seq_store(pcpu->seq_queue, this_cpu, cpu, CFD_SEQ_QUEUE); in __smp_call_single_queue_debug() 423 cfd_seq_store(pcpu->seq_ipi, this_cpu, cpu, CFD_SEQ_IPI); in __smp_call_single_queue_debug() 914 struct cfd_percpu *pcpu = per_cpu_ptr(cfd->pcpu, cpu); in smp_call_function_many_cond() local 915 call_single_data_t *csd = &pcpu->csd; in smp_call_function_many_cond() [all …]
|
/linux/arch/alpha/kernel/ |
A D | err_common.c | 242 cdl_process_console_data_log(int cpu, struct percpu_struct *pcpu) in cdl_process_console_data_log() argument 245 (IDENT_ADDR | pcpu->console_data_log_pa); in cdl_process_console_data_log() 256 pcpu->console_data_log_pa = 0; in cdl_process_console_data_log() 266 struct percpu_struct *pcpu; in cdl_check_console_data_log() local 270 pcpu = (struct percpu_struct *) in cdl_check_console_data_log() 273 if (pcpu->console_data_log_pa) in cdl_check_console_data_log() 274 cdl_process_console_data_log(cpu, pcpu); in cdl_check_console_data_log()
|
/linux/arch/powerpc/platforms/powernv/ |
A D | smp.c | 66 unsigned int pcpu; in pnv_smp_kick_cpu() local 75 pcpu = get_hard_smp_processor_id(nr); in pnv_smp_kick_cpu() 89 rc = opal_query_cpu_status(pcpu, &status); in pnv_smp_kick_cpu() 106 pr_devel("OPAL: Starting CPU %d (HW 0x%x)...\n", nr, pcpu); in pnv_smp_kick_cpu() 107 rc = opal_start_cpu(pcpu, start_here); in pnv_smp_kick_cpu() 120 " (status %d)...\n", nr, pcpu, status); in pnv_smp_kick_cpu()
|
/linux/net/netfilter/ |
A D | nf_conntrack_ecache.c | 41 static enum retry_state ecache_work_evict_list(struct ct_pcpu *pcpu) in ecache_work_evict_list() argument 49 spin_lock(&pcpu->lock); in ecache_work_evict_list() 51 hlist_nulls_for_each_entry(h, n, &pcpu->dying, hnnode) { in ecache_work_evict_list() 88 spin_unlock(&pcpu->lock); in ecache_work_evict_list() 102 struct ct_pcpu *pcpu; in ecache_work() local 109 pcpu = per_cpu_ptr(ctnet->pcpu_lists, cpu); in ecache_work() 111 ret = ecache_work_evict_list(pcpu); in ecache_work()
|
A D | nf_conntrack_core.c | 518 struct ct_pcpu *pcpu; in nf_ct_add_to_dying_list() local 524 spin_lock(&pcpu->lock); in nf_ct_add_to_dying_list() 526 &pcpu->dying); in nf_ct_add_to_dying_list() 527 spin_unlock(&pcpu->lock); in nf_ct_add_to_dying_list() 533 struct ct_pcpu *pcpu; in nf_ct_add_to_unconfirmed_list() local 539 spin_lock(&pcpu->lock); in nf_ct_add_to_unconfirmed_list() 542 spin_unlock(&pcpu->lock); in nf_ct_add_to_unconfirmed_list() 548 struct ct_pcpu *pcpu; in nf_ct_del_from_dying_or_unconfirmed_list() local 553 spin_lock(&pcpu->lock); in nf_ct_del_from_dying_or_unconfirmed_list() 556 spin_unlock(&pcpu->lock); in nf_ct_del_from_dying_or_unconfirmed_list() [all …]
|
/linux/lib/ |
A D | test_vmalloc.c | 286 void __percpu **pcpu; in pcpu_alloc_test() local 290 pcpu = vmalloc(sizeof(void __percpu *) * 35000); in pcpu_alloc_test() 291 if (!pcpu) in pcpu_alloc_test() 306 pcpu[i] = __alloc_percpu(size, align); in pcpu_alloc_test() 307 if (!pcpu[i]) in pcpu_alloc_test() 312 free_percpu(pcpu[i]); in pcpu_alloc_test() 314 vfree(pcpu); in pcpu_alloc_test()
|
/linux/arch/powerpc/platforms/cell/ |
A D | smp.c | 72 unsigned int pcpu; in smp_startup_cpu() local 79 pcpu = get_hard_smp_processor_id(lcpu); in smp_startup_cpu() 89 status = rtas_call(start_cpu, 3, 1, NULL, pcpu, start_here, lcpu); in smp_startup_cpu()
|
/linux/arch/powerpc/kvm/ |
A D | book3s_hv_builtin.c | 236 int cpu = vc->pcpu; in kvmhv_interrupt_vcore() 715 void kvmppc_check_need_tlb_flush(struct kvm *kvm, int pcpu, in kvmppc_check_need_tlb_flush() argument 727 pcpu = cpu_first_tlb_thread_sibling(pcpu); in kvmppc_check_need_tlb_flush() 734 if (cpumask_test_cpu(pcpu, need_tlb_flush)) { in kvmppc_check_need_tlb_flush() 738 cpumask_clear_cpu(pcpu, need_tlb_flush); in kvmppc_check_need_tlb_flush()
|
A D | book3s_hv_rm_xics.c | 713 int pcpu = raw_smp_processor_id(); in ics_rm_eoi() local 715 pcpu = cpu_first_thread_sibling(pcpu); in ics_rm_eoi() 717 if (state->intr_cpu != pcpu) { in ics_rm_eoi() 719 xics_opal_set_server(state->host_irq, pcpu); in ics_rm_eoi()
|
A D | book3s_hv.c | 1353 pcpu = READ_ONCE(v->cpu); in kvmppc_read_dpdes() 1354 if (pcpu >= 0) in kvmppc_read_dpdes() 3000 if (prev_cpu != pcpu) { in kvmppc_prepare_radix_vcpu() 3018 cpu = vc->pcpu; in kvmppc_start_thread() 3025 vcpu->cpu = vc->pcpu; in kvmppc_start_thread() 3465 int pcpu, thr; in kvmppc_run_core() local 3517 pcpu = smp_processor_id(); in kvmppc_run_core() 3635 pvc->pcpu = pcpu + thr; in kvmppc_run_core() 4421 int trap, r, pcpu; in kvmhv_run_single_vcpu() local 4464 pcpu = smp_processor_id(); in kvmhv_run_single_vcpu() [all …]
|
/linux/arch/s390/include/asm/ |
A D | smp.h | 44 u16 pcpu = stap(); in smp_stop_cpu() local 47 __pcpu_sigp(pcpu, SIGP_STOP, 0, NULL); in smp_stop_cpu()
|
/linux/tools/testing/selftests/kvm/aarch64/ |
A D | arch_timer.c | 253 uint32_t pcpu; in test_get_pcpu() local 262 pcpu = rand() % nproc_conf; in test_get_pcpu() 263 } while (!CPU_ISSET(pcpu, &online_cpuset)); in test_get_pcpu() 265 return pcpu; in test_get_pcpu()
|
/linux/drivers/crypto/caam/ |
A D | qi.c | 436 int *pcpu = &get_cpu_var(last_cpu); in caam_drv_ctx_init() local 438 *pcpu = cpumask_next(*pcpu, cpus); in caam_drv_ctx_init() 439 if (*pcpu >= nr_cpu_ids) in caam_drv_ctx_init() 440 *pcpu = cpumask_first(cpus); in caam_drv_ctx_init() 441 *cpu = *pcpu; in caam_drv_ctx_init()
|
/linux/kernel/sched/ |
A D | psi.c | 181 .pcpu = &system_group_pcpu, 193 seqcount_init(&per_cpu_ptr(group->pcpu, cpu)->seq); in group_init() 253 struct psi_group_cpu *groupc = per_cpu_ptr(group->pcpu, cpu); in get_recent_times() 693 groupc = per_cpu_ptr(group->pcpu, cpu); in psi_group_change() 844 per_cpu_ptr(group->pcpu, cpu)->tasks[NR_ONCPU]) { in psi_task_switch() 951 cgroup->psi.pcpu = alloc_percpu(struct psi_group_cpu); in psi_cgroup_alloc() 952 if (!cgroup->psi.pcpu) in psi_cgroup_alloc() 964 free_percpu(cgroup->psi.pcpu); in psi_cgroup_free()
|
/linux/drivers/base/ |
A D | devres.c | 1190 void __percpu *pcpu; in __devm_alloc_percpu() local 1192 pcpu = __alloc_percpu(size, align); in __devm_alloc_percpu() 1193 if (!pcpu) in __devm_alloc_percpu() 1198 free_percpu(pcpu); in __devm_alloc_percpu() 1202 *(void __percpu **)p = pcpu; in __devm_alloc_percpu() 1206 return pcpu; in __devm_alloc_percpu()
|
/linux/include/linux/ |
A D | notifier.h | 106 #define SRCU_NOTIFIER_INIT(name, pcpu) \ argument 110 .srcu = __SRCU_STRUCT_INIT(name.srcu, pcpu), \
|
A D | cpufreq.h | 1125 static inline int of_perf_domain_get_sharing_cpumask(int pcpu, const char *list_name, in of_perf_domain_get_sharing_cpumask() argument 1131 ret = parse_perf_domain(pcpu, list_name, cell_name); in of_perf_domain_get_sharing_cpumask() 1136 cpumask_set_cpu(pcpu, cpumask); in of_perf_domain_get_sharing_cpumask() 1139 if (cpu == pcpu) in of_perf_domain_get_sharing_cpumask() 1179 static inline int of_perf_domain_get_sharing_cpumask(int pcpu, const char *list_name, in of_perf_domain_get_sharing_cpumask() argument
|