/xen/tools/debugger/kdd/ |
A D | kdd-xen.c | 175 if ((void *) cpu + sizeof (*cpu) <= g->hvm_buf + g->hvm_sz) in get_cpu() 333 if (!cpu) in kdd_get_regs() 351 if (!cpu) in kdd_set_regs() 367 c->cr0 = cpu->cr0; in kdd_get_ctrl_x86_32() 368 c->cr2 = cpu->cr2; in kdd_get_ctrl_x86_32() 369 c->cr3 = cpu->cr3; in kdd_get_ctrl_x86_32() 370 c->cr4 = cpu->cr4; in kdd_get_ctrl_x86_32() 371 c->dr0 = cpu->dr0; in kdd_get_ctrl_x86_32() 415 if (!cpu) in kdd_get_ctrl() 435 if (!cpu) in kdd_wrmsr() [all …]
|
/xen/xen/arch/x86/ |
A D | smpboot.c | 253 cpumask_set_cpu(cpu, socket_cpumask[cpu_to_socket(cpu)]); in set_cpu_sibling_map() 254 cpumask_set_cpu(cpu, per_cpu(cpu_core_mask, cpu)); in set_cpu_sibling_map() 255 cpumask_set_cpu(cpu, per_cpu(cpu_sibling_mask, cpu)); in set_cpu_sibling_map() 314 else if ( (i != cpu) && !c[cpu].booted_cores ) in set_cpu_sibling_map() 535 int cpu; in alloc_cpu_id() local 539 return (cpu < nr_cpu_ids) ? cpu : -ENODEV; in alloc_cpu_id() 1047 stub_page = alloc_stub_page(cpu, &per_cpu(stubs.mfn, cpu)); in cpu_smpboot_alloc() 1218 cpumask_clear_cpu(cpu, socket_cpumask[cpu_to_socket(cpu)]); in remove_siblinginfo() 1276 int cpu = -1; in cpu_add() local 1331 cpu); in cpu_add() [all …]
|
A D | percpu.c | 20 unsigned int cpu; in percpu_init_areas() local 22 for ( cpu = 1; cpu < NR_CPUS; cpu++ ) in percpu_init_areas() 26 static int init_percpu_area(unsigned int cpu) in init_percpu_area() argument 43 unsigned int cpu; member 51 unsigned int cpu = info->cpu; in _free_percpu_area() local 58 static void free_percpu_area(unsigned int cpu) in free_percpu_area() argument 62 info->cpu = cpu; in free_percpu_area() 69 unsigned int cpu = (unsigned long)hcpu; in cpu_percpu_callback() local 75 rc = init_percpu_area(cpu); in cpu_percpu_callback() 81 free_percpu_area(cpu); in cpu_percpu_callback() [all …]
|
/xen/xen/include/xen/ |
A D | cpumask.h | 84 return cpu; in cpumask_check() 230 cpu = cpumask_next(cpu, srcp)) in cpumask_last() 231 pcpu = cpu; in cpumask_last() 267 return cpu; in cpumask_any() 375 (cpu) = cpumask_next(cpu, mask)) 378 for ((cpu) = 0; (cpu) < 1; (cpu)++, (void)(mask)) 445 #define cpu_online(cpu) cpumask_test_cpu(cpu, &cpu_online_map) argument 447 #define cpu_present(cpu) cpumask_test_cpu(cpu, &cpu_present_map) argument 452 #define cpu_online(cpu) ((cpu) == 0) argument 453 #define cpu_possible(cpu) ((cpu) == 0) argument [all …]
|
A D | irq_cpustat.h | 22 #define __IRQ_STAT(cpu, member) (irq_stat[cpu].member) argument 25 #define softirq_pending(cpu) __IRQ_STAT((cpu), __softirq_pending) argument 26 #define local_irq_count(cpu) __IRQ_STAT((cpu), __local_irq_count) argument 27 #define mwait_wakeup(cpu) __IRQ_STAT((cpu), __mwait_wakeup) argument
|
/xen/xen/common/ |
A D | cpu.c | 107 if ( (cpu >= nr_cpu_ids) || (cpu == 0) ) in cpu_down() 111 if ( !cpu_online(cpu) ) in cpu_down() 123 __cpu_die(cpu); in cpu_down() 124 err = cpu_online(cpu); in cpu_down() 140 int cpu_up(unsigned int cpu) in cpu_up() argument 148 if ( (cpu >= nr_cpu_ids) || !cpu_present(cpu) ) in cpu_up() 152 if ( cpu_online(cpu) ) in cpu_up() 159 err = __cpu_up(cpu); in cpu_up() 186 int cpu, error = 0; in disable_nonboot_cpus() local 196 if ( cpu == 0 ) in disable_nonboot_cpus() [all …]
|
A D | core_parking.c | 58 unsigned int cpu = -1; in core_parking_performance() local 70 for_each_cpu(cpu, &cpu_online_map) in core_parking_performance() 72 if ( cpu == 0 ) in core_parking_performance() 111 return cpu; in core_parking_performance() 116 unsigned int cpu = -1; in core_parking_power() local 130 if ( cpu == 0 ) in core_parking_power() 169 return cpu; in core_parking_power() 175 unsigned int cpu; in core_parking_helper() local 184 ret = cpu_down(cpu); in core_parking_helper() 197 ret = cpu_up(cpu); in core_parking_helper() [all …]
|
A D | softirq.c | 30 unsigned int i, cpu; in __do_softirq() local 42 cpu = smp_processor_id(); in __do_softirq() 45 rcu_check_callbacks(cpu); in __do_softirq() 48 || cpu_is_offline(cpu) ) in __do_softirq() 52 clear_bit(i, &softirq_pending(cpu)); in __do_softirq() 96 for_each_cpu(cpu, mask) in cpumask_raise_softirq() 98 cpu != this_cpu && in cpumask_raise_softirq() 111 || (cpu == this_cpu) in cpu_raise_softirq() 116 smp_send_event_check_cpu(cpu); in cpu_raise_softirq() 132 for_each_cpu ( cpu, mask ) in cpu_raise_softirq_batch_finish() [all …]
|
A D | tasklet.c | 35 unsigned int cpu = t->scheduled_on; in tasklet_enqueue() local 62 t->scheduled_on = cpu; in tasklet_schedule_on_cpu() 109 unsigned int cpu = smp_processor_id(); in do_tasklet() local 122 ASSERT(tasklet_work_to_do(cpu)); in do_tasklet() 126 do_tasklet_work(cpu, list); in do_tasklet() 140 unsigned int cpu = smp_processor_id(); in tasklet_softirq_action() local 145 do_tasklet_work(cpu, list); in tasklet_softirq_action() 193 BUG_ON(t->scheduled_on != cpu); in migrate_tasklets_from_cpu() 220 unsigned int cpu = (unsigned long)hcpu; in cpu_callback() local 230 migrate_tasklets_from_cpu(cpu, &per_cpu(tasklet_list, cpu)); in cpu_callback() [all …]
|
A D | perfc.c | 40 unsigned int k, cpu; in perfc_printall() local 48 for_each_online_cpu ( cpu ) in perfc_printall() 56 for_each_online_cpu ( cpu ) in perfc_printall() 60 printk(" CPU%02u[%10"PRIperfc"u]", cpu, per_cpu(perfcounters, cpu)[j]); in perfc_printall() 68 for_each_online_cpu ( cpu ) in perfc_printall() 84 for_each_online_cpu ( cpu ) in perfc_printall() 94 for_each_online_cpu ( cpu ) in perfc_printall() 130 unsigned int cpu; in perfc_reset() local 135 for_each_online_cpu ( cpu ) in perfc_reset() 141 for_each_online_cpu ( cpu ) in perfc_reset() [all …]
|
/xen/xen/arch/arm/arm64/ |
A D | smpboot.c | 19 static int __init smp_spin_table_cpu_up(int cpu) in smp_spin_table_cpu_up() argument 23 if (!cpu_release_addr[cpu]) in smp_spin_table_cpu_up() 25 printk("CPU%d: No release addr\n", cpu); in smp_spin_table_cpu_up() 56 static int __init smp_psci_init(int cpu) in smp_psci_init() argument 86 smp_spin_table_init(cpu, dn); in dt_arch_cpu_init() 88 return smp_psci_init(cpu); in dt_arch_cpu_init() 101 return dt_arch_cpu_init(cpu, dn); in arch_cpu_init() 104 return smp_psci_init(cpu); in arch_cpu_init() 107 int arch_cpu_up(int cpu) in arch_cpu_up() argument 109 if ( !smp_enable_ops[cpu].prepare_cpu ) in arch_cpu_up() [all …]
|
/xen/tools/xentrace/ |
A D | formats | 44 0x00022005 CPU%(cpu)d %(tsc)d (+%(reltsc)8d) csched:picked_cpu [ dom:vcpu = 0x%(1)04x%(2)04x,… 45 0x00022006 CPU%(cpu)d %(tsc)d (+%(reltsc)8d) csched:tickle [ cpu = %(1)d ] 48 0x00022009 CPU%(cpu)d %(tsc)d (+%(reltsc)8d) csched:schedule [ cpu[16]:tasklet[8]:idle[8] =… 57 0x00022206 CPU%(cpu)d %(tsc)d (+%(reltsc)8d) csched2:tickle [ cpu = %(1)d ] 75 0x00022801 CPU%(cpu)d %(tsc)d (+%(reltsc)8d) rtds:tickle [ cpu = %(1)d ] 80 0x00022806 CPU%(cpu)d %(tsc)d (+%(reltsc)8d) rtds:schedule [ cpu[16]:tasklet[8]:idle[4]:tic… 83 0x00022A02 CPU%(cpu)d %(tsc)d (+%(reltsc)8d) null:assign [ dom:vcpu = 0x%(1)08x, cpu = %(… 84 0x00022A03 CPU%(cpu)d %(tsc)d (+%(reltsc)8d) null:deassign [ dom:vcpu = 0x%(1)08x, cpu = %(… 85 0x00022A04 CPU%(cpu)d %(tsc)d (+%(reltsc)8d) null:migrate [ dom:vcpu = 0x%(1)08x, new_cpu:… 86 0x00022A05 CPU%(cpu)d %(tsc)d (+%(reltsc)8d) null:schedule [ cpu[16]:tasklet[16] = %(1)08x,… [all …]
|
/xen/xen/arch/arm/ |
A D | percpu.c | 13 unsigned int cpu; in percpu_init_areas() local 14 for ( cpu = 1; cpu < NR_CPUS; cpu++ ) in percpu_init_areas() 18 static int init_percpu_area(unsigned int cpu) in init_percpu_area() argument 26 __per_cpu_offset[cpu] = p - __per_cpu_start; in init_percpu_area() 31 unsigned int cpu; member 39 unsigned int cpu = info->cpu; in _free_percpu_area() local 45 static void free_percpu_area(unsigned int cpu) in free_percpu_area() argument 48 info->cpu = cpu; in free_percpu_area() 55 unsigned int cpu = (unsigned long)hcpu; in cpu_percpu_callback() local 61 rc = init_percpu_area(cpu); in cpu_percpu_callback() [all …]
|
A D | smpboot.c | 89 cpumask_set_cpu(cpu, per_cpu(cpu_sibling_mask, cpu)); in setup_cpu_sibling_map() 90 cpumask_set_cpu(cpu, per_cpu(cpu_core_mask, cpu)); in setup_cpu_sibling_map() 117 struct dt_device_node *cpu; in dt_smp_init_cpus() local 148 dt_node_full_name(cpu), dt_n_size_cells(cpu)); in dt_smp_init_cpus() 407 int __init cpu_up_send_sgi(int cpu) in cpu_up_send_sgi() argument 418 int __cpu_up(unsigned int cpu) in __cpu_up() argument 435 init_data.cpuid = cpu; in __cpu_up() 441 rc = arch_cpu_up(cpu); in __cpu_up() 478 if ( !cpu_online(cpu) ) in __cpu_up() 488 void __cpu_die(unsigned int cpu) in __cpu_die() argument [all …]
|
/xen/xen/arch/x86/genapic/ |
A D | x2apic.c | 55 for_each_online_cpu ( cpu ) in init_apic_ldr_x2apic_cluster() 57 if (this_cpu == cpu || x2apic_cluster(this_cpu) != x2apic_cluster(cpu)) in init_apic_ldr_x2apic_cluster() 74 return per_cpu(cluster_cpus, cpu); in vector_allocation_cpumask_x2apic_cluster() 83 for_each_cpu ( cpu, cluster_cpus ) in cpu_mask_to_apicid_x2apic_cluster() 97 unsigned int cpu; in send_IPI_mask_x2apic_phys() local 115 for_each_cpu ( cpu, cpumask ) in send_IPI_mask_x2apic_phys() 117 if ( !cpu_online(cpu) || (cpu == smp_processor_id()) ) in send_IPI_mask_x2apic_phys() 130 unsigned int cpu = smp_processor_id(); in send_IPI_mask_x2apic_cluster() local 147 for_each_cpu ( cpu, cluster_cpus ) in send_IPI_mask_x2apic_cluster() 207 if ( per_cpu(cluster_cpus, cpu) ) in update_clusterinfo() [all …]
|
/xen/xen/common/sched/ |
A D | null.c | 127 cpumask_and(cpumask_scratch_cpu(cpu), cpumask_scratch_cpu(cpu), in unit_check_affinity() 130 return cpumask_test_cpu(cpu, cpumask_scratch_cpu(cpu)); in unit_check_affinity() 287 cpumask_and(cpumask_scratch_cpu(cpu), cpumask_scratch_cpu(cpu), cpus); in pick_res() 297 && cpumask_test_cpu(cpu, cpumask_scratch_cpu(cpu))) ) in pick_res() 304 cpumask_and(cpumask_scratch_cpu(cpu), cpumask_scratch_cpu(cpu), in pick_res() 359 uint32_t cpu; in unit_assign() member 363 d.cpu = cpu; in unit_assign() 394 d.cpu = cpu; in unit_deassign() 459 unsigned int cpu; in null_unit_insert() local 625 cpumask_and(cpumask_scratch_cpu(cpu), cpumask_scratch_cpu(cpu), in null_unit_wake() [all …]
|
A D | private.h | 76 return get_sched_res(cpu)->curr; in curr_on_cpu() 154 ASSERT(cpu < nr_cpu_ids); in sched_set_res() 155 v->processor = cpu; in sched_set_res() 156 cpu = cpumask_next(cpu, res->cpus); in sched_set_res() 239 sched_lock(pcpu, unsigned int cpu, cpu, ) 241 sched_lock(pcpu, unsigned int cpu, cpu, _irq) in sched_lock() argument 243 sched_unlock(pcpu, unsigned int cpu, cpu, ) in sched_lock() 245 sched_unlock(pcpu, unsigned int cpu, cpu, _irq) in sched_lock() 251 sched_lock(pcpu, unsigned int cpu, cpu, _irqsave, *flags) in sched_lock() 254 sched_unlock(pcpu, unsigned int cpu, cpu, _irqrestore, flags) in sched_lock() [all …]
|
A D | cpupool.c | 93 unsigned int cpu; in cpupool_check_granularity() local 99 for_each_online_cpu ( cpu ) in cpupool_check_granularity() 504 cpupool_moving_cpu = cpu; in cpupool_unassign_cpu_start() 556 if ( !cpu_online(cpu) ) in cpupool_unassign_cpu() 752 sched_rm_cpu(cpu); in cpupool_cpu_remove_forced() 809 unsigned cpu; in cpupool_do_sysctl() local 812 cpu = op->cpu; in cpupool_do_sysctl() 854 unsigned cpu; in cpupool_do_sysctl() local 860 cpu = op->cpu; in cpupool_do_sysctl() 863 ret = (cpu < nr_cpu_ids) ? cpupool_unassign_cpu(c, cpu) : -EINVAL; in cpupool_do_sysctl() [all …]
|
A D | credit.c | 383 ASSERT(cpumask_cycle(cpu, unit->cpu_hard_affinity) == cpu); in __runq_tickle() 490 __trace_var(TRC_CSCHED_TICKLE, 1, sizeof(cpu), &cpu); in __runq_tickle() 595 prv->master = cpu; in init_pdata() 602 prv->balance_bias[cpu_to_node(cpu)] = cpu; in init_pdata() 604 init_timer(&spc->ticker, csched_tick, (void *)(unsigned long)cpu, cpu); in init_pdata() 769 if ( sched_unit_master(unit) == cpu && is_runq_idle(cpu) ) in _csched_cpu_pick() 786 cpu = cpumask_cycle(cpu, cpus); in _csched_cpu_pick() 843 spc->idle_bias = cpu; in _csched_cpu_pick() 846 cpu); in _csched_cpu_pick() 848 return cpu; in _csched_cpu_pick() [all …]
|
A D | credit2.c | 740 cpumask_and(cpumask_scratch_cpu(cpu), cpumask_scratch_cpu(cpu), in get_fallback_cpu() 747 if ( likely(cpumask_test_cpu(cpu, cpumask_scratch_cpu(cpu))) ) in get_fallback_cpu() 789 cpu = cpumask_first(cpumask_scratch_cpu(cpu)); in get_fallback_cpu() 909 cpumask_or(cpumask_scratch_cpu(cpu), cpumask_scratch_cpu(cpu), in cpu_runqueue_siblings_match() 1628 ASSERT(cpumask_cycle(cpu, unit->cpu_hard_affinity) == cpu); in runq_tickle() 1630 ipid = cpu; in runq_tickle() 2500 cpumask_and(cpumask_scratch_cpu(cpu), cpumask_scratch_cpu(cpu), in csched2_res_pick() 2502 cpumask_and(cpumask_scratch_cpu(cpu), cpumask_scratch_cpu(cpu), in csched2_res_pick() 2516 cpumask_and(cpumask_scratch_cpu(cpu), cpumask_scratch_cpu(cpu), in csched2_res_pick() 2648 cpumask_and(cpumask_scratch_cpu(cpu), cpumask_scratch_cpu(cpu), in migrate() [all …]
|
/xen/xen/drivers/cpufreq/ |
A D | cpufreq_misc_governors.c | 33 unsigned int cpu; in cpufreq_governor_userspace() local 36 unlikely(!cpu_online(cpu = policy->cpu))) in cpufreq_governor_userspace() 41 if (!per_cpu(cpu_set_freq, cpu)) in cpufreq_governor_userspace() 42 per_cpu(cpu_set_freq, cpu) = policy->cur; in cpufreq_governor_userspace() 45 per_cpu(cpu_set_freq, cpu) = 0; in cpufreq_governor_userspace() 48 if (policy->max < per_cpu(cpu_set_freq, cpu)) in cpufreq_governor_userspace() 71 if (!cpu_online(cpu) || !(policy = per_cpu(cpufreq_cpu_policy, cpu))) in write_userspace_scaling_setspeed() 74 per_cpu(cpu_set_freq, cpu) = freq; in write_userspace_scaling_setspeed() 97 unsigned int cpu = (unsigned long)hcpu; in cpufreq_userspace_cpu_callback() local 121 unsigned int cpu; in cpufreq_gov_userspace_init() local [all …]
|
A D | cpufreq.c | 138 if (!cpu_online(cpu) || !(data = per_cpu(cpufreq_cpu_policy, cpu)) || in cpufreq_limit_change() 139 !processor_pminfo[cpu]) in cpufreq_limit_change() 155 int cpufreq_add_cpu(unsigned int cpu) in cpufreq_add_cpu() argument 168 if ( !processor_pminfo[cpu] || !cpu_online(cpu) ) in cpufreq_add_cpu() 240 policy->cpu = cpu; in cpufreq_add_cpu() 295 cpufreq_statistic_exit(cpu); in cpufreq_add_cpu() 316 int cpufreq_del_cpu(unsigned int cpu) in cpufreq_del_cpu() argument 326 if ( !processor_pminfo[cpu] || !cpu_online(cpu) ) in cpufreq_del_cpu() 360 cpufreq_statistic_exit(cpu); in cpufreq_del_cpu() 644 (void)cpufreq_add_cpu(cpu); in cpu_callback() [all …]
|
/xen/ |
A D | config.sub | 939 cpu=i386 1017 cpu=bfin 1037 cpu=mt 1040 cpu=m68k 1093 cpu=arm 1105 cpu=`echo "$cpu" | sed 's/^xscale/arm/'` 1125 cpu=neo 1129 cpu=nse 1133 cpu=nsr 1137 cpu=nsv [all …]
|
/xen/xen/arch/x86/acpi/cpufreq/ |
A D | cpufreq.c | 205 if (!cpumask_test_cpu(cpu, mask)) in get_cur_val() 206 cpu = cpumask_first(mask); in get_cur_val() 207 if (cpu >= nr_cpu_ids || !cpu_online(cpu)) in get_cur_val() 229 cmd.mask = cpumask_of(cpu); in get_cur_val() 274 if (!cpu_online(cpu)) in get_measured_perf() 297 if (cpu == smp_processor_id()) { in get_measured_perf() 328 if (!cpu_online(cpu)) in get_cur_freq_on_cpu() 508 unsigned int cpu = policy->cpu; in acpi_cpufreq_cpu_init() local 518 cpufreq_drv_data[cpu] = data; in acpi_cpufreq_cpu_init() 536 if (!check_est_cpu(cpu)) { in acpi_cpufreq_cpu_init() [all …]
|
/xen/xen/include/asm-x86/ |
A D | smp.h | 44 #define cpu_physical_id(cpu) x86_cpu_to_apicid[cpu] argument 46 #define cpu_is_offline(cpu) unlikely(!cpu_online(cpu)) argument 47 extern void cpu_exit_clear(unsigned int cpu); 48 extern void cpu_uninit(unsigned int cpu); 64 bool core_parking_remove(unsigned int cpu); 82 #define get_cpu_current(cpu) \ argument 83 (get_cpu_info_from_stack((unsigned long)stack_base[cpu])->current_vcpu)
|