/linux/arch/x86/kvm/ |
A D | trace.h | 27 __entry->vcpu_id = vcpu->vcpu_id; 1027 __entry->vcpu_id = vcpu_id; 1058 __entry->vcpu_id = vcpu_id; 1088 __entry->vcpu_id = vcpu_id; 1110 __entry->vcpu_id = vcpu_id; 1136 __entry->vcpu_id = vcpu_id; 1158 __entry->vcpu_id = vcpu_id; 1183 __entry->vcpu_id = vcpu_id; 1209 __entry->vcpu_id = vcpu_id; 1235 __entry->vcpu_id = vcpu_id; [all …]
|
A D | ioapic.c | 124 old_val = test_bit(vcpu->vcpu_id, dest_map->map); in __rtc_irq_eoi_tracking_restore_one() 130 __set_bit(vcpu->vcpu_id, dest_map->map); in __rtc_irq_eoi_tracking_restore_one() 131 dest_map->vectors[vcpu->vcpu_id] = e->fields.vector; in __rtc_irq_eoi_tracking_restore_one() 134 __clear_bit(vcpu->vcpu_id, dest_map->map); in __rtc_irq_eoi_tracking_restore_one() 168 if (test_bit(vcpu->vcpu_id, dest_map->map) && in rtc_irq_eoi() 169 (vector == dest_map->vectors[vcpu->vcpu_id]) && in rtc_irq_eoi() 170 (test_and_clear_bit(vcpu->vcpu_id, in rtc_irq_eoi() 289 if (test_bit(vcpu->vcpu_id, dest_map->map)) in kvm_ioapic_scan_entry() 290 __set_bit(dest_map->vectors[vcpu->vcpu_id], in kvm_ioapic_scan_entry()
|
/linux/tools/testing/selftests/kvm/lib/ |
A D | perf_test_util.c | 21 int vcpu_id; member 52 GUEST_ASSERT(vcpu_args->vcpu_id == vcpu_id); in guest_code() 77 int vcpu_id; in perf_test_setup_vcpus() local 79 for (vcpu_id = 0; vcpu_id < vcpus; vcpu_id++) { in perf_test_setup_vcpus() 82 vcpu_args->vcpu_id = vcpu_id; in perf_test_setup_vcpus() 96 vcpu_args_set(vm, vcpu_id, 1, vcpu_id); in perf_test_setup_vcpus() 223 int vcpu_id; in perf_test_start_vcpu_threads() local 228 for (vcpu_id = 0; vcpu_id < vcpus; vcpu_id++) { in perf_test_start_vcpu_threads() 231 vcpu->vcpu_id = vcpu_id; in perf_test_start_vcpu_threads() 237 for (vcpu_id = 0; vcpu_id < vcpus; vcpu_id++) { in perf_test_start_vcpu_threads() [all …]
|
/linux/tools/testing/selftests/kvm/ |
A D | kvm_page_table_test.c | 50 int vcpu_id; member 104 GUEST_ASSERT(vcpu_args->vcpu_id == vcpu_id); in guest_code() 199 int vcpu_id = vcpu_args->vcpu_id; in vcpu_worker() local 205 vcpu_args_set(vm, vcpu_id, 1, vcpu_id); in vcpu_worker() 263 int vcpu_id; in pre_init_before_test() local 295 for (vcpu_id = 0; vcpu_id < KVM_MAX_VCPUS; vcpu_id++) { in pre_init_before_test() 297 vcpu_args->vcpu_id = vcpu_id; in pre_init_before_test() 369 int vcpu_id; in run_test() local 382 for (vcpu_id = 0; vcpu_id < nr_vcpus; vcpu_id++) { in run_test() 427 for (vcpu_id = 0; vcpu_id < nr_vcpus; vcpu_id++) { in run_test() [all …]
|
A D | dirty_log_perf_test.c | 44 int vcpu_id = vcpu_args->vcpu_id; in vcpu_worker() local 46 run = vcpu_state(vm, vcpu_id); in vcpu_worker() 52 ret = _vcpu_run(vm, vcpu_id); in vcpu_worker() 63 vcpu_id, vcpu_last_completed_iteration[vcpu_id]); in vcpu_worker() 69 vcpu_id, current_iteration, ts_diff.tv_sec, in vcpu_worker() 73 vcpu_id, current_iteration, ts_diff.tv_sec, in vcpu_worker() 83 vcpu_id, pages_count, vcpu_last_completed_iteration[vcpu_id], in vcpu_worker() 175 int vcpu_id; in run_test() local 208 for (vcpu_id = 0; vcpu_id < nr_vcpus; vcpu_id++) in run_test() 215 for (vcpu_id = 0; vcpu_id < nr_vcpus; vcpu_id++) { in run_test() [all …]
|
A D | access_tracking_perf_test.c | 141 if (overlap_memory_access && vcpu_id) in mark_vcpu_memory_idle() 173 vcpu_id, no_pfn, pages); in mark_vcpu_memory_idle() 186 vcpu_id, still_idle, pages); in mark_vcpu_memory_idle() 196 uint64_t actual_ucall = get_ucall(vm, vcpu_id, &uc); in assert_ucall() 221 int vcpu_id = vcpu_args->vcpu_id; in vcpu_thread_main() local 227 vcpu_run(vm, vcpu_id); in vcpu_thread_main() 228 assert_ucall(vm, vcpu_id, UCALL_SYNC); in vcpu_thread_main() 231 mark_vcpu_memory_idle(vm, vcpu_id); in vcpu_thread_main() 258 int vcpu_id; in run_iteration() local 266 for (vcpu_id = 0; vcpu_id < vcpus; vcpu_id++) in run_iteration() [all …]
|
A D | demand_paging_test.c | 48 int vcpu_id = vcpu_args->vcpu_id; in vcpu_worker() local 54 run = vcpu_state(vm, vcpu_id); in vcpu_worker() 59 ret = _vcpu_run(vm, vcpu_id); in vcpu_worker() 61 if (get_ucall(vm, vcpu_id, NULL) != UCALL_SYNC) { in vcpu_worker() 288 int vcpu_id; in run_test() local 312 for (vcpu_id = 0; vcpu_id < nr_vcpus; vcpu_id++) { in run_test() 327 r = pipe2(&pipefds[vcpu_id * 2], in run_test() 332 pipefds[vcpu_id * 2], p->uffd_mode, in run_test() 333 p->uffd_delay, &uffd_args[vcpu_id], in run_test() 353 for (vcpu_id = 0; vcpu_id < nr_vcpus; vcpu_id++) { in run_test() [all …]
|
A D | memslot_modification_stress_test.c | 42 int vcpu_id = vcpu_args->vcpu_id; in vcpu_worker() local 46 run = vcpu_state(vm, vcpu_id); in vcpu_worker() 50 ret = _vcpu_run(vm, vcpu_id); in vcpu_worker() 53 if (get_ucall(vm, vcpu_id, NULL) == UCALL_SYNC) in vcpu_worker()
|
/linux/arch/powerpc/kvm/ |
A D | trace_hv.h | 233 __entry->vcpu_id = vcpu->vcpu_id; 240 __entry->vcpu_id, 258 __entry->vcpu_id = vcpu->vcpu_id; 266 __entry->vcpu_id, 291 __entry->vcpu_id = vcpu->vcpu_id; 302 __entry->vcpu_id, 321 __entry->vcpu_id = vcpu->vcpu_id; 347 __entry->vcpu_id = vcpu->vcpu_id; 373 __entry->vcpu_id = vcpu->vcpu_id; 467 __entry->vcpu_id = vcpu->vcpu_id; [all …]
|
/linux/arch/arm64/kvm/vgic/ |
A D | trace.h | 11 TP_PROTO(unsigned long vcpu_id, __u32 irq, bool level), 12 TP_ARGS(vcpu_id, irq, level), 15 __field( unsigned long, vcpu_id ) 21 __entry->vcpu_id = vcpu_id; 27 __entry->vcpu_id, __entry->irq, __entry->level)
|
A D | vgic-debug.c | 30 int vcpu_id; member 45 ++iter->vcpu_id < iter->nr_cpus) in iter_next() 78 iter->vcpu_id == iter->nr_cpus && in end_of_vgic() 169 id = vcpu->vcpu_id; in print_header() 215 (irq->target_vcpu) ? irq->target_vcpu->vcpu_id : -1, in print_irq_state() 227 (irq->vcpu) ? irq->vcpu->vcpu_id : -1); in print_irq_state() 246 if (iter->vcpu_id < iter->nr_cpus) in vgic_debug_show() 247 vcpu = kvm_get_vcpu(kvm, iter->vcpu_id); in vgic_debug_show()
|
/linux/arch/arm64/kvm/ |
A D | trace_arm.h | 181 TP_PROTO(unsigned long vcpu_id, __u32 irq, int level), 182 TP_ARGS(vcpu_id, irq, level), 185 __field( unsigned long, vcpu_id ) 191 __entry->vcpu_id = vcpu_id; 197 __entry->vcpu_id, __entry->irq, __entry->level) 201 TP_PROTO(unsigned long vcpu_id, struct timer_map *map), 202 TP_ARGS(vcpu_id, map), 205 __field( unsigned long, vcpu_id ) 212 __entry->vcpu_id = vcpu_id; 221 __entry->vcpu_id,
|
/linux/tools/testing/selftests/kvm/x86_64/ |
A D | xapic_ipi_test.c | 83 uint32_t vcpu_id; member 209 params->vcpu_id, r); in vcpu_thread() 211 fprintf(stderr, "vCPU thread running vCPU %u\n", params->vcpu_id); in vcpu_thread() 212 vcpu_run(params->vm, params->vcpu_id); in vcpu_thread() 217 params->vcpu_id, exit_reason, exit_reason_str(exit_reason)); in vcpu_thread() 219 if (get_ucall(params->vm, params->vcpu_id, &uc) == UCALL_ABORT) { in vcpu_thread() 227 params->vcpu_id, (const char *)uc.args[0], in vcpu_thread() 247 vcpu_id, r); in cancel_join_vcpu_thread() 252 vcpu_id, r); in cancel_join_vcpu_thread() 445 params[0].vcpu_id = HALTER_VCPU_ID; in main() [all …]
|
/linux/drivers/virt/nitro_enclaves/ |
A D | ne_misc_dev.c | 517 int core_id, u32 vcpu_id) in ne_set_enclave_threads_per_core() argument 521 if (core_id < 0 && vcpu_id == 0) { in ne_set_enclave_threads_per_core() 577 *vcpu_id = cpu; in ne_get_cpu_from_cpu_pool() 647 if (ne_donated_cpu(ne_enclave, vcpu_id)) { in ne_check_cpu_in_cpu_pool() 649 "CPU %d already used\n", vcpu_id); in ne_check_cpu_in_cpu_pool() 706 slot_add_vcpu_req.vcpu_id = vcpu_id; in ne_add_vcpu_ioctl() 1097 u32 vcpu_id = 0; in ne_enclave_ioctl() local 1099 if (copy_from_user(&vcpu_id, (void __user *)arg, sizeof(vcpu_id))) in ne_enclave_ioctl() 1123 if (!vcpu_id) { in ne_enclave_ioctl() 1141 vcpu_id, rc); in ne_enclave_ioctl() [all …]
|
/linux/samples/acrn/ |
A D | vm-sample.c | 49 int vcpu_id, ret; in main() local 86 regs.vcpu_id = 0; in main() 112 for (vcpu_id = 0; vcpu_id < vcpu_num; vcpu_id++) { in main() 113 io_req = &io_req_buf[vcpu_id]; in main() 125 notify.vcpu = vcpu_id; in main()
|
/linux/include/trace/events/ |
A D | kvm.h | 350 TP_ARGS(grow, vcpu_id, new, old), 354 __field(unsigned int, vcpu_id) 361 __entry->vcpu_id = vcpu_id; 367 __entry->vcpu_id, 373 #define trace_kvm_halt_poll_ns_grow(vcpu_id, new, old) \ argument 374 trace_kvm_halt_poll_ns(true, vcpu_id, new, old) 375 #define trace_kvm_halt_poll_ns_shrink(vcpu_id, new, old) \ argument 376 trace_kvm_halt_poll_ns(false, vcpu_id, new, old) 431 __field(int, vcpu_id) 435 __entry->vcpu_id = vcpu->vcpu_id; [all …]
|
/linux/tools/testing/selftests/kvm/lib/x86_64/ |
A D | ucall.c | 38 uint64_t get_ucall(struct kvm_vm *vm, uint32_t vcpu_id, struct ucall *uc) in get_ucall() argument 40 struct kvm_run *run = vcpu_state(vm, vcpu_id); in get_ucall() 49 vcpu_regs_get(vm, vcpu_id, ®s); in get_ucall() 53 vcpu_run_complete_io(vm, vcpu_id); in get_ucall()
|
/linux/arch/s390/kvm/ |
A D | sigp.c | 39 VCPU_EVENT(vcpu, 4, "sensed status of cpu %x rc %x", dst_vcpu->vcpu_id, in __sigp_sense() 49 .u.emerg.code = vcpu->vcpu_id, in __inject_sigp_emergency() 56 dst_vcpu->vcpu_id); in __inject_sigp_emergency() 98 .u.extcall.code = vcpu->vcpu_id, in __sigp_external_call() 109 dst_vcpu->vcpu_id); in __sigp_external_call() 127 dst_vcpu->vcpu_id); in __sigp_stop() 146 dst_vcpu->vcpu_id); in __sigp_stop_and_store_status() 235 dst_vcpu->vcpu_id, rc); in __sigp_sense_running() 342 order_code, dst_vcpu->vcpu_id); in handle_sigp_dst()
|
A D | interrupt.c | 54 sca->cpu[vcpu->vcpu_id].sigp_ctrl; in sca_ext_call_pending() 61 sca->cpu[vcpu->vcpu_id].sigp_ctrl; in sca_ext_call_pending() 83 &(sca->cpu[vcpu->vcpu_id].sigp_ctrl); in sca_inject_ext_call() 95 &(sca->cpu[vcpu->vcpu_id].sigp_ctrl); in sca_inject_ext_call() 126 &(sca->cpu[vcpu->vcpu_id].sigp_ctrl); in sca_clear_ext_call() 134 &(sca->cpu[vcpu->vcpu_id].sigp_ctrl); in sca_clear_ext_call() 552 trace_kvm_s390_deliver_interrupt(vcpu->vcpu_id, in __deliver_pfault_init() 720 trace_kvm_s390_deliver_interrupt(vcpu->vcpu_id, in __deliver_machine_check() 763 trace_kvm_s390_deliver_interrupt(vcpu->vcpu_id, in __deliver_set_prefix() 819 trace_kvm_s390_deliver_interrupt(vcpu->vcpu_id, in __deliver_external_call() [all …]
|
A D | pv.c | 27 vcpu->vcpu_id, *rc, *rrc); in kvm_s390_pv_destroy_cpu() 86 vcpu->vcpu_id, uvcb.cpu_handle, uvcb.header.rc, in kvm_s390_pv_create_cpu() 303 vcpu->vcpu_id, state, uvcb.header.rc, uvcb.header.rrc); in kvm_s390_pv_set_cpu_state()
|
/linux/arch/x86/kvm/svm/ |
A D | avic.c | 93 u32 vcpu_id = AVIC_GATAG_TO_VCPUID(ga_tag); in avic_ga_log_notifier() local 96 trace_kvm_avic_ga_log(vm_id, vcpu_id); in avic_ga_log_notifier() 102 vcpu = kvm_get_vcpu_by_id(&kvm_svm->kvm, vcpu_id); in avic_ga_log_notifier() 258 int id = vcpu->vcpu_id; in avic_init_backing_page() 346 index, vcpu->vcpu_id, icrh, icrl); in avic_incomplete_ipi_interception() 443 if (vcpu->vcpu_id == id) in avic_handle_apic_id_update() 446 old = avic_get_physical_id_entry(vcpu, vcpu->vcpu_id); in avic_handle_apic_id_update() 546 trace_kvm_avic_unaccelerated_access(vcpu->vcpu_id, offset, in avic_unaccelerated_access_interception() 723 u32 vcpu_id = AVIC_GATAG_TO_VCPUID(pi->prev_ga_tag); in svm_ir_list_add() local 839 svm->vcpu.vcpu_id); in svm_update_pi_irte() [all …]
|
/linux/tools/testing/selftests/kvm/lib/s390x/ |
A D | ucall.c | 36 uint64_t get_ucall(struct kvm_vm *vm, uint32_t vcpu_id, struct ucall *uc) in get_ucall() argument 38 struct kvm_run *run = vcpu_state(vm, vcpu_id); in get_ucall() 53 vcpu_run_complete_io(vm, vcpu_id); in get_ucall()
|
/linux/tools/testing/selftests/kvm/lib/aarch64/ |
A D | ucall.c | 92 uint64_t get_ucall(struct kvm_vm *vm, uint32_t vcpu_id, struct ucall *uc) in get_ucall() argument 94 struct kvm_run *run = vcpu_state(vm, vcpu_id); in get_ucall() 109 vcpu_run_complete_io(vm, vcpu_id); in get_ucall()
|
/linux/tools/perf/ |
A D | builtin-kvm.c | 143 int vcpu_id; member 191 if (vcpu_id < event->max_vcpu) in kvm_event_expand() 194 while (event->max_vcpu <= vcpu_id) in kvm_event_expand() 273 if (vcpu_id != -1) in kvm_event_rel_stddev() 274 kvm_stats = &event->vcpu[vcpu_id]; in kvm_event_rel_stddev() 283 if (vcpu_id == -1) { in update_kvm_event() 288 if (!kvm_event_expand(event, vcpu_id)) in update_kvm_event() 291 kvm_update_event_stats(&event->vcpu[vcpu_id], time_diff); in update_kvm_event() 355 vcpu = vcpu_record->vcpu_id; in handle_end_event() 396 sample->time, sample->pid, vcpu_record->vcpu_id, in handle_end_event() [all …]
|
/linux/arch/x86/kvm/mmu/ |
A D | mmutrace.h | 260 __field(int, vcpu_id) 270 __entry->vcpu_id = vcpu->vcpu_id; 280 " new %llx spurious %d fixed %d", __entry->vcpu_id,
|