/linux/tools/testing/selftests/kvm/s390x/ |
A D | resets.c | 16 #define VCPU_ID 3 macro 68 vcpu_get_reg(vm, VCPU_ID, ®); in test_one_reg() 101 vcpu_fpu_get(vm, VCPU_ID, &fpu); in assert_clear() 208 run = vcpu_state(vm, VCPU_ID); in test_normal() 211 vcpu_run(vm, VCPU_ID); in test_normal() 213 inject_irq(VCPU_ID); in test_normal() 230 run = vcpu_state(vm, VCPU_ID); in test_initial() 233 vcpu_run(vm, VCPU_ID); in test_initial() 235 inject_irq(VCPU_ID); in test_initial() 255 vcpu_run(vm, VCPU_ID); in test_clear() [all …]
|
A D | sync_regs_test.c | 25 #define VCPU_ID 5 macro 97 run = vcpu_state(vm, VCPU_ID); in main() 101 rv = _vcpu_run(vm, VCPU_ID); in main() 108 rv = _vcpu_run(vm, VCPU_ID); in main() 116 rv = _vcpu_run(vm, VCPU_ID); in main() 123 rv = _vcpu_run(vm, VCPU_ID); in main() 131 rv = _vcpu_run(vm, VCPU_ID); in main() 144 vcpu_regs_get(vm, VCPU_ID, ®s); in main() 162 rv = _vcpu_run(vm, VCPU_ID); in main() 178 vcpu_regs_get(vm, VCPU_ID, ®s); in main() [all …]
|
A D | memop.c | 16 #define VCPU_ID 1 macro 50 vm = vm_create_default(VCPU_ID, 0, guest_code); in main() 51 run = vcpu_state(vm, VCPU_ID); in main() 63 vcpu_ioctl(vm, VCPU_ID, KVM_S390_MEM_OP, &ksmo); in main() 66 vcpu_run(vm, VCPU_ID); in main() 81 vcpu_ioctl(vm, VCPU_ID, KVM_S390_MEM_OP, &ksmo); in main() 93 rv = _vcpu_ioctl(vm, VCPU_ID, KVM_S390_MEM_OP, &ksmo); in main() 103 rv = _vcpu_ioctl(vm, VCPU_ID, KVM_S390_MEM_OP, &ksmo); in main() 114 rv = _vcpu_ioctl(vm, VCPU_ID, KVM_S390_MEM_OP, &ksmo); in main() 124 rv = _vcpu_ioctl(vm, VCPU_ID, KVM_S390_MEM_OP, &ksmo); in main() [all …]
|
/linux/tools/testing/selftests/kvm/x86_64/ |
A D | sync_regs_test.c | 23 #define VCPU_ID 5 macro 110 run = vcpu_state(vm, VCPU_ID); in main() 114 rv = _vcpu_run(vm, VCPU_ID); in main() 121 rv = _vcpu_run(vm, VCPU_ID); in main() 129 rv = _vcpu_run(vm, VCPU_ID); in main() 136 rv = _vcpu_run(vm, VCPU_ID); in main() 145 rv = _vcpu_run(vm, VCPU_ID); in main() 167 rv = _vcpu_run(vm, VCPU_ID); in main() 194 rv = _vcpu_run(vm, VCPU_ID); in main() 212 rv = _vcpu_run(vm, VCPU_ID); in main() [all …]
|
A D | evmcs_test.c | 20 #define VCPU_ID 5 macro 122 vcpu_regs_get(vm, VCPU_ID, ®s1); in save_restore_vm() 128 vm_vcpu_add(vm, VCPU_ID); in save_restore_vm() 129 vcpu_set_hv_cpuid(vm, VCPU_ID); in save_restore_vm() 130 vcpu_enable_evmcs(vm, VCPU_ID); in save_restore_vm() 131 vcpu_load_state(vm, VCPU_ID, state); in save_restore_vm() 135 vcpu_regs_get(vm, VCPU_ID, ®s2); in save_restore_vm() 160 vcpu_set_hv_cpuid(vm, VCPU_ID); in main() 161 vcpu_enable_evmcs(vm, VCPU_ID); in main() 174 run = vcpu_state(vm, VCPU_ID); in main() [all …]
|
A D | smm_test.c | 22 #define VCPU_ID 1 macro 126 vcpu_events_get(vm, VCPU_ID, &events); in inject_smi() 131 vcpu_events_set(vm, VCPU_ID, &events); in inject_smi() 147 run = vcpu_state(vm, VCPU_ID); in main() 170 vcpu_args_set(vm, VCPU_ID, 1, nested_gva); in main() 173 _vcpu_run(vm, VCPU_ID); in main() 180 vcpu_regs_get(vm, VCPU_ID, ®s); in main() 208 state = vcpu_save_state(vm, VCPU_ID); in main() 211 vm_vcpu_add(vm, VCPU_ID); in main() 213 vcpu_load_state(vm, VCPU_ID, state); in main() [all …]
|
A D | emulator_error_test.c | 14 #define VCPU_ID 1 macro 35 rc = _vcpu_run(vm, VCPU_ID); in run_guest() 62 struct kvm_run *run = vcpu_state(vm, VCPU_ID); in process_exit_on_emulation_error() 96 vcpu_regs_get(vm, VCPU_ID, ®s); in process_exit_on_emulation_error() 98 vcpu_regs_set(vm, VCPU_ID, ®s); in process_exit_on_emulation_error() 111 struct kvm_run *run = vcpu_state(vm, VCPU_ID); in check_for_guest_assert() 122 struct kvm_run *run = vcpu_state(vm, VCPU_ID); in process_ucall_done() 139 struct kvm_run *run = vcpu_state(vm, VCPU_ID); in process_ucall() 147 switch (get_ucall(vm, VCPU_ID, &uc)) { in process_ucall() 179 vm = vm_create_default(VCPU_ID, 0, guest_code); in main() [all …]
|
A D | platform_info_test.c | 24 #define VCPU_ID 0 macro 50 struct kvm_run *run = vcpu_state(vm, VCPU_ID); in test_msr_platform_info_enabled() 54 vcpu_run(vm, VCPU_ID); in test_msr_platform_info_enabled() 59 get_ucall(vm, VCPU_ID, &uc); in test_msr_platform_info_enabled() 70 struct kvm_run *run = vcpu_state(vm, VCPU_ID); in test_msr_platform_info_disabled() 73 vcpu_run(vm, VCPU_ID); in test_msr_platform_info_disabled() 95 vm = vm_create_default(VCPU_ID, 0, guest_code); in main() 97 msr_platform_info = vcpu_get_msr(vm, VCPU_ID, MSR_PLATFORM_INFO); in main() 98 vcpu_set_msr(vm, VCPU_ID, MSR_PLATFORM_INFO, in main() 102 vcpu_set_msr(vm, VCPU_ID, MSR_PLATFORM_INFO, msr_platform_info); in main()
|
A D | vmx_preemption_timer_test.c | 25 #define VCPU_ID 5 macro 179 run = vcpu_state(vm, VCPU_ID); in main() 181 vcpu_regs_get(vm, VCPU_ID, ®s1); in main() 187 _vcpu_run(vm, VCPU_ID); in main() 193 switch (get_ucall(vm, VCPU_ID, &uc)) { in main() 235 state = vcpu_save_state(vm, VCPU_ID); in main() 237 vcpu_regs_get(vm, VCPU_ID, ®s1); in main() 243 vm_vcpu_add(vm, VCPU_ID); in main() 245 vcpu_load_state(vm, VCPU_ID, state); in main() 246 run = vcpu_state(vm, VCPU_ID); in main() [all …]
|
A D | vmx_invalid_nested_guest_state.c | 12 #define VCPU_ID 0 macro 63 vm = vm_create_default(VCPU_ID, 0, (void *) l1_guest_code); in main() 67 vcpu_args_set(vm, VCPU_ID, 1, vmx_pages_gva); in main() 69 vcpu_run(vm, VCPU_ID); in main() 71 run = vcpu_state(vm, VCPU_ID); in main() 91 vcpu_sregs_get(vm, VCPU_ID, &sregs); in main() 93 vcpu_sregs_set(vm, VCPU_ID, &sregs); in main() 95 vcpu_run(vm, VCPU_ID); in main() 97 switch (get_ucall(vm, VCPU_ID, &uc)) { in main()
|
A D | state_test.c | 23 #define VCPU_ID 5 macro 168 run = vcpu_state(vm, VCPU_ID); in main() 170 vcpu_regs_get(vm, VCPU_ID, ®s1); in main() 185 _vcpu_run(vm, VCPU_ID); in main() 191 switch (get_ucall(vm, VCPU_ID, &uc)) { in main() 209 state = vcpu_save_state(vm, VCPU_ID); in main() 211 vcpu_regs_get(vm, VCPU_ID, ®s1); in main() 217 vm_vcpu_add(vm, VCPU_ID); in main() 219 vcpu_load_state(vm, VCPU_ID, state); in main() 220 run = vcpu_state(vm, VCPU_ID); in main() [all …]
|
A D | set_sregs_test.c | 25 #define VCPU_ID 5 macro 40 rc = _vcpu_sregs_set(vm, VCPU_ID, &sregs); in test_cr4_feature_bit() 44 vcpu_sregs_get(vm, VCPU_ID, &sregs); in test_cr4_feature_bit() 99 vm_vcpu_add(vm, VCPU_ID); in main() 101 vcpu_sregs_get(vm, VCPU_ID, &sregs); in main() 106 rc = _vcpu_sregs_set(vm, VCPU_ID, &sregs); in main() 109 vcpu_sregs_get(vm, VCPU_ID, &sregs); in main() 127 vm = vm_create_default(VCPU_ID, 0, NULL); in main() 129 vcpu_sregs_get(vm, VCPU_ID, &sregs); in main() 131 rc = _vcpu_sregs_set(vm, VCPU_ID, &sregs); in main() [all …]
|
A D | debug_regs.c | 13 #define VCPU_ID 0 macro 73 vcpu_regs_get(vm, VCPU_ID, ®s); \ 105 run = vcpu_state(vm, VCPU_ID); in main() 111 vcpu_run(vm, VCPU_ID); in main() 127 vcpu_run(vm, VCPU_ID); in main() 150 vcpu_run(vm, VCPU_ID); in main() 170 vcpu_regs_get(vm, VCPU_ID, ®s); in main() 178 vcpu_run(vm, VCPU_ID); in main() 195 vcpu_run(vm, VCPU_ID); in main() 211 vcpu_run(vm, VCPU_ID); in main() [all …]
|
A D | vmx_pmu_msrs_test.c | 21 #define VCPU_ID 0 macro 73 vm = vm_create_default(VCPU_ID, 0, guest_code); in main() 92 vcpu_set_cpuid(vm, VCPU_ID, cpuid); in main() 96 ASSERT_EQ(vcpu_get_msr(vm, VCPU_ID, MSR_IA32_PERF_CAPABILITIES), PMU_CAP_FW_WRITES); in main() 99 vcpu_run(vm, VCPU_ID); in main() 100 ASSERT_EQ(vcpu_get_msr(vm, VCPU_ID, MSR_IA32_PERF_CAPABILITIES), PMU_CAP_FW_WRITES); in main() 104 ASSERT_EQ(vcpu_get_msr(vm, VCPU_ID, MSR_IA32_PERF_CAPABILITIES), 0); in main() 107 ASSERT_EQ(vcpu_get_msr(vm, VCPU_ID, MSR_IA32_PERF_CAPABILITIES), (u64)host_cap.lbr_format); in main()
|
A D | mmu_role_test.c | 6 #define VCPU_ID 1 macro 34 vm = vm_create_default(VCPU_ID, 0, guest_code); in mmu_role_test() 35 run = vcpu_state(vm, VCPU_ID); in mmu_role_test() 40 r = _vcpu_run(vm, VCPU_ID); in mmu_role_test() 60 vcpu_set_cpuid(vm, VCPU_ID, kvm_get_supported_cpuid()); in mmu_role_test() 73 vcpu_init_descriptor_tables(vm, VCPU_ID); in mmu_role_test() 76 r = _vcpu_run(vm, VCPU_ID); in mmu_role_test() 79 cmd = get_ucall(vm, VCPU_ID, NULL); in mmu_role_test()
|
A D | cr4_cpuid_sync_test.c | 24 #define VCPU_ID 1 macro 83 vm = vm_create_default(VCPU_ID, 0, guest_code); in main() 84 run = vcpu_state(vm, VCPU_ID); in main() 87 rc = _vcpu_run(vm, VCPU_ID); in main() 95 switch (get_ucall(vm, VCPU_ID, &uc)) { in main() 98 vcpu_sregs_get(vm, VCPU_ID, &sregs); in main() 100 vcpu_sregs_set(vm, VCPU_ID, &sregs); in main()
|
A D | xen_shinfo_test.c | 18 #define VCPU_ID 5 macro 161 vm = vm_create_default(VCPU_ID, 0, (void *) guest_code); in main() 162 vcpu_set_cpuid(vm, VCPU_ID, kvm_get_supported_cpuid()); in main() 191 vcpu_ioctl(vm, VCPU_ID, KVM_XEN_VCPU_SET_ATTR, &vi); in main() 206 vcpu_init_descriptor_tables(vm, VCPU_ID); in main() 214 vcpu_ioctl(vm, VCPU_ID, KVM_XEN_VCPU_SET_ATTR, &st); in main() 226 volatile struct kvm_run *run = vcpu_state(vm, VCPU_ID); in main() 229 vcpu_run(vm, VCPU_ID); in main() 236 switch (get_ucall(vm, VCPU_ID, &uc)) { in main() 261 vcpu_ioctl(vm, VCPU_ID, KVM_XEN_VCPU_SET_ATTR, &rst); in main() [all …]
|
A D | hyperv_features.c | 16 #define VCPU_ID 0 macro 165 vcpu_set_cpuid(vm, VCPU_ID, cpuid); in hv_set_cpuid() 199 vcpu_enable_cap(vm, VCPU_ID, &cap); in guest_test_msrs_access() 201 vcpu_set_hv_cpuid(vm, VCPU_ID); in guest_test_msrs_access() 209 run = vcpu_state(vm, VCPU_ID); in guest_test_msrs_access() 342 vcpu_enable_cap(vm, VCPU_ID, &cap); in guest_test_msrs_access() 480 r = _vcpu_run(vm, VCPU_ID); in guest_test_msrs_access() 545 vcpu_enable_cap(vm, VCPU_ID, &cap); in guest_test_hcalls_access() 547 vcpu_set_hv_cpuid(vm, VCPU_ID); in guest_test_hcalls_access() 551 run = vcpu_state(vm, VCPU_ID); in guest_test_hcalls_access() [all …]
|
A D | userspace_io_test.c | 13 #define VCPU_ID 1 macro 65 vm = vm_create_default(VCPU_ID, 0, guest_code); in main() 66 run = vcpu_state(vm, VCPU_ID); in main() 71 rc = _vcpu_run(vm, VCPU_ID); in main() 79 if (get_ucall(vm, VCPU_ID, &uc)) in main() 92 vcpu_regs_get(vm, VCPU_ID, ®s); in main() 98 vcpu_regs_set(vm, VCPU_ID, ®s); in main()
|
A D | vmx_close_while_nested_test.c | 21 #define VCPU_ID 5 macro 59 vm = vm_create_default(VCPU_ID, 0, (void *) l1_guest_code); in main() 63 vcpu_args_set(vm, VCPU_ID, 1, vmx_pages_gva); in main() 66 volatile struct kvm_run *run = vcpu_state(vm, VCPU_ID); in main() 69 vcpu_run(vm, VCPU_ID); in main() 78 switch (get_ucall(vm, VCPU_ID, &uc)) { in main()
|
A D | svm_int_ctl_test.c | 16 #define VCPU_ID 0 macro 95 vm = vm_create_default(VCPU_ID, 0, (void *) l1_guest_code); in main() 98 vcpu_init_descriptor_tables(vm, VCPU_ID); in main() 104 vcpu_args_set(vm, VCPU_ID, 1, svm_gva); in main() 106 struct kvm_run *run = vcpu_state(vm, VCPU_ID); in main() 109 vcpu_run(vm, VCPU_ID); in main() 115 switch (get_ucall(vm, VCPU_ID, &uc)) { in main()
|
A D | svm_vmcall_test.c | 15 #define VCPU_ID 5 macro 46 vm = vm_create_default(VCPU_ID, 0, (void *) l1_guest_code); in main() 49 vcpu_args_set(vm, VCPU_ID, 1, svm_gva); in main() 52 volatile struct kvm_run *run = vcpu_state(vm, VCPU_ID); in main() 55 vcpu_run(vm, VCPU_ID); in main() 61 switch (get_ucall(vm, VCPU_ID, &uc)) { in main()
|
A D | userspace_msr_exit_test.c | 20 #define VCPU_ID 1 macro 402 rc = _vcpu_run(vm, VCPU_ID); in run_guest() 408 struct kvm_run *run = vcpu_state(vm, VCPU_ID); in check_for_guest_assert() 412 get_ucall(vm, VCPU_ID, &uc) == UCALL_ABORT) { in check_for_guest_assert() 420 struct kvm_run *run = vcpu_state(vm, VCPU_ID); in process_rdmsr() 455 struct kvm_run *run = vcpu_state(vm, VCPU_ID); in process_wrmsr() 513 switch (get_ucall(vm, VCPU_ID, &uc)) { in process_ucall() 575 vcpu_init_descriptor_tables(vm, VCPU_ID); in test_msr_filter_allow() 623 switch (get_ucall(vm, VCPU_ID, &uc)) { in handle_ucall() 690 run = vcpu_state(vm, VCPU_ID); in test_msr_filter_deny() [all …]
|
A D | hyperv_clock.c | 174 #define VCPU_ID 0 macro 181 tsc_freq = vcpu_get_msr(vm, VCPU_ID, HV_X64_MSR_TSC_FREQUENCY); in host_check_tsc_msr_rdtsc() 186 t1 = vcpu_get_msr(vm, VCPU_ID, HV_X64_MSR_TIME_REF_COUNT); in host_check_tsc_msr_rdtsc() 189 t2 = vcpu_get_msr(vm, VCPU_ID, HV_X64_MSR_TIME_REF_COUNT); in host_check_tsc_msr_rdtsc() 212 vm = vm_create_default(VCPU_ID, 0, guest_main); in main() 213 run = vcpu_state(vm, VCPU_ID); in main() 215 vcpu_set_hv_cpuid(vm, VCPU_ID); in main() 221 vcpu_args_set(vm, VCPU_ID, 2, tsc_page_gva, addr_gva2gpa(vm, tsc_page_gva)); in main() 226 _vcpu_run(vm, VCPU_ID); in main() 232 switch (get_ucall(vm, VCPU_ID, &uc)) { in main()
|
/linux/tools/testing/selftests/kvm/ |
A D | system_counter_offset_test.c | 17 #define VCPU_ID 0 macro 33 if (!_vcpu_has_device_attr(vm, VCPU_ID, KVM_VCPU_TSC_CTRL, KVM_VCPU_TSC_OFFSET)) in check_preconditions() 42 vcpu_access_device_attr(vm, VCPU_ID, KVM_VCPU_TSC_CTRL, in setup_system_counter() 105 vcpu_run(vm, VCPU_ID); in enter_guest() 108 switch (get_ucall(vm, VCPU_ID, &uc)) { in enter_guest() 117 get_ucall(vm, VCPU_ID, &uc)); in enter_guest() 126 vm = vm_create_default(VCPU_ID, 0, guest_main); in main()
|