/linux/arch/x86/include/asm/ |
A D | virtext.h | 125 uint64_t efer; in cpu_svm_disable() local 128 rdmsrl(MSR_EFER, efer); in cpu_svm_disable() 129 wrmsrl(MSR_EFER, efer & ~EFER_SVME); in cpu_svm_disable()
|
A D | suspend_64.h | 41 unsigned long efer; member
|
A D | realmode.h | 51 u64 efer;
|
A D | svm.h | 244 u64 efer; member
|
A D | kvm_host.h | 629 u64 efer; member 1342 int (*set_efer)(struct kvm_vcpu *vcpu, u64 efer); 1680 bool kvm_valid_efer(struct kvm_vcpu *vcpu, u64 efer);
|
/linux/arch/x86/realmode/ |
A D | init.c | 74 u64 efer; in setup_real_mode() local 122 rdmsrl(MSR_EFER, efer); in setup_real_mode() 123 trampoline_header->efer = efer & ~EFER_LMA; in setup_real_mode()
|
/linux/tools/testing/selftests/kvm/lib/x86_64/ |
A D | svm.c | 81 uint64_t efer; in generic_svm_setup() local 83 efer = rdmsr(MSR_EFER); in generic_svm_setup() 84 wrmsr(MSR_EFER, efer | EFER_SVME); in generic_svm_setup() 98 save->efer = rdmsr(MSR_EFER); in generic_svm_setup()
|
A D | processor.c | 168 sregs->cr8, sregs->efer, sregs->apic_base); in sregs_dump() 309 if ((sregs.efer & EFER_NX) == 0) { in _vm_get_page_table_entry() 636 sregs.efer |= (EFER_LME | EFER_LMA | EFER_NX); in vcpu_setup()
|
/linux/include/xen/interface/hvm/ |
A D | hvm_vcpu.h | 35 uint64_t efer; member 89 uint64_t efer; member
|
/linux/arch/x86/kvm/svm/ |
A D | nested.c | 108 svm->vmcb01.ptr->save.efer, in nested_svm_init_mmu_context() 286 if ((save->efer & EFER_LME) && (save->cr0 & X86_CR0_PG)) { in nested_vmcb_check_cr3_cr4() 310 if (CC(!(save->efer & EFER_SVME))) in nested_vmcb_valid_sregs() 323 if (CC(!kvm_valid_efer(vcpu, save->efer))) in nested_vmcb_valid_sregs() 498 svm_set_efer(&svm->vcpu, vmcb12->save.efer | EFER_SVME); in nested_vmcb02_prepare_save() 696 svm->vmcb01.ptr->save.efer = vcpu->arch.efer; in nested_svm_vmrun() 740 to_save->efer = from_save->efer; in svm_copy_vmrun_state() 804 vmcb12->save.efer = svm->vcpu.arch.efer; in nested_svm_vmexit() 863 svm_set_efer(vcpu, svm->vmcb->save.efer); in nested_svm_vmexit() 1094 if (!(vcpu->arch.efer & EFER_SVME) || !is_paging(vcpu)) { in nested_svm_check_permissions() [all …]
|
A D | svm.c | 281 vcpu->arch.efer = efer; in svm_set_efer() 285 efer |= EFER_NX; in svm_set_efer() 287 if (!(efer & EFER_LMA)) in svm_set_efer() 288 efer &= ~EFER_LME; in svm_set_efer() 292 if (!(efer & EFER_SVME)) { in svm_set_efer() 311 vcpu->arch.efer = old_efer; in svm_set_efer() 320 svm->vmcb->save.efer = efer | EFER_SVME; in svm_set_efer() 490 uint64_t efer; in svm_hardware_enable() local 494 rdmsrl(MSR_EFER, efer); in svm_hardware_enable() 495 if (efer & EFER_SVME) in svm_hardware_enable() [all …]
|
A D | svm.h | 430 int svm_set_efer(struct kvm_vcpu *vcpu, u64 efer);
|
/linux/arch/x86/power/ |
A D | cpu.c | 114 rdmsrl(MSR_EFER, ctxt->efer); in __save_processor_state() 207 wrmsrl(MSR_EFER, ctxt->efer); in __restore_processor_state()
|
/linux/arch/x86/include/uapi/asm/ |
A D | kvm.h | 157 __u64 efer; member 168 __u64 efer; member
|
/linux/tools/arch/x86/include/uapi/asm/ |
A D | kvm.h | 157 __u64 efer; member 168 __u64 efer; member
|
/linux/arch/x86/kvm/ |
A D | emulate.c | 813 u64 efer = 0; in assign_eip_far() local 816 if (efer & EFER_LMA) in assign_eip_far() 1497 u64 efer = 0; in get_descriptor_ptr() local 1648 u64 efer = 0; in __load_segment_descriptor() local 1651 if (efer & EFER_LMA) in __load_segment_descriptor() 2571 efer = 0; in em_rsm() 2685 u64 efer = 0; in em_syscall() local 2705 if (efer & EFER_LMA) { in em_syscall() 2744 u64 efer = 0; in em_sysenter() local 2785 if (efer & EFER_LMA) in em_sysenter() [all …]
|
A D | x86.c | 1582 if (efer & efer_reserved_bits) in kvm_valid_efer() 1591 u64 old_efer = vcpu->arch.efer; in set_efer() 1592 u64 efer = msr_info->data; in set_efer() local 1595 if (efer & efer_reserved_bits) in set_efer() 1603 (vcpu->arch.efer & EFER_LME) != (efer & EFER_LME)) in set_efer() 1607 efer &= ~EFER_LMA; in set_efer() 1608 efer |= vcpu->arch.efer & EFER_LMA; in set_efer() 1617 if ((efer ^ old_efer) & EFER_NX) in set_efer() 10381 sregs->efer = vcpu->arch.efer; in __get_sregs_common() 10563 *mmu_reset_needed |= vcpu->arch.efer != sregs->efer; in __set_sregs_common() [all …]
|
A D | mmu.h | 72 unsigned long cr4, u64 efer, gpa_t nested_cr3);
|
A D | x86.h | 147 return vcpu->arch.efer & EFER_LMA; in is_long_mode()
|
/linux/tools/testing/selftests/kvm/include/x86_64/ |
A D | svm.h | 178 u64 efer; member
|
/linux/drivers/parport/ |
A D | parport_pc.c | 1095 static void decode_winbond(int efer, int key, int devid, int devrev, int oldid) in decode_winbond() argument 1140 efer, key, devid, devrev, oldid, type); in decode_winbond() 1143 show_parconfig_winbond(efer, key); in decode_winbond() 1146 static void decode_smsc(int efer, int key, int devid, int devrev) in decode_smsc() argument 1172 efer, key, devid, devrev, type); in decode_smsc() 1175 func(efer, key); in decode_smsc()
|
/linux/arch/x86/kvm/vmx/ |
A D | vmx.c | 919 u64 guest_efer = vmx->vcpu.arch.efer; in update_transition_efer() 1663 (vmx->vcpu.arch.efer & EFER_SCE); in vmx_setup_uret_msrs() 2845 int vmx_set_efer(struct kvm_vcpu *vcpu, u64 efer) in vmx_set_efer() argument 2854 vcpu->arch.efer = efer; in vmx_set_efer() 2855 if (efer & EFER_LMA) { in vmx_set_efer() 2857 msr->data = efer; in vmx_set_efer() 2861 msr->data = efer & ~EFER_LME; in vmx_set_efer() 2883 vmx_set_efer(vcpu, vcpu->arch.efer | EFER_LMA); in enter_lmode() 2889 vmx_set_efer(vcpu, vcpu->arch.efer & ~EFER_LMA); in exit_lmode() 3023 if (vcpu->arch.efer & EFER_LME) { in vmx_set_cr0() [all …]
|
A D | nested.c | 2102 return vmx->vcpu.arch.efer | (EFER_LMA | EFER_LME); in nested_vmx_calc_efer() 2104 return vmx->vcpu.arch.efer & ~(EFER_LMA | EFER_LME); in nested_vmx_calc_efer() 2546 vcpu->arch.efer = nested_vmx_calc_efer(vmx, vmcs12); in prepare_vmcs02() 2548 vmx_set_efer(vcpu, vcpu->arch.efer); in prepare_vmcs02() 2843 !!(vcpu->arch.efer & EFER_LMA))) in nested_vmx_check_address_space_size() 4151 vmcs12->guest_ia32_efer = vcpu->arch.efer; in sync_vmcs02_to_vmcs12() 4231 vcpu->arch.efer = vmcs12->host_ia32_efer; in load_vmcs12_host_state() 4233 vcpu->arch.efer |= (EFER_LMA | EFER_LME); in load_vmcs12_host_state() 4235 vcpu->arch.efer &= ~(EFER_LMA | EFER_LME); in load_vmcs12_host_state() 4236 vmx_set_efer(vcpu, vcpu->arch.efer); in load_vmcs12_host_state()
|
A D | vmx.h | 373 int vmx_set_efer(struct kvm_vcpu *vcpu, u64 efer);
|
/linux/arch/x86/kvm/mmu/ |
A D | mmu.c | 199 const u64 efer; member 223 BUILD_MMU_ROLE_REGS_ACCESSOR(efer, nx, EFER_NX); 224 BUILD_MMU_ROLE_REGS_ACCESSOR(efer, lma, EFER_LMA); 245 BUILD_MMU_ROLE_ACCESSOR(base, efer, nx); 252 .efer = vcpu->arch.efer, in vcpu_to_role_regs() 4886 unsigned long cr4, u64 efer, gpa_t nested_cr3) in kvm_init_shadow_npt_mmu() argument 4892 .efer = efer, in kvm_init_shadow_npt_mmu()
|