/linux/arch/x86/kvm/vmx/ |
A D | nested.c | 286 if (!vmx->nested.vmxon && !vmx->nested.smm.vmxon) in free_nested() 291 vmx->nested.vmxon = false; in free_nested() 314 vmx->nested.pi_desc = NULL; in free_nested() 1363 if (vmx->nested.vmxon) in vmx_set_vmx_msr() 1973 vmx->nested.hv_evmcs = vmx->nested.hv_evmcs_map.hva; in nested_vmx_handle_enlightened_vmptrld() 2184 if (vmx->nested.dirty_vmcs12 || evmptr_is_valid(vmx->nested.hv_evmcs_vmptr)) in prepare_vmcs02_early() 2476 if (vmx->nested.dirty_vmcs12 || evmptr_is_valid(vmx->nested.hv_evmcs_vmptr)) { in prepare_vmcs02() 3210 vmx->nested.pi_desc = in nested_get_vmcs12_pages() 4924 if (vmx->nested.vmxon) in handle_vmon() 5040 } else if (vmx->nested.hv_evmcs && vmptr == vmx->nested.hv_evmcs_vmptr) { in handle_vmclear() [all …]
|
A D | nested.h | 42 return to_vmx(vcpu)->nested.cached_vmcs12; in get_vmcs12() 47 return to_vmx(vcpu)->nested.cached_shadow_vmcs12; in get_shadow_vmcs12() 60 return vmx->nested.current_vmptr != -1ull || in vmx_has_valid_vmcs12() 61 vmx->nested.hv_evmcs_vmptr != EVMPTR_INVALID; in vmx_has_valid_vmcs12() 68 return vmx->nested.vpid02 ? vmx->nested.vpid02 : vmx->vpid; in nested_get_vpid02() 110 return to_vmx(vcpu)->nested.msrs.misc_low & in nested_cpu_has_vmwrite_any_field() 258 u64 fixed0 = to_vmx(vcpu)->nested.msrs.cr0_fixed0; in nested_guest_cr0_valid() 259 u64 fixed1 = to_vmx(vcpu)->nested.msrs.cr0_fixed1; in nested_guest_cr0_valid() 272 u64 fixed0 = to_vmx(vcpu)->nested.msrs.cr0_fixed0; in nested_host_cr0_valid() 273 u64 fixed1 = to_vmx(vcpu)->nested.msrs.cr0_fixed1; in nested_host_cr0_valid() [all …]
|
A D | vmx.c | 1744 if (!nested) in vmx_get_msr_feature() 4159 if (nested) { in vmx_adjust_secondary_exec_control() 4284 if (nested) in init_vmcs() 4392 if (nested) in __vmx_vcpu_reset() 4393 memcpy(&vmx->nested.msrs, &vmcs_config.nested, sizeof(vmx->nested.msrs)); in __vmx_vcpu_reset() 6408 return nested; in vmx_has_emulated_msr() 6959 if (nested) in vmx_check_processor_compat() 7231 if (nested) in vmx_set_cpu_caps() 7491 vmx->nested.smm.vmxon = vmx->nested.vmxon; in vmx_enter_smm() 7541 if (nested) in hardware_unsetup() [all …]
|
/linux/arch/x86/kvm/svm/ |
A D | nested.c | 109 svm->nested.ctl.nested_cr3); in nested_svm_init_mmu_context() 134 g = &svm->nested.ctl; in recalc_intercepts() 457 if (!svm->nested.vmcb02.ptr) in nested_vmcb02_compute_g_pat() 471 if (svm->nested.vmcb12_gpa != svm->nested.last_vmcb12_gpa) { in nested_vmcb02_prepare_save() 473 svm->nested.last_vmcb12_gpa = svm->nested.vmcb12_gpa; in nested_vmcb02_prepare_save() 558 svm->nested.ctl.tsc_offset, in nested_vmcb02_prepare_control() 655 if (!svm->nested.hsave_msr) { in nested_svm_vmrun() 788 svm->nested.vmcb12_gpa = 0; in nested_svm_vmexit() 919 if (svm->nested.initialized) in svm_allocate_nested() 929 if (!svm->nested.msrpm) in svm_allocate_nested() [all …]
|
A D | svm.h | 172 struct svm_nested_state nested; member 453 return is_guest_mode(vcpu) && (svm->nested.ctl.int_ctl & V_INTR_MASKING_MASK); in nested_svm_virtualize_tpr() 458 return vmcb_is_intercept(&svm->nested.ctl, INTERCEPT_SMI); in nested_exit_on_smi() 463 return vmcb_is_intercept(&svm->nested.ctl, INTERCEPT_INTR); in nested_exit_on_intr() 468 return vmcb_is_intercept(&svm->nested.ctl, INTERCEPT_NMI); in nested_exit_on_nmi()
|
A D | svm.c | 176 static int nested = true; variable 177 module_param(nested, int, S_IRUGO); 936 if (nested) { in svm_set_cpu_caps() 1016 if (nested) { in svm_hardware_setup() 1135 return svm->nested.ctl.tsc_offset; in svm_get_l2_tsc_offset() 1320 svm->nested.vmcb12_gpa = INVALID_GPA; in init_vmcb() 2871 svm->nested.vm_cr_msr &= ~chg_mask; in svm_set_vm_cr() 3542 if (svm->nested.nested_run_pending) in svm_nmi_allowed() 3601 if (svm->nested.nested_run_pending) in svm_interrupt_allowed() 3957 svm->nested.nested_run_pending = 0; in svm_vcpu_run() [all …]
|
/linux/Documentation/virt/kvm/ |
A D | running-nested-guests.rst | 2 Running nested guests with KVM 74 Enabling "nested" (x86) 96 parm: nested:bool 103 options kvm-intel nested=y 142 Starting a nested guest (x86) 157 nested guest with accelerated KVM. 160 Enabling "nested" (s390x) 167 $ modprobe kvm nested=1 182 Live migration with nested KVM 211 Reporting bugs from nested setups [all …]
|
A D | nested-vmx.rst | 16 hypervisors (which use VMX) with their own nested guests. It does so by 20 We describe in much greater detail the theory behind the nested VMX feature, 32 In nested virtualization, we have three levels: The host (KVM), which we call 33 L0, the guest hypervisor, which we call L1, and its nested guest, which we 37 Running nested VMX 40 The nested VMX feature is enabled by default since Linux kernel v4.20. For 41 older Linux kernel, it can be enabled by giving the "nested=1" option to the 65 As a VMX implementation, nested VMX presents a VMCS structure to L1.
|
A D | index.rst | 18 nested-vmx 32 running-nested-guests
|
/linux/Documentation/networking/ |
A D | ethtool-netlink.rst | 109 Compact form: nested (bitset) atrribute contents: 134 Bit-by-bit form: nested (bitset) attribute contents: 371 ``ETHTOOL_A_LINKINFO_HEADER`` nested reply header 483 ``ETHTOOL_A_LINKSTATE_HEADER`` nested reply header 649 ``ETHTOOL_A_DEBUG_HEADER`` nested reply header 697 ``ETHTOOL_A_WOL_HEADER`` nested reply header 738 ``ETHTOOL_A_FEATURES_HEADER`` nested reply header 770 ``ETHTOOL_A_FEATURES_HEADER`` nested reply header 810 ``ETHTOOL_A_PRIVFLAGS_HEADER`` nested reply header 853 ``ETHTOOL_A_RINGS_HEADER`` nested reply header [all …]
|
/linux/tools/perf/bench/ |
A D | epoll-wait.c | 104 static unsigned int nested = 0; variable 137 …OPT_UINTEGER( 'N', "nested", &nested, "Nesting level epoll hierarchy (default is 0, no nesting)… 245 if (nested > EPOLL_MAXNESTS) in nest_epollfd() 246 nested = EPOLL_MAXNESTS; in nest_epollfd() 248 epollfdp = calloc(nested, sizeof(*epollfdp)); in nest_epollfd() 252 for (i = 0; i < nested; i++) { in nest_epollfd() 261 for (i = nested - 1; i; i--) { in nest_epollfd() 317 if (nested) in do_threads() 446 if (nested) in bench_epoll_wait() 451 printinfo("Nesting level(s): %d\n", nested); in bench_epoll_wait()
|
A D | epoll-ctl.c | 56 static unsigned int nested = 0; variable 78 …OPT_UINTEGER( 'N', "nested", &nested, "Nesting level epoll hierarchy (default is 0, no nesting)… 104 if (nested > EPOLL_MAXNESTS) in nest_epollfd() 105 nested = EPOLL_MAXNESTS; in nest_epollfd() 106 printinfo("Nesting level(s): %d\n", nested); in nest_epollfd() 108 epollfdp = calloc(nested, sizeof(int)); in nest_epollfd() 112 for (i = 0; i < nested; i++) { in nest_epollfd() 121 for (i = nested - 1; i; i--) { in nest_epollfd() 331 if (nested) in bench_epoll_ctl()
|
/linux/net/netfilter/ipset/ |
A D | ip_set_bitmap_gen.h | 95 struct nlattr *nested; in mtype_head() local 98 nested = nla_nest_start(skb, IPSET_ATTR_DATA); in mtype_head() 99 if (!nested) in mtype_head() 108 nla_nest_end(skb, nested); in mtype_head() 207 struct nlattr *adt, *nested; in mtype_list() local 229 nested = nla_nest_start(skb, IPSET_ATTR_DATA); in mtype_list() 230 if (!nested) { in mtype_list() 243 nla_nest_end(skb, nested); in mtype_list() 253 nla_nest_cancel(skb, nested); in mtype_list()
|
A D | ip_set_list_set.c | 462 struct nlattr *nested; in list_set_head() local 465 nested = nla_nest_start(skb, IPSET_ATTR_DATA); in list_set_head() 466 if (!nested) in list_set_head() 475 nla_nest_end(skb, nested); in list_set_head() 487 struct nlattr *atd, *nested; in list_set_list() local 505 nested = nla_nest_start(skb, IPSET_ATTR_DATA); in list_set_list() 506 if (!nested) in list_set_list() 513 nla_nest_end(skb, nested); in list_set_list() 523 nla_nest_cancel(skb, nested); in list_set_list()
|
/linux/scripts/coccinelle/locks/ |
A D | flags.cocci | 2 /// Find nested lock+irqsave functions that use the same flags variables 73 cocci.print_secs("nested lock+irqsave that reuses flags",p2) 80 msg="ERROR: nested lock+irqsave that reuses flags from line %s." % (p1[0].line)
|
/linux/arch/um/os-Linux/ |
A D | signal.c | 184 int nested, bail; in hard_handler() local 200 nested = pending & 1; in hard_handler() 215 if (!nested) in hard_handler() 216 pending = from_irq_stack(nested); in hard_handler()
|
/linux/arch/powerpc/kvm/ |
A D | book3s_hv_p9_entry.c | 97 struct kvm_nested_guest *nested = vcpu->arch.nested; in switch_mmu_to_guest_radix() local 100 lpid = nested ? nested->shadow_lpid : kvm->arch.lpid; in switch_mmu_to_guest_radix() 189 struct kvm_nested_guest *nested = vcpu->arch.nested; in kvmhv_vcpu_entry_p9() local 312 kvmppc_check_need_tlb_flush(kvm, vc->pcpu, nested); in kvmhv_vcpu_entry_p9()
|
A D | book3s_64_mmu_radix.c | 101 if (vcpu->arch.nested) in kvmhv_copy_tofrom_guest_radix() 797 if (nested) in kvmppc_hv_handle_set_rc() 1293 struct kvm_nested_guest *nested; in debugfs_radix_read() local 1329 nested = NULL; in debugfs_radix_read() 1335 if (nested) { in debugfs_radix_read() 1336 kvmhv_put_nested(nested); in debugfs_radix_read() 1337 nested = NULL; in debugfs_radix_read() 1349 if (!nested) { in debugfs_radix_read() 1353 pgt = nested->shadow_pgtable; in debugfs_radix_read() 1428 if (nested) in debugfs_radix_read() [all …]
|
/linux/Documentation/x86/ |
A D | kernel-stacks.rst | 28 hardware interrupt (i.e. not a nested hardware interrupt) then the 54 nested IST interrupts then the handler must adjust the IST values on 59 nested. For example, a debug interrupt can safely be interrupted by an 62 IST events with the same code to be nested. However in most cases, the 95 To handle nested #DB correctly there exist two instances of DB stacks. On 97 so a nested #DB starts from a clean stack. The nested #DB switches
|
/linux/Documentation/filesystems/ |
A D | ceph.rst | 23 * Recursive accounting (nested files, directories, bytes) 56 a snapshot on any subdirectory (and its nested contents) in the 60 Ceph also provides some recursive accounting on directories for nested 62 system will reveal the total number of nested regular files and 63 subdirectories, and a summation of all nested file sizes. This makes 129 the summation of file sizes over all files nested beneath that
|
/linux/arch/um/kernel/ |
A D | irq.c | 711 int nested; in to_irq_stack() local 733 nested = (ti->real_thread != NULL); in to_irq_stack() 734 if (!nested) { in to_irq_stack() 747 *mask_out |= mask | nested; in to_irq_stack() 751 unsigned long from_irq_stack(int nested) in from_irq_stack() argument
|
/linux/arch/x86/kvm/ |
A D | Makefile | 30 vmx/evmcs.o vmx/nested.o vmx/posted_intr.o 33 kvm-amd-y += svm/svm.o svm/vmenter.o svm/pmu.o svm/nested.o svm/avic.o svm/sev.o
|
/linux/scripts/coccinelle/misc/ |
A D | irqf_oneshot.cocci | 106 …ING: Threaded IRQ with no primary handler requested without IRQF_ONESHOT (unless it is nested IRQ)" 112 …ING: Threaded IRQ with no primary handler requested without IRQF_ONESHOT (unless it is nested IRQ)"
|
/linux/Documentation/admin-guide/hw-vuln/ |
A D | l1tf.rst | 556 When nested virtualization is in use, three operating systems are involved: 557 the bare metal hypervisor, the nested hypervisor and the nested virtual 558 machine. VMENTER operations from the nested hypervisor into the nested 562 - Flush the L1D cache on every switch from the nested hypervisor to the 563 nested virtual machine, so that the nested hypervisor's secrets are not 564 exposed to the nested virtual machine; 566 - Flush the L1D cache on every switch from the nested virtual machine to 567 the nested hypervisor; this is a complex operation, and flushing the L1D 569 nested virtual machine; 571 - Instruct the nested hypervisor to not perform any L1D cache flush. This
|
/linux/Documentation/trace/ |
A D | ring-buffer-design.rst | 50 - a pointer to the page with the last finished non-nested write. 708 against nested writers:: 732 tail page (moved by nested writer) 765 But if a nested writer preempts here, it will see that the next 766 page is a head page, but it is also nested. It will detect that 767 it is nested and will save that information. The detection is the 771 The nested writer will set the new head page pointer:: 793 After the nested writer finishes, the outermost writer will convert 806 It can be even more complex if several nested writes came in and moved 843 The nested writer moves the tail page forward. But does not set the old [all …]
|