Searched refs:is_hvm_vcpu (Results 1 – 25 of 25) sorted by relevance
17 (is_hvm_vcpu(current) ? \21 (is_hvm_vcpu(current) ? \25 (is_hvm_vcpu(current) ? \29 (is_hvm_vcpu(current) ? \33 (is_hvm_vcpu(current) ? \37 (is_hvm_vcpu(current) ? \
29 return (is_hvm_vcpu(v) ? hvm_local_events_need_delivery(v) : in local_events_need_delivery()
287 if ( !is_hvm_vcpu(v) ) in __core2_vpmu_save()295 if ( !is_hvm_vcpu(v) ) in core2_vpmu_save()343 if ( !is_hvm_vcpu(v) ) in __core2_vpmu_load()394 !(is_hvm_vcpu(v) in core2_vpmu_verify()455 if ( is_hvm_vcpu(v) ) in core2_vpmu_alloc_resource()519 if ( is_hvm_vcpu(current) && cpu_has_vmx_msr_bitmap ) in core2_vpmu_msr_common_check()593 if ( is_hvm_vcpu(v) ) in core2_vpmu_do_wrmsr()662 if ( is_hvm_vcpu(v) ) in core2_vpmu_do_wrmsr()681 if ( is_hvm_vcpu(v) ) in core2_vpmu_do_wrmsr()715 if ( is_hvm_vcpu(v) ) in core2_vpmu_do_rdmsr()[all …]
308 if ( !vpmu_is_set(vpmu, VPMU_RUNNING) && is_hvm_vcpu(v) && in amd_vpmu_save()370 if ( is_hvm_vcpu(v) && (type == MSR_TYPE_CTRL) && in amd_vpmu_do_wrmsr()384 if ( is_hvm_vcpu(v) && is_msr_bitmap_on(vpmu) ) in amd_vpmu_do_wrmsr()393 if ( is_hvm_vcpu(v) && is_msr_bitmap_on(vpmu) ) in amd_vpmu_do_wrmsr()436 if ( is_hvm_vcpu(v) && is_msr_bitmap_on(vpmu) ) in amd_vpmu_destroy()
224 if ( is_hvm_vcpu(sampled) ) in vpmu_do_interrupt()275 if ( !is_hvm_vcpu(sampled) ) in vpmu_do_interrupt()
174 (is_hvm_vcpu(v) ? __linear_l1_table : __sh_linear_l1_table) + \182 (is_hvm_vcpu(v) ? __linear_l2_table : __sh_linear_l2_table) + \191 (is_hvm_vcpu(v) ? __linear_l3_table : __sh_linear_l3_table) + \
211 if ( (addr & (bytes - 1)) && !is_hvm_vcpu(v) ) in hvm_emulate_write()265 if ( (addr & (bytes - 1)) && !is_hvm_vcpu(v) ) in hvm_emulate_cmpxchg()320 ASSERT(is_hvm_vcpu(curr)); in shadow_init_emulation()366 ASSERT(is_hvm_vcpu(current)); in shadow_continue_emulation()
2898 ASSERT(is_hvm_vcpu(v)); in sh_page_fault()3488 ASSERT(is_hvm_vcpu(v)); in sh_page_fault()4597 if ( (sflags & _PAGE_GLOBAL) && !is_hvm_vcpu(v) ) in sh_audit_flags()
49 if (is_hvm_vcpu(vcpu)) in is_32bit_vcpu()
82 if ( !is_hvm_vcpu(curr) ) in xenoprofile_get_mode()
329 ASSERT(is_hvm_vcpu(v)); in athlon_check_ctrs()
52 if ( is_hvm_vcpu(v) ) in compat_vcpu_op()
57 (is_hvm_vcpu(v) ? hvm_get_guest_time(v) : (u64)get_s_time())184 if ( (v == NULL) || !is_hvm_vcpu(v) || (channel != 0) ) in pit_load_count()
441 if ( is_hvm_vcpu(v) ) in arch_vcpu_destroy()1228 if ( is_hvm_vcpu(v) ) in arch_initialise_vcpu()1728 if ( cpu_has_xsaves && is_hvm_vcpu(n) ) in __context_switch()2217 if ( is_hvm_vcpu(v) ) in cpuid_policy_updated()2259 if ( is_hvm_vcpu(v) ) in vcpu_mark_events_pending()
204 ASSERT(is_hvm_vcpu(curr)); in vm_event_fill_regs()
304 if ( is_hvm_vcpu(v) ) in show_guest_stack()
103 if ( guest_mode(regs) && is_hvm_vcpu(v) ) in show_registers()
87 is_hvm_vcpu(v) ? "HVM" : "PV", ctxt->caps, in vmce_restore_vcpu()
155 if ( is_hvm_vcpu(v) ) in elf_memcpy()
45 ASSERT(is_hvm_vcpu(v)); in hvm_get_guest_time_fixed()
3020 if ( !is_hvm_vcpu(v) || cache->num_ents > cache->max_ents ) in hvmemul_read_cache()3052 if ( !is_hvm_vcpu(v) || cache->num_ents > cache->max_ents ) in hvmemul_write_cache()
3272 ASSERT(is_hvm_vcpu(v)); in __hvm_copy()
1036 static always_inline bool is_hvm_vcpu(const struct vcpu *v) in is_hvm_vcpu() function
854 if ( is_hvm_vcpu(current) ) in vmx_vmcs_exit()
1775 if ( is_hvm_vcpu(v) && paging_mode_hap(v->domain) && nestedhvm_is_n2(v) ) in paging_gva_to_gfn()
Completed in 87 milliseconds