/xen/xen/include/xsm/ |
A D | dummy.h | 110 return xsm_default_action(action, current->domain, d); in xsm_domain_create() 116 return xsm_default_action(action, current->domain, d); in xsm_getdomaininfo() 122 return xsm_default_action(action, current->domain, d); in xsm_domctl_scheduler_op() 216 return xsm_default_action(action, current->domain, d); in xsm_memory_exchange() 272 return xsm_default_action(action, current->domain, d); in xsm_claim_pages() 279 return xsm_default_action(action, current->domain, d); in xsm_evtchn_unbound() 303 return xsm_default_action(action, current->domain, d); in xsm_evtchn_status() 330 return xsm_default_action(action, current->domain, d); in xsm_init_hardware_domain() 336 return xsm_default_action(action, current->domain, d); in xsm_get_pod_target() 342 return xsm_default_action(action, current->domain, d); in xsm_set_pod_target() [all …]
|
/xen/xen/include/asm-arm/ |
A D | event.h | 27 if ( vgic_vcpu_pending_irq(current) ) in local_events_need_delivery_nomask() 30 if ( !vcpu_info(current, evtchn_upcall_pending) ) in local_events_need_delivery_nomask() 33 return vgic_evtchn_irq_pending(current); in local_events_need_delivery_nomask() 38 if ( !vcpu_event_delivery_is_enabled(current) ) in local_events_need_delivery()
|
/xen/xen/arch/x86/ |
A D | x86_emulate.c | 46 #define FXSAVE_AREA current->arch.fpu_ctxt 62 *val = current->arch.xcr0; in x86emul_read_xcr() 66 if ( current->domain->arch.cpuid->xstate.xgetbv1 ) in x86emul_read_xcr() 106 struct vcpu *curr = current; in x86emul_read_dr() 149 struct vcpu *curr = current; in x86emul_write_dr() 173 guest_cpuid(current, leaf, subleaf, res); in x86emul_cpuid()
|
A D | trace.c | 11 if ( is_pv_32bit_vcpu(current) ) in __trace_pv_trap() 52 if ( is_pv_32bit_vcpu(current) ) in __trace_pv_page_fault() 83 if ( is_pv_32bit_vcpu(current) ) in __trace_trap_one_addr() 98 if ( is_pv_32bit_vcpu(current) ) in __trace_trap_two_addr() 131 if ( is_pv_32bit_vcpu(current) ) in __trace_ptwr_emulation()
|
A D | compat.c | 46 current->domain->domain_id, current->vcpu_id, arg); in do_sched_op_compat() 47 domain_shutdown(current->domain, (u8)arg); in do_sched_op_compat()
|
/xen/xen/include/asm-x86/ |
A D | guest_access.h | 17 (is_hvm_vcpu(current) ? \ 21 (is_hvm_vcpu(current) ? \ 25 (is_hvm_vcpu(current) ? \ 29 (is_hvm_vcpu(current) ? \ 33 (is_hvm_vcpu(current) ? \ 37 (is_hvm_vcpu(current) ? \ 121 (paging_mode_external(current->domain) || \ 124 (paging_mode_external(current->domain) || \
|
A D | event.h | 25 struct vcpu *v = current; in local_events_need_delivery() 36 vcpu_info(current, evtchn_upcall_mask) = 1; in local_event_delivery_disable() 41 vcpu_info(current, evtchn_upcall_mask) = 0; in local_event_delivery_enable()
|
/xen/xen/arch/x86/mm/ |
A D | altp2m.c | 26 if ( v != current ) in altp2m_vcpu_initialise() 34 if ( v != current ) in altp2m_vcpu_initialise() 43 if ( v != current ) in altp2m_vcpu_destroy() 54 if ( v != current ) in altp2m_vcpu_destroy()
|
/xen/xen/arch/x86/hvm/ |
A D | monitor.c | 38 struct vcpu *curr = current; in hvm_monitor_cr() 65 struct vcpu *curr = current; in hvm_monitor_emul_unimplemented() 82 struct vcpu *curr = current; in hvm_monitor_msr() 118 monitor_traps(current, true, &req); in hvm_monitor_descriptor_access() 123 struct vcpu *curr = current; in gfn_of_rip() 144 struct vcpu *curr = current; in hvm_monitor_debug() 198 struct vcpu *curr = current; in hvm_monitor_cpuid() 224 monitor_traps(current, 1, &req); in hvm_monitor_interrupt() 236 struct vcpu *curr = current; in hvm_monitor_check_p2m()
|
A D | intercept.c | 41 if ( !handler->mmio.ops->check(current, first) ) in hvm_mmio_accept() 47 !handler->mmio.ops->check(current, last) ) in hvm_mmio_accept() 48 domain_crash(current->domain); in hvm_mmio_accept() 58 return handler->mmio.ops->read(current, addr, size, data); in hvm_mmio_read() 66 return handler->mmio.ops->write(current, addr, size, data); in hvm_mmio_write() 139 &data, p->size, current) ) in hvm_process_io_intercept() 153 domain_crash(current->domain); in hvm_process_io_intercept() 167 struct vcpu *curr = current; in hvm_process_io_intercept() 224 struct domain *curr_d = current->domain; in hvm_find_io_handler()
|
/xen/xen/xsm/flask/ |
A D | flask_op.c | 107 rv = domain_has_security(current->domain, SECURITY__COMPUTE_CREATE); in flask_security_create() 121 rv = domain_has_security(current->domain, SECURITY__COMPUTE_AV); in flask_security_access() 158 rv = domain_has_security(current->domain, SECURITY__SETENFORCE); in flask_security_setenforce() 177 rv = domain_has_security(current->domain, SECURITY__CHECK_CONTEXT); in flask_security_context() 201 rv = domain_has_security(current->domain, SECURITY__CHECK_CONTEXT); in flask_security_sid() 293 rv = domain_has_security(current->domain, SECURITY__SETBOOL); in flask_security_set_bool() 398 rv = domain_has_security(current->domain, SECURITY__SETBOOL); in flask_security_commit_bools() 506 rv = domain_has_security(current->domain, perm); in flask_devicetree_label() 553 struct domain *d = current->domain; in flask_get_peer_sid() 584 struct domain_security_struct *csec = current->domain->ssid; in flask_relabel_domain() [all …]
|
A D | hooks.c | 82 u32 csid = domain_sid(current->domain); in avc_current_has_perm() 88 return domain_has_perm(current->domain, d, class, perms); in current_has_perm() 486 return domain_has_xen(current->domain, XEN__KEXEC); in flask_kexec() 506 if ( is_idle_domain(current->domain) && !dom0_created ) in flask_domain_create() 783 return domain_has_xen(current->domain, XEN__HEAP); in flask_sysctl() 838 return domain_has_xen(current->domain, perms); in flask_readconsole() 1045 data.ssid = domain_sid(current->domain); in flask_iomem_permission() 1238 if ( d == current->domain ) in flask_hvm_altp2mhvm_op() 1477 return domain_has_xen(current->domain, XEN__MCA_OP); in flask_do_mca() 1548 data.ssid = domain_sid(current->domain); in flask_ioport_permission() [all …]
|
/xen/xen/arch/x86/pv/ |
A D | mm.c | 37 ASSERT(!paging_mode_translate(current->domain)); in map_guest_l1e() 38 ASSERT(!paging_mode_external(current->domain)); in map_guest_l1e() 64 struct vcpu *curr = current; in guest_get_eff_kern_l1e() 86 struct vcpu *curr = current; in pv_map_ldt_shadow_page()
|
A D | descriptor-tables.c | 40 ASSERT(v == current || !vcpu_cpu_dirty(v)); in pv_destroy_ldt() 68 ASSERT(v == current || !vcpu_cpu_dirty(v)); in pv_destroy_gdt() 91 ASSERT(v == current || !vcpu_cpu_dirty(v)); in pv_set_gdt() 132 struct vcpu *curr = current; in do_set_gdt() 155 struct vcpu *curr = current; in compat_set_gdt() 190 struct domain *currd = current->domain; in do_update_descriptor()
|
A D | emul-inv-op.c | 31 const struct vcpu_msrs *msrs = current->arch.msrs; in emulate_forced_invalid_op() 56 !guest_kernel_mode(current, regs) ) in emulate_forced_invalid_op() 65 guest_cpuid(current, regs->eax, regs->ecx, &res); in emulate_forced_invalid_op()
|
/xen/xen/common/ |
A D | wait.c | 125 struct vcpu *curr = current; in __prepare_to_wait() 181 vcpu_temporary_affinity(current, NR_CPUS, VCPU_AFFINITY_WAIT); in __finish_wait() 186 struct vcpu *curr = current; in check_wakeup_from_wait() 229 struct vcpu *curr = current; in prepare_to_wait() 245 struct vcpu *curr = current; in finish_wait()
|
A D | xenoprof.c | 218 BUG_ON(page_get_owner(page) != current->domain); in unshare_xenoprof_page_with_guest() 501 if ( !paging_mode_translate(current->domain) ) in add_passive_list() 505 current->domain, __pa(d->xenoprof->rawbuf), in add_passive_list() 648 struct domain *d = current->domain; in xenoprof_op_init() 669 xenoprof_primary_profiler = current->domain; in xenoprof_op_init() 681 struct domain *d = current->domain; in xenoprof_op_get_buffer() 740 ret = xsm_profile(XSM_HOOK, current->domain, op); in do_xenoprof_op() 848 xenoprof_reset_buf(current->domain); in do_xenoprof_op() 849 ret = set_active(current->domain); in do_xenoprof_op() 892 (is_active(current->domain)) ) in do_xenoprof_op() [all …]
|
A D | domctl.c | 226 spin_unlock(¤t->domain->hypercall_deadlock_mutex); in domctl_lock_acquire() 233 spin_unlock(¤t->domain->hypercall_deadlock_mutex); in domctl_lock_release() 424 if ( (d == current->domain) || /* no domain_pause() */ in do_domctl() 474 if ( d != current->domain ) in do_domctl() 483 if ( d == current->domain ) /* no domain_pause() */ in do_domctl() 538 if ( (d == current->domain) || /* no domain_pause() */ in do_domctl() 566 if ( d == current->domain ) /* no domain_pause() */ in do_domctl() 668 v == current ) /* no vcpu_pause() */ in do_domctl() 770 if ( pirq >= current->domain->nr_pirqs ) in do_domctl() 775 irq = pirq_access_permitted(current->domain, pirq); in do_domctl() [all …]
|
/xen/xen/arch/arm/ |
A D | time.c | 258 if ( unlikely(is_idle_vcpu(current)) ) in vtimer_interrupt() 263 current->arch.virt_timer.ctl = READ_SYSREG32(CNTV_CTL_EL0); in vtimer_interrupt() 264 WRITE_SYSREG32(current->arch.virt_timer.ctl | CNTx_CTL_MASK, CNTV_CTL_EL0); in vtimer_interrupt() 265 vgic_inject_irq(current->domain, current, current->arch.virt_timer.irq, true); in vtimer_interrupt()
|
A D | domain.c | 312 ASSERT(prev != current); in schedule_tail() 316 ctxt_switch_to(current); in schedule_tail() 322 update_runstate_area(current); in schedule_tail() 325 update_vcpu_system_time(current); in schedule_tail() 331 processor_vcpu_initialise(current); in continue_new_vcpu() 335 if ( is_idle_vcpu(current) ) in continue_new_vcpu() 413 current->hcall_preempted = true; in hypercall_create_continuation() 485 domain_crash(current->domain); in hypercall_create_continuation() 493 struct vcpu *v = current; in startup_cpu_idle_loop() 1099 vcpu_unblock(current); in vcpu_block_unless_event_pending() [all …]
|
A D | vpsci.c | 28 struct domain *d = current->domain; in do_common_cpu_on() 115 struct vcpu *v = current; in do_psci_cpu_off() 134 struct vcpu *v = current; in do_psci_0_2_cpu_suspend() 169 struct domain *d = current->domain; in do_psci_0_2_affinity_info() 201 struct domain *d = current->domain; in do_psci_0_2_system_off() 207 struct domain *d = current->domain; in do_psci_0_2_system_reset()
|
A D | guestcopy.c | 111 GVA_INFO(current), COPY_to_guest | COPY_linear); in raw_copy_to_guest() 117 return copy_guest((void *)from, (vaddr_t)to, len, GVA_INFO(current), in raw_copy_to_guest_flush_dcache() 123 return copy_guest(NULL, (vaddr_t)to, len, GVA_INFO(current), in raw_clear_guest() 129 return copy_guest(to, (vaddr_t)from, len, GVA_INFO(current), in raw_copy_from_guest()
|
A D | traps.c | 645 ASSERT(v == current); in inject_vabt_exception() 1186 if ( v == current ) in vcpu_show_execution_state() 1294 domain_crash(current->domain); in do_trap_brk() 1458 current->hcall_preempted = false; in do_trap_hypercall() 1471 if ( !current->hcall_preempted ) in do_trap_hypercall() 1488 if ( current->hcall_preempted ) in do_trap_hypercall() 1854 struct domain *d = current->domain; in try_map_mmio() 2006 struct vcpu *v = current; in enter_hypervisor_from_guest_preirq() 2020 struct vcpu *v = current; in enter_hypervisor_from_guest() 2255 struct vcpu *v = current; in check_for_vcpu_work() [all …]
|
/xen/xen/arch/x86/oprofile/ |
A D | nmi_int.c | 43 struct vpmu_struct *vpmu = vcpu_vpmu(current); in passive_domain_msr_op_checks() 52 if ( ! model->allocated_msr(current) ) in passive_domain_msr_op_checks() 64 model->load_msr(current, type, index, msr_content); in passive_domain_do_rdmsr() 75 model->save_msr(current, type, index, msr_content); in passive_domain_do_wrmsr() 92 if ( ovf && is_active(current->domain) && !xen_mode ) in nmi_callback() 93 send_guest_vcpu_virq(current, VIRQ_XENOPROF); in nmi_callback() 96 current->arch.nmi_pending = true; in nmi_callback()
|
/xen/xen/arch/arm/platforms/ |
A D | xilinx-zynqmp-eemi.c | 88 if ( !domain_has_node_access(current->domain, nodeid) ) in zynqmp_eemi() 99 if ( !domain_has_reset_access(current->domain, nodeid) ) in zynqmp_eemi() 146 if ( !is_hardware_domain(current->domain) ) in zynqmp_eemi() 166 if ( !is_hardware_domain(current->domain) ) in zynqmp_eemi()
|