Home
last modified time | relevance | path

Searched refs:current (Results 1 – 25 of 244) sorted by relevance

12345678910

/xen/xen/include/xsm/
A Ddummy.h110 return xsm_default_action(action, current->domain, d); in xsm_domain_create()
116 return xsm_default_action(action, current->domain, d); in xsm_getdomaininfo()
122 return xsm_default_action(action, current->domain, d); in xsm_domctl_scheduler_op()
216 return xsm_default_action(action, current->domain, d); in xsm_memory_exchange()
272 return xsm_default_action(action, current->domain, d); in xsm_claim_pages()
279 return xsm_default_action(action, current->domain, d); in xsm_evtchn_unbound()
303 return xsm_default_action(action, current->domain, d); in xsm_evtchn_status()
330 return xsm_default_action(action, current->domain, d); in xsm_init_hardware_domain()
336 return xsm_default_action(action, current->domain, d); in xsm_get_pod_target()
342 return xsm_default_action(action, current->domain, d); in xsm_set_pod_target()
[all …]
/xen/xen/include/asm-arm/
A Devent.h27 if ( vgic_vcpu_pending_irq(current) ) in local_events_need_delivery_nomask()
30 if ( !vcpu_info(current, evtchn_upcall_pending) ) in local_events_need_delivery_nomask()
33 return vgic_evtchn_irq_pending(current); in local_events_need_delivery_nomask()
38 if ( !vcpu_event_delivery_is_enabled(current) ) in local_events_need_delivery()
/xen/xen/arch/x86/
A Dx86_emulate.c46 #define FXSAVE_AREA current->arch.fpu_ctxt
62 *val = current->arch.xcr0; in x86emul_read_xcr()
66 if ( current->domain->arch.cpuid->xstate.xgetbv1 ) in x86emul_read_xcr()
106 struct vcpu *curr = current; in x86emul_read_dr()
149 struct vcpu *curr = current; in x86emul_write_dr()
173 guest_cpuid(current, leaf, subleaf, res); in x86emul_cpuid()
A Dtrace.c11 if ( is_pv_32bit_vcpu(current) ) in __trace_pv_trap()
52 if ( is_pv_32bit_vcpu(current) ) in __trace_pv_page_fault()
83 if ( is_pv_32bit_vcpu(current) ) in __trace_trap_one_addr()
98 if ( is_pv_32bit_vcpu(current) ) in __trace_trap_two_addr()
131 if ( is_pv_32bit_vcpu(current) ) in __trace_ptwr_emulation()
A Dcompat.c46 current->domain->domain_id, current->vcpu_id, arg); in do_sched_op_compat()
47 domain_shutdown(current->domain, (u8)arg); in do_sched_op_compat()
/xen/xen/include/asm-x86/
A Dguest_access.h17 (is_hvm_vcpu(current) ? \
21 (is_hvm_vcpu(current) ? \
25 (is_hvm_vcpu(current) ? \
29 (is_hvm_vcpu(current) ? \
33 (is_hvm_vcpu(current) ? \
37 (is_hvm_vcpu(current) ? \
121 (paging_mode_external(current->domain) || \
124 (paging_mode_external(current->domain) || \
A Devent.h25 struct vcpu *v = current; in local_events_need_delivery()
36 vcpu_info(current, evtchn_upcall_mask) = 1; in local_event_delivery_disable()
41 vcpu_info(current, evtchn_upcall_mask) = 0; in local_event_delivery_enable()
/xen/xen/arch/x86/mm/
A Daltp2m.c26 if ( v != current ) in altp2m_vcpu_initialise()
34 if ( v != current ) in altp2m_vcpu_initialise()
43 if ( v != current ) in altp2m_vcpu_destroy()
54 if ( v != current ) in altp2m_vcpu_destroy()
/xen/xen/arch/x86/hvm/
A Dmonitor.c38 struct vcpu *curr = current; in hvm_monitor_cr()
65 struct vcpu *curr = current; in hvm_monitor_emul_unimplemented()
82 struct vcpu *curr = current; in hvm_monitor_msr()
118 monitor_traps(current, true, &req); in hvm_monitor_descriptor_access()
123 struct vcpu *curr = current; in gfn_of_rip()
144 struct vcpu *curr = current; in hvm_monitor_debug()
198 struct vcpu *curr = current; in hvm_monitor_cpuid()
224 monitor_traps(current, 1, &req); in hvm_monitor_interrupt()
236 struct vcpu *curr = current; in hvm_monitor_check_p2m()
A Dintercept.c41 if ( !handler->mmio.ops->check(current, first) ) in hvm_mmio_accept()
47 !handler->mmio.ops->check(current, last) ) in hvm_mmio_accept()
48 domain_crash(current->domain); in hvm_mmio_accept()
58 return handler->mmio.ops->read(current, addr, size, data); in hvm_mmio_read()
66 return handler->mmio.ops->write(current, addr, size, data); in hvm_mmio_write()
139 &data, p->size, current) ) in hvm_process_io_intercept()
153 domain_crash(current->domain); in hvm_process_io_intercept()
167 struct vcpu *curr = current; in hvm_process_io_intercept()
224 struct domain *curr_d = current->domain; in hvm_find_io_handler()
/xen/xen/xsm/flask/
A Dflask_op.c107 rv = domain_has_security(current->domain, SECURITY__COMPUTE_CREATE); in flask_security_create()
121 rv = domain_has_security(current->domain, SECURITY__COMPUTE_AV); in flask_security_access()
158 rv = domain_has_security(current->domain, SECURITY__SETENFORCE); in flask_security_setenforce()
177 rv = domain_has_security(current->domain, SECURITY__CHECK_CONTEXT); in flask_security_context()
201 rv = domain_has_security(current->domain, SECURITY__CHECK_CONTEXT); in flask_security_sid()
293 rv = domain_has_security(current->domain, SECURITY__SETBOOL); in flask_security_set_bool()
398 rv = domain_has_security(current->domain, SECURITY__SETBOOL); in flask_security_commit_bools()
506 rv = domain_has_security(current->domain, perm); in flask_devicetree_label()
553 struct domain *d = current->domain; in flask_get_peer_sid()
584 struct domain_security_struct *csec = current->domain->ssid; in flask_relabel_domain()
[all …]
A Dhooks.c82 u32 csid = domain_sid(current->domain); in avc_current_has_perm()
88 return domain_has_perm(current->domain, d, class, perms); in current_has_perm()
486 return domain_has_xen(current->domain, XEN__KEXEC); in flask_kexec()
506 if ( is_idle_domain(current->domain) && !dom0_created ) in flask_domain_create()
783 return domain_has_xen(current->domain, XEN__HEAP); in flask_sysctl()
838 return domain_has_xen(current->domain, perms); in flask_readconsole()
1045 data.ssid = domain_sid(current->domain); in flask_iomem_permission()
1238 if ( d == current->domain ) in flask_hvm_altp2mhvm_op()
1477 return domain_has_xen(current->domain, XEN__MCA_OP); in flask_do_mca()
1548 data.ssid = domain_sid(current->domain); in flask_ioport_permission()
[all …]
/xen/xen/arch/x86/pv/
A Dmm.c37 ASSERT(!paging_mode_translate(current->domain)); in map_guest_l1e()
38 ASSERT(!paging_mode_external(current->domain)); in map_guest_l1e()
64 struct vcpu *curr = current; in guest_get_eff_kern_l1e()
86 struct vcpu *curr = current; in pv_map_ldt_shadow_page()
A Ddescriptor-tables.c40 ASSERT(v == current || !vcpu_cpu_dirty(v)); in pv_destroy_ldt()
68 ASSERT(v == current || !vcpu_cpu_dirty(v)); in pv_destroy_gdt()
91 ASSERT(v == current || !vcpu_cpu_dirty(v)); in pv_set_gdt()
132 struct vcpu *curr = current; in do_set_gdt()
155 struct vcpu *curr = current; in compat_set_gdt()
190 struct domain *currd = current->domain; in do_update_descriptor()
A Demul-inv-op.c31 const struct vcpu_msrs *msrs = current->arch.msrs; in emulate_forced_invalid_op()
56 !guest_kernel_mode(current, regs) ) in emulate_forced_invalid_op()
65 guest_cpuid(current, regs->eax, regs->ecx, &res); in emulate_forced_invalid_op()
/xen/xen/common/
A Dwait.c125 struct vcpu *curr = current; in __prepare_to_wait()
181 vcpu_temporary_affinity(current, NR_CPUS, VCPU_AFFINITY_WAIT); in __finish_wait()
186 struct vcpu *curr = current; in check_wakeup_from_wait()
229 struct vcpu *curr = current; in prepare_to_wait()
245 struct vcpu *curr = current; in finish_wait()
A Dxenoprof.c218 BUG_ON(page_get_owner(page) != current->domain); in unshare_xenoprof_page_with_guest()
501 if ( !paging_mode_translate(current->domain) ) in add_passive_list()
505 current->domain, __pa(d->xenoprof->rawbuf), in add_passive_list()
648 struct domain *d = current->domain; in xenoprof_op_init()
669 xenoprof_primary_profiler = current->domain; in xenoprof_op_init()
681 struct domain *d = current->domain; in xenoprof_op_get_buffer()
740 ret = xsm_profile(XSM_HOOK, current->domain, op); in do_xenoprof_op()
848 xenoprof_reset_buf(current->domain); in do_xenoprof_op()
849 ret = set_active(current->domain); in do_xenoprof_op()
892 (is_active(current->domain)) ) in do_xenoprof_op()
[all …]
A Ddomctl.c226 spin_unlock(&current->domain->hypercall_deadlock_mutex); in domctl_lock_acquire()
233 spin_unlock(&current->domain->hypercall_deadlock_mutex); in domctl_lock_release()
424 if ( (d == current->domain) || /* no domain_pause() */ in do_domctl()
474 if ( d != current->domain ) in do_domctl()
483 if ( d == current->domain ) /* no domain_pause() */ in do_domctl()
538 if ( (d == current->domain) || /* no domain_pause() */ in do_domctl()
566 if ( d == current->domain ) /* no domain_pause() */ in do_domctl()
668 v == current ) /* no vcpu_pause() */ in do_domctl()
770 if ( pirq >= current->domain->nr_pirqs ) in do_domctl()
775 irq = pirq_access_permitted(current->domain, pirq); in do_domctl()
[all …]
/xen/xen/arch/arm/
A Dtime.c258 if ( unlikely(is_idle_vcpu(current)) ) in vtimer_interrupt()
263 current->arch.virt_timer.ctl = READ_SYSREG32(CNTV_CTL_EL0); in vtimer_interrupt()
264 WRITE_SYSREG32(current->arch.virt_timer.ctl | CNTx_CTL_MASK, CNTV_CTL_EL0); in vtimer_interrupt()
265 vgic_inject_irq(current->domain, current, current->arch.virt_timer.irq, true); in vtimer_interrupt()
A Ddomain.c312 ASSERT(prev != current); in schedule_tail()
316 ctxt_switch_to(current); in schedule_tail()
322 update_runstate_area(current); in schedule_tail()
325 update_vcpu_system_time(current); in schedule_tail()
331 processor_vcpu_initialise(current); in continue_new_vcpu()
335 if ( is_idle_vcpu(current) ) in continue_new_vcpu()
413 current->hcall_preempted = true; in hypercall_create_continuation()
485 domain_crash(current->domain); in hypercall_create_continuation()
493 struct vcpu *v = current; in startup_cpu_idle_loop()
1099 vcpu_unblock(current); in vcpu_block_unless_event_pending()
[all …]
A Dvpsci.c28 struct domain *d = current->domain; in do_common_cpu_on()
115 struct vcpu *v = current; in do_psci_cpu_off()
134 struct vcpu *v = current; in do_psci_0_2_cpu_suspend()
169 struct domain *d = current->domain; in do_psci_0_2_affinity_info()
201 struct domain *d = current->domain; in do_psci_0_2_system_off()
207 struct domain *d = current->domain; in do_psci_0_2_system_reset()
A Dguestcopy.c111 GVA_INFO(current), COPY_to_guest | COPY_linear); in raw_copy_to_guest()
117 return copy_guest((void *)from, (vaddr_t)to, len, GVA_INFO(current), in raw_copy_to_guest_flush_dcache()
123 return copy_guest(NULL, (vaddr_t)to, len, GVA_INFO(current), in raw_clear_guest()
129 return copy_guest(to, (vaddr_t)from, len, GVA_INFO(current), in raw_copy_from_guest()
A Dtraps.c645 ASSERT(v == current); in inject_vabt_exception()
1186 if ( v == current ) in vcpu_show_execution_state()
1294 domain_crash(current->domain); in do_trap_brk()
1458 current->hcall_preempted = false; in do_trap_hypercall()
1471 if ( !current->hcall_preempted ) in do_trap_hypercall()
1488 if ( current->hcall_preempted ) in do_trap_hypercall()
1854 struct domain *d = current->domain; in try_map_mmio()
2006 struct vcpu *v = current; in enter_hypervisor_from_guest_preirq()
2020 struct vcpu *v = current; in enter_hypervisor_from_guest()
2255 struct vcpu *v = current; in check_for_vcpu_work()
[all …]
/xen/xen/arch/x86/oprofile/
A Dnmi_int.c43 struct vpmu_struct *vpmu = vcpu_vpmu(current); in passive_domain_msr_op_checks()
52 if ( ! model->allocated_msr(current) ) in passive_domain_msr_op_checks()
64 model->load_msr(current, type, index, msr_content); in passive_domain_do_rdmsr()
75 model->save_msr(current, type, index, msr_content); in passive_domain_do_wrmsr()
92 if ( ovf && is_active(current->domain) && !xen_mode ) in nmi_callback()
93 send_guest_vcpu_virq(current, VIRQ_XENOPROF); in nmi_callback()
96 current->arch.nmi_pending = true; in nmi_callback()
/xen/xen/arch/arm/platforms/
A Dxilinx-zynqmp-eemi.c88 if ( !domain_has_node_access(current->domain, nodeid) ) in zynqmp_eemi()
99 if ( !domain_has_reset_access(current->domain, nodeid) ) in zynqmp_eemi()
146 if ( !is_hardware_domain(current->domain) ) in zynqmp_eemi()
166 if ( !is_hardware_domain(current->domain) ) in zynqmp_eemi()

Completed in 50 milliseconds

12345678910