Home
last modified time | relevance | path

Searched refs:d (Results 1 – 25 of 507) sorted by relevance

12345678910>>...21

/xen/tools/xentrace/
A Dformats3 0x0001f001 CPU%(cpu)d %(tsc)d (+%(reltsc)8d) lost_records 0x%(1)08x
29 0x00028008 CPU%(cpu)d %(tsc)d (+%(reltsc)8d) sched_ctl
45 0x00022006 CPU%(cpu)d %(tsc)d (+%(reltsc)8d) csched:tickle [ cpu = %(1)d ]
50 0x0002200B CPU%(cpu)d %(tsc)d (+%(reltsc)8d) csched:steal_check [ peer_cpu = %(1)d, checked = …
117 0x00082010 CPU%(cpu)d %(tsc)d (+%(reltsc)8d) NMI
118 0x00082011 CPU%(cpu)d %(tsc)d (+%(reltsc)8d) SMI
123 0x00082015 CPU%(cpu)d %(tsc)d (+%(reltsc)8d) MCE
198 0x00801003 CPU%(cpu)d %(tsc)d (+%(reltsc)8d) cpu_idle_exit [ C%(1)d -> C0, acpi_pm_tick = %(2)
213 0x00084005 CPU%(cpu)d %(tsc)d (+%(reltsc)8d) hpet destroy [ tn = %(1)d ]
219 0x0008400b CPU%(cpu)d %(tsc)d (+%(reltsc)8d) vpic_update_int_output [ int_output = %(1)d, is_mas…
[all …]
/xen/xen/common/
A Ddomain.c100 evtchn_send(d, d->suspend_evtchn); in __domain_finalise_shutdown()
149 if ( vcpu_id >= d->max_vcpus || d->vcpu[vcpu_id] || in vcpu_create()
220 if ( d != hardware_domain || d->domain_id == 0 ) in late_hwdom_init()
437 if ( is_system_domain(d) && !is_idle_domain(d) ) in domain_create()
447 d->nr_pirqs = min(d->nr_pirqs, nr_irqs); in domain_create()
466 if ( !d->iomem_caps || !d->irq_caps ) in domain_create()
518 return d; in domain_create()
626 d = rcu_dereference(d->next_in_hashbucket) ) in get_domain_by_id()
638 return d; in get_domain_by_id()
650 d = rcu_dereference(d->next_in_hashbucket) ) in rcu_lock_domain_by_id()
[all …]
A Devent_2l.c30 if ( guest_test_and_set_bit(d, port, &shared_info(d, evtchn_pending)) ) in evtchn_2l_set_pending()
33 if ( !guest_test_bit(d, port, &shared_info(d, evtchn_mask)) && in evtchn_2l_set_pending()
34 !guest_test_and_set_bit(d, port / BITS_PER_EVTCHN_WORD(d), in evtchn_2l_set_pending()
45 guest_clear_bit(d, evtchn->port, &shared_info(d, evtchn_pending)); in evtchn_2l_clear_pending()
57 if ( guest_test_and_clear_bit(d, port, &shared_info(d, evtchn_mask)) && in evtchn_2l_unmask()
58 guest_test_bit(d, port, &shared_info(d, evtchn_pending)) && in evtchn_2l_unmask()
59 !guest_test_and_set_bit(d, port / BITS_PER_EVTCHN_WORD(d), in evtchn_2l_unmask()
70 unsigned int max_ports = BITS_PER_EVTCHN_WORD(d) * BITS_PER_EVTCHN_WORD(d); in evtchn_2l_is_pending()
74 guest_test_bit(d, port, &shared_info(d, evtchn_pending))); in evtchn_2l_is_pending()
81 unsigned int max_ports = BITS_PER_EVTCHN_WORD(d) * BITS_PER_EVTCHN_WORD(d); in evtchn_2l_is_masked()
[all …]
A Devent_fifo.c507 if ( !d->evtchn_fifo ) in setup_event_array()
517 if ( !d->evtchn_fifo ) in cleanup_event_array()
522 xfree(d->evtchn_fifo); in cleanup_event_array()
523 d->evtchn_fifo = NULL; in cleanup_event_array()
545 if ( guest_test_bit(d, port, &shared_info(d, evtchn_pending)) ) in setup_ports()
585 spin_lock(&d->event_lock); in evtchn_fifo_init_control()
591 if ( !d->evtchn_fifo ) in evtchn_fifo_init_control()
626 evtchn_fifo_destroy(d); in evtchn_fifo_init_control()
677 if ( !d->evtchn_fifo ) in evtchn_fifo_expand_array()
691 for_each_vcpu( d, v ) in evtchn_fifo_destroy()
[all …]
A Dvm_event.c40 struct domain *d, in vm_event_enable() argument
300 if( curr->domain == d && avail_req < d->max_vcpus && in vm_event_put_request()
557 (void)vm_event_disable(d, &d->vm_event_paging); in vm_event_cleanup()
563 (void)vm_event_disable(d, &d->vm_event_monitor); in vm_event_cleanup()
569 (void)vm_event_disable(d, &d->vm_event_share); in vm_event_cleanup()
653 rc = vm_event_disable(d, &d->vm_event_paging); in vm_event_domctl()
659 rc = vm_event_resume(d, d->vm_event_paging); in vm_event_domctl()
690 rc = vm_event_disable(d, &d->vm_event_monitor); in vm_event_domctl()
697 rc = vm_event_resume(d, d->vm_event_monitor); in vm_event_domctl()
735 rc = vm_event_disable(d, &d->vm_event_share); in vm_event_domctl()
[all …]
A Devent_channel.c188 if ( port > d->max_evtchn_port || port >= max_evtchns(d) ) in evtchn_allocate_port()
226 write_atomic(&d->active_evtchns, d->active_evtchns + 1); in evtchn_allocate_port()
268 write_atomic(&d->xen_evtchns, d->xen_evtchns - 1); in evtchn_free()
269 write_atomic(&d->active_evtchns, d->active_evtchns - 1); in evtchn_free()
1103 i = resuming ? d->next_evtchn : !d->next_evtchn; in evtchn_reset()
1130 if ( d->active_evtchns > d->xen_evtchns ) in evtchn_reset()
1437 d->evtchn = alloc_evtchn_bucket(d, 0); in evtchn_init()
1445 free_evtchn_bucket(d, d->evtchn); in evtchn_init()
1455 free_evtchn_bucket(d, d->evtchn); in evtchn_init()
1511 free_evtchn_bucket(d, d->evtchn); in evtchn_destroy_final()
[all …]
A Dxenoprof.c47 #define xenoprof_buf(d, b, field) ACCESS_ONCE(*(!(d)->xenoprof->is_compat \ argument
145 return (is_active(d) || is_passive(d)); in is_profiled()
333 x = d->xenoprof; in free_xenoprof_pages()
368 x = d->xenoprof; in set_active()
388 x = d->xenoprof; in reset_active()
396 put_domain(d); in reset_active()
408 if ( d == NULL ) in reset_passive()
411 x = d->xenoprof; in reset_passive()
453 if ( d == NULL ) in add_active_list()
578 return xenoprof_add_sample(d, &d->xenoprof->vcpu[vcpu->vcpu_id], in xenoprof_add_trace()
[all …]
A Ddomctl.c138 struct domain *d; in is_free_domid() local
199 info->shared_info_frame = mfn_to_gmfn(d, virt_to_mfn(d->shared_info)); in getdomaininfo()
366 struct domain *d; in do_domctl() local
392 d = NULL; in do_domctl()
408 if ( d && d != dom_io ) in do_domctl()
425 (vcpu >= d->max_vcpus) || ((v = d->vcpu[vcpu]) == NULL) ) in do_domctl()
529 d = NULL; in do_domctl()
623 if ( !d ) in do_domctl()
656 d = NULL; in do_domctl()
881 if ( (d == e) || (d->target != NULL) ) in do_domctl()
[all …]
/xen/xen/arch/x86/
A Dtrace.c18 } d; in __trace_pv_trap() local
25 __trace_var(TRC_PV_TRAP, 1, sizeof(d), &d); in __trace_pv_trap()
34 } d; in __trace_pv_trap() local
44 __trace_var(event, 1, sizeof(d), &d); in __trace_pv_trap()
56 } d; in __trace_pv_page_fault() local
69 } d; in __trace_pv_page_fault() local
77 __trace_var(event, 1, sizeof(d), &d); in __trace_pv_page_fault()
86 __trace_var(event, 1, sizeof(d), &d); in __trace_trap_one_addr()
105 __trace_var(event, 1, sizeof(d), &d); in __trace_trap_two_addr()
115 __trace_var(event, 1, sizeof(d), &d); in __trace_trap_two_addr()
[all …]
A Dmonitor.c41 memset(&d->arch.monitor, 0, sizeof(d->arch.monitor)); in arch_monitor_cleanup_domain()
42 memset(&d->monitor, 0, sizeof(d->monitor)); in arch_monitor_cleanup_domain()
169 domain_pause(d); in arch_monitor_domctl_event()
212 domain_pause(d); in arch_monitor_domctl_event()
239 domain_pause(d); in arch_monitor_domctl_event()
252 domain_pause(d); in arch_monitor_domctl_event()
269 domain_pause(d); in arch_monitor_domctl_event()
286 domain_pause(d); in arch_monitor_domctl_event()
299 domain_pause(d); in arch_monitor_domctl_event()
315 domain_pause(d); in arch_monitor_domctl_event()
[all …]
/xen/xen/arch/x86/mm/hap/
A Dhap.c121 guest_flush_tlb_mask(d, d->dirty_cpumask); in hap_track_dirty_vram()
194 paging_lock(d); in hap_enable_log_dirty()
196 paging_unlock(d); in hap_enable_log_dirty()
208 guest_flush_tlb_mask(d, d->dirty_cpumask); in hap_enable_log_dirty()
215 paging_lock(d); in hap_disable_log_dirty()
237 guest_flush_tlb_mask(d, d->dirty_cpumask); in hap_clean_dirty_bitmap()
411 domain_crash(d); in hap_make_monitor_table()
559 paging_lock(d); in hap_final_teardown()
757 paging_lock(d); in hap_update_paging_modes()
790 paging_lock(d); in hap_write_p2m_entry()
[all …]
/xen/xen/arch/x86/mm/
A Dpaging.c58 page = d->arch.paging.alloc_page(d); in paging_new_log_dirty_page()
107 d->arch.paging.free_page(d, mfn_to_page(mfn)); in paging_free_log_dirty_page()
115 paging_lock(d); in paging_free_log_dirty_bitmap()
192 paging_free_log_dirty_page(d, d->arch.paging.log_dirty.top); in paging_free_log_dirty_bitmap()
229 ret = d->arch.paging.log_dirty.ops->enable(d, log_global); in paging_log_dirty_enable()
245 ret = d->arch.paging.log_dirty.ops->disable(d); in paging_log_dirty_disable()
291 d->arch.paging.log_dirty.top = paging_new_log_dirty_node(d); in paging_mark_pfn_dirty()
431 paging_lock(d); in paging_log_dirty_op()
567 d->arch.paging.log_dirty.ops->clean(d); in paging_log_dirty_op()
616 guest_flush_tlb_mask(d, d->dirty_cpumask); in paging_log_dirty_range()
[all …]
A Dmm-locks.h68 return l + (d && is_control_domain(d) ? MM_LOCK_ORDER_MAX : 0); in _lock_level()
194 #define mm_lock(name, d, l) mm_lock_##name(d, l, __func__, 0) argument
195 #define mm_lock_recursive(name, d, l) mm_lock_##name(d, l, __func__, 1) argument
196 #define mm_write_lock(name, d, l) mm_write_lock_##name(d, l, __func__) argument
197 #define mm_read_lock(name, d, l) mm_read_lock_##name(d, l) argument
252 #define nestedp2m_lock(d) mm_lock(nestedp2m, d, &(d)->arch.nested_p2m_lock) argument
253 #define nestedp2m_unlock(d) mm_unlock(&(d)->arch.nested_p2m_lock) argument
307 #define altp2m_list_lock(d) mm_lock(altp2mlist, d, \ argument
309 #define altp2m_list_unlock(d) mm_unlock(&(d)->arch.altp2m_list_lock) argument
390 #define paging_lock(d) mm_lock(paging, d, &(d)->arch.paging.lock) argument
[all …]
/xen/xen/include/xen/
A Devent.h103 #define group_from_port(d, p) \ argument
111 : BITS_PER_EVTCHN_WORD(d) * BITS_PER_EVTCHN_WORD(d); in max_evtchns()
227 d->evtchn_port_ops->init(d, evtchn); in evtchn_port_init()
235 d->evtchn_port_ops->set_pending(d->vcpu[vcpu_id], evtchn); in evtchn_port_set_pending()
242 d->evtchn_port_ops->clear_pending(d, evtchn); in evtchn_port_clear_pending()
249 d->evtchn_port_ops->unmask(d, evtchn); in evtchn_port_unmask()
255 return evtchn_usable(evtchn) && d->evtchn_port_ops->is_pending(d, evtchn); in evtchn_is_pending()
273 return !evtchn_usable(evtchn) || d->evtchn_port_ops->is_masked(d, evtchn); in evtchn_is_masked()
294 d->evtchn_port_ops->is_busy(d, evtchn); in evtchn_is_busy()
305 return d->evtchn_port_ops->set_priority(d, evtchn, priority); in evtchn_port_set_priority()
[all …]
A Dsched.h545 ASSERT(d->extra_pages <= d->tot_pages); in domain_tot_pages()
547 return d->tot_pages - d->extra_pages; in domain_tot_pages()
555 #define is_idle_domain(d) ((d)->domain_id == DOMID_IDLE) argument
648 return d; in rcu_lock_domain()
787 for (d = rcu_dereference(domain_list); d && d->cpupool != c; in first_domain_in_cpupool()
788 d = rcu_dereference(d->next_in_list)); in first_domain_in_cpupool()
789 return d; in first_domain_in_cpupool()
794 for (d = rcu_dereference(d->next_in_list); d && d->cpupool != c; in next_domain_in_cpupool()
795 d = rcu_dereference(d->next_in_list)); in next_domain_in_cpupool()
1002 return is_pv_domain(d) && d->arch.pv.is_32bit; in is_pv_32bit_domain()
[all …]
A Diommu.h113 void iommu_hwdom_init(struct domain *d);
237 int (*init)(struct domain *d);
238 void (*hwdom_init)(struct domain *d);
252 void (*teardown)(struct domain *d);
283 void (*share_p2m)(struct domain *d);
341 #define dom_iommu(d) (&(d)->iommu) argument
342 #define iommu_set_feature(d, f) set_bit(f, dom_iommu(d)->features) argument
343 #define iommu_clear_feature(d, f) clear_bit(f, dom_iommu(d)->features) argument
346 #define iommu_use_hap_pt(d) (dom_iommu(d)->hap_pt_share) argument
350 #define need_iommu_pt_sync(d) (dom_iommu(d)->need_sync) argument
[all …]
/xen/xen/include/xsm/
A Dxsm.h180 int (*dm_op) (struct domain *d);
232 return xsm_ops->domctl(d, cmd); in xsm_domctl()
388 return xsm_ops->claim_pages(d); in xsm_claim_pages()
398 return xsm_ops->profile(d, op); in xsm_profile()
595 return xsm_ops->mem_access(d); in xsm_mem_access()
602 return xsm_ops->mem_paging(d); in xsm_mem_paging()
609 return xsm_ops->mem_sharing(d); in xsm_mem_sharing()
636 return xsm_ops->apic(d, cmd); in xsm_apic()
688 return xsm_ops->pmu_op(d, op); in xsm_pmu_op()
693 return xsm_ops->dm_op(d); in xsm_dm_op()
[all …]
A Ddummy.h235 if ( d->is_console ) in xsm_console_io()
241 return xsm_default_action(XSM_PRIV, d, NULL); in xsm_console_io()
247 return xsm_default_action(action, d, NULL); in xsm_profile()
297 return xsm_default_action(action, d, NULL); in xsm_evtchn_send()
539 return xsm_default_action(action, d, t); in xsm_map_gmfn_foreign()
643 return xsm_default_action(action, d, NULL); in xsm_apic()
664 rc = xsm_default_action(action, d, f); in xsm_mmu_update()
666 rc = xsm_default_action(action, d, t); in xsm_mmu_update()
673 return xsm_default_action(action, d, f); in xsm_mmuext_op()
680 return xsm_default_action(action, d, f); in xsm_update_va_mapping()
[all …]
/xen/xen/include/asm-x86/
A Dshadow.h53 int shadow_domain_init(struct domain *d);
73 int shadow_domctl(struct domain *d,
101 #define shadow_teardown(d, p) ASSERT(is_pv_domain(d)) argument
102 #define shadow_final_teardown(d) ASSERT(is_pv_domain(d)) argument
103 #define shadow_enable(d, mode) \ argument
180 ASSERT(is_pv_domain(d)); in pv_l1tf_check_pte()
183 if ( d->arch.pv.check_l1tf && !paging_mode_sh_forced(d) && in pv_l1tf_check_pte()
191 d->domain_id, level, pte); in pv_l1tf_check_pte()
206 domain_crash(d); in pv_l1tf_check_pte()
238 d->arch.pv.check_l1tf = is_hardware_domain(d) ? opt_pv_l1tf_hwdom in pv_l1tf_domain_init()
[all …]
/xen/xen/drivers/passthrough/
A Diommu.c174 if ( !is_iommu_enabled(d) ) in iommu_domain_init()
237 iommu_teardown(d); in iommu_domain_destroy()
277 domain_crash(d); in iommu_map()
333 domain_crash(d); in iommu_unmap()
409 domain_crash(d); in iommu_iotlb_flush()
437 domain_crash(d); in iommu_iotlb_flush_all()
552 ASSERT(hap_enabled(d)); in iommu_share_p2m_table()
590 return is_iommu_enabled(d) && test_bit(feature, dom_iommu(d)->features); in iommu_has_feature()
595 struct domain *d; in iommu_dump_p2m_table() local
608 for_each_domain(d) in iommu_dump_p2m_table()
[all …]
/xen/xen/arch/x86/mm/shadow/
A Dcommon.c395 guest_flush_tlb_mask(d, d->dirty_cpumask); in oos_remove_write_access()
940 ASSERT(d->vcpu && d->vcpu[0]); in _shadow_prealloc()
973 guest_flush_tlb_mask(d, d->dirty_cpumask); in _shadow_prealloc()
1010 ASSERT(d->vcpu && d->vcpu[0]); in shadow_blow_tables()
1027 guest_flush_tlb_mask(d, d->dirty_cpumask); in shadow_blow_tables()
1032 if ( shadow_mode_enabled(d) && domain_vcpu(d, 0) ) in shadow_blow_tables_per_domain()
2295 guest_flush_tlb_mask(d, d->dirty_cpumask); in sh_remove_shadows()
2890 sh_new_mode(d, d->arch.paging.mode | mode); in shadow_one_bit_enable()
3012 guest_flush_tlb_mask(d, d->dirty_cpumask); in sh_unshadow_for_p2m_change()
3337 guest_flush_tlb_mask(d, d->dirty_cpumask); in shadow_track_dirty_vram()
[all …]
/xen/xen/arch/x86/hvm/
A Dirq.c40 pirq_info(d, domain_emuirq_to_pirq(d, ioapic_gsi)); in assert_gsi()
63 assert_gsi(d, gsi); in hvm_ioapic_assert()
87 assert_gsi(d, ioapic_gsi); in assert_irq()
95 pirq_info(d, domain_emuirq_to_pirq(d, isa_irq)); in deassert_irq()
119 assert_gsi(d, gsi); in __hvm_pci_intx_assert()
194 assert_gsi(d, gsi); in hvm_gsi_assert()
440 (!has_vlapic(d) || !has_vioapic(d) || !has_vpic(d)) ) in hvm_set_callback_via()
495 for_each_vcpu ( d, v ) in hvm_set_callback_via()
640 struct domain *d; in dump_irq_info() local
646 for_each_domain ( d ) in dump_irq_info()
[all …]
A Dsave.c73 if ( d->arch.vtsc ) in arch_hvm_load()
143 if ( d->is_dying || in hvm_save_one()
150 v = d->vcpu[0]; in hvm_save_one()
151 else if ( instance >= d->max_vcpus || !d->vcpu[instance] ) in hvm_save_one()
213 if ( d->is_dying ) in hvm_save()
363 d->typecode = tc; in _hvm_init_entry()
365 d->length = len; in _hvm_init_entry()
393 (strict_length ? (len != d->length) : (len < d->length)) || in _hvm_check_entry()
394 (d->length > (h->size - h->cur - sizeof(*d))) ) in _hvm_check_entry()
400 d->typecode, d->length, h->size - h->cur - sizeof(*d)); in _hvm_check_entry()
[all …]
/xen/xen/include/acpi/
A Dacmacros.h155 #define ACPI_MOVE_16_TO_32(d,s) {(*(u32 *)(void *)(d))=0;\ argument
159 #define ACPI_MOVE_16_TO_64(d,s) {(*(u64 *)(void *)(d))=0;\ argument
172 #define ACPI_MOVE_32_TO_64(d,s) {(*(u64 *)(void *)(d))=0;\ argument
235 #define ACPI_MOVE_16_TO_32(d,s) {(*(u32 *)(void *)(d)) = 0; ACPI_MOVE_16_TO_16(d,s);} argument
236 #define ACPI_MOVE_16_TO_64(d,s) {(*(u64 *)(void *)(d)) = 0; ACPI_MOVE_16_TO_16(d,s);} argument
247 #define ACPI_MOVE_32_TO_64(d,s) {(*(u64 *)(void *)(d)) = 0; ACPI_MOVE_32_TO_32(d,s);} argument
267 #define ACPI_MOVE_SIZE_TO_16(d,s) ACPI_MOVE_32_TO_16(d,s) argument
270 #define ACPI_MOVE_SIZE_TO_16(d,s) ACPI_MOVE_64_TO_16(d,s) argument
343 #define ACPI_IS_OCTAL_DIGIT(d) (((char)(d) >= '0') && ((char)(d) <= '7')) argument
404 #define ARGI_LIST4(a,b,c,d) (ARG_1(d)|ARG_2(c)|ARG_3(b)|ARG_4(a)) argument
[all …]
/xen/stubdom/
A Dpciutils.patch105 + d->domain = domain;
106 + d->bus = bus;
107 + d->dev = slot;
108 + d->func = fun;
122 + if (pcifront_conf_read(NULL, d->domain, d->bus, d->dev, d->func, pos, len, &val))
127 + if (pcifront_conf_read(NULL, d->domain, d->bus, d->dev, d->func, pos, len, &val))
132 + if (pcifront_conf_read(NULL, d->domain, d->bus, d->dev, d->func, pos, len, &val))
158 + return !pcifront_conf_write(NULL, d->domain, d->bus, d->dev, d->func, pos, len, val);
228 …a->warning("%04x:%02x:%02x.%d: Invalid 64-bit address seen for BAR %d.", d->domain, d->bus, d->dev…
247 …a->warning("%04x:%02x:%02x.%d 64-bit device address ignored.", d->domain, d->bus, d->dev, d->func);
[all …]

Completed in 103 milliseconds

12345678910>>...21