/xen/tools/xentrace/ |
A D | formats | 3 0x0001f001 CPU%(cpu)d %(tsc)d (+%(reltsc)8d) lost_records 0x%(1)08x 29 0x00028008 CPU%(cpu)d %(tsc)d (+%(reltsc)8d) sched_ctl 45 0x00022006 CPU%(cpu)d %(tsc)d (+%(reltsc)8d) csched:tickle [ cpu = %(1)d ] 50 0x0002200B CPU%(cpu)d %(tsc)d (+%(reltsc)8d) csched:steal_check [ peer_cpu = %(1)d, checked = … 117 0x00082010 CPU%(cpu)d %(tsc)d (+%(reltsc)8d) NMI 118 0x00082011 CPU%(cpu)d %(tsc)d (+%(reltsc)8d) SMI 123 0x00082015 CPU%(cpu)d %(tsc)d (+%(reltsc)8d) MCE 198 0x00801003 CPU%(cpu)d %(tsc)d (+%(reltsc)8d) cpu_idle_exit [ C%(1)d -> C0, acpi_pm_tick = %(2)… 213 0x00084005 CPU%(cpu)d %(tsc)d (+%(reltsc)8d) hpet destroy [ tn = %(1)d ] 219 0x0008400b CPU%(cpu)d %(tsc)d (+%(reltsc)8d) vpic_update_int_output [ int_output = %(1)d, is_mas… [all …]
|
/xen/xen/common/ |
A D | domain.c | 100 evtchn_send(d, d->suspend_evtchn); in __domain_finalise_shutdown() 149 if ( vcpu_id >= d->max_vcpus || d->vcpu[vcpu_id] || in vcpu_create() 220 if ( d != hardware_domain || d->domain_id == 0 ) in late_hwdom_init() 437 if ( is_system_domain(d) && !is_idle_domain(d) ) in domain_create() 447 d->nr_pirqs = min(d->nr_pirqs, nr_irqs); in domain_create() 466 if ( !d->iomem_caps || !d->irq_caps ) in domain_create() 518 return d; in domain_create() 626 d = rcu_dereference(d->next_in_hashbucket) ) in get_domain_by_id() 638 return d; in get_domain_by_id() 650 d = rcu_dereference(d->next_in_hashbucket) ) in rcu_lock_domain_by_id() [all …]
|
A D | event_2l.c | 30 if ( guest_test_and_set_bit(d, port, &shared_info(d, evtchn_pending)) ) in evtchn_2l_set_pending() 33 if ( !guest_test_bit(d, port, &shared_info(d, evtchn_mask)) && in evtchn_2l_set_pending() 34 !guest_test_and_set_bit(d, port / BITS_PER_EVTCHN_WORD(d), in evtchn_2l_set_pending() 45 guest_clear_bit(d, evtchn->port, &shared_info(d, evtchn_pending)); in evtchn_2l_clear_pending() 57 if ( guest_test_and_clear_bit(d, port, &shared_info(d, evtchn_mask)) && in evtchn_2l_unmask() 58 guest_test_bit(d, port, &shared_info(d, evtchn_pending)) && in evtchn_2l_unmask() 59 !guest_test_and_set_bit(d, port / BITS_PER_EVTCHN_WORD(d), in evtchn_2l_unmask() 70 unsigned int max_ports = BITS_PER_EVTCHN_WORD(d) * BITS_PER_EVTCHN_WORD(d); in evtchn_2l_is_pending() 74 guest_test_bit(d, port, &shared_info(d, evtchn_pending))); in evtchn_2l_is_pending() 81 unsigned int max_ports = BITS_PER_EVTCHN_WORD(d) * BITS_PER_EVTCHN_WORD(d); in evtchn_2l_is_masked() [all …]
|
A D | event_fifo.c | 507 if ( !d->evtchn_fifo ) in setup_event_array() 517 if ( !d->evtchn_fifo ) in cleanup_event_array() 522 xfree(d->evtchn_fifo); in cleanup_event_array() 523 d->evtchn_fifo = NULL; in cleanup_event_array() 545 if ( guest_test_bit(d, port, &shared_info(d, evtchn_pending)) ) in setup_ports() 585 spin_lock(&d->event_lock); in evtchn_fifo_init_control() 591 if ( !d->evtchn_fifo ) in evtchn_fifo_init_control() 626 evtchn_fifo_destroy(d); in evtchn_fifo_init_control() 677 if ( !d->evtchn_fifo ) in evtchn_fifo_expand_array() 691 for_each_vcpu( d, v ) in evtchn_fifo_destroy() [all …]
|
A D | vm_event.c | 40 struct domain *d, in vm_event_enable() argument 300 if( curr->domain == d && avail_req < d->max_vcpus && in vm_event_put_request() 557 (void)vm_event_disable(d, &d->vm_event_paging); in vm_event_cleanup() 563 (void)vm_event_disable(d, &d->vm_event_monitor); in vm_event_cleanup() 569 (void)vm_event_disable(d, &d->vm_event_share); in vm_event_cleanup() 653 rc = vm_event_disable(d, &d->vm_event_paging); in vm_event_domctl() 659 rc = vm_event_resume(d, d->vm_event_paging); in vm_event_domctl() 690 rc = vm_event_disable(d, &d->vm_event_monitor); in vm_event_domctl() 697 rc = vm_event_resume(d, d->vm_event_monitor); in vm_event_domctl() 735 rc = vm_event_disable(d, &d->vm_event_share); in vm_event_domctl() [all …]
|
A D | event_channel.c | 188 if ( port > d->max_evtchn_port || port >= max_evtchns(d) ) in evtchn_allocate_port() 226 write_atomic(&d->active_evtchns, d->active_evtchns + 1); in evtchn_allocate_port() 268 write_atomic(&d->xen_evtchns, d->xen_evtchns - 1); in evtchn_free() 269 write_atomic(&d->active_evtchns, d->active_evtchns - 1); in evtchn_free() 1103 i = resuming ? d->next_evtchn : !d->next_evtchn; in evtchn_reset() 1130 if ( d->active_evtchns > d->xen_evtchns ) in evtchn_reset() 1437 d->evtchn = alloc_evtchn_bucket(d, 0); in evtchn_init() 1445 free_evtchn_bucket(d, d->evtchn); in evtchn_init() 1455 free_evtchn_bucket(d, d->evtchn); in evtchn_init() 1511 free_evtchn_bucket(d, d->evtchn); in evtchn_destroy_final() [all …]
|
A D | xenoprof.c | 47 #define xenoprof_buf(d, b, field) ACCESS_ONCE(*(!(d)->xenoprof->is_compat \ argument 145 return (is_active(d) || is_passive(d)); in is_profiled() 333 x = d->xenoprof; in free_xenoprof_pages() 368 x = d->xenoprof; in set_active() 388 x = d->xenoprof; in reset_active() 396 put_domain(d); in reset_active() 408 if ( d == NULL ) in reset_passive() 411 x = d->xenoprof; in reset_passive() 453 if ( d == NULL ) in add_active_list() 578 return xenoprof_add_sample(d, &d->xenoprof->vcpu[vcpu->vcpu_id], in xenoprof_add_trace() [all …]
|
A D | domctl.c | 138 struct domain *d; in is_free_domid() local 199 info->shared_info_frame = mfn_to_gmfn(d, virt_to_mfn(d->shared_info)); in getdomaininfo() 366 struct domain *d; in do_domctl() local 392 d = NULL; in do_domctl() 408 if ( d && d != dom_io ) in do_domctl() 425 (vcpu >= d->max_vcpus) || ((v = d->vcpu[vcpu]) == NULL) ) in do_domctl() 529 d = NULL; in do_domctl() 623 if ( !d ) in do_domctl() 656 d = NULL; in do_domctl() 881 if ( (d == e) || (d->target != NULL) ) in do_domctl() [all …]
|
/xen/xen/arch/x86/ |
A D | trace.c | 18 } d; in __trace_pv_trap() local 25 __trace_var(TRC_PV_TRAP, 1, sizeof(d), &d); in __trace_pv_trap() 34 } d; in __trace_pv_trap() local 44 __trace_var(event, 1, sizeof(d), &d); in __trace_pv_trap() 56 } d; in __trace_pv_page_fault() local 69 } d; in __trace_pv_page_fault() local 77 __trace_var(event, 1, sizeof(d), &d); in __trace_pv_page_fault() 86 __trace_var(event, 1, sizeof(d), &d); in __trace_trap_one_addr() 105 __trace_var(event, 1, sizeof(d), &d); in __trace_trap_two_addr() 115 __trace_var(event, 1, sizeof(d), &d); in __trace_trap_two_addr() [all …]
|
A D | monitor.c | 41 memset(&d->arch.monitor, 0, sizeof(d->arch.monitor)); in arch_monitor_cleanup_domain() 42 memset(&d->monitor, 0, sizeof(d->monitor)); in arch_monitor_cleanup_domain() 169 domain_pause(d); in arch_monitor_domctl_event() 212 domain_pause(d); in arch_monitor_domctl_event() 239 domain_pause(d); in arch_monitor_domctl_event() 252 domain_pause(d); in arch_monitor_domctl_event() 269 domain_pause(d); in arch_monitor_domctl_event() 286 domain_pause(d); in arch_monitor_domctl_event() 299 domain_pause(d); in arch_monitor_domctl_event() 315 domain_pause(d); in arch_monitor_domctl_event() [all …]
|
/xen/xen/arch/x86/mm/hap/ |
A D | hap.c | 121 guest_flush_tlb_mask(d, d->dirty_cpumask); in hap_track_dirty_vram() 194 paging_lock(d); in hap_enable_log_dirty() 196 paging_unlock(d); in hap_enable_log_dirty() 208 guest_flush_tlb_mask(d, d->dirty_cpumask); in hap_enable_log_dirty() 215 paging_lock(d); in hap_disable_log_dirty() 237 guest_flush_tlb_mask(d, d->dirty_cpumask); in hap_clean_dirty_bitmap() 411 domain_crash(d); in hap_make_monitor_table() 559 paging_lock(d); in hap_final_teardown() 757 paging_lock(d); in hap_update_paging_modes() 790 paging_lock(d); in hap_write_p2m_entry() [all …]
|
/xen/xen/arch/x86/mm/ |
A D | paging.c | 58 page = d->arch.paging.alloc_page(d); in paging_new_log_dirty_page() 107 d->arch.paging.free_page(d, mfn_to_page(mfn)); in paging_free_log_dirty_page() 115 paging_lock(d); in paging_free_log_dirty_bitmap() 192 paging_free_log_dirty_page(d, d->arch.paging.log_dirty.top); in paging_free_log_dirty_bitmap() 229 ret = d->arch.paging.log_dirty.ops->enable(d, log_global); in paging_log_dirty_enable() 245 ret = d->arch.paging.log_dirty.ops->disable(d); in paging_log_dirty_disable() 291 d->arch.paging.log_dirty.top = paging_new_log_dirty_node(d); in paging_mark_pfn_dirty() 431 paging_lock(d); in paging_log_dirty_op() 567 d->arch.paging.log_dirty.ops->clean(d); in paging_log_dirty_op() 616 guest_flush_tlb_mask(d, d->dirty_cpumask); in paging_log_dirty_range() [all …]
|
A D | mm-locks.h | 68 return l + (d && is_control_domain(d) ? MM_LOCK_ORDER_MAX : 0); in _lock_level() 194 #define mm_lock(name, d, l) mm_lock_##name(d, l, __func__, 0) argument 195 #define mm_lock_recursive(name, d, l) mm_lock_##name(d, l, __func__, 1) argument 196 #define mm_write_lock(name, d, l) mm_write_lock_##name(d, l, __func__) argument 197 #define mm_read_lock(name, d, l) mm_read_lock_##name(d, l) argument 252 #define nestedp2m_lock(d) mm_lock(nestedp2m, d, &(d)->arch.nested_p2m_lock) argument 253 #define nestedp2m_unlock(d) mm_unlock(&(d)->arch.nested_p2m_lock) argument 307 #define altp2m_list_lock(d) mm_lock(altp2mlist, d, \ argument 309 #define altp2m_list_unlock(d) mm_unlock(&(d)->arch.altp2m_list_lock) argument 390 #define paging_lock(d) mm_lock(paging, d, &(d)->arch.paging.lock) argument [all …]
|
/xen/xen/include/xen/ |
A D | event.h | 103 #define group_from_port(d, p) \ argument 111 : BITS_PER_EVTCHN_WORD(d) * BITS_PER_EVTCHN_WORD(d); in max_evtchns() 227 d->evtchn_port_ops->init(d, evtchn); in evtchn_port_init() 235 d->evtchn_port_ops->set_pending(d->vcpu[vcpu_id], evtchn); in evtchn_port_set_pending() 242 d->evtchn_port_ops->clear_pending(d, evtchn); in evtchn_port_clear_pending() 249 d->evtchn_port_ops->unmask(d, evtchn); in evtchn_port_unmask() 255 return evtchn_usable(evtchn) && d->evtchn_port_ops->is_pending(d, evtchn); in evtchn_is_pending() 273 return !evtchn_usable(evtchn) || d->evtchn_port_ops->is_masked(d, evtchn); in evtchn_is_masked() 294 d->evtchn_port_ops->is_busy(d, evtchn); in evtchn_is_busy() 305 return d->evtchn_port_ops->set_priority(d, evtchn, priority); in evtchn_port_set_priority() [all …]
|
A D | sched.h | 545 ASSERT(d->extra_pages <= d->tot_pages); in domain_tot_pages() 547 return d->tot_pages - d->extra_pages; in domain_tot_pages() 555 #define is_idle_domain(d) ((d)->domain_id == DOMID_IDLE) argument 648 return d; in rcu_lock_domain() 787 for (d = rcu_dereference(domain_list); d && d->cpupool != c; in first_domain_in_cpupool() 788 d = rcu_dereference(d->next_in_list)); in first_domain_in_cpupool() 789 return d; in first_domain_in_cpupool() 794 for (d = rcu_dereference(d->next_in_list); d && d->cpupool != c; in next_domain_in_cpupool() 795 d = rcu_dereference(d->next_in_list)); in next_domain_in_cpupool() 1002 return is_pv_domain(d) && d->arch.pv.is_32bit; in is_pv_32bit_domain() [all …]
|
A D | iommu.h | 113 void iommu_hwdom_init(struct domain *d); 237 int (*init)(struct domain *d); 238 void (*hwdom_init)(struct domain *d); 252 void (*teardown)(struct domain *d); 283 void (*share_p2m)(struct domain *d); 341 #define dom_iommu(d) (&(d)->iommu) argument 342 #define iommu_set_feature(d, f) set_bit(f, dom_iommu(d)->features) argument 343 #define iommu_clear_feature(d, f) clear_bit(f, dom_iommu(d)->features) argument 346 #define iommu_use_hap_pt(d) (dom_iommu(d)->hap_pt_share) argument 350 #define need_iommu_pt_sync(d) (dom_iommu(d)->need_sync) argument [all …]
|
/xen/xen/include/xsm/ |
A D | xsm.h | 180 int (*dm_op) (struct domain *d); 232 return xsm_ops->domctl(d, cmd); in xsm_domctl() 388 return xsm_ops->claim_pages(d); in xsm_claim_pages() 398 return xsm_ops->profile(d, op); in xsm_profile() 595 return xsm_ops->mem_access(d); in xsm_mem_access() 602 return xsm_ops->mem_paging(d); in xsm_mem_paging() 609 return xsm_ops->mem_sharing(d); in xsm_mem_sharing() 636 return xsm_ops->apic(d, cmd); in xsm_apic() 688 return xsm_ops->pmu_op(d, op); in xsm_pmu_op() 693 return xsm_ops->dm_op(d); in xsm_dm_op() [all …]
|
A D | dummy.h | 235 if ( d->is_console ) in xsm_console_io() 241 return xsm_default_action(XSM_PRIV, d, NULL); in xsm_console_io() 247 return xsm_default_action(action, d, NULL); in xsm_profile() 297 return xsm_default_action(action, d, NULL); in xsm_evtchn_send() 539 return xsm_default_action(action, d, t); in xsm_map_gmfn_foreign() 643 return xsm_default_action(action, d, NULL); in xsm_apic() 664 rc = xsm_default_action(action, d, f); in xsm_mmu_update() 666 rc = xsm_default_action(action, d, t); in xsm_mmu_update() 673 return xsm_default_action(action, d, f); in xsm_mmuext_op() 680 return xsm_default_action(action, d, f); in xsm_update_va_mapping() [all …]
|
/xen/xen/include/asm-x86/ |
A D | shadow.h | 53 int shadow_domain_init(struct domain *d); 73 int shadow_domctl(struct domain *d, 101 #define shadow_teardown(d, p) ASSERT(is_pv_domain(d)) argument 102 #define shadow_final_teardown(d) ASSERT(is_pv_domain(d)) argument 103 #define shadow_enable(d, mode) \ argument 180 ASSERT(is_pv_domain(d)); in pv_l1tf_check_pte() 183 if ( d->arch.pv.check_l1tf && !paging_mode_sh_forced(d) && in pv_l1tf_check_pte() 191 d->domain_id, level, pte); in pv_l1tf_check_pte() 206 domain_crash(d); in pv_l1tf_check_pte() 238 d->arch.pv.check_l1tf = is_hardware_domain(d) ? opt_pv_l1tf_hwdom in pv_l1tf_domain_init() [all …]
|
/xen/xen/drivers/passthrough/ |
A D | iommu.c | 174 if ( !is_iommu_enabled(d) ) in iommu_domain_init() 237 iommu_teardown(d); in iommu_domain_destroy() 277 domain_crash(d); in iommu_map() 333 domain_crash(d); in iommu_unmap() 409 domain_crash(d); in iommu_iotlb_flush() 437 domain_crash(d); in iommu_iotlb_flush_all() 552 ASSERT(hap_enabled(d)); in iommu_share_p2m_table() 590 return is_iommu_enabled(d) && test_bit(feature, dom_iommu(d)->features); in iommu_has_feature() 595 struct domain *d; in iommu_dump_p2m_table() local 608 for_each_domain(d) in iommu_dump_p2m_table() [all …]
|
/xen/xen/arch/x86/mm/shadow/ |
A D | common.c | 395 guest_flush_tlb_mask(d, d->dirty_cpumask); in oos_remove_write_access() 940 ASSERT(d->vcpu && d->vcpu[0]); in _shadow_prealloc() 973 guest_flush_tlb_mask(d, d->dirty_cpumask); in _shadow_prealloc() 1010 ASSERT(d->vcpu && d->vcpu[0]); in shadow_blow_tables() 1027 guest_flush_tlb_mask(d, d->dirty_cpumask); in shadow_blow_tables() 1032 if ( shadow_mode_enabled(d) && domain_vcpu(d, 0) ) in shadow_blow_tables_per_domain() 2295 guest_flush_tlb_mask(d, d->dirty_cpumask); in sh_remove_shadows() 2890 sh_new_mode(d, d->arch.paging.mode | mode); in shadow_one_bit_enable() 3012 guest_flush_tlb_mask(d, d->dirty_cpumask); in sh_unshadow_for_p2m_change() 3337 guest_flush_tlb_mask(d, d->dirty_cpumask); in shadow_track_dirty_vram() [all …]
|
/xen/xen/arch/x86/hvm/ |
A D | irq.c | 40 pirq_info(d, domain_emuirq_to_pirq(d, ioapic_gsi)); in assert_gsi() 63 assert_gsi(d, gsi); in hvm_ioapic_assert() 87 assert_gsi(d, ioapic_gsi); in assert_irq() 95 pirq_info(d, domain_emuirq_to_pirq(d, isa_irq)); in deassert_irq() 119 assert_gsi(d, gsi); in __hvm_pci_intx_assert() 194 assert_gsi(d, gsi); in hvm_gsi_assert() 440 (!has_vlapic(d) || !has_vioapic(d) || !has_vpic(d)) ) in hvm_set_callback_via() 495 for_each_vcpu ( d, v ) in hvm_set_callback_via() 640 struct domain *d; in dump_irq_info() local 646 for_each_domain ( d ) in dump_irq_info() [all …]
|
A D | save.c | 73 if ( d->arch.vtsc ) in arch_hvm_load() 143 if ( d->is_dying || in hvm_save_one() 150 v = d->vcpu[0]; in hvm_save_one() 151 else if ( instance >= d->max_vcpus || !d->vcpu[instance] ) in hvm_save_one() 213 if ( d->is_dying ) in hvm_save() 363 d->typecode = tc; in _hvm_init_entry() 365 d->length = len; in _hvm_init_entry() 393 (strict_length ? (len != d->length) : (len < d->length)) || in _hvm_check_entry() 394 (d->length > (h->size - h->cur - sizeof(*d))) ) in _hvm_check_entry() 400 d->typecode, d->length, h->size - h->cur - sizeof(*d)); in _hvm_check_entry() [all …]
|
/xen/xen/include/acpi/ |
A D | acmacros.h | 155 #define ACPI_MOVE_16_TO_32(d,s) {(*(u32 *)(void *)(d))=0;\ argument 159 #define ACPI_MOVE_16_TO_64(d,s) {(*(u64 *)(void *)(d))=0;\ argument 172 #define ACPI_MOVE_32_TO_64(d,s) {(*(u64 *)(void *)(d))=0;\ argument 235 #define ACPI_MOVE_16_TO_32(d,s) {(*(u32 *)(void *)(d)) = 0; ACPI_MOVE_16_TO_16(d,s);} argument 236 #define ACPI_MOVE_16_TO_64(d,s) {(*(u64 *)(void *)(d)) = 0; ACPI_MOVE_16_TO_16(d,s);} argument 247 #define ACPI_MOVE_32_TO_64(d,s) {(*(u64 *)(void *)(d)) = 0; ACPI_MOVE_32_TO_32(d,s);} argument 267 #define ACPI_MOVE_SIZE_TO_16(d,s) ACPI_MOVE_32_TO_16(d,s) argument 270 #define ACPI_MOVE_SIZE_TO_16(d,s) ACPI_MOVE_64_TO_16(d,s) argument 343 #define ACPI_IS_OCTAL_DIGIT(d) (((char)(d) >= '0') && ((char)(d) <= '7')) argument 404 #define ARGI_LIST4(a,b,c,d) (ARG_1(d)|ARG_2(c)|ARG_3(b)|ARG_4(a)) argument [all …]
|
/xen/stubdom/ |
A D | pciutils.patch | 105 + d->domain = domain; 106 + d->bus = bus; 107 + d->dev = slot; 108 + d->func = fun; 122 + if (pcifront_conf_read(NULL, d->domain, d->bus, d->dev, d->func, pos, len, &val)) 127 + if (pcifront_conf_read(NULL, d->domain, d->bus, d->dev, d->func, pos, len, &val)) 132 + if (pcifront_conf_read(NULL, d->domain, d->bus, d->dev, d->func, pos, len, &val)) 158 + return !pcifront_conf_write(NULL, d->domain, d->bus, d->dev, d->func, pos, len, val); 228 …a->warning("%04x:%02x:%02x.%d: Invalid 64-bit address seen for BAR %d.", d->domain, d->bus, d->dev… 247 …a->warning("%04x:%02x:%02x.%d 64-bit device address ignored.", d->domain, d->bus, d->dev, d->func); [all …]
|