Searched refs:irq_lock (Results 1 – 11 of 11) sorted by relevance
/xen/xen/arch/arm/vgic/ |
A D | vgic.c | 219 spin_lock(&irqa->irq_lock); in vgic_irq_cmp() 220 spin_lock(&irqb->irq_lock); in vgic_irq_cmp() 240 spin_unlock(&irqb->irq_lock); in vgic_irq_cmp() 241 spin_unlock(&irqa->irq_lock); in vgic_irq_cmp() 327 spin_lock(&irq->irq_lock); in vgic_queue_irq_unlock() 358 spin_unlock(&irq->irq_lock); in vgic_queue_irq_unlock() 436 spin_lock(&irq->irq_lock); in vgic_prune_ap_list() 492 spin_lock(&irq->irq_lock); in vgic_prune_ap_list() 571 spin_lock(&irq->irq_lock); in vgic_flush_lr_state() 918 spin_lock(&irq->irq_lock); in vgic_sync_hardware_irq() [all …]
|
A D | vgic-mmio.c | 80 spin_lock_irqsave(&irq->irq_lock, flags); in vgic_mmio_write_senable() 126 spin_lock_irqsave(&irq->irq_lock, flags); in vgic_mmio_write_cenable() 150 spin_unlock_irqrestore(&irq->irq_lock, flags); in vgic_mmio_write_cenable() 193 spin_lock_irqsave(&irq->irq_lock, flags); in vgic_mmio_write_spending() 213 spin_lock(&irq->irq_lock); in vgic_mmio_write_spending() 220 spin_unlock(&irq->irq_lock); in vgic_mmio_write_spending() 241 spin_lock_irqsave(&irq->irq_lock, flags); in vgic_mmio_write_cpending() 266 spin_lock(&irq->irq_lock); in vgic_mmio_write_cpending() 279 spin_unlock(&irq->irq_lock); in vgic_mmio_write_cpending() 417 spin_lock_irqsave(&irq->irq_lock, flags); in vgic_mmio_write_priority() [all …]
|
A D | vgic-mmio-v2.c | 117 spin_lock_irqsave(&irq->irq_lock, flags); in vgic_mmio_write_sgir() 163 spin_lock_irqsave(&irq->irq_lock, flags); in vgic_mmio_write_target() 179 spin_unlock_irqrestore(&irq->irq_lock, flags); in vgic_mmio_write_target() 219 spin_lock_irqsave(&irq->irq_lock, flags); in vgic_mmio_write_sgipendc() 225 spin_unlock_irqrestore(&irq->irq_lock, flags); in vgic_mmio_write_sgipendc() 244 spin_lock_irqsave(&irq->irq_lock, flags); in vgic_mmio_write_sgipends() 255 spin_unlock_irqrestore(&irq->irq_lock, flags); in vgic_mmio_write_sgipends()
|
A D | vgic-init.c | 75 spin_lock_init(&irq->irq_lock); in vgic_vcpu_early_init() 161 spin_lock_init(&irq->irq_lock); in domain_vgic_init()
|
A D | vgic-v2.c | 111 spin_lock(&irq->irq_lock); in vgic_v2_fold_lr_state() 162 spin_unlock(&irq->irq_lock); in vgic_v2_fold_lr_state()
|
/xen/xen/arch/x86/hvm/ |
A D | irq.c | 61 spin_lock(&d->arch.hvm.irq_lock); in hvm_ioapic_assert() 65 spin_unlock(&d->arch.hvm.irq_lock); in hvm_ioapic_assert() 80 spin_lock(&d->arch.hvm.irq_lock); in hvm_ioapic_deassert() 82 spin_unlock(&d->arch.hvm.irq_lock); in hvm_ioapic_deassert() 131 spin_lock(&d->arch.hvm.irq_lock); in hvm_pci_intx_assert() 165 spin_lock(&d->arch.hvm.irq_lock); in hvm_pci_intx_deassert() 189 spin_lock(&d->arch.hvm.irq_lock); in hvm_gsi_assert() 210 spin_lock(&d->arch.hvm.irq_lock); in hvm_gsi_deassert() 225 spin_lock(&d->arch.hvm.irq_lock); in hvm_isa_irq_assert() 247 spin_lock(&d->arch.hvm.irq_lock); in hvm_isa_irq_deassert() [all …]
|
A D | vioapic.c | 233 spin_lock(&d->arch.hvm.irq_lock); in vioapic_write_redirent() 269 spin_unlock(&d->arch.hvm.irq_lock); in vioapic_write_redirent() 411 ASSERT(spin_is_locked(&d->arch.hvm.irq_lock)); in vioapic_deliver() 499 ASSERT(spin_is_locked(&d->arch.hvm.irq_lock)); in vioapic_irq_positive_edge() 524 spin_lock(&d->arch.hvm.irq_lock); in vioapic_update_EOI() 541 spin_unlock(&d->arch.hvm.irq_lock); in vioapic_update_EOI() 543 spin_lock(&d->arch.hvm.irq_lock); in vioapic_update_EOI() 556 spin_unlock(&d->arch.hvm.irq_lock); in vioapic_update_EOI()
|
A D | vpic.c | 40 vpic[!(v)->is_master])->irq_lock
|
A D | hvm.c | 621 spin_lock_init(&d->arch.hvm.irq_lock); in hvm_domain_initialise()
|
/xen/xen/include/asm-arm/ |
A D | new_vgic.h | 66 spinlock_t irq_lock; /* Protects the content of the struct */ member
|
/xen/xen/include/asm-x86/hvm/ |
A D | domain.h | 125 spinlock_t irq_lock; member
|
Completed in 33 milliseconds