Home
last modified time | relevance | path

Searched refs:irq_lock (Results 1 – 11 of 11) sorted by relevance

/xen/xen/arch/arm/vgic/
A Dvgic.c219 spin_lock(&irqa->irq_lock); in vgic_irq_cmp()
220 spin_lock(&irqb->irq_lock); in vgic_irq_cmp()
240 spin_unlock(&irqb->irq_lock); in vgic_irq_cmp()
241 spin_unlock(&irqa->irq_lock); in vgic_irq_cmp()
327 spin_lock(&irq->irq_lock); in vgic_queue_irq_unlock()
358 spin_unlock(&irq->irq_lock); in vgic_queue_irq_unlock()
436 spin_lock(&irq->irq_lock); in vgic_prune_ap_list()
492 spin_lock(&irq->irq_lock); in vgic_prune_ap_list()
571 spin_lock(&irq->irq_lock); in vgic_flush_lr_state()
918 spin_lock(&irq->irq_lock); in vgic_sync_hardware_irq()
[all …]
A Dvgic-mmio.c80 spin_lock_irqsave(&irq->irq_lock, flags); in vgic_mmio_write_senable()
126 spin_lock_irqsave(&irq->irq_lock, flags); in vgic_mmio_write_cenable()
150 spin_unlock_irqrestore(&irq->irq_lock, flags); in vgic_mmio_write_cenable()
193 spin_lock_irqsave(&irq->irq_lock, flags); in vgic_mmio_write_spending()
213 spin_lock(&irq->irq_lock); in vgic_mmio_write_spending()
220 spin_unlock(&irq->irq_lock); in vgic_mmio_write_spending()
241 spin_lock_irqsave(&irq->irq_lock, flags); in vgic_mmio_write_cpending()
266 spin_lock(&irq->irq_lock); in vgic_mmio_write_cpending()
279 spin_unlock(&irq->irq_lock); in vgic_mmio_write_cpending()
417 spin_lock_irqsave(&irq->irq_lock, flags); in vgic_mmio_write_priority()
[all …]
A Dvgic-mmio-v2.c117 spin_lock_irqsave(&irq->irq_lock, flags); in vgic_mmio_write_sgir()
163 spin_lock_irqsave(&irq->irq_lock, flags); in vgic_mmio_write_target()
179 spin_unlock_irqrestore(&irq->irq_lock, flags); in vgic_mmio_write_target()
219 spin_lock_irqsave(&irq->irq_lock, flags); in vgic_mmio_write_sgipendc()
225 spin_unlock_irqrestore(&irq->irq_lock, flags); in vgic_mmio_write_sgipendc()
244 spin_lock_irqsave(&irq->irq_lock, flags); in vgic_mmio_write_sgipends()
255 spin_unlock_irqrestore(&irq->irq_lock, flags); in vgic_mmio_write_sgipends()
A Dvgic-init.c75 spin_lock_init(&irq->irq_lock); in vgic_vcpu_early_init()
161 spin_lock_init(&irq->irq_lock); in domain_vgic_init()
A Dvgic-v2.c111 spin_lock(&irq->irq_lock); in vgic_v2_fold_lr_state()
162 spin_unlock(&irq->irq_lock); in vgic_v2_fold_lr_state()
/xen/xen/arch/x86/hvm/
A Dirq.c61 spin_lock(&d->arch.hvm.irq_lock); in hvm_ioapic_assert()
65 spin_unlock(&d->arch.hvm.irq_lock); in hvm_ioapic_assert()
80 spin_lock(&d->arch.hvm.irq_lock); in hvm_ioapic_deassert()
82 spin_unlock(&d->arch.hvm.irq_lock); in hvm_ioapic_deassert()
131 spin_lock(&d->arch.hvm.irq_lock); in hvm_pci_intx_assert()
165 spin_lock(&d->arch.hvm.irq_lock); in hvm_pci_intx_deassert()
189 spin_lock(&d->arch.hvm.irq_lock); in hvm_gsi_assert()
210 spin_lock(&d->arch.hvm.irq_lock); in hvm_gsi_deassert()
225 spin_lock(&d->arch.hvm.irq_lock); in hvm_isa_irq_assert()
247 spin_lock(&d->arch.hvm.irq_lock); in hvm_isa_irq_deassert()
[all …]
A Dvioapic.c233 spin_lock(&d->arch.hvm.irq_lock); in vioapic_write_redirent()
269 spin_unlock(&d->arch.hvm.irq_lock); in vioapic_write_redirent()
411 ASSERT(spin_is_locked(&d->arch.hvm.irq_lock)); in vioapic_deliver()
499 ASSERT(spin_is_locked(&d->arch.hvm.irq_lock)); in vioapic_irq_positive_edge()
524 spin_lock(&d->arch.hvm.irq_lock); in vioapic_update_EOI()
541 spin_unlock(&d->arch.hvm.irq_lock); in vioapic_update_EOI()
543 spin_lock(&d->arch.hvm.irq_lock); in vioapic_update_EOI()
556 spin_unlock(&d->arch.hvm.irq_lock); in vioapic_update_EOI()
A Dvpic.c40 vpic[!(v)->is_master])->irq_lock
A Dhvm.c621 spin_lock_init(&d->arch.hvm.irq_lock); in hvm_domain_initialise()
/xen/xen/include/asm-arm/
A Dnew_vgic.h66 spinlock_t irq_lock; /* Protects the content of the struct */ member
/xen/xen/include/asm-x86/hvm/
A Ddomain.h125 spinlock_t irq_lock; member

Completed in 27 milliseconds