/linux/drivers/s390/net/ |
A D | ism_drv.c | 222 dmb->cpu_addr, dmb->dma_addr); in ism_free_dmb() 232 if (!dmb->sba_idx) { in ism_alloc_dmb() 238 dmb->sba_idx = bit; in ism_alloc_dmb() 244 dmb->cpu_addr = dma_alloc_coherent(&ism->pdev->dev, dmb->dmb_len, in ism_alloc_dmb() 245 &dmb->dma_addr, in ism_alloc_dmb() 247 if (!dmb->cpu_addr) in ism_alloc_dmb() 259 ret = ism_alloc_dmb(ism, dmb); in ism_register_dmb() 267 cmd.request.dmb = dmb->dma_addr; in ism_register_dmb() 272 cmd.request.rgid = dmb->rgid; in ism_register_dmb() 276 ism_free_dmb(ism, dmb); in ism_register_dmb() [all …]
|
A D | ism.h | 116 u64 dmb; member 207 #define ISM_CREATE_REQ(dmb, idx, sf, offset) \ argument 208 ((dmb) | (idx) << 24 | (sf) << 23 | (offset))
|
/linux/net/smc/ |
A D | smc_ism.c | 172 struct smcd_dmb dmb; in smc_ism_unregister_dmb() local 178 memset(&dmb, 0, sizeof(dmb)); in smc_ism_unregister_dmb() 179 dmb.dmb_tok = dmb_desc->token; in smc_ism_unregister_dmb() 183 dmb.dmb_len = dmb_desc->len; in smc_ism_unregister_dmb() 196 struct smcd_dmb dmb; in smc_ism_register_dmb() local 199 memset(&dmb, 0, sizeof(dmb)); in smc_ism_register_dmb() 200 dmb.dmb_len = dmb_len; in smc_ism_register_dmb() 202 dmb.vlan_id = lgr->vlan_id; in smc_ism_register_dmb() 203 dmb.rgid = lgr->peer_gid; in smc_ism_register_dmb() 207 dmb_desc->token = dmb.dmb_tok; in smc_ism_register_dmb() [all …]
|
/linux/arch/arm/include/asm/ |
A D | barrier.h | 21 #define dmb(option) __asm__ __volatile__ ("dmb " #option : : : "memory") macro 33 #define dmb(x) __asm__ __volatile__ ("mcr p15, 0, %0, c7, c10, 5" \ macro 40 #define dmb(x) __asm__ __volatile__ ("" : : : "memory") macro 45 #define dmb(x) __asm__ __volatile__ ("" : : : "memory") macro 67 #define dma_rmb() dmb(osh) 68 #define dma_wmb() dmb(oshst) 77 #define __smp_mb() dmb(ish) 79 #define __smp_wmb() dmb(ishst)
|
A D | assembler.h | 329 ALT_SMP(dmb ish) 331 ALT_SMP(W(dmb) ish) 334 ALT_SMP(mcr p15, 0, r0, c7, c10, 5) @ dmb
|
/linux/arch/arm64/include/asm/vdso/ |
A D | compat_barrier.h | 17 #ifdef dmb 18 #undef dmb 21 #define dmb(option) __asm__ __volatile__ ("dmb " #option : : : "memory") macro 23 #define aarch32_smp_mb() dmb(ish) 24 #define aarch32_smp_rmb() dmb(ishld) 25 #define aarch32_smp_wmb() dmb(ishst)
|
/linux/arch/arm64/include/asm/ |
A D | barrier.h | 22 #define dmb(opt) asm volatile("dmb " #opt : : : "memory") macro 45 #define dma_mb() dmb(osh) 46 #define dma_rmb() dmb(oshld) 47 #define dma_wmb() dmb(oshst) 103 #define __smp_mb() dmb(ish) 104 #define __smp_rmb() dmb(ishld) 105 #define __smp_wmb() dmb(ishst)
|
A D | atomic_ll_sc.h | 102 ATOMIC_OP_RETURN( , dmb ish, , l, "memory", __VA_ARGS__)\ 106 ATOMIC_FETCH_OP ( , dmb ish, , l, "memory", __VA_ARGS__)\ 117 ATOMIC_FETCH_OP ( , dmb ish, , l, "memory", __VA_ARGS__)\ in ATOMIC_OPS() 201 ATOMIC64_OP_RETURN(, dmb ish, , l, "memory", __VA_ARGS__) \ 205 ATOMIC64_FETCH_OP (, dmb ish, , l, "memory", __VA_ARGS__) \ 216 ATOMIC64_FETCH_OP (, dmb ish, , l, "memory", __VA_ARGS__) \ 311 __CMPXCHG_CASE(w, b, mb_, 8, dmb ish, , l, "memory", K) 312 __CMPXCHG_CASE(w, h, mb_, 16, dmb ish, , l, "memory", K) 313 __CMPXCHG_CASE(w, , mb_, 32, dmb ish, , l, "memory", K) 314 __CMPXCHG_CASE( , , mb_, 64, dmb ish, , l, "memory", L) [all …]
|
A D | cmpxchg.h | 57 __XCHG_CASE(w, b, mb_, 8, dmb ish, nop, , a, l, "memory") 58 __XCHG_CASE(w, h, mb_, 16, dmb ish, nop, , a, l, "memory") 59 __XCHG_CASE(w, , mb_, 32, dmb ish, nop, , a, l, "memory") 60 __XCHG_CASE( , , mb_, 64, dmb ish, nop, , a, l, "memory")
|
/linux/tools/virtio/asm/ |
A D | barrier.h | 20 #define dmb(opt) asm volatile("dmb " #opt : : : "memory") macro 22 #define virt_rmb() dmb(ishld) 23 #define virt_wmb() dmb(ishst) 24 #define virt_store_mb(var, value) do { WRITE_ONCE(var, value); dmb(ish); } while (0)
|
/linux/arch/arm/common/ |
A D | mcpm_head.S | 121 dmb 136 dmb 148 dmb 152 dmb 173 dmb 182 dmb 196 dmb
|
A D | vlock.S | 29 dmb 33 dmb 80 dmb 93 dmb
|
A D | mcpm_entry.c | 49 dmb(); in __mcpm_cpu_down() 65 dmb(); in __mcpm_outbound_leave_critical()
|
/linux/tools/testing/selftests/kvm/include/aarch64/ |
A D | processor.h | 132 #define dmb(opt) asm volatile("dmb " #opt : : : "memory") macro 134 #define dma_wmb() dmb(oshst) 137 #define dma_rmb() dmb(oshld)
|
/linux/include/net/ |
A D | smc.h | 57 int (*register_dmb)(struct smcd_dev *dev, struct smcd_dmb *dmb); 58 int (*unregister_dmb)(struct smcd_dev *dev, struct smcd_dmb *dmb);
|
/linux/arch/arm64/kernel/ |
A D | head.S | 118 dmb sy // needed before dc ivac with 306 dmb sy 328 dmb sy 381 dmb sy 572 dmb sy 685 dmb sy
|
/linux/arch/arm/mach-omap2/ |
A D | sleep34xx.S | 97 dmb @ data memory barrier 213 dmb 418 dmb @ data memory barrier 429 dmb @ data memory barrier 444 dmb @ data memory barrier
|
A D | omap-smc.S | 52 dmb
|
A D | sleep33xx.S | 133 dmb
|
/linux/arch/arm/mm/ |
A D | cache-b15-rac.c | 66 dmb(); in __b15_rac_disable() 80 dmb(); in __b15_rac_flush()
|
/linux/arch/arm64/mm/ |
A D | flush.c | 85 dmb(osh); in arch_wb_cache_pmem()
|
A D | proc.S | 207 dmb sy // lines are written back before 216 dmb sy // that it is visible to all
|
/linux/arch/arm/mach-socfpga/ |
A D | self-refresh.S | 85 dmb
|
/linux/arch/arm/mach-tegra/ |
A D | sleep.S | 35 dmb @ ensure ordering
|
/linux/arch/arm/kernel/ |
A D | smp_tlb.c | 153 dmb(); in ipi_flush_tlb_a15_erratum()
|