/linux/arch/mips/include/asm/ |
A D | barrier.h | 19 static inline void rmb(void) in rmb() function 21 asm volatile(__SYNC(rmb, always) ::: "memory"); in rmb() 23 #define rmb rmb macro 83 # define __smp_rmb() rmb()
|
A D | io.h | 87 #define iobarrier_r() rmb() 305 rmb(); \ 347 rmb(); \
|
/linux/arch/csky/include/asm/ |
A D | io.h | 21 #define readb(c) ({ u8 __v = readb_relaxed(c); rmb(); __v; }) 22 #define readw(c) ({ u16 __v = readw_relaxed(c); rmb(); __v; }) 23 #define readl(c) ({ u32 __v = readl_relaxed(c); rmb(); __v; })
|
/linux/include/asm-generic/ |
A D | barrier.h | 34 #ifndef rmb 35 #define rmb() mb() macro 43 #define dma_rmb() rmb() 55 #define __smp_rmb() rmb()
|
/linux/tools/include/asm-generic/ |
A D | barrier.h | 31 #ifndef rmb 32 #define rmb() mb() macro
|
/linux/arch/x86/um/asm/ |
A D | barrier.h | 15 #define rmb() alternative("lock; addl $0,0(%%esp)", "lfence", X86_FEATURE_XMM2) macro 21 #define rmb() asm volatile("lfence" : : : "memory") macro
|
/linux/tools/arch/x86/include/asm/ |
A D | barrier.h | 21 #define rmb() asm volatile("lock; addl $0,0(%%esp)" ::: "memory") macro 25 #define rmb() asm volatile("lfence" ::: "memory") macro
|
A D | pvclock.h | 22 rmb(); in pvclock_read_begin() 31 rmb(); in pvclock_read_retry()
|
/linux/arch/powerpc/lib/ |
A D | locks.c | 34 rmb(); in splpar_spin_yield() 60 rmb(); in splpar_rw_yield()
|
/linux/arch/x86/include/asm/ |
A D | barrier.h | 17 #define rmb() asm volatile(ALTERNATIVE("lock; addl $0,-4(%%esp)", "lfence", \ macro 23 #define rmb() asm volatile("lfence":::"memory") macro
|
/linux/arch/s390/include/asm/ |
A D | barrier.h | 30 #define rmb() barrier() macro 35 #define __smp_rmb() rmb()
|
/linux/drivers/video/fbdev/ |
A D | bt455.h | 46 rmb(); in bt455_read_cmap_next() 48 rmb(); in bt455_read_cmap_next()
|
/linux/arch/parisc/include/asm/ |
A D | barrier.h | 17 #define rmb() mb() macro 23 #define rmb() barrier() macro
|
/linux/arch/sh/include/asm/ |
A D | io.h | 49 #define readb(a) ({ u8 r_ = readb_relaxed(a); rmb(); r_; }) 50 #define readw(a) ({ u16 r_ = readw_relaxed(a); rmb(); r_; }) 51 #define readl(a) ({ u32 r_ = readl_relaxed(a); rmb(); r_; }) 52 #define readq(a) ({ u64 r_ = readq_relaxed(a); rmb(); r_; })
|
/linux/arch/arm/vfp/ |
A D | vfp.h | 73 u64 rh, rma, rmb, rl; in mul64to128() local 83 rmb = (u64)nl * mh; in mul64to128() 84 rma += rmb; in mul64to128() 87 rh += ((u64)(rma < rmb) << 32) + (rma >> 32); in mul64to128()
|
/linux/tools/testing/selftests/powerpc/dscr/ |
A D | dscr_default_test.c | 30 rmb(); in do_test() 36 rmb(); in do_test()
|
/linux/arch/arm/include/asm/ |
A D | barrier.h | 65 #define rmb() dsb() macro 71 #define rmb() barrier() macro
|
/linux/arch/x86/kernel/ |
A D | pvclock.c | 124 rmb(); /* fetch version before time */ in pvclock_read_wallclock() 134 rmb(); /* fetch time before checking version */ in pvclock_read_wallclock()
|
/linux/drivers/xen/ |
A D | time.c | 67 rmb(); /* Hypervisor might update data. */ in xen_get_runstate_snapshot_cpu_delta() 69 rmb(); /* Hypervisor might update data. */ in xen_get_runstate_snapshot_cpu_delta()
|
/linux/arch/x86/kernel/cpu/resctrl/ |
A D | pseudo_lock.c | 493 rmb(); in pseudo_lock_fn() 504 rmb(); in pseudo_lock_fn() 1002 rmb(); in measure_residency_fn() 1009 rmb(); in measure_residency_fn() 1015 rmb(); in measure_residency_fn() 1025 rmb(); in measure_residency_fn() 1032 rmb(); in measure_residency_fn()
|
/linux/arch/powerpc/kernel/ |
A D | smp-tbsync.c | 54 rmb(); in smp_generic_take_timebase() 60 rmb(); in smp_generic_take_timebase()
|
/linux/tools/perf/util/ |
A D | tsc.c | 51 rmb(); in perf_read_tsc_conversion() 59 rmb(); in perf_read_tsc_conversion()
|
/linux/tools/arch/arm/include/asm/ |
A D | barrier.h | 10 #define rmb() ((void(*)(void))0xffff0fa0)() macro
|
/linux/drivers/scsi/aic94xx/ |
A D | aic94xx_reg.c | 60 rmb(); in asd_read_byte() 73 rmb(); in asd_read_word() 86 rmb(); in asd_read_dword()
|
/linux/arch/nds32/include/asm/ |
A D | barrier.h | 9 #define rmb() asm volatile("msync all":::"memory") macro
|