__smp_mb           77 arch/arm/include/asm/barrier.h #define __smp_mb()	dmb(ish)
__smp_mb           60 arch/arm64/include/asm/barrier.h #define __smp_mb()	dmb(ish)
__smp_mb           34 arch/csky/include/asm/barrier.h #define __smp_mb()	asm volatile ("bar.brwarws\n":::"memory")
__smp_mb           46 arch/ia64/include/asm/barrier.h # define __smp_mb()	mb()
__smp_mb          200 arch/mips/include/asm/barrier.h #  define __smp_mb()	__sync()
__smp_mb          204 arch/mips/include/asm/barrier.h #  define __smp_mb()	__asm__ __volatile__("sync" : : :"memory")
__smp_mb          209 arch/mips/include/asm/barrier.h #define __smp_mb()	barrier()
__smp_mb           25 arch/parisc/include/asm/barrier.h #define __smp_mb()	mb()
__smp_mb           53 arch/powerpc/include/asm/barrier.h #define __smp_mb()	mb()
__smp_mb           26 arch/riscv/include/asm/barrier.h #define __smp_mb()	RISCV_FENCE(rw,rw)
__smp_mb           30 arch/s390/include/asm/barrier.h #define __smp_mb()			mb()
__smp_mb           34 arch/sh/include/asm/barrier.h #define __smp_mb()	do { int tmp = 0; __asm__ __volatile__ ("cas.l %0,%0,@%1" : "+r"(tmp) : "z"(&tmp) : "memory", "t"); } while(0)
__smp_mb           58 arch/x86/include/asm/barrier.h #define __smp_mb()	asm volatile("lock; addl $0,-4(%%esp)" ::: "memory", "cc")
__smp_mb           60 arch/x86/include/asm/barrier.h #define __smp_mb()	asm volatile("lock; addl $0,-4(%%rsp)" ::: "memory", "cc")
__smp_mb           54 include/asm-generic/barrier.h #define __smp_mb()	mb()