__smp_mb 77 arch/arm/include/asm/barrier.h #define __smp_mb() dmb(ish) __smp_mb 60 arch/arm64/include/asm/barrier.h #define __smp_mb() dmb(ish) __smp_mb 34 arch/csky/include/asm/barrier.h #define __smp_mb() asm volatile ("bar.brwarws\n":::"memory") __smp_mb 46 arch/ia64/include/asm/barrier.h # define __smp_mb() mb() __smp_mb 200 arch/mips/include/asm/barrier.h # define __smp_mb() __sync() __smp_mb 204 arch/mips/include/asm/barrier.h # define __smp_mb() __asm__ __volatile__("sync" : : :"memory") __smp_mb 209 arch/mips/include/asm/barrier.h #define __smp_mb() barrier() __smp_mb 25 arch/parisc/include/asm/barrier.h #define __smp_mb() mb() __smp_mb 53 arch/powerpc/include/asm/barrier.h #define __smp_mb() mb() __smp_mb 26 arch/riscv/include/asm/barrier.h #define __smp_mb() RISCV_FENCE(rw,rw) __smp_mb 30 arch/s390/include/asm/barrier.h #define __smp_mb() mb() __smp_mb 34 arch/sh/include/asm/barrier.h #define __smp_mb() do { int tmp = 0; __asm__ __volatile__ ("cas.l %0,%0,@%1" : "+r"(tmp) : "z"(&tmp) : "memory", "t"); } while(0) __smp_mb 58 arch/x86/include/asm/barrier.h #define __smp_mb() asm volatile("lock; addl $0,-4(%%esp)" ::: "memory", "cc") __smp_mb 60 arch/x86/include/asm/barrier.h #define __smp_mb() asm volatile("lock; addl $0,-4(%%rsp)" ::: "memory", "cc") __smp_mb 54 include/asm-generic/barrier.h #define __smp_mb() mb()