xchg_relaxed 131 arch/arm/include/asm/cmpxchg.h #define xchg xchg_relaxed xchg_relaxed 211 arch/arm64/include/asm/percpu.h _pcp_protect_return(xchg_relaxed, pcp, val) xchg_relaxed 213 arch/arm64/include/asm/percpu.h _pcp_protect_return(xchg_relaxed, pcp, val) xchg_relaxed 215 arch/arm64/include/asm/percpu.h _pcp_protect_return(xchg_relaxed, pcp, val) xchg_relaxed 217 arch/arm64/include/asm/percpu.h _pcp_protect_return(xchg_relaxed, pcp, val) xchg_relaxed 760 arch/arm64/include/asm/pgtable.h return __pte(xchg_relaxed(&pte_val(*ptep), 0)); xchg_relaxed 808 arch/arm64/include/asm/pgtable.h return __pmd(xchg_relaxed(&pmd_val(*pmdp), pmd_val(pmd))); xchg_relaxed 76 arch/mips/mm/context.c mmid = xchg_relaxed(&cpu_data[cpu].asid_cache, 0); xchg_relaxed 202 arch/powerpc/include/asm/atomic.h #define atomic_xchg_relaxed(v, new) xchg_relaxed(&((v)->counter), (new)) xchg_relaxed 494 arch/powerpc/include/asm/atomic.h #define atomic64_xchg_relaxed(v, new) xchg_relaxed(&((v)->counter), (new)) xchg_relaxed 9 include/linux/atomic-fallback.h #ifndef xchg_relaxed xchg_relaxed 181 kernel/locking/qspinlock.c return (u32)xchg_relaxed(&lock->tail,