xchg_relaxed      131 arch/arm/include/asm/cmpxchg.h #define xchg xchg_relaxed
xchg_relaxed      211 arch/arm64/include/asm/percpu.h 	_pcp_protect_return(xchg_relaxed, pcp, val)
xchg_relaxed      213 arch/arm64/include/asm/percpu.h 	_pcp_protect_return(xchg_relaxed, pcp, val)
xchg_relaxed      215 arch/arm64/include/asm/percpu.h 	_pcp_protect_return(xchg_relaxed, pcp, val)
xchg_relaxed      217 arch/arm64/include/asm/percpu.h 	_pcp_protect_return(xchg_relaxed, pcp, val)
xchg_relaxed      760 arch/arm64/include/asm/pgtable.h 	return __pte(xchg_relaxed(&pte_val(*ptep), 0));
xchg_relaxed      808 arch/arm64/include/asm/pgtable.h 	return __pmd(xchg_relaxed(&pmd_val(*pmdp), pmd_val(pmd)));
xchg_relaxed       76 arch/mips/mm/context.c 		mmid = xchg_relaxed(&cpu_data[cpu].asid_cache, 0);
xchg_relaxed      202 arch/powerpc/include/asm/atomic.h #define atomic_xchg_relaxed(v, new) xchg_relaxed(&((v)->counter), (new))
xchg_relaxed      494 arch/powerpc/include/asm/atomic.h #define atomic64_xchg_relaxed(v, new) xchg_relaxed(&((v)->counter), (new))
xchg_relaxed        9 include/linux/atomic-fallback.h #ifndef xchg_relaxed
xchg_relaxed      181 kernel/locking/qspinlock.c 	return (u32)xchg_relaxed(&lock->tail,