cmpxchg_relaxed   233 arch/arm64/include/asm/kvm_mmu.h 		pteval = cmpxchg_relaxed(&pte_val(*ptep), old_pteval, pteval);
cmpxchg_relaxed   220 arch/arm64/include/asm/percpu.h 	_pcp_protect_return(cmpxchg_relaxed, pcp, o, n)
cmpxchg_relaxed   222 arch/arm64/include/asm/percpu.h 	_pcp_protect_return(cmpxchg_relaxed, pcp, o, n)
cmpxchg_relaxed   224 arch/arm64/include/asm/percpu.h 	_pcp_protect_return(cmpxchg_relaxed, pcp, o, n)
cmpxchg_relaxed   226 arch/arm64/include/asm/percpu.h 	_pcp_protect_return(cmpxchg_relaxed, pcp, o, n)
cmpxchg_relaxed   711 arch/arm64/include/asm/pgtable.h 		pte_val(pte) = cmpxchg_relaxed(&pte_val(*ptep),
cmpxchg_relaxed   791 arch/arm64/include/asm/pgtable.h 		pte_val(pte) = cmpxchg_relaxed(&pte_val(*ptep),
cmpxchg_relaxed   223 arch/arm64/mm/fault.c 		pteval = cmpxchg_relaxed(&pte_val(*ptep), old_pteval, pteval);
cmpxchg_relaxed   213 arch/mips/mm/context.c 	    !cmpxchg_relaxed(&cpu_data[cpu].asid_cache, old_active_mmid, ctx)) {
cmpxchg_relaxed   197 arch/powerpc/include/asm/atomic.h 	cmpxchg_relaxed(&((v)->counter), (o), (n))
cmpxchg_relaxed   489 arch/powerpc/include/asm/atomic.h 	cmpxchg_relaxed(&((v)->counter), (o), (n))
cmpxchg_relaxed   529 arch/powerpc/include/asm/cmpxchg.h 	cmpxchg_relaxed((ptr), (o), (n));				\
cmpxchg_relaxed   360 arch/riscv/include/asm/cmpxchg.h 	cmpxchg_relaxed((ptr), (o), (n))				\
cmpxchg_relaxed   372 arch/riscv/include/asm/cmpxchg.h 	cmpxchg_relaxed((ptr), (o), (n));				\
cmpxchg_relaxed  1337 drivers/iommu/arm-smmu-v3.c 		old = cmpxchg_relaxed(&cmdq->q.llq.val, llq.val, head.val);
cmpxchg_relaxed   467 drivers/iommu/io-pgtable-arm-v7s.c 	old = cmpxchg_relaxed(ptep, curr, new);
cmpxchg_relaxed   199 fs/erofs/zdata.c 		if (!cmpxchg_relaxed(pages, NULL, tagptr_cast_ptr(t)))
cmpxchg_relaxed   382 fs/erofs/zdata.c 		       length != cmpxchg_relaxed(&pcl->length, length, llen)) {
cmpxchg_relaxed    32 include/linux/atomic-fallback.h #ifndef cmpxchg_relaxed
cmpxchg_relaxed   380 kernel/locking/qspinlock_paravirt.h 	if (cmpxchg_relaxed(&pn->state, vcpu_halted, vcpu_hashed)
cmpxchg_relaxed   144 kernel/locking/rtmutex.c # define rt_mutex_cmpxchg_relaxed(l,c,n) (cmpxchg_relaxed(&l->owner, c, n) == c)
cmpxchg_relaxed   159 kernel/locking/rtmutex.c 	} while (cmpxchg_relaxed(p, owner,
cmpxchg_relaxed   426 kernel/sched/core.c 	if (unlikely(cmpxchg_relaxed(&node->next, NULL, WAKE_Q_TAIL)))