LOCK_PREFIX 10 arch/x86/include/asm/alternative-asm.h .macro LOCK_PREFIX LOCK_PREFIX 18 arch/x86/include/asm/alternative-asm.h .macro LOCK_PREFIX LOCK_PREFIX 55 arch/x86/include/asm/atomic.h asm volatile(LOCK_PREFIX "addl %1,%0" LOCK_PREFIX 69 arch/x86/include/asm/atomic.h asm volatile(LOCK_PREFIX "subl %1,%0" LOCK_PREFIX 85 arch/x86/include/asm/atomic.h return GEN_BINARY_RMWcc(LOCK_PREFIX "subl", v->counter, e, "er", i); LOCK_PREFIX 97 arch/x86/include/asm/atomic.h asm volatile(LOCK_PREFIX "incl %0" LOCK_PREFIX 110 arch/x86/include/asm/atomic.h asm volatile(LOCK_PREFIX "decl %0" LOCK_PREFIX 125 arch/x86/include/asm/atomic.h return GEN_UNARY_RMWcc(LOCK_PREFIX "decl", v->counter, e); LOCK_PREFIX 139 arch/x86/include/asm/atomic.h return GEN_UNARY_RMWcc(LOCK_PREFIX "incl", v->counter, e); LOCK_PREFIX 154 arch/x86/include/asm/atomic.h return GEN_BINARY_RMWcc(LOCK_PREFIX "addl", v->counter, s, "er", i); LOCK_PREFIX 210 arch/x86/include/asm/atomic.h asm volatile(LOCK_PREFIX "andl %1,%0" LOCK_PREFIX 227 arch/x86/include/asm/atomic.h asm volatile(LOCK_PREFIX "orl %1,%0" LOCK_PREFIX 244 arch/x86/include/asm/atomic.h asm volatile(LOCK_PREFIX "xorl %1,%0" LOCK_PREFIX 46 arch/x86/include/asm/atomic64_64.h asm volatile(LOCK_PREFIX "addq %1,%0" LOCK_PREFIX 60 arch/x86/include/asm/atomic64_64.h asm volatile(LOCK_PREFIX "subq %1,%0" LOCK_PREFIX 76 arch/x86/include/asm/atomic64_64.h return GEN_BINARY_RMWcc(LOCK_PREFIX "subq", v->counter, e, "er", i); LOCK_PREFIX 88 arch/x86/include/asm/atomic64_64.h asm volatile(LOCK_PREFIX "incq %0" LOCK_PREFIX 102 arch/x86/include/asm/atomic64_64.h asm volatile(LOCK_PREFIX "decq %0" LOCK_PREFIX 118 arch/x86/include/asm/atomic64_64.h return GEN_UNARY_RMWcc(LOCK_PREFIX "decq", v->counter, e); LOCK_PREFIX 132 arch/x86/include/asm/atomic64_64.h return GEN_UNARY_RMWcc(LOCK_PREFIX "incq", v->counter, e); LOCK_PREFIX 147 arch/x86/include/asm/atomic64_64.h return GEN_BINARY_RMWcc(LOCK_PREFIX "addq", v->counter, s, "er", i); LOCK_PREFIX 196 arch/x86/include/asm/atomic64_64.h asm volatile(LOCK_PREFIX "andq %1,%0" LOCK_PREFIX 213 arch/x86/include/asm/atomic64_64.h asm volatile(LOCK_PREFIX "orq %1,%0" LOCK_PREFIX 230 arch/x86/include/asm/atomic64_64.h asm volatile(LOCK_PREFIX "xorq %1,%0" LOCK_PREFIX 55 arch/x86/include/asm/bitops.h asm volatile(LOCK_PREFIX "orb %1,%0" LOCK_PREFIX 60 arch/x86/include/asm/bitops.h asm volatile(LOCK_PREFIX __ASM_SIZE(bts) " %1,%0" LOCK_PREFIX 75 arch/x86/include/asm/bitops.h asm volatile(LOCK_PREFIX "andb %1,%0" LOCK_PREFIX 79 arch/x86/include/asm/bitops.h asm volatile(LOCK_PREFIX __ASM_SIZE(btr) " %1,%0" LOCK_PREFIX 101 arch/x86/include/asm/bitops.h asm volatile(LOCK_PREFIX "andb %2,%1" LOCK_PREFIX 126 arch/x86/include/asm/bitops.h asm volatile(LOCK_PREFIX "xorb %1,%0" LOCK_PREFIX 130 arch/x86/include/asm/bitops.h asm volatile(LOCK_PREFIX __ASM_SIZE(btc) " %1,%0" LOCK_PREFIX 138 arch/x86/include/asm/bitops.h return GEN_BINARY_RMWcc(LOCK_PREFIX __ASM_SIZE(bts), *addr, c, "Ir", nr); LOCK_PREFIX 162 arch/x86/include/asm/bitops.h return GEN_BINARY_RMWcc(LOCK_PREFIX __ASM_SIZE(btr), *addr, c, "Ir", nr); LOCK_PREFIX 201 arch/x86/include/asm/bitops.h return GEN_BINARY_RMWcc(LOCK_PREFIX __ASM_SIZE(btc), *addr, c, "Ir", nr); LOCK_PREFIX 134 arch/x86/include/asm/cmpxchg.h __raw_cmpxchg((ptr), (old), (new), (size), LOCK_PREFIX) LOCK_PREFIX 222 arch/x86/include/asm/cmpxchg.h __raw_try_cmpxchg((ptr), (pold), (new), (size), LOCK_PREFIX) LOCK_PREFIX 234 arch/x86/include/asm/cmpxchg.h #define xadd(ptr, inc) __xadd((ptr), (inc), LOCK_PREFIX) LOCK_PREFIX 256 arch/x86/include/asm/cmpxchg.h __cmpxchg_double(LOCK_PREFIX, p1, p2, o1, o2, n1, n2) LOCK_PREFIX 31 arch/x86/include/asm/cmpxchg_32.h LOCK_PREFIX "cmpxchg8b %0\n\t" LOCK_PREFIX 50 arch/x86/include/asm/cmpxchg_32.h asm volatile(LOCK_PREFIX "cmpxchg8b %1" LOCK_PREFIX 32 arch/x86/include/asm/futex.h "2:\t" LOCK_PREFIX "cmpxchgl %3, %2\n" \ LOCK_PREFIX 57 arch/x86/include/asm/futex.h __futex_atomic_op1(LOCK_PREFIX "xaddl %0, %2", ret, oldval, LOCK_PREFIX 23 arch/x86/include/asm/qspinlock.h val = GEN_BINARY_RMWcc(LOCK_PREFIX "btsl", lock->val.counter, c, LOCK_PREFIX 46 arch/x86/include/asm/qspinlock_paravirt.h LOCK_PREFIX "cmpxchg %dl,(%rdi);" LOCK_PREFIX 44 arch/x86/include/asm/refcount.h asm volatile(LOCK_PREFIX "addl %1,%0\n\t" LOCK_PREFIX 53 arch/x86/include/asm/refcount.h asm volatile(LOCK_PREFIX "incl %0\n\t" LOCK_PREFIX 61 arch/x86/include/asm/refcount.h asm volatile(LOCK_PREFIX "decl %0\n\t" LOCK_PREFIX 70 arch/x86/include/asm/refcount.h bool ret = GEN_BINARY_SUFFIXED_RMWcc(LOCK_PREFIX "subl", LOCK_PREFIX 84 arch/x86/include/asm/refcount.h bool ret = GEN_UNARY_SUFFIXED_RMWcc(LOCK_PREFIX "decl", LOCK_PREFIX 600 arch/x86/include/asm/uaccess.h "1:\t" LOCK_PREFIX "cmpxchgb %4, %2\n" \ LOCK_PREFIX 616 arch/x86/include/asm/uaccess.h "1:\t" LOCK_PREFIX "cmpxchgw %4, %2\n" \ LOCK_PREFIX 632 arch/x86/include/asm/uaccess.h "1:\t" LOCK_PREFIX "cmpxchgl %4, %2\n" \ LOCK_PREFIX 651 arch/x86/include/asm/uaccess.h "1:\t" LOCK_PREFIX "cmpxchgq %4, %2\n" \ LOCK_PREFIX 833 arch/x86/include/asm/uv/uv_bau.h asm volatile(LOCK_PREFIX "xaddw %0, %1" LOCK_PREFIX 51 tools/arch/x86/include/asm/atomic.h asm volatile(LOCK_PREFIX "incl %0" LOCK_PREFIX 65 tools/arch/x86/include/asm/atomic.h GEN_UNARY_RMWcc(LOCK_PREFIX "decl", v->counter, "%0", "e"); LOCK_PREFIX 84 tools/arch/x86/include/asm/cmpxchg.h __raw_cmpxchg((ptr), (old), (new), (size), LOCK_PREFIX)