LOCK_PREFIX        10 arch/x86/include/asm/alternative-asm.h 	.macro LOCK_PREFIX
LOCK_PREFIX        18 arch/x86/include/asm/alternative-asm.h 	.macro LOCK_PREFIX
LOCK_PREFIX        55 arch/x86/include/asm/atomic.h 	asm volatile(LOCK_PREFIX "addl %1,%0"
LOCK_PREFIX        69 arch/x86/include/asm/atomic.h 	asm volatile(LOCK_PREFIX "subl %1,%0"
LOCK_PREFIX        85 arch/x86/include/asm/atomic.h 	return GEN_BINARY_RMWcc(LOCK_PREFIX "subl", v->counter, e, "er", i);
LOCK_PREFIX        97 arch/x86/include/asm/atomic.h 	asm volatile(LOCK_PREFIX "incl %0"
LOCK_PREFIX       110 arch/x86/include/asm/atomic.h 	asm volatile(LOCK_PREFIX "decl %0"
LOCK_PREFIX       125 arch/x86/include/asm/atomic.h 	return GEN_UNARY_RMWcc(LOCK_PREFIX "decl", v->counter, e);
LOCK_PREFIX       139 arch/x86/include/asm/atomic.h 	return GEN_UNARY_RMWcc(LOCK_PREFIX "incl", v->counter, e);
LOCK_PREFIX       154 arch/x86/include/asm/atomic.h 	return GEN_BINARY_RMWcc(LOCK_PREFIX "addl", v->counter, s, "er", i);
LOCK_PREFIX       210 arch/x86/include/asm/atomic.h 	asm volatile(LOCK_PREFIX "andl %1,%0"
LOCK_PREFIX       227 arch/x86/include/asm/atomic.h 	asm volatile(LOCK_PREFIX "orl %1,%0"
LOCK_PREFIX       244 arch/x86/include/asm/atomic.h 	asm volatile(LOCK_PREFIX "xorl %1,%0"
LOCK_PREFIX        46 arch/x86/include/asm/atomic64_64.h 	asm volatile(LOCK_PREFIX "addq %1,%0"
LOCK_PREFIX        60 arch/x86/include/asm/atomic64_64.h 	asm volatile(LOCK_PREFIX "subq %1,%0"
LOCK_PREFIX        76 arch/x86/include/asm/atomic64_64.h 	return GEN_BINARY_RMWcc(LOCK_PREFIX "subq", v->counter, e, "er", i);
LOCK_PREFIX        88 arch/x86/include/asm/atomic64_64.h 	asm volatile(LOCK_PREFIX "incq %0"
LOCK_PREFIX       102 arch/x86/include/asm/atomic64_64.h 	asm volatile(LOCK_PREFIX "decq %0"
LOCK_PREFIX       118 arch/x86/include/asm/atomic64_64.h 	return GEN_UNARY_RMWcc(LOCK_PREFIX "decq", v->counter, e);
LOCK_PREFIX       132 arch/x86/include/asm/atomic64_64.h 	return GEN_UNARY_RMWcc(LOCK_PREFIX "incq", v->counter, e);
LOCK_PREFIX       147 arch/x86/include/asm/atomic64_64.h 	return GEN_BINARY_RMWcc(LOCK_PREFIX "addq", v->counter, s, "er", i);
LOCK_PREFIX       196 arch/x86/include/asm/atomic64_64.h 	asm volatile(LOCK_PREFIX "andq %1,%0"
LOCK_PREFIX       213 arch/x86/include/asm/atomic64_64.h 	asm volatile(LOCK_PREFIX "orq %1,%0"
LOCK_PREFIX       230 arch/x86/include/asm/atomic64_64.h 	asm volatile(LOCK_PREFIX "xorq %1,%0"
LOCK_PREFIX        55 arch/x86/include/asm/bitops.h 		asm volatile(LOCK_PREFIX "orb %1,%0"
LOCK_PREFIX        60 arch/x86/include/asm/bitops.h 		asm volatile(LOCK_PREFIX __ASM_SIZE(bts) " %1,%0"
LOCK_PREFIX        75 arch/x86/include/asm/bitops.h 		asm volatile(LOCK_PREFIX "andb %1,%0"
LOCK_PREFIX        79 arch/x86/include/asm/bitops.h 		asm volatile(LOCK_PREFIX __ASM_SIZE(btr) " %1,%0"
LOCK_PREFIX       101 arch/x86/include/asm/bitops.h 	asm volatile(LOCK_PREFIX "andb %2,%1"
LOCK_PREFIX       126 arch/x86/include/asm/bitops.h 		asm volatile(LOCK_PREFIX "xorb %1,%0"
LOCK_PREFIX       130 arch/x86/include/asm/bitops.h 		asm volatile(LOCK_PREFIX __ASM_SIZE(btc) " %1,%0"
LOCK_PREFIX       138 arch/x86/include/asm/bitops.h 	return GEN_BINARY_RMWcc(LOCK_PREFIX __ASM_SIZE(bts), *addr, c, "Ir", nr);
LOCK_PREFIX       162 arch/x86/include/asm/bitops.h 	return GEN_BINARY_RMWcc(LOCK_PREFIX __ASM_SIZE(btr), *addr, c, "Ir", nr);
LOCK_PREFIX       201 arch/x86/include/asm/bitops.h 	return GEN_BINARY_RMWcc(LOCK_PREFIX __ASM_SIZE(btc), *addr, c, "Ir", nr);
LOCK_PREFIX       134 arch/x86/include/asm/cmpxchg.h 	__raw_cmpxchg((ptr), (old), (new), (size), LOCK_PREFIX)
LOCK_PREFIX       222 arch/x86/include/asm/cmpxchg.h 	__raw_try_cmpxchg((ptr), (pold), (new), (size), LOCK_PREFIX)
LOCK_PREFIX       234 arch/x86/include/asm/cmpxchg.h #define xadd(ptr, inc)		__xadd((ptr), (inc), LOCK_PREFIX)
LOCK_PREFIX       256 arch/x86/include/asm/cmpxchg.h 	__cmpxchg_double(LOCK_PREFIX, p1, p2, o1, o2, n1, n2)
LOCK_PREFIX        31 arch/x86/include/asm/cmpxchg_32.h 		     LOCK_PREFIX "cmpxchg8b %0\n\t"
LOCK_PREFIX        50 arch/x86/include/asm/cmpxchg_32.h 	asm volatile(LOCK_PREFIX "cmpxchg8b %1"
LOCK_PREFIX        32 arch/x86/include/asm/futex.h 		     "2:\t" LOCK_PREFIX "cmpxchgl %3, %2\n"	\
LOCK_PREFIX        57 arch/x86/include/asm/futex.h 		__futex_atomic_op1(LOCK_PREFIX "xaddl %0, %2", ret, oldval,
LOCK_PREFIX        23 arch/x86/include/asm/qspinlock.h 	val = GEN_BINARY_RMWcc(LOCK_PREFIX "btsl", lock->val.counter, c,
LOCK_PREFIX        46 arch/x86/include/asm/qspinlock_paravirt.h 	LOCK_PREFIX "cmpxchg %dl,(%rdi);"
LOCK_PREFIX        44 arch/x86/include/asm/refcount.h 	asm volatile(LOCK_PREFIX "addl %1,%0\n\t"
LOCK_PREFIX        53 arch/x86/include/asm/refcount.h 	asm volatile(LOCK_PREFIX "incl %0\n\t"
LOCK_PREFIX        61 arch/x86/include/asm/refcount.h 	asm volatile(LOCK_PREFIX "decl %0\n\t"
LOCK_PREFIX        70 arch/x86/include/asm/refcount.h 	bool ret = GEN_BINARY_SUFFIXED_RMWcc(LOCK_PREFIX "subl",
LOCK_PREFIX        84 arch/x86/include/asm/refcount.h 	bool ret = GEN_UNARY_SUFFIXED_RMWcc(LOCK_PREFIX "decl",
LOCK_PREFIX       600 arch/x86/include/asm/uaccess.h 			"1:\t" LOCK_PREFIX "cmpxchgb %4, %2\n"		\
LOCK_PREFIX       616 arch/x86/include/asm/uaccess.h 			"1:\t" LOCK_PREFIX "cmpxchgw %4, %2\n"		\
LOCK_PREFIX       632 arch/x86/include/asm/uaccess.h 			"1:\t" LOCK_PREFIX "cmpxchgl %4, %2\n"		\
LOCK_PREFIX       651 arch/x86/include/asm/uaccess.h 			"1:\t" LOCK_PREFIX "cmpxchgq %4, %2\n"		\
LOCK_PREFIX       833 arch/x86/include/asm/uv/uv_bau.h 	asm volatile(LOCK_PREFIX "xaddw %0, %1"
LOCK_PREFIX        51 tools/arch/x86/include/asm/atomic.h 	asm volatile(LOCK_PREFIX "incl %0"
LOCK_PREFIX        65 tools/arch/x86/include/asm/atomic.h 	GEN_UNARY_RMWcc(LOCK_PREFIX "decl", v->counter, "%0", "e");
LOCK_PREFIX        84 tools/arch/x86/include/asm/cmpxchg.h 	__raw_cmpxchg((ptr), (old), (new), (size), LOCK_PREFIX)