Lines Matching refs:lock
42 #define __xchg_op(ptr, arg, op, lock) \ argument
47 asm volatile (lock #op "b %b0, %1\n" \
52 asm volatile (lock #op "w %w0, %1\n" \
57 asm volatile (lock #op "l %0, %1\n" \
62 asm volatile (lock #op "q %q0, %1\n" \
85 #define __raw_cmpxchg(ptr, old, new, size, lock) \ argument
94 asm volatile(lock "cmpxchgb %2,%1" \
103 asm volatile(lock "cmpxchgw %2,%1" \
112 asm volatile(lock "cmpxchgl %2,%1" \
121 asm volatile(lock "cmpxchgq %2,%1" \
165 #define __xadd(ptr, inc, lock) __xchg_op((ptr), (inc), xadd, lock) argument
170 #define __add(ptr, inc, lock) \ argument
175 asm volatile (lock "addb %b1, %0\n" \
180 asm volatile (lock "addw %w1, %0\n" \
185 asm volatile (lock "addl %1, %0\n" \
190 asm volatile (lock "addq %1, %0\n" \