Lines Matching refs:lock
40 #define __xchg_op(ptr, arg, op, lock) \ argument
45 asm volatile (lock #op "b %b0, %1\n" \
50 asm volatile (lock #op "w %w0, %1\n" \
55 asm volatile (lock #op "l %0, %1\n" \
60 asm volatile (lock #op "q %q0, %1\n" \
83 #define __raw_cmpxchg(ptr, old, new, size, lock) \ argument
92 asm volatile(lock "cmpxchgb %2,%1" \
101 asm volatile(lock "cmpxchgw %2,%1" \
110 asm volatile(lock "cmpxchgl %2,%1" \
119 asm volatile(lock "cmpxchgq %2,%1" \
163 #define __xadd(ptr, inc, lock) __xchg_op((ptr), (inc), xadd, lock) argument
168 #define __add(ptr, inc, lock) \ argument
173 asm volatile (lock "addb %b1, %0\n" \
178 asm volatile (lock "addw %w1, %0\n" \
183 asm volatile (lock "addl %1, %0\n" \
188 asm volatile (lock "addq %1, %0\n" \