Searched refs:_Q_LOCKED_VAL (Results 1 – 5 of 5) sorted by relevance
/linux-4.4.14/kernel/locking/ |
D | qspinlock.c | 162 WRITE_ONCE(l->locked_pending, _Q_LOCKED_VAL); in clear_pending_set_locked() 192 atomic_add(-_Q_PENDING_VAL + _Q_LOCKED_VAL, &lock->val); in clear_pending_set_locked() 231 WRITE_ONCE(l->locked, _Q_LOCKED_VAL); in set_locked() 318 new = _Q_LOCKED_VAL; in queued_spin_lock_slowpath() 332 if (new == _Q_LOCKED_VAL) in queued_spin_lock_slowpath() 429 old = atomic_cmpxchg(&lock->val, val, _Q_LOCKED_VAL); in queued_spin_lock_slowpath()
|
D | qspinlock_paravirt.h | 283 if (!cmpxchg(&l->locked, _Q_LOCKED_VAL, _Q_SLOW_VAL)) { in pv_wait_head() 325 locked = cmpxchg(&l->locked, _Q_LOCKED_VAL, 0); in __pv_queued_spin_unlock() 326 if (likely(locked == _Q_LOCKED_VAL)) in __pv_queued_spin_unlock()
|
/linux-4.4.14/include/asm-generic/ |
D | qspinlock.h | 88 (atomic_cmpxchg(&lock->val, 0, _Q_LOCKED_VAL) == 0)) in queued_spin_trylock() 103 val = atomic_cmpxchg(&lock->val, 0, _Q_LOCKED_VAL); in queued_spin_lock() 120 atomic_sub(_Q_LOCKED_VAL, &lock->val); in queued_spin_unlock()
|
D | qspinlock_types.h | 76 #define _Q_LOCKED_VAL (1U << _Q_LOCKED_OFFSET) macro
|
/linux-4.4.14/arch/x86/include/asm/ |
D | qspinlock.h | 58 } while (atomic_cmpxchg(&lock->val, 0, _Q_LOCKED_VAL) != 0); in virt_spin_lock()
|