Lines Matching refs:v

24 #define atomic_read(v)		READ_ONCE((v)->counter)  argument
25 #define atomic64_read(v) READ_ONCE((v)->counter) argument
27 #define atomic_set(v,i) WRITE_ONCE(((v)->counter), (i)) argument
28 #define atomic64_set(v,i) WRITE_ONCE(((v)->counter), (i)) argument
32 ia64_atomic_##op (int i, atomic_t *v) \
38 CMPXCHG_BUGCHECK(v); \
39 old = atomic_read(v); \
41 } while (ia64_cmpxchg(acq, v, old, new, sizeof(atomic_t)) != old); \
48 #define atomic_add_return(i,v) \ argument
56 ? ia64_fetch_and_add(__ia64_aar_i, &(v)->counter) \
57 : ia64_atomic_add(__ia64_aar_i, v); \
60 #define atomic_sub_return(i,v) \ argument
68 ? ia64_fetch_and_add(-__ia64_asr_i, &(v)->counter) \
69 : ia64_atomic_sub(__ia64_asr_i, v); \
76 #define atomic_and(i,v) (void)ia64_atomic_and(i,v) argument
77 #define atomic_or(i,v) (void)ia64_atomic_or(i,v) argument
78 #define atomic_xor(i,v) (void)ia64_atomic_xor(i,v) argument
84 ia64_atomic64_##op (__s64 i, atomic64_t *v) \
90 CMPXCHG_BUGCHECK(v); \
91 old = atomic64_read(v); \
93 } while (ia64_cmpxchg(acq, v, old, new, sizeof(atomic64_t)) != old); \
100 #define atomic64_add_return(i,v) \ argument
108 ? ia64_fetch_and_add(__ia64_aar_i, &(v)->counter) \
109 : ia64_atomic64_add(__ia64_aar_i, v); \
112 #define atomic64_sub_return(i,v) \ argument
120 ? ia64_fetch_and_add(-__ia64_asr_i, &(v)->counter) \
121 : ia64_atomic64_sub(__ia64_asr_i, v); \
128 #define atomic64_and(i,v) (void)ia64_atomic64_and(i,v) argument
129 #define atomic64_or(i,v) (void)ia64_atomic64_or(i,v) argument
130 #define atomic64_xor(i,v) (void)ia64_atomic64_xor(i,v) argument
134 #define atomic_cmpxchg(v, old, new) (cmpxchg(&((v)->counter), old, new)) argument
135 #define atomic_xchg(v, new) (xchg(&((v)->counter), new)) argument
137 #define atomic64_cmpxchg(v, old, new) \ argument
138 (cmpxchg(&((v)->counter), old, new))
139 #define atomic64_xchg(v, new) (xchg(&((v)->counter), new)) argument
141 static __inline__ int __atomic_add_unless(atomic_t *v, int a, int u) in __atomic_add_unless() argument
144 c = atomic_read(v); in __atomic_add_unless()
148 old = atomic_cmpxchg((v), c, c + (a)); in __atomic_add_unless()
157 static __inline__ long atomic64_add_unless(atomic64_t *v, long a, long u) in atomic64_add_unless() argument
160 c = atomic64_read(v); in atomic64_add_unless()
164 old = atomic64_cmpxchg((v), c, c + (a)); in atomic64_add_unless()
172 #define atomic64_inc_not_zero(v) atomic64_add_unless((v), 1, 0) argument
179 atomic_add_negative (int i, atomic_t *v) in atomic_add_negative() argument
181 return atomic_add_return(i, v) < 0; in atomic_add_negative()
185 atomic64_add_negative (__s64 i, atomic64_t *v) in atomic64_add_negative() argument
187 return atomic64_add_return(i, v) < 0; in atomic64_add_negative()
190 #define atomic_dec_return(v) atomic_sub_return(1, (v)) argument
191 #define atomic_inc_return(v) atomic_add_return(1, (v)) argument
192 #define atomic64_dec_return(v) atomic64_sub_return(1, (v)) argument
193 #define atomic64_inc_return(v) atomic64_add_return(1, (v)) argument
195 #define atomic_sub_and_test(i,v) (atomic_sub_return((i), (v)) == 0) argument
196 #define atomic_dec_and_test(v) (atomic_sub_return(1, (v)) == 0) argument
197 #define atomic_inc_and_test(v) (atomic_add_return(1, (v)) == 0) argument
198 #define atomic64_sub_and_test(i,v) (atomic64_sub_return((i), (v)) == 0) argument
199 #define atomic64_dec_and_test(v) (atomic64_sub_return(1, (v)) == 0) argument
200 #define atomic64_inc_and_test(v) (atomic64_add_return(1, (v)) == 0) argument
202 #define atomic_add(i,v) (void)atomic_add_return((i), (v)) argument
203 #define atomic_sub(i,v) (void)atomic_sub_return((i), (v)) argument
204 #define atomic_inc(v) atomic_add(1, (v)) argument
205 #define atomic_dec(v) atomic_sub(1, (v)) argument
207 #define atomic64_add(i,v) (void)atomic64_add_return((i), (v)) argument
208 #define atomic64_sub(i,v) (void)atomic64_sub_return((i), (v)) argument
209 #define atomic64_inc(v) atomic64_add(1, (v)) argument
210 #define atomic64_dec(v) atomic64_sub(1, (v)) argument