Searched refs:local_read (Results 1 – 14 of 14) sorted by relevance
28 #define local_read(l) atomic_long_read(&(l)->a) macro50 #define __local_inc(l) local_set((l), local_read(l) + 1)51 #define __local_dec(l) local_set((l), local_read(l) - 1)52 #define __local_add(i,l) local_set((l), local_read(l) + (i))53 #define __local_sub(i,l) local_set((l), local_read(l) - (i))
29 #define local64_read(l) local_read(&(l)->a)
42 handle->wakeup = local_read(&rb->wakeup); in perf_output_get_handle()51 head = local_read(&rb->head); in perf_output_put_handle()93 if (unlikely(head != local_read(&rb->head))) { in perf_output_put_handle()98 if (handle->wakeup != local_read(&rb->wakeup)) in perf_output_put_handle()134 have_lost = local_read(&rb->lost); in perf_output_begin()145 offset = head = local_read(&rb->head); in perf_output_begin()170 if (unlikely(head - local_read(&rb->wakeup) > rb->watermark)) in perf_output_begin()297 aux_head = local_read(&rb->aux_head); in perf_aux_output_begin()311 handle->wakeup = local_read(&rb->aux_wakeup) + rb->aux_watermark; in perf_aux_output_begin()364 aux_head = local_read(&rb->aux_head); in perf_aux_output_end()[all …]
146 index = local_read(&buf->head); in bts_config_buffer()182 if (local_read(&buf->data_size) >= bts->handle.size || in bts_buffer_is_full()183 bts->handle.size - local_read(&buf->data_size) < BTS_RECORD_SIZE) in bts_buffer_is_full()313 if (WARN_ON_ONCE(head != local_read(&buf->head))) in bts_buffer_reset()399 old_head = local_read(&buf->head); in intel_bts_interrupt()403 if (old_head == local_read(&buf->head)) in intel_bts_interrupt()
855 if (local_read(&buf->data_size) >= pt->handle.size) in pt_buffer_is_full()
13 #define local_read(l) atomic_long_read(&(l)->a) macro70 c = local_read(l); \
15 #define local_read(l) atomic_long_read(&(l)->a) macro139 c = local_read((l)); \
37 #define local_read(l) ((l)->counter) macro261 c = local_read(l); in local_add_unless()
386 return local_read(&((struct buffer_data_page *)page)->commit) in ring_buffer_page_len()1448 return local_read(&bpage->entries) & RB_WRITE_MASK; in rb_page_entries()1453 return local_read(&bpage->write) & RB_WRITE_MASK; in rb_page_write()1891 return local_read(&bpage->page->commit); in rb_page_commit()1964 local_read(&cpu_buffer->commit_page->page->commit) & in rb_set_commit_to_write()2493 local_read(&bpage->write) & ~RB_WRITE_MASK; in rb_try_to_discard()2526 !local_read(&cpu_buffer->committing))) in rb_end_commit()2530 commits = local_read(&cpu_buffer->commits); in rb_end_commit()2533 if (local_read(&cpu_buffer->committing) == 1) in rb_end_commit()2546 if (unlikely(local_read(&cpu_buffer->commits) != commits) && in rb_end_commit()[all …]
117 commit = local_read(&rpage->commit) & 0xfffff; in read_page()
18 #define local_read(l) atomic_long_read(&(l)->a) macro135 c = local_read(l); \
105 the data seen by local_read across CPUs must be considered to be out of order110 sum += local_read(&per_cpu(counters, cpu));112 If you want to use a remote local_read to synchronize access to a resource162 local_read(&per_cpu(counters, cpu)));
23 sum += local_read(pv); in fpuemu_stat_get()
14 #define local_read(l) atomic_long_read(&(l)->a) macro