Lines Matching refs:ring_buffer_per_cpu
457 struct ring_buffer_per_cpu { struct
504 struct ring_buffer_per_cpu **buffers; argument
515 struct ring_buffer_per_cpu *cpu_buffer;
552 struct ring_buffer_per_cpu *uninitialized_var(cpu_buffer); in ring_buffer_wait()
657 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_poll_wait()
698 if (__same_type(*(b), struct ring_buffer_per_cpu)) { \
699 struct ring_buffer_per_cpu *__b = \
836 rb_is_head_page(struct ring_buffer_per_cpu *cpu_buffer, in rb_is_head_page()
866 static void rb_set_list_to_head(struct ring_buffer_per_cpu *cpu_buffer, in rb_set_list_to_head()
879 static void rb_head_page_activate(struct ring_buffer_per_cpu *cpu_buffer) in rb_head_page_activate()
904 rb_head_page_deactivate(struct ring_buffer_per_cpu *cpu_buffer) in rb_head_page_deactivate()
915 static int rb_head_page_set(struct ring_buffer_per_cpu *cpu_buffer, in rb_head_page_set()
938 static int rb_head_page_set_update(struct ring_buffer_per_cpu *cpu_buffer, in rb_head_page_set_update()
947 static int rb_head_page_set_head(struct ring_buffer_per_cpu *cpu_buffer, in rb_head_page_set_head()
956 static int rb_head_page_set_normal(struct ring_buffer_per_cpu *cpu_buffer, in rb_head_page_set_normal()
965 static inline void rb_inc_page(struct ring_buffer_per_cpu *cpu_buffer, in rb_inc_page()
974 rb_set_head_page(struct ring_buffer_per_cpu *cpu_buffer) in rb_set_head_page()
1031 static int rb_tail_page_update(struct ring_buffer_per_cpu *cpu_buffer, in rb_tail_page_update()
1098 static int rb_check_bpage(struct ring_buffer_per_cpu *cpu_buffer, in rb_check_bpage()
1112 static int rb_check_list(struct ring_buffer_per_cpu *cpu_buffer, in rb_check_list()
1129 static int rb_check_pages(struct ring_buffer_per_cpu *cpu_buffer) in rb_check_pages()
1203 static int rb_allocate_pages(struct ring_buffer_per_cpu *cpu_buffer, in rb_allocate_pages()
1228 static struct ring_buffer_per_cpu *
1231 struct ring_buffer_per_cpu *cpu_buffer; in rb_allocate_cpu_buffer()
1289 static void rb_free_cpu_buffer(struct ring_buffer_per_cpu *cpu_buffer) in rb_free_cpu_buffer()
1444 static void rb_reset_cpu(struct ring_buffer_per_cpu *cpu_buffer);
1457 rb_remove_pages(struct ring_buffer_per_cpu *cpu_buffer, unsigned long nr_pages) in rb_remove_pages()
1567 rb_insert_pages(struct ring_buffer_per_cpu *cpu_buffer) in rb_insert_pages()
1643 static void rb_update_pages(struct ring_buffer_per_cpu *cpu_buffer) in rb_update_pages()
1659 struct ring_buffer_per_cpu *cpu_buffer = container_of(work, in update_pages_handler()
1660 struct ring_buffer_per_cpu, update_pages_work); in update_pages_handler()
1678 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_resize()
1877 rb_reader_event(struct ring_buffer_per_cpu *cpu_buffer) in rb_reader_event()
1901 rb_commit_index(struct ring_buffer_per_cpu *cpu_buffer) in rb_commit_index()
1915 rb_event_is_commit(struct ring_buffer_per_cpu *cpu_buffer, in rb_event_is_commit()
1929 rb_set_commit_to_write(struct ring_buffer_per_cpu *cpu_buffer) in rb_set_commit_to_write()
1981 static void rb_reset_reader_page(struct ring_buffer_per_cpu *cpu_buffer) in rb_reset_reader_page()
1989 struct ring_buffer_per_cpu *cpu_buffer = iter->cpu_buffer; in rb_inc_iter()
2037 rb_update_event(struct ring_buffer_per_cpu *cpu_buffer, in rb_update_event()
2072 rb_handle_head_page(struct ring_buffer_per_cpu *cpu_buffer, in rb_handle_head_page()
2240 rb_reset_tail(struct ring_buffer_per_cpu *cpu_buffer, in rb_reset_tail()
2313 rb_move_tail(struct ring_buffer_per_cpu *cpu_buffer, in rb_move_tail()
2419 __rb_reserve_next(struct ring_buffer_per_cpu *cpu_buffer, in __rb_reserve_next()
2476 rb_try_to_discard(struct ring_buffer_per_cpu *cpu_buffer, in rb_try_to_discard()
2515 static void rb_start_commit(struct ring_buffer_per_cpu *cpu_buffer) in rb_start_commit()
2521 static inline void rb_end_commit(struct ring_buffer_per_cpu *cpu_buffer) in rb_end_commit()
2555 struct ring_buffer_per_cpu *cpu_buffer, in rb_reserve_next_event()
2681 trace_recursive_lock(struct ring_buffer_per_cpu *cpu_buffer) in trace_recursive_lock()
2706 trace_recursive_unlock(struct ring_buffer_per_cpu *cpu_buffer) in trace_recursive_unlock()
2736 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_lock_reserve()
2780 rb_update_write_stamp(struct ring_buffer_per_cpu *cpu_buffer, in rb_update_write_stamp()
2807 static void rb_commit(struct ring_buffer_per_cpu *cpu_buffer, in rb_commit()
2816 rb_wakeups(struct ring_buffer *buffer, struct ring_buffer_per_cpu *cpu_buffer) in rb_wakeups()
2854 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_unlock_commit()
2891 rb_decrement_entry(struct ring_buffer_per_cpu *cpu_buffer, in rb_decrement_entry()
2946 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_discard_commit()
2998 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_write()
3045 static int rb_per_cpu_empty(struct ring_buffer_per_cpu *cpu_buffer) in rb_per_cpu_empty()
3158 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_record_disable_cpu()
3178 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_record_enable_cpu()
3195 rb_num_of_entries(struct ring_buffer_per_cpu *cpu_buffer) in rb_num_of_entries()
3209 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_oldest_event_ts()
3241 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_bytes_cpu()
3261 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_entries_cpu()
3280 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_overrun_cpu()
3303 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_commit_overrun_cpu()
3325 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_dropped_events_cpu()
3346 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_read_events_cpu()
3365 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_entries()
3388 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_overruns()
3404 struct ring_buffer_per_cpu *cpu_buffer = iter->cpu_buffer; in rb_iter_reset()
3428 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_iter_reset()
3448 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_iter_empty()
3458 rb_update_read_stamp(struct ring_buffer_per_cpu *cpu_buffer, in rb_update_read_stamp()
3520 rb_get_reader_page(struct ring_buffer_per_cpu *cpu_buffer) in rb_get_reader_page()
3648 static void rb_advance_reader(struct ring_buffer_per_cpu *cpu_buffer) in rb_advance_reader()
3673 struct ring_buffer_per_cpu *cpu_buffer; in rb_advance_iter()
3713 static int rb_lost_events(struct ring_buffer_per_cpu *cpu_buffer) in rb_lost_events()
3719 rb_buffer_peek(struct ring_buffer_per_cpu *cpu_buffer, u64 *ts, in rb_buffer_peek()
3788 struct ring_buffer_per_cpu *cpu_buffer; in rb_iter_peek()
3893 struct ring_buffer_per_cpu *cpu_buffer = buffer->buffers[cpu]; in ring_buffer_peek()
3930 struct ring_buffer_per_cpu *cpu_buffer = iter->cpu_buffer; in ring_buffer_iter_peek()
3960 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_consume()
4022 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_read_prepare()
4071 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_read_start()
4097 struct ring_buffer_per_cpu *cpu_buffer = iter->cpu_buffer; in ring_buffer_read_finish()
4127 struct ring_buffer_per_cpu *cpu_buffer = iter->cpu_buffer; in ring_buffer_read()
4167 rb_reset_cpu(struct ring_buffer_per_cpu *cpu_buffer) in rb_reset_cpu()
4215 struct ring_buffer_per_cpu *cpu_buffer = buffer->buffers[cpu]; in ring_buffer_reset_cpu()
4265 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_empty()
4299 struct ring_buffer_per_cpu *cpu_buffer; in ring_buffer_empty_cpu()
4336 struct ring_buffer_per_cpu *cpu_buffer_a; in ring_buffer_swap_cpu()
4337 struct ring_buffer_per_cpu *cpu_buffer_b; in ring_buffer_swap_cpu()
4483 struct ring_buffer_per_cpu *cpu_buffer = buffer->buffers[cpu]; in ring_buffer_read_page()