bpage 343 kernel/trace/ring_buffer.c static void rb_init_page(struct buffer_data_page *bpage) bpage 345 kernel/trace/ring_buffer.c local_set(&bpage->commit, 0); bpage 352 kernel/trace/ring_buffer.c static void free_buffer_page(struct buffer_page *bpage) bpage 354 kernel/trace/ring_buffer.c free_page((unsigned long)bpage->page); bpage 355 kernel/trace/ring_buffer.c kfree(bpage); bpage 999 kernel/trace/ring_buffer.c struct buffer_page **bpage) bpage 1001 kernel/trace/ring_buffer.c struct list_head *p = rb_list_head((*bpage)->list.next); bpage 1003 kernel/trace/ring_buffer.c *bpage = list_entry(p, struct buffer_page, list); bpage 1124 kernel/trace/ring_buffer.c struct buffer_page *bpage) bpage 1126 kernel/trace/ring_buffer.c unsigned long val = (unsigned long)bpage; bpage 1157 kernel/trace/ring_buffer.c struct buffer_page *bpage, *tmp; bpage 1173 kernel/trace/ring_buffer.c list_for_each_entry_safe(bpage, tmp, head, list) { bpage 1175 kernel/trace/ring_buffer.c bpage->list.next->prev != &bpage->list)) bpage 1178 kernel/trace/ring_buffer.c bpage->list.prev->next != &bpage->list)) bpage 1180 kernel/trace/ring_buffer.c if (rb_check_list(cpu_buffer, &bpage->list)) bpage 1191 kernel/trace/ring_buffer.c struct buffer_page *bpage, *tmp; bpage 1228 kernel/trace/ring_buffer.c bpage = kzalloc_node(ALIGN(sizeof(*bpage), cache_line_size()), bpage 1230 kernel/trace/ring_buffer.c if (!bpage) bpage 1233 kernel/trace/ring_buffer.c list_add(&bpage->list, pages); bpage 1238 kernel/trace/ring_buffer.c bpage->page = page_address(page); bpage 1239 kernel/trace/ring_buffer.c rb_init_page(bpage->page); bpage 1250 kernel/trace/ring_buffer.c list_for_each_entry_safe(bpage, tmp, pages, list) { bpage 1251 kernel/trace/ring_buffer.c list_del_init(&bpage->list); bpage 1252 kernel/trace/ring_buffer.c free_buffer_page(bpage); bpage 1289 kernel/trace/ring_buffer.c struct buffer_page *bpage; bpage 1309 kernel/trace/ring_buffer.c bpage = kzalloc_node(ALIGN(sizeof(*bpage), cache_line_size()), bpage 1311 kernel/trace/ring_buffer.c if (!bpage) bpage 1314 kernel/trace/ring_buffer.c rb_check_bpage(cpu_buffer, bpage); bpage 1316 kernel/trace/ring_buffer.c cpu_buffer->reader_page = bpage; bpage 1320 kernel/trace/ring_buffer.c bpage->page = page_address(page); bpage 1321 kernel/trace/ring_buffer.c rb_init_page(bpage->page); bpage 1349 kernel/trace/ring_buffer.c struct buffer_page *bpage, *tmp; bpage 1356 kernel/trace/ring_buffer.c list_for_each_entry_safe(bpage, tmp, head, list) { bpage 1357 kernel/trace/ring_buffer.c list_del_init(&bpage->list); bpage 1358 kernel/trace/ring_buffer.c free_buffer_page(bpage); bpage 1360 kernel/trace/ring_buffer.c bpage = list_entry(head, struct buffer_page, list); bpage 1361 kernel/trace/ring_buffer.c free_buffer_page(bpage); bpage 1484 kernel/trace/ring_buffer.c static inline unsigned long rb_page_entries(struct buffer_page *bpage) bpage 1486 kernel/trace/ring_buffer.c return local_read(&bpage->entries) & RB_WRITE_MASK; bpage 1489 kernel/trace/ring_buffer.c static inline unsigned long rb_page_write(struct buffer_page *bpage) bpage 1491 kernel/trace/ring_buffer.c return local_read(&bpage->write) & RB_WRITE_MASK; bpage 1673 kernel/trace/ring_buffer.c struct buffer_page *bpage, *tmp; bpage 1674 kernel/trace/ring_buffer.c list_for_each_entry_safe(bpage, tmp, &cpu_buffer->new_pages, bpage 1676 kernel/trace/ring_buffer.c list_del_init(&bpage->list); bpage 1677 kernel/trace/ring_buffer.c free_buffer_page(bpage); bpage 1875 kernel/trace/ring_buffer.c struct buffer_page *bpage, *tmp; bpage 1883 kernel/trace/ring_buffer.c list_for_each_entry_safe(bpage, tmp, &cpu_buffer->new_pages, bpage 1885 kernel/trace/ring_buffer.c list_del_init(&bpage->list); bpage 1886 kernel/trace/ring_buffer.c free_buffer_page(bpage); bpage 1905 kernel/trace/ring_buffer.c static __always_inline void *__rb_page_index(struct buffer_page *bpage, unsigned index) bpage 1907 kernel/trace/ring_buffer.c return bpage->page->data + index; bpage 1923 kernel/trace/ring_buffer.c static __always_inline unsigned rb_page_commit(struct buffer_page *bpage) bpage 1925 kernel/trace/ring_buffer.c return local_read(&bpage->page->commit); bpage 1929 kernel/trace/ring_buffer.c static __always_inline unsigned rb_page_size(struct buffer_page *bpage) bpage 1931 kernel/trace/ring_buffer.c return rb_page_commit(bpage); bpage 2419 kernel/trace/ring_buffer.c struct buffer_page *bpage; bpage 2428 kernel/trace/ring_buffer.c bpage = READ_ONCE(cpu_buffer->tail_page); bpage 2430 kernel/trace/ring_buffer.c if (bpage->page == (void *)addr && rb_page_write(bpage) == old_index) { bpage 2432 kernel/trace/ring_buffer.c local_read(&bpage->write) & ~RB_WRITE_MASK; bpage 2442 kernel/trace/ring_buffer.c index = local_cmpxchg(&bpage->write, old_index, new_index); bpage 3017 kernel/trace/ring_buffer.c struct buffer_page *bpage = cpu_buffer->commit_page; bpage 3023 kernel/trace/ring_buffer.c if (likely(bpage->page == (void *)addr)) { bpage 3024 kernel/trace/ring_buffer.c local_dec(&bpage->entries); bpage 3032 kernel/trace/ring_buffer.c rb_inc_page(cpu_buffer, &bpage); bpage 3033 kernel/trace/ring_buffer.c start = bpage; bpage 3035 kernel/trace/ring_buffer.c if (bpage->page == (void *)addr) { bpage 3036 kernel/trace/ring_buffer.c local_dec(&bpage->entries); bpage 3039 kernel/trace/ring_buffer.c rb_inc_page(cpu_buffer, &bpage); bpage 3040 kernel/trace/ring_buffer.c } while (bpage != start); bpage 3352 kernel/trace/ring_buffer.c struct buffer_page *bpage; bpage 3365 kernel/trace/ring_buffer.c bpage = cpu_buffer->reader_page; bpage 3367 kernel/trace/ring_buffer.c bpage = rb_set_head_page(cpu_buffer); bpage 3368 kernel/trace/ring_buffer.c if (bpage) bpage 3369 kernel/trace/ring_buffer.c ret = bpage->page->time_stamp; bpage 4596 kernel/trace/ring_buffer.c struct buffer_data_page *bpage = NULL; bpage 4608 kernel/trace/ring_buffer.c bpage = cpu_buffer->free_page; bpage 4615 kernel/trace/ring_buffer.c if (bpage) bpage 4623 kernel/trace/ring_buffer.c bpage = page_address(page); bpage 4626 kernel/trace/ring_buffer.c rb_init_page(bpage); bpage 4628 kernel/trace/ring_buffer.c return bpage; bpage 4643 kernel/trace/ring_buffer.c struct buffer_data_page *bpage = data; bpage 4644 kernel/trace/ring_buffer.c struct page *page = virt_to_page(bpage); bpage 4655 kernel/trace/ring_buffer.c cpu_buffer->free_page = bpage; bpage 4656 kernel/trace/ring_buffer.c bpage = NULL; bpage 4663 kernel/trace/ring_buffer.c free_page((unsigned long)bpage); bpage 4705 kernel/trace/ring_buffer.c struct buffer_data_page *bpage; bpage 4729 kernel/trace/ring_buffer.c bpage = *data_page; bpage 4730 kernel/trace/ring_buffer.c if (!bpage) bpage 4785 kernel/trace/ring_buffer.c memcpy(bpage->data + pos, rpage->data + rpos, size); bpage 4802 kernel/trace/ring_buffer.c local_set(&bpage->commit, pos); bpage 4803 kernel/trace/ring_buffer.c bpage->time_stamp = save_timestamp; bpage 4813 kernel/trace/ring_buffer.c rb_init_page(bpage); bpage 4814 kernel/trace/ring_buffer.c bpage = reader->page; bpage 4819 kernel/trace/ring_buffer.c *data_page = bpage; bpage 4827 kernel/trace/ring_buffer.c local_set(&bpage->commit, reader->real_end); bpage 4833 kernel/trace/ring_buffer.c commit = local_read(&bpage->commit); bpage 4842 kernel/trace/ring_buffer.c memcpy(&bpage->data[commit], &missed_events, bpage 4844 kernel/trace/ring_buffer.c local_add(RB_MISSED_STORED, &bpage->commit); bpage 4847 kernel/trace/ring_buffer.c local_add(RB_MISSED_EVENTS, &bpage->commit); bpage 4854 kernel/trace/ring_buffer.c memset(&bpage->data[commit], 0, BUF_PAGE_SIZE - commit); bpage 110 kernel/trace/ring_buffer_benchmark.c void *bpage; bpage 116 kernel/trace/ring_buffer_benchmark.c bpage = ring_buffer_alloc_read_page(buffer, cpu); bpage 117 kernel/trace/ring_buffer_benchmark.c if (IS_ERR(bpage)) bpage 120 kernel/trace/ring_buffer_benchmark.c ret = ring_buffer_read_page(buffer, &bpage, PAGE_SIZE, cpu, 1); bpage 122 kernel/trace/ring_buffer_benchmark.c rpage = bpage; bpage 175 kernel/trace/ring_buffer_benchmark.c ring_buffer_free_read_page(buffer, cpu, bpage);