bpage             343 kernel/trace/ring_buffer.c static void rb_init_page(struct buffer_data_page *bpage)
bpage             345 kernel/trace/ring_buffer.c 	local_set(&bpage->commit, 0);
bpage             352 kernel/trace/ring_buffer.c static void free_buffer_page(struct buffer_page *bpage)
bpage             354 kernel/trace/ring_buffer.c 	free_page((unsigned long)bpage->page);
bpage             355 kernel/trace/ring_buffer.c 	kfree(bpage);
bpage             999 kernel/trace/ring_buffer.c 			       struct buffer_page **bpage)
bpage            1001 kernel/trace/ring_buffer.c 	struct list_head *p = rb_list_head((*bpage)->list.next);
bpage            1003 kernel/trace/ring_buffer.c 	*bpage = list_entry(p, struct buffer_page, list);
bpage            1124 kernel/trace/ring_buffer.c 			  struct buffer_page *bpage)
bpage            1126 kernel/trace/ring_buffer.c 	unsigned long val = (unsigned long)bpage;
bpage            1157 kernel/trace/ring_buffer.c 	struct buffer_page *bpage, *tmp;
bpage            1173 kernel/trace/ring_buffer.c 	list_for_each_entry_safe(bpage, tmp, head, list) {
bpage            1175 kernel/trace/ring_buffer.c 			       bpage->list.next->prev != &bpage->list))
bpage            1178 kernel/trace/ring_buffer.c 			       bpage->list.prev->next != &bpage->list))
bpage            1180 kernel/trace/ring_buffer.c 		if (rb_check_list(cpu_buffer, &bpage->list))
bpage            1191 kernel/trace/ring_buffer.c 	struct buffer_page *bpage, *tmp;
bpage            1228 kernel/trace/ring_buffer.c 		bpage = kzalloc_node(ALIGN(sizeof(*bpage), cache_line_size()),
bpage            1230 kernel/trace/ring_buffer.c 		if (!bpage)
bpage            1233 kernel/trace/ring_buffer.c 		list_add(&bpage->list, pages);
bpage            1238 kernel/trace/ring_buffer.c 		bpage->page = page_address(page);
bpage            1239 kernel/trace/ring_buffer.c 		rb_init_page(bpage->page);
bpage            1250 kernel/trace/ring_buffer.c 	list_for_each_entry_safe(bpage, tmp, pages, list) {
bpage            1251 kernel/trace/ring_buffer.c 		list_del_init(&bpage->list);
bpage            1252 kernel/trace/ring_buffer.c 		free_buffer_page(bpage);
bpage            1289 kernel/trace/ring_buffer.c 	struct buffer_page *bpage;
bpage            1309 kernel/trace/ring_buffer.c 	bpage = kzalloc_node(ALIGN(sizeof(*bpage), cache_line_size()),
bpage            1311 kernel/trace/ring_buffer.c 	if (!bpage)
bpage            1314 kernel/trace/ring_buffer.c 	rb_check_bpage(cpu_buffer, bpage);
bpage            1316 kernel/trace/ring_buffer.c 	cpu_buffer->reader_page = bpage;
bpage            1320 kernel/trace/ring_buffer.c 	bpage->page = page_address(page);
bpage            1321 kernel/trace/ring_buffer.c 	rb_init_page(bpage->page);
bpage            1349 kernel/trace/ring_buffer.c 	struct buffer_page *bpage, *tmp;
bpage            1356 kernel/trace/ring_buffer.c 		list_for_each_entry_safe(bpage, tmp, head, list) {
bpage            1357 kernel/trace/ring_buffer.c 			list_del_init(&bpage->list);
bpage            1358 kernel/trace/ring_buffer.c 			free_buffer_page(bpage);
bpage            1360 kernel/trace/ring_buffer.c 		bpage = list_entry(head, struct buffer_page, list);
bpage            1361 kernel/trace/ring_buffer.c 		free_buffer_page(bpage);
bpage            1484 kernel/trace/ring_buffer.c static inline unsigned long rb_page_entries(struct buffer_page *bpage)
bpage            1486 kernel/trace/ring_buffer.c 	return local_read(&bpage->entries) & RB_WRITE_MASK;
bpage            1489 kernel/trace/ring_buffer.c static inline unsigned long rb_page_write(struct buffer_page *bpage)
bpage            1491 kernel/trace/ring_buffer.c 	return local_read(&bpage->write) & RB_WRITE_MASK;
bpage            1673 kernel/trace/ring_buffer.c 		struct buffer_page *bpage, *tmp;
bpage            1674 kernel/trace/ring_buffer.c 		list_for_each_entry_safe(bpage, tmp, &cpu_buffer->new_pages,
bpage            1676 kernel/trace/ring_buffer.c 			list_del_init(&bpage->list);
bpage            1677 kernel/trace/ring_buffer.c 			free_buffer_page(bpage);
bpage            1875 kernel/trace/ring_buffer.c 		struct buffer_page *bpage, *tmp;
bpage            1883 kernel/trace/ring_buffer.c 		list_for_each_entry_safe(bpage, tmp, &cpu_buffer->new_pages,
bpage            1885 kernel/trace/ring_buffer.c 			list_del_init(&bpage->list);
bpage            1886 kernel/trace/ring_buffer.c 			free_buffer_page(bpage);
bpage            1905 kernel/trace/ring_buffer.c static __always_inline void *__rb_page_index(struct buffer_page *bpage, unsigned index)
bpage            1907 kernel/trace/ring_buffer.c 	return bpage->page->data + index;
bpage            1923 kernel/trace/ring_buffer.c static __always_inline unsigned rb_page_commit(struct buffer_page *bpage)
bpage            1925 kernel/trace/ring_buffer.c 	return local_read(&bpage->page->commit);
bpage            1929 kernel/trace/ring_buffer.c static __always_inline unsigned rb_page_size(struct buffer_page *bpage)
bpage            1931 kernel/trace/ring_buffer.c 	return rb_page_commit(bpage);
bpage            2419 kernel/trace/ring_buffer.c 	struct buffer_page *bpage;
bpage            2428 kernel/trace/ring_buffer.c 	bpage = READ_ONCE(cpu_buffer->tail_page);
bpage            2430 kernel/trace/ring_buffer.c 	if (bpage->page == (void *)addr && rb_page_write(bpage) == old_index) {
bpage            2432 kernel/trace/ring_buffer.c 			local_read(&bpage->write) & ~RB_WRITE_MASK;
bpage            2442 kernel/trace/ring_buffer.c 		index = local_cmpxchg(&bpage->write, old_index, new_index);
bpage            3017 kernel/trace/ring_buffer.c 	struct buffer_page *bpage = cpu_buffer->commit_page;
bpage            3023 kernel/trace/ring_buffer.c 	if (likely(bpage->page == (void *)addr)) {
bpage            3024 kernel/trace/ring_buffer.c 		local_dec(&bpage->entries);
bpage            3032 kernel/trace/ring_buffer.c 	rb_inc_page(cpu_buffer, &bpage);
bpage            3033 kernel/trace/ring_buffer.c 	start = bpage;
bpage            3035 kernel/trace/ring_buffer.c 		if (bpage->page == (void *)addr) {
bpage            3036 kernel/trace/ring_buffer.c 			local_dec(&bpage->entries);
bpage            3039 kernel/trace/ring_buffer.c 		rb_inc_page(cpu_buffer, &bpage);
bpage            3040 kernel/trace/ring_buffer.c 	} while (bpage != start);
bpage            3352 kernel/trace/ring_buffer.c 	struct buffer_page *bpage;
bpage            3365 kernel/trace/ring_buffer.c 		bpage = cpu_buffer->reader_page;
bpage            3367 kernel/trace/ring_buffer.c 		bpage = rb_set_head_page(cpu_buffer);
bpage            3368 kernel/trace/ring_buffer.c 	if (bpage)
bpage            3369 kernel/trace/ring_buffer.c 		ret = bpage->page->time_stamp;
bpage            4596 kernel/trace/ring_buffer.c 	struct buffer_data_page *bpage = NULL;
bpage            4608 kernel/trace/ring_buffer.c 		bpage = cpu_buffer->free_page;
bpage            4615 kernel/trace/ring_buffer.c 	if (bpage)
bpage            4623 kernel/trace/ring_buffer.c 	bpage = page_address(page);
bpage            4626 kernel/trace/ring_buffer.c 	rb_init_page(bpage);
bpage            4628 kernel/trace/ring_buffer.c 	return bpage;
bpage            4643 kernel/trace/ring_buffer.c 	struct buffer_data_page *bpage = data;
bpage            4644 kernel/trace/ring_buffer.c 	struct page *page = virt_to_page(bpage);
bpage            4655 kernel/trace/ring_buffer.c 		cpu_buffer->free_page = bpage;
bpage            4656 kernel/trace/ring_buffer.c 		bpage = NULL;
bpage            4663 kernel/trace/ring_buffer.c 	free_page((unsigned long)bpage);
bpage            4705 kernel/trace/ring_buffer.c 	struct buffer_data_page *bpage;
bpage            4729 kernel/trace/ring_buffer.c 	bpage = *data_page;
bpage            4730 kernel/trace/ring_buffer.c 	if (!bpage)
bpage            4785 kernel/trace/ring_buffer.c 			memcpy(bpage->data + pos, rpage->data + rpos, size);
bpage            4802 kernel/trace/ring_buffer.c 		local_set(&bpage->commit, pos);
bpage            4803 kernel/trace/ring_buffer.c 		bpage->time_stamp = save_timestamp;
bpage            4813 kernel/trace/ring_buffer.c 		rb_init_page(bpage);
bpage            4814 kernel/trace/ring_buffer.c 		bpage = reader->page;
bpage            4819 kernel/trace/ring_buffer.c 		*data_page = bpage;
bpage            4827 kernel/trace/ring_buffer.c 			local_set(&bpage->commit, reader->real_end);
bpage            4833 kernel/trace/ring_buffer.c 	commit = local_read(&bpage->commit);
bpage            4842 kernel/trace/ring_buffer.c 			memcpy(&bpage->data[commit], &missed_events,
bpage            4844 kernel/trace/ring_buffer.c 			local_add(RB_MISSED_STORED, &bpage->commit);
bpage            4847 kernel/trace/ring_buffer.c 		local_add(RB_MISSED_EVENTS, &bpage->commit);
bpage            4854 kernel/trace/ring_buffer.c 		memset(&bpage->data[commit], 0, BUF_PAGE_SIZE - commit);
bpage             110 kernel/trace/ring_buffer_benchmark.c 	void *bpage;
bpage             116 kernel/trace/ring_buffer_benchmark.c 	bpage = ring_buffer_alloc_read_page(buffer, cpu);
bpage             117 kernel/trace/ring_buffer_benchmark.c 	if (IS_ERR(bpage))
bpage             120 kernel/trace/ring_buffer_benchmark.c 	ret = ring_buffer_read_page(buffer, &bpage, PAGE_SIZE, cpu, 1);
bpage             122 kernel/trace/ring_buffer_benchmark.c 		rpage = bpage;
bpage             175 kernel/trace/ring_buffer_benchmark.c 	ring_buffer_free_read_page(buffer, cpu, bpage);