aux_nr_pages     5615 kernel/events/core.c 		atomic_long_sub(rb->aux_nr_pages - rb->aux_mmap_locked, &mmap_user->locked_vm);
aux_nr_pages     5768 kernel/events/core.c 		if (rb_has_aux(rb) && rb->aux_nr_pages != nr_pages)
aux_nr_pages       47 kernel/events/internal.h 	int				aux_nr_pages;
aux_nr_pages       89 kernel/events/internal.h 	return !!rb->aux_nr_pages;
aux_nr_pages      125 kernel/events/internal.h 	return rb->aux_nr_pages << PAGE_SHIFT;
aux_nr_pages      620 kernel/events/ring_buffer.c 	if (rb->aux_nr_pages) {
aux_nr_pages      621 kernel/events/ring_buffer.c 		for (pg = 0; pg < rb->aux_nr_pages; pg++)
aux_nr_pages      625 kernel/events/ring_buffer.c 		rb->aux_nr_pages = 0;
aux_nr_pages      662 kernel/events/ring_buffer.c 	for (rb->aux_nr_pages = 0; rb->aux_nr_pages < nr_pages;) {
aux_nr_pages      666 kernel/events/ring_buffer.c 		order = min(max_order, ilog2(nr_pages - rb->aux_nr_pages));
aux_nr_pages      671 kernel/events/ring_buffer.c 		for (last = rb->aux_nr_pages + (1 << page_private(page));
aux_nr_pages      672 kernel/events/ring_buffer.c 		     last > rb->aux_nr_pages; rb->aux_nr_pages++)
aux_nr_pages      673 kernel/events/ring_buffer.c 			rb->aux_pages[rb->aux_nr_pages] = page_address(page++);
aux_nr_pages      908 kernel/events/ring_buffer.c 	if (rb->aux_nr_pages) {
aux_nr_pages      910 kernel/events/ring_buffer.c 		if (pgoff > rb->aux_pgoff + rb->aux_nr_pages)
aux_nr_pages      915 kernel/events/ring_buffer.c 			int aux_pgoff = array_index_nospec(pgoff - rb->aux_pgoff, rb->aux_nr_pages);