aux_nr_pages 5615 kernel/events/core.c atomic_long_sub(rb->aux_nr_pages - rb->aux_mmap_locked, &mmap_user->locked_vm); aux_nr_pages 5768 kernel/events/core.c if (rb_has_aux(rb) && rb->aux_nr_pages != nr_pages) aux_nr_pages 47 kernel/events/internal.h int aux_nr_pages; aux_nr_pages 89 kernel/events/internal.h return !!rb->aux_nr_pages; aux_nr_pages 125 kernel/events/internal.h return rb->aux_nr_pages << PAGE_SHIFT; aux_nr_pages 620 kernel/events/ring_buffer.c if (rb->aux_nr_pages) { aux_nr_pages 621 kernel/events/ring_buffer.c for (pg = 0; pg < rb->aux_nr_pages; pg++) aux_nr_pages 625 kernel/events/ring_buffer.c rb->aux_nr_pages = 0; aux_nr_pages 662 kernel/events/ring_buffer.c for (rb->aux_nr_pages = 0; rb->aux_nr_pages < nr_pages;) { aux_nr_pages 666 kernel/events/ring_buffer.c order = min(max_order, ilog2(nr_pages - rb->aux_nr_pages)); aux_nr_pages 671 kernel/events/ring_buffer.c for (last = rb->aux_nr_pages + (1 << page_private(page)); aux_nr_pages 672 kernel/events/ring_buffer.c last > rb->aux_nr_pages; rb->aux_nr_pages++) aux_nr_pages 673 kernel/events/ring_buffer.c rb->aux_pages[rb->aux_nr_pages] = page_address(page++); aux_nr_pages 908 kernel/events/ring_buffer.c if (rb->aux_nr_pages) { aux_nr_pages 910 kernel/events/ring_buffer.c if (pgoff > rb->aux_pgoff + rb->aux_nr_pages) aux_nr_pages 915 kernel/events/ring_buffer.c int aux_pgoff = array_index_nospec(pgoff - rb->aux_pgoff, rb->aux_nr_pages);