n_pages 126 arch/arc/include/asm/arcregs.h #define PAGES_TO_KB(n_pages) ((n_pages) << (PAGE_SHIFT - 10)) n_pages 127 arch/arc/include/asm/arcregs.h #define PAGES_TO_MB(n_pages) (PAGES_TO_KB(n_pages) >> 10) n_pages 8 arch/ia64/include/asm/uncached.h extern unsigned long uncached_alloc_page(int starting_nid, int n_pages); n_pages 9 arch/ia64/include/asm/uncached.h extern void uncached_free_page(unsigned long uc_addr, int n_pages); n_pages 179 arch/ia64/kernel/uncached.c unsigned long uncached_alloc_page(int starting_nid, int n_pages) n_pages 200 arch/ia64/kernel/uncached.c n_pages * PAGE_SIZE); n_pages 220 arch/ia64/kernel/uncached.c void uncached_free_page(unsigned long uc_addr, int n_pages) n_pages 231 arch/ia64/kernel/uncached.c gen_pool_free(pool, uc_addr, n_pages * PAGE_SIZE); n_pages 149 arch/sh/mm/cache-sh5.c int n_pages; n_pages 154 arch/sh/mm/cache-sh5.c n_pages = ((end - start) >> PAGE_SHIFT); n_pages 155 arch/sh/mm/cache-sh5.c if (n_pages >= 64) { n_pages 470 arch/sh/mm/cache-sh5.c int n_pages = ((end - start) >> PAGE_SHIFT); n_pages 472 arch/sh/mm/cache-sh5.c if (n_pages >= 64 || ((start ^ (end - 1)) & PMD_MASK)) { n_pages 40 arch/xtensa/mm/kasan_init.c unsigned long n_pages = (end - start) / PAGE_SIZE; n_pages 41 arch/xtensa/mm/kasan_init.c unsigned long n_pmds = n_pages / PTRS_PER_PTE; n_pages 46 arch/xtensa/mm/kasan_init.c pte_t *pte = memblock_alloc(n_pages * sizeof(pte_t), PAGE_SIZE); n_pages 50 arch/xtensa/mm/kasan_init.c __func__, n_pages * sizeof(pte_t), PAGE_SIZE); n_pages 22 arch/xtensa/mm/mmu.c static void * __init init_pmd(unsigned long vaddr, unsigned long n_pages) n_pages 29 arch/xtensa/mm/mmu.c n_pages = ALIGN(n_pages, PTRS_PER_PTE); n_pages 32 arch/xtensa/mm/mmu.c __func__, vaddr, n_pages); n_pages 34 arch/xtensa/mm/mmu.c pte = memblock_alloc_low(n_pages * sizeof(pte_t), PAGE_SIZE); n_pages 37 arch/xtensa/mm/mmu.c __func__, n_pages * sizeof(pte_t), PAGE_SIZE); n_pages 39 arch/xtensa/mm/mmu.c for (i = 0; i < n_pages; ++i) n_pages 42 arch/xtensa/mm/mmu.c for (i = 0; i < n_pages; i += PTRS_PER_PTE, ++pmd) { n_pages 230 drivers/gpu/drm/i915/gem/i915_gem_pages.c unsigned long n_pages = obj->base.size >> PAGE_SHIFT; n_pages 241 drivers/gpu/drm/i915/gem/i915_gem_pages.c if (n_pages == 1 && type == I915_MAP_WB) n_pages 244 drivers/gpu/drm/i915/gem/i915_gem_pages.c if (n_pages > ARRAY_SIZE(stack_pages)) { n_pages 246 drivers/gpu/drm/i915/gem/i915_gem_pages.c pages = kvmalloc_array(n_pages, sizeof(*pages), GFP_KERNEL); n_pages 255 drivers/gpu/drm/i915/gem/i915_gem_pages.c GEM_BUG_ON(i != n_pages); n_pages 268 drivers/gpu/drm/i915/gem/i915_gem_pages.c addr = vmap(pages, n_pages, 0, pgprot); n_pages 323 drivers/gpu/drm/vgem/vgem_drv.c long n_pages = obj->size >> PAGE_SHIFT; n_pages 333 drivers/gpu/drm/vgem/vgem_drv.c drm_clflush_pages(pages, n_pages); n_pages 388 drivers/gpu/drm/vgem/vgem_drv.c long n_pages = obj->size >> PAGE_SHIFT; n_pages 395 drivers/gpu/drm/vgem/vgem_drv.c return vmap(pages, n_pages, 0, pgprot_writecombine(PAGE_KERNEL)); n_pages 198 drivers/gpu/drm/vkms/vkms_gem.c unsigned int n_pages = obj->size >> PAGE_SHIFT; n_pages 206 drivers/gpu/drm/vkms/vkms_gem.c vkms_obj->vaddr = vmap(pages, n_pages, VM_MAP, PAGE_KERNEL); n_pages 3156 drivers/md/dm-integrity.c static struct page_list *dm_integrity_alloc_page_list(unsigned n_pages) n_pages 3161 drivers/md/dm-integrity.c pl = kvmalloc_array(n_pages + 1, sizeof(struct page_list), GFP_KERNEL | __GFP_ZERO); n_pages 3165 drivers/md/dm-integrity.c for (i = 0; i < n_pages; i++) { n_pages 3204 drivers/md/dm-integrity.c unsigned n_pages; n_pages 3211 drivers/md/dm-integrity.c n_pages = (end_index - start_index + 1); n_pages 3213 drivers/md/dm-integrity.c s = kvmalloc_array(n_pages, sizeof(struct scatterlist), n_pages 3220 drivers/md/dm-integrity.c sg_init_table(s, n_pages); n_pages 459 drivers/media/common/videobuf2/videobuf2-dma-contig.c int n_pages, i; n_pages 493 drivers/media/common/videobuf2/videobuf2-dma-contig.c n_pages = frame_vector_count(vec); n_pages 502 drivers/media/common/videobuf2/videobuf2-dma-contig.c for (i = 1; i < n_pages; i++) n_pages 521 drivers/media/common/videobuf2/videobuf2-dma-contig.c ret = sg_alloc_table_from_pages(sgt, frame_vector_pages(vec), n_pages, n_pages 80 drivers/media/common/videobuf2/videobuf2-vmalloc.c int n_pages, offset, i; n_pages 96 drivers/media/common/videobuf2/videobuf2-vmalloc.c n_pages = frame_vector_count(vec); n_pages 104 drivers/media/common/videobuf2/videobuf2-vmalloc.c for (i = 1; i < n_pages; i++) n_pages 110 drivers/media/common/videobuf2/videobuf2-vmalloc.c buf->vaddr = vm_map_ram(frame_vector_pages(vec), n_pages, -1, n_pages 133 drivers/media/common/videobuf2/videobuf2-vmalloc.c unsigned int n_pages; n_pages 136 drivers/media/common/videobuf2/videobuf2-vmalloc.c n_pages = frame_vector_count(buf->vec); n_pages 139 drivers/media/common/videobuf2/videobuf2-vmalloc.c vm_unmap_ram((void *)vaddr, n_pages); n_pages 142 drivers/media/common/videobuf2/videobuf2-vmalloc.c for (i = 0; i < n_pages; i++) n_pages 245 drivers/misc/vmw_balloon.c unsigned int n_pages; n_pages 705 drivers/misc/vmw_balloon.c ctl->n_pages = i; n_pages 707 drivers/misc/vmw_balloon.c return req_n_pages == ctl->n_pages ? 0 : -ENOMEM; n_pages 865 drivers/misc/vmw_balloon.c num_pages = ctl->n_pages; n_pages 876 drivers/misc/vmw_balloon.c batch_status = vmballoon_lock_op(b, ctl->n_pages, ctl->page_size, n_pages 906 drivers/misc/vmw_balloon.c ctl->n_pages--; n_pages 925 drivers/misc/vmw_balloon.c int *n_pages, n_pages 935 drivers/misc/vmw_balloon.c if (n_pages) n_pages 936 drivers/misc/vmw_balloon.c *n_pages = 0; n_pages 1002 drivers/misc/vmw_balloon.c unsigned int *n_pages, n_pages 1022 drivers/misc/vmw_balloon.c __count_vm_events(BALLOON_INFLATE, *n_pages * n_pages 1027 drivers/misc/vmw_balloon.c *n_pages = 0; n_pages 1045 drivers/misc/vmw_balloon.c unsigned int *n_pages, n_pages 1055 drivers/misc/vmw_balloon.c *n_pages = balloon_page_list_dequeue(&b->b_dev_info, pages, n_pages 1073 drivers/misc/vmw_balloon.c *n_pages = i; n_pages 1123 drivers/misc/vmw_balloon.c VM_BUG_ON(ctl.n_pages != 0); n_pages 1146 drivers/misc/vmw_balloon.c atomic64_add(ctl.n_pages * page_in_frames, &b->size); n_pages 1148 drivers/misc/vmw_balloon.c vmballoon_enqueue_page_list(b, &ctl.pages, &ctl.n_pages, n_pages 1216 drivers/misc/vmw_balloon.c VM_BUG_ON(ctl.n_pages); n_pages 1241 drivers/misc/vmw_balloon.c vmballoon_dequeue_page_list(b, &ctl.pages, &ctl.n_pages, n_pages 1248 drivers/misc/vmw_balloon.c tried_frames += ctl.n_pages * page_in_frames; n_pages 1266 drivers/misc/vmw_balloon.c deflated_all = (ctl.n_pages == to_deflate_pages); n_pages 1269 drivers/misc/vmw_balloon.c n_unlocked_frames = ctl.n_pages * page_in_frames; n_pages 1274 drivers/misc/vmw_balloon.c ctl.page_size, ctl.n_pages); n_pages 1277 drivers/misc/vmw_balloon.c vmballoon_release_page_list(&ctl.pages, &ctl.n_pages, n_pages 309 drivers/net/ethernet/sfc/mcdi_mon.c unsigned int n_pages, n_sensors, n_attrs, page; n_pages 332 drivers/net/ethernet/sfc/mcdi_mon.c n_pages = page; n_pages 378 drivers/net/ethernet/sfc/mcdi_mon.c if (page == n_pages) n_pages 37 drivers/staging/comedi/comedi_buf.c PAGE_SIZE * bm->n_pages, n_pages 40 drivers/staging/comedi/comedi_buf.c for (i = 0; i < bm->n_pages; i++) { n_pages 76 drivers/staging/comedi/comedi_buf.c unsigned int n_pages) n_pages 93 drivers/staging/comedi/comedi_buf.c bm->page_list = vzalloc(sizeof(*buf) * n_pages); n_pages 106 drivers/staging/comedi/comedi_buf.c PAGE_SIZE * n_pages, &dma_addr, n_pages 111 drivers/staging/comedi/comedi_buf.c for (i = 0; i < n_pages; i++) { n_pages 117 drivers/staging/comedi/comedi_buf.c bm->n_pages = i; n_pages 119 drivers/staging/comedi/comedi_buf.c for (i = 0; i < n_pages; i++) { n_pages 128 drivers/staging/comedi/comedi_buf.c bm->n_pages = i; n_pages 129 drivers/staging/comedi/comedi_buf.c if (i < n_pages) n_pages 142 drivers/staging/comedi/comedi_buf.c unsigned int n_pages) n_pages 157 drivers/staging/comedi/comedi_buf.c bm = comedi_buf_map_alloc(dev, s->async_dma_dir, n_pages); n_pages 173 drivers/staging/comedi/comedi_buf.c pages = vmalloc(sizeof(struct page *) * n_pages); n_pages 177 drivers/staging/comedi/comedi_buf.c for (i = 0; i < n_pages; i++) { n_pages 183 drivers/staging/comedi/comedi_buf.c async->prealloc_buf = vmap(pages, n_pages, VM_MAP, n_pages 211 drivers/staging/comedi/comedi_buf.c while (done < len && pg < bm->n_pages) { n_pages 241 drivers/staging/comedi/comedi_buf.c if (bm && bm->n_pages) n_pages 276 drivers/staging/comedi/comedi_buf.c unsigned int n_pages = new_size >> PAGE_SHIFT; n_pages 278 drivers/staging/comedi/comedi_buf.c __comedi_buf_alloc(dev, s, n_pages); n_pages 2307 drivers/staging/comedi/comedi_fops.c int n_pages; n_pages 2357 drivers/staging/comedi/comedi_fops.c n_pages = vma_pages(vma); n_pages 2361 drivers/staging/comedi/comedi_fops.c if (!bm || n_pages > bm->n_pages) { n_pages 2372 drivers/staging/comedi/comedi_fops.c buf->dma_addr, n_pages * PAGE_SIZE); n_pages 2374 drivers/staging/comedi/comedi_fops.c for (i = 0; i < n_pages; ++i) { n_pages 252 drivers/staging/comedi/comedidev.h unsigned int n_pages; n_pages 768 drivers/usb/gadget/function/f_fs.c unsigned int n_pages; n_pages 775 drivers/usb/gadget/function/f_fs.c n_pages = PAGE_ALIGN(sz) >> PAGE_SHIFT; n_pages 776 drivers/usb/gadget/function/f_fs.c pages = kvmalloc_array(n_pages, sizeof(struct page *), GFP_KERNEL); n_pages 782 drivers/usb/gadget/function/f_fs.c for (i = 0, ptr = vaddr; i < n_pages; ++i, ptr += PAGE_SIZE) n_pages 785 drivers/usb/gadget/function/f_fs.c if (sg_alloc_table_from_pages(sgt, pages, n_pages, 0, sz, GFP_KERNEL)) { n_pages 535 drivers/xen/balloon.c long n_pages; n_pages 537 drivers/xen/balloon.c n_pages = min(-credit, si_mem_available()); n_pages 538 drivers/xen/balloon.c state = decrease_reservation(n_pages, GFP_BALLOON); n_pages 539 drivers/xen/balloon.c if (state == BP_DONE && n_pages != -credit && n_pages 540 drivers/xen/balloon.c n_pages < totalreserve_pages) n_pages 33 drivers/xen/privcmd-buf.c unsigned int n_pages; n_pages 59 drivers/xen/privcmd-buf.c for (i = 0; i < vma_priv->n_pages; i++) n_pages 151 drivers/xen/privcmd-buf.c vma_priv->n_pages++; n_pages 165 drivers/xen/privcmd-buf.c if (vma_priv->n_pages != count) n_pages 169 drivers/xen/privcmd-buf.c vma_priv->n_pages); n_pages 276 fs/fscache/page.c atomic_set(&op->n_pages, 0); n_pages 290 fs/fscache/page.c atomic_read(&op->n_pages), ==, 0); n_pages 464 fs/fscache/page.c atomic_set(&op->n_pages, 1); n_pages 593 fs/fscache/page.c atomic_set(&op->n_pages, *nr_pages); n_pages 705 fs/fscache/page.c atomic_set(&op->n_pages, 1); n_pages 151 include/linux/fscache-cache.h atomic_t n_pages; /* number of pages to be retrieved */ n_pages 193 include/linux/fscache-cache.h int n_pages) n_pages 195 include/linux/fscache-cache.h if (atomic_sub_return_relaxed(n_pages, &op->n_pages) <= 0) n_pages 276 include/linux/scatterlist.h unsigned int n_pages, unsigned int offset, n_pages 280 include/linux/scatterlist.h unsigned int n_pages, unsigned int offset, n_pages 71 kernel/relay.c static struct page **relay_alloc_page_array(unsigned int n_pages) n_pages 73 kernel/relay.c const size_t pa_size = n_pages * sizeof(struct page *); n_pages 126 kernel/relay.c unsigned int i, j, n_pages; n_pages 129 kernel/relay.c n_pages = *size >> PAGE_SHIFT; n_pages 131 kernel/relay.c buf->page_array = relay_alloc_page_array(n_pages); n_pages 135 kernel/relay.c for (i = 0; i < n_pages; i++) { n_pages 141 kernel/relay.c mem = vmap(buf->page_array, n_pages, VM_MAP, PAGE_KERNEL); n_pages 146 kernel/relay.c buf->page_count = n_pages; n_pages 293 kernel/trace/tracing_map.c for (i = 0; i < a->n_pages; i++) n_pages 307 kernel/trace/tracing_map.c for (i = 0; i < a->n_pages; i++) { n_pages 331 kernel/trace/tracing_map.c a->n_pages = n_elts / a->entries_per_page; n_pages 332 kernel/trace/tracing_map.c if (!a->n_pages) n_pages 333 kernel/trace/tracing_map.c a->n_pages = 1; n_pages 337 kernel/trace/tracing_map.c a->pages = kcalloc(a->n_pages, sizeof(void *), GFP_KERNEL); n_pages 341 kernel/trace/tracing_map.c for (i = 0; i < a->n_pages; i++) { n_pages 169 kernel/trace/tracing_map.h unsigned int n_pages; n_pages 390 lib/scatterlist.c unsigned int n_pages, unsigned int offset, n_pages 404 lib/scatterlist.c for (i = 1; i < n_pages; i++) { n_pages 424 lib/scatterlist.c for (j = cur_page + 1; j < n_pages; j++) { n_pages 465 lib/scatterlist.c unsigned int n_pages, unsigned int offset, n_pages 468 lib/scatterlist.c return __sg_alloc_table_from_pages(sgt, pages, n_pages, offset, size, n_pages 45 mm/balloon_compaction.c size_t n_pages = 0; n_pages 51 mm/balloon_compaction.c n_pages++; n_pages 54 mm/balloon_compaction.c return n_pages; n_pages 81 mm/balloon_compaction.c size_t n_pages = 0; n_pages 85 mm/balloon_compaction.c if (n_pages == n_req_pages) n_pages 106 mm/balloon_compaction.c n_pages++; n_pages 110 mm/balloon_compaction.c return n_pages; n_pages 181 mm/balloon_compaction.c int n_pages; n_pages 183 mm/balloon_compaction.c n_pages = balloon_page_list_dequeue(b_dev_info, &pages, 1); n_pages 185 mm/balloon_compaction.c if (n_pages != 1) {