n_pages           126 arch/arc/include/asm/arcregs.h #define PAGES_TO_KB(n_pages)	((n_pages) << (PAGE_SHIFT - 10))
n_pages           127 arch/arc/include/asm/arcregs.h #define PAGES_TO_MB(n_pages)	(PAGES_TO_KB(n_pages) >> 10)
n_pages             8 arch/ia64/include/asm/uncached.h extern unsigned long uncached_alloc_page(int starting_nid, int n_pages);
n_pages             9 arch/ia64/include/asm/uncached.h extern void uncached_free_page(unsigned long uc_addr, int n_pages);
n_pages           179 arch/ia64/kernel/uncached.c unsigned long uncached_alloc_page(int starting_nid, int n_pages)
n_pages           200 arch/ia64/kernel/uncached.c 						 n_pages * PAGE_SIZE);
n_pages           220 arch/ia64/kernel/uncached.c void uncached_free_page(unsigned long uc_addr, int n_pages)
n_pages           231 arch/ia64/kernel/uncached.c 	gen_pool_free(pool, uc_addr, n_pages * PAGE_SIZE);
n_pages           149 arch/sh/mm/cache-sh5.c 	int n_pages;
n_pages           154 arch/sh/mm/cache-sh5.c 	n_pages = ((end - start) >> PAGE_SHIFT);
n_pages           155 arch/sh/mm/cache-sh5.c 	if (n_pages >= 64) {
n_pages           470 arch/sh/mm/cache-sh5.c 	int n_pages = ((end - start) >> PAGE_SHIFT);
n_pages           472 arch/sh/mm/cache-sh5.c 	if (n_pages >= 64 || ((start ^ (end - 1)) & PMD_MASK)) {
n_pages            40 arch/xtensa/mm/kasan_init.c 	unsigned long n_pages = (end - start) / PAGE_SIZE;
n_pages            41 arch/xtensa/mm/kasan_init.c 	unsigned long n_pmds = n_pages / PTRS_PER_PTE;
n_pages            46 arch/xtensa/mm/kasan_init.c 	pte_t *pte = memblock_alloc(n_pages * sizeof(pte_t), PAGE_SIZE);
n_pages            50 arch/xtensa/mm/kasan_init.c 		      __func__, n_pages * sizeof(pte_t), PAGE_SIZE);
n_pages            22 arch/xtensa/mm/mmu.c static void * __init init_pmd(unsigned long vaddr, unsigned long n_pages)
n_pages            29 arch/xtensa/mm/mmu.c 	n_pages = ALIGN(n_pages, PTRS_PER_PTE);
n_pages            32 arch/xtensa/mm/mmu.c 		 __func__, vaddr, n_pages);
n_pages            34 arch/xtensa/mm/mmu.c 	pte = memblock_alloc_low(n_pages * sizeof(pte_t), PAGE_SIZE);
n_pages            37 arch/xtensa/mm/mmu.c 		      __func__, n_pages * sizeof(pte_t), PAGE_SIZE);
n_pages            39 arch/xtensa/mm/mmu.c 	for (i = 0; i < n_pages; ++i)
n_pages            42 arch/xtensa/mm/mmu.c 	for (i = 0; i < n_pages; i += PTRS_PER_PTE, ++pmd) {
n_pages           230 drivers/gpu/drm/i915/gem/i915_gem_pages.c 	unsigned long n_pages = obj->base.size >> PAGE_SHIFT;
n_pages           241 drivers/gpu/drm/i915/gem/i915_gem_pages.c 	if (n_pages == 1 && type == I915_MAP_WB)
n_pages           244 drivers/gpu/drm/i915/gem/i915_gem_pages.c 	if (n_pages > ARRAY_SIZE(stack_pages)) {
n_pages           246 drivers/gpu/drm/i915/gem/i915_gem_pages.c 		pages = kvmalloc_array(n_pages, sizeof(*pages), GFP_KERNEL);
n_pages           255 drivers/gpu/drm/i915/gem/i915_gem_pages.c 	GEM_BUG_ON(i != n_pages);
n_pages           268 drivers/gpu/drm/i915/gem/i915_gem_pages.c 	addr = vmap(pages, n_pages, 0, pgprot);
n_pages           323 drivers/gpu/drm/vgem/vgem_drv.c 	long n_pages = obj->size >> PAGE_SHIFT;
n_pages           333 drivers/gpu/drm/vgem/vgem_drv.c 	drm_clflush_pages(pages, n_pages);
n_pages           388 drivers/gpu/drm/vgem/vgem_drv.c 	long n_pages = obj->size >> PAGE_SHIFT;
n_pages           395 drivers/gpu/drm/vgem/vgem_drv.c 	return vmap(pages, n_pages, 0, pgprot_writecombine(PAGE_KERNEL));
n_pages           198 drivers/gpu/drm/vkms/vkms_gem.c 		unsigned int n_pages = obj->size >> PAGE_SHIFT;
n_pages           206 drivers/gpu/drm/vkms/vkms_gem.c 		vkms_obj->vaddr = vmap(pages, n_pages, VM_MAP, PAGE_KERNEL);
n_pages          3156 drivers/md/dm-integrity.c static struct page_list *dm_integrity_alloc_page_list(unsigned n_pages)
n_pages          3161 drivers/md/dm-integrity.c 	pl = kvmalloc_array(n_pages + 1, sizeof(struct page_list), GFP_KERNEL | __GFP_ZERO);
n_pages          3165 drivers/md/dm-integrity.c 	for (i = 0; i < n_pages; i++) {
n_pages          3204 drivers/md/dm-integrity.c 		unsigned n_pages;
n_pages          3211 drivers/md/dm-integrity.c 		n_pages = (end_index - start_index + 1);
n_pages          3213 drivers/md/dm-integrity.c 		s = kvmalloc_array(n_pages, sizeof(struct scatterlist),
n_pages          3220 drivers/md/dm-integrity.c 		sg_init_table(s, n_pages);
n_pages           459 drivers/media/common/videobuf2/videobuf2-dma-contig.c 	int n_pages, i;
n_pages           493 drivers/media/common/videobuf2/videobuf2-dma-contig.c 	n_pages = frame_vector_count(vec);
n_pages           502 drivers/media/common/videobuf2/videobuf2-dma-contig.c 		for (i = 1; i < n_pages; i++)
n_pages           521 drivers/media/common/videobuf2/videobuf2-dma-contig.c 	ret = sg_alloc_table_from_pages(sgt, frame_vector_pages(vec), n_pages,
n_pages            80 drivers/media/common/videobuf2/videobuf2-vmalloc.c 	int n_pages, offset, i;
n_pages            96 drivers/media/common/videobuf2/videobuf2-vmalloc.c 	n_pages = frame_vector_count(vec);
n_pages           104 drivers/media/common/videobuf2/videobuf2-vmalloc.c 		for (i = 1; i < n_pages; i++)
n_pages           110 drivers/media/common/videobuf2/videobuf2-vmalloc.c 		buf->vaddr = vm_map_ram(frame_vector_pages(vec), n_pages, -1,
n_pages           133 drivers/media/common/videobuf2/videobuf2-vmalloc.c 	unsigned int n_pages;
n_pages           136 drivers/media/common/videobuf2/videobuf2-vmalloc.c 		n_pages = frame_vector_count(buf->vec);
n_pages           139 drivers/media/common/videobuf2/videobuf2-vmalloc.c 			vm_unmap_ram((void *)vaddr, n_pages);
n_pages           142 drivers/media/common/videobuf2/videobuf2-vmalloc.c 			for (i = 0; i < n_pages; i++)
n_pages           245 drivers/misc/vmw_balloon.c 	unsigned int n_pages;
n_pages           705 drivers/misc/vmw_balloon.c 	ctl->n_pages = i;
n_pages           707 drivers/misc/vmw_balloon.c 	return req_n_pages == ctl->n_pages ? 0 : -ENOMEM;
n_pages           865 drivers/misc/vmw_balloon.c 	num_pages = ctl->n_pages;
n_pages           876 drivers/misc/vmw_balloon.c 	batch_status = vmballoon_lock_op(b, ctl->n_pages, ctl->page_size,
n_pages           906 drivers/misc/vmw_balloon.c 		ctl->n_pages--;
n_pages           925 drivers/misc/vmw_balloon.c 				       int *n_pages,
n_pages           935 drivers/misc/vmw_balloon.c 	if (n_pages)
n_pages           936 drivers/misc/vmw_balloon.c 		*n_pages = 0;
n_pages          1002 drivers/misc/vmw_balloon.c 					unsigned int *n_pages,
n_pages          1022 drivers/misc/vmw_balloon.c 		__count_vm_events(BALLOON_INFLATE, *n_pages *
n_pages          1027 drivers/misc/vmw_balloon.c 	*n_pages = 0;
n_pages          1045 drivers/misc/vmw_balloon.c 					unsigned int *n_pages,
n_pages          1055 drivers/misc/vmw_balloon.c 		*n_pages = balloon_page_list_dequeue(&b->b_dev_info, pages,
n_pages          1073 drivers/misc/vmw_balloon.c 	*n_pages = i;
n_pages          1123 drivers/misc/vmw_balloon.c 		VM_BUG_ON(ctl.n_pages != 0);
n_pages          1146 drivers/misc/vmw_balloon.c 		atomic64_add(ctl.n_pages * page_in_frames, &b->size);
n_pages          1148 drivers/misc/vmw_balloon.c 		vmballoon_enqueue_page_list(b, &ctl.pages, &ctl.n_pages,
n_pages          1216 drivers/misc/vmw_balloon.c 		VM_BUG_ON(ctl.n_pages);
n_pages          1241 drivers/misc/vmw_balloon.c 		vmballoon_dequeue_page_list(b, &ctl.pages, &ctl.n_pages,
n_pages          1248 drivers/misc/vmw_balloon.c 		tried_frames += ctl.n_pages * page_in_frames;
n_pages          1266 drivers/misc/vmw_balloon.c 		deflated_all = (ctl.n_pages == to_deflate_pages);
n_pages          1269 drivers/misc/vmw_balloon.c 		n_unlocked_frames = ctl.n_pages * page_in_frames;
n_pages          1274 drivers/misc/vmw_balloon.c 					 ctl.page_size, ctl.n_pages);
n_pages          1277 drivers/misc/vmw_balloon.c 		vmballoon_release_page_list(&ctl.pages, &ctl.n_pages,
n_pages           309 drivers/net/ethernet/sfc/mcdi_mon.c 	unsigned int n_pages, n_sensors, n_attrs, page;
n_pages           332 drivers/net/ethernet/sfc/mcdi_mon.c 	n_pages = page;
n_pages           378 drivers/net/ethernet/sfc/mcdi_mon.c 				if (page == n_pages)
n_pages            37 drivers/staging/comedi/comedi_buf.c 					  PAGE_SIZE * bm->n_pages,
n_pages            40 drivers/staging/comedi/comedi_buf.c 			for (i = 0; i < bm->n_pages; i++) {
n_pages            76 drivers/staging/comedi/comedi_buf.c 		     unsigned int n_pages)
n_pages            93 drivers/staging/comedi/comedi_buf.c 	bm->page_list = vzalloc(sizeof(*buf) * n_pages);
n_pages           106 drivers/staging/comedi/comedi_buf.c 					       PAGE_SIZE * n_pages, &dma_addr,
n_pages           111 drivers/staging/comedi/comedi_buf.c 		for (i = 0; i < n_pages; i++) {
n_pages           117 drivers/staging/comedi/comedi_buf.c 		bm->n_pages = i;
n_pages           119 drivers/staging/comedi/comedi_buf.c 		for (i = 0; i < n_pages; i++) {
n_pages           128 drivers/staging/comedi/comedi_buf.c 		bm->n_pages = i;
n_pages           129 drivers/staging/comedi/comedi_buf.c 		if (i < n_pages)
n_pages           142 drivers/staging/comedi/comedi_buf.c 			       unsigned int n_pages)
n_pages           157 drivers/staging/comedi/comedi_buf.c 	bm = comedi_buf_map_alloc(dev, s->async_dma_dir, n_pages);
n_pages           173 drivers/staging/comedi/comedi_buf.c 		pages = vmalloc(sizeof(struct page *) * n_pages);
n_pages           177 drivers/staging/comedi/comedi_buf.c 		for (i = 0; i < n_pages; i++) {
n_pages           183 drivers/staging/comedi/comedi_buf.c 		async->prealloc_buf = vmap(pages, n_pages, VM_MAP,
n_pages           211 drivers/staging/comedi/comedi_buf.c 	while (done < len && pg < bm->n_pages) {
n_pages           241 drivers/staging/comedi/comedi_buf.c 	if (bm && bm->n_pages)
n_pages           276 drivers/staging/comedi/comedi_buf.c 		unsigned int n_pages = new_size >> PAGE_SHIFT;
n_pages           278 drivers/staging/comedi/comedi_buf.c 		__comedi_buf_alloc(dev, s, n_pages);
n_pages          2307 drivers/staging/comedi/comedi_fops.c 	int n_pages;
n_pages          2357 drivers/staging/comedi/comedi_fops.c 	n_pages = vma_pages(vma);
n_pages          2361 drivers/staging/comedi/comedi_fops.c 	if (!bm || n_pages > bm->n_pages) {
n_pages          2372 drivers/staging/comedi/comedi_fops.c 					   buf->dma_addr, n_pages * PAGE_SIZE);
n_pages          2374 drivers/staging/comedi/comedi_fops.c 		for (i = 0; i < n_pages; ++i) {
n_pages           252 drivers/staging/comedi/comedidev.h 	unsigned int n_pages;
n_pages           768 drivers/usb/gadget/function/f_fs.c 	unsigned int n_pages;
n_pages           775 drivers/usb/gadget/function/f_fs.c 	n_pages = PAGE_ALIGN(sz) >> PAGE_SHIFT;
n_pages           776 drivers/usb/gadget/function/f_fs.c 	pages = kvmalloc_array(n_pages, sizeof(struct page *), GFP_KERNEL);
n_pages           782 drivers/usb/gadget/function/f_fs.c 	for (i = 0, ptr = vaddr; i < n_pages; ++i, ptr += PAGE_SIZE)
n_pages           785 drivers/usb/gadget/function/f_fs.c 	if (sg_alloc_table_from_pages(sgt, pages, n_pages, 0, sz, GFP_KERNEL)) {
n_pages           535 drivers/xen/balloon.c 			long n_pages;
n_pages           537 drivers/xen/balloon.c 			n_pages = min(-credit, si_mem_available());
n_pages           538 drivers/xen/balloon.c 			state = decrease_reservation(n_pages, GFP_BALLOON);
n_pages           539 drivers/xen/balloon.c 			if (state == BP_DONE && n_pages != -credit &&
n_pages           540 drivers/xen/balloon.c 			    n_pages < totalreserve_pages)
n_pages            33 drivers/xen/privcmd-buf.c 	unsigned int n_pages;
n_pages            59 drivers/xen/privcmd-buf.c 	for (i = 0; i < vma_priv->n_pages; i++)
n_pages           151 drivers/xen/privcmd-buf.c 		vma_priv->n_pages++;
n_pages           165 drivers/xen/privcmd-buf.c 	if (vma_priv->n_pages != count)
n_pages           169 drivers/xen/privcmd-buf.c 						vma_priv->n_pages);
n_pages           276 fs/fscache/page.c 	atomic_set(&op->n_pages, 0);
n_pages           290 fs/fscache/page.c 		    atomic_read(&op->n_pages), ==, 0);
n_pages           464 fs/fscache/page.c 	atomic_set(&op->n_pages, 1);
n_pages           593 fs/fscache/page.c 	atomic_set(&op->n_pages, *nr_pages);
n_pages           705 fs/fscache/page.c 	atomic_set(&op->n_pages, 1);
n_pages           151 include/linux/fscache-cache.h 	atomic_t		n_pages;	/* number of pages to be retrieved */
n_pages           193 include/linux/fscache-cache.h 					      int n_pages)
n_pages           195 include/linux/fscache-cache.h 	if (atomic_sub_return_relaxed(n_pages, &op->n_pages) <= 0)
n_pages           276 include/linux/scatterlist.h 				unsigned int n_pages, unsigned int offset,
n_pages           280 include/linux/scatterlist.h 			      unsigned int n_pages, unsigned int offset,
n_pages            71 kernel/relay.c static struct page **relay_alloc_page_array(unsigned int n_pages)
n_pages            73 kernel/relay.c 	const size_t pa_size = n_pages * sizeof(struct page *);
n_pages           126 kernel/relay.c 	unsigned int i, j, n_pages;
n_pages           129 kernel/relay.c 	n_pages = *size >> PAGE_SHIFT;
n_pages           131 kernel/relay.c 	buf->page_array = relay_alloc_page_array(n_pages);
n_pages           135 kernel/relay.c 	for (i = 0; i < n_pages; i++) {
n_pages           141 kernel/relay.c 	mem = vmap(buf->page_array, n_pages, VM_MAP, PAGE_KERNEL);
n_pages           146 kernel/relay.c 	buf->page_count = n_pages;
n_pages           293 kernel/trace/tracing_map.c 	for (i = 0; i < a->n_pages; i++)
n_pages           307 kernel/trace/tracing_map.c 	for (i = 0; i < a->n_pages; i++) {
n_pages           331 kernel/trace/tracing_map.c 	a->n_pages = n_elts / a->entries_per_page;
n_pages           332 kernel/trace/tracing_map.c 	if (!a->n_pages)
n_pages           333 kernel/trace/tracing_map.c 		a->n_pages = 1;
n_pages           337 kernel/trace/tracing_map.c 	a->pages = kcalloc(a->n_pages, sizeof(void *), GFP_KERNEL);
n_pages           341 kernel/trace/tracing_map.c 	for (i = 0; i < a->n_pages; i++) {
n_pages           169 kernel/trace/tracing_map.h 	unsigned int n_pages;
n_pages           390 lib/scatterlist.c 				unsigned int n_pages, unsigned int offset,
n_pages           404 lib/scatterlist.c 	for (i = 1; i < n_pages; i++) {
n_pages           424 lib/scatterlist.c 		for (j = cur_page + 1; j < n_pages; j++) {
n_pages           465 lib/scatterlist.c 			      unsigned int n_pages, unsigned int offset,
n_pages           468 lib/scatterlist.c 	return __sg_alloc_table_from_pages(sgt, pages, n_pages, offset, size,
n_pages            45 mm/balloon_compaction.c 	size_t n_pages = 0;
n_pages            51 mm/balloon_compaction.c 		n_pages++;
n_pages            54 mm/balloon_compaction.c 	return n_pages;
n_pages            81 mm/balloon_compaction.c 	size_t n_pages = 0;
n_pages            85 mm/balloon_compaction.c 		if (n_pages == n_req_pages)
n_pages           106 mm/balloon_compaction.c 		n_pages++;
n_pages           110 mm/balloon_compaction.c 	return n_pages;
n_pages           181 mm/balloon_compaction.c 	int n_pages;
n_pages           183 mm/balloon_compaction.c 	n_pages = balloon_page_list_dequeue(b_dev_info, &pages, 1);
n_pages           185 mm/balloon_compaction.c 	if (n_pages != 1) {