nr_pages 193 arch/alpha/mm/init.c unsigned long nr_pages = 0; nr_pages 199 arch/alpha/mm/init.c nr_pages += crb->map[i].count; nr_pages 203 arch/alpha/mm/init.c console_remap_vm.size = nr_pages << PAGE_SHIFT; nr_pages 43 arch/arm/kernel/atags_compat.c unsigned long nr_pages; /* 4 */ nr_pages 104 arch/arm/kernel/atags_compat.c if (params->u1.s.nr_pages != 0x02000 && nr_pages 105 arch/arm/kernel/atags_compat.c params->u1.s.nr_pages != 0x04000 && nr_pages 106 arch/arm/kernel/atags_compat.c params->u1.s.nr_pages != 0x08000 && nr_pages 107 arch/arm/kernel/atags_compat.c params->u1.s.nr_pages != 0x10000) { nr_pages 110 arch/arm/kernel/atags_compat.c params->u1.s.nr_pages = 0x1000; /* 16MB */ nr_pages 158 arch/arm/kernel/atags_compat.c tag = memtag(tag, PHYS_OFFSET, params->u1.s.nr_pages * PAGE_SIZE); nr_pages 20 arch/arm/mach-rpc/include/mach/uncompress.h unsigned long nr_pages; nr_pages 116 arch/arm/mach-rpc/include/mach/uncompress.h unsigned int nr_pages = 0, page_size = PAGE_SIZE; nr_pages 130 arch/arm/mach-rpc/include/mach/uncompress.h nr_pages += (t->u.mem.size / PAGE_SIZE); nr_pages 134 arch/arm/mach-rpc/include/mach/uncompress.h nr_pages = params->nr_pages; nr_pages 179 arch/arm/mach-rpc/include/mach/uncompress.h if (nr_pages * page_size < 4096*1024) error("<4M of mem\n"); nr_pages 793 arch/arm/mm/dma-mapping.c unsigned long nr_pages = PAGE_ALIGN(size) >> PAGE_SHIFT; nr_pages 800 arch/arm/mm/dma-mapping.c if (off < nr_pages && nr_vma_pages <= (nr_pages - off)) { nr_pages 1527 arch/arm/mm/dma-mapping.c unsigned long nr_pages = PAGE_ALIGN(size) >> PAGE_SHIFT; nr_pages 1533 arch/arm/mm/dma-mapping.c if (vma->vm_pgoff >= nr_pages) nr_pages 1536 arch/arm/mm/dma-mapping.c err = vm_map_pages(vma, pages, nr_pages); nr_pages 24 arch/arm/xen/p2m.c unsigned long nr_pages; nr_pages 73 arch/arm/xen/p2m.c entry->pfn + entry->nr_pages > pfn) { nr_pages 122 arch/arm/xen/p2m.c unsigned long mfn, unsigned long nr_pages) nr_pages 134 arch/arm/xen/p2m.c p2m_entry->pfn + p2m_entry->nr_pages > pfn) { nr_pages 154 arch/arm/xen/p2m.c p2m_entry->nr_pages = nr_pages; nr_pages 1071 arch/arm64/mm/mmu.c unsigned long nr_pages = size >> PAGE_SHIFT; nr_pages 1080 arch/arm64/mm/mmu.c __remove_pages(start_pfn, nr_pages, altmap); nr_pages 96 arch/arm64/mm/pageattr.c for (i = 0; i < area->nr_pages; i++) { nr_pages 676 arch/ia64/mm/init.c unsigned long nr_pages = size >> PAGE_SHIFT; nr_pages 679 arch/ia64/mm/init.c ret = __add_pages(nid, start_pfn, nr_pages, restrictions); nr_pages 691 arch/ia64/mm/init.c unsigned long nr_pages = size >> PAGE_SHIFT; nr_pages 693 arch/ia64/mm/init.c __remove_pages(start_pfn, nr_pages, altmap); nr_pages 277 arch/mips/mm/init.c static int maar_res_walk(unsigned long start_pfn, unsigned long nr_pages, nr_pages 289 arch/mips/mm/init.c cfg->upper = ALIGN_DOWN(PFN_PHYS(start_pfn + nr_pages), maar_align) - 1; nr_pages 102 arch/mips/mm/ioremap.c static int __ioremap_check_ram(unsigned long start_pfn, unsigned long nr_pages, nr_pages 107 arch/mips/mm/ioremap.c for (i = 0; i < nr_pages; i++) { nr_pages 199 arch/powerpc/include/asm/kvm_ppc.h extern struct page *kvm_alloc_hpt_cma(unsigned long nr_pages); nr_pages 200 arch/powerpc/include/asm/kvm_ppc.h extern void kvm_free_hpt_cma(struct page *page, unsigned long nr_pages); nr_pages 72 arch/powerpc/kvm/book3s_hv_builtin.c struct page *kvm_alloc_hpt_cma(unsigned long nr_pages) nr_pages 74 arch/powerpc/kvm/book3s_hv_builtin.c VM_BUG_ON(order_base_2(nr_pages) < KVM_CMA_CHUNK_ORDER - PAGE_SHIFT); nr_pages 76 arch/powerpc/kvm/book3s_hv_builtin.c return cma_alloc(kvm_cma, nr_pages, order_base_2(HPT_ALIGN_PAGES), nr_pages 81 arch/powerpc/kvm/book3s_hv_builtin.c void kvm_free_hpt_cma(struct page *page, unsigned long nr_pages) nr_pages 83 arch/powerpc/kvm/book3s_hv_builtin.c cma_release(kvm_cma, page, nr_pages); nr_pages 842 arch/powerpc/mm/book3s64/radix_tlb.c unsigned long nr_pages = (end - start) >> page_shift; nr_pages 860 arch/powerpc/mm/book3s64/radix_tlb.c nr_pages > tlb_single_page_flush_ceiling); nr_pages 865 arch/powerpc/mm/book3s64/radix_tlb.c nr_pages > tlb_local_single_page_flush_ceiling); nr_pages 1080 arch/powerpc/mm/book3s64/radix_tlb.c unsigned long nr_pages = (end - start) >> page_shift; nr_pages 1098 arch/powerpc/mm/book3s64/radix_tlb.c nr_pages > tlb_single_page_flush_ceiling); nr_pages 1103 arch/powerpc/mm/book3s64/radix_tlb.c nr_pages > tlb_local_single_page_flush_ceiling); nr_pages 289 arch/powerpc/mm/init_64.c unsigned long nr_pages, addr; nr_pages 305 arch/powerpc/mm/init_64.c nr_pages = 1 << page_order; nr_pages 309 arch/powerpc/mm/init_64.c vmem_altmap_free(altmap, nr_pages); nr_pages 319 arch/powerpc/mm/init_64.c while (nr_pages--) nr_pages 132 arch/powerpc/mm/mem.c unsigned long nr_pages = size >> PAGE_SHIFT; nr_pages 147 arch/powerpc/mm/mem.c return __add_pages(nid, start_pfn, nr_pages, restrictions); nr_pages 154 arch/powerpc/mm/mem.c unsigned long nr_pages = size >> PAGE_SHIFT; nr_pages 157 arch/powerpc/mm/mem.c __remove_pages(start_pfn, nr_pages, altmap); nr_pages 71 arch/powerpc/platforms/powernv/memtrace.c static bool memtrace_offline_pages(u32 nid, u64 start_pfn, u64 nr_pages) nr_pages 74 arch/powerpc/platforms/powernv/memtrace.c const unsigned long size = PFN_PHYS(nr_pages); nr_pages 82 arch/powerpc/platforms/powernv/memtrace.c if (offline_pages(start_pfn, nr_pages)) { nr_pages 97 arch/powerpc/platforms/powernv/memtrace.c u64 start_pfn, end_pfn, nr_pages, pfn; nr_pages 106 arch/powerpc/platforms/powernv/memtrace.c nr_pages = size >> PAGE_SHIFT; nr_pages 109 arch/powerpc/platforms/powernv/memtrace.c end_pfn = round_down(end_pfn - nr_pages, nr_pages); nr_pages 112 arch/powerpc/platforms/powernv/memtrace.c for (base_pfn = end_pfn; base_pfn > start_pfn; base_pfn -= nr_pages) { nr_pages 113 arch/powerpc/platforms/powernv/memtrace.c if (memtrace_offline_pages(nid, base_pfn, nr_pages) == true) { nr_pages 120 arch/powerpc/platforms/powernv/memtrace.c end_pfn = base_pfn + nr_pages; nr_pages 501 arch/powerpc/platforms/pseries/cmm.c unsigned long end = start + (marg->nr_pages << PAGE_SHIFT); nr_pages 554 arch/powerpc/platforms/pseries/cmm.c unsigned long end_page = start_page + (marg->nr_pages << PAGE_SHIFT); nr_pages 560 arch/powerpc/platforms/pseries/cmm.c start_page, marg->nr_pages); nr_pages 1245 arch/powerpc/platforms/pseries/iommu.c arg->nr_pages, window->prop); nr_pages 1255 arch/powerpc/platforms/pseries/iommu.c arg->nr_pages, window->prop); nr_pages 370 arch/s390/include/asm/cio.h struct gen_pool *cio_gp_dma_create(struct device *dma_dev, int nr_pages); nr_pages 1692 arch/s390/kernel/perf_cpum_sf.c int nr_pages, bool snapshot) nr_pages 1699 arch/s390/kernel/perf_cpum_sf.c if (!nr_pages || !pages) nr_pages 1702 arch/s390/kernel/perf_cpum_sf.c if (nr_pages > CPUM_SF_MAX_SDB * CPUM_SF_SDB_DIAG_FACTOR) { nr_pages 1705 arch/s390/kernel/perf_cpum_sf.c nr_pages); nr_pages 1707 arch/s390/kernel/perf_cpum_sf.c } else if (nr_pages < CPUM_SF_MIN_SDB * CPUM_SF_SDB_DIAG_FACTOR) { nr_pages 1710 arch/s390/kernel/perf_cpum_sf.c nr_pages); nr_pages 1721 arch/s390/kernel/perf_cpum_sf.c n_sdbt = (nr_pages + CPUM_SF_SDB_PER_TABLE - 1) / CPUM_SF_SDB_PER_TABLE; nr_pages 1727 arch/s390/kernel/perf_cpum_sf.c aux->sdb_index = kmalloc_array(nr_pages, sizeof(void *), GFP_KERNEL); nr_pages 1743 arch/s390/kernel/perf_cpum_sf.c for (i = 0; i < nr_pages; i++, tail++) { nr_pages 1758 arch/s390/kernel/perf_cpum_sf.c sfb->num_sdb = nr_pages; nr_pages 623 arch/s390/kernel/setup.c if (arg->start_pfn + arg->nr_pages - 1 < PFN_DOWN(crashk_res.start)) nr_pages 798 arch/s390/kvm/gaccess.c unsigned long *pages, unsigned long nr_pages, nr_pages 806 arch/s390/kvm/gaccess.c while (nr_pages) { nr_pages 825 arch/s390/kvm/gaccess.c nr_pages--; nr_pages 834 arch/s390/kvm/gaccess.c unsigned long _len, nr_pages, gpa, idx; nr_pages 847 arch/s390/kvm/gaccess.c nr_pages = (((ga & ~PAGE_MASK) + len - 1) >> PAGE_SHIFT) + 1; nr_pages 849 arch/s390/kvm/gaccess.c if (nr_pages > ARRAY_SIZE(pages_array)) nr_pages 850 arch/s390/kvm/gaccess.c pages = vmalloc(array_size(nr_pages, sizeof(unsigned long))); nr_pages 856 arch/s390/kvm/gaccess.c rc = guest_page_range(vcpu, ga, ar, pages, nr_pages, asce, mode); nr_pages 857 arch/s390/kvm/gaccess.c for (idx = 0; idx < nr_pages && !rc; idx++) { nr_pages 870 arch/s390/kvm/gaccess.c if (nr_pages > ARRAY_SIZE(pages_array)) nr_pages 251 arch/s390/mm/init.c mem_data.end = mem_data.start + (arg->nr_pages << PAGE_SHIFT); nr_pages 293 arch/s390/mm/init.c unsigned long nr_pages = size >> PAGE_SHIFT; nr_pages 295 arch/s390/mm/init.c __remove_pages(start_pfn, nr_pages, altmap); nr_pages 136 arch/s390/pci/pci_dma.c unsigned int nr_pages = PAGE_ALIGN(size) >> PAGE_SHIFT; nr_pages 142 arch/s390/pci/pci_dma.c if (!nr_pages) nr_pages 151 arch/s390/pci/pci_dma.c for (i = 0; i < nr_pages; i++) { nr_pages 343 arch/s390/pci/pci_dma.c unsigned long nr_pages; nr_pages 348 arch/s390/pci/pci_dma.c nr_pages = iommu_num_pages(pa, size, PAGE_SIZE); nr_pages 349 arch/s390/pci/pci_dma.c dma_addr = dma_alloc_address(dev, nr_pages); nr_pages 356 arch/s390/pci/pci_dma.c size = nr_pages * PAGE_SIZE; nr_pages 365 arch/s390/pci/pci_dma.c atomic64_add(nr_pages, &zdev->mapped_pages); nr_pages 369 arch/s390/pci/pci_dma.c dma_free_address(dev, dma_addr, nr_pages); nr_pages 441 arch/s390/pci/pci_dma.c unsigned long nr_pages = PAGE_ALIGN(size) >> PAGE_SHIFT; nr_pages 449 arch/s390/pci/pci_dma.c dma_addr_base = dma_alloc_address(dev, nr_pages); nr_pages 471 arch/s390/pci/pci_dma.c atomic64_add(nr_pages, &zdev->mapped_pages); nr_pages 478 arch/s390/pci/pci_dma.c dma_free_address(dev, dma_addr_base, nr_pages); nr_pages 372 arch/sh/kernel/cpu/sh4/sq.c unsigned int nr_pages = 0x04000000 >> PAGE_SHIFT; nr_pages 373 arch/sh/kernel/cpu/sh4/sq.c unsigned int size = (nr_pages + (BITS_PER_LONG - 1)) / BITS_PER_LONG; nr_pages 412 arch/sh/mm/init.c unsigned long nr_pages = size >> PAGE_SHIFT; nr_pages 416 arch/sh/mm/init.c ret = __add_pages(nid, start_pfn, nr_pages, restrictions); nr_pages 436 arch/sh/mm/init.c unsigned long nr_pages = size >> PAGE_SHIFT; nr_pages 438 arch/sh/mm/init.c __remove_pages(start_pfn, nr_pages, altmap); nr_pages 53 arch/x86/events/intel/bts.c unsigned int nr_pages; nr_pages 81 arch/x86/events/intel/bts.c int nr_pages, bool overwrite) nr_pages 88 arch/x86/events/intel/bts.c size_t size = nr_pages << PAGE_SHIFT; nr_pages 92 arch/x86/events/intel/bts.c for (pg = 0, nbuf = 0; pg < nr_pages;) { nr_pages 108 arch/x86/events/intel/bts.c buf->nr_pages = nr_pages; nr_pages 321 arch/x86/events/intel/bts.c buf->nr_pages << PAGE_SHIFT); nr_pages 379 arch/x86/events/intel/bts.c head = handle->head & ((buf->nr_pages << PAGE_SHIFT) - 1); nr_pages 707 arch/x86/events/intel/pt.c p = virt_to_page(buf->data_pages[buf->nr_pages]); nr_pages 735 arch/x86/events/intel/pt.c buf->nr_pages += 1ul << order; nr_pages 816 arch/x86/events/intel/pt.c ((buf->nr_pages << PAGE_SHIFT) - 1)); nr_pages 818 arch/x86/events/intel/pt.c base += buf->nr_pages << PAGE_SHIFT; nr_pages 930 arch/x86/events/intel/pt.c if (WARN_ON_ONCE(pg >= buf->nr_pages)) nr_pages 1061 arch/x86/events/intel/pt.c idx &= buf->nr_pages - 1; nr_pages 1076 arch/x86/events/intel/pt.c idx &= buf->nr_pages - 1; nr_pages 1112 arch/x86/events/intel/pt.c head &= (buf->nr_pages << PAGE_SHIFT) - 1; nr_pages 1114 arch/x86/events/intel/pt.c pg = (head >> PAGE_SHIFT) & (buf->nr_pages - 1); nr_pages 1150 arch/x86/events/intel/pt.c unsigned long nr_pages, gfp_t gfp) nr_pages 1161 arch/x86/events/intel/pt.c while (buf->nr_pages < nr_pages) { nr_pages 1193 arch/x86/events/intel/pt.c int nr_pages, bool snapshot) nr_pages 1198 arch/x86/events/intel/pt.c if (!nr_pages) nr_pages 1216 arch/x86/events/intel/pt.c ret = pt_buffer_init_topa(buf, cpu, nr_pages, GFP_KERNEL); nr_pages 1494 arch/x86/events/intel/pt.c buf->nr_pages << PAGE_SHIFT); nr_pages 79 arch/x86/events/intel/pt.h unsigned long nr_pages; nr_pages 434 arch/x86/kernel/cpu/microcode/core.c unsigned long nr_pages = totalram_pages(); nr_pages 436 arch/x86/kernel/cpu/microcode/core.c if ((len >> PAGE_SHIFT) > nr_pages) { nr_pages 437 arch/x86/kernel/cpu/microcode/core.c pr_err("too much data (max %ld pages)\n", nr_pages); nr_pages 209 arch/x86/kernel/ldt.c int i, nr_pages; nr_pages 225 arch/x86/kernel/ldt.c nr_pages = DIV_ROUND_UP(ldt->nr_entries * LDT_ENTRY_SIZE, PAGE_SIZE); nr_pages 227 arch/x86/kernel/ldt.c for (i = 0; i < nr_pages; i++) { nr_pages 268 arch/x86/kernel/ldt.c int i, nr_pages; nr_pages 277 arch/x86/kernel/ldt.c nr_pages = DIV_ROUND_UP(ldt->nr_entries * LDT_ENTRY_SIZE, PAGE_SIZE); nr_pages 279 arch/x86/kernel/ldt.c for (i = 0; i < nr_pages; i++) { nr_pages 291 arch/x86/kernel/ldt.c flush_tlb_mm_range(mm, va, va + nr_pages * PAGE_SIZE, PAGE_SHIFT, false); nr_pages 616 arch/x86/kernel/machine_kexec_64.c unsigned int nr_pages; nr_pages 626 arch/x86/kernel/machine_kexec_64.c nr_pages = (end >> PAGE_SHIFT) - (start >> PAGE_SHIFT) + 1; nr_pages 628 arch/x86/kernel/machine_kexec_64.c return set_pages_ro(page, nr_pages); nr_pages 630 arch/x86/kernel/machine_kexec_64.c return set_pages_rw(page, nr_pages); nr_pages 6364 arch/x86/kvm/mmu.c unsigned long nr_pages = 0; nr_pages 6373 arch/x86/kvm/mmu.c nr_pages += memslot->npages; nr_pages 6376 arch/x86/kvm/mmu.c nr_mmu_pages = nr_pages * KVM_PERMILLE_MMU_PAGES / 1000; nr_pages 898 arch/x86/mm/init.c u64 nr_pages = 0, nr_free_pages = 0; nr_pages 908 arch/x86/mm/init.c nr_pages = 0; nr_pages 913 arch/x86/mm/init.c nr_pages += end_pfn - start_pfn; nr_pages 930 arch/x86/mm/init.c set_dma_reserve(nr_pages - nr_free_pages); nr_pages 858 arch/x86/mm/init_32.c unsigned long nr_pages = size >> PAGE_SHIFT; nr_pages 860 arch/x86/mm/init_32.c return __add_pages(nid, start_pfn, nr_pages, restrictions); nr_pages 867 arch/x86/mm/init_32.c unsigned long nr_pages = size >> PAGE_SHIFT; nr_pages 869 arch/x86/mm/init_32.c __remove_pages(start_pfn, nr_pages, altmap); nr_pages 846 arch/x86/mm/init_64.c int add_pages(int nid, unsigned long start_pfn, unsigned long nr_pages, nr_pages 851 arch/x86/mm/init_64.c ret = __add_pages(nid, start_pfn, nr_pages, restrictions); nr_pages 856 arch/x86/mm/init_64.c nr_pages << PAGE_SHIFT); nr_pages 865 arch/x86/mm/init_64.c unsigned long nr_pages = size >> PAGE_SHIFT; nr_pages 869 arch/x86/mm/init_64.c return add_pages(nid, start_pfn, nr_pages, restrictions); nr_pages 877 arch/x86/mm/init_64.c unsigned int nr_pages = 1 << order; nr_pages 885 arch/x86/mm/init_64.c while (nr_pages--) nr_pages 888 arch/x86/mm/init_64.c while (nr_pages--) nr_pages 1214 arch/x86/mm/init_64.c unsigned long nr_pages = size >> PAGE_SHIFT; nr_pages 1216 arch/x86/mm/init_64.c __remove_pages(start_pfn, nr_pages, altmap); nr_pages 1536 arch/x86/mm/init_64.c struct page *start_page, unsigned long nr_pages) nr_pages 1539 arch/x86/mm/init_64.c unsigned long end = (unsigned long)(start_page + nr_pages); nr_pages 421 arch/x86/mm/mpx.c int nr_pages = 1; nr_pages 423 arch/x86/mm/mpx.c gup_ret = get_user_pages((unsigned long)addr, nr_pages, nr_pages 1306 arch/x86/xen/enlighten_pv.c xen_start_info->nr_pages); nr_pages 1227 arch/x86/xen/mmu_pv.c size = PAGE_ALIGN(xen_start_info->nr_pages * sizeof(unsigned long)); nr_pages 1248 arch/x86/xen/mmu_pv.c size = PAGE_ALIGN(xen_start_info->nr_pages * nr_pages 2067 arch/x86/xen/mmu_pv.c size = PAGE_ALIGN(xen_start_info->nr_pages * sizeof(unsigned long)); nr_pages 300 arch/x86/xen/p2m.c xen_p2m_size = ALIGN(xen_start_info->nr_pages, P2M_PER_PAGE); nr_pages 302 arch/x86/xen/p2m.c for (pfn = xen_start_info->nr_pages; pfn < xen_p2m_size; pfn++) nr_pages 255 arch/x86/xen/setup.c unsigned long end_pfn, unsigned long nr_pages) nr_pages 263 arch/x86/xen/setup.c end = min(end_pfn, nr_pages); nr_pages 388 arch/x86/xen/setup.c unsigned long start_pfn, unsigned long end_pfn, unsigned long nr_pages, nr_pages 396 arch/x86/xen/setup.c remap_pfn = nr_pages; nr_pages 405 arch/x86/xen/setup.c if (cur_pfn >= nr_pages) { nr_pages 410 arch/x86/xen/setup.c if (cur_pfn + size > nr_pages) nr_pages 411 arch/x86/xen/setup.c size = nr_pages - cur_pfn; nr_pages 417 arch/x86/xen/setup.c cur_pfn + left, nr_pages); nr_pages 444 arch/x86/xen/setup.c unsigned long start_pfn, unsigned long end_pfn, unsigned long nr_pages, nr_pages 447 arch/x86/xen/setup.c if (start_pfn >= nr_pages) nr_pages 450 arch/x86/xen/setup.c return remap_pages + min(end_pfn, nr_pages) - start_pfn; nr_pages 453 arch/x86/xen/setup.c static unsigned long __init xen_foreach_remap_area(unsigned long nr_pages, nr_pages 455 arch/x86/xen/setup.c unsigned long nr_pages, unsigned long last_val)) nr_pages 483 arch/x86/xen/setup.c ret_val = func(start_pfn, end_pfn, nr_pages, nr_pages 713 arch/x86/xen/setup.c size = PFN_ALIGN(xen_start_info->nr_pages * nr_pages 754 arch/x86/xen/setup.c max_pfn = min(max_pfn, xen_start_info->nr_pages); nr_pages 206 block/bio-integrity.c unsigned int len, nr_pages; nr_pages 246 block/bio-integrity.c nr_pages = end - start; nr_pages 249 block/bio-integrity.c bip = bio_integrity_alloc(bio, GFP_NOIO, nr_pages); nr_pages 266 block/bio-integrity.c for (i = 0 ; i < nr_pages ; i++) { nr_pages 922 block/bio.c unsigned short nr_pages = bio->bi_max_vecs - bio->bi_vcnt; nr_pages 939 block/bio.c size = iov_iter_get_pages(iter, pages, LONG_MAX, nr_pages, &offset); nr_pages 1282 block/bio.c int nr_pages; nr_pages 1297 block/bio.c nr_pages = DIV_ROUND_UP(offset + len, PAGE_SIZE); nr_pages 1298 block/bio.c if (nr_pages > BIO_MAX_PAGES) nr_pages 1299 block/bio.c nr_pages = BIO_MAX_PAGES; nr_pages 1302 block/bio.c bio = bio_kmalloc(gfp_mask, nr_pages); nr_pages 1309 block/bio.c nr_pages = 1 << map_data->page_order; nr_pages 1321 block/bio.c if (i == map_data->nr_entries * nr_pages) { nr_pages 1326 block/bio.c page = map_data->pages[i / nr_pages]; nr_pages 1327 block/bio.c page += (i % nr_pages); nr_pages 1523 block/bio.c const int nr_pages = end - start; nr_pages 1529 block/bio.c bio = bio_kmalloc(gfp_mask, nr_pages); nr_pages 1539 block/bio.c for (i = 0; i < nr_pages; i++) { nr_pages 1607 block/bio.c int nr_pages = 0; nr_pages 1615 block/bio.c nr_pages = end - start; nr_pages 1616 block/bio.c bio = bio_kmalloc(gfp_mask, nr_pages); nr_pages 13 block/blk-lib.c struct bio *blk_next_bio(struct bio *bio, unsigned int nr_pages, gfp_t gfp) nr_pages 15 block/blk-lib.c struct bio *new = bio_alloc(gfp, nr_pages); nr_pages 352 block/blk.h struct bio *blk_next_bio(struct bio *bio, unsigned int nr_pages, gfp_t gfp); nr_pages 70 drivers/base/firmware_loader/firmware.h int nr_pages; nr_pages 278 drivers/base/firmware_loader/main.c for (i = 0; i < fw_priv->nr_pages; i++) nr_pages 283 drivers/base/firmware_loader/main.c fw_priv->nr_pages = 0; nr_pages 307 drivers/base/firmware_loader/main.c while (fw_priv->nr_pages < pages_needed) { nr_pages 308 drivers/base/firmware_loader/main.c fw_priv->pages[fw_priv->nr_pages] = nr_pages 311 drivers/base/firmware_loader/main.c if (!fw_priv->pages[fw_priv->nr_pages]) nr_pages 313 drivers/base/firmware_loader/main.c fw_priv->nr_pages++; nr_pages 326 drivers/base/firmware_loader/main.c fw_priv->data = vmap(fw_priv->pages, fw_priv->nr_pages, 0, nr_pages 400 drivers/base/firmware_loader/main.c if (fw_grow_paged_buf(fw_priv, fw_priv->nr_pages + 1)) { nr_pages 406 drivers/base/firmware_loader/main.c page = fw_priv->pages[fw_priv->nr_pages - 1]; nr_pages 217 drivers/base/memory.c unsigned long nr_pages = PAGES_PER_SECTION * sections_per_block; nr_pages 227 drivers/base/memory.c ret = online_pages(start_pfn, nr_pages, online_type); nr_pages 230 drivers/base/memory.c ret = offline_pages(start_pfn, nr_pages); nr_pages 366 drivers/base/memory.c unsigned long nr_pages, int online_type, nr_pages 371 drivers/base/memory.c zone = zone_for_pfn_range(online_type, nid, start_pfn, nr_pages); nr_pages 383 drivers/base/memory.c unsigned long nr_pages = PAGES_PER_SECTION * sections_per_block; nr_pages 397 drivers/base/memory.c if (!test_pages_in_a_zone(start_pfn, start_pfn + nr_pages, nr_pages 406 drivers/base/memory.c default_zone = zone_for_pfn_range(MMOP_ONLINE_KEEP, nid, start_pfn, nr_pages); nr_pages 409 drivers/base/memory.c print_allowed_zone(buf, nid, start_pfn, nr_pages, MMOP_ONLINE_KERNEL, nr_pages 411 drivers/base/memory.c print_allowed_zone(buf, nid, start_pfn, nr_pages, MMOP_ONLINE_MOVABLE, nr_pages 136 drivers/block/brd.c int nr_pages; nr_pages 141 drivers/block/brd.c nr_pages = radix_tree_gang_lookup(&brd->brd_pages, nr_pages 144 drivers/block/brd.c for (i = 0; i < nr_pages; i++) { nr_pages 167 drivers/block/brd.c } while (nr_pages == FREE_BATCH); nr_pages 364 drivers/block/drbd/drbd_receiver.c unsigned nr_pages = (payload_size + PAGE_SIZE -1) >> PAGE_SHIFT; nr_pages 376 drivers/block/drbd/drbd_receiver.c if (nr_pages) { nr_pages 377 drivers/block/drbd/drbd_receiver.c page = drbd_alloc_pages(peer_device, nr_pages, nr_pages 1647 drivers/block/drbd/drbd_receiver.c unsigned nr_pages = (data_size + PAGE_SIZE -1) >> PAGE_SHIFT; nr_pages 1689 drivers/block/drbd/drbd_receiver.c bio = bio_alloc(GFP_NOIO, nr_pages); nr_pages 1691 drivers/block/drbd/drbd_receiver.c drbd_err(device, "submit_ee: Allocation of a bio failed (nr_pages=%u)\n", nr_pages); nr_pages 1711 drivers/block/drbd/drbd_receiver.c --nr_pages; nr_pages 735 drivers/block/null_blk_main.c int nr_pages; nr_pages 744 drivers/block/null_blk_main.c nr_pages = radix_tree_gang_lookup(root, nr_pages 747 drivers/block/null_blk_main.c for (i = 0; i < nr_pages; i++) { nr_pages 755 drivers/block/null_blk_main.c } while (nr_pages == FREE_BATCH); nr_pages 880 drivers/block/null_blk_main.c int i, err, nr_pages; nr_pages 889 drivers/block/null_blk_main.c nr_pages = radix_tree_gang_lookup(&nullb->dev->cache, nr_pages 895 drivers/block/null_blk_main.c for (i = 0; i < nr_pages; i++) { nr_pages 908 drivers/block/null_blk_main.c for (i = 0; i < nr_pages; i++) { nr_pages 919 drivers/block/null_blk_main.c if (nr_pages == 0) nr_pages 291 drivers/block/zram/zram_drv.c unsigned long nr_pages = zram->disksize >> PAGE_SHIFT; nr_pages 303 drivers/block/zram/zram_drv.c for (index = 0; index < nr_pages; index++) { nr_pages 452 drivers/block/zram/zram_drv.c unsigned long nr_pages, *bitmap = NULL; nr_pages 497 drivers/block/zram/zram_drv.c nr_pages = i_size_read(inode) >> PAGE_SHIFT; nr_pages 498 drivers/block/zram/zram_drv.c bitmap_sz = BITS_TO_LONGS(nr_pages) * sizeof(long); nr_pages 516 drivers/block/zram/zram_drv.c zram->nr_pages = nr_pages; nr_pages 557 drivers/block/zram/zram_drv.c blk_idx = find_next_zero_bit(zram->bitmap, zram->nr_pages, blk_idx); nr_pages 558 drivers/block/zram/zram_drv.c if (blk_idx == zram->nr_pages) nr_pages 624 drivers/block/zram/zram_drv.c unsigned long nr_pages = zram->disksize >> PAGE_SHIFT; nr_pages 657 drivers/block/zram/zram_drv.c for (index = 0; index < nr_pages; index++) { nr_pages 868 drivers/block/zram/zram_drv.c unsigned long nr_pages = zram->disksize >> PAGE_SHIFT; nr_pages 882 drivers/block/zram/zram_drv.c for (index = *ppos; index < nr_pages; index++) { nr_pages 123 drivers/block/zram/zram_drv.h unsigned long nr_pages; nr_pages 296 drivers/dax/super.c long dax_direct_access(struct dax_device *dax_dev, pgoff_t pgoff, long nr_pages, nr_pages 307 drivers/dax/super.c if (nr_pages < 0) nr_pages 308 drivers/dax/super.c return nr_pages; nr_pages 310 drivers/dax/super.c avail = dax_dev->ops->direct_access(dax_dev, pgoff, nr_pages, nr_pages 314 drivers/dax/super.c return min(avail, nr_pages); nr_pages 391 drivers/edac/altera_edac.c dimm->nr_pages = ((mem_size - 1) >> PAGE_SHIFT) + 1; nr_pages 2883 drivers/edac/amd64_edac.c u32 cs_mode, nr_pages; nr_pages 2892 drivers/edac/amd64_edac.c nr_pages = pvt->ops->dbam_to_cs(pvt, dct, cs_mode, csrow_nr); nr_pages 2893 drivers/edac/amd64_edac.c nr_pages <<= 20 - PAGE_SHIFT; nr_pages 2897 drivers/edac/amd64_edac.c edac_dbg(0, "nr_pages/channel: %u\n", nr_pages); nr_pages 2899 drivers/edac/amd64_edac.c return nr_pages; nr_pages 2935 drivers/edac/amd64_edac.c dimm->nr_pages = get_csrow_nr_pages(pvt, umc, cs); nr_pages 2957 drivers/edac/amd64_edac.c int nr_pages = 0; nr_pages 2991 drivers/edac/amd64_edac.c nr_pages = get_csrow_nr_pages(pvt, 0, i); nr_pages 2992 drivers/edac/amd64_edac.c csrow->channels[0]->dimm->nr_pages = nr_pages; nr_pages 2999 drivers/edac/amd64_edac.c csrow->channels[1]->dimm->nr_pages = row_dct1_pages; nr_pages 3000 drivers/edac/amd64_edac.c nr_pages += row_dct1_pages; nr_pages 3003 drivers/edac/amd64_edac.c edac_dbg(1, "Total csrow%d pages: %u\n", i, nr_pages); nr_pages 210 drivers/edac/amd76x_edac.c dimm->nr_pages = (mba_mask + 1) >> PAGE_SHIFT; nr_pages 211 drivers/edac/amd76x_edac.c csrow->last_page = csrow->first_page + dimm->nr_pages - 1; nr_pages 213 drivers/edac/amd76x_edac.c dimm->grain = dimm->nr_pages << PAGE_SHIFT; nr_pages 251 drivers/edac/armada_xp_edac.c dimm->nr_pages = 524288; nr_pages 254 drivers/edac/armada_xp_edac.c dimm->nr_pages = 65536; nr_pages 257 drivers/edac/armada_xp_edac.c dimm->nr_pages = 131072; nr_pages 260 drivers/edac/armada_xp_edac.c dimm->nr_pages = 262144; nr_pages 263 drivers/edac/armada_xp_edac.c dimm->nr_pages = 1048576; nr_pages 266 drivers/edac/armada_xp_edac.c dimm->nr_pages = 2097152; nr_pages 234 drivers/edac/aspeed_edac.c u32 nr_pages, dram_type; nr_pages 261 drivers/edac/aspeed_edac.c nr_pages = resource_size(&r) >> PAGE_SHIFT; nr_pages 262 drivers/edac/aspeed_edac.c csrow->last_page = csrow->first_page + nr_pages - 1; nr_pages 270 drivers/edac/aspeed_edac.c dimm->nr_pages = nr_pages / csrow->nr_channels; nr_pages 273 drivers/edac/aspeed_edac.c csrow->first_page, nr_pages); nr_pages 212 drivers/edac/bluefield_edac.c dimm->nr_pages = nr_pages 135 drivers/edac/cell_edac.c u32 nr_pages; nr_pages 149 drivers/edac/cell_edac.c nr_pages = resource_size(&r) >> PAGE_SHIFT; nr_pages 150 drivers/edac/cell_edac.c csrow->last_page = csrow->first_page + nr_pages - 1; nr_pages 156 drivers/edac/cell_edac.c dimm->nr_pages = nr_pages / csrow->nr_channels; nr_pages 162 drivers/edac/cell_edac.c csrow->first_page, nr_pages); nr_pages 322 drivers/edac/cpc925_edac.c unsigned long row_size, nr_pages, last_nr_pages = 0; nr_pages 341 drivers/edac/cpc925_edac.c nr_pages = row_size >> PAGE_SHIFT; nr_pages 342 drivers/edac/cpc925_edac.c csrow->last_page = csrow->first_page + nr_pages - 1; nr_pages 370 drivers/edac/cpc925_edac.c dimm->nr_pages = nr_pages / csrow->nr_channels; nr_pages 1076 drivers/edac/e752x_edac.c u32 dra, drc, cumul_size, i, nr_pages; nr_pages 1109 drivers/edac/e752x_edac.c nr_pages = cumul_size - last_cumul_size; nr_pages 1130 drivers/edac/e752x_edac.c dimm->nr_pages = nr_pages / csrow->nr_channels; nr_pages 360 drivers/edac/e7xxx_edac.c u32 dra, cumul_size, nr_pages; nr_pages 391 drivers/edac/e7xxx_edac.c nr_pages = cumul_size - last_cumul_size; nr_pages 412 drivers/edac/e7xxx_edac.c dimm->nr_pages = nr_pages / (drc_chan + 1); nr_pages 159 drivers/edac/edac_mc.c edac_dbg(4, " dimm->nr_pages = 0x%x\n", dimm->nr_pages); nr_pages 161 drivers/edac/edac_mc.c edac_dbg(4, " dimm->nr_pages = 0x%x\n", dimm->nr_pages); nr_pages 715 drivers/edac/edac_mc.c u32 nr_pages = 0; nr_pages 719 drivers/edac/edac_mc.c nr_pages += csrow->channels[j]->dimm->nr_pages; nr_pages 720 drivers/edac/edac_mc.c if (!nr_pages) nr_pages 724 drivers/edac/edac_mc.c if (csrow->channels[j]->dimm->nr_pages) nr_pages 728 drivers/edac/edac_mc.c if (mci->dimms[i]->nr_pages) nr_pages 865 drivers/edac/edac_mc.c n += dimm->nr_pages; nr_pages 1167 drivers/edac/edac_mc.c if (e->enable_per_layer_report && dimm->nr_pages) { nr_pages 165 drivers/edac/edac_mc_sysfs.c u32 nr_pages = 0; nr_pages 168 drivers/edac/edac_mc_sysfs.c nr_pages += csrow->channels[i]->dimm->nr_pages; nr_pages 169 drivers/edac/edac_mc_sysfs.c return sprintf(data, "%u\n", PAGES_TO_MiB(nr_pages)); nr_pages 367 drivers/edac/edac_mc_sysfs.c if (!csrow->channels[idx]->dimm->nr_pages) nr_pages 392 drivers/edac/edac_mc_sysfs.c int chan, nr_pages = 0; nr_pages 395 drivers/edac/edac_mc_sysfs.c nr_pages += csrow->channels[chan]->dimm->nr_pages; nr_pages 397 drivers/edac/edac_mc_sysfs.c return nr_pages; nr_pages 522 drivers/edac/edac_mc_sysfs.c return sprintf(data, "%u\n", PAGES_TO_MiB(dimm->nr_pages)); nr_pages 819 drivers/edac/edac_mc_sysfs.c total_pages += dimm->nr_pages; nr_pages 948 drivers/edac/edac_mc_sysfs.c if (!dimm->nr_pages) nr_pages 968 drivers/edac/edac_mc_sysfs.c if (!dimm->nr_pages) nr_pages 996 drivers/edac/edac_mc_sysfs.c if (dimm->nr_pages == 0) nr_pages 466 drivers/edac/fsl_ddr_edac.c dimm->nr_pages = end + 1 - start; nr_pages 117 drivers/edac/ghes_edac.c dimm->nr_pages = MiB_TO_PAGES(32);/* Unknown */ nr_pages 119 drivers/edac/ghes_edac.c dimm->nr_pages = MiB_TO_PAGES(entry->extended_size); nr_pages 122 drivers/edac/ghes_edac.c dimm->nr_pages = MiB_TO_PAGES((entry->size & 0x7fff) << 10); nr_pages 124 drivers/edac/ghes_edac.c dimm->nr_pages = MiB_TO_PAGES(entry->size); nr_pages 188 drivers/edac/ghes_edac.c if (dimm->nr_pages) { nr_pages 191 drivers/edac/ghes_edac.c PAGES_TO_MiB(dimm->nr_pages), nr_pages 556 drivers/edac/ghes_edac.c dimm->nr_pages = 1; nr_pages 222 drivers/edac/highbank_mc_edac.c dimm->nr_pages = (~0UL >> PAGE_SHIFT) + 1; nr_pages 315 drivers/edac/i3000_edac.c unsigned long last_cumul_size, nr_pages; nr_pages 405 drivers/edac/i3000_edac.c nr_pages = cumul_size - last_cumul_size; nr_pages 411 drivers/edac/i3000_edac.c dimm->nr_pages = nr_pages / nr_channels; nr_pages 392 drivers/edac/i3200_edac.c unsigned long nr_pages; nr_pages 398 drivers/edac/i3200_edac.c nr_pages = drb_to_nr_pages(drbs, stacked, j, i); nr_pages 399 drivers/edac/i3200_edac.c if (nr_pages == 0) nr_pages 403 drivers/edac/i3200_edac.c stacked ? " (stacked)" : "", PAGES_TO_MiB(nr_pages)); nr_pages 405 drivers/edac/i3200_edac.c dimm->nr_pages = nr_pages; nr_pages 406 drivers/edac/i3200_edac.c dimm->grain = nr_pages << PAGE_SHIFT; nr_pages 1295 drivers/edac/i5000_edac.c dimm->nr_pages = csrow_megs << 8; nr_pages 866 drivers/edac/i5100_edac.c dimm->nr_pages = npages; nr_pages 1208 drivers/edac/i5400_edac.c dimm->nr_pages = size_mb << 8; nr_pages 649 drivers/edac/i7300_edac.c dimm->nr_pages = MiB_TO_PAGES(dinfo->megabytes); nr_pages 605 drivers/edac/i7core_edac.c dimm->nr_pages = npages; nr_pages 220 drivers/edac/i82443bxgx_edac.c dimm->nr_pages = csrow->last_page - csrow->first_page + 1; nr_pages 176 drivers/edac/i82860_edac.c dimm->nr_pages = cumul_size - last_cumul_size; nr_pages 350 drivers/edac/i82875p_edac.c u32 cumul_size, nr_pages; nr_pages 373 drivers/edac/i82875p_edac.c nr_pages = cumul_size - last_cumul_size; nr_pages 379 drivers/edac/i82875p_edac.c dimm->nr_pages = nr_pages / nr_chans; nr_pages 367 drivers/edac/i82975x_edac.c u32 cumul_size, nr_pages; nr_pages 397 drivers/edac/i82975x_edac.c nr_pages = cumul_size - last_cumul_size; nr_pages 398 drivers/edac/i82975x_edac.c if (!nr_pages) nr_pages 410 drivers/edac/i82975x_edac.c dimm->nr_pages = nr_pages / csrow->nr_channels; nr_pages 485 drivers/edac/ie31200_edac.c unsigned long nr_pages; nr_pages 487 drivers/edac/ie31200_edac.c nr_pages = IE31200_PAGES(dimm_info[j][i].size, skl); nr_pages 488 drivers/edac/ie31200_edac.c if (nr_pages == 0) nr_pages 492 drivers/edac/ie31200_edac.c nr_pages = nr_pages / 2; nr_pages 496 drivers/edac/ie31200_edac.c dimm->nr_pages = nr_pages; nr_pages 497 drivers/edac/ie31200_edac.c edac_dbg(0, "set nr pages: 0x%lx\n", nr_pages); nr_pages 508 drivers/edac/ie31200_edac.c dimm->nr_pages = nr_pages; nr_pages 509 drivers/edac/ie31200_edac.c edac_dbg(0, "set nr pages: 0x%lx\n", nr_pages); nr_pages 672 drivers/edac/mv64x60_edac.c dimm->nr_pages = pdata->total_mem >> PAGE_SHIFT; nr_pages 145 drivers/edac/pasemi_edac.c dimm->nr_pages = 128 << (20 - PAGE_SHIFT); nr_pages 148 drivers/edac/pasemi_edac.c dimm->nr_pages = 256 << (20 - PAGE_SHIFT); nr_pages 152 drivers/edac/pasemi_edac.c dimm->nr_pages = 512 << (20 - PAGE_SHIFT); nr_pages 155 drivers/edac/pasemi_edac.c dimm->nr_pages = 1024 << (20 - PAGE_SHIFT); nr_pages 158 drivers/edac/pasemi_edac.c dimm->nr_pages = 2048 << (20 - PAGE_SHIFT); nr_pages 168 drivers/edac/pasemi_edac.c csrow->last_page = csrow->first_page + dimm->nr_pages - 1; nr_pages 169 drivers/edac/pasemi_edac.c last_page_in_mmc += dimm->nr_pages; nr_pages 1256 drivers/edac/pnd2_edac.c dimm->nr_pages = MiB_TO_PAGES(capacity >> (20 - 3)); nr_pages 1322 drivers/edac/pnd2_edac.c dimm->nr_pages = MiB_TO_PAGES(capacity >> (20 - 3)); nr_pages 897 drivers/edac/ppc4xx_edac.c u32 mbxcf, size, nr_pages; nr_pages 948 drivers/edac/ppc4xx_edac.c nr_pages = SDRAM_MBCF_SZ_TO_PAGES(size); nr_pages 974 drivers/edac/ppc4xx_edac.c dimm->nr_pages = nr_pages / csi->nr_channels; nr_pages 255 drivers/edac/r82600_edac.c dimm->nr_pages = csrow->last_page - csrow->first_page + 1; nr_pages 1661 drivers/edac/sb_edac.c dimm->nr_pages = npages; nr_pages 312 drivers/edac/skx_common.c dimm->nr_pages = npages; nr_pages 356 drivers/edac/skx_common.c dimm->nr_pages = size >> PAGE_SHIFT; nr_pages 363 drivers/edac/skx_common.c imc->mc, chan, dimmno, size >> 20, dimm->nr_pages); nr_pages 787 drivers/edac/synopsys_edac.c dimm->nr_pages = (size >> PAGE_SHIFT) / csi->nr_channels; nr_pages 181 drivers/edac/ti_edac.c dimm->nr_pages = memsize >> PAGE_SHIFT; nr_pages 373 drivers/edac/x38_edac.c unsigned long nr_pages; nr_pages 376 drivers/edac/x38_edac.c nr_pages = drb_to_nr_pages(drbs, stacked, nr_pages 380 drivers/edac/x38_edac.c if (nr_pages == 0) nr_pages 386 drivers/edac/x38_edac.c dimm->nr_pages = nr_pages / x38_channel_num; nr_pages 387 drivers/edac/x38_edac.c dimm->grain = nr_pages << PAGE_SHIFT; nr_pages 71 drivers/firmware/efi/libstub/arm32-stub.c unsigned long nr_pages, map_size, desc_size, buff_size; nr_pages 94 drivers/firmware/efi/libstub/arm32-stub.c nr_pages = MAX_UNCOMP_KERNEL_SIZE / EFI_PAGE_SIZE; nr_pages 96 drivers/firmware/efi/libstub/arm32-stub.c EFI_BOOT_SERVICES_DATA, nr_pages, &alloc_addr); nr_pages 177 drivers/firmware/efi/libstub/efi-stub-helper.c unsigned long nr_pages; nr_pages 203 drivers/firmware/efi/libstub/efi-stub-helper.c nr_pages = size / EFI_PAGE_SIZE; nr_pages 214 drivers/firmware/efi/libstub/efi-stub-helper.c if (desc->num_pages < nr_pages) nr_pages 247 drivers/firmware/efi/libstub/efi-stub-helper.c nr_pages, &max_addr); nr_pages 272 drivers/firmware/efi/libstub/efi-stub-helper.c unsigned long nr_pages; nr_pages 297 drivers/firmware/efi/libstub/efi-stub-helper.c nr_pages = size / EFI_PAGE_SIZE; nr_pages 308 drivers/firmware/efi/libstub/efi-stub-helper.c if (desc->num_pages < nr_pages) nr_pages 323 drivers/firmware/efi/libstub/efi-stub-helper.c nr_pages, &start); nr_pages 341 drivers/firmware/efi/libstub/efi-stub-helper.c unsigned long nr_pages; nr_pages 346 drivers/firmware/efi/libstub/efi-stub-helper.c nr_pages = round_up(size, EFI_ALLOC_ALIGN) / EFI_PAGE_SIZE; nr_pages 347 drivers/firmware/efi/libstub/efi-stub-helper.c efi_call_early(free_pages, addr, nr_pages); nr_pages 702 drivers/firmware/efi/libstub/efi-stub-helper.c unsigned long nr_pages; nr_pages 720 drivers/firmware/efi/libstub/efi-stub-helper.c nr_pages = round_up(alloc_size, EFI_ALLOC_ALIGN) / EFI_PAGE_SIZE; nr_pages 723 drivers/firmware/efi/libstub/efi-stub-helper.c nr_pages, &efi_addr); nr_pages 254 drivers/fpga/fpga-mgr.c int nr_pages; nr_pages 270 drivers/fpga/fpga-mgr.c nr_pages = DIV_ROUND_UP((unsigned long)buf + count, PAGE_SIZE) - nr_pages 272 drivers/fpga/fpga-mgr.c pages = kmalloc_array(nr_pages, sizeof(struct page *), GFP_KERNEL); nr_pages 277 drivers/fpga/fpga-mgr.c for (index = 0; index < nr_pages; index++) { nr_pages 793 drivers/gpu/drm/drm_prime.c struct sg_table *drm_prime_pages_to_sg(struct page **pages, unsigned int nr_pages) nr_pages 804 drivers/gpu/drm/drm_prime.c ret = sg_alloc_table_from_pages(sg, pages, nr_pages, 0, nr_pages 805 drivers/gpu/drm/drm_prime.c nr_pages << PAGE_SHIFT, GFP_KERNEL); nr_pages 79 drivers/gpu/drm/exynos/exynos_drm_fbdev.c unsigned int nr_pages; nr_pages 93 drivers/gpu/drm/exynos/exynos_drm_fbdev.c nr_pages = exynos_gem->size >> PAGE_SHIFT; nr_pages 95 drivers/gpu/drm/exynos/exynos_drm_fbdev.c exynos_gem->kvaddr = (void __iomem *) vmap(exynos_gem->pages, nr_pages, nr_pages 24 drivers/gpu/drm/exynos/exynos_drm_gem.c unsigned int nr_pages; nr_pages 56 drivers/gpu/drm/exynos/exynos_drm_gem.c nr_pages = exynos_gem->size >> PAGE_SHIFT; nr_pages 58 drivers/gpu/drm/exynos/exynos_drm_gem.c exynos_gem->pages = kvmalloc_array(nr_pages, sizeof(struct page *), nr_pages 82 drivers/gpu/drm/exynos/exynos_drm_gem.c nr_pages)) { nr_pages 377 drivers/gpu/drm/savage/savage_bci.c unsigned int nr_pages = (n - rest + SAVAGE_DMA_PAGE_SIZE - 1) / nr_pages 383 drivers/gpu/drm/savage/savage_bci.c cur, dev_priv->dma_pages[cur].used, n, rest, nr_pages); nr_pages 385 drivers/gpu/drm/savage/savage_bci.c if (cur + nr_pages < dev_priv->nr_dma_pages) { nr_pages 395 drivers/gpu/drm/savage/savage_bci.c nr_pages = nr_pages 405 drivers/gpu/drm/savage/savage_bci.c for (i = cur; nr_pages > 0; ++i, --nr_pages) { nr_pages 208 drivers/gpu/drm/virtio/virtgpu_object.c int nr_pages = bo->tbo.num_pages; nr_pages 229 drivers/gpu/drm/virtio/virtgpu_object.c ret = __sg_alloc_table_from_pages(bo->pages, pages, nr_pages, 0, nr_pages 230 drivers/gpu/drm/virtio/virtgpu_object.c nr_pages << PAGE_SHIFT, nr_pages 591 drivers/hv/hv_balloon.c unsigned long nr_pages) nr_pages 597 drivers/hv/hv_balloon.c while (pfn < start_pfn + nr_pages) { nr_pages 606 drivers/hv/hv_balloon.c (pfn < start_pfn + nr_pages)) { nr_pages 644 drivers/hv/hv_balloon.c mem->nr_pages); nr_pages 1091 drivers/hv/hv_balloon.c unsigned long nr_pages = totalram_pages(); nr_pages 1104 drivers/hv/hv_balloon.c if (nr_pages < MB2PAGES(128)) nr_pages 1105 drivers/hv/hv_balloon.c min_pages = MB2PAGES(8) + (nr_pages >> 1); nr_pages 1106 drivers/hv/hv_balloon.c else if (nr_pages < MB2PAGES(512)) nr_pages 1107 drivers/hv/hv_balloon.c min_pages = MB2PAGES(40) + (nr_pages >> 2); nr_pages 1108 drivers/hv/hv_balloon.c else if (nr_pages < MB2PAGES(2048)) nr_pages 1109 drivers/hv/hv_balloon.c min_pages = MB2PAGES(104) + (nr_pages >> 3); nr_pages 1110 drivers/hv/hv_balloon.c else if (nr_pages < MB2PAGES(8192)) nr_pages 1111 drivers/hv/hv_balloon.c min_pages = MB2PAGES(232) + (nr_pages >> 4); nr_pages 1113 drivers/hv/hv_balloon.c min_pages = MB2PAGES(488) + (nr_pages >> 5); nr_pages 375 drivers/hwtracing/coresight/coresight-etb10.c int nr_pages, bool overwrite) nr_pages 387 drivers/hwtracing/coresight/coresight-etb10.c buf->nr_pages = nr_pages; nr_pages 411 drivers/hwtracing/coresight/coresight-etb10.c head = handle->head & ((buf->nr_pages << PAGE_SHIFT) - 1); nr_pages 547 drivers/hwtracing/coresight/coresight-etb10.c cur &= buf->nr_pages - 1; nr_pages 208 drivers/hwtracing/coresight/coresight-etm-perf.c int nr_pages, bool overwrite) nr_pages 281 drivers/hwtracing/coresight/coresight-etm-perf.c nr_pages, overwrite); nr_pages 96 drivers/hwtracing/coresight/coresight-priv.h unsigned int nr_pages; nr_pages 390 drivers/hwtracing/coresight/coresight-tmc-etf.c int nr_pages, bool overwrite) nr_pages 403 drivers/hwtracing/coresight/coresight-tmc-etf.c buf->nr_pages = nr_pages; nr_pages 427 drivers/hwtracing/coresight/coresight-tmc-etf.c head = handle->head & ((buf->nr_pages << PAGE_SHIFT) - 1); nr_pages 539 drivers/hwtracing/coresight/coresight-tmc-etf.c cur &= buf->nr_pages - 1; nr_pages 45 drivers/hwtracing/coresight/coresight-tmc-etr.c int nr_pages; nr_pages 50 drivers/hwtracing/coresight/coresight-tmc-etr.c #define PERF_IDX2OFF(idx, buf) ((idx) % ((buf)->nr_pages << PAGE_SHIFT)) nr_pages 122 drivers/hwtracing/coresight/coresight-tmc-etr.c tmc_etr_sg_table_entries(int nr_pages) nr_pages 124 drivers/hwtracing/coresight/coresight-tmc-etr.c unsigned long nr_sgpages = nr_pages * ETR_SG_PAGES_PER_SYSPAGE; nr_pages 147 drivers/hwtracing/coresight/coresight-tmc-etr.c for (i = 0; i < tmc_pages->nr_pages; i++) { nr_pages 167 drivers/hwtracing/coresight/coresight-tmc-etr.c for (i = 0; i < tmc_pages->nr_pages; i++) { nr_pages 179 drivers/hwtracing/coresight/coresight-tmc-etr.c tmc_pages->nr_pages = 0; nr_pages 194 drivers/hwtracing/coresight/coresight-tmc-etr.c int i, nr_pages; nr_pages 199 drivers/hwtracing/coresight/coresight-tmc-etr.c nr_pages = tmc_pages->nr_pages; nr_pages 200 drivers/hwtracing/coresight/coresight-tmc-etr.c tmc_pages->daddrs = kcalloc(nr_pages, sizeof(*tmc_pages->daddrs), nr_pages 204 drivers/hwtracing/coresight/coresight-tmc-etr.c tmc_pages->pages = kcalloc(nr_pages, sizeof(*tmc_pages->pages), nr_pages 212 drivers/hwtracing/coresight/coresight-tmc-etr.c for (i = 0; i < nr_pages; i++) { nr_pages 275 drivers/hwtracing/coresight/coresight-tmc-etr.c table_pages->nr_pages, nr_pages 295 drivers/hwtracing/coresight/coresight-tmc-etr.c sg_table->data_pages.nr_pages, nr_pages 327 drivers/hwtracing/coresight/coresight-tmc-etr.c sg_table->data_pages.nr_pages = nr_dpages; nr_pages 328 drivers/hwtracing/coresight/coresight-tmc-etr.c sg_table->table_pages.nr_pages = nr_tpages; nr_pages 358 drivers/hwtracing/coresight/coresight-tmc-etr.c index = i % data->nr_pages; nr_pages 371 drivers/hwtracing/coresight/coresight-tmc-etr.c for (i = 0; i < table_pages->nr_pages; i++) nr_pages 494 drivers/hwtracing/coresight/coresight-tmc-etr.c nr_entries = tmc_etr_sg_table_entries(sg_table->data_pages.nr_pages); nr_pages 1204 drivers/hwtracing/coresight/coresight-tmc-etr.c int nr_pages, void **pages, bool snapshot) nr_pages 1215 drivers/hwtracing/coresight/coresight-tmc-etr.c if ((nr_pages << PAGE_SHIFT) > drvdata->size) { nr_pages 1216 drivers/hwtracing/coresight/coresight-tmc-etr.c etr_buf = tmc_alloc_etr_buf(drvdata, (nr_pages << PAGE_SHIFT), nr_pages 1242 drivers/hwtracing/coresight/coresight-tmc-etr.c struct perf_event *event, int nr_pages, nr_pages 1279 drivers/hwtracing/coresight/coresight-tmc-etr.c etr_buf = alloc_etr_buf(drvdata, event, nr_pages, pages, snapshot); nr_pages 1306 drivers/hwtracing/coresight/coresight-tmc-etr.c struct perf_event *event, int nr_pages, nr_pages 1313 drivers/hwtracing/coresight/coresight-tmc-etr.c return alloc_etr_buf(drvdata, event, nr_pages, pages, snapshot); nr_pages 1318 drivers/hwtracing/coresight/coresight-tmc-etr.c int nr_pages, void **pages, bool snapshot) nr_pages 1321 drivers/hwtracing/coresight/coresight-tmc-etr.c return get_perf_etr_buf_per_thread(drvdata, event, nr_pages, nr_pages 1324 drivers/hwtracing/coresight/coresight-tmc-etr.c return get_perf_etr_buf_cpu_wide(drvdata, event, nr_pages, nr_pages 1330 drivers/hwtracing/coresight/coresight-tmc-etr.c int nr_pages, void **pages, bool snapshot) nr_pages 1342 drivers/hwtracing/coresight/coresight-tmc-etr.c etr_buf = get_perf_etr_buf(drvdata, event, nr_pages, pages, snapshot); nr_pages 1363 drivers/hwtracing/coresight/coresight-tmc-etr.c int nr_pages, bool snapshot) nr_pages 1369 drivers/hwtracing/coresight/coresight-tmc-etr.c nr_pages, pages, snapshot); nr_pages 1377 drivers/hwtracing/coresight/coresight-tmc-etr.c etr_perf->nr_pages = nr_pages; nr_pages 1460 drivers/hwtracing/coresight/coresight-tmc-etr.c if (++pg_idx == etr_perf->nr_pages) nr_pages 228 drivers/hwtracing/coresight/coresight-tmc.h int nr_pages; nr_pages 323 drivers/hwtracing/coresight/coresight-tmc.h return sg_table->data_pages.nr_pages << PAGE_SHIFT; nr_pages 141 drivers/hwtracing/intel_th/msu.c unsigned long nr_pages; nr_pages 780 drivers/hwtracing/intel_th/msu.c reg = msc->nr_pages; nr_pages 835 drivers/hwtracing/intel_th/msu.c msc->single_sz = reg & ((msc->nr_pages << PAGE_SHIFT) - 1); nr_pages 910 drivers/hwtracing/intel_th/msu.c unsigned long nr_pages = size >> PAGE_SHIFT; nr_pages 935 drivers/hwtracing/intel_th/msu.c msc->nr_pages = nr_pages; nr_pages 963 drivers/hwtracing/intel_th/msu.c for (off = 0; off < msc->nr_pages << PAGE_SHIFT; off += PAGE_SIZE) { nr_pages 970 drivers/hwtracing/intel_th/msu.c msc->nr_pages = 0; nr_pages 983 drivers/hwtracing/intel_th/msu.c if (pgoff >= msc->nr_pages) nr_pages 1108 drivers/hwtracing/intel_th/msu.c msc->nr_pages += nr_blocks; nr_pages 1143 drivers/hwtracing/intel_th/msu.c msc->nr_pages -= win->nr_blocks; nr_pages 1230 drivers/hwtracing/intel_th/msu.c static int msc_buffer_multi_alloc(struct msc *msc, unsigned long *nr_pages, nr_pages 1236 drivers/hwtracing/intel_th/msu.c ret = msc_buffer_win_alloc(msc, nr_pages[i]); nr_pages 1281 drivers/hwtracing/intel_th/msu.c static int msc_buffer_alloc(struct msc *msc, unsigned long *nr_pages, nr_pages 1294 drivers/hwtracing/intel_th/msu.c ret = msc_buffer_contig_alloc(msc, nr_pages[0] << PAGE_SHIFT); nr_pages 1296 drivers/hwtracing/intel_th/msu.c ret = msc_buffer_multi_alloc(msc, nr_pages, nr_wins); nr_pages 1461 drivers/hwtracing/intel_th/msu.c unsigned long size = msc->nr_pages << PAGE_SHIFT, rem = len; nr_pages 1509 drivers/hwtracing/intel_th/msu.c size = msc->nr_pages << PAGE_SHIFT; nr_pages 1565 drivers/hwtracing/intel_th/msu.c for (pg = 0; pg < msc->nr_pages; pg++) { nr_pages 1623 drivers/hwtracing/intel_th/msu.c if (size >> PAGE_SHIFT != msc->nr_pages) nr_pages 1931 drivers/hwtracing/intel_th/msu.c count = scnprintf(buf, PAGE_SIZE, "%ld\n", msc->nr_pages); nr_pages 2017 drivers/hwtracing/intel_th/msu.c static DEVICE_ATTR_RW(nr_pages); nr_pages 1032 drivers/iommu/dma-iommu.c unsigned long nr_pages = PAGE_ALIGN(size) >> PAGE_SHIFT; nr_pages 1041 drivers/iommu/dma-iommu.c if (off >= nr_pages || vma_pages(vma) > nr_pages - off) nr_pages 2225 drivers/iommu/intel-iommu.c unsigned long nr_pages, int prot) nr_pages 2233 drivers/iommu/intel-iommu.c BUG_ON(!domain_pfn_supported(domain, iov_pfn + nr_pages - 1)); nr_pages 2241 drivers/iommu/intel-iommu.c sg_res = nr_pages; nr_pages 2245 drivers/iommu/intel-iommu.c while (nr_pages > 0) { nr_pages 2304 drivers/iommu/intel-iommu.c BUG_ON(nr_pages < lvl_pages); nr_pages 2307 drivers/iommu/intel-iommu.c nr_pages -= lvl_pages; nr_pages 2325 drivers/iommu/intel-iommu.c if (!nr_pages || first_pte_in_page(pte) || nr_pages 2332 drivers/iommu/intel-iommu.c if (!sg_res && nr_pages) nr_pages 2340 drivers/iommu/intel-iommu.c unsigned long nr_pages, int prot) nr_pages 2346 drivers/iommu/intel-iommu.c ret = __domain_mapping(domain, iov_pfn, sg, phys_pfn, nr_pages, prot); nr_pages 2352 drivers/iommu/intel-iommu.c __mapping_notify_one(iommu, domain, iov_pfn, nr_pages); nr_pages 2359 drivers/iommu/intel-iommu.c struct scatterlist *sg, unsigned long nr_pages, nr_pages 2362 drivers/iommu/intel-iommu.c return domain_mapping(domain, iov_pfn, sg, 0, nr_pages, prot); nr_pages 2366 drivers/iommu/intel-iommu.c unsigned long phys_pfn, unsigned long nr_pages, nr_pages 2369 drivers/iommu/intel-iommu.c return domain_mapping(domain, iov_pfn, NULL, phys_pfn, nr_pages, prot); nr_pages 4659 drivers/iommu/intel-iommu.c end = ((mhp->start_pfn + mhp->nr_pages) << PAGE_SHIFT) - 1; nr_pages 4670 drivers/iommu/intel-iommu.c last_vpfn = mm_to_dma_pfn(mhp->start_pfn + mhp->nr_pages - 1); nr_pages 216 drivers/iommu/s390-iommu.c unsigned long irq_flags, nr_pages, i; nr_pages 224 drivers/iommu/s390-iommu.c nr_pages = PAGE_ALIGN(size) >> PAGE_SHIFT; nr_pages 225 drivers/iommu/s390-iommu.c if (!nr_pages) nr_pages 229 drivers/iommu/s390-iommu.c for (i = 0; i < nr_pages; i++) { nr_pages 243 drivers/iommu/s390-iommu.c start_dma_addr, nr_pages * PAGE_SIZE); nr_pages 324 drivers/lightnvm/pblk-core.c int nr_pages) nr_pages 340 drivers/lightnvm/pblk-core.c int nr_pages) nr_pages 346 drivers/lightnvm/pblk-core.c for (i = 0; i < nr_pages; i++) { nr_pages 823 drivers/lightnvm/pblk.h int nr_pages); nr_pages 825 drivers/lightnvm/pblk.h int nr_pages); nr_pages 306 drivers/md/dm-kcopyd.c static int client_reserve_pages(struct dm_kcopyd_client *kc, unsigned nr_pages) nr_pages 311 drivers/md/dm-kcopyd.c for (i = 0; i < nr_pages; i++) { nr_pages 322 drivers/md/dm-kcopyd.c kc->nr_reserved_pages += nr_pages; nr_pages 587 drivers/md/dm-kcopyd.c unsigned nr_pages = dm_div_up(job->dests[0].count, PAGE_SIZE >> 9); nr_pages 589 drivers/md/dm-kcopyd.c r = kcopyd_get_pages(job->kc, nr_pages, &job->pages); nr_pages 167 drivers/md/dm-linear.c long nr_pages, void **kaddr, pfn_t *pfn) nr_pages 176 drivers/md/dm-linear.c ret = bdev_dax_pgoff(bdev, dev_sector, nr_pages * PAGE_SIZE, &pgoff); nr_pages 179 drivers/md/dm-linear.c return dax_direct_access(dax_dev, pgoff, nr_pages, kaddr, pfn); nr_pages 949 drivers/md/dm-log-writes.c long nr_pages, void **kaddr, pfn_t *pfn) nr_pages 955 drivers/md/dm-log-writes.c ret = bdev_dax_pgoff(lc->dev->bdev, sector, nr_pages * PAGE_SIZE, &pgoff); nr_pages 958 drivers/md/dm-log-writes.c return dax_direct_access(lc->dev->dax_dev, pgoff, nr_pages, kaddr, pfn); nr_pages 318 drivers/md/dm-stripe.c long nr_pages, void **kaddr, pfn_t *pfn) nr_pages 332 drivers/md/dm-stripe.c ret = bdev_dax_pgoff(bdev, dev_sector, nr_pages * PAGE_SIZE, &pgoff); nr_pages 335 drivers/md/dm-stripe.c return dax_direct_access(dax_dev, pgoff, nr_pages, kaddr, pfn); nr_pages 145 drivers/md/dm-target.c long nr_pages, void **kaddr, pfn_t *pfn) nr_pages 1082 drivers/md/dm.c long nr_pages, void **kaddr, pfn_t *pfn) nr_pages 1099 drivers/md/dm.c nr_pages = min(len, nr_pages); nr_pages 1100 drivers/md/dm.c ret = ti->type->direct_access(ti, pgoff, nr_pages, kaddr, pfn); nr_pages 137 drivers/media/common/saa7146/saa7146_core.c static struct scatterlist* vmalloc_to_sg(unsigned char *virt, int nr_pages) nr_pages 143 drivers/media/common/saa7146/saa7146_core.c sglist = kcalloc(nr_pages, sizeof(struct scatterlist), GFP_KERNEL); nr_pages 146 drivers/media/common/saa7146/saa7146_core.c sg_init_table(sglist, nr_pages); nr_pages 147 drivers/media/common/saa7146/saa7146_core.c for (i = 0; i < nr_pages; i++, virt += PAGE_SIZE) { nr_pages 239 drivers/media/common/saa7146/saa7146_core.c int nr_pages = 0; nr_pages 258 drivers/media/common/saa7146/saa7146_core.c nr_pages++; nr_pages 265 drivers/media/common/saa7146/saa7146_core.c for(i=nr_pages;i<1024;i++) { nr_pages 71 drivers/media/pci/cx23885/cx23885-alsa.c static int cx23885_alsa_dma_init(struct cx23885_audio_dev *chip, int nr_pages) nr_pages 77 drivers/media/pci/cx23885/cx23885-alsa.c buf->vaddr = vmalloc_32(nr_pages << PAGE_SHIFT); nr_pages 79 drivers/media/pci/cx23885/cx23885-alsa.c dprintk(1, "vmalloc_32(%d pages) failed\n", nr_pages); nr_pages 84 drivers/media/pci/cx23885/cx23885-alsa.c buf->vaddr, nr_pages << PAGE_SHIFT); nr_pages 86 drivers/media/pci/cx23885/cx23885-alsa.c memset(buf->vaddr, 0, nr_pages << PAGE_SHIFT); nr_pages 87 drivers/media/pci/cx23885/cx23885-alsa.c buf->nr_pages = nr_pages; nr_pages 89 drivers/media/pci/cx23885/cx23885-alsa.c buf->sglist = vzalloc(array_size(sizeof(*buf->sglist), buf->nr_pages)); nr_pages 93 drivers/media/pci/cx23885/cx23885-alsa.c sg_init_table(buf->sglist, buf->nr_pages); nr_pages 94 drivers/media/pci/cx23885/cx23885-alsa.c for (i = 0; i < buf->nr_pages; i++) { nr_pages 116 drivers/media/pci/cx23885/cx23885-alsa.c buf->nr_pages, PCI_DMA_FROMDEVICE); nr_pages 329 drivers/media/pci/cx23885/cx23885.h int nr_pages; nr_pages 57 drivers/media/pci/cx25821/cx25821-alsa.c int nr_pages; nr_pages 134 drivers/media/pci/cx25821/cx25821-alsa.c static int cx25821_alsa_dma_init(struct cx25821_audio_dev *chip, int nr_pages) nr_pages 140 drivers/media/pci/cx25821/cx25821-alsa.c buf->vaddr = vmalloc_32(nr_pages << PAGE_SHIFT); nr_pages 142 drivers/media/pci/cx25821/cx25821-alsa.c dprintk(1, "vmalloc_32(%d pages) failed\n", nr_pages); nr_pages 148 drivers/media/pci/cx25821/cx25821-alsa.c nr_pages << PAGE_SHIFT); nr_pages 150 drivers/media/pci/cx25821/cx25821-alsa.c memset(buf->vaddr, 0, nr_pages << PAGE_SHIFT); nr_pages 151 drivers/media/pci/cx25821/cx25821-alsa.c buf->nr_pages = nr_pages; nr_pages 153 drivers/media/pci/cx25821/cx25821-alsa.c buf->sglist = vzalloc(array_size(sizeof(*buf->sglist), buf->nr_pages)); nr_pages 157 drivers/media/pci/cx25821/cx25821-alsa.c sg_init_table(buf->sglist, buf->nr_pages); nr_pages 158 drivers/media/pci/cx25821/cx25821-alsa.c for (i = 0; i < buf->nr_pages; i++) { nr_pages 180 drivers/media/pci/cx25821/cx25821-alsa.c buf->nr_pages, PCI_DMA_FROMDEVICE); nr_pages 50 drivers/media/pci/cx88/cx88-alsa.c int nr_pages; nr_pages 274 drivers/media/pci/cx88/cx88-alsa.c static int cx88_alsa_dma_init(struct cx88_audio_dev *chip, int nr_pages) nr_pages 280 drivers/media/pci/cx88/cx88-alsa.c buf->vaddr = vmalloc_32(nr_pages << PAGE_SHIFT); nr_pages 282 drivers/media/pci/cx88/cx88-alsa.c dprintk(1, "vmalloc_32(%d pages) failed\n", nr_pages); nr_pages 287 drivers/media/pci/cx88/cx88-alsa.c buf->vaddr, nr_pages << PAGE_SHIFT); nr_pages 289 drivers/media/pci/cx88/cx88-alsa.c memset(buf->vaddr, 0, nr_pages << PAGE_SHIFT); nr_pages 290 drivers/media/pci/cx88/cx88-alsa.c buf->nr_pages = nr_pages; nr_pages 292 drivers/media/pci/cx88/cx88-alsa.c buf->sglist = vzalloc(array_size(sizeof(*buf->sglist), buf->nr_pages)); nr_pages 296 drivers/media/pci/cx88/cx88-alsa.c sg_init_table(buf->sglist, buf->nr_pages); nr_pages 297 drivers/media/pci/cx88/cx88-alsa.c for (i = 0; i < buf->nr_pages; i++) { nr_pages 319 drivers/media/pci/cx88/cx88-alsa.c buf->nr_pages, PCI_DMA_FROMDEVICE); nr_pages 255 drivers/media/pci/saa7134/saa7134-alsa.c static int saa7134_alsa_dma_init(struct saa7134_dev *dev, int nr_pages) nr_pages 261 drivers/media/pci/saa7134/saa7134-alsa.c dma->vaddr = vmalloc_32(nr_pages << PAGE_SHIFT); nr_pages 263 drivers/media/pci/saa7134/saa7134-alsa.c pr_debug("vmalloc_32(%d pages) failed\n", nr_pages); nr_pages 268 drivers/media/pci/saa7134/saa7134-alsa.c dma->vaddr, nr_pages << PAGE_SHIFT); nr_pages 270 drivers/media/pci/saa7134/saa7134-alsa.c memset(dma->vaddr, 0, nr_pages << PAGE_SHIFT); nr_pages 271 drivers/media/pci/saa7134/saa7134-alsa.c dma->nr_pages = nr_pages; nr_pages 273 drivers/media/pci/saa7134/saa7134-alsa.c dma->sglist = vzalloc(array_size(sizeof(*dma->sglist), dma->nr_pages)); nr_pages 277 drivers/media/pci/saa7134/saa7134-alsa.c sg_init_table(dma->sglist, dma->nr_pages); nr_pages 278 drivers/media/pci/saa7134/saa7134-alsa.c for (i = 0; i < dma->nr_pages; i++) { nr_pages 300 drivers/media/pci/saa7134/saa7134-alsa.c dma->nr_pages, PCI_DMA_FROMDEVICE); nr_pages 512 drivers/media/pci/saa7134/saa7134.h int nr_pages; nr_pages 63 drivers/media/v4l2-core/videobuf-dma-sg.c int nr_pages) nr_pages 69 drivers/media/v4l2-core/videobuf-dma-sg.c sglist = vzalloc(array_size(nr_pages, sizeof(*sglist))); nr_pages 72 drivers/media/v4l2-core/videobuf-dma-sg.c sg_init_table(sglist, nr_pages); nr_pages 73 drivers/media/v4l2-core/videobuf-dma-sg.c for (i = 0; i < nr_pages; i++, virt += PAGE_SIZE) { nr_pages 93 drivers/media/v4l2-core/videobuf-dma-sg.c int nr_pages, int offset, size_t size) nr_pages 100 drivers/media/v4l2-core/videobuf-dma-sg.c sglist = vmalloc(array_size(nr_pages, sizeof(*sglist))); nr_pages 103 drivers/media/v4l2-core/videobuf-dma-sg.c sg_init_table(sglist, nr_pages); nr_pages 111 drivers/media/v4l2-core/videobuf-dma-sg.c for (i = 1; i < nr_pages; i++) { nr_pages 174 drivers/media/v4l2-core/videobuf-dma-sg.c dma->nr_pages = last-first+1; nr_pages 175 drivers/media/v4l2-core/videobuf-dma-sg.c dma->pages = kmalloc_array(dma->nr_pages, sizeof(struct page *), nr_pages 184 drivers/media/v4l2-core/videobuf-dma-sg.c data, size, dma->nr_pages); nr_pages 186 drivers/media/v4l2-core/videobuf-dma-sg.c err = get_user_pages(data & PAGE_MASK, dma->nr_pages, nr_pages 189 drivers/media/v4l2-core/videobuf-dma-sg.c if (err != dma->nr_pages) { nr_pages 190 drivers/media/v4l2-core/videobuf-dma-sg.c dma->nr_pages = (err >= 0) ? err : 0; nr_pages 192 drivers/media/v4l2-core/videobuf-dma-sg.c dma->nr_pages); nr_pages 211 drivers/media/v4l2-core/videobuf-dma-sg.c int nr_pages) nr_pages 215 drivers/media/v4l2-core/videobuf-dma-sg.c dprintk(1, "init kernel [%d pages]\n", nr_pages); nr_pages 218 drivers/media/v4l2-core/videobuf-dma-sg.c dma->vaddr_pages = kcalloc(nr_pages, sizeof(*dma->vaddr_pages), nr_pages 223 drivers/media/v4l2-core/videobuf-dma-sg.c dma->dma_addr = kcalloc(nr_pages, sizeof(*dma->dma_addr), GFP_KERNEL); nr_pages 228 drivers/media/v4l2-core/videobuf-dma-sg.c for (i = 0; i < nr_pages; i++) { nr_pages 238 drivers/media/v4l2-core/videobuf-dma-sg.c dma->vaddr = vmap(dma->vaddr_pages, nr_pages, VM_MAP | VM_IOREMAP, nr_pages 241 drivers/media/v4l2-core/videobuf-dma-sg.c dprintk(1, "vmalloc_32(%d pages) failed\n", nr_pages); nr_pages 246 drivers/media/v4l2-core/videobuf-dma-sg.c dma->vaddr, nr_pages << PAGE_SHIFT); nr_pages 248 drivers/media/v4l2-core/videobuf-dma-sg.c memset(dma->vaddr, 0, nr_pages << PAGE_SHIFT); nr_pages 249 drivers/media/v4l2-core/videobuf-dma-sg.c dma->nr_pages = nr_pages; nr_pages 270 drivers/media/v4l2-core/videobuf-dma-sg.c dma_addr_t addr, int nr_pages) nr_pages 273 drivers/media/v4l2-core/videobuf-dma-sg.c nr_pages, (unsigned long)addr); nr_pages 280 drivers/media/v4l2-core/videobuf-dma-sg.c dma->nr_pages = nr_pages; nr_pages 288 drivers/media/v4l2-core/videobuf-dma-sg.c BUG_ON(0 == dma->nr_pages); nr_pages 291 drivers/media/v4l2-core/videobuf-dma-sg.c dma->sglist = videobuf_pages_to_sg(dma->pages, dma->nr_pages, nr_pages 296 drivers/media/v4l2-core/videobuf-dma-sg.c dma->nr_pages); nr_pages 305 drivers/media/v4l2-core/videobuf-dma-sg.c sg_dma_len(&dma->sglist[0]) = dma->nr_pages * PAGE_SIZE; nr_pages 314 drivers/media/v4l2-core/videobuf-dma-sg.c dma->nr_pages, dma->direction); nr_pages 335 drivers/media/v4l2-core/videobuf-dma-sg.c dma_unmap_sg(dev, dma->sglist, dma->nr_pages, dma->direction); nr_pages 352 drivers/media/v4l2-core/videobuf-dma-sg.c for (i = 0; i < dma->nr_pages; i++) { nr_pages 362 drivers/media/v4l2-core/videobuf-dma-sg.c for (i = 0; i < dma->nr_pages; i++) { nr_pages 585 drivers/media/v4l2-core/videobuf-dma-sg.c mem->dma.nr_pages, mem->dma.direction); nr_pages 172 drivers/misc/genwqe/card_base.h unsigned int nr_pages; /* number of pages */ nr_pages 356 drivers/misc/genwqe/card_base.h unsigned long nr_pages; nr_pages 450 drivers/misc/genwqe/card_dev.c dma_map->nr_pages = DIV_ROUND_UP(vsize, PAGE_SIZE); nr_pages 296 drivers/misc/genwqe/card_utils.c sgl->nr_pages = DIV_ROUND_UP(sgl->fpage_offs + user_size, PAGE_SIZE); nr_pages 300 drivers/misc/genwqe/card_utils.c __func__, user_addr, user_size, sgl->nr_pages, nr_pages 306 drivers/misc/genwqe/card_utils.c sgl->sgl_size = genwqe_sgl_size(sgl->nr_pages); nr_pages 385 drivers/misc/genwqe/card_utils.c while (p < sgl->nr_pages) { nr_pages 403 drivers/misc/genwqe/card_utils.c } else if ((p == sgl->nr_pages - 1) && nr_pages 424 drivers/misc/genwqe/card_utils.c if (p == sgl->nr_pages) nr_pages 440 drivers/misc/genwqe/card_utils.c if (p == sgl->nr_pages) nr_pages 527 drivers/misc/genwqe/card_utils.c unsigned int nr_pages, int dirty) nr_pages 531 drivers/misc/genwqe/card_utils.c for (i = 0; i < nr_pages; i++) { nr_pages 585 drivers/misc/genwqe/card_utils.c m->nr_pages = DIV_ROUND_UP(offs + size, PAGE_SIZE); nr_pages 587 drivers/misc/genwqe/card_utils.c m->page_list = kcalloc(m->nr_pages, nr_pages 592 drivers/misc/genwqe/card_utils.c m->nr_pages = 0; nr_pages 597 drivers/misc/genwqe/card_utils.c m->dma_list = (dma_addr_t *)(m->page_list + m->nr_pages); nr_pages 601 drivers/misc/genwqe/card_utils.c m->nr_pages, nr_pages 608 drivers/misc/genwqe/card_utils.c if (rc < m->nr_pages) { nr_pages 614 drivers/misc/genwqe/card_utils.c rc = genwqe_map_pages(cd, m->page_list, m->nr_pages, m->dma_list); nr_pages 621 drivers/misc/genwqe/card_utils.c genwqe_free_user_pages(m->page_list, m->nr_pages, m->write); nr_pages 627 drivers/misc/genwqe/card_utils.c m->nr_pages = 0; nr_pages 650 drivers/misc/genwqe/card_utils.c genwqe_unmap_pages(cd, m->dma_list, m->nr_pages); nr_pages 653 drivers/misc/genwqe/card_utils.c genwqe_free_user_pages(m->page_list, m->nr_pages, m->write); nr_pages 658 drivers/misc/genwqe/card_utils.c m->nr_pages = 0; nr_pages 49 drivers/misc/mic/scif/scif_debugfs.c window->nr_pages, window->nr_contig_chunks, window->prot); nr_pages 60 drivers/misc/mic/scif/scif_debugfs.c for (j = 0; j < window->nr_pages; j++) nr_pages 124 drivers/misc/mic/scif/scif_dma.c end_va = start_va + (window->nr_pages << PAGE_SHIFT); nr_pages 376 drivers/misc/mic/scif/scif_dma.c (*out_window)->nr_pages = pinned_pages->nr_pages; nr_pages 631 drivers/misc/mic/scif/scif_dma.c atomic_sub(window->nr_pages, nr_pages 754 drivers/misc/mic/scif/scif_dma.c if (window->nr_pages == window->nr_contig_chunks) { nr_pages 818 drivers/misc/mic/scif/scif_dma.c (window->nr_pages << PAGE_SHIFT); nr_pages 823 drivers/misc/mic/scif/scif_dma.c (window->nr_pages << PAGE_SHIFT); nr_pages 921 drivers/misc/mic/scif/scif_dma.c (window->nr_pages << PAGE_SHIFT); nr_pages 929 drivers/misc/mic/scif/scif_dma.c (window->nr_pages << PAGE_SHIFT); nr_pages 1034 drivers/misc/mic/scif/scif_dma.c (window->nr_pages << PAGE_SHIFT); nr_pages 1110 drivers/misc/mic/scif/scif_dma.c (src_window->nr_pages << PAGE_SHIFT); nr_pages 1112 drivers/misc/mic/scif/scif_dma.c (dst_window->nr_pages << PAGE_SHIFT); nr_pages 1117 drivers/misc/mic/scif/scif_dma.c (src_window->nr_pages << PAGE_SHIFT); nr_pages 1123 drivers/misc/mic/scif/scif_dma.c (dst_window->nr_pages << PAGE_SHIFT); nr_pages 1275 drivers/misc/mic/scif/scif_dma.c (src_window->nr_pages << PAGE_SHIFT); nr_pages 1277 drivers/misc/mic/scif/scif_dma.c (dst_window->nr_pages << PAGE_SHIFT); nr_pages 1284 drivers/misc/mic/scif/scif_dma.c (src_window->nr_pages << PAGE_SHIFT); nr_pages 1290 drivers/misc/mic/scif/scif_dma.c (dst_window->nr_pages << PAGE_SHIFT); nr_pages 1514 drivers/misc/mic/scif/scif_dma.c (src_window->nr_pages << PAGE_SHIFT); nr_pages 1516 drivers/misc/mic/scif/scif_dma.c (dst_window->nr_pages << PAGE_SHIFT); nr_pages 1727 drivers/misc/mic/scif/scif_dma.c atomic_add_return(local_window->nr_pages, nr_pages 34 drivers/misc/mic/scif/scif_mmap.c req.nr_bytes = recv_window->nr_pages << PAGE_SHIFT; nr_pages 48 drivers/misc/mic/scif/scif_mmap.c scif_put_window(window, window->nr_pages); nr_pages 130 drivers/misc/mic/scif/scif_mmap.c scif_put_window(window, window->nr_pages); nr_pages 218 drivers/misc/mic/scif/scif_mmap.c int nr_pages, err, i; nr_pages 233 drivers/misc/mic/scif/scif_mmap.c nr_pages = len >> PAGE_SHIFT; nr_pages 259 drivers/misc/mic/scif/scif_mmap.c (*pages)->phys_addr = scif_zalloc(nr_pages * sizeof(dma_addr_t)); nr_pages 267 drivers/misc/mic/scif/scif_mmap.c ((*pages)->va = scif_zalloc(nr_pages * sizeof(void *))); nr_pages 275 drivers/misc/mic/scif/scif_mmap.c (*pages)->nr_pages = nr_pages; nr_pages 278 drivers/misc/mic/scif/scif_mmap.c for (i = 0; i < nr_pages; i++) { nr_pages 291 drivers/misc/mic/scif/scif_mmap.c scif_get_window(window, nr_pages); nr_pages 297 drivers/misc/mic/scif/scif_mmap.c nr_pages * sizeof(dma_addr_t)); nr_pages 299 drivers/misc/mic/scif/scif_mmap.c nr_pages * sizeof(void *)); nr_pages 336 drivers/misc/mic/scif/scif_mmap.c scif_put_window(window, pages->nr_pages); nr_pages 356 drivers/misc/mic/scif/scif_mmap.c scif_free(pages->phys_addr, pages->nr_pages * sizeof(dma_addr_t)); nr_pages 357 drivers/misc/mic/scif/scif_mmap.c scif_free(pages->va, pages->nr_pages * sizeof(void *)); nr_pages 373 drivers/misc/mic/scif/scif_mmap.c int nr_pages, struct vm_area_struct *vma) nr_pages 377 drivers/misc/mic/scif/scif_mmap.c int loop_nr_pages, nr_pages_left = nr_pages; nr_pages 388 drivers/misc/mic/scif/scif_mmap.c (window->nr_pages << PAGE_SHIFT); nr_pages 419 drivers/misc/mic/scif/scif_mmap.c nr_pages_left = nr_pages; nr_pages 424 drivers/misc/mic/scif/scif_mmap.c (window->nr_pages << PAGE_SHIFT); nr_pages 450 drivers/misc/mic/scif/scif_mmap.c s64 offset, int nr_pages) nr_pages 454 drivers/misc/mic/scif/scif_mmap.c int loop_nr_pages, nr_pages_left = nr_pages; nr_pages 462 drivers/misc/mic/scif/scif_mmap.c nr_pages_left = nr_pages; nr_pages 465 drivers/misc/mic/scif/scif_mmap.c (window->nr_pages << PAGE_SHIFT); nr_pages 546 drivers/misc/mic/scif/scif_mmap.c int nr_pages = vma_pages(vma); nr_pages 561 drivers/misc/mic/scif/scif_mmap.c ep, nr_pages, offset); nr_pages 576 drivers/misc/mic/scif/scif_mmap.c scif_rma_list_munmap(window, offset, nr_pages); nr_pages 608 drivers/misc/mic/scif/scif_mmap.c int nr_pages = vma_pages(vma); nr_pages 614 drivers/misc/mic/scif/scif_mmap.c ep, start_offset, nr_pages); nr_pages 672 drivers/misc/mic/scif/scif_mmap.c err = scif_rma_list_mmap(window, start_offset, nr_pages, vma); nr_pages 81 drivers/misc/mic/scif/scif_rma.c scif_create_pinned_pages(int nr_pages, int prot) nr_pages 90 drivers/misc/mic/scif/scif_rma.c pin->pages = scif_zalloc(nr_pages * sizeof(*pin->pages)); nr_pages 116 drivers/misc/mic/scif/scif_rma.c for (j = 0; j < pin->nr_pages; j++) { nr_pages 125 drivers/misc/mic/scif/scif_rma.c pin->nr_pages * sizeof(*pin->pages)); nr_pages 139 drivers/misc/mic/scif/scif_rma.c struct scif_window *scif_create_window(struct scif_endpt *ep, int nr_pages, nr_pages 149 drivers/misc/mic/scif/scif_rma.c window->dma_addr = scif_zalloc(nr_pages * sizeof(*window->dma_addr)); nr_pages 153 drivers/misc/mic/scif/scif_rma.c window->num_pages = scif_zalloc(nr_pages * sizeof(*window->num_pages)); nr_pages 171 drivers/misc/mic/scif/scif_rma.c nr_pages * sizeof(*window->dma_addr)); nr_pages 188 drivers/misc/mic/scif/scif_rma.c int nr_pages = window->nr_pages; nr_pages 213 drivers/misc/mic/scif/scif_rma.c scif_free(window->dma_addr, nr_pages * sizeof(*window->dma_addr)); nr_pages 214 drivers/misc/mic/scif/scif_rma.c scif_free(window->num_pages, nr_pages * sizeof(*window->num_pages)); nr_pages 266 drivers/misc/mic/scif/scif_rma.c int nr_pages) nr_pages 268 drivers/misc/mic/scif/scif_rma.c if (!mm || !nr_pages || !scif_ulimit_check) nr_pages 271 drivers/misc/mic/scif/scif_rma.c atomic64_sub(nr_pages, &mm->pinned_vm); nr_pages 276 drivers/misc/mic/scif/scif_rma.c int nr_pages) nr_pages 280 drivers/misc/mic/scif/scif_rma.c if (!mm || !nr_pages || !scif_ulimit_check) nr_pages 284 drivers/misc/mic/scif/scif_rma.c locked = atomic64_add_return(nr_pages, &mm->pinned_vm); nr_pages 287 drivers/misc/mic/scif/scif_rma.c atomic64_sub(nr_pages, &mm->pinned_vm); nr_pages 307 drivers/misc/mic/scif/scif_rma.c int nr_pages = window->nr_pages; nr_pages 311 drivers/misc/mic/scif/scif_rma.c __scif_dec_pinned_vm_lock(window->mm, window->nr_pages); nr_pages 334 drivers/misc/mic/scif/scif_rma.c scif_free(window->dma_addr, nr_pages * sizeof(*window->dma_addr)); nr_pages 335 drivers/misc/mic/scif/scif_rma.c scif_free(window->num_pages, nr_pages * sizeof(*window->num_pages)); nr_pages 354 drivers/misc/mic/scif/scif_rma.c int nr_pages = window->nr_pages; nr_pages 365 drivers/misc/mic/scif/scif_rma.c window->nr_lookup = ALIGN(nr_pages * PAGE_SIZE, nr_pages 392 drivers/misc/mic/scif/scif_rma.c for (i = 0, j = 0; i < nr_pages; i += SCIF_NR_ADDR_IN_PAGE, j++) { nr_pages 430 drivers/misc/mic/scif/scif_rma.c for (i = 0, j = 0; i < window->nr_pages; nr_pages 464 drivers/misc/mic/scif/scif_rma.c scif_create_remote_window(struct scif_dev *scifdev, int nr_pages) nr_pages 474 drivers/misc/mic/scif/scif_rma.c window->nr_pages = nr_pages; nr_pages 476 drivers/misc/mic/scif/scif_rma.c window->dma_addr = scif_zalloc(nr_pages * sizeof(*window->dma_addr)); nr_pages 480 drivers/misc/mic/scif/scif_rma.c window->num_pages = scif_zalloc(nr_pages * nr_pages 508 drivers/misc/mic/scif/scif_rma.c scif_free(window->dma_addr, window->nr_pages * nr_pages 510 drivers/misc/mic/scif/scif_rma.c scif_free(window->num_pages, window->nr_pages * nr_pages 535 drivers/misc/mic/scif/scif_rma.c err = sg_alloc_table(window->st, window->nr_pages, GFP_KERNEL); nr_pages 587 drivers/misc/mic/scif/scif_rma.c for (i = 0, j = 0; i < window->nr_pages; i += nr_contig_pages, j++) { nr_pages 592 drivers/misc/mic/scif/scif_rma.c for (k = i + 1; k < window->nr_pages; k++) { nr_pages 709 drivers/misc/mic/scif/scif_rma.c scif_put_window(window, window->nr_pages); nr_pages 722 drivers/misc/mic/scif/scif_rma.c window->nr_pages)) { nr_pages 752 drivers/misc/mic/scif/scif_rma.c msg.payload[1] = window->nr_pages; nr_pages 1045 drivers/misc/mic/scif/scif_rma.c int nr_pages = msg->payload[1]; nr_pages 1047 drivers/misc/mic/scif/scif_rma.c window = scif_create_remote_window(scifdev, nr_pages); nr_pages 1065 drivers/misc/mic/scif/scif_rma.c __func__, __LINE__, err, window, nr_pages); nr_pages 1187 drivers/misc/mic/scif/scif_rma.c req.nr_bytes = recv_window->nr_pages << PAGE_SHIFT; nr_pages 1202 drivers/misc/mic/scif/scif_rma.c scif_put_window(window, window->nr_pages); nr_pages 1311 drivers/misc/mic/scif/scif_rma.c int nr_pages, err = 0, i; nr_pages 1335 drivers/misc/mic/scif/scif_rma.c nr_pages = len >> PAGE_SHIFT; nr_pages 1338 drivers/misc/mic/scif/scif_rma.c pinned_pages = scif_create_pinned_pages(nr_pages, prot); nr_pages 1346 drivers/misc/mic/scif/scif_rma.c for (i = 0; i < nr_pages; i++) { nr_pages 1354 drivers/misc/mic/scif/scif_rma.c pinned_pages->nr_pages = nr_pages; nr_pages 1371 drivers/misc/mic/scif/scif_rma.c err = __scif_check_inc_pinned_vm(mm, nr_pages); nr_pages 1373 drivers/misc/mic/scif/scif_rma.c pinned_pages->nr_pages = 0; nr_pages 1378 drivers/misc/mic/scif/scif_rma.c pinned_pages->nr_pages = get_user_pages_fast( nr_pages 1380 drivers/misc/mic/scif/scif_rma.c nr_pages, nr_pages 1383 drivers/misc/mic/scif/scif_rma.c if (nr_pages != pinned_pages->nr_pages) { nr_pages 1386 drivers/misc/mic/scif/scif_rma.c __scif_dec_pinned_vm_lock(mm, nr_pages); nr_pages 1388 drivers/misc/mic/scif/scif_rma.c for (i = 0; i < pinned_pages->nr_pages; i++) { nr_pages 1401 drivers/misc/mic/scif/scif_rma.c if (pinned_pages->nr_pages < nr_pages) { nr_pages 1403 drivers/misc/mic/scif/scif_rma.c pinned_pages->nr_pages = nr_pages; nr_pages 1413 drivers/misc/mic/scif/scif_rma.c __scif_dec_pinned_vm_lock(mm, nr_pages); nr_pages 1416 drivers/misc/mic/scif/scif_rma.c pinned_pages->nr_pages = nr_pages; nr_pages 1481 drivers/misc/mic/scif/scif_rma.c len = pinned_pages->nr_pages << PAGE_SHIFT; nr_pages 1514 drivers/misc/mic/scif/scif_rma.c window = scif_create_window(ep, pinned_pages->nr_pages, nr_pages 1523 drivers/misc/mic/scif/scif_rma.c window->nr_pages = pinned_pages->nr_pages; nr_pages 1632 drivers/misc/mic/scif/scif_rma.c window->nr_pages = len >> PAGE_SHIFT; nr_pages 1698 drivers/misc/mic/scif/scif_rma.c int nr_pages, err; nr_pages 1720 drivers/misc/mic/scif/scif_rma.c nr_pages = len >> PAGE_SHIFT; nr_pages 1743 drivers/misc/mic/scif/scif_rma.c err = scif_rma_list_unregister(window, offset, nr_pages); nr_pages 187 drivers/misc/mic/scif/scif_rma.h s64 nr_pages; nr_pages 256 drivers/misc/mic/scif/scif_rma.h s64 nr_pages; nr_pages 323 drivers/misc/mic/scif/scif_rma.h s64 offset, int nr_pages, s64 *out_offset); nr_pages 328 drivers/misc/mic/scif/scif_rma.h struct scif_window *scif_create_window(struct scif_endpt *ep, int nr_pages, nr_pages 439 drivers/misc/mic/scif/scif_rma.h static inline void scif_get_window(struct scif_window *window, int nr_pages) nr_pages 441 drivers/misc/mic/scif/scif_rma.h window->ref_count += nr_pages; nr_pages 444 drivers/misc/mic/scif/scif_rma.h static inline void scif_put_window(struct scif_window *window, int nr_pages) nr_pages 446 drivers/misc/mic/scif/scif_rma.h window->ref_count -= nr_pages; nr_pages 449 drivers/misc/mic/scif/scif_rma.h static inline void scif_set_window_ref(struct scif_window *window, int nr_pages) nr_pages 451 drivers/misc/mic/scif/scif_rma.h window->ref_count = nr_pages; nr_pages 65 drivers/misc/mic/scif/scif_rma_list.c scif_set_window_ref(window, window->nr_pages); nr_pages 92 drivers/misc/mic/scif/scif_rma_list.c (window->nr_pages << PAGE_SHIFT); nr_pages 100 drivers/misc/mic/scif/scif_rma_list.c (window->nr_pages << PAGE_SHIFT); nr_pages 148 drivers/misc/mic/scif/scif_rma_list.c (window->nr_pages << PAGE_SHIFT); nr_pages 195 drivers/misc/mic/scif/scif_rma_list.c s64 offset, int nr_pages) nr_pages 205 drivers/misc/mic/scif/scif_rma_list.c end_offset = window->offset + (window->nr_pages << PAGE_SHIFT); nr_pages 207 drivers/misc/mic/scif/scif_rma_list.c nr_pages); nr_pages 211 drivers/misc/mic/scif/scif_rma_list.c nr_pages -= loop_nr_pages; nr_pages 213 drivers/misc/mic/scif/scif_rma_list.c if (!nr_pages) nr_pages 47 drivers/misc/mic/scif/scif_rma_list.h int nr_pages); nr_pages 615 drivers/mtd/devices/mtd_dataflash.c static int add_dataflash_otp(struct spi_device *spi, char *name, int nr_pages, nr_pages 640 drivers/mtd/devices/mtd_dataflash.c device->size = nr_pages * pagesize; nr_pages 673 drivers/mtd/devices/mtd_dataflash.c int nr_pages, int pagesize, int pageoffset) nr_pages 675 drivers/mtd/devices/mtd_dataflash.c return add_dataflash_otp(spi, name, nr_pages, pagesize, nr_pages 688 drivers/mtd/devices/mtd_dataflash.c unsigned nr_pages; nr_pages 856 drivers/mtd/devices/mtd_dataflash.c return add_dataflash_otp(spi, info->name, info->nr_pages, nr_pages 56 drivers/mtd/devices/sst25l.c unsigned nr_pages; nr_pages 377 drivers/mtd/devices/sst25l.c flash->mtd.size = flash_info->page_size * flash_info->nr_pages; nr_pages 1143 drivers/mtd/mtdswap.c unsigned nr_pages) nr_pages 1152 drivers/mtd/mtdswap.c for (page = first; page < first + nr_pages; page++) { nr_pages 2641 drivers/net/ethernet/broadcom/bnxt/bnxt.c for (i = 0; i < rmem->nr_pages; i++) { nr_pages 2651 drivers/net/ethernet/broadcom/bnxt/bnxt.c size_t pg_tbl_size = rmem->nr_pages * 8; nr_pages 2673 drivers/net/ethernet/broadcom/bnxt/bnxt.c if ((rmem->nr_pages > 1 || rmem->depth > 0) && !rmem->pg_tbl) { nr_pages 2674 drivers/net/ethernet/broadcom/bnxt/bnxt.c size_t pg_tbl_size = rmem->nr_pages * 8; nr_pages 2685 drivers/net/ethernet/broadcom/bnxt/bnxt.c for (i = 0; i < rmem->nr_pages; i++) { nr_pages 2695 drivers/net/ethernet/broadcom/bnxt/bnxt.c if (rmem->nr_pages > 1 || rmem->depth > 0) { nr_pages 2696 drivers/net/ethernet/broadcom/bnxt/bnxt.c if (i == rmem->nr_pages - 2 && nr_pages 2699 drivers/net/ethernet/broadcom/bnxt/bnxt.c else if (i == rmem->nr_pages - 1 && nr_pages 3011 drivers/net/ethernet/broadcom/bnxt/bnxt.c rmem->nr_pages = bp->cp_nr_pages; nr_pages 3097 drivers/net/ethernet/broadcom/bnxt/bnxt.c rmem->nr_pages = bp->cp_nr_pages; nr_pages 3109 drivers/net/ethernet/broadcom/bnxt/bnxt.c rmem->nr_pages = bp->rx_nr_pages; nr_pages 3118 drivers/net/ethernet/broadcom/bnxt/bnxt.c rmem->nr_pages = bp->rx_agg_nr_pages; nr_pages 3132 drivers/net/ethernet/broadcom/bnxt/bnxt.c rmem->nr_pages = bp->tx_nr_pages; nr_pages 3148 drivers/net/ethernet/broadcom/bnxt/bnxt.c for (i = 0, prod = 0; i < ring->ring_mem.nr_pages; i++) { nr_pages 5231 drivers/net/ethernet/broadcom/bnxt/bnxt.c if (rmem->nr_pages > 1) { nr_pages 6654 drivers/net/ethernet/broadcom/bnxt/bnxt.c ctx_pg->nr_pages = DIV_ROUND_UP(mem_size, BNXT_PAGE_SIZE); nr_pages 6655 drivers/net/ethernet/broadcom/bnxt/bnxt.c if (ctx_pg->nr_pages > MAX_CTX_TOTAL_PAGES) { nr_pages 6656 drivers/net/ethernet/broadcom/bnxt/bnxt.c ctx_pg->nr_pages = 0; nr_pages 6659 drivers/net/ethernet/broadcom/bnxt/bnxt.c if (ctx_pg->nr_pages > MAX_CTX_PAGES || depth > 1) { nr_pages 6667 drivers/net/ethernet/broadcom/bnxt/bnxt.c nr_tbls = DIV_ROUND_UP(ctx_pg->nr_pages, MAX_CTX_PAGES); nr_pages 6668 drivers/net/ethernet/broadcom/bnxt/bnxt.c rmem->nr_pages = nr_tbls; nr_pages 6683 drivers/net/ethernet/broadcom/bnxt/bnxt.c rmem->nr_pages = MAX_CTX_PAGES; nr_pages 6685 drivers/net/ethernet/broadcom/bnxt/bnxt.c int rem = ctx_pg->nr_pages % MAX_CTX_PAGES; nr_pages 6688 drivers/net/ethernet/broadcom/bnxt/bnxt.c rmem->nr_pages = rem; nr_pages 6695 drivers/net/ethernet/broadcom/bnxt/bnxt.c rmem->nr_pages = DIV_ROUND_UP(mem_size, BNXT_PAGE_SIZE); nr_pages 6696 drivers/net/ethernet/broadcom/bnxt/bnxt.c if (rmem->nr_pages > 1 || depth) nr_pages 6708 drivers/net/ethernet/broadcom/bnxt/bnxt.c if (rmem->depth > 1 || ctx_pg->nr_pages > MAX_CTX_PAGES || nr_pages 6710 drivers/net/ethernet/broadcom/bnxt/bnxt.c int i, nr_tbls = rmem->nr_pages; nr_pages 6729 drivers/net/ethernet/broadcom/bnxt/bnxt.c ctx_pg->nr_pages = 0; nr_pages 711 drivers/net/ethernet/broadcom/bnxt/bnxt.h int nr_pages; nr_pages 1309 drivers/net/ethernet/broadcom/bnxt/bnxt.h u32 nr_pages; nr_pages 408 drivers/net/ethernet/broadcom/bnxt/bnxt_sriov.c u32 nr_pages, size, i, j, k = 0; nr_pages 417 drivers/net/ethernet/broadcom/bnxt/bnxt_sriov.c nr_pages = size / BNXT_PAGE_SIZE; nr_pages 419 drivers/net/ethernet/broadcom/bnxt/bnxt_sriov.c nr_pages++; nr_pages 421 drivers/net/ethernet/broadcom/bnxt/bnxt_sriov.c for (i = 0; i < nr_pages; i++) { nr_pages 447 drivers/net/ethernet/broadcom/bnxt/bnxt_sriov.c bp->pf.hwrm_cmd_req_pages = nr_pages; nr_pages 247 drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c int cxgbi_ppm_ppods_reserve(struct cxgbi_ppm *ppm, unsigned short nr_pages, nr_pages 257 drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c npods = (nr_pages + PPOD_PAGES_MAX - 1) >> PPOD_PAGES_SHIFT; nr_pages 260 drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c __func__, nr_pages, npods); nr_pages 271 drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c nr_pages, npods, ppm->next, caller_data); nr_pages 287 drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.c nr_pages, tag, idx, npods, caller_data); nr_pages 94 drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.h int nr_pages; nr_pages 323 drivers/net/ethernet/chelsio/libcxgb/libcxgb_ppm.h int cxgbi_ppm_ppods_reserve(struct cxgbi_ppm *, unsigned short nr_pages, nr_pages 246 drivers/net/ethernet/ibm/ehea/ehea.h u32 nr_pages; nr_pages 291 drivers/net/ethernet/ibm/ehea/ehea.h u32 nr_pages; nr_pages 3255 drivers/net/ethernet/ibm/ehea/ehea_main.c if (ehea_add_sect_bmap(arg->start_pfn, arg->nr_pages)) nr_pages 3263 drivers/net/ethernet/ibm/ehea/ehea_main.c if (ehea_rem_sect_bmap(arg->start_pfn, arg->nr_pages)) nr_pages 314 drivers/net/ethernet/ibm/ehea/ehea_phyp.c cq_attr->nr_pages = outs[4]; nr_pages 381 drivers/net/ethernet/ibm/ehea/ehea_phyp.c eq_attr->nr_pages = outs[4]; nr_pages 93 drivers/net/ethernet/ibm/ehea/ehea_qmr.c int i, nr_pages; nr_pages 100 drivers/net/ethernet/ibm/ehea/ehea_qmr.c nr_pages = queue->queue_length / queue->pagesize; nr_pages 102 drivers/net/ethernet/ibm/ehea/ehea_qmr.c for (i = 0; i < nr_pages; i += pages_per_kpage) nr_pages 134 drivers/net/ethernet/ibm/ehea/ehea_qmr.c ret = hw_queue_ctor(&cq->hw_queue, cq->attr.nr_pages, nr_pages 139 drivers/net/ethernet/ibm/ehea/ehea_qmr.c for (counter = 0; counter < cq->attr.nr_pages; counter++) { nr_pages 152 drivers/net/ethernet/ibm/ehea/ehea_qmr.c cq, hret, counter, cq->attr.nr_pages); nr_pages 156 drivers/net/ethernet/ibm/ehea/ehea_qmr.c if (counter == (cq->attr.nr_pages - 1)) { nr_pages 255 drivers/net/ethernet/ibm/ehea/ehea_qmr.c ret = hw_queue_ctor(&eq->hw_queue, eq->attr.nr_pages, nr_pages 262 drivers/net/ethernet/ibm/ehea/ehea_qmr.c for (i = 0; i < eq->attr.nr_pages; i++) { nr_pages 276 drivers/net/ethernet/ibm/ehea/ehea_qmr.c if (i == (eq->attr.nr_pages - 1)) { nr_pages 358 drivers/net/ethernet/ibm/ehea/ehea_qmr.c int nr_pages, int wqe_size, int act_nr_sges, nr_pages 365 drivers/net/ethernet/ibm/ehea/ehea_qmr.c ret = hw_queue_ctor(hw_queue, nr_pages, EHEA_PAGESIZE, wqe_size); nr_pages 369 drivers/net/ethernet/ibm/ehea/ehea_qmr.c for (cnt = 0; cnt < nr_pages; cnt++) { nr_pages 608 drivers/net/ethernet/ibm/ehea/ehea_qmr.c static int ehea_update_busmap(unsigned long pfn, unsigned long nr_pages, int add) nr_pages 612 drivers/net/ethernet/ibm/ehea/ehea_qmr.c if (!nr_pages) nr_pages 622 drivers/net/ethernet/ibm/ehea/ehea_qmr.c end_section = start_section + ((nr_pages * PAGE_SIZE) / EHEA_SECTSIZE); nr_pages 651 drivers/net/ethernet/ibm/ehea/ehea_qmr.c int ehea_add_sect_bmap(unsigned long pfn, unsigned long nr_pages) nr_pages 656 drivers/net/ethernet/ibm/ehea/ehea_qmr.c ret = ehea_update_busmap(pfn, nr_pages, EHEA_BUSMAP_ADD_SECT); nr_pages 661 drivers/net/ethernet/ibm/ehea/ehea_qmr.c int ehea_rem_sect_bmap(unsigned long pfn, unsigned long nr_pages) nr_pages 666 drivers/net/ethernet/ibm/ehea/ehea_qmr.c ret = ehea_update_busmap(pfn, nr_pages, EHEA_BUSMAP_REM_SECT); nr_pages 689 drivers/net/ethernet/ibm/ehea/ehea_qmr.c unsigned long pfn, start_pfn, end_pfn, nr_pages; nr_pages 703 drivers/net/ethernet/ibm/ehea/ehea_qmr.c nr_pages = pfn - start_pfn; nr_pages 704 drivers/net/ethernet/ibm/ehea/ehea_qmr.c ret = ehea_update_busmap(start_pfn, nr_pages, nr_pages 717 drivers/net/ethernet/ibm/ehea/ehea_qmr.c nr_pages = pfn - start_pfn; nr_pages 718 drivers/net/ethernet/ibm/ehea/ehea_qmr.c return ehea_update_busmap(start_pfn, nr_pages, EHEA_BUSMAP_ADD_SECT); nr_pages 384 drivers/net/ethernet/ibm/ehea/ehea_qmr.h int ehea_add_sect_bmap(unsigned long pfn, unsigned long nr_pages); nr_pages 385 drivers/net/ethernet/ibm/ehea/ehea_qmr.h int ehea_rem_sect_bmap(unsigned long pfn, unsigned long nr_pages); nr_pages 244 drivers/nvdimm/pmem.c long nr_pages, void **kaddr, pfn_t *pfn) nr_pages 249 drivers/nvdimm/pmem.c PFN_PHYS(nr_pages)))) nr_pages 262 drivers/nvdimm/pmem.c return nr_pages; nr_pages 273 drivers/nvdimm/pmem.c pgoff_t pgoff, long nr_pages, void **kaddr, pfn_t *pfn) nr_pages 277 drivers/nvdimm/pmem.c return __pmem_direct_access(pmem, pgoff, nr_pages, kaddr, pfn); nr_pages 30 drivers/nvdimm/pmem.h long nr_pages, void **kaddr, pfn_t *pfn); nr_pages 45 drivers/perf/arm_spe_pmu.c int nr_pages; nr_pages 78 drivers/perf/arm_spe_pmu.c #define PERF_IDX2OFF(idx, buf) ((idx) % ((buf)->nr_pages << PAGE_SHIFT)) nr_pages 362 drivers/perf/arm_spe_pmu.c u64 limit = buf->nr_pages * PAGE_SIZE; nr_pages 380 drivers/perf/arm_spe_pmu.c limit = ((buf->nr_pages * PAGE_SIZE) >> 1) + handle->head; nr_pages 390 drivers/perf/arm_spe_pmu.c const u64 bufsize = buf->nr_pages * PAGE_SIZE; nr_pages 818 drivers/perf/arm_spe_pmu.c int nr_pages, bool snapshot) nr_pages 825 drivers/perf/arm_spe_pmu.c if (nr_pages < 2) nr_pages 834 drivers/perf/arm_spe_pmu.c if (!nr_pages || (snapshot && (nr_pages & 1))) nr_pages 844 drivers/perf/arm_spe_pmu.c pglist = kcalloc(nr_pages, sizeof(*pglist), GFP_KERNEL); nr_pages 848 drivers/perf/arm_spe_pmu.c for (i = 0; i < nr_pages; ++i) nr_pages 851 drivers/perf/arm_spe_pmu.c buf->base = vmap(pglist, nr_pages, VM_MAP, PAGE_KERNEL); nr_pages 855 drivers/perf/arm_spe_pmu.c buf->nr_pages = nr_pages; nr_pages 542 drivers/rapidio/devices/rio_mport_cdev.c unsigned int nr_pages; nr_pages 581 drivers/rapidio/devices/rio_mport_cdev.c for (i = 0; i < req->nr_pages; i++) nr_pages 813 drivers/rapidio/devices/rio_mport_cdev.c unsigned long nr_pages = 0; nr_pages 856 drivers/rapidio/devices/rio_mport_cdev.c nr_pages = PAGE_ALIGN(xfer->length + offset) >> PAGE_SHIFT; nr_pages 858 drivers/rapidio/devices/rio_mport_cdev.c page_list = kmalloc_array(nr_pages, nr_pages 867 drivers/rapidio/devices/rio_mport_cdev.c nr_pages, nr_pages 871 drivers/rapidio/devices/rio_mport_cdev.c if (pinned != nr_pages) { nr_pages 875 drivers/rapidio/devices/rio_mport_cdev.c nr_pages = 0; nr_pages 878 drivers/rapidio/devices/rio_mport_cdev.c pinned, nr_pages); nr_pages 884 drivers/rapidio/devices/rio_mport_cdev.c nr_pages = pinned; nr_pages 888 drivers/rapidio/devices/rio_mport_cdev.c ret = sg_alloc_table_from_pages(&req->sgt, page_list, nr_pages, nr_pages 896 drivers/rapidio/devices/rio_mport_cdev.c req->nr_pages = nr_pages; nr_pages 954 drivers/rapidio/devices/rio_mport_cdev.c for (i = 0; i < nr_pages; i++) nr_pages 37 drivers/s390/block/dcssblk.c long nr_pages, void **kaddr, pfn_t *pfn); nr_pages 920 drivers/s390/block/dcssblk.c long nr_pages, void **kaddr, pfn_t *pfn) nr_pages 937 drivers/s390/block/dcssblk.c long nr_pages, void **kaddr, pfn_t *pfn) nr_pages 941 drivers/s390/block/dcssblk.c return __dcssblk_direct_access(dev_info, pgoff, nr_pages, kaddr, pfn); nr_pages 328 drivers/s390/char/sclp_cmd.c size = arg->nr_pages << PAGE_SHIFT; nr_pages 63 drivers/s390/char/vmcp.c int nr_pages, order; nr_pages 66 drivers/s390/char/vmcp.c nr_pages = ALIGN(session->bufsize, PAGE_SIZE) >> PAGE_SHIFT; nr_pages 73 drivers/s390/char/vmcp.c page = cma_alloc(vmcp_cma, nr_pages, 0, false); nr_pages 84 drivers/s390/char/vmcp.c int nr_pages, order; nr_pages 90 drivers/s390/char/vmcp.c nr_pages = ALIGN(session->bufsize, PAGE_SIZE) >> PAGE_SHIFT; nr_pages 93 drivers/s390/char/vmcp.c cma_release(vmcp_cma, page, nr_pages); nr_pages 1099 drivers/s390/cio/css.c struct gen_pool *cio_gp_dma_create(struct device *dma_dev, int nr_pages) nr_pages 1109 drivers/s390/cio/css.c for (i = 0; i < nr_pages; ++i) { nr_pages 1368 drivers/scsi/cxgbi/libcxgbi.c ttinfo->nr_pages = (xferlen + sgl->offset + (1 << PAGE_SHIFT) - 1) >> nr_pages 1376 drivers/scsi/cxgbi/libcxgbi.c err = cxgbi_ppm_ppods_reserve(ppm, ttinfo->nr_pages, 0, &ttinfo->idx, nr_pages 1393 drivers/scsi/cxgbi/libcxgbi.c if (err != ttinfo->nr_pages) { nr_pages 4903 drivers/scsi/st.c const int nr_pages = end - start; nr_pages 4913 drivers/scsi/st.c if (nr_pages > max_pages) nr_pages 4926 drivers/scsi/st.c res = get_user_pages_fast(uaddr, nr_pages, rw == READ ? FOLL_WRITE : 0, nr_pages 4930 drivers/scsi/st.c if (res < nr_pages) nr_pages 4933 drivers/scsi/st.c for (i=0; i < nr_pages; i++) { nr_pages 4943 drivers/scsi/st.c return nr_pages; nr_pages 4957 drivers/scsi/st.c const unsigned int nr_pages, int dirtied) nr_pages 4961 drivers/scsi/st.c for (i=0; i < nr_pages; i++) { nr_pages 35 drivers/staging/android/ion/ion_cma_heap.c unsigned long nr_pages = size >> PAGE_SHIFT; nr_pages 42 drivers/staging/android/ion/ion_cma_heap.c pages = cma_alloc(cma_heap->cma, nr_pages, align, false); nr_pages 47 drivers/staging/android/ion/ion_cma_heap.c unsigned long nr_clear_pages = nr_pages; nr_pages 79 drivers/staging/android/ion/ion_cma_heap.c cma_release(cma_heap->cma, pages, nr_pages); nr_pages 87 drivers/staging/android/ion/ion_cma_heap.c unsigned long nr_pages = PAGE_ALIGN(buffer->size) >> PAGE_SHIFT; nr_pages 90 drivers/staging/android/ion/ion_cma_heap.c cma_release(cma_heap->cma, pages, nr_pages); nr_pages 3169 drivers/staging/exfat/exfat_super.c struct list_head *pages, unsigned int nr_pages) nr_pages 3171 drivers/staging/exfat/exfat_super.c return mpage_readpages(mapping, pages, nr_pages, exfat_get_block); nr_pages 191 drivers/target/iscsi/cxgbit/cxgbit_ddp.c ttinfo->nr_pages = (xferlen + sgl->offset + nr_pages 197 drivers/target/iscsi/cxgbit/cxgbit_ddp.c ret = cxgbi_ppm_ppods_reserve(ppm, ttinfo->nr_pages, 0, &ttinfo->idx, nr_pages 854 drivers/target/target_core_pscsi.c int nr_pages = (cmd->data_length + sgl[0].offset + nr_pages 861 drivers/target/target_core_pscsi.c pr_debug("PSCSI: nr_pages: %d\n", nr_pages); nr_pages 884 drivers/target/target_core_pscsi.c nr_vecs = min_t(int, BIO_MAX_PAGES, nr_pages); nr_pages 885 drivers/target/target_core_pscsi.c nr_pages -= nr_vecs; nr_pages 31 drivers/tee/optee/shm_pool.c unsigned int nr_pages = 1 << order, i; nr_pages 34 drivers/tee/optee/shm_pool.c pages = kcalloc(nr_pages, sizeof(pages), GFP_KERNEL); nr_pages 38 drivers/tee/optee/shm_pool.c for (i = 0; i < nr_pages; i++) { nr_pages 44 drivers/tee/optee/shm_pool.c rc = optee_shm_register(shm->ctx, shm, pages, nr_pages, nr_pages 645 drivers/video/fbdev/pvr2fb.c unsigned int nr_pages; nr_pages 649 drivers/video/fbdev/pvr2fb.c nr_pages = (count + PAGE_SIZE - 1) >> PAGE_SHIFT; nr_pages 651 drivers/video/fbdev/pvr2fb.c pages = kmalloc_array(nr_pages, sizeof(struct page *), GFP_KERNEL); nr_pages 655 drivers/video/fbdev/pvr2fb.c ret = get_user_pages_fast((unsigned long)buf, nr_pages, FOLL_WRITE, pages); nr_pages 656 drivers/video/fbdev/pvr2fb.c if (ret < nr_pages) { nr_pages 657 drivers/video/fbdev/pvr2fb.c nr_pages = ret; nr_pages 666 drivers/video/fbdev/pvr2fb.c end = (unsigned long)page_address(pages[nr_pages]); nr_pages 667 drivers/video/fbdev/pvr2fb.c len = nr_pages << PAGE_SHIFT; nr_pages 685 drivers/video/fbdev/pvr2fb.c for (i = 0; i < nr_pages; i++, dst += PAGE_SIZE) { nr_pages 701 drivers/video/fbdev/pvr2fb.c for (i = 0; i < nr_pages; i++) nr_pages 447 drivers/video/fbdev/sh_mobile_lcdcfb.c int nr_pages = 0; nr_pages 452 drivers/video/fbdev/sh_mobile_lcdcfb.c sg_set_page(&ch->sglist[nr_pages++], page, PAGE_SIZE, 0); nr_pages 454 drivers/video/fbdev/sh_mobile_lcdcfb.c return nr_pages; nr_pages 482 drivers/video/fbdev/sh_mobile_lcdcfb.c unsigned int nr_pages = sh_mobile_lcdc_sginit(info, pagelist); nr_pages 485 drivers/video/fbdev/sh_mobile_lcdcfb.c dma_map_sg(ch->lcdc->dev, ch->sglist, nr_pages, DMA_TO_DEVICE); nr_pages 489 drivers/video/fbdev/sh_mobile_lcdcfb.c dma_unmap_sg(ch->lcdc->dev, ch->sglist, nr_pages, nr_pages 48 drivers/video/fbdev/xen-fbfront.c int nr_pages; nr_pages 413 drivers/video/fbdev/xen-fbfront.c info->nr_pages = (fb_size + PAGE_SIZE - 1) >> PAGE_SHIFT; nr_pages 415 drivers/video/fbdev/xen-fbfront.c info->gfns = vmalloc(array_size(sizeof(unsigned long), info->nr_pages)); nr_pages 561 drivers/video/fbdev/xen-fbfront.c for (i = 0; i < info->nr_pages; i++) nr_pages 564 drivers/video/fbdev/xen-fbfront.c for (i = 0; i * epd < info->nr_pages; i++) nr_pages 414 drivers/xen/balloon.c static enum bp_state increase_reservation(unsigned long nr_pages) nr_pages 420 drivers/xen/balloon.c if (nr_pages > ARRAY_SIZE(frame_list)) nr_pages 421 drivers/xen/balloon.c nr_pages = ARRAY_SIZE(frame_list); nr_pages 424 drivers/xen/balloon.c for (i = 0; i < nr_pages; i++) { nr_pages 426 drivers/xen/balloon.c nr_pages = i; nr_pages 434 drivers/xen/balloon.c rc = xenmem_reservation_increase(nr_pages, frame_list); nr_pages 453 drivers/xen/balloon.c static enum bp_state decrease_reservation(unsigned long nr_pages, gfp_t gfp) nr_pages 461 drivers/xen/balloon.c if (nr_pages > ARRAY_SIZE(frame_list)) nr_pages 462 drivers/xen/balloon.c nr_pages = ARRAY_SIZE(frame_list); nr_pages 464 drivers/xen/balloon.c for (i = 0; i < nr_pages; i++) { nr_pages 467 drivers/xen/balloon.c nr_pages = i; nr_pages 502 drivers/xen/balloon.c ret = xenmem_reservation_decrease(nr_pages, frame_list); nr_pages 503 drivers/xen/balloon.c BUG_ON(ret != nr_pages); nr_pages 505 drivers/xen/balloon.c balloon_stats.current_pages -= nr_pages; nr_pages 566 drivers/xen/balloon.c static int add_ballooned_pages(int nr_pages) nr_pages 581 drivers/xen/balloon.c if (si_mem_available() < nr_pages) nr_pages 584 drivers/xen/balloon.c st = decrease_reservation(nr_pages, GFP_USER); nr_pages 597 drivers/xen/balloon.c int alloc_xenballooned_pages(int nr_pages, struct page **pages) nr_pages 605 drivers/xen/balloon.c balloon_stats.target_unpopulated += nr_pages; nr_pages 607 drivers/xen/balloon.c while (pgno < nr_pages) { nr_pages 625 drivers/xen/balloon.c ret = add_ballooned_pages(nr_pages - pgno); nr_pages 644 drivers/xen/balloon.c void free_xenballooned_pages(int nr_pages, struct page **pages) nr_pages 650 drivers/xen/balloon.c for (i = 0; i < nr_pages; i++) { nr_pages 655 drivers/xen/balloon.c balloon_stats.target_unpopulated -= nr_pages; nr_pages 698 drivers/xen/balloon.c ? min(xen_start_info->nr_pages - xen_released_pages, max_pfn) nr_pages 58 drivers/xen/gntdev-dmabuf.c int nr_pages; nr_pages 201 drivers/xen/gntdev-dmabuf.c dmabuf_pages_to_sgt(struct page **pages, unsigned int nr_pages) nr_pages 212 drivers/xen/gntdev-dmabuf.c ret = sg_alloc_table_from_pages(sgt, pages, nr_pages, 0, nr_pages 213 drivers/xen/gntdev-dmabuf.c nr_pages << PAGE_SHIFT, nr_pages 271 drivers/xen/gntdev-dmabuf.c pr_debug("Mapping %d pages for dev %p\n", gntdev_dmabuf->nr_pages, nr_pages 289 drivers/xen/gntdev-dmabuf.c gntdev_dmabuf->nr_pages); nr_pages 399 drivers/xen/gntdev-dmabuf.c gntdev_dmabuf->nr_pages = args->count; nr_pages 593 drivers/xen/gntdev-dmabuf.c gntdev_dmabuf->nr_pages = count; nr_pages 645 drivers/xen/gntdev-dmabuf.c if (attach->dmabuf->size != gntdev_dmabuf->nr_pages << PAGE_SHIFT) { nr_pages 648 drivers/xen/gntdev-dmabuf.c attach->dmabuf->size, gntdev_dmabuf->nr_pages); nr_pages 733 drivers/xen/gntdev-dmabuf.c gntdev_dmabuf->nr_pages); nr_pages 105 drivers/xen/gntdev.c args.nr_pages = map->count; nr_pages 173 drivers/xen/gntdev.c args.nr_pages = count; nr_pages 833 drivers/xen/gntdev.c unsigned int nr_pages; nr_pages 848 drivers/xen/gntdev.c batch->pages[batch->nr_pages++] = page; nr_pages 860 drivers/xen/gntdev.c for (i = 0; i < batch->nr_pages; i++) nr_pages 862 drivers/xen/gntdev.c batch->nr_pages = 0; nr_pages 1001 drivers/xen/gntdev.c batch.nr_pages = 0; nr_pages 776 drivers/xen/grant-table.c int gnttab_pages_set_private(int nr_pages, struct page **pages) nr_pages 780 drivers/xen/grant-table.c for (i = 0; i < nr_pages; i++) { nr_pages 802 drivers/xen/grant-table.c int gnttab_alloc_pages(int nr_pages, struct page **pages) nr_pages 806 drivers/xen/grant-table.c ret = alloc_xenballooned_pages(nr_pages, pages); nr_pages 810 drivers/xen/grant-table.c ret = gnttab_pages_set_private(nr_pages, pages); nr_pages 812 drivers/xen/grant-table.c gnttab_free_pages(nr_pages, pages); nr_pages 818 drivers/xen/grant-table.c void gnttab_pages_clear_private(int nr_pages, struct page **pages) nr_pages 822 drivers/xen/grant-table.c for (i = 0; i < nr_pages; i++) { nr_pages 838 drivers/xen/grant-table.c void gnttab_free_pages(int nr_pages, struct page **pages) nr_pages 840 drivers/xen/grant-table.c gnttab_pages_clear_private(nr_pages, pages); nr_pages 841 drivers/xen/grant-table.c free_xenballooned_pages(nr_pages, pages); nr_pages 856 drivers/xen/grant-table.c size = args->nr_pages << PAGE_SHIFT; nr_pages 871 drivers/xen/grant-table.c for (pfn = start_pfn, i = 0; pfn < start_pfn + args->nr_pages; nr_pages 880 drivers/xen/grant-table.c xenmem_reservation_va_mapping_reset(args->nr_pages, args->pages); nr_pages 882 drivers/xen/grant-table.c ret = xenmem_reservation_decrease(args->nr_pages, args->frames); nr_pages 883 drivers/xen/grant-table.c if (ret != args->nr_pages) { nr_pages 889 drivers/xen/grant-table.c ret = gnttab_pages_set_private(args->nr_pages, args->pages); nr_pages 910 drivers/xen/grant-table.c gnttab_pages_clear_private(args->nr_pages, args->pages); nr_pages 912 drivers/xen/grant-table.c for (i = 0; i < args->nr_pages; i++) nr_pages 915 drivers/xen/grant-table.c ret = xenmem_reservation_increase(args->nr_pages, args->frames); nr_pages 916 drivers/xen/grant-table.c if (ret != args->nr_pages) { nr_pages 923 drivers/xen/grant-table.c xenmem_reservation_va_mapping_update(args->nr_pages, args->pages, nr_pages 926 drivers/xen/grant-table.c size = args->nr_pages << PAGE_SHIFT; nr_pages 68 drivers/xen/privcmd.c unsigned long nr_pages); nr_pages 453 drivers/xen/privcmd.c unsigned long nr_pages; nr_pages 481 drivers/xen/privcmd.c nr_pages = DIV_ROUND_UP(m.num, XEN_PFN_PER_PAGE); nr_pages 482 drivers/xen/privcmd.c if ((m.num <= 0) || (nr_pages > (LONG_MAX >> PAGE_SHIFT))) nr_pages 524 drivers/xen/privcmd.c m.addr + (nr_pages << PAGE_SHIFT) != vma->vm_end) { nr_pages 529 drivers/xen/privcmd.c ret = alloc_empty_pages(vma, nr_pages); nr_pages 536 drivers/xen/privcmd.c m.addr + (nr_pages << PAGE_SHIFT) > vma->vm_end) { nr_pages 540 drivers/xen/privcmd.c if (privcmd_vma_range_is_mapped(vma, m.addr, nr_pages)) { nr_pages 585 drivers/xen/privcmd.c struct page *pages[], unsigned int nr_pages) nr_pages 596 drivers/xen/privcmd.c if (requested > nr_pages) nr_pages 605 drivers/xen/privcmd.c nr_pages -= pinned; nr_pages 612 drivers/xen/privcmd.c static void unlock_pages(struct page *pages[], unsigned int nr_pages) nr_pages 619 drivers/xen/privcmd.c for (i = 0; i < nr_pages; i++) { nr_pages 630 drivers/xen/privcmd.c unsigned int nr_pages = 0; nr_pages 671 drivers/xen/privcmd.c nr_pages += DIV_ROUND_UP( nr_pages 676 drivers/xen/privcmd.c pages = kcalloc(nr_pages, sizeof(*pages), GFP_KERNEL); nr_pages 688 drivers/xen/privcmd.c rc = lock_pages(kbufs, kdata.num, pages, nr_pages); nr_pages 702 drivers/xen/privcmd.c unlock_pages(pages, nr_pages); nr_pages 951 drivers/xen/privcmd.c unsigned long nr_pages) nr_pages 953 drivers/xen/privcmd.c return apply_to_page_range(vma->vm_mm, addr, nr_pages << PAGE_SHIFT, nr_pages 91 drivers/xen/swiotlb-xen.c unsigned int i, nr_pages = XEN_PFN_UP(xen_offset_in_page(p) + size); nr_pages 95 drivers/xen/swiotlb-xen.c for (i = 1; i < nr_pages; i++) nr_pages 360 drivers/xen/xenbus/xenbus_client.c unsigned int nr_pages, grant_ref_t *grefs) nr_pages 365 drivers/xen/xenbus/xenbus_client.c for (i = 0; i < nr_pages; i++) { nr_pages 562 drivers/xen/xenbus/xenbus_client.c unsigned int nr_pages = XENBUS_PAGES(nr_grefs); nr_pages 573 drivers/xen/xenbus/xenbus_client.c err = alloc_xenballooned_pages(nr_pages, node->hvm.pages); nr_pages 588 drivers/xen/xenbus/xenbus_client.c addr = vmap(node->hvm.pages, nr_pages, VM_MAP | VM_IOREMAP, nr_pages 609 drivers/xen/xenbus/xenbus_client.c addr, nr_pages); nr_pages 612 drivers/xen/xenbus/xenbus_client.c free_xenballooned_pages(nr_pages, node->hvm.pages); nr_pages 829 drivers/xen/xenbus/xenbus_client.c unsigned int nr_pages; nr_pages 849 drivers/xen/xenbus/xenbus_client.c nr_pages = XENBUS_PAGES(node->nr_handles); nr_pages 859 drivers/xen/xenbus/xenbus_client.c free_xenballooned_pages(nr_pages, node->hvm.pages); nr_pages 862 drivers/xen/xenbus/xenbus_client.c WARN(1, "Leaking %p, size %u page(s)\n", vaddr, nr_pages); nr_pages 222 drivers/xen/xlate_mmu.c unsigned long nr_pages; nr_pages 225 drivers/xen/xlate_mmu.c nr_pages = DIV_ROUND_UP(nr_grant_frames, XEN_PFN_PER_PAGE); nr_pages 226 drivers/xen/xlate_mmu.c pages = kcalloc(nr_pages, sizeof(pages[0]), GFP_KERNEL); nr_pages 235 drivers/xen/xlate_mmu.c rc = alloc_xenballooned_pages(nr_pages, pages); nr_pages 238 drivers/xen/xlate_mmu.c nr_pages, rc); nr_pages 248 drivers/xen/xlate_mmu.c vaddr = vmap(pages, nr_pages, 0, PAGE_KERNEL); nr_pages 251 drivers/xen/xlate_mmu.c nr_pages, rc); nr_pages 252 drivers/xen/xlate_mmu.c free_xenballooned_pages(nr_pages, pages); nr_pages 281 fs/9p/cache.c unsigned *nr_pages) nr_pages 286 fs/9p/cache.c p9_debug(P9_DEBUG_FSC, "inode %p pages %u\n", inode, *nr_pages); nr_pages 291 fs/9p/cache.c mapping, pages, nr_pages, nr_pages 302 fs/9p/cache.c BUG_ON(*nr_pages != 0); nr_pages 37 fs/9p/cache.h unsigned *nr_pages); nr_pages 62 fs/9p/cache.h unsigned *nr_pages) nr_pages 65 fs/9p/cache.h nr_pages); nr_pages 118 fs/9p/cache.h unsigned *nr_pages) nr_pages 99 fs/9p/vfs_addr.c struct list_head *pages, unsigned nr_pages) nr_pages 107 fs/9p/vfs_addr.c ret = v9fs_readpages_from_fscache(inode, mapping, pages, &nr_pages); nr_pages 169 fs/afs/dir.c for (i = 0; i < req->nr_pages; i++) nr_pages 179 fs/afs/dir.c req->pos, req->index, req->nr_pages, req->offset); nr_pages 181 fs/afs/dir.c for (i = 0; i < req->nr_pages; i++) { nr_pages 219 fs/afs/dir.c int nr_pages, nr_inline, i, n; nr_pages 236 fs/afs/dir.c nr_pages = (i_size + PAGE_SIZE - 1) / PAGE_SIZE; nr_pages 237 fs/afs/dir.c nr_inline = nr_pages; nr_pages 246 fs/afs/dir.c req->nr_pages = nr_pages; nr_pages 248 fs/afs/dir.c req->len = nr_pages * PAGE_SIZE; /* We can ask for more than there is */ nr_pages 253 fs/afs/dir.c req->pages = kcalloc(nr_pages, sizeof(struct page *), nr_pages 267 fs/afs/dir.c req->nr_pages - i, nr_pages 269 fs/afs/dir.c _debug("find %u at %u/%u", n, i, req->nr_pages); nr_pages 293 fs/afs/dir.c } while (i < req->nr_pages); nr_pages 315 fs/afs/dir.c task_io_account_read(PAGE_SIZE * req->nr_pages); nr_pages 26 fs/afs/file.c struct list_head *pages, unsigned nr_pages); nr_pages 195 fs/afs/file.c for (i = 0; i < req->nr_pages; i++) nr_pages 326 fs/afs/file.c req->nr_pages = 1; nr_pages 498 fs/afs/file.c req->pages[req->nr_pages++] = page; nr_pages 500 fs/afs/file.c } while (req->nr_pages < n); nr_pages 502 fs/afs/file.c if (req->nr_pages == 0) { nr_pages 511 fs/afs/file.c task_io_account_read(PAGE_SIZE * req->nr_pages); nr_pages 523 fs/afs/file.c for (i = 0; i < req->nr_pages; i++) { nr_pages 542 fs/afs/file.c struct list_head *pages, unsigned nr_pages) nr_pages 549 fs/afs/file.c key_serial(key), mapping->host->i_ino, nr_pages); nr_pages 564 fs/afs/file.c &nr_pages, nr_pages 576 fs/afs/file.c BUG_ON(nr_pages != 0); nr_pages 363 fs/afs/fsclient.c ASSERTCMP(req->index, <, req->nr_pages); nr_pages 437 fs/afs/fsclient.c for (; req->index < req->nr_pages; req->index++) { nr_pages 445 fs/afs/fsclient.c for (req->index = 0; req->index < req->nr_pages; req->index++) nr_pages 233 fs/afs/internal.h unsigned int nr_pages; nr_pages 54 fs/afs/write.c req->nr_pages = 1; nr_pages 475 fs/afs/yfsclient.c ASSERTCMP(req->index, <, req->nr_pages); nr_pages 553 fs/afs/yfsclient.c for (; req->index < req->nr_pages; req->index++) { nr_pages 561 fs/afs/yfsclient.c for (req->index = 0; req->index < req->nr_pages; req->index++) nr_pages 128 fs/aio.c long nr_pages; nr_pages 236 fs/aio.c static struct file *aio_private_file(struct kioctx *ctx, loff_t nr_pages) nr_pages 245 fs/aio.c inode->i_size = PAGE_SIZE * nr_pages; nr_pages 311 fs/aio.c for (i = 0; i < ctx->nr_pages; i++) { nr_pages 413 fs/aio.c if (idx < (pgoff_t)ctx->nr_pages) { nr_pages 466 fs/aio.c int nr_pages; nr_pages 476 fs/aio.c nr_pages = PFN_UP(size); nr_pages 477 fs/aio.c if (nr_pages < 0) nr_pages 480 fs/aio.c file = aio_private_file(ctx, nr_pages); nr_pages 487 fs/aio.c nr_events = (PAGE_SIZE * nr_pages - sizeof(struct aio_ring)) nr_pages 491 fs/aio.c if (nr_pages > AIO_RING_PAGES) { nr_pages 492 fs/aio.c ctx->ring_pages = kcalloc(nr_pages, sizeof(struct page *), nr_pages 500 fs/aio.c for (i = 0; i < nr_pages; i++) { nr_pages 513 fs/aio.c ctx->nr_pages = i; nr_pages 515 fs/aio.c if (unlikely(i != nr_pages)) { nr_pages 520 fs/aio.c ctx->mmap_size = nr_pages * PAGE_SIZE; nr_pages 204 fs/block_dev.c int nr_pages) nr_pages 219 fs/block_dev.c if (nr_pages <= DIO_INLINE_BIO_VECS) nr_pages 222 fs/block_dev.c vecs = kmalloc_array(nr_pages, sizeof(struct bio_vec), nr_pages 228 fs/block_dev.c bio_init(&bio, vecs, nr_pages); nr_pages 339 fs/block_dev.c __blkdev_direct_IO(struct kiocb *iocb, struct iov_iter *iter, int nr_pages) nr_pages 357 fs/block_dev.c bio = bio_alloc_bioset(GFP_KERNEL, nr_pages, &blkdev_dio_pool); nr_pages 406 fs/block_dev.c nr_pages = iov_iter_npages(iter, BIO_MAX_PAGES); nr_pages 407 fs/block_dev.c if (!nr_pages) { nr_pages 437 fs/block_dev.c bio = bio_alloc(GFP_KERNEL, nr_pages); nr_pages 469 fs/block_dev.c int nr_pages; nr_pages 471 fs/block_dev.c nr_pages = iov_iter_npages(iter, BIO_MAX_PAGES + 1); nr_pages 472 fs/block_dev.c if (!nr_pages) nr_pages 474 fs/block_dev.c if (is_sync_kiocb(iocb) && nr_pages <= BIO_MAX_PAGES) nr_pages 475 fs/block_dev.c return __blkdev_direct_IO_simple(iocb, iter, nr_pages); nr_pages 477 fs/block_dev.c return __blkdev_direct_IO(iocb, iter, min(nr_pages, BIO_MAX_PAGES)); nr_pages 619 fs/block_dev.c struct list_head *pages, unsigned nr_pages) nr_pages 621 fs/block_dev.c return mpage_readpages(mapping, pages, nr_pages, blkdev_get_block); nr_pages 93 fs/btrfs/compression.c for (i = 0; i < cb->nr_pages; i++) { nr_pages 176 fs/btrfs/compression.c for (index = 0; index < cb->nr_pages; index++) { nr_pages 217 fs/btrfs/compression.c unsigned long nr_pages = end_index - index + 1; nr_pages 224 fs/btrfs/compression.c while (nr_pages > 0) { nr_pages 227 fs/btrfs/compression.c nr_pages, ARRAY_SIZE(pages)), pages); nr_pages 229 fs/btrfs/compression.c nr_pages -= 1; nr_pages 239 fs/btrfs/compression.c nr_pages -= ret; nr_pages 287 fs/btrfs/compression.c for (index = 0; index < cb->nr_pages; index++) { nr_pages 313 fs/btrfs/compression.c unsigned long nr_pages, nr_pages 340 fs/btrfs/compression.c cb->nr_pages = nr_pages; nr_pages 353 fs/btrfs/compression.c for (pg_index = 0; pg_index < cb->nr_pages; pg_index++) { nr_pages 397 fs/btrfs/compression.c bytes_left, cb->compressed_len, cb->nr_pages); nr_pages 438 fs/btrfs/compression.c unsigned long nr_pages = 0; nr_pages 521 fs/btrfs/compression.c nr_pages++; nr_pages 553 fs/btrfs/compression.c unsigned long nr_pages; nr_pages 601 fs/btrfs/compression.c nr_pages = DIV_ROUND_UP(compressed_len, PAGE_SIZE); nr_pages 602 fs/btrfs/compression.c cb->compressed_pages = kcalloc(nr_pages, sizeof(struct page *), nr_pages 609 fs/btrfs/compression.c for (pg_index = 0; pg_index < nr_pages; pg_index++) { nr_pages 618 fs/btrfs/compression.c faili = nr_pages - 1; nr_pages 619 fs/btrfs/compression.c cb->nr_pages = nr_pages; nr_pages 633 fs/btrfs/compression.c for (pg_index = 0; pg_index < nr_pages; pg_index++) { nr_pages 51 fs/btrfs/compression.h unsigned long nr_pages; nr_pages 95 fs/btrfs/compression.h unsigned long nr_pages, nr_pages 1872 fs/btrfs/extent_io.c unsigned long nr_pages = end_index - start_index + 1; nr_pages 1885 fs/btrfs/extent_io.c if ((page_ops & PAGE_SET_ERROR) && nr_pages > 0) nr_pages 1888 fs/btrfs/extent_io.c while (nr_pages > 0) { nr_pages 1891 fs/btrfs/extent_io.c nr_pages, ARRAY_SIZE(pages)), pages); nr_pages 1934 fs/btrfs/extent_io.c nr_pages -= ret; nr_pages 3265 fs/btrfs/extent_io.c struct page *pages[], int nr_pages, nr_pages 3277 fs/btrfs/extent_io.c for (index = 0; index < nr_pages; index++) { nr_pages 3936 fs/btrfs/extent_io.c int nr_pages; nr_pages 3964 fs/btrfs/extent_io.c (nr_pages = pagevec_lookup_range_tag(&pvec, mapping, &index, end, nr_pages 3968 fs/btrfs/extent_io.c for (i = 0; i < nr_pages; i++) { nr_pages 4104 fs/btrfs/extent_io.c int nr_pages; nr_pages 4162 fs/btrfs/extent_io.c (nr_pages = pagevec_lookup_range_tag(&pvec, mapping, nr_pages 4166 fs/btrfs/extent_io.c for (i = 0; i < nr_pages; i++) { nr_pages 4273 fs/btrfs/extent_io.c unsigned long nr_pages = (end - start + PAGE_SIZE) >> nr_pages 4284 fs/btrfs/extent_io.c .nr_to_write = nr_pages * 2, nr_pages 4333 fs/btrfs/extent_io.c unsigned nr_pages) nr_pages 416 fs/btrfs/extent_io.h unsigned nr_pages); nr_pages 359 fs/btrfs/inode.c unsigned long nr_pages; nr_pages 385 fs/btrfs/inode.c unsigned long nr_pages, nr_pages 396 fs/btrfs/inode.c async_extent->nr_pages = nr_pages; nr_pages 480 fs/btrfs/inode.c unsigned long nr_pages; nr_pages 507 fs/btrfs/inode.c nr_pages = (end >> PAGE_SHIFT) - (start >> PAGE_SHIFT) + 1; nr_pages 509 fs/btrfs/inode.c nr_pages = min_t(unsigned long, nr_pages, nr_pages 547 fs/btrfs/inode.c pages = kcalloc(nr_pages, sizeof(struct page *), GFP_NOFS); nr_pages 550 fs/btrfs/inode.c nr_pages = 0; nr_pages 581 fs/btrfs/inode.c &nr_pages, nr_pages 587 fs/btrfs/inode.c struct page *page = pages[nr_pages - 1]; nr_pages 643 fs/btrfs/inode.c for (i = 0; i < nr_pages; i++) { nr_pages 676 fs/btrfs/inode.c total_compressed, pages, nr_pages, nr_pages 693 fs/btrfs/inode.c for (i = 0; i < nr_pages; i++) { nr_pages 700 fs/btrfs/inode.c nr_pages = 0; nr_pages 738 fs/btrfs/inode.c for (i = 0; i < async_extent->nr_pages; i++) { nr_pages 743 fs/btrfs/inode.c async_extent->nr_pages = 0; nr_pages 882 fs/btrfs/inode.c async_extent->nr_pages, nr_pages 1174 fs/btrfs/inode.c unsigned long nr_pages; nr_pages 1176 fs/btrfs/inode.c nr_pages = (async_chunk->end - async_chunk->start + PAGE_SIZE) >> nr_pages 1180 fs/btrfs/inode.c if (atomic_sub_return(nr_pages, &fs_info->async_delalloc_pages) < nr_pages 1217 fs/btrfs/inode.c unsigned long nr_pages; nr_pages 1291 fs/btrfs/inode.c nr_pages = DIV_ROUND_UP(cur_end - start, PAGE_SIZE); nr_pages 1292 fs/btrfs/inode.c atomic_add(nr_pages, &fs_info->async_delalloc_pages); nr_pages 1296 fs/btrfs/inode.c *nr_written += nr_pages; nr_pages 8895 fs/btrfs/inode.c struct list_head *pages, unsigned nr_pages) nr_pages 8897 fs/btrfs/inode.c return extent_readpages(mapping, pages, nr_pages); nr_pages 10801 fs/btrfs/inode.c unsigned long nr_pages; nr_pages 10808 fs/btrfs/inode.c unsigned long nr_pages; nr_pages 10818 fs/btrfs/inode.c nr_pages = next_ppage - first_ppage; nr_pages 10828 fs/btrfs/inode.c ret = add_swap_extent(sis, bsi->nr_pages, nr_pages, first_ppage); nr_pages 10832 fs/btrfs/inode.c bsi->nr_pages += nr_pages; nr_pages 11048 fs/btrfs/inode.c sis->max = bsi.nr_pages; nr_pages 11049 fs/btrfs/inode.c sis->pages = bsi.nr_pages - 1; nr_pages 11050 fs/btrfs/inode.c sis->highest_bit = bsi.nr_pages - 1; nr_pages 146 fs/btrfs/lzo.c int nr_pages = 0; nr_pages 182 fs/btrfs/lzo.c nr_pages = 1; nr_pages 237 fs/btrfs/lzo.c if (nr_pages == nr_dest_pages) { nr_pages 249 fs/btrfs/lzo.c pages[nr_pages++] = out_page; nr_pages 294 fs/btrfs/lzo.c *out_pages = nr_pages; nr_pages 132 fs/btrfs/raid56.c int nr_pages; nr_pages 263 fs/btrfs/raid56.c for (i = 0; i < rbio->nr_pages; i++) { nr_pages 310 fs/btrfs/raid56.c for (i = 0; i < dest->nr_pages; i++) { nr_pages 846 fs/btrfs/raid56.c for (i = 0; i < rbio->nr_pages; i++) { nr_pages 1005 fs/btrfs/raid56.c rbio->nr_pages = num_pages; nr_pages 1047 fs/btrfs/raid56.c for (i = 0; i < rbio->nr_pages; i++) { nr_pages 1066 fs/btrfs/raid56.c for (; i < rbio->nr_pages; i++) { nr_pages 316 fs/btrfs/space-info.c unsigned long nr_pages, int nr_items) nr_pages 321 fs/btrfs/space-info.c writeback_inodes_sb_nr(sb, nr_pages, WB_REASON_FS_FREE_SPACE); nr_pages 365 fs/btrfs/space-info.c unsigned long nr_pages; nr_pages 396 fs/btrfs/space-info.c nr_pages = min(delalloc_bytes, to_reclaim) >> PAGE_SHIFT; nr_pages 403 fs/btrfs/space-info.c btrfs_writeback_inodes_sb_nr(fs_info, nr_pages, items); nr_pages 418 fs/btrfs/space-info.c if (async_pages <= nr_pages) nr_pages 421 fs/btrfs/space-info.c async_pages -= nr_pages; nr_pages 26 fs/btrfs/tests/extent-io-tests.c unsigned long nr_pages = end_index - index + 1; nr_pages 31 fs/btrfs/tests/extent-io-tests.c while (nr_pages > 0) { nr_pages 33 fs/btrfs/tests/extent-io-tests.c min_t(unsigned long, nr_pages, nr_pages 45 fs/btrfs/tests/extent-io-tests.c nr_pages -= ret; nr_pages 52 fs/btrfs/tests/extent-io-tests.c start, end, nr_pages, ret); nr_pages 103 fs/btrfs/zlib.c int nr_pages = 0; nr_pages 134 fs/btrfs/zlib.c nr_pages = 1; nr_pages 164 fs/btrfs/zlib.c if (nr_pages == nr_dest_pages) { nr_pages 175 fs/btrfs/zlib.c pages[nr_pages] = out_page; nr_pages 176 fs/btrfs/zlib.c nr_pages++; nr_pages 220 fs/btrfs/zlib.c *out_pages = nr_pages; nr_pages 381 fs/btrfs/zstd.c int nr_pages = 0; nr_pages 418 fs/btrfs/zstd.c pages[nr_pages++] = out_page; nr_pages 455 fs/btrfs/zstd.c if (nr_pages == nr_dest_pages) { nr_pages 465 fs/btrfs/zstd.c pages[nr_pages++] = out_page; nr_pages 515 fs/btrfs/zstd.c if (nr_pages == nr_dest_pages) { nr_pages 525 fs/btrfs/zstd.c pages[nr_pages++] = out_page; nr_pages 540 fs/btrfs/zstd.c *out_pages = nr_pages; nr_pages 684 fs/cachefiles/rdwr.c unsigned *nr_pages, nr_pages 703 fs/cachefiles/rdwr.c *nr_pages); nr_pages 709 fs/cachefiles/rdwr.c if (cachefiles_has_space(cache, 0, *nr_pages) < 0) nr_pages 752 fs/cachefiles/rdwr.c (*nr_pages)--; nr_pages 778 fs/cachefiles/rdwr.c ret, *nr_pages, list_empty(pages) ? " empty" : ""); nr_pages 782 fs/cachefiles/rdwr.c fscache_retrieval_complete(op, *nr_pages); nr_pages 836 fs/cachefiles/rdwr.c unsigned *nr_pages, nr_pages 850 fs/cachefiles/rdwr.c _enter("%p,,,%d,", object, *nr_pages); nr_pages 852 fs/cachefiles/rdwr.c ret = cachefiles_has_space(cache, 0, *nr_pages); nr_pages 868 fs/cachefiles/rdwr.c fscache_retrieval_complete(op, *nr_pages); nr_pages 321 fs/ceph/addr.c int nr_pages = 0; nr_pages 356 fs/ceph/addr.c nr_pages++; nr_pages 358 fs/ceph/addr.c if (max && nr_pages == max) nr_pages 361 fs/ceph/addr.c len = nr_pages << PAGE_SHIFT; nr_pages 362 fs/ceph/addr.c dout("start_read %p nr_pages %d is %lld~%lld\n", inode, nr_pages, nr_pages 376 fs/ceph/addr.c nr_pages = calc_pages_for(0, len); nr_pages 377 fs/ceph/addr.c pages = kmalloc_array(nr_pages, sizeof(*pages), GFP_KERNEL); nr_pages 382 fs/ceph/addr.c for (i = 0; i < nr_pages; ++i) { nr_pages 395 fs/ceph/addr.c nr_pages = i; nr_pages 396 fs/ceph/addr.c if (nr_pages > 0) { nr_pages 397 fs/ceph/addr.c len = nr_pages << PAGE_SHIFT; nr_pages 420 fs/ceph/addr.c return nr_pages; nr_pages 423 fs/ceph/addr.c for (i = 0; i < nr_pages; ++i) { nr_pages 427 fs/ceph/addr.c ceph_put_page_vector(pages, nr_pages, false); nr_pages 442 fs/ceph/addr.c struct list_head *page_list, unsigned nr_pages) nr_pages 455 fs/ceph/addr.c &nr_pages); nr_pages 463 fs/ceph/addr.c inode, file, rw_ctx, nr_pages, max); nr_pages 255 fs/ceph/cache.c unsigned *nr_pages) nr_pages 263 fs/ceph/cache.c ret = fscache_read_or_alloc_pages(ci->fscache, mapping, pages, nr_pages, nr_pages 31 fs/ceph/cache.h unsigned *nr_pages); nr_pages 135 fs/ceph/cache.h unsigned *nr_pages) nr_pages 1334 fs/cifs/cifsglob.h unsigned int nr_pages; nr_pages 1360 fs/cifs/cifsglob.h unsigned int nr_pages; nr_pages 559 fs/cifs/cifsproto.h struct cifs_writedata *cifs_writedata_alloc(unsigned int nr_pages, nr_pages 1702 fs/cifs/cifssmb.c .rq_npages = rdata->nr_pages, nr_pages 2080 fs/cifs/cifssmb.c unsigned int j, nr_pages, wsize, tailsz, cur_len; nr_pages 2084 fs/cifs/cifssmb.c nr_pages = wsize / PAGE_SIZE; nr_pages 2085 fs/cifs/cifssmb.c if (!nr_pages) { nr_pages 2089 fs/cifs/cifssmb.c cur_len = nr_pages * PAGE_SIZE; nr_pages 2092 fs/cifs/cifssmb.c nr_pages = DIV_ROUND_UP(rest_len, PAGE_SIZE); nr_pages 2094 fs/cifs/cifssmb.c tailsz = rest_len - (nr_pages - 1) * PAGE_SIZE; nr_pages 2097 fs/cifs/cifssmb.c wdata2 = cifs_writedata_alloc(nr_pages, cifs_writev_complete); nr_pages 2103 fs/cifs/cifssmb.c for (j = 0; j < nr_pages; j++) { nr_pages 2110 fs/cifs/cifssmb.c wdata2->nr_pages = nr_pages; nr_pages 2129 fs/cifs/cifssmb.c for (j = 0; j < nr_pages; j++) { nr_pages 2142 fs/cifs/cifssmb.c i += nr_pages; nr_pages 2147 fs/cifs/cifssmb.c i += nr_pages; nr_pages 2148 fs/cifs/cifssmb.c } while (i < wdata->nr_pages); nr_pages 2151 fs/cifs/cifssmb.c for (; i < wdata->nr_pages; i++) { nr_pages 2179 fs/cifs/cifssmb.c for (i = 0; i < wdata->nr_pages; i++) { nr_pages 2194 fs/cifs/cifssmb.c cifs_writedata_alloc(unsigned int nr_pages, work_func_t complete) nr_pages 2197 fs/cifs/cifssmb.c kcalloc(nr_pages, sizeof(struct page *), GFP_NOFS); nr_pages 2321 fs/cifs/cifssmb.c rqst.rq_npages = wdata->nr_pages; nr_pages 2175 fs/cifs/file.c unsigned int nr_pages = 0, i; nr_pages 2187 fs/cifs/file.c if (nr_pages == 0) nr_pages 2232 fs/cifs/file.c ++nr_pages; nr_pages 2236 fs/cifs/file.c if (nr_pages == 0) nr_pages 2240 fs/cifs/file.c for (i = nr_pages; i < found_pages; i++) { nr_pages 2245 fs/cifs/file.c return nr_pages; nr_pages 2249 fs/cifs/file.c wdata_send_pages(struct cifs_writedata *wdata, unsigned int nr_pages, nr_pages 2257 fs/cifs/file.c wdata->nr_pages = nr_pages; nr_pages 2261 fs/cifs/file.c page_offset(wdata->pages[nr_pages - 1]), nr_pages 2263 fs/cifs/file.c wdata->bytes = ((nr_pages - 1) * PAGE_SIZE) + wdata->tailsz; nr_pages 2313 fs/cifs/file.c unsigned int i, nr_pages, found_pages, wsize; nr_pages 2352 fs/cifs/file.c nr_pages = wdata_prepare_pages(wdata, found_pages, mapping, wbc, nr_pages 2356 fs/cifs/file.c if (nr_pages == 0) { nr_pages 2374 fs/cifs/file.c rc = wdata_send_pages(wdata, nr_pages, mapping, wbc); nr_pages 2376 fs/cifs/file.c for (i = 0; i < nr_pages; ++i) nr_pages 2382 fs/cifs/file.c for (i = 0; i < nr_pages; ++i) { nr_pages 2410 fs/cifs/file.c wbc->nr_to_write -= nr_pages; nr_pages 2686 fs/cifs/file.c for (i = 0; i < wdata->nr_pages; i++) nr_pages 2718 fs/cifs/file.c unsigned long i, nr_pages = *num_pages; nr_pages 2721 fs/cifs/file.c for (i = 0; i < nr_pages; i++) { nr_pages 2826 fs/cifs/file.c unsigned long nr_pages, num_pages, i; nr_pages 2885 fs/cifs/file.c nr_pages = nr_pages 2899 fs/cifs/file.c nr_pages > 1 ? nr_pages 2901 fs/cifs/file.c (nr_pages - 2) * PAGE_SIZE : nr_pages 2904 fs/cifs/file.c nr_pages = get_numpages(wsize, len, &cur_len); nr_pages 2905 fs/cifs/file.c wdata = cifs_writedata_alloc(nr_pages, nr_pages 2913 fs/cifs/file.c rc = cifs_write_allocate_pages(wdata->pages, nr_pages); nr_pages 2921 fs/cifs/file.c num_pages = nr_pages; nr_pages 2925 fs/cifs/file.c for (i = 0; i < nr_pages; i++) nr_pages 2937 fs/cifs/file.c for ( ; nr_pages > num_pages; nr_pages--) nr_pages 2938 fs/cifs/file.c put_page(wdata->pages[nr_pages - 1]); nr_pages 2940 fs/cifs/file.c wdata->tailsz = cur_len - ((nr_pages - 1) * PAGE_SIZE); nr_pages 2944 fs/cifs/file.c wdata->nr_pages = nr_pages; nr_pages 3285 fs/cifs/file.c cifs_readdata_alloc(unsigned int nr_pages, work_func_t complete) nr_pages 3288 fs/cifs/file.c kcalloc(nr_pages, sizeof(struct page *), GFP_KERNEL); nr_pages 3319 fs/cifs/file.c cifs_read_allocate_pages(struct cifs_readdata *rdata, unsigned int nr_pages) nr_pages 3325 fs/cifs/file.c for (i = 0; i < nr_pages; i++) { nr_pages 3353 fs/cifs/file.c for (i = 0; i < rdata->nr_pages; i++) { nr_pages 3374 fs/cifs/file.c for (i = 0; i < rdata->nr_pages; i++) { nr_pages 3414 fs/cifs/file.c unsigned int nr_pages = rdata->nr_pages; nr_pages 3419 fs/cifs/file.c for (i = 0; i < nr_pages; i++) { nr_pages 3433 fs/cifs/file.c rdata->nr_pages--; nr_pages 3647 fs/cifs/file.c rdata->nr_pages = npages; nr_pages 4099 fs/cifs/file.c for (i = 0; i < rdata->nr_pages; i++) { nr_pages 4133 fs/cifs/file.c unsigned int nr_pages = rdata->nr_pages; nr_pages 4143 fs/cifs/file.c for (i = 0; i < nr_pages; i++) { nr_pages 4178 fs/cifs/file.c rdata->nr_pages--; nr_pages 4186 fs/cifs/file.c rdata->nr_pages--; nr_pages 4228 fs/cifs/file.c unsigned int *nr_pages, loff_t *offset, unsigned int *bytes) nr_pages 4257 fs/cifs/file.c *nr_pages = 1; nr_pages 4279 fs/cifs/file.c (*nr_pages)++; nr_pages 4333 fs/cifs/file.c unsigned int i, nr_pages, bytes, rsize; nr_pages 4366 fs/cifs/file.c &nr_pages, &offset, &bytes); nr_pages 4372 fs/cifs/file.c rdata = cifs_readdata_alloc(nr_pages, cifs_readv_complete); nr_pages 4399 fs/cifs/file.c rdata->pages[rdata->nr_pages++] = page; nr_pages 4413 fs/cifs/file.c for (i = 0; i < rdata->nr_pages; i++) { nr_pages 278 fs/cifs/fscache.c unsigned *nr_pages) nr_pages 283 fs/cifs/fscache.c __func__, CIFS_I(inode)->fscache, *nr_pages, inode); nr_pages 285 fs/cifs/fscache.c pages, nr_pages, nr_pages 95 fs/cifs/fscache.h unsigned *nr_pages) nr_pages 99 fs/cifs/fscache.h nr_pages); nr_pages 149 fs/cifs/fscache.h unsigned *nr_pages) nr_pages 3578 fs/cifs/smb2pdu.c rdata->nr_pages, rdata->page_offset, nr_pages 3639 fs/cifs/smb2pdu.c .rq_npages = rdata->nr_pages, nr_pages 3986 fs/cifs/smb2pdu.c wdata->nr_pages, wdata->page_offset, nr_pages 3994 fs/cifs/smb2pdu.c if (wdata->nr_pages > 1) nr_pages 3997 fs/cifs/smb2pdu.c (wdata->nr_pages - 1) * wdata->pagesz - nr_pages 4022 fs/cifs/smb2pdu.c rqst.rq_npages = wdata->nr_pages; nr_pages 728 fs/direct-io.c int ret, nr_pages; nr_pages 734 fs/direct-io.c nr_pages = min(sdio->pages_in_io, BIO_MAX_PAGES); nr_pages 735 fs/direct-io.c BUG_ON(nr_pages <= 0); nr_pages 736 fs/direct-io.c dio_bio_alloc(dio, sdio, map_bh->b_bdev, sector, nr_pages); nr_pages 286 fs/erofs/data.c unsigned int nr_pages) nr_pages 293 fs/erofs/data.c trace_erofs_readpages(mapping->host, page, nr_pages, true); nr_pages 295 fs/erofs/data.c for (; nr_pages; --nr_pages) { nr_pages 303 fs/erofs/data.c &last_block, nr_pages, true); nr_pages 64 fs/erofs/zdata.c cl->nr_pages = 0; nr_pages 76 fs/erofs/zdata.c DBG_BUGON(cl->nr_pages); nr_pages 679 fs/erofs/zdata.c clt->cl->nr_pages = max_t(pgoff_t, clt->cl->nr_pages, index + 1); nr_pages 762 fs/erofs/zdata.c unsigned int i, outputsize, llen, nr_pages; nr_pages 773 fs/erofs/zdata.c DBG_BUGON(!READ_ONCE(cl->nr_pages)); nr_pages 776 fs/erofs/zdata.c nr_pages = cl->nr_pages; nr_pages 778 fs/erofs/zdata.c if (nr_pages <= Z_EROFS_VMAP_ONSTACK_PAGES) { nr_pages 780 fs/erofs/zdata.c } else if (nr_pages <= Z_EROFS_VMAP_GLOBAL_PAGES && nr_pages 786 fs/erofs/zdata.c if (nr_pages > Z_EROFS_VMAP_GLOBAL_PAGES) nr_pages 789 fs/erofs/zdata.c pages = kvmalloc_array(nr_pages, sizeof(struct page *), nr_pages 799 fs/erofs/zdata.c for (i = 0; i < nr_pages; ++i) nr_pages 823 fs/erofs/zdata.c DBG_BUGON(pagenr >= nr_pages); nr_pages 864 fs/erofs/zdata.c DBG_BUGON(pagenr >= nr_pages); nr_pages 887 fs/erofs/zdata.c if (nr_pages << PAGE_SHIFT >= cl->pageofs + llen) { nr_pages 891 fs/erofs/zdata.c outputsize = (nr_pages << PAGE_SHIFT) - cl->pageofs; nr_pages 921 fs/erofs/zdata.c for (i = 0; i < nr_pages; ++i) { nr_pages 943 fs/erofs/zdata.c cl->nr_pages = 0; nr_pages 1370 fs/erofs/zdata.c unsigned int nr_pages) nr_pages 1375 fs/erofs/zdata.c bool sync = should_decompress_synchronously(sbi, nr_pages); nr_pages 1382 fs/erofs/zdata.c nr_pages, false); nr_pages 1386 fs/erofs/zdata.c for (; nr_pages; --nr_pages) { nr_pages 32 fs/erofs/zdata.h unsigned short nr_pages; nr_pages 882 fs/ext2/inode.c struct list_head *pages, unsigned nr_pages) nr_pages 884 fs/ext2/inode.c return mpage_readpages(mapping, pages, nr_pages, ext2_get_block); nr_pages 3241 fs/ext4/ext4.h unsigned nr_pages, bool is_readahead); nr_pages 1692 fs/ext4/inode.c int nr_pages, i; nr_pages 1713 fs/ext4/inode.c nr_pages = pagevec_lookup_range(&pvec, mapping, &index, end); nr_pages 1714 fs/ext4/inode.c if (nr_pages == 0) nr_pages 1716 fs/ext4/inode.c for (i = 0; i < nr_pages; i++) { nr_pages 2369 fs/ext4/inode.c int nr_pages, i; nr_pages 2385 fs/ext4/inode.c nr_pages = pagevec_lookup_range(&pvec, inode->i_mapping, nr_pages 2387 fs/ext4/inode.c if (nr_pages == 0) nr_pages 2389 fs/ext4/inode.c for (i = 0; i < nr_pages; i++) { nr_pages 2635 fs/ext4/inode.c unsigned int nr_pages; nr_pages 2654 fs/ext4/inode.c nr_pages = pagevec_lookup_range_tag(&pvec, mapping, &index, end, nr_pages 2656 fs/ext4/inode.c if (nr_pages == 0) nr_pages 2659 fs/ext4/inode.c for (i = 0; i < nr_pages; i++) { nr_pages 3343 fs/ext4/inode.c struct list_head *pages, unsigned nr_pages) nr_pages 3351 fs/ext4/inode.c return ext4_mpage_readpages(mapping, pages, NULL, nr_pages, true); nr_pages 227 fs/ext4/readpage.c unsigned nr_pages, bool is_readahead) nr_pages 251 fs/ext4/readpage.c for (; nr_pages; nr_pages--) { nr_pages 269 fs/ext4/readpage.c last_block = block_in_file + nr_pages * blocks_per_page; nr_pages 377 fs/ext4/readpage.c min_t(int, nr_pages, BIO_MAX_PAGES)); nr_pages 369 fs/f2fs/checkpoint.c int nr_pages; nr_pages 379 fs/f2fs/checkpoint.c while ((nr_pages = pagevec_lookup_tag(&pvec, mapping, &index, nr_pages 383 fs/f2fs/checkpoint.c for (i = 0; i < nr_pages; i++) { nr_pages 677 fs/f2fs/data.c unsigned nr_pages, unsigned op_flag, nr_pages 685 fs/f2fs/data.c bio = f2fs_bio_alloc(sbi, min_t(int, nr_pages, BIO_MAX_PAGES), false); nr_pages 1639 fs/f2fs/data.c unsigned nr_pages, nr_pages 1655 fs/f2fs/data.c last_block = block_in_file + nr_pages; nr_pages 1723 fs/f2fs/data.c bio = f2fs_grab_read_bio(inode, block_nr, nr_pages, nr_pages 1767 fs/f2fs/data.c unsigned nr_pages, bool is_readahead) nr_pages 1784 fs/f2fs/data.c for (; nr_pages; nr_pages--) { nr_pages 1796 fs/f2fs/data.c ret = f2fs_read_single_page(inode, page, nr_pages, &map, &bio, nr_pages 1831 fs/f2fs/data.c struct list_head *pages, unsigned nr_pages) nr_pages 1836 fs/f2fs/data.c trace_f2fs_readpages(inode, page, nr_pages); nr_pages 1842 fs/f2fs/data.c return f2fs_mpage_readpages(mapping, pages, NULL, nr_pages, true); nr_pages 2256 fs/f2fs/data.c int nr_pages; nr_pages 2300 fs/f2fs/data.c nr_pages = pagevec_lookup_range_tag(&pvec, mapping, &index, end, nr_pages 2302 fs/f2fs/data.c if (nr_pages == 0) nr_pages 2305 fs/f2fs/data.c for (i = 0; i < nr_pages; i++) { nr_pages 1274 fs/f2fs/f2fs.h atomic_t nr_pages[NR_COUNT_TYPE]; nr_pages 1870 fs/f2fs/f2fs.h atomic_inc(&sbi->nr_pages[count_type]); nr_pages 1891 fs/f2fs/f2fs.h atomic_dec(&sbi->nr_pages[count_type]); nr_pages 1909 fs/f2fs/f2fs.h return atomic_read(&sbi->nr_pages[count_type]); nr_pages 347 fs/f2fs/file.c int nr_pages; nr_pages 353 fs/f2fs/file.c nr_pages = find_get_pages_tag(mapping, &pgofs, PAGECACHE_TAG_DIRTY, nr_pages 355 fs/f2fs/file.c if (!nr_pages) nr_pages 1447 fs/f2fs/node.c int nr_pages; nr_pages 1452 fs/f2fs/node.c while ((nr_pages = pagevec_lookup_tag(&pvec, NODE_MAPPING(sbi), &index, nr_pages 1456 fs/f2fs/node.c for (i = 0; i < nr_pages; i++) { nr_pages 1659 fs/f2fs/node.c int nr_pages; nr_pages 1671 fs/f2fs/node.c while ((nr_pages = pagevec_lookup_tag(&pvec, NODE_MAPPING(sbi), &index, nr_pages 1675 fs/f2fs/node.c for (i = 0; i < nr_pages; i++) { nr_pages 1817 fs/f2fs/node.c int nr_pages, done = 0; nr_pages 1824 fs/f2fs/node.c while (!done && (nr_pages = pagevec_lookup_tag(&pvec, nr_pages 1828 fs/f2fs/node.c for (i = 0; i < nr_pages; i++) { nr_pages 2832 fs/f2fs/super.c atomic_set(&sbi->nr_pages[i], 0); nr_pages 213 fs/fat/inode.c struct list_head *pages, unsigned nr_pages) nr_pages 215 fs/fat/inode.c return mpage_readpages(mapping, pages, nr_pages, fat_get_block); nr_pages 393 fs/file_table.c unsigned long nr_pages = totalram_pages(); nr_pages 394 fs/file_table.c unsigned long memreserve = (nr_pages - nr_free_pages()) * 3/2; nr_pages 396 fs/file_table.c memreserve = min(memreserve, nr_pages - 1); nr_pages 397 fs/file_table.c n = ((nr_pages - memreserve) * (PAGE_SIZE / 1024)) / 10; nr_pages 43 fs/fs-writeback.c long nr_pages; nr_pages 809 fs/fs-writeback.c static long wb_split_bdi_pages(struct bdi_writeback *wb, long nr_pages) nr_pages 814 fs/fs-writeback.c if (nr_pages == LONG_MAX) nr_pages 823 fs/fs-writeback.c return nr_pages; nr_pages 825 fs/fs-writeback.c return DIV_ROUND_UP_ULL((u64)nr_pages * this_bw, tot_bw); nr_pages 854 fs/fs-writeback.c long nr_pages; nr_pages 869 fs/fs-writeback.c nr_pages = wb_split_bdi_pages(wb, base_work->nr_pages); nr_pages 874 fs/fs-writeback.c work->nr_pages = nr_pages; nr_pages 883 fs/fs-writeback.c work->nr_pages = nr_pages; nr_pages 970 fs/fs-writeback.c work->nr_pages = nr; nr_pages 1047 fs/fs-writeback.c static long wb_split_bdi_pages(struct bdi_writeback *wb, long nr_pages) nr_pages 1049 fs/fs-writeback.c return nr_pages; nr_pages 1607 fs/fs-writeback.c pages = min(pages, work->nr_pages); nr_pages 1719 fs/fs-writeback.c work->nr_pages -= write_chunk - wbc.nr_to_write; nr_pages 1759 fs/fs-writeback.c if (work->nr_pages <= 0) nr_pages 1792 fs/fs-writeback.c if (work->nr_pages <= 0) nr_pages 1800 fs/fs-writeback.c static long writeback_inodes_wb(struct bdi_writeback *wb, long nr_pages, nr_pages 1804 fs/fs-writeback.c .nr_pages = nr_pages, nr_pages 1819 fs/fs-writeback.c return nr_pages - work.nr_pages; nr_pages 1841 fs/fs-writeback.c long nr_pages = work->nr_pages; nr_pages 1856 fs/fs-writeback.c if (work->nr_pages <= 0) nr_pages 1930 fs/fs-writeback.c return nr_pages - work->nr_pages; nr_pages 1955 fs/fs-writeback.c .nr_pages = LONG_MAX, nr_pages 1971 fs/fs-writeback.c long nr_pages; nr_pages 1985 fs/fs-writeback.c nr_pages = get_nr_dirty_pages(); nr_pages 1987 fs/fs-writeback.c if (nr_pages) { nr_pages 1989 fs/fs-writeback.c .nr_pages = nr_pages, nr_pages 2004 fs/fs-writeback.c long nr_pages; nr_pages 2009 fs/fs-writeback.c nr_pages = get_nr_dirty_pages(); nr_pages 2010 fs/fs-writeback.c if (nr_pages) { nr_pages 2012 fs/fs-writeback.c .nr_pages = wb_split_bdi_pages(wb, nr_pages), nr_pages 2018 fs/fs-writeback.c nr_pages = wb_writeback(wb, &work); nr_pages 2022 fs/fs-writeback.c return nr_pages; nr_pages 2460 fs/fs-writeback.c .nr_pages = nr, nr_pages 2536 fs/fs-writeback.c .nr_pages = LONG_MAX, nr_pages 561 fs/fscache/page.c unsigned *nr_pages, nr_pages 571 fs/fscache/page.c _enter("%p,,%d,,,", cookie, *nr_pages); nr_pages 584 fs/fscache/page.c ASSERTCMP(*nr_pages, >, 0); nr_pages 593 fs/fscache/page.c atomic_set(&op->n_pages, *nr_pages); nr_pages 627 fs/fscache/page.c op, pages, nr_pages, gfp); nr_pages 632 fs/fscache/page.c op, pages, nr_pages, gfp); nr_pages 922 fs/fuse/file.c unsigned int nr_pages; nr_pages 940 fs/fuse/file.c data->max_pages = min_t(unsigned int, data->nr_pages, nr_pages 961 fs/fuse/file.c data->nr_pages--; nr_pages 966 fs/fuse/file.c struct list_head *pages, unsigned nr_pages) nr_pages 979 fs/fuse/file.c data.nr_pages = nr_pages; nr_pages 980 fs/fuse/file.c data.max_pages = min_t(unsigned int, nr_pages, fc->max_pages); nr_pages 1230 fs/fuse/file.c unsigned int nr_pages = fuse_wr_pages(pos, iov_iter_count(ii), nr_pages 1233 fs/fuse/file.c ap->pages = fuse_pages_alloc(nr_pages, GFP_KERNEL, &ap->descs); nr_pages 1239 fs/fuse/file.c count = fuse_fill_write_pages(ap, mapping, ii, pos, nr_pages); nr_pages 1346 fs/fuse/file.c unsigned int nr_pages) nr_pages 1350 fs/fuse/file.c for (i = index; i < index + nr_pages; i++) nr_pages 241 fs/gfs2/aops.c int nr_pages, nr_pages 246 fs/gfs2/aops.c unsigned nrblocks = nr_pages * (PAGE_SIZE >> inode->i_blkbits); nr_pages 254 fs/gfs2/aops.c for(i = 0; i < nr_pages; i++) { nr_pages 339 fs/gfs2/aops.c int nr_pages; nr_pages 374 fs/gfs2/aops.c nr_pages = pagevec_lookup_range_tag(&pvec, mapping, &index, end, nr_pages 376 fs/gfs2/aops.c if (nr_pages == 0) nr_pages 379 fs/gfs2/aops.c ret = gfs2_write_jdata_pagevec(mapping, wbc, &pvec, nr_pages, &done_index); nr_pages 600 fs/gfs2/aops.c struct list_head *pages, unsigned nr_pages) nr_pages 613 fs/gfs2/aops.c ret = mpage_readpages(mapping, pages, nr_pages, gfs2_block_map); nr_pages 129 fs/hpfs/file.c struct list_head *pages, unsigned nr_pages) nr_pages 131 fs/hpfs/file.c return mpage_readpages(mapping, pages, nr_pages, hpfs_get_block); nr_pages 3312 fs/io_uring.c static void io_unaccount_mem(struct user_struct *user, unsigned long nr_pages) nr_pages 3314 fs/io_uring.c atomic_long_sub(nr_pages, &user->locked_vm); nr_pages 3317 fs/io_uring.c static int io_account_mem(struct user_struct *user, unsigned long nr_pages) nr_pages 3326 fs/io_uring.c new_pages = cur_pages + nr_pages; nr_pages 3467 fs/io_uring.c int pret, nr_pages; nr_pages 3491 fs/io_uring.c nr_pages = end - start; nr_pages 3494 fs/io_uring.c ret = io_account_mem(ctx->user, nr_pages); nr_pages 3500 fs/io_uring.c if (!pages || nr_pages > got_pages) { nr_pages 3503 fs/io_uring.c pages = kvmalloc_array(nr_pages, sizeof(struct page *), nr_pages 3505 fs/io_uring.c vmas = kvmalloc_array(nr_pages, nr_pages 3511 fs/io_uring.c io_unaccount_mem(ctx->user, nr_pages); nr_pages 3514 fs/io_uring.c got_pages = nr_pages; nr_pages 3517 fs/io_uring.c imu->bvec = kvmalloc_array(nr_pages, sizeof(struct bio_vec), nr_pages 3522 fs/io_uring.c io_unaccount_mem(ctx->user, nr_pages); nr_pages 3528 fs/io_uring.c pret = get_user_pages(ubuf, nr_pages, nr_pages 3531 fs/io_uring.c if (pret == nr_pages) { nr_pages 3533 fs/io_uring.c for (j = 0; j < nr_pages; j++) { nr_pages 3554 fs/io_uring.c io_unaccount_mem(ctx->user, nr_pages); nr_pages 3561 fs/io_uring.c for (j = 0; j < nr_pages; j++) { nr_pages 3574 fs/io_uring.c imu->nr_bvecs = nr_pages; nr_pages 382 fs/iomap/buffered-io.c unsigned nr_pages, const struct iomap_ops *ops) nr_pages 208 fs/iomap/direct-io.c int nr_pages, ret = 0; nr_pages 245 fs/iomap/direct-io.c nr_pages = iov_iter_npages(&iter, BIO_MAX_PAGES); nr_pages 246 fs/iomap/direct-io.c if (nr_pages <= 0) nr_pages 247 fs/iomap/direct-io.c return nr_pages; nr_pages 263 fs/iomap/direct-io.c bio = bio_alloc(GFP_KERNEL, nr_pages); nr_pages 303 fs/iomap/direct-io.c nr_pages = iov_iter_npages(&iter, BIO_MAX_PAGES); nr_pages 305 fs/iomap/direct-io.c } while (nr_pages); nr_pages 88 fs/iomap/seek.c unsigned nr_pages, i; nr_pages 90 fs/iomap/seek.c nr_pages = pagevec_lookup_range(&pvec, inode->i_mapping, &index, nr_pages 92 fs/iomap/seek.c if (nr_pages == 0) nr_pages 95 fs/iomap/seek.c for (i = 0; i < nr_pages; i++) { nr_pages 19 fs/iomap/swapfile.c unsigned long nr_pages; /* number of pages collected */ nr_pages 32 fs/iomap/swapfile.c unsigned long nr_pages; nr_pages 49 fs/iomap/swapfile.c nr_pages = next_ppage - first_ppage; nr_pages 65 fs/iomap/swapfile.c error = add_swap_extent(isi->sis, isi->nr_pages, nr_pages, first_ppage); nr_pages 69 fs/iomap/swapfile.c isi->nr_pages += nr_pages; nr_pages 173 fs/iomap/swapfile.c sis->max = isi.nr_pages; nr_pages 174 fs/iomap/swapfile.c sis->pages = isi.nr_pages - 1; nr_pages 175 fs/iomap/swapfile.c sis->highest_bit = isi.nr_pages - 1; nr_pages 1189 fs/isofs/inode.c struct list_head *pages, unsigned nr_pages) nr_pages 1191 fs/isofs/inode.c return mpage_readpages(mapping, pages, nr_pages, isofs_get_block); nr_pages 300 fs/jfs/inode.c struct list_head *pages, unsigned nr_pages) nr_pages 302 fs/jfs/inode.c return mpage_readpages(mapping, pages, nr_pages, jfs_get_block); nr_pages 139 fs/mpage.c unsigned int nr_pages; nr_pages 190 fs/mpage.c last_block = block_in_file + args->nr_pages * blocks_per_page; nr_pages 307 fs/mpage.c min_t(int, args->nr_pages, nr_pages 386 fs/mpage.c unsigned nr_pages, get_block_t get_block) nr_pages 394 fs/mpage.c for (page_idx = 0; page_idx < nr_pages; page_idx++) { nr_pages 403 fs/mpage.c args.nr_pages = nr_pages - page_idx; nr_pages 422 fs/mpage.c .nr_pages = 1, nr_pages 495 fs/nfs/blocklayout/extent_tree.c int nr_pages = DIV_ROUND_UP(buffer_size, PAGE_SIZE), i; nr_pages 497 fs/nfs/blocklayout/extent_tree.c for (i = 0; i < nr_pages; i++) nr_pages 448 fs/nfs/fscache.c unsigned *nr_pages) nr_pages 450 fs/nfs/fscache.c unsigned npages = *nr_pages; nr_pages 457 fs/nfs/fscache.c mapping, pages, nr_pages, nr_pages 461 fs/nfs/fscache.c if (*nr_pages < npages) nr_pages 464 fs/nfs/fscache.c if (*nr_pages > 0) nr_pages 466 fs/nfs/fscache.c *nr_pages); nr_pages 471 fs/nfs/fscache.c BUG_ON(*nr_pages != 0); nr_pages 146 fs/nfs/fscache.h unsigned *nr_pages) nr_pages 150 fs/nfs/fscache.h nr_pages); nr_pages 225 fs/nfs/fscache.h unsigned *nr_pages) nr_pages 405 fs/nfs/read.c struct list_head *pages, unsigned nr_pages) nr_pages 419 fs/nfs/read.c nr_pages); nr_pages 436 fs/nfs/read.c pages, &nr_pages); nr_pages 157 fs/nilfs2/inode.c struct list_head *pages, unsigned int nr_pages) nr_pages 159 fs/nilfs2/inode.c return mpage_readpages(mapping, pages, nr_pages, nilfs_get_block); nr_pages 497 fs/ntfs/compress.c unsigned int nr_pages = (end_vcn - start_vcn) << nr_pages 507 fs/ntfs/compress.c "%i.", index, cb_size, nr_pages); nr_pages 515 fs/ntfs/compress.c pages = kmalloc_array(nr_pages, sizeof(struct page *), GFP_NOFS); nr_pages 516 fs/ntfs/compress.c completed_pages = kmalloc_array(nr_pages + 1, sizeof(int), GFP_NOFS); nr_pages 559 fs/ntfs/compress.c if (nr_pages < max_page) nr_pages 560 fs/ntfs/compress.c max_page = nr_pages; nr_pages 495 fs/ntfs/file.c pgoff_t index, const unsigned nr_pages, struct page **pages, nr_pages 500 fs/ntfs/file.c BUG_ON(!nr_pages); nr_pages 526 fs/ntfs/file.c } while (nr < nr_pages); nr_pages 571 fs/ntfs/file.c unsigned nr_pages, s64 pos, size_t bytes) nr_pages 599 fs/ntfs/file.c BUG_ON(!nr_pages); nr_pages 607 fs/ntfs/file.c vi->i_ino, ni->type, pages[0]->index, nr_pages, nr_pages 624 fs/ntfs/file.c } while (++u < nr_pages); nr_pages 1176 fs/ntfs/file.c if (likely(!err && ++u < nr_pages)) nr_pages 1222 fs/ntfs/file.c } while (++u < nr_pages); nr_pages 1330 fs/ntfs/file.c nr_pages = u; nr_pages 1337 fs/ntfs/file.c if (u == nr_pages && nr_pages 1355 fs/ntfs/file.c } while (++u <= nr_pages); nr_pages 1361 fs/ntfs/file.c unsigned nr_pages) nr_pages 1363 fs/ntfs/file.c BUG_ON(!nr_pages); nr_pages 1370 fs/ntfs/file.c --nr_pages; nr_pages 1371 fs/ntfs/file.c flush_dcache_page(pages[nr_pages]); nr_pages 1372 fs/ntfs/file.c } while (nr_pages > 0); nr_pages 1385 fs/ntfs/file.c struct page **pages, const unsigned nr_pages, nr_pages 1431 fs/ntfs/file.c } while (++u < nr_pages); nr_pages 1540 fs/ntfs/file.c const unsigned nr_pages, s64 pos, size_t bytes) nr_pages 1555 fs/ntfs/file.c BUG_ON(!nr_pages); nr_pages 1563 fs/ntfs/file.c vi->i_ino, ni->type, page->index, nr_pages, nr_pages 1567 fs/ntfs/file.c nr_pages, pos, bytes); nr_pages 1568 fs/ntfs/file.c BUG_ON(nr_pages > 1); nr_pages 1685 fs/ntfs/file.c static size_t ntfs_copy_from_user_iter(struct page **pages, unsigned nr_pages, nr_pages 1688 fs/ntfs/file.c struct page **last_page = pages + nr_pages; nr_pages 1743 fs/ntfs/file.c unsigned nr_pages; nr_pages 1774 fs/ntfs/file.c nr_pages = 1; nr_pages 1776 fs/ntfs/file.c nr_pages = vol->cluster_size >> PAGE_SHIFT; nr_pages 1788 fs/ntfs/file.c if (nr_pages > 1) { nr_pages 1823 fs/ntfs/file.c do_pages = nr_pages; nr_pages 355 fs/ocfs2/aops.c struct list_head *pages, unsigned nr_pages) nr_pages 392 fs/ocfs2/aops.c err = mpage_readpages(mapping, pages, nr_pages, ocfs2_get_block); nr_pages 293 fs/omfs/file.c struct list_head *pages, unsigned nr_pages) nr_pages 295 fs/omfs/file.c return mpage_readpages(mapping, pages, nr_pages, omfs_get_block); nr_pages 1057 fs/pipe.c unsigned int size, nr_pages; nr_pages 1062 fs/pipe.c nr_pages = size >> PAGE_SHIFT; nr_pages 1064 fs/pipe.c if (!nr_pages) nr_pages 1074 fs/pipe.c if (nr_pages > pipe->buffers && nr_pages 1078 fs/pipe.c user_bufs = account_pipe_buffers(pipe->user, pipe->buffers, nr_pages); nr_pages 1080 fs/pipe.c if (nr_pages > pipe->buffers && nr_pages 1094 fs/pipe.c if (nr_pages < pipe->nrbufs) { nr_pages 1099 fs/pipe.c bufs = kcalloc(nr_pages, sizeof(*bufs), nr_pages 1130 fs/pipe.c pipe->buffers = nr_pages; nr_pages 1131 fs/pipe.c return nr_pages * PAGE_SIZE; nr_pages 1134 fs/pipe.c (void) account_pipe_buffers(pipe->user, nr_pages, pipe->buffers); nr_pages 148 fs/proc/kcore.c unsigned long nr_pages = ent->size >> PAGE_SHIFT; nr_pages 154 fs/proc/kcore.c end = ((unsigned long)pfn_to_page(pfn + nr_pages)) - 1; nr_pages 186 fs/proc/kcore.c kclist_add_private(unsigned long pfn, unsigned long nr_pages, void *arg) nr_pages 203 fs/proc/kcore.c ent->size = nr_pages << PAGE_SHIFT; nr_pages 1668 fs/proc/task_mmu.c unsigned long nr_pages) nr_pages 1672 fs/proc/task_mmu.c md->pages += nr_pages; nr_pages 1674 fs/proc/task_mmu.c md->dirty += nr_pages; nr_pages 1677 fs/proc/task_mmu.c md->swapcache += nr_pages; nr_pages 1680 fs/proc/task_mmu.c md->active += nr_pages; nr_pages 1683 fs/proc/task_mmu.c md->writeback += nr_pages; nr_pages 1686 fs/proc/task_mmu.c md->anon += nr_pages; nr_pages 1691 fs/proc/task_mmu.c md->node[page_to_nid(page)] += nr_pages; nr_pages 103 fs/qnx6/inode.c struct list_head *pages, unsigned nr_pages) nr_pages 105 fs/qnx6/inode.c return mpage_readpages(mapping, pages, nr_pages, qnx6_get_block); nr_pages 1165 fs/reiserfs/inode.c struct list_head *pages, unsigned nr_pages) nr_pages 1167 fs/reiserfs/inode.c return mpage_readpages(mapping, pages, nr_pages, reiserfs_get_block); nr_pages 187 fs/splice.c unsigned int spd_pages = spd->nr_pages; nr_pages 214 fs/splice.c if (!--spd->nr_pages) nr_pages 372 fs/splice.c unsigned int nr_pages; nr_pages 392 fs/splice.c nr_pages = DIV_ROUND_UP(res + base, PAGE_SIZE); nr_pages 395 fs/splice.c if (nr_pages > PIPE_DEF_BUFFERS) { nr_pages 396 fs/splice.c vec = kmalloc_array(nr_pages, sizeof(struct kvec), GFP_KERNEL); nr_pages 406 fs/splice.c for (i = 0; i < nr_pages; i++) { nr_pages 414 fs/splice.c res = kernel_readv(in, vec, nr_pages, *ppos); nr_pages 423 fs/splice.c for (i = 0; i < nr_pages; i++) nr_pages 199 fs/udf/inode.c struct list_head *pages, unsigned nr_pages) nr_pages 201 fs/udf/inode.c return mpage_readpages(mapping, pages, nr_pages, udf_get_block); nr_pages 1172 fs/xfs/xfs_aops.c unsigned nr_pages) nr_pages 1174 fs/xfs/xfs_aops.c trace_xfs_vm_readpages(mapping->host, nr_pages); nr_pages 1175 fs/xfs/xfs_aops.c return iomap_readpages(mapping, pages, nr_pages, &xfs_iomap_ops); nr_pages 1269 fs/xfs/xfs_buf.c int nr_pages; nr_pages 1293 fs/xfs/xfs_buf.c nr_pages = min(total_nr_pages, BIO_MAX_PAGES); nr_pages 1295 fs/xfs/xfs_buf.c bio = bio_alloc(GFP_NOIO, nr_pages); nr_pages 1302 fs/xfs/xfs_buf.c for (; size && nr_pages; nr_pages--, page_index++) { nr_pages 1201 fs/xfs/xfs_trace.h TP_PROTO(struct inode *inode, int nr_pages), nr_pages 1202 fs/xfs/xfs_trace.h TP_ARGS(inode, nr_pages), nr_pages 1206 fs/xfs/xfs_trace.h __field(int, nr_pages) nr_pages 1211 fs/xfs/xfs_trace.h __entry->nr_pages = nr_pages; nr_pages 1216 fs/xfs/xfs_trace.h __entry->nr_pages) nr_pages 1221 fs/xfs/xfs_trace.h TP_PROTO(struct inode *inode, int nr_pages), \ nr_pages 1222 fs/xfs/xfs_trace.h TP_ARGS(inode, nr_pages)) nr_pages 93 include/drm/drm_prime.h struct sg_table *drm_prime_pages_to_sg(struct page **pages, unsigned int nr_pages); nr_pages 517 include/linux/ceph/osd_client.h struct page **pages, int nr_pages, nr_pages 527 include/linux/ceph/osd_client.h struct page **pages, int nr_pages); nr_pages 804 include/linux/compat.h asmlinkage long compat_sys_move_pages(pid_t pid, compat_ulong_t nr_pages, nr_pages 214 include/linux/coresight.h int nr_pages, bool overwrite); nr_pages 204 include/linux/dax.h long dax_direct_access(struct dax_device *dax_dev, pgoff_t pgoff, long nr_pages, nr_pages 140 include/linux/device-mapper.h long nr_pages, void **kaddr, pfn_t *pfn); nr_pages 452 include/linux/edac.h u32 nr_pages; /* number of pages on this dimm */ nr_pages 377 include/linux/fs.h struct list_head *pages, unsigned nr_pages); nr_pages 160 include/linux/fscache-cache.h unsigned *nr_pages, nr_pages 592 include/linux/fscache.h unsigned *nr_pages, nr_pages 599 include/linux/fscache.h nr_pages, end_io_func, nr_pages 616 include/linux/gfp.h void free_contig_range(unsigned long pfn, unsigned int nr_pages); nr_pages 54 include/linux/hugetlb_cgroup.h extern int hugetlb_cgroup_charge_cgroup(int idx, unsigned long nr_pages, nr_pages 56 include/linux/hugetlb_cgroup.h extern void hugetlb_cgroup_commit_charge(int idx, unsigned long nr_pages, nr_pages 59 include/linux/hugetlb_cgroup.h extern void hugetlb_cgroup_uncharge_page(int idx, unsigned long nr_pages, nr_pages 61 include/linux/hugetlb_cgroup.h extern void hugetlb_cgroup_uncharge_cgroup(int idx, unsigned long nr_pages, nr_pages 85 include/linux/hugetlb_cgroup.h hugetlb_cgroup_charge_cgroup(int idx, unsigned long nr_pages, nr_pages 92 include/linux/hugetlb_cgroup.h hugetlb_cgroup_commit_charge(int idx, unsigned long nr_pages, nr_pages 99 include/linux/hugetlb_cgroup.h hugetlb_cgroup_uncharge_page(int idx, unsigned long nr_pages, struct page *page) nr_pages 104 include/linux/hugetlb_cgroup.h hugetlb_cgroup_uncharge_cgroup(int idx, unsigned long nr_pages, nr_pages 156 include/linux/iomap.h unsigned nr_pages, const struct iomap_ops *ops); nr_pages 280 include/linux/ioport.h walk_system_ram_range(unsigned long start_pfn, unsigned long nr_pages, nr_pages 703 include/linux/kvm_host.h struct page **pages, int nr_pages); nr_pages 537 include/linux/memcontrol.h int zid, int nr_pages); nr_pages 1336 include/linux/memcontrol.h bool mem_cgroup_charge_skmem(struct mem_cgroup *memcg, unsigned int nr_pages); nr_pages 1337 include/linux/memcontrol.h void mem_cgroup_uncharge_skmem(struct mem_cgroup *memcg, unsigned int nr_pages); nr_pages 1382 include/linux/memcontrol.h unsigned int nr_pages); nr_pages 52 include/linux/memory.h unsigned long nr_pages; nr_pages 67 include/linux/memory.h unsigned int nr_pages; /* # pages in range to check */ nr_pages 94 include/linux/memory_hotplug.h extern int zone_grow_waitqueues(struct zone *zone, unsigned long nr_pages); nr_pages 128 include/linux/memory_hotplug.h extern void __remove_pages(unsigned long start_pfn, unsigned long nr_pages, nr_pages 132 include/linux/memory_hotplug.h extern int __add_pages(int nid, unsigned long start_pfn, unsigned long nr_pages, nr_pages 137 include/linux/memory_hotplug.h unsigned long nr_pages, struct mhp_restrictions *restrictions) nr_pages 139 include/linux/memory_hotplug.h return __add_pages(nid, start_pfn, nr_pages, restrictions); nr_pages 142 include/linux/memory_hotplug.h int add_pages(int nid, unsigned long start_pfn, unsigned long nr_pages, nr_pages 314 include/linux/memory_hotplug.h extern bool is_mem_section_removable(unsigned long pfn, unsigned long nr_pages); nr_pages 316 include/linux/memory_hotplug.h extern int offline_pages(unsigned long start_pfn, unsigned long nr_pages); nr_pages 322 include/linux/memory_hotplug.h unsigned long nr_pages) nr_pages 329 include/linux/memory_hotplug.h static inline int offline_pages(unsigned long start_pfn, unsigned long nr_pages) nr_pages 347 include/linux/memory_hotplug.h unsigned long nr_pages, struct vmem_altmap *altmap); nr_pages 350 include/linux/memory_hotplug.h unsigned long nr_pages); nr_pages 353 include/linux/memory_hotplug.h unsigned long nr_pages, struct vmem_altmap *altmap); nr_pages 355 include/linux/memory_hotplug.h unsigned long pfn, unsigned long nr_pages, nr_pages 359 include/linux/memory_hotplug.h extern bool allow_online_pfn_range(int nid, unsigned long pfn, unsigned long nr_pages, nr_pages 362 include/linux/memory_hotplug.h unsigned long nr_pages); nr_pages 1533 include/linux/mm.h unsigned long start, unsigned long nr_pages, nr_pages 1536 include/linux/mm.h long get_user_pages(unsigned long start, unsigned long nr_pages, nr_pages 1539 include/linux/mm.h long get_user_pages_locked(unsigned long start, unsigned long nr_pages, nr_pages 1541 include/linux/mm.h long get_user_pages_unlocked(unsigned long start, unsigned long nr_pages, nr_pages 1544 include/linux/mm.h int get_user_pages_fast(unsigned long start, int nr_pages, nr_pages 1594 include/linux/mm.h int get_kernel_pages(const struct kvec *iov, int nr_pages, int write, nr_pages 1638 include/linux/mm.h int __get_user_pages_fast(unsigned long start, int nr_pages, int write, nr_pages 2774 include/linux/mm.h unsigned long nr_pages, int nid, struct vmem_altmap *altmap); nr_pages 2795 include/linux/mm.h unsigned long nr_pages); nr_pages 28 include/linux/mm_inline.h int nr_pages) nr_pages 32 include/linux/mm_inline.h __mod_lruvec_state(lruvec, NR_LRU_BASE + lru, nr_pages); nr_pages 34 include/linux/mm_inline.h NR_ZONE_LRU_BASE + lru, nr_pages); nr_pages 39 include/linux/mm_inline.h int nr_pages) nr_pages 41 include/linux/mm_inline.h __update_lru_size(lruvec, lru, zid, nr_pages); nr_pages 43 include/linux/mm_inline.h mem_cgroup_update_lru_size(lruvec, lru, zid, nr_pages); nr_pages 616 include/linux/mmzone.h unsigned long start_pfn, unsigned long nr_pages) nr_pages 621 include/linux/mmzone.h start_pfn + nr_pages <= zone->zone_start_pfn) nr_pages 1180 include/linux/mmzone.h void subsection_map_init(unsigned long pfn, unsigned long nr_pages); nr_pages 18 include/linux/mpage.h unsigned nr_pages, get_block_t get_block); nr_pages 50 include/linux/page_counter.h void page_counter_cancel(struct page_counter *counter, unsigned long nr_pages); nr_pages 51 include/linux/page_counter.h void page_counter_charge(struct page_counter *counter, unsigned long nr_pages); nr_pages 53 include/linux/page_counter.h unsigned long nr_pages, nr_pages 55 include/linux/page_counter.h void page_counter_uncharge(struct page_counter *counter, unsigned long nr_pages); nr_pages 56 include/linux/page_counter.h void page_counter_set_min(struct page_counter *counter, unsigned long nr_pages); nr_pages 57 include/linux/page_counter.h void page_counter_set_low(struct page_counter *counter, unsigned long nr_pages); nr_pages 58 include/linux/page_counter.h int page_counter_set_max(struct page_counter *counter, unsigned long nr_pages); nr_pages 60 include/linux/page_counter.h unsigned long *nr_pages); nr_pages 352 include/linux/pagemap.h pgoff_t end, unsigned int nr_pages, nr_pages 355 include/linux/pagemap.h pgoff_t *start, unsigned int nr_pages, nr_pages 358 include/linux/pagemap.h return find_get_pages_range(mapping, start, (pgoff_t)-1, nr_pages, nr_pages 362 include/linux/pagemap.h unsigned int nr_pages, struct page **pages); nr_pages 364 include/linux/pagemap.h pgoff_t end, xa_mark_t tag, unsigned int nr_pages, nr_pages 367 include/linux/pagemap.h pgoff_t *index, xa_mark_t tag, unsigned int nr_pages, nr_pages 371 include/linux/pagemap.h nr_pages, pages); nr_pages 417 include/linux/perf_event.h int nr_pages, bool overwrite); nr_pages 112 include/linux/scif.h int nr_pages; nr_pages 59 include/linux/splice.h int nr_pages; /* number of populated pages in map */ nr_pages 153 include/linux/swap.h pgoff_t nr_pages; nr_pages 356 include/linux/swap.h unsigned long nr_pages, nr_pages 363 include/linux/swap.h extern unsigned long shrink_all_memory(unsigned long nr_pages); nr_pages 396 include/linux/swap.h unsigned long nr_pages, sector_t start_block); nr_pages 473 include/linux/swap.h extern int init_swap_address_space(unsigned int type, unsigned long nr_pages); nr_pages 661 include/linux/swap.h extern void mem_cgroup_uncharge_swap(swp_entry_t entry, unsigned int nr_pages); nr_pages 676 include/linux/swap.h unsigned int nr_pages) nr_pages 894 include/linux/syscalls.h asmlinkage long sys_move_pages(pid_t pid, unsigned long nr_pages, nr_pages 47 include/linux/vmalloc.h unsigned int nr_pages; nr_pages 374 include/linux/vmstat.h static inline void __mod_zone_freepage_state(struct zone *zone, int nr_pages, nr_pages 377 include/linux/vmstat.h __mod_zone_page_state(zone, NR_FREE_PAGES, nr_pages); nr_pages 379 include/linux/vmstat.h __mod_zone_page_state(zone, NR_FREE_CMA_PAGES, nr_pages); nr_pages 220 include/linux/writeback.h int cgroup_writeback_by_id(u64 bdi_id, int memcg_id, unsigned long nr_pages, nr_pages 63 include/media/videobuf-dma-sg.h int nr_pages; nr_pages 353 include/trace/events/writeback.h __field(long, nr_pages) nr_pages 364 include/trace/events/writeback.h __entry->nr_pages = work->nr_pages; nr_pages 377 include/trace/events/writeback.h __entry->nr_pages, nr_pages 98 include/xen/arm/page.h unsigned long nr_pages); nr_pages 27 include/xen/balloon.h int alloc_xenballooned_pages(int nr_pages, struct page **pages); nr_pages 28 include/xen/balloon.h void free_xenballooned_pages(int nr_pages, struct page **pages); nr_pages 198 include/xen/grant_table.h int gnttab_alloc_pages(int nr_pages, struct page **pages); nr_pages 199 include/xen/grant_table.h void gnttab_free_pages(int nr_pages, struct page **pages); nr_pages 208 include/xen/grant_table.h int nr_pages; nr_pages 219 include/xen/grant_table.h int gnttab_pages_set_private(int nr_pages, struct page **pages); nr_pages 220 include/xen/grant_table.h void gnttab_pages_clear_private(int nr_pages, struct page **pages); nr_pages 638 include/xen/interface/xen.h unsigned long nr_pages; /* Total pages allocated to this domain. */ nr_pages 205 include/xen/xenbus.h unsigned int nr_pages, grant_ref_t *grefs); nr_pages 934 kernel/dma/debug.c int i, nr_pages; nr_pages 949 kernel/dma/debug.c nr_pages = DIV_ROUND_UP(nr_prealloc_entries, DMA_DEBUG_DYNAMIC_ENTRIES); nr_pages 950 kernel/dma/debug.c for (i = 0; i < nr_pages; ++i) nr_pages 1110 kernel/dma/debug.c for (i = 0; i < stack_vm_area->nr_pages; i++) { nr_pages 126 kernel/dma/remap.c unsigned long nr_pages = atomic_pool_size >> PAGE_SHIFT; nr_pages 132 kernel/dma/remap.c page = dma_alloc_from_contiguous(NULL, nr_pages, nr_pages 167 kernel/dma/remap.c if (!dma_release_from_contiguous(NULL, page, nr_pages)) nr_pages 5224 kernel/events/core.c if (!rb || !rb->nr_pages) { nr_pages 5712 kernel/events/core.c unsigned long nr_pages; nr_pages 5730 kernel/events/core.c nr_pages = (vma_size / PAGE_SIZE) - 1; nr_pages 5742 kernel/events/core.c nr_pages = vma_size / PAGE_SIZE; nr_pages 5764 kernel/events/core.c if (aux_size != vma_size || aux_size != nr_pages * PAGE_SIZE) nr_pages 5768 kernel/events/core.c if (rb_has_aux(rb) && rb->aux_nr_pages != nr_pages) nr_pages 5771 kernel/events/core.c if (!is_power_of_2(nr_pages)) nr_pages 5784 kernel/events/core.c user_extra = nr_pages; nr_pages 5793 kernel/events/core.c if (nr_pages != 0 && !is_power_of_2(nr_pages)) nr_pages 5796 kernel/events/core.c if (vma_size != PAGE_SIZE * (1 + nr_pages)) nr_pages 5803 kernel/events/core.c if (event->rb->nr_pages != nr_pages) { nr_pages 5821 kernel/events/core.c user_extra = nr_pages + 1; nr_pages 5872 kernel/events/core.c rb = rb_alloc(nr_pages, nr_pages 5890 kernel/events/core.c ret = rb_alloc_aux(rb, event, vma->vm_pgoff, nr_pages, nr_pages 20 kernel/events/internal.h int nr_pages; /* nr of data pages */ nr_pages 72 kernel/events/internal.h if (!pause && rb->nr_pages) nr_pages 79 kernel/events/internal.h rb_alloc(int nr_pages, long watermark, int cpu, int flags); nr_pages 82 kernel/events/internal.h pgoff_t pgoff, int nr_pages, long watermark, int flags); nr_pages 120 kernel/events/internal.h return rb->nr_pages << (PAGE_SHIFT + page_order(rb)); nr_pages 146 kernel/events/internal.h handle->page &= rb->nr_pages - 1; \ nr_pages 174 kernel/events/ring_buffer.c if (rb->nr_pages) nr_pages 234 kernel/events/ring_buffer.c handle->page = (offset >> page_shift) & (rb->nr_pages - 1); nr_pages 328 kernel/events/ring_buffer.c if (!rb->nr_pages) nr_pages 630 kernel/events/ring_buffer.c pgoff_t pgoff, int nr_pages, long watermark, int flags) nr_pages 643 kernel/events/ring_buffer.c max_order = ilog2(nr_pages); nr_pages 656 kernel/events/ring_buffer.c rb->aux_pages = kcalloc_node(nr_pages, sizeof(void *), GFP_KERNEL, nr_pages 662 kernel/events/ring_buffer.c for (rb->aux_nr_pages = 0; rb->aux_nr_pages < nr_pages;) { nr_pages 666 kernel/events/ring_buffer.c order = min(max_order, ilog2(nr_pages - rb->aux_nr_pages)); nr_pages 690 kernel/events/ring_buffer.c rb->aux_priv = event->pmu->setup_aux(event, rb->aux_pages, nr_pages, nr_pages 709 kernel/events/ring_buffer.c rb->aux_watermark = nr_pages << (PAGE_SHIFT - 1); nr_pages 735 kernel/events/ring_buffer.c if (pgoff > rb->nr_pages) nr_pages 757 kernel/events/ring_buffer.c struct ring_buffer *rb_alloc(int nr_pages, long watermark, int cpu, int flags) nr_pages 764 kernel/events/ring_buffer.c size += nr_pages * sizeof(void *); nr_pages 777 kernel/events/ring_buffer.c for (i = 0; i < nr_pages; i++) { nr_pages 783 kernel/events/ring_buffer.c rb->nr_pages = nr_pages; nr_pages 815 kernel/events/ring_buffer.c for (i = 0; i < rb->nr_pages; i++) nr_pages 823 kernel/events/ring_buffer.c return rb->nr_pages << page_order(rb); nr_pages 866 kernel/events/ring_buffer.c struct ring_buffer *rb_alloc(int nr_pages, long watermark, int cpu, int flags) nr_pages 881 kernel/events/ring_buffer.c all_buf = vmalloc_user((nr_pages + 1) * PAGE_SIZE); nr_pages 887 kernel/events/ring_buffer.c if (nr_pages) { nr_pages 888 kernel/events/ring_buffer.c rb->nr_pages = 1; nr_pages 889 kernel/events/ring_buffer.c rb->page_order = ilog2(nr_pages); nr_pages 380 kernel/fork.c BUG_ON(vm->nr_pages != THREAD_SIZE / PAGE_SIZE); nr_pages 760 kernel/fork.c unsigned long nr_pages = totalram_pages(); nr_pages 766 kernel/fork.c if (fls64(nr_pages) + fls64(PAGE_SIZE) > 64) nr_pages 769 kernel/fork.c threads = div64_u64((u64) nr_pages * (u64) PAGE_SIZE, nr_pages 153 kernel/kexec_core.c unsigned long nr_pages = totalram_pages(); nr_pages 219 kernel/kexec_core.c if (PAGE_COUNT(image->segment[i].memsz) > nr_pages / 2) nr_pages 225 kernel/kexec_core.c if (total_pages > nr_pages / 2) nr_pages 240 kernel/power/hibernate.c unsigned nr_pages, char *msg) nr_pages 253 kernel/power/hibernate.c k = nr_pages * (PAGE_SIZE / 1024); nr_pages 1521 kernel/power/snapshot.c static unsigned long preallocate_image_pages(unsigned long nr_pages, gfp_t mask) nr_pages 1525 kernel/power/snapshot.c while (nr_pages > 0) { nr_pages 1536 kernel/power/snapshot.c nr_pages--; nr_pages 1543 kernel/power/snapshot.c static unsigned long preallocate_image_memory(unsigned long nr_pages, nr_pages 1552 kernel/power/snapshot.c if (nr_pages < alloc) nr_pages 1553 kernel/power/snapshot.c alloc = nr_pages; nr_pages 1559 kernel/power/snapshot.c static unsigned long preallocate_image_highmem(unsigned long nr_pages) nr_pages 1561 kernel/power/snapshot.c return preallocate_image_pages(nr_pages, GFP_IMAGE | __GFP_HIGHMEM); nr_pages 1574 kernel/power/snapshot.c static unsigned long preallocate_highmem_fraction(unsigned long nr_pages, nr_pages 1578 kernel/power/snapshot.c unsigned long alloc = __fraction(nr_pages, highmem, total); nr_pages 1583 kernel/power/snapshot.c static inline unsigned long preallocate_image_highmem(unsigned long nr_pages) nr_pages 1588 kernel/power/snapshot.c static inline unsigned long preallocate_highmem_fraction(unsigned long nr_pages, nr_pages 1875 kernel/power/snapshot.c static int enough_free_mem(unsigned int nr_pages, unsigned int nr_highmem) nr_pages 1884 kernel/power/snapshot.c nr_pages += count_pages_for_highmem(nr_highmem); nr_pages 1886 kernel/power/snapshot.c nr_pages, PAGES_FOR_IO, free); nr_pages 1888 kernel/power/snapshot.c return free > nr_pages + PAGES_FOR_IO; nr_pages 1946 kernel/power/snapshot.c unsigned int nr_pages, unsigned int nr_highmem) nr_pages 1953 kernel/power/snapshot.c nr_pages += alloc_highmem_pages(copy_bm, nr_highmem); nr_pages 1956 kernel/power/snapshot.c if (nr_pages > alloc_normal) { nr_pages 1957 kernel/power/snapshot.c nr_pages -= alloc_normal; nr_pages 1958 kernel/power/snapshot.c while (nr_pages-- > 0) { nr_pages 1977 kernel/power/snapshot.c unsigned int nr_pages, nr_highmem; nr_pages 1982 kernel/power/snapshot.c nr_pages = count_data_pages(); nr_pages 1984 kernel/power/snapshot.c pr_info("Need to copy %u pages\n", nr_pages + nr_highmem); nr_pages 1986 kernel/power/snapshot.c if (!enough_free_mem(nr_pages, nr_highmem)) { nr_pages 1991 kernel/power/snapshot.c if (swsusp_alloc(©_bm, nr_pages, nr_highmem)) { nr_pages 2009 kernel/power/snapshot.c nr_pages += nr_highmem; nr_pages 2010 kernel/power/snapshot.c nr_copy_pages = nr_pages; nr_pages 2011 kernel/power/snapshot.c nr_meta_pages = DIV_ROUND_UP(nr_pages * sizeof(long), PAGE_SIZE); nr_pages 2013 kernel/power/snapshot.c pr_info("Hibernation image created (%d pages copied)\n", nr_pages); nr_pages 2464 kernel/power/snapshot.c unsigned int nr_pages, nr_highmem; nr_pages 2495 kernel/power/snapshot.c nr_pages = nr_copy_pages - nr_highmem - allocated_unsafe_pages; nr_pages 2496 kernel/power/snapshot.c nr_pages = DIV_ROUND_UP(nr_pages, PBES_PER_LINKED_PAGE); nr_pages 2497 kernel/power/snapshot.c while (nr_pages > 0) { nr_pages 2505 kernel/power/snapshot.c nr_pages--; nr_pages 2508 kernel/power/snapshot.c nr_pages = nr_copy_pages - nr_highmem - allocated_unsafe_pages; nr_pages 2509 kernel/power/snapshot.c while (nr_pages > 0) { nr_pages 2523 kernel/power/snapshot.c nr_pages--; nr_pages 536 kernel/power/swap.c int nr_pages; nr_pages 549 kernel/power/swap.c nr_pages = 0; nr_pages 558 kernel/power/swap.c if (!(nr_pages % m)) nr_pages 560 kernel/power/swap.c nr_pages / m * 10); nr_pages 561 kernel/power/swap.c nr_pages++; nr_pages 672 kernel/power/swap.c int nr_pages; nr_pages 767 kernel/power/swap.c nr_pages = 0; nr_pages 782 kernel/power/swap.c if (!(nr_pages % m)) nr_pages 784 kernel/power/swap.c nr_pages / m * 10); nr_pages 785 kernel/power/swap.c nr_pages++; nr_pages 880 kernel/power/swap.c static int enough_swap(unsigned int nr_pages) nr_pages 887 kernel/power/swap.c required = PAGES_FOR_IO + nr_pages; nr_pages 1059 kernel/power/swap.c unsigned nr_pages; nr_pages 1068 kernel/power/swap.c nr_pages = 0; nr_pages 1081 kernel/power/swap.c if (!(nr_pages % m)) nr_pages 1083 kernel/power/swap.c nr_pages / m * 10); nr_pages 1084 kernel/power/swap.c nr_pages++; nr_pages 1164 kernel/power/swap.c unsigned nr_pages; nr_pages 1284 kernel/power/swap.c nr_pages = 0; nr_pages 1408 kernel/power/swap.c if (!(nr_pages % m)) nr_pages 1410 kernel/power/swap.c nr_pages / m * 10); nr_pages 1411 kernel/power/swap.c nr_pages++; nr_pages 1205 kernel/relay.c unsigned int pidx, poff, total_len, subbuf_pages, nr_pages; nr_pages 1218 kernel/relay.c .nr_pages = 0, nr_pages 1240 kernel/relay.c nr_pages = min_t(unsigned int, subbuf_pages, spd.nr_pages_max); nr_pages 1242 kernel/relay.c for (total_len = 0; spd.nr_pages < nr_pages; spd.nr_pages++) { nr_pages 1252 kernel/relay.c spd.pages[spd.nr_pages] = rbuf->page_array[pidx]; nr_pages 1253 kernel/relay.c spd.partial[spd.nr_pages].offset = poff; nr_pages 1260 kernel/relay.c spd.partial[spd.nr_pages].len = this_len; nr_pages 1261 kernel/relay.c spd.partial[spd.nr_pages].private = private; nr_pages 1269 kernel/relay.c spd.nr_pages++; nr_pages 1275 kernel/relay.c if (!spd.nr_pages) nr_pages 475 kernel/resource.c int walk_system_ram_range(unsigned long start_pfn, unsigned long nr_pages, nr_pages 485 kernel/resource.c end = ((u64)(start_pfn + nr_pages) << PAGE_SHIFT) - 1; nr_pages 501 kernel/resource.c static int __is_ram(unsigned long pfn, unsigned long nr_pages, void *arg) nr_pages 451 kernel/trace/ring_buffer.c unsigned long nr_pages; nr_pages 523 kernel/trace/ring_buffer.c return buffer->buffers[cpu]->nr_pages; nr_pages 643 kernel/trace/ring_buffer.c size_t nr_pages; nr_pages 651 kernel/trace/ring_buffer.c nr_pages = cpu_buffer->nr_pages; nr_pages 658 kernel/trace/ring_buffer.c (!nr_pages || (dirty * 100) > full * nr_pages)) nr_pages 1189 kernel/trace/ring_buffer.c static int __rb_allocate_pages(long nr_pages, struct list_head *pages, int cpu) nr_pages 1204 kernel/trace/ring_buffer.c if (i < nr_pages) nr_pages 1225 kernel/trace/ring_buffer.c for (i = 0; i < nr_pages; i++) { nr_pages 1261 kernel/trace/ring_buffer.c unsigned long nr_pages) nr_pages 1265 kernel/trace/ring_buffer.c WARN_ON(!nr_pages); nr_pages 1267 kernel/trace/ring_buffer.c if (__rb_allocate_pages(nr_pages, &pages, cpu_buffer->cpu)) nr_pages 1278 kernel/trace/ring_buffer.c cpu_buffer->nr_pages = nr_pages; nr_pages 1286 kernel/trace/ring_buffer.c rb_allocate_cpu_buffer(struct ring_buffer *buffer, long nr_pages, int cpu) nr_pages 1326 kernel/trace/ring_buffer.c ret = rb_allocate_pages(cpu_buffer, nr_pages); nr_pages 1381 kernel/trace/ring_buffer.c long nr_pages; nr_pages 1395 kernel/trace/ring_buffer.c nr_pages = DIV_ROUND_UP(size, BUF_PAGE_SIZE); nr_pages 1404 kernel/trace/ring_buffer.c if (nr_pages < 2) nr_pages 1405 kernel/trace/ring_buffer.c nr_pages = 2; nr_pages 1417 kernel/trace/ring_buffer.c buffer->buffers[cpu] = rb_allocate_cpu_buffer(buffer, nr_pages, cpu); nr_pages 1495 kernel/trace/ring_buffer.c rb_remove_pages(struct ring_buffer_per_cpu *cpu_buffer, unsigned long nr_pages) nr_pages 1531 kernel/trace/ring_buffer.c for (nr_removed = 0; nr_removed < nr_pages; nr_removed++) { nr_pages 1694 kernel/trace/ring_buffer.c cpu_buffer->nr_pages += cpu_buffer->nr_pages_to_update; nr_pages 1719 kernel/trace/ring_buffer.c unsigned long nr_pages; nr_pages 1733 kernel/trace/ring_buffer.c nr_pages = DIV_ROUND_UP(size, BUF_PAGE_SIZE); nr_pages 1736 kernel/trace/ring_buffer.c if (nr_pages < 2) nr_pages 1737 kernel/trace/ring_buffer.c nr_pages = 2; nr_pages 1739 kernel/trace/ring_buffer.c size = nr_pages * BUF_PAGE_SIZE; nr_pages 1757 kernel/trace/ring_buffer.c cpu_buffer->nr_pages_to_update = nr_pages - nr_pages 1758 kernel/trace/ring_buffer.c cpu_buffer->nr_pages; nr_pages 1817 kernel/trace/ring_buffer.c if (nr_pages == cpu_buffer->nr_pages) nr_pages 1820 kernel/trace/ring_buffer.c cpu_buffer->nr_pages_to_update = nr_pages - nr_pages 1821 kernel/trace/ring_buffer.c cpu_buffer->nr_pages; nr_pages 2474 kernel/trace/ring_buffer.c max_count = cpu_buffer->nr_pages * 100; nr_pages 2614 kernel/trace/ring_buffer.c size_t nr_pages; nr_pages 2642 kernel/trace/ring_buffer.c nr_pages = cpu_buffer->nr_pages; nr_pages 2644 kernel/trace/ring_buffer.c if (full && nr_pages && (dirty * 100) <= full * nr_pages) nr_pages 4346 kernel/trace/ring_buffer.c return BUF_PAGE_SIZE * buffer->buffers[cpu]->nr_pages; nr_pages 4528 kernel/trace/ring_buffer.c if (cpu_buffer_a->nr_pages != cpu_buffer_b->nr_pages) nr_pages 4874 kernel/trace/ring_buffer.c unsigned long nr_pages; nr_pages 4880 kernel/trace/ring_buffer.c nr_pages = 0; nr_pages 4885 kernel/trace/ring_buffer.c if (nr_pages == 0) nr_pages 4886 kernel/trace/ring_buffer.c nr_pages = buffer->buffers[cpu_i]->nr_pages; nr_pages 4887 kernel/trace/ring_buffer.c if (nr_pages != buffer->buffers[cpu_i]->nr_pages) { nr_pages 4894 kernel/trace/ring_buffer.c nr_pages = 2; nr_pages 4896 kernel/trace/ring_buffer.c rb_allocate_cpu_buffer(buffer, nr_pages, cpu); nr_pages 6175 kernel/trace/trace.c .nr_pages = 0, /* This gets updated below. */ nr_pages 6234 kernel/trace/trace.c spd.nr_pages = i; nr_pages 7482 kernel/trace/trace.c spd.nr_pages++; nr_pages 7489 kernel/trace/trace.c spd.nr_pages = i; nr_pages 7492 kernel/trace/trace.c if (!spd.nr_pages) { nr_pages 71 mm/compaction.c unsigned int i, order, nr_pages; nr_pages 79 mm/compaction.c nr_pages = 1 << order; nr_pages 85 mm/compaction.c for (i = 0; i < nr_pages; i++) { nr_pages 509 mm/filemap.c int nr_pages; nr_pages 518 mm/filemap.c nr_pages = pagevec_lookup_range_tag(&pvec, mapping, &index, nr_pages 520 mm/filemap.c if (!nr_pages) nr_pages 523 mm/filemap.c for (i = 0; i < nr_pages; i++) { nr_pages 1792 mm/filemap.c pgoff_t end, unsigned int nr_pages, nr_pages 1799 mm/filemap.c if (unlikely(!nr_pages)) nr_pages 1818 mm/filemap.c if (++ret == nr_pages) { nr_pages 1858 mm/filemap.c unsigned int nr_pages, struct page **pages) nr_pages 1864 mm/filemap.c if (unlikely(!nr_pages)) nr_pages 1886 mm/filemap.c if (++ret == nr_pages) nr_pages 1914 mm/filemap.c pgoff_t end, xa_mark_t tag, unsigned int nr_pages, nr_pages 1921 mm/filemap.c if (unlikely(!nr_pages)) nr_pages 1944 mm/filemap.c if (++ret == nr_pages) { nr_pages 789 mm/gup.c unsigned long start, unsigned long nr_pages, nr_pages 797 mm/gup.c if (!nr_pages) nr_pages 838 mm/gup.c &start, &nr_pages, i, nr_pages 898 mm/gup.c if (page_increm > nr_pages) nr_pages 899 mm/gup.c page_increm = nr_pages; nr_pages 902 mm/gup.c nr_pages -= page_increm; nr_pages 903 mm/gup.c } while (nr_pages); nr_pages 1016 mm/gup.c unsigned long nr_pages, nr_pages 1038 mm/gup.c ret = __get_user_pages(tsk, mm, start, nr_pages, flags, pages, nr_pages 1047 mm/gup.c BUG_ON(ret >= nr_pages); nr_pages 1051 mm/gup.c nr_pages -= ret; nr_pages 1053 mm/gup.c if (!nr_pages) nr_pages 1089 mm/gup.c nr_pages--; nr_pages 1091 mm/gup.c if (!nr_pages) nr_pages 1165 mm/gup.c unsigned long start, unsigned long nr_pages, nr_pages 1178 mm/gup.c return __get_user_pages_locked(tsk, mm, start, nr_pages, pages, vmas, nr_pages 1207 mm/gup.c unsigned long nr_pages = (end - start) / PAGE_SIZE; nr_pages 1238 mm/gup.c return __get_user_pages(current, mm, start, nr_pages, gup_flags, nr_pages 1333 mm/gup.c unsigned long nr_pages, struct page **pages, nr_pages 1349 mm/gup.c for (i = 0; i < nr_pages; i++) { nr_pages 1377 mm/gup.c static bool check_dax_vmas(struct vm_area_struct **vmas, long nr_pages) nr_pages 1382 mm/gup.c for (i = 0; i < nr_pages; i++) { nr_pages 1451 mm/gup.c unsigned long nr_pages, nr_pages 1463 mm/gup.c for (i = 0; i < nr_pages;) { nr_pages 1503 mm/gup.c for (i = 0; i < nr_pages; i++) nr_pages 1522 mm/gup.c nr_pages = __get_user_pages_locked(tsk, mm, start, nr_pages, nr_pages 1526 mm/gup.c if ((nr_pages > 0) && migrate_allow) { nr_pages 1532 mm/gup.c return nr_pages; nr_pages 1538 mm/gup.c unsigned long nr_pages, nr_pages 1543 mm/gup.c return nr_pages; nr_pages 1554 mm/gup.c unsigned long nr_pages, nr_pages 1568 mm/gup.c vmas_tmp = kcalloc(nr_pages, nr_pages 1577 mm/gup.c rc = __get_user_pages_locked(tsk, mm, start, nr_pages, pages, nr_pages 1605 mm/gup.c unsigned long nr_pages, nr_pages 1610 mm/gup.c return __get_user_pages_locked(tsk, mm, start, nr_pages, pages, vmas, nr_pages 1622 mm/gup.c long get_user_pages(unsigned long start, unsigned long nr_pages, nr_pages 1626 mm/gup.c return __gup_longterm_locked(current, current->mm, start, nr_pages, nr_pages 1652 mm/gup.c long get_user_pages_locked(unsigned long start, unsigned long nr_pages, nr_pages 1665 mm/gup.c return __get_user_pages_locked(current, current->mm, start, nr_pages, nr_pages 1686 mm/gup.c long get_user_pages_unlocked(unsigned long start, unsigned long nr_pages, nr_pages 1703 mm/gup.c ret = __get_user_pages_locked(current, mm, start, nr_pages, pages, NULL, nr_pages 2339 mm/gup.c int __get_user_pages_fast(unsigned long start, int nr_pages, int write, nr_pages 2347 mm/gup.c len = (unsigned long) nr_pages << PAGE_SHIFT; nr_pages 2384 mm/gup.c static int __gup_longterm_unlocked(unsigned long start, int nr_pages, nr_pages 2396 mm/gup.c start, nr_pages, nr_pages 2400 mm/gup.c ret = get_user_pages_unlocked(start, nr_pages, nr_pages 2423 mm/gup.c int get_user_pages_fast(unsigned long start, int nr_pages, nr_pages 2435 mm/gup.c len = (unsigned long) nr_pages << PAGE_SHIFT; nr_pages 2458 mm/gup.c if (nr < nr_pages) { nr_pages 2463 mm/gup.c ret = __gup_longterm_unlocked(start, nr_pages - nr, nr_pages 26 mm/gup_benchmark.c unsigned long i, nr_pages, addr, next; nr_pages 34 mm/gup_benchmark.c nr_pages = gup->size / PAGE_SIZE; nr_pages 35 mm/gup_benchmark.c pages = kvcalloc(nr_pages, sizeof(void *), GFP_KERNEL); nr_pages 82 mm/gup_benchmark.c for (i = 0; i < nr_pages; i++) { nr_pages 1054 mm/hugetlb.c int nr_pages = 1 << order; nr_pages 1058 mm/hugetlb.c for (i = 1; i < nr_pages; i++, p = mem_map_next(p, page, i)) { nr_pages 1074 mm/hugetlb.c unsigned long nr_pages, gfp_t gfp_mask) nr_pages 1076 mm/hugetlb.c unsigned long end_pfn = start_pfn + nr_pages; nr_pages 1082 mm/hugetlb.c unsigned long start_pfn, unsigned long nr_pages) nr_pages 1084 mm/hugetlb.c unsigned long i, end_pfn = start_pfn + nr_pages; nr_pages 1109 mm/hugetlb.c unsigned long start_pfn, unsigned long nr_pages) nr_pages 1111 mm/hugetlb.c unsigned long last_pfn = start_pfn + nr_pages - 1; nr_pages 1119 mm/hugetlb.c unsigned long nr_pages = 1 << order; nr_pages 1129 mm/hugetlb.c pfn = ALIGN(zone->zone_start_pfn, nr_pages); nr_pages 1130 mm/hugetlb.c while (zone_spans_last_pfn(zone, pfn, nr_pages)) { nr_pages 1131 mm/hugetlb.c if (pfn_range_valid_gigantic(zone, pfn, nr_pages)) { nr_pages 1140 mm/hugetlb.c ret = __alloc_gigantic_page(pfn, nr_pages, gfp_mask); nr_pages 1145 mm/hugetlb.c pfn += nr_pages; nr_pages 1384 mm/hugetlb.c int nr_pages = 1 << order; nr_pages 1391 mm/hugetlb.c for (i = 1; i < nr_pages; i++, p = mem_map_next(p, page, i)) { nr_pages 1917 mm/hugetlb.c unsigned long nr_pages; nr_pages 1927 mm/hugetlb.c nr_pages = min(unused_resv_pages, h->surplus_huge_pages); nr_pages 1941 mm/hugetlb.c while (nr_pages--) { nr_pages 4389 mm/hugetlb.c unsigned long *position, unsigned long *nr_pages, nr_pages 4394 mm/hugetlb.c unsigned long remainder = *nr_pages; nr_pages 4481 mm/hugetlb.c *nr_pages = 0; nr_pages 4534 mm/hugetlb.c *nr_pages = remainder; nr_pages 128 mm/hugetlb_cgroup.c unsigned int nr_pages; nr_pages 142 mm/hugetlb_cgroup.c nr_pages = compound_nr(page); nr_pages 146 mm/hugetlb_cgroup.c page_counter_charge(&parent->hugepage[idx], nr_pages); nr_pages 150 mm/hugetlb_cgroup.c page_counter_cancel(counter, nr_pages); nr_pages 181 mm/hugetlb_cgroup.c int hugetlb_cgroup_charge_cgroup(int idx, unsigned long nr_pages, nr_pages 205 mm/hugetlb_cgroup.c if (!page_counter_try_charge(&h_cg->hugepage[idx], nr_pages, &counter)) nr_pages 214 mm/hugetlb_cgroup.c void hugetlb_cgroup_commit_charge(int idx, unsigned long nr_pages, nr_pages 228 mm/hugetlb_cgroup.c void hugetlb_cgroup_uncharge_page(int idx, unsigned long nr_pages, nr_pages 240 mm/hugetlb_cgroup.c page_counter_uncharge(&h_cg->hugepage[idx], nr_pages); nr_pages 244 mm/hugetlb_cgroup.c void hugetlb_cgroup_uncharge_cgroup(int idx, unsigned long nr_pages, nr_pages 253 mm/hugetlb_cgroup.c page_counter_uncharge(&h_cg->hugepage[idx], nr_pages); nr_pages 292 mm/hugetlb_cgroup.c unsigned long nr_pages; nr_pages 299 mm/hugetlb_cgroup.c ret = page_counter_memparse(buf, "-1", &nr_pages); nr_pages 304 mm/hugetlb_cgroup.c nr_pages = round_down(nr_pages, 1 << huge_page_order(&hstates[idx])); nr_pages 309 mm/hugetlb_cgroup.c ret = page_counter_set_max(&h_cg->hugepage[idx], nr_pages); nr_pages 330 mm/internal.h int nr_pages = hpage_nr_pages(page); nr_pages 333 mm/internal.h __mod_zone_page_state(page_zone(page), NR_MLOCK, -nr_pages); nr_pages 335 mm/internal.h __mod_zone_page_state(page_zone(newpage), NR_MLOCK, nr_pages); nr_pages 680 mm/kasan/common.c nr_shadow_pages = mem_data->nr_pages >> KASAN_SHADOW_SCALE_SHIFT; nr_pages 686 mm/kasan/common.c if (WARN_ON(mem_data->nr_pages % KASAN_SHADOW_SCALE_SIZE) || nr_pages 2822 mm/ksm.c mn->start_pfn + mn->nr_pages); nr_pages 2888 mm/ksm.c unsigned long nr_pages; nr_pages 2890 mm/ksm.c err = kstrtoul(buf, 10, &nr_pages); nr_pages 2891 mm/ksm.c if (err || nr_pages > UINT_MAX) nr_pages 2894 mm/ksm.c ksm_thread_pages_to_scan = nr_pages; nr_pages 588 mm/memcontrol.c unsigned long nr_pages = page_counter_read(&memcg->memory); nr_pages 592 mm/memcontrol.c if (nr_pages > soft_limit) nr_pages 593 mm/memcontrol.c excess = nr_pages - soft_limit; nr_pages 847 mm/memcontrol.c bool compound, int nr_pages) nr_pages 854 mm/memcontrol.c __mod_memcg_state(memcg, MEMCG_RSS, nr_pages); nr_pages 856 mm/memcontrol.c __mod_memcg_state(memcg, MEMCG_CACHE, nr_pages); nr_pages 858 mm/memcontrol.c __mod_memcg_state(memcg, NR_SHMEM, nr_pages); nr_pages 863 mm/memcontrol.c __mod_memcg_state(memcg, MEMCG_RSS_HUGE, nr_pages); nr_pages 867 mm/memcontrol.c if (nr_pages > 0) nr_pages 871 mm/memcontrol.c nr_pages = -nr_pages; /* for event */ nr_pages 874 mm/memcontrol.c __this_cpu_add(memcg->vmstats_percpu->nr_page_events, nr_pages); nr_pages 1291 mm/memcontrol.c int zid, int nr_pages) nr_pages 1303 mm/memcontrol.c if (nr_pages < 0) nr_pages 1304 mm/memcontrol.c *lru_size += nr_pages; nr_pages 1309 mm/memcontrol.c __func__, lruvec, lru, nr_pages, size)) { nr_pages 1314 mm/memcontrol.c if (nr_pages > 0) nr_pages 1315 mm/memcontrol.c *lru_size += nr_pages; nr_pages 2164 mm/memcontrol.c unsigned int nr_pages; nr_pages 2183 mm/memcontrol.c static bool consume_stock(struct mem_cgroup *memcg, unsigned int nr_pages) nr_pages 2189 mm/memcontrol.c if (nr_pages > MEMCG_CHARGE_BATCH) nr_pages 2195 mm/memcontrol.c if (memcg == stock->cached && stock->nr_pages >= nr_pages) { nr_pages 2196 mm/memcontrol.c stock->nr_pages -= nr_pages; nr_pages 2212 mm/memcontrol.c if (stock->nr_pages) { nr_pages 2213 mm/memcontrol.c page_counter_uncharge(&old->memory, stock->nr_pages); nr_pages 2215 mm/memcontrol.c page_counter_uncharge(&old->memsw, stock->nr_pages); nr_pages 2216 mm/memcontrol.c css_put_many(&old->css, stock->nr_pages); nr_pages 2217 mm/memcontrol.c stock->nr_pages = 0; nr_pages 2244 mm/memcontrol.c static void refill_stock(struct mem_cgroup *memcg, unsigned int nr_pages) nr_pages 2256 mm/memcontrol.c stock->nr_pages += nr_pages; nr_pages 2258 mm/memcontrol.c if (stock->nr_pages > MEMCG_CHARGE_BATCH) nr_pages 2289 mm/memcontrol.c if (memcg && stock->nr_pages && nr_pages 2355 mm/memcontrol.c unsigned int nr_pages, nr_pages 2362 mm/memcontrol.c try_to_free_mem_cgroup_pages(memcg, nr_pages, gfp_mask, true); nr_pages 2432 mm/memcontrol.c unsigned int nr_pages) nr_pages 2485 mm/memcontrol.c penalty_jiffies = penalty_jiffies * nr_pages / MEMCG_CHARGE_BATCH; nr_pages 2503 mm/memcontrol.c unsigned int nr_pages = current->memcg_nr_pages_over_high; nr_pages 2506 mm/memcontrol.c if (likely(!nr_pages)) nr_pages 2510 mm/memcontrol.c reclaim_high(memcg, nr_pages, GFP_KERNEL); nr_pages 2517 mm/memcontrol.c penalty_jiffies = calculate_high_delay(memcg, nr_pages); nr_pages 2542 mm/memcontrol.c unsigned int nr_pages) nr_pages 2544 mm/memcontrol.c unsigned int batch = max(MEMCG_CHARGE_BATCH, nr_pages); nr_pages 2556 mm/memcontrol.c if (consume_stock(memcg, nr_pages)) nr_pages 2571 mm/memcontrol.c if (batch > nr_pages) { nr_pages 2572 mm/memcontrol.c batch = nr_pages; nr_pages 2611 mm/memcontrol.c nr_reclaimed = try_to_free_mem_cgroup_pages(mem_over_limit, nr_pages, nr_pages 2614 mm/memcontrol.c if (mem_cgroup_margin(mem_over_limit) >= nr_pages) nr_pages 2634 mm/memcontrol.c if (nr_reclaimed && nr_pages <= (1 << PAGE_ALLOC_COSTLY_ORDER)) nr_pages 2661 mm/memcontrol.c get_order(nr_pages * PAGE_SIZE)); nr_pages 2680 mm/memcontrol.c page_counter_charge(&memcg->memory, nr_pages); nr_pages 2682 mm/memcontrol.c page_counter_charge(&memcg->memsw, nr_pages); nr_pages 2683 mm/memcontrol.c css_get_many(&memcg->css, nr_pages); nr_pages 2689 mm/memcontrol.c if (batch > nr_pages) nr_pages 2690 mm/memcontrol.c refill_stock(memcg, batch - nr_pages); nr_pages 2717 mm/memcontrol.c static void cancel_charge(struct mem_cgroup *memcg, unsigned int nr_pages) nr_pages 2722 mm/memcontrol.c page_counter_uncharge(&memcg->memory, nr_pages); nr_pages 2724 mm/memcontrol.c page_counter_uncharge(&memcg->memsw, nr_pages); nr_pages 2726 mm/memcontrol.c css_put_many(&memcg->css, nr_pages); nr_pages 3016 mm/memcontrol.c unsigned int nr_pages = 1 << order; nr_pages 3020 mm/memcontrol.c ret = try_charge(memcg, gfp, nr_pages); nr_pages 3025 mm/memcontrol.c !page_counter_try_charge(&memcg->kmem, nr_pages, &counter)) { nr_pages 3033 mm/memcontrol.c page_counter_charge(&memcg->kmem, nr_pages); nr_pages 3036 mm/memcontrol.c cancel_charge(memcg, nr_pages); nr_pages 3076 mm/memcontrol.c unsigned int nr_pages) nr_pages 3079 mm/memcontrol.c page_counter_uncharge(&memcg->kmem, nr_pages); nr_pages 3081 mm/memcontrol.c page_counter_uncharge(&memcg->memory, nr_pages); nr_pages 3083 mm/memcontrol.c page_counter_uncharge(&memcg->memsw, nr_pages); nr_pages 3093 mm/memcontrol.c unsigned int nr_pages = 1 << order; nr_pages 3099 mm/memcontrol.c __memcg_kmem_uncharge_memcg(memcg, nr_pages); nr_pages 3106 mm/memcontrol.c css_put_many(&memcg->css, nr_pages); nr_pages 3682 mm/memcontrol.c unsigned long nr_pages; nr_pages 3686 mm/memcontrol.c ret = page_counter_memparse(buf, "-1", &nr_pages); nr_pages 3698 mm/memcontrol.c ret = mem_cgroup_resize_max(memcg, nr_pages, false); nr_pages 3701 mm/memcontrol.c ret = mem_cgroup_resize_max(memcg, nr_pages, true); nr_pages 3707 mm/memcontrol.c ret = memcg_update_kmem_max(memcg, nr_pages); nr_pages 3710 mm/memcontrol.c ret = memcg_update_tcp_max(memcg, nr_pages); nr_pages 3715 mm/memcontrol.c memcg->soft_limit = nr_pages; nr_pages 5491 mm/memcontrol.c unsigned int nr_pages = compound ? hpage_nr_pages(page) : 1; nr_pages 5520 mm/memcontrol.c __mod_lruvec_state(from_vec, NR_FILE_MAPPED, -nr_pages); nr_pages 5521 mm/memcontrol.c __mod_lruvec_state(to_vec, NR_FILE_MAPPED, nr_pages); nr_pages 5533 mm/memcontrol.c __mod_lruvec_state(from_vec, NR_FILE_DIRTY, -nr_pages); nr_pages 5534 mm/memcontrol.c __mod_lruvec_state(to_vec, NR_FILE_DIRTY, nr_pages); nr_pages 5539 mm/memcontrol.c __mod_lruvec_state(from_vec, NR_WRITEBACK, -nr_pages); nr_pages 5540 mm/memcontrol.c __mod_lruvec_state(to_vec, NR_WRITEBACK, nr_pages); nr_pages 5557 mm/memcontrol.c mem_cgroup_charge_statistics(to, page, compound, nr_pages); nr_pages 5559 mm/memcontrol.c mem_cgroup_charge_statistics(from, page, compound, -nr_pages); nr_pages 6137 mm/memcontrol.c unsigned long nr_pages; nr_pages 6148 mm/memcontrol.c nr_pages = page_counter_read(&memcg->memory); nr_pages 6149 mm/memcontrol.c if (nr_pages > high) nr_pages 6150 mm/memcontrol.c try_to_free_mem_cgroup_pages(memcg, nr_pages - high, nr_pages 6180 mm/memcontrol.c unsigned long nr_pages = page_counter_read(&memcg->memory); nr_pages 6182 mm/memcontrol.c if (nr_pages <= max) nr_pages 6197 mm/memcontrol.c if (!try_to_free_mem_cgroup_pages(memcg, nr_pages - max, nr_pages 6518 mm/memcontrol.c unsigned int nr_pages = compound ? hpage_nr_pages(page) : 1; nr_pages 6551 mm/memcontrol.c ret = try_charge(memcg, gfp_mask, nr_pages); nr_pages 6592 mm/memcontrol.c unsigned int nr_pages = compound ? hpage_nr_pages(page) : 1; nr_pages 6610 mm/memcontrol.c mem_cgroup_charge_statistics(memcg, page, compound, nr_pages); nr_pages 6621 mm/memcontrol.c mem_cgroup_uncharge_swap(entry, nr_pages); nr_pages 6636 mm/memcontrol.c unsigned int nr_pages = compound ? hpage_nr_pages(page) : 1; nr_pages 6648 mm/memcontrol.c cancel_charge(memcg, nr_pages); nr_pages 6669 mm/memcontrol.c unsigned long nr_pages = ug->nr_anon + ug->nr_file + ug->nr_kmem; nr_pages 6673 mm/memcontrol.c page_counter_uncharge(&ug->memcg->memory, nr_pages); nr_pages 6675 mm/memcontrol.c page_counter_uncharge(&ug->memcg->memsw, nr_pages); nr_pages 6687 mm/memcontrol.c __this_cpu_add(ug->memcg->vmstats_percpu->nr_page_events, nr_pages); nr_pages 6692 mm/memcontrol.c css_put_many(&ug->memcg->css, nr_pages); nr_pages 6719 mm/memcontrol.c unsigned int nr_pages = 1; nr_pages 6722 mm/memcontrol.c nr_pages = compound_nr(page); nr_pages 6723 mm/memcontrol.c ug->nr_huge += nr_pages; nr_pages 6726 mm/memcontrol.c ug->nr_anon += nr_pages; nr_pages 6728 mm/memcontrol.c ug->nr_file += nr_pages; nr_pages 6730 mm/memcontrol.c ug->nr_shmem += nr_pages; nr_pages 6819 mm/memcontrol.c unsigned int nr_pages; nr_pages 6843 mm/memcontrol.c nr_pages = compound ? hpage_nr_pages(newpage) : 1; nr_pages 6845 mm/memcontrol.c page_counter_charge(&memcg->memory, nr_pages); nr_pages 6847 mm/memcontrol.c page_counter_charge(&memcg->memsw, nr_pages); nr_pages 6848 mm/memcontrol.c css_get_many(&memcg->css, nr_pages); nr_pages 6853 mm/memcontrol.c mem_cgroup_charge_statistics(memcg, newpage, compound, nr_pages); nr_pages 6898 mm/memcontrol.c bool mem_cgroup_charge_skmem(struct mem_cgroup *memcg, unsigned int nr_pages) nr_pages 6905 mm/memcontrol.c if (page_counter_try_charge(&memcg->tcpmem, nr_pages, &fail)) { nr_pages 6909 mm/memcontrol.c page_counter_charge(&memcg->tcpmem, nr_pages); nr_pages 6918 mm/memcontrol.c mod_memcg_state(memcg, MEMCG_SOCK, nr_pages); nr_pages 6920 mm/memcontrol.c if (try_charge(memcg, gfp_mask, nr_pages) == 0) nr_pages 6923 mm/memcontrol.c try_charge(memcg, gfp_mask|__GFP_NOFAIL, nr_pages); nr_pages 6932 mm/memcontrol.c void mem_cgroup_uncharge_skmem(struct mem_cgroup *memcg, unsigned int nr_pages) nr_pages 6935 mm/memcontrol.c page_counter_uncharge(&memcg->tcpmem, nr_pages); nr_pages 6939 mm/memcontrol.c mod_memcg_state(memcg, MEMCG_SOCK, -nr_pages); nr_pages 6941 mm/memcontrol.c refill_stock(memcg, nr_pages); nr_pages 7102 mm/memcontrol.c unsigned int nr_pages = hpage_nr_pages(page); nr_pages 7124 mm/memcontrol.c !page_counter_try_charge(&memcg->swap, nr_pages, &counter)) { nr_pages 7132 mm/memcontrol.c if (nr_pages > 1) nr_pages 7133 mm/memcontrol.c mem_cgroup_id_get_many(memcg, nr_pages - 1); nr_pages 7134 mm/memcontrol.c oldid = swap_cgroup_record(entry, mem_cgroup_id(memcg), nr_pages); nr_pages 7136 mm/memcontrol.c mod_memcg_state(memcg, MEMCG_SWAP, nr_pages); nr_pages 7146 mm/memcontrol.c void mem_cgroup_uncharge_swap(swp_entry_t entry, unsigned int nr_pages) nr_pages 7154 mm/memcontrol.c id = swap_cgroup_record(entry, 0, nr_pages); nr_pages 7160 mm/memcontrol.c page_counter_uncharge(&memcg->swap, nr_pages); nr_pages 7162 mm/memcontrol.c page_counter_uncharge(&memcg->memsw, nr_pages); nr_pages 7164 mm/memcontrol.c mod_memcg_state(memcg, MEMCG_SWAP, -nr_pages); nr_pages 7165 mm/memcontrol.c mem_cgroup_id_put_many(memcg, nr_pages); nr_pages 3436 mm/memory.c unsigned long address = vmf->address, nr_pages, mask; nr_pages 3442 mm/memory.c nr_pages = READ_ONCE(fault_around_bytes) >> PAGE_SHIFT; nr_pages 3443 mm/memory.c mask = ~(nr_pages * PAGE_SIZE - 1) & PAGE_MASK; nr_pages 3457 mm/memory.c start_pgoff + nr_pages - 1); nr_pages 227 mm/memory_hotplug.c unsigned long i, pfn, end_pfn, nr_pages; nr_pages 231 mm/memory_hotplug.c nr_pages = PAGE_ALIGN(sizeof(struct pglist_data)) >> PAGE_SHIFT; nr_pages 234 mm/memory_hotplug.c for (i = 0; i < nr_pages; i++, page++) nr_pages 254 mm/memory_hotplug.c static int check_pfn_span(unsigned long pfn, unsigned long nr_pages, nr_pages 273 mm/memory_hotplug.c || !IS_ALIGNED(nr_pages, min_align)) { nr_pages 275 mm/memory_hotplug.c reason, pfn, pfn + nr_pages - 1); nr_pages 287 mm/memory_hotplug.c int __ref __add_pages(int nid, unsigned long pfn, unsigned long nr_pages, nr_pages 299 mm/memory_hotplug.c || vmem_altmap_offset(altmap) > nr_pages) { nr_pages 306 mm/memory_hotplug.c err = check_pfn_span(pfn, nr_pages, "add"); nr_pages 311 mm/memory_hotplug.c end_sec = pfn_to_section_nr(pfn + nr_pages - 1); nr_pages 315 mm/memory_hotplug.c pfns = min(nr_pages, PAGES_PER_SECTION nr_pages 321 mm/memory_hotplug.c nr_pages -= pfns; nr_pages 470 mm/memory_hotplug.c unsigned long nr_pages) nr_pages 488 mm/memory_hotplug.c shrink_zone_span(zone, start_pfn, start_pfn + nr_pages); nr_pages 495 mm/memory_hotplug.c static void __remove_section(unsigned long pfn, unsigned long nr_pages, nr_pages 504 mm/memory_hotplug.c sparse_remove_section(ms, pfn, nr_pages, map_offset, altmap); nr_pages 518 mm/memory_hotplug.c void __remove_pages(unsigned long pfn, unsigned long nr_pages, nr_pages 526 mm/memory_hotplug.c if (check_pfn_span(pfn, nr_pages, "remove")) nr_pages 530 mm/memory_hotplug.c end_sec = pfn_to_section_nr(pfn + nr_pages - 1); nr_pages 535 mm/memory_hotplug.c pfns = min(nr_pages, PAGES_PER_SECTION nr_pages 539 mm/memory_hotplug.c nr_pages -= pfns; nr_pages 616 mm/memory_hotplug.c static int online_pages_range(unsigned long start_pfn, unsigned long nr_pages, nr_pages 619 mm/memory_hotplug.c const unsigned long end_pfn = start_pfn + nr_pages; nr_pages 639 mm/memory_hotplug.c *(unsigned long *)arg += nr_pages; nr_pages 644 mm/memory_hotplug.c static void node_states_check_changes_online(unsigned long nr_pages, nr_pages 676 mm/memory_hotplug.c unsigned long nr_pages) nr_pages 683 mm/memory_hotplug.c zone->spanned_pages = max(start_pfn + nr_pages, old_end_pfn) - zone->zone_start_pfn; nr_pages 687 mm/memory_hotplug.c unsigned long nr_pages) nr_pages 694 mm/memory_hotplug.c pgdat->node_spanned_pages = max(start_pfn + nr_pages, old_end_pfn) - pgdat->node_start_pfn; nr_pages 703 mm/memory_hotplug.c unsigned long nr_pages, struct vmem_altmap *altmap) nr_pages 715 mm/memory_hotplug.c init_currently_empty_zone(zone, start_pfn, nr_pages); nr_pages 716 mm/memory_hotplug.c resize_zone_range(zone, start_pfn, nr_pages); nr_pages 718 mm/memory_hotplug.c resize_pgdat_range(pgdat, start_pfn, nr_pages); nr_pages 727 mm/memory_hotplug.c memmap_init_zone(nr_pages, nid, zone_idx(zone), start_pfn, nr_pages 739 mm/memory_hotplug.c unsigned long nr_pages) nr_pages 747 mm/memory_hotplug.c if (zone_intersects(zone, start_pfn, nr_pages)) nr_pages 755 mm/memory_hotplug.c unsigned long nr_pages) nr_pages 758 mm/memory_hotplug.c nr_pages); nr_pages 760 mm/memory_hotplug.c bool in_kernel = zone_intersects(kernel_zone, start_pfn, nr_pages); nr_pages 761 mm/memory_hotplug.c bool in_movable = zone_intersects(movable_zone, start_pfn, nr_pages); nr_pages 779 mm/memory_hotplug.c unsigned long nr_pages) nr_pages 782 mm/memory_hotplug.c return default_kernel_zone_for_pfn(nid, start_pfn, nr_pages); nr_pages 787 mm/memory_hotplug.c return default_zone_for_pfn(nid, start_pfn, nr_pages); nr_pages 790 mm/memory_hotplug.c int __ref online_pages(unsigned long pfn, unsigned long nr_pages, int online_type) nr_pages 812 mm/memory_hotplug.c zone = zone_for_pfn_range(online_type, nid, pfn, nr_pages); nr_pages 813 mm/memory_hotplug.c move_pfn_range_to_zone(zone, pfn, nr_pages, NULL); nr_pages 816 mm/memory_hotplug.c arg.nr_pages = nr_pages; nr_pages 817 mm/memory_hotplug.c node_states_check_changes_online(nr_pages, zone, &arg); nr_pages 834 mm/memory_hotplug.c ret = walk_system_ram_range(pfn, nr_pages, &onlined_pages, nr_pages 873 mm/memory_hotplug.c (((unsigned long long) pfn + nr_pages) << PAGE_SHIFT) - 1); nr_pages 875 mm/memory_hotplug.c remove_pfn_range_from_zone(zone, pfn, nr_pages); nr_pages 1193 mm/memory_hotplug.c bool is_mem_section_removable(unsigned long start_pfn, unsigned long nr_pages) nr_pages 1197 mm/memory_hotplug.c end_pfn = min(start_pfn + nr_pages, nr_pages 1390 mm/memory_hotplug.c offline_isolated_pages_cb(unsigned long start, unsigned long nr_pages, nr_pages 1395 mm/memory_hotplug.c *offlined_pages += __offline_isolated_pages(start, start + nr_pages); nr_pages 1403 mm/memory_hotplug.c check_pages_isolated_cb(unsigned long start_pfn, unsigned long nr_pages, nr_pages 1406 mm/memory_hotplug.c return test_pages_isolated(start_pfn, start_pfn + nr_pages, true); nr_pages 1421 mm/memory_hotplug.c static void node_states_check_changes_offline(unsigned long nr_pages, nr_pages 1442 mm/memory_hotplug.c if (zone_idx(zone) <= ZONE_NORMAL && nr_pages >= present_pages) nr_pages 1455 mm/memory_hotplug.c if (zone_idx(zone) <= ZONE_HIGHMEM && nr_pages >= present_pages) nr_pages 1471 mm/memory_hotplug.c if (nr_pages >= present_pages) nr_pages 1490 mm/memory_hotplug.c unsigned long pfn, nr_pages; nr_pages 1512 mm/memory_hotplug.c nr_pages = end_pfn - start_pfn; nr_pages 1525 mm/memory_hotplug.c arg.nr_pages = nr_pages; nr_pages 1526 mm/memory_hotplug.c node_states_check_changes_offline(nr_pages, zone, &arg); nr_pages 1611 mm/memory_hotplug.c remove_pfn_range_from_zone(zone, start_pfn, nr_pages); nr_pages 1628 mm/memory_hotplug.c int offline_pages(unsigned long start_pfn, unsigned long nr_pages) nr_pages 1630 mm/memory_hotplug.c return __offline_pages(start_pfn, start_pfn + nr_pages); nr_pages 549 mm/migrate.c int nr_pages) nr_pages 555 mm/migrate.c for (i = 0; i < nr_pages; ) { nr_pages 568 mm/migrate.c int nr_pages; nr_pages 573 mm/migrate.c nr_pages = pages_per_huge_page(h); nr_pages 575 mm/migrate.c if (unlikely(nr_pages > MAX_ORDER_NR_PAGES)) { nr_pages 576 mm/migrate.c __copy_gigantic_page(dst, src, nr_pages); nr_pages 582 mm/migrate.c nr_pages = hpage_nr_pages(src); nr_pages 585 mm/migrate.c for (i = 0; i < nr_pages; i++) { nr_pages 1595 mm/migrate.c unsigned long nr_pages, nr_pages 1607 mm/migrate.c for (i = start = 0; i < nr_pages; i++) { nr_pages 1644 mm/migrate.c err += nr_pages - i - 1; nr_pages 1679 mm/migrate.c err += nr_pages - i - 1; nr_pages 1713 mm/migrate.c static void do_pages_stat_array(struct mm_struct *mm, unsigned long nr_pages, nr_pages 1720 mm/migrate.c for (i = 0; i < nr_pages; i++) { nr_pages 1752 mm/migrate.c static int do_pages_stat(struct mm_struct *mm, unsigned long nr_pages, nr_pages 1760 mm/migrate.c while (nr_pages) { nr_pages 1763 mm/migrate.c chunk_nr = nr_pages; nr_pages 1777 mm/migrate.c nr_pages -= chunk_nr; nr_pages 1779 mm/migrate.c return nr_pages ? -EFAULT : 0; nr_pages 1786 mm/migrate.c static int kernel_move_pages(pid_t pid, unsigned long nr_pages, nr_pages 1835 mm/migrate.c err = do_pages_move(mm, task_nodes, nr_pages, pages, nr_pages 1838 mm/migrate.c err = do_pages_stat(mm, nr_pages, pages, status); nr_pages 1848 mm/migrate.c SYSCALL_DEFINE6(move_pages, pid_t, pid, unsigned long, nr_pages, nr_pages 1853 mm/migrate.c return kernel_move_pages(pid, nr_pages, pages, nodes, status, flags); nr_pages 1857 mm/migrate.c COMPAT_SYSCALL_DEFINE6(move_pages, pid_t, pid, compat_ulong_t, nr_pages, nr_pages 1866 mm/migrate.c pages = compat_alloc_user_space(nr_pages * sizeof(void *)); nr_pages 1867 mm/migrate.c for (i = 0; i < nr_pages; i++) { nr_pages 1874 mm/migrate.c return kernel_move_pages(pid, nr_pages, pages, nodes, status, flags); nr_pages 2665 mm/migrate.c long nr_pages = (args->end - args->start) >> PAGE_SHIFT; nr_pages 2672 mm/migrate.c if (nr_pages <= 0) nr_pages 2682 mm/migrate.c memset(args->src, 0, sizeof(*args->src) * nr_pages); nr_pages 184 mm/mlock.c int nr_pages; nr_pages 201 mm/mlock.c nr_pages = 1; nr_pages 205 mm/mlock.c nr_pages = hpage_nr_pages(page); nr_pages 206 mm/mlock.c __mod_zone_page_state(page_zone(page), NR_MLOCK, -nr_pages); nr_pages 219 mm/mlock.c return nr_pages - 1; nr_pages 524 mm/mlock.c int nr_pages; nr_pages 560 mm/mlock.c nr_pages = (end - start) >> PAGE_SHIFT; nr_pages 562 mm/mlock.c nr_pages = -nr_pages; nr_pages 564 mm/mlock.c nr_pages = 0; nr_pages 565 mm/mlock.c mm->locked_vm += nr_pages; nr_pages 563 mm/mmap.c unsigned long nr_pages = 0; nr_pages 571 mm/mmap.c nr_pages = (min(end, vma->vm_end) - nr_pages 582 mm/mmap.c nr_pages += overlap_len >> PAGE_SHIFT; nr_pages 585 mm/mmap.c return nr_pages; nr_pages 1723 mm/mmap.c unsigned long nr_pages; nr_pages 1729 mm/mmap.c nr_pages = count_vma_pages_range(mm, addr, addr + len); nr_pages 1732 mm/mmap.c (len >> PAGE_SHIFT) - nr_pages)) nr_pages 279 mm/page-writeback.c unsigned long nr_pages = 0; nr_pages 288 mm/page-writeback.c nr_pages += zone_page_state(zone, NR_FREE_PAGES); nr_pages 296 mm/page-writeback.c nr_pages -= min(nr_pages, pgdat->totalreserve_pages); nr_pages 298 mm/page-writeback.c nr_pages += node_page_state(pgdat, NR_INACTIVE_FILE); nr_pages 299 mm/page-writeback.c nr_pages += node_page_state(pgdat, NR_ACTIVE_FILE); nr_pages 301 mm/page-writeback.c return nr_pages; nr_pages 314 mm/page-writeback.c unsigned long nr_pages; nr_pages 323 mm/page-writeback.c nr_pages = zone_page_state(z, NR_FREE_PAGES); nr_pages 325 mm/page-writeback.c nr_pages -= min(nr_pages, high_wmark_pages(z)); nr_pages 326 mm/page-writeback.c nr_pages += zone_page_state(z, NR_ZONE_INACTIVE_FILE); nr_pages 327 mm/page-writeback.c nr_pages += zone_page_state(z, NR_ZONE_ACTIVE_FILE); nr_pages 328 mm/page-writeback.c x += nr_pages; nr_pages 505 mm/page-writeback.c unsigned long nr_pages = 0; nr_pages 507 mm/page-writeback.c nr_pages += node_page_state(pgdat, NR_FILE_DIRTY); nr_pages 508 mm/page-writeback.c nr_pages += node_page_state(pgdat, NR_UNSTABLE_NFS); nr_pages 509 mm/page-writeback.c nr_pages += node_page_state(pgdat, NR_WRITEBACK); nr_pages 511 mm/page-writeback.c return nr_pages <= limit; nr_pages 2166 mm/page-writeback.c int nr_pages; nr_pages 2195 mm/page-writeback.c nr_pages = pagevec_lookup_range_tag(&pvec, mapping, &index, end, nr_pages 2197 mm/page-writeback.c if (nr_pages == 0) nr_pages 2200 mm/page-writeback.c for (i = 0; i < nr_pages; i++) { nr_pages 680 mm/page_alloc.c int nr_pages = 1 << order; nr_pages 685 mm/page_alloc.c for (i = 1; i < nr_pages; i++) { nr_pages 1432 mm/page_alloc.c unsigned int nr_pages = 1 << order; nr_pages 1437 mm/page_alloc.c for (loop = 0; loop < (nr_pages - 1); loop++, p++) { nr_pages 1445 mm/page_alloc.c atomic_long_add(nr_pages, &page_zone(page)->managed_pages); nr_pages 1572 mm/page_alloc.c unsigned long nr_pages) nr_pages 1577 mm/page_alloc.c if (!nr_pages) nr_pages 1583 mm/page_alloc.c if (nr_pages == pageblock_nr_pages && nr_pages 1590 mm/page_alloc.c for (i = 0; i < nr_pages; i++, page++, pfn++) { nr_pages 1663 mm/page_alloc.c unsigned long nr_pages = 0; nr_pages 1678 mm/page_alloc.c nr_pages++; nr_pages 1680 mm/page_alloc.c return (nr_pages); nr_pages 1729 mm/page_alloc.c unsigned long nr_pages = 0; nr_pages 1740 mm/page_alloc.c nr_pages += deferred_init_pages(zone, *start_pfn, t); nr_pages 1764 mm/page_alloc.c return nr_pages; nr_pages 1772 mm/page_alloc.c unsigned long spfn = 0, epfn = 0, nr_pages = 0; nr_pages 1814 mm/page_alloc.c nr_pages += deferred_init_maxorder(&i, zone, &spfn, &epfn); nr_pages 1822 mm/page_alloc.c pgdat->node_id, nr_pages, jiffies_to_msecs(jiffies - start)); nr_pages 1850 mm/page_alloc.c unsigned long nr_pages = 0; nr_pages 1897 mm/page_alloc.c nr_pages += deferred_init_maxorder(&i, zone, &spfn, &epfn); nr_pages 1904 mm/page_alloc.c if (nr_pages >= nr_pages_needed) nr_pages 1911 mm/page_alloc.c return nr_pages > 0; nr_pages 8515 mm/page_alloc.c void free_contig_range(unsigned long pfn, unsigned int nr_pages) nr_pages 8519 mm/page_alloc.c for (; nr_pages--; pfn++) { nr_pages 55 mm/page_counter.c void page_counter_cancel(struct page_counter *counter, unsigned long nr_pages) nr_pages 59 mm/page_counter.c new = atomic_long_sub_return(nr_pages, &counter->usage); nr_pages 72 mm/page_counter.c void page_counter_charge(struct page_counter *counter, unsigned long nr_pages) nr_pages 79 mm/page_counter.c new = atomic_long_add_return(nr_pages, &c->usage); nr_pages 100 mm/page_counter.c unsigned long nr_pages, nr_pages 121 mm/page_counter.c new = atomic_long_add_return(nr_pages, &c->usage); nr_pages 123 mm/page_counter.c atomic_long_sub(nr_pages, &c->usage); nr_pages 145 mm/page_counter.c page_counter_cancel(c, nr_pages); nr_pages 155 mm/page_counter.c void page_counter_uncharge(struct page_counter *counter, unsigned long nr_pages) nr_pages 160 mm/page_counter.c page_counter_cancel(c, nr_pages); nr_pages 173 mm/page_counter.c int page_counter_set_max(struct page_counter *counter, unsigned long nr_pages) nr_pages 192 mm/page_counter.c if (usage > nr_pages) nr_pages 195 mm/page_counter.c old = xchg(&counter->max, nr_pages); nr_pages 212 mm/page_counter.c void page_counter_set_min(struct page_counter *counter, unsigned long nr_pages) nr_pages 216 mm/page_counter.c counter->min = nr_pages; nr_pages 229 mm/page_counter.c void page_counter_set_low(struct page_counter *counter, unsigned long nr_pages) nr_pages 233 mm/page_counter.c counter->low = nr_pages; nr_pages 249 mm/page_counter.c unsigned long *nr_pages) nr_pages 255 mm/page_counter.c *nr_pages = PAGE_COUNTER_MAX; nr_pages 263 mm/page_counter.c *nr_pages = min(bytes / PAGE_SIZE, (u64)PAGE_COUNTER_MAX); nr_pages 139 mm/page_ext.c unsigned long nr_pages; nr_pages 141 mm/page_ext.c nr_pages = NODE_DATA(nid)->node_spanned_pages; nr_pages 142 mm/page_ext.c if (!nr_pages) nr_pages 152 mm/page_ext.c nr_pages += MAX_ORDER_NR_PAGES; nr_pages 154 mm/page_ext.c table_size = page_ext_size * nr_pages; nr_pages 287 mm/page_ext.c unsigned long nr_pages, nr_pages 294 mm/page_ext.c end = SECTION_ALIGN_UP(start_pfn + nr_pages); nr_pages 322 mm/page_ext.c unsigned long nr_pages, int nid) nr_pages 327 mm/page_ext.c end = SECTION_ALIGN_UP(start_pfn + nr_pages); nr_pages 344 mm/page_ext.c mn->nr_pages, mn->status_change_nid); nr_pages 348 mm/page_ext.c mn->nr_pages, mn->status_change_nid); nr_pages 352 mm/page_ext.c mn->nr_pages, mn->status_change_nid); nr_pages 40 mm/page_isolation.c arg.nr_pages = pageblock_nr_pages; nr_pages 73 mm/page_isolation.c unsigned long nr_pages; nr_pages 78 mm/page_isolation.c nr_pages = move_freepages_block(zone, page, MIGRATE_ISOLATE, nr_pages 81 mm/page_isolation.c __mod_zone_freepage_state(zone, -nr_pages, mt); nr_pages 93 mm/page_isolation.c unsigned long flags, nr_pages; nr_pages 133 mm/page_isolation.c nr_pages = move_freepages_block(zone, page, migratetype, NULL); nr_pages 134 mm/page_isolation.c __mod_zone_freepage_state(zone, nr_pages, migratetype); nr_pages 147 mm/page_isolation.c __first_valid_page(unsigned long pfn, unsigned long nr_pages) nr_pages 151 mm/page_isolation.c for (i = 0; i < nr_pages; i++) { nr_pages 58 mm/percpu-internal.h int nr_pages; /* # of pages served by this chunk */ nr_pages 82 mm/percpu-internal.h return chunk->nr_pages * PAGE_SIZE / PCPU_BITMAP_BLOCK_SIZE; nr_pages 106 mm/percpu-internal.h return pcpu_nr_pages_to_map_bits(chunk->nr_pages); nr_pages 49 mm/percpu-km.c const int nr_pages = pcpu_group_sizes[0] >> PAGE_SHIFT; nr_pages 59 mm/percpu-km.c pages = alloc_pages(gfp, order_base_2(nr_pages)); nr_pages 65 mm/percpu-km.c for (i = 0; i < nr_pages; i++) nr_pages 72 mm/percpu-km.c pcpu_chunk_populated(chunk, 0, nr_pages); nr_pages 83 mm/percpu-km.c const int nr_pages = pcpu_group_sizes[0] >> PAGE_SHIFT; nr_pages 92 mm/percpu-km.c __free_pages(chunk->data, order_base_2(nr_pages)); nr_pages 103 mm/percpu-km.c size_t nr_pages, alloc_pages; nr_pages 111 mm/percpu-km.c nr_pages = (ai->groups[0].nr_units * ai->unit_size) >> PAGE_SHIFT; nr_pages 112 mm/percpu-km.c alloc_pages = roundup_pow_of_two(nr_pages); nr_pages 114 mm/percpu-km.c if (alloc_pages > nr_pages) nr_pages 116 mm/percpu-km.c alloc_pages - nr_pages); nr_pages 134 mm/percpu-vm.c static void __pcpu_unmap_pages(unsigned long addr, int nr_pages) nr_pages 136 mm/percpu-vm.c unmap_kernel_range_noflush(addr, nr_pages << PAGE_SHIFT); nr_pages 193 mm/percpu-vm.c int nr_pages) nr_pages 195 mm/percpu-vm.c return map_kernel_range_noflush(addr, nr_pages << PAGE_SHIFT, nr_pages 214 mm/percpu.c end_addr = chunk->base_addr + chunk->nr_pages * PAGE_SIZE - nr_pages 1343 mm/percpu.c chunk->nr_pages = region_size >> PAGE_SHIFT; nr_pages 1369 mm/percpu.c bitmap_fill(chunk->populated, chunk->nr_pages); nr_pages 1370 mm/percpu.c chunk->nr_populated = chunk->nr_pages; nr_pages 1371 mm/percpu.c chunk->nr_empty_pop_pages = chunk->nr_pages; nr_pages 1414 mm/percpu.c chunk->nr_pages = pcpu_unit_pages; nr_pages 1435 mm/percpu.c chunk->free_bytes = chunk->nr_pages * PAGE_SIZE; nr_pages 1864 mm/percpu.c chunk->nr_pages) { nr_pages 1903 mm/percpu.c nr_unpop = chunk->nr_pages - chunk->nr_populated; nr_pages 1914 mm/percpu.c chunk->nr_pages) { nr_pages 84 mm/process_vm_access.c unsigned long nr_pages; nr_pages 93 mm/process_vm_access.c nr_pages = (addr + len - 1) / PAGE_SIZE - addr / PAGE_SIZE + 1; nr_pages 98 mm/process_vm_access.c while (!rc && nr_pages && iov_iter_count(iter)) { nr_pages 99 mm/process_vm_access.c int pages = min(nr_pages, max_pages_per_loop); nr_pages 125 mm/process_vm_access.c nr_pages -= pages; nr_pages 162 mm/process_vm_access.c unsigned long nr_pages = 0; nr_pages 178 mm/process_vm_access.c nr_pages = max(nr_pages, nr_pages_iov); nr_pages 182 mm/process_vm_access.c if (nr_pages == 0) nr_pages 185 mm/process_vm_access.c if (nr_pages > PVM_MAX_PP_ARRAY_COUNT) { nr_pages 189 mm/process_vm_access.c sizeof(struct pages *)*nr_pages), nr_pages 117 mm/readahead.c struct list_head *pages, unsigned int nr_pages, gfp_t gfp) nr_pages 126 mm/readahead.c ret = mapping->a_ops->readpages(filp, mapping, pages, nr_pages); nr_pages 132 mm/readahead.c for (page_idx = 0; page_idx < nr_pages; page_idx++) { nr_pages 164 mm/readahead.c unsigned int nr_pages = 0; nr_pages 189 mm/readahead.c if (nr_pages) nr_pages 190 mm/readahead.c read_pages(mapping, filp, &page_pool, nr_pages, nr_pages 192 mm/readahead.c nr_pages = 0; nr_pages 203 mm/readahead.c nr_pages++; nr_pages 211 mm/readahead.c if (nr_pages) nr_pages 212 mm/readahead.c read_pages(mapping, filp, &page_pool, nr_pages, gfp_mask); nr_pages 215 mm/readahead.c return nr_pages; nr_pages 133 mm/shmem.c unsigned long nr_pages = totalram_pages(); nr_pages 135 mm/shmem.c return min(nr_pages - totalhigh_pages(), nr_pages / 2); nr_pages 249 mm/sparse-vmemmap.c unsigned long nr_pages, int nid, struct vmem_altmap *altmap) nr_pages 259 mm/sparse-vmemmap.c end = ALIGN(pfn + nr_pages, PAGES_PER_SUBSECTION); nr_pages 261 mm/sparse-vmemmap.c nr_pages = end - pfn; nr_pages 264 mm/sparse-vmemmap.c end = start + nr_pages * sizeof(struct page); nr_pages 223 mm/sparse.c unsigned long nr_pages) nr_pages 226 mm/sparse.c int end = subsection_map_index(pfn + nr_pages - 1); nr_pages 231 mm/sparse.c void __init subsection_map_init(unsigned long pfn, unsigned long nr_pages) nr_pages 233 mm/sparse.c int end_sec = pfn_to_section_nr(pfn + nr_pages - 1); nr_pages 236 mm/sparse.c if (!nr_pages) nr_pages 243 mm/sparse.c pfns = min(nr_pages, PAGES_PER_SECTION nr_pages 253 mm/sparse.c nr_pages -= pfns; nr_pages 452 mm/sparse.c unsigned long nr_pages, int nid, struct vmem_altmap *altmap) nr_pages 651 mm/sparse.c unsigned long nr_pages, int nid, struct vmem_altmap *altmap) nr_pages 653 mm/sparse.c return __populate_section_memmap(pfn, nr_pages, nid, altmap); nr_pages 656 mm/sparse.c static void depopulate_section_memmap(unsigned long pfn, unsigned long nr_pages, nr_pages 660 mm/sparse.c unsigned long end = start + nr_pages * sizeof(struct page); nr_pages 673 mm/sparse.c unsigned long nr_pages, int nid, struct vmem_altmap *altmap) nr_pages 694 mm/sparse.c static void depopulate_section_memmap(unsigned long pfn, unsigned long nr_pages, nr_pages 709 mm/sparse.c unsigned long magic, nr_pages; nr_pages 712 mm/sparse.c nr_pages = PAGE_ALIGN(PAGES_PER_SECTION * sizeof(struct page)) nr_pages 715 mm/sparse.c for (i = 0; i < nr_pages; i++, page++) { nr_pages 737 mm/sparse.c static void section_deactivate(unsigned long pfn, unsigned long nr_pages, nr_pages 749 mm/sparse.c subsection_mask_set(map, pfn, nr_pages); nr_pages 755 mm/sparse.c pfn, nr_pages)) nr_pages 803 mm/sparse.c depopulate_section_memmap(pfn, nr_pages, altmap); nr_pages 810 mm/sparse.c unsigned long nr_pages, struct vmem_altmap *altmap) nr_pages 819 mm/sparse.c subsection_mask_set(map, pfn, nr_pages); nr_pages 851 mm/sparse.c if (nr_pages < PAGES_PER_SECTION && early_section(ms)) nr_pages 854 mm/sparse.c memmap = populate_section_memmap(pfn, nr_pages, nid, altmap); nr_pages 856 mm/sparse.c section_deactivate(pfn, nr_pages, altmap); nr_pages 878 mm/sparse.c unsigned long nr_pages, struct vmem_altmap *altmap) nr_pages 889 mm/sparse.c memmap = section_activate(nid, start_pfn, nr_pages, altmap); nr_pages 897 mm/sparse.c page_init_poison(memmap, sizeof(struct page) * nr_pages); nr_pages 912 mm/sparse.c static void clear_hwpoisoned_pages(struct page *memmap, int nr_pages) nr_pages 925 mm/sparse.c for (i = 0; i < nr_pages; i++) { nr_pages 933 mm/sparse.c static inline void clear_hwpoisoned_pages(struct page *memmap, int nr_pages) nr_pages 939 mm/sparse.c unsigned long nr_pages, unsigned long map_offset, nr_pages 943 mm/sparse.c nr_pages - map_offset); nr_pages 944 mm/sparse.c section_deactivate(pfn, nr_pages, altmap); nr_pages 597 mm/swap_state.c int init_swap_address_space(unsigned int type, unsigned long nr_pages) nr_pages 602 mm/swap_state.c nr = DIV_ROUND_UP(nr_pages, SWAP_ADDRESS_SPACE_PAGES); nr_pages 181 mm/swapfile.c nr_blocks = ((sector_t)se->nr_pages - 1) << (PAGE_SHIFT - 9); nr_pages 192 mm/swapfile.c nr_blocks = (sector_t)se->nr_pages << (PAGE_SHIFT - 9); nr_pages 215 mm/swapfile.c else if (offset >= se->start_page + se->nr_pages) nr_pages 229 mm/swapfile.c pgoff_t start_page, pgoff_t nr_pages) nr_pages 233 mm/swapfile.c while (nr_pages) { nr_pages 236 mm/swapfile.c sector_t nr_blocks = se->nr_pages - offset; nr_pages 238 mm/swapfile.c if (nr_blocks > nr_pages) nr_pages 239 mm/swapfile.c nr_blocks = nr_pages; nr_pages 241 mm/swapfile.c nr_pages -= nr_blocks; nr_pages 2312 mm/swapfile.c unsigned long nr_pages, sector_t start_block) nr_pages 2329 mm/swapfile.c BUG_ON(se->start_page + se->nr_pages != start_page); nr_pages 2330 mm/swapfile.c if (se->start_block + se->nr_pages == start_block) { nr_pages 2332 mm/swapfile.c se->nr_pages += nr_pages; nr_pages 2342 mm/swapfile.c new_se->nr_pages = nr_pages; nr_pages 2176 mm/vmalloc.c for (i = 0; i < area->nr_pages; i++) nr_pages 2209 mm/vmalloc.c for (i = 0; i < area->nr_pages; i++) { nr_pages 2254 mm/vmalloc.c for (i = 0; i < area->nr_pages; i++) { nr_pages 2260 mm/vmalloc.c atomic_long_sub(area->nr_pages, &nr_vmalloc_pages); nr_pages 2402 mm/vmalloc.c unsigned int nr_pages, array_size, i; nr_pages 2409 mm/vmalloc.c nr_pages = get_vm_area_size(area) >> PAGE_SHIFT; nr_pages 2410 mm/vmalloc.c array_size = (nr_pages * sizeof(struct page *)); nr_pages 2427 mm/vmalloc.c area->nr_pages = nr_pages; nr_pages 2429 mm/vmalloc.c for (i = 0; i < area->nr_pages; i++) { nr_pages 2439 mm/vmalloc.c area->nr_pages = i; nr_pages 2440 mm/vmalloc.c atomic_long_add(area->nr_pages, &nr_vmalloc_pages); nr_pages 2447 mm/vmalloc.c atomic_long_add(area->nr_pages, &nr_vmalloc_pages); nr_pages 2456 mm/vmalloc.c (area->nr_pages*PAGE_SIZE), area->size); nr_pages 3465 mm/vmalloc.c for (nr = 0; nr < v->nr_pages; nr++) nr_pages 3517 mm/vmalloc.c if (v->nr_pages) nr_pages 3518 mm/vmalloc.c seq_printf(m, " pages=%d", v->nr_pages); nr_pages 1140 mm/vmscan.c unsigned int nr_pages; nr_pages 1152 mm/vmscan.c nr_pages = compound_nr(page); nr_pages 1155 mm/vmscan.c sc->nr_scanned += nr_pages; nr_pages 1276 mm/vmscan.c stat->nr_ref_keep += nr_pages; nr_pages 1338 mm/vmscan.c if ((nr_pages > 1) && !PageTransHuge(page)) { nr_pages 1339 mm/vmscan.c sc->nr_scanned -= (nr_pages - 1); nr_pages 1340 mm/vmscan.c nr_pages = 1; nr_pages 1353 mm/vmscan.c stat->nr_unmap_fail += nr_pages; nr_pages 1484 mm/vmscan.c nr_reclaimed += nr_pages; nr_pages 1501 mm/vmscan.c if (nr_pages > 1) { nr_pages 1502 mm/vmscan.c sc->nr_scanned -= (nr_pages - 1); nr_pages 1503 mm/vmscan.c nr_pages = 1; nr_pages 1514 mm/vmscan.c stat->nr_activate[type] += nr_pages; nr_pages 1693 mm/vmscan.c unsigned long scan, total_scan, nr_pages; nr_pages 1707 mm/vmscan.c nr_pages = compound_nr(page); nr_pages 1708 mm/vmscan.c total_scan += nr_pages; nr_pages 1712 mm/vmscan.c nr_skipped[page_zonenum(page)] += nr_pages; nr_pages 1726 mm/vmscan.c scan += nr_pages; nr_pages 1729 mm/vmscan.c nr_taken += nr_pages; nr_pages 1730 mm/vmscan.c nr_zone_taken[page_zonenum(page)] += nr_pages; nr_pages 1882 mm/vmscan.c int nr_pages, nr_moved = 0; nr_pages 1902 mm/vmscan.c nr_pages = hpage_nr_pages(page); nr_pages 1903 mm/vmscan.c update_lru_size(lruvec, lru, page_zonenum(page), nr_pages); nr_pages 1918 mm/vmscan.c nr_moved += nr_pages; nr_pages 3350 mm/vmscan.c unsigned long nr_pages, nr_pages 3360 mm/vmscan.c .nr_to_reclaim = max(nr_pages, SWAP_CLUSTER_MAX), nr_pages 4212 mm/vmscan.c const unsigned long nr_pages = 1 << order; nr_pages 4216 mm/vmscan.c .nr_to_reclaim = max(nr_pages, SWAP_CLUSTER_MAX), nr_pages 4247 mm/vmscan.c } while (sc.nr_reclaimed < nr_pages && --sc.priority >= 0); nr_pages 4257 mm/vmscan.c return sc.nr_reclaimed >= nr_pages; nr_pages 1033 mm/zsmalloc.c int nr_pages = class->pages_per_zspage; nr_pages 1043 mm/zsmalloc.c for (i = 0; i < nr_pages; i++) { nr_pages 22 net/9p/trans_common.c void p9_release_pages(struct page **pages, int nr_pages) nr_pages 26 net/9p/trans_common.c for (i = 0; i < nr_pages; i++) nr_pages 214 net/9p/trans_virtio.c struct page **pdata, int nr_pages, size_t offs, int count) nr_pages 220 net/9p/trans_virtio.c BUG_ON(nr_pages > (limit - start)); nr_pages 225 net/9p/trans_virtio.c while (nr_pages) { nr_pages 235 net/9p/trans_virtio.c nr_pages--; nr_pages 311 net/9p/trans_virtio.c int nr_pages; nr_pages 333 net/9p/trans_virtio.c nr_pages = DIV_ROUND_UP(n + *offs, PAGE_SIZE); nr_pages 334 net/9p/trans_virtio.c atomic_add(nr_pages, &vp_pinned); nr_pages 354 net/9p/trans_virtio.c nr_pages = DIV_ROUND_UP((unsigned long)p + len, PAGE_SIZE) - nr_pages 357 net/9p/trans_virtio.c *pages = kmalloc_array(nr_pages, sizeof(struct page *), nr_pages 364 net/9p/trans_virtio.c for (index = 0; index < nr_pages; index++) { nr_pages 2298 net/core/skbuff.c return spd->nr_pages && nr_pages 2299 net/core/skbuff.c spd->pages[spd->nr_pages - 1] == page && nr_pages 2300 net/core/skbuff.c (spd->partial[spd->nr_pages - 1].offset + nr_pages 2301 net/core/skbuff.c spd->partial[spd->nr_pages - 1].len == offset); nr_pages 2313 net/core/skbuff.c if (unlikely(spd->nr_pages == MAX_SKB_FRAGS)) nr_pages 2322 net/core/skbuff.c spd->partial[spd->nr_pages - 1].len += *len; nr_pages 2326 net/core/skbuff.c spd->pages[spd->nr_pages] = page; nr_pages 2327 net/core/skbuff.c spd->partial[spd->nr_pages].len = *len; nr_pages 2328 net/core/skbuff.c spd->partial[spd->nr_pages].offset = offset; nr_pages 2329 net/core/skbuff.c spd->nr_pages++; nr_pages 2442 net/core/skbuff.c if (spd.nr_pages) nr_pages 1130 net/dccp/proto.c unsigned long nr_pages = totalram_pages(); nr_pages 1157 net/dccp/proto.c if (nr_pages >= (128 * 1024)) nr_pages 1158 net/dccp/proto.c goal = nr_pages >> (21 - PAGE_SHIFT); nr_pages 1160 net/dccp/proto.c goal = nr_pages >> (23 - PAGE_SHIFT); nr_pages 2438 net/netfilter/nf_conntrack_core.c unsigned long nr_pages = totalram_pages(); nr_pages 2458 net/netfilter/nf_conntrack_core.c = (((nr_pages << PAGE_SHIFT) / 16384) nr_pages 2460 net/netfilter/nf_conntrack_core.c if (nr_pages > (4 * (1024 * 1024 * 1024 / PAGE_SIZE))) nr_pages 2462 net/netfilter/nf_conntrack_core.c else if (nr_pages > (1024 * 1024 * 1024 / PAGE_SIZE)) nr_pages 284 net/netfilter/xt_hashlimit.c unsigned long nr_pages = totalram_pages(); nr_pages 290 net/netfilter/xt_hashlimit.c size = (nr_pages << PAGE_SHIFT) / 16384 / nr_pages 292 net/netfilter/xt_hashlimit.c if (nr_pages > 1024 * 1024 * 1024 / PAGE_SIZE) nr_pages 163 net/rds/info.c unsigned long nr_pages = 0; nr_pages 188 net/rds/info.c nr_pages = (PAGE_ALIGN(start + len) - (start & PAGE_MASK)) nr_pages 191 net/rds/info.c pages = kmalloc_array(nr_pages, sizeof(struct page *), GFP_KERNEL); nr_pages 196 net/rds/info.c ret = get_user_pages_fast(start, nr_pages, FOLL_WRITE, pages); nr_pages 197 net/rds/info.c if (ret != nr_pages) { nr_pages 199 net/rds/info.c nr_pages = ret; nr_pages 201 net/rds/info.c nr_pages = 0; nr_pages 206 net/rds/info.c rdsdebug("len %d nr_pages %lu\n", len, nr_pages); nr_pages 238 net/rds/info.c for (i = 0; pages && i < nr_pages; i++) nr_pages 156 net/rds/rdma.c static int rds_pin_pages(unsigned long user_addr, unsigned int nr_pages, nr_pages 161 net/rds/rdma.c ret = get_user_pages_fast(user_addr, nr_pages, write ? FOLL_WRITE : 0, nr_pages 164 net/rds/rdma.c if (ret >= 0 && ret < nr_pages) { nr_pages 178 net/rds/rdma.c unsigned int nr_pages; nr_pages 198 net/rds/rdma.c nr_pages = rds_pages_in_vec(&args->vec); nr_pages 199 net/rds/rdma.c if (nr_pages == 0) { nr_pages 207 net/rds/rdma.c if ((nr_pages - 1) > (RDS_MAX_MSG_SIZE >> PAGE_SHIFT)) { nr_pages 213 net/rds/rdma.c args->vec.addr, args->vec.bytes, nr_pages); nr_pages 216 net/rds/rdma.c pages = kcalloc(nr_pages, sizeof(struct page *), GFP_KERNEL); nr_pages 250 net/rds/rdma.c ret = rds_pin_pages(args->vec.addr, nr_pages, pages, 1); nr_pages 499 net/rds/rdma.c unsigned int nr_pages; nr_pages 504 net/rds/rdma.c nr_pages = rds_pages_in_vec(&iov[i]); nr_pages 505 net/rds/rdma.c if (nr_pages == 0) nr_pages 508 net/rds/rdma.c tot_pages += nr_pages; nr_pages 527 net/rds/rdma.c unsigned int nr_pages; nr_pages 551 net/rds/rdma.c nr_pages = rds_pages_in_vec(vec); nr_pages 552 net/rds/rdma.c if (nr_pages == 0) nr_pages 555 net/rds/rdma.c tot_pages += nr_pages; nr_pages 578 net/rds/rdma.c int nr_pages; nr_pages 608 net/rds/rdma.c nr_pages = rds_rdma_pages(iovs, args->nr_local); nr_pages 609 net/rds/rdma.c if (nr_pages < 0) { nr_pages 614 net/rds/rdma.c pages = kcalloc(nr_pages, sizeof(struct page *), GFP_KERNEL); nr_pages 626 net/rds/rdma.c WARN_ON(!nr_pages); nr_pages 627 net/rds/rdma.c op->op_sg = rds_message_alloc_sgs(rm, nr_pages); nr_pages 1375 net/sctp/protocol.c unsigned long nr_pages = totalram_pages(); nr_pages 1434 net/sctp/protocol.c if (nr_pages >= (128 * 1024)) nr_pages 1435 net/sctp/protocol.c goal = nr_pages >> (22 - PAGE_SHIFT); nr_pages 1437 net/sctp/protocol.c goal = nr_pages >> (24 - PAGE_SHIFT); nr_pages 169 net/smc/smc_rx.c spd.nr_pages = 1; nr_pages 24 security/keys/big_key.c unsigned int nr_pages; nr_pages 144 security/keys/big_key.c memset(buf->virt, 0, buf->nr_pages * PAGE_SIZE); nr_pages 148 security/keys/big_key.c for (i = 0; i < buf->nr_pages; i++) nr_pages 172 security/keys/big_key.c buf->nr_pages = npg; nr_pages 176 security/keys/big_key.c for (i = 0; i < buf->nr_pages; i++) { nr_pages 186 security/keys/big_key.c buf->virt = vmap(buf->pages, buf->nr_pages, VM_MAP, PAGE_KERNEL); nr_pages 11 tools/testing/nvdimm/pmem-dax.c long nr_pages, void **kaddr, pfn_t *pfn) nr_pages 16 tools/testing/nvdimm/pmem-dax.c PFN_PHYS(nr_pages)))) nr_pages 47 tools/testing/nvdimm/pmem-dax.c return nr_pages; nr_pages 80 tools/testing/radix-tree/regression1.c unsigned int nr_pages, struct page **pages) nr_pages 35 tools/testing/selftests/vm/gup_benchmark.c int i, fd, filed, opt, nr_pages = 1, thp = -1, repeats = 1, write = 0; nr_pages 49 tools/testing/selftests/vm/gup_benchmark.c nr_pages = atoi(optarg); nr_pages 87 tools/testing/selftests/vm/gup_benchmark.c gup.nr_pages_per_call = nr_pages; nr_pages 104 tools/testing/selftests/vm/transhuge-stress.c int nr_succeed = 0, nr_failed = 0, nr_pages = 0; nr_pages 128 tools/testing/selftests/vm/transhuge-stress.c nr_pages++; nr_pages 142 tools/testing/selftests/vm/transhuge-stress.c nr_succeed, nr_failed, nr_pages); nr_pages 62 tools/testing/selftests/vm/userfaultfd.c static unsigned long nr_cpus, nr_pages, nr_pages_per_cpu, page_size; nr_pages 132 tools/testing/selftests/vm/userfaultfd.c if (madvise(rel_area, nr_pages * page_size, MADV_DONTNEED)) { nr_pages 142 tools/testing/selftests/vm/userfaultfd.c if (posix_memalign(alloc_area, page_size, nr_pages * page_size)) { nr_pages 159 tools/testing/selftests/vm/userfaultfd.c nr_pages * page_size, nr_pages 160 tools/testing/selftests/vm/userfaultfd.c nr_pages * page_size)) { nr_pages 173 tools/testing/selftests/vm/userfaultfd.c *alloc_area = mmap(NULL, nr_pages * page_size, PROT_READ | PROT_WRITE, nr_pages 177 tools/testing/selftests/vm/userfaultfd.c nr_pages * page_size); nr_pages 184 tools/testing/selftests/vm/userfaultfd.c area_alias = mmap(NULL, nr_pages * page_size, PROT_READ | PROT_WRITE, nr_pages 187 tools/testing/selftests/vm/userfaultfd.c nr_pages * page_size); nr_pages 189 tools/testing/selftests/vm/userfaultfd.c if (munmap(*alloc_area, nr_pages * page_size) < 0) nr_pages 223 tools/testing/selftests/vm/userfaultfd.c if (madvise(rel_area, nr_pages * page_size, MADV_REMOVE)) { nr_pages 233 tools/testing/selftests/vm/userfaultfd.c *alloc_area = mmap(NULL, nr_pages * page_size, PROT_READ | PROT_WRITE, nr_pages 322 tools/testing/selftests/vm/userfaultfd.c page_nr %= nr_pages; nr_pages 416 tools/testing/selftests/vm/userfaultfd.c if (offset >= nr_pages * page_size) nr_pages 721 tools/testing/selftests/vm/userfaultfd.c split_nr_pages = (nr_pages + 1) / 2; nr_pages 723 tools/testing/selftests/vm/userfaultfd.c split_nr_pages = nr_pages; nr_pages 771 tools/testing/selftests/vm/userfaultfd.c area_dst = mremap(area_dst, nr_pages * page_size, nr_pages * page_size, nr_pages 776 tools/testing/selftests/vm/userfaultfd.c for (; nr < nr_pages; nr++) { nr_pages 789 tools/testing/selftests/vm/userfaultfd.c for (nr = 0; nr < nr_pages; nr++) { nr_pages 822 tools/testing/selftests/vm/userfaultfd.c if (offset >= nr_pages * page_size) nr_pages 885 tools/testing/selftests/vm/userfaultfd.c uffdio_register.range.len = nr_pages * page_size; nr_pages 930 tools/testing/selftests/vm/userfaultfd.c uffdio_register.range.len = nr_pages * page_size; nr_pages 964 tools/testing/selftests/vm/userfaultfd.c return userfaults != nr_pages; nr_pages 989 tools/testing/selftests/vm/userfaultfd.c uffdio_register.range.len = nr_pages * page_size; nr_pages 1053 tools/testing/selftests/vm/userfaultfd.c count_verify = malloc(nr_pages * sizeof(unsigned long long)); nr_pages 1059 tools/testing/selftests/vm/userfaultfd.c for (nr = 0; nr < nr_pages; nr++) { nr_pages 1120 tools/testing/selftests/vm/userfaultfd.c uffdio_register.range.len = nr_pages * page_size; nr_pages 1190 tools/testing/selftests/vm/userfaultfd.c for (nr = 0; nr < nr_pages; nr++) { nr_pages 1316 tools/testing/selftests/vm/userfaultfd.c nr_pages = nr_pages_per_cpu * nr_cpus; nr_pages 1334 tools/testing/selftests/vm/userfaultfd.c nr_pages, nr_pages_per_cpu); nr_pages 201 tools/vm/page-types.c static unsigned long nr_pages[HASH_SIZE]; nr_pages 435 tools/vm/page-types.c for (i = 0; i < ARRAY_SIZE(nr_pages); i++) { nr_pages 436 tools/vm/page-types.c if (nr_pages[i]) nr_pages 439 tools/vm/page-types.c nr_pages[i], nr_pages 440 tools/vm/page-types.c pages2mb(nr_pages[i]), nr_pages 656 tools/vm/page-types.c nr_pages[hash_slot(flags)]++; nr_pages 720 tools/vm/page-types.c nr_pages[hash_slot(flags)]++; nr_pages 974 tools/vm/page-types.c unsigned long nr_pages, pfn, i; nr_pages 984 tools/vm/page-types.c nr_pages = (end - off + page_size - 1) / page_size; nr_pages 985 tools/vm/page-types.c if (nr_pages > PAGEMAP_BATCH) nr_pages 986 tools/vm/page-types.c nr_pages = PAGEMAP_BATCH; nr_pages 987 tools/vm/page-types.c len = nr_pages * page_size; nr_pages 1009 tools/vm/page-types.c for (i = 0; i < nr_pages ; i++) { nr_pages 1020 tools/vm/page-types.c nr_pages) != nr_pages) nr_pages 1025 tools/vm/page-types.c for (i = 0; i < nr_pages; i++) { nr_pages 1433 virt/kvm/kvm_main.c gfn_t *nr_pages, bool write) nr_pages 1441 virt/kvm/kvm_main.c if (nr_pages) nr_pages 1442 virt/kvm/kvm_main.c *nr_pages = slot->npages - (gfn - slot->base_gfn); nr_pages 1448 virt/kvm/kvm_main.c gfn_t *nr_pages) nr_pages 1450 virt/kvm/kvm_main.c return __gfn_to_hva_many(slot, gfn, nr_pages, true); nr_pages 1779 virt/kvm/kvm_main.c struct page **pages, int nr_pages) nr_pages 1788 virt/kvm/kvm_main.c if (entry < nr_pages) nr_pages 1791 virt/kvm/kvm_main.c return __get_user_pages_fast(addr, nr_pages, 1, pages);