Home
last modified time | relevance | path

Searched refs:num_pages (Results 1 – 162 of 162) sorted by relevance

/linux-4.1.27/drivers/infiniband/hw/ipath/
Dipath_user_pages.c41 static void __ipath_release_user_pages(struct page **p, size_t num_pages, in __ipath_release_user_pages() argument
46 for (i = 0; i < num_pages; i++) { in __ipath_release_user_pages()
48 (unsigned long) num_pages, p[i]); in __ipath_release_user_pages()
56 static int __ipath_get_user_pages(unsigned long start_page, size_t num_pages, in __ipath_get_user_pages() argument
65 if (num_pages > lock_limit) { in __ipath_get_user_pages()
71 (unsigned long) num_pages, start_page); in __ipath_get_user_pages()
73 for (got = 0; got < num_pages; got += ret) { in __ipath_get_user_pages()
76 num_pages - got, 1, 1, in __ipath_get_user_pages()
82 current->mm->pinned_vm += num_pages; in __ipath_get_user_pages()
161 int ipath_get_user_pages(unsigned long start_page, size_t num_pages, in ipath_get_user_pages() argument
[all …]
/linux-4.1.27/drivers/infiniband/hw/qib/
Dqib_user_pages.c39 static void __qib_release_user_pages(struct page **p, size_t num_pages, in __qib_release_user_pages() argument
44 for (i = 0; i < num_pages; i++) { in __qib_release_user_pages()
54 static int __qib_get_user_pages(unsigned long start_page, size_t num_pages, in __qib_get_user_pages() argument
63 if (num_pages > lock_limit && !capable(CAP_IPC_LOCK)) { in __qib_get_user_pages()
68 for (got = 0; got < num_pages; got += ret) { in __qib_get_user_pages()
71 num_pages - got, 1, 1, in __qib_get_user_pages()
77 current->mm->pinned_vm += num_pages; in __qib_get_user_pages()
132 int qib_get_user_pages(unsigned long start_page, size_t num_pages, in qib_get_user_pages() argument
139 ret = __qib_get_user_pages(start_page, num_pages, p); in qib_get_user_pages()
146 void qib_release_user_pages(struct page **p, size_t num_pages) in qib_release_user_pages() argument
[all …]
/linux-4.1.27/drivers/gpu/drm/vmwgfx/
Dvmwgfx_gmr.c40 unsigned long num_pages, in vmw_gmr2_bind() argument
48 uint32_t remap_num = num_pages / VMW_PPN_PER_REMAP + ((num_pages % VMW_PPN_PER_REMAP) > 0); in vmw_gmr2_bind()
49 uint32_t remap_size = VMW_PPN_SIZE * num_pages + (sizeof(remap_cmd) + sizeof(*cmd)) * remap_num; in vmw_gmr2_bind()
59 define_cmd.numPages = num_pages; in vmw_gmr2_bind()
74 while (num_pages > 0) { in vmw_gmr2_bind()
75 unsigned long nr = min(num_pages, (unsigned long)VMW_PPN_PER_REMAP); in vmw_gmr2_bind()
95 num_pages -= nr; in vmw_gmr2_bind()
130 unsigned long num_pages, in vmw_gmr_bind() argument
143 return vmw_gmr2_bind(dev_priv, &data_iter, num_pages, gmr_id); in vmw_gmr_bind()
Dvmwgfx_buffer.c228 return ++(viter->i) < viter->num_pages; in __vmw_piter_non_sg_next()
296 viter->num_pages = vsgt->num_pages; in vmw_piter_start()
395 vsgt->num_pages = vmw_tt->dma_ttm.ttm.num_pages; in vmw_ttm_map_dma()
406 vmw_tt->sg_alloc_size = sgt_size + sgl_size * vsgt->num_pages; in vmw_ttm_map_dma()
413 vsgt->num_pages, 0, in vmw_ttm_map_dma()
415 vsgt->num_pages << PAGE_SHIFT, in vmw_ttm_map_dma()
420 if (vsgt->num_pages > vmw_tt->sgt.nents) { in vmw_ttm_map_dma()
422 sgl_size * (vsgt->num_pages - in vmw_ttm_map_dma()
563 ttm->num_pages, vmw_be->gmr_id); in vmw_ttm_bind()
567 vmw_mob_create(ttm->num_pages); in vmw_ttm_bind()
[all …]
Dvmwgfx_gmrid_manager.c62 gman->used_gmr_pages += bo->num_pages; in vmw_gmrid_man_get_node()
86 mem->num_pages = bo->num_pages; in vmw_gmrid_man_get_node()
96 gman->used_gmr_pages -= bo->num_pages; in vmw_gmrid_man_get_node()
110 gman->used_gmr_pages -= mem->num_pages; in vmw_gmrid_man_put_node()
Dvmwgfx_reg.h43 __le32 num_pages; member
Dvmwgfx_dmabuf.c208 place.lpfn = bo->num_pages; in vmw_dmabuf_to_start_of_vram()
227 bo->mem.start < bo->num_pages && in vmw_dmabuf_to_start_of_vram()
Dvmwgfx_mob.c58 unsigned long num_pages; member
372 mob->num_pages = vmw_mob_calculate_pt_pages(data_pages); in vmw_mob_create()
394 ret = ttm_bo_create(&dev_priv->bdev, mob->num_pages * PAGE_SIZE, in vmw_mob_pt_populate()
Dvmwgfx_drv.h238 unsigned long num_pages; member
260 unsigned long num_pages; member
599 unsigned long num_pages,
Dvmwgfx_resource.c378 size_t num_pages = PAGE_ALIGN(size) >> PAGE_SHIFT; in vmw_dmabuf_acc_size() local
379 size_t page_array_size = ttm_round_pot(num_pages * sizeof(void *)); in vmw_dmabuf_acc_size()
392 ttm_round_pot(num_pages * sizeof(dma_addr_t)); in vmw_dmabuf_acc_size()
1092 BUG_ON(res->backup->base.num_pages * PAGE_SIZE < size); in vmw_resource_buf_alloc()
Dvmwgfx_fb.c469 par->vmw_bo->base.num_pages, in vmw_fb_init()
643 par->vmw_bo->base.num_pages, in vmw_fb_on()
Dvmwgfx_shader.c480 if ((u64)buffer->base.num_pages * PAGE_SIZE < in vmw_shader_define_ioctl()
Dvmwgfx_surface.c1351 rep->buffer_size = res->backup->base.num_pages * PAGE_SIZE; in vmw_gb_surface_define_ioctl()
1431 rep->crep.buffer_size = srf->res.backup->base.num_pages * PAGE_SIZE; in vmw_gb_surface_reference_ioctl()
Dvmwgfx_execbuf.c767 if (unlikely(new_query_bo->num_pages > 4)) { in vmw_query_bo_switch_prepare()
1255 bo_size = vmw_bo->base.num_pages * PAGE_SIZE; in vmw_cmd_dma()
Dvmwgfx_kms.c1053 if (unlikely(requested_size > dmabuf->base.num_pages * PAGE_SIZE)) { in vmw_kms_new_framebuffer_dmabuf()
/linux-4.1.27/net/ceph/
Dpagevec.c16 int num_pages, bool write_page) in ceph_get_direct_page_vector() argument
22 pages = kmalloc(sizeof(*pages) * num_pages, GFP_NOFS); in ceph_get_direct_page_vector()
26 while (got < num_pages) { in ceph_get_direct_page_vector()
29 num_pages - got, write_page, 0, pages + got); in ceph_get_direct_page_vector()
45 void ceph_put_page_vector(struct page **pages, int num_pages, bool dirty) in ceph_put_page_vector() argument
49 for (i = 0; i < num_pages; i++) { in ceph_put_page_vector()
61 void ceph_release_page_vector(struct page **pages, int num_pages) in ceph_release_page_vector() argument
65 for (i = 0; i < num_pages; i++) in ceph_release_page_vector()
74 struct page **ceph_alloc_page_vector(int num_pages, gfp_t flags) in ceph_alloc_page_vector() argument
79 pages = kmalloc(sizeof(*pages) * num_pages, flags); in ceph_alloc_page_vector()
[all …]
Dosd_client.c268 int num_pages; in ceph_osd_data_release() local
270 num_pages = calc_pages_for((u64)osd_data->alignment, in ceph_osd_data_release()
272 ceph_release_page_vector(osd_data->pages, num_pages); in ceph_osd_data_release()
2682 struct page **pages, int num_pages, int page_align) in ceph_osdc_readpages() argument
2725 struct page **pages, int num_pages) in ceph_osdc_writepages() argument
/linux-4.1.27/drivers/media/v4l2-core/
Dvideobuf2-dma-sg.c51 unsigned int num_pages; member
109 int num_pages; in vb2_dma_sg_alloc() local
125 buf->num_pages = size >> PAGE_SHIFT; in vb2_dma_sg_alloc()
128 buf->pages = kzalloc(buf->num_pages * sizeof(struct page *), in vb2_dma_sg_alloc()
138 buf->num_pages, 0, size, GFP_KERNEL); in vb2_dma_sg_alloc()
161 __func__, buf->num_pages); in vb2_dma_sg_alloc()
168 num_pages = buf->num_pages; in vb2_dma_sg_alloc()
169 while (num_pages--) in vb2_dma_sg_alloc()
170 __free_page(buf->pages[num_pages]); in vb2_dma_sg_alloc()
182 int i = buf->num_pages; in vb2_dma_sg_put()
[all …]
Dvideobuf2-vmalloc.c229 int num_pages = PAGE_ALIGN(buf->size) / PAGE_SIZE; in vb2_vmalloc_dmabuf_ops_attach() local
241 ret = sg_alloc_table(sgt, num_pages, GFP_KERNEL); in vb2_vmalloc_dmabuf_ops_attach()
/linux-4.1.27/drivers/gpu/drm/ttm/
Dttm_tt.c53 ttm->pages = drm_calloc_large(ttm->num_pages, sizeof(void*)); in ttm_tt_alloc_page_directory()
58 ttm->ttm.pages = drm_calloc_large(ttm->ttm.num_pages, in ttm_dma_tt_alloc_page_directory()
62 ttm->cpu_address = (void *) (ttm->ttm.pages + ttm->ttm.num_pages); in ttm_dma_tt_alloc_page_directory()
63 ttm->dma_address = (void *) (ttm->cpu_address + ttm->ttm.num_pages); in ttm_dma_tt_alloc_page_directory()
123 drm_clflush_pages(ttm->pages, ttm->num_pages); in ttm_tt_set_caching()
125 for (i = 0; i < ttm->num_pages; ++i) { in ttm_tt_set_caching()
193 ttm->num_pages = (size + PAGE_SIZE - 1) >> PAGE_SHIFT; in ttm_tt_init()
225 ttm->num_pages = (size + PAGE_SIZE - 1) >> PAGE_SHIFT; in ttm_dma_tt_init()
303 for (i = 0; i < ttm->num_pages; ++i) { in ttm_tt_swapin()
341 ttm->num_pages << PAGE_SHIFT, in ttm_tt_swapout()
[all …]
Dttm_bo_util.c359 memset_io(new_iomap, 0, new_mem->num_pages*PAGE_SIZE); in ttm_bo_move_memcpy()
378 add = new_mem->num_pages - 1; in ttm_bo_move_memcpy()
381 for (i = 0; i < new_mem->num_pages; ++i) { in ttm_bo_move_memcpy()
530 unsigned long num_pages, in ttm_bo_kmap_ttm() argument
545 if (num_pages == 1 && (mem->placement & TTM_PL_FLAG_CACHED)) { in ttm_bo_kmap_ttm()
561 map->virtual = vmap(ttm->pages + start_page, num_pages, in ttm_bo_kmap_ttm()
568 unsigned long start_page, unsigned long num_pages, in ttm_bo_kmap() argument
579 if (num_pages > bo->num_pages) in ttm_bo_kmap()
581 if (start_page > bo->num_pages) in ttm_bo_kmap()
584 if (num_pages > 1 && !capable(CAP_SYS_ADMIN)) in ttm_bo_kmap()
[all …]
Dttm_bo.c91 bo, bo->mem.num_pages, bo->mem.size >> 10, in ttm_bo_mem_space_debug()
252 bo->ttm = bdev->driver->ttm_tt_create(bdev, bo->num_pages << PAGE_SHIFT, in ttm_bo_add_ttm()
258 bo->ttm = bdev->driver->ttm_tt_create(bdev, bo->num_pages << PAGE_SHIFT, in ttm_bo_add_ttm()
983 mem.num_pages = bo->num_pages; in ttm_bo_move_buffer()
984 mem.size = mem.num_pages << PAGE_SHIFT; in ttm_bo_move_buffer()
1013 (heap->lpfn != 0 && (mem->start + mem->num_pages) > heap->lpfn))) in ttm_bo_mem_compat()
1026 (heap->lpfn != 0 && (mem->start + mem->num_pages) > heap->lpfn))) in ttm_bo_mem_compat()
1089 unsigned long num_pages; in ttm_bo_init() local
1103 num_pages = (size + PAGE_SIZE - 1) >> PAGE_SHIFT; in ttm_bo_init()
1104 if (num_pages == 0) { in ttm_bo_init()
[all …]
Dttm_agp_backend.c59 mem = agp_allocate_memory(agp_be->bridge, ttm->num_pages, AGP_USER_MEMORY); in ttm_agp_bind()
64 for (i = 0; i < ttm->num_pages; i++) { in ttm_agp_bind()
Dttm_bo_manager.c77 ret = drm_mm_insert_node_in_range_generic(mm, node, mem->num_pages, in ttm_bo_man_get_node()
Dttm_page_alloc.c873 for (i = 0; i < ttm->num_pages; ++i) { in ttm_pool_populate()
907 for (i = 0; i < ttm->num_pages; ++i) { in ttm_pool_unpopulate()
Dttm_bo_vm.c182 if (unlikely(page_offset >= bo->num_pages)) { in ttm_bo_vm_fault()
Dttm_page_alloc_dma.c903 for (i = 0; i < ttm->num_pages; ++i) { in ttm_dma_populate()
988 for (i = 0; i < ttm->num_pages; i++) { in ttm_dma_unpopulate()
/linux-4.1.27/drivers/gpu/drm/
Ddrm_cache.c59 unsigned long num_pages) in drm_cache_flush_clflush() argument
64 for (i = 0; i < num_pages; i++) in drm_cache_flush_clflush()
71 drm_clflush_pages(struct page *pages[], unsigned long num_pages) in drm_clflush_pages() argument
76 drm_cache_flush_clflush(pages, num_pages); in drm_clflush_pages()
85 for (i = 0; i < num_pages; i++) { in drm_clflush_pages()
Ddrm_memory.c56 unsigned long i, num_pages = in agp_remap() local
83 page_map = vmalloc(num_pages * sizeof(struct page *)); in agp_remap()
88 for (i = 0; i < num_pages; ++i) in agp_remap()
90 addr = vmap(page_map, num_pages, VM_IOREMAP, PAGE_AGP); in agp_remap()
Ddrm_agpsupport.c473 unsigned long num_pages, in drm_agp_bind_pages() argument
482 mem = agp_allocate_memory(dev->agp->bridge, num_pages, in drm_agp_bind_pages()
486 num_pages); in drm_agp_bind_pages()
490 for (i = 0; i < num_pages; i++) in drm_agp_bind_pages()
492 mem->page_count = num_pages; in drm_agp_bind_pages()
/linux-4.1.27/drivers/virtio/
Dvirtio_balloon.c62 unsigned int num_pages; member
159 vb->num_pages += VIRTIO_BALLOON_PAGES_PER_PAGE; in fill_balloon()
197 vb->num_pages -= VIRTIO_BALLOON_PAGES_PER_PAGE; in leak_balloon()
287 u32 num_pages; in towards_target() local
289 virtio_cread(vb->vdev, struct virtio_balloon_config, num_pages, in towards_target()
290 &num_pages); in towards_target()
294 num_pages = le32_to_cpu((__force __le32)num_pages); in towards_target()
296 target = num_pages; in towards_target()
297 return target - vb->num_pages; in towards_target()
302 u32 actual = vb->num_pages; in update_balloon_size()
[all …]
/linux-4.1.27/drivers/gpu/drm/i915/
Di915_gem_userptr.c513 st_set_pages(struct sg_table **st, struct page **pvec, int num_pages) in st_set_pages() argument
523 ret = sg_alloc_table(*st, num_pages, GFP_KERNEL); in st_set_pages()
527 for_each_sg((*st)->sgl, sg, num_pages, n) in st_set_pages()
530 ret = sg_alloc_table_from_pages(*st, pvec, num_pages, in st_set_pages()
531 0, num_pages << PAGE_SHIFT, in st_set_pages()
551 const int num_pages = obj->base.size >> PAGE_SHIFT; in __i915_gem_userptr_get_pages_worker() local
558 pvec = kmalloc(num_pages*sizeof(struct page *), in __i915_gem_userptr_get_pages_worker()
561 pvec = drm_malloc_ab(num_pages, sizeof(struct page *)); in __i915_gem_userptr_get_pages_worker()
566 while (pinned < num_pages) { in __i915_gem_userptr_get_pages_worker()
569 num_pages - pinned, in __i915_gem_userptr_get_pages_worker()
[all …]
Di915_gpu_error.c583 int num_pages; in i915_error_object_create() local
591 num_pages = src->base.size >> PAGE_SHIFT; in i915_error_object_create()
593 dst = kmalloc(sizeof(*dst) + num_pages * sizeof(u32 *), GFP_ATOMIC); in i915_error_object_create()
607 reloc_offset + num_pages * PAGE_SIZE <= dev_priv->gtt.mappable_end); in i915_error_object_create()
617 if (reloc_offset + num_pages * PAGE_SIZE > dev_priv->gtt.mappable_end) in i915_error_object_create()
625 dst->page_count = num_pages; in i915_error_object_create()
626 while (num_pages--) { in i915_error_object_create()
/linux-4.1.27/drivers/gpu/drm/gma500/
Dmmu.c510 uint32_t num_pages, uint32_t desired_tile_stride, in psb_mmu_flush_ptes() argument
528 rows = num_pages / desired_tile_stride; in psb_mmu_flush_ptes()
530 desired_tile_stride = num_pages; in psb_mmu_flush_ptes()
558 uint32_t num_pages, uint32_t desired_tile_stride, in psb_mmu_flush_ptes() argument
566 unsigned long address, uint32_t num_pages) in psb_mmu_remove_pfn_sequence() argument
577 end = addr + (num_pages << PAGE_SHIFT); in psb_mmu_remove_pfn_sequence()
594 psb_mmu_flush_ptes(pd, f_address, num_pages, 1, 1); in psb_mmu_remove_pfn_sequence()
605 uint32_t num_pages, uint32_t desired_tile_stride, in psb_mmu_remove_pages() argument
619 rows = num_pages / desired_tile_stride; in psb_mmu_remove_pages()
621 desired_tile_stride = num_pages; in psb_mmu_remove_pages()
[all …]
Dmmu.h76 uint32_t num_pages);
80 uint32_t num_pages, int type);
85 unsigned long address, uint32_t num_pages,
89 unsigned long address, uint32_t num_pages,
Dgtt.c419 unsigned i, num_pages; in psb_gtt_init() local
536 num_pages = vram_stolen_size >> PAGE_SHIFT; in psb_gtt_init()
538 num_pages, pfn_base << PAGE_SHIFT, 0); in psb_gtt_init()
539 for (i = 0; i < num_pages; ++i) { in psb_gtt_init()
/linux-4.1.27/drivers/hv/
Dhv_balloon.c283 __u32 num_pages; member
444 __u32 num_pages; member
1070 int num_pages = range_array->finfo.page_cnt; in free_balloon_pages() local
1075 for (i = 0; i < num_pages; i++) { in free_balloon_pages()
1085 unsigned int num_pages, in alloc_balloon_pages() argument
1092 if (num_pages < alloc_unit) in alloc_balloon_pages()
1095 for (i = 0; (i * alloc_unit) < num_pages; i++) { in alloc_balloon_pages()
1129 return num_pages; in alloc_balloon_pages()
1136 unsigned int num_pages = dm_device.balloon_wrk.num_pages; in balloon_up() local
1147 WARN_ON_ONCE(num_pages % PAGES_IN_2M != 0); in balloon_up()
[all …]
/linux-4.1.27/fs/fuse/
Dfile.c352 if (idx_from < curr_index + req->num_pages && in fuse_range_is_writeback()
523 for (i = 0; i < req->num_pages; i++) { in fuse_release_user_pages()
689 for (i = start_idx; i < req->num_pages; i++) { in fuse_short_read()
726 req->num_pages = 1; in fuse_do_readpage()
770 for (i = 0; mapping == NULL && i < req->num_pages; i++) in fuse_readpages_end()
785 for (i = 0; i < req->num_pages; i++) { in fuse_readpages_end()
803 size_t count = req->num_pages << PAGE_CACHE_SHIFT; in fuse_send_readpages()
837 if (req->num_pages && in fuse_readpages_fill()
838 (req->num_pages == FUSE_MAX_PAGES_PER_REQ || in fuse_readpages_fill()
839 (req->num_pages + 1) * PAGE_CACHE_SIZE > fc->max_read || in fuse_readpages_fill()
[all …]
Ddev.c1021 for (i = 0; i < req->num_pages && (nbytes || zeroing); i++) { in fuse_copy_pages()
1692 release_pages(req->pages, req->num_pages, false); in fuse_retrieve_end()
1706 int num_pages; in fuse_retrieve() local
1717 num_pages = (num + offset + PAGE_SIZE - 1) >> PAGE_SHIFT; in fuse_retrieve()
1718 num_pages = min(num_pages, FUSE_MAX_PAGES_PER_REQ); in fuse_retrieve()
1720 req = fuse_get_req(fc, num_pages); in fuse_retrieve()
1733 while (num && req->num_pages < num_pages) { in fuse_retrieve()
1742 req->pages[req->num_pages] = page; in fuse_retrieve()
1743 req->page_descs[req->num_pages].length = this_num; in fuse_retrieve()
1744 req->num_pages++; in fuse_retrieve()
Dfuse_i.h371 unsigned num_pages; member
Dcuse.c455 req->num_pages = 1; in cuse_send_init()
Ddir.c1337 req->num_pages = 1; in fuse_readdir()
/linux-4.1.27/drivers/gpu/drm/radeon/
Dradeon_ttm.c261 unsigned num_pages; in radeon_move_blit() local
298 num_pages = new_mem->num_pages * (PAGE_SIZE / RADEON_GPU_PAGE_SIZE); in radeon_move_blit()
299 fence = radeon_copy(rdev, old_start, new_start, num_pages, bo->resv); in radeon_move_blit()
445 atomic64_add((u64)bo->num_pages << PAGE_SHIFT, &rdev->num_bytes_moved); in radeon_bo_move()
456 mem->bus.size = mem->num_pages << PAGE_SHIFT; in radeon_ttm_io_mem_reserve()
547 unsigned long end = gtt->userptr + ttm->num_pages * PAGE_SIZE; in radeon_ttm_tt_pin_userptr()
555 unsigned num_pages = ttm->num_pages - pinned; in radeon_ttm_tt_pin_userptr() local
559 r = get_user_pages(current, current->mm, userptr, num_pages, in radeon_ttm_tt_pin_userptr()
566 } while (pinned < ttm->num_pages); in radeon_ttm_tt_pin_userptr()
568 r = sg_alloc_table_from_pages(ttm->sg, ttm->pages, ttm->num_pages, 0, in radeon_ttm_tt_pin_userptr()
[all …]
Dradeon_prime.c35 int npages = bo->tbo.num_pages; in radeon_gem_prime_get_sg_table()
45 ret = ttm_bo_kmap(&bo->tbo, 0, bo->tbo.num_pages, in radeon_gem_prime_vmap()
Dradeon_object.h98 return bo->tbo.num_pages << PAGE_SHIFT; in radeon_bo_size()
103 return (bo->tbo.num_pages << PAGE_SHIFT) / RADEON_GPU_PAGE_SIZE; in radeon_bo_ngpu_pages()
Dradeon_object.c54 u64 size = (u64)bo->tbo.num_pages << PAGE_SHIFT; in radeon_update_memory_usage()
278 r = ttm_bo_kmap(&bo->tbo, 0, bo->tbo.num_pages, &bo->kmap); in radeon_bo_kmap()
637 bo->tbo.num_pages << PAGE_SHIFT); in radeon_bo_get_surface_reg()
792 size = bo->mem.num_pages << PAGE_SHIFT; in radeon_bo_fault_reserve_notify()
Dradeon_trace.h24 __entry->pages = bo->tbo.num_pages;
Dradeon_cs.c385 return (int)la->robj->tbo.num_pages - (int)lb->robj->tbo.num_pages; in cmp_size_smaller_first()
/linux-4.1.27/include/linux/ceph/
Dlibceph.h210 extern void ceph_release_page_vector(struct page **pages, int num_pages);
213 int num_pages,
215 extern void ceph_put_page_vector(struct page **pages, int num_pages,
217 extern struct page **ceph_alloc_page_vector(int num_pages, gfp_t flags);
/linux-4.1.27/drivers/scsi/bfa/
Dbfa_fcbuild.c675 int num_pages = 0; in fc_logout_params_pages() local
681 num_pages = (be16_to_cpu(prlo->payload_len) - 4) / 16; in fc_logout_params_pages()
684 num_pages = (be16_to_cpu(tprlo->payload_len) - 4) / 16; in fc_logout_params_pages()
686 return num_pages; in fc_logout_params_pages()
691 u32 d_id, u32 s_id, __be16 ox_id, int num_pages) in fc_tprlo_acc_build() argument
697 memset(tprlo_acc, 0, (num_pages * 16) + 4); in fc_tprlo_acc_build()
701 tprlo_acc->payload_len = cpu_to_be16((num_pages * 16) + 4); in fc_tprlo_acc_build()
703 for (page = 0; page < num_pages; page++) { in fc_tprlo_acc_build()
715 u32 s_id, __be16 ox_id, int num_pages) in fc_prlo_acc_build() argument
721 memset(prlo_acc, 0, (num_pages * 16) + 4); in fc_prlo_acc_build()
[all …]
Dbfa_fcbuild.h294 u32 d_id, u32 s_id, __be16 ox_id, int num_pages);
297 u32 d_id, u32 s_id, __be16 ox_id, int num_pages);
308 u16 ox_id, int num_pages);
313 u16 ox_id, int num_pages, enum fc_tprlo_type tprlo_type,
/linux-4.1.27/arch/x86/platform/efi/
Defi.c133 unsigned long long size = md->num_pages << EFI_PAGE_SHIFT; in do_add_efi_memmap()
217 md->phys_addr + (md->num_pages << EFI_PAGE_SHIFT), in print_efi_memmap()
218 (md->num_pages >> (20 - EFI_PAGE_SHIFT))); in print_efi_memmap()
516 npages = md->num_pages; in efi_set_executable()
559 size = md->num_pages << PAGE_SHIFT; in old_map_region()
599 prev_size = prev_md->num_pages << EFI_PAGE_SHIFT; in efi_merge_regions()
602 prev_md->num_pages += md->num_pages; in efi_merge_regions()
616 size = md->num_pages << EFI_PAGE_SHIFT; in get_systab_virt_addr()
988 (md->num_pages << EFI_PAGE_SHIFT)))) in efi_mem_type()
1006 (md->num_pages << EFI_PAGE_SHIFT)))) in efi_mem_attributes()
Defi_32.c43 int __init efi_setup_page_tables(unsigned long pa_memmap, unsigned num_pages) in efi_setup_page_tables() argument
47 void __init efi_cleanup_page_tables(unsigned long pa_memmap, unsigned num_pages) in efi_cleanup_page_tables() argument
Dquirks.c148 u64 size = md->num_pages << EFI_PAGE_SHIFT; in efi_reserve_boot_services()
164 md->num_pages = 0; in efi_reserve_boot_services()
179 unsigned long long size = md->num_pages << EFI_PAGE_SHIFT; in efi_free_boot_services()
Defi_64.c144 int __init efi_setup_page_tables(unsigned long pa_memmap, unsigned num_pages) in efi_setup_page_tables() argument
163 if (kernel_map_pages_in_pgd(pgd, pa_memmap, pa_memmap, num_pages, _PAGE_NX)) { in efi_setup_page_tables()
197 void __init efi_cleanup_page_tables(unsigned long pa_memmap, unsigned num_pages) in efi_cleanup_page_tables() argument
201 kernel_unmap_pages_in_pgd(pgd, pa_memmap, num_pages); in efi_cleanup_page_tables()
212 if (kernel_map_pages_in_pgd(pgd, md->phys_addr, va, md->num_pages, pf)) in __map_region()
219 unsigned long size = md->num_pages << PAGE_SHIFT; in efi_map_region()
/linux-4.1.27/drivers/scsi/be2iscsi/
Dbe_cmds.c840 req->num_pages = cpu_to_le16(PAGES_4K_SPANNED(q_mem->va, q_mem->size)); in beiscsi_cmd_eq_create()
964 req->num_pages = cpu_to_le16(PAGES_4K_SPANNED(q_mem->va, q_mem->size)); in beiscsi_cmd_cq_create()
1042 req->num_pages = PAGES_4K_SPANNED(q_mem->va, q_mem->size); in beiscsi_cmd_mccq_create()
1162 req->num_pages = PAGES_4K_SPANNED(q_mem->va, q_mem->size); in be_cmd_create_default_pdu_queue()
1261 req->num_pages = PAGES_4K_SPANNED(q_mem->va, q_mem->size); in be_cmd_wrbq_create()
1304 req->num_pages = PAGES_4K_SPANNED(q_mem->va, q_mem->size); in be_cmd_iscsi_post_template_hdr()
1336 u32 page_offset, u32 num_pages) in be_cmd_iscsi_post_sgl_pages() argument
1344 u32 temp_num_pages = num_pages; in be_cmd_iscsi_post_sgl_pages()
1346 if (num_pages == 0xff) in be_cmd_iscsi_post_sgl_pages()
1347 num_pages = 1; in be_cmd_iscsi_post_sgl_pages()
[all …]
Dbe_cmds.h271 u16 num_pages; /* sword */ member
546 u16 num_pages; member
581 u16 num_pages; member
760 u32 num_pages);
810 u16 num_pages; member
831 u16 num_pages; member
848 u16 num_pages; member
857 u16 num_pages; member
/linux-4.1.27/drivers/gpu/drm/nouveau/
Dnouveau_bo.c227 nvbo->bo.mem.num_pages = size >> PAGE_SHIFT; in nouveau_bo_new()
268 nvbo->bo.mem.num_pages < vram_pages / 4) { in set_placement_range()
437 ret = ttm_bo_kmap(&nvbo->bo, 0, nvbo->bo.mem.num_pages, in nouveau_bo_map()
473 for (i = 0; i < ttm_dma->ttm.num_pages; i++) in nouveau_bo_sync_for_device()
493 for (i = 0; i < ttm_dma->ttm.num_pages; i++) in nouveau_bo_sync_for_cpu()
709 OUT_RING (chan, new_mem->num_pages); in nve0_bo_move_copy()
733 u32 page_count = new_mem->num_pages; in nvc0_bo_move_copy()
736 page_count = new_mem->num_pages; in nvc0_bo_move_copy()
771 u32 page_count = new_mem->num_pages; in nvc0_bo_move_m2mf()
774 page_count = new_mem->num_pages; in nvc0_bo_move_m2mf()
[all …]
Dnouveau_prime.c34 int npages = nvbo->bo.num_pages; in nouveau_gem_prime_get_sg_table()
44 ret = ttm_bo_kmap(&nvbo->bo, 0, nvbo->bo.num_pages, in nouveau_gem_prime_vmap()
Dnouveau_sgdma.c39 node->size = (mem->num_pages << PAGE_SHIFT) >> 12; in nv04_sgdma_bind()
74 node->size = (mem->num_pages << PAGE_SHIFT) >> 12; in nv50_sgdma_bind()
Dnouveau_ttm.c91 ret = pfb->ram->get(pfb, mem->num_pages << PAGE_SHIFT, in nouveau_vram_manager_new()
256 ret = nvkm_vm_get(man->priv, mem->num_pages << 12, node->page_shift, in nv04_gart_manager_new()
Dnouveau_gem.c245 rep->size = nvbo->bo.mem.num_pages << PAGE_SHIFT; in nouveau_gem_info()
624 nvbo->bo.mem.num_pages << PAGE_SHIFT)) { in nouveau_gem_pushbuf_reloc_apply()
631 ret = ttm_bo_kmap(&nvbo->bo, 0, nvbo->bo.mem.num_pages, in nouveau_gem_pushbuf_reloc_apply()
804 num_pages, in nouveau_gem_ioctl_pushbuf()
/linux-4.1.27/arch/m68k/sun3/
Dconfig.c48 unsigned long num_pages; variable
121 num_pages = __pa(memory_end) >> PAGE_SHIFT; in sun3_bootmem_alloc()
127 availmem += init_bootmem_node(NODE_DATA(0), start_page, 0, num_pages); in sun3_bootmem_alloc()
/linux-4.1.27/drivers/virt/
Dfsl_hypervisor.c155 unsigned int num_pages; in ioctl_memcpy() local
218 num_pages = (param.count + lb_offset + PAGE_SIZE - 1) >> PAGE_SHIFT; in ioctl_memcpy()
226 pages = kzalloc(num_pages * sizeof(struct page *), GFP_KERNEL); in ioctl_memcpy()
236 sg_list_unaligned = kmalloc(num_pages * sizeof(struct fh_sg_list) + in ioctl_memcpy()
248 param.local_vaddr - lb_offset, num_pages, in ioctl_memcpy()
253 if (num_pinned != num_pages) { in ioctl_memcpy()
276 for (i = 1; i < num_pages; i++) { in ioctl_memcpy()
293 virt_to_phys(sg_list), num_pages); in ioctl_memcpy()
297 for (i = 0; i < num_pages; i++) in ioctl_memcpy()
/linux-4.1.27/fs/ceph/
Dfile.c349 struct page **pages, int num_pages, in striped_read() argument
369 pages_left = num_pages; in striped_read()
440 int num_pages, ret; in ceph_sync_read() local
469 num_pages = (n + start + PAGE_SIZE - 1) / PAGE_SIZE; in ceph_sync_read()
472 pages, num_pages, checkeof, in ceph_sync_read()
475 ceph_put_page_vector(pages, num_pages, true); in ceph_sync_read()
485 num_pages = calc_pages_for(off, len); in ceph_sync_read()
486 pages = ceph_alloc_page_vector(num_pages, GFP_NOFS); in ceph_sync_read()
490 num_pages, checkeof, 0, 0); in ceph_sync_read()
507 ceph_release_page_vector(pages, num_pages); in ceph_sync_read()
[all …]
Daddr.c262 int num_pages; in finish_read() local
270 num_pages = calc_pages_for((u64)osd_data->alignment, in finish_read()
272 for (i = 0; i < num_pages; i++) { in finish_read()
295 static void ceph_unlock_page_vector(struct page **pages, int num_pages) in ceph_unlock_page_vector() argument
299 for (i = 0; i < num_pages; i++) in ceph_unlock_page_vector()
606 int num_pages; in writepages_finish() local
618 num_pages = calc_pages_for((u64)osd_data->alignment, in writepages_finish()
627 wrote = num_pages; in writepages_finish()
636 for (i = 0; i < num_pages; i++) { in writepages_finish()
666 ceph_put_wrbuffer_cap_refs(ci, num_pages, snapc); in writepages_finish()
[all …]
/linux-4.1.27/arch/ia64/kernel/
Defi.c273 u64 num_pages; member
278 #define efi_md_size(md) (md->num_pages << EFI_PAGE_SHIFT)
283 return (kmd->start + (kmd->num_pages << EFI_PAGE_SHIFT)); in kmd_end()
315 end = (k->start + (k->num_pages << EFI_PAGE_SHIFT)) & PAGE_MASK; in walk()
574 size = md->num_pages << EFI_PAGE_SHIFT; in efi_init()
715 if (phys_addr - md->start < (md->num_pages << EFI_PAGE_SHIFT)) in kern_memory_descriptor()
1073 k->num_pages = md->num_pages; in efi_memmap_init()
1106 (k-1)->num_pages += in efi_memmap_init()
1112 k->num_pages = (lim - md->phys_addr) in efi_memmap_init()
1127 (k-1)->num_pages += md->num_pages; in efi_memmap_init()
[all …]
/linux-4.1.27/arch/m68k/mm/
Dsun3mmu.c31 extern unsigned long num_pages;
57 size = num_pages * sizeof(pte_t); in paging_init()
Dmcfmmu.c30 extern unsigned long num_pages;
51 size = num_pages * sizeof(pte_t); in paging_init()
80 zones_size[ZONE_DMA] = num_pages; in paging_init()
/linux-4.1.27/include/drm/ttm/
Dttm_bo_api.h122 unsigned long num_pages; member
204 unsigned long num_pages; member
633 unsigned long num_pages, struct ttm_bo_kmap_obj *map);
Dttm_bo_driver.h120 unsigned long num_pages; member
/linux-4.1.27/fs/btrfs/
Dcheck-integrity.c335 char **datav, unsigned int num_pages);
338 unsigned int num_pages,
1614 unsigned int num_pages; in btrfsic_release_block_ctx() local
1618 num_pages = (block_ctx->len + (u64)PAGE_CACHE_SIZE - 1) >> in btrfsic_release_block_ctx()
1620 while (num_pages > 0) { in btrfsic_release_block_ctx()
1621 num_pages--; in btrfsic_release_block_ctx()
1622 if (block_ctx->datav[num_pages]) { in btrfsic_release_block_ctx()
1623 kunmap(block_ctx->pagev[num_pages]); in btrfsic_release_block_ctx()
1624 block_ctx->datav[num_pages] = NULL; in btrfsic_release_block_ctx()
1626 if (block_ctx->pagev[num_pages]) { in btrfsic_release_block_ctx()
[all …]
Dfile.c409 static noinline int btrfs_copy_from_user(loff_t pos, int num_pages, in btrfs_copy_from_user() argument
464 static void btrfs_drop_pages(struct page **pages, size_t num_pages) in btrfs_drop_pages() argument
467 for (i = 0; i < num_pages; i++) { in btrfs_drop_pages()
489 struct page **pages, size_t num_pages, in btrfs_dirty_pages() argument
510 for (i = 0; i < num_pages; i++) { in btrfs_dirty_pages()
1317 size_t num_pages, loff_t pos, in prepare_pages() argument
1326 for (i = 0; i < num_pages; i++) { in prepare_pages()
1338 if (i == num_pages - 1) in prepare_pages()
1372 size_t num_pages, loff_t pos, in lock_and_cleanup_extent_if_need() argument
1382 last_pos = start_pos + ((u64)num_pages << PAGE_CACHE_SHIFT) - 1; in lock_and_cleanup_extent_if_need()
[all …]
Dextent_io.c2088 unsigned long i, num_pages = num_extent_pages(eb->start, eb->len); in repair_eb_io_failure() local
2094 for (i = 0; i < num_pages; i++) { in repair_eb_io_failure()
3612 unsigned long i, num_pages; in lock_extent_buffer_for_io() local
3662 num_pages = num_extent_pages(eb->start, eb->len); in lock_extent_buffer_for_io()
3663 for (i = 0; i < num_pages; i++) { in lock_extent_buffer_for_io()
3784 unsigned long i, num_pages; in write_one_eb() local
3790 num_pages = num_extent_pages(eb->start, eb->len); in write_one_eb()
3791 atomic_set(&eb->io_pages, num_pages); in write_one_eb()
3795 for (i = 0; i < num_pages; i++) { in write_one_eb()
3808 if (atomic_sub_and_test(num_pages - i, &eb->io_pages)) in write_one_eb()
[all …]
Dfree-space-cache.c312 int num_pages; in io_ctl_init() local
315 num_pages = DIV_ROUND_UP(i_size_read(inode), PAGE_CACHE_SIZE); in io_ctl_init()
322 (num_pages * sizeof(u32)) >= PAGE_CACHE_SIZE) in io_ctl_init()
327 io_ctl->pages = kcalloc(num_pages, sizeof(struct page *), GFP_NOFS); in io_ctl_init()
331 io_ctl->num_pages = num_pages; in io_ctl_init()
355 ASSERT(io_ctl->index < io_ctl->num_pages); in io_ctl_map_page()
370 for (i = 0; i < io_ctl->num_pages; i++) { in io_ctl_drop_pages()
386 for (i = 0; i < io_ctl->num_pages; i++) { in io_ctl_prepare_pages()
405 for (i = 0; i < io_ctl->num_pages; i++) { in io_ctl_prepare_pages()
424 io_ctl->cur += (sizeof(u32) * io_ctl->num_pages); in io_ctl_set_generation()
[all …]
Draid56.c954 int num_pages = rbio_nr_pages(stripe_len, real_stripes); in alloc_rbio() local
958 rbio = kzalloc(sizeof(*rbio) + num_pages * sizeof(struct page *) * 2 + in alloc_rbio()
972 rbio->nr_pages = num_pages; in alloc_rbio()
987 rbio->bio_pages = p + sizeof(struct page *) * num_pages; in alloc_rbio()
988 rbio->dbitmap = p + sizeof(struct page *) * num_pages * 2; in alloc_rbio()
Dextent-tree.c3225 u64 num_pages = 0; in cache_save_setup() local
3324 num_pages = div_u64(block_group->key.offset, 256 * 1024 * 1024); in cache_save_setup()
3325 if (!num_pages) in cache_save_setup()
3326 num_pages = 1; in cache_save_setup()
3328 num_pages *= 16; in cache_save_setup()
3329 num_pages *= PAGE_CACHE_SIZE; in cache_save_setup()
3331 ret = btrfs_check_data_free_space(inode, num_pages, num_pages); in cache_save_setup()
3335 ret = btrfs_prealloc_file_range_trans(inode, trans, 0, 0, num_pages, in cache_save_setup()
3336 num_pages, num_pages, in cache_save_setup()
3340 btrfs_free_reserved_data_space(inode, num_pages); in cache_save_setup()
Dctree.h1273 int num_pages; member
3997 struct page **pages, size_t num_pages,
Dioctl.c1100 unsigned long num_pages) in cluster_pages_for_defrag() argument
1119 page_cnt = min_t(u64, (u64)num_pages, (u64)file_end - start_index + 1); in cluster_pages_for_defrag()
/linux-4.1.27/arch/m68k/coldfire/
Dm54xx.c82 unsigned long num_pages; variable
94 num_pages = (_ramend - _rambase) >> PAGE_SHIFT; in mcf54xx_bootmem_alloc()
/linux-4.1.27/drivers/misc/vmw_vmci/
Dvmci_queue_pair.c153 size_t num_pages; /* Number of pages incl. header. */ member
301 const u64 num_pages = DIV_ROUND_UP(size, PAGE_SIZE) + 1; in qp_alloc_queue() local
303 if (num_pages > in qp_alloc_queue()
309 pas_size = num_pages * sizeof(*queue->kernel_if->u.g.pas); in qp_alloc_queue()
310 vas_size = num_pages * sizeof(*queue->kernel_if->u.g.vas); in qp_alloc_queue()
321 queue->kernel_if->num_pages = num_pages; in qp_alloc_queue()
327 for (i = 0; i < num_pages; i++) { in qp_alloc_queue()
627 const u64 num_pages = DIV_ROUND_UP(size, PAGE_SIZE) + 1; in qp_host_alloc_queue() local
630 if (num_pages > (SIZE_MAX - queue_size) / in qp_host_alloc_queue()
634 queue_page_size = num_pages * sizeof(*queue->kernel_if->u.h.page); in qp_host_alloc_queue()
[all …]
/linux-4.1.27/include/trace/events/
Dhswadsp.h161 __field( int, num_pages )
170 __entry->num_pages = stream->request.ringinfo.num_pages;
178 (int)__entry->num_pages, (int)__entry->ring_size,
/linux-4.1.27/drivers/misc/genwqe/
Dcard_utils.c236 int num_pages) in genwqe_unmap_pages() argument
241 for (i = 0; (i < num_pages) && (dma_list[i] != 0x0); i++) { in genwqe_unmap_pages()
249 struct page **page_list, int num_pages, in genwqe_map_pages() argument
256 for (i = 0; i < num_pages; i++) { in genwqe_map_pages()
277 genwqe_unmap_pages(cd, dma_list, num_pages); in genwqe_map_pages()
281 static int genwqe_sgl_size(int num_pages) in genwqe_sgl_size() argument
283 int len, num_tlb = num_pages / 7; in genwqe_sgl_size()
285 len = sizeof(struct sg_entry) * (num_pages+num_tlb + 1); in genwqe_sgl_size()
/linux-4.1.27/include/drm/
Ddrm_cache.h36 void drm_clflush_pages(struct page *pages[], unsigned long num_pages);
Ddrm_agpsupport.h38 unsigned long num_pages,
87 unsigned long num_pages, in drm_agp_bind_pages() argument
DdrmP.h903 void drm_clflush_pages(struct page *pages[], unsigned long num_pages);
/linux-4.1.27/arch/arm64/kernel/
Defi.c86 (addr - md->virt_addr) < (md->num_pages << EFI_PAGE_SHIFT)) in efi_to_phys()
178 npages = md->num_pages; in reserve_regions()
246 npages = md->num_pages; in efi_virtmap_init()
/linux-4.1.27/include/uapi/linux/
Dvirtio_balloon.h43 __u32 num_pages; member
/linux-4.1.27/arch/x86/include/asm/
Defi.h101 extern int __init efi_setup_page_tables(unsigned long pa_memmap, unsigned num_pages);
102 extern void __init efi_cleanup_page_tables(unsigned long pa_memmap, unsigned num_pages);
/linux-4.1.27/drivers/gpu/drm/vgem/
Dvgem_drv.c93 loff_t num_pages; in vgem_gem_fault() local
101 num_pages = DIV_ROUND_UP(obj->base.size, PAGE_SIZE); in vgem_gem_fault()
103 if (page_offset > num_pages) in vgem_gem_fault()
/linux-4.1.27/arch/powerpc/kvm/
De500_mmu.c749 int num_pages, ret, i; in kvm_vcpu_ioctl_config_tlb() local
780 num_pages = DIV_ROUND_UP(cfg->array + array_len - 1, PAGE_SIZE) - in kvm_vcpu_ioctl_config_tlb()
782 pages = kmalloc(sizeof(struct page *) * num_pages, GFP_KERNEL); in kvm_vcpu_ioctl_config_tlb()
786 ret = get_user_pages_fast(cfg->array, num_pages, 1, pages); in kvm_vcpu_ioctl_config_tlb()
790 if (ret != num_pages) { in kvm_vcpu_ioctl_config_tlb()
791 num_pages = ret; in kvm_vcpu_ioctl_config_tlb()
796 virt = vmap(pages, num_pages, VM_MAP, PAGE_KERNEL); in kvm_vcpu_ioctl_config_tlb()
838 vcpu_e500->num_shared_tlb_pages = num_pages; in kvm_vcpu_ioctl_config_tlb()
854 for (i = 0; i < num_pages; i++) in kvm_vcpu_ioctl_config_tlb()
/linux-4.1.27/drivers/firmware/efi/
Druntime-map.c53 EFI_RUNTIME_U64_ATTR_SHOW(num_pages);
73 static struct map_attribute map_num_pages_attr = __ATTR_RO(num_pages);
Defi.c255 u64 size = md->num_pages << EFI_PAGE_SHIFT; in efi_lookup_mapped_addr()
/linux-4.1.27/drivers/firmware/efi/libstub/
Defi-stub-helper.c176 if (desc->num_pages < nr_pages) in efi_high_alloc()
180 end = start + desc->num_pages * (1UL << EFI_PAGE_SHIFT); in efi_high_alloc()
261 if (desc->num_pages < nr_pages) in efi_low_alloc()
265 end = start + desc->num_pages * (1UL << EFI_PAGE_SHIFT); in efi_low_alloc()
Darm-stub.c328 left_end = left->phys_addr + left->num_pages * EFI_PAGE_SIZE; in regions_are_adjacent()
381 size = in->num_pages * EFI_PAGE_SIZE; in efi_get_virtmap()
/linux-4.1.27/drivers/gpu/drm/via/
Dvia_dmablit.c187 for (i = 0; i < vsg->num_pages; ++i) { in via_free_sg_info()
235 vsg->num_pages = VIA_PFN(xfer->mem_addr + (xfer->num_lines * xfer->mem_stride - 1)) - in via_lock_all_dma_pages()
238 vsg->pages = vzalloc(sizeof(struct page *) * vsg->num_pages); in via_lock_all_dma_pages()
244 vsg->num_pages, in via_lock_all_dma_pages()
249 if (ret != vsg->num_pages) { in via_lock_all_dma_pages()
Dvia_dmablit.h42 unsigned long num_pages; member
/linux-4.1.27/drivers/infiniband/hw/mlx5/
Dmem.c151 int page_shift, size_t offset, size_t num_pages, in __mlx5_ib_populate_pas() argument
170 for (i = 0; i < num_pages; ++i) { in __mlx5_ib_populate_pas()
Dmlx5_ib.h604 int page_shift, size_t offset, size_t num_pages,
/linux-4.1.27/fs/exofs/
Dore_raid.c639 unsigned num_pages; in _ore_add_parity_unit() local
645 num_pages = _sp2d_max_pg(sp2d) + 1 - si->cur_pg; in _ore_add_parity_unit()
660 for (i = 0; i < num_pages; i++) { in _ore_add_parity_unit()
669 BUG_ON(si->cur_pg + num_pages > sp2d->pages_in_unit); in _ore_add_parity_unit()
672 per_dev, num_pages * PAGE_SIZE); in _ore_add_parity_unit()
/linux-4.1.27/drivers/block/xen-blkback/
Dblkback.c145 unsigned int num_pages = 0; in shrink_free_pagepool() local
151 page[num_pages] = list_first_entry(&blkif->free_pages, in shrink_free_pagepool()
153 list_del(&page[num_pages]->lru); in shrink_free_pagepool()
155 if (++num_pages == NUM_BATCH_FREE_PAGES) { in shrink_free_pagepool()
157 gnttab_free_pages(num_pages, page); in shrink_free_pagepool()
159 num_pages = 0; in shrink_free_pagepool()
163 if (num_pages != 0) in shrink_free_pagepool()
164 gnttab_free_pages(num_pages, page); in shrink_free_pagepool()
/linux-4.1.27/drivers/gpu/drm/tegra/
Dgem.h42 unsigned long num_pages; member
Dgem.c199 bo->num_pages = bo->gem.size >> PAGE_SHIFT; in tegra_bo_get_pages()
201 sgt = drm_prime_pages_to_sg(bo->pages, bo->num_pages); in tegra_bo_get_pages()
531 if (sg_alloc_table(sgt, bo->num_pages, GFP_KERNEL)) in tegra_gem_prime_map_dma_buf()
534 for_each_sg(sgt->sgl, sg, bo->num_pages, i) in tegra_gem_prime_map_dma_buf()
Dfb.c264 bo->vaddr = vmap(bo->pages, bo->num_pages, VM_MAP, in tegra_fbdev_probe()
/linux-4.1.27/drivers/gpu/drm/qxl/
Dqxl_ttm.c224 mem->bus.size = mem->num_pages << PAGE_SHIFT; in qxl_ttm_io_mem_reserve()
269 if (!ttm->num_pages) { in qxl_ttm_backend_bind()
271 ttm->num_pages, bo_mem, ttm); in qxl_ttm_backend_bind()
Dqxl_object.h57 return bo->tbo.num_pages << PAGE_SHIFT; in qxl_bo_size()
Dqxl_object.c135 r = ttm_bo_kmap(&bo->tbo, 0, bo->tbo.num_pages, &bo->kmap); in qxl_bo_kmap()
/linux-4.1.27/arch/x86/mm/
Dpageattr.c880 unsigned num_pages, pmd_t *pmd, pgprot_t pgprot) in populate_pte() argument
886 while (num_pages-- && start < end) { in populate_pte()
902 unsigned num_pages, pud_t *pud, pgprot_t pgprot) in populate_pmd() argument
912 unsigned long pre_end = start + (num_pages << PAGE_SHIFT); in populate_pmd()
917 cur_pages = min_t(unsigned int, num_pages, cur_pages); in populate_pmd()
935 if (num_pages == cur_pages) in populate_pmd()
968 populate_pte(cpa, start, end, num_pages - cur_pages, in populate_pmd()
971 return num_pages; in populate_pmd()
/linux-4.1.27/drivers/gpu/drm/mgag200/
Dmgag200_cursor.c126 ret = ttm_bo_kmap(&bo->bo, 0, bo->bo.num_pages, &bo->kmap); in mga_crtc_cursor_set()
190 pixels_prev->bo.num_pages, in mga_crtc_cursor_set()
Dmgag200_fb.c80 ret = ttm_bo_kmap(&bo->bo, 0, bo->bo.num_pages, &bo->kmap); in mga_dirty_update()
Dmgag200_ttm.c164 mem->bus.size = mem->num_pages << PAGE_SHIFT; in mgag200_ttm_io_mem_reserve()
Dmgag200_mode.c762 ret = ttm_bo_kmap(&bo->bo, 0, bo->bo.num_pages, &bo->kmap); in mga_crtc_do_set_base()
/linux-4.1.27/drivers/misc/carma/
Dcarma-fpga-program.c534 size_t num_pages, len, avail = 0; in fpga_program_dma() local
546 num_pages = DIV_ROUND_UP(priv->bytes, FIFO_MAX_LEN); in fpga_program_dma()
547 ret = sg_alloc_table(&table, num_pages, GFP_KERNEL); in fpga_program_dma()
565 for_each_sg(table.sgl, sg, num_pages, i) { in fpga_program_dma()
602 tx = dmaengine_prep_dma_sg(chan, table.sgl, num_pages, in fpga_program_dma()
/linux-4.1.27/drivers/scsi/bnx2fc/
Dbnx2fc_tgt.c673 int num_pages; in bnx2fc_alloc_session_resc() local
731 num_pages = tgt->rq_mem_size / CNIC_PAGE_SIZE; in bnx2fc_alloc_session_resc()
735 while (num_pages--) { in bnx2fc_alloc_session_resc()
786 num_pages = tgt->confq_mem_size / CNIC_PAGE_SIZE; in bnx2fc_alloc_session_resc()
790 while (num_pages--) { in bnx2fc_alloc_session_resc()
/linux-4.1.27/drivers/infiniband/hw/ocrdma/
Docrdma_abi.h70 u32 num_pages; member
Docrdma_hw.c425 cmd->num_pages = 4; in ocrdma_mbx_create_eq()
429 ocrdma_build_q_pages(&cmd->pa[0], cmd->num_pages, eq->q.dma, in ocrdma_mbx_create_eq()
549 int num_pages, status; in ocrdma_mbx_create_mq() local
555 num_pages = PAGES_4K_SPANNED(mq->va, mq->size); in ocrdma_mbx_create_mq()
560 cmd->cqid_pages = num_pages; in ocrdma_mbx_create_mq()
573 ocrdma_build_q_pages(pa, num_pages, mq->dma, PAGE_SIZE_4K); in ocrdma_mbx_create_mq()
1550 int *num_pages, int *page_size) in ocrdma_build_q_conf() argument
1566 *num_pages = in ocrdma_build_q_conf()
Docrdma_sli.h299 u32 num_pages; member
Docrdma_verbs.c977 uresp.num_pages = 1; in ocrdma_copy_cq_uresp()
/linux-4.1.27/fs/ocfs2/
Daops.h36 void ocfs2_unlock_and_free_pages(struct page **pages, int num_pages);
Daops.c1238 void ocfs2_unlock_and_free_pages(struct page **pages, int num_pages) in ocfs2_unlock_and_free_pages() argument
1242 for(i = 0; i < num_pages; i++) { in ocfs2_unlock_and_free_pages()
Dalloc.c6836 int ret, i, has_data, num_pages = 0; in ocfs2_convert_inline_data_to_extents() local
6916 ret = ocfs2_grab_eof_pages(inode, 0, end, pages, &num_pages); in ocfs2_convert_inline_data_to_extents()
6938 for (i = 0; i < num_pages; i++) in ocfs2_convert_inline_data_to_extents()
6972 ocfs2_unlock_and_free_pages(pages, num_pages); in ocfs2_convert_inline_data_to_extents()
/linux-4.1.27/drivers/net/ethernet/8390/
Dsmc-ultra.c212 unsigned char num_pages, irqreg, addr, piomode; in ultra_probe1() local
290 num_pages = num_pages_tbl[(addr >> 4) & 3]; in ultra_probe1()
297 ei_status.stop_page = num_pages; in ultra_probe1()
/linux-4.1.27/fs/cifs/
Dfile.c2351 cifs_write_allocate_pages(struct page **pages, unsigned long num_pages) in cifs_write_allocate_pages() argument
2356 for (i = 0; i < num_pages; i++) { in cifs_write_allocate_pages()
2363 num_pages = i; in cifs_write_allocate_pages()
2370 for (i = 0; i < num_pages; i++) in cifs_write_allocate_pages()
2379 size_t num_pages; in get_numpages() local
2383 num_pages = DIV_ROUND_UP(clen, PAGE_SIZE); in get_numpages()
2388 return num_pages; in get_numpages()
2424 size_t *len, unsigned long *num_pages) in wdata_fill_from_iovec() argument
2427 unsigned long i, nr_pages = *num_pages; in wdata_fill_from_iovec()
2460 *num_pages = i + 1; in wdata_fill_from_iovec()
[all …]
/linux-4.1.27/drivers/scsi/
Dstorvsc_drv.c606 int num_pages; in create_bounce_buffer() local
611 num_pages = ALIGN(len, PAGE_SIZE) >> PAGE_SHIFT; in create_bounce_buffer()
613 bounce_sgl = kcalloc(num_pages, sizeof(struct scatterlist), GFP_ATOMIC); in create_bounce_buffer()
617 sg_init_table(bounce_sgl, num_pages); in create_bounce_buffer()
618 for (i = 0; i < num_pages; i++) { in create_bounce_buffer()
628 destroy_bounce_buffer(bounce_sgl, num_pages); in create_bounce_buffer()
/linux-4.1.27/Documentation/ABI/testing/
Dsysfs-firmware-efi-runtime-map28 num_pages : The size of the memory range in pages.
/linux-4.1.27/drivers/block/zram/
Dzram_drv.c348 size_t num_pages = disksize >> PAGE_SHIFT; in zram_meta_free() local
352 for (index = 0; index < num_pages; index++) { in zram_meta_free()
368 size_t num_pages; in zram_meta_alloc() local
375 num_pages = disksize >> PAGE_SHIFT; in zram_meta_alloc()
376 meta->table = vzalloc(num_pages * sizeof(*meta->table)); in zram_meta_alloc()
/linux-4.1.27/Documentation/device-mapper/
Dkcopyd.txt12 int kcopyd_client_create(unsigned int num_pages,
/linux-4.1.27/sound/soc/intel/haswell/
Dsst-haswell-ipc.h277 u32 num_pages; member
436 u32 ring_pt_address, u32 num_pages,
Dsst-haswell-ipc.c1137 u32 ring_pt_address, u32 num_pages, in sst_hsw_stream_buffer() argument
1146 stream->request.ringinfo.num_pages = num_pages; in sst_hsw_stream_buffer()
/linux-4.1.27/drivers/net/ethernet/mellanox/mlx5/core/
Dpagealloc.c78 __be32 num_pages; member
184 *npages = be32_to_cpu(out.num_pages); in mlx5_cmd_query_pages()
Deq.c276 s32 npages = be32_to_cpu(eqe->data.req_pages.num_pages); in mlx5_eq_int()
/linux-4.1.27/drivers/s390/char/
Dsclp_vt220.c639 static int __init __sclp_vt220_init(int num_pages) in __sclp_vt220_init() argument
659 for (i = 0; i < num_pages; i++) { in __sclp_vt220_init()
Dzcore.c190 u32 num_pages; member
592 hdr->num_pages = mem_size / PAGE_SIZE; in zcore_header_init()
/linux-4.1.27/drivers/net/ethernet/emulex/benet/
Dbe_cmds.h328 u16 num_pages; /* sword */ member
430 u16 num_pages; member
493 u16 num_pages; member
501 u16 num_pages; member
520 u8 num_pages; member
546 u8 num_pages; member
Dbe_cmds.c919 req->num_pages = cpu_to_le16(PAGES_4K_SPANNED(q_mem->va, q_mem->size)); in be_cmd_eq_create()
1081 req->num_pages = cpu_to_le16(PAGES_4K_SPANNED(q_mem->va, q_mem->size)); in be_cmd_cq_create()
1159 req->num_pages = cpu_to_le16(PAGES_4K_SPANNED(q_mem->va, q_mem->size)); in be_cmd_mccq_ext_create()
1224 req->num_pages = cpu_to_le16(PAGES_4K_SPANNED(q_mem->va, q_mem->size)); in be_cmd_mccq_org_create()
1286 req->num_pages = PAGES_4K_SPANNED(q_mem->va, q_mem->size); in be_cmd_txq_create()
1333 req->num_pages = 2; in be_cmd_rxq_create()
/linux-4.1.27/drivers/gpu/drm/bochs/
Dbochs_fbdev.c98 ret = ttm_bo_kmap(&bo->bo, 0, bo->bo.num_pages, in bochsfb_create()
Dbochs_mm.c142 mem->bus.size = mem->num_pages << PAGE_SHIFT; in bochs_ttm_io_mem_reserve()
/linux-4.1.27/drivers/net/ethernet/smsc/
Dsmc91c92_cs.c1200 u_short num_pages; in smc_start_xmit() local
1217 num_pages = skb->len >> 8; in smc_start_xmit()
1219 if (num_pages > 7) { in smc_start_xmit()
1220 netdev_err(dev, "Far too big packet error: %d pages\n", num_pages); in smc_start_xmit()
1239 outw(MC_ALLOC | num_pages, ioaddr + MMU_CMD); in smc_start_xmit()
/linux-4.1.27/drivers/char/agp/
Dgeneric.c1195 …t agp_generic_alloc_pages(struct agp_bridge_data *bridge, struct agp_memory *mem, size_t num_pages) in agp_generic_alloc_pages() argument
1200 for (i = 0; i < num_pages; i++) { in agp_generic_alloc_pages()
1217 set_pages_array_uc(mem->pages, num_pages); in agp_generic_alloc_pages()
/linux-4.1.27/drivers/md/
Dbitmap.c758 unsigned long num_pages; in bitmap_storage_alloc() local
765 num_pages = DIV_ROUND_UP(bytes, PAGE_SIZE); in bitmap_storage_alloc()
766 offset = slot_number * (num_pages - 1); in bitmap_storage_alloc()
769 * num_pages, GFP_KERNEL); in bitmap_storage_alloc()
786 for ( ; pnum < num_pages; pnum++) { in bitmap_storage_alloc()
799 roundup(DIV_ROUND_UP(num_pages*4, 8), sizeof(unsigned long)), in bitmap_storage_alloc()
/linux-4.1.27/drivers/scsi/bnx2i/
Dbnx2i_hwi.c952 int num_pages; in setup_qp_page_tables() local
964 num_pages = ep->qp.sq_mem_size / CNIC_PAGE_SIZE; in setup_qp_page_tables()
971 while (num_pages--) { in setup_qp_page_tables()
992 num_pages = ep->qp.rq_mem_size / CNIC_PAGE_SIZE; in setup_qp_page_tables()
999 while (num_pages--) { in setup_qp_page_tables()
1020 num_pages = ep->qp.cq_mem_size / CNIC_PAGE_SIZE; in setup_qp_page_tables()
1027 while (num_pages--) { in setup_qp_page_tables()
/linux-4.1.27/drivers/net/ethernet/broadcom/
Dcnic.h132 int num_pages; member
Dcnic.c728 for (i = 0; i < dma->num_pages; i++) { in cnic_free_dma()
742 dma->num_pages = 0; in cnic_free_dma()
750 for (i = 0; i < dma->num_pages; i++) { in cnic_setup_page_tbl()
764 for (i = 0; i < dma->num_pages; i++) { in cnic_setup_page_tbl_le()
785 dma->num_pages = pages; in cnic_alloc_dma()
2312 fcoe_init->eq_pbl_size = cp->kcq2.dma.num_pages; in cnic_bnx2x_fcoe_init1()
/linux-4.1.27/arch/ia64/hp/sim/boot/
Dfw-emu.c252 md->num_pages = (end - start) >> 12; \ in sys_fw_init()
/linux-4.1.27/drivers/gpu/drm/ast/
Dast_mode.c547 ret = ttm_bo_kmap(&bo->bo, 0, bo->bo.num_pages, &bo->kmap); in ast_crtc_do_set_base()
931 ret = ttm_bo_kmap(&bo->bo, 0, bo->bo.num_pages, &ast->cache_kmap); in ast_cursor_init()
1169 ret = ttm_bo_kmap(&bo->bo, 0, bo->bo.num_pages, &uobj_map); in ast_cursor_set()
Dast_fb.c104 ret = ttm_bo_kmap(&bo->bo, 0, bo->bo.num_pages, &bo->kmap); in ast_dirty_update()
Dast_ttm.c164 mem->bus.size = mem->num_pages << PAGE_SHIFT; in ast_ttm_io_mem_reserve()
/linux-4.1.27/drivers/gpu/drm/cirrus/
Dcirrus_fbdev.c77 ret = ttm_bo_kmap(&bo->bo, 0, bo->bo.num_pages, &bo->kmap); in cirrus_dirty_update()
Dcirrus_ttm.c164 mem->bus.size = mem->num_pages << PAGE_SHIFT; in cirrus_ttm_io_mem_reserve()
Dcirrus_mode.c169 ret = ttm_bo_kmap(&bo->bo, 0, bo->bo.num_pages, &bo->kmap); in cirrus_crtc_do_set_base()
/linux-4.1.27/drivers/block/drbd/
Ddrbd_bitmap.c1050 int num_pages, i, count = 0; in bm_rw() local
1093 num_pages = b->bm_number_of_pages; in bm_rw()
1098 for (i = 0; i < num_pages; i++) { in bm_rw()
/linux-4.1.27/drivers/staging/slicoss/
Dslic.h70 u32 num_pages; member
Dslicoss.c1168 for (i = 0; i < rspq->num_pages; i++) { in slic_rspqueue_free()
1190 rspq->num_pages = SLIC_RSPQ_PAGES_GB; in slic_rspqueue_init()
1192 for (i = 0; i < rspq->num_pages; i++) { in slic_rspqueue_init()
1235 rspq->pageindex = (rspq->pageindex + 1) % rspq->num_pages; in slic_rspqueue_getnext()
/linux-4.1.27/drivers/gpu/drm/omapdrm/
Domap_dmm_tiler.c396 int num_pages = (size + PAGE_SIZE - 1) >> PAGE_SHIFT; in tiler_reserve_1d() local
404 if (tcm_reserve_1d(containers[TILFMT_PAGE], num_pages, in tiler_reserve_1d()
/linux-4.1.27/drivers/infiniband/hw/ehca/
Dehca_mrmw.c1899 int num_pages, in ehca_check_kpages_per_ate() argument
1902 for (; *sg && num_pages > 0; *sg = sg_next(*sg), num_pages--) { in ehca_check_kpages_per_ate()
1910 pgaddr, *prev_pgaddr, num_pages); in ehca_check_kpages_per_ate()
/linux-4.1.27/mm/
Dmemory_hotplug.c339 unsigned long start_pfn, unsigned long num_pages) in ensure_zone_is_initialized() argument
342 return init_currently_empty_zone(zone, start_pfn, num_pages, in ensure_zone_is_initialized()
/linux-4.1.27/arch/x86/boot/compressed/
Deboot.c1242 prev->size += d->num_pages << 12; in setup_e820()
1258 e820_map->size = d->num_pages << PAGE_SHIFT; in setup_e820()
/linux-4.1.27/include/linux/mlx5/
Ddevice.h482 __be32 num_pages; member
/linux-4.1.27/drivers/staging/android/ion/
Dion.c222 int num_pages = PAGE_ALIGN(buffer->size) / PAGE_SIZE; in ion_buffer_create() local
226 buffer->pages = vmalloc(sizeof(struct page *) * num_pages); in ion_buffer_create()
/linux-4.1.27/include/linux/
Defi.h110 u64 num_pages; member
/linux-4.1.27/drivers/net/ethernet/broadcom/bnx2x/
Dbnx2x_ethtool.c810 int num_pages = __bnx2x_get_page_reg_num(bp); in bnx2x_read_pages_regs() local
821 for (i = 0; i < num_pages; i++) { in bnx2x_read_pages_regs()
/linux-4.1.27/kernel/
Dmodule.c1710 void set_page_attributes(void *start, void *end, int (*set)(unsigned long start, int num_pages)) in set_page_attributes() argument