Home
last modified time | relevance | path

Searched refs:cur_page (Results 1 – 9 of 9) sorted by relevance

/linux-4.1.27/fs/ntfs/
Dcompress.c519 unsigned int xpage, max_page, cur_page, cur_ofs, i; in ntfs_read_compressed_block() local
603 cur_page = 0; in ntfs_read_compressed_block()
738 cb_max_page = (cur_page << PAGE_CACHE_SHIFT) + cur_ofs + cb_size; in ntfs_read_compressed_block()
753 for (; cur_page < cb_max_page; cur_page++) { in ntfs_read_compressed_block()
754 page = pages[cur_page]; in ntfs_read_compressed_block()
771 if (cur_page == xpage) in ntfs_read_compressed_block()
775 pages[cur_page] = NULL; in ntfs_read_compressed_block()
784 page = pages[cur_page]; in ntfs_read_compressed_block()
796 unsigned int cur2_page = cur_page; in ntfs_read_compressed_block()
815 for (; cur_page < cb_max_page; cur_page++) { in ntfs_read_compressed_block()
[all …]
/linux-4.1.27/drivers/gpu/drm/ttm/
Dttm_tt.c110 struct page *cur_page; in ttm_tt_set_caching() local
126 cur_page = ttm->pages[i]; in ttm_tt_set_caching()
127 if (likely(cur_page != NULL)) { in ttm_tt_set_caching()
128 ret = ttm_tt_set_page_caching(cur_page, in ttm_tt_set_caching()
142 cur_page = ttm->pages[j]; in ttm_tt_set_caching()
143 if (likely(cur_page != NULL)) { in ttm_tt_set_caching()
144 (void)ttm_tt_set_page_caching(cur_page, c_state, in ttm_tt_set_caching()
/linux-4.1.27/fs/
Ddirect-io.c92 struct page *cur_page; /* The page */ member
675 ret = bio_add_page(sdio->bio, sdio->cur_page, in dio_bio_add_page()
683 page_cache_get(sdio->cur_page); in dio_bio_add_page()
784 if (sdio->cur_page == page && in submit_page_section()
795 if (sdio->cur_page) { in submit_page_section()
797 page_cache_release(sdio->cur_page); in submit_page_section()
798 sdio->cur_page = NULL; in submit_page_section()
804 sdio->cur_page = page; in submit_page_section()
817 page_cache_release(sdio->cur_page); in submit_page_section()
818 sdio->cur_page = NULL; in submit_page_section()
[all …]
/linux-4.1.27/lib/
Dscatterlist.c371 unsigned int cur_page; in sg_alloc_table_from_pages() local
386 cur_page = 0; in sg_alloc_table_from_pages()
392 for (j = cur_page + 1; j < n_pages; ++j) in sg_alloc_table_from_pages()
397 chunk_size = ((j - cur_page) << PAGE_SHIFT) - offset; in sg_alloc_table_from_pages()
398 sg_set_page(s, pages[cur_page], min(size, chunk_size), offset); in sg_alloc_table_from_pages()
401 cur_page = j; in sg_alloc_table_from_pages()
/linux-4.1.27/fs/f2fs/
Dcheckpoint.c613 struct page *cp1, *cp2, *cur_page; in get_valid_checkpoint() local
638 cur_page = cp2; in get_valid_checkpoint()
640 cur_page = cp1; in get_valid_checkpoint()
642 cur_page = cp1; in get_valid_checkpoint()
644 cur_page = cp2; in get_valid_checkpoint()
649 cp_block = (struct f2fs_checkpoint *)page_address(cur_page); in get_valid_checkpoint()
656 if (cur_page == cp2) in get_valid_checkpoint()
663 cur_page = get_meta_page(sbi, cp_blk_no + i); in get_valid_checkpoint()
664 sit_bitmap_ptr = page_address(cur_page); in get_valid_checkpoint()
666 f2fs_put_page(cur_page, 1); in get_valid_checkpoint()
/linux-4.1.27/drivers/infiniband/ulp/iser/
Diser_memory.c254 int i, new_chunk, cur_page, last_ent = data->dma_nents - 1; in iser_sg_to_page_vec() local
260 cur_page = 0; in iser_sg_to_page_vec()
281 pages[cur_page++] = page; in iser_sg_to_page_vec()
288 *data_size, cur_page); in iser_sg_to_page_vec()
289 return cur_page; in iser_sg_to_page_vec()
/linux-4.1.27/block/
Dbio.c1289 int cur_page = 0; in bio_map_user_iov() local
1332 const int page_limit = cur_page + local_nr_pages; in bio_map_user_iov()
1336 &pages[cur_page]); in bio_map_user_iov()
1343 for (j = cur_page; j < page_limit; j++) { in bio_map_user_iov()
1363 cur_page = j; in bio_map_user_iov()
/linux-4.1.27/drivers/crypto/qat/qat_common/
Dicp_qat_uclo.h129 struct icp_qat_uclo_page *cur_page[ICP_QAT_UCLO_MAX_CTX]; member
Dqat_uclo.c1156 obj_handle->ae_data[ae].ae_slices[s].cur_page[ctx] = in qat_uclo_wr_uimage_page()