Searched refs:cur_page (Results 1 – 9 of 9) sorted by relevance
/linux-4.1.27/fs/ntfs/ |
D | compress.c | 519 unsigned int xpage, max_page, cur_page, cur_ofs, i; in ntfs_read_compressed_block() local 603 cur_page = 0; in ntfs_read_compressed_block() 738 cb_max_page = (cur_page << PAGE_CACHE_SHIFT) + cur_ofs + cb_size; in ntfs_read_compressed_block() 753 for (; cur_page < cb_max_page; cur_page++) { in ntfs_read_compressed_block() 754 page = pages[cur_page]; in ntfs_read_compressed_block() 771 if (cur_page == xpage) in ntfs_read_compressed_block() 775 pages[cur_page] = NULL; in ntfs_read_compressed_block() 784 page = pages[cur_page]; in ntfs_read_compressed_block() 796 unsigned int cur2_page = cur_page; in ntfs_read_compressed_block() 815 for (; cur_page < cb_max_page; cur_page++) { in ntfs_read_compressed_block() [all …]
|
/linux-4.1.27/drivers/gpu/drm/ttm/ |
D | ttm_tt.c | 110 struct page *cur_page; in ttm_tt_set_caching() local 126 cur_page = ttm->pages[i]; in ttm_tt_set_caching() 127 if (likely(cur_page != NULL)) { in ttm_tt_set_caching() 128 ret = ttm_tt_set_page_caching(cur_page, in ttm_tt_set_caching() 142 cur_page = ttm->pages[j]; in ttm_tt_set_caching() 143 if (likely(cur_page != NULL)) { in ttm_tt_set_caching() 144 (void)ttm_tt_set_page_caching(cur_page, c_state, in ttm_tt_set_caching()
|
/linux-4.1.27/fs/ |
D | direct-io.c | 92 struct page *cur_page; /* The page */ member 675 ret = bio_add_page(sdio->bio, sdio->cur_page, in dio_bio_add_page() 683 page_cache_get(sdio->cur_page); in dio_bio_add_page() 784 if (sdio->cur_page == page && in submit_page_section() 795 if (sdio->cur_page) { in submit_page_section() 797 page_cache_release(sdio->cur_page); in submit_page_section() 798 sdio->cur_page = NULL; in submit_page_section() 804 sdio->cur_page = page; in submit_page_section() 817 page_cache_release(sdio->cur_page); in submit_page_section() 818 sdio->cur_page = NULL; in submit_page_section() [all …]
|
/linux-4.1.27/lib/ |
D | scatterlist.c | 371 unsigned int cur_page; in sg_alloc_table_from_pages() local 386 cur_page = 0; in sg_alloc_table_from_pages() 392 for (j = cur_page + 1; j < n_pages; ++j) in sg_alloc_table_from_pages() 397 chunk_size = ((j - cur_page) << PAGE_SHIFT) - offset; in sg_alloc_table_from_pages() 398 sg_set_page(s, pages[cur_page], min(size, chunk_size), offset); in sg_alloc_table_from_pages() 401 cur_page = j; in sg_alloc_table_from_pages()
|
/linux-4.1.27/fs/f2fs/ |
D | checkpoint.c | 613 struct page *cp1, *cp2, *cur_page; in get_valid_checkpoint() local 638 cur_page = cp2; in get_valid_checkpoint() 640 cur_page = cp1; in get_valid_checkpoint() 642 cur_page = cp1; in get_valid_checkpoint() 644 cur_page = cp2; in get_valid_checkpoint() 649 cp_block = (struct f2fs_checkpoint *)page_address(cur_page); in get_valid_checkpoint() 656 if (cur_page == cp2) in get_valid_checkpoint() 663 cur_page = get_meta_page(sbi, cp_blk_no + i); in get_valid_checkpoint() 664 sit_bitmap_ptr = page_address(cur_page); in get_valid_checkpoint() 666 f2fs_put_page(cur_page, 1); in get_valid_checkpoint()
|
/linux-4.1.27/drivers/infiniband/ulp/iser/ |
D | iser_memory.c | 254 int i, new_chunk, cur_page, last_ent = data->dma_nents - 1; in iser_sg_to_page_vec() local 260 cur_page = 0; in iser_sg_to_page_vec() 281 pages[cur_page++] = page; in iser_sg_to_page_vec() 288 *data_size, cur_page); in iser_sg_to_page_vec() 289 return cur_page; in iser_sg_to_page_vec()
|
/linux-4.1.27/block/ |
D | bio.c | 1289 int cur_page = 0; in bio_map_user_iov() local 1332 const int page_limit = cur_page + local_nr_pages; in bio_map_user_iov() 1336 &pages[cur_page]); in bio_map_user_iov() 1343 for (j = cur_page; j < page_limit; j++) { in bio_map_user_iov() 1363 cur_page = j; in bio_map_user_iov()
|
/linux-4.1.27/drivers/crypto/qat/qat_common/ |
D | icp_qat_uclo.h | 129 struct icp_qat_uclo_page *cur_page[ICP_QAT_UCLO_MAX_CTX]; member
|
D | qat_uclo.c | 1156 obj_handle->ae_data[ae].ae_slices[s].cur_page[ctx] = in qat_uclo_wr_uimage_page()
|