Searched refs:cur_page (Results 1 – 9 of 9) sorted by relevance
/linux-4.4.14/fs/ntfs/ |
D | compress.c | 519 unsigned int xpage, max_page, cur_page, cur_ofs, i; in ntfs_read_compressed_block() local 603 cur_page = 0; in ntfs_read_compressed_block() 738 cb_max_page = (cur_page << PAGE_CACHE_SHIFT) + cur_ofs + cb_size; in ntfs_read_compressed_block() 753 for (; cur_page < cb_max_page; cur_page++) { in ntfs_read_compressed_block() 754 page = pages[cur_page]; in ntfs_read_compressed_block() 771 if (cur_page == xpage) in ntfs_read_compressed_block() 775 pages[cur_page] = NULL; in ntfs_read_compressed_block() 784 page = pages[cur_page]; in ntfs_read_compressed_block() 796 unsigned int cur2_page = cur_page; in ntfs_read_compressed_block() 815 for (; cur_page < cb_max_page; cur_page++) { in ntfs_read_compressed_block() [all …]
|
/linux-4.4.14/drivers/gpu/drm/ttm/ |
D | ttm_tt.c | 110 struct page *cur_page; in ttm_tt_set_caching() local 126 cur_page = ttm->pages[i]; in ttm_tt_set_caching() 127 if (likely(cur_page != NULL)) { in ttm_tt_set_caching() 128 ret = ttm_tt_set_page_caching(cur_page, in ttm_tt_set_caching() 142 cur_page = ttm->pages[j]; in ttm_tt_set_caching() 143 if (likely(cur_page != NULL)) { in ttm_tt_set_caching() 144 (void)ttm_tt_set_page_caching(cur_page, c_state, in ttm_tt_set_caching()
|
/linux-4.4.14/fs/ |
D | direct-io.c | 92 struct page *cur_page; /* The page */ member 685 ret = bio_add_page(sdio->bio, sdio->cur_page, in dio_bio_add_page() 693 page_cache_get(sdio->cur_page); in dio_bio_add_page() 794 if (sdio->cur_page == page && in submit_page_section() 805 if (sdio->cur_page) { in submit_page_section() 807 page_cache_release(sdio->cur_page); in submit_page_section() 808 sdio->cur_page = NULL; in submit_page_section() 814 sdio->cur_page = page; in submit_page_section() 827 page_cache_release(sdio->cur_page); in submit_page_section() 828 sdio->cur_page = NULL; in submit_page_section() [all …]
|
/linux-4.4.14/lib/ |
D | scatterlist.c | 399 unsigned int cur_page; in sg_alloc_table_from_pages() local 414 cur_page = 0; in sg_alloc_table_from_pages() 420 for (j = cur_page + 1; j < n_pages; ++j) in sg_alloc_table_from_pages() 425 chunk_size = ((j - cur_page) << PAGE_SHIFT) - offset; in sg_alloc_table_from_pages() 426 sg_set_page(s, pages[cur_page], min(size, chunk_size), offset); in sg_alloc_table_from_pages() 429 cur_page = j; in sg_alloc_table_from_pages()
|
/linux-4.4.14/fs/f2fs/ |
D | checkpoint.c | 660 struct page *cp1, *cp2, *cur_page; in get_valid_checkpoint() local 685 cur_page = cp2; in get_valid_checkpoint() 687 cur_page = cp1; in get_valid_checkpoint() 689 cur_page = cp1; in get_valid_checkpoint() 691 cur_page = cp2; in get_valid_checkpoint() 696 cp_block = (struct f2fs_checkpoint *)page_address(cur_page); in get_valid_checkpoint() 703 if (cur_page == cp2) in get_valid_checkpoint() 710 cur_page = get_meta_page(sbi, cp_blk_no + i); in get_valid_checkpoint() 711 sit_bitmap_ptr = page_address(cur_page); in get_valid_checkpoint() 713 f2fs_put_page(cur_page, 1); in get_valid_checkpoint()
|
/linux-4.4.14/drivers/infiniband/ulp/iser/ |
D | iser_memory.c | 157 int i, new_chunk, cur_page, last_ent = data->dma_nents - 1; in iser_sg_to_page_vec() local 163 cur_page = 0; in iser_sg_to_page_vec() 184 pages[cur_page++] = page; in iser_sg_to_page_vec() 191 *data_size, cur_page); in iser_sg_to_page_vec() 192 return cur_page; in iser_sg_to_page_vec()
|
/linux-4.4.14/block/ |
D | bio.c | 1257 int cur_page = 0; in bio_map_user_iov() local 1300 const int page_limit = cur_page + local_nr_pages; in bio_map_user_iov() 1304 &pages[cur_page]); in bio_map_user_iov() 1311 for (j = cur_page; j < page_limit; j++) { in bio_map_user_iov() 1331 cur_page = j; in bio_map_user_iov()
|
/linux-4.4.14/drivers/crypto/qat/qat_common/ |
D | icp_qat_uclo.h | 129 struct icp_qat_uclo_page *cur_page[ICP_QAT_UCLO_MAX_CTX]; member
|
D | qat_uclo.c | 1141 obj_handle->ae_data[ae].ae_slices[s].cur_page[ctx] = in qat_uclo_wr_uimage_page()
|