Lines Matching refs:page
50 struct page **compressed_pages;
84 static int btrfs_decompress_biovec(int type, struct page **pages_in,
111 struct page *page; in check_compressed_csum() local
121 page = cb->compressed_pages[i]; in check_compressed_csum()
124 kaddr = kmap_atomic(page); in check_compressed_csum()
159 struct page *page; in end_compressed_bio_read() local
194 page = cb->compressed_pages[index]; in end_compressed_bio_read()
195 page->mapping = NULL; in end_compressed_bio_read()
196 page_cache_release(page); in end_compressed_bio_read()
232 struct page *pages[16]; in end_compressed_writeback()
274 struct page *page; in end_compressed_bio_write() local
308 page = cb->compressed_pages[index]; in end_compressed_bio_write()
309 page->mapping = NULL; in end_compressed_bio_write()
310 page_cache_release(page); in end_compressed_bio_write()
332 struct page **compressed_pages, in btrfs_submit_compressed_write()
341 struct page *page; in btrfs_submit_compressed_write() local
376 page = compressed_pages[pg_index]; in btrfs_submit_compressed_write()
377 page->mapping = inode->i_mapping; in btrfs_submit_compressed_write()
379 ret = io_tree->ops->merge_bio_hook(WRITE, page, 0, in btrfs_submit_compressed_write()
385 page->mapping = NULL; in btrfs_submit_compressed_write()
386 if (ret || bio_add_page(bio, page, PAGE_CACHE_SIZE, 0) < in btrfs_submit_compressed_write()
416 bio_add_page(bio, page, PAGE_CACHE_SIZE, 0); in btrfs_submit_compressed_write()
453 struct page *page; in add_ra_bio_pages() local
462 page = cb->orig_bio->bi_io_vec[cb->orig_bio->bi_vcnt - 1].bv_page; in add_ra_bio_pages()
463 last_offset = (page_offset(page) + PAGE_CACHE_SIZE); in add_ra_bio_pages()
479 page = radix_tree_lookup(&mapping->page_tree, pg_index); in add_ra_bio_pages()
481 if (page && !radix_tree_exceptional_entry(page)) { in add_ra_bio_pages()
488 page = __page_cache_alloc(mapping_gfp_mask(mapping) & in add_ra_bio_pages()
490 if (!page) in add_ra_bio_pages()
493 if (add_to_page_cache_lru(page, mapping, pg_index, in add_ra_bio_pages()
495 page_cache_release(page); in add_ra_bio_pages()
505 set_page_extent_mapped(page); in add_ra_bio_pages()
517 unlock_page(page); in add_ra_bio_pages()
518 page_cache_release(page); in add_ra_bio_pages()
523 if (page->index == end_index) { in add_ra_bio_pages()
530 userpage = kmap_atomic(page); in add_ra_bio_pages()
532 flush_dcache_page(page); in add_ra_bio_pages()
537 ret = bio_add_page(cb->orig_bio, page, in add_ra_bio_pages()
542 page_cache_release(page); in add_ra_bio_pages()
545 unlock_page(page); in add_ra_bio_pages()
546 page_cache_release(page); in add_ra_bio_pages()
578 struct page *page; in btrfs_submit_compressed_read() local
625 cb->compressed_pages = kcalloc(nr_pages, sizeof(struct page *), in btrfs_submit_compressed_read()
662 page = cb->compressed_pages[pg_index]; in btrfs_submit_compressed_read()
663 page->mapping = inode->i_mapping; in btrfs_submit_compressed_read()
664 page->index = em_start >> PAGE_CACHE_SHIFT; in btrfs_submit_compressed_read()
667 ret = tree->ops->merge_bio_hook(READ, page, 0, in btrfs_submit_compressed_read()
673 page->mapping = NULL; in btrfs_submit_compressed_read()
674 if (ret || bio_add_page(comp_bio, page, PAGE_CACHE_SIZE, 0) < in btrfs_submit_compressed_read()
711 bio_add_page(comp_bio, page, PAGE_CACHE_SIZE, 0); in btrfs_submit_compressed_read()
885 struct page **pages, in btrfs_compress_pages()
924 static int btrfs_decompress_biovec(int type, struct page **pages_in, in btrfs_decompress_biovec()
947 int btrfs_decompress(int type, unsigned char *data_in, struct page *dest_page, in btrfs_decompress()
989 struct page *page_out = bvec[*pg_index].bv_page; in btrfs_decompress_buf2page()
1074 struct page *page = bvec[pg_index].bv_page; in btrfs_clear_biovec_end() local
1084 kaddr = kmap_atomic(page); in btrfs_clear_biovec_end()