/linux-4.1.27/fs/btrfs/ |
D | zlib.c | 228 unsigned long pg_offset; in zlib_decompress_biovec() local 238 pg_offset = 0; in zlib_decompress_biovec() 270 &page_out_index, &pg_offset); in zlib_decompress_biovec() 303 btrfs_clear_biovec_end(bvec, vcnt, page_out_index, pg_offset); in zlib_decompress_biovec() 317 unsigned long pg_offset = 0; in zlib_decompress() local 371 bytes = min(PAGE_CACHE_SIZE - pg_offset, in zlib_decompress() 376 memcpy(kaddr + pg_offset, workspace->buf + buf_offset, bytes); in zlib_decompress() 379 pg_offset += bytes; in zlib_decompress() 398 if (pg_offset < destlen) { in zlib_decompress() 400 memset(kaddr + pg_offset, 0, destlen - pg_offset); in zlib_decompress()
|
D | compression.c | 981 unsigned long *pg_offset) in btrfs_decompress_buf2page() argument 1015 bytes = min(PAGE_CACHE_SIZE - *pg_offset, in btrfs_decompress_buf2page() 1019 memcpy(kaddr + *pg_offset, buf + buf_offset, bytes); in btrfs_decompress_buf2page() 1023 *pg_offset += bytes; in btrfs_decompress_buf2page() 1029 if (*pg_offset == PAGE_CACHE_SIZE) { in btrfs_decompress_buf2page() 1035 *pg_offset = 0; in btrfs_decompress_buf2page() 1071 unsigned long pg_offset) in btrfs_clear_biovec_end() argument 1078 if (pg_offset < off) in btrfs_clear_biovec_end() 1079 pg_offset = off; in btrfs_clear_biovec_end() 1080 if (pg_offset < off + len) { in btrfs_clear_biovec_end() [all …]
|
D | compression.h | 39 unsigned long *pg_offset); 50 unsigned long pg_offset);
|
D | lzo.c | 274 unsigned long pg_offset; in lzo_decompress_biovec() local 295 pg_offset = 0; in lzo_decompress_biovec() 370 &page_out_index, &pg_offset); in lzo_decompress_biovec() 377 btrfs_clear_biovec_end(bvec, vcnt, page_out_index, pg_offset); in lzo_decompress_biovec()
|
D | extent_io.h | 179 size_t pg_offset, 337 struct page *page, unsigned int pg_offset, 340 unsigned int pg_offset); 371 struct page *page, int pg_offset, int icsum,
|
D | extent_io.c | 2027 struct page *page, unsigned int pg_offset, int mirror_num) in repair_io_failure() argument 2067 bio_add_page(bio, page, length, pg_offset); in repair_io_failure() 2113 unsigned int pg_offset) in clean_io_failure() argument 2159 pg_offset, failrec->failed_mirror); in clean_io_failure() 2355 struct page *page, int pg_offset, int icsum, in btrfs_create_repair_bio() argument 2384 bio_add_page(bio, page, failrec->len, pg_offset); in btrfs_create_repair_bio() 2874 __get_extent_map(struct inode *inode, struct page *page, size_t pg_offset, in __get_extent_map() argument 2892 em = get_extent(inode, page, pg_offset, start, len, 0); in __get_extent_map() 2929 size_t pg_offset = 0; in __do_readpage() local 2966 iosize = PAGE_CACHE_SIZE - pg_offset; in __do_readpage() [all …]
|
D | inode.c | 6633 size_t pg_offset, u64 extent_offset, in uncompress_inline() argument 6644 WARN_ON(pg_offset != 0); in uncompress_inline() 6673 size_t pg_offset, u64 start, u64 len, in btrfs_get_extent() argument 6817 extent_offset = page_offset(page) + pg_offset - extent_start; in btrfs_get_extent() 6818 copy_size = min_t(u64, PAGE_CACHE_SIZE - pg_offset, in btrfs_get_extent() 6829 pg_offset, in btrfs_get_extent() 6837 read_extent_buffer(leaf, map + pg_offset, ptr, in btrfs_get_extent() 6839 if (pg_offset + copy_size < PAGE_CACHE_SIZE) { in btrfs_get_extent() 6840 memset(map + pg_offset + copy_size, 0, in btrfs_get_extent() 6841 PAGE_CACHE_SIZE - pg_offset - in btrfs_get_extent() [all …]
|
D | send.c | 4373 unsigned pg_offset = offset & ~PAGE_CACHE_MASK; in fill_read_buf() local 4403 PAGE_CACHE_SIZE - pg_offset); in fill_read_buf() 4422 memcpy(sctx->read_buf + ret, addr + pg_offset, cur_len); in fill_read_buf() 4427 pg_offset = 0; in fill_read_buf()
|
D | ctree.h | 3863 size_t pg_offset, u64 start, u64 len, 3929 size_t pg_offset, u64 start, u64 end,
|
D | disk-io.c | 218 struct page *page, size_t pg_offset, u64 start, u64 len, in btree_get_extent() argument
|
/linux-4.1.27/fs/squashfs/ |
D | block.c | 177 int in, pg_offset = 0; in squashfs_read_data() local 184 if (pg_offset == PAGE_CACHE_SIZE) { in squashfs_read_data() 186 pg_offset = 0; in squashfs_read_data() 189 pg_offset); in squashfs_read_data() 190 memcpy(data + pg_offset, bh[k]->b_data + offset, in squashfs_read_data() 193 pg_offset += avail; in squashfs_read_data()
|
/linux-4.1.27/drivers/char/agp/ |
D | nvidia-agp.c | 33 off_t pg_offset; member 133 nvidia_private.pg_offset = 0; in nvidia_configure() 137 nvidia_private.pg_offset = (apbase & (64 * 1024 * 1024 - 1) & in nvidia_configure() 218 if (!PGE_EMPTY(agp_bridge, readl(agp_bridge->gatt_table+nvidia_private.pg_offset+j))) in nvidia_insert_memory() 229 agp_bridge->gatt_table+nvidia_private.pg_offset+j); in nvidia_insert_memory() 233 readl(agp_bridge->gatt_table+nvidia_private.pg_offset+j - 1); in nvidia_insert_memory() 254 writel(agp_bridge->scratch_page, agp_bridge->gatt_table+nvidia_private.pg_offset+i); in nvidia_remove_memory()
|
/linux-4.1.27/fs/nfs/blocklayout/ |
D | blocklayout.c | 233 unsigned int pg_offset, pg_len; in bl_read_pagelist() local 266 pg_offset = f_offset & ~PAGE_CACHE_MASK; in bl_read_pagelist() 268 if (pg_offset + bytes_left > PAGE_CACHE_SIZE) in bl_read_pagelist() 269 pg_len = PAGE_CACHE_SIZE - pg_offset; in bl_read_pagelist() 273 BUG_ON(pg_offset != 0); in bl_read_pagelist() 277 isect += (pg_offset >> SECTOR_SHIFT); in bl_read_pagelist() 278 extent_length -= (pg_offset >> SECTOR_SHIFT); in bl_read_pagelist() 284 zero_user_segment(pages[i], pg_offset, pg_len); in bl_read_pagelist() 294 pg_offset, &pg_len); in bl_read_pagelist()
|
/linux-4.1.27/fs/nfs/ |
D | pnfs.c | 1461 unsigned pg_offset; in pnfs_update_layout() local 1549 pg_offset = arg.offset & ~PAGE_CACHE_MASK; in pnfs_update_layout() 1550 if (pg_offset) { in pnfs_update_layout() 1551 arg.offset -= pg_offset; in pnfs_update_layout() 1552 arg.length += pg_offset; in pnfs_update_layout()
|
D | nfs4xdr.c | 5210 unsigned int pg_offset; in decode_getacl() local 5219 pg_offset = xdr->buf->head[0].iov_len; in decode_getacl() 5233 res->acl_data_offset = xdr_stream_pos(xdr) - pg_offset; in decode_getacl()
|