Home
last modified time | relevance | path

Searched refs:PAGE_CACHE_SIZE (Results 1 – 200 of 272) sorted by relevance

12

/linux-4.1.27/net/ceph/
Dpagevec.c106 l = min_t(int, PAGE_CACHE_SIZE-po, left); in ceph_copy_user_to_page_vector()
113 if (po == PAGE_CACHE_SIZE) { in ceph_copy_user_to_page_vector()
131 size_t l = min_t(size_t, PAGE_CACHE_SIZE-po, left); in ceph_copy_to_page_vector()
137 if (po == PAGE_CACHE_SIZE) { in ceph_copy_to_page_vector()
154 size_t l = min_t(size_t, PAGE_CACHE_SIZE-po, left); in ceph_copy_from_page_vector()
160 if (po == PAGE_CACHE_SIZE) { in ceph_copy_from_page_vector()
182 int end = min((int)PAGE_CACHE_SIZE, off + len); in ceph_zero_page_vector_range()
189 while (len >= PAGE_CACHE_SIZE) { in ceph_zero_page_vector_range()
191 zero_user_segment(pages[i], 0, PAGE_CACHE_SIZE); in ceph_zero_page_vector_range()
192 len -= PAGE_CACHE_SIZE; in ceph_zero_page_vector_range()
/linux-4.1.27/fs/btrfs/
Dzlib.c62 workspace->buf = kmalloc(PAGE_CACHE_SIZE, GFP_NOFS); in zlib_alloc_workspace()
120 workspace->strm.avail_out = PAGE_CACHE_SIZE; in zlib_compress_pages()
121 workspace->strm.avail_in = min(len, PAGE_CACHE_SIZE); in zlib_compress_pages()
159 workspace->strm.avail_out = PAGE_CACHE_SIZE; in zlib_compress_pages()
175 start += PAGE_CACHE_SIZE; in zlib_compress_pages()
180 PAGE_CACHE_SIZE); in zlib_compress_pages()
226 unsigned long total_pages_in = DIV_ROUND_UP(srclen, PAGE_CACHE_SIZE); in zlib_decompress_biovec()
232 workspace->strm.avail_in = min_t(size_t, srclen, PAGE_CACHE_SIZE); in zlib_decompress_biovec()
237 workspace->strm.avail_out = PAGE_CACHE_SIZE; in zlib_decompress_biovec()
277 workspace->strm.avail_out = PAGE_CACHE_SIZE; in zlib_decompress_biovec()
[all …]
Dcompression.c125 csum = btrfs_csum_data(kaddr, csum, PAGE_CACHE_SIZE); in check_compressed_csum()
347 WARN_ON(start & ((u64)PAGE_CACHE_SIZE - 1)); in btrfs_submit_compressed_write()
380 PAGE_CACHE_SIZE, in btrfs_submit_compressed_write()
386 if (ret || bio_add_page(bio, page, PAGE_CACHE_SIZE, 0) < in btrfs_submit_compressed_write()
387 PAGE_CACHE_SIZE) { in btrfs_submit_compressed_write()
416 bio_add_page(bio, page, PAGE_CACHE_SIZE, 0); in btrfs_submit_compressed_write()
418 if (bytes_left < PAGE_CACHE_SIZE) { in btrfs_submit_compressed_write()
423 bytes_left -= PAGE_CACHE_SIZE; in btrfs_submit_compressed_write()
424 first_byte += PAGE_CACHE_SIZE; in btrfs_submit_compressed_write()
463 last_offset = (page_offset(page) + PAGE_CACHE_SIZE); in add_ra_bio_pages()
[all …]
Dlzo.c58 workspace->buf = vmalloc(lzo1x_worst_compress(PAGE_CACHE_SIZE)); in lzo_alloc_workspace()
59 workspace->cbuf = vmalloc(lzo1x_worst_compress(PAGE_CACHE_SIZE)); in lzo_alloc_workspace()
136 pg_bytes_left = PAGE_CACHE_SIZE - LZO_LEN; in lzo_compress_pages()
139 in_len = min(len, PAGE_CACHE_SIZE); in lzo_compress_pages()
204 pg_bytes_left = PAGE_CACHE_SIZE; in lzo_compress_pages()
226 start += PAGE_CACHE_SIZE; in lzo_compress_pages()
229 in_len = min(bytes_left, PAGE_CACHE_SIZE); in lzo_compress_pages()
269 unsigned long total_pages_in = DIV_ROUND_UP(srclen, PAGE_CACHE_SIZE); in lzo_decompress_biovec()
292 in_page_bytes_left = PAGE_CACHE_SIZE - LZO_LEN; in lzo_decompress_biovec()
348 in_page_bytes_left = PAGE_CACHE_SIZE; in lzo_decompress_biovec()
[all …]
Dextent_io.c1724 max_bytes = PAGE_CACHE_SIZE; in find_lock_delalloc_range()
1988 u64 end = start + PAGE_CACHE_SIZE - 1; in check_page_uptodate()
2098 PAGE_CACHE_SIZE, start, p, in repair_eb_io_failure()
2102 start += PAGE_CACHE_SIZE; in repair_eb_io_failure()
2499 if (bvec->bv_offset || bvec->bv_len != PAGE_CACHE_SIZE) { in end_bio_extent_writepage()
2500 if (bvec->bv_offset + bvec->bv_len != PAGE_CACHE_SIZE) in end_bio_extent_writepage()
2579 if (bvec->bv_offset || bvec->bv_len != PAGE_CACHE_SIZE) { in end_bio_extent_readpage()
2580 if (bvec->bv_offset + bvec->bv_len != PAGE_CACHE_SIZE) in end_bio_extent_readpage()
2644 off = i_size & (PAGE_CACHE_SIZE-1); in end_bio_extent_readpage()
2646 zero_user_segment(page, off, PAGE_CACHE_SIZE); in end_bio_extent_readpage()
[all …]
Dinode-map.c286 #define INODES_PER_BITMAP (PAGE_CACHE_SIZE * 8)
320 PAGE_CACHE_SIZE / sizeof(*info); in recalculate_thresholds()
484 prealloc = ALIGN(prealloc, PAGE_CACHE_SIZE); in btrfs_save_ino_cache()
485 prealloc += ctl->total_bitmaps * PAGE_CACHE_SIZE; in btrfs_save_ino_cache()
489 prealloc += 8 * PAGE_CACHE_SIZE; in btrfs_save_ino_cache()
Dstruct-funcs.c69 (token->offset + PAGE_CACHE_SIZE >= offset + size)) { \
107 (token->offset + PAGE_CACHE_SIZE >= offset + size)) { \
Dcheck-integrity.c770 if (ret < (int)PAGE_CACHE_SIZE) { in btrfsic_process_superblock()
1244 size_t start_offset = block_ctx->start & ((u64)PAGE_CACHE_SIZE - 1); in btrfsic_read_from_block_data()
1248 offset_in_page = (start_offset + offset) & (PAGE_CACHE_SIZE - 1); in btrfsic_read_from_block_data()
1251 cur = min(len, ((size_t)PAGE_CACHE_SIZE - offset_in_page)); in btrfsic_read_from_block_data()
1252 BUG_ON(i >= DIV_ROUND_UP(block_ctx->len, PAGE_CACHE_SIZE)); in btrfsic_read_from_block_data()
1618 num_pages = (block_ctx->len + (u64)PAGE_CACHE_SIZE - 1) >> in btrfsic_release_block_ctx()
1650 if (block_ctx->dev_bytenr & ((u64)PAGE_CACHE_SIZE - 1)) { in btrfsic_read_block()
1657 num_pages = (block_ctx->len + (u64)PAGE_CACHE_SIZE - 1) >> in btrfsic_read_block()
1689 PAGE_CACHE_SIZE, 0); in btrfsic_read_block()
1690 if (PAGE_CACHE_SIZE != ret) in btrfsic_read_block()
[all …]
Draid56.c272 memcpy(d, s, PAGE_CACHE_SIZE); in cache_rbio_pages()
941 return DIV_ROUND_UP(nr, PAGE_CACHE_SIZE); in rbio_nr_pages()
1076 ret = bio_add_page(last, page, PAGE_CACHE_SIZE, 0); in rbio_add_io_page()
1077 if (ret == PAGE_CACHE_SIZE) in rbio_add_io_page()
1092 bio_add_page(bio, page, PAGE_CACHE_SIZE, 0); in rbio_add_io_page()
1248 run_xor(pointers + 1, nr_data - 1, PAGE_CACHE_SIZE); in finish_rmw()
1502 int nr_pages = DIV_ROUND_UP(rbio->stripe_len, PAGE_CACHE_SIZE); in raid56_rmw_stripe()
1798 int nr_pages = DIV_ROUND_UP(rbio->stripe_len, PAGE_CACHE_SIZE); in __raid_recover_end_io()
1912 PAGE_CACHE_SIZE); in __raid_recover_end_io()
1921 run_xor(pointers, rbio->nr_data - 1, PAGE_CACHE_SIZE); in __raid_recover_end_io()
[all …]
Dfile.c417 int offset = pos & (PAGE_CACHE_SIZE - 1); in btrfs_copy_from_user()
421 PAGE_CACHE_SIZE - offset, write_bytes); in btrfs_copy_from_user()
451 if (copied < PAGE_CACHE_SIZE - offset) { in btrfs_copy_from_user()
1299 if (((pos & (PAGE_CACHE_SIZE - 1)) || force_uptodate) && in prepare_uptodate_page()
1381 start_pos = pos & ~((u64)PAGE_CACHE_SIZE - 1); in lock_and_cleanup_extent_if_need()
1488 nrptrs = min(DIV_ROUND_UP(iov_iter_count(i), PAGE_CACHE_SIZE), in __btrfs_buffered_write()
1489 PAGE_CACHE_SIZE / (sizeof(struct page *))); in __btrfs_buffered_write()
1499 size_t offset = pos & (PAGE_CACHE_SIZE - 1); in __btrfs_buffered_write()
1501 nrptrs * (size_t)PAGE_CACHE_SIZE - in __btrfs_buffered_write()
1504 PAGE_CACHE_SIZE); in __btrfs_buffered_write()
[all …]
Dfree-space-cache.c32 #define BITS_PER_BITMAP (PAGE_CACHE_SIZE * 8)
315 num_pages = DIV_ROUND_UP(i_size_read(inode), PAGE_CACHE_SIZE); in io_ctl_init()
322 (num_pages * sizeof(u32)) >= PAGE_CACHE_SIZE) in io_ctl_init()
359 io_ctl->size = PAGE_CACHE_SIZE; in io_ctl_map_page()
361 memset(io_ctl->cur, 0, PAGE_CACHE_SIZE); in io_ctl_map_page()
480 PAGE_CACHE_SIZE - offset); in io_ctl_set_crc()
508 PAGE_CACHE_SIZE - offset); in io_ctl_check_crc()
566 memcpy(io_ctl->cur, bitmap, PAGE_CACHE_SIZE); in io_ctl_add_bitmap()
626 memcpy(entry->bitmap, io_ctl->cur, PAGE_CACHE_SIZE); in io_ctl_read_bitmap()
780 e->bitmap = kzalloc(PAGE_CACHE_SIZE, GFP_NOFS); in __load_free_space_cache()
[all …]
Dinode.c191 PAGE_CACHE_SIZE); in insert_inline_extent()
208 offset = start & (PAGE_CACHE_SIZE - 1); in insert_inline_extent()
260 actual_end > PAGE_CACHE_SIZE || in cow_file_range_inline()
426 nr_pages = min(nr_pages, (128 * 1024UL) / PAGE_CACHE_SIZE); in compress_file_range()
504 (PAGE_CACHE_SIZE - 1); in compress_file_range()
514 PAGE_CACHE_SIZE - offset); in compress_file_range()
570 total_in = ALIGN(total_in, PAGE_CACHE_SIZE); in compress_file_range()
956 (end - start + PAGE_CACHE_SIZE) / PAGE_CACHE_SIZE; in cow_file_range()
1096 nr_pages = (async_cow->end - async_cow->start + PAGE_CACHE_SIZE) >> in async_cow_submit()
1151 nr_pages = (cur_end - start + PAGE_CACHE_SIZE) >> in cow_file_range_async()
[all …]
Dextent_io.h122 #define MAX_INLINE_EXTENT_BUFFER_SIZE (INLINE_EXTENT_BUFFER_PAGES * PAGE_CACHE_SIZE)
283 return ((start + len + PAGE_CACHE_SIZE - 1) >> PAGE_CACHE_SHIFT) - in num_extent_pages()
Dioctl.c893 em = lookup_extent_mapping(em_tree, offset, PAGE_CACHE_SIZE); in check_defrag_in_cache()
983 u64 len = PAGE_CACHE_SIZE; in defrag_lookup_extent()
1138 page_end = page_start + PAGE_CACHE_SIZE - 1; in cluster_pages_for_defrag()
1198 page_end = page_offset(pages[i_done - 1]) + PAGE_CACHE_SIZE; in cluster_pages_for_defrag()
1335 (i < DIV_ROUND_UP(i_size_read(inode), PAGE_CACHE_SIZE))) { in btrfs_defrag_file()
1358 next = DIV_ROUND_UP(skip, PAGE_CACHE_SIZE); in btrfs_defrag_file()
1710 if (vol_args->size > PAGE_CACHE_SIZE) { in btrfs_ioctl_snap_create_v2()
2844 unsigned int cmp_len = PAGE_CACHE_SIZE; in btrfs_cmp_data()
2848 if (len < PAGE_CACHE_SIZE) in btrfs_cmp_data()
2990 if (WARN_ON_ONCE(bs < PAGE_CACHE_SIZE)) { in btrfs_ioctl_file_extent_same()
[all …]
Dscrub.c1696 memset(mapped_buffer, 0, PAGE_CACHE_SIZE); in scrub_write_page_to_dev_replace()
4119 while (len >= PAGE_CACHE_SIZE) { in copy_nocow_pages_for_inode()
4178 offset += PAGE_CACHE_SIZE; in copy_nocow_pages_for_inode()
4179 physical_for_dev_replace += PAGE_CACHE_SIZE; in copy_nocow_pages_for_inode()
4180 nocow_ctx_logical += PAGE_CACHE_SIZE; in copy_nocow_pages_for_inode()
4181 len -= PAGE_CACHE_SIZE; in copy_nocow_pages_for_inode()
4215 ret = bio_add_page(bio, page, PAGE_CACHE_SIZE, 0); in write_page_nocow()
4216 if (ret != PAGE_CACHE_SIZE) { in write_page_nocow()
Dfile-item.c34 PAGE_CACHE_SIZE))
204 if (bio->bi_iter.bi_size > PAGE_CACHE_SIZE * 8) in __btrfs_lookup_bio_sums()
/linux-4.1.27/fs/ecryptfs/
Dmmap.c126 loff_t num_extents_per_page = (PAGE_CACHE_SIZE in ecryptfs_copy_up_encrypted_with_header()
142 memset(page_virt, 0, PAGE_CACHE_SIZE); in ecryptfs_copy_up_encrypted_with_header()
202 PAGE_CACHE_SIZE, in ecryptfs_readpage()
219 page, page->index, 0, PAGE_CACHE_SIZE, in ecryptfs_readpage()
254 if ((i_size_read(inode) / PAGE_CACHE_SIZE) != page->index) in fill_zeros_to_end_of_page()
256 end_byte_in_page = i_size_read(inode) % PAGE_CACHE_SIZE; in fill_zeros_to_end_of_page()
259 zero_user_segment(page, end_byte_in_page, PAGE_CACHE_SIZE); in fill_zeros_to_end_of_page()
300 page, index, 0, PAGE_CACHE_SIZE, mapping->host); in ecryptfs_write_begin()
326 page, index, 0, PAGE_CACHE_SIZE, in ecryptfs_write_begin()
340 zero_user(page, 0, PAGE_CACHE_SIZE); in ecryptfs_write_begin()
[all …]
Dread_write.c128 size_t num_bytes = (PAGE_CACHE_SIZE - start_offset_in_page); in ecryptfs_write()
168 PAGE_CACHE_SIZE - start_offset_in_page); in ecryptfs_write()
Dcrypto.c297 remainder_of_page = PAGE_CACHE_SIZE - offset; in virt_to_scatterlist()
439 extent_base = (((loff_t)page_index) * (PAGE_CACHE_SIZE / extent_size)); in crypt_extent()
509 extent_offset < (PAGE_CACHE_SIZE / crypt_stat->extent_size); in ecryptfs_encrypt_page()
523 PAGE_CACHE_SIZE); in ecryptfs_encrypt_page()
571 rc = ecryptfs_read_lower(page_virt, lower_offset, PAGE_CACHE_SIZE, in ecryptfs_decrypt_page()
582 extent_offset < (PAGE_CACHE_SIZE / crypt_stat->extent_size); in ecryptfs_decrypt_page()
670 if (PAGE_CACHE_SIZE <= ECRYPTFS_MINIMUM_HEADER_EXTENT_SIZE) in ecryptfs_set_default_sizes()
674 crypt_stat->metadata_size = PAGE_CACHE_SIZE; in ecryptfs_set_default_sizes()
1453 memset(page_virt, 0, PAGE_CACHE_SIZE); in ecryptfs_read_metadata()
1486 memset(page_virt, 0, PAGE_CACHE_SIZE); in ecryptfs_read_metadata()
Dmain.c697 .size = PAGE_CACHE_SIZE,
702 .size = PAGE_CACHE_SIZE,
820 if (ECRYPTFS_DEFAULT_EXTENT_SIZE > PAGE_CACHE_SIZE) { in ecryptfs_init()
828 (unsigned long)PAGE_CACHE_SIZE); in ecryptfs_init()
/linux-4.1.27/fs/hfsplus/
Dbnode.c30 l = min_t(int, len, PAGE_CACHE_SIZE - off); in hfs_bnode_read()
36 l = min_t(int, len, PAGE_CACHE_SIZE); in hfs_bnode_read()
83 l = min_t(int, len, PAGE_CACHE_SIZE - off); in hfs_bnode_write()
90 l = min_t(int, len, PAGE_CACHE_SIZE); in hfs_bnode_write()
113 l = min_t(int, len, PAGE_CACHE_SIZE - off); in hfs_bnode_clear()
119 l = min_t(int, len, PAGE_CACHE_SIZE); in hfs_bnode_clear()
145 l = min_t(int, len, PAGE_CACHE_SIZE - src); in hfs_bnode_copy()
152 l = min_t(int, len, PAGE_CACHE_SIZE); in hfs_bnode_copy()
164 if (PAGE_CACHE_SIZE - src < PAGE_CACHE_SIZE - dst) { in hfs_bnode_copy()
165 l = PAGE_CACHE_SIZE - src; in hfs_bnode_copy()
[all …]
Dbtree.c239 (tree->node_size + PAGE_CACHE_SIZE - 1) >> in hfs_btree_open()
406 if (++off >= PAGE_CACHE_SIZE) { in hfs_bmap_alloc()
Dbitmap.c16 #define PAGE_CACHE_BITS (PAGE_CACHE_SIZE * 8)
Dxattr.c223 for (; written < node_size; index++, written += PAGE_CACHE_SIZE) { in hfsplus_create_attributes_file()
234 min_t(size_t, PAGE_CACHE_SIZE, node_size - written)); in hfsplus_create_attributes_file()
/linux-4.1.27/fs/squashfs/
Dfile_direct.c99 bytes = res % PAGE_CACHE_SIZE; in squashfs_readpage_block()
102 memset(pageaddr + bytes, 0, PAGE_CACHE_SIZE - bytes); in squashfs_readpage_block()
156 bytes -= PAGE_CACHE_SIZE, offset += PAGE_CACHE_SIZE) { in squashfs_read_cache()
157 int avail = min_t(int, bytes, PAGE_CACHE_SIZE); in squashfs_read_cache()
164 memset(pageaddr + avail, 0, PAGE_CACHE_SIZE - avail); in squashfs_read_cache()
Dlzo_wrapper.c105 if (bytes <= PAGE_CACHE_SIZE) { in lzo_uncompress()
109 memcpy(data, buff, PAGE_CACHE_SIZE); in lzo_uncompress()
110 buff += PAGE_CACHE_SIZE; in lzo_uncompress()
111 bytes -= PAGE_CACHE_SIZE; in lzo_uncompress()
Dlz4_wrapper.c120 if (bytes <= PAGE_CACHE_SIZE) { in lz4_uncompress()
124 memcpy(data, buff, PAGE_CACHE_SIZE); in lz4_uncompress()
125 buff += PAGE_CACHE_SIZE; in lz4_uncompress()
126 bytes -= PAGE_CACHE_SIZE; in lz4_uncompress()
Dcache.c278 entry->data[j] = kmalloc(PAGE_CACHE_SIZE, GFP_KERNEL); in squashfs_cache_init()
317 void *buff = entry->data[offset / PAGE_CACHE_SIZE] in squashfs_copy_data()
318 + (offset % PAGE_CACHE_SIZE); in squashfs_copy_data()
320 PAGE_CACHE_SIZE - (offset % PAGE_CACHE_SIZE)); in squashfs_copy_data()
418 int pages = (length + PAGE_CACHE_SIZE - 1) >> PAGE_CACHE_SHIFT; in squashfs_read_table()
439 for (i = 0; i < pages; i++, buffer += PAGE_CACHE_SIZE) in squashfs_read_table()
Dfile.c178 __le32 *blist = kmalloc(PAGE_CACHE_SIZE, GFP_KERNEL); in read_indexes()
186 int blocks = min_t(int, n, PAGE_CACHE_SIZE >> 2); in read_indexes()
390 bytes -= PAGE_CACHE_SIZE, offset += PAGE_CACHE_SIZE) { in squashfs_copy_cache()
392 int avail = buffer ? min_t(int, bytes, PAGE_CACHE_SIZE) : 0; in squashfs_copy_cache()
407 memset(pageaddr + avail, 0, PAGE_CACHE_SIZE - avail); in squashfs_copy_cache()
465 if (page->index >= ((i_size_read(inode) + PAGE_CACHE_SIZE - 1) >> in squashfs_readpage()
490 memset(pageaddr, 0, PAGE_CACHE_SIZE); in squashfs_readpage()
Dpage_actor.c51 actor->length = length ? : pages * PAGE_CACHE_SIZE; in squashfs_page_actor_init()
91 actor->length = length ? : pages * PAGE_CACHE_SIZE; in squashfs_page_actor_init_special()
Dsymlink.c54 int length = min_t(int, i_size_read(inode) - index, PAGE_CACHE_SIZE); in squashfs_symlink_readpage()
97 memset(pageaddr + length, 0, PAGE_CACHE_SIZE - length); in squashfs_symlink_readpage()
Dzlib_wrapper.c72 stream->avail_out = PAGE_CACHE_SIZE; in zlib_uncompress()
88 stream->avail_out = PAGE_CACHE_SIZE; in zlib_uncompress()
Dxz_wrapper.c144 stream->buf.out_size = PAGE_CACHE_SIZE; in squashfs_xz_uncompress()
161 total += PAGE_CACHE_SIZE; in squashfs_xz_uncompress()
Dblock.c184 if (pg_offset == PAGE_CACHE_SIZE) { in squashfs_read_data()
188 avail = min_t(int, in, PAGE_CACHE_SIZE - in squashfs_read_data()
Dpage_actor.h27 actor->length = length ? : pages * PAGE_CACHE_SIZE; in squashfs_page_actor_init()
Ddecompressor.c105 buffer = kmalloc(PAGE_CACHE_SIZE, GFP_KERNEL); in get_comp_opts()
Dsuper.c156 if (PAGE_CACHE_SIZE > msblk->block_size) { in squashfs_fill_super()
/linux-4.1.27/fs/ceph/
Daddr.c142 if (offset != 0 || length != PAGE_CACHE_SIZE) { in ceph_invalidatepage()
196 u64 len = PAGE_CACHE_SIZE; in readpage_nounlock()
199 zero_user_segment(page, 0, PAGE_CACHE_SIZE); in readpage_nounlock()
211 zero_user_segment(page, 0, PAGE_CACHE_SIZE); in readpage_nounlock()
233 if (err < PAGE_CACHE_SIZE) in readpage_nounlock()
235 zero_user_segment(page, err, PAGE_CACHE_SIZE); in readpage_nounlock()
277 if (bytes < (int)PAGE_CACHE_SIZE) { in finish_read()
280 zero_user_segment(page, s, PAGE_CACHE_SIZE); in finish_read()
290 bytes -= PAGE_CACHE_SIZE; in finish_read()
414 if (fsc->mount_options->rsize >= PAGE_CACHE_SIZE) in ceph_readpages()
[all …]
Dfile.c632 (pos+n) | (PAGE_CACHE_SIZE-1)); in ceph_sync_direct_write()
737 num_pages = (len + PAGE_CACHE_SIZE - 1) >> PAGE_CACHE_SHIFT; in ceph_sync_write()
883 iocb->ki_pos < PAGE_CACHE_SIZE) { in ceph_read_iter()
886 end = min_t(loff_t, end, PAGE_CACHE_SIZE); in ceph_read_iter()
1132 zero_user(page, offset & (PAGE_CACHE_SIZE - 1), size); in ceph_zero_partial_page()
1141 loff_t nearly = round_up(offset, PAGE_CACHE_SIZE); in ceph_zero_pagecache_range()
1150 if (length >= PAGE_CACHE_SIZE) { in ceph_zero_pagecache_range()
1151 loff_t size = round_down(length, PAGE_CACHE_SIZE); in ceph_zero_pagecache_range()
Dsuper.c896 if (fsc->mount_options->rasize >= PAGE_CACHE_SIZE) in ceph_register_bdi()
898 (fsc->mount_options->rasize + PAGE_CACHE_SIZE - 1) in ceph_register_bdi()
902 VM_MAX_READAHEAD * 1024 / PAGE_CACHE_SIZE; in ceph_register_bdi()
Dmds_client.h99 #define CEPH_CAPS_PER_RELEASE ((PAGE_CACHE_SIZE - \
/linux-4.1.27/fs/cramfs/
DREADME89 PAGE_CACHE_SIZE for cramfs_readpage's convenience.)
93 PAGE_CACHE_SIZE may grow in future (if I interpret the comment
96 Currently, mkcramfs #define's PAGE_CACHE_SIZE as 4096 and uses that
97 for blksize, whereas Linux-2.3.39 uses its PAGE_CACHE_SIZE, which in
102 One option is to change mkcramfs to take its PAGE_CACHE_SIZE from
105 PAGE_CACHE_SIZE (4096)' to `#include <asm/page.h>'. The disadvantage
108 PAGE_CACHE_SIZE is subject to change between kernel versions
129 PAGE_CACHE_SIZE.
132 PAGE_CACHE_SIZE.
135 PAGE_CACHE_SIZE: just make cramfs_readpage read multiple blocks.
[all …]
Dinode.c154 #define BUFFER_SIZE (BLKS_PER_BUF*PAGE_CACHE_SIZE)
176 offset &= PAGE_CACHE_SIZE - 1; in cramfs_read()
231 memcpy(data, kmap(page), PAGE_CACHE_SIZE); in cramfs_read()
235 memset(data, 0, PAGE_CACHE_SIZE); in cramfs_read()
236 data += PAGE_CACHE_SIZE; in cramfs_read()
355 buf->f_bsize = PAGE_CACHE_SIZE; in cramfs_statfs()
498 maxblock = (inode->i_size + PAGE_CACHE_SIZE - 1) >> PAGE_CACHE_SHIFT; in cramfs_readpage()
518 else if (unlikely(compr_len > (PAGE_CACHE_SIZE << 1))) { in cramfs_readpage()
525 PAGE_CACHE_SIZE, in cramfs_readpage()
534 memset(pgdata + bytes_filled, 0, PAGE_CACHE_SIZE - bytes_filled); in cramfs_readpage()
/linux-4.1.27/fs/logfs/
Dfile.c25 if ((len == PAGE_CACHE_SIZE) || PageUptodate(page)) in logfs_write_begin()
28 unsigned start = pos & (PAGE_CACHE_SIZE - 1); in logfs_write_begin()
32 zero_user_segments(page, 0, start, end, PAGE_CACHE_SIZE); in logfs_write_begin()
44 unsigned start = pos & (PAGE_CACHE_SIZE - 1); in logfs_write_end()
48 BUG_ON(PAGE_CACHE_SIZE != inode->i_sb->s_blocksize); in logfs_write_end()
145 offset = i_size & (PAGE_CACHE_SIZE-1); in logfs_writepage()
158 zero_user_segment(page, offset, PAGE_CACHE_SIZE); in logfs_writepage()
Ddev_mtd.c49 BUG_ON(len > PAGE_CACHE_SIZE); in loffs_mtd_write()
/linux-4.1.27/drivers/staging/lustre/lustre/obdclass/
Dcapa.c90 OBD_ALLOC(hash, PAGE_CACHE_SIZE); in init_capa_hash()
94 nr_hash = PAGE_CACHE_SIZE / sizeof(struct hlist_head); in init_capa_hash()
132 OBD_FREE(hash, PAGE_CACHE_SIZE); in cleanup_capa_hash()
283 (unsigned long)(capa) % PAGE_CACHE_SIZE); in capa_hmac()
325 (unsigned long)(d) % PAGE_CACHE_SIZE); in capa_encrypt_id()
329 (unsigned long)(s) % PAGE_CACHE_SIZE); in capa_encrypt_id()
378 (unsigned long)(d) % PAGE_CACHE_SIZE); in capa_decrypt_id()
382 (unsigned long)(s) % PAGE_CACHE_SIZE); in capa_decrypt_id()
Dclass_obd.c500 if ((u64val & ~CFS_PAGE_MASK) >= PAGE_CACHE_SIZE) { in obd_init_checks()
502 (__u64)PAGE_CACHE_SIZE); in obd_init_checks()
/linux-4.1.27/fs/isofs/
Dcompress.c29 static char zisofs_sink_page[PAGE_CACHE_SIZE];
73 memset(page_address(pages[i]), 0, PAGE_CACHE_SIZE); in zisofs_uncompress_block()
124 stream.avail_out = PAGE_CACHE_SIZE - poffset; in zisofs_uncompress_block()
128 stream.avail_out = PAGE_CACHE_SIZE; in zisofs_uncompress_block()
223 end_off = min_t(loff_t, start_off + PAGE_CACHE_SIZE, inode->i_size); in zisofs_fill_pages()
285 PAGE_CACHE_SIZE - poffset); in zisofs_fill_pages()
310 end_index = (inode->i_size + PAGE_CACHE_SIZE - 1) >> PAGE_CACHE_SHIFT; in zisofs_readpage()
/linux-4.1.27/mm/
Dtruncate.c112 do_invalidatepage(page, 0, PAGE_CACHE_SIZE); in truncate_complete_page()
156 PAGE_CACHE_SIZE, 0); in truncate_inode_page()
237 partial_start = lstart & (PAGE_CACHE_SIZE - 1); in truncate_inode_pages_range()
238 partial_end = (lend + 1) & (PAGE_CACHE_SIZE - 1); in truncate_inode_pages_range()
246 start = (lstart + PAGE_CACHE_SIZE - 1) >> PAGE_CACHE_SHIFT; in truncate_inode_pages_range()
294 unsigned int top = PAGE_CACHE_SIZE; in truncate_inode_pages_range()
613 PAGE_CACHE_SIZE, 0); in invalidate_inode_pages2_range()
738 if (from >= to || bsize == PAGE_CACHE_SIZE) in pagecache_isize_extended()
742 if (to <= rounded_from || !(rounded_from & (PAGE_CACHE_SIZE - 1))) in pagecache_isize_extended()
Dreadahead.c51 do_invalidatepage(page, 0, PAGE_CACHE_SIZE); in read_cache_pages_invalidate_page()
104 task_io_account_read(PAGE_CACHE_SIZE); in read_cache_pages()
220 unsigned long this_chunk = (2 * 1024 * 1024) / PAGE_CACHE_SIZE; in force_page_cache_readahead()
235 #define MAX_READAHEAD ((512*4096)/PAGE_CACHE_SIZE)
Dshmem.c76 #define BLOCKS_PER_PAGE (PAGE_CACHE_SIZE/512)
177 security_vm_enough_memory_mm(current->mm, VM_ACCT(PAGE_CACHE_SIZE)) : 0; in shmem_acct_block()
183 vm_unacct_memory(pages * VM_ACCT(PAGE_CACHE_SIZE)); in shmem_unacct_blocks()
399 pgoff_t start = (lstart + PAGE_CACHE_SIZE - 1) >> PAGE_CACHE_SHIFT; in shmem_undo_range()
401 unsigned int partial_start = lstart & (PAGE_CACHE_SIZE - 1); in shmem_undo_range()
402 unsigned int partial_end = (lend + 1) & (PAGE_CACHE_SIZE - 1); in shmem_undo_range()
455 unsigned int top = PAGE_CACHE_SIZE; in shmem_undo_range()
1505 if (copied < PAGE_CACHE_SIZE) { in shmem_write_end()
1506 unsigned from = pos & (PAGE_CACHE_SIZE - 1); in shmem_write_end()
1508 from + copied, PAGE_CACHE_SIZE); in shmem_write_end()
[all …]
Dfilemap.c1476 prev_offset = ra->prev_pos & (PAGE_CACHE_SIZE-1); in do_generic_file_read()
1477 last_index = (*ppos + iter->count + PAGE_CACHE_SIZE-1) >> PAGE_CACHE_SHIFT; in do_generic_file_read()
1534 nr = PAGE_CACHE_SIZE; in do_generic_file_read()
1877 size = round_up(i_size_read(inode), PAGE_CACHE_SIZE); in filemap_fault()
1927 size = round_up(i_size_read(inode), PAGE_CACHE_SIZE); in filemap_fault()
2032 size = round_up(i_size_read(mapping->host), PAGE_CACHE_SIZE); in filemap_map_pages()
2444 offset = (pos & (PAGE_CACHE_SIZE - 1)); in generic_perform_write()
2445 bytes = min_t(unsigned long, PAGE_CACHE_SIZE - offset, in generic_perform_write()
2498 bytes = min_t(unsigned long, PAGE_CACHE_SIZE - offset, in generic_perform_write()
Dfadvise.c127 start_index = (offset+(PAGE_CACHE_SIZE-1)) >> PAGE_CACHE_SHIFT; in SYSCALL_DEFINE4()
Dpage-writeback.c2106 task_io_account_write(PAGE_CACHE_SIZE); in account_page_dirtied()
2127 task_io_account_cancelled_write(PAGE_CACHE_SIZE); in account_page_cleaned()
/linux-4.1.27/drivers/staging/lustre/lnet/selftest/
Dbrw_test.c91 len = npg * PAGE_CACHE_SIZE; in brw_client_init()
103 npg = (len + PAGE_CACHE_SIZE - 1) >> PAGE_CACHE_SHIFT; in brw_client_init()
166 addr += PAGE_CACHE_SIZE - BRW_MSIZE; in brw_fill_page()
172 for (i = 0; i < PAGE_CACHE_SIZE / BRW_MSIZE; i++) in brw_fill_page()
197 addr += PAGE_CACHE_SIZE - BRW_MSIZE; in brw_check_page()
206 for (i = 0; i < PAGE_CACHE_SIZE / BRW_MSIZE; i++) { in brw_check_page()
277 len = npg * PAGE_CACHE_SIZE; in brw_client_prep_rpc()
289 npg = (len + PAGE_CACHE_SIZE - 1) >> PAGE_CACHE_SHIFT; in brw_client_prep_rpc()
466 npg = (reqst->brw_len + PAGE_CACHE_SIZE - 1) >> PAGE_CACHE_SHIFT; in brw_server_handle()
Dconrpc.c792 brq->blk_npg = (param->blk_size + PAGE_CACHE_SIZE - 1) / PAGE_CACHE_SIZE; in lstcon_bulkrpc_v0_prep()
827 npg * PAGE_CACHE_SIZE : in lstcon_testrpc_prep()
854 PAGE_CACHE_SIZE : min_t(int, nob, PAGE_CACHE_SIZE); in lstcon_testrpc_prep()
Dconctl.c755 args->lstio_tes_param_len > PAGE_CACHE_SIZE - sizeof(lstcon_test_t))) in lst_test_add_ioctl()
825 if (data->ioc_plen1 > PAGE_CACHE_SIZE) in lstcon_ioctl_entry()
Dselftest.h392 #define SFW_ID_PER_PAGE (PAGE_CACHE_SIZE / sizeof(lnet_process_id_packed_t))
/linux-4.1.27/fs/nfs/
Dread.c49 zero_user(page, 0, PAGE_CACHE_SIZE); in nfs_return_empty_page()
104 if (len < PAGE_CACHE_SIZE) in nfs_readpage_async()
105 zero_user_segment(page, len, PAGE_CACHE_SIZE); in nfs_readpage_async()
288 page, PAGE_CACHE_SIZE, page_file_index(page)); in nfs_readpage()
354 if (len < PAGE_CACHE_SIZE) in readpage_async_filler()
355 zero_user_segment(page, len, PAGE_CACHE_SIZE); in readpage_async_filler()
415 npages = (pgm->pg_bytes_written + PAGE_CACHE_SIZE - 1) >> in nfs_readpages()
Dfile.c328 unsigned int offset = pos & (PAGE_CACHE_SIZE - 1); in nfs_want_read_modify_write()
404 unsigned offset = pos & (PAGE_CACHE_SIZE - 1); in nfs_write_end()
421 end, PAGE_CACHE_SIZE); in nfs_write_end()
424 zero_user_segment(page, end, PAGE_CACHE_SIZE); in nfs_write_end()
428 zero_user_segment(page, pglen, PAGE_CACHE_SIZE); in nfs_write_end()
462 if (offset != 0 || length < PAGE_CACHE_SIZE) in nfs_invalidate_page()
Dclient.c849 server->rpages = (server->rsize + PAGE_CACHE_SIZE - 1) >> PAGE_CACHE_SHIFT; in nfs_server_set_fsinfo()
858 server->wpages = (server->wsize + PAGE_CACHE_SIZE - 1) >> PAGE_CACHE_SHIFT; in nfs_server_set_fsinfo()
863 if (server->dtsize > PAGE_CACHE_SIZE * NFS_MAX_READDIR_PAGES) in nfs_server_set_fsinfo()
864 server->dtsize = PAGE_CACHE_SIZE * NFS_MAX_READDIR_PAGES; in nfs_server_set_fsinfo()
/linux-4.1.27/fs/gfs2/
Daops.c112 offset = i_size & (PAGE_CACHE_SIZE-1); in gfs2_writepage_common()
114 page->mapping->a_ops->invalidatepage(page, 0, PAGE_CACHE_SIZE); in gfs2_writepage_common()
239 unsigned nrblocks = nr_pages * (PAGE_CACHE_SIZE/inode->i_sb->s_blocksize); in gfs2_write_jdata_pagevec()
461 zero_user(page, 0, PAGE_CACHE_SIZE); in stuffed_readpage()
474 memset(kaddr + dsize, 0, PAGE_CACHE_SIZE - dsize); in stuffed_readpage()
563 unsigned long index = *pos / PAGE_CACHE_SIZE; in gfs2_internal_read()
564 unsigned offset = *pos & (PAGE_CACHE_SIZE - 1); in gfs2_internal_read()
572 if (offset + size > PAGE_CACHE_SIZE) in gfs2_internal_read()
573 amt = PAGE_CACHE_SIZE - offset; in gfs2_internal_read()
651 unsigned from = pos & (PAGE_CACHE_SIZE - 1); in gfs2_write_begin()
[all …]
Dfile.c357 unsigned long size = PAGE_CACHE_SIZE; in gfs2_allocate_page_backing()
408 gfs2_size_hint(vma->vm_file, pos, PAGE_CACHE_SIZE); in gfs2_page_mkwrite()
418 if (!gfs2_write_alloc_required(ip, pos, PAGE_CACHE_SIZE)) { in gfs2_page_mkwrite()
431 gfs2_write_calc_reserv(ip, PAGE_CACHE_SIZE, &data_blocks, &ind_blocks); in gfs2_page_mkwrite()
882 PAGE_CACHE_SIZE/sdp->sd_sb.sb_bsize); in __gfs2_fallocate()
Dquota.c677 unsigned offset = loc & (PAGE_CACHE_SIZE - 1); in gfs2_adjust_quota()
759 if (offset + sizeof(struct gfs2_quota) > PAGE_CACHE_SIZE) in gfs2_adjust_quota()
760 nbytes = PAGE_CACHE_SIZE - offset; in gfs2_adjust_quota()
769 if ((offset + sizeof(struct gfs2_quota)) > PAGE_CACHE_SIZE) { in gfs2_adjust_quota()
Dbmap.c78 memset(kaddr + dsize, 0, PAGE_CACHE_SIZE - dsize); in gfs2_unstuffer_page()
936 unsigned offset = from & (PAGE_CACHE_SIZE-1); in gfs2_block_truncate_page()
/linux-4.1.27/drivers/staging/lustre/lustre/llite/
Drw26.c90 if (offset == 0 && length == PAGE_CACHE_SIZE) { in ll_invalidatepage()
200 *max_pages = (user_addr + size + PAGE_CACHE_SIZE - 1) >> PAGE_CACHE_SHIFT; in ll_get_user_pages()
360 #define MAX_DIO_SIZE ((MAX_MALLOC / sizeof(struct brw_page) * PAGE_CACHE_SIZE) & \
434 size > (PAGE_CACHE_SIZE / sizeof(*pages)) * in ll_direct_IO_26()
435 PAGE_CACHE_SIZE) { in ll_direct_IO_26()
479 unsigned from = pos & (PAGE_CACHE_SIZE - 1); in ll_write_begin()
499 unsigned from = pos & (PAGE_CACHE_SIZE - 1); in ll_write_end()
Dlloop.c222 BUG_ON(bvec.bv_len != PAGE_CACHE_SIZE); in do_bio_lustrebacked()
506 lo->lo_blocksize = PAGE_CACHE_SIZE; in loop_set_fd()
524 CLASSERT(PAGE_CACHE_SIZE < (1 << (sizeof(unsigned short) * 8))); in loop_set_fd()
526 (unsigned short)PAGE_CACHE_SIZE); in loop_set_fd()
Drw.c151 result = cl_io_rw_init(env, io, CIT_WRITE, pos, PAGE_CACHE_SIZE); in ll_cl_init()
1010 kms_pages = (i_size_read(inode) + PAGE_CACHE_SIZE - 1) >> in ras_update()
1192 offset + PAGE_CACHE_SIZE - 1, in ll_writepage()
Dvvp_io.c517 bead->lrr_count = cl_index(obj, tot + PAGE_CACHE_SIZE - 1); in vvp_io_read_start()
958 if (from == 0 && to == PAGE_CACHE_SIZE) { in vvp_io_prepare_write()
1039 to = PAGE_CACHE_SIZE; in vvp_io_commit_write()
Dllite_mmap.c460 unmap_mapping_range(mapping, first + PAGE_CACHE_SIZE - 1, in ll_teardown_mmaps()
Dvvp_page.c173 ll_teardown_mmaps(vmpage->mapping, offset, offset + PAGE_CACHE_SIZE); in vvp_page_unmap()
Dllite_lib.c294 buf = kzalloc(PAGE_CACHE_SIZE, GFP_KERNEL); in client_common_fill_super()
295 obd_connect_flags2str(buf, PAGE_CACHE_SIZE, in client_common_fill_super()
299 OBD_FREE(buf, PAGE_CACHE_SIZE); in client_common_fill_super()
368 sbi->ll_md_brw_size = PAGE_CACHE_SIZE; in client_common_fill_super()
/linux-4.1.27/fs/sysv/
Ddir.c38 return (inode->i_size+PAGE_CACHE_SIZE-1)>>PAGE_CACHE_SHIFT; in dir_pages()
93 limit = kaddr + PAGE_CACHE_SIZE - SYSV_DIRSIZE; in sysv_readdir()
154 kaddr += PAGE_CACHE_SIZE - SYSV_DIRSIZE; in sysv_find_entry()
198 kaddr += PAGE_CACHE_SIZE - SYSV_DIRSIZE; in sysv_add_link()
269 memset(base, 0, PAGE_CACHE_SIZE); in sysv_make_empty()
304 kaddr += PAGE_CACHE_SIZE-SYSV_DIRSIZE; in sysv_empty_dir()
/linux-4.1.27/fs/jfs/
Djfs_metapage.c83 #define MPS_PER_PAGE (PAGE_CACHE_SIZE >> L2PSIZE)
319 for (offset = 0; offset < PAGE_CACHE_SIZE; offset += PSIZE) { in last_write_complete()
374 for (offset = 0; offset < PAGE_CACHE_SIZE; offset += PSIZE) { in metapage_writepage()
419 xlen = (PAGE_CACHE_SIZE - offset) >> inode->i_blkbits; in metapage_writepage()
488 int blocks_per_page = PAGE_CACHE_SIZE >> inode->i_blkbits; in metapage_readpage()
545 for (offset = 0; offset < PAGE_CACHE_SIZE; offset += PSIZE) { in metapage_releasepage()
571 BUG_ON(offset || length < PAGE_CACHE_SIZE); in metapage_invalidatepage()
605 if ((page_offset + size) > PAGE_CACHE_SIZE) { in __get_metapage()
624 if (new && (PSIZE == PAGE_CACHE_SIZE)) { in __get_metapage()
791 for (offset = 0; offset < PAGE_CACHE_SIZE; offset += PSIZE) { in __invalidate_metapages()
/linux-4.1.27/fs/ntfs/
Dcompress.c116 memset(kp + kp_ofs, 0, PAGE_CACHE_SIZE - kp_ofs); in zero_partial_compressed_page()
563 max_page = ((i_size + PAGE_CACHE_SIZE - 1) >> PAGE_CACHE_SHIFT) - in ntfs_read_compressed_block()
569 zero_user(page, 0, PAGE_CACHE_SIZE); in ntfs_read_compressed_block()
765 PAGE_CACHE_SIZE - in ntfs_read_compressed_block()
777 cb_pos += PAGE_CACHE_SIZE - cur_ofs; in ntfs_read_compressed_block()
819 PAGE_CACHE_SIZE - cur_ofs); in ntfs_read_compressed_block()
820 cb_pos += PAGE_CACHE_SIZE - cur_ofs; in ntfs_read_compressed_block()
856 cb_pos2 += PAGE_CACHE_SIZE - cur_ofs2; in ntfs_read_compressed_block()
Daops.c145 recs = PAGE_CACHE_SIZE / rec_size; in ntfs_end_buffer_async_read()
415 if (unlikely(page->index >= (i_size + PAGE_CACHE_SIZE - 1) >> in ntfs_readpage()
417 zero_user(page, 0, PAGE_CACHE_SIZE); in ntfs_readpage()
466 zero_user(page, 0, PAGE_CACHE_SIZE); in ntfs_readpage()
512 memset(addr + attr_len, 0, PAGE_CACHE_SIZE - attr_len); in ntfs_readpage()
928 ntfs_inode *locked_nis[PAGE_CACHE_SIZE / rec_size]; in ntfs_write_mst_block()
952 max_bhs = PAGE_CACHE_SIZE / bh_size; in ntfs_write_mst_block()
964 BUG_ON(!(PAGE_CACHE_SIZE >> rec_size_bits)); in ntfs_write_mst_block()
1303 if (ni->itype.index.block_size == PAGE_CACHE_SIZE) in ntfs_write_mst_block()
1368 if (unlikely(page->index >= (i_size + PAGE_CACHE_SIZE - 1) >> in ntfs_writepage()
[all …]
Dbitmap.c111 len = min_t(s64, cnt >> 3, PAGE_CACHE_SIZE - pos); in __ntfs_bitmap_set_bits_in_run()
135 len = min_t(s64, cnt >> 3, PAGE_CACHE_SIZE); in __ntfs_bitmap_set_bits_in_run()
Ddir.c336 if ((u8*)ia < kaddr || (u8*)ia > kaddr + PAGE_CACHE_SIZE) { in ntfs_lookup_inode_by_name()
369 if (index_end > kaddr + PAGE_CACHE_SIZE) { in ntfs_lookup_inode_by_name()
814 if ((u8*)ia < kaddr || (u8*)ia > kaddr + PAGE_CACHE_SIZE) {
847 if (index_end > kaddr + PAGE_CACHE_SIZE) {
1249 cur_bmp_pos = bmp_pos & ((PAGE_CACHE_SIZE * 8) - 1); in ntfs_readdir()
1250 bmp_pos &= ~(u64)((PAGE_CACHE_SIZE * 8) - 1); in ntfs_readdir()
1255 (unsigned long long)((PAGE_CACHE_SIZE * 8) - 1)); in ntfs_readdir()
1273 if (unlikely((cur_bmp_pos >> 3) >= PAGE_CACHE_SIZE)) { in ntfs_readdir()
1275 bmp_pos += PAGE_CACHE_SIZE * 8; in ntfs_readdir()
1313 if (unlikely((u8*)ia < kaddr || (u8*)ia > kaddr + PAGE_CACHE_SIZE)) { in ntfs_readdir()
[all …]
Dlogfile.c384 size = PAGE_CACHE_SIZE - (pos & ~PAGE_CACHE_MASK); in ntfs_check_and_load_restart_page()
409 size = min_t(int, to_read, PAGE_CACHE_SIZE); in ntfs_check_and_load_restart_page()
512 if (PAGE_CACHE_SIZE >= DefaultLogPageSize && PAGE_CACHE_SIZE <= in ntfs_check_logfile()
516 log_page_size = PAGE_CACHE_SIZE; in ntfs_check_logfile()
Dsuper.c829 if (vol->mft_record_size > PAGE_CACHE_SIZE) { in parse_ntfs_boot_sector()
833 vol->mft_record_size, PAGE_CACHE_SIZE); in parse_ntfs_boot_sector()
1099 mrecs_per_page = PAGE_CACHE_SIZE / vol->mft_record_size; in check_mft_mirror()
1619 size = PAGE_CACHE_SIZE; in load_and_init_attrdef()
1630 if (size == PAGE_CACHE_SIZE) { in load_and_init_attrdef()
1688 size = PAGE_CACHE_SIZE; in load_and_init_upcase()
1699 if (size == PAGE_CACHE_SIZE) { in load_and_init_upcase()
2484 max_index = (((vol->nr_clusters + 7) >> 3) + PAGE_CACHE_SIZE - 1) >> in get_nr_free_clusters()
2488 max_index, PAGE_CACHE_SIZE / 4); in get_nr_free_clusters()
2501 nr_free -= PAGE_CACHE_SIZE * 8; in get_nr_free_clusters()
[all …]
Dfile.c224 end_index = (new_init_size + PAGE_CACHE_SIZE - 1) >> PAGE_CACHE_SHIFT; in ntfs_attr_extend_initialized()
944 if (unlikely(vol->cluster_size < PAGE_CACHE_SIZE)) { in ntfs_prepare_pages_for_non_resident_write()
1641 memset(kaddr + attr_len, 0, PAGE_CACHE_SIZE - attr_len); in ntfs_commit_pages_after_write()
1708 len = PAGE_CACHE_SIZE - ofs; in ntfs_copy_from_user_iter()
1726 len = PAGE_CACHE_SIZE - copied; in ntfs_copy_from_user_iter()
1733 len = PAGE_CACHE_SIZE; in ntfs_copy_from_user_iter()
1789 if (vol->cluster_size > PAGE_CACHE_SIZE && NInoNonResident(ni)) in ntfs_perform_write()
1800 bytes = PAGE_CACHE_SIZE - ofs; in ntfs_perform_write()
Dntfs.h46 NTFS_MAX_PAGES_PER_CLUSTER = NTFS_MAX_CLUSTER_SIZE / PAGE_CACHE_SIZE,
Dindex.c293 if ((u8*)ia < kaddr || (u8*)ia > kaddr + PAGE_CACHE_SIZE) { in ntfs_index_lookup()
326 if (index_end > kaddr + PAGE_CACHE_SIZE) { in ntfs_index_lookup()
Dinode.c871 if (ni->itype.index.block_size > PAGE_CACHE_SIZE) { in ntfs_read_locked_inode()
876 PAGE_CACHE_SIZE); in ntfs_read_locked_inode()
1588 if (ni->itype.index.block_size > PAGE_CACHE_SIZE) { in ntfs_read_locked_index_inode()
1591 ni->itype.index.block_size, PAGE_CACHE_SIZE); in ntfs_read_locked_index_inode()
Dattrib.c1663 memset(kaddr + attr_size, 0, PAGE_CACHE_SIZE - attr_size); in ntfs_attr_make_non_resident()
2539 size = PAGE_CACHE_SIZE; in ntfs_attr_set()
2564 memset(kaddr, val, PAGE_CACHE_SIZE); in ntfs_attr_set()
Dlcnalloc.c294 buf_size = PAGE_CACHE_SIZE - buf_size; in ntfs_cluster_alloc()
/linux-4.1.27/fs/9p/
Dvfs_addr.c156 if (offset == 0 && length == PAGE_CACHE_SIZE) in v9fs_invalidate_page()
172 len = PAGE_CACHE_SIZE; in v9fs_vfs_writepage_locked()
291 if (len == PAGE_CACHE_SIZE) in v9fs_write_begin()
316 unsigned from = pos & (PAGE_CACHE_SIZE - 1); in v9fs_write_end()
Dvfs_super.c90 sb->s_bdi->ra_pages = (VM_MAX_READAHEAD * 1024)/PAGE_CACHE_SIZE; in v9fs_fill_super()
/linux-4.1.27/fs/ext2/
Ddir.c40 #if (PAGE_CACHE_SIZE >= 65536) in ext2_rec_len_from_disk()
49 #if (PAGE_CACHE_SIZE >= 65536) in ext2_rec_len_to_disk()
75 return (inode->i_size+PAGE_CACHE_SIZE-1)>>PAGE_CACHE_SHIFT; in dir_pages()
88 if (last_byte > PAGE_CACHE_SIZE) in ext2_last_byte()
89 last_byte = PAGE_CACHE_SIZE; in ext2_last_byte()
126 unsigned limit = PAGE_CACHE_SIZE; in ext2_check_page()
317 ctx->pos += PAGE_CACHE_SIZE - offset; in ext2_readdir()
519 kaddr += PAGE_CACHE_SIZE - reclen; in ext2_add_link()
/linux-4.1.27/fs/f2fs/
Dinline.c42 zero_user_segment(page, MAX_INLINE_DATA, PAGE_CACHE_SIZE); in read_inline_data()
84 zero_user_segment(page, 0, PAGE_CACHE_SIZE); in f2fs_read_inline_data()
117 zero_user_segment(page, MAX_INLINE_DATA, PAGE_CACHE_SIZE); in f2fs_convert_inline_page()
359 zero_user_segment(page, 0, PAGE_CACHE_SIZE); in f2fs_convert_inline_dir()
381 if (i_size_read(dir) < PAGE_CACHE_SIZE) { in f2fs_convert_inline_dir()
382 i_size_write(dir, PAGE_CACHE_SIZE); in f2fs_convert_inline_dir()
Dfile.c78 zero_user_segment(page, offset, PAGE_CACHE_SIZE); in f2fs_vm_page_mkwrite()
466 unsigned offset = from & (PAGE_CACHE_SIZE - 1); in truncate_partial_data_page()
482 zero_user(page, offset, PAGE_CACHE_SIZE - offset); in truncate_partial_data_page()
734 off_start = offset & (PAGE_CACHE_SIZE - 1); in punch_hole()
735 off_end = (offset + len) & (PAGE_CACHE_SIZE - 1); in punch_hole()
743 PAGE_CACHE_SIZE - off_start); in punch_hole()
792 off_start = offset & (PAGE_CACHE_SIZE - 1); in expand_inode_data()
793 off_end = (offset + len) & (PAGE_CACHE_SIZE - 1); in expand_inode_data()
815 new_size += PAGE_CACHE_SIZE; in expand_inode_data()
Ddata.c147 if (bio_add_page(bio, page, PAGE_CACHE_SIZE, 0) < PAGE_CACHE_SIZE) { in f2fs_submit_page_bio()
184 if (bio_add_page(io->bio, page, PAGE_CACHE_SIZE, 0) < in f2fs_submit_page_mbio()
185 PAGE_CACHE_SIZE) { in f2fs_submit_page_mbio()
1026 zero_user_segment(page, 0, PAGE_CACHE_SIZE); in get_lock_data_page()
1079 zero_user_segment(page, 0, PAGE_CACHE_SIZE); in get_new_data_page()
1441 offset = i_size & (PAGE_CACHE_SIZE - 1); in f2fs_write_data_page()
1445 zero_user_segment(page, offset, PAGE_CACHE_SIZE); in f2fs_write_data_page()
1630 if ((len == PAGE_CACHE_SIZE) || PageUptodate(page)) in f2fs_write_begin()
1636 unsigned start = pos & (PAGE_CACHE_SIZE - 1); in f2fs_write_begin()
1640 zero_user_segments(page, 0, start, end, PAGE_CACHE_SIZE); in f2fs_write_begin()
[all …]
Dsegment.c754 sum_in_page = (PAGE_CACHE_SIZE - 2 * SUM_JOURNAL_SIZE - in npages_for_summary_flush()
759 (PAGE_CACHE_SIZE - SUM_FOOTER_SIZE) / SUMMARY_SIZE) in npages_for_summary_flush()
777 memcpy(kaddr, sum_blk, PAGE_CACHE_SIZE); in write_sum_page()
1385 if (offset + SUMMARY_SIZE <= PAGE_CACHE_SIZE - in read_compacted_summaries()
1457 memcpy(curseg->sum_blk, sum, PAGE_CACHE_SIZE); in read_normal_summaries()
1540 if (written_size + SUMMARY_SIZE <= PAGE_CACHE_SIZE - in write_compacted_summaries()
1631 memcpy(dst_addr, src_addr, PAGE_CACHE_SIZE); in get_next_sit_page()
1950 array[i].sum_blk = kzalloc(PAGE_CACHE_SIZE, GFP_KERNEL); in build_curseg()
Ddebug.c159 si->base_mem += PAGE_CACHE_SIZE * NR_CURSEG_TYPE; in update_mem_info()
Dsuper.c835 if (F2FS_BLKSIZE != PAGE_CACHE_SIZE) { in sanity_check_raw_super()
838 PAGE_CACHE_SIZE); in sanity_check_raw_super()
/linux-4.1.27/drivers/staging/lustre/lustre/include/
Dlustre_mdc.h156 min_t(__u32, body->max_mdsize, PAGE_CACHE_SIZE); in mdc_update_max_ea_from_body()
161 min_t(__u32, body->max_cookiesize, PAGE_CACHE_SIZE); in mdc_update_max_ea_from_body()
Dlustre_disk.h274 #if (128 * 1024UL) > (PAGE_CACHE_SIZE * 8)
277 #define LR_MAX_CLIENTS (PAGE_CACHE_SIZE * 8)
Dobd_support.h686 #define OBD_ALLOC_BIG (4 * PAGE_CACHE_SIZE)
722 #define POISON_PAGE(page, val) do { memset(kmap(page), val, PAGE_CACHE_SIZE); \
Dlu_object.h1128 CLASSERT(PAGE_CACHE_SIZE >= sizeof (*value)); \
/linux-4.1.27/fs/ufs/
Ddir.c70 return (inode->i_size+PAGE_CACHE_SIZE-1)>>PAGE_CACHE_SHIFT; in ufs_dir_pages()
117 unsigned limit = PAGE_CACHE_SIZE; in ufs_check_page()
218 if (last_byte > PAGE_CACHE_SIZE) in ufs_last_byte()
219 last_byte = PAGE_CACHE_SIZE; in ufs_last_byte()
347 kaddr += PAGE_CACHE_SIZE - reclen; in ufs_add_link()
460 ctx->pos += PAGE_CACHE_SIZE - offset; in ufs_readdir()
580 memset(base, 0, PAGE_CACHE_SIZE); in ufs_make_empty()
/linux-4.1.27/fs/udf/
Dfile.c49 memset(kaddr + inode->i_size, 0, PAGE_CACHE_SIZE - inode->i_size); in __udf_adinicb_readpage()
90 if (WARN_ON_ONCE(pos >= PAGE_CACHE_SIZE)) in udf_adinicb_write_begin()
97 if (!PageUptodate(page) && len != PAGE_CACHE_SIZE) in udf_adinicb_write_begin()
/linux-4.1.27/fs/ext4/
Dreadpage.c143 const unsigned blocks_per_page = PAGE_CACHE_SIZE >> blkbits; in ext4_mpage_readpages()
220 PAGE_CACHE_SIZE); in ext4_mpage_readpages()
253 PAGE_CACHE_SIZE); in ext4_mpage_readpages()
Dcrypto.c362 sg_set_page(&dst, dest_page, PAGE_CACHE_SIZE, 0); in ext4_page_crypto()
364 sg_set_page(&src, src_page, PAGE_CACHE_SIZE, 0); in ext4_page_crypto()
365 ablkcipher_request_set_crypt(req, &src, &dst, PAGE_CACHE_SIZE, in ext4_page_crypto()
490 BUG_ON(inode->i_sb->s_blocksize != PAGE_CACHE_SIZE); in ext4_encrypted_zeroout()
Dinode.c916 unsigned from = pos & (PAGE_CACHE_SIZE - 1); in ext4_block_write_begin()
928 BUG_ON(from > PAGE_CACHE_SIZE); in ext4_block_write_begin()
929 BUG_ON(to > PAGE_CACHE_SIZE); in ext4_block_write_begin()
1019 from = pos & (PAGE_CACHE_SIZE - 1); in ext4_write_begin()
1225 from = pos & (PAGE_CACHE_SIZE - 1); in ext4_journalled_write_end()
1375 BUG_ON(stop > PAGE_CACHE_SIZE || stop < length); in ext4_da_page_release_reservation()
1474 block_invalidatepage(page, 0, PAGE_CACHE_SIZE); in mpage_release_unused_pages()
1848 len = PAGE_CACHE_SIZE; in ext4_writepage()
1906 len = PAGE_CACHE_SIZE; in mpage_submit_page()
2101 mpd->io_submit.io_end->size += PAGE_CACHE_SIZE; in mpage_map_and_submit_buffers()
[all …]
Dpage-io.c451 if (len < PAGE_CACHE_SIZE) in ext4_bio_write_page()
452 zero_user_segment(page, len, PAGE_CACHE_SIZE); in ext4_bio_write_page()
Dmove_extent.c274 int blocks_per_page = PAGE_CACHE_SIZE >> orig_inode->i_blkbits; in move_extent_per_page()
567 int blocks_per_page = PAGE_CACHE_SIZE >> orig_inode->i_blkbits; in ext4_move_extents()
/linux-4.1.27/fs/ubifs/
Dfile.c125 memset(addr, 0, PAGE_CACHE_SIZE); in do_readpage()
258 if (!(pos & ~PAGE_CACHE_MASK) && len == PAGE_CACHE_SIZE) in write_begin_slow()
450 if (!(pos & ~PAGE_CACHE_MASK) && len == PAGE_CACHE_SIZE) { in ubifs_write_begin()
553 if (unlikely(copied < len && len == PAGE_CACHE_SIZE)) { in ubifs_write_end()
628 memset(addr, 0, PAGE_CACHE_SIZE); in populate_page()
677 int len = i_size & (PAGE_CACHE_SIZE - 1); in populate_page()
1006 int err, len = i_size & (PAGE_CACHE_SIZE - 1); in ubifs_writepage()
1038 return do_writepage(page, PAGE_CACHE_SIZE); in ubifs_writepage()
1049 memset(kaddr + len, 0, PAGE_CACHE_SIZE - len); in ubifs_writepage()
1161 (PAGE_CACHE_SIZE - 1); in do_truncation()
[all …]
/linux-4.1.27/fs/nilfs2/
Ddir.c66 return (inode->i_size+PAGE_CACHE_SIZE-1)>>PAGE_CACHE_SHIFT; in dir_pages()
78 if (last_byte > PAGE_CACHE_SIZE) in nilfs_last_byte()
79 last_byte = PAGE_CACHE_SIZE; in nilfs_last_byte()
117 unsigned limit = PAGE_CACHE_SIZE; in nilfs_check_page()
280 ctx->pos += PAGE_CACHE_SIZE - offset; in nilfs_readdir()
468 kaddr += PAGE_CACHE_SIZE - reclen; in nilfs_add_link()
/linux-4.1.27/drivers/staging/lustre/lustre/mgc/
Dmgc_request.c1132 OBD_ALLOC(inst, PAGE_CACHE_SIZE); in mgc_apply_recover_logs()
1137 pos = snprintf(inst, PAGE_CACHE_SIZE, "%p", cfg->cfg_instance); in mgc_apply_recover_logs()
1138 if (pos >= PAGE_CACHE_SIZE) { in mgc_apply_recover_logs()
1139 OBD_FREE(inst, PAGE_CACHE_SIZE); in mgc_apply_recover_logs()
1145 PAGE_CACHE_SIZE); in mgc_apply_recover_logs()
1147 OBD_FREE(inst, PAGE_CACHE_SIZE); in mgc_apply_recover_logs()
1155 bufsz = PAGE_CACHE_SIZE - pos; in mgc_apply_recover_logs()
1187 if (entry->mne_length > PAGE_CACHE_SIZE) { in mgc_apply_recover_logs()
1305 OBD_FREE(inst, PAGE_CACHE_SIZE); in mgc_apply_recover_logs()
1390 ptlrpc_prep_bulk_page_pin(desc, pages[i], 0, PAGE_CACHE_SIZE); in mgc_process_recover_log()
[all …]
/linux-4.1.27/block/
Dblk-settings.c254 if ((max_hw_sectors << 9) < PAGE_CACHE_SIZE) { in blk_limits_max_hw_sectors()
354 if (max_size < PAGE_CACHE_SIZE) { in blk_queue_max_segment_size()
355 max_size = PAGE_CACHE_SIZE; in blk_queue_max_segment_size()
780 if (mask < PAGE_CACHE_SIZE - 1) { in blk_queue_segment_boundary()
781 mask = PAGE_CACHE_SIZE - 1; in blk_queue_segment_boundary()
Dioctl.c359 return put_long(arg, (bdi->ra_pages * PAGE_CACHE_SIZE) / 512); in blkdev_ioctl()
387 bdi->ra_pages = (arg * 512) / PAGE_CACHE_SIZE; in blkdev_ioctl()
Dcompat_ioctl.c713 (bdi->ra_pages * PAGE_CACHE_SIZE) / 512); in compat_blkdev_ioctl()
732 bdi->ra_pages = (arg * 512) / PAGE_CACHE_SIZE; in compat_blkdev_ioctl()
Dblk-sysfs.c119 return queue_var_show(PAGE_CACHE_SIZE, (page)); in queue_max_segment_size_show()
/linux-4.1.27/fs/
Dmpage.c146 const unsigned blocks_per_page = PAGE_CACHE_SIZE >> blkbits; in do_mpage_readpage()
250 zero_user_segment(page, first_hole << blkbits, PAGE_CACHE_SIZE); in do_mpage_readpage()
471 const unsigned blocks_per_page = PAGE_CACHE_SIZE >> blkbits; in __mpage_writepage()
582 unsigned offset = i_size & (PAGE_CACHE_SIZE - 1); in __mpage_writepage()
586 zero_user_segment(page, offset, PAGE_CACHE_SIZE); in __mpage_writepage()
Dbuffer.c1523 BUG_ON(stop > PAGE_CACHE_SIZE || stop < length); in block_invalidatepage()
1885 unsigned from = pos & (PAGE_CACHE_SIZE - 1); in __block_write_begin()
1895 BUG_ON(from > PAGE_CACHE_SIZE); in __block_write_begin()
1896 BUG_ON(to > PAGE_CACHE_SIZE); in __block_write_begin()
2038 start = pos & (PAGE_CACHE_SIZE - 1); in block_write_end()
2127 to = min_t(unsigned, PAGE_CACHE_SIZE - from, count); in block_is_partially_uptodate()
2129 if (from < blocksize && to > PAGE_CACHE_SIZE - blocksize) in block_is_partially_uptodate()
2295 len = PAGE_CACHE_SIZE - zerofrom; in cont_expand_zero()
2424 end = PAGE_CACHE_SIZE; in __block_page_mkwrite()
2520 from = pos & (PAGE_CACHE_SIZE - 1); in nobh_write_begin()
[all …]
Ddax.c411 PAGE_CACHE_SIZE, 0); in do_dax_fault()
514 unsigned offset = from & (PAGE_CACHE_SIZE-1); in dax_zero_page_range()
520 BUG_ON((offset + length) > PAGE_CACHE_SIZE); in dax_zero_page_range()
523 bh.b_size = PAGE_CACHE_SIZE; in dax_zero_page_range()
Dlibfs.c41 buf->f_bsize = PAGE_CACHE_SIZE; in simple_statfs()
411 if (!PageUptodate(page) && (len != PAGE_CACHE_SIZE)) { in simple_write_begin()
412 unsigned from = pos & (PAGE_CACHE_SIZE - 1); in simple_write_begin()
414 zero_user_segments(page, 0, from, from + len, PAGE_CACHE_SIZE); in simple_write_begin()
450 unsigned from = pos & (PAGE_CACHE_SIZE - 1); in simple_write_end()
485 s->s_blocksize = PAGE_CACHE_SIZE; in simple_fill_super()
Dsplice.c332 req_pages = (len + loff + PAGE_CACHE_SIZE - 1) >> PAGE_CACHE_SHIFT; in __generic_file_splice_read()
399 this_len = min_t(unsigned long, len, PAGE_CACHE_SIZE - loff); in __generic_file_splice_read()
641 nr_pages = (len + offset + PAGE_CACHE_SIZE - 1) >> PAGE_CACHE_SHIFT; in default_file_splice_read()
651 this_len = min_t(size_t, len, PAGE_CACHE_SIZE - offset); in default_file_splice_read()
/linux-4.1.27/fs/cifs/
Dfile.c1852 if ((to > PAGE_CACHE_SIZE) || (from > to)) { in cifs_partialpagewrite()
2012 wdata->pagesz = PAGE_CACHE_SIZE; in wdata_send_pages()
2015 (loff_t)PAGE_CACHE_SIZE); in wdata_send_pages()
2016 wdata->bytes = ((nr_pages - 1) * PAGE_CACHE_SIZE) + wdata->tailsz; in wdata_send_pages()
2050 if (cifs_sb->wsize < PAGE_CACHE_SIZE) in cifs_writepages()
2074 tofind = min((wsize / PAGE_CACHE_SIZE) - 1, end - index) + 1; in cifs_writepages()
2173 rc = cifs_partialpagewrite(page, 0, PAGE_CACHE_SIZE); in cifs_writepage_locked()
2217 } else if (!PageUptodate(page) && copied == PAGE_CACHE_SIZE) in cifs_write_end()
2222 unsigned offset = pos & (PAGE_CACHE_SIZE - 1); in cifs_write_end()
3289 got_bytes -= min_t(unsigned int, PAGE_CACHE_SIZE, got_bytes); in cifs_readv_complete()
[all …]
/linux-4.1.27/fs/qnx6/
Ddir.c37 return (inode->i_size+PAGE_CACHE_SIZE-1)>>PAGE_CACHE_SHIFT; in dir_pages()
44 if (last_byte > PAGE_CACHE_SIZE) in last_entry()
45 last_byte = PAGE_CACHE_SIZE; in last_entry()
/linux-4.1.27/fs/minix/
Ddir.c41 unsigned last_byte = PAGE_CACHE_SIZE; in minix_last_byte()
44 last_byte = inode->i_size & (PAGE_CACHE_SIZE - 1); in minix_last_byte()
50 return (inode->i_size+PAGE_CACHE_SIZE-1)>>PAGE_CACHE_SHIFT; in dir_pages()
237 limit = kaddr + PAGE_CACHE_SIZE - sbi->s_dirsize; in minix_add_link()
335 memset(kaddr, 0, PAGE_CACHE_SIZE); in minix_make_empty()
/linux-4.1.27/fs/exofs/
Ddir.c50 return (inode->i_size + PAGE_CACHE_SIZE - 1) >> PAGE_CACHE_SHIFT; in dir_pages()
58 if (last_byte > PAGE_CACHE_SIZE) in exofs_last_byte()
59 last_byte = PAGE_CACHE_SIZE; in exofs_last_byte()
94 unsigned limit = PAGE_CACHE_SIZE; in exofs_check_page()
263 ctx->pos += PAGE_CACHE_SIZE - offset; in exofs_readdir()
458 kaddr += PAGE_CACHE_SIZE - reclen; in exofs_add_link()
Dinode.c400 len = PAGE_CACHE_SIZE; in readpage_strip()
445 if (len != PAGE_CACHE_SIZE) in readpage_strip()
446 zero_user(page, len, PAGE_CACHE_SIZE - len); in readpage_strip()
714 len = PAGE_CACHE_SIZE; in writepage_strip()
887 if (!PageUptodate(page) && (len != PAGE_CACHE_SIZE)) { in exofs_write_begin()
893 rlen = PAGE_CACHE_SIZE; in exofs_write_begin()
/linux-4.1.27/drivers/staging/lustre/lustre/libcfs/
Dmodule.c125 while (count1 < PAGE_CACHE_SIZE/sizeof(struct page *) && in kportal_memhog_free()
131 while (count2 < PAGE_CACHE_SIZE/sizeof(struct page *) && in kportal_memhog_free()
181 memset(level1p, 0, PAGE_CACHE_SIZE); in kportal_memhog_alloc()
184 count1 < PAGE_CACHE_SIZE/sizeof(struct page *)) { in kportal_memhog_alloc()
196 memset(level2p, 0, PAGE_CACHE_SIZE); in kportal_memhog_alloc()
199 count2 < PAGE_CACHE_SIZE/sizeof(struct page *)) { in kportal_memhog_alloc()
Dtracefile.c147 if (tage->used + len <= PAGE_CACHE_SIZE) in cfs_trace_get_tage_try()
226 if (len > PAGE_CACHE_SIZE) { in cfs_trace_get_tage()
314 if (needed + known_size > PAGE_CACHE_SIZE) in libcfs_debug_vmsg2()
325 max_nob = PAGE_CACHE_SIZE - tage->used - known_size; in libcfs_debug_vmsg2()
389 __LASSERT (tage->used <= PAGE_CACHE_SIZE); in libcfs_debug_vmsg2()
810 if (nob > 2 * PAGE_CACHE_SIZE) /* string must be "sensible" */ in cfs_trace_allocate_string_buffer()
Dtracefile.h334 __LASSERT(tage->used <= PAGE_CACHE_SIZE); \
/linux-4.1.27/drivers/staging/lustre/lustre/include/linux/
Dlustre_patchless_compat.h56 page->mapping->a_ops->invalidatepage(page, 0, PAGE_CACHE_SIZE); in truncate_complete_page()
/linux-4.1.27/drivers/staging/lustre/include/linux/libcfs/linux/
Dlinux-mem.h60 #define CFS_PAGE_MASK (~((__u64)PAGE_CACHE_SIZE-1))
/linux-4.1.27/fs/reiserfs/
Dxattr.c529 size_t page_offset = (file_pos & (PAGE_CACHE_SIZE - 1)); in reiserfs_xattr_set_handle()
531 if (buffer_size - buffer_pos > PAGE_CACHE_SIZE) in reiserfs_xattr_set_handle()
532 chunk = PAGE_CACHE_SIZE; in reiserfs_xattr_set_handle()
549 if (chunk + skip > PAGE_CACHE_SIZE) in reiserfs_xattr_set_handle()
550 chunk = PAGE_CACHE_SIZE - skip; in reiserfs_xattr_set_handle()
678 if (isize - file_pos > PAGE_CACHE_SIZE) in reiserfs_xattr_get()
679 chunk = PAGE_CACHE_SIZE; in reiserfs_xattr_get()
Dtail_conversion.c154 (tail_offset + total_tail - 1) & (PAGE_CACHE_SIZE - 1); in direct2indirect()
274 tail = tail + (pos & (PAGE_CACHE_SIZE - 1)); in indirect2direct()
Dinode.c389 offset = (cpu_key_k_offset(&key) - 1) & (PAGE_CACHE_SIZE - 1); in _get_block_create_0()
590 tail_start = tail_offset & (PAGE_CACHE_SIZE - 1); in convert_tail_for_hole()
2195 unsigned long offset = (inode->i_size) & (PAGE_CACHE_SIZE - 1); in grab_tail_page()
2267 unsigned long offset = inode->i_size & (PAGE_CACHE_SIZE - 1); in reiserfs_truncate_file()
2428 p += (byte_offset - 1) & (PAGE_CACHE_SIZE - 1); in map_block_for_writepage()
2537 int bh_per_page = PAGE_CACHE_SIZE / s->s_blocksize; in reiserfs_write_full_page()
2566 last_offset = inode->i_size & (PAGE_CACHE_SIZE - 1); in reiserfs_write_full_page()
2572 zero_user_segment(page, last_offset, PAGE_CACHE_SIZE); in reiserfs_write_full_page()
2911 start = pos & (PAGE_CACHE_SIZE - 1); in reiserfs_write_end()
3183 int partial_page = (offset || length < PAGE_CACHE_SIZE); in reiserfs_invalidatepage()
Dfile.c187 int bh_per_page = PAGE_CACHE_SIZE / s->s_blocksize; in reiserfs_commit_page()
/linux-4.1.27/net/sunrpc/
Dxdr.c193 pgto_base = PAGE_CACHE_SIZE; in _shift_data_right_pages()
197 pgfrom_base = PAGE_CACHE_SIZE; in _shift_data_right_pages()
243 copy = PAGE_CACHE_SIZE - pgbase; in _copy_to_pages()
256 if (pgbase == PAGE_CACHE_SIZE) { in _copy_to_pages()
287 copy = PAGE_CACHE_SIZE - pgbase; in _copy_from_pages()
296 if (pgbase == PAGE_CACHE_SIZE) { in _copy_from_pages()
1302 avail_page = min_t(unsigned int, PAGE_CACHE_SIZE - base, in xdr_xcode_array2()
1386 (unsigned int) PAGE_CACHE_SIZE); in xdr_xcode_array2()
1482 page_offset = (offset + buf->page_base) & (PAGE_CACHE_SIZE - 1); in xdr_process_buf()
1484 thislen = PAGE_CACHE_SIZE - page_offset; in xdr_process_buf()
[all …]
Dsocklib.c116 len = PAGE_CACHE_SIZE; in xdr_partial_copy_from_skb()
/linux-4.1.27/fs/ocfs2/
Daops.c237 if (size > PAGE_CACHE_SIZE || in ocfs2_read_inline_data()
250 memset(kaddr + size, 0, PAGE_CACHE_SIZE - size); in ocfs2_read_inline_data()
984 unsigned int cluster_start = 0, cluster_end = PAGE_CACHE_SIZE; in ocfs2_figure_cluster_boundaries()
1157 #if (PAGE_CACHE_SIZE >= OCFS2_MAX_CLUSTERSIZE)
1160 #define OCFS2_MAX_CTXT_PAGES (OCFS2_MAX_CLUSTERSIZE / PAGE_CACHE_SIZE)
1163 #define OCFS2_MAX_CLUSTERS_PER_PAGE (PAGE_CACHE_SIZE / OCFS2_MIN_CLUSTERSIZE)
1361 unsigned from = user_pos & (PAGE_CACHE_SIZE - 1), in ocfs2_write_failure()
1400 map_from = user_pos & (PAGE_CACHE_SIZE - 1); in ocfs2_prepare_page_for_write()
1700 wc->w_target_from = pos & (PAGE_CACHE_SIZE - 1); in ocfs2_set_target_boundaries()
1737 wc->w_target_to = PAGE_CACHE_SIZE; in ocfs2_set_target_boundaries()
[all …]
Dmmap.c68 unsigned int len = PAGE_CACHE_SIZE; in __ocfs2_page_mkwrite()
Dfile.c793 zero_from = abs_from & (PAGE_CACHE_SIZE - 1); in ocfs2_write_zero_page()
794 zero_to = abs_to & (PAGE_CACHE_SIZE - 1); in ocfs2_write_zero_page()
796 zero_to = PAGE_CACHE_SIZE; in ocfs2_write_zero_page()
958 next_pos = (zero_pos & PAGE_CACHE_MASK) + PAGE_CACHE_SIZE; in ocfs2_zero_extend_range()
Dalloc.c6634 unsigned int from, to = PAGE_CACHE_SIZE; in ocfs2_zero_cluster_pages()
6642 to = PAGE_CACHE_SIZE; in ocfs2_zero_cluster_pages()
6646 from = start & (PAGE_CACHE_SIZE - 1); in ocfs2_zero_cluster_pages()
6648 to = end & (PAGE_CACHE_SIZE - 1); in ocfs2_zero_cluster_pages()
6650 BUG_ON(from > PAGE_CACHE_SIZE); in ocfs2_zero_cluster_pages()
6651 BUG_ON(to > PAGE_CACHE_SIZE); in ocfs2_zero_cluster_pages()
6913 PAGE_CACHE_SIZE < osb->s_clustersize) in ocfs2_convert_inline_data_to_extents()
6914 end = PAGE_CACHE_SIZE; in ocfs2_convert_inline_data_to_extents()
6934 page_end = PAGE_CACHE_SIZE; in ocfs2_convert_inline_data_to_extents()
6935 if (PAGE_CACHE_SIZE > osb->s_clustersize) in ocfs2_convert_inline_data_to_extents()
/linux-4.1.27/fs/freevxfs/
Dvxfs_immed.c107 memcpy(kaddr, vip->vii_immed.vi_immed + offset, PAGE_CACHE_SIZE); in vxfs_immed_readpage()
Dvxfs_lookup.c48 #define VXFS_BLOCK_PER_PAGE(sbp) ((PAGE_CACHE_SIZE / (sbp)->s_blocksize))
68 return (inode->i_size + PAGE_CACHE_SIZE - 1) >> PAGE_CACHE_SHIFT; in dir_pages()
/linux-4.1.27/fs/jffs2/
Dfile.c97 ret = jffs2_read_inode_range(c, f, pg_buf, pg->index << PAGE_CACHE_SHIFT, PAGE_CACHE_SIZE); in jffs2_do_readpage_nolock()
248 unsigned start = pos & (PAGE_CACHE_SIZE - 1); in jffs2_write_end()
264 if (end == PAGE_CACHE_SIZE) { in jffs2_write_end()
Dwrite.c175 if ((je32_to_cpu(ri->dsize) >= PAGE_CACHE_SIZE) || in jffs2_write_dnode()
176 ( ((je32_to_cpu(ri->offset)&(PAGE_CACHE_SIZE-1))==0) && in jffs2_write_dnode()
369 datalen = min_t(uint32_t, writelen, PAGE_CACHE_SIZE - (offset & (PAGE_CACHE_SIZE-1))); in jffs2_write_inode_range()
Ddebug.c98 if (frag->ofs & (PAGE_CACHE_SIZE-1) && frag_prev(frag) in __jffs2_dbg_fragtree_paranoia_check_nolock()
99 && frag_prev(frag)->size < PAGE_CACHE_SIZE && frag_prev(frag)->node) { in __jffs2_dbg_fragtree_paranoia_check_nolock()
105 if ((frag->ofs+frag->size) & (PAGE_CACHE_SIZE-1) && frag_next(frag) in __jffs2_dbg_fragtree_paranoia_check_nolock()
106 && frag_next(frag)->size < PAGE_CACHE_SIZE && frag_next(frag)->node) { in __jffs2_dbg_fragtree_paranoia_check_nolock()
Dnodelist.c93 if (frag->node && (frag->ofs & (PAGE_CACHE_SIZE - 1)) == 0) { in jffs2_truncate_fragtree()
385 if (newfrag->ofs & (PAGE_CACHE_SIZE-1)) { in jffs2_add_full_dnode_to_inode()
394 if ((newfrag->ofs+newfrag->size) & (PAGE_CACHE_SIZE-1)) { in jffs2_add_full_dnode_to_inode()
Dgc.c1175 min = start & ~(PAGE_CACHE_SIZE-1); in jffs2_garbage_collect_dnode()
1176 max = min + PAGE_CACHE_SIZE; in jffs2_garbage_collect_dnode()
1334 writebuf = pg_ptr + (offset & (PAGE_CACHE_SIZE -1)); in jffs2_garbage_collect_dnode()
/linux-4.1.27/fs/btrfs/tests/
Dextent-io-tests.c248 if (start != test_start && end != test_start + PAGE_CACHE_SIZE - 1) { in test_find_delalloc()
250 test_start, test_start + PAGE_CACHE_SIZE - 1, start, in test_find_delalloc()
Dfree-space-tests.c24 #define BITS_PER_BITMAP (PAGE_CACHE_SIZE * 8)
/linux-4.1.27/fs/nfs/blocklayout/
Dblocklayout.c268 if (pg_offset + bytes_left > PAGE_CACHE_SIZE) in bl_read_pagelist()
269 pg_len = PAGE_CACHE_SIZE - pg_offset; in bl_read_pagelist()
274 pg_len = PAGE_CACHE_SIZE; in bl_read_pagelist()
349 PAGE_CACHE_SIZE - 1) & (loff_t)PAGE_CACHE_MASK; in bl_write_cleanup()
416 pg_len = PAGE_CACHE_SIZE; in bl_write_pagelist()
814 end = DIV_ROUND_UP(i_size_read(inode), PAGE_CACHE_SIZE); in pnfs_num_cont_bytes()
Dblocklayout.h43 #define PAGE_CACHE_SECTORS (PAGE_CACHE_SIZE >> SECTOR_SHIFT)
/linux-4.1.27/fs/xfs/
Dxfs_aops.c747 p_offset = min_t(unsigned long, end_offset & (PAGE_CACHE_SIZE - 1), in xfs_convert_page()
748 PAGE_CACHE_SIZE); in xfs_convert_page()
749 p_offset = p_offset ? roundup(p_offset, len) : PAGE_CACHE_SIZE; in xfs_convert_page()
925 xfs_vm_invalidatepage(page, 0, PAGE_CACHE_SIZE); in xfs_aops_discard_page()
1010 unsigned offset_into_page = offset & (PAGE_CACHE_SIZE - 1); in xfs_vm_writepage()
1041 zero_user_segment(page, offset_into_page, PAGE_CACHE_SIZE); in xfs_vm_writepage()
1675 loff_t from = pos & (PAGE_CACHE_SIZE - 1); in xfs_vm_write_failed()
1751 ASSERT(len <= PAGE_CACHE_SIZE); in xfs_vm_write_begin()
1804 ASSERT(len <= PAGE_CACHE_SIZE); in xfs_vm_write_end()
Dxfs_mount.h219 return PAGE_CACHE_SIZE; in xfs_preferred_iosize()
224 PAGE_CACHE_SIZE)); in xfs_preferred_iosize()
Dxfs_file.c106 offset = (pos & (PAGE_CACHE_SIZE -1)); /* Within page */ in xfs_iozero()
107 bytes = PAGE_CACHE_SIZE - offset; in xfs_iozero()
/linux-4.1.27/drivers/mmc/core/
Dhost.c513 host->max_seg_size = PAGE_CACHE_SIZE; in mmc_alloc_host()
515 host->max_req_size = PAGE_CACHE_SIZE; in mmc_alloc_host()
517 host->max_blk_count = PAGE_CACHE_SIZE / 512; in mmc_alloc_host()
/linux-4.1.27/fs/hostfs/
Dhostfs_kern.c413 int count = PAGE_CACHE_SIZE; in hostfs_writepage()
418 count = inode->i_size & (PAGE_CACHE_SIZE-1); in hostfs_writepage()
450 PAGE_CACHE_SIZE); in hostfs_readpage()
458 memset(buffer + bytes_read, 0, PAGE_CACHE_SIZE - bytes_read); in hostfs_readpage()
488 unsigned from = pos & (PAGE_CACHE_SIZE - 1); in hostfs_write_end()
495 if (!PageUptodate(page) && err == PAGE_CACHE_SIZE) in hostfs_write_end()
/linux-4.1.27/drivers/staging/lustre/lustre/osc/
Dosc_cache.c863 } else if (blocksize < PAGE_CACHE_SIZE && in osc_extent_finish()
864 last_count != PAGE_CACHE_SIZE) { in osc_extent_finish()
874 lost_grant = PAGE_CACHE_SIZE - count; in osc_extent_finish()
1111 LASSERT(last->oap_page_off + last->oap_count <= PAGE_CACHE_SIZE); in osc_extent_make_ready()
1119 oap->oap_count = PAGE_CACHE_SIZE - oap->oap_page_off; in osc_extent_make_ready()
1278 return kms % PAGE_CACHE_SIZE; in osc_refresh_count()
1280 return PAGE_CACHE_SIZE; in osc_refresh_count()
1361 cli->cl_dirty += PAGE_CACHE_SIZE; in osc_consume_write_grant()
1364 PAGE_CACHE_SIZE, pga, pga->pg); in osc_consume_write_grant()
1380 cli->cl_dirty -= PAGE_CACHE_SIZE; in osc_release_write_grant()
[all …]
Dosc_request.c1338 (ergo(i == 0, poff + pg->count == PAGE_CACHE_SIZE) && in osc_brw_prep_request()
1340 poff == 0 && pg->count == PAGE_CACHE_SIZE) && in osc_brw_prep_request()
1904 PAGE_CACHE_SIZE); in osc_build_rpc()
2807 fm_key->fiemap.fm_start + PAGE_CACHE_SIZE - 1) in osc_get_info()
2812 PAGE_CACHE_SIZE - 1) & CFS_PAGE_MASK; in osc_get_info()
/linux-4.1.27/fs/afs/
Dwrite.c96 if (pos + PAGE_CACHE_SIZE > i_size) in afs_fill_page()
99 len = PAGE_CACHE_SIZE; in afs_fill_page()
126 unsigned from = pos & (PAGE_CACHE_SIZE - 1); in afs_write_begin()
154 if (!PageUptodate(page) && len != PAGE_CACHE_SIZE) { in afs_write_begin()
Dfile.c322 if (offset == 0 && length == PAGE_CACHE_SIZE) { in afs_invalidatepage()
Dsuper.c318 sb->s_blocksize = PAGE_CACHE_SIZE; in afs_fill_super()
/linux-4.1.27/drivers/staging/lustre/lustre/ptlrpc/
Dsec_bulk.c62 #define POINTERS_PER_PAGE (PAGE_CACHE_SIZE / sizeof(void *))
213 OBD_FREE(page_pools.epp_pools[p_idx_max2], PAGE_CACHE_SIZE); in enc_pools_release_free_pages()
297 OBD_FREE(pools[i], PAGE_CACHE_SIZE); in enc_pools_cleanup()
417 OBD_ALLOC(pools[i], PAGE_CACHE_SIZE); in enc_pools_add_pages()
Drecover.c197 LASSERTF((long)req > PAGE_CACHE_SIZE && req != LP_POISON, in ptlrpc_resend()
Dlproc_ptlrpc.c310 bufpages = (svc->srv_buf_size + PAGE_CACHE_SIZE - 1) >> PAGE_CACHE_SHIFT; in ptlrpc_lprocfs_req_history_max_seq_write()
1275 if (count > PAGE_CACHE_SIZE - 1 || count <= prefix_len) in lprocfs_wr_import()
/linux-4.1.27/drivers/staging/lustre/lustre/obdecho/
Decho_client.c1207 offset + npages * PAGE_CACHE_SIZE - 1, in cl_echo_object_brw()
1354 lsm->lsm_stripe_size = PAGE_CACHE_SIZE; in echo_create_object()
1491 LASSERT(count == PAGE_CACHE_SIZE); in echo_client_page_debug_setup()
1495 for (delta = 0; delta < PAGE_CACHE_SIZE; delta += OBD_ECHO_BLOCK_SIZE) { in echo_client_page_debug_setup()
1523 LASSERT(count == PAGE_CACHE_SIZE); in echo_client_page_debug_check()
1527 for (rc = delta = 0; delta < PAGE_CACHE_SIZE; delta += OBD_ECHO_BLOCK_SIZE) { in echo_client_page_debug_check()
1594 i++, pgp++, off += PAGE_CACHE_SIZE) { in echo_client_kbrw()
1604 pgp->count = PAGE_CACHE_SIZE; in echo_client_kbrw()
1685 for (i = 0; i < npages; i++, off += PAGE_CACHE_SIZE) { in echo_client_prep_commit()
1687 rnb[i].len = PAGE_CACHE_SIZE; in echo_client_prep_commit()
[all …]
/linux-4.1.27/include/linux/
Df2fs_fs.h256 #define NAT_ENTRY_PER_BLOCK (PAGE_CACHE_SIZE / sizeof(struct f2fs_nat_entry))
276 #define SIT_ENTRY_PER_BLOCK (PAGE_CACHE_SIZE / sizeof(struct f2fs_sit_entry))
Dpagemap.h91 #define PAGE_CACHE_SIZE PAGE_SIZE macro
93 #define PAGE_CACHE_ALIGN(addr) (((addr)+PAGE_CACHE_SIZE-1)&PAGE_CACHE_MASK)
Dnilfs2_fs.h334 #if !defined(__KERNEL__) || (PAGE_CACHE_SIZE >= 65536) in nilfs_rec_len_from_disk()
343 #if !defined(__KERNEL__) || (PAGE_CACHE_SIZE >= 65536) in nilfs_rec_len_to_disk()
Dbuffer_head.h46 #define MAX_BUF_PER_PAGE (PAGE_CACHE_SIZE / 512)
/linux-4.1.27/fs/ext3/
Dinode.c1268 from = pos & (PAGE_CACHE_SIZE - 1); in ext3_write_begin()
1385 from = pos & (PAGE_CACHE_SIZE - 1); in ext3_ordered_write_end()
1449 from = pos & (PAGE_CACHE_SIZE - 1); in ext3_journalled_write_end()
1606 if (!walk_page_buffers(NULL, page_bufs, 0, PAGE_CACHE_SIZE, in ext3_ordered_writepage()
1621 PAGE_CACHE_SIZE, NULL, bget_one); in ext3_ordered_writepage()
1638 ret = walk_page_buffers(handle, page_bufs, 0, PAGE_CACHE_SIZE, in ext3_ordered_writepage()
1641 PAGE_CACHE_SIZE, NULL, bput_one); in ext3_ordered_writepage()
1676 PAGE_CACHE_SIZE, NULL, buffer_unmapped)) { in ext3_writeback_writepage()
1735 ret = __block_write_begin(page, 0, PAGE_CACHE_SIZE, in ext3_journalled_writepage()
1742 PAGE_CACHE_SIZE, NULL, do_journal_get_write_access); in ext3_journalled_writepage()
[all …]
Dext3.h882 #if (PAGE_CACHE_SIZE >= 65536) in ext3_rec_len_from_disk()
891 #if (PAGE_CACHE_SIZE >= 65536) in ext3_rec_len_to_disk()
/linux-4.1.27/fs/configfs/
Dmount.c74 sb->s_blocksize = PAGE_CACHE_SIZE; in configfs_fill_super()
/linux-4.1.27/drivers/mmc/host/
Dtmio_mmc_dma.c66 if ((!aligned && (host->sg_len > 1 || sg->length > PAGE_CACHE_SIZE || in tmio_mmc_start_dma_rx()
142 if ((!aligned && (host->sg_len > 1 || sg->length > PAGE_CACHE_SIZE || in tmio_mmc_start_dma_tx()
/linux-4.1.27/fs/fuse/
Dfile.c687 size_t off = num_read & (PAGE_CACHE_SIZE - 1); in fuse_short_read()
690 zero_user_segment(req->pages[i], off, PAGE_CACHE_SIZE); in fuse_short_read()
707 size_t count = PAGE_CACHE_SIZE; in fuse_do_readpage()
839 (req->num_pages + 1) * PAGE_CACHE_SIZE > fc->max_read || in fuse_readpages_fill()
1006 if (!req->out.h.error && !offset && count >= PAGE_CACHE_SIZE) in fuse_send_write_pages()
1009 if (count > PAGE_CACHE_SIZE - offset) in fuse_send_write_pages()
1010 count -= PAGE_CACHE_SIZE - offset; in fuse_send_write_pages()
1027 unsigned offset = pos & (PAGE_CACHE_SIZE - 1); in fuse_fill_write_pages()
1038 size_t bytes = min_t(size_t, PAGE_CACHE_SIZE - offset, in fuse_fill_write_pages()
1075 if (offset == PAGE_CACHE_SIZE) in fuse_fill_write_pages()
[all …]
Dinode.c851 ra_pages = arg->max_readahead / PAGE_CACHE_SIZE; in process_init_reply()
888 ra_pages = fc->max_read / PAGE_CACHE_SIZE; in process_init_reply()
909 arg->max_readahead = fc->bdi.ra_pages * PAGE_CACHE_SIZE; in fuse_send_init()
941 fc->bdi.ra_pages = (VM_MAX_READAHEAD * 1024) / PAGE_CACHE_SIZE; in fuse_bdi_init()
1005 sb->s_blocksize = PAGE_CACHE_SIZE; in fuse_fill_super()
/linux-4.1.27/include/linux/ceph/
Dlibceph.h175 return ((off+len+PAGE_CACHE_SIZE-1) >> PAGE_CACHE_SHIFT) - in calc_pages_for()
/linux-4.1.27/fs/affs/
Dfile.c513 BUG_ON(to > PAGE_CACHE_SIZE); in affs_do_readpage_ofs()
617 to = PAGE_CACHE_SIZE; in affs_readpage_ofs()
620 memset(page_address(page) + to, 0, PAGE_CACHE_SIZE - to); in affs_readpage_ofs()
660 err = affs_do_readpage_ofs(page, PAGE_CACHE_SIZE); in affs_write_begin_ofs()
681 from = pos & (PAGE_CACHE_SIZE - 1); in affs_write_end_ofs()
/linux-4.1.27/fs/hfs/
Dbtree.c119 tree->pages_per_bnode = (tree->node_size + PAGE_CACHE_SIZE - 1) >> PAGE_CACHE_SHIFT; in hfs_btree_open()
282 if (++off >= PAGE_CACHE_SIZE) { in hfs_bmap_alloc()
Dbnode.c432 min((int)PAGE_CACHE_SIZE, (int)tree->node_size)); in hfs_bnode_create()
436 memset(kmap(*++pagep), 0, PAGE_CACHE_SIZE); in hfs_bnode_create()
/linux-4.1.27/drivers/staging/lustre/lustre/libcfs/linux/
Dlinux-tcpip.c157 if (nalloc * sizeof(*ifr) > PAGE_CACHE_SIZE) { in libcfs_ipif_enumerate()
159 nalloc = PAGE_CACHE_SIZE/sizeof(*ifr); in libcfs_ipif_enumerate()
/linux-4.1.27/fs/kernfs/
Dmount.c72 sb->s_blocksize = PAGE_CACHE_SIZE; in kernfs_fill_super()
/linux-4.1.27/drivers/md/
Dmultipath.c271 PAGE_CACHE_SIZE - 1); in multipath_add_disk()
443 PAGE_CACHE_SIZE - 1); in multipath_run()
/linux-4.1.27/fs/ramfs/
Dinode.c225 sb->s_blocksize = PAGE_CACHE_SIZE; in ramfs_fill_super()
/linux-4.1.27/fs/efivarfs/
Dsuper.c200 sb->s_blocksize = PAGE_CACHE_SIZE; in efivarfs_fill_super()
/linux-4.1.27/fs/ncpfs/
Dncplib_kernel.h194 #define NCP_DIRCACHE_SIZE ((int)(PAGE_CACHE_SIZE/sizeof(struct dentry *)))
/linux-4.1.27/drivers/block/
Dpmem.c94 pmem_do_bvec(pmem, page, PAGE_CACHE_SIZE, 0, rw, sector); in pmem_rw_page()
Dbrd.c367 int err = brd_do_bvec(brd, page, PAGE_CACHE_SIZE, 0, rw, sector); in brd_rw_page()
/linux-4.1.27/Documentation/filesystems/
Dcramfs.txt41 same endianness, and can be read only by kernels with PAGE_CACHE_SIZE
Dtmpfs.txt63 nr_blocks: The same as size, but in blocks of PAGE_CACHE_SIZE.
/linux-4.1.27/drivers/oprofile/
Doprofilefs.c242 sb->s_blocksize = PAGE_CACHE_SIZE; in oprofilefs_fill_super()
/linux-4.1.27/fs/dlm/
Dlowcomms.c813 cbuf_init(&con->cb, PAGE_CACHE_SIZE); in receive_from_sock()
835 iov[0].iov_len = PAGE_CACHE_SIZE - cbuf_data(&con->cb); in receive_from_sock()
865 PAGE_CACHE_SIZE); in receive_from_sock()
1503 (PAGE_CACHE_SIZE - e->end < len)) { in dlm_lowcomms_get_buffer()
/linux-4.1.27/fs/ocfs2/cluster/
Dheartbeat.c421 vec_start = (cs << bits) % PAGE_CACHE_SIZE; in o2hb_setup_one_bio()
426 vec_len = min(PAGE_CACHE_SIZE - vec_start, in o2hb_setup_one_bio()
427 (max_slots-cs) * (PAGE_CACHE_SIZE/spp) ); in o2hb_setup_one_bio()
435 cs += vec_len / (PAGE_CACHE_SIZE/spp); in o2hb_setup_one_bio()
1602 reg->hr_slots_per_page = PAGE_CACHE_SIZE >> reg->hr_block_bits; in o2hb_init_region_params()
/linux-4.1.27/drivers/usb/storage/
Dscsiglue.c126 max_sectors = PAGE_CACHE_SIZE >> 9; in slave_configure()
/linux-4.1.27/drivers/staging/lustre/lnet/lnet/
Dlib-md.c139 lmd->md_iov.kiov[i].kiov_len > PAGE_CACHE_SIZE) in lnet_md_build()
/linux-4.1.27/fs/pstore/
Dinode.c422 sb->s_blocksize = PAGE_CACHE_SIZE; in pstore_fill_super()
/linux-4.1.27/arch/s390/hypfs/
Dinode.c286 sb->s_blocksize = PAGE_CACHE_SIZE; in hypfs_fill_super()
/linux-4.1.27/fs/ocfs2/dlmfs/
Ddlmfs.c574 sb->s_blocksize = PAGE_CACHE_SIZE; in dlmfs_fill_super()
/linux-4.1.27/drivers/block/aoe/
Daoeblk.c400 q->backing_dev_info.ra_pages = READ_AHEAD / PAGE_CACHE_SIZE; in aoeblk_gdalloc()
/linux-4.1.27/drivers/misc/ibmasm/
Dibmasmfs.c119 sb->s_blocksize = PAGE_CACHE_SIZE; in ibmasmfs_fill_super()
/linux-4.1.27/net/sunrpc/auth_gss/
Dgss_krb5_wrap.c84 & (PAGE_CACHE_SIZE - 1); in gss_krb5_remove_padding()
/linux-4.1.27/fs/jbd/
Dtransaction.c2035 int partial_page = (offset || length < PAGE_CACHE_SIZE); in journal_invalidatepage()
2043 BUG_ON(stop > PAGE_CACHE_SIZE || stop < length); in journal_invalidatepage()
/linux-4.1.27/drivers/staging/lustre/lnet/klnds/socklnd/
Dsocklnd_lib-linux.c289 (kiov[i].kiov_offset + kiov[i].kiov_len != PAGE_CACHE_SIZE && i < niov - 1)) in ksocknal_lib_kiov_vmap()

12