Lines Matching refs:inode

38 	struct inode *inode = file_inode(vma->vm_file);  in f2fs_vm_page_mkwrite()  local
39 struct f2fs_sb_info *sbi = F2FS_I_SB(inode); in f2fs_vm_page_mkwrite()
45 sb_start_pagefault(inode->i_sb); in f2fs_vm_page_mkwrite()
47 f2fs_bug_on(sbi, f2fs_has_inline_data(inode)); in f2fs_vm_page_mkwrite()
51 set_new_dnode(&dn, inode, NULL, NULL, 0); in f2fs_vm_page_mkwrite()
62 if (unlikely(page->mapping != inode->i_mapping || in f2fs_vm_page_mkwrite()
63 page_offset(page) > i_size_read(inode) || in f2fs_vm_page_mkwrite()
78 i_size_read(inode)) { in f2fs_vm_page_mkwrite()
80 offset = i_size_read(inode) & ~PAGE_CACHE_MASK; in f2fs_vm_page_mkwrite()
92 if (f2fs_encrypted_inode(inode) && S_ISREG(inode->i_mode)) in f2fs_vm_page_mkwrite()
98 sb_end_pagefault(inode->i_sb); in f2fs_vm_page_mkwrite()
108 static int get_parent_ino(struct inode *inode, nid_t *pino) in get_parent_ino() argument
112 inode = igrab(inode); in get_parent_ino()
113 dentry = d_find_any_alias(inode); in get_parent_ino()
114 iput(inode); in get_parent_ino()
118 if (update_dent_inode(inode, inode, &dentry->d_name)) { in get_parent_ino()
128 static inline bool need_do_checkpoint(struct inode *inode) in need_do_checkpoint() argument
130 struct f2fs_sb_info *sbi = F2FS_I_SB(inode); in need_do_checkpoint()
133 if (!S_ISREG(inode->i_mode) || inode->i_nlink != 1) in need_do_checkpoint()
135 else if (file_enc_name(inode) && need_dentry_mark(sbi, inode->i_ino)) in need_do_checkpoint()
137 else if (file_wrong_pino(inode)) in need_do_checkpoint()
141 else if (!is_checkpointed_node(sbi, F2FS_I(inode)->i_pino)) in need_do_checkpoint()
143 else if (F2FS_I(inode)->xattr_ver == cur_cp_version(F2FS_CKPT(sbi))) in need_do_checkpoint()
164 static void try_to_fix_pino(struct inode *inode) in try_to_fix_pino() argument
166 struct f2fs_inode_info *fi = F2FS_I(inode); in try_to_fix_pino()
171 if (file_wrong_pino(inode) && inode->i_nlink == 1 && in try_to_fix_pino()
172 get_parent_ino(inode, &pino)) { in try_to_fix_pino()
174 file_got_pino(inode); in try_to_fix_pino()
177 mark_inode_dirty_sync(inode); in try_to_fix_pino()
178 f2fs_write_inode(inode, NULL); in try_to_fix_pino()
186 struct inode *inode = file->f_mapping->host; in f2fs_sync_file() local
187 struct f2fs_inode_info *fi = F2FS_I(inode); in f2fs_sync_file()
188 struct f2fs_sb_info *sbi = F2FS_I_SB(inode); in f2fs_sync_file()
189 nid_t ino = inode->i_ino; in f2fs_sync_file()
198 if (unlikely(f2fs_readonly(inode->i_sb))) in f2fs_sync_file()
201 trace_f2fs_sync_file_enter(inode); in f2fs_sync_file()
204 if (get_dirty_pages(inode) <= SM_I(sbi)->min_fsync_blocks) in f2fs_sync_file()
206 ret = filemap_write_and_wait_range(inode->i_mapping, start, end); in f2fs_sync_file()
210 trace_f2fs_sync_file_exit(inode, need_cp, datasync, ret); in f2fs_sync_file()
216 f2fs_write_inode(inode, NULL); in f2fs_sync_file()
244 need_cp = need_do_checkpoint(inode); in f2fs_sync_file()
249 ret = f2fs_sync_fs(inode->i_sb, 1); in f2fs_sync_file()
255 try_to_fix_pino(inode); in f2fs_sync_file()
268 mark_inode_dirty_sync(inode); in f2fs_sync_file()
269 f2fs_write_inode(inode, NULL); in f2fs_sync_file()
285 trace_f2fs_sync_file_exit(inode, need_cp, datasync, ret); in f2fs_sync_file()
327 struct inode *inode = file->f_mapping->host; in f2fs_seek_block() local
328 loff_t maxbytes = inode->i_sb->s_maxbytes; in f2fs_seek_block()
335 mutex_lock(&inode->i_mutex); in f2fs_seek_block()
337 isize = i_size_read(inode); in f2fs_seek_block()
342 if (f2fs_has_inline_data(inode) || f2fs_has_inline_dentry(inode)) { in f2fs_seek_block()
350 dirty = __get_first_dirty_index(inode->i_mapping, pgofs, whence); in f2fs_seek_block()
353 set_new_dnode(&dn, inode, NULL, NULL, 0); in f2fs_seek_block()
361 F2FS_I(inode)); in f2fs_seek_block()
368 end_offset = ADDRS_PER_PAGE(dn.node_page, F2FS_I(inode)); in f2fs_seek_block()
390 mutex_unlock(&inode->i_mutex); in f2fs_seek_block()
393 mutex_unlock(&inode->i_mutex); in f2fs_seek_block()
399 struct inode *inode = file->f_mapping->host; in f2fs_llseek() local
400 loff_t maxbytes = inode->i_sb->s_maxbytes; in f2fs_llseek()
407 maxbytes, i_size_read(inode)); in f2fs_llseek()
420 struct inode *inode = file_inode(file); in f2fs_file_mmap() local
422 if (f2fs_encrypted_inode(inode)) { in f2fs_file_mmap()
423 int err = f2fs_get_encryption_info(inode); in f2fs_file_mmap()
429 if (f2fs_has_inline_data(inode)) { in f2fs_file_mmap()
430 int err = f2fs_convert_inline_inode(inode); in f2fs_file_mmap()
440 static int f2fs_file_open(struct inode *inode, struct file *filp) in f2fs_file_open() argument
442 int ret = generic_file_open(inode, filp); in f2fs_file_open()
444 if (!ret && f2fs_encrypted_inode(inode)) { in f2fs_file_open()
445 ret = f2fs_get_encryption_info(inode); in f2fs_file_open()
454 struct f2fs_sb_info *sbi = F2FS_I_SB(dn->inode); in truncate_data_blocks_range()
471 clear_inode_flag(F2FS_I(dn->inode), in truncate_data_blocks_range()
483 F2FS_I(dn->inode)) + ofs; in truncate_data_blocks_range()
485 dec_valid_block_count(sbi, dn->inode, nr_free); in truncate_data_blocks_range()
491 trace_f2fs_truncate_data_blocks_range(dn->inode, dn->nid, in truncate_data_blocks_range()
501 static int truncate_partial_data_page(struct inode *inode, u64 from, in truncate_partial_data_page() argument
506 struct address_space *mapping = inode->i_mapping; in truncate_partial_data_page()
520 page = get_lock_data_page(inode, index, true); in truncate_partial_data_page()
526 if (!cache_only || !f2fs_encrypted_inode(inode) || !S_ISREG(inode->i_mode)) in truncate_partial_data_page()
532 int truncate_blocks(struct inode *inode, u64 from, bool lock) in truncate_blocks() argument
534 struct f2fs_sb_info *sbi = F2FS_I_SB(inode); in truncate_blocks()
535 unsigned int blocksize = inode->i_sb->s_blocksize; in truncate_blocks()
542 trace_f2fs_truncate_blocks_enter(inode, from); in truncate_blocks()
549 ipage = get_node_page(sbi, inode->i_ino); in truncate_blocks()
555 if (f2fs_has_inline_data(inode)) { in truncate_blocks()
563 set_new_dnode(&dn, inode, ipage, NULL, 0); in truncate_blocks()
571 count = ADDRS_PER_PAGE(dn.node_page, F2FS_I(inode)); in truncate_blocks()
583 err = truncate_inode_blocks(inode, free_from); in truncate_blocks()
590 err = truncate_partial_data_page(inode, from, truncate_page); in truncate_blocks()
592 trace_f2fs_truncate_blocks_exit(inode, err); in truncate_blocks()
596 int f2fs_truncate(struct inode *inode, bool lock) in f2fs_truncate() argument
600 if (!(S_ISREG(inode->i_mode) || S_ISDIR(inode->i_mode) || in f2fs_truncate()
601 S_ISLNK(inode->i_mode))) in f2fs_truncate()
604 trace_f2fs_truncate(inode); in f2fs_truncate()
607 if (f2fs_has_inline_data(inode) && !f2fs_may_inline_data(inode)) { in f2fs_truncate()
608 err = f2fs_convert_inline_inode(inode); in f2fs_truncate()
613 err = truncate_blocks(inode, i_size_read(inode), lock); in f2fs_truncate()
617 inode->i_mtime = inode->i_ctime = CURRENT_TIME; in f2fs_truncate()
618 mark_inode_dirty(inode); in f2fs_truncate()
625 struct inode *inode = d_inode(dentry); in f2fs_getattr() local
626 generic_fillattr(inode, stat); in f2fs_getattr()
632 static void __setattr_copy(struct inode *inode, const struct iattr *attr) in __setattr_copy() argument
634 struct f2fs_inode_info *fi = F2FS_I(inode); in __setattr_copy()
638 inode->i_uid = attr->ia_uid; in __setattr_copy()
640 inode->i_gid = attr->ia_gid; in __setattr_copy()
642 inode->i_atime = timespec_trunc(attr->ia_atime, in __setattr_copy()
643 inode->i_sb->s_time_gran); in __setattr_copy()
645 inode->i_mtime = timespec_trunc(attr->ia_mtime, in __setattr_copy()
646 inode->i_sb->s_time_gran); in __setattr_copy()
648 inode->i_ctime = timespec_trunc(attr->ia_ctime, in __setattr_copy()
649 inode->i_sb->s_time_gran); in __setattr_copy()
653 if (!in_group_p(inode->i_gid) && !capable(CAP_FSETID)) in __setattr_copy()
664 struct inode *inode = d_inode(dentry); in f2fs_setattr() local
665 struct f2fs_inode_info *fi = F2FS_I(inode); in f2fs_setattr()
668 err = inode_change_ok(inode, attr); in f2fs_setattr()
673 if (f2fs_encrypted_inode(inode) && in f2fs_setattr()
674 f2fs_get_encryption_info(inode)) in f2fs_setattr()
677 if (attr->ia_size <= i_size_read(inode)) { in f2fs_setattr()
678 truncate_setsize(inode, attr->ia_size); in f2fs_setattr()
679 err = f2fs_truncate(inode, true); in f2fs_setattr()
682 f2fs_balance_fs(F2FS_I_SB(inode)); in f2fs_setattr()
688 truncate_setsize(inode, attr->ia_size); in f2fs_setattr()
689 inode->i_mtime = inode->i_ctime = CURRENT_TIME; in f2fs_setattr()
693 __setattr_copy(inode, attr); in f2fs_setattr()
696 err = posix_acl_chmod(inode, get_inode_mode(inode)); in f2fs_setattr()
698 inode->i_mode = fi->i_acl_mode; in f2fs_setattr()
703 mark_inode_dirty(inode); in f2fs_setattr()
721 static int fill_zero(struct inode *inode, pgoff_t index, in fill_zero() argument
724 struct f2fs_sb_info *sbi = F2FS_I_SB(inode); in fill_zero()
733 page = get_new_data_page(inode, NULL, index, false); in fill_zero()
746 int truncate_hole(struct inode *inode, pgoff_t pg_start, pgoff_t pg_end) in truncate_hole() argument
754 set_new_dnode(&dn, inode, NULL, NULL, 0); in truncate_hole()
764 end_offset = ADDRS_PER_PAGE(dn.node_page, F2FS_I(inode)); in truncate_hole()
767 f2fs_bug_on(F2FS_I_SB(inode), count == 0 || count > end_offset); in truncate_hole()
777 static int punch_hole(struct inode *inode, loff_t offset, loff_t len) in punch_hole() argument
783 if (f2fs_has_inline_data(inode)) { in punch_hole()
784 ret = f2fs_convert_inline_inode(inode); in punch_hole()
796 ret = fill_zero(inode, pg_start, off_start, in punch_hole()
802 ret = fill_zero(inode, pg_start++, off_start, in punch_hole()
808 ret = fill_zero(inode, pg_end, 0, off_end); in punch_hole()
814 struct address_space *mapping = inode->i_mapping; in punch_hole()
816 struct f2fs_sb_info *sbi = F2FS_I_SB(inode); in punch_hole()
826 ret = truncate_hole(inode, pg_start, pg_end); in punch_hole()
834 static int __exchange_data_block(struct inode *inode, pgoff_t src, in __exchange_data_block() argument
837 struct f2fs_sb_info *sbi = F2FS_I_SB(inode); in __exchange_data_block()
843 set_new_dnode(&dn, inode, NULL, NULL, 0); in __exchange_data_block()
862 return full ? truncate_hole(inode, dst, dst + 1) : 0; in __exchange_data_block()
865 struct page *ipage = get_node_page(sbi, inode->i_ino); in __exchange_data_block()
873 set_new_dnode(&dn, inode, ipage, NULL, 0); in __exchange_data_block()
887 psrc = get_lock_data_page(inode, src, true); in __exchange_data_block()
890 pdst = get_new_data_page(inode, NULL, dst, false); in __exchange_data_block()
900 return truncate_hole(inode, src, src + 1); in __exchange_data_block()
914 static int f2fs_do_collapse(struct inode *inode, pgoff_t start, pgoff_t end) in f2fs_do_collapse() argument
916 struct f2fs_sb_info *sbi = F2FS_I_SB(inode); in f2fs_do_collapse()
917 pgoff_t nrpages = (i_size_read(inode) + PAGE_SIZE - 1) / PAGE_SIZE; in f2fs_do_collapse()
923 ret = __exchange_data_block(inode, end, start, true); in f2fs_do_collapse()
931 static int f2fs_collapse_range(struct inode *inode, loff_t offset, loff_t len) in f2fs_collapse_range() argument
937 if (offset + len >= i_size_read(inode)) in f2fs_collapse_range()
944 f2fs_balance_fs(F2FS_I_SB(inode)); in f2fs_collapse_range()
946 if (f2fs_has_inline_data(inode)) { in f2fs_collapse_range()
947 ret = f2fs_convert_inline_inode(inode); in f2fs_collapse_range()
956 ret = filemap_write_and_wait_range(inode->i_mapping, offset, LLONG_MAX); in f2fs_collapse_range()
960 truncate_pagecache(inode, offset); in f2fs_collapse_range()
962 ret = f2fs_do_collapse(inode, pg_start, pg_end); in f2fs_collapse_range()
967 filemap_write_and_wait_range(inode->i_mapping, offset, LLONG_MAX); in f2fs_collapse_range()
968 truncate_pagecache(inode, offset); in f2fs_collapse_range()
970 new_size = i_size_read(inode) - len; in f2fs_collapse_range()
971 truncate_pagecache(inode, new_size); in f2fs_collapse_range()
973 ret = truncate_blocks(inode, new_size, true); in f2fs_collapse_range()
975 i_size_write(inode, new_size); in f2fs_collapse_range()
980 static int f2fs_zero_range(struct inode *inode, loff_t offset, loff_t len, in f2fs_zero_range() argument
983 struct f2fs_sb_info *sbi = F2FS_I_SB(inode); in f2fs_zero_range()
984 struct address_space *mapping = inode->i_mapping; in f2fs_zero_range()
986 loff_t new_size = i_size_read(inode); in f2fs_zero_range()
990 ret = inode_newsize_ok(inode, (len + offset)); in f2fs_zero_range()
996 if (f2fs_has_inline_data(inode)) { in f2fs_zero_range()
997 ret = f2fs_convert_inline_inode(inode); in f2fs_zero_range()
1006 truncate_pagecache_range(inode, offset, offset + len - 1); in f2fs_zero_range()
1015 ret = fill_zero(inode, pg_start, off_start, in f2fs_zero_range()
1025 ret = fill_zero(inode, pg_start++, off_start, in f2fs_zero_range()
1040 ipage = get_node_page(sbi, inode->i_ino); in f2fs_zero_range()
1047 set_new_dnode(&dn, inode, ipage, NULL, 0); in f2fs_zero_range()
1071 ret = fill_zero(inode, pg_end, 0, off_end); in f2fs_zero_range()
1080 if (!(mode & FALLOC_FL_KEEP_SIZE) && i_size_read(inode) < new_size) { in f2fs_zero_range()
1081 i_size_write(inode, new_size); in f2fs_zero_range()
1082 mark_inode_dirty(inode); in f2fs_zero_range()
1083 update_inode_page(inode); in f2fs_zero_range()
1089 static int f2fs_insert_range(struct inode *inode, loff_t offset, loff_t len) in f2fs_insert_range() argument
1091 struct f2fs_sb_info *sbi = F2FS_I_SB(inode); in f2fs_insert_range()
1096 new_size = i_size_read(inode) + len; in f2fs_insert_range()
1097 if (new_size > inode->i_sb->s_maxbytes) in f2fs_insert_range()
1100 if (offset >= i_size_read(inode)) in f2fs_insert_range()
1109 if (f2fs_has_inline_data(inode)) { in f2fs_insert_range()
1110 ret = f2fs_convert_inline_inode(inode); in f2fs_insert_range()
1115 ret = truncate_blocks(inode, i_size_read(inode), true); in f2fs_insert_range()
1120 ret = filemap_write_and_wait_range(inode->i_mapping, offset, LLONG_MAX); in f2fs_insert_range()
1124 truncate_pagecache(inode, offset); in f2fs_insert_range()
1129 nrpages = (i_size_read(inode) + PAGE_SIZE - 1) / PAGE_SIZE; in f2fs_insert_range()
1133 ret = __exchange_data_block(inode, idx, idx + delta, false); in f2fs_insert_range()
1140 filemap_write_and_wait_range(inode->i_mapping, offset, LLONG_MAX); in f2fs_insert_range()
1141 truncate_pagecache(inode, offset); in f2fs_insert_range()
1144 i_size_write(inode, new_size); in f2fs_insert_range()
1148 static int expand_inode_data(struct inode *inode, loff_t offset, in expand_inode_data() argument
1151 struct f2fs_sb_info *sbi = F2FS_I_SB(inode); in expand_inode_data()
1153 loff_t new_size = i_size_read(inode); in expand_inode_data()
1159 ret = inode_newsize_ok(inode, (len + offset)); in expand_inode_data()
1163 if (f2fs_has_inline_data(inode)) { in expand_inode_data()
1164 ret = f2fs_convert_inline_inode(inode); in expand_inode_data()
1183 set_new_dnode(&dn, inode, NULL, NULL, 0); in expand_inode_data()
1200 i_size_read(inode) < new_size) { in expand_inode_data()
1201 i_size_write(inode, new_size); in expand_inode_data()
1202 mark_inode_dirty(inode); in expand_inode_data()
1203 update_inode_page(inode); in expand_inode_data()
1213 struct inode *inode = file_inode(file); in f2fs_fallocate() local
1217 if (!S_ISREG(inode->i_mode)) in f2fs_fallocate()
1220 if (f2fs_encrypted_inode(inode) && in f2fs_fallocate()
1229 mutex_lock(&inode->i_mutex); in f2fs_fallocate()
1232 if (offset >= inode->i_size) in f2fs_fallocate()
1235 ret = punch_hole(inode, offset, len); in f2fs_fallocate()
1237 ret = f2fs_collapse_range(inode, offset, len); in f2fs_fallocate()
1239 ret = f2fs_zero_range(inode, offset, len, mode); in f2fs_fallocate()
1241 ret = f2fs_insert_range(inode, offset, len); in f2fs_fallocate()
1243 ret = expand_inode_data(inode, offset, len, mode); in f2fs_fallocate()
1247 inode->i_mtime = inode->i_ctime = CURRENT_TIME; in f2fs_fallocate()
1248 mark_inode_dirty(inode); in f2fs_fallocate()
1252 mutex_unlock(&inode->i_mutex); in f2fs_fallocate()
1254 trace_f2fs_fallocate(inode, mode, offset, len, ret); in f2fs_fallocate()
1258 static int f2fs_release_file(struct inode *inode, struct file *filp) in f2fs_release_file() argument
1261 if (f2fs_is_atomic_file(inode)) in f2fs_release_file()
1262 commit_inmem_pages(inode, true); in f2fs_release_file()
1263 if (f2fs_is_volatile_file(inode)) { in f2fs_release_file()
1264 set_inode_flag(F2FS_I(inode), FI_DROP_CACHE); in f2fs_release_file()
1265 filemap_fdatawrite(inode->i_mapping); in f2fs_release_file()
1266 clear_inode_flag(F2FS_I(inode), FI_DROP_CACHE); in f2fs_release_file()
1286 struct inode *inode = file_inode(filp); in f2fs_ioc_getflags() local
1287 struct f2fs_inode_info *fi = F2FS_I(inode); in f2fs_ioc_getflags()
1294 struct inode *inode = file_inode(filp); in f2fs_ioc_setflags() local
1295 struct f2fs_inode_info *fi = F2FS_I(inode); in f2fs_ioc_setflags()
1304 if (!inode_owner_or_capable(inode)) { in f2fs_ioc_setflags()
1314 flags = f2fs_mask_flags(inode->i_mode, flags); in f2fs_ioc_setflags()
1316 mutex_lock(&inode->i_mutex); in f2fs_ioc_setflags()
1322 mutex_unlock(&inode->i_mutex); in f2fs_ioc_setflags()
1331 mutex_unlock(&inode->i_mutex); in f2fs_ioc_setflags()
1333 f2fs_set_inode_flags(inode); in f2fs_ioc_setflags()
1334 inode->i_ctime = CURRENT_TIME; in f2fs_ioc_setflags()
1335 mark_inode_dirty(inode); in f2fs_ioc_setflags()
1343 struct inode *inode = file_inode(filp); in f2fs_ioc_getversion() local
1345 return put_user(inode->i_generation, (int __user *)arg); in f2fs_ioc_getversion()
1350 struct inode *inode = file_inode(filp); in f2fs_ioc_start_atomic_write() local
1353 if (!inode_owner_or_capable(inode)) in f2fs_ioc_start_atomic_write()
1356 f2fs_balance_fs(F2FS_I_SB(inode)); in f2fs_ioc_start_atomic_write()
1358 if (f2fs_is_atomic_file(inode)) in f2fs_ioc_start_atomic_write()
1361 ret = f2fs_convert_inline_inode(inode); in f2fs_ioc_start_atomic_write()
1365 set_inode_flag(F2FS_I(inode), FI_ATOMIC_FILE); in f2fs_ioc_start_atomic_write()
1371 struct inode *inode = file_inode(filp); in f2fs_ioc_commit_atomic_write() local
1374 if (!inode_owner_or_capable(inode)) in f2fs_ioc_commit_atomic_write()
1377 if (f2fs_is_volatile_file(inode)) in f2fs_ioc_commit_atomic_write()
1384 if (f2fs_is_atomic_file(inode)) { in f2fs_ioc_commit_atomic_write()
1385 clear_inode_flag(F2FS_I(inode), FI_ATOMIC_FILE); in f2fs_ioc_commit_atomic_write()
1386 ret = commit_inmem_pages(inode, false); in f2fs_ioc_commit_atomic_write()
1399 struct inode *inode = file_inode(filp); in f2fs_ioc_start_volatile_write() local
1402 if (!inode_owner_or_capable(inode)) in f2fs_ioc_start_volatile_write()
1405 if (f2fs_is_volatile_file(inode)) in f2fs_ioc_start_volatile_write()
1408 ret = f2fs_convert_inline_inode(inode); in f2fs_ioc_start_volatile_write()
1412 set_inode_flag(F2FS_I(inode), FI_VOLATILE_FILE); in f2fs_ioc_start_volatile_write()
1418 struct inode *inode = file_inode(filp); in f2fs_ioc_release_volatile_write() local
1420 if (!inode_owner_or_capable(inode)) in f2fs_ioc_release_volatile_write()
1423 if (!f2fs_is_volatile_file(inode)) in f2fs_ioc_release_volatile_write()
1426 if (!f2fs_is_first_block_written(inode)) in f2fs_ioc_release_volatile_write()
1427 return truncate_partial_data_page(inode, 0, true); in f2fs_ioc_release_volatile_write()
1429 return punch_hole(inode, 0, F2FS_BLKSIZE); in f2fs_ioc_release_volatile_write()
1434 struct inode *inode = file_inode(filp); in f2fs_ioc_abort_volatile_write() local
1437 if (!inode_owner_or_capable(inode)) in f2fs_ioc_abort_volatile_write()
1444 f2fs_balance_fs(F2FS_I_SB(inode)); in f2fs_ioc_abort_volatile_write()
1446 clear_inode_flag(F2FS_I(inode), FI_ATOMIC_FILE); in f2fs_ioc_abort_volatile_write()
1447 clear_inode_flag(F2FS_I(inode), FI_VOLATILE_FILE); in f2fs_ioc_abort_volatile_write()
1448 commit_inmem_pages(inode, true); in f2fs_ioc_abort_volatile_write()
1456 struct inode *inode = file_inode(filp); in f2fs_ioc_shutdown() local
1457 struct f2fs_sb_info *sbi = F2FS_I_SB(inode); in f2fs_ioc_shutdown()
1495 struct inode *inode = file_inode(filp); in f2fs_ioc_fitrim() local
1496 struct super_block *sb = inode->i_sb; in f2fs_ioc_fitrim()
1537 struct inode *inode = file_inode(filp); in f2fs_ioc_set_encryption_policy() local
1543 return f2fs_process_policy(&policy, inode); in f2fs_ioc_set_encryption_policy()
1553 struct inode *inode = file_inode(filp); in f2fs_ioc_get_encryption_policy() local
1556 err = f2fs_get_policy(inode, &policy); in f2fs_ioc_get_encryption_policy()
1571 struct inode *inode = file_inode(filp); in f2fs_ioc_get_encryption_pwsalt() local
1572 struct f2fs_sb_info *sbi = F2FS_I_SB(inode); in f2fs_ioc_get_encryption_pwsalt()
1575 if (!f2fs_sb_has_crypto(inode->i_sb)) in f2fs_ioc_get_encryption_pwsalt()
1605 struct inode *inode = file_inode(filp); in f2fs_ioc_gc() local
1606 struct f2fs_sb_info *sbi = F2FS_I_SB(inode); in f2fs_ioc_gc()
1630 struct inode *inode = file_inode(filp); in f2fs_ioc_write_checkpoint() local
1631 struct f2fs_sb_info *sbi = F2FS_I_SB(inode); in f2fs_ioc_write_checkpoint()
1689 struct inode *inode = file_inode(iocb->ki_filp); in f2fs_file_write_iter() local
1691 if (f2fs_encrypted_inode(inode) && in f2fs_file_write_iter()
1692 !f2fs_has_encryption_key(inode) && in f2fs_file_write_iter()
1693 f2fs_get_encryption_info(inode)) in f2fs_file_write_iter()