1/* 2 * fs/logfs/file.c - prepare_write, commit_write and friends 3 * 4 * As should be obvious for Linux kernel code, license is GPLv2 5 * 6 * Copyright (c) 2005-2008 Joern Engel <joern@logfs.org> 7 */ 8#include "logfs.h" 9#include <linux/sched.h> 10#include <linux/writeback.h> 11 12static int logfs_write_begin(struct file *file, struct address_space *mapping, 13 loff_t pos, unsigned len, unsigned flags, 14 struct page **pagep, void **fsdata) 15{ 16 struct inode *inode = mapping->host; 17 struct page *page; 18 pgoff_t index = pos >> PAGE_CACHE_SHIFT; 19 20 page = grab_cache_page_write_begin(mapping, index, flags); 21 if (!page) 22 return -ENOMEM; 23 *pagep = page; 24 25 if ((len == PAGE_CACHE_SIZE) || PageUptodate(page)) 26 return 0; 27 if ((pos & PAGE_CACHE_MASK) >= i_size_read(inode)) { 28 unsigned start = pos & (PAGE_CACHE_SIZE - 1); 29 unsigned end = start + len; 30 31 /* Reading beyond i_size is simple: memset to zero */ 32 zero_user_segments(page, 0, start, end, PAGE_CACHE_SIZE); 33 return 0; 34 } 35 return logfs_readpage_nolock(page); 36} 37 38static int logfs_write_end(struct file *file, struct address_space *mapping, 39 loff_t pos, unsigned len, unsigned copied, struct page *page, 40 void *fsdata) 41{ 42 struct inode *inode = mapping->host; 43 pgoff_t index = page->index; 44 unsigned start = pos & (PAGE_CACHE_SIZE - 1); 45 unsigned end = start + copied; 46 int ret = 0; 47 48 BUG_ON(PAGE_CACHE_SIZE != inode->i_sb->s_blocksize); 49 BUG_ON(page->index > I3_BLOCKS); 50 51 if (copied < len) { 52 /* 53 * Short write of a non-initialized paged. Just tell userspace 54 * to retry the entire page. 55 */ 56 if (!PageUptodate(page)) { 57 copied = 0; 58 goto out; 59 } 60 } 61 if (copied == 0) 62 goto out; /* FIXME: do we need to update inode? */ 63 64 if (i_size_read(inode) < (index << PAGE_CACHE_SHIFT) + end) { 65 i_size_write(inode, (index << PAGE_CACHE_SHIFT) + end); 66 mark_inode_dirty_sync(inode); 67 } 68 69 SetPageUptodate(page); 70 if (!PageDirty(page)) { 71 if (!get_page_reserve(inode, page)) 72 __set_page_dirty_nobuffers(page); 73 else 74 ret = logfs_write_buf(inode, page, WF_LOCK); 75 } 76out: 77 unlock_page(page); 78 page_cache_release(page); 79 return ret ? ret : copied; 80} 81 82int logfs_readpage(struct file *file, struct page *page) 83{ 84 int ret; 85 86 ret = logfs_readpage_nolock(page); 87 unlock_page(page); 88 return ret; 89} 90 91/* Clear the page's dirty flag in the radix tree. */ 92/* TODO: mucking with PageWriteback is silly. Add a generic function to clear 93 * the dirty bit from the radix tree for filesystems that don't have to wait 94 * for page writeback to finish (i.e. any compressing filesystem). 95 */ 96static void clear_radix_tree_dirty(struct page *page) 97{ 98 BUG_ON(PagePrivate(page) || page->private); 99 set_page_writeback(page); 100 end_page_writeback(page); 101} 102 103static int __logfs_writepage(struct page *page) 104{ 105 struct inode *inode = page->mapping->host; 106 int err; 107 108 err = logfs_write_buf(inode, page, WF_LOCK); 109 if (err) 110 set_page_dirty(page); 111 else 112 clear_radix_tree_dirty(page); 113 unlock_page(page); 114 return err; 115} 116 117static int logfs_writepage(struct page *page, struct writeback_control *wbc) 118{ 119 struct inode *inode = page->mapping->host; 120 loff_t i_size = i_size_read(inode); 121 pgoff_t end_index = i_size >> PAGE_CACHE_SHIFT; 122 unsigned offset; 123 u64 bix; 124 level_t level; 125 126 log_file("logfs_writepage(%lx, %lx, %p)\n", inode->i_ino, page->index, 127 page); 128 129 logfs_unpack_index(page->index, &bix, &level); 130 131 /* Indirect blocks are never truncated */ 132 if (level != 0) 133 return __logfs_writepage(page); 134 135 /* 136 * TODO: everything below is a near-verbatim copy of nobh_writepage(). 137 * The relevant bits should be factored out after logfs is merged. 138 */ 139 140 /* Is the page fully inside i_size? */ 141 if (bix < end_index) 142 return __logfs_writepage(page); 143 144 /* Is the page fully outside i_size? (truncate in progress) */ 145 offset = i_size & (PAGE_CACHE_SIZE-1); 146 if (bix > end_index || offset == 0) { 147 unlock_page(page); 148 return 0; /* don't care */ 149 } 150 151 /* 152 * The page straddles i_size. It must be zeroed out on each and every 153 * writepage invokation because it may be mmapped. "A file is mapped 154 * in multiples of the page size. For a file that is not a multiple of 155 * the page size, the remaining memory is zeroed when mapped, and 156 * writes to that region are not written out to the file." 157 */ 158 zero_user_segment(page, offset, PAGE_CACHE_SIZE); 159 return __logfs_writepage(page); 160} 161 162static void logfs_invalidatepage(struct page *page, unsigned int offset, 163 unsigned int length) 164{ 165 struct logfs_block *block = logfs_block(page); 166 167 if (block->reserved_bytes) { 168 struct super_block *sb = page->mapping->host->i_sb; 169 struct logfs_super *super = logfs_super(sb); 170 171 super->s_dirty_pages -= block->reserved_bytes; 172 block->ops->free_block(sb, block); 173 BUG_ON(bitmap_weight(block->alias_map, LOGFS_BLOCK_FACTOR)); 174 } else 175 move_page_to_btree(page); 176 BUG_ON(PagePrivate(page) || page->private); 177} 178 179static int logfs_releasepage(struct page *page, gfp_t only_xfs_uses_this) 180{ 181 return 0; /* None of these are easy to release */ 182} 183 184 185long logfs_ioctl(struct file *file, unsigned int cmd, unsigned long arg) 186{ 187 struct inode *inode = file_inode(file); 188 struct logfs_inode *li = logfs_inode(inode); 189 unsigned int oldflags, flags; 190 int err; 191 192 switch (cmd) { 193 case FS_IOC_GETFLAGS: 194 flags = li->li_flags & LOGFS_FL_USER_VISIBLE; 195 return put_user(flags, (int __user *)arg); 196 case FS_IOC_SETFLAGS: 197 if (IS_RDONLY(inode)) 198 return -EROFS; 199 200 if (!inode_owner_or_capable(inode)) 201 return -EACCES; 202 203 err = get_user(flags, (int __user *)arg); 204 if (err) 205 return err; 206 207 mutex_lock(&inode->i_mutex); 208 oldflags = li->li_flags; 209 flags &= LOGFS_FL_USER_MODIFIABLE; 210 flags |= oldflags & ~LOGFS_FL_USER_MODIFIABLE; 211 li->li_flags = flags; 212 mutex_unlock(&inode->i_mutex); 213 214 inode->i_ctime = CURRENT_TIME; 215 mark_inode_dirty_sync(inode); 216 return 0; 217 218 default: 219 return -ENOTTY; 220 } 221} 222 223int logfs_fsync(struct file *file, loff_t start, loff_t end, int datasync) 224{ 225 struct super_block *sb = file->f_mapping->host->i_sb; 226 struct inode *inode = file->f_mapping->host; 227 int ret; 228 229 ret = filemap_write_and_wait_range(inode->i_mapping, start, end); 230 if (ret) 231 return ret; 232 233 mutex_lock(&inode->i_mutex); 234 logfs_get_wblocks(sb, NULL, WF_LOCK); 235 logfs_write_anchor(sb); 236 logfs_put_wblocks(sb, NULL, WF_LOCK); 237 mutex_unlock(&inode->i_mutex); 238 239 return 0; 240} 241 242static int logfs_setattr(struct dentry *dentry, struct iattr *attr) 243{ 244 struct inode *inode = d_inode(dentry); 245 int err = 0; 246 247 err = inode_change_ok(inode, attr); 248 if (err) 249 return err; 250 251 if (attr->ia_valid & ATTR_SIZE) { 252 err = logfs_truncate(inode, attr->ia_size); 253 if (err) 254 return err; 255 } 256 257 setattr_copy(inode, attr); 258 mark_inode_dirty(inode); 259 return 0; 260} 261 262const struct inode_operations logfs_reg_iops = { 263 .setattr = logfs_setattr, 264}; 265 266const struct file_operations logfs_reg_fops = { 267 .read_iter = generic_file_read_iter, 268 .write_iter = generic_file_write_iter, 269 .fsync = logfs_fsync, 270 .unlocked_ioctl = logfs_ioctl, 271 .llseek = generic_file_llseek, 272 .mmap = generic_file_readonly_mmap, 273 .open = generic_file_open, 274}; 275 276const struct address_space_operations logfs_reg_aops = { 277 .invalidatepage = logfs_invalidatepage, 278 .readpage = logfs_readpage, 279 .releasepage = logfs_releasepage, 280 .set_page_dirty = __set_page_dirty_nobuffers, 281 .writepage = logfs_writepage, 282 .writepages = generic_writepages, 283 .write_begin = logfs_write_begin, 284 .write_end = logfs_write_end, 285}; 286