Joern Engel | 5db53f3 | 2009-11-20 20:13:39 +0100 | [diff] [blame] | 1 | /* |
| 2 | * fs/logfs/file.c - prepare_write, commit_write and friends |
| 3 | * |
| 4 | * As should be obvious for Linux kernel code, license is GPLv2 |
| 5 | * |
| 6 | * Copyright (c) 2005-2008 Joern Engel <joern@logfs.org> |
| 7 | */ |
| 8 | #include "logfs.h" |
| 9 | #include <linux/sched.h> |
| 10 | #include <linux/writeback.h> |
| 11 | |
| 12 | static int logfs_write_begin(struct file *file, struct address_space *mapping, |
| 13 | loff_t pos, unsigned len, unsigned flags, |
| 14 | struct page **pagep, void **fsdata) |
| 15 | { |
| 16 | struct inode *inode = mapping->host; |
| 17 | struct page *page; |
| 18 | pgoff_t index = pos >> PAGE_CACHE_SHIFT; |
| 19 | |
| 20 | page = grab_cache_page_write_begin(mapping, index, flags); |
| 21 | if (!page) |
| 22 | return -ENOMEM; |
| 23 | *pagep = page; |
| 24 | |
| 25 | if ((len == PAGE_CACHE_SIZE) || PageUptodate(page)) |
| 26 | return 0; |
| 27 | if ((pos & PAGE_CACHE_MASK) >= i_size_read(inode)) { |
| 28 | unsigned start = pos & (PAGE_CACHE_SIZE - 1); |
| 29 | unsigned end = start + len; |
| 30 | |
| 31 | /* Reading beyond i_size is simple: memset to zero */ |
| 32 | zero_user_segments(page, 0, start, end, PAGE_CACHE_SIZE); |
| 33 | return 0; |
| 34 | } |
| 35 | return logfs_readpage_nolock(page); |
| 36 | } |
| 37 | |
| 38 | static int logfs_write_end(struct file *file, struct address_space *mapping, |
| 39 | loff_t pos, unsigned len, unsigned copied, struct page *page, |
| 40 | void *fsdata) |
| 41 | { |
| 42 | struct inode *inode = mapping->host; |
| 43 | pgoff_t index = page->index; |
| 44 | unsigned start = pos & (PAGE_CACHE_SIZE - 1); |
| 45 | unsigned end = start + copied; |
| 46 | int ret = 0; |
| 47 | |
| 48 | BUG_ON(PAGE_CACHE_SIZE != inode->i_sb->s_blocksize); |
| 49 | BUG_ON(page->index > I3_BLOCKS); |
| 50 | |
| 51 | if (copied < len) { |
| 52 | /* |
| 53 | * Short write of a non-initialized paged. Just tell userspace |
| 54 | * to retry the entire page. |
| 55 | */ |
| 56 | if (!PageUptodate(page)) { |
| 57 | copied = 0; |
| 58 | goto out; |
| 59 | } |
| 60 | } |
| 61 | if (copied == 0) |
| 62 | goto out; /* FIXME: do we need to update inode? */ |
| 63 | |
| 64 | if (i_size_read(inode) < (index << PAGE_CACHE_SHIFT) + end) { |
| 65 | i_size_write(inode, (index << PAGE_CACHE_SHIFT) + end); |
| 66 | mark_inode_dirty_sync(inode); |
| 67 | } |
| 68 | |
| 69 | SetPageUptodate(page); |
| 70 | if (!PageDirty(page)) { |
| 71 | if (!get_page_reserve(inode, page)) |
| 72 | __set_page_dirty_nobuffers(page); |
| 73 | else |
| 74 | ret = logfs_write_buf(inode, page, WF_LOCK); |
| 75 | } |
| 76 | out: |
| 77 | unlock_page(page); |
| 78 | page_cache_release(page); |
| 79 | return ret ? ret : copied; |
| 80 | } |
| 81 | |
| 82 | int logfs_readpage(struct file *file, struct page *page) |
| 83 | { |
| 84 | int ret; |
| 85 | |
| 86 | ret = logfs_readpage_nolock(page); |
| 87 | unlock_page(page); |
| 88 | return ret; |
| 89 | } |
| 90 | |
| 91 | /* Clear the page's dirty flag in the radix tree. */ |
| 92 | /* TODO: mucking with PageWriteback is silly. Add a generic function to clear |
| 93 | * the dirty bit from the radix tree for filesystems that don't have to wait |
| 94 | * for page writeback to finish (i.e. any compressing filesystem). |
| 95 | */ |
| 96 | static void clear_radix_tree_dirty(struct page *page) |
| 97 | { |
| 98 | BUG_ON(PagePrivate(page) || page->private); |
| 99 | set_page_writeback(page); |
| 100 | end_page_writeback(page); |
| 101 | } |
| 102 | |
| 103 | static int __logfs_writepage(struct page *page) |
| 104 | { |
| 105 | struct inode *inode = page->mapping->host; |
| 106 | int err; |
| 107 | |
| 108 | err = logfs_write_buf(inode, page, WF_LOCK); |
| 109 | if (err) |
| 110 | set_page_dirty(page); |
| 111 | else |
| 112 | clear_radix_tree_dirty(page); |
| 113 | unlock_page(page); |
| 114 | return err; |
| 115 | } |
| 116 | |
| 117 | static int logfs_writepage(struct page *page, struct writeback_control *wbc) |
| 118 | { |
| 119 | struct inode *inode = page->mapping->host; |
| 120 | loff_t i_size = i_size_read(inode); |
| 121 | pgoff_t end_index = i_size >> PAGE_CACHE_SHIFT; |
| 122 | unsigned offset; |
| 123 | u64 bix; |
| 124 | level_t level; |
| 125 | |
| 126 | log_file("logfs_writepage(%lx, %lx, %p)\n", inode->i_ino, page->index, |
| 127 | page); |
| 128 | |
| 129 | logfs_unpack_index(page->index, &bix, &level); |
| 130 | |
| 131 | /* Indirect blocks are never truncated */ |
| 132 | if (level != 0) |
| 133 | return __logfs_writepage(page); |
| 134 | |
| 135 | /* |
| 136 | * TODO: everything below is a near-verbatim copy of nobh_writepage(). |
| 137 | * The relevant bits should be factored out after logfs is merged. |
| 138 | */ |
| 139 | |
| 140 | /* Is the page fully inside i_size? */ |
| 141 | if (bix < end_index) |
| 142 | return __logfs_writepage(page); |
| 143 | |
| 144 | /* Is the page fully outside i_size? (truncate in progress) */ |
| 145 | offset = i_size & (PAGE_CACHE_SIZE-1); |
| 146 | if (bix > end_index || offset == 0) { |
| 147 | unlock_page(page); |
| 148 | return 0; /* don't care */ |
| 149 | } |
| 150 | |
| 151 | /* |
| 152 | * The page straddles i_size. It must be zeroed out on each and every |
| 153 | * writepage invokation because it may be mmapped. "A file is mapped |
| 154 | * in multiples of the page size. For a file that is not a multiple of |
| 155 | * the page size, the remaining memory is zeroed when mapped, and |
| 156 | * writes to that region are not written out to the file." |
| 157 | */ |
| 158 | zero_user_segment(page, offset, PAGE_CACHE_SIZE); |
| 159 | return __logfs_writepage(page); |
| 160 | } |
| 161 | |
Lukas Czerner | d47992f | 2013-05-21 23:17:23 -0400 | [diff] [blame] | 162 | static void logfs_invalidatepage(struct page *page, unsigned int offset, |
| 163 | unsigned int length) |
Joern Engel | 5db53f3 | 2009-11-20 20:13:39 +0100 | [diff] [blame] | 164 | { |
Joern Engel | 05ebad8 | 2010-05-04 19:41:09 +0200 | [diff] [blame] | 165 | struct logfs_block *block = logfs_block(page); |
| 166 | |
| 167 | if (block->reserved_bytes) { |
| 168 | struct super_block *sb = page->mapping->host->i_sb; |
| 169 | struct logfs_super *super = logfs_super(sb); |
| 170 | |
| 171 | super->s_dirty_pages -= block->reserved_bytes; |
| 172 | block->ops->free_block(sb, block); |
| 173 | BUG_ON(bitmap_weight(block->alias_map, LOGFS_BLOCK_FACTOR)); |
| 174 | } else |
| 175 | move_page_to_btree(page); |
Joern Engel | 5db53f3 | 2009-11-20 20:13:39 +0100 | [diff] [blame] | 176 | BUG_ON(PagePrivate(page) || page->private); |
| 177 | } |
| 178 | |
| 179 | static int logfs_releasepage(struct page *page, gfp_t only_xfs_uses_this) |
| 180 | { |
| 181 | return 0; /* None of these are easy to release */ |
| 182 | } |
| 183 | |
| 184 | |
Arnd Bergmann | 02d6d68 | 2010-04-27 22:30:06 +0200 | [diff] [blame] | 185 | long logfs_ioctl(struct file *file, unsigned int cmd, unsigned long arg) |
Joern Engel | 5db53f3 | 2009-11-20 20:13:39 +0100 | [diff] [blame] | 186 | { |
Al Viro | 496ad9a | 2013-01-23 17:07:38 -0500 | [diff] [blame] | 187 | struct inode *inode = file_inode(file); |
Joern Engel | 5db53f3 | 2009-11-20 20:13:39 +0100 | [diff] [blame] | 188 | struct logfs_inode *li = logfs_inode(inode); |
| 189 | unsigned int oldflags, flags; |
| 190 | int err; |
| 191 | |
| 192 | switch (cmd) { |
| 193 | case FS_IOC_GETFLAGS: |
| 194 | flags = li->li_flags & LOGFS_FL_USER_VISIBLE; |
| 195 | return put_user(flags, (int __user *)arg); |
| 196 | case FS_IOC_SETFLAGS: |
| 197 | if (IS_RDONLY(inode)) |
| 198 | return -EROFS; |
| 199 | |
Serge E. Hallyn | 2e14967 | 2011-03-23 16:43:26 -0700 | [diff] [blame] | 200 | if (!inode_owner_or_capable(inode)) |
Joern Engel | 5db53f3 | 2009-11-20 20:13:39 +0100 | [diff] [blame] | 201 | return -EACCES; |
| 202 | |
| 203 | err = get_user(flags, (int __user *)arg); |
| 204 | if (err) |
| 205 | return err; |
| 206 | |
| 207 | mutex_lock(&inode->i_mutex); |
| 208 | oldflags = li->li_flags; |
| 209 | flags &= LOGFS_FL_USER_MODIFIABLE; |
| 210 | flags |= oldflags & ~LOGFS_FL_USER_MODIFIABLE; |
| 211 | li->li_flags = flags; |
| 212 | mutex_unlock(&inode->i_mutex); |
| 213 | |
| 214 | inode->i_ctime = CURRENT_TIME; |
| 215 | mark_inode_dirty_sync(inode); |
| 216 | return 0; |
| 217 | |
| 218 | default: |
| 219 | return -ENOTTY; |
| 220 | } |
| 221 | } |
| 222 | |
Josef Bacik | 02c24a8 | 2011-07-16 20:44:56 -0400 | [diff] [blame] | 223 | int logfs_fsync(struct file *file, loff_t start, loff_t end, int datasync) |
Joern Engel | 5db53f3 | 2009-11-20 20:13:39 +0100 | [diff] [blame] | 224 | { |
Christoph Hellwig | 7ea8085 | 2010-05-26 17:53:25 +0200 | [diff] [blame] | 225 | struct super_block *sb = file->f_mapping->host->i_sb; |
Josef Bacik | 02c24a8 | 2011-07-16 20:44:56 -0400 | [diff] [blame] | 226 | struct inode *inode = file->f_mapping->host; |
| 227 | int ret; |
Joern Engel | 5db53f3 | 2009-11-20 20:13:39 +0100 | [diff] [blame] | 228 | |
Josef Bacik | 02c24a8 | 2011-07-16 20:44:56 -0400 | [diff] [blame] | 229 | ret = filemap_write_and_wait_range(inode->i_mapping, start, end); |
| 230 | if (ret) |
| 231 | return ret; |
| 232 | |
| 233 | mutex_lock(&inode->i_mutex); |
Prasad Joshi | 13ced29 | 2012-01-28 11:36:06 +0530 | [diff] [blame] | 234 | logfs_get_wblocks(sb, NULL, WF_LOCK); |
Joern Engel | c0c79c3 | 2010-05-05 22:33:36 +0200 | [diff] [blame] | 235 | logfs_write_anchor(sb); |
Prasad Joshi | 13ced29 | 2012-01-28 11:36:06 +0530 | [diff] [blame] | 236 | logfs_put_wblocks(sb, NULL, WF_LOCK); |
Josef Bacik | 02c24a8 | 2011-07-16 20:44:56 -0400 | [diff] [blame] | 237 | mutex_unlock(&inode->i_mutex); |
| 238 | |
Joern Engel | 5db53f3 | 2009-11-20 20:13:39 +0100 | [diff] [blame] | 239 | return 0; |
| 240 | } |
| 241 | |
| 242 | static int logfs_setattr(struct dentry *dentry, struct iattr *attr) |
| 243 | { |
David Howells | 2b0143b | 2015-03-17 22:25:59 +0000 | [diff] [blame] | 244 | struct inode *inode = d_inode(dentry); |
Joern Engel | 5db53f3 | 2009-11-20 20:13:39 +0100 | [diff] [blame] | 245 | int err = 0; |
| 246 | |
Christoph Hellwig | db78b87 | 2010-06-04 11:30:03 +0200 | [diff] [blame] | 247 | err = inode_change_ok(inode, attr); |
| 248 | if (err) |
| 249 | return err; |
| 250 | |
Christoph Hellwig | 1025774 | 2010-06-04 11:30:02 +0200 | [diff] [blame] | 251 | if (attr->ia_valid & ATTR_SIZE) { |
Joern Engel | 5db53f3 | 2009-11-20 20:13:39 +0100 | [diff] [blame] | 252 | err = logfs_truncate(inode, attr->ia_size); |
Christoph Hellwig | 1025774 | 2010-06-04 11:30:02 +0200 | [diff] [blame] | 253 | if (err) |
| 254 | return err; |
| 255 | } |
Joern Engel | 5db53f3 | 2009-11-20 20:13:39 +0100 | [diff] [blame] | 256 | |
Christoph Hellwig | 1025774 | 2010-06-04 11:30:02 +0200 | [diff] [blame] | 257 | setattr_copy(inode, attr); |
| 258 | mark_inode_dirty(inode); |
| 259 | return 0; |
Joern Engel | 5db53f3 | 2009-11-20 20:13:39 +0100 | [diff] [blame] | 260 | } |
| 261 | |
| 262 | const struct inode_operations logfs_reg_iops = { |
| 263 | .setattr = logfs_setattr, |
| 264 | }; |
| 265 | |
| 266 | const struct file_operations logfs_reg_fops = { |
Al Viro | aad4f8b | 2014-04-02 14:33:16 -0400 | [diff] [blame] | 267 | .read_iter = generic_file_read_iter, |
Al Viro | 8174202 | 2014-04-03 03:17:43 -0400 | [diff] [blame] | 268 | .write_iter = generic_file_write_iter, |
Joern Engel | 5db53f3 | 2009-11-20 20:13:39 +0100 | [diff] [blame] | 269 | .fsync = logfs_fsync, |
Arnd Bergmann | 02d6d68 | 2010-04-27 22:30:06 +0200 | [diff] [blame] | 270 | .unlocked_ioctl = logfs_ioctl, |
Joern Engel | 5db53f3 | 2009-11-20 20:13:39 +0100 | [diff] [blame] | 271 | .llseek = generic_file_llseek, |
| 272 | .mmap = generic_file_readonly_mmap, |
| 273 | .open = generic_file_open, |
Joern Engel | 5db53f3 | 2009-11-20 20:13:39 +0100 | [diff] [blame] | 274 | }; |
| 275 | |
| 276 | const struct address_space_operations logfs_reg_aops = { |
| 277 | .invalidatepage = logfs_invalidatepage, |
| 278 | .readpage = logfs_readpage, |
| 279 | .releasepage = logfs_releasepage, |
| 280 | .set_page_dirty = __set_page_dirty_nobuffers, |
| 281 | .writepage = logfs_writepage, |
| 282 | .writepages = generic_writepages, |
| 283 | .write_begin = logfs_write_begin, |
| 284 | .write_end = logfs_write_end, |
| 285 | }; |