| /* |
| * fs/logfs/file.c - prepare_write, commit_write and friends |
| * |
| * As should be obvious for Linux kernel code, license is GPLv2 |
| * |
| * Copyright (c) 2005-2008 Joern Engel <joern@logfs.org> |
| */ |
| #include "logfs.h" |
| #include <linux/sched.h> |
| #include <linux/writeback.h> |
| |
| static int logfs_write_begin(struct file *file, struct address_space *mapping, |
| loff_t pos, unsigned len, unsigned flags, |
| struct page **pagep, void **fsdata) |
| { |
| struct inode *inode = mapping->host; |
| struct page *page; |
| pgoff_t index = pos >> PAGE_CACHE_SHIFT; |
| |
| page = grab_cache_page_write_begin(mapping, index, flags); |
| if (!page) |
| return -ENOMEM; |
| *pagep = page; |
| |
| if ((len == PAGE_CACHE_SIZE) || PageUptodate(page)) |
| return 0; |
| if ((pos & PAGE_CACHE_MASK) >= i_size_read(inode)) { |
| unsigned start = pos & (PAGE_CACHE_SIZE - 1); |
| unsigned end = start + len; |
| |
| /* Reading beyond i_size is simple: memset to zero */ |
| zero_user_segments(page, 0, start, end, PAGE_CACHE_SIZE); |
| return 0; |
| } |
| return logfs_readpage_nolock(page); |
| } |
| |
| static int logfs_write_end(struct file *file, struct address_space *mapping, |
| loff_t pos, unsigned len, unsigned copied, struct page *page, |
| void *fsdata) |
| { |
| struct inode *inode = mapping->host; |
| pgoff_t index = page->index; |
| unsigned start = pos & (PAGE_CACHE_SIZE - 1); |
| unsigned end = start + copied; |
| int ret = 0; |
| |
| BUG_ON(PAGE_CACHE_SIZE != inode->i_sb->s_blocksize); |
| BUG_ON(page->index > I3_BLOCKS); |
| |
| if (copied < len) { |
| /* |
| * Short write of a non-initialized paged. Just tell userspace |
| * to retry the entire page. |
| */ |
| if (!PageUptodate(page)) { |
| copied = 0; |
| goto out; |
| } |
| } |
| if (copied == 0) |
| goto out; /* FIXME: do we need to update inode? */ |
| |
| if (i_size_read(inode) < (index << PAGE_CACHE_SHIFT) + end) { |
| i_size_write(inode, (index << PAGE_CACHE_SHIFT) + end); |
| mark_inode_dirty_sync(inode); |
| } |
| |
| SetPageUptodate(page); |
| if (!PageDirty(page)) { |
| if (!get_page_reserve(inode, page)) |
| __set_page_dirty_nobuffers(page); |
| else |
| ret = logfs_write_buf(inode, page, WF_LOCK); |
| } |
| out: |
| unlock_page(page); |
| page_cache_release(page); |
| return ret ? ret : copied; |
| } |
| |
| int logfs_readpage(struct file *file, struct page *page) |
| { |
| int ret; |
| |
| ret = logfs_readpage_nolock(page); |
| unlock_page(page); |
| return ret; |
| } |
| |
| /* Clear the page's dirty flag in the radix tree. */ |
| /* TODO: mucking with PageWriteback is silly. Add a generic function to clear |
| * the dirty bit from the radix tree for filesystems that don't have to wait |
| * for page writeback to finish (i.e. any compressing filesystem). |
| */ |
| static void clear_radix_tree_dirty(struct page *page) |
| { |
| BUG_ON(PagePrivate(page) || page->private); |
| set_page_writeback(page); |
| end_page_writeback(page); |
| } |
| |
| static int __logfs_writepage(struct page *page) |
| { |
| struct inode *inode = page->mapping->host; |
| int err; |
| |
| err = logfs_write_buf(inode, page, WF_LOCK); |
| if (err) |
| set_page_dirty(page); |
| else |
| clear_radix_tree_dirty(page); |
| unlock_page(page); |
| return err; |
| } |
| |
| static int logfs_writepage(struct page *page, struct writeback_control *wbc) |
| { |
| struct inode *inode = page->mapping->host; |
| loff_t i_size = i_size_read(inode); |
| pgoff_t end_index = i_size >> PAGE_CACHE_SHIFT; |
| unsigned offset; |
| u64 bix; |
| level_t level; |
| |
| log_file("logfs_writepage(%lx, %lx, %p)\n", inode->i_ino, page->index, |
| page); |
| |
| logfs_unpack_index(page->index, &bix, &level); |
| |
| /* Indirect blocks are never truncated */ |
| if (level != 0) |
| return __logfs_writepage(page); |
| |
| /* |
| * TODO: everything below is a near-verbatim copy of nobh_writepage(). |
| * The relevant bits should be factored out after logfs is merged. |
| */ |
| |
| /* Is the page fully inside i_size? */ |
| if (bix < end_index) |
| return __logfs_writepage(page); |
| |
| /* Is the page fully outside i_size? (truncate in progress) */ |
| offset = i_size & (PAGE_CACHE_SIZE-1); |
| if (bix > end_index || offset == 0) { |
| unlock_page(page); |
| return 0; /* don't care */ |
| } |
| |
| /* |
| * The page straddles i_size. It must be zeroed out on each and every |
| * writepage invokation because it may be mmapped. "A file is mapped |
| * in multiples of the page size. For a file that is not a multiple of |
| * the page size, the remaining memory is zeroed when mapped, and |
| * writes to that region are not written out to the file." |
| */ |
| zero_user_segment(page, offset, PAGE_CACHE_SIZE); |
| return __logfs_writepage(page); |
| } |
| |
| static void logfs_invalidatepage(struct page *page, unsigned int offset, |
| unsigned int length) |
| { |
| struct logfs_block *block = logfs_block(page); |
| |
| if (block->reserved_bytes) { |
| struct super_block *sb = page->mapping->host->i_sb; |
| struct logfs_super *super = logfs_super(sb); |
| |
| super->s_dirty_pages -= block->reserved_bytes; |
| block->ops->free_block(sb, block); |
| BUG_ON(bitmap_weight(block->alias_map, LOGFS_BLOCK_FACTOR)); |
| } else |
| move_page_to_btree(page); |
| BUG_ON(PagePrivate(page) || page->private); |
| } |
| |
| static int logfs_releasepage(struct page *page, gfp_t only_xfs_uses_this) |
| { |
| return 0; /* None of these are easy to release */ |
| } |
| |
| |
| long logfs_ioctl(struct file *file, unsigned int cmd, unsigned long arg) |
| { |
| struct inode *inode = file_inode(file); |
| struct logfs_inode *li = logfs_inode(inode); |
| unsigned int oldflags, flags; |
| int err; |
| |
| switch (cmd) { |
| case FS_IOC_GETFLAGS: |
| flags = li->li_flags & LOGFS_FL_USER_VISIBLE; |
| return put_user(flags, (int __user *)arg); |
| case FS_IOC_SETFLAGS: |
| if (IS_RDONLY(inode)) |
| return -EROFS; |
| |
| if (!inode_owner_or_capable(inode)) |
| return -EACCES; |
| |
| err = get_user(flags, (int __user *)arg); |
| if (err) |
| return err; |
| |
| mutex_lock(&inode->i_mutex); |
| oldflags = li->li_flags; |
| flags &= LOGFS_FL_USER_MODIFIABLE; |
| flags |= oldflags & ~LOGFS_FL_USER_MODIFIABLE; |
| li->li_flags = flags; |
| mutex_unlock(&inode->i_mutex); |
| |
| inode->i_ctime = CURRENT_TIME; |
| mark_inode_dirty_sync(inode); |
| return 0; |
| |
| default: |
| return -ENOTTY; |
| } |
| } |
| |
| int logfs_fsync(struct file *file, loff_t start, loff_t end, int datasync) |
| { |
| struct super_block *sb = file->f_mapping->host->i_sb; |
| struct inode *inode = file->f_mapping->host; |
| int ret; |
| |
| ret = filemap_write_and_wait_range(inode->i_mapping, start, end); |
| if (ret) |
| return ret; |
| |
| mutex_lock(&inode->i_mutex); |
| logfs_get_wblocks(sb, NULL, WF_LOCK); |
| logfs_write_anchor(sb); |
| logfs_put_wblocks(sb, NULL, WF_LOCK); |
| mutex_unlock(&inode->i_mutex); |
| |
| return 0; |
| } |
| |
| static int logfs_setattr(struct dentry *dentry, struct iattr *attr) |
| { |
| struct inode *inode = dentry->d_inode; |
| int err = 0; |
| |
| err = inode_change_ok(inode, attr); |
| if (err) |
| return err; |
| |
| if (attr->ia_valid & ATTR_SIZE) { |
| err = logfs_truncate(inode, attr->ia_size); |
| if (err) |
| return err; |
| } |
| |
| setattr_copy(inode, attr); |
| mark_inode_dirty(inode); |
| return 0; |
| } |
| |
| const struct inode_operations logfs_reg_iops = { |
| .setattr = logfs_setattr, |
| }; |
| |
| const struct file_operations logfs_reg_fops = { |
| .read_iter = generic_file_read_iter, |
| .write_iter = generic_file_write_iter, |
| .fsync = logfs_fsync, |
| .unlocked_ioctl = logfs_ioctl, |
| .llseek = generic_file_llseek, |
| .mmap = generic_file_readonly_mmap, |
| .open = generic_file_open, |
| }; |
| |
| const struct address_space_operations logfs_reg_aops = { |
| .invalidatepage = logfs_invalidatepage, |
| .readpage = logfs_readpage, |
| .releasepage = logfs_releasepage, |
| .set_page_dirty = __set_page_dirty_nobuffers, |
| .writepage = logfs_writepage, |
| .writepages = generic_writepages, |
| .write_begin = logfs_write_begin, |
| .write_end = logfs_write_end, |
| }; |