Btrfs: add helper btrfs_fdatawrite_range
To avoid duplicating this double filemap_fdatawrite_range() call for inodes with async extents (compressed writes) so often. Signed-off-by: Filipe Manana <fdmanana@suse.com> Signed-off-by: Chris Mason <clm@fb.com>
This commit is contained in:
parent
075bdbdbe9
commit
728404dacf
|
@ -3901,6 +3901,7 @@ int btrfs_dirty_pages(struct btrfs_root *root, struct inode *inode,
|
||||||
struct page **pages, size_t num_pages,
|
struct page **pages, size_t num_pages,
|
||||||
loff_t pos, size_t write_bytes,
|
loff_t pos, size_t write_bytes,
|
||||||
struct extent_state **cached);
|
struct extent_state **cached);
|
||||||
|
int btrfs_fdatawrite_range(struct inode *inode, loff_t start, loff_t end);
|
||||||
|
|
||||||
/* tree-defrag.c */
|
/* tree-defrag.c */
|
||||||
int btrfs_defrag_leaves(struct btrfs_trans_handle *trans,
|
int btrfs_defrag_leaves(struct btrfs_trans_handle *trans,
|
||||||
|
|
|
@ -1676,6 +1676,7 @@ static ssize_t __btrfs_direct_write(struct kiocb *iocb,
|
||||||
loff_t pos)
|
loff_t pos)
|
||||||
{
|
{
|
||||||
struct file *file = iocb->ki_filp;
|
struct file *file = iocb->ki_filp;
|
||||||
|
struct inode *inode = file_inode(file);
|
||||||
ssize_t written;
|
ssize_t written;
|
||||||
ssize_t written_buffered;
|
ssize_t written_buffered;
|
||||||
loff_t endbyte;
|
loff_t endbyte;
|
||||||
|
@ -1697,13 +1698,10 @@ static ssize_t __btrfs_direct_write(struct kiocb *iocb,
|
||||||
* able to read what was just written.
|
* able to read what was just written.
|
||||||
*/
|
*/
|
||||||
endbyte = pos + written_buffered - 1;
|
endbyte = pos + written_buffered - 1;
|
||||||
err = filemap_fdatawrite_range(file->f_mapping, pos, endbyte);
|
err = btrfs_fdatawrite_range(inode, pos, endbyte);
|
||||||
if (!err && test_bit(BTRFS_INODE_HAS_ASYNC_EXTENT,
|
|
||||||
&BTRFS_I(file_inode(file))->runtime_flags))
|
|
||||||
err = filemap_fdatawrite_range(file->f_mapping, pos, endbyte);
|
|
||||||
if (err)
|
if (err)
|
||||||
goto out;
|
goto out;
|
||||||
err = filemap_fdatawait_range(file->f_mapping, pos, endbyte);
|
err = filemap_fdatawait_range(inode->i_mapping, pos, endbyte);
|
||||||
if (err)
|
if (err)
|
||||||
goto out;
|
goto out;
|
||||||
written += written_buffered;
|
written += written_buffered;
|
||||||
|
@ -1864,10 +1862,7 @@ static int start_ordered_ops(struct inode *inode, loff_t start, loff_t end)
|
||||||
int ret;
|
int ret;
|
||||||
|
|
||||||
atomic_inc(&BTRFS_I(inode)->sync_writers);
|
atomic_inc(&BTRFS_I(inode)->sync_writers);
|
||||||
ret = filemap_fdatawrite_range(inode->i_mapping, start, end);
|
ret = btrfs_fdatawrite_range(inode, start, end);
|
||||||
if (!ret && test_bit(BTRFS_INODE_HAS_ASYNC_EXTENT,
|
|
||||||
&BTRFS_I(inode)->runtime_flags))
|
|
||||||
ret = filemap_fdatawrite_range(inode->i_mapping, start, end);
|
|
||||||
atomic_dec(&BTRFS_I(inode)->sync_writers);
|
atomic_dec(&BTRFS_I(inode)->sync_writers);
|
||||||
|
|
||||||
return ret;
|
return ret;
|
||||||
|
@ -2820,3 +2815,29 @@ int btrfs_auto_defrag_init(void)
|
||||||
|
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
int btrfs_fdatawrite_range(struct inode *inode, loff_t start, loff_t end)
|
||||||
|
{
|
||||||
|
int ret;
|
||||||
|
|
||||||
|
/*
|
||||||
|
* So with compression we will find and lock a dirty page and clear the
|
||||||
|
* first one as dirty, setup an async extent, and immediately return
|
||||||
|
* with the entire range locked but with nobody actually marked with
|
||||||
|
* writeback. So we can't just filemap_write_and_wait_range() and
|
||||||
|
* expect it to work since it will just kick off a thread to do the
|
||||||
|
* actual work. So we need to call filemap_fdatawrite_range _again_
|
||||||
|
* since it will wait on the page lock, which won't be unlocked until
|
||||||
|
* after the pages have been marked as writeback and so we're good to go
|
||||||
|
* from there. We have to do this otherwise we'll miss the ordered
|
||||||
|
* extents and that results in badness. Please Josef, do not think you
|
||||||
|
* know better and pull this out at some point in the future, it is
|
||||||
|
* right and you are wrong.
|
||||||
|
*/
|
||||||
|
ret = filemap_fdatawrite_range(inode->i_mapping, start, end);
|
||||||
|
if (!ret && test_bit(BTRFS_INODE_HAS_ASYNC_EXTENT,
|
||||||
|
&BTRFS_I(inode)->runtime_flags))
|
||||||
|
ret = filemap_fdatawrite_range(inode->i_mapping, start, end);
|
||||||
|
|
||||||
|
return ret;
|
||||||
|
}
|
||||||
|
|
|
@ -7015,14 +7015,7 @@ static int lock_extent_direct(struct inode *inode, u64 lockstart, u64 lockend,
|
||||||
btrfs_put_ordered_extent(ordered);
|
btrfs_put_ordered_extent(ordered);
|
||||||
} else {
|
} else {
|
||||||
/* Screw you mmap */
|
/* Screw you mmap */
|
||||||
ret = filemap_fdatawrite_range(inode->i_mapping,
|
ret = btrfs_fdatawrite_range(inode, lockstart, lockend);
|
||||||
lockstart,
|
|
||||||
lockend);
|
|
||||||
if (!ret && test_bit(BTRFS_INODE_HAS_ASYNC_EXTENT,
|
|
||||||
&BTRFS_I(inode)->runtime_flags))
|
|
||||||
ret = filemap_fdatawrite_range(inode->i_mapping,
|
|
||||||
lockstart,
|
|
||||||
lockend);
|
|
||||||
if (ret)
|
if (ret)
|
||||||
break;
|
break;
|
||||||
ret = filemap_fdatawait_range(inode->i_mapping,
|
ret = filemap_fdatawait_range(inode->i_mapping,
|
||||||
|
|
|
@ -725,30 +725,10 @@ int btrfs_wait_ordered_range(struct inode *inode, u64 start, u64 len)
|
||||||
/* start IO across the range first to instantiate any delalloc
|
/* start IO across the range first to instantiate any delalloc
|
||||||
* extents
|
* extents
|
||||||
*/
|
*/
|
||||||
ret = filemap_fdatawrite_range(inode->i_mapping, start, orig_end);
|
ret = btrfs_fdatawrite_range(inode, start, orig_end);
|
||||||
if (ret)
|
if (ret)
|
||||||
return ret;
|
return ret;
|
||||||
/*
|
|
||||||
* So with compression we will find and lock a dirty page and clear the
|
|
||||||
* first one as dirty, setup an async extent, and immediately return
|
|
||||||
* with the entire range locked but with nobody actually marked with
|
|
||||||
* writeback. So we can't just filemap_write_and_wait_range() and
|
|
||||||
* expect it to work since it will just kick off a thread to do the
|
|
||||||
* actual work. So we need to call filemap_fdatawrite_range _again_
|
|
||||||
* since it will wait on the page lock, which won't be unlocked until
|
|
||||||
* after the pages have been marked as writeback and so we're good to go
|
|
||||||
* from there. We have to do this otherwise we'll miss the ordered
|
|
||||||
* extents and that results in badness. Please Josef, do not think you
|
|
||||||
* know better and pull this out at some point in the future, it is
|
|
||||||
* right and you are wrong.
|
|
||||||
*/
|
|
||||||
if (test_bit(BTRFS_INODE_HAS_ASYNC_EXTENT,
|
|
||||||
&BTRFS_I(inode)->runtime_flags)) {
|
|
||||||
ret = filemap_fdatawrite_range(inode->i_mapping, start,
|
|
||||||
orig_end);
|
|
||||||
if (ret)
|
|
||||||
return ret;
|
|
||||||
}
|
|
||||||
ret = filemap_fdatawait_range(inode->i_mapping, start, orig_end);
|
ret = filemap_fdatawait_range(inode->i_mapping, start, orig_end);
|
||||||
if (ret)
|
if (ret)
|
||||||
return ret;
|
return ret;
|
||||||
|
|
Loading…
Reference in New Issue