--- Revision 353336633239 +++ Revision 633936356530 @@ -1,54 +1,95 @@ diff --git a/fs/btrfs/inode.c b/fs/btrfs/inode.c -index 3f180b857e20..eea7d68fdcf2 100644 +index 3f180b857e20..1e7f2e2ba4f0 100644 --- a/fs/btrfs/inode.c +++ b/fs/btrfs/inode.c -@@ -2928,7 +2928,7 @@ static int btrfs_finish_ordered_io(struct btrfs_ordered_extent *ordered_extent) - int compress_type = 0; - int ret = 0; - u64 logical_len = ordered_extent->len; -- bool nolock; -+ bool nolock = false; - bool truncated = false; - bool range_locked = false; - bool clear_new_delalloc_bytes = false; -@@ -2939,7 +2939,16 @@ static int btrfs_finish_ordered_io(struct btrfs_ordered_extent *ordered_extent) - !test_bit(BTRFS_ORDERED_DIRECT, &ordered_extent->flags)) - clear_new_delalloc_bytes = true; +@@ -2912,6 +2912,26 @@ static void btrfs_release_delalloc_bytes(struct btrfs_fs_info *fs_info, + btrfs_put_block_group(cache); + } -- nolock = btrfs_is_free_space_inode(BTRFS_I(inode)); -+ if (btrfs_is_free_space_inode(BTRFS_I(inode))) { -+ nolock = true; -+ } else if (btrfs_test_opt(fs_info, FLUSHONCOMMIT)) { -+ spin_lock(&fs_info->trans_lock); -+ if (fs_info->running_transaction && -+ (fs_info->running_transaction->state >= -+ TRANS_STATE_COMMIT_START)) -+ nolock = true; -+ spin_unlock(&fs_info->trans_lock); -+ } ++static struct btrfs_trans_handle *finish_io_join_trans(struct inode *inode) ++{ ++ struct btrfs_root *root = BTRFS_I(inode)->root; ++ bool nolock = false; ++ ++ if (btrfs_is_free_space_inode(BTRFS_I(inode))) { ++ nolock = true; ++ } else if (btrfs_test_opt(fs_info, FLUSHONCOMMIT)) { ++ spin_lock(&fs_info->trans_lock); ++ if (fs_info->running_transaction) ++ nolock = true; ++ spin_unlock(&fs_info->trans_lock); ++ } ++ ++ if (nolock) ++ return btrfs_join_transaction_nolock(root); ++ ++ return btrfs_join_transaction(root); ++} ++ + /* as ordered data IO finishes, this gets called so we can finish + * an ordered extent if the range of bytes in the file it covers are + * fully written. +@@ -2928,7 +2948,6 @@ static int btrfs_finish_ordered_io(struct btrfs_ordered_extent *ordered_extent) + int compress_type = 0; + int ret = 0; + u64 logical_len = ordered_extent->len; +- bool nolock; + bool truncated = false; + bool range_locked = false; + bool clear_new_delalloc_bytes = false; +@@ -2939,8 +2958,6 @@ static int btrfs_finish_ordered_io(struct btrfs_ordered_extent *ordered_extent) + !test_bit(BTRFS_ORDERED_DIRECT, &ordered_extent->flags)) + clear_new_delalloc_bytes = true; - if (test_bit(BTRFS_ORDERED_IOERR, &ordered_extent->flags)) { - ret = -EIO; +- nolock = btrfs_is_free_space_inode(BTRFS_I(inode)); +- + if (test_bit(BTRFS_ORDERED_IOERR, &ordered_extent->flags)) { + ret = -EIO; + goto out; +@@ -2970,10 +2987,7 @@ static int btrfs_finish_ordered_io(struct btrfs_ordered_extent *ordered_extent) + btrfs_qgroup_free_data(inode, NULL, ordered_extent->file_offset, + ordered_extent->len); + btrfs_ordered_update_i_size(inode, 0, ordered_extent); +- if (nolock) +- trans = btrfs_join_transaction_nolock(root); +- else +- trans = btrfs_join_transaction(root); ++ trans = finish_io_join_trans(inode); + if (IS_ERR(trans)) { + ret = PTR_ERR(trans); + trans = NULL; +@@ -3005,10 +3019,7 @@ static int btrfs_finish_ordered_io(struct btrfs_ordered_extent *ordered_extent) + EXTENT_DEFRAG, 0, 0, &cached_state); + } + +- if (nolock) +- trans = btrfs_join_transaction_nolock(root); +- else +- trans = btrfs_join_transaction(root); ++ trans = finish_io_join_trans(inode); + if (IS_ERR(trans)) { + ret = PTR_ERR(trans); + trans = NULL; diff --git a/fs/btrfs/transaction.c b/fs/btrfs/transaction.c -index acdad6d658f5..42d756267e9a 100644 +index acdad6d658f5..d776990fe8e5 100644 --- a/fs/btrfs/transaction.c +++ b/fs/btrfs/transaction.c @@ -1888,17 +1888,8 @@ static void btrfs_cleanup_pending_block_groups(struct btrfs_trans_handle *trans) static inline int btrfs_start_delalloc_flush(struct btrfs_fs_info *fs_info) { -- /* -- * We use writeback_inodes_sb here because if we used -- * btrfs_start_delalloc_roots we would deadlock with fs freeze. -- * Currently are holding the fs freeze lock, if we do an async flush -- * we'll do btrfs_join_transaction() and deadlock because we need to -- * wait for the fs freeze lock. Using the direct flushing we benefit -- * from already being in a transaction and our join_transaction doesn't -- * have to re-take the fs freeze lock. -- */ - if (btrfs_test_opt(fs_info, FLUSHONCOMMIT)) -- writeback_inodes_sb(fs_info->sb, WB_REASON_SYNC); -+ return btrfs_start_delalloc_roots(fs_info, -1); - return 0; +- /* +- * We use writeback_inodes_sb here because if we used +- * btrfs_start_delalloc_roots we would deadlock with fs freeze. +- * Currently are holding the fs freeze lock, if we do an async flush +- * we'll do btrfs_join_transaction() and deadlock because we need to +- * wait for the fs freeze lock. Using the direct flushing we benefit +- * from already being in a transaction and our join_transaction doesn't +- * have to re-take the fs freeze lock. +- */ + if (btrfs_test_opt(fs_info, FLUSHONCOMMIT)) +- writeback_inodes_sb(fs_info->sb, WB_REASON_SYNC); ++ return btrfs_start_delalloc_roots(fs_info, -1); + return 0; }