]> asedeno.scripts.mit.edu Git - linux.git/blob - fs/btrfs/transaction.h
Merge tag 'armsoc-soc' of git://git.kernel.org/pub/scm/linux/kernel/git/soc/soc
[linux.git] / fs / btrfs / transaction.h
1 /* SPDX-License-Identifier: GPL-2.0 */
2 /*
3  * Copyright (C) 2007 Oracle.  All rights reserved.
4  */
5
6 #ifndef BTRFS_TRANSACTION_H
7 #define BTRFS_TRANSACTION_H
8
9 #include <linux/refcount.h>
10 #include "btrfs_inode.h"
11 #include "delayed-ref.h"
12 #include "ctree.h"
13
14 enum btrfs_trans_state {
15         TRANS_STATE_RUNNING,
16         TRANS_STATE_COMMIT_START,
17         TRANS_STATE_COMMIT_DOING,
18         TRANS_STATE_UNBLOCKED,
19         TRANS_STATE_COMPLETED,
20         TRANS_STATE_MAX,
21 };
22
23 #define BTRFS_TRANS_HAVE_FREE_BGS       0
24 #define BTRFS_TRANS_DIRTY_BG_RUN        1
25 #define BTRFS_TRANS_CACHE_ENOSPC        2
26
27 struct btrfs_transaction {
28         u64 transid;
29         /*
30          * total external writers(USERSPACE/START/ATTACH) in this
31          * transaction, it must be zero before the transaction is
32          * being committed
33          */
34         atomic_t num_extwriters;
35         /*
36          * total writers in this transaction, it must be zero before the
37          * transaction can end
38          */
39         atomic_t num_writers;
40         refcount_t use_count;
41
42         unsigned long flags;
43
44         /* Be protected by fs_info->trans_lock when we want to change it. */
45         enum btrfs_trans_state state;
46         int aborted;
47         struct list_head list;
48         struct extent_io_tree dirty_pages;
49         time64_t start_time;
50         wait_queue_head_t writer_wait;
51         wait_queue_head_t commit_wait;
52         struct list_head pending_snapshots;
53         struct list_head dev_update_list;
54         struct list_head switch_commits;
55         struct list_head dirty_bgs;
56
57         /*
58          * There is no explicit lock which protects io_bgs, rather its
59          * consistency is implied by the fact that all the sites which modify
60          * it do so under some form of transaction critical section, namely:
61          *
62          * - btrfs_start_dirty_block_groups - This function can only ever be
63          *   run by one of the transaction committers. Refer to
64          *   BTRFS_TRANS_DIRTY_BG_RUN usage in btrfs_commit_transaction
65          *
66          * - btrfs_write_dirty_blockgroups - this is called by
67          *   commit_cowonly_roots from transaction critical section
68          *   (TRANS_STATE_COMMIT_DOING)
69          *
70          * - btrfs_cleanup_dirty_bgs - called on transaction abort
71          */
72         struct list_head io_bgs;
73         struct list_head dropped_roots;
74
75         /*
76          * we need to make sure block group deletion doesn't race with
77          * free space cache writeout.  This mutex keeps them from stomping
78          * on each other
79          */
80         struct mutex cache_write_mutex;
81         spinlock_t dirty_bgs_lock;
82         /* Protected by spin lock fs_info->unused_bgs_lock. */
83         struct list_head deleted_bgs;
84         spinlock_t dropped_roots_lock;
85         struct btrfs_delayed_ref_root delayed_refs;
86         struct btrfs_fs_info *fs_info;
87 };
88
89 #define __TRANS_FREEZABLE       (1U << 0)
90
91 #define __TRANS_START           (1U << 9)
92 #define __TRANS_ATTACH          (1U << 10)
93 #define __TRANS_JOIN            (1U << 11)
94 #define __TRANS_JOIN_NOLOCK     (1U << 12)
95 #define __TRANS_DUMMY           (1U << 13)
96 #define __TRANS_JOIN_NOSTART    (1U << 14)
97
98 #define TRANS_START             (__TRANS_START | __TRANS_FREEZABLE)
99 #define TRANS_ATTACH            (__TRANS_ATTACH)
100 #define TRANS_JOIN              (__TRANS_JOIN | __TRANS_FREEZABLE)
101 #define TRANS_JOIN_NOLOCK       (__TRANS_JOIN_NOLOCK)
102 #define TRANS_JOIN_NOSTART      (__TRANS_JOIN_NOSTART)
103
104 #define TRANS_EXTWRITERS        (__TRANS_START | __TRANS_ATTACH)
105
106 #define BTRFS_SEND_TRANS_STUB   ((void *)1)
107
108 struct btrfs_trans_handle {
109         u64 transid;
110         u64 bytes_reserved;
111         u64 chunk_bytes_reserved;
112         unsigned long delayed_ref_updates;
113         struct btrfs_transaction *transaction;
114         struct btrfs_block_rsv *block_rsv;
115         struct btrfs_block_rsv *orig_rsv;
116         refcount_t use_count;
117         unsigned int type;
118         short aborted;
119         bool adding_csums;
120         bool allocating_chunk;
121         bool can_flush_pending_bgs;
122         bool reloc_reserved;
123         bool dirty;
124         struct btrfs_root *root;
125         struct btrfs_fs_info *fs_info;
126         struct list_head new_bgs;
127 };
128
129 struct btrfs_pending_snapshot {
130         struct dentry *dentry;
131         struct inode *dir;
132         struct btrfs_root *root;
133         struct btrfs_root_item *root_item;
134         struct btrfs_root *snap;
135         struct btrfs_qgroup_inherit *inherit;
136         struct btrfs_path *path;
137         /* block reservation for the operation */
138         struct btrfs_block_rsv block_rsv;
139         /* extra metadata reservation for relocation */
140         int error;
141         bool readonly;
142         struct list_head list;
143 };
144
145 static inline void btrfs_set_inode_last_trans(struct btrfs_trans_handle *trans,
146                                               struct inode *inode)
147 {
148         spin_lock(&BTRFS_I(inode)->lock);
149         BTRFS_I(inode)->last_trans = trans->transaction->transid;
150         BTRFS_I(inode)->last_sub_trans = BTRFS_I(inode)->root->log_transid;
151         BTRFS_I(inode)->last_log_commit = BTRFS_I(inode)->root->last_log_commit;
152         spin_unlock(&BTRFS_I(inode)->lock);
153 }
154
155 /*
156  * Make qgroup codes to skip given qgroupid, means the old/new_roots for
157  * qgroup won't contain the qgroupid in it.
158  */
159 static inline void btrfs_set_skip_qgroup(struct btrfs_trans_handle *trans,
160                                          u64 qgroupid)
161 {
162         struct btrfs_delayed_ref_root *delayed_refs;
163
164         delayed_refs = &trans->transaction->delayed_refs;
165         WARN_ON(delayed_refs->qgroup_to_skip);
166         delayed_refs->qgroup_to_skip = qgroupid;
167 }
168
169 static inline void btrfs_clear_skip_qgroup(struct btrfs_trans_handle *trans)
170 {
171         struct btrfs_delayed_ref_root *delayed_refs;
172
173         delayed_refs = &trans->transaction->delayed_refs;
174         WARN_ON(!delayed_refs->qgroup_to_skip);
175         delayed_refs->qgroup_to_skip = 0;
176 }
177
178 int btrfs_end_transaction(struct btrfs_trans_handle *trans);
179 struct btrfs_trans_handle *btrfs_start_transaction(struct btrfs_root *root,
180                                                    unsigned int num_items);
181 struct btrfs_trans_handle *btrfs_start_transaction_fallback_global_rsv(
182                                         struct btrfs_root *root,
183                                         unsigned int num_items,
184                                         int min_factor);
185 struct btrfs_trans_handle *btrfs_join_transaction(struct btrfs_root *root);
186 struct btrfs_trans_handle *btrfs_join_transaction_spacecache(struct btrfs_root *root);
187 struct btrfs_trans_handle *btrfs_join_transaction_nostart(struct btrfs_root *root);
188 struct btrfs_trans_handle *btrfs_attach_transaction(struct btrfs_root *root);
189 struct btrfs_trans_handle *btrfs_attach_transaction_barrier(
190                                         struct btrfs_root *root);
191 int btrfs_wait_for_commit(struct btrfs_fs_info *fs_info, u64 transid);
192
193 void btrfs_add_dead_root(struct btrfs_root *root);
194 int btrfs_defrag_root(struct btrfs_root *root);
195 int btrfs_clean_one_deleted_snapshot(struct btrfs_root *root);
196 int btrfs_commit_transaction(struct btrfs_trans_handle *trans);
197 int btrfs_commit_transaction_async(struct btrfs_trans_handle *trans,
198                                    int wait_for_unblock);
199
200 /*
201  * Try to commit transaction asynchronously, so this is safe to call
202  * even holding a spinlock.
203  *
204  * It's done by informing transaction_kthread to commit transaction without
205  * waiting for commit interval.
206  */
207 static inline void btrfs_commit_transaction_locksafe(
208                 struct btrfs_fs_info *fs_info)
209 {
210         set_bit(BTRFS_FS_NEED_ASYNC_COMMIT, &fs_info->flags);
211         wake_up_process(fs_info->transaction_kthread);
212 }
213 int btrfs_end_transaction_throttle(struct btrfs_trans_handle *trans);
214 int btrfs_should_end_transaction(struct btrfs_trans_handle *trans);
215 void btrfs_throttle(struct btrfs_fs_info *fs_info);
216 int btrfs_record_root_in_trans(struct btrfs_trans_handle *trans,
217                                 struct btrfs_root *root);
218 int btrfs_write_marked_extents(struct btrfs_fs_info *fs_info,
219                                 struct extent_io_tree *dirty_pages, int mark);
220 int btrfs_wait_tree_log_extents(struct btrfs_root *root, int mark);
221 int btrfs_transaction_blocked(struct btrfs_fs_info *info);
222 int btrfs_transaction_in_commit(struct btrfs_fs_info *info);
223 void btrfs_put_transaction(struct btrfs_transaction *transaction);
224 void btrfs_apply_pending_changes(struct btrfs_fs_info *fs_info);
225 void btrfs_add_dropped_root(struct btrfs_trans_handle *trans,
226                             struct btrfs_root *root);
227 void btrfs_trans_release_chunk_metadata(struct btrfs_trans_handle *trans);
228
229 #endif