]>
Commit | Line | Data |
---|---|---|
b2441318 | 1 | /* SPDX-License-Identifier: GPL-2.0 */ |
9888c340 DS |
2 | |
3 | #ifndef BTRFS_EXTENT_IO_H | |
4 | #define BTRFS_EXTENT_IO_H | |
d1310b2e CM |
5 | |
6 | #include <linux/rbtree.h> | |
b7ac31b7 | 7 | #include <linux/refcount.h> |
ac467772 | 8 | #include "ulist.h" |
d1310b2e CM |
9 | |
10 | /* bits for the extent state */ | |
9ee49a04 | 11 | #define EXTENT_DIRTY (1U << 0) |
4e586ca3 NB |
12 | #define EXTENT_UPTODATE (1U << 1) |
13 | #define EXTENT_LOCKED (1U << 2) | |
14 | #define EXTENT_NEW (1U << 3) | |
15 | #define EXTENT_DELALLOC (1U << 4) | |
16 | #define EXTENT_DEFRAG (1U << 5) | |
17 | #define EXTENT_BOUNDARY (1U << 6) | |
18 | #define EXTENT_NODATASUM (1U << 7) | |
19 | #define EXTENT_CLEAR_META_RESV (1U << 8) | |
20 | #define EXTENT_NEED_WAIT (1U << 9) | |
21 | #define EXTENT_DAMAGED (1U << 10) | |
22 | #define EXTENT_NORESERVE (1U << 11) | |
23 | #define EXTENT_QGROUP_RESERVED (1U << 12) | |
24 | #define EXTENT_CLEAR_DATA_RESV (1U << 13) | |
25 | #define EXTENT_DELALLOC_NEW (1U << 14) | |
a315e68f FM |
26 | #define EXTENT_DO_ACCOUNTING (EXTENT_CLEAR_META_RESV | \ |
27 | EXTENT_CLEAR_DATA_RESV) | |
ba8f5206 | 28 | #define EXTENT_CTLBITS (EXTENT_DO_ACCOUNTING) |
d1310b2e | 29 | |
8811133d NB |
30 | /* |
31 | * Redefined bits above which are used only in the device allocation tree, | |
32 | * shouldn't be using EXTENT_LOCKED / EXTENT_BOUNDARY / EXTENT_CLEAR_META_RESV | |
33 | * / EXTENT_CLEAR_DATA_RESV because they have special meaning to the bit | |
34 | * manipulation functions | |
35 | */ | |
930b0907 | 36 | #define CHUNK_ALLOCATED EXTENT_DIRTY |
8811133d | 37 | #define CHUNK_TRIMMED EXTENT_DEFRAG |
930b0907 | 38 | |
261507a0 LZ |
39 | /* |
40 | * flags for bio submission. The high bits indicate the compression | |
41 | * type for this bio | |
42 | */ | |
c8b97818 | 43 | #define EXTENT_BIO_COMPRESSED 1 |
261507a0 | 44 | #define EXTENT_BIO_FLAG_SHIFT 16 |
c8b97818 | 45 | |
80cb3836 DS |
46 | enum { |
47 | EXTENT_BUFFER_UPTODATE, | |
48 | EXTENT_BUFFER_DIRTY, | |
49 | EXTENT_BUFFER_CORRUPT, | |
50 | /* this got triggered by readahead */ | |
51 | EXTENT_BUFFER_READAHEAD, | |
52 | EXTENT_BUFFER_TREE_REF, | |
53 | EXTENT_BUFFER_STALE, | |
54 | EXTENT_BUFFER_WRITEBACK, | |
55 | /* read IO error */ | |
56 | EXTENT_BUFFER_READ_ERR, | |
57 | EXTENT_BUFFER_UNMAPPED, | |
58 | EXTENT_BUFFER_IN_TREE, | |
59 | /* write IO error */ | |
60 | EXTENT_BUFFER_WRITE_ERR, | |
61 | }; | |
b4ce94de | 62 | |
da2c7009 | 63 | /* these are flags for __process_pages_contig */ |
c2790a2e JB |
64 | #define PAGE_UNLOCK (1 << 0) |
65 | #define PAGE_CLEAR_DIRTY (1 << 1) | |
66 | #define PAGE_SET_WRITEBACK (1 << 2) | |
67 | #define PAGE_END_WRITEBACK (1 << 3) | |
68 | #define PAGE_SET_PRIVATE2 (1 << 4) | |
704de49d | 69 | #define PAGE_SET_ERROR (1 << 5) |
da2c7009 | 70 | #define PAGE_LOCK (1 << 6) |
a791e35e | 71 | |
d1310b2e CM |
72 | /* |
73 | * page->private values. Every page that is controlled by the extent | |
74 | * map has page->private set to one. | |
75 | */ | |
76 | #define EXTENT_PAGE_PRIVATE 1 | |
d1310b2e | 77 | |
2fe1d551 OS |
78 | /* |
79 | * The extent buffer bitmap operations are done with byte granularity instead of | |
80 | * word granularity for two reasons: | |
81 | * 1. The bitmaps must be little-endian on disk. | |
82 | * 2. Bitmap items are not guaranteed to be aligned to a word and therefore a | |
83 | * single word in a bitmap may straddle two pages in the extent buffer. | |
84 | */ | |
85 | #define BIT_BYTE(nr) ((nr) / BITS_PER_BYTE) | |
86 | #define BYTE_MASK ((1 << BITS_PER_BYTE) - 1) | |
87 | #define BITMAP_FIRST_BYTE_MASK(start) \ | |
88 | ((BYTE_MASK << ((start) & (BITS_PER_BYTE - 1))) & BYTE_MASK) | |
89 | #define BITMAP_LAST_BYTE_MASK(nbits) \ | |
90 | (BYTE_MASK >> (-(nbits) & (BITS_PER_BYTE - 1))) | |
91 | ||
70dec807 | 92 | struct extent_state; |
ea466794 | 93 | struct btrfs_root; |
6fc0ef68 | 94 | struct btrfs_inode; |
facc8a22 | 95 | struct btrfs_io_bio; |
47dc196a | 96 | struct io_failure_record; |
70dec807 | 97 | |
a758781d DS |
98 | |
99 | typedef blk_status_t (extent_submit_bio_start_t)(void *private_data, | |
d0779291 | 100 | struct bio *bio, u64 bio_offset); |
a758781d | 101 | |
d1310b2e | 102 | struct extent_io_ops { |
4d53dddb | 103 | /* |
52042d8e | 104 | * The following callbacks must be always defined, the function |
4d53dddb DS |
105 | * pointer will be called unconditionally. |
106 | */ | |
a56b1c7b | 107 | blk_status_t (*submit_bio_hook)(struct inode *inode, struct bio *bio, |
50489a57 | 108 | int mirror_num, unsigned long bio_flags); |
4d53dddb DS |
109 | int (*readpage_end_io_hook)(struct btrfs_io_bio *io_bio, u64 phy_offset, |
110 | struct page *page, u64 start, u64 end, | |
111 | int mirror); | |
d1310b2e CM |
112 | }; |
113 | ||
43eb5f29 QW |
114 | enum { |
115 | IO_TREE_FS_INFO_FREED_EXTENTS0, | |
116 | IO_TREE_FS_INFO_FREED_EXTENTS1, | |
117 | IO_TREE_INODE_IO, | |
118 | IO_TREE_INODE_IO_FAILURE, | |
119 | IO_TREE_RELOC_BLOCKS, | |
120 | IO_TREE_TRANS_DIRTY_PAGES, | |
121 | IO_TREE_ROOT_DIRTY_LOG_PAGES, | |
122 | IO_TREE_SELFTEST, | |
123 | }; | |
124 | ||
d1310b2e CM |
125 | struct extent_io_tree { |
126 | struct rb_root state; | |
c258d6e3 | 127 | struct btrfs_fs_info *fs_info; |
c6100a4b | 128 | void *private_data; |
d1310b2e | 129 | u64 dirty_bytes; |
7b439738 | 130 | bool track_uptodate; |
43eb5f29 QW |
131 | |
132 | /* Who owns this io tree, should be one of IO_TREE_* */ | |
133 | u8 owner; | |
134 | ||
70dec807 | 135 | spinlock_t lock; |
e8c9f186 | 136 | const struct extent_io_ops *ops; |
d1310b2e CM |
137 | }; |
138 | ||
139 | struct extent_state { | |
140 | u64 start; | |
141 | u64 end; /* inclusive */ | |
d1310b2e | 142 | struct rb_node rb_node; |
9ed74f2d JB |
143 | |
144 | /* ADD NEW ELEMENTS AFTER THIS */ | |
d1310b2e | 145 | wait_queue_head_t wq; |
b7ac31b7 | 146 | refcount_t refs; |
9ee49a04 | 147 | unsigned state; |
d1310b2e | 148 | |
47dc196a | 149 | struct io_failure_record *failrec; |
d1310b2e | 150 | |
6d49ba1b | 151 | #ifdef CONFIG_BTRFS_DEBUG |
2d2ae547 | 152 | struct list_head leak_list; |
6d49ba1b | 153 | #endif |
d1310b2e CM |
154 | }; |
155 | ||
727011e0 | 156 | #define INLINE_EXTENT_BUFFER_PAGES 16 |
09cbfeaf | 157 | #define MAX_INLINE_EXTENT_BUFFER_SIZE (INLINE_EXTENT_BUFFER_PAGES * PAGE_SIZE) |
d1310b2e CM |
158 | struct extent_buffer { |
159 | u64 start; | |
160 | unsigned long len; | |
b4ce94de | 161 | unsigned long bflags; |
f28491e0 | 162 | struct btrfs_fs_info *fs_info; |
3083ee2e | 163 | spinlock_t refs_lock; |
727011e0 | 164 | atomic_t refs; |
0b32f4bb | 165 | atomic_t io_pages; |
5cf1ab56 | 166 | int read_mirror; |
19fe0a8b | 167 | struct rcu_head rcu_head; |
5b25f70f | 168 | pid_t lock_owner; |
b4ce94de | 169 | |
06297d8c | 170 | int blocking_writers; |
bd681513 | 171 | atomic_t blocking_readers; |
ed1b4ed7 | 172 | bool lock_nested; |
656f30db FM |
173 | /* >= 0 if eb belongs to a log tree, -1 otherwise */ |
174 | short log_index; | |
bd681513 CM |
175 | |
176 | /* protects write locks */ | |
177 | rwlock_t lock; | |
178 | ||
179 | /* readers use lock_wq while they wait for the write | |
180 | * lock holders to unlock | |
181 | */ | |
182 | wait_queue_head_t write_lock_wq; | |
b4ce94de | 183 | |
bd681513 CM |
184 | /* writers use read_lock_wq while they wait for readers |
185 | * to unlock | |
b4ce94de | 186 | */ |
bd681513 | 187 | wait_queue_head_t read_lock_wq; |
b8dae313 | 188 | struct page *pages[INLINE_EXTENT_BUFFER_PAGES]; |
6d49ba1b | 189 | #ifdef CONFIG_BTRFS_DEBUG |
f3dc24c5 | 190 | int spinning_writers; |
afd495a8 | 191 | atomic_t spinning_readers; |
5c9c799a | 192 | atomic_t read_locks; |
c79adfc0 | 193 | atomic_t write_locks; |
6d49ba1b ES |
194 | struct list_head leak_list; |
195 | #endif | |
d1310b2e CM |
196 | }; |
197 | ||
ac467772 QW |
198 | /* |
199 | * Structure to record how many bytes and which ranges are set/cleared | |
200 | */ | |
201 | struct extent_changeset { | |
202 | /* How many bytes are set/cleared in this operation */ | |
7bc329c1 | 203 | unsigned int bytes_changed; |
ac467772 QW |
204 | |
205 | /* Changed ranges */ | |
53d32359 | 206 | struct ulist range_changed; |
ac467772 QW |
207 | }; |
208 | ||
364ecf36 QW |
209 | static inline void extent_changeset_init(struct extent_changeset *changeset) |
210 | { | |
211 | changeset->bytes_changed = 0; | |
212 | ulist_init(&changeset->range_changed); | |
213 | } | |
214 | ||
215 | static inline struct extent_changeset *extent_changeset_alloc(void) | |
216 | { | |
217 | struct extent_changeset *ret; | |
218 | ||
219 | ret = kmalloc(sizeof(*ret), GFP_KERNEL); | |
220 | if (!ret) | |
221 | return NULL; | |
222 | ||
223 | extent_changeset_init(ret); | |
224 | return ret; | |
225 | } | |
226 | ||
227 | static inline void extent_changeset_release(struct extent_changeset *changeset) | |
228 | { | |
229 | if (!changeset) | |
230 | return; | |
231 | changeset->bytes_changed = 0; | |
232 | ulist_release(&changeset->range_changed); | |
233 | } | |
234 | ||
235 | static inline void extent_changeset_free(struct extent_changeset *changeset) | |
236 | { | |
237 | if (!changeset) | |
238 | return; | |
239 | extent_changeset_release(changeset); | |
240 | kfree(changeset); | |
241 | } | |
242 | ||
261507a0 LZ |
243 | static inline void extent_set_compress_type(unsigned long *bio_flags, |
244 | int compress_type) | |
245 | { | |
246 | *bio_flags |= compress_type << EXTENT_BIO_FLAG_SHIFT; | |
247 | } | |
248 | ||
249 | static inline int extent_compress_type(unsigned long bio_flags) | |
250 | { | |
251 | return bio_flags >> EXTENT_BIO_FLAG_SHIFT; | |
252 | } | |
253 | ||
d1310b2e CM |
254 | struct extent_map_tree; |
255 | ||
fc4f21b1 | 256 | typedef struct extent_map *(get_extent_t)(struct btrfs_inode *inode, |
d1310b2e | 257 | struct page *page, |
306e16ce | 258 | size_t pg_offset, |
d1310b2e CM |
259 | u64 start, u64 len, |
260 | int create); | |
261 | ||
c258d6e3 | 262 | void extent_io_tree_init(struct btrfs_fs_info *fs_info, |
43eb5f29 QW |
263 | struct extent_io_tree *tree, unsigned int owner, |
264 | void *private_data); | |
41e7acd3 | 265 | void extent_io_tree_release(struct extent_io_tree *tree); |
477a30ba | 266 | int try_release_extent_mapping(struct page *page, gfp_t mask); |
f7a52a40 | 267 | int try_release_extent_buffer(struct page *page); |
1edbb734 | 268 | int lock_extent_bits(struct extent_io_tree *tree, u64 start, u64 end, |
ff13db41 | 269 | struct extent_state **cached); |
cd716d8f DS |
270 | |
271 | static inline int lock_extent(struct extent_io_tree *tree, u64 start, u64 end) | |
272 | { | |
273 | return lock_extent_bits(tree, start, end, NULL); | |
274 | } | |
275 | ||
d0082371 | 276 | int try_lock_extent(struct extent_io_tree *tree, u64 start, u64 end); |
d1310b2e | 277 | int extent_read_full_page(struct extent_io_tree *tree, struct page *page, |
8ddc7d9c | 278 | get_extent_t *get_extent, int mirror_num); |
d1310b2e | 279 | int __init extent_io_init(void); |
e67c718b | 280 | void __cold extent_io_exit(void); |
d1310b2e CM |
281 | |
282 | u64 count_range_bits(struct extent_io_tree *tree, | |
283 | u64 *start, u64 search_end, | |
9ee49a04 | 284 | u64 max_bytes, unsigned bits, int contig); |
d1310b2e | 285 | |
4845e44f | 286 | void free_extent_state(struct extent_state *state); |
d1310b2e | 287 | int test_range_bit(struct extent_io_tree *tree, u64 start, u64 end, |
9ee49a04 | 288 | unsigned bits, int filled, |
41074888 | 289 | struct extent_state *cached_state); |
fefdc557 | 290 | int clear_record_extent_bits(struct extent_io_tree *tree, u64 start, u64 end, |
f734c44a | 291 | unsigned bits, struct extent_changeset *changeset); |
e6dcd2dc | 292 | int clear_extent_bit(struct extent_io_tree *tree, u64 start, u64 end, |
9ee49a04 | 293 | unsigned bits, int wake, int delete, |
ae0f1625 | 294 | struct extent_state **cached); |
66b0c887 DS |
295 | int __clear_extent_bit(struct extent_io_tree *tree, u64 start, u64 end, |
296 | unsigned bits, int wake, int delete, | |
297 | struct extent_state **cached, gfp_t mask, | |
298 | struct extent_changeset *changeset); | |
c6317955 | 299 | |
e83b1d91 DS |
300 | static inline int unlock_extent(struct extent_io_tree *tree, u64 start, u64 end) |
301 | { | |
ae0f1625 | 302 | return clear_extent_bit(tree, start, end, EXTENT_LOCKED, 1, 0, NULL); |
e83b1d91 DS |
303 | } |
304 | ||
305 | static inline int unlock_extent_cached(struct extent_io_tree *tree, u64 start, | |
e43bbe5e | 306 | u64 end, struct extent_state **cached) |
e83b1d91 | 307 | { |
66b0c887 | 308 | return __clear_extent_bit(tree, start, end, EXTENT_LOCKED, 1, 0, cached, |
e43bbe5e | 309 | GFP_NOFS, NULL); |
e83b1d91 DS |
310 | } |
311 | ||
d810a4be DS |
312 | static inline int unlock_extent_cached_atomic(struct extent_io_tree *tree, |
313 | u64 start, u64 end, struct extent_state **cached) | |
e83b1d91 | 314 | { |
d810a4be DS |
315 | return __clear_extent_bit(tree, start, end, EXTENT_LOCKED, 1, 0, cached, |
316 | GFP_ATOMIC, NULL); | |
e83b1d91 DS |
317 | } |
318 | ||
319 | static inline int clear_extent_bits(struct extent_io_tree *tree, u64 start, | |
91166212 | 320 | u64 end, unsigned bits) |
e83b1d91 DS |
321 | { |
322 | int wake = 0; | |
323 | ||
324 | if (bits & EXTENT_LOCKED) | |
325 | wake = 1; | |
326 | ||
ae0f1625 | 327 | return clear_extent_bit(tree, start, end, bits, wake, 0, NULL); |
e83b1d91 DS |
328 | } |
329 | ||
d38ed27f | 330 | int set_record_extent_bits(struct extent_io_tree *tree, u64 start, u64 end, |
2c53b912 | 331 | unsigned bits, struct extent_changeset *changeset); |
4845e44f | 332 | int set_extent_bit(struct extent_io_tree *tree, u64 start, u64 end, |
9ee49a04 | 333 | unsigned bits, u64 *failed_start, |
4845e44f | 334 | struct extent_state **cached_state, gfp_t mask); |
4ca73656 NB |
335 | int set_extent_bits_nowait(struct extent_io_tree *tree, u64 start, u64 end, |
336 | unsigned bits); | |
c6317955 DS |
337 | |
338 | static inline int set_extent_bits(struct extent_io_tree *tree, u64 start, | |
ceeb0ae7 | 339 | u64 end, unsigned bits) |
c6317955 | 340 | { |
ceeb0ae7 | 341 | return set_extent_bit(tree, start, end, bits, NULL, NULL, GFP_NOFS); |
c6317955 DS |
342 | } |
343 | ||
e83b1d91 | 344 | static inline int clear_extent_uptodate(struct extent_io_tree *tree, u64 start, |
f08dc36f | 345 | u64 end, struct extent_state **cached_state) |
e83b1d91 | 346 | { |
66b0c887 | 347 | return __clear_extent_bit(tree, start, end, EXTENT_UPTODATE, 0, 0, |
f08dc36f | 348 | cached_state, GFP_NOFS, NULL); |
e83b1d91 | 349 | } |
c6317955 DS |
350 | |
351 | static inline int set_extent_dirty(struct extent_io_tree *tree, u64 start, | |
352 | u64 end, gfp_t mask) | |
353 | { | |
354 | return set_extent_bit(tree, start, end, EXTENT_DIRTY, NULL, | |
355 | NULL, mask); | |
356 | } | |
357 | ||
e83b1d91 | 358 | static inline int clear_extent_dirty(struct extent_io_tree *tree, u64 start, |
0e6ec385 | 359 | u64 end, struct extent_state **cached) |
e83b1d91 DS |
360 | { |
361 | return clear_extent_bit(tree, start, end, | |
362 | EXTENT_DIRTY | EXTENT_DELALLOC | | |
0e6ec385 | 363 | EXTENT_DO_ACCOUNTING, 0, 0, cached); |
e83b1d91 DS |
364 | } |
365 | ||
462d6fac | 366 | int convert_extent_bit(struct extent_io_tree *tree, u64 start, u64 end, |
9ee49a04 | 367 | unsigned bits, unsigned clear_bits, |
210aa277 | 368 | struct extent_state **cached_state); |
c6317955 DS |
369 | |
370 | static inline int set_extent_delalloc(struct extent_io_tree *tree, u64 start, | |
e3b8a485 FM |
371 | u64 end, unsigned int extra_bits, |
372 | struct extent_state **cached_state) | |
c6317955 DS |
373 | { |
374 | return set_extent_bit(tree, start, end, | |
e3b8a485 | 375 | EXTENT_DELALLOC | EXTENT_UPTODATE | extra_bits, |
7cd8c752 | 376 | NULL, cached_state, GFP_NOFS); |
c6317955 DS |
377 | } |
378 | ||
379 | static inline int set_extent_defrag(struct extent_io_tree *tree, u64 start, | |
018ed4f7 | 380 | u64 end, struct extent_state **cached_state) |
c6317955 DS |
381 | { |
382 | return set_extent_bit(tree, start, end, | |
383 | EXTENT_DELALLOC | EXTENT_UPTODATE | EXTENT_DEFRAG, | |
018ed4f7 | 384 | NULL, cached_state, GFP_NOFS); |
c6317955 DS |
385 | } |
386 | ||
387 | static inline int set_extent_new(struct extent_io_tree *tree, u64 start, | |
3744dbeb | 388 | u64 end) |
c6317955 | 389 | { |
3744dbeb DS |
390 | return set_extent_bit(tree, start, end, EXTENT_NEW, NULL, NULL, |
391 | GFP_NOFS); | |
c6317955 DS |
392 | } |
393 | ||
394 | static inline int set_extent_uptodate(struct extent_io_tree *tree, u64 start, | |
395 | u64 end, struct extent_state **cached_state, gfp_t mask) | |
396 | { | |
397 | return set_extent_bit(tree, start, end, EXTENT_UPTODATE, NULL, | |
398 | cached_state, mask); | |
399 | } | |
400 | ||
d1310b2e | 401 | int find_first_extent_bit(struct extent_io_tree *tree, u64 start, |
9ee49a04 | 402 | u64 *start_ret, u64 *end_ret, unsigned bits, |
e6138876 | 403 | struct extent_state **cached_state); |
45bfcfc1 NB |
404 | void find_first_clear_extent_bit(struct extent_io_tree *tree, u64 start, |
405 | u64 *start_ret, u64 *end_ret, unsigned bits); | |
d1310b2e CM |
406 | int extent_invalidatepage(struct extent_io_tree *tree, |
407 | struct page *page, unsigned long offset); | |
0a9b0e53 | 408 | int extent_write_full_page(struct page *page, struct writeback_control *wbc); |
5e3ee236 | 409 | int extent_write_locked_range(struct inode *inode, u64 start, u64 end, |
771ed689 | 410 | int mode); |
8ae225a8 | 411 | int extent_writepages(struct address_space *mapping, |
d1310b2e | 412 | struct writeback_control *wbc); |
0b32f4bb JB |
413 | int btree_write_cache_pages(struct address_space *mapping, |
414 | struct writeback_control *wbc); | |
2a3ff0ad NB |
415 | int extent_readpages(struct address_space *mapping, struct list_head *pages, |
416 | unsigned nr_pages); | |
1506fcc8 | 417 | int extent_fiemap(struct inode *inode, struct fiemap_extent_info *fieinfo, |
2135fb9b | 418 | __u64 start, __u64 len); |
d1310b2e CM |
419 | void set_page_extent_mapped(struct page *page); |
420 | ||
f28491e0 | 421 | struct extent_buffer *alloc_extent_buffer(struct btrfs_fs_info *fs_info, |
ce3e6984 | 422 | u64 start); |
0f331229 OS |
423 | struct extent_buffer *__alloc_dummy_extent_buffer(struct btrfs_fs_info *fs_info, |
424 | u64 start, unsigned long len); | |
3f556f78 | 425 | struct extent_buffer *alloc_dummy_extent_buffer(struct btrfs_fs_info *fs_info, |
da17066c | 426 | u64 start); |
815a51c7 | 427 | struct extent_buffer *btrfs_clone_extent_buffer(struct extent_buffer *src); |
f28491e0 | 428 | struct extent_buffer *find_extent_buffer(struct btrfs_fs_info *fs_info, |
452c75c3 | 429 | u64 start); |
d1310b2e | 430 | void free_extent_buffer(struct extent_buffer *eb); |
3083ee2e | 431 | void free_extent_buffer_stale(struct extent_buffer *eb); |
bb82ab88 AJ |
432 | #define WAIT_NONE 0 |
433 | #define WAIT_COMPLETE 1 | |
434 | #define WAIT_PAGE_LOCK 2 | |
c2ccfbc6 | 435 | int read_extent_buffer_pages(struct extent_buffer *eb, int wait, |
6af49dbd | 436 | int mirror_num); |
fd8b2b61 | 437 | void wait_on_extent_buffer_writeback(struct extent_buffer *eb); |
479ed9ab | 438 | |
cc5e31a4 | 439 | static inline int num_extent_pages(const struct extent_buffer *eb) |
479ed9ab | 440 | { |
8791d432 DS |
441 | return (round_up(eb->start + eb->len, PAGE_SIZE) >> PAGE_SHIFT) - |
442 | (eb->start >> PAGE_SHIFT); | |
479ed9ab RD |
443 | } |
444 | ||
d1310b2e CM |
445 | static inline void extent_buffer_get(struct extent_buffer *eb) |
446 | { | |
447 | atomic_inc(&eb->refs); | |
448 | } | |
449 | ||
ba020491 AJ |
450 | static inline int extent_buffer_uptodate(struct extent_buffer *eb) |
451 | { | |
452 | return test_bit(EXTENT_BUFFER_UPTODATE, &eb->bflags); | |
453 | } | |
454 | ||
1cbb1f45 JM |
455 | int memcmp_extent_buffer(const struct extent_buffer *eb, const void *ptrv, |
456 | unsigned long start, unsigned long len); | |
457 | void read_extent_buffer(const struct extent_buffer *eb, void *dst, | |
d1310b2e CM |
458 | unsigned long start, |
459 | unsigned long len); | |
1cbb1f45 JM |
460 | int read_extent_buffer_to_user(const struct extent_buffer *eb, |
461 | void __user *dst, unsigned long start, | |
550ac1d8 | 462 | unsigned long len); |
f157bf76 DS |
463 | void write_extent_buffer_fsid(struct extent_buffer *eb, const void *src); |
464 | void write_extent_buffer_chunk_tree_uuid(struct extent_buffer *eb, | |
465 | const void *src); | |
d1310b2e CM |
466 | void write_extent_buffer(struct extent_buffer *eb, const void *src, |
467 | unsigned long start, unsigned long len); | |
58e8012c DS |
468 | void copy_extent_buffer_full(struct extent_buffer *dst, |
469 | struct extent_buffer *src); | |
d1310b2e CM |
470 | void copy_extent_buffer(struct extent_buffer *dst, struct extent_buffer *src, |
471 | unsigned long dst_offset, unsigned long src_offset, | |
472 | unsigned long len); | |
473 | void memcpy_extent_buffer(struct extent_buffer *dst, unsigned long dst_offset, | |
474 | unsigned long src_offset, unsigned long len); | |
475 | void memmove_extent_buffer(struct extent_buffer *dst, unsigned long dst_offset, | |
476 | unsigned long src_offset, unsigned long len); | |
b159fa28 DS |
477 | void memzero_extent_buffer(struct extent_buffer *eb, unsigned long start, |
478 | unsigned long len); | |
3e1e8bb7 OS |
479 | int extent_buffer_test_bit(struct extent_buffer *eb, unsigned long start, |
480 | unsigned long pos); | |
481 | void extent_buffer_bitmap_set(struct extent_buffer *eb, unsigned long start, | |
482 | unsigned long pos, unsigned long len); | |
483 | void extent_buffer_bitmap_clear(struct extent_buffer *eb, unsigned long start, | |
484 | unsigned long pos, unsigned long len); | |
1d4284bd | 485 | void clear_extent_buffer_dirty(struct extent_buffer *eb); |
abb57ef3 | 486 | bool set_extent_buffer_dirty(struct extent_buffer *eb); |
09c25a8c | 487 | void set_extent_buffer_uptodate(struct extent_buffer *eb); |
69ba3927 | 488 | void clear_extent_buffer_uptodate(struct extent_buffer *eb); |
a26e8c9f | 489 | int extent_buffer_under_io(struct extent_buffer *eb); |
1cbb1f45 JM |
490 | int map_private_extent_buffer(const struct extent_buffer *eb, |
491 | unsigned long offset, unsigned long min_len, | |
492 | char **map, unsigned long *map_start, | |
493 | unsigned long *map_len); | |
bd1fa4f0 | 494 | void extent_range_clear_dirty_for_io(struct inode *inode, u64 start, u64 end); |
f6311572 | 495 | void extent_range_redirty_for_io(struct inode *inode, u64 start, u64 end); |
a9d93e17 | 496 | void extent_clear_unlock_delalloc(struct inode *inode, u64 start, u64 end, |
ba8b04c1 | 497 | u64 delalloc_end, struct page *locked_page, |
9ee49a04 | 498 | unsigned bits_to_clear, |
c2790a2e | 499 | unsigned long page_ops); |
c821e7f3 | 500 | struct bio *btrfs_bio_alloc(struct block_device *bdev, u64 first_byte); |
c5e4c3d7 | 501 | struct bio *btrfs_io_bio_alloc(unsigned int nr_iovecs); |
8b6c1d56 | 502 | struct bio *btrfs_bio_clone(struct bio *bio); |
e477094f | 503 | struct bio *btrfs_bio_clone_partial(struct bio *orig, int offset, int size); |
4a54c8c1 | 504 | |
3ec706c8 | 505 | struct btrfs_fs_info; |
9d4f7f8a | 506 | struct btrfs_inode; |
4a54c8c1 | 507 | |
6ec656bc JB |
508 | int repair_io_failure(struct btrfs_fs_info *fs_info, u64 ino, u64 start, |
509 | u64 length, u64 logical, struct page *page, | |
510 | unsigned int pg_offset, int mirror_num); | |
7870d082 JB |
511 | int clean_io_failure(struct btrfs_fs_info *fs_info, |
512 | struct extent_io_tree *failure_tree, | |
513 | struct extent_io_tree *io_tree, u64 start, | |
514 | struct page *page, u64 ino, unsigned int pg_offset); | |
b5227c07 | 515 | void end_extent_writepage(struct page *page, int err, u64 start, u64 end); |
20a1fbf9 | 516 | int btrfs_repair_eb_io_failure(struct extent_buffer *eb, int mirror_num); |
2fe6303e MX |
517 | |
518 | /* | |
519 | * When IO fails, either with EIO or csum verification fails, we | |
520 | * try other mirrors that might have a good copy of the data. This | |
521 | * io_failure_record is used to record state as we go through all the | |
522 | * mirrors. If another mirror has good data, the page is set up to date | |
523 | * and things continue. If a good mirror can't be found, the original | |
524 | * bio end_io callback is called to indicate things have failed. | |
525 | */ | |
526 | struct io_failure_record { | |
527 | struct page *page; | |
528 | u64 start; | |
529 | u64 len; | |
530 | u64 logical; | |
531 | unsigned long bio_flags; | |
532 | int this_mirror; | |
533 | int failed_mirror; | |
534 | int in_validation; | |
535 | }; | |
536 | ||
4ac1f4ac | 537 | |
7ab7956e NB |
538 | void btrfs_free_io_failure_record(struct btrfs_inode *inode, u64 start, |
539 | u64 end); | |
2fe6303e MX |
540 | int btrfs_get_io_failure_record(struct inode *inode, u64 start, u64 end, |
541 | struct io_failure_record **failrec_ret); | |
a0b60d72 | 542 | bool btrfs_check_repairable(struct inode *inode, unsigned failed_bio_pages, |
c3cfb656 | 543 | struct io_failure_record *failrec, int fail_mirror); |
2fe6303e MX |
544 | struct bio *btrfs_create_repair_bio(struct inode *inode, struct bio *failed_bio, |
545 | struct io_failure_record *failrec, | |
546 | struct page *page, int pg_offset, int icsum, | |
8b110e39 | 547 | bio_end_io_t *endio_func, void *data); |
7870d082 JB |
548 | int free_io_failure(struct extent_io_tree *failure_tree, |
549 | struct extent_io_tree *io_tree, | |
550 | struct io_failure_record *rec); | |
294e30fe | 551 | #ifdef CONFIG_BTRFS_FS_RUN_SANITY_TESTS |
3522e903 | 552 | bool find_lock_delalloc_range(struct inode *inode, struct extent_io_tree *tree, |
ce9f967f JT |
553 | struct page *locked_page, u64 *start, |
554 | u64 *end); | |
0d4cf4e6 | 555 | #endif |
faa2dbf0 | 556 | struct extent_buffer *alloc_test_extent_buffer(struct btrfs_fs_info *fs_info, |
da17066c | 557 | u64 start); |
9888c340 | 558 | |
294e30fe | 559 | #endif |