1 // SPDX-License-Identifier: GPL-2.0
3 * Copyright (C) 2007 Oracle. All rights reserved.
7 #include <linux/slab.h>
8 #include <linux/pagemap.h>
9 #include <linux/highmem.h>
10 #include <linux/sched/mm.h>
11 #include <crypto/hash.h>
14 #include "transaction.h"
16 #include "print-tree.h"
17 #include "compression.h"
19 #define __MAX_CSUM_ITEMS(r, size) ((unsigned long)(((BTRFS_LEAF_DATA_SIZE(r) - \
20 sizeof(struct btrfs_item) * 2) / \
23 #define MAX_CSUM_ITEMS(r, size) (min_t(u32, __MAX_CSUM_ITEMS(r, size), \
26 static inline u32
max_ordered_sum_bytes(struct btrfs_fs_info
*fs_info
,
29 u32 ncsums
= (PAGE_SIZE
- sizeof(struct btrfs_ordered_sum
)) / csum_size
;
31 return ncsums
* fs_info
->sectorsize
;
34 int btrfs_insert_file_extent(struct btrfs_trans_handle
*trans
,
35 struct btrfs_root
*root
,
36 u64 objectid
, u64 pos
,
37 u64 disk_offset
, u64 disk_num_bytes
,
38 u64 num_bytes
, u64 offset
, u64 ram_bytes
,
39 u8 compression
, u8 encryption
, u16 other_encoding
)
42 struct btrfs_file_extent_item
*item
;
43 struct btrfs_key file_key
;
44 struct btrfs_path
*path
;
45 struct extent_buffer
*leaf
;
47 path
= btrfs_alloc_path();
50 file_key
.objectid
= objectid
;
51 file_key
.offset
= pos
;
52 file_key
.type
= BTRFS_EXTENT_DATA_KEY
;
54 path
->leave_spinning
= 1;
55 ret
= btrfs_insert_empty_item(trans
, root
, path
, &file_key
,
59 BUG_ON(ret
); /* Can't happen */
60 leaf
= path
->nodes
[0];
61 item
= btrfs_item_ptr(leaf
, path
->slots
[0],
62 struct btrfs_file_extent_item
);
63 btrfs_set_file_extent_disk_bytenr(leaf
, item
, disk_offset
);
64 btrfs_set_file_extent_disk_num_bytes(leaf
, item
, disk_num_bytes
);
65 btrfs_set_file_extent_offset(leaf
, item
, offset
);
66 btrfs_set_file_extent_num_bytes(leaf
, item
, num_bytes
);
67 btrfs_set_file_extent_ram_bytes(leaf
, item
, ram_bytes
);
68 btrfs_set_file_extent_generation(leaf
, item
, trans
->transid
);
69 btrfs_set_file_extent_type(leaf
, item
, BTRFS_FILE_EXTENT_REG
);
70 btrfs_set_file_extent_compression(leaf
, item
, compression
);
71 btrfs_set_file_extent_encryption(leaf
, item
, encryption
);
72 btrfs_set_file_extent_other_encoding(leaf
, item
, other_encoding
);
74 btrfs_mark_buffer_dirty(leaf
);
76 btrfs_free_path(path
);
80 static struct btrfs_csum_item
*
81 btrfs_lookup_csum(struct btrfs_trans_handle
*trans
,
82 struct btrfs_root
*root
,
83 struct btrfs_path
*path
,
86 struct btrfs_fs_info
*fs_info
= root
->fs_info
;
88 struct btrfs_key file_key
;
89 struct btrfs_key found_key
;
90 struct btrfs_csum_item
*item
;
91 struct extent_buffer
*leaf
;
93 u16 csum_size
= btrfs_super_csum_size(fs_info
->super_copy
);
96 file_key
.objectid
= BTRFS_EXTENT_CSUM_OBJECTID
;
97 file_key
.offset
= bytenr
;
98 file_key
.type
= BTRFS_EXTENT_CSUM_KEY
;
99 ret
= btrfs_search_slot(trans
, root
, &file_key
, path
, 0, cow
);
102 leaf
= path
->nodes
[0];
105 if (path
->slots
[0] == 0)
108 btrfs_item_key_to_cpu(leaf
, &found_key
, path
->slots
[0]);
109 if (found_key
.type
!= BTRFS_EXTENT_CSUM_KEY
)
112 csum_offset
= (bytenr
- found_key
.offset
) >>
113 fs_info
->sb
->s_blocksize_bits
;
114 csums_in_item
= btrfs_item_size_nr(leaf
, path
->slots
[0]);
115 csums_in_item
/= csum_size
;
117 if (csum_offset
== csums_in_item
) {
120 } else if (csum_offset
> csums_in_item
) {
124 item
= btrfs_item_ptr(leaf
, path
->slots
[0], struct btrfs_csum_item
);
125 item
= (struct btrfs_csum_item
*)((unsigned char *)item
+
126 csum_offset
* csum_size
);
134 int btrfs_lookup_file_extent(struct btrfs_trans_handle
*trans
,
135 struct btrfs_root
*root
,
136 struct btrfs_path
*path
, u64 objectid
,
140 struct btrfs_key file_key
;
141 int ins_len
= mod
< 0 ? -1 : 0;
144 file_key
.objectid
= objectid
;
145 file_key
.offset
= offset
;
146 file_key
.type
= BTRFS_EXTENT_DATA_KEY
;
147 ret
= btrfs_search_slot(trans
, root
, &file_key
, path
, ins_len
, cow
);
151 static blk_status_t
__btrfs_lookup_bio_sums(struct inode
*inode
, struct bio
*bio
,
152 u64 logical_offset
, u8
*dst
, int dio
)
154 struct btrfs_fs_info
*fs_info
= btrfs_sb(inode
->i_sb
);
156 struct bvec_iter iter
;
157 struct btrfs_io_bio
*btrfs_bio
= btrfs_io_bio(bio
);
158 struct btrfs_csum_item
*item
= NULL
;
159 struct extent_io_tree
*io_tree
= &BTRFS_I(inode
)->io_tree
;
160 struct btrfs_path
*path
;
163 u64 item_start_offset
= 0;
164 u64 item_last_offset
= 0;
170 u16 csum_size
= btrfs_super_csum_size(fs_info
->super_copy
);
172 path
= btrfs_alloc_path();
174 return BLK_STS_RESOURCE
;
176 nblocks
= bio
->bi_iter
.bi_size
>> inode
->i_sb
->s_blocksize_bits
;
178 if (nblocks
* csum_size
> BTRFS_BIO_INLINE_CSUM_SIZE
) {
179 btrfs_bio
->csum
= kmalloc_array(nblocks
, csum_size
,
181 if (!btrfs_bio
->csum
) {
182 btrfs_free_path(path
);
183 return BLK_STS_RESOURCE
;
186 btrfs_bio
->csum
= btrfs_bio
->csum_inline
;
188 csum
= btrfs_bio
->csum
;
193 if (bio
->bi_iter
.bi_size
> PAGE_SIZE
* 8)
194 path
->reada
= READA_FORWARD
;
197 * the free space stuff is only read when it hasn't been
198 * updated in the current transaction. So, we can safely
199 * read from the commit root and sidestep a nasty deadlock
200 * between reading the free space cache and updating the csum tree.
202 if (btrfs_is_free_space_inode(BTRFS_I(inode
))) {
203 path
->search_commit_root
= 1;
204 path
->skip_locking
= 1;
207 disk_bytenr
= (u64
)bio
->bi_iter
.bi_sector
<< 9;
209 offset
= logical_offset
;
211 bio_for_each_segment(bvec
, bio
, iter
) {
212 page_bytes_left
= bvec
.bv_len
;
217 offset
= page_offset(bvec
.bv_page
) + bvec
.bv_offset
;
218 count
= btrfs_find_ordered_sum(inode
, offset
, disk_bytenr
,
223 if (!item
|| disk_bytenr
< item_start_offset
||
224 disk_bytenr
>= item_last_offset
) {
225 struct btrfs_key found_key
;
229 btrfs_release_path(path
);
230 item
= btrfs_lookup_csum(NULL
, fs_info
->csum_root
,
231 path
, disk_bytenr
, 0);
234 memset(csum
, 0, csum_size
);
235 if (BTRFS_I(inode
)->root
->root_key
.objectid
==
236 BTRFS_DATA_RELOC_TREE_OBJECTID
) {
237 set_extent_bits(io_tree
, offset
,
238 offset
+ fs_info
->sectorsize
- 1,
241 btrfs_info_rl(fs_info
,
242 "no csum found for inode %llu start %llu",
243 btrfs_ino(BTRFS_I(inode
)), offset
);
246 btrfs_release_path(path
);
249 btrfs_item_key_to_cpu(path
->nodes
[0], &found_key
,
252 item_start_offset
= found_key
.offset
;
253 item_size
= btrfs_item_size_nr(path
->nodes
[0],
255 item_last_offset
= item_start_offset
+
256 (item_size
/ csum_size
) *
258 item
= btrfs_item_ptr(path
->nodes
[0], path
->slots
[0],
259 struct btrfs_csum_item
);
262 * this byte range must be able to fit inside
263 * a single leaf so it will also fit inside a u32
265 diff
= disk_bytenr
- item_start_offset
;
266 diff
= diff
/ fs_info
->sectorsize
;
267 diff
= diff
* csum_size
;
268 count
= min_t(int, nblocks
, (item_last_offset
- disk_bytenr
) >>
269 inode
->i_sb
->s_blocksize_bits
);
270 read_extent_buffer(path
->nodes
[0], csum
,
271 ((unsigned long)item
) + diff
,
274 csum
+= count
* csum_size
;
278 disk_bytenr
+= fs_info
->sectorsize
;
279 offset
+= fs_info
->sectorsize
;
280 page_bytes_left
-= fs_info
->sectorsize
;
281 if (!page_bytes_left
)
282 break; /* move to next bio */
287 btrfs_free_path(path
);
291 blk_status_t
btrfs_lookup_bio_sums(struct inode
*inode
, struct bio
*bio
,
294 return __btrfs_lookup_bio_sums(inode
, bio
, 0, dst
, 0);
297 blk_status_t
btrfs_lookup_bio_sums_dio(struct inode
*inode
, struct bio
*bio
, u64 offset
)
299 return __btrfs_lookup_bio_sums(inode
, bio
, offset
, NULL
, 1);
302 int btrfs_lookup_csums_range(struct btrfs_root
*root
, u64 start
, u64 end
,
303 struct list_head
*list
, int search_commit
)
305 struct btrfs_fs_info
*fs_info
= root
->fs_info
;
306 struct btrfs_key key
;
307 struct btrfs_path
*path
;
308 struct extent_buffer
*leaf
;
309 struct btrfs_ordered_sum
*sums
;
310 struct btrfs_csum_item
*item
;
312 unsigned long offset
;
316 u16 csum_size
= btrfs_super_csum_size(fs_info
->super_copy
);
318 ASSERT(IS_ALIGNED(start
, fs_info
->sectorsize
) &&
319 IS_ALIGNED(end
+ 1, fs_info
->sectorsize
));
321 path
= btrfs_alloc_path();
326 path
->skip_locking
= 1;
327 path
->reada
= READA_FORWARD
;
328 path
->search_commit_root
= 1;
331 key
.objectid
= BTRFS_EXTENT_CSUM_OBJECTID
;
333 key
.type
= BTRFS_EXTENT_CSUM_KEY
;
335 ret
= btrfs_search_slot(NULL
, root
, &key
, path
, 0, 0);
338 if (ret
> 0 && path
->slots
[0] > 0) {
339 leaf
= path
->nodes
[0];
340 btrfs_item_key_to_cpu(leaf
, &key
, path
->slots
[0] - 1);
341 if (key
.objectid
== BTRFS_EXTENT_CSUM_OBJECTID
&&
342 key
.type
== BTRFS_EXTENT_CSUM_KEY
) {
343 offset
= (start
- key
.offset
) >>
344 fs_info
->sb
->s_blocksize_bits
;
345 if (offset
* csum_size
<
346 btrfs_item_size_nr(leaf
, path
->slots
[0] - 1))
351 while (start
<= end
) {
352 leaf
= path
->nodes
[0];
353 if (path
->slots
[0] >= btrfs_header_nritems(leaf
)) {
354 ret
= btrfs_next_leaf(root
, path
);
359 leaf
= path
->nodes
[0];
362 btrfs_item_key_to_cpu(leaf
, &key
, path
->slots
[0]);
363 if (key
.objectid
!= BTRFS_EXTENT_CSUM_OBJECTID
||
364 key
.type
!= BTRFS_EXTENT_CSUM_KEY
||
368 if (key
.offset
> start
)
371 size
= btrfs_item_size_nr(leaf
, path
->slots
[0]);
372 csum_end
= key
.offset
+ (size
/ csum_size
) * fs_info
->sectorsize
;
373 if (csum_end
<= start
) {
378 csum_end
= min(csum_end
, end
+ 1);
379 item
= btrfs_item_ptr(path
->nodes
[0], path
->slots
[0],
380 struct btrfs_csum_item
);
381 while (start
< csum_end
) {
382 size
= min_t(size_t, csum_end
- start
,
383 max_ordered_sum_bytes(fs_info
, csum_size
));
384 sums
= kzalloc(btrfs_ordered_sum_size(fs_info
, size
),
391 sums
->bytenr
= start
;
392 sums
->len
= (int)size
;
394 offset
= (start
- key
.offset
) >>
395 fs_info
->sb
->s_blocksize_bits
;
397 size
>>= fs_info
->sb
->s_blocksize_bits
;
399 read_extent_buffer(path
->nodes
[0],
401 ((unsigned long)item
) + offset
,
404 start
+= fs_info
->sectorsize
* size
;
405 list_add_tail(&sums
->list
, &tmplist
);
411 while (ret
< 0 && !list_empty(&tmplist
)) {
412 sums
= list_entry(tmplist
.next
, struct btrfs_ordered_sum
, list
);
413 list_del(&sums
->list
);
416 list_splice_tail(&tmplist
, list
);
418 btrfs_free_path(path
);
423 * btrfs_csum_one_bio - Calculates checksums of the data contained inside a bio
424 * @inode: Owner of the data inside the bio
425 * @bio: Contains the data to be checksummed
426 * @file_start: offset in file this bio begins to describe
427 * @contig: Boolean. If true/1 means all bio vecs in this bio are
428 * contiguous and they begin at @file_start in the file. False/0
429 * means this bio can contains potentially discontigous bio vecs
430 * so the logical offset of each should be calculated separately.
432 blk_status_t
btrfs_csum_one_bio(struct inode
*inode
, struct bio
*bio
,
433 u64 file_start
, int contig
)
435 struct btrfs_fs_info
*fs_info
= btrfs_sb(inode
->i_sb
);
436 SHASH_DESC_ON_STACK(shash
, fs_info
->csum_shash
);
437 struct btrfs_ordered_sum
*sums
;
438 struct btrfs_ordered_extent
*ordered
= NULL
;
440 struct bvec_iter iter
;
444 unsigned long total_bytes
= 0;
445 unsigned long this_sum_bytes
= 0;
449 const u16 csum_size
= btrfs_super_csum_size(fs_info
->super_copy
);
451 nofs_flag
= memalloc_nofs_save();
452 sums
= kvzalloc(btrfs_ordered_sum_size(fs_info
, bio
->bi_iter
.bi_size
),
454 memalloc_nofs_restore(nofs_flag
);
457 return BLK_STS_RESOURCE
;
459 sums
->len
= bio
->bi_iter
.bi_size
;
460 INIT_LIST_HEAD(&sums
->list
);
465 offset
= 0; /* shut up gcc */
467 sums
->bytenr
= (u64
)bio
->bi_iter
.bi_sector
<< 9;
470 shash
->tfm
= fs_info
->csum_shash
;
472 bio_for_each_segment(bvec
, bio
, iter
) {
474 offset
= page_offset(bvec
.bv_page
) + bvec
.bv_offset
;
477 ordered
= btrfs_lookup_ordered_extent(inode
, offset
);
478 BUG_ON(!ordered
); /* Logic error */
481 nr_sectors
= BTRFS_BYTES_TO_BLKS(fs_info
,
482 bvec
.bv_len
+ fs_info
->sectorsize
485 for (i
= 0; i
< nr_sectors
; i
++) {
486 if (offset
>= ordered
->file_offset
+ ordered
->len
||
487 offset
< ordered
->file_offset
) {
488 unsigned long bytes_left
;
490 sums
->len
= this_sum_bytes
;
492 btrfs_add_ordered_sum(ordered
, sums
);
493 btrfs_put_ordered_extent(ordered
);
495 bytes_left
= bio
->bi_iter
.bi_size
- total_bytes
;
497 nofs_flag
= memalloc_nofs_save();
498 sums
= kvzalloc(btrfs_ordered_sum_size(fs_info
,
499 bytes_left
), GFP_KERNEL
);
500 memalloc_nofs_restore(nofs_flag
);
501 BUG_ON(!sums
); /* -ENOMEM */
502 sums
->len
= bytes_left
;
503 ordered
= btrfs_lookup_ordered_extent(inode
,
505 ASSERT(ordered
); /* Logic error */
506 sums
->bytenr
= ((u64
)bio
->bi_iter
.bi_sector
<< 9)
511 crypto_shash_init(shash
);
512 data
= kmap_atomic(bvec
.bv_page
);
513 crypto_shash_update(shash
, data
+ bvec
.bv_offset
514 + (i
* fs_info
->sectorsize
),
515 fs_info
->sectorsize
);
517 crypto_shash_final(shash
, (char *)(sums
->sums
+ index
));
519 offset
+= fs_info
->sectorsize
;
520 this_sum_bytes
+= fs_info
->sectorsize
;
521 total_bytes
+= fs_info
->sectorsize
;
526 btrfs_add_ordered_sum(ordered
, sums
);
527 btrfs_put_ordered_extent(ordered
);
532 * helper function for csum removal, this expects the
533 * key to describe the csum pointed to by the path, and it expects
534 * the csum to overlap the range [bytenr, len]
536 * The csum should not be entirely contained in the range and the
537 * range should not be entirely contained in the csum.
539 * This calls btrfs_truncate_item with the correct args based on the
540 * overlap, and fixes up the key as required.
542 static noinline
void truncate_one_csum(struct btrfs_fs_info
*fs_info
,
543 struct btrfs_path
*path
,
544 struct btrfs_key
*key
,
547 struct extent_buffer
*leaf
;
548 u16 csum_size
= btrfs_super_csum_size(fs_info
->super_copy
);
550 u64 end_byte
= bytenr
+ len
;
551 u32 blocksize_bits
= fs_info
->sb
->s_blocksize_bits
;
553 leaf
= path
->nodes
[0];
554 csum_end
= btrfs_item_size_nr(leaf
, path
->slots
[0]) / csum_size
;
555 csum_end
<<= fs_info
->sb
->s_blocksize_bits
;
556 csum_end
+= key
->offset
;
558 if (key
->offset
< bytenr
&& csum_end
<= end_byte
) {
563 * A simple truncate off the end of the item
565 u32 new_size
= (bytenr
- key
->offset
) >> blocksize_bits
;
566 new_size
*= csum_size
;
567 btrfs_truncate_item(path
, new_size
, 1);
568 } else if (key
->offset
>= bytenr
&& csum_end
> end_byte
&&
569 end_byte
> key
->offset
) {
574 * we need to truncate from the beginning of the csum
576 u32 new_size
= (csum_end
- end_byte
) >> blocksize_bits
;
577 new_size
*= csum_size
;
579 btrfs_truncate_item(path
, new_size
, 0);
581 key
->offset
= end_byte
;
582 btrfs_set_item_key_safe(fs_info
, path
, key
);
589 * deletes the csum items from the csum tree for a given
592 int btrfs_del_csums(struct btrfs_trans_handle
*trans
,
593 struct btrfs_fs_info
*fs_info
, u64 bytenr
, u64 len
)
595 struct btrfs_root
*root
= fs_info
->csum_root
;
596 struct btrfs_path
*path
;
597 struct btrfs_key key
;
598 u64 end_byte
= bytenr
+ len
;
600 struct extent_buffer
*leaf
;
602 u16 csum_size
= btrfs_super_csum_size(fs_info
->super_copy
);
603 int blocksize_bits
= fs_info
->sb
->s_blocksize_bits
;
605 path
= btrfs_alloc_path();
610 key
.objectid
= BTRFS_EXTENT_CSUM_OBJECTID
;
611 key
.offset
= end_byte
- 1;
612 key
.type
= BTRFS_EXTENT_CSUM_KEY
;
614 path
->leave_spinning
= 1;
615 ret
= btrfs_search_slot(trans
, root
, &key
, path
, -1, 1);
617 if (path
->slots
[0] == 0)
620 } else if (ret
< 0) {
624 leaf
= path
->nodes
[0];
625 btrfs_item_key_to_cpu(leaf
, &key
, path
->slots
[0]);
627 if (key
.objectid
!= BTRFS_EXTENT_CSUM_OBJECTID
||
628 key
.type
!= BTRFS_EXTENT_CSUM_KEY
) {
632 if (key
.offset
>= end_byte
)
635 csum_end
= btrfs_item_size_nr(leaf
, path
->slots
[0]) / csum_size
;
636 csum_end
<<= blocksize_bits
;
637 csum_end
+= key
.offset
;
639 /* this csum ends before we start, we're done */
640 if (csum_end
<= bytenr
)
643 /* delete the entire item, it is inside our range */
644 if (key
.offset
>= bytenr
&& csum_end
<= end_byte
) {
648 * Check how many csum items preceding this one in this
649 * leaf correspond to our range and then delete them all
652 if (key
.offset
> bytenr
&& path
->slots
[0] > 0) {
653 int slot
= path
->slots
[0] - 1;
658 btrfs_item_key_to_cpu(leaf
, &pk
, slot
);
659 if (pk
.offset
< bytenr
||
660 pk
.type
!= BTRFS_EXTENT_CSUM_KEY
||
662 BTRFS_EXTENT_CSUM_OBJECTID
)
664 path
->slots
[0] = slot
;
666 key
.offset
= pk
.offset
;
670 ret
= btrfs_del_items(trans
, root
, path
,
671 path
->slots
[0], del_nr
);
674 if (key
.offset
== bytenr
)
676 } else if (key
.offset
< bytenr
&& csum_end
> end_byte
) {
677 unsigned long offset
;
678 unsigned long shift_len
;
679 unsigned long item_offset
;
684 * Our bytes are in the middle of the csum,
685 * we need to split this item and insert a new one.
687 * But we can't drop the path because the
688 * csum could change, get removed, extended etc.
690 * The trick here is the max size of a csum item leaves
691 * enough room in the tree block for a single
692 * item header. So, we split the item in place,
693 * adding a new header pointing to the existing
694 * bytes. Then we loop around again and we have
695 * a nicely formed csum item that we can neatly
698 offset
= (bytenr
- key
.offset
) >> blocksize_bits
;
701 shift_len
= (len
>> blocksize_bits
) * csum_size
;
703 item_offset
= btrfs_item_ptr_offset(leaf
,
706 memzero_extent_buffer(leaf
, item_offset
+ offset
,
711 * btrfs_split_item returns -EAGAIN when the
712 * item changed size or key
714 ret
= btrfs_split_item(trans
, root
, path
, &key
, offset
);
715 if (ret
&& ret
!= -EAGAIN
) {
716 btrfs_abort_transaction(trans
, ret
);
720 key
.offset
= end_byte
- 1;
722 truncate_one_csum(fs_info
, path
, &key
, bytenr
, len
);
723 if (key
.offset
< bytenr
)
726 btrfs_release_path(path
);
730 btrfs_free_path(path
);
734 int btrfs_csum_file_blocks(struct btrfs_trans_handle
*trans
,
735 struct btrfs_root
*root
,
736 struct btrfs_ordered_sum
*sums
)
738 struct btrfs_fs_info
*fs_info
= root
->fs_info
;
739 struct btrfs_key file_key
;
740 struct btrfs_key found_key
;
741 struct btrfs_path
*path
;
742 struct btrfs_csum_item
*item
;
743 struct btrfs_csum_item
*item_end
;
744 struct extent_buffer
*leaf
= NULL
;
754 u16 csum_size
= btrfs_super_csum_size(fs_info
->super_copy
);
756 path
= btrfs_alloc_path();
760 next_offset
= (u64
)-1;
762 bytenr
= sums
->bytenr
+ total_bytes
;
763 file_key
.objectid
= BTRFS_EXTENT_CSUM_OBJECTID
;
764 file_key
.offset
= bytenr
;
765 file_key
.type
= BTRFS_EXTENT_CSUM_KEY
;
767 item
= btrfs_lookup_csum(trans
, root
, path
, bytenr
, 1);
770 leaf
= path
->nodes
[0];
771 item_end
= btrfs_item_ptr(leaf
, path
->slots
[0],
772 struct btrfs_csum_item
);
773 item_end
= (struct btrfs_csum_item
*)((char *)item_end
+
774 btrfs_item_size_nr(leaf
, path
->slots
[0]));
778 if (ret
!= -EFBIG
&& ret
!= -ENOENT
)
783 /* we found one, but it isn't big enough yet */
784 leaf
= path
->nodes
[0];
785 item_size
= btrfs_item_size_nr(leaf
, path
->slots
[0]);
786 if ((item_size
/ csum_size
) >=
787 MAX_CSUM_ITEMS(fs_info
, csum_size
)) {
788 /* already at max size, make a new one */
792 int slot
= path
->slots
[0] + 1;
793 /* we didn't find a csum item, insert one */
794 nritems
= btrfs_header_nritems(path
->nodes
[0]);
795 if (!nritems
|| (path
->slots
[0] >= nritems
- 1)) {
796 ret
= btrfs_next_leaf(root
, path
);
801 slot
= path
->slots
[0];
803 btrfs_item_key_to_cpu(path
->nodes
[0], &found_key
, slot
);
804 if (found_key
.objectid
!= BTRFS_EXTENT_CSUM_OBJECTID
||
805 found_key
.type
!= BTRFS_EXTENT_CSUM_KEY
) {
809 next_offset
= found_key
.offset
;
815 * at this point, we know the tree has an item, but it isn't big
816 * enough yet to put our csum in. Grow it
818 btrfs_release_path(path
);
819 ret
= btrfs_search_slot(trans
, root
, &file_key
, path
,
825 if (path
->slots
[0] == 0)
830 leaf
= path
->nodes
[0];
831 btrfs_item_key_to_cpu(leaf
, &found_key
, path
->slots
[0]);
832 csum_offset
= (bytenr
- found_key
.offset
) >>
833 fs_info
->sb
->s_blocksize_bits
;
835 if (found_key
.type
!= BTRFS_EXTENT_CSUM_KEY
||
836 found_key
.objectid
!= BTRFS_EXTENT_CSUM_OBJECTID
||
837 csum_offset
>= MAX_CSUM_ITEMS(fs_info
, csum_size
)) {
841 if (csum_offset
== btrfs_item_size_nr(leaf
, path
->slots
[0]) /
848 if (btrfs_leaf_free_space(leaf
) <
849 sizeof(struct btrfs_item
) + csum_size
* 2)
852 free_space
= btrfs_leaf_free_space(leaf
) -
853 sizeof(struct btrfs_item
) - csum_size
;
854 tmp
= sums
->len
- total_bytes
;
855 tmp
>>= fs_info
->sb
->s_blocksize_bits
;
858 extend_nr
= max_t(int, 1, (int)tmp
);
859 diff
= (csum_offset
+ extend_nr
) * csum_size
;
861 MAX_CSUM_ITEMS(fs_info
, csum_size
) * csum_size
);
863 diff
= diff
- btrfs_item_size_nr(leaf
, path
->slots
[0]);
864 diff
= min(free_space
, diff
);
868 btrfs_extend_item(path
, diff
);
874 btrfs_release_path(path
);
879 tmp
= sums
->len
- total_bytes
;
880 tmp
>>= fs_info
->sb
->s_blocksize_bits
;
881 tmp
= min(tmp
, (next_offset
- file_key
.offset
) >>
882 fs_info
->sb
->s_blocksize_bits
);
884 tmp
= max_t(u64
, 1, tmp
);
885 tmp
= min_t(u64
, tmp
, MAX_CSUM_ITEMS(fs_info
, csum_size
));
886 ins_size
= csum_size
* tmp
;
888 ins_size
= csum_size
;
890 path
->leave_spinning
= 1;
891 ret
= btrfs_insert_empty_item(trans
, root
, path
, &file_key
,
893 path
->leave_spinning
= 0;
896 if (WARN_ON(ret
!= 0))
898 leaf
= path
->nodes
[0];
900 item
= btrfs_item_ptr(leaf
, path
->slots
[0], struct btrfs_csum_item
);
901 item_end
= (struct btrfs_csum_item
*)((unsigned char *)item
+
902 btrfs_item_size_nr(leaf
, path
->slots
[0]));
903 item
= (struct btrfs_csum_item
*)((unsigned char *)item
+
904 csum_offset
* csum_size
);
906 ins_size
= (u32
)(sums
->len
- total_bytes
) >>
907 fs_info
->sb
->s_blocksize_bits
;
908 ins_size
*= csum_size
;
909 ins_size
= min_t(u32
, (unsigned long)item_end
- (unsigned long)item
,
911 write_extent_buffer(leaf
, sums
->sums
+ index
, (unsigned long)item
,
915 ins_size
/= csum_size
;
916 total_bytes
+= ins_size
* fs_info
->sectorsize
;
918 btrfs_mark_buffer_dirty(path
->nodes
[0]);
919 if (total_bytes
< sums
->len
) {
920 btrfs_release_path(path
);
925 btrfs_free_path(path
);
932 void btrfs_extent_item_to_extent_map(struct btrfs_inode
*inode
,
933 const struct btrfs_path
*path
,
934 struct btrfs_file_extent_item
*fi
,
935 const bool new_inline
,
936 struct extent_map
*em
)
938 struct btrfs_fs_info
*fs_info
= inode
->root
->fs_info
;
939 struct btrfs_root
*root
= inode
->root
;
940 struct extent_buffer
*leaf
= path
->nodes
[0];
941 const int slot
= path
->slots
[0];
942 struct btrfs_key key
;
943 u64 extent_start
, extent_end
;
945 u8 type
= btrfs_file_extent_type(leaf
, fi
);
946 int compress_type
= btrfs_file_extent_compression(leaf
, fi
);
948 em
->bdev
= fs_info
->fs_devices
->latest_bdev
;
949 btrfs_item_key_to_cpu(leaf
, &key
, slot
);
950 extent_start
= key
.offset
;
952 if (type
== BTRFS_FILE_EXTENT_REG
||
953 type
== BTRFS_FILE_EXTENT_PREALLOC
) {
954 extent_end
= extent_start
+
955 btrfs_file_extent_num_bytes(leaf
, fi
);
956 } else if (type
== BTRFS_FILE_EXTENT_INLINE
) {
958 size
= btrfs_file_extent_ram_bytes(leaf
, fi
);
959 extent_end
= ALIGN(extent_start
+ size
,
960 fs_info
->sectorsize
);
963 em
->ram_bytes
= btrfs_file_extent_ram_bytes(leaf
, fi
);
964 if (type
== BTRFS_FILE_EXTENT_REG
||
965 type
== BTRFS_FILE_EXTENT_PREALLOC
) {
966 em
->start
= extent_start
;
967 em
->len
= extent_end
- extent_start
;
968 em
->orig_start
= extent_start
-
969 btrfs_file_extent_offset(leaf
, fi
);
970 em
->orig_block_len
= btrfs_file_extent_disk_num_bytes(leaf
, fi
);
971 bytenr
= btrfs_file_extent_disk_bytenr(leaf
, fi
);
973 em
->block_start
= EXTENT_MAP_HOLE
;
976 if (compress_type
!= BTRFS_COMPRESS_NONE
) {
977 set_bit(EXTENT_FLAG_COMPRESSED
, &em
->flags
);
978 em
->compress_type
= compress_type
;
979 em
->block_start
= bytenr
;
980 em
->block_len
= em
->orig_block_len
;
982 bytenr
+= btrfs_file_extent_offset(leaf
, fi
);
983 em
->block_start
= bytenr
;
984 em
->block_len
= em
->len
;
985 if (type
== BTRFS_FILE_EXTENT_PREALLOC
)
986 set_bit(EXTENT_FLAG_PREALLOC
, &em
->flags
);
988 } else if (type
== BTRFS_FILE_EXTENT_INLINE
) {
989 em
->block_start
= EXTENT_MAP_INLINE
;
990 em
->start
= extent_start
;
991 em
->len
= extent_end
- extent_start
;
993 * Initialize orig_start and block_len with the same values
994 * as in inode.c:btrfs_get_extent().
996 em
->orig_start
= EXTENT_MAP_HOLE
;
997 em
->block_len
= (u64
)-1;
998 if (!new_inline
&& compress_type
!= BTRFS_COMPRESS_NONE
) {
999 set_bit(EXTENT_FLAG_COMPRESSED
, &em
->flags
);
1000 em
->compress_type
= compress_type
;
1004 "unknown file extent item type %d, inode %llu, offset %llu, "
1005 "root %llu", type
, btrfs_ino(inode
), extent_start
,
1006 root
->root_key
.objectid
);