]> git.proxmox.com Git - mirror_ubuntu-hirsute-kernel.git/blob - fs/btrfs/zoned.c
Merge tag 'arm64-fixes' of git://git.kernel.org/pub/scm/linux/kernel/git/arm64/linux
[mirror_ubuntu-hirsute-kernel.git] / fs / btrfs / zoned.c
1 // SPDX-License-Identifier: GPL-2.0
2
3 #include <linux/slab.h>
4 #include <linux/blkdev.h>
5 #include "ctree.h"
6 #include "volumes.h"
7 #include "zoned.h"
8 #include "rcu-string.h"
9
10 /* Maximum number of zones to report per blkdev_report_zones() call */
11 #define BTRFS_REPORT_NR_ZONES 4096
12
13 /* Number of superblock log zones */
14 #define BTRFS_NR_SB_LOG_ZONES 2
15
16 static int copy_zone_info_cb(struct blk_zone *zone, unsigned int idx, void *data)
17 {
18 struct blk_zone *zones = data;
19
20 memcpy(&zones[idx], zone, sizeof(*zone));
21
22 return 0;
23 }
24
25 static int sb_write_pointer(struct block_device *bdev, struct blk_zone *zones,
26 u64 *wp_ret)
27 {
28 bool empty[BTRFS_NR_SB_LOG_ZONES];
29 bool full[BTRFS_NR_SB_LOG_ZONES];
30 sector_t sector;
31
32 ASSERT(zones[0].type != BLK_ZONE_TYPE_CONVENTIONAL &&
33 zones[1].type != BLK_ZONE_TYPE_CONVENTIONAL);
34
35 empty[0] = (zones[0].cond == BLK_ZONE_COND_EMPTY);
36 empty[1] = (zones[1].cond == BLK_ZONE_COND_EMPTY);
37 full[0] = (zones[0].cond == BLK_ZONE_COND_FULL);
38 full[1] = (zones[1].cond == BLK_ZONE_COND_FULL);
39
40 /*
41 * Possible states of log buffer zones
42 *
43 * Empty[0] In use[0] Full[0]
44 * Empty[1] * x 0
45 * In use[1] 0 x 0
46 * Full[1] 1 1 C
47 *
48 * Log position:
49 * *: Special case, no superblock is written
50 * 0: Use write pointer of zones[0]
51 * 1: Use write pointer of zones[1]
52 * C: Compare super blcoks from zones[0] and zones[1], use the latest
53 * one determined by generation
54 * x: Invalid state
55 */
56
57 if (empty[0] && empty[1]) {
58 /* Special case to distinguish no superblock to read */
59 *wp_ret = zones[0].start << SECTOR_SHIFT;
60 return -ENOENT;
61 } else if (full[0] && full[1]) {
62 /* Compare two super blocks */
63 struct address_space *mapping = bdev->bd_inode->i_mapping;
64 struct page *page[BTRFS_NR_SB_LOG_ZONES];
65 struct btrfs_super_block *super[BTRFS_NR_SB_LOG_ZONES];
66 int i;
67
68 for (i = 0; i < BTRFS_NR_SB_LOG_ZONES; i++) {
69 u64 bytenr;
70
71 bytenr = ((zones[i].start + zones[i].len)
72 << SECTOR_SHIFT) - BTRFS_SUPER_INFO_SIZE;
73
74 page[i] = read_cache_page_gfp(mapping,
75 bytenr >> PAGE_SHIFT, GFP_NOFS);
76 if (IS_ERR(page[i])) {
77 if (i == 1)
78 btrfs_release_disk_super(super[0]);
79 return PTR_ERR(page[i]);
80 }
81 super[i] = page_address(page[i]);
82 }
83
84 if (super[0]->generation > super[1]->generation)
85 sector = zones[1].start;
86 else
87 sector = zones[0].start;
88
89 for (i = 0; i < BTRFS_NR_SB_LOG_ZONES; i++)
90 btrfs_release_disk_super(super[i]);
91 } else if (!full[0] && (empty[1] || full[1])) {
92 sector = zones[0].wp;
93 } else if (full[0]) {
94 sector = zones[1].wp;
95 } else {
96 return -EUCLEAN;
97 }
98 *wp_ret = sector << SECTOR_SHIFT;
99 return 0;
100 }
101
102 /*
103 * The following zones are reserved as the circular buffer on ZONED btrfs.
104 * - The primary superblock: zones 0 and 1
105 * - The first copy: zones 16 and 17
106 * - The second copy: zones 1024 or zone at 256GB which is minimum, and
107 * the following one
108 */
109 static inline u32 sb_zone_number(int shift, int mirror)
110 {
111 ASSERT(mirror < BTRFS_SUPER_MIRROR_MAX);
112
113 switch (mirror) {
114 case 0: return 0;
115 case 1: return 16;
116 case 2: return min_t(u64, btrfs_sb_offset(mirror) >> shift, 1024);
117 }
118
119 return 0;
120 }
121
122 static int btrfs_get_dev_zones(struct btrfs_device *device, u64 pos,
123 struct blk_zone *zones, unsigned int *nr_zones)
124 {
125 int ret;
126
127 if (!*nr_zones)
128 return 0;
129
130 ret = blkdev_report_zones(device->bdev, pos >> SECTOR_SHIFT, *nr_zones,
131 copy_zone_info_cb, zones);
132 if (ret < 0) {
133 btrfs_err_in_rcu(device->fs_info,
134 "zoned: failed to read zone %llu on %s (devid %llu)",
135 pos, rcu_str_deref(device->name),
136 device->devid);
137 return ret;
138 }
139 *nr_zones = ret;
140 if (!ret)
141 return -EIO;
142
143 return 0;
144 }
145
146 int btrfs_get_dev_zone_info(struct btrfs_device *device)
147 {
148 struct btrfs_zoned_device_info *zone_info = NULL;
149 struct block_device *bdev = device->bdev;
150 struct request_queue *queue = bdev_get_queue(bdev);
151 sector_t nr_sectors;
152 sector_t sector = 0;
153 struct blk_zone *zones = NULL;
154 unsigned int i, nreported = 0, nr_zones;
155 unsigned int zone_sectors;
156 int ret;
157
158 if (!bdev_is_zoned(bdev))
159 return 0;
160
161 if (device->zone_info)
162 return 0;
163
164 zone_info = kzalloc(sizeof(*zone_info), GFP_KERNEL);
165 if (!zone_info)
166 return -ENOMEM;
167
168 nr_sectors = bdev_nr_sectors(bdev);
169 zone_sectors = bdev_zone_sectors(bdev);
170 /* Check if it's power of 2 (see is_power_of_2) */
171 ASSERT(zone_sectors != 0 && (zone_sectors & (zone_sectors - 1)) == 0);
172 zone_info->zone_size = zone_sectors << SECTOR_SHIFT;
173 zone_info->zone_size_shift = ilog2(zone_info->zone_size);
174 zone_info->max_zone_append_size =
175 (u64)queue_max_zone_append_sectors(queue) << SECTOR_SHIFT;
176 zone_info->nr_zones = nr_sectors >> ilog2(zone_sectors);
177 if (!IS_ALIGNED(nr_sectors, zone_sectors))
178 zone_info->nr_zones++;
179
180 zone_info->seq_zones = bitmap_zalloc(zone_info->nr_zones, GFP_KERNEL);
181 if (!zone_info->seq_zones) {
182 ret = -ENOMEM;
183 goto out;
184 }
185
186 zone_info->empty_zones = bitmap_zalloc(zone_info->nr_zones, GFP_KERNEL);
187 if (!zone_info->empty_zones) {
188 ret = -ENOMEM;
189 goto out;
190 }
191
192 zones = kcalloc(BTRFS_REPORT_NR_ZONES, sizeof(struct blk_zone), GFP_KERNEL);
193 if (!zones) {
194 ret = -ENOMEM;
195 goto out;
196 }
197
198 /* Get zones type */
199 while (sector < nr_sectors) {
200 nr_zones = BTRFS_REPORT_NR_ZONES;
201 ret = btrfs_get_dev_zones(device, sector << SECTOR_SHIFT, zones,
202 &nr_zones);
203 if (ret)
204 goto out;
205
206 for (i = 0; i < nr_zones; i++) {
207 if (zones[i].type == BLK_ZONE_TYPE_SEQWRITE_REQ)
208 __set_bit(nreported, zone_info->seq_zones);
209 if (zones[i].cond == BLK_ZONE_COND_EMPTY)
210 __set_bit(nreported, zone_info->empty_zones);
211 nreported++;
212 }
213 sector = zones[nr_zones - 1].start + zones[nr_zones - 1].len;
214 }
215
216 if (nreported != zone_info->nr_zones) {
217 btrfs_err_in_rcu(device->fs_info,
218 "inconsistent number of zones on %s (%u/%u)",
219 rcu_str_deref(device->name), nreported,
220 zone_info->nr_zones);
221 ret = -EIO;
222 goto out;
223 }
224
225 /* Validate superblock log */
226 nr_zones = BTRFS_NR_SB_LOG_ZONES;
227 for (i = 0; i < BTRFS_SUPER_MIRROR_MAX; i++) {
228 u32 sb_zone;
229 u64 sb_wp;
230 int sb_pos = BTRFS_NR_SB_LOG_ZONES * i;
231
232 sb_zone = sb_zone_number(zone_info->zone_size_shift, i);
233 if (sb_zone + 1 >= zone_info->nr_zones)
234 continue;
235
236 sector = sb_zone << (zone_info->zone_size_shift - SECTOR_SHIFT);
237 ret = btrfs_get_dev_zones(device, sector << SECTOR_SHIFT,
238 &zone_info->sb_zones[sb_pos],
239 &nr_zones);
240 if (ret)
241 goto out;
242
243 if (nr_zones != BTRFS_NR_SB_LOG_ZONES) {
244 btrfs_err_in_rcu(device->fs_info,
245 "zoned: failed to read super block log zone info at devid %llu zone %u",
246 device->devid, sb_zone);
247 ret = -EUCLEAN;
248 goto out;
249 }
250
251 /*
252 * If zones[0] is conventional, always use the beggining of the
253 * zone to record superblock. No need to validate in that case.
254 */
255 if (zone_info->sb_zones[BTRFS_NR_SB_LOG_ZONES * i].type ==
256 BLK_ZONE_TYPE_CONVENTIONAL)
257 continue;
258
259 ret = sb_write_pointer(device->bdev,
260 &zone_info->sb_zones[sb_pos], &sb_wp);
261 if (ret != -ENOENT && ret) {
262 btrfs_err_in_rcu(device->fs_info,
263 "zoned: super block log zone corrupted devid %llu zone %u",
264 device->devid, sb_zone);
265 ret = -EUCLEAN;
266 goto out;
267 }
268 }
269
270
271 kfree(zones);
272
273 device->zone_info = zone_info;
274
275 /* device->fs_info is not safe to use for printing messages */
276 btrfs_info_in_rcu(NULL,
277 "host-%s zoned block device %s, %u zones of %llu bytes",
278 bdev_zoned_model(bdev) == BLK_ZONED_HM ? "managed" : "aware",
279 rcu_str_deref(device->name), zone_info->nr_zones,
280 zone_info->zone_size);
281
282 return 0;
283
284 out:
285 kfree(zones);
286 bitmap_free(zone_info->empty_zones);
287 bitmap_free(zone_info->seq_zones);
288 kfree(zone_info);
289
290 return ret;
291 }
292
293 void btrfs_destroy_dev_zone_info(struct btrfs_device *device)
294 {
295 struct btrfs_zoned_device_info *zone_info = device->zone_info;
296
297 if (!zone_info)
298 return;
299
300 bitmap_free(zone_info->seq_zones);
301 bitmap_free(zone_info->empty_zones);
302 kfree(zone_info);
303 device->zone_info = NULL;
304 }
305
306 int btrfs_get_dev_zone(struct btrfs_device *device, u64 pos,
307 struct blk_zone *zone)
308 {
309 unsigned int nr_zones = 1;
310 int ret;
311
312 ret = btrfs_get_dev_zones(device, pos, zone, &nr_zones);
313 if (ret != 0 || !nr_zones)
314 return ret ? ret : -EIO;
315
316 return 0;
317 }
318
319 int btrfs_check_zoned_mode(struct btrfs_fs_info *fs_info)
320 {
321 struct btrfs_fs_devices *fs_devices = fs_info->fs_devices;
322 struct btrfs_device *device;
323 u64 zoned_devices = 0;
324 u64 nr_devices = 0;
325 u64 zone_size = 0;
326 u64 max_zone_append_size = 0;
327 const bool incompat_zoned = btrfs_is_zoned(fs_info);
328 int ret = 0;
329
330 /* Count zoned devices */
331 list_for_each_entry(device, &fs_devices->devices, dev_list) {
332 enum blk_zoned_model model;
333
334 if (!device->bdev)
335 continue;
336
337 model = bdev_zoned_model(device->bdev);
338 if (model == BLK_ZONED_HM ||
339 (model == BLK_ZONED_HA && incompat_zoned)) {
340 struct btrfs_zoned_device_info *zone_info;
341
342 zone_info = device->zone_info;
343 zoned_devices++;
344 if (!zone_size) {
345 zone_size = zone_info->zone_size;
346 } else if (zone_info->zone_size != zone_size) {
347 btrfs_err(fs_info,
348 "zoned: unequal block device zone sizes: have %llu found %llu",
349 device->zone_info->zone_size,
350 zone_size);
351 ret = -EINVAL;
352 goto out;
353 }
354 if (!max_zone_append_size ||
355 (zone_info->max_zone_append_size &&
356 zone_info->max_zone_append_size < max_zone_append_size))
357 max_zone_append_size =
358 zone_info->max_zone_append_size;
359 }
360 nr_devices++;
361 }
362
363 if (!zoned_devices && !incompat_zoned)
364 goto out;
365
366 if (!zoned_devices && incompat_zoned) {
367 /* No zoned block device found on ZONED filesystem */
368 btrfs_err(fs_info,
369 "zoned: no zoned devices found on a zoned filesystem");
370 ret = -EINVAL;
371 goto out;
372 }
373
374 if (zoned_devices && !incompat_zoned) {
375 btrfs_err(fs_info,
376 "zoned: mode not enabled but zoned device found");
377 ret = -EINVAL;
378 goto out;
379 }
380
381 if (zoned_devices != nr_devices) {
382 btrfs_err(fs_info,
383 "zoned: cannot mix zoned and regular devices");
384 ret = -EINVAL;
385 goto out;
386 }
387
388 /*
389 * stripe_size is always aligned to BTRFS_STRIPE_LEN in
390 * __btrfs_alloc_chunk(). Since we want stripe_len == zone_size,
391 * check the alignment here.
392 */
393 if (!IS_ALIGNED(zone_size, BTRFS_STRIPE_LEN)) {
394 btrfs_err(fs_info,
395 "zoned: zone size %llu not aligned to stripe %u",
396 zone_size, BTRFS_STRIPE_LEN);
397 ret = -EINVAL;
398 goto out;
399 }
400
401 if (btrfs_fs_incompat(fs_info, MIXED_GROUPS)) {
402 btrfs_err(fs_info, "zoned: mixed block groups not supported");
403 ret = -EINVAL;
404 goto out;
405 }
406
407 fs_info->zone_size = zone_size;
408 fs_info->max_zone_append_size = max_zone_append_size;
409
410 btrfs_info(fs_info, "zoned mode enabled with zone size %llu", zone_size);
411 out:
412 return ret;
413 }
414
415 int btrfs_check_mountopts_zoned(struct btrfs_fs_info *info)
416 {
417 if (!btrfs_is_zoned(info))
418 return 0;
419
420 /*
421 * Space cache writing is not COWed. Disable that to avoid write errors
422 * in sequential zones.
423 */
424 if (btrfs_test_opt(info, SPACE_CACHE)) {
425 btrfs_err(info, "zoned: space cache v1 is not supported");
426 return -EINVAL;
427 }
428
429 if (btrfs_test_opt(info, NODATACOW)) {
430 btrfs_err(info, "zoned: NODATACOW not supported");
431 return -EINVAL;
432 }
433
434 return 0;
435 }
436
437 static int sb_log_location(struct block_device *bdev, struct blk_zone *zones,
438 int rw, u64 *bytenr_ret)
439 {
440 u64 wp;
441 int ret;
442
443 if (zones[0].type == BLK_ZONE_TYPE_CONVENTIONAL) {
444 *bytenr_ret = zones[0].start << SECTOR_SHIFT;
445 return 0;
446 }
447
448 ret = sb_write_pointer(bdev, zones, &wp);
449 if (ret != -ENOENT && ret < 0)
450 return ret;
451
452 if (rw == WRITE) {
453 struct blk_zone *reset = NULL;
454
455 if (wp == zones[0].start << SECTOR_SHIFT)
456 reset = &zones[0];
457 else if (wp == zones[1].start << SECTOR_SHIFT)
458 reset = &zones[1];
459
460 if (reset && reset->cond != BLK_ZONE_COND_EMPTY) {
461 ASSERT(reset->cond == BLK_ZONE_COND_FULL);
462
463 ret = blkdev_zone_mgmt(bdev, REQ_OP_ZONE_RESET,
464 reset->start, reset->len,
465 GFP_NOFS);
466 if (ret)
467 return ret;
468
469 reset->cond = BLK_ZONE_COND_EMPTY;
470 reset->wp = reset->start;
471 }
472 } else if (ret != -ENOENT) {
473 /* For READ, we want the precious one */
474 if (wp == zones[0].start << SECTOR_SHIFT)
475 wp = (zones[1].start + zones[1].len) << SECTOR_SHIFT;
476 wp -= BTRFS_SUPER_INFO_SIZE;
477 }
478
479 *bytenr_ret = wp;
480 return 0;
481
482 }
483
484 int btrfs_sb_log_location_bdev(struct block_device *bdev, int mirror, int rw,
485 u64 *bytenr_ret)
486 {
487 struct blk_zone zones[BTRFS_NR_SB_LOG_ZONES];
488 unsigned int zone_sectors;
489 u32 sb_zone;
490 int ret;
491 u64 zone_size;
492 u8 zone_sectors_shift;
493 sector_t nr_sectors;
494 u32 nr_zones;
495
496 if (!bdev_is_zoned(bdev)) {
497 *bytenr_ret = btrfs_sb_offset(mirror);
498 return 0;
499 }
500
501 ASSERT(rw == READ || rw == WRITE);
502
503 zone_sectors = bdev_zone_sectors(bdev);
504 if (!is_power_of_2(zone_sectors))
505 return -EINVAL;
506 zone_size = zone_sectors << SECTOR_SHIFT;
507 zone_sectors_shift = ilog2(zone_sectors);
508 nr_sectors = bdev_nr_sectors(bdev);
509 nr_zones = nr_sectors >> zone_sectors_shift;
510
511 sb_zone = sb_zone_number(zone_sectors_shift + SECTOR_SHIFT, mirror);
512 if (sb_zone + 1 >= nr_zones)
513 return -ENOENT;
514
515 ret = blkdev_report_zones(bdev, sb_zone << zone_sectors_shift,
516 BTRFS_NR_SB_LOG_ZONES, copy_zone_info_cb,
517 zones);
518 if (ret < 0)
519 return ret;
520 if (ret != BTRFS_NR_SB_LOG_ZONES)
521 return -EIO;
522
523 return sb_log_location(bdev, zones, rw, bytenr_ret);
524 }
525
526 int btrfs_sb_log_location(struct btrfs_device *device, int mirror, int rw,
527 u64 *bytenr_ret)
528 {
529 struct btrfs_zoned_device_info *zinfo = device->zone_info;
530 u32 zone_num;
531
532 if (!zinfo) {
533 *bytenr_ret = btrfs_sb_offset(mirror);
534 return 0;
535 }
536
537 zone_num = sb_zone_number(zinfo->zone_size_shift, mirror);
538 if (zone_num + 1 >= zinfo->nr_zones)
539 return -ENOENT;
540
541 return sb_log_location(device->bdev,
542 &zinfo->sb_zones[BTRFS_NR_SB_LOG_ZONES * mirror],
543 rw, bytenr_ret);
544 }
545
546 static inline bool is_sb_log_zone(struct btrfs_zoned_device_info *zinfo,
547 int mirror)
548 {
549 u32 zone_num;
550
551 if (!zinfo)
552 return false;
553
554 zone_num = sb_zone_number(zinfo->zone_size_shift, mirror);
555 if (zone_num + 1 >= zinfo->nr_zones)
556 return false;
557
558 if (!test_bit(zone_num, zinfo->seq_zones))
559 return false;
560
561 return true;
562 }
563
564 void btrfs_advance_sb_log(struct btrfs_device *device, int mirror)
565 {
566 struct btrfs_zoned_device_info *zinfo = device->zone_info;
567 struct blk_zone *zone;
568
569 if (!is_sb_log_zone(zinfo, mirror))
570 return;
571
572 zone = &zinfo->sb_zones[BTRFS_NR_SB_LOG_ZONES * mirror];
573 if (zone->cond != BLK_ZONE_COND_FULL) {
574 if (zone->cond == BLK_ZONE_COND_EMPTY)
575 zone->cond = BLK_ZONE_COND_IMP_OPEN;
576
577 zone->wp += (BTRFS_SUPER_INFO_SIZE >> SECTOR_SHIFT);
578
579 if (zone->wp == zone->start + zone->len)
580 zone->cond = BLK_ZONE_COND_FULL;
581
582 return;
583 }
584
585 zone++;
586 ASSERT(zone->cond != BLK_ZONE_COND_FULL);
587 if (zone->cond == BLK_ZONE_COND_EMPTY)
588 zone->cond = BLK_ZONE_COND_IMP_OPEN;
589
590 zone->wp += (BTRFS_SUPER_INFO_SIZE >> SECTOR_SHIFT);
591
592 if (zone->wp == zone->start + zone->len)
593 zone->cond = BLK_ZONE_COND_FULL;
594 }
595
596 int btrfs_reset_sb_log_zones(struct block_device *bdev, int mirror)
597 {
598 sector_t zone_sectors;
599 sector_t nr_sectors;
600 u8 zone_sectors_shift;
601 u32 sb_zone;
602 u32 nr_zones;
603
604 zone_sectors = bdev_zone_sectors(bdev);
605 zone_sectors_shift = ilog2(zone_sectors);
606 nr_sectors = bdev_nr_sectors(bdev);
607 nr_zones = nr_sectors >> zone_sectors_shift;
608
609 sb_zone = sb_zone_number(zone_sectors_shift + SECTOR_SHIFT, mirror);
610 if (sb_zone + 1 >= nr_zones)
611 return -ENOENT;
612
613 return blkdev_zone_mgmt(bdev, REQ_OP_ZONE_RESET,
614 sb_zone << zone_sectors_shift,
615 zone_sectors * BTRFS_NR_SB_LOG_ZONES, GFP_NOFS);
616 }