]> git.proxmox.com Git - mirror_qemu.git/blob - block/file-posix.c
Merge remote-tracking branch 'remotes/thuth-gitlab/tags/pull-request-2022-03-07'...
[mirror_qemu.git] / block / file-posix.c
1 /*
2 * Block driver for RAW files (posix)
3 *
4 * Copyright (c) 2006 Fabrice Bellard
5 *
6 * Permission is hereby granted, free of charge, to any person obtaining a copy
7 * of this software and associated documentation files (the "Software"), to deal
8 * in the Software without restriction, including without limitation the rights
9 * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
10 * copies of the Software, and to permit persons to whom the Software is
11 * furnished to do so, subject to the following conditions:
12 *
13 * The above copyright notice and this permission notice shall be included in
14 * all copies or substantial portions of the Software.
15 *
16 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
17 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
18 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
19 * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
20 * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
21 * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
22 * THE SOFTWARE.
23 */
24
25 #include "qemu/osdep.h"
26 #include "qemu-common.h"
27 #include "qapi/error.h"
28 #include "qemu/cutils.h"
29 #include "qemu/error-report.h"
30 #include "block/block_int.h"
31 #include "qemu/module.h"
32 #include "qemu/option.h"
33 #include "qemu/units.h"
34 #include "qemu/memalign.h"
35 #include "trace.h"
36 #include "block/thread-pool.h"
37 #include "qemu/iov.h"
38 #include "block/raw-aio.h"
39 #include "qapi/qmp/qdict.h"
40 #include "qapi/qmp/qstring.h"
41
42 #include "scsi/pr-manager.h"
43 #include "scsi/constants.h"
44
45 #if defined(__APPLE__) && (__MACH__)
46 #include <sys/ioctl.h>
47 #if defined(HAVE_HOST_BLOCK_DEVICE)
48 #include <paths.h>
49 #include <sys/param.h>
50 #include <sys/mount.h>
51 #include <IOKit/IOKitLib.h>
52 #include <IOKit/IOBSD.h>
53 #include <IOKit/storage/IOMediaBSDClient.h>
54 #include <IOKit/storage/IOMedia.h>
55 #include <IOKit/storage/IOCDMedia.h>
56 //#include <IOKit/storage/IOCDTypes.h>
57 #include <IOKit/storage/IODVDMedia.h>
58 #include <CoreFoundation/CoreFoundation.h>
59 #endif /* defined(HAVE_HOST_BLOCK_DEVICE) */
60 #endif
61
62 #ifdef __sun__
63 #define _POSIX_PTHREAD_SEMANTICS 1
64 #include <sys/dkio.h>
65 #endif
66 #ifdef __linux__
67 #include <sys/ioctl.h>
68 #include <sys/param.h>
69 #include <sys/syscall.h>
70 #include <sys/vfs.h>
71 #include <linux/cdrom.h>
72 #include <linux/fd.h>
73 #include <linux/fs.h>
74 #include <linux/hdreg.h>
75 #include <linux/magic.h>
76 #include <scsi/sg.h>
77 #ifdef __s390__
78 #include <asm/dasd.h>
79 #endif
80 #ifndef FS_NOCOW_FL
81 #define FS_NOCOW_FL 0x00800000 /* Do not cow file */
82 #endif
83 #endif
84 #if defined(CONFIG_FALLOCATE_PUNCH_HOLE) || defined(CONFIG_FALLOCATE_ZERO_RANGE)
85 #include <linux/falloc.h>
86 #endif
87 #if defined (__FreeBSD__) || defined(__FreeBSD_kernel__)
88 #include <sys/disk.h>
89 #include <sys/cdio.h>
90 #endif
91
92 #ifdef __OpenBSD__
93 #include <sys/ioctl.h>
94 #include <sys/disklabel.h>
95 #include <sys/dkio.h>
96 #endif
97
98 #ifdef __NetBSD__
99 #include <sys/ioctl.h>
100 #include <sys/disklabel.h>
101 #include <sys/dkio.h>
102 #include <sys/disk.h>
103 #endif
104
105 #ifdef __DragonFly__
106 #include <sys/ioctl.h>
107 #include <sys/diskslice.h>
108 #endif
109
110 /* OS X does not have O_DSYNC */
111 #ifndef O_DSYNC
112 #ifdef O_SYNC
113 #define O_DSYNC O_SYNC
114 #elif defined(O_FSYNC)
115 #define O_DSYNC O_FSYNC
116 #endif
117 #endif
118
119 /* Approximate O_DIRECT with O_DSYNC if O_DIRECT isn't available */
120 #ifndef O_DIRECT
121 #define O_DIRECT O_DSYNC
122 #endif
123
124 #define FTYPE_FILE 0
125 #define FTYPE_CD 1
126
127 #define MAX_BLOCKSIZE 4096
128
129 /* Posix file locking bytes. Libvirt takes byte 0, we start from higher bytes,
130 * leaving a few more bytes for its future use. */
131 #define RAW_LOCK_PERM_BASE 100
132 #define RAW_LOCK_SHARED_BASE 200
133
134 typedef struct BDRVRawState {
135 int fd;
136 bool use_lock;
137 int type;
138 int open_flags;
139 size_t buf_align;
140
141 /* The current permissions. */
142 uint64_t perm;
143 uint64_t shared_perm;
144
145 /* The perms bits whose corresponding bytes are already locked in
146 * s->fd. */
147 uint64_t locked_perm;
148 uint64_t locked_shared_perm;
149
150 uint64_t aio_max_batch;
151
152 int perm_change_fd;
153 int perm_change_flags;
154 BDRVReopenState *reopen_state;
155
156 bool has_discard:1;
157 bool has_write_zeroes:1;
158 bool discard_zeroes:1;
159 bool use_linux_aio:1;
160 bool use_linux_io_uring:1;
161 int page_cache_inconsistent; /* errno from fdatasync failure */
162 bool has_fallocate;
163 bool needs_alignment;
164 bool force_alignment;
165 bool drop_cache;
166 bool check_cache_dropped;
167 struct {
168 uint64_t discard_nb_ok;
169 uint64_t discard_nb_failed;
170 uint64_t discard_bytes_ok;
171 } stats;
172
173 PRManager *pr_mgr;
174 } BDRVRawState;
175
176 typedef struct BDRVRawReopenState {
177 int open_flags;
178 bool drop_cache;
179 bool check_cache_dropped;
180 } BDRVRawReopenState;
181
182 static int fd_open(BlockDriverState *bs)
183 {
184 BDRVRawState *s = bs->opaque;
185
186 /* this is just to ensure s->fd is sane (its called by io ops) */
187 if (s->fd >= 0) {
188 return 0;
189 }
190 return -EIO;
191 }
192
193 static int64_t raw_getlength(BlockDriverState *bs);
194
195 typedef struct RawPosixAIOData {
196 BlockDriverState *bs;
197 int aio_type;
198 int aio_fildes;
199
200 off_t aio_offset;
201 uint64_t aio_nbytes;
202
203 union {
204 struct {
205 struct iovec *iov;
206 int niov;
207 } io;
208 struct {
209 uint64_t cmd;
210 void *buf;
211 } ioctl;
212 struct {
213 int aio_fd2;
214 off_t aio_offset2;
215 } copy_range;
216 struct {
217 PreallocMode prealloc;
218 Error **errp;
219 } truncate;
220 };
221 } RawPosixAIOData;
222
223 #if defined(__FreeBSD__) || defined(__FreeBSD_kernel__)
224 static int cdrom_reopen(BlockDriverState *bs);
225 #endif
226
227 /*
228 * Elide EAGAIN and EACCES details when failing to lock, as this
229 * indicates that the specified file region is already locked by
230 * another process, which is considered a common scenario.
231 */
232 #define raw_lock_error_setg_errno(errp, err, fmt, ...) \
233 do { \
234 if ((err) == EAGAIN || (err) == EACCES) { \
235 error_setg((errp), (fmt), ## __VA_ARGS__); \
236 } else { \
237 error_setg_errno((errp), (err), (fmt), ## __VA_ARGS__); \
238 } \
239 } while (0)
240
241 #if defined(__NetBSD__)
242 static int raw_normalize_devicepath(const char **filename, Error **errp)
243 {
244 static char namebuf[PATH_MAX];
245 const char *dp, *fname;
246 struct stat sb;
247
248 fname = *filename;
249 dp = strrchr(fname, '/');
250 if (lstat(fname, &sb) < 0) {
251 error_setg_file_open(errp, errno, fname);
252 return -errno;
253 }
254
255 if (!S_ISBLK(sb.st_mode)) {
256 return 0;
257 }
258
259 if (dp == NULL) {
260 snprintf(namebuf, PATH_MAX, "r%s", fname);
261 } else {
262 snprintf(namebuf, PATH_MAX, "%.*s/r%s",
263 (int)(dp - fname), fname, dp + 1);
264 }
265 *filename = namebuf;
266 warn_report("%s is a block device, using %s", fname, *filename);
267
268 return 0;
269 }
270 #else
271 static int raw_normalize_devicepath(const char **filename, Error **errp)
272 {
273 return 0;
274 }
275 #endif
276
277 /*
278 * Get logical block size via ioctl. On success store it in @sector_size_p.
279 */
280 static int probe_logical_blocksize(int fd, unsigned int *sector_size_p)
281 {
282 unsigned int sector_size;
283 bool success = false;
284 int i;
285
286 errno = ENOTSUP;
287 static const unsigned long ioctl_list[] = {
288 #ifdef BLKSSZGET
289 BLKSSZGET,
290 #endif
291 #ifdef DKIOCGETBLOCKSIZE
292 DKIOCGETBLOCKSIZE,
293 #endif
294 #ifdef DIOCGSECTORSIZE
295 DIOCGSECTORSIZE,
296 #endif
297 };
298
299 /* Try a few ioctls to get the right size */
300 for (i = 0; i < (int)ARRAY_SIZE(ioctl_list); i++) {
301 if (ioctl(fd, ioctl_list[i], &sector_size) >= 0) {
302 *sector_size_p = sector_size;
303 success = true;
304 }
305 }
306
307 return success ? 0 : -errno;
308 }
309
310 /**
311 * Get physical block size of @fd.
312 * On success, store it in @blk_size and return 0.
313 * On failure, return -errno.
314 */
315 static int probe_physical_blocksize(int fd, unsigned int *blk_size)
316 {
317 #ifdef BLKPBSZGET
318 if (ioctl(fd, BLKPBSZGET, blk_size) < 0) {
319 return -errno;
320 }
321 return 0;
322 #else
323 return -ENOTSUP;
324 #endif
325 }
326
327 /*
328 * Returns true if no alignment restrictions are necessary even for files
329 * opened with O_DIRECT.
330 *
331 * raw_probe_alignment() probes the required alignment and assume that 1 means
332 * the probing failed, so it falls back to a safe default of 4k. This can be
333 * avoided if we know that byte alignment is okay for the file.
334 */
335 static bool dio_byte_aligned(int fd)
336 {
337 #ifdef __linux__
338 struct statfs buf;
339 int ret;
340
341 ret = fstatfs(fd, &buf);
342 if (ret == 0 && buf.f_type == NFS_SUPER_MAGIC) {
343 return true;
344 }
345 #endif
346 return false;
347 }
348
349 static bool raw_needs_alignment(BlockDriverState *bs)
350 {
351 BDRVRawState *s = bs->opaque;
352
353 if ((bs->open_flags & BDRV_O_NOCACHE) != 0 && !dio_byte_aligned(s->fd)) {
354 return true;
355 }
356
357 return s->force_alignment;
358 }
359
360 /* Check if read is allowed with given memory buffer and length.
361 *
362 * This function is used to check O_DIRECT memory buffer and request alignment.
363 */
364 static bool raw_is_io_aligned(int fd, void *buf, size_t len)
365 {
366 ssize_t ret = pread(fd, buf, len, 0);
367
368 if (ret >= 0) {
369 return true;
370 }
371
372 #ifdef __linux__
373 /* The Linux kernel returns EINVAL for misaligned O_DIRECT reads. Ignore
374 * other errors (e.g. real I/O error), which could happen on a failed
375 * drive, since we only care about probing alignment.
376 */
377 if (errno != EINVAL) {
378 return true;
379 }
380 #endif
381
382 return false;
383 }
384
385 static void raw_probe_alignment(BlockDriverState *bs, int fd, Error **errp)
386 {
387 BDRVRawState *s = bs->opaque;
388 char *buf;
389 size_t max_align = MAX(MAX_BLOCKSIZE, qemu_real_host_page_size);
390 size_t alignments[] = {1, 512, 1024, 2048, 4096};
391
392 /* For SCSI generic devices the alignment is not really used.
393 With buffered I/O, we don't have any restrictions. */
394 if (bdrv_is_sg(bs) || !s->needs_alignment) {
395 bs->bl.request_alignment = 1;
396 s->buf_align = 1;
397 return;
398 }
399
400 bs->bl.request_alignment = 0;
401 s->buf_align = 0;
402 /* Let's try to use the logical blocksize for the alignment. */
403 if (probe_logical_blocksize(fd, &bs->bl.request_alignment) < 0) {
404 bs->bl.request_alignment = 0;
405 }
406
407 #ifdef __linux__
408 /*
409 * The XFS ioctl definitions are shipped in extra packages that might
410 * not always be available. Since we just need the XFS_IOC_DIOINFO ioctl
411 * here, we simply use our own definition instead:
412 */
413 struct xfs_dioattr {
414 uint32_t d_mem;
415 uint32_t d_miniosz;
416 uint32_t d_maxiosz;
417 } da;
418 if (ioctl(fd, _IOR('X', 30, struct xfs_dioattr), &da) >= 0) {
419 bs->bl.request_alignment = da.d_miniosz;
420 /* The kernel returns wrong information for d_mem */
421 /* s->buf_align = da.d_mem; */
422 }
423 #endif
424
425 /*
426 * If we could not get the sizes so far, we can only guess them. First try
427 * to detect request alignment, since it is more likely to succeed. Then
428 * try to detect buf_align, which cannot be detected in some cases (e.g.
429 * Gluster). If buf_align cannot be detected, we fallback to the value of
430 * request_alignment.
431 */
432
433 if (!bs->bl.request_alignment) {
434 int i;
435 size_t align;
436 buf = qemu_memalign(max_align, max_align);
437 for (i = 0; i < ARRAY_SIZE(alignments); i++) {
438 align = alignments[i];
439 if (raw_is_io_aligned(fd, buf, align)) {
440 /* Fallback to safe value. */
441 bs->bl.request_alignment = (align != 1) ? align : max_align;
442 break;
443 }
444 }
445 qemu_vfree(buf);
446 }
447
448 if (!s->buf_align) {
449 int i;
450 size_t align;
451 buf = qemu_memalign(max_align, 2 * max_align);
452 for (i = 0; i < ARRAY_SIZE(alignments); i++) {
453 align = alignments[i];
454 if (raw_is_io_aligned(fd, buf + align, max_align)) {
455 /* Fallback to request_alignment. */
456 s->buf_align = (align != 1) ? align : bs->bl.request_alignment;
457 break;
458 }
459 }
460 qemu_vfree(buf);
461 }
462
463 if (!s->buf_align || !bs->bl.request_alignment) {
464 error_setg(errp, "Could not find working O_DIRECT alignment");
465 error_append_hint(errp, "Try cache.direct=off\n");
466 }
467 }
468
469 static int check_hdev_writable(int fd)
470 {
471 #if defined(BLKROGET)
472 /* Linux block devices can be configured "read-only" using blockdev(8).
473 * This is independent of device node permissions and therefore open(2)
474 * with O_RDWR succeeds. Actual writes fail with EPERM.
475 *
476 * bdrv_open() is supposed to fail if the disk is read-only. Explicitly
477 * check for read-only block devices so that Linux block devices behave
478 * properly.
479 */
480 struct stat st;
481 int readonly = 0;
482
483 if (fstat(fd, &st)) {
484 return -errno;
485 }
486
487 if (!S_ISBLK(st.st_mode)) {
488 return 0;
489 }
490
491 if (ioctl(fd, BLKROGET, &readonly) < 0) {
492 return -errno;
493 }
494
495 if (readonly) {
496 return -EACCES;
497 }
498 #endif /* defined(BLKROGET) */
499 return 0;
500 }
501
502 static void raw_parse_flags(int bdrv_flags, int *open_flags, bool has_writers)
503 {
504 bool read_write = false;
505 assert(open_flags != NULL);
506
507 *open_flags |= O_BINARY;
508 *open_flags &= ~O_ACCMODE;
509
510 if (bdrv_flags & BDRV_O_AUTO_RDONLY) {
511 read_write = has_writers;
512 } else if (bdrv_flags & BDRV_O_RDWR) {
513 read_write = true;
514 }
515
516 if (read_write) {
517 *open_flags |= O_RDWR;
518 } else {
519 *open_flags |= O_RDONLY;
520 }
521
522 /* Use O_DSYNC for write-through caching, no flags for write-back caching,
523 * and O_DIRECT for no caching. */
524 if ((bdrv_flags & BDRV_O_NOCACHE)) {
525 *open_flags |= O_DIRECT;
526 }
527 }
528
529 static void raw_parse_filename(const char *filename, QDict *options,
530 Error **errp)
531 {
532 bdrv_parse_filename_strip_prefix(filename, "file:", options);
533 }
534
535 static QemuOptsList raw_runtime_opts = {
536 .name = "raw",
537 .head = QTAILQ_HEAD_INITIALIZER(raw_runtime_opts.head),
538 .desc = {
539 {
540 .name = "filename",
541 .type = QEMU_OPT_STRING,
542 .help = "File name of the image",
543 },
544 {
545 .name = "aio",
546 .type = QEMU_OPT_STRING,
547 .help = "host AIO implementation (threads, native, io_uring)",
548 },
549 {
550 .name = "aio-max-batch",
551 .type = QEMU_OPT_NUMBER,
552 .help = "AIO max batch size (0 = auto handled by AIO backend, default: 0)",
553 },
554 {
555 .name = "locking",
556 .type = QEMU_OPT_STRING,
557 .help = "file locking mode (on/off/auto, default: auto)",
558 },
559 {
560 .name = "pr-manager",
561 .type = QEMU_OPT_STRING,
562 .help = "id of persistent reservation manager object (default: none)",
563 },
564 #if defined(__linux__)
565 {
566 .name = "drop-cache",
567 .type = QEMU_OPT_BOOL,
568 .help = "invalidate page cache during live migration (default: on)",
569 },
570 #endif
571 {
572 .name = "x-check-cache-dropped",
573 .type = QEMU_OPT_BOOL,
574 .help = "check that page cache was dropped on live migration (default: off)"
575 },
576 { /* end of list */ }
577 },
578 };
579
580 static const char *const mutable_opts[] = { "x-check-cache-dropped", NULL };
581
582 static int raw_open_common(BlockDriverState *bs, QDict *options,
583 int bdrv_flags, int open_flags,
584 bool device, Error **errp)
585 {
586 BDRVRawState *s = bs->opaque;
587 QemuOpts *opts;
588 Error *local_err = NULL;
589 const char *filename = NULL;
590 const char *str;
591 BlockdevAioOptions aio, aio_default;
592 int fd, ret;
593 struct stat st;
594 OnOffAuto locking;
595
596 opts = qemu_opts_create(&raw_runtime_opts, NULL, 0, &error_abort);
597 if (!qemu_opts_absorb_qdict(opts, options, errp)) {
598 ret = -EINVAL;
599 goto fail;
600 }
601
602 filename = qemu_opt_get(opts, "filename");
603
604 ret = raw_normalize_devicepath(&filename, errp);
605 if (ret != 0) {
606 goto fail;
607 }
608
609 if (bdrv_flags & BDRV_O_NATIVE_AIO) {
610 aio_default = BLOCKDEV_AIO_OPTIONS_NATIVE;
611 #ifdef CONFIG_LINUX_IO_URING
612 } else if (bdrv_flags & BDRV_O_IO_URING) {
613 aio_default = BLOCKDEV_AIO_OPTIONS_IO_URING;
614 #endif
615 } else {
616 aio_default = BLOCKDEV_AIO_OPTIONS_THREADS;
617 }
618
619 aio = qapi_enum_parse(&BlockdevAioOptions_lookup,
620 qemu_opt_get(opts, "aio"),
621 aio_default, &local_err);
622 if (local_err) {
623 error_propagate(errp, local_err);
624 ret = -EINVAL;
625 goto fail;
626 }
627
628 s->use_linux_aio = (aio == BLOCKDEV_AIO_OPTIONS_NATIVE);
629 #ifdef CONFIG_LINUX_IO_URING
630 s->use_linux_io_uring = (aio == BLOCKDEV_AIO_OPTIONS_IO_URING);
631 #endif
632
633 s->aio_max_batch = qemu_opt_get_number(opts, "aio-max-batch", 0);
634
635 locking = qapi_enum_parse(&OnOffAuto_lookup,
636 qemu_opt_get(opts, "locking"),
637 ON_OFF_AUTO_AUTO, &local_err);
638 if (local_err) {
639 error_propagate(errp, local_err);
640 ret = -EINVAL;
641 goto fail;
642 }
643 switch (locking) {
644 case ON_OFF_AUTO_ON:
645 s->use_lock = true;
646 if (!qemu_has_ofd_lock()) {
647 warn_report("File lock requested but OFD locking syscall is "
648 "unavailable, falling back to POSIX file locks");
649 error_printf("Due to the implementation, locks can be lost "
650 "unexpectedly.\n");
651 }
652 break;
653 case ON_OFF_AUTO_OFF:
654 s->use_lock = false;
655 break;
656 case ON_OFF_AUTO_AUTO:
657 s->use_lock = qemu_has_ofd_lock();
658 break;
659 default:
660 abort();
661 }
662
663 str = qemu_opt_get(opts, "pr-manager");
664 if (str) {
665 s->pr_mgr = pr_manager_lookup(str, &local_err);
666 if (local_err) {
667 error_propagate(errp, local_err);
668 ret = -EINVAL;
669 goto fail;
670 }
671 }
672
673 s->drop_cache = qemu_opt_get_bool(opts, "drop-cache", true);
674 s->check_cache_dropped = qemu_opt_get_bool(opts, "x-check-cache-dropped",
675 false);
676
677 s->open_flags = open_flags;
678 raw_parse_flags(bdrv_flags, &s->open_flags, false);
679
680 s->fd = -1;
681 fd = qemu_open(filename, s->open_flags, errp);
682 ret = fd < 0 ? -errno : 0;
683
684 if (ret < 0) {
685 if (ret == -EROFS) {
686 ret = -EACCES;
687 }
688 goto fail;
689 }
690 s->fd = fd;
691
692 /* Check s->open_flags rather than bdrv_flags due to auto-read-only */
693 if (s->open_flags & O_RDWR) {
694 ret = check_hdev_writable(s->fd);
695 if (ret < 0) {
696 error_setg_errno(errp, -ret, "The device is not writable");
697 goto fail;
698 }
699 }
700
701 s->perm = 0;
702 s->shared_perm = BLK_PERM_ALL;
703
704 #ifdef CONFIG_LINUX_AIO
705 /* Currently Linux does AIO only for files opened with O_DIRECT */
706 if (s->use_linux_aio) {
707 if (!(s->open_flags & O_DIRECT)) {
708 error_setg(errp, "aio=native was specified, but it requires "
709 "cache.direct=on, which was not specified.");
710 ret = -EINVAL;
711 goto fail;
712 }
713 if (!aio_setup_linux_aio(bdrv_get_aio_context(bs), errp)) {
714 error_prepend(errp, "Unable to use native AIO: ");
715 goto fail;
716 }
717 }
718 #else
719 if (s->use_linux_aio) {
720 error_setg(errp, "aio=native was specified, but is not supported "
721 "in this build.");
722 ret = -EINVAL;
723 goto fail;
724 }
725 #endif /* !defined(CONFIG_LINUX_AIO) */
726
727 #ifdef CONFIG_LINUX_IO_URING
728 if (s->use_linux_io_uring) {
729 if (!aio_setup_linux_io_uring(bdrv_get_aio_context(bs), errp)) {
730 error_prepend(errp, "Unable to use io_uring: ");
731 goto fail;
732 }
733 }
734 #else
735 if (s->use_linux_io_uring) {
736 error_setg(errp, "aio=io_uring was specified, but is not supported "
737 "in this build.");
738 ret = -EINVAL;
739 goto fail;
740 }
741 #endif /* !defined(CONFIG_LINUX_IO_URING) */
742
743 s->has_discard = true;
744 s->has_write_zeroes = true;
745
746 if (fstat(s->fd, &st) < 0) {
747 ret = -errno;
748 error_setg_errno(errp, errno, "Could not stat file");
749 goto fail;
750 }
751
752 if (!device) {
753 if (!S_ISREG(st.st_mode)) {
754 error_setg(errp, "'%s' driver requires '%s' to be a regular file",
755 bs->drv->format_name, bs->filename);
756 ret = -EINVAL;
757 goto fail;
758 } else {
759 s->discard_zeroes = true;
760 s->has_fallocate = true;
761 }
762 } else {
763 if (!(S_ISCHR(st.st_mode) || S_ISBLK(st.st_mode))) {
764 error_setg(errp, "'%s' driver requires '%s' to be either "
765 "a character or block device",
766 bs->drv->format_name, bs->filename);
767 ret = -EINVAL;
768 goto fail;
769 }
770 }
771
772 if (S_ISBLK(st.st_mode)) {
773 #ifdef BLKDISCARDZEROES
774 unsigned int arg;
775 if (ioctl(s->fd, BLKDISCARDZEROES, &arg) == 0 && arg) {
776 s->discard_zeroes = true;
777 }
778 #endif
779 #ifdef __linux__
780 /* On Linux 3.10, BLKDISCARD leaves stale data in the page cache. Do
781 * not rely on the contents of discarded blocks unless using O_DIRECT.
782 * Same for BLKZEROOUT.
783 */
784 if (!(bs->open_flags & BDRV_O_NOCACHE)) {
785 s->discard_zeroes = false;
786 s->has_write_zeroes = false;
787 }
788 #endif
789 }
790 #ifdef __FreeBSD__
791 if (S_ISCHR(st.st_mode)) {
792 /*
793 * The file is a char device (disk), which on FreeBSD isn't behind
794 * a pager, so force all requests to be aligned. This is needed
795 * so QEMU makes sure all IO operations on the device are aligned
796 * to sector size, or else FreeBSD will reject them with EINVAL.
797 */
798 s->force_alignment = true;
799 }
800 #endif
801 s->needs_alignment = raw_needs_alignment(bs);
802
803 bs->supported_zero_flags = BDRV_REQ_MAY_UNMAP | BDRV_REQ_NO_FALLBACK;
804 if (S_ISREG(st.st_mode)) {
805 /* When extending regular files, we get zeros from the OS */
806 bs->supported_truncate_flags = BDRV_REQ_ZERO_WRITE;
807 }
808 ret = 0;
809 fail:
810 if (ret < 0 && s->fd != -1) {
811 qemu_close(s->fd);
812 }
813 if (filename && (bdrv_flags & BDRV_O_TEMPORARY)) {
814 unlink(filename);
815 }
816 qemu_opts_del(opts);
817 return ret;
818 }
819
820 static int raw_open(BlockDriverState *bs, QDict *options, int flags,
821 Error **errp)
822 {
823 BDRVRawState *s = bs->opaque;
824
825 s->type = FTYPE_FILE;
826 return raw_open_common(bs, options, flags, 0, false, errp);
827 }
828
829 typedef enum {
830 RAW_PL_PREPARE,
831 RAW_PL_COMMIT,
832 RAW_PL_ABORT,
833 } RawPermLockOp;
834
835 #define PERM_FOREACH(i) \
836 for ((i) = 0; (1ULL << (i)) <= BLK_PERM_ALL; i++)
837
838 /* Lock bytes indicated by @perm_lock_bits and @shared_perm_lock_bits in the
839 * file; if @unlock == true, also unlock the unneeded bytes.
840 * @shared_perm_lock_bits is the mask of all permissions that are NOT shared.
841 */
842 static int raw_apply_lock_bytes(BDRVRawState *s, int fd,
843 uint64_t perm_lock_bits,
844 uint64_t shared_perm_lock_bits,
845 bool unlock, Error **errp)
846 {
847 int ret;
848 int i;
849 uint64_t locked_perm, locked_shared_perm;
850
851 if (s) {
852 locked_perm = s->locked_perm;
853 locked_shared_perm = s->locked_shared_perm;
854 } else {
855 /*
856 * We don't have the previous bits, just lock/unlock for each of the
857 * requested bits.
858 */
859 if (unlock) {
860 locked_perm = BLK_PERM_ALL;
861 locked_shared_perm = BLK_PERM_ALL;
862 } else {
863 locked_perm = 0;
864 locked_shared_perm = 0;
865 }
866 }
867
868 PERM_FOREACH(i) {
869 int off = RAW_LOCK_PERM_BASE + i;
870 uint64_t bit = (1ULL << i);
871 if ((perm_lock_bits & bit) && !(locked_perm & bit)) {
872 ret = qemu_lock_fd(fd, off, 1, false);
873 if (ret) {
874 raw_lock_error_setg_errno(errp, -ret, "Failed to lock byte %d",
875 off);
876 return ret;
877 } else if (s) {
878 s->locked_perm |= bit;
879 }
880 } else if (unlock && (locked_perm & bit) && !(perm_lock_bits & bit)) {
881 ret = qemu_unlock_fd(fd, off, 1);
882 if (ret) {
883 error_setg_errno(errp, -ret, "Failed to unlock byte %d", off);
884 return ret;
885 } else if (s) {
886 s->locked_perm &= ~bit;
887 }
888 }
889 }
890 PERM_FOREACH(i) {
891 int off = RAW_LOCK_SHARED_BASE + i;
892 uint64_t bit = (1ULL << i);
893 if ((shared_perm_lock_bits & bit) && !(locked_shared_perm & bit)) {
894 ret = qemu_lock_fd(fd, off, 1, false);
895 if (ret) {
896 raw_lock_error_setg_errno(errp, -ret, "Failed to lock byte %d",
897 off);
898 return ret;
899 } else if (s) {
900 s->locked_shared_perm |= bit;
901 }
902 } else if (unlock && (locked_shared_perm & bit) &&
903 !(shared_perm_lock_bits & bit)) {
904 ret = qemu_unlock_fd(fd, off, 1);
905 if (ret) {
906 error_setg_errno(errp, -ret, "Failed to unlock byte %d", off);
907 return ret;
908 } else if (s) {
909 s->locked_shared_perm &= ~bit;
910 }
911 }
912 }
913 return 0;
914 }
915
916 /* Check "unshared" bytes implied by @perm and ~@shared_perm in the file. */
917 static int raw_check_lock_bytes(int fd, uint64_t perm, uint64_t shared_perm,
918 Error **errp)
919 {
920 int ret;
921 int i;
922
923 PERM_FOREACH(i) {
924 int off = RAW_LOCK_SHARED_BASE + i;
925 uint64_t p = 1ULL << i;
926 if (perm & p) {
927 ret = qemu_lock_fd_test(fd, off, 1, true);
928 if (ret) {
929 char *perm_name = bdrv_perm_names(p);
930
931 raw_lock_error_setg_errno(errp, -ret,
932 "Failed to get \"%s\" lock",
933 perm_name);
934 g_free(perm_name);
935 return ret;
936 }
937 }
938 }
939 PERM_FOREACH(i) {
940 int off = RAW_LOCK_PERM_BASE + i;
941 uint64_t p = 1ULL << i;
942 if (!(shared_perm & p)) {
943 ret = qemu_lock_fd_test(fd, off, 1, true);
944 if (ret) {
945 char *perm_name = bdrv_perm_names(p);
946
947 raw_lock_error_setg_errno(errp, -ret,
948 "Failed to get shared \"%s\" lock",
949 perm_name);
950 g_free(perm_name);
951 return ret;
952 }
953 }
954 }
955 return 0;
956 }
957
958 static int raw_handle_perm_lock(BlockDriverState *bs,
959 RawPermLockOp op,
960 uint64_t new_perm, uint64_t new_shared,
961 Error **errp)
962 {
963 BDRVRawState *s = bs->opaque;
964 int ret = 0;
965 Error *local_err = NULL;
966
967 if (!s->use_lock) {
968 return 0;
969 }
970
971 if (bdrv_get_flags(bs) & BDRV_O_INACTIVE) {
972 return 0;
973 }
974
975 switch (op) {
976 case RAW_PL_PREPARE:
977 if ((s->perm | new_perm) == s->perm &&
978 (s->shared_perm & new_shared) == s->shared_perm)
979 {
980 /*
981 * We are going to unlock bytes, it should not fail. If it fail due
982 * to some fs-dependent permission-unrelated reasons (which occurs
983 * sometimes on NFS and leads to abort in bdrv_replace_child) we
984 * can't prevent such errors by any check here. And we ignore them
985 * anyway in ABORT and COMMIT.
986 */
987 return 0;
988 }
989 ret = raw_apply_lock_bytes(s, s->fd, s->perm | new_perm,
990 ~s->shared_perm | ~new_shared,
991 false, errp);
992 if (!ret) {
993 ret = raw_check_lock_bytes(s->fd, new_perm, new_shared, errp);
994 if (!ret) {
995 return 0;
996 }
997 error_append_hint(errp,
998 "Is another process using the image [%s]?\n",
999 bs->filename);
1000 }
1001 /* fall through to unlock bytes. */
1002 case RAW_PL_ABORT:
1003 raw_apply_lock_bytes(s, s->fd, s->perm, ~s->shared_perm,
1004 true, &local_err);
1005 if (local_err) {
1006 /* Theoretically the above call only unlocks bytes and it cannot
1007 * fail. Something weird happened, report it.
1008 */
1009 warn_report_err(local_err);
1010 }
1011 break;
1012 case RAW_PL_COMMIT:
1013 raw_apply_lock_bytes(s, s->fd, new_perm, ~new_shared,
1014 true, &local_err);
1015 if (local_err) {
1016 /* Theoretically the above call only unlocks bytes and it cannot
1017 * fail. Something weird happened, report it.
1018 */
1019 warn_report_err(local_err);
1020 }
1021 break;
1022 }
1023 return ret;
1024 }
1025
1026 static int raw_reconfigure_getfd(BlockDriverState *bs, int flags,
1027 int *open_flags, uint64_t perm, bool force_dup,
1028 Error **errp)
1029 {
1030 BDRVRawState *s = bs->opaque;
1031 int fd = -1;
1032 int ret;
1033 bool has_writers = perm &
1034 (BLK_PERM_WRITE | BLK_PERM_WRITE_UNCHANGED | BLK_PERM_RESIZE);
1035 int fcntl_flags = O_APPEND | O_NONBLOCK;
1036 #ifdef O_NOATIME
1037 fcntl_flags |= O_NOATIME;
1038 #endif
1039
1040 *open_flags = 0;
1041 if (s->type == FTYPE_CD) {
1042 *open_flags |= O_NONBLOCK;
1043 }
1044
1045 raw_parse_flags(flags, open_flags, has_writers);
1046
1047 #ifdef O_ASYNC
1048 /* Not all operating systems have O_ASYNC, and those that don't
1049 * will not let us track the state into rs->open_flags (typically
1050 * you achieve the same effect with an ioctl, for example I_SETSIG
1051 * on Solaris). But we do not use O_ASYNC, so that's fine.
1052 */
1053 assert((s->open_flags & O_ASYNC) == 0);
1054 #endif
1055
1056 if (!force_dup && *open_flags == s->open_flags) {
1057 /* We're lucky, the existing fd is fine */
1058 return s->fd;
1059 }
1060
1061 if ((*open_flags & ~fcntl_flags) == (s->open_flags & ~fcntl_flags)) {
1062 /* dup the original fd */
1063 fd = qemu_dup(s->fd);
1064 if (fd >= 0) {
1065 ret = fcntl_setfl(fd, *open_flags);
1066 if (ret) {
1067 qemu_close(fd);
1068 fd = -1;
1069 }
1070 }
1071 }
1072
1073 /* If we cannot use fcntl, or fcntl failed, fall back to qemu_open() */
1074 if (fd == -1) {
1075 const char *normalized_filename = bs->filename;
1076 ret = raw_normalize_devicepath(&normalized_filename, errp);
1077 if (ret >= 0) {
1078 fd = qemu_open(normalized_filename, *open_flags, errp);
1079 if (fd == -1) {
1080 return -1;
1081 }
1082 }
1083 }
1084
1085 if (fd != -1 && (*open_flags & O_RDWR)) {
1086 ret = check_hdev_writable(fd);
1087 if (ret < 0) {
1088 qemu_close(fd);
1089 error_setg_errno(errp, -ret, "The device is not writable");
1090 return -1;
1091 }
1092 }
1093
1094 return fd;
1095 }
1096
1097 static int raw_reopen_prepare(BDRVReopenState *state,
1098 BlockReopenQueue *queue, Error **errp)
1099 {
1100 BDRVRawState *s;
1101 BDRVRawReopenState *rs;
1102 QemuOpts *opts;
1103 int ret;
1104
1105 assert(state != NULL);
1106 assert(state->bs != NULL);
1107
1108 s = state->bs->opaque;
1109
1110 state->opaque = g_new0(BDRVRawReopenState, 1);
1111 rs = state->opaque;
1112
1113 /* Handle options changes */
1114 opts = qemu_opts_create(&raw_runtime_opts, NULL, 0, &error_abort);
1115 if (!qemu_opts_absorb_qdict(opts, state->options, errp)) {
1116 ret = -EINVAL;
1117 goto out;
1118 }
1119
1120 rs->drop_cache = qemu_opt_get_bool_del(opts, "drop-cache", true);
1121 rs->check_cache_dropped =
1122 qemu_opt_get_bool_del(opts, "x-check-cache-dropped", false);
1123
1124 /* This driver's reopen function doesn't currently allow changing
1125 * other options, so let's put them back in the original QDict and
1126 * bdrv_reopen_prepare() will detect changes and complain. */
1127 qemu_opts_to_qdict(opts, state->options);
1128
1129 /*
1130 * As part of reopen prepare we also want to create new fd by
1131 * raw_reconfigure_getfd(). But it wants updated "perm", when in
1132 * bdrv_reopen_multiple() .bdrv_reopen_prepare() callback called prior to
1133 * permission update. Happily, permission update is always a part (a seprate
1134 * stage) of bdrv_reopen_multiple() so we can rely on this fact and
1135 * reconfigure fd in raw_check_perm().
1136 */
1137
1138 s->reopen_state = state;
1139 ret = 0;
1140
1141 out:
1142 qemu_opts_del(opts);
1143 return ret;
1144 }
1145
1146 static void raw_reopen_commit(BDRVReopenState *state)
1147 {
1148 BDRVRawReopenState *rs = state->opaque;
1149 BDRVRawState *s = state->bs->opaque;
1150
1151 s->drop_cache = rs->drop_cache;
1152 s->check_cache_dropped = rs->check_cache_dropped;
1153 s->open_flags = rs->open_flags;
1154 g_free(state->opaque);
1155 state->opaque = NULL;
1156
1157 assert(s->reopen_state == state);
1158 s->reopen_state = NULL;
1159 }
1160
1161
1162 static void raw_reopen_abort(BDRVReopenState *state)
1163 {
1164 BDRVRawReopenState *rs = state->opaque;
1165 BDRVRawState *s = state->bs->opaque;
1166
1167 /* nothing to do if NULL, we didn't get far enough */
1168 if (rs == NULL) {
1169 return;
1170 }
1171
1172 g_free(state->opaque);
1173 state->opaque = NULL;
1174
1175 assert(s->reopen_state == state);
1176 s->reopen_state = NULL;
1177 }
1178
1179 static int hdev_get_max_hw_transfer(int fd, struct stat *st)
1180 {
1181 #ifdef BLKSECTGET
1182 if (S_ISBLK(st->st_mode)) {
1183 unsigned short max_sectors = 0;
1184 if (ioctl(fd, BLKSECTGET, &max_sectors) == 0) {
1185 return max_sectors * 512;
1186 }
1187 } else {
1188 int max_bytes = 0;
1189 if (ioctl(fd, BLKSECTGET, &max_bytes) == 0) {
1190 return max_bytes;
1191 }
1192 }
1193 return -errno;
1194 #else
1195 return -ENOSYS;
1196 #endif
1197 }
1198
1199 static int hdev_get_max_segments(int fd, struct stat *st)
1200 {
1201 #ifdef CONFIG_LINUX
1202 char buf[32];
1203 const char *end;
1204 char *sysfspath = NULL;
1205 int ret;
1206 int sysfd = -1;
1207 long max_segments;
1208
1209 if (S_ISCHR(st->st_mode)) {
1210 if (ioctl(fd, SG_GET_SG_TABLESIZE, &ret) == 0) {
1211 return ret;
1212 }
1213 return -ENOTSUP;
1214 }
1215
1216 if (!S_ISBLK(st->st_mode)) {
1217 return -ENOTSUP;
1218 }
1219
1220 sysfspath = g_strdup_printf("/sys/dev/block/%u:%u/queue/max_segments",
1221 major(st->st_rdev), minor(st->st_rdev));
1222 sysfd = open(sysfspath, O_RDONLY);
1223 if (sysfd == -1) {
1224 ret = -errno;
1225 goto out;
1226 }
1227 do {
1228 ret = read(sysfd, buf, sizeof(buf) - 1);
1229 } while (ret == -1 && errno == EINTR);
1230 if (ret < 0) {
1231 ret = -errno;
1232 goto out;
1233 } else if (ret == 0) {
1234 ret = -EIO;
1235 goto out;
1236 }
1237 buf[ret] = 0;
1238 /* The file is ended with '\n', pass 'end' to accept that. */
1239 ret = qemu_strtol(buf, &end, 10, &max_segments);
1240 if (ret == 0 && end && *end == '\n') {
1241 ret = max_segments;
1242 }
1243
1244 out:
1245 if (sysfd != -1) {
1246 close(sysfd);
1247 }
1248 g_free(sysfspath);
1249 return ret;
1250 #else
1251 return -ENOTSUP;
1252 #endif
1253 }
1254
1255 static void raw_refresh_limits(BlockDriverState *bs, Error **errp)
1256 {
1257 BDRVRawState *s = bs->opaque;
1258 struct stat st;
1259
1260 s->needs_alignment = raw_needs_alignment(bs);
1261 raw_probe_alignment(bs, s->fd, errp);
1262
1263 bs->bl.min_mem_alignment = s->buf_align;
1264 bs->bl.opt_mem_alignment = MAX(s->buf_align, qemu_real_host_page_size);
1265
1266 /*
1267 * Maximum transfers are best effort, so it is okay to ignore any
1268 * errors. That said, based on the man page errors in fstat would be
1269 * very much unexpected; the only possible case seems to be ENOMEM.
1270 */
1271 if (fstat(s->fd, &st)) {
1272 return;
1273 }
1274
1275 #if defined(__APPLE__) && (__MACH__)
1276 struct statfs buf;
1277
1278 if (!fstatfs(s->fd, &buf)) {
1279 bs->bl.opt_transfer = buf.f_iosize;
1280 bs->bl.pdiscard_alignment = buf.f_bsize;
1281 }
1282 #endif
1283
1284 if (bs->sg || S_ISBLK(st.st_mode)) {
1285 int ret = hdev_get_max_hw_transfer(s->fd, &st);
1286
1287 if (ret > 0 && ret <= BDRV_REQUEST_MAX_BYTES) {
1288 bs->bl.max_hw_transfer = ret;
1289 }
1290
1291 ret = hdev_get_max_segments(s->fd, &st);
1292 if (ret > 0) {
1293 bs->bl.max_hw_iov = ret;
1294 }
1295 }
1296 }
1297
1298 static int check_for_dasd(int fd)
1299 {
1300 #ifdef BIODASDINFO2
1301 struct dasd_information2_t info = {0};
1302
1303 return ioctl(fd, BIODASDINFO2, &info);
1304 #else
1305 return -1;
1306 #endif
1307 }
1308
1309 /**
1310 * Try to get @bs's logical and physical block size.
1311 * On success, store them in @bsz and return zero.
1312 * On failure, return negative errno.
1313 */
1314 static int hdev_probe_blocksizes(BlockDriverState *bs, BlockSizes *bsz)
1315 {
1316 BDRVRawState *s = bs->opaque;
1317 int ret;
1318
1319 /* If DASD, get blocksizes */
1320 if (check_for_dasd(s->fd) < 0) {
1321 return -ENOTSUP;
1322 }
1323 ret = probe_logical_blocksize(s->fd, &bsz->log);
1324 if (ret < 0) {
1325 return ret;
1326 }
1327 return probe_physical_blocksize(s->fd, &bsz->phys);
1328 }
1329
1330 /**
1331 * Try to get @bs's geometry: cyls, heads, sectors.
1332 * On success, store them in @geo and return 0.
1333 * On failure return -errno.
1334 * (Allows block driver to assign default geometry values that guest sees)
1335 */
1336 #ifdef __linux__
1337 static int hdev_probe_geometry(BlockDriverState *bs, HDGeometry *geo)
1338 {
1339 BDRVRawState *s = bs->opaque;
1340 struct hd_geometry ioctl_geo = {0};
1341
1342 /* If DASD, get its geometry */
1343 if (check_for_dasd(s->fd) < 0) {
1344 return -ENOTSUP;
1345 }
1346 if (ioctl(s->fd, HDIO_GETGEO, &ioctl_geo) < 0) {
1347 return -errno;
1348 }
1349 /* HDIO_GETGEO may return success even though geo contains zeros
1350 (e.g. certain multipath setups) */
1351 if (!ioctl_geo.heads || !ioctl_geo.sectors || !ioctl_geo.cylinders) {
1352 return -ENOTSUP;
1353 }
1354 /* Do not return a geometry for partition */
1355 if (ioctl_geo.start != 0) {
1356 return -ENOTSUP;
1357 }
1358 geo->heads = ioctl_geo.heads;
1359 geo->sectors = ioctl_geo.sectors;
1360 geo->cylinders = ioctl_geo.cylinders;
1361
1362 return 0;
1363 }
1364 #else /* __linux__ */
1365 static int hdev_probe_geometry(BlockDriverState *bs, HDGeometry *geo)
1366 {
1367 return -ENOTSUP;
1368 }
1369 #endif
1370
1371 #if defined(__linux__)
1372 static int handle_aiocb_ioctl(void *opaque)
1373 {
1374 RawPosixAIOData *aiocb = opaque;
1375 int ret;
1376
1377 do {
1378 ret = ioctl(aiocb->aio_fildes, aiocb->ioctl.cmd, aiocb->ioctl.buf);
1379 } while (ret == -1 && errno == EINTR);
1380 if (ret == -1) {
1381 return -errno;
1382 }
1383
1384 return 0;
1385 }
1386 #endif /* linux */
1387
1388 static int handle_aiocb_flush(void *opaque)
1389 {
1390 RawPosixAIOData *aiocb = opaque;
1391 BDRVRawState *s = aiocb->bs->opaque;
1392 int ret;
1393
1394 if (s->page_cache_inconsistent) {
1395 return -s->page_cache_inconsistent;
1396 }
1397
1398 ret = qemu_fdatasync(aiocb->aio_fildes);
1399 if (ret == -1) {
1400 trace_file_flush_fdatasync_failed(errno);
1401
1402 /* There is no clear definition of the semantics of a failing fsync(),
1403 * so we may have to assume the worst. The sad truth is that this
1404 * assumption is correct for Linux. Some pages are now probably marked
1405 * clean in the page cache even though they are inconsistent with the
1406 * on-disk contents. The next fdatasync() call would succeed, but no
1407 * further writeback attempt will be made. We can't get back to a state
1408 * in which we know what is on disk (we would have to rewrite
1409 * everything that was touched since the last fdatasync() at least), so
1410 * make bdrv_flush() fail permanently. Given that the behaviour isn't
1411 * really defined, I have little hope that other OSes are doing better.
1412 *
1413 * Obviously, this doesn't affect O_DIRECT, which bypasses the page
1414 * cache. */
1415 if ((s->open_flags & O_DIRECT) == 0) {
1416 s->page_cache_inconsistent = errno;
1417 }
1418 return -errno;
1419 }
1420 return 0;
1421 }
1422
1423 #ifdef CONFIG_PREADV
1424
1425 static bool preadv_present = true;
1426
1427 static ssize_t
1428 qemu_preadv(int fd, const struct iovec *iov, int nr_iov, off_t offset)
1429 {
1430 return preadv(fd, iov, nr_iov, offset);
1431 }
1432
1433 static ssize_t
1434 qemu_pwritev(int fd, const struct iovec *iov, int nr_iov, off_t offset)
1435 {
1436 return pwritev(fd, iov, nr_iov, offset);
1437 }
1438
1439 #else
1440
1441 static bool preadv_present = false;
1442
1443 static ssize_t
1444 qemu_preadv(int fd, const struct iovec *iov, int nr_iov, off_t offset)
1445 {
1446 return -ENOSYS;
1447 }
1448
1449 static ssize_t
1450 qemu_pwritev(int fd, const struct iovec *iov, int nr_iov, off_t offset)
1451 {
1452 return -ENOSYS;
1453 }
1454
1455 #endif
1456
1457 static ssize_t handle_aiocb_rw_vector(RawPosixAIOData *aiocb)
1458 {
1459 ssize_t len;
1460
1461 do {
1462 if (aiocb->aio_type & QEMU_AIO_WRITE)
1463 len = qemu_pwritev(aiocb->aio_fildes,
1464 aiocb->io.iov,
1465 aiocb->io.niov,
1466 aiocb->aio_offset);
1467 else
1468 len = qemu_preadv(aiocb->aio_fildes,
1469 aiocb->io.iov,
1470 aiocb->io.niov,
1471 aiocb->aio_offset);
1472 } while (len == -1 && errno == EINTR);
1473
1474 if (len == -1) {
1475 return -errno;
1476 }
1477 return len;
1478 }
1479
1480 /*
1481 * Read/writes the data to/from a given linear buffer.
1482 *
1483 * Returns the number of bytes handles or -errno in case of an error. Short
1484 * reads are only returned if the end of the file is reached.
1485 */
1486 static ssize_t handle_aiocb_rw_linear(RawPosixAIOData *aiocb, char *buf)
1487 {
1488 ssize_t offset = 0;
1489 ssize_t len;
1490
1491 while (offset < aiocb->aio_nbytes) {
1492 if (aiocb->aio_type & QEMU_AIO_WRITE) {
1493 len = pwrite(aiocb->aio_fildes,
1494 (const char *)buf + offset,
1495 aiocb->aio_nbytes - offset,
1496 aiocb->aio_offset + offset);
1497 } else {
1498 len = pread(aiocb->aio_fildes,
1499 buf + offset,
1500 aiocb->aio_nbytes - offset,
1501 aiocb->aio_offset + offset);
1502 }
1503 if (len == -1 && errno == EINTR) {
1504 continue;
1505 } else if (len == -1 && errno == EINVAL &&
1506 (aiocb->bs->open_flags & BDRV_O_NOCACHE) &&
1507 !(aiocb->aio_type & QEMU_AIO_WRITE) &&
1508 offset > 0) {
1509 /* O_DIRECT pread() may fail with EINVAL when offset is unaligned
1510 * after a short read. Assume that O_DIRECT short reads only occur
1511 * at EOF. Therefore this is a short read, not an I/O error.
1512 */
1513 break;
1514 } else if (len == -1) {
1515 offset = -errno;
1516 break;
1517 } else if (len == 0) {
1518 break;
1519 }
1520 offset += len;
1521 }
1522
1523 return offset;
1524 }
1525
1526 static int handle_aiocb_rw(void *opaque)
1527 {
1528 RawPosixAIOData *aiocb = opaque;
1529 ssize_t nbytes;
1530 char *buf;
1531
1532 if (!(aiocb->aio_type & QEMU_AIO_MISALIGNED)) {
1533 /*
1534 * If there is just a single buffer, and it is properly aligned
1535 * we can just use plain pread/pwrite without any problems.
1536 */
1537 if (aiocb->io.niov == 1) {
1538 nbytes = handle_aiocb_rw_linear(aiocb, aiocb->io.iov->iov_base);
1539 goto out;
1540 }
1541 /*
1542 * We have more than one iovec, and all are properly aligned.
1543 *
1544 * Try preadv/pwritev first and fall back to linearizing the
1545 * buffer if it's not supported.
1546 */
1547 if (preadv_present) {
1548 nbytes = handle_aiocb_rw_vector(aiocb);
1549 if (nbytes == aiocb->aio_nbytes ||
1550 (nbytes < 0 && nbytes != -ENOSYS)) {
1551 goto out;
1552 }
1553 preadv_present = false;
1554 }
1555
1556 /*
1557 * XXX(hch): short read/write. no easy way to handle the reminder
1558 * using these interfaces. For now retry using plain
1559 * pread/pwrite?
1560 */
1561 }
1562
1563 /*
1564 * Ok, we have to do it the hard way, copy all segments into
1565 * a single aligned buffer.
1566 */
1567 buf = qemu_try_blockalign(aiocb->bs, aiocb->aio_nbytes);
1568 if (buf == NULL) {
1569 nbytes = -ENOMEM;
1570 goto out;
1571 }
1572
1573 if (aiocb->aio_type & QEMU_AIO_WRITE) {
1574 char *p = buf;
1575 int i;
1576
1577 for (i = 0; i < aiocb->io.niov; ++i) {
1578 memcpy(p, aiocb->io.iov[i].iov_base, aiocb->io.iov[i].iov_len);
1579 p += aiocb->io.iov[i].iov_len;
1580 }
1581 assert(p - buf == aiocb->aio_nbytes);
1582 }
1583
1584 nbytes = handle_aiocb_rw_linear(aiocb, buf);
1585 if (!(aiocb->aio_type & QEMU_AIO_WRITE)) {
1586 char *p = buf;
1587 size_t count = aiocb->aio_nbytes, copy;
1588 int i;
1589
1590 for (i = 0; i < aiocb->io.niov && count; ++i) {
1591 copy = count;
1592 if (copy > aiocb->io.iov[i].iov_len) {
1593 copy = aiocb->io.iov[i].iov_len;
1594 }
1595 memcpy(aiocb->io.iov[i].iov_base, p, copy);
1596 assert(count >= copy);
1597 p += copy;
1598 count -= copy;
1599 }
1600 assert(count == 0);
1601 }
1602 qemu_vfree(buf);
1603
1604 out:
1605 if (nbytes == aiocb->aio_nbytes) {
1606 return 0;
1607 } else if (nbytes >= 0 && nbytes < aiocb->aio_nbytes) {
1608 if (aiocb->aio_type & QEMU_AIO_WRITE) {
1609 return -EINVAL;
1610 } else {
1611 iov_memset(aiocb->io.iov, aiocb->io.niov, nbytes,
1612 0, aiocb->aio_nbytes - nbytes);
1613 return 0;
1614 }
1615 } else {
1616 assert(nbytes < 0);
1617 return nbytes;
1618 }
1619 }
1620
1621 #if defined(CONFIG_FALLOCATE) || defined(BLKZEROOUT) || defined(BLKDISCARD)
1622 static int translate_err(int err)
1623 {
1624 if (err == -ENODEV || err == -ENOSYS || err == -EOPNOTSUPP ||
1625 err == -ENOTTY) {
1626 err = -ENOTSUP;
1627 }
1628 return err;
1629 }
1630 #endif
1631
1632 #ifdef CONFIG_FALLOCATE
1633 static int do_fallocate(int fd, int mode, off_t offset, off_t len)
1634 {
1635 do {
1636 if (fallocate(fd, mode, offset, len) == 0) {
1637 return 0;
1638 }
1639 } while (errno == EINTR);
1640 return translate_err(-errno);
1641 }
1642 #endif
1643
1644 static ssize_t handle_aiocb_write_zeroes_block(RawPosixAIOData *aiocb)
1645 {
1646 int ret = -ENOTSUP;
1647 BDRVRawState *s = aiocb->bs->opaque;
1648
1649 if (!s->has_write_zeroes) {
1650 return -ENOTSUP;
1651 }
1652
1653 #ifdef BLKZEROOUT
1654 /* The BLKZEROOUT implementation in the kernel doesn't set
1655 * BLKDEV_ZERO_NOFALLBACK, so we can't call this if we have to avoid slow
1656 * fallbacks. */
1657 if (!(aiocb->aio_type & QEMU_AIO_NO_FALLBACK)) {
1658 do {
1659 uint64_t range[2] = { aiocb->aio_offset, aiocb->aio_nbytes };
1660 if (ioctl(aiocb->aio_fildes, BLKZEROOUT, range) == 0) {
1661 return 0;
1662 }
1663 } while (errno == EINTR);
1664
1665 ret = translate_err(-errno);
1666 if (ret == -ENOTSUP) {
1667 s->has_write_zeroes = false;
1668 }
1669 }
1670 #endif
1671
1672 return ret;
1673 }
1674
1675 static int handle_aiocb_write_zeroes(void *opaque)
1676 {
1677 RawPosixAIOData *aiocb = opaque;
1678 #ifdef CONFIG_FALLOCATE
1679 BDRVRawState *s = aiocb->bs->opaque;
1680 int64_t len;
1681 #endif
1682
1683 if (aiocb->aio_type & QEMU_AIO_BLKDEV) {
1684 return handle_aiocb_write_zeroes_block(aiocb);
1685 }
1686
1687 #ifdef CONFIG_FALLOCATE_ZERO_RANGE
1688 if (s->has_write_zeroes) {
1689 int ret = do_fallocate(s->fd, FALLOC_FL_ZERO_RANGE,
1690 aiocb->aio_offset, aiocb->aio_nbytes);
1691 if (ret == -ENOTSUP) {
1692 s->has_write_zeroes = false;
1693 } else if (ret == 0 || ret != -EINVAL) {
1694 return ret;
1695 }
1696 /*
1697 * Note: Some file systems do not like unaligned byte ranges, and
1698 * return EINVAL in such a case, though they should not do it according
1699 * to the man-page of fallocate(). Thus we simply ignore this return
1700 * value and try the other fallbacks instead.
1701 */
1702 }
1703 #endif
1704
1705 #ifdef CONFIG_FALLOCATE_PUNCH_HOLE
1706 if (s->has_discard && s->has_fallocate) {
1707 int ret = do_fallocate(s->fd,
1708 FALLOC_FL_PUNCH_HOLE | FALLOC_FL_KEEP_SIZE,
1709 aiocb->aio_offset, aiocb->aio_nbytes);
1710 if (ret == 0) {
1711 ret = do_fallocate(s->fd, 0, aiocb->aio_offset, aiocb->aio_nbytes);
1712 if (ret == 0 || ret != -ENOTSUP) {
1713 return ret;
1714 }
1715 s->has_fallocate = false;
1716 } else if (ret == -EINVAL) {
1717 /*
1718 * Some file systems like older versions of GPFS do not like un-
1719 * aligned byte ranges, and return EINVAL in such a case, though
1720 * they should not do it according to the man-page of fallocate().
1721 * Warn about the bad filesystem and try the final fallback instead.
1722 */
1723 warn_report_once("Your file system is misbehaving: "
1724 "fallocate(FALLOC_FL_PUNCH_HOLE) returned EINVAL. "
1725 "Please report this bug to your file system "
1726 "vendor.");
1727 } else if (ret != -ENOTSUP) {
1728 return ret;
1729 } else {
1730 s->has_discard = false;
1731 }
1732 }
1733 #endif
1734
1735 #ifdef CONFIG_FALLOCATE
1736 /* Last resort: we are trying to extend the file with zeroed data. This
1737 * can be done via fallocate(fd, 0) */
1738 len = bdrv_getlength(aiocb->bs);
1739 if (s->has_fallocate && len >= 0 && aiocb->aio_offset >= len) {
1740 int ret = do_fallocate(s->fd, 0, aiocb->aio_offset, aiocb->aio_nbytes);
1741 if (ret == 0 || ret != -ENOTSUP) {
1742 return ret;
1743 }
1744 s->has_fallocate = false;
1745 }
1746 #endif
1747
1748 return -ENOTSUP;
1749 }
1750
1751 static int handle_aiocb_write_zeroes_unmap(void *opaque)
1752 {
1753 RawPosixAIOData *aiocb = opaque;
1754 BDRVRawState *s G_GNUC_UNUSED = aiocb->bs->opaque;
1755
1756 /* First try to write zeros and unmap at the same time */
1757
1758 #ifdef CONFIG_FALLOCATE_PUNCH_HOLE
1759 int ret = do_fallocate(s->fd, FALLOC_FL_PUNCH_HOLE | FALLOC_FL_KEEP_SIZE,
1760 aiocb->aio_offset, aiocb->aio_nbytes);
1761 switch (ret) {
1762 case -ENOTSUP:
1763 case -EINVAL:
1764 case -EBUSY:
1765 break;
1766 default:
1767 return ret;
1768 }
1769 #endif
1770
1771 /* If we couldn't manage to unmap while guaranteed that the area reads as
1772 * all-zero afterwards, just write zeroes without unmapping */
1773 return handle_aiocb_write_zeroes(aiocb);
1774 }
1775
1776 #ifndef HAVE_COPY_FILE_RANGE
1777 static off_t copy_file_range(int in_fd, off_t *in_off, int out_fd,
1778 off_t *out_off, size_t len, unsigned int flags)
1779 {
1780 #ifdef __NR_copy_file_range
1781 return syscall(__NR_copy_file_range, in_fd, in_off, out_fd,
1782 out_off, len, flags);
1783 #else
1784 errno = ENOSYS;
1785 return -1;
1786 #endif
1787 }
1788 #endif
1789
1790 static int handle_aiocb_copy_range(void *opaque)
1791 {
1792 RawPosixAIOData *aiocb = opaque;
1793 uint64_t bytes = aiocb->aio_nbytes;
1794 off_t in_off = aiocb->aio_offset;
1795 off_t out_off = aiocb->copy_range.aio_offset2;
1796
1797 while (bytes) {
1798 ssize_t ret = copy_file_range(aiocb->aio_fildes, &in_off,
1799 aiocb->copy_range.aio_fd2, &out_off,
1800 bytes, 0);
1801 trace_file_copy_file_range(aiocb->bs, aiocb->aio_fildes, in_off,
1802 aiocb->copy_range.aio_fd2, out_off, bytes,
1803 0, ret);
1804 if (ret == 0) {
1805 /* No progress (e.g. when beyond EOF), let the caller fall back to
1806 * buffer I/O. */
1807 return -ENOSPC;
1808 }
1809 if (ret < 0) {
1810 switch (errno) {
1811 case ENOSYS:
1812 return -ENOTSUP;
1813 case EINTR:
1814 continue;
1815 default:
1816 return -errno;
1817 }
1818 }
1819 bytes -= ret;
1820 }
1821 return 0;
1822 }
1823
1824 static int handle_aiocb_discard(void *opaque)
1825 {
1826 RawPosixAIOData *aiocb = opaque;
1827 int ret = -ENOTSUP;
1828 BDRVRawState *s = aiocb->bs->opaque;
1829
1830 if (!s->has_discard) {
1831 return -ENOTSUP;
1832 }
1833
1834 if (aiocb->aio_type & QEMU_AIO_BLKDEV) {
1835 #ifdef BLKDISCARD
1836 do {
1837 uint64_t range[2] = { aiocb->aio_offset, aiocb->aio_nbytes };
1838 if (ioctl(aiocb->aio_fildes, BLKDISCARD, range) == 0) {
1839 return 0;
1840 }
1841 } while (errno == EINTR);
1842
1843 ret = translate_err(-errno);
1844 #endif
1845 } else {
1846 #ifdef CONFIG_FALLOCATE_PUNCH_HOLE
1847 ret = do_fallocate(s->fd, FALLOC_FL_PUNCH_HOLE | FALLOC_FL_KEEP_SIZE,
1848 aiocb->aio_offset, aiocb->aio_nbytes);
1849 ret = translate_err(ret);
1850 #elif defined(__APPLE__) && (__MACH__)
1851 fpunchhole_t fpunchhole;
1852 fpunchhole.fp_flags = 0;
1853 fpunchhole.reserved = 0;
1854 fpunchhole.fp_offset = aiocb->aio_offset;
1855 fpunchhole.fp_length = aiocb->aio_nbytes;
1856 if (fcntl(s->fd, F_PUNCHHOLE, &fpunchhole) == -1) {
1857 ret = errno == ENODEV ? -ENOTSUP : -errno;
1858 } else {
1859 ret = 0;
1860 }
1861 #endif
1862 }
1863
1864 if (ret == -ENOTSUP) {
1865 s->has_discard = false;
1866 }
1867 return ret;
1868 }
1869
1870 /*
1871 * Help alignment probing by allocating the first block.
1872 *
1873 * When reading with direct I/O from unallocated area on Gluster backed by XFS,
1874 * reading succeeds regardless of request length. In this case we fallback to
1875 * safe alignment which is not optimal. Allocating the first block avoids this
1876 * fallback.
1877 *
1878 * fd may be opened with O_DIRECT, but we don't know the buffer alignment or
1879 * request alignment, so we use safe values.
1880 *
1881 * Returns: 0 on success, -errno on failure. Since this is an optimization,
1882 * caller may ignore failures.
1883 */
1884 static int allocate_first_block(int fd, size_t max_size)
1885 {
1886 size_t write_size = (max_size < MAX_BLOCKSIZE)
1887 ? BDRV_SECTOR_SIZE
1888 : MAX_BLOCKSIZE;
1889 size_t max_align = MAX(MAX_BLOCKSIZE, qemu_real_host_page_size);
1890 void *buf;
1891 ssize_t n;
1892 int ret;
1893
1894 buf = qemu_memalign(max_align, write_size);
1895 memset(buf, 0, write_size);
1896
1897 do {
1898 n = pwrite(fd, buf, write_size, 0);
1899 } while (n == -1 && errno == EINTR);
1900
1901 ret = (n == -1) ? -errno : 0;
1902
1903 qemu_vfree(buf);
1904 return ret;
1905 }
1906
1907 static int handle_aiocb_truncate(void *opaque)
1908 {
1909 RawPosixAIOData *aiocb = opaque;
1910 int result = 0;
1911 int64_t current_length = 0;
1912 char *buf = NULL;
1913 struct stat st;
1914 int fd = aiocb->aio_fildes;
1915 int64_t offset = aiocb->aio_offset;
1916 PreallocMode prealloc = aiocb->truncate.prealloc;
1917 Error **errp = aiocb->truncate.errp;
1918
1919 if (fstat(fd, &st) < 0) {
1920 result = -errno;
1921 error_setg_errno(errp, -result, "Could not stat file");
1922 return result;
1923 }
1924
1925 current_length = st.st_size;
1926 if (current_length > offset && prealloc != PREALLOC_MODE_OFF) {
1927 error_setg(errp, "Cannot use preallocation for shrinking files");
1928 return -ENOTSUP;
1929 }
1930
1931 switch (prealloc) {
1932 #ifdef CONFIG_POSIX_FALLOCATE
1933 case PREALLOC_MODE_FALLOC:
1934 /*
1935 * Truncating before posix_fallocate() makes it about twice slower on
1936 * file systems that do not support fallocate(), trying to check if a
1937 * block is allocated before allocating it, so don't do that here.
1938 */
1939 if (offset != current_length) {
1940 result = -posix_fallocate(fd, current_length,
1941 offset - current_length);
1942 if (result != 0) {
1943 /* posix_fallocate() doesn't set errno. */
1944 error_setg_errno(errp, -result,
1945 "Could not preallocate new data");
1946 } else if (current_length == 0) {
1947 /*
1948 * posix_fallocate() uses fallocate() if the filesystem
1949 * supports it, or fallback to manually writing zeroes. If
1950 * fallocate() was used, unaligned reads from the fallocated
1951 * area in raw_probe_alignment() will succeed, hence we need to
1952 * allocate the first block.
1953 *
1954 * Optimize future alignment probing; ignore failures.
1955 */
1956 allocate_first_block(fd, offset);
1957 }
1958 } else {
1959 result = 0;
1960 }
1961 goto out;
1962 #endif
1963 case PREALLOC_MODE_FULL:
1964 {
1965 int64_t num = 0, left = offset - current_length;
1966 off_t seek_result;
1967
1968 /*
1969 * Knowing the final size from the beginning could allow the file
1970 * system driver to do less allocations and possibly avoid
1971 * fragmentation of the file.
1972 */
1973 if (ftruncate(fd, offset) != 0) {
1974 result = -errno;
1975 error_setg_errno(errp, -result, "Could not resize file");
1976 goto out;
1977 }
1978
1979 buf = g_malloc0(65536);
1980
1981 seek_result = lseek(fd, current_length, SEEK_SET);
1982 if (seek_result < 0) {
1983 result = -errno;
1984 error_setg_errno(errp, -result,
1985 "Failed to seek to the old end of file");
1986 goto out;
1987 }
1988
1989 while (left > 0) {
1990 num = MIN(left, 65536);
1991 result = write(fd, buf, num);
1992 if (result < 0) {
1993 if (errno == EINTR) {
1994 continue;
1995 }
1996 result = -errno;
1997 error_setg_errno(errp, -result,
1998 "Could not write zeros for preallocation");
1999 goto out;
2000 }
2001 left -= result;
2002 }
2003 if (result >= 0) {
2004 result = fsync(fd);
2005 if (result < 0) {
2006 result = -errno;
2007 error_setg_errno(errp, -result,
2008 "Could not flush file to disk");
2009 goto out;
2010 }
2011 }
2012 goto out;
2013 }
2014 case PREALLOC_MODE_OFF:
2015 if (ftruncate(fd, offset) != 0) {
2016 result = -errno;
2017 error_setg_errno(errp, -result, "Could not resize file");
2018 } else if (current_length == 0 && offset > current_length) {
2019 /* Optimize future alignment probing; ignore failures. */
2020 allocate_first_block(fd, offset);
2021 }
2022 return result;
2023 default:
2024 result = -ENOTSUP;
2025 error_setg(errp, "Unsupported preallocation mode: %s",
2026 PreallocMode_str(prealloc));
2027 return result;
2028 }
2029
2030 out:
2031 if (result < 0) {
2032 if (ftruncate(fd, current_length) < 0) {
2033 error_report("Failed to restore old file length: %s",
2034 strerror(errno));
2035 }
2036 }
2037
2038 g_free(buf);
2039 return result;
2040 }
2041
2042 static int coroutine_fn raw_thread_pool_submit(BlockDriverState *bs,
2043 ThreadPoolFunc func, void *arg)
2044 {
2045 /* @bs can be NULL, bdrv_get_aio_context() returns the main context then */
2046 ThreadPool *pool = aio_get_thread_pool(bdrv_get_aio_context(bs));
2047 return thread_pool_submit_co(pool, func, arg);
2048 }
2049
2050 static int coroutine_fn raw_co_prw(BlockDriverState *bs, uint64_t offset,
2051 uint64_t bytes, QEMUIOVector *qiov, int type)
2052 {
2053 BDRVRawState *s = bs->opaque;
2054 RawPosixAIOData acb;
2055
2056 if (fd_open(bs) < 0)
2057 return -EIO;
2058
2059 /*
2060 * When using O_DIRECT, the request must be aligned to be able to use
2061 * either libaio or io_uring interface. If not fail back to regular thread
2062 * pool read/write code which emulates this for us if we
2063 * set QEMU_AIO_MISALIGNED.
2064 */
2065 if (s->needs_alignment && !bdrv_qiov_is_aligned(bs, qiov)) {
2066 type |= QEMU_AIO_MISALIGNED;
2067 #ifdef CONFIG_LINUX_IO_URING
2068 } else if (s->use_linux_io_uring) {
2069 LuringState *aio = aio_get_linux_io_uring(bdrv_get_aio_context(bs));
2070 assert(qiov->size == bytes);
2071 return luring_co_submit(bs, aio, s->fd, offset, qiov, type);
2072 #endif
2073 #ifdef CONFIG_LINUX_AIO
2074 } else if (s->use_linux_aio) {
2075 LinuxAioState *aio = aio_get_linux_aio(bdrv_get_aio_context(bs));
2076 assert(qiov->size == bytes);
2077 return laio_co_submit(bs, aio, s->fd, offset, qiov, type,
2078 s->aio_max_batch);
2079 #endif
2080 }
2081
2082 acb = (RawPosixAIOData) {
2083 .bs = bs,
2084 .aio_fildes = s->fd,
2085 .aio_type = type,
2086 .aio_offset = offset,
2087 .aio_nbytes = bytes,
2088 .io = {
2089 .iov = qiov->iov,
2090 .niov = qiov->niov,
2091 },
2092 };
2093
2094 assert(qiov->size == bytes);
2095 return raw_thread_pool_submit(bs, handle_aiocb_rw, &acb);
2096 }
2097
2098 static int coroutine_fn raw_co_preadv(BlockDriverState *bs, int64_t offset,
2099 int64_t bytes, QEMUIOVector *qiov,
2100 BdrvRequestFlags flags)
2101 {
2102 return raw_co_prw(bs, offset, bytes, qiov, QEMU_AIO_READ);
2103 }
2104
2105 static int coroutine_fn raw_co_pwritev(BlockDriverState *bs, int64_t offset,
2106 int64_t bytes, QEMUIOVector *qiov,
2107 BdrvRequestFlags flags)
2108 {
2109 assert(flags == 0);
2110 return raw_co_prw(bs, offset, bytes, qiov, QEMU_AIO_WRITE);
2111 }
2112
2113 static void raw_aio_plug(BlockDriverState *bs)
2114 {
2115 BDRVRawState __attribute__((unused)) *s = bs->opaque;
2116 #ifdef CONFIG_LINUX_AIO
2117 if (s->use_linux_aio) {
2118 LinuxAioState *aio = aio_get_linux_aio(bdrv_get_aio_context(bs));
2119 laio_io_plug(bs, aio);
2120 }
2121 #endif
2122 #ifdef CONFIG_LINUX_IO_URING
2123 if (s->use_linux_io_uring) {
2124 LuringState *aio = aio_get_linux_io_uring(bdrv_get_aio_context(bs));
2125 luring_io_plug(bs, aio);
2126 }
2127 #endif
2128 }
2129
2130 static void raw_aio_unplug(BlockDriverState *bs)
2131 {
2132 BDRVRawState __attribute__((unused)) *s = bs->opaque;
2133 #ifdef CONFIG_LINUX_AIO
2134 if (s->use_linux_aio) {
2135 LinuxAioState *aio = aio_get_linux_aio(bdrv_get_aio_context(bs));
2136 laio_io_unplug(bs, aio, s->aio_max_batch);
2137 }
2138 #endif
2139 #ifdef CONFIG_LINUX_IO_URING
2140 if (s->use_linux_io_uring) {
2141 LuringState *aio = aio_get_linux_io_uring(bdrv_get_aio_context(bs));
2142 luring_io_unplug(bs, aio);
2143 }
2144 #endif
2145 }
2146
2147 static int raw_co_flush_to_disk(BlockDriverState *bs)
2148 {
2149 BDRVRawState *s = bs->opaque;
2150 RawPosixAIOData acb;
2151 int ret;
2152
2153 ret = fd_open(bs);
2154 if (ret < 0) {
2155 return ret;
2156 }
2157
2158 acb = (RawPosixAIOData) {
2159 .bs = bs,
2160 .aio_fildes = s->fd,
2161 .aio_type = QEMU_AIO_FLUSH,
2162 };
2163
2164 #ifdef CONFIG_LINUX_IO_URING
2165 if (s->use_linux_io_uring) {
2166 LuringState *aio = aio_get_linux_io_uring(bdrv_get_aio_context(bs));
2167 return luring_co_submit(bs, aio, s->fd, 0, NULL, QEMU_AIO_FLUSH);
2168 }
2169 #endif
2170 return raw_thread_pool_submit(bs, handle_aiocb_flush, &acb);
2171 }
2172
2173 static void raw_aio_attach_aio_context(BlockDriverState *bs,
2174 AioContext *new_context)
2175 {
2176 BDRVRawState __attribute__((unused)) *s = bs->opaque;
2177 #ifdef CONFIG_LINUX_AIO
2178 if (s->use_linux_aio) {
2179 Error *local_err = NULL;
2180 if (!aio_setup_linux_aio(new_context, &local_err)) {
2181 error_reportf_err(local_err, "Unable to use native AIO, "
2182 "falling back to thread pool: ");
2183 s->use_linux_aio = false;
2184 }
2185 }
2186 #endif
2187 #ifdef CONFIG_LINUX_IO_URING
2188 if (s->use_linux_io_uring) {
2189 Error *local_err = NULL;
2190 if (!aio_setup_linux_io_uring(new_context, &local_err)) {
2191 error_reportf_err(local_err, "Unable to use linux io_uring, "
2192 "falling back to thread pool: ");
2193 s->use_linux_io_uring = false;
2194 }
2195 }
2196 #endif
2197 }
2198
2199 static void raw_close(BlockDriverState *bs)
2200 {
2201 BDRVRawState *s = bs->opaque;
2202
2203 if (s->fd >= 0) {
2204 qemu_close(s->fd);
2205 s->fd = -1;
2206 }
2207 }
2208
2209 /**
2210 * Truncates the given regular file @fd to @offset and, when growing, fills the
2211 * new space according to @prealloc.
2212 *
2213 * Returns: 0 on success, -errno on failure.
2214 */
2215 static int coroutine_fn
2216 raw_regular_truncate(BlockDriverState *bs, int fd, int64_t offset,
2217 PreallocMode prealloc, Error **errp)
2218 {
2219 RawPosixAIOData acb;
2220
2221 acb = (RawPosixAIOData) {
2222 .bs = bs,
2223 .aio_fildes = fd,
2224 .aio_type = QEMU_AIO_TRUNCATE,
2225 .aio_offset = offset,
2226 .truncate = {
2227 .prealloc = prealloc,
2228 .errp = errp,
2229 },
2230 };
2231
2232 return raw_thread_pool_submit(bs, handle_aiocb_truncate, &acb);
2233 }
2234
2235 static int coroutine_fn raw_co_truncate(BlockDriverState *bs, int64_t offset,
2236 bool exact, PreallocMode prealloc,
2237 BdrvRequestFlags flags, Error **errp)
2238 {
2239 BDRVRawState *s = bs->opaque;
2240 struct stat st;
2241 int ret;
2242
2243 if (fstat(s->fd, &st)) {
2244 ret = -errno;
2245 error_setg_errno(errp, -ret, "Failed to fstat() the file");
2246 return ret;
2247 }
2248
2249 if (S_ISREG(st.st_mode)) {
2250 /* Always resizes to the exact @offset */
2251 return raw_regular_truncate(bs, s->fd, offset, prealloc, errp);
2252 }
2253
2254 if (prealloc != PREALLOC_MODE_OFF) {
2255 error_setg(errp, "Preallocation mode '%s' unsupported for this "
2256 "non-regular file", PreallocMode_str(prealloc));
2257 return -ENOTSUP;
2258 }
2259
2260 if (S_ISCHR(st.st_mode) || S_ISBLK(st.st_mode)) {
2261 int64_t cur_length = raw_getlength(bs);
2262
2263 if (offset != cur_length && exact) {
2264 error_setg(errp, "Cannot resize device files");
2265 return -ENOTSUP;
2266 } else if (offset > cur_length) {
2267 error_setg(errp, "Cannot grow device files");
2268 return -EINVAL;
2269 }
2270 } else {
2271 error_setg(errp, "Resizing this file is not supported");
2272 return -ENOTSUP;
2273 }
2274
2275 return 0;
2276 }
2277
2278 #ifdef __OpenBSD__
2279 static int64_t raw_getlength(BlockDriverState *bs)
2280 {
2281 BDRVRawState *s = bs->opaque;
2282 int fd = s->fd;
2283 struct stat st;
2284
2285 if (fstat(fd, &st))
2286 return -errno;
2287 if (S_ISCHR(st.st_mode) || S_ISBLK(st.st_mode)) {
2288 struct disklabel dl;
2289
2290 if (ioctl(fd, DIOCGDINFO, &dl))
2291 return -errno;
2292 return (uint64_t)dl.d_secsize *
2293 dl.d_partitions[DISKPART(st.st_rdev)].p_size;
2294 } else
2295 return st.st_size;
2296 }
2297 #elif defined(__NetBSD__)
2298 static int64_t raw_getlength(BlockDriverState *bs)
2299 {
2300 BDRVRawState *s = bs->opaque;
2301 int fd = s->fd;
2302 struct stat st;
2303
2304 if (fstat(fd, &st))
2305 return -errno;
2306 if (S_ISCHR(st.st_mode) || S_ISBLK(st.st_mode)) {
2307 struct dkwedge_info dkw;
2308
2309 if (ioctl(fd, DIOCGWEDGEINFO, &dkw) != -1) {
2310 return dkw.dkw_size * 512;
2311 } else {
2312 struct disklabel dl;
2313
2314 if (ioctl(fd, DIOCGDINFO, &dl))
2315 return -errno;
2316 return (uint64_t)dl.d_secsize *
2317 dl.d_partitions[DISKPART(st.st_rdev)].p_size;
2318 }
2319 } else
2320 return st.st_size;
2321 }
2322 #elif defined(__sun__)
2323 static int64_t raw_getlength(BlockDriverState *bs)
2324 {
2325 BDRVRawState *s = bs->opaque;
2326 struct dk_minfo minfo;
2327 int ret;
2328 int64_t size;
2329
2330 ret = fd_open(bs);
2331 if (ret < 0) {
2332 return ret;
2333 }
2334
2335 /*
2336 * Use the DKIOCGMEDIAINFO ioctl to read the size.
2337 */
2338 ret = ioctl(s->fd, DKIOCGMEDIAINFO, &minfo);
2339 if (ret != -1) {
2340 return minfo.dki_lbsize * minfo.dki_capacity;
2341 }
2342
2343 /*
2344 * There are reports that lseek on some devices fails, but
2345 * irc discussion said that contingency on contingency was overkill.
2346 */
2347 size = lseek(s->fd, 0, SEEK_END);
2348 if (size < 0) {
2349 return -errno;
2350 }
2351 return size;
2352 }
2353 #elif defined(CONFIG_BSD)
2354 static int64_t raw_getlength(BlockDriverState *bs)
2355 {
2356 BDRVRawState *s = bs->opaque;
2357 int fd = s->fd;
2358 int64_t size;
2359 struct stat sb;
2360 #if defined (__FreeBSD__) || defined(__FreeBSD_kernel__)
2361 int reopened = 0;
2362 #endif
2363 int ret;
2364
2365 ret = fd_open(bs);
2366 if (ret < 0)
2367 return ret;
2368
2369 #if defined (__FreeBSD__) || defined(__FreeBSD_kernel__)
2370 again:
2371 #endif
2372 if (!fstat(fd, &sb) && (S_IFCHR & sb.st_mode)) {
2373 size = 0;
2374 #ifdef DIOCGMEDIASIZE
2375 if (ioctl(fd, DIOCGMEDIASIZE, (off_t *)&size)) {
2376 size = 0;
2377 }
2378 #endif
2379 #ifdef DIOCGPART
2380 if (size == 0) {
2381 struct partinfo pi;
2382 if (ioctl(fd, DIOCGPART, &pi) == 0) {
2383 size = pi.media_size;
2384 }
2385 }
2386 #endif
2387 #if defined(DKIOCGETBLOCKCOUNT) && defined(DKIOCGETBLOCKSIZE)
2388 if (size == 0) {
2389 uint64_t sectors = 0;
2390 uint32_t sector_size = 0;
2391
2392 if (ioctl(fd, DKIOCGETBLOCKCOUNT, &sectors) == 0
2393 && ioctl(fd, DKIOCGETBLOCKSIZE, &sector_size) == 0) {
2394 size = sectors * sector_size;
2395 }
2396 }
2397 #endif
2398 if (size == 0) {
2399 size = lseek(fd, 0LL, SEEK_END);
2400 }
2401 if (size < 0) {
2402 return -errno;
2403 }
2404 #if defined(__FreeBSD__) || defined(__FreeBSD_kernel__)
2405 switch(s->type) {
2406 case FTYPE_CD:
2407 /* XXX FreeBSD acd returns UINT_MAX sectors for an empty drive */
2408 if (size == 2048LL * (unsigned)-1)
2409 size = 0;
2410 /* XXX no disc? maybe we need to reopen... */
2411 if (size <= 0 && !reopened && cdrom_reopen(bs) >= 0) {
2412 reopened = 1;
2413 goto again;
2414 }
2415 }
2416 #endif
2417 } else {
2418 size = lseek(fd, 0, SEEK_END);
2419 if (size < 0) {
2420 return -errno;
2421 }
2422 }
2423 return size;
2424 }
2425 #else
2426 static int64_t raw_getlength(BlockDriverState *bs)
2427 {
2428 BDRVRawState *s = bs->opaque;
2429 int ret;
2430 int64_t size;
2431
2432 ret = fd_open(bs);
2433 if (ret < 0) {
2434 return ret;
2435 }
2436
2437 size = lseek(s->fd, 0, SEEK_END);
2438 if (size < 0) {
2439 return -errno;
2440 }
2441 return size;
2442 }
2443 #endif
2444
2445 static int64_t raw_get_allocated_file_size(BlockDriverState *bs)
2446 {
2447 struct stat st;
2448 BDRVRawState *s = bs->opaque;
2449
2450 if (fstat(s->fd, &st) < 0) {
2451 return -errno;
2452 }
2453 return (int64_t)st.st_blocks * 512;
2454 }
2455
2456 static int coroutine_fn
2457 raw_co_create(BlockdevCreateOptions *options, Error **errp)
2458 {
2459 BlockdevCreateOptionsFile *file_opts;
2460 Error *local_err = NULL;
2461 int fd;
2462 uint64_t perm, shared;
2463 int result = 0;
2464
2465 /* Validate options and set default values */
2466 assert(options->driver == BLOCKDEV_DRIVER_FILE);
2467 file_opts = &options->u.file;
2468
2469 if (!file_opts->has_nocow) {
2470 file_opts->nocow = false;
2471 }
2472 if (!file_opts->has_preallocation) {
2473 file_opts->preallocation = PREALLOC_MODE_OFF;
2474 }
2475 if (!file_opts->has_extent_size_hint) {
2476 file_opts->extent_size_hint = 1 * MiB;
2477 }
2478 if (file_opts->extent_size_hint > UINT32_MAX) {
2479 result = -EINVAL;
2480 error_setg(errp, "Extent size hint is too large");
2481 goto out;
2482 }
2483
2484 /* Create file */
2485 fd = qemu_create(file_opts->filename, O_RDWR | O_BINARY, 0644, errp);
2486 if (fd < 0) {
2487 result = -errno;
2488 goto out;
2489 }
2490
2491 /* Take permissions: We want to discard everything, so we need
2492 * BLK_PERM_WRITE; and truncation to the desired size requires
2493 * BLK_PERM_RESIZE.
2494 * On the other hand, we cannot share the RESIZE permission
2495 * because we promise that after this function, the file has the
2496 * size given in the options. If someone else were to resize it
2497 * concurrently, we could not guarantee that.
2498 * Note that after this function, we can no longer guarantee that
2499 * the file is not touched by a third party, so it may be resized
2500 * then. */
2501 perm = BLK_PERM_WRITE | BLK_PERM_RESIZE;
2502 shared = BLK_PERM_ALL & ~BLK_PERM_RESIZE;
2503
2504 /* Step one: Take locks */
2505 result = raw_apply_lock_bytes(NULL, fd, perm, ~shared, false, errp);
2506 if (result < 0) {
2507 goto out_close;
2508 }
2509
2510 /* Step two: Check that nobody else has taken conflicting locks */
2511 result = raw_check_lock_bytes(fd, perm, shared, errp);
2512 if (result < 0) {
2513 error_append_hint(errp,
2514 "Is another process using the image [%s]?\n",
2515 file_opts->filename);
2516 goto out_unlock;
2517 }
2518
2519 /* Clear the file by truncating it to 0 */
2520 result = raw_regular_truncate(NULL, fd, 0, PREALLOC_MODE_OFF, errp);
2521 if (result < 0) {
2522 goto out_unlock;
2523 }
2524
2525 if (file_opts->nocow) {
2526 #ifdef __linux__
2527 /* Set NOCOW flag to solve performance issue on fs like btrfs.
2528 * This is an optimisation. The FS_IOC_SETFLAGS ioctl return value
2529 * will be ignored since any failure of this operation should not
2530 * block the left work.
2531 */
2532 int attr;
2533 if (ioctl(fd, FS_IOC_GETFLAGS, &attr) == 0) {
2534 attr |= FS_NOCOW_FL;
2535 ioctl(fd, FS_IOC_SETFLAGS, &attr);
2536 }
2537 #endif
2538 }
2539 #ifdef FS_IOC_FSSETXATTR
2540 /*
2541 * Try to set the extent size hint. Failure is not fatal, and a warning is
2542 * only printed if the option was explicitly specified.
2543 */
2544 {
2545 struct fsxattr attr;
2546 result = ioctl(fd, FS_IOC_FSGETXATTR, &attr);
2547 if (result == 0) {
2548 attr.fsx_xflags |= FS_XFLAG_EXTSIZE;
2549 attr.fsx_extsize = file_opts->extent_size_hint;
2550 result = ioctl(fd, FS_IOC_FSSETXATTR, &attr);
2551 }
2552 if (result < 0 && file_opts->has_extent_size_hint &&
2553 file_opts->extent_size_hint)
2554 {
2555 warn_report("Failed to set extent size hint: %s",
2556 strerror(errno));
2557 }
2558 }
2559 #endif
2560
2561 /* Resize and potentially preallocate the file to the desired
2562 * final size */
2563 result = raw_regular_truncate(NULL, fd, file_opts->size,
2564 file_opts->preallocation, errp);
2565 if (result < 0) {
2566 goto out_unlock;
2567 }
2568
2569 out_unlock:
2570 raw_apply_lock_bytes(NULL, fd, 0, 0, true, &local_err);
2571 if (local_err) {
2572 /* The above call should not fail, and if it does, that does
2573 * not mean the whole creation operation has failed. So
2574 * report it the user for their convenience, but do not report
2575 * it to the caller. */
2576 warn_report_err(local_err);
2577 }
2578
2579 out_close:
2580 if (qemu_close(fd) != 0 && result == 0) {
2581 result = -errno;
2582 error_setg_errno(errp, -result, "Could not close the new file");
2583 }
2584 out:
2585 return result;
2586 }
2587
2588 static int coroutine_fn raw_co_create_opts(BlockDriver *drv,
2589 const char *filename,
2590 QemuOpts *opts,
2591 Error **errp)
2592 {
2593 BlockdevCreateOptions options;
2594 int64_t total_size = 0;
2595 int64_t extent_size_hint = 0;
2596 bool has_extent_size_hint = false;
2597 bool nocow = false;
2598 PreallocMode prealloc;
2599 char *buf = NULL;
2600 Error *local_err = NULL;
2601
2602 /* Skip file: protocol prefix */
2603 strstart(filename, "file:", &filename);
2604
2605 /* Read out options */
2606 total_size = ROUND_UP(qemu_opt_get_size_del(opts, BLOCK_OPT_SIZE, 0),
2607 BDRV_SECTOR_SIZE);
2608 if (qemu_opt_get(opts, BLOCK_OPT_EXTENT_SIZE_HINT)) {
2609 has_extent_size_hint = true;
2610 extent_size_hint =
2611 qemu_opt_get_size_del(opts, BLOCK_OPT_EXTENT_SIZE_HINT, -1);
2612 }
2613 nocow = qemu_opt_get_bool(opts, BLOCK_OPT_NOCOW, false);
2614 buf = qemu_opt_get_del(opts, BLOCK_OPT_PREALLOC);
2615 prealloc = qapi_enum_parse(&PreallocMode_lookup, buf,
2616 PREALLOC_MODE_OFF, &local_err);
2617 g_free(buf);
2618 if (local_err) {
2619 error_propagate(errp, local_err);
2620 return -EINVAL;
2621 }
2622
2623 options = (BlockdevCreateOptions) {
2624 .driver = BLOCKDEV_DRIVER_FILE,
2625 .u.file = {
2626 .filename = (char *) filename,
2627 .size = total_size,
2628 .has_preallocation = true,
2629 .preallocation = prealloc,
2630 .has_nocow = true,
2631 .nocow = nocow,
2632 .has_extent_size_hint = has_extent_size_hint,
2633 .extent_size_hint = extent_size_hint,
2634 },
2635 };
2636 return raw_co_create(&options, errp);
2637 }
2638
2639 static int coroutine_fn raw_co_delete_file(BlockDriverState *bs,
2640 Error **errp)
2641 {
2642 struct stat st;
2643 int ret;
2644
2645 if (!(stat(bs->filename, &st) == 0) || !S_ISREG(st.st_mode)) {
2646 error_setg_errno(errp, ENOENT, "%s is not a regular file",
2647 bs->filename);
2648 return -ENOENT;
2649 }
2650
2651 ret = unlink(bs->filename);
2652 if (ret < 0) {
2653 ret = -errno;
2654 error_setg_errno(errp, -ret, "Error when deleting file %s",
2655 bs->filename);
2656 }
2657
2658 return ret;
2659 }
2660
2661 /*
2662 * Find allocation range in @bs around offset @start.
2663 * May change underlying file descriptor's file offset.
2664 * If @start is not in a hole, store @start in @data, and the
2665 * beginning of the next hole in @hole, and return 0.
2666 * If @start is in a non-trailing hole, store @start in @hole and the
2667 * beginning of the next non-hole in @data, and return 0.
2668 * If @start is in a trailing hole or beyond EOF, return -ENXIO.
2669 * If we can't find out, return a negative errno other than -ENXIO.
2670 */
2671 static int find_allocation(BlockDriverState *bs, off_t start,
2672 off_t *data, off_t *hole)
2673 {
2674 #if defined SEEK_HOLE && defined SEEK_DATA
2675 BDRVRawState *s = bs->opaque;
2676 off_t offs;
2677
2678 /*
2679 * SEEK_DATA cases:
2680 * D1. offs == start: start is in data
2681 * D2. offs > start: start is in a hole, next data at offs
2682 * D3. offs < 0, errno = ENXIO: either start is in a trailing hole
2683 * or start is beyond EOF
2684 * If the latter happens, the file has been truncated behind
2685 * our back since we opened it. All bets are off then.
2686 * Treating like a trailing hole is simplest.
2687 * D4. offs < 0, errno != ENXIO: we learned nothing
2688 */
2689 offs = lseek(s->fd, start, SEEK_DATA);
2690 if (offs < 0) {
2691 return -errno; /* D3 or D4 */
2692 }
2693
2694 if (offs < start) {
2695 /* This is not a valid return by lseek(). We are safe to just return
2696 * -EIO in this case, and we'll treat it like D4. */
2697 return -EIO;
2698 }
2699
2700 if (offs > start) {
2701 /* D2: in hole, next data at offs */
2702 *hole = start;
2703 *data = offs;
2704 return 0;
2705 }
2706
2707 /* D1: in data, end not yet known */
2708
2709 /*
2710 * SEEK_HOLE cases:
2711 * H1. offs == start: start is in a hole
2712 * If this happens here, a hole has been dug behind our back
2713 * since the previous lseek().
2714 * H2. offs > start: either start is in data, next hole at offs,
2715 * or start is in trailing hole, EOF at offs
2716 * Linux treats trailing holes like any other hole: offs ==
2717 * start. Solaris seeks to EOF instead: offs > start (blech).
2718 * If that happens here, a hole has been dug behind our back
2719 * since the previous lseek().
2720 * H3. offs < 0, errno = ENXIO: start is beyond EOF
2721 * If this happens, the file has been truncated behind our
2722 * back since we opened it. Treat it like a trailing hole.
2723 * H4. offs < 0, errno != ENXIO: we learned nothing
2724 * Pretend we know nothing at all, i.e. "forget" about D1.
2725 */
2726 offs = lseek(s->fd, start, SEEK_HOLE);
2727 if (offs < 0) {
2728 return -errno; /* D1 and (H3 or H4) */
2729 }
2730
2731 if (offs < start) {
2732 /* This is not a valid return by lseek(). We are safe to just return
2733 * -EIO in this case, and we'll treat it like H4. */
2734 return -EIO;
2735 }
2736
2737 if (offs > start) {
2738 /*
2739 * D1 and H2: either in data, next hole at offs, or it was in
2740 * data but is now in a trailing hole. In the latter case,
2741 * all bets are off. Treating it as if it there was data all
2742 * the way to EOF is safe, so simply do that.
2743 */
2744 *data = start;
2745 *hole = offs;
2746 return 0;
2747 }
2748
2749 /* D1 and H1 */
2750 return -EBUSY;
2751 #else
2752 return -ENOTSUP;
2753 #endif
2754 }
2755
2756 /*
2757 * Returns the allocation status of the specified offset.
2758 *
2759 * The block layer guarantees 'offset' and 'bytes' are within bounds.
2760 *
2761 * 'pnum' is set to the number of bytes (including and immediately following
2762 * the specified offset) that are known to be in the same
2763 * allocated/unallocated state.
2764 *
2765 * 'bytes' is a soft cap for 'pnum'. If the information is free, 'pnum' may
2766 * well exceed it.
2767 */
2768 static int coroutine_fn raw_co_block_status(BlockDriverState *bs,
2769 bool want_zero,
2770 int64_t offset,
2771 int64_t bytes, int64_t *pnum,
2772 int64_t *map,
2773 BlockDriverState **file)
2774 {
2775 off_t data = 0, hole = 0;
2776 int ret;
2777
2778 assert(QEMU_IS_ALIGNED(offset | bytes, bs->bl.request_alignment));
2779
2780 ret = fd_open(bs);
2781 if (ret < 0) {
2782 return ret;
2783 }
2784
2785 if (!want_zero) {
2786 *pnum = bytes;
2787 *map = offset;
2788 *file = bs;
2789 return BDRV_BLOCK_DATA | BDRV_BLOCK_OFFSET_VALID;
2790 }
2791
2792 ret = find_allocation(bs, offset, &data, &hole);
2793 if (ret == -ENXIO) {
2794 /* Trailing hole */
2795 *pnum = bytes;
2796 ret = BDRV_BLOCK_ZERO;
2797 } else if (ret < 0) {
2798 /* No info available, so pretend there are no holes */
2799 *pnum = bytes;
2800 ret = BDRV_BLOCK_DATA;
2801 } else if (data == offset) {
2802 /* On a data extent, compute bytes to the end of the extent,
2803 * possibly including a partial sector at EOF. */
2804 *pnum = hole - offset;
2805
2806 /*
2807 * We are not allowed to return partial sectors, though, so
2808 * round up if necessary.
2809 */
2810 if (!QEMU_IS_ALIGNED(*pnum, bs->bl.request_alignment)) {
2811 int64_t file_length = raw_getlength(bs);
2812 if (file_length > 0) {
2813 /* Ignore errors, this is just a safeguard */
2814 assert(hole == file_length);
2815 }
2816 *pnum = ROUND_UP(*pnum, bs->bl.request_alignment);
2817 }
2818
2819 ret = BDRV_BLOCK_DATA;
2820 } else {
2821 /* On a hole, compute bytes to the beginning of the next extent. */
2822 assert(hole == offset);
2823 *pnum = data - offset;
2824 ret = BDRV_BLOCK_ZERO;
2825 }
2826 *map = offset;
2827 *file = bs;
2828 return ret | BDRV_BLOCK_OFFSET_VALID;
2829 }
2830
2831 #if defined(__linux__)
2832 /* Verify that the file is not in the page cache */
2833 static void check_cache_dropped(BlockDriverState *bs, Error **errp)
2834 {
2835 const size_t window_size = 128 * 1024 * 1024;
2836 BDRVRawState *s = bs->opaque;
2837 void *window = NULL;
2838 size_t length = 0;
2839 unsigned char *vec;
2840 size_t page_size;
2841 off_t offset;
2842 off_t end;
2843
2844 /* mincore(2) page status information requires 1 byte per page */
2845 page_size = sysconf(_SC_PAGESIZE);
2846 vec = g_malloc(DIV_ROUND_UP(window_size, page_size));
2847
2848 end = raw_getlength(bs);
2849
2850 for (offset = 0; offset < end; offset += window_size) {
2851 void *new_window;
2852 size_t new_length;
2853 size_t vec_end;
2854 size_t i;
2855 int ret;
2856
2857 /* Unmap previous window if size has changed */
2858 new_length = MIN(end - offset, window_size);
2859 if (new_length != length) {
2860 munmap(window, length);
2861 window = NULL;
2862 length = 0;
2863 }
2864
2865 new_window = mmap(window, new_length, PROT_NONE, MAP_PRIVATE,
2866 s->fd, offset);
2867 if (new_window == MAP_FAILED) {
2868 error_setg_errno(errp, errno, "mmap failed");
2869 break;
2870 }
2871
2872 window = new_window;
2873 length = new_length;
2874
2875 ret = mincore(window, length, vec);
2876 if (ret < 0) {
2877 error_setg_errno(errp, errno, "mincore failed");
2878 break;
2879 }
2880
2881 vec_end = DIV_ROUND_UP(length, page_size);
2882 for (i = 0; i < vec_end; i++) {
2883 if (vec[i] & 0x1) {
2884 break;
2885 }
2886 }
2887 if (i < vec_end) {
2888 error_setg(errp, "page cache still in use!");
2889 break;
2890 }
2891 }
2892
2893 if (window) {
2894 munmap(window, length);
2895 }
2896
2897 g_free(vec);
2898 }
2899 #endif /* __linux__ */
2900
2901 static void coroutine_fn raw_co_invalidate_cache(BlockDriverState *bs,
2902 Error **errp)
2903 {
2904 BDRVRawState *s = bs->opaque;
2905 int ret;
2906
2907 ret = fd_open(bs);
2908 if (ret < 0) {
2909 error_setg_errno(errp, -ret, "The file descriptor is not open");
2910 return;
2911 }
2912
2913 if (!s->drop_cache) {
2914 return;
2915 }
2916
2917 if (s->open_flags & O_DIRECT) {
2918 return; /* No host kernel page cache */
2919 }
2920
2921 #if defined(__linux__)
2922 /* This sets the scene for the next syscall... */
2923 ret = bdrv_co_flush(bs);
2924 if (ret < 0) {
2925 error_setg_errno(errp, -ret, "flush failed");
2926 return;
2927 }
2928
2929 /* Linux does not invalidate pages that are dirty, locked, or mmapped by a
2930 * process. These limitations are okay because we just fsynced the file,
2931 * we don't use mmap, and the file should not be in use by other processes.
2932 */
2933 ret = posix_fadvise(s->fd, 0, 0, POSIX_FADV_DONTNEED);
2934 if (ret != 0) { /* the return value is a positive errno */
2935 error_setg_errno(errp, ret, "fadvise failed");
2936 return;
2937 }
2938
2939 if (s->check_cache_dropped) {
2940 check_cache_dropped(bs, errp);
2941 }
2942 #else /* __linux__ */
2943 /* Do nothing. Live migration to a remote host with cache.direct=off is
2944 * unsupported on other host operating systems. Cache consistency issues
2945 * may occur but no error is reported here, partly because that's the
2946 * historical behavior and partly because it's hard to differentiate valid
2947 * configurations that should not cause errors.
2948 */
2949 #endif /* !__linux__ */
2950 }
2951
2952 static void raw_account_discard(BDRVRawState *s, uint64_t nbytes, int ret)
2953 {
2954 if (ret) {
2955 s->stats.discard_nb_failed++;
2956 } else {
2957 s->stats.discard_nb_ok++;
2958 s->stats.discard_bytes_ok += nbytes;
2959 }
2960 }
2961
2962 static coroutine_fn int
2963 raw_do_pdiscard(BlockDriverState *bs, int64_t offset, int64_t bytes,
2964 bool blkdev)
2965 {
2966 BDRVRawState *s = bs->opaque;
2967 RawPosixAIOData acb;
2968 int ret;
2969
2970 acb = (RawPosixAIOData) {
2971 .bs = bs,
2972 .aio_fildes = s->fd,
2973 .aio_type = QEMU_AIO_DISCARD,
2974 .aio_offset = offset,
2975 .aio_nbytes = bytes,
2976 };
2977
2978 if (blkdev) {
2979 acb.aio_type |= QEMU_AIO_BLKDEV;
2980 }
2981
2982 ret = raw_thread_pool_submit(bs, handle_aiocb_discard, &acb);
2983 raw_account_discard(s, bytes, ret);
2984 return ret;
2985 }
2986
2987 static coroutine_fn int
2988 raw_co_pdiscard(BlockDriverState *bs, int64_t offset, int64_t bytes)
2989 {
2990 return raw_do_pdiscard(bs, offset, bytes, false);
2991 }
2992
2993 static int coroutine_fn
2994 raw_do_pwrite_zeroes(BlockDriverState *bs, int64_t offset, int64_t bytes,
2995 BdrvRequestFlags flags, bool blkdev)
2996 {
2997 BDRVRawState *s = bs->opaque;
2998 RawPosixAIOData acb;
2999 ThreadPoolFunc *handler;
3000
3001 #ifdef CONFIG_FALLOCATE
3002 if (offset + bytes > bs->total_sectors * BDRV_SECTOR_SIZE) {
3003 BdrvTrackedRequest *req;
3004
3005 /*
3006 * This is a workaround for a bug in the Linux XFS driver,
3007 * where writes submitted through the AIO interface will be
3008 * discarded if they happen beyond a concurrently running
3009 * fallocate() that increases the file length (i.e., both the
3010 * write and the fallocate() happen beyond the EOF).
3011 *
3012 * To work around it, we extend the tracked request for this
3013 * zero write until INT64_MAX (effectively infinity), and mark
3014 * it as serializing.
3015 *
3016 * We have to enable this workaround for all filesystems and
3017 * AIO modes (not just XFS with aio=native), because for
3018 * remote filesystems we do not know the host configuration.
3019 */
3020
3021 req = bdrv_co_get_self_request(bs);
3022 assert(req);
3023 assert(req->type == BDRV_TRACKED_WRITE);
3024 assert(req->offset <= offset);
3025 assert(req->offset + req->bytes >= offset + bytes);
3026
3027 req->bytes = BDRV_MAX_LENGTH - req->offset;
3028
3029 bdrv_check_request(req->offset, req->bytes, &error_abort);
3030
3031 bdrv_make_request_serialising(req, bs->bl.request_alignment);
3032 }
3033 #endif
3034
3035 acb = (RawPosixAIOData) {
3036 .bs = bs,
3037 .aio_fildes = s->fd,
3038 .aio_type = QEMU_AIO_WRITE_ZEROES,
3039 .aio_offset = offset,
3040 .aio_nbytes = bytes,
3041 };
3042
3043 if (blkdev) {
3044 acb.aio_type |= QEMU_AIO_BLKDEV;
3045 }
3046 if (flags & BDRV_REQ_NO_FALLBACK) {
3047 acb.aio_type |= QEMU_AIO_NO_FALLBACK;
3048 }
3049
3050 if (flags & BDRV_REQ_MAY_UNMAP) {
3051 acb.aio_type |= QEMU_AIO_DISCARD;
3052 handler = handle_aiocb_write_zeroes_unmap;
3053 } else {
3054 handler = handle_aiocb_write_zeroes;
3055 }
3056
3057 return raw_thread_pool_submit(bs, handler, &acb);
3058 }
3059
3060 static int coroutine_fn raw_co_pwrite_zeroes(
3061 BlockDriverState *bs, int64_t offset,
3062 int64_t bytes, BdrvRequestFlags flags)
3063 {
3064 return raw_do_pwrite_zeroes(bs, offset, bytes, flags, false);
3065 }
3066
3067 static int raw_get_info(BlockDriverState *bs, BlockDriverInfo *bdi)
3068 {
3069 return 0;
3070 }
3071
3072 static BlockStatsSpecificFile get_blockstats_specific_file(BlockDriverState *bs)
3073 {
3074 BDRVRawState *s = bs->opaque;
3075 return (BlockStatsSpecificFile) {
3076 .discard_nb_ok = s->stats.discard_nb_ok,
3077 .discard_nb_failed = s->stats.discard_nb_failed,
3078 .discard_bytes_ok = s->stats.discard_bytes_ok,
3079 };
3080 }
3081
3082 static BlockStatsSpecific *raw_get_specific_stats(BlockDriverState *bs)
3083 {
3084 BlockStatsSpecific *stats = g_new(BlockStatsSpecific, 1);
3085
3086 stats->driver = BLOCKDEV_DRIVER_FILE;
3087 stats->u.file = get_blockstats_specific_file(bs);
3088
3089 return stats;
3090 }
3091
3092 #if defined(HAVE_HOST_BLOCK_DEVICE)
3093 static BlockStatsSpecific *hdev_get_specific_stats(BlockDriverState *bs)
3094 {
3095 BlockStatsSpecific *stats = g_new(BlockStatsSpecific, 1);
3096
3097 stats->driver = BLOCKDEV_DRIVER_HOST_DEVICE;
3098 stats->u.host_device = get_blockstats_specific_file(bs);
3099
3100 return stats;
3101 }
3102 #endif /* HAVE_HOST_BLOCK_DEVICE */
3103
3104 static QemuOptsList raw_create_opts = {
3105 .name = "raw-create-opts",
3106 .head = QTAILQ_HEAD_INITIALIZER(raw_create_opts.head),
3107 .desc = {
3108 {
3109 .name = BLOCK_OPT_SIZE,
3110 .type = QEMU_OPT_SIZE,
3111 .help = "Virtual disk size"
3112 },
3113 {
3114 .name = BLOCK_OPT_NOCOW,
3115 .type = QEMU_OPT_BOOL,
3116 .help = "Turn off copy-on-write (valid only on btrfs)"
3117 },
3118 {
3119 .name = BLOCK_OPT_PREALLOC,
3120 .type = QEMU_OPT_STRING,
3121 .help = "Preallocation mode (allowed values: off"
3122 #ifdef CONFIG_POSIX_FALLOCATE
3123 ", falloc"
3124 #endif
3125 ", full)"
3126 },
3127 {
3128 .name = BLOCK_OPT_EXTENT_SIZE_HINT,
3129 .type = QEMU_OPT_SIZE,
3130 .help = "Extent size hint for the image file, 0 to disable"
3131 },
3132 { /* end of list */ }
3133 }
3134 };
3135
3136 static int raw_check_perm(BlockDriverState *bs, uint64_t perm, uint64_t shared,
3137 Error **errp)
3138 {
3139 BDRVRawState *s = bs->opaque;
3140 int input_flags = s->reopen_state ? s->reopen_state->flags : bs->open_flags;
3141 int open_flags;
3142 int ret;
3143
3144 /* We may need a new fd if auto-read-only switches the mode */
3145 ret = raw_reconfigure_getfd(bs, input_flags, &open_flags, perm,
3146 false, errp);
3147 if (ret < 0) {
3148 return ret;
3149 } else if (ret != s->fd) {
3150 Error *local_err = NULL;
3151
3152 /*
3153 * Fail already check_perm() if we can't get a working O_DIRECT
3154 * alignment with the new fd.
3155 */
3156 raw_probe_alignment(bs, ret, &local_err);
3157 if (local_err) {
3158 error_propagate(errp, local_err);
3159 return -EINVAL;
3160 }
3161
3162 s->perm_change_fd = ret;
3163 s->perm_change_flags = open_flags;
3164 }
3165
3166 /* Prepare permissions on old fd to avoid conflicts between old and new,
3167 * but keep everything locked that new will need. */
3168 ret = raw_handle_perm_lock(bs, RAW_PL_PREPARE, perm, shared, errp);
3169 if (ret < 0) {
3170 goto fail;
3171 }
3172
3173 /* Copy locks to the new fd */
3174 if (s->perm_change_fd && s->use_lock) {
3175 ret = raw_apply_lock_bytes(NULL, s->perm_change_fd, perm, ~shared,
3176 false, errp);
3177 if (ret < 0) {
3178 raw_handle_perm_lock(bs, RAW_PL_ABORT, 0, 0, NULL);
3179 goto fail;
3180 }
3181 }
3182 return 0;
3183
3184 fail:
3185 if (s->perm_change_fd) {
3186 qemu_close(s->perm_change_fd);
3187 }
3188 s->perm_change_fd = 0;
3189 return ret;
3190 }
3191
3192 static void raw_set_perm(BlockDriverState *bs, uint64_t perm, uint64_t shared)
3193 {
3194 BDRVRawState *s = bs->opaque;
3195
3196 /* For reopen, we have already switched to the new fd (.bdrv_set_perm is
3197 * called after .bdrv_reopen_commit) */
3198 if (s->perm_change_fd && s->fd != s->perm_change_fd) {
3199 qemu_close(s->fd);
3200 s->fd = s->perm_change_fd;
3201 s->open_flags = s->perm_change_flags;
3202 }
3203 s->perm_change_fd = 0;
3204
3205 raw_handle_perm_lock(bs, RAW_PL_COMMIT, perm, shared, NULL);
3206 s->perm = perm;
3207 s->shared_perm = shared;
3208 }
3209
3210 static void raw_abort_perm_update(BlockDriverState *bs)
3211 {
3212 BDRVRawState *s = bs->opaque;
3213
3214 /* For reopen, .bdrv_reopen_abort is called afterwards and will close
3215 * the file descriptor. */
3216 if (s->perm_change_fd) {
3217 qemu_close(s->perm_change_fd);
3218 }
3219 s->perm_change_fd = 0;
3220
3221 raw_handle_perm_lock(bs, RAW_PL_ABORT, 0, 0, NULL);
3222 }
3223
3224 static int coroutine_fn raw_co_copy_range_from(
3225 BlockDriverState *bs, BdrvChild *src, int64_t src_offset,
3226 BdrvChild *dst, int64_t dst_offset, int64_t bytes,
3227 BdrvRequestFlags read_flags, BdrvRequestFlags write_flags)
3228 {
3229 return bdrv_co_copy_range_to(src, src_offset, dst, dst_offset, bytes,
3230 read_flags, write_flags);
3231 }
3232
3233 static int coroutine_fn raw_co_copy_range_to(BlockDriverState *bs,
3234 BdrvChild *src,
3235 int64_t src_offset,
3236 BdrvChild *dst,
3237 int64_t dst_offset,
3238 int64_t bytes,
3239 BdrvRequestFlags read_flags,
3240 BdrvRequestFlags write_flags)
3241 {
3242 RawPosixAIOData acb;
3243 BDRVRawState *s = bs->opaque;
3244 BDRVRawState *src_s;
3245
3246 assert(dst->bs == bs);
3247 if (src->bs->drv->bdrv_co_copy_range_to != raw_co_copy_range_to) {
3248 return -ENOTSUP;
3249 }
3250
3251 src_s = src->bs->opaque;
3252 if (fd_open(src->bs) < 0 || fd_open(dst->bs) < 0) {
3253 return -EIO;
3254 }
3255
3256 acb = (RawPosixAIOData) {
3257 .bs = bs,
3258 .aio_type = QEMU_AIO_COPY_RANGE,
3259 .aio_fildes = src_s->fd,
3260 .aio_offset = src_offset,
3261 .aio_nbytes = bytes,
3262 .copy_range = {
3263 .aio_fd2 = s->fd,
3264 .aio_offset2 = dst_offset,
3265 },
3266 };
3267
3268 return raw_thread_pool_submit(bs, handle_aiocb_copy_range, &acb);
3269 }
3270
3271 BlockDriver bdrv_file = {
3272 .format_name = "file",
3273 .protocol_name = "file",
3274 .instance_size = sizeof(BDRVRawState),
3275 .bdrv_needs_filename = true,
3276 .bdrv_probe = NULL, /* no probe for protocols */
3277 .bdrv_parse_filename = raw_parse_filename,
3278 .bdrv_file_open = raw_open,
3279 .bdrv_reopen_prepare = raw_reopen_prepare,
3280 .bdrv_reopen_commit = raw_reopen_commit,
3281 .bdrv_reopen_abort = raw_reopen_abort,
3282 .bdrv_close = raw_close,
3283 .bdrv_co_create = raw_co_create,
3284 .bdrv_co_create_opts = raw_co_create_opts,
3285 .bdrv_has_zero_init = bdrv_has_zero_init_1,
3286 .bdrv_co_block_status = raw_co_block_status,
3287 .bdrv_co_invalidate_cache = raw_co_invalidate_cache,
3288 .bdrv_co_pwrite_zeroes = raw_co_pwrite_zeroes,
3289 .bdrv_co_delete_file = raw_co_delete_file,
3290
3291 .bdrv_co_preadv = raw_co_preadv,
3292 .bdrv_co_pwritev = raw_co_pwritev,
3293 .bdrv_co_flush_to_disk = raw_co_flush_to_disk,
3294 .bdrv_co_pdiscard = raw_co_pdiscard,
3295 .bdrv_co_copy_range_from = raw_co_copy_range_from,
3296 .bdrv_co_copy_range_to = raw_co_copy_range_to,
3297 .bdrv_refresh_limits = raw_refresh_limits,
3298 .bdrv_io_plug = raw_aio_plug,
3299 .bdrv_io_unplug = raw_aio_unplug,
3300 .bdrv_attach_aio_context = raw_aio_attach_aio_context,
3301
3302 .bdrv_co_truncate = raw_co_truncate,
3303 .bdrv_getlength = raw_getlength,
3304 .bdrv_get_info = raw_get_info,
3305 .bdrv_get_allocated_file_size
3306 = raw_get_allocated_file_size,
3307 .bdrv_get_specific_stats = raw_get_specific_stats,
3308 .bdrv_check_perm = raw_check_perm,
3309 .bdrv_set_perm = raw_set_perm,
3310 .bdrv_abort_perm_update = raw_abort_perm_update,
3311 .create_opts = &raw_create_opts,
3312 .mutable_opts = mutable_opts,
3313 };
3314
3315 /***********************************************/
3316 /* host device */
3317
3318 #if defined(HAVE_HOST_BLOCK_DEVICE)
3319
3320 #if defined(__APPLE__) && defined(__MACH__)
3321 static kern_return_t GetBSDPath(io_iterator_t mediaIterator, char *bsdPath,
3322 CFIndex maxPathSize, int flags);
3323 static char *FindEjectableOpticalMedia(io_iterator_t *mediaIterator)
3324 {
3325 kern_return_t kernResult = KERN_FAILURE;
3326 mach_port_t masterPort;
3327 CFMutableDictionaryRef classesToMatch;
3328 const char *matching_array[] = {kIODVDMediaClass, kIOCDMediaClass};
3329 char *mediaType = NULL;
3330
3331 kernResult = IOMasterPort( MACH_PORT_NULL, &masterPort );
3332 if ( KERN_SUCCESS != kernResult ) {
3333 printf( "IOMasterPort returned %d\n", kernResult );
3334 }
3335
3336 int index;
3337 for (index = 0; index < ARRAY_SIZE(matching_array); index++) {
3338 classesToMatch = IOServiceMatching(matching_array[index]);
3339 if (classesToMatch == NULL) {
3340 error_report("IOServiceMatching returned NULL for %s",
3341 matching_array[index]);
3342 continue;
3343 }
3344 CFDictionarySetValue(classesToMatch, CFSTR(kIOMediaEjectableKey),
3345 kCFBooleanTrue);
3346 kernResult = IOServiceGetMatchingServices(masterPort, classesToMatch,
3347 mediaIterator);
3348 if (kernResult != KERN_SUCCESS) {
3349 error_report("Note: IOServiceGetMatchingServices returned %d",
3350 kernResult);
3351 continue;
3352 }
3353
3354 /* If a match was found, leave the loop */
3355 if (*mediaIterator != 0) {
3356 trace_file_FindEjectableOpticalMedia(matching_array[index]);
3357 mediaType = g_strdup(matching_array[index]);
3358 break;
3359 }
3360 }
3361 return mediaType;
3362 }
3363
3364 kern_return_t GetBSDPath(io_iterator_t mediaIterator, char *bsdPath,
3365 CFIndex maxPathSize, int flags)
3366 {
3367 io_object_t nextMedia;
3368 kern_return_t kernResult = KERN_FAILURE;
3369 *bsdPath = '\0';
3370 nextMedia = IOIteratorNext( mediaIterator );
3371 if ( nextMedia )
3372 {
3373 CFTypeRef bsdPathAsCFString;
3374 bsdPathAsCFString = IORegistryEntryCreateCFProperty( nextMedia, CFSTR( kIOBSDNameKey ), kCFAllocatorDefault, 0 );
3375 if ( bsdPathAsCFString ) {
3376 size_t devPathLength;
3377 strcpy( bsdPath, _PATH_DEV );
3378 if (flags & BDRV_O_NOCACHE) {
3379 strcat(bsdPath, "r");
3380 }
3381 devPathLength = strlen( bsdPath );
3382 if ( CFStringGetCString( bsdPathAsCFString, bsdPath + devPathLength, maxPathSize - devPathLength, kCFStringEncodingASCII ) ) {
3383 kernResult = KERN_SUCCESS;
3384 }
3385 CFRelease( bsdPathAsCFString );
3386 }
3387 IOObjectRelease( nextMedia );
3388 }
3389
3390 return kernResult;
3391 }
3392
3393 /* Sets up a real cdrom for use in QEMU */
3394 static bool setup_cdrom(char *bsd_path, Error **errp)
3395 {
3396 int index, num_of_test_partitions = 2, fd;
3397 char test_partition[MAXPATHLEN];
3398 bool partition_found = false;
3399
3400 /* look for a working partition */
3401 for (index = 0; index < num_of_test_partitions; index++) {
3402 snprintf(test_partition, sizeof(test_partition), "%ss%d", bsd_path,
3403 index);
3404 fd = qemu_open(test_partition, O_RDONLY | O_BINARY | O_LARGEFILE, NULL);
3405 if (fd >= 0) {
3406 partition_found = true;
3407 qemu_close(fd);
3408 break;
3409 }
3410 }
3411
3412 /* if a working partition on the device was not found */
3413 if (partition_found == false) {
3414 error_setg(errp, "Failed to find a working partition on disc");
3415 } else {
3416 trace_file_setup_cdrom(test_partition);
3417 pstrcpy(bsd_path, MAXPATHLEN, test_partition);
3418 }
3419 return partition_found;
3420 }
3421
3422 /* Prints directions on mounting and unmounting a device */
3423 static void print_unmounting_directions(const char *file_name)
3424 {
3425 error_report("If device %s is mounted on the desktop, unmount"
3426 " it first before using it in QEMU", file_name);
3427 error_report("Command to unmount device: diskutil unmountDisk %s",
3428 file_name);
3429 error_report("Command to mount device: diskutil mountDisk %s", file_name);
3430 }
3431
3432 #endif /* defined(__APPLE__) && defined(__MACH__) */
3433
3434 static int hdev_probe_device(const char *filename)
3435 {
3436 struct stat st;
3437
3438 /* allow a dedicated CD-ROM driver to match with a higher priority */
3439 if (strstart(filename, "/dev/cdrom", NULL))
3440 return 50;
3441
3442 if (stat(filename, &st) >= 0 &&
3443 (S_ISCHR(st.st_mode) || S_ISBLK(st.st_mode))) {
3444 return 100;
3445 }
3446
3447 return 0;
3448 }
3449
3450 static void hdev_parse_filename(const char *filename, QDict *options,
3451 Error **errp)
3452 {
3453 bdrv_parse_filename_strip_prefix(filename, "host_device:", options);
3454 }
3455
3456 static bool hdev_is_sg(BlockDriverState *bs)
3457 {
3458
3459 #if defined(__linux__)
3460
3461 BDRVRawState *s = bs->opaque;
3462 struct stat st;
3463 struct sg_scsi_id scsiid;
3464 int sg_version;
3465 int ret;
3466
3467 if (stat(bs->filename, &st) < 0 || !S_ISCHR(st.st_mode)) {
3468 return false;
3469 }
3470
3471 ret = ioctl(s->fd, SG_GET_VERSION_NUM, &sg_version);
3472 if (ret < 0) {
3473 return false;
3474 }
3475
3476 ret = ioctl(s->fd, SG_GET_SCSI_ID, &scsiid);
3477 if (ret >= 0) {
3478 trace_file_hdev_is_sg(scsiid.scsi_type, sg_version);
3479 return true;
3480 }
3481
3482 #endif
3483
3484 return false;
3485 }
3486
3487 static int hdev_open(BlockDriverState *bs, QDict *options, int flags,
3488 Error **errp)
3489 {
3490 BDRVRawState *s = bs->opaque;
3491 int ret;
3492
3493 #if defined(__APPLE__) && defined(__MACH__)
3494 /*
3495 * Caution: while qdict_get_str() is fine, getting non-string types
3496 * would require more care. When @options come from -blockdev or
3497 * blockdev_add, its members are typed according to the QAPI
3498 * schema, but when they come from -drive, they're all QString.
3499 */
3500 const char *filename = qdict_get_str(options, "filename");
3501 char bsd_path[MAXPATHLEN] = "";
3502 bool error_occurred = false;
3503
3504 /* If using a real cdrom */
3505 if (strcmp(filename, "/dev/cdrom") == 0) {
3506 char *mediaType = NULL;
3507 kern_return_t ret_val;
3508 io_iterator_t mediaIterator = 0;
3509
3510 mediaType = FindEjectableOpticalMedia(&mediaIterator);
3511 if (mediaType == NULL) {
3512 error_setg(errp, "Please make sure your CD/DVD is in the optical"
3513 " drive");
3514 error_occurred = true;
3515 goto hdev_open_Mac_error;
3516 }
3517
3518 ret_val = GetBSDPath(mediaIterator, bsd_path, sizeof(bsd_path), flags);
3519 if (ret_val != KERN_SUCCESS) {
3520 error_setg(errp, "Could not get BSD path for optical drive");
3521 error_occurred = true;
3522 goto hdev_open_Mac_error;
3523 }
3524
3525 /* If a real optical drive was not found */
3526 if (bsd_path[0] == '\0') {
3527 error_setg(errp, "Failed to obtain bsd path for optical drive");
3528 error_occurred = true;
3529 goto hdev_open_Mac_error;
3530 }
3531
3532 /* If using a cdrom disc and finding a partition on the disc failed */
3533 if (strncmp(mediaType, kIOCDMediaClass, 9) == 0 &&
3534 setup_cdrom(bsd_path, errp) == false) {
3535 print_unmounting_directions(bsd_path);
3536 error_occurred = true;
3537 goto hdev_open_Mac_error;
3538 }
3539
3540 qdict_put_str(options, "filename", bsd_path);
3541
3542 hdev_open_Mac_error:
3543 g_free(mediaType);
3544 if (mediaIterator) {
3545 IOObjectRelease(mediaIterator);
3546 }
3547 if (error_occurred) {
3548 return -ENOENT;
3549 }
3550 }
3551 #endif /* defined(__APPLE__) && defined(__MACH__) */
3552
3553 s->type = FTYPE_FILE;
3554
3555 ret = raw_open_common(bs, options, flags, 0, true, errp);
3556 if (ret < 0) {
3557 #if defined(__APPLE__) && defined(__MACH__)
3558 if (*bsd_path) {
3559 filename = bsd_path;
3560 }
3561 /* if a physical device experienced an error while being opened */
3562 if (strncmp(filename, "/dev/", 5) == 0) {
3563 print_unmounting_directions(filename);
3564 }
3565 #endif /* defined(__APPLE__) && defined(__MACH__) */
3566 return ret;
3567 }
3568
3569 /* Since this does ioctl the device must be already opened */
3570 bs->sg = hdev_is_sg(bs);
3571
3572 return ret;
3573 }
3574
3575 #if defined(__linux__)
3576 static int coroutine_fn
3577 hdev_co_ioctl(BlockDriverState *bs, unsigned long int req, void *buf)
3578 {
3579 BDRVRawState *s = bs->opaque;
3580 RawPosixAIOData acb;
3581 int ret;
3582
3583 ret = fd_open(bs);
3584 if (ret < 0) {
3585 return ret;
3586 }
3587
3588 if (req == SG_IO && s->pr_mgr) {
3589 struct sg_io_hdr *io_hdr = buf;
3590 if (io_hdr->cmdp[0] == PERSISTENT_RESERVE_OUT ||
3591 io_hdr->cmdp[0] == PERSISTENT_RESERVE_IN) {
3592 return pr_manager_execute(s->pr_mgr, bdrv_get_aio_context(bs),
3593 s->fd, io_hdr);
3594 }
3595 }
3596
3597 acb = (RawPosixAIOData) {
3598 .bs = bs,
3599 .aio_type = QEMU_AIO_IOCTL,
3600 .aio_fildes = s->fd,
3601 .aio_offset = 0,
3602 .ioctl = {
3603 .buf = buf,
3604 .cmd = req,
3605 },
3606 };
3607
3608 return raw_thread_pool_submit(bs, handle_aiocb_ioctl, &acb);
3609 }
3610 #endif /* linux */
3611
3612 static coroutine_fn int
3613 hdev_co_pdiscard(BlockDriverState *bs, int64_t offset, int64_t bytes)
3614 {
3615 BDRVRawState *s = bs->opaque;
3616 int ret;
3617
3618 ret = fd_open(bs);
3619 if (ret < 0) {
3620 raw_account_discard(s, bytes, ret);
3621 return ret;
3622 }
3623 return raw_do_pdiscard(bs, offset, bytes, true);
3624 }
3625
3626 static coroutine_fn int hdev_co_pwrite_zeroes(BlockDriverState *bs,
3627 int64_t offset, int64_t bytes, BdrvRequestFlags flags)
3628 {
3629 int rc;
3630
3631 rc = fd_open(bs);
3632 if (rc < 0) {
3633 return rc;
3634 }
3635
3636 return raw_do_pwrite_zeroes(bs, offset, bytes, flags, true);
3637 }
3638
3639 static BlockDriver bdrv_host_device = {
3640 .format_name = "host_device",
3641 .protocol_name = "host_device",
3642 .instance_size = sizeof(BDRVRawState),
3643 .bdrv_needs_filename = true,
3644 .bdrv_probe_device = hdev_probe_device,
3645 .bdrv_parse_filename = hdev_parse_filename,
3646 .bdrv_file_open = hdev_open,
3647 .bdrv_close = raw_close,
3648 .bdrv_reopen_prepare = raw_reopen_prepare,
3649 .bdrv_reopen_commit = raw_reopen_commit,
3650 .bdrv_reopen_abort = raw_reopen_abort,
3651 .bdrv_co_create_opts = bdrv_co_create_opts_simple,
3652 .create_opts = &bdrv_create_opts_simple,
3653 .mutable_opts = mutable_opts,
3654 .bdrv_co_invalidate_cache = raw_co_invalidate_cache,
3655 .bdrv_co_pwrite_zeroes = hdev_co_pwrite_zeroes,
3656
3657 .bdrv_co_preadv = raw_co_preadv,
3658 .bdrv_co_pwritev = raw_co_pwritev,
3659 .bdrv_co_flush_to_disk = raw_co_flush_to_disk,
3660 .bdrv_co_pdiscard = hdev_co_pdiscard,
3661 .bdrv_co_copy_range_from = raw_co_copy_range_from,
3662 .bdrv_co_copy_range_to = raw_co_copy_range_to,
3663 .bdrv_refresh_limits = raw_refresh_limits,
3664 .bdrv_io_plug = raw_aio_plug,
3665 .bdrv_io_unplug = raw_aio_unplug,
3666 .bdrv_attach_aio_context = raw_aio_attach_aio_context,
3667
3668 .bdrv_co_truncate = raw_co_truncate,
3669 .bdrv_getlength = raw_getlength,
3670 .bdrv_get_info = raw_get_info,
3671 .bdrv_get_allocated_file_size
3672 = raw_get_allocated_file_size,
3673 .bdrv_get_specific_stats = hdev_get_specific_stats,
3674 .bdrv_check_perm = raw_check_perm,
3675 .bdrv_set_perm = raw_set_perm,
3676 .bdrv_abort_perm_update = raw_abort_perm_update,
3677 .bdrv_probe_blocksizes = hdev_probe_blocksizes,
3678 .bdrv_probe_geometry = hdev_probe_geometry,
3679
3680 /* generic scsi device */
3681 #ifdef __linux__
3682 .bdrv_co_ioctl = hdev_co_ioctl,
3683 #endif
3684 };
3685
3686 #if defined(__linux__) || defined(__FreeBSD__) || defined(__FreeBSD_kernel__)
3687 static void cdrom_parse_filename(const char *filename, QDict *options,
3688 Error **errp)
3689 {
3690 bdrv_parse_filename_strip_prefix(filename, "host_cdrom:", options);
3691 }
3692 #endif
3693
3694 #ifdef __linux__
3695 static int cdrom_open(BlockDriverState *bs, QDict *options, int flags,
3696 Error **errp)
3697 {
3698 BDRVRawState *s = bs->opaque;
3699
3700 s->type = FTYPE_CD;
3701
3702 /* open will not fail even if no CD is inserted, so add O_NONBLOCK */
3703 return raw_open_common(bs, options, flags, O_NONBLOCK, true, errp);
3704 }
3705
3706 static int cdrom_probe_device(const char *filename)
3707 {
3708 int fd, ret;
3709 int prio = 0;
3710 struct stat st;
3711
3712 fd = qemu_open(filename, O_RDONLY | O_NONBLOCK, NULL);
3713 if (fd < 0) {
3714 goto out;
3715 }
3716 ret = fstat(fd, &st);
3717 if (ret == -1 || !S_ISBLK(st.st_mode)) {
3718 goto outc;
3719 }
3720
3721 /* Attempt to detect via a CDROM specific ioctl */
3722 ret = ioctl(fd, CDROM_DRIVE_STATUS, CDSL_CURRENT);
3723 if (ret >= 0)
3724 prio = 100;
3725
3726 outc:
3727 qemu_close(fd);
3728 out:
3729 return prio;
3730 }
3731
3732 static bool cdrom_is_inserted(BlockDriverState *bs)
3733 {
3734 BDRVRawState *s = bs->opaque;
3735 int ret;
3736
3737 ret = ioctl(s->fd, CDROM_DRIVE_STATUS, CDSL_CURRENT);
3738 return ret == CDS_DISC_OK;
3739 }
3740
3741 static void cdrom_eject(BlockDriverState *bs, bool eject_flag)
3742 {
3743 BDRVRawState *s = bs->opaque;
3744
3745 if (eject_flag) {
3746 if (ioctl(s->fd, CDROMEJECT, NULL) < 0)
3747 perror("CDROMEJECT");
3748 } else {
3749 if (ioctl(s->fd, CDROMCLOSETRAY, NULL) < 0)
3750 perror("CDROMEJECT");
3751 }
3752 }
3753
3754 static void cdrom_lock_medium(BlockDriverState *bs, bool locked)
3755 {
3756 BDRVRawState *s = bs->opaque;
3757
3758 if (ioctl(s->fd, CDROM_LOCKDOOR, locked) < 0) {
3759 /*
3760 * Note: an error can happen if the distribution automatically
3761 * mounts the CD-ROM
3762 */
3763 /* perror("CDROM_LOCKDOOR"); */
3764 }
3765 }
3766
3767 static BlockDriver bdrv_host_cdrom = {
3768 .format_name = "host_cdrom",
3769 .protocol_name = "host_cdrom",
3770 .instance_size = sizeof(BDRVRawState),
3771 .bdrv_needs_filename = true,
3772 .bdrv_probe_device = cdrom_probe_device,
3773 .bdrv_parse_filename = cdrom_parse_filename,
3774 .bdrv_file_open = cdrom_open,
3775 .bdrv_close = raw_close,
3776 .bdrv_reopen_prepare = raw_reopen_prepare,
3777 .bdrv_reopen_commit = raw_reopen_commit,
3778 .bdrv_reopen_abort = raw_reopen_abort,
3779 .bdrv_co_create_opts = bdrv_co_create_opts_simple,
3780 .create_opts = &bdrv_create_opts_simple,
3781 .mutable_opts = mutable_opts,
3782 .bdrv_co_invalidate_cache = raw_co_invalidate_cache,
3783
3784 .bdrv_co_preadv = raw_co_preadv,
3785 .bdrv_co_pwritev = raw_co_pwritev,
3786 .bdrv_co_flush_to_disk = raw_co_flush_to_disk,
3787 .bdrv_refresh_limits = raw_refresh_limits,
3788 .bdrv_io_plug = raw_aio_plug,
3789 .bdrv_io_unplug = raw_aio_unplug,
3790 .bdrv_attach_aio_context = raw_aio_attach_aio_context,
3791
3792 .bdrv_co_truncate = raw_co_truncate,
3793 .bdrv_getlength = raw_getlength,
3794 .has_variable_length = true,
3795 .bdrv_get_allocated_file_size
3796 = raw_get_allocated_file_size,
3797
3798 /* removable device support */
3799 .bdrv_is_inserted = cdrom_is_inserted,
3800 .bdrv_eject = cdrom_eject,
3801 .bdrv_lock_medium = cdrom_lock_medium,
3802
3803 /* generic scsi device */
3804 .bdrv_co_ioctl = hdev_co_ioctl,
3805 };
3806 #endif /* __linux__ */
3807
3808 #if defined (__FreeBSD__) || defined(__FreeBSD_kernel__)
3809 static int cdrom_open(BlockDriverState *bs, QDict *options, int flags,
3810 Error **errp)
3811 {
3812 BDRVRawState *s = bs->opaque;
3813 int ret;
3814
3815 s->type = FTYPE_CD;
3816
3817 ret = raw_open_common(bs, options, flags, 0, true, errp);
3818 if (ret) {
3819 return ret;
3820 }
3821
3822 /* make sure the door isn't locked at this time */
3823 ioctl(s->fd, CDIOCALLOW);
3824 return 0;
3825 }
3826
3827 static int cdrom_probe_device(const char *filename)
3828 {
3829 if (strstart(filename, "/dev/cd", NULL) ||
3830 strstart(filename, "/dev/acd", NULL))
3831 return 100;
3832 return 0;
3833 }
3834
3835 static int cdrom_reopen(BlockDriverState *bs)
3836 {
3837 BDRVRawState *s = bs->opaque;
3838 int fd;
3839
3840 /*
3841 * Force reread of possibly changed/newly loaded disc,
3842 * FreeBSD seems to not notice sometimes...
3843 */
3844 if (s->fd >= 0)
3845 qemu_close(s->fd);
3846 fd = qemu_open(bs->filename, s->open_flags, NULL);
3847 if (fd < 0) {
3848 s->fd = -1;
3849 return -EIO;
3850 }
3851 s->fd = fd;
3852
3853 /* make sure the door isn't locked at this time */
3854 ioctl(s->fd, CDIOCALLOW);
3855 return 0;
3856 }
3857
3858 static bool cdrom_is_inserted(BlockDriverState *bs)
3859 {
3860 return raw_getlength(bs) > 0;
3861 }
3862
3863 static void cdrom_eject(BlockDriverState *bs, bool eject_flag)
3864 {
3865 BDRVRawState *s = bs->opaque;
3866
3867 if (s->fd < 0)
3868 return;
3869
3870 (void) ioctl(s->fd, CDIOCALLOW);
3871
3872 if (eject_flag) {
3873 if (ioctl(s->fd, CDIOCEJECT) < 0)
3874 perror("CDIOCEJECT");
3875 } else {
3876 if (ioctl(s->fd, CDIOCCLOSE) < 0)
3877 perror("CDIOCCLOSE");
3878 }
3879
3880 cdrom_reopen(bs);
3881 }
3882
3883 static void cdrom_lock_medium(BlockDriverState *bs, bool locked)
3884 {
3885 BDRVRawState *s = bs->opaque;
3886
3887 if (s->fd < 0)
3888 return;
3889 if (ioctl(s->fd, (locked ? CDIOCPREVENT : CDIOCALLOW)) < 0) {
3890 /*
3891 * Note: an error can happen if the distribution automatically
3892 * mounts the CD-ROM
3893 */
3894 /* perror("CDROM_LOCKDOOR"); */
3895 }
3896 }
3897
3898 static BlockDriver bdrv_host_cdrom = {
3899 .format_name = "host_cdrom",
3900 .protocol_name = "host_cdrom",
3901 .instance_size = sizeof(BDRVRawState),
3902 .bdrv_needs_filename = true,
3903 .bdrv_probe_device = cdrom_probe_device,
3904 .bdrv_parse_filename = cdrom_parse_filename,
3905 .bdrv_file_open = cdrom_open,
3906 .bdrv_close = raw_close,
3907 .bdrv_reopen_prepare = raw_reopen_prepare,
3908 .bdrv_reopen_commit = raw_reopen_commit,
3909 .bdrv_reopen_abort = raw_reopen_abort,
3910 .bdrv_co_create_opts = bdrv_co_create_opts_simple,
3911 .create_opts = &bdrv_create_opts_simple,
3912 .mutable_opts = mutable_opts,
3913
3914 .bdrv_co_preadv = raw_co_preadv,
3915 .bdrv_co_pwritev = raw_co_pwritev,
3916 .bdrv_co_flush_to_disk = raw_co_flush_to_disk,
3917 .bdrv_refresh_limits = raw_refresh_limits,
3918 .bdrv_io_plug = raw_aio_plug,
3919 .bdrv_io_unplug = raw_aio_unplug,
3920 .bdrv_attach_aio_context = raw_aio_attach_aio_context,
3921
3922 .bdrv_co_truncate = raw_co_truncate,
3923 .bdrv_getlength = raw_getlength,
3924 .has_variable_length = true,
3925 .bdrv_get_allocated_file_size
3926 = raw_get_allocated_file_size,
3927
3928 /* removable device support */
3929 .bdrv_is_inserted = cdrom_is_inserted,
3930 .bdrv_eject = cdrom_eject,
3931 .bdrv_lock_medium = cdrom_lock_medium,
3932 };
3933 #endif /* __FreeBSD__ */
3934
3935 #endif /* HAVE_HOST_BLOCK_DEVICE */
3936
3937 static void bdrv_file_init(void)
3938 {
3939 /*
3940 * Register all the drivers. Note that order is important, the driver
3941 * registered last will get probed first.
3942 */
3943 bdrv_register(&bdrv_file);
3944 #if defined(HAVE_HOST_BLOCK_DEVICE)
3945 bdrv_register(&bdrv_host_device);
3946 #ifdef __linux__
3947 bdrv_register(&bdrv_host_cdrom);
3948 #endif
3949 #if defined(__FreeBSD__) || defined(__FreeBSD_kernel__)
3950 bdrv_register(&bdrv_host_cdrom);
3951 #endif
3952 #endif /* HAVE_HOST_BLOCK_DEVICE */
3953 }
3954
3955 block_init(bdrv_file_init);