]> git.proxmox.com Git - qemu.git/blame - block.c
block: make bdrv_set_enable_write_cache() modify open_flags
[qemu.git] / block.c
CommitLineData
fc01f7e7
FB
1/*
2 * QEMU System Emulator block driver
5fafdf24 3 *
fc01f7e7 4 * Copyright (c) 2003 Fabrice Bellard
5fafdf24 5 *
fc01f7e7
FB
6 * Permission is hereby granted, free of charge, to any person obtaining a copy
7 * of this software and associated documentation files (the "Software"), to deal
8 * in the Software without restriction, including without limitation the rights
9 * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
10 * copies of the Software, and to permit persons to whom the Software is
11 * furnished to do so, subject to the following conditions:
12 *
13 * The above copyright notice and this permission notice shall be included in
14 * all copies or substantial portions of the Software.
15 *
16 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
17 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
18 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
19 * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
20 * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
21 * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
22 * THE SOFTWARE.
23 */
3990d09a 24#include "config-host.h"
faf07963 25#include "qemu-common.h"
6d519a5f 26#include "trace.h"
376253ec 27#include "monitor.h"
ea2384d3 28#include "block_int.h"
5efa9d5a 29#include "module.h"
f795e743 30#include "qjson.h"
68485420 31#include "qemu-coroutine.h"
b2023818 32#include "qmp-commands.h"
0563e191 33#include "qemu-timer.h"
fc01f7e7 34
71e72a19 35#ifdef CONFIG_BSD
7674e7bf
FB
36#include <sys/types.h>
37#include <sys/stat.h>
38#include <sys/ioctl.h>
72cf2d4f 39#include <sys/queue.h>
c5e97233 40#ifndef __DragonFly__
7674e7bf
FB
41#include <sys/disk.h>
42#endif
c5e97233 43#endif
7674e7bf 44
49dc768d
AL
45#ifdef _WIN32
46#include <windows.h>
47#endif
48
1c9805a3
SH
49#define NOT_DONE 0x7fffffff /* used while emulated sync operation in progress */
50
470c0504
SH
51typedef enum {
52 BDRV_REQ_COPY_ON_READ = 0x1,
f08f2dda 53 BDRV_REQ_ZERO_WRITE = 0x2,
470c0504
SH
54} BdrvRequestFlags;
55
7d4b4ba5 56static void bdrv_dev_change_media_cb(BlockDriverState *bs, bool load);
f141eafe
AL
57static BlockDriverAIOCB *bdrv_aio_readv_em(BlockDriverState *bs,
58 int64_t sector_num, QEMUIOVector *qiov, int nb_sectors,
c87c0672 59 BlockDriverCompletionFunc *cb, void *opaque);
f141eafe
AL
60static BlockDriverAIOCB *bdrv_aio_writev_em(BlockDriverState *bs,
61 int64_t sector_num, QEMUIOVector *qiov, int nb_sectors,
ce1a14dc 62 BlockDriverCompletionFunc *cb, void *opaque);
f9f05dc5
KW
63static int coroutine_fn bdrv_co_readv_em(BlockDriverState *bs,
64 int64_t sector_num, int nb_sectors,
65 QEMUIOVector *iov);
66static int coroutine_fn bdrv_co_writev_em(BlockDriverState *bs,
67 int64_t sector_num, int nb_sectors,
68 QEMUIOVector *iov);
c5fbe571 69static int coroutine_fn bdrv_co_do_readv(BlockDriverState *bs,
470c0504
SH
70 int64_t sector_num, int nb_sectors, QEMUIOVector *qiov,
71 BdrvRequestFlags flags);
1c9805a3 72static int coroutine_fn bdrv_co_do_writev(BlockDriverState *bs,
f08f2dda
SH
73 int64_t sector_num, int nb_sectors, QEMUIOVector *qiov,
74 BdrvRequestFlags flags);
b2a61371
SH
75static BlockDriverAIOCB *bdrv_co_aio_rw_vector(BlockDriverState *bs,
76 int64_t sector_num,
77 QEMUIOVector *qiov,
78 int nb_sectors,
79 BlockDriverCompletionFunc *cb,
80 void *opaque,
8c5873d6 81 bool is_write);
b2a61371 82static void coroutine_fn bdrv_co_do_rw(void *opaque);
621f0589
KW
83static int coroutine_fn bdrv_co_do_write_zeroes(BlockDriverState *bs,
84 int64_t sector_num, int nb_sectors);
ec530c81 85
98f90dba
ZYW
86static bool bdrv_exceed_bps_limits(BlockDriverState *bs, int nb_sectors,
87 bool is_write, double elapsed_time, uint64_t *wait);
88static bool bdrv_exceed_iops_limits(BlockDriverState *bs, bool is_write,
89 double elapsed_time, uint64_t *wait);
90static bool bdrv_exceed_io_limits(BlockDriverState *bs, int nb_sectors,
91 bool is_write, int64_t *wait);
92
1b7bdbc1
SH
93static QTAILQ_HEAD(, BlockDriverState) bdrv_states =
94 QTAILQ_HEAD_INITIALIZER(bdrv_states);
7ee930d0 95
8a22f02a
SH
96static QLIST_HEAD(, BlockDriver) bdrv_drivers =
97 QLIST_HEAD_INITIALIZER(bdrv_drivers);
ea2384d3 98
f9092b10
MA
99/* The device to use for VM snapshots */
100static BlockDriverState *bs_snapshots;
101
eb852011
MA
102/* If non-zero, use only whitelisted block drivers */
103static int use_bdrv_whitelist;
104
9e0b22f4
SH
105#ifdef _WIN32
106static int is_windows_drive_prefix(const char *filename)
107{
108 return (((filename[0] >= 'a' && filename[0] <= 'z') ||
109 (filename[0] >= 'A' && filename[0] <= 'Z')) &&
110 filename[1] == ':');
111}
112
113int is_windows_drive(const char *filename)
114{
115 if (is_windows_drive_prefix(filename) &&
116 filename[2] == '\0')
117 return 1;
118 if (strstart(filename, "\\\\.\\", NULL) ||
119 strstart(filename, "//./", NULL))
120 return 1;
121 return 0;
122}
123#endif
124
0563e191 125/* throttling disk I/O limits */
98f90dba
ZYW
126void bdrv_io_limits_disable(BlockDriverState *bs)
127{
128 bs->io_limits_enabled = false;
129
130 while (qemu_co_queue_next(&bs->throttled_reqs));
131
132 if (bs->block_timer) {
133 qemu_del_timer(bs->block_timer);
134 qemu_free_timer(bs->block_timer);
135 bs->block_timer = NULL;
136 }
137
138 bs->slice_start = 0;
139 bs->slice_end = 0;
140 bs->slice_time = 0;
141 memset(&bs->io_base, 0, sizeof(bs->io_base));
142}
143
0563e191
ZYW
144static void bdrv_block_timer(void *opaque)
145{
146 BlockDriverState *bs = opaque;
147
148 qemu_co_queue_next(&bs->throttled_reqs);
149}
150
151void bdrv_io_limits_enable(BlockDriverState *bs)
152{
153 qemu_co_queue_init(&bs->throttled_reqs);
154 bs->block_timer = qemu_new_timer_ns(vm_clock, bdrv_block_timer, bs);
155 bs->slice_time = 5 * BLOCK_IO_SLICE_TIME;
156 bs->slice_start = qemu_get_clock_ns(vm_clock);
157 bs->slice_end = bs->slice_start + bs->slice_time;
158 memset(&bs->io_base, 0, sizeof(bs->io_base));
159 bs->io_limits_enabled = true;
160}
161
162bool bdrv_io_limits_enabled(BlockDriverState *bs)
163{
164 BlockIOLimit *io_limits = &bs->io_limits;
165 return io_limits->bps[BLOCK_IO_LIMIT_READ]
166 || io_limits->bps[BLOCK_IO_LIMIT_WRITE]
167 || io_limits->bps[BLOCK_IO_LIMIT_TOTAL]
168 || io_limits->iops[BLOCK_IO_LIMIT_READ]
169 || io_limits->iops[BLOCK_IO_LIMIT_WRITE]
170 || io_limits->iops[BLOCK_IO_LIMIT_TOTAL];
171}
172
98f90dba
ZYW
173static void bdrv_io_limits_intercept(BlockDriverState *bs,
174 bool is_write, int nb_sectors)
175{
176 int64_t wait_time = -1;
177
178 if (!qemu_co_queue_empty(&bs->throttled_reqs)) {
179 qemu_co_queue_wait(&bs->throttled_reqs);
180 }
181
182 /* In fact, we hope to keep each request's timing, in FIFO mode. The next
183 * throttled requests will not be dequeued until the current request is
184 * allowed to be serviced. So if the current request still exceeds the
185 * limits, it will be inserted to the head. All requests followed it will
186 * be still in throttled_reqs queue.
187 */
188
189 while (bdrv_exceed_io_limits(bs, nb_sectors, is_write, &wait_time)) {
190 qemu_mod_timer(bs->block_timer,
191 wait_time + qemu_get_clock_ns(vm_clock));
192 qemu_co_queue_wait_insert_head(&bs->throttled_reqs);
193 }
194
195 qemu_co_queue_next(&bs->throttled_reqs);
196}
197
9e0b22f4
SH
198/* check if the path starts with "<protocol>:" */
199static int path_has_protocol(const char *path)
200{
947995c0
PB
201 const char *p;
202
9e0b22f4
SH
203#ifdef _WIN32
204 if (is_windows_drive(path) ||
205 is_windows_drive_prefix(path)) {
206 return 0;
207 }
947995c0
PB
208 p = path + strcspn(path, ":/\\");
209#else
210 p = path + strcspn(path, ":/");
9e0b22f4
SH
211#endif
212
947995c0 213 return *p == ':';
9e0b22f4
SH
214}
215
83f64091 216int path_is_absolute(const char *path)
3b0d4f61 217{
21664424
FB
218#ifdef _WIN32
219 /* specific case for names like: "\\.\d:" */
f53f4da9 220 if (is_windows_drive(path) || is_windows_drive_prefix(path)) {
21664424 221 return 1;
f53f4da9
PB
222 }
223 return (*path == '/' || *path == '\\');
3b9f94e1 224#else
f53f4da9 225 return (*path == '/');
3b9f94e1 226#endif
3b0d4f61
FB
227}
228
83f64091
FB
229/* if filename is absolute, just copy it to dest. Otherwise, build a
230 path to it by considering it is relative to base_path. URL are
231 supported. */
232void path_combine(char *dest, int dest_size,
233 const char *base_path,
234 const char *filename)
3b0d4f61 235{
83f64091
FB
236 const char *p, *p1;
237 int len;
238
239 if (dest_size <= 0)
240 return;
241 if (path_is_absolute(filename)) {
242 pstrcpy(dest, dest_size, filename);
243 } else {
244 p = strchr(base_path, ':');
245 if (p)
246 p++;
247 else
248 p = base_path;
3b9f94e1
FB
249 p1 = strrchr(base_path, '/');
250#ifdef _WIN32
251 {
252 const char *p2;
253 p2 = strrchr(base_path, '\\');
254 if (!p1 || p2 > p1)
255 p1 = p2;
256 }
257#endif
83f64091
FB
258 if (p1)
259 p1++;
260 else
261 p1 = base_path;
262 if (p1 > p)
263 p = p1;
264 len = p - base_path;
265 if (len > dest_size - 1)
266 len = dest_size - 1;
267 memcpy(dest, base_path, len);
268 dest[len] = '\0';
269 pstrcat(dest, dest_size, filename);
3b0d4f61 270 }
3b0d4f61
FB
271}
272
dc5a1371
PB
273void bdrv_get_full_backing_filename(BlockDriverState *bs, char *dest, size_t sz)
274{
275 if (bs->backing_file[0] == '\0' || path_has_protocol(bs->backing_file)) {
276 pstrcpy(dest, sz, bs->backing_file);
277 } else {
278 path_combine(dest, sz, bs->filename, bs->backing_file);
279 }
280}
281
5efa9d5a 282void bdrv_register(BlockDriver *bdrv)
ea2384d3 283{
8c5873d6
SH
284 /* Block drivers without coroutine functions need emulation */
285 if (!bdrv->bdrv_co_readv) {
f9f05dc5
KW
286 bdrv->bdrv_co_readv = bdrv_co_readv_em;
287 bdrv->bdrv_co_writev = bdrv_co_writev_em;
288
f8c35c1d
SH
289 /* bdrv_co_readv_em()/brdv_co_writev_em() work in terms of aio, so if
290 * the block driver lacks aio we need to emulate that too.
291 */
f9f05dc5
KW
292 if (!bdrv->bdrv_aio_readv) {
293 /* add AIO emulation layer */
294 bdrv->bdrv_aio_readv = bdrv_aio_readv_em;
295 bdrv->bdrv_aio_writev = bdrv_aio_writev_em;
f9f05dc5 296 }
83f64091 297 }
b2e12bc6 298
8a22f02a 299 QLIST_INSERT_HEAD(&bdrv_drivers, bdrv, list);
ea2384d3 300}
b338082b
FB
301
302/* create a new block device (by default it is empty) */
303BlockDriverState *bdrv_new(const char *device_name)
304{
1b7bdbc1 305 BlockDriverState *bs;
b338082b 306
7267c094 307 bs = g_malloc0(sizeof(BlockDriverState));
b338082b 308 pstrcpy(bs->device_name, sizeof(bs->device_name), device_name);
ea2384d3 309 if (device_name[0] != '\0') {
1b7bdbc1 310 QTAILQ_INSERT_TAIL(&bdrv_states, bs, list);
ea2384d3 311 }
28a7282a 312 bdrv_iostatus_disable(bs);
b338082b
FB
313 return bs;
314}
315
ea2384d3
FB
316BlockDriver *bdrv_find_format(const char *format_name)
317{
318 BlockDriver *drv1;
8a22f02a
SH
319 QLIST_FOREACH(drv1, &bdrv_drivers, list) {
320 if (!strcmp(drv1->format_name, format_name)) {
ea2384d3 321 return drv1;
8a22f02a 322 }
ea2384d3
FB
323 }
324 return NULL;
325}
326
eb852011
MA
327static int bdrv_is_whitelisted(BlockDriver *drv)
328{
329 static const char *whitelist[] = {
330 CONFIG_BDRV_WHITELIST
331 };
332 const char **p;
333
334 if (!whitelist[0])
335 return 1; /* no whitelist, anything goes */
336
337 for (p = whitelist; *p; p++) {
338 if (!strcmp(drv->format_name, *p)) {
339 return 1;
340 }
341 }
342 return 0;
343}
344
345BlockDriver *bdrv_find_whitelisted_format(const char *format_name)
346{
347 BlockDriver *drv = bdrv_find_format(format_name);
348 return drv && bdrv_is_whitelisted(drv) ? drv : NULL;
349}
350
5b7e1542
ZYW
351typedef struct CreateCo {
352 BlockDriver *drv;
353 char *filename;
354 QEMUOptionParameter *options;
355 int ret;
356} CreateCo;
357
358static void coroutine_fn bdrv_create_co_entry(void *opaque)
359{
360 CreateCo *cco = opaque;
361 assert(cco->drv);
362
363 cco->ret = cco->drv->bdrv_create(cco->filename, cco->options);
364}
365
0e7e1989
KW
366int bdrv_create(BlockDriver *drv, const char* filename,
367 QEMUOptionParameter *options)
ea2384d3 368{
5b7e1542
ZYW
369 int ret;
370
371 Coroutine *co;
372 CreateCo cco = {
373 .drv = drv,
374 .filename = g_strdup(filename),
375 .options = options,
376 .ret = NOT_DONE,
377 };
378
379 if (!drv->bdrv_create) {
ea2384d3 380 return -ENOTSUP;
5b7e1542
ZYW
381 }
382
383 if (qemu_in_coroutine()) {
384 /* Fast-path if already in coroutine context */
385 bdrv_create_co_entry(&cco);
386 } else {
387 co = qemu_coroutine_create(bdrv_create_co_entry);
388 qemu_coroutine_enter(co, &cco);
389 while (cco.ret == NOT_DONE) {
390 qemu_aio_wait();
391 }
392 }
393
394 ret = cco.ret;
395 g_free(cco.filename);
0e7e1989 396
5b7e1542 397 return ret;
ea2384d3
FB
398}
399
84a12e66
CH
400int bdrv_create_file(const char* filename, QEMUOptionParameter *options)
401{
402 BlockDriver *drv;
403
b50cbabc 404 drv = bdrv_find_protocol(filename);
84a12e66 405 if (drv == NULL) {
16905d71 406 return -ENOENT;
84a12e66
CH
407 }
408
409 return bdrv_create(drv, filename, options);
410}
411
eba25057
JM
412/*
413 * Create a uniquely-named empty temporary file.
414 * Return 0 upon success, otherwise a negative errno value.
415 */
416int get_tmp_filename(char *filename, int size)
d5249393 417{
eba25057 418#ifdef _WIN32
3b9f94e1 419 char temp_dir[MAX_PATH];
eba25057
JM
420 /* GetTempFileName requires that its output buffer (4th param)
421 have length MAX_PATH or greater. */
422 assert(size >= MAX_PATH);
423 return (GetTempPath(MAX_PATH, temp_dir)
424 && GetTempFileName(temp_dir, "qem", 0, filename)
425 ? 0 : -GetLastError());
d5249393 426#else
67b915a5 427 int fd;
7ccfb2eb 428 const char *tmpdir;
0badc1ee
AJ
429 tmpdir = getenv("TMPDIR");
430 if (!tmpdir)
431 tmpdir = "/tmp";
eba25057
JM
432 if (snprintf(filename, size, "%s/vl.XXXXXX", tmpdir) >= size) {
433 return -EOVERFLOW;
434 }
ea2384d3 435 fd = mkstemp(filename);
fe235a06
DH
436 if (fd < 0) {
437 return -errno;
438 }
439 if (close(fd) != 0) {
440 unlink(filename);
eba25057
JM
441 return -errno;
442 }
443 return 0;
d5249393 444#endif
eba25057 445}
fc01f7e7 446
84a12e66
CH
447/*
448 * Detect host devices. By convention, /dev/cdrom[N] is always
449 * recognized as a host CDROM.
450 */
451static BlockDriver *find_hdev_driver(const char *filename)
452{
453 int score_max = 0, score;
454 BlockDriver *drv = NULL, *d;
455
456 QLIST_FOREACH(d, &bdrv_drivers, list) {
457 if (d->bdrv_probe_device) {
458 score = d->bdrv_probe_device(filename);
459 if (score > score_max) {
460 score_max = score;
461 drv = d;
462 }
463 }
464 }
465
466 return drv;
467}
468
b50cbabc 469BlockDriver *bdrv_find_protocol(const char *filename)
83f64091
FB
470{
471 BlockDriver *drv1;
472 char protocol[128];
1cec71e3 473 int len;
83f64091 474 const char *p;
19cb3738 475
66f82cee
KW
476 /* TODO Drivers without bdrv_file_open must be specified explicitly */
477
39508e7a
CH
478 /*
479 * XXX(hch): we really should not let host device detection
480 * override an explicit protocol specification, but moving this
481 * later breaks access to device names with colons in them.
482 * Thanks to the brain-dead persistent naming schemes on udev-
483 * based Linux systems those actually are quite common.
484 */
485 drv1 = find_hdev_driver(filename);
486 if (drv1) {
487 return drv1;
488 }
489
9e0b22f4 490 if (!path_has_protocol(filename)) {
39508e7a 491 return bdrv_find_format("file");
84a12e66 492 }
9e0b22f4
SH
493 p = strchr(filename, ':');
494 assert(p != NULL);
1cec71e3
AL
495 len = p - filename;
496 if (len > sizeof(protocol) - 1)
497 len = sizeof(protocol) - 1;
498 memcpy(protocol, filename, len);
499 protocol[len] = '\0';
8a22f02a 500 QLIST_FOREACH(drv1, &bdrv_drivers, list) {
5fafdf24 501 if (drv1->protocol_name &&
8a22f02a 502 !strcmp(drv1->protocol_name, protocol)) {
83f64091 503 return drv1;
8a22f02a 504 }
83f64091
FB
505 }
506 return NULL;
507}
508
c98ac35d 509static int find_image_format(const char *filename, BlockDriver **pdrv)
f3a5d3f8
CH
510{
511 int ret, score, score_max;
512 BlockDriver *drv1, *drv;
513 uint8_t buf[2048];
514 BlockDriverState *bs;
515
f5edb014 516 ret = bdrv_file_open(&bs, filename, 0);
c98ac35d
SW
517 if (ret < 0) {
518 *pdrv = NULL;
519 return ret;
520 }
f8ea0b00 521
08a00559
KW
522 /* Return the raw BlockDriver * to scsi-generic devices or empty drives */
523 if (bs->sg || !bdrv_is_inserted(bs)) {
1a396859 524 bdrv_delete(bs);
c98ac35d
SW
525 drv = bdrv_find_format("raw");
526 if (!drv) {
527 ret = -ENOENT;
528 }
529 *pdrv = drv;
530 return ret;
1a396859 531 }
f8ea0b00 532
83f64091
FB
533 ret = bdrv_pread(bs, 0, buf, sizeof(buf));
534 bdrv_delete(bs);
535 if (ret < 0) {
c98ac35d
SW
536 *pdrv = NULL;
537 return ret;
83f64091
FB
538 }
539
ea2384d3 540 score_max = 0;
84a12e66 541 drv = NULL;
8a22f02a 542 QLIST_FOREACH(drv1, &bdrv_drivers, list) {
83f64091
FB
543 if (drv1->bdrv_probe) {
544 score = drv1->bdrv_probe(buf, ret, filename);
545 if (score > score_max) {
546 score_max = score;
547 drv = drv1;
548 }
0849bf08 549 }
fc01f7e7 550 }
c98ac35d
SW
551 if (!drv) {
552 ret = -ENOENT;
553 }
554 *pdrv = drv;
555 return ret;
ea2384d3
FB
556}
557
51762288
SH
558/**
559 * Set the current 'total_sectors' value
560 */
561static int refresh_total_sectors(BlockDriverState *bs, int64_t hint)
562{
563 BlockDriver *drv = bs->drv;
564
396759ad
NB
565 /* Do not attempt drv->bdrv_getlength() on scsi-generic devices */
566 if (bs->sg)
567 return 0;
568
51762288
SH
569 /* query actual device if possible, otherwise just trust the hint */
570 if (drv->bdrv_getlength) {
571 int64_t length = drv->bdrv_getlength(bs);
572 if (length < 0) {
573 return length;
574 }
575 hint = length >> BDRV_SECTOR_BITS;
576 }
577
578 bs->total_sectors = hint;
579 return 0;
580}
581
c3993cdc
SH
582/**
583 * Set open flags for a given cache mode
584 *
585 * Return 0 on success, -1 if the cache mode was invalid.
586 */
587int bdrv_parse_cache_flags(const char *mode, int *flags)
588{
589 *flags &= ~BDRV_O_CACHE_MASK;
590
591 if (!strcmp(mode, "off") || !strcmp(mode, "none")) {
592 *flags |= BDRV_O_NOCACHE | BDRV_O_CACHE_WB;
92196b2f
SH
593 } else if (!strcmp(mode, "directsync")) {
594 *flags |= BDRV_O_NOCACHE;
c3993cdc
SH
595 } else if (!strcmp(mode, "writeback")) {
596 *flags |= BDRV_O_CACHE_WB;
597 } else if (!strcmp(mode, "unsafe")) {
598 *flags |= BDRV_O_CACHE_WB;
599 *flags |= BDRV_O_NO_FLUSH;
600 } else if (!strcmp(mode, "writethrough")) {
601 /* this is the default */
602 } else {
603 return -1;
604 }
605
606 return 0;
607}
608
53fec9d3
SH
609/**
610 * The copy-on-read flag is actually a reference count so multiple users may
611 * use the feature without worrying about clobbering its previous state.
612 * Copy-on-read stays enabled until all users have called to disable it.
613 */
614void bdrv_enable_copy_on_read(BlockDriverState *bs)
615{
616 bs->copy_on_read++;
617}
618
619void bdrv_disable_copy_on_read(BlockDriverState *bs)
620{
621 assert(bs->copy_on_read > 0);
622 bs->copy_on_read--;
623}
624
57915332
KW
625/*
626 * Common part for opening disk images and files
627 */
628static int bdrv_open_common(BlockDriverState *bs, const char *filename,
629 int flags, BlockDriver *drv)
630{
631 int ret, open_flags;
632
633 assert(drv != NULL);
6405875c 634 assert(bs->file == NULL);
57915332 635
28dcee10
SH
636 trace_bdrv_open_common(bs, filename, flags, drv->format_name);
637
57915332 638 bs->open_flags = flags;
57915332
KW
639 bs->buffer_alignment = 512;
640
53fec9d3
SH
641 assert(bs->copy_on_read == 0); /* bdrv_new() and bdrv_close() make it so */
642 if ((flags & BDRV_O_RDWR) && (flags & BDRV_O_COPY_ON_READ)) {
643 bdrv_enable_copy_on_read(bs);
644 }
645
57915332
KW
646 pstrcpy(bs->filename, sizeof(bs->filename), filename);
647
648 if (use_bdrv_whitelist && !bdrv_is_whitelisted(drv)) {
649 return -ENOTSUP;
650 }
651
652 bs->drv = drv;
7267c094 653 bs->opaque = g_malloc0(drv->instance_size);
57915332 654
03f541bd 655 bs->enable_write_cache = !!(flags & BDRV_O_CACHE_WB);
e1e9b0ac 656 open_flags = flags | BDRV_O_CACHE_WB;
57915332
KW
657
658 /*
659 * Clear flags that are internal to the block layer before opening the
660 * image.
661 */
e1e9b0ac 662 open_flags &= ~(BDRV_O_SNAPSHOT | BDRV_O_NO_BACKING);
57915332
KW
663
664 /*
ebabb67a 665 * Snapshots should be writable.
57915332
KW
666 */
667 if (bs->is_temporary) {
668 open_flags |= BDRV_O_RDWR;
669 }
670
be028adc 671 bs->read_only = !(open_flags & BDRV_O_RDWR);
e7c63796 672
66f82cee
KW
673 /* Open the image, either directly or using a protocol */
674 if (drv->bdrv_file_open) {
675 ret = drv->bdrv_file_open(bs, filename, open_flags);
676 } else {
677 ret = bdrv_file_open(&bs->file, filename, open_flags);
678 if (ret >= 0) {
679 ret = drv->bdrv_open(bs, open_flags);
680 }
681 }
682
57915332
KW
683 if (ret < 0) {
684 goto free_and_fail;
685 }
686
51762288
SH
687 ret = refresh_total_sectors(bs, bs->total_sectors);
688 if (ret < 0) {
689 goto free_and_fail;
57915332 690 }
51762288 691
57915332
KW
692#ifndef _WIN32
693 if (bs->is_temporary) {
694 unlink(filename);
695 }
696#endif
697 return 0;
698
699free_and_fail:
66f82cee
KW
700 if (bs->file) {
701 bdrv_delete(bs->file);
702 bs->file = NULL;
703 }
7267c094 704 g_free(bs->opaque);
57915332
KW
705 bs->opaque = NULL;
706 bs->drv = NULL;
707 return ret;
708}
709
b6ce07aa
KW
710/*
711 * Opens a file using a protocol (file, host_device, nbd, ...)
712 */
83f64091 713int bdrv_file_open(BlockDriverState **pbs, const char *filename, int flags)
ea2384d3 714{
83f64091 715 BlockDriverState *bs;
6db95603 716 BlockDriver *drv;
83f64091
FB
717 int ret;
718
b50cbabc 719 drv = bdrv_find_protocol(filename);
6db95603
CH
720 if (!drv) {
721 return -ENOENT;
722 }
723
83f64091 724 bs = bdrv_new("");
b6ce07aa 725 ret = bdrv_open_common(bs, filename, flags, drv);
83f64091
FB
726 if (ret < 0) {
727 bdrv_delete(bs);
728 return ret;
3b0d4f61 729 }
71d0770c 730 bs->growable = 1;
83f64091
FB
731 *pbs = bs;
732 return 0;
733}
734
b6ce07aa
KW
735/*
736 * Opens a disk image (raw, qcow2, vmdk, ...)
737 */
d6e9098e
KW
738int bdrv_open(BlockDriverState *bs, const char *filename, int flags,
739 BlockDriver *drv)
ea2384d3 740{
b6ce07aa 741 int ret;
2b572816 742 char tmp_filename[PATH_MAX];
712e7874 743
83f64091 744 if (flags & BDRV_O_SNAPSHOT) {
ea2384d3
FB
745 BlockDriverState *bs1;
746 int64_t total_size;
7c96d46e 747 int is_protocol = 0;
91a073a9
KW
748 BlockDriver *bdrv_qcow2;
749 QEMUOptionParameter *options;
b6ce07aa 750 char backing_filename[PATH_MAX];
3b46e624 751
ea2384d3
FB
752 /* if snapshot, we create a temporary backing file and open it
753 instead of opening 'filename' directly */
33e3963e 754
ea2384d3
FB
755 /* if there is a backing file, use it */
756 bs1 = bdrv_new("");
d6e9098e 757 ret = bdrv_open(bs1, filename, 0, drv);
51d7c00c 758 if (ret < 0) {
ea2384d3 759 bdrv_delete(bs1);
51d7c00c 760 return ret;
ea2384d3 761 }
3e82990b 762 total_size = bdrv_getlength(bs1) & BDRV_SECTOR_MASK;
7c96d46e
AL
763
764 if (bs1->drv && bs1->drv->protocol_name)
765 is_protocol = 1;
766
ea2384d3 767 bdrv_delete(bs1);
3b46e624 768
eba25057
JM
769 ret = get_tmp_filename(tmp_filename, sizeof(tmp_filename));
770 if (ret < 0) {
771 return ret;
772 }
7c96d46e
AL
773
774 /* Real path is meaningless for protocols */
775 if (is_protocol)
776 snprintf(backing_filename, sizeof(backing_filename),
777 "%s", filename);
114cdfa9
KS
778 else if (!realpath(filename, backing_filename))
779 return -errno;
7c96d46e 780
91a073a9
KW
781 bdrv_qcow2 = bdrv_find_format("qcow2");
782 options = parse_option_parameters("", bdrv_qcow2->create_options, NULL);
783
3e82990b 784 set_option_parameter_int(options, BLOCK_OPT_SIZE, total_size);
91a073a9
KW
785 set_option_parameter(options, BLOCK_OPT_BACKING_FILE, backing_filename);
786 if (drv) {
787 set_option_parameter(options, BLOCK_OPT_BACKING_FMT,
788 drv->format_name);
789 }
790
791 ret = bdrv_create(bdrv_qcow2, tmp_filename, options);
d748768c 792 free_option_parameters(options);
51d7c00c
AL
793 if (ret < 0) {
794 return ret;
ea2384d3 795 }
91a073a9 796
ea2384d3 797 filename = tmp_filename;
91a073a9 798 drv = bdrv_qcow2;
ea2384d3
FB
799 bs->is_temporary = 1;
800 }
712e7874 801
b6ce07aa 802 /* Find the right image format driver */
6db95603 803 if (!drv) {
c98ac35d 804 ret = find_image_format(filename, &drv);
51d7c00c 805 }
6987307c 806
51d7c00c 807 if (!drv) {
51d7c00c 808 goto unlink_and_fail;
ea2384d3 809 }
b6ce07aa 810
be028adc
JC
811 if (flags & BDRV_O_RDWR) {
812 flags |= BDRV_O_ALLOW_RDWR;
813 }
814
815 bs->keep_read_only = !(flags & BDRV_O_ALLOW_RDWR);
816
b6ce07aa
KW
817 /* Open the image */
818 ret = bdrv_open_common(bs, filename, flags, drv);
819 if (ret < 0) {
6987307c
CH
820 goto unlink_and_fail;
821 }
822
b6ce07aa
KW
823 /* If there is a backing file, use it */
824 if ((flags & BDRV_O_NO_BACKING) == 0 && bs->backing_file[0] != '\0') {
825 char backing_filename[PATH_MAX];
826 int back_flags;
827 BlockDriver *back_drv = NULL;
828
829 bs->backing_hd = bdrv_new("");
dc5a1371
PB
830 bdrv_get_full_backing_filename(bs, backing_filename,
831 sizeof(backing_filename));
df2dbb4a
SH
832
833 if (bs->backing_format[0] != '\0') {
b6ce07aa 834 back_drv = bdrv_find_format(bs->backing_format);
df2dbb4a 835 }
b6ce07aa
KW
836
837 /* backing files always opened read-only */
838 back_flags =
839 flags & ~(BDRV_O_RDWR | BDRV_O_SNAPSHOT | BDRV_O_NO_BACKING);
840
841 ret = bdrv_open(bs->backing_hd, backing_filename, back_flags, back_drv);
842 if (ret < 0) {
843 bdrv_close(bs);
844 return ret;
845 }
b6ce07aa
KW
846 }
847
848 if (!bdrv_key_required(bs)) {
7d4b4ba5 849 bdrv_dev_change_media_cb(bs, true);
b6ce07aa
KW
850 }
851
98f90dba
ZYW
852 /* throttling disk I/O limits */
853 if (bs->io_limits_enabled) {
854 bdrv_io_limits_enable(bs);
855 }
856
b6ce07aa
KW
857 return 0;
858
859unlink_and_fail:
860 if (bs->is_temporary) {
861 unlink(filename);
862 }
863 return ret;
864}
865
fc01f7e7
FB
866void bdrv_close(BlockDriverState *bs)
867{
80ccf93b 868 bdrv_flush(bs);
19cb3738 869 if (bs->drv) {
3e914655
PB
870 if (bs->job) {
871 block_job_cancel_sync(bs->job);
872 }
7094f12f
KW
873 bdrv_drain_all();
874
f9092b10
MA
875 if (bs == bs_snapshots) {
876 bs_snapshots = NULL;
877 }
557df6ac 878 if (bs->backing_hd) {
ea2384d3 879 bdrv_delete(bs->backing_hd);
557df6ac
SH
880 bs->backing_hd = NULL;
881 }
ea2384d3 882 bs->drv->bdrv_close(bs);
7267c094 883 g_free(bs->opaque);
ea2384d3
FB
884#ifdef _WIN32
885 if (bs->is_temporary) {
886 unlink(bs->filename);
887 }
67b915a5 888#endif
ea2384d3
FB
889 bs->opaque = NULL;
890 bs->drv = NULL;
53fec9d3 891 bs->copy_on_read = 0;
a275fa42
PB
892 bs->backing_file[0] = '\0';
893 bs->backing_format[0] = '\0';
6405875c
PB
894 bs->total_sectors = 0;
895 bs->encrypted = 0;
896 bs->valid_key = 0;
897 bs->sg = 0;
898 bs->growable = 0;
b338082b 899
66f82cee 900 if (bs->file != NULL) {
0ac9377d
PB
901 bdrv_delete(bs->file);
902 bs->file = NULL;
66f82cee 903 }
b338082b 904 }
98f90dba 905
9ca11154
PH
906 bdrv_dev_change_media_cb(bs, false);
907
98f90dba
ZYW
908 /*throttling disk I/O limits*/
909 if (bs->io_limits_enabled) {
910 bdrv_io_limits_disable(bs);
911 }
b338082b
FB
912}
913
2bc93fed
MK
914void bdrv_close_all(void)
915{
916 BlockDriverState *bs;
917
918 QTAILQ_FOREACH(bs, &bdrv_states, list) {
919 bdrv_close(bs);
920 }
921}
922
922453bc
SH
923/*
924 * Wait for pending requests to complete across all BlockDriverStates
925 *
926 * This function does not flush data to disk, use bdrv_flush_all() for that
927 * after calling this function.
4c355d53
ZYW
928 *
929 * Note that completion of an asynchronous I/O operation can trigger any
930 * number of other I/O operations on other devices---for example a coroutine
931 * can be arbitrarily complex and a constant flow of I/O can come until the
932 * coroutine is complete. Because of this, it is not possible to have a
933 * function to drain a single device's I/O queue.
922453bc
SH
934 */
935void bdrv_drain_all(void)
936{
937 BlockDriverState *bs;
4c355d53
ZYW
938 bool busy;
939
940 do {
941 busy = qemu_aio_wait();
922453bc 942
4c355d53
ZYW
943 /* FIXME: We do not have timer support here, so this is effectively
944 * a busy wait.
945 */
946 QTAILQ_FOREACH(bs, &bdrv_states, list) {
947 if (!qemu_co_queue_empty(&bs->throttled_reqs)) {
948 qemu_co_queue_restart_all(&bs->throttled_reqs);
949 busy = true;
950 }
951 }
952 } while (busy);
922453bc
SH
953
954 /* If requests are still pending there is a bug somewhere */
955 QTAILQ_FOREACH(bs, &bdrv_states, list) {
956 assert(QLIST_EMPTY(&bs->tracked_requests));
957 assert(qemu_co_queue_empty(&bs->throttled_reqs));
958 }
959}
960
d22b2f41
RH
961/* make a BlockDriverState anonymous by removing from bdrv_state list.
962 Also, NULL terminate the device_name to prevent double remove */
963void bdrv_make_anon(BlockDriverState *bs)
964{
965 if (bs->device_name[0] != '\0') {
966 QTAILQ_REMOVE(&bdrv_states, bs, list);
967 }
968 bs->device_name[0] = '\0';
969}
970
e023b2e2
PB
971static void bdrv_rebind(BlockDriverState *bs)
972{
973 if (bs->drv && bs->drv->bdrv_rebind) {
974 bs->drv->bdrv_rebind(bs);
975 }
976}
977
4ddc07ca
PB
978static void bdrv_move_feature_fields(BlockDriverState *bs_dest,
979 BlockDriverState *bs_src)
8802d1fd 980{
4ddc07ca
PB
981 /* move some fields that need to stay attached to the device */
982 bs_dest->open_flags = bs_src->open_flags;
8802d1fd
JC
983
984 /* dev info */
4ddc07ca
PB
985 bs_dest->dev_ops = bs_src->dev_ops;
986 bs_dest->dev_opaque = bs_src->dev_opaque;
987 bs_dest->dev = bs_src->dev;
988 bs_dest->buffer_alignment = bs_src->buffer_alignment;
989 bs_dest->copy_on_read = bs_src->copy_on_read;
8802d1fd 990
4ddc07ca 991 bs_dest->enable_write_cache = bs_src->enable_write_cache;
c4a248a1 992
8802d1fd 993 /* i/o timing parameters */
4ddc07ca
PB
994 bs_dest->slice_time = bs_src->slice_time;
995 bs_dest->slice_start = bs_src->slice_start;
996 bs_dest->slice_end = bs_src->slice_end;
997 bs_dest->io_limits = bs_src->io_limits;
998 bs_dest->io_base = bs_src->io_base;
999 bs_dest->throttled_reqs = bs_src->throttled_reqs;
1000 bs_dest->block_timer = bs_src->block_timer;
1001 bs_dest->io_limits_enabled = bs_src->io_limits_enabled;
8802d1fd 1002
8802d1fd 1003 /* r/w error */
4ddc07ca
PB
1004 bs_dest->on_read_error = bs_src->on_read_error;
1005 bs_dest->on_write_error = bs_src->on_write_error;
8802d1fd
JC
1006
1007 /* i/o status */
4ddc07ca
PB
1008 bs_dest->iostatus_enabled = bs_src->iostatus_enabled;
1009 bs_dest->iostatus = bs_src->iostatus;
8802d1fd 1010
a9fc4408 1011 /* dirty bitmap */
4ddc07ca
PB
1012 bs_dest->dirty_count = bs_src->dirty_count;
1013 bs_dest->dirty_bitmap = bs_src->dirty_bitmap;
a9fc4408
PB
1014
1015 /* job */
4ddc07ca
PB
1016 bs_dest->in_use = bs_src->in_use;
1017 bs_dest->job = bs_src->job;
a9fc4408 1018
8802d1fd 1019 /* keep the same entry in bdrv_states */
4ddc07ca
PB
1020 pstrcpy(bs_dest->device_name, sizeof(bs_dest->device_name),
1021 bs_src->device_name);
1022 bs_dest->list = bs_src->list;
1023}
8802d1fd 1024
4ddc07ca
PB
1025/*
1026 * Swap bs contents for two image chains while they are live,
1027 * while keeping required fields on the BlockDriverState that is
1028 * actually attached to a device.
1029 *
1030 * This will modify the BlockDriverState fields, and swap contents
1031 * between bs_new and bs_old. Both bs_new and bs_old are modified.
1032 *
1033 * bs_new is required to be anonymous.
1034 *
1035 * This function does not create any image files.
1036 */
1037void bdrv_swap(BlockDriverState *bs_new, BlockDriverState *bs_old)
1038{
1039 BlockDriverState tmp;
f6801b83 1040
4ddc07ca
PB
1041 /* bs_new must be anonymous and shouldn't have anything fancy enabled */
1042 assert(bs_new->device_name[0] == '\0');
1043 assert(bs_new->dirty_bitmap == NULL);
1044 assert(bs_new->job == NULL);
1045 assert(bs_new->dev == NULL);
1046 assert(bs_new->in_use == 0);
1047 assert(bs_new->io_limits_enabled == false);
1048 assert(bs_new->block_timer == NULL);
8802d1fd 1049
4ddc07ca
PB
1050 tmp = *bs_new;
1051 *bs_new = *bs_old;
1052 *bs_old = tmp;
a9fc4408 1053
4ddc07ca
PB
1054 /* there are some fields that should not be swapped, move them back */
1055 bdrv_move_feature_fields(&tmp, bs_old);
1056 bdrv_move_feature_fields(bs_old, bs_new);
1057 bdrv_move_feature_fields(bs_new, &tmp);
8802d1fd 1058
4ddc07ca
PB
1059 /* bs_new shouldn't be in bdrv_states even after the swap! */
1060 assert(bs_new->device_name[0] == '\0');
1061
1062 /* Check a few fields that should remain attached to the device */
1063 assert(bs_new->dev == NULL);
1064 assert(bs_new->job == NULL);
1065 assert(bs_new->in_use == 0);
1066 assert(bs_new->io_limits_enabled == false);
1067 assert(bs_new->block_timer == NULL);
e023b2e2
PB
1068
1069 bdrv_rebind(bs_new);
4ddc07ca
PB
1070 bdrv_rebind(bs_old);
1071}
1072
1073/*
1074 * Add new bs contents at the top of an image chain while the chain is
1075 * live, while keeping required fields on the top layer.
1076 *
1077 * This will modify the BlockDriverState fields, and swap contents
1078 * between bs_new and bs_top. Both bs_new and bs_top are modified.
1079 *
1080 * bs_new is required to be anonymous.
1081 *
1082 * This function does not create any image files.
1083 */
1084void bdrv_append(BlockDriverState *bs_new, BlockDriverState *bs_top)
1085{
1086 bdrv_swap(bs_new, bs_top);
1087
1088 /* The contents of 'tmp' will become bs_top, as we are
1089 * swapping bs_new and bs_top contents. */
1090 bs_top->backing_hd = bs_new;
1091 bs_top->open_flags &= ~BDRV_O_NO_BACKING;
1092 pstrcpy(bs_top->backing_file, sizeof(bs_top->backing_file),
1093 bs_new->filename);
1094 pstrcpy(bs_top->backing_format, sizeof(bs_top->backing_format),
1095 bs_new->drv ? bs_new->drv->format_name : "");
8802d1fd
JC
1096}
1097
b338082b
FB
1098void bdrv_delete(BlockDriverState *bs)
1099{
fa879d62 1100 assert(!bs->dev);
3e914655
PB
1101 assert(!bs->job);
1102 assert(!bs->in_use);
18846dee 1103
1b7bdbc1 1104 /* remove from list, if necessary */
d22b2f41 1105 bdrv_make_anon(bs);
34c6f050 1106
b338082b 1107 bdrv_close(bs);
66f82cee 1108
f9092b10 1109 assert(bs != bs_snapshots);
7267c094 1110 g_free(bs);
fc01f7e7
FB
1111}
1112
fa879d62
MA
1113int bdrv_attach_dev(BlockDriverState *bs, void *dev)
1114/* TODO change to DeviceState *dev when all users are qdevified */
18846dee 1115{
fa879d62 1116 if (bs->dev) {
18846dee
MA
1117 return -EBUSY;
1118 }
fa879d62 1119 bs->dev = dev;
28a7282a 1120 bdrv_iostatus_reset(bs);
18846dee
MA
1121 return 0;
1122}
1123
fa879d62
MA
1124/* TODO qdevified devices don't use this, remove when devices are qdevified */
1125void bdrv_attach_dev_nofail(BlockDriverState *bs, void *dev)
18846dee 1126{
fa879d62
MA
1127 if (bdrv_attach_dev(bs, dev) < 0) {
1128 abort();
1129 }
1130}
1131
1132void bdrv_detach_dev(BlockDriverState *bs, void *dev)
1133/* TODO change to DeviceState *dev when all users are qdevified */
1134{
1135 assert(bs->dev == dev);
1136 bs->dev = NULL;
0e49de52
MA
1137 bs->dev_ops = NULL;
1138 bs->dev_opaque = NULL;
29e05f20 1139 bs->buffer_alignment = 512;
18846dee
MA
1140}
1141
fa879d62
MA
1142/* TODO change to return DeviceState * when all users are qdevified */
1143void *bdrv_get_attached_dev(BlockDriverState *bs)
18846dee 1144{
fa879d62 1145 return bs->dev;
18846dee
MA
1146}
1147
0e49de52
MA
1148void bdrv_set_dev_ops(BlockDriverState *bs, const BlockDevOps *ops,
1149 void *opaque)
1150{
1151 bs->dev_ops = ops;
1152 bs->dev_opaque = opaque;
2c6942fa
MA
1153 if (bdrv_dev_has_removable_media(bs) && bs == bs_snapshots) {
1154 bs_snapshots = NULL;
1155 }
0e49de52
MA
1156}
1157
329c0a48
LC
1158void bdrv_emit_qmp_error_event(const BlockDriverState *bdrv,
1159 BlockQMPEventAction action, int is_read)
1160{
1161 QObject *data;
1162 const char *action_str;
1163
1164 switch (action) {
1165 case BDRV_ACTION_REPORT:
1166 action_str = "report";
1167 break;
1168 case BDRV_ACTION_IGNORE:
1169 action_str = "ignore";
1170 break;
1171 case BDRV_ACTION_STOP:
1172 action_str = "stop";
1173 break;
1174 default:
1175 abort();
1176 }
1177
1178 data = qobject_from_jsonf("{ 'device': %s, 'action': %s, 'operation': %s }",
1179 bdrv->device_name,
1180 action_str,
1181 is_read ? "read" : "write");
1182 monitor_protocol_event(QEVENT_BLOCK_IO_ERROR, data);
1183
1184 qobject_decref(data);
1185}
1186
6f382ed2
LC
1187static void bdrv_emit_qmp_eject_event(BlockDriverState *bs, bool ejected)
1188{
1189 QObject *data;
1190
1191 data = qobject_from_jsonf("{ 'device': %s, 'tray-open': %i }",
1192 bdrv_get_device_name(bs), ejected);
1193 monitor_protocol_event(QEVENT_DEVICE_TRAY_MOVED, data);
1194
1195 qobject_decref(data);
1196}
1197
7d4b4ba5 1198static void bdrv_dev_change_media_cb(BlockDriverState *bs, bool load)
0e49de52 1199{
145feb17 1200 if (bs->dev_ops && bs->dev_ops->change_media_cb) {
6f382ed2 1201 bool tray_was_closed = !bdrv_dev_is_tray_open(bs);
7d4b4ba5 1202 bs->dev_ops->change_media_cb(bs->dev_opaque, load);
6f382ed2
LC
1203 if (tray_was_closed) {
1204 /* tray open */
1205 bdrv_emit_qmp_eject_event(bs, true);
1206 }
1207 if (load) {
1208 /* tray close */
1209 bdrv_emit_qmp_eject_event(bs, false);
1210 }
145feb17
MA
1211 }
1212}
1213
2c6942fa
MA
1214bool bdrv_dev_has_removable_media(BlockDriverState *bs)
1215{
1216 return !bs->dev || (bs->dev_ops && bs->dev_ops->change_media_cb);
1217}
1218
025ccaa7
PB
1219void bdrv_dev_eject_request(BlockDriverState *bs, bool force)
1220{
1221 if (bs->dev_ops && bs->dev_ops->eject_request_cb) {
1222 bs->dev_ops->eject_request_cb(bs->dev_opaque, force);
1223 }
1224}
1225
e4def80b
MA
1226bool bdrv_dev_is_tray_open(BlockDriverState *bs)
1227{
1228 if (bs->dev_ops && bs->dev_ops->is_tray_open) {
1229 return bs->dev_ops->is_tray_open(bs->dev_opaque);
1230 }
1231 return false;
1232}
1233
145feb17
MA
1234static void bdrv_dev_resize_cb(BlockDriverState *bs)
1235{
1236 if (bs->dev_ops && bs->dev_ops->resize_cb) {
1237 bs->dev_ops->resize_cb(bs->dev_opaque);
0e49de52
MA
1238 }
1239}
1240
f107639a
MA
1241bool bdrv_dev_is_medium_locked(BlockDriverState *bs)
1242{
1243 if (bs->dev_ops && bs->dev_ops->is_medium_locked) {
1244 return bs->dev_ops->is_medium_locked(bs->dev_opaque);
1245 }
1246 return false;
1247}
1248
e97fc193
AL
1249/*
1250 * Run consistency checks on an image
1251 *
e076f338 1252 * Returns 0 if the check could be completed (it doesn't mean that the image is
a1c7273b 1253 * free of errors) or -errno when an internal error occurred. The results of the
e076f338 1254 * check are stored in res.
e97fc193 1255 */
4534ff54 1256int bdrv_check(BlockDriverState *bs, BdrvCheckResult *res, BdrvCheckMode fix)
e97fc193
AL
1257{
1258 if (bs->drv->bdrv_check == NULL) {
1259 return -ENOTSUP;
1260 }
1261
e076f338 1262 memset(res, 0, sizeof(*res));
4534ff54 1263 return bs->drv->bdrv_check(bs, res, fix);
e97fc193
AL
1264}
1265
8a426614
KW
1266#define COMMIT_BUF_SECTORS 2048
1267
33e3963e
FB
1268/* commit COW file into the raw image */
1269int bdrv_commit(BlockDriverState *bs)
1270{
19cb3738 1271 BlockDriver *drv = bs->drv;
ee181196 1272 BlockDriver *backing_drv;
8a426614
KW
1273 int64_t sector, total_sectors;
1274 int n, ro, open_flags;
4dca4b63 1275 int ret = 0, rw_ret = 0;
8a426614 1276 uint8_t *buf;
4dca4b63
NS
1277 char filename[1024];
1278 BlockDriverState *bs_rw, *bs_ro;
33e3963e 1279
19cb3738
FB
1280 if (!drv)
1281 return -ENOMEDIUM;
4dca4b63
NS
1282
1283 if (!bs->backing_hd) {
1284 return -ENOTSUP;
33e3963e
FB
1285 }
1286
4dca4b63
NS
1287 if (bs->backing_hd->keep_read_only) {
1288 return -EACCES;
1289 }
ee181196 1290
2d3735d3
SH
1291 if (bdrv_in_use(bs) || bdrv_in_use(bs->backing_hd)) {
1292 return -EBUSY;
1293 }
1294
ee181196 1295 backing_drv = bs->backing_hd->drv;
4dca4b63
NS
1296 ro = bs->backing_hd->read_only;
1297 strncpy(filename, bs->backing_hd->filename, sizeof(filename));
1298 open_flags = bs->backing_hd->open_flags;
1299
1300 if (ro) {
1301 /* re-open as RW */
1302 bdrv_delete(bs->backing_hd);
1303 bs->backing_hd = NULL;
1304 bs_rw = bdrv_new("");
ee181196
KW
1305 rw_ret = bdrv_open(bs_rw, filename, open_flags | BDRV_O_RDWR,
1306 backing_drv);
4dca4b63
NS
1307 if (rw_ret < 0) {
1308 bdrv_delete(bs_rw);
1309 /* try to re-open read-only */
1310 bs_ro = bdrv_new("");
ee181196
KW
1311 ret = bdrv_open(bs_ro, filename, open_flags & ~BDRV_O_RDWR,
1312 backing_drv);
4dca4b63
NS
1313 if (ret < 0) {
1314 bdrv_delete(bs_ro);
1315 /* drive not functional anymore */
1316 bs->drv = NULL;
1317 return ret;
1318 }
1319 bs->backing_hd = bs_ro;
1320 return rw_ret;
1321 }
1322 bs->backing_hd = bs_rw;
ea2384d3 1323 }
33e3963e 1324
6ea44308 1325 total_sectors = bdrv_getlength(bs) >> BDRV_SECTOR_BITS;
7267c094 1326 buf = g_malloc(COMMIT_BUF_SECTORS * BDRV_SECTOR_SIZE);
8a426614
KW
1327
1328 for (sector = 0; sector < total_sectors; sector += n) {
05c4af54 1329 if (bdrv_is_allocated(bs, sector, COMMIT_BUF_SECTORS, &n)) {
8a426614
KW
1330
1331 if (bdrv_read(bs, sector, buf, n) != 0) {
1332 ret = -EIO;
1333 goto ro_cleanup;
1334 }
1335
1336 if (bdrv_write(bs->backing_hd, sector, buf, n) != 0) {
1337 ret = -EIO;
1338 goto ro_cleanup;
1339 }
ea2384d3 1340 }
33e3963e 1341 }
95389c86 1342
1d44952f
CH
1343 if (drv->bdrv_make_empty) {
1344 ret = drv->bdrv_make_empty(bs);
1345 bdrv_flush(bs);
1346 }
95389c86 1347
3f5075ae
CH
1348 /*
1349 * Make sure all data we wrote to the backing device is actually
1350 * stable on disk.
1351 */
1352 if (bs->backing_hd)
1353 bdrv_flush(bs->backing_hd);
4dca4b63
NS
1354
1355ro_cleanup:
7267c094 1356 g_free(buf);
4dca4b63
NS
1357
1358 if (ro) {
1359 /* re-open as RO */
1360 bdrv_delete(bs->backing_hd);
1361 bs->backing_hd = NULL;
1362 bs_ro = bdrv_new("");
ee181196
KW
1363 ret = bdrv_open(bs_ro, filename, open_flags & ~BDRV_O_RDWR,
1364 backing_drv);
4dca4b63
NS
1365 if (ret < 0) {
1366 bdrv_delete(bs_ro);
1367 /* drive not functional anymore */
1368 bs->drv = NULL;
1369 return ret;
1370 }
1371 bs->backing_hd = bs_ro;
1372 bs->backing_hd->keep_read_only = 0;
1373 }
1374
1d44952f 1375 return ret;
33e3963e
FB
1376}
1377
e8877497 1378int bdrv_commit_all(void)
6ab4b5ab
MA
1379{
1380 BlockDriverState *bs;
1381
1382 QTAILQ_FOREACH(bs, &bdrv_states, list) {
e8877497
SH
1383 int ret = bdrv_commit(bs);
1384 if (ret < 0) {
1385 return ret;
1386 }
6ab4b5ab 1387 }
e8877497 1388 return 0;
6ab4b5ab
MA
1389}
1390
dbffbdcf
SH
1391struct BdrvTrackedRequest {
1392 BlockDriverState *bs;
1393 int64_t sector_num;
1394 int nb_sectors;
1395 bool is_write;
1396 QLIST_ENTRY(BdrvTrackedRequest) list;
5f8b6491 1397 Coroutine *co; /* owner, used for deadlock detection */
f4658285 1398 CoQueue wait_queue; /* coroutines blocked on this request */
dbffbdcf
SH
1399};
1400
1401/**
1402 * Remove an active request from the tracked requests list
1403 *
1404 * This function should be called when a tracked request is completing.
1405 */
1406static void tracked_request_end(BdrvTrackedRequest *req)
1407{
1408 QLIST_REMOVE(req, list);
f4658285 1409 qemu_co_queue_restart_all(&req->wait_queue);
dbffbdcf
SH
1410}
1411
1412/**
1413 * Add an active request to the tracked requests list
1414 */
1415static void tracked_request_begin(BdrvTrackedRequest *req,
1416 BlockDriverState *bs,
1417 int64_t sector_num,
1418 int nb_sectors, bool is_write)
1419{
1420 *req = (BdrvTrackedRequest){
1421 .bs = bs,
1422 .sector_num = sector_num,
1423 .nb_sectors = nb_sectors,
1424 .is_write = is_write,
5f8b6491 1425 .co = qemu_coroutine_self(),
dbffbdcf
SH
1426 };
1427
f4658285
SH
1428 qemu_co_queue_init(&req->wait_queue);
1429
dbffbdcf
SH
1430 QLIST_INSERT_HEAD(&bs->tracked_requests, req, list);
1431}
1432
d83947ac
SH
1433/**
1434 * Round a region to cluster boundaries
1435 */
1436static void round_to_clusters(BlockDriverState *bs,
1437 int64_t sector_num, int nb_sectors,
1438 int64_t *cluster_sector_num,
1439 int *cluster_nb_sectors)
1440{
1441 BlockDriverInfo bdi;
1442
1443 if (bdrv_get_info(bs, &bdi) < 0 || bdi.cluster_size == 0) {
1444 *cluster_sector_num = sector_num;
1445 *cluster_nb_sectors = nb_sectors;
1446 } else {
1447 int64_t c = bdi.cluster_size / BDRV_SECTOR_SIZE;
1448 *cluster_sector_num = QEMU_ALIGN_DOWN(sector_num, c);
1449 *cluster_nb_sectors = QEMU_ALIGN_UP(sector_num - *cluster_sector_num +
1450 nb_sectors, c);
1451 }
1452}
1453
f4658285
SH
1454static bool tracked_request_overlaps(BdrvTrackedRequest *req,
1455 int64_t sector_num, int nb_sectors) {
d83947ac
SH
1456 /* aaaa bbbb */
1457 if (sector_num >= req->sector_num + req->nb_sectors) {
1458 return false;
1459 }
1460 /* bbbb aaaa */
1461 if (req->sector_num >= sector_num + nb_sectors) {
1462 return false;
1463 }
1464 return true;
f4658285
SH
1465}
1466
1467static void coroutine_fn wait_for_overlapping_requests(BlockDriverState *bs,
1468 int64_t sector_num, int nb_sectors)
1469{
1470 BdrvTrackedRequest *req;
d83947ac
SH
1471 int64_t cluster_sector_num;
1472 int cluster_nb_sectors;
f4658285
SH
1473 bool retry;
1474
d83947ac
SH
1475 /* If we touch the same cluster it counts as an overlap. This guarantees
1476 * that allocating writes will be serialized and not race with each other
1477 * for the same cluster. For example, in copy-on-read it ensures that the
1478 * CoR read and write operations are atomic and guest writes cannot
1479 * interleave between them.
1480 */
1481 round_to_clusters(bs, sector_num, nb_sectors,
1482 &cluster_sector_num, &cluster_nb_sectors);
1483
f4658285
SH
1484 do {
1485 retry = false;
1486 QLIST_FOREACH(req, &bs->tracked_requests, list) {
d83947ac
SH
1487 if (tracked_request_overlaps(req, cluster_sector_num,
1488 cluster_nb_sectors)) {
5f8b6491
SH
1489 /* Hitting this means there was a reentrant request, for
1490 * example, a block driver issuing nested requests. This must
1491 * never happen since it means deadlock.
1492 */
1493 assert(qemu_coroutine_self() != req->co);
1494
f4658285
SH
1495 qemu_co_queue_wait(&req->wait_queue);
1496 retry = true;
1497 break;
1498 }
1499 }
1500 } while (retry);
1501}
1502
756e6736
KW
1503/*
1504 * Return values:
1505 * 0 - success
1506 * -EINVAL - backing format specified, but no file
1507 * -ENOSPC - can't update the backing file because no space is left in the
1508 * image file header
1509 * -ENOTSUP - format driver doesn't support changing the backing file
1510 */
1511int bdrv_change_backing_file(BlockDriverState *bs,
1512 const char *backing_file, const char *backing_fmt)
1513{
1514 BlockDriver *drv = bs->drv;
469ef350 1515 int ret;
756e6736 1516
5f377794
PB
1517 /* Backing file format doesn't make sense without a backing file */
1518 if (backing_fmt && !backing_file) {
1519 return -EINVAL;
1520 }
1521
756e6736 1522 if (drv->bdrv_change_backing_file != NULL) {
469ef350 1523 ret = drv->bdrv_change_backing_file(bs, backing_file, backing_fmt);
756e6736 1524 } else {
469ef350 1525 ret = -ENOTSUP;
756e6736 1526 }
469ef350
PB
1527
1528 if (ret == 0) {
1529 pstrcpy(bs->backing_file, sizeof(bs->backing_file), backing_file ?: "");
1530 pstrcpy(bs->backing_format, sizeof(bs->backing_format), backing_fmt ?: "");
1531 }
1532 return ret;
756e6736
KW
1533}
1534
71d0770c
AL
1535static int bdrv_check_byte_request(BlockDriverState *bs, int64_t offset,
1536 size_t size)
1537{
1538 int64_t len;
1539
1540 if (!bdrv_is_inserted(bs))
1541 return -ENOMEDIUM;
1542
1543 if (bs->growable)
1544 return 0;
1545
1546 len = bdrv_getlength(bs);
1547
fbb7b4e0
KW
1548 if (offset < 0)
1549 return -EIO;
1550
1551 if ((offset > len) || (len - offset < size))
71d0770c
AL
1552 return -EIO;
1553
1554 return 0;
1555}
1556
1557static int bdrv_check_request(BlockDriverState *bs, int64_t sector_num,
1558 int nb_sectors)
1559{
eb5a3165
JS
1560 return bdrv_check_byte_request(bs, sector_num * BDRV_SECTOR_SIZE,
1561 nb_sectors * BDRV_SECTOR_SIZE);
71d0770c
AL
1562}
1563
1c9805a3
SH
1564typedef struct RwCo {
1565 BlockDriverState *bs;
1566 int64_t sector_num;
1567 int nb_sectors;
1568 QEMUIOVector *qiov;
1569 bool is_write;
1570 int ret;
1571} RwCo;
1572
1573static void coroutine_fn bdrv_rw_co_entry(void *opaque)
fc01f7e7 1574{
1c9805a3 1575 RwCo *rwco = opaque;
ea2384d3 1576
1c9805a3
SH
1577 if (!rwco->is_write) {
1578 rwco->ret = bdrv_co_do_readv(rwco->bs, rwco->sector_num,
470c0504 1579 rwco->nb_sectors, rwco->qiov, 0);
1c9805a3
SH
1580 } else {
1581 rwco->ret = bdrv_co_do_writev(rwco->bs, rwco->sector_num,
f08f2dda 1582 rwco->nb_sectors, rwco->qiov, 0);
1c9805a3
SH
1583 }
1584}
e7a8a783 1585
1c9805a3
SH
1586/*
1587 * Process a synchronous request using coroutines
1588 */
1589static int bdrv_rw_co(BlockDriverState *bs, int64_t sector_num, uint8_t *buf,
1590 int nb_sectors, bool is_write)
1591{
1592 QEMUIOVector qiov;
1593 struct iovec iov = {
1594 .iov_base = (void *)buf,
1595 .iov_len = nb_sectors * BDRV_SECTOR_SIZE,
1596 };
1597 Coroutine *co;
1598 RwCo rwco = {
1599 .bs = bs,
1600 .sector_num = sector_num,
1601 .nb_sectors = nb_sectors,
1602 .qiov = &qiov,
1603 .is_write = is_write,
1604 .ret = NOT_DONE,
1605 };
e7a8a783 1606
1c9805a3 1607 qemu_iovec_init_external(&qiov, &iov, 1);
e7a8a783 1608
498e386c
ZYW
1609 /**
1610 * In sync call context, when the vcpu is blocked, this throttling timer
1611 * will not fire; so the I/O throttling function has to be disabled here
1612 * if it has been enabled.
1613 */
1614 if (bs->io_limits_enabled) {
1615 fprintf(stderr, "Disabling I/O throttling on '%s' due "
1616 "to synchronous I/O.\n", bdrv_get_device_name(bs));
1617 bdrv_io_limits_disable(bs);
1618 }
1619
1c9805a3
SH
1620 if (qemu_in_coroutine()) {
1621 /* Fast-path if already in coroutine context */
1622 bdrv_rw_co_entry(&rwco);
1623 } else {
1624 co = qemu_coroutine_create(bdrv_rw_co_entry);
1625 qemu_coroutine_enter(co, &rwco);
1626 while (rwco.ret == NOT_DONE) {
1627 qemu_aio_wait();
1628 }
1629 }
1630 return rwco.ret;
1631}
b338082b 1632
1c9805a3
SH
1633/* return < 0 if error. See bdrv_write() for the return codes */
1634int bdrv_read(BlockDriverState *bs, int64_t sector_num,
1635 uint8_t *buf, int nb_sectors)
1636{
1637 return bdrv_rw_co(bs, sector_num, buf, nb_sectors, false);
fc01f7e7
FB
1638}
1639
07d27a44
MA
1640/* Just like bdrv_read(), but with I/O throttling temporarily disabled */
1641int bdrv_read_unthrottled(BlockDriverState *bs, int64_t sector_num,
1642 uint8_t *buf, int nb_sectors)
1643{
1644 bool enabled;
1645 int ret;
1646
1647 enabled = bs->io_limits_enabled;
1648 bs->io_limits_enabled = false;
1649 ret = bdrv_read(bs, 0, buf, 1);
1650 bs->io_limits_enabled = enabled;
1651 return ret;
1652}
1653
71df14fc
PB
1654#define BITS_PER_LONG (sizeof(unsigned long) * 8)
1655
7cd1e32a 1656static void set_dirty_bitmap(BlockDriverState *bs, int64_t sector_num,
a55eb92c 1657 int nb_sectors, int dirty)
7cd1e32a 1658{
1659 int64_t start, end;
c6d22830 1660 unsigned long val, idx, bit;
a55eb92c 1661
6ea44308 1662 start = sector_num / BDRV_SECTORS_PER_DIRTY_CHUNK;
c6d22830 1663 end = (sector_num + nb_sectors - 1) / BDRV_SECTORS_PER_DIRTY_CHUNK;
a55eb92c
JK
1664
1665 for (; start <= end; start++) {
71df14fc
PB
1666 idx = start / BITS_PER_LONG;
1667 bit = start % BITS_PER_LONG;
c6d22830
JK
1668 val = bs->dirty_bitmap[idx];
1669 if (dirty) {
6d59fec1 1670 if (!(val & (1UL << bit))) {
aaa0eb75 1671 bs->dirty_count++;
6d59fec1 1672 val |= 1UL << bit;
aaa0eb75 1673 }
c6d22830 1674 } else {
6d59fec1 1675 if (val & (1UL << bit)) {
aaa0eb75 1676 bs->dirty_count--;
6d59fec1 1677 val &= ~(1UL << bit);
aaa0eb75 1678 }
c6d22830
JK
1679 }
1680 bs->dirty_bitmap[idx] = val;
7cd1e32a 1681 }
1682}
1683
5fafdf24 1684/* Return < 0 if error. Important errors are:
19cb3738
FB
1685 -EIO generic I/O error (may happen for all errors)
1686 -ENOMEDIUM No media inserted.
1687 -EINVAL Invalid sector number or nb_sectors
1688 -EACCES Trying to write a read-only device
1689*/
5fafdf24 1690int bdrv_write(BlockDriverState *bs, int64_t sector_num,
fc01f7e7
FB
1691 const uint8_t *buf, int nb_sectors)
1692{
1c9805a3 1693 return bdrv_rw_co(bs, sector_num, (uint8_t *)buf, nb_sectors, true);
83f64091
FB
1694}
1695
eda578e5
AL
1696int bdrv_pread(BlockDriverState *bs, int64_t offset,
1697 void *buf, int count1)
83f64091 1698{
6ea44308 1699 uint8_t tmp_buf[BDRV_SECTOR_SIZE];
83f64091
FB
1700 int len, nb_sectors, count;
1701 int64_t sector_num;
9a8c4cce 1702 int ret;
83f64091
FB
1703
1704 count = count1;
1705 /* first read to align to sector start */
6ea44308 1706 len = (BDRV_SECTOR_SIZE - offset) & (BDRV_SECTOR_SIZE - 1);
83f64091
FB
1707 if (len > count)
1708 len = count;
6ea44308 1709 sector_num = offset >> BDRV_SECTOR_BITS;
83f64091 1710 if (len > 0) {
9a8c4cce
KW
1711 if ((ret = bdrv_read(bs, sector_num, tmp_buf, 1)) < 0)
1712 return ret;
6ea44308 1713 memcpy(buf, tmp_buf + (offset & (BDRV_SECTOR_SIZE - 1)), len);
83f64091
FB
1714 count -= len;
1715 if (count == 0)
1716 return count1;
1717 sector_num++;
1718 buf += len;
1719 }
1720
1721 /* read the sectors "in place" */
6ea44308 1722 nb_sectors = count >> BDRV_SECTOR_BITS;
83f64091 1723 if (nb_sectors > 0) {
9a8c4cce
KW
1724 if ((ret = bdrv_read(bs, sector_num, buf, nb_sectors)) < 0)
1725 return ret;
83f64091 1726 sector_num += nb_sectors;
6ea44308 1727 len = nb_sectors << BDRV_SECTOR_BITS;
83f64091
FB
1728 buf += len;
1729 count -= len;
1730 }
1731
1732 /* add data from the last sector */
1733 if (count > 0) {
9a8c4cce
KW
1734 if ((ret = bdrv_read(bs, sector_num, tmp_buf, 1)) < 0)
1735 return ret;
83f64091
FB
1736 memcpy(buf, tmp_buf, count);
1737 }
1738 return count1;
1739}
1740
eda578e5
AL
1741int bdrv_pwrite(BlockDriverState *bs, int64_t offset,
1742 const void *buf, int count1)
83f64091 1743{
6ea44308 1744 uint8_t tmp_buf[BDRV_SECTOR_SIZE];
83f64091
FB
1745 int len, nb_sectors, count;
1746 int64_t sector_num;
9a8c4cce 1747 int ret;
83f64091
FB
1748
1749 count = count1;
1750 /* first write to align to sector start */
6ea44308 1751 len = (BDRV_SECTOR_SIZE - offset) & (BDRV_SECTOR_SIZE - 1);
83f64091
FB
1752 if (len > count)
1753 len = count;
6ea44308 1754 sector_num = offset >> BDRV_SECTOR_BITS;
83f64091 1755 if (len > 0) {
9a8c4cce
KW
1756 if ((ret = bdrv_read(bs, sector_num, tmp_buf, 1)) < 0)
1757 return ret;
6ea44308 1758 memcpy(tmp_buf + (offset & (BDRV_SECTOR_SIZE - 1)), buf, len);
9a8c4cce
KW
1759 if ((ret = bdrv_write(bs, sector_num, tmp_buf, 1)) < 0)
1760 return ret;
83f64091
FB
1761 count -= len;
1762 if (count == 0)
1763 return count1;
1764 sector_num++;
1765 buf += len;
1766 }
1767
1768 /* write the sectors "in place" */
6ea44308 1769 nb_sectors = count >> BDRV_SECTOR_BITS;
83f64091 1770 if (nb_sectors > 0) {
9a8c4cce
KW
1771 if ((ret = bdrv_write(bs, sector_num, buf, nb_sectors)) < 0)
1772 return ret;
83f64091 1773 sector_num += nb_sectors;
6ea44308 1774 len = nb_sectors << BDRV_SECTOR_BITS;
83f64091
FB
1775 buf += len;
1776 count -= len;
1777 }
1778
1779 /* add data from the last sector */
1780 if (count > 0) {
9a8c4cce
KW
1781 if ((ret = bdrv_read(bs, sector_num, tmp_buf, 1)) < 0)
1782 return ret;
83f64091 1783 memcpy(tmp_buf, buf, count);
9a8c4cce
KW
1784 if ((ret = bdrv_write(bs, sector_num, tmp_buf, 1)) < 0)
1785 return ret;
83f64091
FB
1786 }
1787 return count1;
1788}
83f64091 1789
f08145fe
KW
1790/*
1791 * Writes to the file and ensures that no writes are reordered across this
1792 * request (acts as a barrier)
1793 *
1794 * Returns 0 on success, -errno in error cases.
1795 */
1796int bdrv_pwrite_sync(BlockDriverState *bs, int64_t offset,
1797 const void *buf, int count)
1798{
1799 int ret;
1800
1801 ret = bdrv_pwrite(bs, offset, buf, count);
1802 if (ret < 0) {
1803 return ret;
1804 }
1805
f05fa4ad
PB
1806 /* No flush needed for cache modes that already do it */
1807 if (bs->enable_write_cache) {
f08145fe
KW
1808 bdrv_flush(bs);
1809 }
1810
1811 return 0;
1812}
1813
470c0504 1814static int coroutine_fn bdrv_co_do_copy_on_readv(BlockDriverState *bs,
ab185921
SH
1815 int64_t sector_num, int nb_sectors, QEMUIOVector *qiov)
1816{
1817 /* Perform I/O through a temporary buffer so that users who scribble over
1818 * their read buffer while the operation is in progress do not end up
1819 * modifying the image file. This is critical for zero-copy guest I/O
1820 * where anything might happen inside guest memory.
1821 */
1822 void *bounce_buffer;
1823
79c053bd 1824 BlockDriver *drv = bs->drv;
ab185921
SH
1825 struct iovec iov;
1826 QEMUIOVector bounce_qiov;
1827 int64_t cluster_sector_num;
1828 int cluster_nb_sectors;
1829 size_t skip_bytes;
1830 int ret;
1831
1832 /* Cover entire cluster so no additional backing file I/O is required when
1833 * allocating cluster in the image file.
1834 */
1835 round_to_clusters(bs, sector_num, nb_sectors,
1836 &cluster_sector_num, &cluster_nb_sectors);
1837
470c0504
SH
1838 trace_bdrv_co_do_copy_on_readv(bs, sector_num, nb_sectors,
1839 cluster_sector_num, cluster_nb_sectors);
ab185921
SH
1840
1841 iov.iov_len = cluster_nb_sectors * BDRV_SECTOR_SIZE;
1842 iov.iov_base = bounce_buffer = qemu_blockalign(bs, iov.iov_len);
1843 qemu_iovec_init_external(&bounce_qiov, &iov, 1);
1844
79c053bd
SH
1845 ret = drv->bdrv_co_readv(bs, cluster_sector_num, cluster_nb_sectors,
1846 &bounce_qiov);
ab185921
SH
1847 if (ret < 0) {
1848 goto err;
1849 }
1850
79c053bd
SH
1851 if (drv->bdrv_co_write_zeroes &&
1852 buffer_is_zero(bounce_buffer, iov.iov_len)) {
621f0589
KW
1853 ret = bdrv_co_do_write_zeroes(bs, cluster_sector_num,
1854 cluster_nb_sectors);
79c053bd 1855 } else {
f05fa4ad
PB
1856 /* This does not change the data on the disk, it is not necessary
1857 * to flush even in cache=writethrough mode.
1858 */
79c053bd 1859 ret = drv->bdrv_co_writev(bs, cluster_sector_num, cluster_nb_sectors,
ab185921 1860 &bounce_qiov);
79c053bd
SH
1861 }
1862
ab185921
SH
1863 if (ret < 0) {
1864 /* It might be okay to ignore write errors for guest requests. If this
1865 * is a deliberate copy-on-read then we don't want to ignore the error.
1866 * Simply report it in all cases.
1867 */
1868 goto err;
1869 }
1870
1871 skip_bytes = (sector_num - cluster_sector_num) * BDRV_SECTOR_SIZE;
03396148
MT
1872 qemu_iovec_from_buf(qiov, 0, bounce_buffer + skip_bytes,
1873 nb_sectors * BDRV_SECTOR_SIZE);
ab185921
SH
1874
1875err:
1876 qemu_vfree(bounce_buffer);
1877 return ret;
1878}
1879
c5fbe571
SH
1880/*
1881 * Handle a read request in coroutine context
1882 */
1883static int coroutine_fn bdrv_co_do_readv(BlockDriverState *bs,
470c0504
SH
1884 int64_t sector_num, int nb_sectors, QEMUIOVector *qiov,
1885 BdrvRequestFlags flags)
da1fa91d
KW
1886{
1887 BlockDriver *drv = bs->drv;
dbffbdcf
SH
1888 BdrvTrackedRequest req;
1889 int ret;
da1fa91d 1890
da1fa91d
KW
1891 if (!drv) {
1892 return -ENOMEDIUM;
1893 }
1894 if (bdrv_check_request(bs, sector_num, nb_sectors)) {
1895 return -EIO;
1896 }
1897
98f90dba
ZYW
1898 /* throttling disk read I/O */
1899 if (bs->io_limits_enabled) {
1900 bdrv_io_limits_intercept(bs, false, nb_sectors);
1901 }
1902
f4658285 1903 if (bs->copy_on_read) {
470c0504
SH
1904 flags |= BDRV_REQ_COPY_ON_READ;
1905 }
1906 if (flags & BDRV_REQ_COPY_ON_READ) {
1907 bs->copy_on_read_in_flight++;
1908 }
1909
1910 if (bs->copy_on_read_in_flight) {
f4658285
SH
1911 wait_for_overlapping_requests(bs, sector_num, nb_sectors);
1912 }
1913
dbffbdcf 1914 tracked_request_begin(&req, bs, sector_num, nb_sectors, false);
ab185921 1915
470c0504 1916 if (flags & BDRV_REQ_COPY_ON_READ) {
ab185921
SH
1917 int pnum;
1918
1919 ret = bdrv_co_is_allocated(bs, sector_num, nb_sectors, &pnum);
1920 if (ret < 0) {
1921 goto out;
1922 }
1923
1924 if (!ret || pnum != nb_sectors) {
470c0504 1925 ret = bdrv_co_do_copy_on_readv(bs, sector_num, nb_sectors, qiov);
ab185921
SH
1926 goto out;
1927 }
1928 }
1929
dbffbdcf 1930 ret = drv->bdrv_co_readv(bs, sector_num, nb_sectors, qiov);
ab185921
SH
1931
1932out:
dbffbdcf 1933 tracked_request_end(&req);
470c0504
SH
1934
1935 if (flags & BDRV_REQ_COPY_ON_READ) {
1936 bs->copy_on_read_in_flight--;
1937 }
1938
dbffbdcf 1939 return ret;
da1fa91d
KW
1940}
1941
c5fbe571 1942int coroutine_fn bdrv_co_readv(BlockDriverState *bs, int64_t sector_num,
da1fa91d
KW
1943 int nb_sectors, QEMUIOVector *qiov)
1944{
c5fbe571 1945 trace_bdrv_co_readv(bs, sector_num, nb_sectors);
da1fa91d 1946
470c0504
SH
1947 return bdrv_co_do_readv(bs, sector_num, nb_sectors, qiov, 0);
1948}
1949
1950int coroutine_fn bdrv_co_copy_on_readv(BlockDriverState *bs,
1951 int64_t sector_num, int nb_sectors, QEMUIOVector *qiov)
1952{
1953 trace_bdrv_co_copy_on_readv(bs, sector_num, nb_sectors);
1954
1955 return bdrv_co_do_readv(bs, sector_num, nb_sectors, qiov,
1956 BDRV_REQ_COPY_ON_READ);
c5fbe571
SH
1957}
1958
f08f2dda
SH
1959static int coroutine_fn bdrv_co_do_write_zeroes(BlockDriverState *bs,
1960 int64_t sector_num, int nb_sectors)
1961{
1962 BlockDriver *drv = bs->drv;
1963 QEMUIOVector qiov;
1964 struct iovec iov;
1965 int ret;
1966
621f0589
KW
1967 /* TODO Emulate only part of misaligned requests instead of letting block
1968 * drivers return -ENOTSUP and emulate everything */
1969
f08f2dda
SH
1970 /* First try the efficient write zeroes operation */
1971 if (drv->bdrv_co_write_zeroes) {
621f0589
KW
1972 ret = drv->bdrv_co_write_zeroes(bs, sector_num, nb_sectors);
1973 if (ret != -ENOTSUP) {
1974 return ret;
1975 }
f08f2dda
SH
1976 }
1977
1978 /* Fall back to bounce buffer if write zeroes is unsupported */
1979 iov.iov_len = nb_sectors * BDRV_SECTOR_SIZE;
1980 iov.iov_base = qemu_blockalign(bs, iov.iov_len);
1981 memset(iov.iov_base, 0, iov.iov_len);
1982 qemu_iovec_init_external(&qiov, &iov, 1);
1983
1984 ret = drv->bdrv_co_writev(bs, sector_num, nb_sectors, &qiov);
1985
1986 qemu_vfree(iov.iov_base);
1987 return ret;
1988}
1989
c5fbe571
SH
1990/*
1991 * Handle a write request in coroutine context
1992 */
1993static int coroutine_fn bdrv_co_do_writev(BlockDriverState *bs,
f08f2dda
SH
1994 int64_t sector_num, int nb_sectors, QEMUIOVector *qiov,
1995 BdrvRequestFlags flags)
c5fbe571
SH
1996{
1997 BlockDriver *drv = bs->drv;
dbffbdcf 1998 BdrvTrackedRequest req;
6b7cb247 1999 int ret;
da1fa91d
KW
2000
2001 if (!bs->drv) {
2002 return -ENOMEDIUM;
2003 }
2004 if (bs->read_only) {
2005 return -EACCES;
2006 }
2007 if (bdrv_check_request(bs, sector_num, nb_sectors)) {
2008 return -EIO;
2009 }
2010
98f90dba
ZYW
2011 /* throttling disk write I/O */
2012 if (bs->io_limits_enabled) {
2013 bdrv_io_limits_intercept(bs, true, nb_sectors);
2014 }
2015
470c0504 2016 if (bs->copy_on_read_in_flight) {
f4658285
SH
2017 wait_for_overlapping_requests(bs, sector_num, nb_sectors);
2018 }
2019
dbffbdcf
SH
2020 tracked_request_begin(&req, bs, sector_num, nb_sectors, true);
2021
f08f2dda
SH
2022 if (flags & BDRV_REQ_ZERO_WRITE) {
2023 ret = bdrv_co_do_write_zeroes(bs, sector_num, nb_sectors);
2024 } else {
2025 ret = drv->bdrv_co_writev(bs, sector_num, nb_sectors, qiov);
2026 }
6b7cb247 2027
f05fa4ad
PB
2028 if (ret == 0 && !bs->enable_write_cache) {
2029 ret = bdrv_co_flush(bs);
2030 }
2031
da1fa91d
KW
2032 if (bs->dirty_bitmap) {
2033 set_dirty_bitmap(bs, sector_num, nb_sectors, 1);
2034 }
2035
2036 if (bs->wr_highest_sector < sector_num + nb_sectors - 1) {
2037 bs->wr_highest_sector = sector_num + nb_sectors - 1;
2038 }
2039
dbffbdcf
SH
2040 tracked_request_end(&req);
2041
6b7cb247 2042 return ret;
da1fa91d
KW
2043}
2044
c5fbe571
SH
2045int coroutine_fn bdrv_co_writev(BlockDriverState *bs, int64_t sector_num,
2046 int nb_sectors, QEMUIOVector *qiov)
2047{
2048 trace_bdrv_co_writev(bs, sector_num, nb_sectors);
2049
f08f2dda
SH
2050 return bdrv_co_do_writev(bs, sector_num, nb_sectors, qiov, 0);
2051}
2052
2053int coroutine_fn bdrv_co_write_zeroes(BlockDriverState *bs,
2054 int64_t sector_num, int nb_sectors)
2055{
2056 trace_bdrv_co_write_zeroes(bs, sector_num, nb_sectors);
2057
2058 return bdrv_co_do_writev(bs, sector_num, nb_sectors, NULL,
2059 BDRV_REQ_ZERO_WRITE);
c5fbe571
SH
2060}
2061
83f64091
FB
2062/**
2063 * Truncate file to 'offset' bytes (needed only for file protocols)
2064 */
2065int bdrv_truncate(BlockDriverState *bs, int64_t offset)
2066{
2067 BlockDriver *drv = bs->drv;
51762288 2068 int ret;
83f64091 2069 if (!drv)
19cb3738 2070 return -ENOMEDIUM;
83f64091
FB
2071 if (!drv->bdrv_truncate)
2072 return -ENOTSUP;
59f2689d
NS
2073 if (bs->read_only)
2074 return -EACCES;
8591675f
MT
2075 if (bdrv_in_use(bs))
2076 return -EBUSY;
51762288
SH
2077 ret = drv->bdrv_truncate(bs, offset);
2078 if (ret == 0) {
2079 ret = refresh_total_sectors(bs, offset >> BDRV_SECTOR_BITS);
145feb17 2080 bdrv_dev_resize_cb(bs);
51762288
SH
2081 }
2082 return ret;
83f64091
FB
2083}
2084
4a1d5e1f
FZ
2085/**
2086 * Length of a allocated file in bytes. Sparse files are counted by actual
2087 * allocated space. Return < 0 if error or unknown.
2088 */
2089int64_t bdrv_get_allocated_file_size(BlockDriverState *bs)
2090{
2091 BlockDriver *drv = bs->drv;
2092 if (!drv) {
2093 return -ENOMEDIUM;
2094 }
2095 if (drv->bdrv_get_allocated_file_size) {
2096 return drv->bdrv_get_allocated_file_size(bs);
2097 }
2098 if (bs->file) {
2099 return bdrv_get_allocated_file_size(bs->file);
2100 }
2101 return -ENOTSUP;
2102}
2103
83f64091
FB
2104/**
2105 * Length of a file in bytes. Return < 0 if error or unknown.
2106 */
2107int64_t bdrv_getlength(BlockDriverState *bs)
2108{
2109 BlockDriver *drv = bs->drv;
2110 if (!drv)
19cb3738 2111 return -ENOMEDIUM;
51762288 2112
2c6942fa 2113 if (bs->growable || bdrv_dev_has_removable_media(bs)) {
46a4e4e6
SH
2114 if (drv->bdrv_getlength) {
2115 return drv->bdrv_getlength(bs);
2116 }
83f64091 2117 }
46a4e4e6 2118 return bs->total_sectors * BDRV_SECTOR_SIZE;
fc01f7e7
FB
2119}
2120
19cb3738 2121/* return 0 as number of sectors if no device present or error */
96b8f136 2122void bdrv_get_geometry(BlockDriverState *bs, uint64_t *nb_sectors_ptr)
fc01f7e7 2123{
19cb3738
FB
2124 int64_t length;
2125 length = bdrv_getlength(bs);
2126 if (length < 0)
2127 length = 0;
2128 else
6ea44308 2129 length = length >> BDRV_SECTOR_BITS;
19cb3738 2130 *nb_sectors_ptr = length;
fc01f7e7 2131}
cf98951b 2132
0563e191
ZYW
2133/* throttling disk io limits */
2134void bdrv_set_io_limits(BlockDriverState *bs,
2135 BlockIOLimit *io_limits)
2136{
2137 bs->io_limits = *io_limits;
2138 bs->io_limits_enabled = bdrv_io_limits_enabled(bs);
2139}
2140
abd7f68d
MA
2141void bdrv_set_on_error(BlockDriverState *bs, BlockErrorAction on_read_error,
2142 BlockErrorAction on_write_error)
2143{
2144 bs->on_read_error = on_read_error;
2145 bs->on_write_error = on_write_error;
2146}
2147
2148BlockErrorAction bdrv_get_on_error(BlockDriverState *bs, int is_read)
2149{
2150 return is_read ? bs->on_read_error : bs->on_write_error;
2151}
2152
b338082b
FB
2153int bdrv_is_read_only(BlockDriverState *bs)
2154{
2155 return bs->read_only;
2156}
2157
985a03b0
TS
2158int bdrv_is_sg(BlockDriverState *bs)
2159{
2160 return bs->sg;
2161}
2162
e900a7b7
CH
2163int bdrv_enable_write_cache(BlockDriverState *bs)
2164{
2165 return bs->enable_write_cache;
2166}
2167
425b0148
PB
2168void bdrv_set_enable_write_cache(BlockDriverState *bs, bool wce)
2169{
2170 bs->enable_write_cache = wce;
55b110f2
JC
2171
2172 /* so a reopen() will preserve wce */
2173 if (wce) {
2174 bs->open_flags |= BDRV_O_CACHE_WB;
2175 } else {
2176 bs->open_flags &= ~BDRV_O_CACHE_WB;
2177 }
425b0148
PB
2178}
2179
ea2384d3
FB
2180int bdrv_is_encrypted(BlockDriverState *bs)
2181{
2182 if (bs->backing_hd && bs->backing_hd->encrypted)
2183 return 1;
2184 return bs->encrypted;
2185}
2186
c0f4ce77
AL
2187int bdrv_key_required(BlockDriverState *bs)
2188{
2189 BlockDriverState *backing_hd = bs->backing_hd;
2190
2191 if (backing_hd && backing_hd->encrypted && !backing_hd->valid_key)
2192 return 1;
2193 return (bs->encrypted && !bs->valid_key);
2194}
2195
ea2384d3
FB
2196int bdrv_set_key(BlockDriverState *bs, const char *key)
2197{
2198 int ret;
2199 if (bs->backing_hd && bs->backing_hd->encrypted) {
2200 ret = bdrv_set_key(bs->backing_hd, key);
2201 if (ret < 0)
2202 return ret;
2203 if (!bs->encrypted)
2204 return 0;
2205 }
fd04a2ae
SH
2206 if (!bs->encrypted) {
2207 return -EINVAL;
2208 } else if (!bs->drv || !bs->drv->bdrv_set_key) {
2209 return -ENOMEDIUM;
2210 }
c0f4ce77 2211 ret = bs->drv->bdrv_set_key(bs, key);
bb5fc20f
AL
2212 if (ret < 0) {
2213 bs->valid_key = 0;
2214 } else if (!bs->valid_key) {
2215 bs->valid_key = 1;
2216 /* call the change callback now, we skipped it on open */
7d4b4ba5 2217 bdrv_dev_change_media_cb(bs, true);
bb5fc20f 2218 }
c0f4ce77 2219 return ret;
ea2384d3
FB
2220}
2221
f8d6bba1 2222const char *bdrv_get_format_name(BlockDriverState *bs)
ea2384d3 2223{
f8d6bba1 2224 return bs->drv ? bs->drv->format_name : NULL;
ea2384d3
FB
2225}
2226
5fafdf24 2227void bdrv_iterate_format(void (*it)(void *opaque, const char *name),
ea2384d3
FB
2228 void *opaque)
2229{
2230 BlockDriver *drv;
2231
8a22f02a 2232 QLIST_FOREACH(drv, &bdrv_drivers, list) {
ea2384d3
FB
2233 it(opaque, drv->format_name);
2234 }
2235}
2236
b338082b
FB
2237BlockDriverState *bdrv_find(const char *name)
2238{
2239 BlockDriverState *bs;
2240
1b7bdbc1
SH
2241 QTAILQ_FOREACH(bs, &bdrv_states, list) {
2242 if (!strcmp(name, bs->device_name)) {
b338082b 2243 return bs;
1b7bdbc1 2244 }
b338082b
FB
2245 }
2246 return NULL;
2247}
2248
2f399b0a
MA
2249BlockDriverState *bdrv_next(BlockDriverState *bs)
2250{
2251 if (!bs) {
2252 return QTAILQ_FIRST(&bdrv_states);
2253 }
2254 return QTAILQ_NEXT(bs, list);
2255}
2256
51de9760 2257void bdrv_iterate(void (*it)(void *opaque, BlockDriverState *bs), void *opaque)
81d0912d
FB
2258{
2259 BlockDriverState *bs;
2260
1b7bdbc1 2261 QTAILQ_FOREACH(bs, &bdrv_states, list) {
51de9760 2262 it(opaque, bs);
81d0912d
FB
2263 }
2264}
2265
ea2384d3
FB
2266const char *bdrv_get_device_name(BlockDriverState *bs)
2267{
2268 return bs->device_name;
2269}
2270
c8433287
MA
2271int bdrv_get_flags(BlockDriverState *bs)
2272{
2273 return bs->open_flags;
2274}
2275
c6ca28d6
AL
2276void bdrv_flush_all(void)
2277{
2278 BlockDriverState *bs;
2279
1b7bdbc1 2280 QTAILQ_FOREACH(bs, &bdrv_states, list) {
29cdb251 2281 bdrv_flush(bs);
1b7bdbc1 2282 }
c6ca28d6
AL
2283}
2284
f2feebbd
KW
2285int bdrv_has_zero_init(BlockDriverState *bs)
2286{
2287 assert(bs->drv);
2288
336c1c12
KW
2289 if (bs->drv->bdrv_has_zero_init) {
2290 return bs->drv->bdrv_has_zero_init(bs);
f2feebbd
KW
2291 }
2292
2293 return 1;
2294}
2295
376ae3f1
SH
2296typedef struct BdrvCoIsAllocatedData {
2297 BlockDriverState *bs;
2298 int64_t sector_num;
2299 int nb_sectors;
2300 int *pnum;
2301 int ret;
2302 bool done;
2303} BdrvCoIsAllocatedData;
2304
f58c7b35
TS
2305/*
2306 * Returns true iff the specified sector is present in the disk image. Drivers
2307 * not implementing the functionality are assumed to not support backing files,
2308 * hence all their sectors are reported as allocated.
2309 *
bd9533e3
SH
2310 * If 'sector_num' is beyond the end of the disk image the return value is 0
2311 * and 'pnum' is set to 0.
2312 *
f58c7b35
TS
2313 * 'pnum' is set to the number of sectors (including and immediately following
2314 * the specified sector) that are known to be in the same
2315 * allocated/unallocated state.
2316 *
bd9533e3
SH
2317 * 'nb_sectors' is the max value 'pnum' should be set to. If nb_sectors goes
2318 * beyond the end of the disk image it will be clamped.
f58c7b35 2319 */
060f51c9
SH
2320int coroutine_fn bdrv_co_is_allocated(BlockDriverState *bs, int64_t sector_num,
2321 int nb_sectors, int *pnum)
f58c7b35 2322{
bd9533e3
SH
2323 int64_t n;
2324
2325 if (sector_num >= bs->total_sectors) {
2326 *pnum = 0;
2327 return 0;
2328 }
2329
2330 n = bs->total_sectors - sector_num;
2331 if (n < nb_sectors) {
2332 nb_sectors = n;
2333 }
2334
6aebab14 2335 if (!bs->drv->bdrv_co_is_allocated) {
bd9533e3 2336 *pnum = nb_sectors;
f58c7b35
TS
2337 return 1;
2338 }
6aebab14 2339
060f51c9
SH
2340 return bs->drv->bdrv_co_is_allocated(bs, sector_num, nb_sectors, pnum);
2341}
2342
2343/* Coroutine wrapper for bdrv_is_allocated() */
2344static void coroutine_fn bdrv_is_allocated_co_entry(void *opaque)
2345{
2346 BdrvCoIsAllocatedData *data = opaque;
2347 BlockDriverState *bs = data->bs;
2348
2349 data->ret = bdrv_co_is_allocated(bs, data->sector_num, data->nb_sectors,
2350 data->pnum);
2351 data->done = true;
2352}
2353
2354/*
2355 * Synchronous wrapper around bdrv_co_is_allocated().
2356 *
2357 * See bdrv_co_is_allocated() for details.
2358 */
2359int bdrv_is_allocated(BlockDriverState *bs, int64_t sector_num, int nb_sectors,
2360 int *pnum)
2361{
6aebab14
SH
2362 Coroutine *co;
2363 BdrvCoIsAllocatedData data = {
2364 .bs = bs,
2365 .sector_num = sector_num,
2366 .nb_sectors = nb_sectors,
2367 .pnum = pnum,
2368 .done = false,
2369 };
2370
2371 co = qemu_coroutine_create(bdrv_is_allocated_co_entry);
2372 qemu_coroutine_enter(co, &data);
2373 while (!data.done) {
2374 qemu_aio_wait();
2375 }
2376 return data.ret;
f58c7b35
TS
2377}
2378
188a7bbf
PB
2379/*
2380 * Given an image chain: ... -> [BASE] -> [INTER1] -> [INTER2] -> [TOP]
2381 *
2382 * Return true if the given sector is allocated in any image between
2383 * BASE and TOP (inclusive). BASE can be NULL to check if the given
2384 * sector is allocated in any image of the chain. Return false otherwise.
2385 *
2386 * 'pnum' is set to the number of sectors (including and immediately following
2387 * the specified sector) that are known to be in the same
2388 * allocated/unallocated state.
2389 *
2390 */
2391int coroutine_fn bdrv_co_is_allocated_above(BlockDriverState *top,
2392 BlockDriverState *base,
2393 int64_t sector_num,
2394 int nb_sectors, int *pnum)
2395{
2396 BlockDriverState *intermediate;
2397 int ret, n = nb_sectors;
2398
2399 intermediate = top;
2400 while (intermediate && intermediate != base) {
2401 int pnum_inter;
2402 ret = bdrv_co_is_allocated(intermediate, sector_num, nb_sectors,
2403 &pnum_inter);
2404 if (ret < 0) {
2405 return ret;
2406 } else if (ret) {
2407 *pnum = pnum_inter;
2408 return 1;
2409 }
2410
2411 /*
2412 * [sector_num, nb_sectors] is unallocated on top but intermediate
2413 * might have
2414 *
2415 * [sector_num+x, nr_sectors] allocated.
2416 */
2417 if (n > pnum_inter) {
2418 n = pnum_inter;
2419 }
2420
2421 intermediate = intermediate->backing_hd;
2422 }
2423
2424 *pnum = n;
2425 return 0;
2426}
2427
b2023818 2428BlockInfoList *qmp_query_block(Error **errp)
b338082b 2429{
b2023818 2430 BlockInfoList *head = NULL, *cur_item = NULL;
b338082b
FB
2431 BlockDriverState *bs;
2432
1b7bdbc1 2433 QTAILQ_FOREACH(bs, &bdrv_states, list) {
b2023818 2434 BlockInfoList *info = g_malloc0(sizeof(*info));
d15e5465 2435
b2023818
LC
2436 info->value = g_malloc0(sizeof(*info->value));
2437 info->value->device = g_strdup(bs->device_name);
2438 info->value->type = g_strdup("unknown");
2439 info->value->locked = bdrv_dev_is_medium_locked(bs);
2440 info->value->removable = bdrv_dev_has_removable_media(bs);
d15e5465 2441
e4def80b 2442 if (bdrv_dev_has_removable_media(bs)) {
b2023818
LC
2443 info->value->has_tray_open = true;
2444 info->value->tray_open = bdrv_dev_is_tray_open(bs);
e4def80b 2445 }
f04ef601
LC
2446
2447 if (bdrv_iostatus_is_enabled(bs)) {
b2023818
LC
2448 info->value->has_io_status = true;
2449 info->value->io_status = bs->iostatus;
f04ef601
LC
2450 }
2451
19cb3738 2452 if (bs->drv) {
b2023818
LC
2453 info->value->has_inserted = true;
2454 info->value->inserted = g_malloc0(sizeof(*info->value->inserted));
2455 info->value->inserted->file = g_strdup(bs->filename);
2456 info->value->inserted->ro = bs->read_only;
2457 info->value->inserted->drv = g_strdup(bs->drv->format_name);
2458 info->value->inserted->encrypted = bs->encrypted;
c75a1a8a 2459 info->value->inserted->encryption_key_missing = bdrv_key_required(bs);
b2023818
LC
2460 if (bs->backing_file[0]) {
2461 info->value->inserted->has_backing_file = true;
2462 info->value->inserted->backing_file = g_strdup(bs->backing_file);
376253ec 2463 }
727f005e 2464
2e3e3317
BC
2465 info->value->inserted->backing_file_depth =
2466 bdrv_get_backing_file_depth(bs);
2467
727f005e
ZYW
2468 if (bs->io_limits_enabled) {
2469 info->value->inserted->bps =
2470 bs->io_limits.bps[BLOCK_IO_LIMIT_TOTAL];
2471 info->value->inserted->bps_rd =
2472 bs->io_limits.bps[BLOCK_IO_LIMIT_READ];
2473 info->value->inserted->bps_wr =
2474 bs->io_limits.bps[BLOCK_IO_LIMIT_WRITE];
2475 info->value->inserted->iops =
2476 bs->io_limits.iops[BLOCK_IO_LIMIT_TOTAL];
2477 info->value->inserted->iops_rd =
2478 bs->io_limits.iops[BLOCK_IO_LIMIT_READ];
2479 info->value->inserted->iops_wr =
2480 bs->io_limits.iops[BLOCK_IO_LIMIT_WRITE];
2481 }
b2023818 2482 }
d15e5465 2483
b2023818
LC
2484 /* XXX: waiting for the qapi to support GSList */
2485 if (!cur_item) {
2486 head = cur_item = info;
2487 } else {
2488 cur_item->next = info;
2489 cur_item = info;
b338082b 2490 }
b338082b 2491 }
d15e5465 2492
b2023818 2493 return head;
b338082b 2494}
a36e69dd 2495
f11f57e4
LC
2496/* Consider exposing this as a full fledged QMP command */
2497static BlockStats *qmp_query_blockstat(const BlockDriverState *bs, Error **errp)
2498{
2499 BlockStats *s;
2500
2501 s = g_malloc0(sizeof(*s));
2502
2503 if (bs->device_name[0]) {
2504 s->has_device = true;
2505 s->device = g_strdup(bs->device_name);
294cc35f
KW
2506 }
2507
f11f57e4
LC
2508 s->stats = g_malloc0(sizeof(*s->stats));
2509 s->stats->rd_bytes = bs->nr_bytes[BDRV_ACCT_READ];
2510 s->stats->wr_bytes = bs->nr_bytes[BDRV_ACCT_WRITE];
2511 s->stats->rd_operations = bs->nr_ops[BDRV_ACCT_READ];
2512 s->stats->wr_operations = bs->nr_ops[BDRV_ACCT_WRITE];
2513 s->stats->wr_highest_offset = bs->wr_highest_sector * BDRV_SECTOR_SIZE;
2514 s->stats->flush_operations = bs->nr_ops[BDRV_ACCT_FLUSH];
2515 s->stats->wr_total_time_ns = bs->total_time_ns[BDRV_ACCT_WRITE];
2516 s->stats->rd_total_time_ns = bs->total_time_ns[BDRV_ACCT_READ];
2517 s->stats->flush_total_time_ns = bs->total_time_ns[BDRV_ACCT_FLUSH];
2518
294cc35f 2519 if (bs->file) {
f11f57e4
LC
2520 s->has_parent = true;
2521 s->parent = qmp_query_blockstat(bs->file, NULL);
294cc35f
KW
2522 }
2523
f11f57e4 2524 return s;
294cc35f
KW
2525}
2526
f11f57e4 2527BlockStatsList *qmp_query_blockstats(Error **errp)
218a536a 2528{
f11f57e4 2529 BlockStatsList *head = NULL, *cur_item = NULL;
a36e69dd
TS
2530 BlockDriverState *bs;
2531
1b7bdbc1 2532 QTAILQ_FOREACH(bs, &bdrv_states, list) {
f11f57e4
LC
2533 BlockStatsList *info = g_malloc0(sizeof(*info));
2534 info->value = qmp_query_blockstat(bs, NULL);
2535
2536 /* XXX: waiting for the qapi to support GSList */
2537 if (!cur_item) {
2538 head = cur_item = info;
2539 } else {
2540 cur_item->next = info;
2541 cur_item = info;
2542 }
a36e69dd 2543 }
218a536a 2544
f11f57e4 2545 return head;
a36e69dd 2546}
ea2384d3 2547
045df330
AL
2548const char *bdrv_get_encrypted_filename(BlockDriverState *bs)
2549{
2550 if (bs->backing_hd && bs->backing_hd->encrypted)
2551 return bs->backing_file;
2552 else if (bs->encrypted)
2553 return bs->filename;
2554 else
2555 return NULL;
2556}
2557
5fafdf24 2558void bdrv_get_backing_filename(BlockDriverState *bs,
83f64091
FB
2559 char *filename, int filename_size)
2560{
3574c608 2561 pstrcpy(filename, filename_size, bs->backing_file);
83f64091
FB
2562}
2563
5fafdf24 2564int bdrv_write_compressed(BlockDriverState *bs, int64_t sector_num,
faea38e7
FB
2565 const uint8_t *buf, int nb_sectors)
2566{
2567 BlockDriver *drv = bs->drv;
2568 if (!drv)
19cb3738 2569 return -ENOMEDIUM;
faea38e7
FB
2570 if (!drv->bdrv_write_compressed)
2571 return -ENOTSUP;
fbb7b4e0
KW
2572 if (bdrv_check_request(bs, sector_num, nb_sectors))
2573 return -EIO;
a55eb92c 2574
c6d22830 2575 if (bs->dirty_bitmap) {
7cd1e32a 2576 set_dirty_bitmap(bs, sector_num, nb_sectors, 1);
2577 }
a55eb92c 2578
faea38e7
FB
2579 return drv->bdrv_write_compressed(bs, sector_num, buf, nb_sectors);
2580}
3b46e624 2581
faea38e7
FB
2582int bdrv_get_info(BlockDriverState *bs, BlockDriverInfo *bdi)
2583{
2584 BlockDriver *drv = bs->drv;
2585 if (!drv)
19cb3738 2586 return -ENOMEDIUM;
faea38e7
FB
2587 if (!drv->bdrv_get_info)
2588 return -ENOTSUP;
2589 memset(bdi, 0, sizeof(*bdi));
2590 return drv->bdrv_get_info(bs, bdi);
2591}
2592
45566e9c
CH
2593int bdrv_save_vmstate(BlockDriverState *bs, const uint8_t *buf,
2594 int64_t pos, int size)
178e08a5
AL
2595{
2596 BlockDriver *drv = bs->drv;
2597 if (!drv)
2598 return -ENOMEDIUM;
7cdb1f6d
MK
2599 if (drv->bdrv_save_vmstate)
2600 return drv->bdrv_save_vmstate(bs, buf, pos, size);
2601 if (bs->file)
2602 return bdrv_save_vmstate(bs->file, buf, pos, size);
2603 return -ENOTSUP;
178e08a5
AL
2604}
2605
45566e9c
CH
2606int bdrv_load_vmstate(BlockDriverState *bs, uint8_t *buf,
2607 int64_t pos, int size)
178e08a5
AL
2608{
2609 BlockDriver *drv = bs->drv;
2610 if (!drv)
2611 return -ENOMEDIUM;
7cdb1f6d
MK
2612 if (drv->bdrv_load_vmstate)
2613 return drv->bdrv_load_vmstate(bs, buf, pos, size);
2614 if (bs->file)
2615 return bdrv_load_vmstate(bs->file, buf, pos, size);
2616 return -ENOTSUP;
178e08a5
AL
2617}
2618
8b9b0cc2
KW
2619void bdrv_debug_event(BlockDriverState *bs, BlkDebugEvent event)
2620{
2621 BlockDriver *drv = bs->drv;
2622
2623 if (!drv || !drv->bdrv_debug_event) {
2624 return;
2625 }
2626
0ed8b6f6 2627 drv->bdrv_debug_event(bs, event);
8b9b0cc2
KW
2628
2629}
2630
faea38e7
FB
2631/**************************************************************/
2632/* handling of snapshots */
2633
feeee5ac
MDCF
2634int bdrv_can_snapshot(BlockDriverState *bs)
2635{
2636 BlockDriver *drv = bs->drv;
07b70bfb 2637 if (!drv || !bdrv_is_inserted(bs) || bdrv_is_read_only(bs)) {
feeee5ac
MDCF
2638 return 0;
2639 }
2640
2641 if (!drv->bdrv_snapshot_create) {
2642 if (bs->file != NULL) {
2643 return bdrv_can_snapshot(bs->file);
2644 }
2645 return 0;
2646 }
2647
2648 return 1;
2649}
2650
199630b6
BS
2651int bdrv_is_snapshot(BlockDriverState *bs)
2652{
2653 return !!(bs->open_flags & BDRV_O_SNAPSHOT);
2654}
2655
f9092b10
MA
2656BlockDriverState *bdrv_snapshots(void)
2657{
2658 BlockDriverState *bs;
2659
3ac906f7 2660 if (bs_snapshots) {
f9092b10 2661 return bs_snapshots;
3ac906f7 2662 }
f9092b10
MA
2663
2664 bs = NULL;
2665 while ((bs = bdrv_next(bs))) {
2666 if (bdrv_can_snapshot(bs)) {
3ac906f7
MA
2667 bs_snapshots = bs;
2668 return bs;
f9092b10
MA
2669 }
2670 }
2671 return NULL;
f9092b10
MA
2672}
2673
5fafdf24 2674int bdrv_snapshot_create(BlockDriverState *bs,
faea38e7
FB
2675 QEMUSnapshotInfo *sn_info)
2676{
2677 BlockDriver *drv = bs->drv;
2678 if (!drv)
19cb3738 2679 return -ENOMEDIUM;
7cdb1f6d
MK
2680 if (drv->bdrv_snapshot_create)
2681 return drv->bdrv_snapshot_create(bs, sn_info);
2682 if (bs->file)
2683 return bdrv_snapshot_create(bs->file, sn_info);
2684 return -ENOTSUP;
faea38e7
FB
2685}
2686
5fafdf24 2687int bdrv_snapshot_goto(BlockDriverState *bs,
faea38e7
FB
2688 const char *snapshot_id)
2689{
2690 BlockDriver *drv = bs->drv;
7cdb1f6d
MK
2691 int ret, open_ret;
2692
faea38e7 2693 if (!drv)
19cb3738 2694 return -ENOMEDIUM;
7cdb1f6d
MK
2695 if (drv->bdrv_snapshot_goto)
2696 return drv->bdrv_snapshot_goto(bs, snapshot_id);
2697
2698 if (bs->file) {
2699 drv->bdrv_close(bs);
2700 ret = bdrv_snapshot_goto(bs->file, snapshot_id);
2701 open_ret = drv->bdrv_open(bs, bs->open_flags);
2702 if (open_ret < 0) {
2703 bdrv_delete(bs->file);
2704 bs->drv = NULL;
2705 return open_ret;
2706 }
2707 return ret;
2708 }
2709
2710 return -ENOTSUP;
faea38e7
FB
2711}
2712
2713int bdrv_snapshot_delete(BlockDriverState *bs, const char *snapshot_id)
2714{
2715 BlockDriver *drv = bs->drv;
2716 if (!drv)
19cb3738 2717 return -ENOMEDIUM;
7cdb1f6d
MK
2718 if (drv->bdrv_snapshot_delete)
2719 return drv->bdrv_snapshot_delete(bs, snapshot_id);
2720 if (bs->file)
2721 return bdrv_snapshot_delete(bs->file, snapshot_id);
2722 return -ENOTSUP;
faea38e7
FB
2723}
2724
5fafdf24 2725int bdrv_snapshot_list(BlockDriverState *bs,
faea38e7
FB
2726 QEMUSnapshotInfo **psn_info)
2727{
2728 BlockDriver *drv = bs->drv;
2729 if (!drv)
19cb3738 2730 return -ENOMEDIUM;
7cdb1f6d
MK
2731 if (drv->bdrv_snapshot_list)
2732 return drv->bdrv_snapshot_list(bs, psn_info);
2733 if (bs->file)
2734 return bdrv_snapshot_list(bs->file, psn_info);
2735 return -ENOTSUP;
faea38e7
FB
2736}
2737
51ef6727 2738int bdrv_snapshot_load_tmp(BlockDriverState *bs,
2739 const char *snapshot_name)
2740{
2741 BlockDriver *drv = bs->drv;
2742 if (!drv) {
2743 return -ENOMEDIUM;
2744 }
2745 if (!bs->read_only) {
2746 return -EINVAL;
2747 }
2748 if (drv->bdrv_snapshot_load_tmp) {
2749 return drv->bdrv_snapshot_load_tmp(bs, snapshot_name);
2750 }
2751 return -ENOTSUP;
2752}
2753
e8a6bb9c
MT
2754BlockDriverState *bdrv_find_backing_image(BlockDriverState *bs,
2755 const char *backing_file)
2756{
2757 if (!bs->drv) {
2758 return NULL;
2759 }
2760
2761 if (bs->backing_hd) {
2762 if (strcmp(bs->backing_file, backing_file) == 0) {
2763 return bs->backing_hd;
2764 } else {
2765 return bdrv_find_backing_image(bs->backing_hd, backing_file);
2766 }
2767 }
2768
2769 return NULL;
2770}
2771
f198fd1c
BC
2772int bdrv_get_backing_file_depth(BlockDriverState *bs)
2773{
2774 if (!bs->drv) {
2775 return 0;
2776 }
2777
2778 if (!bs->backing_hd) {
2779 return 0;
2780 }
2781
2782 return 1 + bdrv_get_backing_file_depth(bs->backing_hd);
2783}
2784
faea38e7
FB
2785#define NB_SUFFIXES 4
2786
2787char *get_human_readable_size(char *buf, int buf_size, int64_t size)
2788{
2789 static const char suffixes[NB_SUFFIXES] = "KMGT";
2790 int64_t base;
2791 int i;
2792
2793 if (size <= 999) {
2794 snprintf(buf, buf_size, "%" PRId64, size);
2795 } else {
2796 base = 1024;
2797 for(i = 0; i < NB_SUFFIXES; i++) {
2798 if (size < (10 * base)) {
5fafdf24 2799 snprintf(buf, buf_size, "%0.1f%c",
faea38e7
FB
2800 (double)size / base,
2801 suffixes[i]);
2802 break;
2803 } else if (size < (1000 * base) || i == (NB_SUFFIXES - 1)) {
5fafdf24 2804 snprintf(buf, buf_size, "%" PRId64 "%c",
faea38e7
FB
2805 ((size + (base >> 1)) / base),
2806 suffixes[i]);
2807 break;
2808 }
2809 base = base * 1024;
2810 }
2811 }
2812 return buf;
2813}
2814
2815char *bdrv_snapshot_dump(char *buf, int buf_size, QEMUSnapshotInfo *sn)
2816{
2817 char buf1[128], date_buf[128], clock_buf[128];
3b9f94e1
FB
2818#ifdef _WIN32
2819 struct tm *ptm;
2820#else
faea38e7 2821 struct tm tm;
3b9f94e1 2822#endif
faea38e7
FB
2823 time_t ti;
2824 int64_t secs;
2825
2826 if (!sn) {
5fafdf24
TS
2827 snprintf(buf, buf_size,
2828 "%-10s%-20s%7s%20s%15s",
faea38e7
FB
2829 "ID", "TAG", "VM SIZE", "DATE", "VM CLOCK");
2830 } else {
2831 ti = sn->date_sec;
3b9f94e1
FB
2832#ifdef _WIN32
2833 ptm = localtime(&ti);
2834 strftime(date_buf, sizeof(date_buf),
2835 "%Y-%m-%d %H:%M:%S", ptm);
2836#else
faea38e7
FB
2837 localtime_r(&ti, &tm);
2838 strftime(date_buf, sizeof(date_buf),
2839 "%Y-%m-%d %H:%M:%S", &tm);
3b9f94e1 2840#endif
faea38e7
FB
2841 secs = sn->vm_clock_nsec / 1000000000;
2842 snprintf(clock_buf, sizeof(clock_buf),
2843 "%02d:%02d:%02d.%03d",
2844 (int)(secs / 3600),
2845 (int)((secs / 60) % 60),
5fafdf24 2846 (int)(secs % 60),
faea38e7
FB
2847 (int)((sn->vm_clock_nsec / 1000000) % 1000));
2848 snprintf(buf, buf_size,
5fafdf24 2849 "%-10s%-20s%7s%20s%15s",
faea38e7
FB
2850 sn->id_str, sn->name,
2851 get_human_readable_size(buf1, sizeof(buf1), sn->vm_state_size),
2852 date_buf,
2853 clock_buf);
2854 }
2855 return buf;
2856}
2857
ea2384d3 2858/**************************************************************/
83f64091 2859/* async I/Os */
ea2384d3 2860
3b69e4b9 2861BlockDriverAIOCB *bdrv_aio_readv(BlockDriverState *bs, int64_t sector_num,
f141eafe 2862 QEMUIOVector *qiov, int nb_sectors,
3b69e4b9 2863 BlockDriverCompletionFunc *cb, void *opaque)
83f64091 2864{
bbf0a440
SH
2865 trace_bdrv_aio_readv(bs, sector_num, nb_sectors, opaque);
2866
b2a61371 2867 return bdrv_co_aio_rw_vector(bs, sector_num, qiov, nb_sectors,
8c5873d6 2868 cb, opaque, false);
ea2384d3
FB
2869}
2870
f141eafe
AL
2871BlockDriverAIOCB *bdrv_aio_writev(BlockDriverState *bs, int64_t sector_num,
2872 QEMUIOVector *qiov, int nb_sectors,
2873 BlockDriverCompletionFunc *cb, void *opaque)
ea2384d3 2874{
bbf0a440
SH
2875 trace_bdrv_aio_writev(bs, sector_num, nb_sectors, opaque);
2876
1a6e115b 2877 return bdrv_co_aio_rw_vector(bs, sector_num, qiov, nb_sectors,
8c5873d6 2878 cb, opaque, true);
83f64091
FB
2879}
2880
40b4f539
KW
2881
2882typedef struct MultiwriteCB {
2883 int error;
2884 int num_requests;
2885 int num_callbacks;
2886 struct {
2887 BlockDriverCompletionFunc *cb;
2888 void *opaque;
2889 QEMUIOVector *free_qiov;
40b4f539
KW
2890 } callbacks[];
2891} MultiwriteCB;
2892
2893static void multiwrite_user_cb(MultiwriteCB *mcb)
2894{
2895 int i;
2896
2897 for (i = 0; i < mcb->num_callbacks; i++) {
2898 mcb->callbacks[i].cb(mcb->callbacks[i].opaque, mcb->error);
1e1ea48d
SH
2899 if (mcb->callbacks[i].free_qiov) {
2900 qemu_iovec_destroy(mcb->callbacks[i].free_qiov);
2901 }
7267c094 2902 g_free(mcb->callbacks[i].free_qiov);
40b4f539
KW
2903 }
2904}
2905
2906static void multiwrite_cb(void *opaque, int ret)
2907{
2908 MultiwriteCB *mcb = opaque;
2909
6d519a5f
SH
2910 trace_multiwrite_cb(mcb, ret);
2911
cb6d3ca0 2912 if (ret < 0 && !mcb->error) {
40b4f539 2913 mcb->error = ret;
40b4f539
KW
2914 }
2915
2916 mcb->num_requests--;
2917 if (mcb->num_requests == 0) {
de189a1b 2918 multiwrite_user_cb(mcb);
7267c094 2919 g_free(mcb);
40b4f539
KW
2920 }
2921}
2922
2923static int multiwrite_req_compare(const void *a, const void *b)
2924{
77be4366
CH
2925 const BlockRequest *req1 = a, *req2 = b;
2926
2927 /*
2928 * Note that we can't simply subtract req2->sector from req1->sector
2929 * here as that could overflow the return value.
2930 */
2931 if (req1->sector > req2->sector) {
2932 return 1;
2933 } else if (req1->sector < req2->sector) {
2934 return -1;
2935 } else {
2936 return 0;
2937 }
40b4f539
KW
2938}
2939
2940/*
2941 * Takes a bunch of requests and tries to merge them. Returns the number of
2942 * requests that remain after merging.
2943 */
2944static int multiwrite_merge(BlockDriverState *bs, BlockRequest *reqs,
2945 int num_reqs, MultiwriteCB *mcb)
2946{
2947 int i, outidx;
2948
2949 // Sort requests by start sector
2950 qsort(reqs, num_reqs, sizeof(*reqs), &multiwrite_req_compare);
2951
2952 // Check if adjacent requests touch the same clusters. If so, combine them,
2953 // filling up gaps with zero sectors.
2954 outidx = 0;
2955 for (i = 1; i < num_reqs; i++) {
2956 int merge = 0;
2957 int64_t oldreq_last = reqs[outidx].sector + reqs[outidx].nb_sectors;
2958
b6a127a1 2959 // Handle exactly sequential writes and overlapping writes.
40b4f539
KW
2960 if (reqs[i].sector <= oldreq_last) {
2961 merge = 1;
2962 }
2963
e2a305fb
CH
2964 if (reqs[outidx].qiov->niov + reqs[i].qiov->niov + 1 > IOV_MAX) {
2965 merge = 0;
2966 }
2967
40b4f539
KW
2968 if (merge) {
2969 size_t size;
7267c094 2970 QEMUIOVector *qiov = g_malloc0(sizeof(*qiov));
40b4f539
KW
2971 qemu_iovec_init(qiov,
2972 reqs[outidx].qiov->niov + reqs[i].qiov->niov + 1);
2973
2974 // Add the first request to the merged one. If the requests are
2975 // overlapping, drop the last sectors of the first request.
2976 size = (reqs[i].sector - reqs[outidx].sector) << 9;
1b093c48 2977 qemu_iovec_concat(qiov, reqs[outidx].qiov, 0, size);
40b4f539 2978
b6a127a1
PB
2979 // We should need to add any zeros between the two requests
2980 assert (reqs[i].sector <= oldreq_last);
40b4f539
KW
2981
2982 // Add the second request
1b093c48 2983 qemu_iovec_concat(qiov, reqs[i].qiov, 0, reqs[i].qiov->size);
40b4f539 2984
cbf1dff2 2985 reqs[outidx].nb_sectors = qiov->size >> 9;
40b4f539
KW
2986 reqs[outidx].qiov = qiov;
2987
2988 mcb->callbacks[i].free_qiov = reqs[outidx].qiov;
2989 } else {
2990 outidx++;
2991 reqs[outidx].sector = reqs[i].sector;
2992 reqs[outidx].nb_sectors = reqs[i].nb_sectors;
2993 reqs[outidx].qiov = reqs[i].qiov;
2994 }
2995 }
2996
2997 return outidx + 1;
2998}
2999
3000/*
3001 * Submit multiple AIO write requests at once.
3002 *
3003 * On success, the function returns 0 and all requests in the reqs array have
3004 * been submitted. In error case this function returns -1, and any of the
3005 * requests may or may not be submitted yet. In particular, this means that the
3006 * callback will be called for some of the requests, for others it won't. The
3007 * caller must check the error field of the BlockRequest to wait for the right
3008 * callbacks (if error != 0, no callback will be called).
3009 *
3010 * The implementation may modify the contents of the reqs array, e.g. to merge
3011 * requests. However, the fields opaque and error are left unmodified as they
3012 * are used to signal failure for a single request to the caller.
3013 */
3014int bdrv_aio_multiwrite(BlockDriverState *bs, BlockRequest *reqs, int num_reqs)
3015{
40b4f539
KW
3016 MultiwriteCB *mcb;
3017 int i;
3018
301db7c2
RH
3019 /* don't submit writes if we don't have a medium */
3020 if (bs->drv == NULL) {
3021 for (i = 0; i < num_reqs; i++) {
3022 reqs[i].error = -ENOMEDIUM;
3023 }
3024 return -1;
3025 }
3026
40b4f539
KW
3027 if (num_reqs == 0) {
3028 return 0;
3029 }
3030
3031 // Create MultiwriteCB structure
7267c094 3032 mcb = g_malloc0(sizeof(*mcb) + num_reqs * sizeof(*mcb->callbacks));
40b4f539
KW
3033 mcb->num_requests = 0;
3034 mcb->num_callbacks = num_reqs;
3035
3036 for (i = 0; i < num_reqs; i++) {
3037 mcb->callbacks[i].cb = reqs[i].cb;
3038 mcb->callbacks[i].opaque = reqs[i].opaque;
3039 }
3040
3041 // Check for mergable requests
3042 num_reqs = multiwrite_merge(bs, reqs, num_reqs, mcb);
3043
6d519a5f
SH
3044 trace_bdrv_aio_multiwrite(mcb, mcb->num_callbacks, num_reqs);
3045
df9309fb
PB
3046 /* Run the aio requests. */
3047 mcb->num_requests = num_reqs;
40b4f539 3048 for (i = 0; i < num_reqs; i++) {
ad54ae80 3049 bdrv_aio_writev(bs, reqs[i].sector, reqs[i].qiov,
40b4f539 3050 reqs[i].nb_sectors, multiwrite_cb, mcb);
40b4f539
KW
3051 }
3052
3053 return 0;
40b4f539
KW
3054}
3055
83f64091 3056void bdrv_aio_cancel(BlockDriverAIOCB *acb)
83f64091 3057{
6bbff9a0 3058 acb->pool->cancel(acb);
83f64091
FB
3059}
3060
98f90dba
ZYW
3061/* block I/O throttling */
3062static bool bdrv_exceed_bps_limits(BlockDriverState *bs, int nb_sectors,
3063 bool is_write, double elapsed_time, uint64_t *wait)
3064{
3065 uint64_t bps_limit = 0;
3066 double bytes_limit, bytes_base, bytes_res;
3067 double slice_time, wait_time;
3068
3069 if (bs->io_limits.bps[BLOCK_IO_LIMIT_TOTAL]) {
3070 bps_limit = bs->io_limits.bps[BLOCK_IO_LIMIT_TOTAL];
3071 } else if (bs->io_limits.bps[is_write]) {
3072 bps_limit = bs->io_limits.bps[is_write];
3073 } else {
3074 if (wait) {
3075 *wait = 0;
3076 }
3077
3078 return false;
3079 }
3080
3081 slice_time = bs->slice_end - bs->slice_start;
3082 slice_time /= (NANOSECONDS_PER_SECOND);
3083 bytes_limit = bps_limit * slice_time;
3084 bytes_base = bs->nr_bytes[is_write] - bs->io_base.bytes[is_write];
3085 if (bs->io_limits.bps[BLOCK_IO_LIMIT_TOTAL]) {
3086 bytes_base += bs->nr_bytes[!is_write] - bs->io_base.bytes[!is_write];
3087 }
3088
3089 /* bytes_base: the bytes of data which have been read/written; and
3090 * it is obtained from the history statistic info.
3091 * bytes_res: the remaining bytes of data which need to be read/written.
3092 * (bytes_base + bytes_res) / bps_limit: used to calcuate
3093 * the total time for completing reading/writting all data.
3094 */
3095 bytes_res = (unsigned) nb_sectors * BDRV_SECTOR_SIZE;
3096
3097 if (bytes_base + bytes_res <= bytes_limit) {
3098 if (wait) {
3099 *wait = 0;
3100 }
3101
3102 return false;
3103 }
3104
3105 /* Calc approx time to dispatch */
3106 wait_time = (bytes_base + bytes_res) / bps_limit - elapsed_time;
3107
3108 /* When the I/O rate at runtime exceeds the limits,
3109 * bs->slice_end need to be extended in order that the current statistic
3110 * info can be kept until the timer fire, so it is increased and tuned
3111 * based on the result of experiment.
3112 */
3113 bs->slice_time = wait_time * BLOCK_IO_SLICE_TIME * 10;
3114 bs->slice_end += bs->slice_time - 3 * BLOCK_IO_SLICE_TIME;
3115 if (wait) {
3116 *wait = wait_time * BLOCK_IO_SLICE_TIME * 10;
3117 }
3118
3119 return true;
3120}
3121
3122static bool bdrv_exceed_iops_limits(BlockDriverState *bs, bool is_write,
3123 double elapsed_time, uint64_t *wait)
3124{
3125 uint64_t iops_limit = 0;
3126 double ios_limit, ios_base;
3127 double slice_time, wait_time;
3128
3129 if (bs->io_limits.iops[BLOCK_IO_LIMIT_TOTAL]) {
3130 iops_limit = bs->io_limits.iops[BLOCK_IO_LIMIT_TOTAL];
3131 } else if (bs->io_limits.iops[is_write]) {
3132 iops_limit = bs->io_limits.iops[is_write];
3133 } else {
3134 if (wait) {
3135 *wait = 0;
3136 }
3137
3138 return false;
3139 }
3140
3141 slice_time = bs->slice_end - bs->slice_start;
3142 slice_time /= (NANOSECONDS_PER_SECOND);
3143 ios_limit = iops_limit * slice_time;
3144 ios_base = bs->nr_ops[is_write] - bs->io_base.ios[is_write];
3145 if (bs->io_limits.iops[BLOCK_IO_LIMIT_TOTAL]) {
3146 ios_base += bs->nr_ops[!is_write] - bs->io_base.ios[!is_write];
3147 }
3148
3149 if (ios_base + 1 <= ios_limit) {
3150 if (wait) {
3151 *wait = 0;
3152 }
3153
3154 return false;
3155 }
3156
3157 /* Calc approx time to dispatch */
3158 wait_time = (ios_base + 1) / iops_limit;
3159 if (wait_time > elapsed_time) {
3160 wait_time = wait_time - elapsed_time;
3161 } else {
3162 wait_time = 0;
3163 }
3164
3165 bs->slice_time = wait_time * BLOCK_IO_SLICE_TIME * 10;
3166 bs->slice_end += bs->slice_time - 3 * BLOCK_IO_SLICE_TIME;
3167 if (wait) {
3168 *wait = wait_time * BLOCK_IO_SLICE_TIME * 10;
3169 }
3170
3171 return true;
3172}
3173
3174static bool bdrv_exceed_io_limits(BlockDriverState *bs, int nb_sectors,
3175 bool is_write, int64_t *wait)
3176{
3177 int64_t now, max_wait;
3178 uint64_t bps_wait = 0, iops_wait = 0;
3179 double elapsed_time;
3180 int bps_ret, iops_ret;
3181
3182 now = qemu_get_clock_ns(vm_clock);
3183 if ((bs->slice_start < now)
3184 && (bs->slice_end > now)) {
3185 bs->slice_end = now + bs->slice_time;
3186 } else {
3187 bs->slice_time = 5 * BLOCK_IO_SLICE_TIME;
3188 bs->slice_start = now;
3189 bs->slice_end = now + bs->slice_time;
3190
3191 bs->io_base.bytes[is_write] = bs->nr_bytes[is_write];
3192 bs->io_base.bytes[!is_write] = bs->nr_bytes[!is_write];
3193
3194 bs->io_base.ios[is_write] = bs->nr_ops[is_write];
3195 bs->io_base.ios[!is_write] = bs->nr_ops[!is_write];
3196 }
3197
3198 elapsed_time = now - bs->slice_start;
3199 elapsed_time /= (NANOSECONDS_PER_SECOND);
3200
3201 bps_ret = bdrv_exceed_bps_limits(bs, nb_sectors,
3202 is_write, elapsed_time, &bps_wait);
3203 iops_ret = bdrv_exceed_iops_limits(bs, is_write,
3204 elapsed_time, &iops_wait);
3205 if (bps_ret || iops_ret) {
3206 max_wait = bps_wait > iops_wait ? bps_wait : iops_wait;
3207 if (wait) {
3208 *wait = max_wait;
3209 }
3210
3211 now = qemu_get_clock_ns(vm_clock);
3212 if (bs->slice_end < now + max_wait) {
3213 bs->slice_end = now + max_wait;
3214 }
3215
3216 return true;
3217 }
3218
3219 if (wait) {
3220 *wait = 0;
3221 }
3222
3223 return false;
3224}
ce1a14dc 3225
83f64091
FB
3226/**************************************************************/
3227/* async block device emulation */
3228
c16b5a2c
CH
3229typedef struct BlockDriverAIOCBSync {
3230 BlockDriverAIOCB common;
3231 QEMUBH *bh;
3232 int ret;
3233 /* vector translation state */
3234 QEMUIOVector *qiov;
3235 uint8_t *bounce;
3236 int is_write;
3237} BlockDriverAIOCBSync;
3238
3239static void bdrv_aio_cancel_em(BlockDriverAIOCB *blockacb)
3240{
b666d239
KW
3241 BlockDriverAIOCBSync *acb =
3242 container_of(blockacb, BlockDriverAIOCBSync, common);
6a7ad299 3243 qemu_bh_delete(acb->bh);
36afc451 3244 acb->bh = NULL;
c16b5a2c
CH
3245 qemu_aio_release(acb);
3246}
3247
3248static AIOPool bdrv_em_aio_pool = {
3249 .aiocb_size = sizeof(BlockDriverAIOCBSync),
3250 .cancel = bdrv_aio_cancel_em,
3251};
3252
ce1a14dc 3253static void bdrv_aio_bh_cb(void *opaque)
83f64091 3254{
ce1a14dc 3255 BlockDriverAIOCBSync *acb = opaque;
f141eafe 3256
f141eafe 3257 if (!acb->is_write)
03396148 3258 qemu_iovec_from_buf(acb->qiov, 0, acb->bounce, acb->qiov->size);
ceb42de8 3259 qemu_vfree(acb->bounce);
ce1a14dc 3260 acb->common.cb(acb->common.opaque, acb->ret);
6a7ad299 3261 qemu_bh_delete(acb->bh);
36afc451 3262 acb->bh = NULL;
ce1a14dc 3263 qemu_aio_release(acb);
83f64091 3264}
beac80cd 3265
f141eafe
AL
3266static BlockDriverAIOCB *bdrv_aio_rw_vector(BlockDriverState *bs,
3267 int64_t sector_num,
3268 QEMUIOVector *qiov,
3269 int nb_sectors,
3270 BlockDriverCompletionFunc *cb,
3271 void *opaque,
3272 int is_write)
3273
83f64091 3274{
ce1a14dc 3275 BlockDriverAIOCBSync *acb;
ce1a14dc 3276
c16b5a2c 3277 acb = qemu_aio_get(&bdrv_em_aio_pool, bs, cb, opaque);
f141eafe
AL
3278 acb->is_write = is_write;
3279 acb->qiov = qiov;
e268ca52 3280 acb->bounce = qemu_blockalign(bs, qiov->size);
3f3aace8 3281 acb->bh = qemu_bh_new(bdrv_aio_bh_cb, acb);
f141eafe
AL
3282
3283 if (is_write) {
d5e6b161 3284 qemu_iovec_to_buf(acb->qiov, 0, acb->bounce, qiov->size);
1ed20acf 3285 acb->ret = bs->drv->bdrv_write(bs, sector_num, acb->bounce, nb_sectors);
f141eafe 3286 } else {
1ed20acf 3287 acb->ret = bs->drv->bdrv_read(bs, sector_num, acb->bounce, nb_sectors);
f141eafe
AL
3288 }
3289
ce1a14dc 3290 qemu_bh_schedule(acb->bh);
f141eafe 3291
ce1a14dc 3292 return &acb->common;
beac80cd
FB
3293}
3294
f141eafe
AL
3295static BlockDriverAIOCB *bdrv_aio_readv_em(BlockDriverState *bs,
3296 int64_t sector_num, QEMUIOVector *qiov, int nb_sectors,
ce1a14dc 3297 BlockDriverCompletionFunc *cb, void *opaque)
beac80cd 3298{
f141eafe
AL
3299 return bdrv_aio_rw_vector(bs, sector_num, qiov, nb_sectors, cb, opaque, 0);
3300}
83f64091 3301
f141eafe
AL
3302static BlockDriverAIOCB *bdrv_aio_writev_em(BlockDriverState *bs,
3303 int64_t sector_num, QEMUIOVector *qiov, int nb_sectors,
3304 BlockDriverCompletionFunc *cb, void *opaque)
3305{
3306 return bdrv_aio_rw_vector(bs, sector_num, qiov, nb_sectors, cb, opaque, 1);
beac80cd 3307}
beac80cd 3308
68485420
KW
3309
3310typedef struct BlockDriverAIOCBCoroutine {
3311 BlockDriverAIOCB common;
3312 BlockRequest req;
3313 bool is_write;
3314 QEMUBH* bh;
3315} BlockDriverAIOCBCoroutine;
3316
3317static void bdrv_aio_co_cancel_em(BlockDriverAIOCB *blockacb)
3318{
3319 qemu_aio_flush();
3320}
3321
3322static AIOPool bdrv_em_co_aio_pool = {
3323 .aiocb_size = sizeof(BlockDriverAIOCBCoroutine),
3324 .cancel = bdrv_aio_co_cancel_em,
3325};
3326
35246a68 3327static void bdrv_co_em_bh(void *opaque)
68485420
KW
3328{
3329 BlockDriverAIOCBCoroutine *acb = opaque;
3330
3331 acb->common.cb(acb->common.opaque, acb->req.error);
3332 qemu_bh_delete(acb->bh);
3333 qemu_aio_release(acb);
3334}
3335
b2a61371
SH
3336/* Invoke bdrv_co_do_readv/bdrv_co_do_writev */
3337static void coroutine_fn bdrv_co_do_rw(void *opaque)
3338{
3339 BlockDriverAIOCBCoroutine *acb = opaque;
3340 BlockDriverState *bs = acb->common.bs;
3341
3342 if (!acb->is_write) {
3343 acb->req.error = bdrv_co_do_readv(bs, acb->req.sector,
470c0504 3344 acb->req.nb_sectors, acb->req.qiov, 0);
b2a61371
SH
3345 } else {
3346 acb->req.error = bdrv_co_do_writev(bs, acb->req.sector,
f08f2dda 3347 acb->req.nb_sectors, acb->req.qiov, 0);
b2a61371
SH
3348 }
3349
35246a68 3350 acb->bh = qemu_bh_new(bdrv_co_em_bh, acb);
b2a61371
SH
3351 qemu_bh_schedule(acb->bh);
3352}
3353
68485420
KW
3354static BlockDriverAIOCB *bdrv_co_aio_rw_vector(BlockDriverState *bs,
3355 int64_t sector_num,
3356 QEMUIOVector *qiov,
3357 int nb_sectors,
3358 BlockDriverCompletionFunc *cb,
3359 void *opaque,
8c5873d6 3360 bool is_write)
68485420
KW
3361{
3362 Coroutine *co;
3363 BlockDriverAIOCBCoroutine *acb;
3364
3365 acb = qemu_aio_get(&bdrv_em_co_aio_pool, bs, cb, opaque);
3366 acb->req.sector = sector_num;
3367 acb->req.nb_sectors = nb_sectors;
3368 acb->req.qiov = qiov;
3369 acb->is_write = is_write;
3370
8c5873d6 3371 co = qemu_coroutine_create(bdrv_co_do_rw);
68485420
KW
3372 qemu_coroutine_enter(co, acb);
3373
3374 return &acb->common;
3375}
3376
07f07615 3377static void coroutine_fn bdrv_aio_flush_co_entry(void *opaque)
b2e12bc6 3378{
07f07615
PB
3379 BlockDriverAIOCBCoroutine *acb = opaque;
3380 BlockDriverState *bs = acb->common.bs;
b2e12bc6 3381
07f07615
PB
3382 acb->req.error = bdrv_co_flush(bs);
3383 acb->bh = qemu_bh_new(bdrv_co_em_bh, acb);
b2e12bc6 3384 qemu_bh_schedule(acb->bh);
b2e12bc6
CH
3385}
3386
07f07615 3387BlockDriverAIOCB *bdrv_aio_flush(BlockDriverState *bs,
016f5cf6
AG
3388 BlockDriverCompletionFunc *cb, void *opaque)
3389{
07f07615 3390 trace_bdrv_aio_flush(bs, opaque);
016f5cf6 3391
07f07615
PB
3392 Coroutine *co;
3393 BlockDriverAIOCBCoroutine *acb;
016f5cf6 3394
07f07615
PB
3395 acb = qemu_aio_get(&bdrv_em_co_aio_pool, bs, cb, opaque);
3396 co = qemu_coroutine_create(bdrv_aio_flush_co_entry);
3397 qemu_coroutine_enter(co, acb);
016f5cf6 3398
016f5cf6
AG
3399 return &acb->common;
3400}
3401
4265d620
PB
3402static void coroutine_fn bdrv_aio_discard_co_entry(void *opaque)
3403{
3404 BlockDriverAIOCBCoroutine *acb = opaque;
3405 BlockDriverState *bs = acb->common.bs;
3406
3407 acb->req.error = bdrv_co_discard(bs, acb->req.sector, acb->req.nb_sectors);
3408 acb->bh = qemu_bh_new(bdrv_co_em_bh, acb);
3409 qemu_bh_schedule(acb->bh);
3410}
3411
3412BlockDriverAIOCB *bdrv_aio_discard(BlockDriverState *bs,
3413 int64_t sector_num, int nb_sectors,
3414 BlockDriverCompletionFunc *cb, void *opaque)
3415{
3416 Coroutine *co;
3417 BlockDriverAIOCBCoroutine *acb;
3418
3419 trace_bdrv_aio_discard(bs, sector_num, nb_sectors, opaque);
3420
3421 acb = qemu_aio_get(&bdrv_em_co_aio_pool, bs, cb, opaque);
3422 acb->req.sector = sector_num;
3423 acb->req.nb_sectors = nb_sectors;
3424 co = qemu_coroutine_create(bdrv_aio_discard_co_entry);
3425 qemu_coroutine_enter(co, acb);
3426
3427 return &acb->common;
3428}
3429
ea2384d3
FB
3430void bdrv_init(void)
3431{
5efa9d5a 3432 module_call_init(MODULE_INIT_BLOCK);
ea2384d3 3433}
ce1a14dc 3434
eb852011
MA
3435void bdrv_init_with_whitelist(void)
3436{
3437 use_bdrv_whitelist = 1;
3438 bdrv_init();
3439}
3440
c16b5a2c
CH
3441void *qemu_aio_get(AIOPool *pool, BlockDriverState *bs,
3442 BlockDriverCompletionFunc *cb, void *opaque)
ce1a14dc 3443{
ce1a14dc
PB
3444 BlockDriverAIOCB *acb;
3445
6bbff9a0
AL
3446 if (pool->free_aiocb) {
3447 acb = pool->free_aiocb;
3448 pool->free_aiocb = acb->next;
ce1a14dc 3449 } else {
7267c094 3450 acb = g_malloc0(pool->aiocb_size);
6bbff9a0 3451 acb->pool = pool;
ce1a14dc
PB
3452 }
3453 acb->bs = bs;
3454 acb->cb = cb;
3455 acb->opaque = opaque;
3456 return acb;
3457}
3458
3459void qemu_aio_release(void *p)
3460{
6bbff9a0
AL
3461 BlockDriverAIOCB *acb = (BlockDriverAIOCB *)p;
3462 AIOPool *pool = acb->pool;
3463 acb->next = pool->free_aiocb;
3464 pool->free_aiocb = acb;
ce1a14dc 3465}
19cb3738 3466
f9f05dc5
KW
3467/**************************************************************/
3468/* Coroutine block device emulation */
3469
3470typedef struct CoroutineIOCompletion {
3471 Coroutine *coroutine;
3472 int ret;
3473} CoroutineIOCompletion;
3474
3475static void bdrv_co_io_em_complete(void *opaque, int ret)
3476{
3477 CoroutineIOCompletion *co = opaque;
3478
3479 co->ret = ret;
3480 qemu_coroutine_enter(co->coroutine, NULL);
3481}
3482
3483static int coroutine_fn bdrv_co_io_em(BlockDriverState *bs, int64_t sector_num,
3484 int nb_sectors, QEMUIOVector *iov,
3485 bool is_write)
3486{
3487 CoroutineIOCompletion co = {
3488 .coroutine = qemu_coroutine_self(),
3489 };
3490 BlockDriverAIOCB *acb;
3491
3492 if (is_write) {
a652d160
SH
3493 acb = bs->drv->bdrv_aio_writev(bs, sector_num, iov, nb_sectors,
3494 bdrv_co_io_em_complete, &co);
f9f05dc5 3495 } else {
a652d160
SH
3496 acb = bs->drv->bdrv_aio_readv(bs, sector_num, iov, nb_sectors,
3497 bdrv_co_io_em_complete, &co);
f9f05dc5
KW
3498 }
3499
59370aaa 3500 trace_bdrv_co_io_em(bs, sector_num, nb_sectors, is_write, acb);
f9f05dc5
KW
3501 if (!acb) {
3502 return -EIO;
3503 }
3504 qemu_coroutine_yield();
3505
3506 return co.ret;
3507}
3508
3509static int coroutine_fn bdrv_co_readv_em(BlockDriverState *bs,
3510 int64_t sector_num, int nb_sectors,
3511 QEMUIOVector *iov)
3512{
3513 return bdrv_co_io_em(bs, sector_num, nb_sectors, iov, false);
3514}
3515
3516static int coroutine_fn bdrv_co_writev_em(BlockDriverState *bs,
3517 int64_t sector_num, int nb_sectors,
3518 QEMUIOVector *iov)
3519{
3520 return bdrv_co_io_em(bs, sector_num, nb_sectors, iov, true);
3521}
3522
07f07615 3523static void coroutine_fn bdrv_flush_co_entry(void *opaque)
e7a8a783 3524{
07f07615
PB
3525 RwCo *rwco = opaque;
3526
3527 rwco->ret = bdrv_co_flush(rwco->bs);
3528}
3529
3530int coroutine_fn bdrv_co_flush(BlockDriverState *bs)
3531{
eb489bb1
KW
3532 int ret;
3533
29cdb251 3534 if (!bs || !bdrv_is_inserted(bs) || bdrv_is_read_only(bs)) {
07f07615 3535 return 0;
eb489bb1
KW
3536 }
3537
ca716364 3538 /* Write back cached data to the OS even with cache=unsafe */
eb489bb1
KW
3539 if (bs->drv->bdrv_co_flush_to_os) {
3540 ret = bs->drv->bdrv_co_flush_to_os(bs);
3541 if (ret < 0) {
3542 return ret;
3543 }
3544 }
3545
ca716364
KW
3546 /* But don't actually force it to the disk with cache=unsafe */
3547 if (bs->open_flags & BDRV_O_NO_FLUSH) {
d4c82329 3548 goto flush_parent;
ca716364
KW
3549 }
3550
eb489bb1 3551 if (bs->drv->bdrv_co_flush_to_disk) {
29cdb251 3552 ret = bs->drv->bdrv_co_flush_to_disk(bs);
07f07615
PB
3553 } else if (bs->drv->bdrv_aio_flush) {
3554 BlockDriverAIOCB *acb;
3555 CoroutineIOCompletion co = {
3556 .coroutine = qemu_coroutine_self(),
3557 };
3558
3559 acb = bs->drv->bdrv_aio_flush(bs, bdrv_co_io_em_complete, &co);
3560 if (acb == NULL) {
29cdb251 3561 ret = -EIO;
07f07615
PB
3562 } else {
3563 qemu_coroutine_yield();
29cdb251 3564 ret = co.ret;
07f07615 3565 }
07f07615
PB
3566 } else {
3567 /*
3568 * Some block drivers always operate in either writethrough or unsafe
3569 * mode and don't support bdrv_flush therefore. Usually qemu doesn't
3570 * know how the server works (because the behaviour is hardcoded or
3571 * depends on server-side configuration), so we can't ensure that
3572 * everything is safe on disk. Returning an error doesn't work because
3573 * that would break guests even if the server operates in writethrough
3574 * mode.
3575 *
3576 * Let's hope the user knows what he's doing.
3577 */
29cdb251 3578 ret = 0;
07f07615 3579 }
29cdb251
PB
3580 if (ret < 0) {
3581 return ret;
3582 }
3583
3584 /* Now flush the underlying protocol. It will also have BDRV_O_NO_FLUSH
3585 * in the case of cache=unsafe, so there are no useless flushes.
3586 */
d4c82329 3587flush_parent:
29cdb251 3588 return bdrv_co_flush(bs->file);
07f07615
PB
3589}
3590
0f15423c
AL
3591void bdrv_invalidate_cache(BlockDriverState *bs)
3592{
3593 if (bs->drv && bs->drv->bdrv_invalidate_cache) {
3594 bs->drv->bdrv_invalidate_cache(bs);
3595 }
3596}
3597
3598void bdrv_invalidate_cache_all(void)
3599{
3600 BlockDriverState *bs;
3601
3602 QTAILQ_FOREACH(bs, &bdrv_states, list) {
3603 bdrv_invalidate_cache(bs);
3604 }
3605}
3606
07789269
BC
3607void bdrv_clear_incoming_migration_all(void)
3608{
3609 BlockDriverState *bs;
3610
3611 QTAILQ_FOREACH(bs, &bdrv_states, list) {
3612 bs->open_flags = bs->open_flags & ~(BDRV_O_INCOMING);
3613 }
3614}
3615
07f07615
PB
3616int bdrv_flush(BlockDriverState *bs)
3617{
3618 Coroutine *co;
3619 RwCo rwco = {
3620 .bs = bs,
3621 .ret = NOT_DONE,
e7a8a783 3622 };
e7a8a783 3623
07f07615
PB
3624 if (qemu_in_coroutine()) {
3625 /* Fast-path if already in coroutine context */
3626 bdrv_flush_co_entry(&rwco);
3627 } else {
3628 co = qemu_coroutine_create(bdrv_flush_co_entry);
3629 qemu_coroutine_enter(co, &rwco);
3630 while (rwco.ret == NOT_DONE) {
3631 qemu_aio_wait();
3632 }
e7a8a783 3633 }
07f07615
PB
3634
3635 return rwco.ret;
e7a8a783
KW
3636}
3637
4265d620
PB
3638static void coroutine_fn bdrv_discard_co_entry(void *opaque)
3639{
3640 RwCo *rwco = opaque;
3641
3642 rwco->ret = bdrv_co_discard(rwco->bs, rwco->sector_num, rwco->nb_sectors);
3643}
3644
3645int coroutine_fn bdrv_co_discard(BlockDriverState *bs, int64_t sector_num,
3646 int nb_sectors)
3647{
3648 if (!bs->drv) {
3649 return -ENOMEDIUM;
3650 } else if (bdrv_check_request(bs, sector_num, nb_sectors)) {
3651 return -EIO;
3652 } else if (bs->read_only) {
3653 return -EROFS;
3654 } else if (bs->drv->bdrv_co_discard) {
3655 return bs->drv->bdrv_co_discard(bs, sector_num, nb_sectors);
3656 } else if (bs->drv->bdrv_aio_discard) {
3657 BlockDriverAIOCB *acb;
3658 CoroutineIOCompletion co = {
3659 .coroutine = qemu_coroutine_self(),
3660 };
3661
3662 acb = bs->drv->bdrv_aio_discard(bs, sector_num, nb_sectors,
3663 bdrv_co_io_em_complete, &co);
3664 if (acb == NULL) {
3665 return -EIO;
3666 } else {
3667 qemu_coroutine_yield();
3668 return co.ret;
3669 }
4265d620
PB
3670 } else {
3671 return 0;
3672 }
3673}
3674
3675int bdrv_discard(BlockDriverState *bs, int64_t sector_num, int nb_sectors)
3676{
3677 Coroutine *co;
3678 RwCo rwco = {
3679 .bs = bs,
3680 .sector_num = sector_num,
3681 .nb_sectors = nb_sectors,
3682 .ret = NOT_DONE,
3683 };
3684
3685 if (qemu_in_coroutine()) {
3686 /* Fast-path if already in coroutine context */
3687 bdrv_discard_co_entry(&rwco);
3688 } else {
3689 co = qemu_coroutine_create(bdrv_discard_co_entry);
3690 qemu_coroutine_enter(co, &rwco);
3691 while (rwco.ret == NOT_DONE) {
3692 qemu_aio_wait();
3693 }
3694 }
3695
3696 return rwco.ret;
3697}
3698
19cb3738
FB
3699/**************************************************************/
3700/* removable device support */
3701
3702/**
3703 * Return TRUE if the media is present
3704 */
3705int bdrv_is_inserted(BlockDriverState *bs)
3706{
3707 BlockDriver *drv = bs->drv;
a1aff5bf 3708
19cb3738
FB
3709 if (!drv)
3710 return 0;
3711 if (!drv->bdrv_is_inserted)
a1aff5bf
MA
3712 return 1;
3713 return drv->bdrv_is_inserted(bs);
19cb3738
FB
3714}
3715
3716/**
8e49ca46
MA
3717 * Return whether the media changed since the last call to this
3718 * function, or -ENOTSUP if we don't know. Most drivers don't know.
19cb3738
FB
3719 */
3720int bdrv_media_changed(BlockDriverState *bs)
3721{
3722 BlockDriver *drv = bs->drv;
19cb3738 3723
8e49ca46
MA
3724 if (drv && drv->bdrv_media_changed) {
3725 return drv->bdrv_media_changed(bs);
3726 }
3727 return -ENOTSUP;
19cb3738
FB
3728}
3729
3730/**
3731 * If eject_flag is TRUE, eject the media. Otherwise, close the tray
3732 */
f36f3949 3733void bdrv_eject(BlockDriverState *bs, bool eject_flag)
19cb3738
FB
3734{
3735 BlockDriver *drv = bs->drv;
19cb3738 3736
822e1cd1
MA
3737 if (drv && drv->bdrv_eject) {
3738 drv->bdrv_eject(bs, eject_flag);
19cb3738 3739 }
6f382ed2
LC
3740
3741 if (bs->device_name[0] != '\0') {
3742 bdrv_emit_qmp_eject_event(bs, eject_flag);
3743 }
19cb3738
FB
3744}
3745
19cb3738
FB
3746/**
3747 * Lock or unlock the media (if it is locked, the user won't be able
3748 * to eject it manually).
3749 */
025e849a 3750void bdrv_lock_medium(BlockDriverState *bs, bool locked)
19cb3738
FB
3751{
3752 BlockDriver *drv = bs->drv;
3753
025e849a 3754 trace_bdrv_lock_medium(bs, locked);
b8c6d095 3755
025e849a
MA
3756 if (drv && drv->bdrv_lock_medium) {
3757 drv->bdrv_lock_medium(bs, locked);
19cb3738
FB
3758 }
3759}
985a03b0
TS
3760
3761/* needed for generic scsi interface */
3762
3763int bdrv_ioctl(BlockDriverState *bs, unsigned long int req, void *buf)
3764{
3765 BlockDriver *drv = bs->drv;
3766
3767 if (drv && drv->bdrv_ioctl)
3768 return drv->bdrv_ioctl(bs, req, buf);
3769 return -ENOTSUP;
3770}
7d780669 3771
221f715d
AL
3772BlockDriverAIOCB *bdrv_aio_ioctl(BlockDriverState *bs,
3773 unsigned long int req, void *buf,
3774 BlockDriverCompletionFunc *cb, void *opaque)
7d780669 3775{
221f715d 3776 BlockDriver *drv = bs->drv;
7d780669 3777
221f715d
AL
3778 if (drv && drv->bdrv_aio_ioctl)
3779 return drv->bdrv_aio_ioctl(bs, req, buf, cb, opaque);
3780 return NULL;
7d780669 3781}
e268ca52 3782
7b6f9300
MA
3783void bdrv_set_buffer_alignment(BlockDriverState *bs, int align)
3784{
3785 bs->buffer_alignment = align;
3786}
7cd1e32a 3787
e268ca52
AL
3788void *qemu_blockalign(BlockDriverState *bs, size_t size)
3789{
3790 return qemu_memalign((bs && bs->buffer_alignment) ? bs->buffer_alignment : 512, size);
3791}
7cd1e32a 3792
3793void bdrv_set_dirty_tracking(BlockDriverState *bs, int enable)
3794{
3795 int64_t bitmap_size;
a55eb92c 3796
aaa0eb75 3797 bs->dirty_count = 0;
a55eb92c 3798 if (enable) {
c6d22830
JK
3799 if (!bs->dirty_bitmap) {
3800 bitmap_size = (bdrv_getlength(bs) >> BDRV_SECTOR_BITS) +
71df14fc
PB
3801 BDRV_SECTORS_PER_DIRTY_CHUNK * BITS_PER_LONG - 1;
3802 bitmap_size /= BDRV_SECTORS_PER_DIRTY_CHUNK * BITS_PER_LONG;
a55eb92c 3803
71df14fc 3804 bs->dirty_bitmap = g_new0(unsigned long, bitmap_size);
a55eb92c 3805 }
7cd1e32a 3806 } else {
c6d22830 3807 if (bs->dirty_bitmap) {
7267c094 3808 g_free(bs->dirty_bitmap);
c6d22830 3809 bs->dirty_bitmap = NULL;
a55eb92c 3810 }
7cd1e32a 3811 }
3812}
3813
3814int bdrv_get_dirty(BlockDriverState *bs, int64_t sector)
3815{
6ea44308 3816 int64_t chunk = sector / (int64_t)BDRV_SECTORS_PER_DIRTY_CHUNK;
a55eb92c 3817
c6d22830
JK
3818 if (bs->dirty_bitmap &&
3819 (sector << BDRV_SECTOR_BITS) < bdrv_getlength(bs)) {
6d59fec1
MT
3820 return !!(bs->dirty_bitmap[chunk / (sizeof(unsigned long) * 8)] &
3821 (1UL << (chunk % (sizeof(unsigned long) * 8))));
7cd1e32a 3822 } else {
3823 return 0;
3824 }
3825}
3826
a55eb92c
JK
3827void bdrv_reset_dirty(BlockDriverState *bs, int64_t cur_sector,
3828 int nr_sectors)
7cd1e32a 3829{
3830 set_dirty_bitmap(bs, cur_sector, nr_sectors, 0);
3831}
aaa0eb75
LS
3832
3833int64_t bdrv_get_dirty_count(BlockDriverState *bs)
3834{
3835 return bs->dirty_count;
3836}
f88e1a42 3837
db593f25
MT
3838void bdrv_set_in_use(BlockDriverState *bs, int in_use)
3839{
3840 assert(bs->in_use != in_use);
3841 bs->in_use = in_use;
3842}
3843
3844int bdrv_in_use(BlockDriverState *bs)
3845{
3846 return bs->in_use;
3847}
3848
28a7282a
LC
3849void bdrv_iostatus_enable(BlockDriverState *bs)
3850{
d6bf279e 3851 bs->iostatus_enabled = true;
58e21ef5 3852 bs->iostatus = BLOCK_DEVICE_IO_STATUS_OK;
28a7282a
LC
3853}
3854
3855/* The I/O status is only enabled if the drive explicitly
3856 * enables it _and_ the VM is configured to stop on errors */
3857bool bdrv_iostatus_is_enabled(const BlockDriverState *bs)
3858{
d6bf279e 3859 return (bs->iostatus_enabled &&
28a7282a
LC
3860 (bs->on_write_error == BLOCK_ERR_STOP_ENOSPC ||
3861 bs->on_write_error == BLOCK_ERR_STOP_ANY ||
3862 bs->on_read_error == BLOCK_ERR_STOP_ANY));
3863}
3864
3865void bdrv_iostatus_disable(BlockDriverState *bs)
3866{
d6bf279e 3867 bs->iostatus_enabled = false;
28a7282a
LC
3868}
3869
3870void bdrv_iostatus_reset(BlockDriverState *bs)
3871{
3872 if (bdrv_iostatus_is_enabled(bs)) {
58e21ef5 3873 bs->iostatus = BLOCK_DEVICE_IO_STATUS_OK;
28a7282a
LC
3874 }
3875}
3876
3877/* XXX: Today this is set by device models because it makes the implementation
3878 quite simple. However, the block layer knows about the error, so it's
3879 possible to implement this without device models being involved */
3880void bdrv_iostatus_set_err(BlockDriverState *bs, int error)
3881{
58e21ef5
LC
3882 if (bdrv_iostatus_is_enabled(bs) &&
3883 bs->iostatus == BLOCK_DEVICE_IO_STATUS_OK) {
28a7282a 3884 assert(error >= 0);
58e21ef5
LC
3885 bs->iostatus = error == ENOSPC ? BLOCK_DEVICE_IO_STATUS_NOSPACE :
3886 BLOCK_DEVICE_IO_STATUS_FAILED;
28a7282a
LC
3887 }
3888}
3889
a597e79c
CH
3890void
3891bdrv_acct_start(BlockDriverState *bs, BlockAcctCookie *cookie, int64_t bytes,
3892 enum BlockAcctType type)
3893{
3894 assert(type < BDRV_MAX_IOTYPE);
3895
3896 cookie->bytes = bytes;
c488c7f6 3897 cookie->start_time_ns = get_clock();
a597e79c
CH
3898 cookie->type = type;
3899}
3900
3901void
3902bdrv_acct_done(BlockDriverState *bs, BlockAcctCookie *cookie)
3903{
3904 assert(cookie->type < BDRV_MAX_IOTYPE);
3905
3906 bs->nr_bytes[cookie->type] += cookie->bytes;
3907 bs->nr_ops[cookie->type]++;
c488c7f6 3908 bs->total_time_ns[cookie->type] += get_clock() - cookie->start_time_ns;
a597e79c
CH
3909}
3910
f88e1a42
JS
3911int bdrv_img_create(const char *filename, const char *fmt,
3912 const char *base_filename, const char *base_fmt,
3913 char *options, uint64_t img_size, int flags)
3914{
3915 QEMUOptionParameter *param = NULL, *create_options = NULL;
d220894e 3916 QEMUOptionParameter *backing_fmt, *backing_file, *size;
f88e1a42
JS
3917 BlockDriverState *bs = NULL;
3918 BlockDriver *drv, *proto_drv;
96df67d1 3919 BlockDriver *backing_drv = NULL;
f88e1a42
JS
3920 int ret = 0;
3921
3922 /* Find driver and parse its options */
3923 drv = bdrv_find_format(fmt);
3924 if (!drv) {
3925 error_report("Unknown file format '%s'", fmt);
4f70f249 3926 ret = -EINVAL;
f88e1a42
JS
3927 goto out;
3928 }
3929
3930 proto_drv = bdrv_find_protocol(filename);
3931 if (!proto_drv) {
3932 error_report("Unknown protocol '%s'", filename);
4f70f249 3933 ret = -EINVAL;
f88e1a42
JS
3934 goto out;
3935 }
3936
3937 create_options = append_option_parameters(create_options,
3938 drv->create_options);
3939 create_options = append_option_parameters(create_options,
3940 proto_drv->create_options);
3941
3942 /* Create parameter list with default values */
3943 param = parse_option_parameters("", create_options, param);
3944
3945 set_option_parameter_int(param, BLOCK_OPT_SIZE, img_size);
3946
3947 /* Parse -o options */
3948 if (options) {
3949 param = parse_option_parameters(options, create_options, param);
3950 if (param == NULL) {
3951 error_report("Invalid options for file format '%s'.", fmt);
4f70f249 3952 ret = -EINVAL;
f88e1a42
JS
3953 goto out;
3954 }
3955 }
3956
3957 if (base_filename) {
3958 if (set_option_parameter(param, BLOCK_OPT_BACKING_FILE,
3959 base_filename)) {
3960 error_report("Backing file not supported for file format '%s'",
3961 fmt);
4f70f249 3962 ret = -EINVAL;
f88e1a42
JS
3963 goto out;
3964 }
3965 }
3966
3967 if (base_fmt) {
3968 if (set_option_parameter(param, BLOCK_OPT_BACKING_FMT, base_fmt)) {
3969 error_report("Backing file format not supported for file "
3970 "format '%s'", fmt);
4f70f249 3971 ret = -EINVAL;
f88e1a42
JS
3972 goto out;
3973 }
3974 }
3975
792da93a
JS
3976 backing_file = get_option_parameter(param, BLOCK_OPT_BACKING_FILE);
3977 if (backing_file && backing_file->value.s) {
3978 if (!strcmp(filename, backing_file->value.s)) {
3979 error_report("Error: Trying to create an image with the "
3980 "same filename as the backing file");
4f70f249 3981 ret = -EINVAL;
792da93a
JS
3982 goto out;
3983 }
3984 }
3985
f88e1a42
JS
3986 backing_fmt = get_option_parameter(param, BLOCK_OPT_BACKING_FMT);
3987 if (backing_fmt && backing_fmt->value.s) {
96df67d1
SH
3988 backing_drv = bdrv_find_format(backing_fmt->value.s);
3989 if (!backing_drv) {
f88e1a42
JS
3990 error_report("Unknown backing file format '%s'",
3991 backing_fmt->value.s);
4f70f249 3992 ret = -EINVAL;
f88e1a42
JS
3993 goto out;
3994 }
3995 }
3996
3997 // The size for the image must always be specified, with one exception:
3998 // If we are using a backing file, we can obtain the size from there
d220894e
KW
3999 size = get_option_parameter(param, BLOCK_OPT_SIZE);
4000 if (size && size->value.n == -1) {
f88e1a42
JS
4001 if (backing_file && backing_file->value.s) {
4002 uint64_t size;
f88e1a42 4003 char buf[32];
63090dac
PB
4004 int back_flags;
4005
4006 /* backing files always opened read-only */
4007 back_flags =
4008 flags & ~(BDRV_O_RDWR | BDRV_O_SNAPSHOT | BDRV_O_NO_BACKING);
f88e1a42 4009
f88e1a42
JS
4010 bs = bdrv_new("");
4011
63090dac 4012 ret = bdrv_open(bs, backing_file->value.s, back_flags, backing_drv);
f88e1a42 4013 if (ret < 0) {
96df67d1 4014 error_report("Could not open '%s'", backing_file->value.s);
f88e1a42
JS
4015 goto out;
4016 }
4017 bdrv_get_geometry(bs, &size);
4018 size *= 512;
4019
4020 snprintf(buf, sizeof(buf), "%" PRId64, size);
4021 set_option_parameter(param, BLOCK_OPT_SIZE, buf);
4022 } else {
4023 error_report("Image creation needs a size parameter");
4f70f249 4024 ret = -EINVAL;
f88e1a42
JS
4025 goto out;
4026 }
4027 }
4028
4029 printf("Formatting '%s', fmt=%s ", filename, fmt);
4030 print_option_parameters(param);
4031 puts("");
4032
4033 ret = bdrv_create(drv, filename, param);
4034
4035 if (ret < 0) {
4036 if (ret == -ENOTSUP) {
4037 error_report("Formatting or formatting option not supported for "
4038 "file format '%s'", fmt);
4039 } else if (ret == -EFBIG) {
4040 error_report("The image size is too large for file format '%s'",
4041 fmt);
4042 } else {
4043 error_report("%s: error while creating %s: %s", filename, fmt,
4044 strerror(-ret));
4045 }
4046 }
4047
4048out:
4049 free_option_parameters(create_options);
4050 free_option_parameters(param);
4051
4052 if (bs) {
4053 bdrv_delete(bs);
4054 }
4f70f249
JS
4055
4056 return ret;
f88e1a42 4057}
eeec61f2
SH
4058
4059void *block_job_create(const BlockJobType *job_type, BlockDriverState *bs,
c83c66c3
SH
4060 int64_t speed, BlockDriverCompletionFunc *cb,
4061 void *opaque, Error **errp)
eeec61f2
SH
4062{
4063 BlockJob *job;
4064
4065 if (bs->job || bdrv_in_use(bs)) {
fd7f8c65 4066 error_set(errp, QERR_DEVICE_IN_USE, bdrv_get_device_name(bs));
eeec61f2
SH
4067 return NULL;
4068 }
4069 bdrv_set_in_use(bs, 1);
4070
4071 job = g_malloc0(job_type->instance_size);
4072 job->job_type = job_type;
4073 job->bs = bs;
4074 job->cb = cb;
4075 job->opaque = opaque;
4513eafe 4076 job->busy = true;
eeec61f2 4077 bs->job = job;
c83c66c3
SH
4078
4079 /* Only set speed when necessary to avoid NotSupported error */
4080 if (speed != 0) {
4081 Error *local_err = NULL;
4082
4083 block_job_set_speed(job, speed, &local_err);
4084 if (error_is_set(&local_err)) {
4085 bs->job = NULL;
4086 g_free(job);
4087 bdrv_set_in_use(bs, 0);
4088 error_propagate(errp, local_err);
4089 return NULL;
4090 }
4091 }
eeec61f2
SH
4092 return job;
4093}
4094
4095void block_job_complete(BlockJob *job, int ret)
4096{
4097 BlockDriverState *bs = job->bs;
4098
4099 assert(bs->job == job);
4100 job->cb(job->opaque, ret);
4101 bs->job = NULL;
4102 g_free(job);
4103 bdrv_set_in_use(bs, 0);
4104}
4105
882ec7ce 4106void block_job_set_speed(BlockJob *job, int64_t speed, Error **errp)
eeec61f2 4107{
9e6636c7 4108 Error *local_err = NULL;
9f25eccc 4109
eeec61f2 4110 if (!job->job_type->set_speed) {
9e6636c7
SH
4111 error_set(errp, QERR_NOT_SUPPORTED);
4112 return;
eeec61f2 4113 }
882ec7ce 4114 job->job_type->set_speed(job, speed, &local_err);
9e6636c7
SH
4115 if (error_is_set(&local_err)) {
4116 error_propagate(errp, local_err);
4117 return;
9f25eccc 4118 }
9e6636c7 4119
882ec7ce 4120 job->speed = speed;
eeec61f2
SH
4121}
4122
4123void block_job_cancel(BlockJob *job)
4124{
4125 job->cancelled = true;
fa4478d5
PB
4126 if (job->co && !job->busy) {
4127 qemu_coroutine_enter(job->co, NULL);
4128 }
eeec61f2
SH
4129}
4130
4131bool block_job_is_cancelled(BlockJob *job)
4132{
4133 return job->cancelled;
4134}
3e914655 4135
fa4478d5
PB
4136struct BlockCancelData {
4137 BlockJob *job;
4138 BlockDriverCompletionFunc *cb;
4139 void *opaque;
4140 bool cancelled;
4141 int ret;
4142};
4143
4144static void block_job_cancel_cb(void *opaque, int ret)
3e914655 4145{
fa4478d5
PB
4146 struct BlockCancelData *data = opaque;
4147
4148 data->cancelled = block_job_is_cancelled(data->job);
4149 data->ret = ret;
4150 data->cb(data->opaque, ret);
4151}
4152
4153int block_job_cancel_sync(BlockJob *job)
4154{
4155 struct BlockCancelData data;
3e914655
PB
4156 BlockDriverState *bs = job->bs;
4157
4158 assert(bs->job == job);
fa4478d5
PB
4159
4160 /* Set up our own callback to store the result and chain to
4161 * the original callback.
4162 */
4163 data.job = job;
4164 data.cb = job->cb;
4165 data.opaque = job->opaque;
4166 data.ret = -EINPROGRESS;
4167 job->cb = block_job_cancel_cb;
4168 job->opaque = &data;
3e914655 4169 block_job_cancel(job);
fa4478d5 4170 while (data.ret == -EINPROGRESS) {
3e914655
PB
4171 qemu_aio_wait();
4172 }
fa4478d5 4173 return (data.cancelled && data.ret == 0) ? -ECANCELED : data.ret;
3e914655 4174}
4513eafe
PB
4175
4176void block_job_sleep_ns(BlockJob *job, QEMUClock *clock, int64_t ns)
4177{
4178 /* Check cancellation *before* setting busy = false, too! */
4179 if (!block_job_is_cancelled(job)) {
4180 job->busy = false;
4181 co_sleep_ns(clock, ns);
4182 job->busy = true;
4183 }
4184}