]>
Commit | Line | Data |
---|---|---|
3b491a90 EGE |
1 | /* |
2 | * QEMU System Emulator block driver | |
3 | * | |
4 | * Copyright (c) 2003 Fabrice Bellard | |
5 | * | |
6 | * Permission is hereby granted, free of charge, to any person obtaining a copy | |
7 | * of this software and associated documentation files (the "Software"), to deal | |
8 | * in the Software without restriction, including without limitation the rights | |
9 | * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell | |
10 | * copies of the Software, and to permit persons to whom the Software is | |
11 | * furnished to do so, subject to the following conditions: | |
12 | * | |
13 | * The above copyright notice and this permission notice shall be included in | |
14 | * all copies or substantial portions of the Software. | |
15 | * | |
16 | * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR | |
17 | * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, | |
18 | * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL | |
19 | * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER | |
20 | * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, | |
21 | * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN | |
22 | * THE SOFTWARE. | |
23 | */ | |
24 | #ifndef BLOCK_IO_H | |
25 | #define BLOCK_IO_H | |
26 | ||
e2c1c34f MA |
27 | #include "block/aio-wait.h" |
28 | #include "block/block-common.h" | |
29 | #include "qemu/coroutine.h" | |
30 | #include "qemu/iov.h" | |
3b491a90 EGE |
31 | |
32 | /* | |
33 | * I/O API functions. These functions are thread-safe, and therefore | |
34 | * can run in any thread as long as the thread has called | |
35 | * aio_context_acquire/release(). | |
36 | * | |
37 | * These functions can only call functions from I/O and Common categories, | |
38 | * but can be invoked by GS, "I/O or GS" and I/O APIs. | |
39 | * | |
40 | * All functions in this category must use the macro | |
41 | * IO_CODE(); | |
42 | * to catch when they are accidentally called by the wrong API. | |
43 | */ | |
44 | ||
90830f59 EGE |
45 | int co_wrapper_mixed_bdrv_rdlock |
46 | bdrv_pwrite_zeroes(BdrvChild *child, int64_t offset, int64_t bytes, | |
47 | BdrvRequestFlags flags); | |
48 | ||
3b491a90 | 49 | int bdrv_make_zero(BdrvChild *child, BdrvRequestFlags flags); |
90830f59 EGE |
50 | |
51 | int co_wrapper_mixed_bdrv_rdlock | |
52 | bdrv_pread(BdrvChild *child, int64_t offset, int64_t bytes, void *buf, | |
53 | BdrvRequestFlags flags); | |
54 | ||
55 | int co_wrapper_mixed_bdrv_rdlock | |
56 | bdrv_pwrite(BdrvChild *child, int64_t offset,int64_t bytes, | |
57 | const void *buf, BdrvRequestFlags flags); | |
58 | ||
59 | int co_wrapper_mixed_bdrv_rdlock | |
60 | bdrv_pwrite_sync(BdrvChild *child, int64_t offset, int64_t bytes, | |
61 | const void *buf, BdrvRequestFlags flags); | |
62 | ||
e97190a4 AF |
63 | int coroutine_fn bdrv_co_pwrite_sync(BdrvChild *child, int64_t offset, |
64 | int64_t bytes, const void *buf, | |
65 | BdrvRequestFlags flags); | |
3b491a90 EGE |
66 | /* |
67 | * Efficiently zero a region of the disk image. Note that this is a regular | |
68 | * I/O request like read or write and should have a reasonable size. This | |
69 | * function is not suitable for zeroing the entire image in a single request | |
70 | * because it may allocate memory for the entire region. | |
71 | */ | |
72 | int coroutine_fn bdrv_co_pwrite_zeroes(BdrvChild *child, int64_t offset, | |
73 | int64_t bytes, BdrvRequestFlags flags); | |
74 | ||
75 | int coroutine_fn bdrv_co_truncate(BdrvChild *child, int64_t offset, bool exact, | |
76 | PreallocMode prealloc, BdrvRequestFlags flags, | |
77 | Error **errp); | |
78 | ||
c86422c5 EGE |
79 | int64_t coroutine_fn bdrv_co_nb_sectors(BlockDriverState *bs); |
80 | int64_t co_wrapper_mixed bdrv_nb_sectors(BlockDriverState *bs); | |
81 | ||
82 | int64_t coroutine_fn bdrv_co_getlength(BlockDriverState *bs); | |
83 | int64_t co_wrapper_mixed bdrv_getlength(BlockDriverState *bs); | |
84 | ||
82618d7b EGE |
85 | int64_t coroutine_fn bdrv_co_get_allocated_file_size(BlockDriverState *bs); |
86 | int64_t co_wrapper bdrv_get_allocated_file_size(BlockDriverState *bs); | |
87 | ||
3b491a90 EGE |
88 | BlockMeasureInfo *bdrv_measure(BlockDriver *drv, QemuOpts *opts, |
89 | BlockDriverState *in_bs, Error **errp); | |
90 | void bdrv_get_geometry(BlockDriverState *bs, uint64_t *nb_sectors_ptr); | |
91 | int coroutine_fn bdrv_co_delete_file(BlockDriverState *bs, Error **errp); | |
92 | void coroutine_fn bdrv_co_delete_file_noerr(BlockDriverState *bs); | |
93 | ||
94 | ||
95 | /* async block I/O */ | |
96 | void bdrv_aio_cancel(BlockAIOCB *acb); | |
97 | void bdrv_aio_cancel_async(BlockAIOCB *acb); | |
98 | ||
99 | /* sg packet commands */ | |
16bb776f | 100 | int coroutine_fn bdrv_co_ioctl(BlockDriverState *bs, int req, void *buf); |
3b491a90 EGE |
101 | |
102 | /* Ensure contents are flushed to disk. */ | |
103 | int coroutine_fn bdrv_co_flush(BlockDriverState *bs); | |
104 | ||
16bb776f AF |
105 | int coroutine_fn bdrv_co_pdiscard(BdrvChild *child, int64_t offset, |
106 | int64_t bytes); | |
3b491a90 EGE |
107 | bool bdrv_can_write_zeroes_with_unmap(BlockDriverState *bs); |
108 | int bdrv_block_status(BlockDriverState *bs, int64_t offset, | |
109 | int64_t bytes, int64_t *pnum, int64_t *map, | |
110 | BlockDriverState **file); | |
7b52a921 EGE |
111 | |
112 | int coroutine_fn bdrv_co_block_status_above(BlockDriverState *bs, | |
113 | BlockDriverState *base, | |
114 | int64_t offset, int64_t bytes, | |
115 | int64_t *pnum, int64_t *map, | |
116 | BlockDriverState **file); | |
3b491a90 EGE |
117 | int bdrv_block_status_above(BlockDriverState *bs, BlockDriverState *base, |
118 | int64_t offset, int64_t bytes, int64_t *pnum, | |
119 | int64_t *map, BlockDriverState **file); | |
7b52a921 EGE |
120 | |
121 | int coroutine_fn bdrv_co_is_allocated(BlockDriverState *bs, int64_t offset, | |
122 | int64_t bytes, int64_t *pnum); | |
3b491a90 EGE |
123 | int bdrv_is_allocated(BlockDriverState *bs, int64_t offset, int64_t bytes, |
124 | int64_t *pnum); | |
7b52a921 EGE |
125 | |
126 | int coroutine_fn bdrv_co_is_allocated_above(BlockDriverState *top, | |
127 | BlockDriverState *base, | |
128 | bool include_base, int64_t offset, | |
129 | int64_t bytes, int64_t *pnum); | |
3b491a90 EGE |
130 | int bdrv_is_allocated_above(BlockDriverState *top, BlockDriverState *base, |
131 | bool include_base, int64_t offset, int64_t bytes, | |
132 | int64_t *pnum); | |
7b52a921 | 133 | |
3b491a90 EGE |
134 | int coroutine_fn bdrv_co_is_zero_fast(BlockDriverState *bs, int64_t offset, |
135 | int64_t bytes); | |
136 | ||
137 | int bdrv_can_set_read_only(BlockDriverState *bs, bool read_only, | |
138 | bool ignore_allow_rdw, Error **errp); | |
139 | int bdrv_apply_auto_read_only(BlockDriverState *bs, const char *errmsg, | |
140 | Error **errp); | |
141 | bool bdrv_is_read_only(BlockDriverState *bs); | |
142 | bool bdrv_is_writable(BlockDriverState *bs); | |
143 | bool bdrv_is_sg(BlockDriverState *bs); | |
15aee7ac | 144 | int bdrv_get_flags(BlockDriverState *bs); |
1e97be91 EGE |
145 | |
146 | bool coroutine_fn bdrv_co_is_inserted(BlockDriverState *bs); | |
147 | bool co_wrapper bdrv_is_inserted(BlockDriverState *bs); | |
148 | ||
3b491a90 EGE |
149 | void bdrv_lock_medium(BlockDriverState *bs, bool locked); |
150 | void bdrv_eject(BlockDriverState *bs, bool eject_flag); | |
151 | const char *bdrv_get_format_name(BlockDriverState *bs); | |
152 | ||
153 | bool bdrv_supports_compressed_writes(BlockDriverState *bs); | |
154 | const char *bdrv_get_node_name(const BlockDriverState *bs); | |
155 | const char *bdrv_get_device_name(const BlockDriverState *bs); | |
156 | const char *bdrv_get_device_or_node_name(const BlockDriverState *bs); | |
3d47eb0a EGE |
157 | |
158 | int coroutine_fn bdrv_co_get_info(BlockDriverState *bs, BlockDriverInfo *bdi); | |
159 | int co_wrapper_mixed bdrv_get_info(BlockDriverState *bs, BlockDriverInfo *bdi); | |
160 | ||
3b491a90 EGE |
161 | ImageInfoSpecific *bdrv_get_specific_info(BlockDriverState *bs, |
162 | Error **errp); | |
163 | BlockStatsSpecific *bdrv_get_specific_stats(BlockDriverState *bs); | |
164 | void bdrv_round_to_clusters(BlockDriverState *bs, | |
165 | int64_t offset, int64_t bytes, | |
166 | int64_t *cluster_offset, | |
167 | int64_t *cluster_bytes); | |
168 | ||
169 | void bdrv_get_backing_filename(BlockDriverState *bs, | |
170 | char *filename, int filename_size); | |
171 | ||
172 | int bdrv_save_vmstate(BlockDriverState *bs, const uint8_t *buf, | |
173 | int64_t pos, int size); | |
174 | ||
175 | int bdrv_load_vmstate(BlockDriverState *bs, uint8_t *buf, | |
176 | int64_t pos, int size); | |
177 | ||
178 | /* | |
179 | * Returns the alignment in bytes that is required so that no bounce buffer | |
180 | * is required throughout the stack | |
181 | */ | |
182 | size_t bdrv_min_mem_align(BlockDriverState *bs); | |
183 | /* Returns optimal alignment in bytes for bounce buffer */ | |
184 | size_t bdrv_opt_mem_align(BlockDriverState *bs); | |
185 | void *qemu_blockalign(BlockDriverState *bs, size_t size); | |
186 | void *qemu_blockalign0(BlockDriverState *bs, size_t size); | |
187 | void *qemu_try_blockalign(BlockDriverState *bs, size_t size); | |
188 | void *qemu_try_blockalign0(BlockDriverState *bs, size_t size); | |
3b491a90 EGE |
189 | |
190 | void bdrv_enable_copy_on_read(BlockDriverState *bs); | |
191 | void bdrv_disable_copy_on_read(BlockDriverState *bs); | |
192 | ||
193 | void bdrv_debug_event(BlockDriverState *bs, BlkdebugEvent event); | |
194 | ||
195 | #define BLKDBG_EVENT(child, evt) \ | |
196 | do { \ | |
197 | if (child) { \ | |
198 | bdrv_debug_event(child->bs, evt); \ | |
199 | } \ | |
200 | } while (0) | |
201 | ||
202 | /** | |
203 | * bdrv_get_aio_context: | |
204 | * | |
205 | * Returns: the currently bound #AioContext | |
206 | */ | |
207 | AioContext *bdrv_get_aio_context(BlockDriverState *bs); | |
208 | ||
d5f8d79c HR |
209 | AioContext *bdrv_child_get_parent_aio_context(BdrvChild *c); |
210 | ||
3b491a90 EGE |
211 | /** |
212 | * Move the current coroutine to the AioContext of @bs and return the old | |
213 | * AioContext of the coroutine. Increase bs->in_flight so that draining @bs | |
214 | * will wait for the operation to proceed until the corresponding | |
215 | * bdrv_co_leave(). | |
216 | * | |
217 | * Consequently, you can't call drain inside a bdrv_co_enter/leave() section as | |
218 | * this will deadlock. | |
219 | */ | |
220 | AioContext *coroutine_fn bdrv_co_enter(BlockDriverState *bs); | |
221 | ||
222 | /** | |
223 | * Ends a section started by bdrv_co_enter(). Move the current coroutine back | |
224 | * to old_ctx and decrease bs->in_flight again. | |
225 | */ | |
226 | void coroutine_fn bdrv_co_leave(BlockDriverState *bs, AioContext *old_ctx); | |
227 | ||
3b491a90 EGE |
228 | AioContext *child_of_bds_get_parent_aio_context(BdrvChild *c); |
229 | ||
8f497454 | 230 | void coroutine_fn bdrv_co_io_plug(BlockDriverState *bs); |
09d9fc97 | 231 | void coroutine_fn bdrv_co_io_unplug(BlockDriverState *bs); |
3b491a90 | 232 | |
0508d0be EGE |
233 | bool coroutine_fn bdrv_co_can_store_new_dirty_bitmap(BlockDriverState *bs, |
234 | const char *name, | |
235 | uint32_t granularity, | |
236 | Error **errp); | |
237 | bool co_wrapper bdrv_can_store_new_dirty_bitmap(BlockDriverState *bs, | |
238 | const char *name, | |
239 | uint32_t granularity, | |
240 | Error **errp); | |
3b491a90 EGE |
241 | |
242 | /** | |
243 | * | |
244 | * bdrv_co_copy_range: | |
245 | * | |
246 | * Do offloaded copy between two children. If the operation is not implemented | |
247 | * by the driver, or if the backend storage doesn't support it, a negative | |
248 | * error code will be returned. | |
249 | * | |
250 | * Note: block layer doesn't emulate or fallback to a bounce buffer approach | |
251 | * because usually the caller shouldn't attempt offloaded copy any more (e.g. | |
252 | * calling copy_file_range(2)) after the first error, thus it should fall back | |
253 | * to a read+write path in the caller level. | |
254 | * | |
255 | * @src: Source child to copy data from | |
256 | * @src_offset: offset in @src image to read data | |
257 | * @dst: Destination child to copy data to | |
258 | * @dst_offset: offset in @dst image to write data | |
259 | * @bytes: number of bytes to copy | |
260 | * @flags: request flags. Supported flags: | |
261 | * BDRV_REQ_ZERO_WRITE - treat the @src range as zero data and do zero | |
262 | * write on @dst as if bdrv_co_pwrite_zeroes is | |
263 | * called. Used to simplify caller code, or | |
264 | * during BlockDriver.bdrv_co_copy_range_from() | |
265 | * recursion. | |
266 | * BDRV_REQ_NO_SERIALISING - do not serialize with other overlapping | |
267 | * requests currently in flight. | |
268 | * | |
269 | * Returns: 0 if succeeded; negative error code if failed. | |
270 | **/ | |
271 | int coroutine_fn bdrv_co_copy_range(BdrvChild *src, int64_t src_offset, | |
272 | BdrvChild *dst, int64_t dst_offset, | |
273 | int64_t bytes, BdrvRequestFlags read_flags, | |
274 | BdrvRequestFlags write_flags); | |
275 | ||
3b491a90 EGE |
276 | /* |
277 | * "I/O or GS" API functions. These functions can run without | |
278 | * the BQL, but only in one specific iothread/main loop. | |
279 | * | |
280 | * More specifically, these functions use BDRV_POLL_WHILE(bs), which | |
281 | * requires the caller to be either in the main thread and hold | |
282 | * the BlockdriverState (bs) AioContext lock, or directly in the | |
283 | * home thread that runs the bs AioContext. Calling them from | |
284 | * another thread in another AioContext would cause deadlocks. | |
285 | * | |
286 | * Therefore, these functions are not proper I/O, because they | |
287 | * can't run in *any* iothreads, but only in a specific one. | |
288 | * | |
289 | * These functions can call any function from I/O, Common and this | |
290 | * categories, but must be invoked only by other "I/O or GS" and GS APIs. | |
291 | * | |
292 | * All functions in this category must use the macro | |
293 | * IO_OR_GS_CODE(); | |
294 | * to catch when they are accidentally called by the wrong API. | |
295 | */ | |
296 | ||
297 | #define BDRV_POLL_WHILE(bs, cond) ({ \ | |
298 | BlockDriverState *bs_ = (bs); \ | |
384a48fb | 299 | IO_OR_GS_CODE(); \ |
3b491a90 EGE |
300 | AIO_WAIT_WHILE(bdrv_get_aio_context(bs_), \ |
301 | cond); }) | |
302 | ||
303 | void bdrv_drain(BlockDriverState *bs); | |
3b491a90 | 304 | |
90830f59 | 305 | int co_wrapper_mixed_bdrv_rdlock |
3b491a90 EGE |
306 | bdrv_truncate(BdrvChild *child, int64_t offset, bool exact, |
307 | PreallocMode prealloc, BdrvRequestFlags flags, Error **errp); | |
308 | ||
90830f59 EGE |
309 | int co_wrapper_mixed_bdrv_rdlock |
310 | bdrv_check(BlockDriverState *bs, BdrvCheckResult *res, BdrvCheckMode fix); | |
3b491a90 EGE |
311 | |
312 | /* Invalidate any cached metadata used by image formats */ | |
90830f59 EGE |
313 | int co_wrapper_mixed_bdrv_rdlock |
314 | bdrv_invalidate_cache(BlockDriverState *bs, Error **errp); | |
315 | ||
316 | int co_wrapper_mixed_bdrv_rdlock bdrv_flush(BlockDriverState *bs); | |
317 | ||
318 | int co_wrapper_mixed_bdrv_rdlock | |
319 | bdrv_pdiscard(BdrvChild *child, int64_t offset, int64_t bytes); | |
320 | ||
321 | int co_wrapper_mixed_bdrv_rdlock | |
3b491a90 | 322 | bdrv_readv_vmstate(BlockDriverState *bs, QEMUIOVector *qiov, int64_t pos); |
90830f59 EGE |
323 | |
324 | int co_wrapper_mixed_bdrv_rdlock | |
3b491a90 EGE |
325 | bdrv_writev_vmstate(BlockDriverState *bs, QEMUIOVector *qiov, int64_t pos); |
326 | ||
327 | /** | |
328 | * bdrv_parent_drained_begin_single: | |
329 | * | |
606ed756 | 330 | * Begin a quiesced section for the parent of @c. |
3b491a90 | 331 | */ |
606ed756 | 332 | void bdrv_parent_drained_begin_single(BdrvChild *c); |
3b491a90 | 333 | |
23987471 KW |
334 | /** |
335 | * bdrv_parent_drained_poll_single: | |
336 | * | |
337 | * Returns true if there is any pending activity to cease before @c can be | |
338 | * called quiesced, false otherwise. | |
339 | */ | |
340 | bool bdrv_parent_drained_poll_single(BdrvChild *c); | |
341 | ||
3b491a90 EGE |
342 | /** |
343 | * bdrv_parent_drained_end_single: | |
344 | * | |
345 | * End a quiesced section for the parent of @c. | |
3b491a90 EGE |
346 | */ |
347 | void bdrv_parent_drained_end_single(BdrvChild *c); | |
348 | ||
349 | /** | |
350 | * bdrv_drain_poll: | |
351 | * | |
299403ae | 352 | * Poll for pending requests in @bs and its parents (except for @ignore_parent). |
3b491a90 EGE |
353 | * |
354 | * If @ignore_bds_parents is true, parents that are BlockDriverStates must | |
355 | * ignore the drain request because they will be drained separately (used for | |
356 | * drain_all). | |
357 | * | |
358 | * This is part of bdrv_drained_begin. | |
359 | */ | |
299403ae KW |
360 | bool bdrv_drain_poll(BlockDriverState *bs, BdrvChild *ignore_parent, |
361 | bool ignore_bds_parents); | |
3b491a90 EGE |
362 | |
363 | /** | |
364 | * bdrv_drained_begin: | |
365 | * | |
366 | * Begin a quiesced section for exclusive access to the BDS, by disabling | |
367 | * external request sources including NBD server, block jobs, and device model. | |
368 | * | |
369 | * This function can be recursive. | |
370 | */ | |
371 | void bdrv_drained_begin(BlockDriverState *bs); | |
372 | ||
373 | /** | |
374 | * bdrv_do_drained_begin_quiesce: | |
375 | * | |
376 | * Quiesces a BDS like bdrv_drained_begin(), but does not wait for already | |
377 | * running requests to complete. | |
378 | */ | |
a82a3bd1 | 379 | void bdrv_do_drained_begin_quiesce(BlockDriverState *bs, BdrvChild *parent); |
3b491a90 | 380 | |
3b491a90 EGE |
381 | /** |
382 | * bdrv_drained_end: | |
383 | * | |
384 | * End a quiescent section started by bdrv_drained_begin(). | |
3b491a90 EGE |
385 | */ |
386 | void bdrv_drained_end(BlockDriverState *bs); | |
387 | ||
3b491a90 | 388 | #endif /* BLOCK_IO_H */ |