]>
Commit | Line | Data |
---|---|---|
1 | From 0000000000000000000000000000000000000000 Mon Sep 17 00:00:00 2001 | |
2 | From: Dietmar Maurer <dietmar@proxmox.com> | |
3 | Date: Mon, 6 Apr 2020 12:16:46 +0200 | |
4 | Subject: [PATCH] PVE: add savevm-async for background state snapshots | |
5 | ||
6 | Put qemu_savevm_state_{header,setup} into the main loop and the rest | |
7 | of the iteration into a coroutine. The former need to lock the | |
8 | iothread (and we can't unlock it in the coroutine), and the latter | |
9 | can't deal with being in a separate thread, so a coroutine it must | |
10 | be. | |
11 | ||
12 | Truncate output file at 1024 boundary. | |
13 | ||
14 | Do not block the VM and save the state on aborting a snapshot, as the | |
15 | snapshot will be invalid anyway. | |
16 | ||
17 | Also, when aborting, wait for the target file to be closed, otherwise a | |
18 | client might run into race-conditions when trying to remove the file | |
19 | still opened by QEMU. | |
20 | ||
21 | Signed-off-by: Thomas Lamprecht <t.lamprecht@proxmox.com> | |
22 | Signed-off-by: Dietmar Maurer <dietmar@proxmox.com> | |
23 | Signed-off-by: Wolfgang Bumiller <w.bumiller@proxmox.com> | |
24 | [improve aborting] | |
25 | Signed-off-by: Stefan Reiter <s.reiter@proxmox.com> | |
26 | [FE: further improve aborting | |
27 | adapt to removal of QEMUFileOps] | |
28 | Signed-off-by: Fiona Ebner <f.ebner@proxmox.com> | |
29 | --- | |
30 | hmp-commands-info.hx | 13 + | |
31 | hmp-commands.hx | 33 +++ | |
32 | include/migration/snapshot.h | 2 + | |
33 | include/monitor/hmp.h | 5 + | |
34 | migration/meson.build | 1 + | |
35 | migration/savevm-async.c | 531 +++++++++++++++++++++++++++++++++++ | |
36 | monitor/hmp-cmds.c | 57 ++++ | |
37 | qapi/migration.json | 34 +++ | |
38 | qapi/misc.json | 32 +++ | |
39 | qemu-options.hx | 12 + | |
40 | softmmu/vl.c | 10 + | |
41 | 11 files changed, 730 insertions(+) | |
42 | create mode 100644 migration/savevm-async.c | |
43 | ||
44 | diff --git a/hmp-commands-info.hx b/hmp-commands-info.hx | |
45 | index 754b1e8408..489c524e9e 100644 | |
46 | --- a/hmp-commands-info.hx | |
47 | +++ b/hmp-commands-info.hx | |
48 | @@ -540,6 +540,19 @@ SRST | |
49 | Show current migration parameters. | |
50 | ERST | |
51 | ||
52 | + { | |
53 | + .name = "savevm", | |
54 | + .args_type = "", | |
55 | + .params = "", | |
56 | + .help = "show savevm status", | |
57 | + .cmd = hmp_info_savevm, | |
58 | + }, | |
59 | + | |
60 | +SRST | |
61 | + ``info savevm`` | |
62 | + Show savevm status. | |
63 | +ERST | |
64 | + | |
65 | { | |
66 | .name = "balloon", | |
67 | .args_type = "", | |
68 | diff --git a/hmp-commands.hx b/hmp-commands.hx | |
69 | index 673e39a697..039be0033d 100644 | |
70 | --- a/hmp-commands.hx | |
71 | +++ b/hmp-commands.hx | |
72 | @@ -1815,3 +1815,36 @@ SRST | |
73 | Dump the FDT in dtb format to *filename*. | |
74 | ERST | |
75 | #endif | |
76 | + | |
77 | + { | |
78 | + .name = "savevm-start", | |
79 | + .args_type = "statefile:s?", | |
80 | + .params = "[statefile]", | |
81 | + .help = "Prepare for snapshot and halt VM. Save VM state to statefile.", | |
82 | + .cmd = hmp_savevm_start, | |
83 | + }, | |
84 | + | |
85 | + { | |
86 | + .name = "snapshot-drive", | |
87 | + .args_type = "device:s,name:s", | |
88 | + .params = "device name", | |
89 | + .help = "Create internal snapshot.", | |
90 | + .cmd = hmp_snapshot_drive, | |
91 | + }, | |
92 | + | |
93 | + { | |
94 | + .name = "delete-drive-snapshot", | |
95 | + .args_type = "device:s,name:s", | |
96 | + .params = "device name", | |
97 | + .help = "Delete internal snapshot.", | |
98 | + .cmd = hmp_delete_drive_snapshot, | |
99 | + }, | |
100 | + | |
101 | + { | |
102 | + .name = "savevm-end", | |
103 | + .args_type = "", | |
104 | + .params = "", | |
105 | + .help = "Resume VM after snaphot.", | |
106 | + .cmd = hmp_savevm_end, | |
107 | + .coroutine = true, | |
108 | + }, | |
109 | diff --git a/include/migration/snapshot.h b/include/migration/snapshot.h | |
110 | index e72083b117..c846d37806 100644 | |
111 | --- a/include/migration/snapshot.h | |
112 | +++ b/include/migration/snapshot.h | |
113 | @@ -61,4 +61,6 @@ bool delete_snapshot(const char *name, | |
114 | bool has_devices, strList *devices, | |
115 | Error **errp); | |
116 | ||
117 | +int load_snapshot_from_blockdev(const char *filename, Error **errp); | |
118 | + | |
119 | #endif | |
120 | diff --git a/include/monitor/hmp.h b/include/monitor/hmp.h | |
121 | index dfbc0c9a2f..440f86aba8 100644 | |
122 | --- a/include/monitor/hmp.h | |
123 | +++ b/include/monitor/hmp.h | |
124 | @@ -27,6 +27,7 @@ void hmp_info_status(Monitor *mon, const QDict *qdict); | |
125 | void hmp_info_uuid(Monitor *mon, const QDict *qdict); | |
126 | void hmp_info_chardev(Monitor *mon, const QDict *qdict); | |
127 | void hmp_info_mice(Monitor *mon, const QDict *qdict); | |
128 | +void hmp_info_savevm(Monitor *mon, const QDict *qdict); | |
129 | void hmp_info_migrate(Monitor *mon, const QDict *qdict); | |
130 | void hmp_info_migrate_capabilities(Monitor *mon, const QDict *qdict); | |
131 | void hmp_info_migrate_parameters(Monitor *mon, const QDict *qdict); | |
132 | @@ -81,6 +82,10 @@ void hmp_netdev_add(Monitor *mon, const QDict *qdict); | |
133 | void hmp_netdev_del(Monitor *mon, const QDict *qdict); | |
134 | void hmp_getfd(Monitor *mon, const QDict *qdict); | |
135 | void hmp_closefd(Monitor *mon, const QDict *qdict); | |
136 | +void hmp_savevm_start(Monitor *mon, const QDict *qdict); | |
137 | +void hmp_snapshot_drive(Monitor *mon, const QDict *qdict); | |
138 | +void hmp_delete_drive_snapshot(Monitor *mon, const QDict *qdict); | |
139 | +void hmp_savevm_end(Monitor *mon, const QDict *qdict); | |
140 | void hmp_sendkey(Monitor *mon, const QDict *qdict); | |
141 | void coroutine_fn hmp_screendump(Monitor *mon, const QDict *qdict); | |
142 | void hmp_chardev_add(Monitor *mon, const QDict *qdict); | |
143 | diff --git a/migration/meson.build b/migration/meson.build | |
144 | index 8cac83c06c..0842d00cd2 100644 | |
145 | --- a/migration/meson.build | |
146 | +++ b/migration/meson.build | |
147 | @@ -24,6 +24,7 @@ softmmu_ss.add(files( | |
148 | 'multifd-zlib.c', | |
149 | 'postcopy-ram.c', | |
150 | 'savevm.c', | |
151 | + 'savevm-async.c', | |
152 | 'socket.c', | |
153 | 'tls.c', | |
154 | ), gnutls) | |
155 | diff --git a/migration/savevm-async.c b/migration/savevm-async.c | |
156 | new file mode 100644 | |
157 | index 0000000000..05d394c0e2 | |
158 | --- /dev/null | |
159 | +++ b/migration/savevm-async.c | |
160 | @@ -0,0 +1,531 @@ | |
161 | +#include "qemu/osdep.h" | |
162 | +#include "migration/channel-savevm-async.h" | |
163 | +#include "migration/migration.h" | |
164 | +#include "migration/savevm.h" | |
165 | +#include "migration/snapshot.h" | |
166 | +#include "migration/global_state.h" | |
167 | +#include "migration/ram.h" | |
168 | +#include "migration/qemu-file.h" | |
169 | +#include "sysemu/sysemu.h" | |
170 | +#include "sysemu/runstate.h" | |
171 | +#include "block/block.h" | |
172 | +#include "sysemu/block-backend.h" | |
173 | +#include "qapi/error.h" | |
174 | +#include "qapi/qmp/qerror.h" | |
175 | +#include "qapi/qmp/qdict.h" | |
176 | +#include "qapi/qapi-commands-migration.h" | |
177 | +#include "qapi/qapi-commands-misc.h" | |
178 | +#include "qapi/qapi-commands-block.h" | |
179 | +#include "qemu/cutils.h" | |
180 | +#include "qemu/timer.h" | |
181 | +#include "qemu/main-loop.h" | |
182 | +#include "qemu/rcu.h" | |
183 | + | |
184 | +/* #define DEBUG_SAVEVM_STATE */ | |
185 | + | |
186 | +#ifdef DEBUG_SAVEVM_STATE | |
187 | +#define DPRINTF(fmt, ...) \ | |
188 | + do { printf("savevm-async: " fmt, ## __VA_ARGS__); } while (0) | |
189 | +#else | |
190 | +#define DPRINTF(fmt, ...) \ | |
191 | + do { } while (0) | |
192 | +#endif | |
193 | + | |
194 | +enum { | |
195 | + SAVE_STATE_DONE, | |
196 | + SAVE_STATE_ERROR, | |
197 | + SAVE_STATE_ACTIVE, | |
198 | + SAVE_STATE_COMPLETED, | |
199 | + SAVE_STATE_CANCELLED | |
200 | +}; | |
201 | + | |
202 | + | |
203 | +static struct SnapshotState { | |
204 | + BlockBackend *target; | |
205 | + size_t bs_pos; | |
206 | + int state; | |
207 | + Error *error; | |
208 | + Error *blocker; | |
209 | + int saved_vm_running; | |
210 | + QEMUFile *file; | |
211 | + int64_t total_time; | |
212 | + QEMUBH *finalize_bh; | |
213 | + Coroutine *co; | |
214 | + QemuCoSleep target_close_wait; | |
215 | +} snap_state; | |
216 | + | |
217 | +static bool savevm_aborted(void) | |
218 | +{ | |
219 | + return snap_state.state == SAVE_STATE_CANCELLED || | |
220 | + snap_state.state == SAVE_STATE_ERROR; | |
221 | +} | |
222 | + | |
223 | +SaveVMInfo *qmp_query_savevm(Error **errp) | |
224 | +{ | |
225 | + SaveVMInfo *info = g_malloc0(sizeof(*info)); | |
226 | + struct SnapshotState *s = &snap_state; | |
227 | + | |
228 | + if (s->state != SAVE_STATE_DONE) { | |
229 | + info->has_bytes = true; | |
230 | + info->bytes = s->bs_pos; | |
231 | + switch (s->state) { | |
232 | + case SAVE_STATE_ERROR: | |
233 | + info->has_status = true; | |
234 | + info->status = g_strdup("failed"); | |
235 | + info->has_total_time = true; | |
236 | + info->total_time = s->total_time; | |
237 | + if (s->error) { | |
238 | + info->has_error = true; | |
239 | + info->error = g_strdup(error_get_pretty(s->error)); | |
240 | + } | |
241 | + break; | |
242 | + case SAVE_STATE_ACTIVE: | |
243 | + info->has_status = true; | |
244 | + info->status = g_strdup("active"); | |
245 | + info->has_total_time = true; | |
246 | + info->total_time = qemu_clock_get_ms(QEMU_CLOCK_REALTIME) | |
247 | + - s->total_time; | |
248 | + break; | |
249 | + case SAVE_STATE_COMPLETED: | |
250 | + info->has_status = true; | |
251 | + info->status = g_strdup("completed"); | |
252 | + info->has_total_time = true; | |
253 | + info->total_time = s->total_time; | |
254 | + break; | |
255 | + } | |
256 | + } | |
257 | + | |
258 | + return info; | |
259 | +} | |
260 | + | |
261 | +static int save_snapshot_cleanup(void) | |
262 | +{ | |
263 | + int ret = 0; | |
264 | + | |
265 | + DPRINTF("save_snapshot_cleanup\n"); | |
266 | + | |
267 | + snap_state.total_time = qemu_clock_get_ms(QEMU_CLOCK_REALTIME) - | |
268 | + snap_state.total_time; | |
269 | + | |
270 | + if (snap_state.file) { | |
271 | + ret = qemu_fclose(snap_state.file); | |
272 | + snap_state.file = NULL; | |
273 | + } | |
274 | + | |
275 | + if (snap_state.target) { | |
276 | + if (!savevm_aborted()) { | |
277 | + /* try to truncate, but ignore errors (will fail on block devices). | |
278 | + * note1: bdrv_read() need whole blocks, so we need to round up | |
279 | + * note2: PVE requires 1024 (BDRV_SECTOR_SIZE*2) alignment | |
280 | + */ | |
281 | + size_t size = QEMU_ALIGN_UP(snap_state.bs_pos, BDRV_SECTOR_SIZE*2); | |
282 | + blk_truncate(snap_state.target, size, false, PREALLOC_MODE_OFF, 0, NULL); | |
283 | + } | |
284 | + blk_op_unblock_all(snap_state.target, snap_state.blocker); | |
285 | + error_free(snap_state.blocker); | |
286 | + snap_state.blocker = NULL; | |
287 | + blk_unref(snap_state.target); | |
288 | + snap_state.target = NULL; | |
289 | + | |
290 | + qemu_co_sleep_wake(&snap_state.target_close_wait); | |
291 | + } | |
292 | + | |
293 | + return ret; | |
294 | +} | |
295 | + | |
296 | +static void save_snapshot_error(const char *fmt, ...) | |
297 | +{ | |
298 | + va_list ap; | |
299 | + char *msg; | |
300 | + | |
301 | + va_start(ap, fmt); | |
302 | + msg = g_strdup_vprintf(fmt, ap); | |
303 | + va_end(ap); | |
304 | + | |
305 | + DPRINTF("save_snapshot_error: %s\n", msg); | |
306 | + | |
307 | + if (!snap_state.error) { | |
308 | + error_set(&snap_state.error, ERROR_CLASS_GENERIC_ERROR, "%s", msg); | |
309 | + } | |
310 | + | |
311 | + g_free (msg); | |
312 | + | |
313 | + snap_state.state = SAVE_STATE_ERROR; | |
314 | +} | |
315 | + | |
316 | +static void process_savevm_finalize(void *opaque) | |
317 | +{ | |
318 | + int ret; | |
319 | + AioContext *iohandler_ctx = iohandler_get_aio_context(); | |
320 | + MigrationState *ms = migrate_get_current(); | |
321 | + | |
322 | + bool aborted = savevm_aborted(); | |
323 | + | |
324 | +#ifdef DEBUG_SAVEVM_STATE | |
325 | + int64_t start_time = qemu_clock_get_ms(QEMU_CLOCK_REALTIME); | |
326 | +#endif | |
327 | + | |
328 | + qemu_bh_delete(snap_state.finalize_bh); | |
329 | + snap_state.finalize_bh = NULL; | |
330 | + snap_state.co = NULL; | |
331 | + | |
332 | + /* We need to own the target bdrv's context for the following functions, | |
333 | + * so move it back. It can stay in the main context and live out its live | |
334 | + * there, since we're done with it after this method ends anyway. | |
335 | + */ | |
336 | + aio_context_acquire(iohandler_ctx); | |
337 | + blk_set_aio_context(snap_state.target, qemu_get_aio_context(), NULL); | |
338 | + aio_context_release(iohandler_ctx); | |
339 | + | |
340 | + ret = vm_stop_force_state(RUN_STATE_FINISH_MIGRATE); | |
341 | + if (ret < 0) { | |
342 | + save_snapshot_error("vm_stop_force_state error %d", ret); | |
343 | + } | |
344 | + | |
345 | + if (!aborted) { | |
346 | + /* skip state saving if we aborted, snapshot will be invalid anyway */ | |
347 | + (void)qemu_savevm_state_complete_precopy(snap_state.file, false, false); | |
348 | + ret = qemu_file_get_error(snap_state.file); | |
349 | + if (ret < 0) { | |
350 | + save_snapshot_error("qemu_savevm_state_complete_precopy error %d", ret); | |
351 | + } | |
352 | + } | |
353 | + | |
354 | + DPRINTF("state saving complete\n"); | |
355 | + DPRINTF("timing: process_savevm_finalize (state saving) took %ld ms\n", | |
356 | + qemu_clock_get_ms(QEMU_CLOCK_REALTIME) - start_time); | |
357 | + | |
358 | + /* clear migration state */ | |
359 | + migrate_set_state(&ms->state, MIGRATION_STATUS_SETUP, | |
360 | + ret || aborted ? MIGRATION_STATUS_FAILED : MIGRATION_STATUS_COMPLETED); | |
361 | + ms->to_dst_file = NULL; | |
362 | + | |
363 | + qemu_savevm_state_cleanup(); | |
364 | + | |
365 | + ret = save_snapshot_cleanup(); | |
366 | + if (ret < 0) { | |
367 | + save_snapshot_error("save_snapshot_cleanup error %d", ret); | |
368 | + } else if (snap_state.state == SAVE_STATE_ACTIVE) { | |
369 | + snap_state.state = SAVE_STATE_COMPLETED; | |
370 | + } else if (aborted) { | |
371 | + /* | |
372 | + * If there was an error, there's no need to set a new one here. | |
373 | + * If the snapshot was canceled, leave setting the state to | |
374 | + * qmp_savevm_end(), which is waked by save_snapshot_cleanup(). | |
375 | + */ | |
376 | + } else { | |
377 | + save_snapshot_error("process_savevm_cleanup: invalid state: %d", | |
378 | + snap_state.state); | |
379 | + } | |
380 | + if (snap_state.saved_vm_running) { | |
381 | + vm_start(); | |
382 | + snap_state.saved_vm_running = false; | |
383 | + } | |
384 | + | |
385 | + DPRINTF("timing: process_savevm_finalize (full) took %ld ms\n", | |
386 | + qemu_clock_get_ms(QEMU_CLOCK_REALTIME) - start_time); | |
387 | +} | |
388 | + | |
389 | +static void coroutine_fn process_savevm_co(void *opaque) | |
390 | +{ | |
391 | + int ret; | |
392 | + int64_t maxlen; | |
393 | + BdrvNextIterator it; | |
394 | + BlockDriverState *bs = NULL; | |
395 | + | |
396 | +#ifdef DEBUG_SAVEVM_STATE | |
397 | + int64_t start_time = qemu_clock_get_ms(QEMU_CLOCK_REALTIME); | |
398 | +#endif | |
399 | + | |
400 | + ret = qemu_file_get_error(snap_state.file); | |
401 | + if (ret < 0) { | |
402 | + save_snapshot_error("qemu_savevm_state_setup failed"); | |
403 | + return; | |
404 | + } | |
405 | + | |
406 | + while (snap_state.state == SAVE_STATE_ACTIVE) { | |
407 | + uint64_t pending_size, pend_precopy, pend_compatible, pend_postcopy; | |
408 | + | |
409 | + /* pending is expected to be called without iothread lock */ | |
410 | + qemu_mutex_unlock_iothread(); | |
411 | + qemu_savevm_state_pending(snap_state.file, 0, &pend_precopy, &pend_compatible, &pend_postcopy); | |
412 | + qemu_mutex_lock_iothread(); | |
413 | + | |
414 | + pending_size = pend_precopy + pend_compatible + pend_postcopy; | |
415 | + | |
416 | + maxlen = blk_getlength(snap_state.target) - 30*1024*1024; | |
417 | + | |
418 | + if (pending_size > 400000 && snap_state.bs_pos + pending_size < maxlen) { | |
419 | + ret = qemu_savevm_state_iterate(snap_state.file, false); | |
420 | + if (ret < 0) { | |
421 | + save_snapshot_error("qemu_savevm_state_iterate error %d", ret); | |
422 | + break; | |
423 | + } | |
424 | + DPRINTF("savevm iterate pending size %lu ret %d\n", pending_size, ret); | |
425 | + } else { | |
426 | + qemu_system_wakeup_request(QEMU_WAKEUP_REASON_OTHER, NULL); | |
427 | + ret = global_state_store(); | |
428 | + if (ret) { | |
429 | + save_snapshot_error("global_state_store error %d", ret); | |
430 | + break; | |
431 | + } | |
432 | + | |
433 | + DPRINTF("savevm iterate complete\n"); | |
434 | + break; | |
435 | + } | |
436 | + } | |
437 | + | |
438 | + DPRINTF("timing: process_savevm_co took %ld ms\n", | |
439 | + qemu_clock_get_ms(QEMU_CLOCK_REALTIME) - start_time); | |
440 | + | |
441 | +#ifdef DEBUG_SAVEVM_STATE | |
442 | + int64_t start_time_flush = qemu_clock_get_ms(QEMU_CLOCK_REALTIME); | |
443 | +#endif | |
444 | + /* If a drive runs in an IOThread we can flush it async, and only | |
445 | + * need to sync-flush whatever IO happens between now and | |
446 | + * vm_stop_force_state. bdrv_next can only be called from main AioContext, | |
447 | + * so move there now and after every flush. | |
448 | + */ | |
449 | + aio_co_reschedule_self(qemu_get_aio_context()); | |
450 | + for (bs = bdrv_first(&it); bs; bs = bdrv_next(&it)) { | |
451 | + /* target has BDRV_O_NO_FLUSH, no sense calling bdrv_flush on it */ | |
452 | + if (bs == blk_bs(snap_state.target)) { | |
453 | + continue; | |
454 | + } | |
455 | + | |
456 | + AioContext *bs_ctx = bdrv_get_aio_context(bs); | |
457 | + if (bs_ctx != qemu_get_aio_context()) { | |
458 | + DPRINTF("savevm: async flushing drive %s\n", bs->filename); | |
459 | + aio_co_reschedule_self(bs_ctx); | |
460 | + bdrv_flush(bs); | |
461 | + aio_co_reschedule_self(qemu_get_aio_context()); | |
462 | + } | |
463 | + } | |
464 | + | |
465 | + DPRINTF("timing: async flushing took %ld ms\n", | |
466 | + qemu_clock_get_ms(QEMU_CLOCK_REALTIME) - start_time_flush); | |
467 | + | |
468 | + qemu_bh_schedule(snap_state.finalize_bh); | |
469 | +} | |
470 | + | |
471 | +void qmp_savevm_start(bool has_statefile, const char *statefile, Error **errp) | |
472 | +{ | |
473 | + Error *local_err = NULL; | |
474 | + MigrationState *ms = migrate_get_current(); | |
475 | + AioContext *iohandler_ctx = iohandler_get_aio_context(); | |
476 | + | |
477 | + int bdrv_oflags = BDRV_O_RDWR | BDRV_O_RESIZE | BDRV_O_NO_FLUSH; | |
478 | + | |
479 | + if (snap_state.state != SAVE_STATE_DONE) { | |
480 | + error_set(errp, ERROR_CLASS_GENERIC_ERROR, | |
481 | + "VM snapshot already started\n"); | |
482 | + return; | |
483 | + } | |
484 | + | |
485 | + if (migration_is_running(ms->state)) { | |
486 | + error_set(errp, ERROR_CLASS_GENERIC_ERROR, QERR_MIGRATION_ACTIVE); | |
487 | + return; | |
488 | + } | |
489 | + | |
490 | + if (migrate_use_block()) { | |
491 | + error_set(errp, ERROR_CLASS_GENERIC_ERROR, | |
492 | + "Block migration and snapshots are incompatible"); | |
493 | + return; | |
494 | + } | |
495 | + | |
496 | + /* initialize snapshot info */ | |
497 | + snap_state.saved_vm_running = runstate_is_running(); | |
498 | + snap_state.bs_pos = 0; | |
499 | + snap_state.total_time = qemu_clock_get_ms(QEMU_CLOCK_REALTIME); | |
500 | + snap_state.blocker = NULL; | |
501 | + snap_state.target_close_wait = (QemuCoSleep){ .to_wake = NULL }; | |
502 | + | |
503 | + if (snap_state.error) { | |
504 | + error_free(snap_state.error); | |
505 | + snap_state.error = NULL; | |
506 | + } | |
507 | + | |
508 | + if (!has_statefile) { | |
509 | + vm_stop(RUN_STATE_SAVE_VM); | |
510 | + snap_state.state = SAVE_STATE_COMPLETED; | |
511 | + return; | |
512 | + } | |
513 | + | |
514 | + if (qemu_savevm_state_blocked(errp)) { | |
515 | + return; | |
516 | + } | |
517 | + | |
518 | + /* Open the image */ | |
519 | + QDict *options = NULL; | |
520 | + options = qdict_new(); | |
521 | + qdict_put_str(options, "driver", "raw"); | |
522 | + snap_state.target = blk_new_open(statefile, NULL, options, bdrv_oflags, &local_err); | |
523 | + if (!snap_state.target) { | |
524 | + error_set(errp, ERROR_CLASS_GENERIC_ERROR, "failed to open '%s'", statefile); | |
525 | + goto restart; | |
526 | + } | |
527 | + | |
528 | + QIOChannel *ioc = QIO_CHANNEL(qio_channel_savevm_async_new(snap_state.target, | |
529 | + &snap_state.bs_pos)); | |
530 | + snap_state.file = qemu_file_new_output(ioc); | |
531 | + | |
532 | + if (!snap_state.file) { | |
533 | + error_set(errp, ERROR_CLASS_GENERIC_ERROR, "failed to open '%s'", statefile); | |
534 | + goto restart; | |
535 | + } | |
536 | + | |
537 | + /* | |
538 | + * qemu_savevm_* paths use migration code and expect a migration state. | |
539 | + * State is cleared in process_savevm_co, but has to be initialized | |
540 | + * here (blocking main thread, from QMP) to avoid race conditions. | |
541 | + */ | |
542 | + migrate_init(ms); | |
543 | + memset(&ram_counters, 0, sizeof(ram_counters)); | |
544 | + ms->to_dst_file = snap_state.file; | |
545 | + | |
546 | + error_setg(&snap_state.blocker, "block device is in use by savevm"); | |
547 | + blk_op_block_all(snap_state.target, snap_state.blocker); | |
548 | + | |
549 | + snap_state.state = SAVE_STATE_ACTIVE; | |
550 | + snap_state.finalize_bh = qemu_bh_new(process_savevm_finalize, &snap_state); | |
551 | + snap_state.co = qemu_coroutine_create(&process_savevm_co, NULL); | |
552 | + qemu_mutex_unlock_iothread(); | |
553 | + qemu_savevm_state_header(snap_state.file); | |
554 | + qemu_savevm_state_setup(snap_state.file); | |
555 | + qemu_mutex_lock_iothread(); | |
556 | + | |
557 | + /* Async processing from here on out happens in iohandler context, so let | |
558 | + * the target bdrv have its home there. | |
559 | + */ | |
560 | + blk_set_aio_context(snap_state.target, iohandler_ctx, &local_err); | |
561 | + | |
562 | + aio_co_schedule(iohandler_ctx, snap_state.co); | |
563 | + | |
564 | + return; | |
565 | + | |
566 | +restart: | |
567 | + | |
568 | + save_snapshot_error("setup failed"); | |
569 | + | |
570 | + if (snap_state.saved_vm_running) { | |
571 | + vm_start(); | |
572 | + snap_state.saved_vm_running = false; | |
573 | + } | |
574 | +} | |
575 | + | |
576 | +void coroutine_fn qmp_savevm_end(Error **errp) | |
577 | +{ | |
578 | + int64_t timeout; | |
579 | + | |
580 | + if (snap_state.state == SAVE_STATE_DONE) { | |
581 | + error_set(errp, ERROR_CLASS_GENERIC_ERROR, | |
582 | + "VM snapshot not started\n"); | |
583 | + return; | |
584 | + } | |
585 | + | |
586 | + if (snap_state.state == SAVE_STATE_ACTIVE) { | |
587 | + snap_state.state = SAVE_STATE_CANCELLED; | |
588 | + goto wait_for_close; | |
589 | + } | |
590 | + | |
591 | + if (snap_state.saved_vm_running) { | |
592 | + vm_start(); | |
593 | + snap_state.saved_vm_running = false; | |
594 | + } | |
595 | + | |
596 | + snap_state.state = SAVE_STATE_DONE; | |
597 | + | |
598 | +wait_for_close: | |
599 | + if (!snap_state.target) { | |
600 | + DPRINTF("savevm-end: no target file open\n"); | |
601 | + return; | |
602 | + } | |
603 | + | |
604 | + /* wait until cleanup is done before returning, this ensures that after this | |
605 | + * call exits the statefile will be closed and can be removed immediately */ | |
606 | + DPRINTF("savevm-end: waiting for cleanup\n"); | |
607 | + timeout = 30L * 1000 * 1000 * 1000; | |
608 | + qemu_co_sleep_ns_wakeable(&snap_state.target_close_wait, | |
609 | + QEMU_CLOCK_REALTIME, timeout); | |
610 | + if (snap_state.target) { | |
611 | + save_snapshot_error("timeout waiting for target file close in " | |
612 | + "qmp_savevm_end"); | |
613 | + /* we cannot assume the snapshot finished in this case, so leave the | |
614 | + * state alone - caller has to figure something out */ | |
615 | + return; | |
616 | + } | |
617 | + | |
618 | + // File closed and no other error, so ensure next snapshot can be started. | |
619 | + if (snap_state.state != SAVE_STATE_ERROR) { | |
620 | + snap_state.state = SAVE_STATE_DONE; | |
621 | + } | |
622 | + | |
623 | + DPRINTF("savevm-end: cleanup done\n"); | |
624 | +} | |
625 | + | |
626 | +// FIXME: Deprecated | |
627 | +void qmp_snapshot_drive(const char *device, const char *name, Error **errp) | |
628 | +{ | |
629 | + // Compatibility to older qemu-server. | |
630 | + qmp_blockdev_snapshot_internal_sync(device, name, errp); | |
631 | +} | |
632 | + | |
633 | +// FIXME: Deprecated | |
634 | +void qmp_delete_drive_snapshot(const char *device, const char *name, | |
635 | + Error **errp) | |
636 | +{ | |
637 | + // Compatibility to older qemu-server. | |
638 | + (void)qmp_blockdev_snapshot_delete_internal_sync(device, false, NULL, | |
639 | + true, name, errp); | |
640 | +} | |
641 | + | |
642 | +int load_snapshot_from_blockdev(const char *filename, Error **errp) | |
643 | +{ | |
644 | + BlockBackend *be; | |
645 | + Error *local_err = NULL; | |
646 | + Error *blocker = NULL; | |
647 | + | |
648 | + QEMUFile *f; | |
649 | + size_t bs_pos = 0; | |
650 | + int ret = -EINVAL; | |
651 | + | |
652 | + be = blk_new_open(filename, NULL, NULL, 0, &local_err); | |
653 | + | |
654 | + if (!be) { | |
655 | + error_setg(errp, "Could not open VM state file"); | |
656 | + goto the_end; | |
657 | + } | |
658 | + | |
659 | + error_setg(&blocker, "block device is in use by load state"); | |
660 | + blk_op_block_all(be, blocker); | |
661 | + | |
662 | + /* restore the VM state */ | |
663 | + f = qemu_file_new_input(QIO_CHANNEL(qio_channel_savevm_async_new(be, &bs_pos))); | |
664 | + if (!f) { | |
665 | + error_setg(errp, "Could not open VM state file"); | |
666 | + goto the_end; | |
667 | + } | |
668 | + | |
669 | + qemu_system_reset(SHUTDOWN_CAUSE_NONE); | |
670 | + ret = qemu_loadvm_state(f); | |
671 | + | |
672 | + /* dirty bitmap migration has a special case we need to trigger manually */ | |
673 | + dirty_bitmap_mig_before_vm_start(); | |
674 | + | |
675 | + qemu_fclose(f); | |
676 | + migration_incoming_state_destroy(); | |
677 | + if (ret < 0) { | |
678 | + error_setg_errno(errp, -ret, "Error while loading VM state"); | |
679 | + goto the_end; | |
680 | + } | |
681 | + | |
682 | + ret = 0; | |
683 | + | |
684 | + the_end: | |
685 | + if (be) { | |
686 | + blk_op_unblock_all(be, blocker); | |
687 | + error_free(blocker); | |
688 | + blk_unref(be); | |
689 | + } | |
690 | + return ret; | |
691 | +} | |
692 | diff --git a/monitor/hmp-cmds.c b/monitor/hmp-cmds.c | |
693 | index 480b798963..cfebfd1db5 100644 | |
694 | --- a/monitor/hmp-cmds.c | |
695 | +++ b/monitor/hmp-cmds.c | |
696 | @@ -1906,6 +1906,63 @@ void hmp_info_memory_devices(Monitor *mon, const QDict *qdict) | |
697 | hmp_handle_error(mon, err); | |
698 | } | |
699 | ||
700 | +void hmp_savevm_start(Monitor *mon, const QDict *qdict) | |
701 | +{ | |
702 | + Error *errp = NULL; | |
703 | + const char *statefile = qdict_get_try_str(qdict, "statefile"); | |
704 | + | |
705 | + qmp_savevm_start(statefile != NULL, statefile, &errp); | |
706 | + hmp_handle_error(mon, errp); | |
707 | +} | |
708 | + | |
709 | +void hmp_snapshot_drive(Monitor *mon, const QDict *qdict) | |
710 | +{ | |
711 | + Error *errp = NULL; | |
712 | + const char *name = qdict_get_str(qdict, "name"); | |
713 | + const char *device = qdict_get_str(qdict, "device"); | |
714 | + | |
715 | + qmp_snapshot_drive(device, name, &errp); | |
716 | + hmp_handle_error(mon, errp); | |
717 | +} | |
718 | + | |
719 | +void hmp_delete_drive_snapshot(Monitor *mon, const QDict *qdict) | |
720 | +{ | |
721 | + Error *errp = NULL; | |
722 | + const char *name = qdict_get_str(qdict, "name"); | |
723 | + const char *device = qdict_get_str(qdict, "device"); | |
724 | + | |
725 | + qmp_delete_drive_snapshot(device, name, &errp); | |
726 | + hmp_handle_error(mon, errp); | |
727 | +} | |
728 | + | |
729 | +void coroutine_fn hmp_savevm_end(Monitor *mon, const QDict *qdict) | |
730 | +{ | |
731 | + Error *errp = NULL; | |
732 | + | |
733 | + qmp_savevm_end(&errp); | |
734 | + hmp_handle_error(mon, errp); | |
735 | +} | |
736 | + | |
737 | +void hmp_info_savevm(Monitor *mon, const QDict *qdict) | |
738 | +{ | |
739 | + SaveVMInfo *info; | |
740 | + info = qmp_query_savevm(NULL); | |
741 | + | |
742 | + if (info->has_status) { | |
743 | + monitor_printf(mon, "savevm status: %s\n", info->status); | |
744 | + monitor_printf(mon, "total time: %" PRIu64 " milliseconds\n", | |
745 | + info->total_time); | |
746 | + } else { | |
747 | + monitor_printf(mon, "savevm status: not running\n"); | |
748 | + } | |
749 | + if (info->has_bytes) { | |
750 | + monitor_printf(mon, "Bytes saved: %"PRIu64"\n", info->bytes); | |
751 | + } | |
752 | + if (info->has_error) { | |
753 | + monitor_printf(mon, "Error: %s\n", info->error); | |
754 | + } | |
755 | +} | |
756 | + | |
757 | void hmp_info_iothreads(Monitor *mon, const QDict *qdict) | |
758 | { | |
759 | IOThreadInfoList *info_list = qmp_query_iothreads(NULL); | |
760 | diff --git a/qapi/migration.json b/qapi/migration.json | |
761 | index 88ecf86ac8..4435866379 100644 | |
762 | --- a/qapi/migration.json | |
763 | +++ b/qapi/migration.json | |
764 | @@ -261,6 +261,40 @@ | |
765 | '*compression': 'CompressionStats', | |
766 | '*socket-address': ['SocketAddress'] } } | |
767 | ||
768 | +## | |
769 | +# @SaveVMInfo: | |
770 | +# | |
771 | +# Information about current migration process. | |
772 | +# | |
773 | +# @status: string describing the current savevm status. | |
774 | +# This can be 'active', 'completed', 'failed'. | |
775 | +# If this field is not returned, no savevm process | |
776 | +# has been initiated | |
777 | +# | |
778 | +# @error: string containing error message is status is failed. | |
779 | +# | |
780 | +# @total-time: total amount of milliseconds since savevm started. | |
781 | +# If savevm has ended, it returns the total save time | |
782 | +# | |
783 | +# @bytes: total amount of data transfered | |
784 | +# | |
785 | +# Since: 1.3 | |
786 | +## | |
787 | +{ 'struct': 'SaveVMInfo', | |
788 | + 'data': {'*status': 'str', '*error': 'str', | |
789 | + '*total-time': 'int', '*bytes': 'int'} } | |
790 | + | |
791 | +## | |
792 | +# @query-savevm: | |
793 | +# | |
794 | +# Returns information about current savevm process. | |
795 | +# | |
796 | +# Returns: @SaveVMInfo | |
797 | +# | |
798 | +# Since: 1.3 | |
799 | +## | |
800 | +{ 'command': 'query-savevm', 'returns': 'SaveVMInfo' } | |
801 | + | |
802 | ## | |
803 | # @query-migrate: | |
804 | # | |
805 | diff --git a/qapi/misc.json b/qapi/misc.json | |
806 | index 27ef5a2b20..b3ce75dcae 100644 | |
807 | --- a/qapi/misc.json | |
808 | +++ b/qapi/misc.json | |
809 | @@ -435,6 +435,38 @@ | |
810 | ## | |
811 | { 'command': 'query-fdsets', 'returns': ['FdsetInfo'] } | |
812 | ||
813 | +## | |
814 | +# @savevm-start: | |
815 | +# | |
816 | +# Prepare for snapshot and halt VM. Save VM state to statefile. | |
817 | +# | |
818 | +## | |
819 | +{ 'command': 'savevm-start', 'data': { '*statefile': 'str' } } | |
820 | + | |
821 | +## | |
822 | +# @snapshot-drive: | |
823 | +# | |
824 | +# Create an internal drive snapshot. | |
825 | +# | |
826 | +## | |
827 | +{ 'command': 'snapshot-drive', 'data': { 'device': 'str', 'name': 'str' } } | |
828 | + | |
829 | +## | |
830 | +# @delete-drive-snapshot: | |
831 | +# | |
832 | +# Delete a drive snapshot. | |
833 | +# | |
834 | +## | |
835 | +{ 'command': 'delete-drive-snapshot', 'data': { 'device': 'str', 'name': 'str' } } | |
836 | + | |
837 | +## | |
838 | +# @savevm-end: | |
839 | +# | |
840 | +# Resume VM after a snapshot. | |
841 | +# | |
842 | +## | |
843 | +{ 'command': 'savevm-end', 'coroutine': true } | |
844 | + | |
845 | ## | |
846 | # @CommandLineParameterType: | |
847 | # | |
848 | diff --git a/qemu-options.hx b/qemu-options.hx | |
849 | index 7f99d15b23..54efb127c4 100644 | |
850 | --- a/qemu-options.hx | |
851 | +++ b/qemu-options.hx | |
852 | @@ -4391,6 +4391,18 @@ SRST | |
853 | Start right away with a saved state (``loadvm`` in monitor) | |
854 | ERST | |
855 | ||
856 | +DEF("loadstate", HAS_ARG, QEMU_OPTION_loadstate, \ | |
857 | + "-loadstate file\n" \ | |
858 | + " start right away with a saved state\n", | |
859 | + QEMU_ARCH_ALL) | |
860 | +SRST | |
861 | +``-loadstate file`` | |
862 | + Start right away with a saved state. This option does not rollback | |
863 | + disk state like @code{loadvm}, so user must make sure that disk | |
864 | + have correct state. @var{file} can be any valid device URL. See the section | |
865 | + for "Device URL Syntax" for more information. | |
866 | +ERST | |
867 | + | |
868 | #ifndef _WIN32 | |
869 | DEF("daemonize", 0, QEMU_OPTION_daemonize, \ | |
870 | "-daemonize daemonize QEMU after initializing\n", QEMU_ARCH_ALL) | |
871 | diff --git a/softmmu/vl.c b/softmmu/vl.c | |
872 | index 5f7f6ca981..21f067d115 100644 | |
873 | --- a/softmmu/vl.c | |
874 | +++ b/softmmu/vl.c | |
875 | @@ -164,6 +164,7 @@ static const char *accelerators; | |
876 | static bool have_custom_ram_size; | |
877 | static const char *ram_memdev_id; | |
878 | static QDict *machine_opts_dict; | |
879 | +static const char *loadstate; | |
880 | static QTAILQ_HEAD(, ObjectOption) object_opts = QTAILQ_HEAD_INITIALIZER(object_opts); | |
881 | static QTAILQ_HEAD(, DeviceOption) device_opts = QTAILQ_HEAD_INITIALIZER(device_opts); | |
882 | static int display_remote; | |
883 | @@ -2607,6 +2608,12 @@ void qmp_x_exit_preconfig(Error **errp) | |
884 | ||
885 | if (loadvm) { | |
886 | load_snapshot(loadvm, NULL, false, NULL, &error_fatal); | |
887 | + } else if (loadstate) { | |
888 | + Error *local_err = NULL; | |
889 | + if (load_snapshot_from_blockdev(loadstate, &local_err) < 0) { | |
890 | + error_report_err(local_err); | |
891 | + autostart = 0; | |
892 | + } | |
893 | } | |
894 | if (replay_mode != REPLAY_MODE_NONE) { | |
895 | replay_vmstate_init(); | |
896 | @@ -3151,6 +3158,9 @@ void qemu_init(int argc, char **argv) | |
897 | case QEMU_OPTION_loadvm: | |
898 | loadvm = optarg; | |
899 | break; | |
900 | + case QEMU_OPTION_loadstate: | |
901 | + loadstate = optarg; | |
902 | + break; | |
903 | case QEMU_OPTION_full_screen: | |
904 | dpy.has_full_screen = true; | |
905 | dpy.full_screen = true; |