]> git.proxmox.com Git - pve-qemu-kvm.git/blobdiff - debian/patches/pve/0046-convert-savevm-async-to-threads.patch
fix backup jobs for multiple disks
[pve-qemu-kvm.git] / debian / patches / pve / 0046-convert-savevm-async-to-threads.patch
index 6a9d9f4310fb2add2bbde1507a99621f73a9e61e..e216979e0830999c343c03599a42a73da02868ec 100644 (file)
@@ -1,14 +1,14 @@
-From f18d2aee91bca252a6b90583784f49236ec17d58 Mon Sep 17 00:00:00 2001
+From 190e9321e1657ec0b956ecece21d6a037487cd14 Mon Sep 17 00:00:00 2001
 From: Wolfgang Bumiller <w.bumiller@proxmox.com>
 Date: Tue, 8 Nov 2016 11:13:06 +0100
-Subject: [PATCH 46/46] convert savevm-async to threads
+Subject: [PATCH 46/47] convert savevm-async to threads
 
 ---
- savevm-async.c | 149 +++++++++++++++++++++++++++++++++++----------------------
- 1 file changed, 93 insertions(+), 56 deletions(-)
+ savevm-async.c | 144 +++++++++++++++++++++++++++++++++++----------------------
+ 1 file changed, 88 insertions(+), 56 deletions(-)
 
 diff --git a/savevm-async.c b/savevm-async.c
-index 05b5b19..e293a00 100644
+index 3adf89f..9f839fa 100644
 --- a/savevm-async.c
 +++ b/savevm-async.c
 @@ -48,6 +48,8 @@ static struct SnapshotState {
@@ -40,7 +40,7 @@ index 05b5b19..e293a00 100644
  }
  
  static int block_state_close(void *opaque)
-@@ -156,36 +145,76 @@ static int block_state_close(void *opaque)
+@@ -156,51 +145,90 @@ static int block_state_close(void *opaque)
      return blk_flush(snap_state.target);
  }
  
@@ -77,13 +77,14 @@ index 05b5b19..e293a00 100644
 -    if (ret < 0) {
 -        return ret;
 +
++    aio_context = blk_get_aio_context(snap_state.target);
++    aio_context_acquire(aio_context);
 +    co = qemu_coroutine_create(&block_state_write_entry, &rwco);
 +    qemu_coroutine_enter(co);
-+
-+    aio_context = blk_get_aio_context(snap_state.target);
 +    while (rwco.ret == NOT_DONE) {
 +        aio_poll(aio_context, true);
      }
++    aio_context_release(aio_context);
 +
      snap_state.bs_pos += qiov.size;
      return qiov.size;
@@ -124,18 +125,16 @@ index 05b5b19..e293a00 100644
  {
      int ret;
      int64_t maxlen;
-+    AioContext *aio_context;
 +
      MigrationParams params = {
          .blk = 0,
          .shared = 0
-@@ -193,57 +222,64 @@ static void process_savevm_co(void *opaque)
+     };
  
-     snap_state.state = SAVE_STATE_ACTIVE;
+-    snap_state.state = SAVE_STATE_ACTIVE;
++    rcu_register_thread();
  
 -    qemu_mutex_unlock_iothread();
-+    rcu_register_thread();
-+
      qemu_savevm_state_header(snap_state.file);
      ret = qemu_savevm_state_begin(snap_state.file, &params);
 -    qemu_mutex_lock_iothread();
@@ -147,12 +146,8 @@ index 05b5b19..e293a00 100644
 +        return NULL;
      }
  
-+    aio_context = bdrv_get_aio_context(blk_bs(snap_state.target));
-+    aio_context_acquire(aio_context);
-+
      while (snap_state.state == SAVE_STATE_ACTIVE) {
-         uint64_t pending_size, pend_post, pend_nonpost;
+@@ -209,41 +237,43 @@ static void process_savevm_co(void *opaque)
          qemu_savevm_state_pending(snap_state.file, 0, &pend_nonpost, &pend_post);
          pending_size = pend_post + pend_nonpost;
  
@@ -166,28 +161,26 @@ index 05b5b19..e293a00 100644
 +        maxlen = blk_getlength(snap_state.target) - 30*1024*1024;
 +
 +        if (pending_size > 400000 && snap_state.bs_pos + pending_size < maxlen) {
++            qemu_mutex_lock_iothread();
 +            ret = qemu_savevm_state_iterate(snap_state.file, false);
 +            if (ret < 0) {
-+                aio_context_release(aio_context);
 +                save_snapshot_error("qemu_savevm_state_iterate error %d", ret);
 +                break;
 +            }
++            qemu_mutex_unlock_iothread();
 +            DPRINTF("savevm inerate pending size %lu ret %d\n", pending_size, ret);
          } else {
 -            DPRINTF("done iterating\n");
 -            if (store_and_stop())
-+            aio_context_release(aio_context);
 +            qemu_mutex_lock_iothread();
 +            qemu_system_wakeup_request(QEMU_WAKEUP_REASON_OTHER);
 +            ret = global_state_store();
 +            if (ret) {
-+                qemu_mutex_unlock_iothread();
 +                save_snapshot_error("global_state_store error %d", ret);
 +                break;
 +            }
 +            ret = vm_stop_force_state(RUN_STATE_FINISH_MIGRATE);
 +            if (ret < 0) {
-+                qemu_mutex_unlock_iothread();
 +                save_snapshot_error("vm_stop_force_state error %d", ret);
                  break;
 +            }
@@ -196,7 +189,6 @@ index 05b5b19..e293a00 100644
 +            qemu_savevm_state_cleanup();
              DPRINTF("save complete\n");
 -            save_snapshot_completed();
-+            qemu_mutex_unlock_iothread();
              break;
          }
 -
@@ -209,26 +201,28 @@ index 05b5b19..e293a00 100644
 -            if (store_and_stop())
 -                break;
 -        }
--    }
--
+     }
 -    if(snap_state.state == SAVE_STATE_CANCELLED) {
 -        save_snapshot_completed();
 -        Error *errp = NULL;
 -        qmp_savevm_end(&errp);
-     }
+-    }
 +    qemu_bh_schedule(snap_state.cleanup_bh);
++    qemu_mutex_unlock_iothread();
 +    rcu_unregister_thread();
 +    return NULL;
  }
  
  static const QEMUFileOps block_file_ops = {
-@@ -306,8 +342,9 @@ void qmp_savevm_start(bool has_statefile, const char *statefile, Error **errp)
+@@ -306,8 +336,10 @@ void qmp_savevm_start(bool has_statefile, const char *statefile, Error **errp)
      error_setg(&snap_state.blocker, "block device is in use by savevm");
      blk_op_block_all(snap_state.target, snap_state.blocker);
  
 -    Coroutine *co = qemu_coroutine_create(process_savevm_co, NULL);
 -    qemu_coroutine_enter(co);
++    snap_state.state = SAVE_STATE_ACTIVE;
 +    snap_state.cleanup_bh = qemu_bh_new(process_savevm_cleanup, &snap_state);
 +    qemu_thread_create(&snap_state.thread, "savevm-async", process_savevm_thread,
 +                       NULL, QEMU_THREAD_JOINABLE);