]> git.proxmox.com Git - mirror_qemu.git/commitdiff
Merge remote-tracking branch 'remotes/huth-gitlab/tags/pull-request-2020-01-12' into...
authorPeter Maydell <peter.maydell@linaro.org>
Mon, 13 Jan 2020 13:06:49 +0000 (13:06 +0000)
committerPeter Maydell <peter.maydell@linaro.org>
Mon, 13 Jan 2020 13:06:49 +0000 (13:06 +0000)
* Move qtests into a separate directory
* Build index.html for docs

# gpg: Signature made Sun 12 Jan 2020 11:21:41 GMT
# gpg:                using RSA key 27B88847EEE0250118F3EAB92ED9D774FE702DB5
# gpg:                issuer "thuth@redhat.com"
# gpg: Good signature from "Thomas Huth <th.huth@gmx.de>" [full]
# gpg:                 aka "Thomas Huth <thuth@redhat.com>" [full]
# gpg:                 aka "Thomas Huth <huth@tuxfamily.org>" [full]
# gpg:                 aka "Thomas Huth <th.huth@posteo.de>" [unknown]
# Primary key fingerprint: 27B8 8847 EEE0 2501 18F3  EAB9 2ED9 D774 FE70 2DB5

* remotes/huth-gitlab/tags/pull-request-2020-01-12:
  docs: build an index page for the HTML docs
  tests/libqos: Move the libqos files under tests/qtest/
  tests/Makefile: Move qtest-related settings to a separate Makefile.include
  test: Move qtests to a separate directory
  tests/Makefile: Separate unit test dependencies from qtest dependencies
  tests/Makefile: Remove 'tests/' and '$(EXESUF)' from the check-qtest variables
  tests/ptimer: Remove unnecessary inclusion of libqtest.h
  tests/Makefile: test-char does not need libqtest

Signed-off-by: Peter Maydell <peter.maydell@linaro.org>
1  2 
MAINTAINERS
tests/qtest/vhost-user-test.c

diff --cc MAINTAINERS
Simple merge
index 0000000000000000000000000000000000000000,91ea373ba5915c37216a36f0f7cc376e66c945fa..2324b964adeb9bf4bf0fe3e867778376149c8845
mode 000000,100644..100644
--- /dev/null
@@@ -1,0 -1,967 +1,967 @@@
 -static void chr_event(void *opaque, int event)
+ /*
+  * QTest testcase for the vhost-user
+  *
+  * Copyright (c) 2014 Virtual Open Systems Sarl.
+  *
+  * This work is licensed under the terms of the GNU GPL, version 2 or later.
+  * See the COPYING file in the top-level directory.
+  *
+  */
+ #include "qemu/osdep.h"
+ #include "libqtest-single.h"
+ #include "qapi/error.h"
+ #include "qapi/qmp/qdict.h"
+ #include "qemu/config-file.h"
+ #include "qemu/option.h"
+ #include "qemu/range.h"
+ #include "qemu/sockets.h"
+ #include "chardev/char-fe.h"
+ #include "qemu/memfd.h"
+ #include "qemu/module.h"
+ #include "sysemu/sysemu.h"
+ #include "libqos/libqos.h"
+ #include "libqos/pci-pc.h"
+ #include "libqos/virtio-pci.h"
+ #include "libqos/malloc-pc.h"
+ #include "hw/virtio/virtio-net.h"
+ #include "standard-headers/linux/vhost_types.h"
+ #include "standard-headers/linux/virtio_ids.h"
+ #include "standard-headers/linux/virtio_net.h"
+ #ifdef CONFIG_LINUX
+ #include <sys/vfs.h>
+ #endif
+ #define QEMU_CMD_MEM    " -m %d -object memory-backend-file,id=mem,size=%dM," \
+                         "mem-path=%s,share=on -numa node,memdev=mem"
+ #define QEMU_CMD_MEMFD  " -m %d -object memory-backend-memfd,id=mem,size=%dM," \
+                         " -numa node,memdev=mem"
+ #define QEMU_CMD_CHR    " -chardev socket,id=%s,path=%s%s"
+ #define QEMU_CMD_NETDEV " -netdev vhost-user,id=hs0,chardev=%s,vhostforce"
+ #define HUGETLBFS_MAGIC       0x958458f6
+ /*********** FROM hw/virtio/vhost-user.c *************************************/
+ #define VHOST_MEMORY_MAX_NREGIONS    8
+ #define VHOST_MAX_VIRTQUEUES    0x100
+ #define VHOST_USER_F_PROTOCOL_FEATURES 30
+ #define VHOST_USER_PROTOCOL_F_MQ 0
+ #define VHOST_USER_PROTOCOL_F_LOG_SHMFD 1
+ #define VHOST_USER_PROTOCOL_F_CROSS_ENDIAN   6
+ #define VHOST_LOG_PAGE 0x1000
+ typedef enum VhostUserRequest {
+     VHOST_USER_NONE = 0,
+     VHOST_USER_GET_FEATURES = 1,
+     VHOST_USER_SET_FEATURES = 2,
+     VHOST_USER_SET_OWNER = 3,
+     VHOST_USER_RESET_OWNER = 4,
+     VHOST_USER_SET_MEM_TABLE = 5,
+     VHOST_USER_SET_LOG_BASE = 6,
+     VHOST_USER_SET_LOG_FD = 7,
+     VHOST_USER_SET_VRING_NUM = 8,
+     VHOST_USER_SET_VRING_ADDR = 9,
+     VHOST_USER_SET_VRING_BASE = 10,
+     VHOST_USER_GET_VRING_BASE = 11,
+     VHOST_USER_SET_VRING_KICK = 12,
+     VHOST_USER_SET_VRING_CALL = 13,
+     VHOST_USER_SET_VRING_ERR = 14,
+     VHOST_USER_GET_PROTOCOL_FEATURES = 15,
+     VHOST_USER_SET_PROTOCOL_FEATURES = 16,
+     VHOST_USER_GET_QUEUE_NUM = 17,
+     VHOST_USER_SET_VRING_ENABLE = 18,
+     VHOST_USER_MAX
+ } VhostUserRequest;
+ typedef struct VhostUserMemoryRegion {
+     uint64_t guest_phys_addr;
+     uint64_t memory_size;
+     uint64_t userspace_addr;
+     uint64_t mmap_offset;
+ } VhostUserMemoryRegion;
+ typedef struct VhostUserMemory {
+     uint32_t nregions;
+     uint32_t padding;
+     VhostUserMemoryRegion regions[VHOST_MEMORY_MAX_NREGIONS];
+ } VhostUserMemory;
+ typedef struct VhostUserLog {
+     uint64_t mmap_size;
+     uint64_t mmap_offset;
+ } VhostUserLog;
+ typedef struct VhostUserMsg {
+     VhostUserRequest request;
+ #define VHOST_USER_VERSION_MASK     (0x3)
+ #define VHOST_USER_REPLY_MASK       (0x1<<2)
+     uint32_t flags;
+     uint32_t size; /* the following payload size */
+     union {
+ #define VHOST_USER_VRING_IDX_MASK   (0xff)
+ #define VHOST_USER_VRING_NOFD_MASK  (0x1<<8)
+         uint64_t u64;
+         struct vhost_vring_state state;
+         struct vhost_vring_addr addr;
+         VhostUserMemory memory;
+         VhostUserLog log;
+     } payload;
+ } QEMU_PACKED VhostUserMsg;
+ static VhostUserMsg m __attribute__ ((unused));
+ #define VHOST_USER_HDR_SIZE (sizeof(m.request) \
+                             + sizeof(m.flags) \
+                             + sizeof(m.size))
+ #define VHOST_USER_PAYLOAD_SIZE (sizeof(m) - VHOST_USER_HDR_SIZE)
+ /* The version of the protocol we support */
+ #define VHOST_USER_VERSION    (0x1)
+ /*****************************************************************************/
+ enum {
+     TEST_FLAGS_OK,
+     TEST_FLAGS_DISCONNECT,
+     TEST_FLAGS_BAD,
+     TEST_FLAGS_END,
+ };
+ typedef struct TestServer {
+     gchar *socket_path;
+     gchar *mig_path;
+     gchar *chr_name;
+     gchar *tmpfs;
+     CharBackend chr;
+     int fds_num;
+     int fds[VHOST_MEMORY_MAX_NREGIONS];
+     VhostUserMemory memory;
+     GMainContext *context;
+     GMainLoop *loop;
+     GThread *thread;
+     GMutex data_mutex;
+     GCond data_cond;
+     int log_fd;
+     uint64_t rings;
+     bool test_fail;
+     int test_flags;
+     int queues;
+ } TestServer;
+ static const char *init_hugepagefs(void);
+ static TestServer *test_server_new(const gchar *name);
+ static void test_server_free(TestServer *server);
+ static void test_server_listen(TestServer *server);
+ enum test_memfd {
+     TEST_MEMFD_AUTO,
+     TEST_MEMFD_YES,
+     TEST_MEMFD_NO,
+ };
+ static void append_vhost_opts(TestServer *s, GString *cmd_line,
+                              const char *chr_opts)
+ {
+     g_string_append_printf(cmd_line, QEMU_CMD_CHR QEMU_CMD_NETDEV,
+                            s->chr_name, s->socket_path,
+                            chr_opts, s->chr_name);
+ }
+ static void append_mem_opts(TestServer *server, GString *cmd_line,
+                             int size, enum test_memfd memfd)
+ {
+     if (memfd == TEST_MEMFD_AUTO) {
+         memfd = qemu_memfd_check(MFD_ALLOW_SEALING) ? TEST_MEMFD_YES
+                                                     : TEST_MEMFD_NO;
+     }
+     if (memfd == TEST_MEMFD_YES) {
+         g_string_append_printf(cmd_line, QEMU_CMD_MEMFD, size, size);
+     } else {
+         const char *root = init_hugepagefs() ? : server->tmpfs;
+         g_string_append_printf(cmd_line, QEMU_CMD_MEM, size, size, root);
+     }
+ }
+ static bool wait_for_fds(TestServer *s)
+ {
+     gint64 end_time;
+     bool got_region;
+     int i;
+     g_mutex_lock(&s->data_mutex);
+     end_time = g_get_monotonic_time() + 5 * G_TIME_SPAN_SECOND;
+     while (!s->fds_num) {
+         if (!g_cond_wait_until(&s->data_cond, &s->data_mutex, end_time)) {
+             /* timeout has passed */
+             g_assert(s->fds_num);
+             break;
+         }
+     }
+     /* check for sanity */
+     g_assert_cmpint(s->fds_num, >, 0);
+     g_assert_cmpint(s->fds_num, ==, s->memory.nregions);
+     g_mutex_unlock(&s->data_mutex);
+     got_region = false;
+     for (i = 0; i < s->memory.nregions; ++i) {
+         VhostUserMemoryRegion *reg = &s->memory.regions[i];
+         if (reg->guest_phys_addr == 0) {
+             got_region = true;
+             break;
+         }
+     }
+     if (!got_region) {
+         g_test_skip("No memory at address 0x0");
+     }
+     return got_region;
+ }
+ static void read_guest_mem_server(QTestState *qts, TestServer *s)
+ {
+     uint8_t *guest_mem;
+     int i, j;
+     size_t size;
+     g_mutex_lock(&s->data_mutex);
+     /* iterate all regions */
+     for (i = 0; i < s->fds_num; i++) {
+         /* We'll check only the region statring at 0x0*/
+         if (s->memory.regions[i].guest_phys_addr != 0x0) {
+             continue;
+         }
+         g_assert_cmpint(s->memory.regions[i].memory_size, >, 1024);
+         size = s->memory.regions[i].memory_size +
+             s->memory.regions[i].mmap_offset;
+         guest_mem = mmap(0, size, PROT_READ | PROT_WRITE,
+                          MAP_SHARED, s->fds[i], 0);
+         g_assert(guest_mem != MAP_FAILED);
+         guest_mem += (s->memory.regions[i].mmap_offset / sizeof(*guest_mem));
+         for (j = 0; j < 1024; j++) {
+             uint32_t a = qtest_readb(qts, s->memory.regions[i].guest_phys_addr + j);
+             uint32_t b = guest_mem[j];
+             g_assert_cmpint(a, ==, b);
+         }
+         munmap(guest_mem, s->memory.regions[i].memory_size);
+     }
+     g_mutex_unlock(&s->data_mutex);
+ }
+ static void *thread_function(void *data)
+ {
+     GMainLoop *loop = data;
+     g_main_loop_run(loop);
+     return NULL;
+ }
+ static int chr_can_read(void *opaque)
+ {
+     return VHOST_USER_HDR_SIZE;
+ }
+ static void chr_read(void *opaque, const uint8_t *buf, int size)
+ {
+     TestServer *s = opaque;
+     CharBackend *chr = &s->chr;
+     VhostUserMsg msg;
+     uint8_t *p = (uint8_t *) &msg;
+     int fd = -1;
+     if (s->test_fail) {
+         qemu_chr_fe_disconnect(chr);
+         /* now switch to non-failure */
+         s->test_fail = false;
+     }
+     if (size != VHOST_USER_HDR_SIZE) {
+         g_test_message("Wrong message size received %d", size);
+         return;
+     }
+     g_mutex_lock(&s->data_mutex);
+     memcpy(p, buf, VHOST_USER_HDR_SIZE);
+     if (msg.size) {
+         p += VHOST_USER_HDR_SIZE;
+         size = qemu_chr_fe_read_all(chr, p, msg.size);
+         if (size != msg.size) {
+             g_test_message("Wrong message size received %d != %d",
+                            size, msg.size);
+             return;
+         }
+     }
+     switch (msg.request) {
+     case VHOST_USER_GET_FEATURES:
+         /* send back features to qemu */
+         msg.flags |= VHOST_USER_REPLY_MASK;
+         msg.size = sizeof(m.payload.u64);
+         msg.payload.u64 = 0x1ULL << VHOST_F_LOG_ALL |
+             0x1ULL << VHOST_USER_F_PROTOCOL_FEATURES;
+         if (s->queues > 1) {
+             msg.payload.u64 |= 0x1ULL << VIRTIO_NET_F_MQ;
+         }
+         if (s->test_flags >= TEST_FLAGS_BAD) {
+             msg.payload.u64 = 0;
+             s->test_flags = TEST_FLAGS_END;
+         }
+         p = (uint8_t *) &msg;
+         qemu_chr_fe_write_all(chr, p, VHOST_USER_HDR_SIZE + msg.size);
+         break;
+     case VHOST_USER_SET_FEATURES:
+         g_assert_cmpint(msg.payload.u64 & (0x1ULL << VHOST_USER_F_PROTOCOL_FEATURES),
+                         !=, 0ULL);
+         if (s->test_flags == TEST_FLAGS_DISCONNECT) {
+             qemu_chr_fe_disconnect(chr);
+             s->test_flags = TEST_FLAGS_BAD;
+         }
+         break;
+     case VHOST_USER_GET_PROTOCOL_FEATURES:
+         /* send back features to qemu */
+         msg.flags |= VHOST_USER_REPLY_MASK;
+         msg.size = sizeof(m.payload.u64);
+         msg.payload.u64 = 1 << VHOST_USER_PROTOCOL_F_LOG_SHMFD;
+         msg.payload.u64 |= 1 << VHOST_USER_PROTOCOL_F_CROSS_ENDIAN;
+         if (s->queues > 1) {
+             msg.payload.u64 |= 1 << VHOST_USER_PROTOCOL_F_MQ;
+         }
+         p = (uint8_t *) &msg;
+         qemu_chr_fe_write_all(chr, p, VHOST_USER_HDR_SIZE + msg.size);
+         break;
+     case VHOST_USER_GET_VRING_BASE:
+         /* send back vring base to qemu */
+         msg.flags |= VHOST_USER_REPLY_MASK;
+         msg.size = sizeof(m.payload.state);
+         msg.payload.state.num = 0;
+         p = (uint8_t *) &msg;
+         qemu_chr_fe_write_all(chr, p, VHOST_USER_HDR_SIZE + msg.size);
+         assert(msg.payload.state.index < s->queues * 2);
+         s->rings &= ~(0x1ULL << msg.payload.state.index);
+         g_cond_broadcast(&s->data_cond);
+         break;
+     case VHOST_USER_SET_MEM_TABLE:
+         /* received the mem table */
+         memcpy(&s->memory, &msg.payload.memory, sizeof(msg.payload.memory));
+         s->fds_num = qemu_chr_fe_get_msgfds(chr, s->fds,
+                                             G_N_ELEMENTS(s->fds));
+         /* signal the test that it can continue */
+         g_cond_broadcast(&s->data_cond);
+         break;
+     case VHOST_USER_SET_VRING_KICK:
+     case VHOST_USER_SET_VRING_CALL:
+         /* consume the fd */
+         qemu_chr_fe_get_msgfds(chr, &fd, 1);
+         /*
+          * This is a non-blocking eventfd.
+          * The receive function forces it to be blocking,
+          * so revert it back to non-blocking.
+          */
+         qemu_set_nonblock(fd);
+         break;
+     case VHOST_USER_SET_LOG_BASE:
+         if (s->log_fd != -1) {
+             close(s->log_fd);
+             s->log_fd = -1;
+         }
+         qemu_chr_fe_get_msgfds(chr, &s->log_fd, 1);
+         msg.flags |= VHOST_USER_REPLY_MASK;
+         msg.size = 0;
+         p = (uint8_t *) &msg;
+         qemu_chr_fe_write_all(chr, p, VHOST_USER_HDR_SIZE);
+         g_cond_broadcast(&s->data_cond);
+         break;
+     case VHOST_USER_SET_VRING_BASE:
+         assert(msg.payload.state.index < s->queues * 2);
+         s->rings |= 0x1ULL << msg.payload.state.index;
+         g_cond_broadcast(&s->data_cond);
+         break;
+     case VHOST_USER_GET_QUEUE_NUM:
+         msg.flags |= VHOST_USER_REPLY_MASK;
+         msg.size = sizeof(m.payload.u64);
+         msg.payload.u64 = s->queues;
+         p = (uint8_t *) &msg;
+         qemu_chr_fe_write_all(chr, p, VHOST_USER_HDR_SIZE + msg.size);
+         break;
+     default:
+         break;
+     }
+     g_mutex_unlock(&s->data_mutex);
+ }
+ static const char *init_hugepagefs(void)
+ {
+ #ifdef CONFIG_LINUX
+     static const char *hugepagefs;
+     const char *path = getenv("QTEST_HUGETLBFS_PATH");
+     struct statfs fs;
+     int ret;
+     if (hugepagefs) {
+         return hugepagefs;
+     }
+     if (!path) {
+         return NULL;
+     }
+     if (access(path, R_OK | W_OK | X_OK)) {
+         g_test_message("access on path (%s): %s", path, strerror(errno));
+         g_test_fail();
+         return NULL;
+     }
+     do {
+         ret = statfs(path, &fs);
+     } while (ret != 0 && errno == EINTR);
+     if (ret != 0) {
+         g_test_message("statfs on path (%s): %s", path, strerror(errno));
+         g_test_fail();
+         return NULL;
+     }
+     if (fs.f_type != HUGETLBFS_MAGIC) {
+         g_test_message("Warning: path not on HugeTLBFS: %s", path);
+         g_test_fail();
+         return NULL;
+     }
+     hugepagefs = path;
+     return hugepagefs;
+ #else
+     return NULL;
+ #endif
+ }
+ static TestServer *test_server_new(const gchar *name)
+ {
+     TestServer *server = g_new0(TestServer, 1);
+     char template[] = "/tmp/vhost-test-XXXXXX";
+     const char *tmpfs;
+     server->context = g_main_context_new();
+     server->loop = g_main_loop_new(server->context, FALSE);
+     /* run the main loop thread so the chardev may operate */
+     server->thread = g_thread_new(NULL, thread_function, server->loop);
+     tmpfs = mkdtemp(template);
+     if (!tmpfs) {
+         g_test_message("mkdtemp on path (%s): %s", template, strerror(errno));
+     }
+     g_assert(tmpfs);
+     server->tmpfs = g_strdup(tmpfs);
+     server->socket_path = g_strdup_printf("%s/%s.sock", tmpfs, name);
+     server->mig_path = g_strdup_printf("%s/%s.mig", tmpfs, name);
+     server->chr_name = g_strdup_printf("chr-%s", name);
+     g_mutex_init(&server->data_mutex);
+     g_cond_init(&server->data_cond);
+     server->log_fd = -1;
+     server->queues = 1;
+     return server;
+ }
++static void chr_event(void *opaque, QEMUChrEvent event)
+ {
+     TestServer *s = opaque;
+     if (s->test_flags == TEST_FLAGS_END &&
+         event == CHR_EVENT_CLOSED) {
+         s->test_flags = TEST_FLAGS_OK;
+     }
+ }
+ static void test_server_create_chr(TestServer *server, const gchar *opt)
+ {
+     gchar *chr_path;
+     Chardev *chr;
+     chr_path = g_strdup_printf("unix:%s%s", server->socket_path, opt);
+     chr = qemu_chr_new(server->chr_name, chr_path, server->context);
+     g_free(chr_path);
+     g_assert_nonnull(chr);
+     qemu_chr_fe_init(&server->chr, chr, &error_abort);
+     qemu_chr_fe_set_handlers(&server->chr, chr_can_read, chr_read,
+                              chr_event, NULL, server, server->context, true);
+ }
+ static void test_server_listen(TestServer *server)
+ {
+     test_server_create_chr(server, ",server,nowait");
+ }
+ static void test_server_free(TestServer *server)
+ {
+     int i, ret;
+     /* finish the helper thread and dispatch pending sources */
+     g_main_loop_quit(server->loop);
+     g_thread_join(server->thread);
+     while (g_main_context_pending(NULL)) {
+         g_main_context_iteration(NULL, TRUE);
+     }
+     unlink(server->socket_path);
+     g_free(server->socket_path);
+     unlink(server->mig_path);
+     g_free(server->mig_path);
+     ret = rmdir(server->tmpfs);
+     if (ret != 0) {
+         g_test_message("unable to rmdir: path (%s): %s",
+                        server->tmpfs, strerror(errno));
+     }
+     g_free(server->tmpfs);
+     qemu_chr_fe_deinit(&server->chr, true);
+     for (i = 0; i < server->fds_num; i++) {
+         close(server->fds[i]);
+     }
+     if (server->log_fd != -1) {
+         close(server->log_fd);
+     }
+     g_free(server->chr_name);
+     g_main_loop_unref(server->loop);
+     g_main_context_unref(server->context);
+     g_cond_clear(&server->data_cond);
+     g_mutex_clear(&server->data_mutex);
+     g_free(server);
+ }
+ static void wait_for_log_fd(TestServer *s)
+ {
+     gint64 end_time;
+     g_mutex_lock(&s->data_mutex);
+     end_time = g_get_monotonic_time() + 5 * G_TIME_SPAN_SECOND;
+     while (s->log_fd == -1) {
+         if (!g_cond_wait_until(&s->data_cond, &s->data_mutex, end_time)) {
+             /* timeout has passed */
+             g_assert(s->log_fd != -1);
+             break;
+         }
+     }
+     g_mutex_unlock(&s->data_mutex);
+ }
+ static void write_guest_mem(TestServer *s, uint32_t seed)
+ {
+     uint32_t *guest_mem;
+     int i, j;
+     size_t size;
+     /* iterate all regions */
+     for (i = 0; i < s->fds_num; i++) {
+         /* We'll write only the region statring at 0x0 */
+         if (s->memory.regions[i].guest_phys_addr != 0x0) {
+             continue;
+         }
+         g_assert_cmpint(s->memory.regions[i].memory_size, >, 1024);
+         size = s->memory.regions[i].memory_size +
+             s->memory.regions[i].mmap_offset;
+         guest_mem = mmap(0, size, PROT_READ | PROT_WRITE,
+                          MAP_SHARED, s->fds[i], 0);
+         g_assert(guest_mem != MAP_FAILED);
+         guest_mem += (s->memory.regions[i].mmap_offset / sizeof(*guest_mem));
+         for (j = 0; j < 256; j++) {
+             guest_mem[j] = seed + j;
+         }
+         munmap(guest_mem, s->memory.regions[i].memory_size);
+         break;
+     }
+ }
+ static guint64 get_log_size(TestServer *s)
+ {
+     guint64 log_size = 0;
+     int i;
+     for (i = 0; i < s->memory.nregions; ++i) {
+         VhostUserMemoryRegion *reg = &s->memory.regions[i];
+         guint64 last = range_get_last(reg->guest_phys_addr,
+                                        reg->memory_size);
+         log_size = MAX(log_size, last / (8 * VHOST_LOG_PAGE) + 1);
+     }
+     return log_size;
+ }
+ typedef struct TestMigrateSource {
+     GSource source;
+     TestServer *src;
+     TestServer *dest;
+ } TestMigrateSource;
+ static gboolean
+ test_migrate_source_check(GSource *source)
+ {
+     TestMigrateSource *t = (TestMigrateSource *)source;
+     gboolean overlap = t->src->rings && t->dest->rings;
+     g_assert(!overlap);
+     return FALSE;
+ }
+ GSourceFuncs test_migrate_source_funcs = {
+     .check = test_migrate_source_check,
+ };
+ static void vhost_user_test_cleanup(void *s)
+ {
+     TestServer *server = s;
+     qos_invalidate_command_line();
+     test_server_free(server);
+ }
+ static void *vhost_user_test_setup(GString *cmd_line, void *arg)
+ {
+     TestServer *server = test_server_new("vhost-user-test");
+     test_server_listen(server);
+     append_mem_opts(server, cmd_line, 256, TEST_MEMFD_AUTO);
+     append_vhost_opts(server, cmd_line, "");
+     g_test_queue_destroy(vhost_user_test_cleanup, server);
+     return server;
+ }
+ static void *vhost_user_test_setup_memfd(GString *cmd_line, void *arg)
+ {
+     TestServer *server = test_server_new("vhost-user-test");
+     test_server_listen(server);
+     append_mem_opts(server, cmd_line, 256, TEST_MEMFD_YES);
+     append_vhost_opts(server, cmd_line, "");
+     g_test_queue_destroy(vhost_user_test_cleanup, server);
+     return server;
+ }
+ static void test_read_guest_mem(void *obj, void *arg, QGuestAllocator *alloc)
+ {
+     TestServer *server = arg;
+     if (!wait_for_fds(server)) {
+         return;
+     }
+     read_guest_mem_server(global_qtest, server);
+ }
+ static void test_migrate(void *obj, void *arg, QGuestAllocator *alloc)
+ {
+     TestServer *s = arg;
+     TestServer *dest = test_server_new("dest");
+     GString *dest_cmdline = g_string_new(qos_get_current_command_line());
+     char *uri = g_strdup_printf("%s%s", "unix:", dest->mig_path);
+     QTestState *to;
+     GSource *source;
+     QDict *rsp;
+     guint8 *log;
+     guint64 size;
+     if (!wait_for_fds(s)) {
+         return;
+     }
+     size = get_log_size(s);
+     g_assert_cmpint(size, ==, (256 * 1024 * 1024) / (VHOST_LOG_PAGE * 8));
+     test_server_listen(dest);
+     g_string_append_printf(dest_cmdline, " -incoming %s", uri);
+     append_mem_opts(dest, dest_cmdline, 256, TEST_MEMFD_AUTO);
+     append_vhost_opts(dest, dest_cmdline, "");
+     to = qtest_init(dest_cmdline->str);
+     /* This would be where you call qos_allocate_objects(to, NULL), if you want
+      * to talk to the QVirtioNet object on the destination.
+      */
+     source = g_source_new(&test_migrate_source_funcs,
+                           sizeof(TestMigrateSource));
+     ((TestMigrateSource *)source)->src = s;
+     ((TestMigrateSource *)source)->dest = dest;
+     g_source_attach(source, s->context);
+     /* slow down migration to have time to fiddle with log */
+     /* TODO: qtest could learn to break on some places */
+     rsp = qmp("{ 'execute': 'migrate_set_speed',"
+               "'arguments': { 'value': 10 } }");
+     g_assert(qdict_haskey(rsp, "return"));
+     qobject_unref(rsp);
+     rsp = qmp("{ 'execute': 'migrate', 'arguments': { 'uri': %s } }", uri);
+     g_assert(qdict_haskey(rsp, "return"));
+     qobject_unref(rsp);
+     wait_for_log_fd(s);
+     log = mmap(0, size, PROT_READ | PROT_WRITE, MAP_SHARED, s->log_fd, 0);
+     g_assert(log != MAP_FAILED);
+     /* modify first page */
+     write_guest_mem(s, 0x42);
+     log[0] = 1;
+     munmap(log, size);
+     /* speed things up */
+     rsp = qmp("{ 'execute': 'migrate_set_speed',"
+               "'arguments': { 'value': 0 } }");
+     g_assert(qdict_haskey(rsp, "return"));
+     qobject_unref(rsp);
+     qmp_eventwait("STOP");
+     qtest_qmp_eventwait(to, "RESUME");
+     g_assert(wait_for_fds(dest));
+     read_guest_mem_server(to, dest);
+     g_source_destroy(source);
+     g_source_unref(source);
+     qtest_quit(to);
+     test_server_free(dest);
+     g_free(uri);
+ }
+ static void wait_for_rings_started(TestServer *s, size_t count)
+ {
+     gint64 end_time;
+     g_mutex_lock(&s->data_mutex);
+     end_time = g_get_monotonic_time() + 5 * G_TIME_SPAN_SECOND;
+     while (ctpop64(s->rings) != count) {
+         if (!g_cond_wait_until(&s->data_cond, &s->data_mutex, end_time)) {
+             /* timeout has passed */
+             g_assert_cmpint(ctpop64(s->rings), ==, count);
+             break;
+         }
+     }
+     g_mutex_unlock(&s->data_mutex);
+ }
+ static inline void test_server_connect(TestServer *server)
+ {
+     test_server_create_chr(server, ",reconnect=1");
+ }
+ static gboolean
+ reconnect_cb(gpointer user_data)
+ {
+     TestServer *s = user_data;
+     qemu_chr_fe_disconnect(&s->chr);
+     return FALSE;
+ }
+ static gpointer
+ connect_thread(gpointer data)
+ {
+     TestServer *s = data;
+     /* wait for qemu to start before first try, to avoid extra warnings */
+     g_usleep(G_USEC_PER_SEC);
+     test_server_connect(s);
+     return NULL;
+ }
+ static void *vhost_user_test_setup_reconnect(GString *cmd_line, void *arg)
+ {
+     TestServer *s = test_server_new("reconnect");
+     g_thread_new("connect", connect_thread, s);
+     append_mem_opts(s, cmd_line, 256, TEST_MEMFD_AUTO);
+     append_vhost_opts(s, cmd_line, ",server");
+     g_test_queue_destroy(vhost_user_test_cleanup, s);
+     return s;
+ }
+ static void test_reconnect(void *obj, void *arg, QGuestAllocator *alloc)
+ {
+     TestServer *s = arg;
+     GSource *src;
+     if (!wait_for_fds(s)) {
+         return;
+     }
+     wait_for_rings_started(s, 2);
+     /* reconnect */
+     s->fds_num = 0;
+     s->rings = 0;
+     src = g_idle_source_new();
+     g_source_set_callback(src, reconnect_cb, s, NULL);
+     g_source_attach(src, s->context);
+     g_source_unref(src);
+     g_assert(wait_for_fds(s));
+     wait_for_rings_started(s, 2);
+ }
+ static void *vhost_user_test_setup_connect_fail(GString *cmd_line, void *arg)
+ {
+     TestServer *s = test_server_new("connect-fail");
+     s->test_fail = true;
+     g_thread_new("connect", connect_thread, s);
+     append_mem_opts(s, cmd_line, 256, TEST_MEMFD_AUTO);
+     append_vhost_opts(s, cmd_line, ",server");
+     g_test_queue_destroy(vhost_user_test_cleanup, s);
+     return s;
+ }
+ static void *vhost_user_test_setup_flags_mismatch(GString *cmd_line, void *arg)
+ {
+     TestServer *s = test_server_new("flags-mismatch");
+     s->test_flags = TEST_FLAGS_DISCONNECT;
+     g_thread_new("connect", connect_thread, s);
+     append_mem_opts(s, cmd_line, 256, TEST_MEMFD_AUTO);
+     append_vhost_opts(s, cmd_line, ",server");
+     g_test_queue_destroy(vhost_user_test_cleanup, s);
+     return s;
+ }
+ static void test_vhost_user_started(void *obj, void *arg, QGuestAllocator *alloc)
+ {
+     TestServer *s = arg;
+     if (!wait_for_fds(s)) {
+         return;
+     }
+     wait_for_rings_started(s, 2);
+ }
+ static void *vhost_user_test_setup_multiqueue(GString *cmd_line, void *arg)
+ {
+     TestServer *s = vhost_user_test_setup(cmd_line, arg);
+     s->queues = 2;
+     g_string_append_printf(cmd_line,
+                            " -set netdev.hs0.queues=%d"
+                            " -global virtio-net-pci.vectors=%d",
+                            s->queues, s->queues * 2 + 2);
+     return s;
+ }
+ static void test_multiqueue(void *obj, void *arg, QGuestAllocator *alloc)
+ {
+     TestServer *s = arg;
+     wait_for_rings_started(s, s->queues * 2);
+ }
+ static void register_vhost_user_test(void)
+ {
+     QOSGraphTestOptions opts = {
+         .before = vhost_user_test_setup,
+         .subprocess = true,
+     };
+     qemu_add_opts(&qemu_chardev_opts);
+     qos_add_test("vhost-user/read-guest-mem/memfile",
+                  "virtio-net",
+                  test_read_guest_mem, &opts);
+     if (qemu_memfd_check(MFD_ALLOW_SEALING)) {
+         opts.before = vhost_user_test_setup_memfd;
+         qos_add_test("vhost-user/read-guest-mem/memfd",
+                      "virtio-net",
+                      test_read_guest_mem, &opts);
+     }
+     qos_add_test("vhost-user/migrate",
+                  "virtio-net",
+                  test_migrate, &opts);
+     /* keeps failing on build-system since Aug 15 2017 */
+     if (getenv("QTEST_VHOST_USER_FIXME")) {
+         opts.before = vhost_user_test_setup_reconnect;
+         qos_add_test("vhost-user/reconnect", "virtio-net",
+                      test_reconnect, &opts);
+         opts.before = vhost_user_test_setup_connect_fail;
+         qos_add_test("vhost-user/connect-fail", "virtio-net",
+                      test_vhost_user_started, &opts);
+         opts.before = vhost_user_test_setup_flags_mismatch;
+         qos_add_test("vhost-user/flags-mismatch", "virtio-net",
+                      test_vhost_user_started, &opts);
+     }
+     opts.before = vhost_user_test_setup_multiqueue;
+     opts.edge.extra_device_opts = "mq=on";
+     qos_add_test("vhost-user/multiqueue",
+                  "virtio-net",
+                  test_multiqueue, &opts);
+ }
+ libqos_init(register_vhost_user_test);