]> git.proxmox.com Git - mirror_qemu.git/blob - hw/display/vhost-user-gpu.c
Merge tag 'for_upstream' of git://git.kernel.org/pub/scm/virt/kvm/mst/qemu into staging
[mirror_qemu.git] / hw / display / vhost-user-gpu.c
1 /*
2 * vhost-user GPU Device
3 *
4 * Copyright Red Hat, Inc. 2018
5 *
6 * Authors:
7 * Marc-André Lureau <marcandre.lureau@redhat.com>
8 *
9 * This work is licensed under the terms of the GNU GPL, version 2 or later.
10 * See the COPYING file in the top-level directory.
11 */
12
13 #include "qemu/osdep.h"
14 #include "hw/qdev-properties.h"
15 #include "hw/virtio/virtio-gpu.h"
16 #include "chardev/char-fe.h"
17 #include "qapi/error.h"
18 #include "migration/blocker.h"
19
20 typedef enum VhostUserGpuRequest {
21 VHOST_USER_GPU_NONE = 0,
22 VHOST_USER_GPU_GET_PROTOCOL_FEATURES,
23 VHOST_USER_GPU_SET_PROTOCOL_FEATURES,
24 VHOST_USER_GPU_GET_DISPLAY_INFO,
25 VHOST_USER_GPU_CURSOR_POS,
26 VHOST_USER_GPU_CURSOR_POS_HIDE,
27 VHOST_USER_GPU_CURSOR_UPDATE,
28 VHOST_USER_GPU_SCANOUT,
29 VHOST_USER_GPU_UPDATE,
30 VHOST_USER_GPU_DMABUF_SCANOUT,
31 VHOST_USER_GPU_DMABUF_UPDATE,
32 } VhostUserGpuRequest;
33
34 typedef struct VhostUserGpuDisplayInfoReply {
35 struct virtio_gpu_resp_display_info info;
36 } VhostUserGpuDisplayInfoReply;
37
38 typedef struct VhostUserGpuCursorPos {
39 uint32_t scanout_id;
40 uint32_t x;
41 uint32_t y;
42 } QEMU_PACKED VhostUserGpuCursorPos;
43
44 typedef struct VhostUserGpuCursorUpdate {
45 VhostUserGpuCursorPos pos;
46 uint32_t hot_x;
47 uint32_t hot_y;
48 uint32_t data[64 * 64];
49 } QEMU_PACKED VhostUserGpuCursorUpdate;
50
51 typedef struct VhostUserGpuScanout {
52 uint32_t scanout_id;
53 uint32_t width;
54 uint32_t height;
55 } QEMU_PACKED VhostUserGpuScanout;
56
57 typedef struct VhostUserGpuUpdate {
58 uint32_t scanout_id;
59 uint32_t x;
60 uint32_t y;
61 uint32_t width;
62 uint32_t height;
63 uint8_t data[];
64 } QEMU_PACKED VhostUserGpuUpdate;
65
66 typedef struct VhostUserGpuDMABUFScanout {
67 uint32_t scanout_id;
68 uint32_t x;
69 uint32_t y;
70 uint32_t width;
71 uint32_t height;
72 uint32_t fd_width;
73 uint32_t fd_height;
74 uint32_t fd_stride;
75 uint32_t fd_flags;
76 int fd_drm_fourcc;
77 } QEMU_PACKED VhostUserGpuDMABUFScanout;
78
79 typedef struct VhostUserGpuMsg {
80 uint32_t request; /* VhostUserGpuRequest */
81 uint32_t flags;
82 uint32_t size; /* the following payload size */
83 union {
84 VhostUserGpuCursorPos cursor_pos;
85 VhostUserGpuCursorUpdate cursor_update;
86 VhostUserGpuScanout scanout;
87 VhostUserGpuUpdate update;
88 VhostUserGpuDMABUFScanout dmabuf_scanout;
89 struct virtio_gpu_resp_display_info display_info;
90 uint64_t u64;
91 } payload;
92 } QEMU_PACKED VhostUserGpuMsg;
93
94 static VhostUserGpuMsg m __attribute__ ((unused));
95 #define VHOST_USER_GPU_HDR_SIZE \
96 (sizeof(m.request) + sizeof(m.size) + sizeof(m.flags))
97
98 #define VHOST_USER_GPU_MSG_FLAG_REPLY 0x4
99
100 static void vhost_user_gpu_update_blocked(VhostUserGPU *g, bool blocked);
101
102 static void
103 vhost_user_gpu_handle_cursor(VhostUserGPU *g, VhostUserGpuMsg *msg)
104 {
105 VhostUserGpuCursorPos *pos = &msg->payload.cursor_pos;
106 struct virtio_gpu_scanout *s;
107
108 if (pos->scanout_id >= g->parent_obj.conf.max_outputs) {
109 return;
110 }
111 s = &g->parent_obj.scanout[pos->scanout_id];
112
113 if (msg->request == VHOST_USER_GPU_CURSOR_UPDATE) {
114 VhostUserGpuCursorUpdate *up = &msg->payload.cursor_update;
115 if (!s->current_cursor) {
116 s->current_cursor = cursor_alloc(64, 64);
117 }
118
119 s->current_cursor->hot_x = up->hot_x;
120 s->current_cursor->hot_y = up->hot_y;
121
122 memcpy(s->current_cursor->data, up->data,
123 64 * 64 * sizeof(uint32_t));
124
125 dpy_cursor_define(s->con, s->current_cursor);
126 }
127
128 dpy_mouse_set(s->con, pos->x, pos->y,
129 msg->request != VHOST_USER_GPU_CURSOR_POS_HIDE);
130 }
131
132 static void
133 vhost_user_gpu_send_msg(VhostUserGPU *g, const VhostUserGpuMsg *msg)
134 {
135 qemu_chr_fe_write(&g->vhost_chr, (uint8_t *)msg,
136 VHOST_USER_GPU_HDR_SIZE + msg->size);
137 }
138
139 static void
140 vhost_user_gpu_unblock(VhostUserGPU *g)
141 {
142 VhostUserGpuMsg msg = {
143 .request = VHOST_USER_GPU_DMABUF_UPDATE,
144 .flags = VHOST_USER_GPU_MSG_FLAG_REPLY,
145 };
146
147 vhost_user_gpu_send_msg(g, &msg);
148 }
149
150 static void
151 vhost_user_gpu_handle_display(VhostUserGPU *g, VhostUserGpuMsg *msg)
152 {
153 QemuConsole *con = NULL;
154 struct virtio_gpu_scanout *s;
155
156 switch (msg->request) {
157 case VHOST_USER_GPU_GET_PROTOCOL_FEATURES: {
158 VhostUserGpuMsg reply = {
159 .request = msg->request,
160 .flags = VHOST_USER_GPU_MSG_FLAG_REPLY,
161 .size = sizeof(uint64_t),
162 };
163
164 vhost_user_gpu_send_msg(g, &reply);
165 break;
166 }
167 case VHOST_USER_GPU_SET_PROTOCOL_FEATURES: {
168 break;
169 }
170 case VHOST_USER_GPU_GET_DISPLAY_INFO: {
171 struct virtio_gpu_resp_display_info display_info = { {} };
172 VhostUserGpuMsg reply = {
173 .request = msg->request,
174 .flags = VHOST_USER_GPU_MSG_FLAG_REPLY,
175 .size = sizeof(struct virtio_gpu_resp_display_info),
176 };
177
178 display_info.hdr.type = VIRTIO_GPU_RESP_OK_DISPLAY_INFO;
179 virtio_gpu_base_fill_display_info(VIRTIO_GPU_BASE(g), &display_info);
180 memcpy(&reply.payload.display_info, &display_info,
181 sizeof(display_info));
182 vhost_user_gpu_send_msg(g, &reply);
183 break;
184 }
185 case VHOST_USER_GPU_SCANOUT: {
186 VhostUserGpuScanout *m = &msg->payload.scanout;
187
188 if (m->scanout_id >= g->parent_obj.conf.max_outputs) {
189 return;
190 }
191
192 g->parent_obj.enable = 1;
193 s = &g->parent_obj.scanout[m->scanout_id];
194 con = s->con;
195
196 if (m->width == 0) {
197 dpy_gfx_replace_surface(con, NULL);
198 } else {
199 s->ds = qemu_create_displaysurface(m->width, m->height);
200 /* replace surface on next update */
201 }
202
203 break;
204 }
205 case VHOST_USER_GPU_DMABUF_SCANOUT: {
206 VhostUserGpuDMABUFScanout *m = &msg->payload.dmabuf_scanout;
207 int fd = qemu_chr_fe_get_msgfd(&g->vhost_chr);
208 QemuDmaBuf *dmabuf;
209
210 if (m->scanout_id >= g->parent_obj.conf.max_outputs) {
211 error_report("invalid scanout: %d", m->scanout_id);
212 if (fd >= 0) {
213 close(fd);
214 }
215 break;
216 }
217
218 g->parent_obj.enable = 1;
219 con = g->parent_obj.scanout[m->scanout_id].con;
220 dmabuf = &g->dmabuf[m->scanout_id];
221 if (dmabuf->fd >= 0) {
222 close(dmabuf->fd);
223 dmabuf->fd = -1;
224 }
225 dpy_gl_release_dmabuf(con, dmabuf);
226 if (fd == -1) {
227 dpy_gl_scanout_disable(con);
228 break;
229 }
230 *dmabuf = (QemuDmaBuf) {
231 .fd = fd,
232 .width = m->fd_width,
233 .height = m->fd_height,
234 .stride = m->fd_stride,
235 .fourcc = m->fd_drm_fourcc,
236 .y0_top = m->fd_flags & VIRTIO_GPU_RESOURCE_FLAG_Y_0_TOP,
237 };
238 dpy_gl_scanout_dmabuf(con, dmabuf);
239 break;
240 }
241 case VHOST_USER_GPU_DMABUF_UPDATE: {
242 VhostUserGpuUpdate *m = &msg->payload.update;
243
244 if (m->scanout_id >= g->parent_obj.conf.max_outputs ||
245 !g->parent_obj.scanout[m->scanout_id].con) {
246 error_report("invalid scanout update: %d", m->scanout_id);
247 vhost_user_gpu_unblock(g);
248 break;
249 }
250
251 con = g->parent_obj.scanout[m->scanout_id].con;
252 if (!console_has_gl(con)) {
253 error_report("console doesn't support GL!");
254 vhost_user_gpu_unblock(g);
255 break;
256 }
257 g->backend_blocked = true;
258 dpy_gl_update(con, m->x, m->y, m->width, m->height);
259 break;
260 }
261 case VHOST_USER_GPU_UPDATE: {
262 VhostUserGpuUpdate *m = &msg->payload.update;
263
264 if (m->scanout_id >= g->parent_obj.conf.max_outputs) {
265 break;
266 }
267 s = &g->parent_obj.scanout[m->scanout_id];
268 con = s->con;
269 pixman_image_t *image =
270 pixman_image_create_bits(PIXMAN_x8r8g8b8,
271 m->width,
272 m->height,
273 (uint32_t *)m->data,
274 m->width * 4);
275
276 pixman_image_composite(PIXMAN_OP_SRC,
277 image, NULL, s->ds->image,
278 0, 0, 0, 0, m->x, m->y, m->width, m->height);
279
280 pixman_image_unref(image);
281 if (qemu_console_surface(con) != s->ds) {
282 dpy_gfx_replace_surface(con, s->ds);
283 } else {
284 dpy_gfx_update(con, m->x, m->y, m->width, m->height);
285 }
286 break;
287 }
288 default:
289 g_warning("unhandled message %d %d", msg->request, msg->size);
290 }
291
292 if (con && qemu_console_is_gl_blocked(con)) {
293 vhost_user_gpu_update_blocked(g, true);
294 }
295 }
296
297 static void
298 vhost_user_gpu_chr_read(void *opaque)
299 {
300 VhostUserGPU *g = opaque;
301 VhostUserGpuMsg *msg = NULL;
302 VhostUserGpuRequest request;
303 uint32_t size, flags;
304 int r;
305
306 r = qemu_chr_fe_read_all(&g->vhost_chr,
307 (uint8_t *)&request, sizeof(uint32_t));
308 if (r != sizeof(uint32_t)) {
309 error_report("failed to read msg header: %d, %d", r, errno);
310 goto end;
311 }
312
313 r = qemu_chr_fe_read_all(&g->vhost_chr,
314 (uint8_t *)&flags, sizeof(uint32_t));
315 if (r != sizeof(uint32_t)) {
316 error_report("failed to read msg flags");
317 goto end;
318 }
319
320 r = qemu_chr_fe_read_all(&g->vhost_chr,
321 (uint8_t *)&size, sizeof(uint32_t));
322 if (r != sizeof(uint32_t)) {
323 error_report("failed to read msg size");
324 goto end;
325 }
326
327 msg = g_malloc(VHOST_USER_GPU_HDR_SIZE + size);
328
329 r = qemu_chr_fe_read_all(&g->vhost_chr,
330 (uint8_t *)&msg->payload, size);
331 if (r != size) {
332 error_report("failed to read msg payload %d != %d", r, size);
333 goto end;
334 }
335
336 msg->request = request;
337 msg->flags = size;
338 msg->size = size;
339
340 if (request == VHOST_USER_GPU_CURSOR_UPDATE ||
341 request == VHOST_USER_GPU_CURSOR_POS ||
342 request == VHOST_USER_GPU_CURSOR_POS_HIDE) {
343 vhost_user_gpu_handle_cursor(g, msg);
344 } else {
345 vhost_user_gpu_handle_display(g, msg);
346 }
347
348 end:
349 g_free(msg);
350 }
351
352 static void
353 vhost_user_gpu_update_blocked(VhostUserGPU *g, bool blocked)
354 {
355 qemu_set_fd_handler(g->vhost_gpu_fd,
356 blocked ? NULL : vhost_user_gpu_chr_read, NULL, g);
357 }
358
359 static void
360 vhost_user_gpu_gl_flushed(VirtIOGPUBase *b)
361 {
362 VhostUserGPU *g = VHOST_USER_GPU(b);
363
364 if (g->backend_blocked) {
365 vhost_user_gpu_unblock(VHOST_USER_GPU(g));
366 g->backend_blocked = false;
367 }
368
369 vhost_user_gpu_update_blocked(VHOST_USER_GPU(g), false);
370 }
371
372 static bool
373 vhost_user_gpu_do_set_socket(VhostUserGPU *g, Error **errp)
374 {
375 Chardev *chr;
376 int sv[2];
377
378 if (socketpair(PF_UNIX, SOCK_STREAM, 0, sv) == -1) {
379 error_setg_errno(errp, errno, "socketpair() failed");
380 return false;
381 }
382
383 chr = CHARDEV(object_new(TYPE_CHARDEV_SOCKET));
384 if (!chr || qemu_chr_add_client(chr, sv[0]) == -1) {
385 error_setg(errp, "Failed to make socket chardev");
386 goto err;
387 }
388 if (!qemu_chr_fe_init(&g->vhost_chr, chr, errp)) {
389 goto err;
390 }
391 if (vhost_user_gpu_set_socket(&g->vhost->dev, sv[1]) < 0) {
392 error_setg(errp, "Failed to set vhost-user-gpu socket");
393 qemu_chr_fe_deinit(&g->vhost_chr, false);
394 goto err;
395 }
396
397 g->vhost_gpu_fd = sv[0];
398 vhost_user_gpu_update_blocked(g, false);
399 close(sv[1]);
400 return true;
401
402 err:
403 close(sv[0]);
404 close(sv[1]);
405 if (chr) {
406 object_unref(OBJECT(chr));
407 }
408 return false;
409 }
410
411 static void
412 vhost_user_gpu_get_config(VirtIODevice *vdev, uint8_t *config_data)
413 {
414 VhostUserGPU *g = VHOST_USER_GPU(vdev);
415 VirtIOGPUBase *b = VIRTIO_GPU_BASE(vdev);
416 struct virtio_gpu_config *vgconfig =
417 (struct virtio_gpu_config *)config_data;
418 Error *local_err = NULL;
419 int ret;
420
421 memset(config_data, 0, sizeof(struct virtio_gpu_config));
422
423 ret = vhost_dev_get_config(&g->vhost->dev,
424 config_data, sizeof(struct virtio_gpu_config),
425 &local_err);
426 if (ret) {
427 error_report_err(local_err);
428 return;
429 }
430
431 /* those fields are managed by qemu */
432 vgconfig->num_scanouts = b->virtio_config.num_scanouts;
433 vgconfig->events_read = b->virtio_config.events_read;
434 vgconfig->events_clear = b->virtio_config.events_clear;
435 }
436
437 static void
438 vhost_user_gpu_set_config(VirtIODevice *vdev,
439 const uint8_t *config_data)
440 {
441 VhostUserGPU *g = VHOST_USER_GPU(vdev);
442 VirtIOGPUBase *b = VIRTIO_GPU_BASE(vdev);
443 const struct virtio_gpu_config *vgconfig =
444 (const struct virtio_gpu_config *)config_data;
445 int ret;
446
447 if (vgconfig->events_clear) {
448 b->virtio_config.events_read &= ~vgconfig->events_clear;
449 }
450
451 ret = vhost_dev_set_config(&g->vhost->dev, config_data,
452 0, sizeof(struct virtio_gpu_config),
453 VHOST_SET_CONFIG_TYPE_MASTER);
454 if (ret) {
455 error_report("vhost-user-gpu: set device config space failed");
456 return;
457 }
458 }
459
460 static void
461 vhost_user_gpu_set_status(VirtIODevice *vdev, uint8_t val)
462 {
463 VhostUserGPU *g = VHOST_USER_GPU(vdev);
464 Error *err = NULL;
465
466 if (val & VIRTIO_CONFIG_S_DRIVER_OK && vdev->vm_running) {
467 if (!vhost_user_gpu_do_set_socket(g, &err)) {
468 error_report_err(err);
469 return;
470 }
471 vhost_user_backend_start(g->vhost);
472 } else {
473 /* unblock any wait and stop processing */
474 if (g->vhost_gpu_fd != -1) {
475 vhost_user_gpu_update_blocked(g, true);
476 qemu_chr_fe_deinit(&g->vhost_chr, true);
477 g->vhost_gpu_fd = -1;
478 }
479 vhost_user_backend_stop(g->vhost);
480 }
481 }
482
483 static bool
484 vhost_user_gpu_guest_notifier_pending(VirtIODevice *vdev, int idx)
485 {
486 VhostUserGPU *g = VHOST_USER_GPU(vdev);
487
488 return vhost_virtqueue_pending(&g->vhost->dev, idx);
489 }
490
491 static void
492 vhost_user_gpu_guest_notifier_mask(VirtIODevice *vdev, int idx, bool mask)
493 {
494 VhostUserGPU *g = VHOST_USER_GPU(vdev);
495
496 vhost_virtqueue_mask(&g->vhost->dev, vdev, idx, mask);
497 }
498
499 static void
500 vhost_user_gpu_instance_init(Object *obj)
501 {
502 VhostUserGPU *g = VHOST_USER_GPU(obj);
503
504 g->vhost = VHOST_USER_BACKEND(object_new(TYPE_VHOST_USER_BACKEND));
505 object_property_add_alias(obj, "chardev",
506 OBJECT(g->vhost), "chardev");
507 }
508
509 static void
510 vhost_user_gpu_instance_finalize(Object *obj)
511 {
512 VhostUserGPU *g = VHOST_USER_GPU(obj);
513
514 object_unref(OBJECT(g->vhost));
515 }
516
517 static void
518 vhost_user_gpu_reset(VirtIODevice *vdev)
519 {
520 VhostUserGPU *g = VHOST_USER_GPU(vdev);
521
522 virtio_gpu_base_reset(VIRTIO_GPU_BASE(vdev));
523
524 vhost_user_backend_stop(g->vhost);
525 }
526
527 static int
528 vhost_user_gpu_config_change(struct vhost_dev *dev)
529 {
530 error_report("vhost-user-gpu: unhandled backend config change");
531 return -1;
532 }
533
534 static const VhostDevConfigOps config_ops = {
535 .vhost_dev_config_notifier = vhost_user_gpu_config_change,
536 };
537
538 static void
539 vhost_user_gpu_device_realize(DeviceState *qdev, Error **errp)
540 {
541 VhostUserGPU *g = VHOST_USER_GPU(qdev);
542 VirtIODevice *vdev = VIRTIO_DEVICE(g);
543
544 vhost_dev_set_config_notifier(&g->vhost->dev, &config_ops);
545 if (vhost_user_backend_dev_init(g->vhost, vdev, 2, errp) < 0) {
546 return;
547 }
548
549 /* existing backend may send DMABUF, so let's add that requirement */
550 g->parent_obj.conf.flags |= 1 << VIRTIO_GPU_FLAG_DMABUF_ENABLED;
551 if (virtio_has_feature(g->vhost->dev.features, VIRTIO_GPU_F_VIRGL)) {
552 g->parent_obj.conf.flags |= 1 << VIRTIO_GPU_FLAG_VIRGL_ENABLED;
553 }
554 if (virtio_has_feature(g->vhost->dev.features, VIRTIO_GPU_F_EDID)) {
555 g->parent_obj.conf.flags |= 1 << VIRTIO_GPU_FLAG_EDID_ENABLED;
556 } else {
557 error_report("EDID requested but the backend doesn't support it.");
558 g->parent_obj.conf.flags &= ~(1 << VIRTIO_GPU_FLAG_EDID_ENABLED);
559 }
560
561 if (!virtio_gpu_base_device_realize(qdev, NULL, NULL, errp)) {
562 return;
563 }
564
565 g->vhost_gpu_fd = -1;
566 }
567
568 static struct vhost_dev *vhost_user_gpu_get_vhost(VirtIODevice *vdev)
569 {
570 VhostUserGPU *g = VHOST_USER_GPU(vdev);
571 return &g->vhost->dev;
572 }
573
574 static Property vhost_user_gpu_properties[] = {
575 VIRTIO_GPU_BASE_PROPERTIES(VhostUserGPU, parent_obj.conf),
576 DEFINE_PROP_END_OF_LIST(),
577 };
578
579 static void
580 vhost_user_gpu_class_init(ObjectClass *klass, void *data)
581 {
582 DeviceClass *dc = DEVICE_CLASS(klass);
583 VirtioDeviceClass *vdc = VIRTIO_DEVICE_CLASS(klass);
584 VirtIOGPUBaseClass *vgc = VIRTIO_GPU_BASE_CLASS(klass);
585
586 vgc->gl_flushed = vhost_user_gpu_gl_flushed;
587
588 vdc->realize = vhost_user_gpu_device_realize;
589 vdc->reset = vhost_user_gpu_reset;
590 vdc->set_status = vhost_user_gpu_set_status;
591 vdc->guest_notifier_mask = vhost_user_gpu_guest_notifier_mask;
592 vdc->guest_notifier_pending = vhost_user_gpu_guest_notifier_pending;
593 vdc->get_config = vhost_user_gpu_get_config;
594 vdc->set_config = vhost_user_gpu_set_config;
595 vdc->get_vhost = vhost_user_gpu_get_vhost;
596
597 device_class_set_props(dc, vhost_user_gpu_properties);
598 }
599
600 static const TypeInfo vhost_user_gpu_info = {
601 .name = TYPE_VHOST_USER_GPU,
602 .parent = TYPE_VIRTIO_GPU_BASE,
603 .instance_size = sizeof(VhostUserGPU),
604 .instance_init = vhost_user_gpu_instance_init,
605 .instance_finalize = vhost_user_gpu_instance_finalize,
606 .class_init = vhost_user_gpu_class_init,
607 };
608 module_obj(TYPE_VHOST_USER_GPU);
609
610 static void vhost_user_gpu_register_types(void)
611 {
612 type_register_static(&vhost_user_gpu_info);
613 }
614
615 type_init(vhost_user_gpu_register_types)