]> git.proxmox.com Git - mirror_qemu.git/blob - hw/hyperv/hyperv.c
Merge tag 'for-upstream' of https://repo.or.cz/qemu/kevin into staging
[mirror_qemu.git] / hw / hyperv / hyperv.c
1 /*
2 * Hyper-V guest/hypervisor interaction
3 *
4 * Copyright (c) 2015-2018 Virtuozzo International GmbH.
5 *
6 * This work is licensed under the terms of the GNU GPL, version 2 or later.
7 * See the COPYING file in the top-level directory.
8 */
9
10 #include "qemu/osdep.h"
11 #include "qemu/main-loop.h"
12 #include "qemu/module.h"
13 #include "qapi/error.h"
14 #include "exec/address-spaces.h"
15 #include "sysemu/kvm.h"
16 #include "qemu/bitops.h"
17 #include "qemu/error-report.h"
18 #include "qemu/lockable.h"
19 #include "qemu/queue.h"
20 #include "qemu/rcu.h"
21 #include "qemu/rcu_queue.h"
22 #include "hw/hyperv/hyperv.h"
23 #include "qom/object.h"
24
25 struct SynICState {
26 DeviceState parent_obj;
27
28 CPUState *cs;
29
30 bool sctl_enabled;
31 hwaddr msg_page_addr;
32 hwaddr event_page_addr;
33 MemoryRegion msg_page_mr;
34 MemoryRegion event_page_mr;
35 struct hyperv_message_page *msg_page;
36 struct hyperv_event_flags_page *event_page;
37
38 QemuMutex sint_routes_mutex;
39 QLIST_HEAD(, HvSintRoute) sint_routes;
40 };
41
42 #define TYPE_SYNIC "hyperv-synic"
43 OBJECT_DECLARE_SIMPLE_TYPE(SynICState, SYNIC)
44
45 static bool synic_enabled;
46
47 bool hyperv_is_synic_enabled(void)
48 {
49 return synic_enabled;
50 }
51
52 static SynICState *get_synic(CPUState *cs)
53 {
54 return SYNIC(object_resolve_path_component(OBJECT(cs), "synic"));
55 }
56
57 static void synic_update(SynICState *synic, bool sctl_enable,
58 hwaddr msg_page_addr, hwaddr event_page_addr)
59 {
60
61 synic->sctl_enabled = sctl_enable;
62 if (synic->msg_page_addr != msg_page_addr) {
63 if (synic->msg_page_addr) {
64 memory_region_del_subregion(get_system_memory(),
65 &synic->msg_page_mr);
66 }
67 if (msg_page_addr) {
68 memory_region_add_subregion(get_system_memory(), msg_page_addr,
69 &synic->msg_page_mr);
70 }
71 synic->msg_page_addr = msg_page_addr;
72 }
73 if (synic->event_page_addr != event_page_addr) {
74 if (synic->event_page_addr) {
75 memory_region_del_subregion(get_system_memory(),
76 &synic->event_page_mr);
77 }
78 if (event_page_addr) {
79 memory_region_add_subregion(get_system_memory(), event_page_addr,
80 &synic->event_page_mr);
81 }
82 synic->event_page_addr = event_page_addr;
83 }
84 }
85
86 void hyperv_synic_update(CPUState *cs, bool sctl_enable,
87 hwaddr msg_page_addr, hwaddr event_page_addr)
88 {
89 SynICState *synic = get_synic(cs);
90
91 if (!synic) {
92 return;
93 }
94
95 synic_update(synic, sctl_enable, msg_page_addr, event_page_addr);
96 }
97
98 static void synic_realize(DeviceState *dev, Error **errp)
99 {
100 Object *obj = OBJECT(dev);
101 SynICState *synic = SYNIC(dev);
102 char *msgp_name, *eventp_name;
103 uint32_t vp_index;
104
105 /* memory region names have to be globally unique */
106 vp_index = hyperv_vp_index(synic->cs);
107 msgp_name = g_strdup_printf("synic-%u-msg-page", vp_index);
108 eventp_name = g_strdup_printf("synic-%u-event-page", vp_index);
109
110 memory_region_init_ram(&synic->msg_page_mr, obj, msgp_name,
111 sizeof(*synic->msg_page), &error_abort);
112 memory_region_init_ram(&synic->event_page_mr, obj, eventp_name,
113 sizeof(*synic->event_page), &error_abort);
114 synic->msg_page = memory_region_get_ram_ptr(&synic->msg_page_mr);
115 synic->event_page = memory_region_get_ram_ptr(&synic->event_page_mr);
116 qemu_mutex_init(&synic->sint_routes_mutex);
117 QLIST_INIT(&synic->sint_routes);
118
119 g_free(msgp_name);
120 g_free(eventp_name);
121 }
122
123 static void synic_reset(DeviceState *dev)
124 {
125 SynICState *synic = SYNIC(dev);
126 memset(synic->msg_page, 0, sizeof(*synic->msg_page));
127 memset(synic->event_page, 0, sizeof(*synic->event_page));
128 synic_update(synic, false, 0, 0);
129 assert(QLIST_EMPTY(&synic->sint_routes));
130 }
131
132 static void synic_class_init(ObjectClass *klass, void *data)
133 {
134 DeviceClass *dc = DEVICE_CLASS(klass);
135
136 dc->realize = synic_realize;
137 dc->reset = synic_reset;
138 dc->user_creatable = false;
139 }
140
141 void hyperv_synic_add(CPUState *cs)
142 {
143 Object *obj;
144 SynICState *synic;
145
146 obj = object_new(TYPE_SYNIC);
147 synic = SYNIC(obj);
148 synic->cs = cs;
149 object_property_add_child(OBJECT(cs), "synic", obj);
150 object_unref(obj);
151 qdev_realize(DEVICE(obj), NULL, &error_abort);
152 synic_enabled = true;
153 }
154
155 void hyperv_synic_reset(CPUState *cs)
156 {
157 SynICState *synic = get_synic(cs);
158
159 if (synic) {
160 device_cold_reset(DEVICE(synic));
161 }
162 }
163
164 static const TypeInfo synic_type_info = {
165 .name = TYPE_SYNIC,
166 .parent = TYPE_DEVICE,
167 .instance_size = sizeof(SynICState),
168 .class_init = synic_class_init,
169 };
170
171 static void synic_register_types(void)
172 {
173 type_register_static(&synic_type_info);
174 }
175
176 type_init(synic_register_types)
177
178 /*
179 * KVM has its own message producers (SynIC timers). To guarantee
180 * serialization with both KVM vcpu and the guest cpu, the messages are first
181 * staged in an intermediate area and then posted to the SynIC message page in
182 * the vcpu thread.
183 */
184 typedef struct HvSintStagedMessage {
185 /* message content staged by hyperv_post_msg */
186 struct hyperv_message msg;
187 /* callback + data (r/o) to complete the processing in a BH */
188 HvSintMsgCb cb;
189 void *cb_data;
190 /* message posting status filled by cpu_post_msg */
191 int status;
192 /* passing the buck: */
193 enum {
194 /* initial state */
195 HV_STAGED_MSG_FREE,
196 /*
197 * hyperv_post_msg (e.g. in main loop) grabs the staged area (FREE ->
198 * BUSY), copies msg, and schedules cpu_post_msg on the assigned cpu
199 */
200 HV_STAGED_MSG_BUSY,
201 /*
202 * cpu_post_msg (vcpu thread) tries to copy staged msg to msg slot,
203 * notify the guest, records the status, marks the posting done (BUSY
204 * -> POSTED), and schedules sint_msg_bh BH
205 */
206 HV_STAGED_MSG_POSTED,
207 /*
208 * sint_msg_bh (BH) verifies that the posting is done, runs the
209 * callback, and starts over (POSTED -> FREE)
210 */
211 } state;
212 } HvSintStagedMessage;
213
214 struct HvSintRoute {
215 uint32_t sint;
216 SynICState *synic;
217 int gsi;
218 EventNotifier sint_set_notifier;
219 EventNotifier sint_ack_notifier;
220
221 HvSintStagedMessage *staged_msg;
222
223 unsigned refcount;
224 QLIST_ENTRY(HvSintRoute) link;
225 };
226
227 static CPUState *hyperv_find_vcpu(uint32_t vp_index)
228 {
229 CPUState *cs = qemu_get_cpu(vp_index);
230 assert(hyperv_vp_index(cs) == vp_index);
231 return cs;
232 }
233
234 /*
235 * BH to complete the processing of a staged message.
236 */
237 static void sint_msg_bh(void *opaque)
238 {
239 HvSintRoute *sint_route = opaque;
240 HvSintStagedMessage *staged_msg = sint_route->staged_msg;
241
242 if (qatomic_read(&staged_msg->state) != HV_STAGED_MSG_POSTED) {
243 /* status nor ready yet (spurious ack from guest?), ignore */
244 return;
245 }
246
247 staged_msg->cb(staged_msg->cb_data, staged_msg->status);
248 staged_msg->status = 0;
249
250 /* staged message processing finished, ready to start over */
251 qatomic_set(&staged_msg->state, HV_STAGED_MSG_FREE);
252 /* drop the reference taken in hyperv_post_msg */
253 hyperv_sint_route_unref(sint_route);
254 }
255
256 /*
257 * Worker to transfer the message from the staging area into the SynIC message
258 * page in vcpu context.
259 */
260 static void cpu_post_msg(CPUState *cs, run_on_cpu_data data)
261 {
262 HvSintRoute *sint_route = data.host_ptr;
263 HvSintStagedMessage *staged_msg = sint_route->staged_msg;
264 SynICState *synic = sint_route->synic;
265 struct hyperv_message *dst_msg;
266 bool wait_for_sint_ack = false;
267
268 assert(staged_msg->state == HV_STAGED_MSG_BUSY);
269
270 if (!synic->msg_page_addr) {
271 staged_msg->status = -ENXIO;
272 goto posted;
273 }
274
275 dst_msg = &synic->msg_page->slot[sint_route->sint];
276
277 if (dst_msg->header.message_type != HV_MESSAGE_NONE) {
278 dst_msg->header.message_flags |= HV_MESSAGE_FLAG_PENDING;
279 staged_msg->status = -EAGAIN;
280 wait_for_sint_ack = true;
281 } else {
282 memcpy(dst_msg, &staged_msg->msg, sizeof(*dst_msg));
283 staged_msg->status = hyperv_sint_route_set_sint(sint_route);
284 }
285
286 memory_region_set_dirty(&synic->msg_page_mr, 0, sizeof(*synic->msg_page));
287
288 posted:
289 qatomic_set(&staged_msg->state, HV_STAGED_MSG_POSTED);
290 /*
291 * Notify the msg originator of the progress made; if the slot was busy we
292 * set msg_pending flag in it so it will be the guest who will do EOM and
293 * trigger the notification from KVM via sint_ack_notifier
294 */
295 if (!wait_for_sint_ack) {
296 aio_bh_schedule_oneshot(qemu_get_aio_context(), sint_msg_bh,
297 sint_route);
298 }
299 }
300
301 /*
302 * Post a Hyper-V message to the staging area, for delivery to guest in the
303 * vcpu thread.
304 */
305 int hyperv_post_msg(HvSintRoute *sint_route, struct hyperv_message *src_msg)
306 {
307 HvSintStagedMessage *staged_msg = sint_route->staged_msg;
308
309 assert(staged_msg);
310
311 /* grab the staging area */
312 if (qatomic_cmpxchg(&staged_msg->state, HV_STAGED_MSG_FREE,
313 HV_STAGED_MSG_BUSY) != HV_STAGED_MSG_FREE) {
314 return -EAGAIN;
315 }
316
317 memcpy(&staged_msg->msg, src_msg, sizeof(*src_msg));
318
319 /* hold a reference on sint_route until the callback is finished */
320 hyperv_sint_route_ref(sint_route);
321
322 /* schedule message posting attempt in vcpu thread */
323 async_run_on_cpu(sint_route->synic->cs, cpu_post_msg,
324 RUN_ON_CPU_HOST_PTR(sint_route));
325 return 0;
326 }
327
328 static void sint_ack_handler(EventNotifier *notifier)
329 {
330 HvSintRoute *sint_route = container_of(notifier, HvSintRoute,
331 sint_ack_notifier);
332 event_notifier_test_and_clear(notifier);
333
334 /*
335 * the guest consumed the previous message so complete the current one with
336 * -EAGAIN and let the msg originator retry
337 */
338 aio_bh_schedule_oneshot(qemu_get_aio_context(), sint_msg_bh, sint_route);
339 }
340
341 /*
342 * Set given event flag for a given sint on a given vcpu, and signal the sint.
343 */
344 int hyperv_set_event_flag(HvSintRoute *sint_route, unsigned eventno)
345 {
346 int ret;
347 SynICState *synic = sint_route->synic;
348 unsigned long *flags, set_mask;
349 unsigned set_idx;
350
351 if (eventno > HV_EVENT_FLAGS_COUNT) {
352 return -EINVAL;
353 }
354 if (!synic->sctl_enabled || !synic->event_page_addr) {
355 return -ENXIO;
356 }
357
358 set_idx = BIT_WORD(eventno);
359 set_mask = BIT_MASK(eventno);
360 flags = synic->event_page->slot[sint_route->sint].flags;
361
362 if ((qatomic_fetch_or(&flags[set_idx], set_mask) & set_mask) != set_mask) {
363 memory_region_set_dirty(&synic->event_page_mr, 0,
364 sizeof(*synic->event_page));
365 ret = hyperv_sint_route_set_sint(sint_route);
366 } else {
367 ret = 0;
368 }
369 return ret;
370 }
371
372 HvSintRoute *hyperv_sint_route_new(uint32_t vp_index, uint32_t sint,
373 HvSintMsgCb cb, void *cb_data)
374 {
375 HvSintRoute *sint_route = NULL;
376 EventNotifier *ack_notifier = NULL;
377 int r, gsi;
378 CPUState *cs;
379 SynICState *synic;
380 bool ack_event_initialized = false;
381
382 cs = hyperv_find_vcpu(vp_index);
383 if (!cs) {
384 return NULL;
385 }
386
387 synic = get_synic(cs);
388 if (!synic) {
389 return NULL;
390 }
391
392 sint_route = g_new0(HvSintRoute, 1);
393 if (!sint_route) {
394 return NULL;
395 }
396
397 sint_route->synic = synic;
398 sint_route->sint = sint;
399 sint_route->refcount = 1;
400
401 ack_notifier = cb ? &sint_route->sint_ack_notifier : NULL;
402 if (ack_notifier) {
403 sint_route->staged_msg = g_new0(HvSintStagedMessage, 1);
404 if (!sint_route->staged_msg) {
405 goto cleanup_err_sint;
406 }
407 sint_route->staged_msg->cb = cb;
408 sint_route->staged_msg->cb_data = cb_data;
409
410 r = event_notifier_init(ack_notifier, false);
411 if (r) {
412 goto cleanup_err_sint;
413 }
414 event_notifier_set_handler(ack_notifier, sint_ack_handler);
415 ack_event_initialized = true;
416 }
417
418 /* See if we are done or we need to setup a GSI for this SintRoute */
419 if (!synic->sctl_enabled) {
420 goto cleanup;
421 }
422
423 /* We need to setup a GSI for this SintRoute */
424 r = event_notifier_init(&sint_route->sint_set_notifier, false);
425 if (r) {
426 goto cleanup_err_sint;
427 }
428
429 gsi = kvm_irqchip_add_hv_sint_route(kvm_state, vp_index, sint);
430 if (gsi < 0) {
431 goto cleanup_err_sint_notifier;
432 }
433
434 r = kvm_irqchip_add_irqfd_notifier_gsi(kvm_state,
435 &sint_route->sint_set_notifier,
436 ack_notifier, gsi);
437 if (r) {
438 goto cleanup_err_irqfd;
439 }
440 sint_route->gsi = gsi;
441 cleanup:
442 qemu_mutex_lock(&synic->sint_routes_mutex);
443 QLIST_INSERT_HEAD(&synic->sint_routes, sint_route, link);
444 qemu_mutex_unlock(&synic->sint_routes_mutex);
445 return sint_route;
446
447 cleanup_err_irqfd:
448 kvm_irqchip_release_virq(kvm_state, gsi);
449
450 cleanup_err_sint_notifier:
451 event_notifier_cleanup(&sint_route->sint_set_notifier);
452
453 cleanup_err_sint:
454 if (ack_notifier) {
455 if (ack_event_initialized) {
456 event_notifier_set_handler(ack_notifier, NULL);
457 event_notifier_cleanup(ack_notifier);
458 }
459
460 g_free(sint_route->staged_msg);
461 }
462
463 g_free(sint_route);
464 return NULL;
465 }
466
467 void hyperv_sint_route_ref(HvSintRoute *sint_route)
468 {
469 sint_route->refcount++;
470 }
471
472 void hyperv_sint_route_unref(HvSintRoute *sint_route)
473 {
474 SynICState *synic;
475
476 if (!sint_route) {
477 return;
478 }
479
480 assert(sint_route->refcount > 0);
481
482 if (--sint_route->refcount) {
483 return;
484 }
485
486 synic = sint_route->synic;
487 qemu_mutex_lock(&synic->sint_routes_mutex);
488 QLIST_REMOVE(sint_route, link);
489 qemu_mutex_unlock(&synic->sint_routes_mutex);
490
491 if (sint_route->gsi) {
492 kvm_irqchip_remove_irqfd_notifier_gsi(kvm_state,
493 &sint_route->sint_set_notifier,
494 sint_route->gsi);
495 kvm_irqchip_release_virq(kvm_state, sint_route->gsi);
496 event_notifier_cleanup(&sint_route->sint_set_notifier);
497 }
498
499 if (sint_route->staged_msg) {
500 event_notifier_set_handler(&sint_route->sint_ack_notifier, NULL);
501 event_notifier_cleanup(&sint_route->sint_ack_notifier);
502 g_free(sint_route->staged_msg);
503 }
504 g_free(sint_route);
505 }
506
507 int hyperv_sint_route_set_sint(HvSintRoute *sint_route)
508 {
509 if (!sint_route->gsi) {
510 return 0;
511 }
512
513 return event_notifier_set(&sint_route->sint_set_notifier);
514 }
515
516 typedef struct MsgHandler {
517 struct rcu_head rcu;
518 QLIST_ENTRY(MsgHandler) link;
519 uint32_t conn_id;
520 HvMsgHandler handler;
521 void *data;
522 } MsgHandler;
523
524 typedef struct EventFlagHandler {
525 struct rcu_head rcu;
526 QLIST_ENTRY(EventFlagHandler) link;
527 uint32_t conn_id;
528 EventNotifier *notifier;
529 } EventFlagHandler;
530
531 static QLIST_HEAD(, MsgHandler) msg_handlers;
532 static QLIST_HEAD(, EventFlagHandler) event_flag_handlers;
533 static QemuMutex handlers_mutex;
534
535 static void __attribute__((constructor)) hv_init(void)
536 {
537 QLIST_INIT(&msg_handlers);
538 QLIST_INIT(&event_flag_handlers);
539 qemu_mutex_init(&handlers_mutex);
540 }
541
542 int hyperv_set_msg_handler(uint32_t conn_id, HvMsgHandler handler, void *data)
543 {
544 int ret;
545 MsgHandler *mh;
546
547 QEMU_LOCK_GUARD(&handlers_mutex);
548 QLIST_FOREACH(mh, &msg_handlers, link) {
549 if (mh->conn_id == conn_id) {
550 if (handler) {
551 ret = -EEXIST;
552 } else {
553 QLIST_REMOVE_RCU(mh, link);
554 g_free_rcu(mh, rcu);
555 ret = 0;
556 }
557 return ret;
558 }
559 }
560
561 if (handler) {
562 mh = g_new(MsgHandler, 1);
563 mh->conn_id = conn_id;
564 mh->handler = handler;
565 mh->data = data;
566 QLIST_INSERT_HEAD_RCU(&msg_handlers, mh, link);
567 ret = 0;
568 } else {
569 ret = -ENOENT;
570 }
571
572 return ret;
573 }
574
575 uint16_t hyperv_hcall_post_message(uint64_t param, bool fast)
576 {
577 uint16_t ret;
578 hwaddr len;
579 struct hyperv_post_message_input *msg;
580 MsgHandler *mh;
581
582 if (fast) {
583 return HV_STATUS_INVALID_HYPERCALL_CODE;
584 }
585 if (param & (__alignof__(*msg) - 1)) {
586 return HV_STATUS_INVALID_ALIGNMENT;
587 }
588
589 len = sizeof(*msg);
590 msg = cpu_physical_memory_map(param, &len, 0);
591 if (len < sizeof(*msg)) {
592 ret = HV_STATUS_INSUFFICIENT_MEMORY;
593 goto unmap;
594 }
595 if (msg->payload_size > sizeof(msg->payload)) {
596 ret = HV_STATUS_INVALID_HYPERCALL_INPUT;
597 goto unmap;
598 }
599
600 ret = HV_STATUS_INVALID_CONNECTION_ID;
601 WITH_RCU_READ_LOCK_GUARD() {
602 QLIST_FOREACH_RCU(mh, &msg_handlers, link) {
603 if (mh->conn_id == (msg->connection_id & HV_CONNECTION_ID_MASK)) {
604 ret = mh->handler(msg, mh->data);
605 break;
606 }
607 }
608 }
609
610 unmap:
611 cpu_physical_memory_unmap(msg, len, 0, 0);
612 return ret;
613 }
614
615 static int set_event_flag_handler(uint32_t conn_id, EventNotifier *notifier)
616 {
617 int ret;
618 EventFlagHandler *handler;
619
620 QEMU_LOCK_GUARD(&handlers_mutex);
621 QLIST_FOREACH(handler, &event_flag_handlers, link) {
622 if (handler->conn_id == conn_id) {
623 if (notifier) {
624 ret = -EEXIST;
625 } else {
626 QLIST_REMOVE_RCU(handler, link);
627 g_free_rcu(handler, rcu);
628 ret = 0;
629 }
630 return ret;
631 }
632 }
633
634 if (notifier) {
635 handler = g_new(EventFlagHandler, 1);
636 handler->conn_id = conn_id;
637 handler->notifier = notifier;
638 QLIST_INSERT_HEAD_RCU(&event_flag_handlers, handler, link);
639 ret = 0;
640 } else {
641 ret = -ENOENT;
642 }
643
644 return ret;
645 }
646
647 static bool process_event_flags_userspace;
648
649 int hyperv_set_event_flag_handler(uint32_t conn_id, EventNotifier *notifier)
650 {
651 if (!process_event_flags_userspace &&
652 !kvm_check_extension(kvm_state, KVM_CAP_HYPERV_EVENTFD)) {
653 process_event_flags_userspace = true;
654
655 warn_report("Hyper-V event signaling is not supported by this kernel; "
656 "using slower userspace hypercall processing");
657 }
658
659 if (!process_event_flags_userspace) {
660 struct kvm_hyperv_eventfd hvevfd = {
661 .conn_id = conn_id,
662 .fd = notifier ? event_notifier_get_fd(notifier) : -1,
663 .flags = notifier ? 0 : KVM_HYPERV_EVENTFD_DEASSIGN,
664 };
665
666 return kvm_vm_ioctl(kvm_state, KVM_HYPERV_EVENTFD, &hvevfd);
667 }
668 return set_event_flag_handler(conn_id, notifier);
669 }
670
671 uint16_t hyperv_hcall_signal_event(uint64_t param, bool fast)
672 {
673 EventFlagHandler *handler;
674
675 if (unlikely(!fast)) {
676 hwaddr addr = param;
677
678 if (addr & (__alignof__(addr) - 1)) {
679 return HV_STATUS_INVALID_ALIGNMENT;
680 }
681
682 param = ldq_phys(&address_space_memory, addr);
683 }
684
685 /*
686 * Per spec, bits 32-47 contain the extra "flag number". However, we
687 * have no use for it, and in all known usecases it is zero, so just
688 * report lookup failure if it isn't.
689 */
690 if (param & 0xffff00000000ULL) {
691 return HV_STATUS_INVALID_PORT_ID;
692 }
693 /* remaining bits are reserved-zero */
694 if (param & ~HV_CONNECTION_ID_MASK) {
695 return HV_STATUS_INVALID_HYPERCALL_INPUT;
696 }
697
698 RCU_READ_LOCK_GUARD();
699 QLIST_FOREACH_RCU(handler, &event_flag_handlers, link) {
700 if (handler->conn_id == param) {
701 event_notifier_set(handler->notifier);
702 return 0;
703 }
704 }
705 return HV_STATUS_INVALID_CONNECTION_ID;
706 }
707
708 static HvSynDbgHandler hv_syndbg_handler;
709 static void *hv_syndbg_context;
710
711 void hyperv_set_syndbg_handler(HvSynDbgHandler handler, void *context)
712 {
713 assert(!hv_syndbg_handler);
714 hv_syndbg_handler = handler;
715 hv_syndbg_context = context;
716 }
717
718 uint16_t hyperv_hcall_reset_dbg_session(uint64_t outgpa)
719 {
720 uint16_t ret;
721 HvSynDbgMsg msg;
722 struct hyperv_reset_debug_session_output *reset_dbg_session = NULL;
723 hwaddr len;
724
725 if (!hv_syndbg_handler) {
726 ret = HV_STATUS_INVALID_HYPERCALL_CODE;
727 goto cleanup;
728 }
729
730 len = sizeof(*reset_dbg_session);
731 reset_dbg_session = cpu_physical_memory_map(outgpa, &len, 1);
732 if (!reset_dbg_session || len < sizeof(*reset_dbg_session)) {
733 ret = HV_STATUS_INSUFFICIENT_MEMORY;
734 goto cleanup;
735 }
736
737 msg.type = HV_SYNDBG_MSG_CONNECTION_INFO;
738 ret = hv_syndbg_handler(hv_syndbg_context, &msg);
739 if (ret) {
740 goto cleanup;
741 }
742
743 reset_dbg_session->host_ip = msg.u.connection_info.host_ip;
744 reset_dbg_session->host_port = msg.u.connection_info.host_port;
745 /* The following fields are only used as validation for KDVM */
746 memset(&reset_dbg_session->host_mac, 0,
747 sizeof(reset_dbg_session->host_mac));
748 reset_dbg_session->target_ip = msg.u.connection_info.host_ip;
749 reset_dbg_session->target_port = msg.u.connection_info.host_port;
750 memset(&reset_dbg_session->target_mac, 0,
751 sizeof(reset_dbg_session->target_mac));
752 cleanup:
753 if (reset_dbg_session) {
754 cpu_physical_memory_unmap(reset_dbg_session,
755 sizeof(*reset_dbg_session), 1, len);
756 }
757
758 return ret;
759 }
760
761 uint16_t hyperv_hcall_retreive_dbg_data(uint64_t ingpa, uint64_t outgpa,
762 bool fast)
763 {
764 uint16_t ret;
765 struct hyperv_retrieve_debug_data_input *debug_data_in = NULL;
766 struct hyperv_retrieve_debug_data_output *debug_data_out = NULL;
767 hwaddr in_len, out_len;
768 HvSynDbgMsg msg;
769
770 if (fast || !hv_syndbg_handler) {
771 ret = HV_STATUS_INVALID_HYPERCALL_CODE;
772 goto cleanup;
773 }
774
775 in_len = sizeof(*debug_data_in);
776 debug_data_in = cpu_physical_memory_map(ingpa, &in_len, 0);
777 if (!debug_data_in || in_len < sizeof(*debug_data_in)) {
778 ret = HV_STATUS_INSUFFICIENT_MEMORY;
779 goto cleanup;
780 }
781
782 out_len = sizeof(*debug_data_out);
783 debug_data_out = cpu_physical_memory_map(outgpa, &out_len, 1);
784 if (!debug_data_out || out_len < sizeof(*debug_data_out)) {
785 ret = HV_STATUS_INSUFFICIENT_MEMORY;
786 goto cleanup;
787 }
788
789 msg.type = HV_SYNDBG_MSG_RECV;
790 msg.u.recv.buf_gpa = outgpa + sizeof(*debug_data_out);
791 msg.u.recv.count = TARGET_PAGE_SIZE - sizeof(*debug_data_out);
792 msg.u.recv.options = debug_data_in->options;
793 msg.u.recv.timeout = debug_data_in->timeout;
794 msg.u.recv.is_raw = true;
795 ret = hv_syndbg_handler(hv_syndbg_context, &msg);
796 if (ret == HV_STATUS_NO_DATA) {
797 debug_data_out->retrieved_count = 0;
798 debug_data_out->remaining_count = debug_data_in->count;
799 goto cleanup;
800 } else if (ret != HV_STATUS_SUCCESS) {
801 goto cleanup;
802 }
803
804 debug_data_out->retrieved_count = msg.u.recv.retrieved_count;
805 debug_data_out->remaining_count =
806 debug_data_in->count - msg.u.recv.retrieved_count;
807 cleanup:
808 if (debug_data_out) {
809 cpu_physical_memory_unmap(debug_data_out, sizeof(*debug_data_out), 1,
810 out_len);
811 }
812
813 if (debug_data_in) {
814 cpu_physical_memory_unmap(debug_data_in, sizeof(*debug_data_in), 0,
815 in_len);
816 }
817
818 return ret;
819 }
820
821 uint16_t hyperv_hcall_post_dbg_data(uint64_t ingpa, uint64_t outgpa, bool fast)
822 {
823 uint16_t ret;
824 struct hyperv_post_debug_data_input *post_data_in = NULL;
825 struct hyperv_post_debug_data_output *post_data_out = NULL;
826 hwaddr in_len, out_len;
827 HvSynDbgMsg msg;
828
829 if (fast || !hv_syndbg_handler) {
830 ret = HV_STATUS_INVALID_HYPERCALL_CODE;
831 goto cleanup;
832 }
833
834 in_len = sizeof(*post_data_in);
835 post_data_in = cpu_physical_memory_map(ingpa, &in_len, 0);
836 if (!post_data_in || in_len < sizeof(*post_data_in)) {
837 ret = HV_STATUS_INSUFFICIENT_MEMORY;
838 goto cleanup;
839 }
840
841 if (post_data_in->count > TARGET_PAGE_SIZE - sizeof(*post_data_in)) {
842 ret = HV_STATUS_INVALID_PARAMETER;
843 goto cleanup;
844 }
845
846 out_len = sizeof(*post_data_out);
847 post_data_out = cpu_physical_memory_map(outgpa, &out_len, 1);
848 if (!post_data_out || out_len < sizeof(*post_data_out)) {
849 ret = HV_STATUS_INSUFFICIENT_MEMORY;
850 goto cleanup;
851 }
852
853 msg.type = HV_SYNDBG_MSG_SEND;
854 msg.u.send.buf_gpa = ingpa + sizeof(*post_data_in);
855 msg.u.send.count = post_data_in->count;
856 msg.u.send.is_raw = true;
857 ret = hv_syndbg_handler(hv_syndbg_context, &msg);
858 if (ret != HV_STATUS_SUCCESS) {
859 goto cleanup;
860 }
861
862 post_data_out->pending_count = msg.u.send.pending_count;
863 ret = post_data_out->pending_count ? HV_STATUS_INSUFFICIENT_BUFFERS :
864 HV_STATUS_SUCCESS;
865 cleanup:
866 if (post_data_out) {
867 cpu_physical_memory_unmap(post_data_out,
868 sizeof(*post_data_out), 1, out_len);
869 }
870
871 if (post_data_in) {
872 cpu_physical_memory_unmap(post_data_in,
873 sizeof(*post_data_in), 0, in_len);
874 }
875
876 return ret;
877 }
878
879 uint32_t hyperv_syndbg_send(uint64_t ingpa, uint32_t count)
880 {
881 HvSynDbgMsg msg;
882
883 if (!hv_syndbg_handler) {
884 return HV_SYNDBG_STATUS_INVALID;
885 }
886
887 msg.type = HV_SYNDBG_MSG_SEND;
888 msg.u.send.buf_gpa = ingpa;
889 msg.u.send.count = count;
890 msg.u.send.is_raw = false;
891 if (hv_syndbg_handler(hv_syndbg_context, &msg)) {
892 return HV_SYNDBG_STATUS_INVALID;
893 }
894
895 return HV_SYNDBG_STATUS_SEND_SUCCESS;
896 }
897
898 uint32_t hyperv_syndbg_recv(uint64_t ingpa, uint32_t count)
899 {
900 uint16_t ret;
901 HvSynDbgMsg msg;
902
903 if (!hv_syndbg_handler) {
904 return HV_SYNDBG_STATUS_INVALID;
905 }
906
907 msg.type = HV_SYNDBG_MSG_RECV;
908 msg.u.recv.buf_gpa = ingpa;
909 msg.u.recv.count = count;
910 msg.u.recv.options = 0;
911 msg.u.recv.timeout = 0;
912 msg.u.recv.is_raw = false;
913 ret = hv_syndbg_handler(hv_syndbg_context, &msg);
914 if (ret != HV_STATUS_SUCCESS) {
915 return 0;
916 }
917
918 return HV_SYNDBG_STATUS_SET_SIZE(HV_SYNDBG_STATUS_RECV_SUCCESS,
919 msg.u.recv.retrieved_count);
920 }
921
922 void hyperv_syndbg_set_pending_page(uint64_t ingpa)
923 {
924 HvSynDbgMsg msg;
925
926 if (!hv_syndbg_handler) {
927 return;
928 }
929
930 msg.type = HV_SYNDBG_MSG_SET_PENDING_PAGE;
931 msg.u.pending_page.buf_gpa = ingpa;
932 hv_syndbg_handler(hv_syndbg_context, &msg);
933 }
934
935 uint64_t hyperv_syndbg_query_options(void)
936 {
937 HvSynDbgMsg msg;
938
939 if (!hv_syndbg_handler) {
940 return 0;
941 }
942
943 msg.type = HV_SYNDBG_MSG_QUERY_OPTIONS;
944 if (hv_syndbg_handler(hv_syndbg_context, &msg) != HV_STATUS_SUCCESS) {
945 return 0;
946 }
947
948 return msg.u.query_options.options;
949 }