]> git.proxmox.com Git - mirror_qemu.git/blob - accel/hvf/hvf-accel-ops.c
accel: Do not set CPUState::can_do_io in non-TCG accels
[mirror_qemu.git] / accel / hvf / hvf-accel-ops.c
1 /*
2 * Copyright 2008 IBM Corporation
3 * 2008 Red Hat, Inc.
4 * Copyright 2011 Intel Corporation
5 * Copyright 2016 Veertu, Inc.
6 * Copyright 2017 The Android Open Source Project
7 *
8 * QEMU Hypervisor.framework support
9 *
10 * This program is free software; you can redistribute it and/or
11 * modify it under the terms of version 2 of the GNU General Public
12 * License as published by the Free Software Foundation.
13 *
14 * This program is distributed in the hope that it will be useful,
15 * but WITHOUT ANY WARRANTY; without even the implied warranty of
16 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
17 * General Public License for more details.
18 *
19 * You should have received a copy of the GNU General Public License
20 * along with this program; if not, see <http://www.gnu.org/licenses/>.
21 *
22 * This file contain code under public domain from the hvdos project:
23 * https://github.com/mist64/hvdos
24 *
25 * Parts Copyright (c) 2011 NetApp, Inc.
26 * All rights reserved.
27 *
28 * Redistribution and use in source and binary forms, with or without
29 * modification, are permitted provided that the following conditions
30 * are met:
31 * 1. Redistributions of source code must retain the above copyright
32 * notice, this list of conditions and the following disclaimer.
33 * 2. Redistributions in binary form must reproduce the above copyright
34 * notice, this list of conditions and the following disclaimer in the
35 * documentation and/or other materials provided with the distribution.
36 *
37 * THIS SOFTWARE IS PROVIDED BY NETAPP, INC ``AS IS'' AND
38 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
39 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
40 * ARE DISCLAIMED. IN NO EVENT SHALL NETAPP, INC OR CONTRIBUTORS BE LIABLE
41 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
42 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
43 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
44 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
45 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
46 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
47 * SUCH DAMAGE.
48 */
49
50 #include "qemu/osdep.h"
51 #include "qemu/error-report.h"
52 #include "qemu/main-loop.h"
53 #include "exec/address-spaces.h"
54 #include "exec/exec-all.h"
55 #include "exec/gdbstub.h"
56 #include "sysemu/cpus.h"
57 #include "sysemu/hvf.h"
58 #include "sysemu/hvf_int.h"
59 #include "sysemu/runstate.h"
60 #include "qemu/guest-random.h"
61
62 HVFState *hvf_state;
63
64 #ifdef __aarch64__
65 #define HV_VM_DEFAULT NULL
66 #endif
67
68 /* Memory slots */
69
70 hvf_slot *hvf_find_overlap_slot(uint64_t start, uint64_t size)
71 {
72 hvf_slot *slot;
73 int x;
74 for (x = 0; x < hvf_state->num_slots; ++x) {
75 slot = &hvf_state->slots[x];
76 if (slot->size && start < (slot->start + slot->size) &&
77 (start + size) > slot->start) {
78 return slot;
79 }
80 }
81 return NULL;
82 }
83
84 struct mac_slot {
85 int present;
86 uint64_t size;
87 uint64_t gpa_start;
88 uint64_t gva;
89 };
90
91 struct mac_slot mac_slots[32];
92
93 static int do_hvf_set_memory(hvf_slot *slot, hv_memory_flags_t flags)
94 {
95 struct mac_slot *macslot;
96 hv_return_t ret;
97
98 macslot = &mac_slots[slot->slot_id];
99
100 if (macslot->present) {
101 if (macslot->size != slot->size) {
102 macslot->present = 0;
103 ret = hv_vm_unmap(macslot->gpa_start, macslot->size);
104 assert_hvf_ok(ret);
105 }
106 }
107
108 if (!slot->size) {
109 return 0;
110 }
111
112 macslot->present = 1;
113 macslot->gpa_start = slot->start;
114 macslot->size = slot->size;
115 ret = hv_vm_map(slot->mem, slot->start, slot->size, flags);
116 assert_hvf_ok(ret);
117 return 0;
118 }
119
120 static void hvf_set_phys_mem(MemoryRegionSection *section, bool add)
121 {
122 hvf_slot *mem;
123 MemoryRegion *area = section->mr;
124 bool writable = !area->readonly && !area->rom_device;
125 hv_memory_flags_t flags;
126 uint64_t page_size = qemu_real_host_page_size();
127
128 if (!memory_region_is_ram(area)) {
129 if (writable) {
130 return;
131 } else if (!memory_region_is_romd(area)) {
132 /*
133 * If the memory device is not in romd_mode, then we actually want
134 * to remove the hvf memory slot so all accesses will trap.
135 */
136 add = false;
137 }
138 }
139
140 if (!QEMU_IS_ALIGNED(int128_get64(section->size), page_size) ||
141 !QEMU_IS_ALIGNED(section->offset_within_address_space, page_size)) {
142 /* Not page aligned, so we can not map as RAM */
143 add = false;
144 }
145
146 mem = hvf_find_overlap_slot(
147 section->offset_within_address_space,
148 int128_get64(section->size));
149
150 if (mem && add) {
151 if (mem->size == int128_get64(section->size) &&
152 mem->start == section->offset_within_address_space &&
153 mem->mem == (memory_region_get_ram_ptr(area) +
154 section->offset_within_region)) {
155 return; /* Same region was attempted to register, go away. */
156 }
157 }
158
159 /* Region needs to be reset. set the size to 0 and remap it. */
160 if (mem) {
161 mem->size = 0;
162 if (do_hvf_set_memory(mem, 0)) {
163 error_report("Failed to reset overlapping slot");
164 abort();
165 }
166 }
167
168 if (!add) {
169 return;
170 }
171
172 if (area->readonly ||
173 (!memory_region_is_ram(area) && memory_region_is_romd(area))) {
174 flags = HV_MEMORY_READ | HV_MEMORY_EXEC;
175 } else {
176 flags = HV_MEMORY_READ | HV_MEMORY_WRITE | HV_MEMORY_EXEC;
177 }
178
179 /* Now make a new slot. */
180 int x;
181
182 for (x = 0; x < hvf_state->num_slots; ++x) {
183 mem = &hvf_state->slots[x];
184 if (!mem->size) {
185 break;
186 }
187 }
188
189 if (x == hvf_state->num_slots) {
190 error_report("No free slots");
191 abort();
192 }
193
194 mem->size = int128_get64(section->size);
195 mem->mem = memory_region_get_ram_ptr(area) + section->offset_within_region;
196 mem->start = section->offset_within_address_space;
197 mem->region = area;
198
199 if (do_hvf_set_memory(mem, flags)) {
200 error_report("Error registering new memory slot");
201 abort();
202 }
203 }
204
205 static void do_hvf_cpu_synchronize_state(CPUState *cpu, run_on_cpu_data arg)
206 {
207 if (!cpu->vcpu_dirty) {
208 hvf_get_registers(cpu);
209 cpu->vcpu_dirty = true;
210 }
211 }
212
213 static void hvf_cpu_synchronize_state(CPUState *cpu)
214 {
215 if (!cpu->vcpu_dirty) {
216 run_on_cpu(cpu, do_hvf_cpu_synchronize_state, RUN_ON_CPU_NULL);
217 }
218 }
219
220 static void do_hvf_cpu_synchronize_set_dirty(CPUState *cpu,
221 run_on_cpu_data arg)
222 {
223 /* QEMU state is the reference, push it to HVF now and on next entry */
224 cpu->vcpu_dirty = true;
225 }
226
227 static void hvf_cpu_synchronize_post_reset(CPUState *cpu)
228 {
229 run_on_cpu(cpu, do_hvf_cpu_synchronize_set_dirty, RUN_ON_CPU_NULL);
230 }
231
232 static void hvf_cpu_synchronize_post_init(CPUState *cpu)
233 {
234 run_on_cpu(cpu, do_hvf_cpu_synchronize_set_dirty, RUN_ON_CPU_NULL);
235 }
236
237 static void hvf_cpu_synchronize_pre_loadvm(CPUState *cpu)
238 {
239 run_on_cpu(cpu, do_hvf_cpu_synchronize_set_dirty, RUN_ON_CPU_NULL);
240 }
241
242 static void hvf_set_dirty_tracking(MemoryRegionSection *section, bool on)
243 {
244 hvf_slot *slot;
245
246 slot = hvf_find_overlap_slot(
247 section->offset_within_address_space,
248 int128_get64(section->size));
249
250 /* protect region against writes; begin tracking it */
251 if (on) {
252 slot->flags |= HVF_SLOT_LOG;
253 hv_vm_protect((uintptr_t)slot->start, (size_t)slot->size,
254 HV_MEMORY_READ | HV_MEMORY_EXEC);
255 /* stop tracking region*/
256 } else {
257 slot->flags &= ~HVF_SLOT_LOG;
258 hv_vm_protect((uintptr_t)slot->start, (size_t)slot->size,
259 HV_MEMORY_READ | HV_MEMORY_WRITE | HV_MEMORY_EXEC);
260 }
261 }
262
263 static void hvf_log_start(MemoryListener *listener,
264 MemoryRegionSection *section, int old, int new)
265 {
266 if (old != 0) {
267 return;
268 }
269
270 hvf_set_dirty_tracking(section, 1);
271 }
272
273 static void hvf_log_stop(MemoryListener *listener,
274 MemoryRegionSection *section, int old, int new)
275 {
276 if (new != 0) {
277 return;
278 }
279
280 hvf_set_dirty_tracking(section, 0);
281 }
282
283 static void hvf_log_sync(MemoryListener *listener,
284 MemoryRegionSection *section)
285 {
286 /*
287 * sync of dirty pages is handled elsewhere; just make sure we keep
288 * tracking the region.
289 */
290 hvf_set_dirty_tracking(section, 1);
291 }
292
293 static void hvf_region_add(MemoryListener *listener,
294 MemoryRegionSection *section)
295 {
296 hvf_set_phys_mem(section, true);
297 }
298
299 static void hvf_region_del(MemoryListener *listener,
300 MemoryRegionSection *section)
301 {
302 hvf_set_phys_mem(section, false);
303 }
304
305 static MemoryListener hvf_memory_listener = {
306 .name = "hvf",
307 .priority = MEMORY_LISTENER_PRIORITY_ACCEL,
308 .region_add = hvf_region_add,
309 .region_del = hvf_region_del,
310 .log_start = hvf_log_start,
311 .log_stop = hvf_log_stop,
312 .log_sync = hvf_log_sync,
313 };
314
315 static void dummy_signal(int sig)
316 {
317 }
318
319 bool hvf_allowed;
320
321 static int hvf_accel_init(MachineState *ms)
322 {
323 int x;
324 hv_return_t ret;
325 HVFState *s;
326
327 ret = hv_vm_create(HV_VM_DEFAULT);
328 assert_hvf_ok(ret);
329
330 s = g_new0(HVFState, 1);
331
332 s->num_slots = ARRAY_SIZE(s->slots);
333 for (x = 0; x < s->num_slots; ++x) {
334 s->slots[x].size = 0;
335 s->slots[x].slot_id = x;
336 }
337
338 QTAILQ_INIT(&s->hvf_sw_breakpoints);
339
340 hvf_state = s;
341 memory_listener_register(&hvf_memory_listener, &address_space_memory);
342
343 return hvf_arch_init();
344 }
345
346 static inline int hvf_gdbstub_sstep_flags(void)
347 {
348 return SSTEP_ENABLE | SSTEP_NOIRQ;
349 }
350
351 static void hvf_accel_class_init(ObjectClass *oc, void *data)
352 {
353 AccelClass *ac = ACCEL_CLASS(oc);
354 ac->name = "HVF";
355 ac->init_machine = hvf_accel_init;
356 ac->allowed = &hvf_allowed;
357 ac->gdbstub_supported_sstep_flags = hvf_gdbstub_sstep_flags;
358 }
359
360 static const TypeInfo hvf_accel_type = {
361 .name = TYPE_HVF_ACCEL,
362 .parent = TYPE_ACCEL,
363 .class_init = hvf_accel_class_init,
364 };
365
366 static void hvf_type_init(void)
367 {
368 type_register_static(&hvf_accel_type);
369 }
370
371 type_init(hvf_type_init);
372
373 static void hvf_vcpu_destroy(CPUState *cpu)
374 {
375 hv_return_t ret = hv_vcpu_destroy(cpu->accel->fd);
376 assert_hvf_ok(ret);
377
378 hvf_arch_vcpu_destroy(cpu);
379 g_free(cpu->accel);
380 cpu->accel = NULL;
381 }
382
383 static int hvf_init_vcpu(CPUState *cpu)
384 {
385 int r;
386
387 cpu->accel = g_new0(AccelCPUState, 1);
388
389 /* init cpu signals */
390 struct sigaction sigact;
391
392 memset(&sigact, 0, sizeof(sigact));
393 sigact.sa_handler = dummy_signal;
394 sigaction(SIG_IPI, &sigact, NULL);
395
396 pthread_sigmask(SIG_BLOCK, NULL, &cpu->accel->unblock_ipi_mask);
397 sigdelset(&cpu->accel->unblock_ipi_mask, SIG_IPI);
398
399 #ifdef __aarch64__
400 r = hv_vcpu_create(&cpu->accel->fd,
401 (hv_vcpu_exit_t **)&cpu->accel->exit, NULL);
402 #else
403 r = hv_vcpu_create((hv_vcpuid_t *)&cpu->accel->fd, HV_VCPU_DEFAULT);
404 #endif
405 cpu->vcpu_dirty = 1;
406 assert_hvf_ok(r);
407
408 cpu->accel->guest_debug_enabled = false;
409
410 return hvf_arch_init_vcpu(cpu);
411 }
412
413 /*
414 * The HVF-specific vCPU thread function. This one should only run when the host
415 * CPU supports the VMX "unrestricted guest" feature.
416 */
417 static void *hvf_cpu_thread_fn(void *arg)
418 {
419 CPUState *cpu = arg;
420
421 int r;
422
423 assert(hvf_enabled());
424
425 rcu_register_thread();
426
427 bql_lock();
428 qemu_thread_get_self(cpu->thread);
429
430 cpu->thread_id = qemu_get_thread_id();
431 current_cpu = cpu;
432
433 hvf_init_vcpu(cpu);
434
435 /* signal CPU creation */
436 cpu_thread_signal_created(cpu);
437 qemu_guest_random_seed_thread_part2(cpu->random_seed);
438
439 do {
440 if (cpu_can_run(cpu)) {
441 r = hvf_vcpu_exec(cpu);
442 if (r == EXCP_DEBUG) {
443 cpu_handle_guest_debug(cpu);
444 }
445 }
446 qemu_wait_io_event(cpu);
447 } while (!cpu->unplug || cpu_can_run(cpu));
448
449 hvf_vcpu_destroy(cpu);
450 cpu_thread_signal_destroyed(cpu);
451 bql_unlock();
452 rcu_unregister_thread();
453 return NULL;
454 }
455
456 static void hvf_start_vcpu_thread(CPUState *cpu)
457 {
458 char thread_name[VCPU_THREAD_NAME_SIZE];
459
460 /*
461 * HVF currently does not support TCG, and only runs in
462 * unrestricted-guest mode.
463 */
464 assert(hvf_enabled());
465
466 cpu->thread = g_malloc0(sizeof(QemuThread));
467 cpu->halt_cond = g_malloc0(sizeof(QemuCond));
468 qemu_cond_init(cpu->halt_cond);
469
470 snprintf(thread_name, VCPU_THREAD_NAME_SIZE, "CPU %d/HVF",
471 cpu->cpu_index);
472 qemu_thread_create(cpu->thread, thread_name, hvf_cpu_thread_fn,
473 cpu, QEMU_THREAD_JOINABLE);
474 }
475
476 static int hvf_insert_breakpoint(CPUState *cpu, int type, vaddr addr, vaddr len)
477 {
478 struct hvf_sw_breakpoint *bp;
479 int err;
480
481 if (type == GDB_BREAKPOINT_SW) {
482 bp = hvf_find_sw_breakpoint(cpu, addr);
483 if (bp) {
484 bp->use_count++;
485 return 0;
486 }
487
488 bp = g_new(struct hvf_sw_breakpoint, 1);
489 bp->pc = addr;
490 bp->use_count = 1;
491 err = hvf_arch_insert_sw_breakpoint(cpu, bp);
492 if (err) {
493 g_free(bp);
494 return err;
495 }
496
497 QTAILQ_INSERT_HEAD(&hvf_state->hvf_sw_breakpoints, bp, entry);
498 } else {
499 err = hvf_arch_insert_hw_breakpoint(addr, len, type);
500 if (err) {
501 return err;
502 }
503 }
504
505 CPU_FOREACH(cpu) {
506 err = hvf_update_guest_debug(cpu);
507 if (err) {
508 return err;
509 }
510 }
511 return 0;
512 }
513
514 static int hvf_remove_breakpoint(CPUState *cpu, int type, vaddr addr, vaddr len)
515 {
516 struct hvf_sw_breakpoint *bp;
517 int err;
518
519 if (type == GDB_BREAKPOINT_SW) {
520 bp = hvf_find_sw_breakpoint(cpu, addr);
521 if (!bp) {
522 return -ENOENT;
523 }
524
525 if (bp->use_count > 1) {
526 bp->use_count--;
527 return 0;
528 }
529
530 err = hvf_arch_remove_sw_breakpoint(cpu, bp);
531 if (err) {
532 return err;
533 }
534
535 QTAILQ_REMOVE(&hvf_state->hvf_sw_breakpoints, bp, entry);
536 g_free(bp);
537 } else {
538 err = hvf_arch_remove_hw_breakpoint(addr, len, type);
539 if (err) {
540 return err;
541 }
542 }
543
544 CPU_FOREACH(cpu) {
545 err = hvf_update_guest_debug(cpu);
546 if (err) {
547 return err;
548 }
549 }
550 return 0;
551 }
552
553 static void hvf_remove_all_breakpoints(CPUState *cpu)
554 {
555 struct hvf_sw_breakpoint *bp, *next;
556 CPUState *tmpcpu;
557
558 QTAILQ_FOREACH_SAFE(bp, &hvf_state->hvf_sw_breakpoints, entry, next) {
559 if (hvf_arch_remove_sw_breakpoint(cpu, bp) != 0) {
560 /* Try harder to find a CPU that currently sees the breakpoint. */
561 CPU_FOREACH(tmpcpu)
562 {
563 if (hvf_arch_remove_sw_breakpoint(tmpcpu, bp) == 0) {
564 break;
565 }
566 }
567 }
568 QTAILQ_REMOVE(&hvf_state->hvf_sw_breakpoints, bp, entry);
569 g_free(bp);
570 }
571 hvf_arch_remove_all_hw_breakpoints();
572
573 CPU_FOREACH(cpu) {
574 hvf_update_guest_debug(cpu);
575 }
576 }
577
578 static void hvf_accel_ops_class_init(ObjectClass *oc, void *data)
579 {
580 AccelOpsClass *ops = ACCEL_OPS_CLASS(oc);
581
582 ops->create_vcpu_thread = hvf_start_vcpu_thread;
583 ops->kick_vcpu_thread = hvf_kick_vcpu_thread;
584
585 ops->synchronize_post_reset = hvf_cpu_synchronize_post_reset;
586 ops->synchronize_post_init = hvf_cpu_synchronize_post_init;
587 ops->synchronize_state = hvf_cpu_synchronize_state;
588 ops->synchronize_pre_loadvm = hvf_cpu_synchronize_pre_loadvm;
589
590 ops->insert_breakpoint = hvf_insert_breakpoint;
591 ops->remove_breakpoint = hvf_remove_breakpoint;
592 ops->remove_all_breakpoints = hvf_remove_all_breakpoints;
593 ops->update_guest_debug = hvf_update_guest_debug;
594 ops->supports_guest_debug = hvf_arch_supports_guest_debug;
595 };
596 static const TypeInfo hvf_accel_ops_type = {
597 .name = ACCEL_OPS_NAME("hvf"),
598
599 .parent = TYPE_ACCEL_OPS,
600 .class_init = hvf_accel_ops_class_init,
601 .abstract = true,
602 };
603 static void hvf_accel_ops_register_types(void)
604 {
605 type_register_static(&hvf_accel_ops_type);
606 }
607 type_init(hvf_accel_ops_register_types);