]> git.proxmox.com Git - mirror_qemu.git/blame - cpus.c
target/i386: move cpu_tmp2_i32 to DisasContext
[mirror_qemu.git] / cpus.c
CommitLineData
296af7c9
BS
1/*
2 * QEMU System Emulator
3 *
4 * Copyright (c) 2003-2008 Fabrice Bellard
5 *
6 * Permission is hereby granted, free of charge, to any person obtaining a copy
7 * of this software and associated documentation files (the "Software"), to deal
8 * in the Software without restriction, including without limitation the rights
9 * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
10 * copies of the Software, and to permit persons to whom the Software is
11 * furnished to do so, subject to the following conditions:
12 *
13 * The above copyright notice and this permission notice shall be included in
14 * all copies or substantial portions of the Software.
15 *
16 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
17 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
18 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
19 * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
20 * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
21 * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
22 * THE SOFTWARE.
23 */
24
7b31bbc2 25#include "qemu/osdep.h"
8d4e9146 26#include "qemu/config-file.h"
33c11879 27#include "cpu.h"
83c9089e 28#include "monitor/monitor.h"
e688df6b 29#include "qapi/error.h"
112ed241 30#include "qapi/qapi-commands-misc.h"
9af23989 31#include "qapi/qapi-events-run-state.h"
a4e15de9 32#include "qapi/qmp/qerror.h"
d49b6836 33#include "qemu/error-report.h"
9c17d615 34#include "sysemu/sysemu.h"
da31d594 35#include "sysemu/block-backend.h"
022c62cb 36#include "exec/gdbstub.h"
9c17d615 37#include "sysemu/dma.h"
b3946626 38#include "sysemu/hw_accel.h"
9c17d615 39#include "sysemu/kvm.h"
b0cb0a66 40#include "sysemu/hax.h"
c97d6d2c 41#include "sysemu/hvf.h"
19306806 42#include "sysemu/whpx.h"
63c91552 43#include "exec/exec-all.h"
296af7c9 44
1de7afc9 45#include "qemu/thread.h"
9c17d615
PB
46#include "sysemu/cpus.h"
47#include "sysemu/qtest.h"
1de7afc9 48#include "qemu/main-loop.h"
922a01a0 49#include "qemu/option.h"
1de7afc9 50#include "qemu/bitmap.h"
cb365646 51#include "qemu/seqlock.h"
8d4e9146 52#include "tcg.h"
9cb805fd 53#include "hw/nmi.h"
8b427044 54#include "sysemu/replay.h"
afed5a5a 55#include "hw/boards.h"
0ff0fc19 56
6d9cb73c
JK
57#ifdef CONFIG_LINUX
58
59#include <sys/prctl.h>
60
c0532a76
MT
61#ifndef PR_MCE_KILL
62#define PR_MCE_KILL 33
63#endif
64
6d9cb73c
JK
65#ifndef PR_MCE_KILL_SET
66#define PR_MCE_KILL_SET 1
67#endif
68
69#ifndef PR_MCE_KILL_EARLY
70#define PR_MCE_KILL_EARLY 1
71#endif
72
73#endif /* CONFIG_LINUX */
74
27498bef
ST
75int64_t max_delay;
76int64_t max_advance;
296af7c9 77
2adcc85d
JH
78/* vcpu throttling controls */
79static QEMUTimer *throttle_timer;
80static unsigned int throttle_percentage;
81
82#define CPU_THROTTLE_PCT_MIN 1
83#define CPU_THROTTLE_PCT_MAX 99
84#define CPU_THROTTLE_TIMESLICE_NS 10000000
85
321bc0b2
TC
86bool cpu_is_stopped(CPUState *cpu)
87{
88 return cpu->stopped || !runstate_is_running();
89}
90
a98ae1d8 91static bool cpu_thread_is_idle(CPUState *cpu)
ac873f1e 92{
c64ca814 93 if (cpu->stop || cpu->queued_work_first) {
ac873f1e
PM
94 return false;
95 }
321bc0b2 96 if (cpu_is_stopped(cpu)) {
ac873f1e
PM
97 return true;
98 }
8c2e1b00 99 if (!cpu->halted || cpu_has_work(cpu) ||
215e79c0 100 kvm_halt_in_kernel()) {
ac873f1e
PM
101 return false;
102 }
103 return true;
104}
105
106static bool all_cpu_threads_idle(void)
107{
182735ef 108 CPUState *cpu;
ac873f1e 109
bdc44640 110 CPU_FOREACH(cpu) {
182735ef 111 if (!cpu_thread_is_idle(cpu)) {
ac873f1e
PM
112 return false;
113 }
114 }
115 return true;
116}
117
946fb27c
PB
118/***********************************************************/
119/* guest cycle counter */
120
a3270e19
PB
121/* Protected by TimersState seqlock */
122
5045e9d9 123static bool icount_sleep = true;
946fb27c
PB
124/* Arbitrarily pick 1MIPS as the minimum allowable speed. */
125#define MAX_ICOUNT_SHIFT 10
a3270e19 126
946fb27c 127typedef struct TimersState {
cb365646 128 /* Protected by BQL. */
946fb27c
PB
129 int64_t cpu_ticks_prev;
130 int64_t cpu_ticks_offset;
cb365646 131
94377115
PB
132 /* Protect fields that can be respectively read outside the
133 * BQL, and written from multiple threads.
cb365646
LPF
134 */
135 QemuSeqLock vm_clock_seqlock;
94377115
PB
136 QemuSpin vm_clock_lock;
137
138 int16_t cpu_ticks_enabled;
c96778bb 139
c1ff073c 140 /* Conversion factor from emulated instructions to virtual clock ticks. */
94377115
PB
141 int16_t icount_time_shift;
142
c96778bb
FK
143 /* Compensate for varying guest execution speed. */
144 int64_t qemu_icount_bias;
94377115
PB
145
146 int64_t vm_clock_warp_start;
147 int64_t cpu_clock_offset;
148
c96778bb
FK
149 /* Only written by TCG thread */
150 int64_t qemu_icount;
94377115 151
b39e3f34 152 /* for adjusting icount */
b39e3f34
PD
153 QEMUTimer *icount_rt_timer;
154 QEMUTimer *icount_vm_timer;
155 QEMUTimer *icount_warp_timer;
946fb27c
PB
156} TimersState;
157
d9cd4007 158static TimersState timers_state;
8d4e9146
FK
159bool mttcg_enabled;
160
161/*
162 * We default to false if we know other options have been enabled
163 * which are currently incompatible with MTTCG. Otherwise when each
164 * guest (target) has been updated to support:
165 * - atomic instructions
166 * - memory ordering primitives (barriers)
167 * they can set the appropriate CONFIG flags in ${target}-softmmu.mak
168 *
169 * Once a guest architecture has been converted to the new primitives
170 * there are two remaining limitations to check.
171 *
172 * - The guest can't be oversized (e.g. 64 bit guest on 32 bit host)
173 * - The host must have a stronger memory order than the guest
174 *
175 * It may be possible in future to support strong guests on weak hosts
176 * but that will require tagging all load/stores in a guest with their
177 * implicit memory order requirements which would likely slow things
178 * down a lot.
179 */
180
181static bool check_tcg_memory_orders_compatible(void)
182{
183#if defined(TCG_GUEST_DEFAULT_MO) && defined(TCG_TARGET_DEFAULT_MO)
184 return (TCG_GUEST_DEFAULT_MO & ~TCG_TARGET_DEFAULT_MO) == 0;
185#else
186 return false;
187#endif
188}
189
190static bool default_mttcg_enabled(void)
191{
83fd9629 192 if (use_icount || TCG_OVERSIZED_GUEST) {
8d4e9146
FK
193 return false;
194 } else {
195#ifdef TARGET_SUPPORTS_MTTCG
196 return check_tcg_memory_orders_compatible();
197#else
198 return false;
199#endif
200 }
201}
202
203void qemu_tcg_configure(QemuOpts *opts, Error **errp)
204{
205 const char *t = qemu_opt_get(opts, "thread");
206 if (t) {
207 if (strcmp(t, "multi") == 0) {
208 if (TCG_OVERSIZED_GUEST) {
209 error_setg(errp, "No MTTCG when guest word size > hosts");
83fd9629
AB
210 } else if (use_icount) {
211 error_setg(errp, "No MTTCG when icount is enabled");
8d4e9146 212 } else {
86953503 213#ifndef TARGET_SUPPORTS_MTTCG
c34c7620
AB
214 error_report("Guest not yet converted to MTTCG - "
215 "you may get unexpected results");
216#endif
8d4e9146
FK
217 if (!check_tcg_memory_orders_compatible()) {
218 error_report("Guest expects a stronger memory ordering "
219 "than the host provides");
8cfef892 220 error_printf("This may cause strange/hard to debug errors\n");
8d4e9146
FK
221 }
222 mttcg_enabled = true;
223 }
224 } else if (strcmp(t, "single") == 0) {
225 mttcg_enabled = false;
226 } else {
227 error_setg(errp, "Invalid 'thread' setting %s", t);
228 }
229 } else {
230 mttcg_enabled = default_mttcg_enabled();
231 }
232}
946fb27c 233
e4cd9657
AB
234/* The current number of executed instructions is based on what we
235 * originally budgeted minus the current state of the decrementing
236 * icount counters in extra/u16.low.
237 */
238static int64_t cpu_get_icount_executed(CPUState *cpu)
239{
240 return cpu->icount_budget - (cpu->icount_decr.u16.low + cpu->icount_extra);
241}
242
512d3c80
AB
243/*
244 * Update the global shared timer_state.qemu_icount to take into
245 * account executed instructions. This is done by the TCG vCPU
246 * thread so the main-loop can see time has moved forward.
247 */
9b4e6f49 248static void cpu_update_icount_locked(CPUState *cpu)
512d3c80
AB
249{
250 int64_t executed = cpu_get_icount_executed(cpu);
251 cpu->icount_budget -= executed;
252
38adcb6e
EC
253 atomic_set_i64(&timers_state.qemu_icount,
254 timers_state.qemu_icount + executed);
9b4e6f49
PB
255}
256
257/*
258 * Update the global shared timer_state.qemu_icount to take into
259 * account executed instructions. This is done by the TCG vCPU
260 * thread so the main-loop can see time has moved forward.
261 */
262void cpu_update_icount(CPUState *cpu)
263{
264 seqlock_write_lock(&timers_state.vm_clock_seqlock,
265 &timers_state.vm_clock_lock);
266 cpu_update_icount_locked(cpu);
94377115
PB
267 seqlock_write_unlock(&timers_state.vm_clock_seqlock,
268 &timers_state.vm_clock_lock);
512d3c80
AB
269}
270
c1ff073c 271static int64_t cpu_get_icount_raw_locked(void)
946fb27c 272{
4917cf44 273 CPUState *cpu = current_cpu;
946fb27c 274
243c5f77 275 if (cpu && cpu->running) {
414b15c9 276 if (!cpu->can_do_io) {
493d89bf 277 error_report("Bad icount read");
2a62914b 278 exit(1);
946fb27c 279 }
e4cd9657 280 /* Take into account what has run */
9b4e6f49 281 cpu_update_icount_locked(cpu);
946fb27c 282 }
38adcb6e
EC
283 /* The read is protected by the seqlock, but needs atomic64 to avoid UB */
284 return atomic_read_i64(&timers_state.qemu_icount);
2a62914b
PD
285}
286
2a62914b
PD
287static int64_t cpu_get_icount_locked(void)
288{
c1ff073c 289 int64_t icount = cpu_get_icount_raw_locked();
c97595d1
EC
290 return atomic_read_i64(&timers_state.qemu_icount_bias) +
291 cpu_icount_to_ns(icount);
c1ff073c
PB
292}
293
294int64_t cpu_get_icount_raw(void)
295{
296 int64_t icount;
297 unsigned start;
298
299 do {
300 start = seqlock_read_begin(&timers_state.vm_clock_seqlock);
301 icount = cpu_get_icount_raw_locked();
302 } while (seqlock_read_retry(&timers_state.vm_clock_seqlock, start));
303
304 return icount;
946fb27c
PB
305}
306
c1ff073c 307/* Return the virtual CPU time, based on the instruction counter. */
17a15f1b
PB
308int64_t cpu_get_icount(void)
309{
310 int64_t icount;
311 unsigned start;
312
313 do {
314 start = seqlock_read_begin(&timers_state.vm_clock_seqlock);
315 icount = cpu_get_icount_locked();
316 } while (seqlock_read_retry(&timers_state.vm_clock_seqlock, start));
317
318 return icount;
319}
320
3f031313
FK
321int64_t cpu_icount_to_ns(int64_t icount)
322{
c1ff073c 323 return icount << atomic_read(&timers_state.icount_time_shift);
3f031313
FK
324}
325
f2a4ad6d
PB
326static int64_t cpu_get_ticks_locked(void)
327{
328 int64_t ticks = timers_state.cpu_ticks_offset;
329 if (timers_state.cpu_ticks_enabled) {
330 ticks += cpu_get_host_ticks();
331 }
332
333 if (timers_state.cpu_ticks_prev > ticks) {
334 /* Non increasing ticks may happen if the host uses software suspend. */
335 timers_state.cpu_ticks_offset += timers_state.cpu_ticks_prev - ticks;
336 ticks = timers_state.cpu_ticks_prev;
337 }
338
339 timers_state.cpu_ticks_prev = ticks;
340 return ticks;
341}
342
d90f3cca
C
343/* return the time elapsed in VM between vm_start and vm_stop. Unless
344 * icount is active, cpu_get_ticks() uses units of the host CPU cycle
345 * counter.
d90f3cca 346 */
946fb27c
PB
347int64_t cpu_get_ticks(void)
348{
5f3e3101
PB
349 int64_t ticks;
350
946fb27c
PB
351 if (use_icount) {
352 return cpu_get_icount();
353 }
5f3e3101 354
f2a4ad6d
PB
355 qemu_spin_lock(&timers_state.vm_clock_lock);
356 ticks = cpu_get_ticks_locked();
357 qemu_spin_unlock(&timers_state.vm_clock_lock);
5f3e3101 358 return ticks;
946fb27c
PB
359}
360
cb365646 361static int64_t cpu_get_clock_locked(void)
946fb27c 362{
1d45cea5 363 int64_t time;
cb365646 364
1d45cea5 365 time = timers_state.cpu_clock_offset;
5f3e3101 366 if (timers_state.cpu_ticks_enabled) {
1d45cea5 367 time += get_clock();
946fb27c 368 }
cb365646 369
1d45cea5 370 return time;
cb365646
LPF
371}
372
d90f3cca 373/* Return the monotonic time elapsed in VM, i.e.,
8212ff86
PM
374 * the time between vm_start and vm_stop
375 */
cb365646
LPF
376int64_t cpu_get_clock(void)
377{
378 int64_t ti;
379 unsigned start;
380
381 do {
382 start = seqlock_read_begin(&timers_state.vm_clock_seqlock);
383 ti = cpu_get_clock_locked();
384 } while (seqlock_read_retry(&timers_state.vm_clock_seqlock, start));
385
386 return ti;
946fb27c
PB
387}
388
cb365646 389/* enable cpu_get_ticks()
3224e878 390 * Caller must hold BQL which serves as mutex for vm_clock_seqlock.
cb365646 391 */
946fb27c
PB
392void cpu_enable_ticks(void)
393{
94377115
PB
394 seqlock_write_lock(&timers_state.vm_clock_seqlock,
395 &timers_state.vm_clock_lock);
946fb27c 396 if (!timers_state.cpu_ticks_enabled) {
4a7428c5 397 timers_state.cpu_ticks_offset -= cpu_get_host_ticks();
946fb27c
PB
398 timers_state.cpu_clock_offset -= get_clock();
399 timers_state.cpu_ticks_enabled = 1;
400 }
94377115
PB
401 seqlock_write_unlock(&timers_state.vm_clock_seqlock,
402 &timers_state.vm_clock_lock);
946fb27c
PB
403}
404
405/* disable cpu_get_ticks() : the clock is stopped. You must not call
cb365646 406 * cpu_get_ticks() after that.
3224e878 407 * Caller must hold BQL which serves as mutex for vm_clock_seqlock.
cb365646 408 */
946fb27c
PB
409void cpu_disable_ticks(void)
410{
94377115
PB
411 seqlock_write_lock(&timers_state.vm_clock_seqlock,
412 &timers_state.vm_clock_lock);
946fb27c 413 if (timers_state.cpu_ticks_enabled) {
4a7428c5 414 timers_state.cpu_ticks_offset += cpu_get_host_ticks();
cb365646 415 timers_state.cpu_clock_offset = cpu_get_clock_locked();
946fb27c
PB
416 timers_state.cpu_ticks_enabled = 0;
417 }
94377115
PB
418 seqlock_write_unlock(&timers_state.vm_clock_seqlock,
419 &timers_state.vm_clock_lock);
946fb27c
PB
420}
421
422/* Correlation between real and virtual time is always going to be
423 fairly approximate, so ignore small variation.
424 When the guest is idle real and virtual time will be aligned in
425 the IO wait loop. */
73bcb24d 426#define ICOUNT_WOBBLE (NANOSECONDS_PER_SECOND / 10)
946fb27c
PB
427
428static void icount_adjust(void)
429{
430 int64_t cur_time;
431 int64_t cur_icount;
432 int64_t delta;
a3270e19
PB
433
434 /* Protected by TimersState mutex. */
946fb27c 435 static int64_t last_delta;
468cc7cf 436
946fb27c
PB
437 /* If the VM is not running, then do nothing. */
438 if (!runstate_is_running()) {
439 return;
440 }
468cc7cf 441
94377115
PB
442 seqlock_write_lock(&timers_state.vm_clock_seqlock,
443 &timers_state.vm_clock_lock);
17a15f1b
PB
444 cur_time = cpu_get_clock_locked();
445 cur_icount = cpu_get_icount_locked();
468cc7cf 446
946fb27c
PB
447 delta = cur_icount - cur_time;
448 /* FIXME: This is a very crude algorithm, somewhat prone to oscillation. */
449 if (delta > 0
450 && last_delta + ICOUNT_WOBBLE < delta * 2
c1ff073c 451 && timers_state.icount_time_shift > 0) {
946fb27c 452 /* The guest is getting too far ahead. Slow time down. */
c1ff073c
PB
453 atomic_set(&timers_state.icount_time_shift,
454 timers_state.icount_time_shift - 1);
946fb27c
PB
455 }
456 if (delta < 0
457 && last_delta - ICOUNT_WOBBLE > delta * 2
c1ff073c 458 && timers_state.icount_time_shift < MAX_ICOUNT_SHIFT) {
946fb27c 459 /* The guest is getting too far behind. Speed time up. */
c1ff073c
PB
460 atomic_set(&timers_state.icount_time_shift,
461 timers_state.icount_time_shift + 1);
946fb27c
PB
462 }
463 last_delta = delta;
c97595d1
EC
464 atomic_set_i64(&timers_state.qemu_icount_bias,
465 cur_icount - (timers_state.qemu_icount
466 << timers_state.icount_time_shift));
94377115
PB
467 seqlock_write_unlock(&timers_state.vm_clock_seqlock,
468 &timers_state.vm_clock_lock);
946fb27c
PB
469}
470
471static void icount_adjust_rt(void *opaque)
472{
b39e3f34 473 timer_mod(timers_state.icount_rt_timer,
1979b908 474 qemu_clock_get_ms(QEMU_CLOCK_VIRTUAL_RT) + 1000);
946fb27c
PB
475 icount_adjust();
476}
477
478static void icount_adjust_vm(void *opaque)
479{
b39e3f34 480 timer_mod(timers_state.icount_vm_timer,
40daca54 481 qemu_clock_get_ns(QEMU_CLOCK_VIRTUAL) +
73bcb24d 482 NANOSECONDS_PER_SECOND / 10);
946fb27c
PB
483 icount_adjust();
484}
485
486static int64_t qemu_icount_round(int64_t count)
487{
c1ff073c
PB
488 int shift = atomic_read(&timers_state.icount_time_shift);
489 return (count + (1 << shift) - 1) >> shift;
946fb27c
PB
490}
491
efab87cf 492static void icount_warp_rt(void)
946fb27c 493{
ccffff48
AB
494 unsigned seq;
495 int64_t warp_start;
496
17a15f1b
PB
497 /* The icount_warp_timer is rescheduled soon after vm_clock_warp_start
498 * changes from -1 to another value, so the race here is okay.
499 */
ccffff48
AB
500 do {
501 seq = seqlock_read_begin(&timers_state.vm_clock_seqlock);
b39e3f34 502 warp_start = timers_state.vm_clock_warp_start;
ccffff48
AB
503 } while (seqlock_read_retry(&timers_state.vm_clock_seqlock, seq));
504
505 if (warp_start == -1) {
946fb27c
PB
506 return;
507 }
508
94377115
PB
509 seqlock_write_lock(&timers_state.vm_clock_seqlock,
510 &timers_state.vm_clock_lock);
946fb27c 511 if (runstate_is_running()) {
8eda206e
PD
512 int64_t clock = REPLAY_CLOCK(REPLAY_CLOCK_VIRTUAL_RT,
513 cpu_get_clock_locked());
8ed961d9
PB
514 int64_t warp_delta;
515
b39e3f34 516 warp_delta = clock - timers_state.vm_clock_warp_start;
8ed961d9 517 if (use_icount == 2) {
946fb27c 518 /*
40daca54 519 * In adaptive mode, do not let QEMU_CLOCK_VIRTUAL run too
946fb27c
PB
520 * far ahead of real time.
521 */
17a15f1b 522 int64_t cur_icount = cpu_get_icount_locked();
bf2a7ddb 523 int64_t delta = clock - cur_icount;
8ed961d9 524 warp_delta = MIN(warp_delta, delta);
946fb27c 525 }
c97595d1
EC
526 atomic_set_i64(&timers_state.qemu_icount_bias,
527 timers_state.qemu_icount_bias + warp_delta);
946fb27c 528 }
b39e3f34 529 timers_state.vm_clock_warp_start = -1;
94377115
PB
530 seqlock_write_unlock(&timers_state.vm_clock_seqlock,
531 &timers_state.vm_clock_lock);
8ed961d9
PB
532
533 if (qemu_clock_expired(QEMU_CLOCK_VIRTUAL)) {
534 qemu_clock_notify(QEMU_CLOCK_VIRTUAL);
535 }
946fb27c
PB
536}
537
e76d1798 538static void icount_timer_cb(void *opaque)
efab87cf 539{
e76d1798
PD
540 /* No need for a checkpoint because the timer already synchronizes
541 * with CHECKPOINT_CLOCK_VIRTUAL_RT.
542 */
543 icount_warp_rt();
efab87cf
PD
544}
545
8156be56
PB
546void qtest_clock_warp(int64_t dest)
547{
40daca54 548 int64_t clock = qemu_clock_get_ns(QEMU_CLOCK_VIRTUAL);
efef88b3 549 AioContext *aio_context;
8156be56 550 assert(qtest_enabled());
efef88b3 551 aio_context = qemu_get_aio_context();
8156be56 552 while (clock < dest) {
40daca54 553 int64_t deadline = qemu_clock_deadline_ns_all(QEMU_CLOCK_VIRTUAL);
c9299e2f 554 int64_t warp = qemu_soonest_timeout(dest - clock, deadline);
efef88b3 555
94377115
PB
556 seqlock_write_lock(&timers_state.vm_clock_seqlock,
557 &timers_state.vm_clock_lock);
c97595d1
EC
558 atomic_set_i64(&timers_state.qemu_icount_bias,
559 timers_state.qemu_icount_bias + warp);
94377115
PB
560 seqlock_write_unlock(&timers_state.vm_clock_seqlock,
561 &timers_state.vm_clock_lock);
17a15f1b 562
40daca54 563 qemu_clock_run_timers(QEMU_CLOCK_VIRTUAL);
efef88b3 564 timerlist_run_timers(aio_context->tlg.tl[QEMU_CLOCK_VIRTUAL]);
40daca54 565 clock = qemu_clock_get_ns(QEMU_CLOCK_VIRTUAL);
8156be56 566 }
40daca54 567 qemu_clock_notify(QEMU_CLOCK_VIRTUAL);
8156be56
PB
568}
569
e76d1798 570void qemu_start_warp_timer(void)
946fb27c 571{
ce78d18c 572 int64_t clock;
946fb27c
PB
573 int64_t deadline;
574
e76d1798 575 if (!use_icount) {
946fb27c
PB
576 return;
577 }
578
8bd7f71d
PD
579 /* Nothing to do if the VM is stopped: QEMU_CLOCK_VIRTUAL timers
580 * do not fire, so computing the deadline does not make sense.
581 */
582 if (!runstate_is_running()) {
583 return;
584 }
585
586 /* warp clock deterministically in record/replay mode */
e76d1798 587 if (!replay_checkpoint(CHECKPOINT_CLOCK_WARP_START)) {
8bd7f71d
PD
588 return;
589 }
590
ce78d18c 591 if (!all_cpu_threads_idle()) {
946fb27c
PB
592 return;
593 }
594
8156be56
PB
595 if (qtest_enabled()) {
596 /* When testing, qtest commands advance icount. */
e76d1798 597 return;
8156be56
PB
598 }
599
ac70aafc 600 /* We want to use the earliest deadline from ALL vm_clocks */
bf2a7ddb 601 clock = qemu_clock_get_ns(QEMU_CLOCK_VIRTUAL_RT);
40daca54 602 deadline = qemu_clock_deadline_ns_all(QEMU_CLOCK_VIRTUAL);
ce78d18c 603 if (deadline < 0) {
d7a0f71d
VC
604 static bool notified;
605 if (!icount_sleep && !notified) {
3dc6f869 606 warn_report("icount sleep disabled and no active timers");
d7a0f71d
VC
607 notified = true;
608 }
ce78d18c 609 return;
ac70aafc
AB
610 }
611
946fb27c
PB
612 if (deadline > 0) {
613 /*
40daca54 614 * Ensure QEMU_CLOCK_VIRTUAL proceeds even when the virtual CPU goes to
946fb27c
PB
615 * sleep. Otherwise, the CPU might be waiting for a future timer
616 * interrupt to wake it up, but the interrupt never comes because
617 * the vCPU isn't running any insns and thus doesn't advance the
40daca54 618 * QEMU_CLOCK_VIRTUAL.
946fb27c 619 */
5045e9d9
VC
620 if (!icount_sleep) {
621 /*
622 * We never let VCPUs sleep in no sleep icount mode.
623 * If there is a pending QEMU_CLOCK_VIRTUAL timer we just advance
624 * to the next QEMU_CLOCK_VIRTUAL event and notify it.
625 * It is useful when we want a deterministic execution time,
626 * isolated from host latencies.
627 */
94377115
PB
628 seqlock_write_lock(&timers_state.vm_clock_seqlock,
629 &timers_state.vm_clock_lock);
c97595d1
EC
630 atomic_set_i64(&timers_state.qemu_icount_bias,
631 timers_state.qemu_icount_bias + deadline);
94377115
PB
632 seqlock_write_unlock(&timers_state.vm_clock_seqlock,
633 &timers_state.vm_clock_lock);
5045e9d9
VC
634 qemu_clock_notify(QEMU_CLOCK_VIRTUAL);
635 } else {
636 /*
637 * We do stop VCPUs and only advance QEMU_CLOCK_VIRTUAL after some
638 * "real" time, (related to the time left until the next event) has
639 * passed. The QEMU_CLOCK_VIRTUAL_RT clock will do this.
640 * This avoids that the warps are visible externally; for example,
641 * you will not be sending network packets continuously instead of
642 * every 100ms.
643 */
94377115
PB
644 seqlock_write_lock(&timers_state.vm_clock_seqlock,
645 &timers_state.vm_clock_lock);
b39e3f34
PD
646 if (timers_state.vm_clock_warp_start == -1
647 || timers_state.vm_clock_warp_start > clock) {
648 timers_state.vm_clock_warp_start = clock;
5045e9d9 649 }
94377115
PB
650 seqlock_write_unlock(&timers_state.vm_clock_seqlock,
651 &timers_state.vm_clock_lock);
b39e3f34
PD
652 timer_mod_anticipate(timers_state.icount_warp_timer,
653 clock + deadline);
ce78d18c 654 }
ac70aafc 655 } else if (deadline == 0) {
40daca54 656 qemu_clock_notify(QEMU_CLOCK_VIRTUAL);
946fb27c
PB
657 }
658}
659
e76d1798
PD
660static void qemu_account_warp_timer(void)
661{
662 if (!use_icount || !icount_sleep) {
663 return;
664 }
665
666 /* Nothing to do if the VM is stopped: QEMU_CLOCK_VIRTUAL timers
667 * do not fire, so computing the deadline does not make sense.
668 */
669 if (!runstate_is_running()) {
670 return;
671 }
672
673 /* warp clock deterministically in record/replay mode */
674 if (!replay_checkpoint(CHECKPOINT_CLOCK_WARP_ACCOUNT)) {
675 return;
676 }
677
b39e3f34 678 timer_del(timers_state.icount_warp_timer);
e76d1798
PD
679 icount_warp_rt();
680}
681
d09eae37
FK
682static bool icount_state_needed(void *opaque)
683{
684 return use_icount;
685}
686
b39e3f34
PD
687static bool warp_timer_state_needed(void *opaque)
688{
689 TimersState *s = opaque;
690 return s->icount_warp_timer != NULL;
691}
692
693static bool adjust_timers_state_needed(void *opaque)
694{
695 TimersState *s = opaque;
696 return s->icount_rt_timer != NULL;
697}
698
699/*
700 * Subsection for warp timer migration is optional, because may not be created
701 */
702static const VMStateDescription icount_vmstate_warp_timer = {
703 .name = "timer/icount/warp_timer",
704 .version_id = 1,
705 .minimum_version_id = 1,
706 .needed = warp_timer_state_needed,
707 .fields = (VMStateField[]) {
708 VMSTATE_INT64(vm_clock_warp_start, TimersState),
709 VMSTATE_TIMER_PTR(icount_warp_timer, TimersState),
710 VMSTATE_END_OF_LIST()
711 }
712};
713
714static const VMStateDescription icount_vmstate_adjust_timers = {
715 .name = "timer/icount/timers",
716 .version_id = 1,
717 .minimum_version_id = 1,
718 .needed = adjust_timers_state_needed,
719 .fields = (VMStateField[]) {
720 VMSTATE_TIMER_PTR(icount_rt_timer, TimersState),
721 VMSTATE_TIMER_PTR(icount_vm_timer, TimersState),
722 VMSTATE_END_OF_LIST()
723 }
724};
725
d09eae37
FK
726/*
727 * This is a subsection for icount migration.
728 */
729static const VMStateDescription icount_vmstate_timers = {
730 .name = "timer/icount",
731 .version_id = 1,
732 .minimum_version_id = 1,
5cd8cada 733 .needed = icount_state_needed,
d09eae37
FK
734 .fields = (VMStateField[]) {
735 VMSTATE_INT64(qemu_icount_bias, TimersState),
736 VMSTATE_INT64(qemu_icount, TimersState),
737 VMSTATE_END_OF_LIST()
b39e3f34
PD
738 },
739 .subsections = (const VMStateDescription*[]) {
740 &icount_vmstate_warp_timer,
741 &icount_vmstate_adjust_timers,
742 NULL
d09eae37
FK
743 }
744};
745
946fb27c
PB
746static const VMStateDescription vmstate_timers = {
747 .name = "timer",
748 .version_id = 2,
749 .minimum_version_id = 1,
35d08458 750 .fields = (VMStateField[]) {
946fb27c 751 VMSTATE_INT64(cpu_ticks_offset, TimersState),
c1ff073c 752 VMSTATE_UNUSED(8),
946fb27c
PB
753 VMSTATE_INT64_V(cpu_clock_offset, TimersState, 2),
754 VMSTATE_END_OF_LIST()
d09eae37 755 },
5cd8cada
JQ
756 .subsections = (const VMStateDescription*[]) {
757 &icount_vmstate_timers,
758 NULL
946fb27c
PB
759 }
760};
761
14e6fe12 762static void cpu_throttle_thread(CPUState *cpu, run_on_cpu_data opaque)
2adcc85d 763{
2adcc85d
JH
764 double pct;
765 double throttle_ratio;
766 long sleeptime_ns;
767
768 if (!cpu_throttle_get_percentage()) {
769 return;
770 }
771
772 pct = (double)cpu_throttle_get_percentage()/100;
773 throttle_ratio = pct / (1 - pct);
774 sleeptime_ns = (long)(throttle_ratio * CPU_THROTTLE_TIMESLICE_NS);
775
776 qemu_mutex_unlock_iothread();
2adcc85d
JH
777 g_usleep(sleeptime_ns / 1000); /* Convert ns to us for usleep call */
778 qemu_mutex_lock_iothread();
90bb0c04 779 atomic_set(&cpu->throttle_thread_scheduled, 0);
2adcc85d
JH
780}
781
782static void cpu_throttle_timer_tick(void *opaque)
783{
784 CPUState *cpu;
785 double pct;
786
787 /* Stop the timer if needed */
788 if (!cpu_throttle_get_percentage()) {
789 return;
790 }
791 CPU_FOREACH(cpu) {
792 if (!atomic_xchg(&cpu->throttle_thread_scheduled, 1)) {
14e6fe12
PB
793 async_run_on_cpu(cpu, cpu_throttle_thread,
794 RUN_ON_CPU_NULL);
2adcc85d
JH
795 }
796 }
797
798 pct = (double)cpu_throttle_get_percentage()/100;
799 timer_mod(throttle_timer, qemu_clock_get_ns(QEMU_CLOCK_VIRTUAL_RT) +
800 CPU_THROTTLE_TIMESLICE_NS / (1-pct));
801}
802
803void cpu_throttle_set(int new_throttle_pct)
804{
805 /* Ensure throttle percentage is within valid range */
806 new_throttle_pct = MIN(new_throttle_pct, CPU_THROTTLE_PCT_MAX);
807 new_throttle_pct = MAX(new_throttle_pct, CPU_THROTTLE_PCT_MIN);
808
809 atomic_set(&throttle_percentage, new_throttle_pct);
810
811 timer_mod(throttle_timer, qemu_clock_get_ns(QEMU_CLOCK_VIRTUAL_RT) +
812 CPU_THROTTLE_TIMESLICE_NS);
813}
814
815void cpu_throttle_stop(void)
816{
817 atomic_set(&throttle_percentage, 0);
818}
819
820bool cpu_throttle_active(void)
821{
822 return (cpu_throttle_get_percentage() != 0);
823}
824
825int cpu_throttle_get_percentage(void)
826{
827 return atomic_read(&throttle_percentage);
828}
829
4603ea01
PD
830void cpu_ticks_init(void)
831{
ccdb3c1f 832 seqlock_init(&timers_state.vm_clock_seqlock);
87a09cdc 833 qemu_spin_init(&timers_state.vm_clock_lock);
4603ea01 834 vmstate_register(NULL, 0, &vmstate_timers, &timers_state);
2adcc85d
JH
835 throttle_timer = timer_new_ns(QEMU_CLOCK_VIRTUAL_RT,
836 cpu_throttle_timer_tick, NULL);
4603ea01
PD
837}
838
1ad9580b 839void configure_icount(QemuOpts *opts, Error **errp)
946fb27c 840{
1ad9580b 841 const char *option;
a8bfac37 842 char *rem_str = NULL;
1ad9580b 843
1ad9580b 844 option = qemu_opt_get(opts, "shift");
946fb27c 845 if (!option) {
a8bfac37
ST
846 if (qemu_opt_get(opts, "align") != NULL) {
847 error_setg(errp, "Please specify shift option when using align");
848 }
946fb27c
PB
849 return;
850 }
f1f4b57e
VC
851
852 icount_sleep = qemu_opt_get_bool(opts, "sleep", true);
5045e9d9 853 if (icount_sleep) {
b39e3f34 854 timers_state.icount_warp_timer = timer_new_ns(QEMU_CLOCK_VIRTUAL_RT,
e76d1798 855 icount_timer_cb, NULL);
5045e9d9 856 }
f1f4b57e 857
a8bfac37 858 icount_align_option = qemu_opt_get_bool(opts, "align", false);
f1f4b57e
VC
859
860 if (icount_align_option && !icount_sleep) {
778d9f9b 861 error_setg(errp, "align=on and sleep=off are incompatible");
f1f4b57e 862 }
946fb27c 863 if (strcmp(option, "auto") != 0) {
a8bfac37 864 errno = 0;
c1ff073c 865 timers_state.icount_time_shift = strtol(option, &rem_str, 0);
a8bfac37
ST
866 if (errno != 0 || *rem_str != '\0' || !strlen(option)) {
867 error_setg(errp, "icount: Invalid shift value");
868 }
946fb27c
PB
869 use_icount = 1;
870 return;
a8bfac37
ST
871 } else if (icount_align_option) {
872 error_setg(errp, "shift=auto and align=on are incompatible");
f1f4b57e 873 } else if (!icount_sleep) {
778d9f9b 874 error_setg(errp, "shift=auto and sleep=off are incompatible");
946fb27c
PB
875 }
876
877 use_icount = 2;
878
879 /* 125MIPS seems a reasonable initial guess at the guest speed.
880 It will be corrected fairly quickly anyway. */
c1ff073c 881 timers_state.icount_time_shift = 3;
946fb27c
PB
882
883 /* Have both realtime and virtual time triggers for speed adjustment.
884 The realtime trigger catches emulated time passing too slowly,
885 the virtual time trigger catches emulated time passing too fast.
886 Realtime triggers occur even when idle, so use them less frequently
887 than VM triggers. */
b39e3f34
PD
888 timers_state.vm_clock_warp_start = -1;
889 timers_state.icount_rt_timer = timer_new_ms(QEMU_CLOCK_VIRTUAL_RT,
bf2a7ddb 890 icount_adjust_rt, NULL);
b39e3f34 891 timer_mod(timers_state.icount_rt_timer,
bf2a7ddb 892 qemu_clock_get_ms(QEMU_CLOCK_VIRTUAL_RT) + 1000);
b39e3f34 893 timers_state.icount_vm_timer = timer_new_ns(QEMU_CLOCK_VIRTUAL,
40daca54 894 icount_adjust_vm, NULL);
b39e3f34 895 timer_mod(timers_state.icount_vm_timer,
40daca54 896 qemu_clock_get_ns(QEMU_CLOCK_VIRTUAL) +
73bcb24d 897 NANOSECONDS_PER_SECOND / 10);
946fb27c
PB
898}
899
6546706d
AB
900/***********************************************************/
901/* TCG vCPU kick timer
902 *
903 * The kick timer is responsible for moving single threaded vCPU
904 * emulation on to the next vCPU. If more than one vCPU is running a
905 * timer event with force a cpu->exit so the next vCPU can get
906 * scheduled.
907 *
908 * The timer is removed if all vCPUs are idle and restarted again once
909 * idleness is complete.
910 */
911
912static QEMUTimer *tcg_kick_vcpu_timer;
791158d9 913static CPUState *tcg_current_rr_cpu;
6546706d
AB
914
915#define TCG_KICK_PERIOD (NANOSECONDS_PER_SECOND / 10)
916
917static inline int64_t qemu_tcg_next_kick(void)
918{
919 return qemu_clock_get_ns(QEMU_CLOCK_VIRTUAL) + TCG_KICK_PERIOD;
920}
921
791158d9
AB
922/* Kick the currently round-robin scheduled vCPU */
923static void qemu_cpu_kick_rr_cpu(void)
924{
925 CPUState *cpu;
791158d9
AB
926 do {
927 cpu = atomic_mb_read(&tcg_current_rr_cpu);
928 if (cpu) {
929 cpu_exit(cpu);
930 }
931 } while (cpu != atomic_mb_read(&tcg_current_rr_cpu));
932}
933
6b8f0187
PB
934static void do_nothing(CPUState *cpu, run_on_cpu_data unused)
935{
936}
937
3f53bc61
PB
938void qemu_timer_notify_cb(void *opaque, QEMUClockType type)
939{
6b8f0187
PB
940 if (!use_icount || type != QEMU_CLOCK_VIRTUAL) {
941 qemu_notify_event();
942 return;
943 }
944
c52e7132
PM
945 if (qemu_in_vcpu_thread()) {
946 /* A CPU is currently running; kick it back out to the
947 * tcg_cpu_exec() loop so it will recalculate its
948 * icount deadline immediately.
949 */
950 qemu_cpu_kick(current_cpu);
951 } else if (first_cpu) {
6b8f0187
PB
952 /* qemu_cpu_kick is not enough to kick a halted CPU out of
953 * qemu_tcg_wait_io_event. async_run_on_cpu, instead,
954 * causes cpu_thread_is_idle to return false. This way,
955 * handle_icount_deadline can run.
c52e7132
PM
956 * If we have no CPUs at all for some reason, we don't
957 * need to do anything.
6b8f0187
PB
958 */
959 async_run_on_cpu(first_cpu, do_nothing, RUN_ON_CPU_NULL);
960 }
3f53bc61
PB
961}
962
6546706d
AB
963static void kick_tcg_thread(void *opaque)
964{
965 timer_mod(tcg_kick_vcpu_timer, qemu_tcg_next_kick());
791158d9 966 qemu_cpu_kick_rr_cpu();
6546706d
AB
967}
968
969static void start_tcg_kick_timer(void)
970{
db08b687
PB
971 assert(!mttcg_enabled);
972 if (!tcg_kick_vcpu_timer && CPU_NEXT(first_cpu)) {
6546706d
AB
973 tcg_kick_vcpu_timer = timer_new_ns(QEMU_CLOCK_VIRTUAL,
974 kick_tcg_thread, NULL);
975 timer_mod(tcg_kick_vcpu_timer, qemu_tcg_next_kick());
976 }
977}
978
979static void stop_tcg_kick_timer(void)
980{
db08b687 981 assert(!mttcg_enabled);
6546706d
AB
982 if (tcg_kick_vcpu_timer) {
983 timer_del(tcg_kick_vcpu_timer);
984 tcg_kick_vcpu_timer = NULL;
985 }
986}
987
296af7c9
BS
988/***********************************************************/
989void hw_error(const char *fmt, ...)
990{
991 va_list ap;
55e5c285 992 CPUState *cpu;
296af7c9
BS
993
994 va_start(ap, fmt);
995 fprintf(stderr, "qemu: hardware error: ");
996 vfprintf(stderr, fmt, ap);
997 fprintf(stderr, "\n");
bdc44640 998 CPU_FOREACH(cpu) {
55e5c285 999 fprintf(stderr, "CPU #%d:\n", cpu->cpu_index);
878096ee 1000 cpu_dump_state(cpu, stderr, fprintf, CPU_DUMP_FPU);
296af7c9
BS
1001 }
1002 va_end(ap);
1003 abort();
1004}
1005
1006void cpu_synchronize_all_states(void)
1007{
182735ef 1008 CPUState *cpu;
296af7c9 1009
bdc44640 1010 CPU_FOREACH(cpu) {
182735ef 1011 cpu_synchronize_state(cpu);
c97d6d2c
SAGDR
1012 /* TODO: move to cpu_synchronize_state() */
1013 if (hvf_enabled()) {
1014 hvf_cpu_synchronize_state(cpu);
1015 }
296af7c9
BS
1016 }
1017}
1018
1019void cpu_synchronize_all_post_reset(void)
1020{
182735ef 1021 CPUState *cpu;
296af7c9 1022
bdc44640 1023 CPU_FOREACH(cpu) {
182735ef 1024 cpu_synchronize_post_reset(cpu);
c97d6d2c
SAGDR
1025 /* TODO: move to cpu_synchronize_post_reset() */
1026 if (hvf_enabled()) {
1027 hvf_cpu_synchronize_post_reset(cpu);
1028 }
296af7c9
BS
1029 }
1030}
1031
1032void cpu_synchronize_all_post_init(void)
1033{
182735ef 1034 CPUState *cpu;
296af7c9 1035
bdc44640 1036 CPU_FOREACH(cpu) {
182735ef 1037 cpu_synchronize_post_init(cpu);
c97d6d2c
SAGDR
1038 /* TODO: move to cpu_synchronize_post_init() */
1039 if (hvf_enabled()) {
1040 hvf_cpu_synchronize_post_init(cpu);
1041 }
296af7c9
BS
1042 }
1043}
1044
75e972da
DG
1045void cpu_synchronize_all_pre_loadvm(void)
1046{
1047 CPUState *cpu;
1048
1049 CPU_FOREACH(cpu) {
1050 cpu_synchronize_pre_loadvm(cpu);
1051 }
1052}
1053
4486e89c 1054static int do_vm_stop(RunState state, bool send_stop)
296af7c9 1055{
56983463
KW
1056 int ret = 0;
1057
1354869c 1058 if (runstate_is_running()) {
296af7c9 1059 cpu_disable_ticks();
296af7c9 1060 pause_all_vcpus();
f5bbfba1 1061 runstate_set(state);
1dfb4dd9 1062 vm_state_notify(0, state);
4486e89c 1063 if (send_stop) {
3ab72385 1064 qapi_event_send_stop();
4486e89c 1065 }
296af7c9 1066 }
56983463 1067
594a45ce 1068 bdrv_drain_all();
6d0ceb80 1069 replay_disable_events();
22af08ea 1070 ret = bdrv_flush_all();
594a45ce 1071
56983463 1072 return ret;
296af7c9
BS
1073}
1074
4486e89c
SH
1075/* Special vm_stop() variant for terminating the process. Historically clients
1076 * did not expect a QMP STOP event and so we need to retain compatibility.
1077 */
1078int vm_shutdown(void)
1079{
1080 return do_vm_stop(RUN_STATE_SHUTDOWN, false);
1081}
1082
a1fcaa73 1083static bool cpu_can_run(CPUState *cpu)
296af7c9 1084{
4fdeee7c 1085 if (cpu->stop) {
a1fcaa73 1086 return false;
0ab07c62 1087 }
321bc0b2 1088 if (cpu_is_stopped(cpu)) {
a1fcaa73 1089 return false;
0ab07c62 1090 }
a1fcaa73 1091 return true;
296af7c9
BS
1092}
1093
91325046 1094static void cpu_handle_guest_debug(CPUState *cpu)
83f338f7 1095{
64f6b346 1096 gdb_set_stop_cpu(cpu);
8cf71710 1097 qemu_system_debug_request();
f324e766 1098 cpu->stopped = true;
3c638d06
JK
1099}
1100
6d9cb73c
JK
1101#ifdef CONFIG_LINUX
1102static void sigbus_reraise(void)
1103{
1104 sigset_t set;
1105 struct sigaction action;
1106
1107 memset(&action, 0, sizeof(action));
1108 action.sa_handler = SIG_DFL;
1109 if (!sigaction(SIGBUS, &action, NULL)) {
1110 raise(SIGBUS);
1111 sigemptyset(&set);
1112 sigaddset(&set, SIGBUS);
a2d1761d 1113 pthread_sigmask(SIG_UNBLOCK, &set, NULL);
6d9cb73c
JK
1114 }
1115 perror("Failed to re-raise SIGBUS!\n");
1116 abort();
1117}
1118
d98d4072 1119static void sigbus_handler(int n, siginfo_t *siginfo, void *ctx)
6d9cb73c 1120{
a16fc07e
PB
1121 if (siginfo->si_code != BUS_MCEERR_AO && siginfo->si_code != BUS_MCEERR_AR) {
1122 sigbus_reraise();
1123 }
1124
2ae41db2
PB
1125 if (current_cpu) {
1126 /* Called asynchronously in VCPU thread. */
1127 if (kvm_on_sigbus_vcpu(current_cpu, siginfo->si_code, siginfo->si_addr)) {
1128 sigbus_reraise();
1129 }
1130 } else {
1131 /* Called synchronously (via signalfd) in main thread. */
1132 if (kvm_on_sigbus(siginfo->si_code, siginfo->si_addr)) {
1133 sigbus_reraise();
1134 }
6d9cb73c
JK
1135 }
1136}
1137
1138static void qemu_init_sigbus(void)
1139{
1140 struct sigaction action;
1141
1142 memset(&action, 0, sizeof(action));
1143 action.sa_flags = SA_SIGINFO;
d98d4072 1144 action.sa_sigaction = sigbus_handler;
6d9cb73c
JK
1145 sigaction(SIGBUS, &action, NULL);
1146
1147 prctl(PR_MCE_KILL, PR_MCE_KILL_SET, PR_MCE_KILL_EARLY, 0, 0);
1148}
6d9cb73c 1149#else /* !CONFIG_LINUX */
6d9cb73c
JK
1150static void qemu_init_sigbus(void)
1151{
1152}
a16fc07e 1153#endif /* !CONFIG_LINUX */
ff48eb5f 1154
b2532d88 1155static QemuMutex qemu_global_mutex;
296af7c9
BS
1156
1157static QemuThread io_thread;
1158
296af7c9
BS
1159/* cpu creation */
1160static QemuCond qemu_cpu_cond;
1161/* system init */
296af7c9
BS
1162static QemuCond qemu_pause_cond;
1163
d3b12f5d 1164void qemu_init_cpu_loop(void)
296af7c9 1165{
6d9cb73c 1166 qemu_init_sigbus();
ed94592b 1167 qemu_cond_init(&qemu_cpu_cond);
ed94592b 1168 qemu_cond_init(&qemu_pause_cond);
296af7c9 1169 qemu_mutex_init(&qemu_global_mutex);
296af7c9 1170
b7680cb6 1171 qemu_thread_get_self(&io_thread);
296af7c9
BS
1172}
1173
14e6fe12 1174void run_on_cpu(CPUState *cpu, run_on_cpu_func func, run_on_cpu_data data)
e82bcec2 1175{
d148d90e 1176 do_run_on_cpu(cpu, func, data, &qemu_global_mutex);
3c02270d
CV
1177}
1178
4c055ab5
GZ
1179static void qemu_kvm_destroy_vcpu(CPUState *cpu)
1180{
1181 if (kvm_destroy_vcpu(cpu) < 0) {
1182 error_report("kvm_destroy_vcpu failed");
1183 exit(EXIT_FAILURE);
1184 }
1185}
1186
1187static void qemu_tcg_destroy_vcpu(CPUState *cpu)
1188{
1189}
1190
ebd05fea
DH
1191static void qemu_cpu_stop(CPUState *cpu, bool exit)
1192{
1193 g_assert(qemu_cpu_is_self(cpu));
1194 cpu->stop = false;
1195 cpu->stopped = true;
1196 if (exit) {
1197 cpu_exit(cpu);
1198 }
1199 qemu_cond_broadcast(&qemu_pause_cond);
1200}
1201
509a0d78 1202static void qemu_wait_io_event_common(CPUState *cpu)
296af7c9 1203{
37257942 1204 atomic_mb_set(&cpu->thread_kicked, false);
4fdeee7c 1205 if (cpu->stop) {
ebd05fea 1206 qemu_cpu_stop(cpu, false);
296af7c9 1207 }
a5403c69 1208 process_queued_cpu_work(cpu);
37257942
AB
1209}
1210
db08b687 1211static void qemu_tcg_rr_wait_io_event(CPUState *cpu)
37257942 1212{
db08b687 1213 while (all_cpu_threads_idle()) {
6546706d 1214 stop_tcg_kick_timer();
d5f8d613 1215 qemu_cond_wait(cpu->halt_cond, &qemu_global_mutex);
16400322 1216 }
296af7c9 1217
6546706d
AB
1218 start_tcg_kick_timer();
1219
37257942 1220 qemu_wait_io_event_common(cpu);
296af7c9
BS
1221}
1222
db08b687 1223static void qemu_wait_io_event(CPUState *cpu)
296af7c9 1224{
a98ae1d8 1225 while (cpu_thread_is_idle(cpu)) {
f5c121b8 1226 qemu_cond_wait(cpu->halt_cond, &qemu_global_mutex);
16400322 1227 }
296af7c9 1228
db08b687
PB
1229#ifdef _WIN32
1230 /* Eat dummy APC queued by qemu_cpu_kick_thread. */
1231 if (!tcg_enabled()) {
1232 SleepEx(0, TRUE);
c97d6d2c 1233 }
db08b687 1234#endif
c97d6d2c
SAGDR
1235 qemu_wait_io_event_common(cpu);
1236}
1237
7e97cd88 1238static void *qemu_kvm_cpu_thread_fn(void *arg)
296af7c9 1239{
48a106bd 1240 CPUState *cpu = arg;
84b4915d 1241 int r;
296af7c9 1242
ab28bd23
PB
1243 rcu_register_thread();
1244
2e7f7a3c 1245 qemu_mutex_lock_iothread();
814e612e 1246 qemu_thread_get_self(cpu->thread);
9f09e18a 1247 cpu->thread_id = qemu_get_thread_id();
626cf8f4 1248 cpu->can_do_io = 1;
4917cf44 1249 current_cpu = cpu;
296af7c9 1250
504134d2 1251 r = kvm_init_vcpu(cpu);
84b4915d 1252 if (r < 0) {
493d89bf 1253 error_report("kvm_init_vcpu failed: %s", strerror(-r));
84b4915d
JK
1254 exit(1);
1255 }
296af7c9 1256
18268b60 1257 kvm_init_cpu_signals(cpu);
296af7c9
BS
1258
1259 /* signal CPU creation */
61a46217 1260 cpu->created = true;
296af7c9
BS
1261 qemu_cond_signal(&qemu_cpu_cond);
1262
4c055ab5 1263 do {
a1fcaa73 1264 if (cpu_can_run(cpu)) {
1458c363 1265 r = kvm_cpu_exec(cpu);
83f338f7 1266 if (r == EXCP_DEBUG) {
91325046 1267 cpu_handle_guest_debug(cpu);
83f338f7 1268 }
0ab07c62 1269 }
db08b687 1270 qemu_wait_io_event(cpu);
4c055ab5 1271 } while (!cpu->unplug || cpu_can_run(cpu));
296af7c9 1272
4c055ab5 1273 qemu_kvm_destroy_vcpu(cpu);
2c579042
BR
1274 cpu->created = false;
1275 qemu_cond_signal(&qemu_cpu_cond);
4c055ab5 1276 qemu_mutex_unlock_iothread();
57615ed5 1277 rcu_unregister_thread();
296af7c9
BS
1278 return NULL;
1279}
1280
c7f0f3b1
AL
1281static void *qemu_dummy_cpu_thread_fn(void *arg)
1282{
1283#ifdef _WIN32
493d89bf 1284 error_report("qtest is not supported under Windows");
c7f0f3b1
AL
1285 exit(1);
1286#else
10a9021d 1287 CPUState *cpu = arg;
c7f0f3b1
AL
1288 sigset_t waitset;
1289 int r;
1290
ab28bd23
PB
1291 rcu_register_thread();
1292
c7f0f3b1 1293 qemu_mutex_lock_iothread();
814e612e 1294 qemu_thread_get_self(cpu->thread);
9f09e18a 1295 cpu->thread_id = qemu_get_thread_id();
626cf8f4 1296 cpu->can_do_io = 1;
37257942 1297 current_cpu = cpu;
c7f0f3b1
AL
1298
1299 sigemptyset(&waitset);
1300 sigaddset(&waitset, SIG_IPI);
1301
1302 /* signal CPU creation */
61a46217 1303 cpu->created = true;
c7f0f3b1
AL
1304 qemu_cond_signal(&qemu_cpu_cond);
1305
d2831ab0 1306 do {
c7f0f3b1
AL
1307 qemu_mutex_unlock_iothread();
1308 do {
1309 int sig;
1310 r = sigwait(&waitset, &sig);
1311 } while (r == -1 && (errno == EAGAIN || errno == EINTR));
1312 if (r == -1) {
1313 perror("sigwait");
1314 exit(1);
1315 }
1316 qemu_mutex_lock_iothread();
db08b687 1317 qemu_wait_io_event(cpu);
d2831ab0 1318 } while (!cpu->unplug);
c7f0f3b1 1319
d2831ab0 1320 rcu_unregister_thread();
c7f0f3b1
AL
1321 return NULL;
1322#endif
1323}
1324
1be7fcb8
AB
1325static int64_t tcg_get_icount_limit(void)
1326{
1327 int64_t deadline;
1328
1329 if (replay_mode != REPLAY_MODE_PLAY) {
1330 deadline = qemu_clock_deadline_ns_all(QEMU_CLOCK_VIRTUAL);
1331
1332 /* Maintain prior (possibly buggy) behaviour where if no deadline
1333 * was set (as there is no QEMU_CLOCK_VIRTUAL timer) or it is more than
1334 * INT32_MAX nanoseconds ahead, we still use INT32_MAX
1335 * nanoseconds.
1336 */
1337 if ((deadline < 0) || (deadline > INT32_MAX)) {
1338 deadline = INT32_MAX;
1339 }
1340
1341 return qemu_icount_round(deadline);
1342 } else {
1343 return replay_get_instructions();
1344 }
1345}
1346
12e9700d
AB
1347static void handle_icount_deadline(void)
1348{
6b8f0187 1349 assert(qemu_in_vcpu_thread());
12e9700d
AB
1350 if (use_icount) {
1351 int64_t deadline =
1352 qemu_clock_deadline_ns_all(QEMU_CLOCK_VIRTUAL);
1353
1354 if (deadline == 0) {
6b8f0187 1355 /* Wake up other AioContexts. */
12e9700d 1356 qemu_clock_notify(QEMU_CLOCK_VIRTUAL);
6b8f0187 1357 qemu_clock_run_timers(QEMU_CLOCK_VIRTUAL);
12e9700d
AB
1358 }
1359 }
1360}
1361
05248382 1362static void prepare_icount_for_run(CPUState *cpu)
1be7fcb8 1363{
1be7fcb8 1364 if (use_icount) {
eda5f7c6 1365 int insns_left;
05248382
AB
1366
1367 /* These should always be cleared by process_icount_data after
1368 * each vCPU execution. However u16.high can be raised
1369 * asynchronously by cpu_exit/cpu_interrupt/tcg_handle_interrupt
1370 */
1371 g_assert(cpu->icount_decr.u16.low == 0);
1372 g_assert(cpu->icount_extra == 0);
1373
eda5f7c6
AB
1374 cpu->icount_budget = tcg_get_icount_limit();
1375 insns_left = MIN(0xffff, cpu->icount_budget);
1376 cpu->icount_decr.u16.low = insns_left;
1377 cpu->icount_extra = cpu->icount_budget - insns_left;
d759c951
AB
1378
1379 replay_mutex_lock();
1be7fcb8 1380 }
05248382
AB
1381}
1382
1383static void process_icount_data(CPUState *cpu)
1384{
1be7fcb8 1385 if (use_icount) {
e4cd9657 1386 /* Account for executed instructions */
512d3c80 1387 cpu_update_icount(cpu);
05248382
AB
1388
1389 /* Reset the counters */
1390 cpu->icount_decr.u16.low = 0;
1be7fcb8 1391 cpu->icount_extra = 0;
e4cd9657
AB
1392 cpu->icount_budget = 0;
1393
1be7fcb8 1394 replay_account_executed_instructions();
d759c951
AB
1395
1396 replay_mutex_unlock();
1be7fcb8 1397 }
05248382
AB
1398}
1399
1400
1401static int tcg_cpu_exec(CPUState *cpu)
1402{
1403 int ret;
1404#ifdef CONFIG_PROFILER
1405 int64_t ti;
1406#endif
1407
f28d0dfd 1408 assert(tcg_enabled());
05248382
AB
1409#ifdef CONFIG_PROFILER
1410 ti = profile_getclock();
1411#endif
05248382
AB
1412 cpu_exec_start(cpu);
1413 ret = cpu_exec(cpu);
1414 cpu_exec_end(cpu);
05248382
AB
1415#ifdef CONFIG_PROFILER
1416 tcg_time += profile_getclock() - ti;
1417#endif
1be7fcb8
AB
1418 return ret;
1419}
1420
c93bbbef
AB
1421/* Destroy any remaining vCPUs which have been unplugged and have
1422 * finished running
1423 */
1424static void deal_with_unplugged_cpus(void)
1be7fcb8 1425{
c93bbbef 1426 CPUState *cpu;
1be7fcb8 1427
c93bbbef
AB
1428 CPU_FOREACH(cpu) {
1429 if (cpu->unplug && !cpu_can_run(cpu)) {
1430 qemu_tcg_destroy_vcpu(cpu);
1431 cpu->created = false;
1432 qemu_cond_signal(&qemu_cpu_cond);
1be7fcb8
AB
1433 break;
1434 }
1435 }
1be7fcb8 1436}
bdb7ca67 1437
6546706d
AB
1438/* Single-threaded TCG
1439 *
1440 * In the single-threaded case each vCPU is simulated in turn. If
1441 * there is more than a single vCPU we create a simple timer to kick
1442 * the vCPU and ensure we don't get stuck in a tight loop in one vCPU.
1443 * This is done explicitly rather than relying on side-effects
1444 * elsewhere.
1445 */
1446
37257942 1447static void *qemu_tcg_rr_cpu_thread_fn(void *arg)
296af7c9 1448{
c3586ba7 1449 CPUState *cpu = arg;
296af7c9 1450
f28d0dfd 1451 assert(tcg_enabled());
ab28bd23 1452 rcu_register_thread();
3468b59e 1453 tcg_register_thread();
ab28bd23 1454
2e7f7a3c 1455 qemu_mutex_lock_iothread();
814e612e 1456 qemu_thread_get_self(cpu->thread);
296af7c9 1457
5a9c973b
DH
1458 cpu->thread_id = qemu_get_thread_id();
1459 cpu->created = true;
1460 cpu->can_do_io = 1;
296af7c9
BS
1461 qemu_cond_signal(&qemu_cpu_cond);
1462
fa7d1867 1463 /* wait for initial kick-off after machine start */
c28e399c 1464 while (first_cpu->stopped) {
d5f8d613 1465 qemu_cond_wait(first_cpu->halt_cond, &qemu_global_mutex);
8e564b4e
JK
1466
1467 /* process any pending work */
bdc44640 1468 CPU_FOREACH(cpu) {
37257942 1469 current_cpu = cpu;
182735ef 1470 qemu_wait_io_event_common(cpu);
8e564b4e 1471 }
0ab07c62 1472 }
296af7c9 1473
6546706d
AB
1474 start_tcg_kick_timer();
1475
c93bbbef
AB
1476 cpu = first_cpu;
1477
e5143e30
AB
1478 /* process any pending work */
1479 cpu->exit_request = 1;
1480
296af7c9 1481 while (1) {
d759c951
AB
1482 qemu_mutex_unlock_iothread();
1483 replay_mutex_lock();
1484 qemu_mutex_lock_iothread();
c93bbbef
AB
1485 /* Account partial waits to QEMU_CLOCK_VIRTUAL. */
1486 qemu_account_warp_timer();
1487
6b8f0187
PB
1488 /* Run the timers here. This is much more efficient than
1489 * waking up the I/O thread and waiting for completion.
1490 */
1491 handle_icount_deadline();
1492
d759c951
AB
1493 replay_mutex_unlock();
1494
c93bbbef
AB
1495 if (!cpu) {
1496 cpu = first_cpu;
1497 }
1498
e5143e30
AB
1499 while (cpu && !cpu->queued_work_first && !cpu->exit_request) {
1500
791158d9 1501 atomic_mb_set(&tcg_current_rr_cpu, cpu);
37257942 1502 current_cpu = cpu;
c93bbbef
AB
1503
1504 qemu_clock_enable(QEMU_CLOCK_VIRTUAL,
1505 (cpu->singlestep_enabled & SSTEP_NOTIMER) == 0);
1506
1507 if (cpu_can_run(cpu)) {
1508 int r;
05248382 1509
d759c951 1510 qemu_mutex_unlock_iothread();
05248382
AB
1511 prepare_icount_for_run(cpu);
1512
c93bbbef 1513 r = tcg_cpu_exec(cpu);
05248382
AB
1514
1515 process_icount_data(cpu);
d759c951 1516 qemu_mutex_lock_iothread();
05248382 1517
c93bbbef
AB
1518 if (r == EXCP_DEBUG) {
1519 cpu_handle_guest_debug(cpu);
1520 break;
08e73c48
PK
1521 } else if (r == EXCP_ATOMIC) {
1522 qemu_mutex_unlock_iothread();
1523 cpu_exec_step_atomic(cpu);
1524 qemu_mutex_lock_iothread();
1525 break;
c93bbbef 1526 }
37257942 1527 } else if (cpu->stop) {
c93bbbef
AB
1528 if (cpu->unplug) {
1529 cpu = CPU_NEXT(cpu);
1530 }
1531 break;
1532 }
1533
e5143e30
AB
1534 cpu = CPU_NEXT(cpu);
1535 } /* while (cpu && !cpu->exit_request).. */
1536
791158d9
AB
1537 /* Does not need atomic_mb_set because a spurious wakeup is okay. */
1538 atomic_set(&tcg_current_rr_cpu, NULL);
c93bbbef 1539
e5143e30
AB
1540 if (cpu && cpu->exit_request) {
1541 atomic_mb_set(&cpu->exit_request, 0);
1542 }
ac70aafc 1543
068a5ea0 1544 qemu_tcg_rr_wait_io_event(cpu ? cpu : first_cpu);
c93bbbef 1545 deal_with_unplugged_cpus();
296af7c9
BS
1546 }
1547
9b0605f9 1548 rcu_unregister_thread();
296af7c9
BS
1549 return NULL;
1550}
1551
b0cb0a66
VP
1552static void *qemu_hax_cpu_thread_fn(void *arg)
1553{
1554 CPUState *cpu = arg;
1555 int r;
b3d3a426 1556
9857c2d2 1557 rcu_register_thread();
b3d3a426 1558 qemu_mutex_lock_iothread();
b0cb0a66 1559 qemu_thread_get_self(cpu->thread);
b0cb0a66
VP
1560
1561 cpu->thread_id = qemu_get_thread_id();
1562 cpu->created = true;
1563 cpu->halted = 0;
1564 current_cpu = cpu;
1565
1566 hax_init_vcpu(cpu);
1567 qemu_cond_signal(&qemu_cpu_cond);
1568
9857c2d2 1569 do {
b0cb0a66
VP
1570 if (cpu_can_run(cpu)) {
1571 r = hax_smp_cpu_exec(cpu);
1572 if (r == EXCP_DEBUG) {
1573 cpu_handle_guest_debug(cpu);
1574 }
1575 }
1576
db08b687 1577 qemu_wait_io_event(cpu);
9857c2d2
PB
1578 } while (!cpu->unplug || cpu_can_run(cpu));
1579 rcu_unregister_thread();
b0cb0a66
VP
1580 return NULL;
1581}
1582
c97d6d2c
SAGDR
1583/* The HVF-specific vCPU thread function. This one should only run when the host
1584 * CPU supports the VMX "unrestricted guest" feature. */
1585static void *qemu_hvf_cpu_thread_fn(void *arg)
1586{
1587 CPUState *cpu = arg;
1588
1589 int r;
1590
1591 assert(hvf_enabled());
1592
1593 rcu_register_thread();
1594
1595 qemu_mutex_lock_iothread();
1596 qemu_thread_get_self(cpu->thread);
1597
1598 cpu->thread_id = qemu_get_thread_id();
1599 cpu->can_do_io = 1;
1600 current_cpu = cpu;
1601
1602 hvf_init_vcpu(cpu);
1603
1604 /* signal CPU creation */
1605 cpu->created = true;
1606 qemu_cond_signal(&qemu_cpu_cond);
1607
1608 do {
1609 if (cpu_can_run(cpu)) {
1610 r = hvf_vcpu_exec(cpu);
1611 if (r == EXCP_DEBUG) {
1612 cpu_handle_guest_debug(cpu);
1613 }
1614 }
db08b687 1615 qemu_wait_io_event(cpu);
c97d6d2c
SAGDR
1616 } while (!cpu->unplug || cpu_can_run(cpu));
1617
1618 hvf_vcpu_destroy(cpu);
1619 cpu->created = false;
1620 qemu_cond_signal(&qemu_cpu_cond);
1621 qemu_mutex_unlock_iothread();
8178e637 1622 rcu_unregister_thread();
c97d6d2c
SAGDR
1623 return NULL;
1624}
1625
19306806
JTV
1626static void *qemu_whpx_cpu_thread_fn(void *arg)
1627{
1628 CPUState *cpu = arg;
1629 int r;
1630
1631 rcu_register_thread();
1632
1633 qemu_mutex_lock_iothread();
1634 qemu_thread_get_self(cpu->thread);
1635 cpu->thread_id = qemu_get_thread_id();
1636 current_cpu = cpu;
1637
1638 r = whpx_init_vcpu(cpu);
1639 if (r < 0) {
1640 fprintf(stderr, "whpx_init_vcpu failed: %s\n", strerror(-r));
1641 exit(1);
1642 }
1643
1644 /* signal CPU creation */
1645 cpu->created = true;
1646 qemu_cond_signal(&qemu_cpu_cond);
1647
1648 do {
1649 if (cpu_can_run(cpu)) {
1650 r = whpx_vcpu_exec(cpu);
1651 if (r == EXCP_DEBUG) {
1652 cpu_handle_guest_debug(cpu);
1653 }
1654 }
1655 while (cpu_thread_is_idle(cpu)) {
1656 qemu_cond_wait(cpu->halt_cond, &qemu_global_mutex);
1657 }
1658 qemu_wait_io_event_common(cpu);
1659 } while (!cpu->unplug || cpu_can_run(cpu));
1660
1661 whpx_destroy_vcpu(cpu);
1662 cpu->created = false;
1663 qemu_cond_signal(&qemu_cpu_cond);
1664 qemu_mutex_unlock_iothread();
1665 rcu_unregister_thread();
c97d6d2c
SAGDR
1666 return NULL;
1667}
1668
b0cb0a66
VP
1669#ifdef _WIN32
1670static void CALLBACK dummy_apc_func(ULONG_PTR unused)
1671{
1672}
1673#endif
1674
37257942
AB
1675/* Multi-threaded TCG
1676 *
1677 * In the multi-threaded case each vCPU has its own thread. The TLS
1678 * variable current_cpu can be used deep in the code to find the
1679 * current CPUState for a given thread.
1680 */
1681
1682static void *qemu_tcg_cpu_thread_fn(void *arg)
1683{
1684 CPUState *cpu = arg;
1685
f28d0dfd 1686 assert(tcg_enabled());
bf51c720
AB
1687 g_assert(!use_icount);
1688
37257942 1689 rcu_register_thread();
3468b59e 1690 tcg_register_thread();
37257942
AB
1691
1692 qemu_mutex_lock_iothread();
1693 qemu_thread_get_self(cpu->thread);
1694
1695 cpu->thread_id = qemu_get_thread_id();
1696 cpu->created = true;
1697 cpu->can_do_io = 1;
1698 current_cpu = cpu;
1699 qemu_cond_signal(&qemu_cpu_cond);
1700
1701 /* process any pending work */
1702 cpu->exit_request = 1;
1703
54961aac 1704 do {
37257942
AB
1705 if (cpu_can_run(cpu)) {
1706 int r;
d759c951 1707 qemu_mutex_unlock_iothread();
37257942 1708 r = tcg_cpu_exec(cpu);
d759c951 1709 qemu_mutex_lock_iothread();
37257942
AB
1710 switch (r) {
1711 case EXCP_DEBUG:
1712 cpu_handle_guest_debug(cpu);
1713 break;
1714 case EXCP_HALTED:
1715 /* during start-up the vCPU is reset and the thread is
1716 * kicked several times. If we don't ensure we go back
1717 * to sleep in the halted state we won't cleanly
1718 * start-up when the vCPU is enabled.
1719 *
1720 * cpu->halted should ensure we sleep in wait_io_event
1721 */
1722 g_assert(cpu->halted);
1723 break;
08e73c48
PK
1724 case EXCP_ATOMIC:
1725 qemu_mutex_unlock_iothread();
1726 cpu_exec_step_atomic(cpu);
1727 qemu_mutex_lock_iothread();
37257942
AB
1728 default:
1729 /* Ignore everything else? */
1730 break;
1731 }
1732 }
1733
37257942 1734 atomic_mb_set(&cpu->exit_request, 0);
db08b687 1735 qemu_wait_io_event(cpu);
9b0605f9 1736 } while (!cpu->unplug || cpu_can_run(cpu));
37257942 1737
9b0605f9
PB
1738 qemu_tcg_destroy_vcpu(cpu);
1739 cpu->created = false;
1740 qemu_cond_signal(&qemu_cpu_cond);
1741 qemu_mutex_unlock_iothread();
1742 rcu_unregister_thread();
37257942
AB
1743 return NULL;
1744}
1745
2ff09a40 1746static void qemu_cpu_kick_thread(CPUState *cpu)
cc015e9a
PB
1747{
1748#ifndef _WIN32
1749 int err;
1750
e0c38211
PB
1751 if (cpu->thread_kicked) {
1752 return;
9102deda 1753 }
e0c38211 1754 cpu->thread_kicked = true;
814e612e 1755 err = pthread_kill(cpu->thread->thread, SIG_IPI);
cc015e9a
PB
1756 if (err) {
1757 fprintf(stderr, "qemu:%s: %s", __func__, strerror(err));
1758 exit(1);
1759 }
1760#else /* _WIN32 */
b0cb0a66 1761 if (!qemu_cpu_is_self(cpu)) {
19306806
JTV
1762 if (whpx_enabled()) {
1763 whpx_vcpu_kick(cpu);
1764 } else if (!QueueUserAPC(dummy_apc_func, cpu->hThread, 0)) {
b0cb0a66
VP
1765 fprintf(stderr, "%s: QueueUserAPC failed with error %lu\n",
1766 __func__, GetLastError());
1767 exit(1);
1768 }
1769 }
e0c38211
PB
1770#endif
1771}
ed9164a3 1772
c08d7424 1773void qemu_cpu_kick(CPUState *cpu)
296af7c9 1774{
f5c121b8 1775 qemu_cond_broadcast(cpu->halt_cond);
e0c38211 1776 if (tcg_enabled()) {
791158d9 1777 cpu_exit(cpu);
37257942 1778 /* NOP unless doing single-thread RR */
791158d9 1779 qemu_cpu_kick_rr_cpu();
e0c38211 1780 } else {
b0cb0a66
VP
1781 if (hax_enabled()) {
1782 /*
1783 * FIXME: race condition with the exit_request check in
1784 * hax_vcpu_hax_exec
1785 */
1786 cpu->exit_request = 1;
1787 }
e0c38211
PB
1788 qemu_cpu_kick_thread(cpu);
1789 }
296af7c9
BS
1790}
1791
46d62fac 1792void qemu_cpu_kick_self(void)
296af7c9 1793{
4917cf44 1794 assert(current_cpu);
9102deda 1795 qemu_cpu_kick_thread(current_cpu);
296af7c9
BS
1796}
1797
60e82579 1798bool qemu_cpu_is_self(CPUState *cpu)
296af7c9 1799{
814e612e 1800 return qemu_thread_is_self(cpu->thread);
296af7c9
BS
1801}
1802
79e2b9ae 1803bool qemu_in_vcpu_thread(void)
aa723c23 1804{
4917cf44 1805 return current_cpu && qemu_cpu_is_self(current_cpu);
aa723c23
JQ
1806}
1807
afbe7053
PB
1808static __thread bool iothread_locked = false;
1809
1810bool qemu_mutex_iothread_locked(void)
1811{
1812 return iothread_locked;
1813}
1814
cb764d06
EC
1815/*
1816 * The BQL is taken from so many places that it is worth profiling the
1817 * callers directly, instead of funneling them all through a single function.
1818 */
1819void qemu_mutex_lock_iothread_impl(const char *file, int line)
296af7c9 1820{
cb764d06
EC
1821 QemuMutexLockFunc bql_lock = atomic_read(&qemu_bql_mutex_lock_func);
1822
8d04fb55 1823 g_assert(!qemu_mutex_iothread_locked());
cb764d06 1824 bql_lock(&qemu_global_mutex, file, line);
afbe7053 1825 iothread_locked = true;
296af7c9
BS
1826}
1827
1828void qemu_mutex_unlock_iothread(void)
1829{
8d04fb55 1830 g_assert(qemu_mutex_iothread_locked());
afbe7053 1831 iothread_locked = false;
296af7c9
BS
1832 qemu_mutex_unlock(&qemu_global_mutex);
1833}
1834
e8faee06 1835static bool all_vcpus_paused(void)
296af7c9 1836{
bdc44640 1837 CPUState *cpu;
296af7c9 1838
bdc44640 1839 CPU_FOREACH(cpu) {
182735ef 1840 if (!cpu->stopped) {
e8faee06 1841 return false;
0ab07c62 1842 }
296af7c9
BS
1843 }
1844
e8faee06 1845 return true;
296af7c9
BS
1846}
1847
1848void pause_all_vcpus(void)
1849{
bdc44640 1850 CPUState *cpu;
296af7c9 1851
40daca54 1852 qemu_clock_enable(QEMU_CLOCK_VIRTUAL, false);
bdc44640 1853 CPU_FOREACH(cpu) {
ebd05fea
DH
1854 if (qemu_cpu_is_self(cpu)) {
1855 qemu_cpu_stop(cpu, true);
1856 } else {
1857 cpu->stop = true;
1858 qemu_cpu_kick(cpu);
1859 }
d798e974
JK
1860 }
1861
d759c951
AB
1862 /* We need to drop the replay_lock so any vCPU threads woken up
1863 * can finish their replay tasks
1864 */
1865 replay_mutex_unlock();
1866
296af7c9 1867 while (!all_vcpus_paused()) {
be7d6c57 1868 qemu_cond_wait(&qemu_pause_cond, &qemu_global_mutex);
bdc44640 1869 CPU_FOREACH(cpu) {
182735ef 1870 qemu_cpu_kick(cpu);
296af7c9
BS
1871 }
1872 }
d759c951
AB
1873
1874 qemu_mutex_unlock_iothread();
1875 replay_mutex_lock();
1876 qemu_mutex_lock_iothread();
296af7c9
BS
1877}
1878
2993683b
IM
1879void cpu_resume(CPUState *cpu)
1880{
1881 cpu->stop = false;
1882 cpu->stopped = false;
1883 qemu_cpu_kick(cpu);
1884}
1885
296af7c9
BS
1886void resume_all_vcpus(void)
1887{
bdc44640 1888 CPUState *cpu;
296af7c9 1889
40daca54 1890 qemu_clock_enable(QEMU_CLOCK_VIRTUAL, true);
bdc44640 1891 CPU_FOREACH(cpu) {
182735ef 1892 cpu_resume(cpu);
296af7c9
BS
1893 }
1894}
1895
dbadee4f 1896void cpu_remove_sync(CPUState *cpu)
4c055ab5
GZ
1897{
1898 cpu->stop = true;
1899 cpu->unplug = true;
1900 qemu_cpu_kick(cpu);
dbadee4f
PB
1901 qemu_mutex_unlock_iothread();
1902 qemu_thread_join(cpu->thread);
1903 qemu_mutex_lock_iothread();
2c579042
BR
1904}
1905
4900116e
DDAG
1906/* For temporary buffers for forming a name */
1907#define VCPU_THREAD_NAME_SIZE 16
1908
e5ab30a2 1909static void qemu_tcg_init_vcpu(CPUState *cpu)
296af7c9 1910{
4900116e 1911 char thread_name[VCPU_THREAD_NAME_SIZE];
37257942
AB
1912 static QemuCond *single_tcg_halt_cond;
1913 static QemuThread *single_tcg_cpu_thread;
e8feb96f
EC
1914 static int tcg_region_inited;
1915
f28d0dfd 1916 assert(tcg_enabled());
e8feb96f
EC
1917 /*
1918 * Initialize TCG regions--once. Now is a good time, because:
1919 * (1) TCG's init context, prologue and target globals have been set up.
1920 * (2) qemu_tcg_mttcg_enabled() works now (TCG init code runs before the
1921 * -accel flag is processed, so the check doesn't work then).
1922 */
1923 if (!tcg_region_inited) {
1924 tcg_region_inited = 1;
1925 tcg_region_init();
1926 }
4900116e 1927
37257942 1928 if (qemu_tcg_mttcg_enabled() || !single_tcg_cpu_thread) {
814e612e 1929 cpu->thread = g_malloc0(sizeof(QemuThread));
f5c121b8
AF
1930 cpu->halt_cond = g_malloc0(sizeof(QemuCond));
1931 qemu_cond_init(cpu->halt_cond);
37257942
AB
1932
1933 if (qemu_tcg_mttcg_enabled()) {
1934 /* create a thread per vCPU with TCG (MTTCG) */
1935 parallel_cpus = true;
1936 snprintf(thread_name, VCPU_THREAD_NAME_SIZE, "CPU %d/TCG",
4900116e 1937 cpu->cpu_index);
37257942
AB
1938
1939 qemu_thread_create(cpu->thread, thread_name, qemu_tcg_cpu_thread_fn,
1940 cpu, QEMU_THREAD_JOINABLE);
1941
1942 } else {
1943 /* share a single thread for all cpus with TCG */
1944 snprintf(thread_name, VCPU_THREAD_NAME_SIZE, "ALL CPUs/TCG");
1945 qemu_thread_create(cpu->thread, thread_name,
1946 qemu_tcg_rr_cpu_thread_fn,
1947 cpu, QEMU_THREAD_JOINABLE);
1948
1949 single_tcg_halt_cond = cpu->halt_cond;
1950 single_tcg_cpu_thread = cpu->thread;
1951 }
1ecf47bf 1952#ifdef _WIN32
814e612e 1953 cpu->hThread = qemu_thread_get_handle(cpu->thread);
1ecf47bf 1954#endif
296af7c9 1955 } else {
37257942
AB
1956 /* For non-MTTCG cases we share the thread */
1957 cpu->thread = single_tcg_cpu_thread;
1958 cpu->halt_cond = single_tcg_halt_cond;
a342173a
DH
1959 cpu->thread_id = first_cpu->thread_id;
1960 cpu->can_do_io = 1;
1961 cpu->created = true;
296af7c9
BS
1962 }
1963}
1964
b0cb0a66
VP
1965static void qemu_hax_start_vcpu(CPUState *cpu)
1966{
1967 char thread_name[VCPU_THREAD_NAME_SIZE];
1968
1969 cpu->thread = g_malloc0(sizeof(QemuThread));
1970 cpu->halt_cond = g_malloc0(sizeof(QemuCond));
1971 qemu_cond_init(cpu->halt_cond);
1972
1973 snprintf(thread_name, VCPU_THREAD_NAME_SIZE, "CPU %d/HAX",
1974 cpu->cpu_index);
1975 qemu_thread_create(cpu->thread, thread_name, qemu_hax_cpu_thread_fn,
1976 cpu, QEMU_THREAD_JOINABLE);
1977#ifdef _WIN32
1978 cpu->hThread = qemu_thread_get_handle(cpu->thread);
1979#endif
b0cb0a66
VP
1980}
1981
48a106bd 1982static void qemu_kvm_start_vcpu(CPUState *cpu)
296af7c9 1983{
4900116e
DDAG
1984 char thread_name[VCPU_THREAD_NAME_SIZE];
1985
814e612e 1986 cpu->thread = g_malloc0(sizeof(QemuThread));
f5c121b8
AF
1987 cpu->halt_cond = g_malloc0(sizeof(QemuCond));
1988 qemu_cond_init(cpu->halt_cond);
4900116e
DDAG
1989 snprintf(thread_name, VCPU_THREAD_NAME_SIZE, "CPU %d/KVM",
1990 cpu->cpu_index);
1991 qemu_thread_create(cpu->thread, thread_name, qemu_kvm_cpu_thread_fn,
1992 cpu, QEMU_THREAD_JOINABLE);
296af7c9
BS
1993}
1994
c97d6d2c
SAGDR
1995static void qemu_hvf_start_vcpu(CPUState *cpu)
1996{
1997 char thread_name[VCPU_THREAD_NAME_SIZE];
1998
1999 /* HVF currently does not support TCG, and only runs in
2000 * unrestricted-guest mode. */
2001 assert(hvf_enabled());
2002
2003 cpu->thread = g_malloc0(sizeof(QemuThread));
2004 cpu->halt_cond = g_malloc0(sizeof(QemuCond));
2005 qemu_cond_init(cpu->halt_cond);
2006
2007 snprintf(thread_name, VCPU_THREAD_NAME_SIZE, "CPU %d/HVF",
2008 cpu->cpu_index);
2009 qemu_thread_create(cpu->thread, thread_name, qemu_hvf_cpu_thread_fn,
2010 cpu, QEMU_THREAD_JOINABLE);
c97d6d2c
SAGDR
2011}
2012
19306806
JTV
2013static void qemu_whpx_start_vcpu(CPUState *cpu)
2014{
2015 char thread_name[VCPU_THREAD_NAME_SIZE];
2016
2017 cpu->thread = g_malloc0(sizeof(QemuThread));
2018 cpu->halt_cond = g_malloc0(sizeof(QemuCond));
2019 qemu_cond_init(cpu->halt_cond);
2020 snprintf(thread_name, VCPU_THREAD_NAME_SIZE, "CPU %d/WHPX",
2021 cpu->cpu_index);
2022 qemu_thread_create(cpu->thread, thread_name, qemu_whpx_cpu_thread_fn,
2023 cpu, QEMU_THREAD_JOINABLE);
2024#ifdef _WIN32
2025 cpu->hThread = qemu_thread_get_handle(cpu->thread);
2026#endif
19306806
JTV
2027}
2028
10a9021d 2029static void qemu_dummy_start_vcpu(CPUState *cpu)
c7f0f3b1 2030{
4900116e
DDAG
2031 char thread_name[VCPU_THREAD_NAME_SIZE];
2032
814e612e 2033 cpu->thread = g_malloc0(sizeof(QemuThread));
f5c121b8
AF
2034 cpu->halt_cond = g_malloc0(sizeof(QemuCond));
2035 qemu_cond_init(cpu->halt_cond);
4900116e
DDAG
2036 snprintf(thread_name, VCPU_THREAD_NAME_SIZE, "CPU %d/DUMMY",
2037 cpu->cpu_index);
2038 qemu_thread_create(cpu->thread, thread_name, qemu_dummy_cpu_thread_fn, cpu,
c7f0f3b1 2039 QEMU_THREAD_JOINABLE);
c7f0f3b1
AL
2040}
2041
c643bed9 2042void qemu_init_vcpu(CPUState *cpu)
296af7c9 2043{
ce3960eb
AF
2044 cpu->nr_cores = smp_cores;
2045 cpu->nr_threads = smp_threads;
f324e766 2046 cpu->stopped = true;
56943e8c
PM
2047
2048 if (!cpu->as) {
2049 /* If the target cpu hasn't set up any address spaces itself,
2050 * give it the default one.
2051 */
12ebc9a7 2052 cpu->num_ases = 1;
80ceb07a 2053 cpu_address_space_init(cpu, 0, "cpu-memory", cpu->memory);
56943e8c
PM
2054 }
2055
0ab07c62 2056 if (kvm_enabled()) {
48a106bd 2057 qemu_kvm_start_vcpu(cpu);
b0cb0a66
VP
2058 } else if (hax_enabled()) {
2059 qemu_hax_start_vcpu(cpu);
c97d6d2c
SAGDR
2060 } else if (hvf_enabled()) {
2061 qemu_hvf_start_vcpu(cpu);
c7f0f3b1 2062 } else if (tcg_enabled()) {
e5ab30a2 2063 qemu_tcg_init_vcpu(cpu);
19306806
JTV
2064 } else if (whpx_enabled()) {
2065 qemu_whpx_start_vcpu(cpu);
c7f0f3b1 2066 } else {
10a9021d 2067 qemu_dummy_start_vcpu(cpu);
0ab07c62 2068 }
81e96311
DH
2069
2070 while (!cpu->created) {
2071 qemu_cond_wait(&qemu_cpu_cond, &qemu_global_mutex);
2072 }
296af7c9
BS
2073}
2074
b4a3d965 2075void cpu_stop_current(void)
296af7c9 2076{
4917cf44 2077 if (current_cpu) {
ebd05fea 2078 qemu_cpu_stop(current_cpu, true);
b4a3d965 2079 }
296af7c9
BS
2080}
2081
56983463 2082int vm_stop(RunState state)
296af7c9 2083{
aa723c23 2084 if (qemu_in_vcpu_thread()) {
74892d24 2085 qemu_system_vmstop_request_prepare();
1dfb4dd9 2086 qemu_system_vmstop_request(state);
296af7c9
BS
2087 /*
2088 * FIXME: should not return to device code in case
2089 * vm_stop() has been requested.
2090 */
b4a3d965 2091 cpu_stop_current();
56983463 2092 return 0;
296af7c9 2093 }
56983463 2094
4486e89c 2095 return do_vm_stop(state, true);
296af7c9
BS
2096}
2097
2d76e823
CI
2098/**
2099 * Prepare for (re)starting the VM.
2100 * Returns -1 if the vCPUs are not to be restarted (e.g. if they are already
2101 * running or in case of an error condition), 0 otherwise.
2102 */
2103int vm_prepare_start(void)
2104{
2105 RunState requested;
2d76e823
CI
2106
2107 qemu_vmstop_requested(&requested);
2108 if (runstate_is_running() && requested == RUN_STATE__MAX) {
2109 return -1;
2110 }
2111
2112 /* Ensure that a STOP/RESUME pair of events is emitted if a
2113 * vmstop request was pending. The BLOCK_IO_ERROR event, for
2114 * example, according to documentation is always followed by
2115 * the STOP event.
2116 */
2117 if (runstate_is_running()) {
3ab72385
PX
2118 qapi_event_send_stop();
2119 qapi_event_send_resume();
f056158d 2120 return -1;
2d76e823
CI
2121 }
2122
2123 /* We are sending this now, but the CPUs will be resumed shortly later */
3ab72385 2124 qapi_event_send_resume();
f056158d
MA
2125
2126 replay_enable_events();
2127 cpu_enable_ticks();
2128 runstate_set(RUN_STATE_RUNNING);
2129 vm_state_notify(1, RUN_STATE_RUNNING);
2130 return 0;
2d76e823
CI
2131}
2132
2133void vm_start(void)
2134{
2135 if (!vm_prepare_start()) {
2136 resume_all_vcpus();
2137 }
2138}
2139
8a9236f1
LC
2140/* does a state transition even if the VM is already stopped,
2141 current state is forgotten forever */
56983463 2142int vm_stop_force_state(RunState state)
8a9236f1
LC
2143{
2144 if (runstate_is_running()) {
56983463 2145 return vm_stop(state);
8a9236f1
LC
2146 } else {
2147 runstate_set(state);
b2780d32
WC
2148
2149 bdrv_drain_all();
594a45ce
KW
2150 /* Make sure to return an error if the flush in a previous vm_stop()
2151 * failed. */
22af08ea 2152 return bdrv_flush_all();
8a9236f1
LC
2153 }
2154}
2155
9a78eead 2156void list_cpus(FILE *f, fprintf_function cpu_fprintf, const char *optarg)
262353cb
BS
2157{
2158 /* XXX: implement xxx_cpu_list for targets that still miss it */
e916cbf8
PM
2159#if defined(cpu_list)
2160 cpu_list(f, cpu_fprintf);
262353cb
BS
2161#endif
2162}
de0b36b6
LC
2163
2164CpuInfoList *qmp_query_cpus(Error **errp)
2165{
afed5a5a
IM
2166 MachineState *ms = MACHINE(qdev_get_machine());
2167 MachineClass *mc = MACHINE_GET_CLASS(ms);
de0b36b6 2168 CpuInfoList *head = NULL, *cur_item = NULL;
182735ef 2169 CPUState *cpu;
de0b36b6 2170
bdc44640 2171 CPU_FOREACH(cpu) {
de0b36b6 2172 CpuInfoList *info;
182735ef
AF
2173#if defined(TARGET_I386)
2174 X86CPU *x86_cpu = X86_CPU(cpu);
2175 CPUX86State *env = &x86_cpu->env;
2176#elif defined(TARGET_PPC)
2177 PowerPCCPU *ppc_cpu = POWERPC_CPU(cpu);
2178 CPUPPCState *env = &ppc_cpu->env;
2179#elif defined(TARGET_SPARC)
2180 SPARCCPU *sparc_cpu = SPARC_CPU(cpu);
2181 CPUSPARCState *env = &sparc_cpu->env;
25fa194b
MC
2182#elif defined(TARGET_RISCV)
2183 RISCVCPU *riscv_cpu = RISCV_CPU(cpu);
2184 CPURISCVState *env = &riscv_cpu->env;
182735ef
AF
2185#elif defined(TARGET_MIPS)
2186 MIPSCPU *mips_cpu = MIPS_CPU(cpu);
2187 CPUMIPSState *env = &mips_cpu->env;
48e06fe0
BK
2188#elif defined(TARGET_TRICORE)
2189 TriCoreCPU *tricore_cpu = TRICORE_CPU(cpu);
2190 CPUTriCoreState *env = &tricore_cpu->env;
9d0306df
VM
2191#elif defined(TARGET_S390X)
2192 S390CPU *s390_cpu = S390_CPU(cpu);
2193 CPUS390XState *env = &s390_cpu->env;
182735ef 2194#endif
de0b36b6 2195
cb446eca 2196 cpu_synchronize_state(cpu);
de0b36b6
LC
2197
2198 info = g_malloc0(sizeof(*info));
2199 info->value = g_malloc0(sizeof(*info->value));
55e5c285 2200 info->value->CPU = cpu->cpu_index;
182735ef 2201 info->value->current = (cpu == first_cpu);
259186a7 2202 info->value->halted = cpu->halted;
58f88d4b 2203 info->value->qom_path = object_get_canonical_path(OBJECT(cpu));
9f09e18a 2204 info->value->thread_id = cpu->thread_id;
de0b36b6 2205#if defined(TARGET_I386)
86f4b687 2206 info->value->arch = CPU_INFO_ARCH_X86;
544a3731 2207 info->value->u.x86.pc = env->eip + env->segs[R_CS].base;
de0b36b6 2208#elif defined(TARGET_PPC)
86f4b687 2209 info->value->arch = CPU_INFO_ARCH_PPC;
544a3731 2210 info->value->u.ppc.nip = env->nip;
de0b36b6 2211#elif defined(TARGET_SPARC)
86f4b687 2212 info->value->arch = CPU_INFO_ARCH_SPARC;
544a3731
EB
2213 info->value->u.q_sparc.pc = env->pc;
2214 info->value->u.q_sparc.npc = env->npc;
de0b36b6 2215#elif defined(TARGET_MIPS)
86f4b687 2216 info->value->arch = CPU_INFO_ARCH_MIPS;
544a3731 2217 info->value->u.q_mips.PC = env->active_tc.PC;
48e06fe0 2218#elif defined(TARGET_TRICORE)
86f4b687 2219 info->value->arch = CPU_INFO_ARCH_TRICORE;
544a3731 2220 info->value->u.tricore.PC = env->PC;
9d0306df
VM
2221#elif defined(TARGET_S390X)
2222 info->value->arch = CPU_INFO_ARCH_S390;
2223 info->value->u.s390.cpu_state = env->cpu_state;
25fa194b
MC
2224#elif defined(TARGET_RISCV)
2225 info->value->arch = CPU_INFO_ARCH_RISCV;
2226 info->value->u.riscv.pc = env->pc;
86f4b687
EB
2227#else
2228 info->value->arch = CPU_INFO_ARCH_OTHER;
de0b36b6 2229#endif
afed5a5a
IM
2230 info->value->has_props = !!mc->cpu_index_to_instance_props;
2231 if (info->value->has_props) {
2232 CpuInstanceProperties *props;
2233 props = g_malloc0(sizeof(*props));
2234 *props = mc->cpu_index_to_instance_props(ms, cpu->cpu_index);
2235 info->value->props = props;
2236 }
de0b36b6
LC
2237
2238 /* XXX: waiting for the qapi to support GSList */
2239 if (!cur_item) {
2240 head = cur_item = info;
2241 } else {
2242 cur_item->next = info;
2243 cur_item = info;
2244 }
2245 }
2246
2247 return head;
2248}
0cfd6a9a 2249
daa9d2bc
LE
2250static CpuInfoArch sysemu_target_to_cpuinfo_arch(SysEmuTarget target)
2251{
2252 /*
2253 * The @SysEmuTarget -> @CpuInfoArch mapping below is based on the
2254 * TARGET_ARCH -> TARGET_BASE_ARCH mapping in the "configure" script.
2255 */
2256 switch (target) {
2257 case SYS_EMU_TARGET_I386:
2258 case SYS_EMU_TARGET_X86_64:
2259 return CPU_INFO_ARCH_X86;
2260
2261 case SYS_EMU_TARGET_PPC:
daa9d2bc
LE
2262 case SYS_EMU_TARGET_PPC64:
2263 return CPU_INFO_ARCH_PPC;
2264
2265 case SYS_EMU_TARGET_SPARC:
2266 case SYS_EMU_TARGET_SPARC64:
2267 return CPU_INFO_ARCH_SPARC;
2268
2269 case SYS_EMU_TARGET_MIPS:
2270 case SYS_EMU_TARGET_MIPSEL:
2271 case SYS_EMU_TARGET_MIPS64:
2272 case SYS_EMU_TARGET_MIPS64EL:
2273 return CPU_INFO_ARCH_MIPS;
2274
2275 case SYS_EMU_TARGET_TRICORE:
2276 return CPU_INFO_ARCH_TRICORE;
2277
2278 case SYS_EMU_TARGET_S390X:
2279 return CPU_INFO_ARCH_S390;
2280
2281 case SYS_EMU_TARGET_RISCV32:
2282 case SYS_EMU_TARGET_RISCV64:
2283 return CPU_INFO_ARCH_RISCV;
2284
2285 default:
2286 return CPU_INFO_ARCH_OTHER;
2287 }
2288}
2289
2290static void cpustate_to_cpuinfo_s390(CpuInfoS390 *info, const CPUState *cpu)
2291{
2292#ifdef TARGET_S390X
2293 S390CPU *s390_cpu = S390_CPU(cpu);
2294 CPUS390XState *env = &s390_cpu->env;
2295
2296 info->cpu_state = env->cpu_state;
2297#else
2298 abort();
2299#endif
2300}
2301
ce74ee3d
LC
2302/*
2303 * fast means: we NEVER interrupt vCPU threads to retrieve
2304 * information from KVM.
2305 */
2306CpuInfoFastList *qmp_query_cpus_fast(Error **errp)
2307{
2308 MachineState *ms = MACHINE(qdev_get_machine());
2309 MachineClass *mc = MACHINE_GET_CLASS(ms);
2310 CpuInfoFastList *head = NULL, *cur_item = NULL;
daa9d2bc
LE
2311 SysEmuTarget target = qapi_enum_parse(&SysEmuTarget_lookup, TARGET_NAME,
2312 -1, &error_abort);
ce74ee3d
LC
2313 CPUState *cpu;
2314
2315 CPU_FOREACH(cpu) {
2316 CpuInfoFastList *info = g_malloc0(sizeof(*info));
2317 info->value = g_malloc0(sizeof(*info->value));
2318
2319 info->value->cpu_index = cpu->cpu_index;
2320 info->value->qom_path = object_get_canonical_path(OBJECT(cpu));
2321 info->value->thread_id = cpu->thread_id;
2322
2323 info->value->has_props = !!mc->cpu_index_to_instance_props;
2324 if (info->value->has_props) {
2325 CpuInstanceProperties *props;
2326 props = g_malloc0(sizeof(*props));
2327 *props = mc->cpu_index_to_instance_props(ms, cpu->cpu_index);
2328 info->value->props = props;
2329 }
2330
daa9d2bc
LE
2331 info->value->arch = sysemu_target_to_cpuinfo_arch(target);
2332 info->value->target = target;
2333 if (target == SYS_EMU_TARGET_S390X) {
2334 cpustate_to_cpuinfo_s390(&info->value->u.s390x, cpu);
daa9d2bc
LE
2335 }
2336
ce74ee3d
LC
2337 if (!cur_item) {
2338 head = cur_item = info;
2339 } else {
2340 cur_item->next = info;
2341 cur_item = info;
2342 }
2343 }
2344
2345 return head;
2346}
2347
0cfd6a9a
LC
2348void qmp_memsave(int64_t addr, int64_t size, const char *filename,
2349 bool has_cpu, int64_t cpu_index, Error **errp)
2350{
2351 FILE *f;
2352 uint32_t l;
55e5c285 2353 CPUState *cpu;
0cfd6a9a 2354 uint8_t buf[1024];
0dc9daf0 2355 int64_t orig_addr = addr, orig_size = size;
0cfd6a9a
LC
2356
2357 if (!has_cpu) {
2358 cpu_index = 0;
2359 }
2360
151d1322
AF
2361 cpu = qemu_get_cpu(cpu_index);
2362 if (cpu == NULL) {
c6bd8c70
MA
2363 error_setg(errp, QERR_INVALID_PARAMETER_VALUE, "cpu-index",
2364 "a CPU number");
0cfd6a9a
LC
2365 return;
2366 }
2367
2368 f = fopen(filename, "wb");
2369 if (!f) {
618da851 2370 error_setg_file_open(errp, errno, filename);
0cfd6a9a
LC
2371 return;
2372 }
2373
2374 while (size != 0) {
2375 l = sizeof(buf);
2376 if (l > size)
2377 l = size;
2f4d0f59 2378 if (cpu_memory_rw_debug(cpu, addr, buf, l, 0) != 0) {
0dc9daf0
BP
2379 error_setg(errp, "Invalid addr 0x%016" PRIx64 "/size %" PRId64
2380 " specified", orig_addr, orig_size);
2f4d0f59
AK
2381 goto exit;
2382 }
0cfd6a9a 2383 if (fwrite(buf, 1, l, f) != l) {
c6bd8c70 2384 error_setg(errp, QERR_IO_ERROR);
0cfd6a9a
LC
2385 goto exit;
2386 }
2387 addr += l;
2388 size -= l;
2389 }
2390
2391exit:
2392 fclose(f);
2393}
6d3962bf
LC
2394
2395void qmp_pmemsave(int64_t addr, int64_t size, const char *filename,
2396 Error **errp)
2397{
2398 FILE *f;
2399 uint32_t l;
2400 uint8_t buf[1024];
2401
2402 f = fopen(filename, "wb");
2403 if (!f) {
618da851 2404 error_setg_file_open(errp, errno, filename);
6d3962bf
LC
2405 return;
2406 }
2407
2408 while (size != 0) {
2409 l = sizeof(buf);
2410 if (l > size)
2411 l = size;
eb6282f2 2412 cpu_physical_memory_read(addr, buf, l);
6d3962bf 2413 if (fwrite(buf, 1, l, f) != l) {
c6bd8c70 2414 error_setg(errp, QERR_IO_ERROR);
6d3962bf
LC
2415 goto exit;
2416 }
2417 addr += l;
2418 size -= l;
2419 }
2420
2421exit:
2422 fclose(f);
2423}
ab49ab5c
LC
2424
2425void qmp_inject_nmi(Error **errp)
2426{
9cb805fd 2427 nmi_monitor_handle(monitor_get_cpu_index(), errp);
ab49ab5c 2428}
27498bef
ST
2429
2430void dump_drift_info(FILE *f, fprintf_function cpu_fprintf)
2431{
2432 if (!use_icount) {
2433 return;
2434 }
2435
2436 cpu_fprintf(f, "Host - Guest clock %"PRIi64" ms\n",
2437 (cpu_get_clock() - cpu_get_icount())/SCALE_MS);
2438 if (icount_align_option) {
2439 cpu_fprintf(f, "Max guest delay %"PRIi64" ms\n", -max_delay/SCALE_MS);
2440 cpu_fprintf(f, "Max guest advance %"PRIi64" ms\n", max_advance/SCALE_MS);
2441 } else {
2442 cpu_fprintf(f, "Max guest delay NA\n");
2443 cpu_fprintf(f, "Max guest advance NA\n");
2444 }
2445}