]> git.proxmox.com Git - qemu.git/blame - qemu-timer.h
target-arm: Handle UNDEF cases for Neon 2 regs + scalar forms
[qemu.git] / qemu-timer.h
CommitLineData
87ecb68b
PB
1#ifndef QEMU_TIMER_H
2#define QEMU_TIMER_H
3
29e922b6 4#include "qemu-common.h"
c57c846a
BS
5#include <time.h>
6#include <sys/time.h>
7
8#ifdef _WIN32
9#include <windows.h>
10#include <mmsystem.h>
11#endif
29e922b6 12
87ecb68b
PB
13/* timers */
14
0ce1b948
PB
15#define SCALE_MS 1000000
16#define SCALE_US 1000
17#define SCALE_NS 1
18
87ecb68b
PB
19typedef struct QEMUClock QEMUClock;
20typedef void QEMUTimerCB(void *opaque);
21
22/* The real time clock should be used only for stuff which does not
23 change the virtual machine state, as it is run even if the virtual
24 machine is stopped. The real time clock has a frequency of 1000
25 Hz. */
26extern QEMUClock *rt_clock;
27
28/* The virtual clock is only run during the emulation. It is stopped
29 when the virtual machine is stopped. Virtual timers use a high
30 precision clock, usually cpu cycles (use ticks_per_sec). */
31extern QEMUClock *vm_clock;
32
21d5d12b
JK
33/* The host clock should be use for device models that emulate accurate
34 real time sources. It will continue to run when the virtual machine
35 is suspended, and it will reflect system time changes the host may
36 undergo (e.g. due to NTP). The host clock has the same precision as
37 the virtual clock. */
38extern QEMUClock *host_clock;
39
41c872b6 40int64_t qemu_get_clock_ns(QEMUClock *clock);
db1a4972 41void qemu_clock_enable(QEMUClock *clock, int enabled);
87ecb68b 42
4a998740
PB
43QEMUTimer *qemu_new_timer(QEMUClock *clock, int scale,
44 QEMUTimerCB *cb, void *opaque);
87ecb68b
PB
45void qemu_free_timer(QEMUTimer *ts);
46void qemu_del_timer(QEMUTimer *ts);
47void qemu_mod_timer(QEMUTimer *ts, int64_t expire_time);
48int qemu_timer_pending(QEMUTimer *ts);
2430ffe4 49int qemu_timer_expired(QEMUTimer *timer_head, int64_t current_time);
87ecb68b 50
db1a4972
PB
51void qemu_run_all_timers(void);
52int qemu_alarm_pending(void);
53int64_t qemu_next_deadline(void);
54void configure_alarms(char const *opt);
55void configure_icount(const char *option);
56int qemu_calculate_timeout(void);
57void init_clocks(void);
58int init_timer_alarm(void);
59void quit_timers(void);
60
0ce1b948
PB
61static inline QEMUTimer *qemu_new_timer_ns(QEMUClock *clock, QEMUTimerCB *cb,
62 void *opaque)
63{
4a998740 64 return qemu_new_timer(clock, SCALE_NS, cb, opaque);
0ce1b948
PB
65}
66
67static inline QEMUTimer *qemu_new_timer_ms(QEMUClock *clock, QEMUTimerCB *cb,
68 void *opaque)
69{
4a998740 70 return qemu_new_timer(clock, SCALE_MS, cb, opaque);
0ce1b948
PB
71}
72
73static inline int64_t qemu_get_clock_ms(QEMUClock *clock)
74{
75 return qemu_get_clock_ns(clock) / SCALE_MS;
76}
77
274dfed8
AL
78static inline int64_t get_ticks_per_sec(void)
79{
80 return 1000000000LL;
81}
87ecb68b 82
c57c846a
BS
83/* real time host monotonic timer */
84static inline int64_t get_clock_realtime(void)
85{
86 struct timeval tv;
87
88 gettimeofday(&tv, NULL);
89 return tv.tv_sec * 1000000000LL + (tv.tv_usec * 1000);
90}
91
92/* Warning: don't insert tracepoints into these functions, they are
93 also used by simpletrace backend and tracepoints would cause
94 an infinite recursion! */
95#ifdef _WIN32
96extern int64_t clock_freq;
97
98static inline int64_t get_clock(void)
99{
100 LARGE_INTEGER ti;
101 QueryPerformanceCounter(&ti);
102 return muldiv64(ti.QuadPart, get_ticks_per_sec(), clock_freq);
103}
104
105#else
106
107extern int use_rt_clock;
108
109static inline int64_t get_clock(void)
110{
111#if defined(__linux__) || (defined(__FreeBSD__) && __FreeBSD_version >= 500000) \
112 || defined(__DragonFly__) || defined(__FreeBSD_kernel__)
113 if (use_rt_clock) {
114 struct timespec ts;
115 clock_gettime(CLOCK_MONOTONIC, &ts);
116 return ts.tv_sec * 1000000000LL + ts.tv_nsec;
117 } else
118#endif
119 {
120 /* XXX: using gettimeofday leads to problems if the date
121 changes, so it should be avoided. */
122 return get_clock_realtime();
123 }
124}
125#endif
db1a4972 126
87ecb68b
PB
127void qemu_get_timer(QEMUFile *f, QEMUTimer *ts);
128void qemu_put_timer(QEMUFile *f, QEMUTimer *ts);
129
130/* ptimer.c */
131typedef struct ptimer_state ptimer_state;
132typedef void (*ptimer_cb)(void *opaque);
133
134ptimer_state *ptimer_init(QEMUBH *bh);
135void ptimer_set_period(ptimer_state *s, int64_t period);
136void ptimer_set_freq(ptimer_state *s, uint32_t freq);
137void ptimer_set_limit(ptimer_state *s, uint64_t limit, int reload);
138uint64_t ptimer_get_count(ptimer_state *s);
139void ptimer_set_count(ptimer_state *s, uint64_t count);
140void ptimer_run(ptimer_state *s, int oneshot);
141void ptimer_stop(ptimer_state *s);
142void qemu_put_ptimer(QEMUFile *f, ptimer_state *s);
143void qemu_get_ptimer(QEMUFile *f, ptimer_state *s);
144
29e922b6
BS
145/* icount */
146int64_t qemu_icount_round(int64_t count);
147extern int64_t qemu_icount;
148extern int use_icount;
149extern int icount_time_shift;
150extern int64_t qemu_icount_bias;
151int64_t cpu_get_icount(void);
152
153/*******************************************/
154/* host CPU ticks (if available) */
155
156#if defined(_ARCH_PPC)
157
158static inline int64_t cpu_get_real_ticks(void)
159{
160 int64_t retval;
161#ifdef _ARCH_PPC64
162 /* This reads timebase in one 64bit go and includes Cell workaround from:
163 http://ozlabs.org/pipermail/linuxppc-dev/2006-October/027052.html
164 */
165 __asm__ __volatile__ ("mftb %0\n\t"
166 "cmpwi %0,0\n\t"
167 "beq- $-8"
168 : "=r" (retval));
169#else
170 /* http://ozlabs.org/pipermail/linuxppc-dev/1999-October/003889.html */
171 unsigned long junk;
4a9590f3
AG
172 __asm__ __volatile__ ("mfspr %1,269\n\t" /* mftbu */
173 "mfspr %L0,268\n\t" /* mftb */
174 "mfspr %0,269\n\t" /* mftbu */
29e922b6
BS
175 "cmpw %0,%1\n\t"
176 "bne $-16"
177 : "=r" (retval), "=r" (junk));
178#endif
179 return retval;
180}
181
182#elif defined(__i386__)
183
184static inline int64_t cpu_get_real_ticks(void)
185{
186 int64_t val;
187 asm volatile ("rdtsc" : "=A" (val));
188 return val;
189}
190
191#elif defined(__x86_64__)
192
193static inline int64_t cpu_get_real_ticks(void)
194{
195 uint32_t low,high;
196 int64_t val;
197 asm volatile("rdtsc" : "=a" (low), "=d" (high));
198 val = high;
199 val <<= 32;
200 val |= low;
201 return val;
202}
203
204#elif defined(__hppa__)
205
206static inline int64_t cpu_get_real_ticks(void)
207{
208 int val;
209 asm volatile ("mfctl %%cr16, %0" : "=r"(val));
210 return val;
211}
212
213#elif defined(__ia64)
214
215static inline int64_t cpu_get_real_ticks(void)
216{
217 int64_t val;
218 asm volatile ("mov %0 = ar.itc" : "=r"(val) :: "memory");
219 return val;
220}
221
222#elif defined(__s390__)
223
224static inline int64_t cpu_get_real_ticks(void)
225{
226 int64_t val;
227 asm volatile("stck 0(%1)" : "=m" (val) : "a" (&val) : "cc");
228 return val;
229}
230
231#elif defined(__sparc_v8plus__) || defined(__sparc_v8plusa__) || defined(__sparc_v9__)
232
233static inline int64_t cpu_get_real_ticks (void)
234{
235#if defined(_LP64)
236 uint64_t rval;
237 asm volatile("rd %%tick,%0" : "=r"(rval));
238 return rval;
239#else
240 union {
241 uint64_t i64;
242 struct {
243 uint32_t high;
244 uint32_t low;
245 } i32;
246 } rval;
247 asm volatile("rd %%tick,%1; srlx %1,32,%0"
248 : "=r"(rval.i32.high), "=r"(rval.i32.low));
249 return rval.i64;
250#endif
251}
252
253#elif defined(__mips__) && \
254 ((defined(__mips_isa_rev) && __mips_isa_rev >= 2) || defined(__linux__))
255/*
256 * binutils wants to use rdhwr only on mips32r2
257 * but as linux kernel emulate it, it's fine
258 * to use it.
259 *
260 */
261#define MIPS_RDHWR(rd, value) { \
262 __asm__ __volatile__ (".set push\n\t" \
263 ".set mips32r2\n\t" \
264 "rdhwr %0, "rd"\n\t" \
265 ".set pop" \
266 : "=r" (value)); \
267 }
268
269static inline int64_t cpu_get_real_ticks(void)
270{
271 /* On kernels >= 2.6.25 rdhwr <reg>, $2 and $3 are emulated */
272 uint32_t count;
273 static uint32_t cyc_per_count = 0;
274
275 if (!cyc_per_count) {
276 MIPS_RDHWR("$3", cyc_per_count);
277 }
278
279 MIPS_RDHWR("$2", count);
280 return (int64_t)(count * cyc_per_count);
281}
282
14a6063a
RH
283#elif defined(__alpha__)
284
285static inline int64_t cpu_get_real_ticks(void)
286{
287 uint64_t cc;
288 uint32_t cur, ofs;
289
290 asm volatile("rpcc %0" : "=r"(cc));
291 cur = cc;
292 ofs = cc >> 32;
293 return cur - ofs;
294}
295
29e922b6
BS
296#else
297/* The host CPU doesn't have an easily accessible cycle counter.
298 Just return a monotonically increasing value. This will be
299 totally wrong, but hopefully better than nothing. */
300static inline int64_t cpu_get_real_ticks (void)
301{
302 static int64_t ticks = 0;
303 return ticks++;
304}
305#endif
306
307#ifdef NEED_CPU_H
308/* Deterministic execution requires that IO only be performed on the last
309 instruction of a TB so that interrupts take effect immediately. */
310static inline int can_do_io(CPUState *env)
311{
312 if (!use_icount)
313 return 1;
314
315 /* If not executing code then assume we are ok. */
316 if (!env->current_tb)
317 return 1;
318
319 return env->can_do_io != 0;
320}
321#endif
322
2d8ebcf9
RH
323#ifdef CONFIG_PROFILER
324static inline int64_t profile_getclock(void)
325{
326 return cpu_get_real_ticks();
327}
328
329extern int64_t qemu_time, qemu_time_start;
330extern int64_t tlb_flush_time;
331extern int64_t dev_time;
332#endif
333
87ecb68b 334#endif