]> git.proxmox.com Git - qemu.git/blame - cpu-exec.c
Add machine type aliases
[qemu.git] / cpu-exec.c
CommitLineData
7d13299d
FB
1/*
2 * i386 emulator main execution loop
5fafdf24 3 *
66321a11 4 * Copyright (c) 2003-2005 Fabrice Bellard
7d13299d 5 *
3ef693a0
FB
6 * This library is free software; you can redistribute it and/or
7 * modify it under the terms of the GNU Lesser General Public
8 * License as published by the Free Software Foundation; either
9 * version 2 of the License, or (at your option) any later version.
7d13299d 10 *
3ef693a0
FB
11 * This library is distributed in the hope that it will be useful,
12 * but WITHOUT ANY WARRANTY; without even the implied warranty of
13 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
14 * Lesser General Public License for more details.
7d13299d 15 *
3ef693a0 16 * You should have received a copy of the GNU Lesser General Public
8167ee88 17 * License along with this library; if not, see <http://www.gnu.org/licenses/>.
7d13299d 18 */
e4533c7a 19#include "config.h"
93ac68bc 20#include "exec.h"
956034d7 21#include "disas.h"
7cb69cae 22#include "tcg.h"
7ba1e619 23#include "kvm.h"
7d13299d 24
fbf9eeb3
FB
25#if !defined(CONFIG_SOFTMMU)
26#undef EAX
27#undef ECX
28#undef EDX
29#undef EBX
30#undef ESP
31#undef EBP
32#undef ESI
33#undef EDI
34#undef EIP
35#include <signal.h>
84778508 36#ifdef __linux__
fbf9eeb3
FB
37#include <sys/ucontext.h>
38#endif
84778508 39#endif
fbf9eeb3 40
572a9d4a
BS
41#if defined(__sparc__) && !defined(HOST_SOLARIS)
42// Work around ugly bugs in glibc that mangle global register contents
43#undef env
44#define env cpu_single_env
45#endif
46
36bdbe54
FB
47int tb_invalidated_flag;
48
dc99065b 49//#define DEBUG_EXEC
9de5e440 50//#define DEBUG_SIGNAL
7d13299d 51
6a4955a8
AL
52int qemu_cpu_has_work(CPUState *env)
53{
54 return cpu_has_work(env);
55}
56
e4533c7a
FB
57void cpu_loop_exit(void)
58{
bfed01fc
TS
59 /* NOTE: the register at this point must be saved by hand because
60 longjmp restore them */
61 regs_to_env();
e4533c7a
FB
62 longjmp(env->jmp_env, 1);
63}
bfed01fc 64
fbf9eeb3
FB
65/* exit the current TB from a signal handler. The host registers are
66 restored in a state compatible with the CPU emulator
67 */
5fafdf24 68void cpu_resume_from_signal(CPUState *env1, void *puc)
fbf9eeb3
FB
69{
70#if !defined(CONFIG_SOFTMMU)
84778508 71#ifdef __linux__
fbf9eeb3 72 struct ucontext *uc = puc;
84778508
BS
73#elif defined(__OpenBSD__)
74 struct sigcontext *uc = puc;
75#endif
fbf9eeb3
FB
76#endif
77
78 env = env1;
79
80 /* XXX: restore cpu registers saved in host registers */
81
82#if !defined(CONFIG_SOFTMMU)
83 if (puc) {
84 /* XXX: use siglongjmp ? */
84778508 85#ifdef __linux__
fbf9eeb3 86 sigprocmask(SIG_SETMASK, &uc->uc_sigmask, NULL);
84778508
BS
87#elif defined(__OpenBSD__)
88 sigprocmask(SIG_SETMASK, &uc->sc_mask, NULL);
89#endif
fbf9eeb3
FB
90 }
91#endif
9a3ea654 92 env->exception_index = -1;
fbf9eeb3
FB
93 longjmp(env->jmp_env, 1);
94}
95
2e70f6ef
PB
96/* Execute the code without caching the generated code. An interpreter
97 could be used if available. */
98static void cpu_exec_nocache(int max_cycles, TranslationBlock *orig_tb)
99{
100 unsigned long next_tb;
101 TranslationBlock *tb;
102
103 /* Should never happen.
104 We only end up here when an existing TB is too long. */
105 if (max_cycles > CF_COUNT_MASK)
106 max_cycles = CF_COUNT_MASK;
107
108 tb = tb_gen_code(env, orig_tb->pc, orig_tb->cs_base, orig_tb->flags,
109 max_cycles);
110 env->current_tb = tb;
111 /* execute the generated code */
112 next_tb = tcg_qemu_tb_exec(tb->tc_ptr);
113
114 if ((next_tb & 3) == 2) {
115 /* Restore PC. This may happen if async event occurs before
116 the TB starts executing. */
622ed360 117 cpu_pc_from_tb(env, tb);
2e70f6ef
PB
118 }
119 tb_phys_invalidate(tb, -1);
120 tb_free(tb);
121}
122
8a40a180
FB
123static TranslationBlock *tb_find_slow(target_ulong pc,
124 target_ulong cs_base,
c068688b 125 uint64_t flags)
8a40a180
FB
126{
127 TranslationBlock *tb, **ptb1;
8a40a180
FB
128 unsigned int h;
129 target_ulong phys_pc, phys_page1, phys_page2, virt_page2;
3b46e624 130
8a40a180 131 tb_invalidated_flag = 0;
3b46e624 132
8a40a180 133 regs_to_env(); /* XXX: do it just before cpu_gen_code() */
3b46e624 134
8a40a180
FB
135 /* find translated block using physical mappings */
136 phys_pc = get_phys_addr_code(env, pc);
137 phys_page1 = phys_pc & TARGET_PAGE_MASK;
138 phys_page2 = -1;
139 h = tb_phys_hash_func(phys_pc);
140 ptb1 = &tb_phys_hash[h];
141 for(;;) {
142 tb = *ptb1;
143 if (!tb)
144 goto not_found;
5fafdf24 145 if (tb->pc == pc &&
8a40a180 146 tb->page_addr[0] == phys_page1 &&
5fafdf24 147 tb->cs_base == cs_base &&
8a40a180
FB
148 tb->flags == flags) {
149 /* check next page if needed */
150 if (tb->page_addr[1] != -1) {
5fafdf24 151 virt_page2 = (pc & TARGET_PAGE_MASK) +
8a40a180
FB
152 TARGET_PAGE_SIZE;
153 phys_page2 = get_phys_addr_code(env, virt_page2);
154 if (tb->page_addr[1] == phys_page2)
155 goto found;
156 } else {
157 goto found;
158 }
159 }
160 ptb1 = &tb->phys_hash_next;
161 }
162 not_found:
2e70f6ef
PB
163 /* if no translated code available, then translate it now */
164 tb = tb_gen_code(env, pc, cs_base, flags, 0);
3b46e624 165
8a40a180 166 found:
8a40a180
FB
167 /* we add the TB in the virtual pc hash table */
168 env->tb_jmp_cache[tb_jmp_cache_hash_func(pc)] = tb;
8a40a180
FB
169 return tb;
170}
171
172static inline TranslationBlock *tb_find_fast(void)
173{
174 TranslationBlock *tb;
175 target_ulong cs_base, pc;
6b917547 176 int flags;
8a40a180
FB
177
178 /* we record a subset of the CPU state. It will
179 always be the same before a given translated block
180 is executed. */
6b917547 181 cpu_get_tb_cpu_state(env, &pc, &cs_base, &flags);
bce61846 182 tb = env->tb_jmp_cache[tb_jmp_cache_hash_func(pc)];
551bd27f
TS
183 if (unlikely(!tb || tb->pc != pc || tb->cs_base != cs_base ||
184 tb->flags != flags)) {
8a40a180
FB
185 tb = tb_find_slow(pc, cs_base, flags);
186 }
187 return tb;
188}
189
dde2367e
AL
190static CPUDebugExcpHandler *debug_excp_handler;
191
192CPUDebugExcpHandler *cpu_set_debug_excp_handler(CPUDebugExcpHandler *handler)
193{
194 CPUDebugExcpHandler *old_handler = debug_excp_handler;
195
196 debug_excp_handler = handler;
197 return old_handler;
198}
199
6e140f28
AL
200static void cpu_handle_debug_exception(CPUState *env)
201{
202 CPUWatchpoint *wp;
203
204 if (!env->watchpoint_hit)
c0ce998e 205 TAILQ_FOREACH(wp, &env->watchpoints, entry)
6e140f28 206 wp->flags &= ~BP_WATCHPOINT_HIT;
dde2367e
AL
207
208 if (debug_excp_handler)
209 debug_excp_handler(env);
6e140f28
AL
210}
211
7d13299d
FB
212/* main execution loop */
213
e4533c7a 214int cpu_exec(CPUState *env1)
7d13299d 215{
1057eaa7
PB
216#define DECLARE_HOST_REGS 1
217#include "hostregs_helper.h"
8a40a180 218 int ret, interrupt_request;
8a40a180 219 TranslationBlock *tb;
c27004ec 220 uint8_t *tc_ptr;
d5975363 221 unsigned long next_tb;
8c6939c0 222
bfed01fc
TS
223 if (cpu_halted(env1) == EXCP_HALTED)
224 return EXCP_HALTED;
5a1e3cfc 225
5fafdf24 226 cpu_single_env = env1;
6a00d601 227
7d13299d 228 /* first we save global registers */
1057eaa7
PB
229#define SAVE_HOST_REGS 1
230#include "hostregs_helper.h"
c27004ec 231 env = env1;
e4533c7a 232
0d1a29f9 233 env_to_regs();
ecb644f4 234#if defined(TARGET_I386)
9de5e440 235 /* put eflags in CPU temporary format */
fc2b4c48
FB
236 CC_SRC = env->eflags & (CC_O | CC_S | CC_Z | CC_A | CC_P | CC_C);
237 DF = 1 - (2 * ((env->eflags >> 10) & 1));
9de5e440 238 CC_OP = CC_OP_EFLAGS;
fc2b4c48 239 env->eflags &= ~(DF_MASK | CC_O | CC_S | CC_Z | CC_A | CC_P | CC_C);
93ac68bc 240#elif defined(TARGET_SPARC)
e6e5906b
PB
241#elif defined(TARGET_M68K)
242 env->cc_op = CC_OP_FLAGS;
243 env->cc_dest = env->sr & 0xf;
244 env->cc_x = (env->sr >> 4) & 1;
ecb644f4
TS
245#elif defined(TARGET_ALPHA)
246#elif defined(TARGET_ARM)
247#elif defined(TARGET_PPC)
b779e29e 248#elif defined(TARGET_MICROBLAZE)
6af0bf9c 249#elif defined(TARGET_MIPS)
fdf9b3e8 250#elif defined(TARGET_SH4)
f1ccf904 251#elif defined(TARGET_CRIS)
fdf9b3e8 252 /* XXXXX */
e4533c7a
FB
253#else
254#error unsupported target CPU
255#endif
3fb2ded1 256 env->exception_index = -1;
9d27abd9 257
7d13299d 258 /* prepare setjmp context for exception handling */
3fb2ded1
FB
259 for(;;) {
260 if (setjmp(env->jmp_env) == 0) {
9ddff3d2
BS
261#if defined(__sparc__) && !defined(HOST_SOLARIS)
262#undef env
263 env = cpu_single_env;
264#define env cpu_single_env
265#endif
ee8b7021 266 env->current_tb = NULL;
3fb2ded1
FB
267 /* if an exception is pending, we execute it here */
268 if (env->exception_index >= 0) {
269 if (env->exception_index >= EXCP_INTERRUPT) {
270 /* exit request from the cpu execution loop */
271 ret = env->exception_index;
6e140f28
AL
272 if (ret == EXCP_DEBUG)
273 cpu_handle_debug_exception(env);
3fb2ded1 274 break;
72d239ed
AJ
275 } else {
276#if defined(CONFIG_USER_ONLY)
3fb2ded1 277 /* if user mode only, we simulate a fake exception
9f083493 278 which will be handled outside the cpu execution
3fb2ded1 279 loop */
83479e77 280#if defined(TARGET_I386)
5fafdf24
TS
281 do_interrupt_user(env->exception_index,
282 env->exception_is_int,
283 env->error_code,
3fb2ded1 284 env->exception_next_eip);
eba01623
FB
285 /* successfully delivered */
286 env->old_exception = -1;
83479e77 287#endif
3fb2ded1
FB
288 ret = env->exception_index;
289 break;
72d239ed 290#else
83479e77 291#if defined(TARGET_I386)
3fb2ded1
FB
292 /* simulate a real cpu exception. On i386, it can
293 trigger new exceptions, but we do not handle
294 double or triple faults yet. */
5fafdf24
TS
295 do_interrupt(env->exception_index,
296 env->exception_is_int,
297 env->error_code,
d05e66d2 298 env->exception_next_eip, 0);
678dde13
TS
299 /* successfully delivered */
300 env->old_exception = -1;
ce09776b
FB
301#elif defined(TARGET_PPC)
302 do_interrupt(env);
b779e29e
EI
303#elif defined(TARGET_MICROBLAZE)
304 do_interrupt(env);
6af0bf9c
FB
305#elif defined(TARGET_MIPS)
306 do_interrupt(env);
e95c8d51 307#elif defined(TARGET_SPARC)
f2bc7e7f 308 do_interrupt(env);
b5ff1b31
FB
309#elif defined(TARGET_ARM)
310 do_interrupt(env);
fdf9b3e8
FB
311#elif defined(TARGET_SH4)
312 do_interrupt(env);
eddf68a6
JM
313#elif defined(TARGET_ALPHA)
314 do_interrupt(env);
f1ccf904
TS
315#elif defined(TARGET_CRIS)
316 do_interrupt(env);
0633879f
PB
317#elif defined(TARGET_M68K)
318 do_interrupt(0);
72d239ed 319#endif
83479e77 320#endif
3fb2ded1
FB
321 }
322 env->exception_index = -1;
5fafdf24 323 }
640f42e4 324#ifdef CONFIG_KQEMU
be214e6c 325 if (kqemu_is_ok(env) && env->interrupt_request == 0 && env->exit_request == 0) {
9df217a3 326 int ret;
a7812ae4 327 env->eflags = env->eflags | helper_cc_compute_all(CC_OP) | (DF & DF_MASK);
9df217a3
FB
328 ret = kqemu_cpu_exec(env);
329 /* put eflags in CPU temporary format */
330 CC_SRC = env->eflags & (CC_O | CC_S | CC_Z | CC_A | CC_P | CC_C);
331 DF = 1 - (2 * ((env->eflags >> 10) & 1));
332 CC_OP = CC_OP_EFLAGS;
333 env->eflags &= ~(DF_MASK | CC_O | CC_S | CC_Z | CC_A | CC_P | CC_C);
334 if (ret == 1) {
335 /* exception */
336 longjmp(env->jmp_env, 1);
337 } else if (ret == 2) {
338 /* softmmu execution needed */
339 } else {
be214e6c 340 if (env->interrupt_request != 0 || env->exit_request != 0) {
9df217a3
FB
341 /* hardware interrupt will be executed just after */
342 } else {
343 /* otherwise, we restart */
344 longjmp(env->jmp_env, 1);
345 }
346 }
3fb2ded1 347 }
9df217a3
FB
348#endif
349
7ba1e619 350 if (kvm_enabled()) {
becfc390
AL
351 kvm_cpu_exec(env);
352 longjmp(env->jmp_env, 1);
7ba1e619
AL
353 }
354
b5fc09ae 355 next_tb = 0; /* force lookup of first TB */
3fb2ded1 356 for(;;) {
68a79315 357 interrupt_request = env->interrupt_request;
e1638bd8 358 if (unlikely(interrupt_request)) {
359 if (unlikely(env->singlestep_enabled & SSTEP_NOIRQ)) {
360 /* Mask out external interrupts for this step. */
361 interrupt_request &= ~(CPU_INTERRUPT_HARD |
362 CPU_INTERRUPT_FIQ |
363 CPU_INTERRUPT_SMI |
364 CPU_INTERRUPT_NMI);
365 }
6658ffb8
PB
366 if (interrupt_request & CPU_INTERRUPT_DEBUG) {
367 env->interrupt_request &= ~CPU_INTERRUPT_DEBUG;
368 env->exception_index = EXCP_DEBUG;
369 cpu_loop_exit();
370 }
a90b7318 371#if defined(TARGET_ARM) || defined(TARGET_SPARC) || defined(TARGET_MIPS) || \
b779e29e
EI
372 defined(TARGET_PPC) || defined(TARGET_ALPHA) || defined(TARGET_CRIS) || \
373 defined(TARGET_MICROBLAZE)
a90b7318
AZ
374 if (interrupt_request & CPU_INTERRUPT_HALT) {
375 env->interrupt_request &= ~CPU_INTERRUPT_HALT;
376 env->halted = 1;
377 env->exception_index = EXCP_HLT;
378 cpu_loop_exit();
379 }
380#endif
68a79315 381#if defined(TARGET_I386)
b09ea7d5
GN
382 if (interrupt_request & CPU_INTERRUPT_INIT) {
383 svm_check_intercept(SVM_EXIT_INIT);
384 do_cpu_init(env);
385 env->exception_index = EXCP_HALTED;
386 cpu_loop_exit();
387 } else if (interrupt_request & CPU_INTERRUPT_SIPI) {
388 do_cpu_sipi(env);
389 } else if (env->hflags2 & HF2_GIF_MASK) {
db620f46
FB
390 if ((interrupt_request & CPU_INTERRUPT_SMI) &&
391 !(env->hflags & HF_SMM_MASK)) {
392 svm_check_intercept(SVM_EXIT_SMI);
393 env->interrupt_request &= ~CPU_INTERRUPT_SMI;
394 do_smm_enter();
395 next_tb = 0;
396 } else if ((interrupt_request & CPU_INTERRUPT_NMI) &&
397 !(env->hflags2 & HF2_NMI_MASK)) {
398 env->interrupt_request &= ~CPU_INTERRUPT_NMI;
399 env->hflags2 |= HF2_NMI_MASK;
400 do_interrupt(EXCP02_NMI, 0, 0, 0, 1);
401 next_tb = 0;
79c4f6b0
HY
402 } else if (interrupt_request & CPU_INTERRUPT_MCE) {
403 env->interrupt_request &= ~CPU_INTERRUPT_MCE;
404 do_interrupt(EXCP12_MCHK, 0, 0, 0, 0);
405 next_tb = 0;
db620f46
FB
406 } else if ((interrupt_request & CPU_INTERRUPT_HARD) &&
407 (((env->hflags2 & HF2_VINTR_MASK) &&
408 (env->hflags2 & HF2_HIF_MASK)) ||
409 (!(env->hflags2 & HF2_VINTR_MASK) &&
410 (env->eflags & IF_MASK &&
411 !(env->hflags & HF_INHIBIT_IRQ_MASK))))) {
412 int intno;
413 svm_check_intercept(SVM_EXIT_INTR);
414 env->interrupt_request &= ~(CPU_INTERRUPT_HARD | CPU_INTERRUPT_VIRQ);
415 intno = cpu_get_pic_interrupt(env);
93fcfe39 416 qemu_log_mask(CPU_LOG_TB_IN_ASM, "Servicing hardware INT=0x%02x\n", intno);
9ddff3d2
BS
417#if defined(__sparc__) && !defined(HOST_SOLARIS)
418#undef env
419 env = cpu_single_env;
420#define env cpu_single_env
421#endif
db620f46
FB
422 do_interrupt(intno, 0, 0, 0, 1);
423 /* ensure that no TB jump will be modified as
424 the program flow was changed */
425 next_tb = 0;
0573fbfc 426#if !defined(CONFIG_USER_ONLY)
db620f46
FB
427 } else if ((interrupt_request & CPU_INTERRUPT_VIRQ) &&
428 (env->eflags & IF_MASK) &&
429 !(env->hflags & HF_INHIBIT_IRQ_MASK)) {
430 int intno;
431 /* FIXME: this should respect TPR */
432 svm_check_intercept(SVM_EXIT_VINTR);
db620f46 433 intno = ldl_phys(env->vm_vmcb + offsetof(struct vmcb, control.int_vector));
93fcfe39 434 qemu_log_mask(CPU_LOG_TB_IN_ASM, "Servicing virtual hardware INT=0x%02x\n", intno);
db620f46 435 do_interrupt(intno, 0, 0, 0, 1);
d40c54d6 436 env->interrupt_request &= ~CPU_INTERRUPT_VIRQ;
db620f46 437 next_tb = 0;
907a5b26 438#endif
db620f46 439 }
68a79315 440 }
ce09776b 441#elif defined(TARGET_PPC)
9fddaa0c
FB
442#if 0
443 if ((interrupt_request & CPU_INTERRUPT_RESET)) {
444 cpu_ppc_reset(env);
445 }
446#endif
47103572 447 if (interrupt_request & CPU_INTERRUPT_HARD) {
e9df014c
JM
448 ppc_hw_interrupt(env);
449 if (env->pending_interrupts == 0)
450 env->interrupt_request &= ~CPU_INTERRUPT_HARD;
b5fc09ae 451 next_tb = 0;
ce09776b 452 }
b779e29e
EI
453#elif defined(TARGET_MICROBLAZE)
454 if ((interrupt_request & CPU_INTERRUPT_HARD)
455 && (env->sregs[SR_MSR] & MSR_IE)
456 && !(env->sregs[SR_MSR] & (MSR_EIP | MSR_BIP))
457 && !(env->iflags & (D_FLAG | IMM_FLAG))) {
458 env->exception_index = EXCP_IRQ;
459 do_interrupt(env);
460 next_tb = 0;
461 }
6af0bf9c
FB
462#elif defined(TARGET_MIPS)
463 if ((interrupt_request & CPU_INTERRUPT_HARD) &&
24c7b0e3 464 (env->CP0_Status & env->CP0_Cause & CP0Ca_IP_mask) &&
6af0bf9c 465 (env->CP0_Status & (1 << CP0St_IE)) &&
24c7b0e3
TS
466 !(env->CP0_Status & (1 << CP0St_EXL)) &&
467 !(env->CP0_Status & (1 << CP0St_ERL)) &&
6af0bf9c
FB
468 !(env->hflags & MIPS_HFLAG_DM)) {
469 /* Raise it */
470 env->exception_index = EXCP_EXT_INTERRUPT;
471 env->error_code = 0;
472 do_interrupt(env);
b5fc09ae 473 next_tb = 0;
6af0bf9c 474 }
e95c8d51 475#elif defined(TARGET_SPARC)
66321a11 476 if ((interrupt_request & CPU_INTERRUPT_HARD) &&
5210977a 477 cpu_interrupts_enabled(env)) {
66321a11
FB
478 int pil = env->interrupt_index & 15;
479 int type = env->interrupt_index & 0xf0;
480
481 if (((type == TT_EXTINT) &&
482 (pil == 15 || pil > env->psrpil)) ||
483 type != TT_EXTINT) {
484 env->interrupt_request &= ~CPU_INTERRUPT_HARD;
f2bc7e7f
BS
485 env->exception_index = env->interrupt_index;
486 do_interrupt(env);
66321a11 487 env->interrupt_index = 0;
5210977a 488#if !defined(CONFIG_USER_ONLY)
327ac2e7
BS
489 cpu_check_irqs(env);
490#endif
b5fc09ae 491 next_tb = 0;
66321a11 492 }
e95c8d51
FB
493 } else if (interrupt_request & CPU_INTERRUPT_TIMER) {
494 //do_interrupt(0, 0, 0, 0, 0);
495 env->interrupt_request &= ~CPU_INTERRUPT_TIMER;
a90b7318 496 }
b5ff1b31
FB
497#elif defined(TARGET_ARM)
498 if (interrupt_request & CPU_INTERRUPT_FIQ
499 && !(env->uncached_cpsr & CPSR_F)) {
500 env->exception_index = EXCP_FIQ;
501 do_interrupt(env);
b5fc09ae 502 next_tb = 0;
b5ff1b31 503 }
9ee6e8bb
PB
504 /* ARMv7-M interrupt return works by loading a magic value
505 into the PC. On real hardware the load causes the
506 return to occur. The qemu implementation performs the
507 jump normally, then does the exception return when the
508 CPU tries to execute code at the magic address.
509 This will cause the magic PC value to be pushed to
510 the stack if an interrupt occured at the wrong time.
511 We avoid this by disabling interrupts when
512 pc contains a magic address. */
b5ff1b31 513 if (interrupt_request & CPU_INTERRUPT_HARD
9ee6e8bb
PB
514 && ((IS_M(env) && env->regs[15] < 0xfffffff0)
515 || !(env->uncached_cpsr & CPSR_I))) {
b5ff1b31
FB
516 env->exception_index = EXCP_IRQ;
517 do_interrupt(env);
b5fc09ae 518 next_tb = 0;
b5ff1b31 519 }
fdf9b3e8 520#elif defined(TARGET_SH4)
e96e2044
TS
521 if (interrupt_request & CPU_INTERRUPT_HARD) {
522 do_interrupt(env);
b5fc09ae 523 next_tb = 0;
e96e2044 524 }
eddf68a6
JM
525#elif defined(TARGET_ALPHA)
526 if (interrupt_request & CPU_INTERRUPT_HARD) {
527 do_interrupt(env);
b5fc09ae 528 next_tb = 0;
eddf68a6 529 }
f1ccf904 530#elif defined(TARGET_CRIS)
1b1a38b0
EI
531 if (interrupt_request & CPU_INTERRUPT_HARD
532 && (env->pregs[PR_CCS] & I_FLAG)) {
533 env->exception_index = EXCP_IRQ;
534 do_interrupt(env);
535 next_tb = 0;
536 }
537 if (interrupt_request & CPU_INTERRUPT_NMI
538 && (env->pregs[PR_CCS] & M_FLAG)) {
539 env->exception_index = EXCP_NMI;
f1ccf904 540 do_interrupt(env);
b5fc09ae 541 next_tb = 0;
f1ccf904 542 }
0633879f
PB
543#elif defined(TARGET_M68K)
544 if (interrupt_request & CPU_INTERRUPT_HARD
545 && ((env->sr & SR_I) >> SR_I_SHIFT)
546 < env->pending_level) {
547 /* Real hardware gets the interrupt vector via an
548 IACK cycle at this point. Current emulated
549 hardware doesn't rely on this, so we
550 provide/save the vector when the interrupt is
551 first signalled. */
552 env->exception_index = env->pending_vector;
553 do_interrupt(1);
b5fc09ae 554 next_tb = 0;
0633879f 555 }
68a79315 556#endif
9d05095e
FB
557 /* Don't use the cached interupt_request value,
558 do_interrupt may have updated the EXITTB flag. */
b5ff1b31 559 if (env->interrupt_request & CPU_INTERRUPT_EXITTB) {
bf3e8bf1
FB
560 env->interrupt_request &= ~CPU_INTERRUPT_EXITTB;
561 /* ensure that no TB jump will be modified as
562 the program flow was changed */
b5fc09ae 563 next_tb = 0;
bf3e8bf1 564 }
be214e6c
AJ
565 }
566 if (unlikely(env->exit_request)) {
567 env->exit_request = 0;
568 env->exception_index = EXCP_INTERRUPT;
569 cpu_loop_exit();
3fb2ded1 570 }
7d13299d 571#ifdef DEBUG_EXEC
8fec2b8c 572 if (qemu_loglevel_mask(CPU_LOG_TB_CPU)) {
3fb2ded1 573 /* restore flags in standard format */
ecb644f4
TS
574 regs_to_env();
575#if defined(TARGET_I386)
a7812ae4 576 env->eflags = env->eflags | helper_cc_compute_all(CC_OP) | (DF & DF_MASK);
93fcfe39 577 log_cpu_state(env, X86_DUMP_CCOP);
3fb2ded1 578 env->eflags &= ~(DF_MASK | CC_O | CC_S | CC_Z | CC_A | CC_P | CC_C);
e4533c7a 579#elif defined(TARGET_ARM)
93fcfe39 580 log_cpu_state(env, 0);
93ac68bc 581#elif defined(TARGET_SPARC)
93fcfe39 582 log_cpu_state(env, 0);
67867308 583#elif defined(TARGET_PPC)
93fcfe39 584 log_cpu_state(env, 0);
e6e5906b
PB
585#elif defined(TARGET_M68K)
586 cpu_m68k_flush_flags(env, env->cc_op);
587 env->cc_op = CC_OP_FLAGS;
588 env->sr = (env->sr & 0xffe0)
589 | env->cc_dest | (env->cc_x << 4);
93fcfe39 590 log_cpu_state(env, 0);
b779e29e
EI
591#elif defined(TARGET_MICROBLAZE)
592 log_cpu_state(env, 0);
6af0bf9c 593#elif defined(TARGET_MIPS)
93fcfe39 594 log_cpu_state(env, 0);
fdf9b3e8 595#elif defined(TARGET_SH4)
93fcfe39 596 log_cpu_state(env, 0);
eddf68a6 597#elif defined(TARGET_ALPHA)
93fcfe39 598 log_cpu_state(env, 0);
f1ccf904 599#elif defined(TARGET_CRIS)
93fcfe39 600 log_cpu_state(env, 0);
e4533c7a 601#else
5fafdf24 602#error unsupported target CPU
e4533c7a 603#endif
3fb2ded1 604 }
7d13299d 605#endif
d5975363 606 spin_lock(&tb_lock);
8a40a180 607 tb = tb_find_fast();
d5975363
PB
608 /* Note: we do it here to avoid a gcc bug on Mac OS X when
609 doing it in tb_find_slow */
610 if (tb_invalidated_flag) {
611 /* as some TB could have been invalidated because
612 of memory exceptions while generating the code, we
613 must recompute the hash index here */
614 next_tb = 0;
2e70f6ef 615 tb_invalidated_flag = 0;
d5975363 616 }
9d27abd9 617#ifdef DEBUG_EXEC
93fcfe39
AL
618 qemu_log_mask(CPU_LOG_EXEC, "Trace 0x%08lx [" TARGET_FMT_lx "] %s\n",
619 (long)tb->tc_ptr, tb->pc,
620 lookup_symbol(tb->pc));
9d27abd9 621#endif
8a40a180
FB
622 /* see if we can patch the calling TB. When the TB
623 spans two pages, we cannot safely do a direct
624 jump. */
c27004ec 625 {
b5fc09ae 626 if (next_tb != 0 &&
640f42e4 627#ifdef CONFIG_KQEMU
f32fc648
FB
628 (env->kqemu_enabled != 2) &&
629#endif
ec6338ba 630 tb->page_addr[1] == -1) {
b5fc09ae 631 tb_add_jump((TranslationBlock *)(next_tb & ~3), next_tb & 3, tb);
3fb2ded1 632 }
c27004ec 633 }
d5975363 634 spin_unlock(&tb_lock);
83479e77 635 env->current_tb = tb;
55e8b85e 636
637 /* cpu_interrupt might be called while translating the
638 TB, but before it is linked into a potentially
639 infinite loop and becomes env->current_tb. Avoid
640 starting execution if there is a pending interrupt. */
be214e6c 641 if (unlikely (env->exit_request))
55e8b85e 642 env->current_tb = NULL;
643
2e70f6ef
PB
644 while (env->current_tb) {
645 tc_ptr = tb->tc_ptr;
3fb2ded1 646 /* execute the generated code */
572a9d4a
BS
647#if defined(__sparc__) && !defined(HOST_SOLARIS)
648#undef env
2e70f6ef 649 env = cpu_single_env;
572a9d4a
BS
650#define env cpu_single_env
651#endif
2e70f6ef
PB
652 next_tb = tcg_qemu_tb_exec(tc_ptr);
653 env->current_tb = NULL;
654 if ((next_tb & 3) == 2) {
bf20dc07 655 /* Instruction counter expired. */
2e70f6ef
PB
656 int insns_left;
657 tb = (TranslationBlock *)(long)(next_tb & ~3);
658 /* Restore PC. */
622ed360 659 cpu_pc_from_tb(env, tb);
2e70f6ef
PB
660 insns_left = env->icount_decr.u32;
661 if (env->icount_extra && insns_left >= 0) {
662 /* Refill decrementer and continue execution. */
663 env->icount_extra += insns_left;
664 if (env->icount_extra > 0xffff) {
665 insns_left = 0xffff;
666 } else {
667 insns_left = env->icount_extra;
668 }
669 env->icount_extra -= insns_left;
670 env->icount_decr.u16.low = insns_left;
671 } else {
672 if (insns_left > 0) {
673 /* Execute remaining instructions. */
674 cpu_exec_nocache(insns_left, tb);
675 }
676 env->exception_index = EXCP_INTERRUPT;
677 next_tb = 0;
678 cpu_loop_exit();
679 }
680 }
681 }
4cbf74b6
FB
682 /* reset soft MMU for next block (it can currently
683 only be set by a memory fault) */
640f42e4 684#if defined(CONFIG_KQEMU)
f32fc648
FB
685#define MIN_CYCLE_BEFORE_SWITCH (100 * 1000)
686 if (kqemu_is_ok(env) &&
687 (cpu_get_time_fast() - env->last_io_time) >= MIN_CYCLE_BEFORE_SWITCH) {
688 cpu_loop_exit();
689 }
4cbf74b6 690#endif
50a518e3 691 } /* for(;;) */
3fb2ded1 692 } else {
0d1a29f9 693 env_to_regs();
7d13299d 694 }
3fb2ded1
FB
695 } /* for(;;) */
696
7d13299d 697
e4533c7a 698#if defined(TARGET_I386)
9de5e440 699 /* restore flags in standard format */
a7812ae4 700 env->eflags = env->eflags | helper_cc_compute_all(CC_OP) | (DF & DF_MASK);
e4533c7a 701#elif defined(TARGET_ARM)
b7bcbe95 702 /* XXX: Save/restore host fpu exception state?. */
93ac68bc 703#elif defined(TARGET_SPARC)
67867308 704#elif defined(TARGET_PPC)
e6e5906b
PB
705#elif defined(TARGET_M68K)
706 cpu_m68k_flush_flags(env, env->cc_op);
707 env->cc_op = CC_OP_FLAGS;
708 env->sr = (env->sr & 0xffe0)
709 | env->cc_dest | (env->cc_x << 4);
b779e29e 710#elif defined(TARGET_MICROBLAZE)
6af0bf9c 711#elif defined(TARGET_MIPS)
fdf9b3e8 712#elif defined(TARGET_SH4)
eddf68a6 713#elif defined(TARGET_ALPHA)
f1ccf904 714#elif defined(TARGET_CRIS)
fdf9b3e8 715 /* XXXXX */
e4533c7a
FB
716#else
717#error unsupported target CPU
718#endif
1057eaa7
PB
719
720 /* restore global registers */
1057eaa7
PB
721#include "hostregs_helper.h"
722
6a00d601 723 /* fail safe : never use cpu_single_env outside cpu_exec() */
5fafdf24 724 cpu_single_env = NULL;
7d13299d
FB
725 return ret;
726}
6dbad63e 727
fbf9eeb3
FB
728/* must only be called from the generated code as an exception can be
729 generated */
730void tb_invalidate_page_range(target_ulong start, target_ulong end)
731{
dc5d0b3d
FB
732 /* XXX: cannot enable it yet because it yields to MMU exception
733 where NIP != read address on PowerPC */
734#if 0
fbf9eeb3
FB
735 target_ulong phys_addr;
736 phys_addr = get_phys_addr_code(env, start);
737 tb_invalidate_phys_page_range(phys_addr, phys_addr + end - start, 0);
dc5d0b3d 738#endif
fbf9eeb3
FB
739}
740
1a18c71b 741#if defined(TARGET_I386) && defined(CONFIG_USER_ONLY)
e4533c7a 742
6dbad63e
FB
743void cpu_x86_load_seg(CPUX86State *s, int seg_reg, int selector)
744{
745 CPUX86State *saved_env;
746
747 saved_env = env;
748 env = s;
a412ac57 749 if (!(env->cr[0] & CR0_PE_MASK) || (env->eflags & VM_MASK)) {
a513fe19 750 selector &= 0xffff;
5fafdf24 751 cpu_x86_load_seg_cache(env, seg_reg, selector,
c27004ec 752 (selector << 4), 0xffff, 0);
a513fe19 753 } else {
5d97559d 754 helper_load_seg(seg_reg, selector);
a513fe19 755 }
6dbad63e
FB
756 env = saved_env;
757}
9de5e440 758
6f12a2a6 759void cpu_x86_fsave(CPUX86State *s, target_ulong ptr, int data32)
d0a1ffc9
FB
760{
761 CPUX86State *saved_env;
762
763 saved_env = env;
764 env = s;
3b46e624 765
6f12a2a6 766 helper_fsave(ptr, data32);
d0a1ffc9
FB
767
768 env = saved_env;
769}
770
6f12a2a6 771void cpu_x86_frstor(CPUX86State *s, target_ulong ptr, int data32)
d0a1ffc9
FB
772{
773 CPUX86State *saved_env;
774
775 saved_env = env;
776 env = s;
3b46e624 777
6f12a2a6 778 helper_frstor(ptr, data32);
d0a1ffc9
FB
779
780 env = saved_env;
781}
782
e4533c7a
FB
783#endif /* TARGET_I386 */
784
67b915a5
FB
785#if !defined(CONFIG_SOFTMMU)
786
3fb2ded1
FB
787#if defined(TARGET_I386)
788
b56dad1c 789/* 'pc' is the host PC at which the exception was raised. 'address' is
fd6ce8f6
FB
790 the effective address of the memory exception. 'is_write' is 1 if a
791 write caused the exception and otherwise 0'. 'old_set' is the
792 signal set which should be restored */
2b413144 793static inline int handle_cpu_signal(unsigned long pc, unsigned long address,
5fafdf24 794 int is_write, sigset_t *old_set,
bf3e8bf1 795 void *puc)
9de5e440 796{
a513fe19
FB
797 TranslationBlock *tb;
798 int ret;
68a79315 799
83479e77
FB
800 if (cpu_single_env)
801 env = cpu_single_env; /* XXX: find a correct solution for multithread */
fd6ce8f6 802#if defined(DEBUG_SIGNAL)
5fafdf24 803 qemu_printf("qemu: SIGSEGV pc=0x%08lx address=%08lx w=%d oldset=0x%08lx\n",
bf3e8bf1 804 pc, address, is_write, *(unsigned long *)old_set);
9de5e440 805#endif
25eb4484 806 /* XXX: locking issue */
53a5960a 807 if (is_write && page_unprotect(h2g(address), pc, puc)) {
fd6ce8f6
FB
808 return 1;
809 }
fbf9eeb3 810
3fb2ded1 811 /* see if it is an MMU fault */
6ebbf390 812 ret = cpu_x86_handle_mmu_fault(env, address, is_write, MMU_USER_IDX, 0);
3fb2ded1
FB
813 if (ret < 0)
814 return 0; /* not an MMU fault */
815 if (ret == 0)
816 return 1; /* the MMU fault was handled without causing real CPU fault */
817 /* now we have a real cpu fault */
a513fe19
FB
818 tb = tb_find_pc(pc);
819 if (tb) {
9de5e440
FB
820 /* the PC is inside the translated code. It means that we have
821 a virtual CPU fault */
bf3e8bf1 822 cpu_restore_state(tb, env, pc, puc);
3fb2ded1 823 }
4cbf74b6 824 if (ret == 1) {
3fb2ded1 825#if 0
5fafdf24 826 printf("PF exception: EIP=0x%08x CR2=0x%08x error=0x%x\n",
4cbf74b6 827 env->eip, env->cr[2], env->error_code);
3fb2ded1 828#endif
4cbf74b6
FB
829 /* we restore the process signal mask as the sigreturn should
830 do it (XXX: use sigsetjmp) */
831 sigprocmask(SIG_SETMASK, old_set, NULL);
54ca9095 832 raise_exception_err(env->exception_index, env->error_code);
4cbf74b6
FB
833 } else {
834 /* activate soft MMU for this block */
3f337316 835 env->hflags |= HF_SOFTMMU_MASK;
fbf9eeb3 836 cpu_resume_from_signal(env, puc);
4cbf74b6 837 }
3fb2ded1
FB
838 /* never comes here */
839 return 1;
840}
841
e4533c7a 842#elif defined(TARGET_ARM)
3fb2ded1 843static inline int handle_cpu_signal(unsigned long pc, unsigned long address,
bf3e8bf1
FB
844 int is_write, sigset_t *old_set,
845 void *puc)
3fb2ded1 846{
68016c62
FB
847 TranslationBlock *tb;
848 int ret;
849
850 if (cpu_single_env)
851 env = cpu_single_env; /* XXX: find a correct solution for multithread */
852#if defined(DEBUG_SIGNAL)
5fafdf24 853 printf("qemu: SIGSEGV pc=0x%08lx address=%08lx w=%d oldset=0x%08lx\n",
68016c62
FB
854 pc, address, is_write, *(unsigned long *)old_set);
855#endif
9f0777ed 856 /* XXX: locking issue */
53a5960a 857 if (is_write && page_unprotect(h2g(address), pc, puc)) {
9f0777ed
FB
858 return 1;
859 }
68016c62 860 /* see if it is an MMU fault */
6ebbf390 861 ret = cpu_arm_handle_mmu_fault(env, address, is_write, MMU_USER_IDX, 0);
68016c62
FB
862 if (ret < 0)
863 return 0; /* not an MMU fault */
864 if (ret == 0)
865 return 1; /* the MMU fault was handled without causing real CPU fault */
866 /* now we have a real cpu fault */
867 tb = tb_find_pc(pc);
868 if (tb) {
869 /* the PC is inside the translated code. It means that we have
870 a virtual CPU fault */
871 cpu_restore_state(tb, env, pc, puc);
872 }
873 /* we restore the process signal mask as the sigreturn should
874 do it (XXX: use sigsetjmp) */
875 sigprocmask(SIG_SETMASK, old_set, NULL);
876 cpu_loop_exit();
968c74da
AJ
877 /* never comes here */
878 return 1;
3fb2ded1 879}
93ac68bc
FB
880#elif defined(TARGET_SPARC)
881static inline int handle_cpu_signal(unsigned long pc, unsigned long address,
bf3e8bf1
FB
882 int is_write, sigset_t *old_set,
883 void *puc)
93ac68bc 884{
68016c62
FB
885 TranslationBlock *tb;
886 int ret;
887
888 if (cpu_single_env)
889 env = cpu_single_env; /* XXX: find a correct solution for multithread */
890#if defined(DEBUG_SIGNAL)
5fafdf24 891 printf("qemu: SIGSEGV pc=0x%08lx address=%08lx w=%d oldset=0x%08lx\n",
68016c62
FB
892 pc, address, is_write, *(unsigned long *)old_set);
893#endif
b453b70b 894 /* XXX: locking issue */
53a5960a 895 if (is_write && page_unprotect(h2g(address), pc, puc)) {
b453b70b
FB
896 return 1;
897 }
68016c62 898 /* see if it is an MMU fault */
6ebbf390 899 ret = cpu_sparc_handle_mmu_fault(env, address, is_write, MMU_USER_IDX, 0);
68016c62
FB
900 if (ret < 0)
901 return 0; /* not an MMU fault */
902 if (ret == 0)
903 return 1; /* the MMU fault was handled without causing real CPU fault */
904 /* now we have a real cpu fault */
905 tb = tb_find_pc(pc);
906 if (tb) {
907 /* the PC is inside the translated code. It means that we have
908 a virtual CPU fault */
909 cpu_restore_state(tb, env, pc, puc);
910 }
911 /* we restore the process signal mask as the sigreturn should
912 do it (XXX: use sigsetjmp) */
913 sigprocmask(SIG_SETMASK, old_set, NULL);
914 cpu_loop_exit();
968c74da
AJ
915 /* never comes here */
916 return 1;
93ac68bc 917}
67867308
FB
918#elif defined (TARGET_PPC)
919static inline int handle_cpu_signal(unsigned long pc, unsigned long address,
bf3e8bf1
FB
920 int is_write, sigset_t *old_set,
921 void *puc)
67867308
FB
922{
923 TranslationBlock *tb;
ce09776b 924 int ret;
3b46e624 925
67867308
FB
926 if (cpu_single_env)
927 env = cpu_single_env; /* XXX: find a correct solution for multithread */
67867308 928#if defined(DEBUG_SIGNAL)
5fafdf24 929 printf("qemu: SIGSEGV pc=0x%08lx address=%08lx w=%d oldset=0x%08lx\n",
67867308
FB
930 pc, address, is_write, *(unsigned long *)old_set);
931#endif
932 /* XXX: locking issue */
53a5960a 933 if (is_write && page_unprotect(h2g(address), pc, puc)) {
67867308
FB
934 return 1;
935 }
936
ce09776b 937 /* see if it is an MMU fault */
6ebbf390 938 ret = cpu_ppc_handle_mmu_fault(env, address, is_write, MMU_USER_IDX, 0);
ce09776b
FB
939 if (ret < 0)
940 return 0; /* not an MMU fault */
941 if (ret == 0)
942 return 1; /* the MMU fault was handled without causing real CPU fault */
943
67867308
FB
944 /* now we have a real cpu fault */
945 tb = tb_find_pc(pc);
946 if (tb) {
947 /* the PC is inside the translated code. It means that we have
948 a virtual CPU fault */
bf3e8bf1 949 cpu_restore_state(tb, env, pc, puc);
67867308 950 }
ce09776b 951 if (ret == 1) {
67867308 952#if 0
5fafdf24 953 printf("PF exception: NIP=0x%08x error=0x%x %p\n",
ce09776b 954 env->nip, env->error_code, tb);
67867308
FB
955#endif
956 /* we restore the process signal mask as the sigreturn should
957 do it (XXX: use sigsetjmp) */
bf3e8bf1 958 sigprocmask(SIG_SETMASK, old_set, NULL);
e06fcd75 959 cpu_loop_exit();
ce09776b
FB
960 } else {
961 /* activate soft MMU for this block */
fbf9eeb3 962 cpu_resume_from_signal(env, puc);
ce09776b 963 }
67867308 964 /* never comes here */
e6e5906b
PB
965 return 1;
966}
967
968#elif defined(TARGET_M68K)
969static inline int handle_cpu_signal(unsigned long pc, unsigned long address,
970 int is_write, sigset_t *old_set,
971 void *puc)
972{
973 TranslationBlock *tb;
974 int ret;
975
976 if (cpu_single_env)
977 env = cpu_single_env; /* XXX: find a correct solution for multithread */
978#if defined(DEBUG_SIGNAL)
5fafdf24 979 printf("qemu: SIGSEGV pc=0x%08lx address=%08lx w=%d oldset=0x%08lx\n",
e6e5906b
PB
980 pc, address, is_write, *(unsigned long *)old_set);
981#endif
982 /* XXX: locking issue */
983 if (is_write && page_unprotect(address, pc, puc)) {
984 return 1;
985 }
986 /* see if it is an MMU fault */
6ebbf390 987 ret = cpu_m68k_handle_mmu_fault(env, address, is_write, MMU_USER_IDX, 0);
e6e5906b
PB
988 if (ret < 0)
989 return 0; /* not an MMU fault */
990 if (ret == 0)
991 return 1; /* the MMU fault was handled without causing real CPU fault */
992 /* now we have a real cpu fault */
993 tb = tb_find_pc(pc);
994 if (tb) {
995 /* the PC is inside the translated code. It means that we have
996 a virtual CPU fault */
997 cpu_restore_state(tb, env, pc, puc);
998 }
999 /* we restore the process signal mask as the sigreturn should
1000 do it (XXX: use sigsetjmp) */
1001 sigprocmask(SIG_SETMASK, old_set, NULL);
1002 cpu_loop_exit();
1003 /* never comes here */
67867308
FB
1004 return 1;
1005}
6af0bf9c
FB
1006
1007#elif defined (TARGET_MIPS)
1008static inline int handle_cpu_signal(unsigned long pc, unsigned long address,
1009 int is_write, sigset_t *old_set,
1010 void *puc)
1011{
1012 TranslationBlock *tb;
1013 int ret;
3b46e624 1014
6af0bf9c
FB
1015 if (cpu_single_env)
1016 env = cpu_single_env; /* XXX: find a correct solution for multithread */
1017#if defined(DEBUG_SIGNAL)
5fafdf24 1018 printf("qemu: SIGSEGV pc=0x%08lx address=%08lx w=%d oldset=0x%08lx\n",
6af0bf9c
FB
1019 pc, address, is_write, *(unsigned long *)old_set);
1020#endif
1021 /* XXX: locking issue */
53a5960a 1022 if (is_write && page_unprotect(h2g(address), pc, puc)) {
6af0bf9c
FB
1023 return 1;
1024 }
1025
1026 /* see if it is an MMU fault */
6ebbf390 1027 ret = cpu_mips_handle_mmu_fault(env, address, is_write, MMU_USER_IDX, 0);
6af0bf9c
FB
1028 if (ret < 0)
1029 return 0; /* not an MMU fault */
1030 if (ret == 0)
1031 return 1; /* the MMU fault was handled without causing real CPU fault */
1032
1033 /* now we have a real cpu fault */
1034 tb = tb_find_pc(pc);
1035 if (tb) {
1036 /* the PC is inside the translated code. It means that we have
1037 a virtual CPU fault */
1038 cpu_restore_state(tb, env, pc, puc);
1039 }
1040 if (ret == 1) {
1041#if 0
5fafdf24 1042 printf("PF exception: PC=0x" TARGET_FMT_lx " error=0x%x %p\n",
1eb5207b 1043 env->PC, env->error_code, tb);
b779e29e
EI
1044#endif
1045 /* we restore the process signal mask as the sigreturn should
1046 do it (XXX: use sigsetjmp) */
1047 sigprocmask(SIG_SETMASK, old_set, NULL);
1048 cpu_loop_exit();
1049 } else {
1050 /* activate soft MMU for this block */
1051 cpu_resume_from_signal(env, puc);
1052 }
1053 /* never comes here */
1054 return 1;
1055}
1056
1057#elif defined (TARGET_MICROBLAZE)
1058static inline int handle_cpu_signal(unsigned long pc, unsigned long address,
1059 int is_write, sigset_t *old_set,
1060 void *puc)
1061{
1062 TranslationBlock *tb;
1063 int ret;
1064
1065 if (cpu_single_env)
1066 env = cpu_single_env; /* XXX: find a correct solution for multithread */
1067#if defined(DEBUG_SIGNAL)
1068 printf("qemu: SIGSEGV pc=0x%08lx address=%08lx w=%d oldset=0x%08lx\n",
1069 pc, address, is_write, *(unsigned long *)old_set);
1070#endif
1071 /* XXX: locking issue */
1072 if (is_write && page_unprotect(h2g(address), pc, puc)) {
1073 return 1;
1074 }
1075
1076 /* see if it is an MMU fault */
1077 ret = cpu_mb_handle_mmu_fault(env, address, is_write, MMU_USER_IDX, 0);
1078 if (ret < 0)
1079 return 0; /* not an MMU fault */
1080 if (ret == 0)
1081 return 1; /* the MMU fault was handled without causing real CPU fault */
1082
1083 /* now we have a real cpu fault */
1084 tb = tb_find_pc(pc);
1085 if (tb) {
1086 /* the PC is inside the translated code. It means that we have
1087 a virtual CPU fault */
1088 cpu_restore_state(tb, env, pc, puc);
1089 }
1090 if (ret == 1) {
1091#if 0
1092 printf("PF exception: PC=0x" TARGET_FMT_lx " error=0x%x %p\n",
1093 env->PC, env->error_code, tb);
6af0bf9c
FB
1094#endif
1095 /* we restore the process signal mask as the sigreturn should
1096 do it (XXX: use sigsetjmp) */
1097 sigprocmask(SIG_SETMASK, old_set, NULL);
f9480ffc 1098 cpu_loop_exit();
6af0bf9c
FB
1099 } else {
1100 /* activate soft MMU for this block */
1101 cpu_resume_from_signal(env, puc);
1102 }
1103 /* never comes here */
1104 return 1;
1105}
1106
fdf9b3e8
FB
1107#elif defined (TARGET_SH4)
1108static inline int handle_cpu_signal(unsigned long pc, unsigned long address,
1109 int is_write, sigset_t *old_set,
1110 void *puc)
1111{
1112 TranslationBlock *tb;
1113 int ret;
3b46e624 1114
fdf9b3e8
FB
1115 if (cpu_single_env)
1116 env = cpu_single_env; /* XXX: find a correct solution for multithread */
1117#if defined(DEBUG_SIGNAL)
5fafdf24 1118 printf("qemu: SIGSEGV pc=0x%08lx address=%08lx w=%d oldset=0x%08lx\n",
fdf9b3e8
FB
1119 pc, address, is_write, *(unsigned long *)old_set);
1120#endif
1121 /* XXX: locking issue */
1122 if (is_write && page_unprotect(h2g(address), pc, puc)) {
1123 return 1;
1124 }
1125
1126 /* see if it is an MMU fault */
6ebbf390 1127 ret = cpu_sh4_handle_mmu_fault(env, address, is_write, MMU_USER_IDX, 0);
fdf9b3e8
FB
1128 if (ret < 0)
1129 return 0; /* not an MMU fault */
1130 if (ret == 0)
1131 return 1; /* the MMU fault was handled without causing real CPU fault */
1132
1133 /* now we have a real cpu fault */
eddf68a6
JM
1134 tb = tb_find_pc(pc);
1135 if (tb) {
1136 /* the PC is inside the translated code. It means that we have
1137 a virtual CPU fault */
1138 cpu_restore_state(tb, env, pc, puc);
1139 }
1140#if 0
5fafdf24 1141 printf("PF exception: NIP=0x%08x error=0x%x %p\n",
eddf68a6
JM
1142 env->nip, env->error_code, tb);
1143#endif
1144 /* we restore the process signal mask as the sigreturn should
1145 do it (XXX: use sigsetjmp) */
1146 sigprocmask(SIG_SETMASK, old_set, NULL);
1147 cpu_loop_exit();
1148 /* never comes here */
1149 return 1;
1150}
1151
1152#elif defined (TARGET_ALPHA)
1153static inline int handle_cpu_signal(unsigned long pc, unsigned long address,
1154 int is_write, sigset_t *old_set,
1155 void *puc)
1156{
1157 TranslationBlock *tb;
1158 int ret;
3b46e624 1159
eddf68a6
JM
1160 if (cpu_single_env)
1161 env = cpu_single_env; /* XXX: find a correct solution for multithread */
1162#if defined(DEBUG_SIGNAL)
5fafdf24 1163 printf("qemu: SIGSEGV pc=0x%08lx address=%08lx w=%d oldset=0x%08lx\n",
eddf68a6
JM
1164 pc, address, is_write, *(unsigned long *)old_set);
1165#endif
1166 /* XXX: locking issue */
1167 if (is_write && page_unprotect(h2g(address), pc, puc)) {
1168 return 1;
1169 }
1170
1171 /* see if it is an MMU fault */
6ebbf390 1172 ret = cpu_alpha_handle_mmu_fault(env, address, is_write, MMU_USER_IDX, 0);
eddf68a6
JM
1173 if (ret < 0)
1174 return 0; /* not an MMU fault */
1175 if (ret == 0)
1176 return 1; /* the MMU fault was handled without causing real CPU fault */
1177
1178 /* now we have a real cpu fault */
fdf9b3e8
FB
1179 tb = tb_find_pc(pc);
1180 if (tb) {
1181 /* the PC is inside the translated code. It means that we have
1182 a virtual CPU fault */
1183 cpu_restore_state(tb, env, pc, puc);
1184 }
fdf9b3e8 1185#if 0
5fafdf24 1186 printf("PF exception: NIP=0x%08x error=0x%x %p\n",
fdf9b3e8
FB
1187 env->nip, env->error_code, tb);
1188#endif
1189 /* we restore the process signal mask as the sigreturn should
1190 do it (XXX: use sigsetjmp) */
355fb23d
PB
1191 sigprocmask(SIG_SETMASK, old_set, NULL);
1192 cpu_loop_exit();
fdf9b3e8
FB
1193 /* never comes here */
1194 return 1;
1195}
f1ccf904
TS
1196#elif defined (TARGET_CRIS)
1197static inline int handle_cpu_signal(unsigned long pc, unsigned long address,
1198 int is_write, sigset_t *old_set,
1199 void *puc)
1200{
1201 TranslationBlock *tb;
1202 int ret;
1203
1204 if (cpu_single_env)
1205 env = cpu_single_env; /* XXX: find a correct solution for multithread */
1206#if defined(DEBUG_SIGNAL)
1207 printf("qemu: SIGSEGV pc=0x%08lx address=%08lx w=%d oldset=0x%08lx\n",
1208 pc, address, is_write, *(unsigned long *)old_set);
1209#endif
1210 /* XXX: locking issue */
1211 if (is_write && page_unprotect(h2g(address), pc, puc)) {
1212 return 1;
1213 }
1214
1215 /* see if it is an MMU fault */
6ebbf390 1216 ret = cpu_cris_handle_mmu_fault(env, address, is_write, MMU_USER_IDX, 0);
f1ccf904
TS
1217 if (ret < 0)
1218 return 0; /* not an MMU fault */
1219 if (ret == 0)
1220 return 1; /* the MMU fault was handled without causing real CPU fault */
1221
1222 /* now we have a real cpu fault */
1223 tb = tb_find_pc(pc);
1224 if (tb) {
1225 /* the PC is inside the translated code. It means that we have
1226 a virtual CPU fault */
1227 cpu_restore_state(tb, env, pc, puc);
1228 }
f1ccf904
TS
1229 /* we restore the process signal mask as the sigreturn should
1230 do it (XXX: use sigsetjmp) */
1231 sigprocmask(SIG_SETMASK, old_set, NULL);
1232 cpu_loop_exit();
1233 /* never comes here */
1234 return 1;
1235}
1236
e4533c7a
FB
1237#else
1238#error unsupported target CPU
1239#endif
9de5e440 1240
2b413144
FB
1241#if defined(__i386__)
1242
d8ecc0b9
FB
1243#if defined(__APPLE__)
1244# include <sys/ucontext.h>
1245
1246# define EIP_sig(context) (*((unsigned long*)&(context)->uc_mcontext->ss.eip))
1247# define TRAP_sig(context) ((context)->uc_mcontext->es.trapno)
1248# define ERROR_sig(context) ((context)->uc_mcontext->es.err)
d39bb24a
BS
1249# define MASK_sig(context) ((context)->uc_sigmask)
1250#elif defined(__OpenBSD__)
1251# define EIP_sig(context) ((context)->sc_eip)
1252# define TRAP_sig(context) ((context)->sc_trapno)
1253# define ERROR_sig(context) ((context)->sc_err)
1254# define MASK_sig(context) ((context)->sc_mask)
d8ecc0b9
FB
1255#else
1256# define EIP_sig(context) ((context)->uc_mcontext.gregs[REG_EIP])
1257# define TRAP_sig(context) ((context)->uc_mcontext.gregs[REG_TRAPNO])
1258# define ERROR_sig(context) ((context)->uc_mcontext.gregs[REG_ERR])
d39bb24a 1259# define MASK_sig(context) ((context)->uc_sigmask)
d8ecc0b9
FB
1260#endif
1261
5fafdf24 1262int cpu_signal_handler(int host_signum, void *pinfo,
e4533c7a 1263 void *puc)
9de5e440 1264{
5a7b542b 1265 siginfo_t *info = pinfo;
d39bb24a
BS
1266#if defined(__OpenBSD__)
1267 struct sigcontext *uc = puc;
1268#else
9de5e440 1269 struct ucontext *uc = puc;
d39bb24a 1270#endif
9de5e440 1271 unsigned long pc;
bf3e8bf1 1272 int trapno;
97eb5b14 1273
d691f669
FB
1274#ifndef REG_EIP
1275/* for glibc 2.1 */
fd6ce8f6
FB
1276#define REG_EIP EIP
1277#define REG_ERR ERR
1278#define REG_TRAPNO TRAPNO
d691f669 1279#endif
d8ecc0b9
FB
1280 pc = EIP_sig(uc);
1281 trapno = TRAP_sig(uc);
ec6338ba
FB
1282 return handle_cpu_signal(pc, (unsigned long)info->si_addr,
1283 trapno == 0xe ?
1284 (ERROR_sig(uc) >> 1) & 1 : 0,
d39bb24a 1285 &MASK_sig(uc), puc);
2b413144
FB
1286}
1287
bc51c5c9
FB
1288#elif defined(__x86_64__)
1289
b3efe5c8 1290#ifdef __NetBSD__
d397abbd
BS
1291#define PC_sig(context) _UC_MACHINE_PC(context)
1292#define TRAP_sig(context) ((context)->uc_mcontext.__gregs[_REG_TRAPNO])
1293#define ERROR_sig(context) ((context)->uc_mcontext.__gregs[_REG_ERR])
1294#define MASK_sig(context) ((context)->uc_sigmask)
1295#elif defined(__OpenBSD__)
1296#define PC_sig(context) ((context)->sc_rip)
1297#define TRAP_sig(context) ((context)->sc_trapno)
1298#define ERROR_sig(context) ((context)->sc_err)
1299#define MASK_sig(context) ((context)->sc_mask)
b3efe5c8 1300#else
d397abbd
BS
1301#define PC_sig(context) ((context)->uc_mcontext.gregs[REG_RIP])
1302#define TRAP_sig(context) ((context)->uc_mcontext.gregs[REG_TRAPNO])
1303#define ERROR_sig(context) ((context)->uc_mcontext.gregs[REG_ERR])
1304#define MASK_sig(context) ((context)->uc_sigmask)
b3efe5c8
BS
1305#endif
1306
5a7b542b 1307int cpu_signal_handler(int host_signum, void *pinfo,
bc51c5c9
FB
1308 void *puc)
1309{
5a7b542b 1310 siginfo_t *info = pinfo;
bc51c5c9 1311 unsigned long pc;
b3efe5c8
BS
1312#ifdef __NetBSD__
1313 ucontext_t *uc = puc;
d397abbd
BS
1314#elif defined(__OpenBSD__)
1315 struct sigcontext *uc = puc;
b3efe5c8
BS
1316#else
1317 struct ucontext *uc = puc;
1318#endif
bc51c5c9 1319
d397abbd 1320 pc = PC_sig(uc);
5fafdf24 1321 return handle_cpu_signal(pc, (unsigned long)info->si_addr,
d397abbd
BS
1322 TRAP_sig(uc) == 0xe ?
1323 (ERROR_sig(uc) >> 1) & 1 : 0,
1324 &MASK_sig(uc), puc);
bc51c5c9
FB
1325}
1326
e58ffeb3 1327#elif defined(_ARCH_PPC)
2b413144 1328
83fb7adf
FB
1329/***********************************************************************
1330 * signal context platform-specific definitions
1331 * From Wine
1332 */
1333#ifdef linux
1334/* All Registers access - only for local access */
1335# define REG_sig(reg_name, context) ((context)->uc_mcontext.regs->reg_name)
1336/* Gpr Registers access */
1337# define GPR_sig(reg_num, context) REG_sig(gpr[reg_num], context)
1338# define IAR_sig(context) REG_sig(nip, context) /* Program counter */
1339# define MSR_sig(context) REG_sig(msr, context) /* Machine State Register (Supervisor) */
1340# define CTR_sig(context) REG_sig(ctr, context) /* Count register */
1341# define XER_sig(context) REG_sig(xer, context) /* User's integer exception register */
1342# define LR_sig(context) REG_sig(link, context) /* Link register */
1343# define CR_sig(context) REG_sig(ccr, context) /* Condition register */
1344/* Float Registers access */
1345# define FLOAT_sig(reg_num, context) (((double*)((char*)((context)->uc_mcontext.regs+48*4)))[reg_num])
1346# define FPSCR_sig(context) (*(int*)((char*)((context)->uc_mcontext.regs+(48+32*2)*4)))
1347/* Exception Registers access */
1348# define DAR_sig(context) REG_sig(dar, context)
1349# define DSISR_sig(context) REG_sig(dsisr, context)
1350# define TRAP_sig(context) REG_sig(trap, context)
1351#endif /* linux */
1352
1353#ifdef __APPLE__
1354# include <sys/ucontext.h>
1355typedef struct ucontext SIGCONTEXT;
1356/* All Registers access - only for local access */
1357# define REG_sig(reg_name, context) ((context)->uc_mcontext->ss.reg_name)
1358# define FLOATREG_sig(reg_name, context) ((context)->uc_mcontext->fs.reg_name)
1359# define EXCEPREG_sig(reg_name, context) ((context)->uc_mcontext->es.reg_name)
1360# define VECREG_sig(reg_name, context) ((context)->uc_mcontext->vs.reg_name)
1361/* Gpr Registers access */
1362# define GPR_sig(reg_num, context) REG_sig(r##reg_num, context)
1363# define IAR_sig(context) REG_sig(srr0, context) /* Program counter */
1364# define MSR_sig(context) REG_sig(srr1, context) /* Machine State Register (Supervisor) */
1365# define CTR_sig(context) REG_sig(ctr, context)
1366# define XER_sig(context) REG_sig(xer, context) /* Link register */
1367# define LR_sig(context) REG_sig(lr, context) /* User's integer exception register */
1368# define CR_sig(context) REG_sig(cr, context) /* Condition register */
1369/* Float Registers access */
1370# define FLOAT_sig(reg_num, context) FLOATREG_sig(fpregs[reg_num], context)
1371# define FPSCR_sig(context) ((double)FLOATREG_sig(fpscr, context))
1372/* Exception Registers access */
1373# define DAR_sig(context) EXCEPREG_sig(dar, context) /* Fault registers for coredump */
1374# define DSISR_sig(context) EXCEPREG_sig(dsisr, context)
1375# define TRAP_sig(context) EXCEPREG_sig(exception, context) /* number of powerpc exception taken */
1376#endif /* __APPLE__ */
1377
5fafdf24 1378int cpu_signal_handler(int host_signum, void *pinfo,
e4533c7a 1379 void *puc)
2b413144 1380{
5a7b542b 1381 siginfo_t *info = pinfo;
25eb4484 1382 struct ucontext *uc = puc;
25eb4484 1383 unsigned long pc;
25eb4484
FB
1384 int is_write;
1385
83fb7adf 1386 pc = IAR_sig(uc);
25eb4484
FB
1387 is_write = 0;
1388#if 0
1389 /* ppc 4xx case */
83fb7adf 1390 if (DSISR_sig(uc) & 0x00800000)
25eb4484
FB
1391 is_write = 1;
1392#else
83fb7adf 1393 if (TRAP_sig(uc) != 0x400 && (DSISR_sig(uc) & 0x02000000))
25eb4484
FB
1394 is_write = 1;
1395#endif
5fafdf24 1396 return handle_cpu_signal(pc, (unsigned long)info->si_addr,
bf3e8bf1 1397 is_write, &uc->uc_sigmask, puc);
2b413144
FB
1398}
1399
2f87c607
FB
1400#elif defined(__alpha__)
1401
5fafdf24 1402int cpu_signal_handler(int host_signum, void *pinfo,
2f87c607
FB
1403 void *puc)
1404{
5a7b542b 1405 siginfo_t *info = pinfo;
2f87c607
FB
1406 struct ucontext *uc = puc;
1407 uint32_t *pc = uc->uc_mcontext.sc_pc;
1408 uint32_t insn = *pc;
1409 int is_write = 0;
1410
8c6939c0 1411 /* XXX: need kernel patch to get write flag faster */
2f87c607
FB
1412 switch (insn >> 26) {
1413 case 0x0d: // stw
1414 case 0x0e: // stb
1415 case 0x0f: // stq_u
1416 case 0x24: // stf
1417 case 0x25: // stg
1418 case 0x26: // sts
1419 case 0x27: // stt
1420 case 0x2c: // stl
1421 case 0x2d: // stq
1422 case 0x2e: // stl_c
1423 case 0x2f: // stq_c
1424 is_write = 1;
1425 }
1426
5fafdf24 1427 return handle_cpu_signal(pc, (unsigned long)info->si_addr,
bf3e8bf1 1428 is_write, &uc->uc_sigmask, puc);
2f87c607 1429}
8c6939c0
FB
1430#elif defined(__sparc__)
1431
5fafdf24 1432int cpu_signal_handler(int host_signum, void *pinfo,
e4533c7a 1433 void *puc)
8c6939c0 1434{
5a7b542b 1435 siginfo_t *info = pinfo;
8c6939c0
FB
1436 int is_write;
1437 uint32_t insn;
6b4c11cd 1438#if !defined(__arch64__) || defined(HOST_SOLARIS)
c9e1e2b0
BS
1439 uint32_t *regs = (uint32_t *)(info + 1);
1440 void *sigmask = (regs + 20);
8c6939c0 1441 /* XXX: is there a standard glibc define ? */
c9e1e2b0
BS
1442 unsigned long pc = regs[1];
1443#else
84778508 1444#ifdef __linux__
c9e1e2b0
BS
1445 struct sigcontext *sc = puc;
1446 unsigned long pc = sc->sigc_regs.tpc;
1447 void *sigmask = (void *)sc->sigc_mask;
84778508
BS
1448#elif defined(__OpenBSD__)
1449 struct sigcontext *uc = puc;
1450 unsigned long pc = uc->sc_pc;
1451 void *sigmask = (void *)(long)uc->sc_mask;
1452#endif
c9e1e2b0
BS
1453#endif
1454
8c6939c0
FB
1455 /* XXX: need kernel patch to get write flag faster */
1456 is_write = 0;
1457 insn = *(uint32_t *)pc;
1458 if ((insn >> 30) == 3) {
1459 switch((insn >> 19) & 0x3f) {
1460 case 0x05: // stb
d877fa5a 1461 case 0x15: // stba
8c6939c0 1462 case 0x06: // sth
d877fa5a 1463 case 0x16: // stha
8c6939c0 1464 case 0x04: // st
d877fa5a 1465 case 0x14: // sta
8c6939c0 1466 case 0x07: // std
d877fa5a
BS
1467 case 0x17: // stda
1468 case 0x0e: // stx
1469 case 0x1e: // stxa
8c6939c0 1470 case 0x24: // stf
d877fa5a 1471 case 0x34: // stfa
8c6939c0 1472 case 0x27: // stdf
d877fa5a
BS
1473 case 0x37: // stdfa
1474 case 0x26: // stqf
1475 case 0x36: // stqfa
8c6939c0 1476 case 0x25: // stfsr
d877fa5a
BS
1477 case 0x3c: // casa
1478 case 0x3e: // casxa
8c6939c0
FB
1479 is_write = 1;
1480 break;
1481 }
1482 }
5fafdf24 1483 return handle_cpu_signal(pc, (unsigned long)info->si_addr,
bf3e8bf1 1484 is_write, sigmask, NULL);
8c6939c0
FB
1485}
1486
1487#elif defined(__arm__)
1488
5fafdf24 1489int cpu_signal_handler(int host_signum, void *pinfo,
e4533c7a 1490 void *puc)
8c6939c0 1491{
5a7b542b 1492 siginfo_t *info = pinfo;
8c6939c0
FB
1493 struct ucontext *uc = puc;
1494 unsigned long pc;
1495 int is_write;
3b46e624 1496
48bbf11b 1497#if (__GLIBC__ < 2 || (__GLIBC__ == 2 && __GLIBC_MINOR__ <= 3))
5c49b363
AZ
1498 pc = uc->uc_mcontext.gregs[R15];
1499#else
4eee57f5 1500 pc = uc->uc_mcontext.arm_pc;
5c49b363 1501#endif
8c6939c0
FB
1502 /* XXX: compute is_write */
1503 is_write = 0;
5fafdf24 1504 return handle_cpu_signal(pc, (unsigned long)info->si_addr,
8c6939c0 1505 is_write,
f3a9676a 1506 &uc->uc_sigmask, puc);
8c6939c0
FB
1507}
1508
38e584a0
FB
1509#elif defined(__mc68000)
1510
5fafdf24 1511int cpu_signal_handler(int host_signum, void *pinfo,
38e584a0
FB
1512 void *puc)
1513{
5a7b542b 1514 siginfo_t *info = pinfo;
38e584a0
FB
1515 struct ucontext *uc = puc;
1516 unsigned long pc;
1517 int is_write;
3b46e624 1518
38e584a0
FB
1519 pc = uc->uc_mcontext.gregs[16];
1520 /* XXX: compute is_write */
1521 is_write = 0;
5fafdf24 1522 return handle_cpu_signal(pc, (unsigned long)info->si_addr,
38e584a0 1523 is_write,
bf3e8bf1 1524 &uc->uc_sigmask, puc);
38e584a0
FB
1525}
1526
b8076a74
FB
1527#elif defined(__ia64)
1528
1529#ifndef __ISR_VALID
1530 /* This ought to be in <bits/siginfo.h>... */
1531# define __ISR_VALID 1
b8076a74
FB
1532#endif
1533
5a7b542b 1534int cpu_signal_handler(int host_signum, void *pinfo, void *puc)
b8076a74 1535{
5a7b542b 1536 siginfo_t *info = pinfo;
b8076a74
FB
1537 struct ucontext *uc = puc;
1538 unsigned long ip;
1539 int is_write = 0;
1540
1541 ip = uc->uc_mcontext.sc_ip;
1542 switch (host_signum) {
1543 case SIGILL:
1544 case SIGFPE:
1545 case SIGSEGV:
1546 case SIGBUS:
1547 case SIGTRAP:
fd4a43e4 1548 if (info->si_code && (info->si_segvflags & __ISR_VALID))
b8076a74
FB
1549 /* ISR.W (write-access) is bit 33: */
1550 is_write = (info->si_isr >> 33) & 1;
1551 break;
1552
1553 default:
1554 break;
1555 }
1556 return handle_cpu_signal(ip, (unsigned long)info->si_addr,
1557 is_write,
1558 &uc->uc_sigmask, puc);
1559}
1560
90cb9493
FB
1561#elif defined(__s390__)
1562
5fafdf24 1563int cpu_signal_handler(int host_signum, void *pinfo,
90cb9493
FB
1564 void *puc)
1565{
5a7b542b 1566 siginfo_t *info = pinfo;
90cb9493
FB
1567 struct ucontext *uc = puc;
1568 unsigned long pc;
1569 int is_write;
3b46e624 1570
90cb9493
FB
1571 pc = uc->uc_mcontext.psw.addr;
1572 /* XXX: compute is_write */
1573 is_write = 0;
5fafdf24 1574 return handle_cpu_signal(pc, (unsigned long)info->si_addr,
c4b89d18
TS
1575 is_write, &uc->uc_sigmask, puc);
1576}
1577
1578#elif defined(__mips__)
1579
5fafdf24 1580int cpu_signal_handler(int host_signum, void *pinfo,
c4b89d18
TS
1581 void *puc)
1582{
9617efe8 1583 siginfo_t *info = pinfo;
c4b89d18
TS
1584 struct ucontext *uc = puc;
1585 greg_t pc = uc->uc_mcontext.pc;
1586 int is_write;
3b46e624 1587
c4b89d18
TS
1588 /* XXX: compute is_write */
1589 is_write = 0;
5fafdf24 1590 return handle_cpu_signal(pc, (unsigned long)info->si_addr,
c4b89d18 1591 is_write, &uc->uc_sigmask, puc);
90cb9493
FB
1592}
1593
f54b3f92
AJ
1594#elif defined(__hppa__)
1595
1596int cpu_signal_handler(int host_signum, void *pinfo,
1597 void *puc)
1598{
1599 struct siginfo *info = pinfo;
1600 struct ucontext *uc = puc;
1601 unsigned long pc;
1602 int is_write;
1603
1604 pc = uc->uc_mcontext.sc_iaoq[0];
1605 /* FIXME: compute is_write */
1606 is_write = 0;
1607 return handle_cpu_signal(pc, (unsigned long)info->si_addr,
1608 is_write,
1609 &uc->uc_sigmask, puc);
1610}
1611
9de5e440 1612#else
2b413144 1613
3fb2ded1 1614#error host CPU specific signal handler needed
2b413144 1615
9de5e440 1616#endif
67b915a5
FB
1617
1618#endif /* !defined(CONFIG_SOFTMMU) */