]> git.proxmox.com Git - mirror_qemu.git/blob - target/alpha/translate.c
f6247bf38db52e8817411df2627c6a5535d10a32
[mirror_qemu.git] / target / alpha / translate.c
1 /*
2 * Alpha emulation cpu translation for qemu.
3 *
4 * Copyright (c) 2007 Jocelyn Mayer
5 *
6 * This library is free software; you can redistribute it and/or
7 * modify it under the terms of the GNU Lesser General Public
8 * License as published by the Free Software Foundation; either
9 * version 2 of the License, or (at your option) any later version.
10 *
11 * This library is distributed in the hope that it will be useful,
12 * but WITHOUT ANY WARRANTY; without even the implied warranty of
13 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
14 * Lesser General Public License for more details.
15 *
16 * You should have received a copy of the GNU Lesser General Public
17 * License along with this library; if not, see <http://www.gnu.org/licenses/>.
18 */
19
20 #include "qemu/osdep.h"
21 #include "cpu.h"
22 #include "sysemu/cpus.h"
23 #include "disas/disas.h"
24 #include "qemu/host-utils.h"
25 #include "exec/exec-all.h"
26 #include "tcg-op.h"
27 #include "exec/cpu_ldst.h"
28 #include "exec/helper-proto.h"
29 #include "exec/helper-gen.h"
30 #include "trace-tcg.h"
31 #include "exec/translator.h"
32 #include "exec/log.h"
33
34
35 #undef ALPHA_DEBUG_DISAS
36 #define CONFIG_SOFTFLOAT_INLINE
37
38 #ifdef ALPHA_DEBUG_DISAS
39 # define LOG_DISAS(...) qemu_log_mask(CPU_LOG_TB_IN_ASM, ## __VA_ARGS__)
40 #else
41 # define LOG_DISAS(...) do { } while (0)
42 #endif
43
44 typedef struct DisasContext DisasContext;
45 struct DisasContext {
46 DisasContextBase base;
47
48 #ifndef CONFIG_USER_ONLY
49 uint64_t palbr;
50 #endif
51 uint32_t tbflags;
52 int mem_idx;
53
54 /* implver and amask values for this CPU. */
55 int implver;
56 int amask;
57
58 /* Current rounding mode for this TB. */
59 int tb_rm;
60 /* Current flush-to-zero setting for this TB. */
61 int tb_ftz;
62
63 /* The set of registers active in the current context. */
64 TCGv *ir;
65
66 /* Temporaries for $31 and $f31 as source and destination. */
67 TCGv zero;
68 TCGv sink;
69 /* Temporary for immediate constants. */
70 TCGv lit;
71 };
72
73 /* Target-specific return values from translate_one, indicating the
74 state of the TB. Note that DISAS_NEXT indicates that we are not
75 exiting the TB. */
76 #define DISAS_PC_UPDATED_NOCHAIN DISAS_TARGET_0
77 #define DISAS_PC_UPDATED DISAS_TARGET_1
78 #define DISAS_PC_STALE DISAS_TARGET_2
79
80 /* global register indexes */
81 static TCGv_env cpu_env;
82 static TCGv cpu_std_ir[31];
83 static TCGv cpu_fir[31];
84 static TCGv cpu_pc;
85 static TCGv cpu_lock_addr;
86 static TCGv cpu_lock_value;
87
88 #ifndef CONFIG_USER_ONLY
89 static TCGv cpu_pal_ir[31];
90 #endif
91
92 #include "exec/gen-icount.h"
93
94 void alpha_translate_init(void)
95 {
96 #define DEF_VAR(V) { &cpu_##V, #V, offsetof(CPUAlphaState, V) }
97
98 typedef struct { TCGv *var; const char *name; int ofs; } GlobalVar;
99 static const GlobalVar vars[] = {
100 DEF_VAR(pc),
101 DEF_VAR(lock_addr),
102 DEF_VAR(lock_value),
103 };
104
105 #undef DEF_VAR
106
107 /* Use the symbolic register names that match the disassembler. */
108 static const char greg_names[31][4] = {
109 "v0", "t0", "t1", "t2", "t3", "t4", "t5", "t6",
110 "t7", "s0", "s1", "s2", "s3", "s4", "s5", "fp",
111 "a0", "a1", "a2", "a3", "a4", "a5", "t8", "t9",
112 "t10", "t11", "ra", "t12", "at", "gp", "sp"
113 };
114 static const char freg_names[31][4] = {
115 "f0", "f1", "f2", "f3", "f4", "f5", "f6", "f7",
116 "f8", "f9", "f10", "f11", "f12", "f13", "f14", "f15",
117 "f16", "f17", "f18", "f19", "f20", "f21", "f22", "f23",
118 "f24", "f25", "f26", "f27", "f28", "f29", "f30"
119 };
120 #ifndef CONFIG_USER_ONLY
121 static const char shadow_names[8][8] = {
122 "pal_t7", "pal_s0", "pal_s1", "pal_s2",
123 "pal_s3", "pal_s4", "pal_s5", "pal_t11"
124 };
125 #endif
126
127 int i;
128
129 cpu_env = tcg_global_reg_new_ptr(TCG_AREG0, "env");
130 tcg_ctx->tcg_env = cpu_env;
131
132 for (i = 0; i < 31; i++) {
133 cpu_std_ir[i] = tcg_global_mem_new_i64(cpu_env,
134 offsetof(CPUAlphaState, ir[i]),
135 greg_names[i]);
136 }
137
138 for (i = 0; i < 31; i++) {
139 cpu_fir[i] = tcg_global_mem_new_i64(cpu_env,
140 offsetof(CPUAlphaState, fir[i]),
141 freg_names[i]);
142 }
143
144 #ifndef CONFIG_USER_ONLY
145 memcpy(cpu_pal_ir, cpu_std_ir, sizeof(cpu_pal_ir));
146 for (i = 0; i < 8; i++) {
147 int r = (i == 7 ? 25 : i + 8);
148 cpu_pal_ir[r] = tcg_global_mem_new_i64(cpu_env,
149 offsetof(CPUAlphaState,
150 shadow[i]),
151 shadow_names[i]);
152 }
153 #endif
154
155 for (i = 0; i < ARRAY_SIZE(vars); ++i) {
156 const GlobalVar *v = &vars[i];
157 *v->var = tcg_global_mem_new_i64(cpu_env, v->ofs, v->name);
158 }
159 }
160
161 static TCGv load_zero(DisasContext *ctx)
162 {
163 if (TCGV_IS_UNUSED_I64(ctx->zero)) {
164 ctx->zero = tcg_const_i64(0);
165 }
166 return ctx->zero;
167 }
168
169 static TCGv dest_sink(DisasContext *ctx)
170 {
171 if (TCGV_IS_UNUSED_I64(ctx->sink)) {
172 ctx->sink = tcg_temp_new();
173 }
174 return ctx->sink;
175 }
176
177 static void free_context_temps(DisasContext *ctx)
178 {
179 if (!TCGV_IS_UNUSED_I64(ctx->sink)) {
180 tcg_gen_discard_i64(ctx->sink);
181 tcg_temp_free(ctx->sink);
182 TCGV_UNUSED_I64(ctx->sink);
183 }
184 if (!TCGV_IS_UNUSED_I64(ctx->zero)) {
185 tcg_temp_free(ctx->zero);
186 TCGV_UNUSED_I64(ctx->zero);
187 }
188 if (!TCGV_IS_UNUSED_I64(ctx->lit)) {
189 tcg_temp_free(ctx->lit);
190 TCGV_UNUSED_I64(ctx->lit);
191 }
192 }
193
194 static TCGv load_gpr(DisasContext *ctx, unsigned reg)
195 {
196 if (likely(reg < 31)) {
197 return ctx->ir[reg];
198 } else {
199 return load_zero(ctx);
200 }
201 }
202
203 static TCGv load_gpr_lit(DisasContext *ctx, unsigned reg,
204 uint8_t lit, bool islit)
205 {
206 if (islit) {
207 ctx->lit = tcg_const_i64(lit);
208 return ctx->lit;
209 } else if (likely(reg < 31)) {
210 return ctx->ir[reg];
211 } else {
212 return load_zero(ctx);
213 }
214 }
215
216 static TCGv dest_gpr(DisasContext *ctx, unsigned reg)
217 {
218 if (likely(reg < 31)) {
219 return ctx->ir[reg];
220 } else {
221 return dest_sink(ctx);
222 }
223 }
224
225 static TCGv load_fpr(DisasContext *ctx, unsigned reg)
226 {
227 if (likely(reg < 31)) {
228 return cpu_fir[reg];
229 } else {
230 return load_zero(ctx);
231 }
232 }
233
234 static TCGv dest_fpr(DisasContext *ctx, unsigned reg)
235 {
236 if (likely(reg < 31)) {
237 return cpu_fir[reg];
238 } else {
239 return dest_sink(ctx);
240 }
241 }
242
243 static int get_flag_ofs(unsigned shift)
244 {
245 int ofs = offsetof(CPUAlphaState, flags);
246 #ifdef HOST_WORDS_BIGENDIAN
247 ofs += 3 - (shift / 8);
248 #else
249 ofs += shift / 8;
250 #endif
251 return ofs;
252 }
253
254 static void ld_flag_byte(TCGv val, unsigned shift)
255 {
256 tcg_gen_ld8u_i64(val, cpu_env, get_flag_ofs(shift));
257 }
258
259 static void st_flag_byte(TCGv val, unsigned shift)
260 {
261 tcg_gen_st8_i64(val, cpu_env, get_flag_ofs(shift));
262 }
263
264 static void gen_excp_1(int exception, int error_code)
265 {
266 TCGv_i32 tmp1, tmp2;
267
268 tmp1 = tcg_const_i32(exception);
269 tmp2 = tcg_const_i32(error_code);
270 gen_helper_excp(cpu_env, tmp1, tmp2);
271 tcg_temp_free_i32(tmp2);
272 tcg_temp_free_i32(tmp1);
273 }
274
275 static DisasJumpType gen_excp(DisasContext *ctx, int exception, int error_code)
276 {
277 tcg_gen_movi_i64(cpu_pc, ctx->base.pc_next);
278 gen_excp_1(exception, error_code);
279 return DISAS_NORETURN;
280 }
281
282 static inline DisasJumpType gen_invalid(DisasContext *ctx)
283 {
284 return gen_excp(ctx, EXCP_OPCDEC, 0);
285 }
286
287 static inline void gen_qemu_ldf(TCGv t0, TCGv t1, int flags)
288 {
289 TCGv_i32 tmp32 = tcg_temp_new_i32();
290 tcg_gen_qemu_ld_i32(tmp32, t1, flags, MO_LEUL);
291 gen_helper_memory_to_f(t0, tmp32);
292 tcg_temp_free_i32(tmp32);
293 }
294
295 static inline void gen_qemu_ldg(TCGv t0, TCGv t1, int flags)
296 {
297 TCGv tmp = tcg_temp_new();
298 tcg_gen_qemu_ld_i64(tmp, t1, flags, MO_LEQ);
299 gen_helper_memory_to_g(t0, tmp);
300 tcg_temp_free(tmp);
301 }
302
303 static inline void gen_qemu_lds(TCGv t0, TCGv t1, int flags)
304 {
305 TCGv_i32 tmp32 = tcg_temp_new_i32();
306 tcg_gen_qemu_ld_i32(tmp32, t1, flags, MO_LEUL);
307 gen_helper_memory_to_s(t0, tmp32);
308 tcg_temp_free_i32(tmp32);
309 }
310
311 static inline void gen_qemu_ldl_l(TCGv t0, TCGv t1, int flags)
312 {
313 tcg_gen_qemu_ld_i64(t0, t1, flags, MO_LESL);
314 tcg_gen_mov_i64(cpu_lock_addr, t1);
315 tcg_gen_mov_i64(cpu_lock_value, t0);
316 }
317
318 static inline void gen_qemu_ldq_l(TCGv t0, TCGv t1, int flags)
319 {
320 tcg_gen_qemu_ld_i64(t0, t1, flags, MO_LEQ);
321 tcg_gen_mov_i64(cpu_lock_addr, t1);
322 tcg_gen_mov_i64(cpu_lock_value, t0);
323 }
324
325 static inline void gen_load_mem(DisasContext *ctx,
326 void (*tcg_gen_qemu_load)(TCGv t0, TCGv t1,
327 int flags),
328 int ra, int rb, int32_t disp16, bool fp,
329 bool clear)
330 {
331 TCGv tmp, addr, va;
332
333 /* LDQ_U with ra $31 is UNOP. Other various loads are forms of
334 prefetches, which we can treat as nops. No worries about
335 missed exceptions here. */
336 if (unlikely(ra == 31)) {
337 return;
338 }
339
340 tmp = tcg_temp_new();
341 addr = load_gpr(ctx, rb);
342
343 if (disp16) {
344 tcg_gen_addi_i64(tmp, addr, disp16);
345 addr = tmp;
346 }
347 if (clear) {
348 tcg_gen_andi_i64(tmp, addr, ~0x7);
349 addr = tmp;
350 }
351
352 va = (fp ? cpu_fir[ra] : ctx->ir[ra]);
353 tcg_gen_qemu_load(va, addr, ctx->mem_idx);
354
355 tcg_temp_free(tmp);
356 }
357
358 static inline void gen_qemu_stf(TCGv t0, TCGv t1, int flags)
359 {
360 TCGv_i32 tmp32 = tcg_temp_new_i32();
361 gen_helper_f_to_memory(tmp32, t0);
362 tcg_gen_qemu_st_i32(tmp32, t1, flags, MO_LEUL);
363 tcg_temp_free_i32(tmp32);
364 }
365
366 static inline void gen_qemu_stg(TCGv t0, TCGv t1, int flags)
367 {
368 TCGv tmp = tcg_temp_new();
369 gen_helper_g_to_memory(tmp, t0);
370 tcg_gen_qemu_st_i64(tmp, t1, flags, MO_LEQ);
371 tcg_temp_free(tmp);
372 }
373
374 static inline void gen_qemu_sts(TCGv t0, TCGv t1, int flags)
375 {
376 TCGv_i32 tmp32 = tcg_temp_new_i32();
377 gen_helper_s_to_memory(tmp32, t0);
378 tcg_gen_qemu_st_i32(tmp32, t1, flags, MO_LEUL);
379 tcg_temp_free_i32(tmp32);
380 }
381
382 static inline void gen_store_mem(DisasContext *ctx,
383 void (*tcg_gen_qemu_store)(TCGv t0, TCGv t1,
384 int flags),
385 int ra, int rb, int32_t disp16, bool fp,
386 bool clear)
387 {
388 TCGv tmp, addr, va;
389
390 tmp = tcg_temp_new();
391 addr = load_gpr(ctx, rb);
392
393 if (disp16) {
394 tcg_gen_addi_i64(tmp, addr, disp16);
395 addr = tmp;
396 }
397 if (clear) {
398 tcg_gen_andi_i64(tmp, addr, ~0x7);
399 addr = tmp;
400 }
401
402 va = (fp ? load_fpr(ctx, ra) : load_gpr(ctx, ra));
403 tcg_gen_qemu_store(va, addr, ctx->mem_idx);
404
405 tcg_temp_free(tmp);
406 }
407
408 static DisasJumpType gen_store_conditional(DisasContext *ctx, int ra, int rb,
409 int32_t disp16, int mem_idx,
410 TCGMemOp op)
411 {
412 TCGLabel *lab_fail, *lab_done;
413 TCGv addr, val;
414
415 addr = tcg_temp_new_i64();
416 tcg_gen_addi_i64(addr, load_gpr(ctx, rb), disp16);
417 free_context_temps(ctx);
418
419 lab_fail = gen_new_label();
420 lab_done = gen_new_label();
421 tcg_gen_brcond_i64(TCG_COND_NE, addr, cpu_lock_addr, lab_fail);
422 tcg_temp_free_i64(addr);
423
424 val = tcg_temp_new_i64();
425 tcg_gen_atomic_cmpxchg_i64(val, cpu_lock_addr, cpu_lock_value,
426 load_gpr(ctx, ra), mem_idx, op);
427 free_context_temps(ctx);
428
429 if (ra != 31) {
430 tcg_gen_setcond_i64(TCG_COND_EQ, ctx->ir[ra], val, cpu_lock_value);
431 }
432 tcg_temp_free_i64(val);
433 tcg_gen_br(lab_done);
434
435 gen_set_label(lab_fail);
436 if (ra != 31) {
437 tcg_gen_movi_i64(ctx->ir[ra], 0);
438 }
439
440 gen_set_label(lab_done);
441 tcg_gen_movi_i64(cpu_lock_addr, -1);
442 return DISAS_NEXT;
443 }
444
445 static bool in_superpage(DisasContext *ctx, int64_t addr)
446 {
447 #ifndef CONFIG_USER_ONLY
448 return ((ctx->tbflags & ENV_FLAG_PS_USER) == 0
449 && addr >> TARGET_VIRT_ADDR_SPACE_BITS == -1
450 && ((addr >> 41) & 3) == 2);
451 #else
452 return false;
453 #endif
454 }
455
456 static bool use_exit_tb(DisasContext *ctx)
457 {
458 return ((tb_cflags(ctx->base.tb) & CF_LAST_IO)
459 || ctx->base.singlestep_enabled
460 || singlestep);
461 }
462
463 static bool use_goto_tb(DisasContext *ctx, uint64_t dest)
464 {
465 /* Suppress goto_tb in the case of single-steping and IO. */
466 if (unlikely(use_exit_tb(ctx))) {
467 return false;
468 }
469 #ifndef CONFIG_USER_ONLY
470 /* If the destination is in the superpage, the page perms can't change. */
471 if (in_superpage(ctx, dest)) {
472 return true;
473 }
474 /* Check for the dest on the same page as the start of the TB. */
475 return ((ctx->base.tb->pc ^ dest) & TARGET_PAGE_MASK) == 0;
476 #else
477 return true;
478 #endif
479 }
480
481 static DisasJumpType gen_bdirect(DisasContext *ctx, int ra, int32_t disp)
482 {
483 uint64_t dest = ctx->base.pc_next + (disp << 2);
484
485 if (ra != 31) {
486 tcg_gen_movi_i64(ctx->ir[ra], ctx->base.pc_next);
487 }
488
489 /* Notice branch-to-next; used to initialize RA with the PC. */
490 if (disp == 0) {
491 return 0;
492 } else if (use_goto_tb(ctx, dest)) {
493 tcg_gen_goto_tb(0);
494 tcg_gen_movi_i64(cpu_pc, dest);
495 tcg_gen_exit_tb((uintptr_t)ctx->base.tb);
496 return DISAS_NORETURN;
497 } else {
498 tcg_gen_movi_i64(cpu_pc, dest);
499 return DISAS_PC_UPDATED;
500 }
501 }
502
503 static DisasJumpType gen_bcond_internal(DisasContext *ctx, TCGCond cond,
504 TCGv cmp, int32_t disp)
505 {
506 uint64_t dest = ctx->base.pc_next + (disp << 2);
507 TCGLabel *lab_true = gen_new_label();
508
509 if (use_goto_tb(ctx, dest)) {
510 tcg_gen_brcondi_i64(cond, cmp, 0, lab_true);
511
512 tcg_gen_goto_tb(0);
513 tcg_gen_movi_i64(cpu_pc, ctx->base.pc_next);
514 tcg_gen_exit_tb((uintptr_t)ctx->base.tb);
515
516 gen_set_label(lab_true);
517 tcg_gen_goto_tb(1);
518 tcg_gen_movi_i64(cpu_pc, dest);
519 tcg_gen_exit_tb((uintptr_t)ctx->base.tb + 1);
520
521 return DISAS_NORETURN;
522 } else {
523 TCGv_i64 z = tcg_const_i64(0);
524 TCGv_i64 d = tcg_const_i64(dest);
525 TCGv_i64 p = tcg_const_i64(ctx->base.pc_next);
526
527 tcg_gen_movcond_i64(cond, cpu_pc, cmp, z, d, p);
528
529 tcg_temp_free_i64(z);
530 tcg_temp_free_i64(d);
531 tcg_temp_free_i64(p);
532 return DISAS_PC_UPDATED;
533 }
534 }
535
536 static DisasJumpType gen_bcond(DisasContext *ctx, TCGCond cond, int ra,
537 int32_t disp, int mask)
538 {
539 if (mask) {
540 TCGv tmp = tcg_temp_new();
541 DisasJumpType ret;
542
543 tcg_gen_andi_i64(tmp, load_gpr(ctx, ra), 1);
544 ret = gen_bcond_internal(ctx, cond, tmp, disp);
545 tcg_temp_free(tmp);
546 return ret;
547 }
548 return gen_bcond_internal(ctx, cond, load_gpr(ctx, ra), disp);
549 }
550
551 /* Fold -0.0 for comparison with COND. */
552
553 static void gen_fold_mzero(TCGCond cond, TCGv dest, TCGv src)
554 {
555 uint64_t mzero = 1ull << 63;
556
557 switch (cond) {
558 case TCG_COND_LE:
559 case TCG_COND_GT:
560 /* For <= or >, the -0.0 value directly compares the way we want. */
561 tcg_gen_mov_i64(dest, src);
562 break;
563
564 case TCG_COND_EQ:
565 case TCG_COND_NE:
566 /* For == or !=, we can simply mask off the sign bit and compare. */
567 tcg_gen_andi_i64(dest, src, mzero - 1);
568 break;
569
570 case TCG_COND_GE:
571 case TCG_COND_LT:
572 /* For >= or <, map -0.0 to +0.0 via comparison and mask. */
573 tcg_gen_setcondi_i64(TCG_COND_NE, dest, src, mzero);
574 tcg_gen_neg_i64(dest, dest);
575 tcg_gen_and_i64(dest, dest, src);
576 break;
577
578 default:
579 abort();
580 }
581 }
582
583 static DisasJumpType gen_fbcond(DisasContext *ctx, TCGCond cond, int ra,
584 int32_t disp)
585 {
586 TCGv cmp_tmp = tcg_temp_new();
587 DisasJumpType ret;
588
589 gen_fold_mzero(cond, cmp_tmp, load_fpr(ctx, ra));
590 ret = gen_bcond_internal(ctx, cond, cmp_tmp, disp);
591 tcg_temp_free(cmp_tmp);
592 return ret;
593 }
594
595 static void gen_fcmov(DisasContext *ctx, TCGCond cond, int ra, int rb, int rc)
596 {
597 TCGv_i64 va, vb, z;
598
599 z = load_zero(ctx);
600 vb = load_fpr(ctx, rb);
601 va = tcg_temp_new();
602 gen_fold_mzero(cond, va, load_fpr(ctx, ra));
603
604 tcg_gen_movcond_i64(cond, dest_fpr(ctx, rc), va, z, vb, load_fpr(ctx, rc));
605
606 tcg_temp_free(va);
607 }
608
609 #define QUAL_RM_N 0x080 /* Round mode nearest even */
610 #define QUAL_RM_C 0x000 /* Round mode chopped */
611 #define QUAL_RM_M 0x040 /* Round mode minus infinity */
612 #define QUAL_RM_D 0x0c0 /* Round mode dynamic */
613 #define QUAL_RM_MASK 0x0c0
614
615 #define QUAL_U 0x100 /* Underflow enable (fp output) */
616 #define QUAL_V 0x100 /* Overflow enable (int output) */
617 #define QUAL_S 0x400 /* Software completion enable */
618 #define QUAL_I 0x200 /* Inexact detection enable */
619
620 static void gen_qual_roundmode(DisasContext *ctx, int fn11)
621 {
622 TCGv_i32 tmp;
623
624 fn11 &= QUAL_RM_MASK;
625 if (fn11 == ctx->tb_rm) {
626 return;
627 }
628 ctx->tb_rm = fn11;
629
630 tmp = tcg_temp_new_i32();
631 switch (fn11) {
632 case QUAL_RM_N:
633 tcg_gen_movi_i32(tmp, float_round_nearest_even);
634 break;
635 case QUAL_RM_C:
636 tcg_gen_movi_i32(tmp, float_round_to_zero);
637 break;
638 case QUAL_RM_M:
639 tcg_gen_movi_i32(tmp, float_round_down);
640 break;
641 case QUAL_RM_D:
642 tcg_gen_ld8u_i32(tmp, cpu_env,
643 offsetof(CPUAlphaState, fpcr_dyn_round));
644 break;
645 }
646
647 #if defined(CONFIG_SOFTFLOAT_INLINE)
648 /* ??? The "fpu/softfloat.h" interface is to call set_float_rounding_mode.
649 With CONFIG_SOFTFLOAT that expands to an out-of-line call that just
650 sets the one field. */
651 tcg_gen_st8_i32(tmp, cpu_env,
652 offsetof(CPUAlphaState, fp_status.float_rounding_mode));
653 #else
654 gen_helper_setroundmode(tmp);
655 #endif
656
657 tcg_temp_free_i32(tmp);
658 }
659
660 static void gen_qual_flushzero(DisasContext *ctx, int fn11)
661 {
662 TCGv_i32 tmp;
663
664 fn11 &= QUAL_U;
665 if (fn11 == ctx->tb_ftz) {
666 return;
667 }
668 ctx->tb_ftz = fn11;
669
670 tmp = tcg_temp_new_i32();
671 if (fn11) {
672 /* Underflow is enabled, use the FPCR setting. */
673 tcg_gen_ld8u_i32(tmp, cpu_env,
674 offsetof(CPUAlphaState, fpcr_flush_to_zero));
675 } else {
676 /* Underflow is disabled, force flush-to-zero. */
677 tcg_gen_movi_i32(tmp, 1);
678 }
679
680 #if defined(CONFIG_SOFTFLOAT_INLINE)
681 tcg_gen_st8_i32(tmp, cpu_env,
682 offsetof(CPUAlphaState, fp_status.flush_to_zero));
683 #else
684 gen_helper_setflushzero(tmp);
685 #endif
686
687 tcg_temp_free_i32(tmp);
688 }
689
690 static TCGv gen_ieee_input(DisasContext *ctx, int reg, int fn11, int is_cmp)
691 {
692 TCGv val;
693
694 if (unlikely(reg == 31)) {
695 val = load_zero(ctx);
696 } else {
697 val = cpu_fir[reg];
698 if ((fn11 & QUAL_S) == 0) {
699 if (is_cmp) {
700 gen_helper_ieee_input_cmp(cpu_env, val);
701 } else {
702 gen_helper_ieee_input(cpu_env, val);
703 }
704 } else {
705 #ifndef CONFIG_USER_ONLY
706 /* In system mode, raise exceptions for denormals like real
707 hardware. In user mode, proceed as if the OS completion
708 handler is handling the denormal as per spec. */
709 gen_helper_ieee_input_s(cpu_env, val);
710 #endif
711 }
712 }
713 return val;
714 }
715
716 static void gen_fp_exc_raise(int rc, int fn11)
717 {
718 /* ??? We ought to be able to do something with imprecise exceptions.
719 E.g. notice we're still in the trap shadow of something within the
720 TB and do not generate the code to signal the exception; end the TB
721 when an exception is forced to arrive, either by consumption of a
722 register value or TRAPB or EXCB. */
723 TCGv_i32 reg, ign;
724 uint32_t ignore = 0;
725
726 if (!(fn11 & QUAL_U)) {
727 /* Note that QUAL_U == QUAL_V, so ignore either. */
728 ignore |= FPCR_UNF | FPCR_IOV;
729 }
730 if (!(fn11 & QUAL_I)) {
731 ignore |= FPCR_INE;
732 }
733 ign = tcg_const_i32(ignore);
734
735 /* ??? Pass in the regno of the destination so that the helper can
736 set EXC_MASK, which contains a bitmask of destination registers
737 that have caused arithmetic traps. A simple userspace emulation
738 does not require this. We do need it for a guest kernel's entArith,
739 or if we were to do something clever with imprecise exceptions. */
740 reg = tcg_const_i32(rc + 32);
741 if (fn11 & QUAL_S) {
742 gen_helper_fp_exc_raise_s(cpu_env, ign, reg);
743 } else {
744 gen_helper_fp_exc_raise(cpu_env, ign, reg);
745 }
746
747 tcg_temp_free_i32(reg);
748 tcg_temp_free_i32(ign);
749 }
750
751 static void gen_cvtlq(TCGv vc, TCGv vb)
752 {
753 TCGv tmp = tcg_temp_new();
754
755 /* The arithmetic right shift here, plus the sign-extended mask below
756 yields a sign-extended result without an explicit ext32s_i64. */
757 tcg_gen_shri_i64(tmp, vb, 29);
758 tcg_gen_sari_i64(vc, vb, 32);
759 tcg_gen_deposit_i64(vc, vc, tmp, 0, 30);
760
761 tcg_temp_free(tmp);
762 }
763
764 static void gen_ieee_arith2(DisasContext *ctx,
765 void (*helper)(TCGv, TCGv_ptr, TCGv),
766 int rb, int rc, int fn11)
767 {
768 TCGv vb;
769
770 gen_qual_roundmode(ctx, fn11);
771 gen_qual_flushzero(ctx, fn11);
772
773 vb = gen_ieee_input(ctx, rb, fn11, 0);
774 helper(dest_fpr(ctx, rc), cpu_env, vb);
775
776 gen_fp_exc_raise(rc, fn11);
777 }
778
779 #define IEEE_ARITH2(name) \
780 static inline void glue(gen_, name)(DisasContext *ctx, \
781 int rb, int rc, int fn11) \
782 { \
783 gen_ieee_arith2(ctx, gen_helper_##name, rb, rc, fn11); \
784 }
785 IEEE_ARITH2(sqrts)
786 IEEE_ARITH2(sqrtt)
787 IEEE_ARITH2(cvtst)
788 IEEE_ARITH2(cvtts)
789
790 static void gen_cvttq(DisasContext *ctx, int rb, int rc, int fn11)
791 {
792 TCGv vb, vc;
793
794 /* No need to set flushzero, since we have an integer output. */
795 vb = gen_ieee_input(ctx, rb, fn11, 0);
796 vc = dest_fpr(ctx, rc);
797
798 /* Almost all integer conversions use cropped rounding;
799 special case that. */
800 if ((fn11 & QUAL_RM_MASK) == QUAL_RM_C) {
801 gen_helper_cvttq_c(vc, cpu_env, vb);
802 } else {
803 gen_qual_roundmode(ctx, fn11);
804 gen_helper_cvttq(vc, cpu_env, vb);
805 }
806 gen_fp_exc_raise(rc, fn11);
807 }
808
809 static void gen_ieee_intcvt(DisasContext *ctx,
810 void (*helper)(TCGv, TCGv_ptr, TCGv),
811 int rb, int rc, int fn11)
812 {
813 TCGv vb, vc;
814
815 gen_qual_roundmode(ctx, fn11);
816 vb = load_fpr(ctx, rb);
817 vc = dest_fpr(ctx, rc);
818
819 /* The only exception that can be raised by integer conversion
820 is inexact. Thus we only need to worry about exceptions when
821 inexact handling is requested. */
822 if (fn11 & QUAL_I) {
823 helper(vc, cpu_env, vb);
824 gen_fp_exc_raise(rc, fn11);
825 } else {
826 helper(vc, cpu_env, vb);
827 }
828 }
829
830 #define IEEE_INTCVT(name) \
831 static inline void glue(gen_, name)(DisasContext *ctx, \
832 int rb, int rc, int fn11) \
833 { \
834 gen_ieee_intcvt(ctx, gen_helper_##name, rb, rc, fn11); \
835 }
836 IEEE_INTCVT(cvtqs)
837 IEEE_INTCVT(cvtqt)
838
839 static void gen_cpy_mask(TCGv vc, TCGv va, TCGv vb, bool inv_a, uint64_t mask)
840 {
841 TCGv vmask = tcg_const_i64(mask);
842 TCGv tmp = tcg_temp_new_i64();
843
844 if (inv_a) {
845 tcg_gen_andc_i64(tmp, vmask, va);
846 } else {
847 tcg_gen_and_i64(tmp, va, vmask);
848 }
849
850 tcg_gen_andc_i64(vc, vb, vmask);
851 tcg_gen_or_i64(vc, vc, tmp);
852
853 tcg_temp_free(vmask);
854 tcg_temp_free(tmp);
855 }
856
857 static void gen_ieee_arith3(DisasContext *ctx,
858 void (*helper)(TCGv, TCGv_ptr, TCGv, TCGv),
859 int ra, int rb, int rc, int fn11)
860 {
861 TCGv va, vb, vc;
862
863 gen_qual_roundmode(ctx, fn11);
864 gen_qual_flushzero(ctx, fn11);
865
866 va = gen_ieee_input(ctx, ra, fn11, 0);
867 vb = gen_ieee_input(ctx, rb, fn11, 0);
868 vc = dest_fpr(ctx, rc);
869 helper(vc, cpu_env, va, vb);
870
871 gen_fp_exc_raise(rc, fn11);
872 }
873
874 #define IEEE_ARITH3(name) \
875 static inline void glue(gen_, name)(DisasContext *ctx, \
876 int ra, int rb, int rc, int fn11) \
877 { \
878 gen_ieee_arith3(ctx, gen_helper_##name, ra, rb, rc, fn11); \
879 }
880 IEEE_ARITH3(adds)
881 IEEE_ARITH3(subs)
882 IEEE_ARITH3(muls)
883 IEEE_ARITH3(divs)
884 IEEE_ARITH3(addt)
885 IEEE_ARITH3(subt)
886 IEEE_ARITH3(mult)
887 IEEE_ARITH3(divt)
888
889 static void gen_ieee_compare(DisasContext *ctx,
890 void (*helper)(TCGv, TCGv_ptr, TCGv, TCGv),
891 int ra, int rb, int rc, int fn11)
892 {
893 TCGv va, vb, vc;
894
895 va = gen_ieee_input(ctx, ra, fn11, 1);
896 vb = gen_ieee_input(ctx, rb, fn11, 1);
897 vc = dest_fpr(ctx, rc);
898 helper(vc, cpu_env, va, vb);
899
900 gen_fp_exc_raise(rc, fn11);
901 }
902
903 #define IEEE_CMP3(name) \
904 static inline void glue(gen_, name)(DisasContext *ctx, \
905 int ra, int rb, int rc, int fn11) \
906 { \
907 gen_ieee_compare(ctx, gen_helper_##name, ra, rb, rc, fn11); \
908 }
909 IEEE_CMP3(cmptun)
910 IEEE_CMP3(cmpteq)
911 IEEE_CMP3(cmptlt)
912 IEEE_CMP3(cmptle)
913
914 static inline uint64_t zapnot_mask(uint8_t lit)
915 {
916 uint64_t mask = 0;
917 int i;
918
919 for (i = 0; i < 8; ++i) {
920 if ((lit >> i) & 1) {
921 mask |= 0xffull << (i * 8);
922 }
923 }
924 return mask;
925 }
926
927 /* Implement zapnot with an immediate operand, which expands to some
928 form of immediate AND. This is a basic building block in the
929 definition of many of the other byte manipulation instructions. */
930 static void gen_zapnoti(TCGv dest, TCGv src, uint8_t lit)
931 {
932 switch (lit) {
933 case 0x00:
934 tcg_gen_movi_i64(dest, 0);
935 break;
936 case 0x01:
937 tcg_gen_ext8u_i64(dest, src);
938 break;
939 case 0x03:
940 tcg_gen_ext16u_i64(dest, src);
941 break;
942 case 0x0f:
943 tcg_gen_ext32u_i64(dest, src);
944 break;
945 case 0xff:
946 tcg_gen_mov_i64(dest, src);
947 break;
948 default:
949 tcg_gen_andi_i64(dest, src, zapnot_mask(lit));
950 break;
951 }
952 }
953
954 /* EXTWH, EXTLH, EXTQH */
955 static void gen_ext_h(DisasContext *ctx, TCGv vc, TCGv va, int rb, bool islit,
956 uint8_t lit, uint8_t byte_mask)
957 {
958 if (islit) {
959 int pos = (64 - lit * 8) & 0x3f;
960 int len = cto32(byte_mask) * 8;
961 if (pos < len) {
962 tcg_gen_deposit_z_i64(vc, va, pos, len - pos);
963 } else {
964 tcg_gen_movi_i64(vc, 0);
965 }
966 } else {
967 TCGv tmp = tcg_temp_new();
968 tcg_gen_shli_i64(tmp, load_gpr(ctx, rb), 3);
969 tcg_gen_neg_i64(tmp, tmp);
970 tcg_gen_andi_i64(tmp, tmp, 0x3f);
971 tcg_gen_shl_i64(vc, va, tmp);
972 tcg_temp_free(tmp);
973 }
974 gen_zapnoti(vc, vc, byte_mask);
975 }
976
977 /* EXTBL, EXTWL, EXTLL, EXTQL */
978 static void gen_ext_l(DisasContext *ctx, TCGv vc, TCGv va, int rb, bool islit,
979 uint8_t lit, uint8_t byte_mask)
980 {
981 if (islit) {
982 int pos = (lit & 7) * 8;
983 int len = cto32(byte_mask) * 8;
984 if (pos + len >= 64) {
985 len = 64 - pos;
986 }
987 tcg_gen_extract_i64(vc, va, pos, len);
988 } else {
989 TCGv tmp = tcg_temp_new();
990 tcg_gen_andi_i64(tmp, load_gpr(ctx, rb), 7);
991 tcg_gen_shli_i64(tmp, tmp, 3);
992 tcg_gen_shr_i64(vc, va, tmp);
993 tcg_temp_free(tmp);
994 gen_zapnoti(vc, vc, byte_mask);
995 }
996 }
997
998 /* INSWH, INSLH, INSQH */
999 static void gen_ins_h(DisasContext *ctx, TCGv vc, TCGv va, int rb, bool islit,
1000 uint8_t lit, uint8_t byte_mask)
1001 {
1002 if (islit) {
1003 int pos = 64 - (lit & 7) * 8;
1004 int len = cto32(byte_mask) * 8;
1005 if (pos < len) {
1006 tcg_gen_extract_i64(vc, va, pos, len - pos);
1007 } else {
1008 tcg_gen_movi_i64(vc, 0);
1009 }
1010 } else {
1011 TCGv tmp = tcg_temp_new();
1012 TCGv shift = tcg_temp_new();
1013
1014 /* The instruction description has us left-shift the byte mask
1015 and extract bits <15:8> and apply that zap at the end. This
1016 is equivalent to simply performing the zap first and shifting
1017 afterward. */
1018 gen_zapnoti(tmp, va, byte_mask);
1019
1020 /* If (B & 7) == 0, we need to shift by 64 and leave a zero. Do this
1021 portably by splitting the shift into two parts: shift_count-1 and 1.
1022 Arrange for the -1 by using ones-complement instead of
1023 twos-complement in the negation: ~(B * 8) & 63. */
1024
1025 tcg_gen_shli_i64(shift, load_gpr(ctx, rb), 3);
1026 tcg_gen_not_i64(shift, shift);
1027 tcg_gen_andi_i64(shift, shift, 0x3f);
1028
1029 tcg_gen_shr_i64(vc, tmp, shift);
1030 tcg_gen_shri_i64(vc, vc, 1);
1031 tcg_temp_free(shift);
1032 tcg_temp_free(tmp);
1033 }
1034 }
1035
1036 /* INSBL, INSWL, INSLL, INSQL */
1037 static void gen_ins_l(DisasContext *ctx, TCGv vc, TCGv va, int rb, bool islit,
1038 uint8_t lit, uint8_t byte_mask)
1039 {
1040 if (islit) {
1041 int pos = (lit & 7) * 8;
1042 int len = cto32(byte_mask) * 8;
1043 if (pos + len > 64) {
1044 len = 64 - pos;
1045 }
1046 tcg_gen_deposit_z_i64(vc, va, pos, len);
1047 } else {
1048 TCGv tmp = tcg_temp_new();
1049 TCGv shift = tcg_temp_new();
1050
1051 /* The instruction description has us left-shift the byte mask
1052 and extract bits <15:8> and apply that zap at the end. This
1053 is equivalent to simply performing the zap first and shifting
1054 afterward. */
1055 gen_zapnoti(tmp, va, byte_mask);
1056
1057 tcg_gen_andi_i64(shift, load_gpr(ctx, rb), 7);
1058 tcg_gen_shli_i64(shift, shift, 3);
1059 tcg_gen_shl_i64(vc, tmp, shift);
1060 tcg_temp_free(shift);
1061 tcg_temp_free(tmp);
1062 }
1063 }
1064
1065 /* MSKWH, MSKLH, MSKQH */
1066 static void gen_msk_h(DisasContext *ctx, TCGv vc, TCGv va, int rb, bool islit,
1067 uint8_t lit, uint8_t byte_mask)
1068 {
1069 if (islit) {
1070 gen_zapnoti(vc, va, ~((byte_mask << (lit & 7)) >> 8));
1071 } else {
1072 TCGv shift = tcg_temp_new();
1073 TCGv mask = tcg_temp_new();
1074
1075 /* The instruction description is as above, where the byte_mask
1076 is shifted left, and then we extract bits <15:8>. This can be
1077 emulated with a right-shift on the expanded byte mask. This
1078 requires extra care because for an input <2:0> == 0 we need a
1079 shift of 64 bits in order to generate a zero. This is done by
1080 splitting the shift into two parts, the variable shift - 1
1081 followed by a constant 1 shift. The code we expand below is
1082 equivalent to ~(B * 8) & 63. */
1083
1084 tcg_gen_shli_i64(shift, load_gpr(ctx, rb), 3);
1085 tcg_gen_not_i64(shift, shift);
1086 tcg_gen_andi_i64(shift, shift, 0x3f);
1087 tcg_gen_movi_i64(mask, zapnot_mask (byte_mask));
1088 tcg_gen_shr_i64(mask, mask, shift);
1089 tcg_gen_shri_i64(mask, mask, 1);
1090
1091 tcg_gen_andc_i64(vc, va, mask);
1092
1093 tcg_temp_free(mask);
1094 tcg_temp_free(shift);
1095 }
1096 }
1097
1098 /* MSKBL, MSKWL, MSKLL, MSKQL */
1099 static void gen_msk_l(DisasContext *ctx, TCGv vc, TCGv va, int rb, bool islit,
1100 uint8_t lit, uint8_t byte_mask)
1101 {
1102 if (islit) {
1103 gen_zapnoti(vc, va, ~(byte_mask << (lit & 7)));
1104 } else {
1105 TCGv shift = tcg_temp_new();
1106 TCGv mask = tcg_temp_new();
1107
1108 tcg_gen_andi_i64(shift, load_gpr(ctx, rb), 7);
1109 tcg_gen_shli_i64(shift, shift, 3);
1110 tcg_gen_movi_i64(mask, zapnot_mask(byte_mask));
1111 tcg_gen_shl_i64(mask, mask, shift);
1112
1113 tcg_gen_andc_i64(vc, va, mask);
1114
1115 tcg_temp_free(mask);
1116 tcg_temp_free(shift);
1117 }
1118 }
1119
1120 static void gen_rx(DisasContext *ctx, int ra, int set)
1121 {
1122 TCGv tmp;
1123
1124 if (ra != 31) {
1125 ld_flag_byte(ctx->ir[ra], ENV_FLAG_RX_SHIFT);
1126 }
1127
1128 tmp = tcg_const_i64(set);
1129 st_flag_byte(ctx->ir[ra], ENV_FLAG_RX_SHIFT);
1130 tcg_temp_free(tmp);
1131 }
1132
1133 static DisasJumpType gen_call_pal(DisasContext *ctx, int palcode)
1134 {
1135 /* We're emulating OSF/1 PALcode. Many of these are trivial access
1136 to internal cpu registers. */
1137
1138 /* Unprivileged PAL call */
1139 if (palcode >= 0x80 && palcode < 0xC0) {
1140 switch (palcode) {
1141 case 0x86:
1142 /* IMB */
1143 /* No-op inside QEMU. */
1144 break;
1145 case 0x9E:
1146 /* RDUNIQUE */
1147 tcg_gen_ld_i64(ctx->ir[IR_V0], cpu_env,
1148 offsetof(CPUAlphaState, unique));
1149 break;
1150 case 0x9F:
1151 /* WRUNIQUE */
1152 tcg_gen_st_i64(ctx->ir[IR_A0], cpu_env,
1153 offsetof(CPUAlphaState, unique));
1154 break;
1155 default:
1156 palcode &= 0xbf;
1157 goto do_call_pal;
1158 }
1159 return DISAS_NEXT;
1160 }
1161
1162 #ifndef CONFIG_USER_ONLY
1163 /* Privileged PAL code */
1164 if (palcode < 0x40 && (ctx->tbflags & ENV_FLAG_PS_USER) == 0) {
1165 switch (palcode) {
1166 case 0x01:
1167 /* CFLUSH */
1168 /* No-op inside QEMU. */
1169 break;
1170 case 0x02:
1171 /* DRAINA */
1172 /* No-op inside QEMU. */
1173 break;
1174 case 0x2D:
1175 /* WRVPTPTR */
1176 tcg_gen_st_i64(ctx->ir[IR_A0], cpu_env,
1177 offsetof(CPUAlphaState, vptptr));
1178 break;
1179 case 0x31:
1180 /* WRVAL */
1181 tcg_gen_st_i64(ctx->ir[IR_A0], cpu_env,
1182 offsetof(CPUAlphaState, sysval));
1183 break;
1184 case 0x32:
1185 /* RDVAL */
1186 tcg_gen_ld_i64(ctx->ir[IR_V0], cpu_env,
1187 offsetof(CPUAlphaState, sysval));
1188 break;
1189
1190 case 0x35:
1191 /* SWPIPL */
1192 /* Note that we already know we're in kernel mode, so we know
1193 that PS only contains the 3 IPL bits. */
1194 ld_flag_byte(ctx->ir[IR_V0], ENV_FLAG_PS_SHIFT);
1195
1196 /* But make sure and store only the 3 IPL bits from the user. */
1197 {
1198 TCGv tmp = tcg_temp_new();
1199 tcg_gen_andi_i64(tmp, ctx->ir[IR_A0], PS_INT_MASK);
1200 st_flag_byte(tmp, ENV_FLAG_PS_SHIFT);
1201 tcg_temp_free(tmp);
1202 }
1203
1204 /* Allow interrupts to be recognized right away. */
1205 tcg_gen_movi_i64(cpu_pc, ctx->base.pc_next);
1206 return DISAS_PC_UPDATED_NOCHAIN;
1207
1208 case 0x36:
1209 /* RDPS */
1210 ld_flag_byte(ctx->ir[IR_V0], ENV_FLAG_PS_SHIFT);
1211 break;
1212
1213 case 0x38:
1214 /* WRUSP */
1215 tcg_gen_st_i64(ctx->ir[IR_A0], cpu_env,
1216 offsetof(CPUAlphaState, usp));
1217 break;
1218 case 0x3A:
1219 /* RDUSP */
1220 tcg_gen_ld_i64(ctx->ir[IR_V0], cpu_env,
1221 offsetof(CPUAlphaState, usp));
1222 break;
1223 case 0x3C:
1224 /* WHAMI */
1225 tcg_gen_ld32s_i64(ctx->ir[IR_V0], cpu_env,
1226 -offsetof(AlphaCPU, env) + offsetof(CPUState, cpu_index));
1227 break;
1228
1229 case 0x3E:
1230 /* WTINT */
1231 {
1232 TCGv_i32 tmp = tcg_const_i32(1);
1233 tcg_gen_st_i32(tmp, cpu_env, -offsetof(AlphaCPU, env) +
1234 offsetof(CPUState, halted));
1235 tcg_temp_free_i32(tmp);
1236 }
1237 tcg_gen_movi_i64(ctx->ir[IR_V0], 0);
1238 return gen_excp(ctx, EXCP_HALTED, 0);
1239
1240 default:
1241 palcode &= 0x3f;
1242 goto do_call_pal;
1243 }
1244 return DISAS_NEXT;
1245 }
1246 #endif
1247 return gen_invalid(ctx);
1248
1249 do_call_pal:
1250 #ifdef CONFIG_USER_ONLY
1251 return gen_excp(ctx, EXCP_CALL_PAL, palcode);
1252 #else
1253 {
1254 TCGv tmp = tcg_temp_new();
1255 uint64_t exc_addr = ctx->base.pc_next;
1256 uint64_t entry = ctx->palbr;
1257
1258 if (ctx->tbflags & ENV_FLAG_PAL_MODE) {
1259 exc_addr |= 1;
1260 } else {
1261 tcg_gen_movi_i64(tmp, 1);
1262 st_flag_byte(tmp, ENV_FLAG_PAL_SHIFT);
1263 }
1264
1265 tcg_gen_movi_i64(tmp, exc_addr);
1266 tcg_gen_st_i64(tmp, cpu_env, offsetof(CPUAlphaState, exc_addr));
1267 tcg_temp_free(tmp);
1268
1269 entry += (palcode & 0x80
1270 ? 0x2000 + (palcode - 0x80) * 64
1271 : 0x1000 + palcode * 64);
1272
1273 /* Since the destination is running in PALmode, we don't really
1274 need the page permissions check. We'll see the existence of
1275 the page when we create the TB, and we'll flush all TBs if
1276 we change the PAL base register. */
1277 if (!use_exit_tb(ctx)) {
1278 tcg_gen_goto_tb(0);
1279 tcg_gen_movi_i64(cpu_pc, entry);
1280 tcg_gen_exit_tb((uintptr_t)ctx->base.tb);
1281 return DISAS_NORETURN;
1282 } else {
1283 tcg_gen_movi_i64(cpu_pc, entry);
1284 return DISAS_PC_UPDATED;
1285 }
1286 }
1287 #endif
1288 }
1289
1290 #ifndef CONFIG_USER_ONLY
1291
1292 #define PR_LONG 0x200000
1293
1294 static int cpu_pr_data(int pr)
1295 {
1296 switch (pr) {
1297 case 2: return offsetof(CPUAlphaState, pcc_ofs) | PR_LONG;
1298 case 3: return offsetof(CPUAlphaState, trap_arg0);
1299 case 4: return offsetof(CPUAlphaState, trap_arg1);
1300 case 5: return offsetof(CPUAlphaState, trap_arg2);
1301 case 6: return offsetof(CPUAlphaState, exc_addr);
1302 case 7: return offsetof(CPUAlphaState, palbr);
1303 case 8: return offsetof(CPUAlphaState, ptbr);
1304 case 9: return offsetof(CPUAlphaState, vptptr);
1305 case 10: return offsetof(CPUAlphaState, unique);
1306 case 11: return offsetof(CPUAlphaState, sysval);
1307 case 12: return offsetof(CPUAlphaState, usp);
1308
1309 case 40 ... 63:
1310 return offsetof(CPUAlphaState, scratch[pr - 40]);
1311
1312 case 251:
1313 return offsetof(CPUAlphaState, alarm_expire);
1314 }
1315 return 0;
1316 }
1317
1318 static DisasJumpType gen_mfpr(DisasContext *ctx, TCGv va, int regno)
1319 {
1320 void (*helper)(TCGv);
1321 int data;
1322
1323 switch (regno) {
1324 case 32 ... 39:
1325 /* Accessing the "non-shadow" general registers. */
1326 regno = regno == 39 ? 25 : regno - 32 + 8;
1327 tcg_gen_mov_i64(va, cpu_std_ir[regno]);
1328 break;
1329
1330 case 250: /* WALLTIME */
1331 helper = gen_helper_get_walltime;
1332 goto do_helper;
1333 case 249: /* VMTIME */
1334 helper = gen_helper_get_vmtime;
1335 do_helper:
1336 if (use_icount) {
1337 gen_io_start();
1338 helper(va);
1339 gen_io_end();
1340 return DISAS_PC_STALE;
1341 } else {
1342 helper(va);
1343 }
1344 break;
1345
1346 case 0: /* PS */
1347 ld_flag_byte(va, ENV_FLAG_PS_SHIFT);
1348 break;
1349 case 1: /* FEN */
1350 ld_flag_byte(va, ENV_FLAG_FEN_SHIFT);
1351 break;
1352
1353 default:
1354 /* The basic registers are data only, and unknown registers
1355 are read-zero, write-ignore. */
1356 data = cpu_pr_data(regno);
1357 if (data == 0) {
1358 tcg_gen_movi_i64(va, 0);
1359 } else if (data & PR_LONG) {
1360 tcg_gen_ld32s_i64(va, cpu_env, data & ~PR_LONG);
1361 } else {
1362 tcg_gen_ld_i64(va, cpu_env, data);
1363 }
1364 break;
1365 }
1366
1367 return DISAS_NEXT;
1368 }
1369
1370 static DisasJumpType gen_mtpr(DisasContext *ctx, TCGv vb, int regno)
1371 {
1372 int data;
1373
1374 switch (regno) {
1375 case 255:
1376 /* TBIA */
1377 gen_helper_tbia(cpu_env);
1378 break;
1379
1380 case 254:
1381 /* TBIS */
1382 gen_helper_tbis(cpu_env, vb);
1383 break;
1384
1385 case 253:
1386 /* WAIT */
1387 {
1388 TCGv_i32 tmp = tcg_const_i32(1);
1389 tcg_gen_st_i32(tmp, cpu_env, -offsetof(AlphaCPU, env) +
1390 offsetof(CPUState, halted));
1391 tcg_temp_free_i32(tmp);
1392 }
1393 return gen_excp(ctx, EXCP_HALTED, 0);
1394
1395 case 252:
1396 /* HALT */
1397 gen_helper_halt(vb);
1398 return DISAS_PC_STALE;
1399
1400 case 251:
1401 /* ALARM */
1402 gen_helper_set_alarm(cpu_env, vb);
1403 break;
1404
1405 case 7:
1406 /* PALBR */
1407 tcg_gen_st_i64(vb, cpu_env, offsetof(CPUAlphaState, palbr));
1408 /* Changing the PAL base register implies un-chaining all of the TBs
1409 that ended with a CALL_PAL. Since the base register usually only
1410 changes during boot, flushing everything works well. */
1411 gen_helper_tb_flush(cpu_env);
1412 return DISAS_PC_STALE;
1413
1414 case 32 ... 39:
1415 /* Accessing the "non-shadow" general registers. */
1416 regno = regno == 39 ? 25 : regno - 32 + 8;
1417 tcg_gen_mov_i64(cpu_std_ir[regno], vb);
1418 break;
1419
1420 case 0: /* PS */
1421 st_flag_byte(vb, ENV_FLAG_PS_SHIFT);
1422 break;
1423 case 1: /* FEN */
1424 st_flag_byte(vb, ENV_FLAG_FEN_SHIFT);
1425 break;
1426
1427 default:
1428 /* The basic registers are data only, and unknown registers
1429 are read-zero, write-ignore. */
1430 data = cpu_pr_data(regno);
1431 if (data != 0) {
1432 if (data & PR_LONG) {
1433 tcg_gen_st32_i64(vb, cpu_env, data & ~PR_LONG);
1434 } else {
1435 tcg_gen_st_i64(vb, cpu_env, data);
1436 }
1437 }
1438 break;
1439 }
1440
1441 return DISAS_NEXT;
1442 }
1443 #endif /* !USER_ONLY*/
1444
1445 #define REQUIRE_NO_LIT \
1446 do { \
1447 if (real_islit) { \
1448 goto invalid_opc; \
1449 } \
1450 } while (0)
1451
1452 #define REQUIRE_AMASK(FLAG) \
1453 do { \
1454 if ((ctx->amask & AMASK_##FLAG) == 0) { \
1455 goto invalid_opc; \
1456 } \
1457 } while (0)
1458
1459 #define REQUIRE_TB_FLAG(FLAG) \
1460 do { \
1461 if ((ctx->tbflags & (FLAG)) == 0) { \
1462 goto invalid_opc; \
1463 } \
1464 } while (0)
1465
1466 #define REQUIRE_REG_31(WHICH) \
1467 do { \
1468 if (WHICH != 31) { \
1469 goto invalid_opc; \
1470 } \
1471 } while (0)
1472
1473 static DisasJumpType translate_one(DisasContext *ctx, uint32_t insn)
1474 {
1475 int32_t disp21, disp16, disp12 __attribute__((unused));
1476 uint16_t fn11;
1477 uint8_t opc, ra, rb, rc, fpfn, fn7, lit;
1478 bool islit, real_islit;
1479 TCGv va, vb, vc, tmp, tmp2;
1480 TCGv_i32 t32;
1481 DisasJumpType ret;
1482
1483 /* Decode all instruction fields */
1484 opc = extract32(insn, 26, 6);
1485 ra = extract32(insn, 21, 5);
1486 rb = extract32(insn, 16, 5);
1487 rc = extract32(insn, 0, 5);
1488 real_islit = islit = extract32(insn, 12, 1);
1489 lit = extract32(insn, 13, 8);
1490
1491 disp21 = sextract32(insn, 0, 21);
1492 disp16 = sextract32(insn, 0, 16);
1493 disp12 = sextract32(insn, 0, 12);
1494
1495 fn11 = extract32(insn, 5, 11);
1496 fpfn = extract32(insn, 5, 6);
1497 fn7 = extract32(insn, 5, 7);
1498
1499 if (rb == 31 && !islit) {
1500 islit = true;
1501 lit = 0;
1502 }
1503
1504 ret = DISAS_NEXT;
1505 switch (opc) {
1506 case 0x00:
1507 /* CALL_PAL */
1508 ret = gen_call_pal(ctx, insn & 0x03ffffff);
1509 break;
1510 case 0x01:
1511 /* OPC01 */
1512 goto invalid_opc;
1513 case 0x02:
1514 /* OPC02 */
1515 goto invalid_opc;
1516 case 0x03:
1517 /* OPC03 */
1518 goto invalid_opc;
1519 case 0x04:
1520 /* OPC04 */
1521 goto invalid_opc;
1522 case 0x05:
1523 /* OPC05 */
1524 goto invalid_opc;
1525 case 0x06:
1526 /* OPC06 */
1527 goto invalid_opc;
1528 case 0x07:
1529 /* OPC07 */
1530 goto invalid_opc;
1531
1532 case 0x09:
1533 /* LDAH */
1534 disp16 = (uint32_t)disp16 << 16;
1535 /* fall through */
1536 case 0x08:
1537 /* LDA */
1538 va = dest_gpr(ctx, ra);
1539 /* It's worth special-casing immediate loads. */
1540 if (rb == 31) {
1541 tcg_gen_movi_i64(va, disp16);
1542 } else {
1543 tcg_gen_addi_i64(va, load_gpr(ctx, rb), disp16);
1544 }
1545 break;
1546
1547 case 0x0A:
1548 /* LDBU */
1549 REQUIRE_AMASK(BWX);
1550 gen_load_mem(ctx, &tcg_gen_qemu_ld8u, ra, rb, disp16, 0, 0);
1551 break;
1552 case 0x0B:
1553 /* LDQ_U */
1554 gen_load_mem(ctx, &tcg_gen_qemu_ld64, ra, rb, disp16, 0, 1);
1555 break;
1556 case 0x0C:
1557 /* LDWU */
1558 REQUIRE_AMASK(BWX);
1559 gen_load_mem(ctx, &tcg_gen_qemu_ld16u, ra, rb, disp16, 0, 0);
1560 break;
1561 case 0x0D:
1562 /* STW */
1563 REQUIRE_AMASK(BWX);
1564 gen_store_mem(ctx, &tcg_gen_qemu_st16, ra, rb, disp16, 0, 0);
1565 break;
1566 case 0x0E:
1567 /* STB */
1568 REQUIRE_AMASK(BWX);
1569 gen_store_mem(ctx, &tcg_gen_qemu_st8, ra, rb, disp16, 0, 0);
1570 break;
1571 case 0x0F:
1572 /* STQ_U */
1573 gen_store_mem(ctx, &tcg_gen_qemu_st64, ra, rb, disp16, 0, 1);
1574 break;
1575
1576 case 0x10:
1577 vc = dest_gpr(ctx, rc);
1578 vb = load_gpr_lit(ctx, rb, lit, islit);
1579
1580 if (ra == 31) {
1581 if (fn7 == 0x00) {
1582 /* Special case ADDL as SEXTL. */
1583 tcg_gen_ext32s_i64(vc, vb);
1584 break;
1585 }
1586 if (fn7 == 0x29) {
1587 /* Special case SUBQ as NEGQ. */
1588 tcg_gen_neg_i64(vc, vb);
1589 break;
1590 }
1591 }
1592
1593 va = load_gpr(ctx, ra);
1594 switch (fn7) {
1595 case 0x00:
1596 /* ADDL */
1597 tcg_gen_add_i64(vc, va, vb);
1598 tcg_gen_ext32s_i64(vc, vc);
1599 break;
1600 case 0x02:
1601 /* S4ADDL */
1602 tmp = tcg_temp_new();
1603 tcg_gen_shli_i64(tmp, va, 2);
1604 tcg_gen_add_i64(tmp, tmp, vb);
1605 tcg_gen_ext32s_i64(vc, tmp);
1606 tcg_temp_free(tmp);
1607 break;
1608 case 0x09:
1609 /* SUBL */
1610 tcg_gen_sub_i64(vc, va, vb);
1611 tcg_gen_ext32s_i64(vc, vc);
1612 break;
1613 case 0x0B:
1614 /* S4SUBL */
1615 tmp = tcg_temp_new();
1616 tcg_gen_shli_i64(tmp, va, 2);
1617 tcg_gen_sub_i64(tmp, tmp, vb);
1618 tcg_gen_ext32s_i64(vc, tmp);
1619 tcg_temp_free(tmp);
1620 break;
1621 case 0x0F:
1622 /* CMPBGE */
1623 if (ra == 31) {
1624 /* Special case 0 >= X as X == 0. */
1625 gen_helper_cmpbe0(vc, vb);
1626 } else {
1627 gen_helper_cmpbge(vc, va, vb);
1628 }
1629 break;
1630 case 0x12:
1631 /* S8ADDL */
1632 tmp = tcg_temp_new();
1633 tcg_gen_shli_i64(tmp, va, 3);
1634 tcg_gen_add_i64(tmp, tmp, vb);
1635 tcg_gen_ext32s_i64(vc, tmp);
1636 tcg_temp_free(tmp);
1637 break;
1638 case 0x1B:
1639 /* S8SUBL */
1640 tmp = tcg_temp_new();
1641 tcg_gen_shli_i64(tmp, va, 3);
1642 tcg_gen_sub_i64(tmp, tmp, vb);
1643 tcg_gen_ext32s_i64(vc, tmp);
1644 tcg_temp_free(tmp);
1645 break;
1646 case 0x1D:
1647 /* CMPULT */
1648 tcg_gen_setcond_i64(TCG_COND_LTU, vc, va, vb);
1649 break;
1650 case 0x20:
1651 /* ADDQ */
1652 tcg_gen_add_i64(vc, va, vb);
1653 break;
1654 case 0x22:
1655 /* S4ADDQ */
1656 tmp = tcg_temp_new();
1657 tcg_gen_shli_i64(tmp, va, 2);
1658 tcg_gen_add_i64(vc, tmp, vb);
1659 tcg_temp_free(tmp);
1660 break;
1661 case 0x29:
1662 /* SUBQ */
1663 tcg_gen_sub_i64(vc, va, vb);
1664 break;
1665 case 0x2B:
1666 /* S4SUBQ */
1667 tmp = tcg_temp_new();
1668 tcg_gen_shli_i64(tmp, va, 2);
1669 tcg_gen_sub_i64(vc, tmp, vb);
1670 tcg_temp_free(tmp);
1671 break;
1672 case 0x2D:
1673 /* CMPEQ */
1674 tcg_gen_setcond_i64(TCG_COND_EQ, vc, va, vb);
1675 break;
1676 case 0x32:
1677 /* S8ADDQ */
1678 tmp = tcg_temp_new();
1679 tcg_gen_shli_i64(tmp, va, 3);
1680 tcg_gen_add_i64(vc, tmp, vb);
1681 tcg_temp_free(tmp);
1682 break;
1683 case 0x3B:
1684 /* S8SUBQ */
1685 tmp = tcg_temp_new();
1686 tcg_gen_shli_i64(tmp, va, 3);
1687 tcg_gen_sub_i64(vc, tmp, vb);
1688 tcg_temp_free(tmp);
1689 break;
1690 case 0x3D:
1691 /* CMPULE */
1692 tcg_gen_setcond_i64(TCG_COND_LEU, vc, va, vb);
1693 break;
1694 case 0x40:
1695 /* ADDL/V */
1696 tmp = tcg_temp_new();
1697 tcg_gen_ext32s_i64(tmp, va);
1698 tcg_gen_ext32s_i64(vc, vb);
1699 tcg_gen_add_i64(tmp, tmp, vc);
1700 tcg_gen_ext32s_i64(vc, tmp);
1701 gen_helper_check_overflow(cpu_env, vc, tmp);
1702 tcg_temp_free(tmp);
1703 break;
1704 case 0x49:
1705 /* SUBL/V */
1706 tmp = tcg_temp_new();
1707 tcg_gen_ext32s_i64(tmp, va);
1708 tcg_gen_ext32s_i64(vc, vb);
1709 tcg_gen_sub_i64(tmp, tmp, vc);
1710 tcg_gen_ext32s_i64(vc, tmp);
1711 gen_helper_check_overflow(cpu_env, vc, tmp);
1712 tcg_temp_free(tmp);
1713 break;
1714 case 0x4D:
1715 /* CMPLT */
1716 tcg_gen_setcond_i64(TCG_COND_LT, vc, va, vb);
1717 break;
1718 case 0x60:
1719 /* ADDQ/V */
1720 tmp = tcg_temp_new();
1721 tmp2 = tcg_temp_new();
1722 tcg_gen_eqv_i64(tmp, va, vb);
1723 tcg_gen_mov_i64(tmp2, va);
1724 tcg_gen_add_i64(vc, va, vb);
1725 tcg_gen_xor_i64(tmp2, tmp2, vc);
1726 tcg_gen_and_i64(tmp, tmp, tmp2);
1727 tcg_gen_shri_i64(tmp, tmp, 63);
1728 tcg_gen_movi_i64(tmp2, 0);
1729 gen_helper_check_overflow(cpu_env, tmp, tmp2);
1730 tcg_temp_free(tmp);
1731 tcg_temp_free(tmp2);
1732 break;
1733 case 0x69:
1734 /* SUBQ/V */
1735 tmp = tcg_temp_new();
1736 tmp2 = tcg_temp_new();
1737 tcg_gen_xor_i64(tmp, va, vb);
1738 tcg_gen_mov_i64(tmp2, va);
1739 tcg_gen_sub_i64(vc, va, vb);
1740 tcg_gen_xor_i64(tmp2, tmp2, vc);
1741 tcg_gen_and_i64(tmp, tmp, tmp2);
1742 tcg_gen_shri_i64(tmp, tmp, 63);
1743 tcg_gen_movi_i64(tmp2, 0);
1744 gen_helper_check_overflow(cpu_env, tmp, tmp2);
1745 tcg_temp_free(tmp);
1746 tcg_temp_free(tmp2);
1747 break;
1748 case 0x6D:
1749 /* CMPLE */
1750 tcg_gen_setcond_i64(TCG_COND_LE, vc, va, vb);
1751 break;
1752 default:
1753 goto invalid_opc;
1754 }
1755 break;
1756
1757 case 0x11:
1758 if (fn7 == 0x20) {
1759 if (rc == 31) {
1760 /* Special case BIS as NOP. */
1761 break;
1762 }
1763 if (ra == 31) {
1764 /* Special case BIS as MOV. */
1765 vc = dest_gpr(ctx, rc);
1766 if (islit) {
1767 tcg_gen_movi_i64(vc, lit);
1768 } else {
1769 tcg_gen_mov_i64(vc, load_gpr(ctx, rb));
1770 }
1771 break;
1772 }
1773 }
1774
1775 vc = dest_gpr(ctx, rc);
1776 vb = load_gpr_lit(ctx, rb, lit, islit);
1777
1778 if (fn7 == 0x28 && ra == 31) {
1779 /* Special case ORNOT as NOT. */
1780 tcg_gen_not_i64(vc, vb);
1781 break;
1782 }
1783
1784 va = load_gpr(ctx, ra);
1785 switch (fn7) {
1786 case 0x00:
1787 /* AND */
1788 tcg_gen_and_i64(vc, va, vb);
1789 break;
1790 case 0x08:
1791 /* BIC */
1792 tcg_gen_andc_i64(vc, va, vb);
1793 break;
1794 case 0x14:
1795 /* CMOVLBS */
1796 tmp = tcg_temp_new();
1797 tcg_gen_andi_i64(tmp, va, 1);
1798 tcg_gen_movcond_i64(TCG_COND_NE, vc, tmp, load_zero(ctx),
1799 vb, load_gpr(ctx, rc));
1800 tcg_temp_free(tmp);
1801 break;
1802 case 0x16:
1803 /* CMOVLBC */
1804 tmp = tcg_temp_new();
1805 tcg_gen_andi_i64(tmp, va, 1);
1806 tcg_gen_movcond_i64(TCG_COND_EQ, vc, tmp, load_zero(ctx),
1807 vb, load_gpr(ctx, rc));
1808 tcg_temp_free(tmp);
1809 break;
1810 case 0x20:
1811 /* BIS */
1812 tcg_gen_or_i64(vc, va, vb);
1813 break;
1814 case 0x24:
1815 /* CMOVEQ */
1816 tcg_gen_movcond_i64(TCG_COND_EQ, vc, va, load_zero(ctx),
1817 vb, load_gpr(ctx, rc));
1818 break;
1819 case 0x26:
1820 /* CMOVNE */
1821 tcg_gen_movcond_i64(TCG_COND_NE, vc, va, load_zero(ctx),
1822 vb, load_gpr(ctx, rc));
1823 break;
1824 case 0x28:
1825 /* ORNOT */
1826 tcg_gen_orc_i64(vc, va, vb);
1827 break;
1828 case 0x40:
1829 /* XOR */
1830 tcg_gen_xor_i64(vc, va, vb);
1831 break;
1832 case 0x44:
1833 /* CMOVLT */
1834 tcg_gen_movcond_i64(TCG_COND_LT, vc, va, load_zero(ctx),
1835 vb, load_gpr(ctx, rc));
1836 break;
1837 case 0x46:
1838 /* CMOVGE */
1839 tcg_gen_movcond_i64(TCG_COND_GE, vc, va, load_zero(ctx),
1840 vb, load_gpr(ctx, rc));
1841 break;
1842 case 0x48:
1843 /* EQV */
1844 tcg_gen_eqv_i64(vc, va, vb);
1845 break;
1846 case 0x61:
1847 /* AMASK */
1848 REQUIRE_REG_31(ra);
1849 tcg_gen_andi_i64(vc, vb, ~ctx->amask);
1850 break;
1851 case 0x64:
1852 /* CMOVLE */
1853 tcg_gen_movcond_i64(TCG_COND_LE, vc, va, load_zero(ctx),
1854 vb, load_gpr(ctx, rc));
1855 break;
1856 case 0x66:
1857 /* CMOVGT */
1858 tcg_gen_movcond_i64(TCG_COND_GT, vc, va, load_zero(ctx),
1859 vb, load_gpr(ctx, rc));
1860 break;
1861 case 0x6C:
1862 /* IMPLVER */
1863 REQUIRE_REG_31(ra);
1864 tcg_gen_movi_i64(vc, ctx->implver);
1865 break;
1866 default:
1867 goto invalid_opc;
1868 }
1869 break;
1870
1871 case 0x12:
1872 vc = dest_gpr(ctx, rc);
1873 va = load_gpr(ctx, ra);
1874 switch (fn7) {
1875 case 0x02:
1876 /* MSKBL */
1877 gen_msk_l(ctx, vc, va, rb, islit, lit, 0x01);
1878 break;
1879 case 0x06:
1880 /* EXTBL */
1881 gen_ext_l(ctx, vc, va, rb, islit, lit, 0x01);
1882 break;
1883 case 0x0B:
1884 /* INSBL */
1885 gen_ins_l(ctx, vc, va, rb, islit, lit, 0x01);
1886 break;
1887 case 0x12:
1888 /* MSKWL */
1889 gen_msk_l(ctx, vc, va, rb, islit, lit, 0x03);
1890 break;
1891 case 0x16:
1892 /* EXTWL */
1893 gen_ext_l(ctx, vc, va, rb, islit, lit, 0x03);
1894 break;
1895 case 0x1B:
1896 /* INSWL */
1897 gen_ins_l(ctx, vc, va, rb, islit, lit, 0x03);
1898 break;
1899 case 0x22:
1900 /* MSKLL */
1901 gen_msk_l(ctx, vc, va, rb, islit, lit, 0x0f);
1902 break;
1903 case 0x26:
1904 /* EXTLL */
1905 gen_ext_l(ctx, vc, va, rb, islit, lit, 0x0f);
1906 break;
1907 case 0x2B:
1908 /* INSLL */
1909 gen_ins_l(ctx, vc, va, rb, islit, lit, 0x0f);
1910 break;
1911 case 0x30:
1912 /* ZAP */
1913 if (islit) {
1914 gen_zapnoti(vc, va, ~lit);
1915 } else {
1916 gen_helper_zap(vc, va, load_gpr(ctx, rb));
1917 }
1918 break;
1919 case 0x31:
1920 /* ZAPNOT */
1921 if (islit) {
1922 gen_zapnoti(vc, va, lit);
1923 } else {
1924 gen_helper_zapnot(vc, va, load_gpr(ctx, rb));
1925 }
1926 break;
1927 case 0x32:
1928 /* MSKQL */
1929 gen_msk_l(ctx, vc, va, rb, islit, lit, 0xff);
1930 break;
1931 case 0x34:
1932 /* SRL */
1933 if (islit) {
1934 tcg_gen_shri_i64(vc, va, lit & 0x3f);
1935 } else {
1936 tmp = tcg_temp_new();
1937 vb = load_gpr(ctx, rb);
1938 tcg_gen_andi_i64(tmp, vb, 0x3f);
1939 tcg_gen_shr_i64(vc, va, tmp);
1940 tcg_temp_free(tmp);
1941 }
1942 break;
1943 case 0x36:
1944 /* EXTQL */
1945 gen_ext_l(ctx, vc, va, rb, islit, lit, 0xff);
1946 break;
1947 case 0x39:
1948 /* SLL */
1949 if (islit) {
1950 tcg_gen_shli_i64(vc, va, lit & 0x3f);
1951 } else {
1952 tmp = tcg_temp_new();
1953 vb = load_gpr(ctx, rb);
1954 tcg_gen_andi_i64(tmp, vb, 0x3f);
1955 tcg_gen_shl_i64(vc, va, tmp);
1956 tcg_temp_free(tmp);
1957 }
1958 break;
1959 case 0x3B:
1960 /* INSQL */
1961 gen_ins_l(ctx, vc, va, rb, islit, lit, 0xff);
1962 break;
1963 case 0x3C:
1964 /* SRA */
1965 if (islit) {
1966 tcg_gen_sari_i64(vc, va, lit & 0x3f);
1967 } else {
1968 tmp = tcg_temp_new();
1969 vb = load_gpr(ctx, rb);
1970 tcg_gen_andi_i64(tmp, vb, 0x3f);
1971 tcg_gen_sar_i64(vc, va, tmp);
1972 tcg_temp_free(tmp);
1973 }
1974 break;
1975 case 0x52:
1976 /* MSKWH */
1977 gen_msk_h(ctx, vc, va, rb, islit, lit, 0x03);
1978 break;
1979 case 0x57:
1980 /* INSWH */
1981 gen_ins_h(ctx, vc, va, rb, islit, lit, 0x03);
1982 break;
1983 case 0x5A:
1984 /* EXTWH */
1985 gen_ext_h(ctx, vc, va, rb, islit, lit, 0x03);
1986 break;
1987 case 0x62:
1988 /* MSKLH */
1989 gen_msk_h(ctx, vc, va, rb, islit, lit, 0x0f);
1990 break;
1991 case 0x67:
1992 /* INSLH */
1993 gen_ins_h(ctx, vc, va, rb, islit, lit, 0x0f);
1994 break;
1995 case 0x6A:
1996 /* EXTLH */
1997 gen_ext_h(ctx, vc, va, rb, islit, lit, 0x0f);
1998 break;
1999 case 0x72:
2000 /* MSKQH */
2001 gen_msk_h(ctx, vc, va, rb, islit, lit, 0xff);
2002 break;
2003 case 0x77:
2004 /* INSQH */
2005 gen_ins_h(ctx, vc, va, rb, islit, lit, 0xff);
2006 break;
2007 case 0x7A:
2008 /* EXTQH */
2009 gen_ext_h(ctx, vc, va, rb, islit, lit, 0xff);
2010 break;
2011 default:
2012 goto invalid_opc;
2013 }
2014 break;
2015
2016 case 0x13:
2017 vc = dest_gpr(ctx, rc);
2018 vb = load_gpr_lit(ctx, rb, lit, islit);
2019 va = load_gpr(ctx, ra);
2020 switch (fn7) {
2021 case 0x00:
2022 /* MULL */
2023 tcg_gen_mul_i64(vc, va, vb);
2024 tcg_gen_ext32s_i64(vc, vc);
2025 break;
2026 case 0x20:
2027 /* MULQ */
2028 tcg_gen_mul_i64(vc, va, vb);
2029 break;
2030 case 0x30:
2031 /* UMULH */
2032 tmp = tcg_temp_new();
2033 tcg_gen_mulu2_i64(tmp, vc, va, vb);
2034 tcg_temp_free(tmp);
2035 break;
2036 case 0x40:
2037 /* MULL/V */
2038 tmp = tcg_temp_new();
2039 tcg_gen_ext32s_i64(tmp, va);
2040 tcg_gen_ext32s_i64(vc, vb);
2041 tcg_gen_mul_i64(tmp, tmp, vc);
2042 tcg_gen_ext32s_i64(vc, tmp);
2043 gen_helper_check_overflow(cpu_env, vc, tmp);
2044 tcg_temp_free(tmp);
2045 break;
2046 case 0x60:
2047 /* MULQ/V */
2048 tmp = tcg_temp_new();
2049 tmp2 = tcg_temp_new();
2050 tcg_gen_muls2_i64(vc, tmp, va, vb);
2051 tcg_gen_sari_i64(tmp2, vc, 63);
2052 gen_helper_check_overflow(cpu_env, tmp, tmp2);
2053 tcg_temp_free(tmp);
2054 tcg_temp_free(tmp2);
2055 break;
2056 default:
2057 goto invalid_opc;
2058 }
2059 break;
2060
2061 case 0x14:
2062 REQUIRE_AMASK(FIX);
2063 vc = dest_fpr(ctx, rc);
2064 switch (fpfn) { /* fn11 & 0x3F */
2065 case 0x04:
2066 /* ITOFS */
2067 REQUIRE_REG_31(rb);
2068 t32 = tcg_temp_new_i32();
2069 va = load_gpr(ctx, ra);
2070 tcg_gen_extrl_i64_i32(t32, va);
2071 gen_helper_memory_to_s(vc, t32);
2072 tcg_temp_free_i32(t32);
2073 break;
2074 case 0x0A:
2075 /* SQRTF */
2076 REQUIRE_REG_31(ra);
2077 vb = load_fpr(ctx, rb);
2078 gen_helper_sqrtf(vc, cpu_env, vb);
2079 break;
2080 case 0x0B:
2081 /* SQRTS */
2082 REQUIRE_REG_31(ra);
2083 gen_sqrts(ctx, rb, rc, fn11);
2084 break;
2085 case 0x14:
2086 /* ITOFF */
2087 REQUIRE_REG_31(rb);
2088 t32 = tcg_temp_new_i32();
2089 va = load_gpr(ctx, ra);
2090 tcg_gen_extrl_i64_i32(t32, va);
2091 gen_helper_memory_to_f(vc, t32);
2092 tcg_temp_free_i32(t32);
2093 break;
2094 case 0x24:
2095 /* ITOFT */
2096 REQUIRE_REG_31(rb);
2097 va = load_gpr(ctx, ra);
2098 tcg_gen_mov_i64(vc, va);
2099 break;
2100 case 0x2A:
2101 /* SQRTG */
2102 REQUIRE_REG_31(ra);
2103 vb = load_fpr(ctx, rb);
2104 gen_helper_sqrtg(vc, cpu_env, vb);
2105 break;
2106 case 0x02B:
2107 /* SQRTT */
2108 REQUIRE_REG_31(ra);
2109 gen_sqrtt(ctx, rb, rc, fn11);
2110 break;
2111 default:
2112 goto invalid_opc;
2113 }
2114 break;
2115
2116 case 0x15:
2117 /* VAX floating point */
2118 /* XXX: rounding mode and trap are ignored (!) */
2119 vc = dest_fpr(ctx, rc);
2120 vb = load_fpr(ctx, rb);
2121 va = load_fpr(ctx, ra);
2122 switch (fpfn) { /* fn11 & 0x3F */
2123 case 0x00:
2124 /* ADDF */
2125 gen_helper_addf(vc, cpu_env, va, vb);
2126 break;
2127 case 0x01:
2128 /* SUBF */
2129 gen_helper_subf(vc, cpu_env, va, vb);
2130 break;
2131 case 0x02:
2132 /* MULF */
2133 gen_helper_mulf(vc, cpu_env, va, vb);
2134 break;
2135 case 0x03:
2136 /* DIVF */
2137 gen_helper_divf(vc, cpu_env, va, vb);
2138 break;
2139 case 0x1E:
2140 /* CVTDG -- TODO */
2141 REQUIRE_REG_31(ra);
2142 goto invalid_opc;
2143 case 0x20:
2144 /* ADDG */
2145 gen_helper_addg(vc, cpu_env, va, vb);
2146 break;
2147 case 0x21:
2148 /* SUBG */
2149 gen_helper_subg(vc, cpu_env, va, vb);
2150 break;
2151 case 0x22:
2152 /* MULG */
2153 gen_helper_mulg(vc, cpu_env, va, vb);
2154 break;
2155 case 0x23:
2156 /* DIVG */
2157 gen_helper_divg(vc, cpu_env, va, vb);
2158 break;
2159 case 0x25:
2160 /* CMPGEQ */
2161 gen_helper_cmpgeq(vc, cpu_env, va, vb);
2162 break;
2163 case 0x26:
2164 /* CMPGLT */
2165 gen_helper_cmpglt(vc, cpu_env, va, vb);
2166 break;
2167 case 0x27:
2168 /* CMPGLE */
2169 gen_helper_cmpgle(vc, cpu_env, va, vb);
2170 break;
2171 case 0x2C:
2172 /* CVTGF */
2173 REQUIRE_REG_31(ra);
2174 gen_helper_cvtgf(vc, cpu_env, vb);
2175 break;
2176 case 0x2D:
2177 /* CVTGD -- TODO */
2178 REQUIRE_REG_31(ra);
2179 goto invalid_opc;
2180 case 0x2F:
2181 /* CVTGQ */
2182 REQUIRE_REG_31(ra);
2183 gen_helper_cvtgq(vc, cpu_env, vb);
2184 break;
2185 case 0x3C:
2186 /* CVTQF */
2187 REQUIRE_REG_31(ra);
2188 gen_helper_cvtqf(vc, cpu_env, vb);
2189 break;
2190 case 0x3E:
2191 /* CVTQG */
2192 REQUIRE_REG_31(ra);
2193 gen_helper_cvtqg(vc, cpu_env, vb);
2194 break;
2195 default:
2196 goto invalid_opc;
2197 }
2198 break;
2199
2200 case 0x16:
2201 /* IEEE floating-point */
2202 switch (fpfn) { /* fn11 & 0x3F */
2203 case 0x00:
2204 /* ADDS */
2205 gen_adds(ctx, ra, rb, rc, fn11);
2206 break;
2207 case 0x01:
2208 /* SUBS */
2209 gen_subs(ctx, ra, rb, rc, fn11);
2210 break;
2211 case 0x02:
2212 /* MULS */
2213 gen_muls(ctx, ra, rb, rc, fn11);
2214 break;
2215 case 0x03:
2216 /* DIVS */
2217 gen_divs(ctx, ra, rb, rc, fn11);
2218 break;
2219 case 0x20:
2220 /* ADDT */
2221 gen_addt(ctx, ra, rb, rc, fn11);
2222 break;
2223 case 0x21:
2224 /* SUBT */
2225 gen_subt(ctx, ra, rb, rc, fn11);
2226 break;
2227 case 0x22:
2228 /* MULT */
2229 gen_mult(ctx, ra, rb, rc, fn11);
2230 break;
2231 case 0x23:
2232 /* DIVT */
2233 gen_divt(ctx, ra, rb, rc, fn11);
2234 break;
2235 case 0x24:
2236 /* CMPTUN */
2237 gen_cmptun(ctx, ra, rb, rc, fn11);
2238 break;
2239 case 0x25:
2240 /* CMPTEQ */
2241 gen_cmpteq(ctx, ra, rb, rc, fn11);
2242 break;
2243 case 0x26:
2244 /* CMPTLT */
2245 gen_cmptlt(ctx, ra, rb, rc, fn11);
2246 break;
2247 case 0x27:
2248 /* CMPTLE */
2249 gen_cmptle(ctx, ra, rb, rc, fn11);
2250 break;
2251 case 0x2C:
2252 REQUIRE_REG_31(ra);
2253 if (fn11 == 0x2AC || fn11 == 0x6AC) {
2254 /* CVTST */
2255 gen_cvtst(ctx, rb, rc, fn11);
2256 } else {
2257 /* CVTTS */
2258 gen_cvtts(ctx, rb, rc, fn11);
2259 }
2260 break;
2261 case 0x2F:
2262 /* CVTTQ */
2263 REQUIRE_REG_31(ra);
2264 gen_cvttq(ctx, rb, rc, fn11);
2265 break;
2266 case 0x3C:
2267 /* CVTQS */
2268 REQUIRE_REG_31(ra);
2269 gen_cvtqs(ctx, rb, rc, fn11);
2270 break;
2271 case 0x3E:
2272 /* CVTQT */
2273 REQUIRE_REG_31(ra);
2274 gen_cvtqt(ctx, rb, rc, fn11);
2275 break;
2276 default:
2277 goto invalid_opc;
2278 }
2279 break;
2280
2281 case 0x17:
2282 switch (fn11) {
2283 case 0x010:
2284 /* CVTLQ */
2285 REQUIRE_REG_31(ra);
2286 vc = dest_fpr(ctx, rc);
2287 vb = load_fpr(ctx, rb);
2288 gen_cvtlq(vc, vb);
2289 break;
2290 case 0x020:
2291 /* CPYS */
2292 if (rc == 31) {
2293 /* Special case CPYS as FNOP. */
2294 } else {
2295 vc = dest_fpr(ctx, rc);
2296 va = load_fpr(ctx, ra);
2297 if (ra == rb) {
2298 /* Special case CPYS as FMOV. */
2299 tcg_gen_mov_i64(vc, va);
2300 } else {
2301 vb = load_fpr(ctx, rb);
2302 gen_cpy_mask(vc, va, vb, 0, 0x8000000000000000ULL);
2303 }
2304 }
2305 break;
2306 case 0x021:
2307 /* CPYSN */
2308 vc = dest_fpr(ctx, rc);
2309 vb = load_fpr(ctx, rb);
2310 va = load_fpr(ctx, ra);
2311 gen_cpy_mask(vc, va, vb, 1, 0x8000000000000000ULL);
2312 break;
2313 case 0x022:
2314 /* CPYSE */
2315 vc = dest_fpr(ctx, rc);
2316 vb = load_fpr(ctx, rb);
2317 va = load_fpr(ctx, ra);
2318 gen_cpy_mask(vc, va, vb, 0, 0xFFF0000000000000ULL);
2319 break;
2320 case 0x024:
2321 /* MT_FPCR */
2322 va = load_fpr(ctx, ra);
2323 gen_helper_store_fpcr(cpu_env, va);
2324 if (ctx->tb_rm == QUAL_RM_D) {
2325 /* Re-do the copy of the rounding mode to fp_status
2326 the next time we use dynamic rounding. */
2327 ctx->tb_rm = -1;
2328 }
2329 break;
2330 case 0x025:
2331 /* MF_FPCR */
2332 va = dest_fpr(ctx, ra);
2333 gen_helper_load_fpcr(va, cpu_env);
2334 break;
2335 case 0x02A:
2336 /* FCMOVEQ */
2337 gen_fcmov(ctx, TCG_COND_EQ, ra, rb, rc);
2338 break;
2339 case 0x02B:
2340 /* FCMOVNE */
2341 gen_fcmov(ctx, TCG_COND_NE, ra, rb, rc);
2342 break;
2343 case 0x02C:
2344 /* FCMOVLT */
2345 gen_fcmov(ctx, TCG_COND_LT, ra, rb, rc);
2346 break;
2347 case 0x02D:
2348 /* FCMOVGE */
2349 gen_fcmov(ctx, TCG_COND_GE, ra, rb, rc);
2350 break;
2351 case 0x02E:
2352 /* FCMOVLE */
2353 gen_fcmov(ctx, TCG_COND_LE, ra, rb, rc);
2354 break;
2355 case 0x02F:
2356 /* FCMOVGT */
2357 gen_fcmov(ctx, TCG_COND_GT, ra, rb, rc);
2358 break;
2359 case 0x030: /* CVTQL */
2360 case 0x130: /* CVTQL/V */
2361 case 0x530: /* CVTQL/SV */
2362 REQUIRE_REG_31(ra);
2363 vc = dest_fpr(ctx, rc);
2364 vb = load_fpr(ctx, rb);
2365 gen_helper_cvtql(vc, cpu_env, vb);
2366 gen_fp_exc_raise(rc, fn11);
2367 break;
2368 default:
2369 goto invalid_opc;
2370 }
2371 break;
2372
2373 case 0x18:
2374 switch ((uint16_t)disp16) {
2375 case 0x0000:
2376 /* TRAPB */
2377 /* No-op. */
2378 break;
2379 case 0x0400:
2380 /* EXCB */
2381 /* No-op. */
2382 break;
2383 case 0x4000:
2384 /* MB */
2385 tcg_gen_mb(TCG_MO_ALL | TCG_BAR_SC);
2386 break;
2387 case 0x4400:
2388 /* WMB */
2389 tcg_gen_mb(TCG_MO_ST_ST | TCG_BAR_SC);
2390 break;
2391 case 0x8000:
2392 /* FETCH */
2393 /* No-op */
2394 break;
2395 case 0xA000:
2396 /* FETCH_M */
2397 /* No-op */
2398 break;
2399 case 0xC000:
2400 /* RPCC */
2401 va = dest_gpr(ctx, ra);
2402 if (tb_cflags(ctx->base.tb) & CF_USE_ICOUNT) {
2403 gen_io_start();
2404 gen_helper_load_pcc(va, cpu_env);
2405 gen_io_end();
2406 ret = DISAS_PC_STALE;
2407 } else {
2408 gen_helper_load_pcc(va, cpu_env);
2409 }
2410 break;
2411 case 0xE000:
2412 /* RC */
2413 gen_rx(ctx, ra, 0);
2414 break;
2415 case 0xE800:
2416 /* ECB */
2417 break;
2418 case 0xF000:
2419 /* RS */
2420 gen_rx(ctx, ra, 1);
2421 break;
2422 case 0xF800:
2423 /* WH64 */
2424 /* No-op */
2425 break;
2426 case 0xFC00:
2427 /* WH64EN */
2428 /* No-op */
2429 break;
2430 default:
2431 goto invalid_opc;
2432 }
2433 break;
2434
2435 case 0x19:
2436 /* HW_MFPR (PALcode) */
2437 #ifndef CONFIG_USER_ONLY
2438 REQUIRE_TB_FLAG(ENV_FLAG_PAL_MODE);
2439 va = dest_gpr(ctx, ra);
2440 ret = gen_mfpr(ctx, va, insn & 0xffff);
2441 break;
2442 #else
2443 goto invalid_opc;
2444 #endif
2445
2446 case 0x1A:
2447 /* JMP, JSR, RET, JSR_COROUTINE. These only differ by the branch
2448 prediction stack action, which of course we don't implement. */
2449 vb = load_gpr(ctx, rb);
2450 tcg_gen_andi_i64(cpu_pc, vb, ~3);
2451 if (ra != 31) {
2452 tcg_gen_movi_i64(ctx->ir[ra], ctx->base.pc_next);
2453 }
2454 ret = DISAS_PC_UPDATED;
2455 break;
2456
2457 case 0x1B:
2458 /* HW_LD (PALcode) */
2459 #ifndef CONFIG_USER_ONLY
2460 REQUIRE_TB_FLAG(ENV_FLAG_PAL_MODE);
2461 {
2462 TCGv addr = tcg_temp_new();
2463 vb = load_gpr(ctx, rb);
2464 va = dest_gpr(ctx, ra);
2465
2466 tcg_gen_addi_i64(addr, vb, disp12);
2467 switch ((insn >> 12) & 0xF) {
2468 case 0x0:
2469 /* Longword physical access (hw_ldl/p) */
2470 tcg_gen_qemu_ld_i64(va, addr, MMU_PHYS_IDX, MO_LESL);
2471 break;
2472 case 0x1:
2473 /* Quadword physical access (hw_ldq/p) */
2474 tcg_gen_qemu_ld_i64(va, addr, MMU_PHYS_IDX, MO_LEQ);
2475 break;
2476 case 0x2:
2477 /* Longword physical access with lock (hw_ldl_l/p) */
2478 gen_qemu_ldl_l(va, addr, MMU_PHYS_IDX);
2479 break;
2480 case 0x3:
2481 /* Quadword physical access with lock (hw_ldq_l/p) */
2482 gen_qemu_ldq_l(va, addr, MMU_PHYS_IDX);
2483 break;
2484 case 0x4:
2485 /* Longword virtual PTE fetch (hw_ldl/v) */
2486 goto invalid_opc;
2487 case 0x5:
2488 /* Quadword virtual PTE fetch (hw_ldq/v) */
2489 goto invalid_opc;
2490 break;
2491 case 0x6:
2492 /* Invalid */
2493 goto invalid_opc;
2494 case 0x7:
2495 /* Invaliid */
2496 goto invalid_opc;
2497 case 0x8:
2498 /* Longword virtual access (hw_ldl) */
2499 goto invalid_opc;
2500 case 0x9:
2501 /* Quadword virtual access (hw_ldq) */
2502 goto invalid_opc;
2503 case 0xA:
2504 /* Longword virtual access with protection check (hw_ldl/w) */
2505 tcg_gen_qemu_ld_i64(va, addr, MMU_KERNEL_IDX, MO_LESL);
2506 break;
2507 case 0xB:
2508 /* Quadword virtual access with protection check (hw_ldq/w) */
2509 tcg_gen_qemu_ld_i64(va, addr, MMU_KERNEL_IDX, MO_LEQ);
2510 break;
2511 case 0xC:
2512 /* Longword virtual access with alt access mode (hw_ldl/a)*/
2513 goto invalid_opc;
2514 case 0xD:
2515 /* Quadword virtual access with alt access mode (hw_ldq/a) */
2516 goto invalid_opc;
2517 case 0xE:
2518 /* Longword virtual access with alternate access mode and
2519 protection checks (hw_ldl/wa) */
2520 tcg_gen_qemu_ld_i64(va, addr, MMU_USER_IDX, MO_LESL);
2521 break;
2522 case 0xF:
2523 /* Quadword virtual access with alternate access mode and
2524 protection checks (hw_ldq/wa) */
2525 tcg_gen_qemu_ld_i64(va, addr, MMU_USER_IDX, MO_LEQ);
2526 break;
2527 }
2528 tcg_temp_free(addr);
2529 break;
2530 }
2531 #else
2532 goto invalid_opc;
2533 #endif
2534
2535 case 0x1C:
2536 vc = dest_gpr(ctx, rc);
2537 if (fn7 == 0x70) {
2538 /* FTOIT */
2539 REQUIRE_AMASK(FIX);
2540 REQUIRE_REG_31(rb);
2541 va = load_fpr(ctx, ra);
2542 tcg_gen_mov_i64(vc, va);
2543 break;
2544 } else if (fn7 == 0x78) {
2545 /* FTOIS */
2546 REQUIRE_AMASK(FIX);
2547 REQUIRE_REG_31(rb);
2548 t32 = tcg_temp_new_i32();
2549 va = load_fpr(ctx, ra);
2550 gen_helper_s_to_memory(t32, va);
2551 tcg_gen_ext_i32_i64(vc, t32);
2552 tcg_temp_free_i32(t32);
2553 break;
2554 }
2555
2556 vb = load_gpr_lit(ctx, rb, lit, islit);
2557 switch (fn7) {
2558 case 0x00:
2559 /* SEXTB */
2560 REQUIRE_AMASK(BWX);
2561 REQUIRE_REG_31(ra);
2562 tcg_gen_ext8s_i64(vc, vb);
2563 break;
2564 case 0x01:
2565 /* SEXTW */
2566 REQUIRE_AMASK(BWX);
2567 REQUIRE_REG_31(ra);
2568 tcg_gen_ext16s_i64(vc, vb);
2569 break;
2570 case 0x30:
2571 /* CTPOP */
2572 REQUIRE_AMASK(CIX);
2573 REQUIRE_REG_31(ra);
2574 REQUIRE_NO_LIT;
2575 tcg_gen_ctpop_i64(vc, vb);
2576 break;
2577 case 0x31:
2578 /* PERR */
2579 REQUIRE_AMASK(MVI);
2580 REQUIRE_NO_LIT;
2581 va = load_gpr(ctx, ra);
2582 gen_helper_perr(vc, va, vb);
2583 break;
2584 case 0x32:
2585 /* CTLZ */
2586 REQUIRE_AMASK(CIX);
2587 REQUIRE_REG_31(ra);
2588 REQUIRE_NO_LIT;
2589 tcg_gen_clzi_i64(vc, vb, 64);
2590 break;
2591 case 0x33:
2592 /* CTTZ */
2593 REQUIRE_AMASK(CIX);
2594 REQUIRE_REG_31(ra);
2595 REQUIRE_NO_LIT;
2596 tcg_gen_ctzi_i64(vc, vb, 64);
2597 break;
2598 case 0x34:
2599 /* UNPKBW */
2600 REQUIRE_AMASK(MVI);
2601 REQUIRE_REG_31(ra);
2602 REQUIRE_NO_LIT;
2603 gen_helper_unpkbw(vc, vb);
2604 break;
2605 case 0x35:
2606 /* UNPKBL */
2607 REQUIRE_AMASK(MVI);
2608 REQUIRE_REG_31(ra);
2609 REQUIRE_NO_LIT;
2610 gen_helper_unpkbl(vc, vb);
2611 break;
2612 case 0x36:
2613 /* PKWB */
2614 REQUIRE_AMASK(MVI);
2615 REQUIRE_REG_31(ra);
2616 REQUIRE_NO_LIT;
2617 gen_helper_pkwb(vc, vb);
2618 break;
2619 case 0x37:
2620 /* PKLB */
2621 REQUIRE_AMASK(MVI);
2622 REQUIRE_REG_31(ra);
2623 REQUIRE_NO_LIT;
2624 gen_helper_pklb(vc, vb);
2625 break;
2626 case 0x38:
2627 /* MINSB8 */
2628 REQUIRE_AMASK(MVI);
2629 va = load_gpr(ctx, ra);
2630 gen_helper_minsb8(vc, va, vb);
2631 break;
2632 case 0x39:
2633 /* MINSW4 */
2634 REQUIRE_AMASK(MVI);
2635 va = load_gpr(ctx, ra);
2636 gen_helper_minsw4(vc, va, vb);
2637 break;
2638 case 0x3A:
2639 /* MINUB8 */
2640 REQUIRE_AMASK(MVI);
2641 va = load_gpr(ctx, ra);
2642 gen_helper_minub8(vc, va, vb);
2643 break;
2644 case 0x3B:
2645 /* MINUW4 */
2646 REQUIRE_AMASK(MVI);
2647 va = load_gpr(ctx, ra);
2648 gen_helper_minuw4(vc, va, vb);
2649 break;
2650 case 0x3C:
2651 /* MAXUB8 */
2652 REQUIRE_AMASK(MVI);
2653 va = load_gpr(ctx, ra);
2654 gen_helper_maxub8(vc, va, vb);
2655 break;
2656 case 0x3D:
2657 /* MAXUW4 */
2658 REQUIRE_AMASK(MVI);
2659 va = load_gpr(ctx, ra);
2660 gen_helper_maxuw4(vc, va, vb);
2661 break;
2662 case 0x3E:
2663 /* MAXSB8 */
2664 REQUIRE_AMASK(MVI);
2665 va = load_gpr(ctx, ra);
2666 gen_helper_maxsb8(vc, va, vb);
2667 break;
2668 case 0x3F:
2669 /* MAXSW4 */
2670 REQUIRE_AMASK(MVI);
2671 va = load_gpr(ctx, ra);
2672 gen_helper_maxsw4(vc, va, vb);
2673 break;
2674 default:
2675 goto invalid_opc;
2676 }
2677 break;
2678
2679 case 0x1D:
2680 /* HW_MTPR (PALcode) */
2681 #ifndef CONFIG_USER_ONLY
2682 REQUIRE_TB_FLAG(ENV_FLAG_PAL_MODE);
2683 vb = load_gpr(ctx, rb);
2684 ret = gen_mtpr(ctx, vb, insn & 0xffff);
2685 break;
2686 #else
2687 goto invalid_opc;
2688 #endif
2689
2690 case 0x1E:
2691 /* HW_RET (PALcode) */
2692 #ifndef CONFIG_USER_ONLY
2693 REQUIRE_TB_FLAG(ENV_FLAG_PAL_MODE);
2694 if (rb == 31) {
2695 /* Pre-EV6 CPUs interpreted this as HW_REI, loading the return
2696 address from EXC_ADDR. This turns out to be useful for our
2697 emulation PALcode, so continue to accept it. */
2698 ctx->lit = vb = tcg_temp_new();
2699 tcg_gen_ld_i64(vb, cpu_env, offsetof(CPUAlphaState, exc_addr));
2700 } else {
2701 vb = load_gpr(ctx, rb);
2702 }
2703 tcg_gen_movi_i64(cpu_lock_addr, -1);
2704 tmp = tcg_temp_new();
2705 tcg_gen_movi_i64(tmp, 0);
2706 st_flag_byte(tmp, ENV_FLAG_RX_SHIFT);
2707 tcg_gen_andi_i64(tmp, vb, 1);
2708 st_flag_byte(tmp, ENV_FLAG_PAL_SHIFT);
2709 tcg_temp_free(tmp);
2710 tcg_gen_andi_i64(cpu_pc, vb, ~3);
2711 /* Allow interrupts to be recognized right away. */
2712 ret = DISAS_PC_UPDATED_NOCHAIN;
2713 break;
2714 #else
2715 goto invalid_opc;
2716 #endif
2717
2718 case 0x1F:
2719 /* HW_ST (PALcode) */
2720 #ifndef CONFIG_USER_ONLY
2721 REQUIRE_TB_FLAG(ENV_FLAG_PAL_MODE);
2722 {
2723 switch ((insn >> 12) & 0xF) {
2724 case 0x0:
2725 /* Longword physical access */
2726 va = load_gpr(ctx, ra);
2727 vb = load_gpr(ctx, rb);
2728 tmp = tcg_temp_new();
2729 tcg_gen_addi_i64(tmp, vb, disp12);
2730 tcg_gen_qemu_st_i64(va, tmp, MMU_PHYS_IDX, MO_LESL);
2731 tcg_temp_free(tmp);
2732 break;
2733 case 0x1:
2734 /* Quadword physical access */
2735 va = load_gpr(ctx, ra);
2736 vb = load_gpr(ctx, rb);
2737 tmp = tcg_temp_new();
2738 tcg_gen_addi_i64(tmp, vb, disp12);
2739 tcg_gen_qemu_st_i64(va, tmp, MMU_PHYS_IDX, MO_LEQ);
2740 tcg_temp_free(tmp);
2741 break;
2742 case 0x2:
2743 /* Longword physical access with lock */
2744 ret = gen_store_conditional(ctx, ra, rb, disp12,
2745 MMU_PHYS_IDX, MO_LESL);
2746 break;
2747 case 0x3:
2748 /* Quadword physical access with lock */
2749 ret = gen_store_conditional(ctx, ra, rb, disp12,
2750 MMU_PHYS_IDX, MO_LEQ);
2751 break;
2752 case 0x4:
2753 /* Longword virtual access */
2754 goto invalid_opc;
2755 case 0x5:
2756 /* Quadword virtual access */
2757 goto invalid_opc;
2758 case 0x6:
2759 /* Invalid */
2760 goto invalid_opc;
2761 case 0x7:
2762 /* Invalid */
2763 goto invalid_opc;
2764 case 0x8:
2765 /* Invalid */
2766 goto invalid_opc;
2767 case 0x9:
2768 /* Invalid */
2769 goto invalid_opc;
2770 case 0xA:
2771 /* Invalid */
2772 goto invalid_opc;
2773 case 0xB:
2774 /* Invalid */
2775 goto invalid_opc;
2776 case 0xC:
2777 /* Longword virtual access with alternate access mode */
2778 goto invalid_opc;
2779 case 0xD:
2780 /* Quadword virtual access with alternate access mode */
2781 goto invalid_opc;
2782 case 0xE:
2783 /* Invalid */
2784 goto invalid_opc;
2785 case 0xF:
2786 /* Invalid */
2787 goto invalid_opc;
2788 }
2789 break;
2790 }
2791 #else
2792 goto invalid_opc;
2793 #endif
2794 case 0x20:
2795 /* LDF */
2796 gen_load_mem(ctx, &gen_qemu_ldf, ra, rb, disp16, 1, 0);
2797 break;
2798 case 0x21:
2799 /* LDG */
2800 gen_load_mem(ctx, &gen_qemu_ldg, ra, rb, disp16, 1, 0);
2801 break;
2802 case 0x22:
2803 /* LDS */
2804 gen_load_mem(ctx, &gen_qemu_lds, ra, rb, disp16, 1, 0);
2805 break;
2806 case 0x23:
2807 /* LDT */
2808 gen_load_mem(ctx, &tcg_gen_qemu_ld64, ra, rb, disp16, 1, 0);
2809 break;
2810 case 0x24:
2811 /* STF */
2812 gen_store_mem(ctx, &gen_qemu_stf, ra, rb, disp16, 1, 0);
2813 break;
2814 case 0x25:
2815 /* STG */
2816 gen_store_mem(ctx, &gen_qemu_stg, ra, rb, disp16, 1, 0);
2817 break;
2818 case 0x26:
2819 /* STS */
2820 gen_store_mem(ctx, &gen_qemu_sts, ra, rb, disp16, 1, 0);
2821 break;
2822 case 0x27:
2823 /* STT */
2824 gen_store_mem(ctx, &tcg_gen_qemu_st64, ra, rb, disp16, 1, 0);
2825 break;
2826 case 0x28:
2827 /* LDL */
2828 gen_load_mem(ctx, &tcg_gen_qemu_ld32s, ra, rb, disp16, 0, 0);
2829 break;
2830 case 0x29:
2831 /* LDQ */
2832 gen_load_mem(ctx, &tcg_gen_qemu_ld64, ra, rb, disp16, 0, 0);
2833 break;
2834 case 0x2A:
2835 /* LDL_L */
2836 gen_load_mem(ctx, &gen_qemu_ldl_l, ra, rb, disp16, 0, 0);
2837 break;
2838 case 0x2B:
2839 /* LDQ_L */
2840 gen_load_mem(ctx, &gen_qemu_ldq_l, ra, rb, disp16, 0, 0);
2841 break;
2842 case 0x2C:
2843 /* STL */
2844 gen_store_mem(ctx, &tcg_gen_qemu_st32, ra, rb, disp16, 0, 0);
2845 break;
2846 case 0x2D:
2847 /* STQ */
2848 gen_store_mem(ctx, &tcg_gen_qemu_st64, ra, rb, disp16, 0, 0);
2849 break;
2850 case 0x2E:
2851 /* STL_C */
2852 ret = gen_store_conditional(ctx, ra, rb, disp16,
2853 ctx->mem_idx, MO_LESL);
2854 break;
2855 case 0x2F:
2856 /* STQ_C */
2857 ret = gen_store_conditional(ctx, ra, rb, disp16,
2858 ctx->mem_idx, MO_LEQ);
2859 break;
2860 case 0x30:
2861 /* BR */
2862 ret = gen_bdirect(ctx, ra, disp21);
2863 break;
2864 case 0x31: /* FBEQ */
2865 ret = gen_fbcond(ctx, TCG_COND_EQ, ra, disp21);
2866 break;
2867 case 0x32: /* FBLT */
2868 ret = gen_fbcond(ctx, TCG_COND_LT, ra, disp21);
2869 break;
2870 case 0x33: /* FBLE */
2871 ret = gen_fbcond(ctx, TCG_COND_LE, ra, disp21);
2872 break;
2873 case 0x34:
2874 /* BSR */
2875 ret = gen_bdirect(ctx, ra, disp21);
2876 break;
2877 case 0x35: /* FBNE */
2878 ret = gen_fbcond(ctx, TCG_COND_NE, ra, disp21);
2879 break;
2880 case 0x36: /* FBGE */
2881 ret = gen_fbcond(ctx, TCG_COND_GE, ra, disp21);
2882 break;
2883 case 0x37: /* FBGT */
2884 ret = gen_fbcond(ctx, TCG_COND_GT, ra, disp21);
2885 break;
2886 case 0x38:
2887 /* BLBC */
2888 ret = gen_bcond(ctx, TCG_COND_EQ, ra, disp21, 1);
2889 break;
2890 case 0x39:
2891 /* BEQ */
2892 ret = gen_bcond(ctx, TCG_COND_EQ, ra, disp21, 0);
2893 break;
2894 case 0x3A:
2895 /* BLT */
2896 ret = gen_bcond(ctx, TCG_COND_LT, ra, disp21, 0);
2897 break;
2898 case 0x3B:
2899 /* BLE */
2900 ret = gen_bcond(ctx, TCG_COND_LE, ra, disp21, 0);
2901 break;
2902 case 0x3C:
2903 /* BLBS */
2904 ret = gen_bcond(ctx, TCG_COND_NE, ra, disp21, 1);
2905 break;
2906 case 0x3D:
2907 /* BNE */
2908 ret = gen_bcond(ctx, TCG_COND_NE, ra, disp21, 0);
2909 break;
2910 case 0x3E:
2911 /* BGE */
2912 ret = gen_bcond(ctx, TCG_COND_GE, ra, disp21, 0);
2913 break;
2914 case 0x3F:
2915 /* BGT */
2916 ret = gen_bcond(ctx, TCG_COND_GT, ra, disp21, 0);
2917 break;
2918 invalid_opc:
2919 ret = gen_invalid(ctx);
2920 break;
2921 }
2922
2923 return ret;
2924 }
2925
2926 static int alpha_tr_init_disas_context(DisasContextBase *dcbase,
2927 CPUState *cpu, int max_insns)
2928 {
2929 DisasContext *ctx = container_of(dcbase, DisasContext, base);
2930 CPUAlphaState *env = cpu->env_ptr;
2931 int64_t bound, mask;
2932
2933 ctx->tbflags = ctx->base.tb->flags;
2934 ctx->mem_idx = cpu_mmu_index(env, false);
2935 ctx->implver = env->implver;
2936 ctx->amask = env->amask;
2937
2938 #ifdef CONFIG_USER_ONLY
2939 ctx->ir = cpu_std_ir;
2940 #else
2941 ctx->palbr = env->palbr;
2942 ctx->ir = (ctx->tbflags & ENV_FLAG_PAL_MODE ? cpu_pal_ir : cpu_std_ir);
2943 #endif
2944
2945 /* ??? Every TB begins with unset rounding mode, to be initialized on
2946 the first fp insn of the TB. Alternately we could define a proper
2947 default for every TB (e.g. QUAL_RM_N or QUAL_RM_D) and make sure
2948 to reset the FP_STATUS to that default at the end of any TB that
2949 changes the default. We could even (gasp) dynamiclly figure out
2950 what default would be most efficient given the running program. */
2951 ctx->tb_rm = -1;
2952 /* Similarly for flush-to-zero. */
2953 ctx->tb_ftz = -1;
2954
2955 TCGV_UNUSED_I64(ctx->zero);
2956 TCGV_UNUSED_I64(ctx->sink);
2957 TCGV_UNUSED_I64(ctx->lit);
2958
2959 /* Bound the number of insns to execute to those left on the page. */
2960 if (in_superpage(ctx, ctx->base.pc_first)) {
2961 mask = -1ULL << 41;
2962 } else {
2963 mask = TARGET_PAGE_MASK;
2964 }
2965 bound = -(ctx->base.pc_first | mask) / 4;
2966
2967 return MIN(max_insns, bound);
2968 }
2969
2970 static void alpha_tr_tb_start(DisasContextBase *db, CPUState *cpu)
2971 {
2972 }
2973
2974 static void alpha_tr_insn_start(DisasContextBase *dcbase, CPUState *cpu)
2975 {
2976 tcg_gen_insn_start(dcbase->pc_next);
2977 }
2978
2979 static bool alpha_tr_breakpoint_check(DisasContextBase *dcbase, CPUState *cpu,
2980 const CPUBreakpoint *bp)
2981 {
2982 DisasContext *ctx = container_of(dcbase, DisasContext, base);
2983
2984 ctx->base.is_jmp = gen_excp(ctx, EXCP_DEBUG, 0);
2985
2986 /* The address covered by the breakpoint must be included in
2987 [tb->pc, tb->pc + tb->size) in order to for it to be
2988 properly cleared -- thus we increment the PC here so that
2989 the logic setting tb->size below does the right thing. */
2990 ctx->base.pc_next += 4;
2991 return true;
2992 }
2993
2994 static void alpha_tr_translate_insn(DisasContextBase *dcbase, CPUState *cpu)
2995 {
2996 DisasContext *ctx = container_of(dcbase, DisasContext, base);
2997 CPUAlphaState *env = cpu->env_ptr;
2998 uint32_t insn = cpu_ldl_code(env, ctx->base.pc_next);
2999
3000 ctx->base.pc_next += 4;
3001 ctx->base.is_jmp = translate_one(ctx, insn);
3002
3003 free_context_temps(ctx);
3004 translator_loop_temp_check(&ctx->base);
3005 }
3006
3007 static void alpha_tr_tb_stop(DisasContextBase *dcbase, CPUState *cpu)
3008 {
3009 DisasContext *ctx = container_of(dcbase, DisasContext, base);
3010
3011 switch (ctx->base.is_jmp) {
3012 case DISAS_NORETURN:
3013 break;
3014 case DISAS_TOO_MANY:
3015 if (use_goto_tb(ctx, ctx->base.pc_next)) {
3016 tcg_gen_goto_tb(0);
3017 tcg_gen_movi_i64(cpu_pc, ctx->base.pc_next);
3018 tcg_gen_exit_tb((uintptr_t)ctx->base.tb);
3019 }
3020 /* FALLTHRU */
3021 case DISAS_PC_STALE:
3022 tcg_gen_movi_i64(cpu_pc, ctx->base.pc_next);
3023 /* FALLTHRU */
3024 case DISAS_PC_UPDATED:
3025 if (!use_exit_tb(ctx)) {
3026 tcg_gen_lookup_and_goto_ptr();
3027 break;
3028 }
3029 /* FALLTHRU */
3030 case DISAS_PC_UPDATED_NOCHAIN:
3031 if (ctx->base.singlestep_enabled) {
3032 gen_excp_1(EXCP_DEBUG, 0);
3033 } else {
3034 tcg_gen_exit_tb(0);
3035 }
3036 break;
3037 default:
3038 g_assert_not_reached();
3039 }
3040 }
3041
3042 static void alpha_tr_disas_log(const DisasContextBase *dcbase, CPUState *cpu)
3043 {
3044 qemu_log("IN: %s\n", lookup_symbol(dcbase->pc_first));
3045 log_target_disas(cpu, dcbase->pc_first, dcbase->tb->size, 1);
3046 }
3047
3048 static const TranslatorOps alpha_tr_ops = {
3049 .init_disas_context = alpha_tr_init_disas_context,
3050 .tb_start = alpha_tr_tb_start,
3051 .insn_start = alpha_tr_insn_start,
3052 .breakpoint_check = alpha_tr_breakpoint_check,
3053 .translate_insn = alpha_tr_translate_insn,
3054 .tb_stop = alpha_tr_tb_stop,
3055 .disas_log = alpha_tr_disas_log,
3056 };
3057
3058 void gen_intermediate_code(CPUState *cpu, TranslationBlock *tb)
3059 {
3060 DisasContext dc;
3061 translator_loop(&alpha_tr_ops, &dc.base, cpu, tb);
3062 }
3063
3064 void restore_state_to_opc(CPUAlphaState *env, TranslationBlock *tb,
3065 target_ulong *data)
3066 {
3067 env->pc = data[0];
3068 }