4 * Copyright (c) 2009 Ulrich Hecht
5 * Copyright (c) 2010 Alexander Graf
7 * This library is free software; you can redistribute it and/or
8 * modify it under the terms of the GNU Lesser General Public
9 * License as published by the Free Software Foundation; either
10 * version 2 of the License, or (at your option) any later version.
12 * This library is distributed in the hope that it will be useful,
13 * but WITHOUT ANY WARRANTY; without even the implied warranty of
14 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
15 * Lesser General Public License for more details.
17 * You should have received a copy of the GNU Lesser General Public
18 * License along with this library; if not, see <http://www.gnu.org/licenses/>.
21 /* #define DEBUG_INLINE_BRANCHES */
22 #define S390X_DEBUG_DISAS
23 /* #define S390X_DEBUG_DISAS_VERBOSE */
25 #ifdef S390X_DEBUG_DISAS_VERBOSE
26 # define LOG_DISAS(...) qemu_log(__VA_ARGS__)
28 # define LOG_DISAS(...) do { } while (0)
32 #include "disas/disas.h"
35 #include "qemu/host-utils.h"
37 /* global register indexes */
38 static TCGv_ptr cpu_env
;
40 #include "exec/gen-icount.h"
46 /* Information that (most) every instruction needs to manipulate. */
47 typedef struct DisasContext DisasContext
;
48 typedef struct DisasInsn DisasInsn
;
49 typedef struct DisasFields DisasFields
;
52 struct TranslationBlock
*tb
;
53 const DisasInsn
*insn
;
57 bool singlestep_enabled
;
61 /* Information carried about a condition to be evaluated. */
68 struct { TCGv_i64 a
, b
; } s64
;
69 struct { TCGv_i32 a
, b
; } s32
;
75 static void gen_op_calc_cc(DisasContext
*s
);
77 #ifdef DEBUG_INLINE_BRANCHES
78 static uint64_t inline_branch_hit
[CC_OP_MAX
];
79 static uint64_t inline_branch_miss
[CC_OP_MAX
];
82 static inline void debug_insn(uint64_t insn
)
84 LOG_DISAS("insn: 0x%" PRIx64
"\n", insn
);
87 static inline uint64_t pc_to_link_info(DisasContext
*s
, uint64_t pc
)
89 if (!(s
->tb
->flags
& FLAG_MASK_64
)) {
90 if (s
->tb
->flags
& FLAG_MASK_32
) {
91 return pc
| 0x80000000;
97 void cpu_dump_state(CPUS390XState
*env
, FILE *f
, fprintf_function cpu_fprintf
,
102 if (env
->cc_op
> 3) {
103 cpu_fprintf(f
, "PSW=mask %016" PRIx64
" addr %016" PRIx64
" cc %15s\n",
104 env
->psw
.mask
, env
->psw
.addr
, cc_name(env
->cc_op
));
106 cpu_fprintf(f
, "PSW=mask %016" PRIx64
" addr %016" PRIx64
" cc %02x\n",
107 env
->psw
.mask
, env
->psw
.addr
, env
->cc_op
);
110 for (i
= 0; i
< 16; i
++) {
111 cpu_fprintf(f
, "R%02d=%016" PRIx64
, i
, env
->regs
[i
]);
113 cpu_fprintf(f
, "\n");
119 for (i
= 0; i
< 16; i
++) {
120 cpu_fprintf(f
, "F%02d=%016" PRIx64
, i
, env
->fregs
[i
].ll
);
122 cpu_fprintf(f
, "\n");
128 #ifndef CONFIG_USER_ONLY
129 for (i
= 0; i
< 16; i
++) {
130 cpu_fprintf(f
, "C%02d=%016" PRIx64
, i
, env
->cregs
[i
]);
132 cpu_fprintf(f
, "\n");
139 #ifdef DEBUG_INLINE_BRANCHES
140 for (i
= 0; i
< CC_OP_MAX
; i
++) {
141 cpu_fprintf(f
, " %15s = %10ld\t%10ld\n", cc_name(i
),
142 inline_branch_miss
[i
], inline_branch_hit
[i
]);
146 cpu_fprintf(f
, "\n");
149 static TCGv_i64 psw_addr
;
150 static TCGv_i64 psw_mask
;
152 static TCGv_i32 cc_op
;
153 static TCGv_i64 cc_src
;
154 static TCGv_i64 cc_dst
;
155 static TCGv_i64 cc_vr
;
157 static char cpu_reg_names
[32][4];
158 static TCGv_i64 regs
[16];
159 static TCGv_i64 fregs
[16];
161 static uint8_t gen_opc_cc_op
[OPC_BUF_SIZE
];
163 void s390x_translate_init(void)
167 cpu_env
= tcg_global_reg_new_ptr(TCG_AREG0
, "env");
168 psw_addr
= tcg_global_mem_new_i64(TCG_AREG0
,
169 offsetof(CPUS390XState
, psw
.addr
),
171 psw_mask
= tcg_global_mem_new_i64(TCG_AREG0
,
172 offsetof(CPUS390XState
, psw
.mask
),
175 cc_op
= tcg_global_mem_new_i32(TCG_AREG0
, offsetof(CPUS390XState
, cc_op
),
177 cc_src
= tcg_global_mem_new_i64(TCG_AREG0
, offsetof(CPUS390XState
, cc_src
),
179 cc_dst
= tcg_global_mem_new_i64(TCG_AREG0
, offsetof(CPUS390XState
, cc_dst
),
181 cc_vr
= tcg_global_mem_new_i64(TCG_AREG0
, offsetof(CPUS390XState
, cc_vr
),
184 for (i
= 0; i
< 16; i
++) {
185 snprintf(cpu_reg_names
[i
], sizeof(cpu_reg_names
[0]), "r%d", i
);
186 regs
[i
] = tcg_global_mem_new(TCG_AREG0
,
187 offsetof(CPUS390XState
, regs
[i
]),
191 for (i
= 0; i
< 16; i
++) {
192 snprintf(cpu_reg_names
[i
+ 16], sizeof(cpu_reg_names
[0]), "f%d", i
);
193 fregs
[i
] = tcg_global_mem_new(TCG_AREG0
,
194 offsetof(CPUS390XState
, fregs
[i
].d
),
195 cpu_reg_names
[i
+ 16]);
198 /* register helpers */
203 static inline TCGv_i64
load_reg(int reg
)
205 TCGv_i64 r
= tcg_temp_new_i64();
206 tcg_gen_mov_i64(r
, regs
[reg
]);
210 static inline TCGv_i64
load_freg(int reg
)
212 TCGv_i64 r
= tcg_temp_new_i64();
213 tcg_gen_mov_i64(r
, fregs
[reg
]);
217 static inline TCGv_i32
load_freg32(int reg
)
219 TCGv_i32 r
= tcg_temp_new_i32();
220 #if HOST_LONG_BITS == 32
221 tcg_gen_mov_i32(r
, TCGV_HIGH(fregs
[reg
]));
223 tcg_gen_shri_i64(MAKE_TCGV_I64(GET_TCGV_I32(r
)), fregs
[reg
], 32);
228 static inline TCGv_i64
load_freg32_i64(int reg
)
230 TCGv_i64 r
= tcg_temp_new_i64();
231 tcg_gen_shri_i64(r
, fregs
[reg
], 32);
235 static inline TCGv_i32
load_reg32(int reg
)
237 TCGv_i32 r
= tcg_temp_new_i32();
238 tcg_gen_trunc_i64_i32(r
, regs
[reg
]);
242 static inline TCGv_i64
load_reg32_i64(int reg
)
244 TCGv_i64 r
= tcg_temp_new_i64();
245 tcg_gen_ext32s_i64(r
, regs
[reg
]);
249 static inline void store_reg(int reg
, TCGv_i64 v
)
251 tcg_gen_mov_i64(regs
[reg
], v
);
254 static inline void store_freg(int reg
, TCGv_i64 v
)
256 tcg_gen_mov_i64(fregs
[reg
], v
);
259 static inline void store_reg32(int reg
, TCGv_i32 v
)
261 /* 32 bit register writes keep the upper half */
262 #if HOST_LONG_BITS == 32
263 tcg_gen_mov_i32(TCGV_LOW(regs
[reg
]), v
);
265 tcg_gen_deposit_i64(regs
[reg
], regs
[reg
],
266 MAKE_TCGV_I64(GET_TCGV_I32(v
)), 0, 32);
270 static inline void store_reg32_i64(int reg
, TCGv_i64 v
)
272 /* 32 bit register writes keep the upper half */
273 tcg_gen_deposit_i64(regs
[reg
], regs
[reg
], v
, 0, 32);
276 static inline void store_reg32h_i64(int reg
, TCGv_i64 v
)
278 tcg_gen_deposit_i64(regs
[reg
], regs
[reg
], v
, 32, 32);
281 static inline void store_freg32(int reg
, TCGv_i32 v
)
283 /* 32 bit register writes keep the lower half */
284 #if HOST_LONG_BITS == 32
285 tcg_gen_mov_i32(TCGV_HIGH(fregs
[reg
]), v
);
287 tcg_gen_deposit_i64(fregs
[reg
], fregs
[reg
],
288 MAKE_TCGV_I64(GET_TCGV_I32(v
)), 32, 32);
292 static inline void store_freg32_i64(int reg
, TCGv_i64 v
)
294 tcg_gen_deposit_i64(fregs
[reg
], fregs
[reg
], v
, 32, 32);
297 static inline void return_low128(TCGv_i64 dest
)
299 tcg_gen_ld_i64(dest
, cpu_env
, offsetof(CPUS390XState
, retxl
));
302 static inline void update_psw_addr(DisasContext
*s
)
305 tcg_gen_movi_i64(psw_addr
, s
->pc
);
308 static inline void potential_page_fault(DisasContext
*s
)
310 #ifndef CONFIG_USER_ONLY
316 static inline uint64_t ld_code2(CPUS390XState
*env
, uint64_t pc
)
318 return (uint64_t)cpu_lduw_code(env
, pc
);
321 static inline uint64_t ld_code4(CPUS390XState
*env
, uint64_t pc
)
323 return (uint64_t)(uint32_t)cpu_ldl_code(env
, pc
);
326 static inline uint64_t ld_code6(CPUS390XState
*env
, uint64_t pc
)
328 return (ld_code2(env
, pc
) << 32) | ld_code4(env
, pc
+ 2);
331 static inline int get_mem_index(DisasContext
*s
)
333 switch (s
->tb
->flags
& FLAG_MASK_ASC
) {
334 case PSW_ASC_PRIMARY
>> 32:
336 case PSW_ASC_SECONDARY
>> 32:
338 case PSW_ASC_HOME
>> 32:
346 static void gen_exception(int excp
)
348 TCGv_i32 tmp
= tcg_const_i32(excp
);
349 gen_helper_exception(cpu_env
, tmp
);
350 tcg_temp_free_i32(tmp
);
353 static void gen_program_exception(DisasContext
*s
, int code
)
357 /* Remember what pgm exeption this was. */
358 tmp
= tcg_const_i32(code
);
359 tcg_gen_st_i32(tmp
, cpu_env
, offsetof(CPUS390XState
, int_pgm_code
));
360 tcg_temp_free_i32(tmp
);
362 tmp
= tcg_const_i32(s
->next_pc
- s
->pc
);
363 tcg_gen_st_i32(tmp
, cpu_env
, offsetof(CPUS390XState
, int_pgm_ilen
));
364 tcg_temp_free_i32(tmp
);
366 /* Advance past instruction. */
373 /* Trigger exception. */
374 gen_exception(EXCP_PGM
);
377 s
->is_jmp
= DISAS_EXCP
;
380 static inline void gen_illegal_opcode(DisasContext
*s
)
382 gen_program_exception(s
, PGM_SPECIFICATION
);
385 static inline void check_privileged(DisasContext
*s
)
387 if (s
->tb
->flags
& (PSW_MASK_PSTATE
>> 32)) {
388 gen_program_exception(s
, PGM_PRIVILEGED
);
392 static TCGv_i64
get_address(DisasContext
*s
, int x2
, int b2
, int d2
)
396 /* 31-bitify the immediate part; register contents are dealt with below */
397 if (!(s
->tb
->flags
& FLAG_MASK_64
)) {
403 tmp
= tcg_const_i64(d2
);
404 tcg_gen_add_i64(tmp
, tmp
, regs
[x2
]);
409 tcg_gen_add_i64(tmp
, tmp
, regs
[b2
]);
413 tmp
= tcg_const_i64(d2
);
414 tcg_gen_add_i64(tmp
, tmp
, regs
[b2
]);
419 tmp
= tcg_const_i64(d2
);
422 /* 31-bit mode mask if there are values loaded from registers */
423 if (!(s
->tb
->flags
& FLAG_MASK_64
) && (x2
|| b2
)) {
424 tcg_gen_andi_i64(tmp
, tmp
, 0x7fffffffUL
);
430 static inline void gen_op_movi_cc(DisasContext
*s
, uint32_t val
)
432 s
->cc_op
= CC_OP_CONST0
+ val
;
435 static void gen_op_update1_cc_i64(DisasContext
*s
, enum cc_op op
, TCGv_i64 dst
)
437 tcg_gen_discard_i64(cc_src
);
438 tcg_gen_mov_i64(cc_dst
, dst
);
439 tcg_gen_discard_i64(cc_vr
);
443 static void gen_op_update1_cc_i32(DisasContext
*s
, enum cc_op op
, TCGv_i32 dst
)
445 tcg_gen_discard_i64(cc_src
);
446 tcg_gen_extu_i32_i64(cc_dst
, dst
);
447 tcg_gen_discard_i64(cc_vr
);
451 static void gen_op_update2_cc_i64(DisasContext
*s
, enum cc_op op
, TCGv_i64 src
,
454 tcg_gen_mov_i64(cc_src
, src
);
455 tcg_gen_mov_i64(cc_dst
, dst
);
456 tcg_gen_discard_i64(cc_vr
);
460 static void gen_op_update2_cc_i32(DisasContext
*s
, enum cc_op op
, TCGv_i32 src
,
463 tcg_gen_extu_i32_i64(cc_src
, src
);
464 tcg_gen_extu_i32_i64(cc_dst
, dst
);
465 tcg_gen_discard_i64(cc_vr
);
469 static void gen_op_update3_cc_i64(DisasContext
*s
, enum cc_op op
, TCGv_i64 src
,
470 TCGv_i64 dst
, TCGv_i64 vr
)
472 tcg_gen_mov_i64(cc_src
, src
);
473 tcg_gen_mov_i64(cc_dst
, dst
);
474 tcg_gen_mov_i64(cc_vr
, vr
);
478 static inline void set_cc_nz_u32(DisasContext
*s
, TCGv_i32 val
)
480 gen_op_update1_cc_i32(s
, CC_OP_NZ
, val
);
483 static inline void set_cc_nz_u64(DisasContext
*s
, TCGv_i64 val
)
485 gen_op_update1_cc_i64(s
, CC_OP_NZ
, val
);
488 static inline void gen_set_cc_nz_f32(DisasContext
*s
, TCGv_i64 val
)
490 gen_op_update1_cc_i64(s
, CC_OP_NZ_F32
, val
);
493 static inline void gen_set_cc_nz_f64(DisasContext
*s
, TCGv_i64 val
)
495 gen_op_update1_cc_i64(s
, CC_OP_NZ_F64
, val
);
498 static inline void gen_set_cc_nz_f128(DisasContext
*s
, TCGv_i64 vh
, TCGv_i64 vl
)
500 gen_op_update2_cc_i64(s
, CC_OP_NZ_F128
, vh
, vl
);
503 static inline void cmp_32(DisasContext
*s
, TCGv_i32 v1
, TCGv_i32 v2
,
506 gen_op_update2_cc_i32(s
, cond
, v1
, v2
);
509 static inline void cmp_64(DisasContext
*s
, TCGv_i64 v1
, TCGv_i64 v2
,
512 gen_op_update2_cc_i64(s
, cond
, v1
, v2
);
515 static inline void cmp_s32(DisasContext
*s
, TCGv_i32 v1
, TCGv_i32 v2
)
517 cmp_32(s
, v1
, v2
, CC_OP_LTGT_32
);
520 static inline void cmp_u32(DisasContext
*s
, TCGv_i32 v1
, TCGv_i32 v2
)
522 cmp_32(s
, v1
, v2
, CC_OP_LTUGTU_32
);
525 static inline void cmp_s32c(DisasContext
*s
, TCGv_i32 v1
, int32_t v2
)
527 /* XXX optimize for the constant? put it in s? */
528 TCGv_i32 tmp
= tcg_const_i32(v2
);
529 cmp_32(s
, v1
, tmp
, CC_OP_LTGT_32
);
530 tcg_temp_free_i32(tmp
);
533 static inline void cmp_u32c(DisasContext
*s
, TCGv_i32 v1
, uint32_t v2
)
535 TCGv_i32 tmp
= tcg_const_i32(v2
);
536 cmp_32(s
, v1
, tmp
, CC_OP_LTUGTU_32
);
537 tcg_temp_free_i32(tmp
);
540 static inline void cmp_s64(DisasContext
*s
, TCGv_i64 v1
, TCGv_i64 v2
)
542 cmp_64(s
, v1
, v2
, CC_OP_LTGT_64
);
545 static inline void cmp_u64(DisasContext
*s
, TCGv_i64 v1
, TCGv_i64 v2
)
547 cmp_64(s
, v1
, v2
, CC_OP_LTUGTU_64
);
550 static inline void cmp_s64c(DisasContext
*s
, TCGv_i64 v1
, int64_t v2
)
552 TCGv_i64 tmp
= tcg_const_i64(v2
);
554 tcg_temp_free_i64(tmp
);
557 static inline void cmp_u64c(DisasContext
*s
, TCGv_i64 v1
, uint64_t v2
)
559 TCGv_i64 tmp
= tcg_const_i64(v2
);
561 tcg_temp_free_i64(tmp
);
564 static inline void set_cc_s32(DisasContext
*s
, TCGv_i32 val
)
566 gen_op_update1_cc_i32(s
, CC_OP_LTGT0_32
, val
);
569 static inline void set_cc_s64(DisasContext
*s
, TCGv_i64 val
)
571 gen_op_update1_cc_i64(s
, CC_OP_LTGT0_64
, val
);
574 /* CC value is in env->cc_op */
575 static inline void set_cc_static(DisasContext
*s
)
577 tcg_gen_discard_i64(cc_src
);
578 tcg_gen_discard_i64(cc_dst
);
579 tcg_gen_discard_i64(cc_vr
);
580 s
->cc_op
= CC_OP_STATIC
;
583 static inline void gen_op_set_cc_op(DisasContext
*s
)
585 if (s
->cc_op
!= CC_OP_DYNAMIC
&& s
->cc_op
!= CC_OP_STATIC
) {
586 tcg_gen_movi_i32(cc_op
, s
->cc_op
);
590 static inline void gen_update_cc_op(DisasContext
*s
)
595 /* calculates cc into cc_op */
596 static void gen_op_calc_cc(DisasContext
*s
)
598 TCGv_i32 local_cc_op
= tcg_const_i32(s
->cc_op
);
599 TCGv_i64 dummy
= tcg_const_i64(0);
606 /* s->cc_op is the cc value */
607 tcg_gen_movi_i32(cc_op
, s
->cc_op
- CC_OP_CONST0
);
610 /* env->cc_op already is the cc value */
625 gen_helper_calc_cc(cc_op
, cpu_env
, local_cc_op
, dummy
, cc_dst
, dummy
);
630 case CC_OP_LTUGTU_32
:
631 case CC_OP_LTUGTU_64
:
638 gen_helper_calc_cc(cc_op
, cpu_env
, local_cc_op
, cc_src
, cc_dst
, dummy
);
653 gen_helper_calc_cc(cc_op
, cpu_env
, local_cc_op
, cc_src
, cc_dst
, cc_vr
);
656 /* unknown operation - assume 3 arguments and cc_op in env */
657 gen_helper_calc_cc(cc_op
, cpu_env
, cc_op
, cc_src
, cc_dst
, cc_vr
);
663 tcg_temp_free_i32(local_cc_op
);
664 tcg_temp_free_i64(dummy
);
666 /* We now have cc in cc_op as constant */
670 static inline void decode_rr(DisasContext
*s
, uint64_t insn
, int *r1
, int *r2
)
674 *r1
= (insn
>> 4) & 0xf;
678 static inline TCGv_i64
decode_rx(DisasContext
*s
, uint64_t insn
, int *r1
,
679 int *x2
, int *b2
, int *d2
)
683 *r1
= (insn
>> 20) & 0xf;
684 *x2
= (insn
>> 16) & 0xf;
685 *b2
= (insn
>> 12) & 0xf;
688 return get_address(s
, *x2
, *b2
, *d2
);
691 static inline void decode_rs(DisasContext
*s
, uint64_t insn
, int *r1
, int *r3
,
696 *r1
= (insn
>> 20) & 0xf;
698 *r3
= (insn
>> 16) & 0xf;
699 *b2
= (insn
>> 12) & 0xf;
703 static inline TCGv_i64
decode_si(DisasContext
*s
, uint64_t insn
, int *i2
,
708 *i2
= (insn
>> 16) & 0xff;
709 *b1
= (insn
>> 12) & 0xf;
712 return get_address(s
, 0, *b1
, *d1
);
715 static int use_goto_tb(DisasContext
*s
, uint64_t dest
)
717 /* NOTE: we handle the case where the TB spans two pages here */
718 return (((dest
& TARGET_PAGE_MASK
) == (s
->tb
->pc
& TARGET_PAGE_MASK
)
719 || (dest
& TARGET_PAGE_MASK
) == ((s
->pc
- 1) & TARGET_PAGE_MASK
))
720 && !s
->singlestep_enabled
721 && !(s
->tb
->cflags
& CF_LAST_IO
));
724 static inline void gen_goto_tb(DisasContext
*s
, int tb_num
, target_ulong pc
)
728 if (use_goto_tb(s
, pc
)) {
729 tcg_gen_goto_tb(tb_num
);
730 tcg_gen_movi_i64(psw_addr
, pc
);
731 tcg_gen_exit_tb((tcg_target_long
)s
->tb
+ tb_num
);
733 /* jump to another page: currently not optimized */
734 tcg_gen_movi_i64(psw_addr
, pc
);
739 static inline void account_noninline_branch(DisasContext
*s
, int cc_op
)
741 #ifdef DEBUG_INLINE_BRANCHES
742 inline_branch_miss
[cc_op
]++;
746 static inline void account_inline_branch(DisasContext
*s
, int cc_op
)
748 #ifdef DEBUG_INLINE_BRANCHES
749 inline_branch_hit
[cc_op
]++;
753 /* Table of mask values to comparison codes, given a comparison as input.
754 For a true comparison CC=3 will never be set, but we treat this
755 conservatively for possible use when CC=3 indicates overflow. */
756 static const TCGCond ltgt_cond
[16] = {
757 TCG_COND_NEVER
, TCG_COND_NEVER
, /* | | | x */
758 TCG_COND_GT
, TCG_COND_NEVER
, /* | | GT | x */
759 TCG_COND_LT
, TCG_COND_NEVER
, /* | LT | | x */
760 TCG_COND_NE
, TCG_COND_NEVER
, /* | LT | GT | x */
761 TCG_COND_EQ
, TCG_COND_NEVER
, /* EQ | | | x */
762 TCG_COND_GE
, TCG_COND_NEVER
, /* EQ | | GT | x */
763 TCG_COND_LE
, TCG_COND_NEVER
, /* EQ | LT | | x */
764 TCG_COND_ALWAYS
, TCG_COND_ALWAYS
, /* EQ | LT | GT | x */
767 /* Table of mask values to comparison codes, given a logic op as input.
768 For such, only CC=0 and CC=1 should be possible. */
769 static const TCGCond nz_cond
[16] = {
771 TCG_COND_NEVER
, TCG_COND_NEVER
, TCG_COND_NEVER
, TCG_COND_NEVER
,
773 TCG_COND_NE
, TCG_COND_NE
, TCG_COND_NE
, TCG_COND_NE
,
775 TCG_COND_EQ
, TCG_COND_EQ
, TCG_COND_EQ
, TCG_COND_EQ
,
776 /* EQ | NE | x | x */
777 TCG_COND_ALWAYS
, TCG_COND_ALWAYS
, TCG_COND_ALWAYS
, TCG_COND_ALWAYS
,
780 /* Interpret MASK in terms of S->CC_OP, and fill in C with all the
781 details required to generate a TCG comparison. */
782 static void disas_jcc(DisasContext
*s
, DisasCompare
*c
, uint32_t mask
)
785 enum cc_op old_cc_op
= s
->cc_op
;
787 if (mask
== 15 || mask
== 0) {
788 c
->cond
= (mask
? TCG_COND_ALWAYS
: TCG_COND_NEVER
);
791 c
->g1
= c
->g2
= true;
796 /* Find the TCG condition for the mask + cc op. */
802 cond
= ltgt_cond
[mask
];
803 if (cond
== TCG_COND_NEVER
) {
806 account_inline_branch(s
, old_cc_op
);
809 case CC_OP_LTUGTU_32
:
810 case CC_OP_LTUGTU_64
:
811 cond
= tcg_unsigned_cond(ltgt_cond
[mask
]);
812 if (cond
== TCG_COND_NEVER
) {
815 account_inline_branch(s
, old_cc_op
);
819 cond
= nz_cond
[mask
];
820 if (cond
== TCG_COND_NEVER
) {
823 account_inline_branch(s
, old_cc_op
);
838 account_inline_branch(s
, old_cc_op
);
853 account_inline_branch(s
, old_cc_op
);
857 switch (mask
& 0xa) {
858 case 8: /* src == 0 -> no one bit found */
861 case 2: /* src != 0 -> one bit found */
867 account_inline_branch(s
, old_cc_op
);
872 /* Calculate cc value. */
877 /* Jump based on CC. We'll load up the real cond below;
878 the assignment here merely avoids a compiler warning. */
879 account_noninline_branch(s
, old_cc_op
);
880 old_cc_op
= CC_OP_STATIC
;
881 cond
= TCG_COND_NEVER
;
885 /* Load up the arguments of the comparison. */
887 c
->g1
= c
->g2
= false;
891 c
->u
.s32
.a
= tcg_temp_new_i32();
892 tcg_gen_trunc_i64_i32(c
->u
.s32
.a
, cc_dst
);
893 c
->u
.s32
.b
= tcg_const_i32(0);
896 case CC_OP_LTUGTU_32
:
898 c
->u
.s32
.a
= tcg_temp_new_i32();
899 tcg_gen_trunc_i64_i32(c
->u
.s32
.a
, cc_src
);
900 c
->u
.s32
.b
= tcg_temp_new_i32();
901 tcg_gen_trunc_i64_i32(c
->u
.s32
.b
, cc_dst
);
908 c
->u
.s64
.b
= tcg_const_i64(0);
912 case CC_OP_LTUGTU_64
:
915 c
->g1
= c
->g2
= true;
921 c
->u
.s64
.a
= tcg_temp_new_i64();
922 c
->u
.s64
.b
= tcg_const_i64(0);
923 tcg_gen_and_i64(c
->u
.s64
.a
, cc_src
, cc_dst
);
931 case 0x8 | 0x4 | 0x2: /* cc != 3 */
933 c
->u
.s32
.b
= tcg_const_i32(3);
935 case 0x8 | 0x4 | 0x1: /* cc != 2 */
937 c
->u
.s32
.b
= tcg_const_i32(2);
939 case 0x8 | 0x2 | 0x1: /* cc != 1 */
941 c
->u
.s32
.b
= tcg_const_i32(1);
943 case 0x8 | 0x2: /* cc == 0 ||Â cc == 2 => (cc & 1) == 0 */
946 c
->u
.s32
.a
= tcg_temp_new_i32();
947 c
->u
.s32
.b
= tcg_const_i32(0);
948 tcg_gen_andi_i32(c
->u
.s32
.a
, cc_op
, 1);
950 case 0x8 | 0x4: /* cc < 2 */
952 c
->u
.s32
.b
= tcg_const_i32(2);
954 case 0x8: /* cc == 0 */
956 c
->u
.s32
.b
= tcg_const_i32(0);
958 case 0x4 | 0x2 | 0x1: /* cc != 0 */
960 c
->u
.s32
.b
= tcg_const_i32(0);
962 case 0x4 | 0x1: /* cc == 1 ||Â cc == 3 => (cc & 1) != 0 */
965 c
->u
.s32
.a
= tcg_temp_new_i32();
966 c
->u
.s32
.b
= tcg_const_i32(0);
967 tcg_gen_andi_i32(c
->u
.s32
.a
, cc_op
, 1);
969 case 0x4: /* cc == 1 */
971 c
->u
.s32
.b
= tcg_const_i32(1);
973 case 0x2 | 0x1: /* cc > 1 */
975 c
->u
.s32
.b
= tcg_const_i32(1);
977 case 0x2: /* cc == 2 */
979 c
->u
.s32
.b
= tcg_const_i32(2);
981 case 0x1: /* cc == 3 */
983 c
->u
.s32
.b
= tcg_const_i32(3);
986 /* CC is masked by something else: (8 >> cc) & mask. */
989 c
->u
.s32
.a
= tcg_const_i32(8);
990 c
->u
.s32
.b
= tcg_const_i32(0);
991 tcg_gen_shr_i32(c
->u
.s32
.a
, c
->u
.s32
.a
, cc_op
);
992 tcg_gen_andi_i32(c
->u
.s32
.a
, c
->u
.s32
.a
, mask
);
1003 static void free_compare(DisasCompare
*c
)
1007 tcg_temp_free_i64(c
->u
.s64
.a
);
1009 tcg_temp_free_i32(c
->u
.s32
.a
);
1014 tcg_temp_free_i64(c
->u
.s64
.b
);
1016 tcg_temp_free_i32(c
->u
.s32
.b
);
1021 static void disas_b2(CPUS390XState
*env
, DisasContext
*s
, int op
,
1024 #ifndef CONFIG_USER_ONLY
1025 TCGv_i64 tmp
, tmp2
, tmp3
;
1026 TCGv_i32 tmp32_1
, tmp32_2
;
1030 r1
= (insn
>> 4) & 0xf;
1033 LOG_DISAS("disas_b2: op 0x%x r1 %d r2 %d\n", op
, r1
, r2
);
1036 case 0x78: /* STCKE D2(B2) [S] */
1037 /* Store Clock Extended */
1038 decode_rs(s
, insn
, &r1
, &r3
, &b2
, &d2
);
1039 tmp
= get_address(s
, 0, b2
, d2
);
1040 potential_page_fault(s
);
1041 gen_helper_stcke(cc_op
, cpu_env
, tmp
);
1043 tcg_temp_free_i64(tmp
);
1045 case 0x79: /* SACF D2(B2) [S] */
1046 /* Set Address Space Control Fast */
1047 check_privileged(s
);
1048 decode_rs(s
, insn
, &r1
, &r3
, &b2
, &d2
);
1049 tmp
= get_address(s
, 0, b2
, d2
);
1050 potential_page_fault(s
);
1051 gen_helper_sacf(cpu_env
, tmp
);
1052 tcg_temp_free_i64(tmp
);
1053 /* addressing mode has changed, so end the block */
1056 s
->is_jmp
= DISAS_JUMP
;
1058 case 0x7d: /* STSI D2,(B2) [S] */
1059 check_privileged(s
);
1060 decode_rs(s
, insn
, &r1
, &r3
, &b2
, &d2
);
1061 tmp
= get_address(s
, 0, b2
, d2
);
1062 tmp32_1
= load_reg32(0);
1063 tmp32_2
= load_reg32(1);
1064 potential_page_fault(s
);
1065 gen_helper_stsi(cc_op
, cpu_env
, tmp
, tmp32_1
, tmp32_2
);
1067 tcg_temp_free_i64(tmp
);
1068 tcg_temp_free_i32(tmp32_1
);
1069 tcg_temp_free_i32(tmp32_2
);
1071 case 0xb1: /* STFL D2(B2) [S] */
1072 /* Store Facility List (CPU features) at 200 */
1073 check_privileged(s
);
1074 tmp2
= tcg_const_i64(0xc0000000);
1075 tmp
= tcg_const_i64(200);
1076 tcg_gen_qemu_st32(tmp2
, tmp
, get_mem_index(s
));
1077 tcg_temp_free_i64(tmp2
);
1078 tcg_temp_free_i64(tmp
);
1080 case 0xb2: /* LPSWE D2(B2) [S] */
1081 /* Load PSW Extended */
1082 check_privileged(s
);
1083 decode_rs(s
, insn
, &r1
, &r3
, &b2
, &d2
);
1084 tmp
= get_address(s
, 0, b2
, d2
);
1085 tmp2
= tcg_temp_new_i64();
1086 tmp3
= tcg_temp_new_i64();
1087 tcg_gen_qemu_ld64(tmp2
, tmp
, get_mem_index(s
));
1088 tcg_gen_addi_i64(tmp
, tmp
, 8);
1089 tcg_gen_qemu_ld64(tmp3
, tmp
, get_mem_index(s
));
1090 gen_helper_load_psw(cpu_env
, tmp2
, tmp3
);
1091 /* we need to keep cc_op intact */
1092 s
->is_jmp
= DISAS_JUMP
;
1093 tcg_temp_free_i64(tmp
);
1094 tcg_temp_free_i64(tmp2
);
1095 tcg_temp_free_i64(tmp3
);
1097 case 0x20: /* SERVC R1,R2 [RRE] */
1098 /* SCLP Service call (PV hypercall) */
1099 check_privileged(s
);
1100 potential_page_fault(s
);
1101 tmp32_1
= load_reg32(r2
);
1103 gen_helper_servc(cc_op
, cpu_env
, tmp32_1
, tmp
);
1105 tcg_temp_free_i32(tmp32_1
);
1106 tcg_temp_free_i64(tmp
);
1110 LOG_DISAS("illegal b2 operation 0x%x\n", op
);
1111 gen_illegal_opcode(s
);
1112 #ifndef CONFIG_USER_ONLY
1118 static void disas_s390_insn(CPUS390XState
*env
, DisasContext
*s
)
1124 opc
= cpu_ldub_code(env
, s
->pc
);
1125 LOG_DISAS("opc 0x%x\n", opc
);
1129 insn
= ld_code4(env
, s
->pc
);
1130 op
= (insn
>> 16) & 0xff;
1131 disas_b2(env
, s
, op
, insn
);
1134 qemu_log_mask(LOG_UNIMP
, "unimplemented opcode 0x%x\n", opc
);
1135 gen_illegal_opcode(s
);
1140 /* ====================================================================== */
1141 /* Define the insn format enumeration. */
1142 #define F0(N) FMT_##N,
1143 #define F1(N, X1) F0(N)
1144 #define F2(N, X1, X2) F0(N)
1145 #define F3(N, X1, X2, X3) F0(N)
1146 #define F4(N, X1, X2, X3, X4) F0(N)
1147 #define F5(N, X1, X2, X3, X4, X5) F0(N)
1150 #include "insn-format.def"
1160 /* Define a structure to hold the decoded fields. We'll store each inside
1161 an array indexed by an enum. In order to conserve memory, we'll arrange
1162 for fields that do not exist at the same time to overlap, thus the "C"
1163 for compact. For checking purposes there is an "O" for original index
1164 as well that will be applied to availability bitmaps. */
1166 enum DisasFieldIndexO
{
1189 enum DisasFieldIndexC
{
1220 struct DisasFields
{
1223 unsigned presentC
:16;
1224 unsigned int presentO
;
1228 /* This is the way fields are to be accessed out of DisasFields. */
1229 #define have_field(S, F) have_field1((S), FLD_O_##F)
1230 #define get_field(S, F) get_field1((S), FLD_O_##F, FLD_C_##F)
1232 static bool have_field1(const DisasFields
*f
, enum DisasFieldIndexO c
)
1234 return (f
->presentO
>> c
) & 1;
1237 static int get_field1(const DisasFields
*f
, enum DisasFieldIndexO o
,
1238 enum DisasFieldIndexC c
)
1240 assert(have_field1(f
, o
));
1244 /* Describe the layout of each field in each format. */
1245 typedef struct DisasField
{
1247 unsigned int size
:8;
1248 unsigned int type
:2;
1249 unsigned int indexC
:6;
1250 enum DisasFieldIndexO indexO
:8;
1253 typedef struct DisasFormatInfo
{
1254 DisasField op
[NUM_C_FIELD
];
1257 #define R(N, B) { B, 4, 0, FLD_C_r##N, FLD_O_r##N }
1258 #define M(N, B) { B, 4, 0, FLD_C_m##N, FLD_O_m##N }
1259 #define BD(N, BB, BD) { BB, 4, 0, FLD_C_b##N, FLD_O_b##N }, \
1260 { BD, 12, 0, FLD_C_d##N, FLD_O_d##N }
1261 #define BXD(N) { 16, 4, 0, FLD_C_b##N, FLD_O_b##N }, \
1262 { 12, 4, 0, FLD_C_x##N, FLD_O_x##N }, \
1263 { 20, 12, 0, FLD_C_d##N, FLD_O_d##N }
1264 #define BDL(N) { 16, 4, 0, FLD_C_b##N, FLD_O_b##N }, \
1265 { 20, 20, 2, FLD_C_d##N, FLD_O_d##N }
1266 #define BXDL(N) { 16, 4, 0, FLD_C_b##N, FLD_O_b##N }, \
1267 { 12, 4, 0, FLD_C_x##N, FLD_O_x##N }, \
1268 { 20, 20, 2, FLD_C_d##N, FLD_O_d##N }
1269 #define I(N, B, S) { B, S, 1, FLD_C_i##N, FLD_O_i##N }
1270 #define L(N, B, S) { B, S, 0, FLD_C_l##N, FLD_O_l##N }
1272 #define F0(N) { { } },
1273 #define F1(N, X1) { { X1 } },
1274 #define F2(N, X1, X2) { { X1, X2 } },
1275 #define F3(N, X1, X2, X3) { { X1, X2, X3 } },
1276 #define F4(N, X1, X2, X3, X4) { { X1, X2, X3, X4 } },
1277 #define F5(N, X1, X2, X3, X4, X5) { { X1, X2, X3, X4, X5 } },
1279 static const DisasFormatInfo format_info
[] = {
1280 #include "insn-format.def"
1298 /* Generally, we'll extract operands into this structures, operate upon
1299 them, and store them back. See the "in1", "in2", "prep", "wout" sets
1300 of routines below for more details. */
1302 bool g_out
, g_out2
, g_in1
, g_in2
;
1303 TCGv_i64 out
, out2
, in1
, in2
;
1307 /* Return values from translate_one, indicating the state of the TB. */
1309 /* Continue the TB. */
1311 /* We have emitted one or more goto_tb. No fixup required. */
1313 /* We are not using a goto_tb (for whatever reason), but have updated
1314 the PC (for whatever reason), so there's no need to do it again on
1317 /* We are exiting the TB, but have neither emitted a goto_tb, nor
1318 updated the PC for the next instruction to be executed. */
1320 /* We are ending the TB with a noreturn function call, e.g. longjmp.
1321 No following code will be executed. */
1325 typedef enum DisasFacility
{
1326 FAC_Z
, /* zarch (default) */
1327 FAC_CASS
, /* compare and swap and store */
1328 FAC_CASS2
, /* compare and swap and store 2*/
1329 FAC_DFP
, /* decimal floating point */
1330 FAC_DFPR
, /* decimal floating point rounding */
1331 FAC_DO
, /* distinct operands */
1332 FAC_EE
, /* execute extensions */
1333 FAC_EI
, /* extended immediate */
1334 FAC_FPE
, /* floating point extension */
1335 FAC_FPSSH
, /* floating point support sign handling */
1336 FAC_FPRGR
, /* FPR-GR transfer */
1337 FAC_GIE
, /* general instructions extension */
1338 FAC_HFP_MA
, /* HFP multiply-and-add/subtract */
1339 FAC_HW
, /* high-word */
1340 FAC_IEEEE_SIM
, /* IEEE exception sumilation */
1341 FAC_LOC
, /* load/store on condition */
1342 FAC_LD
, /* long displacement */
1343 FAC_PC
, /* population count */
1344 FAC_SCF
, /* store clock fast */
1345 FAC_SFLE
, /* store facility list extended */
1351 DisasFacility fac
:6;
1355 void (*help_in1
)(DisasContext
*, DisasFields
*, DisasOps
*);
1356 void (*help_in2
)(DisasContext
*, DisasFields
*, DisasOps
*);
1357 void (*help_prep
)(DisasContext
*, DisasFields
*, DisasOps
*);
1358 void (*help_wout
)(DisasContext
*, DisasFields
*, DisasOps
*);
1359 void (*help_cout
)(DisasContext
*, DisasOps
*);
1360 ExitStatus (*help_op
)(DisasContext
*, DisasOps
*);
1365 /* ====================================================================== */
1366 /* Miscelaneous helpers, used by several operations. */
1368 static void help_l2_shift(DisasContext
*s
, DisasFields
*f
,
1369 DisasOps
*o
, int mask
)
1371 int b2
= get_field(f
, b2
);
1372 int d2
= get_field(f
, d2
);
1375 o
->in2
= tcg_const_i64(d2
& mask
);
1377 o
->in2
= get_address(s
, 0, b2
, d2
);
1378 tcg_gen_andi_i64(o
->in2
, o
->in2
, mask
);
1382 static ExitStatus
help_goto_direct(DisasContext
*s
, uint64_t dest
)
1384 if (dest
== s
->next_pc
) {
1387 if (use_goto_tb(s
, dest
)) {
1388 gen_update_cc_op(s
);
1390 tcg_gen_movi_i64(psw_addr
, dest
);
1391 tcg_gen_exit_tb((tcg_target_long
)s
->tb
);
1392 return EXIT_GOTO_TB
;
1394 tcg_gen_movi_i64(psw_addr
, dest
);
1395 return EXIT_PC_UPDATED
;
1399 static ExitStatus
help_branch(DisasContext
*s
, DisasCompare
*c
,
1400 bool is_imm
, int imm
, TCGv_i64 cdest
)
1403 uint64_t dest
= s
->pc
+ 2 * imm
;
1406 /* Take care of the special cases first. */
1407 if (c
->cond
== TCG_COND_NEVER
) {
1412 if (dest
== s
->next_pc
) {
1413 /* Branch to next. */
1417 if (c
->cond
== TCG_COND_ALWAYS
) {
1418 ret
= help_goto_direct(s
, dest
);
1422 if (TCGV_IS_UNUSED_I64(cdest
)) {
1423 /* E.g. bcr %r0 -> no branch. */
1427 if (c
->cond
== TCG_COND_ALWAYS
) {
1428 tcg_gen_mov_i64(psw_addr
, cdest
);
1429 ret
= EXIT_PC_UPDATED
;
1434 if (use_goto_tb(s
, s
->next_pc
)) {
1435 if (is_imm
&& use_goto_tb(s
, dest
)) {
1436 /* Both exits can use goto_tb. */
1437 gen_update_cc_op(s
);
1439 lab
= gen_new_label();
1441 tcg_gen_brcond_i64(c
->cond
, c
->u
.s64
.a
, c
->u
.s64
.b
, lab
);
1443 tcg_gen_brcond_i32(c
->cond
, c
->u
.s32
.a
, c
->u
.s32
.b
, lab
);
1446 /* Branch not taken. */
1448 tcg_gen_movi_i64(psw_addr
, s
->next_pc
);
1449 tcg_gen_exit_tb((tcg_target_long
)s
->tb
+ 0);
1454 tcg_gen_movi_i64(psw_addr
, dest
);
1455 tcg_gen_exit_tb((tcg_target_long
)s
->tb
+ 1);
1459 /* Fallthru can use goto_tb, but taken branch cannot. */
1460 /* Store taken branch destination before the brcond. This
1461 avoids having to allocate a new local temp to hold it.
1462 We'll overwrite this in the not taken case anyway. */
1464 tcg_gen_mov_i64(psw_addr
, cdest
);
1467 lab
= gen_new_label();
1469 tcg_gen_brcond_i64(c
->cond
, c
->u
.s64
.a
, c
->u
.s64
.b
, lab
);
1471 tcg_gen_brcond_i32(c
->cond
, c
->u
.s32
.a
, c
->u
.s32
.b
, lab
);
1474 /* Branch not taken. */
1475 gen_update_cc_op(s
);
1477 tcg_gen_movi_i64(psw_addr
, s
->next_pc
);
1478 tcg_gen_exit_tb((tcg_target_long
)s
->tb
+ 0);
1482 tcg_gen_movi_i64(psw_addr
, dest
);
1484 ret
= EXIT_PC_UPDATED
;
1487 /* Fallthru cannot use goto_tb. This by itself is vanishingly rare.
1488 Most commonly we're single-stepping or some other condition that
1489 disables all use of goto_tb. Just update the PC and exit. */
1491 TCGv_i64 next
= tcg_const_i64(s
->next_pc
);
1493 cdest
= tcg_const_i64(dest
);
1497 tcg_gen_movcond_i64(c
->cond
, psw_addr
, c
->u
.s64
.a
, c
->u
.s64
.b
,
1500 TCGv_i32 t0
= tcg_temp_new_i32();
1501 TCGv_i64 t1
= tcg_temp_new_i64();
1502 TCGv_i64 z
= tcg_const_i64(0);
1503 tcg_gen_setcond_i32(c
->cond
, t0
, c
->u
.s32
.a
, c
->u
.s32
.b
);
1504 tcg_gen_extu_i32_i64(t1
, t0
);
1505 tcg_temp_free_i32(t0
);
1506 tcg_gen_movcond_i64(TCG_COND_NE
, psw_addr
, t1
, z
, cdest
, next
);
1507 tcg_temp_free_i64(t1
);
1508 tcg_temp_free_i64(z
);
1512 tcg_temp_free_i64(cdest
);
1514 tcg_temp_free_i64(next
);
1516 ret
= EXIT_PC_UPDATED
;
1524 /* ====================================================================== */
1525 /* The operations. These perform the bulk of the work for any insn,
1526 usually after the operands have been loaded and output initialized. */
1528 static ExitStatus
op_abs(DisasContext
*s
, DisasOps
*o
)
1530 gen_helper_abs_i64(o
->out
, o
->in2
);
1534 static ExitStatus
op_absf32(DisasContext
*s
, DisasOps
*o
)
1536 tcg_gen_andi_i64(o
->out
, o
->in2
, 0x7fffffffull
);
1540 static ExitStatus
op_absf64(DisasContext
*s
, DisasOps
*o
)
1542 tcg_gen_andi_i64(o
->out
, o
->in2
, 0x7fffffffffffffffull
);
1546 static ExitStatus
op_absf128(DisasContext
*s
, DisasOps
*o
)
1548 tcg_gen_andi_i64(o
->out
, o
->in1
, 0x7fffffffffffffffull
);
1549 tcg_gen_mov_i64(o
->out2
, o
->in2
);
1553 static ExitStatus
op_add(DisasContext
*s
, DisasOps
*o
)
1555 tcg_gen_add_i64(o
->out
, o
->in1
, o
->in2
);
1559 static ExitStatus
op_addc(DisasContext
*s
, DisasOps
*o
)
1563 tcg_gen_add_i64(o
->out
, o
->in1
, o
->in2
);
1565 /* XXX possible optimization point */
1567 cc
= tcg_temp_new_i64();
1568 tcg_gen_extu_i32_i64(cc
, cc_op
);
1569 tcg_gen_shri_i64(cc
, cc
, 1);
1571 tcg_gen_add_i64(o
->out
, o
->out
, cc
);
1572 tcg_temp_free_i64(cc
);
1576 static ExitStatus
op_aeb(DisasContext
*s
, DisasOps
*o
)
1578 gen_helper_aeb(o
->out
, cpu_env
, o
->in1
, o
->in2
);
1582 static ExitStatus
op_adb(DisasContext
*s
, DisasOps
*o
)
1584 gen_helper_adb(o
->out
, cpu_env
, o
->in1
, o
->in2
);
1588 static ExitStatus
op_axb(DisasContext
*s
, DisasOps
*o
)
1590 gen_helper_axb(o
->out
, cpu_env
, o
->out
, o
->out2
, o
->in1
, o
->in2
);
1591 return_low128(o
->out2
);
1595 static ExitStatus
op_and(DisasContext
*s
, DisasOps
*o
)
1597 tcg_gen_and_i64(o
->out
, o
->in1
, o
->in2
);
1601 static ExitStatus
op_andi(DisasContext
*s
, DisasOps
*o
)
1603 int shift
= s
->insn
->data
& 0xff;
1604 int size
= s
->insn
->data
>> 8;
1605 uint64_t mask
= ((1ull << size
) - 1) << shift
;
1608 tcg_gen_shli_i64(o
->in2
, o
->in2
, shift
);
1609 tcg_gen_ori_i64(o
->in2
, o
->in2
, ~mask
);
1610 tcg_gen_and_i64(o
->out
, o
->in1
, o
->in2
);
1612 /* Produce the CC from only the bits manipulated. */
1613 tcg_gen_andi_i64(cc_dst
, o
->out
, mask
);
1614 set_cc_nz_u64(s
, cc_dst
);
1618 static ExitStatus
op_bas(DisasContext
*s
, DisasOps
*o
)
1620 tcg_gen_movi_i64(o
->out
, pc_to_link_info(s
, s
->next_pc
));
1621 if (!TCGV_IS_UNUSED_I64(o
->in2
)) {
1622 tcg_gen_mov_i64(psw_addr
, o
->in2
);
1623 return EXIT_PC_UPDATED
;
1629 static ExitStatus
op_basi(DisasContext
*s
, DisasOps
*o
)
1631 tcg_gen_movi_i64(o
->out
, pc_to_link_info(s
, s
->next_pc
));
1632 return help_goto_direct(s
, s
->pc
+ 2 * get_field(s
->fields
, i2
));
1635 static ExitStatus
op_bc(DisasContext
*s
, DisasOps
*o
)
1637 int m1
= get_field(s
->fields
, m1
);
1638 bool is_imm
= have_field(s
->fields
, i2
);
1639 int imm
= is_imm
? get_field(s
->fields
, i2
) : 0;
1642 disas_jcc(s
, &c
, m1
);
1643 return help_branch(s
, &c
, is_imm
, imm
, o
->in2
);
1646 static ExitStatus
op_bct32(DisasContext
*s
, DisasOps
*o
)
1648 int r1
= get_field(s
->fields
, r1
);
1649 bool is_imm
= have_field(s
->fields
, i2
);
1650 int imm
= is_imm
? get_field(s
->fields
, i2
) : 0;
1654 c
.cond
= TCG_COND_NE
;
1659 t
= tcg_temp_new_i64();
1660 tcg_gen_subi_i64(t
, regs
[r1
], 1);
1661 store_reg32_i64(r1
, t
);
1662 c
.u
.s32
.a
= tcg_temp_new_i32();
1663 c
.u
.s32
.b
= tcg_const_i32(0);
1664 tcg_gen_trunc_i64_i32(c
.u
.s32
.a
, t
);
1665 tcg_temp_free_i64(t
);
1667 return help_branch(s
, &c
, is_imm
, imm
, o
->in2
);
1670 static ExitStatus
op_bct64(DisasContext
*s
, DisasOps
*o
)
1672 int r1
= get_field(s
->fields
, r1
);
1673 bool is_imm
= have_field(s
->fields
, i2
);
1674 int imm
= is_imm
? get_field(s
->fields
, i2
) : 0;
1677 c
.cond
= TCG_COND_NE
;
1682 tcg_gen_subi_i64(regs
[r1
], regs
[r1
], 1);
1683 c
.u
.s64
.a
= regs
[r1
];
1684 c
.u
.s64
.b
= tcg_const_i64(0);
1686 return help_branch(s
, &c
, is_imm
, imm
, o
->in2
);
1689 static ExitStatus
op_ceb(DisasContext
*s
, DisasOps
*o
)
1691 gen_helper_ceb(cc_op
, cpu_env
, o
->in1
, o
->in2
);
1696 static ExitStatus
op_cdb(DisasContext
*s
, DisasOps
*o
)
1698 gen_helper_cdb(cc_op
, cpu_env
, o
->in1
, o
->in2
);
1703 static ExitStatus
op_cxb(DisasContext
*s
, DisasOps
*o
)
1705 gen_helper_cxb(cc_op
, cpu_env
, o
->out
, o
->out2
, o
->in1
, o
->in2
);
1710 static ExitStatus
op_cfeb(DisasContext
*s
, DisasOps
*o
)
1712 TCGv_i32 m3
= tcg_const_i32(get_field(s
->fields
, m3
));
1713 gen_helper_cfeb(o
->out
, cpu_env
, o
->in2
, m3
);
1714 tcg_temp_free_i32(m3
);
1715 gen_set_cc_nz_f32(s
, o
->in2
);
1719 static ExitStatus
op_cfdb(DisasContext
*s
, DisasOps
*o
)
1721 TCGv_i32 m3
= tcg_const_i32(get_field(s
->fields
, m3
));
1722 gen_helper_cfdb(o
->out
, cpu_env
, o
->in2
, m3
);
1723 tcg_temp_free_i32(m3
);
1724 gen_set_cc_nz_f64(s
, o
->in2
);
1728 static ExitStatus
op_cfxb(DisasContext
*s
, DisasOps
*o
)
1730 TCGv_i32 m3
= tcg_const_i32(get_field(s
->fields
, m3
));
1731 gen_helper_cfxb(o
->out
, cpu_env
, o
->in1
, o
->in2
, m3
);
1732 tcg_temp_free_i32(m3
);
1733 gen_set_cc_nz_f128(s
, o
->in1
, o
->in2
);
1737 static ExitStatus
op_cgeb(DisasContext
*s
, DisasOps
*o
)
1739 TCGv_i32 m3
= tcg_const_i32(get_field(s
->fields
, m3
));
1740 gen_helper_cgeb(o
->out
, cpu_env
, o
->in2
, m3
);
1741 tcg_temp_free_i32(m3
);
1742 gen_set_cc_nz_f32(s
, o
->in2
);
1746 static ExitStatus
op_cgdb(DisasContext
*s
, DisasOps
*o
)
1748 TCGv_i32 m3
= tcg_const_i32(get_field(s
->fields
, m3
));
1749 gen_helper_cgdb(o
->out
, cpu_env
, o
->in2
, m3
);
1750 tcg_temp_free_i32(m3
);
1751 gen_set_cc_nz_f64(s
, o
->in2
);
1755 static ExitStatus
op_cgxb(DisasContext
*s
, DisasOps
*o
)
1757 TCGv_i32 m3
= tcg_const_i32(get_field(s
->fields
, m3
));
1758 gen_helper_cgxb(o
->out
, cpu_env
, o
->in1
, o
->in2
, m3
);
1759 tcg_temp_free_i32(m3
);
1760 gen_set_cc_nz_f128(s
, o
->in1
, o
->in2
);
1764 static ExitStatus
op_cegb(DisasContext
*s
, DisasOps
*o
)
1766 TCGv_i32 m3
= tcg_const_i32(get_field(s
->fields
, m3
));
1767 gen_helper_cegb(o
->out
, cpu_env
, o
->in2
, m3
);
1768 tcg_temp_free_i32(m3
);
1772 static ExitStatus
op_cdgb(DisasContext
*s
, DisasOps
*o
)
1774 TCGv_i32 m3
= tcg_const_i32(get_field(s
->fields
, m3
));
1775 gen_helper_cdgb(o
->out
, cpu_env
, o
->in2
, m3
);
1776 tcg_temp_free_i32(m3
);
1780 static ExitStatus
op_cxgb(DisasContext
*s
, DisasOps
*o
)
1782 TCGv_i32 m3
= tcg_const_i32(get_field(s
->fields
, m3
));
1783 gen_helper_cxgb(o
->out
, cpu_env
, o
->in2
, m3
);
1784 tcg_temp_free_i32(m3
);
1785 return_low128(o
->out2
);
1789 static ExitStatus
op_cksm(DisasContext
*s
, DisasOps
*o
)
1791 int r2
= get_field(s
->fields
, r2
);
1792 TCGv_i64 len
= tcg_temp_new_i64();
1794 potential_page_fault(s
);
1795 gen_helper_cksm(len
, cpu_env
, o
->in1
, o
->in2
, regs
[r2
+ 1]);
1797 return_low128(o
->out
);
1799 tcg_gen_add_i64(regs
[r2
], regs
[r2
], len
);
1800 tcg_gen_sub_i64(regs
[r2
+ 1], regs
[r2
+ 1], len
);
1801 tcg_temp_free_i64(len
);
1806 static ExitStatus
op_clc(DisasContext
*s
, DisasOps
*o
)
1808 int l
= get_field(s
->fields
, l1
);
1813 tcg_gen_qemu_ld8u(cc_src
, o
->addr1
, get_mem_index(s
));
1814 tcg_gen_qemu_ld8u(cc_dst
, o
->in2
, get_mem_index(s
));
1817 tcg_gen_qemu_ld16u(cc_src
, o
->addr1
, get_mem_index(s
));
1818 tcg_gen_qemu_ld16u(cc_dst
, o
->in2
, get_mem_index(s
));
1821 tcg_gen_qemu_ld32u(cc_src
, o
->addr1
, get_mem_index(s
));
1822 tcg_gen_qemu_ld32u(cc_dst
, o
->in2
, get_mem_index(s
));
1825 tcg_gen_qemu_ld64(cc_src
, o
->addr1
, get_mem_index(s
));
1826 tcg_gen_qemu_ld64(cc_dst
, o
->in2
, get_mem_index(s
));
1829 potential_page_fault(s
);
1830 vl
= tcg_const_i32(l
);
1831 gen_helper_clc(cc_op
, cpu_env
, vl
, o
->addr1
, o
->in2
);
1832 tcg_temp_free_i32(vl
);
1836 gen_op_update2_cc_i64(s
, CC_OP_LTUGTU_64
, cc_src
, cc_dst
);
1840 static ExitStatus
op_clcle(DisasContext
*s
, DisasOps
*o
)
1842 TCGv_i32 r1
= tcg_const_i32(get_field(s
->fields
, r1
));
1843 TCGv_i32 r3
= tcg_const_i32(get_field(s
->fields
, r3
));
1844 potential_page_fault(s
);
1845 gen_helper_clcle(cc_op
, cpu_env
, r1
, o
->in2
, r3
);
1846 tcg_temp_free_i32(r1
);
1847 tcg_temp_free_i32(r3
);
1852 static ExitStatus
op_clm(DisasContext
*s
, DisasOps
*o
)
1854 TCGv_i32 m3
= tcg_const_i32(get_field(s
->fields
, m3
));
1855 TCGv_i32 t1
= tcg_temp_new_i32();
1856 tcg_gen_trunc_i64_i32(t1
, o
->in1
);
1857 potential_page_fault(s
);
1858 gen_helper_clm(cc_op
, cpu_env
, t1
, m3
, o
->in2
);
1860 tcg_temp_free_i32(t1
);
1861 tcg_temp_free_i32(m3
);
1865 static ExitStatus
op_clst(DisasContext
*s
, DisasOps
*o
)
1867 potential_page_fault(s
);
1868 gen_helper_clst(o
->in1
, cpu_env
, regs
[0], o
->in1
, o
->in2
);
1870 return_low128(o
->in2
);
1874 static ExitStatus
op_cs(DisasContext
*s
, DisasOps
*o
)
1876 int r3
= get_field(s
->fields
, r3
);
1877 potential_page_fault(s
);
1878 gen_helper_cs(o
->out
, cpu_env
, o
->in1
, o
->in2
, regs
[r3
]);
1883 static ExitStatus
op_csg(DisasContext
*s
, DisasOps
*o
)
1885 int r3
= get_field(s
->fields
, r3
);
1886 potential_page_fault(s
);
1887 gen_helper_csg(o
->out
, cpu_env
, o
->in1
, o
->in2
, regs
[r3
]);
1892 #ifndef CONFIG_USER_ONLY
1893 static ExitStatus
op_csp(DisasContext
*s
, DisasOps
*o
)
1895 TCGv_i32 r1
= tcg_const_i32(get_field(s
->fields
, r1
));
1896 check_privileged(s
);
1897 gen_helper_csp(cc_op
, cpu_env
, r1
, o
->in2
);
1898 tcg_temp_free_i32(r1
);
1904 static ExitStatus
op_cds(DisasContext
*s
, DisasOps
*o
)
1906 int r3
= get_field(s
->fields
, r3
);
1907 TCGv_i64 in3
= tcg_temp_new_i64();
1908 tcg_gen_deposit_i64(in3
, regs
[r3
+ 1], regs
[r3
], 32, 32);
1909 potential_page_fault(s
);
1910 gen_helper_csg(o
->out
, cpu_env
, o
->in1
, o
->in2
, in3
);
1911 tcg_temp_free_i64(in3
);
1916 static ExitStatus
op_cdsg(DisasContext
*s
, DisasOps
*o
)
1918 TCGv_i32 r1
= tcg_const_i32(get_field(s
->fields
, r1
));
1919 TCGv_i32 r3
= tcg_const_i32(get_field(s
->fields
, r3
));
1920 potential_page_fault(s
);
1921 /* XXX rewrite in tcg */
1922 gen_helper_cdsg(cc_op
, cpu_env
, r1
, o
->in2
, r3
);
1927 static ExitStatus
op_cvd(DisasContext
*s
, DisasOps
*o
)
1929 TCGv_i64 t1
= tcg_temp_new_i64();
1930 TCGv_i32 t2
= tcg_temp_new_i32();
1931 tcg_gen_trunc_i64_i32(t2
, o
->in1
);
1932 gen_helper_cvd(t1
, t2
);
1933 tcg_temp_free_i32(t2
);
1934 tcg_gen_qemu_st64(t1
, o
->in2
, get_mem_index(s
));
1935 tcg_temp_free_i64(t1
);
1939 #ifndef CONFIG_USER_ONLY
1940 static ExitStatus
op_diag(DisasContext
*s
, DisasOps
*o
)
1944 check_privileged(s
);
1945 potential_page_fault(s
);
1947 /* We pretend the format is RX_a so that D2 is the field we want. */
1948 tmp
= tcg_const_i32(get_field(s
->fields
, d2
) & 0xfff);
1949 gen_helper_diag(regs
[2], cpu_env
, tmp
, regs
[2], regs
[1]);
1950 tcg_temp_free_i32(tmp
);
1955 static ExitStatus
op_divs32(DisasContext
*s
, DisasOps
*o
)
1957 gen_helper_divs32(o
->out2
, cpu_env
, o
->in1
, o
->in2
);
1958 return_low128(o
->out
);
1962 static ExitStatus
op_divu32(DisasContext
*s
, DisasOps
*o
)
1964 gen_helper_divu32(o
->out2
, cpu_env
, o
->in1
, o
->in2
);
1965 return_low128(o
->out
);
1969 static ExitStatus
op_divs64(DisasContext
*s
, DisasOps
*o
)
1971 gen_helper_divs64(o
->out2
, cpu_env
, o
->in1
, o
->in2
);
1972 return_low128(o
->out
);
1976 static ExitStatus
op_divu64(DisasContext
*s
, DisasOps
*o
)
1978 gen_helper_divu64(o
->out2
, cpu_env
, o
->out
, o
->out2
, o
->in2
);
1979 return_low128(o
->out
);
1983 static ExitStatus
op_deb(DisasContext
*s
, DisasOps
*o
)
1985 gen_helper_deb(o
->out
, cpu_env
, o
->in1
, o
->in2
);
1989 static ExitStatus
op_ddb(DisasContext
*s
, DisasOps
*o
)
1991 gen_helper_ddb(o
->out
, cpu_env
, o
->in1
, o
->in2
);
1995 static ExitStatus
op_dxb(DisasContext
*s
, DisasOps
*o
)
1997 gen_helper_dxb(o
->out
, cpu_env
, o
->out
, o
->out2
, o
->in1
, o
->in2
);
1998 return_low128(o
->out2
);
2002 static ExitStatus
op_ear(DisasContext
*s
, DisasOps
*o
)
2004 int r2
= get_field(s
->fields
, r2
);
2005 tcg_gen_ld32u_i64(o
->out
, cpu_env
, offsetof(CPUS390XState
, aregs
[r2
]));
2009 static ExitStatus
op_efpc(DisasContext
*s
, DisasOps
*o
)
2011 tcg_gen_ld32u_i64(o
->out
, cpu_env
, offsetof(CPUS390XState
, fpc
));
2015 static ExitStatus
op_ex(DisasContext
*s
, DisasOps
*o
)
2017 /* ??? Perhaps a better way to implement EXECUTE is to set a bit in
2018 tb->flags, (ab)use the tb->cs_base field as the address of
2019 the template in memory, and grab 8 bits of tb->flags/cflags for
2020 the contents of the register. We would then recognize all this
2021 in gen_intermediate_code_internal, generating code for exactly
2022 one instruction. This new TB then gets executed normally.
2024 On the other hand, this seems to be mostly used for modifying
2025 MVC inside of memcpy, which needs a helper call anyway. So
2026 perhaps this doesn't bear thinking about any further. */
2033 tmp
= tcg_const_i64(s
->next_pc
);
2034 gen_helper_ex(cc_op
, cpu_env
, cc_op
, o
->in1
, o
->in2
, tmp
);
2035 tcg_temp_free_i64(tmp
);
2041 static ExitStatus
op_flogr(DisasContext
*s
, DisasOps
*o
)
2043 /* We'll use the original input for cc computation, since we get to
2044 compare that against 0, which ought to be better than comparing
2045 the real output against 64. It also lets cc_dst be a convenient
2046 temporary during our computation. */
2047 gen_op_update1_cc_i64(s
, CC_OP_FLOGR
, o
->in2
);
2049 /* R1 = IN ? CLZ(IN) : 64. */
2050 gen_helper_clz(o
->out
, o
->in2
);
2052 /* R1+1 = IN & ~(found bit). Note that we may attempt to shift this
2053 value by 64, which is undefined. But since the shift is 64 iff the
2054 input is zero, we still get the correct result after and'ing. */
2055 tcg_gen_movi_i64(o
->out2
, 0x8000000000000000ull
);
2056 tcg_gen_shr_i64(o
->out2
, o
->out2
, o
->out
);
2057 tcg_gen_andc_i64(o
->out2
, cc_dst
, o
->out2
);
2061 static ExitStatus
op_icm(DisasContext
*s
, DisasOps
*o
)
2063 int m3
= get_field(s
->fields
, m3
);
2064 int pos
, len
, base
= s
->insn
->data
;
2065 TCGv_i64 tmp
= tcg_temp_new_i64();
2070 /* Effectively a 32-bit load. */
2071 tcg_gen_qemu_ld32u(tmp
, o
->in2
, get_mem_index(s
));
2078 /* Effectively a 16-bit load. */
2079 tcg_gen_qemu_ld16u(tmp
, o
->in2
, get_mem_index(s
));
2087 /* Effectively an 8-bit load. */
2088 tcg_gen_qemu_ld8u(tmp
, o
->in2
, get_mem_index(s
));
2093 pos
= base
+ ctz32(m3
) * 8;
2094 tcg_gen_deposit_i64(o
->out
, o
->out
, tmp
, pos
, len
);
2095 ccm
= ((1ull << len
) - 1) << pos
;
2099 /* This is going to be a sequence of loads and inserts. */
2100 pos
= base
+ 32 - 8;
2104 tcg_gen_qemu_ld8u(tmp
, o
->in2
, get_mem_index(s
));
2105 tcg_gen_addi_i64(o
->in2
, o
->in2
, 1);
2106 tcg_gen_deposit_i64(o
->out
, o
->out
, tmp
, pos
, 8);
2109 m3
= (m3
<< 1) & 0xf;
2115 tcg_gen_movi_i64(tmp
, ccm
);
2116 gen_op_update2_cc_i64(s
, CC_OP_ICM
, tmp
, o
->out
);
2117 tcg_temp_free_i64(tmp
);
2121 static ExitStatus
op_insi(DisasContext
*s
, DisasOps
*o
)
2123 int shift
= s
->insn
->data
& 0xff;
2124 int size
= s
->insn
->data
>> 8;
2125 tcg_gen_deposit_i64(o
->out
, o
->in1
, o
->in2
, shift
, size
);
2129 static ExitStatus
op_ipm(DisasContext
*s
, DisasOps
*o
)
2134 tcg_gen_andi_i64(o
->out
, o
->out
, ~0xff000000ull
);
2136 t1
= tcg_temp_new_i64();
2137 tcg_gen_shli_i64(t1
, psw_mask
, 20);
2138 tcg_gen_shri_i64(t1
, t1
, 36);
2139 tcg_gen_or_i64(o
->out
, o
->out
, t1
);
2141 tcg_gen_extu_i32_i64(t1
, cc_op
);
2142 tcg_gen_shli_i64(t1
, t1
, 28);
2143 tcg_gen_or_i64(o
->out
, o
->out
, t1
);
2144 tcg_temp_free_i64(t1
);
2148 #ifndef CONFIG_USER_ONLY
2149 static ExitStatus
op_ipte(DisasContext
*s
, DisasOps
*o
)
2151 check_privileged(s
);
2152 gen_helper_ipte(cpu_env
, o
->in1
, o
->in2
);
2156 static ExitStatus
op_iske(DisasContext
*s
, DisasOps
*o
)
2158 check_privileged(s
);
2159 gen_helper_iske(o
->out
, cpu_env
, o
->in2
);
2164 static ExitStatus
op_ldeb(DisasContext
*s
, DisasOps
*o
)
2166 gen_helper_ldeb(o
->out
, cpu_env
, o
->in2
);
2170 static ExitStatus
op_ledb(DisasContext
*s
, DisasOps
*o
)
2172 gen_helper_ledb(o
->out
, cpu_env
, o
->in2
);
2176 static ExitStatus
op_ldxb(DisasContext
*s
, DisasOps
*o
)
2178 gen_helper_ldxb(o
->out
, cpu_env
, o
->in1
, o
->in2
);
2182 static ExitStatus
op_lexb(DisasContext
*s
, DisasOps
*o
)
2184 gen_helper_lexb(o
->out
, cpu_env
, o
->in1
, o
->in2
);
2188 static ExitStatus
op_lxdb(DisasContext
*s
, DisasOps
*o
)
2190 gen_helper_lxdb(o
->out
, cpu_env
, o
->in2
);
2191 return_low128(o
->out2
);
2195 static ExitStatus
op_lxeb(DisasContext
*s
, DisasOps
*o
)
2197 gen_helper_lxeb(o
->out
, cpu_env
, o
->in2
);
2198 return_low128(o
->out2
);
2202 static ExitStatus
op_llgt(DisasContext
*s
, DisasOps
*o
)
2204 tcg_gen_andi_i64(o
->out
, o
->in2
, 0x7fffffff);
2208 static ExitStatus
op_ld8s(DisasContext
*s
, DisasOps
*o
)
2210 tcg_gen_qemu_ld8s(o
->out
, o
->in2
, get_mem_index(s
));
2214 static ExitStatus
op_ld8u(DisasContext
*s
, DisasOps
*o
)
2216 tcg_gen_qemu_ld8u(o
->out
, o
->in2
, get_mem_index(s
));
2220 static ExitStatus
op_ld16s(DisasContext
*s
, DisasOps
*o
)
2222 tcg_gen_qemu_ld16s(o
->out
, o
->in2
, get_mem_index(s
));
2226 static ExitStatus
op_ld16u(DisasContext
*s
, DisasOps
*o
)
2228 tcg_gen_qemu_ld16u(o
->out
, o
->in2
, get_mem_index(s
));
2232 static ExitStatus
op_ld32s(DisasContext
*s
, DisasOps
*o
)
2234 tcg_gen_qemu_ld32s(o
->out
, o
->in2
, get_mem_index(s
));
2238 static ExitStatus
op_ld32u(DisasContext
*s
, DisasOps
*o
)
2240 tcg_gen_qemu_ld32u(o
->out
, o
->in2
, get_mem_index(s
));
2244 static ExitStatus
op_ld64(DisasContext
*s
, DisasOps
*o
)
2246 tcg_gen_qemu_ld64(o
->out
, o
->in2
, get_mem_index(s
));
2250 #ifndef CONFIG_USER_ONLY
2251 static ExitStatus
op_lctl(DisasContext
*s
, DisasOps
*o
)
2253 TCGv_i32 r1
= tcg_const_i32(get_field(s
->fields
, r1
));
2254 TCGv_i32 r3
= tcg_const_i32(get_field(s
->fields
, r3
));
2255 check_privileged(s
);
2256 potential_page_fault(s
);
2257 gen_helper_lctl(cpu_env
, r1
, o
->in2
, r3
);
2258 tcg_temp_free_i32(r1
);
2259 tcg_temp_free_i32(r3
);
2263 static ExitStatus
op_lctlg(DisasContext
*s
, DisasOps
*o
)
2265 TCGv_i32 r1
= tcg_const_i32(get_field(s
->fields
, r1
));
2266 TCGv_i32 r3
= tcg_const_i32(get_field(s
->fields
, r3
));
2267 check_privileged(s
);
2268 potential_page_fault(s
);
2269 gen_helper_lctlg(cpu_env
, r1
, o
->in2
, r3
);
2270 tcg_temp_free_i32(r1
);
2271 tcg_temp_free_i32(r3
);
2274 static ExitStatus
op_lra(DisasContext
*s
, DisasOps
*o
)
2276 check_privileged(s
);
2277 potential_page_fault(s
);
2278 gen_helper_lra(o
->out
, cpu_env
, o
->in2
);
2283 static ExitStatus
op_lpsw(DisasContext
*s
, DisasOps
*o
)
2287 check_privileged(s
);
2289 t1
= tcg_temp_new_i64();
2290 t2
= tcg_temp_new_i64();
2291 tcg_gen_qemu_ld32u(t1
, o
->in2
, get_mem_index(s
));
2292 tcg_gen_addi_i64(o
->in2
, o
->in2
, 4);
2293 tcg_gen_qemu_ld32u(t2
, o
->in2
, get_mem_index(s
));
2294 /* Convert the 32-bit PSW_MASK into the 64-bit PSW_MASK. */
2295 tcg_gen_shli_i64(t1
, t1
, 32);
2296 gen_helper_load_psw(cpu_env
, t1
, t2
);
2297 tcg_temp_free_i64(t1
);
2298 tcg_temp_free_i64(t2
);
2299 return EXIT_NORETURN
;
2303 static ExitStatus
op_lam(DisasContext
*s
, DisasOps
*o
)
2305 TCGv_i32 r1
= tcg_const_i32(get_field(s
->fields
, r1
));
2306 TCGv_i32 r3
= tcg_const_i32(get_field(s
->fields
, r3
));
2307 potential_page_fault(s
);
2308 gen_helper_lam(cpu_env
, r1
, o
->in2
, r3
);
2309 tcg_temp_free_i32(r1
);
2310 tcg_temp_free_i32(r3
);
2314 static ExitStatus
op_lm32(DisasContext
*s
, DisasOps
*o
)
2316 int r1
= get_field(s
->fields
, r1
);
2317 int r3
= get_field(s
->fields
, r3
);
2318 TCGv_i64 t
= tcg_temp_new_i64();
2319 TCGv_i64 t4
= tcg_const_i64(4);
2322 tcg_gen_qemu_ld32u(t
, o
->in2
, get_mem_index(s
));
2323 store_reg32_i64(r1
, t
);
2327 tcg_gen_add_i64(o
->in2
, o
->in2
, t4
);
2331 tcg_temp_free_i64(t
);
2332 tcg_temp_free_i64(t4
);
2336 static ExitStatus
op_lmh(DisasContext
*s
, DisasOps
*o
)
2338 int r1
= get_field(s
->fields
, r1
);
2339 int r3
= get_field(s
->fields
, r3
);
2340 TCGv_i64 t
= tcg_temp_new_i64();
2341 TCGv_i64 t4
= tcg_const_i64(4);
2344 tcg_gen_qemu_ld32u(t
, o
->in2
, get_mem_index(s
));
2345 store_reg32h_i64(r1
, t
);
2349 tcg_gen_add_i64(o
->in2
, o
->in2
, t4
);
2353 tcg_temp_free_i64(t
);
2354 tcg_temp_free_i64(t4
);
2358 static ExitStatus
op_lm64(DisasContext
*s
, DisasOps
*o
)
2360 int r1
= get_field(s
->fields
, r1
);
2361 int r3
= get_field(s
->fields
, r3
);
2362 TCGv_i64 t8
= tcg_const_i64(8);
2365 tcg_gen_qemu_ld64(regs
[r1
], o
->in2
, get_mem_index(s
));
2369 tcg_gen_add_i64(o
->in2
, o
->in2
, t8
);
2373 tcg_temp_free_i64(t8
);
2377 static ExitStatus
op_mov2(DisasContext
*s
, DisasOps
*o
)
2380 o
->g_out
= o
->g_in2
;
2381 TCGV_UNUSED_I64(o
->in2
);
2386 static ExitStatus
op_movx(DisasContext
*s
, DisasOps
*o
)
2390 o
->g_out
= o
->g_in1
;
2391 o
->g_out2
= o
->g_in2
;
2392 TCGV_UNUSED_I64(o
->in1
);
2393 TCGV_UNUSED_I64(o
->in2
);
2394 o
->g_in1
= o
->g_in2
= false;
2398 static ExitStatus
op_mvc(DisasContext
*s
, DisasOps
*o
)
2400 TCGv_i32 l
= tcg_const_i32(get_field(s
->fields
, l1
));
2401 potential_page_fault(s
);
2402 gen_helper_mvc(cpu_env
, l
, o
->addr1
, o
->in2
);
2403 tcg_temp_free_i32(l
);
2407 static ExitStatus
op_mvcl(DisasContext
*s
, DisasOps
*o
)
2409 TCGv_i32 r1
= tcg_const_i32(get_field(s
->fields
, r1
));
2410 TCGv_i32 r2
= tcg_const_i32(get_field(s
->fields
, r2
));
2411 potential_page_fault(s
);
2412 gen_helper_mvcl(cc_op
, cpu_env
, r1
, r2
);
2413 tcg_temp_free_i32(r1
);
2414 tcg_temp_free_i32(r2
);
2419 static ExitStatus
op_mvcle(DisasContext
*s
, DisasOps
*o
)
2421 TCGv_i32 r1
= tcg_const_i32(get_field(s
->fields
, r1
));
2422 TCGv_i32 r3
= tcg_const_i32(get_field(s
->fields
, r3
));
2423 potential_page_fault(s
);
2424 gen_helper_mvcle(cc_op
, cpu_env
, r1
, o
->in2
, r3
);
2425 tcg_temp_free_i32(r1
);
2426 tcg_temp_free_i32(r3
);
2431 #ifndef CONFIG_USER_ONLY
2432 static ExitStatus
op_mvcp(DisasContext
*s
, DisasOps
*o
)
2434 int r1
= get_field(s
->fields
, l1
);
2435 check_privileged(s
);
2436 potential_page_fault(s
);
2437 gen_helper_mvcp(cc_op
, cpu_env
, regs
[r1
], o
->addr1
, o
->in2
);
2442 static ExitStatus
op_mvcs(DisasContext
*s
, DisasOps
*o
)
2444 int r1
= get_field(s
->fields
, l1
);
2445 check_privileged(s
);
2446 potential_page_fault(s
);
2447 gen_helper_mvcs(cc_op
, cpu_env
, regs
[r1
], o
->addr1
, o
->in2
);
2453 static ExitStatus
op_mvpg(DisasContext
*s
, DisasOps
*o
)
2455 potential_page_fault(s
);
2456 gen_helper_mvpg(cpu_env
, regs
[0], o
->in1
, o
->in2
);
2461 static ExitStatus
op_mvst(DisasContext
*s
, DisasOps
*o
)
2463 potential_page_fault(s
);
2464 gen_helper_mvst(o
->in1
, cpu_env
, regs
[0], o
->in1
, o
->in2
);
2466 return_low128(o
->in2
);
2470 static ExitStatus
op_mul(DisasContext
*s
, DisasOps
*o
)
2472 tcg_gen_mul_i64(o
->out
, o
->in1
, o
->in2
);
2476 static ExitStatus
op_mul128(DisasContext
*s
, DisasOps
*o
)
2478 gen_helper_mul128(o
->out
, cpu_env
, o
->in1
, o
->in2
);
2479 return_low128(o
->out2
);
2483 static ExitStatus
op_meeb(DisasContext
*s
, DisasOps
*o
)
2485 gen_helper_meeb(o
->out
, cpu_env
, o
->in1
, o
->in2
);
2489 static ExitStatus
op_mdeb(DisasContext
*s
, DisasOps
*o
)
2491 gen_helper_mdeb(o
->out
, cpu_env
, o
->in1
, o
->in2
);
2495 static ExitStatus
op_mdb(DisasContext
*s
, DisasOps
*o
)
2497 gen_helper_mdb(o
->out
, cpu_env
, o
->in1
, o
->in2
);
2501 static ExitStatus
op_mxb(DisasContext
*s
, DisasOps
*o
)
2503 gen_helper_mxb(o
->out
, cpu_env
, o
->out
, o
->out2
, o
->in1
, o
->in2
);
2504 return_low128(o
->out2
);
2508 static ExitStatus
op_mxdb(DisasContext
*s
, DisasOps
*o
)
2510 gen_helper_mxdb(o
->out
, cpu_env
, o
->out
, o
->out2
, o
->in2
);
2511 return_low128(o
->out2
);
2515 static ExitStatus
op_maeb(DisasContext
*s
, DisasOps
*o
)
2517 TCGv_i64 r3
= load_freg32_i64(get_field(s
->fields
, r3
));
2518 gen_helper_maeb(o
->out
, cpu_env
, o
->in1
, o
->in2
, r3
);
2519 tcg_temp_free_i64(r3
);
2523 static ExitStatus
op_madb(DisasContext
*s
, DisasOps
*o
)
2525 int r3
= get_field(s
->fields
, r3
);
2526 gen_helper_madb(o
->out
, cpu_env
, o
->in1
, o
->in2
, fregs
[r3
]);
2530 static ExitStatus
op_mseb(DisasContext
*s
, DisasOps
*o
)
2532 TCGv_i64 r3
= load_freg32_i64(get_field(s
->fields
, r3
));
2533 gen_helper_mseb(o
->out
, cpu_env
, o
->in1
, o
->in2
, r3
);
2534 tcg_temp_free_i64(r3
);
2538 static ExitStatus
op_msdb(DisasContext
*s
, DisasOps
*o
)
2540 int r3
= get_field(s
->fields
, r3
);
2541 gen_helper_msdb(o
->out
, cpu_env
, o
->in1
, o
->in2
, fregs
[r3
]);
2545 static ExitStatus
op_nabs(DisasContext
*s
, DisasOps
*o
)
2547 gen_helper_nabs_i64(o
->out
, o
->in2
);
2551 static ExitStatus
op_nabsf32(DisasContext
*s
, DisasOps
*o
)
2553 tcg_gen_ori_i64(o
->out
, o
->in2
, 0x80000000ull
);
2557 static ExitStatus
op_nabsf64(DisasContext
*s
, DisasOps
*o
)
2559 tcg_gen_ori_i64(o
->out
, o
->in2
, 0x8000000000000000ull
);
2563 static ExitStatus
op_nabsf128(DisasContext
*s
, DisasOps
*o
)
2565 tcg_gen_ori_i64(o
->out
, o
->in1
, 0x8000000000000000ull
);
2566 tcg_gen_mov_i64(o
->out2
, o
->in2
);
2570 static ExitStatus
op_nc(DisasContext
*s
, DisasOps
*o
)
2572 TCGv_i32 l
= tcg_const_i32(get_field(s
->fields
, l1
));
2573 potential_page_fault(s
);
2574 gen_helper_nc(cc_op
, cpu_env
, l
, o
->addr1
, o
->in2
);
2575 tcg_temp_free_i32(l
);
2580 static ExitStatus
op_neg(DisasContext
*s
, DisasOps
*o
)
2582 tcg_gen_neg_i64(o
->out
, o
->in2
);
2586 static ExitStatus
op_negf32(DisasContext
*s
, DisasOps
*o
)
2588 tcg_gen_xori_i64(o
->out
, o
->in2
, 0x80000000ull
);
2592 static ExitStatus
op_negf64(DisasContext
*s
, DisasOps
*o
)
2594 tcg_gen_xori_i64(o
->out
, o
->in2
, 0x8000000000000000ull
);
2598 static ExitStatus
op_negf128(DisasContext
*s
, DisasOps
*o
)
2600 tcg_gen_xori_i64(o
->out
, o
->in1
, 0x8000000000000000ull
);
2601 tcg_gen_mov_i64(o
->out2
, o
->in2
);
2605 static ExitStatus
op_oc(DisasContext
*s
, DisasOps
*o
)
2607 TCGv_i32 l
= tcg_const_i32(get_field(s
->fields
, l1
));
2608 potential_page_fault(s
);
2609 gen_helper_oc(cc_op
, cpu_env
, l
, o
->addr1
, o
->in2
);
2610 tcg_temp_free_i32(l
);
2615 static ExitStatus
op_or(DisasContext
*s
, DisasOps
*o
)
2617 tcg_gen_or_i64(o
->out
, o
->in1
, o
->in2
);
2621 static ExitStatus
op_ori(DisasContext
*s
, DisasOps
*o
)
2623 int shift
= s
->insn
->data
& 0xff;
2624 int size
= s
->insn
->data
>> 8;
2625 uint64_t mask
= ((1ull << size
) - 1) << shift
;
2628 tcg_gen_shli_i64(o
->in2
, o
->in2
, shift
);
2629 tcg_gen_or_i64(o
->out
, o
->in1
, o
->in2
);
2631 /* Produce the CC from only the bits manipulated. */
2632 tcg_gen_andi_i64(cc_dst
, o
->out
, mask
);
2633 set_cc_nz_u64(s
, cc_dst
);
2637 #ifndef CONFIG_USER_ONLY
2638 static ExitStatus
op_ptlb(DisasContext
*s
, DisasOps
*o
)
2640 check_privileged(s
);
2641 gen_helper_ptlb(cpu_env
);
2646 static ExitStatus
op_rev16(DisasContext
*s
, DisasOps
*o
)
2648 tcg_gen_bswap16_i64(o
->out
, o
->in2
);
2652 static ExitStatus
op_rev32(DisasContext
*s
, DisasOps
*o
)
2654 tcg_gen_bswap32_i64(o
->out
, o
->in2
);
2658 static ExitStatus
op_rev64(DisasContext
*s
, DisasOps
*o
)
2660 tcg_gen_bswap64_i64(o
->out
, o
->in2
);
2664 static ExitStatus
op_rll32(DisasContext
*s
, DisasOps
*o
)
2666 TCGv_i32 t1
= tcg_temp_new_i32();
2667 TCGv_i32 t2
= tcg_temp_new_i32();
2668 TCGv_i32 to
= tcg_temp_new_i32();
2669 tcg_gen_trunc_i64_i32(t1
, o
->in1
);
2670 tcg_gen_trunc_i64_i32(t2
, o
->in2
);
2671 tcg_gen_rotl_i32(to
, t1
, t2
);
2672 tcg_gen_extu_i32_i64(o
->out
, to
);
2673 tcg_temp_free_i32(t1
);
2674 tcg_temp_free_i32(t2
);
2675 tcg_temp_free_i32(to
);
2679 static ExitStatus
op_rll64(DisasContext
*s
, DisasOps
*o
)
2681 tcg_gen_rotl_i64(o
->out
, o
->in1
, o
->in2
);
2685 #ifndef CONFIG_USER_ONLY
2686 static ExitStatus
op_rrbe(DisasContext
*s
, DisasOps
*o
)
2688 check_privileged(s
);
2689 gen_helper_rrbe(cc_op
, cpu_env
, o
->in2
);
2695 static ExitStatus
op_sar(DisasContext
*s
, DisasOps
*o
)
2697 int r1
= get_field(s
->fields
, r1
);
2698 tcg_gen_st32_i64(o
->in2
, cpu_env
, offsetof(CPUS390XState
, aregs
[r1
]));
2702 static ExitStatus
op_seb(DisasContext
*s
, DisasOps
*o
)
2704 gen_helper_seb(o
->out
, cpu_env
, o
->in1
, o
->in2
);
2708 static ExitStatus
op_sdb(DisasContext
*s
, DisasOps
*o
)
2710 gen_helper_sdb(o
->out
, cpu_env
, o
->in1
, o
->in2
);
2714 static ExitStatus
op_sxb(DisasContext
*s
, DisasOps
*o
)
2716 gen_helper_sxb(o
->out
, cpu_env
, o
->out
, o
->out2
, o
->in1
, o
->in2
);
2717 return_low128(o
->out2
);
2721 static ExitStatus
op_sqeb(DisasContext
*s
, DisasOps
*o
)
2723 gen_helper_sqeb(o
->out
, cpu_env
, o
->in2
);
2727 static ExitStatus
op_sqdb(DisasContext
*s
, DisasOps
*o
)
2729 gen_helper_sqdb(o
->out
, cpu_env
, o
->in2
);
2733 static ExitStatus
op_sqxb(DisasContext
*s
, DisasOps
*o
)
2735 gen_helper_sqxb(o
->out
, cpu_env
, o
->in1
, o
->in2
);
2736 return_low128(o
->out2
);
2740 #ifndef CONFIG_USER_ONLY
2741 static ExitStatus
op_sigp(DisasContext
*s
, DisasOps
*o
)
2743 TCGv_i32 r1
= tcg_const_i32(get_field(s
->fields
, r1
));
2744 check_privileged(s
);
2745 potential_page_fault(s
);
2746 gen_helper_sigp(cc_op
, cpu_env
, o
->in2
, r1
, o
->in1
);
2747 tcg_temp_free_i32(r1
);
2752 static ExitStatus
op_sla(DisasContext
*s
, DisasOps
*o
)
2754 uint64_t sign
= 1ull << s
->insn
->data
;
2755 enum cc_op cco
= s
->insn
->data
== 31 ? CC_OP_SLA_32
: CC_OP_SLA_64
;
2756 gen_op_update2_cc_i64(s
, cco
, o
->in1
, o
->in2
);
2757 tcg_gen_shl_i64(o
->out
, o
->in1
, o
->in2
);
2758 /* The arithmetic left shift is curious in that it does not affect
2759 the sign bit. Copy that over from the source unchanged. */
2760 tcg_gen_andi_i64(o
->out
, o
->out
, ~sign
);
2761 tcg_gen_andi_i64(o
->in1
, o
->in1
, sign
);
2762 tcg_gen_or_i64(o
->out
, o
->out
, o
->in1
);
2766 static ExitStatus
op_sll(DisasContext
*s
, DisasOps
*o
)
2768 tcg_gen_shl_i64(o
->out
, o
->in1
, o
->in2
);
2772 static ExitStatus
op_sra(DisasContext
*s
, DisasOps
*o
)
2774 tcg_gen_sar_i64(o
->out
, o
->in1
, o
->in2
);
2778 static ExitStatus
op_srl(DisasContext
*s
, DisasOps
*o
)
2780 tcg_gen_shr_i64(o
->out
, o
->in1
, o
->in2
);
2784 static ExitStatus
op_sfpc(DisasContext
*s
, DisasOps
*o
)
2786 gen_helper_sfpc(cpu_env
, o
->in2
);
2790 #ifndef CONFIG_USER_ONLY
2791 static ExitStatus
op_spka(DisasContext
*s
, DisasOps
*o
)
2793 check_privileged(s
);
2794 tcg_gen_shri_i64(o
->in2
, o
->in2
, 4);
2795 tcg_gen_deposit_i64(psw_mask
, psw_mask
, o
->in2
, PSW_SHIFT_KEY
- 4, 4);
2799 static ExitStatus
op_sske(DisasContext
*s
, DisasOps
*o
)
2801 check_privileged(s
);
2802 gen_helper_sske(cpu_env
, o
->in1
, o
->in2
);
2806 static ExitStatus
op_ssm(DisasContext
*s
, DisasOps
*o
)
2808 check_privileged(s
);
2809 tcg_gen_deposit_i64(psw_mask
, psw_mask
, o
->in2
, 56, 8);
2813 static ExitStatus
op_stap(DisasContext
*s
, DisasOps
*o
)
2815 check_privileged(s
);
2816 /* ??? Surely cpu address != cpu number. In any case the previous
2817 version of this stored more than the required half-word, so it
2818 is unlikely this has ever been tested. */
2819 tcg_gen_ld32u_i64(o
->out
, cpu_env
, offsetof(CPUS390XState
, cpu_num
));
2823 static ExitStatus
op_stck(DisasContext
*s
, DisasOps
*o
)
2825 gen_helper_stck(o
->out
, cpu_env
);
2826 /* ??? We don't implement clock states. */
2827 gen_op_movi_cc(s
, 0);
2831 static ExitStatus
op_sckc(DisasContext
*s
, DisasOps
*o
)
2833 check_privileged(s
);
2834 gen_helper_sckc(cpu_env
, o
->in2
);
2838 static ExitStatus
op_stckc(DisasContext
*s
, DisasOps
*o
)
2840 check_privileged(s
);
2841 gen_helper_stckc(o
->out
, cpu_env
);
2845 static ExitStatus
op_stctg(DisasContext
*s
, DisasOps
*o
)
2847 TCGv_i32 r1
= tcg_const_i32(get_field(s
->fields
, r1
));
2848 TCGv_i32 r3
= tcg_const_i32(get_field(s
->fields
, r3
));
2849 check_privileged(s
);
2850 potential_page_fault(s
);
2851 gen_helper_stctg(cpu_env
, r1
, o
->in2
, r3
);
2852 tcg_temp_free_i32(r1
);
2853 tcg_temp_free_i32(r3
);
2857 static ExitStatus
op_stctl(DisasContext
*s
, DisasOps
*o
)
2859 TCGv_i32 r1
= tcg_const_i32(get_field(s
->fields
, r1
));
2860 TCGv_i32 r3
= tcg_const_i32(get_field(s
->fields
, r3
));
2861 check_privileged(s
);
2862 potential_page_fault(s
);
2863 gen_helper_stctl(cpu_env
, r1
, o
->in2
, r3
);
2864 tcg_temp_free_i32(r1
);
2865 tcg_temp_free_i32(r3
);
2869 static ExitStatus
op_stidp(DisasContext
*s
, DisasOps
*o
)
2871 check_privileged(s
);
2872 tcg_gen_ld32u_i64(o
->out
, cpu_env
, offsetof(CPUS390XState
, cpu_num
));
2876 static ExitStatus
op_spt(DisasContext
*s
, DisasOps
*o
)
2878 check_privileged(s
);
2879 gen_helper_spt(cpu_env
, o
->in2
);
2883 static ExitStatus
op_stpt(DisasContext
*s
, DisasOps
*o
)
2885 check_privileged(s
);
2886 gen_helper_stpt(o
->out
, cpu_env
);
2890 static ExitStatus
op_spx(DisasContext
*s
, DisasOps
*o
)
2892 check_privileged(s
);
2893 gen_helper_spx(cpu_env
, o
->in2
);
2897 static ExitStatus
op_subchannel(DisasContext
*s
, DisasOps
*o
)
2899 check_privileged(s
);
2900 /* Not operational. */
2901 gen_op_movi_cc(s
, 3);
2905 static ExitStatus
op_stpx(DisasContext
*s
, DisasOps
*o
)
2907 check_privileged(s
);
2908 tcg_gen_ld_i64(o
->out
, cpu_env
, offsetof(CPUS390XState
, psa
));
2909 tcg_gen_andi_i64(o
->out
, o
->out
, 0x7fffe000);
2913 static ExitStatus
op_stnosm(DisasContext
*s
, DisasOps
*o
)
2915 uint64_t i2
= get_field(s
->fields
, i2
);
2918 check_privileged(s
);
2920 /* It is important to do what the instruction name says: STORE THEN.
2921 If we let the output hook perform the store then if we fault and
2922 restart, we'll have the wrong SYSTEM MASK in place. */
2923 t
= tcg_temp_new_i64();
2924 tcg_gen_shri_i64(t
, psw_mask
, 56);
2925 tcg_gen_qemu_st8(t
, o
->addr1
, get_mem_index(s
));
2926 tcg_temp_free_i64(t
);
2928 if (s
->fields
->op
== 0xac) {
2929 tcg_gen_andi_i64(psw_mask
, psw_mask
,
2930 (i2
<< 56) | 0x00ffffffffffffffull
);
2932 tcg_gen_ori_i64(psw_mask
, psw_mask
, i2
<< 56);
2937 static ExitStatus
op_stura(DisasContext
*s
, DisasOps
*o
)
2939 check_privileged(s
);
2940 potential_page_fault(s
);
2941 gen_helper_stura(cpu_env
, o
->in2
, o
->in1
);
2946 static ExitStatus
op_st8(DisasContext
*s
, DisasOps
*o
)
2948 tcg_gen_qemu_st8(o
->in1
, o
->in2
, get_mem_index(s
));
2952 static ExitStatus
op_st16(DisasContext
*s
, DisasOps
*o
)
2954 tcg_gen_qemu_st16(o
->in1
, o
->in2
, get_mem_index(s
));
2958 static ExitStatus
op_st32(DisasContext
*s
, DisasOps
*o
)
2960 tcg_gen_qemu_st32(o
->in1
, o
->in2
, get_mem_index(s
));
2964 static ExitStatus
op_st64(DisasContext
*s
, DisasOps
*o
)
2966 tcg_gen_qemu_st64(o
->in1
, o
->in2
, get_mem_index(s
));
2970 static ExitStatus
op_stam(DisasContext
*s
, DisasOps
*o
)
2972 TCGv_i32 r1
= tcg_const_i32(get_field(s
->fields
, r1
));
2973 TCGv_i32 r3
= tcg_const_i32(get_field(s
->fields
, r3
));
2974 potential_page_fault(s
);
2975 gen_helper_stam(cpu_env
, r1
, o
->in2
, r3
);
2976 tcg_temp_free_i32(r1
);
2977 tcg_temp_free_i32(r3
);
2981 static ExitStatus
op_stcm(DisasContext
*s
, DisasOps
*o
)
2983 int m3
= get_field(s
->fields
, m3
);
2984 int pos
, base
= s
->insn
->data
;
2985 TCGv_i64 tmp
= tcg_temp_new_i64();
2987 pos
= base
+ ctz32(m3
) * 8;
2990 /* Effectively a 32-bit store. */
2991 tcg_gen_shri_i64(tmp
, o
->in1
, pos
);
2992 tcg_gen_qemu_st32(tmp
, o
->in2
, get_mem_index(s
));
2998 /* Effectively a 16-bit store. */
2999 tcg_gen_shri_i64(tmp
, o
->in1
, pos
);
3000 tcg_gen_qemu_st16(tmp
, o
->in2
, get_mem_index(s
));
3007 /* Effectively an 8-bit store. */
3008 tcg_gen_shri_i64(tmp
, o
->in1
, pos
);
3009 tcg_gen_qemu_st8(tmp
, o
->in2
, get_mem_index(s
));
3013 /* This is going to be a sequence of shifts and stores. */
3014 pos
= base
+ 32 - 8;
3017 tcg_gen_shri_i64(tmp
, o
->in1
, pos
);
3018 tcg_gen_qemu_st8(tmp
, o
->in2
, get_mem_index(s
));
3019 tcg_gen_addi_i64(o
->in2
, o
->in2
, 1);
3021 m3
= (m3
<< 1) & 0xf;
3026 tcg_temp_free_i64(tmp
);
3030 static ExitStatus
op_stm(DisasContext
*s
, DisasOps
*o
)
3032 int r1
= get_field(s
->fields
, r1
);
3033 int r3
= get_field(s
->fields
, r3
);
3034 int size
= s
->insn
->data
;
3035 TCGv_i64 tsize
= tcg_const_i64(size
);
3039 tcg_gen_qemu_st64(regs
[r1
], o
->in2
, get_mem_index(s
));
3041 tcg_gen_qemu_st32(regs
[r1
], o
->in2
, get_mem_index(s
));
3046 tcg_gen_add_i64(o
->in2
, o
->in2
, tsize
);
3050 tcg_temp_free_i64(tsize
);
3054 static ExitStatus
op_stmh(DisasContext
*s
, DisasOps
*o
)
3056 int r1
= get_field(s
->fields
, r1
);
3057 int r3
= get_field(s
->fields
, r3
);
3058 TCGv_i64 t
= tcg_temp_new_i64();
3059 TCGv_i64 t4
= tcg_const_i64(4);
3060 TCGv_i64 t32
= tcg_const_i64(32);
3063 tcg_gen_shl_i64(t
, regs
[r1
], t32
);
3064 tcg_gen_qemu_st32(t
, o
->in2
, get_mem_index(s
));
3068 tcg_gen_add_i64(o
->in2
, o
->in2
, t4
);
3072 tcg_temp_free_i64(t
);
3073 tcg_temp_free_i64(t4
);
3074 tcg_temp_free_i64(t32
);
3078 static ExitStatus
op_srst(DisasContext
*s
, DisasOps
*o
)
3080 potential_page_fault(s
);
3081 gen_helper_srst(o
->in1
, cpu_env
, regs
[0], o
->in1
, o
->in2
);
3083 return_low128(o
->in2
);
3087 static ExitStatus
op_sub(DisasContext
*s
, DisasOps
*o
)
3089 tcg_gen_sub_i64(o
->out
, o
->in1
, o
->in2
);
3093 static ExitStatus
op_subb(DisasContext
*s
, DisasOps
*o
)
3098 tcg_gen_not_i64(o
->in2
, o
->in2
);
3099 tcg_gen_add_i64(o
->out
, o
->in1
, o
->in2
);
3101 /* XXX possible optimization point */
3103 cc
= tcg_temp_new_i64();
3104 tcg_gen_extu_i32_i64(cc
, cc_op
);
3105 tcg_gen_shri_i64(cc
, cc
, 1);
3106 tcg_gen_add_i64(o
->out
, o
->out
, cc
);
3107 tcg_temp_free_i64(cc
);
3111 static ExitStatus
op_svc(DisasContext
*s
, DisasOps
*o
)
3118 t
= tcg_const_i32(get_field(s
->fields
, i1
) & 0xff);
3119 tcg_gen_st_i32(t
, cpu_env
, offsetof(CPUS390XState
, int_svc_code
));
3120 tcg_temp_free_i32(t
);
3122 t
= tcg_const_i32(s
->next_pc
- s
->pc
);
3123 tcg_gen_st_i32(t
, cpu_env
, offsetof(CPUS390XState
, int_svc_ilen
));
3124 tcg_temp_free_i32(t
);
3126 gen_exception(EXCP_SVC
);
3127 return EXIT_NORETURN
;
3130 static ExitStatus
op_tceb(DisasContext
*s
, DisasOps
*o
)
3132 gen_helper_tceb(cc_op
, o
->in1
, o
->in2
);
3137 static ExitStatus
op_tcdb(DisasContext
*s
, DisasOps
*o
)
3139 gen_helper_tcdb(cc_op
, o
->in1
, o
->in2
);
3144 static ExitStatus
op_tcxb(DisasContext
*s
, DisasOps
*o
)
3146 gen_helper_tcxb(cc_op
, o
->out
, o
->out2
, o
->in2
);
3151 #ifndef CONFIG_USER_ONLY
3152 static ExitStatus
op_tprot(DisasContext
*s
, DisasOps
*o
)
3154 potential_page_fault(s
);
3155 gen_helper_tprot(cc_op
, o
->addr1
, o
->in2
);
3161 static ExitStatus
op_tr(DisasContext
*s
, DisasOps
*o
)
3163 TCGv_i32 l
= tcg_const_i32(get_field(s
->fields
, l1
));
3164 potential_page_fault(s
);
3165 gen_helper_tr(cpu_env
, l
, o
->addr1
, o
->in2
);
3166 tcg_temp_free_i32(l
);
3171 static ExitStatus
op_unpk(DisasContext
*s
, DisasOps
*o
)
3173 TCGv_i32 l
= tcg_const_i32(get_field(s
->fields
, l1
));
3174 potential_page_fault(s
);
3175 gen_helper_unpk(cpu_env
, l
, o
->addr1
, o
->in2
);
3176 tcg_temp_free_i32(l
);
3180 static ExitStatus
op_xc(DisasContext
*s
, DisasOps
*o
)
3182 TCGv_i32 l
= tcg_const_i32(get_field(s
->fields
, l1
));
3183 potential_page_fault(s
);
3184 gen_helper_xc(cc_op
, cpu_env
, l
, o
->addr1
, o
->in2
);
3185 tcg_temp_free_i32(l
);
3190 static ExitStatus
op_xor(DisasContext
*s
, DisasOps
*o
)
3192 tcg_gen_xor_i64(o
->out
, o
->in1
, o
->in2
);
3196 static ExitStatus
op_xori(DisasContext
*s
, DisasOps
*o
)
3198 int shift
= s
->insn
->data
& 0xff;
3199 int size
= s
->insn
->data
>> 8;
3200 uint64_t mask
= ((1ull << size
) - 1) << shift
;
3203 tcg_gen_shli_i64(o
->in2
, o
->in2
, shift
);
3204 tcg_gen_xor_i64(o
->out
, o
->in1
, o
->in2
);
3206 /* Produce the CC from only the bits manipulated. */
3207 tcg_gen_andi_i64(cc_dst
, o
->out
, mask
);
3208 set_cc_nz_u64(s
, cc_dst
);
3212 static ExitStatus
op_zero(DisasContext
*s
, DisasOps
*o
)
3214 o
->out
= tcg_const_i64(0);
3218 static ExitStatus
op_zero2(DisasContext
*s
, DisasOps
*o
)
3220 o
->out
= tcg_const_i64(0);
3226 /* ====================================================================== */
3227 /* The "Cc OUTput" generators. Given the generated output (and in some cases
3228 the original inputs), update the various cc data structures in order to
3229 be able to compute the new condition code. */
3231 static void cout_abs32(DisasContext
*s
, DisasOps
*o
)
3233 gen_op_update1_cc_i64(s
, CC_OP_ABS_32
, o
->out
);
3236 static void cout_abs64(DisasContext
*s
, DisasOps
*o
)
3238 gen_op_update1_cc_i64(s
, CC_OP_ABS_64
, o
->out
);
3241 static void cout_adds32(DisasContext
*s
, DisasOps
*o
)
3243 gen_op_update3_cc_i64(s
, CC_OP_ADD_32
, o
->in1
, o
->in2
, o
->out
);
3246 static void cout_adds64(DisasContext
*s
, DisasOps
*o
)
3248 gen_op_update3_cc_i64(s
, CC_OP_ADD_64
, o
->in1
, o
->in2
, o
->out
);
3251 static void cout_addu32(DisasContext
*s
, DisasOps
*o
)
3253 gen_op_update3_cc_i64(s
, CC_OP_ADDU_32
, o
->in1
, o
->in2
, o
->out
);
3256 static void cout_addu64(DisasContext
*s
, DisasOps
*o
)
3258 gen_op_update3_cc_i64(s
, CC_OP_ADDU_64
, o
->in1
, o
->in2
, o
->out
);
3261 static void cout_addc32(DisasContext
*s
, DisasOps
*o
)
3263 gen_op_update3_cc_i64(s
, CC_OP_ADDC_32
, o
->in1
, o
->in2
, o
->out
);
3266 static void cout_addc64(DisasContext
*s
, DisasOps
*o
)
3268 gen_op_update3_cc_i64(s
, CC_OP_ADDC_64
, o
->in1
, o
->in2
, o
->out
);
3271 static void cout_cmps32(DisasContext
*s
, DisasOps
*o
)
3273 gen_op_update2_cc_i64(s
, CC_OP_LTGT_32
, o
->in1
, o
->in2
);
3276 static void cout_cmps64(DisasContext
*s
, DisasOps
*o
)
3278 gen_op_update2_cc_i64(s
, CC_OP_LTGT_64
, o
->in1
, o
->in2
);
3281 static void cout_cmpu32(DisasContext
*s
, DisasOps
*o
)
3283 gen_op_update2_cc_i64(s
, CC_OP_LTUGTU_32
, o
->in1
, o
->in2
);
3286 static void cout_cmpu64(DisasContext
*s
, DisasOps
*o
)
3288 gen_op_update2_cc_i64(s
, CC_OP_LTUGTU_64
, o
->in1
, o
->in2
);
3291 static void cout_f32(DisasContext
*s
, DisasOps
*o
)
3293 gen_op_update1_cc_i64(s
, CC_OP_NZ_F32
, o
->out
);
3296 static void cout_f64(DisasContext
*s
, DisasOps
*o
)
3298 gen_op_update1_cc_i64(s
, CC_OP_NZ_F64
, o
->out
);
3301 static void cout_f128(DisasContext
*s
, DisasOps
*o
)
3303 gen_op_update2_cc_i64(s
, CC_OP_NZ_F128
, o
->out
, o
->out2
);
3306 static void cout_nabs32(DisasContext
*s
, DisasOps
*o
)
3308 gen_op_update1_cc_i64(s
, CC_OP_NABS_32
, o
->out
);
3311 static void cout_nabs64(DisasContext
*s
, DisasOps
*o
)
3313 gen_op_update1_cc_i64(s
, CC_OP_NABS_64
, o
->out
);
3316 static void cout_neg32(DisasContext
*s
, DisasOps
*o
)
3318 gen_op_update1_cc_i64(s
, CC_OP_COMP_32
, o
->out
);
3321 static void cout_neg64(DisasContext
*s
, DisasOps
*o
)
3323 gen_op_update1_cc_i64(s
, CC_OP_COMP_64
, o
->out
);
3326 static void cout_nz32(DisasContext
*s
, DisasOps
*o
)
3328 tcg_gen_ext32u_i64(cc_dst
, o
->out
);
3329 gen_op_update1_cc_i64(s
, CC_OP_NZ
, cc_dst
);
3332 static void cout_nz64(DisasContext
*s
, DisasOps
*o
)
3334 gen_op_update1_cc_i64(s
, CC_OP_NZ
, o
->out
);
3337 static void cout_s32(DisasContext
*s
, DisasOps
*o
)
3339 gen_op_update1_cc_i64(s
, CC_OP_LTGT0_32
, o
->out
);
3342 static void cout_s64(DisasContext
*s
, DisasOps
*o
)
3344 gen_op_update1_cc_i64(s
, CC_OP_LTGT0_64
, o
->out
);
3347 static void cout_subs32(DisasContext
*s
, DisasOps
*o
)
3349 gen_op_update3_cc_i64(s
, CC_OP_SUB_32
, o
->in1
, o
->in2
, o
->out
);
3352 static void cout_subs64(DisasContext
*s
, DisasOps
*o
)
3354 gen_op_update3_cc_i64(s
, CC_OP_SUB_64
, o
->in1
, o
->in2
, o
->out
);
3357 static void cout_subu32(DisasContext
*s
, DisasOps
*o
)
3359 gen_op_update3_cc_i64(s
, CC_OP_SUBU_32
, o
->in1
, o
->in2
, o
->out
);
3362 static void cout_subu64(DisasContext
*s
, DisasOps
*o
)
3364 gen_op_update3_cc_i64(s
, CC_OP_SUBU_64
, o
->in1
, o
->in2
, o
->out
);
3367 static void cout_subb32(DisasContext
*s
, DisasOps
*o
)
3369 gen_op_update3_cc_i64(s
, CC_OP_SUBB_32
, o
->in1
, o
->in2
, o
->out
);
3372 static void cout_subb64(DisasContext
*s
, DisasOps
*o
)
3374 gen_op_update3_cc_i64(s
, CC_OP_SUBB_64
, o
->in1
, o
->in2
, o
->out
);
3377 static void cout_tm32(DisasContext
*s
, DisasOps
*o
)
3379 gen_op_update2_cc_i64(s
, CC_OP_TM_32
, o
->in1
, o
->in2
);
3382 static void cout_tm64(DisasContext
*s
, DisasOps
*o
)
3384 gen_op_update2_cc_i64(s
, CC_OP_TM_64
, o
->in1
, o
->in2
);
3387 /* ====================================================================== */
3388 /* The "PREPeration" generators. These initialize the DisasOps.OUT fields
3389 with the TCG register to which we will write. Used in combination with
3390 the "wout" generators, in some cases we need a new temporary, and in
3391 some cases we can write to a TCG global. */
3393 static void prep_new(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
3395 o
->out
= tcg_temp_new_i64();
3398 static void prep_new_P(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
3400 o
->out
= tcg_temp_new_i64();
3401 o
->out2
= tcg_temp_new_i64();
3404 static void prep_r1(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
3406 o
->out
= regs
[get_field(f
, r1
)];
3410 static void prep_r1_P(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
3412 /* ??? Specification exception: r1 must be even. */
3413 int r1
= get_field(f
, r1
);
3415 o
->out2
= regs
[(r1
+ 1) & 15];
3416 o
->g_out
= o
->g_out2
= true;
3419 static void prep_f1(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
3421 o
->out
= fregs
[get_field(f
, r1
)];
3425 static void prep_x1(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
3427 /* ??? Specification exception: r1 must be < 14. */
3428 int r1
= get_field(f
, r1
);
3430 o
->out2
= fregs
[(r1
+ 2) & 15];
3431 o
->g_out
= o
->g_out2
= true;
3434 /* ====================================================================== */
3435 /* The "Write OUTput" generators. These generally perform some non-trivial
3436 copy of data to TCG globals, or to main memory. The trivial cases are
3437 generally handled by having a "prep" generator install the TCG global
3438 as the destination of the operation. */
3440 static void wout_r1(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
3442 store_reg(get_field(f
, r1
), o
->out
);
3445 static void wout_r1_8(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
3447 int r1
= get_field(f
, r1
);
3448 tcg_gen_deposit_i64(regs
[r1
], regs
[r1
], o
->out
, 0, 8);
3451 static void wout_r1_16(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
3453 int r1
= get_field(f
, r1
);
3454 tcg_gen_deposit_i64(regs
[r1
], regs
[r1
], o
->out
, 0, 16);
3457 static void wout_r1_32(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
3459 store_reg32_i64(get_field(f
, r1
), o
->out
);
3462 static void wout_r1_P32(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
3464 /* ??? Specification exception: r1 must be even. */
3465 int r1
= get_field(f
, r1
);
3466 store_reg32_i64(r1
, o
->out
);
3467 store_reg32_i64((r1
+ 1) & 15, o
->out2
);
3470 static void wout_r1_D32(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
3472 /* ??? Specification exception: r1 must be even. */
3473 int r1
= get_field(f
, r1
);
3474 store_reg32_i64((r1
+ 1) & 15, o
->out
);
3475 tcg_gen_shri_i64(o
->out
, o
->out
, 32);
3476 store_reg32_i64(r1
, o
->out
);
3479 static void wout_e1(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
3481 store_freg32_i64(get_field(f
, r1
), o
->out
);
3484 static void wout_f1(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
3486 store_freg(get_field(f
, r1
), o
->out
);
3489 static void wout_x1(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
3491 /* ??? Specification exception: r1 must be < 14. */
3492 int f1
= get_field(s
->fields
, r1
);
3493 store_freg(f1
, o
->out
);
3494 store_freg((f1
+ 2) & 15, o
->out2
);
3497 static void wout_cond_r1r2_32(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
3499 if (get_field(f
, r1
) != get_field(f
, r2
)) {
3500 store_reg32_i64(get_field(f
, r1
), o
->out
);
3504 static void wout_cond_e1e2(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
3506 if (get_field(f
, r1
) != get_field(f
, r2
)) {
3507 store_freg32_i64(get_field(f
, r1
), o
->out
);
3511 static void wout_m1_8(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
3513 tcg_gen_qemu_st8(o
->out
, o
->addr1
, get_mem_index(s
));
3516 static void wout_m1_16(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
3518 tcg_gen_qemu_st16(o
->out
, o
->addr1
, get_mem_index(s
));
3521 static void wout_m1_32(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
3523 tcg_gen_qemu_st32(o
->out
, o
->addr1
, get_mem_index(s
));
3526 static void wout_m1_64(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
3528 tcg_gen_qemu_st64(o
->out
, o
->addr1
, get_mem_index(s
));
3531 static void wout_m2_32(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
3533 tcg_gen_qemu_st32(o
->out
, o
->in2
, get_mem_index(s
));
3536 /* ====================================================================== */
3537 /* The "INput 1" generators. These load the first operand to an insn. */
3539 static void in1_r1(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
3541 o
->in1
= load_reg(get_field(f
, r1
));
3544 static void in1_r1_o(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
3546 o
->in1
= regs
[get_field(f
, r1
)];
3550 static void in1_r1_32s(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
3552 o
->in1
= tcg_temp_new_i64();
3553 tcg_gen_ext32s_i64(o
->in1
, regs
[get_field(f
, r1
)]);
3556 static void in1_r1_32u(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
3558 o
->in1
= tcg_temp_new_i64();
3559 tcg_gen_ext32u_i64(o
->in1
, regs
[get_field(f
, r1
)]);
3562 static void in1_r1_sr32(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
3564 o
->in1
= tcg_temp_new_i64();
3565 tcg_gen_shri_i64(o
->in1
, regs
[get_field(f
, r1
)], 32);
3568 static void in1_r1p1(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
3570 /* ??? Specification exception: r1 must be even. */
3571 int r1
= get_field(f
, r1
);
3572 o
->in1
= load_reg((r1
+ 1) & 15);
3575 static void in1_r1p1_32s(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
3577 /* ??? Specification exception: r1 must be even. */
3578 int r1
= get_field(f
, r1
);
3579 o
->in1
= tcg_temp_new_i64();
3580 tcg_gen_ext32s_i64(o
->in1
, regs
[(r1
+ 1) & 15]);
3583 static void in1_r1p1_32u(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
3585 /* ??? Specification exception: r1 must be even. */
3586 int r1
= get_field(f
, r1
);
3587 o
->in1
= tcg_temp_new_i64();
3588 tcg_gen_ext32u_i64(o
->in1
, regs
[(r1
+ 1) & 15]);
3591 static void in1_r1_D32(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
3593 /* ??? Specification exception: r1 must be even. */
3594 int r1
= get_field(f
, r1
);
3595 o
->in1
= tcg_temp_new_i64();
3596 tcg_gen_concat32_i64(o
->in1
, regs
[r1
+ 1], regs
[r1
]);
3599 static void in1_r2(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
3601 o
->in1
= load_reg(get_field(f
, r2
));
3604 static void in1_r3(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
3606 o
->in1
= load_reg(get_field(f
, r3
));
3609 static void in1_r3_o(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
3611 o
->in1
= regs
[get_field(f
, r3
)];
3615 static void in1_r3_32s(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
3617 o
->in1
= tcg_temp_new_i64();
3618 tcg_gen_ext32s_i64(o
->in1
, regs
[get_field(f
, r3
)]);
3621 static void in1_r3_32u(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
3623 o
->in1
= tcg_temp_new_i64();
3624 tcg_gen_ext32u_i64(o
->in1
, regs
[get_field(f
, r3
)]);
3627 static void in1_e1(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
3629 o
->in1
= load_freg32_i64(get_field(f
, r1
));
3632 static void in1_f1_o(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
3634 o
->in1
= fregs
[get_field(f
, r1
)];
3638 static void in1_x1_o(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
3640 /* ??? Specification exception: r1 must be < 14. */
3641 int r1
= get_field(f
, r1
);
3643 o
->out2
= fregs
[(r1
+ 2) & 15];
3644 o
->g_out
= o
->g_out2
= true;
3647 static void in1_la1(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
3649 o
->addr1
= get_address(s
, 0, get_field(f
, b1
), get_field(f
, d1
));
3652 static void in1_la2(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
3654 int x2
= have_field(f
, x2
) ? get_field(f
, x2
) : 0;
3655 o
->addr1
= get_address(s
, x2
, get_field(f
, b2
), get_field(f
, d2
));
3658 static void in1_m1_8u(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
3661 o
->in1
= tcg_temp_new_i64();
3662 tcg_gen_qemu_ld8u(o
->in1
, o
->addr1
, get_mem_index(s
));
3665 static void in1_m1_16s(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
3668 o
->in1
= tcg_temp_new_i64();
3669 tcg_gen_qemu_ld16s(o
->in1
, o
->addr1
, get_mem_index(s
));
3672 static void in1_m1_16u(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
3675 o
->in1
= tcg_temp_new_i64();
3676 tcg_gen_qemu_ld16u(o
->in1
, o
->addr1
, get_mem_index(s
));
3679 static void in1_m1_32s(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
3682 o
->in1
= tcg_temp_new_i64();
3683 tcg_gen_qemu_ld32s(o
->in1
, o
->addr1
, get_mem_index(s
));
3686 static void in1_m1_32u(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
3689 o
->in1
= tcg_temp_new_i64();
3690 tcg_gen_qemu_ld32u(o
->in1
, o
->addr1
, get_mem_index(s
));
3693 static void in1_m1_64(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
3696 o
->in1
= tcg_temp_new_i64();
3697 tcg_gen_qemu_ld64(o
->in1
, o
->addr1
, get_mem_index(s
));
3700 /* ====================================================================== */
3701 /* The "INput 2" generators. These load the second operand to an insn. */
3703 static void in2_r1_o(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
3705 o
->in2
= regs
[get_field(f
, r1
)];
3709 static void in2_r1_16u(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
3711 o
->in2
= tcg_temp_new_i64();
3712 tcg_gen_ext16u_i64(o
->in2
, regs
[get_field(f
, r1
)]);
3715 static void in2_r1_32u(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
3717 o
->in2
= tcg_temp_new_i64();
3718 tcg_gen_ext32u_i64(o
->in2
, regs
[get_field(f
, r1
)]);
3721 static void in2_r2(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
3723 o
->in2
= load_reg(get_field(f
, r2
));
3726 static void in2_r2_o(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
3728 o
->in2
= regs
[get_field(f
, r2
)];
3732 static void in2_r2_nz(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
3734 int r2
= get_field(f
, r2
);
3736 o
->in2
= load_reg(r2
);
3740 static void in2_r2_8s(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
3742 o
->in2
= tcg_temp_new_i64();
3743 tcg_gen_ext8s_i64(o
->in2
, regs
[get_field(f
, r2
)]);
3746 static void in2_r2_8u(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
3748 o
->in2
= tcg_temp_new_i64();
3749 tcg_gen_ext8u_i64(o
->in2
, regs
[get_field(f
, r2
)]);
3752 static void in2_r2_16s(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
3754 o
->in2
= tcg_temp_new_i64();
3755 tcg_gen_ext16s_i64(o
->in2
, regs
[get_field(f
, r2
)]);
3758 static void in2_r2_16u(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
3760 o
->in2
= tcg_temp_new_i64();
3761 tcg_gen_ext16u_i64(o
->in2
, regs
[get_field(f
, r2
)]);
3764 static void in2_r3(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
3766 o
->in2
= load_reg(get_field(f
, r3
));
3769 static void in2_r2_32s(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
3771 o
->in2
= tcg_temp_new_i64();
3772 tcg_gen_ext32s_i64(o
->in2
, regs
[get_field(f
, r2
)]);
3775 static void in2_r2_32u(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
3777 o
->in2
= tcg_temp_new_i64();
3778 tcg_gen_ext32u_i64(o
->in2
, regs
[get_field(f
, r2
)]);
3781 static void in2_e2(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
3783 o
->in2
= load_freg32_i64(get_field(f
, r2
));
3786 static void in2_f2_o(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
3788 o
->in2
= fregs
[get_field(f
, r2
)];
3792 static void in2_x2_o(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
3794 /* ??? Specification exception: r1 must be < 14. */
3795 int r2
= get_field(f
, r2
);
3797 o
->in2
= fregs
[(r2
+ 2) & 15];
3798 o
->g_in1
= o
->g_in2
= true;
3801 static void in2_ra2(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
3803 o
->in2
= get_address(s
, 0, get_field(f
, r2
), 0);
3806 static void in2_a2(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
3808 int x2
= have_field(f
, x2
) ? get_field(f
, x2
) : 0;
3809 o
->in2
= get_address(s
, x2
, get_field(f
, b2
), get_field(f
, d2
));
3812 static void in2_ri2(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
3814 o
->in2
= tcg_const_i64(s
->pc
+ (int64_t)get_field(f
, i2
) * 2);
3817 static void in2_sh32(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
3819 help_l2_shift(s
, f
, o
, 31);
3822 static void in2_sh64(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
3824 help_l2_shift(s
, f
, o
, 63);
3827 static void in2_m2_8u(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
3830 tcg_gen_qemu_ld8u(o
->in2
, o
->in2
, get_mem_index(s
));
3833 static void in2_m2_16s(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
3836 tcg_gen_qemu_ld16s(o
->in2
, o
->in2
, get_mem_index(s
));
3839 static void in2_m2_16u(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
3842 tcg_gen_qemu_ld16u(o
->in2
, o
->in2
, get_mem_index(s
));
3845 static void in2_m2_32s(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
3848 tcg_gen_qemu_ld32s(o
->in2
, o
->in2
, get_mem_index(s
));
3851 static void in2_m2_32u(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
3854 tcg_gen_qemu_ld32u(o
->in2
, o
->in2
, get_mem_index(s
));
3857 static void in2_m2_64(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
3860 tcg_gen_qemu_ld64(o
->in2
, o
->in2
, get_mem_index(s
));
3863 static void in2_mri2_16u(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
3866 tcg_gen_qemu_ld16u(o
->in2
, o
->in2
, get_mem_index(s
));
3869 static void in2_mri2_32s(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
3872 tcg_gen_qemu_ld32s(o
->in2
, o
->in2
, get_mem_index(s
));
3875 static void in2_mri2_32u(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
3878 tcg_gen_qemu_ld32u(o
->in2
, o
->in2
, get_mem_index(s
));
3881 static void in2_mri2_64(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
3884 tcg_gen_qemu_ld64(o
->in2
, o
->in2
, get_mem_index(s
));
3887 static void in2_i2(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
3889 o
->in2
= tcg_const_i64(get_field(f
, i2
));
3892 static void in2_i2_8u(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
3894 o
->in2
= tcg_const_i64((uint8_t)get_field(f
, i2
));
3897 static void in2_i2_16u(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
3899 o
->in2
= tcg_const_i64((uint16_t)get_field(f
, i2
));
3902 static void in2_i2_32u(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
3904 o
->in2
= tcg_const_i64((uint32_t)get_field(f
, i2
));
3907 static void in2_i2_16u_shl(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
3909 uint64_t i2
= (uint16_t)get_field(f
, i2
);
3910 o
->in2
= tcg_const_i64(i2
<< s
->insn
->data
);
3913 static void in2_i2_32u_shl(DisasContext
*s
, DisasFields
*f
, DisasOps
*o
)
3915 uint64_t i2
= (uint32_t)get_field(f
, i2
);
3916 o
->in2
= tcg_const_i64(i2
<< s
->insn
->data
);
3919 /* ====================================================================== */
3921 /* Find opc within the table of insns. This is formulated as a switch
3922 statement so that (1) we get compile-time notice of cut-paste errors
3923 for duplicated opcodes, and (2) the compiler generates the binary
3924 search tree, rather than us having to post-process the table. */
3926 #define C(OPC, NM, FT, FC, I1, I2, P, W, OP, CC) \
3927 D(OPC, NM, FT, FC, I1, I2, P, W, OP, CC, 0)
3929 #define D(OPC, NM, FT, FC, I1, I2, P, W, OP, CC, D) insn_ ## NM,
3931 enum DisasInsnEnum
{
3932 #include "insn-data.def"
3936 #define D(OPC, NM, FT, FC, I1, I2, P, W, OP, CC, D) { \
3941 .help_in1 = in1_##I1, \
3942 .help_in2 = in2_##I2, \
3943 .help_prep = prep_##P, \
3944 .help_wout = wout_##W, \
3945 .help_cout = cout_##CC, \
3946 .help_op = op_##OP, \
3950 /* Allow 0 to be used for NULL in the table below. */
3958 static const DisasInsn insn_info
[] = {
3959 #include "insn-data.def"
3963 #define D(OPC, NM, FT, FC, I1, I2, P, W, OP, CC, D) \
3964 case OPC: return &insn_info[insn_ ## NM];
3966 static const DisasInsn
*lookup_opc(uint16_t opc
)
3969 #include "insn-data.def"
3978 /* Extract a field from the insn. The INSN should be left-aligned in
3979 the uint64_t so that we can more easily utilize the big-bit-endian
3980 definitions we extract from the Principals of Operation. */
3982 static void extract_field(DisasFields
*o
, const DisasField
*f
, uint64_t insn
)
3990 /* Zero extract the field from the insn. */
3991 r
= (insn
<< f
->beg
) >> (64 - f
->size
);
3993 /* Sign-extend, or un-swap the field as necessary. */
3995 case 0: /* unsigned */
3997 case 1: /* signed */
3998 assert(f
->size
<= 32);
3999 m
= 1u << (f
->size
- 1);
4002 case 2: /* dl+dh split, signed 20 bit. */
4003 r
= ((int8_t)r
<< 12) | (r
>> 8);
4009 /* Validate that the "compressed" encoding we selected above is valid.
4010 I.e. we havn't make two different original fields overlap. */
4011 assert(((o
->presentC
>> f
->indexC
) & 1) == 0);
4012 o
->presentC
|= 1 << f
->indexC
;
4013 o
->presentO
|= 1 << f
->indexO
;
4015 o
->c
[f
->indexC
] = r
;
4018 /* Lookup the insn at the current PC, extracting the operands into O and
4019 returning the info struct for the insn. Returns NULL for invalid insn. */
4021 static const DisasInsn
*extract_insn(CPUS390XState
*env
, DisasContext
*s
,
4024 uint64_t insn
, pc
= s
->pc
;
4026 const DisasInsn
*info
;
4028 insn
= ld_code2(env
, pc
);
4029 op
= (insn
>> 8) & 0xff;
4030 ilen
= get_ilen(op
);
4031 s
->next_pc
= s
->pc
+ ilen
;
4038 insn
= ld_code4(env
, pc
) << 32;
4041 insn
= (insn
<< 48) | (ld_code4(env
, pc
+ 2) << 16);
4047 /* We can't actually determine the insn format until we've looked up
4048 the full insn opcode. Which we can't do without locating the
4049 secondary opcode. Assume by default that OP2 is at bit 40; for
4050 those smaller insns that don't actually have a secondary opcode
4051 this will correctly result in OP2 = 0. */
4057 case 0xb2: /* S, RRF, RRE */
4058 case 0xb3: /* RRE, RRD, RRF */
4059 case 0xb9: /* RRE, RRF */
4060 case 0xe5: /* SSE, SIL */
4061 op2
= (insn
<< 8) >> 56;
4065 case 0xc0: /* RIL */
4066 case 0xc2: /* RIL */
4067 case 0xc4: /* RIL */
4068 case 0xc6: /* RIL */
4069 case 0xc8: /* SSF */
4070 case 0xcc: /* RIL */
4071 op2
= (insn
<< 12) >> 60;
4073 case 0xd0 ... 0xdf: /* SS */
4079 case 0xee ... 0xf3: /* SS */
4080 case 0xf8 ... 0xfd: /* SS */
4084 op2
= (insn
<< 40) >> 56;
4088 memset(f
, 0, sizeof(*f
));
4092 /* Lookup the instruction. */
4093 info
= lookup_opc(op
<< 8 | op2
);
4095 /* If we found it, extract the operands. */
4097 DisasFormat fmt
= info
->fmt
;
4100 for (i
= 0; i
< NUM_C_FIELD
; ++i
) {
4101 extract_field(f
, &format_info
[fmt
].op
[i
], insn
);
4107 static ExitStatus
translate_one(CPUS390XState
*env
, DisasContext
*s
)
4109 const DisasInsn
*insn
;
4110 ExitStatus ret
= NO_EXIT
;
4114 insn
= extract_insn(env
, s
, &f
);
4116 /* If not found, try the old interpreter. This includes ILLOPC. */
4118 disas_s390_insn(env
, s
);
4119 switch (s
->is_jmp
) {
4127 ret
= EXIT_PC_UPDATED
;
4130 ret
= EXIT_NORETURN
;
4140 /* Set up the strutures we use to communicate with the helpers. */
4143 o
.g_out
= o
.g_out2
= o
.g_in1
= o
.g_in2
= false;
4144 TCGV_UNUSED_I64(o
.out
);
4145 TCGV_UNUSED_I64(o
.out2
);
4146 TCGV_UNUSED_I64(o
.in1
);
4147 TCGV_UNUSED_I64(o
.in2
);
4148 TCGV_UNUSED_I64(o
.addr1
);
4150 /* Implement the instruction. */
4151 if (insn
->help_in1
) {
4152 insn
->help_in1(s
, &f
, &o
);
4154 if (insn
->help_in2
) {
4155 insn
->help_in2(s
, &f
, &o
);
4157 if (insn
->help_prep
) {
4158 insn
->help_prep(s
, &f
, &o
);
4160 if (insn
->help_op
) {
4161 ret
= insn
->help_op(s
, &o
);
4163 if (insn
->help_wout
) {
4164 insn
->help_wout(s
, &f
, &o
);
4166 if (insn
->help_cout
) {
4167 insn
->help_cout(s
, &o
);
4170 /* Free any temporaries created by the helpers. */
4171 if (!TCGV_IS_UNUSED_I64(o
.out
) && !o
.g_out
) {
4172 tcg_temp_free_i64(o
.out
);
4174 if (!TCGV_IS_UNUSED_I64(o
.out2
) && !o
.g_out2
) {
4175 tcg_temp_free_i64(o
.out2
);
4177 if (!TCGV_IS_UNUSED_I64(o
.in1
) && !o
.g_in1
) {
4178 tcg_temp_free_i64(o
.in1
);
4180 if (!TCGV_IS_UNUSED_I64(o
.in2
) && !o
.g_in2
) {
4181 tcg_temp_free_i64(o
.in2
);
4183 if (!TCGV_IS_UNUSED_I64(o
.addr1
)) {
4184 tcg_temp_free_i64(o
.addr1
);
4187 /* Advance to the next instruction. */
4192 static inline void gen_intermediate_code_internal(CPUS390XState
*env
,
4193 TranslationBlock
*tb
,
4197 target_ulong pc_start
;
4198 uint64_t next_page_start
;
4199 uint16_t *gen_opc_end
;
4201 int num_insns
, max_insns
;
4209 if (!(tb
->flags
& FLAG_MASK_64
)) {
4210 pc_start
&= 0x7fffffff;
4215 dc
.cc_op
= CC_OP_DYNAMIC
;
4216 do_debug
= dc
.singlestep_enabled
= env
->singlestep_enabled
;
4217 dc
.is_jmp
= DISAS_NEXT
;
4219 gen_opc_end
= tcg_ctx
.gen_opc_buf
+ OPC_MAX_SIZE
;
4221 next_page_start
= (pc_start
& TARGET_PAGE_MASK
) + TARGET_PAGE_SIZE
;
4224 max_insns
= tb
->cflags
& CF_COUNT_MASK
;
4225 if (max_insns
== 0) {
4226 max_insns
= CF_COUNT_MASK
;
4233 j
= tcg_ctx
.gen_opc_ptr
- tcg_ctx
.gen_opc_buf
;
4237 tcg_ctx
.gen_opc_instr_start
[lj
++] = 0;
4240 tcg_ctx
.gen_opc_pc
[lj
] = dc
.pc
;
4241 gen_opc_cc_op
[lj
] = dc
.cc_op
;
4242 tcg_ctx
.gen_opc_instr_start
[lj
] = 1;
4243 tcg_ctx
.gen_opc_icount
[lj
] = num_insns
;
4245 if (++num_insns
== max_insns
&& (tb
->cflags
& CF_LAST_IO
)) {
4249 if (unlikely(qemu_loglevel_mask(CPU_LOG_TB_OP
| CPU_LOG_TB_OP_OPT
))) {
4250 tcg_gen_debug_insn_start(dc
.pc
);
4254 if (unlikely(!QTAILQ_EMPTY(&env
->breakpoints
))) {
4255 QTAILQ_FOREACH(bp
, &env
->breakpoints
, entry
) {
4256 if (bp
->pc
== dc
.pc
) {
4257 status
= EXIT_PC_STALE
;
4263 if (status
== NO_EXIT
) {
4264 status
= translate_one(env
, &dc
);
4267 /* If we reach a page boundary, are single stepping,
4268 or exhaust instruction count, stop generation. */
4269 if (status
== NO_EXIT
4270 && (dc
.pc
>= next_page_start
4271 || tcg_ctx
.gen_opc_ptr
>= gen_opc_end
4272 || num_insns
>= max_insns
4274 || env
->singlestep_enabled
)) {
4275 status
= EXIT_PC_STALE
;
4277 } while (status
== NO_EXIT
);
4279 if (tb
->cflags
& CF_LAST_IO
) {
4288 update_psw_addr(&dc
);
4290 case EXIT_PC_UPDATED
:
4291 if (singlestep
&& dc
.cc_op
!= CC_OP_DYNAMIC
) {
4292 gen_op_calc_cc(&dc
);
4294 /* Next TB starts off with CC_OP_DYNAMIC,
4295 so make sure the cc op type is in env */
4296 gen_op_set_cc_op(&dc
);
4299 gen_exception(EXCP_DEBUG
);
4301 /* Generate the return instruction */
4309 gen_icount_end(tb
, num_insns
);
4310 *tcg_ctx
.gen_opc_ptr
= INDEX_op_end
;
4312 j
= tcg_ctx
.gen_opc_ptr
- tcg_ctx
.gen_opc_buf
;
4315 tcg_ctx
.gen_opc_instr_start
[lj
++] = 0;
4318 tb
->size
= dc
.pc
- pc_start
;
4319 tb
->icount
= num_insns
;
4322 #if defined(S390X_DEBUG_DISAS)
4323 if (qemu_loglevel_mask(CPU_LOG_TB_IN_ASM
)) {
4324 qemu_log("IN: %s\n", lookup_symbol(pc_start
));
4325 log_target_disas(env
, pc_start
, dc
.pc
- pc_start
, 1);
4331 void gen_intermediate_code (CPUS390XState
*env
, struct TranslationBlock
*tb
)
4333 gen_intermediate_code_internal(env
, tb
, 0);
4336 void gen_intermediate_code_pc (CPUS390XState
*env
, struct TranslationBlock
*tb
)
4338 gen_intermediate_code_internal(env
, tb
, 1);
4341 void restore_state_to_opc(CPUS390XState
*env
, TranslationBlock
*tb
, int pc_pos
)
4344 env
->psw
.addr
= tcg_ctx
.gen_opc_pc
[pc_pos
];
4345 cc_op
= gen_opc_cc_op
[pc_pos
];
4346 if ((cc_op
!= CC_OP_DYNAMIC
) && (cc_op
!= CC_OP_STATIC
)) {