]> git.proxmox.com Git - mirror_qemu.git/blame - tcg/s390/tcg-target.inc.c
cpu: Move the softmmu tlb to CPUNegativeOffsetState
[mirror_qemu.git] / tcg / s390 / tcg-target.inc.c
CommitLineData
2827822e
AG
1/*
2 * Tiny Code Generator for QEMU
3 *
4 * Copyright (c) 2009 Ulrich Hecht <uli@suse.de>
48bb3750
RH
5 * Copyright (c) 2009 Alexander Graf <agraf@suse.de>
6 * Copyright (c) 2010 Richard Henderson <rth@twiddle.net>
2827822e
AG
7 *
8 * Permission is hereby granted, free of charge, to any person obtaining a copy
9 * of this software and associated documentation files (the "Software"), to deal
10 * in the Software without restriction, including without limitation the rights
11 * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
12 * copies of the Software, and to permit persons to whom the Software is
13 * furnished to do so, subject to the following conditions:
14 *
15 * The above copyright notice and this permission notice shall be included in
16 * all copies or substantial portions of the Software.
17 *
18 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
19 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
20 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
21 * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
22 * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
23 * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
24 * THE SOFTWARE.
25 */
26
a01fc30d
RH
27/* We only support generating code for 64-bit mode. */
28#if TCG_TARGET_REG_BITS != 64
29#error "unsupported code generation mode"
30#endif
31
28eef8aa 32#include "tcg-pool.inc.c"
c9baa30f
RH
33#include "elf.h"
34
48bb3750
RH
35/* ??? The translation blocks produced by TCG are generally small enough to
36 be entirely reachable with a 16-bit displacement. Leaving the option for
37 a 32-bit displacement here Just In Case. */
38#define USE_LONG_BRANCHES 0
39
a8f0269e
RH
40#define TCG_CT_CONST_S16 0x100
41#define TCG_CT_CONST_S32 0x200
a534bb15
RH
42#define TCG_CT_CONST_S33 0x400
43#define TCG_CT_CONST_ZERO 0x800
48bb3750
RH
44
45/* Several places within the instruction set 0 means "no register"
46 rather than TCG_REG_R0. */
47#define TCG_REG_NONE 0
48
49/* A scratch register that may be be used throughout the backend. */
ce411066 50#define TCG_TMP0 TCG_REG_R1
48bb3750 51
829e1376
RH
52/* A scratch register that holds a pointer to the beginning of the TB.
53 We don't need this when we have pc-relative loads with the general
54 instructions extension facility. */
55#define TCG_REG_TB TCG_REG_R12
56#define USE_REG_TB (!(s390_facilities & FACILITY_GEN_INST_EXT))
57
4cbea598 58#ifndef CONFIG_SOFTMMU
48bb3750 59#define TCG_GUEST_BASE_REG TCG_REG_R13
48bb3750
RH
60#endif
61
48bb3750
RH
62/* All of the following instructions are prefixed with their instruction
63 format, and are defined as 8- or 16-bit quantities, even when the two
64 halves of the 16-bit quantity may appear 32 bits apart in the insn.
65 This makes it easy to copy the values from the tables in Appendix B. */
66typedef enum S390Opcode {
67 RIL_AFI = 0xc209,
68 RIL_AGFI = 0xc208,
3790b918 69 RIL_ALFI = 0xc20b,
48bb3750
RH
70 RIL_ALGFI = 0xc20a,
71 RIL_BRASL = 0xc005,
72 RIL_BRCL = 0xc004,
73 RIL_CFI = 0xc20d,
74 RIL_CGFI = 0xc20c,
75 RIL_CLFI = 0xc20f,
76 RIL_CLGFI = 0xc20e,
a534bb15
RH
77 RIL_CLRL = 0xc60f,
78 RIL_CLGRL = 0xc60a,
79 RIL_CRL = 0xc60d,
80 RIL_CGRL = 0xc608,
48bb3750
RH
81 RIL_IIHF = 0xc008,
82 RIL_IILF = 0xc009,
83 RIL_LARL = 0xc000,
84 RIL_LGFI = 0xc001,
85 RIL_LGRL = 0xc408,
86 RIL_LLIHF = 0xc00e,
87 RIL_LLILF = 0xc00f,
88 RIL_LRL = 0xc40d,
89 RIL_MSFI = 0xc201,
90 RIL_MSGFI = 0xc200,
91 RIL_NIHF = 0xc00a,
92 RIL_NILF = 0xc00b,
93 RIL_OIHF = 0xc00c,
94 RIL_OILF = 0xc00d,
3790b918 95 RIL_SLFI = 0xc205,
0db921e6 96 RIL_SLGFI = 0xc204,
48bb3750
RH
97 RIL_XIHF = 0xc006,
98 RIL_XILF = 0xc007,
99
100 RI_AGHI = 0xa70b,
101 RI_AHI = 0xa70a,
102 RI_BRC = 0xa704,
a534bb15
RH
103 RI_CHI = 0xa70e,
104 RI_CGHI = 0xa70f,
48bb3750
RH
105 RI_IIHH = 0xa500,
106 RI_IIHL = 0xa501,
107 RI_IILH = 0xa502,
108 RI_IILL = 0xa503,
109 RI_LGHI = 0xa709,
110 RI_LLIHH = 0xa50c,
111 RI_LLIHL = 0xa50d,
112 RI_LLILH = 0xa50e,
113 RI_LLILL = 0xa50f,
114 RI_MGHI = 0xa70d,
115 RI_MHI = 0xa70c,
116 RI_NIHH = 0xa504,
117 RI_NIHL = 0xa505,
118 RI_NILH = 0xa506,
119 RI_NILL = 0xa507,
120 RI_OIHH = 0xa508,
121 RI_OIHL = 0xa509,
122 RI_OILH = 0xa50a,
123 RI_OILL = 0xa50b,
124
125 RIE_CGIJ = 0xec7c,
126 RIE_CGRJ = 0xec64,
127 RIE_CIJ = 0xec7e,
128 RIE_CLGRJ = 0xec65,
129 RIE_CLIJ = 0xec7f,
130 RIE_CLGIJ = 0xec7d,
131 RIE_CLRJ = 0xec77,
132 RIE_CRJ = 0xec76,
7af525af 133 RIE_LOCGHI = 0xec46,
d5690ea4 134 RIE_RISBG = 0xec55,
48bb3750
RH
135
136 RRE_AGR = 0xb908,
3790b918
RH
137 RRE_ALGR = 0xb90a,
138 RRE_ALCR = 0xb998,
139 RRE_ALCGR = 0xb988,
48bb3750
RH
140 RRE_CGR = 0xb920,
141 RRE_CLGR = 0xb921,
142 RRE_DLGR = 0xb987,
143 RRE_DLR = 0xb997,
144 RRE_DSGFR = 0xb91d,
145 RRE_DSGR = 0xb90d,
ce411066 146 RRE_FLOGR = 0xb983,
48bb3750
RH
147 RRE_LGBR = 0xb906,
148 RRE_LCGR = 0xb903,
149 RRE_LGFR = 0xb914,
150 RRE_LGHR = 0xb907,
151 RRE_LGR = 0xb904,
152 RRE_LLGCR = 0xb984,
153 RRE_LLGFR = 0xb916,
154 RRE_LLGHR = 0xb985,
155 RRE_LRVR = 0xb91f,
156 RRE_LRVGR = 0xb90f,
157 RRE_LTGR = 0xb902,
36017dc6 158 RRE_MLGR = 0xb986,
48bb3750
RH
159 RRE_MSGR = 0xb90c,
160 RRE_MSR = 0xb252,
161 RRE_NGR = 0xb980,
162 RRE_OGR = 0xb981,
163 RRE_SGR = 0xb909,
3790b918
RH
164 RRE_SLGR = 0xb90b,
165 RRE_SLBR = 0xb999,
166 RRE_SLBGR = 0xb989,
48bb3750
RH
167 RRE_XGR = 0xb982,
168
96a9f093
RH
169 RRF_LOCR = 0xb9f2,
170 RRF_LOCGR = 0xb9e2,
c2097136
RH
171 RRF_NRK = 0xb9f4,
172 RRF_NGRK = 0xb9e4,
173 RRF_ORK = 0xb9f6,
174 RRF_OGRK = 0xb9e6,
175 RRF_SRK = 0xb9f9,
176 RRF_SGRK = 0xb9e9,
177 RRF_SLRK = 0xb9fb,
178 RRF_SLGRK = 0xb9eb,
179 RRF_XRK = 0xb9f7,
180 RRF_XGRK = 0xb9e7,
96a9f093 181
48bb3750 182 RR_AR = 0x1a,
3790b918 183 RR_ALR = 0x1e,
48bb3750
RH
184 RR_BASR = 0x0d,
185 RR_BCR = 0x07,
186 RR_CLR = 0x15,
187 RR_CR = 0x19,
188 RR_DR = 0x1d,
189 RR_LCR = 0x13,
190 RR_LR = 0x18,
191 RR_LTR = 0x12,
192 RR_NR = 0x14,
193 RR_OR = 0x16,
194 RR_SR = 0x1b,
3790b918 195 RR_SLR = 0x1f,
48bb3750
RH
196 RR_XR = 0x17,
197
198 RSY_RLL = 0xeb1d,
199 RSY_RLLG = 0xeb1c,
200 RSY_SLLG = 0xeb0d,
c2097136 201 RSY_SLLK = 0xebdf,
48bb3750 202 RSY_SRAG = 0xeb0a,
c2097136 203 RSY_SRAK = 0xebdc,
48bb3750 204 RSY_SRLG = 0xeb0c,
c2097136 205 RSY_SRLK = 0xebde,
48bb3750
RH
206
207 RS_SLL = 0x89,
208 RS_SRA = 0x8a,
209 RS_SRL = 0x88,
210
211 RXY_AG = 0xe308,
212 RXY_AY = 0xe35a,
213 RXY_CG = 0xe320,
a534bb15
RH
214 RXY_CLG = 0xe321,
215 RXY_CLY = 0xe355,
48bb3750 216 RXY_CY = 0xe359,
0db921e6 217 RXY_LAY = 0xe371,
48bb3750
RH
218 RXY_LB = 0xe376,
219 RXY_LG = 0xe304,
220 RXY_LGB = 0xe377,
221 RXY_LGF = 0xe314,
222 RXY_LGH = 0xe315,
223 RXY_LHY = 0xe378,
224 RXY_LLGC = 0xe390,
225 RXY_LLGF = 0xe316,
226 RXY_LLGH = 0xe391,
227 RXY_LMG = 0xeb04,
228 RXY_LRV = 0xe31e,
229 RXY_LRVG = 0xe30f,
230 RXY_LRVH = 0xe31f,
231 RXY_LY = 0xe358,
bdcd5d19 232 RXY_NG = 0xe380,
4046d9ca 233 RXY_OG = 0xe381,
48bb3750
RH
234 RXY_STCY = 0xe372,
235 RXY_STG = 0xe324,
236 RXY_STHY = 0xe370,
237 RXY_STMG = 0xeb24,
238 RXY_STRV = 0xe33e,
239 RXY_STRVG = 0xe32f,
240 RXY_STRVH = 0xe33f,
241 RXY_STY = 0xe350,
5bf67a92 242 RXY_XG = 0xe382,
48bb3750
RH
243
244 RX_A = 0x5a,
245 RX_C = 0x59,
246 RX_L = 0x58,
0db921e6 247 RX_LA = 0x41,
48bb3750
RH
248 RX_LH = 0x48,
249 RX_ST = 0x50,
250 RX_STC = 0x42,
251 RX_STH = 0x40,
ed3d51ec
SF
252
253 NOP = 0x0707,
48bb3750
RH
254} S390Opcode;
255
8d8fdbae 256#ifdef CONFIG_DEBUG_TCG
48bb3750
RH
257static const char * const tcg_target_reg_names[TCG_TARGET_NB_REGS] = {
258 "%r0", "%r1", "%r2", "%r3", "%r4", "%r5", "%r6", "%r7",
259 "%r8", "%r9", "%r10" "%r11" "%r12" "%r13" "%r14" "%r15"
260};
261#endif
262
263/* Since R6 is a potential argument register, choose it last of the
264 call-saved registers. Likewise prefer the call-clobbered registers
265 in reverse order to maximize the chance of avoiding the arguments. */
2827822e 266static const int tcg_target_reg_alloc_order[] = {
f24efee4 267 /* Call saved registers. */
48bb3750
RH
268 TCG_REG_R13,
269 TCG_REG_R12,
270 TCG_REG_R11,
271 TCG_REG_R10,
272 TCG_REG_R9,
273 TCG_REG_R8,
274 TCG_REG_R7,
275 TCG_REG_R6,
f24efee4 276 /* Call clobbered registers. */
48bb3750
RH
277 TCG_REG_R14,
278 TCG_REG_R0,
279 TCG_REG_R1,
f24efee4 280 /* Argument registers, in reverse order of allocation. */
48bb3750
RH
281 TCG_REG_R5,
282 TCG_REG_R4,
283 TCG_REG_R3,
284 TCG_REG_R2,
2827822e
AG
285};
286
287static const int tcg_target_call_iarg_regs[] = {
48bb3750
RH
288 TCG_REG_R2,
289 TCG_REG_R3,
290 TCG_REG_R4,
291 TCG_REG_R5,
292 TCG_REG_R6,
2827822e
AG
293};
294
295static const int tcg_target_call_oarg_regs[] = {
48bb3750 296 TCG_REG_R2,
48bb3750
RH
297};
298
299#define S390_CC_EQ 8
300#define S390_CC_LT 4
301#define S390_CC_GT 2
302#define S390_CC_OV 1
303#define S390_CC_NE (S390_CC_LT | S390_CC_GT)
304#define S390_CC_LE (S390_CC_LT | S390_CC_EQ)
305#define S390_CC_GE (S390_CC_GT | S390_CC_EQ)
306#define S390_CC_NEVER 0
307#define S390_CC_ALWAYS 15
308
309/* Condition codes that result from a COMPARE and COMPARE LOGICAL. */
0aed257f 310static const uint8_t tcg_cond_to_s390_cond[] = {
48bb3750
RH
311 [TCG_COND_EQ] = S390_CC_EQ,
312 [TCG_COND_NE] = S390_CC_NE,
313 [TCG_COND_LT] = S390_CC_LT,
314 [TCG_COND_LE] = S390_CC_LE,
315 [TCG_COND_GT] = S390_CC_GT,
316 [TCG_COND_GE] = S390_CC_GE,
317 [TCG_COND_LTU] = S390_CC_LT,
318 [TCG_COND_LEU] = S390_CC_LE,
319 [TCG_COND_GTU] = S390_CC_GT,
320 [TCG_COND_GEU] = S390_CC_GE,
321};
322
323/* Condition codes that result from a LOAD AND TEST. Here, we have no
324 unsigned instruction variation, however since the test is vs zero we
325 can re-map the outcomes appropriately. */
0aed257f 326static const uint8_t tcg_cond_to_ltr_cond[] = {
48bb3750
RH
327 [TCG_COND_EQ] = S390_CC_EQ,
328 [TCG_COND_NE] = S390_CC_NE,
329 [TCG_COND_LT] = S390_CC_LT,
330 [TCG_COND_LE] = S390_CC_LE,
331 [TCG_COND_GT] = S390_CC_GT,
332 [TCG_COND_GE] = S390_CC_GE,
333 [TCG_COND_LTU] = S390_CC_NEVER,
334 [TCG_COND_LEU] = S390_CC_EQ,
335 [TCG_COND_GTU] = S390_CC_NE,
336 [TCG_COND_GEU] = S390_CC_ALWAYS,
337};
338
339#ifdef CONFIG_SOFTMMU
f24efee4
RH
340static void * const qemu_ld_helpers[16] = {
341 [MO_UB] = helper_ret_ldub_mmu,
342 [MO_SB] = helper_ret_ldsb_mmu,
343 [MO_LEUW] = helper_le_lduw_mmu,
344 [MO_LESW] = helper_le_ldsw_mmu,
345 [MO_LEUL] = helper_le_ldul_mmu,
346 [MO_LESL] = helper_le_ldsl_mmu,
347 [MO_LEQ] = helper_le_ldq_mmu,
348 [MO_BEUW] = helper_be_lduw_mmu,
349 [MO_BESW] = helper_be_ldsw_mmu,
350 [MO_BEUL] = helper_be_ldul_mmu,
351 [MO_BESL] = helper_be_ldsl_mmu,
352 [MO_BEQ] = helper_be_ldq_mmu,
e141ab52
BS
353};
354
f24efee4
RH
355static void * const qemu_st_helpers[16] = {
356 [MO_UB] = helper_ret_stb_mmu,
357 [MO_LEUW] = helper_le_stw_mmu,
358 [MO_LEUL] = helper_le_stl_mmu,
359 [MO_LEQ] = helper_le_stq_mmu,
360 [MO_BEUW] = helper_be_stw_mmu,
361 [MO_BEUL] = helper_be_stl_mmu,
362 [MO_BEQ] = helper_be_stq_mmu,
e141ab52 363};
e141ab52 364#endif
48bb3750 365
8c081b18 366static tcg_insn_unit *tb_ret_addr;
b2c98d9d 367uint64_t s390_facilities;
2827822e 368
6ac17786 369static bool patch_reloc(tcg_insn_unit *code_ptr, int type,
2ba7fae2 370 intptr_t value, intptr_t addend)
2827822e 371{
e692a349 372 intptr_t pcrel2;
28eef8aa 373 uint32_t old;
e692a349
RH
374
375 value += addend;
376 pcrel2 = (tcg_insn_unit *)value - code_ptr;
48bb3750
RH
377
378 switch (type) {
379 case R_390_PC16DBL:
55dfd8fe
RH
380 if (pcrel2 == (int16_t)pcrel2) {
381 tcg_patch16(code_ptr, pcrel2);
382 return true;
383 }
48bb3750
RH
384 break;
385 case R_390_PC32DBL:
55dfd8fe
RH
386 if (pcrel2 == (int32_t)pcrel2) {
387 tcg_patch32(code_ptr, pcrel2);
388 return true;
389 }
48bb3750 390 break;
28eef8aa 391 case R_390_20:
55dfd8fe
RH
392 if (value == sextract64(value, 0, 20)) {
393 old = *(uint32_t *)code_ptr & 0xf00000ff;
394 old |= ((value & 0xfff) << 16) | ((value & 0xff000) >> 4);
395 tcg_patch32(code_ptr, old);
396 return true;
397 }
28eef8aa 398 break;
48bb3750 399 default:
e692a349 400 g_assert_not_reached();
48bb3750 401 }
55dfd8fe 402 return false;
2827822e
AG
403}
404
2827822e 405/* parse target specific constraints */
069ea736
RH
406static const char *target_parse_constraint(TCGArgConstraint *ct,
407 const char *ct_str, TCGType type)
2827822e 408{
069ea736 409 switch (*ct_str++) {
48bb3750
RH
410 case 'r': /* all registers */
411 ct->ct |= TCG_CT_REG;
f46934df 412 ct->u.regs = 0xffff;
48bb3750 413 break;
48bb3750
RH
414 case 'L': /* qemu_ld/st constraint */
415 ct->ct |= TCG_CT_REG;
f46934df
RH
416 ct->u.regs = 0xffff;
417 tcg_regset_reset_reg(ct->u.regs, TCG_REG_R2);
418 tcg_regset_reset_reg(ct->u.regs, TCG_REG_R3);
419 tcg_regset_reset_reg(ct->u.regs, TCG_REG_R4);
48bb3750
RH
420 break;
421 case 'a': /* force R2 for division */
422 ct->ct |= TCG_CT_REG;
ccb1bb66 423 ct->u.regs = 0;
48bb3750
RH
424 tcg_regset_set_reg(ct->u.regs, TCG_REG_R2);
425 break;
426 case 'b': /* force R3 for division */
427 ct->ct |= TCG_CT_REG;
ccb1bb66 428 ct->u.regs = 0;
48bb3750
RH
429 tcg_regset_set_reg(ct->u.regs, TCG_REG_R3);
430 break;
ad19b358 431 case 'A':
ba18b07d 432 ct->ct |= TCG_CT_CONST_S33;
ad19b358 433 break;
a8f0269e
RH
434 case 'I':
435 ct->ct |= TCG_CT_CONST_S16;
436 break;
437 case 'J':
438 ct->ct |= TCG_CT_CONST_S32;
48bb3750 439 break;
752b1be9
RH
440 case 'Z':
441 ct->ct |= TCG_CT_CONST_ZERO;
442 break;
48bb3750 443 default:
069ea736 444 return NULL;
48bb3750 445 }
069ea736 446 return ct_str;
2827822e
AG
447}
448
449/* Test if a constant matches the constraint. */
f6c6afc1 450static int tcg_target_const_match(tcg_target_long val, TCGType type,
48bb3750 451 const TCGArgConstraint *arg_ct)
2827822e 452{
48bb3750
RH
453 int ct = arg_ct->ct;
454
455 if (ct & TCG_CT_CONST) {
456 return 1;
457 }
458
671c835b 459 if (type == TCG_TYPE_I32) {
48bb3750
RH
460 val = (int32_t)val;
461 }
462
463 /* The following are mutually exclusive. */
a8f0269e
RH
464 if (ct & TCG_CT_CONST_S16) {
465 return val == (int16_t)val;
466 } else if (ct & TCG_CT_CONST_S32) {
467 return val == (int32_t)val;
ba18b07d
RH
468 } else if (ct & TCG_CT_CONST_S33) {
469 return val >= -0xffffffffll && val <= 0xffffffffll;
752b1be9
RH
470 } else if (ct & TCG_CT_CONST_ZERO) {
471 return val == 0;
48bb3750
RH
472 }
473
2827822e
AG
474 return 0;
475}
476
48bb3750
RH
477/* Emit instructions according to the given instruction format. */
478
479static void tcg_out_insn_RR(TCGContext *s, S390Opcode op, TCGReg r1, TCGReg r2)
480{
481 tcg_out16(s, (op << 8) | (r1 << 4) | r2);
482}
483
484static void tcg_out_insn_RRE(TCGContext *s, S390Opcode op,
485 TCGReg r1, TCGReg r2)
486{
487 tcg_out32(s, (op << 16) | (r1 << 4) | r2);
488}
489
96a9f093
RH
490static void tcg_out_insn_RRF(TCGContext *s, S390Opcode op,
491 TCGReg r1, TCGReg r2, int m3)
492{
493 tcg_out32(s, (op << 16) | (m3 << 12) | (r1 << 4) | r2);
494}
495
48bb3750
RH
496static void tcg_out_insn_RI(TCGContext *s, S390Opcode op, TCGReg r1, int i2)
497{
498 tcg_out32(s, (op << 16) | (r1 << 20) | (i2 & 0xffff));
499}
500
7af525af
RH
501static void tcg_out_insn_RIE(TCGContext *s, S390Opcode op, TCGReg r1,
502 int i2, int m3)
503{
504 tcg_out16(s, (op & 0xff00) | (r1 << 4) | m3);
505 tcg_out32(s, (i2 << 16) | (op & 0xff));
506}
507
48bb3750
RH
508static void tcg_out_insn_RIL(TCGContext *s, S390Opcode op, TCGReg r1, int i2)
509{
510 tcg_out16(s, op | (r1 << 4));
511 tcg_out32(s, i2);
512}
513
514static void tcg_out_insn_RS(TCGContext *s, S390Opcode op, TCGReg r1,
515 TCGReg b2, TCGReg r3, int disp)
516{
517 tcg_out32(s, (op << 24) | (r1 << 20) | (r3 << 16) | (b2 << 12)
518 | (disp & 0xfff));
519}
520
521static void tcg_out_insn_RSY(TCGContext *s, S390Opcode op, TCGReg r1,
522 TCGReg b2, TCGReg r3, int disp)
523{
524 tcg_out16(s, (op & 0xff00) | (r1 << 4) | r3);
525 tcg_out32(s, (op & 0xff) | (b2 << 28)
526 | ((disp & 0xfff) << 16) | ((disp & 0xff000) >> 4));
527}
528
529#define tcg_out_insn_RX tcg_out_insn_RS
530#define tcg_out_insn_RXY tcg_out_insn_RSY
531
532/* Emit an opcode with "type-checking" of the format. */
533#define tcg_out_insn(S, FMT, OP, ...) \
534 glue(tcg_out_insn_,FMT)(S, glue(glue(FMT,_),OP), ## __VA_ARGS__)
535
536
537/* emit 64-bit shifts */
538static void tcg_out_sh64(TCGContext* s, S390Opcode op, TCGReg dest,
539 TCGReg src, TCGReg sh_reg, int sh_imm)
540{
541 tcg_out_insn_RSY(s, op, dest, sh_reg, src, sh_imm);
542}
543
544/* emit 32-bit shifts */
545static void tcg_out_sh32(TCGContext* s, S390Opcode op, TCGReg dest,
546 TCGReg sh_reg, int sh_imm)
547{
548 tcg_out_insn_RS(s, op, dest, sh_reg, 0, sh_imm);
549}
550
78113e83 551static bool tcg_out_mov(TCGContext *s, TCGType type, TCGReg dst, TCGReg src)
48bb3750
RH
552{
553 if (src != dst) {
554 if (type == TCG_TYPE_I32) {
555 tcg_out_insn(s, RR, LR, dst, src);
556 } else {
557 tcg_out_insn(s, RRE, LGR, dst, src);
558 }
559 }
78113e83 560 return true;
48bb3750
RH
561}
562
28eef8aa
RH
563static const S390Opcode lli_insns[4] = {
564 RI_LLILL, RI_LLILH, RI_LLIHL, RI_LLIHH
565};
48bb3750 566
28eef8aa
RH
567static bool maybe_out_small_movi(TCGContext *s, TCGType type,
568 TCGReg ret, tcg_target_long sval)
569{
48bb3750
RH
570 tcg_target_ulong uval = sval;
571 int i;
572
573 if (type == TCG_TYPE_I32) {
574 uval = (uint32_t)sval;
575 sval = (int32_t)sval;
576 }
577
578 /* Try all 32-bit insns that can load it in one go. */
579 if (sval >= -0x8000 && sval < 0x8000) {
580 tcg_out_insn(s, RI, LGHI, ret, sval);
28eef8aa 581 return true;
48bb3750
RH
582 }
583
584 for (i = 0; i < 4; i++) {
585 tcg_target_long mask = 0xffffull << i*16;
586 if ((uval & mask) == uval) {
587 tcg_out_insn_RI(s, lli_insns[i], ret, uval >> i*16);
28eef8aa 588 return true;
48bb3750
RH
589 }
590 }
591
28eef8aa
RH
592 return false;
593}
594
595/* load a register with an immediate value */
596static void tcg_out_movi_int(TCGContext *s, TCGType type, TCGReg ret,
597 tcg_target_long sval, bool in_prologue)
598{
599 tcg_target_ulong uval;
600
601 /* Try all 32-bit insns that can load it in one go. */
602 if (maybe_out_small_movi(s, type, ret, sval)) {
603 return;
604 }
605
606 uval = sval;
607 if (type == TCG_TYPE_I32) {
608 uval = (uint32_t)sval;
609 sval = (int32_t)sval;
610 }
611
48bb3750 612 /* Try all 48-bit insns that can load it in one go. */
b2c98d9d 613 if (s390_facilities & FACILITY_EXT_IMM) {
48bb3750
RH
614 if (sval == (int32_t)sval) {
615 tcg_out_insn(s, RIL, LGFI, ret, sval);
616 return;
617 }
618 if (uval <= 0xffffffff) {
619 tcg_out_insn(s, RIL, LLILF, ret, uval);
620 return;
621 }
622 if ((uval & 0xffffffff) == 0) {
28eef8aa 623 tcg_out_insn(s, RIL, LLIHF, ret, uval >> 32);
48bb3750
RH
624 return;
625 }
626 }
627
829e1376
RH
628 /* Try for PC-relative address load. For odd addresses,
629 attempt to use an offset from the start of the TB. */
48bb3750 630 if ((sval & 1) == 0) {
8c081b18 631 ptrdiff_t off = tcg_pcrel_diff(s, (void *)sval) >> 1;
48bb3750
RH
632 if (off == (int32_t)off) {
633 tcg_out_insn(s, RIL, LARL, ret, off);
634 return;
635 }
829e1376
RH
636 } else if (USE_REG_TB && !in_prologue) {
637 ptrdiff_t off = sval - (uintptr_t)s->code_gen_ptr;
638 if (off == sextract64(off, 0, 20)) {
639 /* This is certain to be an address within TB, and therefore
640 OFF will be negative; don't try RX_LA. */
641 tcg_out_insn(s, RXY, LAY, ret, TCG_REG_TB, TCG_REG_NONE, off);
642 return;
643 }
48bb3750
RH
644 }
645
28eef8aa
RH
646 /* A 32-bit unsigned value can be loaded in 2 insns. And given
647 that LLILL, LLIHL, LLILF above did not succeed, we know that
648 both insns are required. */
649 if (uval <= 0xffffffff) {
650 tcg_out_insn(s, RI, LLILL, ret, uval);
651 tcg_out_insn(s, RI, IILH, ret, uval >> 16);
652 return;
653 }
48bb3750 654
ba2c7479
RH
655 /* Otherwise, stuff it in the constant pool. */
656 if (s390_facilities & FACILITY_GEN_INST_EXT) {
657 tcg_out_insn(s, RIL, LGRL, ret, 0);
658 new_pool_label(s, sval, R_390_PC32DBL, s->code_ptr - 2, 2);
659 } else if (USE_REG_TB && !in_prologue) {
660 tcg_out_insn(s, RXY, LG, ret, TCG_REG_TB, TCG_REG_NONE, 0);
661 new_pool_label(s, sval, R_390_20, s->code_ptr - 2,
662 -(intptr_t)s->code_gen_ptr);
48bb3750 663 } else {
ba2c7479
RH
664 TCGReg base = ret ? ret : TCG_TMP0;
665 tcg_out_insn(s, RIL, LARL, base, 0);
666 new_pool_label(s, sval, R_390_PC32DBL, s->code_ptr - 2, 2);
667 tcg_out_insn(s, RXY, LG, ret, base, TCG_REG_NONE, 0);
48bb3750
RH
668 }
669}
670
829e1376
RH
671static void tcg_out_movi(TCGContext *s, TCGType type,
672 TCGReg ret, tcg_target_long sval)
673{
674 tcg_out_movi_int(s, type, ret, sval, false);
675}
48bb3750
RH
676
677/* Emit a load/store type instruction. Inputs are:
678 DATA: The register to be loaded or stored.
679 BASE+OFS: The effective address.
680 OPC_RX: If the operation has an RX format opcode (e.g. STC), otherwise 0.
681 OPC_RXY: The RXY format opcode for the operation (e.g. STCY). */
682
683static void tcg_out_mem(TCGContext *s, S390Opcode opc_rx, S390Opcode opc_rxy,
684 TCGReg data, TCGReg base, TCGReg index,
685 tcg_target_long ofs)
686{
687 if (ofs < -0x80000 || ofs >= 0x80000) {
78c9f7c5
RH
688 /* Combine the low 20 bits of the offset with the actual load insn;
689 the high 44 bits must come from an immediate load. */
690 tcg_target_long low = ((ofs & 0xfffff) ^ 0x80000) - 0x80000;
691 tcg_out_movi(s, TCG_TYPE_PTR, TCG_TMP0, ofs - low);
692 ofs = low;
48bb3750
RH
693
694 /* If we were already given an index register, add it in. */
695 if (index != TCG_REG_NONE) {
696 tcg_out_insn(s, RRE, AGR, TCG_TMP0, index);
697 }
698 index = TCG_TMP0;
699 }
700
701 if (opc_rx && ofs >= 0 && ofs < 0x1000) {
702 tcg_out_insn_RX(s, opc_rx, data, base, index, ofs);
703 } else {
704 tcg_out_insn_RXY(s, opc_rxy, data, base, index, ofs);
705 }
2827822e
AG
706}
707
48bb3750 708
2827822e 709/* load data without address translation or endianness conversion */
48bb3750 710static inline void tcg_out_ld(TCGContext *s, TCGType type, TCGReg data,
a05b5b9b 711 TCGReg base, intptr_t ofs)
2827822e 712{
48bb3750
RH
713 if (type == TCG_TYPE_I32) {
714 tcg_out_mem(s, RX_L, RXY_LY, data, base, TCG_REG_NONE, ofs);
715 } else {
716 tcg_out_mem(s, 0, RXY_LG, data, base, TCG_REG_NONE, ofs);
717 }
2827822e
AG
718}
719
48bb3750 720static inline void tcg_out_st(TCGContext *s, TCGType type, TCGReg data,
a05b5b9b 721 TCGReg base, intptr_t ofs)
2827822e 722{
48bb3750
RH
723 if (type == TCG_TYPE_I32) {
724 tcg_out_mem(s, RX_ST, RXY_STY, data, base, TCG_REG_NONE, ofs);
725 } else {
726 tcg_out_mem(s, 0, RXY_STG, data, base, TCG_REG_NONE, ofs);
727 }
728}
729
59d7c14e
RH
730static inline bool tcg_out_sti(TCGContext *s, TCGType type, TCGArg val,
731 TCGReg base, intptr_t ofs)
732{
733 return false;
734}
735
48bb3750
RH
736/* load data from an absolute host address */
737static void tcg_out_ld_abs(TCGContext *s, TCGType type, TCGReg dest, void *abs)
738{
8c081b18 739 intptr_t addr = (intptr_t)abs;
48bb3750 740
b2c98d9d 741 if ((s390_facilities & FACILITY_GEN_INST_EXT) && !(addr & 1)) {
8c081b18 742 ptrdiff_t disp = tcg_pcrel_diff(s, abs) >> 1;
48bb3750
RH
743 if (disp == (int32_t)disp) {
744 if (type == TCG_TYPE_I32) {
745 tcg_out_insn(s, RIL, LRL, dest, disp);
746 } else {
747 tcg_out_insn(s, RIL, LGRL, dest, disp);
748 }
749 return;
750 }
751 }
829e1376
RH
752 if (USE_REG_TB) {
753 ptrdiff_t disp = abs - (void *)s->code_gen_ptr;
754 if (disp == sextract64(disp, 0, 20)) {
755 tcg_out_ld(s, type, dest, TCG_REG_TB, disp);
756 return;
757 }
758 }
48bb3750
RH
759
760 tcg_out_movi(s, TCG_TYPE_PTR, dest, addr & ~0xffff);
761 tcg_out_ld(s, type, dest, dest, addr & 0xffff);
762}
763
f0bffc27
RH
764static inline void tcg_out_risbg(TCGContext *s, TCGReg dest, TCGReg src,
765 int msb, int lsb, int ofs, int z)
766{
767 /* Format RIE-f */
768 tcg_out16(s, (RIE_RISBG & 0xff00) | (dest << 4) | src);
769 tcg_out16(s, (msb << 8) | (z << 7) | lsb);
770 tcg_out16(s, (ofs << 8) | (RIE_RISBG & 0xff));
771}
772
48bb3750
RH
773static void tgen_ext8s(TCGContext *s, TCGType type, TCGReg dest, TCGReg src)
774{
b2c98d9d 775 if (s390_facilities & FACILITY_EXT_IMM) {
48bb3750
RH
776 tcg_out_insn(s, RRE, LGBR, dest, src);
777 return;
778 }
779
780 if (type == TCG_TYPE_I32) {
781 if (dest == src) {
782 tcg_out_sh32(s, RS_SLL, dest, TCG_REG_NONE, 24);
783 } else {
784 tcg_out_sh64(s, RSY_SLLG, dest, src, TCG_REG_NONE, 24);
785 }
786 tcg_out_sh32(s, RS_SRA, dest, TCG_REG_NONE, 24);
787 } else {
788 tcg_out_sh64(s, RSY_SLLG, dest, src, TCG_REG_NONE, 56);
789 tcg_out_sh64(s, RSY_SRAG, dest, dest, TCG_REG_NONE, 56);
790 }
791}
792
793static void tgen_ext8u(TCGContext *s, TCGType type, TCGReg dest, TCGReg src)
794{
b2c98d9d 795 if (s390_facilities & FACILITY_EXT_IMM) {
48bb3750
RH
796 tcg_out_insn(s, RRE, LLGCR, dest, src);
797 return;
798 }
799
800 if (dest == src) {
801 tcg_out_movi(s, type, TCG_TMP0, 0xff);
802 src = TCG_TMP0;
803 } else {
804 tcg_out_movi(s, type, dest, 0xff);
805 }
806 if (type == TCG_TYPE_I32) {
807 tcg_out_insn(s, RR, NR, dest, src);
808 } else {
809 tcg_out_insn(s, RRE, NGR, dest, src);
810 }
811}
812
813static void tgen_ext16s(TCGContext *s, TCGType type, TCGReg dest, TCGReg src)
814{
b2c98d9d 815 if (s390_facilities & FACILITY_EXT_IMM) {
48bb3750
RH
816 tcg_out_insn(s, RRE, LGHR, dest, src);
817 return;
818 }
819
820 if (type == TCG_TYPE_I32) {
821 if (dest == src) {
822 tcg_out_sh32(s, RS_SLL, dest, TCG_REG_NONE, 16);
823 } else {
824 tcg_out_sh64(s, RSY_SLLG, dest, src, TCG_REG_NONE, 16);
825 }
826 tcg_out_sh32(s, RS_SRA, dest, TCG_REG_NONE, 16);
827 } else {
828 tcg_out_sh64(s, RSY_SLLG, dest, src, TCG_REG_NONE, 48);
829 tcg_out_sh64(s, RSY_SRAG, dest, dest, TCG_REG_NONE, 48);
830 }
831}
832
833static void tgen_ext16u(TCGContext *s, TCGType type, TCGReg dest, TCGReg src)
834{
b2c98d9d 835 if (s390_facilities & FACILITY_EXT_IMM) {
48bb3750
RH
836 tcg_out_insn(s, RRE, LLGHR, dest, src);
837 return;
838 }
839
840 if (dest == src) {
841 tcg_out_movi(s, type, TCG_TMP0, 0xffff);
842 src = TCG_TMP0;
843 } else {
844 tcg_out_movi(s, type, dest, 0xffff);
845 }
846 if (type == TCG_TYPE_I32) {
847 tcg_out_insn(s, RR, NR, dest, src);
848 } else {
849 tcg_out_insn(s, RRE, NGR, dest, src);
850 }
851}
852
853static inline void tgen_ext32s(TCGContext *s, TCGReg dest, TCGReg src)
854{
855 tcg_out_insn(s, RRE, LGFR, dest, src);
856}
857
858static inline void tgen_ext32u(TCGContext *s, TCGReg dest, TCGReg src)
859{
860 tcg_out_insn(s, RRE, LLGFR, dest, src);
861}
862
f0bffc27
RH
863/* Accept bit patterns like these:
864 0....01....1
865 1....10....0
866 1..10..01..1
867 0..01..10..0
868 Copied from gcc sources. */
869static inline bool risbg_mask(uint64_t c)
870{
871 uint64_t lsb;
872 /* We don't change the number of transitions by inverting,
873 so make sure we start with the LSB zero. */
874 if (c & 1) {
875 c = ~c;
876 }
877 /* Reject all zeros or all ones. */
878 if (c == 0) {
879 return false;
880 }
881 /* Find the first transition. */
882 lsb = c & -c;
883 /* Invert to look for a second transition. */
884 c = ~c;
885 /* Erase the first transition. */
886 c &= -lsb;
887 /* Find the second transition, if any. */
888 lsb = c & -c;
889 /* Match if all the bits are 1's, or if c is zero. */
890 return c == -lsb;
891}
892
547ec121
RH
893static void tgen_andi_risbg(TCGContext *s, TCGReg out, TCGReg in, uint64_t val)
894{
895 int msb, lsb;
896 if ((val & 0x8000000000000001ull) == 0x8000000000000001ull) {
897 /* Achieve wraparound by swapping msb and lsb. */
898 msb = 64 - ctz64(~val);
899 lsb = clz64(~val) - 1;
900 } else {
901 msb = clz64(val);
902 lsb = 63 - ctz64(val);
903 }
904 tcg_out_risbg(s, out, in, msb, lsb, 0, 1);
905}
906
07ff7983 907static void tgen_andi(TCGContext *s, TCGType type, TCGReg dest, uint64_t val)
48bb3750
RH
908{
909 static const S390Opcode ni_insns[4] = {
910 RI_NILL, RI_NILH, RI_NIHL, RI_NIHH
911 };
912 static const S390Opcode nif_insns[2] = {
913 RIL_NILF, RIL_NIHF
914 };
07ff7983 915 uint64_t valid = (type == TCG_TYPE_I32 ? 0xffffffffull : -1ull);
48bb3750
RH
916 int i;
917
48bb3750 918 /* Look for the zero-extensions. */
07ff7983 919 if ((val & valid) == 0xffffffff) {
48bb3750
RH
920 tgen_ext32u(s, dest, dest);
921 return;
922 }
b2c98d9d 923 if (s390_facilities & FACILITY_EXT_IMM) {
07ff7983 924 if ((val & valid) == 0xff) {
48bb3750
RH
925 tgen_ext8u(s, TCG_TYPE_I64, dest, dest);
926 return;
927 }
07ff7983 928 if ((val & valid) == 0xffff) {
48bb3750
RH
929 tgen_ext16u(s, TCG_TYPE_I64, dest, dest);
930 return;
931 }
07ff7983 932 }
48bb3750 933
07ff7983
RH
934 /* Try all 32-bit insns that can perform it in one go. */
935 for (i = 0; i < 4; i++) {
936 tcg_target_ulong mask = ~(0xffffull << i*16);
937 if (((val | ~valid) & mask) == mask) {
938 tcg_out_insn_RI(s, ni_insns[i], dest, val >> i*16);
939 return;
48bb3750 940 }
07ff7983 941 }
48bb3750 942
07ff7983 943 /* Try all 48-bit insns that can perform it in one go. */
b2c98d9d 944 if (s390_facilities & FACILITY_EXT_IMM) {
07ff7983
RH
945 for (i = 0; i < 2; i++) {
946 tcg_target_ulong mask = ~(0xffffffffull << i*32);
947 if (((val | ~valid) & mask) == mask) {
948 tcg_out_insn_RIL(s, nif_insns[i], dest, val >> i*32);
949 return;
48bb3750
RH
950 }
951 }
07ff7983 952 }
b2c98d9d 953 if ((s390_facilities & FACILITY_GEN_INST_EXT) && risbg_mask(val)) {
547ec121 954 tgen_andi_risbg(s, dest, dest, val);
f0bffc27
RH
955 return;
956 }
48bb3750 957
bdcd5d19
RH
958 /* Use the constant pool if USE_REG_TB, but not for small constants. */
959 if (USE_REG_TB) {
960 if (!maybe_out_small_movi(s, type, TCG_TMP0, val)) {
961 tcg_out_insn(s, RXY, NG, dest, TCG_REG_TB, TCG_REG_NONE, 0);
962 new_pool_label(s, val & valid, R_390_20, s->code_ptr - 2,
963 -(intptr_t)s->code_gen_ptr);
964 return;
965 }
966 } else {
967 tcg_out_movi(s, type, TCG_TMP0, val);
968 }
07ff7983
RH
969 if (type == TCG_TYPE_I32) {
970 tcg_out_insn(s, RR, NR, dest, TCG_TMP0);
48bb3750 971 } else {
07ff7983 972 tcg_out_insn(s, RRE, NGR, dest, TCG_TMP0);
48bb3750
RH
973 }
974}
975
4046d9ca 976static void tgen_ori(TCGContext *s, TCGType type, TCGReg dest, uint64_t val)
48bb3750
RH
977{
978 static const S390Opcode oi_insns[4] = {
979 RI_OILL, RI_OILH, RI_OIHL, RI_OIHH
980 };
4046d9ca 981 static const S390Opcode oif_insns[2] = {
48bb3750
RH
982 RIL_OILF, RIL_OIHF
983 };
984
985 int i;
986
987 /* Look for no-op. */
4046d9ca 988 if (unlikely(val == 0)) {
48bb3750
RH
989 return;
990 }
991
4046d9ca
RH
992 /* Try all 32-bit insns that can perform it in one go. */
993 for (i = 0; i < 4; i++) {
994 tcg_target_ulong mask = (0xffffull << i*16);
995 if ((val & mask) != 0 && (val & ~mask) == 0) {
996 tcg_out_insn_RI(s, oi_insns[i], dest, val >> i*16);
997 return;
48bb3750 998 }
4046d9ca 999 }
48bb3750 1000
4046d9ca
RH
1001 /* Try all 48-bit insns that can perform it in one go. */
1002 if (s390_facilities & FACILITY_EXT_IMM) {
48bb3750
RH
1003 for (i = 0; i < 2; i++) {
1004 tcg_target_ulong mask = (0xffffffffull << i*32);
1005 if ((val & mask) != 0 && (val & ~mask) == 0) {
4046d9ca 1006 tcg_out_insn_RIL(s, oif_insns[i], dest, val >> i*32);
48bb3750
RH
1007 return;
1008 }
1009 }
4046d9ca 1010 }
48bb3750 1011
4046d9ca
RH
1012 /* Use the constant pool if USE_REG_TB, but not for small constants. */
1013 if (maybe_out_small_movi(s, type, TCG_TMP0, val)) {
1014 if (type == TCG_TYPE_I32) {
1015 tcg_out_insn(s, RR, OR, dest, TCG_TMP0);
1016 } else {
1017 tcg_out_insn(s, RRE, OGR, dest, TCG_TMP0);
1018 }
1019 } else if (USE_REG_TB) {
1020 tcg_out_insn(s, RXY, OG, dest, TCG_REG_TB, TCG_REG_NONE, 0);
1021 new_pool_label(s, val, R_390_20, s->code_ptr - 2,
1022 -(intptr_t)s->code_gen_ptr);
1023 } else {
48bb3750
RH
1024 /* Perform the OR via sequential modifications to the high and
1025 low parts. Do this via recursion to handle 16-bit vs 32-bit
1026 masks in each half. */
4046d9ca
RH
1027 tcg_debug_assert(s390_facilities & FACILITY_EXT_IMM);
1028 tgen_ori(s, type, dest, val & 0x00000000ffffffffull);
1029 tgen_ori(s, type, dest, val & 0xffffffff00000000ull);
48bb3750
RH
1030 }
1031}
1032
5bf67a92 1033static void tgen_xori(TCGContext *s, TCGType type, TCGReg dest, uint64_t val)
48bb3750 1034{
5bf67a92
RH
1035 /* Try all 48-bit insns that can perform it in one go. */
1036 if (s390_facilities & FACILITY_EXT_IMM) {
1037 if ((val & 0xffffffff00000000ull) == 0) {
1038 tcg_out_insn(s, RIL, XILF, dest, val);
1039 return;
1040 }
1041 if ((val & 0x00000000ffffffffull) == 0) {
1042 tcg_out_insn(s, RIL, XIHF, dest, val >> 32);
1043 return;
1044 }
48bb3750 1045 }
5bf67a92
RH
1046
1047 /* Use the constant pool if USE_REG_TB, but not for small constants. */
1048 if (maybe_out_small_movi(s, type, TCG_TMP0, val)) {
1049 if (type == TCG_TYPE_I32) {
1050 tcg_out_insn(s, RR, XR, dest, TCG_TMP0);
1051 } else {
1052 tcg_out_insn(s, RRE, XGR, dest, TCG_TMP0);
1053 }
1054 } else if (USE_REG_TB) {
1055 tcg_out_insn(s, RXY, XG, dest, TCG_REG_TB, TCG_REG_NONE, 0);
1056 new_pool_label(s, val, R_390_20, s->code_ptr - 2,
1057 -(intptr_t)s->code_gen_ptr);
1058 } else {
1059 /* Perform the xor by parts. */
1060 tcg_debug_assert(s390_facilities & FACILITY_EXT_IMM);
1061 if (val & 0xffffffff) {
1062 tcg_out_insn(s, RIL, XILF, dest, val);
1063 }
1064 if (val > 0xffffffff) {
1065 tcg_out_insn(s, RIL, XIHF, dest, val >> 32);
1066 }
48bb3750
RH
1067 }
1068}
1069
1070static int tgen_cmp(TCGContext *s, TCGType type, TCGCond c, TCGReg r1,
65839b56 1071 TCGArg c2, bool c2const, bool need_carry)
48bb3750 1072{
bcc66562 1073 bool is_unsigned = is_unsigned_cond(c);
a534bb15
RH
1074 S390Opcode op;
1075
48bb3750
RH
1076 if (c2const) {
1077 if (c2 == 0) {
65839b56
RH
1078 if (!(is_unsigned && need_carry)) {
1079 if (type == TCG_TYPE_I32) {
1080 tcg_out_insn(s, RR, LTR, r1, r1);
1081 } else {
1082 tcg_out_insn(s, RRE, LTGR, r1, r1);
1083 }
1084 return tcg_cond_to_ltr_cond[c];
1085 }
65839b56 1086 }
a534bb15
RH
1087
1088 if (!is_unsigned && c2 == (int16_t)c2) {
1089 op = (type == TCG_TYPE_I32 ? RI_CHI : RI_CGHI);
1090 tcg_out_insn_RI(s, op, r1, c2);
1091 goto exit;
1092 }
1093
1094 if (s390_facilities & FACILITY_EXT_IMM) {
65839b56 1095 if (type == TCG_TYPE_I32) {
a534bb15
RH
1096 op = (is_unsigned ? RIL_CLFI : RIL_CFI);
1097 tcg_out_insn_RIL(s, op, r1, c2);
1098 goto exit;
1099 } else if (c2 == (is_unsigned ? (uint32_t)c2 : (int32_t)c2)) {
1100 op = (is_unsigned ? RIL_CLGFI : RIL_CGFI);
1101 tcg_out_insn_RIL(s, op, r1, c2);
1102 goto exit;
48bb3750
RH
1103 }
1104 }
a534bb15
RH
1105
1106 /* Use the constant pool, but not for small constants. */
1107 if (maybe_out_small_movi(s, type, TCG_TMP0, c2)) {
1108 c2 = TCG_TMP0;
1109 /* fall through to reg-reg */
1110 } else if (USE_REG_TB) {
48bb3750 1111 if (type == TCG_TYPE_I32) {
a534bb15
RH
1112 op = (is_unsigned ? RXY_CLY : RXY_CY);
1113 tcg_out_insn_RXY(s, op, r1, TCG_REG_TB, TCG_REG_NONE, 0);
1114 new_pool_label(s, (uint32_t)c2, R_390_20, s->code_ptr - 2,
1115 4 - (intptr_t)s->code_gen_ptr);
48bb3750 1116 } else {
a534bb15
RH
1117 op = (is_unsigned ? RXY_CLG : RXY_CG);
1118 tcg_out_insn_RXY(s, op, r1, TCG_REG_TB, TCG_REG_NONE, 0);
1119 new_pool_label(s, c2, R_390_20, s->code_ptr - 2,
1120 -(intptr_t)s->code_gen_ptr);
48bb3750 1121 }
a534bb15 1122 goto exit;
48bb3750
RH
1123 } else {
1124 if (type == TCG_TYPE_I32) {
a534bb15
RH
1125 op = (is_unsigned ? RIL_CLRL : RIL_CRL);
1126 tcg_out_insn_RIL(s, op, r1, 0);
1127 new_pool_label(s, (uint32_t)c2, R_390_PC32DBL,
1128 s->code_ptr - 2, 2 + 4);
48bb3750 1129 } else {
a534bb15
RH
1130 op = (is_unsigned ? RIL_CLGRL : RIL_CGRL);
1131 tcg_out_insn_RIL(s, op, r1, 0);
1132 new_pool_label(s, c2, R_390_PC32DBL, s->code_ptr - 2, 2);
48bb3750 1133 }
a534bb15 1134 goto exit;
48bb3750
RH
1135 }
1136 }
a534bb15
RH
1137
1138 if (type == TCG_TYPE_I32) {
1139 op = (is_unsigned ? RR_CLR : RR_CR);
1140 tcg_out_insn_RR(s, op, r1, c2);
1141 } else {
1142 op = (is_unsigned ? RRE_CLGR : RRE_CGR);
1143 tcg_out_insn_RRE(s, op, r1, c2);
1144 }
1145
1146 exit:
48bb3750
RH
1147 return tcg_cond_to_s390_cond[c];
1148}
1149
7b7066b1 1150static void tgen_setcond(TCGContext *s, TCGType type, TCGCond cond,
96a9f093 1151 TCGReg dest, TCGReg c1, TCGArg c2, int c2const)
48bb3750 1152{
7b7066b1 1153 int cc;
7af525af 1154 bool have_loc;
7b7066b1 1155
7af525af
RH
1156 /* With LOC2, we can always emit the minimum 3 insns. */
1157 if (s390_facilities & FACILITY_LOAD_ON_COND2) {
1158 /* Emit: d = 0, d = (cc ? 1 : d). */
1159 cc = tgen_cmp(s, type, cond, c1, c2, c2const, false);
1160 tcg_out_movi(s, TCG_TYPE_I64, dest, 0);
1161 tcg_out_insn(s, RIE, LOCGHI, dest, 1, cc);
1162 return;
1163 }
1164
1165 have_loc = (s390_facilities & FACILITY_LOAD_ON_COND) != 0;
1166
4609190b
RH
1167 /* For HAVE_LOC, only the paths through GTU/GT/LEU/LE are smaller. */
1168 restart:
7b7066b1 1169 switch (cond) {
4609190b
RH
1170 case TCG_COND_NE:
1171 /* X != 0 is X > 0. */
1172 if (c2const && c2 == 0) {
1173 cond = TCG_COND_GTU;
1174 } else {
1175 break;
1176 }
1177 /* fallthru */
1178
7b7066b1
RH
1179 case TCG_COND_GTU:
1180 case TCG_COND_GT:
7b7066b1
RH
1181 /* The result of a compare has CC=2 for GT and CC=3 unused.
1182 ADD LOGICAL WITH CARRY considers (CC & 2) the carry bit. */
65839b56 1183 tgen_cmp(s, type, cond, c1, c2, c2const, true);
7b7066b1
RH
1184 tcg_out_movi(s, type, dest, 0);
1185 tcg_out_insn(s, RRE, ALCGR, dest, dest);
1186 return;
1187
4609190b
RH
1188 case TCG_COND_EQ:
1189 /* X == 0 is X <= 0. */
1190 if (c2const && c2 == 0) {
1191 cond = TCG_COND_LEU;
7b7066b1 1192 } else {
4609190b 1193 break;
7b7066b1 1194 }
4609190b 1195 /* fallthru */
7b7066b1
RH
1196
1197 case TCG_COND_LEU:
4609190b
RH
1198 case TCG_COND_LE:
1199 /* As above, but we're looking for borrow, or !carry.
1200 The second insn computes d - d - borrow, or -1 for true
1201 and 0 for false. So we must mask to 1 bit afterward. */
1202 tgen_cmp(s, type, cond, c1, c2, c2const, true);
1203 tcg_out_insn(s, RRE, SLBGR, dest, dest);
1204 tgen_andi(s, type, dest, 1);
1205 return;
1206
1207 case TCG_COND_GEU:
7b7066b1
RH
1208 case TCG_COND_LTU:
1209 case TCG_COND_LT:
4609190b
RH
1210 case TCG_COND_GE:
1211 /* Swap operands so that we can use LEU/GTU/GT/LE. */
7b7066b1 1212 if (c2const) {
7af525af 1213 if (have_loc) {
4609190b 1214 break;
7af525af 1215 }
7b7066b1
RH
1216 tcg_out_movi(s, type, TCG_TMP0, c2);
1217 c2 = c1;
1218 c2const = 0;
1219 c1 = TCG_TMP0;
1220 } else {
1221 TCGReg t = c1;
1222 c1 = c2;
1223 c2 = t;
1224 }
7b7066b1 1225 cond = tcg_swap_cond(cond);
4609190b 1226 goto restart;
48bb3750 1227
7b7066b1 1228 default:
4609190b 1229 g_assert_not_reached();
7b7066b1
RH
1230 }
1231
65839b56 1232 cc = tgen_cmp(s, type, cond, c1, c2, c2const, false);
4609190b
RH
1233 if (have_loc) {
1234 /* Emit: d = 0, t = 1, d = (cc ? t : d). */
1235 tcg_out_movi(s, TCG_TYPE_I64, dest, 0);
1236 tcg_out_movi(s, TCG_TYPE_I64, TCG_TMP0, 1);
1237 tcg_out_insn(s, RRF, LOCGR, dest, TCG_TMP0, cc);
1238 } else {
1239 /* Emit: d = 1; if (cc) goto over; d = 0; over: */
1240 tcg_out_movi(s, type, dest, 1);
1241 tcg_out_insn(s, RI, BRC, cc, (4 + 4) >> 1);
1242 tcg_out_movi(s, type, dest, 0);
1243 }
48bb3750
RH
1244}
1245
96a9f093 1246static void tgen_movcond(TCGContext *s, TCGType type, TCGCond c, TCGReg dest,
7af525af
RH
1247 TCGReg c1, TCGArg c2, int c2const,
1248 TCGArg v3, int v3const)
96a9f093
RH
1249{
1250 int cc;
b2c98d9d 1251 if (s390_facilities & FACILITY_LOAD_ON_COND) {
65839b56 1252 cc = tgen_cmp(s, type, c, c1, c2, c2const, false);
7af525af
RH
1253 if (v3const) {
1254 tcg_out_insn(s, RIE, LOCGHI, dest, v3, cc);
1255 } else {
1256 tcg_out_insn(s, RRF, LOCGR, dest, v3, cc);
1257 }
96a9f093
RH
1258 } else {
1259 c = tcg_invert_cond(c);
65839b56 1260 cc = tgen_cmp(s, type, c, c1, c2, c2const, false);
96a9f093
RH
1261
1262 /* Emit: if (cc) goto over; dest = r3; over: */
1263 tcg_out_insn(s, RI, BRC, cc, (4 + 4) >> 1);
7af525af 1264 tcg_out_insn(s, RRE, LGR, dest, v3);
96a9f093
RH
1265 }
1266}
1267
ce411066
RH
1268static void tgen_clz(TCGContext *s, TCGReg dest, TCGReg a1,
1269 TCGArg a2, int a2const)
1270{
1271 /* Since this sets both R and R+1, we have no choice but to store the
1272 result into R0, allowing R1 == TCG_TMP0 to be clobbered as well. */
1273 QEMU_BUILD_BUG_ON(TCG_TMP0 != TCG_REG_R1);
1274 tcg_out_insn(s, RRE, FLOGR, TCG_REG_R0, a1);
1275
1276 if (a2const && a2 == 64) {
1277 tcg_out_mov(s, TCG_TYPE_I64, dest, TCG_REG_R0);
1278 } else {
1279 if (a2const) {
1280 tcg_out_movi(s, TCG_TYPE_I64, dest, a2);
1281 } else {
1282 tcg_out_mov(s, TCG_TYPE_I64, dest, a2);
1283 }
1284 if (s390_facilities & FACILITY_LOAD_ON_COND) {
1285 /* Emit: if (one bit found) dest = r0. */
1286 tcg_out_insn(s, RRF, LOCGR, dest, TCG_REG_R0, 2);
1287 } else {
1288 /* Emit: if (no one bit found) goto over; dest = r0; over: */
1289 tcg_out_insn(s, RI, BRC, 8, (4 + 4) >> 1);
1290 tcg_out_insn(s, RRE, LGR, dest, TCG_REG_R0);
1291 }
1292 }
1293}
1294
d5690ea4 1295static void tgen_deposit(TCGContext *s, TCGReg dest, TCGReg src,
752b1be9 1296 int ofs, int len, int z)
d5690ea4
RH
1297{
1298 int lsb = (63 - ofs);
1299 int msb = lsb - (len - 1);
752b1be9 1300 tcg_out_risbg(s, dest, src, msb, lsb, ofs, z);
d5690ea4
RH
1301}
1302
b0bf5fe8
RH
1303static void tgen_extract(TCGContext *s, TCGReg dest, TCGReg src,
1304 int ofs, int len)
1305{
1306 tcg_out_risbg(s, dest, src, 64 - len, 63, 64 - ofs, 1);
1307}
1308
8c081b18 1309static void tgen_gotoi(TCGContext *s, int cc, tcg_insn_unit *dest)
48bb3750 1310{
8c081b18
RH
1311 ptrdiff_t off = dest - s->code_ptr;
1312 if (off == (int16_t)off) {
48bb3750
RH
1313 tcg_out_insn(s, RI, BRC, cc, off);
1314 } else if (off == (int32_t)off) {
1315 tcg_out_insn(s, RIL, BRCL, cc, off);
1316 } else {
8c081b18 1317 tcg_out_movi(s, TCG_TYPE_PTR, TCG_TMP0, (uintptr_t)dest);
48bb3750
RH
1318 tcg_out_insn(s, RR, BCR, cc, TCG_TMP0);
1319 }
1320}
1321
bec16311 1322static void tgen_branch(TCGContext *s, int cc, TCGLabel *l)
48bb3750 1323{
48bb3750 1324 if (l->has_value) {
8c081b18 1325 tgen_gotoi(s, cc, l->u.value_ptr);
48bb3750
RH
1326 } else if (USE_LONG_BRANCHES) {
1327 tcg_out16(s, RIL_BRCL | (cc << 4));
e692a349 1328 tcg_out_reloc(s, s->code_ptr, R_390_PC32DBL, l, 2);
8c081b18 1329 s->code_ptr += 2;
48bb3750
RH
1330 } else {
1331 tcg_out16(s, RI_BRC | (cc << 4));
e692a349 1332 tcg_out_reloc(s, s->code_ptr, R_390_PC16DBL, l, 2);
8c081b18 1333 s->code_ptr += 1;
48bb3750
RH
1334 }
1335}
1336
1337static void tgen_compare_branch(TCGContext *s, S390Opcode opc, int cc,
bec16311 1338 TCGReg r1, TCGReg r2, TCGLabel *l)
48bb3750 1339{
3661612f 1340 intptr_t off = 0;
48bb3750
RH
1341
1342 if (l->has_value) {
8c081b18 1343 off = l->u.value_ptr - s->code_ptr;
55dfd8fe 1344 tcg_debug_assert(off == (int16_t)off);
48bb3750 1345 } else {
e692a349 1346 tcg_out_reloc(s, s->code_ptr + 1, R_390_PC16DBL, l, 2);
48bb3750
RH
1347 }
1348
1349 tcg_out16(s, (opc & 0xff00) | (r1 << 4) | r2);
1350 tcg_out16(s, off);
1351 tcg_out16(s, cc << 12 | (opc & 0xff));
1352}
1353
1354static void tgen_compare_imm_branch(TCGContext *s, S390Opcode opc, int cc,
bec16311 1355 TCGReg r1, int i2, TCGLabel *l)
48bb3750 1356{
3661612f 1357 tcg_target_long off = 0;
48bb3750
RH
1358
1359 if (l->has_value) {
8c081b18 1360 off = l->u.value_ptr - s->code_ptr;
55dfd8fe 1361 tcg_debug_assert(off == (int16_t)off);
48bb3750 1362 } else {
e692a349 1363 tcg_out_reloc(s, s->code_ptr + 1, R_390_PC16DBL, l, 2);
48bb3750
RH
1364 }
1365
1366 tcg_out16(s, (opc & 0xff00) | (r1 << 4) | cc);
1367 tcg_out16(s, off);
1368 tcg_out16(s, (i2 << 8) | (opc & 0xff));
1369}
1370
1371static void tgen_brcond(TCGContext *s, TCGType type, TCGCond c,
bec16311 1372 TCGReg r1, TCGArg c2, int c2const, TCGLabel *l)
48bb3750
RH
1373{
1374 int cc;
1375
b2c98d9d 1376 if (s390_facilities & FACILITY_GEN_INST_EXT) {
b879f308 1377 bool is_unsigned = is_unsigned_cond(c);
48bb3750
RH
1378 bool in_range;
1379 S390Opcode opc;
1380
1381 cc = tcg_cond_to_s390_cond[c];
1382
1383 if (!c2const) {
1384 opc = (type == TCG_TYPE_I32
1385 ? (is_unsigned ? RIE_CLRJ : RIE_CRJ)
1386 : (is_unsigned ? RIE_CLGRJ : RIE_CGRJ));
bec16311 1387 tgen_compare_branch(s, opc, cc, r1, c2, l);
48bb3750
RH
1388 return;
1389 }
1390
1391 /* COMPARE IMMEDIATE AND BRANCH RELATIVE has an 8-bit immediate field.
1392 If the immediate we've been given does not fit that range, we'll
1393 fall back to separate compare and branch instructions using the
1394 larger comparison range afforded by COMPARE IMMEDIATE. */
1395 if (type == TCG_TYPE_I32) {
1396 if (is_unsigned) {
1397 opc = RIE_CLIJ;
1398 in_range = (uint32_t)c2 == (uint8_t)c2;
1399 } else {
1400 opc = RIE_CIJ;
1401 in_range = (int32_t)c2 == (int8_t)c2;
1402 }
1403 } else {
1404 if (is_unsigned) {
1405 opc = RIE_CLGIJ;
1406 in_range = (uint64_t)c2 == (uint8_t)c2;
1407 } else {
1408 opc = RIE_CGIJ;
1409 in_range = (int64_t)c2 == (int8_t)c2;
1410 }
1411 }
1412 if (in_range) {
bec16311 1413 tgen_compare_imm_branch(s, opc, cc, r1, c2, l);
48bb3750
RH
1414 return;
1415 }
1416 }
1417
65839b56 1418 cc = tgen_cmp(s, type, c, r1, c2, c2const, false);
bec16311 1419 tgen_branch(s, cc, l);
48bb3750
RH
1420}
1421
a8111212 1422static void tcg_out_call(TCGContext *s, tcg_insn_unit *dest)
48bb3750 1423{
8c081b18 1424 ptrdiff_t off = dest - s->code_ptr;
48bb3750
RH
1425 if (off == (int32_t)off) {
1426 tcg_out_insn(s, RIL, BRASL, TCG_REG_R14, off);
1427 } else {
8c081b18 1428 tcg_out_movi(s, TCG_TYPE_PTR, TCG_TMP0, (uintptr_t)dest);
48bb3750
RH
1429 tcg_out_insn(s, RR, BASR, TCG_REG_R14, TCG_TMP0);
1430 }
1431}
1432
a5a04f28 1433static void tcg_out_qemu_ld_direct(TCGContext *s, TCGMemOp opc, TCGReg data,
48bb3750
RH
1434 TCGReg base, TCGReg index, int disp)
1435{
3c8691f5 1436 switch (opc & (MO_SSIZE | MO_BSWAP)) {
a5a04f28 1437 case MO_UB:
48bb3750
RH
1438 tcg_out_insn(s, RXY, LLGC, data, base, index, disp);
1439 break;
a5a04f28 1440 case MO_SB:
48bb3750
RH
1441 tcg_out_insn(s, RXY, LGB, data, base, index, disp);
1442 break;
b8dd88b8
RH
1443
1444 case MO_UW | MO_BSWAP:
1445 /* swapped unsigned halfword load with upper bits zeroed */
1446 tcg_out_insn(s, RXY, LRVH, data, base, index, disp);
1447 tgen_ext16u(s, TCG_TYPE_I64, data, data);
1448 break;
a5a04f28 1449 case MO_UW:
b8dd88b8
RH
1450 tcg_out_insn(s, RXY, LLGH, data, base, index, disp);
1451 break;
1452
1453 case MO_SW | MO_BSWAP:
1454 /* swapped sign-extended halfword load */
1455 tcg_out_insn(s, RXY, LRVH, data, base, index, disp);
1456 tgen_ext16s(s, TCG_TYPE_I64, data, data);
48bb3750 1457 break;
a5a04f28 1458 case MO_SW:
b8dd88b8
RH
1459 tcg_out_insn(s, RXY, LGH, data, base, index, disp);
1460 break;
1461
1462 case MO_UL | MO_BSWAP:
1463 /* swapped unsigned int load with upper bits zeroed */
1464 tcg_out_insn(s, RXY, LRV, data, base, index, disp);
1465 tgen_ext32u(s, data, data);
48bb3750 1466 break;
a5a04f28 1467 case MO_UL:
b8dd88b8
RH
1468 tcg_out_insn(s, RXY, LLGF, data, base, index, disp);
1469 break;
1470
1471 case MO_SL | MO_BSWAP:
1472 /* swapped sign-extended int load */
1473 tcg_out_insn(s, RXY, LRV, data, base, index, disp);
1474 tgen_ext32s(s, data, data);
48bb3750 1475 break;
a5a04f28 1476 case MO_SL:
b8dd88b8
RH
1477 tcg_out_insn(s, RXY, LGF, data, base, index, disp);
1478 break;
1479
1480 case MO_Q | MO_BSWAP:
1481 tcg_out_insn(s, RXY, LRVG, data, base, index, disp);
48bb3750 1482 break;
a5a04f28 1483 case MO_Q:
b8dd88b8 1484 tcg_out_insn(s, RXY, LG, data, base, index, disp);
48bb3750 1485 break;
b8dd88b8 1486
48bb3750
RH
1487 default:
1488 tcg_abort();
1489 }
1490}
1491
a5a04f28 1492static void tcg_out_qemu_st_direct(TCGContext *s, TCGMemOp opc, TCGReg data,
48bb3750
RH
1493 TCGReg base, TCGReg index, int disp)
1494{
3c8691f5 1495 switch (opc & (MO_SIZE | MO_BSWAP)) {
a5a04f28 1496 case MO_UB:
48bb3750
RH
1497 if (disp >= 0 && disp < 0x1000) {
1498 tcg_out_insn(s, RX, STC, data, base, index, disp);
1499 } else {
1500 tcg_out_insn(s, RXY, STCY, data, base, index, disp);
1501 }
1502 break;
b8dd88b8
RH
1503
1504 case MO_UW | MO_BSWAP:
1505 tcg_out_insn(s, RXY, STRVH, data, base, index, disp);
1506 break;
a5a04f28 1507 case MO_UW:
b8dd88b8 1508 if (disp >= 0 && disp < 0x1000) {
48bb3750
RH
1509 tcg_out_insn(s, RX, STH, data, base, index, disp);
1510 } else {
1511 tcg_out_insn(s, RXY, STHY, data, base, index, disp);
1512 }
1513 break;
b8dd88b8
RH
1514
1515 case MO_UL | MO_BSWAP:
1516 tcg_out_insn(s, RXY, STRV, data, base, index, disp);
1517 break;
a5a04f28 1518 case MO_UL:
b8dd88b8 1519 if (disp >= 0 && disp < 0x1000) {
48bb3750
RH
1520 tcg_out_insn(s, RX, ST, data, base, index, disp);
1521 } else {
1522 tcg_out_insn(s, RXY, STY, data, base, index, disp);
1523 }
1524 break;
b8dd88b8
RH
1525
1526 case MO_Q | MO_BSWAP:
1527 tcg_out_insn(s, RXY, STRVG, data, base, index, disp);
1528 break;
a5a04f28 1529 case MO_Q:
b8dd88b8 1530 tcg_out_insn(s, RXY, STG, data, base, index, disp);
48bb3750 1531 break;
b8dd88b8 1532
48bb3750
RH
1533 default:
1534 tcg_abort();
1535 }
1536}
1537
1538#if defined(CONFIG_SOFTMMU)
659ef5cb
RH
1539#include "tcg-ldst.inc.c"
1540
269bd5d8
RH
1541/* We're expecting to use a 20-bit negative offset on the tlb memory ops. */
1542QEMU_BUILD_BUG_ON(TLB_MASK_TABLE_OFS(0) > 0);
1543QEMU_BUILD_BUG_ON(TLB_MASK_TABLE_OFS(0) < -(1 << 19));
fb596415
RH
1544
1545/* Load and compare a TLB entry, leaving the flags set. Loads the TLB
1546 addend into R2. Returns a register with the santitized guest address. */
1547static TCGReg tcg_out_tlb_read(TCGContext* s, TCGReg addr_reg, TCGMemOp opc,
1548 int mem_index, bool is_ld)
48bb3750 1549{
85aa8081
RH
1550 unsigned s_bits = opc & MO_SIZE;
1551 unsigned a_bits = get_alignment_bits(opc);
1552 unsigned s_mask = (1 << s_bits) - 1;
1553 unsigned a_mask = (1 << a_bits) - 1;
269bd5d8
RH
1554 int fast_off = TLB_MASK_TABLE_OFS(mem_index);
1555 int mask_off = fast_off + offsetof(CPUTLBDescFast, mask);
1556 int table_off = fast_off + offsetof(CPUTLBDescFast, table);
a5e39810
RH
1557 int ofs, a_off;
1558 uint64_t tlb_mask;
1559
4f47e338
RH
1560 tcg_out_sh64(s, RSY_SRLG, TCG_REG_R2, addr_reg, TCG_REG_NONE,
1561 TARGET_PAGE_BITS - CPU_TLB_ENTRY_BITS);
1562 tcg_out_insn(s, RXY, NG, TCG_REG_R2, TCG_AREG0, TCG_REG_NONE, mask_off);
1563 tcg_out_insn(s, RXY, AG, TCG_REG_R2, TCG_AREG0, TCG_REG_NONE, table_off);
1564
a5e39810
RH
1565 /* For aligned accesses, we check the first byte and include the alignment
1566 bits within the address. For unaligned access, we check that we don't
1567 cross pages using the address of the last byte of the access. */
85aa8081
RH
1568 a_off = (a_bits >= s_bits ? 0 : s_mask - a_mask);
1569 tlb_mask = (uint64_t)TARGET_PAGE_MASK | a_mask;
4f47e338
RH
1570 if ((s390_facilities & FACILITY_GEN_INST_EXT) && a_off == 0) {
1571 tgen_andi_risbg(s, TCG_REG_R3, addr_reg, tlb_mask);
48bb3750 1572 } else {
a5e39810 1573 tcg_out_insn(s, RX, LA, TCG_REG_R3, addr_reg, TCG_REG_NONE, a_off);
547ec121 1574 tgen_andi(s, TCG_TYPE_TL, TCG_REG_R3, tlb_mask);
48bb3750
RH
1575 }
1576
fb596415 1577 if (is_ld) {
4f47e338 1578 ofs = offsetof(CPUTLBEntry, addr_read);
fb596415 1579 } else {
4f47e338 1580 ofs = offsetof(CPUTLBEntry, addr_write);
48bb3750 1581 }
48bb3750 1582 if (TARGET_LONG_BITS == 32) {
4f47e338 1583 tcg_out_insn(s, RX, C, TCG_REG_R3, TCG_REG_R2, TCG_REG_NONE, ofs);
48bb3750 1584 } else {
4f47e338 1585 tcg_out_insn(s, RXY, CG, TCG_REG_R3, TCG_REG_R2, TCG_REG_NONE, ofs);
48bb3750
RH
1586 }
1587
4f47e338
RH
1588 tcg_out_insn(s, RXY, LG, TCG_REG_R2, TCG_REG_R2, TCG_REG_NONE,
1589 offsetof(CPUTLBEntry, addend));
fb596415 1590
48bb3750 1591 if (TARGET_LONG_BITS == 32) {
fb596415
RH
1592 tgen_ext32u(s, TCG_REG_R3, addr_reg);
1593 return TCG_REG_R3;
48bb3750 1594 }
fb596415
RH
1595 return addr_reg;
1596}
48bb3750 1597
3972ef6f
RH
1598static void add_qemu_ldst_label(TCGContext *s, bool is_ld, TCGMemOpIdx oi,
1599 TCGReg data, TCGReg addr,
fb596415
RH
1600 tcg_insn_unit *raddr, tcg_insn_unit *label_ptr)
1601{
1602 TCGLabelQemuLdst *label = new_ldst_label(s);
1603
1604 label->is_ld = is_ld;
3972ef6f 1605 label->oi = oi;
fb596415
RH
1606 label->datalo_reg = data;
1607 label->addrlo_reg = addr;
fb596415
RH
1608 label->raddr = raddr;
1609 label->label_ptr[0] = label_ptr;
1610}
48bb3750 1611
aeee05f5 1612static bool tcg_out_qemu_ld_slow_path(TCGContext *s, TCGLabelQemuLdst *lb)
fb596415
RH
1613{
1614 TCGReg addr_reg = lb->addrlo_reg;
1615 TCGReg data_reg = lb->datalo_reg;
3972ef6f
RH
1616 TCGMemOpIdx oi = lb->oi;
1617 TCGMemOp opc = get_memop(oi);
48bb3750 1618
aeee05f5
RH
1619 if (!patch_reloc(lb->label_ptr[0], R_390_PC16DBL,
1620 (intptr_t)s->code_ptr, 2)) {
1621 return false;
1622 }
48bb3750 1623
fb596415
RH
1624 tcg_out_mov(s, TCG_TYPE_PTR, TCG_REG_R2, TCG_AREG0);
1625 if (TARGET_LONG_BITS == 64) {
1626 tcg_out_mov(s, TCG_TYPE_I64, TCG_REG_R3, addr_reg);
1627 }
3972ef6f 1628 tcg_out_movi(s, TCG_TYPE_I32, TCG_REG_R4, oi);
fb596415 1629 tcg_out_movi(s, TCG_TYPE_PTR, TCG_REG_R5, (uintptr_t)lb->raddr);
2b7ec66f 1630 tcg_out_call(s, qemu_ld_helpers[opc & (MO_BSWAP | MO_SSIZE)]);
fb596415 1631 tcg_out_mov(s, TCG_TYPE_I64, data_reg, TCG_REG_R2);
65a62a75 1632
fb596415 1633 tgen_gotoi(s, S390_CC_ALWAYS, lb->raddr);
aeee05f5 1634 return true;
48bb3750
RH
1635}
1636
aeee05f5 1637static bool tcg_out_qemu_st_slow_path(TCGContext *s, TCGLabelQemuLdst *lb)
48bb3750 1638{
fb596415
RH
1639 TCGReg addr_reg = lb->addrlo_reg;
1640 TCGReg data_reg = lb->datalo_reg;
3972ef6f
RH
1641 TCGMemOpIdx oi = lb->oi;
1642 TCGMemOp opc = get_memop(oi);
fb596415 1643
aeee05f5
RH
1644 if (!patch_reloc(lb->label_ptr[0], R_390_PC16DBL,
1645 (intptr_t)s->code_ptr, 2)) {
1646 return false;
1647 }
fb596415
RH
1648
1649 tcg_out_mov(s, TCG_TYPE_PTR, TCG_REG_R2, TCG_AREG0);
1650 if (TARGET_LONG_BITS == 64) {
1651 tcg_out_mov(s, TCG_TYPE_I64, TCG_REG_R3, addr_reg);
1652 }
1653 switch (opc & MO_SIZE) {
1654 case MO_UB:
1655 tgen_ext8u(s, TCG_TYPE_I64, TCG_REG_R4, data_reg);
1656 break;
1657 case MO_UW:
1658 tgen_ext16u(s, TCG_TYPE_I64, TCG_REG_R4, data_reg);
1659 break;
1660 case MO_UL:
1661 tgen_ext32u(s, TCG_REG_R4, data_reg);
1662 break;
1663 case MO_Q:
1664 tcg_out_mov(s, TCG_TYPE_I64, TCG_REG_R4, data_reg);
1665 break;
1666 default:
1667 tcg_abort();
1668 }
3972ef6f 1669 tcg_out_movi(s, TCG_TYPE_I32, TCG_REG_R5, oi);
fb596415 1670 tcg_out_movi(s, TCG_TYPE_PTR, TCG_REG_R6, (uintptr_t)lb->raddr);
2b7ec66f 1671 tcg_out_call(s, qemu_st_helpers[opc & (MO_BSWAP | MO_SIZE)]);
fb596415
RH
1672
1673 tgen_gotoi(s, S390_CC_ALWAYS, lb->raddr);
aeee05f5 1674 return true;
48bb3750
RH
1675}
1676#else
1677static void tcg_prepare_user_ldst(TCGContext *s, TCGReg *addr_reg,
1678 TCGReg *index_reg, tcg_target_long *disp)
1679{
1680 if (TARGET_LONG_BITS == 32) {
1681 tgen_ext32u(s, TCG_TMP0, *addr_reg);
1682 *addr_reg = TCG_TMP0;
1683 }
b76f21a7 1684 if (guest_base < 0x80000) {
48bb3750 1685 *index_reg = TCG_REG_NONE;
b76f21a7 1686 *disp = guest_base;
48bb3750
RH
1687 } else {
1688 *index_reg = TCG_GUEST_BASE_REG;
1689 *disp = 0;
1690 }
1691}
1692#endif /* CONFIG_SOFTMMU */
1693
f24efee4 1694static void tcg_out_qemu_ld(TCGContext* s, TCGReg data_reg, TCGReg addr_reg,
59227d5d 1695 TCGMemOpIdx oi)
48bb3750 1696{
59227d5d 1697 TCGMemOp opc = get_memop(oi);
fb596415 1698#ifdef CONFIG_SOFTMMU
59227d5d 1699 unsigned mem_index = get_mmuidx(oi);
fb596415
RH
1700 tcg_insn_unit *label_ptr;
1701 TCGReg base_reg;
1702
1703 base_reg = tcg_out_tlb_read(s, addr_reg, opc, mem_index, 1);
1704
cd3b29b7
AJ
1705 tcg_out16(s, RI_BRC | (S390_CC_NE << 4));
1706 label_ptr = s->code_ptr;
1707 s->code_ptr += 1;
fb596415
RH
1708
1709 tcg_out_qemu_ld_direct(s, opc, data_reg, base_reg, TCG_REG_R2, 0);
48bb3750 1710
3972ef6f 1711 add_qemu_ldst_label(s, 1, oi, data_reg, addr_reg, s->code_ptr, label_ptr);
48bb3750 1712#else
f24efee4
RH
1713 TCGReg index_reg;
1714 tcg_target_long disp;
1715
48bb3750
RH
1716 tcg_prepare_user_ldst(s, &addr_reg, &index_reg, &disp);
1717 tcg_out_qemu_ld_direct(s, opc, data_reg, addr_reg, index_reg, disp);
1718#endif
1719}
1720
f24efee4 1721static void tcg_out_qemu_st(TCGContext* s, TCGReg data_reg, TCGReg addr_reg,
59227d5d 1722 TCGMemOpIdx oi)
48bb3750 1723{
59227d5d 1724 TCGMemOp opc = get_memop(oi);
fb596415 1725#ifdef CONFIG_SOFTMMU
59227d5d 1726 unsigned mem_index = get_mmuidx(oi);
fb596415
RH
1727 tcg_insn_unit *label_ptr;
1728 TCGReg base_reg;
1729
1730 base_reg = tcg_out_tlb_read(s, addr_reg, opc, mem_index, 0);
1731
cd3b29b7
AJ
1732 tcg_out16(s, RI_BRC | (S390_CC_NE << 4));
1733 label_ptr = s->code_ptr;
1734 s->code_ptr += 1;
fb596415
RH
1735
1736 tcg_out_qemu_st_direct(s, opc, data_reg, base_reg, TCG_REG_R2, 0);
48bb3750 1737
3972ef6f 1738 add_qemu_ldst_label(s, 0, oi, data_reg, addr_reg, s->code_ptr, label_ptr);
48bb3750 1739#else
f24efee4
RH
1740 TCGReg index_reg;
1741 tcg_target_long disp;
1742
48bb3750
RH
1743 tcg_prepare_user_ldst(s, &addr_reg, &index_reg, &disp);
1744 tcg_out_qemu_st_direct(s, opc, data_reg, addr_reg, index_reg, disp);
1745#endif
2827822e
AG
1746}
1747
48bb3750
RH
1748# define OP_32_64(x) \
1749 case glue(glue(INDEX_op_,x),_i32): \
1750 case glue(glue(INDEX_op_,x),_i64)
48bb3750 1751
a9751609 1752static inline void tcg_out_op(TCGContext *s, TCGOpcode opc,
2827822e
AG
1753 const TCGArg *args, const int *const_args)
1754{
c2097136 1755 S390Opcode op, op2;
0db921e6 1756 TCGArg a0, a1, a2;
48bb3750
RH
1757
1758 switch (opc) {
1759 case INDEX_op_exit_tb:
46644483
RH
1760 /* Reuse the zeroing that exists for goto_ptr. */
1761 a0 = args[0];
1762 if (a0 == 0) {
1763 tgen_gotoi(s, S390_CC_ALWAYS, s->code_gen_epilogue);
1764 } else {
1765 tcg_out_movi(s, TCG_TYPE_PTR, TCG_REG_R2, a0);
1766 tgen_gotoi(s, S390_CC_ALWAYS, tb_ret_addr);
1767 }
48bb3750
RH
1768 break;
1769
1770 case INDEX_op_goto_tb:
829e1376 1771 a0 = args[0];
f309101c 1772 if (s->tb_jmp_insn_offset) {
ed3d51ec
SF
1773 /* branch displacement must be aligned for atomic patching;
1774 * see if we need to add extra nop before branch
1775 */
1776 if (!QEMU_PTR_IS_ALIGNED(s->code_ptr + 1, 4)) {
1777 tcg_out16(s, NOP);
1778 }
829e1376 1779 tcg_debug_assert(!USE_REG_TB);
a10c64e0 1780 tcg_out16(s, RIL_BRCL | (S390_CC_ALWAYS << 4));
829e1376 1781 s->tb_jmp_insn_offset[a0] = tcg_current_code_size(s);
a10c64e0 1782 s->code_ptr += 2;
48bb3750 1783 } else {
829e1376
RH
1784 /* load address stored at s->tb_jmp_target_addr + a0 */
1785 tcg_out_ld_abs(s, TCG_TYPE_PTR, TCG_REG_TB,
1786 s->tb_jmp_target_addr + a0);
48bb3750 1787 /* and go there */
829e1376
RH
1788 tcg_out_insn(s, RR, BCR, S390_CC_ALWAYS, TCG_REG_TB);
1789 }
9f754620 1790 set_jmp_reset_offset(s, a0);
829e1376
RH
1791
1792 /* For the unlinked path of goto_tb, we need to reset
1793 TCG_REG_TB to the beginning of this TB. */
1794 if (USE_REG_TB) {
1795 int ofs = -tcg_current_code_size(s);
1796 assert(ofs == (int16_t)ofs);
1797 tcg_out_insn(s, RI, AGHI, TCG_REG_TB, ofs);
48bb3750 1798 }
48bb3750
RH
1799 break;
1800
46644483 1801 case INDEX_op_goto_ptr:
829e1376
RH
1802 a0 = args[0];
1803 if (USE_REG_TB) {
1804 tcg_out_mov(s, TCG_TYPE_PTR, TCG_REG_TB, a0);
1805 }
1806 tcg_out_insn(s, RR, BCR, S390_CC_ALWAYS, a0);
46644483
RH
1807 break;
1808
48bb3750
RH
1809 OP_32_64(ld8u):
1810 /* ??? LLC (RXY format) is only present with the extended-immediate
1811 facility, whereas LLGC is always present. */
1812 tcg_out_mem(s, 0, RXY_LLGC, args[0], args[1], TCG_REG_NONE, args[2]);
1813 break;
1814
1815 OP_32_64(ld8s):
1816 /* ??? LB is no smaller than LGB, so no point to using it. */
1817 tcg_out_mem(s, 0, RXY_LGB, args[0], args[1], TCG_REG_NONE, args[2]);
1818 break;
1819
1820 OP_32_64(ld16u):
1821 /* ??? LLH (RXY format) is only present with the extended-immediate
1822 facility, whereas LLGH is always present. */
1823 tcg_out_mem(s, 0, RXY_LLGH, args[0], args[1], TCG_REG_NONE, args[2]);
1824 break;
1825
1826 case INDEX_op_ld16s_i32:
1827 tcg_out_mem(s, RX_LH, RXY_LHY, args[0], args[1], TCG_REG_NONE, args[2]);
1828 break;
1829
1830 case INDEX_op_ld_i32:
1831 tcg_out_ld(s, TCG_TYPE_I32, args[0], args[1], args[2]);
1832 break;
1833
1834 OP_32_64(st8):
1835 tcg_out_mem(s, RX_STC, RXY_STCY, args[0], args[1],
1836 TCG_REG_NONE, args[2]);
1837 break;
1838
1839 OP_32_64(st16):
1840 tcg_out_mem(s, RX_STH, RXY_STHY, args[0], args[1],
1841 TCG_REG_NONE, args[2]);
1842 break;
1843
1844 case INDEX_op_st_i32:
1845 tcg_out_st(s, TCG_TYPE_I32, args[0], args[1], args[2]);
1846 break;
1847
1848 case INDEX_op_add_i32:
0db921e6 1849 a0 = args[0], a1 = args[1], a2 = (int32_t)args[2];
48bb3750 1850 if (const_args[2]) {
0db921e6
RH
1851 do_addi_32:
1852 if (a0 == a1) {
1853 if (a2 == (int16_t)a2) {
1854 tcg_out_insn(s, RI, AHI, a0, a2);
1855 break;
1856 }
b2c98d9d 1857 if (s390_facilities & FACILITY_EXT_IMM) {
0db921e6
RH
1858 tcg_out_insn(s, RIL, AFI, a0, a2);
1859 break;
1860 }
1861 }
1862 tcg_out_mem(s, RX_LA, RXY_LAY, a0, a1, TCG_REG_NONE, a2);
1863 } else if (a0 == a1) {
1864 tcg_out_insn(s, RR, AR, a0, a2);
48bb3750 1865 } else {
0db921e6 1866 tcg_out_insn(s, RX, LA, a0, a1, a2, 0);
48bb3750
RH
1867 }
1868 break;
1869 case INDEX_op_sub_i32:
0db921e6 1870 a0 = args[0], a1 = args[1], a2 = (int32_t)args[2];
48bb3750 1871 if (const_args[2]) {
0db921e6
RH
1872 a2 = -a2;
1873 goto do_addi_32;
c2097136
RH
1874 } else if (a0 == a1) {
1875 tcg_out_insn(s, RR, SR, a0, a2);
1876 } else {
1877 tcg_out_insn(s, RRF, SRK, a0, a1, a2);
48bb3750
RH
1878 }
1879 break;
1880
1881 case INDEX_op_and_i32:
c2097136 1882 a0 = args[0], a1 = args[1], a2 = (uint32_t)args[2];
48bb3750 1883 if (const_args[2]) {
c2097136
RH
1884 tcg_out_mov(s, TCG_TYPE_I32, a0, a1);
1885 tgen_andi(s, TCG_TYPE_I32, a0, a2);
1886 } else if (a0 == a1) {
1887 tcg_out_insn(s, RR, NR, a0, a2);
48bb3750 1888 } else {
c2097136 1889 tcg_out_insn(s, RRF, NRK, a0, a1, a2);
48bb3750
RH
1890 }
1891 break;
1892 case INDEX_op_or_i32:
c2097136 1893 a0 = args[0], a1 = args[1], a2 = (uint32_t)args[2];
48bb3750 1894 if (const_args[2]) {
c2097136 1895 tcg_out_mov(s, TCG_TYPE_I32, a0, a1);
4046d9ca 1896 tgen_ori(s, TCG_TYPE_I32, a0, a2);
c2097136
RH
1897 } else if (a0 == a1) {
1898 tcg_out_insn(s, RR, OR, a0, a2);
48bb3750 1899 } else {
c2097136 1900 tcg_out_insn(s, RRF, ORK, a0, a1, a2);
48bb3750
RH
1901 }
1902 break;
1903 case INDEX_op_xor_i32:
c2097136 1904 a0 = args[0], a1 = args[1], a2 = (uint32_t)args[2];
48bb3750 1905 if (const_args[2]) {
c2097136 1906 tcg_out_mov(s, TCG_TYPE_I32, a0, a1);
5bf67a92 1907 tgen_xori(s, TCG_TYPE_I32, a0, a2);
c2097136 1908 } else if (a0 == a1) {
48bb3750 1909 tcg_out_insn(s, RR, XR, args[0], args[2]);
c2097136
RH
1910 } else {
1911 tcg_out_insn(s, RRF, XRK, a0, a1, a2);
48bb3750
RH
1912 }
1913 break;
1914
1915 case INDEX_op_neg_i32:
1916 tcg_out_insn(s, RR, LCR, args[0], args[1]);
1917 break;
1918
1919 case INDEX_op_mul_i32:
1920 if (const_args[2]) {
1921 if ((int32_t)args[2] == (int16_t)args[2]) {
1922 tcg_out_insn(s, RI, MHI, args[0], args[2]);
1923 } else {
1924 tcg_out_insn(s, RIL, MSFI, args[0], args[2]);
1925 }
1926 } else {
1927 tcg_out_insn(s, RRE, MSR, args[0], args[2]);
1928 }
1929 break;
1930
1931 case INDEX_op_div2_i32:
1932 tcg_out_insn(s, RR, DR, TCG_REG_R2, args[4]);
1933 break;
1934 case INDEX_op_divu2_i32:
1935 tcg_out_insn(s, RRE, DLR, TCG_REG_R2, args[4]);
1936 break;
1937
1938 case INDEX_op_shl_i32:
1939 op = RS_SLL;
c2097136 1940 op2 = RSY_SLLK;
48bb3750 1941 do_shift32:
c2097136
RH
1942 a0 = args[0], a1 = args[1], a2 = (int32_t)args[2];
1943 if (a0 == a1) {
1944 if (const_args[2]) {
1945 tcg_out_sh32(s, op, a0, TCG_REG_NONE, a2);
1946 } else {
1947 tcg_out_sh32(s, op, a0, a2, 0);
1948 }
48bb3750 1949 } else {
c2097136
RH
1950 /* Using tcg_out_sh64 here for the format; it is a 32-bit shift. */
1951 if (const_args[2]) {
1952 tcg_out_sh64(s, op2, a0, a1, TCG_REG_NONE, a2);
1953 } else {
1954 tcg_out_sh64(s, op2, a0, a1, a2, 0);
1955 }
48bb3750
RH
1956 }
1957 break;
1958 case INDEX_op_shr_i32:
1959 op = RS_SRL;
c2097136 1960 op2 = RSY_SRLK;
48bb3750
RH
1961 goto do_shift32;
1962 case INDEX_op_sar_i32:
1963 op = RS_SRA;
c2097136 1964 op2 = RSY_SRAK;
48bb3750
RH
1965 goto do_shift32;
1966
1967 case INDEX_op_rotl_i32:
1968 /* ??? Using tcg_out_sh64 here for the format; it is a 32-bit rol. */
1969 if (const_args[2]) {
1970 tcg_out_sh64(s, RSY_RLL, args[0], args[1], TCG_REG_NONE, args[2]);
1971 } else {
1972 tcg_out_sh64(s, RSY_RLL, args[0], args[1], args[2], 0);
1973 }
1974 break;
1975 case INDEX_op_rotr_i32:
1976 if (const_args[2]) {
1977 tcg_out_sh64(s, RSY_RLL, args[0], args[1],
1978 TCG_REG_NONE, (32 - args[2]) & 31);
1979 } else {
1980 tcg_out_insn(s, RR, LCR, TCG_TMP0, args[2]);
1981 tcg_out_sh64(s, RSY_RLL, args[0], args[1], TCG_TMP0, 0);
1982 }
1983 break;
1984
1985 case INDEX_op_ext8s_i32:
1986 tgen_ext8s(s, TCG_TYPE_I32, args[0], args[1]);
1987 break;
1988 case INDEX_op_ext16s_i32:
1989 tgen_ext16s(s, TCG_TYPE_I32, args[0], args[1]);
1990 break;
1991 case INDEX_op_ext8u_i32:
1992 tgen_ext8u(s, TCG_TYPE_I32, args[0], args[1]);
1993 break;
1994 case INDEX_op_ext16u_i32:
1995 tgen_ext16u(s, TCG_TYPE_I32, args[0], args[1]);
1996 break;
1997
1998 OP_32_64(bswap16):
1999 /* The TCG bswap definition requires bits 0-47 already be zero.
2000 Thus we don't need the G-type insns to implement bswap16_i64. */
2001 tcg_out_insn(s, RRE, LRVR, args[0], args[1]);
2002 tcg_out_sh32(s, RS_SRL, args[0], TCG_REG_NONE, 16);
2003 break;
2004 OP_32_64(bswap32):
2005 tcg_out_insn(s, RRE, LRVR, args[0], args[1]);
2006 break;
2007
3790b918 2008 case INDEX_op_add2_i32:
ad19b358
RH
2009 if (const_args[4]) {
2010 tcg_out_insn(s, RIL, ALFI, args[0], args[4]);
2011 } else {
2012 tcg_out_insn(s, RR, ALR, args[0], args[4]);
2013 }
3790b918
RH
2014 tcg_out_insn(s, RRE, ALCR, args[1], args[5]);
2015 break;
2016 case INDEX_op_sub2_i32:
ad19b358
RH
2017 if (const_args[4]) {
2018 tcg_out_insn(s, RIL, SLFI, args[0], args[4]);
2019 } else {
2020 tcg_out_insn(s, RR, SLR, args[0], args[4]);
2021 }
3790b918
RH
2022 tcg_out_insn(s, RRE, SLBR, args[1], args[5]);
2023 break;
2024
48bb3750 2025 case INDEX_op_br:
bec16311 2026 tgen_branch(s, S390_CC_ALWAYS, arg_label(args[0]));
48bb3750
RH
2027 break;
2028
2029 case INDEX_op_brcond_i32:
2030 tgen_brcond(s, TCG_TYPE_I32, args[2], args[0],
bec16311 2031 args[1], const_args[1], arg_label(args[3]));
48bb3750
RH
2032 break;
2033 case INDEX_op_setcond_i32:
2034 tgen_setcond(s, TCG_TYPE_I32, args[3], args[0], args[1],
2035 args[2], const_args[2]);
2036 break;
96a9f093
RH
2037 case INDEX_op_movcond_i32:
2038 tgen_movcond(s, TCG_TYPE_I32, args[5], args[0], args[1],
7af525af 2039 args[2], const_args[2], args[3], const_args[3]);
96a9f093 2040 break;
48bb3750 2041
f24efee4 2042 case INDEX_op_qemu_ld_i32:
48bb3750 2043 /* ??? Technically we can use a non-extending instruction. */
f24efee4 2044 case INDEX_op_qemu_ld_i64:
59227d5d 2045 tcg_out_qemu_ld(s, args[0], args[1], args[2]);
48bb3750 2046 break;
f24efee4
RH
2047 case INDEX_op_qemu_st_i32:
2048 case INDEX_op_qemu_st_i64:
59227d5d 2049 tcg_out_qemu_st(s, args[0], args[1], args[2]);
48bb3750
RH
2050 break;
2051
48bb3750
RH
2052 case INDEX_op_ld16s_i64:
2053 tcg_out_mem(s, 0, RXY_LGH, args[0], args[1], TCG_REG_NONE, args[2]);
2054 break;
2055 case INDEX_op_ld32u_i64:
2056 tcg_out_mem(s, 0, RXY_LLGF, args[0], args[1], TCG_REG_NONE, args[2]);
2057 break;
2058 case INDEX_op_ld32s_i64:
2059 tcg_out_mem(s, 0, RXY_LGF, args[0], args[1], TCG_REG_NONE, args[2]);
2060 break;
2061 case INDEX_op_ld_i64:
2062 tcg_out_ld(s, TCG_TYPE_I64, args[0], args[1], args[2]);
2063 break;
2064
2065 case INDEX_op_st32_i64:
2066 tcg_out_st(s, TCG_TYPE_I32, args[0], args[1], args[2]);
2067 break;
2068 case INDEX_op_st_i64:
2069 tcg_out_st(s, TCG_TYPE_I64, args[0], args[1], args[2]);
2070 break;
2071
2072 case INDEX_op_add_i64:
0db921e6 2073 a0 = args[0], a1 = args[1], a2 = args[2];
48bb3750 2074 if (const_args[2]) {
0db921e6
RH
2075 do_addi_64:
2076 if (a0 == a1) {
2077 if (a2 == (int16_t)a2) {
2078 tcg_out_insn(s, RI, AGHI, a0, a2);
2079 break;
2080 }
b2c98d9d 2081 if (s390_facilities & FACILITY_EXT_IMM) {
0db921e6
RH
2082 if (a2 == (int32_t)a2) {
2083 tcg_out_insn(s, RIL, AGFI, a0, a2);
2084 break;
2085 } else if (a2 == (uint32_t)a2) {
2086 tcg_out_insn(s, RIL, ALGFI, a0, a2);
2087 break;
2088 } else if (-a2 == (uint32_t)-a2) {
2089 tcg_out_insn(s, RIL, SLGFI, a0, -a2);
2090 break;
2091 }
2092 }
2093 }
2094 tcg_out_mem(s, RX_LA, RXY_LAY, a0, a1, TCG_REG_NONE, a2);
2095 } else if (a0 == a1) {
2096 tcg_out_insn(s, RRE, AGR, a0, a2);
48bb3750 2097 } else {
0db921e6 2098 tcg_out_insn(s, RX, LA, a0, a1, a2, 0);
48bb3750
RH
2099 }
2100 break;
2101 case INDEX_op_sub_i64:
0db921e6 2102 a0 = args[0], a1 = args[1], a2 = args[2];
48bb3750 2103 if (const_args[2]) {
0db921e6
RH
2104 a2 = -a2;
2105 goto do_addi_64;
c2097136
RH
2106 } else if (a0 == a1) {
2107 tcg_out_insn(s, RRE, SGR, a0, a2);
48bb3750 2108 } else {
c2097136 2109 tcg_out_insn(s, RRF, SGRK, a0, a1, a2);
48bb3750
RH
2110 }
2111 break;
2112
2113 case INDEX_op_and_i64:
c2097136 2114 a0 = args[0], a1 = args[1], a2 = args[2];
48bb3750 2115 if (const_args[2]) {
c2097136 2116 tcg_out_mov(s, TCG_TYPE_I64, a0, a1);
07ff7983 2117 tgen_andi(s, TCG_TYPE_I64, args[0], args[2]);
c2097136 2118 } else if (a0 == a1) {
48bb3750 2119 tcg_out_insn(s, RRE, NGR, args[0], args[2]);
c2097136
RH
2120 } else {
2121 tcg_out_insn(s, RRF, NGRK, a0, a1, a2);
48bb3750
RH
2122 }
2123 break;
2124 case INDEX_op_or_i64:
c2097136 2125 a0 = args[0], a1 = args[1], a2 = args[2];
48bb3750 2126 if (const_args[2]) {
c2097136 2127 tcg_out_mov(s, TCG_TYPE_I64, a0, a1);
4046d9ca 2128 tgen_ori(s, TCG_TYPE_I64, a0, a2);
c2097136
RH
2129 } else if (a0 == a1) {
2130 tcg_out_insn(s, RRE, OGR, a0, a2);
48bb3750 2131 } else {
c2097136 2132 tcg_out_insn(s, RRF, OGRK, a0, a1, a2);
48bb3750
RH
2133 }
2134 break;
2135 case INDEX_op_xor_i64:
c2097136 2136 a0 = args[0], a1 = args[1], a2 = args[2];
48bb3750 2137 if (const_args[2]) {
c2097136 2138 tcg_out_mov(s, TCG_TYPE_I64, a0, a1);
5bf67a92 2139 tgen_xori(s, TCG_TYPE_I64, a0, a2);
c2097136
RH
2140 } else if (a0 == a1) {
2141 tcg_out_insn(s, RRE, XGR, a0, a2);
48bb3750 2142 } else {
c2097136 2143 tcg_out_insn(s, RRF, XGRK, a0, a1, a2);
48bb3750
RH
2144 }
2145 break;
2146
2147 case INDEX_op_neg_i64:
2148 tcg_out_insn(s, RRE, LCGR, args[0], args[1]);
2149 break;
2150 case INDEX_op_bswap64_i64:
2151 tcg_out_insn(s, RRE, LRVGR, args[0], args[1]);
2152 break;
2153
2154 case INDEX_op_mul_i64:
2155 if (const_args[2]) {
2156 if (args[2] == (int16_t)args[2]) {
2157 tcg_out_insn(s, RI, MGHI, args[0], args[2]);
2158 } else {
2159 tcg_out_insn(s, RIL, MSGFI, args[0], args[2]);
2160 }
2161 } else {
2162 tcg_out_insn(s, RRE, MSGR, args[0], args[2]);
2163 }
2164 break;
2165
2166 case INDEX_op_div2_i64:
2167 /* ??? We get an unnecessary sign-extension of the dividend
2168 into R3 with this definition, but as we do in fact always
2169 produce both quotient and remainder using INDEX_op_div_i64
2170 instead requires jumping through even more hoops. */
2171 tcg_out_insn(s, RRE, DSGR, TCG_REG_R2, args[4]);
2172 break;
2173 case INDEX_op_divu2_i64:
2174 tcg_out_insn(s, RRE, DLGR, TCG_REG_R2, args[4]);
2175 break;
36017dc6
RH
2176 case INDEX_op_mulu2_i64:
2177 tcg_out_insn(s, RRE, MLGR, TCG_REG_R2, args[3]);
2178 break;
48bb3750
RH
2179
2180 case INDEX_op_shl_i64:
2181 op = RSY_SLLG;
2182 do_shift64:
2183 if (const_args[2]) {
2184 tcg_out_sh64(s, op, args[0], args[1], TCG_REG_NONE, args[2]);
2185 } else {
2186 tcg_out_sh64(s, op, args[0], args[1], args[2], 0);
2187 }
2188 break;
2189 case INDEX_op_shr_i64:
2190 op = RSY_SRLG;
2191 goto do_shift64;
2192 case INDEX_op_sar_i64:
2193 op = RSY_SRAG;
2194 goto do_shift64;
2195
2196 case INDEX_op_rotl_i64:
2197 if (const_args[2]) {
2198 tcg_out_sh64(s, RSY_RLLG, args[0], args[1],
2199 TCG_REG_NONE, args[2]);
2200 } else {
2201 tcg_out_sh64(s, RSY_RLLG, args[0], args[1], args[2], 0);
2202 }
2203 break;
2204 case INDEX_op_rotr_i64:
2205 if (const_args[2]) {
2206 tcg_out_sh64(s, RSY_RLLG, args[0], args[1],
2207 TCG_REG_NONE, (64 - args[2]) & 63);
2208 } else {
2209 /* We can use the smaller 32-bit negate because only the
2210 low 6 bits are examined for the rotate. */
2211 tcg_out_insn(s, RR, LCR, TCG_TMP0, args[2]);
2212 tcg_out_sh64(s, RSY_RLLG, args[0], args[1], TCG_TMP0, 0);
2213 }
2214 break;
2215
2216 case INDEX_op_ext8s_i64:
2217 tgen_ext8s(s, TCG_TYPE_I64, args[0], args[1]);
2218 break;
2219 case INDEX_op_ext16s_i64:
2220 tgen_ext16s(s, TCG_TYPE_I64, args[0], args[1]);
2221 break;
4f2331e5 2222 case INDEX_op_ext_i32_i64:
48bb3750
RH
2223 case INDEX_op_ext32s_i64:
2224 tgen_ext32s(s, args[0], args[1]);
2225 break;
2226 case INDEX_op_ext8u_i64:
2227 tgen_ext8u(s, TCG_TYPE_I64, args[0], args[1]);
2228 break;
2229 case INDEX_op_ext16u_i64:
2230 tgen_ext16u(s, TCG_TYPE_I64, args[0], args[1]);
2231 break;
4f2331e5 2232 case INDEX_op_extu_i32_i64:
48bb3750
RH
2233 case INDEX_op_ext32u_i64:
2234 tgen_ext32u(s, args[0], args[1]);
2235 break;
2236
3790b918 2237 case INDEX_op_add2_i64:
ad19b358
RH
2238 if (const_args[4]) {
2239 if ((int64_t)args[4] >= 0) {
2240 tcg_out_insn(s, RIL, ALGFI, args[0], args[4]);
2241 } else {
2242 tcg_out_insn(s, RIL, SLGFI, args[0], -args[4]);
2243 }
2244 } else {
2245 tcg_out_insn(s, RRE, ALGR, args[0], args[4]);
2246 }
3790b918
RH
2247 tcg_out_insn(s, RRE, ALCGR, args[1], args[5]);
2248 break;
2249 case INDEX_op_sub2_i64:
ad19b358
RH
2250 if (const_args[4]) {
2251 if ((int64_t)args[4] >= 0) {
2252 tcg_out_insn(s, RIL, SLGFI, args[0], args[4]);
2253 } else {
2254 tcg_out_insn(s, RIL, ALGFI, args[0], -args[4]);
2255 }
2256 } else {
2257 tcg_out_insn(s, RRE, SLGR, args[0], args[4]);
2258 }
3790b918
RH
2259 tcg_out_insn(s, RRE, SLBGR, args[1], args[5]);
2260 break;
2261
48bb3750
RH
2262 case INDEX_op_brcond_i64:
2263 tgen_brcond(s, TCG_TYPE_I64, args[2], args[0],
bec16311 2264 args[1], const_args[1], arg_label(args[3]));
48bb3750
RH
2265 break;
2266 case INDEX_op_setcond_i64:
2267 tgen_setcond(s, TCG_TYPE_I64, args[3], args[0], args[1],
2268 args[2], const_args[2]);
2269 break;
96a9f093
RH
2270 case INDEX_op_movcond_i64:
2271 tgen_movcond(s, TCG_TYPE_I64, args[5], args[0], args[1],
7af525af 2272 args[2], const_args[2], args[3], const_args[3]);
96a9f093 2273 break;
48bb3750 2274
d5690ea4 2275 OP_32_64(deposit):
752b1be9
RH
2276 a0 = args[0], a1 = args[1], a2 = args[2];
2277 if (const_args[1]) {
2278 tgen_deposit(s, a0, a2, args[3], args[4], 1);
2279 } else {
2280 /* Since we can't support "0Z" as a constraint, we allow a1 in
2281 any register. Fix things up as if a matching constraint. */
2282 if (a0 != a1) {
2283 TCGType type = (opc == INDEX_op_deposit_i64);
2284 if (a0 == a2) {
2285 tcg_out_mov(s, type, TCG_TMP0, a2);
2286 a2 = TCG_TMP0;
2287 }
2288 tcg_out_mov(s, type, a0, a1);
2289 }
2290 tgen_deposit(s, a0, a2, args[3], args[4], 0);
2291 }
d5690ea4 2292 break;
752b1be9 2293
b0bf5fe8
RH
2294 OP_32_64(extract):
2295 tgen_extract(s, args[0], args[1], args[2], args[3]);
2296 break;
d5690ea4 2297
ce411066
RH
2298 case INDEX_op_clz_i64:
2299 tgen_clz(s, args[0], args[1], args[2], const_args[2]);
2300 break;
2301
c9314d61
PK
2302 case INDEX_op_mb:
2303 /* The host memory model is quite strong, we simply need to
2304 serialize the instruction stream. */
2305 if (args[0] & TCG_MO_ST_LD) {
2306 tcg_out_insn(s, RR, BCR,
b2c98d9d 2307 s390_facilities & FACILITY_FAST_BCR_SER ? 14 : 15, 0);
c9314d61
PK
2308 }
2309 break;
2310
96d0ee7f
RH
2311 case INDEX_op_mov_i32: /* Always emitted via tcg_out_mov. */
2312 case INDEX_op_mov_i64:
2313 case INDEX_op_movi_i32: /* Always emitted via tcg_out_movi. */
2314 case INDEX_op_movi_i64:
2315 case INDEX_op_call: /* Always emitted via tcg_out_call. */
48bb3750 2316 default:
48bb3750
RH
2317 tcg_abort();
2318 }
2827822e
AG
2319}
2320
f69d277e
RH
2321static const TCGTargetOpDef *tcg_target_op_def(TCGOpcode op)
2322{
9b5500b6
RH
2323 static const TCGTargetOpDef r = { .args_ct_str = { "r" } };
2324 static const TCGTargetOpDef r_r = { .args_ct_str = { "r", "r" } };
2325 static const TCGTargetOpDef r_L = { .args_ct_str = { "r", "L" } };
2326 static const TCGTargetOpDef L_L = { .args_ct_str = { "L", "L" } };
07952d95 2327 static const TCGTargetOpDef r_ri = { .args_ct_str = { "r", "ri" } };
9b5500b6
RH
2328 static const TCGTargetOpDef r_r_ri = { .args_ct_str = { "r", "r", "ri" } };
2329 static const TCGTargetOpDef r_0_ri = { .args_ct_str = { "r", "0", "ri" } };
a8f0269e
RH
2330 static const TCGTargetOpDef r_0_rI = { .args_ct_str = { "r", "0", "rI" } };
2331 static const TCGTargetOpDef r_0_rJ = { .args_ct_str = { "r", "0", "rJ" } };
ba18b07d
RH
2332 static const TCGTargetOpDef a2_r
2333 = { .args_ct_str = { "r", "r", "0", "1", "r", "r" } };
2334 static const TCGTargetOpDef a2_ri
2335 = { .args_ct_str = { "r", "r", "0", "1", "ri", "r" } };
2336 static const TCGTargetOpDef a2_rA
2337 = { .args_ct_str = { "r", "r", "0", "1", "rA", "r" } };
9b5500b6
RH
2338
2339 switch (op) {
2340 case INDEX_op_goto_ptr:
2341 return &r;
2342
2343 case INDEX_op_ld8u_i32:
2344 case INDEX_op_ld8u_i64:
2345 case INDEX_op_ld8s_i32:
2346 case INDEX_op_ld8s_i64:
2347 case INDEX_op_ld16u_i32:
2348 case INDEX_op_ld16u_i64:
2349 case INDEX_op_ld16s_i32:
2350 case INDEX_op_ld16s_i64:
2351 case INDEX_op_ld_i32:
2352 case INDEX_op_ld32u_i64:
2353 case INDEX_op_ld32s_i64:
2354 case INDEX_op_ld_i64:
2355 case INDEX_op_st8_i32:
2356 case INDEX_op_st8_i64:
2357 case INDEX_op_st16_i32:
2358 case INDEX_op_st16_i64:
2359 case INDEX_op_st_i32:
2360 case INDEX_op_st32_i64:
2361 case INDEX_op_st_i64:
2362 return &r_r;
2363
2364 case INDEX_op_add_i32:
2365 case INDEX_op_add_i64:
2366 return &r_r_ri;
2367 case INDEX_op_sub_i32:
2368 case INDEX_op_sub_i64:
bdcd5d19
RH
2369 case INDEX_op_and_i32:
2370 case INDEX_op_and_i64:
4046d9ca
RH
2371 case INDEX_op_or_i32:
2372 case INDEX_op_or_i64:
5bf67a92
RH
2373 case INDEX_op_xor_i32:
2374 case INDEX_op_xor_i64:
c2097136 2375 return (s390_facilities & FACILITY_DISTINCT_OPS ? &r_r_ri : &r_0_ri);
a8f0269e 2376
9b5500b6 2377 case INDEX_op_mul_i32:
a8f0269e
RH
2378 /* If we have the general-instruction-extensions, then we have
2379 MULTIPLY SINGLE IMMEDIATE with a signed 32-bit, otherwise we
2380 have only MULTIPLY HALFWORD IMMEDIATE, with a signed 16-bit. */
2381 return (s390_facilities & FACILITY_GEN_INST_EXT ? &r_0_ri : &r_0_rI);
9b5500b6 2382 case INDEX_op_mul_i64:
a8f0269e
RH
2383 return (s390_facilities & FACILITY_GEN_INST_EXT ? &r_0_rJ : &r_0_rI);
2384
9b5500b6
RH
2385 case INDEX_op_shl_i32:
2386 case INDEX_op_shr_i32:
2387 case INDEX_op_sar_i32:
c2097136 2388 return (s390_facilities & FACILITY_DISTINCT_OPS ? &r_r_ri : &r_0_ri);
9b5500b6
RH
2389
2390 case INDEX_op_shl_i64:
2391 case INDEX_op_shr_i64:
2392 case INDEX_op_sar_i64:
2393 return &r_r_ri;
2394
2395 case INDEX_op_rotl_i32:
2396 case INDEX_op_rotl_i64:
2397 case INDEX_op_rotr_i32:
2398 case INDEX_op_rotr_i64:
2399 return &r_r_ri;
2400
2401 case INDEX_op_brcond_i32:
2402 case INDEX_op_brcond_i64:
a534bb15 2403 return &r_ri;
9b5500b6
RH
2404
2405 case INDEX_op_bswap16_i32:
2406 case INDEX_op_bswap16_i64:
2407 case INDEX_op_bswap32_i32:
2408 case INDEX_op_bswap32_i64:
2409 case INDEX_op_bswap64_i64:
2410 case INDEX_op_neg_i32:
2411 case INDEX_op_neg_i64:
2412 case INDEX_op_ext8s_i32:
2413 case INDEX_op_ext8s_i64:
2414 case INDEX_op_ext8u_i32:
2415 case INDEX_op_ext8u_i64:
2416 case INDEX_op_ext16s_i32:
2417 case INDEX_op_ext16s_i64:
2418 case INDEX_op_ext16u_i32:
2419 case INDEX_op_ext16u_i64:
2420 case INDEX_op_ext32s_i64:
2421 case INDEX_op_ext32u_i64:
2422 case INDEX_op_ext_i32_i64:
2423 case INDEX_op_extu_i32_i64:
2424 case INDEX_op_extract_i32:
2425 case INDEX_op_extract_i64:
2426 return &r_r;
2427
2428 case INDEX_op_clz_i64:
a534bb15
RH
2429 case INDEX_op_setcond_i32:
2430 case INDEX_op_setcond_i64:
9b5500b6
RH
2431 return &r_r_ri;
2432
2433 case INDEX_op_qemu_ld_i32:
2434 case INDEX_op_qemu_ld_i64:
2435 return &r_L;
2436 case INDEX_op_qemu_st_i64:
2437 case INDEX_op_qemu_st_i32:
2438 return &L_L;
f69d277e 2439
9b5500b6
RH
2440 case INDEX_op_deposit_i32:
2441 case INDEX_op_deposit_i64:
2442 {
2443 static const TCGTargetOpDef dep
2444 = { .args_ct_str = { "r", "rZ", "r" } };
2445 return &dep;
f69d277e 2446 }
9b5500b6
RH
2447 case INDEX_op_movcond_i32:
2448 case INDEX_op_movcond_i64:
2449 {
a534bb15
RH
2450 static const TCGTargetOpDef movc
2451 = { .args_ct_str = { "r", "r", "ri", "r", "0" } };
7af525af 2452 static const TCGTargetOpDef movc_l
a534bb15
RH
2453 = { .args_ct_str = { "r", "r", "ri", "rI", "0" } };
2454 return (s390_facilities & FACILITY_LOAD_ON_COND2 ? &movc_l : &movc);
9b5500b6
RH
2455 }
2456 case INDEX_op_div2_i32:
2457 case INDEX_op_div2_i64:
2458 case INDEX_op_divu2_i32:
2459 case INDEX_op_divu2_i64:
2460 {
2461 static const TCGTargetOpDef div2
2462 = { .args_ct_str = { "b", "a", "0", "1", "r" } };
2463 return &div2;
2464 }
2465 case INDEX_op_mulu2_i64:
2466 {
2467 static const TCGTargetOpDef mul2
2468 = { .args_ct_str = { "b", "a", "0", "r" } };
2469 return &mul2;
2470 }
ba18b07d 2471
9b5500b6 2472 case INDEX_op_add2_i32:
9b5500b6 2473 case INDEX_op_sub2_i32:
ba18b07d
RH
2474 return (s390_facilities & FACILITY_EXT_IMM ? &a2_ri : &a2_r);
2475 case INDEX_op_add2_i64:
9b5500b6 2476 case INDEX_op_sub2_i64:
ba18b07d 2477 return (s390_facilities & FACILITY_EXT_IMM ? &a2_rA : &a2_r);
9b5500b6
RH
2478
2479 default:
2480 break;
f69d277e
RH
2481 }
2482 return NULL;
2483}
2484
b2c98d9d 2485static void query_s390_facilities(void)
48bb3750 2486{
c9baa30f 2487 unsigned long hwcap = qemu_getauxval(AT_HWCAP);
48bb3750 2488
c9baa30f
RH
2489 /* Is STORE FACILITY LIST EXTENDED available? Honestly, I believe this
2490 is present on all 64-bit systems, but let's check for it anyway. */
2491 if (hwcap & HWCAP_S390_STFLE) {
2492 register int r0 __asm__("0");
2493 register void *r1 __asm__("1");
48bb3750 2494
c9baa30f 2495 /* stfle 0(%r1) */
b2c98d9d 2496 r1 = &s390_facilities;
c9baa30f
RH
2497 asm volatile(".word 0xb2b0,0x1000"
2498 : "=r"(r0) : "0"(0), "r"(r1) : "memory", "cc");
48bb3750
RH
2499 }
2500}
2501
2502static void tcg_target_init(TCGContext *s)
2827822e 2503{
b2c98d9d 2504 query_s390_facilities();
48bb3750 2505
f46934df
RH
2506 tcg_target_available_regs[TCG_TYPE_I32] = 0xffff;
2507 tcg_target_available_regs[TCG_TYPE_I64] = 0xffff;
48bb3750 2508
ccb1bb66 2509 tcg_target_call_clobber_regs = 0;
48bb3750
RH
2510 tcg_regset_set_reg(tcg_target_call_clobber_regs, TCG_REG_R0);
2511 tcg_regset_set_reg(tcg_target_call_clobber_regs, TCG_REG_R1);
2512 tcg_regset_set_reg(tcg_target_call_clobber_regs, TCG_REG_R2);
2513 tcg_regset_set_reg(tcg_target_call_clobber_regs, TCG_REG_R3);
2514 tcg_regset_set_reg(tcg_target_call_clobber_regs, TCG_REG_R4);
2515 tcg_regset_set_reg(tcg_target_call_clobber_regs, TCG_REG_R5);
f24efee4
RH
2516 /* The r6 register is technically call-saved, but it's also a parameter
2517 register, so it can get killed by setup for the qemu_st helper. */
2518 tcg_regset_set_reg(tcg_target_call_clobber_regs, TCG_REG_R6);
48bb3750
RH
2519 /* The return register can be considered call-clobbered. */
2520 tcg_regset_set_reg(tcg_target_call_clobber_regs, TCG_REG_R14);
2521
ccb1bb66 2522 s->reserved_regs = 0;
48bb3750
RH
2523 tcg_regset_set_reg(s->reserved_regs, TCG_TMP0);
2524 /* XXX many insns can't be used with R0, so we better avoid it for now */
2525 tcg_regset_set_reg(s->reserved_regs, TCG_REG_R0);
2526 tcg_regset_set_reg(s->reserved_regs, TCG_REG_CALL_STACK);
829e1376
RH
2527 if (USE_REG_TB) {
2528 tcg_regset_set_reg(s->reserved_regs, TCG_REG_TB);
2529 }
2827822e
AG
2530}
2531
f167dc37
RH
2532#define FRAME_SIZE ((int)(TCG_TARGET_CALL_STACK_OFFSET \
2533 + TCG_STATIC_CALL_ARGS_SIZE \
2534 + CPU_TEMP_BUF_NLONGS * sizeof(long)))
2535
48bb3750 2536static void tcg_target_qemu_prologue(TCGContext *s)
2827822e 2537{
48bb3750
RH
2538 /* stmg %r6,%r15,48(%r15) (save registers) */
2539 tcg_out_insn(s, RXY, STMG, TCG_REG_R6, TCG_REG_R15, TCG_REG_R15, 48);
2540
a4924e8b 2541 /* aghi %r15,-frame_size */
f167dc37 2542 tcg_out_insn(s, RI, AGHI, TCG_REG_R15, -FRAME_SIZE);
a4924e8b
RH
2543
2544 tcg_set_frame(s, TCG_REG_CALL_STACK,
2545 TCG_STATIC_CALL_ARGS_SIZE + TCG_TARGET_CALL_STACK_OFFSET,
2546 CPU_TEMP_BUF_NLONGS * sizeof(long));
48bb3750 2547
090d0bfd 2548#ifndef CONFIG_SOFTMMU
b76f21a7 2549 if (guest_base >= 0x80000) {
829e1376 2550 tcg_out_movi_int(s, TCG_TYPE_PTR, TCG_GUEST_BASE_REG, guest_base, true);
48bb3750
RH
2551 tcg_regset_set_reg(s->reserved_regs, TCG_GUEST_BASE_REG);
2552 }
090d0bfd 2553#endif
48bb3750 2554
cea5f9a2 2555 tcg_out_mov(s, TCG_TYPE_PTR, TCG_AREG0, tcg_target_call_iarg_regs[0]);
829e1376
RH
2556 if (USE_REG_TB) {
2557 tcg_out_mov(s, TCG_TYPE_PTR, TCG_REG_TB,
2558 tcg_target_call_iarg_regs[1]);
2559 }
2560
cea5f9a2
BS
2561 /* br %r3 (go to TB) */
2562 tcg_out_insn(s, RR, BCR, S390_CC_ALWAYS, tcg_target_call_iarg_regs[1]);
48bb3750 2563
46644483
RH
2564 /*
2565 * Return path for goto_ptr. Set return value to 0, a-la exit_tb,
2566 * and fall through to the rest of the epilogue.
2567 */
2568 s->code_gen_epilogue = s->code_ptr;
2569 tcg_out_movi(s, TCG_TYPE_PTR, TCG_REG_R2, 0);
2570
2571 /* TB epilogue */
48bb3750
RH
2572 tb_ret_addr = s->code_ptr;
2573
a4924e8b
RH
2574 /* lmg %r6,%r15,fs+48(%r15) (restore registers) */
2575 tcg_out_insn(s, RXY, LMG, TCG_REG_R6, TCG_REG_R15, TCG_REG_R15,
f167dc37 2576 FRAME_SIZE + 48);
48bb3750
RH
2577
2578 /* br %r14 (return) */
2579 tcg_out_insn(s, RR, BCR, S390_CC_ALWAYS, TCG_REG_R14);
2827822e 2580}
f167dc37 2581
28eef8aa
RH
2582static void tcg_out_nop_fill(tcg_insn_unit *p, int count)
2583{
2584 memset(p, 0x07, count * sizeof(tcg_insn_unit));
2585}
2586
f167dc37 2587typedef struct {
d2e16f2c 2588 DebugFrameHeader h;
f167dc37
RH
2589 uint8_t fde_def_cfa[4];
2590 uint8_t fde_reg_ofs[18];
2591} DebugFrame;
2592
2593/* We're expecting a 2 byte uleb128 encoded value. */
2594QEMU_BUILD_BUG_ON(FRAME_SIZE >= (1 << 14));
2595
2596#define ELF_HOST_MACHINE EM_S390
2597
d2e16f2c
RH
2598static const DebugFrame debug_frame = {
2599 .h.cie.len = sizeof(DebugFrameCIE)-4, /* length after .len member */
2600 .h.cie.id = -1,
2601 .h.cie.version = 1,
2602 .h.cie.code_align = 1,
2603 .h.cie.data_align = 8, /* sleb128 8 */
2604 .h.cie.return_column = TCG_REG_R14,
f167dc37
RH
2605
2606 /* Total FDE size does not include the "len" member. */
d2e16f2c 2607 .h.fde.len = sizeof(DebugFrame) - offsetof(DebugFrame, h.fde.cie_offset),
f167dc37
RH
2608
2609 .fde_def_cfa = {
2610 12, TCG_REG_CALL_STACK, /* DW_CFA_def_cfa %r15, ... */
2611 (FRAME_SIZE & 0x7f) | 0x80, /* ... uleb128 FRAME_SIZE */
2612 (FRAME_SIZE >> 7)
2613 },
2614 .fde_reg_ofs = {
2615 0x86, 6, /* DW_CFA_offset, %r6, 48 */
2616 0x87, 7, /* DW_CFA_offset, %r7, 56 */
2617 0x88, 8, /* DW_CFA_offset, %r8, 64 */
2618 0x89, 9, /* DW_CFA_offset, %r92, 72 */
2619 0x8a, 10, /* DW_CFA_offset, %r10, 80 */
2620 0x8b, 11, /* DW_CFA_offset, %r11, 88 */
2621 0x8c, 12, /* DW_CFA_offset, %r12, 96 */
2622 0x8d, 13, /* DW_CFA_offset, %r13, 104 */
2623 0x8e, 14, /* DW_CFA_offset, %r14, 112 */
2624 }
2625};
2626
2627void tcg_register_jit(void *buf, size_t buf_size)
2628{
f167dc37
RH
2629 tcg_register_jit_int(buf, buf_size, &debug_frame, sizeof(debug_frame));
2630}