]> git.proxmox.com Git - mirror_qemu.git/blame - tcg/hppa/tcg-target.c
xtensa hw/: Don't use CPUState
[mirror_qemu.git] / tcg / hppa / tcg-target.c
CommitLineData
f54b3f92
AJ
1/*
2 * Tiny Code Generator for QEMU
3 *
4 * Copyright (c) 2008 Fabrice Bellard
5 *
6 * Permission is hereby granted, free of charge, to any person obtaining a copy
7 * of this software and associated documentation files (the "Software"), to deal
8 * in the Software without restriction, including without limitation the rights
9 * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
10 * copies of the Software, and to permit persons to whom the Software is
11 * furnished to do so, subject to the following conditions:
12 *
13 * The above copyright notice and this permission notice shall be included in
14 * all copies or substantial portions of the Software.
15 *
16 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
17 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
18 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
19 * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
20 * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
21 * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
22 * THE SOFTWARE.
23 */
24
d4a9eb1f 25#ifndef NDEBUG
f54b3f92 26static const char * const tcg_target_reg_names[TCG_TARGET_NB_REGS] = {
fd76e73a
RH
27 "%r0", "%r1", "%rp", "%r3", "%r4", "%r5", "%r6", "%r7",
28 "%r8", "%r9", "%r10", "%r11", "%r12", "%r13", "%r14", "%r15",
29 "%r16", "%r17", "%r18", "%r19", "%r20", "%r21", "%r22", "%r23",
30 "%r24", "%r25", "%r26", "%dp", "%ret0", "%ret1", "%sp", "%r31",
f54b3f92 31};
d4a9eb1f 32#endif
f54b3f92 33
fd76e73a
RH
34/* This is an 8 byte temp slot in the stack frame. */
35#define STACK_TEMP_OFS -16
36
fd76e73a
RH
37#ifdef CONFIG_USE_GUEST_BASE
38#define TCG_GUEST_BASE_REG TCG_REG_R16
39#else
40#define TCG_GUEST_BASE_REG TCG_REG_R0
41#endif
42
f54b3f92
AJ
43static const int tcg_target_reg_alloc_order[] = {
44 TCG_REG_R4,
45 TCG_REG_R5,
46 TCG_REG_R6,
47 TCG_REG_R7,
48 TCG_REG_R8,
49 TCG_REG_R9,
50 TCG_REG_R10,
51 TCG_REG_R11,
52 TCG_REG_R12,
53 TCG_REG_R13,
54
55 TCG_REG_R17,
56 TCG_REG_R14,
57 TCG_REG_R15,
58 TCG_REG_R16,
fd76e73a
RH
59
60 TCG_REG_R26,
61 TCG_REG_R25,
62 TCG_REG_R24,
63 TCG_REG_R23,
64
65 TCG_REG_RET0,
66 TCG_REG_RET1,
f54b3f92
AJ
67};
68
69static const int tcg_target_call_iarg_regs[4] = {
70 TCG_REG_R26,
71 TCG_REG_R25,
72 TCG_REG_R24,
73 TCG_REG_R23,
74};
75
76static const int tcg_target_call_oarg_regs[2] = {
77 TCG_REG_RET0,
78 TCG_REG_RET1,
79};
80
fd76e73a
RH
81/* True iff val fits a signed field of width BITS. */
82static inline int check_fit_tl(tcg_target_long val, unsigned int bits)
83{
84 return (val << ((sizeof(tcg_target_long) * 8 - bits))
85 >> (sizeof(tcg_target_long) * 8 - bits)) == val;
86}
87
88/* True iff depi can be used to compute (reg | MASK).
89 Accept a bit pattern like:
90 0....01....1
91 1....10....0
92 0..01..10..0
93 Copied from gcc sources. */
94static inline int or_mask_p(tcg_target_ulong mask)
95{
0085bd51
RH
96 if (mask == 0 || mask == -1) {
97 return 0;
98 }
fd76e73a
RH
99 mask += mask & -mask;
100 return (mask & (mask - 1)) == 0;
101}
102
103/* True iff depi or extru can be used to compute (reg & mask).
104 Accept a bit pattern like these:
105 0....01....1
106 1....10....0
107 1..10..01..1
108 Copied from gcc sources. */
109static inline int and_mask_p(tcg_target_ulong mask)
110{
111 return or_mask_p(~mask);
112}
113
114static int low_sign_ext(int val, int len)
115{
116 return (((val << 1) & ~(-1u << len)) | ((val >> (len - 1)) & 1));
117}
118
119static int reassemble_12(int as12)
120{
121 return (((as12 & 0x800) >> 11) |
122 ((as12 & 0x400) >> 8) |
123 ((as12 & 0x3ff) << 3));
124}
125
126static int reassemble_17(int as17)
127{
128 return (((as17 & 0x10000) >> 16) |
129 ((as17 & 0x0f800) << 5) |
130 ((as17 & 0x00400) >> 8) |
131 ((as17 & 0x003ff) << 3));
132}
133
134static int reassemble_21(int as21)
135{
136 return (((as21 & 0x100000) >> 20) |
137 ((as21 & 0x0ffe00) >> 8) |
138 ((as21 & 0x000180) << 7) |
139 ((as21 & 0x00007c) << 14) |
140 ((as21 & 0x000003) << 12));
141}
142
143/* ??? Bizzarely, there is no PCREL12F relocation type. I guess all
144 such relocations are simply fully handled by the assembler. */
145#define R_PARISC_PCREL12F R_PARISC_NONE
146
f54b3f92
AJ
147static void patch_reloc(uint8_t *code_ptr, int type,
148 tcg_target_long value, tcg_target_long addend)
149{
fd76e73a
RH
150 uint32_t *insn_ptr = (uint32_t *)code_ptr;
151 uint32_t insn = *insn_ptr;
152 tcg_target_long pcrel;
153
154 value += addend;
155 pcrel = (value - ((tcg_target_long)code_ptr + 8)) >> 2;
156
f54b3f92 157 switch (type) {
fd76e73a
RH
158 case R_PARISC_PCREL12F:
159 assert(check_fit_tl(pcrel, 12));
160 /* ??? We assume all patches are forward. See tcg_out_brcond
161 re setting the NUL bit on the branch and eliding the nop. */
162 assert(pcrel >= 0);
163 insn &= ~0x1ffdu;
164 insn |= reassemble_12(pcrel);
165 break;
f54b3f92 166 case R_PARISC_PCREL17F:
fd76e73a
RH
167 assert(check_fit_tl(pcrel, 17));
168 insn &= ~0x1f1ffdu;
169 insn |= reassemble_17(pcrel);
f54b3f92
AJ
170 break;
171 default:
172 tcg_abort();
173 }
fd76e73a
RH
174
175 *insn_ptr = insn;
f54b3f92
AJ
176}
177
178/* maximum number of register used for input function arguments */
179static inline int tcg_target_get_call_iarg_regs_count(int flags)
180{
181 return 4;
182}
183
184/* parse target specific constraints */
d4a9eb1f 185static int target_parse_constraint(TCGArgConstraint *ct, const char **pct_str)
f54b3f92
AJ
186{
187 const char *ct_str;
188
189 ct_str = *pct_str;
190 switch (ct_str[0]) {
191 case 'r':
192 ct->ct |= TCG_CT_REG;
193 tcg_regset_set32(ct->u.regs, 0, 0xffffffff);
194 break;
195 case 'L': /* qemu_ld/st constraint */
196 ct->ct |= TCG_CT_REG;
197 tcg_regset_set32(ct->u.regs, 0, 0xffffffff);
198 tcg_regset_reset_reg(ct->u.regs, TCG_REG_R26);
199 tcg_regset_reset_reg(ct->u.regs, TCG_REG_R25);
200 tcg_regset_reset_reg(ct->u.regs, TCG_REG_R24);
201 tcg_regset_reset_reg(ct->u.regs, TCG_REG_R23);
202 break;
fd76e73a
RH
203 case 'Z':
204 ct->ct |= TCG_CT_CONST_0;
205 break;
206 case 'I':
207 ct->ct |= TCG_CT_CONST_S11;
208 break;
209 case 'J':
210 ct->ct |= TCG_CT_CONST_S5;
211 break;
91493631
RH
212 case 'K':
213 ct->ct |= TCG_CT_CONST_MS11;
214 break;
0085bd51
RH
215 case 'M':
216 ct->ct |= TCG_CT_CONST_AND;
217 break;
218 case 'O':
219 ct->ct |= TCG_CT_CONST_OR;
220 break;
f54b3f92
AJ
221 default:
222 return -1;
223 }
224 ct_str++;
225 *pct_str = ct_str;
226 return 0;
227}
228
229/* test if a constant matches the constraint */
fd76e73a
RH
230static int tcg_target_const_match(tcg_target_long val,
231 const TCGArgConstraint *arg_ct)
f54b3f92 232{
fd76e73a
RH
233 int ct = arg_ct->ct;
234 if (ct & TCG_CT_CONST) {
235 return 1;
236 } else if (ct & TCG_CT_CONST_0) {
237 return val == 0;
238 } else if (ct & TCG_CT_CONST_S5) {
239 return check_fit_tl(val, 5);
240 } else if (ct & TCG_CT_CONST_S11) {
241 return check_fit_tl(val, 11);
91493631
RH
242 } else if (ct & TCG_CT_CONST_MS11) {
243 return check_fit_tl(-val, 11);
0085bd51
RH
244 } else if (ct & TCG_CT_CONST_AND) {
245 return and_mask_p(val);
246 } else if (ct & TCG_CT_CONST_OR) {
247 return or_mask_p(val);
fd76e73a 248 }
f54b3f92
AJ
249 return 0;
250}
251
252#define INSN_OP(x) ((x) << 26)
253#define INSN_EXT3BR(x) ((x) << 13)
254#define INSN_EXT3SH(x) ((x) << 10)
255#define INSN_EXT4(x) ((x) << 6)
256#define INSN_EXT5(x) (x)
257#define INSN_EXT6(x) ((x) << 6)
258#define INSN_EXT7(x) ((x) << 6)
259#define INSN_EXT8A(x) ((x) << 6)
260#define INSN_EXT8B(x) ((x) << 5)
261#define INSN_T(x) (x)
262#define INSN_R1(x) ((x) << 16)
263#define INSN_R2(x) ((x) << 21)
264#define INSN_DEP_LEN(x) (32 - (x))
265#define INSN_SHDEP_CP(x) ((31 - (x)) << 5)
266#define INSN_SHDEP_P(x) ((x) << 5)
267#define INSN_COND(x) ((x) << 13)
fd76e73a
RH
268#define INSN_IM11(x) low_sign_ext(x, 11)
269#define INSN_IM14(x) low_sign_ext(x, 14)
270#define INSN_IM5(x) (low_sign_ext(x, 5) << 16)
271
272#define COND_NEVER 0
273#define COND_EQ 1
274#define COND_LT 2
275#define COND_LE 3
276#define COND_LTU 4
277#define COND_LEU 5
278#define COND_SV 6
279#define COND_OD 7
280#define COND_FALSE 8
281
282#define INSN_ADD (INSN_OP(0x02) | INSN_EXT6(0x18))
283#define INSN_ADDC (INSN_OP(0x02) | INSN_EXT6(0x1c))
284#define INSN_ADDI (INSN_OP(0x2d))
285#define INSN_ADDIL (INSN_OP(0x0a))
286#define INSN_ADDL (INSN_OP(0x02) | INSN_EXT6(0x28))
287#define INSN_AND (INSN_OP(0x02) | INSN_EXT6(0x08))
288#define INSN_ANDCM (INSN_OP(0x02) | INSN_EXT6(0x00))
289#define INSN_COMCLR (INSN_OP(0x02) | INSN_EXT6(0x22))
290#define INSN_COMICLR (INSN_OP(0x24))
291#define INSN_DEP (INSN_OP(0x35) | INSN_EXT3SH(3))
292#define INSN_DEPI (INSN_OP(0x35) | INSN_EXT3SH(7))
293#define INSN_EXTRS (INSN_OP(0x34) | INSN_EXT3SH(7))
294#define INSN_EXTRU (INSN_OP(0x34) | INSN_EXT3SH(6))
295#define INSN_LDIL (INSN_OP(0x08))
296#define INSN_LDO (INSN_OP(0x0d))
297#define INSN_MTCTL (INSN_OP(0x00) | INSN_EXT8B(0xc2))
298#define INSN_OR (INSN_OP(0x02) | INSN_EXT6(0x09))
299#define INSN_SHD (INSN_OP(0x34) | INSN_EXT3SH(2))
300#define INSN_SUB (INSN_OP(0x02) | INSN_EXT6(0x10))
301#define INSN_SUBB (INSN_OP(0x02) | INSN_EXT6(0x14))
302#define INSN_SUBI (INSN_OP(0x25))
303#define INSN_VEXTRS (INSN_OP(0x34) | INSN_EXT3SH(5))
304#define INSN_VEXTRU (INSN_OP(0x34) | INSN_EXT3SH(4))
305#define INSN_VSHD (INSN_OP(0x34) | INSN_EXT3SH(0))
306#define INSN_XOR (INSN_OP(0x02) | INSN_EXT6(0x0a))
307#define INSN_ZDEP (INSN_OP(0x35) | INSN_EXT3SH(2))
308#define INSN_ZVDEP (INSN_OP(0x35) | INSN_EXT3SH(0))
309
310#define INSN_BL (INSN_OP(0x3a) | INSN_EXT3BR(0))
311#define INSN_BL_N (INSN_OP(0x3a) | INSN_EXT3BR(0) | 2)
312#define INSN_BLR (INSN_OP(0x3a) | INSN_EXT3BR(2))
313#define INSN_BV (INSN_OP(0x3a) | INSN_EXT3BR(6))
314#define INSN_BV_N (INSN_OP(0x3a) | INSN_EXT3BR(6) | 2)
315#define INSN_BLE_SR4 (INSN_OP(0x39) | (1 << 13))
316
317#define INSN_LDB (INSN_OP(0x10))
318#define INSN_LDH (INSN_OP(0x11))
319#define INSN_LDW (INSN_OP(0x12))
320#define INSN_LDWM (INSN_OP(0x13))
321#define INSN_FLDDS (INSN_OP(0x0b) | INSN_EXT4(0) | (1 << 12))
322
323#define INSN_LDBX (INSN_OP(0x03) | INSN_EXT4(0))
324#define INSN_LDHX (INSN_OP(0x03) | INSN_EXT4(1))
325#define INSN_LDWX (INSN_OP(0x03) | INSN_EXT4(2))
326
327#define INSN_STB (INSN_OP(0x18))
328#define INSN_STH (INSN_OP(0x19))
329#define INSN_STW (INSN_OP(0x1a))
330#define INSN_STWM (INSN_OP(0x1b))
331#define INSN_FSTDS (INSN_OP(0x0b) | INSN_EXT4(8) | (1 << 12))
332
333#define INSN_COMBT (INSN_OP(0x20))
334#define INSN_COMBF (INSN_OP(0x22))
335#define INSN_COMIBT (INSN_OP(0x21))
336#define INSN_COMIBF (INSN_OP(0x23))
337
338/* supplied by libgcc */
339extern void *__canonicalize_funcptr_for_compare(void *);
340
2a534aff 341static void tcg_out_mov(TCGContext *s, TCGType type, TCGReg ret, TCGReg arg)
fd76e73a
RH
342{
343 /* PA1.1 defines COPY as OR r,0,t; PA2.0 defines COPY as LDO 0(r),t
344 but hppa-dis.c is unaware of this definition */
345 if (ret != arg) {
346 tcg_out32(s, INSN_OR | INSN_T(ret) | INSN_R1(arg)
347 | INSN_R2(TCG_REG_R0));
348 }
349}
f54b3f92 350
fd76e73a 351static void tcg_out_movi(TCGContext *s, TCGType type,
2a534aff 352 TCGReg ret, tcg_target_long arg)
fd76e73a
RH
353{
354 if (check_fit_tl(arg, 14)) {
355 tcg_out32(s, INSN_LDO | INSN_R1(ret)
356 | INSN_R2(TCG_REG_R0) | INSN_IM14(arg));
357 } else {
358 uint32_t hi, lo;
359 hi = arg >> 11;
360 lo = arg & 0x7ff;
361
362 tcg_out32(s, INSN_LDIL | INSN_R2(ret) | reassemble_21(hi));
363 if (lo) {
364 tcg_out32(s, INSN_LDO | INSN_R1(ret)
365 | INSN_R2(ret) | INSN_IM14(lo));
366 }
367 }
368}
f54b3f92 369
fd76e73a
RH
370static void tcg_out_ldst(TCGContext *s, int ret, int addr,
371 tcg_target_long offset, int op)
372{
373 if (!check_fit_tl(offset, 14)) {
374 uint32_t hi, lo, op;
f54b3f92 375
fd76e73a
RH
376 hi = offset >> 11;
377 lo = offset & 0x7ff;
f54b3f92 378
fd76e73a
RH
379 if (addr == TCG_REG_R0) {
380 op = INSN_LDIL | INSN_R2(TCG_REG_R1);
381 } else {
382 op = INSN_ADDIL | INSN_R2(addr);
383 }
384 tcg_out32(s, op | reassemble_21(hi));
f54b3f92 385
fd76e73a
RH
386 addr = TCG_REG_R1;
387 offset = lo;
388 }
f54b3f92 389
fd76e73a
RH
390 if (ret != addr || offset != 0 || op != INSN_LDO) {
391 tcg_out32(s, op | INSN_R1(ret) | INSN_R2(addr) | INSN_IM14(offset));
392 }
393}
f54b3f92 394
fd76e73a 395/* This function is required by tcg.c. */
2a534aff
RH
396static inline void tcg_out_ld(TCGContext *s, TCGType type, TCGReg ret,
397 TCGReg arg1, tcg_target_long arg2)
fd76e73a
RH
398{
399 tcg_out_ldst(s, ret, arg1, arg2, INSN_LDW);
400}
401
402/* This function is required by tcg.c. */
2a534aff
RH
403static inline void tcg_out_st(TCGContext *s, TCGType type, TCGReg ret,
404 TCGReg arg1, tcg_target_long arg2)
fd76e73a
RH
405{
406 tcg_out_ldst(s, ret, arg1, arg2, INSN_STW);
407}
408
409static void tcg_out_ldst_index(TCGContext *s, int data,
410 int base, int index, int op)
411{
412 tcg_out32(s, op | INSN_T(data) | INSN_R1(index) | INSN_R2(base));
413}
414
415static inline void tcg_out_addi2(TCGContext *s, int ret, int arg1,
416 tcg_target_long val)
417{
418 tcg_out_ldst(s, ret, arg1, val, INSN_LDO);
419}
f54b3f92 420
fd76e73a
RH
421/* This function is required by tcg.c. */
422static inline void tcg_out_addi(TCGContext *s, int reg, tcg_target_long val)
423{
424 tcg_out_addi2(s, reg, reg, val);
425}
f54b3f92 426
fd76e73a
RH
427static inline void tcg_out_arith(TCGContext *s, int t, int r1, int r2, int op)
428{
429 tcg_out32(s, op | INSN_T(t) | INSN_R1(r1) | INSN_R2(r2));
430}
f54b3f92 431
fd76e73a
RH
432static inline void tcg_out_arithi(TCGContext *s, int t, int r1,
433 tcg_target_long val, int op)
f54b3f92 434{
fd76e73a
RH
435 assert(check_fit_tl(val, 11));
436 tcg_out32(s, op | INSN_R1(t) | INSN_R2(r1) | INSN_IM11(val));
f54b3f92
AJ
437}
438
fd76e73a 439static inline void tcg_out_nop(TCGContext *s)
f54b3f92 440{
fd76e73a
RH
441 tcg_out_arith(s, TCG_REG_R0, TCG_REG_R0, TCG_REG_R0, INSN_OR);
442}
f54b3f92 443
fd76e73a
RH
444static inline void tcg_out_mtctl_sar(TCGContext *s, int arg)
445{
446 tcg_out32(s, INSN_MTCTL | INSN_R2(11) | INSN_R1(arg));
447}
448
449/* Extract LEN bits at position OFS from ARG and place in RET.
450 Note that here the bit ordering is reversed from the PA-RISC
451 standard, such that the right-most bit is 0. */
452static inline void tcg_out_extr(TCGContext *s, int ret, int arg,
453 unsigned ofs, unsigned len, int sign)
454{
455 assert(ofs < 32 && len <= 32 - ofs);
456 tcg_out32(s, (sign ? INSN_EXTRS : INSN_EXTRU)
457 | INSN_R1(ret) | INSN_R2(arg)
458 | INSN_SHDEP_P(31 - ofs) | INSN_DEP_LEN(len));
f54b3f92
AJ
459}
460
fd76e73a
RH
461/* Likewise with OFS interpreted little-endian. */
462static inline void tcg_out_dep(TCGContext *s, int ret, int arg,
463 unsigned ofs, unsigned len)
f54b3f92 464{
fd76e73a
RH
465 assert(ofs < 32 && len <= 32 - ofs);
466 tcg_out32(s, INSN_DEP | INSN_R2(ret) | INSN_R1(arg)
467 | INSN_SHDEP_CP(31 - ofs) | INSN_DEP_LEN(len));
468}
469
ec188429
RH
470static inline void tcg_out_depi(TCGContext *s, int ret, int arg,
471 unsigned ofs, unsigned len)
472{
473 assert(ofs < 32 && len <= 32 - ofs);
474 tcg_out32(s, INSN_DEPI | INSN_R2(ret) | INSN_IM5(arg)
475 | INSN_SHDEP_CP(31 - ofs) | INSN_DEP_LEN(len));
476}
477
fd76e73a
RH
478static inline void tcg_out_shd(TCGContext *s, int ret, int hi, int lo,
479 unsigned count)
480{
481 assert(count < 32);
482 tcg_out32(s, INSN_SHD | INSN_R1(hi) | INSN_R2(lo) | INSN_T(ret)
483 | INSN_SHDEP_CP(count));
484}
485
486static void tcg_out_vshd(TCGContext *s, int ret, int hi, int lo, int creg)
487{
488 tcg_out_mtctl_sar(s, creg);
489 tcg_out32(s, INSN_VSHD | INSN_T(ret) | INSN_R1(hi) | INSN_R2(lo));
490}
491
492static void tcg_out_ori(TCGContext *s, int ret, int arg, tcg_target_ulong m)
493{
0085bd51
RH
494 int bs0, bs1;
495
496 /* Note that the argument is constrained to match or_mask_p. */
497 for (bs0 = 0; bs0 < 32; bs0++) {
498 if ((m & (1u << bs0)) != 0) {
499 break;
fd76e73a 500 }
0085bd51
RH
501 }
502 for (bs1 = bs0; bs1 < 32; bs1++) {
503 if ((m & (1u << bs1)) == 0) {
504 break;
fd76e73a 505 }
fd76e73a 506 }
0085bd51
RH
507 assert(bs1 == 32 || (1ul << bs1) > m);
508
3b6dac34 509 tcg_out_mov(s, TCG_TYPE_I32, ret, arg);
ec188429 510 tcg_out_depi(s, ret, -1, bs0, bs1 - bs0);
fd76e73a
RH
511}
512
513static void tcg_out_andi(TCGContext *s, int ret, int arg, tcg_target_ulong m)
514{
0085bd51 515 int ls0, ls1, ms0;
fd76e73a 516
0085bd51
RH
517 /* Note that the argument is constrained to match and_mask_p. */
518 for (ls0 = 0; ls0 < 32; ls0++) {
519 if ((m & (1u << ls0)) == 0) {
520 break;
fd76e73a 521 }
0085bd51
RH
522 }
523 for (ls1 = ls0; ls1 < 32; ls1++) {
524 if ((m & (1u << ls1)) != 0) {
525 break;
fd76e73a 526 }
0085bd51
RH
527 }
528 for (ms0 = ls1; ms0 < 32; ms0++) {
529 if ((m & (1u << ms0)) == 0) {
530 break;
fd76e73a 531 }
0085bd51
RH
532 }
533 assert (ms0 == 32);
fd76e73a 534
0085bd51
RH
535 if (ls1 == 32) {
536 tcg_out_extr(s, ret, arg, 0, ls0, 0);
f54b3f92 537 } else {
3b6dac34 538 tcg_out_mov(s, TCG_TYPE_I32, ret, arg);
ec188429 539 tcg_out_depi(s, ret, 0, ls0, ls1 - ls0);
f54b3f92
AJ
540 }
541}
542
fd76e73a 543static inline void tcg_out_ext8s(TCGContext *s, int ret, int arg)
f54b3f92 544{
fd76e73a 545 tcg_out_extr(s, ret, arg, 0, 8, 1);
f54b3f92
AJ
546}
547
fd76e73a 548static inline void tcg_out_ext16s(TCGContext *s, int ret, int arg)
f54b3f92 549{
fd76e73a 550 tcg_out_extr(s, ret, arg, 0, 16, 1);
f54b3f92
AJ
551}
552
fd76e73a 553static void tcg_out_shli(TCGContext *s, int ret, int arg, int count)
f54b3f92 554{
fd76e73a
RH
555 count &= 31;
556 tcg_out32(s, INSN_ZDEP | INSN_R2(ret) | INSN_R1(arg)
557 | INSN_SHDEP_CP(31 - count) | INSN_DEP_LEN(32 - count));
f54b3f92
AJ
558}
559
fd76e73a 560static void tcg_out_shl(TCGContext *s, int ret, int arg, int creg)
f54b3f92 561{
fd76e73a
RH
562 tcg_out_arithi(s, TCG_REG_R20, creg, 31, INSN_SUBI);
563 tcg_out_mtctl_sar(s, TCG_REG_R20);
564 tcg_out32(s, INSN_ZVDEP | INSN_R2(ret) | INSN_R1(arg) | INSN_DEP_LEN(32));
f54b3f92
AJ
565}
566
fd76e73a 567static void tcg_out_shri(TCGContext *s, int ret, int arg, int count)
f54b3f92 568{
fd76e73a
RH
569 count &= 31;
570 tcg_out_extr(s, ret, arg, count, 32 - count, 0);
f54b3f92
AJ
571}
572
fd76e73a 573static void tcg_out_shr(TCGContext *s, int ret, int arg, int creg)
f54b3f92 574{
fd76e73a 575 tcg_out_vshd(s, ret, TCG_REG_R0, arg, creg);
f54b3f92
AJ
576}
577
fd76e73a 578static void tcg_out_sari(TCGContext *s, int ret, int arg, int count)
f54b3f92 579{
fd76e73a
RH
580 count &= 31;
581 tcg_out_extr(s, ret, arg, count, 32 - count, 1);
f54b3f92
AJ
582}
583
fd76e73a 584static void tcg_out_sar(TCGContext *s, int ret, int arg, int creg)
f54b3f92 585{
fd76e73a
RH
586 tcg_out_arithi(s, TCG_REG_R20, creg, 31, INSN_SUBI);
587 tcg_out_mtctl_sar(s, TCG_REG_R20);
588 tcg_out32(s, INSN_VEXTRS | INSN_R1(ret) | INSN_R2(arg) | INSN_DEP_LEN(32));
f54b3f92
AJ
589}
590
fd76e73a 591static void tcg_out_rotli(TCGContext *s, int ret, int arg, int count)
f54b3f92 592{
fd76e73a
RH
593 count &= 31;
594 tcg_out_shd(s, ret, arg, arg, 32 - count);
f54b3f92
AJ
595}
596
fd76e73a
RH
597static void tcg_out_rotl(TCGContext *s, int ret, int arg, int creg)
598{
599 tcg_out_arithi(s, TCG_REG_R20, creg, 32, INSN_SUBI);
600 tcg_out_vshd(s, ret, arg, arg, TCG_REG_R20);
f54b3f92
AJ
601}
602
fd76e73a
RH
603static void tcg_out_rotri(TCGContext *s, int ret, int arg, int count)
604{
605 count &= 31;
606 tcg_out_shd(s, ret, arg, arg, count);
f54b3f92
AJ
607}
608
fd76e73a
RH
609static void tcg_out_rotr(TCGContext *s, int ret, int arg, int creg)
610{
611 tcg_out_vshd(s, ret, arg, arg, creg);
f54b3f92
AJ
612}
613
fd76e73a
RH
614static void tcg_out_bswap16(TCGContext *s, int ret, int arg, int sign)
615{
616 if (ret != arg) {
3b6dac34 617 tcg_out_mov(s, TCG_TYPE_I32, ret, arg); /* arg = xxAB */
fd76e73a
RH
618 }
619 tcg_out_dep(s, ret, ret, 16, 8); /* ret = xBAB */
620 tcg_out_extr(s, ret, ret, 8, 16, sign); /* ret = ..BA */
f54b3f92
AJ
621}
622
fd76e73a 623static void tcg_out_bswap32(TCGContext *s, int ret, int arg, int temp)
f54b3f92 624{
fd76e73a
RH
625 /* arg = ABCD */
626 tcg_out_rotri(s, temp, arg, 16); /* temp = CDAB */
627 tcg_out_dep(s, temp, temp, 16, 8); /* temp = CBAB */
628 tcg_out_shd(s, ret, arg, temp, 8); /* ret = DCBA */
f54b3f92
AJ
629}
630
fd76e73a
RH
631static void tcg_out_call(TCGContext *s, void *func)
632{
633 tcg_target_long val, hi, lo, disp;
634
635 val = (uint32_t)__canonicalize_funcptr_for_compare(func);
636 disp = (val - ((tcg_target_long)s->code_ptr + 8)) >> 2;
637
638 if (check_fit_tl(disp, 17)) {
639 tcg_out32(s, INSN_BL_N | INSN_R2(TCG_REG_RP) | reassemble_17(disp));
640 } else {
641 hi = val >> 11;
642 lo = val & 0x7ff;
643
644 tcg_out32(s, INSN_LDIL | INSN_R2(TCG_REG_R20) | reassemble_21(hi));
645 tcg_out32(s, INSN_BLE_SR4 | INSN_R2(TCG_REG_R20)
646 | reassemble_17(lo >> 2));
3b6dac34 647 tcg_out_mov(s, TCG_TYPE_I32, TCG_REG_RP, TCG_REG_R31);
fd76e73a
RH
648 }
649}
79383c9c 650
fd76e73a
RH
651static void tcg_out_xmpyu(TCGContext *s, int retl, int reth,
652 int arg1, int arg2)
653{
654 /* Store both words into the stack for copy to the FPU. */
a42bceec
BS
655 tcg_out_ldst(s, arg1, TCG_REG_CALL_STACK, STACK_TEMP_OFS, INSN_STW);
656 tcg_out_ldst(s, arg2, TCG_REG_CALL_STACK, STACK_TEMP_OFS + 4, INSN_STW);
fd76e73a
RH
657
658 /* Load both words into the FPU at the same time. We get away
659 with this because we can address the left and right half of the
660 FPU registers individually once loaded. */
661 /* fldds stack_temp(sp),fr22 */
a42bceec 662 tcg_out32(s, INSN_FLDDS | INSN_R2(TCG_REG_CALL_STACK)
fd76e73a
RH
663 | INSN_IM5(STACK_TEMP_OFS) | INSN_T(22));
664
665 /* xmpyu fr22r,fr22,fr22 */
666 tcg_out32(s, 0x3ad64796);
667
668 /* Store the 64-bit result back into the stack. */
669 /* fstds stack_temp(sp),fr22 */
a42bceec 670 tcg_out32(s, INSN_FSTDS | INSN_R2(TCG_REG_CALL_STACK)
fd76e73a
RH
671 | INSN_IM5(STACK_TEMP_OFS) | INSN_T(22));
672
673 /* Load the pieces of the result that the caller requested. */
674 if (reth) {
a42bceec 675 tcg_out_ldst(s, reth, TCG_REG_CALL_STACK, STACK_TEMP_OFS, INSN_LDW);
fd76e73a
RH
676 }
677 if (retl) {
a42bceec
BS
678 tcg_out_ldst(s, retl, TCG_REG_CALL_STACK, STACK_TEMP_OFS + 4,
679 INSN_LDW);
fd76e73a
RH
680 }
681}
682
91493631
RH
683static void tcg_out_add2(TCGContext *s, int destl, int desth,
684 int al, int ah, int bl, int bh, int blconst)
685{
686 int tmp = (destl == ah || destl == bh ? TCG_REG_R20 : destl);
687
688 if (blconst) {
689 tcg_out_arithi(s, tmp, al, bl, INSN_ADDI);
690 } else {
691 tcg_out_arith(s, tmp, al, bl, INSN_ADD);
692 }
693 tcg_out_arith(s, desth, ah, bh, INSN_ADDC);
694
3b6dac34 695 tcg_out_mov(s, TCG_TYPE_I32, destl, tmp);
91493631
RH
696}
697
698static void tcg_out_sub2(TCGContext *s, int destl, int desth, int al, int ah,
699 int bl, int bh, int alconst, int blconst)
700{
701 int tmp = (destl == ah || destl == bh ? TCG_REG_R20 : destl);
702
703 if (alconst) {
704 if (blconst) {
705 tcg_out_movi(s, TCG_TYPE_I32, TCG_REG_R20, bl);
706 bl = TCG_REG_R20;
707 }
708 tcg_out_arithi(s, tmp, bl, al, INSN_SUBI);
709 } else if (blconst) {
710 tcg_out_arithi(s, tmp, al, -bl, INSN_ADDI);
711 } else {
712 tcg_out_arith(s, tmp, al, bl, INSN_SUB);
713 }
714 tcg_out_arith(s, desth, ah, bh, INSN_SUBB);
715
3b6dac34 716 tcg_out_mov(s, TCG_TYPE_I32, destl, tmp);
91493631
RH
717}
718
fd76e73a
RH
719static void tcg_out_branch(TCGContext *s, int label_index, int nul)
720{
721 TCGLabel *l = &s->labels[label_index];
722 uint32_t op = nul ? INSN_BL_N : INSN_BL;
723
724 if (l->has_value) {
725 tcg_target_long val = l->u.value;
726
727 val -= (tcg_target_long)s->code_ptr + 8;
728 val >>= 2;
729 assert(check_fit_tl(val, 17));
730
731 tcg_out32(s, op | reassemble_17(val));
732 } else {
2d097a83
RH
733 /* We need to keep the offset unchanged for retranslation. */
734 uint32_t old_insn = *(uint32_t *)s->code_ptr;
735
fd76e73a 736 tcg_out_reloc(s, s->code_ptr, R_PARISC_PCREL17F, label_index, 0);
2d097a83 737 tcg_out32(s, op | (old_insn & 0x1f1ffdu));
fd76e73a
RH
738 }
739}
740
741static const uint8_t tcg_cond_to_cmp_cond[10] =
742{
743 [TCG_COND_EQ] = COND_EQ,
744 [TCG_COND_NE] = COND_EQ | COND_FALSE,
745 [TCG_COND_LT] = COND_LT,
746 [TCG_COND_GE] = COND_LT | COND_FALSE,
747 [TCG_COND_LE] = COND_LE,
748 [TCG_COND_GT] = COND_LE | COND_FALSE,
749 [TCG_COND_LTU] = COND_LTU,
750 [TCG_COND_GEU] = COND_LTU | COND_FALSE,
751 [TCG_COND_LEU] = COND_LEU,
752 [TCG_COND_GTU] = COND_LEU | COND_FALSE,
753};
754
755static void tcg_out_brcond(TCGContext *s, int cond, TCGArg c1,
756 TCGArg c2, int c2const, int label_index)
757{
758 TCGLabel *l = &s->labels[label_index];
759 int op, pacond;
760
761 /* Note that COMIB operates as if the immediate is the first
762 operand. We model brcond with the immediate in the second
763 to better match what targets are likely to give us. For
764 consistency, model COMB with reversed operands as well. */
765 pacond = tcg_cond_to_cmp_cond[tcg_swap_cond(cond)];
766
767 if (c2const) {
768 op = (pacond & COND_FALSE ? INSN_COMIBF : INSN_COMIBT);
769 op |= INSN_IM5(c2);
770 } else {
771 op = (pacond & COND_FALSE ? INSN_COMBF : INSN_COMBT);
772 op |= INSN_R1(c2);
773 }
774 op |= INSN_R2(c1);
775 op |= INSN_COND(pacond & 7);
776
777 if (l->has_value) {
778 tcg_target_long val = l->u.value;
779
780 val -= (tcg_target_long)s->code_ptr + 8;
781 val >>= 2;
782 assert(check_fit_tl(val, 12));
783
784 /* ??? Assume that all branches to defined labels are backward.
785 Which means that if the nul bit is set, the delay slot is
786 executed if the branch is taken, and not executed in fallthru. */
787 tcg_out32(s, op | reassemble_12(val));
788 tcg_out_nop(s);
789 } else {
2d097a83
RH
790 /* We need to keep the offset unchanged for retranslation. */
791 uint32_t old_insn = *(uint32_t *)s->code_ptr;
792
fd76e73a
RH
793 tcg_out_reloc(s, s->code_ptr, R_PARISC_PCREL12F, label_index, 0);
794 /* ??? Assume that all branches to undefined labels are forward.
795 Which means that if the nul bit is set, the delay slot is
796 not executed if the branch is taken, which is what we want. */
2d097a83 797 tcg_out32(s, op | 2 | (old_insn & 0x1ffdu));
fd76e73a
RH
798 }
799}
800
801static void tcg_out_comclr(TCGContext *s, int cond, TCGArg ret,
802 TCGArg c1, TCGArg c2, int c2const)
803{
804 int op, pacond;
805
806 /* Note that COMICLR operates as if the immediate is the first
807 operand. We model setcond with the immediate in the second
808 to better match what targets are likely to give us. For
809 consistency, model COMCLR with reversed operands as well. */
810 pacond = tcg_cond_to_cmp_cond[tcg_swap_cond(cond)];
811
812 if (c2const) {
813 op = INSN_COMICLR | INSN_R2(c1) | INSN_R1(ret) | INSN_IM11(c2);
814 } else {
815 op = INSN_COMCLR | INSN_R2(c1) | INSN_R1(c2) | INSN_T(ret);
816 }
817 op |= INSN_COND(pacond & 7);
818 op |= pacond & COND_FALSE ? 1 << 12 : 0;
819
820 tcg_out32(s, op);
821}
822
823static void tcg_out_brcond2(TCGContext *s, int cond, TCGArg al, TCGArg ah,
824 TCGArg bl, int blconst, TCGArg bh, int bhconst,
825 int label_index)
826{
827 switch (cond) {
828 case TCG_COND_EQ:
829 case TCG_COND_NE:
830 tcg_out_comclr(s, tcg_invert_cond(cond), TCG_REG_R0, al, bl, blconst);
831 tcg_out_brcond(s, cond, ah, bh, bhconst, label_index);
832 break;
833
834 default:
835 tcg_out_brcond(s, cond, ah, bh, bhconst, label_index);
836 tcg_out_comclr(s, TCG_COND_NE, TCG_REG_R0, ah, bh, bhconst);
837 tcg_out_brcond(s, tcg_unsigned_cond(cond),
838 al, bl, blconst, label_index);
839 break;
840 }
841}
842
843static void tcg_out_setcond(TCGContext *s, int cond, TCGArg ret,
844 TCGArg c1, TCGArg c2, int c2const)
845{
846 tcg_out_comclr(s, tcg_invert_cond(cond), ret, c1, c2, c2const);
847 tcg_out_movi(s, TCG_TYPE_I32, ret, 1);
848}
849
850static void tcg_out_setcond2(TCGContext *s, int cond, TCGArg ret,
851 TCGArg al, TCGArg ah, TCGArg bl, int blconst,
852 TCGArg bh, int bhconst)
853{
854 int scratch = TCG_REG_R20;
855
856 if (ret != al && ret != ah
857 && (blconst || ret != bl)
858 && (bhconst || ret != bh)) {
859 scratch = ret;
860 }
861
862 switch (cond) {
863 case TCG_COND_EQ:
864 case TCG_COND_NE:
865 tcg_out_setcond(s, cond, scratch, al, bl, blconst);
866 tcg_out_comclr(s, TCG_COND_EQ, TCG_REG_R0, ah, bh, bhconst);
867 tcg_out_movi(s, TCG_TYPE_I32, scratch, cond == TCG_COND_NE);
868 break;
869
870 default:
871 tcg_out_setcond(s, tcg_unsigned_cond(cond), scratch, al, bl, blconst);
872 tcg_out_comclr(s, TCG_COND_EQ, TCG_REG_R0, ah, bh, bhconst);
873 tcg_out_movi(s, TCG_TYPE_I32, scratch, 0);
874 tcg_out_comclr(s, cond, TCG_REG_R0, ah, bh, bhconst);
875 tcg_out_movi(s, TCG_TYPE_I32, scratch, 1);
876 break;
877 }
878
3b6dac34 879 tcg_out_mov(s, TCG_TYPE_I32, ret, scratch);
fd76e73a
RH
880}
881
882#if defined(CONFIG_SOFTMMU)
79383c9c 883#include "../../softmmu_defs.h"
f54b3f92
AJ
884
885static void *qemu_ld_helpers[4] = {
886 __ldb_mmu,
887 __ldw_mmu,
888 __ldl_mmu,
889 __ldq_mmu,
890};
891
892static void *qemu_st_helpers[4] = {
893 __stb_mmu,
894 __stw_mmu,
895 __stl_mmu,
896 __stq_mmu,
897};
fd76e73a
RH
898
899/* Load and compare a TLB entry, and branch if TLB miss. OFFSET is set to
900 the offset of the first ADDR_READ or ADDR_WRITE member of the appropriate
901 TLB for the memory index. The return value is the offset from ENV
902 contained in R1 afterward (to be used when loading ADDEND); if the
903 return value is 0, R1 is not used. */
904
905static int tcg_out_tlb_read(TCGContext *s, int r0, int r1, int addrlo,
906 int addrhi, int s_bits, int lab_miss, int offset)
907{
908 int ret;
909
910 /* Extracting the index into the TLB. The "normal C operation" is
911 r1 = addr_reg >> TARGET_PAGE_BITS;
912 r1 &= CPU_TLB_SIZE - 1;
913 r1 <<= CPU_TLB_ENTRY_BITS;
914 What this does is extract CPU_TLB_BITS beginning at TARGET_PAGE_BITS
915 and place them at CPU_TLB_ENTRY_BITS. We can combine the first two
916 operations with an EXTRU. Unfortunately, the current value of
917 CPU_TLB_ENTRY_BITS is > 3, so we can't merge that shift with the
918 add that follows. */
919 tcg_out_extr(s, r1, addrlo, TARGET_PAGE_BITS, CPU_TLB_BITS, 0);
fd76e73a
RH
920 tcg_out_shli(s, r1, r1, CPU_TLB_ENTRY_BITS);
921 tcg_out_arith(s, r1, r1, TCG_AREG0, INSN_ADDL);
922
923 /* Make sure that both the addr_{read,write} and addend can be
924 read with a 14-bit offset from the same base register. */
925 if (check_fit_tl(offset + CPU_TLB_SIZE, 14)) {
926 ret = 0;
927 } else {
928 ret = (offset + 0x400) & ~0x7ff;
929 offset = ret - offset;
930 tcg_out_addi2(s, TCG_REG_R1, r1, ret);
931 r1 = TCG_REG_R1;
932 }
933
934 /* Load the entry from the computed slot. */
935 if (TARGET_LONG_BITS == 64) {
936 tcg_out_ld(s, TCG_TYPE_PTR, TCG_REG_R23, r1, offset);
937 tcg_out_ld(s, TCG_TYPE_PTR, TCG_REG_R20, r1, offset + 4);
938 } else {
939 tcg_out_ld(s, TCG_TYPE_PTR, TCG_REG_R20, r1, offset);
940 }
941
739734cb
RH
942 /* Compute the value that ought to appear in the TLB for a hit, namely, the page
943 of the address. We include the low N bits of the address to catch unaligned
944 accesses and force them onto the slow path. Do this computation after having
945 issued the load from the TLB slot to give the load time to complete. */
946 tcg_out_andi(s, r0, addrlo, TARGET_PAGE_MASK | ((1 << s_bits) - 1));
947
fd76e73a
RH
948 /* If not equal, jump to lab_miss. */
949 if (TARGET_LONG_BITS == 64) {
950 tcg_out_brcond2(s, TCG_COND_NE, TCG_REG_R20, TCG_REG_R23,
951 r0, 0, addrhi, 0, lab_miss);
952 } else {
953 tcg_out_brcond(s, TCG_COND_NE, TCG_REG_R20, r0, 0, lab_miss);
954 }
955
956 return ret;
957}
f54b3f92
AJ
958#endif
959
f061b40e
RH
960static void tcg_out_qemu_ld_direct(TCGContext *s, int datalo_reg, int datahi_reg,
961 int addr_reg, int addend_reg, int opc)
f54b3f92 962{
f54b3f92 963#ifdef TARGET_WORDS_BIGENDIAN
f061b40e 964 const int bswap = 0;
f54b3f92 965#else
f061b40e 966 const int bswap = 1;
f54b3f92 967#endif
f061b40e 968
f54b3f92 969 switch (opc) {
fd76e73a 970 case 0:
f061b40e 971 tcg_out_ldst_index(s, datalo_reg, addr_reg, addend_reg, INSN_LDBX);
fd76e73a
RH
972 break;
973 case 0 | 4:
f061b40e
RH
974 tcg_out_ldst_index(s, datalo_reg, addr_reg, addend_reg, INSN_LDBX);
975 tcg_out_ext8s(s, datalo_reg, datalo_reg);
fd76e73a
RH
976 break;
977 case 1:
f061b40e 978 tcg_out_ldst_index(s, datalo_reg, addr_reg, addend_reg, INSN_LDHX);
fd76e73a 979 if (bswap) {
f061b40e 980 tcg_out_bswap16(s, datalo_reg, datalo_reg, 0);
fd76e73a
RH
981 }
982 break;
983 case 1 | 4:
f061b40e 984 tcg_out_ldst_index(s, datalo_reg, addr_reg, addend_reg, INSN_LDHX);
fd76e73a 985 if (bswap) {
f061b40e 986 tcg_out_bswap16(s, datalo_reg, datalo_reg, 1);
fd76e73a 987 } else {
f061b40e 988 tcg_out_ext16s(s, datalo_reg, datalo_reg);
fd76e73a
RH
989 }
990 break;
991 case 2:
f061b40e 992 tcg_out_ldst_index(s, datalo_reg, addr_reg, addend_reg, INSN_LDWX);
fd76e73a 993 if (bswap) {
f061b40e 994 tcg_out_bswap32(s, datalo_reg, datalo_reg, TCG_REG_R20);
fd76e73a
RH
995 }
996 break;
997 case 3:
998 if (bswap) {
f061b40e
RH
999 int t = datahi_reg;
1000 datahi_reg = datalo_reg;
1001 datalo_reg = t;
fd76e73a 1002 }
f061b40e
RH
1003 /* We can't access the low-part with a reg+reg addressing mode,
1004 so perform the addition now and use reg_ofs addressing mode. */
1005 if (addend_reg != TCG_REG_R0) {
1006 tcg_out_arith(s, TCG_REG_R20, addr_reg, addend_reg, INSN_ADD);
1007 addr_reg = TCG_REG_R20;
1008 }
1009 /* Make sure not to clobber the base register. */
1010 if (datahi_reg == addr_reg) {
1011 tcg_out_ldst(s, datalo_reg, addr_reg, 4, INSN_LDW);
1012 tcg_out_ldst(s, datahi_reg, addr_reg, 0, INSN_LDW);
fd76e73a 1013 } else {
f061b40e
RH
1014 tcg_out_ldst(s, datahi_reg, addr_reg, 0, INSN_LDW);
1015 tcg_out_ldst(s, datalo_reg, addr_reg, 4, INSN_LDW);
fd76e73a
RH
1016 }
1017 if (bswap) {
f061b40e
RH
1018 tcg_out_bswap32(s, datalo_reg, datalo_reg, TCG_REG_R20);
1019 tcg_out_bswap32(s, datahi_reg, datahi_reg, TCG_REG_R20);
fd76e73a
RH
1020 }
1021 break;
1022 default:
1023 tcg_abort();
f54b3f92 1024 }
f061b40e
RH
1025}
1026
1027static void tcg_out_qemu_ld(TCGContext *s, const TCGArg *args, int opc)
1028{
1029 int datalo_reg = *args++;
1030 /* Note that datahi_reg is only used for 64-bit loads. */
1031 int datahi_reg = (opc == 3 ? *args++ : TCG_REG_R0);
1032 int addrlo_reg = *args++;
f54b3f92
AJ
1033
1034#if defined(CONFIG_SOFTMMU)
f061b40e
RH
1035 /* Note that addrhi_reg is only used for 64-bit guests. */
1036 int addrhi_reg = (TARGET_LONG_BITS == 64 ? *args++ : TCG_REG_R0);
1037 int mem_index = *args;
1038 int lab1, lab2, argreg, offset;
1039
1040 lab1 = gen_new_label();
1041 lab2 = gen_new_label();
1042
1043 offset = offsetof(CPUState, tlb_table[mem_index][0].addr_read);
1044 offset = tcg_out_tlb_read(s, TCG_REG_R26, TCG_REG_R25, addrlo_reg, addrhi_reg,
1045 opc & 3, lab1, offset);
1046
1047 /* TLB Hit. */
1048 tcg_out_ld(s, TCG_TYPE_PTR, TCG_REG_R20, (offset ? TCG_REG_R1 : TCG_REG_R25),
1049 offsetof(CPUState, tlb_table[mem_index][0].addend) - offset);
1050 tcg_out_qemu_ld_direct(s, datalo_reg, datahi_reg, addrlo_reg, TCG_REG_R20, opc);
fd76e73a
RH
1051 tcg_out_branch(s, lab2, 1);
1052
1053 /* TLB Miss. */
1054 /* label1: */
9d6fca70 1055 tcg_out_label(s, lab1, s->code_ptr);
fd76e73a
RH
1056
1057 argreg = TCG_REG_R26;
3b6dac34 1058 tcg_out_mov(s, TCG_TYPE_I32, argreg--, addrlo_reg);
fd76e73a 1059 if (TARGET_LONG_BITS == 64) {
3b6dac34 1060 tcg_out_mov(s, TCG_TYPE_I32, argreg--, addrhi_reg);
fd76e73a
RH
1061 }
1062 tcg_out_movi(s, TCG_TYPE_I32, argreg, mem_index);
1063
f061b40e 1064 tcg_out_call(s, qemu_ld_helpers[opc & 3]);
fd76e73a
RH
1065
1066 switch (opc) {
1067 case 0:
f061b40e 1068 tcg_out_andi(s, datalo_reg, TCG_REG_RET0, 0xff);
fd76e73a
RH
1069 break;
1070 case 0 | 4:
f061b40e 1071 tcg_out_ext8s(s, datalo_reg, TCG_REG_RET0);
fd76e73a
RH
1072 break;
1073 case 1:
f061b40e 1074 tcg_out_andi(s, datalo_reg, TCG_REG_RET0, 0xffff);
fd76e73a
RH
1075 break;
1076 case 1 | 4:
f061b40e 1077 tcg_out_ext16s(s, datalo_reg, TCG_REG_RET0);
fd76e73a
RH
1078 break;
1079 case 2:
1080 case 2 | 4:
3b6dac34 1081 tcg_out_mov(s, TCG_TYPE_I32, datalo_reg, TCG_REG_RET0);
fd76e73a
RH
1082 break;
1083 case 3:
3b6dac34
RH
1084 tcg_out_mov(s, TCG_TYPE_I32, datahi_reg, TCG_REG_RET0);
1085 tcg_out_mov(s, TCG_TYPE_I32, datalo_reg, TCG_REG_RET1);
fd76e73a
RH
1086 break;
1087 default:
1088 tcg_abort();
1089 }
1090
f54b3f92 1091 /* label2: */
9d6fca70 1092 tcg_out_label(s, lab2, s->code_ptr);
f061b40e
RH
1093#else
1094 tcg_out_qemu_ld_direct(s, datalo_reg, datahi_reg, addrlo_reg,
1095 (GUEST_BASE ? TCG_GUEST_BASE_REG : TCG_REG_R0), opc);
f54b3f92
AJ
1096#endif
1097}
1098
f061b40e
RH
1099static void tcg_out_qemu_st_direct(TCGContext *s, int datalo_reg, int datahi_reg,
1100 int addr_reg, int opc)
f54b3f92 1101{
f54b3f92 1102#ifdef TARGET_WORDS_BIGENDIAN
f061b40e 1103 const int bswap = 0;
f54b3f92 1104#else
f061b40e 1105 const int bswap = 1;
f54b3f92 1106#endif
f061b40e 1107
f54b3f92
AJ
1108 switch (opc) {
1109 case 0:
f061b40e 1110 tcg_out_ldst(s, datalo_reg, addr_reg, 0, INSN_STB);
f54b3f92
AJ
1111 break;
1112 case 1:
1113 if (bswap) {
f061b40e
RH
1114 tcg_out_bswap16(s, TCG_REG_R20, datalo_reg, 0);
1115 datalo_reg = TCG_REG_R20;
f54b3f92 1116 }
f061b40e 1117 tcg_out_ldst(s, datalo_reg, addr_reg, 0, INSN_STH);
f54b3f92
AJ
1118 break;
1119 case 2:
1120 if (bswap) {
f061b40e
RH
1121 tcg_out_bswap32(s, TCG_REG_R20, datalo_reg, TCG_REG_R20);
1122 datalo_reg = TCG_REG_R20;
f54b3f92 1123 }
f061b40e 1124 tcg_out_ldst(s, datalo_reg, addr_reg, 0, INSN_STW);
f54b3f92
AJ
1125 break;
1126 case 3:
fd76e73a 1127 if (bswap) {
f061b40e
RH
1128 tcg_out_bswap32(s, TCG_REG_R20, datalo_reg, TCG_REG_R20);
1129 tcg_out_bswap32(s, TCG_REG_R23, datahi_reg, TCG_REG_R23);
1130 datahi_reg = TCG_REG_R20;
1131 datalo_reg = TCG_REG_R23;
f54b3f92 1132 }
f061b40e
RH
1133 tcg_out_ldst(s, datahi_reg, addr_reg, 0, INSN_STW);
1134 tcg_out_ldst(s, datalo_reg, addr_reg, 4, INSN_STW);
f54b3f92
AJ
1135 break;
1136 default:
1137 tcg_abort();
1138 }
1139
f061b40e
RH
1140}
1141
1142static void tcg_out_qemu_st(TCGContext *s, const TCGArg *args, int opc)
1143{
1144 int datalo_reg = *args++;
1145 /* Note that datahi_reg is only used for 64-bit loads. */
1146 int datahi_reg = (opc == 3 ? *args++ : TCG_REG_R0);
1147 int addrlo_reg = *args++;
1148
f54b3f92 1149#if defined(CONFIG_SOFTMMU)
f061b40e
RH
1150 /* Note that addrhi_reg is only used for 64-bit guests. */
1151 int addrhi_reg = (TARGET_LONG_BITS == 64 ? *args++ : TCG_REG_R0);
1152 int mem_index = *args;
1153 int lab1, lab2, argreg, offset;
1154
1155 lab1 = gen_new_label();
1156 lab2 = gen_new_label();
1157
1158 offset = offsetof(CPUState, tlb_table[mem_index][0].addr_write);
1159 offset = tcg_out_tlb_read(s, TCG_REG_R26, TCG_REG_R25, addrlo_reg, addrhi_reg,
1160 opc, lab1, offset);
1161
1162 /* TLB Hit. */
1163 tcg_out_ld(s, TCG_TYPE_PTR, TCG_REG_R20, (offset ? TCG_REG_R1 : TCG_REG_R25),
1164 offsetof(CPUState, tlb_table[mem_index][0].addend) - offset);
1165
1166 /* There are no indexed stores, so we must do this addition explitly.
1167 Careful to avoid R20, which is used for the bswaps to follow. */
1168 tcg_out_arith(s, TCG_REG_R31, addrlo_reg, TCG_REG_R20, INSN_ADDL);
1169 tcg_out_qemu_st_direct(s, datalo_reg, datahi_reg, TCG_REG_R31, opc);
fd76e73a
RH
1170 tcg_out_branch(s, lab2, 1);
1171
1172 /* TLB Miss. */
1173 /* label1: */
9d6fca70 1174 tcg_out_label(s, lab1, s->code_ptr);
fd76e73a
RH
1175
1176 argreg = TCG_REG_R26;
3b6dac34 1177 tcg_out_mov(s, TCG_TYPE_I32, argreg--, addrlo_reg);
fd76e73a 1178 if (TARGET_LONG_BITS == 64) {
3b6dac34 1179 tcg_out_mov(s, TCG_TYPE_I32, argreg--, addrhi_reg);
fd76e73a
RH
1180 }
1181
1182 switch(opc) {
1183 case 0:
f061b40e 1184 tcg_out_andi(s, argreg--, datalo_reg, 0xff);
fd76e73a
RH
1185 tcg_out_movi(s, TCG_TYPE_I32, argreg, mem_index);
1186 break;
1187 case 1:
f061b40e 1188 tcg_out_andi(s, argreg--, datalo_reg, 0xffff);
fd76e73a
RH
1189 tcg_out_movi(s, TCG_TYPE_I32, argreg, mem_index);
1190 break;
1191 case 2:
3b6dac34 1192 tcg_out_mov(s, TCG_TYPE_I32, argreg--, datalo_reg);
fd76e73a
RH
1193 tcg_out_movi(s, TCG_TYPE_I32, argreg, mem_index);
1194 break;
1195 case 3:
1196 /* Because of the alignment required by the 64-bit data argument,
1197 we will always use R23/R24. Also, we will always run out of
1198 argument registers for storing mem_index, so that will have
1199 to go on the stack. */
1200 if (mem_index == 0) {
1201 argreg = TCG_REG_R0;
1202 } else {
1203 argreg = TCG_REG_R20;
1204 tcg_out_movi(s, TCG_TYPE_I32, argreg, mem_index);
1205 }
3b6dac34
RH
1206 tcg_out_mov(s, TCG_TYPE_I32, TCG_REG_R23, datahi_reg);
1207 tcg_out_mov(s, TCG_TYPE_I32, TCG_REG_R24, datalo_reg);
a42bceec 1208 tcg_out_st(s, TCG_TYPE_I32, argreg, TCG_REG_CALL_STACK,
fd76e73a
RH
1209 TCG_TARGET_CALL_STACK_OFFSET - 4);
1210 break;
1211 default:
1212 tcg_abort();
1213 }
1214
f061b40e 1215 tcg_out_call(s, qemu_st_helpers[opc]);
fd76e73a 1216
f54b3f92 1217 /* label2: */
9d6fca70 1218 tcg_out_label(s, lab2, s->code_ptr);
f061b40e
RH
1219#else
1220 /* There are no indexed stores, so if GUEST_BASE is set we must do the add
1221 explicitly. Careful to avoid R20, which is used for the bswaps to follow. */
1222 if (GUEST_BASE != 0) {
1223 tcg_out_arith(s, TCG_REG_R31, addrlo_reg, TCG_GUEST_BASE_REG, INSN_ADDL);
1224 addrlo_reg = TCG_REG_R31;
1225 }
1226 tcg_out_qemu_st_direct(s, datalo_reg, datahi_reg, addrlo_reg, opc);
f54b3f92
AJ
1227#endif
1228}
1229
fd76e73a
RH
1230static void tcg_out_exit_tb(TCGContext *s, TCGArg arg)
1231{
1232 if (!check_fit_tl(arg, 14)) {
1233 uint32_t hi, lo;
1234 hi = arg & ~0x7ff;
1235 lo = arg & 0x7ff;
1236 if (lo) {
1237 tcg_out_movi(s, TCG_TYPE_PTR, TCG_REG_RET0, hi);
1238 tcg_out32(s, INSN_BV | INSN_R2(TCG_REG_R18));
1239 tcg_out_addi(s, TCG_REG_RET0, lo);
1240 return;
1241 }
1242 arg = hi;
1243 }
1244 tcg_out32(s, INSN_BV | INSN_R2(TCG_REG_R18));
1245 tcg_out_movi(s, TCG_TYPE_PTR, TCG_REG_RET0, arg);
1246}
1247
1248static void tcg_out_goto_tb(TCGContext *s, TCGArg arg)
1249{
1250 if (s->tb_jmp_offset) {
1251 /* direct jump method */
1252 fprintf(stderr, "goto_tb direct\n");
1253 tcg_abort();
1254 } else {
1255 /* indirect jump method */
1256 tcg_out_ld(s, TCG_TYPE_PTR, TCG_REG_R20, TCG_REG_R0,
1257 (tcg_target_long)(s->tb_next + arg));
1258 tcg_out32(s, INSN_BV_N | INSN_R2(TCG_REG_R20));
1259 }
1260 s->tb_next_offset[arg] = s->code_ptr - s->code_buf;
1261}
1262
a9751609 1263static inline void tcg_out_op(TCGContext *s, TCGOpcode opc, const TCGArg *args,
f54b3f92
AJ
1264 const int *const_args)
1265{
f54b3f92
AJ
1266 switch (opc) {
1267 case INDEX_op_exit_tb:
fd76e73a 1268 tcg_out_exit_tb(s, args[0]);
f54b3f92
AJ
1269 break;
1270 case INDEX_op_goto_tb:
fd76e73a 1271 tcg_out_goto_tb(s, args[0]);
f54b3f92 1272 break;
fd76e73a 1273
f54b3f92 1274 case INDEX_op_call:
fd76e73a
RH
1275 if (const_args[0]) {
1276 tcg_out_call(s, (void *)args[0]);
1277 } else {
3e1f46ea
RH
1278 /* ??? FIXME: the value in the register in args[0] is almost
1279 certainly a procedure descriptor, not a code address. We
1280 probably need to use the millicode $$dyncall routine. */
1281 tcg_abort();
fd76e73a 1282 }
f54b3f92 1283 break;
fd76e73a 1284
f54b3f92
AJ
1285 case INDEX_op_jmp:
1286 fprintf(stderr, "unimplemented jmp\n");
1287 tcg_abort();
1288 break;
fd76e73a 1289
f54b3f92 1290 case INDEX_op_br:
fd76e73a 1291 tcg_out_branch(s, args[0], 1);
f54b3f92 1292 break;
fd76e73a 1293
f54b3f92
AJ
1294 case INDEX_op_movi_i32:
1295 tcg_out_movi(s, TCG_TYPE_I32, args[0], (uint32_t)args[1]);
1296 break;
1297
1298 case INDEX_op_ld8u_i32:
fd76e73a 1299 tcg_out_ldst(s, args[0], args[1], args[2], INSN_LDB);
f54b3f92
AJ
1300 break;
1301 case INDEX_op_ld8s_i32:
fd76e73a 1302 tcg_out_ldst(s, args[0], args[1], args[2], INSN_LDB);
f54b3f92
AJ
1303 tcg_out_ext8s(s, args[0], args[0]);
1304 break;
1305 case INDEX_op_ld16u_i32:
fd76e73a 1306 tcg_out_ldst(s, args[0], args[1], args[2], INSN_LDH);
f54b3f92
AJ
1307 break;
1308 case INDEX_op_ld16s_i32:
fd76e73a 1309 tcg_out_ldst(s, args[0], args[1], args[2], INSN_LDH);
f54b3f92
AJ
1310 tcg_out_ext16s(s, args[0], args[0]);
1311 break;
1312 case INDEX_op_ld_i32:
fd76e73a 1313 tcg_out_ldst(s, args[0], args[1], args[2], INSN_LDW);
f54b3f92
AJ
1314 break;
1315
1316 case INDEX_op_st8_i32:
fd76e73a 1317 tcg_out_ldst(s, args[0], args[1], args[2], INSN_STB);
f54b3f92
AJ
1318 break;
1319 case INDEX_op_st16_i32:
fd76e73a 1320 tcg_out_ldst(s, args[0], args[1], args[2], INSN_STH);
f54b3f92
AJ
1321 break;
1322 case INDEX_op_st_i32:
fd76e73a
RH
1323 tcg_out_ldst(s, args[0], args[1], args[2], INSN_STW);
1324 break;
1325
1326 case INDEX_op_add_i32:
1327 if (const_args[2]) {
1328 tcg_out_addi2(s, args[0], args[1], args[2]);
1329 } else {
1330 tcg_out_arith(s, args[0], args[1], args[2], INSN_ADDL);
1331 }
f54b3f92
AJ
1332 break;
1333
1334 case INDEX_op_sub_i32:
fd76e73a
RH
1335 if (const_args[1]) {
1336 if (const_args[2]) {
1337 tcg_out_movi(s, TCG_TYPE_I32, args[0], args[1] - args[2]);
1338 } else {
1339 /* Recall that SUBI is a reversed subtract. */
1340 tcg_out_arithi(s, args[0], args[2], args[1], INSN_SUBI);
1341 }
1342 } else if (const_args[2]) {
1343 tcg_out_addi2(s, args[0], args[1], -args[2]);
1344 } else {
1345 tcg_out_arith(s, args[0], args[1], args[2], INSN_SUB);
1346 }
1347 break;
1348
f54b3f92 1349 case INDEX_op_and_i32:
fd76e73a
RH
1350 if (const_args[2]) {
1351 tcg_out_andi(s, args[0], args[1], args[2]);
1352 } else {
1353 tcg_out_arith(s, args[0], args[1], args[2], INSN_AND);
1354 }
1355 break;
1356
f54b3f92 1357 case INDEX_op_or_i32:
fd76e73a
RH
1358 if (const_args[2]) {
1359 tcg_out_ori(s, args[0], args[1], args[2]);
1360 } else {
1361 tcg_out_arith(s, args[0], args[1], args[2], INSN_OR);
1362 }
1363 break;
1364
f54b3f92 1365 case INDEX_op_xor_i32:
fd76e73a
RH
1366 tcg_out_arith(s, args[0], args[1], args[2], INSN_XOR);
1367 break;
1368
1369 case INDEX_op_andc_i32:
1370 if (const_args[2]) {
1371 tcg_out_andi(s, args[0], args[1], ~args[2]);
1372 } else {
1373 tcg_out_arith(s, args[0], args[1], args[2], INSN_ANDCM);
1374 }
1375 break;
f54b3f92
AJ
1376
1377 case INDEX_op_shl_i32:
fd76e73a
RH
1378 if (const_args[2]) {
1379 tcg_out_shli(s, args[0], args[1], args[2]);
1380 } else {
1381 tcg_out_shl(s, args[0], args[1], args[2]);
1382 }
f54b3f92 1383 break;
fd76e73a 1384
f54b3f92 1385 case INDEX_op_shr_i32:
fd76e73a
RH
1386 if (const_args[2]) {
1387 tcg_out_shri(s, args[0], args[1], args[2]);
1388 } else {
1389 tcg_out_shr(s, args[0], args[1], args[2]);
1390 }
f54b3f92 1391 break;
fd76e73a 1392
f54b3f92 1393 case INDEX_op_sar_i32:
fd76e73a
RH
1394 if (const_args[2]) {
1395 tcg_out_sari(s, args[0], args[1], args[2]);
1396 } else {
1397 tcg_out_sar(s, args[0], args[1], args[2]);
1398 }
1399 break;
1400
1401 case INDEX_op_rotl_i32:
1402 if (const_args[2]) {
1403 tcg_out_rotli(s, args[0], args[1], args[2]);
1404 } else {
1405 tcg_out_rotl(s, args[0], args[1], args[2]);
1406 }
1407 break;
1408
1409 case INDEX_op_rotr_i32:
1410 if (const_args[2]) {
1411 tcg_out_rotri(s, args[0], args[1], args[2]);
1412 } else {
1413 tcg_out_rotr(s, args[0], args[1], args[2]);
1414 }
f54b3f92
AJ
1415 break;
1416
1417 case INDEX_op_mul_i32:
fd76e73a 1418 tcg_out_xmpyu(s, args[0], TCG_REG_R0, args[1], args[2]);
f54b3f92
AJ
1419 break;
1420 case INDEX_op_mulu2_i32:
fd76e73a 1421 tcg_out_xmpyu(s, args[0], args[1], args[2], args[3]);
f54b3f92 1422 break;
fd76e73a
RH
1423
1424 case INDEX_op_bswap16_i32:
1425 tcg_out_bswap16(s, args[0], args[1], 0);
f54b3f92 1426 break;
fd76e73a
RH
1427 case INDEX_op_bswap32_i32:
1428 tcg_out_bswap32(s, args[0], args[1], TCG_REG_R20);
1429 break;
1430
1431 case INDEX_op_not_i32:
1432 tcg_out_arithi(s, args[0], args[1], -1, INSN_SUBI);
1433 break;
1434 case INDEX_op_ext8s_i32:
1435 tcg_out_ext8s(s, args[0], args[1]);
1436 break;
1437 case INDEX_op_ext16s_i32:
1438 tcg_out_ext16s(s, args[0], args[1]);
1439 break;
1440
f54b3f92 1441 case INDEX_op_brcond_i32:
fd76e73a
RH
1442 tcg_out_brcond(s, args[2], args[0], args[1], const_args[1], args[3]);
1443 break;
1444 case INDEX_op_brcond2_i32:
1445 tcg_out_brcond2(s, args[4], args[0], args[1],
1446 args[2], const_args[2],
1447 args[3], const_args[3], args[5]);
1448 break;
1449
1450 case INDEX_op_setcond_i32:
1451 tcg_out_setcond(s, args[3], args[0], args[1], args[2], const_args[2]);
1452 break;
1453 case INDEX_op_setcond2_i32:
1454 tcg_out_setcond2(s, args[5], args[0], args[1], args[2],
1455 args[3], const_args[3], args[4], const_args[4]);
1456 break;
1457
1458 case INDEX_op_add2_i32:
91493631
RH
1459 tcg_out_add2(s, args[0], args[1], args[2], args[3],
1460 args[4], args[5], const_args[4]);
fd76e73a
RH
1461 break;
1462
1463 case INDEX_op_sub2_i32:
91493631
RH
1464 tcg_out_sub2(s, args[0], args[1], args[2], args[3],
1465 args[4], args[5], const_args[2], const_args[4]);
f54b3f92
AJ
1466 break;
1467
ec188429
RH
1468 case INDEX_op_deposit_i32:
1469 if (const_args[2]) {
1470 tcg_out_depi(s, args[0], args[2], args[3], args[4]);
1471 } else {
1472 tcg_out_dep(s, args[0], args[2], args[3], args[4]);
1473 }
1474 break;
1475
f54b3f92
AJ
1476 case INDEX_op_qemu_ld8u:
1477 tcg_out_qemu_ld(s, args, 0);
1478 break;
1479 case INDEX_op_qemu_ld8s:
1480 tcg_out_qemu_ld(s, args, 0 | 4);
1481 break;
1482 case INDEX_op_qemu_ld16u:
1483 tcg_out_qemu_ld(s, args, 1);
1484 break;
1485 case INDEX_op_qemu_ld16s:
1486 tcg_out_qemu_ld(s, args, 1 | 4);
1487 break;
86feb1c8 1488 case INDEX_op_qemu_ld32:
f54b3f92
AJ
1489 tcg_out_qemu_ld(s, args, 2);
1490 break;
fd76e73a
RH
1491 case INDEX_op_qemu_ld64:
1492 tcg_out_qemu_ld(s, args, 3);
1493 break;
f54b3f92
AJ
1494
1495 case INDEX_op_qemu_st8:
1496 tcg_out_qemu_st(s, args, 0);
1497 break;
1498 case INDEX_op_qemu_st16:
1499 tcg_out_qemu_st(s, args, 1);
1500 break;
1501 case INDEX_op_qemu_st32:
1502 tcg_out_qemu_st(s, args, 2);
1503 break;
fd76e73a
RH
1504 case INDEX_op_qemu_st64:
1505 tcg_out_qemu_st(s, args, 3);
1506 break;
f54b3f92
AJ
1507
1508 default:
1509 fprintf(stderr, "unknown opcode 0x%x\n", opc);
1510 tcg_abort();
1511 }
f54b3f92
AJ
1512}
1513
1514static const TCGTargetOpDef hppa_op_defs[] = {
1515 { INDEX_op_exit_tb, { } },
1516 { INDEX_op_goto_tb, { } },
1517
fd76e73a 1518 { INDEX_op_call, { "ri" } },
f54b3f92
AJ
1519 { INDEX_op_jmp, { "r" } },
1520 { INDEX_op_br, { } },
1521
1522 { INDEX_op_mov_i32, { "r", "r" } },
1523 { INDEX_op_movi_i32, { "r" } },
fd76e73a 1524
f54b3f92
AJ
1525 { INDEX_op_ld8u_i32, { "r", "r" } },
1526 { INDEX_op_ld8s_i32, { "r", "r" } },
1527 { INDEX_op_ld16u_i32, { "r", "r" } },
1528 { INDEX_op_ld16s_i32, { "r", "r" } },
1529 { INDEX_op_ld_i32, { "r", "r" } },
fd76e73a
RH
1530 { INDEX_op_st8_i32, { "rZ", "r" } },
1531 { INDEX_op_st16_i32, { "rZ", "r" } },
1532 { INDEX_op_st_i32, { "rZ", "r" } },
1533
1534 { INDEX_op_add_i32, { "r", "rZ", "ri" } },
1535 { INDEX_op_sub_i32, { "r", "rI", "ri" } },
0085bd51
RH
1536 { INDEX_op_and_i32, { "r", "rZ", "rM" } },
1537 { INDEX_op_or_i32, { "r", "rZ", "rO" } },
fd76e73a 1538 { INDEX_op_xor_i32, { "r", "rZ", "rZ" } },
0085bd51
RH
1539 /* Note that the second argument will be inverted, which means
1540 we want a constant whose inversion matches M, and that O = ~M.
1541 See the implementation of and_mask_p. */
1542 { INDEX_op_andc_i32, { "r", "rZ", "rO" } },
fd76e73a
RH
1543
1544 { INDEX_op_mul_i32, { "r", "r", "r" } },
1545 { INDEX_op_mulu2_i32, { "r", "r", "r", "r" } },
f54b3f92 1546
fd76e73a
RH
1547 { INDEX_op_shl_i32, { "r", "r", "ri" } },
1548 { INDEX_op_shr_i32, { "r", "r", "ri" } },
1549 { INDEX_op_sar_i32, { "r", "r", "ri" } },
1550 { INDEX_op_rotl_i32, { "r", "r", "ri" } },
1551 { INDEX_op_rotr_i32, { "r", "r", "ri" } },
f54b3f92 1552
fd76e73a
RH
1553 { INDEX_op_bswap16_i32, { "r", "r" } },
1554 { INDEX_op_bswap32_i32, { "r", "r" } },
fd76e73a 1555 { INDEX_op_not_i32, { "r", "r" } },
f54b3f92 1556
fd76e73a 1557 { INDEX_op_ext8s_i32, { "r", "r" } },
fd76e73a 1558 { INDEX_op_ext16s_i32, { "r", "r" } },
fd76e73a
RH
1559
1560 { INDEX_op_brcond_i32, { "rZ", "rJ" } },
1561 { INDEX_op_brcond2_i32, { "rZ", "rZ", "rJ", "rJ" } },
1562
1563 { INDEX_op_setcond_i32, { "r", "rZ", "rI" } },
1564 { INDEX_op_setcond2_i32, { "r", "rZ", "rZ", "rI", "rI" } },
1565
1566 { INDEX_op_add2_i32, { "r", "r", "rZ", "rZ", "rI", "rZ" } },
91493631 1567 { INDEX_op_sub2_i32, { "r", "r", "rI", "rZ", "rK", "rZ" } },
f54b3f92 1568
ec188429
RH
1569 { INDEX_op_deposit_i32, { "r", "0", "rJ" } },
1570
f54b3f92
AJ
1571#if TARGET_LONG_BITS == 32
1572 { INDEX_op_qemu_ld8u, { "r", "L" } },
1573 { INDEX_op_qemu_ld8s, { "r", "L" } },
1574 { INDEX_op_qemu_ld16u, { "r", "L" } },
1575 { INDEX_op_qemu_ld16s, { "r", "L" } },
86feb1c8 1576 { INDEX_op_qemu_ld32, { "r", "L" } },
f54b3f92
AJ
1577 { INDEX_op_qemu_ld64, { "r", "r", "L" } },
1578
fd76e73a
RH
1579 { INDEX_op_qemu_st8, { "LZ", "L" } },
1580 { INDEX_op_qemu_st16, { "LZ", "L" } },
1581 { INDEX_op_qemu_st32, { "LZ", "L" } },
1582 { INDEX_op_qemu_st64, { "LZ", "LZ", "L" } },
f54b3f92
AJ
1583#else
1584 { INDEX_op_qemu_ld8u, { "r", "L", "L" } },
1585 { INDEX_op_qemu_ld8s, { "r", "L", "L" } },
1586 { INDEX_op_qemu_ld16u, { "r", "L", "L" } },
1587 { INDEX_op_qemu_ld16s, { "r", "L", "L" } },
86feb1c8 1588 { INDEX_op_qemu_ld32, { "r", "L", "L" } },
f54b3f92
AJ
1589 { INDEX_op_qemu_ld64, { "r", "r", "L", "L" } },
1590
fd76e73a
RH
1591 { INDEX_op_qemu_st8, { "LZ", "L", "L" } },
1592 { INDEX_op_qemu_st16, { "LZ", "L", "L" } },
1593 { INDEX_op_qemu_st32, { "LZ", "L", "L" } },
1594 { INDEX_op_qemu_st64, { "LZ", "LZ", "L", "L" } },
f54b3f92
AJ
1595#endif
1596 { -1 },
1597};
1598
fd76e73a
RH
1599static int tcg_target_callee_save_regs[] = {
1600 /* R2, the return address register, is saved specially
1601 in the caller's frame. */
1602 /* R3, the frame pointer, is not currently modified. */
1603 TCG_REG_R4,
1604 TCG_REG_R5,
1605 TCG_REG_R6,
1606 TCG_REG_R7,
1607 TCG_REG_R8,
1608 TCG_REG_R9,
1609 TCG_REG_R10,
1610 TCG_REG_R11,
1611 TCG_REG_R12,
1612 TCG_REG_R13,
1613 TCG_REG_R14,
1614 TCG_REG_R15,
1615 TCG_REG_R16,
cea5f9a2 1616 TCG_REG_R17, /* R17 is the global env. */
fd76e73a
RH
1617 TCG_REG_R18
1618};
1619
e4d58b41 1620static void tcg_target_qemu_prologue(TCGContext *s)
fd76e73a
RH
1621{
1622 int frame_size, i;
1623
1624 /* Allocate space for the fixed frame marker. */
1625 frame_size = -TCG_TARGET_CALL_STACK_OFFSET;
1626 frame_size += TCG_TARGET_STATIC_CALL_ARGS_SIZE;
1627
1628 /* Allocate space for the saved registers. */
1629 frame_size += ARRAY_SIZE(tcg_target_callee_save_regs) * 4;
1630
2a6a665f
BS
1631 /* Allocate space for the TCG temps. */
1632 frame_size += CPU_TEMP_BUF_NLONGS * sizeof(long);
1633
fd76e73a
RH
1634 /* Align the allocated space. */
1635 frame_size = ((frame_size + TCG_TARGET_STACK_ALIGN - 1)
1636 & -TCG_TARGET_STACK_ALIGN);
1637
1638 /* The return address is stored in the caller's frame. */
a42bceec 1639 tcg_out_st(s, TCG_TYPE_PTR, TCG_REG_RP, TCG_REG_CALL_STACK, -20);
fd76e73a
RH
1640
1641 /* Allocate stack frame, saving the first register at the same time. */
1642 tcg_out_ldst(s, tcg_target_callee_save_regs[0],
a42bceec 1643 TCG_REG_CALL_STACK, frame_size, INSN_STWM);
fd76e73a
RH
1644
1645 /* Save all callee saved registers. */
1646 for (i = 1; i < ARRAY_SIZE(tcg_target_callee_save_regs); i++) {
1647 tcg_out_st(s, TCG_TYPE_PTR, tcg_target_callee_save_regs[i],
a42bceec 1648 TCG_REG_CALL_STACK, -frame_size + i * 4);
fd76e73a
RH
1649 }
1650
2a6a665f
BS
1651 /* Record the location of the TCG temps. */
1652 tcg_set_frame(s, TCG_REG_CALL_STACK, -frame_size + i * 4,
6e6a9924 1653 CPU_TEMP_BUF_NLONGS * sizeof(long));
2a6a665f 1654
884d348b 1655#ifdef CONFIG_USE_GUEST_BASE
4b31713d
RH
1656 if (GUEST_BASE != 0) {
1657 tcg_out_movi(s, TCG_TYPE_PTR, TCG_GUEST_BASE_REG, GUEST_BASE);
1658 tcg_regset_set_reg(s->reserved_regs, TCG_GUEST_BASE_REG);
1659 }
884d348b 1660#endif
fd76e73a 1661
cea5f9a2
BS
1662 tcg_out_mov(s, TCG_TYPE_PTR, TCG_AREG0, tcg_target_call_iarg_regs[0]);
1663
fd76e73a 1664 /* Jump to TB, and adjust R18 to be the return address. */
cea5f9a2 1665 tcg_out32(s, INSN_BLE_SR4 | INSN_R2(tcg_target_call_iarg_regs[1]));
3b6dac34 1666 tcg_out_mov(s, TCG_TYPE_I32, TCG_REG_R18, TCG_REG_R31);
fd76e73a
RH
1667
1668 /* Restore callee saved registers. */
a42bceec
BS
1669 tcg_out_ld(s, TCG_TYPE_PTR, TCG_REG_RP, TCG_REG_CALL_STACK,
1670 -frame_size - 20);
fd76e73a
RH
1671 for (i = 1; i < ARRAY_SIZE(tcg_target_callee_save_regs); i++) {
1672 tcg_out_ld(s, TCG_TYPE_PTR, tcg_target_callee_save_regs[i],
a42bceec 1673 TCG_REG_CALL_STACK, -frame_size + i * 4);
fd76e73a
RH
1674 }
1675
1676 /* Deallocate stack frame and return. */
1677 tcg_out32(s, INSN_BV | INSN_R2(TCG_REG_RP));
1678 tcg_out_ldst(s, tcg_target_callee_save_regs[0],
a42bceec 1679 TCG_REG_CALL_STACK, -frame_size, INSN_LDWM);
fd76e73a
RH
1680}
1681
e4d58b41 1682static void tcg_target_init(TCGContext *s)
f54b3f92
AJ
1683{
1684 tcg_regset_set32(tcg_target_available_regs[TCG_TYPE_I32], 0, 0xffffffff);
fd76e73a
RH
1685
1686 tcg_regset_clear(tcg_target_call_clobber_regs);
1687 tcg_regset_set_reg(tcg_target_call_clobber_regs, TCG_REG_R20);
1688 tcg_regset_set_reg(tcg_target_call_clobber_regs, TCG_REG_R21);
1689 tcg_regset_set_reg(tcg_target_call_clobber_regs, TCG_REG_R22);
1690 tcg_regset_set_reg(tcg_target_call_clobber_regs, TCG_REG_R23);
1691 tcg_regset_set_reg(tcg_target_call_clobber_regs, TCG_REG_R24);
1692 tcg_regset_set_reg(tcg_target_call_clobber_regs, TCG_REG_R25);
1693 tcg_regset_set_reg(tcg_target_call_clobber_regs, TCG_REG_R26);
1694 tcg_regset_set_reg(tcg_target_call_clobber_regs, TCG_REG_RET0);
1695 tcg_regset_set_reg(tcg_target_call_clobber_regs, TCG_REG_RET1);
f54b3f92
AJ
1696
1697 tcg_regset_clear(s->reserved_regs);
1698 tcg_regset_set_reg(s->reserved_regs, TCG_REG_R0); /* hardwired to zero */
1699 tcg_regset_set_reg(s->reserved_regs, TCG_REG_R1); /* addil target */
1700 tcg_regset_set_reg(s->reserved_regs, TCG_REG_RP); /* link register */
1701 tcg_regset_set_reg(s->reserved_regs, TCG_REG_R3); /* frame pointer */
1702 tcg_regset_set_reg(s->reserved_regs, TCG_REG_R18); /* return pointer */
1703 tcg_regset_set_reg(s->reserved_regs, TCG_REG_R19); /* clobbered w/o pic */
1704 tcg_regset_set_reg(s->reserved_regs, TCG_REG_R20); /* reserved */
1705 tcg_regset_set_reg(s->reserved_regs, TCG_REG_DP); /* data pointer */
a42bceec 1706 tcg_regset_set_reg(s->reserved_regs, TCG_REG_CALL_STACK); /* stack pointer */
f54b3f92
AJ
1707 tcg_regset_set_reg(s->reserved_regs, TCG_REG_R31); /* ble link reg */
1708
1709 tcg_add_target_add_op_defs(hppa_op_defs);
1710}