]> git.proxmox.com Git - mirror_qemu.git/blame - tcg/hppa/tcg-target.c
tcg: Initialize the prologue after GUEST_BASE is fixed.
[mirror_qemu.git] / tcg / hppa / tcg-target.c
CommitLineData
f54b3f92
AJ
1/*
2 * Tiny Code Generator for QEMU
3 *
4 * Copyright (c) 2008 Fabrice Bellard
5 *
6 * Permission is hereby granted, free of charge, to any person obtaining a copy
7 * of this software and associated documentation files (the "Software"), to deal
8 * in the Software without restriction, including without limitation the rights
9 * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
10 * copies of the Software, and to permit persons to whom the Software is
11 * furnished to do so, subject to the following conditions:
12 *
13 * The above copyright notice and this permission notice shall be included in
14 * all copies or substantial portions of the Software.
15 *
16 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
17 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
18 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
19 * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
20 * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
21 * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
22 * THE SOFTWARE.
23 */
24
d4a9eb1f 25#ifndef NDEBUG
f54b3f92 26static const char * const tcg_target_reg_names[TCG_TARGET_NB_REGS] = {
fd76e73a
RH
27 "%r0", "%r1", "%rp", "%r3", "%r4", "%r5", "%r6", "%r7",
28 "%r8", "%r9", "%r10", "%r11", "%r12", "%r13", "%r14", "%r15",
29 "%r16", "%r17", "%r18", "%r19", "%r20", "%r21", "%r22", "%r23",
30 "%r24", "%r25", "%r26", "%dp", "%ret0", "%ret1", "%sp", "%r31",
f54b3f92 31};
d4a9eb1f 32#endif
f54b3f92 33
fd76e73a
RH
34/* This is an 8 byte temp slot in the stack frame. */
35#define STACK_TEMP_OFS -16
36
fd76e73a
RH
37#ifdef CONFIG_USE_GUEST_BASE
38#define TCG_GUEST_BASE_REG TCG_REG_R16
39#else
40#define TCG_GUEST_BASE_REG TCG_REG_R0
41#endif
42
f54b3f92
AJ
43static const int tcg_target_reg_alloc_order[] = {
44 TCG_REG_R4,
45 TCG_REG_R5,
46 TCG_REG_R6,
47 TCG_REG_R7,
48 TCG_REG_R8,
49 TCG_REG_R9,
50 TCG_REG_R10,
51 TCG_REG_R11,
52 TCG_REG_R12,
53 TCG_REG_R13,
54
55 TCG_REG_R17,
56 TCG_REG_R14,
57 TCG_REG_R15,
58 TCG_REG_R16,
fd76e73a
RH
59
60 TCG_REG_R26,
61 TCG_REG_R25,
62 TCG_REG_R24,
63 TCG_REG_R23,
64
65 TCG_REG_RET0,
66 TCG_REG_RET1,
f54b3f92
AJ
67};
68
69static const int tcg_target_call_iarg_regs[4] = {
70 TCG_REG_R26,
71 TCG_REG_R25,
72 TCG_REG_R24,
73 TCG_REG_R23,
74};
75
76static const int tcg_target_call_oarg_regs[2] = {
77 TCG_REG_RET0,
78 TCG_REG_RET1,
79};
80
fd76e73a
RH
81/* True iff val fits a signed field of width BITS. */
82static inline int check_fit_tl(tcg_target_long val, unsigned int bits)
83{
84 return (val << ((sizeof(tcg_target_long) * 8 - bits))
85 >> (sizeof(tcg_target_long) * 8 - bits)) == val;
86}
87
88/* True iff depi can be used to compute (reg | MASK).
89 Accept a bit pattern like:
90 0....01....1
91 1....10....0
92 0..01..10..0
93 Copied from gcc sources. */
94static inline int or_mask_p(tcg_target_ulong mask)
95{
0085bd51
RH
96 if (mask == 0 || mask == -1) {
97 return 0;
98 }
fd76e73a
RH
99 mask += mask & -mask;
100 return (mask & (mask - 1)) == 0;
101}
102
103/* True iff depi or extru can be used to compute (reg & mask).
104 Accept a bit pattern like these:
105 0....01....1
106 1....10....0
107 1..10..01..1
108 Copied from gcc sources. */
109static inline int and_mask_p(tcg_target_ulong mask)
110{
111 return or_mask_p(~mask);
112}
113
114static int low_sign_ext(int val, int len)
115{
116 return (((val << 1) & ~(-1u << len)) | ((val >> (len - 1)) & 1));
117}
118
119static int reassemble_12(int as12)
120{
121 return (((as12 & 0x800) >> 11) |
122 ((as12 & 0x400) >> 8) |
123 ((as12 & 0x3ff) << 3));
124}
125
126static int reassemble_17(int as17)
127{
128 return (((as17 & 0x10000) >> 16) |
129 ((as17 & 0x0f800) << 5) |
130 ((as17 & 0x00400) >> 8) |
131 ((as17 & 0x003ff) << 3));
132}
133
134static int reassemble_21(int as21)
135{
136 return (((as21 & 0x100000) >> 20) |
137 ((as21 & 0x0ffe00) >> 8) |
138 ((as21 & 0x000180) << 7) |
139 ((as21 & 0x00007c) << 14) |
140 ((as21 & 0x000003) << 12));
141}
142
143/* ??? Bizzarely, there is no PCREL12F relocation type. I guess all
144 such relocations are simply fully handled by the assembler. */
145#define R_PARISC_PCREL12F R_PARISC_NONE
146
f54b3f92
AJ
147static void patch_reloc(uint8_t *code_ptr, int type,
148 tcg_target_long value, tcg_target_long addend)
149{
fd76e73a
RH
150 uint32_t *insn_ptr = (uint32_t *)code_ptr;
151 uint32_t insn = *insn_ptr;
152 tcg_target_long pcrel;
153
154 value += addend;
155 pcrel = (value - ((tcg_target_long)code_ptr + 8)) >> 2;
156
f54b3f92 157 switch (type) {
fd76e73a
RH
158 case R_PARISC_PCREL12F:
159 assert(check_fit_tl(pcrel, 12));
160 /* ??? We assume all patches are forward. See tcg_out_brcond
161 re setting the NUL bit on the branch and eliding the nop. */
162 assert(pcrel >= 0);
163 insn &= ~0x1ffdu;
164 insn |= reassemble_12(pcrel);
165 break;
f54b3f92 166 case R_PARISC_PCREL17F:
fd76e73a
RH
167 assert(check_fit_tl(pcrel, 17));
168 insn &= ~0x1f1ffdu;
169 insn |= reassemble_17(pcrel);
f54b3f92
AJ
170 break;
171 default:
172 tcg_abort();
173 }
fd76e73a
RH
174
175 *insn_ptr = insn;
f54b3f92
AJ
176}
177
178/* maximum number of register used for input function arguments */
179static inline int tcg_target_get_call_iarg_regs_count(int flags)
180{
181 return 4;
182}
183
184/* parse target specific constraints */
d4a9eb1f 185static int target_parse_constraint(TCGArgConstraint *ct, const char **pct_str)
f54b3f92
AJ
186{
187 const char *ct_str;
188
189 ct_str = *pct_str;
190 switch (ct_str[0]) {
191 case 'r':
192 ct->ct |= TCG_CT_REG;
193 tcg_regset_set32(ct->u.regs, 0, 0xffffffff);
194 break;
195 case 'L': /* qemu_ld/st constraint */
196 ct->ct |= TCG_CT_REG;
197 tcg_regset_set32(ct->u.regs, 0, 0xffffffff);
198 tcg_regset_reset_reg(ct->u.regs, TCG_REG_R26);
199 tcg_regset_reset_reg(ct->u.regs, TCG_REG_R25);
200 tcg_regset_reset_reg(ct->u.regs, TCG_REG_R24);
201 tcg_regset_reset_reg(ct->u.regs, TCG_REG_R23);
202 break;
fd76e73a
RH
203 case 'Z':
204 ct->ct |= TCG_CT_CONST_0;
205 break;
206 case 'I':
207 ct->ct |= TCG_CT_CONST_S11;
208 break;
209 case 'J':
210 ct->ct |= TCG_CT_CONST_S5;
211 break;
91493631
RH
212 case 'K':
213 ct->ct |= TCG_CT_CONST_MS11;
214 break;
0085bd51
RH
215 case 'M':
216 ct->ct |= TCG_CT_CONST_AND;
217 break;
218 case 'O':
219 ct->ct |= TCG_CT_CONST_OR;
220 break;
f54b3f92
AJ
221 default:
222 return -1;
223 }
224 ct_str++;
225 *pct_str = ct_str;
226 return 0;
227}
228
229/* test if a constant matches the constraint */
fd76e73a
RH
230static int tcg_target_const_match(tcg_target_long val,
231 const TCGArgConstraint *arg_ct)
f54b3f92 232{
fd76e73a
RH
233 int ct = arg_ct->ct;
234 if (ct & TCG_CT_CONST) {
235 return 1;
236 } else if (ct & TCG_CT_CONST_0) {
237 return val == 0;
238 } else if (ct & TCG_CT_CONST_S5) {
239 return check_fit_tl(val, 5);
240 } else if (ct & TCG_CT_CONST_S11) {
241 return check_fit_tl(val, 11);
91493631
RH
242 } else if (ct & TCG_CT_CONST_MS11) {
243 return check_fit_tl(-val, 11);
0085bd51
RH
244 } else if (ct & TCG_CT_CONST_AND) {
245 return and_mask_p(val);
246 } else if (ct & TCG_CT_CONST_OR) {
247 return or_mask_p(val);
fd76e73a 248 }
f54b3f92
AJ
249 return 0;
250}
251
252#define INSN_OP(x) ((x) << 26)
253#define INSN_EXT3BR(x) ((x) << 13)
254#define INSN_EXT3SH(x) ((x) << 10)
255#define INSN_EXT4(x) ((x) << 6)
256#define INSN_EXT5(x) (x)
257#define INSN_EXT6(x) ((x) << 6)
258#define INSN_EXT7(x) ((x) << 6)
259#define INSN_EXT8A(x) ((x) << 6)
260#define INSN_EXT8B(x) ((x) << 5)
261#define INSN_T(x) (x)
262#define INSN_R1(x) ((x) << 16)
263#define INSN_R2(x) ((x) << 21)
264#define INSN_DEP_LEN(x) (32 - (x))
265#define INSN_SHDEP_CP(x) ((31 - (x)) << 5)
266#define INSN_SHDEP_P(x) ((x) << 5)
267#define INSN_COND(x) ((x) << 13)
fd76e73a
RH
268#define INSN_IM11(x) low_sign_ext(x, 11)
269#define INSN_IM14(x) low_sign_ext(x, 14)
270#define INSN_IM5(x) (low_sign_ext(x, 5) << 16)
271
272#define COND_NEVER 0
273#define COND_EQ 1
274#define COND_LT 2
275#define COND_LE 3
276#define COND_LTU 4
277#define COND_LEU 5
278#define COND_SV 6
279#define COND_OD 7
280#define COND_FALSE 8
281
282#define INSN_ADD (INSN_OP(0x02) | INSN_EXT6(0x18))
283#define INSN_ADDC (INSN_OP(0x02) | INSN_EXT6(0x1c))
284#define INSN_ADDI (INSN_OP(0x2d))
285#define INSN_ADDIL (INSN_OP(0x0a))
286#define INSN_ADDL (INSN_OP(0x02) | INSN_EXT6(0x28))
287#define INSN_AND (INSN_OP(0x02) | INSN_EXT6(0x08))
288#define INSN_ANDCM (INSN_OP(0x02) | INSN_EXT6(0x00))
289#define INSN_COMCLR (INSN_OP(0x02) | INSN_EXT6(0x22))
290#define INSN_COMICLR (INSN_OP(0x24))
291#define INSN_DEP (INSN_OP(0x35) | INSN_EXT3SH(3))
292#define INSN_DEPI (INSN_OP(0x35) | INSN_EXT3SH(7))
293#define INSN_EXTRS (INSN_OP(0x34) | INSN_EXT3SH(7))
294#define INSN_EXTRU (INSN_OP(0x34) | INSN_EXT3SH(6))
295#define INSN_LDIL (INSN_OP(0x08))
296#define INSN_LDO (INSN_OP(0x0d))
297#define INSN_MTCTL (INSN_OP(0x00) | INSN_EXT8B(0xc2))
298#define INSN_OR (INSN_OP(0x02) | INSN_EXT6(0x09))
299#define INSN_SHD (INSN_OP(0x34) | INSN_EXT3SH(2))
300#define INSN_SUB (INSN_OP(0x02) | INSN_EXT6(0x10))
301#define INSN_SUBB (INSN_OP(0x02) | INSN_EXT6(0x14))
302#define INSN_SUBI (INSN_OP(0x25))
303#define INSN_VEXTRS (INSN_OP(0x34) | INSN_EXT3SH(5))
304#define INSN_VEXTRU (INSN_OP(0x34) | INSN_EXT3SH(4))
305#define INSN_VSHD (INSN_OP(0x34) | INSN_EXT3SH(0))
306#define INSN_XOR (INSN_OP(0x02) | INSN_EXT6(0x0a))
307#define INSN_ZDEP (INSN_OP(0x35) | INSN_EXT3SH(2))
308#define INSN_ZVDEP (INSN_OP(0x35) | INSN_EXT3SH(0))
309
310#define INSN_BL (INSN_OP(0x3a) | INSN_EXT3BR(0))
311#define INSN_BL_N (INSN_OP(0x3a) | INSN_EXT3BR(0) | 2)
312#define INSN_BLR (INSN_OP(0x3a) | INSN_EXT3BR(2))
313#define INSN_BV (INSN_OP(0x3a) | INSN_EXT3BR(6))
314#define INSN_BV_N (INSN_OP(0x3a) | INSN_EXT3BR(6) | 2)
315#define INSN_BLE_SR4 (INSN_OP(0x39) | (1 << 13))
316
317#define INSN_LDB (INSN_OP(0x10))
318#define INSN_LDH (INSN_OP(0x11))
319#define INSN_LDW (INSN_OP(0x12))
320#define INSN_LDWM (INSN_OP(0x13))
321#define INSN_FLDDS (INSN_OP(0x0b) | INSN_EXT4(0) | (1 << 12))
322
323#define INSN_LDBX (INSN_OP(0x03) | INSN_EXT4(0))
324#define INSN_LDHX (INSN_OP(0x03) | INSN_EXT4(1))
325#define INSN_LDWX (INSN_OP(0x03) | INSN_EXT4(2))
326
327#define INSN_STB (INSN_OP(0x18))
328#define INSN_STH (INSN_OP(0x19))
329#define INSN_STW (INSN_OP(0x1a))
330#define INSN_STWM (INSN_OP(0x1b))
331#define INSN_FSTDS (INSN_OP(0x0b) | INSN_EXT4(8) | (1 << 12))
332
333#define INSN_COMBT (INSN_OP(0x20))
334#define INSN_COMBF (INSN_OP(0x22))
335#define INSN_COMIBT (INSN_OP(0x21))
336#define INSN_COMIBF (INSN_OP(0x23))
337
338/* supplied by libgcc */
339extern void *__canonicalize_funcptr_for_compare(void *);
340
341static void tcg_out_mov(TCGContext *s, int ret, int arg)
342{
343 /* PA1.1 defines COPY as OR r,0,t; PA2.0 defines COPY as LDO 0(r),t
344 but hppa-dis.c is unaware of this definition */
345 if (ret != arg) {
346 tcg_out32(s, INSN_OR | INSN_T(ret) | INSN_R1(arg)
347 | INSN_R2(TCG_REG_R0));
348 }
349}
f54b3f92 350
fd76e73a
RH
351static void tcg_out_movi(TCGContext *s, TCGType type,
352 int ret, tcg_target_long arg)
353{
354 if (check_fit_tl(arg, 14)) {
355 tcg_out32(s, INSN_LDO | INSN_R1(ret)
356 | INSN_R2(TCG_REG_R0) | INSN_IM14(arg));
357 } else {
358 uint32_t hi, lo;
359 hi = arg >> 11;
360 lo = arg & 0x7ff;
361
362 tcg_out32(s, INSN_LDIL | INSN_R2(ret) | reassemble_21(hi));
363 if (lo) {
364 tcg_out32(s, INSN_LDO | INSN_R1(ret)
365 | INSN_R2(ret) | INSN_IM14(lo));
366 }
367 }
368}
f54b3f92 369
fd76e73a
RH
370static void tcg_out_ldst(TCGContext *s, int ret, int addr,
371 tcg_target_long offset, int op)
372{
373 if (!check_fit_tl(offset, 14)) {
374 uint32_t hi, lo, op;
f54b3f92 375
fd76e73a
RH
376 hi = offset >> 11;
377 lo = offset & 0x7ff;
f54b3f92 378
fd76e73a
RH
379 if (addr == TCG_REG_R0) {
380 op = INSN_LDIL | INSN_R2(TCG_REG_R1);
381 } else {
382 op = INSN_ADDIL | INSN_R2(addr);
383 }
384 tcg_out32(s, op | reassemble_21(hi));
f54b3f92 385
fd76e73a
RH
386 addr = TCG_REG_R1;
387 offset = lo;
388 }
f54b3f92 389
fd76e73a
RH
390 if (ret != addr || offset != 0 || op != INSN_LDO) {
391 tcg_out32(s, op | INSN_R1(ret) | INSN_R2(addr) | INSN_IM14(offset));
392 }
393}
f54b3f92 394
fd76e73a
RH
395/* This function is required by tcg.c. */
396static inline void tcg_out_ld(TCGContext *s, TCGType type, int ret,
397 int arg1, tcg_target_long arg2)
398{
399 tcg_out_ldst(s, ret, arg1, arg2, INSN_LDW);
400}
401
402/* This function is required by tcg.c. */
403static inline void tcg_out_st(TCGContext *s, TCGType type, int ret,
404 int arg1, tcg_target_long arg2)
405{
406 tcg_out_ldst(s, ret, arg1, arg2, INSN_STW);
407}
408
409static void tcg_out_ldst_index(TCGContext *s, int data,
410 int base, int index, int op)
411{
412 tcg_out32(s, op | INSN_T(data) | INSN_R1(index) | INSN_R2(base));
413}
414
415static inline void tcg_out_addi2(TCGContext *s, int ret, int arg1,
416 tcg_target_long val)
417{
418 tcg_out_ldst(s, ret, arg1, val, INSN_LDO);
419}
f54b3f92 420
fd76e73a
RH
421/* This function is required by tcg.c. */
422static inline void tcg_out_addi(TCGContext *s, int reg, tcg_target_long val)
423{
424 tcg_out_addi2(s, reg, reg, val);
425}
f54b3f92 426
fd76e73a
RH
427static inline void tcg_out_arith(TCGContext *s, int t, int r1, int r2, int op)
428{
429 tcg_out32(s, op | INSN_T(t) | INSN_R1(r1) | INSN_R2(r2));
430}
f54b3f92 431
fd76e73a
RH
432static inline void tcg_out_arithi(TCGContext *s, int t, int r1,
433 tcg_target_long val, int op)
f54b3f92 434{
fd76e73a
RH
435 assert(check_fit_tl(val, 11));
436 tcg_out32(s, op | INSN_R1(t) | INSN_R2(r1) | INSN_IM11(val));
f54b3f92
AJ
437}
438
fd76e73a 439static inline void tcg_out_nop(TCGContext *s)
f54b3f92 440{
fd76e73a
RH
441 tcg_out_arith(s, TCG_REG_R0, TCG_REG_R0, TCG_REG_R0, INSN_OR);
442}
f54b3f92 443
fd76e73a
RH
444static inline void tcg_out_mtctl_sar(TCGContext *s, int arg)
445{
446 tcg_out32(s, INSN_MTCTL | INSN_R2(11) | INSN_R1(arg));
447}
448
449/* Extract LEN bits at position OFS from ARG and place in RET.
450 Note that here the bit ordering is reversed from the PA-RISC
451 standard, such that the right-most bit is 0. */
452static inline void tcg_out_extr(TCGContext *s, int ret, int arg,
453 unsigned ofs, unsigned len, int sign)
454{
455 assert(ofs < 32 && len <= 32 - ofs);
456 tcg_out32(s, (sign ? INSN_EXTRS : INSN_EXTRU)
457 | INSN_R1(ret) | INSN_R2(arg)
458 | INSN_SHDEP_P(31 - ofs) | INSN_DEP_LEN(len));
f54b3f92
AJ
459}
460
fd76e73a
RH
461/* Likewise with OFS interpreted little-endian. */
462static inline void tcg_out_dep(TCGContext *s, int ret, int arg,
463 unsigned ofs, unsigned len)
f54b3f92 464{
fd76e73a
RH
465 assert(ofs < 32 && len <= 32 - ofs);
466 tcg_out32(s, INSN_DEP | INSN_R2(ret) | INSN_R1(arg)
467 | INSN_SHDEP_CP(31 - ofs) | INSN_DEP_LEN(len));
468}
469
470static inline void tcg_out_shd(TCGContext *s, int ret, int hi, int lo,
471 unsigned count)
472{
473 assert(count < 32);
474 tcg_out32(s, INSN_SHD | INSN_R1(hi) | INSN_R2(lo) | INSN_T(ret)
475 | INSN_SHDEP_CP(count));
476}
477
478static void tcg_out_vshd(TCGContext *s, int ret, int hi, int lo, int creg)
479{
480 tcg_out_mtctl_sar(s, creg);
481 tcg_out32(s, INSN_VSHD | INSN_T(ret) | INSN_R1(hi) | INSN_R2(lo));
482}
483
484static void tcg_out_ori(TCGContext *s, int ret, int arg, tcg_target_ulong m)
485{
0085bd51
RH
486 int bs0, bs1;
487
488 /* Note that the argument is constrained to match or_mask_p. */
489 for (bs0 = 0; bs0 < 32; bs0++) {
490 if ((m & (1u << bs0)) != 0) {
491 break;
fd76e73a 492 }
0085bd51
RH
493 }
494 for (bs1 = bs0; bs1 < 32; bs1++) {
495 if ((m & (1u << bs1)) == 0) {
496 break;
fd76e73a 497 }
fd76e73a 498 }
0085bd51
RH
499 assert(bs1 == 32 || (1ul << bs1) > m);
500
501 tcg_out_mov(s, ret, arg);
502 tcg_out32(s, INSN_DEPI | INSN_R2(ret) | INSN_IM5(-1)
503 | INSN_SHDEP_CP(31 - bs0) | INSN_DEP_LEN(bs1 - bs0));
fd76e73a
RH
504}
505
506static void tcg_out_andi(TCGContext *s, int ret, int arg, tcg_target_ulong m)
507{
0085bd51 508 int ls0, ls1, ms0;
fd76e73a 509
0085bd51
RH
510 /* Note that the argument is constrained to match and_mask_p. */
511 for (ls0 = 0; ls0 < 32; ls0++) {
512 if ((m & (1u << ls0)) == 0) {
513 break;
fd76e73a 514 }
0085bd51
RH
515 }
516 for (ls1 = ls0; ls1 < 32; ls1++) {
517 if ((m & (1u << ls1)) != 0) {
518 break;
fd76e73a 519 }
0085bd51
RH
520 }
521 for (ms0 = ls1; ms0 < 32; ms0++) {
522 if ((m & (1u << ms0)) == 0) {
523 break;
fd76e73a 524 }
0085bd51
RH
525 }
526 assert (ms0 == 32);
fd76e73a 527
0085bd51
RH
528 if (ls1 == 32) {
529 tcg_out_extr(s, ret, arg, 0, ls0, 0);
f54b3f92 530 } else {
0085bd51
RH
531 tcg_out_mov(s, ret, arg);
532 tcg_out32(s, INSN_DEPI | INSN_R2(ret) | INSN_IM5(0)
533 | INSN_SHDEP_CP(31 - ls0) | INSN_DEP_LEN(ls1 - ls0));
f54b3f92
AJ
534 }
535}
536
fd76e73a 537static inline void tcg_out_ext8s(TCGContext *s, int ret, int arg)
f54b3f92 538{
fd76e73a 539 tcg_out_extr(s, ret, arg, 0, 8, 1);
f54b3f92
AJ
540}
541
fd76e73a 542static inline void tcg_out_ext16s(TCGContext *s, int ret, int arg)
f54b3f92 543{
fd76e73a 544 tcg_out_extr(s, ret, arg, 0, 16, 1);
f54b3f92
AJ
545}
546
fd76e73a 547static void tcg_out_shli(TCGContext *s, int ret, int arg, int count)
f54b3f92 548{
fd76e73a
RH
549 count &= 31;
550 tcg_out32(s, INSN_ZDEP | INSN_R2(ret) | INSN_R1(arg)
551 | INSN_SHDEP_CP(31 - count) | INSN_DEP_LEN(32 - count));
f54b3f92
AJ
552}
553
fd76e73a 554static void tcg_out_shl(TCGContext *s, int ret, int arg, int creg)
f54b3f92 555{
fd76e73a
RH
556 tcg_out_arithi(s, TCG_REG_R20, creg, 31, INSN_SUBI);
557 tcg_out_mtctl_sar(s, TCG_REG_R20);
558 tcg_out32(s, INSN_ZVDEP | INSN_R2(ret) | INSN_R1(arg) | INSN_DEP_LEN(32));
f54b3f92
AJ
559}
560
fd76e73a 561static void tcg_out_shri(TCGContext *s, int ret, int arg, int count)
f54b3f92 562{
fd76e73a
RH
563 count &= 31;
564 tcg_out_extr(s, ret, arg, count, 32 - count, 0);
f54b3f92
AJ
565}
566
fd76e73a 567static void tcg_out_shr(TCGContext *s, int ret, int arg, int creg)
f54b3f92 568{
fd76e73a 569 tcg_out_vshd(s, ret, TCG_REG_R0, arg, creg);
f54b3f92
AJ
570}
571
fd76e73a 572static void tcg_out_sari(TCGContext *s, int ret, int arg, int count)
f54b3f92 573{
fd76e73a
RH
574 count &= 31;
575 tcg_out_extr(s, ret, arg, count, 32 - count, 1);
f54b3f92
AJ
576}
577
fd76e73a 578static void tcg_out_sar(TCGContext *s, int ret, int arg, int creg)
f54b3f92 579{
fd76e73a
RH
580 tcg_out_arithi(s, TCG_REG_R20, creg, 31, INSN_SUBI);
581 tcg_out_mtctl_sar(s, TCG_REG_R20);
582 tcg_out32(s, INSN_VEXTRS | INSN_R1(ret) | INSN_R2(arg) | INSN_DEP_LEN(32));
f54b3f92
AJ
583}
584
fd76e73a 585static void tcg_out_rotli(TCGContext *s, int ret, int arg, int count)
f54b3f92 586{
fd76e73a
RH
587 count &= 31;
588 tcg_out_shd(s, ret, arg, arg, 32 - count);
f54b3f92
AJ
589}
590
fd76e73a
RH
591static void tcg_out_rotl(TCGContext *s, int ret, int arg, int creg)
592{
593 tcg_out_arithi(s, TCG_REG_R20, creg, 32, INSN_SUBI);
594 tcg_out_vshd(s, ret, arg, arg, TCG_REG_R20);
f54b3f92
AJ
595}
596
fd76e73a
RH
597static void tcg_out_rotri(TCGContext *s, int ret, int arg, int count)
598{
599 count &= 31;
600 tcg_out_shd(s, ret, arg, arg, count);
f54b3f92
AJ
601}
602
fd76e73a
RH
603static void tcg_out_rotr(TCGContext *s, int ret, int arg, int creg)
604{
605 tcg_out_vshd(s, ret, arg, arg, creg);
f54b3f92
AJ
606}
607
fd76e73a
RH
608static void tcg_out_bswap16(TCGContext *s, int ret, int arg, int sign)
609{
610 if (ret != arg) {
611 tcg_out_mov(s, ret, arg); /* arg = xxAB */
612 }
613 tcg_out_dep(s, ret, ret, 16, 8); /* ret = xBAB */
614 tcg_out_extr(s, ret, ret, 8, 16, sign); /* ret = ..BA */
f54b3f92
AJ
615}
616
fd76e73a 617static void tcg_out_bswap32(TCGContext *s, int ret, int arg, int temp)
f54b3f92 618{
fd76e73a
RH
619 /* arg = ABCD */
620 tcg_out_rotri(s, temp, arg, 16); /* temp = CDAB */
621 tcg_out_dep(s, temp, temp, 16, 8); /* temp = CBAB */
622 tcg_out_shd(s, ret, arg, temp, 8); /* ret = DCBA */
f54b3f92
AJ
623}
624
fd76e73a
RH
625static void tcg_out_call(TCGContext *s, void *func)
626{
627 tcg_target_long val, hi, lo, disp;
628
629 val = (uint32_t)__canonicalize_funcptr_for_compare(func);
630 disp = (val - ((tcg_target_long)s->code_ptr + 8)) >> 2;
631
632 if (check_fit_tl(disp, 17)) {
633 tcg_out32(s, INSN_BL_N | INSN_R2(TCG_REG_RP) | reassemble_17(disp));
634 } else {
635 hi = val >> 11;
636 lo = val & 0x7ff;
637
638 tcg_out32(s, INSN_LDIL | INSN_R2(TCG_REG_R20) | reassemble_21(hi));
639 tcg_out32(s, INSN_BLE_SR4 | INSN_R2(TCG_REG_R20)
640 | reassemble_17(lo >> 2));
641 tcg_out_mov(s, TCG_REG_RP, TCG_REG_R31);
642 }
643}
79383c9c 644
fd76e73a
RH
645static void tcg_out_xmpyu(TCGContext *s, int retl, int reth,
646 int arg1, int arg2)
647{
648 /* Store both words into the stack for copy to the FPU. */
649 tcg_out_ldst(s, arg1, TCG_REG_SP, STACK_TEMP_OFS, INSN_STW);
650 tcg_out_ldst(s, arg2, TCG_REG_SP, STACK_TEMP_OFS + 4, INSN_STW);
651
652 /* Load both words into the FPU at the same time. We get away
653 with this because we can address the left and right half of the
654 FPU registers individually once loaded. */
655 /* fldds stack_temp(sp),fr22 */
656 tcg_out32(s, INSN_FLDDS | INSN_R2(TCG_REG_SP)
657 | INSN_IM5(STACK_TEMP_OFS) | INSN_T(22));
658
659 /* xmpyu fr22r,fr22,fr22 */
660 tcg_out32(s, 0x3ad64796);
661
662 /* Store the 64-bit result back into the stack. */
663 /* fstds stack_temp(sp),fr22 */
664 tcg_out32(s, INSN_FSTDS | INSN_R2(TCG_REG_SP)
665 | INSN_IM5(STACK_TEMP_OFS) | INSN_T(22));
666
667 /* Load the pieces of the result that the caller requested. */
668 if (reth) {
669 tcg_out_ldst(s, reth, TCG_REG_SP, STACK_TEMP_OFS, INSN_LDW);
670 }
671 if (retl) {
672 tcg_out_ldst(s, retl, TCG_REG_SP, STACK_TEMP_OFS + 4, INSN_LDW);
673 }
674}
675
91493631
RH
676static void tcg_out_add2(TCGContext *s, int destl, int desth,
677 int al, int ah, int bl, int bh, int blconst)
678{
679 int tmp = (destl == ah || destl == bh ? TCG_REG_R20 : destl);
680
681 if (blconst) {
682 tcg_out_arithi(s, tmp, al, bl, INSN_ADDI);
683 } else {
684 tcg_out_arith(s, tmp, al, bl, INSN_ADD);
685 }
686 tcg_out_arith(s, desth, ah, bh, INSN_ADDC);
687
688 tcg_out_mov(s, destl, tmp);
689}
690
691static void tcg_out_sub2(TCGContext *s, int destl, int desth, int al, int ah,
692 int bl, int bh, int alconst, int blconst)
693{
694 int tmp = (destl == ah || destl == bh ? TCG_REG_R20 : destl);
695
696 if (alconst) {
697 if (blconst) {
698 tcg_out_movi(s, TCG_TYPE_I32, TCG_REG_R20, bl);
699 bl = TCG_REG_R20;
700 }
701 tcg_out_arithi(s, tmp, bl, al, INSN_SUBI);
702 } else if (blconst) {
703 tcg_out_arithi(s, tmp, al, -bl, INSN_ADDI);
704 } else {
705 tcg_out_arith(s, tmp, al, bl, INSN_SUB);
706 }
707 tcg_out_arith(s, desth, ah, bh, INSN_SUBB);
708
709 tcg_out_mov(s, destl, tmp);
710}
711
fd76e73a
RH
712static void tcg_out_branch(TCGContext *s, int label_index, int nul)
713{
714 TCGLabel *l = &s->labels[label_index];
715 uint32_t op = nul ? INSN_BL_N : INSN_BL;
716
717 if (l->has_value) {
718 tcg_target_long val = l->u.value;
719
720 val -= (tcg_target_long)s->code_ptr + 8;
721 val >>= 2;
722 assert(check_fit_tl(val, 17));
723
724 tcg_out32(s, op | reassemble_17(val));
725 } else {
2d097a83
RH
726 /* We need to keep the offset unchanged for retranslation. */
727 uint32_t old_insn = *(uint32_t *)s->code_ptr;
728
fd76e73a 729 tcg_out_reloc(s, s->code_ptr, R_PARISC_PCREL17F, label_index, 0);
2d097a83 730 tcg_out32(s, op | (old_insn & 0x1f1ffdu));
fd76e73a
RH
731 }
732}
733
734static const uint8_t tcg_cond_to_cmp_cond[10] =
735{
736 [TCG_COND_EQ] = COND_EQ,
737 [TCG_COND_NE] = COND_EQ | COND_FALSE,
738 [TCG_COND_LT] = COND_LT,
739 [TCG_COND_GE] = COND_LT | COND_FALSE,
740 [TCG_COND_LE] = COND_LE,
741 [TCG_COND_GT] = COND_LE | COND_FALSE,
742 [TCG_COND_LTU] = COND_LTU,
743 [TCG_COND_GEU] = COND_LTU | COND_FALSE,
744 [TCG_COND_LEU] = COND_LEU,
745 [TCG_COND_GTU] = COND_LEU | COND_FALSE,
746};
747
748static void tcg_out_brcond(TCGContext *s, int cond, TCGArg c1,
749 TCGArg c2, int c2const, int label_index)
750{
751 TCGLabel *l = &s->labels[label_index];
752 int op, pacond;
753
754 /* Note that COMIB operates as if the immediate is the first
755 operand. We model brcond with the immediate in the second
756 to better match what targets are likely to give us. For
757 consistency, model COMB with reversed operands as well. */
758 pacond = tcg_cond_to_cmp_cond[tcg_swap_cond(cond)];
759
760 if (c2const) {
761 op = (pacond & COND_FALSE ? INSN_COMIBF : INSN_COMIBT);
762 op |= INSN_IM5(c2);
763 } else {
764 op = (pacond & COND_FALSE ? INSN_COMBF : INSN_COMBT);
765 op |= INSN_R1(c2);
766 }
767 op |= INSN_R2(c1);
768 op |= INSN_COND(pacond & 7);
769
770 if (l->has_value) {
771 tcg_target_long val = l->u.value;
772
773 val -= (tcg_target_long)s->code_ptr + 8;
774 val >>= 2;
775 assert(check_fit_tl(val, 12));
776
777 /* ??? Assume that all branches to defined labels are backward.
778 Which means that if the nul bit is set, the delay slot is
779 executed if the branch is taken, and not executed in fallthru. */
780 tcg_out32(s, op | reassemble_12(val));
781 tcg_out_nop(s);
782 } else {
2d097a83
RH
783 /* We need to keep the offset unchanged for retranslation. */
784 uint32_t old_insn = *(uint32_t *)s->code_ptr;
785
fd76e73a
RH
786 tcg_out_reloc(s, s->code_ptr, R_PARISC_PCREL12F, label_index, 0);
787 /* ??? Assume that all branches to undefined labels are forward.
788 Which means that if the nul bit is set, the delay slot is
789 not executed if the branch is taken, which is what we want. */
2d097a83 790 tcg_out32(s, op | 2 | (old_insn & 0x1ffdu));
fd76e73a
RH
791 }
792}
793
794static void tcg_out_comclr(TCGContext *s, int cond, TCGArg ret,
795 TCGArg c1, TCGArg c2, int c2const)
796{
797 int op, pacond;
798
799 /* Note that COMICLR operates as if the immediate is the first
800 operand. We model setcond with the immediate in the second
801 to better match what targets are likely to give us. For
802 consistency, model COMCLR with reversed operands as well. */
803 pacond = tcg_cond_to_cmp_cond[tcg_swap_cond(cond)];
804
805 if (c2const) {
806 op = INSN_COMICLR | INSN_R2(c1) | INSN_R1(ret) | INSN_IM11(c2);
807 } else {
808 op = INSN_COMCLR | INSN_R2(c1) | INSN_R1(c2) | INSN_T(ret);
809 }
810 op |= INSN_COND(pacond & 7);
811 op |= pacond & COND_FALSE ? 1 << 12 : 0;
812
813 tcg_out32(s, op);
814}
815
816static void tcg_out_brcond2(TCGContext *s, int cond, TCGArg al, TCGArg ah,
817 TCGArg bl, int blconst, TCGArg bh, int bhconst,
818 int label_index)
819{
820 switch (cond) {
821 case TCG_COND_EQ:
822 case TCG_COND_NE:
823 tcg_out_comclr(s, tcg_invert_cond(cond), TCG_REG_R0, al, bl, blconst);
824 tcg_out_brcond(s, cond, ah, bh, bhconst, label_index);
825 break;
826
827 default:
828 tcg_out_brcond(s, cond, ah, bh, bhconst, label_index);
829 tcg_out_comclr(s, TCG_COND_NE, TCG_REG_R0, ah, bh, bhconst);
830 tcg_out_brcond(s, tcg_unsigned_cond(cond),
831 al, bl, blconst, label_index);
832 break;
833 }
834}
835
836static void tcg_out_setcond(TCGContext *s, int cond, TCGArg ret,
837 TCGArg c1, TCGArg c2, int c2const)
838{
839 tcg_out_comclr(s, tcg_invert_cond(cond), ret, c1, c2, c2const);
840 tcg_out_movi(s, TCG_TYPE_I32, ret, 1);
841}
842
843static void tcg_out_setcond2(TCGContext *s, int cond, TCGArg ret,
844 TCGArg al, TCGArg ah, TCGArg bl, int blconst,
845 TCGArg bh, int bhconst)
846{
847 int scratch = TCG_REG_R20;
848
849 if (ret != al && ret != ah
850 && (blconst || ret != bl)
851 && (bhconst || ret != bh)) {
852 scratch = ret;
853 }
854
855 switch (cond) {
856 case TCG_COND_EQ:
857 case TCG_COND_NE:
858 tcg_out_setcond(s, cond, scratch, al, bl, blconst);
859 tcg_out_comclr(s, TCG_COND_EQ, TCG_REG_R0, ah, bh, bhconst);
860 tcg_out_movi(s, TCG_TYPE_I32, scratch, cond == TCG_COND_NE);
861 break;
862
863 default:
864 tcg_out_setcond(s, tcg_unsigned_cond(cond), scratch, al, bl, blconst);
865 tcg_out_comclr(s, TCG_COND_EQ, TCG_REG_R0, ah, bh, bhconst);
866 tcg_out_movi(s, TCG_TYPE_I32, scratch, 0);
867 tcg_out_comclr(s, cond, TCG_REG_R0, ah, bh, bhconst);
868 tcg_out_movi(s, TCG_TYPE_I32, scratch, 1);
869 break;
870 }
871
872 tcg_out_mov(s, ret, scratch);
873}
874
875#if defined(CONFIG_SOFTMMU)
79383c9c 876#include "../../softmmu_defs.h"
f54b3f92
AJ
877
878static void *qemu_ld_helpers[4] = {
879 __ldb_mmu,
880 __ldw_mmu,
881 __ldl_mmu,
882 __ldq_mmu,
883};
884
885static void *qemu_st_helpers[4] = {
886 __stb_mmu,
887 __stw_mmu,
888 __stl_mmu,
889 __stq_mmu,
890};
fd76e73a
RH
891
892/* Load and compare a TLB entry, and branch if TLB miss. OFFSET is set to
893 the offset of the first ADDR_READ or ADDR_WRITE member of the appropriate
894 TLB for the memory index. The return value is the offset from ENV
895 contained in R1 afterward (to be used when loading ADDEND); if the
896 return value is 0, R1 is not used. */
897
898static int tcg_out_tlb_read(TCGContext *s, int r0, int r1, int addrlo,
899 int addrhi, int s_bits, int lab_miss, int offset)
900{
901 int ret;
902
903 /* Extracting the index into the TLB. The "normal C operation" is
904 r1 = addr_reg >> TARGET_PAGE_BITS;
905 r1 &= CPU_TLB_SIZE - 1;
906 r1 <<= CPU_TLB_ENTRY_BITS;
907 What this does is extract CPU_TLB_BITS beginning at TARGET_PAGE_BITS
908 and place them at CPU_TLB_ENTRY_BITS. We can combine the first two
909 operations with an EXTRU. Unfortunately, the current value of
910 CPU_TLB_ENTRY_BITS is > 3, so we can't merge that shift with the
911 add that follows. */
912 tcg_out_extr(s, r1, addrlo, TARGET_PAGE_BITS, CPU_TLB_BITS, 0);
fd76e73a
RH
913 tcg_out_shli(s, r1, r1, CPU_TLB_ENTRY_BITS);
914 tcg_out_arith(s, r1, r1, TCG_AREG0, INSN_ADDL);
915
916 /* Make sure that both the addr_{read,write} and addend can be
917 read with a 14-bit offset from the same base register. */
918 if (check_fit_tl(offset + CPU_TLB_SIZE, 14)) {
919 ret = 0;
920 } else {
921 ret = (offset + 0x400) & ~0x7ff;
922 offset = ret - offset;
923 tcg_out_addi2(s, TCG_REG_R1, r1, ret);
924 r1 = TCG_REG_R1;
925 }
926
927 /* Load the entry from the computed slot. */
928 if (TARGET_LONG_BITS == 64) {
929 tcg_out_ld(s, TCG_TYPE_PTR, TCG_REG_R23, r1, offset);
930 tcg_out_ld(s, TCG_TYPE_PTR, TCG_REG_R20, r1, offset + 4);
931 } else {
932 tcg_out_ld(s, TCG_TYPE_PTR, TCG_REG_R20, r1, offset);
933 }
934
739734cb
RH
935 /* Compute the value that ought to appear in the TLB for a hit, namely, the page
936 of the address. We include the low N bits of the address to catch unaligned
937 accesses and force them onto the slow path. Do this computation after having
938 issued the load from the TLB slot to give the load time to complete. */
939 tcg_out_andi(s, r0, addrlo, TARGET_PAGE_MASK | ((1 << s_bits) - 1));
940
fd76e73a
RH
941 /* If not equal, jump to lab_miss. */
942 if (TARGET_LONG_BITS == 64) {
943 tcg_out_brcond2(s, TCG_COND_NE, TCG_REG_R20, TCG_REG_R23,
944 r0, 0, addrhi, 0, lab_miss);
945 } else {
946 tcg_out_brcond(s, TCG_COND_NE, TCG_REG_R20, r0, 0, lab_miss);
947 }
948
949 return ret;
950}
f54b3f92
AJ
951#endif
952
f061b40e
RH
953static void tcg_out_qemu_ld_direct(TCGContext *s, int datalo_reg, int datahi_reg,
954 int addr_reg, int addend_reg, int opc)
f54b3f92 955{
f54b3f92 956#ifdef TARGET_WORDS_BIGENDIAN
f061b40e 957 const int bswap = 0;
f54b3f92 958#else
f061b40e 959 const int bswap = 1;
f54b3f92 960#endif
f061b40e 961
f54b3f92 962 switch (opc) {
fd76e73a 963 case 0:
f061b40e 964 tcg_out_ldst_index(s, datalo_reg, addr_reg, addend_reg, INSN_LDBX);
fd76e73a
RH
965 break;
966 case 0 | 4:
f061b40e
RH
967 tcg_out_ldst_index(s, datalo_reg, addr_reg, addend_reg, INSN_LDBX);
968 tcg_out_ext8s(s, datalo_reg, datalo_reg);
fd76e73a
RH
969 break;
970 case 1:
f061b40e 971 tcg_out_ldst_index(s, datalo_reg, addr_reg, addend_reg, INSN_LDHX);
fd76e73a 972 if (bswap) {
f061b40e 973 tcg_out_bswap16(s, datalo_reg, datalo_reg, 0);
fd76e73a
RH
974 }
975 break;
976 case 1 | 4:
f061b40e 977 tcg_out_ldst_index(s, datalo_reg, addr_reg, addend_reg, INSN_LDHX);
fd76e73a 978 if (bswap) {
f061b40e 979 tcg_out_bswap16(s, datalo_reg, datalo_reg, 1);
fd76e73a 980 } else {
f061b40e 981 tcg_out_ext16s(s, datalo_reg, datalo_reg);
fd76e73a
RH
982 }
983 break;
984 case 2:
f061b40e 985 tcg_out_ldst_index(s, datalo_reg, addr_reg, addend_reg, INSN_LDWX);
fd76e73a 986 if (bswap) {
f061b40e 987 tcg_out_bswap32(s, datalo_reg, datalo_reg, TCG_REG_R20);
fd76e73a
RH
988 }
989 break;
990 case 3:
991 if (bswap) {
f061b40e
RH
992 int t = datahi_reg;
993 datahi_reg = datalo_reg;
994 datalo_reg = t;
fd76e73a 995 }
f061b40e
RH
996 /* We can't access the low-part with a reg+reg addressing mode,
997 so perform the addition now and use reg_ofs addressing mode. */
998 if (addend_reg != TCG_REG_R0) {
999 tcg_out_arith(s, TCG_REG_R20, addr_reg, addend_reg, INSN_ADD);
1000 addr_reg = TCG_REG_R20;
1001 }
1002 /* Make sure not to clobber the base register. */
1003 if (datahi_reg == addr_reg) {
1004 tcg_out_ldst(s, datalo_reg, addr_reg, 4, INSN_LDW);
1005 tcg_out_ldst(s, datahi_reg, addr_reg, 0, INSN_LDW);
fd76e73a 1006 } else {
f061b40e
RH
1007 tcg_out_ldst(s, datahi_reg, addr_reg, 0, INSN_LDW);
1008 tcg_out_ldst(s, datalo_reg, addr_reg, 4, INSN_LDW);
fd76e73a
RH
1009 }
1010 if (bswap) {
f061b40e
RH
1011 tcg_out_bswap32(s, datalo_reg, datalo_reg, TCG_REG_R20);
1012 tcg_out_bswap32(s, datahi_reg, datahi_reg, TCG_REG_R20);
fd76e73a
RH
1013 }
1014 break;
1015 default:
1016 tcg_abort();
f54b3f92 1017 }
f061b40e
RH
1018}
1019
1020static void tcg_out_qemu_ld(TCGContext *s, const TCGArg *args, int opc)
1021{
1022 int datalo_reg = *args++;
1023 /* Note that datahi_reg is only used for 64-bit loads. */
1024 int datahi_reg = (opc == 3 ? *args++ : TCG_REG_R0);
1025 int addrlo_reg = *args++;
f54b3f92
AJ
1026
1027#if defined(CONFIG_SOFTMMU)
f061b40e
RH
1028 /* Note that addrhi_reg is only used for 64-bit guests. */
1029 int addrhi_reg = (TARGET_LONG_BITS == 64 ? *args++ : TCG_REG_R0);
1030 int mem_index = *args;
1031 int lab1, lab2, argreg, offset;
1032
1033 lab1 = gen_new_label();
1034 lab2 = gen_new_label();
1035
1036 offset = offsetof(CPUState, tlb_table[mem_index][0].addr_read);
1037 offset = tcg_out_tlb_read(s, TCG_REG_R26, TCG_REG_R25, addrlo_reg, addrhi_reg,
1038 opc & 3, lab1, offset);
1039
1040 /* TLB Hit. */
1041 tcg_out_ld(s, TCG_TYPE_PTR, TCG_REG_R20, (offset ? TCG_REG_R1 : TCG_REG_R25),
1042 offsetof(CPUState, tlb_table[mem_index][0].addend) - offset);
1043 tcg_out_qemu_ld_direct(s, datalo_reg, datahi_reg, addrlo_reg, TCG_REG_R20, opc);
fd76e73a
RH
1044 tcg_out_branch(s, lab2, 1);
1045
1046 /* TLB Miss. */
1047 /* label1: */
1048 tcg_out_label(s, lab1, (tcg_target_long)s->code_ptr);
1049
1050 argreg = TCG_REG_R26;
f061b40e 1051 tcg_out_mov(s, argreg--, addrlo_reg);
fd76e73a 1052 if (TARGET_LONG_BITS == 64) {
f061b40e 1053 tcg_out_mov(s, argreg--, addrhi_reg);
fd76e73a
RH
1054 }
1055 tcg_out_movi(s, TCG_TYPE_I32, argreg, mem_index);
1056
f061b40e 1057 tcg_out_call(s, qemu_ld_helpers[opc & 3]);
fd76e73a
RH
1058
1059 switch (opc) {
1060 case 0:
f061b40e 1061 tcg_out_andi(s, datalo_reg, TCG_REG_RET0, 0xff);
fd76e73a
RH
1062 break;
1063 case 0 | 4:
f061b40e 1064 tcg_out_ext8s(s, datalo_reg, TCG_REG_RET0);
fd76e73a
RH
1065 break;
1066 case 1:
f061b40e 1067 tcg_out_andi(s, datalo_reg, TCG_REG_RET0, 0xffff);
fd76e73a
RH
1068 break;
1069 case 1 | 4:
f061b40e 1070 tcg_out_ext16s(s, datalo_reg, TCG_REG_RET0);
fd76e73a
RH
1071 break;
1072 case 2:
1073 case 2 | 4:
f061b40e 1074 tcg_out_mov(s, datalo_reg, TCG_REG_RET0);
fd76e73a
RH
1075 break;
1076 case 3:
f061b40e
RH
1077 tcg_out_mov(s, datahi_reg, TCG_REG_RET0);
1078 tcg_out_mov(s, datalo_reg, TCG_REG_RET1);
fd76e73a
RH
1079 break;
1080 default:
1081 tcg_abort();
1082 }
1083
f54b3f92 1084 /* label2: */
fd76e73a 1085 tcg_out_label(s, lab2, (tcg_target_long)s->code_ptr);
f061b40e
RH
1086#else
1087 tcg_out_qemu_ld_direct(s, datalo_reg, datahi_reg, addrlo_reg,
1088 (GUEST_BASE ? TCG_GUEST_BASE_REG : TCG_REG_R0), opc);
f54b3f92
AJ
1089#endif
1090}
1091
f061b40e
RH
1092static void tcg_out_qemu_st_direct(TCGContext *s, int datalo_reg, int datahi_reg,
1093 int addr_reg, int opc)
f54b3f92 1094{
f54b3f92 1095#ifdef TARGET_WORDS_BIGENDIAN
f061b40e 1096 const int bswap = 0;
f54b3f92 1097#else
f061b40e 1098 const int bswap = 1;
f54b3f92 1099#endif
f061b40e 1100
f54b3f92
AJ
1101 switch (opc) {
1102 case 0:
f061b40e 1103 tcg_out_ldst(s, datalo_reg, addr_reg, 0, INSN_STB);
f54b3f92
AJ
1104 break;
1105 case 1:
1106 if (bswap) {
f061b40e
RH
1107 tcg_out_bswap16(s, TCG_REG_R20, datalo_reg, 0);
1108 datalo_reg = TCG_REG_R20;
f54b3f92 1109 }
f061b40e 1110 tcg_out_ldst(s, datalo_reg, addr_reg, 0, INSN_STH);
f54b3f92
AJ
1111 break;
1112 case 2:
1113 if (bswap) {
f061b40e
RH
1114 tcg_out_bswap32(s, TCG_REG_R20, datalo_reg, TCG_REG_R20);
1115 datalo_reg = TCG_REG_R20;
f54b3f92 1116 }
f061b40e 1117 tcg_out_ldst(s, datalo_reg, addr_reg, 0, INSN_STW);
f54b3f92
AJ
1118 break;
1119 case 3:
fd76e73a 1120 if (bswap) {
f061b40e
RH
1121 tcg_out_bswap32(s, TCG_REG_R20, datalo_reg, TCG_REG_R20);
1122 tcg_out_bswap32(s, TCG_REG_R23, datahi_reg, TCG_REG_R23);
1123 datahi_reg = TCG_REG_R20;
1124 datalo_reg = TCG_REG_R23;
f54b3f92 1125 }
f061b40e
RH
1126 tcg_out_ldst(s, datahi_reg, addr_reg, 0, INSN_STW);
1127 tcg_out_ldst(s, datalo_reg, addr_reg, 4, INSN_STW);
f54b3f92
AJ
1128 break;
1129 default:
1130 tcg_abort();
1131 }
1132
f061b40e
RH
1133}
1134
1135static void tcg_out_qemu_st(TCGContext *s, const TCGArg *args, int opc)
1136{
1137 int datalo_reg = *args++;
1138 /* Note that datahi_reg is only used for 64-bit loads. */
1139 int datahi_reg = (opc == 3 ? *args++ : TCG_REG_R0);
1140 int addrlo_reg = *args++;
1141
f54b3f92 1142#if defined(CONFIG_SOFTMMU)
f061b40e
RH
1143 /* Note that addrhi_reg is only used for 64-bit guests. */
1144 int addrhi_reg = (TARGET_LONG_BITS == 64 ? *args++ : TCG_REG_R0);
1145 int mem_index = *args;
1146 int lab1, lab2, argreg, offset;
1147
1148 lab1 = gen_new_label();
1149 lab2 = gen_new_label();
1150
1151 offset = offsetof(CPUState, tlb_table[mem_index][0].addr_write);
1152 offset = tcg_out_tlb_read(s, TCG_REG_R26, TCG_REG_R25, addrlo_reg, addrhi_reg,
1153 opc, lab1, offset);
1154
1155 /* TLB Hit. */
1156 tcg_out_ld(s, TCG_TYPE_PTR, TCG_REG_R20, (offset ? TCG_REG_R1 : TCG_REG_R25),
1157 offsetof(CPUState, tlb_table[mem_index][0].addend) - offset);
1158
1159 /* There are no indexed stores, so we must do this addition explitly.
1160 Careful to avoid R20, which is used for the bswaps to follow. */
1161 tcg_out_arith(s, TCG_REG_R31, addrlo_reg, TCG_REG_R20, INSN_ADDL);
1162 tcg_out_qemu_st_direct(s, datalo_reg, datahi_reg, TCG_REG_R31, opc);
fd76e73a
RH
1163 tcg_out_branch(s, lab2, 1);
1164
1165 /* TLB Miss. */
1166 /* label1: */
1167 tcg_out_label(s, lab1, (tcg_target_long)s->code_ptr);
1168
1169 argreg = TCG_REG_R26;
f061b40e 1170 tcg_out_mov(s, argreg--, addrlo_reg);
fd76e73a 1171 if (TARGET_LONG_BITS == 64) {
f061b40e 1172 tcg_out_mov(s, argreg--, addrhi_reg);
fd76e73a
RH
1173 }
1174
1175 switch(opc) {
1176 case 0:
f061b40e 1177 tcg_out_andi(s, argreg--, datalo_reg, 0xff);
fd76e73a
RH
1178 tcg_out_movi(s, TCG_TYPE_I32, argreg, mem_index);
1179 break;
1180 case 1:
f061b40e 1181 tcg_out_andi(s, argreg--, datalo_reg, 0xffff);
fd76e73a
RH
1182 tcg_out_movi(s, TCG_TYPE_I32, argreg, mem_index);
1183 break;
1184 case 2:
f061b40e 1185 tcg_out_mov(s, argreg--, datalo_reg);
fd76e73a
RH
1186 tcg_out_movi(s, TCG_TYPE_I32, argreg, mem_index);
1187 break;
1188 case 3:
1189 /* Because of the alignment required by the 64-bit data argument,
1190 we will always use R23/R24. Also, we will always run out of
1191 argument registers for storing mem_index, so that will have
1192 to go on the stack. */
1193 if (mem_index == 0) {
1194 argreg = TCG_REG_R0;
1195 } else {
1196 argreg = TCG_REG_R20;
1197 tcg_out_movi(s, TCG_TYPE_I32, argreg, mem_index);
1198 }
f061b40e
RH
1199 tcg_out_mov(s, TCG_REG_R23, datahi_reg);
1200 tcg_out_mov(s, TCG_REG_R24, datalo_reg);
fd76e73a
RH
1201 tcg_out_st(s, TCG_TYPE_I32, argreg, TCG_REG_SP,
1202 TCG_TARGET_CALL_STACK_OFFSET - 4);
1203 break;
1204 default:
1205 tcg_abort();
1206 }
1207
f061b40e 1208 tcg_out_call(s, qemu_st_helpers[opc]);
fd76e73a 1209
f54b3f92 1210 /* label2: */
fd76e73a 1211 tcg_out_label(s, lab2, (tcg_target_long)s->code_ptr);
f061b40e
RH
1212#else
1213 /* There are no indexed stores, so if GUEST_BASE is set we must do the add
1214 explicitly. Careful to avoid R20, which is used for the bswaps to follow. */
1215 if (GUEST_BASE != 0) {
1216 tcg_out_arith(s, TCG_REG_R31, addrlo_reg, TCG_GUEST_BASE_REG, INSN_ADDL);
1217 addrlo_reg = TCG_REG_R31;
1218 }
1219 tcg_out_qemu_st_direct(s, datalo_reg, datahi_reg, addrlo_reg, opc);
f54b3f92
AJ
1220#endif
1221}
1222
fd76e73a
RH
1223static void tcg_out_exit_tb(TCGContext *s, TCGArg arg)
1224{
1225 if (!check_fit_tl(arg, 14)) {
1226 uint32_t hi, lo;
1227 hi = arg & ~0x7ff;
1228 lo = arg & 0x7ff;
1229 if (lo) {
1230 tcg_out_movi(s, TCG_TYPE_PTR, TCG_REG_RET0, hi);
1231 tcg_out32(s, INSN_BV | INSN_R2(TCG_REG_R18));
1232 tcg_out_addi(s, TCG_REG_RET0, lo);
1233 return;
1234 }
1235 arg = hi;
1236 }
1237 tcg_out32(s, INSN_BV | INSN_R2(TCG_REG_R18));
1238 tcg_out_movi(s, TCG_TYPE_PTR, TCG_REG_RET0, arg);
1239}
1240
1241static void tcg_out_goto_tb(TCGContext *s, TCGArg arg)
1242{
1243 if (s->tb_jmp_offset) {
1244 /* direct jump method */
1245 fprintf(stderr, "goto_tb direct\n");
1246 tcg_abort();
1247 } else {
1248 /* indirect jump method */
1249 tcg_out_ld(s, TCG_TYPE_PTR, TCG_REG_R20, TCG_REG_R0,
1250 (tcg_target_long)(s->tb_next + arg));
1251 tcg_out32(s, INSN_BV_N | INSN_R2(TCG_REG_R20));
1252 }
1253 s->tb_next_offset[arg] = s->code_ptr - s->code_buf;
1254}
1255
a9751609 1256static inline void tcg_out_op(TCGContext *s, TCGOpcode opc, const TCGArg *args,
f54b3f92
AJ
1257 const int *const_args)
1258{
f54b3f92
AJ
1259 switch (opc) {
1260 case INDEX_op_exit_tb:
fd76e73a 1261 tcg_out_exit_tb(s, args[0]);
f54b3f92
AJ
1262 break;
1263 case INDEX_op_goto_tb:
fd76e73a 1264 tcg_out_goto_tb(s, args[0]);
f54b3f92 1265 break;
fd76e73a 1266
f54b3f92 1267 case INDEX_op_call:
fd76e73a
RH
1268 if (const_args[0]) {
1269 tcg_out_call(s, (void *)args[0]);
1270 } else {
3e1f46ea
RH
1271 /* ??? FIXME: the value in the register in args[0] is almost
1272 certainly a procedure descriptor, not a code address. We
1273 probably need to use the millicode $$dyncall routine. */
1274 tcg_abort();
fd76e73a 1275 }
f54b3f92 1276 break;
fd76e73a 1277
f54b3f92
AJ
1278 case INDEX_op_jmp:
1279 fprintf(stderr, "unimplemented jmp\n");
1280 tcg_abort();
1281 break;
fd76e73a 1282
f54b3f92 1283 case INDEX_op_br:
fd76e73a 1284 tcg_out_branch(s, args[0], 1);
f54b3f92 1285 break;
fd76e73a 1286
f54b3f92
AJ
1287 case INDEX_op_movi_i32:
1288 tcg_out_movi(s, TCG_TYPE_I32, args[0], (uint32_t)args[1]);
1289 break;
1290
1291 case INDEX_op_ld8u_i32:
fd76e73a 1292 tcg_out_ldst(s, args[0], args[1], args[2], INSN_LDB);
f54b3f92
AJ
1293 break;
1294 case INDEX_op_ld8s_i32:
fd76e73a 1295 tcg_out_ldst(s, args[0], args[1], args[2], INSN_LDB);
f54b3f92
AJ
1296 tcg_out_ext8s(s, args[0], args[0]);
1297 break;
1298 case INDEX_op_ld16u_i32:
fd76e73a 1299 tcg_out_ldst(s, args[0], args[1], args[2], INSN_LDH);
f54b3f92
AJ
1300 break;
1301 case INDEX_op_ld16s_i32:
fd76e73a 1302 tcg_out_ldst(s, args[0], args[1], args[2], INSN_LDH);
f54b3f92
AJ
1303 tcg_out_ext16s(s, args[0], args[0]);
1304 break;
1305 case INDEX_op_ld_i32:
fd76e73a 1306 tcg_out_ldst(s, args[0], args[1], args[2], INSN_LDW);
f54b3f92
AJ
1307 break;
1308
1309 case INDEX_op_st8_i32:
fd76e73a 1310 tcg_out_ldst(s, args[0], args[1], args[2], INSN_STB);
f54b3f92
AJ
1311 break;
1312 case INDEX_op_st16_i32:
fd76e73a 1313 tcg_out_ldst(s, args[0], args[1], args[2], INSN_STH);
f54b3f92
AJ
1314 break;
1315 case INDEX_op_st_i32:
fd76e73a
RH
1316 tcg_out_ldst(s, args[0], args[1], args[2], INSN_STW);
1317 break;
1318
1319 case INDEX_op_add_i32:
1320 if (const_args[2]) {
1321 tcg_out_addi2(s, args[0], args[1], args[2]);
1322 } else {
1323 tcg_out_arith(s, args[0], args[1], args[2], INSN_ADDL);
1324 }
f54b3f92
AJ
1325 break;
1326
1327 case INDEX_op_sub_i32:
fd76e73a
RH
1328 if (const_args[1]) {
1329 if (const_args[2]) {
1330 tcg_out_movi(s, TCG_TYPE_I32, args[0], args[1] - args[2]);
1331 } else {
1332 /* Recall that SUBI is a reversed subtract. */
1333 tcg_out_arithi(s, args[0], args[2], args[1], INSN_SUBI);
1334 }
1335 } else if (const_args[2]) {
1336 tcg_out_addi2(s, args[0], args[1], -args[2]);
1337 } else {
1338 tcg_out_arith(s, args[0], args[1], args[2], INSN_SUB);
1339 }
1340 break;
1341
f54b3f92 1342 case INDEX_op_and_i32:
fd76e73a
RH
1343 if (const_args[2]) {
1344 tcg_out_andi(s, args[0], args[1], args[2]);
1345 } else {
1346 tcg_out_arith(s, args[0], args[1], args[2], INSN_AND);
1347 }
1348 break;
1349
f54b3f92 1350 case INDEX_op_or_i32:
fd76e73a
RH
1351 if (const_args[2]) {
1352 tcg_out_ori(s, args[0], args[1], args[2]);
1353 } else {
1354 tcg_out_arith(s, args[0], args[1], args[2], INSN_OR);
1355 }
1356 break;
1357
f54b3f92 1358 case INDEX_op_xor_i32:
fd76e73a
RH
1359 tcg_out_arith(s, args[0], args[1], args[2], INSN_XOR);
1360 break;
1361
1362 case INDEX_op_andc_i32:
1363 if (const_args[2]) {
1364 tcg_out_andi(s, args[0], args[1], ~args[2]);
1365 } else {
1366 tcg_out_arith(s, args[0], args[1], args[2], INSN_ANDCM);
1367 }
1368 break;
f54b3f92
AJ
1369
1370 case INDEX_op_shl_i32:
fd76e73a
RH
1371 if (const_args[2]) {
1372 tcg_out_shli(s, args[0], args[1], args[2]);
1373 } else {
1374 tcg_out_shl(s, args[0], args[1], args[2]);
1375 }
f54b3f92 1376 break;
fd76e73a 1377
f54b3f92 1378 case INDEX_op_shr_i32:
fd76e73a
RH
1379 if (const_args[2]) {
1380 tcg_out_shri(s, args[0], args[1], args[2]);
1381 } else {
1382 tcg_out_shr(s, args[0], args[1], args[2]);
1383 }
f54b3f92 1384 break;
fd76e73a 1385
f54b3f92 1386 case INDEX_op_sar_i32:
fd76e73a
RH
1387 if (const_args[2]) {
1388 tcg_out_sari(s, args[0], args[1], args[2]);
1389 } else {
1390 tcg_out_sar(s, args[0], args[1], args[2]);
1391 }
1392 break;
1393
1394 case INDEX_op_rotl_i32:
1395 if (const_args[2]) {
1396 tcg_out_rotli(s, args[0], args[1], args[2]);
1397 } else {
1398 tcg_out_rotl(s, args[0], args[1], args[2]);
1399 }
1400 break;
1401
1402 case INDEX_op_rotr_i32:
1403 if (const_args[2]) {
1404 tcg_out_rotri(s, args[0], args[1], args[2]);
1405 } else {
1406 tcg_out_rotr(s, args[0], args[1], args[2]);
1407 }
f54b3f92
AJ
1408 break;
1409
1410 case INDEX_op_mul_i32:
fd76e73a 1411 tcg_out_xmpyu(s, args[0], TCG_REG_R0, args[1], args[2]);
f54b3f92
AJ
1412 break;
1413 case INDEX_op_mulu2_i32:
fd76e73a 1414 tcg_out_xmpyu(s, args[0], args[1], args[2], args[3]);
f54b3f92 1415 break;
fd76e73a
RH
1416
1417 case INDEX_op_bswap16_i32:
1418 tcg_out_bswap16(s, args[0], args[1], 0);
f54b3f92 1419 break;
fd76e73a
RH
1420 case INDEX_op_bswap32_i32:
1421 tcg_out_bswap32(s, args[0], args[1], TCG_REG_R20);
1422 break;
1423
1424 case INDEX_op_not_i32:
1425 tcg_out_arithi(s, args[0], args[1], -1, INSN_SUBI);
1426 break;
1427 case INDEX_op_ext8s_i32:
1428 tcg_out_ext8s(s, args[0], args[1]);
1429 break;
1430 case INDEX_op_ext16s_i32:
1431 tcg_out_ext16s(s, args[0], args[1]);
1432 break;
1433
f54b3f92 1434 case INDEX_op_brcond_i32:
fd76e73a
RH
1435 tcg_out_brcond(s, args[2], args[0], args[1], const_args[1], args[3]);
1436 break;
1437 case INDEX_op_brcond2_i32:
1438 tcg_out_brcond2(s, args[4], args[0], args[1],
1439 args[2], const_args[2],
1440 args[3], const_args[3], args[5]);
1441 break;
1442
1443 case INDEX_op_setcond_i32:
1444 tcg_out_setcond(s, args[3], args[0], args[1], args[2], const_args[2]);
1445 break;
1446 case INDEX_op_setcond2_i32:
1447 tcg_out_setcond2(s, args[5], args[0], args[1], args[2],
1448 args[3], const_args[3], args[4], const_args[4]);
1449 break;
1450
1451 case INDEX_op_add2_i32:
91493631
RH
1452 tcg_out_add2(s, args[0], args[1], args[2], args[3],
1453 args[4], args[5], const_args[4]);
fd76e73a
RH
1454 break;
1455
1456 case INDEX_op_sub2_i32:
91493631
RH
1457 tcg_out_sub2(s, args[0], args[1], args[2], args[3],
1458 args[4], args[5], const_args[2], const_args[4]);
f54b3f92
AJ
1459 break;
1460
1461 case INDEX_op_qemu_ld8u:
1462 tcg_out_qemu_ld(s, args, 0);
1463 break;
1464 case INDEX_op_qemu_ld8s:
1465 tcg_out_qemu_ld(s, args, 0 | 4);
1466 break;
1467 case INDEX_op_qemu_ld16u:
1468 tcg_out_qemu_ld(s, args, 1);
1469 break;
1470 case INDEX_op_qemu_ld16s:
1471 tcg_out_qemu_ld(s, args, 1 | 4);
1472 break;
86feb1c8 1473 case INDEX_op_qemu_ld32:
f54b3f92
AJ
1474 tcg_out_qemu_ld(s, args, 2);
1475 break;
fd76e73a
RH
1476 case INDEX_op_qemu_ld64:
1477 tcg_out_qemu_ld(s, args, 3);
1478 break;
f54b3f92
AJ
1479
1480 case INDEX_op_qemu_st8:
1481 tcg_out_qemu_st(s, args, 0);
1482 break;
1483 case INDEX_op_qemu_st16:
1484 tcg_out_qemu_st(s, args, 1);
1485 break;
1486 case INDEX_op_qemu_st32:
1487 tcg_out_qemu_st(s, args, 2);
1488 break;
fd76e73a
RH
1489 case INDEX_op_qemu_st64:
1490 tcg_out_qemu_st(s, args, 3);
1491 break;
f54b3f92
AJ
1492
1493 default:
1494 fprintf(stderr, "unknown opcode 0x%x\n", opc);
1495 tcg_abort();
1496 }
f54b3f92
AJ
1497}
1498
1499static const TCGTargetOpDef hppa_op_defs[] = {
1500 { INDEX_op_exit_tb, { } },
1501 { INDEX_op_goto_tb, { } },
1502
fd76e73a 1503 { INDEX_op_call, { "ri" } },
f54b3f92
AJ
1504 { INDEX_op_jmp, { "r" } },
1505 { INDEX_op_br, { } },
1506
1507 { INDEX_op_mov_i32, { "r", "r" } },
1508 { INDEX_op_movi_i32, { "r" } },
fd76e73a 1509
f54b3f92
AJ
1510 { INDEX_op_ld8u_i32, { "r", "r" } },
1511 { INDEX_op_ld8s_i32, { "r", "r" } },
1512 { INDEX_op_ld16u_i32, { "r", "r" } },
1513 { INDEX_op_ld16s_i32, { "r", "r" } },
1514 { INDEX_op_ld_i32, { "r", "r" } },
fd76e73a
RH
1515 { INDEX_op_st8_i32, { "rZ", "r" } },
1516 { INDEX_op_st16_i32, { "rZ", "r" } },
1517 { INDEX_op_st_i32, { "rZ", "r" } },
1518
1519 { INDEX_op_add_i32, { "r", "rZ", "ri" } },
1520 { INDEX_op_sub_i32, { "r", "rI", "ri" } },
0085bd51
RH
1521 { INDEX_op_and_i32, { "r", "rZ", "rM" } },
1522 { INDEX_op_or_i32, { "r", "rZ", "rO" } },
fd76e73a 1523 { INDEX_op_xor_i32, { "r", "rZ", "rZ" } },
0085bd51
RH
1524 /* Note that the second argument will be inverted, which means
1525 we want a constant whose inversion matches M, and that O = ~M.
1526 See the implementation of and_mask_p. */
1527 { INDEX_op_andc_i32, { "r", "rZ", "rO" } },
fd76e73a
RH
1528
1529 { INDEX_op_mul_i32, { "r", "r", "r" } },
1530 { INDEX_op_mulu2_i32, { "r", "r", "r", "r" } },
f54b3f92 1531
fd76e73a
RH
1532 { INDEX_op_shl_i32, { "r", "r", "ri" } },
1533 { INDEX_op_shr_i32, { "r", "r", "ri" } },
1534 { INDEX_op_sar_i32, { "r", "r", "ri" } },
1535 { INDEX_op_rotl_i32, { "r", "r", "ri" } },
1536 { INDEX_op_rotr_i32, { "r", "r", "ri" } },
f54b3f92 1537
fd76e73a
RH
1538 { INDEX_op_bswap16_i32, { "r", "r" } },
1539 { INDEX_op_bswap32_i32, { "r", "r" } },
fd76e73a 1540 { INDEX_op_not_i32, { "r", "r" } },
f54b3f92 1541
fd76e73a 1542 { INDEX_op_ext8s_i32, { "r", "r" } },
fd76e73a 1543 { INDEX_op_ext16s_i32, { "r", "r" } },
fd76e73a
RH
1544
1545 { INDEX_op_brcond_i32, { "rZ", "rJ" } },
1546 { INDEX_op_brcond2_i32, { "rZ", "rZ", "rJ", "rJ" } },
1547
1548 { INDEX_op_setcond_i32, { "r", "rZ", "rI" } },
1549 { INDEX_op_setcond2_i32, { "r", "rZ", "rZ", "rI", "rI" } },
1550
1551 { INDEX_op_add2_i32, { "r", "r", "rZ", "rZ", "rI", "rZ" } },
91493631 1552 { INDEX_op_sub2_i32, { "r", "r", "rI", "rZ", "rK", "rZ" } },
f54b3f92
AJ
1553
1554#if TARGET_LONG_BITS == 32
1555 { INDEX_op_qemu_ld8u, { "r", "L" } },
1556 { INDEX_op_qemu_ld8s, { "r", "L" } },
1557 { INDEX_op_qemu_ld16u, { "r", "L" } },
1558 { INDEX_op_qemu_ld16s, { "r", "L" } },
86feb1c8 1559 { INDEX_op_qemu_ld32, { "r", "L" } },
f54b3f92
AJ
1560 { INDEX_op_qemu_ld64, { "r", "r", "L" } },
1561
fd76e73a
RH
1562 { INDEX_op_qemu_st8, { "LZ", "L" } },
1563 { INDEX_op_qemu_st16, { "LZ", "L" } },
1564 { INDEX_op_qemu_st32, { "LZ", "L" } },
1565 { INDEX_op_qemu_st64, { "LZ", "LZ", "L" } },
f54b3f92
AJ
1566#else
1567 { INDEX_op_qemu_ld8u, { "r", "L", "L" } },
1568 { INDEX_op_qemu_ld8s, { "r", "L", "L" } },
1569 { INDEX_op_qemu_ld16u, { "r", "L", "L" } },
1570 { INDEX_op_qemu_ld16s, { "r", "L", "L" } },
86feb1c8 1571 { INDEX_op_qemu_ld32, { "r", "L", "L" } },
f54b3f92
AJ
1572 { INDEX_op_qemu_ld64, { "r", "r", "L", "L" } },
1573
fd76e73a
RH
1574 { INDEX_op_qemu_st8, { "LZ", "L", "L" } },
1575 { INDEX_op_qemu_st16, { "LZ", "L", "L" } },
1576 { INDEX_op_qemu_st32, { "LZ", "L", "L" } },
1577 { INDEX_op_qemu_st64, { "LZ", "LZ", "L", "L" } },
f54b3f92
AJ
1578#endif
1579 { -1 },
1580};
1581
fd76e73a
RH
1582static int tcg_target_callee_save_regs[] = {
1583 /* R2, the return address register, is saved specially
1584 in the caller's frame. */
1585 /* R3, the frame pointer, is not currently modified. */
1586 TCG_REG_R4,
1587 TCG_REG_R5,
1588 TCG_REG_R6,
1589 TCG_REG_R7,
1590 TCG_REG_R8,
1591 TCG_REG_R9,
1592 TCG_REG_R10,
1593 TCG_REG_R11,
1594 TCG_REG_R12,
1595 TCG_REG_R13,
1596 TCG_REG_R14,
1597 TCG_REG_R15,
1598 TCG_REG_R16,
1599 /* R17 is the global env, so no need to save. */
1600 TCG_REG_R18
1601};
1602
1603void tcg_target_qemu_prologue(TCGContext *s)
1604{
1605 int frame_size, i;
1606
1607 /* Allocate space for the fixed frame marker. */
1608 frame_size = -TCG_TARGET_CALL_STACK_OFFSET;
1609 frame_size += TCG_TARGET_STATIC_CALL_ARGS_SIZE;
1610
1611 /* Allocate space for the saved registers. */
1612 frame_size += ARRAY_SIZE(tcg_target_callee_save_regs) * 4;
1613
1614 /* Align the allocated space. */
1615 frame_size = ((frame_size + TCG_TARGET_STACK_ALIGN - 1)
1616 & -TCG_TARGET_STACK_ALIGN);
1617
1618 /* The return address is stored in the caller's frame. */
1619 tcg_out_st(s, TCG_TYPE_PTR, TCG_REG_RP, TCG_REG_SP, -20);
1620
1621 /* Allocate stack frame, saving the first register at the same time. */
1622 tcg_out_ldst(s, tcg_target_callee_save_regs[0],
1623 TCG_REG_SP, frame_size, INSN_STWM);
1624
1625 /* Save all callee saved registers. */
1626 for (i = 1; i < ARRAY_SIZE(tcg_target_callee_save_regs); i++) {
1627 tcg_out_st(s, TCG_TYPE_PTR, tcg_target_callee_save_regs[i],
1628 TCG_REG_SP, -frame_size + i * 4);
1629 }
1630
884d348b
RH
1631#ifdef CONFIG_USE_GUEST_BASE
1632 /* Note that GUEST_BASE can change after the prologue is generated.
1633 To combat that, load the value from the variable instead of
1634 embedding a constant here. */
1635 tcg_out_ld(s, TCG_TYPE_PTR, TCG_GUEST_BASE_REG,
1636 TCG_REG_R0, (tcg_target_long)&guest_base);
1637#endif
fd76e73a
RH
1638
1639 /* Jump to TB, and adjust R18 to be the return address. */
1640 tcg_out32(s, INSN_BLE_SR4 | INSN_R2(TCG_REG_R26));
1641 tcg_out_mov(s, TCG_REG_R18, TCG_REG_R31);
1642
1643 /* Restore callee saved registers. */
1644 tcg_out_ld(s, TCG_TYPE_PTR, TCG_REG_RP, TCG_REG_SP, -frame_size - 20);
1645 for (i = 1; i < ARRAY_SIZE(tcg_target_callee_save_regs); i++) {
1646 tcg_out_ld(s, TCG_TYPE_PTR, tcg_target_callee_save_regs[i],
1647 TCG_REG_SP, -frame_size + i * 4);
1648 }
1649
1650 /* Deallocate stack frame and return. */
1651 tcg_out32(s, INSN_BV | INSN_R2(TCG_REG_RP));
1652 tcg_out_ldst(s, tcg_target_callee_save_regs[0],
1653 TCG_REG_SP, -frame_size, INSN_LDWM);
1654}
1655
f54b3f92
AJ
1656void tcg_target_init(TCGContext *s)
1657{
1658 tcg_regset_set32(tcg_target_available_regs[TCG_TYPE_I32], 0, 0xffffffff);
fd76e73a
RH
1659
1660 tcg_regset_clear(tcg_target_call_clobber_regs);
1661 tcg_regset_set_reg(tcg_target_call_clobber_regs, TCG_REG_R20);
1662 tcg_regset_set_reg(tcg_target_call_clobber_regs, TCG_REG_R21);
1663 tcg_regset_set_reg(tcg_target_call_clobber_regs, TCG_REG_R22);
1664 tcg_regset_set_reg(tcg_target_call_clobber_regs, TCG_REG_R23);
1665 tcg_regset_set_reg(tcg_target_call_clobber_regs, TCG_REG_R24);
1666 tcg_regset_set_reg(tcg_target_call_clobber_regs, TCG_REG_R25);
1667 tcg_regset_set_reg(tcg_target_call_clobber_regs, TCG_REG_R26);
1668 tcg_regset_set_reg(tcg_target_call_clobber_regs, TCG_REG_RET0);
1669 tcg_regset_set_reg(tcg_target_call_clobber_regs, TCG_REG_RET1);
f54b3f92
AJ
1670
1671 tcg_regset_clear(s->reserved_regs);
1672 tcg_regset_set_reg(s->reserved_regs, TCG_REG_R0); /* hardwired to zero */
1673 tcg_regset_set_reg(s->reserved_regs, TCG_REG_R1); /* addil target */
1674 tcg_regset_set_reg(s->reserved_regs, TCG_REG_RP); /* link register */
1675 tcg_regset_set_reg(s->reserved_regs, TCG_REG_R3); /* frame pointer */
1676 tcg_regset_set_reg(s->reserved_regs, TCG_REG_R18); /* return pointer */
1677 tcg_regset_set_reg(s->reserved_regs, TCG_REG_R19); /* clobbered w/o pic */
1678 tcg_regset_set_reg(s->reserved_regs, TCG_REG_R20); /* reserved */
1679 tcg_regset_set_reg(s->reserved_regs, TCG_REG_DP); /* data pointer */
1680 tcg_regset_set_reg(s->reserved_regs, TCG_REG_SP); /* stack pointer */
1681 tcg_regset_set_reg(s->reserved_regs, TCG_REG_R31); /* ble link reg */
884d348b
RH
1682#ifdef CONFIG_USE_GUEST_BASE
1683 tcg_regset_set_reg(s->reserved_regs, TCG_GUEST_BASE_REG);
1684#endif
f54b3f92
AJ
1685
1686 tcg_add_target_add_op_defs(hppa_op_defs);
1687}