1/*
2 * Tiny Code Generator for QEMU
3 *
4 * Copyright (c) 2008 Fabrice Bellard
5 *
6 * Permission is hereby granted, free of charge, to any person obtaining a copy
7 * of this software and associated documentation files (the "Software"), to deal
8 * in the Software without restriction, including without limitation the rights
9 * to use, copy, modify, merge, publish, distribute, sublicense, and/or sell
10 * copies of the Software, and to permit persons to whom the Software is
11 * furnished to do so, subject to the following conditions:
12 *
13 * The above copyright notice and this permission notice shall be included in
14 * all copies or substantial portions of the Software.
15 *
16 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
17 * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
18 * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
19 * THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
20 * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
21 * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN
22 * THE SOFTWARE.
23 */
24
25#ifndef NDEBUG
26static const char * const tcg_target_reg_names[TCG_TARGET_NB_REGS] = {
27    "%r0", "%r1", "%rp", "%r3", "%r4", "%r5", "%r6", "%r7",
28    "%r8", "%r9", "%r10", "%r11", "%r12", "%r13", "%r14", "%r15",
29    "%r16", "%r17", "%r18", "%r19", "%r20", "%r21", "%r22", "%r23",
30    "%r24", "%r25", "%r26", "%dp", "%ret0", "%ret1", "%sp", "%r31",
31};
32#endif
33
34/* This is an 8 byte temp slot in the stack frame.  */
35#define STACK_TEMP_OFS -16
36
37#ifdef CONFIG_USE_GUEST_BASE
38#define TCG_GUEST_BASE_REG TCG_REG_R16
39#else
40#define TCG_GUEST_BASE_REG TCG_REG_R0
41#endif
42
43static const int tcg_target_reg_alloc_order[] = {
44    TCG_REG_R4,
45    TCG_REG_R5,
46    TCG_REG_R6,
47    TCG_REG_R7,
48    TCG_REG_R8,
49    TCG_REG_R9,
50    TCG_REG_R10,
51    TCG_REG_R11,
52    TCG_REG_R12,
53    TCG_REG_R13,
54
55    TCG_REG_R17,
56    TCG_REG_R14,
57    TCG_REG_R15,
58    TCG_REG_R16,
59
60    TCG_REG_R26,
61    TCG_REG_R25,
62    TCG_REG_R24,
63    TCG_REG_R23,
64
65    TCG_REG_RET0,
66    TCG_REG_RET1,
67};
68
69static const int tcg_target_call_iarg_regs[4] = {
70    TCG_REG_R26,
71    TCG_REG_R25,
72    TCG_REG_R24,
73    TCG_REG_R23,
74};
75
76static const int tcg_target_call_oarg_regs[2] = {
77    TCG_REG_RET0,
78    TCG_REG_RET1,
79};
80
81/* True iff val fits a signed field of width BITS.  */
82static inline int check_fit_tl(tcg_target_long val, unsigned int bits)
83{
84    return (val << ((sizeof(tcg_target_long) * 8 - bits))
85            >> (sizeof(tcg_target_long) * 8 - bits)) == val;
86}
87
88/* True iff depi can be used to compute (reg | MASK).
89   Accept a bit pattern like:
90      0....01....1
91      1....10....0
92      0..01..10..0
93   Copied from gcc sources.  */
94static inline int or_mask_p(tcg_target_ulong mask)
95{
96    if (mask == 0 || mask == -1) {
97        return 0;
98    }
99    mask += mask & -mask;
100    return (mask & (mask - 1)) == 0;
101}
102
103/* True iff depi or extru can be used to compute (reg & mask).
104   Accept a bit pattern like these:
105      0....01....1
106      1....10....0
107      1..10..01..1
108   Copied from gcc sources.  */
109static inline int and_mask_p(tcg_target_ulong mask)
110{
111    return or_mask_p(~mask);
112}
113
114static int low_sign_ext(int val, int len)
115{
116    return (((val << 1) & ~(-1u << len)) | ((val >> (len - 1)) & 1));
117}
118
119static int reassemble_12(int as12)
120{
121    return (((as12 & 0x800) >> 11) |
122            ((as12 & 0x400) >> 8) |
123            ((as12 & 0x3ff) << 3));
124}
125
126static int reassemble_17(int as17)
127{
128    return (((as17 & 0x10000) >> 16) |
129            ((as17 & 0x0f800) << 5) |
130            ((as17 & 0x00400) >> 8) |
131            ((as17 & 0x003ff) << 3));
132}
133
134static int reassemble_21(int as21)
135{
136    return (((as21 & 0x100000) >> 20) |
137            ((as21 & 0x0ffe00) >> 8) |
138            ((as21 & 0x000180) << 7) |
139            ((as21 & 0x00007c) << 14) |
140            ((as21 & 0x000003) << 12));
141}
142
143/* ??? Bizzarely, there is no PCREL12F relocation type.  I guess all
144   such relocations are simply fully handled by the assembler.  */
145#define R_PARISC_PCREL12F  R_PARISC_NONE
146
147static void patch_reloc(uint8_t *code_ptr, int type,
148                        tcg_target_long value, tcg_target_long addend)
149{
150    uint32_t *insn_ptr = (uint32_t *)code_ptr;
151    uint32_t insn = *insn_ptr;
152    tcg_target_long pcrel;
153
154    value += addend;
155    pcrel = (value - ((tcg_target_long)code_ptr + 8)) >> 2;
156
157    switch (type) {
158    case R_PARISC_PCREL12F:
159        assert(check_fit_tl(pcrel, 12));
160        /* ??? We assume all patches are forward.  See tcg_out_brcond
161           re setting the NUL bit on the branch and eliding the nop.  */
162        assert(pcrel >= 0);
163        insn &= ~0x1ffdu;
164        insn |= reassemble_12(pcrel);
165        break;
166    case R_PARISC_PCREL17F:
167        assert(check_fit_tl(pcrel, 17));
168        insn &= ~0x1f1ffdu;
169        insn |= reassemble_17(pcrel);
170        break;
171    default:
172        tcg_abort();
173    }
174
175    *insn_ptr = insn;
176}
177
178/* maximum number of register used for input function arguments */
179static inline int tcg_target_get_call_iarg_regs_count(int flags)
180{
181    return 4;
182}
183
184/* parse target specific constraints */
185static int target_parse_constraint(TCGArgConstraint *ct, const char **pct_str)
186{
187    const char *ct_str;
188
189    ct_str = *pct_str;
190    switch (ct_str[0]) {
191    case 'r':
192        ct->ct |= TCG_CT_REG;
193        tcg_regset_set32(ct->u.regs, 0, 0xffffffff);
194        break;
195    case 'L': /* qemu_ld/st constraint */
196        ct->ct |= TCG_CT_REG;
197        tcg_regset_set32(ct->u.regs, 0, 0xffffffff);
198        tcg_regset_reset_reg(ct->u.regs, TCG_REG_R26);
199        tcg_regset_reset_reg(ct->u.regs, TCG_REG_R25);
200        tcg_regset_reset_reg(ct->u.regs, TCG_REG_R24);
201        tcg_regset_reset_reg(ct->u.regs, TCG_REG_R23);
202        break;
203    case 'Z':
204        ct->ct |= TCG_CT_CONST_0;
205        break;
206    case 'I':
207        ct->ct |= TCG_CT_CONST_S11;
208        break;
209    case 'J':
210        ct->ct |= TCG_CT_CONST_S5;
211	break;
212    case 'K':
213        ct->ct |= TCG_CT_CONST_MS11;
214        break;
215    case 'M':
216        ct->ct |= TCG_CT_CONST_AND;
217        break;
218    case 'O':
219        ct->ct |= TCG_CT_CONST_OR;
220        break;
221    default:
222        return -1;
223    }
224    ct_str++;
225    *pct_str = ct_str;
226    return 0;
227}
228
229/* test if a constant matches the constraint */
230static int tcg_target_const_match(tcg_target_long val,
231                                  const TCGArgConstraint *arg_ct)
232{
233    int ct = arg_ct->ct;
234    if (ct & TCG_CT_CONST) {
235        return 1;
236    } else if (ct & TCG_CT_CONST_0) {
237        return val == 0;
238    } else if (ct & TCG_CT_CONST_S5) {
239        return check_fit_tl(val, 5);
240    } else if (ct & TCG_CT_CONST_S11) {
241        return check_fit_tl(val, 11);
242    } else if (ct & TCG_CT_CONST_MS11) {
243        return check_fit_tl(-val, 11);
244    } else if (ct & TCG_CT_CONST_AND) {
245        return and_mask_p(val);
246    } else if (ct & TCG_CT_CONST_OR) {
247        return or_mask_p(val);
248    }
249    return 0;
250}
251
252#define INSN_OP(x)       ((x) << 26)
253#define INSN_EXT3BR(x)   ((x) << 13)
254#define INSN_EXT3SH(x)   ((x) << 10)
255#define INSN_EXT4(x)     ((x) << 6)
256#define INSN_EXT5(x)     (x)
257#define INSN_EXT6(x)     ((x) << 6)
258#define INSN_EXT7(x)     ((x) << 6)
259#define INSN_EXT8A(x)    ((x) << 6)
260#define INSN_EXT8B(x)    ((x) << 5)
261#define INSN_T(x)        (x)
262#define INSN_R1(x)       ((x) << 16)
263#define INSN_R2(x)       ((x) << 21)
264#define INSN_DEP_LEN(x)  (32 - (x))
265#define INSN_SHDEP_CP(x) ((31 - (x)) << 5)
266#define INSN_SHDEP_P(x)  ((x) << 5)
267#define INSN_COND(x)     ((x) << 13)
268#define INSN_IM11(x)     low_sign_ext(x, 11)
269#define INSN_IM14(x)     low_sign_ext(x, 14)
270#define INSN_IM5(x)      (low_sign_ext(x, 5) << 16)
271
272#define COND_NEVER   0
273#define COND_EQ      1
274#define COND_LT      2
275#define COND_LE      3
276#define COND_LTU     4
277#define COND_LEU     5
278#define COND_SV      6
279#define COND_OD      7
280#define COND_FALSE   8
281
282#define INSN_ADD	(INSN_OP(0x02) | INSN_EXT6(0x18))
283#define INSN_ADDC	(INSN_OP(0x02) | INSN_EXT6(0x1c))
284#define INSN_ADDI	(INSN_OP(0x2d))
285#define INSN_ADDIL	(INSN_OP(0x0a))
286#define INSN_ADDL	(INSN_OP(0x02) | INSN_EXT6(0x28))
287#define INSN_AND	(INSN_OP(0x02) | INSN_EXT6(0x08))
288#define INSN_ANDCM	(INSN_OP(0x02) | INSN_EXT6(0x00))
289#define INSN_COMCLR	(INSN_OP(0x02) | INSN_EXT6(0x22))
290#define INSN_COMICLR	(INSN_OP(0x24))
291#define INSN_DEP	(INSN_OP(0x35) | INSN_EXT3SH(3))
292#define INSN_DEPI	(INSN_OP(0x35) | INSN_EXT3SH(7))
293#define INSN_EXTRS	(INSN_OP(0x34) | INSN_EXT3SH(7))
294#define INSN_EXTRU	(INSN_OP(0x34) | INSN_EXT3SH(6))
295#define INSN_LDIL	(INSN_OP(0x08))
296#define INSN_LDO	(INSN_OP(0x0d))
297#define INSN_MTCTL	(INSN_OP(0x00) | INSN_EXT8B(0xc2))
298#define INSN_OR		(INSN_OP(0x02) | INSN_EXT6(0x09))
299#define INSN_SHD	(INSN_OP(0x34) | INSN_EXT3SH(2))
300#define INSN_SUB	(INSN_OP(0x02) | INSN_EXT6(0x10))
301#define INSN_SUBB	(INSN_OP(0x02) | INSN_EXT6(0x14))
302#define INSN_SUBI	(INSN_OP(0x25))
303#define INSN_VEXTRS	(INSN_OP(0x34) | INSN_EXT3SH(5))
304#define INSN_VEXTRU	(INSN_OP(0x34) | INSN_EXT3SH(4))
305#define INSN_VSHD	(INSN_OP(0x34) | INSN_EXT3SH(0))
306#define INSN_XOR	(INSN_OP(0x02) | INSN_EXT6(0x0a))
307#define INSN_ZDEP	(INSN_OP(0x35) | INSN_EXT3SH(2))
308#define INSN_ZVDEP	(INSN_OP(0x35) | INSN_EXT3SH(0))
309
310#define INSN_BL         (INSN_OP(0x3a) | INSN_EXT3BR(0))
311#define INSN_BL_N       (INSN_OP(0x3a) | INSN_EXT3BR(0) | 2)
312#define INSN_BLR        (INSN_OP(0x3a) | INSN_EXT3BR(2))
313#define INSN_BV         (INSN_OP(0x3a) | INSN_EXT3BR(6))
314#define INSN_BV_N       (INSN_OP(0x3a) | INSN_EXT3BR(6) | 2)
315#define INSN_BLE_SR4    (INSN_OP(0x39) | (1 << 13))
316
317#define INSN_LDB        (INSN_OP(0x10))
318#define INSN_LDH        (INSN_OP(0x11))
319#define INSN_LDW        (INSN_OP(0x12))
320#define INSN_LDWM       (INSN_OP(0x13))
321#define INSN_FLDDS      (INSN_OP(0x0b) | INSN_EXT4(0) | (1 << 12))
322
323#define INSN_LDBX	(INSN_OP(0x03) | INSN_EXT4(0))
324#define INSN_LDHX	(INSN_OP(0x03) | INSN_EXT4(1))
325#define INSN_LDWX       (INSN_OP(0x03) | INSN_EXT4(2))
326
327#define INSN_STB        (INSN_OP(0x18))
328#define INSN_STH        (INSN_OP(0x19))
329#define INSN_STW        (INSN_OP(0x1a))
330#define INSN_STWM       (INSN_OP(0x1b))
331#define INSN_FSTDS      (INSN_OP(0x0b) | INSN_EXT4(8) | (1 << 12))
332
333#define INSN_COMBT      (INSN_OP(0x20))
334#define INSN_COMBF      (INSN_OP(0x22))
335#define INSN_COMIBT     (INSN_OP(0x21))
336#define INSN_COMIBF     (INSN_OP(0x23))
337
338/* supplied by libgcc */
339extern void *__canonicalize_funcptr_for_compare(void *);
340
341static void tcg_out_mov(TCGContext *s, TCGType type, int ret, int arg)
342{
343    /* PA1.1 defines COPY as OR r,0,t; PA2.0 defines COPY as LDO 0(r),t
344       but hppa-dis.c is unaware of this definition */
345    if (ret != arg) {
346        tcg_out32(s, INSN_OR | INSN_T(ret) | INSN_R1(arg)
347                  | INSN_R2(TCG_REG_R0));
348    }
349}
350
351static void tcg_out_movi(TCGContext *s, TCGType type,
352                         int ret, tcg_target_long arg)
353{
354    if (check_fit_tl(arg, 14)) {
355        tcg_out32(s, INSN_LDO | INSN_R1(ret)
356                  | INSN_R2(TCG_REG_R0) | INSN_IM14(arg));
357    } else {
358        uint32_t hi, lo;
359        hi = arg >> 11;
360        lo = arg & 0x7ff;
361
362        tcg_out32(s, INSN_LDIL | INSN_R2(ret) | reassemble_21(hi));
363        if (lo) {
364            tcg_out32(s, INSN_LDO | INSN_R1(ret)
365                      | INSN_R2(ret) | INSN_IM14(lo));
366        }
367    }
368}
369
370static void tcg_out_ldst(TCGContext *s, int ret, int addr,
371                         tcg_target_long offset, int op)
372{
373    if (!check_fit_tl(offset, 14)) {
374        uint32_t hi, lo, op;
375
376        hi = offset >> 11;
377        lo = offset & 0x7ff;
378
379        if (addr == TCG_REG_R0) {
380            op = INSN_LDIL | INSN_R2(TCG_REG_R1);
381        } else {
382            op = INSN_ADDIL | INSN_R2(addr);
383        }
384        tcg_out32(s, op | reassemble_21(hi));
385
386        addr = TCG_REG_R1;
387	offset = lo;
388    }
389
390    if (ret != addr || offset != 0 || op != INSN_LDO) {
391        tcg_out32(s, op | INSN_R1(ret) | INSN_R2(addr) | INSN_IM14(offset));
392    }
393}
394
395/* This function is required by tcg.c.  */
396static inline void tcg_out_ld(TCGContext *s, TCGType type, int ret,
397                              int arg1, tcg_target_long arg2)
398{
399    tcg_out_ldst(s, ret, arg1, arg2, INSN_LDW);
400}
401
402/* This function is required by tcg.c.  */
403static inline void tcg_out_st(TCGContext *s, TCGType type, int ret,
404                              int arg1, tcg_target_long arg2)
405{
406    tcg_out_ldst(s, ret, arg1, arg2, INSN_STW);
407}
408
409static void tcg_out_ldst_index(TCGContext *s, int data,
410                               int base, int index, int op)
411{
412    tcg_out32(s, op | INSN_T(data) | INSN_R1(index) | INSN_R2(base));
413}
414
415static inline void tcg_out_addi2(TCGContext *s, int ret, int arg1,
416                                 tcg_target_long val)
417{
418    tcg_out_ldst(s, ret, arg1, val, INSN_LDO);
419}
420
421/* This function is required by tcg.c.  */
422static inline void tcg_out_addi(TCGContext *s, int reg, tcg_target_long val)
423{
424    tcg_out_addi2(s, reg, reg, val);
425}
426
427static inline void tcg_out_arith(TCGContext *s, int t, int r1, int r2, int op)
428{
429    tcg_out32(s, op | INSN_T(t) | INSN_R1(r1) | INSN_R2(r2));
430}
431
432static inline void tcg_out_arithi(TCGContext *s, int t, int r1,
433                                  tcg_target_long val, int op)
434{
435    assert(check_fit_tl(val, 11));
436    tcg_out32(s, op | INSN_R1(t) | INSN_R2(r1) | INSN_IM11(val));
437}
438
439static inline void tcg_out_nop(TCGContext *s)
440{
441    tcg_out_arith(s, TCG_REG_R0, TCG_REG_R0, TCG_REG_R0, INSN_OR);
442}
443
444static inline void tcg_out_mtctl_sar(TCGContext *s, int arg)
445{
446    tcg_out32(s, INSN_MTCTL | INSN_R2(11) | INSN_R1(arg));
447}
448
449/* Extract LEN bits at position OFS from ARG and place in RET.
450   Note that here the bit ordering is reversed from the PA-RISC
451   standard, such that the right-most bit is 0.  */
452static inline void tcg_out_extr(TCGContext *s, int ret, int arg,
453                                unsigned ofs, unsigned len, int sign)
454{
455    assert(ofs < 32 && len <= 32 - ofs);
456    tcg_out32(s, (sign ? INSN_EXTRS : INSN_EXTRU)
457              | INSN_R1(ret) | INSN_R2(arg)
458              | INSN_SHDEP_P(31 - ofs) | INSN_DEP_LEN(len));
459}
460
461/* Likewise with OFS interpreted little-endian.  */
462static inline void tcg_out_dep(TCGContext *s, int ret, int arg,
463                               unsigned ofs, unsigned len)
464{
465    assert(ofs < 32 && len <= 32 - ofs);
466    tcg_out32(s, INSN_DEP | INSN_R2(ret) | INSN_R1(arg)
467              | INSN_SHDEP_CP(31 - ofs) | INSN_DEP_LEN(len));
468}
469
470static inline void tcg_out_shd(TCGContext *s, int ret, int hi, int lo,
471                               unsigned count)
472{
473    assert(count < 32);
474    tcg_out32(s, INSN_SHD | INSN_R1(hi) | INSN_R2(lo) | INSN_T(ret)
475              | INSN_SHDEP_CP(count));
476}
477
478static void tcg_out_vshd(TCGContext *s, int ret, int hi, int lo, int creg)
479{
480    tcg_out_mtctl_sar(s, creg);
481    tcg_out32(s, INSN_VSHD | INSN_T(ret) | INSN_R1(hi) | INSN_R2(lo));
482}
483
484static void tcg_out_ori(TCGContext *s, int ret, int arg, tcg_target_ulong m)
485{
486    int bs0, bs1;
487
488    /* Note that the argument is constrained to match or_mask_p.  */
489    for (bs0 = 0; bs0 < 32; bs0++) {
490        if ((m & (1u << bs0)) != 0) {
491            break;
492        }
493    }
494    for (bs1 = bs0; bs1 < 32; bs1++) {
495        if ((m & (1u << bs1)) == 0) {
496            break;
497        }
498    }
499    assert(bs1 == 32 || (1ul << bs1) > m);
500
501    tcg_out_mov(s, TCG_TYPE_I32, ret, arg);
502    tcg_out32(s, INSN_DEPI | INSN_R2(ret) | INSN_IM5(-1)
503              | INSN_SHDEP_CP(31 - bs0) | INSN_DEP_LEN(bs1 - bs0));
504}
505
506static void tcg_out_andi(TCGContext *s, int ret, int arg, tcg_target_ulong m)
507{
508    int ls0, ls1, ms0;
509
510    /* Note that the argument is constrained to match and_mask_p.  */
511    for (ls0 = 0; ls0 < 32; ls0++) {
512        if ((m & (1u << ls0)) == 0) {
513            break;
514        }
515    }
516    for (ls1 = ls0; ls1 < 32; ls1++) {
517        if ((m & (1u << ls1)) != 0) {
518            break;
519        }
520    }
521    for (ms0 = ls1; ms0 < 32; ms0++) {
522        if ((m & (1u << ms0)) == 0) {
523            break;
524        }
525    }
526    assert (ms0 == 32);
527
528    if (ls1 == 32) {
529        tcg_out_extr(s, ret, arg, 0, ls0, 0);
530    } else {
531        tcg_out_mov(s, TCG_TYPE_I32, ret, arg);
532        tcg_out32(s, INSN_DEPI | INSN_R2(ret) | INSN_IM5(0)
533                  | INSN_SHDEP_CP(31 - ls0) | INSN_DEP_LEN(ls1 - ls0));
534    }
535}
536
537static inline void tcg_out_ext8s(TCGContext *s, int ret, int arg)
538{
539    tcg_out_extr(s, ret, arg, 0, 8, 1);
540}
541
542static inline void tcg_out_ext16s(TCGContext *s, int ret, int arg)
543{
544    tcg_out_extr(s, ret, arg, 0, 16, 1);
545}
546
547static void tcg_out_shli(TCGContext *s, int ret, int arg, int count)
548{
549    count &= 31;
550    tcg_out32(s, INSN_ZDEP | INSN_R2(ret) | INSN_R1(arg)
551              | INSN_SHDEP_CP(31 - count) | INSN_DEP_LEN(32 - count));
552}
553
554static void tcg_out_shl(TCGContext *s, int ret, int arg, int creg)
555{
556    tcg_out_arithi(s, TCG_REG_R20, creg, 31, INSN_SUBI);
557    tcg_out_mtctl_sar(s, TCG_REG_R20);
558    tcg_out32(s, INSN_ZVDEP | INSN_R2(ret) | INSN_R1(arg) | INSN_DEP_LEN(32));
559}
560
561static void tcg_out_shri(TCGContext *s, int ret, int arg, int count)
562{
563    count &= 31;
564    tcg_out_extr(s, ret, arg, count, 32 - count, 0);
565}
566
567static void tcg_out_shr(TCGContext *s, int ret, int arg, int creg)
568{
569    tcg_out_vshd(s, ret, TCG_REG_R0, arg, creg);
570}
571
572static void tcg_out_sari(TCGContext *s, int ret, int arg, int count)
573{
574    count &= 31;
575    tcg_out_extr(s, ret, arg, count, 32 - count, 1);
576}
577
578static void tcg_out_sar(TCGContext *s, int ret, int arg, int creg)
579{
580    tcg_out_arithi(s, TCG_REG_R20, creg, 31, INSN_SUBI);
581    tcg_out_mtctl_sar(s, TCG_REG_R20);
582    tcg_out32(s, INSN_VEXTRS | INSN_R1(ret) | INSN_R2(arg) | INSN_DEP_LEN(32));
583}
584
585static void tcg_out_rotli(TCGContext *s, int ret, int arg, int count)
586{
587    count &= 31;
588    tcg_out_shd(s, ret, arg, arg, 32 - count);
589}
590
591static void tcg_out_rotl(TCGContext *s, int ret, int arg, int creg)
592{
593    tcg_out_arithi(s, TCG_REG_R20, creg, 32, INSN_SUBI);
594    tcg_out_vshd(s, ret, arg, arg, TCG_REG_R20);
595}
596
597static void tcg_out_rotri(TCGContext *s, int ret, int arg, int count)
598{
599    count &= 31;
600    tcg_out_shd(s, ret, arg, arg, count);
601}
602
603static void tcg_out_rotr(TCGContext *s, int ret, int arg, int creg)
604{
605    tcg_out_vshd(s, ret, arg, arg, creg);
606}
607
608static void tcg_out_bswap16(TCGContext *s, int ret, int arg, int sign)
609{
610    if (ret != arg) {
611        tcg_out_mov(s, TCG_TYPE_I32, ret, arg); /* arg =  xxAB */
612    }
613    tcg_out_dep(s, ret, ret, 16, 8);          /* ret =  xBAB */
614    tcg_out_extr(s, ret, ret, 8, 16, sign);   /* ret =  ..BA */
615}
616
617static void tcg_out_bswap32(TCGContext *s, int ret, int arg, int temp)
618{
619                                          /* arg =  ABCD */
620    tcg_out_rotri(s, temp, arg, 16);      /* temp = CDAB */
621    tcg_out_dep(s, temp, temp, 16, 8);    /* temp = CBAB */
622    tcg_out_shd(s, ret, arg, temp, 8);    /* ret =  DCBA */
623}
624
625static void tcg_out_call(TCGContext *s, void *func)
626{
627    tcg_target_long val, hi, lo, disp;
628
629    val = (uint32_t)__canonicalize_funcptr_for_compare(func);
630    disp = (val - ((tcg_target_long)s->code_ptr + 8)) >> 2;
631
632    if (check_fit_tl(disp, 17)) {
633        tcg_out32(s, INSN_BL_N | INSN_R2(TCG_REG_RP) | reassemble_17(disp));
634    } else {
635        hi = val >> 11;
636        lo = val & 0x7ff;
637
638        tcg_out32(s, INSN_LDIL | INSN_R2(TCG_REG_R20) | reassemble_21(hi));
639        tcg_out32(s, INSN_BLE_SR4 | INSN_R2(TCG_REG_R20)
640                  | reassemble_17(lo >> 2));
641        tcg_out_mov(s, TCG_TYPE_I32, TCG_REG_RP, TCG_REG_R31);
642    }
643}
644
645static void tcg_out_xmpyu(TCGContext *s, int retl, int reth,
646                          int arg1, int arg2)
647{
648    /* Store both words into the stack for copy to the FPU.  */
649    tcg_out_ldst(s, arg1, TCG_REG_SP, STACK_TEMP_OFS, INSN_STW);
650    tcg_out_ldst(s, arg2, TCG_REG_SP, STACK_TEMP_OFS + 4, INSN_STW);
651
652    /* Load both words into the FPU at the same time.  We get away
653       with this because we can address the left and right half of the
654       FPU registers individually once loaded.  */
655    /* fldds stack_temp(sp),fr22 */
656    tcg_out32(s, INSN_FLDDS | INSN_R2(TCG_REG_SP)
657              | INSN_IM5(STACK_TEMP_OFS) | INSN_T(22));
658
659    /* xmpyu fr22r,fr22,fr22 */
660    tcg_out32(s, 0x3ad64796);
661
662    /* Store the 64-bit result back into the stack.  */
663    /* fstds stack_temp(sp),fr22 */
664    tcg_out32(s, INSN_FSTDS | INSN_R2(TCG_REG_SP)
665              | INSN_IM5(STACK_TEMP_OFS) | INSN_T(22));
666
667    /* Load the pieces of the result that the caller requested.  */
668    if (reth) {
669        tcg_out_ldst(s, reth, TCG_REG_SP, STACK_TEMP_OFS, INSN_LDW);
670    }
671    if (retl) {
672        tcg_out_ldst(s, retl, TCG_REG_SP, STACK_TEMP_OFS + 4, INSN_LDW);
673    }
674}
675
676static void tcg_out_add2(TCGContext *s, int destl, int desth,
677                         int al, int ah, int bl, int bh, int blconst)
678{
679    int tmp = (destl == ah || destl == bh ? TCG_REG_R20 : destl);
680
681    if (blconst) {
682        tcg_out_arithi(s, tmp, al, bl, INSN_ADDI);
683    } else {
684        tcg_out_arith(s, tmp, al, bl, INSN_ADD);
685    }
686    tcg_out_arith(s, desth, ah, bh, INSN_ADDC);
687
688    tcg_out_mov(s, TCG_TYPE_I32, destl, tmp);
689}
690
691static void tcg_out_sub2(TCGContext *s, int destl, int desth, int al, int ah,
692                         int bl, int bh, int alconst, int blconst)
693{
694    int tmp = (destl == ah || destl == bh ? TCG_REG_R20 : destl);
695
696    if (alconst) {
697        if (blconst) {
698            tcg_out_movi(s, TCG_TYPE_I32, TCG_REG_R20, bl);
699            bl = TCG_REG_R20;
700        }
701        tcg_out_arithi(s, tmp, bl, al, INSN_SUBI);
702    } else if (blconst) {
703        tcg_out_arithi(s, tmp, al, -bl, INSN_ADDI);
704    } else {
705        tcg_out_arith(s, tmp, al, bl, INSN_SUB);
706    }
707    tcg_out_arith(s, desth, ah, bh, INSN_SUBB);
708
709    tcg_out_mov(s, TCG_TYPE_I32, destl, tmp);
710}
711
712static void tcg_out_branch(TCGContext *s, int label_index, int nul)
713{
714    TCGLabel *l = &s->labels[label_index];
715    uint32_t op = nul ? INSN_BL_N : INSN_BL;
716
717    if (l->has_value) {
718        tcg_target_long val = l->u.value;
719
720        val -= (tcg_target_long)s->code_ptr + 8;
721        val >>= 2;
722        assert(check_fit_tl(val, 17));
723
724        tcg_out32(s, op | reassemble_17(val));
725    } else {
726        /* We need to keep the offset unchanged for retranslation.  */
727        uint32_t old_insn = *(uint32_t *)s->code_ptr;
728
729        tcg_out_reloc(s, s->code_ptr, R_PARISC_PCREL17F, label_index, 0);
730        tcg_out32(s, op | (old_insn & 0x1f1ffdu));
731    }
732}
733
734static const uint8_t tcg_cond_to_cmp_cond[10] =
735{
736    [TCG_COND_EQ] = COND_EQ,
737    [TCG_COND_NE] = COND_EQ | COND_FALSE,
738    [TCG_COND_LT] = COND_LT,
739    [TCG_COND_GE] = COND_LT | COND_FALSE,
740    [TCG_COND_LE] = COND_LE,
741    [TCG_COND_GT] = COND_LE | COND_FALSE,
742    [TCG_COND_LTU] = COND_LTU,
743    [TCG_COND_GEU] = COND_LTU | COND_FALSE,
744    [TCG_COND_LEU] = COND_LEU,
745    [TCG_COND_GTU] = COND_LEU | COND_FALSE,
746};
747
748static void tcg_out_brcond(TCGContext *s, int cond, TCGArg c1,
749                           TCGArg c2, int c2const, int label_index)
750{
751    TCGLabel *l = &s->labels[label_index];
752    int op, pacond;
753
754    /* Note that COMIB operates as if the immediate is the first
755       operand.  We model brcond with the immediate in the second
756       to better match what targets are likely to give us.  For
757       consistency, model COMB with reversed operands as well.  */
758    pacond = tcg_cond_to_cmp_cond[tcg_swap_cond(cond)];
759
760    if (c2const) {
761        op = (pacond & COND_FALSE ? INSN_COMIBF : INSN_COMIBT);
762        op |= INSN_IM5(c2);
763    } else {
764        op = (pacond & COND_FALSE ? INSN_COMBF : INSN_COMBT);
765        op |= INSN_R1(c2);
766    }
767    op |= INSN_R2(c1);
768    op |= INSN_COND(pacond & 7);
769
770    if (l->has_value) {
771        tcg_target_long val = l->u.value;
772
773        val -= (tcg_target_long)s->code_ptr + 8;
774        val >>= 2;
775        assert(check_fit_tl(val, 12));
776
777        /* ??? Assume that all branches to defined labels are backward.
778           Which means that if the nul bit is set, the delay slot is
779           executed if the branch is taken, and not executed in fallthru.  */
780        tcg_out32(s, op | reassemble_12(val));
781        tcg_out_nop(s);
782    } else {
783        /* We need to keep the offset unchanged for retranslation.  */
784        uint32_t old_insn = *(uint32_t *)s->code_ptr;
785
786        tcg_out_reloc(s, s->code_ptr, R_PARISC_PCREL12F, label_index, 0);
787        /* ??? Assume that all branches to undefined labels are forward.
788           Which means that if the nul bit is set, the delay slot is
789           not executed if the branch is taken, which is what we want.  */
790        tcg_out32(s, op | 2 | (old_insn & 0x1ffdu));
791    }
792}
793
794static void tcg_out_comclr(TCGContext *s, int cond, TCGArg ret,
795                           TCGArg c1, TCGArg c2, int c2const)
796{
797    int op, pacond;
798
799    /* Note that COMICLR operates as if the immediate is the first
800       operand.  We model setcond with the immediate in the second
801       to better match what targets are likely to give us.  For
802       consistency, model COMCLR with reversed operands as well.  */
803    pacond = tcg_cond_to_cmp_cond[tcg_swap_cond(cond)];
804
805    if (c2const) {
806        op = INSN_COMICLR | INSN_R2(c1) | INSN_R1(ret) | INSN_IM11(c2);
807    } else {
808        op = INSN_COMCLR | INSN_R2(c1) | INSN_R1(c2) | INSN_T(ret);
809    }
810    op |= INSN_COND(pacond & 7);
811    op |= pacond & COND_FALSE ? 1 << 12 : 0;
812
813    tcg_out32(s, op);
814}
815
816static void tcg_out_brcond2(TCGContext *s, int cond, TCGArg al, TCGArg ah,
817                            TCGArg bl, int blconst, TCGArg bh, int bhconst,
818                            int label_index)
819{
820    switch (cond) {
821    case TCG_COND_EQ:
822    case TCG_COND_NE:
823        tcg_out_comclr(s, tcg_invert_cond(cond), TCG_REG_R0, al, bl, blconst);
824        tcg_out_brcond(s, cond, ah, bh, bhconst, label_index);
825        break;
826
827    default:
828        tcg_out_brcond(s, cond, ah, bh, bhconst, label_index);
829        tcg_out_comclr(s, TCG_COND_NE, TCG_REG_R0, ah, bh, bhconst);
830        tcg_out_brcond(s, tcg_unsigned_cond(cond),
831                       al, bl, blconst, label_index);
832        break;
833    }
834}
835
836static void tcg_out_setcond(TCGContext *s, int cond, TCGArg ret,
837                            TCGArg c1, TCGArg c2, int c2const)
838{
839    tcg_out_comclr(s, tcg_invert_cond(cond), ret, c1, c2, c2const);
840    tcg_out_movi(s, TCG_TYPE_I32, ret, 1);
841}
842
843static void tcg_out_setcond2(TCGContext *s, int cond, TCGArg ret,
844                             TCGArg al, TCGArg ah, TCGArg bl, int blconst,
845                             TCGArg bh, int bhconst)
846{
847    int scratch = TCG_REG_R20;
848
849    if (ret != al && ret != ah
850        && (blconst || ret != bl)
851        && (bhconst || ret != bh)) {
852        scratch = ret;
853    }
854
855    switch (cond) {
856    case TCG_COND_EQ:
857    case TCG_COND_NE:
858        tcg_out_setcond(s, cond, scratch, al, bl, blconst);
859        tcg_out_comclr(s, TCG_COND_EQ, TCG_REG_R0, ah, bh, bhconst);
860        tcg_out_movi(s, TCG_TYPE_I32, scratch, cond == TCG_COND_NE);
861        break;
862
863    default:
864        tcg_out_setcond(s, tcg_unsigned_cond(cond), scratch, al, bl, blconst);
865        tcg_out_comclr(s, TCG_COND_EQ, TCG_REG_R0, ah, bh, bhconst);
866        tcg_out_movi(s, TCG_TYPE_I32, scratch, 0);
867        tcg_out_comclr(s, cond, TCG_REG_R0, ah, bh, bhconst);
868        tcg_out_movi(s, TCG_TYPE_I32, scratch, 1);
869        break;
870    }
871
872    tcg_out_mov(s, TCG_TYPE_I32, ret, scratch);
873}
874
875#if defined(CONFIG_SOFTMMU)
876#include "../../softmmu_defs.h"
877
878static void *qemu_ld_helpers[4] = {
879    __ldb_mmu,
880    __ldw_mmu,
881    __ldl_mmu,
882    __ldq_mmu,
883};
884
885static void *qemu_st_helpers[4] = {
886    __stb_mmu,
887    __stw_mmu,
888    __stl_mmu,
889    __stq_mmu,
890};
891
892/* Load and compare a TLB entry, and branch if TLB miss.  OFFSET is set to
893   the offset of the first ADDR_READ or ADDR_WRITE member of the appropriate
894   TLB for the memory index.  The return value is the offset from ENV
895   contained in R1 afterward (to be used when loading ADDEND); if the
896   return value is 0, R1 is not used.  */
897
898static int tcg_out_tlb_read(TCGContext *s, int r0, int r1, int addrlo,
899                            int addrhi, int s_bits, int lab_miss, int offset)
900{
901    int ret;
902
903    /* Extracting the index into the TLB.  The "normal C operation" is
904          r1 = addr_reg >> TARGET_PAGE_BITS;
905          r1 &= CPU_TLB_SIZE - 1;
906          r1 <<= CPU_TLB_ENTRY_BITS;
907       What this does is extract CPU_TLB_BITS beginning at TARGET_PAGE_BITS
908       and place them at CPU_TLB_ENTRY_BITS.  We can combine the first two
909       operations with an EXTRU.  Unfortunately, the current value of
910       CPU_TLB_ENTRY_BITS is > 3, so we can't merge that shift with the
911       add that follows.  */
912    tcg_out_extr(s, r1, addrlo, TARGET_PAGE_BITS, CPU_TLB_BITS, 0);
913    tcg_out_shli(s, r1, r1, CPU_TLB_ENTRY_BITS);
914    tcg_out_arith(s, r1, r1, TCG_AREG0, INSN_ADDL);
915
916    /* Make sure that both the addr_{read,write} and addend can be
917       read with a 14-bit offset from the same base register.  */
918    if (check_fit_tl(offset + CPU_TLB_SIZE, 14)) {
919        ret = 0;
920    } else {
921        ret = (offset + 0x400) & ~0x7ff;
922        offset = ret - offset;
923        tcg_out_addi2(s, TCG_REG_R1, r1, ret);
924        r1 = TCG_REG_R1;
925    }
926
927    /* Load the entry from the computed slot.  */
928    if (TARGET_LONG_BITS == 64) {
929        tcg_out_ld(s, TCG_TYPE_PTR, TCG_REG_R23, r1, offset);
930        tcg_out_ld(s, TCG_TYPE_PTR, TCG_REG_R20, r1, offset + 4);
931    } else {
932        tcg_out_ld(s, TCG_TYPE_PTR, TCG_REG_R20, r1, offset);
933    }
934
935    /* Compute the value that ought to appear in the TLB for a hit, namely, the page
936       of the address.  We include the low N bits of the address to catch unaligned
937       accesses and force them onto the slow path.  Do this computation after having
938       issued the load from the TLB slot to give the load time to complete.  */
939    tcg_out_andi(s, r0, addrlo, TARGET_PAGE_MASK | ((1 << s_bits) - 1));
940
941    /* If not equal, jump to lab_miss. */
942    if (TARGET_LONG_BITS == 64) {
943        tcg_out_brcond2(s, TCG_COND_NE, TCG_REG_R20, TCG_REG_R23,
944                        r0, 0, addrhi, 0, lab_miss);
945    } else {
946        tcg_out_brcond(s, TCG_COND_NE, TCG_REG_R20, r0, 0, lab_miss);
947    }
948
949    return ret;
950}
951#endif
952
953static void tcg_out_qemu_ld_direct(TCGContext *s, int datalo_reg, int datahi_reg,
954                                   int addr_reg, int addend_reg, int opc)
955{
956#ifdef TARGET_WORDS_BIGENDIAN
957    const int bswap = 0;
958#else
959    const int bswap = 1;
960#endif
961
962    switch (opc) {
963    case 0:
964        tcg_out_ldst_index(s, datalo_reg, addr_reg, addend_reg, INSN_LDBX);
965        break;
966    case 0 | 4:
967        tcg_out_ldst_index(s, datalo_reg, addr_reg, addend_reg, INSN_LDBX);
968        tcg_out_ext8s(s, datalo_reg, datalo_reg);
969        break;
970    case 1:
971        tcg_out_ldst_index(s, datalo_reg, addr_reg, addend_reg, INSN_LDHX);
972        if (bswap) {
973            tcg_out_bswap16(s, datalo_reg, datalo_reg, 0);
974        }
975        break;
976    case 1 | 4:
977        tcg_out_ldst_index(s, datalo_reg, addr_reg, addend_reg, INSN_LDHX);
978        if (bswap) {
979            tcg_out_bswap16(s, datalo_reg, datalo_reg, 1);
980        } else {
981            tcg_out_ext16s(s, datalo_reg, datalo_reg);
982        }
983        break;
984    case 2:
985        tcg_out_ldst_index(s, datalo_reg, addr_reg, addend_reg, INSN_LDWX);
986        if (bswap) {
987            tcg_out_bswap32(s, datalo_reg, datalo_reg, TCG_REG_R20);
988        }
989        break;
990    case 3:
991        if (bswap) {
992            int t = datahi_reg;
993            datahi_reg = datalo_reg;
994            datalo_reg = t;
995        }
996        /* We can't access the low-part with a reg+reg addressing mode,
997           so perform the addition now and use reg_ofs addressing mode.  */
998        if (addend_reg != TCG_REG_R0) {
999            tcg_out_arith(s, TCG_REG_R20, addr_reg, addend_reg, INSN_ADD);
1000            addr_reg = TCG_REG_R20;
1001	}
1002        /* Make sure not to clobber the base register.  */
1003        if (datahi_reg == addr_reg) {
1004            tcg_out_ldst(s, datalo_reg, addr_reg, 4, INSN_LDW);
1005            tcg_out_ldst(s, datahi_reg, addr_reg, 0, INSN_LDW);
1006        } else {
1007            tcg_out_ldst(s, datahi_reg, addr_reg, 0, INSN_LDW);
1008            tcg_out_ldst(s, datalo_reg, addr_reg, 4, INSN_LDW);
1009        }
1010        if (bswap) {
1011            tcg_out_bswap32(s, datalo_reg, datalo_reg, TCG_REG_R20);
1012            tcg_out_bswap32(s, datahi_reg, datahi_reg, TCG_REG_R20);
1013        }
1014        break;
1015    default:
1016        tcg_abort();
1017    }
1018}
1019
1020static void tcg_out_qemu_ld(TCGContext *s, const TCGArg *args, int opc)
1021{
1022    int datalo_reg = *args++;
1023    /* Note that datahi_reg is only used for 64-bit loads.  */
1024    int datahi_reg = (opc == 3 ? *args++ : TCG_REG_R0);
1025    int addrlo_reg = *args++;
1026
1027#if defined(CONFIG_SOFTMMU)
1028    /* Note that addrhi_reg is only used for 64-bit guests.  */
1029    int addrhi_reg = (TARGET_LONG_BITS == 64 ? *args++ : TCG_REG_R0);
1030    int mem_index = *args;
1031    int lab1, lab2, argreg, offset;
1032
1033    lab1 = gen_new_label();
1034    lab2 = gen_new_label();
1035
1036    offset = offsetof(CPUState, tlb_table[mem_index][0].addr_read);
1037    offset = tcg_out_tlb_read(s, TCG_REG_R26, TCG_REG_R25, addrlo_reg, addrhi_reg,
1038                              opc & 3, lab1, offset);
1039
1040    /* TLB Hit.  */
1041    tcg_out_ld(s, TCG_TYPE_PTR, TCG_REG_R20, (offset ? TCG_REG_R1 : TCG_REG_R25),
1042               offsetof(CPUState, tlb_table[mem_index][0].addend) - offset);
1043    tcg_out_qemu_ld_direct(s, datalo_reg, datahi_reg, addrlo_reg, TCG_REG_R20, opc);
1044    tcg_out_branch(s, lab2, 1);
1045
1046    /* TLB Miss.  */
1047    /* label1: */
1048    tcg_out_label(s, lab1, (tcg_target_long)s->code_ptr);
1049
1050    argreg = TCG_REG_R26;
1051    tcg_out_mov(s, TCG_TYPE_I32, argreg--, addrlo_reg);
1052    if (TARGET_LONG_BITS == 64) {
1053        tcg_out_mov(s, TCG_TYPE_I32, argreg--, addrhi_reg);
1054    }
1055    tcg_out_movi(s, TCG_TYPE_I32, argreg, mem_index);
1056
1057    tcg_out_call(s, qemu_ld_helpers[opc & 3]);
1058
1059    switch (opc) {
1060    case 0:
1061        tcg_out_andi(s, datalo_reg, TCG_REG_RET0, 0xff);
1062        break;
1063    case 0 | 4:
1064        tcg_out_ext8s(s, datalo_reg, TCG_REG_RET0);
1065        break;
1066    case 1:
1067        tcg_out_andi(s, datalo_reg, TCG_REG_RET0, 0xffff);
1068        break;
1069    case 1 | 4:
1070        tcg_out_ext16s(s, datalo_reg, TCG_REG_RET0);
1071        break;
1072    case 2:
1073    case 2 | 4:
1074        tcg_out_mov(s, TCG_TYPE_I32, datalo_reg, TCG_REG_RET0);
1075        break;
1076    case 3:
1077        tcg_out_mov(s, TCG_TYPE_I32, datahi_reg, TCG_REG_RET0);
1078        tcg_out_mov(s, TCG_TYPE_I32, datalo_reg, TCG_REG_RET1);
1079        break;
1080    default:
1081        tcg_abort();
1082    }
1083
1084    /* label2: */
1085    tcg_out_label(s, lab2, (tcg_target_long)s->code_ptr);
1086#else
1087    tcg_out_qemu_ld_direct(s, datalo_reg, datahi_reg, addrlo_reg,
1088                           (GUEST_BASE ? TCG_GUEST_BASE_REG : TCG_REG_R0), opc);
1089#endif
1090}
1091
1092static void tcg_out_qemu_st_direct(TCGContext *s, int datalo_reg, int datahi_reg,
1093                                   int addr_reg, int opc)
1094{
1095#ifdef TARGET_WORDS_BIGENDIAN
1096    const int bswap = 0;
1097#else
1098    const int bswap = 1;
1099#endif
1100
1101    switch (opc) {
1102    case 0:
1103        tcg_out_ldst(s, datalo_reg, addr_reg, 0, INSN_STB);
1104        break;
1105    case 1:
1106        if (bswap) {
1107            tcg_out_bswap16(s, TCG_REG_R20, datalo_reg, 0);
1108            datalo_reg = TCG_REG_R20;
1109        }
1110        tcg_out_ldst(s, datalo_reg, addr_reg, 0, INSN_STH);
1111        break;
1112    case 2:
1113        if (bswap) {
1114            tcg_out_bswap32(s, TCG_REG_R20, datalo_reg, TCG_REG_R20);
1115            datalo_reg = TCG_REG_R20;
1116        }
1117        tcg_out_ldst(s, datalo_reg, addr_reg, 0, INSN_STW);
1118        break;
1119    case 3:
1120        if (bswap) {
1121            tcg_out_bswap32(s, TCG_REG_R20, datalo_reg, TCG_REG_R20);
1122            tcg_out_bswap32(s, TCG_REG_R23, datahi_reg, TCG_REG_R23);
1123            datahi_reg = TCG_REG_R20;
1124            datalo_reg = TCG_REG_R23;
1125        }
1126        tcg_out_ldst(s, datahi_reg, addr_reg, 0, INSN_STW);
1127        tcg_out_ldst(s, datalo_reg, addr_reg, 4, INSN_STW);
1128        break;
1129    default:
1130        tcg_abort();
1131    }
1132
1133}
1134
1135static void tcg_out_qemu_st(TCGContext *s, const TCGArg *args, int opc)
1136{
1137    int datalo_reg = *args++;
1138    /* Note that datahi_reg is only used for 64-bit loads.  */
1139    int datahi_reg = (opc == 3 ? *args++ : TCG_REG_R0);
1140    int addrlo_reg = *args++;
1141
1142#if defined(CONFIG_SOFTMMU)
1143    /* Note that addrhi_reg is only used for 64-bit guests.  */
1144    int addrhi_reg = (TARGET_LONG_BITS == 64 ? *args++ : TCG_REG_R0);
1145    int mem_index = *args;
1146    int lab1, lab2, argreg, offset;
1147
1148    lab1 = gen_new_label();
1149    lab2 = gen_new_label();
1150
1151    offset = offsetof(CPUState, tlb_table[mem_index][0].addr_write);
1152    offset = tcg_out_tlb_read(s, TCG_REG_R26, TCG_REG_R25, addrlo_reg, addrhi_reg,
1153                              opc, lab1, offset);
1154
1155    /* TLB Hit.  */
1156    tcg_out_ld(s, TCG_TYPE_PTR, TCG_REG_R20, (offset ? TCG_REG_R1 : TCG_REG_R25),
1157               offsetof(CPUState, tlb_table[mem_index][0].addend) - offset);
1158
1159    /* There are no indexed stores, so we must do this addition explitly.
1160       Careful to avoid R20, which is used for the bswaps to follow.  */
1161    tcg_out_arith(s, TCG_REG_R31, addrlo_reg, TCG_REG_R20, INSN_ADDL);
1162    tcg_out_qemu_st_direct(s, datalo_reg, datahi_reg, TCG_REG_R31, opc);
1163    tcg_out_branch(s, lab2, 1);
1164
1165    /* TLB Miss.  */
1166    /* label1: */
1167    tcg_out_label(s, lab1, (tcg_target_long)s->code_ptr);
1168
1169    argreg = TCG_REG_R26;
1170    tcg_out_mov(s, TCG_TYPE_I32, argreg--, addrlo_reg);
1171    if (TARGET_LONG_BITS == 64) {
1172        tcg_out_mov(s, TCG_TYPE_I32, argreg--, addrhi_reg);
1173    }
1174
1175    switch(opc) {
1176    case 0:
1177        tcg_out_andi(s, argreg--, datalo_reg, 0xff);
1178        tcg_out_movi(s, TCG_TYPE_I32, argreg, mem_index);
1179        break;
1180    case 1:
1181        tcg_out_andi(s, argreg--, datalo_reg, 0xffff);
1182        tcg_out_movi(s, TCG_TYPE_I32, argreg, mem_index);
1183        break;
1184    case 2:
1185        tcg_out_mov(s, TCG_TYPE_I32, argreg--, datalo_reg);
1186        tcg_out_movi(s, TCG_TYPE_I32, argreg, mem_index);
1187        break;
1188    case 3:
1189        /* Because of the alignment required by the 64-bit data argument,
1190           we will always use R23/R24.  Also, we will always run out of
1191           argument registers for storing mem_index, so that will have
1192           to go on the stack.  */
1193        if (mem_index == 0) {
1194            argreg = TCG_REG_R0;
1195        } else {
1196            argreg = TCG_REG_R20;
1197            tcg_out_movi(s, TCG_TYPE_I32, argreg, mem_index);
1198        }
1199        tcg_out_mov(s, TCG_TYPE_I32, TCG_REG_R23, datahi_reg);
1200        tcg_out_mov(s, TCG_TYPE_I32, TCG_REG_R24, datalo_reg);
1201        tcg_out_st(s, TCG_TYPE_I32, argreg, TCG_REG_SP,
1202                   TCG_TARGET_CALL_STACK_OFFSET - 4);
1203        break;
1204    default:
1205        tcg_abort();
1206    }
1207
1208    tcg_out_call(s, qemu_st_helpers[opc]);
1209
1210    /* label2: */
1211    tcg_out_label(s, lab2, (tcg_target_long)s->code_ptr);
1212#else
1213    /* There are no indexed stores, so if GUEST_BASE is set we must do the add
1214       explicitly.  Careful to avoid R20, which is used for the bswaps to follow.  */
1215    if (GUEST_BASE != 0) {
1216        tcg_out_arith(s, TCG_REG_R31, addrlo_reg, TCG_GUEST_BASE_REG, INSN_ADDL);
1217        addrlo_reg = TCG_REG_R31;
1218    }
1219    tcg_out_qemu_st_direct(s, datalo_reg, datahi_reg, addrlo_reg, opc);
1220#endif
1221}
1222
1223static void tcg_out_exit_tb(TCGContext *s, TCGArg arg)
1224{
1225    if (!check_fit_tl(arg, 14)) {
1226        uint32_t hi, lo;
1227        hi = arg & ~0x7ff;
1228        lo = arg & 0x7ff;
1229        if (lo) {
1230            tcg_out_movi(s, TCG_TYPE_PTR, TCG_REG_RET0, hi);
1231            tcg_out32(s, INSN_BV | INSN_R2(TCG_REG_R18));
1232            tcg_out_addi(s, TCG_REG_RET0, lo);
1233            return;
1234        }
1235        arg = hi;
1236    }
1237    tcg_out32(s, INSN_BV | INSN_R2(TCG_REG_R18));
1238    tcg_out_movi(s, TCG_TYPE_PTR, TCG_REG_RET0, arg);
1239}
1240
1241static void tcg_out_goto_tb(TCGContext *s, TCGArg arg)
1242{
1243    if (s->tb_jmp_offset) {
1244        /* direct jump method */
1245        fprintf(stderr, "goto_tb direct\n");
1246        tcg_abort();
1247    } else {
1248        /* indirect jump method */
1249        tcg_out_ld(s, TCG_TYPE_PTR, TCG_REG_R20, TCG_REG_R0,
1250                   (tcg_target_long)(s->tb_next + arg));
1251        tcg_out32(s, INSN_BV_N | INSN_R2(TCG_REG_R20));
1252    }
1253    s->tb_next_offset[arg] = s->code_ptr - s->code_buf;
1254}
1255
1256static inline void tcg_out_op(TCGContext *s, TCGOpcode opc, const TCGArg *args,
1257                              const int *const_args)
1258{
1259    switch (opc) {
1260    case INDEX_op_exit_tb:
1261        tcg_out_exit_tb(s, args[0]);
1262        break;
1263    case INDEX_op_goto_tb:
1264        tcg_out_goto_tb(s, args[0]);
1265        break;
1266
1267    case INDEX_op_call:
1268        if (const_args[0]) {
1269            tcg_out_call(s, (void *)args[0]);
1270        } else {
1271            /* ??? FIXME: the value in the register in args[0] is almost
1272               certainly a procedure descriptor, not a code address.  We
1273               probably need to use the millicode $$dyncall routine.  */
1274            tcg_abort();
1275        }
1276        break;
1277
1278    case INDEX_op_jmp:
1279        fprintf(stderr, "unimplemented jmp\n");
1280        tcg_abort();
1281        break;
1282
1283    case INDEX_op_br:
1284        tcg_out_branch(s, args[0], 1);
1285        break;
1286
1287    case INDEX_op_movi_i32:
1288        tcg_out_movi(s, TCG_TYPE_I32, args[0], (uint32_t)args[1]);
1289        break;
1290
1291    case INDEX_op_ld8u_i32:
1292        tcg_out_ldst(s, args[0], args[1], args[2], INSN_LDB);
1293        break;
1294    case INDEX_op_ld8s_i32:
1295        tcg_out_ldst(s, args[0], args[1], args[2], INSN_LDB);
1296        tcg_out_ext8s(s, args[0], args[0]);
1297        break;
1298    case INDEX_op_ld16u_i32:
1299        tcg_out_ldst(s, args[0], args[1], args[2], INSN_LDH);
1300        break;
1301    case INDEX_op_ld16s_i32:
1302        tcg_out_ldst(s, args[0], args[1], args[2], INSN_LDH);
1303        tcg_out_ext16s(s, args[0], args[0]);
1304        break;
1305    case INDEX_op_ld_i32:
1306        tcg_out_ldst(s, args[0], args[1], args[2], INSN_LDW);
1307        break;
1308
1309    case INDEX_op_st8_i32:
1310        tcg_out_ldst(s, args[0], args[1], args[2], INSN_STB);
1311        break;
1312    case INDEX_op_st16_i32:
1313        tcg_out_ldst(s, args[0], args[1], args[2], INSN_STH);
1314        break;
1315    case INDEX_op_st_i32:
1316        tcg_out_ldst(s, args[0], args[1], args[2], INSN_STW);
1317        break;
1318
1319    case INDEX_op_add_i32:
1320        if (const_args[2]) {
1321            tcg_out_addi2(s, args[0], args[1], args[2]);
1322        } else {
1323            tcg_out_arith(s, args[0], args[1], args[2], INSN_ADDL);
1324        }
1325        break;
1326
1327    case INDEX_op_sub_i32:
1328        if (const_args[1]) {
1329            if (const_args[2]) {
1330                tcg_out_movi(s, TCG_TYPE_I32, args[0], args[1] - args[2]);
1331            } else {
1332                /* Recall that SUBI is a reversed subtract.  */
1333                tcg_out_arithi(s, args[0], args[2], args[1], INSN_SUBI);
1334            }
1335        } else if (const_args[2]) {
1336            tcg_out_addi2(s, args[0], args[1], -args[2]);
1337        } else {
1338            tcg_out_arith(s, args[0], args[1], args[2], INSN_SUB);
1339        }
1340        break;
1341
1342    case INDEX_op_and_i32:
1343        if (const_args[2]) {
1344            tcg_out_andi(s, args[0], args[1], args[2]);
1345        } else {
1346            tcg_out_arith(s, args[0], args[1], args[2], INSN_AND);
1347        }
1348        break;
1349
1350    case INDEX_op_or_i32:
1351        if (const_args[2]) {
1352            tcg_out_ori(s, args[0], args[1], args[2]);
1353        } else {
1354            tcg_out_arith(s, args[0], args[1], args[2], INSN_OR);
1355        }
1356        break;
1357
1358    case INDEX_op_xor_i32:
1359        tcg_out_arith(s, args[0], args[1], args[2], INSN_XOR);
1360        break;
1361
1362    case INDEX_op_andc_i32:
1363        if (const_args[2]) {
1364            tcg_out_andi(s, args[0], args[1], ~args[2]);
1365        } else {
1366            tcg_out_arith(s, args[0], args[1], args[2], INSN_ANDCM);
1367        }
1368        break;
1369
1370    case INDEX_op_shl_i32:
1371        if (const_args[2]) {
1372            tcg_out_shli(s, args[0], args[1], args[2]);
1373        } else {
1374            tcg_out_shl(s, args[0], args[1], args[2]);
1375        }
1376        break;
1377
1378    case INDEX_op_shr_i32:
1379        if (const_args[2]) {
1380            tcg_out_shri(s, args[0], args[1], args[2]);
1381        } else {
1382            tcg_out_shr(s, args[0], args[1], args[2]);
1383        }
1384        break;
1385
1386    case INDEX_op_sar_i32:
1387        if (const_args[2]) {
1388            tcg_out_sari(s, args[0], args[1], args[2]);
1389        } else {
1390            tcg_out_sar(s, args[0], args[1], args[2]);
1391        }
1392        break;
1393
1394    case INDEX_op_rotl_i32:
1395        if (const_args[2]) {
1396            tcg_out_rotli(s, args[0], args[1], args[2]);
1397        } else {
1398            tcg_out_rotl(s, args[0], args[1], args[2]);
1399        }
1400        break;
1401
1402    case INDEX_op_rotr_i32:
1403        if (const_args[2]) {
1404            tcg_out_rotri(s, args[0], args[1], args[2]);
1405        } else {
1406            tcg_out_rotr(s, args[0], args[1], args[2]);
1407        }
1408        break;
1409
1410    case INDEX_op_mul_i32:
1411        tcg_out_xmpyu(s, args[0], TCG_REG_R0, args[1], args[2]);
1412        break;
1413    case INDEX_op_mulu2_i32:
1414        tcg_out_xmpyu(s, args[0], args[1], args[2], args[3]);
1415        break;
1416
1417    case INDEX_op_bswap16_i32:
1418        tcg_out_bswap16(s, args[0], args[1], 0);
1419        break;
1420    case INDEX_op_bswap32_i32:
1421        tcg_out_bswap32(s, args[0], args[1], TCG_REG_R20);
1422        break;
1423
1424    case INDEX_op_not_i32:
1425        tcg_out_arithi(s, args[0], args[1], -1, INSN_SUBI);
1426        break;
1427    case INDEX_op_ext8s_i32:
1428        tcg_out_ext8s(s, args[0], args[1]);
1429        break;
1430    case INDEX_op_ext16s_i32:
1431        tcg_out_ext16s(s, args[0], args[1]);
1432        break;
1433
1434    case INDEX_op_brcond_i32:
1435        tcg_out_brcond(s, args[2], args[0], args[1], const_args[1], args[3]);
1436        break;
1437    case INDEX_op_brcond2_i32:
1438        tcg_out_brcond2(s, args[4], args[0], args[1],
1439                        args[2], const_args[2],
1440                        args[3], const_args[3], args[5]);
1441        break;
1442
1443    case INDEX_op_setcond_i32:
1444        tcg_out_setcond(s, args[3], args[0], args[1], args[2], const_args[2]);
1445        break;
1446    case INDEX_op_setcond2_i32:
1447        tcg_out_setcond2(s, args[5], args[0], args[1], args[2],
1448                         args[3], const_args[3], args[4], const_args[4]);
1449        break;
1450
1451    case INDEX_op_add2_i32:
1452        tcg_out_add2(s, args[0], args[1], args[2], args[3],
1453                     args[4], args[5], const_args[4]);
1454        break;
1455
1456    case INDEX_op_sub2_i32:
1457        tcg_out_sub2(s, args[0], args[1], args[2], args[3],
1458                     args[4], args[5], const_args[2], const_args[4]);
1459        break;
1460
1461    case INDEX_op_qemu_ld8u:
1462        tcg_out_qemu_ld(s, args, 0);
1463        break;
1464    case INDEX_op_qemu_ld8s:
1465        tcg_out_qemu_ld(s, args, 0 | 4);
1466        break;
1467    case INDEX_op_qemu_ld16u:
1468        tcg_out_qemu_ld(s, args, 1);
1469        break;
1470    case INDEX_op_qemu_ld16s:
1471        tcg_out_qemu_ld(s, args, 1 | 4);
1472        break;
1473    case INDEX_op_qemu_ld32:
1474        tcg_out_qemu_ld(s, args, 2);
1475        break;
1476    case INDEX_op_qemu_ld64:
1477        tcg_out_qemu_ld(s, args, 3);
1478        break;
1479
1480    case INDEX_op_qemu_st8:
1481        tcg_out_qemu_st(s, args, 0);
1482        break;
1483    case INDEX_op_qemu_st16:
1484        tcg_out_qemu_st(s, args, 1);
1485        break;
1486    case INDEX_op_qemu_st32:
1487        tcg_out_qemu_st(s, args, 2);
1488        break;
1489    case INDEX_op_qemu_st64:
1490        tcg_out_qemu_st(s, args, 3);
1491        break;
1492
1493    default:
1494        fprintf(stderr, "unknown opcode 0x%x\n", opc);
1495        tcg_abort();
1496    }
1497}
1498
1499static const TCGTargetOpDef hppa_op_defs[] = {
1500    { INDEX_op_exit_tb, { } },
1501    { INDEX_op_goto_tb, { } },
1502
1503    { INDEX_op_call, { "ri" } },
1504    { INDEX_op_jmp, { "r" } },
1505    { INDEX_op_br, { } },
1506
1507    { INDEX_op_mov_i32, { "r", "r" } },
1508    { INDEX_op_movi_i32, { "r" } },
1509
1510    { INDEX_op_ld8u_i32, { "r", "r" } },
1511    { INDEX_op_ld8s_i32, { "r", "r" } },
1512    { INDEX_op_ld16u_i32, { "r", "r" } },
1513    { INDEX_op_ld16s_i32, { "r", "r" } },
1514    { INDEX_op_ld_i32, { "r", "r" } },
1515    { INDEX_op_st8_i32, { "rZ", "r" } },
1516    { INDEX_op_st16_i32, { "rZ", "r" } },
1517    { INDEX_op_st_i32, { "rZ", "r" } },
1518
1519    { INDEX_op_add_i32, { "r", "rZ", "ri" } },
1520    { INDEX_op_sub_i32, { "r", "rI", "ri" } },
1521    { INDEX_op_and_i32, { "r", "rZ", "rM" } },
1522    { INDEX_op_or_i32, { "r", "rZ", "rO" } },
1523    { INDEX_op_xor_i32, { "r", "rZ", "rZ" } },
1524    /* Note that the second argument will be inverted, which means
1525       we want a constant whose inversion matches M, and that O = ~M.
1526       See the implementation of and_mask_p.  */
1527    { INDEX_op_andc_i32, { "r", "rZ", "rO" } },
1528
1529    { INDEX_op_mul_i32, { "r", "r", "r" } },
1530    { INDEX_op_mulu2_i32, { "r", "r", "r", "r" } },
1531
1532    { INDEX_op_shl_i32, { "r", "r", "ri" } },
1533    { INDEX_op_shr_i32, { "r", "r", "ri" } },
1534    { INDEX_op_sar_i32, { "r", "r", "ri" } },
1535    { INDEX_op_rotl_i32, { "r", "r", "ri" } },
1536    { INDEX_op_rotr_i32, { "r", "r", "ri" } },
1537
1538    { INDEX_op_bswap16_i32, { "r", "r" } },
1539    { INDEX_op_bswap32_i32, { "r", "r" } },
1540    { INDEX_op_not_i32, { "r", "r" } },
1541
1542    { INDEX_op_ext8s_i32, { "r", "r" } },
1543    { INDEX_op_ext16s_i32, { "r", "r" } },
1544
1545    { INDEX_op_brcond_i32, { "rZ", "rJ" } },
1546    { INDEX_op_brcond2_i32,  { "rZ", "rZ", "rJ", "rJ" } },
1547
1548    { INDEX_op_setcond_i32, { "r", "rZ", "rI" } },
1549    { INDEX_op_setcond2_i32, { "r", "rZ", "rZ", "rI", "rI" } },
1550
1551    { INDEX_op_add2_i32, { "r", "r", "rZ", "rZ", "rI", "rZ" } },
1552    { INDEX_op_sub2_i32, { "r", "r", "rI", "rZ", "rK", "rZ" } },
1553
1554#if TARGET_LONG_BITS == 32
1555    { INDEX_op_qemu_ld8u, { "r", "L" } },
1556    { INDEX_op_qemu_ld8s, { "r", "L" } },
1557    { INDEX_op_qemu_ld16u, { "r", "L" } },
1558    { INDEX_op_qemu_ld16s, { "r", "L" } },
1559    { INDEX_op_qemu_ld32, { "r", "L" } },
1560    { INDEX_op_qemu_ld64, { "r", "r", "L" } },
1561
1562    { INDEX_op_qemu_st8, { "LZ", "L" } },
1563    { INDEX_op_qemu_st16, { "LZ", "L" } },
1564    { INDEX_op_qemu_st32, { "LZ", "L" } },
1565    { INDEX_op_qemu_st64, { "LZ", "LZ", "L" } },
1566#else
1567    { INDEX_op_qemu_ld8u, { "r", "L", "L" } },
1568    { INDEX_op_qemu_ld8s, { "r", "L", "L" } },
1569    { INDEX_op_qemu_ld16u, { "r", "L", "L" } },
1570    { INDEX_op_qemu_ld16s, { "r", "L", "L" } },
1571    { INDEX_op_qemu_ld32, { "r", "L", "L" } },
1572    { INDEX_op_qemu_ld64, { "r", "r", "L", "L" } },
1573
1574    { INDEX_op_qemu_st8, { "LZ", "L", "L" } },
1575    { INDEX_op_qemu_st16, { "LZ", "L", "L" } },
1576    { INDEX_op_qemu_st32, { "LZ", "L", "L" } },
1577    { INDEX_op_qemu_st64, { "LZ", "LZ", "L", "L" } },
1578#endif
1579    { -1 },
1580};
1581
1582static int tcg_target_callee_save_regs[] = {
1583    /* R2, the return address register, is saved specially
1584       in the caller's frame.  */
1585    /* R3, the frame pointer, is not currently modified.  */
1586    TCG_REG_R4,
1587    TCG_REG_R5,
1588    TCG_REG_R6,
1589    TCG_REG_R7,
1590    TCG_REG_R8,
1591    TCG_REG_R9,
1592    TCG_REG_R10,
1593    TCG_REG_R11,
1594    TCG_REG_R12,
1595    TCG_REG_R13,
1596    TCG_REG_R14,
1597    TCG_REG_R15,
1598    TCG_REG_R16,
1599    /* R17 is the global env, so no need to save.  */
1600    TCG_REG_R18
1601};
1602
1603static void tcg_target_qemu_prologue(TCGContext *s)
1604{
1605    int frame_size, i;
1606
1607    /* Allocate space for the fixed frame marker.  */
1608    frame_size = -TCG_TARGET_CALL_STACK_OFFSET;
1609    frame_size += TCG_TARGET_STATIC_CALL_ARGS_SIZE;
1610
1611    /* Allocate space for the saved registers.  */
1612    frame_size += ARRAY_SIZE(tcg_target_callee_save_regs) * 4;
1613
1614    /* Align the allocated space.  */
1615    frame_size = ((frame_size + TCG_TARGET_STACK_ALIGN - 1)
1616                  & -TCG_TARGET_STACK_ALIGN);
1617
1618    /* The return address is stored in the caller's frame.  */
1619    tcg_out_st(s, TCG_TYPE_PTR, TCG_REG_RP, TCG_REG_SP, -20);
1620
1621    /* Allocate stack frame, saving the first register at the same time.  */
1622    tcg_out_ldst(s, tcg_target_callee_save_regs[0],
1623                 TCG_REG_SP, frame_size, INSN_STWM);
1624
1625    /* Save all callee saved registers.  */
1626    for (i = 1; i < ARRAY_SIZE(tcg_target_callee_save_regs); i++) {
1627        tcg_out_st(s, TCG_TYPE_PTR, tcg_target_callee_save_regs[i],
1628                   TCG_REG_SP, -frame_size + i * 4);
1629    }
1630
1631#ifdef CONFIG_USE_GUEST_BASE
1632    if (GUEST_BASE != 0) {
1633        tcg_out_movi(s, TCG_TYPE_PTR, TCG_GUEST_BASE_REG, GUEST_BASE);
1634        tcg_regset_set_reg(s->reserved_regs, TCG_GUEST_BASE_REG);
1635    }
1636#endif
1637
1638    /* Jump to TB, and adjust R18 to be the return address.  */
1639    tcg_out32(s, INSN_BLE_SR4 | INSN_R2(TCG_REG_R26));
1640    tcg_out_mov(s, TCG_TYPE_I32, TCG_REG_R18, TCG_REG_R31);
1641
1642    /* Restore callee saved registers.  */
1643    tcg_out_ld(s, TCG_TYPE_PTR, TCG_REG_RP, TCG_REG_SP, -frame_size - 20);
1644    for (i = 1; i < ARRAY_SIZE(tcg_target_callee_save_regs); i++) {
1645        tcg_out_ld(s, TCG_TYPE_PTR, tcg_target_callee_save_regs[i],
1646                   TCG_REG_SP, -frame_size + i * 4);
1647    }
1648
1649    /* Deallocate stack frame and return.  */
1650    tcg_out32(s, INSN_BV | INSN_R2(TCG_REG_RP));
1651    tcg_out_ldst(s, tcg_target_callee_save_regs[0],
1652                 TCG_REG_SP, -frame_size, INSN_LDWM);
1653}
1654
1655static void tcg_target_init(TCGContext *s)
1656{
1657    tcg_regset_set32(tcg_target_available_regs[TCG_TYPE_I32], 0, 0xffffffff);
1658
1659    tcg_regset_clear(tcg_target_call_clobber_regs);
1660    tcg_regset_set_reg(tcg_target_call_clobber_regs, TCG_REG_R20);
1661    tcg_regset_set_reg(tcg_target_call_clobber_regs, TCG_REG_R21);
1662    tcg_regset_set_reg(tcg_target_call_clobber_regs, TCG_REG_R22);
1663    tcg_regset_set_reg(tcg_target_call_clobber_regs, TCG_REG_R23);
1664    tcg_regset_set_reg(tcg_target_call_clobber_regs, TCG_REG_R24);
1665    tcg_regset_set_reg(tcg_target_call_clobber_regs, TCG_REG_R25);
1666    tcg_regset_set_reg(tcg_target_call_clobber_regs, TCG_REG_R26);
1667    tcg_regset_set_reg(tcg_target_call_clobber_regs, TCG_REG_RET0);
1668    tcg_regset_set_reg(tcg_target_call_clobber_regs, TCG_REG_RET1);
1669
1670    tcg_regset_clear(s->reserved_regs);
1671    tcg_regset_set_reg(s->reserved_regs, TCG_REG_R0);  /* hardwired to zero */
1672    tcg_regset_set_reg(s->reserved_regs, TCG_REG_R1);  /* addil target */
1673    tcg_regset_set_reg(s->reserved_regs, TCG_REG_RP);  /* link register */
1674    tcg_regset_set_reg(s->reserved_regs, TCG_REG_R3);  /* frame pointer */
1675    tcg_regset_set_reg(s->reserved_regs, TCG_REG_R18); /* return pointer */
1676    tcg_regset_set_reg(s->reserved_regs, TCG_REG_R19); /* clobbered w/o pic */
1677    tcg_regset_set_reg(s->reserved_regs, TCG_REG_R20); /* reserved */
1678    tcg_regset_set_reg(s->reserved_regs, TCG_REG_DP);  /* data pointer */
1679    tcg_regset_set_reg(s->reserved_regs, TCG_REG_SP);  /* stack pointer */
1680    tcg_regset_set_reg(s->reserved_regs, TCG_REG_R31); /* ble link reg */
1681
1682    tcg_add_target_add_op_defs(hppa_op_defs);
1683}
1684