assemble_arm64.cc revision 9f975bfe091e9592a1b6b5b46d224ec04b1183b6
1/* 2 * Copyright (C) 2011 The Android Open Source Project 3 * 4 * Licensed under the Apache License, Version 2.0 (the "License"); 5 * you may not use this file except in compliance with the License. 6 * You may obtain a copy of the License at 7 * 8 * http://www.apache.org/licenses/LICENSE-2.0 9 * 10 * Unless required by applicable law or agreed to in writing, software 11 * distributed under the License is distributed on an "AS IS" BASIS, 12 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 13 * See the License for the specific language governing permissions and 14 * limitations under the License. 15 */ 16 17#include "arm64_lir.h" 18#include "codegen_arm64.h" 19#include "dex/quick/mir_to_lir-inl.h" 20 21namespace art { 22 23// The macros below are exclusively used in the encoding map. 24 25// Most generic way of providing two variants for one instructions. 26#define CUSTOM_VARIANTS(variant1, variant2) variant1, variant2 27 28// Used for instructions which do not have a wide variant. 29#define NO_VARIANTS(variant) \ 30 CUSTOM_VARIANTS(variant, 0) 31 32// Used for instructions which have a wide variant with the sf bit set to 1. 33#define SF_VARIANTS(sf0_skeleton) \ 34 CUSTOM_VARIANTS(sf0_skeleton, (sf0_skeleton | 0x80000000)) 35 36// Used for instructions which have a wide variant with the size bits set to either x0 or x1. 37#define SIZE_VARIANTS(sizex0_skeleton) \ 38 CUSTOM_VARIANTS(sizex0_skeleton, (sizex0_skeleton | 0x40000000)) 39 40// Used for instructions which have a wide variant with the sf and n bits set to 1. 41#define SF_N_VARIANTS(sf0_n0_skeleton) \ 42 CUSTOM_VARIANTS(sf0_n0_skeleton, (sf0_n0_skeleton | 0x80400000)) 43 44// Used for FP instructions which have a single and double precision variants, with he type bits set 45// to either 00 or 01. 46#define FLOAT_VARIANTS(type00_skeleton) \ 47 CUSTOM_VARIANTS(type00_skeleton, (type00_skeleton | 0x00400000)) 48 49/* 50 * opcode: ArmOpcode enum 51 * variants: instruction skeletons supplied via CUSTOM_VARIANTS or derived macros. 52 * a{n}k: key to applying argument {n} \ 53 * a{n}s: argument {n} start bit position | n = 0, 1, 2, 3 54 * a{n}e: argument {n} end bit position / 55 * flags: instruction attributes (used in optimization) 56 * name: mnemonic name 57 * fmt: for pretty-printing 58 * fixup: used for second-pass fixes (e.g. adresses fixups in branch instructions). 59 */ 60#define ENCODING_MAP(opcode, variants, a0k, a0s, a0e, a1k, a1s, a1e, a2k, a2s, a2e, \ 61 a3k, a3s, a3e, flags, name, fmt, fixup) \ 62 {variants, {{a0k, a0s, a0e}, {a1k, a1s, a1e}, {a2k, a2s, a2e}, \ 63 {a3k, a3s, a3e}}, opcode, flags, name, fmt, 4, fixup} 64 65/* Instruction dump string format keys: !pf, where "!" is the start 66 * of the key, "p" is which numeric operand to use and "f" is the 67 * print format. 68 * 69 * [p]ositions: 70 * 0 -> operands[0] (dest) 71 * 1 -> operands[1] (src1) 72 * 2 -> operands[2] (src2) 73 * 3 -> operands[3] (extra) 74 * 75 * [f]ormats: 76 * d -> decimal 77 * D -> decimal*4 or decimal*8 depending on the instruction width 78 * E -> decimal*4 79 * F -> decimal*2 80 * G -> ", lsl #2" or ", lsl #3" depending on the instruction width 81 * c -> branch condition (eq, ne, etc.) 82 * t -> pc-relative target 83 * p -> pc-relative address 84 * s -> single precision floating point register 85 * S -> double precision floating point register 86 * f -> single or double precision register (depending on instruction width) 87 * I -> 8-bit immediate floating point number 88 * l -> logical immediate 89 * M -> 16-bit shift expression ("" or ", lsl #16" or ", lsl #32"...) 90 * B -> dmb option string (sy, st, ish, ishst, nsh, hshst) 91 * H -> operand shift 92 * T -> register shift (either ", lsl #0" or ", lsl #12") 93 * e -> register extend (e.g. uxtb #1) 94 * o -> register shift (e.g. lsl #1) for Word registers 95 * w -> word (32-bit) register wn, or wzr 96 * W -> word (32-bit) register wn, or wsp 97 * x -> extended (64-bit) register xn, or xzr 98 * X -> extended (64-bit) register xn, or sp 99 * r -> register with same width as instruction, r31 -> wzr, xzr 100 * R -> register with same width as instruction, r31 -> wsp, sp 101 * 102 * [!] escape. To insert "!", use "!!" 103 */ 104/* NOTE: must be kept in sync with enum ArmOpcode from arm64_lir.h */ 105const ArmEncodingMap Arm64Mir2Lir::EncodingMap[kA64Last] = { 106 ENCODING_MAP(WIDE(kA64Adc3rrr), SF_VARIANTS(0x1a000000), 107 kFmtRegR, 4, 0, kFmtRegR, 9, 5, kFmtRegR, 20, 16, 108 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE1, 109 "adc", "!0r, !1r, !2r", kFixupNone), 110 ENCODING_MAP(WIDE(kA64Add4RRdT), SF_VARIANTS(0x11000000), 111 kFmtRegROrSp, 4, 0, kFmtRegROrSp, 9, 5, kFmtBitBlt, 21, 10, 112 kFmtBitBlt, 23, 22, IS_QUAD_OP | REG_DEF0_USE1, 113 "add", "!0R, !1R, #!2d!3T", kFixupNone), 114 ENCODING_MAP(WIDE(kA64Add4rrro), SF_VARIANTS(0x0b000000), 115 kFmtRegR, 4, 0, kFmtRegR, 9, 5, kFmtRegR, 20, 16, 116 kFmtShift, -1, -1, IS_QUAD_OP | REG_DEF0_USE1, 117 "add", "!0r, !1r, !2r!3o", kFixupNone), 118 ENCODING_MAP(WIDE(kA64Add4rrre), SF_VARIANTS(0x0b200000), 119 kFmtRegROrSp, 4, 0, kFmtRegROrSp, 9, 5, kFmtRegR, 20, 16, 120 kFmtExtend, -1, -1, IS_QUAD_OP | REG_DEF0_USE12, 121 "add", "!0r, !1r, !2r!3o", kFixupNone), 122 // Note: adr is binary, but declared as tertiary. The third argument is used while doing the 123 // fixups and contains information to identify the adr label. 124 ENCODING_MAP(kA64Adr2xd, NO_VARIANTS(0x10000000), 125 kFmtRegX, 4, 0, kFmtImm21, -1, -1, kFmtUnused, -1, -1, 126 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0 | NEEDS_FIXUP, 127 "adr", "!0x, #!1d", kFixupAdr), 128 ENCODING_MAP(WIDE(kA64And3Rrl), SF_VARIANTS(0x12000000), 129 kFmtRegROrSp, 4, 0, kFmtRegR, 9, 5, kFmtBitBlt, 22, 10, 130 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE1, 131 "and", "!0R, !1r, #!2l", kFixupNone), 132 ENCODING_MAP(WIDE(kA64And4rrro), SF_VARIANTS(0x0a000000), 133 kFmtRegR, 4, 0, kFmtRegR, 9, 5, kFmtRegR, 20, 16, 134 kFmtShift, -1, -1, IS_QUAD_OP | REG_DEF0_USE12, 135 "and", "!0r, !1r, !2r!3o", kFixupNone), 136 ENCODING_MAP(WIDE(kA64Asr3rrd), CUSTOM_VARIANTS(0x13007c00, 0x9340fc00), 137 kFmtRegR, 4, 0, kFmtRegR, 9, 5, kFmtBitBlt, 21, 16, 138 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE1, 139 "asr", "!0r, !1r, #!2d", kFixupNone), 140 ENCODING_MAP(WIDE(kA64Asr3rrr), SF_VARIANTS(0x1ac02800), 141 kFmtRegR, 4, 0, kFmtRegR, 9, 5, kFmtRegR, 20, 16, 142 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE12, 143 "asr", "!0r, !1r, !2r", kFixupNone), 144 ENCODING_MAP(kA64B2ct, NO_VARIANTS(0x54000000), 145 kFmtBitBlt, 3, 0, kFmtBitBlt, 23, 5, kFmtUnused, -1, -1, 146 kFmtUnused, -1, -1, IS_BINARY_OP | IS_BRANCH | USES_CCODES | 147 NEEDS_FIXUP, "b.!0c", "!1t", kFixupCondBranch), 148 ENCODING_MAP(kA64Blr1x, NO_VARIANTS(0xd63f0000), 149 kFmtRegX, 9, 5, kFmtUnused, -1, -1, kFmtUnused, -1, -1, 150 kFmtUnused, -1, -1, 151 IS_UNARY_OP | REG_USE0 | IS_BRANCH | REG_DEF_LR, 152 "blr", "!0x", kFixupNone), 153 ENCODING_MAP(kA64Br1x, NO_VARIANTS(0xd61f0000), 154 kFmtRegX, 9, 5, kFmtUnused, -1, -1, kFmtUnused, -1, -1, 155 kFmtUnused, -1, -1, IS_UNARY_OP | REG_USE0 | IS_BRANCH, 156 "br", "!0x", kFixupNone), 157 ENCODING_MAP(kA64Brk1d, NO_VARIANTS(0xd4200000), 158 kFmtBitBlt, 20, 5, kFmtUnused, -1, -1, kFmtUnused, -1, -1, 159 kFmtUnused, -1, -1, IS_UNARY_OP | IS_BRANCH, 160 "brk", "!0d", kFixupNone), 161 ENCODING_MAP(kA64B1t, NO_VARIANTS(0x14000000), 162 kFmtBitBlt, 25, 0, kFmtUnused, -1, -1, kFmtUnused, -1, -1, 163 kFmtUnused, -1, -1, IS_UNARY_OP | IS_BRANCH | NEEDS_FIXUP, 164 "b", "!0t", kFixupT1Branch), 165 ENCODING_MAP(WIDE(kA64Cbnz2rt), SF_VARIANTS(0x35000000), 166 kFmtRegR, 4, 0, kFmtBitBlt, 23, 5, kFmtUnused, -1, -1, 167 kFmtUnused, -1, -1, 168 IS_BINARY_OP | REG_USE0 | IS_BRANCH | NEEDS_FIXUP, 169 "cbnz", "!0r, !1t", kFixupCBxZ), 170 ENCODING_MAP(WIDE(kA64Cbz2rt), SF_VARIANTS(0x34000000), 171 kFmtRegR, 4, 0, kFmtBitBlt, 23, 5, kFmtUnused, -1, -1, 172 kFmtUnused, -1, -1, 173 IS_BINARY_OP | REG_USE0 | IS_BRANCH | NEEDS_FIXUP, 174 "cbz", "!0r, !1t", kFixupCBxZ), 175 ENCODING_MAP(WIDE(kA64Cmn3rro), SF_VARIANTS(0x2b00001f), 176 kFmtRegR, 9, 5, kFmtRegR, 20, 16, kFmtShift, -1, -1, 177 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_USE01 | SETS_CCODES, 178 "cmn", "!0r, !1r!2o", kFixupNone), 179 ENCODING_MAP(WIDE(kA64Cmn3Rre), SF_VARIANTS(0x2b20001f), 180 kFmtRegROrSp, 9, 5, kFmtRegR, 20, 16, kFmtExtend, -1, -1, 181 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_USE01 | SETS_CCODES, 182 "cmn", "!0R, !1r!2e", kFixupNone), 183 ENCODING_MAP(WIDE(kA64Cmn3RdT), SF_VARIANTS(0x3100001f), 184 kFmtRegROrSp, 9, 5, kFmtBitBlt, 21, 10, kFmtBitBlt, 23, 22, 185 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_USE0 | SETS_CCODES, 186 "cmn", "!0R, #!1d!2T", kFixupNone), 187 ENCODING_MAP(WIDE(kA64Cmp3rro), SF_VARIANTS(0x6b00001f), 188 kFmtRegR, 9, 5, kFmtRegR, 20, 16, kFmtShift, -1, -1, 189 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_USE01 | SETS_CCODES, 190 "cmp", "!0r, !1r!2o", kFixupNone), 191 ENCODING_MAP(WIDE(kA64Cmp3Rre), SF_VARIANTS(0x6b20001f), 192 kFmtRegROrSp, 9, 5, kFmtRegR, 20, 16, kFmtExtend, -1, -1, 193 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_USE01 | SETS_CCODES, 194 "cmp", "!0R, !1r!2e", kFixupNone), 195 ENCODING_MAP(WIDE(kA64Cmp3RdT), SF_VARIANTS(0x7100001f), 196 kFmtRegROrSp, 9, 5, kFmtBitBlt, 21, 10, kFmtBitBlt, 23, 22, 197 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_USE0 | SETS_CCODES, 198 "cmp", "!0R, #!1d!2T", kFixupNone), 199 ENCODING_MAP(WIDE(kA64Csel4rrrc), SF_VARIANTS(0x1a800000), 200 kFmtRegR, 4, 0, kFmtRegR, 9, 5, kFmtRegR, 20, 16, 201 kFmtBitBlt, 15, 12, IS_QUAD_OP | REG_DEF0_USE12 | USES_CCODES, 202 "csel", "!0r, !1r, !2r, !3c", kFixupNone), 203 ENCODING_MAP(WIDE(kA64Csinc4rrrc), SF_VARIANTS(0x1a800400), 204 kFmtRegR, 4, 0, kFmtRegR, 9, 5, kFmtRegR, 20, 16, 205 kFmtBitBlt, 15, 12, IS_QUAD_OP | REG_DEF0_USE12 | USES_CCODES, 206 "csinc", "!0r, !1r, !2r, !3c", kFixupNone), 207 ENCODING_MAP(WIDE(kA64Csneg4rrrc), SF_VARIANTS(0x5a800400), 208 kFmtRegR, 4, 0, kFmtRegR, 9, 5, kFmtRegR, 20, 16, 209 kFmtBitBlt, 15, 12, IS_QUAD_OP | REG_DEF0_USE12 | USES_CCODES, 210 "csneg", "!0r, !1r, !2r, !3c", kFixupNone), 211 ENCODING_MAP(kA64Dmb1B, NO_VARIANTS(0xd50330bf), 212 kFmtBitBlt, 11, 8, kFmtUnused, -1, -1, kFmtUnused, -1, -1, 213 kFmtUnused, -1, -1, IS_UNARY_OP, 214 "dmb", "#!0B", kFixupNone), 215 ENCODING_MAP(WIDE(kA64Eor3Rrl), SF_VARIANTS(0x52000000), 216 kFmtRegROrSp, 4, 0, kFmtRegR, 9, 5, kFmtBitBlt, 22, 10, 217 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE1, 218 "eor", "!0R, !1r, #!2l", kFixupNone), 219 ENCODING_MAP(WIDE(kA64Eor4rrro), SF_VARIANTS(0x4a000000), 220 kFmtRegR, 4, 0, kFmtRegR, 9, 5, kFmtRegR, 20, 16, 221 kFmtShift, -1, -1, IS_QUAD_OP | REG_DEF0_USE12, 222 "eor", "!0r, !1r, !2r!3o", kFixupNone), 223 ENCODING_MAP(WIDE(kA64Extr4rrrd), SF_N_VARIANTS(0x13800000), 224 kFmtRegR, 4, 0, kFmtRegR, 9, 5, kFmtRegR, 20, 16, 225 kFmtBitBlt, 15, 10, IS_QUAD_OP | REG_DEF0_USE12, 226 "extr", "!0r, !1r, !2r, #!3d", kFixupNone), 227 ENCODING_MAP(FWIDE(kA64Fabs2ff), FLOAT_VARIANTS(0x1e20c000), 228 kFmtRegF, 4, 0, kFmtRegF, 9, 5, kFmtUnused, -1, -1, 229 kFmtUnused, -1, -1, IS_BINARY_OP| REG_DEF0_USE1, 230 "fabs", "!0f, !1f", kFixupNone), 231 ENCODING_MAP(FWIDE(kA64Fadd3fff), FLOAT_VARIANTS(0x1e202800), 232 kFmtRegF, 4, 0, kFmtRegF, 9, 5, kFmtRegF, 20, 16, 233 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE12, 234 "fadd", "!0f, !1f, !2f", kFixupNone), 235 ENCODING_MAP(FWIDE(kA64Fcmp1f), FLOAT_VARIANTS(0x1e202008), 236 kFmtRegF, 9, 5, kFmtUnused, -1, -1, kFmtUnused, -1, -1, 237 kFmtUnused, -1, -1, IS_UNARY_OP | REG_USE0 | SETS_CCODES, 238 "fcmp", "!0f, #0", kFixupNone), 239 ENCODING_MAP(FWIDE(kA64Fcmp2ff), FLOAT_VARIANTS(0x1e202000), 240 kFmtRegF, 9, 5, kFmtRegF, 20, 16, kFmtUnused, -1, -1, 241 kFmtUnused, -1, -1, IS_BINARY_OP | REG_USE01 | SETS_CCODES, 242 "fcmp", "!0f, !1f", kFixupNone), 243 ENCODING_MAP(FWIDE(kA64Fcvtzs2wf), FLOAT_VARIANTS(0x1e380000), 244 kFmtRegW, 4, 0, kFmtRegF, 9, 5, kFmtUnused, -1, -1, 245 kFmtUnused, -1, -1, IS_BINARY_OP | REG_DEF0_USE1, 246 "fcvtzs", "!0w, !1f", kFixupNone), 247 ENCODING_MAP(FWIDE(kA64Fcvtzs2xf), FLOAT_VARIANTS(0x9e380000), 248 kFmtRegX, 4, 0, kFmtRegF, 9, 5, kFmtUnused, -1, -1, 249 kFmtUnused, -1, -1, IS_BINARY_OP | REG_DEF0_USE1, 250 "fcvtzs", "!0x, !1f", kFixupNone), 251 ENCODING_MAP(kA64Fcvt2Ss, NO_VARIANTS(0x1e22C000), 252 kFmtRegD, 4, 0, kFmtRegS, 9, 5, kFmtUnused, -1, -1, 253 kFmtUnused, -1, -1, IS_BINARY_OP | REG_DEF0_USE1, 254 "fcvt", "!0S, !1s", kFixupNone), 255 ENCODING_MAP(kA64Fcvt2sS, NO_VARIANTS(0x1e624000), 256 kFmtRegS, 4, 0, kFmtRegD, 9, 5, kFmtUnused, -1, -1, 257 kFmtUnused, -1, -1, IS_BINARY_OP | REG_DEF0_USE1, 258 "fcvt", "!0s, !1S", kFixupNone), 259 ENCODING_MAP(FWIDE(kA64Fdiv3fff), FLOAT_VARIANTS(0x1e201800), 260 kFmtRegF, 4, 0, kFmtRegF, 9, 5, kFmtRegF, 20, 16, 261 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE12, 262 "fdiv", "!0f, !1f, !2f", kFixupNone), 263 ENCODING_MAP(FWIDE(kA64Fmov2ff), FLOAT_VARIANTS(0x1e204000), 264 kFmtRegF, 4, 0, kFmtRegF, 9, 5, kFmtUnused, -1, -1, 265 kFmtUnused, -1, -1, IS_BINARY_OP | REG_DEF0_USE1, 266 "fmov", "!0f, !1f", kFixupNone), 267 ENCODING_MAP(FWIDE(kA64Fmov2fI), FLOAT_VARIANTS(0x1e201000), 268 kFmtRegF, 4, 0, kFmtBitBlt, 20, 13, kFmtUnused, -1, -1, 269 kFmtUnused, -1, -1, IS_BINARY_OP | REG_DEF0, 270 "fmov", "!0f, #!1I", kFixupNone), 271 ENCODING_MAP(kA64Fmov2sw, NO_VARIANTS(0x1e270000), 272 kFmtRegS, 4, 0, kFmtRegW, 9, 5, kFmtUnused, -1, -1, 273 kFmtUnused, -1, -1, IS_BINARY_OP | REG_DEF0_USE1, 274 "fmov", "!0s, !1w", kFixupNone), 275 ENCODING_MAP(kA64Fmov2Sx, NO_VARIANTS(0x9e670000), 276 kFmtRegD, 4, 0, kFmtRegX, 9, 5, kFmtUnused, -1, -1, 277 kFmtUnused, -1, -1, IS_BINARY_OP | REG_DEF0_USE1, 278 "fmov", "!0S, !1x", kFixupNone), 279 ENCODING_MAP(kA64Fmov2ws, NO_VARIANTS(0x1e260000), 280 kFmtRegW, 4, 0, kFmtRegS, 9, 5, kFmtUnused, -1, -1, 281 kFmtUnused, -1, -1, IS_BINARY_OP | REG_DEF0_USE1, 282 "fmov", "!0w, !1s", kFixupNone), 283 ENCODING_MAP(kA64Fmov2xS, NO_VARIANTS(0x9e6e0000), 284 kFmtRegX, 4, 0, kFmtRegD, 9, 5, kFmtUnused, -1, -1, 285 kFmtUnused, -1, -1, IS_BINARY_OP | REG_DEF0_USE1, 286 "fmov", "!0x, !1S", kFixupNone), 287 ENCODING_MAP(FWIDE(kA64Fmul3fff), FLOAT_VARIANTS(0x1e200800), 288 kFmtRegF, 4, 0, kFmtRegF, 9, 5, kFmtRegF, 20, 16, 289 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE12, 290 "fmul", "!0f, !1f, !2f", kFixupNone), 291 ENCODING_MAP(FWIDE(kA64Fneg2ff), FLOAT_VARIANTS(0x1e214000), 292 kFmtRegF, 4, 0, kFmtRegF, 9, 5, kFmtUnused, -1, -1, 293 kFmtUnused, -1, -1, IS_BINARY_OP | REG_DEF0_USE1, 294 "fneg", "!0f, !1f", kFixupNone), 295 ENCODING_MAP(FWIDE(kA64Frintz2ff), FLOAT_VARIANTS(0x1e25c000), 296 kFmtRegF, 4, 0, kFmtRegF, 9, 5, kFmtUnused, -1, -1, 297 kFmtUnused, -1, -1, IS_BINARY_OP | REG_DEF0_USE1, 298 "frintz", "!0f, !1f", kFixupNone), 299 ENCODING_MAP(FWIDE(kA64Fsqrt2ff), FLOAT_VARIANTS(0x1e61c000), 300 kFmtRegF, 4, 0, kFmtRegF, 9, 5, kFmtUnused, -1, -1, 301 kFmtUnused, -1, -1, IS_BINARY_OP | REG_DEF0_USE1, 302 "fsqrt", "!0f, !1f", kFixupNone), 303 ENCODING_MAP(FWIDE(kA64Fsub3fff), FLOAT_VARIANTS(0x1e203800), 304 kFmtRegF, 4, 0, kFmtRegF, 9, 5, kFmtRegF, 20, 16, 305 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE12, 306 "fsub", "!0f, !1f, !2f", kFixupNone), 307 ENCODING_MAP(kA64Ldrb3wXd, NO_VARIANTS(0x39400000), 308 kFmtRegW, 4, 0, kFmtRegXOrSp, 9, 5, kFmtBitBlt, 21, 10, 309 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE1 | IS_LOAD, 310 "ldrb", "!0w, [!1X, #!2d]", kFixupNone), 311 ENCODING_MAP(kA64Ldrb3wXx, NO_VARIANTS(0x38606800), 312 kFmtRegW, 4, 0, kFmtRegXOrSp, 9, 5, kFmtRegX, 20, 16, 313 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE12 | IS_LOAD, 314 "ldrb", "!0w, [!1X, !2x]", kFixupNone), 315 ENCODING_MAP(WIDE(kA64Ldrsb3rXd), CUSTOM_VARIANTS(0x39c00000, 0x39800000), 316 kFmtRegR, 4, 0, kFmtRegXOrSp, 9, 5, kFmtBitBlt, 21, 10, 317 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE1 | IS_LOAD, 318 "ldrsb", "!0r, [!1X, #!2d]", kFixupNone), 319 ENCODING_MAP(WIDE(kA64Ldrsb3rXx), CUSTOM_VARIANTS(0x38e06800, 0x38a06800), 320 kFmtRegR, 4, 0, kFmtRegXOrSp, 9, 5, kFmtRegX, 20, 16, 321 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE12 | IS_LOAD, 322 "ldrsb", "!0r, [!1X, !2x]", kFixupNone), 323 ENCODING_MAP(kA64Ldrh3wXF, NO_VARIANTS(0x79400000), 324 kFmtRegW, 4, 0, kFmtRegXOrSp, 9, 5, kFmtBitBlt, 21, 10, 325 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE1 | IS_LOAD, 326 "ldrh", "!0w, [!1X, #!2F]", kFixupNone), 327 ENCODING_MAP(kA64Ldrh4wXxd, NO_VARIANTS(0x78606800), 328 kFmtRegW, 4, 0, kFmtRegXOrSp, 9, 5, kFmtRegX, 20, 16, 329 kFmtBitBlt, 12, 12, IS_QUAD_OP | REG_DEF0_USE12 | IS_LOAD, 330 "ldrh", "!0w, [!1X, !2x, lsl #!3d]", kFixupNone), 331 ENCODING_MAP(WIDE(kA64Ldrsh3rXF), CUSTOM_VARIANTS(0x79c00000, 0x79800000), 332 kFmtRegR, 4, 0, kFmtRegXOrSp, 9, 5, kFmtBitBlt, 21, 10, 333 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE1 | IS_LOAD, 334 "ldrsh", "!0r, [!1X, #!2F]", kFixupNone), 335 ENCODING_MAP(WIDE(kA64Ldrsh4rXxd), CUSTOM_VARIANTS(0x78e06800, 0x78906800), 336 kFmtRegR, 4, 0, kFmtRegXOrSp, 9, 5, kFmtRegX, 20, 16, 337 kFmtBitBlt, 12, 12, IS_QUAD_OP | REG_DEF0_USE12 | IS_LOAD, 338 "ldrsh", "!0r, [!1X, !2x, lsl #!3d]", kFixupNone), 339 ENCODING_MAP(FWIDE(kA64Ldr2fp), SIZE_VARIANTS(0x1c000000), 340 kFmtRegF, 4, 0, kFmtBitBlt, 23, 5, kFmtUnused, -1, -1, 341 kFmtUnused, -1, -1, 342 IS_BINARY_OP | REG_DEF0 | REG_USE_PC | IS_LOAD | NEEDS_FIXUP, 343 "ldr", "!0f, !1p", kFixupLoad), 344 ENCODING_MAP(WIDE(kA64Ldr2rp), SIZE_VARIANTS(0x18000000), 345 kFmtRegR, 4, 0, kFmtBitBlt, 23, 5, kFmtUnused, -1, -1, 346 kFmtUnused, -1, -1, 347 IS_BINARY_OP | REG_DEF0 | REG_USE_PC | IS_LOAD | NEEDS_FIXUP, 348 "ldr", "!0r, !1p", kFixupLoad), 349 ENCODING_MAP(FWIDE(kA64Ldr3fXD), SIZE_VARIANTS(0xbd400000), 350 kFmtRegF, 4, 0, kFmtRegXOrSp, 9, 5, kFmtBitBlt, 21, 10, 351 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE1 | IS_LOAD, 352 "ldr", "!0f, [!1X, #!2D]", kFixupNone), 353 ENCODING_MAP(WIDE(kA64Ldr3rXD), SIZE_VARIANTS(0xb9400000), 354 kFmtRegR, 4, 0, kFmtRegXOrSp, 9, 5, kFmtBitBlt, 21, 10, 355 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE1 | IS_LOAD, 356 "ldr", "!0r, [!1X, #!2D]", kFixupNone), 357 ENCODING_MAP(FWIDE(kA64Ldr4fXxG), SIZE_VARIANTS(0xbc606800), 358 kFmtRegF, 4, 0, kFmtRegXOrSp, 9, 5, kFmtRegX, 20, 16, 359 kFmtBitBlt, 12, 12, IS_QUAD_OP | REG_DEF0_USE12 | IS_LOAD, 360 "ldr", "!0f, [!1X, !2x!3G]", kFixupNone), 361 ENCODING_MAP(WIDE(kA64Ldr4rXxG), SIZE_VARIANTS(0xb8606800), 362 kFmtRegR, 4, 0, kFmtRegXOrSp, 9, 5, kFmtRegX, 20, 16, 363 kFmtBitBlt, 12, 12, IS_QUAD_OP | REG_DEF0_USE12 | IS_LOAD, 364 "ldr", "!0r, [!1X, !2x!3G]", kFixupNone), 365 ENCODING_MAP(WIDE(kA64LdrPost3rXd), SIZE_VARIANTS(0xb8400400), 366 kFmtRegR, 4, 0, kFmtRegXOrSp, 9, 5, kFmtBitBlt, 20, 12, 367 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF01 | REG_USE1 | IS_LOAD, 368 "ldr", "!0r, [!1X], #!2d", kFixupNone), 369 ENCODING_MAP(WIDE(kA64Ldp4ffXD), CUSTOM_VARIANTS(0x2d400000, 0x6d400000), 370 kFmtRegF, 4, 0, kFmtRegF, 14, 10, kFmtRegXOrSp, 9, 5, 371 kFmtBitBlt, 21, 15, IS_QUAD_OP | REG_USE2 | REG_DEF01 | IS_LOAD, 372 "ldp", "!0f, !1f, [!2X, #!3D]", kFixupNone), 373 ENCODING_MAP(WIDE(kA64Ldp4rrXD), SF_VARIANTS(0x29400000), 374 kFmtRegR, 4, 0, kFmtRegR, 14, 10, kFmtRegXOrSp, 9, 5, 375 kFmtBitBlt, 21, 15, IS_QUAD_OP | REG_USE2 | REG_DEF01 | IS_LOAD, 376 "ldp", "!0r, !1r, [!2X, #!3D]", kFixupNone), 377 ENCODING_MAP(WIDE(kA64LdpPost4rrXD), CUSTOM_VARIANTS(0x28c00000, 0xa8c00000), 378 kFmtRegR, 4, 0, kFmtRegR, 14, 10, kFmtRegXOrSp, 9, 5, 379 kFmtBitBlt, 21, 15, IS_QUAD_OP | REG_USE2 | REG_DEF012 | IS_LOAD, 380 "ldp", "!0r, !1r, [!2X], #!3D", kFixupNone), 381 ENCODING_MAP(FWIDE(kA64Ldur3fXd), CUSTOM_VARIANTS(0xbc400000, 0xfc400000), 382 kFmtRegF, 4, 0, kFmtRegXOrSp, 9, 5, kFmtBitBlt, 20, 12, 383 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE1 | IS_LOAD, 384 "ldur", "!0f, [!1X, #!2d]", kFixupNone), 385 ENCODING_MAP(WIDE(kA64Ldur3rXd), SIZE_VARIANTS(0xb8400000), 386 kFmtRegR, 4, 0, kFmtRegXOrSp, 9, 5, kFmtBitBlt, 20, 12, 387 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE1 | IS_LOAD, 388 "ldur", "!0r, [!1X, #!2d]", kFixupNone), 389 ENCODING_MAP(WIDE(kA64Ldxr2rX), SIZE_VARIANTS(0x885f7c00), 390 kFmtRegR, 4, 0, kFmtRegXOrSp, 9, 5, kFmtUnused, -1, -1, 391 kFmtUnused, -1, -1, IS_BINARY_OP | REG_DEF0_USE1 | IS_LOAD, 392 "ldxr", "!0r, [!1X]", kFixupNone), 393 ENCODING_MAP(WIDE(kA64Ldaxr2rX), SIZE_VARIANTS(0x885ffc00), 394 kFmtRegR, 4, 0, kFmtRegXOrSp, 9, 5, kFmtUnused, -1, -1, 395 kFmtUnused, -1, -1, IS_BINARY_OP | REG_DEF0_USE1 | IS_LOAD, 396 "ldaxr", "!0r, [!1X]", kFixupNone), 397 ENCODING_MAP(WIDE(kA64Lsl3rrr), SF_VARIANTS(0x1ac02000), 398 kFmtRegR, 4, 0, kFmtRegR, 9, 5, kFmtRegR, 20, 16, 399 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE12, 400 "lsl", "!0r, !1r, !2r", kFixupNone), 401 ENCODING_MAP(WIDE(kA64Lsr3rrd), CUSTOM_VARIANTS(0x53007c00, 0xd340fc00), 402 kFmtRegR, 4, 0, kFmtRegR, 9, 5, kFmtBitBlt, 21, 16, 403 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE1, 404 "lsr", "!0r, !1r, #!2d", kFixupNone), 405 ENCODING_MAP(WIDE(kA64Lsr3rrr), SF_VARIANTS(0x1ac02400), 406 kFmtRegR, 4, 0, kFmtRegR, 9, 5, kFmtRegR, 20, 16, 407 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE12, 408 "lsr", "!0r, !1r, !2r", kFixupNone), 409 ENCODING_MAP(WIDE(kA64Movk3rdM), SF_VARIANTS(0x72800000), 410 kFmtRegR, 4, 0, kFmtBitBlt, 20, 5, kFmtBitBlt, 22, 21, 411 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE0, 412 "movk", "!0r, #!1d!2M", kFixupNone), 413 ENCODING_MAP(WIDE(kA64Movn3rdM), SF_VARIANTS(0x12800000), 414 kFmtRegR, 4, 0, kFmtBitBlt, 20, 5, kFmtBitBlt, 22, 21, 415 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0, 416 "movn", "!0r, #!1d!2M", kFixupNone), 417 ENCODING_MAP(WIDE(kA64Movz3rdM), SF_VARIANTS(0x52800000), 418 kFmtRegR, 4, 0, kFmtBitBlt, 20, 5, kFmtBitBlt, 22, 21, 419 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0, 420 "movz", "!0r, #!1d!2M", kFixupNone), 421 ENCODING_MAP(WIDE(kA64Mov2rr), SF_VARIANTS(0x2a0003e0), 422 kFmtRegR, 4, 0, kFmtRegR, 20, 16, kFmtUnused, -1, -1, 423 kFmtUnused, -1, -1, IS_BINARY_OP | REG_DEF0_USE1, 424 "mov", "!0r, !1r", kFixupNone), 425 ENCODING_MAP(WIDE(kA64Mvn2rr), SF_VARIANTS(0x2a2003e0), 426 kFmtRegR, 4, 0, kFmtRegR, 20, 16, kFmtUnused, -1, -1, 427 kFmtUnused, -1, -1, IS_BINARY_OP | REG_DEF0_USE1, 428 "mvn", "!0r, !1r", kFixupNone), 429 ENCODING_MAP(WIDE(kA64Mul3rrr), SF_VARIANTS(0x1b007c00), 430 kFmtRegR, 4, 0, kFmtRegR, 9, 5, kFmtRegR, 20, 16, 431 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE12, 432 "mul", "!0r, !1r, !2r", kFixupNone), 433 ENCODING_MAP(WIDE(kA64Msub4rrrr), SF_VARIANTS(0x1b008000), 434 kFmtRegR, 4, 0, kFmtRegR, 9, 5, kFmtRegR, 14, 10, 435 kFmtRegR, 20, 16, IS_QUAD_OP | REG_DEF0_USE123, 436 "msub", "!0r, !1r, !3r, !2r", kFixupNone), 437 ENCODING_MAP(WIDE(kA64Neg3rro), SF_VARIANTS(0x4b0003e0), 438 kFmtRegR, 4, 0, kFmtRegR, 20, 16, kFmtShift, -1, -1, 439 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE1, 440 "neg", "!0r, !1r!2o", kFixupNone), 441 ENCODING_MAP(WIDE(kA64Orr3Rrl), SF_VARIANTS(0x32000000), 442 kFmtRegROrSp, 4, 0, kFmtRegR, 9, 5, kFmtBitBlt, 22, 10, 443 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE1, 444 "orr", "!0R, !1r, #!2l", kFixupNone), 445 ENCODING_MAP(WIDE(kA64Orr4rrro), SF_VARIANTS(0x2a000000), 446 kFmtRegR, 4, 0, kFmtRegR, 9, 5, kFmtRegR, 20, 16, 447 kFmtShift, -1, -1, IS_QUAD_OP | REG_DEF0_USE12, 448 "orr", "!0r, !1r, !2r!3o", kFixupNone), 449 ENCODING_MAP(kA64Ret, NO_VARIANTS(0xd65f03c0), 450 kFmtUnused, -1, -1, kFmtUnused, -1, -1, kFmtUnused, -1, -1, 451 kFmtUnused, -1, -1, NO_OPERAND | IS_BRANCH, 452 "ret", "", kFixupNone), 453 ENCODING_MAP(WIDE(kA64Rev2rr), CUSTOM_VARIANTS(0x5ac00800, 0xdac00c00), 454 kFmtRegR, 4, 0, kFmtRegR, 9, 5, kFmtUnused, -1, -1, 455 kFmtUnused, -1, -1, IS_BINARY_OP | REG_DEF0_USE1, 456 "rev", "!0r, !1r", kFixupNone), 457 ENCODING_MAP(WIDE(kA64Rev162rr), SF_VARIANTS(0x5ac00400), 458 kFmtRegR, 4, 0, kFmtRegR, 9, 5, kFmtUnused, -1, -1, 459 kFmtUnused, -1, -1, IS_BINARY_OP | REG_DEF0_USE1, 460 "rev16", "!0r, !1r", kFixupNone), 461 ENCODING_MAP(WIDE(kA64Ror3rrr), SF_VARIANTS(0x1ac02c00), 462 kFmtRegR, 4, 0, kFmtRegR, 9, 5, kFmtRegR, 20, 16, 463 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE12, 464 "ror", "!0r, !1r, !2r", kFixupNone), 465 ENCODING_MAP(WIDE(kA64Sbc3rrr), SF_VARIANTS(0x5a000000), 466 kFmtRegR, 4, 0, kFmtRegR, 9, 5, kFmtRegR, 20, 16, 467 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE12, 468 "sbc", "!0r, !1r, !2r", kFixupNone), 469 ENCODING_MAP(WIDE(kA64Sbfm4rrdd), SF_N_VARIANTS(0x13000000), 470 kFmtRegR, 4, 0, kFmtRegR, 9, 5, kFmtBitBlt, 21, 16, 471 kFmtBitBlt, 15, 10, IS_QUAD_OP | REG_DEF0_USE1, 472 "sbfm", "!0r, !1r, #!2d, #!3d", kFixupNone), 473 ENCODING_MAP(FWIDE(kA64Scvtf2fw), FLOAT_VARIANTS(0x1e220000), 474 kFmtRegF, 4, 0, kFmtRegW, 9, 5, kFmtUnused, -1, -1, 475 kFmtUnused, -1, -1, IS_BINARY_OP | REG_DEF0_USE1, 476 "scvtf", "!0f, !1w", kFixupNone), 477 ENCODING_MAP(FWIDE(kA64Scvtf2fx), FLOAT_VARIANTS(0x9e220000), 478 kFmtRegF, 4, 0, kFmtRegX, 9, 5, kFmtUnused, -1, -1, 479 kFmtUnused, -1, -1, IS_BINARY_OP | REG_DEF0_USE1, 480 "scvtf", "!0f, !1x", kFixupNone), 481 ENCODING_MAP(WIDE(kA64Sdiv3rrr), SF_VARIANTS(0x1ac00c00), 482 kFmtRegR, 4, 0, kFmtRegR, 9, 5, kFmtRegR, 20, 16, 483 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE12, 484 "sdiv", "!0r, !1r, !2r", kFixupNone), 485 ENCODING_MAP(WIDE(kA64Smaddl4xwwx), NO_VARIANTS(0x9b200000), 486 kFmtRegX, 4, 0, kFmtRegW, 9, 5, kFmtRegW, 20, 16, 487 kFmtRegX, -1, -1, IS_QUAD_OP | REG_DEF0_USE123, 488 "smaddl", "!0x, !1w, !2w, !3x", kFixupNone), 489 ENCODING_MAP(WIDE(kA64Stp4ffXD), CUSTOM_VARIANTS(0x2d000000, 0x6d000000), 490 kFmtRegF, 4, 0, kFmtRegF, 14, 10, kFmtRegXOrSp, 9, 5, 491 kFmtBitBlt, 21, 15, IS_QUAD_OP | REG_USE012 | IS_STORE, 492 "stp", "!0f, !1f, [!2X, #!3D]", kFixupNone), 493 ENCODING_MAP(WIDE(kA64Stp4rrXD), SF_VARIANTS(0x29000000), 494 kFmtRegR, 4, 0, kFmtRegR, 14, 10, kFmtRegXOrSp, 9, 5, 495 kFmtBitBlt, 21, 15, IS_QUAD_OP | REG_USE012 | IS_STORE, 496 "stp", "!0r, !1r, [!2X, #!3D]", kFixupNone), 497 ENCODING_MAP(WIDE(kA64StpPost4rrXD), CUSTOM_VARIANTS(0x28800000, 0xa8800000), 498 kFmtRegR, 4, 0, kFmtRegR, 14, 10, kFmtRegXOrSp, 9, 5, 499 kFmtBitBlt, 21, 15, IS_QUAD_OP | REG_DEF2 | REG_USE012 | IS_STORE, 500 "stp", "!0r, !1r, [!2X], #!3D", kFixupNone), 501 ENCODING_MAP(WIDE(kA64StpPre4rrXD), CUSTOM_VARIANTS(0x29800000, 0xa9800000), 502 kFmtRegR, 4, 0, kFmtRegR, 14, 10, kFmtRegXOrSp, 9, 5, 503 kFmtBitBlt, 21, 15, IS_QUAD_OP | REG_DEF2 | REG_USE012 | IS_STORE, 504 "stp", "!0r, !1r, [!2X, #!3D]!!", kFixupNone), 505 ENCODING_MAP(FWIDE(kA64Str3fXD), CUSTOM_VARIANTS(0xbd000000, 0xfd000000), 506 kFmtRegF, 4, 0, kFmtRegXOrSp, 9, 5, kFmtBitBlt, 21, 10, 507 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_USE01 | IS_STORE, 508 "str", "!0f, [!1X, #!2D]", kFixupNone), 509 ENCODING_MAP(FWIDE(kA64Str4fXxG), CUSTOM_VARIANTS(0xbc206800, 0xfc206800), 510 kFmtRegF, 4, 0, kFmtRegXOrSp, 9, 5, kFmtRegX, 20, 16, 511 kFmtBitBlt, 12, 12, IS_QUAD_OP | REG_USE012 | IS_STORE, 512 "str", "!0f, [!1X, !2x!3G]", kFixupNone), 513 ENCODING_MAP(WIDE(kA64Str3rXD), SIZE_VARIANTS(0xb9000000), 514 kFmtRegR, 4, 0, kFmtRegXOrSp, 9, 5, kFmtBitBlt, 21, 10, 515 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_USE01 | IS_STORE, 516 "str", "!0r, [!1X, #!2D]", kFixupNone), 517 ENCODING_MAP(WIDE(kA64Str4rXxG), SIZE_VARIANTS(0xb8206800), 518 kFmtRegR, 4, 0, kFmtRegXOrSp, 9, 5, kFmtRegX, 20, 16, 519 kFmtBitBlt, 12, 12, IS_QUAD_OP | REG_USE012 | IS_STORE, 520 "str", "!0r, [!1X, !2x!3G]", kFixupNone), 521 ENCODING_MAP(kA64Strb3wXd, NO_VARIANTS(0x39000000), 522 kFmtRegW, 4, 0, kFmtRegXOrSp, 9, 5, kFmtBitBlt, 21, 10, 523 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_USE01 | IS_STORE, 524 "strb", "!0w, [!1X, #!2d]", kFixupNone), 525 ENCODING_MAP(kA64Strb3wXx, NO_VARIANTS(0x38206800), 526 kFmtRegW, 4, 0, kFmtRegXOrSp, 9, 5, kFmtRegX, 20, 16, 527 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_USE012 | IS_STORE, 528 "strb", "!0w, [!1X, !2x]", kFixupNone), 529 ENCODING_MAP(kA64Strh3wXF, NO_VARIANTS(0x79000000), 530 kFmtRegW, 4, 0, kFmtRegXOrSp, 9, 5, kFmtBitBlt, 21, 10, 531 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_USE01 | IS_STORE, 532 "strh", "!0w, [!1X, #!2F]", kFixupNone), 533 ENCODING_MAP(kA64Strh4wXxd, NO_VARIANTS(0x78206800), 534 kFmtRegW, 4, 0, kFmtRegXOrSp, 9, 5, kFmtRegX, 20, 16, 535 kFmtBitBlt, 12, 12, IS_QUAD_OP | REG_USE012 | IS_STORE, 536 "strh", "!0w, [!1X, !2x, lsl #!3d]", kFixupNone), 537 ENCODING_MAP(WIDE(kA64StrPost3rXd), SIZE_VARIANTS(0xb8000400), 538 kFmtRegR, 4, 0, kFmtRegXOrSp, 9, 5, kFmtBitBlt, 20, 12, 539 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_USE01 | REG_DEF1 | IS_STORE, 540 "str", "!0r, [!1X], #!2d", kFixupNone), 541 ENCODING_MAP(FWIDE(kA64Stur3fXd), CUSTOM_VARIANTS(0xbc000000, 0xfc000000), 542 kFmtRegF, 4, 0, kFmtRegXOrSp, 9, 5, kFmtBitBlt, 20, 12, 543 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_USE01 | IS_STORE, 544 "stur", "!0f, [!1X, #!2d]", kFixupNone), 545 ENCODING_MAP(WIDE(kA64Stur3rXd), SIZE_VARIANTS(0xb8000000), 546 kFmtRegR, 4, 0, kFmtRegXOrSp, 9, 5, kFmtBitBlt, 20, 12, 547 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_USE01 | IS_STORE, 548 "stur", "!0r, [!1X, #!2d]", kFixupNone), 549 ENCODING_MAP(WIDE(kA64Stxr3wrX), SIZE_VARIANTS(0x88007c00), 550 kFmtRegW, 20, 16, kFmtRegR, 4, 0, kFmtRegXOrSp, 9, 5, 551 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE12 | IS_STORE, 552 "stxr", "!0w, !1r, [!2X]", kFixupNone), 553 ENCODING_MAP(WIDE(kA64Stlxr3wrX), SIZE_VARIANTS(0x8800fc00), 554 kFmtRegW, 20, 16, kFmtRegR, 4, 0, kFmtRegXOrSp, 9, 5, 555 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE12 | IS_STORE, 556 "stlxr", "!0w, !1r, [!2X]", kFixupNone), 557 ENCODING_MAP(WIDE(kA64Sub4RRdT), SF_VARIANTS(0x51000000), 558 kFmtRegROrSp, 4, 0, kFmtRegROrSp, 9, 5, kFmtBitBlt, 21, 10, 559 kFmtBitBlt, 23, 22, IS_QUAD_OP | REG_DEF0_USE1, 560 "sub", "!0R, !1R, #!2d!3T", kFixupNone), 561 ENCODING_MAP(WIDE(kA64Sub4rrro), SF_VARIANTS(0x4b000000), 562 kFmtRegR, 4, 0, kFmtRegR, 9, 5, kFmtRegR, 20, 16, 563 kFmtShift, -1, -1, IS_QUAD_OP | REG_DEF0_USE12, 564 "sub", "!0r, !1r, !2r!3o", kFixupNone), 565 ENCODING_MAP(WIDE(kA64Sub4rrre), SF_VARIANTS(0x4b200000), 566 kFmtRegROrSp, 4, 0, kFmtRegROrSp, 9, 5, kFmtRegR, 20, 16, 567 kFmtExtend, -1, -1, IS_QUAD_OP | REG_DEF0_USE12, 568 "sub", "!0r, !1r, !2r!3o", kFixupNone), 569 ENCODING_MAP(WIDE(kA64Subs3rRd), SF_VARIANTS(0x71000000), 570 kFmtRegR, 4, 0, kFmtRegROrSp, 9, 5, kFmtBitBlt, 21, 10, 571 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE1 | SETS_CCODES, 572 "subs", "!0r, !1R, #!2d", kFixupNone), 573 ENCODING_MAP(WIDE(kA64Tst3rro), SF_VARIANTS(0x6a000000), 574 kFmtRegR, 9, 5, kFmtRegR, 20, 16, kFmtShift, -1, -1, 575 kFmtUnused, -1, -1, IS_QUAD_OP | REG_USE01 | SETS_CCODES, 576 "tst", "!0r, !1r!2o", kFixupNone), 577 ENCODING_MAP(WIDE(kA64Ubfm4rrdd), SF_N_VARIANTS(0x53000000), 578 kFmtRegR, 4, 0, kFmtRegR, 9, 5, kFmtBitBlt, 21, 16, 579 kFmtBitBlt, 15, 10, IS_QUAD_OP | REG_DEF0_USE1, 580 "ubfm", "!0r, !1r, !2d, !3d", kFixupNone), 581}; 582 583// new_lir replaces orig_lir in the pcrel_fixup list. 584void Arm64Mir2Lir::ReplaceFixup(LIR* prev_lir, LIR* orig_lir, LIR* new_lir) { 585 new_lir->u.a.pcrel_next = orig_lir->u.a.pcrel_next; 586 if (UNLIKELY(prev_lir == NULL)) { 587 first_fixup_ = new_lir; 588 } else { 589 prev_lir->u.a.pcrel_next = new_lir; 590 } 591 orig_lir->flags.fixup = kFixupNone; 592} 593 594// new_lir is inserted before orig_lir in the pcrel_fixup list. 595void Arm64Mir2Lir::InsertFixupBefore(LIR* prev_lir, LIR* orig_lir, LIR* new_lir) { 596 new_lir->u.a.pcrel_next = orig_lir; 597 if (UNLIKELY(prev_lir == NULL)) { 598 first_fixup_ = new_lir; 599 } else { 600 DCHECK(prev_lir->u.a.pcrel_next == orig_lir); 601 prev_lir->u.a.pcrel_next = new_lir; 602 } 603} 604 605/* Nop, used for aligning code. Nop is an alias for hint #0. */ 606#define PADDING_NOP (UINT32_C(0xd503201f)) 607 608uint8_t* Arm64Mir2Lir::EncodeLIRs(uint8_t* write_pos, LIR* lir) { 609 for (; lir != nullptr; lir = NEXT_LIR(lir)) { 610 bool opcode_is_wide = IS_WIDE(lir->opcode); 611 ArmOpcode opcode = UNWIDE(lir->opcode); 612 613 if (UNLIKELY(IsPseudoLirOp(opcode))) { 614 continue; 615 } 616 617 if (LIKELY(!lir->flags.is_nop)) { 618 const ArmEncodingMap *encoder = &EncodingMap[opcode]; 619 620 // Select the right variant of the skeleton. 621 uint32_t bits = opcode_is_wide ? encoder->xskeleton : encoder->wskeleton; 622 DCHECK(!opcode_is_wide || IS_WIDE(encoder->opcode)); 623 624 for (int i = 0; i < 4; i++) { 625 ArmEncodingKind kind = encoder->field_loc[i].kind; 626 uint32_t operand = lir->operands[i]; 627 uint32_t value; 628 629 if (LIKELY(static_cast<unsigned>(kind) <= kFmtBitBlt)) { 630 // Note: this will handle kFmtReg* and kFmtBitBlt. 631 632 if (static_cast<unsigned>(kind) < kFmtBitBlt) { 633 bool is_zero = A64_REG_IS_ZR(operand); 634 635 if (kIsDebugBuild) { 636 // Register usage checks: First establish register usage requirements based on the 637 // format in `kind'. 638 bool want_float = false; 639 bool want_64_bit = false; 640 bool want_size_match = false; 641 bool want_zero = false; 642 switch (kind) { 643 case kFmtRegX: 644 want_64_bit = true; 645 // Intentional fall-through. 646 case kFmtRegW: 647 want_size_match = true; 648 // Intentional fall-through. 649 case kFmtRegR: 650 want_zero = true; 651 break; 652 case kFmtRegXOrSp: 653 want_64_bit = true; 654 // Intentional fall-through. 655 case kFmtRegWOrSp: 656 want_size_match = true; 657 break; 658 case kFmtRegROrSp: 659 break; 660 case kFmtRegD: 661 want_64_bit = true; 662 // Intentional fall-through. 663 case kFmtRegS: 664 want_size_match = true; 665 // Intentional fall-through. 666 case kFmtRegF: 667 want_float = true; 668 break; 669 default: 670 LOG(FATAL) << "Bad fmt for arg n. " << i << " of " << encoder->name 671 << " (" << kind << ")"; 672 break; 673 } 674 675 // Now check that the requirements are satisfied. 676 RegStorage reg(operand | RegStorage::kValid); 677 const char *expected = nullptr; 678 if (want_float) { 679 if (!reg.IsFloat()) { 680 expected = "float register"; 681 } else if (want_size_match && (reg.IsDouble() != want_64_bit)) { 682 expected = (want_64_bit) ? "double register" : "single register"; 683 } 684 } else { 685 if (reg.IsFloat()) { 686 expected = "core register"; 687 } else if (want_size_match && (reg.Is64Bit() != want_64_bit)) { 688 expected = (want_64_bit) ? "x-register" : "w-register"; 689 } else if (reg.GetRegNum() == 31 && is_zero != want_zero) { 690 expected = (want_zero) ? "zero-register" : "sp-register"; 691 } 692 } 693 694 // TODO(Arm64): if !want_size_match, then we still should compare the size of the 695 // register with the size required by the instruction width (kA64Wide). 696 697 // Fail, if `expected' contains an unsatisfied requirement. 698 if (expected != nullptr) { 699 LOG(WARNING) << "Method: " << PrettyMethod(cu_->method_idx, *cu_->dex_file) 700 << " @ 0x" << std::hex << lir->dalvik_offset; 701 LOG(FATAL) << "Bad argument n. " << i << " of " << encoder->name 702 << ". Expected " << expected << ", got 0x" << std::hex << operand; 703 } 704 } 705 706 // TODO(Arm64): this may or may not be necessary, depending on how wzr, xzr are 707 // defined. 708 if (is_zero) { 709 operand = 31; 710 } 711 } 712 713 value = (operand << encoder->field_loc[i].start) & 714 ((1 << (encoder->field_loc[i].end + 1)) - 1); 715 bits |= value; 716 } else { 717 switch (kind) { 718 case kFmtSkip: 719 break; // Nothing to do, but continue to next. 720 case kFmtUnused: 721 i = 4; // Done, break out of the enclosing loop. 722 break; 723 case kFmtShift: 724 // Intentional fallthrough. 725 case kFmtExtend: 726 DCHECK_EQ((operand & (1 << 6)) == 0, kind == kFmtShift); 727 value = (operand & 0x3f) << 10; 728 value |= ((operand & 0x1c0) >> 6) << 21; 729 bits |= value; 730 break; 731 case kFmtImm21: 732 value = (operand & 0x3) << 29; 733 value |= ((operand & 0x1ffffc) >> 2) << 5; 734 bits |= value; 735 break; 736 default: 737 LOG(FATAL) << "Bad fmt for arg. " << i << " in " << encoder->name 738 << " (" << kind << ")"; 739 } 740 } 741 } 742 743 DCHECK_EQ(encoder->size, 4); 744 write_pos[0] = (bits & 0xff); 745 write_pos[1] = ((bits >> 8) & 0xff); 746 write_pos[2] = ((bits >> 16) & 0xff); 747 write_pos[3] = ((bits >> 24) & 0xff); 748 write_pos += 4; 749 } 750 } 751 752 return write_pos; 753} 754 755// Align data offset on 8 byte boundary: it will only contain double-word items, as word immediates 756// are better set directly from the code (they will require no more than 2 instructions). 757#define ALIGNED_DATA_OFFSET(offset) (((offset) + 0x7) & ~0x7) 758 759// Assemble the LIR into binary instruction format. 760void Arm64Mir2Lir::AssembleLIR() { 761 LIR* lir; 762 LIR* prev_lir; 763 cu_->NewTimingSplit("Assemble"); 764 int assembler_retries = 0; 765 CodeOffset starting_offset = LinkFixupInsns(first_lir_insn_, last_lir_insn_, 0); 766 data_offset_ = ALIGNED_DATA_OFFSET(starting_offset); 767 int32_t offset_adjustment; 768 AssignDataOffsets(); 769 770 /* 771 * Note: generation must be 1 on first pass (to distinguish from initialized state of 0 772 * for non-visited nodes). Start at zero here, and bit will be flipped to 1 on entry to the loop. 773 */ 774 int generation = 0; 775 while (true) { 776 // TODO(Arm64): check whether passes and offset adjustments are really necessary. 777 // Currently they aren't, as - in the fixups below - LIR are never inserted. 778 // Things can be different if jump ranges above 1 MB need to be supported. 779 // If they are not, then we can get rid of the assembler retry logic. 780 781 offset_adjustment = 0; 782 AssemblerStatus res = kSuccess; // Assume success 783 generation ^= 1; 784 // Note: nodes requiring possible fixup linked in ascending order. 785 lir = first_fixup_; 786 prev_lir = NULL; 787 while (lir != NULL) { 788 /* 789 * NOTE: the lir being considered here will be encoded following the switch (so long as 790 * we're not in a retry situation). However, any new non-pc_rel instructions inserted 791 * due to retry must be explicitly encoded at the time of insertion. Note that 792 * inserted instructions don't need use/def flags, but do need size and pc-rel status 793 * properly updated. 794 */ 795 lir->offset += offset_adjustment; 796 // During pass, allows us to tell whether a node has been updated with offset_adjustment yet. 797 lir->flags.generation = generation; 798 switch (static_cast<FixupKind>(lir->flags.fixup)) { 799 case kFixupLabel: 800 case kFixupNone: 801 case kFixupVLoad: 802 break; 803 case kFixupT1Branch: { 804 LIR *target_lir = lir->target; 805 DCHECK(target_lir); 806 CodeOffset pc = lir->offset; 807 CodeOffset target = target_lir->offset + 808 ((target_lir->flags.generation == lir->flags.generation) ? 0 : offset_adjustment); 809 int32_t delta = target - pc; 810 if (!((delta & 0x3) == 0 && IS_SIGNED_IMM19(delta >> 2))) { 811 LOG(FATAL) << "Invalid jump range in kFixupT1Branch"; 812 } 813 lir->operands[0] = delta >> 2; 814 break; 815 } 816 case kFixupLoad: 817 case kFixupCBxZ: 818 case kFixupCondBranch: { 819 LIR *target_lir = lir->target; 820 DCHECK(target_lir); 821 CodeOffset pc = lir->offset; 822 CodeOffset target = target_lir->offset + 823 ((target_lir->flags.generation == lir->flags.generation) ? 0 : offset_adjustment); 824 int32_t delta = target - pc; 825 if (!((delta & 0x3) == 0 && IS_SIGNED_IMM19(delta >> 2))) { 826 LOG(FATAL) << "Invalid jump range in kFixupLoad"; 827 } 828 lir->operands[1] = delta >> 2; 829 break; 830 } 831 case kFixupAdr: { 832 LIR* target_lir = lir->target; 833 int32_t delta; 834 if (target_lir) { 835 CodeOffset target_offs = ((target_lir->flags.generation == lir->flags.generation) ? 836 0 : offset_adjustment) + target_lir->offset; 837 delta = target_offs - lir->offset; 838 } else if (lir->operands[2] >= 0) { 839 EmbeddedData* tab = reinterpret_cast<EmbeddedData*>(UnwrapPointer(lir->operands[2])); 840 delta = tab->offset + offset_adjustment - lir->offset; 841 } else { 842 // No fixup: this usage allows to retrieve the current PC. 843 delta = lir->operands[1]; 844 } 845 if (!IS_SIGNED_IMM21(delta)) { 846 LOG(FATAL) << "Jump range above 1MB in kFixupAdr"; 847 } 848 lir->operands[1] = delta; 849 break; 850 } 851 default: 852 LOG(FATAL) << "Unexpected case " << lir->flags.fixup; 853 } 854 prev_lir = lir; 855 lir = lir->u.a.pcrel_next; 856 } 857 858 if (res == kSuccess) { 859 break; 860 } else { 861 assembler_retries++; 862 if (assembler_retries > MAX_ASSEMBLER_RETRIES) { 863 CodegenDump(); 864 LOG(FATAL) << "Assembler error - too many retries"; 865 } 866 starting_offset += offset_adjustment; 867 data_offset_ = ALIGNED_DATA_OFFSET(starting_offset); 868 AssignDataOffsets(); 869 } 870 } 871 872 // Build the CodeBuffer. 873 DCHECK_LE(data_offset_, total_size_); 874 code_buffer_.reserve(total_size_); 875 code_buffer_.resize(starting_offset); 876 uint8_t* write_pos = &code_buffer_[0]; 877 write_pos = EncodeLIRs(write_pos, first_lir_insn_); 878 DCHECK_EQ(static_cast<CodeOffset>(write_pos - &code_buffer_[0]), starting_offset); 879 880 DCHECK_EQ(data_offset_, ALIGNED_DATA_OFFSET(code_buffer_.size())); 881 882 // Install literals 883 InstallLiteralPools(); 884 885 // Install switch tables 886 InstallSwitchTables(); 887 888 // Install fill array data 889 InstallFillArrayData(); 890 891 // Create the mapping table and native offset to reference map. 892 cu_->NewTimingSplit("PcMappingTable"); 893 CreateMappingTables(); 894 895 cu_->NewTimingSplit("GcMap"); 896 CreateNativeGcMap(); 897} 898 899size_t Arm64Mir2Lir::GetInsnSize(LIR* lir) { 900 ArmOpcode opcode = UNWIDE(lir->opcode); 901 DCHECK(!IsPseudoLirOp(opcode)); 902 return EncodingMap[opcode].size; 903} 904 905// Encode instruction bit pattern and assign offsets. 906uint32_t Arm64Mir2Lir::LinkFixupInsns(LIR* head_lir, LIR* tail_lir, uint32_t offset) { 907 LIR* end_lir = tail_lir->next; 908 909 LIR* last_fixup = NULL; 910 for (LIR* lir = head_lir; lir != end_lir; lir = NEXT_LIR(lir)) { 911 ArmOpcode opcode = UNWIDE(lir->opcode); 912 if (!lir->flags.is_nop) { 913 if (lir->flags.fixup != kFixupNone) { 914 if (!IsPseudoLirOp(opcode)) { 915 lir->flags.size = EncodingMap[opcode].size; 916 lir->flags.fixup = EncodingMap[opcode].fixup; 917 } else { 918 DCHECK_NE(static_cast<int>(opcode), kPseudoPseudoAlign4); 919 lir->flags.size = 0; 920 lir->flags.fixup = kFixupLabel; 921 } 922 // Link into the fixup chain. 923 lir->flags.use_def_invalid = true; 924 lir->u.a.pcrel_next = NULL; 925 if (first_fixup_ == NULL) { 926 first_fixup_ = lir; 927 } else { 928 last_fixup->u.a.pcrel_next = lir; 929 } 930 last_fixup = lir; 931 lir->offset = offset; 932 } 933 offset += lir->flags.size; 934 } 935 } 936 return offset; 937} 938 939void Arm64Mir2Lir::AssignDataOffsets() { 940 /* Set up offsets for literals */ 941 CodeOffset offset = data_offset_; 942 943 offset = AssignLiteralOffset(offset); 944 945 offset = AssignSwitchTablesOffset(offset); 946 947 total_size_ = AssignFillArrayDataOffset(offset); 948} 949 950} // namespace art 951