assemble_arm64.cc revision 63999683329612292d534e6be09dbde9480f1250
1/* 2 * Copyright (C) 2011 The Android Open Source Project 3 * 4 * Licensed under the Apache License, Version 2.0 (the "License"); 5 * you may not use this file except in compliance with the License. 6 * You may obtain a copy of the License at 7 * 8 * http://www.apache.org/licenses/LICENSE-2.0 9 * 10 * Unless required by applicable law or agreed to in writing, software 11 * distributed under the License is distributed on an "AS IS" BASIS, 12 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 13 * See the License for the specific language governing permissions and 14 * limitations under the License. 15 */ 16 17#include "arm64_lir.h" 18#include "codegen_arm64.h" 19#include "dex/quick/mir_to_lir-inl.h" 20 21namespace art { 22 23// The macros below are exclusively used in the encoding map. 24 25// Most generic way of providing two variants for one instructions. 26#define CUSTOM_VARIANTS(variant1, variant2) variant1, variant2 27 28// Used for instructions which do not have a wide variant. 29#define NO_VARIANTS(variant) \ 30 CUSTOM_VARIANTS(variant, 0) 31 32// Used for instructions which have a wide variant with the sf bit set to 1. 33#define SF_VARIANTS(sf0_skeleton) \ 34 CUSTOM_VARIANTS(sf0_skeleton, (sf0_skeleton | 0x80000000)) 35 36// Used for instructions which have a wide variant with the size bits set to either x0 or x1. 37#define SIZE_VARIANTS(sizex0_skeleton) \ 38 CUSTOM_VARIANTS(sizex0_skeleton, (sizex0_skeleton | 0x40000000)) 39 40// Used for instructions which have a wide variant with the sf and n bits set to 1. 41#define SF_N_VARIANTS(sf0_n0_skeleton) \ 42 CUSTOM_VARIANTS(sf0_n0_skeleton, (sf0_n0_skeleton | 0x80400000)) 43 44// Used for FP instructions which have a single and double precision variants, with he type bits set 45// to either 00 or 01. 46#define FLOAT_VARIANTS(type00_skeleton) \ 47 CUSTOM_VARIANTS(type00_skeleton, (type00_skeleton | 0x00400000)) 48 49/* 50 * opcode: ArmOpcode enum 51 * variants: instruction skeletons supplied via CUSTOM_VARIANTS or derived macros. 52 * a{n}k: key to applying argument {n} \ 53 * a{n}s: argument {n} start bit position | n = 0, 1, 2, 3 54 * a{n}e: argument {n} end bit position / 55 * flags: instruction attributes (used in optimization) 56 * name: mnemonic name 57 * fmt: for pretty-printing 58 * fixup: used for second-pass fixes (e.g. adresses fixups in branch instructions). 59 */ 60#define ENCODING_MAP(opcode, variants, a0k, a0s, a0e, a1k, a1s, a1e, a2k, a2s, a2e, \ 61 a3k, a3s, a3e, flags, name, fmt, fixup) \ 62 {variants, {{a0k, a0s, a0e}, {a1k, a1s, a1e}, {a2k, a2s, a2e}, \ 63 {a3k, a3s, a3e}}, opcode, flags, name, fmt, 4, fixup} 64 65/* Instruction dump string format keys: !pf, where "!" is the start 66 * of the key, "p" is which numeric operand to use and "f" is the 67 * print format. 68 * 69 * [p]ositions: 70 * 0 -> operands[0] (dest) 71 * 1 -> operands[1] (src1) 72 * 2 -> operands[2] (src2) 73 * 3 -> operands[3] (extra) 74 * 75 * [f]ormats: 76 * d -> decimal 77 * D -> decimal*4 or decimal*8 depending on the instruction width 78 * E -> decimal*4 79 * F -> decimal*2 80 * G -> ", lsl #2" or ", lsl #3" depending on the instruction width 81 * c -> branch condition (eq, ne, etc.) 82 * t -> pc-relative target 83 * p -> pc-relative address 84 * s -> single precision floating point register 85 * S -> double precision floating point register 86 * f -> single or double precision register (depending on instruction width) 87 * I -> 8-bit immediate floating point number 88 * l -> logical immediate 89 * M -> 16-bit shift expression ("" or ", lsl #16" or ", lsl #32"...) 90 * B -> dmb option string (sy, st, ish, ishst, nsh, hshst) 91 * H -> operand shift 92 * T -> register shift (either ", lsl #0" or ", lsl #12") 93 * e -> register extend (e.g. uxtb #1) 94 * o -> register shift (e.g. lsl #1) for Word registers 95 * w -> word (32-bit) register wn, or wzr 96 * W -> word (32-bit) register wn, or wsp 97 * x -> extended (64-bit) register xn, or xzr 98 * X -> extended (64-bit) register xn, or sp 99 * r -> register with same width as instruction, r31 -> wzr, xzr 100 * R -> register with same width as instruction, r31 -> wsp, sp 101 * 102 * [!] escape. To insert "!", use "!!" 103 */ 104/* NOTE: must be kept in sync with enum ArmOpcode from arm64_lir.h */ 105const ArmEncodingMap Arm64Mir2Lir::EncodingMap[kA64Last] = { 106 ENCODING_MAP(WIDE(kA64Adc3rrr), SF_VARIANTS(0x1a000000), 107 kFmtRegR, 4, 0, kFmtRegR, 9, 5, kFmtRegR, 20, 16, 108 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE12, 109 "adc", "!0r, !1r, !2r", kFixupNone), 110 ENCODING_MAP(WIDE(kA64Add4RRdT), SF_VARIANTS(0x11000000), 111 kFmtRegROrSp, 4, 0, kFmtRegROrSp, 9, 5, kFmtBitBlt, 21, 10, 112 kFmtBitBlt, 23, 22, IS_QUAD_OP | REG_DEF0_USE1, 113 "add", "!0R, !1R, #!2d!3T", kFixupNone), 114 ENCODING_MAP(WIDE(kA64Add4rrro), SF_VARIANTS(0x0b000000), 115 kFmtRegR, 4, 0, kFmtRegR, 9, 5, kFmtRegR, 20, 16, 116 kFmtShift, -1, -1, IS_QUAD_OP | REG_DEF0_USE12, 117 "add", "!0r, !1r, !2r!3o", kFixupNone), 118 ENCODING_MAP(WIDE(kA64Add4RRre), SF_VARIANTS(0x0b200000), 119 kFmtRegROrSp, 4, 0, kFmtRegROrSp, 9, 5, kFmtRegR, 20, 16, 120 kFmtExtend, -1, -1, IS_QUAD_OP | REG_DEF0_USE12, 121 "add", "!0r, !1r, !2r!3e", kFixupNone), 122 // Note: adr is binary, but declared as tertiary. The third argument is used while doing the 123 // fixups and contains information to identify the adr label. 124 ENCODING_MAP(kA64Adr2xd, NO_VARIANTS(0x10000000), 125 kFmtRegX, 4, 0, kFmtImm21, -1, -1, kFmtUnused, -1, -1, 126 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0 | NEEDS_FIXUP, 127 "adr", "!0x, #!1d", kFixupAdr), 128 ENCODING_MAP(WIDE(kA64And3Rrl), SF_VARIANTS(0x12000000), 129 kFmtRegROrSp, 4, 0, kFmtRegR, 9, 5, kFmtBitBlt, 22, 10, 130 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE1, 131 "and", "!0R, !1r, #!2l", kFixupNone), 132 ENCODING_MAP(WIDE(kA64And4rrro), SF_VARIANTS(0x0a000000), 133 kFmtRegR, 4, 0, kFmtRegR, 9, 5, kFmtRegR, 20, 16, 134 kFmtShift, -1, -1, IS_QUAD_OP | REG_DEF0_USE12, 135 "and", "!0r, !1r, !2r!3o", kFixupNone), 136 ENCODING_MAP(WIDE(kA64Asr3rrd), CUSTOM_VARIANTS(0x13007c00, 0x9340fc00), 137 kFmtRegR, 4, 0, kFmtRegR, 9, 5, kFmtBitBlt, 21, 16, 138 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE1, 139 "asr", "!0r, !1r, #!2d", kFixupNone), 140 ENCODING_MAP(WIDE(kA64Asr3rrr), SF_VARIANTS(0x1ac02800), 141 kFmtRegR, 4, 0, kFmtRegR, 9, 5, kFmtRegR, 20, 16, 142 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE12, 143 "asr", "!0r, !1r, !2r", kFixupNone), 144 ENCODING_MAP(kA64B2ct, NO_VARIANTS(0x54000000), 145 kFmtBitBlt, 3, 0, kFmtBitBlt, 23, 5, kFmtUnused, -1, -1, 146 kFmtUnused, -1, -1, IS_BINARY_OP | IS_BRANCH | USES_CCODES | 147 NEEDS_FIXUP, "b.!0c", "!1t", kFixupCondBranch), 148 ENCODING_MAP(kA64Blr1x, NO_VARIANTS(0xd63f0000), 149 kFmtRegX, 9, 5, kFmtUnused, -1, -1, kFmtUnused, -1, -1, 150 kFmtUnused, -1, -1, 151 IS_UNARY_OP | REG_USE0 | IS_BRANCH | REG_DEF_LR, 152 "blr", "!0x", kFixupNone), 153 ENCODING_MAP(kA64Br1x, NO_VARIANTS(0xd61f0000), 154 kFmtRegX, 9, 5, kFmtUnused, -1, -1, kFmtUnused, -1, -1, 155 kFmtUnused, -1, -1, IS_UNARY_OP | REG_USE0 | IS_BRANCH, 156 "br", "!0x", kFixupNone), 157 ENCODING_MAP(kA64Brk1d, NO_VARIANTS(0xd4200000), 158 kFmtBitBlt, 20, 5, kFmtUnused, -1, -1, kFmtUnused, -1, -1, 159 kFmtUnused, -1, -1, IS_UNARY_OP | IS_BRANCH, 160 "brk", "!0d", kFixupNone), 161 ENCODING_MAP(kA64B1t, NO_VARIANTS(0x14000000), 162 kFmtBitBlt, 25, 0, kFmtUnused, -1, -1, kFmtUnused, -1, -1, 163 kFmtUnused, -1, -1, IS_UNARY_OP | IS_BRANCH | NEEDS_FIXUP, 164 "b", "!0t", kFixupT1Branch), 165 ENCODING_MAP(WIDE(kA64Cbnz2rt), SF_VARIANTS(0x35000000), 166 kFmtRegR, 4, 0, kFmtBitBlt, 23, 5, kFmtUnused, -1, -1, 167 kFmtUnused, -1, -1, 168 IS_BINARY_OP | REG_USE0 | IS_BRANCH | NEEDS_FIXUP, 169 "cbnz", "!0r, !1t", kFixupCBxZ), 170 ENCODING_MAP(WIDE(kA64Cbz2rt), SF_VARIANTS(0x34000000), 171 kFmtRegR, 4, 0, kFmtBitBlt, 23, 5, kFmtUnused, -1, -1, 172 kFmtUnused, -1, -1, 173 IS_BINARY_OP | REG_USE0 | IS_BRANCH | NEEDS_FIXUP, 174 "cbz", "!0r, !1t", kFixupCBxZ), 175 ENCODING_MAP(WIDE(kA64Cmn3rro), SF_VARIANTS(0x2b00001f), 176 kFmtRegR, 9, 5, kFmtRegR, 20, 16, kFmtShift, -1, -1, 177 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_USE01 | SETS_CCODES, 178 "cmn", "!0r, !1r!2o", kFixupNone), 179 ENCODING_MAP(WIDE(kA64Cmn3Rre), SF_VARIANTS(0x2b20001f), 180 kFmtRegROrSp, 9, 5, kFmtRegR, 20, 16, kFmtExtend, -1, -1, 181 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_USE01 | SETS_CCODES, 182 "cmn", "!0R, !1r!2e", kFixupNone), 183 ENCODING_MAP(WIDE(kA64Cmn3RdT), SF_VARIANTS(0x3100001f), 184 kFmtRegROrSp, 9, 5, kFmtBitBlt, 21, 10, kFmtBitBlt, 23, 22, 185 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_USE0 | SETS_CCODES, 186 "cmn", "!0R, #!1d!2T", kFixupNone), 187 ENCODING_MAP(WIDE(kA64Cmp3rro), SF_VARIANTS(0x6b00001f), 188 kFmtRegR, 9, 5, kFmtRegR, 20, 16, kFmtShift, -1, -1, 189 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_USE01 | SETS_CCODES, 190 "cmp", "!0r, !1r!2o", kFixupNone), 191 ENCODING_MAP(WIDE(kA64Cmp3Rre), SF_VARIANTS(0x6b20001f), 192 kFmtRegROrSp, 9, 5, kFmtRegR, 20, 16, kFmtExtend, -1, -1, 193 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_USE01 | SETS_CCODES, 194 "cmp", "!0R, !1r!2e", kFixupNone), 195 ENCODING_MAP(WIDE(kA64Cmp3RdT), SF_VARIANTS(0x7100001f), 196 kFmtRegROrSp, 9, 5, kFmtBitBlt, 21, 10, kFmtBitBlt, 23, 22, 197 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_USE0 | SETS_CCODES, 198 "cmp", "!0R, #!1d!2T", kFixupNone), 199 ENCODING_MAP(WIDE(kA64Csel4rrrc), SF_VARIANTS(0x1a800000), 200 kFmtRegR, 4, 0, kFmtRegR, 9, 5, kFmtRegR, 20, 16, 201 kFmtBitBlt, 15, 12, IS_QUAD_OP | REG_DEF0_USE12 | USES_CCODES, 202 "csel", "!0r, !1r, !2r, !3c", kFixupNone), 203 ENCODING_MAP(WIDE(kA64Csinc4rrrc), SF_VARIANTS(0x1a800400), 204 kFmtRegR, 4, 0, kFmtRegR, 9, 5, kFmtRegR, 20, 16, 205 kFmtBitBlt, 15, 12, IS_QUAD_OP | REG_DEF0_USE12 | USES_CCODES, 206 "csinc", "!0r, !1r, !2r, !3c", kFixupNone), 207 ENCODING_MAP(WIDE(kA64Csinv4rrrc), SF_VARIANTS(0x5a800000), 208 kFmtRegR, 4, 0, kFmtRegR, 9, 5, kFmtRegR, 20, 16, 209 kFmtBitBlt, 15, 12, IS_QUAD_OP | REG_DEF0_USE12 | USES_CCODES, 210 "csinv", "!0r, !1r, !2r, !3c", kFixupNone), 211 ENCODING_MAP(WIDE(kA64Csneg4rrrc), SF_VARIANTS(0x5a800400), 212 kFmtRegR, 4, 0, kFmtRegR, 9, 5, kFmtRegR, 20, 16, 213 kFmtBitBlt, 15, 12, IS_QUAD_OP | REG_DEF0_USE12 | USES_CCODES, 214 "csneg", "!0r, !1r, !2r, !3c", kFixupNone), 215 ENCODING_MAP(kA64Dmb1B, NO_VARIANTS(0xd50330bf), 216 kFmtBitBlt, 11, 8, kFmtUnused, -1, -1, kFmtUnused, -1, -1, 217 kFmtUnused, -1, -1, IS_UNARY_OP | IS_VOLATILE, 218 "dmb", "#!0B", kFixupNone), 219 ENCODING_MAP(WIDE(kA64Eor3Rrl), SF_VARIANTS(0x52000000), 220 kFmtRegROrSp, 4, 0, kFmtRegR, 9, 5, kFmtBitBlt, 22, 10, 221 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE1, 222 "eor", "!0R, !1r, #!2l", kFixupNone), 223 ENCODING_MAP(WIDE(kA64Eor4rrro), SF_VARIANTS(0x4a000000), 224 kFmtRegR, 4, 0, kFmtRegR, 9, 5, kFmtRegR, 20, 16, 225 kFmtShift, -1, -1, IS_QUAD_OP | REG_DEF0_USE12, 226 "eor", "!0r, !1r, !2r!3o", kFixupNone), 227 ENCODING_MAP(WIDE(kA64Extr4rrrd), SF_N_VARIANTS(0x13800000), 228 kFmtRegR, 4, 0, kFmtRegR, 9, 5, kFmtRegR, 20, 16, 229 kFmtBitBlt, 15, 10, IS_QUAD_OP | REG_DEF0_USE12, 230 "extr", "!0r, !1r, !2r, #!3d", kFixupNone), 231 ENCODING_MAP(FWIDE(kA64Fabs2ff), FLOAT_VARIANTS(0x1e20c000), 232 kFmtRegF, 4, 0, kFmtRegF, 9, 5, kFmtUnused, -1, -1, 233 kFmtUnused, -1, -1, IS_BINARY_OP| REG_DEF0_USE1, 234 "fabs", "!0f, !1f", kFixupNone), 235 ENCODING_MAP(FWIDE(kA64Fadd3fff), FLOAT_VARIANTS(0x1e202800), 236 kFmtRegF, 4, 0, kFmtRegF, 9, 5, kFmtRegF, 20, 16, 237 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE12, 238 "fadd", "!0f, !1f, !2f", kFixupNone), 239 ENCODING_MAP(FWIDE(kA64Fcmp1f), FLOAT_VARIANTS(0x1e202008), 240 kFmtRegF, 9, 5, kFmtUnused, -1, -1, kFmtUnused, -1, -1, 241 kFmtUnused, -1, -1, IS_UNARY_OP | REG_USE0 | SETS_CCODES, 242 "fcmp", "!0f, #0", kFixupNone), 243 ENCODING_MAP(FWIDE(kA64Fcmp2ff), FLOAT_VARIANTS(0x1e202000), 244 kFmtRegF, 9, 5, kFmtRegF, 20, 16, kFmtUnused, -1, -1, 245 kFmtUnused, -1, -1, IS_BINARY_OP | REG_USE01 | SETS_CCODES, 246 "fcmp", "!0f, !1f", kFixupNone), 247 ENCODING_MAP(FWIDE(kA64Fcvtzs2wf), FLOAT_VARIANTS(0x1e380000), 248 kFmtRegW, 4, 0, kFmtRegF, 9, 5, kFmtUnused, -1, -1, 249 kFmtUnused, -1, -1, IS_BINARY_OP | REG_DEF0_USE1, 250 "fcvtzs", "!0w, !1f", kFixupNone), 251 ENCODING_MAP(FWIDE(kA64Fcvtzs2xf), FLOAT_VARIANTS(0x9e380000), 252 kFmtRegX, 4, 0, kFmtRegF, 9, 5, kFmtUnused, -1, -1, 253 kFmtUnused, -1, -1, IS_BINARY_OP | REG_DEF0_USE1, 254 "fcvtzs", "!0x, !1f", kFixupNone), 255 ENCODING_MAP(kA64Fcvt2Ss, NO_VARIANTS(0x1e22C000), 256 kFmtRegD, 4, 0, kFmtRegS, 9, 5, kFmtUnused, -1, -1, 257 kFmtUnused, -1, -1, IS_BINARY_OP | REG_DEF0_USE1, 258 "fcvt", "!0S, !1s", kFixupNone), 259 ENCODING_MAP(kA64Fcvt2sS, NO_VARIANTS(0x1e624000), 260 kFmtRegS, 4, 0, kFmtRegD, 9, 5, kFmtUnused, -1, -1, 261 kFmtUnused, -1, -1, IS_BINARY_OP | REG_DEF0_USE1, 262 "fcvt", "!0s, !1S", kFixupNone), 263 ENCODING_MAP(FWIDE(kA64Fdiv3fff), FLOAT_VARIANTS(0x1e201800), 264 kFmtRegF, 4, 0, kFmtRegF, 9, 5, kFmtRegF, 20, 16, 265 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE12, 266 "fdiv", "!0f, !1f, !2f", kFixupNone), 267 ENCODING_MAP(FWIDE(kA64Fmax3fff), FLOAT_VARIANTS(0x1e204800), 268 kFmtRegF, 4, 0, kFmtRegF, 9, 5, kFmtRegF, 20, 16, 269 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE12, 270 "fmax", "!0f, !1f, !2f", kFixupNone), 271 ENCODING_MAP(FWIDE(kA64Fmin3fff), FLOAT_VARIANTS(0x1e205800), 272 kFmtRegF, 4, 0, kFmtRegF, 9, 5, kFmtRegF, 20, 16, 273 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE12, 274 "fmin", "!0f, !1f, !2f", kFixupNone), 275 ENCODING_MAP(FWIDE(kA64Fmov2ff), FLOAT_VARIANTS(0x1e204000), 276 kFmtRegF, 4, 0, kFmtRegF, 9, 5, kFmtUnused, -1, -1, 277 kFmtUnused, -1, -1, IS_BINARY_OP | REG_DEF0_USE1 | IS_MOVE, 278 "fmov", "!0f, !1f", kFixupNone), 279 ENCODING_MAP(FWIDE(kA64Fmov2fI), FLOAT_VARIANTS(0x1e201000), 280 kFmtRegF, 4, 0, kFmtBitBlt, 20, 13, kFmtUnused, -1, -1, 281 kFmtUnused, -1, -1, IS_BINARY_OP | REG_DEF0, 282 "fmov", "!0f, #!1I", kFixupNone), 283 ENCODING_MAP(kA64Fmov2sw, NO_VARIANTS(0x1e270000), 284 kFmtRegS, 4, 0, kFmtRegW, 9, 5, kFmtUnused, -1, -1, 285 kFmtUnused, -1, -1, IS_BINARY_OP | REG_DEF0_USE1, 286 "fmov", "!0s, !1w", kFixupNone), 287 ENCODING_MAP(kA64Fmov2Sx, NO_VARIANTS(0x9e670000), 288 kFmtRegD, 4, 0, kFmtRegX, 9, 5, kFmtUnused, -1, -1, 289 kFmtUnused, -1, -1, IS_BINARY_OP | REG_DEF0_USE1, 290 "fmov", "!0S, !1x", kFixupNone), 291 ENCODING_MAP(kA64Fmov2ws, NO_VARIANTS(0x1e260000), 292 kFmtRegW, 4, 0, kFmtRegS, 9, 5, kFmtUnused, -1, -1, 293 kFmtUnused, -1, -1, IS_BINARY_OP | REG_DEF0_USE1, 294 "fmov", "!0w, !1s", kFixupNone), 295 ENCODING_MAP(kA64Fmov2xS, NO_VARIANTS(0x9e660000), 296 kFmtRegX, 4, 0, kFmtRegD, 9, 5, kFmtUnused, -1, -1, 297 kFmtUnused, -1, -1, IS_BINARY_OP | REG_DEF0_USE1, 298 "fmov", "!0x, !1S", kFixupNone), 299 ENCODING_MAP(FWIDE(kA64Fmul3fff), FLOAT_VARIANTS(0x1e200800), 300 kFmtRegF, 4, 0, kFmtRegF, 9, 5, kFmtRegF, 20, 16, 301 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE12, 302 "fmul", "!0f, !1f, !2f", kFixupNone), 303 ENCODING_MAP(FWIDE(kA64Fneg2ff), FLOAT_VARIANTS(0x1e214000), 304 kFmtRegF, 4, 0, kFmtRegF, 9, 5, kFmtUnused, -1, -1, 305 kFmtUnused, -1, -1, IS_BINARY_OP | REG_DEF0_USE1, 306 "fneg", "!0f, !1f", kFixupNone), 307 ENCODING_MAP(FWIDE(kA64Frintz2ff), FLOAT_VARIANTS(0x1e25c000), 308 kFmtRegF, 4, 0, kFmtRegF, 9, 5, kFmtUnused, -1, -1, 309 kFmtUnused, -1, -1, IS_BINARY_OP | REG_DEF0_USE1, 310 "frintz", "!0f, !1f", kFixupNone), 311 ENCODING_MAP(FWIDE(kA64Fsqrt2ff), FLOAT_VARIANTS(0x1e61c000), 312 kFmtRegF, 4, 0, kFmtRegF, 9, 5, kFmtUnused, -1, -1, 313 kFmtUnused, -1, -1, IS_BINARY_OP | REG_DEF0_USE1, 314 "fsqrt", "!0f, !1f", kFixupNone), 315 ENCODING_MAP(FWIDE(kA64Fsub3fff), FLOAT_VARIANTS(0x1e203800), 316 kFmtRegF, 4, 0, kFmtRegF, 9, 5, kFmtRegF, 20, 16, 317 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE12, 318 "fsub", "!0f, !1f, !2f", kFixupNone), 319 ENCODING_MAP(kA64Ldrb3wXd, NO_VARIANTS(0x39400000), 320 kFmtRegW, 4, 0, kFmtRegXOrSp, 9, 5, kFmtBitBlt, 21, 10, 321 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE1 | IS_LOAD_OFF, 322 "ldrb", "!0w, [!1X, #!2d]", kFixupNone), 323 ENCODING_MAP(kA64Ldrb3wXx, NO_VARIANTS(0x38606800), 324 kFmtRegW, 4, 0, kFmtRegXOrSp, 9, 5, kFmtRegX, 20, 16, 325 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE12 | IS_LOAD, 326 "ldrb", "!0w, [!1X, !2x]", kFixupNone), 327 ENCODING_MAP(WIDE(kA64Ldrsb3rXd), CUSTOM_VARIANTS(0x39c00000, 0x39800000), 328 kFmtRegR, 4, 0, kFmtRegXOrSp, 9, 5, kFmtBitBlt, 21, 10, 329 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE1 | IS_LOAD_OFF, 330 "ldrsb", "!0r, [!1X, #!2d]", kFixupNone), 331 ENCODING_MAP(WIDE(kA64Ldrsb3rXx), CUSTOM_VARIANTS(0x38e06800, 0x38a06800), 332 kFmtRegR, 4, 0, kFmtRegXOrSp, 9, 5, kFmtRegX, 20, 16, 333 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE12 | IS_LOAD, 334 "ldrsb", "!0r, [!1X, !2x]", kFixupNone), 335 ENCODING_MAP(kA64Ldrh3wXF, NO_VARIANTS(0x79400000), 336 kFmtRegW, 4, 0, kFmtRegXOrSp, 9, 5, kFmtBitBlt, 21, 10, 337 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE1 | IS_LOAD_OFF, 338 "ldrh", "!0w, [!1X, #!2F]", kFixupNone), 339 ENCODING_MAP(kA64Ldrh4wXxd, NO_VARIANTS(0x78606800), 340 kFmtRegW, 4, 0, kFmtRegXOrSp, 9, 5, kFmtRegX, 20, 16, 341 kFmtBitBlt, 12, 12, IS_QUAD_OP | REG_DEF0_USE12 | IS_LOAD_OFF, 342 "ldrh", "!0w, [!1X, !2x, lsl #!3d]", kFixupNone), 343 ENCODING_MAP(WIDE(kA64Ldrsh3rXF), CUSTOM_VARIANTS(0x79c00000, 0x79800000), 344 kFmtRegR, 4, 0, kFmtRegXOrSp, 9, 5, kFmtBitBlt, 21, 10, 345 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE1 | IS_LOAD_OFF, 346 "ldrsh", "!0r, [!1X, #!2F]", kFixupNone), 347 ENCODING_MAP(WIDE(kA64Ldrsh4rXxd), CUSTOM_VARIANTS(0x78e06800, 0x78906800), 348 kFmtRegR, 4, 0, kFmtRegXOrSp, 9, 5, kFmtRegX, 20, 16, 349 kFmtBitBlt, 12, 12, IS_QUAD_OP | REG_DEF0_USE12 | IS_LOAD_OFF, 350 "ldrsh", "!0r, [!1X, !2x, lsl #!3d]", kFixupNone), 351 ENCODING_MAP(FWIDE(kA64Ldr2fp), SIZE_VARIANTS(0x1c000000), 352 kFmtRegF, 4, 0, kFmtBitBlt, 23, 5, kFmtUnused, -1, -1, 353 kFmtUnused, -1, -1, 354 IS_BINARY_OP | REG_DEF0 | REG_USE_PC | IS_LOAD | NEEDS_FIXUP, 355 "ldr", "!0f, !1p", kFixupLoad), 356 ENCODING_MAP(WIDE(kA64Ldr2rp), SIZE_VARIANTS(0x18000000), 357 kFmtRegR, 4, 0, kFmtBitBlt, 23, 5, kFmtUnused, -1, -1, 358 kFmtUnused, -1, -1, 359 IS_BINARY_OP | REG_DEF0 | REG_USE_PC | IS_LOAD | NEEDS_FIXUP, 360 "ldr", "!0r, !1p", kFixupLoad), 361 ENCODING_MAP(FWIDE(kA64Ldr3fXD), SIZE_VARIANTS(0xbd400000), 362 kFmtRegF, 4, 0, kFmtRegXOrSp, 9, 5, kFmtBitBlt, 21, 10, 363 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE1 | IS_LOAD_OFF, 364 "ldr", "!0f, [!1X, #!2D]", kFixupNone), 365 ENCODING_MAP(WIDE(kA64Ldr3rXD), SIZE_VARIANTS(0xb9400000), 366 kFmtRegR, 4, 0, kFmtRegXOrSp, 9, 5, kFmtBitBlt, 21, 10, 367 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE1 | IS_LOAD_OFF, 368 "ldr", "!0r, [!1X, #!2D]", kFixupNone), 369 ENCODING_MAP(FWIDE(kA64Ldr4fXxG), SIZE_VARIANTS(0xbc606800), 370 kFmtRegF, 4, 0, kFmtRegXOrSp, 9, 5, kFmtRegX, 20, 16, 371 kFmtBitBlt, 12, 12, IS_QUAD_OP | REG_DEF0_USE12 | IS_LOAD, 372 "ldr", "!0f, [!1X, !2x!3G]", kFixupNone), 373 ENCODING_MAP(WIDE(kA64Ldr4rXxG), SIZE_VARIANTS(0xb8606800), 374 kFmtRegR, 4, 0, kFmtRegXOrSp, 9, 5, kFmtRegX, 20, 16, 375 kFmtBitBlt, 12, 12, IS_QUAD_OP | REG_DEF0_USE12 | IS_LOAD, 376 "ldr", "!0r, [!1X, !2x!3G]", kFixupNone), 377 ENCODING_MAP(WIDE(kA64LdrPost3rXd), SIZE_VARIANTS(0xb8400400), 378 kFmtRegR, 4, 0, kFmtRegXOrSp, 9, 5, kFmtBitBlt, 20, 12, 379 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF01 | REG_USE1 | IS_LOAD, 380 "ldr", "!0r, [!1X], #!2d", kFixupNone), 381 ENCODING_MAP(WIDE(kA64Ldp4ffXD), CUSTOM_VARIANTS(0x2d400000, 0x6d400000), 382 kFmtRegF, 4, 0, kFmtRegF, 14, 10, kFmtRegXOrSp, 9, 5, 383 kFmtBitBlt, 21, 15, IS_QUAD_OP | REG_USE2 | REG_DEF01 | IS_LOAD_OFF, 384 "ldp", "!0f, !1f, [!2X, #!3D]", kFixupNone), 385 ENCODING_MAP(WIDE(kA64Ldp4rrXD), SF_VARIANTS(0x29400000), 386 kFmtRegR, 4, 0, kFmtRegR, 14, 10, kFmtRegXOrSp, 9, 5, 387 kFmtBitBlt, 21, 15, IS_QUAD_OP | REG_USE2 | REG_DEF01 | IS_LOAD_OFF, 388 "ldp", "!0r, !1r, [!2X, #!3D]", kFixupNone), 389 ENCODING_MAP(WIDE(kA64LdpPost4rrXD), CUSTOM_VARIANTS(0x28c00000, 0xa8c00000), 390 kFmtRegR, 4, 0, kFmtRegR, 14, 10, kFmtRegXOrSp, 9, 5, 391 kFmtBitBlt, 21, 15, IS_QUAD_OP | REG_USE2 | REG_DEF012 | IS_LOAD, 392 "ldp", "!0r, !1r, [!2X], #!3D", kFixupNone), 393 ENCODING_MAP(FWIDE(kA64Ldur3fXd), CUSTOM_VARIANTS(0xbc400000, 0xfc400000), 394 kFmtRegF, 4, 0, kFmtRegXOrSp, 9, 5, kFmtBitBlt, 20, 12, 395 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE1 | IS_LOAD, 396 "ldur", "!0f, [!1X, #!2d]", kFixupNone), 397 ENCODING_MAP(WIDE(kA64Ldur3rXd), SIZE_VARIANTS(0xb8400000), 398 kFmtRegR, 4, 0, kFmtRegXOrSp, 9, 5, kFmtBitBlt, 20, 12, 399 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE1 | IS_LOAD, 400 "ldur", "!0r, [!1X, #!2d]", kFixupNone), 401 ENCODING_MAP(WIDE(kA64Ldxr2rX), SIZE_VARIANTS(0x885f7c00), 402 kFmtRegR, 4, 0, kFmtRegXOrSp, 9, 5, kFmtUnused, -1, -1, 403 kFmtUnused, -1, -1, IS_BINARY_OP | REG_DEF0_USE1 | IS_LOADX, 404 "ldxr", "!0r, [!1X]", kFixupNone), 405 ENCODING_MAP(WIDE(kA64Ldaxr2rX), SIZE_VARIANTS(0x885ffc00), 406 kFmtRegR, 4, 0, kFmtRegXOrSp, 9, 5, kFmtUnused, -1, -1, 407 kFmtUnused, -1, -1, IS_BINARY_OP | REG_DEF0_USE1 | IS_LOADX, 408 "ldaxr", "!0r, [!1X]", kFixupNone), 409 ENCODING_MAP(WIDE(kA64Lsl3rrr), SF_VARIANTS(0x1ac02000), 410 kFmtRegR, 4, 0, kFmtRegR, 9, 5, kFmtRegR, 20, 16, 411 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE12, 412 "lsl", "!0r, !1r, !2r", kFixupNone), 413 ENCODING_MAP(WIDE(kA64Lsr3rrd), CUSTOM_VARIANTS(0x53007c00, 0xd340fc00), 414 kFmtRegR, 4, 0, kFmtRegR, 9, 5, kFmtBitBlt, 21, 16, 415 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE1, 416 "lsr", "!0r, !1r, #!2d", kFixupNone), 417 ENCODING_MAP(WIDE(kA64Lsr3rrr), SF_VARIANTS(0x1ac02400), 418 kFmtRegR, 4, 0, kFmtRegR, 9, 5, kFmtRegR, 20, 16, 419 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE12, 420 "lsr", "!0r, !1r, !2r", kFixupNone), 421 ENCODING_MAP(WIDE(kA64Movk3rdM), SF_VARIANTS(0x72800000), 422 kFmtRegR, 4, 0, kFmtBitBlt, 20, 5, kFmtBitBlt, 22, 21, 423 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE0, 424 "movk", "!0r, #!1d!2M", kFixupNone), 425 ENCODING_MAP(WIDE(kA64Movn3rdM), SF_VARIANTS(0x12800000), 426 kFmtRegR, 4, 0, kFmtBitBlt, 20, 5, kFmtBitBlt, 22, 21, 427 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0, 428 "movn", "!0r, #!1d!2M", kFixupNone), 429 ENCODING_MAP(WIDE(kA64Movz3rdM), SF_VARIANTS(0x52800000), 430 kFmtRegR, 4, 0, kFmtBitBlt, 20, 5, kFmtBitBlt, 22, 21, 431 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0, 432 "movz", "!0r, #!1d!2M", kFixupNone), 433 ENCODING_MAP(WIDE(kA64Mov2rr), SF_VARIANTS(0x2a0003e0), 434 kFmtRegR, 4, 0, kFmtRegR, 20, 16, kFmtUnused, -1, -1, 435 kFmtUnused, -1, -1, IS_BINARY_OP | REG_DEF0_USE1 | IS_MOVE, 436 "mov", "!0r, !1r", kFixupNone), 437 ENCODING_MAP(WIDE(kA64Mvn2rr), SF_VARIANTS(0x2a2003e0), 438 kFmtRegR, 4, 0, kFmtRegR, 20, 16, kFmtUnused, -1, -1, 439 kFmtUnused, -1, -1, IS_BINARY_OP | REG_DEF0_USE1, 440 "mvn", "!0r, !1r", kFixupNone), 441 ENCODING_MAP(WIDE(kA64Mul3rrr), SF_VARIANTS(0x1b007c00), 442 kFmtRegR, 4, 0, kFmtRegR, 9, 5, kFmtRegR, 20, 16, 443 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE12, 444 "mul", "!0r, !1r, !2r", kFixupNone), 445 ENCODING_MAP(WIDE(kA64Msub4rrrr), SF_VARIANTS(0x1b008000), 446 kFmtRegR, 4, 0, kFmtRegR, 9, 5, kFmtRegR, 14, 10, 447 kFmtRegR, 20, 16, IS_QUAD_OP | REG_DEF0_USE123, 448 "msub", "!0r, !1r, !3r, !2r", kFixupNone), 449 ENCODING_MAP(WIDE(kA64Neg3rro), SF_VARIANTS(0x4b0003e0), 450 kFmtRegR, 4, 0, kFmtRegR, 20, 16, kFmtShift, -1, -1, 451 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE1, 452 "neg", "!0r, !1r!2o", kFixupNone), 453 ENCODING_MAP(WIDE(kA64Orr3Rrl), SF_VARIANTS(0x32000000), 454 kFmtRegROrSp, 4, 0, kFmtRegR, 9, 5, kFmtBitBlt, 22, 10, 455 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE1, 456 "orr", "!0R, !1r, #!2l", kFixupNone), 457 ENCODING_MAP(WIDE(kA64Orr4rrro), SF_VARIANTS(0x2a000000), 458 kFmtRegR, 4, 0, kFmtRegR, 9, 5, kFmtRegR, 20, 16, 459 kFmtShift, -1, -1, IS_QUAD_OP | REG_DEF0_USE12, 460 "orr", "!0r, !1r, !2r!3o", kFixupNone), 461 ENCODING_MAP(kA64Ret, NO_VARIANTS(0xd65f03c0), 462 kFmtUnused, -1, -1, kFmtUnused, -1, -1, kFmtUnused, -1, -1, 463 kFmtUnused, -1, -1, NO_OPERAND | IS_BRANCH, 464 "ret", "", kFixupNone), 465 ENCODING_MAP(WIDE(kA64Rbit2rr), SF_VARIANTS(0x5ac00000), 466 kFmtRegR, 4, 0, kFmtRegR, 9, 5, kFmtUnused, -1, -1, 467 kFmtUnused, -1, -1, IS_BINARY_OP | REG_DEF0_USE1, 468 "rbit", "!0r, !1r", kFixupNone), 469 ENCODING_MAP(WIDE(kA64Rev2rr), CUSTOM_VARIANTS(0x5ac00800, 0xdac00c00), 470 kFmtRegR, 4, 0, kFmtRegR, 9, 5, kFmtUnused, -1, -1, 471 kFmtUnused, -1, -1, IS_BINARY_OP | REG_DEF0_USE1, 472 "rev", "!0r, !1r", kFixupNone), 473 ENCODING_MAP(WIDE(kA64Rev162rr), SF_VARIANTS(0x5ac00400), 474 kFmtRegR, 4, 0, kFmtRegR, 9, 5, kFmtUnused, -1, -1, 475 kFmtUnused, -1, -1, IS_BINARY_OP | REG_DEF0_USE1, 476 "rev16", "!0r, !1r", kFixupNone), 477 ENCODING_MAP(WIDE(kA64Ror3rrr), SF_VARIANTS(0x1ac02c00), 478 kFmtRegR, 4, 0, kFmtRegR, 9, 5, kFmtRegR, 20, 16, 479 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE12, 480 "ror", "!0r, !1r, !2r", kFixupNone), 481 ENCODING_MAP(WIDE(kA64Sbc3rrr), SF_VARIANTS(0x5a000000), 482 kFmtRegR, 4, 0, kFmtRegR, 9, 5, kFmtRegR, 20, 16, 483 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE12, 484 "sbc", "!0r, !1r, !2r", kFixupNone), 485 ENCODING_MAP(WIDE(kA64Sbfm4rrdd), SF_N_VARIANTS(0x13000000), 486 kFmtRegR, 4, 0, kFmtRegR, 9, 5, kFmtBitBlt, 21, 16, 487 kFmtBitBlt, 15, 10, IS_QUAD_OP | REG_DEF0_USE1, 488 "sbfm", "!0r, !1r, #!2d, #!3d", kFixupNone), 489 ENCODING_MAP(FWIDE(kA64Scvtf2fw), FLOAT_VARIANTS(0x1e220000), 490 kFmtRegF, 4, 0, kFmtRegW, 9, 5, kFmtUnused, -1, -1, 491 kFmtUnused, -1, -1, IS_BINARY_OP | REG_DEF0_USE1, 492 "scvtf", "!0f, !1w", kFixupNone), 493 ENCODING_MAP(FWIDE(kA64Scvtf2fx), FLOAT_VARIANTS(0x9e220000), 494 kFmtRegF, 4, 0, kFmtRegX, 9, 5, kFmtUnused, -1, -1, 495 kFmtUnused, -1, -1, IS_BINARY_OP | REG_DEF0_USE1, 496 "scvtf", "!0f, !1x", kFixupNone), 497 ENCODING_MAP(WIDE(kA64Sdiv3rrr), SF_VARIANTS(0x1ac00c00), 498 kFmtRegR, 4, 0, kFmtRegR, 9, 5, kFmtRegR, 20, 16, 499 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE12, 500 "sdiv", "!0r, !1r, !2r", kFixupNone), 501 ENCODING_MAP(WIDE(kA64Smaddl4xwwx), NO_VARIANTS(0x9b200000), 502 kFmtRegX, 4, 0, kFmtRegW, 9, 5, kFmtRegW, 20, 16, 503 kFmtRegX, 14, 10, IS_QUAD_OP | REG_DEF0_USE123, 504 "smaddl", "!0x, !1w, !2w, !3x", kFixupNone), 505 ENCODING_MAP(kA64Smulh3xxx, NO_VARIANTS(0x9b407c00), 506 kFmtRegX, 4, 0, kFmtRegX, 9, 5, kFmtRegX, 20, 16, 507 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE12, 508 "smulh", "!0x, !1x, !2x", kFixupNone), 509 ENCODING_MAP(WIDE(kA64Stp4ffXD), CUSTOM_VARIANTS(0x2d000000, 0x6d000000), 510 kFmtRegF, 4, 0, kFmtRegF, 14, 10, kFmtRegXOrSp, 9, 5, 511 kFmtBitBlt, 21, 15, IS_QUAD_OP | REG_USE012 | IS_STORE_OFF, 512 "stp", "!0f, !1f, [!2X, #!3D]", kFixupNone), 513 ENCODING_MAP(WIDE(kA64Stp4rrXD), SF_VARIANTS(0x29000000), 514 kFmtRegR, 4, 0, kFmtRegR, 14, 10, kFmtRegXOrSp, 9, 5, 515 kFmtBitBlt, 21, 15, IS_QUAD_OP | REG_USE012 | IS_STORE_OFF, 516 "stp", "!0r, !1r, [!2X, #!3D]", kFixupNone), 517 ENCODING_MAP(WIDE(kA64StpPost4rrXD), CUSTOM_VARIANTS(0x28800000, 0xa8800000), 518 kFmtRegR, 4, 0, kFmtRegR, 14, 10, kFmtRegXOrSp, 9, 5, 519 kFmtBitBlt, 21, 15, IS_QUAD_OP | REG_DEF2 | REG_USE012 | IS_STORE, 520 "stp", "!0r, !1r, [!2X], #!3D", kFixupNone), 521 ENCODING_MAP(WIDE(kA64StpPre4rrXD), CUSTOM_VARIANTS(0x29800000, 0xa9800000), 522 kFmtRegR, 4, 0, kFmtRegR, 14, 10, kFmtRegXOrSp, 9, 5, 523 kFmtBitBlt, 21, 15, IS_QUAD_OP | REG_DEF2 | REG_USE012 | IS_STORE, 524 "stp", "!0r, !1r, [!2X, #!3D]!!", kFixupNone), 525 ENCODING_MAP(FWIDE(kA64Str3fXD), CUSTOM_VARIANTS(0xbd000000, 0xfd000000), 526 kFmtRegF, 4, 0, kFmtRegXOrSp, 9, 5, kFmtBitBlt, 21, 10, 527 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_USE01 | IS_STORE_OFF, 528 "str", "!0f, [!1X, #!2D]", kFixupNone), 529 ENCODING_MAP(FWIDE(kA64Str4fXxG), CUSTOM_VARIANTS(0xbc206800, 0xfc206800), 530 kFmtRegF, 4, 0, kFmtRegXOrSp, 9, 5, kFmtRegX, 20, 16, 531 kFmtBitBlt, 12, 12, IS_QUAD_OP | REG_USE012 | IS_STORE, 532 "str", "!0f, [!1X, !2x!3G]", kFixupNone), 533 ENCODING_MAP(WIDE(kA64Str3rXD), SIZE_VARIANTS(0xb9000000), 534 kFmtRegR, 4, 0, kFmtRegXOrSp, 9, 5, kFmtBitBlt, 21, 10, 535 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_USE01 | IS_STORE_OFF, 536 "str", "!0r, [!1X, #!2D]", kFixupNone), 537 ENCODING_MAP(WIDE(kA64Str4rXxG), SIZE_VARIANTS(0xb8206800), 538 kFmtRegR, 4, 0, kFmtRegXOrSp, 9, 5, kFmtRegX, 20, 16, 539 kFmtBitBlt, 12, 12, IS_QUAD_OP | REG_USE012 | IS_STORE, 540 "str", "!0r, [!1X, !2x!3G]", kFixupNone), 541 ENCODING_MAP(kA64Strb3wXd, NO_VARIANTS(0x39000000), 542 kFmtRegW, 4, 0, kFmtRegXOrSp, 9, 5, kFmtBitBlt, 21, 10, 543 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_USE01 | IS_STORE_OFF, 544 "strb", "!0w, [!1X, #!2d]", kFixupNone), 545 ENCODING_MAP(kA64Strb3wXx, NO_VARIANTS(0x38206800), 546 kFmtRegW, 4, 0, kFmtRegXOrSp, 9, 5, kFmtRegX, 20, 16, 547 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_USE012 | IS_STORE, 548 "strb", "!0w, [!1X, !2x]", kFixupNone), 549 ENCODING_MAP(kA64Strh3wXF, NO_VARIANTS(0x79000000), 550 kFmtRegW, 4, 0, kFmtRegXOrSp, 9, 5, kFmtBitBlt, 21, 10, 551 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_USE01 | IS_STORE_OFF, 552 "strh", "!0w, [!1X, #!2F]", kFixupNone), 553 ENCODING_MAP(kA64Strh4wXxd, NO_VARIANTS(0x78206800), 554 kFmtRegW, 4, 0, kFmtRegXOrSp, 9, 5, kFmtRegX, 20, 16, 555 kFmtBitBlt, 12, 12, IS_QUAD_OP | REG_USE012 | IS_STORE, 556 "strh", "!0w, [!1X, !2x, lsl #!3d]", kFixupNone), 557 ENCODING_MAP(WIDE(kA64StrPost3rXd), SIZE_VARIANTS(0xb8000400), 558 kFmtRegR, 4, 0, kFmtRegXOrSp, 9, 5, kFmtBitBlt, 20, 12, 559 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_USE01 | REG_DEF1 | IS_STORE, 560 "str", "!0r, [!1X], #!2d", kFixupNone), 561 ENCODING_MAP(FWIDE(kA64Stur3fXd), CUSTOM_VARIANTS(0xbc000000, 0xfc000000), 562 kFmtRegF, 4, 0, kFmtRegXOrSp, 9, 5, kFmtBitBlt, 20, 12, 563 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_USE01 | IS_STORE, 564 "stur", "!0f, [!1X, #!2d]", kFixupNone), 565 ENCODING_MAP(WIDE(kA64Stur3rXd), SIZE_VARIANTS(0xb8000000), 566 kFmtRegR, 4, 0, kFmtRegXOrSp, 9, 5, kFmtBitBlt, 20, 12, 567 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_USE01 | IS_STORE, 568 "stur", "!0r, [!1X, #!2d]", kFixupNone), 569 ENCODING_MAP(WIDE(kA64Stxr3wrX), SIZE_VARIANTS(0x88007c00), 570 kFmtRegW, 20, 16, kFmtRegR, 4, 0, kFmtRegXOrSp, 9, 5, 571 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE12 | IS_STOREX, 572 "stxr", "!0w, !1r, [!2X]", kFixupNone), 573 ENCODING_MAP(WIDE(kA64Stlxr3wrX), SIZE_VARIANTS(0x8800fc00), 574 kFmtRegW, 20, 16, kFmtRegR, 4, 0, kFmtRegXOrSp, 9, 5, 575 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE12 | IS_STOREX, 576 "stlxr", "!0w, !1r, [!2X]", kFixupNone), 577 ENCODING_MAP(WIDE(kA64Sub4RRdT), SF_VARIANTS(0x51000000), 578 kFmtRegROrSp, 4, 0, kFmtRegROrSp, 9, 5, kFmtBitBlt, 21, 10, 579 kFmtBitBlt, 23, 22, IS_QUAD_OP | REG_DEF0_USE1, 580 "sub", "!0R, !1R, #!2d!3T", kFixupNone), 581 ENCODING_MAP(WIDE(kA64Sub4rrro), SF_VARIANTS(0x4b000000), 582 kFmtRegR, 4, 0, kFmtRegR, 9, 5, kFmtRegR, 20, 16, 583 kFmtShift, -1, -1, IS_QUAD_OP | REG_DEF0_USE12, 584 "sub", "!0r, !1r, !2r!3o", kFixupNone), 585 ENCODING_MAP(WIDE(kA64Sub4RRre), SF_VARIANTS(0x4b200000), 586 kFmtRegROrSp, 4, 0, kFmtRegROrSp, 9, 5, kFmtRegR, 20, 16, 587 kFmtExtend, -1, -1, IS_QUAD_OP | REG_DEF0_USE12, 588 "sub", "!0r, !1r, !2r!3e", kFixupNone), 589 ENCODING_MAP(WIDE(kA64Subs3rRd), SF_VARIANTS(0x71000000), 590 kFmtRegR, 4, 0, kFmtRegROrSp, 9, 5, kFmtBitBlt, 21, 10, 591 kFmtUnused, -1, -1, IS_TERTIARY_OP | REG_DEF0_USE1 | SETS_CCODES, 592 "subs", "!0r, !1R, #!2d", kFixupNone), 593 ENCODING_MAP(WIDE(kA64Tst3rro), SF_VARIANTS(0x6a000000), 594 kFmtRegR, 9, 5, kFmtRegR, 20, 16, kFmtShift, -1, -1, 595 kFmtUnused, -1, -1, IS_QUAD_OP | REG_USE01 | SETS_CCODES, 596 "tst", "!0r, !1r!2o", kFixupNone), 597 ENCODING_MAP(WIDE(kA64Ubfm4rrdd), SF_N_VARIANTS(0x53000000), 598 kFmtRegR, 4, 0, kFmtRegR, 9, 5, kFmtBitBlt, 21, 16, 599 kFmtBitBlt, 15, 10, IS_QUAD_OP | REG_DEF0_USE1, 600 "ubfm", "!0r, !1r, !2d, !3d", kFixupNone), 601}; 602 603// new_lir replaces orig_lir in the pcrel_fixup list. 604void Arm64Mir2Lir::ReplaceFixup(LIR* prev_lir, LIR* orig_lir, LIR* new_lir) { 605 new_lir->u.a.pcrel_next = orig_lir->u.a.pcrel_next; 606 if (UNLIKELY(prev_lir == NULL)) { 607 first_fixup_ = new_lir; 608 } else { 609 prev_lir->u.a.pcrel_next = new_lir; 610 } 611 orig_lir->flags.fixup = kFixupNone; 612} 613 614// new_lir is inserted before orig_lir in the pcrel_fixup list. 615void Arm64Mir2Lir::InsertFixupBefore(LIR* prev_lir, LIR* orig_lir, LIR* new_lir) { 616 new_lir->u.a.pcrel_next = orig_lir; 617 if (UNLIKELY(prev_lir == NULL)) { 618 first_fixup_ = new_lir; 619 } else { 620 DCHECK(prev_lir->u.a.pcrel_next == orig_lir); 621 prev_lir->u.a.pcrel_next = new_lir; 622 } 623} 624 625/* Nop, used for aligning code. Nop is an alias for hint #0. */ 626#define PADDING_NOP (UINT32_C(0xd503201f)) 627 628uint8_t* Arm64Mir2Lir::EncodeLIRs(uint8_t* write_pos, LIR* lir) { 629 for (; lir != nullptr; lir = NEXT_LIR(lir)) { 630 bool opcode_is_wide = IS_WIDE(lir->opcode); 631 ArmOpcode opcode = UNWIDE(lir->opcode); 632 633 if (UNLIKELY(IsPseudoLirOp(opcode))) { 634 continue; 635 } 636 637 if (LIKELY(!lir->flags.is_nop)) { 638 const ArmEncodingMap *encoder = &EncodingMap[opcode]; 639 640 // Select the right variant of the skeleton. 641 uint32_t bits = opcode_is_wide ? encoder->xskeleton : encoder->wskeleton; 642 DCHECK(!opcode_is_wide || IS_WIDE(encoder->opcode)); 643 644 for (int i = 0; i < 4; i++) { 645 ArmEncodingKind kind = encoder->field_loc[i].kind; 646 uint32_t operand = lir->operands[i]; 647 uint32_t value; 648 649 if (LIKELY(static_cast<unsigned>(kind) <= kFmtBitBlt)) { 650 // Note: this will handle kFmtReg* and kFmtBitBlt. 651 652 if (static_cast<unsigned>(kind) < kFmtBitBlt) { 653 bool is_zero = A64_REG_IS_ZR(operand); 654 655 if (kIsDebugBuild && (kFailOnSizeError || kReportSizeError)) { 656 // Register usage checks: First establish register usage requirements based on the 657 // format in `kind'. 658 bool want_float = false; // Want a float (rather than core) register. 659 bool want_64_bit = false; // Want a 64-bit (rather than 32-bit) register. 660 bool want_var_size = true; // Want register with variable size (kFmtReg{R,F}). 661 bool want_zero = false; // Want the zero (rather than sp) register. 662 switch (kind) { 663 case kFmtRegX: 664 want_64_bit = true; 665 // Intentional fall-through. 666 case kFmtRegW: 667 want_var_size = false; 668 // Intentional fall-through. 669 case kFmtRegR: 670 want_zero = true; 671 break; 672 case kFmtRegXOrSp: 673 want_64_bit = true; 674 // Intentional fall-through. 675 case kFmtRegWOrSp: 676 want_var_size = false; 677 break; 678 case kFmtRegROrSp: 679 break; 680 case kFmtRegD: 681 want_64_bit = true; 682 // Intentional fall-through. 683 case kFmtRegS: 684 want_var_size = false; 685 // Intentional fall-through. 686 case kFmtRegF: 687 want_float = true; 688 break; 689 default: 690 LOG(FATAL) << "Bad fmt for arg n. " << i << " of " << encoder->name 691 << " (" << kind << ")"; 692 break; 693 } 694 695 // want_var_size == true means kind == kFmtReg{R,F}. In these two cases, we want 696 // the register size to be coherent with the instruction width. 697 if (want_var_size) { 698 want_64_bit = opcode_is_wide; 699 } 700 701 // Now check that the requirements are satisfied. 702 RegStorage reg(operand | RegStorage::kValid); 703 const char *expected = nullptr; 704 if (want_float) { 705 if (!reg.IsFloat()) { 706 expected = "float register"; 707 } else if (reg.IsDouble() != want_64_bit) { 708 expected = (want_64_bit) ? "double register" : "single register"; 709 } 710 } else { 711 if (reg.IsFloat()) { 712 expected = "core register"; 713 } else if (reg.Is64Bit() != want_64_bit) { 714 expected = (want_64_bit) ? "x-register" : "w-register"; 715 } else if (A64_REGSTORAGE_IS_SP_OR_ZR(reg) && is_zero != want_zero) { 716 expected = (want_zero) ? "zero-register" : "sp-register"; 717 } 718 } 719 720 // Fail, if `expected' contains an unsatisfied requirement. 721 if (expected != nullptr) { 722 LOG(WARNING) << "Method: " << PrettyMethod(cu_->method_idx, *cu_->dex_file) 723 << " @ 0x" << std::hex << lir->dalvik_offset; 724 if (kFailOnSizeError) { 725 LOG(FATAL) << "Bad argument n. " << i << " of " << encoder->name 726 << ". Expected " << expected << ", got 0x" << std::hex << operand; 727 } else { 728 LOG(WARNING) << "Bad argument n. " << i << " of " << encoder->name 729 << ". Expected " << expected << ", got 0x" << std::hex << operand; 730 } 731 } 732 } 733 734 // In the lines below, we rely on (operand & 0x1f) == 31 to be true for register sp 735 // and zr. This means that these two registers do not need any special treatment, as 736 // their bottom 5 bits are correctly set to 31 == 0b11111, which is the right 737 // value for encoding both sp and zr. 738 COMPILE_ASSERT((rxzr & 0x1f) == 0x1f, rzr_register_number_must_be_31); 739 COMPILE_ASSERT((rsp & 0x1f) == 0x1f, rsp_register_number_must_be_31); 740 } 741 742 value = (operand << encoder->field_loc[i].start) & 743 ((1 << (encoder->field_loc[i].end + 1)) - 1); 744 bits |= value; 745 } else { 746 switch (kind) { 747 case kFmtSkip: 748 break; // Nothing to do, but continue to next. 749 case kFmtUnused: 750 i = 4; // Done, break out of the enclosing loop. 751 break; 752 case kFmtShift: 753 // Intentional fallthrough. 754 case kFmtExtend: 755 DCHECK_EQ((operand & (1 << 6)) == 0, kind == kFmtShift); 756 value = (operand & 0x3f) << 10; 757 value |= ((operand & 0x1c0) >> 6) << 21; 758 bits |= value; 759 break; 760 case kFmtImm21: 761 value = (operand & 0x3) << 29; 762 value |= ((operand & 0x1ffffc) >> 2) << 5; 763 bits |= value; 764 break; 765 default: 766 LOG(FATAL) << "Bad fmt for arg. " << i << " in " << encoder->name 767 << " (" << kind << ")"; 768 } 769 } 770 } 771 772 DCHECK_EQ(encoder->size, 4); 773 write_pos[0] = (bits & 0xff); 774 write_pos[1] = ((bits >> 8) & 0xff); 775 write_pos[2] = ((bits >> 16) & 0xff); 776 write_pos[3] = ((bits >> 24) & 0xff); 777 write_pos += 4; 778 } 779 } 780 781 return write_pos; 782} 783 784// Align data offset on 8 byte boundary: it will only contain double-word items, as word immediates 785// are better set directly from the code (they will require no more than 2 instructions). 786#define ALIGNED_DATA_OFFSET(offset) (((offset) + 0x7) & ~0x7) 787 788// Assemble the LIR into binary instruction format. 789void Arm64Mir2Lir::AssembleLIR() { 790 LIR* lir; 791 LIR* prev_lir; 792 cu_->NewTimingSplit("Assemble"); 793 int assembler_retries = 0; 794 CodeOffset starting_offset = LinkFixupInsns(first_lir_insn_, last_lir_insn_, 0); 795 data_offset_ = ALIGNED_DATA_OFFSET(starting_offset); 796 int32_t offset_adjustment; 797 AssignDataOffsets(); 798 799 /* 800 * Note: generation must be 1 on first pass (to distinguish from initialized state of 0 801 * for non-visited nodes). Start at zero here, and bit will be flipped to 1 on entry to the loop. 802 */ 803 int generation = 0; 804 while (true) { 805 // TODO(Arm64): check whether passes and offset adjustments are really necessary. 806 // Currently they aren't, as - in the fixups below - LIR are never inserted. 807 // Things can be different if jump ranges above 1 MB need to be supported. 808 // If they are not, then we can get rid of the assembler retry logic. 809 810 offset_adjustment = 0; 811 AssemblerStatus res = kSuccess; // Assume success 812 generation ^= 1; 813 // Note: nodes requiring possible fixup linked in ascending order. 814 lir = first_fixup_; 815 prev_lir = NULL; 816 while (lir != NULL) { 817 /* 818 * NOTE: the lir being considered here will be encoded following the switch (so long as 819 * we're not in a retry situation). However, any new non-pc_rel instructions inserted 820 * due to retry must be explicitly encoded at the time of insertion. Note that 821 * inserted instructions don't need use/def flags, but do need size and pc-rel status 822 * properly updated. 823 */ 824 lir->offset += offset_adjustment; 825 // During pass, allows us to tell whether a node has been updated with offset_adjustment yet. 826 lir->flags.generation = generation; 827 switch (static_cast<FixupKind>(lir->flags.fixup)) { 828 case kFixupLabel: 829 case kFixupNone: 830 case kFixupVLoad: 831 break; 832 case kFixupT1Branch: { 833 LIR *target_lir = lir->target; 834 DCHECK(target_lir); 835 CodeOffset pc = lir->offset; 836 CodeOffset target = target_lir->offset + 837 ((target_lir->flags.generation == lir->flags.generation) ? 0 : offset_adjustment); 838 int32_t delta = target - pc; 839 if (!((delta & 0x3) == 0 && IS_SIGNED_IMM19(delta >> 2))) { 840 LOG(FATAL) << "Invalid jump range in kFixupT1Branch"; 841 } 842 lir->operands[0] = delta >> 2; 843 break; 844 } 845 case kFixupLoad: 846 case kFixupCBxZ: 847 case kFixupCondBranch: { 848 LIR *target_lir = lir->target; 849 DCHECK(target_lir); 850 CodeOffset pc = lir->offset; 851 CodeOffset target = target_lir->offset + 852 ((target_lir->flags.generation == lir->flags.generation) ? 0 : offset_adjustment); 853 int32_t delta = target - pc; 854 if (!((delta & 0x3) == 0 && IS_SIGNED_IMM19(delta >> 2))) { 855 LOG(FATAL) << "Invalid jump range in kFixupLoad"; 856 } 857 lir->operands[1] = delta >> 2; 858 break; 859 } 860 case kFixupAdr: { 861 LIR* target_lir = lir->target; 862 int32_t delta; 863 if (target_lir) { 864 CodeOffset target_offs = ((target_lir->flags.generation == lir->flags.generation) ? 865 0 : offset_adjustment) + target_lir->offset; 866 delta = target_offs - lir->offset; 867 } else if (lir->operands[2] >= 0) { 868 EmbeddedData* tab = reinterpret_cast<EmbeddedData*>(UnwrapPointer(lir->operands[2])); 869 delta = tab->offset + offset_adjustment - lir->offset; 870 } else { 871 // No fixup: this usage allows to retrieve the current PC. 872 delta = lir->operands[1]; 873 } 874 if (!IS_SIGNED_IMM21(delta)) { 875 LOG(FATAL) << "Jump range above 1MB in kFixupAdr"; 876 } 877 lir->operands[1] = delta; 878 break; 879 } 880 default: 881 LOG(FATAL) << "Unexpected case " << lir->flags.fixup; 882 } 883 prev_lir = lir; 884 lir = lir->u.a.pcrel_next; 885 } 886 887 if (res == kSuccess) { 888 break; 889 } else { 890 assembler_retries++; 891 if (assembler_retries > MAX_ASSEMBLER_RETRIES) { 892 CodegenDump(); 893 LOG(FATAL) << "Assembler error - too many retries"; 894 } 895 starting_offset += offset_adjustment; 896 data_offset_ = ALIGNED_DATA_OFFSET(starting_offset); 897 AssignDataOffsets(); 898 } 899 } 900 901 // Build the CodeBuffer. 902 DCHECK_LE(data_offset_, total_size_); 903 code_buffer_.reserve(total_size_); 904 code_buffer_.resize(starting_offset); 905 uint8_t* write_pos = &code_buffer_[0]; 906 write_pos = EncodeLIRs(write_pos, first_lir_insn_); 907 DCHECK_EQ(static_cast<CodeOffset>(write_pos - &code_buffer_[0]), starting_offset); 908 909 DCHECK_EQ(data_offset_, ALIGNED_DATA_OFFSET(code_buffer_.size())); 910 911 // Install literals 912 InstallLiteralPools(); 913 914 // Install switch tables 915 InstallSwitchTables(); 916 917 // Install fill array data 918 InstallFillArrayData(); 919 920 // Create the mapping table and native offset to reference map. 921 cu_->NewTimingSplit("PcMappingTable"); 922 CreateMappingTables(); 923 924 cu_->NewTimingSplit("GcMap"); 925 CreateNativeGcMap(); 926} 927 928size_t Arm64Mir2Lir::GetInsnSize(LIR* lir) { 929 ArmOpcode opcode = UNWIDE(lir->opcode); 930 DCHECK(!IsPseudoLirOp(opcode)); 931 return EncodingMap[opcode].size; 932} 933 934// Encode instruction bit pattern and assign offsets. 935uint32_t Arm64Mir2Lir::LinkFixupInsns(LIR* head_lir, LIR* tail_lir, uint32_t offset) { 936 LIR* end_lir = tail_lir->next; 937 938 LIR* last_fixup = NULL; 939 for (LIR* lir = head_lir; lir != end_lir; lir = NEXT_LIR(lir)) { 940 ArmOpcode opcode = UNWIDE(lir->opcode); 941 if (!lir->flags.is_nop) { 942 if (lir->flags.fixup != kFixupNone) { 943 if (!IsPseudoLirOp(opcode)) { 944 lir->flags.size = EncodingMap[opcode].size; 945 lir->flags.fixup = EncodingMap[opcode].fixup; 946 } else { 947 DCHECK_NE(static_cast<int>(opcode), kPseudoPseudoAlign4); 948 lir->flags.size = 0; 949 lir->flags.fixup = kFixupLabel; 950 } 951 // Link into the fixup chain. 952 lir->flags.use_def_invalid = true; 953 lir->u.a.pcrel_next = NULL; 954 if (first_fixup_ == NULL) { 955 first_fixup_ = lir; 956 } else { 957 last_fixup->u.a.pcrel_next = lir; 958 } 959 last_fixup = lir; 960 lir->offset = offset; 961 } 962 offset += lir->flags.size; 963 } 964 } 965 return offset; 966} 967 968void Arm64Mir2Lir::AssignDataOffsets() { 969 /* Set up offsets for literals */ 970 CodeOffset offset = data_offset_; 971 972 offset = AssignLiteralOffset(offset); 973 974 offset = AssignSwitchTablesOffset(offset); 975 976 total_size_ = AssignFillArrayDataOffset(offset); 977} 978 979} // namespace art 980