/art/compiler/dex/quick/x86/ |
H A D | fp_x86.cc | 74 NewLIR2(op, r_dest, r_src2); 129 NewLIR2(op, r_dest, r_src2); 165 NewLIR2(kX86Cvtsi2ssRR, temp_reg, rl_result.low_reg); 166 NewLIR2(kX86ComissRR, src_reg, temp_reg); 167 LIR* branch_pos_overflow = NewLIR2(kX86Jcc8, 0, kX86CondA); 168 LIR* branch_na_n = NewLIR2(kX86Jcc8, 0, kX86CondP); 169 NewLIR2(kX86Cvttss2siRR, rl_result.low_reg, src_reg); 172 NewLIR2(kX86Xor32RR, rl_result.low_reg, rl_result.low_reg); 187 NewLIR2(kX86Cvtsi2sdRR, temp_reg, rl_result.low_reg); 188 NewLIR2(kX86ComisdR [all...] |
H A D | call_x86.cc | 147 NewLIR2(kX86PcRelAdr, rX86_ARG1, reinterpret_cast<uintptr_t>(tab_rec)); 148 NewLIR2(kX86Add32RR, rX86_ARG1, rX86_ARG2); 160 NewLIR2(kX86Mov32RT, rDX, Thread::ThinLockIdOffset().Int32Value()); 161 NewLIR2(kX86Sal32RI, rDX, LW_LOCK_OWNER_SHIFT); 162 NewLIR2(kX86Xor32RR, rAX, rAX); 164 LIR* branch = NewLIR2(kX86Jcc8, 0, kX86CondEq); 177 NewLIR2(kX86Mov32RT, rDX, Thread::ThinLockIdOffset().Int32Value()); 178 NewLIR2(kX86Sal32RI, rDX, LW_LOCK_OWNER_SHIFT); 181 LIR* branch = NewLIR2(kX86Jcc8, 0, kX86CondNe); 193 NewLIR2(kX86Mov32R [all...] |
H A D | int_x86.cc | 55 NewLIR2(kX86Set8R, r2, kX86CondL); // r2 = (r1:r0) < (r3:r2) ? 1 : 0 56 NewLIR2(kX86Movzx8RR, r2, r2); 59 NewLIR2(kX86Set8R, r0, kX86CondNz); // r0 = (r1:r0) != (r3:r2) ? 1 : 0 60 NewLIR2(kX86Movzx8RR, r0, r0); 90 NewLIR2(kX86Cmp32RR, src1, src2); 92 LIR* branch = NewLIR2(kX86Jcc8, 0 /* lir operand for Jcc offset */ , 102 NewLIR2(kX86Test32RR, reg, reg); 104 NewLIR2(IS_SIMM8(check_value) ? kX86Cmp32RI8 : kX86Cmp32RI, reg, check_value); 107 LIR* branch = NewLIR2(kX86Jcc8, 0 /* lir operand for Jcc offset */ , cc); 141 NewLIR2(kX86MovdxrR [all...] |
H A D | utility_x86.cc | 80 return NewLIR2(kX86XorpsRR, r_dest, r_dest); 88 res = NewLIR2(kX86Xor32RR, r_dest, r_dest); 91 res = NewLIR2(kX86Mov32RI, r_dest, value); 95 NewLIR2(kX86MovdxrRR, r_dest_save, r_dest); 109 LIR* branch = NewLIR2(kX86Jcc8, 0 /* offset to be patched */, 150 return NewLIR2(opcode, r_dest_src1, value); 180 NewLIR2(kX86Mov32RR, r_dest_src1, r_src2); 181 NewLIR2(kX86Sal32RI, r_dest_src1, 24); 182 return NewLIR2(kX86Sar32RI, r_dest_src1, 24); 195 return NewLIR2(opcod [all...] |
/art/compiler/dex/quick/arm/ |
H A D | fp_arm.cc | 168 NewLIR2(op, S2d(rl_result.low_reg, rl_result.high_reg), src_reg); 172 NewLIR2(op, rl_result.low_reg, src_reg); 187 NewLIR2(kThumb2Vcmpd, S2d(rl_src1.low_reg, rl_src2.high_reg), 194 NewLIR2(kThumb2Vcmps, rl_src1.low_reg, rl_src2.low_reg); 262 NewLIR2(kThumb2Vcmpd, S2d(rl_src1.low_reg, rl_src2.high_reg), 271 NewLIR2(kThumb2Vcmps, rl_src1.low_reg, rl_src2.low_reg); 277 NewLIR2(kThumb2MovImmShift, rl_result.low_reg, 292 NewLIR2(kThumb2Vnegs, rl_result.low_reg, rl_src.low_reg); 300 NewLIR2(kThumb2Vnegd, S2d(rl_result.low_reg, rl_result.high_reg), 312 NewLIR2(kThumb2Vsqrt [all...] |
H A D | utility_arm.cc | 77 NewLIR2(kThumb2Vmovs_IMM8, r_dest, 0); 83 return NewLIR2(kThumb2Vmovs_IMM8, r_dest, encoded_imm); 183 return NewLIR2(kThumbMovImm, r_dest, value); 188 res = NewLIR2(kThumb2MovImmShift, r_dest, mod_imm); 193 res = NewLIR2(kThumb2MvnImm12, r_dest, mod_imm); 198 res = NewLIR2(kThumb2MovImm16, r_dest, value); 202 res = NewLIR2(kThumb2MovImm16, r_dest, Low16Bits(value)); 203 NewLIR2(kThumb2MovImm16H, r_dest, High16Bits(value)); 214 LIR* branch = NewLIR2(kThumb2BCond, 0 /* offset to be patched */, 333 return NewLIR2(opcod [all...] |
H A D | int_arm.cc | 68 return NewLIR2(kThumb2It, code, mask); 101 NewLIR2(kThumb2MovImmShift, t_reg, ModifiedImmediate(-1)); 318 branch = NewLIR2((arm_cond == kArmCondEq) ? kThumb2Cbz : kThumb2Cbnz, 323 NewLIR2(kThumbCmpRI8, reg, check_value); 325 NewLIR2(kThumb2CmpRI12, reg, mod_imm); 331 branch = NewLIR2(kThumbBCond, 0, arm_cond); 595 NewLIR2(kThumbSubRI8, rARM_SUSPEND, 1);
|
H A D | call_arm.cc | 347 NewLIR2(kThumb2LdmiaWB, rBase, (1 << r_key) | (1 << r_disp));
|
/art/compiler/dex/quick/mips/ |
H A D | int_mips.cc | 116 branch = NewLIR2(br_op, src1, src2); 192 NewLIR2(kMipsMtc1, src_lo, dest_lo); 193 NewLIR2(kMipsMtc1, src_hi, dest_hi); 197 NewLIR2(kMipsMfc1, dest_lo, src_lo); 198 NewLIR2(kMipsMfc1, dest_hi, src_hi); 231 NewLIR2(kMipsMflo, rl_result.low_reg, r_LO); 233 NewLIR2(kMipsMfhi, rl_result.low_reg, r_HI); 245 NewLIR2(kMipsMflo, rl_result.low_reg, r_LO); 247 NewLIR2(kMipsMfhi, rl_result.low_reg, r_HI);
|
H A D | utility_mips.cc | 90 res = NewLIR2(kMipsMove, r_dest, r_ZERO); 96 res = NewLIR2(kMipsLui, r_dest, value>>16); 102 NewLIR2(kMipsMtc1, r_dest, r_dest_save); 127 return NewLIR2(opcode, r_RA, r_dest_src); 149 res = NewLIR2(opcode, r_dest_src1, abs_value); 154 NewLIR2(opcode, r_dest_src1, r_scratch); 305 res = NewLIR2(kMipsSeb, r_dest_src1, r_src2); 313 res = NewLIR2(kMipsSeh, r_dest_src1, r_src2); 325 return NewLIR2(opcode, r_dest_src1, r_src2);
|
H A D | fp_mips.cc | 167 NewLIR2(op, S2d(rl_result.low_reg, rl_result.high_reg), src_reg); 171 NewLIR2(op, rl_result.low_reg, src_reg);
|
H A D | call_mips.cc | 86 NewLIR2(kMipsLui, rEnd, size_hi);
|
/art/compiler/dex/quick/ |
H A D | mir_to_lir-inl.h | 85 inline LIR* Mir2Lir::NewLIR2(int opcode, int dest, int src1) { function in class:art::Mir2Lir
|
H A D | mir_to_lir.h | 266 LIR* NewLIR2(int opcode, int dest, int src1);
|