H A D | intrinsics_arm64.cc | 236 #define __ masm-> 254 static void MoveFPToInt(LocationSummary* locations, bool is64bit, MacroAssembler* masm) { argument 261 static void MoveIntToFP(LocationSummary* locations, bool is64bit, MacroAssembler* masm) { argument 306 MacroAssembler* masm) { 360 MacroAssembler* masm) { 387 MacroAssembler* masm) { 415 MacroAssembler* masm) { 440 static void GenBitCount(HInvoke* instr, Primitive::Type type, MacroAssembler* masm) { argument 445 UseScratchRegisterScope temps(masm); 481 static void MathAbsFP(LocationSummary* locations, bool is64bit, MacroAssembler* masm) { argument 304 GenReverseBytes(LocationSummary* locations, Primitive::Type type, MacroAssembler* masm) argument 358 GenNumberOfLeadingZeros(LocationSummary* locations, Primitive::Type type, MacroAssembler* masm) argument 385 GenNumberOfTrailingZeros(LocationSummary* locations, Primitive::Type type, MacroAssembler* masm) argument 413 GenReverse(LocationSummary* locations, Primitive::Type type, MacroAssembler* masm) argument 515 GenAbsInteger(LocationSummary* locations, bool is64bit, MacroAssembler* masm) argument 544 GenMinMaxFP(LocationSummary* locations, bool is_min, bool is_double, MacroAssembler* masm) argument 604 GenMinMax(LocationSummary* locations, bool is_min, bool is_long, MacroAssembler* masm) argument 658 MacroAssembler* masm = GetVIXLAssembler(); local 668 MacroAssembler* masm = GetVIXLAssembler(); local 678 MacroAssembler* masm = GetVIXLAssembler(); local 688 MacroAssembler* masm = GetVIXLAssembler(); local 701 GenMathRound(HInvoke* invoke, bool is_double, vixl::aarch64::MacroAssembler* masm) argument 759 MacroAssembler* masm = GetVIXLAssembler(); local 769 MacroAssembler* masm = GetVIXLAssembler(); local 779 MacroAssembler* masm = GetVIXLAssembler(); local 789 MacroAssembler* masm = GetVIXLAssembler(); local 807 MacroAssembler* masm = GetVIXLAssembler(); local 817 MacroAssembler* masm = GetVIXLAssembler(); local 827 MacroAssembler* masm = GetVIXLAssembler(); local 837 MacroAssembler* masm = GetVIXLAssembler(); local 1001 MacroAssembler* masm = codegen->GetVIXLAssembler(); local 1131 MacroAssembler* masm = codegen->GetVIXLAssembler(); local 1264 MacroAssembler* masm = GetVIXLAssembler(); local 1505 MacroAssembler* masm = GetVIXLAssembler(); local 1658 GenerateVisitStringIndexOf(HInvoke* invoke, MacroAssembler* masm, CodeGeneratorARM64* codegen, ArenaAllocator* allocator, bool start_at_zero) argument 1755 MacroAssembler* masm = GetVIXLAssembler(); local 1801 MacroAssembler* masm = GetVIXLAssembler(); local 2003 MacroAssembler* masm = GetVIXLAssembler(); local 2156 CheckSystemArrayCopyPosition(MacroAssembler* masm, const Location& pos, const Register& input, const Location& length, SlowPathCodeARM64* slow_path, const Register& temp, bool length_is_input_length = false) argument 2203 GenSystemArrayCopyAddresses(MacroAssembler* masm, Primitive::Type type, const Register& src, const Location& src_pos, const Register& dst, const Location& dst_pos, const Location& copy_length, const Register& src_base, const Register& dst_base, const Register& src_end) argument 2245 MacroAssembler* masm = GetVIXLAssembler(); local 2412 MacroAssembler* masm = GetVIXLAssembler(); local 2864 GenIsInfinite(LocationSummary* locations, bool is64bit, MacroAssembler* masm) argument 2918 MacroAssembler* masm = GetVIXLAssembler(); local 2978 MacroAssembler* masm = GetVIXLAssembler(); local [all...] |