/art/compiler/optimizing/ |
H A D | code_generator_arm.h | 109 ArmAssembler* GetAssembler() const; 170 ArmAssembler* GetAssembler() const { return assembler_; } function in class:art::arm::InstructionCodeGeneratorARM 235 ArmAssembler* GetAssembler() OVERRIDE {
|
H A D | code_generator_arm64.h | 151 Arm64Assembler* GetAssembler() const { return assembler_; } function in class:art::arm64::InstructionCodeGeneratorARM64 152 vixl::MacroAssembler* GetVIXLAssembler() { return GetAssembler()->vixl_masm_; } 211 Arm64Assembler* GetAssembler() const; 213 return GetAssembler()->vixl_masm_; 267 Arm64Assembler* GetAssembler() OVERRIDE { return &assembler_; } 268 vixl::MacroAssembler* GetVIXLAssembler() { return GetAssembler()->vixl_masm_; } 359 inline Arm64Assembler* ParallelMoveResolverARM64::GetAssembler() const { function in class:art::arm64::ParallelMoveResolverARM64 360 return codegen_->GetAssembler();
|
H A D | code_generator_mips64.h | 120 Mips64Assembler* GetAssembler() const; 179 Mips64Assembler* GetAssembler() const { return assembler_; } function in class:art::mips64::InstructionCodeGeneratorMIPS64 229 Mips64Assembler* GetAssembler() OVERRIDE { return &assembler_; }
|
H A D | code_generator_x86.h | 103 X86Assembler* GetAssembler() const; 153 X86Assembler* GetAssembler() const { return assembler_; } function in class:art::x86::InstructionCodeGeneratorX86 228 X86Assembler* GetAssembler() OVERRIDE {
|
H A D | code_generator_x86_64.h | 112 X86_64Assembler* GetAssembler() const; 163 X86_64Assembler* GetAssembler() const { return assembler_; } function in class:art::x86_64::InstructionCodeGeneratorX86_64 228 X86_64Assembler* GetAssembler() OVERRIDE {
|
H A D | intrinsics_arm.cc | 33 ArmAssembler* IntrinsicCodeGeneratorARM::GetAssembler() { function in class:art::arm::IntrinsicCodeGeneratorARM 34 return codegen_->GetAssembler(); 41 #define __ codegen->GetAssembler()-> 188 MoveFPToInt(invoke->GetLocations(), true, GetAssembler()); 191 MoveIntToFP(invoke->GetLocations(), true, GetAssembler()); 202 MoveFPToInt(invoke->GetLocations(), false, GetAssembler()); 205 MoveIntToFP(invoke->GetLocations(), false, GetAssembler()); 241 MathAbsFP(invoke->GetLocations(), true, GetAssembler()); 249 MathAbsFP(invoke->GetLocations(), false, GetAssembler()); 298 GenAbsInteger(invoke->GetLocations(), false, GetAssembler()); [all...] |
H A D | intrinsics_x86.cc | 45 X86Assembler* IntrinsicCodeGeneratorX86::GetAssembler() { function in class:art::x86::IntrinsicCodeGeneratorX86 46 return reinterpret_cast<X86Assembler*>(codegen_->GetAssembler()); 59 #define __ reinterpret_cast<X86Assembler*>(codegen->GetAssembler())-> 232 MoveFPToInt(invoke->GetLocations(), true, GetAssembler()); 235 MoveIntToFP(invoke->GetLocations(), true, GetAssembler()); 246 MoveFPToInt(invoke->GetLocations(), false, GetAssembler()); 249 MoveIntToFP(invoke->GetLocations(), false, GetAssembler()); 301 GenReverseBytes(invoke->GetLocations(), Primitive::kPrimInt, GetAssembler()); 317 X86Assembler* assembler = GetAssembler(); 330 GenReverseBytes(invoke->GetLocations(), Primitive::kPrimShort, GetAssembler()); [all...] |
H A D | intrinsics_x86_64.cc | 41 X86_64Assembler* IntrinsicCodeGeneratorX86_64::GetAssembler() { function in class:art::x86_64::IntrinsicCodeGeneratorX86_64 42 return reinterpret_cast<X86_64Assembler*>(codegen_->GetAssembler()); 55 #define __ reinterpret_cast<X86_64Assembler*>(codegen->GetAssembler())-> 197 MoveFPToInt(invoke->GetLocations(), true, GetAssembler()); 200 MoveIntToFP(invoke->GetLocations(), true, GetAssembler()); 211 MoveFPToInt(invoke->GetLocations(), false, GetAssembler()); 214 MoveIntToFP(invoke->GetLocations(), false, GetAssembler()); 253 GenReverseBytes(invoke->GetLocations(), Primitive::kPrimInt, GetAssembler()); 261 GenReverseBytes(invoke->GetLocations(), Primitive::kPrimLong, GetAssembler()); 269 GenReverseBytes(invoke->GetLocations(), Primitive::kPrimShort, GetAssembler()); [all...] |
H A D | code_generator_arm.cc | 56 #define __ reinterpret_cast<ArmAssembler*>(codegen->GetAssembler())-> 319 #define __ reinterpret_cast<ArmAssembler*>(GetAssembler())-> 493 assembler_(codegen->GetAssembler()), 972 if (GetAssembler()->ShifterOperandCanHold(R0, left, CMP, value, &operand)) { 1051 if (GetAssembler()->ShifterOperandCanHold(R0, left, CMP, value, &operand)) { 3569 ArmAssembler* ParallelMoveResolverARM::GetAssembler() const { function in class:art::arm::ParallelMoveResolverARM 3570 return codegen_->GetAssembler();
|
H A D | code_generator_mips64.cc | 106 #define __ down_cast<CodeGeneratorMIPS64*>(codegen)->GetAssembler()-> 417 #define __ down_cast<Mips64Assembler*>(GetAssembler())-> 424 Mips64Assembler* ParallelMoveResolverMIPS64::GetAssembler() const { function in class:art::mips64::ParallelMoveResolverMIPS64 425 return codegen_->GetAssembler(); 1006 assembler_(codegen->GetAssembler()),
|
H A D | code_generator_x86.cc | 46 #define __ reinterpret_cast<X86Assembler*>(codegen->GetAssembler())-> 326 #define __ reinterpret_cast<X86Assembler*>(GetAssembler())-> 466 assembler_(codegen->GetAssembler()), 4028 X86Assembler* ParallelMoveResolverX86::GetAssembler() const { function in class:art::x86::ParallelMoveResolverX86 4029 return codegen_->GetAssembler();
|
H A D | code_generator_x86_64.cc | 49 #define __ reinterpret_cast<X86_64Assembler*>(codegen->GetAssembler())-> 346 #define __ reinterpret_cast<X86_64Assembler*>(GetAssembler())-> 455 assembler_(codegen->GetAssembler()), 3895 X86_64Assembler* ParallelMoveResolverX86_64::GetAssembler() const { function in class:art::x86_64::ParallelMoveResolverX86_64 3896 return codegen_->GetAssembler(); 4463 X86_64Assembler* assembler = GetAssembler();
|
/art/compiler/utils/ |
H A D | assembler_test.h | 48 Ass* GetAssembler() { function in class:art::AssemblerTest
|