/art/compiler/optimizing/ |
H A D | intrinsics_mips.h | 35 explicit IntrinsicLocationsBuilderMIPS(CodeGeneratorMIPS* codegen); 59 explicit IntrinsicCodeGeneratorMIPS(CodeGeneratorMIPS* codegen) : codegen_(codegen) {} argument
|
H A D | intrinsics_mips64.h | 35 explicit IntrinsicLocationsBuilderMIPS64(CodeGeneratorMIPS64* codegen); 59 explicit IntrinsicCodeGeneratorMIPS64(CodeGeneratorMIPS64* codegen) : codegen_(codegen) {} argument
|
H A D | intrinsics_x86.h | 35 explicit IntrinsicLocationsBuilderX86(CodeGeneratorX86* codegen); 60 explicit IntrinsicCodeGeneratorX86(CodeGeneratorX86* codegen) : codegen_(codegen) {} argument
|
H A D | intrinsics_x86_64.h | 35 explicit IntrinsicLocationsBuilderX86_64(CodeGeneratorX86_64* codegen); 60 explicit IntrinsicCodeGeneratorX86_64(CodeGeneratorX86_64* codegen) : codegen_(codegen) {} argument
|
H A D | pc_relative_fixups_x86.h | 31 PcRelativeFixups(HGraph* graph, CodeGenerator* codegen, OptimizingCompilerStats* stats) argument 33 codegen_(codegen) {}
|
H A D | intrinsics_arm.h | 66 explicit IntrinsicCodeGeneratorARM(CodeGeneratorARM* codegen) : codegen_(codegen) {} argument
|
H A D | intrinsics_arm64.h | 64 explicit IntrinsicCodeGeneratorARM64(CodeGeneratorARM64* codegen) : codegen_(codegen) {} argument
|
H A D | intrinsics_utils.h | 44 Location MoveArguments(CodeGenerator* codegen) { argument 46 IntrinsicVisitor::MoveArguments(invoke_, codegen, &calling_convention_visitor); 50 void EmitNativeCode(CodeGenerator* codegen) OVERRIDE { 51 Assembler* assembler = codegen->GetAssembler(); 54 SaveLiveRegisters(codegen, invoke_->GetLocations()); 56 Location method_loc = MoveArguments(codegen); 59 codegen->GenerateStaticOrDirectCall(invoke_->AsInvokeStaticOrDirect(), method_loc); 61 codegen->GenerateVirtualCall(invoke_->AsInvokeVirtual(), method_loc); 63 codegen->RecordPcInfo(invoke_, invoke_->GetDexPc(), this); 70 codegen [all...] |
H A D | sharpening.h | 36 CodeGenerator* codegen, 40 codegen_(codegen), 35 HSharpening(HGraph* graph, CodeGenerator* codegen, const DexCompilationUnit& compilation_unit, CompilerDriver* compiler_driver) argument
|
H A D | inliner.h | 37 CodeGenerator* codegen, 49 codegen_(codegen), 35 HInliner(HGraph* outer_graph, HGraph* outermost_graph, CodeGenerator* codegen, const DexCompilationUnit& outer_compilation_unit, const DexCompilationUnit& caller_compilation_unit, CompilerDriver* compiler_driver, StackHandleScopeCollection* handles, OptimizingCompilerStats* stats, size_t total_number_of_dex_registers, size_t depth) argument
|
H A D | intrinsics.h | 84 CodeGenerator* codegen, 102 HParallelMove parallel_move(codegen->GetGraph()->GetArena()); 112 codegen->GetMoveResolver()->EmitNativeCode(¶llel_move); 83 MoveArguments(HInvoke* invoke, CodeGenerator* codegen, InvokeDexCallingConventionVisitor* calling_convention_visitor) argument
|
H A D | pc_relative_fixups_x86.cc | 29 PCRelativeHandlerVisitor(HGraph* graph, CodeGenerator* codegen) argument 31 codegen_(down_cast<CodeGeneratorX86*>(codegen)),
|
H A D | code_generator_mips.h | 135 ParallelMoveResolverMIPS(ArenaAllocator* allocator, CodeGeneratorMIPS* codegen) argument 136 : ParallelMoveResolverWithSwap(allocator), codegen_(codegen) {} 170 LocationsBuilderMIPS(HGraph* graph, CodeGeneratorMIPS* codegen) argument 171 : HGraphVisitor(graph), codegen_(codegen) {} 203 InstructionCodeGeneratorMIPS(HGraph* graph, CodeGeneratorMIPS* codegen);
|
H A D | code_generator_mips64.h | 135 ParallelMoveResolverMIPS64(ArenaAllocator* allocator, CodeGeneratorMIPS64* codegen) argument 136 : ParallelMoveResolverWithSwap(allocator), codegen_(codegen) {} 170 LocationsBuilderMIPS64(HGraph* graph, CodeGeneratorMIPS64* codegen) argument 171 : HGraphVisitor(graph), codegen_(codegen) {} 203 InstructionCodeGeneratorMIPS64(HGraph* graph, CodeGeneratorMIPS64* codegen);
|
H A D | code_generator_x86_64.h | 131 ParallelMoveResolverX86_64(ArenaAllocator* allocator, CodeGeneratorX86_64* codegen) argument 132 : ParallelMoveResolverWithSwap(allocator), codegen_(codegen) {} 157 LocationsBuilderX86_64(HGraph* graph, CodeGeneratorX86_64* codegen) argument 158 : HGraphVisitor(graph), codegen_(codegen) {} 189 InstructionCodeGeneratorX86_64(HGraph* graph, CodeGeneratorX86_64* codegen);
|
H A D | code_generator_arm.h | 134 ParallelMoveResolverARM(ArenaAllocator* allocator, CodeGeneratorARM* codegen) argument 135 : ParallelMoveResolverWithSwap(allocator), codegen_(codegen) {} 155 LocationsBuilderARM(HGraph* graph, CodeGeneratorARM* codegen) argument 156 : HGraphVisitor(graph), codegen_(codegen) {} 194 InstructionCodeGeneratorARM(HGraph* graph, CodeGeneratorARM* codegen);
|
H A D | code_generator_arm64.h | 76 void SaveLiveRegisters(CodeGenerator* codegen, LocationSummary* locations) OVERRIDE; 77 void RestoreLiveRegisters(CodeGenerator* codegen, LocationSummary* locations) OVERRIDE; 93 void EmitTable(CodeGeneratorARM64* codegen); 193 InstructionCodeGeneratorARM64(HGraph* graph, CodeGeneratorARM64* codegen); 284 LocationsBuilderARM64(HGraph* graph, CodeGeneratorARM64* codegen) argument 285 : HGraphVisitor(graph), codegen_(codegen) {} 317 ParallelMoveResolverARM64(ArenaAllocator* allocator, CodeGeneratorARM64* codegen) argument 318 : ParallelMoveResolverNoSwap(allocator), codegen_(codegen), vixl_temps_() {}
|
H A D | code_generator_x86.h | 127 ParallelMoveResolverX86(ArenaAllocator* allocator, CodeGeneratorX86* codegen) argument 128 : ParallelMoveResolverWithSwap(allocator), codegen_(codegen) {} 151 LocationsBuilderX86(HGraph* graph, CodeGeneratorX86* codegen) argument 152 : HGraphVisitor(graph), codegen_(codegen) {} 183 InstructionCodeGeneratorX86(HGraph* graph, CodeGeneratorX86* codegen);
|
H A D | codegen_test.cc | 59 // Provide our own codegen, that ensures the C calling conventions 188 const CodeGenerator& codegen, 191 InstructionSet target_isa = codegen.GetInstructionSet(); 204 static void RunCode(CodeGenerator* codegen, argument 218 SsaLivenessAnalysis liveness(graph, codegen); 222 RegisterAllocator(graph->GetArena(), codegen, liveness).AllocateRegisters(); 226 codegen->Compile(&allocator); 227 Run(allocator, *codegen, has_result, expected); 187 Run(const InternalCodeAllocator& allocator, const CodeGenerator& codegen, bool has_result, Expected expected) argument
|
H A D | graph_visualizer.cc | 168 const CodeGenerator& codegen, 175 codegen_(codegen), 771 const CodeGenerator& codegen) 772 : output_(output), graph_(graph), codegen_(codegen) {} 163 HGraphVisualizerPrinter(HGraph* graph, std::ostream& output, const char* pass_name, bool is_after_pass, bool graph_in_bad_state, const CodeGenerator& codegen, const DisassemblyInformation* disasm_info = nullptr) argument 769 HGraphVisualizer(std::ostream* output, HGraph* graph, const CodeGenerator& codegen) argument
|
H A D | code_generator.cc | 166 DisassemblyScope(HInstruction* instruction, const CodeGenerator& codegen) argument 167 : codegen_(codegen), instruction_(instruction), start_offset_(static_cast<size_t>(-1)) { 386 // The transfer between the two will be done at codegen level. 396 // make the transfer at codegen. 1192 void SlowPathCode::SaveLiveRegisters(CodeGenerator* codegen, LocationSummary* locations) { argument 1194 size_t stack_offset = codegen->GetFirstRegisterSlotInSlowPath(); 1196 for (size_t i = 0, e = codegen->GetNumberOfCoreRegisters(); i < e; ++i) { 1197 if (!codegen->IsCoreCalleeSaveRegister(i)) { 1203 DCHECK_LT(stack_offset, codegen->GetFrameSize() - codegen 1223 RestoreLiveRegisters(CodeGenerator* codegen, LocationSummary* locations) argument [all...] |
H A D | optimizing_compiler.cc | 125 CodeGenerator* codegen, 134 visualizer_(visualizer_output, graph, *codegen), 142 codegen->SetDisassemblyInformation(&disasm_info_); 303 CodeGenerator* codegen, 400 CodeGenerator* codegen, 416 codegen, 431 CodeGenerator* codegen, 468 new (arena) x86::PcRelativeFixups(graph, codegen, stats); 483 CodeGenerator* codegen, 490 SsaLivenessAnalysis liveness(graph, codegen); 124 PassObserver(HGraph* graph, CodeGenerator* codegen, std::ostream* visualizer_output, CompilerDriver* compiler_driver) argument 399 MaybeRunInliner(HGraph* graph, CodeGenerator* codegen, CompilerDriver* driver, OptimizingCompilerStats* stats, const DexCompilationUnit& dex_compilation_unit, PassObserver* pass_observer, StackHandleScopeCollection* handles) argument 429 RunArchOptimizations(InstructionSet instruction_set, HGraph* graph, CodeGenerator* codegen, OptimizingCompilerStats* stats, PassObserver* pass_observer) argument 482 AllocateRegisters(HGraph* graph, CodeGenerator* codegen, PassObserver* pass_observer) argument 501 RunOptimizations(HGraph* graph, CodeGenerator* codegen, CompilerDriver* driver, OptimizingCompilerStats* stats, const DexCompilationUnit& dex_compilation_unit, PassObserver* pass_observer, StackHandleScopeCollection* handles) argument 567 EmitAndSortLinkerPatches(CodeGenerator* codegen) argument 580 Emit(ArenaAllocator* arena, CodeVectorAllocator* code_allocator, CodeGenerator* codegen, CompilerDriver* compiler_driver, const DexFile::CodeItem* code_item) const argument 888 std::unique_ptr<CodeGenerator> codegen; local [all...] |
H A D | register_allocator.cc | 41 CodeGenerator* codegen, 44 codegen_(codegen), 64 blocked_core_registers_(codegen->GetBlockedCoreRegisters()), 65 blocked_fp_registers_(codegen->GetBlockedFloatingPointRegisters()), 75 codegen->SetupBlockedRegisters(); 76 physical_core_register_intervals_.resize(codegen->GetNumberOfCoreRegisters(), nullptr); 77 physical_fp_register_intervals_.resize(codegen->GetNumberOfFloatingPointRegisters(), nullptr); 81 reserved_out_slots_ = InstructionSetPointerSize(codegen->GetInstructionSet()) / kVRegSize + 82 codegen->GetGraph()->GetMaximumNumberOfOutVRegs(); 478 const CodeGenerator& codegen, 40 RegisterAllocator(ArenaAllocator* allocator, CodeGenerator* codegen, const SsaLivenessAnalysis& liveness) argument 475 ValidateIntervals(const ArenaVector<LiveInterval*>& intervals, size_t number_of_spill_slots, size_t number_of_out_slots, const CodeGenerator& codegen, ArenaAllocator* allocator, bool processing_core_registers, bool log_fatal_on_failure) argument [all...] |
H A D | code_generator.h | 81 virtual void EmitNativeCode(CodeGenerator* codegen) = 0; 83 virtual void SaveLiveRegisters(CodeGenerator* codegen, LocationSummary* locations); 84 virtual void RestoreLiveRegisters(CodeGenerator* codegen, LocationSummary* locations); 698 SlowPathGenerator(HGraph* graph, CodeGenerator* codegen) argument 700 codegen_(codegen), 790 InstructionCodeGenerator(HGraph* graph, CodeGenerator* codegen) argument 792 deopt_slow_paths_(graph, codegen) {}
|
H A D | intrinsics_arm.cc | 503 CodeGeneratorARM* codegen) { 505 ArmAssembler* assembler = codegen->GetAssembler(); 527 codegen->GenerateArrayLoadWithBakerReadBarrier( 537 codegen->GenerateReadBarrierSlow(invoke, trg_loc, trg_loc, base_loc, 0U, offset_loc); 552 if (is_volatile && !codegen->GetInstructionSetFeatures().HasAtomicLdrdAndStrd()) { 691 CodeGeneratorARM* codegen) { 692 ArmAssembler* assembler = codegen->GetAssembler(); 705 if (is_volatile && !codegen->GetInstructionSetFeatures().HasAtomicLdrdAndStrd()) { 741 codegen->MarkGCCard(temp, card, base, value, value_can_be_null); 832 static void GenCas(LocationSummary* locations, Primitive::Type type, CodeGeneratorARM* codegen) { argument 500 GenUnsafeGet(HInvoke* invoke, Primitive::Type type, bool is_volatile, CodeGeneratorARM* codegen) argument 687 GenUnsafePut(LocationSummary* locations, Primitive::Type type, bool is_volatile, bool is_ordered, CodeGeneratorARM* codegen) argument 1112 GenerateVisitStringIndexOf(HInvoke* invoke, ArmAssembler* assembler, CodeGeneratorARM* codegen, ArenaAllocator* allocator, bool start_at_zero) argument 1659 GenFPToFPCall(HInvoke* invoke, ArmAssembler* assembler, CodeGeneratorARM* codegen, QuickEntrypointEnum entry) argument 1683 GenFPFPToFPCall(HInvoke* invoke, ArmAssembler* assembler, CodeGeneratorARM* codegen, QuickEntrypointEnum entry) argument [all...] |