/art/compiler/optimizing/ |
H A D | prepare_for_register_allocation.h | 47 void VisitDeoptimize(HDeoptimize* deoptimize) OVERRIDE;
|
H A D | prepare_for_register_allocation.cc | 43 void PrepareForRegisterAllocation::VisitDeoptimize(HDeoptimize* deoptimize) { argument 44 if (deoptimize->GuardsAnInput()) { 46 deoptimize->ReplaceWith(deoptimize->GuardedInput()); 47 deoptimize->RemoveGuard();
|
H A D | ssa_liveness_analysis_test.cc | 192 HInstruction* deoptimize = local 194 block->AddInstruction(deoptimize); 199 deoptimize); 201 deoptimize->SetRawEnvironment(deoptimize_env); 211 EXPECT_EQ(20u, deoptimize->GetLifetimePosition());
|
H A D | cha_guard_optimization.cc | 203 // Need a new deoptimize instruction that copies the environment 205 HDeoptimize* deoptimize = new (GetGraph()->GetArena()) HDeoptimize( local 207 pre_header->InsertInstructionBefore(deoptimize, pre_header->GetLastInstruction()); 208 deoptimize->CopyEnvironmentFromWithLoopPhiAdjustment(
|
H A D | inliner.cc | 474 // - the next run, we deoptimize because we miss a type check, but the method 479 // to JIT (e.g. system server). If we deoptimize we will run interpreted code for the 881 HDeoptimize* deoptimize = new (graph_->GetArena()) HDeoptimize( local 889 bb_cursor->InsertInstructionAfter(deoptimize, compare); 890 deoptimize->CopyEnvironmentFrom(invoke_instruction->GetEnvironment()); 892 receiver->ReplaceUsesDominatedBy(deoptimize, deoptimize); 893 deoptimize->SetReferenceTypeInfo(receiver->GetReferenceTypeInfo()); 950 // we deoptimize instead of keeping the original invoke instruction. 951 bool deoptimize local 1167 HDeoptimize* deoptimize = new (graph_->GetArena()) HDeoptimize( local [all...] |
H A D | bounds_check_elimination.cc | 552 // We should never deoptimize from an osr method, otherwise we might wrongly optimize 1192 // if (max_c >= a.length) deoptimize; 1194 // if (base-min_c > base+max_c) deoptimize; 1195 // if (base+max_c >= a.length ) deoptimize; 1422 // if (max_upper >= a.length ) deoptimize; 1424 // if (min_upper > max_upper) deoptimize; unless min_c == max_c 1425 // if (max_upper >= a.length ) deoptimize; 1427 // if (min_lower > max_lower) deoptimize; unless min_c == max_c 1428 // if (max_lower > max_upper) deoptimize; 1429 // if (max_upper >= a.length ) deoptimize; 1627 HDeoptimize* deoptimize = new (GetGraph()->GetArena()) HDeoptimize( local 1640 HDeoptimize* deoptimize = new (GetGraph()->GetArena()) HDeoptimize( local [all...] |
H A D | instruction_simplifier.cc | 104 void VisitDeoptimize(HDeoptimize* deoptimize) OVERRIDE; 2132 void InstructionSimplifierVisitor::VisitDeoptimize(HDeoptimize* deoptimize) { argument 2133 HInstruction* cond = deoptimize->InputAt(0); 2137 if (deoptimize->GuardsAnInput()) { 2138 deoptimize->ReplaceWith(deoptimize->GuardedInput()); 2140 deoptimize->GetBlock()->RemoveInstruction(deoptimize);
|
H A D | graph_visualizer.cc | 508 void VisitDeoptimize(HDeoptimize* deoptimize) OVERRIDE { 509 StartAttributeStream("kind") << deoptimize->GetKind();
|
H A D | code_generator_arm64.cc | 3689 void LocationsBuilderARM64::VisitDeoptimize(HDeoptimize* deoptimize) { argument 3691 LocationSummary(deoptimize, LocationSummary::kCallOnSlowPath); 3696 if (IsBooleanValueOrMaterializedCondition(deoptimize->InputAt(0))) { 3701 void InstructionCodeGeneratorARM64::VisitDeoptimize(HDeoptimize* deoptimize) { argument 3703 deopt_slow_paths_.NewSlowPath<DeoptimizationSlowPathARM64>(deoptimize); 3704 GenerateTestAndBranch(deoptimize,
|
H A D | code_generator_mips64.cc | 3911 void LocationsBuilderMIPS64::VisitDeoptimize(HDeoptimize* deoptimize) { 3913 LocationSummary(deoptimize, LocationSummary::kCallOnSlowPath); 3918 if (IsBooleanValueOrMaterializedCondition(deoptimize->InputAt(0))) { 3923 void InstructionCodeGeneratorMIPS64::VisitDeoptimize(HDeoptimize* deoptimize) { 3925 deopt_slow_paths_.NewSlowPath<DeoptimizationSlowPathMIPS64>(deoptimize); 3926 GenerateTestAndBranch(deoptimize,
|
H A D | code_generator_arm.cc | 2684 void LocationsBuilderARM::VisitDeoptimize(HDeoptimize* deoptimize) { argument 2686 LocationSummary(deoptimize, LocationSummary::kCallOnSlowPath); 2691 if (IsBooleanValueOrMaterializedCondition(deoptimize->InputAt(0))) { 2696 void InstructionCodeGeneratorARM::VisitDeoptimize(HDeoptimize* deoptimize) { argument 2697 SlowPathCodeARM* slow_path = deopt_slow_paths_.NewSlowPath<DeoptimizationSlowPathARM>(deoptimize); 2698 GenerateTestAndBranch(deoptimize,
|
H A D | code_generator_arm_vixl.cc | 2717 void LocationsBuilderARMVIXL::VisitDeoptimize(HDeoptimize* deoptimize) { argument 2719 LocationSummary(deoptimize, LocationSummary::kCallOnSlowPath); 2724 if (IsBooleanValueOrMaterializedCondition(deoptimize->InputAt(0))) { 2729 void InstructionCodeGeneratorARMVIXL::VisitDeoptimize(HDeoptimize* deoptimize) { argument 2731 deopt_slow_paths_.NewSlowPath<DeoptimizationSlowPathARMVIXL>(deoptimize); 2732 GenerateTestAndBranch(deoptimize,
|
H A D | code_generator_mips.cc | 1317 // Initialize should deoptimize flag to 0. 5160 void LocationsBuilderMIPS::VisitDeoptimize(HDeoptimize* deoptimize) { 5162 LocationSummary(deoptimize, LocationSummary::kCallOnSlowPath); 5167 if (IsBooleanValueOrMaterializedCondition(deoptimize->InputAt(0))) { 5172 void InstructionCodeGeneratorMIPS::VisitDeoptimize(HDeoptimize* deoptimize) { 5174 deopt_slow_paths_.NewSlowPath<DeoptimizationSlowPathMIPS>(deoptimize); 5175 GenerateTestAndBranch(deoptimize,
|
H A D | code_generator_x86.cc | 1694 void LocationsBuilderX86::VisitDeoptimize(HDeoptimize* deoptimize) { argument 1696 LocationSummary(deoptimize, LocationSummary::kCallOnSlowPath); 1701 if (IsBooleanValueOrMaterializedCondition(deoptimize->InputAt(0))) { 1706 void InstructionCodeGeneratorX86::VisitDeoptimize(HDeoptimize* deoptimize) { argument 1707 SlowPathCode* slow_path = deopt_slow_paths_.NewSlowPath<DeoptimizationSlowPathX86>(deoptimize); 1708 GenerateTestAndBranch<Label>(deoptimize,
|
H A D | code_generator_x86_64.cc | 1716 void LocationsBuilderX86_64::VisitDeoptimize(HDeoptimize* deoptimize) { argument 1718 LocationSummary(deoptimize, LocationSummary::kCallOnSlowPath); 1723 if (IsBooleanValueOrMaterializedCondition(deoptimize->InputAt(0))) { 1728 void InstructionCodeGeneratorX86_64::VisitDeoptimize(HDeoptimize* deoptimize) { argument 1729 SlowPathCode* slow_path = deopt_slow_paths_.NewSlowPath<DeoptimizationSlowPathX86_64>(deoptimize); 1730 GenerateTestAndBranch<Label>(deoptimize,
|
/art/runtime/ |
H A D | instrumentation.cc | 812 // Restore code and possibly stack only if we did not deoptimize everything. 1106 bool deoptimize = (visitor.caller != nullptr) && local 1109 if (deoptimize && Runtime::Current()->IsAsyncDeoptimizeable(*return_pc)) { 1127 if (deoptimize && !Runtime::Current()->IsAsyncDeoptimizeable(*return_pc)) {
|