compiler_driver.cc revision acab8d485bd9c007b991f482ed01264d527b63d6
1/* 2 * Copyright (C) 2011 The Android Open Source Project 3 * 4 * Licensed under the Apache License, Version 2.0 (the "License"); 5 * you may not use this file except in compliance with the License. 6 * You may obtain a copy of the License at 7 * 8 * http://www.apache.org/licenses/LICENSE-2.0 9 * 10 * Unless required by applicable law or agreed to in writing, software 11 * distributed under the License is distributed on an "AS IS" BASIS, 12 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 13 * See the License for the specific language governing permissions and 14 * limitations under the License. 15 */ 16 17#include "compiler_driver.h" 18 19#include <unordered_set> 20#include <vector> 21#include <unistd.h> 22 23#ifndef __APPLE__ 24#include <malloc.h> // For mallinfo 25#endif 26 27#include "android-base/strings.h" 28 29#include "art_field-inl.h" 30#include "art_method-inl.h" 31#include "base/array_ref.h" 32#include "base/bit_vector.h" 33#include "base/enums.h" 34#include "base/stl_util.h" 35#include "base/systrace.h" 36#include "base/time_utils.h" 37#include "base/timing_logger.h" 38#include "class_linker-inl.h" 39#include "compiled_class.h" 40#include "compiled_method.h" 41#include "compiler.h" 42#include "compiler_callbacks.h" 43#include "compiler_driver-inl.h" 44#include "dex_compilation_unit.h" 45#include "dex_file-inl.h" 46#include "dex_instruction-inl.h" 47#include "dex/dex_to_dex_compiler.h" 48#include "dex/dex_to_dex_decompiler.h" 49#include "dex/verification_results.h" 50#include "dex/verified_method.h" 51#include "driver/compiler_options.h" 52#include "intrinsics_enum.h" 53#include "jni_internal.h" 54#include "object_lock.h" 55#include "runtime.h" 56#include "gc/accounting/card_table-inl.h" 57#include "gc/accounting/heap_bitmap.h" 58#include "gc/space/image_space.h" 59#include "gc/space/space.h" 60#include "mirror/class_loader.h" 61#include "mirror/class-inl.h" 62#include "mirror/dex_cache-inl.h" 63#include "mirror/object-inl.h" 64#include "mirror/object_array-inl.h" 65#include "mirror/throwable.h" 66#include "scoped_thread_state_change-inl.h" 67#include "ScopedLocalRef.h" 68#include "handle_scope-inl.h" 69#include "thread.h" 70#include "thread_list.h" 71#include "thread_pool.h" 72#include "trampolines/trampoline_compiler.h" 73#include "transaction.h" 74#include "utils/atomic_method_ref_map-inl.h" 75#include "utils/dex_cache_arrays_layout-inl.h" 76#include "utils/swap_space.h" 77#include "vdex_file.h" 78#include "verifier/method_verifier.h" 79#include "verifier/method_verifier-inl.h" 80#include "verifier/verifier_log_mode.h" 81#include "verifier/verifier_deps.h" 82 83namespace art { 84 85static constexpr bool kTimeCompileMethod = !kIsDebugBuild; 86 87// Print additional info during profile guided compilation. 88static constexpr bool kDebugProfileGuidedCompilation = false; 89 90static double Percentage(size_t x, size_t y) { 91 return 100.0 * (static_cast<double>(x)) / (static_cast<double>(x + y)); 92} 93 94static void DumpStat(size_t x, size_t y, const char* str) { 95 if (x == 0 && y == 0) { 96 return; 97 } 98 LOG(INFO) << Percentage(x, y) << "% of " << str << " for " << (x + y) << " cases"; 99} 100 101class CompilerDriver::AOTCompilationStats { 102 public: 103 AOTCompilationStats() 104 : stats_lock_("AOT compilation statistics lock"), 105 resolved_types_(0), unresolved_types_(0), 106 resolved_instance_fields_(0), unresolved_instance_fields_(0), 107 resolved_local_static_fields_(0), resolved_static_fields_(0), unresolved_static_fields_(0), 108 type_based_devirtualization_(0), 109 safe_casts_(0), not_safe_casts_(0) { 110 for (size_t i = 0; i <= kMaxInvokeType; i++) { 111 resolved_methods_[i] = 0; 112 unresolved_methods_[i] = 0; 113 virtual_made_direct_[i] = 0; 114 direct_calls_to_boot_[i] = 0; 115 direct_methods_to_boot_[i] = 0; 116 } 117 } 118 119 void Dump() { 120 DumpStat(resolved_types_, unresolved_types_, "types resolved"); 121 DumpStat(resolved_instance_fields_, unresolved_instance_fields_, "instance fields resolved"); 122 DumpStat(resolved_local_static_fields_ + resolved_static_fields_, unresolved_static_fields_, 123 "static fields resolved"); 124 DumpStat(resolved_local_static_fields_, resolved_static_fields_ + unresolved_static_fields_, 125 "static fields local to a class"); 126 DumpStat(safe_casts_, not_safe_casts_, "check-casts removed based on type information"); 127 // Note, the code below subtracts the stat value so that when added to the stat value we have 128 // 100% of samples. TODO: clean this up. 129 DumpStat(type_based_devirtualization_, 130 resolved_methods_[kVirtual] + unresolved_methods_[kVirtual] + 131 resolved_methods_[kInterface] + unresolved_methods_[kInterface] - 132 type_based_devirtualization_, 133 "virtual/interface calls made direct based on type information"); 134 135 for (size_t i = 0; i <= kMaxInvokeType; i++) { 136 std::ostringstream oss; 137 oss << static_cast<InvokeType>(i) << " methods were AOT resolved"; 138 DumpStat(resolved_methods_[i], unresolved_methods_[i], oss.str().c_str()); 139 if (virtual_made_direct_[i] > 0) { 140 std::ostringstream oss2; 141 oss2 << static_cast<InvokeType>(i) << " methods made direct"; 142 DumpStat(virtual_made_direct_[i], 143 resolved_methods_[i] + unresolved_methods_[i] - virtual_made_direct_[i], 144 oss2.str().c_str()); 145 } 146 if (direct_calls_to_boot_[i] > 0) { 147 std::ostringstream oss2; 148 oss2 << static_cast<InvokeType>(i) << " method calls are direct into boot"; 149 DumpStat(direct_calls_to_boot_[i], 150 resolved_methods_[i] + unresolved_methods_[i] - direct_calls_to_boot_[i], 151 oss2.str().c_str()); 152 } 153 if (direct_methods_to_boot_[i] > 0) { 154 std::ostringstream oss2; 155 oss2 << static_cast<InvokeType>(i) << " method calls have methods in boot"; 156 DumpStat(direct_methods_to_boot_[i], 157 resolved_methods_[i] + unresolved_methods_[i] - direct_methods_to_boot_[i], 158 oss2.str().c_str()); 159 } 160 } 161 } 162 163// Allow lossy statistics in non-debug builds. 164#ifndef NDEBUG 165#define STATS_LOCK() MutexLock mu(Thread::Current(), stats_lock_) 166#else 167#define STATS_LOCK() 168#endif 169 170 void TypeDoesntNeedAccessCheck() REQUIRES(!stats_lock_) { 171 STATS_LOCK(); 172 resolved_types_++; 173 } 174 175 void TypeNeedsAccessCheck() REQUIRES(!stats_lock_) { 176 STATS_LOCK(); 177 unresolved_types_++; 178 } 179 180 void ResolvedInstanceField() REQUIRES(!stats_lock_) { 181 STATS_LOCK(); 182 resolved_instance_fields_++; 183 } 184 185 void UnresolvedInstanceField() REQUIRES(!stats_lock_) { 186 STATS_LOCK(); 187 unresolved_instance_fields_++; 188 } 189 190 void ResolvedLocalStaticField() REQUIRES(!stats_lock_) { 191 STATS_LOCK(); 192 resolved_local_static_fields_++; 193 } 194 195 void ResolvedStaticField() REQUIRES(!stats_lock_) { 196 STATS_LOCK(); 197 resolved_static_fields_++; 198 } 199 200 void UnresolvedStaticField() REQUIRES(!stats_lock_) { 201 STATS_LOCK(); 202 unresolved_static_fields_++; 203 } 204 205 // Indicate that type information from the verifier led to devirtualization. 206 void PreciseTypeDevirtualization() REQUIRES(!stats_lock_) { 207 STATS_LOCK(); 208 type_based_devirtualization_++; 209 } 210 211 // A check-cast could be eliminated due to verifier type analysis. 212 void SafeCast() REQUIRES(!stats_lock_) { 213 STATS_LOCK(); 214 safe_casts_++; 215 } 216 217 // A check-cast couldn't be eliminated due to verifier type analysis. 218 void NotASafeCast() REQUIRES(!stats_lock_) { 219 STATS_LOCK(); 220 not_safe_casts_++; 221 } 222 223 private: 224 Mutex stats_lock_; 225 226 size_t resolved_types_; 227 size_t unresolved_types_; 228 229 size_t resolved_instance_fields_; 230 size_t unresolved_instance_fields_; 231 232 size_t resolved_local_static_fields_; 233 size_t resolved_static_fields_; 234 size_t unresolved_static_fields_; 235 // Type based devirtualization for invoke interface and virtual. 236 size_t type_based_devirtualization_; 237 238 size_t resolved_methods_[kMaxInvokeType + 1]; 239 size_t unresolved_methods_[kMaxInvokeType + 1]; 240 size_t virtual_made_direct_[kMaxInvokeType + 1]; 241 size_t direct_calls_to_boot_[kMaxInvokeType + 1]; 242 size_t direct_methods_to_boot_[kMaxInvokeType + 1]; 243 244 size_t safe_casts_; 245 size_t not_safe_casts_; 246 247 DISALLOW_COPY_AND_ASSIGN(AOTCompilationStats); 248}; 249 250class CompilerDriver::DexFileMethodSet { 251 public: 252 explicit DexFileMethodSet(const DexFile& dex_file) 253 : dex_file_(dex_file), 254 method_indexes_(dex_file.NumMethodIds(), false, Allocator::GetMallocAllocator()) { 255 } 256 DexFileMethodSet(DexFileMethodSet&& other) = default; 257 258 const DexFile& GetDexFile() const { return dex_file_; } 259 260 BitVector& GetMethodIndexes() { return method_indexes_; } 261 const BitVector& GetMethodIndexes() const { return method_indexes_; } 262 263 private: 264 const DexFile& dex_file_; 265 BitVector method_indexes_; 266}; 267 268CompilerDriver::CompilerDriver( 269 const CompilerOptions* compiler_options, 270 VerificationResults* verification_results, 271 Compiler::Kind compiler_kind, 272 InstructionSet instruction_set, 273 const InstructionSetFeatures* instruction_set_features, 274 std::unordered_set<std::string>* image_classes, 275 std::unordered_set<std::string>* compiled_classes, 276 std::unordered_set<std::string>* compiled_methods, 277 size_t thread_count, 278 bool dump_stats, 279 bool dump_passes, 280 CumulativeLogger* timer, 281 int swap_fd, 282 const ProfileCompilationInfo* profile_compilation_info) 283 : compiler_options_(compiler_options), 284 verification_results_(verification_results), 285 compiler_(Compiler::Create(this, compiler_kind)), 286 compiler_kind_(compiler_kind), 287 instruction_set_(instruction_set == kArm ? kThumb2: instruction_set), 288 instruction_set_features_(instruction_set_features), 289 requires_constructor_barrier_lock_("constructor barrier lock"), 290 compiled_classes_lock_("compiled classes lock"), 291 non_relative_linker_patch_count_(0u), 292 image_classes_(image_classes), 293 classes_to_compile_(compiled_classes), 294 methods_to_compile_(compiled_methods), 295 had_hard_verifier_failure_(false), 296 parallel_thread_count_(thread_count), 297 stats_(new AOTCompilationStats), 298 dump_stats_(dump_stats), 299 dump_passes_(dump_passes), 300 timings_logger_(timer), 301 compiler_context_(nullptr), 302 support_boot_image_fixup_(instruction_set != kMips64), 303 dex_files_for_oat_file_(nullptr), 304 compiled_method_storage_(swap_fd), 305 profile_compilation_info_(profile_compilation_info), 306 max_arena_alloc_(0), 307 dex_to_dex_references_lock_("dex-to-dex references lock"), 308 dex_to_dex_references_(), 309 current_dex_to_dex_methods_(nullptr) { 310 DCHECK(compiler_options_ != nullptr); 311 312 compiler_->Init(); 313 314 if (compiler_options->VerifyOnlyProfile()) { 315 CHECK(profile_compilation_info_ != nullptr) << "Requires profile"; 316 } 317 if (GetCompilerOptions().IsBootImage()) { 318 CHECK(image_classes_.get() != nullptr) << "Expected image classes for boot image"; 319 } 320} 321 322CompilerDriver::~CompilerDriver() { 323 Thread* self = Thread::Current(); 324 { 325 MutexLock mu(self, compiled_classes_lock_); 326 STLDeleteValues(&compiled_classes_); 327 } 328 compiled_methods_.Visit([this](const MethodReference& ref ATTRIBUTE_UNUSED, 329 CompiledMethod* method) { 330 if (method != nullptr) { 331 CompiledMethod::ReleaseSwapAllocatedCompiledMethod(this, method); 332 } 333 }); 334 compiler_->UnInit(); 335} 336 337 338#define CREATE_TRAMPOLINE(type, abi, offset) \ 339 if (Is64BitInstructionSet(instruction_set_)) { \ 340 return CreateTrampoline64(instruction_set_, abi, \ 341 type ## _ENTRYPOINT_OFFSET(PointerSize::k64, offset)); \ 342 } else { \ 343 return CreateTrampoline32(instruction_set_, abi, \ 344 type ## _ENTRYPOINT_OFFSET(PointerSize::k32, offset)); \ 345 } 346 347std::unique_ptr<const std::vector<uint8_t>> CompilerDriver::CreateJniDlsymLookup() const { 348 CREATE_TRAMPOLINE(JNI, kJniAbi, pDlsymLookup) 349} 350 351std::unique_ptr<const std::vector<uint8_t>> CompilerDriver::CreateQuickGenericJniTrampoline() 352 const { 353 CREATE_TRAMPOLINE(QUICK, kQuickAbi, pQuickGenericJniTrampoline) 354} 355 356std::unique_ptr<const std::vector<uint8_t>> CompilerDriver::CreateQuickImtConflictTrampoline() 357 const { 358 CREATE_TRAMPOLINE(QUICK, kQuickAbi, pQuickImtConflictTrampoline) 359} 360 361std::unique_ptr<const std::vector<uint8_t>> CompilerDriver::CreateQuickResolutionTrampoline() 362 const { 363 CREATE_TRAMPOLINE(QUICK, kQuickAbi, pQuickResolutionTrampoline) 364} 365 366std::unique_ptr<const std::vector<uint8_t>> CompilerDriver::CreateQuickToInterpreterBridge() 367 const { 368 CREATE_TRAMPOLINE(QUICK, kQuickAbi, pQuickToInterpreterBridge) 369} 370#undef CREATE_TRAMPOLINE 371 372static void SetupIntrinsic(Thread* self, 373 Intrinsics intrinsic, 374 InvokeType invoke_type, 375 const char* class_name, 376 const char* method_name, 377 const char* signature) 378 REQUIRES_SHARED(Locks::mutator_lock_) { 379 ClassLinker* class_linker = Runtime::Current()->GetClassLinker(); 380 PointerSize image_size = class_linker->GetImagePointerSize(); 381 mirror::Class* cls = class_linker->FindSystemClass(self, class_name); 382 if (cls == nullptr) { 383 LOG(FATAL) << "Could not find class of intrinsic " << class_name; 384 } 385 ArtMethod* method = (invoke_type == kStatic || invoke_type == kDirect) 386 ? cls->FindDeclaredDirectMethod(method_name, signature, image_size) 387 : cls->FindDeclaredVirtualMethod(method_name, signature, image_size); 388 if (method == nullptr) { 389 LOG(FATAL) << "Could not find method of intrinsic " 390 << class_name << " " << method_name << " " << signature; 391 } 392 DCHECK_EQ(method->GetInvokeType(), invoke_type); 393 method->SetIntrinsic(static_cast<uint32_t>(intrinsic)); 394} 395 396void CompilerDriver::CompileAll(jobject class_loader, 397 const std::vector<const DexFile*>& dex_files, 398 TimingLogger* timings) { 399 DCHECK(!Runtime::Current()->IsStarted()); 400 401 InitializeThreadPools(); 402 403 VLOG(compiler) << "Before precompile " << GetMemoryUsageString(false); 404 // Precompile: 405 // 1) Load image classes 406 // 2) Resolve all classes 407 // 3) Attempt to verify all classes 408 // 4) Attempt to initialize image classes, and trivially initialized classes 409 PreCompile(class_loader, dex_files, timings); 410 if (GetCompilerOptions().IsBootImage()) { 411 // We don't need to setup the intrinsics for non boot image compilation, as 412 // those compilations will pick up a boot image that have the ArtMethod already 413 // set with the intrinsics flag. 414 ScopedObjectAccess soa(Thread::Current()); 415#define SETUP_INTRINSICS(Name, InvokeType, NeedsEnvironmentOrCache, SideEffects, Exceptions, \ 416 ClassName, MethodName, Signature) \ 417 SetupIntrinsic(soa.Self(), Intrinsics::k##Name, InvokeType, ClassName, MethodName, Signature); 418#include "intrinsics_list.h" 419INTRINSICS_LIST(SETUP_INTRINSICS) 420#undef INTRINSICS_LIST 421#undef SETUP_INTRINSICS 422 } 423 // Compile: 424 // 1) Compile all classes and methods enabled for compilation. May fall back to dex-to-dex 425 // compilation. 426 if (!GetCompilerOptions().VerifyAtRuntime()) { 427 Compile(class_loader, dex_files, timings); 428 } 429 if (dump_stats_) { 430 stats_->Dump(); 431 } 432 433 FreeThreadPools(); 434} 435 436// In-place unquicken the given `dex_files` based on `quickening_info`. 437static void Unquicken(const std::vector<const DexFile*>& dex_files, 438 const ArrayRef<const uint8_t>& quickening_info) { 439 const uint8_t* quickening_info_ptr = quickening_info.data(); 440 const uint8_t* const quickening_info_end = quickening_info.data() + quickening_info.size(); 441 for (const DexFile* dex_file : dex_files) { 442 for (uint32_t i = 0; i < dex_file->NumClassDefs(); ++i) { 443 const DexFile::ClassDef& class_def = dex_file->GetClassDef(i); 444 const uint8_t* class_data = dex_file->GetClassData(class_def); 445 if (class_data == nullptr) { 446 continue; 447 } 448 ClassDataItemIterator it(*dex_file, class_data); 449 // Skip fields 450 while (it.HasNextStaticField()) { 451 it.Next(); 452 } 453 while (it.HasNextInstanceField()) { 454 it.Next(); 455 } 456 457 // Unquicken each method. 458 while (it.HasNextDirectMethod()) { 459 const DexFile::CodeItem* code_item = it.GetMethodCodeItem(); 460 if (code_item != nullptr) { 461 uint32_t quickening_size = *reinterpret_cast<const uint32_t*>(quickening_info_ptr); 462 quickening_info_ptr += sizeof(uint32_t); 463 optimizer::ArtDecompileDEX( 464 *code_item, ArrayRef<const uint8_t>(quickening_info_ptr, quickening_size)); 465 quickening_info_ptr += quickening_size; 466 } 467 it.Next(); 468 } 469 470 while (it.HasNextVirtualMethod()) { 471 const DexFile::CodeItem* code_item = it.GetMethodCodeItem(); 472 if (code_item != nullptr) { 473 uint32_t quickening_size = *reinterpret_cast<const uint32_t*>(quickening_info_ptr); 474 quickening_info_ptr += sizeof(uint32_t); 475 optimizer::ArtDecompileDEX( 476 *code_item, ArrayRef<const uint8_t>(quickening_info_ptr, quickening_size)); 477 quickening_info_ptr += quickening_size; 478 } 479 it.Next(); 480 } 481 DCHECK(!it.HasNext()); 482 } 483 } 484 DCHECK_EQ(quickening_info_ptr, quickening_info_end) << "Failed to use all quickening info"; 485} 486 487void CompilerDriver::CompileAll(jobject class_loader, 488 const std::vector<const DexFile*>& dex_files, 489 VdexFile* vdex_file, 490 TimingLogger* timings) { 491 if (vdex_file != nullptr) { 492 // TODO: we unquicken unconditionnally, as we don't know 493 // if the boot image has changed. How exactly we'll know is under 494 // experimentation. 495 Unquicken(dex_files, vdex_file->GetQuickeningInfo()); 496 Runtime::Current()->GetCompilerCallbacks()->SetVerifierDeps( 497 new verifier::VerifierDeps(dex_files, vdex_file->GetVerifierDepsData())); 498 } 499 CompileAll(class_loader, dex_files, timings); 500} 501 502static optimizer::DexToDexCompilationLevel GetDexToDexCompilationLevel( 503 Thread* self, const CompilerDriver& driver, Handle<mirror::ClassLoader> class_loader, 504 const DexFile& dex_file, const DexFile::ClassDef& class_def) 505 REQUIRES_SHARED(Locks::mutator_lock_) { 506 auto* const runtime = Runtime::Current(); 507 if (runtime->UseJitCompilation() || driver.GetCompilerOptions().VerifyAtRuntime()) { 508 // Verify at runtime shouldn't dex to dex since we didn't resolve of verify. 509 return optimizer::DexToDexCompilationLevel::kDontDexToDexCompile; 510 } 511 const char* descriptor = dex_file.GetClassDescriptor(class_def); 512 ClassLinker* class_linker = runtime->GetClassLinker(); 513 mirror::Class* klass = class_linker->FindClass(self, descriptor, class_loader); 514 if (klass == nullptr) { 515 CHECK(self->IsExceptionPending()); 516 self->ClearException(); 517 return optimizer::DexToDexCompilationLevel::kDontDexToDexCompile; 518 } 519 // DexToDex at the kOptimize level may introduce quickened opcodes, which replace symbolic 520 // references with actual offsets. We cannot re-verify such instructions. 521 // 522 // We store the verification information in the class status in the oat file, which the linker 523 // can validate (checksums) and use to skip load-time verification. It is thus safe to 524 // optimize when a class has been fully verified before. 525 if (klass->IsVerified()) { 526 // Class is verified so we can enable DEX-to-DEX compilation for performance. 527 return optimizer::DexToDexCompilationLevel::kOptimize; 528 } else if (klass->IsCompileTimeVerified()) { 529 // Class verification has soft-failed. Anyway, ensure at least correctness. 530 DCHECK_EQ(klass->GetStatus(), mirror::Class::kStatusRetryVerificationAtRuntime); 531 return optimizer::DexToDexCompilationLevel::kRequired; 532 } else { 533 // Class verification has failed: do not run DEX-to-DEX compilation. 534 return optimizer::DexToDexCompilationLevel::kDontDexToDexCompile; 535 } 536} 537 538static optimizer::DexToDexCompilationLevel GetDexToDexCompilationLevel( 539 Thread* self, 540 const CompilerDriver& driver, 541 jobject jclass_loader, 542 const DexFile& dex_file, 543 const DexFile::ClassDef& class_def) { 544 ScopedObjectAccess soa(self); 545 StackHandleScope<1> hs(soa.Self()); 546 Handle<mirror::ClassLoader> class_loader( 547 hs.NewHandle(soa.Decode<mirror::ClassLoader>(jclass_loader))); 548 return GetDexToDexCompilationLevel(self, driver, class_loader, dex_file, class_def); 549} 550 551// Does the runtime for the InstructionSet provide an implementation returned by 552// GetQuickGenericJniStub allowing down calls that aren't compiled using a JNI compiler? 553static bool InstructionSetHasGenericJniStub(InstructionSet isa) { 554 switch (isa) { 555 case kArm: 556 case kArm64: 557 case kThumb2: 558 case kMips: 559 case kMips64: 560 case kX86: 561 case kX86_64: return true; 562 default: return false; 563 } 564} 565 566static void CompileMethod(Thread* self, 567 CompilerDriver* driver, 568 const DexFile::CodeItem* code_item, 569 uint32_t access_flags, 570 InvokeType invoke_type, 571 uint16_t class_def_idx, 572 uint32_t method_idx, 573 jobject class_loader, 574 const DexFile& dex_file, 575 optimizer::DexToDexCompilationLevel dex_to_dex_compilation_level, 576 bool compilation_enabled, 577 Handle<mirror::DexCache> dex_cache) { 578 DCHECK(driver != nullptr); 579 CompiledMethod* compiled_method = nullptr; 580 uint64_t start_ns = kTimeCompileMethod ? NanoTime() : 0; 581 MethodReference method_ref(&dex_file, method_idx); 582 583 if (driver->GetCurrentDexToDexMethods() != nullptr) { 584 // This is the second pass when we dex-to-dex compile previously marked methods. 585 // TODO: Refactor the compilation to avoid having to distinguish the two passes 586 // here. That should be done on a higher level. http://b/29089975 587 if (driver->GetCurrentDexToDexMethods()->IsBitSet(method_idx)) { 588 const VerifiedMethod* verified_method = 589 driver->GetVerificationResults()->GetVerifiedMethod(method_ref); 590 // Do not optimize if a VerifiedMethod is missing. SafeCast elision, 591 // for example, relies on it. 592 compiled_method = optimizer::ArtCompileDEX( 593 driver, 594 code_item, 595 access_flags, 596 invoke_type, 597 class_def_idx, 598 method_idx, 599 class_loader, 600 dex_file, 601 (verified_method != nullptr) 602 ? dex_to_dex_compilation_level 603 : optimizer::DexToDexCompilationLevel::kRequired); 604 } 605 } else if ((access_flags & kAccNative) != 0) { 606 // Are we extracting only and have support for generic JNI down calls? 607 if (!driver->GetCompilerOptions().IsJniCompilationEnabled() && 608 InstructionSetHasGenericJniStub(driver->GetInstructionSet())) { 609 // Leaving this empty will trigger the generic JNI version 610 } else { 611 // Look-up the ArtMethod associated with this code_item (if any) 612 // -- It is later used to lookup any [optimization] annotations for this method. 613 ScopedObjectAccess soa(self); 614 StackHandleScope<1> hs(soa.Self()); 615 Handle<mirror::ClassLoader> class_loader_handle(hs.NewHandle( 616 soa.Decode<mirror::ClassLoader>(class_loader))); 617 618 // TODO: Lookup annotation from DexFile directly without resolving method. 619 ArtMethod* method = 620 Runtime::Current()->GetClassLinker()->ResolveMethod<ClassLinker::kNoICCECheckForCache>( 621 dex_file, 622 method_idx, 623 dex_cache, 624 class_loader_handle, 625 /* referrer */ nullptr, 626 invoke_type); 627 628 // Query any JNI optimization annotations such as @FastNative or @CriticalNative. 629 Compiler::JniOptimizationFlags optimization_flags = Compiler::kNone; 630 if (UNLIKELY(method == nullptr)) { 631 // Failed method resolutions happen very rarely, e.g. ancestor class cannot be resolved. 632 DCHECK(self->IsExceptionPending()); 633 self->ClearException(); 634 } else if (method->IsAnnotatedWithFastNative()) { 635 // TODO: Will no longer need this CHECK once we have verifier checking this. 636 CHECK(!method->IsAnnotatedWithCriticalNative()); 637 optimization_flags = Compiler::kFastNative; 638 } else if (method->IsAnnotatedWithCriticalNative()) { 639 // TODO: Will no longer need this CHECK once we have verifier checking this. 640 CHECK(!method->IsAnnotatedWithFastNative()); 641 optimization_flags = Compiler::kCriticalNative; 642 } 643 644 compiled_method = driver->GetCompiler()->JniCompile(access_flags, 645 method_idx, 646 dex_file, 647 optimization_flags); 648 CHECK(compiled_method != nullptr); 649 } 650 } else if ((access_flags & kAccAbstract) != 0) { 651 // Abstract methods don't have code. 652 } else { 653 const VerifiedMethod* verified_method = 654 driver->GetVerificationResults()->GetVerifiedMethod(method_ref); 655 bool compile = compilation_enabled && 656 // Basic checks, e.g., not <clinit>. 657 driver->GetVerificationResults() 658 ->IsCandidateForCompilation(method_ref, access_flags) && 659 // Did not fail to create VerifiedMethod metadata. 660 verified_method != nullptr && 661 // Do not have failures that should punt to the interpreter. 662 !verified_method->HasRuntimeThrow() && 663 (verified_method->GetEncounteredVerificationFailures() & 664 (verifier::VERIFY_ERROR_FORCE_INTERPRETER | verifier::VERIFY_ERROR_LOCKING)) == 0 && 665 // Is eligable for compilation by methods-to-compile filter. 666 driver->IsMethodToCompile(method_ref) && 667 driver->ShouldCompileBasedOnProfile(method_ref); 668 669 if (compile) { 670 // NOTE: if compiler declines to compile this method, it will return null. 671 compiled_method = driver->GetCompiler()->Compile(code_item, access_flags, invoke_type, 672 class_def_idx, method_idx, class_loader, 673 dex_file, dex_cache); 674 } 675 if (compiled_method == nullptr && 676 dex_to_dex_compilation_level != optimizer::DexToDexCompilationLevel::kDontDexToDexCompile) { 677 DCHECK(!Runtime::Current()->UseJitCompilation()); 678 // TODO: add a command-line option to disable DEX-to-DEX compilation ? 679 driver->MarkForDexToDexCompilation(self, method_ref); 680 } 681 } 682 if (kTimeCompileMethod) { 683 uint64_t duration_ns = NanoTime() - start_ns; 684 if (duration_ns > MsToNs(driver->GetCompiler()->GetMaximumCompilationTimeBeforeWarning())) { 685 LOG(WARNING) << "Compilation of " << dex_file.PrettyMethod(method_idx) 686 << " took " << PrettyDuration(duration_ns); 687 } 688 } 689 690 if (compiled_method != nullptr) { 691 // Count non-relative linker patches. 692 size_t non_relative_linker_patch_count = 0u; 693 for (const LinkerPatch& patch : compiled_method->GetPatches()) { 694 if (!patch.IsPcRelative()) { 695 ++non_relative_linker_patch_count; 696 } 697 } 698 bool compile_pic = driver->GetCompilerOptions().GetCompilePic(); // Off by default 699 // When compiling with PIC, there should be zero non-relative linker patches 700 CHECK(!compile_pic || non_relative_linker_patch_count == 0u); 701 702 driver->AddCompiledMethod(method_ref, compiled_method, non_relative_linker_patch_count); 703 } 704 705 if (self->IsExceptionPending()) { 706 ScopedObjectAccess soa(self); 707 LOG(FATAL) << "Unexpected exception compiling: " << dex_file.PrettyMethod(method_idx) << "\n" 708 << self->GetException()->Dump(); 709 } 710} 711 712void CompilerDriver::CompileOne(Thread* self, ArtMethod* method, TimingLogger* timings) { 713 DCHECK(!Runtime::Current()->IsStarted()); 714 jobject jclass_loader; 715 const DexFile* dex_file; 716 uint16_t class_def_idx; 717 uint32_t method_idx = method->GetDexMethodIndex(); 718 uint32_t access_flags = method->GetAccessFlags(); 719 InvokeType invoke_type = method->GetInvokeType(); 720 StackHandleScope<1> hs(self); 721 Handle<mirror::DexCache> dex_cache(hs.NewHandle(method->GetDexCache())); 722 { 723 ScopedObjectAccessUnchecked soa(self); 724 ScopedLocalRef<jobject> local_class_loader( 725 soa.Env(), soa.AddLocalReference<jobject>(method->GetDeclaringClass()->GetClassLoader())); 726 jclass_loader = soa.Env()->NewGlobalRef(local_class_loader.get()); 727 // Find the dex_file 728 dex_file = method->GetDexFile(); 729 class_def_idx = method->GetClassDefIndex(); 730 } 731 const DexFile::CodeItem* code_item = dex_file->GetCodeItem(method->GetCodeItemOffset()); 732 733 // Go to native so that we don't block GC during compilation. 734 ScopedThreadSuspension sts(self, kNative); 735 736 std::vector<const DexFile*> dex_files; 737 dex_files.push_back(dex_file); 738 739 InitializeThreadPools(); 740 741 PreCompile(jclass_loader, dex_files, timings); 742 743 // Can we run DEX-to-DEX compiler on this class ? 744 optimizer::DexToDexCompilationLevel dex_to_dex_compilation_level = 745 GetDexToDexCompilationLevel(self, 746 *this, 747 jclass_loader, 748 *dex_file, 749 dex_file->GetClassDef(class_def_idx)); 750 751 DCHECK(current_dex_to_dex_methods_ == nullptr); 752 CompileMethod(self, 753 this, 754 code_item, 755 access_flags, 756 invoke_type, 757 class_def_idx, 758 method_idx, 759 jclass_loader, 760 *dex_file, 761 dex_to_dex_compilation_level, 762 true, 763 dex_cache); 764 765 ArrayRef<DexFileMethodSet> dex_to_dex_references; 766 { 767 // From this point on, we shall not modify dex_to_dex_references_, so 768 // just grab a reference to it that we use without holding the mutex. 769 MutexLock lock(Thread::Current(), dex_to_dex_references_lock_); 770 dex_to_dex_references = ArrayRef<DexFileMethodSet>(dex_to_dex_references_); 771 } 772 if (!dex_to_dex_references.empty()) { 773 DCHECK_EQ(dex_to_dex_references.size(), 1u); 774 DCHECK(&dex_to_dex_references[0].GetDexFile() == dex_file); 775 current_dex_to_dex_methods_ = &dex_to_dex_references.front().GetMethodIndexes(); 776 DCHECK(current_dex_to_dex_methods_->IsBitSet(method_idx)); 777 DCHECK_EQ(current_dex_to_dex_methods_->NumSetBits(), 1u); 778 CompileMethod(self, 779 this, 780 code_item, 781 access_flags, 782 invoke_type, 783 class_def_idx, 784 method_idx, 785 jclass_loader, 786 *dex_file, 787 dex_to_dex_compilation_level, 788 true, 789 dex_cache); 790 current_dex_to_dex_methods_ = nullptr; 791 } 792 793 FreeThreadPools(); 794 795 self->GetJniEnv()->DeleteGlobalRef(jclass_loader); 796} 797 798void CompilerDriver::Resolve(jobject class_loader, 799 const std::vector<const DexFile*>& dex_files, 800 TimingLogger* timings) { 801 // Resolution allocates classes and needs to run single-threaded to be deterministic. 802 bool force_determinism = GetCompilerOptions().IsForceDeterminism(); 803 ThreadPool* resolve_thread_pool = force_determinism 804 ? single_thread_pool_.get() 805 : parallel_thread_pool_.get(); 806 size_t resolve_thread_count = force_determinism ? 1U : parallel_thread_count_; 807 808 for (size_t i = 0; i != dex_files.size(); ++i) { 809 const DexFile* dex_file = dex_files[i]; 810 CHECK(dex_file != nullptr); 811 ResolveDexFile(class_loader, 812 *dex_file, 813 dex_files, 814 resolve_thread_pool, 815 resolve_thread_count, 816 timings); 817 } 818} 819 820// Resolve const-strings in the code. Done to have deterministic allocation behavior. Right now 821// this is single-threaded for simplicity. 822// TODO: Collect the relevant string indices in parallel, then allocate them sequentially in a 823// stable order. 824 825static void ResolveConstStrings(Handle<mirror::DexCache> dex_cache, 826 const DexFile& dex_file, 827 const DexFile::CodeItem* code_item) 828 REQUIRES_SHARED(Locks::mutator_lock_) { 829 if (code_item == nullptr) { 830 // Abstract or native method. 831 return; 832 } 833 834 const uint16_t* code_ptr = code_item->insns_; 835 const uint16_t* code_end = code_item->insns_ + code_item->insns_size_in_code_units_; 836 ClassLinker* const class_linker = Runtime::Current()->GetClassLinker(); 837 838 while (code_ptr < code_end) { 839 const Instruction* inst = Instruction::At(code_ptr); 840 switch (inst->Opcode()) { 841 case Instruction::CONST_STRING: 842 case Instruction::CONST_STRING_JUMBO: { 843 dex::StringIndex string_index((inst->Opcode() == Instruction::CONST_STRING) 844 ? inst->VRegB_21c() 845 : inst->VRegB_31c()); 846 mirror::String* string = class_linker->ResolveString(dex_file, string_index, dex_cache); 847 CHECK(string != nullptr) << "Could not allocate a string when forcing determinism"; 848 break; 849 } 850 851 default: 852 break; 853 } 854 855 code_ptr += inst->SizeInCodeUnits(); 856 } 857} 858 859static void ResolveConstStrings(CompilerDriver* driver, 860 const std::vector<const DexFile*>& dex_files, 861 TimingLogger* timings) { 862 ScopedObjectAccess soa(Thread::Current()); 863 StackHandleScope<1> hs(soa.Self()); 864 ClassLinker* const class_linker = Runtime::Current()->GetClassLinker(); 865 MutableHandle<mirror::DexCache> dex_cache(hs.NewHandle<mirror::DexCache>(nullptr)); 866 867 for (const DexFile* dex_file : dex_files) { 868 dex_cache.Assign(class_linker->FindDexCache(soa.Self(), *dex_file, false)); 869 TimingLogger::ScopedTiming t("Resolve const-string Strings", timings); 870 871 size_t class_def_count = dex_file->NumClassDefs(); 872 for (size_t class_def_index = 0; class_def_index < class_def_count; ++class_def_index) { 873 const DexFile::ClassDef& class_def = dex_file->GetClassDef(class_def_index); 874 875 const uint8_t* class_data = dex_file->GetClassData(class_def); 876 if (class_data == nullptr) { 877 // empty class, probably a marker interface 878 continue; 879 } 880 881 ClassDataItemIterator it(*dex_file, class_data); 882 // Skip fields 883 while (it.HasNextStaticField()) { 884 it.Next(); 885 } 886 while (it.HasNextInstanceField()) { 887 it.Next(); 888 } 889 890 bool compilation_enabled = driver->IsClassToCompile( 891 dex_file->StringByTypeIdx(class_def.class_idx_)); 892 if (!compilation_enabled) { 893 // Compilation is skipped, do not resolve const-string in code of this class. 894 // TODO: Make sure that inlining honors this. 895 continue; 896 } 897 898 // Direct methods. 899 int64_t previous_direct_method_idx = -1; 900 while (it.HasNextDirectMethod()) { 901 uint32_t method_idx = it.GetMemberIndex(); 902 if (method_idx == previous_direct_method_idx) { 903 // smali can create dex files with two encoded_methods sharing the same method_idx 904 // http://code.google.com/p/smali/issues/detail?id=119 905 it.Next(); 906 continue; 907 } 908 previous_direct_method_idx = method_idx; 909 ResolveConstStrings(dex_cache, *dex_file, it.GetMethodCodeItem()); 910 it.Next(); 911 } 912 // Virtual methods. 913 int64_t previous_virtual_method_idx = -1; 914 while (it.HasNextVirtualMethod()) { 915 uint32_t method_idx = it.GetMemberIndex(); 916 if (method_idx == previous_virtual_method_idx) { 917 // smali can create dex files with two encoded_methods sharing the same method_idx 918 // http://code.google.com/p/smali/issues/detail?id=119 919 it.Next(); 920 continue; 921 } 922 previous_virtual_method_idx = method_idx; 923 ResolveConstStrings(dex_cache, *dex_file, it.GetMethodCodeItem()); 924 it.Next(); 925 } 926 DCHECK(!it.HasNext()); 927 } 928 } 929} 930 931inline void CompilerDriver::CheckThreadPools() { 932 DCHECK(parallel_thread_pool_ != nullptr); 933 DCHECK(single_thread_pool_ != nullptr); 934} 935 936void CompilerDriver::PreCompile(jobject class_loader, 937 const std::vector<const DexFile*>& dex_files, 938 TimingLogger* timings) { 939 CheckThreadPools(); 940 941 for (const DexFile* dex_file : dex_files) { 942 // Can be already inserted if the caller is CompileOne. This happens for gtests. 943 if (!compiled_methods_.HaveDexFile(dex_file)) { 944 compiled_methods_.AddDexFile(dex_file); 945 } 946 } 947 948 LoadImageClasses(timings); 949 VLOG(compiler) << "LoadImageClasses: " << GetMemoryUsageString(false); 950 951 const bool verification_enabled = compiler_options_->IsVerificationEnabled(); 952 const bool never_verify = compiler_options_->NeverVerify(); 953 const bool verify_only_profile = compiler_options_->VerifyOnlyProfile(); 954 955 // We need to resolve for never_verify since it needs to run dex to dex to add the 956 // RETURN_VOID_NO_BARRIER. 957 // Let the verifier resolve as needed for the verify_only_profile case. 958 if ((never_verify || verification_enabled) && !verify_only_profile) { 959 Resolve(class_loader, dex_files, timings); 960 VLOG(compiler) << "Resolve: " << GetMemoryUsageString(false); 961 } 962 963 if (never_verify) { 964 VLOG(compiler) << "Verify none mode specified, skipping verification."; 965 SetVerified(class_loader, dex_files, timings); 966 } 967 968 if (!verification_enabled) { 969 return; 970 } 971 972 if (GetCompilerOptions().IsForceDeterminism() && GetCompilerOptions().IsBootImage()) { 973 // Resolve strings from const-string. Do this now to have a deterministic image. 974 ResolveConstStrings(this, dex_files, timings); 975 VLOG(compiler) << "Resolve const-strings: " << GetMemoryUsageString(false); 976 } 977 978 Verify(class_loader, dex_files, timings); 979 VLOG(compiler) << "Verify: " << GetMemoryUsageString(false); 980 981 if (had_hard_verifier_failure_ && GetCompilerOptions().AbortOnHardVerifierFailure()) { 982 LOG(FATAL) << "Had a hard failure verifying all classes, and was asked to abort in such " 983 << "situations. Please check the log."; 984 } 985 986 InitializeClasses(class_loader, dex_files, timings); 987 VLOG(compiler) << "InitializeClasses: " << GetMemoryUsageString(false); 988 989 UpdateImageClasses(timings); 990 VLOG(compiler) << "UpdateImageClasses: " << GetMemoryUsageString(false); 991} 992 993bool CompilerDriver::IsImageClass(const char* descriptor) const { 994 if (image_classes_ != nullptr) { 995 // If we have a set of image classes, use those. 996 return image_classes_->find(descriptor) != image_classes_->end(); 997 } 998 // No set of image classes, assume we include all the classes. 999 // NOTE: Currently only reachable from InitImageMethodVisitor for the app image case. 1000 return !GetCompilerOptions().IsBootImage(); 1001} 1002 1003bool CompilerDriver::IsClassToCompile(const char* descriptor) const { 1004 if (classes_to_compile_ == nullptr) { 1005 return true; 1006 } 1007 return classes_to_compile_->find(descriptor) != classes_to_compile_->end(); 1008} 1009 1010bool CompilerDriver::IsMethodToCompile(const MethodReference& method_ref) const { 1011 if (methods_to_compile_ == nullptr) { 1012 return true; 1013 } 1014 1015 std::string tmp = method_ref.dex_file->PrettyMethod(method_ref.dex_method_index, true); 1016 return methods_to_compile_->find(tmp.c_str()) != methods_to_compile_->end(); 1017} 1018 1019bool CompilerDriver::ShouldCompileBasedOnProfile(const MethodReference& method_ref) const { 1020 if (profile_compilation_info_ == nullptr) { 1021 // If we miss profile information it means that we don't do a profile guided compilation. 1022 // Return true, and let the other filters decide if the method should be compiled. 1023 return true; 1024 } 1025 bool result = profile_compilation_info_->ContainsMethod(method_ref); 1026 1027 if (kDebugProfileGuidedCompilation) { 1028 LOG(INFO) << "[ProfileGuidedCompilation] " 1029 << (result ? "Compiled" : "Skipped") << " method:" 1030 << method_ref.dex_file->PrettyMethod(method_ref.dex_method_index, true); 1031 } 1032 return result; 1033} 1034 1035bool CompilerDriver::ShouldVerifyClassBasedOnProfile(const DexFile& dex_file, 1036 uint16_t class_idx) const { 1037 if (!compiler_options_->VerifyOnlyProfile()) { 1038 // No profile, verify everything. 1039 return true; 1040 } 1041 DCHECK(profile_compilation_info_ != nullptr); 1042 const DexFile::ClassDef& class_def = dex_file.GetClassDef(class_idx); 1043 dex::TypeIndex type_idx = class_def.class_idx_; 1044 bool result = profile_compilation_info_->ContainsClass(dex_file, type_idx); 1045 if (kDebugProfileGuidedCompilation) { 1046 LOG(INFO) << "[ProfileGuidedCompilation] " << (result ? "Verified" : "Skipped") << " method:" 1047 << dex_file.GetClassDescriptor(class_def); 1048 } 1049 return result; 1050} 1051 1052class ResolveCatchBlockExceptionsClassVisitor : public ClassVisitor { 1053 public: 1054 explicit ResolveCatchBlockExceptionsClassVisitor( 1055 std::set<std::pair<dex::TypeIndex, const DexFile*>>& exceptions_to_resolve) 1056 : exceptions_to_resolve_(exceptions_to_resolve) {} 1057 1058 virtual bool operator()(ObjPtr<mirror::Class> c) OVERRIDE REQUIRES_SHARED(Locks::mutator_lock_) { 1059 const auto pointer_size = Runtime::Current()->GetClassLinker()->GetImagePointerSize(); 1060 for (auto& m : c->GetMethods(pointer_size)) { 1061 ResolveExceptionsForMethod(&m, pointer_size); 1062 } 1063 return true; 1064 } 1065 1066 private: 1067 void ResolveExceptionsForMethod(ArtMethod* method_handle, PointerSize pointer_size) 1068 REQUIRES_SHARED(Locks::mutator_lock_) { 1069 const DexFile::CodeItem* code_item = method_handle->GetCodeItem(); 1070 if (code_item == nullptr) { 1071 return; // native or abstract method 1072 } 1073 if (code_item->tries_size_ == 0) { 1074 return; // nothing to process 1075 } 1076 const uint8_t* encoded_catch_handler_list = DexFile::GetCatchHandlerData(*code_item, 0); 1077 size_t num_encoded_catch_handlers = DecodeUnsignedLeb128(&encoded_catch_handler_list); 1078 for (size_t i = 0; i < num_encoded_catch_handlers; i++) { 1079 int32_t encoded_catch_handler_size = DecodeSignedLeb128(&encoded_catch_handler_list); 1080 bool has_catch_all = false; 1081 if (encoded_catch_handler_size <= 0) { 1082 encoded_catch_handler_size = -encoded_catch_handler_size; 1083 has_catch_all = true; 1084 } 1085 for (int32_t j = 0; j < encoded_catch_handler_size; j++) { 1086 dex::TypeIndex encoded_catch_handler_handlers_type_idx = 1087 dex::TypeIndex(DecodeUnsignedLeb128(&encoded_catch_handler_list)); 1088 // Add to set of types to resolve if not already in the dex cache resolved types 1089 if (!method_handle->IsResolvedTypeIdx(encoded_catch_handler_handlers_type_idx, 1090 pointer_size)) { 1091 exceptions_to_resolve_.emplace(encoded_catch_handler_handlers_type_idx, 1092 method_handle->GetDexFile()); 1093 } 1094 // ignore address associated with catch handler 1095 DecodeUnsignedLeb128(&encoded_catch_handler_list); 1096 } 1097 if (has_catch_all) { 1098 // ignore catch all address 1099 DecodeUnsignedLeb128(&encoded_catch_handler_list); 1100 } 1101 } 1102 } 1103 1104 std::set<std::pair<dex::TypeIndex, const DexFile*>>& exceptions_to_resolve_; 1105}; 1106 1107class RecordImageClassesVisitor : public ClassVisitor { 1108 public: 1109 explicit RecordImageClassesVisitor(std::unordered_set<std::string>* image_classes) 1110 : image_classes_(image_classes) {} 1111 1112 bool operator()(ObjPtr<mirror::Class> klass) OVERRIDE REQUIRES_SHARED(Locks::mutator_lock_) { 1113 std::string temp; 1114 image_classes_->insert(klass->GetDescriptor(&temp)); 1115 return true; 1116 } 1117 1118 private: 1119 std::unordered_set<std::string>* const image_classes_; 1120}; 1121 1122// Make a list of descriptors for classes to include in the image 1123void CompilerDriver::LoadImageClasses(TimingLogger* timings) { 1124 CHECK(timings != nullptr); 1125 if (!GetCompilerOptions().IsBootImage()) { 1126 return; 1127 } 1128 1129 TimingLogger::ScopedTiming t("LoadImageClasses", timings); 1130 // Make a first class to load all classes explicitly listed in the file 1131 Thread* self = Thread::Current(); 1132 ScopedObjectAccess soa(self); 1133 ClassLinker* class_linker = Runtime::Current()->GetClassLinker(); 1134 CHECK(image_classes_.get() != nullptr); 1135 for (auto it = image_classes_->begin(), end = image_classes_->end(); it != end;) { 1136 const std::string& descriptor(*it); 1137 StackHandleScope<1> hs(self); 1138 Handle<mirror::Class> klass( 1139 hs.NewHandle(class_linker->FindSystemClass(self, descriptor.c_str()))); 1140 if (klass.Get() == nullptr) { 1141 VLOG(compiler) << "Failed to find class " << descriptor; 1142 image_classes_->erase(it++); 1143 self->ClearException(); 1144 } else { 1145 ++it; 1146 } 1147 } 1148 1149 // Resolve exception classes referenced by the loaded classes. The catch logic assumes 1150 // exceptions are resolved by the verifier when there is a catch block in an interested method. 1151 // Do this here so that exception classes appear to have been specified image classes. 1152 std::set<std::pair<dex::TypeIndex, const DexFile*>> unresolved_exception_types; 1153 StackHandleScope<1> hs(self); 1154 Handle<mirror::Class> java_lang_Throwable( 1155 hs.NewHandle(class_linker->FindSystemClass(self, "Ljava/lang/Throwable;"))); 1156 do { 1157 unresolved_exception_types.clear(); 1158 ResolveCatchBlockExceptionsClassVisitor visitor(unresolved_exception_types); 1159 class_linker->VisitClasses(&visitor); 1160 for (const auto& exception_type : unresolved_exception_types) { 1161 dex::TypeIndex exception_type_idx = exception_type.first; 1162 const DexFile* dex_file = exception_type.second; 1163 StackHandleScope<2> hs2(self); 1164 Handle<mirror::DexCache> dex_cache(hs2.NewHandle(class_linker->RegisterDexFile(*dex_file, 1165 nullptr))); 1166 Handle<mirror::Class> klass(hs2.NewHandle( 1167 class_linker->ResolveType(*dex_file, 1168 exception_type_idx, 1169 dex_cache, 1170 ScopedNullHandle<mirror::ClassLoader>()))); 1171 if (klass.Get() == nullptr) { 1172 const DexFile::TypeId& type_id = dex_file->GetTypeId(exception_type_idx); 1173 const char* descriptor = dex_file->GetTypeDescriptor(type_id); 1174 LOG(FATAL) << "Failed to resolve class " << descriptor; 1175 } 1176 DCHECK(java_lang_Throwable->IsAssignableFrom(klass.Get())); 1177 } 1178 // Resolving exceptions may load classes that reference more exceptions, iterate until no 1179 // more are found 1180 } while (!unresolved_exception_types.empty()); 1181 1182 // We walk the roots looking for classes so that we'll pick up the 1183 // above classes plus any classes them depend on such super 1184 // classes, interfaces, and the required ClassLinker roots. 1185 RecordImageClassesVisitor visitor(image_classes_.get()); 1186 class_linker->VisitClasses(&visitor); 1187 1188 CHECK_NE(image_classes_->size(), 0U); 1189} 1190 1191static void MaybeAddToImageClasses(Handle<mirror::Class> c, 1192 std::unordered_set<std::string>* image_classes) 1193 REQUIRES_SHARED(Locks::mutator_lock_) { 1194 Thread* self = Thread::Current(); 1195 StackHandleScope<1> hs(self); 1196 // Make a copy of the handle so that we don't clobber it doing Assign. 1197 MutableHandle<mirror::Class> klass(hs.NewHandle(c.Get())); 1198 std::string temp; 1199 const PointerSize pointer_size = Runtime::Current()->GetClassLinker()->GetImagePointerSize(); 1200 while (!klass->IsObjectClass()) { 1201 const char* descriptor = klass->GetDescriptor(&temp); 1202 std::pair<std::unordered_set<std::string>::iterator, bool> result = 1203 image_classes->insert(descriptor); 1204 if (!result.second) { // Previously inserted. 1205 break; 1206 } 1207 VLOG(compiler) << "Adding " << descriptor << " to image classes"; 1208 for (size_t i = 0; i < klass->NumDirectInterfaces(); ++i) { 1209 StackHandleScope<1> hs2(self); 1210 // May cause thread suspension. 1211 MaybeAddToImageClasses(hs2.NewHandle(mirror::Class::GetDirectInterface(self, klass, i)), 1212 image_classes); 1213 } 1214 for (auto& m : c->GetVirtualMethods(pointer_size)) { 1215 StackHandleScope<1> hs2(self); 1216 MaybeAddToImageClasses(hs2.NewHandle(m.GetDeclaringClass()), image_classes); 1217 } 1218 if (klass->IsArrayClass()) { 1219 StackHandleScope<1> hs2(self); 1220 MaybeAddToImageClasses(hs2.NewHandle(klass->GetComponentType()), image_classes); 1221 } 1222 klass.Assign(klass->GetSuperClass()); 1223 } 1224} 1225 1226// Keeps all the data for the update together. Also doubles as the reference visitor. 1227// Note: we can use object pointers because we suspend all threads. 1228class ClinitImageUpdate { 1229 public: 1230 static ClinitImageUpdate* Create(VariableSizedHandleScope& hs, 1231 std::unordered_set<std::string>* image_class_descriptors, 1232 Thread* self, 1233 ClassLinker* linker) { 1234 std::unique_ptr<ClinitImageUpdate> res(new ClinitImageUpdate(hs, 1235 image_class_descriptors, 1236 self, 1237 linker)); 1238 return res.release(); 1239 } 1240 1241 ~ClinitImageUpdate() { 1242 // Allow others to suspend again. 1243 self_->EndAssertNoThreadSuspension(old_cause_); 1244 } 1245 1246 // Visitor for VisitReferences. 1247 void operator()(ObjPtr<mirror::Object> object, 1248 MemberOffset field_offset, 1249 bool /* is_static */) const 1250 REQUIRES_SHARED(Locks::mutator_lock_) { 1251 mirror::Object* ref = object->GetFieldObject<mirror::Object>(field_offset); 1252 if (ref != nullptr) { 1253 VisitClinitClassesObject(ref); 1254 } 1255 } 1256 1257 // java.lang.Reference visitor for VisitReferences. 1258 void operator()(ObjPtr<mirror::Class> klass ATTRIBUTE_UNUSED, 1259 ObjPtr<mirror::Reference> ref ATTRIBUTE_UNUSED) const {} 1260 1261 // Ignore class native roots. 1262 void VisitRootIfNonNull(mirror::CompressedReference<mirror::Object>* root ATTRIBUTE_UNUSED) 1263 const {} 1264 void VisitRoot(mirror::CompressedReference<mirror::Object>* root ATTRIBUTE_UNUSED) const {} 1265 1266 void Walk() REQUIRES_SHARED(Locks::mutator_lock_) { 1267 // Use the initial classes as roots for a search. 1268 for (Handle<mirror::Class> klass_root : image_classes_) { 1269 VisitClinitClassesObject(klass_root.Get()); 1270 } 1271 for (Handle<mirror::Class> h_klass : to_insert_) { 1272 MaybeAddToImageClasses(h_klass, image_class_descriptors_); 1273 } 1274 } 1275 1276 private: 1277 class FindImageClassesVisitor : public ClassVisitor { 1278 public: 1279 explicit FindImageClassesVisitor(VariableSizedHandleScope& hs, 1280 ClinitImageUpdate* data) 1281 : data_(data), 1282 hs_(hs) {} 1283 1284 bool operator()(ObjPtr<mirror::Class> klass) OVERRIDE REQUIRES_SHARED(Locks::mutator_lock_) { 1285 std::string temp; 1286 const char* name = klass->GetDescriptor(&temp); 1287 if (data_->image_class_descriptors_->find(name) != data_->image_class_descriptors_->end()) { 1288 data_->image_classes_.push_back(hs_.NewHandle(klass)); 1289 } else { 1290 // Check whether it is initialized and has a clinit. They must be kept, too. 1291 if (klass->IsInitialized() && klass->FindClassInitializer( 1292 Runtime::Current()->GetClassLinker()->GetImagePointerSize()) != nullptr) { 1293 data_->image_classes_.push_back(hs_.NewHandle(klass)); 1294 } 1295 } 1296 return true; 1297 } 1298 1299 private: 1300 ClinitImageUpdate* const data_; 1301 VariableSizedHandleScope& hs_; 1302 }; 1303 1304 ClinitImageUpdate(VariableSizedHandleScope& hs, 1305 std::unordered_set<std::string>* image_class_descriptors, 1306 Thread* self, 1307 ClassLinker* linker) REQUIRES_SHARED(Locks::mutator_lock_) 1308 : hs_(hs), 1309 image_class_descriptors_(image_class_descriptors), 1310 self_(self) { 1311 CHECK(linker != nullptr); 1312 CHECK(image_class_descriptors != nullptr); 1313 1314 // Make sure nobody interferes with us. 1315 old_cause_ = self->StartAssertNoThreadSuspension("Boot image closure"); 1316 1317 // Find all the already-marked classes. 1318 WriterMutexLock mu(self, *Locks::heap_bitmap_lock_); 1319 FindImageClassesVisitor visitor(hs_, this); 1320 linker->VisitClasses(&visitor); 1321 } 1322 1323 void VisitClinitClassesObject(mirror::Object* object) const 1324 REQUIRES_SHARED(Locks::mutator_lock_) { 1325 DCHECK(object != nullptr); 1326 if (marked_objects_.find(object) != marked_objects_.end()) { 1327 // Already processed. 1328 return; 1329 } 1330 1331 // Mark it. 1332 marked_objects_.insert(object); 1333 1334 if (object->IsClass()) { 1335 // Add to the TODO list since MaybeAddToImageClasses may cause thread suspension. Thread 1336 // suspensionb is not safe to do in VisitObjects or VisitReferences. 1337 to_insert_.push_back(hs_.NewHandle(object->AsClass())); 1338 } else { 1339 // Else visit the object's class. 1340 VisitClinitClassesObject(object->GetClass()); 1341 } 1342 1343 // If it is not a DexCache, visit all references. 1344 if (!object->IsDexCache()) { 1345 object->VisitReferences(*this, *this); 1346 } 1347 } 1348 1349 VariableSizedHandleScope& hs_; 1350 mutable std::vector<Handle<mirror::Class>> to_insert_; 1351 mutable std::unordered_set<mirror::Object*> marked_objects_; 1352 std::unordered_set<std::string>* const image_class_descriptors_; 1353 std::vector<Handle<mirror::Class>> image_classes_; 1354 Thread* const self_; 1355 const char* old_cause_; 1356 1357 DISALLOW_COPY_AND_ASSIGN(ClinitImageUpdate); 1358}; 1359 1360void CompilerDriver::UpdateImageClasses(TimingLogger* timings) { 1361 if (GetCompilerOptions().IsBootImage()) { 1362 TimingLogger::ScopedTiming t("UpdateImageClasses", timings); 1363 1364 Runtime* runtime = Runtime::Current(); 1365 1366 // Suspend all threads. 1367 ScopedSuspendAll ssa(__FUNCTION__); 1368 1369 VariableSizedHandleScope hs(Thread::Current()); 1370 std::string error_msg; 1371 std::unique_ptr<ClinitImageUpdate> update(ClinitImageUpdate::Create(hs, 1372 image_classes_.get(), 1373 Thread::Current(), 1374 runtime->GetClassLinker())); 1375 1376 // Do the marking. 1377 update->Walk(); 1378 } 1379} 1380 1381bool CompilerDriver::CanAssumeClassIsLoaded(mirror::Class* klass) { 1382 Runtime* runtime = Runtime::Current(); 1383 if (!runtime->IsAotCompiler()) { 1384 DCHECK(runtime->UseJitCompilation()); 1385 // Having the klass reference here implies that the klass is already loaded. 1386 return true; 1387 } 1388 if (!GetCompilerOptions().IsBootImage()) { 1389 // Assume loaded only if klass is in the boot image. App classes cannot be assumed 1390 // loaded because we don't even know what class loader will be used to load them. 1391 bool class_in_image = runtime->GetHeap()->FindSpaceFromObject(klass, false)->IsImageSpace(); 1392 return class_in_image; 1393 } 1394 std::string temp; 1395 const char* descriptor = klass->GetDescriptor(&temp); 1396 return IsImageClass(descriptor); 1397} 1398 1399void CompilerDriver::MarkForDexToDexCompilation(Thread* self, const MethodReference& method_ref) { 1400 MutexLock lock(self, dex_to_dex_references_lock_); 1401 // Since we're compiling one dex file at a time, we need to look for the 1402 // current dex file entry only at the end of dex_to_dex_references_. 1403 if (dex_to_dex_references_.empty() || 1404 &dex_to_dex_references_.back().GetDexFile() != method_ref.dex_file) { 1405 dex_to_dex_references_.emplace_back(*method_ref.dex_file); 1406 } 1407 dex_to_dex_references_.back().GetMethodIndexes().SetBit(method_ref.dex_method_index); 1408} 1409 1410bool CompilerDriver::CanAccessTypeWithoutChecks(uint32_t referrer_idx, 1411 Handle<mirror::DexCache> dex_cache, 1412 dex::TypeIndex type_idx) { 1413 // Get type from dex cache assuming it was populated by the verifier 1414 mirror::Class* resolved_class = dex_cache->GetResolvedType(type_idx); 1415 if (resolved_class == nullptr) { 1416 stats_->TypeNeedsAccessCheck(); 1417 return false; // Unknown class needs access checks. 1418 } 1419 const DexFile::MethodId& method_id = dex_cache->GetDexFile()->GetMethodId(referrer_idx); 1420 bool is_accessible = resolved_class->IsPublic(); // Public classes are always accessible. 1421 if (!is_accessible) { 1422 mirror::Class* referrer_class = dex_cache->GetResolvedType(method_id.class_idx_); 1423 if (referrer_class == nullptr) { 1424 stats_->TypeNeedsAccessCheck(); 1425 return false; // Incomplete referrer knowledge needs access check. 1426 } 1427 // Perform access check, will return true if access is ok or false if we're going to have to 1428 // check this at runtime (for example for class loaders). 1429 is_accessible = referrer_class->CanAccess(resolved_class); 1430 } 1431 if (is_accessible) { 1432 stats_->TypeDoesntNeedAccessCheck(); 1433 } else { 1434 stats_->TypeNeedsAccessCheck(); 1435 } 1436 return is_accessible; 1437} 1438 1439bool CompilerDriver::CanAccessInstantiableTypeWithoutChecks(uint32_t referrer_idx, 1440 Handle<mirror::DexCache> dex_cache, 1441 dex::TypeIndex type_idx, 1442 bool* finalizable) { 1443 // Get type from dex cache assuming it was populated by the verifier. 1444 mirror::Class* resolved_class = dex_cache->GetResolvedType(type_idx); 1445 if (resolved_class == nullptr) { 1446 stats_->TypeNeedsAccessCheck(); 1447 // Be conservative. 1448 *finalizable = true; 1449 return false; // Unknown class needs access checks. 1450 } 1451 *finalizable = resolved_class->IsFinalizable(); 1452 const DexFile::MethodId& method_id = dex_cache->GetDexFile()->GetMethodId(referrer_idx); 1453 bool is_accessible = resolved_class->IsPublic(); // Public classes are always accessible. 1454 if (!is_accessible) { 1455 mirror::Class* referrer_class = dex_cache->GetResolvedType(method_id.class_idx_); 1456 if (referrer_class == nullptr) { 1457 stats_->TypeNeedsAccessCheck(); 1458 return false; // Incomplete referrer knowledge needs access check. 1459 } 1460 // Perform access and instantiable checks, will return true if access is ok or false if we're 1461 // going to have to check this at runtime (for example for class loaders). 1462 is_accessible = referrer_class->CanAccess(resolved_class); 1463 } 1464 bool result = is_accessible && resolved_class->IsInstantiable(); 1465 if (result) { 1466 stats_->TypeDoesntNeedAccessCheck(); 1467 } else { 1468 stats_->TypeNeedsAccessCheck(); 1469 } 1470 return result; 1471} 1472 1473void CompilerDriver::ProcessedInstanceField(bool resolved) { 1474 if (!resolved) { 1475 stats_->UnresolvedInstanceField(); 1476 } else { 1477 stats_->ResolvedInstanceField(); 1478 } 1479} 1480 1481void CompilerDriver::ProcessedStaticField(bool resolved, bool local) { 1482 if (!resolved) { 1483 stats_->UnresolvedStaticField(); 1484 } else if (local) { 1485 stats_->ResolvedLocalStaticField(); 1486 } else { 1487 stats_->ResolvedStaticField(); 1488 } 1489} 1490 1491ArtField* CompilerDriver::ComputeInstanceFieldInfo(uint32_t field_idx, 1492 const DexCompilationUnit* mUnit, bool is_put, 1493 const ScopedObjectAccess& soa) { 1494 // Try to resolve the field and compiling method's class. 1495 ArtField* resolved_field; 1496 mirror::Class* referrer_class; 1497 Handle<mirror::DexCache> dex_cache(mUnit->GetDexCache()); 1498 { 1499 StackHandleScope<1> hs(soa.Self()); 1500 Handle<mirror::ClassLoader> class_loader_handle( 1501 hs.NewHandle(soa.Decode<mirror::ClassLoader>(mUnit->GetClassLoader()))); 1502 resolved_field = ResolveField(soa, dex_cache, class_loader_handle, mUnit, field_idx, false); 1503 referrer_class = resolved_field != nullptr 1504 ? ResolveCompilingMethodsClass(soa, dex_cache, class_loader_handle, mUnit) : nullptr; 1505 } 1506 bool can_link = false; 1507 if (resolved_field != nullptr && referrer_class != nullptr) { 1508 std::pair<bool, bool> fast_path = IsFastInstanceField( 1509 dex_cache.Get(), referrer_class, resolved_field, field_idx); 1510 can_link = is_put ? fast_path.second : fast_path.first; 1511 } 1512 ProcessedInstanceField(can_link); 1513 return can_link ? resolved_field : nullptr; 1514} 1515 1516bool CompilerDriver::ComputeInstanceFieldInfo(uint32_t field_idx, const DexCompilationUnit* mUnit, 1517 bool is_put, MemberOffset* field_offset, 1518 bool* is_volatile) { 1519 ScopedObjectAccess soa(Thread::Current()); 1520 ArtField* resolved_field = ComputeInstanceFieldInfo(field_idx, mUnit, is_put, soa); 1521 1522 if (resolved_field == nullptr) { 1523 // Conservative defaults. 1524 *is_volatile = true; 1525 *field_offset = MemberOffset(static_cast<size_t>(-1)); 1526 return false; 1527 } else { 1528 *is_volatile = resolved_field->IsVolatile(); 1529 *field_offset = resolved_field->GetOffset(); 1530 return true; 1531 } 1532} 1533 1534void CompilerDriver::GetCodeAndMethodForDirectCall(const mirror::Class* referrer_class, 1535 ArtMethod* method, 1536 uintptr_t* direct_code, 1537 uintptr_t* direct_method) { 1538 // For direct and static methods compute possible direct_code and direct_method values, ie 1539 // an address for the Method* being invoked and an address of the code for that Method*. 1540 // For interface calls compute a value for direct_method that is the interface method being 1541 // invoked, so this can be passed to the out-of-line runtime support code. 1542 *direct_code = 0; 1543 *direct_method = 0; 1544 Runtime* const runtime = Runtime::Current(); 1545 gc::Heap* const heap = runtime->GetHeap(); 1546 auto* cl = runtime->GetClassLinker(); 1547 bool use_dex_cache = GetCompilerOptions().GetCompilePic(); // Off by default 1548 const bool compiling_boot = heap->IsCompilingBoot(); 1549 // TODO This is somewhat hacky. We should refactor all of this invoke codepath. 1550 const bool force_relocations = (compiling_boot || 1551 GetCompilerOptions().GetIncludePatchInformation()); 1552 // TODO: support patching on all architectures. 1553 use_dex_cache = use_dex_cache || (force_relocations && !support_boot_image_fixup_); 1554 mirror::Class* declaring_class = method->GetDeclaringClass(); 1555 bool method_code_in_boot = declaring_class->GetClassLoader() == nullptr; 1556 if (!use_dex_cache) { 1557 if (!method_code_in_boot) { 1558 use_dex_cache = true; 1559 } else if (method->IsStatic() && 1560 declaring_class != referrer_class && 1561 !declaring_class->IsInitialized()) { 1562 // Ensure we run the clinit trampoline unless we are invoking a static method in the same 1563 // class. 1564 use_dex_cache = true; 1565 } 1566 } 1567 if (runtime->UseJitCompilation()) { 1568 // If we are the JIT, then don't allow a direct call to the interpreter bridge since this will 1569 // never be updated even after we compile the method. 1570 if (cl->IsQuickToInterpreterBridge( 1571 reinterpret_cast<const void*>(compiler_->GetEntryPointOf(method)))) { 1572 use_dex_cache = true; 1573 } 1574 } 1575 1576 if (!use_dex_cache && force_relocations) { 1577 bool is_in_image; 1578 if (GetCompilerOptions().IsBootImage()) { 1579 is_in_image = IsImageClass(method->GetDeclaringClassDescriptor()); 1580 } else { 1581 is_in_image = instruction_set_ != kX86 && instruction_set_ != kX86_64 && 1582 heap->FindSpaceFromObject(method->GetDeclaringClass(), false)->IsImageSpace() && 1583 !cl->IsQuickToInterpreterBridge( 1584 reinterpret_cast<const void*>(compiler_->GetEntryPointOf(method))); 1585 } 1586 if (!is_in_image) { 1587 // We can only branch directly to Methods that are resolved in the DexCache. 1588 // Otherwise we won't invoke the resolution trampoline. 1589 use_dex_cache = true; 1590 } 1591 } 1592 1593 if (!use_dex_cache) { 1594 bool method_in_image = false; 1595 const std::vector<gc::space::ImageSpace*>& image_spaces = heap->GetBootImageSpaces(); 1596 for (gc::space::ImageSpace* image_space : image_spaces) { 1597 const auto& method_section = image_space->GetImageHeader().GetMethodsSection(); 1598 if (method_section.Contains(reinterpret_cast<uint8_t*>(method) - image_space->Begin())) { 1599 method_in_image = true; 1600 break; 1601 } 1602 } 1603 if (method_in_image || compiling_boot || runtime->UseJitCompilation()) { 1604 // We know we must be able to get to the method in the image, so use that pointer. 1605 // In the case where we are the JIT, we can always use direct pointers since we know where 1606 // the method and its code are / will be. We don't sharpen to interpreter bridge since we 1607 // check IsQuickToInterpreterBridge above. 1608 CHECK(!method->IsAbstract()); 1609 *direct_method = force_relocations ? -1 : reinterpret_cast<uintptr_t>(method); 1610 *direct_code = force_relocations ? -1 : compiler_->GetEntryPointOf(method); 1611 } else { 1612 // Set the code and rely on the dex cache for the method. 1613 *direct_code = force_relocations ? -1 : compiler_->GetEntryPointOf(method); 1614 } 1615 } 1616} 1617 1618const VerifiedMethod* CompilerDriver::GetVerifiedMethod(const DexFile* dex_file, 1619 uint32_t method_idx) const { 1620 MethodReference ref(dex_file, method_idx); 1621 return verification_results_->GetVerifiedMethod(ref); 1622} 1623 1624bool CompilerDriver::IsSafeCast(const DexCompilationUnit* mUnit, uint32_t dex_pc) { 1625 if (!compiler_options_->IsVerificationEnabled()) { 1626 // If we didn't verify, every cast has to be treated as non-safe. 1627 return false; 1628 } 1629 DCHECK(mUnit->GetVerifiedMethod() != nullptr); 1630 bool result = mUnit->GetVerifiedMethod()->IsSafeCast(dex_pc); 1631 if (result) { 1632 stats_->SafeCast(); 1633 } else { 1634 stats_->NotASafeCast(); 1635 } 1636 return result; 1637} 1638 1639class CompilationVisitor { 1640 public: 1641 virtual ~CompilationVisitor() {} 1642 virtual void Visit(size_t index) = 0; 1643}; 1644 1645class ParallelCompilationManager { 1646 public: 1647 ParallelCompilationManager(ClassLinker* class_linker, 1648 jobject class_loader, 1649 CompilerDriver* compiler, 1650 const DexFile* dex_file, 1651 const std::vector<const DexFile*>& dex_files, 1652 ThreadPool* thread_pool) 1653 : index_(0), 1654 class_linker_(class_linker), 1655 class_loader_(class_loader), 1656 compiler_(compiler), 1657 dex_file_(dex_file), 1658 dex_files_(dex_files), 1659 thread_pool_(thread_pool) {} 1660 1661 ClassLinker* GetClassLinker() const { 1662 CHECK(class_linker_ != nullptr); 1663 return class_linker_; 1664 } 1665 1666 jobject GetClassLoader() const { 1667 return class_loader_; 1668 } 1669 1670 CompilerDriver* GetCompiler() const { 1671 CHECK(compiler_ != nullptr); 1672 return compiler_; 1673 } 1674 1675 const DexFile* GetDexFile() const { 1676 CHECK(dex_file_ != nullptr); 1677 return dex_file_; 1678 } 1679 1680 const std::vector<const DexFile*>& GetDexFiles() const { 1681 return dex_files_; 1682 } 1683 1684 void ForAll(size_t begin, size_t end, CompilationVisitor* visitor, size_t work_units) 1685 REQUIRES(!*Locks::mutator_lock_) { 1686 Thread* self = Thread::Current(); 1687 self->AssertNoPendingException(); 1688 CHECK_GT(work_units, 0U); 1689 1690 index_.StoreRelaxed(begin); 1691 for (size_t i = 0; i < work_units; ++i) { 1692 thread_pool_->AddTask(self, new ForAllClosure(this, end, visitor)); 1693 } 1694 thread_pool_->StartWorkers(self); 1695 1696 // Ensure we're suspended while we're blocked waiting for the other threads to finish (worker 1697 // thread destructor's called below perform join). 1698 CHECK_NE(self->GetState(), kRunnable); 1699 1700 // Wait for all the worker threads to finish. 1701 thread_pool_->Wait(self, true, false); 1702 1703 // And stop the workers accepting jobs. 1704 thread_pool_->StopWorkers(self); 1705 } 1706 1707 size_t NextIndex() { 1708 return index_.FetchAndAddSequentiallyConsistent(1); 1709 } 1710 1711 private: 1712 class ForAllClosure : public Task { 1713 public: 1714 ForAllClosure(ParallelCompilationManager* manager, size_t end, CompilationVisitor* visitor) 1715 : manager_(manager), 1716 end_(end), 1717 visitor_(visitor) {} 1718 1719 virtual void Run(Thread* self) { 1720 while (true) { 1721 const size_t index = manager_->NextIndex(); 1722 if (UNLIKELY(index >= end_)) { 1723 break; 1724 } 1725 visitor_->Visit(index); 1726 self->AssertNoPendingException(); 1727 } 1728 } 1729 1730 virtual void Finalize() { 1731 delete this; 1732 } 1733 1734 private: 1735 ParallelCompilationManager* const manager_; 1736 const size_t end_; 1737 CompilationVisitor* const visitor_; 1738 }; 1739 1740 AtomicInteger index_; 1741 ClassLinker* const class_linker_; 1742 const jobject class_loader_; 1743 CompilerDriver* const compiler_; 1744 const DexFile* const dex_file_; 1745 const std::vector<const DexFile*>& dex_files_; 1746 ThreadPool* const thread_pool_; 1747 1748 DISALLOW_COPY_AND_ASSIGN(ParallelCompilationManager); 1749}; 1750 1751// A fast version of SkipClass above if the class pointer is available 1752// that avoids the expensive FindInClassPath search. 1753static bool SkipClass(jobject class_loader, const DexFile& dex_file, mirror::Class* klass) 1754 REQUIRES_SHARED(Locks::mutator_lock_) { 1755 DCHECK(klass != nullptr); 1756 const DexFile& original_dex_file = *klass->GetDexCache()->GetDexFile(); 1757 if (&dex_file != &original_dex_file) { 1758 if (class_loader == nullptr) { 1759 LOG(WARNING) << "Skipping class " << klass->PrettyDescriptor() << " from " 1760 << dex_file.GetLocation() << " previously found in " 1761 << original_dex_file.GetLocation(); 1762 } 1763 return true; 1764 } 1765 return false; 1766} 1767 1768static void CheckAndClearResolveException(Thread* self) 1769 REQUIRES_SHARED(Locks::mutator_lock_) { 1770 CHECK(self->IsExceptionPending()); 1771 mirror::Throwable* exception = self->GetException(); 1772 std::string temp; 1773 const char* descriptor = exception->GetClass()->GetDescriptor(&temp); 1774 const char* expected_exceptions[] = { 1775 "Ljava/lang/IllegalAccessError;", 1776 "Ljava/lang/IncompatibleClassChangeError;", 1777 "Ljava/lang/InstantiationError;", 1778 "Ljava/lang/LinkageError;", 1779 "Ljava/lang/NoClassDefFoundError;", 1780 "Ljava/lang/NoSuchFieldError;", 1781 "Ljava/lang/NoSuchMethodError;" 1782 }; 1783 bool found = false; 1784 for (size_t i = 0; (found == false) && (i < arraysize(expected_exceptions)); ++i) { 1785 if (strcmp(descriptor, expected_exceptions[i]) == 0) { 1786 found = true; 1787 } 1788 } 1789 if (!found) { 1790 LOG(FATAL) << "Unexpected exception " << exception->Dump(); 1791 } 1792 self->ClearException(); 1793} 1794 1795bool CompilerDriver::RequiresConstructorBarrier(const DexFile& dex_file, 1796 uint16_t class_def_idx) const { 1797 const DexFile::ClassDef& class_def = dex_file.GetClassDef(class_def_idx); 1798 const uint8_t* class_data = dex_file.GetClassData(class_def); 1799 if (class_data == nullptr) { 1800 // Empty class such as a marker interface. 1801 return false; 1802 } 1803 ClassDataItemIterator it(dex_file, class_data); 1804 while (it.HasNextStaticField()) { 1805 it.Next(); 1806 } 1807 // We require a constructor barrier if there are final instance fields. 1808 while (it.HasNextInstanceField()) { 1809 if (it.MemberIsFinal()) { 1810 return true; 1811 } 1812 it.Next(); 1813 } 1814 return false; 1815} 1816 1817class ResolveClassFieldsAndMethodsVisitor : public CompilationVisitor { 1818 public: 1819 explicit ResolveClassFieldsAndMethodsVisitor(const ParallelCompilationManager* manager) 1820 : manager_(manager) {} 1821 1822 void Visit(size_t class_def_index) OVERRIDE REQUIRES(!Locks::mutator_lock_) { 1823 ATRACE_CALL(); 1824 Thread* const self = Thread::Current(); 1825 jobject jclass_loader = manager_->GetClassLoader(); 1826 const DexFile& dex_file = *manager_->GetDexFile(); 1827 ClassLinker* class_linker = manager_->GetClassLinker(); 1828 1829 // If an instance field is final then we need to have a barrier on the return, static final 1830 // fields are assigned within the lock held for class initialization. Conservatively assume 1831 // constructor barriers are always required. 1832 bool requires_constructor_barrier = true; 1833 1834 // Method and Field are the worst. We can't resolve without either 1835 // context from the code use (to disambiguate virtual vs direct 1836 // method and instance vs static field) or from class 1837 // definitions. While the compiler will resolve what it can as it 1838 // needs it, here we try to resolve fields and methods used in class 1839 // definitions, since many of them many never be referenced by 1840 // generated code. 1841 const DexFile::ClassDef& class_def = dex_file.GetClassDef(class_def_index); 1842 ScopedObjectAccess soa(self); 1843 StackHandleScope<2> hs(soa.Self()); 1844 Handle<mirror::ClassLoader> class_loader( 1845 hs.NewHandle(soa.Decode<mirror::ClassLoader>(jclass_loader))); 1846 Handle<mirror::DexCache> dex_cache(hs.NewHandle(class_linker->FindDexCache( 1847 soa.Self(), dex_file, false))); 1848 // Resolve the class. 1849 mirror::Class* klass = class_linker->ResolveType(dex_file, class_def.class_idx_, dex_cache, 1850 class_loader); 1851 bool resolve_fields_and_methods; 1852 if (klass == nullptr) { 1853 // Class couldn't be resolved, for example, super-class is in a different dex file. Don't 1854 // attempt to resolve methods and fields when there is no declaring class. 1855 CheckAndClearResolveException(soa.Self()); 1856 resolve_fields_and_methods = false; 1857 } else { 1858 // We successfully resolved a class, should we skip it? 1859 if (SkipClass(jclass_loader, dex_file, klass)) { 1860 return; 1861 } 1862 // We want to resolve the methods and fields eagerly. 1863 resolve_fields_and_methods = true; 1864 } 1865 // Note the class_data pointer advances through the headers, 1866 // static fields, instance fields, direct methods, and virtual 1867 // methods. 1868 const uint8_t* class_data = dex_file.GetClassData(class_def); 1869 if (class_data == nullptr) { 1870 // Empty class such as a marker interface. 1871 requires_constructor_barrier = false; 1872 } else { 1873 ClassDataItemIterator it(dex_file, class_data); 1874 while (it.HasNextStaticField()) { 1875 if (resolve_fields_and_methods) { 1876 ArtField* field = class_linker->ResolveField(dex_file, it.GetMemberIndex(), 1877 dex_cache, class_loader, true); 1878 if (field == nullptr) { 1879 CheckAndClearResolveException(soa.Self()); 1880 } 1881 } 1882 it.Next(); 1883 } 1884 // We require a constructor barrier if there are final instance fields. 1885 requires_constructor_barrier = false; 1886 while (it.HasNextInstanceField()) { 1887 if (it.MemberIsFinal()) { 1888 requires_constructor_barrier = true; 1889 } 1890 if (resolve_fields_and_methods) { 1891 ArtField* field = class_linker->ResolveField(dex_file, it.GetMemberIndex(), 1892 dex_cache, class_loader, false); 1893 if (field == nullptr) { 1894 CheckAndClearResolveException(soa.Self()); 1895 } 1896 } 1897 it.Next(); 1898 } 1899 if (resolve_fields_and_methods) { 1900 while (it.HasNextDirectMethod()) { 1901 ArtMethod* method = class_linker->ResolveMethod<ClassLinker::kNoICCECheckForCache>( 1902 dex_file, it.GetMemberIndex(), dex_cache, class_loader, nullptr, 1903 it.GetMethodInvokeType(class_def)); 1904 if (method == nullptr) { 1905 CheckAndClearResolveException(soa.Self()); 1906 } 1907 it.Next(); 1908 } 1909 while (it.HasNextVirtualMethod()) { 1910 ArtMethod* method = class_linker->ResolveMethod<ClassLinker::kNoICCECheckForCache>( 1911 dex_file, it.GetMemberIndex(), dex_cache, class_loader, nullptr, 1912 it.GetMethodInvokeType(class_def)); 1913 if (method == nullptr) { 1914 CheckAndClearResolveException(soa.Self()); 1915 } 1916 it.Next(); 1917 } 1918 DCHECK(!it.HasNext()); 1919 } 1920 } 1921 manager_->GetCompiler()->SetRequiresConstructorBarrier(self, 1922 &dex_file, 1923 class_def_index, 1924 requires_constructor_barrier); 1925 } 1926 1927 private: 1928 const ParallelCompilationManager* const manager_; 1929}; 1930 1931class ResolveTypeVisitor : public CompilationVisitor { 1932 public: 1933 explicit ResolveTypeVisitor(const ParallelCompilationManager* manager) : manager_(manager) { 1934 } 1935 void Visit(size_t type_idx) OVERRIDE REQUIRES(!Locks::mutator_lock_) { 1936 // Class derived values are more complicated, they require the linker and loader. 1937 ScopedObjectAccess soa(Thread::Current()); 1938 ClassLinker* class_linker = manager_->GetClassLinker(); 1939 const DexFile& dex_file = *manager_->GetDexFile(); 1940 StackHandleScope<2> hs(soa.Self()); 1941 Handle<mirror::ClassLoader> class_loader( 1942 hs.NewHandle(soa.Decode<mirror::ClassLoader>(manager_->GetClassLoader()))); 1943 Handle<mirror::DexCache> dex_cache(hs.NewHandle(class_linker->RegisterDexFile( 1944 dex_file, 1945 class_loader.Get()))); 1946 mirror::Class* klass = class_linker->ResolveType(dex_file, 1947 dex::TypeIndex(type_idx), 1948 dex_cache, 1949 class_loader); 1950 1951 if (klass == nullptr) { 1952 soa.Self()->AssertPendingException(); 1953 mirror::Throwable* exception = soa.Self()->GetException(); 1954 VLOG(compiler) << "Exception during type resolution: " << exception->Dump(); 1955 if (exception->GetClass()->DescriptorEquals("Ljava/lang/OutOfMemoryError;")) { 1956 // There's little point continuing compilation if the heap is exhausted. 1957 LOG(FATAL) << "Out of memory during type resolution for compilation"; 1958 } 1959 soa.Self()->ClearException(); 1960 } 1961 } 1962 1963 private: 1964 const ParallelCompilationManager* const manager_; 1965}; 1966 1967void CompilerDriver::ResolveDexFile(jobject class_loader, 1968 const DexFile& dex_file, 1969 const std::vector<const DexFile*>& dex_files, 1970 ThreadPool* thread_pool, 1971 size_t thread_count, 1972 TimingLogger* timings) { 1973 ClassLinker* class_linker = Runtime::Current()->GetClassLinker(); 1974 1975 // TODO: we could resolve strings here, although the string table is largely filled with class 1976 // and method names. 1977 1978 ParallelCompilationManager context(class_linker, class_loader, this, &dex_file, dex_files, 1979 thread_pool); 1980 if (GetCompilerOptions().IsBootImage()) { 1981 // For images we resolve all types, such as array, whereas for applications just those with 1982 // classdefs are resolved by ResolveClassFieldsAndMethods. 1983 TimingLogger::ScopedTiming t("Resolve Types", timings); 1984 ResolveTypeVisitor visitor(&context); 1985 context.ForAll(0, dex_file.NumTypeIds(), &visitor, thread_count); 1986 } 1987 1988 TimingLogger::ScopedTiming t("Resolve MethodsAndFields", timings); 1989 ResolveClassFieldsAndMethodsVisitor visitor(&context); 1990 context.ForAll(0, dex_file.NumClassDefs(), &visitor, thread_count); 1991} 1992 1993void CompilerDriver::SetVerified(jobject class_loader, 1994 const std::vector<const DexFile*>& dex_files, 1995 TimingLogger* timings) { 1996 // This can be run in parallel. 1997 for (const DexFile* dex_file : dex_files) { 1998 CHECK(dex_file != nullptr); 1999 SetVerifiedDexFile(class_loader, 2000 *dex_file, 2001 dex_files, 2002 parallel_thread_pool_.get(), 2003 parallel_thread_count_, 2004 timings); 2005 } 2006} 2007 2008void CompilerDriver::Verify(jobject jclass_loader, 2009 const std::vector<const DexFile*>& dex_files, 2010 TimingLogger* timings) { 2011 verifier::VerifierDeps* verifier_deps = 2012 Runtime::Current()->GetCompilerCallbacks()->GetVerifierDeps(); 2013 // If there is an existing `VerifierDeps`, try to use it for fast verification. 2014 if (verifier_deps != nullptr) { 2015 TimingLogger::ScopedTiming t("Fast Verify", timings); 2016 ScopedObjectAccess soa(Thread::Current()); 2017 StackHandleScope<2> hs(soa.Self()); 2018 Handle<mirror::ClassLoader> class_loader( 2019 hs.NewHandle(soa.Decode<mirror::ClassLoader>(jclass_loader))); 2020 MutableHandle<mirror::Class> cls(hs.NewHandle<mirror::Class>(nullptr)); 2021 ClassLinker* class_linker = Runtime::Current()->GetClassLinker(); 2022 if (verifier_deps->ValidateDependencies(class_loader, soa.Self())) { 2023 // We successfully validated the dependencies, now update class status 2024 // of verified classes. Note that the dependencies also record which classes 2025 // could not be fully verified; we could try again, but that would hurt verification 2026 // time. So instead we assume these classes still need to be verified at 2027 // runtime. 2028 for (const DexFile* dex_file : dex_files) { 2029 // Fetch the list of unverified classes and turn it into a set for faster 2030 // lookups. 2031 const std::vector<dex::TypeIndex>& unverified_classes = 2032 verifier_deps->GetUnverifiedClasses(*dex_file); 2033 std::set<dex::TypeIndex> set(unverified_classes.begin(), unverified_classes.end()); 2034 for (uint32_t i = 0; i < dex_file->NumClassDefs(); ++i) { 2035 const DexFile::ClassDef& class_def = dex_file->GetClassDef(i); 2036 const char* descriptor = dex_file->GetClassDescriptor(class_def); 2037 cls.Assign(class_linker->FindClass(soa.Self(), descriptor, class_loader)); 2038 if (cls.Get() == nullptr) { 2039 CHECK(soa.Self()->IsExceptionPending()); 2040 soa.Self()->ClearException(); 2041 } else if (set.find(class_def.class_idx_) == set.end()) { 2042 ObjectLock<mirror::Class> lock(soa.Self(), cls); 2043 mirror::Class::SetStatus(cls, mirror::Class::kStatusVerified, soa.Self()); 2044 } 2045 } 2046 } 2047 return; 2048 } 2049 } 2050 2051 // If there is no existing `verifier_deps` (because of non-existing vdex), or 2052 // the existing `verifier_deps` is not valid anymore, create a new one for 2053 // non boot image compilation. The verifier will need it to record the new dependencies. 2054 // Then dex2oat can update the vdex file with these new dependencies. 2055 if (!GetCompilerOptions().IsBootImage()) { 2056 // Create the main VerifierDeps, and set it to this thread. 2057 verifier_deps = new verifier::VerifierDeps(dex_files); 2058 Runtime::Current()->GetCompilerCallbacks()->SetVerifierDeps(verifier_deps); 2059 Thread::Current()->SetVerifierDeps(verifier_deps); 2060 // Create per-thread VerifierDeps to avoid contention on the main one. 2061 // We will merge them after verification. 2062 for (ThreadPoolWorker* worker : parallel_thread_pool_->GetWorkers()) { 2063 worker->GetThread()->SetVerifierDeps(new verifier::VerifierDeps(dex_files)); 2064 } 2065 } 2066 // Note: verification should not be pulling in classes anymore when compiling the boot image, 2067 // as all should have been resolved before. As such, doing this in parallel should still 2068 // be deterministic. 2069 for (const DexFile* dex_file : dex_files) { 2070 CHECK(dex_file != nullptr); 2071 VerifyDexFile(jclass_loader, 2072 *dex_file, 2073 dex_files, 2074 parallel_thread_pool_.get(), 2075 parallel_thread_count_, 2076 timings); 2077 } 2078 2079 if (!GetCompilerOptions().IsBootImage()) { 2080 // Merge all VerifierDeps into the main one. 2081 for (ThreadPoolWorker* worker : parallel_thread_pool_->GetWorkers()) { 2082 verifier::VerifierDeps* thread_deps = worker->GetThread()->GetVerifierDeps(); 2083 worker->GetThread()->SetVerifierDeps(nullptr); 2084 verifier_deps->MergeWith(*thread_deps, dex_files);; 2085 delete thread_deps; 2086 } 2087 Thread::Current()->SetVerifierDeps(nullptr); 2088 } 2089} 2090 2091class VerifyClassVisitor : public CompilationVisitor { 2092 public: 2093 VerifyClassVisitor(const ParallelCompilationManager* manager, verifier::HardFailLogMode log_level) 2094 : manager_(manager), log_level_(log_level) {} 2095 2096 virtual void Visit(size_t class_def_index) REQUIRES(!Locks::mutator_lock_) OVERRIDE { 2097 ATRACE_CALL(); 2098 ScopedObjectAccess soa(Thread::Current()); 2099 const DexFile& dex_file = *manager_->GetDexFile(); 2100 if (!manager_->GetCompiler()->ShouldVerifyClassBasedOnProfile(dex_file, class_def_index)) { 2101 // Skip verification since the class is not in the profile. 2102 return; 2103 } 2104 const DexFile::ClassDef& class_def = dex_file.GetClassDef(class_def_index); 2105 const char* descriptor = dex_file.GetClassDescriptor(class_def); 2106 ClassLinker* class_linker = manager_->GetClassLinker(); 2107 jobject jclass_loader = manager_->GetClassLoader(); 2108 StackHandleScope<3> hs(soa.Self()); 2109 Handle<mirror::ClassLoader> class_loader( 2110 hs.NewHandle(soa.Decode<mirror::ClassLoader>(jclass_loader))); 2111 Handle<mirror::Class> klass( 2112 hs.NewHandle(class_linker->FindClass(soa.Self(), descriptor, class_loader))); 2113 verifier::MethodVerifier::FailureKind failure_kind; 2114 if (klass.Get() == nullptr) { 2115 CHECK(soa.Self()->IsExceptionPending()); 2116 soa.Self()->ClearException(); 2117 2118 /* 2119 * At compile time, we can still structurally verify the class even if FindClass fails. 2120 * This is to ensure the class is structurally sound for compilation. An unsound class 2121 * will be rejected by the verifier and later skipped during compilation in the compiler. 2122 */ 2123 Handle<mirror::DexCache> dex_cache(hs.NewHandle(class_linker->FindDexCache( 2124 soa.Self(), dex_file, false))); 2125 std::string error_msg; 2126 failure_kind = 2127 verifier::MethodVerifier::VerifyClass(soa.Self(), 2128 &dex_file, 2129 dex_cache, 2130 class_loader, 2131 class_def, 2132 Runtime::Current()->GetCompilerCallbacks(), 2133 true /* allow soft failures */, 2134 log_level_, 2135 &error_msg); 2136 if (failure_kind == verifier::MethodVerifier::kHardFailure) { 2137 LOG(ERROR) << "Verification failed on class " << PrettyDescriptor(descriptor) 2138 << " because: " << error_msg; 2139 manager_->GetCompiler()->SetHadHardVerifierFailure(); 2140 } 2141 } else if (!SkipClass(jclass_loader, dex_file, klass.Get())) { 2142 CHECK(klass->IsResolved()) << klass->PrettyClass(); 2143 failure_kind = class_linker->VerifyClass(soa.Self(), klass, log_level_); 2144 2145 if (klass->IsErroneous()) { 2146 // ClassLinker::VerifyClass throws, which isn't useful in the compiler. 2147 CHECK(soa.Self()->IsExceptionPending()); 2148 soa.Self()->ClearException(); 2149 manager_->GetCompiler()->SetHadHardVerifierFailure(); 2150 } 2151 2152 CHECK(klass->IsCompileTimeVerified() || klass->IsErroneous()) 2153 << klass->PrettyDescriptor() << ": state=" << klass->GetStatus(); 2154 2155 // It is *very* problematic if there are verification errors in the boot classpath. For example, 2156 // we rely on things working OK without verification when the decryption dialog is brought up. 2157 // So abort in a debug build if we find this violated. 2158 if (kIsDebugBuild) { 2159 // TODO(narayan): Remove this special case for signature polymorphic 2160 // invokes once verifier support is fully implemented. 2161 if (manager_->GetCompiler()->GetCompilerOptions().IsBootImage() && 2162 !android::base::StartsWith(descriptor, "Ljava/lang/invoke/")) { 2163 DCHECK(klass->IsVerified()) << "Boot classpath class " << klass->PrettyClass() 2164 << " failed to fully verify: state= " << klass->GetStatus(); 2165 } 2166 } 2167 } else { 2168 // Make the skip a soft failure, essentially being considered as verify at runtime. 2169 failure_kind = verifier::MethodVerifier::kSoftFailure; 2170 } 2171 verifier::VerifierDeps::MaybeRecordVerificationStatus( 2172 dex_file, class_def.class_idx_, failure_kind); 2173 soa.Self()->AssertNoPendingException(); 2174 } 2175 2176 private: 2177 const ParallelCompilationManager* const manager_; 2178 const verifier::HardFailLogMode log_level_; 2179}; 2180 2181void CompilerDriver::VerifyDexFile(jobject class_loader, 2182 const DexFile& dex_file, 2183 const std::vector<const DexFile*>& dex_files, 2184 ThreadPool* thread_pool, 2185 size_t thread_count, 2186 TimingLogger* timings) { 2187 TimingLogger::ScopedTiming t("Verify Dex File", timings); 2188 ClassLinker* class_linker = Runtime::Current()->GetClassLinker(); 2189 ParallelCompilationManager context(class_linker, class_loader, this, &dex_file, dex_files, 2190 thread_pool); 2191 verifier::HardFailLogMode log_level = GetCompilerOptions().AbortOnHardVerifierFailure() 2192 ? verifier::HardFailLogMode::kLogInternalFatal 2193 : verifier::HardFailLogMode::kLogWarning; 2194 VerifyClassVisitor visitor(&context, log_level); 2195 context.ForAll(0, dex_file.NumClassDefs(), &visitor, thread_count); 2196} 2197 2198class SetVerifiedClassVisitor : public CompilationVisitor { 2199 public: 2200 explicit SetVerifiedClassVisitor(const ParallelCompilationManager* manager) : manager_(manager) {} 2201 2202 virtual void Visit(size_t class_def_index) REQUIRES(!Locks::mutator_lock_) OVERRIDE { 2203 ATRACE_CALL(); 2204 ScopedObjectAccess soa(Thread::Current()); 2205 const DexFile& dex_file = *manager_->GetDexFile(); 2206 const DexFile::ClassDef& class_def = dex_file.GetClassDef(class_def_index); 2207 const char* descriptor = dex_file.GetClassDescriptor(class_def); 2208 ClassLinker* class_linker = manager_->GetClassLinker(); 2209 jobject jclass_loader = manager_->GetClassLoader(); 2210 StackHandleScope<3> hs(soa.Self()); 2211 Handle<mirror::ClassLoader> class_loader( 2212 hs.NewHandle(soa.Decode<mirror::ClassLoader>(jclass_loader))); 2213 Handle<mirror::Class> klass( 2214 hs.NewHandle(class_linker->FindClass(soa.Self(), descriptor, class_loader))); 2215 // Class might have failed resolution. Then don't set it to verified. 2216 if (klass.Get() != nullptr) { 2217 // Only do this if the class is resolved. If even resolution fails, quickening will go very, 2218 // very wrong. 2219 if (klass->IsResolved()) { 2220 if (klass->GetStatus() < mirror::Class::kStatusVerified) { 2221 ObjectLock<mirror::Class> lock(soa.Self(), klass); 2222 // Set class status to verified. 2223 mirror::Class::SetStatus(klass, mirror::Class::kStatusVerified, soa.Self()); 2224 // Mark methods as pre-verified. If we don't do this, the interpreter will run with 2225 // access checks. 2226 klass->SetSkipAccessChecksFlagOnAllMethods( 2227 GetInstructionSetPointerSize(manager_->GetCompiler()->GetInstructionSet())); 2228 klass->SetVerificationAttempted(); 2229 } 2230 // Record the final class status if necessary. 2231 ClassReference ref(manager_->GetDexFile(), class_def_index); 2232 manager_->GetCompiler()->RecordClassStatus(ref, klass->GetStatus()); 2233 } 2234 } else { 2235 Thread* self = soa.Self(); 2236 DCHECK(self->IsExceptionPending()); 2237 self->ClearException(); 2238 } 2239 } 2240 2241 private: 2242 const ParallelCompilationManager* const manager_; 2243}; 2244 2245void CompilerDriver::SetVerifiedDexFile(jobject class_loader, 2246 const DexFile& dex_file, 2247 const std::vector<const DexFile*>& dex_files, 2248 ThreadPool* thread_pool, 2249 size_t thread_count, 2250 TimingLogger* timings) { 2251 TimingLogger::ScopedTiming t("Verify Dex File", timings); 2252 ClassLinker* class_linker = Runtime::Current()->GetClassLinker(); 2253 ParallelCompilationManager context(class_linker, class_loader, this, &dex_file, dex_files, 2254 thread_pool); 2255 SetVerifiedClassVisitor visitor(&context); 2256 context.ForAll(0, dex_file.NumClassDefs(), &visitor, thread_count); 2257} 2258 2259class InitializeClassVisitor : public CompilationVisitor { 2260 public: 2261 explicit InitializeClassVisitor(const ParallelCompilationManager* manager) : manager_(manager) {} 2262 2263 virtual void Visit(size_t class_def_index) REQUIRES(!Locks::mutator_lock_) OVERRIDE { 2264 ATRACE_CALL(); 2265 jobject jclass_loader = manager_->GetClassLoader(); 2266 const DexFile& dex_file = *manager_->GetDexFile(); 2267 const DexFile::ClassDef& class_def = dex_file.GetClassDef(class_def_index); 2268 const DexFile::TypeId& class_type_id = dex_file.GetTypeId(class_def.class_idx_); 2269 const char* descriptor = dex_file.StringDataByIdx(class_type_id.descriptor_idx_); 2270 2271 ScopedObjectAccess soa(Thread::Current()); 2272 StackHandleScope<3> hs(soa.Self()); 2273 Handle<mirror::ClassLoader> class_loader( 2274 hs.NewHandle(soa.Decode<mirror::ClassLoader>(jclass_loader))); 2275 Handle<mirror::Class> klass( 2276 hs.NewHandle(manager_->GetClassLinker()->FindClass(soa.Self(), descriptor, class_loader))); 2277 2278 if (klass.Get() != nullptr && !SkipClass(jclass_loader, dex_file, klass.Get())) { 2279 // Only try to initialize classes that were successfully verified. 2280 if (klass->IsVerified()) { 2281 // Attempt to initialize the class but bail if we either need to initialize the super-class 2282 // or static fields. 2283 manager_->GetClassLinker()->EnsureInitialized(soa.Self(), klass, false, false); 2284 if (!klass->IsInitialized()) { 2285 // We don't want non-trivial class initialization occurring on multiple threads due to 2286 // deadlock problems. For example, a parent class is initialized (holding its lock) that 2287 // refers to a sub-class in its static/class initializer causing it to try to acquire the 2288 // sub-class' lock. While on a second thread the sub-class is initialized (holding its lock) 2289 // after first initializing its parents, whose locks are acquired. This leads to a 2290 // parent-to-child and a child-to-parent lock ordering and consequent potential deadlock. 2291 // We need to use an ObjectLock due to potential suspension in the interpreting code. Rather 2292 // than use a special Object for the purpose we use the Class of java.lang.Class. 2293 Handle<mirror::Class> h_klass(hs.NewHandle(klass->GetClass())); 2294 ObjectLock<mirror::Class> lock(soa.Self(), h_klass); 2295 // Attempt to initialize allowing initialization of parent classes but still not static 2296 // fields. 2297 manager_->GetClassLinker()->EnsureInitialized(soa.Self(), klass, false, true); 2298 if (!klass->IsInitialized()) { 2299 // We need to initialize static fields, we only do this for image classes that aren't 2300 // marked with the $NoPreloadHolder (which implies this should not be initialized early). 2301 bool can_init_static_fields = 2302 manager_->GetCompiler()->GetCompilerOptions().IsBootImage() && 2303 manager_->GetCompiler()->IsImageClass(descriptor) && 2304 !StringPiece(descriptor).ends_with("$NoPreloadHolder;"); 2305 if (can_init_static_fields) { 2306 VLOG(compiler) << "Initializing: " << descriptor; 2307 // TODO multithreading support. We should ensure the current compilation thread has 2308 // exclusive access to the runtime and the transaction. To achieve this, we could use 2309 // a ReaderWriterMutex but we're holding the mutator lock so we fail mutex sanity 2310 // checks in Thread::AssertThreadSuspensionIsAllowable. 2311 Runtime* const runtime = Runtime::Current(); 2312 Transaction transaction; 2313 2314 // Run the class initializer in transaction mode. 2315 runtime->EnterTransactionMode(&transaction); 2316 const mirror::Class::Status old_status = klass->GetStatus(); 2317 bool success = manager_->GetClassLinker()->EnsureInitialized(soa.Self(), klass, true, 2318 true); 2319 // TODO we detach transaction from runtime to indicate we quit the transactional 2320 // mode which prevents the GC from visiting objects modified during the transaction. 2321 // Ensure GC is not run so don't access freed objects when aborting transaction. 2322 2323 ScopedAssertNoThreadSuspension ants("Transaction end"); 2324 runtime->ExitTransactionMode(); 2325 2326 if (!success) { 2327 CHECK(soa.Self()->IsExceptionPending()); 2328 mirror::Throwable* exception = soa.Self()->GetException(); 2329 VLOG(compiler) << "Initialization of " << descriptor << " aborted because of " 2330 << exception->Dump(); 2331 std::ostream* file_log = manager_->GetCompiler()-> 2332 GetCompilerOptions().GetInitFailureOutput(); 2333 if (file_log != nullptr) { 2334 *file_log << descriptor << "\n"; 2335 *file_log << exception->Dump() << "\n"; 2336 } 2337 soa.Self()->ClearException(); 2338 transaction.Rollback(); 2339 CHECK_EQ(old_status, klass->GetStatus()) << "Previous class status not restored"; 2340 } 2341 } 2342 } 2343 soa.Self()->AssertNoPendingException(); 2344 } 2345 } 2346 // Record the final class status if necessary. 2347 ClassReference ref(manager_->GetDexFile(), class_def_index); 2348 manager_->GetCompiler()->RecordClassStatus(ref, klass->GetStatus()); 2349 } 2350 // Clear any class not found or verification exceptions. 2351 soa.Self()->ClearException(); 2352 } 2353 2354 private: 2355 const ParallelCompilationManager* const manager_; 2356}; 2357 2358void CompilerDriver::InitializeClasses(jobject jni_class_loader, 2359 const DexFile& dex_file, 2360 const std::vector<const DexFile*>& dex_files, 2361 TimingLogger* timings) { 2362 TimingLogger::ScopedTiming t("InitializeNoClinit", timings); 2363 2364 // Initialization allocates objects and needs to run single-threaded to be deterministic. 2365 bool force_determinism = GetCompilerOptions().IsForceDeterminism(); 2366 ThreadPool* init_thread_pool = force_determinism 2367 ? single_thread_pool_.get() 2368 : parallel_thread_pool_.get(); 2369 size_t init_thread_count = force_determinism ? 1U : parallel_thread_count_; 2370 2371 ClassLinker* class_linker = Runtime::Current()->GetClassLinker(); 2372 ParallelCompilationManager context(class_linker, jni_class_loader, this, &dex_file, dex_files, 2373 init_thread_pool); 2374 if (GetCompilerOptions().IsBootImage()) { 2375 // TODO: remove this when transactional mode supports multithreading. 2376 init_thread_count = 1U; 2377 } 2378 InitializeClassVisitor visitor(&context); 2379 context.ForAll(0, dex_file.NumClassDefs(), &visitor, init_thread_count); 2380} 2381 2382class InitializeArrayClassesAndCreateConflictTablesVisitor : public ClassVisitor { 2383 public: 2384 explicit InitializeArrayClassesAndCreateConflictTablesVisitor(VariableSizedHandleScope& hs) 2385 : hs_(hs) {} 2386 2387 virtual bool operator()(ObjPtr<mirror::Class> klass) OVERRIDE 2388 REQUIRES_SHARED(Locks::mutator_lock_) { 2389 if (Runtime::Current()->GetHeap()->ObjectIsInBootImageSpace(klass)) { 2390 return true; 2391 } 2392 if (klass->IsArrayClass()) { 2393 StackHandleScope<1> hs(Thread::Current()); 2394 auto h_klass = hs.NewHandleWrapper(&klass); 2395 Runtime::Current()->GetClassLinker()->EnsureInitialized(hs.Self(), h_klass, true, true); 2396 } 2397 // Collect handles since there may be thread suspension in future EnsureInitialized. 2398 to_visit_.push_back(hs_.NewHandle(klass)); 2399 return true; 2400 } 2401 2402 void FillAllIMTAndConflictTables() REQUIRES_SHARED(Locks::mutator_lock_) { 2403 for (Handle<mirror::Class> c : to_visit_) { 2404 // Create the conflict tables. 2405 FillIMTAndConflictTables(c.Get()); 2406 } 2407 } 2408 2409 private: 2410 void FillIMTAndConflictTables(ObjPtr<mirror::Class> klass) 2411 REQUIRES_SHARED(Locks::mutator_lock_) { 2412 if (!klass->ShouldHaveImt()) { 2413 return; 2414 } 2415 if (visited_classes_.find(klass) != visited_classes_.end()) { 2416 return; 2417 } 2418 if (klass->HasSuperClass()) { 2419 FillIMTAndConflictTables(klass->GetSuperClass()); 2420 } 2421 if (!klass->IsTemp()) { 2422 Runtime::Current()->GetClassLinker()->FillIMTAndConflictTables(klass); 2423 } 2424 visited_classes_.insert(klass); 2425 } 2426 2427 VariableSizedHandleScope& hs_; 2428 std::vector<Handle<mirror::Class>> to_visit_; 2429 std::unordered_set<ObjPtr<mirror::Class>, HashObjPtr> visited_classes_; 2430}; 2431 2432void CompilerDriver::InitializeClasses(jobject class_loader, 2433 const std::vector<const DexFile*>& dex_files, 2434 TimingLogger* timings) { 2435 for (size_t i = 0; i != dex_files.size(); ++i) { 2436 const DexFile* dex_file = dex_files[i]; 2437 CHECK(dex_file != nullptr); 2438 InitializeClasses(class_loader, *dex_file, dex_files, timings); 2439 } 2440 if (GetCompilerOptions().IsBootImage() || GetCompilerOptions().IsAppImage()) { 2441 // Make sure that we call EnsureIntiailized on all the array classes to call 2442 // SetVerificationAttempted so that the access flags are set. If we do not do this they get 2443 // changed at runtime resulting in more dirty image pages. 2444 // Also create conflict tables. 2445 // Only useful if we are compiling an image (image_classes_ is not null). 2446 ScopedObjectAccess soa(Thread::Current()); 2447 VariableSizedHandleScope hs(soa.Self()); 2448 InitializeArrayClassesAndCreateConflictTablesVisitor visitor(hs); 2449 Runtime::Current()->GetClassLinker()->VisitClassesWithoutClassesLock(&visitor); 2450 visitor.FillAllIMTAndConflictTables(); 2451 } 2452 if (GetCompilerOptions().IsBootImage()) { 2453 // Prune garbage objects created during aborted transactions. 2454 Runtime::Current()->GetHeap()->CollectGarbage(true); 2455 } 2456} 2457 2458void CompilerDriver::Compile(jobject class_loader, 2459 const std::vector<const DexFile*>& dex_files, 2460 TimingLogger* timings) { 2461 if (kDebugProfileGuidedCompilation) { 2462 LOG(INFO) << "[ProfileGuidedCompilation] " << 2463 ((profile_compilation_info_ == nullptr) 2464 ? "null" 2465 : profile_compilation_info_->DumpInfo(&dex_files)); 2466 } 2467 2468 DCHECK(current_dex_to_dex_methods_ == nullptr); 2469 for (const DexFile* dex_file : dex_files) { 2470 CHECK(dex_file != nullptr); 2471 CompileDexFile(class_loader, 2472 *dex_file, 2473 dex_files, 2474 parallel_thread_pool_.get(), 2475 parallel_thread_count_, 2476 timings); 2477 const ArenaPool* const arena_pool = Runtime::Current()->GetArenaPool(); 2478 const size_t arena_alloc = arena_pool->GetBytesAllocated(); 2479 max_arena_alloc_ = std::max(arena_alloc, max_arena_alloc_); 2480 Runtime::Current()->ReclaimArenaPoolMemory(); 2481 } 2482 2483 ArrayRef<DexFileMethodSet> dex_to_dex_references; 2484 { 2485 // From this point on, we shall not modify dex_to_dex_references_, so 2486 // just grab a reference to it that we use without holding the mutex. 2487 MutexLock lock(Thread::Current(), dex_to_dex_references_lock_); 2488 dex_to_dex_references = ArrayRef<DexFileMethodSet>(dex_to_dex_references_); 2489 } 2490 for (const auto& method_set : dex_to_dex_references) { 2491 current_dex_to_dex_methods_ = &method_set.GetMethodIndexes(); 2492 CompileDexFile(class_loader, 2493 method_set.GetDexFile(), 2494 dex_files, 2495 parallel_thread_pool_.get(), 2496 parallel_thread_count_, 2497 timings); 2498 } 2499 current_dex_to_dex_methods_ = nullptr; 2500 2501 VLOG(compiler) << "Compile: " << GetMemoryUsageString(false); 2502} 2503 2504class CompileClassVisitor : public CompilationVisitor { 2505 public: 2506 explicit CompileClassVisitor(const ParallelCompilationManager* manager) : manager_(manager) {} 2507 2508 virtual void Visit(size_t class_def_index) REQUIRES(!Locks::mutator_lock_) OVERRIDE { 2509 ATRACE_CALL(); 2510 const DexFile& dex_file = *manager_->GetDexFile(); 2511 const DexFile::ClassDef& class_def = dex_file.GetClassDef(class_def_index); 2512 ClassLinker* class_linker = manager_->GetClassLinker(); 2513 jobject jclass_loader = manager_->GetClassLoader(); 2514 ClassReference ref(&dex_file, class_def_index); 2515 // Skip compiling classes with generic verifier failures since they will still fail at runtime 2516 if (manager_->GetCompiler()->verification_results_->IsClassRejected(ref)) { 2517 return; 2518 } 2519 // Use a scoped object access to perform to the quick SkipClass check. 2520 const char* descriptor = dex_file.GetClassDescriptor(class_def); 2521 ScopedObjectAccess soa(Thread::Current()); 2522 StackHandleScope<3> hs(soa.Self()); 2523 Handle<mirror::ClassLoader> class_loader( 2524 hs.NewHandle(soa.Decode<mirror::ClassLoader>(jclass_loader))); 2525 Handle<mirror::Class> klass( 2526 hs.NewHandle(class_linker->FindClass(soa.Self(), descriptor, class_loader))); 2527 Handle<mirror::DexCache> dex_cache; 2528 if (klass.Get() == nullptr) { 2529 soa.Self()->AssertPendingException(); 2530 soa.Self()->ClearException(); 2531 dex_cache = hs.NewHandle(class_linker->FindDexCache(soa.Self(), dex_file)); 2532 } else if (SkipClass(jclass_loader, dex_file, klass.Get())) { 2533 return; 2534 } else { 2535 dex_cache = hs.NewHandle(klass->GetDexCache()); 2536 } 2537 2538 const uint8_t* class_data = dex_file.GetClassData(class_def); 2539 if (class_data == nullptr) { 2540 // empty class, probably a marker interface 2541 return; 2542 } 2543 2544 // Go to native so that we don't block GC during compilation. 2545 ScopedThreadSuspension sts(soa.Self(), kNative); 2546 2547 CompilerDriver* const driver = manager_->GetCompiler(); 2548 2549 // Can we run DEX-to-DEX compiler on this class ? 2550 optimizer::DexToDexCompilationLevel dex_to_dex_compilation_level = 2551 GetDexToDexCompilationLevel(soa.Self(), *driver, jclass_loader, dex_file, class_def); 2552 2553 ClassDataItemIterator it(dex_file, class_data); 2554 // Skip fields 2555 while (it.HasNextStaticField()) { 2556 it.Next(); 2557 } 2558 while (it.HasNextInstanceField()) { 2559 it.Next(); 2560 } 2561 2562 bool compilation_enabled = driver->IsClassToCompile( 2563 dex_file.StringByTypeIdx(class_def.class_idx_)); 2564 2565 // Compile direct methods 2566 int64_t previous_direct_method_idx = -1; 2567 while (it.HasNextDirectMethod()) { 2568 uint32_t method_idx = it.GetMemberIndex(); 2569 if (method_idx == previous_direct_method_idx) { 2570 // smali can create dex files with two encoded_methods sharing the same method_idx 2571 // http://code.google.com/p/smali/issues/detail?id=119 2572 it.Next(); 2573 continue; 2574 } 2575 previous_direct_method_idx = method_idx; 2576 CompileMethod(soa.Self(), driver, it.GetMethodCodeItem(), it.GetMethodAccessFlags(), 2577 it.GetMethodInvokeType(class_def), class_def_index, 2578 method_idx, jclass_loader, dex_file, dex_to_dex_compilation_level, 2579 compilation_enabled, dex_cache); 2580 it.Next(); 2581 } 2582 // Compile virtual methods 2583 int64_t previous_virtual_method_idx = -1; 2584 while (it.HasNextVirtualMethod()) { 2585 uint32_t method_idx = it.GetMemberIndex(); 2586 if (method_idx == previous_virtual_method_idx) { 2587 // smali can create dex files with two encoded_methods sharing the same method_idx 2588 // http://code.google.com/p/smali/issues/detail?id=119 2589 it.Next(); 2590 continue; 2591 } 2592 previous_virtual_method_idx = method_idx; 2593 CompileMethod(soa.Self(), driver, it.GetMethodCodeItem(), it.GetMethodAccessFlags(), 2594 it.GetMethodInvokeType(class_def), class_def_index, 2595 method_idx, jclass_loader, dex_file, dex_to_dex_compilation_level, 2596 compilation_enabled, dex_cache); 2597 it.Next(); 2598 } 2599 DCHECK(!it.HasNext()); 2600 } 2601 2602 private: 2603 const ParallelCompilationManager* const manager_; 2604}; 2605 2606void CompilerDriver::CompileDexFile(jobject class_loader, 2607 const DexFile& dex_file, 2608 const std::vector<const DexFile*>& dex_files, 2609 ThreadPool* thread_pool, 2610 size_t thread_count, 2611 TimingLogger* timings) { 2612 TimingLogger::ScopedTiming t("Compile Dex File", timings); 2613 ParallelCompilationManager context(Runtime::Current()->GetClassLinker(), class_loader, this, 2614 &dex_file, dex_files, thread_pool); 2615 CompileClassVisitor visitor(&context); 2616 context.ForAll(0, dex_file.NumClassDefs(), &visitor, thread_count); 2617} 2618 2619void CompilerDriver::AddCompiledMethod(const MethodReference& method_ref, 2620 CompiledMethod* const compiled_method, 2621 size_t non_relative_linker_patch_count) { 2622 DCHECK(GetCompiledMethod(method_ref) == nullptr) 2623 << method_ref.dex_file->PrettyMethod(method_ref.dex_method_index); 2624 MethodTable::InsertResult result = compiled_methods_.Insert(method_ref, 2625 /*expected*/ nullptr, 2626 compiled_method); 2627 CHECK(result == MethodTable::kInsertResultSuccess); 2628 non_relative_linker_patch_count_.FetchAndAddRelaxed(non_relative_linker_patch_count); 2629 DCHECK(GetCompiledMethod(method_ref) != nullptr) 2630 << method_ref.dex_file->PrettyMethod(method_ref.dex_method_index); 2631} 2632 2633CompiledClass* CompilerDriver::GetCompiledClass(ClassReference ref) const { 2634 MutexLock mu(Thread::Current(), compiled_classes_lock_); 2635 ClassTable::const_iterator it = compiled_classes_.find(ref); 2636 if (it == compiled_classes_.end()) { 2637 return nullptr; 2638 } 2639 CHECK(it->second != nullptr); 2640 return it->second; 2641} 2642 2643void CompilerDriver::RecordClassStatus(ClassReference ref, mirror::Class::Status status) { 2644 MutexLock mu(Thread::Current(), compiled_classes_lock_); 2645 auto it = compiled_classes_.find(ref); 2646 if (it == compiled_classes_.end() || it->second->GetStatus() != status) { 2647 // An entry doesn't exist or the status is lower than the new status. 2648 if (it != compiled_classes_.end()) { 2649 CHECK_GT(status, it->second->GetStatus()); 2650 delete it->second; 2651 } 2652 switch (status) { 2653 case mirror::Class::kStatusNotReady: 2654 case mirror::Class::kStatusError: 2655 case mirror::Class::kStatusRetryVerificationAtRuntime: 2656 case mirror::Class::kStatusVerified: 2657 case mirror::Class::kStatusInitialized: 2658 case mirror::Class::kStatusResolved: 2659 break; // Expected states. 2660 default: 2661 LOG(FATAL) << "Unexpected class status for class " 2662 << PrettyDescriptor(ref.first->GetClassDescriptor(ref.first->GetClassDef(ref.second))) 2663 << " of " << status; 2664 } 2665 CompiledClass* compiled_class = new CompiledClass(status); 2666 compiled_classes_.Overwrite(ref, compiled_class); 2667 } 2668} 2669 2670CompiledMethod* CompilerDriver::GetCompiledMethod(MethodReference ref) const { 2671 CompiledMethod* compiled_method = nullptr; 2672 compiled_methods_.Get(ref, &compiled_method); 2673 return compiled_method; 2674} 2675 2676bool CompilerDriver::IsMethodVerifiedWithoutFailures(uint32_t method_idx, 2677 uint16_t class_def_idx, 2678 const DexFile& dex_file) const { 2679 const VerifiedMethod* verified_method = GetVerifiedMethod(&dex_file, method_idx); 2680 if (verified_method != nullptr) { 2681 return !verified_method->HasVerificationFailures(); 2682 } 2683 2684 // If we can't find verification metadata, check if this is a system class (we trust that system 2685 // classes have their methods verified). If it's not, be conservative and assume the method 2686 // has not been verified successfully. 2687 2688 // TODO: When compiling the boot image it should be safe to assume that everything is verified, 2689 // even if methods are not found in the verification cache. 2690 const char* descriptor = dex_file.GetClassDescriptor(dex_file.GetClassDef(class_def_idx)); 2691 ClassLinker* class_linker = Runtime::Current()->GetClassLinker(); 2692 Thread* self = Thread::Current(); 2693 ScopedObjectAccess soa(self); 2694 bool is_system_class = class_linker->FindSystemClass(self, descriptor) != nullptr; 2695 if (!is_system_class) { 2696 self->ClearException(); 2697 } 2698 return is_system_class; 2699} 2700 2701size_t CompilerDriver::GetNonRelativeLinkerPatchCount() const { 2702 return non_relative_linker_patch_count_.LoadRelaxed(); 2703} 2704 2705void CompilerDriver::SetRequiresConstructorBarrier(Thread* self, 2706 const DexFile* dex_file, 2707 uint16_t class_def_index, 2708 bool requires) { 2709 WriterMutexLock mu(self, requires_constructor_barrier_lock_); 2710 requires_constructor_barrier_.emplace(ClassReference(dex_file, class_def_index), requires); 2711} 2712 2713bool CompilerDriver::RequiresConstructorBarrier(Thread* self, 2714 const DexFile* dex_file, 2715 uint16_t class_def_index) { 2716 ClassReference class_ref(dex_file, class_def_index); 2717 { 2718 ReaderMutexLock mu(self, requires_constructor_barrier_lock_); 2719 auto it = requires_constructor_barrier_.find(class_ref); 2720 if (it != requires_constructor_barrier_.end()) { 2721 return it->second; 2722 } 2723 } 2724 WriterMutexLock mu(self, requires_constructor_barrier_lock_); 2725 const bool requires = RequiresConstructorBarrier(*dex_file, class_def_index); 2726 requires_constructor_barrier_.emplace(class_ref, requires); 2727 return requires; 2728} 2729 2730std::string CompilerDriver::GetMemoryUsageString(bool extended) const { 2731 std::ostringstream oss; 2732 const gc::Heap* const heap = Runtime::Current()->GetHeap(); 2733 const size_t java_alloc = heap->GetBytesAllocated(); 2734 oss << "arena alloc=" << PrettySize(max_arena_alloc_) << " (" << max_arena_alloc_ << "B)"; 2735 oss << " java alloc=" << PrettySize(java_alloc) << " (" << java_alloc << "B)"; 2736#if defined(__BIONIC__) || defined(__GLIBC__) 2737 const struct mallinfo info = mallinfo(); 2738 const size_t allocated_space = static_cast<size_t>(info.uordblks); 2739 const size_t free_space = static_cast<size_t>(info.fordblks); 2740 oss << " native alloc=" << PrettySize(allocated_space) << " (" << allocated_space << "B)" 2741 << " free=" << PrettySize(free_space) << " (" << free_space << "B)"; 2742#endif 2743 compiled_method_storage_.DumpMemoryUsage(oss, extended); 2744 return oss.str(); 2745} 2746 2747bool CompilerDriver::MayInlineInternal(const DexFile* inlined_from, 2748 const DexFile* inlined_into) const { 2749 // We're not allowed to inline across dex files if we're the no-inline-from dex file. 2750 if (inlined_from != inlined_into && 2751 compiler_options_->GetNoInlineFromDexFile() != nullptr && 2752 ContainsElement(*compiler_options_->GetNoInlineFromDexFile(), inlined_from)) { 2753 return false; 2754 } 2755 2756 return true; 2757} 2758 2759void CompilerDriver::InitializeThreadPools() { 2760 size_t parallel_count = parallel_thread_count_ > 0 ? parallel_thread_count_ - 1 : 0; 2761 parallel_thread_pool_.reset( 2762 new ThreadPool("Compiler driver thread pool", parallel_count)); 2763 single_thread_pool_.reset(new ThreadPool("Single-threaded Compiler driver thread pool", 0)); 2764} 2765 2766void CompilerDriver::FreeThreadPools() { 2767 parallel_thread_pool_.reset(); 2768 single_thread_pool_.reset(); 2769} 2770 2771} // namespace art 2772