/art/compiler/optimizing/ |
H A D | intrinsics_arm.h | 80 ArenaAllocator* GetAllocator();
|
H A D | intrinsics_arm64.h | 78 ArenaAllocator* GetAllocator();
|
H A D | intrinsics_mips.h | 77 ArenaAllocator* GetAllocator();
|
H A D | intrinsics_mips64.h | 73 ArenaAllocator* GetAllocator();
|
H A D | intrinsics_x86.h | 74 ArenaAllocator* GetAllocator();
|
H A D | intrinsics_x86_64.h | 74 ArenaAllocator* GetAllocator();
|
H A D | intrinsics_mips64.cc | 42 ArenaAllocator* IntrinsicCodeGeneratorMIPS64::GetAllocator() { function in class:art::mips64::IntrinsicCodeGeneratorMIPS64 1317 SlowPathCodeMIPS64* slow_path = new (GetAllocator()) IntrinsicSlowPathMIPS64(invoke); 1354 SlowPathCodeMIPS64* slow_path = new (GetAllocator()) IntrinsicSlowPathMIPS64(invoke); 1539 GenerateStringIndexOf(invoke, GetAssembler(), codegen_, GetAllocator(), /* start_at_zero */ true); 1559 invoke, GetAssembler(), codegen_, GetAllocator(), /* start_at_zero */ false); 1581 SlowPathCodeMIPS64* slow_path = new (GetAllocator()) IntrinsicSlowPathMIPS64(invoke); 1646 SlowPathCodeMIPS64* slow_path = new (GetAllocator()) IntrinsicSlowPathMIPS64(invoke);
|
H A D | intrinsics_arm64.cc | 63 ArenaAllocator* IntrinsicCodeGeneratorARM64::GetAllocator() { function in class:art::arm64::IntrinsicCodeGeneratorARM64 1158 SlowPathCodeARM64* slow_path = new (GetAllocator()) IntrinsicSlowPathARM64(invoke); 1193 SlowPathCodeARM64* slow_path = new (GetAllocator()) IntrinsicSlowPathARM64(invoke); 1363 invoke, GetVIXLAssembler(), codegen_, GetAllocator(), /* start_at_zero */ true); 1384 invoke, GetVIXLAssembler(), codegen_, GetAllocator(), /* start_at_zero */ false); 1405 SlowPathCodeARM64* slow_path = new (GetAllocator()) IntrinsicSlowPathARM64(invoke); 1459 SlowPathCodeARM64* slow_path = new (GetAllocator()) IntrinsicSlowPathARM64(invoke); 1860 SlowPathCodeARM64* slow_path = new (GetAllocator()) IntrinsicSlowPathARM64(invoke); 2014 SlowPathCodeARM64* slow_path = new (GetAllocator()) IntrinsicSlowPathARM64(invoke);
|
H A D | intrinsics_arm.cc | 38 ArenaAllocator* IntrinsicCodeGeneratorARM::GetAllocator() { function in class:art::arm::IntrinsicCodeGeneratorARM 971 SlowPathCode* slow_path = new (GetAllocator()) IntrinsicSlowPathARM(invoke); 1007 SlowPathCode* slow_path = new (GetAllocator()) IntrinsicSlowPathARM(invoke); 1179 invoke, GetAssembler(), codegen_, GetAllocator(), /* start_at_zero */ true); 1200 invoke, GetAssembler(), codegen_, GetAllocator(), /* start_at_zero */ false); 1221 SlowPathCode* slow_path = new (GetAllocator()) IntrinsicSlowPathARM(invoke); 1275 SlowPathCode* slow_path = new (GetAllocator()) IntrinsicSlowPathARM(invoke); 1394 SlowPathCode* slow_path = new (GetAllocator()) IntrinsicSlowPathARM(invoke);
|
H A D | bounds_check_elimination.cc | 315 ArenaAllocator* GetAllocator() const { return allocator_; } function in class:art::ValueRange 469 return new (GetAllocator()) ValueRange(GetAllocator(), lower, range->GetUpper()); 489 return new (GetAllocator()) ValueRange(GetAllocator(), range->GetLower(), upper);
|
H A D | intrinsics_x86_64.cc | 47 ArenaAllocator* IntrinsicCodeGeneratorX86_64::GetAllocator() { function in class:art::x86_64::IntrinsicCodeGeneratorX86_64 922 SlowPathCode* slow_path = new (GetAllocator()) IntrinsicSlowPathX86_64(invoke); 1060 SlowPathCode* slow_path = new (GetAllocator()) IntrinsicSlowPathX86_64(invoke); 1150 SlowPathCode* slow_path = new (GetAllocator()) IntrinsicSlowPathX86_64(invoke); 1378 SlowPathCode* slow_path = new (GetAllocator()) IntrinsicSlowPathX86_64(invoke); 1615 GenerateStringIndexOf(invoke, GetAssembler(), codegen_, GetAllocator(), /* start_at_zero */ true); 1624 invoke, GetAssembler(), codegen_, GetAllocator(), /* start_at_zero */ false); 1645 SlowPathCode* slow_path = new (GetAllocator()) IntrinsicSlowPathX86_64(invoke); 1697 SlowPathCode* slow_path = new (GetAllocator()) IntrinsicSlowPathX86_64(invoke);
|
H A D | intrinsics_x86.cc | 53 ArenaAllocator* IntrinsicCodeGeneratorX86::GetAllocator() { function in class:art::x86::IntrinsicCodeGeneratorX86 1060 SlowPathCode* slow_path = new (GetAllocator()) IntrinsicSlowPathX86(invoke); 1188 SlowPathCode* slow_path = new (GetAllocator()) IntrinsicSlowPathX86(invoke); 1272 SlowPathCode* slow_path = new (GetAllocator()) IntrinsicSlowPathX86(invoke); 1514 GenerateStringIndexOf(invoke, GetAssembler(), codegen_, GetAllocator(), /* start_at_zero */ true); 1523 invoke, GetAssembler(), codegen_, GetAllocator(), /* start_at_zero */ false); 1544 SlowPathCode* slow_path = new (GetAllocator()) IntrinsicSlowPathX86(invoke); 1594 SlowPathCode* slow_path = new (GetAllocator()) IntrinsicSlowPathX86(invoke);
|
H A D | intrinsics_mips.cc | 42 ArenaAllocator* IntrinsicCodeGeneratorMIPS::GetAllocator() { function in class:art::mips::IntrinsicCodeGeneratorMIPS 1906 SlowPathCodeMIPS* slow_path = new (GetAllocator()) IntrinsicSlowPathMIPS(invoke); 1943 SlowPathCodeMIPS* slow_path = new (GetAllocator()) IntrinsicSlowPathMIPS(invoke); 2140 GetAllocator()); 2166 GetAllocator()); 2188 SlowPathCodeMIPS* slow_path = new (GetAllocator()) IntrinsicSlowPathMIPS(invoke); 2250 SlowPathCodeMIPS* slow_path = new (GetAllocator()) IntrinsicSlowPathMIPS(invoke);
|
/art/runtime/mirror/ |
H A D | class_loader.h | 53 LinearAlloc* GetAllocator() SHARED_REQUIRES(Locks::mutator_lock_) {
|
H A D | dex_cache_test.cc | 60 LinearAlloc* const linear_alloc = klass->GetClassLoader()->GetAllocator();
|
/art/runtime/base/ |
H A D | bit_vector.h | 274 Allocator* GetAllocator() const;
|
H A D | bit_vector.cc | 371 Allocator* BitVector::GetAllocator() const { function in class:art::BitVector
|
/art/runtime/verifier/ |
H A D | reg_type.cc | 805 unresolved_types_(unresolved, false, unresolved.GetAllocator()) {
|
/art/runtime/ |
H A D | class_linker.cc | 3013 LinearAlloc* allocator = class_loader->GetAllocator(); 3023 LinearAlloc* allocator = class_loader->GetAllocator(); 3026 allocator = class_loader->GetAllocator(); 5061 CHECK(class_loader->GetAllocator() == nullptr);
|