/art/runtime/base/ |
H A D | arena_allocator.cc | 122 for (const Arena* arena = first; arena != nullptr; arena = arena->next_) { 123 malloc_bytes += arena->Size(); 124 lost_bytes += arena->RemainingSpace(); 127 // The lost_bytes_adjustment is used to make up for the fact that the current arena 220 auto* arena = free_arenas_; local 222 delete arena; 254 for (auto* arena [all...] |
H A D | arena_allocator_test.cc | 25 size_t NumberOfArenas(ArenaAllocator* arena) { argument 27 for (Arena* a = arena->arena_head_; a != nullptr; a = a->next_) { 36 ArenaAllocator arena(&pool); 37 ArenaBitVector bv(&arena, 10, true); 51 // Allocate a small array from an arena and release it. 52 ArenaAllocator arena(&pool); 53 small_array = arena.AllocArray<uint32_t>(kSmallArraySize); 57 // Reuse the previous arena and allocate more than previous allocation including red zone. 58 ArenaAllocator arena(&pool); 59 uint32_t* large_array = arena [all...] |
H A D | arena_bit_vector.cc | 55 static ArenaBitVectorAllocator* Create(ArenaAlloc* arena, ArenaAllocKind kind) { argument 56 void* storage = arena->template Alloc<ArenaBitVectorAllocator>(kind); 57 return new (storage) ArenaBitVectorAllocator(arena, kind); 72 ArenaBitVectorAllocator(ArenaAlloc* arena, ArenaAllocKind kind) argument 73 : ArenaBitVectorAllocatorKind(kind), arena_(arena) { } 80 ArenaBitVector::ArenaBitVector(ArenaAllocator* arena, argument 86 ArenaBitVectorAllocator<ArenaAllocator>::Create(arena, kind)) { 89 ArenaBitVector::ArenaBitVector(ScopedArenaAllocator* arena, argument 95 ArenaBitVectorAllocator<ScopedArenaAllocator>::Create(arena, kind)) {
|
H A D | arena_bit_vector.h | 34 static ArenaBitVector* Create(Allocator* arena, argument 38 void* storage = arena->template Alloc<ArenaBitVector>(kind); 39 return new (storage) ArenaBitVector(arena, start_bits, expandable, kind); 42 ArenaBitVector(ArenaAllocator* arena, 46 ArenaBitVector(ScopedArenaAllocator* arena,
|
H A D | arena_object.h | 26 // Parent for arena allocated objects giving appropriate new and delete operators. 35 static void* operator new(size_t size, ScopedArenaAllocator* arena) { argument 36 return arena->Alloc(size, kAllocKind); 50 // Parent for arena allocated objects that get deleted, gives appropriate new and delete operators. 59 static void* operator new(size_t size, ScopedArenaAllocator* arena) { argument 60 return arena->Alloc(size, kAllocKind);
|
/art/runtime/verifier/ |
H A D | method_verifier.cc | 65 PcToRegisterLineTable::PcToRegisterLineTable(ScopedArenaAllocator& arena) argument 66 : register_lines_(arena.Adapter(kArenaAllocVerifier)) {}
|
H A D | method_verifier.h | 117 explicit PcToRegisterLineTable(ScopedArenaAllocator& arena);
|
H A D | reg_type-inl.h | 188 inline void* RegType::operator new(size_t size, ScopedArenaAllocator* arena) { argument 189 return arena->Alloc(size, kArenaAllocMisc);
|
H A D | reg_type.h | 276 static void* operator new(size_t size, ArenaAllocator* arena) = delete; 277 static void* operator new(size_t size, ScopedArenaAllocator* arena);
|
H A D | reg_type_cache.cc | 266 RegTypeCache::RegTypeCache(bool can_load_classes, ScopedArenaAllocator& arena) argument 267 : entries_(arena.Adapter(kArenaAllocVerifier)), 268 klass_entries_(arena.Adapter(kArenaAllocVerifier)), 270 arena_(arena) {
|
H A D | reg_type_cache.h | 42 // Use 8 bytes since that is the default arena allocator alignment. 47 explicit RegTypeCache(bool can_load_classes, ScopedArenaAllocator& arena); 169 // Add a string piece to the arena allocator so that it stays live for the lifetime of the
|
/art/compiler/jni/ |
H A D | jni_cfi_test.cc | 48 ArenaAllocator arena(&pool); 51 JniCallingConvention::Create(&arena, is_static, is_synchronized, shorty, isa)); 53 ManagedRuntimeCallingConvention::Create(&arena, is_static, is_synchronized, shorty, isa)); 58 std::unique_ptr<Assembler> jni_asm(Assembler::Create(&arena, isa));
|
/art/compiler/jni/quick/ |
H A D | calling_convention.cc | 50 ArenaAllocator* arena, 60 new (arena) arm::ArmManagedRuntimeCallingConvention(is_static, is_synchronized, shorty)); 65 new (arena) arm64::Arm64ManagedRuntimeCallingConvention( 71 new (arena) mips::MipsManagedRuntimeCallingConvention( 77 new (arena) mips64::Mips64ManagedRuntimeCallingConvention( 83 new (arena) x86::X86ManagedRuntimeCallingConvention(is_static, is_synchronized, shorty)); 88 new (arena) x86_64::X86_64ManagedRuntimeCallingConvention( 149 std::unique_ptr<JniCallingConvention> JniCallingConvention::Create(ArenaAllocator* arena, argument 159 new (arena) arm::ArmJniCallingConvention(is_static, is_synchronized, shorty)); 164 new (arena) arm6 49 Create( ArenaAllocator* arena, bool is_static, bool is_synchronized, const char* shorty, InstructionSet instruction_set) argument [all...] |
H A D | calling_convention.h | 226 static std::unique_ptr<ManagedRuntimeCallingConvention> Create(ArenaAllocator* arena, 279 static std::unique_ptr<JniCallingConvention> Create(ArenaAllocator* arena,
|
H A D | jni_compiler.cc | 75 ArenaAllocator arena(&pool); 79 JniCallingConvention::Create(&arena, is_static, is_synchronized, shorty, instruction_set)); 84 &arena, is_static, is_synchronized, shorty, instruction_set)); 100 &arena, is_static, is_synchronized, jni_end_shorty, instruction_set)); 104 Assembler::Create(&arena, instruction_set, instruction_set_features));
|
/art/compiler/linker/arm/ |
H A D | relative_patcher_thumb2.cc | 83 ArenaAllocator arena(&pool); 84 arm::Thumb2Assembler assembler(&arena);
|
/art/compiler/linker/arm64/ |
H A D | relative_patcher_arm64.cc | 251 ArenaAllocator arena(&pool); 252 arm64::Arm64Assembler assembler(&arena);
|
/art/compiler/optimizing/ |
H A D | code_generator.cc | 561 ArenaAllocator* arena = graph->GetArena(); local 567 new (arena) arm::CodeGeneratorARM(graph, 576 new (arena) arm64::CodeGeneratorARM64(graph, 585 new (arena) mips::CodeGeneratorMIPS(graph, 594 new (arena) mips64::CodeGeneratorMIPS64(graph, 603 new (arena) x86::CodeGeneratorX86(graph, 612 new (arena) x86_64::CodeGeneratorX86_64(graph, 833 ArenaAllocator* arena = graph_->GetArena(); local 848 ArenaBitVector::Create(arena, 0, /* expandable */ true, kArenaAllocCodeGenerator);
|
H A D | codegen_test.cc | 297 ArenaAllocator arena(&pool); 298 HGraph* graph = CreateCFG(&arena, data); 310 ArenaAllocator arena(&pool); 311 HGraph* graph = CreateCFG(&arena, data, Primitive::kPrimLong);
|
H A D | find_loops_test.cc | 38 ArenaPool arena; local 39 ArenaAllocator allocator(&arena); 51 ArenaPool arena; local 52 ArenaAllocator allocator(&arena); 67 ArenaPool arena; local 68 ArenaAllocator allocator(&arena); 84 ArenaPool arena; local 85 ArenaAllocator allocator(&arena); 99 ArenaPool arena; local 100 ArenaAllocator allocator(&arena); 145 ArenaPool arena; local 173 ArenaPool arena; local 198 ArenaPool arena; local 224 ArenaPool arena; local 250 ArenaPool arena; local 275 ArenaPool arena; local 306 ArenaPool arena; local 336 ArenaPool arena; local 352 ArenaPool arena; local [all...] |
H A D | instruction_simplifier.cc | 557 static HCondition* GetOppositeConditionSwapOps(ArenaAllocator* arena, HInstruction* cond) { argument 562 return new (arena) HEqual(rhs, lhs); 564 return new (arena) HNotEqual(rhs, lhs); 566 return new (arena) HGreaterThan(rhs, lhs); 568 return new (arena) HGreaterThanOrEqual(rhs, lhs); 570 return new (arena) HLessThan(rhs, lhs); 572 return new (arena) HLessThanOrEqual(rhs, lhs); 574 return new (arena) HAbove(rhs, lhs); 576 return new (arena) HAboveOrEqual(rhs, lhs); 578 return new (arena) HBelo [all...] |
H A D | instruction_simplifier_arm64.cc | 56 ArenaAllocator* arena = GetGraph()->GetArena(); local 61 new (arena) HArm64IntermediateAddress(array, offset, kNoDexPc);
|
H A D | instruction_simplifier_shared.cc | 76 ArenaAllocator* arena = mul->GetBlock()->GetGraph()->GetArena(); local 77 HMultiplyAccumulate* mulacc = new(arena) HMultiplyAccumulate( 106 ArenaAllocator* arena = mul->GetBlock()->GetGraph()->GetArena(); local 138 new (arena) HMultiplyAccumulate(type, 151 new (arena) HMultiplyAccumulate(type,
|
H A D | intrinsics_arm.cc | 68 static void CreateFPToIntLocations(ArenaAllocator* arena, HInvoke* invoke) { argument 69 LocationSummary* locations = new (arena) LocationSummary(invoke, 76 static void CreateIntToFPLocations(ArenaAllocator* arena, HInvoke* invoke) { argument 77 LocationSummary* locations = new (arena) LocationSummary(invoke, 136 static void CreateIntToIntLocations(ArenaAllocator* arena, HInvoke* invoke) { argument 137 LocationSummary* locations = new (arena) LocationSummary(invoke, 144 static void CreateFPToFPLocations(ArenaAllocator* arena, HInvoke* invoke) { argument 145 LocationSummary* locations = new (arena) LocationSummary(invoke, 271 static void CreateIntToIntPlusTemp(ArenaAllocator* arena, HInvoke* invoke) { argument 272 LocationSummary* locations = new (arena) LocationSummar 343 CreateIntIntToIntLocations(ArenaAllocator* arena, HInvoke* invoke) argument 433 CreateIntIntToVoidLocations(ArenaAllocator* arena, HInvoke* invoke) argument 570 CreateIntIntIntToIntLocations(ArenaAllocator* arena, HInvoke* invoke, Primitive::Type type) argument 630 CreateIntIntIntIntToVoid(ArenaAllocator* arena, const ArmInstructionSetFeatures& features, Primitive::Type type, bool is_volatile, HInvoke* invoke) argument 809 CreateIntIntIntIntIntToIntPlusTemps(ArenaAllocator* arena, HInvoke* invoke, Primitive::Type type) argument 1603 CreateFPToFPCallLocations(ArenaAllocator* arena, HInvoke* invoke) argument 1629 CreateFPFPToFPCallLocations(ArenaAllocator* arena, HInvoke* invoke) argument [all...] |
H A D | intrinsics_arm.h | 36 IntrinsicLocationsBuilderARM(ArenaAllocator* arena, argument 39 : arena_(arena), assembler_(assembler), features_(features) {}
|