Searched refs:arena (Results 1 - 25 of 49) sorted by last modified time

12

/art/runtime/base/
H A Darena_allocator.cc122 for (const Arena* arena = first; arena != nullptr; arena = arena->next_) {
123 malloc_bytes += arena->Size();
124 lost_bytes += arena->RemainingSpace();
127 // The lost_bytes_adjustment is used to make up for the fact that the current arena
220 auto* arena = free_arenas_; local
222 delete arena;
254 for (auto* arena
[all...]
H A Darena_allocator_test.cc25 size_t NumberOfArenas(ArenaAllocator* arena) { argument
27 for (Arena* a = arena->arena_head_; a != nullptr; a = a->next_) {
36 ArenaAllocator arena(&pool);
37 ArenaBitVector bv(&arena, 10, true);
51 // Allocate a small array from an arena and release it.
52 ArenaAllocator arena(&pool);
53 small_array = arena.AllocArray<uint32_t>(kSmallArraySize);
57 // Reuse the previous arena and allocate more than previous allocation including red zone.
58 ArenaAllocator arena(&pool);
59 uint32_t* large_array = arena
[all...]
H A Darena_bit_vector.cc55 static ArenaBitVectorAllocator* Create(ArenaAlloc* arena, ArenaAllocKind kind) { argument
56 void* storage = arena->template Alloc<ArenaBitVectorAllocator>(kind);
57 return new (storage) ArenaBitVectorAllocator(arena, kind);
72 ArenaBitVectorAllocator(ArenaAlloc* arena, ArenaAllocKind kind) argument
73 : ArenaBitVectorAllocatorKind(kind), arena_(arena) { }
80 ArenaBitVector::ArenaBitVector(ArenaAllocator* arena, argument
86 ArenaBitVectorAllocator<ArenaAllocator>::Create(arena, kind)) {
89 ArenaBitVector::ArenaBitVector(ScopedArenaAllocator* arena, argument
95 ArenaBitVectorAllocator<ScopedArenaAllocator>::Create(arena, kind)) {
H A Darena_bit_vector.h34 static ArenaBitVector* Create(Allocator* arena, argument
38 void* storage = arena->template Alloc<ArenaBitVector>(kind);
39 return new (storage) ArenaBitVector(arena, start_bits, expandable, kind);
42 ArenaBitVector(ArenaAllocator* arena,
46 ArenaBitVector(ScopedArenaAllocator* arena,
H A Darena_object.h26 // Parent for arena allocated objects giving appropriate new and delete operators.
35 static void* operator new(size_t size, ScopedArenaAllocator* arena) { argument
36 return arena->Alloc(size, kAllocKind);
50 // Parent for arena allocated objects that get deleted, gives appropriate new and delete operators.
59 static void* operator new(size_t size, ScopedArenaAllocator* arena) { argument
60 return arena->Alloc(size, kAllocKind);
/art/runtime/verifier/
H A Dmethod_verifier.cc65 PcToRegisterLineTable::PcToRegisterLineTable(ScopedArenaAllocator& arena) argument
66 : register_lines_(arena.Adapter(kArenaAllocVerifier)) {}
H A Dmethod_verifier.h117 explicit PcToRegisterLineTable(ScopedArenaAllocator& arena);
H A Dreg_type-inl.h188 inline void* RegType::operator new(size_t size, ScopedArenaAllocator* arena) { argument
189 return arena->Alloc(size, kArenaAllocMisc);
H A Dreg_type.h276 static void* operator new(size_t size, ArenaAllocator* arena) = delete;
277 static void* operator new(size_t size, ScopedArenaAllocator* arena);
H A Dreg_type_cache.cc266 RegTypeCache::RegTypeCache(bool can_load_classes, ScopedArenaAllocator& arena) argument
267 : entries_(arena.Adapter(kArenaAllocVerifier)),
268 klass_entries_(arena.Adapter(kArenaAllocVerifier)),
270 arena_(arena) {
H A Dreg_type_cache.h42 // Use 8 bytes since that is the default arena allocator alignment.
47 explicit RegTypeCache(bool can_load_classes, ScopedArenaAllocator& arena);
169 // Add a string piece to the arena allocator so that it stays live for the lifetime of the
/art/compiler/jni/
H A Djni_cfi_test.cc48 ArenaAllocator arena(&pool);
51 JniCallingConvention::Create(&arena, is_static, is_synchronized, shorty, isa));
53 ManagedRuntimeCallingConvention::Create(&arena, is_static, is_synchronized, shorty, isa));
58 std::unique_ptr<Assembler> jni_asm(Assembler::Create(&arena, isa));
/art/compiler/jni/quick/
H A Dcalling_convention.cc50 ArenaAllocator* arena,
60 new (arena) arm::ArmManagedRuntimeCallingConvention(is_static, is_synchronized, shorty));
65 new (arena) arm64::Arm64ManagedRuntimeCallingConvention(
71 new (arena) mips::MipsManagedRuntimeCallingConvention(
77 new (arena) mips64::Mips64ManagedRuntimeCallingConvention(
83 new (arena) x86::X86ManagedRuntimeCallingConvention(is_static, is_synchronized, shorty));
88 new (arena) x86_64::X86_64ManagedRuntimeCallingConvention(
149 std::unique_ptr<JniCallingConvention> JniCallingConvention::Create(ArenaAllocator* arena, argument
159 new (arena) arm::ArmJniCallingConvention(is_static, is_synchronized, shorty));
164 new (arena) arm6
49 Create( ArenaAllocator* arena, bool is_static, bool is_synchronized, const char* shorty, InstructionSet instruction_set) argument
[all...]
H A Dcalling_convention.h226 static std::unique_ptr<ManagedRuntimeCallingConvention> Create(ArenaAllocator* arena,
279 static std::unique_ptr<JniCallingConvention> Create(ArenaAllocator* arena,
H A Djni_compiler.cc75 ArenaAllocator arena(&pool);
79 JniCallingConvention::Create(&arena, is_static, is_synchronized, shorty, instruction_set));
84 &arena, is_static, is_synchronized, shorty, instruction_set));
100 &arena, is_static, is_synchronized, jni_end_shorty, instruction_set));
104 Assembler::Create(&arena, instruction_set, instruction_set_features));
/art/compiler/linker/arm/
H A Drelative_patcher_thumb2.cc83 ArenaAllocator arena(&pool);
84 arm::Thumb2Assembler assembler(&arena);
/art/compiler/linker/arm64/
H A Drelative_patcher_arm64.cc251 ArenaAllocator arena(&pool);
252 arm64::Arm64Assembler assembler(&arena);
/art/compiler/optimizing/
H A Dcode_generator.cc561 ArenaAllocator* arena = graph->GetArena(); local
567 new (arena) arm::CodeGeneratorARM(graph,
576 new (arena) arm64::CodeGeneratorARM64(graph,
585 new (arena) mips::CodeGeneratorMIPS(graph,
594 new (arena) mips64::CodeGeneratorMIPS64(graph,
603 new (arena) x86::CodeGeneratorX86(graph,
612 new (arena) x86_64::CodeGeneratorX86_64(graph,
833 ArenaAllocator* arena = graph_->GetArena(); local
848 ArenaBitVector::Create(arena, 0, /* expandable */ true, kArenaAllocCodeGenerator);
H A Dcodegen_test.cc297 ArenaAllocator arena(&pool);
298 HGraph* graph = CreateCFG(&arena, data);
310 ArenaAllocator arena(&pool);
311 HGraph* graph = CreateCFG(&arena, data, Primitive::kPrimLong);
H A Dfind_loops_test.cc38 ArenaPool arena; local
39 ArenaAllocator allocator(&arena);
51 ArenaPool arena; local
52 ArenaAllocator allocator(&arena);
67 ArenaPool arena; local
68 ArenaAllocator allocator(&arena);
84 ArenaPool arena; local
85 ArenaAllocator allocator(&arena);
99 ArenaPool arena; local
100 ArenaAllocator allocator(&arena);
145 ArenaPool arena; local
173 ArenaPool arena; local
198 ArenaPool arena; local
224 ArenaPool arena; local
250 ArenaPool arena; local
275 ArenaPool arena; local
306 ArenaPool arena; local
336 ArenaPool arena; local
352 ArenaPool arena; local
[all...]
H A Dinstruction_simplifier.cc557 static HCondition* GetOppositeConditionSwapOps(ArenaAllocator* arena, HInstruction* cond) { argument
562 return new (arena) HEqual(rhs, lhs);
564 return new (arena) HNotEqual(rhs, lhs);
566 return new (arena) HGreaterThan(rhs, lhs);
568 return new (arena) HGreaterThanOrEqual(rhs, lhs);
570 return new (arena) HLessThan(rhs, lhs);
572 return new (arena) HLessThanOrEqual(rhs, lhs);
574 return new (arena) HAbove(rhs, lhs);
576 return new (arena) HAboveOrEqual(rhs, lhs);
578 return new (arena) HBelo
[all...]
H A Dinstruction_simplifier_arm64.cc56 ArenaAllocator* arena = GetGraph()->GetArena(); local
61 new (arena) HArm64IntermediateAddress(array, offset, kNoDexPc);
H A Dinstruction_simplifier_shared.cc76 ArenaAllocator* arena = mul->GetBlock()->GetGraph()->GetArena(); local
77 HMultiplyAccumulate* mulacc = new(arena) HMultiplyAccumulate(
106 ArenaAllocator* arena = mul->GetBlock()->GetGraph()->GetArena(); local
138 new (arena) HMultiplyAccumulate(type,
151 new (arena) HMultiplyAccumulate(type,
H A Dintrinsics_arm.cc68 static void CreateFPToIntLocations(ArenaAllocator* arena, HInvoke* invoke) { argument
69 LocationSummary* locations = new (arena) LocationSummary(invoke,
76 static void CreateIntToFPLocations(ArenaAllocator* arena, HInvoke* invoke) { argument
77 LocationSummary* locations = new (arena) LocationSummary(invoke,
136 static void CreateIntToIntLocations(ArenaAllocator* arena, HInvoke* invoke) { argument
137 LocationSummary* locations = new (arena) LocationSummary(invoke,
144 static void CreateFPToFPLocations(ArenaAllocator* arena, HInvoke* invoke) { argument
145 LocationSummary* locations = new (arena) LocationSummary(invoke,
271 static void CreateIntToIntPlusTemp(ArenaAllocator* arena, HInvoke* invoke) { argument
272 LocationSummary* locations = new (arena) LocationSummar
343 CreateIntIntToIntLocations(ArenaAllocator* arena, HInvoke* invoke) argument
433 CreateIntIntToVoidLocations(ArenaAllocator* arena, HInvoke* invoke) argument
570 CreateIntIntIntToIntLocations(ArenaAllocator* arena, HInvoke* invoke, Primitive::Type type) argument
630 CreateIntIntIntIntToVoid(ArenaAllocator* arena, const ArmInstructionSetFeatures& features, Primitive::Type type, bool is_volatile, HInvoke* invoke) argument
809 CreateIntIntIntIntIntToIntPlusTemps(ArenaAllocator* arena, HInvoke* invoke, Primitive::Type type) argument
1603 CreateFPToFPCallLocations(ArenaAllocator* arena, HInvoke* invoke) argument
1629 CreateFPFPToFPCallLocations(ArenaAllocator* arena, HInvoke* invoke) argument
[all...]
H A Dintrinsics_arm.h36 IntrinsicLocationsBuilderARM(ArenaAllocator* arena, argument
39 : arena_(arena), assembler_(assembler), features_(features) {}

Completed in 251 milliseconds

12