/art/compiler/utils/arm/ |
H A D | assembler_arm32.h | 33 explicit Arm32Assembler(ArenaAllocator* arena) : ArmAssembler(arena) {} argument
|
H A D | assembler_thumb2.h | 37 explicit Thumb2Assembler(ArenaAllocator* arena, bool can_relocate_branches = true) argument 38 : ArmAssembler(arena), 43 fixups_(arena->Adapter(kArenaAllocAssembler)), 44 fixup_dependents_(arena->Adapter(kArenaAllocAssembler)), 45 literals_(arena->Adapter(kArenaAllocAssembler)), 46 jump_tables_(arena->Adapter(kArenaAllocAssembler)),
|
/art/compiler/optimizing/ |
H A D | intrinsics_arm.h | 36 IntrinsicLocationsBuilderARM(ArenaAllocator* arena, argument 39 : arena_(arena), assembler_(assembler), features_(features) {}
|
H A D | intrinsics_arm64.h | 40 explicit IntrinsicLocationsBuilderARM64(ArenaAllocator* arena) : arena_(arena) {} argument
|
H A D | find_loops_test.cc | 38 ArenaPool arena; local 39 ArenaAllocator allocator(&arena); 51 ArenaPool arena; local 52 ArenaAllocator allocator(&arena); 67 ArenaPool arena; local 68 ArenaAllocator allocator(&arena); 84 ArenaPool arena; local 85 ArenaAllocator allocator(&arena); 99 ArenaPool arena; local 100 ArenaAllocator allocator(&arena); 145 ArenaPool arena; local 173 ArenaPool arena; local 198 ArenaPool arena; local 224 ArenaPool arena; local 250 ArenaPool arena; local 275 ArenaPool arena; local 306 ArenaPool arena; local 336 ArenaPool arena; local 352 ArenaPool arena; local [all...] |
H A D | locations.cc | 39 ArenaAllocator* arena = instruction->GetBlock()->GetGraph()->GetArena(); local 40 stack_mask_ = ArenaBitVector::Create(arena, 0, true, kArenaAllocLocationSummary);
|
H A D | instruction_simplifier_shared.cc | 76 ArenaAllocator* arena = mul->GetBlock()->GetGraph()->GetArena(); local 77 HMultiplyAccumulate* mulacc = new(arena) HMultiplyAccumulate( 106 ArenaAllocator* arena = mul->GetBlock()->GetGraph()->GetArena(); local 138 new (arena) HMultiplyAccumulate(type, 151 new (arena) HMultiplyAccumulate(type,
|
H A D | instruction_simplifier_arm64.cc | 56 ArenaAllocator* arena = GetGraph()->GetArena(); local 61 new (arena) HArm64IntermediateAddress(array, offset, kNoDexPc);
|
/art/runtime/base/ |
H A D | arena_allocator_test.cc | 25 size_t NumberOfArenas(ArenaAllocator* arena) { argument 27 for (Arena* a = arena->arena_head_; a != nullptr; a = a->next_) { 36 ArenaAllocator arena(&pool); 37 ArenaBitVector bv(&arena, 10, true); 51 // Allocate a small array from an arena and release it. 52 ArenaAllocator arena(&pool); 53 small_array = arena.AllocArray<uint32_t>(kSmallArraySize); 57 // Reuse the previous arena and allocate more than previous allocation including red zone. 58 ArenaAllocator arena(&pool); 59 uint32_t* large_array = arena [all...] |
H A D | arena_bit_vector.h | 34 static ArenaBitVector* Create(Allocator* arena, argument 38 void* storage = arena->template Alloc<ArenaBitVector>(kind); 39 return new (storage) ArenaBitVector(arena, start_bits, expandable, kind); 42 ArenaBitVector(ArenaAllocator* arena, 46 ArenaBitVector(ScopedArenaAllocator* arena,
|
H A D | arena_object.h | 26 // Parent for arena allocated objects giving appropriate new and delete operators. 35 static void* operator new(size_t size, ScopedArenaAllocator* arena) { argument 36 return arena->Alloc(size, kAllocKind); 50 // Parent for arena allocated objects that get deleted, gives appropriate new and delete operators. 59 static void* operator new(size_t size, ScopedArenaAllocator* arena) { argument 60 return arena->Alloc(size, kAllocKind);
|
H A D | arena_bit_vector.cc | 55 static ArenaBitVectorAllocator* Create(ArenaAlloc* arena, ArenaAllocKind kind) { argument 56 void* storage = arena->template Alloc<ArenaBitVectorAllocator>(kind); 57 return new (storage) ArenaBitVectorAllocator(arena, kind); 72 ArenaBitVectorAllocator(ArenaAlloc* arena, ArenaAllocKind kind) argument 73 : ArenaBitVectorAllocatorKind(kind), arena_(arena) { } 80 ArenaBitVector::ArenaBitVector(ArenaAllocator* arena, argument 86 ArenaBitVectorAllocator<ArenaAllocator>::Create(arena, kind)) { 89 ArenaBitVector::ArenaBitVector(ScopedArenaAllocator* arena, argument 95 ArenaBitVectorAllocator<ScopedArenaAllocator>::Create(arena, kind)) {
|
H A D | arena_allocator.cc | 122 for (const Arena* arena = first; arena != nullptr; arena = arena->next_) { 123 malloc_bytes += arena->Size(); 124 lost_bytes += arena->RemainingSpace(); 127 // The lost_bytes_adjustment is used to make up for the fact that the current arena 220 auto* arena = free_arenas_; local 222 delete arena; 254 for (auto* arena [all...] |
/art/compiler/trampolines/ |
H A D | trampoline_compiler.cc | 53 ArenaAllocator* arena, EntryPointCallingConvention abi, ThreadOffset<4> offset) { 54 Thumb2Assembler assembler(arena); 83 ArenaAllocator* arena, EntryPointCallingConvention abi, ThreadOffset<8> offset) { 84 Arm64Assembler assembler(arena); 122 ArenaAllocator* arena, EntryPointCallingConvention abi, ThreadOffset<4> offset) { 123 MipsAssembler assembler(arena); 154 ArenaAllocator* arena, EntryPointCallingConvention abi, ThreadOffset<8> offset) { 155 Mips64Assembler assembler(arena); 185 static std::unique_ptr<const std::vector<uint8_t>> CreateTrampoline(ArenaAllocator* arena, argument 187 X86Assembler assembler(arena); 52 CreateTrampoline( ArenaAllocator* arena, EntryPointCallingConvention abi, ThreadOffset<4> offset) argument 82 CreateTrampoline( ArenaAllocator* arena, EntryPointCallingConvention abi, ThreadOffset<8> offset) argument 121 CreateTrampoline( ArenaAllocator* arena, EntryPointCallingConvention abi, ThreadOffset<4> offset) argument 153 CreateTrampoline( ArenaAllocator* arena, EntryPointCallingConvention abi, ThreadOffset<8> offset) argument 206 CreateTrampoline(ArenaAllocator* arena, ThreadOffset<8> offset) argument [all...] |
/art/compiler/jni/quick/ |
H A D | calling_convention.cc | 50 ArenaAllocator* arena, 60 new (arena) arm::ArmManagedRuntimeCallingConvention(is_static, is_synchronized, shorty)); 65 new (arena) arm64::Arm64ManagedRuntimeCallingConvention( 71 new (arena) mips::MipsManagedRuntimeCallingConvention( 77 new (arena) mips64::Mips64ManagedRuntimeCallingConvention( 83 new (arena) x86::X86ManagedRuntimeCallingConvention(is_static, is_synchronized, shorty)); 88 new (arena) x86_64::X86_64ManagedRuntimeCallingConvention( 149 std::unique_ptr<JniCallingConvention> JniCallingConvention::Create(ArenaAllocator* arena, argument 159 new (arena) arm::ArmJniCallingConvention(is_static, is_synchronized, shorty)); 164 new (arena) arm6 49 Create( ArenaAllocator* arena, bool is_static, bool is_synchronized, const char* shorty, InstructionSet instruction_set) argument [all...] |
/art/runtime/verifier/ |
H A D | reg_type-inl.h | 188 inline void* RegType::operator new(size_t size, ScopedArenaAllocator* arena) { argument 189 return arena->Alloc(size, kArenaAllocMisc);
|
H A D | reg_type_cache.cc | 266 RegTypeCache::RegTypeCache(bool can_load_classes, ScopedArenaAllocator& arena) argument 267 : entries_(arena.Adapter(kArenaAllocVerifier)), 268 klass_entries_(arena.Adapter(kArenaAllocVerifier)), 270 arena_(arena) {
|
/art/compiler/utils/arm64/ |
H A D | assembler_arm64.h | 89 explicit Arm64Assembler(ArenaAllocator* arena) argument 90 : Assembler(arena), 91 exception_blocks_(arena->Adapter(kArenaAllocAssembler)),
|
/art/compiler/utils/ |
H A D | assembler.cc | 47 AssemblerBuffer::AssemblerBuffer(ArenaAllocator* arena) argument 48 : arena_(arena) { 125 ArenaAllocator* arena, 131 return std::unique_ptr<Assembler>(new (arena) arm::Arm32Assembler(arena)); 133 return std::unique_ptr<Assembler>(new (arena) arm::Thumb2Assembler(arena)); 137 return std::unique_ptr<Assembler>(new (arena) arm64::Arm64Assembler(arena)); 141 return std::unique_ptr<Assembler>(new (arena) mip 124 Create( ArenaAllocator* arena, InstructionSet instruction_set, const InstructionSetFeatures* instruction_set_features) argument [all...] |
H A D | assembler_thumb_test.cc | 200 Thumb2AssemblerTest() : pool(), arena(&pool), assembler(&arena) { } 203 ArenaAllocator arena; member in class:art::arm::Thumb2AssemblerTest
|
H A D | assembler.h | 90 explicit AssemblerBuffer(ArenaAllocator* arena); 356 ArenaAllocator* arena, 521 explicit Assembler(ArenaAllocator* arena) : buffer_(arena), cfi_(this) {} argument
|
/art/compiler/utils/mips/ |
H A D | assembler_mips.h | 105 explicit MipsAssembler(ArenaAllocator* arena, argument 107 : Assembler(arena),
|
/art/compiler/utils/mips64/ |
H A D | assembler_mips64.h | 105 explicit Mips64Assembler(ArenaAllocator* arena) argument 106 : Assembler(arena),
|
/art/compiler/utils/x86/ |
H A D | assembler_x86.h | 266 explicit ConstantArea(ArenaAllocator* arena) : buffer_(arena->Adapter(kArenaAllocAssembler)) {} argument 307 explicit X86Assembler(ArenaAllocator* arena) : Assembler(arena), constant_area_(arena) {} argument
|
/art/compiler/utils/x86_64/ |
H A D | assembler_x86_64.h | 275 explicit ConstantArea(ArenaAllocator* arena) : buffer_(arena->Adapter(kArenaAllocAssembler)) {} argument 337 explicit X86_64Assembler(ArenaAllocator* arena) : Assembler(arena), constant_area_(arena) {} argument
|