/art/compiler/utils/ |
H A D | jni_macro_assembler.cc | 50 ArenaAllocator* allocator, 61 return MacroAsm32UniquePtr(new (allocator) arm::ArmVIXLJNIMacroAssembler(allocator)); 65 return MacroAsm32UniquePtr(new (allocator) mips::MipsAssembler( 66 allocator, 73 return MacroAsm32UniquePtr(new (allocator) x86::X86JNIMacroAssembler(allocator)); 85 ArenaAllocator* allocator, 95 return MacroAsm64UniquePtr(new (allocator) arm64::Arm64JNIMacroAssembler(allocator)); 49 Create( ArenaAllocator* allocator, InstructionSet instruction_set, const InstructionSetFeatures* instruction_set_features) argument 84 Create( ArenaAllocator* allocator, InstructionSet instruction_set, const InstructionSetFeatures* instruction_set_features) argument [all...] |
H A D | assembler.cc | 28 AssemblerBuffer::AssemblerBuffer(ArenaAllocator* allocator) argument 29 : allocator_(allocator) {
|
H A D | jni_macro_assembler.h | 50 ArenaAllocator* allocator, 279 explicit JNIMacroAssemblerFwd(ArenaAllocator* allocator) : asm_(allocator) {} argument
|
/art/compiler/utils/x86/ |
H A D | jni_macro_assembler_x86.h | 37 explicit X86JNIMacroAssembler(ArenaAllocator* allocator) : JNIMacroAssemblerFwd(allocator) {} argument
|
/art/compiler/utils/x86_64/ |
H A D | jni_macro_assembler_x86_64.h | 37 explicit X86_64JNIMacroAssembler(ArenaAllocator* allocator) argument 38 : JNIMacroAssemblerFwd<X86_64Assembler, PointerSize::k64>(allocator) {}
|
/art/runtime/base/ |
H A D | arena_object.h | 33 void* operator new(size_t size, ArenaAllocator* allocator) { argument 34 return allocator->Alloc(size, kAllocKind); 37 static void* operator new(size_t size, ScopedArenaAllocator* allocator) { argument 38 return allocator->Alloc(size, kAllocKind); 57 void* operator new(size_t size, ArenaAllocator* allocator) { argument 58 return allocator->Alloc(size, kAllocKind); 61 static void* operator new(size_t size, ScopedArenaAllocator* allocator) { argument 62 return allocator->Alloc(size, kAllocKind);
|
H A D | arena_bit_vector.h | 34 static ArenaBitVector* Create(Allocator* allocator, argument 38 void* storage = allocator->template Alloc<ArenaBitVector>(kind); 39 return new (storage) ArenaBitVector(allocator, start_bits, expandable, kind); 42 ArenaBitVector(ArenaAllocator* allocator, 46 ArenaBitVector(ScopedArenaAllocator* allocator,
|
H A D | arena_allocator_test.cc | 26 size_t NumberOfArenas(ArenaAllocator* allocator) { argument 28 for (Arena* a = allocator->arena_head_; a != nullptr; a = a->next_) { 37 ArenaAllocator allocator(&pool); 38 ArenaBitVector bv(&allocator, 10, true); 53 ArenaAllocator allocator(&pool); 54 small_array = allocator.AllocArray<uint32_t>(kSmallArraySize); 59 ArenaAllocator allocator(&pool); 60 uint32_t* large_array = allocator.AllocArray<uint32_t>(kLargeArraySize); 75 ArenaAllocator allocator(&pool); 77 void* alloc1 = allocator [all...] |
H A D | arena_bit_vector.cc | 19 #include "base/allocator.h" 55 static ArenaBitVectorAllocator* Create(ArenaAlloc* allocator, ArenaAllocKind kind) { argument 56 void* storage = allocator->template Alloc<ArenaBitVectorAllocator>(kind); 57 return new (storage) ArenaBitVectorAllocator(allocator, kind); 72 ArenaBitVectorAllocator(ArenaAlloc* allocator, ArenaAllocKind kind) argument 73 : ArenaBitVectorAllocatorKind(kind), allocator_(allocator) { } 80 ArenaBitVector::ArenaBitVector(ArenaAllocator* allocator, argument 86 ArenaBitVectorAllocator<ArenaAllocator>::Create(allocator, kind)) { 89 ArenaBitVector::ArenaBitVector(ScopedArenaAllocator* allocator, argument 95 ArenaBitVectorAllocator<ScopedArenaAllocator>::Create(allocator, kin [all...] |
/art/compiler/optimizing/ |
H A D | intrinsics_arm64.h | 42 explicit IntrinsicLocationsBuilderARM64(ArenaAllocator* allocator, CodeGeneratorARM64* codegen) argument 43 : allocator_(allocator), codegen_(codegen) {}
|
H A D | live_interval_test.cc | 27 ScopedArenaAllocator* allocator = pool.GetScopedAllocator(); local 31 LiveInterval* interval = BuildInterval(ranges, arraysize(ranges), allocator); 37 LiveInterval* interval = BuildInterval(ranges, arraysize(ranges), allocator); 44 ScopedArenaAllocator* allocator = pool.GetScopedAllocator(); local 48 LiveInterval* interval = BuildInterval(ranges, arraysize(ranges), allocator); 58 LiveInterval* interval = BuildInterval(ranges, arraysize(ranges), allocator); 72 ScopedArenaAllocator* allocator = pool.GetScopedAllocator(); local 76 LiveInterval* interval = BuildInterval(ranges, arraysize(ranges), allocator); 86 LiveInterval* interval = BuildInterval(ranges, arraysize(ranges), allocator); 100 ScopedArenaAllocator* allocator local 192 ScopedArenaAllocator* allocator = pool.GetScopedAllocator(); local 283 ScopedArenaAllocator* allocator = pool.GetScopedAllocator(); local [all...] |
H A D | locations.cc | 32 ArenaAllocator* allocator) 33 : inputs_(instruction->InputCount(), allocator->Adapter(kArenaAllocLocationSummary)), 34 temps_(allocator->Adapter(kArenaAllocLocationSummary)), 46 stack_mask_ = ArenaBitVector::Create(allocator, 0, true, kArenaAllocLocationSummary); 29 LocationSummary(HInstruction* instruction, CallKind call_kind, bool intrinsified, ArenaAllocator* allocator) argument
|
H A D | graph_visualizer.h | 57 explicit DisassemblyInformation(ArenaAllocator* allocator) argument 59 instruction_intervals_(std::less<const HInstruction*>(), allocator->Adapter()), 60 slow_path_intervals_(allocator->Adapter()) {}
|
H A D | instruction_simplifier_mips.cc | 75 ArenaAllocator* allocator = graph->GetAllocator(); local 114 new (allocator) HIntermediateArrayAddressIndex(index, shift, kNoDexPc);
|
H A D | optimization.cc | 172 ArenaAllocator* allocator, 179 ArenaVector<HOptimization*> optimizations(allocator->Adapter()); 202 opt = most_recent_side_effects = new (allocator) SideEffectsAnalysis(graph, name); 205 opt = most_recent_induction = new (allocator) HInductionVarAnalysis(graph, name); 208 opt = most_recent_lsa = new (allocator) LoadStoreAnalysis(graph, name); 215 opt = new (allocator) GVNOptimization(graph, *most_recent_side_effects, name); 219 opt = new (allocator) LICM(graph, *most_recent_side_effects, stats, name); 223 opt = new (allocator) HLoopOptimization(graph, driver, most_recent_induction, stats, name); 227 opt = new (allocator) BoundsCheckElimination( 232 opt = new (allocator) LoadStoreEliminatio 169 ConstructOptimizations( const OptimizationDef definitions[], size_t length, ArenaAllocator* allocator, HGraph* graph, OptimizingCompilerStats* stats, CodeGenerator* codegen, CompilerDriver* driver, const DexCompilationUnit& dex_compilation_unit, VariableSizedHandleScope* handles) argument [all...] |
H A D | parallel_move_resolver.h | 34 explicit ParallelMoveResolver(ArenaAllocator* allocator) argument 35 : moves_(allocator->Adapter(kArenaAllocParallelMoveResolver)) { 56 explicit ParallelMoveResolverWithSwap(ArenaAllocator* allocator) argument 57 : ParallelMoveResolver(allocator) {} 125 explicit ParallelMoveResolverNoSwap(ArenaAllocator* allocator) argument 126 : ParallelMoveResolver(allocator), 127 scratches_(allocator->Adapter(kArenaAllocParallelMoveResolver)), 128 pending_moves_(allocator->Adapter(kArenaAllocParallelMoveResolver)), 129 allocator_(allocator) {
|
H A D | parallel_move_test.cc | 55 explicit TestParallelMoveResolverWithSwap(ArenaAllocator* allocator) argument 56 : ParallelMoveResolverWithSwap(allocator) {} 98 explicit TestParallelMoveResolverNoSwap(ArenaAllocator* allocator) argument 99 : ParallelMoveResolverNoSwap(allocator), scratch_index_(kScratchRegisterStartIndexForTest) {} 153 static HParallelMove* BuildParallelMove(ArenaAllocator* allocator, argument 156 HParallelMove* moves = new (allocator) HParallelMove(allocator); 184 ArenaAllocator allocator(&pool); 187 TypeParam resolver(&allocator); 189 resolver.EmitNativeCode(BuildParallelMove(&allocator, move [all...] |
/art/compiler/utils/arm64/ |
H A D | jni_macro_assembler_arm64.h | 45 explicit Arm64JNIMacroAssembler(ArenaAllocator* allocator) argument 46 : JNIMacroAssemblerFwd(allocator), 47 exception_blocks_(allocator->Adapter(kArenaAllocAssembler)) {}
|
/art/runtime/gc/allocator/ |
H A D | dlmalloc.h | 51 namespace allocator { namespace in namespace:art::gc 57 } // namespace allocator
|
H A D | rosalloc-inl.h | 24 namespace allocator { namespace in namespace:art::gc 122 } // namespace allocator
|
/art/runtime/gc/ |
H A D | allocator_type.h | 27 kAllocatorTypeBumpPointer, // Use BumpPointer allocator, has entrypoints. 28 kAllocatorTypeTLAB, // Use TLAB allocator, has entrypoints. 29 kAllocatorTypeRosAlloc, // Use RosAlloc allocator, has entrypoints. 30 kAllocatorTypeDlMalloc, // Use dlmalloc allocator, has entrypoints. 31 kAllocatorTypeNonMoving, // Special allocator for non moving objects, doesn't have entrypoints. 38 inline constexpr bool IsTLABAllocator(AllocatorType allocator) { argument 39 return allocator == kAllocatorTypeTLAB || allocator == kAllocatorTypeRegionTLAB;
|
/art/compiler/trampolines/ |
H A D | trampoline_compiler.cc | 60 ArenaAllocator* allocator, EntryPointCallingConvention abi, ThreadOffset32 offset) { 64 ArmVIXLAssembler assembler(allocator); 101 ArenaAllocator* allocator, EntryPointCallingConvention abi, ThreadOffset64 offset) { 102 Arm64Assembler assembler(allocator); 140 ArenaAllocator* allocator, EntryPointCallingConvention abi, ThreadOffset32 offset) { 141 MipsAssembler assembler(allocator); 172 ArenaAllocator* allocator, EntryPointCallingConvention abi, ThreadOffset64 offset) { 173 Mips64Assembler assembler(allocator); 203 static std::unique_ptr<const std::vector<uint8_t>> CreateTrampoline(ArenaAllocator* allocator, argument 205 X86Assembler assembler(allocator); 59 CreateTrampoline( ArenaAllocator* allocator, EntryPointCallingConvention abi, ThreadOffset32 offset) argument 100 CreateTrampoline( ArenaAllocator* allocator, EntryPointCallingConvention abi, ThreadOffset64 offset) argument 139 CreateTrampoline( ArenaAllocator* allocator, EntryPointCallingConvention abi, ThreadOffset32 offset) argument 171 CreateTrampoline( ArenaAllocator* allocator, EntryPointCallingConvention abi, ThreadOffset64 offset) argument 224 CreateTrampoline(ArenaAllocator* allocator, ThreadOffset64 offset) argument [all...] |
/art/compiler/utils/arm/ |
H A D | jni_macro_assembler_arm_vixl.h | 40 explicit ArmVIXLJNIMacroAssembler(ArenaAllocator* allocator) argument 41 : JNIMacroAssemblerFwd(allocator), 42 exception_blocks_(allocator->Adapter(kArenaAllocAssembler)) {}
|
H A D | assembler_arm_vixl.h | 156 explicit ArmVIXLAssembler(ArenaAllocator* allocator) argument 157 : Assembler(allocator) {
|
/art/runtime/entrypoints/quick/ |
H A D | quick_alloc_entrypoints.cc | 195 void SetQuickAllocEntryPointsAllocator(gc::AllocatorType allocator) { argument 196 entry_points_allocator = allocator;
|