/art/compiler/optimizing/ |
H A D | nodes_test.cc | 31 ArenaAllocator allocator(&pool); 33 HGraph* graph = CreateGraph(&allocator); 34 HBasicBlock* entry = new (&allocator) HBasicBlock(graph); 37 HInstruction* parameter = new (&allocator) HParameterValue( 40 entry->AddInstruction(new (&allocator) HGoto()); 42 HBasicBlock* first_block = new (&allocator) HBasicBlock(graph); 45 HInstruction* null_check = new (&allocator) HNullCheck(parameter, 0); 47 first_block->AddInstruction(new (&allocator) HReturnVoid()); 49 HBasicBlock* exit_block = new (&allocator) HBasicBlock(graph); 52 exit_block->AddInstruction(new (&allocator) HExi [all...] |
H A D | gvn_test.cc | 30 ArenaAllocator allocator(&pool); 33 HGraph* graph = CreateGraph(&allocator); 34 HBasicBlock* entry = new (&allocator) HBasicBlock(graph); 37 HInstruction* parameter = new (&allocator) HParameterValue(graph->GetDexFile(), 43 HBasicBlock* block = new (&allocator) HBasicBlock(graph); 47 block->AddInstruction(new (&allocator) HInstanceFieldGet(parameter, 56 block->AddInstruction(new (&allocator) HInstanceFieldGet(parameter, 66 block->AddInstruction(new (&allocator) HInstanceFieldGet(parameter, 77 block->AddInstruction(new (&allocator) HInstanceFieldSet(parameter, 87 block->AddInstruction(new (&allocator) HInstanceFieldGe [all...] |
H A D | register_allocator_test.cc | 33 // Note: the register allocator tests rely on the fact that constants have live 40 ArenaAllocator allocator(&pool); 41 HGraph* graph = CreateCFG(&allocator, data); 47 RegisterAllocator register_allocator(&allocator, &codegen, liveness); 53 * Unit testing of RegisterAllocator::ValidateIntervals. Register allocator 58 ArenaAllocator allocator(&pool); 59 HGraph* graph = CreateGraph(&allocator); 63 ArenaVector<LiveInterval*> intervals(allocator.Adapter()); 68 intervals.push_back(BuildInterval(ranges, arraysize(ranges), &allocator, 0)); 69 intervals.push_back(BuildInterval(ranges, arraysize(ranges), &allocator, 456 BuildIfElseWithPhi(ArenaAllocator* allocator, HPhi** phi, HInstruction** input1, HInstruction** input2) argument 610 BuildFieldReturn(ArenaAllocator* allocator, HInstruction** field, HInstruction** ret) argument 687 BuildTwoSubs(ArenaAllocator* allocator, HInstruction** first_sub, HInstruction** second_sub) argument 759 BuildDiv(ArenaAllocator* allocator, HInstruction** div) argument [all...] |
H A D | graph_test.cc | 28 static HBasicBlock* createIfBlock(HGraph* graph, ArenaAllocator* allocator) { argument 29 HBasicBlock* if_block = new (allocator) HBasicBlock(graph); 32 HInstruction* equal = new (allocator) HEqual(instr, instr); 34 instr = new (allocator) HIf(equal); 39 static HBasicBlock* createGotoBlock(HGraph* graph, ArenaAllocator* allocator) { argument 40 HBasicBlock* block = new (allocator) HBasicBlock(graph); 42 HInstruction* got = new (allocator) HGoto(); 47 static HBasicBlock* createEntryBlock(HGraph* graph, ArenaAllocator* allocator) { argument 48 HBasicBlock* block = createGotoBlock(graph, allocator); 53 static HBasicBlock* createReturnBlock(HGraph* graph, ArenaAllocator* allocator) { argument 61 createExitBlock(HGraph* graph, ArenaAllocator* allocator) argument [all...] |
H A D | parallel_move_test.cc | 55 explicit TestParallelMoveResolverWithSwap(ArenaAllocator* allocator) argument 56 : ParallelMoveResolverWithSwap(allocator) {} 98 explicit TestParallelMoveResolverNoSwap(ArenaAllocator* allocator) argument 99 : ParallelMoveResolverNoSwap(allocator), scratch_index_(kScratchRegisterStartIndexForTest) {} 153 static HParallelMove* BuildParallelMove(ArenaAllocator* allocator, argument 156 HParallelMove* moves = new (allocator) HParallelMove(allocator); 184 ArenaAllocator allocator(&pool); 187 TypeParam resolver(&allocator); 189 resolver.EmitNativeCode(BuildParallelMove(&allocator, move [all...] |
H A D | graph_checker_test.cc | 30 HGraph* CreateSimpleCFG(ArenaAllocator* allocator) { argument 31 HGraph* graph = CreateGraph(allocator); 32 HBasicBlock* entry_block = new (allocator) HBasicBlock(graph); 33 entry_block->AddInstruction(new (allocator) HReturnVoid()); 36 HBasicBlock* exit_block = new (allocator) HBasicBlock(graph); 37 exit_block->AddInstruction(new (allocator) HExit()); 47 ArenaAllocator allocator(&pool); 48 HGraph* graph = CreateCFG(&allocator, data); 97 ArenaAllocator allocator(&pool); 99 HGraph* graph = CreateSimpleCFG(&allocator); [all...] |
H A D | live_interval_test.cc | 27 ArenaAllocator allocator(&pool); 31 LiveInterval* interval = BuildInterval(ranges, arraysize(ranges), &allocator); 37 LiveInterval* interval = BuildInterval(ranges, arraysize(ranges), &allocator); 44 ArenaAllocator allocator(&pool); 48 LiveInterval* interval = BuildInterval(ranges, arraysize(ranges), &allocator); 58 LiveInterval* interval = BuildInterval(ranges, arraysize(ranges), &allocator); 72 ArenaAllocator allocator(&pool); 76 LiveInterval* interval = BuildInterval(ranges, arraysize(ranges), &allocator); 86 LiveInterval* interval = BuildInterval(ranges, arraysize(ranges), &allocator); 100 ArenaAllocator allocator( [all...] |
H A D | parallel_move_resolver.h | 34 explicit ParallelMoveResolver(ArenaAllocator* allocator) argument 35 : moves_(allocator->Adapter(kArenaAllocParallelMoveResolver)) { 56 explicit ParallelMoveResolverWithSwap(ArenaAllocator* allocator) argument 57 : ParallelMoveResolver(allocator) {} 125 explicit ParallelMoveResolverNoSwap(ArenaAllocator* allocator) argument 126 : ParallelMoveResolver(allocator), 127 scratches_(allocator->Adapter(kArenaAllocParallelMoveResolver)), 128 pending_moves_(allocator->Adapter(kArenaAllocParallelMoveResolver)), 129 allocator_(allocator) {
|
H A D | optimizing_unit_test.h | 50 ArenaAllocator* allocator, 53 LiveInterval* interval = LiveInterval::MakeInterval(allocator, Primitive::kPrimInt, defined_by); 77 inline HGraph* CreateGraph(ArenaAllocator* allocator) { argument 78 return new (allocator) HGraph( 79 allocator, *reinterpret_cast<DexFile*>(allocator->Alloc(sizeof(DexFile))), -1, false, 84 inline HGraph* CreateCFG(ArenaAllocator* allocator, argument 89 HGraph* graph = CreateGraph(allocator); 48 BuildInterval(const size_t ranges[][2], size_t number_of_ranges, ArenaAllocator* allocator, int reg = -1, HInstruction* defined_by = nullptr) argument
|
H A D | bounds_check_elimination_test.cc | 359 ArenaAllocator* allocator, 363 HBasicBlock* entry = new (allocator) HBasicBlock(graph); 366 HInstruction* parameter = new (allocator) HParameterValue( 374 HBasicBlock* block = new (allocator) HBasicBlock(graph); 377 block->AddInstruction(new (allocator) HGoto()); 379 HBasicBlock* loop_header = new (allocator) HBasicBlock(graph); 380 HBasicBlock* loop_body = new (allocator) HBasicBlock(graph); 381 HBasicBlock* exit = new (allocator) HBasicBlock(graph); 391 HPhi* phi = new (allocator) HPhi(allocator, 358 BuildSSAGraph1(HGraph* graph, ArenaAllocator* allocator, int initial, int increment, IfCondition cond = kCondGE) argument 473 BuildSSAGraph2(HGraph *graph, ArenaAllocator* allocator, int initial, int increment = -1, IfCondition cond = kCondLE) argument 583 BuildSSAGraph3(HGraph* graph, ArenaAllocator* allocator, int initial, int increment, IfCondition cond) argument 687 BuildSSAGraph4(HGraph* graph, ArenaAllocator* allocator, int initial, IfCondition cond = kCondGE) argument [all...] |
H A D | find_loops_test.cc | 39 ArenaAllocator allocator(&arena); 40 HGraph* graph = CreateCFG(&allocator, data); 52 ArenaAllocator allocator(&arena); 53 HGraph* graph = CreateCFG(&allocator, data); 68 ArenaAllocator allocator(&arena); 69 HGraph* graph = CreateCFG(&allocator, data); 85 ArenaAllocator allocator(&arena); 86 HGraph* graph = CreateCFG(&allocator, data); 100 ArenaAllocator allocator(&arena); 101 HGraph* graph = CreateCFG(&allocator, dat [all...] |
H A D | stack_map_stream.h | 62 explicit StackMapStream(ArenaAllocator* allocator) argument 63 : allocator_(allocator), 64 stack_maps_(allocator->Adapter(kArenaAllocStackMapStream)), 65 location_catalog_entries_(allocator->Adapter(kArenaAllocStackMapStream)), 66 location_catalog_entries_indices_(allocator->Adapter(kArenaAllocStackMapStream)), 67 dex_register_locations_(allocator->Adapter(kArenaAllocStackMapStream)), 68 inline_infos_(allocator->Adapter(kArenaAllocStackMapStream)), 74 allocator->Adapter(kArenaAllocStackMapStream)), 77 code_info_encoding_(allocator->Adapter(kArenaAllocStackMapStream)),
|
H A D | codegen_test.cc | 187 static void Run(const InternalCodeAllocator& allocator, argument 194 CommonCompilerTest::MakeExecutable(allocator.GetMemory(), allocator.GetSize()); 195 fptr f = reinterpret_cast<fptr>(allocator.GetMemory()); 225 InternalCodeAllocator allocator; local 226 codegen->Compile(&allocator); 227 Run(allocator, *codegen, has_result, expected); 633 ArenaAllocator allocator(&pool); 635 HGraph* graph = CreateGraph(&allocator); 637 HBasicBlock* entry = new (&allocator) HBasicBloc [all...] |
H A D | graph_visualizer.h | 57 explicit DisassemblyInformation(ArenaAllocator* allocator) argument 59 instruction_intervals_(std::less<const HInstruction*>(), allocator->Adapter()), 60 slow_path_intervals_(allocator->Adapter()) {}
|
H A D | suspend_check_test.cc | 33 ArenaAllocator allocator(&pool); 34 HGraph* graph = CreateCFG(&allocator, data);
|
H A D | gvn.cc | 39 explicit ValueSet(ArenaAllocator* allocator) argument 40 : allocator_(allocator), 42 buckets_(allocator->AllocArray<Node*>(num_buckets_, kArenaAllocGvn)), 43 buckets_owned_(allocator, num_buckets_, false, kArenaAllocGvn), 52 ValueSet(ArenaAllocator* allocator, const ValueSet& other) argument 53 : allocator_(allocator), 55 buckets_(allocator->AllocArray<Node*>(num_buckets_, kArenaAllocGvn)), 56 buckets_owned_(allocator, num_buckets_, false, kArenaAllocGvn), 211 Node* Dup(ArenaAllocator* allocator, Node* new_next = nullptr) { argument 212 return new (allocator) Nod 353 GlobalValueNumberer(ArenaAllocator* allocator, HGraph* graph, const SideEffectsAnalysis& side_effects) argument [all...] |
H A D | live_ranges_test.cc | 34 static HGraph* BuildGraph(const uint16_t* data, ArenaAllocator* allocator) { argument 35 HGraph* graph = CreateCFG(allocator, data); 62 ArenaAllocator allocator(&pool); 63 HGraph* graph = BuildGraph(data, &allocator); 111 ArenaAllocator allocator(&pool); 112 HGraph* graph = BuildGraph(data, &allocator); 162 ArenaAllocator allocator(&pool); 163 HGraph* graph = BuildGraph(data, &allocator); 240 ArenaAllocator allocator(&pool); 241 HGraph* graph = BuildGraph(data, &allocator); [all...] |
/art/runtime/base/ |
H A D | arena_object.h | 31 void* operator new(size_t size, ArenaAllocator* allocator) { argument 32 return allocator->Alloc(size, kAllocKind); 55 void* operator new(size_t size, ArenaAllocator* allocator) { argument 56 return allocator->Alloc(size, kAllocKind);
|
H A D | allocator.h | 106 // Tracking allocator for use with STL types, tracks how much memory is used. 108 class TrackingAllocatorImpl : public std::allocator<T> { 110 typedef typename std::allocator<T>::value_type value_type; 111 typedef typename std::allocator<T>::size_type size_type; 112 typedef typename std::allocator<T>::difference_type difference_type; 113 typedef typename std::allocator<T>::pointer pointer; 114 typedef typename std::allocator<T>::const_pointer const_pointer; 115 typedef typename std::allocator<T>::reference reference; 116 typedef typename std::allocator<T>::const_reference const_reference; 127 // Enables an allocator fo [all...] |
/art/runtime/entrypoints/quick/ |
H A D | quick_alloc_entrypoints.h | 33 void SetQuickAllocEntryPointsAllocator(gc::AllocatorType allocator)
|
/art/runtime/gc/allocator/ |
H A D | dlmalloc.h | 57 namespace allocator { namespace in namespace:art::gc 63 } // namespace allocator
|
/art/runtime/gc/space/ |
H A D | rosalloc_space.cc | 43 // template class MemoryToolMallocSpace<RosAllocSpace, allocator::RosAlloc*>; 46 art::gc::allocator::RosAlloc* rosalloc, uint8_t* begin, uint8_t* end, 63 allocator::RosAlloc* rosalloc = CreateRosAlloc(mem_map->Begin(), starting_size, initial_size, 131 allocator::RosAlloc* RosAllocSpace::CreateRosAlloc(void* begin, size_t morecore_start, 140 allocator::RosAlloc* rosalloc = new art::gc::allocator::RosAlloc( 143 art::gc::allocator::RosAlloc::kPageReleaseModeAll : 144 art::gc::allocator::RosAlloc::kPageReleaseModeSizeAndEnd, 177 void* allocator, uint8_t* begin, uint8_t* end, 182 mem_map, initial_size_, name, reinterpret_cast<allocator 176 CreateInstance(MemMap* mem_map, const std::string& name, void* allocator, uint8_t* begin, uint8_t* end, uint8_t* limit, size_t growth_limit, bool can_move_objects) argument 378 namespace allocator { namespace in namespace:art::gc [all...] |
/art/runtime/mirror/ |
H A D | class_loader.h | 58 void SetAllocator(LinearAlloc* allocator) SHARED_REQUIRES(Locks::mutator_lock_) { 60 reinterpret_cast<uint64_t>(allocator));
|
/art/runtime/verifier/ |
H A D | reg_type_test.cc | 38 ScopedArenaAllocator allocator(&stack); 40 RegTypeCache cache(true, allocator); 62 ScopedArenaAllocator allocator(&stack); 64 RegTypeCache cache(true, allocator); 88 ScopedArenaAllocator allocator(&stack); 90 RegTypeCache cache(true, allocator); 363 ScopedArenaAllocator allocator(&stack); 365 RegTypeCache cache(true, allocator); 380 ScopedArenaAllocator allocator(&stack); 382 RegTypeCache cache(true, allocator); [all...] |
/art/runtime/gc/ |
H A D | heap-inl.h | 44 AllocatorType allocator, 53 // Need to check that we arent the large object allocator since the large object allocation code 73 if (allocator == kAllocatorTypeTLAB || allocator == kAllocatorTypeRegionTLAB) { 77 if ((allocator == kAllocatorTypeTLAB || allocator == kAllocatorTypeRegionTLAB) && 92 } else if (!kInstrumented && allocator == kAllocatorTypeRosAlloc && 109 obj = TryToAllocate<kInstrumented, false>(self, allocator, byte_count, &bytes_allocated, 113 // or changes the allocator in a suspend point here, we need to retry the allocation. 115 allocator, 41 AllocObjectWithAllocator(Thread* self, mirror::Class* klass, size_t byte_count, AllocatorType allocator, const PreFenceVisitor& pre_fence_visitor) argument [all...] |