/art/compiler/optimizing/ |
H A D | nodes_test.cc | 31 ArenaAllocator allocator(&pool); 33 HGraph* graph = CreateGraph(&allocator); 34 HBasicBlock* entry = new (&allocator) HBasicBlock(graph); 37 HInstruction* parameter = new (&allocator) HParameterValue(0, Primitive::kPrimNot); 39 entry->AddInstruction(new (&allocator) HGoto()); 41 HBasicBlock* first_block = new (&allocator) HBasicBlock(graph); 44 HInstruction* null_check = new (&allocator) HNullCheck(parameter, 0); 46 first_block->AddInstruction(new (&allocator) HReturnVoid()); 48 HBasicBlock* exit_block = new (&allocator) HBasicBlock(graph); 51 exit_block->AddInstruction(new (&allocator) HExi [all...] |
H A D | bounds_check_elimination_test.cc | 43 ArenaAllocator allocator(&pool); 45 HGraph* graph = CreateGraph(&allocator); 48 HBasicBlock* entry = new (&allocator) HBasicBlock(graph); 51 HInstruction* parameter1 = new (&allocator) 53 HInstruction* parameter2 = new (&allocator) 61 HBasicBlock* block1 = new (&allocator) HBasicBlock(graph); 63 HInstruction* cmp = new (&allocator) HGreaterThanOrEqual(parameter2, constant_0); 64 HIf* if_inst = new (&allocator) HIf(cmp); 69 HBasicBlock* block2 = new (&allocator) HBasicBlock(graph); 71 HNullCheck* null_check = new (&allocator) HNullChec 362 BuildSSAGraph1(ArenaAllocator* allocator, HInstruction** bounds_check, int initial, int increment, IfCondition cond = kCondGE) argument 497 BuildSSAGraph2(ArenaAllocator* allocator, HInstruction** bounds_check, int initial, int increment = -1, IfCondition cond = kCondLE) argument 625 BuildSSAGraph3(ArenaAllocator* allocator, HInstruction** bounds_check, int initial, int increment, IfCondition cond) argument 741 BuildSSAGraph4(ArenaAllocator* allocator, HInstruction** bounds_check, int initial, IfCondition cond = kCondGE) argument [all...] |
H A D | gvn_test.cc | 30 ArenaAllocator allocator(&pool); 32 HGraph* graph = CreateGraph(&allocator); 33 HBasicBlock* entry = new (&allocator) HBasicBlock(graph); 36 HInstruction* parameter = new (&allocator) HParameterValue(0, Primitive::kPrimNot); 39 HBasicBlock* block = new (&allocator) HBasicBlock(graph); 44 new (&allocator) HInstanceFieldGet(parameter, Primitive::kPrimNot, 47 new (&allocator) HInstanceFieldGet(parameter, Primitive::kPrimNot, 51 new (&allocator) HInstanceFieldGet(parameter, Primitive::kPrimNot, 55 block->AddInstruction(new (&allocator) HInstanceFieldSet( 58 new (&allocator) HInstanceFieldGe [all...] |
H A D | register_allocator_test.cc | 35 // Note: the register allocator tests rely on the fact that constants have live 40 ArenaAllocator allocator(&pool); 41 HGraph* graph = CreateGraph(&allocator); 51 RegisterAllocator register_allocator(&allocator, &codegen, liveness); 57 * Unit testing of RegisterAllocator::ValidateIntervals. Register allocator 62 ArenaAllocator allocator(&pool); 63 HGraph* graph = CreateGraph(&allocator); 67 GrowableArray<LiveInterval*> intervals(&allocator, 0); 72 intervals.Add(BuildInterval(ranges, arraysize(ranges), &allocator, 0)); 73 intervals.Add(BuildInterval(ranges, arraysize(ranges), &allocator, 257 BuildSSAGraph(const uint16_t* data, ArenaAllocator* allocator) argument 462 BuildIfElseWithPhi(ArenaAllocator* allocator, HPhi** phi, HInstruction** input1, HInstruction** input2) argument 593 BuildFieldReturn(ArenaAllocator* allocator, HInstruction** field, HInstruction** ret) argument 661 BuildTwoSubs(ArenaAllocator* allocator, HInstruction** first_sub, HInstruction** second_sub) argument 732 BuildDiv(ArenaAllocator* allocator, HInstruction** div) argument [all...] |
H A D | graph_test.cc | 28 static HBasicBlock* createIfBlock(HGraph* graph, ArenaAllocator* allocator) { argument 29 HBasicBlock* if_block = new (allocator) HBasicBlock(graph); 32 HInstruction* equal = new (allocator) HEqual(instr, instr); 34 instr = new (allocator) HIf(equal); 39 static HBasicBlock* createGotoBlock(HGraph* graph, ArenaAllocator* allocator) { argument 40 HBasicBlock* block = new (allocator) HBasicBlock(graph); 42 HInstruction* got = new (allocator) HGoto(); 47 static HBasicBlock* createEntryBlock(HGraph* graph, ArenaAllocator* allocator) { argument 48 HBasicBlock* block = createGotoBlock(graph, allocator); 53 static HBasicBlock* createReturnBlock(HGraph* graph, ArenaAllocator* allocator) { argument 61 createExitBlock(HGraph* graph, ArenaAllocator* allocator) argument [all...] |
H A D | graph_checker_test.cc | 32 HGraph* CreateSimpleCFG(ArenaAllocator* allocator) { argument 33 HGraph* graph = CreateGraph(allocator); 34 HBasicBlock* entry_block = new (allocator) HBasicBlock(graph); 35 entry_block->AddInstruction(new (allocator) HGoto()); 38 HBasicBlock* exit_block = new (allocator) HBasicBlock(graph); 39 exit_block->AddInstruction(new (allocator) HExit()); 49 ArenaAllocator allocator(&pool); 50 HGraph* graph = CreateCFG(&allocator, data); 53 GraphChecker graph_checker(&allocator, graph); 60 ArenaAllocator allocator( [all...] |
H A D | parallel_move_test.cc | 55 explicit TestParallelMoveResolverWithSwap(ArenaAllocator* allocator) argument 56 : ParallelMoveResolverWithSwap(allocator) {} 98 explicit TestParallelMoveResolverNoSwap(ArenaAllocator* allocator) argument 99 : ParallelMoveResolverNoSwap(allocator), scratch_index_(kScratchRegisterStartIndexForTest) {} 153 static HParallelMove* BuildParallelMove(ArenaAllocator* allocator, argument 156 HParallelMove* moves = new (allocator) HParallelMove(allocator); 184 ArenaAllocator allocator(&pool); 187 TypeParam resolver(&allocator); 189 resolver.EmitNativeCode(BuildParallelMove(&allocator, move [all...] |
H A D | live_interval_test.cc | 27 ArenaAllocator allocator(&pool); 31 LiveInterval* interval = BuildInterval(ranges, arraysize(ranges), &allocator); 37 LiveInterval* interval = BuildInterval(ranges, arraysize(ranges), &allocator); 44 ArenaAllocator allocator(&pool); 48 LiveInterval* interval = BuildInterval(ranges, arraysize(ranges), &allocator); 58 LiveInterval* interval = BuildInterval(ranges, arraysize(ranges), &allocator); 72 ArenaAllocator allocator(&pool); 76 LiveInterval* interval = BuildInterval(ranges, arraysize(ranges), &allocator); 86 LiveInterval* interval = BuildInterval(ranges, arraysize(ranges), &allocator); 100 ArenaAllocator allocator( [all...] |
H A D | parallel_move_resolver.h | 34 explicit ParallelMoveResolver(ArenaAllocator* allocator) : moves_(allocator, 32) {} argument 53 explicit ParallelMoveResolverWithSwap(ArenaAllocator* allocator) argument 54 : ParallelMoveResolver(allocator) {} 122 explicit ParallelMoveResolverNoSwap(ArenaAllocator* allocator) argument 123 : ParallelMoveResolver(allocator), scratches_(allocator, 32), 124 pending_moves_(allocator, 8), allocator_(allocator) {}
|
H A D | optimizing_unit_test.h | 48 ArenaAllocator* allocator, 51 LiveInterval* interval = LiveInterval::MakeInterval(allocator, Primitive::kPrimInt, defined_by); 75 inline HGraph* CreateGraph(ArenaAllocator* allocator) { argument 76 return new (allocator) HGraph( 77 allocator, *reinterpret_cast<DexFile*>(allocator->Alloc(sizeof(DexFile))), -1, kRuntimeISA, 82 inline HGraph* CreateCFG(ArenaAllocator* allocator, argument 85 HGraph* graph = CreateGraph(allocator); 46 BuildInterval(const size_t ranges[][2], size_t number_of_ranges, ArenaAllocator* allocator, int reg = -1, HInstruction* defined_by = nullptr) argument
|
H A D | find_loops_test.cc | 30 static HGraph* TestCode(const uint16_t* data, ArenaAllocator* allocator) { argument 31 HGraph* graph = CreateGraph(allocator); 47 ArenaAllocator allocator(&arena); 48 HGraph* graph = TestCode(data, &allocator); 60 ArenaAllocator allocator(&arena); 61 HGraph* graph = TestCode(data, &allocator); 76 ArenaAllocator allocator(&arena); 77 HGraph* graph = TestCode(data, &allocator); 93 ArenaAllocator allocator(&arena); 94 HGraph* graph = TestCode(data, &allocator); [all...] |
H A D | graph_checker.h | 29 GraphChecker(ArenaAllocator* allocator, HGraph* graph, argument 32 allocator_(allocator), 34 seen_ids_(allocator, graph->GetCurrentInstructionId(), false) {} 98 // TODO: There's no need to pass a separate allocator as we could get it from the graph. 99 SSAChecker(ArenaAllocator* allocator, HGraph* graph) argument 100 : GraphChecker(allocator, graph, "art::SSAChecker: ") {}
|
H A D | boolean_simplifier.cc | 67 ArenaAllocator* allocator = graph->GetArena(); local 73 return new (allocator) HNotEqual(lhs, rhs); 75 return new (allocator) HEqual(lhs, rhs); 77 return new (allocator) HGreaterThanOrEqual(lhs, rhs); 79 return new (allocator) HGreaterThan(lhs, rhs); 81 return new (allocator) HLessThanOrEqual(lhs, rhs); 84 return new (allocator) HLessThan(lhs, rhs); 97 return new (allocator) HBooleanNot(cond);
|
H A D | codegen_test.cc | 124 static void Run(const InternalCodeAllocator& allocator, argument 129 CommonCompilerTest::MakeExecutable(allocator.GetMemory(), allocator.GetSize()); 130 fptr f = reinterpret_cast<fptr>(allocator.GetMemory()); 143 InternalCodeAllocator allocator; local 151 codegenX86.CompileBaseline(&allocator, true); 153 Run(allocator, codegenX86, has_result, expected); 159 codegenARM.CompileBaseline(&allocator, true); 161 Run(allocator, codegenARM, has_result, expected); 167 codegenX86_64.CompileBaseline(&allocator, tru 206 InternalCodeAllocator allocator; local [all...] |
H A D | stack_map_stream.h | 62 explicit StackMapStream(ArenaAllocator* allocator) argument 63 : allocator_(allocator), 64 stack_maps_(allocator, 10), 65 location_catalog_entries_(allocator, 4), 66 dex_register_locations_(allocator, 10 * 4), 67 inline_infos_(allocator, 2), 73 dex_map_hash_to_stack_map_indices_(std::less<uint32_t>(), allocator->Adapter()),
|
H A D | gvn.cc | 38 explicit ValueSet(ArenaAllocator* allocator) argument 39 : allocator_(allocator), 41 buckets_(allocator->AllocArray<Node*>(num_buckets_)), 42 buckets_owned_(allocator, num_buckets_, false), 51 ValueSet(ArenaAllocator* allocator, const ValueSet& to_copy) argument 52 : allocator_(allocator), 54 buckets_(allocator->AllocArray<Node*>(num_buckets_)), 55 buckets_owned_(allocator, num_buckets_, false), 156 Node* Dup(ArenaAllocator* allocator, Node* new_next = nullptr) { argument 157 return new (allocator) Nod 303 GlobalValueNumberer(ArenaAllocator* allocator, HGraph* graph, const SideEffectsAnalysis& side_effects) argument [all...] |
H A D | dead_code_elimination_test.cc | 33 ArenaAllocator allocator(&pool); 34 HGraph* graph = CreateCFG(&allocator, data); 48 SSAChecker ssa_checker(&allocator, graph);
|
H A D | suspend_check_test.cc | 32 ArenaAllocator allocator(&pool); 33 HGraph* graph = CreateGraph(&allocator);
|
/art/runtime/base/ |
H A D | arena_object.h | 31 void* operator new(size_t size, ArenaAllocator* allocator) { argument 32 return allocator->Alloc(size, kAllocKind); 52 void* operator new(size_t size, ArenaAllocator* allocator) { argument 53 return allocator->Alloc(size, kAllocKind);
|
H A D | allocator.h | 103 // Tracking allocator for use with STL types, tracks how much memory is used. 105 class TrackingAllocatorImpl : public std::allocator<T> { 107 typedef typename std::allocator<T>::value_type value_type; 108 typedef typename std::allocator<T>::size_type size_type; 109 typedef typename std::allocator<T>::difference_type difference_type; 110 typedef typename std::allocator<T>::pointer pointer; 111 typedef typename std::allocator<T>::const_pointer const_pointer; 112 typedef typename std::allocator<T>::reference reference; 113 typedef typename std::allocator<T>::const_reference const_reference; 126 // Enables an allocator fo [all...] |
/art/compiler/dex/quick/ |
H A D | lazy_debug_frame_opcode_writer.h | 44 ArenaAllocator* allocator) 45 : Base(enable_writes, allocator->Adapter()), 47 advances_(allocator->Adapter()), 43 LazyDebugFrameOpCodeWriter(LIR** last_lir_insn, bool enable_writes, ArenaAllocator* allocator) argument
|
/art/runtime/entrypoints/quick/ |
H A D | quick_alloc_entrypoints.h | 33 void SetQuickAllocEntryPointsAllocator(gc::AllocatorType allocator)
|
/art/compiler/dex/ |
H A D | global_value_numbering.cc | 25 GlobalValueNumbering::GlobalValueNumbering(CompilationUnit* cu, ScopedArenaAllocator* allocator, argument 29 allocator_(allocator), 35 global_value_map_(std::less<uint64_t>(), allocator->Adapter()), 36 array_location_map_(ArrayLocationComparator(), allocator->Adapter()), 37 array_location_reverse_map_(allocator->Adapter()), 38 ref_set_map_(std::less<ValueNameSet>(), allocator->Adapter()), 39 lvns_(mir_graph_->GetNumBlocks(), nullptr, allocator->Adapter()), 41 merge_lvns_(allocator->Adapter()) { 49 ScopedArenaAllocator* allocator) { 67 if (allocator 48 PrepareBasicBlock(BasicBlock* bb, ScopedArenaAllocator* allocator) argument [all...] |
/art/runtime/gc/allocator/ |
H A D | dlmalloc.h | 57 namespace allocator { namespace in namespace:art::gc 63 } // namespace allocator
|
/art/runtime/gc/space/ |
H A D | rosalloc_space.cc | 46 // template class ValgrindMallocSpace<RosAllocSpace, allocator::RosAlloc*>; 49 art::gc::allocator::RosAlloc* rosalloc, uint8_t* begin, uint8_t* end, 66 allocator::RosAlloc* rosalloc = CreateRosAlloc(mem_map->Begin(), starting_size, initial_size, 134 allocator::RosAlloc* RosAllocSpace::CreateRosAlloc(void* begin, size_t morecore_start, 143 allocator::RosAlloc* rosalloc = new art::gc::allocator::RosAlloc( 146 art::gc::allocator::RosAlloc::kPageReleaseModeAll : 147 art::gc::allocator::RosAlloc::kPageReleaseModeSizeAndEnd, 180 void* allocator, uint8_t* begin, uint8_t* end, 185 mem_map, initial_size_, name, reinterpret_cast<allocator 179 CreateInstance(MemMap* mem_map, const std::string& name, void* allocator, uint8_t* begin, uint8_t* end, uint8_t* limit, size_t growth_limit, bool can_move_objects) argument 379 namespace allocator { namespace in namespace:art::gc [all...] |