/art/runtime/base/ |
H A D | bit_vector.cc | 40 Allocator* allocator, 43 : allocator_(allocator), 38 BitVector(uint32_t start_bits, bool expandable, Allocator* allocator, uint32_t storage_size, uint32_t* storage) argument
|
/art/runtime/gc/space/ |
H A D | dlmalloc_space.cc | 152 void* allocator, byte* begin, byte* end, 155 return new DlMallocSpace(name, mem_map, allocator, begin, end, limit, growth_limit, 151 CreateInstance(const std::string& name, MemMap* mem_map, void* allocator, byte* begin, byte* end, byte* limit, size_t growth_limit, bool can_move_objects) argument
|
H A D | rosalloc_space.cc | 42 // template class ValgrindMallocSpace<RosAllocSpace, allocator::RosAlloc*>; 45 art::gc::allocator::RosAlloc* rosalloc, byte* begin, byte* end, 59 allocator::RosAlloc* rosalloc = CreateRosAlloc(mem_map->Begin(), starting_size, initial_size, 125 allocator::RosAlloc* RosAllocSpace::CreateRosAlloc(void* begin, size_t morecore_start, 133 allocator::RosAlloc* rosalloc = new art::gc::allocator::RosAlloc( 136 art::gc::allocator::RosAlloc::kPageReleaseModeAll : 137 art::gc::allocator::RosAlloc::kPageReleaseModeSizeAndEnd); 166 MallocSpace* RosAllocSpace::CreateInstance(const std::string& name, MemMap* mem_map, void* allocator, argument 169 return new RosAllocSpace(name, mem_map, reinterpret_cast<allocator [all...] |
/art/compiler/dex/ |
H A D | local_value_numbering.h | 39 LocalValueNumbering(GlobalValueNumbering* gvn, BasicBlockId id, ScopedArenaAllocator* allocator); 80 static void* operator new(size_t size, ScopedArenaAllocator* allocator) { argument 81 return allocator->Alloc(sizeof(LocalValueNumbering), kArenaAllocMisc);
|
H A D | mir_optimization.cc | 326 std::unique_ptr<ScopedArenaAllocator> allocator; local 330 allocator.reset(ScopedArenaAllocator::Create(&cu_->arena_stack)); 331 global_valnum.reset(new (allocator.get()) GlobalValueNumbering(cu_, allocator.get())); 332 local_valnum.reset(new (allocator.get()) LocalValueNumbering(global_valnum.get(), bb->id, 333 allocator.get())); 993 ScopedArenaAllocator allocator(&cu_->arena_stack); 995 allocator.Adapter()); 1174 ScopedArenaAllocator allocator(&cu_->arena_stack); // Reclaim memory after each LVN. 1175 LocalValueNumbering* lvn = temp_gvn_->PrepareBasicBlock(bb, &allocator); [all...] |
H A D | local_value_numbering.cc | 311 // with our allocator, rather than the allocator of the source. 314 it->second = entry.second; // Map assignments preserve current allocator. 319 ScopedArenaAllocator* allocator) 322 sreg_value_map_(std::less<uint16_t>(), allocator->Adapter()), 323 sreg_wide_value_map_(std::less<uint16_t>(), allocator->Adapter()), 324 sfield_value_map_(std::less<uint16_t>(), allocator->Adapter()), 325 non_aliasing_ifield_value_map_(std::less<uint16_t>(), allocator->Adapter()), 326 aliasing_ifield_value_map_(std::less<uint16_t>(), allocator->Adapter()), 327 non_aliasing_array_value_map_(std::less<uint16_t>(), allocator 318 LocalValueNumbering(GlobalValueNumbering* gvn, uint16_t id, ScopedArenaAllocator* allocator) argument [all...] |
H A D | mir_graph.cc | 1539 const ScopedArenaVector<size_t>* visited_cnt_values, ScopedArenaAllocator* allocator, 1547 ArenaBitVector candidate_reachable(allocator, mir_graph->GetNumBlocks(), false, kBitMapMisc); 1615 ScopedArenaAllocator allocator(&cu_->arena_stack); 1618 ScopedArenaQueue<BasicBlock*> q(allocator.Adapter()); 1619 ScopedArenaVector<size_t> visited_cnt_values(num_blocks, 0u, allocator.Adapter()); 1620 ScopedArenaVector<BasicBlockId> loop_head_stack(allocator.Adapter()); 1622 ArenaBitVector loop_exit_blocks(&allocator, num_blocks, false, kBitMapMisc); 1663 ScopedArenaVector<ArenaBitVector*> loop_head_reachable_from(allocator.Adapter()); 1666 ScopedArenaVector<BasicBlockId> tmp_stack(allocator.Adapter()); 1704 &allocator, 1537 SelectTopologicalSortOrderFallBack( MIRGraph* mir_graph, const ArenaBitVector* current_loop, const ScopedArenaVector<size_t>* visited_cnt_values, ScopedArenaAllocator* allocator, ScopedArenaVector<BasicBlockId>* tmp_stack) argument [all...] |
/art/compiler/optimizing/ |
H A D | register_allocator.cc | 27 RegisterAllocator::RegisterAllocator(ArenaAllocator* allocator, argument 30 : allocator_(allocator), 33 unhandled_(allocator, 0), 34 handled_(allocator, 0), 35 active_(allocator, 0), 36 inactive_(allocator, 0), 37 physical_register_intervals_(allocator, codegen->GetNumberOfRegisters()), 38 spill_slots_(allocator, kDefaultNumberOfSpillSlots), 42 blocked_registers_(allocator->AllocArray<bool>(codegen->GetNumberOfRegisters())) { 103 // TODO: make the register allocator understan 226 ValidateIntervals(const GrowableArray<LiveInterval*>& intervals, size_t number_of_spill_slots, const CodeGenerator& codegen, ArenaAllocator* allocator, bool processing_core_registers, bool log_fatal_on_failure) argument [all...] |
H A D | ssa_liveness_analysis.h | 28 BlockInfo(ArenaAllocator* allocator, const HBasicBlock& block, size_t number_of_ssa_values) argument 30 live_in_(allocator, number_of_ssa_values, false), 31 live_out_(allocator, number_of_ssa_values, false), 32 kill_(allocator, number_of_ssa_values, false) { 136 LiveInterval(ArenaAllocator* allocator, Primitive::Type type, HInstruction* defined_by = nullptr) argument 137 : allocator_(allocator), 149 static LiveInterval* MakeFixedInterval(ArenaAllocator* allocator, int reg, Primitive::Type type) { argument 150 LiveInterval* interval = new (allocator) LiveInterval(allocator, type);
|
/art/runtime/native/ |
H A D | dalvik_system_VMRuntime.cc | 25 #include "gc/allocator/dlmalloc.h" 77 gc::AllocatorType allocator = runtime->GetHeap()->GetCurrentNonMovingAllocator(); local 79 array_class->GetComponentSize(), allocator); 101 gc::AllocatorType allocator = runtime->GetHeap()->GetCurrentAllocator(); local 103 array_class->GetComponentSize(), allocator,
|
/art/runtime/gc/allocator/ |
H A D | rosalloc.cc | 31 namespace allocator { namespace in namespace:art::gc 2216 } // namespace allocator
|
/art/runtime/gc/ |
H A D | heap.h | 75 namespace allocator { namespace in namespace:art::gc 77 } // namespace allocator 194 Thread* self, mirror::Class* klass, size_t byte_count, AllocatorType allocator, 216 // Change the allocator, updates entrypoints. 217 void ChangeAllocator(AllocatorType allocator) 530 space::RosAllocSpace* GetRosAllocSpace(gc::allocator::RosAlloc* rosalloc) const; 652 mirror::Object* AllocateInternalWithGc(Thread* self, AllocatorType allocator, size_t num_bytes,
|
H A D | heap.cc | 26 #include "base/allocator.h" 194 // If we aren't the zygote, switch to the default non zygote allocator. This may update the 497 void Heap::ChangeAllocator(AllocatorType allocator) { argument 498 if (current_allocator_ != allocator) { 500 CHECK_NE(allocator, kAllocatorTypeLOS); 501 CHECK_NE(allocator, kAllocatorTypeNonMoving); 502 current_allocator_ = allocator; 1217 space::RosAllocSpace* Heap::GetRosAllocSpace(gc::allocator::RosAlloc* rosalloc) const { 1228 mirror::Object* Heap::AllocateInternalWithGc(Thread* self, AllocatorType allocator, argument 1232 bool was_default_allocator = allocator [all...] |