/art/runtime/gc/accounting/ |
H A D | read_barrier_table.h | 42 CHECK(mem_map != nullptr && mem_map->Begin() != nullptr) 47 uint8_t* entry_start = EntryFromAddr(space->Begin()); 70 memset(mem_map_->Begin(), kSetEntryValue, mem_map_->Size()); 73 for (uint32_t* p = reinterpret_cast<uint32_t*>(mem_map_->Begin()); 91 uint8_t* entry_addr = mem_map_->Begin() + reinterpret_cast<uintptr_t>(heap_addr) / kRegionSize; 107 uint8_t* begin = mem_map_->Begin();
|
H A D | bitmap.cc | 33 : mem_map_(mem_map), bitmap_begin_(reinterpret_cast<uintptr_t*>(mem_map->Begin())), 73 std::copy(source_bitmap->Begin(), 74 source_bitmap->Begin() + BitmapSize() / kBitsPerBitmapWord, Begin());
|
H A D | atomic_stack.h | 164 StackReference<T>* Begin() const { function in class:art::gc::accounting::AtomicStack 185 std::sort(Begin(), End(), ObjectComparator()); 195 return std::binary_search(Begin(), End(), value, ObjectComparator()); 199 for (auto cur = Begin(), end = End(); cur != end; ++cur) { 242 uint8_t* addr = mem_map_->Begin();
|
H A D | card_table.cc | 72 uint8_t* cardtable_begin = mem_map->Begin(); 100 uint8_t* card_start = CardFromAddr(space->Begin()); 136 uint8_t* begin = mem_map_->Begin() + offset_;
|
H A D | mod_union_table.cc | 162 card_table->ModifyCardsAtomic(space_->Begin(), space_->End(), AgeCardVisitor(), visitor); 359 "mod union bitmap", reinterpret_cast<uintptr_t>(space->Begin()), 399 card_table->ModifyCardsAtomic(space_->Begin(), space_->End(), AgeCardVisitor(), visitor); 417 for (uint8_t* addr = space_->Begin(); addr < AlignUp(space_->End(), CardTable::kCardSize); 429 for (uint8_t* addr = space_->Begin(); addr < AlignUp(space_->End(), CardTable::kCardSize); 440 for (uint8_t* addr = space_->Begin(); addr < AlignUp(space_->End(), CardTable::kCardSize);
|
H A D | remembered_set.cc | 59 card_table->ModifyCardsAtomic(space_->Begin(), space_->End(), AgeCardVisitor(), card_visitor); 170 DCHECK_LE(space_->Begin(), start);
|
H A D | space_bitmap.cc | 46 uintptr_t* bitmap_begin = reinterpret_cast<uintptr_t*>(mem_map->Begin()); 107 std::copy(source_bitmap->Begin(), source_bitmap->Begin() + source_bitmap->Size() / sizeof(intptr_t), Begin());
|
/art/runtime/gc/collector/ |
H A D | immune_region.cc | 41 mirror::Object* space_begin = reinterpret_cast<mirror::Object*>(space->Begin()); 60 begin_ <= reinterpret_cast<mirror::Object*>(space->Begin()) &&
|
H A D | immune_region.h | 60 mirror::Object* Begin() { function in class:art::gc::collector::ImmuneRegion
|
H A D | mark_compact.cc | 121 bump_pointer_ = reinterpret_cast<uint8_t*>(space_->Begin()); 124 objects_before_forwarding_->VisitMarkedRange(reinterpret_cast<uintptr_t>(space_->Begin()), 189 "objects before forwarding", space_->Begin(), space_->Size())); 192 "objects with lock words", space_->Begin(), space_->Size())); 275 std::vector<StackReference<Object>> temp(mark_stack_->Begin(), mark_stack_->End()); 397 bitmap->VisitMarkedRange(reinterpret_cast<uintptr_t>(space->Begin()), 409 objects_before_forwarding_->VisitMarkedRange(reinterpret_cast<uintptr_t>(space_->Begin()), 561 objects_before_forwarding_->VisitMarkedRange(reinterpret_cast<uintptr_t>(space_->Begin()),
|
/art/runtime/gc/space/ |
H A D | bump_pointer_space.cc | 57 : ContinuousMemMapAllocSpace(name, mem_map, mem_map->Begin(), mem_map->Begin(), mem_map->End(), 69 memset(Begin(), 0, Limit() - Begin()); 71 CHECK_NE(madvise(Begin(), Limit() - Begin(), MADV_DONTNEED), -1) << "madvise failed"; 74 SetEnd(Begin()); 87 << reinterpret_cast<void*>(Begin()) << "-" << reinterpret_cast<void*>(End()) << " - " 157 uint8_t* pos = Begin(); 167 main_end = Begin() [all...] |
H A D | malloc_space.cc | 49 CHECK(IsAligned<kGcCardSize>(reinterpret_cast<uintptr_t>(mem_map->Begin()))); 53 Begin(), NonGrowthLimitCapacity())); 58 Begin(), NonGrowthLimitCapacity())); 139 CHECK_LE(new_end, Begin() + Capacity()); 143 CHECK_GE(original_end + increment, Begin()); 178 VLOG(heap) << "Begin " << reinterpret_cast<const void*>(begin_) << "\n" 198 uint8_t* end = mem_map->Begin() + starting_size_; 223 << " begin=" << reinterpret_cast<void*>(Begin()) 261 limit_ = Begin() + new_capacity;
|
H A D | region_space.cc | 48 : ContinuousMemMapAllocSpace(name, mem_map, mem_map->Begin(), mem_map->End(), mem_map->End(), 53 CHECK_ALIGNED(mem_map->Begin(), kRegionSize); 58 uint8_t* region_addr = mem_map->Begin(); 63 CHECK_EQ(regions_[0].Begin(), Begin()); 66 CHECK_EQ(static_cast<size_t>(regions_[i].End() - regions_[i].Begin()), kRegionSize); 68 CHECK_EQ(regions_[i].End(), regions_[i + 1].Begin()); 197 rb_table->Clear(r->Begin(), r->End()); 285 << reinterpret_cast<void*>(Begin()) << "-" << reinterpret_cast<void*>(Limit()); 352 self->SetTlab(r->Begin(), [all...] |
H A D | large_object_space.h | 82 uint8_t* Begin() const { function in class:art::gc::space::LargeObjectSpace 91 return End() - Begin(); 96 return Begin() <= byte_obj && byte_obj < End(); 116 // Begin and end, may change as more large objects are allocated. 177 return (address - reinterpret_cast<uintptr_t>(Begin())) / kAlignment; 183 return reinterpret_cast<uintptr_t>(Begin()) + slot * kAlignment;
|
H A D | zygote_space.cc | 51 live_bitmap->VisitMarkedRange(reinterpret_cast<uintptr_t>(mem_map->Begin()), 67 : ContinuousMemMapAllocSpace(name, mem_map, mem_map->Begin(), mem_map->End(), mem_map->End(), 74 << " begin=" << reinterpret_cast<void*>(Begin())
|
H A D | dlmalloc_space.cc | 51 void* mspace = CreateMspace(mem_map->Begin(), starting_size, initial_size); 58 uint8_t* end = mem_map->Begin() + starting_size; 64 uint8_t* begin = mem_map->Begin(); 278 madvise(GetMemMap()->Begin(), GetMemMap()->Size(), MADV_DONTNEED); 281 SetEnd(Begin() + starting_size_); 282 mspace_ = CreateMspace(mem_map_->Begin(), starting_size_, initial_size_);
|
H A D | bump_pointer_space.h | 43 // guaranteed to be granted, if it is required, the caller should call Begin on the returned 116 return Begin() == End(); 125 return byte_obj >= Begin() && byte_obj < End(); 169 // allocation. The main block starts at the space Begin().
|
H A D | image_space.h | 73 return *reinterpret_cast<ImageHeader*>(Begin());
|
H A D | region_space-inl.h | 238 mirror::Object* obj = reinterpret_cast<mirror::Object*>(r->Begin()); 245 uint8_t* pos = r->Begin(); 304 first_reg->SetTop(first_reg->Begin() + num_bytes); 316 return reinterpret_cast<mirror::Object*>(first_reg->Begin());
|
H A D | space.h | 270 uint8_t* Begin() const { function in class:art::gc::space::ContinuousSpace 296 return End() - Begin(); 304 return Limit() - Begin(); 311 return byte_ptr >= Begin() && byte_ptr < Limit();
|
/art/runtime/base/ |
H A D | scoped_arena_allocator.cc | 72 return top_arena_->Begin(); 87 size_t allocated = static_cast<size_t>(top_ptr_ - top_arena_->Begin()); 141 mark_ptr_ = arena_stack_->top_ptr_ = mark_arena_->Begin();
|
/art/runtime/ |
H A D | mem_map.h | 109 uint8_t* Begin() const { function in class:art::MemMap 121 return Begin() + Size(); 137 return Begin() <= addr && addr < End();
|
H A D | elf_file.cc | 222 program_headers_start_ = Begin() + GetHeader().e_phoff; 302 Begin() + section_header->sh_offset) { 306 << " != " << reinterpret_cast<void*>(Begin() + section_header->sh_offset); 344 if (Begin() + offset >= End()) { 349 *target = Begin() + offset; 368 if (Begin() + section_header->sh_offset == source) { 374 } else if (Begin() + section_header->sh_offset == target) { 479 CHECK(map_->Begin() != nullptr) << file_->GetPath(); 481 header_ = reinterpret_cast<Elf_Ehdr*>(map_->Begin()); 1002 uint8_t* strings = Begin() 1859 uint8_t* ElfFile::Begin() const { function in class:art::ElfFile [all...] |
H A D | elf_file.h | 56 uint8_t* Begin() const;
|
/art/runtime/jit/ |
H A D | jit_code_cache.cc | 48 uint8_t* divider = mem_map->Begin() + RoundUp(mem_map->Size() / 4, kPageSize); 51 data_cache_ptr_ = mem_map->Begin();
|