/art/runtime/gc/space/ |
H A D | malloc_space.cc | 55 CHECK_ALIGNED(reinterpret_cast<uintptr_t>(mem_map->End()), kGcCardSize); 137 uint8_t* original_end = End(); 171 SetEnd(reinterpret_cast<uint8_t*>(RoundUp(reinterpret_cast<uintptr_t>(End()), kPageSize))); 173 DCHECK_ALIGNED(End(), accounting::CardTable::kCardSize); 175 DCHECK_ALIGNED(End(), kPageSize); 184 << "End " << reinterpret_cast<const void*>(End()) << "\n" 197 std::unique_ptr<MemMap> mem_map(GetMemMap()->RemapAtEnd(End(), alloc_space_name, 200 void* allocator = CreateAllocator(End(), starting_size_, initial_size_, capacity, 207 *out_malloc_space = CreateInstance(mem_map.release(), alloc_space_name, allocator, End(), en [all...] |
H A D | bump_pointer_space-walk-inl.h | 33 uint8_t* end = End(); 84 CHECK_LE(reinterpret_cast<const uint8_t*>(end_obj), End());
|
H A D | zygote_space.cc | 53 reinterpret_cast<uintptr_t>(mem_map->End()), visitor); 68 : ContinuousMemMapAllocSpace(name, mem_map, mem_map->Begin(), mem_map->End(), mem_map->End(), 76 << ",end=" << reinterpret_cast<void*>(End())
|
H A D | region_space.cc | 75 CHECK_ALIGNED(mem_map->End(), kRegionSize); 85 : ContinuousMemMapAllocSpace(name, mem_map, mem_map->Begin(), mem_map->End(), mem_map->End(), 110 CHECK_EQ(static_cast<size_t>(regions_[i].End() - regions_[i].Begin()), kRegionSize); 112 CHECK_EQ(regions_[i].End(), regions_[i + 1].Begin()); 115 CHECK_EQ(regions_[num_regions_ - 1].End(), Limit()); 250 rb_table->Clear(r->Begin(), r->End()); 305 clear_block_end = r->End(); 394 if (current_region_->End() - current_region_->Top() > 0) { 395 max_contiguous_allocation = current_region_->End() [all...] |
H A D | bump_pointer_space.cc | 57 : ContinuousMemMapAllocSpace(name, mem_map, mem_map->Begin(), mem_map->Begin(), mem_map->End(), 59 growth_end_(mem_map->End()), 87 << reinterpret_cast<void*>(Begin()) << "-" << reinterpret_cast<void*>(End()) << " - " 216 size_t max_contiguous_allocation = Limit() - End();
|
H A D | large_object_space.h | 87 uint8_t* End() const { function in class:art::gc::space::LargeObjectSpace 92 return End() - Begin(); 97 return Begin() <= byte_obj && byte_obj < End(); 108 // GetRangeAtomic returns Begin() and End() atomically, that is, it never returns Begin() and 109 // End() from different allocations.
|
/art/dexlayout/ |
H A D | dex_container.h | 50 uint8_t* End() { function in class:art::DexContainer::Section
|
/art/runtime/gc/collector/ |
H A D | immune_region.h | 65 mirror::Object* End() const { function in class:art::gc::collector::ImmuneRegion
|
H A D | immune_spaces_test.cc | 51 map->End()), 111 std::unique_ptr<DummyOatFile> oat_file(new DummyOatFile(oat_map->Begin(), oat_map->End())); 203 EXPECT_EQ(reinterpret_cast<uint8_t*>(spaces.GetLargestImmuneRegion().End()), b.Limit()); 235 EXPECT_EQ(reinterpret_cast<uint8_t*>(spaces.GetLargestImmuneRegion().End()), 250 EXPECT_EQ(reinterpret_cast<uint8_t*>(spaces.GetLargestImmuneRegion().End()), space.Limit()); 309 EXPECT_EQ(reinterpret_cast<uint8_t*>(spaces.GetLargestImmuneRegion().End()), 320 EXPECT_EQ(reinterpret_cast<uint8_t*>(spaces.GetLargestImmuneRegion().End()), 344 EXPECT_EQ(reinterpret_cast<uint8_t*>(spaces.GetLargestImmuneRegion().End()), 366 EXPECT_EQ(reinterpret_cast<uint8_t*>(spaces.GetLargestImmuneRegion().End()), space5->Limit());
|
H A D | immune_spaces.cc | 59 reinterpret_cast<uintptr_t>(image_oat_file->End()), 104 << largest_immune_region_.End();
|
H A D | mark_compact.cc | 103 reinterpret_cast<uintptr_t>(space_->End()), 254 std::vector<StackReference<mirror::Object>> temp(mark_stack_->Begin(), mark_stack_->End()); 367 reinterpret_cast<uintptr_t>(space->End()), 379 reinterpret_cast<uintptr_t>(space_->End()), 393 int64_t bytes_freed = reinterpret_cast<int64_t>(space_->End()) - 524 reinterpret_cast<uintptr_t>(space_->End()),
|
/art/compiler/debug/ |
H A D | elf_debug_writer.cc | 93 line_writer.End(write_oat_patches); 104 info_writer.End(write_oat_patches); 185 builder->End(); 219 info_writer.End(false /* write_oat_patches */); 221 builder->End();
|
H A D | elf_symtab_writer.h | 113 strtab->End();
|
/art/runtime/gc/accounting/ |
H A D | read_barrier_table.h | 76 p < reinterpret_cast<uint32_t*>(mem_map_->End()); ++p) { 110 uint8_t* end = mem_map_->End();
|
H A D | atomic_stack.h | 180 StackReference<T>* End() const { function in class:art::gc::accounting::AtomicStack 198 std::sort(Begin(), End(), ObjectComparator()); 208 return std::binary_search(Begin(), End(), value, ObjectComparator()); 212 for (auto cur = Begin(), end = End(); cur != end; ++cur) {
|
H A D | card_table-inl.h | 217 << " end: " << reinterpret_cast<void*>(mem_map_->End()); 232 uint8_t* end = mem_map_->End(); 240 << " end: " << reinterpret_cast<void*>(mem_map_->End());
|
H A D | card_table.cc | 122 uint8_t* end = mem_map_->End();
|
/art/compiler/utils/ |
H A D | swap_space.cc | 204 CHECK_LE(prev->End(), chunk.Start()); 205 if (prev->End() == chunk.Start()) { 216 CHECK_LE(chunk.End(), it->Start()); 217 if (chunk.End() == it->Start()) {
|
/art/dex2oat/linker/ |
H A D | elf_writer.h | 77 virtual bool End() = 0;
|
H A D | elf_writer_quick.cc | 119 bool End() OVERRIDE; 226 builder_->GetRoData()->End(); 239 builder_->GetText()->End(); 289 bool ElfWriterQuick<ElfTypes>::End() { function in class:art::linker::ElfWriterQuick 290 builder_->End(); 311 // End of file.
|
/art/runtime/base/ |
H A D | scoped_arena_allocator.cc | 68 top_end_ = top_arena_->End(); 157 mark_end_ = arena_stack_->top_end_ = mark_arena_->End();
|
/art/runtime/ |
H A D | mem_map.h | 66 // [source->Begin(), source->End()]. 180 uint8_t* End() const { function in class:art::MemMap 197 return Begin() <= addr && addr < End();
|
H A D | elf_file.h | 66 uint8_t* End() const;
|
H A D | oat_file.h | 279 return End() - Begin(); 283 return p >= Begin() && p < End(); 309 const uint8_t* End() const;
|
/art/runtime/gc/ |
H A D | heap-visit-objects-inl.h | 112 for (auto* it = allocation_stack_->Begin(), *end = allocation_stack_->End(); it < end; ++it) {
|