/art/runtime/gc/collector/ |
H A D | immune_region.h | 56 end_ = end; 64 return end_; 72 size_ = reinterpret_cast<uintptr_t>(end_) - reinterpret_cast<uintptr_t>(begin_); 76 mirror::Object* end_; member in class:art::gc::collector::ImmuneRegion
|
H A D | immune_region.cc | 49 } else if (space_begin >= end_) { // Space is after the immune region. 61 end_ >= reinterpret_cast<mirror::Object*>(space->Limit());
|
H A D | mark_sweep.cc | 771 end_(end), 778 uint8_t* const end_; member in class:art::gc::collector::CardScanTask 790 card_table->Scan<true>(bitmap_, begin_, end_, visitor, minimum_age_) : 791 card_table->Scan<false>(bitmap_, begin_, end_, visitor, minimum_age_); 793 << reinterpret_cast<void*>(end_) << " = " << cards_scanned; 913 end_(end) { 919 const uintptr_t end_; member in class:art::gc::collector::RecursiveMarkTask 928 bitmap_->VisitMarkedRange(begin_, end_, visitor);
|
/art/runtime/ |
H A D | mapping_table.h | 69 table_(table), element_(element), end_(table_->DexToPcSize()), encoded_table_ptr_(nullptr), 72 if (end_ > 0) { 90 if (element_ != end_) { // Avoid reading beyond the end of the table. 107 uint32_t element_; // A value in the range 0 to end_. 108 const uint32_t end_; // Equal to table_->DexToPcSize(). member in class:art::MappingTable::DexToPcIterator 146 table_(table), element_(element), end_(table_->PcToDexSize()), encoded_table_ptr_(nullptr), 149 if (end_ > 0) { 167 if (element_ != end_) { // Avoid reading beyond the end of the table. 185 const uint32_t end_; // Equal to table_->PcToDexSize(). member in class:art::MappingTable::PcToDexIterator
|
H A D | oat_file.cc | 97 oat_file->end_ = elf_file->Begin() + size + offset; 202 : location_(location), begin_(nullptr), end_(nullptr), bss_begin_(nullptr), bss_end_(nullptr), 255 end_ = reinterpret_cast<uint8_t*>(dlsym(dlopen_handle_, "oatlastword")); 256 if (end_ == nullptr) { 262 end_ += sizeof(uint32_t); 353 end_ = elf_file_->FindDynamicSymbolAddress("oatlastword"); 354 if (end_ == nullptr) { 359 end_ += sizeof(uint32_t); 516 CHECK(end_ != nullptr); 517 return end_; [all...] |
H A D | oat_file.h | 304 const uint8_t* end_; member in class:art::FINAL
|
/art/runtime/jdwp/ |
H A D | jdwp_request.cc | 30 end_ = bytes + byte_count_; 47 if (p_ < end_) { 48 CHECK(p_ == end_) << "read too few bytes: " << (end_ - p_); 49 } else if (p_ > end_) { 50 CHECK(p_ == end_) << "read too many bytes: " << (p_ - end_);
|
H A D | jdwp.h | 477 size_t size() { return end_ - p_; } 498 const uint8_t* end_; member in class:art::JDWP::Request
|
/art/runtime/gc/space/ |
H A D | bump_pointer_space-inl.h | 48 uint8_t* end = end_.LoadRelaxed(); 53 end_.StoreRelaxed(end + num_bytes); 70 old_end = end_.LoadRelaxed(); 76 } while (!end_.CompareExchangeWeakSequentiallyConsistent(old_end, new_end));
|
H A D | region_space.h | 235 begin_(nullptr), top_(nullptr), end_(nullptr), 241 : idx_(idx), begin_(begin), top_(begin), end_(end), 265 memset(begin_, 0, end_ - begin_); 267 madvise(begin_, end_ - begin_, MADV_DONTNEED); 443 return end_; 447 return begin_ <= reinterpret_cast<uint8_t*>(ref) && reinterpret_cast<uint8_t*>(ref) < end_; 455 DCHECK_EQ(top_, end_); 458 DCHECK_EQ(top_, end_); 466 uint8_t* end_; // The end address of the region. member in class:art::gc::space::FINAL::Region
|
H A D | region_space-inl.h | 125 if (UNLIKELY(new_top > end_)) { 130 DCHECK_LE(atomic_top->LoadRelaxed(), end_); 131 DCHECK_LT(old_top, end_); 132 DCHECK_LE(new_top, end_);
|
H A D | space.h | 276 return end_.LoadRelaxed(); 287 end_.StoreRelaxed(end); 327 Space(name, gc_retention_policy), begin_(begin), end_(end), limit_(limit) { 334 Atomic<uint8_t*> end_; member in class:art::gc::space::ContinuousSpace
|
H A D | large_object_space.cc | 109 total_objects_allocated_(0), begin_(begin), end_(end) { 156 if (end_ == nullptr || obj_end > end_) { 157 end_ = obj_end; 380 const uintptr_t free_end_start = reinterpret_cast<uintptr_t>(end_) - free_end_; 418 uintptr_t free_end_start = reinterpret_cast<uintptr_t>(end_) - free_end_; 536 uintptr_t free_end_start = reinterpret_cast<uintptr_t>(end_) - free_end_; 566 uintptr_t free_end_start = reinterpret_cast<uintptr_t>(end_) - free_end_;
|
H A D | large_object_space.h | 87 return end_; 118 uint8_t* end_; member in class:art::gc::space::LargeObjectSpace
|
H A D | region_space.cc | 412 << "-" << reinterpret_cast<void*>(end_)
|
/art/cmdline/ |
H A D | token_range.h | 43 end_(token_list_->end()) 51 end_(token_list_->end()) 61 end_(token_list_->end()) { 71 end_(token_list_->end()) 78 end_(token_list_->end()) 87 end_(it_end) { 102 end_(token_list_->end()) 150 return end_; 155 return std::distance(begin_, end_); 395 for (auto it = begin_; it != end_; 421 const iterator end_; member in struct:art::TokenRange [all...] |
/art/runtime/gc/accounting/ |
H A D | space_bitmap_test.cc | 46 end_(end) {} 50 EXPECT_TRUE(obj <= end_); 56 const mirror::Object* end_; member in class:art::gc::accounting::BitmapVerify
|
/art/runtime/base/ |
H A D | arena_allocator.cc | 255 end_(nullptr), 271 if (UNLIKELY(ptr_ + rounded_bytes > end_)) { 302 end_ = new_arena->End(); 306 if (ptr >= begin_ && ptr < end_) { 333 (arena_head_ == nullptr) ? 0 : (end_ - ptr_) - arena_head_->RemainingSpace();
|
H A D | arena_allocator.h | 214 if (UNLIKELY(ptr_ + bytes > end_)) { 237 const size_t remain = end_ - ptr_; 280 uint8_t* end_; member in class:art::ArenaAllocator
|
/art/compiler/optimizing/ |
H A D | ssa_liveness_analysis.h | 60 LiveRange(size_t start, size_t end, LiveRange* next) : start_(start), end_(end), next_(next) { 66 size_t GetEnd() const { return end_; } 70 return (start_ >= other.start_ && start_ < other.end_) 71 || (other.start_ >= start_ && other.start_ < end_); 75 return end_ <= other.start_; 79 stream << "[" << start_ << ", " << end_ << ")"; local 84 start_, end_, next_ == nullptr ? nullptr : next_->Dup(allocator)); 93 size_t end_; member in class:art::FINAL 286 first_range_->end_ = position; 378 first_range_->end_ [all...] |
H A D | bounds_check_elimination.cc | 319 // initial_ to end_. So adding deoptimization might be too aggressive and can 503 end_(nullptr), 513 void SetEnd(HInstruction* end) { end_ = end; } 671 // initial_ to end_. We do the following checks to make sure that's the case: 673 // b) The increment_ is 1/-1. An increment of 2, for example, may skip end_. 678 // when the induction variable is at initial_ and end_ must be in a legal range. 680 // at initial_ and end_ values, no deoptimization will be triggered either. 684 // deoptimization, we do a test between initial_ and end_ first and skip over 690 // end_ so that boundaries are covered by the loop. An increment of 2, 691 // for example, may skip end_ 1088 HInstruction* end_; // End value. member in class:art::MonotonicValueRange [all...] |
/art/compiler/driver/ |
H A D | compiler_driver.cc | 1641 end_(end), 1647 if (UNLIKELY(index >= end_)) { 1661 const size_t end_; member in class:art::ParallelCompilationManager::ForAllClosure
|