/art/disassembler/ |
H A D | disassembler_arm64.cc | 42 void DisassemblerArm64::Dump(std::ostream& os, const uint8_t* begin, const uint8_t* end) { argument 43 for (const uint8_t* cur = begin; cur < end; cur += vixl::kInstructionSize) {
|
H A D | disassembler_mips.cc | 264 void DisassemblerMips::Dump(std::ostream& os, const uint8_t* begin, const uint8_t* end) { argument 265 for (const uint8_t* cur = begin; cur < end; cur += 4) {
|
/art/runtime/base/ |
H A D | stl_util.h | 28 std::sort(v->begin(), v->end()); 29 v->erase(std::unique(v->begin(), v->end()), v->end()); 44 ForwardIterator end) { 45 while (begin != end) { 54 // hash_set, or any other STL container which defines sensible begin(), end(), 65 STLDeleteContainerPointers(container->begin(), container->end()); 75 for (typename T::iterator i = v->begin(); i != v->end(); ++i) { 43 STLDeleteContainerPointers(ForwardIterator begin, ForwardIterator end) argument
|
H A D | bit_vector.h | 43 * object with begin() and end() suitable for range-based loops: 78 // Helper function to check for end without comparing with bit_vector.Indexes().end(). 138 IndexIterator end() const { function in class:art::BitVector::IndexContainer 183 // Number of bits set in range [0, end). 184 uint32_t NumSetBits(uint32_t end) const; 204 // Number of bits set in range [0, end) in storage. (No range check.) 205 static uint32_t NumSetBits(const uint32_t* storage, uint32_t end);
|
H A D | stringpiece.h | 118 // Does "this" end with "x" 137 iterator end() const { return ptr_ + length_; } function in class:art::StringPiece
|
/art/test/028-array-write/src/ |
H A D | Main.java | 27 static public void report(long start, long end) { argument 32 System.out.println("Finished in " + ((end - start) / 1000000.0) 42 long start, end; 50 end = System.nanoTime(); 52 report(start, end); 56 long start, end; 68 end = System.nanoTime(); 70 report(start, end);
|
/art/runtime/gc/allocator/ |
H A D | dlmalloc.cc | 53 extern "C" void DlmallocMadviseCallback(void* start, void* end, size_t used_bytes, void* arg) { argument 60 end = reinterpret_cast<void*>(art::RoundDown(reinterpret_cast<uintptr_t>(end), art::kPageSize)); 61 if (end > start) { 62 size_t length = reinterpret_cast<uint8_t*>(end) - reinterpret_cast<uint8_t*>(start); 73 extern "C" void DlmallocBytesAllocatedCallback(void* start, void* end, size_t used_bytes, void* arg) { argument 81 extern "C" void DlmallocObjectsAllocatedCallback(void* start, void* end, size_t used_bytes, void* arg) { argument
|
H A D | rosalloc.h | 71 byte* end = fpr_base + ByteSize(rosalloc); variable 72 return end; 235 // Set the alloc_bit_map_ bits for slots that are past the end of the run. 244 void InspectAllSlots(void (*handler)(void* start, void* end, size_t used_bytes, void* callback_arg), void* arg); 394 kPageReleaseModeEnd, // Release empty pages at the end of the space. 397 // at the end of the space. 417 // the end of the memory region that's currently managed by this allocator. 421 // the end of the memory region that's ever managed by this allocator. 481 // The end address of the memory region that's managed by this allocator. 523 size_t ReleasePageRange(byte* start, byte* end) EXCLUSIVE_LOCKS_REQUIRE [all...] |
/art/runtime/gc/collector/ |
H A D | immune_region.h | 55 void SetEnd(mirror::Object* end) { argument 56 end_ = end;
|
/art/runtime/gc/space/ |
H A D | bump_pointer_space-inl.h | 44 byte* end = end_.LoadRelaxed(); local 45 if (end + num_bytes > growth_end_) { 48 mirror::Object* obj = reinterpret_cast<mirror::Object*>(end); 49 end_.StoreRelaxed(end + num_bytes);
|
H A D | valgrind_malloc_space-inl.h | 97 byte* end, byte* limit, size_t growth_limit, 100 S(name, mem_map, allocator, begin, end, limit, growth_limit, can_move_objects, starting_size, 95 ValgrindMallocSpace(const std::string& name, MemMap* mem_map, A allocator, byte* begin, byte* end, byte* limit, size_t growth_limit, size_t initial_size, bool can_move_objects, size_t starting_size) argument
|
/art/runtime/ |
H A D | leb128_test.cc | 116 uint8_t* end = EncodeUnsignedLeb128(encoded_data, uleb128_tests[i].decoded); local 117 size_t data_size = static_cast<size_t>(end - encoded_data); 155 uint8_t* end = encoded_data; local 157 end = EncodeUnsignedLeb128(end, uleb128_tests[i].decoded); 159 size_t data_size = static_cast<size_t>(end - encoded_data); 197 uint8_t* end = EncodeSignedLeb128(encoded_data, sleb128_tests[i].decoded); local 198 size_t data_size = static_cast<size_t>(end - encoded_data); 236 uint8_t* end = encoded_data; local 238 end [all...] |
H A D | vmap_table.h | 66 size_t end = DecodeUnsignedLeb128(&table); local 73 for (size_t i = 0; i < end; ++i) {
|
H A D | leb128.h | 26 // just past the end of the read value. This function tolerates 54 // just past the end of the read value. This function tolerates 62 // just past the end of the read value. This function tolerates 162 void InsertBackUnsigned(It cur, It end) { argument 163 for (; cur != end; ++cur) { 181 void InsertBackSigned(It cur, It end) { argument 182 for (; cur != end; ++cur) {
|
H A D | memory_region.h | 45 byte* end() const { return start() + size_; } function in class:art::MemoryRegion
|
H A D | thread-inl.h | 192 inline void Thread::SetThreadLocalAllocationStack(mirror::Object** start, mirror::Object** end) { argument 195 DCHECK(end != nullptr); 197 DCHECK_ALIGNED(end, sizeof(mirror::Object*)); 198 DCHECK_LT(start, end); 199 tlsPtr_.thread_local_alloc_stack_end = end;
|
/art/runtime/gc/accounting/ |
H A D | card_table_test.cc | 115 byte* end = cend - end_offset; local 118 card_table_->ModifyCardsAtomic(start, end, visitor, visitor); 123 for (byte* cur = end + kCardSize; cur < HeapLimit(); cur += kCardSize) { 127 for (byte* cur = start; cur < AlignUp(end, kCardSize); cur += kCardSize) { 131 LOG(ERROR) << reinterpret_cast<void*>(start) << " " << reinterpret_cast<void*>(cur) << " " << reinterpret_cast<void*>(end);
|
H A D | card_table.cc | 112 byte* end = mem_map_->End(); local 116 << " end: " << reinterpret_cast<void*>(end) 119 << " heap end: " << AddrFromCard(end)
|
H A D | card_table.h | 73 void VisitClear(const void* start, const void* end, const Visitor& visitor) { argument 75 byte* card_end = CardFromAddr(end); 102 // For every dirty at least minumum age between begin and end invoke the visitor with the 134 byte* end = mem_map_->End(); local 135 return card_addr >= begin && card_addr < end;
|
H A D | remembered_set.cc | 149 DCHECK(dirty_cards_.find(card_addr) != dirty_cards_.end()); 159 auto end = start + CardTable::kCardSize; local 160 os << reinterpret_cast<void*>(start) << "-" << reinterpret_cast<void*>(end) << "\n"; 169 auto end = start + CardTable::kCardSize; local 171 DCHECK_LE(end, space_->Limit());
|
H A D | space_bitmap_test.cc | 43 const mirror::Object* end) 46 end_(end) {} 77 // This handles all the cases, having runs which start and end on the same word, and different 83 mirror::Object* end = local 85 BitmapVerify(space_bitmap.get(), start, end); 143 size_t end = offset + RoundDown(r.next() % (remain + 1), kAlignment); variable 146 reinterpret_cast<uintptr_t>(heap_begin) + end, c); 149 for (uintptr_t k = offset; k < end; k += kAlignment) { 42 BitmapVerify(ContinuousSpaceBitmap* bitmap, const mirror::Object* begin, const mirror::Object* end) argument
|
H A D | mod_union_table.cc | 177 references_.find(ref) == references_.end()) { 247 uintptr_t end = start + CardTable::kCardSize; local 248 os << reinterpret_cast<void*>(start) << "-" << reinterpret_cast<void*>(end) << ","; 254 uintptr_t end = start + CardTable::kCardSize; local 255 os << reinterpret_cast<void*>(start) << "-" << reinterpret_cast<void*>(end) << "->{"; 274 uintptr_t end = start + CardTable::kCardSize; local 278 live_bitmap->VisitMarkedRange(start, end, add_visitor); 282 if (found == references_.end()) { 330 auto end = start + CardTable::kCardSize; local 331 os << reinterpret_cast<void*>(start) << "-" << reinterpret_cast<void*>(end) << "\ [all...] |
H A D | space_bitmap.cc | 80 // Not sure if doing this trim is necessary, since nothing past the end of the heap capacity 108 uintptr_t end = OffsetToIndex(HeapLimit() - heap_begin_ - 1); local 110 for (uintptr_t i = 0; i <= end; ++i) { 154 size_t end = OffsetToIndex(sweep_end - live_bitmap.heap_begin_ - 1); local 155 CHECK_LT(end, live_bitmap.Size() / kWordSize); 158 for (size_t i = start; i <= end; i++) { 252 uintptr_t end = Size() / kWordSize; local 253 for (uintptr_t i = 0; i < end; ++i) {
|
/art/compiler/utils/ |
H A D | array_ref.h | 116 iterator end() { return array_ + size_; } function in class:art::ArrayRef 117 const_iterator end() const { return array_ + size_; } function in class:art::ArrayRef 119 reverse_iterator rbegin() { return reverse_iterator(end()); } 120 const_reverse_iterator rbegin() const { return const_reverse_iterator(end()); }
|
/art/compiler/dex/ |
H A D | verified_method.cc | 72 return (it != devirt_map_.end()) ? &it->second : nullptr; 76 return std::binary_search(safe_cast_set_.begin(), safe_cast_set_.end(), pc); 194 // verification, we might end up sharpening to a wrong implementation. Just abort. 202 const Instruction* end = Instruction::At(insns + code_item->insns_size_in_code_units_); local 204 for (; inst < end; inst = inst->Next()) { 276 const Instruction* end = Instruction::At(code_item->insns_ + local 279 for (; inst < end; inst = inst->Next()) {
|