/art/runtime/gc/accounting/ |
H A D | space_bitmap_test.cc | 43 const mirror::Object* end) 46 end_(end) {} 77 // This handles all the cases, having runs which start and end on the same word, and different 83 mirror::Object* end = local 85 BitmapVerify(space_bitmap.get(), start, end); 143 size_t end = offset + RoundDown(r.next() % (remain + 1), kAlignment); variable 146 reinterpret_cast<uintptr_t>(heap_begin) + end, c); 149 for (uintptr_t k = offset; k < end; k += kAlignment) { 42 BitmapVerify(ContinuousSpaceBitmap* bitmap, const mirror::Object* begin, const mirror::Object* end) argument
|
H A D | card_table-inl.h | 112 // Handle any unaligned cards at the end. 158 // Handle unaligned cards at the end. 214 << " end: " << reinterpret_cast<void*>(mem_map_->End()); 229 uint8_t* end = mem_map_->End(); local 230 return card_addr >= begin && card_addr < end; 237 << " end: " << reinterpret_cast<void*>(mem_map_->End());
|
H A D | space_bitmap.cc | 87 // Not sure if doing this trim is necessary, since nothing past the end of the heap capacity 115 uintptr_t end = OffsetToIndex(HeapLimit() - heap_begin_ - 1); local 117 for (uintptr_t i = 0; i <= end; ++i) { 161 size_t end = OffsetToIndex(sweep_end - live_bitmap.heap_begin_ - 1); local 162 CHECK_LT(end, live_bitmap.Size() / sizeof(intptr_t)); 165 for (size_t i = start; i <= end; i++) { 251 uintptr_t end = Size() / sizeof(intptr_t); local 252 for (uintptr_t i = 0; i < end; ++i) {
|
H A D | mod_union_table.cc | 270 references_.find(ref) == references_.end()) { 351 uintptr_t end = start + CardTable::kCardSize; local 352 os << reinterpret_cast<void*>(start) << "-" << reinterpret_cast<void*>(end) << ","; 358 uintptr_t end = start + CardTable::kCardSize; local 359 os << reinterpret_cast<void*>(start) << "-" << reinterpret_cast<void*>(end) << "->{"; 382 uintptr_t end = start + CardTable::kCardSize; local 387 live_bitmap->VisitMarkedRange(start, end, add_visitor); 390 if (found == references_.end()) { 410 for (auto it = references_.begin(); it != references_.end();) { 519 // Only clean up to the end sinc [all...] |
/art/runtime/ |
H A D | leb128.h | 29 // just past the end of the read value. This function tolerates 57 // just past the end of the read value. This function tolerates 65 // just past the end of the read value. This function tolerates 149 for (uint8_t* end = EncodeUnsignedLeb128(dest, value); end < old_end; end++) { 151 end[-1] |= 0x80; 152 end[0] = 0; 202 void InsertBackUnsigned(It cur, It end) { argument 203 for (; cur != end; 213 InsertBackSigned(It cur, It end) argument [all...] |
H A D | thread-inl.h | 260 StackReference<mirror::Object>* end) { 263 DCHECK(end != nullptr); 265 DCHECK_ALIGNED(end, sizeof(StackReference<mirror::Object>)); 266 DCHECK_LT(start, end); 267 tlsPtr_.thread_local_alloc_stack_end = end; 259 SetThreadLocalAllocationStack(StackReference<mirror::Object>* start, StackReference<mirror::Object>* end) argument
|
H A D | indirect_reference_table.h | 333 IrtIterator end() { function in class:art::IndirectReferenceTable 355 // Release pages past the end of the table that may have previously held references.
|
H A D | memory_region.h | 50 uint8_t* end() const { return start() + size_; } function in class:art::FINAL
|
H A D | safe_map.h | 64 iterator end() { return map_.end(); } function in class:art::SafeMap 65 const_iterator end() const { return map_.end(); } function in class:art::SafeMap 86 DCHECK(it != map_.end()); 105 DCHECK(pos == map_.end() || map_.key_comp()(k, pos->first)); 111 DCHECK(pos == map_.end() || map_.key_comp()(k, pos->first)); 133 if (lb != end() && !key_comp()(k, lb->first)) {
|
H A D | utf.cc | 49 const char* end = utf8 + byte_count; local 50 for (; utf8 < end; ++utf8) { 244 const uint16_t *end = chars + char_count; local 245 while (chars < end) { 256 if (chars < end) {
|
H A D | utils.h | 54 char* end; local 55 unsigned long long int result = strtoull(in, &end, 0); // NOLINT(runtime/int) 56 if (in == end || *end != '\0') { 68 char* end; local 69 long long int result = strtoll(in, &end, 0); // NOLINT(runtime/int) 70 if (in == end || *end != '\0') { 203 // additionally allowing names that begin with '<' and end with '>'. 401 inline void FlushInstructionCache(char* begin, char* end) { argument [all...] |
/art/runtime/native/ |
H A D | java_lang_String.cc | 83 static void String_getCharsNoCheck(JNIEnv* env, jobject java_this, jint start, jint end, argument 88 soa.Decode<mirror::String*>(java_this)->GetChars(start, end, char_array, index);
|
H A D | libcore_util_CharsetUtils.cc | 205 const int end = offset + length; local 206 for (int i = offset; i < end; ++i) { 221 jchar low = (i + 1 != end) ? string->CharAt(i + 1) : 0;
|
/art/compiler/utils/ |
H A D | array_ref.h | 119 iterator end() { return array_ + size_; } function in class:art::ArrayRef 120 const_iterator end() const { return array_ + size_; } function in class:art::ArrayRef 122 reverse_iterator rbegin() { return reverse_iterator(end()); } 123 const_reverse_iterator rbegin() const { return const_reverse_iterator(end()); } 184 return lhs.size() == rhs.size() && std::equal(lhs.begin(), lhs.end(), rhs.begin());
|
H A D | intrusive_forward_list.h | 157 iterator end() { return iterator(nullptr); } function in class:art::IntrusiveForwardList 158 const_iterator end() const { return const_iterator(nullptr); } function in class:art::IntrusiveForwardList 164 bool empty() const { return begin() == end(); } 216 DCHECK(position != end()); 217 splice_after(position, src, src.before_begin(), src.end()); 244 DCHECK(position != end()); 250 // If position is just before end() and last is src.end(), we can finish this quickly. 251 if (++const_iterator(position) == end() && last == src.end()) { [all...] |
/art/runtime/base/ |
H A D | bit_vector.cc | 240 uint32_t BitVector::NumSetBits(uint32_t end) const { 241 DCHECK_LE(end, storage_size_ * kWordBits); 242 return NumSetBits(storage_, end); 315 uint32_t BitVector::NumSetBits(const uint32_t* storage, uint32_t end) { argument 316 uint32_t word_end = WordIndex(end); 317 uint32_t partial_word_bits = end & 0x1f;
|
H A D | bit_vector.h | 44 * object with begin() and end() suitable for range-based loops: 64 // Helper function to check for end without comparing with bit_vector.Indexes().end(). 106 IndexIterator end() const { function in class:art::BitVector::IndexContainer 210 // Count the number of bits that are set in range [0, end). 211 uint32_t NumSetBits(uint32_t end) const; 256 // Number of bits set in range [0, end) in storage. (No range check.) 257 static uint32_t NumSetBits(const uint32_t* storage, uint32_t end);
|
H A D | hash_set_test.cc | 75 ASSERT_TRUE(after_it == hash_set.end()); 79 ASSERT_TRUE(it == hash_set.end()); 91 ASSERT_TRUE(it != hash_set.end()); 98 ASSERT_TRUE(it != hash_set.end()); 105 ASSERT_TRUE(it == hash_set.end()); 109 ASSERT_TRUE(it != hash_set.end()); 116 ASSERT_TRUE(hash_set.begin() == hash_set.end()); 134 for (auto it = hash_set.begin(); it != hash_set.end();) { 194 EXPECT_NE(hash_set.end(), hash_set.Find(initial_string)) 250 ASSERT_EQ(it1 == hash_set.end(), it 291 HashIntSequence(T begin, T end) argument [all...] |
H A D | stringpiece.h | 127 // Does "this" end with "x" 134 iterator end() const { return ptr_ + length_; } function in class:art::StringPiece
|
/art/runtime/base/unix_file/ |
H A D | fd_file.cc | 218 // 0: end of file 280 off_t end = off + sz; local 281 while (off != end) { 283 sendfile(Fd(), input_file->Fd(), &off, end - off));
|
/art/runtime/gc/ |
H A D | allocation_record.cc | 46 char* end; local 47 size_t value = strtoul(allocMaxString, &end, 10); 48 if (*end != '\0') { 62 char* end; local 63 size_t value = strtoul(recentAllocMaxString, &end, 10); 64 if (*end != '\0') { 78 char* end; local 79 size_t value = strtoul(stackDepthString, &end, 10); 80 if (*end != '\0') { 104 for (auto it = entries_.rbegin(), end [all...] |
/art/runtime/gc/collector/ |
H A D | immune_spaces_test.cc | 33 DummyOatFile(uint8_t* begin, uint8_t* end) : OatFile("Location", /*is_executable*/ false) { argument 35 end_ = end; 164 DummySpace(uint8_t* begin, uint8_t* end) argument 168 end, 169 /*limit*/end) {}
|
/art/runtime/gc/space/ |
H A D | bump_pointer_space.cc | 72 // Reset the end of the space back to the beginning, we move the end forward as we allocate 158 uint8_t* end = End(); local 172 end = main_end; 190 while (pos < end) { 198 // assume its the end. TODO: Have a thread update the header when it flushes the block?
|
H A D | malloc_space.cc | 39 uint8_t* begin, uint8_t* end, uint8_t* limit, size_t growth_limit, 42 : ContinuousMemMapAllocSpace(name, mem_map, begin, end, limit, kGcRetentionPolicyAlwaysCollect), 198 uint8_t* end = mem_map->Begin() + starting_size_; local 200 CHECK_MEMORY_CALL(mprotect, (end, capacity - initial_size_, PROT_NONE), alloc_space_name); 202 *out_malloc_space = CreateInstance(mem_map.release(), alloc_space_name, allocator, End(), end, 224 << ",end=" << reinterpret_cast<void*>(End()) 38 MallocSpace(const std::string& name, MemMap* mem_map, uint8_t* begin, uint8_t* end, uint8_t* limit, size_t growth_limit, bool create_bitmaps, bool can_move_objects, size_t starting_size, size_t initial_size) argument
|
/art/compiler/optimizing/ |
H A D | ssa_liveness_analysis.cc | 49 for (auto end = worklist->rend(); insert_pos != end; ++insert_pos) { 109 // start and end position. Non-phi instructions have a distinct lifetime position than 114 // to differentiate between the start and end of an instruction. Adding 2 to 116 // instruction is different than the end of the previous instruction. 237 // A phi input whose last user is the phi dies at the end of the predecessor block, 414 size_t end = GetEnd(); local 415 while (use != nullptr && use->GetPosition() <= end) { 432 if (end == user->GetBlock()->GetPredecessors()[input_index]->GetLifetimeEnd()) { 476 size_t end local [all...] |