/art/runtime/ |
H A D | monitor_pool.cc | 71 Monitor* last = reinterpret_cast<Monitor*>(reinterpret_cast<uintptr_t>(chunk) + local 73 last->next_free_ = nullptr; 75 last->monitor_id_ = OffsetToMonitorId(current_chunk_list_index_* (kMaxListSize * kChunkSize) 78 Monitor* before = reinterpret_cast<Monitor*>(reinterpret_cast<uintptr_t>(last) - 80 before->next_free_ = last; 81 // Derive monitor_id from last. 82 before->monitor_id_ = OffsetToMonitorId(MonitorIdToOffset(last->monitor_id_) - 85 last = before; 87 DCHECK(last == reinterpret_cast<Monitor*>(chunk)); 88 first_free_ = last; [all...] |
H A D | mem_map_test.cc | 152 uintptr_t last = 0; local 155 EXPECT_NE(last, random_start); 156 last = random_start; 307 // Create source from the last 2 pages
|
/art/compiler/utils/ |
H A D | intrusive_forward_list.h | 149 IntrusiveForwardList(InputIterator first, InputIterator last) : IntrusiveForwardList() { argument 150 insert_after(before_begin(), first, last); local 184 void assign(InputIterator first, InputIterator last) { argument 185 IntrusiveForwardList tmp(first, last); 202 iterator insert_after(const_iterator position, InputIterator first, InputIterator last) { argument 203 while (first != last) { 209 const_iterator last = position; local 210 std::advance(last, 2); 211 return erase_after(position, last); 213 iterator erase_after(const_iterator position, const_iterator last) { argument 242 const_iterator last = i; local 251 splice_after(const_iterator position, IntrusiveForwardList& src, const_iterator first, const_iterator last) argument 280 splice_after(const_iterator position, IntrusiveForwardList&& src, const_iterator first, const_iterator last) argument [all...] |
H A D | assembler.h | 314 // Override the last delayed PC. The new PC can be out of order. 345 void AppendRawData(const std::vector<uint8_t>& raw_data, size_t first, size_t last) { argument 347 DCHECK_LE(first, last); 348 DCHECK_LE(last, raw_data.size()); 349 opcodes_.insert(opcodes_.end(), raw_data.begin() + first, raw_data.begin() + last);
|
/art/libartbase/base/ |
H A D | iteration_range.h | 25 // range [first, last) defined by two iterators. 35 IterationRange(iterator first, iterator last) : first_(first), last_(last) { } argument
|
H A D | dchecked_vector.h | 70 InputIterator last, 72 : Base(first, last, alloc) { } 159 iterator insert(const_iterator position, InputIterator first, InputIterator last) { argument 161 return Base::insert(position, first, last); 177 iterator erase(const_iterator first, const_iterator last) { 179 DCHECK(first <= last && last <= cend()); 180 return Base::erase(first, last); 69 dchecked_vector(InputIterator first, InputIterator last, const allocator_type& alloc = allocator_type()) argument
|
H A D | stringpiece.h | 179 const char* last = ptr_ + std::min(ulen - s.length_, pos) + s.length_; local 180 const char* result = std::find_end(ptr_, last, s.ptr_, s.ptr_ + s.length_); 181 return result != last ? result - ptr_ : npos; 239 // Test last byte in case strings share large common prefix 243 // At this point we can, but don't have to, ignore the last byte. We use
|
/art/compiler/optimizing/ |
H A D | constant_folding_test.cc | 761 HInstruction* last; local 762 block->AddInstruction(last = new (GetAllocator()) HAbove(zero, parameter)); 763 block->AddInstruction(new (GetAllocator()) HSelect(last, parameter, parameter, 0)); 764 block->AddInstruction(last = new (GetAllocator()) HAbove(parameter, zero)); 765 block->AddInstruction(new (GetAllocator()) HSelect(last, parameter, parameter, 0)); 766 block->AddInstruction(last = new (GetAllocator()) HAboveOrEqual(zero, parameter)); 767 block->AddInstruction(new (GetAllocator()) HSelect(last, parameter, parameter, 0)); 768 block->AddInstruction(last = new (GetAllocator()) HAboveOrEqual(parameter, zero)); 769 block->AddInstruction(new (GetAllocator()) HSelect(last, parameter, parameter, 0)); 770 block->AddInstruction(last [all...] |
H A D | dead_code_elimination.cc | 150 // Test for an if as last statement. 244 HInstruction* last = block->GetLastInstruction(); local 247 first->GetNext() == last && 248 last->IsGoto() && 326 HInstruction* last = block->GetLastInstruction(); local 329 last->IsIf() && 332 bool has_only_phi_and_if = (last == first) && (last->InputAt(0) == block->GetFirstPhi()); 337 (first->GetNext() == last) && 338 (last [all...] |
H A D | induction_var_range_test.cc | 784 HInstruction* last = range_.GenerateLastValue(phi, graph_, loop_preheader_); local 785 ASSERT_TRUE(last->IsAdd()); 786 ExpectInt(1000, last->InputAt(0)); 787 ExpectInt(0, last->InputAt(1)); 836 HInstruction* last = range_.GenerateLastValue(phi, graph_, loop_preheader_); local 837 ASSERT_TRUE(last->IsSub()); 838 ExpectInt(1000, last->InputAt(0)); 839 ASSERT_TRUE(last->InputAt(1)->IsNeg()); 840 last = last [all...] |
H A D | code_sinking.cc | 37 HInstruction* last = exit_predecessor->GetLastInstruction(); local 39 if (!last->IsReturn() && !last->IsReturnVoid()) {
|
H A D | register_allocation_resolver.cc | 636 HInstruction* last = block->GetLastInstruction(); local 641 DCHECK(!last->IsIf() && !last->IsPackedSwitch()); 642 HInstruction* previous = last->GetPrevious(); 646 size_t position = last->GetLifetimePosition(); 651 block->InsertInstructionBefore(move, last);
|
H A D | ssa_liveness_analysis.h | 199 inline Iterator FindUseAtOrAfterPosition(Iterator first, Iterator last, size_t position) { argument 205 first, last, [position](const value_type& use) { return use.GetPosition() >= position; }); 219 Iterator last, 222 Iterator begin = FindUseAtOrAfterPosition(first, last, position_begin); 223 Iterator end = FindUseAtOrAfterPosition(begin, last, position_end); 1046 // i.e. loop blocks may not be adjacent and a back edge may not be last, 1055 // we need to add subsequent entries after the last inserted entry. 1067 // We're only adding a synthesized use at the last back edge. Adding synthesized uses on 1069 // last back edge. If we want branches in a loop to have better register allocation than 1091 // Ranges of this interval. We need a quick access to the last rang 218 FindMatchingUseRange(Iterator first, Iterator last, size_t position_begin, size_t position_end) argument [all...] |
H A D | nodes.cc | 2079 // a loop header is removed last, a requirement for correct loop information 2204 // If this was the last back edge of the loop, we deliberately leave the 2425 HInstruction* last = body->GetLastInstruction(); local 2432 if (last->IsReturn()) { 2433 return_value = last->InputAt(0); 2435 DCHECK(last->IsReturnVoid()); 2438 invoke->GetBlock()->RemoveInstruction(last); 2507 HInstruction* last = predecessor->GetLastInstruction(); local 2508 if (last->IsThrow()) { 2522 if (last [all...] |
H A D | induction_var_analysis.h | 160 InductionInfo* RotatePeriodicInduction(InductionInfo* induction, InductionInfo* last);
|
H A D | code_generator_arm_vixl.cc | 172 // from a list of contiguous S registers a list of contiguous D registers (processing first/last 189 size_t last, 194 DCHECK_LE(first, last); 195 if ((first == last) && (first == 0)) { 205 if (last % 2 == 0) { 207 --last; 210 if (first < last) { 212 DCHECK_EQ((last - first + 1) % 2, 0u); 213 size_t number_of_d_regs = (last - first + 1) / 2; 230 __ Vstr(vixl32::SRegister(last 188 SaveContiguousSRegisterList(size_t first, size_t last, CodeGenerator* codegen, size_t stack_offset) argument 237 RestoreContiguousSRegisterList(size_t first, size_t last, CodeGenerator* codegen, size_t stack_offset) argument [all...] |
H A D | induction_var_analysis.cc | 522 InductionInfo* last) { 532 last, 539 RotatePeriodicInduction(induction->op_b, last), 865 // A narrowing conversion is allowed as *last* operation of the cycle of a linear induction 520 RotatePeriodicInduction( InductionInfo* induction, InductionInfo* last) argument
|
/art/libdexfile/dex/ |
H A D | primitive_test.cc | 28 int last = static_cast<int>(Primitive::Type::kPrimLast); local 29 for (int i = 0; i <= last; ++i) {
|
/art/runtime/base/ |
H A D | arena_allocator.cc | 369 Arena* last = first; local 370 while (last->next_ != nullptr) { 371 last = last->next_; 375 last->next_ = free_arenas_;
|
/art/runtime/jdwp/ |
H A D | jdwp_main.cc | 721 * Return the time, in milliseconds, since the last debugger activity. 732 int64_t last = last_activity_time_ms_.LoadSequentiallyConsistent(); local 735 if (last == 0) { 736 VLOG(jdwp) << "+++ last=busy"; 742 CHECK_GE(now, last); 744 VLOG(jdwp) << "+++ debugger interval=" << (now - last); 745 return now - last;
|
/art/tools/dexfuzz/src/dexfuzz/program/mutators/ |
H A D | ArithOpChanger.java | 163 Opcode last = opcodeList.get(opcodeList.size() - 1); 164 if (Opcode.isBetween(opcode, first, last)) {
|
/art/runtime/mirror/ |
H A D | var_handle_test.cc | 215 auto last = it + 1; local 216 while (*last != ';') { 217 ++last; 219 descriptors.push_back(prefix + std::string(it, last + 1)); 221 it = last;
|
H A D | var_handle.cc | 1634 auto last = std::cend(kAccessorToAccessMode); local 1636 last, 1639 if (it == last || strcmp(it->method_name, method_name) != 0) {
|
/art/compiler/linker/ |
H A D | elf_builder.h | 75 // The debug sections are written last for easier stripping. 211 Elf_Word last = sections.empty() ? PF_R : sections.back()->phdr_flags_; local 212 if (phdr_flags_ != last) { 390 // The sh_info file must be set to index one-past the last local symbol. 522 // very last thing in the output file creation.
|
/art/oatdump/ |
H A D | oatdump.cc | 921 // If the last thing in the file is code for a method, there won't be an offset for the "next" 1728 StackMap last = code_info_.GetStackMapAt(0u, encoding_); local 1731 if (last.GetNativePcOffset(encoding_.stack_map.encoding, instruction_set) > 1736 last = current;
|