/art/runtime/base/ |
H A D | stl_util.h | 28 std::sort(v->begin(), v->end()); 29 v->erase(std::unique(v->begin(), v->end()), v->end()); 43 void STLDeleteContainerPointers(ForwardIterator begin, argument 45 while (begin != end) { 46 ForwardIterator temp = begin; 47 ++begin; 54 // hash_set, or any other STL container which defines sensible begin(), end(), 65 STLDeleteContainerPointers(container->begin(), container->end()); 75 for (typename T::iterator i = v->begin(); i != v->end(); ++i) {
|
H A D | stringpiece.h | 136 iterator begin() const { return ptr_; } function in class:art::StringPiece
|
/art/runtime/ |
H A D | dex_file_verifier.h | 27 static bool Verify(const DexFile* dex_file, const byte* begin, size_t size); 30 DexFileVerifier(const DexFile* dex_file, const byte* begin, size_t size) argument 31 : dex_file_(dex_file), begin_(begin), size_(size),
|
H A D | disassembler_mips.cc | 263 size_t DisassemblerMips::Dump(std::ostream& os, const uint8_t* begin) { argument 264 DumpMips(os, begin); 268 void DisassemblerMips::Dump(std::ostream& os, const uint8_t* begin, const uint8_t* end) { argument 269 for (const uint8_t* cur = begin; cur < end; cur += 4) {
|
H A D | mem_map.cc | 156 MemMap::MemMap(const std::string& name, byte* begin, size_t size, void* base_begin, argument 158 : name_(name), begin_(begin), size_(size), base_begin_(base_begin), base_size_(base_size),
|
H A D | safe_map.h | 46 iterator begin() { return map_.begin(); } function in class:art::SafeMap 47 const_iterator begin() const { return map_.begin(); } function in class:art::SafeMap
|
H A D | indirect_reference_table.h | 302 IrtIterator begin() { function in class:art::IndirectReferenceTable
|
H A D | disassembler_x86.cc | 30 size_t DisassemblerX86::Dump(std::ostream& os, const uint8_t* begin) { argument 31 return DumpInstruction(os, begin); 34 void DisassemblerX86::Dump(std::ostream& os, const uint8_t* begin, const uint8_t* end) { argument 36 for (const uint8_t* cur = begin; cur < end; cur += length) {
|
H A D | zip_archive.cc | 130 static bool CopyFdToMemory(uint8_t* begin, size_t size, int in, size_t count) { argument 131 uint8_t* dst = begin; 144 DCHECK_EQ(dst, begin + size); 174 static bool InflateToMemory(uint8_t* begin, size_t size, argument 176 uint8_t* dst = begin; 246 DCHECK_EQ(dst, begin + size); 267 bool ZipEntry::ExtractToMemory(uint8_t* begin, size_t size) { argument 286 return CopyFdToMemory(begin, size, zip_archive_->fd_, GetUncompressedLength()); 288 return InflateToMemory(begin, size, zip_archive_->fd_,
|
H A D | dex_file_verifier.cc | 103 bool DexFileVerifier::Verify(const DexFile* dex_file, const byte* begin, size_t size) { argument 104 UniquePtr<DexFileVerifier> verifier(new DexFileVerifier(dex_file, begin, size));
|
H A D | disassembler_arm.cc | 31 size_t DisassemblerArm::Dump(std::ostream& os, const uint8_t* begin) { argument 32 if ((reinterpret_cast<intptr_t>(begin) & 1) == 0) { 33 DumpArm(os, begin); 37 begin = reinterpret_cast<const uint8_t*>(reinterpret_cast<uintptr_t>(begin) & ~1); 38 return DumpThumb16(os, begin); 42 void DisassemblerArm::Dump(std::ostream& os, const uint8_t* begin, const uint8_t* end) { argument 43 if ((reinterpret_cast<intptr_t>(begin) & 1) == 0) { 44 for (const uint8_t* cur = begin; cur < end; cur += 4) { 49 begin [all...] |
H A D | runtime.cc | 313 const char* begin = &s[colon + 1]; local 315 size_t result = strtoul(begin, &end, 10); 316 if (begin == end || *end != '\0') {
|
H A D | thread.cc | 96 uintptr_t* begin = reinterpret_cast<uintptr_t*>(&interpreter_entrypoints_); local 97 uintptr_t* end = reinterpret_cast<uintptr_t*>(reinterpret_cast<uint8_t*>(begin) + sizeof(quick_entrypoints_)); 98 for (uintptr_t* it = begin; it != end; ++it) { 101 begin = reinterpret_cast<uintptr_t*>(&interpreter_entrypoints_); 102 end = reinterpret_cast<uintptr_t*>(reinterpret_cast<uint8_t*>(begin) + sizeof(portable_entrypoints_)); 103 for (uintptr_t* it = begin; it != end; ++it) { 2184 for (It it = instrumentation_stack->begin(), end = instrumentation_stack->end(); it != end; ++it) {
|
H A D | debugger.cc | 1418 const uint8_t* begin = reinterpret_cast<const uint8_t*>(code_item->insns_); local 1419 const uint8_t* end = begin + byte_count; 1420 for (const uint8_t* p = begin; p != end; ++p) { 2386 gBreakpoints.erase(gBreakpoints.begin() + i); 2579 for (std::set<uint32_t>::iterator it = gSingleStepControl.dex_pcs.begin() ; it != gSingleStepControl.dex_pcs.end(); ++it) { 3440 for (It cur = spaces.begin(), end = spaces.end(); cur != end; ++cur) { 3619 return std::distance(table_.begin(), it);
|
/art/runtime/gc/accounting/ |
H A D | space_bitmap_test.cc | 44 BitmapVerify(SpaceBitmap* bitmap, const mirror::Object* begin, const mirror::Object* end) argument 46 begin_(begin),
|
H A D | card_table.cc | 84 byte* __attribute__((unused)) begin = mem_map_->Begin() + offset_; local 106 byte* begin = mem_map_->Begin() + offset_; local 110 << " begin: " << reinterpret_cast<void*>(begin) 113 << " heap begin: " << AddrFromCard(begin)
|
H A D | card_table.h | 102 // For every dirty at least minumum age between begin and end invoke the visitor with the 128 CardTable(MemMap* begin, byte* biased_begin, size_t offset); 132 byte* begin = mem_map_->Begin() + offset_; local 134 return card_addr >= begin && card_addr < end; 146 // Card table doesn't begin at the beginning of the mem_map_, instead it is displaced by offset
|
/art/compiler/utils/ |
H A D | dedupe_set.h | 49 iterator begin() { return keys_.begin(); } function in class:art::DedupeSet 50 const_iterator begin() const { return keys_.begin(); } function in class:art::DedupeSet
|
/art/runtime/gc/space/ |
H A D | large_object_space.cc | 111 for (MemMaps::iterator it = mem_maps_.begin(); it != mem_maps_.end(); ++it) { 137 FreeListSpace::FreeListSpace(const std::string& name, MemMap* mem_map, byte* begin, byte* end) argument 139 begin_(begin), 143 free_end_ = end - begin; 304 << " begin: " << reinterpret_cast<void*>(Begin())
|
H A D | space.h | 216 byte* begin, byte* end) : 217 Space(name, gc_retention_policy), begin_(begin), end_(end) { 215 ContinuousSpace(const std::string& name, GcRetentionPolicy gc_retention_policy, byte* begin, byte* end) argument
|
H A D | dlmalloc_space.cc | 104 ValgrindDlMallocSpace(const std::string& name, MemMap* mem_map, void* mspace, byte* begin, argument 106 DlMallocSpace(name, mem_map, mspace, begin, end, growth_limit) { 119 DlMallocSpace::DlMallocSpace(const std::string& name, MemMap* mem_map, void* mspace, byte* begin, argument 121 : MemMapSpace(name, mem_map, end - begin, kGcRetentionPolicyAlwaysCollect), 223 void* DlMallocSpace::CreateMallocSpace(void* begin, size_t morecore_start, size_t initial_size) { argument 226 // create mspace using our backing storage starting at begin and with a footprint of 229 void* msp = create_mspace_with_base(begin, morecore_start, false /*locked*/); 429 // Should never be asked for negative footprint (ie before begin) 498 << " begin=" << reinterpret_cast<void*>(Begin())
|
/art/runtime/gc/collector/ |
H A D | mark_sweep.cc | 349 void MarkSweep::SetImmuneRange(Object* begin, Object* end) { argument 350 immune_begin_ = begin; 745 byte* begin, byte* end, byte minimum_age, size_t mark_stack_size, 749 begin_(begin), 885 uintptr_t begin = reinterpret_cast<uintptr_t>(image_space->Begin()); local 889 live_bitmap->VisitMarkedRange(begin, end, [this](const Object* obj) { 904 accounting::SpaceBitmap* bitmap, uintptr_t begin, uintptr_t end) 907 begin_(begin), 961 uintptr_t begin = reinterpret_cast<uintptr_t>(space->Begin()); local 967 while (begin ! 903 RecursiveMarkTask(ThreadPool* thread_pool, MarkSweep* mark_sweep, accounting::SpaceBitmap* bitmap, uintptr_t begin, uintptr_t end) argument 983 uintptr_t begin = reinterpret_cast<uintptr_t>(space->Begin()); local 1237 uintptr_t begin = reinterpret_cast<uintptr_t>(space->Begin()); local [all...] |
/art/compiler/driver/ |
H A D | compiler_driver.cc | 684 for (auto it = image_classes_->begin(), end = image_classes_->end(); it != end;) { 1371 void ForAll(size_t begin, size_t end, Callback callback, size_t work_units) { argument 1377 index_ = begin;
|