/art/runtime/base/ |
H A D | variant_map.h | 232 auto* ptr = Get(key); local 233 return (ptr == nullptr) ? key.CreateDefaultValue() : *ptr; 256 TValue* ptr = Get(key); local 257 if (ptr != nullptr) { 258 return std::move(*ptr); 285 TValue* ptr = Get(key); local 286 if (ptr == nullptr) {
|
H A D | hash_set.h | 151 HashSet(const uint8_t* ptr, bool make_copy_of_data, size_t* read_count) { argument 154 offset = ReadFromBytes(ptr, offset, &temp); 156 offset = ReadFromBytes(ptr, offset, &temp); 159 offset = ReadFromBytes(ptr, offset, &temp); 161 offset = ReadFromBytes(ptr, offset, &min_load_factor_); 162 offset = ReadFromBytes(ptr, offset, &max_load_factor_); 165 data_ = const_cast<T*>(reinterpret_cast<const T*>(ptr + offset)); 172 offset = ReadFromBytes(ptr, offset, &data_[i]); 181 size_t WriteToMemory(uint8_t* ptr) { argument 183 offset = WriteToBytes(ptr, offse 507 WriteToBytes(uint8_t* ptr, size_t offset, Elem n) argument 516 ReadFromBytes(const uint8_t* ptr, size_t offset, Elem* out) argument [all...] |
/art/runtime/gc/space/ |
H A D | dlmalloc_space.cc | 165 size_t DlMallocSpace::Free(Thread* self, mirror::Object* ptr) { argument 168 CHECK(ptr != nullptr); 169 CHECK(Contains(ptr)) << "Free (" << ptr << ") not in bounds of heap " << *this; 171 const size_t bytes_freed = AllocationSizeNonvirtual(ptr, nullptr); 173 RegisterRecentFree(ptr); 175 mspace_free(mspace_, ptr); 185 mirror::Object* ptr = ptrs[i]; local 191 bytes_freed += AllocationSizeNonvirtual(ptr, nullptr);
|
H A D | rosalloc_space.cc | 194 size_t RosAllocSpace::Free(Thread* self, mirror::Object* ptr) { argument 196 CHECK(ptr != nullptr); 197 CHECK(Contains(ptr)) << "Free (" << ptr << ") not in bounds of heap " << *this; 201 RegisterRecentFree(ptr); 203 return rosalloc_->Free(self, ptr);
|
H A D | large_object_space.cc | 186 size_t LargeObjectMapSpace::Free(Thread* self, mirror::Object* ptr) { argument 188 auto it = large_objects_.find(ptr); 191 LOG(FATAL) << "Attempted to free large object " << ptr << " which was not live";
|
/art/runtime/ |
H A D | intern_table.cc | 300 const uint8_t* ptr = image_space->Begin() + intern_section.Offset(); local 301 const size_t offset = ReadFromMemory(ptr); 305 size_t InternTable::ReadFromMemory(const uint8_t* ptr) { argument 307 return ReadFromMemoryLocked(ptr); 310 size_t InternTable::ReadFromMemoryLocked(const uint8_t* ptr) { argument 311 return strong_interns_.ReadIntoPreZygoteTable(ptr); 314 size_t InternTable::WriteToMemory(uint8_t* ptr) { argument 316 return strong_interns_.WriteFromPostZygoteTable(ptr); 334 size_t InternTable::Table::ReadIntoPreZygoteTable(const uint8_t* ptr) { argument 337 pre_zygote_table_ = UnorderedSet(ptr, fals 341 WriteFromPostZygoteTable(uint8_t* ptr) argument [all...] |
H A D | mem_map.cc | 141 static bool ContainedWithinExistingMap(uint8_t* ptr, size_t size, argument 143 uintptr_t begin = reinterpret_cast<uintptr_t>(ptr); 244 static inline void* TryMemMapLow4GB(void* ptr, size_t page_aligned_byte_count, int prot, int flags, argument 246 void* actual = mmap(ptr, page_aligned_byte_count, prot, flags, fd, 0); 334 for (uintptr_t ptr = next_mem_pos_; ptr < 4 * GB; ptr += kPageSize) { 336 // Find the first map which is address > ptr. 337 auto it = maps_->upper_bound(reinterpret_cast<void*>(ptr)); 342 ptr [all...] |
H A D | oat.cc | 388 const char* ptr = reinterpret_cast<const char*>(&key_value_store_); local 389 const char* end = ptr + key_value_store_size_; 391 while (ptr < end) { 393 const char* str_end = ParseString(ptr, end); 395 if (strcmp(key, ptr) == 0) { 402 ptr = ParseString(str_end + 1, end) + 1; 414 const char* ptr = reinterpret_cast<const char*>(&key_value_store_); local 415 const char* end = ptr + key_value_store_size_; 418 while (ptr < end && counter >= 0) { 420 const char* str_end = ParseString(ptr, en [all...] |
H A D | trace.cc | 625 uint8_t* ptr = buf + kTraceHeaderLength; local 628 while (ptr < end) { 629 uint32_t tmid = ReadBytes(ptr + 2, sizeof(tmid)); 633 ptr += GetRecordSize(clock_source); 925 uint8_t* ptr; local 929 ptr = stack_buf; 931 ptr = buf_.get() + old_offset; 934 Append2LE(ptr, thread->GetTid()); 935 Append4LE(ptr + 2, method_value); 936 ptr 977 uint8_t* ptr = buf_.get() + kTraceHeaderLength; local [all...] |
H A D | dex_file.cc | 1131 static int32_t ReadSignedInt(const uint8_t* ptr, int zwidth) { argument 1134 val = ((uint32_t)val >> 8) | (((int32_t)*ptr++) << 24); 1142 static uint32_t ReadUnsignedInt(const uint8_t* ptr, int zwidth, bool fill_on_right) { argument 1146 val = (val >> 8) | (((uint32_t)*ptr++) << 24); 1151 val = (val >> 8) | (((uint32_t)*ptr++) << 24); 1158 static int64_t ReadSignedLong(const uint8_t* ptr, int zwidth) { argument 1161 val = ((uint64_t)val >> 8) | (((int64_t)*ptr++) << 56); 1169 static uint64_t ReadUnsignedLong(const uint8_t* ptr, int zwidth, bool fill_on_right) { argument 1173 val = (val >> 8) | (((uint64_t)*ptr++) << 56); 1178 val = (val >> 8) | (((uint64_t)*ptr [all...] |
H A D | dex_file_verifier.cc | 204 bool DexFileVerifier::CheckList(size_t element_size, const char* label, const uint8_t* *ptr) { argument 206 if (!CheckListSize(*ptr, 1, 4U, label)) { 210 uint32_t count = *reinterpret_cast<const uint32_t*>(*ptr); 212 if (!CheckListSize(*ptr + 4, count, element_size, label)) { 217 *ptr += 4 + count * element_size; 1446 uint16_t DexFileVerifier::FindFirstClassDataDefiner(const uint8_t* ptr, bool* success) { argument 1447 ClassDataItemIterator it(*dex_file_, ptr); 1465 uint16_t DexFileVerifier::FindFirstAnnotationsDirectoryDefiner(const uint8_t* ptr, bool* success) { argument 1467 reinterpret_cast<const DexFile::AnnotationsDirectoryItem*>(ptr);
|
H A D | check_jni.cc | 188 bool CheckNonNull(const void* ptr) { argument 189 if (UNLIKELY(ptr == nullptr)) { 2311 void* ptr = baseEnv(env)->GetPrimitiveArrayCritical(env, array, is_copy); local 2312 if (ptr != nullptr && soa.ForceCopy()) { 2313 ptr = GuardedCopy::CreateGuardedPACopy(env, array, is_copy, ptr); 2315 result.p = ptr; 3067 void* ptr; local 3070 ptr = const_cast<char*>(baseEnv(env)->GetStringUTFChars(env, string, is_copy)); 3071 result.u = reinterpret_cast<char*>(ptr); 3180 void* ptr = nullptr; local [all...] |
H A D | class_linker.cc | 2291 auto* ptr = reinterpret_cast<ArtField*>(la->AllocArray<ArtField>(self, length)); local 2292 CHECK(ptr!= nullptr); 2293 std::uninitialized_fill_n(ptr, length, ArtField()); 2294 return ptr; 2299 uintptr_t ptr = reinterpret_cast<uintptr_t>( local 2301 CHECK_NE(ptr, 0u); 2303 new(reinterpret_cast<void*>(ptr + i * method_size)) ArtMethod; 2305 return reinterpret_cast<ArtMethod*>(ptr);
|
H A D | debugger.cc | 2594 uint8_t* ptr = expandBufAddSpace(pReply, width + 1); local 2595 error = Dbg::GetLocalValue(visitor, soa, slot, reqSigByte, ptr, width); 4642 void AppendChunk(uint8_t state, void* ptr, size_t length, bool is_native) 4663 EnsureHeader(ptr);
|
/art/cmdline/ |
H A D | cmdline_parser.h | 75 auto* ptr = variant_map_->Get(key); local 76 if (ptr == nullptr) { 78 ptr = variant_map_->Get(key); 79 assert(ptr != nullptr); 82 return *ptr;
|
/art/disassembler/ |
H A D | disassembler_arm.cc | 102 static uint32_t ReadU16(const uint8_t* ptr) { argument 103 return ptr[0] | (ptr[1] << 8); 106 static uint32_t ReadU32(const uint8_t* ptr) { argument 107 return ptr[0] | (ptr[1] << 8) | (ptr[2] << 16) | (ptr[3] << 24);
|
/art/runtime/gc/allocator/ |
H A D | rosalloc.cc | 294 size_t RosAlloc::FreePages(Thread* self, void* ptr, bool already_zero) { argument 296 size_t pm_idx = ToPageMapIndex(ptr); 310 << static_cast<int>(pm_type) << ", ptr=" << std::hex 311 << reinterpret_cast<intptr_t>(ptr); 327 const uintptr_t* word_ptr = reinterpret_cast<uintptr_t*>(ptr); 333 memset(ptr, 0, byte_size); 337 LOG(INFO) << __PRETTY_FUNCTION__ << " : 0x" << std::hex << reinterpret_cast<intptr_t>(ptr) 338 << "-0x" << (reinterpret_cast<intptr_t>(ptr) + byte_size) 343 FreePageRun* fpr = reinterpret_cast<FreePageRun*>(ptr); 495 size_t RosAlloc::FreeInternal(Thread* self, void* ptr) { argument 539 Free(Thread* self, void* ptr) argument 772 FreeFromRun(Thread* self, void* ptr, Run* run) argument 883 FreeSlot(void* ptr) argument 1005 MarkThreadLocalFreeBitMap(void* ptr) argument 1010 MarkBulkFreeBitMap(void* ptr) argument 1014 MarkFreeBitMapShared(void* ptr, uint32_t* free_bit_map_base, const char* caller_name) argument 1175 void* ptr = ptrs[i]; local 1448 UsableSize(const void* ptr) argument [all...] |
/art/compiler/dex/quick/x86/ |
H A D | target_x86.cc | 968 uintptr_t ptr = reinterpret_cast<uintptr_t>(&id); local 973 static_cast<int>(ptr), type_idx,
|
/art/runtime/gc/ |
H A D | heap.cc | 1518 mirror::Object* ptr = TryToAllocate<true, false>(self, allocator, alloc_size, bytes_allocated, local 1520 if (ptr != nullptr) { 1521 return ptr; 1532 mirror::Object* ptr = TryToAllocate<true, false>(self, allocator, alloc_size, bytes_allocated, local 1534 if (ptr != nullptr) { 1535 return ptr; 1552 mirror::Object* ptr = TryToAllocate<true, false>(self, allocator, alloc_size, bytes_allocated, local 1554 if (ptr != nullptr) { 1555 return ptr; 1561 mirror::Object* ptr local [all...] |