/art/runtime/mirror/ |
H A D | object_reference-inl.h | 28 void ObjectReference<kPoisonReferences, MirrorType>::Assign(ObjPtr<MirrorType> ptr) { argument 29 Assign(ptr.Ptr());
|
/art/libartbase/base/ |
H A D | casts.h | 160 inline Dest reinterpret_cast64(Source* ptr) { argument 165 return static_cast<Dest>(reinterpret_cast<uintptr_t>(ptr));
|
H A D | scoped_flock.h | 78 void operator()(LockedFile* ptr) { argument 79 ptr->ReleaseLock(); 80 UNUSED(ptr->Close()); 82 delete ptr;
|
H A D | variant_map_test.cc | 86 auto* ptr = fm.Get(FruitMap::Label); local 87 ASSERT_TRUE(ptr != nullptr); 88 *ptr = "foobar"; 92 fm.Set(FruitMap::Label, *ptr); 95 ASSERT_TRUE(ptr != nullptr);
|
H A D | stride_iterator.h | 37 StrideIterator(T* ptr, size_t stride) argument 38 : ptr_(reinterpret_cast<uintptr_t>(ptr)),
|
H A D | time_utils.cc | 103 char* ptr = fraction_buffer; local 108 *ptr++ = '0'; 111 snprintf(ptr, avail_digits, "%" PRIu64, fractional_part);
|
/art/openjdkjvmti/ |
H A D | transform.cc | 98 uintptr_t ptr = reinterpret_cast<uintptr_t>(siginfo->si_addr); variable 109 [&](const auto op) { return op->ContainsAddress(ptr); }); 117 // Wait for the ptr to be initialized (if it is currently initializing). 118 while (DefinitionIsInitializing(ptr)) { 126 [&](const auto op) { return op->ContainsAddress(ptr); }) != 182 bool DefinitionIsInitializing(uintptr_t ptr) REQUIRES(uninitialized_class_definitions_lock_) { 185 [&](const auto op) { return op->ContainsAddress(ptr); }) !=
|
/art/runtime/ |
H A D | imtable.h | 46 uint8_t* ptr = AddressOfElement(index, pointer_size); local 48 uint32_t value = *reinterpret_cast<uint32_t*>(ptr); 51 uint64_t value = *reinterpret_cast<uint64_t*>(ptr); 58 uint8_t* ptr = AddressOfElement(index, pointer_size); local 62 *reinterpret_cast<uint32_t*>(ptr) = static_cast<uint32_t>(value); 64 *reinterpret_cast<uint64_t*>(ptr) = reinterpret_cast<uint64_t>(method);
|
H A D | method_info.h | 35 explicit MethodInfo(const uint8_t* ptr) { argument 36 if (ptr != nullptr) { 37 num_method_indices_ = DecodeUnsignedLeb128(&ptr); 38 region_ = MemoryRegion(const_cast<uint8_t*>(ptr), 44 MethodInfo(uint8_t* ptr, size_t num_method_indices) : num_method_indices_(num_method_indices) { argument 45 DCHECK(ptr != nullptr); 46 ptr = EncodeUnsignedLeb128(ptr, num_method_indices_); 47 region_ = MemoryRegion(ptr, num_method_indices_ * sizeof(MethodIndexType)); 52 uint8_t* ptr local [all...] |
H A D | obj_ptr-inl.h | 43 inline uintptr_t ObjPtr<MirrorType>::Encode(MirrorType* ptr) { argument 44 uintptr_t ref = reinterpret_cast<uintptr_t>(ptr); 58 inline std::ostream& operator<<(std::ostream& os, ObjPtr<MirrorType> ptr) { argument 60 return os << ptr.PtrUnchecked();
|
H A D | linear_alloc.cc | 26 void* LinearAlloc::Realloc(Thread* self, void* ptr, size_t old_size, size_t new_size) { argument 28 return allocator_.Realloc(ptr, old_size, new_size); 51 bool LinearAlloc::Contains(void* ptr) const { 53 return allocator_.Contains(ptr); 56 bool LinearAlloc::ContainsUnsafe(void* ptr) const { 57 return allocator_.Contains(ptr);
|
H A D | obj_ptr.h | 60 ALWAYS_INLINE ObjPtr(Type* ptr) 62 : reference_(Encode(static_cast<MirrorType*>(ptr))) { 84 ALWAYS_INLINE ObjPtr& operator=(MirrorType* ptr) REQUIRES_SHARED(Locks::mutator_lock_) { 85 Assign(ptr); 89 ALWAYS_INLINE void Assign(MirrorType* ptr) REQUIRES_SHARED(Locks::mutator_lock_) { 90 reference_ = Encode(ptr); 111 ALWAYS_INLINE bool operator==(const ObjPtr& ptr) const REQUIRES_SHARED(Locks::mutator_lock_) { 112 return Ptr() == ptr.Ptr(); 116 ALWAYS_INLINE bool operator==(const PointerType* ptr) const 118 return Ptr() == ptr; 209 MakeObjPtr(MirrorType* ptr) argument 214 MakeObjPtr(ObjPtr<MirrorType> ptr) argument [all...] |
H A D | type_lookup_table.h | 130 const uint8_t* ptr = dex_data_begin_ + str_offset; local 133 DecodeUnsignedLeb128(&ptr); 135 str, reinterpret_cast<const char*>(ptr)) == 0;
|
H A D | handle_scope-inl.h | 195 inline MutableHandle<MirrorType> VariableSizedHandleScope::NewHandle(ObjPtr<MirrorType> ptr) { argument 196 return NewHandle(ptr.Ptr());
|
/art/test/1909-per-agent-tls/src/art/ |
H A D | Test1909.java | 36 public void setTLS(long jvmtienv, long ptr) { argument 37 Test1909.setTLS(jvmtienv, this.thr, ptr); 75 public void setTLS(long jvmtienv, long ptr) { argument 76 Test1909.setTLS(jvmtienv, this.thr, ptr); 173 public static native void setTLS(long jvmtienv, Thread thr, long ptr); argument
|
/art/runtime/base/ |
H A D | scoped_arena_allocator.cc | 97 uint8_t* ptr = top_ptr_; local 98 if (UNLIKELY(static_cast<size_t>(top_end_ - ptr) < rounded_bytes)) { 99 ptr = AllocateFromNextArena(rounded_bytes); 100 CHECK(ptr != nullptr) << "Failed to allocate memory"; 101 MEMORY_TOOL_MAKE_NOACCESS(ptr, top_end_ - ptr); 104 top_ptr_ = ptr + rounded_bytes; 105 MEMORY_TOOL_MAKE_UNDEFINED(ptr, bytes); 106 return ptr;
|
H A D | scoped_arena_allocator.h | 65 static ArenaFreeTag& ArenaTagForAllocation(void* ptr) { argument 67 return *(reinterpret_cast<ArenaFreeTag*>(ptr) - 1); 101 uint8_t* ptr = top_ptr_; variable 102 if (UNLIKELY(static_cast<size_t>(top_end_ - ptr) < rounded_bytes)) { 103 ptr = AllocateFromNextArena(rounded_bytes); 106 top_ptr_ = ptr + rounded_bytes; 108 ptr += kAlignment; 109 ArenaTagForAllocation(ptr) = ArenaFreeTag::kUsed; variable 111 return ptr; 170 static void operator delete(void* ptr ATTRIBUTE_UNUSE [all...] |
H A D | scoped_arena_containers.h | 230 ALWAYS_INLINE void ProtectMemory(T* ptr, size_t size) const { argument 234 memset(ptr, kMagicFill, size); 235 MEMORY_TOOL_MAKE_NOACCESS(ptr, size); 237 CHECK(ArenaStack::ArenaTagForAllocation(reinterpret_cast<void*>(ptr)) == ArenaFreeTag::kUsed) 238 << "Freeing invalid object " << ptr; 239 ArenaStack::ArenaTagForAllocation(reinterpret_cast<void*>(ptr)) = ArenaFreeTag::kFree; 241 memset(ptr, kMagicFill, size); 246 void operator()(T* ptr) const { 247 if (ptr != nullptr) { 248 ptr [all...] |
/art/test/1900-track-alloc/src/art/ |
H A D | Test1900.java | 143 private static native void doDeallocate(long jvmtienv, long ptr); argument
|
/art/compiler/debug/dwarf/ |
H A D | writer.h | 119 void PushData(const uint8_t* ptr, size_t num_bytes) { argument 120 data_->insert(data_->end(), ptr, ptr + num_bytes); 123 void PushData(const char* ptr, size_t num_bytes) { argument 124 data_->insert(data_->end(), ptr, ptr + num_bytes);
|
/art/compiler/utils/ |
H A D | swap_space.cc | 75 if (munmap(chunk.ptr, chunk.size) != 0) { 77 << static_cast<const void*>(chunk.ptr) << " size=" << chunk.size; 126 // The free_by_start_ map contains disjoint intervals ordered by the `ptr`. 128 it->free_by_start_entry->ptr += size; 131 // The free_by_size_ map is ordered by the `size` and then `free_by_start_entry->ptr`. 132 // Adjusting the `ptr` above does not change that ordering but decreasing `size` can 149 return old_chunk.ptr; 155 SpaceChunk remainder = { new_chunk.ptr + size, new_chunk.size - size }; 158 return new_chunk.ptr; 169 uint8_t* ptr local 190 Free(void* ptr, size_t size) argument [all...] |
/art/runtime/arch/arm/ |
H A D | fault_handler_arm.cc | 76 uint8_t* ptr = reinterpret_cast<uint8_t*>(sc->arm_pc); local 77 VLOG(signals) << "pc: " << std::hex << static_cast<void*>(ptr); 79 if (ptr == nullptr) { 85 uint32_t instr_size = GetInstructionSize(ptr); 103 uint8_t* ptr = reinterpret_cast<uint8_t*>(sc->arm_pc); local 104 uint32_t instr_size = GetInstructionSize(ptr);
|
/art/runtime/gc/space/ |
H A D | large_object_space_test.cc | 134 mirror::Object* ptr = los_->Alloc(self, size_, &alloc_size, nullptr, local 139 los_->Free(self, ptr);
|
H A D | memory_tool_malloc_space-inl.h | 198 Thread* self, mirror::Object* ptr) { 199 void* obj_after_rdz = reinterpret_cast<void*>(ptr); 204 size_t allocation_size = AllocationSize(ptr, &usable_size); 197 Free( Thread* self, mirror::Object* ptr) argument
|
/art/runtime/gc/ |
H A D | verification.cc | 43 uint8_t* ptr = reinterpret_cast<uint8_t*>(p); local 44 oss << std::hex << std::setfill('0') << std::setw(2) << static_cast<uintptr_t>(*ptr);
|