/art/runtime/ |
H A D | atomic.cc | 26 Mutex* QuasiAtomic::GetSwapMutex(const volatile int64_t* addr) { argument 27 return (*gSwapMutexes)[(reinterpret_cast<uintptr_t>(addr) >> 3U) % kSwapMutexCount]; 46 int64_t QuasiAtomic::SwapMutexRead64(volatile const int64_t* addr) { argument 47 MutexLock mu(Thread::Current(), *GetSwapMutex(addr)); 48 return *addr; 51 void QuasiAtomic::SwapMutexWrite64(volatile int64_t* addr, int64_t value) { argument 52 MutexLock mu(Thread::Current(), *GetSwapMutex(addr)); 53 *addr = value; 57 bool QuasiAtomic::SwapMutexCas64(int64_t old_value, int64_t new_value, volatile int64_t* addr) { argument 58 MutexLock mu(Thread::Current(), *GetSwapMutex(addr)); [all...] |
H A D | atomic.h | 61 // Reads the 64-bit value at "addr" without tearing. 62 static int64_t Read64(volatile const int64_t* addr) { argument 66 value = *addr; 74 : "m" (*addr)); 80 : "Q" (*addr)); 86 : "m" (*addr)); 93 return SwapMutexRead64(addr); 97 // Writes to the 64-bit value at "addr" without tearing. 98 static void Write64(volatile int64_t* addr, int64_t value) { argument 101 *addr 144 Cas64(int64_t old_value, int64_t new_value, volatile int64_t* addr) argument [all...] |
H A D | mem_map.h | 64 static MemMap* MapAnonymous(const char* ashmem_name, uint8_t* addr, size_t byte_count, int prot, 71 static MemMap* MapDummy(const char* name, uint8_t* addr, size_t byte_count); 90 static MemMap* MapFileAtAddress(uint8_t* addr, size_t byte_count, int prot, int flags, int fd, 136 bool HasAddress(const void* addr) const { 137 return Begin() <= addr && addr < End();
|
H A D | art_method.h | 597 const auto addr = reinterpret_cast<uintptr_t>(this) + offset.Uint32Value(); local 599 return reinterpret_cast<T>(*reinterpret_cast<const uint32_t*>(addr)); 601 auto v = *reinterpret_cast<const uint64_t*>(addr); 610 const auto addr = reinterpret_cast<uintptr_t>(this) + offset.Uint32Value(); local 614 *reinterpret_cast<uint32_t*>(addr) = static_cast<uint32_t>(ptr); 616 *reinterpret_cast<uint64_t*>(addr) = reinterpret_cast<uintptr_t>(new_value);
|
H A D | stack.cc | 250 const uint8_t* addr = reinterpret_cast<const uint8_t*>(cur_quick_frame_) + offset; local 251 *val = *reinterpret_cast<const uint32_t*>(addr); 338 uint32_t* addr = GetVRegAddrFromQuickCode( local 341 *val = *reinterpret_cast<uint64_t*>(addr); 413 uint32_t* addr = GetVRegAddrFromQuickCode( local 416 *addr = new_value; 495 uint32_t* addr = GetVRegAddrFromQuickCode( local 498 *reinterpret_cast<uint64_t*>(addr) = new_value;
|
/art/runtime/base/ |
H A D | hex_dump.cc | 36 const unsigned char* addr = reinterpret_cast<const unsigned char*>(address_); local 47 offset = reinterpret_cast<size_t>(addr); 83 *hex++ = gHexDigit[*addr >> 4]; 84 *hex++ = gHexDigit[*addr & 0x0f]; 86 if (*addr >= 0x20 && *addr < 0x7f /*isprint(*addr)*/) { 87 *asc++ = *addr; 91 addr++;
|
H A D | scoped_arena_allocator.h | 107 void* addr = arena_stack->Alloc(sizeof(ScopedArenaAllocator), kArenaAllocMisc); local 108 ScopedArenaAllocator* allocator = new(addr) ScopedArenaAllocator(arena_stack); 109 allocator->mark_ptr_ = reinterpret_cast<uint8_t*>(addr);
|
/art/runtime/gc/accounting/ |
H A D | bitmap.h | 148 const uintptr_t addr = CoverBegin() + bit_index * kAlignment; local 149 DCHECK_EQ(BitIndexFromAddr(addr), bit_index); 150 return addr; 154 ALWAYS_INLINE uintptr_t BitIndexFromAddr(uintptr_t addr) const { 155 DCHECK(HasAddress(addr)) << CoverBegin() << " <= " << addr << " < " << CoverEnd(); 156 return (addr - CoverBegin()) / kAlignment; 159 ALWAYS_INLINE bool HasAddress(const uintptr_t addr) const { 160 return cover_begin_ <= addr && addr < cover_end 163 Set(uintptr_t addr) argument 167 Clear(size_t addr) argument 176 AtomicTestAndSet(size_t addr) argument [all...] |
H A D | card_table_test.cc | 64 uint8_t PseudoRandomCard(const uint8_t* addr) const { 65 size_t offset = RoundDown(addr - heap_begin_, CardTable::kCardSize); 69 for (const uint8_t* addr = HeapBegin(); addr != HeapLimit(); addr += CardTable::kCardSize) { 70 EXPECT_TRUE(card_table_->AddrIsInCardTable(addr)); 71 uint8_t* card = card_table_->CardFromAddr(addr); 72 *card = PseudoRandomCard(addr); 83 for (const uint8_t* addr = HeapBegin(); addr < HeapLimi [all...] |
H A D | card_table.h | 59 ALWAYS_INLINE void MarkCard(const void *addr) { argument 60 *CardFromAddr(addr) = kCardDirty; 114 void CheckAddrIsInCardTable(const uint8_t* addr) const; 127 uint8_t* CardFromAddr(const void *addr) const ALWAYS_INLINE; 129 bool AddrIsInCardTable(const void* addr) const;
|
H A D | mod_union_table.cc | 417 for (uint8_t* addr = space_->Begin(); addr < AlignUp(space_->End(), CardTable::kCardSize); 418 addr += CardTable::kCardSize) { 419 if (card_bitmap_->Test(reinterpret_cast<uintptr_t>(addr))) { 420 os << reinterpret_cast<void*>(addr) << "-" 421 << reinterpret_cast<void*>(addr + CardTable::kCardSize) << "\n"; 429 for (uint8_t* addr = space_->Begin(); addr < AlignUp(space_->End(), CardTable::kCardSize); 430 addr += CardTable::kCardSize) { 431 card_bitmap_->Set(reinterpret_cast<uintptr_t>(addr)); 435 ContainsCardFor(uintptr_t addr) argument 446 ContainsCardFor(uintptr_t addr) argument [all...] |
H A D | card_table.cc | 130 bool CardTable::AddrIsInCardTable(const void* addr) const { 131 return IsValidCard(biased_begin_ + ((uintptr_t)addr >> kCardShift)); 134 void CardTable::CheckAddrIsInCardTable(const uint8_t* addr) const { 135 uint8_t* card_addr = biased_begin_ + ((uintptr_t)addr >> kCardShift); 138 CHECK(AddrIsInCardTable(addr)) 145 << " addr: " << reinterpret_cast<const void*>(addr);
|
H A D | space_bitmap-inl.h | 34 uintptr_t addr = reinterpret_cast<uintptr_t>(obj); local 35 DCHECK_GE(addr, heap_begin_); 36 const uintptr_t offset = addr - heap_begin_; 56 uintptr_t addr = reinterpret_cast<uintptr_t>(obj); local 59 DCHECK_GE(addr, heap_begin_); 60 const uintptr_t offset = addr - heap_begin_; 160 uintptr_t addr = reinterpret_cast<uintptr_t>(obj); local 161 DCHECK_GE(addr, heap_begin_); 163 const uintptr_t offset = addr - heap_begin_;
|
H A D | mod_union_table.h | 89 virtual bool ContainsCardFor(uintptr_t addr) = 0; 133 virtual bool ContainsCardFor(uintptr_t addr) OVERRIDE; 169 virtual bool ContainsCardFor(uintptr_t addr) OVERRIDE;
|
H A D | atomic_stack.h | 242 uint8_t* addr = mem_map_->Begin(); local 243 CHECK(addr != nullptr); 245 begin_ = reinterpret_cast<StackReference<T>*>(addr);
|
H A D | card_table-inl.h | 219 inline uint8_t* CardTable::CardFromAddr(const void *addr) const { 220 uint8_t *card_addr = biased_begin_ + (reinterpret_cast<uintptr_t>(addr) >> kCardShift); 222 DCHECK(IsValidCard(card_addr)) << "addr: " << addr
|
/art/tools/dexfuzz/src/dexfuzz/rawdex/ |
H A D | EncodedTypeAddrPair.java | 23 public int addr; field in class:EncodedTypeAddrPair 28 addr = file.readUleb128(); 34 file.writeUleb128(addr);
|
/art/runtime/jdwp/ |
H A D | jdwp_socket.cc | 138 } addr; local 139 addr.addrInet.sin_family = AF_INET; 140 addr.addrInet.sin_port = htons(port); 141 inet_aton("127.0.0.1", &addr.addrInet.sin_addr); 143 if (bind(netState->listenSock, &addr.addrPlain, sizeof(addr)) != 0) { 213 } addr; local 223 addrlen = sizeof(addr); 225 sock = accept(listenSock, &addr.addrPlain, &addrlen); 240 remote_addr_ = addr 265 } addr; local 428 } addr; local [all...] |
/art/compiler/linker/arm/ |
H A D | relative_patcher_thumb2.cc | 94 uint8_t* addr = &(*code)[offset]; local 95 addr[0] = (value >> 16) & 0xff; 96 addr[1] = (value >> 24) & 0xff; 97 addr[2] = (value >> 0) & 0xff; 98 addr[3] = (value >> 8) & 0xff; 104 const uint8_t* addr = &code[offset]; local 106 (static_cast<uint32_t>(addr[0]) << 16) + 107 (static_cast<uint32_t>(addr[1]) << 24) + 108 (static_cast<uint32_t>(addr[2]) << 0)+ 109 (static_cast<uint32_t>(addr[ [all...] |
/art/compiler/ |
H A D | elf_writer_test.cc | 40 void* addr = reinterpret_cast<void*>(ef->FindSymbolAddress(SHT_DYNSYM, \ 43 EXPECT_NE(nullptr, addr); \ 44 EXPECT_LT(static_cast<uintptr_t>(ART_BASE_ADDRESS), reinterpret_cast<uintptr_t>(addr)); \ 46 expected_value = addr; \ 48 EXPECT_EQ(expected_value, addr); \
|
/art/compiler/linker/arm64/ |
H A D | relative_patcher_arm64.cc | 301 uint8_t* addr = &(*code)[offset]; local 302 addr[0] = (value >> 0) & 0xff; 303 addr[1] = (value >> 8) & 0xff; 304 addr[2] = (value >> 16) & 0xff; 305 addr[3] = (value >> 24) & 0xff; 311 const uint8_t* addr = &code[offset]; local 313 (static_cast<uint32_t>(addr[0]) << 0) + 314 (static_cast<uint32_t>(addr[1]) << 8) + 315 (static_cast<uint32_t>(addr[2]) << 16)+ 316 (static_cast<uint32_t>(addr[ [all...] |
/art/runtime/gc/allocator/ |
H A D | rosalloc.h | 375 size_t ToPageMapIndex(const void* addr) const { 376 DCHECK_LE(base_, addr); 377 DCHECK_LT(addr, base_ + capacity_); 378 size_t byte_offset = reinterpret_cast<const uint8_t*>(addr) - base_; 383 size_t RoundDownToPageMapIndex(const void* addr) const { 384 DCHECK(base_ <= addr && addr < reinterpret_cast<uint8_t*>(base_) + capacity_); 385 return (reinterpret_cast<uintptr_t>(addr) - reinterpret_cast<uintptr_t>(base_)) / kPageSize;
|
/art/runtime/gc/space/ |
H A D | region_space.cc | 295 for (uint8_t* addr = begin_addr; addr < end_addr; addr += kRegionSize) { 296 Region* reg = RefToRegionLocked(reinterpret_cast<mirror::Object*>(addr)); 297 if (addr == begin_addr) {
|
/art/runtime/gc/collector/ |
H A D | concurrent_copying.cc | 1163 Atomic<mirror::Object*>* addr = reinterpret_cast<Atomic<mirror::Object*>*>(root); local 1167 if (expected_ref != addr->LoadRelaxed()) { 1171 } while (!addr->CompareExchangeWeakSequentiallyConsistent(expected_ref, new_ref)); 1188 auto* addr = reinterpret_cast<Atomic<mirror::CompressedReference<mirror::Object>>*>(root); local 1192 if (ref != addr->LoadRelaxed().AsMirrorPtr()) { 1196 } while (!addr->CompareExchangeWeakSequentiallyConsistent(expected_ref, new_ref)); 1263 uint8_t* addr = it->second; local 1265 CHECK(region_space_->IsInToSpace(reinterpret_cast<mirror::Object*>(addr))); 1268 LOG(INFO) << "Reusing skipped bytes : " << reinterpret_cast<void*>(addr) << ", " << byte_size; 1271 memset(addr, [all...] |
/art/compiler/optimizing/ |
H A D | intrinsics_arm.cc | 389 Register addr = invoke->GetLocations()->InAt(0).AsRegisterPairLow<Register>(); local 391 // exception. So we can't use ldrd as addr may be unaligned. 394 if (addr == lo) { 395 __ ldr(hi, Address(addr, 4)); 396 __ ldr(lo, Address(addr, 0)); 398 __ ldr(lo, Address(addr, 0)); 399 __ ldr(hi, Address(addr, 4)); 449 Register addr = invoke->GetLocations()->InAt(0).AsRegisterPairLow<Register>(); local 451 // exception. So we can't use ldrd as addr may be unaligned. 452 __ str(invoke->GetLocations()->InAt(1).AsRegisterPairLow<Register>(), Address(addr, [all...] |