/art/runtime/ |
H A D | atomic.cc | 26 Mutex* QuasiAtomic::GetSwapMutex(const volatile int64_t* addr) { argument 27 return (*gSwapMutexes)[(reinterpret_cast<uintptr_t>(addr) >> 3U) % kSwapMutexCount]; 46 int64_t QuasiAtomic::SwapMutexRead64(volatile const int64_t* addr) { argument 47 MutexLock mu(Thread::Current(), *GetSwapMutex(addr)); 48 return *addr; 51 void QuasiAtomic::SwapMutexWrite64(volatile int64_t* addr, int64_t value) { argument 52 MutexLock mu(Thread::Current(), *GetSwapMutex(addr)); 53 *addr = value; 57 bool QuasiAtomic::SwapMutexCas64(int64_t old_value, int64_t new_value, volatile int64_t* addr) { argument 58 MutexLock mu(Thread::Current(), *GetSwapMutex(addr)); [all...] |
H A D | atomic.h | 58 // Reads the 64-bit value at "addr" without tearing. 59 static int64_t Read64(volatile const int64_t* addr) { argument 63 value = *addr; 71 : "m" (*addr)); 77 : "Q" (*addr)); 83 : "m" (*addr)); 90 return SwapMutexRead64(addr); 94 // Writes to the 64-bit value at "addr" without tearing. 95 static void Write64(volatile int64_t* addr, int64_t value) { argument 98 *addr 141 Cas64(int64_t old_value, int64_t new_value, volatile int64_t* addr) argument [all...] |
H A D | mem_map.h | 63 static MemMap* MapAnonymous(const char* ashmem_name, byte* addr, size_t byte_count, int prot, 83 static MemMap* MapFileAtAddress(byte* addr, size_t byte_count, int prot, int flags, int fd, 126 bool HasAddress(const void* addr) const { 127 return Begin() <= addr && addr < End();
|
H A D | dex_file.h | 665 const byte* addr = begin_ + class_def.interfaces_off_; local 666 return reinterpret_cast<const TypeList*>(addr); 684 const byte* addr = begin_ + code_off; local 685 return reinterpret_cast<const CodeItem*>(addr); 737 const byte* addr = begin_ + proto_id.parameters_off_; local 738 return reinterpret_cast<const TypeList*>(addr);
|
H A D | stack.cc | 244 uint32_t* addr = GetVRegAddr(cur_quick_frame_, code_item, frame_info.CoreSpillMask(), local 246 *val = *reinterpret_cast<uint64_t*>(addr); 301 uint32_t* addr = GetVRegAddr(cur_quick_frame_, code_item, frame_info.CoreSpillMask(), local 303 *addr = new_value; 363 uint32_t* addr = GetVRegAddr(cur_quick_frame_, code_item, frame_info.CoreSpillMask(), local 365 *reinterpret_cast<uint64_t*>(addr) = new_value;
|
H A D | dex_file_verifier.cc | 428 uint32_t addr = DecodeUnsignedLeb128(&ptr_); local 429 if (UNLIKELY(addr >= code_item->insns_size_in_code_units_)) { 430 ErrorStringPrintf("Invalid handler addr: %x", addr); 436 uint32_t addr = DecodeUnsignedLeb128(&ptr_); local 437 if (UNLIKELY(addr >= code_item->insns_size_in_code_units_)) { 438 ErrorStringPrintf("Invalid handler catch_all_addr: %x", addr);
|
/art/runtime/base/ |
H A D | hex_dump.cc | 36 const unsigned char* addr = reinterpret_cast<const unsigned char*>(address_); local 47 offset = reinterpret_cast<size_t>(addr); 83 *hex++ = gHexDigit[*addr >> 4]; 84 *hex++ = gHexDigit[*addr & 0x0f]; 86 if (*addr >= 0x20 && *addr < 0x7f /*isprint(*addr)*/) { 87 *asc++ = *addr; 91 addr++;
|
/art/runtime/gc/accounting/ |
H A D | card_table_test.cc | 61 byte PRandCard(const byte* addr) const { 62 size_t offset = RoundDown(addr - heap_begin_, kCardSize); 66 for (const byte* addr = HeapBegin(); addr != HeapLimit(); addr += kCardSize) { 67 EXPECT_TRUE(card_table_->AddrIsInCardTable(addr)); 68 byte* card = card_table_->CardFromAddr(addr); 69 *card = PRandCard(addr); 80 for (const byte* addr = HeapBegin(); addr < HeapLimi [all...] |
H A D | card_table.cc | 105 bool CardTable::AddrIsInCardTable(const void* addr) const { 106 return IsValidCard(biased_begin_ + ((uintptr_t)addr >> kCardShift)); 109 void CardTable::CheckAddrIsInCardTable(const byte* addr) const { 110 byte* card_addr = biased_begin_ + ((uintptr_t)addr >> kCardShift); 113 CHECK(AddrIsInCardTable(addr)) 120 << " addr: " << reinterpret_cast<const void*>(addr);
|
H A D | card_table.h | 57 ALWAYS_INLINE void MarkCard(const void *addr) { argument 58 *CardFromAddr(addr) = kCardDirty; 112 void CheckAddrIsInCardTable(const byte* addr) const; 124 byte* CardFromAddr(const void *addr) const ALWAYS_INLINE; 126 bool AddrIsInCardTable(const void* addr) const;
|
H A D | space_bitmap-inl.h | 34 uintptr_t addr = reinterpret_cast<uintptr_t>(obj); local 35 DCHECK_GE(addr, heap_begin_); 36 const uintptr_t offset = addr - heap_begin_; 56 uintptr_t addr = reinterpret_cast<uintptr_t>(obj); local 59 DCHECK_GE(addr, heap_begin_); 60 const uintptr_t offset = addr - heap_begin_; 160 uintptr_t addr = reinterpret_cast<uintptr_t>(obj); local 161 DCHECK_GE(addr, heap_begin_); 162 const uintptr_t offset = addr - heap_begin_;
|
H A D | atomic_stack.h | 216 byte* addr = mem_map_->Begin(); local 217 CHECK(addr != NULL); 219 begin_ = reinterpret_cast<T*>(addr);
|
H A D | card_table-inl.h | 209 inline byte* CardTable::CardFromAddr(const void *addr) const { 210 byte *card_addr = biased_begin_ + (reinterpret_cast<uintptr_t>(addr) >> kCardShift); 212 DCHECK(IsValidCard(card_addr)) << "addr: " << addr
|
H A D | space_bitmap.h | 186 uintptr_t addr = reinterpret_cast<uintptr_t>(obj); local 187 const uintptr_t offset = addr - heap_begin_;
|
/art/runtime/jdwp/ |
H A D | jdwp_socket.cc | 138 } addr; local 139 addr.addrInet.sin_family = AF_INET; 140 addr.addrInet.sin_port = htons(port); 141 inet_aton("127.0.0.1", &addr.addrInet.sin_addr); 143 if (bind(netState->listenSock, &addr.addrPlain, sizeof(addr)) != 0) { 213 } addr; local 223 addrlen = sizeof(addr); 225 sock = accept(listenSock, &addr.addrPlain, &addrlen); 240 remote_addr_ = addr 265 } addr; local 428 } addr; local [all...] |
/art/compiler/ |
H A D | elf_writer_test.cc | 36 void* addr = reinterpret_cast<void*>(ef->FindSymbolAddress(SHT_DYNSYM, \ 39 EXPECT_NE(nullptr, addr); \ 40 EXPECT_LT(static_cast<uintptr_t>(ART_BASE_ADDRESS), reinterpret_cast<uintptr_t>(addr)); \ 42 expected_value = addr; \ 44 EXPECT_EQ(expected_value, addr); \
|
H A D | elf_writer_mclinker.cc | 408 uint32_t* addr = reinterpret_cast<uint32_t*>(elf_file.Begin() + oatdata_offset); local 409 *addr = compiled_code_offset;
|
H A D | elf_writer_quick.h | 162 Elf32_Addr addr,
|
/art/compiler/utils/ |
H A D | scoped_arena_allocator.h | 107 void* addr = arena_stack->Alloc(sizeof(ScopedArenaAllocator), kArenaAllocMisc); local 108 ScopedArenaAllocator* allocator = new(addr) ScopedArenaAllocator(arena_stack); 109 allocator->mark_ptr_ = reinterpret_cast<uint8_t*>(addr);
|
/art/runtime/gc/allocator/ |
H A D | rosalloc.h | 356 size_t ToPageMapIndex(const void* addr) const { 357 DCHECK(base_ <= addr && addr < base_ + capacity_); 358 size_t byte_offset = reinterpret_cast<const byte*>(addr) - base_; 363 size_t RoundDownToPageMapIndex(void* addr) const { 364 DCHECK(base_ <= addr && addr < reinterpret_cast<byte*>(base_) + capacity_); 365 return (reinterpret_cast<uintptr_t>(addr) - reinterpret_cast<uintptr_t>(base_)) / kPageSize;
|
/art/runtime/mirror/ |
H A D | object-inl.h | 518 const int64_t* addr = reinterpret_cast<const int64_t*>(raw_addr); local 520 return reinterpret_cast<const Atomic<int64_t>*>(addr)->LoadSequentiallyConsistent(); 522 return reinterpret_cast<const Atomic<int64_t>*>(addr)->LoadJavaData(); 546 int64_t* addr = reinterpret_cast<int64_t*>(raw_addr); local 548 reinterpret_cast<Atomic<int64_t>*>(addr)->StoreSequentiallyConsistent(new_value); 550 reinterpret_cast<Atomic<int64_t>*>(addr)->StoreJavaData(new_value);
|
/art/compiler/utils/arm/ |
H A D | assembler_arm32.cc | 550 const Address& addr = static_cast<const Address&>(ad); local 576 addr.encodingArm(); 588 const Address& addr = static_cast<const Address&>(ad); local 593 addr.encoding3(); 877 const Address& addr = static_cast<const Address&>(ad); local 884 B11 | B9 | addr.vencoding(); 890 const Address& addr = static_cast<const Address&>(ad); local 891 CHECK_NE(static_cast<Register>(addr.encodingArm() & (0xf << kRnShift)), PC); 898 B11 | B9 | addr.vencoding(); 904 const Address& addr local 917 const Address& addr = static_cast<const Address&>(ad); local [all...] |
H A D | assembler_thumb2.cc | 1762 const Address& addr = static_cast<const Address&>(ad); local 1769 B11 | B9 | addr.vencoding(); 1775 const Address& addr = static_cast<const Address&>(ad); local 1776 CHECK_NE(static_cast<Register>(addr.encodingArm() & (0xf << kRnShift)), PC); 1783 B11 | B9 | addr.vencoding(); 1789 const Address& addr = static_cast<const Address&>(ad); local 1796 B11 | B9 | B8 | addr.vencoding(); 1802 const Address& addr = static_cast<const Address&>(ad); local 1803 CHECK_NE(static_cast<Register>(addr.encodingArm() & (0xf << kRnShift)), PC); 1810 B11 | B9 | B8 | addr [all...] |
/art/compiler/utils/x86_64/ |
H A D | assembler_x86_64.h | 229 static Address Absolute(uword addr, bool no_rip = false) { argument 234 result.SetDisp32(addr); 237 result.SetDisp32(addr); 243 static Address Absolute(ThreadOffset<8> addr, bool no_rip = false) { argument 244 return Absolute(addr.Int32Value(), no_rip);
|
/art/compiler/utils/x86/ |
H A D | assembler_x86.h | 194 static Address Absolute(uword addr) { argument 197 result.SetDisp32(addr); 201 static Address Absolute(ThreadOffset<4> addr) { argument 202 return Absolute(addr.Int32Value());
|