/art/test/102-concurrent-gc/src/ |
H A D | Main.java | 24 public byte[] bytes; field in class:Main.ByteContainer 37 l[index].bytes = new byte[bufferSize]; 57 byte[] temp = l[a].bytes; 58 l[a].bytes = l[b].bytes; 59 l[b].bytes = temp;
|
/art/runtime/jdwp/ |
H A D | jdwp_bits.h | 35 static inline void Append1BE(std::vector<uint8_t>& bytes, uint8_t value) { argument 36 bytes.push_back(value); 39 static inline void Append2BE(std::vector<uint8_t>& bytes, uint16_t value) { argument 40 bytes.push_back(static_cast<uint8_t>(value >> 8)); 41 bytes.push_back(static_cast<uint8_t>(value)); 44 static inline void Append4BE(std::vector<uint8_t>& bytes, uint32_t value) { argument 45 bytes.push_back(static_cast<uint8_t>(value >> 24)); 46 bytes.push_back(static_cast<uint8_t>(value >> 16)); 47 bytes.push_back(static_cast<uint8_t>(value >> 8)); 48 bytes 51 Append8BE(std::vector<uint8_t>& bytes, uint64_t value) argument 62 AppendUtf16BE(std::vector<uint8_t>& bytes, const uint16_t* chars, size_t char_count) argument [all...] |
H A D | jdwp_request.cc | 28 Request::Request(const uint8_t* bytes, uint32_t available) : p_(bytes) { argument 30 end_ = bytes + byte_count_; 48 CHECK(p_ == end_) << "read too few bytes: " << (end_ - p_); 50 CHECK(p_ == end_) << "read too many bytes: " << (p_ - end_);
|
/art/test/004-NativeAllocations/src/ |
H A D | Main.java | 31 private int bytes; field in class:Main.NativeAllocation 33 NativeAllocation(int bytes, boolean testingDeadlock) throws Exception { argument 34 this.bytes = bytes; 35 register_native_allocation.invoke(runtime, bytes); 38 nativeBytes += bytes; 48 nativeBytes -= bytes; 50 register_native_free.invoke(runtime, bytes);
|
/art/runtime/native/ |
H A D | org_apache_harmony_dalvik_ddmc_DdmVmInternal.cc | 100 * (1b) bytes per entry 121 std::vector<uint8_t>& bytes = *reinterpret_cast<std::vector<uint8_t>*>(context); local 122 JDWP::Append4BE(bytes, t->GetThreadId()); 123 JDWP::Append1BE(bytes, Dbg::ToJdwpThreadStatus(t->GetState())); 124 JDWP::Append4BE(bytes, t->GetTid()); 125 JDWP::Append4BE(bytes, utime); 126 JDWP::Append4BE(bytes, stime); 127 JDWP::Append1BE(bytes, t->IsDaemon()); 131 std::vector<uint8_t> bytes; local 140 JDWP::Append1BE(bytes, kThstHeaderLe [all...] |
H A D | libcore_util_CharsetUtils.cc | 34 * We could avoid this by keeping the UTF-8 bytes on the native heap until we're done and only 112 ScopedByteArrayRO bytes(env, javaBytes); 113 if (bytes.get() == nullptr) { 121 const jbyte* src = &bytes[offset]; 132 ScopedByteArrayRO bytes(env, javaBytes); 133 if (bytes.get() == nullptr) { 141 const jbyte* src = &bytes[offset]; 149 * Translates the given characters to US-ASCII or ISO-8859-1 bytes, using the fact that 163 ScopedByteArrayRW bytes(env, javaBytes); 164 if (bytes [all...] |
/art/cmdline/ |
H A D | memory_representation.h | 28 // An integral representation of bytes of memory. 34 static Memory<kDivisor> FromBytes(size_t bytes) { argument 35 assert(bytes % kDivisor == 0); 36 return Memory<kDivisor>(bytes);
|
/art/test/580-checker-string-factory-intrinsics/src/ |
H A D | Main.java | 26 byte[] bytes = { 'f', 'o', 'o' }; 27 String s = StringFactory.newStringFromBytes(bytes, 0, 0, 3);
|
/art/runtime/arch/x86_64/ |
H A D | memcmp16_x86_64.S | 701 jnc L(16bytes) 706 jnc L(16bytes) 711 jnc L(16bytes) 715 jmp L(16bytes) 720 jmp L(16bytes) 724 jmp L(16bytes) 728 jmp L(16bytes) 732 jmp L(16bytes) 736 jmp L(16bytes) 740 jmp L(16bytes) [all...] |
/art/test/407-arrays/src/ |
H A D | Main.java | 32 static void $opt$testReads(boolean[] bools, byte[] bytes, char[] chars, short[] shorts, argument 38 assertEquals(0, bytes[0]); 39 assertEquals(0, bytes[index]); 63 static void $opt$testWrites(boolean[] bools, byte[] bytes, char[] chars, short[] shorts, argument 71 bytes[0] = -4; 72 assertEquals(-4, bytes[0]); 73 bytes[index] = -8; 74 assertEquals(-8, bytes[index]);
|
/art/runtime/base/ |
H A D | scoped_arena_allocator.h | 88 void* Alloc(size_t bytes, ArenaAllocKind kind) ALWAYS_INLINE { 90 return AllocWithMemoryTool(bytes, kind); 93 size_t rounded_bytes = RoundUp(bytes + (kIsDebugBuild ? kArenaAlignment : 0u), kArenaAlignment); 98 CurrentStats()->RecordAlloc(bytes, kind); 110 void* AllocWithMemoryTool(size_t bytes, ArenaAllocKind kind); 149 void* Alloc(size_t bytes, ArenaAllocKind kind = kArenaAllocMisc) ALWAYS_INLINE { 151 return arena_stack_->Alloc(bytes, kind);
|
H A D | arena_allocator.cc | 100 void ArenaAllocatorStatsImpl<kCount>::RecordAlloc(size_t bytes, ArenaAllocKind kind) { argument 101 alloc_stats_[kind] += bytes; 312 // Update how many bytes we have allocated into the arena so that the arena pool knows how 318 void* ArenaAllocator::AllocWithMemoryTool(size_t bytes, ArenaAllocKind kind) { argument 322 size_t rounded_bytes = RoundUp(bytes + kMemoryToolRedZoneBytes, 8); 327 uint8_t* noaccess_begin = ret + bytes; 344 MEMORY_TOOL_MAKE_DEFINED(ret, bytes); 346 DCHECK(std::all_of(ret, ret + bytes, [](uint8_t val) { return val == 0u; })); 356 uint8_t* ArenaAllocator::AllocFromNewArena(size_t bytes) { argument 357 Arena* new_arena = pool_->AllocArena(std::max(Arena::kDefaultSize, bytes)); [all...] |
H A D | arena_allocator.h | 110 void RecordAlloc(size_t bytes ATTRIBUTE_UNUSED, ArenaAllocKind kind ATTRIBUTE_UNUSED) {} 126 void RecordAlloc(size_t bytes, ArenaAllocKind kind); 303 void* Alloc(size_t bytes, ArenaAllocKind kind = kArenaAllocMisc) ALWAYS_INLINE { 305 return AllocWithMemoryTool(bytes, kind); 307 bytes = RoundUp(bytes, kAlignment); 308 ArenaAllocatorStats::RecordAlloc(bytes, kind); 309 if (UNLIKELY(bytes > static_cast<size_t>(end_ - ptr_))) { 310 return AllocFromNewArena(bytes); 313 ptr_ += bytes; [all...] |
H A D | scoped_arena_allocator.cc | 83 // Update how many bytes we have allocated into the arena so that the arena pool knows how 93 void* ArenaStack::AllocWithMemoryTool(size_t bytes, ArenaAllocKind kind) { argument 97 size_t rounded_bytes = RoundUp(bytes + kMemoryToolRedZoneBytes, 8); 104 CurrentStats()->RecordAlloc(bytes, kind); 106 MEMORY_TOOL_MAKE_UNDEFINED(ptr, bytes);
|
H A D | allocator.h | 80 // Running count of number of bytes used for this kind of allocation. Increased by allocations, 84 // Largest value of bytes used seen. 87 // Total number of bytes allocated of this kind. 92 inline void RegisterAllocation(AllocatorTag tag, size_t bytes) { argument 93 g_total_bytes_used[tag].FetchAndAddSequentiallyConsistent(bytes); 94 size_t new_bytes = g_bytes_used[tag].FetchAndAddSequentiallyConsistent(bytes) + bytes; 100 inline void RegisterFree(AllocatorTag tag, size_t bytes) { argument 101 g_bytes_used[tag].FetchAndSubSequentiallyConsistent(bytes);
|
/art/runtime/gc/collector/ |
H A D | garbage_collector.h | 48 : objects(num_objects), bytes(num_bytes) {} 51 bytes += other.bytes; 55 // Freed bytes are signed since the GC can free negative bytes if it promotes objects to a space 57 int64_t bytes; member in struct:art::gc::collector::ObjectBytePair 77 return freed_.bytes; 80 return freed_los_.bytes; 168 // Returns the estimated throughput in bytes / second.
|
/art/runtime/gc/space/ |
H A D | region_space-inl.h | 157 uint64_t bytes = 0; local 166 bytes += r->BytesAllocated(); 170 bytes += r->BytesAllocated(); 175 bytes += r->BytesAllocated(); 180 bytes += r->BytesAllocated(); 187 return bytes; 192 uint64_t bytes = 0; local 201 bytes += r->ObjectsAllocated(); 205 bytes += r->ObjectsAllocated(); 210 bytes [all...] |
H A D | bump_pointer_space.cc | 140 uint8_t* BumpPointerSpace::AllocBlock(size_t bytes) { argument 141 bytes = RoundUp(bytes, kAlignment); 146 AllocNonvirtualWithoutAccounting(bytes + sizeof(BlockHeader))); 149 header->size_ = bytes; // Write out the block header. 255 bool BumpPointerSpace::AllocNewTlab(Thread* self, size_t bytes) { argument 258 uint8_t* start = AllocBlock(bytes); 262 self->SetTlab(start, start + bytes); 270 << max_contiguous_allocation << " bytes)";
|
H A D | bump_pointer_space.h | 138 bool AllocNewTlab(Thread* self, size_t bytes) REQUIRES(!block_lock_); 150 // Record objects / bytes freed. 151 void RecordFree(int32_t objects, int32_t bytes) { argument 153 bytes_allocated_.FetchAndSubSequentiallyConsistent(bytes); 165 // Allocate a raw block of bytes. 166 uint8_t* AllocBlock(size_t bytes) REQUIRES(block_lock_); 187 size_t size_; // Size of the block in bytes, does not include the header.
|
/art/runtime/arch/x86/ |
H A D | memcmp16_x86.S | 851 je L(8bytes) 853 je L(10bytes) 855 je L(12bytes) 856 jmp L(14bytes) 863 je L(16bytes) 865 je L(18bytes) 867 je L(20bytes) 868 jmp L(22bytes) 875 je L(24bytes) 877 je L(26bytes) [all...] |
/art/tools/ |
H A D | stream-trace-converter.py | 39 bytes = [ (val & 0xFF), ((val >> 8) & 0xFF) ] 40 asbytearray = bytearray(bytes) 59 bytes = [ (val & 0xFF), ((val >> 8) & 0xFF), ((val >> 16) & 0xFF), ((val >> 24) & 0xFF) ] 60 asbytearray = bytearray(bytes) 111 # Skip over offsetToData bytes
|
/art/test/003-omnibus-opcodes/src/ |
H A D | Array.java | 25 static void checkBytes(byte[] bytes) { argument 26 Main.assertTrue(bytes[0] == 0); 27 Main.assertTrue(bytes[1] == -1); 28 Main.assertTrue(bytes[2] == -2); 29 Main.assertTrue(bytes[3] == -3); 30 Main.assertTrue(bytes[4] == -4);
|
/art/runtime/interpreter/mterp/mips/ |
H A D | bincmp.S | 27 addu a2, rINST, rINST # convert to bytes
|
H A D | zcmp.S | 25 addu a1, rINST, rINST # convert to bytes
|
/art/test/093-serialization/src/ |
H A D | Main.java | 52 byte[] bytes = byteStream.toByteArray(); 56 return bytes;
|