/art/test/102-concurrent-gc/src/ |
H A D | Main.java | 24 public byte[] bytes; field in class:Main.ByteContainer 37 l[index].bytes = new byte[bufferSize]; 57 byte[] temp = l[a].bytes; 58 l[a].bytes = l[b].bytes; 59 l[b].bytes = temp;
|
/art/runtime/jdwp/ |
H A D | jdwp_bits.h | 35 static inline void Append1BE(std::vector<uint8_t>& bytes, uint8_t value) { argument 36 bytes.push_back(value); 39 static inline void Append2BE(std::vector<uint8_t>& bytes, uint16_t value) { argument 40 bytes.push_back(static_cast<uint8_t>(value >> 8)); 41 bytes.push_back(static_cast<uint8_t>(value)); 44 static inline void Append4BE(std::vector<uint8_t>& bytes, uint32_t value) { argument 45 bytes.push_back(static_cast<uint8_t>(value >> 24)); 46 bytes.push_back(static_cast<uint8_t>(value >> 16)); 47 bytes.push_back(static_cast<uint8_t>(value >> 8)); 48 bytes 51 Append8BE(std::vector<uint8_t>& bytes, uint64_t value) argument 62 AppendUtf16BE(std::vector<uint8_t>& bytes, const uint16_t* chars, size_t char_count) argument [all...] |
H A D | jdwp_request.cc | 28 Request::Request(const uint8_t* bytes, uint32_t available) : p_(bytes) { argument 30 end_ = bytes + byte_count_; 48 CHECK(p_ == end_) << "read too few bytes: " << (end_ - p_); 50 CHECK(p_ == end_) << "read too many bytes: " << (p_ - end_);
|
/art/test/004-NativeAllocations/src/ |
H A D | Main.java | 29 private int bytes; field in class:Main.NativeAllocation 31 NativeAllocation(int bytes) throws Exception { argument 32 this.bytes = bytes; 33 register_native_allocation.invoke(runtime, bytes); 35 nativeBytes += bytes; 44 nativeBytes -= bytes; 46 register_native_free.invoke(runtime, bytes);
|
/art/runtime/native/ |
H A D | org_apache_harmony_dalvik_ddmc_DdmVmInternal.cc | 103 * (1b) bytes per entry 124 std::vector<uint8_t>& bytes = *reinterpret_cast<std::vector<uint8_t>*>(context); local 125 JDWP::Append4BE(bytes, t->GetThreadId()); 126 JDWP::Append1BE(bytes, Dbg::ToJdwpThreadStatus(t->GetState())); 127 JDWP::Append4BE(bytes, t->GetTid()); 128 JDWP::Append4BE(bytes, utime); 129 JDWP::Append4BE(bytes, stime); 130 JDWP::Append1BE(bytes, t->IsDaemon()); 134 std::vector<uint8_t> bytes; local 143 JDWP::Append1BE(bytes, kThstHeaderLe [all...] |
H A D | dalvik_system_VMRuntime.cc | 185 static void VMRuntime_registerNativeAllocation(JNIEnv* env, jobject, jint bytes) { argument 186 if (UNLIKELY(bytes < 0)) { 188 ThrowRuntimeException("allocation size negative %d", bytes); 191 Runtime::Current()->GetHeap()->RegisterNativeAllocation(env, static_cast<size_t>(bytes)); 194 static void VMRuntime_registerNativeFree(JNIEnv* env, jobject, jint bytes) { argument 195 if (UNLIKELY(bytes < 0)) { 197 ThrowRuntimeException("allocation size negative %d", bytes); 200 Runtime::Current()->GetHeap()->RegisterNativeFree(env, static_cast<size_t>(bytes));
|
/art/runtime/gc/allocator/ |
H A D | rosalloc-inl.h | 39 byte* bytes = reinterpret_cast<byte*>(m); local 41 DCHECK_EQ(bytes[i], 0);
|
H A D | rosalloc.h | 266 // The number of smaller size brackets that are 16 bytes apart. 268 // The sizes (the slot sizes, in bytes) of the size brackets. 274 // The header sizes in bytes of the runs for each size bracket. 412 // The footprint in bytes of the currently allocated portion of the 487 // Returns how many bytes were freed. 542 size_t UsableSize(size_t bytes) { 543 if (UNLIKELY(bytes > kLargeSizeThreshold)) { 544 return RoundUp(bytes, kPageSize); 546 return RoundToBracketSize(bytes); 563 void SetFootprintLimit(size_t bytes) LOCKS_EXCLUDE [all...] |
/art/test/407-arrays/src/ |
H A D | Main.java | 32 static void $opt$testReads(boolean[] bools, byte[] bytes, char[] chars, short[] shorts, argument 37 assertEquals(0, bytes[0]); 38 assertEquals(0, bytes[index]); 56 static void $opt$testWrites(boolean[] bools, byte[] bytes, char[] chars, short[] shorts, argument 63 bytes[0] = -4; 64 assertEquals(-4, bytes[0]); 65 bytes[index] = -8; 66 assertEquals(-8, bytes[index]);
|
/art/compiler/utils/ |
H A D | scoped_arena_allocator.h | 66 void* Alloc(size_t bytes, ArenaAllocKind kind) ALWAYS_INLINE { 68 return AllocValgrind(bytes, kind); 70 size_t rounded_bytes = RoundUp(bytes, 8); 75 CurrentStats()->RecordAlloc(bytes, kind); 83 void* AllocValgrind(size_t bytes, ArenaAllocKind kind); 118 void* Alloc(size_t bytes, ArenaAllocKind kind) ALWAYS_INLINE { 120 return arena_stack_->Alloc(bytes, kind);
|
H A D | arena_allocator.h | 74 void RecordAlloc(size_t bytes, ArenaAllocKind kind) { UNUSED(bytes); UNUSED(kind); } argument 90 void RecordAlloc(size_t bytes, ArenaAllocKind kind); 168 void* Alloc(size_t bytes, ArenaAllocKind kind) ALWAYS_INLINE { 170 return AllocValgrind(bytes, kind); 172 bytes = RoundUp(bytes, 8); 173 if (UNLIKELY(ptr_ + bytes > end_)) { 175 ObtainNewArenaForAllocation(bytes); variable 180 ArenaAllocatorStats::RecordAlloc(bytes, kin [all...] |
H A D | scoped_arena_allocator.cc | 84 // Update how many bytes we have allocated into the arena so that the arena pool knows how 94 void* ArenaStack::AllocValgrind(size_t bytes, ArenaAllocKind kind) { argument 95 size_t rounded_bytes = RoundUp(bytes + kValgrindRedZoneBytes, 8); 100 CurrentStats()->RecordAlloc(bytes, kind); 102 VALGRIND_MAKE_MEM_UNDEFINED(ptr, bytes); 103 VALGRIND_MAKE_MEM_NOACCESS(ptr + bytes, rounded_bytes - bytes);
|
H A D | arena_allocator.cc | 66 void ArenaAllocatorStatsImpl<kCount>::RecordAlloc(size_t bytes, ArenaAllocKind kind) { argument 67 alloc_stats_[kind] += bytes; 222 // Update how many bytes we have allocated into the arena so that the arena pool knows how 228 void* ArenaAllocator::AllocValgrind(size_t bytes, ArenaAllocKind kind) { argument 229 size_t rounded_bytes = RoundUp(bytes + kValgrindRedZoneBytes, 8); 244 VALGRIND_MAKE_MEM_NOACCESS(ret + bytes, rounded_bytes - bytes);
|
/art/runtime/arch/x86_64/ |
H A D | memcmp16_x86_64.S | 701 jnc L(16bytes) 706 jnc L(16bytes) 711 jnc L(16bytes) 715 jmp L(16bytes) 720 jmp L(16bytes) 724 jmp L(16bytes) 728 jmp L(16bytes) 732 jmp L(16bytes) 736 jmp L(16bytes) 740 jmp L(16bytes) [all...] |
/art/runtime/gc/collector/ |
H A D | garbage_collector.h | 38 : objects(num_objects), bytes(num_bytes) {} 41 bytes += other.bytes; 45 // Freed bytes are signed since the GC can free negative bytes if it promotes objects to a space 47 int64_t bytes; member in struct:art::gc::collector::ObjectBytePair 67 return freed_.bytes; 70 return freed_los_.bytes; 161 // Returns the estimated throughput in bytes / second.
|
H A D | garbage_collector.cc | 51 return (static_cast<uint64_t>(freed_.bytes) * 1000) / (NsToMs(GetDurationNs()) + 1); 83 // Update cumulative statistics with how many bytes the GC iteration freed. 167 heap_->RecordFree(freed.objects, freed.bytes); 171 heap_->RecordFree(freed.objects, freed.bytes);
|
/art/runtime/base/ |
H A D | allocator.h | 77 static void RegisterAllocation(AllocatorTag tag, uint64_t bytes) { argument 78 total_bytes_used_[tag].FetchAndAddSequentiallyConsistent(bytes); 79 uint64_t new_bytes = bytes_used_[tag].FetchAndAddSequentiallyConsistent(bytes) + bytes; 82 static void RegisterFree(AllocatorTag tag, uint64_t bytes) { argument 83 bytes_used_[tag].FetchAndSubSequentiallyConsistent(bytes);
|
/art/runtime/arch/x86/ |
H A D | memcmp16_x86.S | 851 je L(8bytes) 853 je L(10bytes) 855 je L(12bytes) 856 jmp L(14bytes) 863 je L(16bytes) 865 je L(18bytes) 867 je L(20bytes) 868 jmp L(22bytes) 875 je L(24bytes) 877 je L(26bytes) [all...] |
/art/runtime/gc/space/ |
H A D | bump_pointer_space.h | 136 bool AllocNewTlab(Thread* self, size_t bytes); 148 // Record objects / bytes freed. 149 void RecordFree(int32_t objects, int32_t bytes) { argument 151 bytes_allocated_.FetchAndSubSequentiallyConsistent(bytes); 163 // Allocate a raw block of bytes. 164 byte* AllocBlock(size_t bytes) EXCLUSIVE_LOCKS_REQUIRED(block_lock_); 185 size_t size_; // Size of the block in bytes, does not include the header.
|
H A D | bump_pointer_space.cc | 137 byte* BumpPointerSpace::AllocBlock(size_t bytes) { argument 138 bytes = RoundUp(bytes, kAlignment); 143 AllocNonvirtualWithoutAccounting(bytes + sizeof(BlockHeader))); 146 header->size_ = bytes; // Write out the block header. 250 bool BumpPointerSpace::AllocNewTlab(Thread* self, size_t bytes) { argument 253 byte* start = AllocBlock(bytes); 257 self->SetTlab(start, start + bytes); 265 << max_contiguous_allocation << " bytes)";
|
/art/test/003-omnibus-opcodes/src/ |
H A D | Array.java | 25 static void checkBytes(byte[] bytes) { argument 26 Main.assertTrue(bytes[0] == 0); 27 Main.assertTrue(bytes[1] == -1); 28 Main.assertTrue(bytes[2] == -2); 29 Main.assertTrue(bytes[3] == -3); 30 Main.assertTrue(bytes[4] == -4);
|
/art/test/093-serialization/src/ |
H A D | Main.java | 52 byte[] bytes = byteStream.toByteArray(); 56 return bytes;
|
/art/runtime/ |
H A D | dex_file_verifier_test.cc | 65 // the final = symbols are read and used to trim the remaining bytes 169 static const DexFile* FixChecksumAndOpen(byte* bytes, size_t length, const char* location, argument 172 CHECK(bytes != nullptr); 175 FixUpChecksum(bytes); 180 if (!file->WriteFully(bytes, length)) {
|
H A D | thread-inl.h | 169 inline mirror::Object* Thread::AllocTlab(size_t bytes) { argument 170 DCHECK_GE(TlabSize(), bytes); local 173 tlsPtr_.thread_local_pos += bytes;
|
H A D | debugger.cc | 3993 std::vector<uint8_t> bytes; local 3994 JDWP::Append4BE(bytes, t->GetThreadId()); 3995 JDWP::AppendUtf16BE(bytes, chars, char_count); 3996 CHECK_EQ(bytes.size(), char_count*2 + sizeof(uint32_t)*2); 3997 Dbg::DdmSendChunk(type, bytes); 4048 void Dbg::DdmSendChunk(uint32_t type, const std::vector<uint8_t>& bytes) { argument 4049 DdmSendChunk(type, bytes.size(), &bytes[0]); 4116 * [u4]: max heap size in bytes (-Xmx) 4117 * [u4]: current heap size in bytes 4123 std::vector<uint8_t> bytes; local 4742 std::vector<uint8_t> bytes; local [all...] |