Searched refs:bytes (Results 1 - 25 of 44) sorted by relevance

12

/art/test/102-concurrent-gc/src/
H A DMain.java24 public byte[] bytes; field in class:Main.ByteContainer
37 l[index].bytes = new byte[bufferSize];
57 byte[] temp = l[a].bytes;
58 l[a].bytes = l[b].bytes;
59 l[b].bytes = temp;
/art/runtime/jdwp/
H A Djdwp_bits.h35 static inline void Append1BE(std::vector<uint8_t>& bytes, uint8_t value) { argument
36 bytes.push_back(value);
39 static inline void Append2BE(std::vector<uint8_t>& bytes, uint16_t value) { argument
40 bytes.push_back(static_cast<uint8_t>(value >> 8));
41 bytes.push_back(static_cast<uint8_t>(value));
44 static inline void Append4BE(std::vector<uint8_t>& bytes, uint32_t value) { argument
45 bytes.push_back(static_cast<uint8_t>(value >> 24));
46 bytes.push_back(static_cast<uint8_t>(value >> 16));
47 bytes.push_back(static_cast<uint8_t>(value >> 8));
48 bytes
51 Append8BE(std::vector<uint8_t>& bytes, uint64_t value) argument
62 AppendUtf16BE(std::vector<uint8_t>& bytes, const uint16_t* chars, size_t char_count) argument
[all...]
H A Djdwp_request.cc28 Request::Request(const uint8_t* bytes, uint32_t available) : p_(bytes) { argument
30 end_ = bytes + byte_count_;
48 CHECK(p_ == end_) << "read too few bytes: " << (end_ - p_);
50 CHECK(p_ == end_) << "read too many bytes: " << (p_ - end_);
/art/test/004-NativeAllocations/src/
H A DMain.java29 private int bytes; field in class:Main.NativeAllocation
31 NativeAllocation(int bytes) throws Exception { argument
32 this.bytes = bytes;
33 register_native_allocation.invoke(runtime, bytes);
35 nativeBytes += bytes;
44 nativeBytes -= bytes;
46 register_native_free.invoke(runtime, bytes);
/art/runtime/native/
H A Dorg_apache_harmony_dalvik_ddmc_DdmVmInternal.cc103 * (1b) bytes per entry
124 std::vector<uint8_t>& bytes = *reinterpret_cast<std::vector<uint8_t>*>(context); local
125 JDWP::Append4BE(bytes, t->GetThreadId());
126 JDWP::Append1BE(bytes, Dbg::ToJdwpThreadStatus(t->GetState()));
127 JDWP::Append4BE(bytes, t->GetTid());
128 JDWP::Append4BE(bytes, utime);
129 JDWP::Append4BE(bytes, stime);
130 JDWP::Append1BE(bytes, t->IsDaemon());
134 std::vector<uint8_t> bytes; local
143 JDWP::Append1BE(bytes, kThstHeaderLe
[all...]
H A Ddalvik_system_VMRuntime.cc185 static void VMRuntime_registerNativeAllocation(JNIEnv* env, jobject, jint bytes) { argument
186 if (UNLIKELY(bytes < 0)) {
188 ThrowRuntimeException("allocation size negative %d", bytes);
191 Runtime::Current()->GetHeap()->RegisterNativeAllocation(env, static_cast<size_t>(bytes));
194 static void VMRuntime_registerNativeFree(JNIEnv* env, jobject, jint bytes) { argument
195 if (UNLIKELY(bytes < 0)) {
197 ThrowRuntimeException("allocation size negative %d", bytes);
200 Runtime::Current()->GetHeap()->RegisterNativeFree(env, static_cast<size_t>(bytes));
/art/runtime/gc/allocator/
H A Drosalloc-inl.h39 byte* bytes = reinterpret_cast<byte*>(m); local
41 DCHECK_EQ(bytes[i], 0);
H A Drosalloc.h266 // The number of smaller size brackets that are 16 bytes apart.
268 // The sizes (the slot sizes, in bytes) of the size brackets.
274 // The header sizes in bytes of the runs for each size bracket.
412 // The footprint in bytes of the currently allocated portion of the
487 // Returns how many bytes were freed.
542 size_t UsableSize(size_t bytes) {
543 if (UNLIKELY(bytes > kLargeSizeThreshold)) {
544 return RoundUp(bytes, kPageSize);
546 return RoundToBracketSize(bytes);
563 void SetFootprintLimit(size_t bytes) LOCKS_EXCLUDE
[all...]
/art/test/407-arrays/src/
H A DMain.java32 static void $opt$testReads(boolean[] bools, byte[] bytes, char[] chars, short[] shorts, argument
37 assertEquals(0, bytes[0]);
38 assertEquals(0, bytes[index]);
56 static void $opt$testWrites(boolean[] bools, byte[] bytes, char[] chars, short[] shorts, argument
63 bytes[0] = -4;
64 assertEquals(-4, bytes[0]);
65 bytes[index] = -8;
66 assertEquals(-8, bytes[index]);
/art/compiler/utils/
H A Dscoped_arena_allocator.h66 void* Alloc(size_t bytes, ArenaAllocKind kind) ALWAYS_INLINE {
68 return AllocValgrind(bytes, kind);
70 size_t rounded_bytes = RoundUp(bytes, 8);
75 CurrentStats()->RecordAlloc(bytes, kind);
83 void* AllocValgrind(size_t bytes, ArenaAllocKind kind);
118 void* Alloc(size_t bytes, ArenaAllocKind kind) ALWAYS_INLINE {
120 return arena_stack_->Alloc(bytes, kind);
H A Darena_allocator.h74 void RecordAlloc(size_t bytes, ArenaAllocKind kind) { UNUSED(bytes); UNUSED(kind); } argument
90 void RecordAlloc(size_t bytes, ArenaAllocKind kind);
168 void* Alloc(size_t bytes, ArenaAllocKind kind) ALWAYS_INLINE {
170 return AllocValgrind(bytes, kind);
172 bytes = RoundUp(bytes, 8);
173 if (UNLIKELY(ptr_ + bytes > end_)) {
175 ObtainNewArenaForAllocation(bytes); variable
180 ArenaAllocatorStats::RecordAlloc(bytes, kin
[all...]
H A Dscoped_arena_allocator.cc84 // Update how many bytes we have allocated into the arena so that the arena pool knows how
94 void* ArenaStack::AllocValgrind(size_t bytes, ArenaAllocKind kind) { argument
95 size_t rounded_bytes = RoundUp(bytes + kValgrindRedZoneBytes, 8);
100 CurrentStats()->RecordAlloc(bytes, kind);
102 VALGRIND_MAKE_MEM_UNDEFINED(ptr, bytes);
103 VALGRIND_MAKE_MEM_NOACCESS(ptr + bytes, rounded_bytes - bytes);
H A Darena_allocator.cc66 void ArenaAllocatorStatsImpl<kCount>::RecordAlloc(size_t bytes, ArenaAllocKind kind) { argument
67 alloc_stats_[kind] += bytes;
222 // Update how many bytes we have allocated into the arena so that the arena pool knows how
228 void* ArenaAllocator::AllocValgrind(size_t bytes, ArenaAllocKind kind) { argument
229 size_t rounded_bytes = RoundUp(bytes + kValgrindRedZoneBytes, 8);
244 VALGRIND_MAKE_MEM_NOACCESS(ret + bytes, rounded_bytes - bytes);
/art/runtime/arch/x86_64/
H A Dmemcmp16_x86_64.S701 jnc L(16bytes)
706 jnc L(16bytes)
711 jnc L(16bytes)
715 jmp L(16bytes)
720 jmp L(16bytes)
724 jmp L(16bytes)
728 jmp L(16bytes)
732 jmp L(16bytes)
736 jmp L(16bytes)
740 jmp L(16bytes)
[all...]
/art/runtime/gc/collector/
H A Dgarbage_collector.h38 : objects(num_objects), bytes(num_bytes) {}
41 bytes += other.bytes;
45 // Freed bytes are signed since the GC can free negative bytes if it promotes objects to a space
47 int64_t bytes; member in struct:art::gc::collector::ObjectBytePair
67 return freed_.bytes;
70 return freed_los_.bytes;
161 // Returns the estimated throughput in bytes / second.
H A Dgarbage_collector.cc51 return (static_cast<uint64_t>(freed_.bytes) * 1000) / (NsToMs(GetDurationNs()) + 1);
83 // Update cumulative statistics with how many bytes the GC iteration freed.
167 heap_->RecordFree(freed.objects, freed.bytes);
171 heap_->RecordFree(freed.objects, freed.bytes);
/art/runtime/base/
H A Dallocator.h77 static void RegisterAllocation(AllocatorTag tag, uint64_t bytes) { argument
78 total_bytes_used_[tag].FetchAndAddSequentiallyConsistent(bytes);
79 uint64_t new_bytes = bytes_used_[tag].FetchAndAddSequentiallyConsistent(bytes) + bytes;
82 static void RegisterFree(AllocatorTag tag, uint64_t bytes) { argument
83 bytes_used_[tag].FetchAndSubSequentiallyConsistent(bytes);
/art/runtime/arch/x86/
H A Dmemcmp16_x86.S851 je L(8bytes)
853 je L(10bytes)
855 je L(12bytes)
856 jmp L(14bytes)
863 je L(16bytes)
865 je L(18bytes)
867 je L(20bytes)
868 jmp L(22bytes)
875 je L(24bytes)
877 je L(26bytes)
[all...]
/art/runtime/gc/space/
H A Dbump_pointer_space.h136 bool AllocNewTlab(Thread* self, size_t bytes);
148 // Record objects / bytes freed.
149 void RecordFree(int32_t objects, int32_t bytes) { argument
151 bytes_allocated_.FetchAndSubSequentiallyConsistent(bytes);
163 // Allocate a raw block of bytes.
164 byte* AllocBlock(size_t bytes) EXCLUSIVE_LOCKS_REQUIRED(block_lock_);
185 size_t size_; // Size of the block in bytes, does not include the header.
H A Dbump_pointer_space.cc137 byte* BumpPointerSpace::AllocBlock(size_t bytes) { argument
138 bytes = RoundUp(bytes, kAlignment);
143 AllocNonvirtualWithoutAccounting(bytes + sizeof(BlockHeader)));
146 header->size_ = bytes; // Write out the block header.
250 bool BumpPointerSpace::AllocNewTlab(Thread* self, size_t bytes) { argument
253 byte* start = AllocBlock(bytes);
257 self->SetTlab(start, start + bytes);
265 << max_contiguous_allocation << " bytes)";
/art/test/003-omnibus-opcodes/src/
H A DArray.java25 static void checkBytes(byte[] bytes) { argument
26 Main.assertTrue(bytes[0] == 0);
27 Main.assertTrue(bytes[1] == -1);
28 Main.assertTrue(bytes[2] == -2);
29 Main.assertTrue(bytes[3] == -3);
30 Main.assertTrue(bytes[4] == -4);
/art/test/093-serialization/src/
H A DMain.java52 byte[] bytes = byteStream.toByteArray();
56 return bytes;
/art/runtime/
H A Ddex_file_verifier_test.cc65 // the final = symbols are read and used to trim the remaining bytes
169 static const DexFile* FixChecksumAndOpen(byte* bytes, size_t length, const char* location, argument
172 CHECK(bytes != nullptr);
175 FixUpChecksum(bytes);
180 if (!file->WriteFully(bytes, length)) {
H A Dthread-inl.h169 inline mirror::Object* Thread::AllocTlab(size_t bytes) { argument
170 DCHECK_GE(TlabSize(), bytes); local
173 tlsPtr_.thread_local_pos += bytes;
H A Ddebugger.cc3993 std::vector<uint8_t> bytes; local
3994 JDWP::Append4BE(bytes, t->GetThreadId());
3995 JDWP::AppendUtf16BE(bytes, chars, char_count);
3996 CHECK_EQ(bytes.size(), char_count*2 + sizeof(uint32_t)*2);
3997 Dbg::DdmSendChunk(type, bytes);
4048 void Dbg::DdmSendChunk(uint32_t type, const std::vector<uint8_t>& bytes) { argument
4049 DdmSendChunk(type, bytes.size(), &bytes[0]);
4116 * [u4]: max heap size in bytes (-Xmx)
4117 * [u4]: current heap size in bytes
4123 std::vector<uint8_t> bytes; local
4742 std::vector<uint8_t> bytes; local
[all...]

Completed in 3204 milliseconds

12