/art/runtime/gc/allocator/ |
H A D | rosalloc-inl.h | 27 inline ALWAYS_INLINE void* RosAlloc::Alloc(Thread* self, size_t size, size_t* bytes_allocated) { argument 29 return AllocLargeObject(self, size, bytes_allocated); 33 m = AllocFromRun(self, size, bytes_allocated); 35 m = AllocFromRunThreadUnsafe(self, size, bytes_allocated);
|
H A D | dlmalloc.cc | 77 size_t* bytes_allocated = reinterpret_cast<size_t*>(arg); local 78 *bytes_allocated += used_bytes + sizeof(size_t);
|
H A D | rosalloc.cc | 453 void* RosAlloc::AllocLargeObject(Thread* self, size_t size, size_t* bytes_allocated) { argument 468 *bytes_allocated = total_bytes; 622 void* RosAlloc::AllocFromRunThreadUnsafe(Thread* self, size_t size, size_t* bytes_allocated) { argument 634 DCHECK(bytes_allocated != nullptr); 635 *bytes_allocated = bracket_size; 641 void* RosAlloc::AllocFromRun(Thread* self, size_t size, size_t* bytes_allocated) { argument 735 DCHECK(bytes_allocated != nullptr); 736 *bytes_allocated = bracket_size; 1860 size_t* bytes_allocated = reinterpret_cast<size_t*>(arg); local 1861 *bytes_allocated [all...] |
/art/runtime/gc/space/ |
H A D | dlmalloc_space-inl.h | 29 size_t* bytes_allocated, 34 obj = AllocWithoutGrowthLocked(self, num_bytes, bytes_allocated, usable_size); 53 size_t* bytes_allocated, 62 DCHECK(bytes_allocated != NULL); 63 *bytes_allocated = allocation_size; 28 AllocNonvirtual(Thread* self, size_t num_bytes, size_t* bytes_allocated, size_t* usable_size) argument 52 AllocWithoutGrowthLocked(Thread* , size_t num_bytes, size_t* bytes_allocated, size_t* usable_size) argument
|
H A D | bump_pointer_space-inl.h | 26 inline mirror::Object* BumpPointerSpace::Alloc(Thread*, size_t num_bytes, size_t* bytes_allocated, argument 31 *bytes_allocated = num_bytes; 40 size_t* bytes_allocated, 50 *bytes_allocated = num_bytes; 39 AllocThreadUnsafe(Thread* self, size_t num_bytes, size_t* bytes_allocated, size_t* usable_size) argument
|
H A D | rosalloc_space-inl.h | 51 size_t* bytes_allocated, size_t* usable_size) { 63 DCHECK(bytes_allocated != NULL); 64 *bytes_allocated = rosalloc_size; 50 AllocCommon(Thread* self, size_t num_bytes, size_t* bytes_allocated, size_t* usable_size) argument
|
H A D | rosalloc_space.h | 49 mirror::Object* AllocWithGrowth(Thread* self, size_t num_bytes, size_t* bytes_allocated, 51 mirror::Object* Alloc(Thread* self, size_t num_bytes, size_t* bytes_allocated, 53 return AllocNonvirtual(self, num_bytes, bytes_allocated, usable_size); 55 mirror::Object* AllocThreadUnsafe(Thread* self, size_t num_bytes, size_t* bytes_allocated, 58 return AllocNonvirtualThreadUnsafe(self, num_bytes, bytes_allocated, usable_size); 68 mirror::Object* AllocNonvirtual(Thread* self, size_t num_bytes, size_t* bytes_allocated, argument 71 return AllocCommon(self, num_bytes, bytes_allocated, usable_size); 74 size_t* bytes_allocated, size_t* usable_size) { 76 return AllocCommon<false>(self, num_bytes, bytes_allocated, usable_size); 134 mirror::Object* AllocCommon(Thread* self, size_t num_bytes, size_t* bytes_allocated, 73 AllocNonvirtualThreadUnsafe(Thread* self, size_t num_bytes, size_t* bytes_allocated, size_t* usable_size) argument [all...] |
H A D | large_object_space_test.cc | 86 size_t bytes_allocated = 0; local 88 mirror::Object* obj = los->Alloc(Thread::Current(), 100 * MB, &bytes_allocated, nullptr);
|
H A D | valgrind_malloc_space-inl.h | 34 size_t* bytes_allocated, 37 bytes_allocated, usable_size); 51 size_t* bytes_allocated, 53 void* obj_with_rdz = S::Alloc(self, num_bytes + 2 * kValgrindRedZoneBytes, bytes_allocated, 33 AllocWithGrowth(Thread* self, size_t num_bytes, size_t* bytes_allocated, size_t* usable_size) argument 50 Alloc(Thread* self, size_t num_bytes, size_t* bytes_allocated, size_t* usable_size) argument
|
H A D | zygote_space.cc | 78 mirror::Object* ZygoteSpace::Alloc(Thread* self, size_t num_bytes, size_t* bytes_allocated, argument
|
H A D | dlmalloc_space.cc | 129 size_t* bytes_allocated, size_t* usable_size) { 137 result = AllocWithoutGrowthLocked(self, num_bytes, bytes_allocated, usable_size); 279 size_t bytes_allocated = 0; local 280 mspace_inspect_all(mspace_, DlmallocBytesAllocatedCallback, &bytes_allocated); 281 return bytes_allocated; 128 AllocWithGrowth(Thread* self, size_t num_bytes, size_t* bytes_allocated, size_t* usable_size) argument
|
H A D | rosalloc_space.cc | 147 size_t* bytes_allocated, size_t* usable_size) { 155 result = AllocCommon(self, num_bytes, bytes_allocated, usable_size); 279 size_t bytes_allocated = 0; local 280 InspectAllRosAlloc(art::gc::allocator::RosAlloc::BytesAllocatedCallback, &bytes_allocated, false); 281 return bytes_allocated; 146 AllocWithGrowth(Thread* self, size_t num_bytes, size_t* bytes_allocated, size_t* usable_size) argument
|
H A D | space_test.h | 64 size_t* bytes_allocated, size_t* usable_size) 68 mirror::Object* obj = alloc_space->Alloc(self, bytes, bytes_allocated, usable_size); 76 size_t* bytes_allocated, size_t* usable_size) 80 mirror::Object* obj = alloc_space->AllocWithGrowth(self, bytes, bytes_allocated, usable_size); 424 size_t bytes_allocated = 0; local 426 object.Assign(Alloc(space, self, alloc_size, &bytes_allocated, nullptr)); 428 object.Assign(AllocWithGrowth(space, self, alloc_size, &bytes_allocated, nullptr)); 435 EXPECT_EQ(bytes_allocated, allocation_size); 517 size_t bytes_allocated = 0; local 519 large_object.Assign(Alloc(space, self, three_quarters_space, &bytes_allocated, nullpt [all...] |
H A D | large_object_space.cc | 40 virtual mirror::Object* Alloc(Thread* self, size_t num_bytes, size_t* bytes_allocated, 43 LargeObjectMapSpace::Alloc(self, num_bytes + kValgrindRedZoneBytes * 2, bytes_allocated, 111 size_t* bytes_allocated, size_t* usable_size) { 124 DCHECK(bytes_allocated != nullptr); 130 *bytes_allocated = allocation_size; 409 mirror::Object* FreeListSpace::Alloc(Thread* self, size_t num_bytes, size_t* bytes_allocated, argument 443 DCHECK(bytes_allocated != nullptr); 444 *bytes_allocated = allocation_size; 110 Alloc(Thread* self, size_t num_bytes, size_t* bytes_allocated, size_t* usable_size) argument
|
/art/runtime/gc/ |
H A D | heap-inl.h | 57 size_t bytes_allocated; local 74 bytes_allocated = byte_count; 75 usable_size = bytes_allocated; 79 obj = TryToAllocate<kInstrumented, false>(self, allocator, byte_count, &bytes_allocated, 83 obj = AllocateInternalWithGc(self, allocator, byte_count, &bytes_allocated, &usable_size, 96 DCHECK_GT(bytes_allocated, 0u); 119 static_cast<size_t>(num_bytes_allocated_.FetchAndAddSequentiallyConsistent(bytes_allocated)) 120 + bytes_allocated; 130 thread_stats->allocated_bytes += bytes_allocated; 133 global_stats->allocated_bytes += bytes_allocated; 183 TryToAllocate(Thread* self, AllocatorType allocator_type, size_t alloc_size, size_t* bytes_allocated, size_t* usable_size) argument [all...] |
H A D | heap.cc | 1229 size_t alloc_size, size_t* bytes_allocated, 1249 mirror::Object* ptr = TryToAllocate<true, false>(self, allocator, alloc_size, bytes_allocated, 1263 mirror::Object* ptr = TryToAllocate<true, false>(self, allocator, alloc_size, bytes_allocated, 1283 mirror::Object* ptr = TryToAllocate<true, false>(self, allocator, alloc_size, bytes_allocated, 1292 mirror::Object* ptr = TryToAllocate<true, true>(self, allocator, alloc_size, bytes_allocated, 1310 ptr = TryToAllocate<true, true>(self, allocator, alloc_size, bytes_allocated, usable_size); 1325 ptr = TryToAllocate<true, true>(self, allocator, alloc_size, bytes_allocated, 1369 ptr = TryToAllocate<true, true>(self, allocator, alloc_size, bytes_allocated, 1858 size_t bytes_allocated; local 1859 forward_address = to_space_->Alloc(self_, object_size, &bytes_allocated, nullpt 1228 AllocateInternalWithGc(Thread* self, AllocatorType allocator, size_t alloc_size, size_t* bytes_allocated, size_t* usable_size, mirror::Class** klass) argument 2865 const uint64_t bytes_allocated = GetBytesAllocated(); local [all...] |
/art/compiler/utils/ |
H A D | arena_allocator.cc | 96 const size_t bytes_allocated = BytesAllocated(); local 97 os << " MEM: used: " << bytes_allocated << ", allocated: " << malloc_bytes 102 << num_allocations << ", avg size: " << bytes_allocated / num_allocations << "\n";
|
/art/runtime/gc/collector/ |
H A D | semi_space.cc | 482 size_t bytes_allocated; local 488 forward_address = promo_dest_space_->AllocThreadUnsafe(self_, object_size, &bytes_allocated, 492 forward_address = to_space_->AllocThreadUnsafe(self_, object_size, &bytes_allocated, nullptr); 496 bytes_promoted_ += bytes_allocated; 537 forward_address = to_space_->AllocThreadUnsafe(self_, object_size, &bytes_allocated, nullptr); 544 forward_address = fallback_space_->AllocThreadUnsafe(self_, object_size, &bytes_allocated, 553 bytes_moved_ += bytes_allocated;
|