/art/runtime/gc/space/ |
H A D | dlmalloc_space-inl.h | 29 size_t* bytes_allocated, 35 obj = AllocWithoutGrowthLocked(self, num_bytes, bytes_allocated, usable_size, 56 size_t* bytes_allocated, 66 DCHECK(bytes_allocated != nullptr); 67 *bytes_allocated = allocation_size; 28 AllocNonvirtual(Thread* self, size_t num_bytes, size_t* bytes_allocated, size_t* usable_size, size_t* bytes_tl_bulk_allocated) argument 54 AllocWithoutGrowthLocked( Thread* , size_t num_bytes, size_t* bytes_allocated, size_t* usable_size, size_t* bytes_tl_bulk_allocated) argument
|
H A D | rosalloc_space-inl.h | 64 size_t* bytes_allocated, size_t* usable_size, 81 DCHECK(bytes_allocated != nullptr); 82 *bytes_allocated = rosalloc_bytes_allocated; 98 size_t* bytes_allocated) { 99 DCHECK(bytes_allocated != nullptr); 101 rosalloc_->AllocFromThreadLocalRun(self, num_bytes, bytes_allocated)); 63 AllocCommon(Thread* self, size_t num_bytes, size_t* bytes_allocated, size_t* usable_size, size_t* bytes_tl_bulk_allocated) argument 97 AllocThreadLocal(Thread* self, size_t num_bytes, size_t* bytes_allocated) argument
|
H A D | region_space-inl.h | 26 inline mirror::Object* RegionSpace::Alloc(Thread*, size_t num_bytes, size_t* bytes_allocated, argument 30 return AllocNonvirtual<false>(num_bytes, bytes_allocated, usable_size, 35 size_t* bytes_allocated, 39 return Alloc(self, num_bytes, bytes_allocated, usable_size, bytes_tl_bulk_allocated); 43 inline mirror::Object* RegionSpace::AllocNonvirtual(size_t num_bytes, size_t* bytes_allocated, argument 51 obj = current_region_->Alloc(num_bytes, bytes_allocated, usable_size, 55 obj = evac_region_->Alloc(num_bytes, bytes_allocated, usable_size, 64 obj = current_region_->Alloc(num_bytes, bytes_allocated, usable_size, 67 obj = evac_region_->Alloc(num_bytes, bytes_allocated, usable_size, 84 obj = r->Alloc(num_bytes, bytes_allocated, usable_siz 34 AllocThreadUnsafe(Thread* self, size_t num_bytes, size_t* bytes_allocated, size_t* usable_size, size_t* bytes_tl_bulk_allocated) argument 114 Alloc(size_t num_bytes, size_t* bytes_allocated, size_t* usable_size, size_t* bytes_tl_bulk_allocated) argument 266 AllocLarge(size_t num_bytes, size_t* bytes_allocated, size_t* usable_size, size_t* bytes_tl_bulk_allocated) argument [all...] |
H A D | valgrind_malloc_space.h | 36 mirror::Object* AllocWithGrowth(Thread* self, size_t num_bytes, size_t* bytes_allocated, 39 mirror::Object* Alloc(Thread* self, size_t num_bytes, size_t* bytes_allocated, 41 mirror::Object* AllocThreadUnsafe(Thread* self, size_t num_bytes, size_t* bytes_allocated,
|
H A D | rosalloc_space.h | 49 mirror::Object* AllocWithGrowth(Thread* self, size_t num_bytes, size_t* bytes_allocated, 52 mirror::Object* Alloc(Thread* self, size_t num_bytes, size_t* bytes_allocated, 54 return AllocNonvirtual(self, num_bytes, bytes_allocated, usable_size, 57 mirror::Object* AllocThreadUnsafe(Thread* self, size_t num_bytes, size_t* bytes_allocated, 60 return AllocNonvirtualThreadUnsafe(self, num_bytes, bytes_allocated, usable_size, 71 mirror::Object* AllocNonvirtual(Thread* self, size_t num_bytes, size_t* bytes_allocated, argument 74 return AllocCommon(self, num_bytes, bytes_allocated, usable_size, 78 size_t* bytes_allocated, size_t* usable_size, 81 return AllocCommon<false>(self, num_bytes, bytes_allocated, usable_size, 91 size_t* bytes_allocated); 77 AllocNonvirtualThreadUnsafe(Thread* self, size_t num_bytes, size_t* bytes_allocated, size_t* usable_size, size_t* bytes_tl_bulk_allocated) argument [all...] |
H A D | valgrind_malloc_space-inl.h | 34 size_t bytes_allocated, size_t usable_size, 39 *bytes_allocated_out = bytes_allocated; 65 // Right redzone. Assumes that if bytes_allocated > usable_size, then the difference is 92 size_t bytes_allocated; local 96 &bytes_allocated, &usable_size, 104 bytes_allocated, usable_size, 121 size_t bytes_allocated; local 125 &bytes_allocated, &usable_size, &bytes_tl_bulk_allocated); 132 bytes_allocated, usable_size, 149 size_t bytes_allocated; local 33 AdjustForValgrind(void* obj_with_rdz, size_t num_bytes, size_t bytes_allocated, size_t usable_size, size_t bytes_tl_bulk_allocated, size_t* bytes_allocated_out, size_t* usable_size_out, size_t* bytes_tl_bulk_allocated_out) argument [all...] |
H A D | bump_pointer_space-inl.h | 27 inline mirror::Object* BumpPointerSpace::Alloc(Thread*, size_t num_bytes, size_t* bytes_allocated, argument 33 *bytes_allocated = num_bytes; 43 size_t* bytes_allocated, 54 *bytes_allocated = num_bytes; 42 AllocThreadUnsafe(Thread* self, size_t num_bytes, size_t* bytes_allocated, size_t* usable_size, size_t* bytes_tl_bulk_allocated) argument
|
H A D | dlmalloc_space.h | 50 virtual mirror::Object* AllocWithGrowth(Thread* self, size_t num_bytes, size_t* bytes_allocated, 55 virtual mirror::Object* Alloc(Thread* self, size_t num_bytes, size_t* bytes_allocated, 58 return AllocNonvirtual(self, num_bytes, bytes_allocated, usable_size, 87 mirror::Object* AllocNonvirtual(Thread* self, size_t num_bytes, size_t* bytes_allocated, 147 mirror::Object* AllocWithoutGrowthLocked(Thread* self, size_t num_bytes, size_t* bytes_allocated,
|
H A D | region_space.h | 44 mirror::Object* Alloc(Thread* self, size_t num_bytes, size_t* bytes_allocated, 47 mirror::Object* AllocThreadUnsafe(Thread* self, size_t num_bytes, size_t* bytes_allocated, 52 ALWAYS_INLINE mirror::Object* AllocNonvirtual(size_t num_bytes, size_t* bytes_allocated, 57 mirror::Object* AllocLarge(size_t num_bytes, size_t* bytes_allocated, size_t* usable_size, 59 void FreeLarge(mirror::Object* large_obj, size_t bytes_allocated); 273 ALWAYS_INLINE mirror::Object* Alloc(size_t num_bytes, size_t* bytes_allocated, 392 size_t bytes_allocated = RoundUp(BytesAllocated(), kRegionSize); local 393 DCHECK_GE(bytes_allocated, 0U); 394 uint result = (live_bytes_ * 100U) / bytes_allocated;
|
H A D | bump_pointer_space.h | 49 mirror::Object* Alloc(Thread* self, size_t num_bytes, size_t* bytes_allocated, 52 mirror::Object* AllocThreadUnsafe(Thread* self, size_t num_bytes, size_t* bytes_allocated,
|
H A D | malloc_space.h | 58 size_t* bytes_allocated, size_t* usable_size, 61 virtual mirror::Object* Alloc(Thread* self, size_t num_bytes, size_t* bytes_allocated,
|
H A D | zygote_space.h | 48 mirror::Object* Alloc(Thread* self, size_t num_bytes, size_t* bytes_allocated,
|
H A D | dlmalloc_space.cc | 127 size_t* bytes_allocated, size_t* usable_size, 136 result = AllocWithoutGrowthLocked(self, num_bytes, bytes_allocated, usable_size, 264 size_t bytes_allocated = 0; local 265 mspace_inspect_all(mspace_, DlmallocBytesAllocatedCallback, &bytes_allocated); 266 return bytes_allocated; 126 AllocWithGrowth(Thread* self, size_t num_bytes, size_t* bytes_allocated, size_t* usable_size, size_t* bytes_tl_bulk_allocated) argument
|
H A D | rosalloc_space.cc | 158 size_t* bytes_allocated, size_t* usable_size, 167 result = AllocCommon(self, num_bytes, bytes_allocated, usable_size, 290 size_t bytes_allocated = 0; local 291 InspectAllRosAlloc(art::gc::allocator::RosAlloc::BytesAllocatedCallback, &bytes_allocated, false); 292 return bytes_allocated; 157 AllocWithGrowth(Thread* self, size_t num_bytes, size_t* bytes_allocated, size_t* usable_size, size_t* bytes_tl_bulk_allocated) argument
|
H A D | space_test.h | 66 size_t* bytes_allocated, size_t* usable_size, 71 mirror::Object* obj = alloc_space->Alloc(self, bytes, bytes_allocated, usable_size, 80 size_t* bytes_allocated, size_t* usable_size, 85 mirror::Object* obj = alloc_space->AllocWithGrowth(self, bytes, bytes_allocated, usable_size, 454 size_t bytes_allocated = 0; local 457 object.Assign(Alloc(space, self, alloc_size, &bytes_allocated, nullptr, 460 object.Assign(AllocWithGrowth(space, self, alloc_size, &bytes_allocated, nullptr, 468 EXPECT_EQ(bytes_allocated, allocation_size); 552 size_t bytes_allocated = 0; local 555 large_object.Assign(Alloc(space, self, three_quarters_space, &bytes_allocated, nullpt [all...] |
H A D | space.h | 203 // succeeds, the output parameter bytes_allocated will be set to the 209 // 1) equal to bytes_allocated if it's not a thread local allocation, 210 // 2) greater than bytes_allocated if it's a thread local 215 virtual mirror::Object* Alloc(Thread* self, size_t num_bytes, size_t* bytes_allocated, 219 virtual mirror::Object* AllocThreadUnsafe(Thread* self, size_t num_bytes, size_t* bytes_allocated, 223 return Alloc(self, num_bytes, bytes_allocated, usable_size, bytes_tl_bulk_allocated);
|
H A D | large_object_space.cc | 49 mirror::Object* Alloc(Thread* self, size_t num_bytes, size_t* bytes_allocated, 53 LargeObjectMapSpace::Alloc(self, num_bytes + kValgrindRedZoneBytes * 2, bytes_allocated, 130 size_t* bytes_allocated, size_t* usable_size, 153 DCHECK(bytes_allocated != nullptr); 159 *bytes_allocated = allocation_size; 474 mirror::Object* FreeListSpace::Alloc(Thread* self, size_t num_bytes, size_t* bytes_allocated, argument 508 DCHECK(bytes_allocated != nullptr); 509 *bytes_allocated = allocation_size; 129 Alloc(Thread* self, size_t num_bytes, size_t* bytes_allocated, size_t* usable_size, size_t* bytes_tl_bulk_allocated) argument
|
H A D | large_object_space.h | 134 mirror::Object* Alloc(Thread* self, size_t num_bytes, size_t* bytes_allocated, 167 mirror::Object* Alloc(Thread* self, size_t num_bytes, size_t* bytes_allocated,
|
H A D | large_object_space_test.cc | 103 size_t bytes_allocated = 0, bytes_tl_bulk_allocated; local 105 mirror::Object* obj = los->Alloc(self, 100 * MB, &bytes_allocated, nullptr,
|
/art/runtime/gc/ |
H A D | heap-inl.h | 69 size_t bytes_allocated; local 87 bytes_allocated = byte_count; 88 usable_size = bytes_allocated; 92 (obj = rosalloc_space_->AllocThreadLocal(self, byte_count, &bytes_allocated)) && 102 usable_size = bytes_allocated; 108 obj = TryToAllocate<kInstrumented, false>(self, allocator, byte_count, &bytes_allocated, 112 obj = AllocateInternalWithGc(self, allocator, byte_count, &bytes_allocated, &usable_size, 125 DCHECK_GT(bytes_allocated, 0u); 159 thread_stats->allocated_bytes += bytes_allocated; 162 global_stats->allocated_bytes += bytes_allocated; 222 TryToAllocate(Thread* self, AllocatorType allocator_type, size_t alloc_size, size_t* bytes_allocated, size_t* usable_size, size_t* bytes_tl_bulk_allocated) argument [all...] |
/art/runtime/gc/allocator/ |
H A D | rosalloc-inl.h | 31 inline ALWAYS_INLINE void* RosAlloc::Alloc(Thread* self, size_t size, size_t* bytes_allocated, argument 35 return AllocLargeObject(self, size, bytes_allocated, usable_size, 40 m = AllocFromRun(self, size, bytes_allocated, usable_size, bytes_tl_bulk_allocated); 42 m = AllocFromRunThreadUnsafe(self, size, bytes_allocated, usable_size, 90 size_t* bytes_allocated) { 91 DCHECK(bytes_allocated != nullptr); 108 *bytes_allocated = bracket_size; 89 AllocFromThreadLocalRun(Thread* self, size_t size, size_t* bytes_allocated) argument
|
H A D | dlmalloc.cc | 85 size_t* bytes_allocated = reinterpret_cast<size_t*>(arg); local 86 *bytes_allocated += used_bytes + sizeof(size_t);
|
/art/runtime/gc/accounting/ |
H A D | mod_union_table_test.cc | 52 size_t bytes_allocated = 0, bytes_tl_bulk_allocated; local 54 space->Alloc(self, size, &bytes_allocated, nullptr, &bytes_tl_bulk_allocated)); 59 EXPECT_GE(bytes_allocated, size); 81 size_t bytes_allocated = 0, bytes_tl_bulk_allocated; local 82 auto* klass = down_cast<mirror::Class*>(space->Alloc(self, class_size, &bytes_allocated,
|
/art/test/099-vmdebug/src/ |
H A D | Main.java | 166 String bytes_allocated = VMDebug.getRuntimeStat("art.gc.bytes-allocated"); 175 checkNumber(bytes_allocated); 193 String bytes_allocated = map.get("art.gc.bytes-allocated"); 202 checkNumber(bytes_allocated);
|
/art/runtime/base/ |
H A D | arena_allocator.cc | 106 const size_t bytes_allocated = BytesAllocated(); local 107 os << " MEM: used: " << bytes_allocated << ", allocated: " << malloc_bytes 112 << num_allocations << ", avg size: " << bytes_allocated / num_allocations << "\n";
|