/art/runtime/gc/space/ |
H A D | bump_pointer_space-inl.h | 27 inline mirror::Object* BumpPointerSpace::Alloc(Thread*, size_t num_bytes, size_t* bytes_allocated, argument 33 *bytes_allocated = num_bytes; 43 size_t* bytes_allocated, 54 *bytes_allocated = num_bytes; 42 AllocThreadUnsafe(Thread* self, size_t num_bytes, size_t* bytes_allocated, size_t* usable_size, size_t* bytes_tl_bulk_allocated) argument
|
H A D | dlmalloc_space-inl.h | 29 size_t* bytes_allocated, 35 obj = AllocWithoutGrowthLocked(self, num_bytes, bytes_allocated, usable_size, 56 size_t* bytes_allocated, 66 DCHECK(bytes_allocated != nullptr); 67 *bytes_allocated = allocation_size; 28 AllocNonvirtual(Thread* self, size_t num_bytes, size_t* bytes_allocated, size_t* usable_size, size_t* bytes_tl_bulk_allocated) argument 54 AllocWithoutGrowthLocked( Thread* , size_t num_bytes, size_t* bytes_allocated, size_t* usable_size, size_t* bytes_tl_bulk_allocated) argument
|
H A D | rosalloc_space.h | 49 mirror::Object* AllocWithGrowth(Thread* self, size_t num_bytes, size_t* bytes_allocated, 52 mirror::Object* Alloc(Thread* self, size_t num_bytes, size_t* bytes_allocated, 54 return AllocNonvirtual(self, num_bytes, bytes_allocated, usable_size, 57 mirror::Object* AllocThreadUnsafe(Thread* self, size_t num_bytes, size_t* bytes_allocated, 60 return AllocNonvirtualThreadUnsafe(self, num_bytes, bytes_allocated, usable_size, 71 mirror::Object* AllocNonvirtual(Thread* self, size_t num_bytes, size_t* bytes_allocated, argument 74 return AllocCommon(self, num_bytes, bytes_allocated, usable_size, 78 size_t* bytes_allocated, size_t* usable_size, 81 return AllocCommon<false>(self, num_bytes, bytes_allocated, usable_size, 91 size_t* bytes_allocated); 77 AllocNonvirtualThreadUnsafe(Thread* self, size_t num_bytes, size_t* bytes_allocated, size_t* usable_size, size_t* bytes_tl_bulk_allocated) argument [all...] |
H A D | rosalloc_space-inl.h | 65 size_t* bytes_allocated, size_t* usable_size, 82 DCHECK(bytes_allocated != nullptr); 83 *bytes_allocated = rosalloc_bytes_allocated; 99 size_t* bytes_allocated) { 100 DCHECK(bytes_allocated != nullptr); 102 rosalloc_->AllocFromThreadLocalRun(self, num_bytes, bytes_allocated)); 64 AllocCommon(Thread* self, size_t num_bytes, size_t* bytes_allocated, size_t* usable_size, size_t* bytes_tl_bulk_allocated) argument 98 AllocThreadLocal(Thread* self, size_t num_bytes, size_t* bytes_allocated) argument
|
H A D | large_object_space_test.cc | 105 size_t bytes_allocated = 0, bytes_tl_bulk_allocated; local 107 mirror::Object* obj = los->Alloc(self, 100 * MB, &bytes_allocated, nullptr,
|
H A D | memory_tool_malloc_space-inl.h | 32 size_t bytes_allocated, size_t usable_size, 37 *bytes_allocated_out = bytes_allocated; 63 // Right redzone. Assumes that if bytes_allocated > usable_size, then the difference is 90 size_t bytes_allocated; local 94 &bytes_allocated, &usable_size, 102 bytes_allocated, usable_size, 119 size_t bytes_allocated; local 123 &bytes_allocated, &usable_size, &bytes_tl_bulk_allocated); 130 bytes_allocated, usable_size, 147 size_t bytes_allocated; local 31 AdjustForValgrind(void* obj_with_rdz, size_t num_bytes, size_t bytes_allocated, size_t usable_size, size_t bytes_tl_bulk_allocated, size_t* bytes_allocated_out, size_t* usable_size_out, size_t* bytes_tl_bulk_allocated_out) argument [all...] |
H A D | region_space-inl.h | 27 inline mirror::Object* RegionSpace::Alloc(Thread*, size_t num_bytes, size_t* bytes_allocated, argument 31 return AllocNonvirtual<false>(num_bytes, bytes_allocated, usable_size, 36 size_t* bytes_allocated, 40 return Alloc(self, num_bytes, bytes_allocated, usable_size, bytes_tl_bulk_allocated); 44 inline mirror::Object* RegionSpace::AllocNonvirtual(size_t num_bytes, size_t* bytes_allocated, argument 52 obj = current_region_->Alloc(num_bytes, bytes_allocated, usable_size, 56 obj = evac_region_->Alloc(num_bytes, bytes_allocated, usable_size, 65 obj = current_region_->Alloc(num_bytes, bytes_allocated, usable_size, 68 obj = evac_region_->Alloc(num_bytes, bytes_allocated, usable_size, 85 obj = r->Alloc(num_bytes, bytes_allocated, usable_siz 35 AllocThreadUnsafe(Thread* self, size_t num_bytes, size_t* bytes_allocated, size_t* usable_size, size_t* bytes_tl_bulk_allocated) argument 115 Alloc(size_t num_bytes, size_t* bytes_allocated, size_t* usable_size, size_t* bytes_tl_bulk_allocated) argument 280 AllocLarge(size_t num_bytes, size_t* bytes_allocated, size_t* usable_size, size_t* bytes_tl_bulk_allocated) argument [all...] |
H A D | space_test.h | 71 size_t* bytes_allocated, 79 bytes_allocated, 91 size_t* bytes_allocated, 97 mirror::Object* obj = alloc_space->AllocWithGrowth(self, bytes, bytes_allocated, usable_size, 192 size_t bytes_allocated = 0; local 195 object.Assign(Alloc(space, self, alloc_size, &bytes_allocated, nullptr, 198 object.Assign(AllocWithGrowth(space, self, alloc_size, &bytes_allocated, nullptr, 206 EXPECT_EQ(bytes_allocated, allocation_size); 290 size_t bytes_allocated = 0; local 293 large_object.Assign(Alloc(space, self, three_quarters_space, &bytes_allocated, nullpt [all...] |
H A D | dlmalloc_space.cc | 129 size_t* bytes_allocated, size_t* usable_size, 138 result = AllocWithoutGrowthLocked(self, num_bytes, bytes_allocated, usable_size, 266 size_t bytes_allocated = 0; local 267 mspace_inspect_all(mspace_, DlmallocBytesAllocatedCallback, &bytes_allocated); 268 return bytes_allocated; 128 AllocWithGrowth(Thread* self, size_t num_bytes, size_t* bytes_allocated, size_t* usable_size, size_t* bytes_tl_bulk_allocated) argument
|
H A D | region_space.cc | 167 const size_t bytes_allocated = RoundUp(BytesAllocated(), kRegionSize); local 168 DCHECK_LE(live_bytes_, bytes_allocated); 173 result = live_bytes_ * 100U < kEvaculateLivePercentThreshold * bytes_allocated; 400 void RegionSpace::FreeLarge(mirror::Object* large_obj, size_t bytes_allocated) { argument 405 uint8_t* end_addr = AlignUp(reinterpret_cast<uint8_t*>(large_obj) + bytes_allocated, kRegionSize);
|
H A D | rosalloc_space.cc | 155 size_t* bytes_allocated, size_t* usable_size, 164 result = AllocCommon(self, num_bytes, bytes_allocated, usable_size, 290 size_t bytes_allocated = 0; local 291 InspectAllRosAlloc(art::gc::allocator::RosAlloc::BytesAllocatedCallback, &bytes_allocated, false); 292 return bytes_allocated; 154 AllocWithGrowth(Thread* self, size_t num_bytes, size_t* bytes_allocated, size_t* usable_size, size_t* bytes_tl_bulk_allocated) argument
|
H A D | large_object_space.cc | 51 mirror::Object* Alloc(Thread* self, size_t num_bytes, size_t* bytes_allocated, 55 LargeObjectMapSpace::Alloc(self, num_bytes + kMemoryToolRedZoneBytes * 2, bytes_allocated, 133 size_t* bytes_allocated, size_t* usable_size, 146 DCHECK(bytes_allocated != nullptr); 153 *bytes_allocated = allocation_size; 469 mirror::Object* FreeListSpace::Alloc(Thread* self, size_t num_bytes, size_t* bytes_allocated, argument 503 DCHECK(bytes_allocated != nullptr); 504 *bytes_allocated = allocation_size; 132 Alloc(Thread* self, size_t num_bytes, size_t* bytes_allocated, size_t* usable_size, size_t* bytes_tl_bulk_allocated) argument
|
/art/runtime/gc/allocator/ |
H A D | dlmalloc.cc | 89 size_t* bytes_allocated = reinterpret_cast<size_t*>(arg); local 90 *bytes_allocated += used_bytes + sizeof(size_t);
|
H A D | rosalloc-inl.h | 31 inline ALWAYS_INLINE void* RosAlloc::Alloc(Thread* self, size_t size, size_t* bytes_allocated, argument 35 return AllocLargeObject(self, size, bytes_allocated, usable_size, 40 m = AllocFromRun(self, size, bytes_allocated, usable_size, bytes_tl_bulk_allocated); 42 m = AllocFromRunThreadUnsafe(self, size, bytes_allocated, usable_size, 79 size_t* bytes_allocated) { 80 DCHECK(bytes_allocated != nullptr); 97 *bytes_allocated = bracket_size; 78 AllocFromThreadLocalRun(Thread* self, size_t size, size_t* bytes_allocated) argument
|
H A D | rosalloc.cc | 459 void* RosAlloc::AllocLargeObject(Thread* self, size_t size, size_t* bytes_allocated, argument 461 DCHECK(bytes_allocated != nullptr); 477 *bytes_allocated = total_bytes; 631 void* RosAlloc::AllocFromRunThreadUnsafe(Thread* self, size_t size, size_t* bytes_allocated, argument 634 DCHECK(bytes_allocated != nullptr); 643 *bytes_allocated = bracket_size; 651 void* RosAlloc::AllocFromRun(Thread* self, size_t size, size_t* bytes_allocated, argument 653 DCHECK(bytes_allocated != nullptr); 734 *bytes_allocated = bracket_size; 747 *bytes_allocated 1726 size_t* bytes_allocated = reinterpret_cast<size_t*>(arg); local [all...] |
/art/runtime/gc/ |
H A D | heap-inl.h | 77 size_t bytes_allocated; local 91 bytes_allocated = byte_count; 92 usable_size = bytes_allocated; 97 (obj = rosalloc_space_->AllocThreadLocal(self, byte_count, &bytes_allocated)) != nullptr && 104 usable_size = bytes_allocated; 110 obj = TryToAllocate<kInstrumented, false>(self, allocator, byte_count, &bytes_allocated, 119 &bytes_allocated, 136 DCHECK_GT(bytes_allocated, 0u); 167 thread_stats->allocated_bytes += bytes_allocated; 170 global_stats->allocated_bytes += bytes_allocated; 240 TryToAllocate(Thread* self, AllocatorType allocator_type, size_t alloc_size, size_t* bytes_allocated, size_t* usable_size, size_t* bytes_tl_bulk_allocated) argument [all...] |
H A D | heap.cc | 1660 size_t* bytes_allocated, 1681 mirror::Object* ptr = TryToAllocate<true, false>(self, allocator, alloc_size, bytes_allocated, 1696 mirror::Object* ptr = TryToAllocate<true, false>(self, allocator, alloc_size, bytes_allocated, 1717 mirror::Object* ptr = TryToAllocate<true, false>(self, allocator, alloc_size, bytes_allocated, 1726 mirror::Object* ptr = TryToAllocate<true, true>(self, allocator, alloc_size, bytes_allocated, 1745 ptr = TryToAllocate<true, true>(self, allocator, alloc_size, bytes_allocated, usable_size, 1766 ptr = TryToAllocate<true, true>(self, allocator, alloc_size, bytes_allocated, 1820 ptr = TryToAllocate<true, true>(self, allocator, alloc_size, bytes_allocated, 2377 size_t bytes_allocated, dummy; local 2378 forward_address = to_space_->Alloc(self_, alloc_size, &bytes_allocated, nullpt 1656 AllocateInternalWithGc(Thread* self, AllocatorType allocator, bool instrumented, size_t alloc_size, size_t* bytes_allocated, size_t* usable_size, size_t* bytes_tl_bulk_allocated, ObjPtr<mirror::Class>* klass) argument 3581 const uint64_t bytes_allocated = GetBytesAllocated(); local 4194 AllocWithNewTLAB(Thread* self, size_t alloc_size, bool grow, size_t* bytes_allocated, size_t* usable_size, size_t* bytes_tl_bulk_allocated) argument [all...] |
/art/runtime/gc/accounting/ |
H A D | mod_union_table_test.cc | 53 size_t bytes_allocated = 0, bytes_tl_bulk_allocated; local 55 space->Alloc(self, size, &bytes_allocated, nullptr, &bytes_tl_bulk_allocated)); 60 EXPECT_GE(bytes_allocated, size); 82 size_t bytes_allocated = 0, bytes_tl_bulk_allocated; local 83 auto* klass = down_cast<mirror::Class*>(space->Alloc(self, class_size, &bytes_allocated,
|
/art/runtime/base/ |
H A D | arena_allocator.cc | 136 const size_t bytes_allocated = BytesAllocated(); local 137 os << " MEM: used: " << bytes_allocated << ", allocated: " << malloc_bytes 142 << num_allocations << ", avg size: " << bytes_allocated / num_allocations << "\n";
|
/art/runtime/gc/collector/ |
H A D | semi_space.cc | 515 size_t bytes_allocated, dummy; local 521 forward_address = promo_dest_space_->AllocThreadUnsafe(self_, object_size, &bytes_allocated, 525 forward_address = to_space_->AllocThreadUnsafe(self_, object_size, &bytes_allocated, nullptr, 530 bytes_promoted_ += bytes_allocated; 571 forward_address = to_space_->AllocThreadUnsafe(self_, object_size, &bytes_allocated, nullptr, 579 forward_address = fallback_space_->AllocThreadUnsafe(self_, object_size, &bytes_allocated, 588 bytes_moved_ += bytes_allocated;
|
H A D | concurrent_copying.cc | 2181 size_t bytes_allocated = 0U; local 2185 bytes_allocated = region_space_bytes_allocated; 2199 bytes_allocated = region_space_alloc_size; 2217 bytes_allocated = non_moving_space_bytes_allocated; 2250 FillWithDummyObject(to_ref, bytes_allocated); 2253 if (bytes_allocated > space::RegionSpace::kRegionSize) { 2255 region_space_->FreeLarge(to_ref, bytes_allocated); 2258 heap_->num_bytes_allocated_.FetchAndAddSequentiallyConsistent(bytes_allocated); 2259 to_space_bytes_skipped_.FetchAndAddSequentiallyConsistent(bytes_allocated); 2262 skipped_blocks_map_.insert(std::make_pair(bytes_allocated, [all...] |