Searched defs:bytes_allocated (Results 1 - 21 of 21) sorted by relevance

/art/runtime/gc/space/
H A Dbump_pointer_space-inl.h27 inline mirror::Object* BumpPointerSpace::Alloc(Thread*, size_t num_bytes, size_t* bytes_allocated, argument
33 *bytes_allocated = num_bytes;
43 size_t* bytes_allocated,
54 *bytes_allocated = num_bytes;
42 AllocThreadUnsafe(Thread* self, size_t num_bytes, size_t* bytes_allocated, size_t* usable_size, size_t* bytes_tl_bulk_allocated) argument
H A Ddlmalloc_space-inl.h29 size_t* bytes_allocated,
35 obj = AllocWithoutGrowthLocked(self, num_bytes, bytes_allocated, usable_size,
56 size_t* bytes_allocated,
66 DCHECK(bytes_allocated != nullptr);
67 *bytes_allocated = allocation_size;
28 AllocNonvirtual(Thread* self, size_t num_bytes, size_t* bytes_allocated, size_t* usable_size, size_t* bytes_tl_bulk_allocated) argument
54 AllocWithoutGrowthLocked( Thread* , size_t num_bytes, size_t* bytes_allocated, size_t* usable_size, size_t* bytes_tl_bulk_allocated) argument
H A Drosalloc_space.h49 mirror::Object* AllocWithGrowth(Thread* self, size_t num_bytes, size_t* bytes_allocated,
52 mirror::Object* Alloc(Thread* self, size_t num_bytes, size_t* bytes_allocated,
54 return AllocNonvirtual(self, num_bytes, bytes_allocated, usable_size,
57 mirror::Object* AllocThreadUnsafe(Thread* self, size_t num_bytes, size_t* bytes_allocated,
60 return AllocNonvirtualThreadUnsafe(self, num_bytes, bytes_allocated, usable_size,
71 mirror::Object* AllocNonvirtual(Thread* self, size_t num_bytes, size_t* bytes_allocated, argument
74 return AllocCommon(self, num_bytes, bytes_allocated, usable_size,
78 size_t* bytes_allocated, size_t* usable_size,
81 return AllocCommon<false>(self, num_bytes, bytes_allocated, usable_size,
91 size_t* bytes_allocated);
77 AllocNonvirtualThreadUnsafe(Thread* self, size_t num_bytes, size_t* bytes_allocated, size_t* usable_size, size_t* bytes_tl_bulk_allocated) argument
[all...]
H A Drosalloc_space-inl.h65 size_t* bytes_allocated, size_t* usable_size,
82 DCHECK(bytes_allocated != nullptr);
83 *bytes_allocated = rosalloc_bytes_allocated;
99 size_t* bytes_allocated) {
100 DCHECK(bytes_allocated != nullptr);
102 rosalloc_->AllocFromThreadLocalRun(self, num_bytes, bytes_allocated));
64 AllocCommon(Thread* self, size_t num_bytes, size_t* bytes_allocated, size_t* usable_size, size_t* bytes_tl_bulk_allocated) argument
98 AllocThreadLocal(Thread* self, size_t num_bytes, size_t* bytes_allocated) argument
H A Dlarge_object_space_test.cc105 size_t bytes_allocated = 0, bytes_tl_bulk_allocated; local
107 mirror::Object* obj = los->Alloc(self, 100 * MB, &bytes_allocated, nullptr,
H A Dmemory_tool_malloc_space-inl.h32 size_t bytes_allocated, size_t usable_size,
37 *bytes_allocated_out = bytes_allocated;
63 // Right redzone. Assumes that if bytes_allocated > usable_size, then the difference is
90 size_t bytes_allocated; local
94 &bytes_allocated, &usable_size,
102 bytes_allocated, usable_size,
119 size_t bytes_allocated; local
123 &bytes_allocated, &usable_size, &bytes_tl_bulk_allocated);
130 bytes_allocated, usable_size,
147 size_t bytes_allocated; local
31 AdjustForValgrind(void* obj_with_rdz, size_t num_bytes, size_t bytes_allocated, size_t usable_size, size_t bytes_tl_bulk_allocated, size_t* bytes_allocated_out, size_t* usable_size_out, size_t* bytes_tl_bulk_allocated_out) argument
[all...]
H A Dregion_space-inl.h27 inline mirror::Object* RegionSpace::Alloc(Thread*, size_t num_bytes, size_t* bytes_allocated, argument
31 return AllocNonvirtual<false>(num_bytes, bytes_allocated, usable_size,
36 size_t* bytes_allocated,
40 return Alloc(self, num_bytes, bytes_allocated, usable_size, bytes_tl_bulk_allocated);
44 inline mirror::Object* RegionSpace::AllocNonvirtual(size_t num_bytes, size_t* bytes_allocated, argument
52 obj = current_region_->Alloc(num_bytes, bytes_allocated, usable_size,
56 obj = evac_region_->Alloc(num_bytes, bytes_allocated, usable_size,
65 obj = current_region_->Alloc(num_bytes, bytes_allocated, usable_size,
68 obj = evac_region_->Alloc(num_bytes, bytes_allocated, usable_size,
85 obj = r->Alloc(num_bytes, bytes_allocated, usable_siz
35 AllocThreadUnsafe(Thread* self, size_t num_bytes, size_t* bytes_allocated, size_t* usable_size, size_t* bytes_tl_bulk_allocated) argument
115 Alloc(size_t num_bytes, size_t* bytes_allocated, size_t* usable_size, size_t* bytes_tl_bulk_allocated) argument
280 AllocLarge(size_t num_bytes, size_t* bytes_allocated, size_t* usable_size, size_t* bytes_tl_bulk_allocated) argument
[all...]
H A Dspace_test.h71 size_t* bytes_allocated,
79 bytes_allocated,
91 size_t* bytes_allocated,
97 mirror::Object* obj = alloc_space->AllocWithGrowth(self, bytes, bytes_allocated, usable_size,
192 size_t bytes_allocated = 0; local
195 object.Assign(Alloc(space, self, alloc_size, &bytes_allocated, nullptr,
198 object.Assign(AllocWithGrowth(space, self, alloc_size, &bytes_allocated, nullptr,
206 EXPECT_EQ(bytes_allocated, allocation_size);
290 size_t bytes_allocated = 0; local
293 large_object.Assign(Alloc(space, self, three_quarters_space, &bytes_allocated, nullpt
[all...]
H A Ddlmalloc_space.cc129 size_t* bytes_allocated, size_t* usable_size,
138 result = AllocWithoutGrowthLocked(self, num_bytes, bytes_allocated, usable_size,
266 size_t bytes_allocated = 0; local
267 mspace_inspect_all(mspace_, DlmallocBytesAllocatedCallback, &bytes_allocated);
268 return bytes_allocated;
128 AllocWithGrowth(Thread* self, size_t num_bytes, size_t* bytes_allocated, size_t* usable_size, size_t* bytes_tl_bulk_allocated) argument
H A Dregion_space.cc167 const size_t bytes_allocated = RoundUp(BytesAllocated(), kRegionSize); local
168 DCHECK_LE(live_bytes_, bytes_allocated);
173 result = live_bytes_ * 100U < kEvaculateLivePercentThreshold * bytes_allocated;
400 void RegionSpace::FreeLarge(mirror::Object* large_obj, size_t bytes_allocated) { argument
405 uint8_t* end_addr = AlignUp(reinterpret_cast<uint8_t*>(large_obj) + bytes_allocated, kRegionSize);
H A Drosalloc_space.cc155 size_t* bytes_allocated, size_t* usable_size,
164 result = AllocCommon(self, num_bytes, bytes_allocated, usable_size,
290 size_t bytes_allocated = 0; local
291 InspectAllRosAlloc(art::gc::allocator::RosAlloc::BytesAllocatedCallback, &bytes_allocated, false);
292 return bytes_allocated;
154 AllocWithGrowth(Thread* self, size_t num_bytes, size_t* bytes_allocated, size_t* usable_size, size_t* bytes_tl_bulk_allocated) argument
H A Dlarge_object_space.cc51 mirror::Object* Alloc(Thread* self, size_t num_bytes, size_t* bytes_allocated,
55 LargeObjectMapSpace::Alloc(self, num_bytes + kMemoryToolRedZoneBytes * 2, bytes_allocated,
133 size_t* bytes_allocated, size_t* usable_size,
146 DCHECK(bytes_allocated != nullptr);
153 *bytes_allocated = allocation_size;
469 mirror::Object* FreeListSpace::Alloc(Thread* self, size_t num_bytes, size_t* bytes_allocated, argument
503 DCHECK(bytes_allocated != nullptr);
504 *bytes_allocated = allocation_size;
132 Alloc(Thread* self, size_t num_bytes, size_t* bytes_allocated, size_t* usable_size, size_t* bytes_tl_bulk_allocated) argument
/art/runtime/gc/allocator/
H A Ddlmalloc.cc89 size_t* bytes_allocated = reinterpret_cast<size_t*>(arg); local
90 *bytes_allocated += used_bytes + sizeof(size_t);
H A Drosalloc-inl.h31 inline ALWAYS_INLINE void* RosAlloc::Alloc(Thread* self, size_t size, size_t* bytes_allocated, argument
35 return AllocLargeObject(self, size, bytes_allocated, usable_size,
40 m = AllocFromRun(self, size, bytes_allocated, usable_size, bytes_tl_bulk_allocated);
42 m = AllocFromRunThreadUnsafe(self, size, bytes_allocated, usable_size,
79 size_t* bytes_allocated) {
80 DCHECK(bytes_allocated != nullptr);
97 *bytes_allocated = bracket_size;
78 AllocFromThreadLocalRun(Thread* self, size_t size, size_t* bytes_allocated) argument
H A Drosalloc.cc459 void* RosAlloc::AllocLargeObject(Thread* self, size_t size, size_t* bytes_allocated, argument
461 DCHECK(bytes_allocated != nullptr);
477 *bytes_allocated = total_bytes;
631 void* RosAlloc::AllocFromRunThreadUnsafe(Thread* self, size_t size, size_t* bytes_allocated, argument
634 DCHECK(bytes_allocated != nullptr);
643 *bytes_allocated = bracket_size;
651 void* RosAlloc::AllocFromRun(Thread* self, size_t size, size_t* bytes_allocated, argument
653 DCHECK(bytes_allocated != nullptr);
734 *bytes_allocated = bracket_size;
747 *bytes_allocated
1726 size_t* bytes_allocated = reinterpret_cast<size_t*>(arg); local
[all...]
/art/runtime/gc/
H A Dheap-inl.h77 size_t bytes_allocated; local
91 bytes_allocated = byte_count;
92 usable_size = bytes_allocated;
97 (obj = rosalloc_space_->AllocThreadLocal(self, byte_count, &bytes_allocated)) != nullptr &&
104 usable_size = bytes_allocated;
110 obj = TryToAllocate<kInstrumented, false>(self, allocator, byte_count, &bytes_allocated,
119 &bytes_allocated,
136 DCHECK_GT(bytes_allocated, 0u);
167 thread_stats->allocated_bytes += bytes_allocated;
170 global_stats->allocated_bytes += bytes_allocated;
240 TryToAllocate(Thread* self, AllocatorType allocator_type, size_t alloc_size, size_t* bytes_allocated, size_t* usable_size, size_t* bytes_tl_bulk_allocated) argument
[all...]
H A Dheap.cc1660 size_t* bytes_allocated,
1681 mirror::Object* ptr = TryToAllocate<true, false>(self, allocator, alloc_size, bytes_allocated,
1696 mirror::Object* ptr = TryToAllocate<true, false>(self, allocator, alloc_size, bytes_allocated,
1717 mirror::Object* ptr = TryToAllocate<true, false>(self, allocator, alloc_size, bytes_allocated,
1726 mirror::Object* ptr = TryToAllocate<true, true>(self, allocator, alloc_size, bytes_allocated,
1745 ptr = TryToAllocate<true, true>(self, allocator, alloc_size, bytes_allocated, usable_size,
1766 ptr = TryToAllocate<true, true>(self, allocator, alloc_size, bytes_allocated,
1820 ptr = TryToAllocate<true, true>(self, allocator, alloc_size, bytes_allocated,
2377 size_t bytes_allocated, dummy; local
2378 forward_address = to_space_->Alloc(self_, alloc_size, &bytes_allocated, nullpt
1656 AllocateInternalWithGc(Thread* self, AllocatorType allocator, bool instrumented, size_t alloc_size, size_t* bytes_allocated, size_t* usable_size, size_t* bytes_tl_bulk_allocated, ObjPtr<mirror::Class>* klass) argument
3581 const uint64_t bytes_allocated = GetBytesAllocated(); local
4194 AllocWithNewTLAB(Thread* self, size_t alloc_size, bool grow, size_t* bytes_allocated, size_t* usable_size, size_t* bytes_tl_bulk_allocated) argument
[all...]
/art/runtime/gc/accounting/
H A Dmod_union_table_test.cc53 size_t bytes_allocated = 0, bytes_tl_bulk_allocated; local
55 space->Alloc(self, size, &bytes_allocated, nullptr, &bytes_tl_bulk_allocated));
60 EXPECT_GE(bytes_allocated, size);
82 size_t bytes_allocated = 0, bytes_tl_bulk_allocated; local
83 auto* klass = down_cast<mirror::Class*>(space->Alloc(self, class_size, &bytes_allocated,
/art/runtime/base/
H A Darena_allocator.cc136 const size_t bytes_allocated = BytesAllocated(); local
137 os << " MEM: used: " << bytes_allocated << ", allocated: " << malloc_bytes
142 << num_allocations << ", avg size: " << bytes_allocated / num_allocations << "\n";
/art/runtime/gc/collector/
H A Dsemi_space.cc515 size_t bytes_allocated, dummy; local
521 forward_address = promo_dest_space_->AllocThreadUnsafe(self_, object_size, &bytes_allocated,
525 forward_address = to_space_->AllocThreadUnsafe(self_, object_size, &bytes_allocated, nullptr,
530 bytes_promoted_ += bytes_allocated;
571 forward_address = to_space_->AllocThreadUnsafe(self_, object_size, &bytes_allocated, nullptr,
579 forward_address = fallback_space_->AllocThreadUnsafe(self_, object_size, &bytes_allocated,
588 bytes_moved_ += bytes_allocated;
H A Dconcurrent_copying.cc2181 size_t bytes_allocated = 0U; local
2185 bytes_allocated = region_space_bytes_allocated;
2199 bytes_allocated = region_space_alloc_size;
2217 bytes_allocated = non_moving_space_bytes_allocated;
2250 FillWithDummyObject(to_ref, bytes_allocated);
2253 if (bytes_allocated > space::RegionSpace::kRegionSize) {
2255 region_space_->FreeLarge(to_ref, bytes_allocated);
2258 heap_->num_bytes_allocated_.FetchAndAddSequentiallyConsistent(bytes_allocated);
2259 to_space_bytes_skipped_.FetchAndAddSequentiallyConsistent(bytes_allocated);
2262 skipped_blocks_map_.insert(std::make_pair(bytes_allocated,
[all...]

Completed in 198 milliseconds