Searched defs:bytes_allocated (Results 1 - 21 of 21) sorted by relevance

/art/runtime/gc/space/
H A Dbump_pointer_space-inl.h28 inline mirror::Object* BumpPointerSpace::Alloc(Thread*, size_t num_bytes, size_t* bytes_allocated, argument
34 *bytes_allocated = num_bytes;
44 size_t* bytes_allocated,
55 *bytes_allocated = num_bytes;
43 AllocThreadUnsafe(Thread* self, size_t num_bytes, size_t* bytes_allocated, size_t* usable_size, size_t* bytes_tl_bulk_allocated) argument
H A Ddlmalloc_space-inl.h29 size_t* bytes_allocated,
35 obj = AllocWithoutGrowthLocked(self, num_bytes, bytes_allocated, usable_size,
56 size_t* bytes_allocated,
66 DCHECK(bytes_allocated != nullptr);
67 *bytes_allocated = allocation_size;
28 AllocNonvirtual(Thread* self, size_t num_bytes, size_t* bytes_allocated, size_t* usable_size, size_t* bytes_tl_bulk_allocated) argument
54 AllocWithoutGrowthLocked( Thread* , size_t num_bytes, size_t* bytes_allocated, size_t* usable_size, size_t* bytes_tl_bulk_allocated) argument
H A Drosalloc_space-inl.h33 size_t* bytes_allocated, size_t* usable_size,
50 DCHECK(bytes_allocated != nullptr);
51 *bytes_allocated = rosalloc_bytes_allocated;
67 size_t* bytes_allocated) {
68 DCHECK(bytes_allocated != nullptr);
70 rosalloc_->AllocFromThreadLocalRun(self, num_bytes, bytes_allocated));
32 AllocCommon(Thread* self, size_t num_bytes, size_t* bytes_allocated, size_t* usable_size, size_t* bytes_tl_bulk_allocated) argument
66 AllocThreadLocal(Thread* self, size_t num_bytes, size_t* bytes_allocated) argument
H A Drosalloc_space.h49 mirror::Object* AllocWithGrowth(Thread* self, size_t num_bytes, size_t* bytes_allocated,
52 mirror::Object* Alloc(Thread* self, size_t num_bytes, size_t* bytes_allocated,
54 return AllocNonvirtual(self, num_bytes, bytes_allocated, usable_size,
57 mirror::Object* AllocThreadUnsafe(Thread* self, size_t num_bytes, size_t* bytes_allocated,
60 return AllocNonvirtualThreadUnsafe(self, num_bytes, bytes_allocated, usable_size,
71 mirror::Object* AllocNonvirtual(Thread* self, size_t num_bytes, size_t* bytes_allocated, argument
74 return AllocCommon(self, num_bytes, bytes_allocated, usable_size,
78 size_t* bytes_allocated, size_t* usable_size,
81 return AllocCommon<false>(self, num_bytes, bytes_allocated, usable_size,
91 size_t* bytes_allocated);
77 AllocNonvirtualThreadUnsafe(Thread* self, size_t num_bytes, size_t* bytes_allocated, size_t* usable_size, size_t* bytes_tl_bulk_allocated) argument
[all...]
H A Dlarge_object_space_test.cc113 size_t bytes_allocated = 0, bytes_tl_bulk_allocated; local
115 mirror::Object* obj = los->Alloc(self, 100 * MB, &bytes_allocated, nullptr,
H A Dmemory_tool_malloc_space-inl.h34 size_t bytes_allocated, size_t usable_size,
39 *bytes_allocated_out = bytes_allocated;
65 // Right redzone. Assumes that if bytes_allocated > usable_size, then the difference is
92 size_t bytes_allocated; local
96 &bytes_allocated, &usable_size,
104 bytes_allocated, usable_size,
121 size_t bytes_allocated; local
125 &bytes_allocated, &usable_size, &bytes_tl_bulk_allocated);
132 bytes_allocated, usable_size,
149 size_t bytes_allocated; local
33 AdjustForValgrind(void* obj_with_rdz, size_t num_bytes, size_t bytes_allocated, size_t usable_size, size_t bytes_tl_bulk_allocated, size_t* bytes_allocated_out, size_t* usable_size_out, size_t* bytes_tl_bulk_allocated_out) argument
[all...]
H A Dregion_space-inl.h29 /* out */ size_t* bytes_allocated,
33 return AllocNonvirtual<false>(num_bytes, bytes_allocated, usable_size,
39 /* out */ size_t* bytes_allocated,
43 return Alloc(self, num_bytes, bytes_allocated, usable_size, bytes_tl_bulk_allocated);
48 /* out */ size_t* bytes_allocated,
56 bytes_allocated,
65 bytes_allocated,
73 obj = r->Alloc(num_bytes, bytes_allocated, usable_size, bytes_tl_bulk_allocated);
86 obj = AllocLarge<kForEvac>(num_bytes, bytes_allocated, usable_size, bytes_tl_bulk_allocated);
95 /* out */ size_t* bytes_allocated,
27 Alloc(Thread* self ATTRIBUTE_UNUSED, size_t num_bytes, size_t* bytes_allocated, size_t* usable_size, size_t* bytes_tl_bulk_allocated) argument
37 AllocThreadUnsafe(Thread* self, size_t num_bytes, size_t* bytes_allocated, size_t* usable_size, size_t* bytes_tl_bulk_allocated) argument
47 AllocNonvirtual(size_t num_bytes, size_t* bytes_allocated, size_t* usable_size, size_t* bytes_tl_bulk_allocated) argument
94 Alloc(size_t num_bytes, size_t* bytes_allocated, size_t* usable_size, size_t* bytes_tl_bulk_allocated) argument
244 AllocLarge(size_t num_bytes, size_t* bytes_allocated, size_t* usable_size, size_t* bytes_tl_bulk_allocated) argument
316 FreeLarge(mirror::Object* large_obj, size_t bytes_allocated) argument
[all...]
H A Dspace_test.h71 size_t* bytes_allocated,
79 bytes_allocated,
91 size_t* bytes_allocated,
97 mirror::Object* obj = alloc_space->AllocWithGrowth(self, bytes, bytes_allocated, usable_size,
192 size_t bytes_allocated = 0; local
195 object.Assign(Alloc(space, self, alloc_size, &bytes_allocated, nullptr,
198 object.Assign(AllocWithGrowth(space, self, alloc_size, &bytes_allocated, nullptr,
206 EXPECT_EQ(bytes_allocated, allocation_size);
290 size_t bytes_allocated = 0; local
293 large_object.Assign(Alloc(space, self, three_quarters_space, &bytes_allocated, nullpt
[all...]
H A Ddlmalloc_space.cc131 size_t* bytes_allocated, size_t* usable_size,
140 result = AllocWithoutGrowthLocked(self, num_bytes, bytes_allocated, usable_size,
268 size_t bytes_allocated = 0; local
269 mspace_inspect_all(mspace_, DlmallocBytesAllocatedCallback, &bytes_allocated);
270 return bytes_allocated;
130 AllocWithGrowth(Thread* self, size_t num_bytes, size_t* bytes_allocated, size_t* usable_size, size_t* bytes_tl_bulk_allocated) argument
H A Drosalloc_space.cc157 size_t* bytes_allocated, size_t* usable_size,
166 result = AllocCommon(self, num_bytes, bytes_allocated, usable_size,
292 size_t bytes_allocated = 0; local
293 InspectAllRosAlloc(art::gc::allocator::RosAlloc::BytesAllocatedCallback, &bytes_allocated, false);
294 return bytes_allocated;
156 AllocWithGrowth(Thread* self, size_t num_bytes, size_t* bytes_allocated, size_t* usable_size, size_t* bytes_tl_bulk_allocated) argument
H A Dlarge_object_space.cc56 mirror::Object* Alloc(Thread* self, size_t num_bytes, size_t* bytes_allocated,
60 LargeObjectMapSpace::Alloc(self, num_bytes + kMemoryToolRedZoneBytes * 2, bytes_allocated,
138 size_t* bytes_allocated, size_t* usable_size,
151 DCHECK(bytes_allocated != nullptr);
158 *bytes_allocated = allocation_size;
474 mirror::Object* FreeListSpace::Alloc(Thread* self, size_t num_bytes, size_t* bytes_allocated, argument
508 DCHECK(bytes_allocated != nullptr);
509 *bytes_allocated = allocation_size;
137 Alloc(Thread* self, size_t num_bytes, size_t* bytes_allocated, size_t* usable_size, size_t* bytes_tl_bulk_allocated) argument
H A Dregion_space.cc174 const size_t bytes_allocated = RoundUp(BytesAllocated(), kRegionSize); local
175 DCHECK_LE(live_bytes_, bytes_allocated);
180 result = (live_bytes_ * 100U < kEvacuateLivePercentThreshold * bytes_allocated);
/art/runtime/gc/allocator/
H A Ddlmalloc.cc92 size_t* bytes_allocated = reinterpret_cast<size_t*>(arg); local
93 *bytes_allocated += used_bytes + sizeof(size_t);
H A Drosalloc-inl.h31 inline ALWAYS_INLINE void* RosAlloc::Alloc(Thread* self, size_t size, size_t* bytes_allocated, argument
35 return AllocLargeObject(self, size, bytes_allocated, usable_size,
40 m = AllocFromRun(self, size, bytes_allocated, usable_size, bytes_tl_bulk_allocated);
42 m = AllocFromRunThreadUnsafe(self, size, bytes_allocated, usable_size,
79 size_t* bytes_allocated) {
80 DCHECK(bytes_allocated != nullptr);
97 *bytes_allocated = bracket_size;
78 AllocFromThreadLocalRun(Thread* self, size_t size, size_t* bytes_allocated) argument
H A Drosalloc.cc460 void* RosAlloc::AllocLargeObject(Thread* self, size_t size, size_t* bytes_allocated, argument
462 DCHECK(bytes_allocated != nullptr);
478 *bytes_allocated = total_bytes;
632 void* RosAlloc::AllocFromRunThreadUnsafe(Thread* self, size_t size, size_t* bytes_allocated, argument
635 DCHECK(bytes_allocated != nullptr);
644 *bytes_allocated = bracket_size;
652 void* RosAlloc::AllocFromRun(Thread* self, size_t size, size_t* bytes_allocated, argument
654 DCHECK(bytes_allocated != nullptr);
735 *bytes_allocated = bracket_size;
748 *bytes_allocated
1727 size_t* bytes_allocated = reinterpret_cast<size_t*>(arg); local
[all...]
/art/runtime/gc/
H A Dheap-inl.h79 size_t bytes_allocated; local
93 bytes_allocated = byte_count;
94 usable_size = bytes_allocated;
99 (obj = rosalloc_space_->AllocThreadLocal(self, byte_count, &bytes_allocated)) != nullptr &&
106 usable_size = bytes_allocated;
112 obj = TryToAllocate<kInstrumented, false>(self, allocator, byte_count, &bytes_allocated,
121 &bytes_allocated,
138 DCHECK_GT(bytes_allocated, 0u);
175 thread_stats->allocated_bytes += bytes_allocated;
178 global_stats->allocated_bytes += bytes_allocated;
248 TryToAllocate(Thread* self, AllocatorType allocator_type, size_t alloc_size, size_t* bytes_allocated, size_t* usable_size, size_t* bytes_tl_bulk_allocated) argument
[all...]
H A Dheap.cc1575 size_t* bytes_allocated,
1596 mirror::Object* ptr = TryToAllocate<true, false>(self, allocator, alloc_size, bytes_allocated,
1611 mirror::Object* ptr = TryToAllocate<true, false>(self, allocator, alloc_size, bytes_allocated,
1632 mirror::Object* ptr = TryToAllocate<true, false>(self, allocator, alloc_size, bytes_allocated,
1641 mirror::Object* ptr = TryToAllocate<true, true>(self, allocator, alloc_size, bytes_allocated,
1660 ptr = TryToAllocate<true, true>(self, allocator, alloc_size, bytes_allocated, usable_size,
1681 ptr = TryToAllocate<true, true>(self, allocator, alloc_size, bytes_allocated,
1735 ptr = TryToAllocate<true, true>(self, allocator, alloc_size, bytes_allocated,
2240 size_t bytes_allocated, dummy; local
2241 forward_address = to_space_->Alloc(self_, alloc_size, &bytes_allocated, nullpt
1571 AllocateInternalWithGc(Thread* self, AllocatorType allocator, bool instrumented, size_t alloc_size, size_t* bytes_allocated, size_t* usable_size, size_t* bytes_tl_bulk_allocated, ObjPtr<mirror::Class>* klass) argument
3444 const uint64_t bytes_allocated = GetBytesAllocated(); local
4020 AllocWithNewTLAB(Thread* self, size_t alloc_size, bool grow, size_t* bytes_allocated, size_t* usable_size, size_t* bytes_tl_bulk_allocated) argument
[all...]
/art/runtime/gc/accounting/
H A Dmod_union_table_test.cc53 size_t bytes_allocated = 0, bytes_tl_bulk_allocated; local
55 space->Alloc(self, size, &bytes_allocated, nullptr, &bytes_tl_bulk_allocated));
60 EXPECT_GE(bytes_allocated, size);
82 size_t bytes_allocated = 0, bytes_tl_bulk_allocated; local
83 auto* klass = down_cast<mirror::Class*>(space->Alloc(self, class_size, &bytes_allocated,
/art/runtime/base/
H A Darena_allocator.cc144 const size_t bytes_allocated = BytesAllocated(); local
145 os << " MEM: used: " << bytes_allocated << ", allocated: " << malloc_bytes
150 << num_allocations << ", avg size: " << bytes_allocated / num_allocations << "\n";
/art/runtime/gc/collector/
H A Dsemi_space.cc515 size_t bytes_allocated, dummy; local
521 forward_address = promo_dest_space_->AllocThreadUnsafe(self_, object_size, &bytes_allocated,
525 forward_address = to_space_->AllocThreadUnsafe(self_, object_size, &bytes_allocated, nullptr,
530 bytes_promoted_ += bytes_allocated;
571 forward_address = to_space_->AllocThreadUnsafe(self_, object_size, &bytes_allocated, nullptr,
579 forward_address = fallback_space_->AllocThreadUnsafe(self_, object_size, &bytes_allocated,
588 bytes_moved_ += bytes_allocated;
H A Dconcurrent_copying.cc2357 size_t bytes_allocated = 0U; local
2362 bytes_allocated = region_space_bytes_allocated;
2374 bytes_allocated = region_space_alloc_size;
2375 heap_->num_bytes_allocated_.fetch_sub(bytes_allocated, std::memory_order_seq_cst);
2376 to_space_bytes_skipped_.fetch_sub(bytes_allocated, std::memory_order_seq_cst);
2394 bytes_allocated = non_moving_space_bytes_allocated;
2427 FillWithDummyObject(to_ref, bytes_allocated);
2430 if (bytes_allocated > space::RegionSpace::kRegionSize) {
2432 region_space_->FreeLarge</*kForEvac*/ true>(to_ref, bytes_allocated);
2435 heap_->num_bytes_allocated_.FetchAndAddSequentiallyConsistent(bytes_allocated);
[all...]

Completed in 4010 milliseconds