Searched refs:bytes_allocated (Results 26 - 32 of 32) sorted by relevance

12

/art/runtime/gc/allocator/
H A Drosalloc.h520 void* AllocFromRun(Thread* self, size_t size, size_t* bytes_allocated, size_t* usable_size,
525 void* AllocFromRunThreadUnsafe(Thread* self, size_t size, size_t* bytes_allocated,
545 void* AllocLargeObject(Thread* self, size_t size, size_t* bytes_allocated,
571 void* Alloc(Thread* self, size_t size, size_t* bytes_allocated, size_t* usable_size,
584 ALWAYS_INLINE void* AllocFromThreadLocalRun(Thread* self, size_t size, size_t* bytes_allocated);
H A Drosalloc.cc458 void* RosAlloc::AllocLargeObject(Thread* self, size_t size, size_t* bytes_allocated, argument
460 DCHECK(bytes_allocated != nullptr);
476 *bytes_allocated = total_bytes;
631 void* RosAlloc::AllocFromRunThreadUnsafe(Thread* self, size_t size, size_t* bytes_allocated, argument
634 DCHECK(bytes_allocated != nullptr);
648 *bytes_allocated = bracket_size;
656 void* RosAlloc::AllocFromRun(Thread* self, size_t size, size_t* bytes_allocated, argument
658 DCHECK(bytes_allocated != nullptr);
750 *bytes_allocated = bracket_size;
763 *bytes_allocated
1876 size_t* bytes_allocated = reinterpret_cast<size_t*>(arg); local
[all...]
/art/runtime/gc/collector/
H A Dsemi_space.cc493 size_t bytes_allocated, dummy; local
499 forward_address = promo_dest_space_->AllocThreadUnsafe(self_, object_size, &bytes_allocated,
503 forward_address = to_space_->AllocThreadUnsafe(self_, object_size, &bytes_allocated, nullptr,
508 bytes_promoted_ += bytes_allocated;
549 forward_address = to_space_->AllocThreadUnsafe(self_, object_size, &bytes_allocated, nullptr,
557 forward_address = fallback_space_->AllocThreadUnsafe(self_, object_size, &bytes_allocated,
566 bytes_moved_ += bytes_allocated;
H A Dconcurrent_copying.cc1293 size_t bytes_allocated = 0U; local
1297 bytes_allocated = region_space_bytes_allocated;
1311 bytes_allocated = region_space_alloc_size;
1324 bytes_allocated = non_moving_space_bytes_allocated;
1347 FillWithDummyObject(to_ref, bytes_allocated);
1350 if (bytes_allocated > space::RegionSpace::kRegionSize) {
1352 region_space_->FreeLarge(to_ref, bytes_allocated);
1355 heap_->num_bytes_allocated_.FetchAndAddSequentiallyConsistent(bytes_allocated);
1356 to_space_bytes_skipped_.FetchAndAddSequentiallyConsistent(bytes_allocated);
1359 skipped_blocks_map_.insert(std::make_pair(bytes_allocated,
[all...]
/art/runtime/gc/
H A Dheap.cc1497 size_t alloc_size, size_t* bytes_allocated,
1518 mirror::Object* ptr = TryToAllocate<true, false>(self, allocator, alloc_size, bytes_allocated,
1532 mirror::Object* ptr = TryToAllocate<true, false>(self, allocator, alloc_size, bytes_allocated,
1552 mirror::Object* ptr = TryToAllocate<true, false>(self, allocator, alloc_size, bytes_allocated,
1561 mirror::Object* ptr = TryToAllocate<true, true>(self, allocator, alloc_size, bytes_allocated,
1579 ptr = TryToAllocate<true, true>(self, allocator, alloc_size, bytes_allocated, usable_size,
1595 ptr = TryToAllocate<true, true>(self, allocator, alloc_size, bytes_allocated,
1640 ptr = TryToAllocate<true, true>(self, allocator, alloc_size, bytes_allocated,
2142 size_t bytes_allocated, dummy; local
2143 forward_address = to_space_->Alloc(self_, alloc_size, &bytes_allocated, nullpt
1496 AllocateInternalWithGc(Thread* self, AllocatorType allocator, size_t alloc_size, size_t* bytes_allocated, size_t* usable_size, size_t* bytes_tl_bulk_allocated, mirror::Class** klass) argument
3275 const uint64_t bytes_allocated = GetBytesAllocated(); local
[all...]
H A Dheap.h748 size_t* bytes_allocated, size_t* usable_size,
767 size_t alloc_size, size_t* bytes_allocated,
/art/runtime/gc/space/
H A Dregion_space.cc288 void RegionSpace::FreeLarge(mirror::Object* large_obj, size_t bytes_allocated) { argument
293 uint8_t* end_addr = AlignUp(reinterpret_cast<uint8_t*>(large_obj) + bytes_allocated, kRegionSize);

Completed in 98 milliseconds

12