/art/runtime/gc/allocator/ |
H A D | rosalloc.h | 520 void* AllocFromRun(Thread* self, size_t size, size_t* bytes_allocated, size_t* usable_size, 525 void* AllocFromRunThreadUnsafe(Thread* self, size_t size, size_t* bytes_allocated, 545 void* AllocLargeObject(Thread* self, size_t size, size_t* bytes_allocated, 571 void* Alloc(Thread* self, size_t size, size_t* bytes_allocated, size_t* usable_size, 584 ALWAYS_INLINE void* AllocFromThreadLocalRun(Thread* self, size_t size, size_t* bytes_allocated);
|
H A D | rosalloc.cc | 458 void* RosAlloc::AllocLargeObject(Thread* self, size_t size, size_t* bytes_allocated, argument 460 DCHECK(bytes_allocated != nullptr); 476 *bytes_allocated = total_bytes; 631 void* RosAlloc::AllocFromRunThreadUnsafe(Thread* self, size_t size, size_t* bytes_allocated, argument 634 DCHECK(bytes_allocated != nullptr); 648 *bytes_allocated = bracket_size; 656 void* RosAlloc::AllocFromRun(Thread* self, size_t size, size_t* bytes_allocated, argument 658 DCHECK(bytes_allocated != nullptr); 750 *bytes_allocated = bracket_size; 763 *bytes_allocated 1876 size_t* bytes_allocated = reinterpret_cast<size_t*>(arg); local [all...] |
/art/runtime/gc/collector/ |
H A D | semi_space.cc | 493 size_t bytes_allocated, dummy; local 499 forward_address = promo_dest_space_->AllocThreadUnsafe(self_, object_size, &bytes_allocated, 503 forward_address = to_space_->AllocThreadUnsafe(self_, object_size, &bytes_allocated, nullptr, 508 bytes_promoted_ += bytes_allocated; 549 forward_address = to_space_->AllocThreadUnsafe(self_, object_size, &bytes_allocated, nullptr, 557 forward_address = fallback_space_->AllocThreadUnsafe(self_, object_size, &bytes_allocated, 566 bytes_moved_ += bytes_allocated;
|
H A D | concurrent_copying.cc | 1293 size_t bytes_allocated = 0U; local 1297 bytes_allocated = region_space_bytes_allocated; 1311 bytes_allocated = region_space_alloc_size; 1324 bytes_allocated = non_moving_space_bytes_allocated; 1347 FillWithDummyObject(to_ref, bytes_allocated); 1350 if (bytes_allocated > space::RegionSpace::kRegionSize) { 1352 region_space_->FreeLarge(to_ref, bytes_allocated); 1355 heap_->num_bytes_allocated_.FetchAndAddSequentiallyConsistent(bytes_allocated); 1356 to_space_bytes_skipped_.FetchAndAddSequentiallyConsistent(bytes_allocated); 1359 skipped_blocks_map_.insert(std::make_pair(bytes_allocated, [all...] |
/art/runtime/gc/ |
H A D | heap.cc | 1497 size_t alloc_size, size_t* bytes_allocated, 1518 mirror::Object* ptr = TryToAllocate<true, false>(self, allocator, alloc_size, bytes_allocated, 1532 mirror::Object* ptr = TryToAllocate<true, false>(self, allocator, alloc_size, bytes_allocated, 1552 mirror::Object* ptr = TryToAllocate<true, false>(self, allocator, alloc_size, bytes_allocated, 1561 mirror::Object* ptr = TryToAllocate<true, true>(self, allocator, alloc_size, bytes_allocated, 1579 ptr = TryToAllocate<true, true>(self, allocator, alloc_size, bytes_allocated, usable_size, 1595 ptr = TryToAllocate<true, true>(self, allocator, alloc_size, bytes_allocated, 1640 ptr = TryToAllocate<true, true>(self, allocator, alloc_size, bytes_allocated, 2142 size_t bytes_allocated, dummy; local 2143 forward_address = to_space_->Alloc(self_, alloc_size, &bytes_allocated, nullpt 1496 AllocateInternalWithGc(Thread* self, AllocatorType allocator, size_t alloc_size, size_t* bytes_allocated, size_t* usable_size, size_t* bytes_tl_bulk_allocated, mirror::Class** klass) argument 3275 const uint64_t bytes_allocated = GetBytesAllocated(); local [all...] |
H A D | heap.h | 748 size_t* bytes_allocated, size_t* usable_size, 767 size_t alloc_size, size_t* bytes_allocated,
|
/art/runtime/gc/space/ |
H A D | region_space.cc | 288 void RegionSpace::FreeLarge(mirror::Object* large_obj, size_t bytes_allocated) { argument 293 uint8_t* end_addr = AlignUp(reinterpret_cast<uint8_t*>(large_obj) + bytes_allocated, kRegionSize);
|