/art/runtime/gc/ |
H A D | heap-inl.h | 238 size_t alloc_size, 245 UNLIKELY(IsOutOfMemoryOnAllocation<kGrow>(allocator_type, alloc_size))) { 252 alloc_size = RoundUp(alloc_size, space::BumpPointerSpace::kAlignment); 253 ret = bump_pointer_space_->AllocNonvirtual(alloc_size); 255 *bytes_allocated = alloc_size; 256 *usable_size = alloc_size; 257 *bytes_tl_bulk_allocated = alloc_size; 264 size_t max_bytes_tl_bulk_allocated = rosalloc_space_->MaxBytesBulkAllocatedFor(alloc_size); 269 ret = rosalloc_space_->Alloc(self, alloc_size, bytes_allocate 236 TryToAllocate(Thread* self, AllocatorType allocator_type, size_t alloc_size, size_t* bytes_allocated, size_t* usable_size, size_t* bytes_tl_bulk_allocated) argument 407 IsOutOfMemoryOnAllocation(AllocatorType allocator_type, size_t alloc_size) argument [all...] |
H A D | heap.cc | 1687 size_t alloc_size, 1710 mirror::Object* ptr = TryToAllocate<true, false>(self, allocator, alloc_size, bytes_allocated, 1725 mirror::Object* ptr = TryToAllocate<true, false>(self, allocator, alloc_size, bytes_allocated, 1746 mirror::Object* ptr = TryToAllocate<true, false>(self, allocator, alloc_size, bytes_allocated, 1755 mirror::Object* ptr = TryToAllocate<true, true>(self, allocator, alloc_size, bytes_allocated, 1764 VLOG(gc) << "Forcing collection of SoftReferences for " << PrettySize(alloc_size) 1774 ptr = TryToAllocate<true, true>(self, allocator, alloc_size, bytes_allocated, usable_size, 1795 ptr = TryToAllocate<true, true>(self, allocator, alloc_size, bytes_allocated, 1828 if (!IsOutOfMemoryOnAllocation<false>(allocator, alloc_size)) { 1845 ptr = TryToAllocate<true, true>(self, allocator, alloc_size, bytes_allocate 1684 AllocateInternalWithGc(Thread* self, AllocatorType allocator, bool instrumented, size_t alloc_size, size_t* bytes_allocated, size_t* usable_size, size_t* bytes_tl_bulk_allocated, mirror::Class** klass) argument 2371 size_t alloc_size = RoundUp(obj_size, kObjectAlignment); local [all...] |
H A D | heap.h | 882 size_t alloc_size, 892 ALWAYS_INLINE bool IsOutOfMemoryOnAllocation(AllocatorType allocator_type, size_t alloc_size);
|
/art/runtime/gc/space/ |
H A D | space_test.h | 182 size_t alloc_size; local 184 alloc_size = object_size; 186 alloc_size = test_rand(&rand_seed) % static_cast<size_t>(-object_size); 189 if (alloc_size < size_of_zero_length_byte_array) { 190 alloc_size = size_of_zero_length_byte_array; 198 object.Assign(Alloc(space, self, alloc_size, &bytes_allocated, nullptr, 201 object.Assign(AllocWithGrowth(space, self, alloc_size, &bytes_allocated, nullptr,
|
H A D | large_object_space.cc | 213 size_t alloc_size = it->second.mem_map->BaseSize(); local 215 *usable_size = alloc_size; 217 return alloc_size; 390 size_t alloc_size = cur_info->ByteSize(); local 392 uint8_t* byte_end = byte_start + alloc_size; 393 callback(byte_start, byte_end, alloc_size, arg); 472 size_t alloc_size = info->ByteSize(); local 474 *usable_size = alloc_size; 476 return alloc_size;
|
H A D | large_object_space_test.cc | 123 size_t alloc_size, bytes_tl_bulk_allocated; local 124 mirror::Object* ptr = los_->Alloc(self, size_, &alloc_size, nullptr,
|
H A D | region_space.h | 213 void AddLiveBytes(mirror::Object* ref, size_t alloc_size) { argument 215 reg->AddLiveBytes(alloc_size);
|
/art/runtime/gc/collector/ |
H A D | concurrent_copying.cc | 1389 size_t alloc_size = RoundUp(obj_size, space::RegionSpace::kAlignment); local 1390 collector_->region_space_->AddLiveBytes(ref, alloc_size); 1733 mirror::Object* ConcurrentCopying::AllocateInSkippedBlock(size_t alloc_size) { argument 1735 CHECK_ALIGNED(alloc_size, space::RegionSpace::kAlignment); 1739 auto it = skipped_blocks_map_.lower_bound(alloc_size); 1746 CHECK_GE(byte_size, alloc_size); 1747 if (byte_size > alloc_size && byte_size - alloc_size < min_object_size) { 1749 it = skipped_blocks_map_.lower_bound(alloc_size + min_object_size); 1754 CHECK_ALIGNED(it->first - alloc_size, spac [all...] |
H A D | mark_compact.cc | 75 const size_t alloc_size = RoundUp(obj->SizeOf(), space::BumpPointerSpace::kAlignment); local 85 bump_pointer_ += alloc_size;
|
H A D | concurrent_copying.h | 167 mirror::Object* AllocateInSkippedBlock(size_t alloc_size)
|
/art/runtime/mirror/ |
H A D | string-inl.h | 167 size_t alloc_size = RoundUp(size, kObjectAlignment); local 186 heap->AllocObjectWithAllocator<kIsInstrumented, true>(self, string_class, alloc_size,
|