Searched defs:alloc_size (Results 1 - 9 of 9) sorted by relevance

/art/runtime/gc/
H A Dheap-inl.h250 size_t alloc_size,
257 UNLIKELY(IsOutOfMemoryOnAllocation(allocator_type, alloc_size, kGrow))) {
264 alloc_size = RoundUp(alloc_size, space::BumpPointerSpace::kAlignment);
265 ret = bump_pointer_space_->AllocNonvirtual(alloc_size);
267 *bytes_allocated = alloc_size;
268 *usable_size = alloc_size;
269 *bytes_tl_bulk_allocated = alloc_size;
276 size_t max_bytes_tl_bulk_allocated = rosalloc_space_->MaxBytesBulkAllocatedFor(alloc_size);
282 ret = rosalloc_space_->Alloc(self, alloc_size, bytes_allocate
248 TryToAllocate(Thread* self, AllocatorType allocator_type, size_t alloc_size, size_t* bytes_allocated, size_t* usable_size, size_t* bytes_tl_bulk_allocated) argument
393 IsOutOfMemoryOnAllocation(AllocatorType allocator_type, size_t alloc_size, bool grow) argument
[all...]
H A Dheap.cc1574 size_t alloc_size,
1596 mirror::Object* ptr = TryToAllocate<true, false>(self, allocator, alloc_size, bytes_allocated,
1611 mirror::Object* ptr = TryToAllocate<true, false>(self, allocator, alloc_size, bytes_allocated,
1632 mirror::Object* ptr = TryToAllocate<true, false>(self, allocator, alloc_size, bytes_allocated,
1641 mirror::Object* ptr = TryToAllocate<true, true>(self, allocator, alloc_size, bytes_allocated,
1650 VLOG(gc) << "Forcing collection of SoftReferences for " << PrettySize(alloc_size)
1660 ptr = TryToAllocate<true, true>(self, allocator, alloc_size, bytes_allocated, usable_size,
1681 ptr = TryToAllocate<true, true>(self, allocator, alloc_size, bytes_allocated,
1718 if (!IsOutOfMemoryOnAllocation(allocator, alloc_size, /*grow*/ false)) {
1735 ptr = TryToAllocate<true, true>(self, allocator, alloc_size, bytes_allocate
1571 AllocateInternalWithGc(Thread* self, AllocatorType allocator, bool instrumented, size_t alloc_size, size_t* bytes_allocated, size_t* usable_size, size_t* bytes_tl_bulk_allocated, ObjPtr<mirror::Class>* klass) argument
2233 size_t alloc_size = RoundUp(obj_size, kObjectAlignment); local
4020 AllocWithNewTLAB(Thread* self, size_t alloc_size, bool grow, size_t* bytes_allocated, size_t* usable_size, size_t* bytes_tl_bulk_allocated) argument
4108 VlogHeapGrowth(size_t max_allowed_footprint, size_t new_footprint, size_t alloc_size) argument
[all...]
/art/runtime/gc/space/
H A Dlarge_object_space_test.cc133 size_t alloc_size, bytes_tl_bulk_allocated; local
134 mirror::Object* ptr = los_->Alloc(self, size_, &alloc_size, nullptr,
H A Dspace_test.h179 size_t alloc_size; local
181 alloc_size = object_size;
183 alloc_size = test_rand(&rand_seed) % static_cast<size_t>(-object_size);
186 if (alloc_size < size_of_zero_length_byte_array) {
187 alloc_size = size_of_zero_length_byte_array;
195 object.Assign(Alloc(space, self, alloc_size, &bytes_allocated, nullptr,
198 object.Assign(AllocWithGrowth(space, self, alloc_size, &bytes_allocated, nullptr,
H A Dlarge_object_space.cc208 size_t alloc_size = it->second.mem_map->BaseSize(); local
210 *usable_size = alloc_size;
212 return alloc_size;
385 size_t alloc_size = cur_info->ByteSize(); local
387 uint8_t* byte_end = byte_start + alloc_size;
388 callback(byte_start, byte_end, alloc_size, arg);
467 size_t alloc_size = info->ByteSize(); local
469 *usable_size = alloc_size;
471 return alloc_size;
H A Dregion_space.h262 void AddLiveBytes(mirror::Object* ref, size_t alloc_size) { argument
264 reg->AddLiveBytes(alloc_size);
/art/runtime/mirror/
H A Dstring-inl.h227 size_t alloc_size = RoundUp(size, kObjectAlignment); local
247 heap->AllocObjectWithAllocator<kIsInstrumented, true>(self, string_class, alloc_size,
/art/runtime/gc/collector/
H A Dmark_compact.cc82 const size_t alloc_size = RoundUp(obj->SizeOf(), space::BumpPointerSpace::kAlignment); local
92 bump_pointer_ += alloc_size;
H A Dconcurrent_copying.cc1563 size_t alloc_size = RoundUp(obj_size, space::RegionSpace::kAlignment); local
1564 region_space_->AddLiveBytes(to_ref, alloc_size);
2280 mirror::Object* ConcurrentCopying::AllocateInSkippedBlock(size_t alloc_size) { argument
2282 CHECK_ALIGNED(alloc_size, space::RegionSpace::kAlignment);
2289 auto it = skipped_blocks_map_.lower_bound(alloc_size);
2295 CHECK_GE(byte_size, alloc_size);
2296 if (byte_size > alloc_size && byte_size - alloc_size < min_object_size) {
2298 it = skipped_blocks_map_.lower_bound(alloc_size + min_object_size);
2303 CHECK_ALIGNED(it->first - alloc_size, spac
[all...]

Completed in 95 milliseconds