/art/runtime/gc/space/ |
H A D | dlmalloc_space-inl.h | 28 inline mirror::Object* DlMallocSpace::AllocNonvirtual(Thread* self, size_t num_bytes, argument 34 obj = AllocWithoutGrowthLocked(self, num_bytes, bytes_allocated, usable_size); 38 memset(obj, 0, num_bytes); 52 inline mirror::Object* DlMallocSpace::AllocWithoutGrowthLocked(Thread* /*self*/, size_t num_bytes, argument 55 mirror::Object* result = reinterpret_cast<mirror::Object*>(mspace_malloc(mspace_, num_bytes));
|
H A D | bump_pointer_space-inl.h | 26 inline mirror::Object* BumpPointerSpace::Alloc(Thread*, size_t num_bytes, size_t* bytes_allocated, argument 28 num_bytes = RoundUp(num_bytes, kAlignment); 29 mirror::Object* ret = AllocNonvirtual(num_bytes); 31 *bytes_allocated = num_bytes; 33 *usable_size = num_bytes; 39 inline mirror::Object* BumpPointerSpace::AllocThreadUnsafe(Thread* self, size_t num_bytes, argument 43 num_bytes = RoundUp(num_bytes, kAlignment); 45 if (end + num_bytes > growth_end 60 AllocNonvirtualWithoutAccounting(size_t num_bytes) argument 75 AllocNonvirtual(size_t num_bytes) argument 86 size_t num_bytes = obj->SizeOf(); local [all...] |
H A D | rosalloc_space-inl.h | 50 inline mirror::Object* RosAllocSpace::AllocCommon(Thread* self, size_t num_bytes, argument 57 rosalloc_->Alloc<kThreadSafe>(self, num_bytes, &rosalloc_size));
|
H A D | rosalloc_space.h | 49 mirror::Object* AllocWithGrowth(Thread* self, size_t num_bytes, size_t* bytes_allocated, 51 mirror::Object* Alloc(Thread* self, size_t num_bytes, size_t* bytes_allocated, 53 return AllocNonvirtual(self, num_bytes, bytes_allocated, usable_size); 55 mirror::Object* AllocThreadUnsafe(Thread* self, size_t num_bytes, size_t* bytes_allocated, 58 return AllocNonvirtualThreadUnsafe(self, num_bytes, bytes_allocated, usable_size); 68 mirror::Object* AllocNonvirtual(Thread* self, size_t num_bytes, size_t* bytes_allocated, argument 71 return AllocCommon(self, num_bytes, bytes_allocated, usable_size); 73 mirror::Object* AllocNonvirtualThreadUnsafe(Thread* self, size_t num_bytes, argument 76 return AllocCommon<false>(self, num_bytes, bytes_allocated, usable_size); 134 mirror::Object* AllocCommon(Thread* self, size_t num_bytes, size_ [all...] |
H A D | valgrind_malloc_space-inl.h | 33 mirror::Object* ValgrindMallocSpace<S, A>::AllocWithGrowth(Thread* self, size_t num_bytes, argument 36 void* obj_with_rdz = S::AllocWithGrowth(self, num_bytes + 2 * kValgrindRedZoneBytes, 45 VALGRIND_MAKE_MEM_NOACCESS(reinterpret_cast<byte*>(result) + num_bytes, kValgrindRedZoneBytes); 50 mirror::Object* ValgrindMallocSpace<S, A>::Alloc(Thread* self, size_t num_bytes, argument 53 void* obj_with_rdz = S::Alloc(self, num_bytes + 2 * kValgrindRedZoneBytes, bytes_allocated, 62 VALGRIND_MAKE_MEM_NOACCESS(reinterpret_cast<byte*>(result) + num_bytes, kValgrindRedZoneBytes);
|
H A D | zygote_space.cc | 78 mirror::Object* ZygoteSpace::Alloc(Thread* self, size_t num_bytes, size_t* bytes_allocated, argument
|
H A D | dlmalloc_space.cc | 128 mirror::Object* DlMallocSpace::AllocWithGrowth(Thread* self, size_t num_bytes, argument 137 result = AllocWithoutGrowthLocked(self, num_bytes, bytes_allocated, usable_size); 144 memset(result, 0, num_bytes); 247 void DlMallocSpace::Walk(void(*callback)(void *start, void *end, size_t num_bytes, void* callback_arg), argument
|
H A D | rosalloc_space.cc | 146 mirror::Object* RosAllocSpace::AllocWithGrowth(Thread* self, size_t num_bytes, argument 155 result = AllocCommon(self, num_bytes, bytes_allocated, usable_size); 250 void RosAllocSpace::Walk(void(*callback)(void *start, void *end, size_t num_bytes, void* callback_arg), argument 291 void (*callback)(void *start, void *end, size_t num_bytes, void* callback_arg), 308 void RosAllocSpace::InspectAllRosAlloc(void (*callback)(void *start, void *end, size_t num_bytes, void* callback_arg),
|
H A D | large_object_space.cc | 40 virtual mirror::Object* Alloc(Thread* self, size_t num_bytes, size_t* bytes_allocated, 43 LargeObjectMapSpace::Alloc(self, num_bytes + kValgrindRedZoneBytes * 2, bytes_allocated, 48 VALGRIND_MAKE_MEM_NOACCESS(reinterpret_cast<byte*>(object_without_rdz) + num_bytes, 51 *usable_size = num_bytes; // Since we have redzones, shrink the usable size. 110 mirror::Object* LargeObjectMapSpace::Alloc(Thread* self, size_t num_bytes, argument 113 MemMap* mem_map = MemMap::MapAnonymous("large object space allocation", NULL, num_bytes, 409 mirror::Object* FreeListSpace::Alloc(Thread* self, size_t num_bytes, size_t* bytes_allocated, argument 412 const size_t allocation_size = RoundUp(num_bytes, kAlignment); 417 // Find the smallest chunk at least num_bytes in size.
|
/art/runtime/mirror/ |
H A D | art_field.cc | 45 void ArtField::SetOffset(MemberOffset num_bytes) { argument 51 DCHECK_ALIGNED(num_bytes.Uint32Value(), 8); 55 SetField32<false>(OFFSET_OF_OBJECT_MEMBER(ArtField, offset_), num_bytes.Uint32Value());
|
H A D | object.cc | 69 size_t num_bytes) { 75 memcpy(dst_bytes + offset, src_bytes + offset, num_bytes - offset); 103 explicit CopyObjectVisitor(Thread* self, Handle<Object>* orig, size_t num_bytes) argument 104 : self_(self), orig_(orig), num_bytes_(num_bytes) { 125 size_t num_bytes = SizeOf(); local 129 CopyObjectVisitor visitor(self, &this_object, num_bytes); 131 copy = heap->AllocObject<true>(self, GetClass(), num_bytes, visitor); 133 copy = heap->AllocNonMovableObject<true>(self, GetClass(), num_bytes, visitor); 68 CopyObject(Thread* self, mirror::Object* dest, mirror::Object* src, size_t num_bytes) argument
|