/art/runtime/gc/space/ |
H A D | dlmalloc_space-inl.h | 30 size_t* usable_size) { 34 obj = AllocWithoutGrowthLocked(self, num_bytes, bytes_allocated, usable_size); 43 inline size_t DlMallocSpace::AllocationSizeNonvirtual(mirror::Object* obj, size_t* usable_size) { argument 46 if (usable_size != nullptr) { 47 *usable_size = size; 54 size_t* usable_size) { 61 size_t allocation_size = AllocationSizeNonvirtual(result, usable_size); 28 AllocNonvirtual(Thread* self, size_t num_bytes, size_t* bytes_allocated, size_t* usable_size) argument 52 AllocWithoutGrowthLocked(Thread* , size_t num_bytes, size_t* bytes_allocated, size_t* usable_size) argument
|
H A D | bump_pointer_space-inl.h | 27 size_t* usable_size) { 32 if (usable_size != nullptr) { 33 *usable_size = num_bytes; 41 size_t* usable_size) { 54 if (UNLIKELY(usable_size != nullptr)) { 55 *usable_size = num_bytes; 84 inline size_t BumpPointerSpace::AllocationSizeNonvirtual(mirror::Object* obj, size_t* usable_size) 87 if (usable_size != nullptr) { 88 *usable_size = RoundUp(num_bytes, kAlignment); 26 Alloc(Thread*, size_t num_bytes, size_t* bytes_allocated, size_t* usable_size) argument 39 AllocThreadUnsafe(Thread* self, size_t num_bytes, size_t* bytes_allocated, size_t* usable_size) argument
|
H A D | rosalloc_space-inl.h | 28 inline size_t RosAllocSpace::AllocationSizeNonvirtual(mirror::Object* obj, size_t* usable_size) { argument 43 if (usable_size != nullptr) { 44 *usable_size = size_by_size; 51 size_t* bytes_allocated, size_t* usable_size) { 66 if (usable_size != nullptr) { 67 *usable_size = rosalloc_size; 50 AllocCommon(Thread* self, size_t num_bytes, size_t* bytes_allocated, size_t* usable_size) argument
|
H A D | rosalloc_space.h | 50 size_t* usable_size) OVERRIDE LOCKS_EXCLUDED(lock_); 52 size_t* usable_size) OVERRIDE { 53 return AllocNonvirtual(self, num_bytes, bytes_allocated, usable_size); 56 size_t* usable_size) 58 return AllocNonvirtualThreadUnsafe(self, num_bytes, bytes_allocated, usable_size); 60 size_t AllocationSize(mirror::Object* obj, size_t* usable_size) OVERRIDE { 61 return AllocationSizeNonvirtual(obj, usable_size); 69 size_t* usable_size) { 71 return AllocCommon(self, num_bytes, bytes_allocated, usable_size); 74 size_t* bytes_allocated, size_t* usable_size) { 68 AllocNonvirtual(Thread* self, size_t num_bytes, size_t* bytes_allocated, size_t* usable_size) argument 73 AllocNonvirtualThreadUnsafe(Thread* self, size_t num_bytes, size_t* bytes_allocated, size_t* usable_size) argument [all...] |
H A D | valgrind_malloc_space-inl.h | 35 size_t* usable_size) { 37 bytes_allocated, usable_size); 52 size_t* usable_size) { 54 usable_size); 67 size_t ValgrindMallocSpace<S, A>::AllocationSize(mirror::Object* obj, size_t* usable_size) { argument 69 reinterpret_cast<byte*>(obj) - kValgrindRedZoneBytes), usable_size); 78 size_t usable_size = 0; local 79 AllocationSize(ptr, &usable_size); 80 VALGRIND_MAKE_MEM_UNDEFINED(obj_with_rdz, usable_size); 33 AllocWithGrowth(Thread* self, size_t num_bytes, size_t* bytes_allocated, size_t* usable_size) argument 50 Alloc(Thread* self, size_t num_bytes, size_t* bytes_allocated, size_t* usable_size) argument
|
H A D | zygote_space.cc | 79 size_t* usable_size) { 84 size_t ZygoteSpace::AllocationSize(mirror::Object* obj, size_t* usable_size) { argument 78 Alloc(Thread* self, size_t num_bytes, size_t* bytes_allocated, size_t* usable_size) argument
|
H A D | dlmalloc_space.cc | 129 size_t* bytes_allocated, size_t* usable_size) { 137 result = AllocWithoutGrowthLocked(self, num_bytes, bytes_allocated, usable_size); 128 AllocWithGrowth(Thread* self, size_t num_bytes, size_t* bytes_allocated, size_t* usable_size) argument
|
H A D | rosalloc_space.cc | 147 size_t* bytes_allocated, size_t* usable_size) { 155 result = AllocCommon(self, num_bytes, bytes_allocated, usable_size); 146 AllocWithGrowth(Thread* self, size_t num_bytes, size_t* bytes_allocated, size_t* usable_size) argument
|
H A D | space_test.h | 64 size_t* bytes_allocated, size_t* usable_size) 68 mirror::Object* obj = alloc_space->Alloc(self, bytes, bytes_allocated, usable_size); 76 size_t* bytes_allocated, size_t* usable_size) 80 mirror::Object* obj = alloc_space->AllocWithGrowth(self, bytes, bytes_allocated, usable_size); 348 size_t allocation_size, usable_size; local 351 &usable_size); 355 EXPECT_EQ(usable_size, computed_usable_size); 363 size_t allocation_size, usable_size; local 364 lots_of_objects[i] = AllocWithGrowth(space, self, 1024, &allocation_size, &usable_size); 368 EXPECT_EQ(usable_size, computed_usable_siz [all...] |
H A D | large_object_space.cc | 41 size_t* usable_size) OVERRIDE { 44 usable_size); 50 if (usable_size != nullptr) { 51 *usable_size = num_bytes; // Since we have redzones, shrink the usable size. 56 virtual size_t AllocationSize(mirror::Object* obj, size_t* usable_size) OVERRIDE { 59 return LargeObjectMapSpace::AllocationSize(object_with_rdz, usable_size); 111 size_t* bytes_allocated, size_t* usable_size) { 131 if (usable_size != nullptr) { 132 *usable_size = allocation_size; 157 size_t LargeObjectMapSpace::AllocationSize(mirror::Object* obj, size_t* usable_size) { argument 110 Alloc(Thread* self, size_t num_bytes, size_t* bytes_allocated, size_t* usable_size) argument 398 AllocationSize(mirror::Object* obj, size_t* usable_size) argument 409 Alloc(Thread* self, size_t num_bytes, size_t* bytes_allocated, size_t* usable_size) argument [all...] |
/art/runtime/gc/ |
H A D | heap-inl.h | 58 size_t usable_size; local 75 usable_size = bytes_allocated; 76 pre_fence_visitor(obj, usable_size); 80 &usable_size); 83 obj = AllocateInternalWithGc(self, allocator, byte_count, &bytes_allocated, &usable_size, 97 DCHECK_GT(usable_size, 0u); 117 pre_fence_visitor(obj, usable_size); 123 CHECK_LE(obj->SizeOf(), usable_size); 185 size_t* usable_size) { 198 *usable_size 183 TryToAllocate(Thread* self, AllocatorType allocator_type, size_t alloc_size, size_t* bytes_allocated, size_t* usable_size) argument [all...] |
H A D | heap.cc | 1230 size_t* usable_size, 1250 usable_size); 1264 usable_size); 1284 usable_size); 1293 usable_size); 1310 ptr = TryToAllocate<true, true>(self, allocator, alloc_size, bytes_allocated, usable_size); 1326 usable_size); 1370 usable_size); 1228 AllocateInternalWithGc(Thread* self, AllocatorType allocator, size_t alloc_size, size_t* bytes_allocated, size_t* usable_size, mirror::Class** klass) argument
|