/art/runtime/gc/space/ |
H A D | bump_pointer_space-inl.h | 27 inline mirror::Object* BumpPointerSpace::Alloc(Thread*, size_t num_bytes, size_t* bytes_allocated, argument 30 num_bytes = RoundUp(num_bytes, kAlignment); 31 mirror::Object* ret = AllocNonvirtual(num_bytes); 33 *bytes_allocated = num_bytes; 35 *usable_size = num_bytes; 37 *bytes_tl_bulk_allocated = num_bytes; 42 inline mirror::Object* BumpPointerSpace::AllocThreadUnsafe(Thread* self, size_t num_bytes, argument 47 num_bytes = RoundUp(num_bytes, kAlignmen 65 AllocNonvirtualWithoutAccounting(size_t num_bytes) argument 80 AllocNonvirtual(size_t num_bytes) argument 91 size_t num_bytes = obj->SizeOf(); local [all...] |
H A D | region_space-inl.h | 26 inline mirror::Object* RegionSpace::Alloc(Thread*, size_t num_bytes, size_t* bytes_allocated, argument 29 num_bytes = RoundUp(num_bytes, kAlignment); 30 return AllocNonvirtual<false>(num_bytes, bytes_allocated, usable_size, 34 inline mirror::Object* RegionSpace::AllocThreadUnsafe(Thread* self, size_t num_bytes, argument 39 return Alloc(self, num_bytes, bytes_allocated, usable_size, bytes_tl_bulk_allocated); 43 inline mirror::Object* RegionSpace::AllocNonvirtual(size_t num_bytes, size_t* bytes_allocated, argument 46 DCHECK_ALIGNED(num_bytes, kAlignment); 48 if (LIKELY(num_bytes <= kRegionSize)) { 51 obj = current_region_->Alloc(num_bytes, bytes_allocate 114 Alloc(size_t num_bytes, size_t* bytes_allocated, size_t* usable_size, size_t* bytes_tl_bulk_allocated) argument 142 size_t num_bytes = obj->SizeOf(); local 265 AllocLarge(size_t num_bytes, size_t* bytes_allocated, size_t* usable_size, size_t* bytes_tl_bulk_allocated) argument [all...] |
H A D | dlmalloc_space-inl.h | 28 inline mirror::Object* DlMallocSpace::AllocNonvirtual(Thread* self, size_t num_bytes, argument 35 obj = AllocWithoutGrowthLocked(self, num_bytes, bytes_allocated, usable_size, 40 memset(obj, 0, num_bytes); 55 Thread* /*self*/, size_t num_bytes, 59 mirror::Object* result = reinterpret_cast<mirror::Object*>(mspace_malloc(mspace_, num_bytes)); 54 AllocWithoutGrowthLocked( Thread* , size_t num_bytes, size_t* bytes_allocated, size_t* usable_size, size_t* bytes_tl_bulk_allocated) argument
|
H A D | rosalloc_space-inl.h | 64 inline mirror::Object* RosAllocSpace::AllocCommon(Thread* self, size_t num_bytes, argument 74 rosalloc_->Alloc<kThreadSafe>(self, num_bytes, &rosalloc_bytes_allocated, 94 inline bool RosAllocSpace::CanAllocThreadLocal(Thread* self, size_t num_bytes) { argument 95 return rosalloc_->CanAllocFromThreadLocalRun(self, num_bytes); 98 inline mirror::Object* RosAllocSpace::AllocThreadLocal(Thread* self, size_t num_bytes, argument 102 rosalloc_->AllocFromThreadLocalRun(self, num_bytes, bytes_allocated)); 105 inline size_t RosAllocSpace::MaxBytesBulkAllocatedForNonvirtual(size_t num_bytes) { argument 106 return rosalloc_->MaxBytesBulkAllocatedFor(num_bytes);
|
H A D | rosalloc_space.h | 49 mirror::Object* AllocWithGrowth(Thread* self, size_t num_bytes, size_t* bytes_allocated, 52 mirror::Object* Alloc(Thread* self, size_t num_bytes, size_t* bytes_allocated, 54 return AllocNonvirtual(self, num_bytes, bytes_allocated, usable_size, 57 mirror::Object* AllocThreadUnsafe(Thread* self, size_t num_bytes, size_t* bytes_allocated, 60 return AllocNonvirtualThreadUnsafe(self, num_bytes, bytes_allocated, usable_size, 71 mirror::Object* AllocNonvirtual(Thread* self, size_t num_bytes, size_t* bytes_allocated, argument 74 return AllocCommon(self, num_bytes, bytes_allocated, usable_size, 77 mirror::Object* AllocNonvirtualThreadUnsafe(Thread* self, size_t num_bytes, argument 81 return AllocCommon<false>(self, num_bytes, bytes_allocated, usable_size, 87 ALWAYS_INLINE bool CanAllocThreadLocal(Thread* self, size_t num_bytes); [all...] |
H A D | memory_tool_malloc_space-inl.h | 31 inline mirror::Object* AdjustForValgrind(void* obj_with_rdz, size_t num_bytes, argument 47 *usable_size_out = num_bytes; 61 MEMORY_TOOL_MAKE_DEFINED(result, num_bytes); 66 // and DlMalloc (allocation_size = (usable_size == num_bytes) + 4, 4 is management) 67 MEMORY_TOOL_MAKE_NOACCESS(reinterpret_cast<uint8_t*>(result) + num_bytes, 68 usable_size - (num_bytes + kMemoryToolRedZoneBytes)); 88 Thread* self, size_t num_bytes, size_t* bytes_allocated_out, size_t* usable_size_out, 93 void* obj_with_rdz = S::AllocWithGrowth(self, num_bytes + 2 * kMemoryToolRedZoneBytes, 101 obj_with_rdz, num_bytes, 117 Thread* self, size_t num_bytes, size_ 87 AllocWithGrowth( Thread* self, size_t num_bytes, size_t* bytes_allocated_out, size_t* usable_size_out, size_t* bytes_tl_bulk_allocated_out) argument 116 Alloc( Thread* self, size_t num_bytes, size_t* bytes_allocated_out, size_t* usable_size_out, size_t* bytes_tl_bulk_allocated_out) argument 144 AllocThreadUnsafe( Thread* self, size_t num_bytes, size_t* bytes_allocated_out, size_t* usable_size_out, size_t* bytes_tl_bulk_allocated_out) argument 255 MaxBytesBulkAllocatedFor(size_t num_bytes) argument [all...] |
H A D | memory_tool_malloc_space.h | 34 mirror::Object* AllocWithGrowth(Thread* self, size_t num_bytes, size_t* bytes_allocated, 37 mirror::Object* Alloc(Thread* self, size_t num_bytes, size_t* bytes_allocated, 39 mirror::Object* AllocThreadUnsafe(Thread* self, size_t num_bytes, size_t* bytes_allocated, 53 size_t MaxBytesBulkAllocatedFor(size_t num_bytes) OVERRIDE;
|
H A D | dlmalloc_space.h | 50 virtual mirror::Object* AllocWithGrowth(Thread* self, size_t num_bytes, size_t* bytes_allocated, 55 virtual mirror::Object* Alloc(Thread* self, size_t num_bytes, size_t* bytes_allocated, 58 return AllocNonvirtual(self, num_bytes, bytes_allocated, usable_size, 74 size_t MaxBytesBulkAllocatedFor(size_t num_bytes) OVERRIDE { 75 return num_bytes; 87 mirror::Object* AllocNonvirtual(Thread* self, size_t num_bytes, size_t* bytes_allocated, 106 // in use, indicated by num_bytes equaling zero. 147 mirror::Object* AllocWithoutGrowthLocked(Thread* self, size_t num_bytes, size_t* bytes_allocated,
|
H A D | malloc_space.h | 49 typedef void(*WalkCallback)(void *start, void *end, size_t num_bytes, void* callback_arg); 55 // Allocate num_bytes allowing the underlying space to grow. 56 virtual mirror::Object* AllocWithGrowth(Thread* self, size_t num_bytes, 59 // Allocate num_bytes without allowing the underlying space to grow. 60 virtual mirror::Object* Alloc(Thread* self, size_t num_bytes, size_t* bytes_allocated, 73 virtual size_t MaxBytesBulkAllocatedFor(size_t num_bytes) = 0; 87 // in use, indicated by num_bytes equaling zero.
|
H A D | bump_pointer_space.h | 36 typedef void(*WalkCallback)(void *start, void *end, size_t num_bytes, void* callback_arg); 48 // Allocate num_bytes, returns null if the space is full. 49 mirror::Object* Alloc(Thread* self, size_t num_bytes, size_t* bytes_allocated, 52 mirror::Object* AllocThreadUnsafe(Thread* self, size_t num_bytes, size_t* bytes_allocated, 56 mirror::Object* AllocNonvirtual(size_t num_bytes); 57 mirror::Object* AllocNonvirtualWithoutAccounting(size_t num_bytes);
|
H A D | region_space.h | 32 typedef void(*WalkCallback)(void *start, void *end, size_t num_bytes, void* callback_arg); 43 // Allocate num_bytes, returns null if the space is full. 44 mirror::Object* Alloc(Thread* self, size_t num_bytes, size_t* bytes_allocated, 48 mirror::Object* AllocThreadUnsafe(Thread* self, size_t num_bytes, size_t* bytes_allocated, 53 ALWAYS_INLINE mirror::Object* AllocNonvirtual(size_t num_bytes, size_t* bytes_allocated, 59 mirror::Object* AllocLarge(size_t num_bytes, size_t* bytes_allocated, size_t* usable_size, 275 ALWAYS_INLINE mirror::Object* Alloc(size_t num_bytes, size_t* bytes_allocated, 454 void RecordThreadLocalAllocations(size_t num_objects, size_t num_bytes) { argument 459 top_ = begin_ + num_bytes;
|
H A D | zygote_space.h | 48 mirror::Object* Alloc(Thread* self, size_t num_bytes, size_t* bytes_allocated,
|
H A D | rosalloc_space.cc | 154 mirror::Object* RosAllocSpace::AllocWithGrowth(Thread* self, size_t num_bytes, argument 164 result = AllocCommon(self, num_bytes, bytes_allocated, usable_size, 261 void RosAllocSpace::Walk(void(*callback)(void *start, void *end, size_t num_bytes, void* callback_arg), argument 302 void (*callback)(void *start, void *end, size_t num_bytes, void* callback_arg), 315 void RosAllocSpace::InspectAllRosAlloc(void (*callback)(void *start, void *end, size_t num_bytes, void* callback_arg),
|
H A D | space.h | 202 // Allocate num_bytes without allowing growth. If the allocation 204 // actually allocated bytes which is >= num_bytes. 215 virtual mirror::Object* Alloc(Thread* self, size_t num_bytes, size_t* bytes_allocated, 219 virtual mirror::Object* AllocThreadUnsafe(Thread* self, size_t num_bytes, size_t* bytes_allocated, 223 return Alloc(self, num_bytes, bytes_allocated, usable_size, bytes_tl_bulk_allocated);
|
H A D | dlmalloc_space.cc | 128 mirror::Object* DlMallocSpace::AllocWithGrowth(Thread* self, size_t num_bytes, argument 138 result = AllocWithoutGrowthLocked(self, num_bytes, bytes_allocated, usable_size, 146 memset(result, 0, num_bytes); 234 void DlMallocSpace::Walk(void(*callback)(void *start, void *end, size_t num_bytes, void* callback_arg), argument
|
H A D | large_object_space.cc | 52 mirror::Object* Alloc(Thread* self, size_t num_bytes, size_t* bytes_allocated, 56 LargeObjectMapSpace::Alloc(self, num_bytes + kMemoryToolRedZoneBytes * 2, bytes_allocated, 62 reinterpret_cast<uint8_t*>(object_without_rdz) + num_bytes, 65 *usable_size = num_bytes; // Since we have redzones, shrink the usable size. 133 mirror::Object* LargeObjectMapSpace::Alloc(Thread* self, size_t num_bytes, argument 137 MemMap* mem_map = MemMap::MapAnonymous("large object space allocation", nullptr, num_bytes, 479 mirror::Object* FreeListSpace::Alloc(Thread* self, size_t num_bytes, size_t* bytes_allocated, argument 482 const size_t allocation_size = RoundUp(num_bytes, kAlignment); 487 // Find the smallest chunk at least num_bytes in size.
|
H A D | large_object_space.h | 134 mirror::Object* Alloc(Thread* self, size_t num_bytes, size_t* bytes_allocated, 168 mirror::Object* Alloc(Thread* self, size_t num_bytes, size_t* bytes_allocated,
|
/art/compiler/debug/dwarf/ |
H A D | expression.h | 75 void WriteOpPiece(uint32_t num_bytes) { argument 77 PushUleb128(num_bytes); 84 void WriteOpDerefSize(uint8_t num_bytes) { argument 86 PushUint8(num_bytes);
|
H A D | writer.h | 117 void PushData(const uint8_t* ptr, size_t num_bytes) { argument 118 data_->insert(data_->end(), ptr, ptr + num_bytes); 121 void PushData(const char* ptr, size_t num_bytes) { argument 122 data_->insert(data_->end(), ptr, ptr + num_bytes);
|
H A D | debug_info_entry_writer.h | 95 void WriteBlock(Attribute attrib, const uint8_t* ptr, size_t num_bytes) { argument 97 this->PushUleb128(num_bytes); 98 this->PushData(ptr, num_bytes);
|
/art/runtime/ |
H A D | art_field.cc | 37 void ArtField::SetOffset(MemberOffset num_bytes) { argument 43 DCHECK_ALIGNED(num_bytes.Uint32Value(), 8); 47 offset_ = num_bytes.Uint32Value();
|
H A D | art_field.h | 86 void SetOffset(MemberOffset num_bytes) SHARED_REQUIRES(Locks::mutator_lock_);
|
/art/runtime/mirror/ |
H A D | object.cc | 76 size_t num_bytes) { 82 memcpy(dst_bytes + offset, src_bytes + offset, num_bytes - offset); 110 CopyObjectVisitor(Thread* self, Handle<Object>* orig, size_t num_bytes) argument 111 : self_(self), orig_(orig), num_bytes_(num_bytes) { 131 size_t num_bytes = SizeOf(); local 135 CopyObjectVisitor visitor(self, &this_object, num_bytes); 137 copy = heap->AllocObject<true>(self, GetClass(), num_bytes, visitor); 139 copy = heap->AllocNonMovableObject<true>(self, GetClass(), num_bytes, visitor); 75 CopyObject(Thread* self, mirror::Object* dest, mirror::Object* src, size_t num_bytes) argument
|
/art/runtime/gc/collector/ |
H A D | garbage_collector.h | 47 ObjectBytePair(uint64_t num_objects = 0, int64_t num_bytes = 0) 48 : objects(num_objects), bytes(num_bytes) {}
|
/art/runtime/gc/ |
H A D | heap.h | 194 size_t num_bytes, 200 self, klass, num_bytes, GetCurrentAllocator(), pre_fence_visitor); 206 size_t num_bytes, 212 self, klass, num_bytes, GetCurrentNonMovingAllocator(), pre_fence_visitor); 858 size_t num_bytes,
|