/art/runtime/mirror/ |
H A D | stack_trace_element.cc | 42 StackTraceElement* StackTraceElement::Alloc(Thread* self, function in class:art::mirror::StackTraceElement
|
H A D | class_ext.cc | 77 ObjectArray<DexCache>::Alloc(self, 104 ClassExt* ClassExt::Alloc(Thread* self) { function in class:art::mirror::ClassExt
|
H A D | object_array-inl.h | 40 inline ObjectArray<T>* ObjectArray<T>::Alloc(Thread* self, function in class:art::mirror::ObjectArray 43 Array* array = Array::Alloc<true>(self, 57 inline ObjectArray<T>* ObjectArray<T>::Alloc(Thread* self, function in class:art::mirror::ObjectArray 60 return Alloc(self, 357 ObjectArray<T>* new_array = Alloc(self, GetClass(), new_length, allocator_type);
|
H A D | array-inl.h | 155 inline Array* Array::Alloc(Thread* self, function in class:art::mirror::Array 215 Handle<PrimitiveArray<T>> arr(hs.NewHandle(PrimitiveArray<T>::Alloc(self, length))); 224 inline PrimitiveArray<T>* PrimitiveArray<T>::Alloc(Thread* self, size_t length) { function in class:art::mirror::PrimitiveArray 225 Array* raw_array = Array::Alloc<true>(self,
|
H A D | string-inl.h | 209 inline String* String::Alloc(Thread* self, int32_t utf16_length_with_flag, function in class:art::mirror::String 251 return Alloc<kIsInstrumented>(self, length_with_flag, allocator_type, visitor); 263 String* string = Alloc<kIsInstrumented>(self, length_with_flag, allocator_type, visitor); 277 String* new_string = Alloc<kIsInstrumented>(self, length_with_flag, allocator_type, visitor); 289 String* new_string = Alloc<kIsInstrumented>(self, length_with_flag, allocator_type, visitor);
|
H A D | class-inl.h | 737 inline ObjPtr<Object> Class::Alloc(Thread* self, gc::AllocatorType allocator_type) { function in class:art::mirror::Class 762 return Alloc<true>(self, Runtime::Current()->GetHeap()->GetCurrentAllocator()); 766 return Alloc<true>(self, Runtime::Current()->GetHeap()->GetCurrentNonMovingAllocator());
|
/art/runtime/base/ |
H A D | arena_bit_vector.cc | 56 void* storage = arena->template Alloc<ArenaBitVectorAllocator>(kind); 65 virtual void* Alloc(size_t size) { function in class:art::FINAL 66 return arena_->Alloc(size, this->Kind());
|
H A D | allocator.cc | 32 void* Alloc(size_t size) { function in class:art::FINAL 51 void* Alloc(size_t size ATTRIBUTE_UNUSED) { function in class:art::FINAL 52 LOG(FATAL) << "NoopAllocator::Alloc should not be called";
|
H A D | scoped_arena_allocator.h | 89 void* Alloc(size_t bytes, ArenaAllocKind kind) ALWAYS_INLINE { 139 void* addr = arena_stack->Alloc(sizeof(ScopedArenaAllocator), kArenaAllocMisc); 150 void* Alloc(size_t bytes, ArenaAllocKind kind = kArenaAllocMisc) ALWAYS_INLINE { 152 return arena_stack_->Alloc(bytes, kind); 156 T* Alloc(ArenaAllocKind kind = kArenaAllocMisc) { function in class:art::ScopedArenaAllocator 162 return static_cast<T*>(Alloc(length * sizeof(T), kind));
|
H A D | arena_allocator.h | 291 void* Alloc(size_t bytes, ArenaAllocKind kind = kArenaAllocMisc) ALWAYS_INLINE { 351 auto* new_ptr = Alloc(new_size, kind); // Note: Alloc will take care of aligning new_size. 358 T* Alloc(ArenaAllocKind kind = kArenaAllocMisc) { function in class:art::ArenaAllocator 364 return static_cast<T*>(Alloc(length * sizeof(T), kind));
|
/art/runtime/gc/space/ |
H A D | bump_pointer_space-inl.h | 27 inline mirror::Object* BumpPointerSpace::Alloc(Thread*, size_t num_bytes, size_t* bytes_allocated, function in class:art::gc::space::BumpPointerSpace
|
H A D | memory_tool_malloc_space-inl.h | 116 kUseObjSizeForUsable>::Alloc( function in class:art::gc::space::MemoryToolMallocSpace 122 void* obj_with_rdz = S::Alloc(self, num_bytes + 2 * kMemoryToolRedZoneBytes,
|
H A D | region_space-inl.h | 27 inline mirror::Object* RegionSpace::Alloc(Thread*, size_t num_bytes, size_t* bytes_allocated, function in class:art::gc::space::RegionSpace 40 return Alloc(self, num_bytes, bytes_allocated, usable_size, bytes_tl_bulk_allocated); 52 obj = current_region_->Alloc(num_bytes, bytes_allocated, usable_size, 56 obj = evac_region_->Alloc(num_bytes, bytes_allocated, usable_size, 65 obj = current_region_->Alloc(num_bytes, bytes_allocated, usable_size, 68 obj = evac_region_->Alloc(num_bytes, bytes_allocated, usable_size, 85 obj = r->Alloc(num_bytes, bytes_allocated, usable_size, bytes_tl_bulk_allocated); 97 obj = r->Alloc(num_bytes, bytes_allocated, usable_size, bytes_tl_bulk_allocated); 115 inline mirror::Object* RegionSpace::Region::Alloc(size_t num_bytes, size_t* bytes_allocated, function in class:art::gc::space::RegionSpace::Region
|
H A D | zygote_space.cc | 79 mirror::Object* ZygoteSpace::Alloc(Thread*, size_t, size_t*, size_t*, size_t*) { function in class:art::gc::space::ZygoteSpace
|
H A D | large_object_space.cc | 51 mirror::Object* Alloc(Thread* self, size_t num_bytes, size_t* bytes_allocated, 55 LargeObjectMapSpace::Alloc(self, num_bytes + kMemoryToolRedZoneBytes * 2, bytes_allocated, 132 mirror::Object* LargeObjectMapSpace::Alloc(Thread* self, size_t num_bytes, function in class:art::gc::space::LargeObjectMapSpace 469 mirror::Object* FreeListSpace::Alloc(Thread* self, size_t num_bytes, size_t* bytes_allocated, function in class:art::gc::space::FreeListSpace
|
/art/runtime/ |
H A D | linear_alloc.cc | 31 void* LinearAlloc::Alloc(Thread* self, size_t size) { function in class:art::LinearAlloc 33 return allocator_.Alloc(size);
|
/art/runtime/gc/allocator/ |
H A D | rosalloc-inl.h | 31 inline ALWAYS_INLINE void* RosAlloc::Alloc(Thread* self, size_t size, size_t* bytes_allocated, function in class:art::gc::allocator::RosAlloc
|
/art/compiler/utils/ |
H A D | swap_space.cc | 107 void* SwapSpace::Alloc(size_t size) { function in class:art::SwapSpace
|
/art/test/004-ThreadStress/src/ |
H A D | Main.java | 40 // -alloc:X ........ frequency of Alloc 118 private final static class Alloc extends Operation { class in class:Main 247 frequencyMap.put(new Alloc(), 0.25); // 50/200 294 op = new Alloc();
|