/art/runtime/mirror/ |
H A D | stack_trace_element.cc | 43 StackTraceElement* StackTraceElement::Alloc(Thread* self, function in class:art::mirror::StackTraceElement
|
H A D | class_ext.cc | 77 ObjectArray<DexCache>::Alloc(self, 104 ClassExt* ClassExt::Alloc(Thread* self) { function in class:art::mirror::ClassExt
|
H A D | object_array-inl.h | 40 inline ObjectArray<T>* ObjectArray<T>::Alloc(Thread* self, function in class:art::mirror::ObjectArray 43 Array* array = Array::Alloc<true>(self, 57 inline ObjectArray<T>* ObjectArray<T>::Alloc(Thread* self, function in class:art::mirror::ObjectArray 60 return Alloc(self, 357 ObjectArray<T>* new_array = Alloc(self, GetClass(), new_length, allocator_type);
|
H A D | string-inl.h | 213 inline String* String::Alloc(Thread* self, int32_t utf16_length_with_flag, function in class:art::mirror::String 255 return Alloc<kIsInstrumented>(self, length_with_flag, allocator_type, visitor); 268 String* string = Alloc<kIsInstrumented>(self, length_with_flag, allocator_type, visitor); 282 String* new_string = Alloc<kIsInstrumented>(self, length_with_flag, allocator_type, visitor); 294 String* new_string = Alloc<kIsInstrumented>(self, length_with_flag, allocator_type, visitor);
|
H A D | array-inl.h | 154 inline Array* Array::Alloc(Thread* self, function in class:art::mirror::Array 214 Handle<PrimitiveArray<T>> arr(hs.NewHandle(PrimitiveArray<T>::Alloc(self, length))); 223 inline PrimitiveArray<T>* PrimitiveArray<T>::Alloc(Thread* self, size_t length) { function in class:art::mirror::PrimitiveArray 224 Array* raw_array = Array::Alloc<true>(self,
|
H A D | class-inl.h | 783 inline ObjPtr<Object> Class::Alloc(Thread* self, gc::AllocatorType allocator_type) { function in class:art::mirror::Class 808 return Alloc<true>(self, Runtime::Current()->GetHeap()->GetCurrentAllocator()); 812 return Alloc<true>(self, Runtime::Current()->GetHeap()->GetCurrentNonMovingAllocator());
|
/art/runtime/base/ |
H A D | arena_bit_vector.cc | 56 void* storage = allocator->template Alloc<ArenaBitVectorAllocator>(kind); 65 virtual void* Alloc(size_t size) { function in class:art::FINAL 66 return allocator_->Alloc(size, this->Kind());
|
H A D | scoped_arena_allocator.h | 95 void* Alloc(size_t bytes, ArenaAllocKind kind) ALWAYS_INLINE { 151 void* Alloc(size_t bytes, ArenaAllocKind kind = kArenaAllocMisc) ALWAYS_INLINE { 153 return arena_stack_->Alloc(bytes, kind); 157 T* Alloc(ArenaAllocKind kind = kArenaAllocMisc) { function in class:art::ScopedArenaAllocator 163 return static_cast<T*>(Alloc(length * sizeof(T), kind));
|
H A D | arena_allocator.h | 294 void* Alloc(size_t bytes, ArenaAllocKind kind = kArenaAllocMisc) ALWAYS_INLINE { 355 auto* new_ptr = Alloc(new_size, kind); // Note: Alloc will take care of aligning new_size. 362 T* Alloc(ArenaAllocKind kind = kArenaAllocMisc) { function in class:art::ArenaAllocator 368 return static_cast<T*>(Alloc(length * sizeof(T), kind));
|
/art/runtime/gc/space/ |
H A D | bump_pointer_space-inl.h | 28 inline mirror::Object* BumpPointerSpace::Alloc(Thread*, size_t num_bytes, size_t* bytes_allocated, function in class:art::gc::space::BumpPointerSpace
|
H A D | memory_tool_malloc_space-inl.h | 118 kUseObjSizeForUsable>::Alloc( function in class:art::gc::space::MemoryToolMallocSpace 124 void* obj_with_rdz = S::Alloc(self, num_bytes + 2 * kMemoryToolRedZoneBytes,
|
H A D | zygote_space.cc | 81 mirror::Object* ZygoteSpace::Alloc(Thread*, size_t, size_t*, size_t*, size_t*) { function in class:art::gc::space::ZygoteSpace
|
H A D | region_space-inl.h | 27 inline mirror::Object* RegionSpace::Alloc(Thread* self ATTRIBUTE_UNUSED, function in class:art::gc::space::RegionSpace 43 return Alloc(self, num_bytes, bytes_allocated, usable_size, bytes_tl_bulk_allocated); 55 obj = (kForEvac ? evac_region_ : current_region_)->Alloc(num_bytes, 64 obj = (kForEvac ? evac_region_ : current_region_)->Alloc(num_bytes, 73 obj = r->Alloc(num_bytes, bytes_allocated, usable_size, bytes_tl_bulk_allocated); 94 inline mirror::Object* RegionSpace::Region::Alloc(size_t num_bytes, function in class:art::gc::space::RegionSpace::Region 289 // allocation may use all of 'usable_size' (see mirror::Array::Alloc).
|
H A D | large_object_space.cc | 56 mirror::Object* Alloc(Thread* self, size_t num_bytes, size_t* bytes_allocated, 60 LargeObjectMapSpace::Alloc(self, num_bytes + kMemoryToolRedZoneBytes * 2, bytes_allocated, 137 mirror::Object* LargeObjectMapSpace::Alloc(Thread* self, size_t num_bytes, function in class:art::gc::space::LargeObjectMapSpace 474 mirror::Object* FreeListSpace::Alloc(Thread* self, size_t num_bytes, size_t* bytes_allocated, function in class:art::gc::space::FreeListSpace
|
/art/runtime/ |
H A D | linear_alloc.cc | 31 void* LinearAlloc::Alloc(Thread* self, size_t size) { function in class:art::LinearAlloc 33 return allocator_.Alloc(size);
|
/art/libartbase/base/ |
H A D | allocator.cc | 33 void* Alloc(size_t size) { function in class:art::FINAL 52 void* Alloc(size_t size ATTRIBUTE_UNUSED) { function in class:art::FINAL 53 LOG(FATAL) << "NoopAllocator::Alloc should not be called";
|
/art/runtime/gc/allocator/ |
H A D | rosalloc-inl.h | 31 inline ALWAYS_INLINE void* RosAlloc::Alloc(Thread* self, size_t size, size_t* bytes_allocated, function in class:art::gc::allocator::RosAlloc
|
/art/compiler/utils/ |
H A D | swap_space.cc | 108 void* SwapSpace::Alloc(size_t size) { function in class:art::SwapSpace
|
/art/test/004-ThreadStress/src-art/ |
H A D | Main.java | 47 // -alloc:X .......... frequency of Alloc (double) 129 private final static class Alloc extends Operation { class in class:Main 318 frequencyMap.put(new Alloc(), 0.225); // 45/200 336 frequencyMap.put(new Alloc(), 0.575); // 115/200 379 op = new Alloc();
|
/art/dt_fd_forward/ |
H A D | dt_fd_forward.cc | 518 out = reinterpret_cast<jbyte*>(transport_->Alloc(rem)); 634 void* FdForwardTransport::Alloc(size_t s) { function in class:dt_fd_forward::FdForwardTransport 644 *err = reinterpret_cast<char*>(Alloc(data.size() + 1)); 705 *actual_address = reinterpret_cast<char*>(AsFdForward(env)->Alloc(strlen(address) + 1));
|