/art/runtime/jit/ |
H A D | jit_code_cache.cc | 196 uint8_t* JitCodeCache::CommitCode(Thread* self, argument 205 uint8_t* result = CommitCodeInternal(self, 216 GarbageCollectCache(self); 217 result = CommitCodeInternal(self, 230 bool JitCodeCache::WaitForPotentialCollectionToComplete(Thread* self) { argument 234 lock_cond_.Wait(self); 261 void JitCodeCache::RemoveMethodsIn(Thread* self, const LinearAlloc& alloc) { argument 263 MutexLock mu(self, lock_); 298 void JitCodeCache::ClearGcRootsInInlineCaches(Thread* self) { argument 299 MutexLock mu(self, lock 307 CommitCodeInternal(Thread* self, ArtMethod* method, const uint8_t* vmap_table, size_t frame_size_in_bytes, size_t core_spill_mask, size_t fp_spill_mask, const uint8_t* code, size_t code_size, bool osr) argument 404 ClearData(Thread* self, void* data) argument 409 ReserveData(Thread* self, size_t size, ArtMethod* method) argument 500 NotifyCollectionDone(Thread* self) argument 541 MarkCompiledCodeOnThreadStacks(Thread* self) argument 571 GarbageCollectCache(Thread* self) argument 658 RemoveUnmarkedCode(Thread* self) argument 676 DoCollection(Thread* self, bool collect_profiling_info) argument 838 GarbageCollectCache(self); variable 912 NotifyCompilationOf(ArtMethod* method, Thread* self, bool osr) argument 940 NotifyCompilerUse(ArtMethod* method, Thread* self) argument 949 DoneCompilerUse(ArtMethod* method, Thread* self) argument [all...] |
H A D | jit_code_cache.h | 72 bool NotifyCompilationOf(ArtMethod* method, Thread* self, bool osr) 80 ProfilingInfo* NotifyCompilerUse(ArtMethod* method, Thread* self) 84 void DoneCompiling(ArtMethod* method, Thread* self, bool osr) 88 void DoneCompilerUse(ArtMethod* method, Thread* self) 93 uint8_t* CommitCode(Thread* self, 112 uint8_t* ReserveData(Thread* self, size_t size, ArtMethod* method) 117 void ClearData(Thread* self, void* data) 131 void GarbageCollectCache(Thread* self) 147 void RemoveMethodsIn(Thread* self, const LinearAlloc& alloc) 152 void ClearGcRootsInInlineCaches(Thread* self) REQUIRE [all...] |
/art/runtime/gc/space/ |
H A D | rosalloc_space-inl.h | 64 inline mirror::Object* RosAllocSpace::AllocCommon(Thread* self, size_t num_bytes, argument 71 Locks::mutator_lock_->AssertExclusiveHeld(self); 74 rosalloc_->Alloc<kThreadSafe>(self, num_bytes, &rosalloc_bytes_allocated, 94 inline bool RosAllocSpace::CanAllocThreadLocal(Thread* self, size_t num_bytes) { argument 95 return rosalloc_->CanAllocFromThreadLocalRun(self, num_bytes); 98 inline mirror::Object* RosAllocSpace::AllocThreadLocal(Thread* self, size_t num_bytes, argument 102 rosalloc_->AllocFromThreadLocalRun(self, num_bytes, bytes_allocated));
|
H A D | rosalloc_space.h | 49 mirror::Object* AllocWithGrowth(Thread* self, size_t num_bytes, size_t* bytes_allocated, 52 mirror::Object* Alloc(Thread* self, size_t num_bytes, size_t* bytes_allocated, 54 return AllocNonvirtual(self, num_bytes, bytes_allocated, usable_size, 57 mirror::Object* AllocThreadUnsafe(Thread* self, size_t num_bytes, size_t* bytes_allocated, 60 return AllocNonvirtualThreadUnsafe(self, num_bytes, bytes_allocated, usable_size, 66 size_t Free(Thread* self, mirror::Object* ptr) OVERRIDE 68 size_t FreeList(Thread* self, size_t num_ptrs, mirror::Object** ptrs) OVERRIDE 71 mirror::Object* AllocNonvirtual(Thread* self, size_t num_bytes, size_t* bytes_allocated, argument 74 return AllocCommon(self, num_bytes, bytes_allocated, usable_size, 77 mirror::Object* AllocNonvirtualThreadUnsafe(Thread* self, size_ argument [all...] |
H A D | large_object_space_test.cc | 37 Thread* const self = Thread::Current(); local 55 mirror::Object* obj = los->Alloc(self, request_size, &allocation_size, nullptr, 78 ASSERT_FALSE(los->IsZygoteLargeObject(self, obj)); 80 los->SetAllLargeObjectsAsZygoteObjects(self); 83 ASSERT_TRUE(los->IsZygoteLargeObject(self, obj)); 105 mirror::Object* obj = los->Alloc(self, 100 * MB, &bytes_allocated, nullptr, 121 void Run(Thread* self) { argument 124 mirror::Object* ptr = los_->Alloc(self, size_, &alloc_size, nullptr, 129 los_->Free(self, ptr); 153 Thread* self local [all...] |
/art/runtime/gc/allocator/ |
H A D | rosalloc-inl.h | 31 inline ALWAYS_INLINE void* RosAlloc::Alloc(Thread* self, size_t size, size_t* bytes_allocated, argument 35 return AllocLargeObject(self, size, bytes_allocated, usable_size, 40 m = AllocFromRun(self, size, bytes_allocated, usable_size, bytes_tl_bulk_allocated); 42 m = AllocFromRunThreadUnsafe(self, size, bytes_allocated, usable_size, 59 inline bool RosAlloc::CanAllocFromThreadLocalRun(Thread* self, size_t size) { argument 66 Run* thread_local_run = reinterpret_cast<Run*>(self->GetRosAllocRun(idx)); 69 MutexLock mu(self, *size_bracket_locks_[idx]); 78 inline void* RosAlloc::AllocFromThreadLocalRun(Thread* self, size_t size, argument 86 Run* thread_local_run = reinterpret_cast<Run*>(self->GetRosAllocRun(idx)); 89 MutexLock mu(self, *size_bracket_locks [all...] |
/art/runtime/entrypoints/ |
H A D | entrypoint_utils-inl.h | 83 inline ArtMethod* GetCalleeSaveMethodCaller(Thread* self, Runtime::CalleeSaveType type) 86 self->GetManagedStack()->GetTopQuickFrame(), type, true /* do_caller_check */); 93 Thread* self, bool* slow_path) { 101 DCHECK(self->IsExceptionPending()); 104 DCHECK(!self->IsExceptionPending()); 109 self->ThrowNewException("Ljava/lang/InstantiationError;", PrettyDescriptor(klass).c_str()); 121 StackHandleScope<1> hs(self); 132 if (!Runtime::Current()->GetClassLinker()->EnsureInitialized(self, h_klass, true, true)) { 133 DCHECK(self->IsExceptionPending()); 136 DCHECK(!self 91 CheckObjectAlloc(uint32_t type_idx, ArtMethod* method, Thread* self, bool* slow_path) argument 144 CheckClassInitializedForObjectAlloc(mirror::Class* klass, Thread* self, bool* slow_path) argument 174 AllocObjectFromCode(uint32_t type_idx, ArtMethod* method, Thread* self, gc::AllocatorType allocator_type) argument 196 AllocObjectFromCodeResolved(mirror::Class* klass, Thread* self, gc::AllocatorType allocator_type) argument 219 AllocObjectFromCodeInitialized(mirror::Class* klass, Thread* self, gc::AllocatorType allocator_type) argument 268 AllocArrayFromCode(uint32_t type_idx, int32_t component_count, ArtMethod* method, Thread* self, gc::AllocatorType allocator_type) argument 294 AllocArrayFromCodeResolved(mirror::Class* klass, int32_t component_count, ArtMethod* method, Thread* self, gc::AllocatorType allocator_type) argument 437 FindMethodFromCode(uint32_t method_idx, mirror::Object** this_object, ArtMethod* referrer, Thread* self) argument 725 ResolveVerifyAndClinit(uint32_t type_idx, ArtMethod* referrer, Thread* self, bool can_run_clinit, bool verify_access) argument 764 UnlockJniSynchronizedMethod(jobject locked, Thread* self) argument [all...] |
H A D | entrypoint_utils.h | 49 Thread* self, bool* slow_path) 53 Thread* self, 64 Thread* self, 71 Thread* self, 78 Thread* self, 98 Thread* self, 106 Thread* self, 111 ArtMethod* method, Thread* self, 119 Thread* self, 138 uint32_t field_idx, ArtMethod* referrer, Thread* self, size_ [all...] |
/art/runtime/ |
H A D | monitor.h | 78 static void Notify(Thread* self, mirror::Object* obj) SHARED_REQUIRES(Locks::mutator_lock_) { 79 DoNotify(self, obj, false); 81 static void NotifyAll(Thread* self, mirror::Object* obj) SHARED_REQUIRES(Locks::mutator_lock_) { 82 DoNotify(self, obj, true); 87 static void Wait(Thread* self, mirror::Object* obj, int64_t ms, int32_t ns, 133 static void InflateThinLocked(Thread* self, Handle<mirror::Object> obj, LockWord lock_word, 139 static bool Deflate(Thread* self, mirror::Object* obj) 157 Monitor(Thread* self, Thread* owner, mirror::Object* obj, int32_t hash_code) 159 Monitor(Thread* self, Thread* owner, mirror::Object* obj, int32_t hash_code, MonitorId id) 164 bool Install(Thread* self) [all...] |
H A D | scoped_thread_state_change.h | 36 ScopedThreadStateChange(Thread* self, ThreadState new_thread_state) 38 : self_(self), thread_state_(new_thread_state), expected_has_no_thread_(false) { 45 DCHECK_EQ(self, Thread::Current()); 48 old_thread_state_ = self->GetState(); 183 explicit ScopedObjectAccessAlreadyRunnable(Thread* self) 185 : self_(self), env_(down_cast<JNIEnvExt*>(self->GetJniEnv())), 229 explicit ScopedObjectAccessUnchecked(Thread* self) 231 : ScopedObjectAccessAlreadyRunnable(self), tsc_(self, kRunnabl [all...] |
H A D | class_linker-inl.h | 32 inline mirror::Class* ClassLinker::FindSystemClass(Thread* self, const char* descriptor) { argument 33 return FindClass(self, descriptor, ScopedNullHandle<mirror::ClassLoader>()); 36 inline mirror::Class* ClassLinker::FindArrayClass(Thread* self, mirror::Class** element_class) { argument 50 mirror::Class* array_class = FindClass(self, descriptor.c_str(), class_loader); 58 self->AssertPendingException(); 141 inline ArtMethod* ClassLinker::ResolveMethod(Thread* self, argument 148 StackHandleScope<2> hs(self); 189 inline mirror::Object* ClassLinker::AllocObject(Thread* self) { argument 191 self, 196 inline mirror::ObjectArray<T>* ClassLinker::AllocObjectArray(Thread* self, size_ argument 200 AllocClassArray(Thread* self, size_t length) argument 205 AllocStringArray(Thread* self, size_t length) argument 212 AllocIfTable(Thread* self, size_t ifcount) argument [all...] |
/art/runtime/mirror/ |
H A D | string.cc | 82 String* String::AllocFromStrings(Thread* self, Handle<String> string, Handle<String> string2) { argument 87 String* new_string = Alloc<true>(self, length + length2, allocator_type, visitor); 97 String* String::AllocFromUtf16(Thread* self, int32_t utf16_length, const uint16_t* utf16_data_in) { argument 101 String* string = Alloc<true>(self, utf16_length, allocator_type, visitor); 110 String* String::AllocFromModifiedUtf8(Thread* self, const char* utf) { argument 114 return AllocFromModifiedUtf8(self, char_count, utf, byte_count); 117 String* String::AllocFromModifiedUtf8(Thread* self, int32_t utf16_length, const char* utf8_data_in) { argument 118 return AllocFromModifiedUtf8(self, utf16_length, utf8_data_in, strlen(utf8_data_in)); 121 String* String::AllocFromModifiedUtf8(Thread* self, int32_t utf16_length, argument 125 String* string = Alloc<true>(self, utf16_lengt 258 ToCharArray(Thread* self) argument [all...] |
/art/runtime/native/ |
H A D | java_lang_VMClassLoader.cc | 47 Thread* self = soa.Self(); local 49 self->DecodeJObject(WellKnownClasses::java_lang_ExceptionInInitializerError)->AsClass(); 51 self->DecodeJObject(WellKnownClasses::java_lang_IllegalAccessError)->AsClass(); 53 self->DecodeJObject(WellKnownClasses::java_lang_NoClassDefFoundError)->AsClass(); 54 mirror::Class* exception = self->GetException()->GetClass(); 56 self->ThrowNewWrappedException("Ljava/lang/ClassNotFoundException;",
|
/art/runtime/interpreter/ |
H A D | interpreter_common.h | 74 extern JValue ExecuteSwitchImpl(Thread* self, const DexFile::CodeItem* code_item, 79 extern JValue ExecuteGotoImpl(Thread* self, const DexFile::CodeItem* code_item, 83 extern "C" bool ExecuteMterpImpl(Thread* self, const DexFile::CodeItem* code_item, 90 static inline void DoMonitorEnter(Thread* self, 95 StackHandleScope<1> hs(self); 97 h_ref->MonitorEnter(self); 99 frame->GetLockCountData().AddMonitor(self, h_ref.Get()); 104 static inline void DoMonitorExit(Thread* self, 109 StackHandleScope<1> hs(self); 111 h_ref->MonitorExit(self); 208 DoCreateLambda(Thread* self, const Instruction* inst, ShadowFrame& shadow_frame, lambda::ClosureBuilder* closure_builder, lambda::Closure* uninitialized_closure) argument 377 DoCaptureVariable(Thread* self, const Instruction* inst, ShadowFrame& shadow_frame, lambda::ClosureBuilder* closure_builder) argument 478 DoLiberateVariable(Thread* self, const Instruction* inst, size_t captured_variable_index, ShadowFrame& shadow_frame) argument 597 DoInvokeLambda(Thread* self, ShadowFrame& shadow_frame, const Instruction* inst, uint16_t inst_data, JValue* result) argument 628 DoInvoke(Thread* self, ShadowFrame& shadow_frame, const Instruction* inst, uint16_t inst_data, JValue* result) argument 670 DoInvokeVirtualQuick(Thread* self, ShadowFrame& shadow_frame, const Instruction* inst, uint16_t inst_data, JValue* result) argument 941 DoUnboxLambda(Thread* self, ShadowFrame& shadow_frame, const Instruction* inst, uint16_t inst_data) argument [all...] |
/art/runtime/interpreter/mterp/arm/ |
H A D | alt_stub.S | 12 b MterpCheckBefore @ (self, shadow_frame) @ Tail call.
|
H A D | op_monitor_enter.S | 8 mov r1, rSELF @ r1<- self
|
H A D | op_new_instance.S | 9 bl MterpNewInstance @ (shadow_frame, self, inst_data)
|
H A D | op_return.S | 12 blne MterpSuspendCheck @ (self)
|
H A D | op_return_wide.S | 10 blne MterpSuspendCheck @ (self)
|
/art/runtime/interpreter/mterp/mips/ |
H A D | alt_stub.S | 13 jalr zero, a2 # Tail call to Mterp(self, shadow_frame)
|
H A D | op_return.S | 13 JAL(MterpSuspendCheck) # (self)
|
/art/runtime/interpreter/mterp/mips64/ |
H A D | alt_stub.S | 14 jalr zero, t9 # (self, shadow_frame) Note: tail call.
|
H A D | op_monitor_enter.S | 9 move a1, rSELF # a1 <- self
|
H A D | op_new_instance.S | 10 jal MterpNewInstance # (shadow_frame, self, inst_data)
|
H A D | op_return.S | 14 jal MterpSuspendCheck # (self)
|