/art/runtime/interpreter/mterp/x86_64/ |
H A D | op_instance_of.S | 15 call SYMBOL(MterpInstanceOf) # (index, &obj, method, self)
|
H A D | op_sget.S | 12 movq rSELF, OUT_ARG2 # self
|
/art/runtime/mirror/ |
H A D | stack_trace_element.cc | 42 StackTraceElement* StackTraceElement::Alloc(Thread* self, Handle<String> declaring_class, argument 46 down_cast<StackTraceElement*>(GetStackTraceElement()->AllocObject(self));
|
H A D | object_array-inl.h | 37 inline ObjectArray<T>* ObjectArray<T>::Alloc(Thread* self, Class* object_array_class, argument 39 Array* array = Array::Alloc<true>(self, object_array_class, length, 52 inline ObjectArray<T>* ObjectArray<T>::Alloc(Thread* self, Class* object_array_class, argument 54 return Alloc(self, object_array_class, length, 239 Thread* self = Thread::Current(); local 241 self->ThrowNewExceptionF("Ljava/lang/ArrayStoreException;", 252 inline ObjectArray<T>* ObjectArray<T>::CopyOf(Thread* self, int32_t new_length) { argument 255 StackHandleScope<1> hs(self); 260 ObjectArray<T>* new_array = Alloc(self, GetClass(), new_length, allocator_type);
|
H A D | object_array.h | 33 static ObjectArray<T>* Alloc(Thread* self, Class* object_array_class, int32_t length, 37 static ObjectArray<T>* Alloc(Thread* self, Class* object_array_class, int32_t length) 86 ObjectArray<T>* CopyOf(Thread* self, int32_t new_length)
|
/art/runtime/gc/space/ |
H A D | large_object_space.cc | 52 mirror::Object* Alloc(Thread* self, size_t num_bytes, size_t* bytes_allocated, 56 LargeObjectMapSpace::Alloc(self, num_bytes + kMemoryToolRedZoneBytes * 2, bytes_allocated, 74 bool IsZygoteLargeObject(Thread* self, mirror::Object* obj) const OVERRIDE { 75 return LargeObjectMapSpace::IsZygoteLargeObject(self, ObjectWithRedzone(obj)); 78 size_t Free(Thread* self, mirror::Object* obj) OVERRIDE { 81 return LargeObjectMapSpace::Free(self, object_with_rdz); 133 mirror::Object* LargeObjectMapSpace::Alloc(Thread* self, size_t num_bytes, argument 154 MutexLock mu(self, lock_); 176 bool LargeObjectMapSpace::IsZygoteLargeObject(Thread* self, mirror::Object* obj) const { argument 177 MutexLock mu(self, lock 183 SetAllLargeObjectsAsZygoteObjects(Thread* self) argument 190 Free(Thread* self, mirror::Object* ptr) argument 220 FreeList(Thread* self, size_t num_ptrs, mirror::Object** ptrs) argument 241 Thread* self = Thread::Current(); local 409 Free(Thread* self, mirror::Object* obj) argument 479 Alloc(Thread* self, size_t num_bytes, size_t* bytes_allocated, size_t* usable_size, size_t* bytes_tl_bulk_allocated) argument 569 SetAllLargeObjectsAsZygoteObjects(Thread* self) argument 584 Thread* self = context->self; local [all...] |
H A D | malloc_space.h | 56 virtual mirror::Object* AllocWithGrowth(Thread* self, size_t num_bytes, 60 virtual mirror::Object* Alloc(Thread* self, size_t num_bytes, size_t* bytes_allocated, 65 virtual size_t Free(Thread* self, mirror::Object* ptr) 67 virtual size_t FreeList(Thread* self, size_t num_ptrs, mirror::Object** ptrs)
|
H A D | dlmalloc_space.cc | 128 mirror::Object* DlMallocSpace::AllocWithGrowth(Thread* self, size_t num_bytes, argument 133 MutexLock mu(self, lock_); 138 result = AllocWithoutGrowthLocked(self, num_bytes, bytes_allocated, usable_size, 167 size_t DlMallocSpace::Free(Thread* self, mirror::Object* ptr) { argument 168 MutexLock mu(self, lock_); 181 size_t DlMallocSpace::FreeList(Thread* self, size_t num_ptrs, mirror::Object** ptrs) { argument 197 MutexLock mu(self, lock_); 218 MutexLock mu(self, lock_); 305 Thread* const self = Thread::Current(); local 310 Locks::mutator_lock_->AssertSharedHeld(self); [all...] |
H A D | region_space.cc | 335 bool RegionSpace::AllocNewTlab(Thread* self) { argument 336 MutexLock mu(self, region_lock_); 337 RevokeThreadLocalBuffersLocked(self); 351 r->thread_ = self; 352 self->SetTlab(r->Begin(), r->End()); 382 Thread* self = Thread::Current(); local 383 MutexLock mu(self, *Locks::runtime_shutdown_lock_); 384 MutexLock mu2(self, *Locks::thread_list_lock_); 400 Thread* self = Thread::Current(); local 401 MutexLock mu(self, *Lock [all...] |
/art/compiler/ |
H A D | common_compiler_test.cc | 113 Thread* self = Thread::Current(); local 114 StackHandleScope<1> hs(self); 116 mirror::Class* klass = class_linker_->FindClass(self, class_descriptor.c_str(), loader); 234 Thread* self = Thread::Current(); local 235 StackHandleScope<1> hs(self); 237 mirror::Class* klass = class_linker_->FindClass(self, class_descriptor.c_str(), loader); 258 Thread* self = Thread::Current(); local 259 mirror::Class* klass = class_linker_->FindClass(self, class_descriptor.c_str(), class_loader); 272 Thread* self = Thread::Current(); local 273 mirror::Class* klass = class_linker_->FindClass(self, class_descripto [all...] |
/art/runtime/ |
H A D | instrumentation_test.cc | 206 void DeoptimizeMethod(Thread* self, ArtMethod* method, bool enable_deoptimization) 210 ScopedThreadSuspension sts(self, kSuspended); 211 gc::ScopedGCCriticalSection gcs(self, 221 void UndeoptimizeMethod(Thread* self, ArtMethod* method, 226 ScopedThreadSuspension sts(self, kSuspended); 227 gc::ScopedGCCriticalSection gcs(self, 237 void DeoptimizeEverything(Thread* self, const char* key, bool enable_deoptimization) 241 ScopedThreadSuspension sts(self, kSuspended); 242 gc::ScopedGCCriticalSection gcs(self, 252 void UndeoptimizeEverything(Thread* self, cons [all...] |
H A D | monitor_android.cc | 52 void Monitor::LogContentionEvent(Thread* self, uint32_t wait_ms, uint32_t sample_percent, 60 int fd = open("/proc/self/cmdline", O_RDONLY); 71 // Emit self thread name string, <= 37 bytes. 73 self->GetThreadName(thread_name); 81 ArtMethod* m = self->GetCurrentMethod(&pc);
|
H A D | thread.cc | 415 Thread* self = reinterpret_cast<Thread*>(arg); local 418 LOG(ERROR) << "Thread attaching to non-existent runtime: " << *self; 422 // TODO: pass self to MutexLock - requires self to equal Thread::Current(), which is only true 423 // after self->Init(). 432 CHECK(self->Init(runtime->GetThreadList(), runtime->GetJavaVM(), self->tlsPtr_.tmp_jni_env)); 433 self->tlsPtr_.tmp_jni_env = nullptr; 437 ScopedObjectAccess soa(self); 438 self 586 Thread* self = static_cast<JNIEnvExt*>(env)->self; local 739 Thread* self; local 842 Thread* self = this; local 1004 ModifySuspendCount(Thread* self, int delta, AtomicInteger* suspend_barrier, bool for_debugger) argument 1052 PassActiveSuspendBarriers(Thread* self) argument 1219 Thread* self = Thread::Current(); local 1541 Thread* self = reinterpret_cast<Thread*>(arg); local 1672 MonitorExitVisitor(Thread* self) argument 1691 Thread* self = this; local 1944 Interrupt(Thread* self) argument 1954 Thread* self = Thread::Current(); local 1959 NotifyLocked(Thread* self) argument 2017 BuildInternalStackTraceVisitor(Thread* self, Thread* thread, int skip_depth) argument [all...] |
H A D | handle_scope-inl.h | 30 inline StackHandleScope<kNumReferences>::StackHandleScope(Thread* self, mirror::Object* fill_value) argument 31 : HandleScope(self->GetTopHandleScope(), kNumReferences), self_(self), pos_(0) { 32 DCHECK_EQ(self, Thread::Current());
|
H A D | signal_catcher.h | 50 int WaitForSignal(Thread* self, SignalSet& signals) REQUIRES(!lock_);
|
/art/compiler/driver/ |
H A D | compiler_driver_test.cc | 191 Thread* self = Thread::Current(); local 194 ScopedObjectAccess soa(self); 208 ScopedObjectAccess soa(self); 209 StackHandleScope<1> hs(self); 211 reinterpret_cast<mirror::ClassLoader*>(self->DecodeJObject(class_loader)))); 212 mirror::Class* klass = class_linker->FindClass(self, "LStaticLeafMethods;", h_loader); 265 Thread* self = Thread::Current(); local 266 ScopedObjectAccess soa(self); 267 StackHandleScope<1> hs(self); 269 reinterpret_cast<mirror::ClassLoader*>(self 295 Thread* self = Thread::Current(); local [all...] |
/art/runtime/interpreter/ |
H A D | interpreter_switch_impl.cc | 30 DCHECK(self->IsExceptionPending()); \ 31 self->AllowThreadSuspension(); \ 32 uint32_t found_dex_pc = FindNextInstructionFollowingException(self, shadow_frame, \ 37 DoMonitorCheckOnExit<do_assignability_check>(self, &shadow_frame); \ 59 if (!DoMonitorCheckOnExit<do_assignability_check>(self, &shadow_frame)) { \ 67 instrumentation->DexPcMovedEvent(self, shadow_frame.GetThisObject(code_item->ins_size_), \ 75 instrumentation->Branch(self, method, dex_pc, offset); \ 78 if (jit::Jit::MaybeDoOnStackReplacement(self, method, dex_pc, offset, &result)) { \ 90 jit->AddSamples(self, method, 1, /*with_backedges*/ true); \ 100 JValue ExecuteSwitchImpl(Thread* self, cons argument [all...] |
/art/runtime/gc/ |
H A D | heap.h | 192 mirror::Object* AllocObject(Thread* self, 200 self, klass, num_bytes, GetCurrentAllocator(), pre_fence_visitor); 204 mirror::Object* AllocNonMovableObject(Thread* self, 212 self, klass, num_bytes, GetCurrentNonMovingAllocator(), pre_fence_visitor); 216 ALWAYS_INLINE mirror::Object* AllocObjectWithAllocator(Thread* self, 294 void IncrementDisableMovingGC(Thread* self) REQUIRES(!*gc_complete_lock_); 295 void DecrementDisableMovingGC(Thread* self) REQUIRES(!*gc_complete_lock_); 298 void IncrementDisableThreadFlip(Thread* self) REQUIRES(!*thread_flip_lock_); 299 void DecrementDisableThreadFlip(Thread* self) REQUIRES(!*thread_flip_lock_); 300 void ThreadFlipBegin(Thread* self) REQUIRE [all...] |
/art/compiler/jit/ |
H A D | jit_compiler.cc | 59 void* handle, ArtMethod* method, Thread* self, bool osr) 63 return jit_compiler->CompileMethod(self, method, osr); 202 bool JitCompiler::CompileMethod(Thread* self, ArtMethod* method, bool osr) { argument 205 StackHandleScope<2> hs(self); 206 self->AssertNoPendingException(); 211 if (!runtime->GetClassLinker()->EnsureInitialized(self, h_class, true, true)) { 221 success = compiler_driver_->GetCompiler()->JitCompile(self, code_cache, method, osr);
|
H A D | jit_compiler.h | 39 bool CompileMethod(Thread* self, ArtMethod* method, bool osr)
|
/art/runtime/jit/ |
H A D | jit.h | 51 bool CompileMethod(ArtMethod* method, Thread* self, bool osr) 100 void WaitForCompilationToFinish(Thread* self); 106 void AddSamples(Thread* self, ArtMethod* method, uint16_t samples, bool with_backedges) 116 void NotifyInterpreterToCompiledCodeTransition(Thread* self, ArtMethod* caller) 118 AddSamples(self, caller, invoke_transition_weight_, false); 121 void NotifyCompiledCodeToInterpreterTransition(Thread* self, ArtMethod* callee) 123 AddSamples(self, callee, invoke_transition_weight_, false);
|
/art/runtime/interpreter/mterp/out/ |
H A D | mterp_x86.S | 344 * 0 Thread* self 793 call SYMBOL(MterpConstString) # (index, tgt_reg, shadow_frame, self) 812 call SYMBOL(MterpConstString) # (index, tgt_reg, shadow_frame, self) 831 call SYMBOL(MterpConstClass) # (index, tgt_reg, shadow_frame, self) 850 call SYMBOL(artLockObjectFromCode) # (object, self) 873 call SYMBOL(artUnlockObjectFromCode) # (object, self) 896 call SYMBOL(MterpCheckCast) # (index, &obj, method, self) 924 call SYMBOL(MterpInstanceOf) # (index, &obj, method, self) 2114 movl %ecx, OUT_ARG3(%esp) # self 2147 movl %ecx, OUT_ARG3(%esp) # self [all...] |
/art/compiler/dex/ |
H A D | verification_results.cc | 40 Thread* self = Thread::Current(); local 42 WriterMutexLock mu(self, verified_methods_lock_);
|
/art/compiler/utils/ |
H A D | dedupe_set-inl.h | 69 const StoreKey* Add(Thread* self, size_t hash, const InKey& in_key) REQUIRES(!lock_) { 70 MutexLock lock(self, lock_); 82 void UpdateStats(Thread* self, Stats* global_stats) REQUIRES(!lock_) { 87 MutexLock lock(self, lock_); 188 Thread* self, const InKey& key) { 200 return shards_[shard_bin]->Add(self, shard_hash, key); 236 Thread* self) const { 239 shards_[shard]->UpdateStats(self, &stats); 187 Add( Thread* self, const InKey& key) argument
|
/art/runtime/entrypoints/quick/ |
H A D | callee_save_frame.h | 39 explicit ScopedQuickEntrypointChecks(Thread *self, 42 SHARED_REQUIRES(Locks::mutator_lock_) : self_(self), exit_check_(exit_check) {
|