/art/runtime/gc/ |
H A D | scoped_gc_critical_section.cc | 30 : self_(self) { 35 self_->EndAssertNoThreadSuspension(old_cause_); 36 Runtime::Current()->GetHeap()->FinishGC(self_, collector::kGcTypeNone);
|
H A D | scoped_gc_critical_section.h | 40 Thread* const self_; member in class:art::gc::ScopedGCCriticalSection
|
H A D | heap.cc | 2379 forward_address = to_space_->Alloc(self_, alloc_size, &bytes_allocated, nullptr, &dummy);
|
/art/runtime/ |
H A D | object_lock.cc | 25 ObjectLock<T>::ObjectLock(Thread* self, Handle<T> object) : self_(self), obj_(object) { 27 obj_->MonitorEnter(self_); 32 obj_->MonitorExit(self_); 37 Monitor::Wait(self_, obj_.Get(), 0, 0, false, kWaiting); 42 obj_->Notify(self_); 47 obj_->NotifyAll(self_); 51 ObjectTryLock<T>::ObjectTryLock(Thread* self, Handle<T> object) : self_(self), obj_(object) { 53 acquired_ = obj_->MonitorTryEnter(self_) != nullptr; 59 obj_->MonitorExit(self_);
|
H A D | scoped_thread_state_change.h | 38 : self_(self), thread_state_(new_thread_state), expected_has_no_thread_(false) { 39 if (UNLIKELY(self_ == nullptr)) { 43 CHECK(runtime == nullptr || !runtime->IsStarted() || runtime->IsShuttingDown(self_)); 51 self_->TransitionFromSuspendedToRunnable(); 53 self_->TransitionFromRunnableToSuspended(new_thread_state); 56 self_->SetState(new_thread_state); 63 if (UNLIKELY(self_ == nullptr)) { 72 self_->TransitionFromSuspendedToRunnable(); 74 self_->TransitionFromRunnableToSuspended(old_thread_state_); 77 self_ 93 Thread* const self_; member in class:art::ScopedThreadStateChange 199 Thread* const self_; member in class:art::ScopedObjectAccessAlreadyRunnable 298 Thread* const self_; member in class:art::ScopedThreadSuspension [all...] |
H A D | object_lock.h | 42 Thread* const self_; member in class:art::ObjectLock 60 Thread* const self_; member in class:art::ObjectTryLock
|
H A D | handle_scope-inl.h | 31 : HandleScope(self->GetTopHandleScope(), kNumReferences), self_(self), pos_(0) { 42 self_->PushHandleScope(this); 47 HandleScope* top_handle_scope = self_->PopHandleScope(); 50 Locks::mutator_lock_->AssertSharedHeld(self_);
|
H A D | quick_exception_handler.cc | 41 : self_(self), 145 self_->DumpStack(LOG(INFO) << "Delivering exception: " << PrettyTypeOf(exception) 148 StackHandleScope<1> hs(self_); 152 CatchBlockStackVisitor visitor(self_, context_, &exception_ref, this); 167 DCHECK(!self_->IsExceptionPending()); 170 self_->SetException(exception_ref.Get()); 218 self_->DumpStack(LOG(INFO) << "Setting catch phis: "); 488 self_->DumpStack(LOG(INFO) << "Deoptimizing: "); 491 DeoptimizeStackVisitor visitor(self_, context_, this, false); 495 self_ [all...] |
H A D | handle_scope.h | 162 return self_; 177 Thread* const self_; variable 199 self_(self), 214 new StackHandleScope<kNumReferencesPerScope>(self_); 225 Thread* const self_; member in class:art::StackHandleScopeCollection
|
H A D | quick_exception_handler.h | 112 Thread* const self_; member in class:art::QuickExceptionHandler
|
H A D | thread.h | 1537 : self_(self), old_cause_(self->StartAssertNoThreadSuspension(cause)) { 1540 self_->EndAssertNoThreadSuspension(old_cause_); 1543 return self_; 1547 Thread* const self_; member in class:art::ScopedAssertNoThreadSuspension 1554 : self_(self), type_(type) { 1555 self_->PushStackedShadowFrame(sf, type); 1558 self_->PopStackedShadowFrame(type_); 1562 Thread* const self_; member in class:art::ScopedStackedShadowFramePusher 1571 explicit ScopedDebugDisallowReadBarriers(Thread* self) : self_(self) { 1572 self_ 1579 Thread* const self_; member in class:art::ScopedDebugDisallowReadBarriers [all...] |
H A D | thread.cc | 1672 explicit MonitorExitVisitor(Thread* self) : self_(self) { } 1677 if (self_->HoldsLock(entered_monitor)) { 1682 entered_monitor->MonitorExit(self_); 1687 Thread* const self_; member in class:art::MonitorExitVisitor 2019 self_(self), 2030 StackHandleScope<1> hs(self_); 2039 self_->StartAssertNoThreadSuspension("Building internal stack trace"); 2040 self_->AssertPendingOOMException(); 2043 mirror::PointerArray* methods_and_pcs = class_linker->AllocPointerArray(self_, depth * 2); 2045 self_ 2095 Thread* const self_; member in class:art::BuildInternalStackTraceVisitor [all...] |
H A D | monitor.cc | 1362 MonitorDeflateVisitor() : self_(Thread::Current()), deflate_count_(0) {} 1366 if (Monitor::Deflate(self_, object)) { 1375 Thread* const self_; member in class:art::MonitorDeflateVisitor 1381 Locks::mutator_lock_->AssertExclusiveHeld(visitor.self_);
|
/art/runtime/entrypoints/quick/ |
H A D | callee_save_frame.h | 42 SHARED_REQUIRES(Locks::mutator_lock_) : self_(self), exit_check_(exit_check) { 49 : self_(kIsDebugBuild ? Thread::Current() : nullptr), exit_check_(kIsDebugBuild) { 63 Locks::mutator_lock_->AssertSharedHeld(self_); 64 self_->VerifyStack(); 68 Locks::mutator_lock_->AssertSharedHeld(self_); 69 self_->VerifyStack(); 72 Thread* const self_; member in class:art::ScopedQuickEntrypointChecks
|
/art/runtime/base/ |
H A D | mutex.h | 484 MutexLock(Thread* self, Mutex& mu) ACQUIRE(mu) : self_(self), mu_(mu) { 485 mu_.ExclusiveLock(self_); 489 mu_.ExclusiveUnlock(self_); 493 Thread* const self_; member in class:art::MutexLock 505 self_(self), mu_(mu) { 506 mu_.SharedLock(self_); 510 mu_.SharedUnlock(self_); 514 Thread* const self_; member in class:art::ReaderMutexLock 527 self_(self), mu_(mu) { 528 mu_.ExclusiveLock(self_); 536 Thread* const self_; member in class:art::WriterMutexLock [all...] |
/art/runtime/gc/collector/ |
H A D | semi_space.cc | 64 WriterMutexLock mu(self_, *Locks::heap_bitmap_lock_); 99 self_(nullptr), 153 self_ = Thread::Current(); 176 CHECK(Locks::mutator_lock_->IsExclusiveHeld(self_)); 178 Locks::mutator_lock_->AssertExclusiveHeld(self_); 181 ThreadState old_state = self_->SetStateUnsafe(kRunnable); 187 CHECK_EQ(self_->SetStateUnsafe(old_state), kRunnable); 214 Locks::mutator_lock_->AssertExclusiveHeld(self_); 238 heap_->RevokeAllThreadLocalAllocationStacks(self_); 242 WriterMutexLock mu(self_, *Lock [all...] |
H A D | semi_space.h | 221 Thread* self_; member in class:art::gc::collector::SemiSpace
|
/art/runtime/mirror/ |
H A D | object.cc | 111 : self_(self), orig_(orig), num_bytes_(num_bytes) { 116 Object::CopyObject(self_, obj, orig_->Get(), num_bytes_); 120 Thread* const self_; member in class:art::mirror::CopyObjectVisitor
|
H A D | class.cc | 966 : self_(self), orig_(orig), new_length_(new_length), 972 StackHandleScope<1> hs(self_); 974 mirror::Object::CopyObject(self_, h_new_class_obj.Get(), orig_->Get(), copy_bytes_); 975 mirror::Class::SetStatus(h_new_class_obj, Class::kStatusResolving, self_); 986 Thread* const self_; member in class:art::mirror::CopyClassVisitor
|
/art/runtime/verifier/ |
H A D | method_verifier.cc | 511 : self_(self), 1053 DCHECK(self_->IsExceptionPending()); 1054 self_->ClearException(); 1814 self_->AllowThreadSuspension(); 2850 DCHECK(!can_load_classes_ || self_->IsExceptionPending()); 2851 self_->ClearException(); 2893 DCHECK(!can_load_classes_ || self_->IsExceptionPending()); 2894 self_->ClearException(); 3536 DCHECK(self_->IsExceptionPending()); 3537 self_ [all...] |
H A D | method_verifier.h | 735 Thread* const self_; member in class:art::verifier::MethodVerifier
|
/art/compiler/optimizing/ |
H A D | inliner.cc | 226 self_(self), 237 Runtime::Current()->GetJit()->GetCodeCache()->DoneCompilerUse(method_, self_); 245 Thread* const self_; member in class:art::ScopedProfilingInfoInlineUse
|
/art/compiler/driver/ |
H A D | compiler_driver.cc | 1195 self_->EndAssertNoThreadSuspension(old_cause_); 1250 image_class_descriptors_(image_class_descriptors), self_(self) { 1280 StackHandleScope<1> hs(self_); 1298 Thread* const self_; member in class:art::ClinitImageUpdate
|