/art/runtime/ |
H A D | thread-inl.h | 39 return full_env->self; 261 DCHECK(Thread::Current() == this) << "Should be called by self"; 273 // Note: self is not necessarily equal to this thread since thread may be suspended. 274 Thread* self = Thread::Current(); local 275 DCHECK(this == self || IsSuspended() || GetState() == kWaitingPerformingGc) 276 << GetState() << " thread " << this << " self " << self;
|
H A D | stack.cc | 953 void LockCountData::AddMonitor(Thread* self, mirror::Object* obj) { argument 960 if (self->IsExceptionPending()) { 970 void LockCountData::RemoveMonitorOrThrow(Thread* self, const mirror::Object* obj) { argument 988 self->ClearException(); 989 self->ThrowNewExceptionF("Ljava/lang/IllegalMonitorStateException;", 997 void MonitorExitHelper(Thread* self, mirror::Object* obj) NO_THREAD_SAFETY_ANALYSIS { 998 DCHECK(self != nullptr); 1000 obj->MonitorExit(self); 1003 bool LockCountData::CheckAllMonitorsReleasedOrThrow(Thread* self) { argument 1004 DCHECK(self ! [all...] |
H A D | oat_file_manager.cc | 554 Thread* const self = Thread::Current(); local 555 Locks::mutator_lock_->AssertNotHeld(self); 638 ScopedObjectAccess soa(self); 639 StackHandleScope<1> hs(self); 648 ScopedThreadSuspension sts(self, kSuspended); 649 gc::ScopedGCCriticalSection gcs(self, 672 ScopedThreadSuspension sts(self, kSuspended); 673 gc::ScopedGCCriticalSection gcs(self,
|
H A D | intern_table.h | 88 mirror::String* LookupStrong(Thread* self, mirror::String* s) 91 mirror::String* LookupStrong(Thread* self, uint32_t utf16_length, const char* utf8_data) 96 mirror::String* LookupWeak(Thread* self, mirror::String* s) 260 void WaitUntilAccessible(Thread* self)
|
H A D | handle_scope.h | 148 explicit ALWAYS_INLINE StackHandleScope(Thread* self, mirror::Object* fill_value = nullptr); 198 explicit StackHandleScopeCollection(Thread* const self) : argument 199 self_(self),
|
H A D | transaction.cc | 74 void Transaction::ThrowAbortError(Thread* self, const std::string* abort_message) { argument 82 self->ThrowNewWrappedException(Transaction::kAbortExceptionSignature, 86 self->ThrowNewWrappedException(Transaction::kAbortExceptionSignature, 196 Thread* self = Thread::Current(); local 197 self->AssertNoPendingException(); 198 MutexLock mu1(self, *Locks::intern_table_lock_); 199 MutexLock mu2(self, log_lock_);
|
H A D | dex_file.cc | 1596 Thread* self = Thread::Current(); local 1597 ScopedObjectAccessUnchecked soa(self); 1598 StackHandleScope<5> hs(self); 1602 hs.NewHandle(mirror::String::AllocFromModifiedUtf8(self, name))); 1621 Handle<mirror::Object> new_member(hs.NewHandle(annotation_member_class->AllocObject(self))); 1623 hs.NewHandle(mirror::Method::CreateFromArtMethod(self, annotation_method))); 1641 annotation_member_init->Invoke(self, args, sizeof(args), &result, "VLLLL"); 1642 if (self->IsExceptionPending()) { 1754 Thread* self = Thread::Current(); local 1755 ScopedObjectAccessUnchecked soa(self); 1808 Thread* self = Thread::Current(); local 1839 Thread* self = Thread::Current(); local 2104 Thread* self = Thread::Current(); local [all...] |
H A D | thread.h | 163 // Get the currently executing thread, frequently referred to as 'self'. This call has reasonably 164 // high cost and so we favor passing self around when possible. 229 bool ModifySuspendCount(Thread* self, int delta, AtomicInteger* suspend_barrier, bool for_debugger) 465 void Interrupt(Thread* self) REQUIRES(!*wait_mutex_); 472 void NotifyLocked(Thread* self) REQUIRES(wait_mutex_); 602 return ThreadOffsetFromTlsPtr<pointer_size>(OFFSETOF_MEMBER(tls_ptr_sized_values, self)); 958 bool PassActiveSuspendBarriers(Thread* self) 1350 self(nullptr), opeer(nullptr), jpeer(nullptr), stack_begin(nullptr), stack_size(0), 1392 Thread* self; variable 1536 ScopedAssertNoThreadSuspension(Thread* self, cons 1553 ScopedStackedShadowFramePusher(Thread* self, ShadowFrame* sf, StackedShadowFrameType type) argument 1571 ScopedDebugDisallowReadBarriers(Thread* self) argument [all...] |
H A D | reflection_test.cc | 91 Thread* self = Thread::Current(); local 92 StackHandleScope<2> hs(self); 95 ScopedObjectAccessUnchecked(self).Decode<mirror::ClassLoader*>(jclass_loader))); 97 MakeExecutable(ScopedObjectAccessUnchecked(self).Decode<mirror::ClassLoader*>(jclass_loader), 102 MakeExecutable(ScopedObjectAccessUnchecked(self).Decode<mirror::ClassLoader*>(jclass_loader), 106 mirror::Class* c = class_linker_->FindClass(self, DotToDescriptor(class_name).c_str(), 118 StackHandleScope<1> hs2(self); 120 bool initialized = class_linker_->EnsureInitialized(self, h_class, true, true); 122 *receiver = c->AllocObject(self); 128 self [all...] |
/art/tools/checker/match/ |
H A D | file.py | 26 def __init__(self, assertion, lineNo, variables): 27 self.assertion = assertion 28 self.lineNo = lineNo 29 self.variables = variables
|
/art/runtime/gc/collector/ |
H A D | mark_compact.cc | 59 Thread* self = Thread::Current(); local 61 CHECK(!Locks::mutator_lock_->IsExclusiveHeld(self)); 126 void MarkCompact::ProcessReferences(Thread* self) { argument 127 WriterMutexLock mu(self, *Locks::heap_bitmap_lock_); 170 Thread* self = Thread::Current(); local 177 CHECK(Locks::mutator_lock_->IsExclusiveHeld(self)); 191 heap_->RevokeAllThreadLocalAllocationStacks(self); 196 WriterMutexLock mu(self, *Locks::heap_bitmap_lock_); 203 ProcessReferences(self); 205 ReaderMutexLock mu(self, *Lock [all...] |
H A D | mark_sweep.h | 91 void MarkRoots(Thread* self) 106 void MarkRootsCheckpoint(Thread* self, bool revoke_ros_alloc_thread_local_buffers_at_checkpoint) 133 void ProcessReferences(Thread* self) 177 void SweepSystemWeaks(Thread* self) 301 void RevokeAllThreadLocalAllocationStacks(Thread* self) NO_THREAD_SAFETY_ANALYSIS;
|
H A D | concurrent_copying.h | 148 void ProcessReferences(Thread* self) SHARED_REQUIRES(Locks::mutator_lock_); 157 void SweepSystemWeaks(Thread* self) 176 void RecordLiveStackFreezeSize(Thread* self); 182 void ReenableWeakRefAccess(Thread* self) SHARED_REQUIRES(Locks::mutator_lock_);
|
H A D | garbage_collector.cc | 83 Thread* self = Thread::Current(); local 97 if (Locks::mutator_lock_->IsExclusiveHeld(self)) { 105 MutexLock mu(self, pause_histogram_lock_);
|
/art/runtime/interpreter/mterp/arm/ |
H A D | footer.S | 97 bl MterpHandleException @ (self, shadow_frame) 167 bl MterpSuspendCheck @ (self) 182 bl MterpMaybeDoOnStackReplacement @ (self, shadow_frame, offset) 201 bl MterpMaybeDoOnStackReplacement @ (self, shadow_frame, offset) 211 bl MterpAddHotnessBatch @ (method, shadow_frame, self) 224 bl MterpMaybeDoOnStackReplacement @ (self, shadow_frame, offset) 293 bl MterpAddHotnessBatch @ (method, shadow_frame, self)
|
/art/runtime/interpreter/mterp/mips/ |
H A D | footer.S | 96 JAL(MterpHandleException) # (self, shadow_frame) 127 JAL(MterpSuspendCheck) # (self)
|
H A D | op_goto.S | 15 JAL(MterpProfileBranch) # (self, shadow_frame, offset)
|
H A D | op_goto_32.S | 20 JAL(MterpProfileBranch) # (self, shadow_frame, offset)
|
/art/compiler/ |
H A D | compiler.h | 62 virtual bool JitCompile(Thread* self ATTRIBUTE_UNUSED,
|
H A D | image_writer.cc | 431 Thread* const self = Thread::Current(); local 432 ReaderMutexLock mu(self, *class_linker->DexLock()); 435 down_cast<mirror::DexCache*>(self->DecodeJObject(data.weak_root)); 843 Thread* self = Thread::Current(); local 869 ScopedAssertNoThreadSuspension sa(self, __FUNCTION__); 870 ReaderMutexLock mu(self, *Locks::classlinker_classes_lock_); // For ClassInClassTable 871 ReaderMutexLock mu2(self, *class_linker->DexLock()); 873 if (self->IsJWeakCleared(data.weak_root)) { 876 mirror::DexCache* dex_cache = self->DecodeJObject(data.weak_root)->AsDexCache(); 951 Thread* const self local 1012 Thread* self = Thread::Current(); local 1333 Thread* const self = Thread::Current(); local [all...] |
/art/runtime/base/ |
H A D | arena_allocator.cc | 232 Thread* self = Thread::Current(); local 235 MutexLock lock(self, lock_); 280 Thread* self = Thread::Current(); local 281 MutexLock lock(self, lock_);
|
/art/runtime/gc/space/ |
H A D | malloc_space.cc | 234 Thread* self = context->self; local 235 Locks::heap_bitmap_lock_->AssertExclusiveHeld(self); 248 context->freed.bytes += space->FreeList(self, num_ptrs, ptrs);
|
H A D | bump_pointer_space.h | 49 mirror::Object* Alloc(Thread* self, size_t num_bytes, size_t* bytes_allocated, 52 mirror::Object* AllocThreadUnsafe(Thread* self, size_t num_bytes, size_t* bytes_allocated, 138 bool AllocNewTlab(Thread* self, size_t bytes) REQUIRES(!block_lock_);
|
/art/runtime/mirror/ |
H A D | stack_trace_element.h | 50 static StackTraceElement* Alloc(Thread* self, Handle<String> declaring_class,
|
/art/runtime/arch/arm/ |
H A D | fault_handler_arm.cc | 58 Thread* self = Thread::Current(); local 59 CHECK(self != nullptr); // This will cause a SIGABRT if self is null. 61 sc->arm_r0 = reinterpret_cast<uintptr_t>(*self->GetNestedSignalState());
|