/art/runtime/ |
H A D | closure.h | 27 virtual void Run(Thread* self) = 0;
|
H A D | monitor_pool.h | 45 static Monitor* CreateMonitor(Thread* self, Thread* owner, mirror::Object* obj, int32_t hash_code) 48 return new Monitor(self, owner, obj, hash_code); 50 return GetMonitorPool()->CreateMonitorInPool(self, owner, obj, hash_code); 54 static void ReleaseMonitor(Thread* self, Monitor* monitor) { argument 58 GetMonitorPool()->ReleaseMonitorToPool(self, monitor); 62 static void ReleaseMonitors(Thread* self, MonitorList::Monitors* monitors) { argument 66 GetMonitorPool()->ReleaseMonitorsToPool(self, monitors); 86 static MonitorId ComputeMonitorId(Monitor* mon, Thread* self) { argument 90 return GetMonitorPool()->ComputeMonitorIdInPool(mon, self); 110 Monitor* CreateMonitorInPool(Thread* self, Threa 131 ComputeMonitorIdInPool(Monitor* mon, Thread* self) argument [all...] |
H A D | signal_catcher.cc | 45 // On Android, /proc/self/cmdline will have been rewritten to something like "system_server". 48 if (ReadFileToString("/proc/self/cmdline", ¤t_cmd_line)) { 74 Thread* self = Thread::Current(); local 75 MutexLock mu(self, lock_); 77 cond_.Wait(self); 127 Thread* self = Thread::Current(); local 128 Locks::mutator_lock_->AssertExclusiveHeld(self); 129 const char* old_cause = self->StartAssertNoThreadSuspension("Handling SIGQUIT"); 130 ThreadState old_state = self->SetStateUnsafe(kRunnable); 147 if (ReadFileToString("/proc/self/map 168 WaitForSignal(Thread* self, SignalSet& signals) argument 196 Thread* self = Thread::Current(); local [all...] |
H A D | handle_scope-inl.h | 28 inline StackHandleScope<kNumReferences>::StackHandleScope(Thread* self) argument 29 : HandleScope(kNumReferences), self_(self), pos_(0) {
|
H A D | monitor.h | 70 static void Notify(Thread* self, mirror::Object* obj) 72 DoNotify(self, obj, false); 74 static void NotifyAll(Thread* self, mirror::Object* obj) 76 DoNotify(self, obj, true); 78 static void Wait(Thread* self, mirror::Object* obj, int64_t ms, int32_t ns, 124 static void InflateThinLocked(Thread* self, Handle<mirror::Object> obj, LockWord lock_word, 127 static bool Deflate(Thread* self, mirror::Object* obj) 131 explicit Monitor(Thread* self, Thread* owner, mirror::Object* obj, int32_t hash_code) 133 explicit Monitor(Thread* self, Thread* owner, mirror::Object* obj, int32_t hash_code, 138 bool Install(Thread* self) [all...] |
H A D | monitor_pool.cc | 91 Monitor* MonitorPool::CreateMonitorInPool(Thread* self, Thread* owner, mirror::Object* obj, 95 MutexLock mu(self, *Locks::allocated_monitor_ids_lock_); 110 Monitor* monitor = new(mon_uninitialized) Monitor(self, owner, obj, hash_code, id); 115 void MonitorPool::ReleaseMonitorToPool(Thread* self, Monitor* monitor) { argument 117 MutexLock mu(self, *Locks::allocated_monitor_ids_lock_); 134 void MonitorPool::ReleaseMonitorsToPool(Thread* self, MonitorList::Monitors* monitors) { argument 136 ReleaseMonitorToPool(self, mon);
|
/art/runtime/entrypoints/quick/ |
H A D | quick_fillarray_entrypoints.cc | 42 Thread* self, StackReference<mirror::ArtMethod>* sp) 44 FinishCalleeSaveFrameSetup(self, sp, Runtime::kRefsOnly); 52 ThrowLocation throw_location = self->GetCurrentLocationForThrow(); 53 self->ThrowNewExceptionF(throw_location, "Ljava/lang/ArrayIndexOutOfBoundsException;",
|
H A D | quick_field_entrypoints.cc | 30 Thread* self, StackReference<mirror::ArtMethod>* sp) 37 FinishCalleeSaveFrameSetup(self, sp, Runtime::kRefsOnly); 38 field = FindFieldFromCode<StaticPrimitiveRead, true>(field_idx, referrer, self, sizeof(int32_t)); 47 Thread* self, StackReference<mirror::ArtMethod>* sp) 54 FinishCalleeSaveFrameSetup(self, sp, Runtime::kRefsOnly); 55 field = FindFieldFromCode<StaticPrimitiveRead, true>(field_idx, referrer, self, sizeof(int64_t)); 64 Thread* self, 72 FinishCalleeSaveFrameSetup(self, sp, Runtime::kRefsOnly); 73 field = FindFieldFromCode<StaticObjectRead, true>(field_idx, referrer, self, 82 mirror::ArtMethod* referrer, Thread* self, [all...] |
H A D | quick_entrypoints.h | 50 extern uint32_t JniMethodStart(Thread* self) NO_THREAD_SAFETY_ANALYSIS HOT_ATTR; 51 extern uint32_t JniMethodStartSynchronized(jobject to_lock, Thread* self) 53 extern void JniMethodEnd(uint32_t saved_local_ref_cookie, Thread* self) 56 Thread* self) 59 Thread* self) 64 jobject locked, Thread* self)
|
H A D | quick_alloc_entrypoints.cc | 30 uint32_t type_idx, mirror::ArtMethod* method, Thread* self, \ 33 FinishCalleeSaveFrameSetup(self, sp, Runtime::kRefsOnly); \ 34 return AllocObjectFromCode<false, instrumented_bool>(type_idx, method, self, allocator_type); \ 37 mirror::Class* klass, mirror::ArtMethod* method, Thread* self, \ 40 FinishCalleeSaveFrameSetup(self, sp, Runtime::kRefsOnly); \ 41 return AllocObjectFromCodeResolved<instrumented_bool>(klass, method, self, allocator_type); \ 44 mirror::Class* klass, mirror::ArtMethod* method, Thread* self, \ 47 FinishCalleeSaveFrameSetup(self, sp, Runtime::kRefsOnly); \ 48 return AllocObjectFromCodeInitialized<instrumented_bool>(klass, method, self, allocator_type); \ 51 uint32_t type_idx, mirror::ArtMethod* method, Thread* self, \ [all...] |
H A D | callee_save_frame.h | 38 static inline void FinishCalleeSaveFrameSetup(Thread* self, StackReference<mirror::ArtMethod>* sp, 42 Locks::mutator_lock_->AssertSharedHeld(self); 44 self->SetTopOfStack(sp, 0); 45 self->VerifyStack();
|
/art/runtime/entrypoints/portable/ |
H A D | portable_thread_entrypoints.cc | 27 explicit ShadowFrameCopyVisitor(Thread* self) : StackVisitor(self, NULL), prev_frame_(NULL), argument 72 extern "C" void art_portable_test_suspend_from_code(Thread* self) 74 CheckSuspend(self); 77 ShadowFrameCopyVisitor visitor(self); 79 self->SetDeoptimizationShadowFrame(visitor.GetShadowFrameCopy()); 80 self->SetDeoptimizationReturnValue(JValue()); 81 self->SetException(ThrowLocation(), Thread::GetDeoptimizationException());
|
H A D | portable_fillarray_entrypoints.cc | 39 Thread* self = Thread::Current(); local 40 ThrowLocation throw_location = self->GetCurrentLocationForThrow(); 41 self->ThrowNewExceptionF(throw_location, "Ljava/lang/ArrayIndexOutOfBoundsException;",
|
/art/runtime/gc/ |
H A D | heap-inl.h | 39 inline mirror::Object* Heap::AllocObjectWithAllocator(Thread* self, mirror::Class* klass, argument 46 CHECK_EQ(self->GetState(), kRunnable); 47 self->AssertThreadSuspensionIsAllowable(); 52 return AllocLargeObject<kInstrumented, PreFenceVisitor>(self, klass, byte_count, 64 if (allocator == kAllocatorTypeTLAB && byte_count <= self->TlabSize()) { 65 obj = self->AllocTlab(byte_count); 79 obj = TryToAllocate<kInstrumented, false>(self, allocator, byte_count, &bytes_allocated, 83 obj = AllocateInternalWithGc(self, allocator, byte_count, &bytes_allocated, &usable_size, 89 if (!self->IsExceptionPending() && is_current_allocator && !after_is_current_allocator) { 91 return AllocObject<kInstrumented>(self, klas 163 PushOnAllocationStack(Thread* self, mirror::Object** obj) argument 174 AllocLargeObject(Thread* self, mirror::Class* klass, size_t byte_count, const PreFenceVisitor& pre_fence_visitor) argument 183 TryToAllocate(Thread* self, AllocatorType allocator_type, size_t alloc_size, size_t* bytes_allocated, size_t* usable_size) argument 310 CheckConcurrentGC(Thread* self, size_t new_num_bytes_allocated, mirror::Object** obj) argument [all...] |
/art/runtime/mirror/ |
H A D | array.cc | 43 static Array* RecursiveCreateMultiArray(Thread* self, 48 StackHandleScope<1> hs(self); 51 Array::Alloc<true>(self, array_class.Get(), array_length, array_class->GetComponentSize(), 54 CHECK(self->IsExceptionPending()); 60 StackHandleScope<1> hs(self); 62 Array* sub_array = RecursiveCreateMultiArray(self, h_component_type, 65 CHECK(self->IsExceptionPending()); 75 Array* Array::CreateMultiArray(Thread* self, Handle<Class> element_class, argument 96 StackHandleScope<1> hs(self); 98 hs.NewHandle(class_linker->FindArrayClass(self, [all...] |
H A D | art_method.cc | 238 Thread* self = Thread::Current(); local 240 StackHandleScope<1> hs(self); 241 Handle<mirror::Throwable> exception(hs.NewHandle(self->GetException(&throw_location))); 242 bool is_exception_reported = self->IsExceptionReportedToInstrumentation(); 243 self->ClearException(); 260 self->ClearException(); 263 delete self->GetLongJumpContext(); 278 self->SetException(throw_location, exception.Get()); 279 self->SetExceptionReportedToInstrumentation(is_exception_reported); 284 void ArtMethod::Invoke(Thread* self, uint32_ argument 374 RegisterNative(Thread* self, const void* native_method, bool is_fast) argument 385 UnregisterNative(Thread* self) argument [all...] |
/art/runtime/interpreter/ |
H A D | interpreter.cc | 27 static void UnstartedRuntimeJni(Thread* self, ArtMethod* method, 36 mirror::Class* array_class = runtime->GetClassLinker()->FindArrayClass(self, &element_class); 39 result->SetL(mirror::Array::Alloc<true>(self, array_class, length, 44 NthCallerVisitor visitor(self, 3); 52 StackHandleScope<1> hs(self); 63 result->SetL(receiver->Clone(self)); 65 receiver->NotifyAll(self); 75 StackHandleScope<2> hs(self); 78 result->SetL(Array::CreateMultiArray(self, h_class, h_dimensions)); 80 ScopedObjectAccessUnchecked soa(self); 325 ExecuteGotoImpl(Thread* self, MethodHelper& mh, const DexFile::CodeItem* code_item, ShadowFrame& shadow_frame, JValue result_register) argument 353 Execute(Thread* self, MethodHelper& mh, const DexFile::CodeItem* code_item, ShadowFrame& shadow_frame, JValue result_register) argument 397 EnterInterpreterFromInvoke(Thread* self, ArtMethod* method, Object* receiver, uint32_t* args, JValue* result) argument 510 EnterInterpreterFromStub(Thread* self, MethodHelper& mh, const DexFile::CodeItem* code_item, ShadowFrame& shadow_frame) argument 522 artInterpreterToInterpreterBridge(Thread* self, MethodHelper& mh, const DexFile::CodeItem* code_item, ShadowFrame* shadow_frame, JValue* result) argument [all...] |
H A D | interpreter.h | 37 extern void EnterInterpreterFromInvoke(Thread* self, mirror::ArtMethod* method, 41 extern void EnterInterpreterFromDeoptimize(Thread* self, ShadowFrame* shadow_frame, 45 extern JValue EnterInterpreterFromStub(Thread* self, MethodHelper& mh, 50 extern "C" void artInterpreterToInterpreterBridge(Thread* self, MethodHelper& mh, 57 extern "C" void artInterpreterToCompiledCodeBridge(Thread* self, MethodHelper& mh,
|
H A D | interpreter_common.cc | 30 bool DoFieldGet(Thread* self, ShadowFrame& shadow_frame, const Instruction* inst, argument 34 ArtField* f = FindFieldFromCode<find_type, do_access_check>(field_idx, shadow_frame.GetMethod(), self, 37 CHECK(self->IsExceptionPending()); 50 f->GetDeclaringClass()->AssertInitializedOrInitializingInThread(self); 55 instrumentation->FieldReadEvent(self, this_object, shadow_frame.GetMethod(), 89 template bool DoFieldGet<_find_type, _field_type, _do_check>(Thread* self, \ 205 bool DoFieldPut(Thread* self, const ShadowFrame& shadow_frame, const Instruction* inst, argument 210 ArtField* f = FindFieldFromCode<find_type, do_access_check>(field_idx, shadow_frame.GetMethod(), self, 213 CHECK(self->IsExceptionPending()); 227 f->GetDeclaringClass()->AssertInitializedOrInitializingInThread(self); 450 FindNextInstructionFollowingException(Thread* self, ShadowFrame& shadow_frame, uint32_t dex_pc, const instrumentation::Instrumentation* instrumentation) argument 513 AbortTransaction(Thread* self, const char* fmt, ...) argument 524 DoCall(ArtMethod* method, Thread* self, ShadowFrame& shadow_frame, const Instruction* inst, uint16_t inst_data, JValue* result) argument 661 DoFilledNewArray(const Instruction* inst, const ShadowFrame& shadow_frame, Thread* self, JValue* result) argument 795 UnstartedRuntimeInvoke(Thread* self, MethodHelper& mh, const DexFile::CodeItem* code_item, ShadowFrame* shadow_frame, JValue* result, size_t arg_offset) argument [all...] |
/art/runtime/jdwp/ |
H A D | object_registry.cc | 91 Thread* self = Thread::Current(); local 92 MutexLock mu(self, lock_); 93 return ContainsLocked(self, o, identity_hash_code, out_entry); 96 bool ObjectRegistry::ContainsLocked(Thread* self, mirror::Object* o, int32_t identity_hash_code, argument 102 if (o == self->DecodeJObject(entry->jni_reference)) { 113 Thread* self = Thread::Current(); local 114 MutexLock mu(self, lock_); 117 JNIEnv* env = self->GetJniEnv(); 133 Thread* self = Thread::Current(); local 134 MutexLock mu(self, lock 147 Thread* self = Thread::Current(); local 156 Thread* self = Thread::Current(); local 164 Thread* self = Thread::Current(); local 173 Thread* self = Thread::Current(); local 184 Thread* self = Thread::Current(); local 194 Thread* self = Thread::Current(); local 208 Thread* self = Thread::Current(); local [all...] |
/art/runtime/base/ |
H A D | mutex.cc | 150 void BaseMutex::CheckSafeToWait(Thread* self) { argument 151 if (self == NULL) { 156 CHECK(self->GetHeldMutex(level_) == this || level_ == kMonitorLock) 161 BaseMutex* held_mutex = self->GetHeldMutex(static_cast<LockLevel>(i)); 305 void Mutex::ExclusiveLock(Thread* self) { argument 306 DCHECK(self == NULL || self == Thread::Current()); 308 AssertNotHeld(self); 310 if (!recursive_ || !IsExclusiveHeld(self)) { 320 ScopedContentionRecorder scr(this, SafeGetTid(self), GetExclusiveOwnerTi 348 ExclusiveTryLock(Thread* self) argument 389 ExclusiveUnlock(Thread* self) argument 484 ExclusiveLock(Thread* self) argument 518 ExclusiveUnlock(Thread* self) argument 553 ExclusiveLockWithTimeout(Thread* self, int64_t ms, int32_t ns) argument 607 SharedTryLock(Thread* self) argument 697 Broadcast(Thread* self) argument 725 Signal(Thread* self) argument 742 Wait(Thread* self) argument 747 WaitHoldingLocks(Thread* self) argument 782 TimedWait(Thread* self, int64_t ms, int32_t ns) argument [all...] |
/art/tools/ |
H A D | cpplint.py | 437 # self._section will move monotonically through this set. If it ever 460 def __init__(self): 461 dict.__init__(self) 463 self._section = self._INITIAL_SECTION 465 self._last_header = '' 467 def CanonicalizeAlphabeticalOrder(self, header_path): 482 def IsInAlphabeticalOrder(self, header_path): 491 canonical_header = self.CanonicalizeAlphabeticalOrder(header_path) 492 if self [all...] |
/art/runtime/entrypoints/interpreter/ |
H A D | interpreter_entrypoints.h | 36 void (*pInterpreterToInterpreterBridge)(Thread* self, MethodHelper& mh, 39 void (*pInterpreterToCompiledCodeBridge)(Thread* self, MethodHelper& mh,
|
/art/compiler/utils/ |
H A D | dedupe_set_test.cc | 36 Thread* self = Thread::Current(); local 46 array1 = deduplicator.Add(self, test1); 57 array2 = deduplicator.Add(self, test1); 69 array3 = deduplicator.Add(self, test1);
|
/art/runtime/gc/space/ |
H A D | valgrind_malloc_space.h | 33 mirror::Object* AllocWithGrowth(Thread* self, size_t num_bytes, size_t* bytes_allocated, 35 mirror::Object* Alloc(Thread* self, size_t num_bytes, size_t* bytes_allocated, 40 size_t Free(Thread* self, mirror::Object* ptr) OVERRIDE 43 size_t FreeList(Thread* self, size_t num_ptrs, mirror::Object** ptrs) OVERRIDE
|