/art/runtime/gc/ |
H A D | task_processor.cc | 36 void TaskProcessor::AddTask(Thread* self, HeapTask* task) { argument 37 ScopedThreadStateChange tsc(self, kBlocked); 38 MutexLock mu(self, *lock_); 40 cond_->Signal(self); 43 HeapTask* TaskProcessor::GetTask(Thread* self) { argument 44 ScopedThreadStateChange tsc(self, kBlocked); 45 MutexLock mu(self, *lock_); 51 cond_->Wait(self); // Empty queue, wait until we are signalled. 68 cond_->TimedWait(self, static_cast<int64_t>(ms_delta), static_cast<int32_t>(ns_delta)); 74 void TaskProcessor::UpdateTargetRunTime(Thread* self, HeapTas argument 106 Stop(Thread* self) argument 113 Start(Thread* self) argument 119 RunAllTasks(Thread* self) argument [all...] |
H A D | reference_processor.cc | 52 void ReferenceProcessor::DisableSlowPath(Thread* self) { argument 54 condition_.Broadcast(self); 57 void ReferenceProcessor::BroadcastForSlowPath(Thread* self) { argument 59 MutexLock mu(self, *Locks::reference_processor_lock_); 60 condition_.Broadcast(self); 63 mirror::Object* ReferenceProcessor::GetReferent(Thread* self, mirror::Reference* reference) { argument 64 if (!kUseReadBarrier || self->GetWeakRefAccessEnabled()) { 75 MutexLock mu(self, *Locks::reference_processor_lock_); 77 (kUseReadBarrier && !self->GetWeakRefAccessEnabled())) { 101 condition_.WaitHoldingLocks(self); 106 StartPreservingReferences(Thread* self) argument 111 StopPreservingReferences(Thread* self) argument 123 Thread* self = Thread::Current(); local 198 Thread* self = Thread::Current(); local 238 EnqueueClearedReferences(Thread* self) argument 264 Thread* self = Thread::Current(); local [all...] |
H A D | task_processor_test.cc | 36 virtual void Run(Thread* self) OVERRIDE { 38 task_processor_->AddTask(self, 55 virtual void Run(Thread* self) OVERRIDE { 56 task_processor_->RunAllTasks(self); 67 Thread* const self = Thread::Current(); local 72 task_processor.AddTask(self, new RecursiveTask(&task_processor, &counter, kRecursion)); 73 task_processor.Start(self); 75 thread_pool.AddTask(self, new WorkUntilDoneTask(&task_processor, &done_running)); 76 thread_pool.StartWorkers(self); 83 task_processor.Stop(self); 121 Thread* const self = Thread::Current(); local [all...] |
H A D | reference_queue_test.cc | 29 Thread* self = Thread::Current(); local 30 ScopedObjectAccess soa(self); 31 StackHandleScope<20> hs(self); 37 Runtime::Current()->GetClassLinker()->FindClass(self, "Ljava/lang/ref/WeakReference;", 40 auto ref1(hs.NewHandle(ref_class->AllocObject(self)->AsReference())); 42 auto ref2(hs.NewHandle(ref_class->AllocObject(self)->AsReference())); 63 Thread* self = Thread::Current(); local 64 ScopedObjectAccess soa(self); 65 StackHandleScope<20> hs(self); 70 Runtime::Current()->GetClassLinker()->FindClass(self, "Ljav [all...] |
/art/runtime/ |
H A D | monitor_test.cc | 60 static void FillHeap(Thread* self, ClassLinker* class_linker, 66 hsp->reset(new StackHandleScope<kMaxHandles>(self)); 68 Handle<mirror::Class> c((*hsp)->NewHandle(class_linker->FindSystemClass(self, 71 Handle<mirror::Class> ca((*hsp)->NewHandle(class_linker->FindSystemClass(self, 78 mirror::ObjectArray<mirror::Object>::Alloc(self, ca.Get(), length / 4))); 79 if (self->IsExceptionPending() || h.Get() == nullptr) { 80 self->ClearException(); 95 while (!self->IsExceptionPending()) { 96 MutableHandle<mirror::Object> h = (*hsp)->NewHandle<mirror::Object>(c->AllocObject(self)); 97 if (!self 113 Run(Thread* self) argument 187 Run(Thread* self) argument 225 Run(Thread* self) argument 267 Run(Thread* self) argument 293 Thread* const self = Thread::Current(); local [all...] |
H A D | thread_pool_test.cc | 31 void Run(Thread* self) { argument 33 LOG(INFO) << "Running: " << *self; 62 Thread* self = Thread::Current(); local 67 thread_pool.AddTask(self, new CountTask(&count)); 69 thread_pool.StartWorkers(self); 71 thread_pool.Wait(self, true, false); 77 Thread* self = Thread::Current(); local 82 thread_pool.AddTask(self, new CountTask(&count)); 88 thread_pool.StartWorkers(self); 90 thread_pool.StopWorkers(self); 108 Run(Thread* self) argument 129 Thread* self = Thread::Current(); local [all...] |
H A D | barrier_test.cc | 35 void Run(Thread* self) { argument 36 LOG(INFO) << "Before barrier" << *self; 38 barrier_->Wait(self); 40 LOG(INFO) << "After barrier" << *self; 62 Thread* self = Thread::Current(); local 69 thread_pool.AddTask(self, new CheckWaitTask(&barrier, &count1, &count2)); 71 thread_pool.StartWorkers(self); 73 timeout_barrier.Increment(self, 1, 100); // sleep 100 msecs 78 barrier.Wait(self); 80 thread_pool.Wait(self, tru 94 Run(Thread* self) argument 113 Thread* self = Thread::Current(); local [all...] |
H A D | barrier.h | 42 void Pass(Thread* self) REQUIRES(!lock_); 45 void Wait(Thread* self) REQUIRES(!lock_); 54 void Increment(Thread* self, int delta) REQUIRES(!lock_); 58 bool Increment(Thread* self, int delta, uint32_t timeout_ms) REQUIRES(!lock_); 62 void Init(Thread* self, int count) REQUIRES(!lock_); 65 void SetCountLocked(Thread* self, int count) REQUIRES(lock_);
|
H A D | monitor_pool_test.cc | 39 static void VerifyMonitor(Monitor* mon, Thread* self) { argument 43 EXPECT_EQ(MonitorPool::ComputeMonitorId(mon, self), mon->GetMonitorId()); 58 Thread* self = Thread::Current(); local 59 ScopedObjectAccess soa(self); 74 Monitor* mon = MonitorPool::CreateMonitor(self, self, nullptr, static_cast<int32_t>(i)); 77 VerifyMonitor(mon, self); 85 VerifyMonitor(mon, self); 87 MonitorPool::ReleaseMonitor(self, mon); 97 Monitor* mon = MonitorPool::CreateMonitor(self, sel [all...] |
H A D | thread_pool.cc | 75 Thread* self = Thread::Current(); local 77 thread_pool_->creation_barier_.Wait(self); 78 while ((task = thread_pool_->GetTask(self)) != nullptr) { 79 task->Run(self); 94 void ThreadPool::AddTask(Thread* self, Task* task) { argument 95 MutexLock mu(self, task_queue_lock_); 99 task_queue_condition_.Signal(self); 103 void ThreadPool::RemoveAllTasks(Thread* self) { argument 104 MutexLock mu(self, task_queue_lock_); 121 Thread* self local 140 Thread* self = Thread::Current(); local 152 StartWorkers(Thread* self) argument 160 StopWorkers(Thread* self) argument 165 GetTask(Thread* self) argument 197 TryGetTask(Thread* self) argument 211 Wait(Thread* self, bool do_work, bool may_hold_locks) argument 230 GetTaskCount(Thread* self) argument [all...] |
H A D | monitor.cc | 76 Monitor::Monitor(Thread* self, Thread* owner, mirror::Object* obj, int32_t hash_code) argument 87 monitor_id_(MonitorPool::ComputeMonitorId(this, self)) { 94 CHECK(owner == nullptr || owner == self || owner->IsSuspended()); 98 Monitor::Monitor(Thread* self, Thread* owner, mirror::Object* obj, int32_t hash_code, argument 116 CHECK(owner == nullptr || owner == self || owner->IsSuspended()); 130 bool Monitor::Install(Thread* self) { argument 131 MutexLock mu(self, monitor_lock_); // Uncontended mutex acquisition as monitor isn't yet public. 132 CHECK(owner_ == nullptr || owner_ == self || owner_->IsSuspended()); 257 void Monitor::AtraceMonitorLock(Thread* self, mirror::Object* obj, bool is_wait) { argument 259 AtraceMonitorLockImpl(self, ob 263 AtraceMonitorLockImpl(Thread* self, mirror::Object* obj, bool is_wait) argument 317 Lock(Thread* self) argument 444 Thread* self = Thread::Current(); local 533 Unlock(Thread* self) argument 563 Wait(Thread* self, int64_t ms, int32_t ns, bool interruptShouldThrow, ThreadState why) argument 702 Notify(Thread* self) argument 725 NotifyAll(Thread* self) argument 742 Deflate(Thread* self, mirror::Object* obj) argument 791 Inflate(Thread* self, Thread* owner, mirror::Object* obj, int32_t hash_code) argument 812 InflateThinLocked(Thread* self, Handle<mirror::Object> obj, LockWord lock_word, uint32_t hash_code) argument 855 MonitorEnter(Thread* self, mirror::Object* obj) argument 934 MonitorExit(Thread* self, mirror::Object* obj) argument 993 Wait(Thread* self, mirror::Object *obj, int64_t ms, int32_t ns, bool interruptShouldThrow, ThreadState why) argument 1030 DoNotify(Thread* self, mirror::Object* obj, bool notify_all) argument 1096 Thread* self = Thread::Current(); local 1282 Thread* self = Thread::Current(); local 1298 Thread* self = Thread::Current(); local 1306 Thread* self = Thread::Current(); local 1312 Thread* self = Thread::Current(); local 1322 Thread* self = Thread::Current(); local [all...] |
/art/test/utils/python/ |
H A D | generate_java_main.py | 64 def __init__(self): 68 self.tests = set() 69 self.global_funcs = set() 71 def add_instance(self, it): 75 self.tests.add(it) 77 def add_func(self, f): 81 self.global_funcs.add(f) 83 def get_name(self): 89 def __str__(self): 93 all_tests = sorted(self [all...] |
/art/test/968-default-partial-compile-generated/util-src/ |
H A D | generate_java.py | 43 def __init__(self, inner): 44 self.inner = inner 46 def get_name(self): 48 return self.inner.get_name() 50 def __str__(self): 52 for line in str(self.inner).splitlines(keepends = True): 58 def __init__(self, sources, javac, temp_dir, classes_dir): 59 self.javac = javac 60 self.temp_dir = temp_dir 61 self [all...] |
/art/test/971-iface-super/util-src/ |
H A D | generate_java.py | 43 def __init__(self, inner): 44 self.inner = inner 46 def get_name(self): 48 return self.inner.get_name() 50 def __str__(self): 52 for line in str(self.inner).splitlines(keepends = True): 58 def __init__(self, sources, javac, temp_dir, classes_dir): 59 self.javac = javac 60 self.temp_dir = temp_dir 61 self [all...] |
/art/tools/ |
H A D | stream-trace-converter.py | 71 def PrintHeader(self, header): 78 def ProcessDataHeader(self, input, body): 103 self._mRecordSize = 9; 105 self._mRecordSize = 10; 107 self._mRecordSize = ReadShortLE(input) 108 WriteShortLE(body, self._mRecordSize) 114 def ProcessMethod(self, input): 117 self._methods.append(str) 120 def ProcessThread(self, input): 124 self [all...] |
/art/runtime/base/ |
H A D | mutex.h | 160 void RegisterAsLocked(Thread* self); 161 void RegisterAsUnlocked(Thread* self); 162 void CheckSafeToWait(Thread* self); 223 void ExclusiveLock(Thread* self) ACQUIRE(); 224 void Lock(Thread* self) ACQUIRE() { ExclusiveLock(self); } 227 bool ExclusiveTryLock(Thread* self) TRY_ACQUIRE(true); 228 bool TryLock(Thread* self) TRY_ACQUIRE(true) { return ExclusiveTryLock(self); } 231 void ExclusiveUnlock(Thread* self) RELEAS 252 AssertNotHeldExclusive(self); variable 355 AssertNotExclusiveHeld(self); variable [all...] |
/art/runtime/interpreter/ |
H A D | unstarted_runtime_test.cc | 47 static void Unstarted ## Name(Thread* self, \ 52 interpreter::UnstartedRuntime::Unstarted ## Name(self, shadow_frame, result, arg_offset); \ 62 static void UnstartedJNI ## Name(Thread* self, \ 68 interpreter::UnstartedRuntime::UnstartedJNI ## Name(self, method, receiver, args, result); \ 82 Thread* self, 87 mirror::Class* array_type = runtime->GetClassLinker()->FindArrayClass(self, &component_type); 90 mirror::ObjectArray<mirror::Object>::Alloc(self, array_type, 3); 94 CHECK(!self->IsExceptionPending()); 109 void RunArrayCopy(Thread* self, 124 UnstartedSystemArraycopy(self, tm 201 Thread* self = Thread::Current(); local 223 Thread* self = Thread::Current(); local 248 Thread* self = Thread::Current(); local 273 Thread* self = Thread::Current(); local 298 Thread* self = Thread::Current(); local 358 Thread* self = Thread::Current(); local 382 Thread* self = Thread::Current(); local 412 Thread* self = Thread::Current(); local 447 Thread* self = Thread::Current(); local 553 Thread* self = Thread::Current(); local 601 Thread* self = Thread::Current(); local 648 Thread* self = Thread::Current(); local 677 Thread* self = Thread::Current(); local 706 Thread* self = Thread::Current(); local 807 Thread* self = Thread::Current(); local 826 Thread* self = Thread::Current(); local 850 Thread* self = Thread::Current(); local [all...] |
H A D | interpreter.h | 38 extern void EnterInterpreterFromInvoke(Thread* self, ArtMethod* method, 44 extern void EnterInterpreterFromDeoptimize(Thread* self, ShadowFrame* shadow_frame, bool from_code, 48 extern JValue EnterInterpreterFromEntryPoint(Thread* self, const DexFile::CodeItem* code_item, 52 void ArtInterpreterToInterpreterBridge(Thread* self, const DexFile::CodeItem* code_item, 59 void InitInterpreterTls(Thread* self);
|
/art/runtime/lambda/ |
H A D | leaking_allocator.cc | 25 void* LeakingAllocator::AllocateMemoryImpl(Thread* self, size_t byte_size, size_t align_size) { argument 27 void* mem = Runtime::Current()->GetLinearAlloc()->Alloc(self, byte_size);
|
H A D | leaking_allocator.h | 48 static AlignedMemoryStorage<T>* AllocateMemory(Thread* self, size_t byte_size = sizeof(T)) { argument 50 AllocateMemoryImpl(self, byte_size, alignof(T))); 55 static T* MakeFlexibleInstance(Thread* self, size_t byte_size, Args&&... args) { argument 56 return new (AllocateMemory<T>(self, byte_size)) T(std::forward<Args>(args)...); 61 static T* MakeInstance(Thread* self, Args&&... args) { argument 62 return new (AllocateMemory<T>(self, sizeof(T))) T(std::forward<Args>(args)...); 66 static void* AllocateMemoryImpl(Thread* self, size_t byte_size, size_t align_size);
|
/art/compiler/dex/quick/ |
H A D | dex_file_to_method_inliner_map.cc | 41 Thread* self = Thread::Current(); local 43 ReaderMutexLock mu(self, lock_); 58 WriterMutexLock mu(self, lock_); 66 locked_inliner->lock_.ExclusiveLock(self); // Acquire inliner's lock_ before releasing lock_. 69 locked_inliner->lock_.ExclusiveUnlock(self);
|
/art/runtime/arch/ |
H A D | stub_test.cc | 63 Thread::tls_ptr_sized_values* GetTlsPtr(Thread* self) { argument 64 return &self->tlsPtr_; 68 size_t Invoke3(size_t arg0, size_t arg1, size_t arg2, uintptr_t code, Thread* self) { argument 69 return Invoke3WithReferrer(arg0, arg1, arg2, code, self, nullptr); 73 size_t Invoke3WithReferrer(size_t arg0, size_t arg1, size_t arg2, uintptr_t code, Thread* self, argument 75 return Invoke3WithReferrerAndHidden(arg0, arg1, arg2, code, self, referrer, 0); 80 Thread* self, ArtMethod* referrer, size_t hidden) { 83 self->PushManagedStackFragment(&fragment); 155 "str %[self], [sp, #16]\n\t" 173 : [arg0] "r"(arg0), [arg1] "r"(arg1), [arg2] "r"(arg2), [code] "r"(code), [self] " 79 Invoke3WithReferrerAndHidden(size_t arg0, size_t arg1, size_t arg2, uintptr_t code, Thread* self, ArtMethod* referrer, size_t hidden) argument 530 GetEntrypoint(Thread* self, QuickEntrypointEnum entrypoint) argument 547 Thread* self = Thread::Current(); local 557 10 * sizeof(uint32_t), StubTest::GetEntrypoint(self, kQuickMemcpy), self); local 587 Thread* self = Thread::Current(); local 602 Invoke3(reinterpret_cast<size_t>(obj.Get()), 0U, 0U, art_quick_lock_object, self); local 610 Invoke3(reinterpret_cast<size_t>(obj.Get()), 0U, 0U, art_quick_lock_object, self); local 626 Invoke3(reinterpret_cast<size_t>(obj2.Get()), 0U, 0U, art_quick_lock_object, self); local 661 Thread* self = Thread::Current(); variable 676 test->Invoke3(reinterpret_cast<size_t>(obj.Get()), 0U, 0U, art_quick_unlock_object, self); variable 685 test->Invoke3(reinterpret_cast<size_t>(obj.Get()), 0U, 0U, art_quick_lock_object, self); variable 691 test->Invoke3(reinterpret_cast<size_t>(obj.Get()), 0U, 0U, art_quick_unlock_object, self); variable 746 self); variable 750 art_quick_unlock_object, self); variable 782 self); variable 816 Thread* self = Thread::Current(); local 833 art_quick_check_cast, self); local 838 art_quick_check_cast, self); local 843 art_quick_check_cast, self); local 850 art_quick_check_cast, self); local 866 Thread* self = Thread::Current(); local 901 art_quick_aput_obj_with_null_and_bound_check, self); local 907 art_quick_aput_obj_with_null_and_bound_check, self); local 913 art_quick_aput_obj_with_null_and_bound_check, self); local 919 art_quick_aput_obj_with_null_and_bound_check, self); local 927 art_quick_aput_obj_with_null_and_bound_check, self); local 933 art_quick_aput_obj_with_null_and_bound_check, self); local 939 art_quick_aput_obj_with_null_and_bound_check, self); local 945 art_quick_aput_obj_with_null_and_bound_check, self); local 974 art_quick_aput_obj_with_null_and_bound_check, self); local 982 art_quick_aput_obj_with_null_and_bound_check, self); local 1003 Thread* self = Thread::Current(); local 1128 Thread* self = Thread::Current(); local 1212 Thread* self = Thread::Current(); local 1685 self, local 1730 self, local 1762 self, local 1787 TestFields(Thread* self, StubTest* test, Primitive::Type test_type) argument 1881 Thread* self = Thread::Current(); local 1893 Thread* self = Thread::Current(); local 1905 Thread* self = Thread::Current(); local 1916 Thread* self = Thread::Current(); local 1927 Thread* self = Thread::Current(); local 1945 Thread* self = Thread::Current(); local 2083 Thread* self = Thread::Current(); local 2158 Thread* self = Thread::Current(); local 2194 Thread* self = Thread::Current(); local [all...] |
/art/runtime/mirror/ |
H A D | array.cc | 43 static Array* RecursiveCreateMultiArray(Thread* self, 48 StackHandleScope<1> hs(self); 51 Array::Alloc<true>(self, array_class.Get(), array_length, 55 CHECK(self->IsExceptionPending()); 61 StackHandleScope<1> hs2(self); 63 Array* sub_array = RecursiveCreateMultiArray(self, h_component_type, 66 CHECK(self->IsExceptionPending()); 76 Array* Array::CreateMultiArray(Thread* self, Handle<Class> element_class, argument 97 StackHandleScope<1> hs(self); 99 hs.NewHandle(class_linker->FindArrayClass(self, 128 CopyOf(Thread* self, int32_t new_length) argument [all...] |
H A D | field-inl.h | 31 inline mirror::Field* Field::CreateFromArtField(Thread* self, ArtField* field, argument 33 StackHandleScope<2> hs(self); 40 self->AssertPendingException(); 45 mirror::Throwable* exception = self->GetException(); 49 self->ClearException(); 52 auto ret = hs.NewHandle(static_cast<Field*>(StaticClass()->AllocObject(self))); 54 self->AssertPendingOOMException();
|
/art/runtime/entrypoints/quick/ |
H A D | quick_instrumentation_entrypoints.cc | 29 Thread* self, 34 ScopedQuickEntrypointChecks sqec(self, kIsDebugBuild, false); 44 instrumentation->PushInstrumentationStackFrame(self, method->IsStatic() ? nullptr : this_object, 50 extern "C" TwoWordReturn artInstrumentationMethodExitFromCode(Thread* self, ArtMethod** sp, 55 CHECK(!self->IsExceptionPending()) << "Enter instrumentation exit stub with pending exception " 56 << self->GetException()->Dump(); 69 self, return_pc, gpr_result, fpr_result);
|