/art/runtime/jit/ |
H A D | debugger_interface.cc | 119 Thread* self = Thread::Current(); local 120 MutexLock mu(self, g_jit_debug_mutex); 125 Thread* self = Thread::Current(); local 126 MutexLock mu(self, g_jit_debug_mutex); 135 Thread* self = Thread::Current(); local 136 MutexLock mu(self, g_jit_debug_mutex); 144 Thread* self = Thread::Current(); local 145 MutexLock mu(self, g_jit_debug_mutex);
|
H A D | profile_compilation_info_test.cc | 38 Thread* self = Thread::Current(); local 39 ScopedObjectAccess soa(self); 40 StackHandleScope<1> hs(self); 42 reinterpret_cast<mirror::ClassLoader*>(self->DecodeJObject(class_loader)))); 43 mirror::Class* klass = class_linker->FindClass(self, clazz.c_str(), h_loader); 96 Thread* self = Thread::Current(); local 99 ScopedObjectAccess soa(self); 114 ScopedObjectAccess soa(self); 130 ScopedObjectAccess soa(self);
|
/art/runtime/native/ |
H A D | java_lang_System.cc | 41 Thread* self = Thread::Current(); local 42 self->ThrowNewExceptionF("Ljava/lang/ArrayStoreException;",
|
H A D | org_apache_harmony_dalvik_ddmc_DdmVmInternal.cc | 50 Thread* const self = Thread::Current(); local 51 if (static_cast<uint32_t>(thin_lock_id) == self->GetThreadId()) { 54 jobject internal_trace = self->CreateInternalStackTrace<false>(soa); 132 Thread* self = static_cast<JNIEnvExt*>(env)->self; local 134 MutexLock mu(self, *Locks::thread_list_lock_);
|
/art/runtime/ |
H A D | reflection_test.cc | 91 Thread* self = Thread::Current(); local 92 StackHandleScope<2> hs(self); 95 ScopedObjectAccessUnchecked(self).Decode<mirror::ClassLoader*>(jclass_loader))); 97 MakeExecutable(ScopedObjectAccessUnchecked(self).Decode<mirror::ClassLoader*>(jclass_loader), 102 MakeExecutable(ScopedObjectAccessUnchecked(self).Decode<mirror::ClassLoader*>(jclass_loader), 106 mirror::Class* c = class_linker_->FindClass(self, DotToDescriptor(class_name).c_str(), 118 StackHandleScope<1> hs2(self); 120 bool initialized = class_linker_->EnsureInitialized(self, h_class, true, true); 122 *receiver = c->AllocObject(self); 128 self [all...] |
H A D | signal_catcher.cc | 48 // On Android, /proc/self/cmdline will have been rewritten to something like "system_server". 51 if (ReadFileToString("/proc/self/cmdline", ¤t_cmd_line)) { 77 Thread* self = Thread::Current(); local 78 MutexLock mu(self, lock_); 80 cond_.Wait(self); 148 if (ReadFileToString("/proc/self/maps", &maps)) { 149 os << "/proc/self/maps:\n" << maps; 161 int SignalCatcher::WaitForSignal(Thread* self, SignalSet& signals) { argument 162 ScopedThreadStateChange tsc(self, kWaitingInMainSignalCatcherLoop); 172 LOG(INFO) << *self << " 189 Thread* self = Thread::Current(); local [all...] |
H A D | thread-inl.h | 39 return full_env->self; 261 DCHECK(Thread::Current() == this) << "Should be called by self"; 273 // Note: self is not necessarily equal to this thread since thread may be suspended. 274 Thread* self = Thread::Current(); local 275 DCHECK(this == self || IsSuspended() || GetState() == kWaitingPerformingGc) 276 << GetState() << " thread " << this << " self " << self;
|
H A D | class_linker-inl.h | 32 inline mirror::Class* ClassLinker::FindSystemClass(Thread* self, const char* descriptor) { argument 33 return FindClass(self, descriptor, ScopedNullHandle<mirror::ClassLoader>()); 36 inline mirror::Class* ClassLinker::FindArrayClass(Thread* self, mirror::Class** element_class) { argument 50 mirror::Class* array_class = FindClass(self, descriptor.c_str(), class_loader); 58 self->AssertPendingException(); 141 inline ArtMethod* ClassLinker::ResolveMethod(Thread* self, argument 148 StackHandleScope<2> hs(self); 189 inline mirror::Object* ClassLinker::AllocObject(Thread* self) { argument 191 self, 196 inline mirror::ObjectArray<T>* ClassLinker::AllocObjectArray(Thread* self, size_ argument 200 AllocClassArray(Thread* self, size_t length) argument 205 AllocStringArray(Thread* self, size_t length) argument 212 AllocIfTable(Thread* self, size_t ifcount) argument [all...] |
H A D | fault_handler.cc | 154 Thread* self = Thread::Current(); local 156 DCHECK(self != nullptr); 213 if (setjmp(*self->GetNestedSignalState()) == 0) { 430 Thread* self = Thread::Current(); local 434 self->SetTopOfStack(reinterpret_cast<ArtMethod**>(sp)); 441 self->DumpJavaStack(LOG(ERROR));
|
H A D | handle_scope.h | 148 explicit ALWAYS_INLINE StackHandleScope(Thread* self, mirror::Object* fill_value = nullptr); 198 explicit StackHandleScopeCollection(Thread* const self) : argument 199 self_(self),
|
H A D | indirect_reference_table.cc | 183 auto* self = Thread::Current(); local 184 if (self->HandleScopeContains(reinterpret_cast<jobject>(iref))) { 185 auto* env = self->GetJniEnv(); 188 ScopedObjectAccess soa(self); 191 self->Dump(LOG(WARNING));
|
/art/compiler/ |
H A D | common_compiler_test.cc | 113 Thread* self = Thread::Current(); local 114 StackHandleScope<1> hs(self); 116 mirror::Class* klass = class_linker_->FindClass(self, class_descriptor.c_str(), loader); 234 Thread* self = Thread::Current(); local 235 StackHandleScope<1> hs(self); 237 mirror::Class* klass = class_linker_->FindClass(self, class_descriptor.c_str(), loader); 258 Thread* self = Thread::Current(); local 259 mirror::Class* klass = class_linker_->FindClass(self, class_descriptor.c_str(), class_loader); 272 Thread* self = Thread::Current(); local 273 mirror::Class* klass = class_linker_->FindClass(self, class_descripto [all...] |
H A D | elf_writer_quick.cc | 224 Thread* self = Thread::Current(); local 233 debug_info_thread_pool_->AddTask(self, debug_info_task_.get()); 234 debug_info_thread_pool_->StartWorkers(self); 248 Thread* self = Thread::Current(); local 250 debug_info_thread_pool_->Wait(self, true, false);
|
/art/compiler/driver/ |
H A D | compiled_method_storage.cc | 194 Thread* self = Thread::Current(); local 195 os << "\nCode dedupe: " << dedupe_code_.DumpStats(self); 196 os << "\nVmap table dedupe: " << dedupe_vmap_table_.DumpStats(self); 197 os << "\nCFI info dedupe: " << dedupe_cfi_info_.DumpStats(self);
|
H A D | compiler_driver_test.cc | 191 Thread* self = Thread::Current(); local 194 ScopedObjectAccess soa(self); 208 ScopedObjectAccess soa(self); 209 StackHandleScope<1> hs(self); 211 reinterpret_cast<mirror::ClassLoader*>(self->DecodeJObject(class_loader)))); 212 mirror::Class* klass = class_linker->FindClass(self, "LStaticLeafMethods;", h_loader); 265 Thread* self = Thread::Current(); local 266 ScopedObjectAccess soa(self); 267 StackHandleScope<1> hs(self); 269 reinterpret_cast<mirror::ClassLoader*>(self 295 Thread* self = Thread::Current(); local [all...] |
/art/compiler/utils/ |
H A D | dedupe_set-inl.h | 69 const StoreKey* Add(Thread* self, size_t hash, const InKey& in_key) REQUIRES(!lock_) { 70 MutexLock lock(self, lock_); 82 void UpdateStats(Thread* self, Stats* global_stats) REQUIRES(!lock_) { 87 MutexLock lock(self, lock_); 188 Thread* self, const InKey& key) { 200 return shards_[shard_bin]->Add(self, shard_hash, key); 236 Thread* self) const { 239 shards_[shard]->UpdateStats(self, &stats); 187 Add( Thread* self, const InKey& key) argument
|
/art/runtime/arch/x86/ |
H A D | fault_handler_x86.cc | 243 Thread* self = Thread::Current(); local 244 CHECK(self != nullptr); // This will cause a SIGABRT if self is null. 247 uc->CTX_JMP_BUF = reinterpret_cast<uintptr_t>(*self->GetNestedSignalState());
|
/art/runtime/entrypoints/ |
H A D | entrypoint_utils.cc | 43 Thread* self, 56 DCHECK(self->IsExceptionPending()); 65 self->ThrowNewExceptionF( 85 ArtMethod* referrer, Thread* self, 88 mirror::Class* klass = CheckFilledNewArrayAlloc(type_idx, component_count, referrer, self, 97 return mirror::Array::Alloc<false>(self, klass, component_count, 106 Thread* self, 109 mirror::Class* klass = CheckFilledNewArrayAlloc(type_idx, component_count, referrer, self, 117 return mirror::Array::Alloc<true>(self, klass, component_count, 122 void CheckReferenceResult(mirror::Object* o, Thread* self) { argument 84 CheckAndAllocArrayFromCode(uint32_t type_idx, int32_t component_count, ArtMethod* referrer, Thread* self, bool access_check, gc::AllocatorType ) argument 103 CheckAndAllocArrayFromCodeInstrumented(uint32_t type_idx, int32_t component_count, ArtMethod* referrer, Thread* self, bool access_check, gc::AllocatorType ) argument 246 Thread* self = Thread::Current(); local [all...] |
/art/runtime/gc/accounting/ |
H A D | mod_union_table_test.cc | 49 Thread* self, space::ContinuousMemMapAllocSpace* space, size_t component_count) 51 auto* klass = GetObjectArrayClass(self, space); 55 space->Alloc(self, size, &bytes_allocated, nullptr, &bytes_tl_bulk_allocated)); 70 mirror::Class* GetObjectArrayClass(Thread* self, space::ContinuousMemMapAllocSpace* space) 83 auto* klass = down_cast<mirror::Class*>(space->Alloc(self, class_size, &bytes_allocated, 177 Thread* const self = Thread::Current(); local 178 ScopedObjectAccess soa(self); 189 ScopedThreadSuspension sts(self, kSuspended); 197 auto* obj1 = AllocObjectArray(self, space, CardTable::kCardSize); 199 auto* obj2 = AllocObjectArray(self, spac [all...] |
/art/runtime/gc/ |
H A D | allocation_record.cc | 215 Thread* self = Thread::Current(); local 219 MutexLock mu(self, *Locks::alloc_tracker_lock_); 231 self->GetThreadName(self_name); 233 records->alloc_ddm_thread_id_ = self->GetTid(); 243 MutexLock mu(self, *Locks::alloc_tracker_lock_); 250 MutexLock mu(self, *Locks::alloc_tracker_lock_); 264 void AllocRecordObjectMap::RecordAllocation(Thread* self, argument 270 AllocRecordStackVisitor visitor(self, max_stack_depth_, /*out*/ &trace); 272 StackHandleScope<1> hs(self); 277 MutexLock mu(self, *Lock [all...] |
/art/runtime/gc/collector/ |
H A D | garbage_collector.cc | 83 Thread* self = Thread::Current(); local 97 if (Locks::mutator_lock_->IsExclusiveHeld(self)) { 105 MutexLock mu(self, pause_histogram_lock_);
|
/art/runtime/gc/space/ |
H A D | bump_pointer_space.cc | 103 Thread* self = Thread::Current(); local 104 MutexLock mu(self, *Locks::runtime_shutdown_lock_); 105 MutexLock mu2(self, *Locks::thread_list_lock_); 123 Thread* self = Thread::Current(); local 124 MutexLock mu(self, *Locks::runtime_shutdown_lock_); 125 MutexLock mu2(self, *Locks::thread_list_lock_); 216 Thread* self = Thread::Current(); local 217 MutexLock mu(self, *Locks::runtime_shutdown_lock_); 218 MutexLock mu2(self, *Locks::thread_list_lock_); 234 Thread* self local 255 AllocNewTlab(Thread* self, size_t bytes) argument [all...] |
H A D | malloc_space.cc | 234 Thread* self = context->self; local 235 Locks::heap_bitmap_lock_->AssertExclusiveHeld(self); 248 context->freed.bytes += space->FreeList(self, num_ptrs, ptrs);
|
H A D | region_space.cc | 335 bool RegionSpace::AllocNewTlab(Thread* self) { argument 336 MutexLock mu(self, region_lock_); 337 RevokeThreadLocalBuffersLocked(self); 351 r->thread_ = self; 352 self->SetTlab(r->Begin(), r->End()); 382 Thread* self = Thread::Current(); local 383 MutexLock mu(self, *Locks::runtime_shutdown_lock_); 384 MutexLock mu2(self, *Locks::thread_list_lock_); 400 Thread* self = Thread::Current(); local 401 MutexLock mu(self, *Lock [all...] |
H A D | space_test.h | 55 mirror::Class* GetByteArrayClass(Thread* self) SHARED_REQUIRES(Locks::mutator_lock_) { 56 StackHandleScope<1> hs(self); 60 Runtime::Current()->GetClassLinker()->FindClass(self, "[B", null_loader); 62 byte_array_class_ = self->GetJniEnv()->NewLocalRef(byte_array_class); 65 return reinterpret_cast<mirror::Class*>(self->DecodeJObject(byte_array_class_)); 69 Thread* self, 75 StackHandleScope<1> hs(self); 76 Handle<mirror::Class> byte_array_class(hs.NewHandle(GetByteArrayClass(self))); 77 mirror::Object* obj = alloc_space->Alloc(self, 89 Thread* self, 175 Thread* self = Thread::Current(); local [all...] |