/art/runtime/gc/space/ |
H A D | large_object_space.h | 63 size_t FreeList(Thread* self, size_t num_ptrs, mirror::Object** ptrs) OVERRIDE; 102 virtual bool IsZygoteLargeObject(Thread* self, mirror::Object* obj) const = 0; 105 virtual void SetAllLargeObjectsAsZygoteObjects(Thread* self) = 0; 134 mirror::Object* Alloc(Thread* self, size_t num_bytes, size_t* bytes_allocated, 137 size_t Free(Thread* self, mirror::Object* ptr) REQUIRES(!lock_); 150 bool IsZygoteLargeObject(Thread* self, mirror::Object* obj) const OVERRIDE REQUIRES(!lock_); 151 void SetAllLargeObjectsAsZygoteObjects(Thread* self) OVERRIDE REQUIRES(!lock_); 168 mirror::Object* Alloc(Thread* self, size_t num_bytes, size_t* bytes_allocated, 171 size_t Free(Thread* self, mirror::Object* obj) OVERRIDE REQUIRES(!lock_); 192 bool IsZygoteLargeObject(Thread* self, mirro [all...] |
H A D | dlmalloc_space.h | 50 virtual mirror::Object* AllocWithGrowth(Thread* self, size_t num_bytes, size_t* bytes_allocated, 55 virtual mirror::Object* Alloc(Thread* self, size_t num_bytes, size_t* bytes_allocated, 58 return AllocNonvirtual(self, num_bytes, bytes_allocated, usable_size, 66 virtual size_t Free(Thread* self, mirror::Object* ptr) OVERRIDE 70 virtual size_t FreeList(Thread* self, size_t num_ptrs, mirror::Object** ptrs) OVERRIDE 87 mirror::Object* AllocNonvirtual(Thread* self, size_t num_bytes, size_t* bytes_allocated, 147 mirror::Object* AllocWithoutGrowthLocked(Thread* self, size_t num_bytes, size_t* bytes_allocated,
|
H A D | rosalloc_space.cc | 154 mirror::Object* RosAllocSpace::AllocWithGrowth(Thread* self, size_t num_bytes, argument 159 MutexLock mu(self, lock_); 164 result = AllocCommon(self, num_bytes, bytes_allocated, usable_size, 191 size_t RosAllocSpace::Free(Thread* self, mirror::Object* ptr) { argument 197 MutexLock mu(self, lock_); 200 return rosalloc_->Free(self, ptr); 203 size_t RosAllocSpace::FreeList(Thread* self, size_t num_ptrs, mirror::Object** ptrs) { argument 217 MutexLock mu(self, lock_); 237 const size_t bytes_freed = rosalloc_->BulkFree(self, reinterpret_cast<void**>(ptrs), num_ptrs); 247 Thread* const self local 305 Thread* self = Thread::Current(); variable 318 Thread* self = Thread::Current(); variable [all...] |
H A D | memory_tool_malloc_space-inl.h | 88 Thread* self, size_t num_bytes, size_t* bytes_allocated_out, size_t* usable_size_out, 93 void* obj_with_rdz = S::AllocWithGrowth(self, num_bytes + 2 * kMemoryToolRedZoneBytes, 117 Thread* self, size_t num_bytes, size_t* bytes_allocated_out, size_t* usable_size_out, 122 void* obj_with_rdz = S::Alloc(self, num_bytes + 2 * kMemoryToolRedZoneBytes, 145 Thread* self, size_t num_bytes, size_t* bytes_allocated_out, size_t* usable_size_out, 150 void* obj_with_rdz = S::AllocThreadUnsafe(self, num_bytes + 2 * kMemoryToolRedZoneBytes, 196 Thread* self, mirror::Object* ptr) { 213 return S::Free(self, reinterpret_cast<mirror::Object*>(obj_with_rdz)); 224 Thread* self, size_t num_ptrs, mirror::Object** ptrs) { 227 freed += Free(self, ptr 87 AllocWithGrowth( Thread* self, size_t num_bytes, size_t* bytes_allocated_out, size_t* usable_size_out, size_t* bytes_tl_bulk_allocated_out) argument 116 Alloc( Thread* self, size_t num_bytes, size_t* bytes_allocated_out, size_t* usable_size_out, size_t* bytes_tl_bulk_allocated_out) argument 144 AllocThreadUnsafe( Thread* self, size_t num_bytes, size_t* bytes_allocated_out, size_t* usable_size_out, size_t* bytes_tl_bulk_allocated_out) argument 195 Free( Thread* self, mirror::Object* ptr) argument 223 FreeList( Thread* self, size_t num_ptrs, mirror::Object** ptrs) argument [all...] |
H A D | zygote_space.h | 48 mirror::Object* Alloc(Thread* self, size_t num_bytes, size_t* bytes_allocated, 53 size_t Free(Thread* self, mirror::Object* ptr) OVERRIDE; 55 size_t FreeList(Thread* self, size_t num_ptrs, mirror::Object** ptrs) OVERRIDE;
|
/art/runtime/mirror/ |
H A D | string-inl.h | 137 Thread* self = Thread::Current(); local 138 self->ThrowNewExceptionF("Ljava/lang/StringIndexOutOfBoundsException;", 156 inline String* String::Alloc(Thread* self, int32_t utf16_length, gc::AllocatorType allocator_type, argument 178 self->ThrowOutOfMemoryError(StringPrintf("%s of length %d would overflow", 186 heap->AllocObjectWithAllocator<kIsInstrumented, true>(self, string_class, alloc_size, 191 inline String* String::AllocFromByteArray(Thread* self, int32_t byte_length, argument 195 String* string = Alloc<kIsInstrumented>(self, byte_length, allocator_type, visitor); 200 inline String* String::AllocFromCharArray(Thread* self, int32_t count, argument 206 String* new_string = Alloc<kIsInstrumented>(self, count, allocator_type, visitor); 211 inline String* String::AllocFromString(Thread* self, int32_ argument [all...] |
H A D | string.h | 86 ALWAYS_INLINE static String* Alloc(Thread* self, int32_t utf16_length, 92 ALWAYS_INLINE static String* AllocFromByteArray(Thread* self, int32_t byte_length, 99 ALWAYS_INLINE static String* AllocFromCharArray(Thread* self, int32_t count, 105 ALWAYS_INLINE static String* AllocFromString(Thread* self, int32_t string_length, 110 static String* AllocFromStrings(Thread* self, Handle<String> string, Handle<String> string2) 113 static String* AllocFromUtf16(Thread* self, int32_t utf16_length, const uint16_t* utf16_data_in) 116 static String* AllocFromModifiedUtf8(Thread* self, const char* utf) 119 static String* AllocFromModifiedUtf8(Thread* self, int32_t utf16_length, 123 static String* AllocFromModifiedUtf8(Thread* self, int32_t utf16_length, const char* utf8_data_in) 154 CharArray* ToCharArray(Thread* self) SHARED_REQUIRE [all...] |
/art/runtime/gc/collector/ |
H A D | concurrent_copying.cc | 61 Thread* self = Thread::Current(); local 63 ReaderMutexLock mu(self, *Locks::heap_bitmap_lock_); 71 MutexLock mu(self, mark_stack_lock_); 97 Thread* self = Thread::Current(); local 98 thread_running_gc_ = self; 99 Locks::mutator_lock_->AssertNotHeld(self); 101 ReaderMutexLock mu(self, *Locks::mutator_lock_); 106 ReaderMutexLock mu(self, *Locks::mutator_lock_); 124 ReaderMutexLock mu(self, *Locks::mutator_lock_); 134 Thread* self local 199 Thread* self = Thread::Current(); local 201 << thread->GetState() << " thread " << thread << " self " << self; local 238 Thread* self = Thread::Current(); local 268 Thread* self = Thread::Current(); local 294 RecordLiveStackFreezeSize(Thread* self) argument 342 Thread* self = Thread::Current(); variable 344 << thread->GetState() << " thread " << thread << " self " << self; variable 347 concurrent_copying_->GetBarrier().Pass(self); variable 400 Thread* self = Thread::Current(); local 471 ReenableWeakRefAccess(Thread* self) argument 498 Thread* self = Thread::Current(); variable 500 << thread->GetState() << " thread " << thread << " self " << self; variable 507 concurrent_copying_->GetBarrier().Pass(self); variable 515 Thread* self = Thread::Current(); local 552 Thread* self = Thread::Current(); local 586 Thread* self = Thread::Current(); // TODO: pass self as an argument from call sites? local 773 Thread* self = Thread::Current(); local 900 Thread* self = Thread::Current(); variable 902 << thread->GetState() << " thread " << thread << " self " << self; variable 916 concurrent_copying_->GetBarrier().Pass(self); variable 925 Thread* self = Thread::Current(); local 944 Thread* self = Thread::Current(); local 971 Thread* self = Thread::Current(); local 1124 Thread* self = Thread::Current(); local 1144 Thread* self = Thread::Current(); local 1159 Thread* self = Thread::Current(); local 1192 SweepSystemWeaks(Thread* self) argument 1289 Thread* self = Thread::Current(); local 1739 Thread* self = Thread::Current(); local 2072 Thread* const self = Thread::Current(); local 2109 ProcessReferences(Thread* self) argument [all...] |
H A D | mark_sweep.cc | 144 Thread* self = Thread::Current(); local 146 Locks::mutator_lock_->AssertNotHeld(self); 150 ReaderMutexLock mu(self, *Locks::mutator_lock_); 167 ReaderMutexLock mu(self, *Locks::mutator_lock_); 174 void MarkSweep::ProcessReferences(Thread* self) { argument 175 WriterMutexLock mu(self, *Locks::heap_bitmap_lock_); 185 Thread* self = Thread::Current(); local 186 Locks::mutator_lock_->AssertExclusiveHeld(self); 189 WriterMutexLock mu(self, *Locks::heap_bitmap_lock_); 197 WriterMutexLock mu(self, *Lock 219 Thread* self = Thread::Current(); local 246 RevokeAllThreadLocalAllocationStacks(Thread* self) argument 256 Thread* self = Thread::Current(); local 316 Thread* const self = Thread::Current(); local 449 Thread* self = Thread::Current(); variable 605 MarkRoots(Thread* self) argument 862 Thread* self = Thread::Current(); local 1009 auto* self = Thread::Current(); local 1077 SweepSystemWeaks(Thread* self) argument 1142 Thread* const self = Thread::Current(); variable 1144 << thread->GetState() << " thread " << thread << " self " << self; variable 1152 mark_sweep_->GetBarrier().Pass(self); variable 1160 MarkRootsCheckpoint(Thread* self, bool revoke_ros_alloc_thread_local_buffers_at_checkpoint) argument 1186 Thread* self = Thread::Current(); local 1370 Thread* self = Thread::Current(); local 1468 Thread* const self = Thread::Current(); local [all...] |
/art/runtime/ |
H A D | monitor_pool.cc | 108 Monitor* MonitorPool::CreateMonitorInPool(Thread* self, Thread* owner, mirror::Object* obj, 112 MutexLock mu(self, *Locks::allocated_monitor_ids_lock_); 127 Monitor* monitor = new(mon_uninitialized) Monitor(self, owner, obj, hash_code, id); 132 void MonitorPool::ReleaseMonitorToPool(Thread* self, Monitor* monitor) { argument 134 MutexLock mu(self, *Locks::allocated_monitor_ids_lock_); 151 void MonitorPool::ReleaseMonitorsToPool(Thread* self, MonitorList::Monitors* monitors) { argument 153 ReleaseMonitorToPool(self, mon);
|
H A D | instrumentation.cc | 593 Thread* const self = Thread::Current(); local 595 Locks::mutator_lock_->AssertExclusiveHeld(self); 596 Locks::thread_list_lock_->AssertNotHeld(self); 609 MutexLock mu(self, *Locks::thread_list_lock_); 619 ReaderMutexLock mu(self, deoptimized_methods_lock_); 623 MutexLock mu(self, *Locks::thread_list_lock_); 637 Thread* self = Thread::Current(); local 639 Locks::mutator_lock_->AssertNotHeld(self); 640 Locks::instrument_entrypoints_lock_->AssertHeld(self); 643 MutexLock mu(self, *Lock 764 Thread* self = Thread::Current(); local 787 Thread* self = Thread::Current(); local 1012 ComputeFrameId(Thread* self, size_t frame_depth, size_t inlined_frames_before_frame) argument 1032 PushInstrumentationStackFrame(Thread* self, mirror::Object* this_object, ArtMethod* method, uintptr_t lr, bool interpreter_entry) argument 1050 PopInstrumentationStackFrame(Thread* self, uintptr_t* return_pc, uint64_t gpr_result, uint64_t fpr_result) argument 1113 PopMethodForUnwind(Thread* self, bool is_deoptimization) const argument [all...] |
H A D | runtime.cc | 238 Thread* self = Thread::Current(); local 239 const bool attach_shutdown_thread = self == nullptr; 242 self = Thread::Current(); 249 MutexLock mu(self, *Locks::runtime_shutdown_lock_); 252 shutdown_cond_->Wait(self); 257 CHECK(self != nullptr); 260 self->ClearException(); 261 self->GetJniEnv()->CallStaticVoidMethod(WellKnownClasses::java_lang_Daemons, 269 self = nullptr; 273 heap_->WaitForGcToComplete(gc::kGcCauseBackground, self); 341 Thread* self = Thread::Current(); local 555 Thread* self = Thread::Current(); local 750 IsShuttingDown(Thread* self) argument 764 Thread* self = Thread::Current(); local 1116 Thread* self = Thread::Attach("main", false, nullptr, false); local 1286 Thread* self = Thread::Current(); local 1332 InitThreadGroups(Thread* self) argument 1429 Thread* self = Thread::Current(); local 1503 Thread* self = Thread::Current(); local 1810 AbortTransactionAndThrowAbortError(Thread* self, const std::string& abort_message) argument 1820 ThrowTransactionAbortError(Thread* self) argument [all...] |
H A D | class_linker.h | 161 mirror::Class* FindClass(Thread* self, 173 Thread* self, 183 mirror::Class* FindSystemClass(Thread* self, const char* descriptor) 188 mirror::Class* FindArrayClass(Thread* self, mirror::Class** element_class) 198 mirror::Class* DefineClass(Thread* self, 209 mirror::Class* LookupClass(Thread* self, 320 ArtMethod* ResolveMethod(Thread* self, uint32_t method_idx, ArtMethod* referrer, InvokeType type) 367 bool EnsureInitialized(Thread* self, 411 mirror::DexCache* FindDexCache(Thread* self, 421 mirror::Object* AllocObject(Thread* self) [all...] |
H A D | intern_table.cc | 90 mirror::String* InternTable::LookupWeak(Thread* self, mirror::String* s) { argument 91 MutexLock mu(self, *Locks::intern_table_lock_); 95 mirror::String* InternTable::LookupStrong(Thread* self, mirror::String* s) { argument 96 MutexLock mu(self, *Locks::intern_table_lock_); 100 mirror::String* InternTable::LookupStrong(Thread* self, argument 107 MutexLock mu(self, *Locks::intern_table_lock_); 239 Thread* self = Thread::Current(); local 240 MutexLock mu(self, *Locks::intern_table_lock_); 241 weak_intern_condition_.Broadcast(self); 244 void InternTable::WaitUntilAccessible(Thread* self) { argument 260 Thread* const self = Thread::Current(); local [all...] |
H A D | jni_env_ext.h | 37 static JNIEnvExt* Create(Thread* self, JavaVMExt* vm); 60 Thread* const self; member in struct:art::JNIEnvExt 107 JNIEnvExt(Thread* self, JavaVMExt* vm);
|
H A D | object_lock.h | 31 ObjectLock(Thread* self, Handle<T> object) SHARED_REQUIRES(Locks::mutator_lock_);
|
H A D | reflection.h | 72 bool VerifyAccess(Thread* self, mirror::Object* obj, mirror::Class* declaring_class, 77 bool VerifyAccess(Thread* self, mirror::Object* obj, mirror::Class* declaring_class, 82 mirror::Class* GetCallingClass(Thread* self, size_t num_frames) 88 void UpdateReference(Thread* self, jobject obj, mirror::Object* result)
|
/art/runtime/entrypoints/ |
H A D | entrypoint_utils.cc | 43 Thread* self, 56 DCHECK(self->IsExceptionPending()); 65 self->ThrowNewExceptionF( 85 ArtMethod* referrer, Thread* self, 88 mirror::Class* klass = CheckFilledNewArrayAlloc(type_idx, component_count, referrer, self, 97 return mirror::Array::Alloc<false>(self, klass, component_count, 106 Thread* self, 109 mirror::Class* klass = CheckFilledNewArrayAlloc(type_idx, component_count, referrer, self, 117 return mirror::Array::Alloc<true>(self, klass, component_count, 122 void CheckReferenceResult(mirror::Object* o, Thread* self) { argument 84 CheckAndAllocArrayFromCode(uint32_t type_idx, int32_t component_count, ArtMethod* referrer, Thread* self, bool access_check, gc::AllocatorType ) argument 103 CheckAndAllocArrayFromCodeInstrumented(uint32_t type_idx, int32_t component_count, ArtMethod* referrer, Thread* self, bool access_check, gc::AllocatorType ) argument 246 Thread* self = Thread::Current(); local [all...] |
/art/runtime/gc/ |
H A D | allocation_record.cc | 215 Thread* self = Thread::Current(); local 219 MutexLock mu(self, *Locks::alloc_tracker_lock_); 231 self->GetThreadName(self_name); 233 records->alloc_ddm_thread_id_ = self->GetTid(); 243 MutexLock mu(self, *Locks::alloc_tracker_lock_); 250 MutexLock mu(self, *Locks::alloc_tracker_lock_); 264 void AllocRecordObjectMap::RecordAllocation(Thread* self, argument 270 AllocRecordStackVisitor visitor(self, max_stack_depth_, /*out*/ &trace); 272 StackHandleScope<1> hs(self); 277 MutexLock mu(self, *Lock [all...] |
/art/runtime/interpreter/mterp/out/ |
H A D | mterp_x86_64.S | 327 * 0 Thread* self 744 call SYMBOL(MterpConstString) # (index, tgt_reg, shadow_frame, self) 759 call SYMBOL(MterpConstString) # (index, tgt_reg, shadow_frame, self) 774 call SYMBOL(MterpConstClass) # (index, tgt_reg, shadow_frame, self) 790 call SYMBOL(artLockObjectFromCode) # (object, self) 810 call SYMBOL(artUnlockObjectFromCode) # (object, self) 828 call SYMBOL(MterpCheckCast) # (index, &obj, method, self) 851 call SYMBOL(MterpInstanceOf) # (index, &obj, method, self) 2452 movq rSELF, OUT_ARG2 # self 2483 movq rSELF, OUT_ARG2 # self [all...] |
/art/runtime/entrypoints/quick/ |
H A D | quick_trampoline_entrypoints.cc | 644 extern "C" uint64_t artQuickToInterpreterBridge(ArtMethod* method, Thread* self, ArtMethod** sp) 648 ScopedQuickEntrypointChecks sqec(self); 656 ShadowFrame* deopt_frame = self->PopStackedShadowFrame( 686 QuickExceptionHandler::DumpFramesWithType(self, true); 691 self->PopDeoptimizationContext(&result, &pending_exception, /* out */ &from_code); 695 self->PushManagedStackFragment(&fragment); 710 DummyStackVisitor dsv(self); 717 self->SetException(pending_exception); 719 interpreter::EnterInterpreterFromDeoptimize(self, deopt_frame, from_code, &result); 721 const char* old_cause = self 1769 BuildGenericJniFrameVisitor(Thread* self, bool is_static, const char* shorty, uint32_t shorty_len, ArtMethod*** sp) argument 1902 FinalizeHandleScope(Thread* self) argument 1915 artQuickGenericJniEndJNIRef(Thread* self, uint32_t cookie, jobject l, jobject lock) argument 1923 artQuickGenericJniEndJNINonRef(Thread* self, uint32_t cookie, jobject lock) argument 2020 artQuickGenericJniEndTrampoline(Thread* self, jvalue result, uint64_t result_f) argument 2042 artInvokeCommon(uint32_t method_idx, mirror::Object* this_object, Thread* self, ArtMethod** sp) argument [all...] |
/art/runtime/jit/ |
H A D | jit.cc | 243 bool Jit::CompileMethod(ArtMethod* method, Thread* self, bool osr) { argument 263 if (!code_cache_->NotifyCompilationOf(method_to_compile, self, osr)) { 270 bool success = jit_compile_method_(jit_compiler_handle_, method_to_compile, self, osr); 271 code_cache_->DoneCompiling(method_to_compile, self, osr); 289 Thread* self = Thread::Current(); local 290 DCHECK(Runtime::Current()->IsShuttingDown(self)); 299 cache->StopWorkers(self); 300 cache->RemoveAllTasks(self); 304 cache->Wait(self, false, false); 383 Thread* self); 590 AddSamples(Thread* self, ArtMethod* method, uint16_t count, bool with_backedges) argument 697 WaitForCompilationToFinish(Thread* self) argument [all...] |
/art/runtime/native/ |
H A D | org_apache_harmony_dalvik_ddmc_DdmVmInternal.cc | 50 Thread* const self = Thread::Current(); local 51 if (static_cast<uint32_t>(thin_lock_id) == self->GetThreadId()) { 54 jobject internal_trace = self->CreateInternalStackTrace<false>(soa); 132 Thread* self = static_cast<JNIEnvExt*>(env)->self; local 134 MutexLock mu(self, *Locks::thread_list_lock_);
|
/art/tools/checker/file_format/c1visualizer/ |
H A D | parser.py | 24 def __init__(self): 25 self.currentState = C1ParserState.OutsideBlock 26 self.lastMethodName = None
|
/art/runtime/arch/x86_64/ |
H A D | thread_x86_64.cc | 48 tlsPtr_.self = this;
|