/art/runtime/ |
H A D | barrier.h | 42 void Pass(Thread* self) REQUIRES(!lock_); 45 void Wait(Thread* self) REQUIRES(!lock_); 54 void Increment(Thread* self, int delta) REQUIRES(!lock_); 58 bool Increment(Thread* self, int delta, uint32_t timeout_ms) REQUIRES(!lock_); 62 void Init(Thread* self, int count) REQUIRES(!lock_); 65 void SetCountLocked(Thread* self, int count) REQUIRES(lock_); 68 int count_ GUARDED_BY(lock_); 70 Mutex lock_ ACQUIRED_AFTER(Locks::abort_lock_); 71 ConditionVariable condition_ GUARDED_BY(lock_);
|
H A D | linear_alloc.h | 31 void* Alloc(Thread* self, size_t size) REQUIRES(!lock_); 34 void* Realloc(Thread* self, void* ptr, size_t old_size, size_t new_size) REQUIRES(!lock_); 38 T* AllocArray(Thread* self, size_t elements) REQUIRES(!lock_) { 43 size_t GetUsedMemory() const REQUIRES(!lock_); 45 ArenaPool* GetArenaPool() REQUIRES(!lock_); 48 bool Contains(void* ptr) const REQUIRES(!lock_); 55 mutable Mutex lock_ DEFAULT_MUTEX_ACQUIRED_AFTER; 56 ArenaAllocator allocator_ GUARDED_BY(lock_);
|
H A D | signal_catcher.h | 48 void SetHaltFlag(bool new_value) REQUIRES(!lock_); 49 bool ShouldHalt() REQUIRES(!lock_); 50 int WaitForSignal(Thread* self, SignalSet& signals) REQUIRES(!lock_); 54 mutable Mutex lock_ DEFAULT_MUTEX_ACQUIRED_AFTER; 55 ConditionVariable cond_ GUARDED_BY(lock_); 56 bool halt_ GUARDED_BY(lock_); 57 pthread_t pthread_ GUARDED_BY(lock_); 58 Thread* thread_ GUARDED_BY(lock_);
|
H A D | linear_alloc.cc | 23 LinearAlloc::LinearAlloc(ArenaPool* pool) : lock_("linear alloc"), allocator_(pool) { 27 MutexLock mu(self, lock_); 32 MutexLock mu(self, lock_); 37 MutexLock mu(Thread::Current(), lock_); local 42 MutexLock mu(Thread::Current(), lock_); local 47 MutexLock mu(Thread::Current(), lock_); local
|
H A D | class_table.h | 74 REQUIRES(!lock_) 80 REQUIRES(!lock_) 84 size_t NumZygoteClasses() const REQUIRES(!lock_); 87 size_t NumNonZygoteClasses() const REQUIRES(!lock_); 91 REQUIRES(!lock_) 98 REQUIRES(!lock_) 104 REQUIRES(!lock_) 110 REQUIRES(!lock_) 115 REQUIRES(!lock_) 120 REQUIRES(!lock_) 167 mutable ReaderWriterMutex lock_; member in class:art::ClassTable [all...] |
H A D | barrier.cc | 28 lock_("GC barrier lock", kThreadSuspendCountLock), 29 condition_("GC barrier condition", lock_) { 33 MutexLock mu(self, lock_); 42 MutexLock mu(self, lock_); 47 MutexLock mu(self, lock_); 63 MutexLock mu(self, lock_);
|
H A D | class_table.cc | 23 ClassTable::ClassTable() : lock_("Class loader classes", kClassLoaderClassesLock) { 30 WriterMutexLock mu(Thread::Current(), lock_); local 35 ReaderMutexLock mu(Thread::Current(), lock_); local 46 ReaderMutexLock mu(Thread::Current(), lock_); local 57 WriterMutexLock mu(Thread::Current(), lock_); local 81 ReaderMutexLock mu(Thread::Current(), lock_); local 90 ReaderMutexLock mu(Thread::Current(), lock_); local 95 ReaderMutexLock mu(Thread::Current(), lock_); local 106 WriterMutexLock mu(Thread::Current(), lock_); local 111 WriterMutexLock mu(Thread::Current(), lock_); local 116 WriterMutexLock mu(Thread::Current(), lock_); local 150 WriterMutexLock mu(Thread::Current(), lock_); local 162 ReaderMutexLock mu(Thread::Current(), lock_); local 188 WriterMutexLock mu(Thread::Current(), lock_); local 193 WriterMutexLock mu(Thread::Current(), lock_); local [all...] |
H A D | class_table-inl.h | 26 ReaderMutexLock mu(Thread::Current(), lock_); local 39 ReaderMutexLock mu(Thread::Current(), lock_); local 52 ReaderMutexLock mu(Thread::Current(), lock_); local
|
H A D | signal_catcher.cc | 69 lock_("SignalCatcher lock"), 70 cond_("SignalCatcher::cond_", lock_), 78 MutexLock mu(self, lock_); 93 MutexLock mu(Thread::Current(), lock_); local 98 MutexLock mu(Thread::Current(), lock_); local 192 MutexLock mu(self, signal_catcher->lock_);
|
/art/runtime/gc/ |
H A D | task_processor.h | 57 void AddTask(Thread* self, HeapTask* task) REQUIRES(!*lock_); 58 HeapTask* GetTask(Thread* self) REQUIRES(!*lock_); 59 void Start(Thread* self) REQUIRES(!*lock_); 62 void Stop(Thread* self) REQUIRES(!*lock_); 63 void RunAllTasks(Thread* self) REQUIRES(!*lock_); 64 bool IsRunning() const REQUIRES(!*lock_); 66 REQUIRES(!*lock_); 67 Thread* GetRunningThread() const REQUIRES(!*lock_); 77 mutable Mutex* lock_ DEFAULT_MUTEX_ACQUIRED_AFTER; 78 bool is_running_ GUARDED_BY(lock_); [all...] |
H A D | task_processor.cc | 26 : lock_(new Mutex("Task processor lock", kReferenceProcessorLock)), is_running_(false), 29 cond_.reset(new ConditionVariable("Task processor condition", *lock_)); 33 delete lock_; 38 MutexLock mu(self, *lock_); 45 MutexLock mu(self, *lock_); 75 MutexLock mu(self, *lock_); 97 MutexLock mu(Thread::Current(), *lock_); local 102 MutexLock mu(Thread::Current(), *lock_); local 107 MutexLock mu(self, *lock_); 114 MutexLock mu(self, *lock_); [all...] |
H A D | reference_queue.h | 58 SHARED_REQUIRES(Locks::mutator_lock_) REQUIRES(!*lock_); 105 Mutex* const lock_; member in class:art::gc::ReferenceQueue
|
/art/runtime/jit/ |
H A D | jit_code_cache.h | 67 size_t CodeCacheSize() REQUIRES(!lock_); 70 size_t DataCacheSize() REQUIRES(!lock_); 74 REQUIRES(!lock_); 82 REQUIRES(!lock_); 86 REQUIRES(!lock_); 90 REQUIRES(!lock_); 103 REQUIRES(!lock_); 109 bool ContainsMethod(ArtMethod* method) REQUIRES(!lock_); 114 REQUIRES(!lock_); 119 REQUIRES(!lock_); 178 MutexLock lock(Thread::Current(), lock_); variable 267 Mutex lock_; member in class:art::jit::JitCodeCache [all...] |
H A D | jit_code_cache.cc | 123 : lock_("Jit code cache", kJitCodeCacheLock), 124 lock_cond_("Jit code cache variable", lock_), 169 MutexLock mu(Thread::Current(), lock_); local 263 MutexLock mu(self, lock_); 299 MutexLock mu(self, lock_); 326 MutexLock mu(self, lock_); 352 MutexLock mu(self, lock_); 387 MutexLock mu(Thread::Current(), lock_); local 396 MutexLock mu(Thread::Current(), lock_); local 405 MutexLock mu(self, lock_); 789 MutexLock mu(Thread::Current(), lock_); local 809 MutexLock mu(Thread::Current(), lock_); local 893 MutexLock mu(Thread::Current(), lock_); local 908 MutexLock mu(Thread::Current(), lock_); local 963 MutexLock mu(Thread::Current(), lock_); local 984 MutexLock mu(Thread::Current(), lock_); local 991 MutexLock mu(Thread::Current(), lock_); local 1023 MutexLock mu(Thread::Current(), lock_); local [all...] |
/art/compiler/dex/quick/ |
H A D | dex_file_to_method_inliner_map.h | 48 ReaderWriterMutex lock_; member in class:art::DexFileToMethodInlinerMap 49 std::map<const DexFile*, DexFileMethodInliner*> inliners_ GUARDED_BY(lock_);
|
H A D | dex_file_to_method_inliner_map.cc | 31 : lock_("DexFileToMethodInlinerMap lock", kDexFileToMethodInlinerMapLock) { 43 ReaderMutexLock mu(self, lock_); 50 // We need to acquire our lock_ to modify inliners_ but we want to release it 52 // new inliner's lock_ before we release our lock_ to prevent another thread 58 WriterMutexLock mu(self, lock_); 66 locked_inliner->lock_.ExclusiveLock(self); // Acquire inliner's lock_ before releasing lock_. 69 locked_inliner->lock_ [all...] |
H A D | dex_file_method_inliner.h | 59 SHARED_REQUIRES(Locks::mutator_lock_) REQUIRES(!lock_); 64 InlineMethodFlags IsIntrinsicOrSpecial(uint32_t method_index) REQUIRES(!lock_); 69 bool IsIntrinsic(uint32_t method_index, InlineMethod* intrinsic) REQUIRES(!lock_); 74 bool IsSpecial(uint32_t method_index) REQUIRES(!lock_); 80 REQUIRES(!lock_); 85 bool IsStringInitMethodIndex(uint32_t method_index) REQUIRES(!lock_); 380 void FindIntrinsics(const DexFile* dex_file) REQUIRES(lock_); 384 bool AddInlineMethod(int32_t method_idx, const InlineMethod& method) REQUIRES(!lock_); 386 ReaderWriterMutex lock_; member in class:art::DexFileMethodInliner 390 SafeMap<uint32_t, InlineMethod> inline_methods_ GUARDED_BY(lock_); [all...] |
/art/runtime/jdwp/ |
H A D | object_registry.h | 67 REQUIRES(!Locks::thread_list_lock_, !Locks::thread_suspend_count_lock_, !lock_); 71 REQUIRES(!Locks::thread_list_lock_, !Locks::thread_suspend_count_lock_, !lock_); 76 REQUIRES(!Locks::thread_list_lock_, !Locks::thread_suspend_count_lock_, !lock_); 80 REQUIRES(!Locks::thread_list_lock_, !Locks::thread_suspend_count_lock_, !lock_); 83 SHARED_REQUIRES(Locks::mutator_lock_) REQUIRES(!lock_) { 91 void Clear() SHARED_REQUIRES(Locks::mutator_lock_) REQUIRES(!lock_); 94 SHARED_REQUIRES(Locks::mutator_lock_) REQUIRES(!lock_); 97 SHARED_REQUIRES(Locks::mutator_lock_) REQUIRES(!lock_); 100 SHARED_REQUIRES(Locks::mutator_lock_) REQUIRES(!lock_); 103 SHARED_REQUIRES(Locks::mutator_lock_) REQUIRES(!lock_); [all...] |
H A D | object_registry.cc | 35 : lock_("ObjectRegistry lock", kJdwpObjectRegistryLock), next_id_(1) { 86 // Call IdentityHashCode here to avoid a lock level violation between lock_ and monitor_lock. 90 MutexLock mu(soa.Self(), lock_); 152 MutexLock mu(self, lock_); 172 MutexLock mu(self, lock_); 188 MutexLock mu(self, lock_); 197 MutexLock mu(self, lock_); 205 MutexLock mu(self, lock_); 235 MutexLock mu(self, lock_); 249 MutexLock mu(self, lock_); [all...] |
/art/runtime/gc/space/ |
H A D | large_object_space.h | 133 size_t AllocationSize(mirror::Object* obj, size_t* usable_size) REQUIRES(!lock_); 136 REQUIRES(!lock_); 137 size_t Free(Thread* self, mirror::Object* ptr) REQUIRES(!lock_); 138 void Walk(DlMallocSpace::WalkCallback, void* arg) OVERRIDE REQUIRES(!lock_); 139 // TODO: disabling thread safety analysis as this may be called when we already hold lock_. 150 bool IsZygoteLargeObject(Thread* self, mirror::Object* obj) const OVERRIDE REQUIRES(!lock_); 151 void SetAllLargeObjectsAsZygoteObjects(Thread* self) OVERRIDE REQUIRES(!lock_); 154 mutable Mutex lock_ DEFAULT_MUTEX_ACQUIRED_AFTER; 156 GUARDED_BY(lock_); 167 REQUIRES(lock_); [all...] |
H A D | dlmalloc_space.h | 53 OVERRIDE REQUIRES(!lock_); 57 OVERRIDE REQUIRES(!lock_) { 67 REQUIRES(!lock_) 71 REQUIRES(!lock_) 89 REQUIRES(!lock_); 107 void Walk(WalkCallback callback, void* arg) OVERRIDE REQUIRES(!lock_); 150 REQUIRES(lock_);
|
/art/runtime/base/ |
H A D | timing_logger.h | 36 void End() REQUIRES(!lock_); 37 void Reset() REQUIRES(!lock_); 38 void Dump(std::ostream& os) const REQUIRES(!lock_); 44 void SetName(const std::string& name) REQUIRES(!lock_); 45 void AddLogger(const TimingLogger& logger) REQUIRES(!lock_); 46 size_t GetIterations() const REQUIRES(!lock_); 61 REQUIRES(lock_); 62 void DumpHistogram(std::ostream &os) const REQUIRES(lock_); 67 std::set<Histogram<uint64_t>*, HistogramComparator> histograms_ GUARDED_BY(lock_); 70 mutable Mutex lock_ DEFAULT_MUTEX_ACQUIRED_AFTE [all...] |
H A D | timing_logger.cc | 41 lock_(lock_name_.c_str(), kDefaultMutexLevel, true) { 50 MutexLock mu(Thread::Current(), lock_); local 58 MutexLock mu(Thread::Current(), lock_); local 63 MutexLock mu(Thread::Current(), lock_); local 70 MutexLock mu(Thread::Current(), lock_); local 82 MutexLock mu(Thread::Current(), lock_); local 87 MutexLock mu(Thread::Current(), lock_); local
|
/art/runtime/gc/allocator/ |
H A D | rosalloc.h | 54 size_t ByteSize(RosAlloc* rosalloc) const REQUIRES(rosalloc->lock_) { 63 REQUIRES(rosalloc->lock_) { 72 void* End(RosAlloc* rosalloc) REQUIRES(rosalloc->lock_) { 78 REQUIRES(rosalloc->lock_) { 82 REQUIRES(rosalloc->lock_) { 85 bool ShouldReleasePages(RosAlloc* rosalloc) REQUIRES(rosalloc->lock_) { 102 void ReleasePages(RosAlloc* rosalloc) REQUIRES(rosalloc->lock_) { 718 AllocationTrackingSet<FreePageRun*, kAllocatorTagRosAlloc> free_page_runs_ GUARDED_BY(lock_); 742 volatile uint8_t* page_map_; // No GUARDED_BY(lock_) for kReadPageMapEntryWithoutLockInBulkFree. 751 GUARDED_BY(lock_); [all...] |
/art/compiler/utils/ |
H A D | swap_space.h | 38 void* Alloc(size_t size) REQUIRES(!lock_); 39 void Free(void* ptr, size_t size) REQUIRES(!lock_); 81 SpaceChunk NewFileChunk(size_t min_size) REQUIRES(lock_); 83 void RemoveChunk(FreeBySizeSet::const_iterator free_by_size_pos) REQUIRES(lock_); 84 void InsertChunk(const SpaceChunk& chunk) REQUIRES(lock_); 92 FreeByStartSet free_by_start_ GUARDED_BY(lock_); 94 FreeBySizeSet free_by_size_ GUARDED_BY(lock_); 96 mutable Mutex lock_ DEFAULT_MUTEX_ACQUIRED_AFTER;
|