/art/runtime/gc/accounting/ |
H A D | bitmap.h | 49 static ALWAYS_INLINE constexpr size_t BitIndexToWordIndex(uintptr_t offset) { 54 static ALWAYS_INLINE constexpr T WordIndexToBitIndex(T word_index) { 58 static ALWAYS_INLINE constexpr uintptr_t BitIndexToMask(uintptr_t bit_index) { 62 ALWAYS_INLINE bool SetBit(size_t bit_index) { 66 ALWAYS_INLINE bool ClearBit(size_t bit_index) { 70 ALWAYS_INLINE bool TestBit(size_t bit_index) const; 73 ALWAYS_INLINE bool AtomicTestAndSetBit(size_t bit_index); 96 ALWAYS_INLINE void CheckValidBitIndex(size_t bit_index) const { 112 ALWAYS_INLINE bool ModifyBit(uintptr_t bit_index); 137 ALWAYS_INLINE uintptr_ [all...] |
H A D | mod_union_table-inl.h | 36 bool ShouldAddReference(const mirror::Object* ref) const OVERRIDE ALWAYS_INLINE {
|
H A D | card_table.h | 59 ALWAYS_INLINE void MarkCard(const void *addr) { 126 void* AddrFromCard(const uint8_t *card_addr) const ALWAYS_INLINE; member in class:art::gc::accounting::CardTable 129 uint8_t* CardFromAddr(const void *addr) const ALWAYS_INLINE; member in class:art::gc::accounting::CardTable 137 bool IsValidCard(const uint8_t* card_addr) const ALWAYS_INLINE; member in class:art::gc::accounting::CardTable 139 void CheckCardValid(uint8_t* card) const ALWAYS_INLINE; member in class:art::gc::accounting::CardTable
|
H A D | heap_bitmap.h | 43 REQUIRES(Locks::heap_bitmap_lock_) ALWAYS_INLINE; member in class:art::gc::accounting::HeapBitmap 47 REQUIRES(Locks::heap_bitmap_lock_) ALWAYS_INLINE; member in class:art::gc::accounting::HeapBitmap
|
/art/runtime/ |
H A D | handle.h | 42 ALWAYS_INLINE Handle(const Handle<T>& handle) : reference_(handle.reference_) { 45 ALWAYS_INLINE Handle<T>& operator=(const Handle<T>& handle) { 50 ALWAYS_INLINE explicit Handle(StackReference<T>* reference) : reference_(reference) { 53 ALWAYS_INLINE T& operator*() const SHARED_REQUIRES(Locks::mutator_lock_) { 57 ALWAYS_INLINE T* operator->() const SHARED_REQUIRES(Locks::mutator_lock_) { 61 ALWAYS_INLINE T* Get() const SHARED_REQUIRES(Locks::mutator_lock_) { 65 ALWAYS_INLINE jobject ToJObject() const SHARED_REQUIRES(Locks::mutator_lock_) { 73 ALWAYS_INLINE StackReference<mirror::Object>* GetReference() { 77 ALWAYS_INLINE const StackReference<mirror::Object>* GetReference() const { 108 ALWAYS_INLINE MutableHandl [all...] |
H A D | verify_object.h | 55 ALWAYS_INLINE void VerifyObject(mirror::Object* obj) NO_THREAD_SAFETY_ANALYSIS; 58 ALWAYS_INLINE bool VerifyClassClass(mirror::Class* c) NO_THREAD_SAFETY_ANALYSIS;
|
H A D | scoped_thread_state_change.h | 37 REQUIRES(!Locks::thread_suspend_count_lock_) ALWAYS_INLINE 62 ~ScopedThreadStateChange() REQUIRES(!Locks::thread_suspend_count_lock_) ALWAYS_INLINE { 179 REQUIRES(!Locks::thread_suspend_count_lock_) ALWAYS_INLINE 184 REQUIRES(!Locks::thread_suspend_count_lock_) ALWAYS_INLINE 195 ~ScopedObjectAccessAlreadyRunnable() ALWAYS_INLINE { 223 REQUIRES(!Locks::thread_suspend_count_lock_) ALWAYS_INLINE 230 REQUIRES(!Locks::thread_suspend_count_lock_) ALWAYS_INLINE 238 explicit ScopedObjectAccessUnchecked(JavaVM* vm) ALWAYS_INLINE 254 SHARED_LOCK_FUNCTION(Locks::mutator_lock_) ALWAYS_INLINE 260 SHARED_LOCK_FUNCTION(Locks::mutator_lock_) ALWAYS_INLINE [all...] |
H A D | handle_scope.h | 62 ALWAYS_INLINE mirror::Object* GetReference(size_t i) const 65 ALWAYS_INLINE Handle<mirror::Object> GetHandle(size_t i); 67 ALWAYS_INLINE MutableHandle<mirror::Object> GetMutableHandle(size_t i) 70 ALWAYS_INLINE void SetReference(size_t i, mirror::Object* object) 73 ALWAYS_INLINE bool Contains(StackReference<mirror::Object>* handle_scope_entry) const; 98 ALWAYS_INLINE StackReference<mirror::Object>* GetReferences() const { 148 explicit ALWAYS_INLINE StackHandleScope(Thread* self, mirror::Object* fill_value = nullptr); 149 ALWAYS_INLINE ~StackHandleScope(); 152 ALWAYS_INLINE MutableHandle<T> NewHandle(T* object) SHARED_REQUIRES(Locks::mutator_lock_); 155 ALWAYS_INLINE HandleWrappe [all...] |
H A D | utf.h | 61 ALWAYS_INLINE int CompareModifiedUtf8ToModifiedUtf8AsUtf16CodePointValues(const char* utf8_1, 116 ALWAYS_INLINE uint16_t GetLeadingUtf16Char(uint32_t maybe_pair); 122 ALWAYS_INLINE uint16_t GetTrailingUtf16Char(uint32_t maybe_pair);
|
H A D | gc_root.h | 93 ALWAYS_INLINE void VisitRoot(mirror::Object** root, const RootInfo& info) 99 ALWAYS_INLINE void VisitRootIfNonNull(mirror::Object** root, const RootInfo& info) 171 ALWAYS_INLINE MirrorType* Read(GcRootSource* gc_root_source = nullptr) const 189 ALWAYS_INLINE mirror::CompressedReference<mirror::Object>* AddressWithoutBarrier() { 193 ALWAYS_INLINE bool IsNull() const { 198 ALWAYS_INLINE GcRoot() {} 199 explicit ALWAYS_INLINE GcRoot(MirrorType* ref) SHARED_REQUIRES(Locks::mutator_lock_); 225 ALWAYS_INLINE void VisitRootIfNonNull(GcRoot<MirrorType>& root) 233 ALWAYS_INLINE void VisitRootIfNonNull(mirror::CompressedReference<MirrorType>* root)
|
H A D | memory_region.h | 55 ALWAYS_INLINE T Load(uintptr_t offset) const { 65 ALWAYS_INLINE void Store(uintptr_t offset, T value) const { 74 ALWAYS_INLINE T LoadUnaligned(uintptr_t offset) const { 89 ALWAYS_INLINE void StoreUnaligned(uintptr_t offset, T value) const { 101 ALWAYS_INLINE T* PointerTo(uintptr_t offset) const { 107 ALWAYS_INLINE bool LoadBit(uintptr_t bit_offset) const { 113 ALWAYS_INLINE void StoreBit(uintptr_t bit_offset, bool value) const { 165 ALWAYS_INLINE T* ComputeInternalPointer(size_t offset) const { 173 ALWAYS_INLINE uint8_t* ComputeBitPointer(uintptr_t bit_offset, uint8_t* bit_mask) const {
|
H A D | read_barrier.h | 50 ALWAYS_INLINE static MirrorType* Barrier( 57 ALWAYS_INLINE static MirrorType* BarrierForRoot(MirrorType** root, 64 ALWAYS_INLINE static MirrorType* BarrierForRoot(mirror::CompressedReference<MirrorType>* root, 83 // ALWAYS_INLINE on this caused a performance regression b/26744236. 96 ALWAYS_INLINE static bool HasGrayReadBarrierPointer(mirror::Object* obj,
|
H A D | art_method.h | 233 ALWAYS_INLINE mirror::Class* GetDeclaringClass() SHARED_REQUIRES(Locks::mutator_lock_); 236 ALWAYS_INLINE mirror::Class* GetDeclaringClassUnchecked() 252 ALWAYS_INLINE uint32_t GetAccessFlags(); 419 ALWAYS_INLINE uint32_t GetDexMethodIndex() SHARED_REQUIRES(Locks::mutator_lock_); 426 ALWAYS_INLINE ArtMethod** GetDexCacheResolvedMethods(size_t pointer_size) 428 ALWAYS_INLINE ArtMethod* GetDexCacheResolvedMethod(uint16_t method_index, size_t ptr_size) 430 ALWAYS_INLINE void SetDexCacheResolvedMethod(uint16_t method_index, 434 ALWAYS_INLINE void SetDexCacheResolvedMethods(ArtMethod** new_dex_cache_methods, size_t ptr_size) 478 ALWAYS_INLINE const void* GetEntryPointFromQuickCompiledCodePtrSize(size_t pointer_size) { 487 ALWAYS_INLINE voi [all...] |
H A D | stack_map.h | 663 ALWAYS_INLINE size_t BitSize() const { return end_offset_ - start_offset_; } 665 ALWAYS_INLINE int32_t Load(const MemoryRegion& region) const { 695 ALWAYS_INLINE void Store(MemoryRegion region, int32_t value) const { 747 ALWAYS_INLINE FieldEncoding GetNativePcEncoding() const { 750 ALWAYS_INLINE FieldEncoding GetDexPcEncoding() const { 753 ALWAYS_INLINE FieldEncoding GetDexRegisterMapEncoding() const { 756 ALWAYS_INLINE FieldEncoding GetInlineInfoEncoding() const { 759 ALWAYS_INLINE FieldEncoding GetRegisterMaskEncoding() const { 762 ALWAYS_INLINE size_t GetStackMaskBitOffset() const { 796 ALWAYS_INLINE boo [all...] |
/art/runtime/native/ |
H A D | scoped_fast_native_object_access.h | 31 SHARED_LOCK_FUNCTION(Locks::mutator_lock_) ALWAYS_INLINE 39 ~ScopedFastNativeObjectAccess() UNLOCK_FUNCTION(Locks::mutator_lock_) ALWAYS_INLINE {
|
/art/compiler/debug/dwarf/ |
H A D | debug_frame_opcode_writer.h | 47 void ALWAYS_INLINE AdvancePC(int absolute_pc) { 73 void ALWAYS_INLINE RelOffset(Reg reg, int offset) { 78 void ALWAYS_INLINE AdjustCFAOffset(int delta) { 83 void ALWAYS_INLINE RelOffsetForMany(Reg reg_base, int offset, 99 void ALWAYS_INLINE RestoreMany(Reg reg_base, uint32_t reg_mask) { 111 void ALWAYS_INLINE Nop() { 117 void ALWAYS_INLINE Offset(Reg reg, int offset) { 139 void ALWAYS_INLINE Restore(Reg reg) { 151 void ALWAYS_INLINE Undefined(Reg reg) { 159 void ALWAYS_INLINE SameValu [all...] |
/art/runtime/mirror/ |
H A D | dex_cache.h | 112 String* GetResolvedString(uint32_t string_idx) ALWAYS_INLINE 115 void SetResolvedString(uint32_t string_idx, String* resolved) ALWAYS_INLINE 122 ALWAYS_INLINE ArtMethod* GetResolvedMethod(uint32_t method_idx, size_t ptr_size) 125 ALWAYS_INLINE void SetResolvedMethod(uint32_t method_idx, ArtMethod* resolved, size_t ptr_size) 129 ALWAYS_INLINE ArtField* GetResolvedField(uint32_t idx, size_t ptr_size) 133 ALWAYS_INLINE void SetResolvedField(uint32_t idx, ArtField* field, size_t ptr_size) 136 GcRoot<String>* GetStrings() ALWAYS_INLINE SHARED_REQUIRES(Locks::mutator_lock_) { 140 void SetStrings(GcRoot<String>* strings) ALWAYS_INLINE SHARED_REQUIRES(Locks::mutator_lock_) { 144 GcRoot<Class>* GetResolvedTypes() ALWAYS_INLINE SHARED_REQUIRES(Locks::mutator_lock_) { 149 ALWAYS_INLINE [all...] |
H A D | object_array.h | 42 ALWAYS_INLINE T* Get(int32_t i) SHARED_REQUIRES(Locks::mutator_lock_); 50 ALWAYS_INLINE void Set(int32_t i, T* object) SHARED_REQUIRES(Locks::mutator_lock_); 54 ALWAYS_INLINE void Set(int32_t i, T* object) NO_THREAD_SAFETY_ANALYSIS; 62 ALWAYS_INLINE void SetWithoutChecks(int32_t i, T* object) NO_THREAD_SAFETY_ANALYSIS; 67 ALWAYS_INLINE void SetWithoutChecksAndWriteBarrier(int32_t i, T* object) 70 ALWAYS_INLINE T* GetWithoutChecks(int32_t i) SHARED_REQUIRES(Locks::mutator_lock_);
|
H A D | array.h | 40 ALWAYS_INLINE static Array* Alloc(Thread* self, Class* array_class, int32_t component_count, 52 ALWAYS_INLINE int32_t GetLength() SHARED_REQUIRES(Locks::mutator_lock_) { 85 ALWAYS_INLINE bool CheckIsValidIndex(int32_t index) SHARED_REQUIRES(Locks::mutator_lock_); 114 const T* GetData() const ALWAYS_INLINE SHARED_REQUIRES(Locks::mutator_lock_) { 118 T* GetData() ALWAYS_INLINE SHARED_REQUIRES(Locks::mutator_lock_) { 122 T Get(int32_t i) ALWAYS_INLINE SHARED_REQUIRES(Locks::mutator_lock_); 124 T GetWithoutChecks(int32_t i) ALWAYS_INLINE SHARED_REQUIRES(Locks::mutator_lock_) { 129 void Set(int32_t i, T value) ALWAYS_INLINE SHARED_REQUIRES(Locks::mutator_lock_); 134 void Set(int32_t i, T value) ALWAYS_INLINE NO_THREAD_SAFETY_ANALYSIS; 141 void SetWithoutChecks(int32_t i, T value) ALWAYS_INLINE NO_THREAD_SAFETY_ANALYSI [all...] |
H A D | object.h | 90 ALWAYS_INLINE Class* GetClass() SHARED_REQUIRES(Locks::mutator_lock_); 103 ALWAYS_INLINE bool AtomicSetReadBarrierPointer(Object* expected_rb_ptr, Object* rb_ptr) 112 ALWAYS_INLINE bool InstanceOf(Class* klass) SHARED_REQUIRES(Locks::mutator_lock_); 266 ALWAYS_INLINE T* GetFieldObject(MemberOffset field_offset) 271 ALWAYS_INLINE T* GetFieldObjectVolatile(MemberOffset field_offset) 276 ALWAYS_INLINE void SetFieldObjectWithoutWriteBarrier(MemberOffset field_offset, Object* new_value) 281 ALWAYS_INLINE void SetFieldObject(MemberOffset field_offset, Object* new_value) 286 ALWAYS_INLINE void SetFieldObjectVolatile(MemberOffset field_offset, Object* new_value) 328 ALWAYS_INLINE uint8_t GetFieldBoolean(MemberOffset field_offset) 332 ALWAYS_INLINE int8_ [all...] |
H A D | class.h | 195 ALWAYS_INLINE uint32_t GetAccessFlags() SHARED_REQUIRES(Locks::mutator_lock_); 201 ALWAYS_INLINE uint32_t GetClassFlags() SHARED_REQUIRES(Locks::mutator_lock_) { 209 ALWAYS_INLINE bool IsInterface() SHARED_REQUIRES(Locks::mutator_lock_) { 214 ALWAYS_INLINE bool IsPublic() SHARED_REQUIRES(Locks::mutator_lock_) { 219 ALWAYS_INLINE bool IsFinal() SHARED_REQUIRES(Locks::mutator_lock_) { 223 ALWAYS_INLINE bool IsFinalizable() SHARED_REQUIRES(Locks::mutator_lock_) { 227 ALWAYS_INLINE void SetRecursivelyInitialized() SHARED_REQUIRES(Locks::mutator_lock_) { 233 ALWAYS_INLINE void SetHasDefaultMethods() SHARED_REQUIRES(Locks::mutator_lock_) { 239 ALWAYS_INLINE void SetFinalizable() SHARED_REQUIRES(Locks::mutator_lock_) { 244 ALWAYS_INLINE boo 890 SHARED_REQUIRES(Locks::mutator_lock_) ALWAYS_INLINE; member in class:art::mirror::FINAL [all...] |
/art/runtime/base/ |
H A D | arena_object.h | 45 ALWAYS_INLINE void* operator new(size_t, void* ptr) noexcept { return ptr; } 46 ALWAYS_INLINE void operator delete(void*, void*) noexcept { }
|
/art/runtime/entrypoints/ |
H A D | entrypoint_utils.h | 47 ALWAYS_INLINE inline mirror::Class* CheckObjectAlloc(uint32_t type_idx, 52 ALWAYS_INLINE inline mirror::Class* CheckClassInitializedForObjectAlloc(mirror::Class* klass, 62 ALWAYS_INLINE inline mirror::Object* AllocObjectFromCode(uint32_t type_idx, 70 ALWAYS_INLINE inline mirror::Object* AllocObjectFromCodeResolved(mirror::Class* klass, 77 ALWAYS_INLINE inline mirror::Object* AllocObjectFromCodeInitialized(mirror::Class* klass, 84 ALWAYS_INLINE inline mirror::Class* CheckArrayAlloc(uint32_t type_idx, 95 ALWAYS_INLINE inline mirror::Array* AllocArrayFromCode(uint32_t type_idx, 103 ALWAYS_INLINE inline mirror::Array* AllocArrayFromCodeResolved(mirror::Class* klass,
|
/art/runtime/gc/collector/ |
H A D | immune_spaces.h | 52 ALWAYS_INLINE bool IsInImmuneRegion(const mirror::Object* obj) const { 70 ALWAYS_INLINE bool ContainsObject(const mirror::Object* obj) const {
|
H A D | immune_region.h | 46 ALWAYS_INLINE bool ContainsObject(const mirror::Object* obj) const {
|