Searched refs:ALWAYS_INLINE (Results 1 - 25 of 69) sorted by relevance

123

/art/runtime/gc/accounting/
H A Dbitmap.h49 static ALWAYS_INLINE constexpr size_t BitIndexToWordIndex(uintptr_t offset) {
54 static ALWAYS_INLINE constexpr T WordIndexToBitIndex(T word_index) {
58 static ALWAYS_INLINE constexpr uintptr_t BitIndexToMask(uintptr_t bit_index) {
62 ALWAYS_INLINE bool SetBit(size_t bit_index) {
66 ALWAYS_INLINE bool ClearBit(size_t bit_index) {
70 ALWAYS_INLINE bool TestBit(size_t bit_index) const;
73 ALWAYS_INLINE bool AtomicTestAndSetBit(size_t bit_index);
96 ALWAYS_INLINE void CheckValidBitIndex(size_t bit_index) const {
112 ALWAYS_INLINE bool ModifyBit(uintptr_t bit_index);
137 ALWAYS_INLINE uintptr_
[all...]
H A Dmod_union_table-inl.h36 bool ShouldAddReference(const mirror::Object* ref) const OVERRIDE ALWAYS_INLINE {
H A Dcard_table.h59 ALWAYS_INLINE void MarkCard(const void *addr) {
126 void* AddrFromCard(const uint8_t *card_addr) const ALWAYS_INLINE; member in class:art::gc::accounting::CardTable
129 uint8_t* CardFromAddr(const void *addr) const ALWAYS_INLINE; member in class:art::gc::accounting::CardTable
137 bool IsValidCard(const uint8_t* card_addr) const ALWAYS_INLINE; member in class:art::gc::accounting::CardTable
139 void CheckCardValid(uint8_t* card) const ALWAYS_INLINE; member in class:art::gc::accounting::CardTable
H A Dheap_bitmap.h43 REQUIRES(Locks::heap_bitmap_lock_) ALWAYS_INLINE; member in class:art::gc::accounting::HeapBitmap
47 REQUIRES(Locks::heap_bitmap_lock_) ALWAYS_INLINE; member in class:art::gc::accounting::HeapBitmap
/art/runtime/
H A Dhandle.h42 ALWAYS_INLINE Handle(const Handle<T>& handle) : reference_(handle.reference_) {
45 ALWAYS_INLINE Handle<T>& operator=(const Handle<T>& handle) {
50 ALWAYS_INLINE explicit Handle(StackReference<T>* reference) : reference_(reference) {
53 ALWAYS_INLINE T& operator*() const SHARED_REQUIRES(Locks::mutator_lock_) {
57 ALWAYS_INLINE T* operator->() const SHARED_REQUIRES(Locks::mutator_lock_) {
61 ALWAYS_INLINE T* Get() const SHARED_REQUIRES(Locks::mutator_lock_) {
65 ALWAYS_INLINE jobject ToJObject() const SHARED_REQUIRES(Locks::mutator_lock_) {
73 ALWAYS_INLINE StackReference<mirror::Object>* GetReference() {
77 ALWAYS_INLINE const StackReference<mirror::Object>* GetReference() const {
108 ALWAYS_INLINE MutableHandl
[all...]
H A Dverify_object.h55 ALWAYS_INLINE void VerifyObject(mirror::Object* obj) NO_THREAD_SAFETY_ANALYSIS;
58 ALWAYS_INLINE bool VerifyClassClass(mirror::Class* c) NO_THREAD_SAFETY_ANALYSIS;
H A Dscoped_thread_state_change.h37 REQUIRES(!Locks::thread_suspend_count_lock_) ALWAYS_INLINE
62 ~ScopedThreadStateChange() REQUIRES(!Locks::thread_suspend_count_lock_) ALWAYS_INLINE {
179 REQUIRES(!Locks::thread_suspend_count_lock_) ALWAYS_INLINE
184 REQUIRES(!Locks::thread_suspend_count_lock_) ALWAYS_INLINE
195 ~ScopedObjectAccessAlreadyRunnable() ALWAYS_INLINE {
223 REQUIRES(!Locks::thread_suspend_count_lock_) ALWAYS_INLINE
230 REQUIRES(!Locks::thread_suspend_count_lock_) ALWAYS_INLINE
238 explicit ScopedObjectAccessUnchecked(JavaVM* vm) ALWAYS_INLINE
254 SHARED_LOCK_FUNCTION(Locks::mutator_lock_) ALWAYS_INLINE
260 SHARED_LOCK_FUNCTION(Locks::mutator_lock_) ALWAYS_INLINE
[all...]
H A Dhandle_scope.h62 ALWAYS_INLINE mirror::Object* GetReference(size_t i) const
65 ALWAYS_INLINE Handle<mirror::Object> GetHandle(size_t i);
67 ALWAYS_INLINE MutableHandle<mirror::Object> GetMutableHandle(size_t i)
70 ALWAYS_INLINE void SetReference(size_t i, mirror::Object* object)
73 ALWAYS_INLINE bool Contains(StackReference<mirror::Object>* handle_scope_entry) const;
98 ALWAYS_INLINE StackReference<mirror::Object>* GetReferences() const {
148 explicit ALWAYS_INLINE StackHandleScope(Thread* self, mirror::Object* fill_value = nullptr);
149 ALWAYS_INLINE ~StackHandleScope();
152 ALWAYS_INLINE MutableHandle<T> NewHandle(T* object) SHARED_REQUIRES(Locks::mutator_lock_);
155 ALWAYS_INLINE HandleWrappe
[all...]
H A Dutf.h61 ALWAYS_INLINE int CompareModifiedUtf8ToModifiedUtf8AsUtf16CodePointValues(const char* utf8_1,
116 ALWAYS_INLINE uint16_t GetLeadingUtf16Char(uint32_t maybe_pair);
122 ALWAYS_INLINE uint16_t GetTrailingUtf16Char(uint32_t maybe_pair);
H A Dgc_root.h93 ALWAYS_INLINE void VisitRoot(mirror::Object** root, const RootInfo& info)
99 ALWAYS_INLINE void VisitRootIfNonNull(mirror::Object** root, const RootInfo& info)
171 ALWAYS_INLINE MirrorType* Read(GcRootSource* gc_root_source = nullptr) const
189 ALWAYS_INLINE mirror::CompressedReference<mirror::Object>* AddressWithoutBarrier() {
193 ALWAYS_INLINE bool IsNull() const {
198 ALWAYS_INLINE GcRoot() {}
199 explicit ALWAYS_INLINE GcRoot(MirrorType* ref) SHARED_REQUIRES(Locks::mutator_lock_);
225 ALWAYS_INLINE void VisitRootIfNonNull(GcRoot<MirrorType>& root)
233 ALWAYS_INLINE void VisitRootIfNonNull(mirror::CompressedReference<MirrorType>* root)
H A Dmemory_region.h55 ALWAYS_INLINE T Load(uintptr_t offset) const {
65 ALWAYS_INLINE void Store(uintptr_t offset, T value) const {
74 ALWAYS_INLINE T LoadUnaligned(uintptr_t offset) const {
89 ALWAYS_INLINE void StoreUnaligned(uintptr_t offset, T value) const {
101 ALWAYS_INLINE T* PointerTo(uintptr_t offset) const {
107 ALWAYS_INLINE bool LoadBit(uintptr_t bit_offset) const {
113 ALWAYS_INLINE void StoreBit(uintptr_t bit_offset, bool value) const {
165 ALWAYS_INLINE T* ComputeInternalPointer(size_t offset) const {
173 ALWAYS_INLINE uint8_t* ComputeBitPointer(uintptr_t bit_offset, uint8_t* bit_mask) const {
H A Dread_barrier.h50 ALWAYS_INLINE static MirrorType* Barrier(
57 ALWAYS_INLINE static MirrorType* BarrierForRoot(MirrorType** root,
64 ALWAYS_INLINE static MirrorType* BarrierForRoot(mirror::CompressedReference<MirrorType>* root,
83 // ALWAYS_INLINE on this caused a performance regression b/26744236.
96 ALWAYS_INLINE static bool HasGrayReadBarrierPointer(mirror::Object* obj,
H A Dart_method.h233 ALWAYS_INLINE mirror::Class* GetDeclaringClass() SHARED_REQUIRES(Locks::mutator_lock_);
236 ALWAYS_INLINE mirror::Class* GetDeclaringClassUnchecked()
252 ALWAYS_INLINE uint32_t GetAccessFlags();
419 ALWAYS_INLINE uint32_t GetDexMethodIndex() SHARED_REQUIRES(Locks::mutator_lock_);
426 ALWAYS_INLINE ArtMethod** GetDexCacheResolvedMethods(size_t pointer_size)
428 ALWAYS_INLINE ArtMethod* GetDexCacheResolvedMethod(uint16_t method_index, size_t ptr_size)
430 ALWAYS_INLINE void SetDexCacheResolvedMethod(uint16_t method_index,
434 ALWAYS_INLINE void SetDexCacheResolvedMethods(ArtMethod** new_dex_cache_methods, size_t ptr_size)
478 ALWAYS_INLINE const void* GetEntryPointFromQuickCompiledCodePtrSize(size_t pointer_size) {
487 ALWAYS_INLINE voi
[all...]
H A Dstack_map.h663 ALWAYS_INLINE size_t BitSize() const { return end_offset_ - start_offset_; }
665 ALWAYS_INLINE int32_t Load(const MemoryRegion& region) const {
695 ALWAYS_INLINE void Store(MemoryRegion region, int32_t value) const {
747 ALWAYS_INLINE FieldEncoding GetNativePcEncoding() const {
750 ALWAYS_INLINE FieldEncoding GetDexPcEncoding() const {
753 ALWAYS_INLINE FieldEncoding GetDexRegisterMapEncoding() const {
756 ALWAYS_INLINE FieldEncoding GetInlineInfoEncoding() const {
759 ALWAYS_INLINE FieldEncoding GetRegisterMaskEncoding() const {
762 ALWAYS_INLINE size_t GetStackMaskBitOffset() const {
796 ALWAYS_INLINE boo
[all...]
/art/runtime/native/
H A Dscoped_fast_native_object_access.h31 SHARED_LOCK_FUNCTION(Locks::mutator_lock_) ALWAYS_INLINE
39 ~ScopedFastNativeObjectAccess() UNLOCK_FUNCTION(Locks::mutator_lock_) ALWAYS_INLINE {
/art/compiler/debug/dwarf/
H A Ddebug_frame_opcode_writer.h47 void ALWAYS_INLINE AdvancePC(int absolute_pc) {
73 void ALWAYS_INLINE RelOffset(Reg reg, int offset) {
78 void ALWAYS_INLINE AdjustCFAOffset(int delta) {
83 void ALWAYS_INLINE RelOffsetForMany(Reg reg_base, int offset,
99 void ALWAYS_INLINE RestoreMany(Reg reg_base, uint32_t reg_mask) {
111 void ALWAYS_INLINE Nop() {
117 void ALWAYS_INLINE Offset(Reg reg, int offset) {
139 void ALWAYS_INLINE Restore(Reg reg) {
151 void ALWAYS_INLINE Undefined(Reg reg) {
159 void ALWAYS_INLINE SameValu
[all...]
/art/runtime/mirror/
H A Ddex_cache.h112 String* GetResolvedString(uint32_t string_idx) ALWAYS_INLINE
115 void SetResolvedString(uint32_t string_idx, String* resolved) ALWAYS_INLINE
122 ALWAYS_INLINE ArtMethod* GetResolvedMethod(uint32_t method_idx, size_t ptr_size)
125 ALWAYS_INLINE void SetResolvedMethod(uint32_t method_idx, ArtMethod* resolved, size_t ptr_size)
129 ALWAYS_INLINE ArtField* GetResolvedField(uint32_t idx, size_t ptr_size)
133 ALWAYS_INLINE void SetResolvedField(uint32_t idx, ArtField* field, size_t ptr_size)
136 GcRoot<String>* GetStrings() ALWAYS_INLINE SHARED_REQUIRES(Locks::mutator_lock_) {
140 void SetStrings(GcRoot<String>* strings) ALWAYS_INLINE SHARED_REQUIRES(Locks::mutator_lock_) {
144 GcRoot<Class>* GetResolvedTypes() ALWAYS_INLINE SHARED_REQUIRES(Locks::mutator_lock_) {
149 ALWAYS_INLINE
[all...]
H A Dobject_array.h42 ALWAYS_INLINE T* Get(int32_t i) SHARED_REQUIRES(Locks::mutator_lock_);
50 ALWAYS_INLINE void Set(int32_t i, T* object) SHARED_REQUIRES(Locks::mutator_lock_);
54 ALWAYS_INLINE void Set(int32_t i, T* object) NO_THREAD_SAFETY_ANALYSIS;
62 ALWAYS_INLINE void SetWithoutChecks(int32_t i, T* object) NO_THREAD_SAFETY_ANALYSIS;
67 ALWAYS_INLINE void SetWithoutChecksAndWriteBarrier(int32_t i, T* object)
70 ALWAYS_INLINE T* GetWithoutChecks(int32_t i) SHARED_REQUIRES(Locks::mutator_lock_);
H A Darray.h40 ALWAYS_INLINE static Array* Alloc(Thread* self, Class* array_class, int32_t component_count,
52 ALWAYS_INLINE int32_t GetLength() SHARED_REQUIRES(Locks::mutator_lock_) {
85 ALWAYS_INLINE bool CheckIsValidIndex(int32_t index) SHARED_REQUIRES(Locks::mutator_lock_);
114 const T* GetData() const ALWAYS_INLINE SHARED_REQUIRES(Locks::mutator_lock_) {
118 T* GetData() ALWAYS_INLINE SHARED_REQUIRES(Locks::mutator_lock_) {
122 T Get(int32_t i) ALWAYS_INLINE SHARED_REQUIRES(Locks::mutator_lock_);
124 T GetWithoutChecks(int32_t i) ALWAYS_INLINE SHARED_REQUIRES(Locks::mutator_lock_) {
129 void Set(int32_t i, T value) ALWAYS_INLINE SHARED_REQUIRES(Locks::mutator_lock_);
134 void Set(int32_t i, T value) ALWAYS_INLINE NO_THREAD_SAFETY_ANALYSIS;
141 void SetWithoutChecks(int32_t i, T value) ALWAYS_INLINE NO_THREAD_SAFETY_ANALYSI
[all...]
H A Dobject.h90 ALWAYS_INLINE Class* GetClass() SHARED_REQUIRES(Locks::mutator_lock_);
103 ALWAYS_INLINE bool AtomicSetReadBarrierPointer(Object* expected_rb_ptr, Object* rb_ptr)
112 ALWAYS_INLINE bool InstanceOf(Class* klass) SHARED_REQUIRES(Locks::mutator_lock_);
266 ALWAYS_INLINE T* GetFieldObject(MemberOffset field_offset)
271 ALWAYS_INLINE T* GetFieldObjectVolatile(MemberOffset field_offset)
276 ALWAYS_INLINE void SetFieldObjectWithoutWriteBarrier(MemberOffset field_offset, Object* new_value)
281 ALWAYS_INLINE void SetFieldObject(MemberOffset field_offset, Object* new_value)
286 ALWAYS_INLINE void SetFieldObjectVolatile(MemberOffset field_offset, Object* new_value)
328 ALWAYS_INLINE uint8_t GetFieldBoolean(MemberOffset field_offset)
332 ALWAYS_INLINE int8_
[all...]
H A Dclass.h195 ALWAYS_INLINE uint32_t GetAccessFlags() SHARED_REQUIRES(Locks::mutator_lock_);
201 ALWAYS_INLINE uint32_t GetClassFlags() SHARED_REQUIRES(Locks::mutator_lock_) {
209 ALWAYS_INLINE bool IsInterface() SHARED_REQUIRES(Locks::mutator_lock_) {
214 ALWAYS_INLINE bool IsPublic() SHARED_REQUIRES(Locks::mutator_lock_) {
219 ALWAYS_INLINE bool IsFinal() SHARED_REQUIRES(Locks::mutator_lock_) {
223 ALWAYS_INLINE bool IsFinalizable() SHARED_REQUIRES(Locks::mutator_lock_) {
227 ALWAYS_INLINE void SetRecursivelyInitialized() SHARED_REQUIRES(Locks::mutator_lock_) {
233 ALWAYS_INLINE void SetHasDefaultMethods() SHARED_REQUIRES(Locks::mutator_lock_) {
239 ALWAYS_INLINE void SetFinalizable() SHARED_REQUIRES(Locks::mutator_lock_) {
244 ALWAYS_INLINE boo
890 SHARED_REQUIRES(Locks::mutator_lock_) ALWAYS_INLINE; member in class:art::mirror::FINAL
[all...]
/art/runtime/base/
H A Darena_object.h45 ALWAYS_INLINE void* operator new(size_t, void* ptr) noexcept { return ptr; }
46 ALWAYS_INLINE void operator delete(void*, void*) noexcept { }
/art/runtime/entrypoints/
H A Dentrypoint_utils.h47 ALWAYS_INLINE inline mirror::Class* CheckObjectAlloc(uint32_t type_idx,
52 ALWAYS_INLINE inline mirror::Class* CheckClassInitializedForObjectAlloc(mirror::Class* klass,
62 ALWAYS_INLINE inline mirror::Object* AllocObjectFromCode(uint32_t type_idx,
70 ALWAYS_INLINE inline mirror::Object* AllocObjectFromCodeResolved(mirror::Class* klass,
77 ALWAYS_INLINE inline mirror::Object* AllocObjectFromCodeInitialized(mirror::Class* klass,
84 ALWAYS_INLINE inline mirror::Class* CheckArrayAlloc(uint32_t type_idx,
95 ALWAYS_INLINE inline mirror::Array* AllocArrayFromCode(uint32_t type_idx,
103 ALWAYS_INLINE inline mirror::Array* AllocArrayFromCodeResolved(mirror::Class* klass,
/art/runtime/gc/collector/
H A Dimmune_spaces.h52 ALWAYS_INLINE bool IsInImmuneRegion(const mirror::Object* obj) const {
70 ALWAYS_INLINE bool ContainsObject(const mirror::Object* obj) const {
H A Dimmune_region.h46 ALWAYS_INLINE bool ContainsObject(const mirror::Object* obj) const {

Completed in 254 milliseconds

123