/art/runtime/ |
H A D | jni_env_ext-inl.h | 29 IndirectRef ref = locals_.Add(local_ref_cookie_, obj, &error_msg); local 30 if (UNLIKELY(ref == nullptr)) { 49 return reinterpret_cast<T>(ref);
|
H A D | gc_root-inl.h | 37 inline GcRoot<MirrorType>::GcRoot(MirrorType* ref) argument 38 : root_(mirror::CompressedReference<mirror::Object>::FromMirrorPtr(ref)) { } 41 inline GcRoot<MirrorType>::GcRoot(ObjPtr<MirrorType> ref) argument 42 : GcRoot(ref.Ptr()) { }
|
H A D | handle.h | 132 StackReference<mirror::Object>* ref = Handle<T>::GetReference(); local 133 T* old = down_cast<T*>(ref->AsMirrorPtr()); 134 ref->Assign(reference); 139 StackReference<mirror::Object>* ref = Handle<T>::GetReference(); local 140 T* old = down_cast<T*>(ref->AsMirrorPtr()); 141 ref->Assign(reference.Ptr());
|
H A D | obj_ptr-inl.h | 44 uintptr_t ref = reinterpret_cast<uintptr_t>(ptr); local 45 DCHECK_ALIGNED(ref, kObjectAlignment); 46 if (kObjPtrPoisoning && ref != 0) { 47 DCHECK_LE(ref, 0xFFFFFFFFU); 48 ref >>= kObjectAlignmentShift; 52 ref |= self->GetPoisonObjectCookie() << kCookieShift; 54 return ref;
|
H A D | handle_scope_test.cc | 79 void VisitRootIfNonNull(StackReference<mirror::Object>* ref) { argument 80 if (!ref->IsNull()) { 81 visited.insert(ref); 117 for (StackReference<mirror::Object>* ref : visitor.visited) { 118 EXPECT_TRUE(base->Contains(ref));
|
H A D | read_barrier-inl.h | 50 // the is_gray load to the ref field (ptr) load to avoid needing a load-load barrier between 59 MirrorType* ref = ref_addr->template AsMirrorPtr<kIsVolatile>(); local 60 MirrorType* old_ref = ref; 63 ref = reinterpret_cast<MirrorType*>(Mark(ref)); 66 if (kAlwaysUpdateField && ref != old_ref) { 68 offset, old_ref, ref); 71 AssertToSpaceInvariant(obj, offset, ref); 72 return ref; 77 MirrorType* ref local 104 MirrorType* ref = *root; local 152 MirrorType* ref = root->AsMirrorPtr(); local 188 IsMarked(MirrorType* ref) argument 225 AssertToSpaceInvariant(mirror::Object* obj, MemberOffset offset, mirror::Object* ref) argument 236 AssertToSpaceInvariant(GcRootSource* gc_root_source, mirror::Object* ref) argument [all...] |
/art/runtime/native/ |
H A D | java_lang_ref_FinalizerReference.cc | 33 ObjPtr<mirror::FinalizerReference> ref = soa.Decode<mirror::FinalizerReference>(javaThis); local 34 return Runtime::Current()->GetHeap()->GetReferenceProcessor()->MakeCircularListIfUnenqueued(ref); 39 ObjPtr<mirror::Reference> ref = soa.Decode<mirror::Reference>(javaThis); local 41 Runtime::Current()->GetHeap()->GetReferenceProcessor()->GetReferent(soa.Self(), ref); 51 REGISTER_NATIVE_METHODS("java/lang/ref/FinalizerReference");
|
H A D | java_lang_ref_Reference.cc | 33 ObjPtr<mirror::Reference> ref = soa.Decode<mirror::Reference>(javaThis); local 35 Runtime::Current()->GetHeap()->GetReferenceProcessor()->GetReferent(soa.Self(), ref); 41 ObjPtr<mirror::Reference> ref = soa.Decode<mirror::Reference>(javaThis); local 42 Runtime::Current()->GetHeap()->GetReferenceProcessor()->ClearReferent(ref); 51 REGISTER_NATIVE_METHODS("java/lang/ref/Reference");
|
/art/tools/ahat/src/main/com/android/ahat/heapdump/ |
H A D | Reference.java | 20 * Reference represents a reference from 'src' to 'ref' through 'field'. 29 public final AhatInstance ref; field in class:Reference 32 public Reference(AhatInstance src, String field, AhatInstance ref, boolean strong) { argument 35 this.ref = ref;
|
/art/benchmark/jobject-benchmark/ |
H A D | jobject_benchmark.cc | 32 jobject ref = soa.Env()->AddLocalReference<jobject>(obj); local 33 soa.Env()->DeleteLocalRef(ref); 42 jobject ref = soa.Env()->AddLocalReference<jobject>(obj); local 44 CHECK_EQ(soa.Decode<mirror::Object>(ref), obj); 46 soa.Env()->DeleteLocalRef(ref); 55 jobject ref = soa.Vm()->AddGlobalRef(soa.Self(), obj); local 56 soa.Vm()->DeleteGlobalRef(soa.Self(), ref); 65 jobject ref = soa.Vm()->AddGlobalRef(soa.Self(), obj); local 67 CHECK_EQ(soa.Decode<mirror::Object>(ref), obj); 69 soa.Vm()->DeleteGlobalRef(soa.Self(), ref); 78 jobject ref = soa.Vm()->AddWeakGlobalRef(soa.Self(), obj); local 88 jobject ref = soa.Vm()->AddWeakGlobalRef(soa.Self(), obj); local [all...] |
/art/compiler/dex/ |
H A D | quick_compiler_callbacks.cc | 35 void QuickCompilerCallbacks::ClassRejected(ClassReference ref) { argument 37 verification_results_->AddRejectedClass(ref); 41 ClassStatus QuickCompilerCallbacks::GetPreviousClassState(ClassReference ref) { argument 51 return compiler_driver_->GetClassStatus(ref); 54 void QuickCompilerCallbacks::UpdateClassState(ClassReference ref, ClassStatus status) { argument 57 compiler_driver_->RecordClassStatus(ref, status);
|
H A D | verification_results.cc | 42 atomic_verified_methods_.Visit([](const DexFileReference& ref ATTRIBUTE_UNUSED, 50 MethodReference ref = method_verifier->GetMethodReference(); local 56 AtomicMap::InsertResult result = atomic_verified_methods_.Insert(ref, 65 CHECK(atomic_verified_methods_.Get(ref, &existing)); 70 auto it = verified_methods_.find(ref); 73 verified_methods_.Put(ref, verified_method.get()); 74 DCHECK(verified_methods_.find(ref) != verified_methods_.end()); 81 DCHECK_EQ(GetVerifiedMethod(ref), verified_method.get()); 85 LOG(WARNING) << "Method processed more than once: " << ref.PrettyMethod(); 100 const VerifiedMethod* VerificationResults::GetVerifiedMethod(MethodReference ref) { argument 110 CreateVerifiedMethodFor(MethodReference ref) argument 124 AddRejectedClass(ClassReference ref) argument 132 IsClassRejected(ClassReference ref) argument 157 MethodReference ref = it->first; local [all...] |
/art/dex2oat/linker/ |
H A D | multi_oat_relative_patcher.cc | 63 MethodReference ref) { 64 auto it = map.find(ref); 62 FindMethodOffset( MethodReference ref) argument
|
/art/test/004-NativeAllocations/src-art/ |
H A D | Main.java | 18 import java.lang.ref.ReferenceQueue; 19 import java.lang.ref.PhantomReference; 27 // Save ref as a static field to ensure it doesn't get GC'd before the 29 static PhantomReference ref = null; field in class:Main 55 ref = allocPhantom(queue); 57 for (int i = 0; !ref.isEnqueued() && i < allocationCount; ++i) {
|
/art/libartbase/base/ |
H A D | transform_array_ref_test.cc | 118 Ref ref; local 122 auto taref = MakeTransformArrayRef(input, ref); 158 taref = MakeTransformArrayRef(transformed, ref); 166 auto ctaref = MakeTransformArrayRef(cinput, ref);
|
H A D | debug_stack.h | 67 explicit DebugStackIndirectTopRefImpl(DebugStackReferenceImpl<false>* ref ATTRIBUTE_UNUSED) {} 121 explicit DebugStackIndirectTopRefImpl(DebugStackReferenceImpl<kIsDebug>* ref) argument 122 : ref_(ref) {
|
H A D | transform_iterator_test.cc | 249 auto ref = [](const ValueHolder& h) -> const int& { return h.value; }; local 253 using vector_titer = decltype(MakeTransformIterator(input.begin(), ref)); 260 using vector_ctiter = decltype(MakeTransformIterator(input.cbegin(), ref)); 267 using vector_rtiter = decltype(MakeTransformIterator(input.rbegin(), ref)); 274 using vector_crtiter = decltype(MakeTransformIterator(input.crbegin(), ref)); 281 std::copy(MakeTransformIterator(input.begin(), ref), 282 MakeTransformIterator(input.end(), ref), 287 std::copy(MakeTransformIterator(input.cbegin(), ref), 288 MakeTransformIterator(input.cend(), ref), 293 std::copy(MakeTransformIterator(input.rbegin(), ref), 342 auto ref = [](ValueHolder& h) -> int& { return h.value; }; local 410 Ref ref; local 522 auto ref = [](ValueHolder& h) -> int& { return h.value; }; local [all...] |
/art/compiler/utils/ |
H A D | atomic_dex_ref_map-inl.h | 53 AtomicDexRefMap<DexFileReferenceType, Value>::Insert(const DexFileReferenceType& ref, argument 56 ElementArray* const array = GetArray(ref.dex_file); 60 DCHECK_LT(ref.index, array->size()); 61 return (*array)[ref.index].CompareAndSetStrongSequentiallyConsistent(expected, desired) 67 inline bool AtomicDexRefMap<DexFileReferenceType, Value>::Get(const DexFileReferenceType& ref, argument 69 const ElementArray* const array = GetArray(ref.dex_file); 73 *out = (*array)[ref.index].LoadRelaxed(); 78 inline bool AtomicDexRefMap<DexFileReferenceType, Value>::Remove(const DexFileReferenceType& ref, argument 80 ElementArray* const array = GetArray(ref.dex_file); 84 *out = (*array)[ref [all...] |
/art/runtime/gc/collector/ |
H A D | concurrent_copying-inl.h | 35 mirror::Object* ref, accounting::ContinuousSpaceBitmap* bitmap) { 45 if (bitmap->Test(ref)) { 46 return ref; 56 success = ref->AtomicSetReadBarrierState(/* expected_rb_state */ ReadBarrier::WhiteState(), 59 success = !bitmap->AtomicTestAndSet(ref); 64 DCHECK_EQ(ref->GetReadBarrierState(), ReadBarrier::GrayState()); 66 PushOntoMarkStack(ref); 68 return ref; 72 inline mirror::Object* ConcurrentCopying::MarkImmuneSpace(mirror::Object* ref) { argument 88 return ref; 34 MarkUnevacFromSpaceRegion( mirror::Object* ref, accounting::ContinuousSpaceBitmap* bitmap) argument [all...] |
/art/runtime/mirror/ |
H A D | object_reference.h | 45 static MirrorType* Decompress(uint32_t ref) { argument 46 uintptr_t as_bits = kPoisonReferences ? -ref : ref;
|
/art/tools/veridex/ |
H A D | hidden_api.h | 62 static std::string GetApiMethodName(MethodReference ref) { argument 63 return HiddenApi::GetApiMethodName(*ref.dex_file, ref.index);
|
H A D | hidden_api_finder.cc | 33 MethodReference ref) { 38 method_locations_[name].push_back(ref); 44 MethodReference ref) { 49 field_locations_[name].push_back(ref); 206 for (const MethodReference& ref : pair.second) { 207 os << kPrefix << HiddenApi::GetApiMethodName(ref) << std::endl; 218 for (const MethodReference& ref : pair.second) { 219 os << kPrefix << HiddenApi::GetApiMethodName(ref) << std::endl; 236 for (const MethodReference& ref : reflection_locations_[name]) { 237 os << kPrefix << HiddenApi::GetApiMethodName(ref) << st 31 CheckMethod(uint32_t method_id, VeridexResolver* resolver, MethodReference ref) argument 42 CheckField(uint32_t field_id, VeridexResolver* resolver, MethodReference ref) argument [all...] |
H A D | precise_hidden_api_finder.cc | 59 MethodReference ref) { 62 concrete_uses_[ref].push_back(info); 64 abstract_uses_[ref].push_back(info); 100 MethodReference ref = it.first; local 107 named_uses[full_name].push_back(ref); 119 for (const MethodReference& ref : it.second) { 120 os << kPrefix << HiddenApi::GetApiMethodName(ref) << std::endl; 58 AddUsesAt(const std::vector<ReflectAccessInfo>& accesses, MethodReference ref) argument
|
/art/runtime/gc/ |
H A D | reference_processor.cc | 68 // weak ref access is disabled as the call includes a read barrier which may push a ref onto the 108 // presence of threads blocking for weak ref access. 140 // Weak ref access is enabled at Zygote compaction by SemiSpace (concurrent == false). 207 // Process the "referent" field in a java.lang.ref.Reference. If the referent has not yet been 210 ObjPtr<mirror::Reference> ref, 212 // klass can be the class of the old object if the visitor already updated the class of ref. 215 mirror::HeapReference<mirror::Object>* referent = ref->GetReferentReferenceAddr(); 233 soft_reference_queue_.AtomicEnqueueIfNotEnqueued(self, ref); 235 weak_reference_queue_.AtomicEnqueueIfNotEnqueued(self, ref); 209 DelayReferenceReferent(ObjPtr<mirror::Class> klass, ObjPtr<mirror::Reference> ref, collector::GarbageCollector* collector) argument 293 ClearReferent(ObjPtr<mirror::Reference> ref) argument [all...] |
H A D | reference_queue.cc | 33 void ReferenceQueue::AtomicEnqueueIfNotEnqueued(Thread* self, ObjPtr<mirror::Reference> ref) { argument 34 DCHECK(ref != nullptr); 36 if (ref->IsUnprocessed()) { 37 EnqueueReference(ref); 41 void ReferenceQueue::EnqueueReference(ObjPtr<mirror::Reference> ref) { argument 42 DCHECK(ref != nullptr); 43 CHECK(ref->IsUnprocessed()); 45 // 1 element cyclic queue, ie: Reference ref = ..; ref.pendingNext = ref; 60 ObjPtr<mirror::Reference> ref = list_->GetPendingNext<kWithoutReadBarrier>(); local 75 DisableReadBarrierForReference(ObjPtr<mirror::Reference> ref) argument 135 ObjPtr<mirror::Reference> ref = DequeuePendingReference(); local 157 ObjPtr<mirror::FinalizerReference> ref = DequeuePendingReference()->AsFinalizerReference(); local 184 ObjPtr<mirror::Reference> ref = head; local [all...] |