/art/tools/ahat/src/main/com/android/ahat/heapdump/ |
H A D | Reference.java | 20 * Reference represents a reference from 'src' to 'ref' through 'field'. 29 public final AhatInstance ref; field in class:Reference 32 public Reference(AhatInstance src, String field, AhatInstance ref, boolean strong) { argument 35 this.ref = ref;
|
H A D | AhatInstance.java | 364 * This is only relevant for instances of java.lang.ref.Reference or its 613 for (Reference ref : root.getReferences()) { 614 strong.add(ref); 618 Reference ref = strong.poll(); 619 assert ref.strong; 621 if (ref.ref.mNextInstanceToGcRoot == null) { 622 // This is the first time we have seen ref.ref. 623 ref [all...] |
H A D | DominatorReferenceIterator.java | 39 Reference ref = mIter.next(); 40 if (ref.strong) { 41 mNext = ref.ref;
|
/art/runtime/ |
H A D | read_barrier-inl.h | 50 // the is_gray load to the ref field (ptr) load to avoid needing a load-load barrier between 59 MirrorType* ref = ref_addr->template AsMirrorPtr<kIsVolatile>(); local 60 MirrorType* old_ref = ref; 63 ref = reinterpret_cast<MirrorType*>(Mark(ref)); 66 if (kAlwaysUpdateField && ref != old_ref) { 68 offset, old_ref, ref); 71 AssertToSpaceInvariant(obj, offset, ref); 72 return ref; 77 MirrorType* ref local 104 MirrorType* ref = *root; local 152 MirrorType* ref = root->AsMirrorPtr(); local 188 IsMarked(MirrorType* ref) argument 225 AssertToSpaceInvariant(mirror::Object* obj, MemberOffset offset, mirror::Object* ref) argument 236 AssertToSpaceInvariant(GcRootSource* gc_root_source, mirror::Object* ref) argument [all...] |
H A D | obj_ptr-inl.h | 44 uintptr_t ref = reinterpret_cast<uintptr_t>(ptr); local 45 DCHECK_ALIGNED(ref, kObjectAlignment); 46 if (kObjPtrPoisoning && ref != 0) { 47 DCHECK_LE(ref, 0xFFFFFFFFU); 48 ref >>= kObjectAlignmentShift; 52 ref |= self->GetPoisonObjectCookie() << kCookieShift; 54 return ref;
|
H A D | jni_env_ext-inl.h | 29 IndirectRef ref = locals_.Add(local_ref_cookie_, obj, &error_msg); local 30 if (UNLIKELY(ref == nullptr)) { 49 return reinterpret_cast<T>(ref);
|
H A D | gc_root-inl.h | 37 inline GcRoot<MirrorType>::GcRoot(MirrorType* ref) argument 38 : root_(mirror::CompressedReference<mirror::Object>::FromMirrorPtr(ref)) { } 41 inline GcRoot<MirrorType>::GcRoot(ObjPtr<MirrorType> ref) argument 42 : GcRoot(ref.Ptr()) { }
|
H A D | read_barrier.h | 76 ALWAYS_INLINE static MirrorType* IsMarked(MirrorType* ref) 82 static void AssertToSpaceInvariant(mirror::Object* ref) 84 AssertToSpaceInvariant(nullptr, MemberOffset(0), ref); 88 mirror::Object* ref) 91 static void AssertToSpaceInvariant(GcRootSource* gc_root_source, mirror::Object* ref) 95 static void MaybeAssertToSpaceInvariant(mirror::Object* ref) 98 AssertToSpaceInvariant(ref);
|
/art/runtime/gc/ |
H A D | reference_queue.cc | 33 void ReferenceQueue::AtomicEnqueueIfNotEnqueued(Thread* self, ObjPtr<mirror::Reference> ref) { argument 34 DCHECK(ref != nullptr); 36 if (ref->IsUnprocessed()) { 37 EnqueueReference(ref); 41 void ReferenceQueue::EnqueueReference(ObjPtr<mirror::Reference> ref) { argument 42 DCHECK(ref != nullptr); 43 CHECK(ref->IsUnprocessed()); 45 // 1 element cyclic queue, ie: Reference ref = ..; ref.pendingNext = ref; 60 ObjPtr<mirror::Reference> ref = list_->GetPendingNext<kWithoutReadBarrier>(); local 75 DisableReadBarrierForReference(ObjPtr<mirror::Reference> ref) argument 135 ObjPtr<mirror::Reference> ref = DequeuePendingReference(); local 157 ObjPtr<mirror::FinalizerReference> ref = DequeuePendingReference()->AsFinalizerReference(); local 184 ObjPtr<mirror::Reference> ref = head; local [all...] |
/art/compiler/utils/ |
H A D | intrusive_forward_list_test.cc | 169 std::forward_list<int> ref({ 1, 2, 7 }); 170 std::vector<ValueType> storage(ref.begin(), ref.end()); 172 ASSERT_LISTS_EQUAL(ref, ifl); 206 std::forward_list<int> ref; local 208 ASSERT_LISTS_EQUAL(ref, ifl); 209 ref.push_front(3); 211 ASSERT_LISTS_EQUAL(ref, ifl); 213 ref.push_front(7); 215 ASSERT_LISTS_EQUAL(ref, if 239 std::forward_list<int> ref; local [all...] |
H A D | atomic_dex_ref_map-inl.h | 53 AtomicDexRefMap<DexFileReferenceType, Value>::Insert(const DexFileReferenceType& ref, argument 56 ElementArray* const array = GetArray(ref.dex_file); 60 DCHECK_LT(ref.index, array->size()); 61 return (*array)[ref.index].CompareAndSetStrongSequentiallyConsistent(expected, desired) 67 inline bool AtomicDexRefMap<DexFileReferenceType, Value>::Get(const DexFileReferenceType& ref, argument 69 const ElementArray* const array = GetArray(ref.dex_file); 73 *out = (*array)[ref.index].LoadRelaxed(); 78 inline bool AtomicDexRefMap<DexFileReferenceType, Value>::Remove(const DexFileReferenceType& ref, argument 80 ElementArray* const array = GetArray(ref.dex_file); 84 *out = (*array)[ref [all...] |
/art/compiler/dex/ |
H A D | quick_compiler_callbacks.cc | 35 void QuickCompilerCallbacks::ClassRejected(ClassReference ref) { argument 37 verification_results_->AddRejectedClass(ref); 41 ClassStatus QuickCompilerCallbacks::GetPreviousClassState(ClassReference ref) { argument 51 return compiler_driver_->GetClassStatus(ref); 54 void QuickCompilerCallbacks::UpdateClassState(ClassReference ref, ClassStatus status) { argument 57 compiler_driver_->RecordClassStatus(ref, status);
|
H A D | verification_results.cc | 42 atomic_verified_methods_.Visit([](const DexFileReference& ref ATTRIBUTE_UNUSED, 50 MethodReference ref = method_verifier->GetMethodReference(); local 56 AtomicMap::InsertResult result = atomic_verified_methods_.Insert(ref, 65 CHECK(atomic_verified_methods_.Get(ref, &existing)); 70 auto it = verified_methods_.find(ref); 73 verified_methods_.Put(ref, verified_method.get()); 74 DCHECK(verified_methods_.find(ref) != verified_methods_.end()); 81 DCHECK_EQ(GetVerifiedMethod(ref), verified_method.get()); 85 LOG(WARNING) << "Method processed more than once: " << ref.PrettyMethod(); 100 const VerifiedMethod* VerificationResults::GetVerifiedMethod(MethodReference ref) { argument 110 CreateVerifiedMethodFor(MethodReference ref) argument 124 AddRejectedClass(ClassReference ref) argument 132 IsClassRejected(ClassReference ref) argument 157 MethodReference ref = it->first; local [all...] |
H A D | verification_results.h | 51 void CreateVerifiedMethodFor(MethodReference ref) 54 const VerifiedMethod* GetVerifiedMethod(MethodReference ref) 57 void AddRejectedClass(ClassReference ref) REQUIRES(!rejected_classes_lock_); 58 bool IsClassRejected(ClassReference ref) REQUIRES(!rejected_classes_lock_);
|
/art/runtime/native/ |
H A D | java_lang_ref_FinalizerReference.cc | 33 ObjPtr<mirror::FinalizerReference> ref = soa.Decode<mirror::FinalizerReference>(javaThis); local 34 return Runtime::Current()->GetHeap()->GetReferenceProcessor()->MakeCircularListIfUnenqueued(ref); 39 ObjPtr<mirror::Reference> ref = soa.Decode<mirror::Reference>(javaThis); local 41 Runtime::Current()->GetHeap()->GetReferenceProcessor()->GetReferent(soa.Self(), ref); 51 REGISTER_NATIVE_METHODS("java/lang/ref/FinalizerReference");
|
H A D | java_lang_ref_Reference.cc | 33 ObjPtr<mirror::Reference> ref = soa.Decode<mirror::Reference>(javaThis); local 35 Runtime::Current()->GetHeap()->GetReferenceProcessor()->GetReferent(soa.Self(), ref); 41 ObjPtr<mirror::Reference> ref = soa.Decode<mirror::Reference>(javaThis); local 42 Runtime::Current()->GetHeap()->GetReferenceProcessor()->ClearReferent(ref); 51 REGISTER_NATIVE_METHODS("java/lang/ref/Reference");
|
/art/profman/ |
H A D | boot_image_profile.cc | 56 MethodReference ref(dex_file.get(), i); 60 Hotness hotness = profile->GetMethodHotness(ref); 63 out_profile->AddMethodHotness(ref, hotness); 64 inferred_classes.emplace(profile.get(), ref.GetMethodId().class_idx_); 73 out_profile->AddMethodHotness(ref, hotness); 79 TypeReference ref(dex_file.get(), class_def.class_idx_); 113 auto it = inferred_classes.find(std::make_pair(profile.get(), ref.TypeIndex())); 115 profile->ContainsClass(*ref.dex_file, ref.TypeIndex())) { 124 out_profile->AddClassForDex(ref); [all...] |
/art/runtime/gc/accounting/ |
H A D | mod_union_table-inl.h | 36 bool ShouldAddReference(const mirror::Object* ref) const OVERRIDE ALWAYS_INLINE { 37 return !space_->HasAddress(ref);
|
/art/benchmark/jobject-benchmark/ |
H A D | jobject_benchmark.cc | 32 jobject ref = soa.Env()->AddLocalReference<jobject>(obj); local 33 soa.Env()->DeleteLocalRef(ref); 42 jobject ref = soa.Env()->AddLocalReference<jobject>(obj); local 44 CHECK_EQ(soa.Decode<mirror::Object>(ref), obj); 46 soa.Env()->DeleteLocalRef(ref); 55 jobject ref = soa.Vm()->AddGlobalRef(soa.Self(), obj); local 56 soa.Vm()->DeleteGlobalRef(soa.Self(), ref); 65 jobject ref = soa.Vm()->AddGlobalRef(soa.Self(), obj); local 67 CHECK_EQ(soa.Decode<mirror::Object>(ref), obj); 69 soa.Vm()->DeleteGlobalRef(soa.Self(), ref); 78 jobject ref = soa.Vm()->AddWeakGlobalRef(soa.Self(), obj); local 88 jobject ref = soa.Vm()->AddWeakGlobalRef(soa.Self(), obj); local [all...] |
/art/libartbase/base/ |
H A D | transform_iterator_test.cc | 249 auto ref = [](const ValueHolder& h) -> const int& { return h.value; }; local 253 using vector_titer = decltype(MakeTransformIterator(input.begin(), ref)); 260 using vector_ctiter = decltype(MakeTransformIterator(input.cbegin(), ref)); 267 using vector_rtiter = decltype(MakeTransformIterator(input.rbegin(), ref)); 274 using vector_crtiter = decltype(MakeTransformIterator(input.crbegin(), ref)); 281 std::copy(MakeTransformIterator(input.begin(), ref), 282 MakeTransformIterator(input.end(), ref), 287 std::copy(MakeTransformIterator(input.cbegin(), ref), 288 MakeTransformIterator(input.cend(), ref), 293 std::copy(MakeTransformIterator(input.rbegin(), ref), 342 auto ref = [](ValueHolder& h) -> int& { return h.value; }; local 410 Ref ref; local 522 auto ref = [](ValueHolder& h) -> int& { return h.value; }; local [all...] |
/art/runtime/gc/collector/ |
H A D | concurrent_copying-inl.h | 35 mirror::Object* ref, accounting::ContinuousSpaceBitmap* bitmap) { 45 if (bitmap->Test(ref)) { 46 return ref; 56 success = ref->AtomicSetReadBarrierState(/* expected_rb_state */ ReadBarrier::WhiteState(), 59 success = !bitmap->AtomicTestAndSet(ref); 64 DCHECK_EQ(ref->GetReadBarrierState(), ReadBarrier::GrayState()); 66 PushOntoMarkStack(ref); 68 return ref; 72 inline mirror::Object* ConcurrentCopying::MarkImmuneSpace(mirror::Object* ref) { argument 88 return ref; 34 MarkUnevacFromSpaceRegion( mirror::Object* ref, accounting::ContinuousSpaceBitmap* bitmap) argument [all...] |
H A D | semi_space-inl.h | 63 auto slow_path = [this](const mirror::Object* ref) { 64 CHECK(!to_space_->HasAddress(ref)) << "Marking " << ref << " in to_space_"; 66 CHECK_ALIGNED(ref, kPageSize);
|
/art/test/004-NativeAllocations/src-art/ |
H A D | Main.java | 18 import java.lang.ref.ReferenceQueue; 19 import java.lang.ref.PhantomReference; 27 // Save ref as a static field to ensure it doesn't get GC'd before the 29 static PhantomReference ref = null; field in class:Main 55 ref = allocPhantom(queue); 57 for (int i = 0; !ref.isEnqueued() && i < allocationCount; ++i) {
|
/art/runtime/gc/space/ |
H A D | region_space.h | 202 bool IsInFromSpace(mirror::Object* ref) { argument 203 if (HasAddress(ref)) { 204 Region* r = RefToRegionUnlocked(ref); 210 bool IsInNewlyAllocatedRegion(mirror::Object* ref) { argument 211 if (HasAddress(ref)) { 212 Region* r = RefToRegionUnlocked(ref); 218 bool IsInUnevacFromSpace(mirror::Object* ref) { argument 219 if (HasAddress(ref)) { 220 Region* r = RefToRegionUnlocked(ref); 226 bool IsInToSpace(mirror::Object* ref) { argument 236 GetRegionType(mirror::Object* ref) argument 245 GetRegionTypeUnsafe(mirror::Object* ref) argument 262 AddLiveBytes(mirror::Object* ref, size_t alloc_size) argument [all...] |
/art/runtime/interpreter/mterp/arm/ |
H A D | op_sput.S | 9 FETCH r0, 1 @ r0<- field ref BBBB
|