/art/test/003-omnibus-opcodes/src/ |
H A D | Monitor.java | 25 Object obj = new Object(); 26 synchronized (obj) { 28 obj = null; // does NOT cause a failure on exit 29 Main.assertTrue(obj == null); 37 Object obj = null; 40 synchronized (obj) { 48 obj = new Object(); 49 synchronized (obj) {
|
H A D | UnresTest2.java | 8 static boolean checkCasts(Object obj) { argument 12 UnresClass un = (UnresClass) obj; 18 foo = obj instanceof UnresClass;
|
/art/runtime/entrypoints/quick/ |
H A D | quick_lock_entrypoints.cc | 23 extern "C" int artLockObjectFromCode(mirror::Object* obj, Thread* self) 27 if (UNLIKELY(obj == nullptr)) { 32 obj = obj->MonitorEnter(self); // May block 33 CHECK(self->HoldsLock(obj)); 36 obj->MonitorEnter(self); // May block 43 extern "C" int artUnlockObjectFromCode(mirror::Object* obj, Thread* self) 47 if (UNLIKELY(obj == nullptr)) { 52 return obj->MonitorExit(self) ? 0 /* Success */ : -1 /* Failure */;
|
/art/test/132-daemon-locks-shutdown/src/ |
H A D | Main.java | 35 private Object obj; field in class:Main.Wait 37 public Wait(Object obj) { argument 38 this.obj = obj; 43 synchronized(obj) { 45 obj.wait(1);
|
/art/test/024-illegal-access/src/ |
H A D | CheckInstanceof.java | 21 public static void main(Object obj) { argument 22 if (obj instanceof otherpkg.Package)
|
/art/runtime/gc/accounting/ |
H A D | heap_bitmap-inl.h | 38 inline bool HeapBitmap::Test(const mirror::Object* obj) { argument 39 ContinuousSpaceBitmap* bitmap = GetContinuousSpaceBitmap(obj); 41 return bitmap->Test(obj); 44 if (LIKELY(lo_bitmap->HasAddress(obj))) { 45 return lo_bitmap->Test(obj); 48 LOG(FATAL) << "Invalid object " << obj; 52 inline void HeapBitmap::Clear(const mirror::Object* obj) { argument 53 ContinuousSpaceBitmap* bitmap = GetContinuousSpaceBitmap(obj); 55 bitmap->Clear(obj); 59 if (LIKELY(lo_bitmap->HasAddress(obj))) { 67 Set(const mirror::Object* obj, const LargeObjectSetVisitor& visitor) argument 83 AtomicTestAndSet(const mirror::Object* obj, const LargeObjectSetVisitor& visitor) argument [all...] |
H A D | space_bitmap-inl.h | 33 inline bool SpaceBitmap<kAlignment>::AtomicTestAndSet(const mirror::Object* obj) { argument 34 uintptr_t addr = reinterpret_cast<uintptr_t>(obj); 46 DCHECK(Test(obj)); 50 DCHECK(Test(obj)); 55 inline bool SpaceBitmap<kAlignment>::Test(const mirror::Object* obj) const { 56 uintptr_t addr = reinterpret_cast<uintptr_t>(obj); 57 DCHECK(HasAddress(obj)) << obj; 70 mirror::Object* obj = reinterpret_cast<mirror::Object*>(i); 71 if (Test(obj)) { 111 mirror::Object* obj = reinterpret_cast<mirror::Object*>(ptr_base + shift * kAlignment); local 124 mirror::Object* obj = reinterpret_cast<mirror::Object*>(ptr_base + shift * kAlignment); local 150 mirror::Object* obj = reinterpret_cast<mirror::Object*>(ptr_base + shift * kAlignment); local 159 Modify(const mirror::Object* obj) argument [all...] |
H A D | heap_bitmap.h | 38 bool Test(const mirror::Object* obj) SHARED_LOCKS_REQUIRED(Locks::heap_bitmap_lock_); 39 void Clear(const mirror::Object* obj) EXCLUSIVE_LOCKS_REQUIRED(Locks::heap_bitmap_lock_); 41 bool Set(const mirror::Object* obj, const LargeObjectSetVisitor& visitor) 45 bool AtomicTestAndSet(const mirror::Object* obj, const LargeObjectSetVisitor& visitor) 48 ContinuousSpaceBitmap* GetContinuousSpaceBitmap(const mirror::Object* obj) const; 49 LargeObjectBitmap* GetLargeObjectBitmap(const mirror::Object* obj) const;
|
/art/test/023-many-interfaces/src/ |
H A D | ManyInterfaces.java | 190 ManyInterfaces obj = new ManyInterfaces(); 206 testIface001(obj, iter); 211 testIface049(obj, iter); 216 testIface099(obj, iter); 221 testVirt001(obj, iter); 226 testVirt049(obj, iter); 231 testVirt099(obj, iter); 236 testInstance001(obj, iter); 241 testInstance049(obj, iter); 246 testInstance099(obj, ite 329 testVirt001(ManyInterfaces obj, int count) argument 338 testVirt049(ManyInterfaces obj, int count) argument 347 testVirt099(ManyInterfaces obj, int count) argument 356 testInstance001(Object obj, int count) argument 380 testInstance049(Object obj, int count) argument 404 testInstance099(Object obj, int count) argument [all...] |
/art/runtime/gc/collector/ |
H A D | semi_space-inl.h | 34 void operator()(const mirror::Object* obj) const { 35 CHECK(!semi_space_->to_space_->HasAddress(obj)) << "Marking " << obj << " in to_space_"; 37 CHECK(IsAligned<kPageSize>(obj)); 44 inline mirror::Object* SemiSpace::GetForwardingAddressInFromSpace(mirror::Object* obj) const { 45 DCHECK(from_space_->HasAddress(obj)); 46 LockWord lock_word = obj->GetLockWord(false); 59 mirror::Object* obj = obj_ptr->AsMirrorPtr(); local 60 if (obj == nullptr) { 63 if (from_space_->HasAddress(obj)) { [all...] |
H A D | mark_sweep-inl.h | 32 inline void MarkSweep::ScanObjectVisit(mirror::Object* obj, const MarkVisitor& visitor, argument 34 DCHECK(IsMarked(obj)) << "Scanning unmarked object " << obj << "\n" << heap_->DumpSpaces(); 35 obj->VisitReferences<false>(visitor, ref_visitor); 37 mirror::Class* klass = obj->GetClass<kVerifyNone>();
|
H A D | mark_compact.cc | 88 void MarkCompact::ForwardObject(mirror::Object* obj) { argument 89 const size_t alloc_size = RoundUp(obj->SizeOf(), space::BumpPointerSpace::kAlignment); 90 LockWord lock_word = obj->GetLockWord(false); 94 objects_with_lockword_->Set(obj); 97 obj->SetLockWord(LockWord::FromForwardingAddress(reinterpret_cast<size_t>(bump_pointer_)), 107 void operator()(mirror::Object* obj) const EXCLUSIVE_LOCKS_REQUIRED(Locks::mutator_lock_, 109 DCHECK_ALIGNED(obj, space::BumpPointerSpace::kAlignment); 110 DCHECK(collector_->IsMarked(obj)); 111 collector_->ForwardObject(obj); 150 void operator()(const mirror::Object* obj) cons 159 MarkObject(mirror::Object* obj) argument 283 MarkStackPush(Object* obj) argument 334 mirror::Object* obj = *roots[i]; local 348 mirror::Object* obj = roots[i]->AsMirrorPtr(); local 439 MarkedForwardingAddressCallback(mirror::Object* obj, void* arg) argument 444 mirror::Object* obj = reference->AsMirrorPtr(); local 479 UpdateObjectReferences(mirror::Object* obj) argument 540 MoveObject(mirror::Object* obj, size_t len) argument 621 ScanObject(Object* obj) argument 630 Object* obj = mark_stack_->PopBack(); local [all...] |
/art/test/064-field-access/src/ |
H A D | GetNonexistent.java | 19 Object obj = Holder.mObject;
|
/art/runtime/ |
H A D | verify_object-inl.h | 27 inline void VerifyObject(mirror::Object* obj) { argument 28 if (kVerifyObjectSupport > kVerifyObjectModeDisabled && obj != nullptr) { 31 Runtime::Current()->GetHeap()->VerifyObjectBody(obj); 35 bool failed = !IsAligned<kObjectAlignment>(obj); 37 mirror::Class* c = obj->GetClass<kVerifyNone>(); 42 Runtime::Current()->GetHeap()->VerifyObjectBody(obj);
|
H A D | jni_env_ext-inl.h | 27 inline T JNIEnvExt::AddLocalReference(mirror::Object* obj) { argument 28 IndirectRef ref = locals.Add(local_ref_cookie, obj); 36 << entry_count << " (most recent was a " << PrettyTypeOf(obj) << ")\n");
|
H A D | indirect_reference_table-inl.h | 81 mirror::Object* obj = table_[idx].GetReference()->Read<kReadBarrierOption>(); local 82 VerifyObject(obj); 83 return obj; 86 inline void IndirectReferenceTable::Update(IndirectRef iref, mirror::Object* obj) { argument 92 table_[idx].SetReference(obj);
|
H A D | runtime-inl.h | 27 inline bool Runtime::IsClearedJniWeakGlobal(mirror::Object* obj) { argument 28 return obj == GetClearedJniWeakGlobal(); 32 mirror::Object* obj = sentinel_.Read(); local 33 DCHECK(obj != nullptr); 34 return obj;
|
H A D | jni_env_ext.cc | 66 jobject JNIEnvExt::NewLocalRef(mirror::Object* obj) SHARED_LOCKS_REQUIRED(Locks::mutator_lock_) { 67 if (obj == nullptr) { 70 return reinterpret_cast<jobject>(locals.Add(local_ref_cookie, obj)); 73 void JNIEnvExt::DeleteLocalRef(jobject obj) SHARED_LOCKS_REQUIRED(Locks::mutator_lock_) { 74 if (obj != nullptr) { 75 locals.Remove(local_ref_cookie, reinterpret_cast<IndirectRef>(obj));
|
/art/test/469-condition-materialization-regression/src/ |
H A D | Main.java | 22 private static Object obj; field in class:Main 36 obj = new Object();
|
/art/runtime/gc/space/ |
H A D | dlmalloc_space-inl.h | 32 mirror::Object* obj; local 35 obj = AllocWithoutGrowthLocked(self, num_bytes, bytes_allocated, usable_size, 38 if (LIKELY(obj != nullptr)) { 40 memset(obj, 0, num_bytes); 42 return obj; 45 inline size_t DlMallocSpace::AllocationSizeNonvirtual(mirror::Object* obj, size_t* usable_size) { argument 46 void* obj_ptr = const_cast<void*>(reinterpret_cast<const void*>(obj));
|
H A D | region_space-inl.h | 47 mirror::Object* obj; local 51 obj = current_region_->Alloc(num_bytes, bytes_allocated, usable_size, 55 obj = evac_region_->Alloc(num_bytes, bytes_allocated, usable_size, 58 if (LIKELY(obj != nullptr)) { 59 return obj; 64 obj = current_region_->Alloc(num_bytes, bytes_allocated, usable_size, 67 obj = evac_region_->Alloc(num_bytes, bytes_allocated, usable_size, 70 if (LIKELY(obj != nullptr)) { 71 return obj; 84 obj 238 mirror::Object* obj = reinterpret_cast<mirror::Object*>(r->Begin()); local 248 mirror::Object* obj = reinterpret_cast<mirror::Object*>(pos); local 260 GetNextObject(mirror::Object* obj) argument [all...] |
/art/runtime/native/ |
H A D | sun_misc_Unsafe.cc | 31 mirror::Object* obj = soa.Decode<mirror::Object*>(javaObj); local 33 bool success = obj->CasFieldStrongSequentiallyConsistent32<false>(MemberOffset(offset), 41 mirror::Object* obj = soa.Decode<mirror::Object*>(javaObj); local 43 bool success = obj->CasFieldStrongSequentiallyConsistent64<false>(MemberOffset(offset), 51 mirror::Object* obj = soa.Decode<mirror::Object*>(javaObj); local 55 bool success = obj->CasFieldStrongSequentiallyConsistentObject<false>(MemberOffset(offset), 62 mirror::Object* obj = soa.Decode<mirror::Object*>(javaObj); local 63 return obj->GetField32(MemberOffset(offset)); 68 mirror::Object* obj = soa.Decode<mirror::Object*>(javaObj); local 69 return obj 74 mirror::Object* obj = soa.Decode<mirror::Object*>(javaObj); local 82 mirror::Object* obj = soa.Decode<mirror::Object*>(javaObj); local 90 mirror::Object* obj = soa.Decode<mirror::Object*>(javaObj); local 98 mirror::Object* obj = soa.Decode<mirror::Object*>(javaObj); local 104 mirror::Object* obj = soa.Decode<mirror::Object*>(javaObj); local 110 mirror::Object* obj = soa.Decode<mirror::Object*>(javaObj); local 118 mirror::Object* obj = soa.Decode<mirror::Object*>(javaObj); local 126 mirror::Object* obj = soa.Decode<mirror::Object*>(javaObj); local 134 mirror::Object* obj = soa.Decode<mirror::Object*>(javaObj); local 141 mirror::Object* obj = soa.Decode<mirror::Object*>(javaObj); local 149 mirror::Object* obj = soa.Decode<mirror::Object*>(javaObj); local 158 mirror::Object* obj = soa.Decode<mirror::Object*>(javaObj); local 167 mirror::Object* obj = soa.Decode<mirror::Object*>(javaObj); local [all...] |
/art/runtime/gc/ |
H A D | heap-inl.h | 53 mirror::Object* obj; local 55 obj = AllocLargeObject<kInstrumented, PreFenceVisitor>(self, &klass, byte_count, 57 if (obj != nullptr) { 58 return obj; 67 AllocationTimer alloc_timer(this, &obj); 78 obj = self->AllocTlab(byte_count); 79 DCHECK(obj != nullptr) << "AllocTlab can't fail"; 80 obj->SetClass(klass); 83 obj->SetReadBarrierPointer(obj); 199 PushOnAllocationStack(Thread* self, mirror::Object** obj) argument 424 CheckConcurrentGC(Thread* self, size_t new_num_bytes_allocated, mirror::Object** obj) argument [all...] |
/art/test/476-checker-ctor-memory-barrier/src/ |
H A D | Main.java | 26 public ClassWithFinals obj; field in class:ClassWithFinals 57 obj = new ClassWithFinals();
|
/art/tools/ |
H A D | run-libcore-tests.sh | 23 test_jar=out/target/common/obj/JAVA_LIBRARIES/core-tests_intermediates/javalib.jar
|