/art/runtime/ |
H A D | object_lock.cc | 26 ObjectLock<T>::ObjectLock(Thread* self, Handle<T> object) : self_(self), obj_(object) { 28 obj_->MonitorEnter(self_); 33 obj_->MonitorExit(self_); 38 Monitor::Wait(self_, obj_.Get(), 0, 0, false, kWaiting); 43 obj_->Notify(self_); 48 obj_->NotifyAll(self_); 52 ObjectTryLock<T>::ObjectTryLock(Thread* self, Handle<T> object) : self_(self), obj_(object) { 54 acquired_ = obj_->MonitorTryEnter(self_) != nullptr; 60 obj_->MonitorExit(self_);
|
H A D | object_lock.h | 43 Handle<T> const obj_; member in class:art::ObjectLock 61 Handle<T> const obj_; member in class:art::ObjectTryLock
|
H A D | handle_scope.h | 181 : MutableHandle<T>(handle), obj_(obj) { 187 *obj_ = MutableHandle<T>::Get(); 191 T** const obj_; member in class:art::HandleWrapper 201 : MutableHandle<T>(handle), obj_(obj) {} 206 *obj_ = ObjPtr<T>(MutableHandle<T>::Get()); 210 ObjPtr<T>* const obj_; member in class:art::HandleWrapperObjPtr
|
H A D | monitor_test.cc | 327 explicit TryLockTask(Handle<mirror::Object> obj) : obj_(obj) {} 332 ObjectTryLock<mirror::Object> lock(self, obj_); 341 Handle<mirror::Object> obj_; member in class:art::TryLockTask
|
H A D | monitor.h | 119 return obj_.Read<kReadBarrierOption>(); 304 GcRoot<mirror::Object> obj_; member in class:art::Monitor
|
H A D | monitor.cc | 98 obj_(GcRoot<mirror::Object>(obj)), 121 obj_(GcRoot<mirror::Object>(obj)), 267 obj_ = GcRoot<mirror::Object>(object); 952 monitor->obj_ = GcRoot<mirror::Object>(nullptr);
|
/art/compiler/optimizing/ |
H A D | code_generator_arm64.cc | 750 obj_(obj), 770 DCHECK(obj_.IsW()); 819 __ Ldr(temp_, HeapOperand(obj_, monitor_offset)); 832 __ Add(obj_.X(), obj_.X(), Operand(temp_.X(), LSR, 32)); 839 obj_, 862 Register obj_; member in class:art::arm64::LoadReferenceWithBakerReadBarrierSlowPathARM64 863 // The offset, index and scale factor to access the reference in `obj_`. 871 // A temporary register used to hold the lock word of `obj_`. 910 obj_(ob 1088 const Register obj_; member in class:art::arm64::LoadReferenceWithBakerReadBarrierAndUpdateFieldSlowPathARM64 1282 const Location obj_; member in class:art::arm64::ReadBarrierForHeapReferenceSlowPathARM64 [all...] |
H A D | code_generator_arm_vixl.cc | 876 obj_(obj), 942 arm_codegen->GetAssembler()->LoadFromOffset(kLoadWord, temp_, obj_, monitor_offset); 955 __ Add(obj_, obj_, Operand(temp_, ShiftType::LSR, 32)); 960 instruction_, ref_, obj_, offset_, index_, scale_factor_, /* needs_null_check */ false); 981 vixl32::Register obj_; member in class:art::arm::LoadReferenceWithBakerReadBarrierSlowPathARMVIXL 982 // The offset, index and scale factor to access the reference in `obj_`. 988 // A temporary register used to hold the lock word of `obj_`. 1027 obj_(obj), 1084 arm_codegen->GetAssembler()->LoadFromOffset(kLoadWord, temp1_, obj_, monitor_offse 1198 const vixl32::Register obj_; member in class:art::arm::LoadReferenceWithBakerReadBarrierAndUpdateFieldSlowPathARMVIXL 1393 const Location obj_; member in class:art::arm::ReadBarrierForHeapReferenceSlowPathARMVIXL [all...] |
H A D | code_generator_mips64.cc | 571 obj_(obj), 633 // update the field in the holder (`*(obj_ + field_offset_)`). 649 GpuRegister base = obj_; 700 const GpuRegister obj_; member in class:art::mips64::ReadBarrierMarkAndUpdateFieldSlowPathMIPS64 701 // The location of the offset of the marked reference field within `obj_`. 721 obj_(obj), 832 parallel_move.AddMove(obj_, 865 size_t obj = static_cast<int>(obj_.AsRegister<GpuRegister>()); 885 const Location obj_; member in class:art::mips64::ReadBarrierForHeapReferenceSlowPathMIPS64
|
H A D | code_generator_x86.cc | 541 obj_(obj), 619 Register base = obj_; 673 const Register obj_; member in class:art::x86::ReadBarrierMarkAndUpdateFieldSlowPathX86 674 // The address of the marked reference field. The base of this address must be `obj_`. 697 obj_(obj), 810 parallel_move.AddMove(obj_, 838 size_t obj = static_cast<int>(obj_.AsRegister<Register>()); 855 const Location obj_; member in class:art::x86::ReadBarrierForHeapReferenceSlowPathX86
|
H A D | code_generator_x86_64.cc | 559 obj_(obj), 642 CpuRegister base = obj_; 697 const CpuRegister obj_; member in class:art::x86_64::ReadBarrierMarkAndUpdateFieldSlowPathX86_64 698 // The address of the marked reference field. The base of this address must be `obj_`. 722 obj_(obj), 832 parallel_move.AddMove(obj_, 865 size_t obj = static_cast<int>(obj_.AsRegister<CpuRegister>().AsRegister()); 882 const Location obj_; member in class:art::x86_64::ReadBarrierForHeapReferenceSlowPathX86_64
|
H A D | code_generator_mips.cc | 615 obj_(obj), 678 // update the field in the holder (`*(obj_ + field_offset_)`). 694 Register base = obj_; 754 const Register obj_; member in class:art::mips::ReadBarrierMarkAndUpdateFieldSlowPathMIPS 755 // The location of the offset of the marked reference field within `obj_`. 775 obj_(obj), 888 parallel_move.AddMove(obj_, 921 size_t obj = static_cast<int>(obj_.AsRegister<Register>()); 941 const Location obj_; member in class:art::mips::ReadBarrierForHeapReferenceSlowPathMIPS
|
/art/dex2oat/ |
H A D | dex2oat.cc | 2999 explicit ScopedGlobalRef(jobject obj) : obj_(obj) {} 3001 if (obj_ != nullptr) { 3003 soa.Env()->GetVm()->DeleteGlobalRef(soa.Self(), obj_); 3008 jobject obj_; member in class:art::ScopedGlobalRef
|
/art/runtime/gc/ |
H A D | heap.cc | 2729 explicit RootMatchesObjectVisitor(const mirror::Object* obj) : obj_(obj) { } 2733 if (root == obj_) { 2734 LOG(INFO) << "Object " << obj_ << " is a root " << info.ToString(); 2739 const mirror::Object* const obj_; member in class:art::gc::RootMatchesObjectVisitor::mirror
|