/art/runtime/ |
D | object_lock.cc | 25 ObjectLock<T>::ObjectLock(Thread* self, Handle<T> object) : self_(self), obj_(object) { in ObjectLock() 27 obj_->MonitorEnter(self_); in ObjectLock() 32 obj_->MonitorExit(self_); in ~ObjectLock() 37 Monitor::Wait(self_, obj_.Get(), 0, 0, false, kWaiting); in WaitIgnoringInterrupts() 42 obj_->Notify(self_); in Notify() 47 obj_->NotifyAll(self_); in NotifyAll()
|
D | handle_scope.h | 131 : MutableHandle<T>(handle), obj_(obj) { in HandleWrapper() 137 *obj_ = MutableHandle<T>::Get(); in ~HandleWrapper() 141 T** const obj_;
|
D | object_lock.h | 43 Handle<T> const obj_; variable
|
D | monitor.h | 111 return obj_.Read<kReadBarrierOption>(); in GetObject() 282 GcRoot<mirror::Object> obj_; variable
|
D | monitor.cc | 82 obj_(GcRoot<mirror::Object>(obj)), in Monitor() 105 obj_(GcRoot<mirror::Object>(obj)), in Monitor() 215 obj_ = GcRoot<mirror::Object>(object); in SetObject() 786 monitor->obj_ = GcRoot<mirror::Object>(nullptr); in Deflate()
|
/art/compiler/optimizing/ |
D | code_generator_arm64.cc | 580 : SlowPathCodeARM64(instruction), out_(out), obj_(obj) { in ReadBarrierMarkSlowPathARM64() 606 arm64_codegen->MoveLocation(LocationFrom(calling_convention.GetRegisterAt(0)), obj_, type); in EmitNativeCode() 620 const Location obj_; member in art::arm64::ReadBarrierMarkSlowPathARM64 637 obj_(obj), in ReadBarrierForHeapReferenceSlowPathARM64() 748 parallel_move.AddMove(obj_, in EmitNativeCode() 780 size_t obj = static_cast<int>(XRegisterFrom(obj_).code()); in FindAvailableCallerSaveRegister() 797 const Location obj_; member in art::arm64::ReadBarrierForHeapReferenceSlowPathARM64
|
D | code_generator_arm.cc | 412 : SlowPathCode(instruction), out_(out), obj_(obj) { in ReadBarrierMarkSlowPathARM() 438 arm_codegen->Move32(Location::RegisterLocation(calling_convention.GetRegisterAt(0)), obj_); in EmitNativeCode() 452 const Location obj_; member in art::arm::ReadBarrierMarkSlowPathARM 469 obj_(obj), in ReadBarrierForHeapReferenceSlowPathARM() 575 parallel_move.AddMove(obj_, in EmitNativeCode() 606 size_t obj = static_cast<int>(obj_.AsRegister<Register>()); in FindAvailableCallerSaveRegister() 623 const Location obj_; member in art::arm::ReadBarrierForHeapReferenceSlowPathARM
|
D | code_generator_x86_64.cc | 450 : SlowPathCode(instruction), out_(out), obj_(obj) { in ReadBarrierMarkSlowPathX86_64() 476 x86_64_codegen->Move(Location::RegisterLocation(calling_convention.GetRegisterAt(0)), obj_); in EmitNativeCode() 490 const Location obj_; member in art::x86_64::ReadBarrierMarkSlowPathX86_64 507 obj_(obj), in ReadBarrierForHeapReferenceSlowPathX86_64() 610 parallel_move.AddMove(obj_, in EmitNativeCode() 643 size_t obj = static_cast<int>(obj_.AsRegister<CpuRegister>().AsRegister()); in FindAvailableCallerSaveRegister() 660 const Location obj_; member in art::x86_64::ReadBarrierForHeapReferenceSlowPathX86_64
|
D | code_generator_x86.cc | 429 : SlowPathCode(instruction), out_(out), obj_(obj) { in ReadBarrierMarkSlowPathX86() 455 x86_codegen->Move32(Location::RegisterLocation(calling_convention.GetRegisterAt(0)), obj_); in EmitNativeCode() 469 const Location obj_; member in art::x86::ReadBarrierMarkSlowPathX86 486 obj_(obj), in ReadBarrierForHeapReferenceSlowPathX86() 592 parallel_move.AddMove(obj_, in EmitNativeCode() 623 size_t obj = static_cast<int>(obj_.AsRegister<Register>()); in FindAvailableCallerSaveRegister() 640 const Location obj_; member in art::x86::ReadBarrierForHeapReferenceSlowPathX86
|
/art/runtime/gc/ |
D | heap.cc | 2831 explicit RootMatchesObjectVisitor(const mirror::Object* obj) : obj_(obj) { } in RootMatchesObjectVisitor() 2835 if (root == obj_) { in VisitRoot() 2836 LOG(INFO) << "Object " << obj_ << " is a root " << info.ToString(); in VisitRoot() 2841 const mirror::Object* const obj_; member in art::gc::RootMatchesObjectVisitor
|