/* * Copyright (C) 2011 The Android Open Source Project * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ #ifndef ART_RUNTIME_MIRROR_OBJECT_READBARRIER_INL_H_ #define ART_RUNTIME_MIRROR_OBJECT_READBARRIER_INL_H_ #include "object.h" #include "atomic.h" #include "lock_word-inl.h" #include "object_reference-inl.h" #include "read_barrier.h" #include "runtime.h" namespace art { namespace mirror { template inline LockWord Object::GetLockWord(bool as_volatile) { if (as_volatile) { return LockWord(GetField32Volatile(OFFSET_OF_OBJECT_MEMBER(Object, monitor_))); } return LockWord(GetField32(OFFSET_OF_OBJECT_MEMBER(Object, monitor_))); } template inline bool Object::CasFieldWeakRelaxed32(MemberOffset field_offset, int32_t old_value, int32_t new_value) { if (kCheckTransaction) { DCHECK_EQ(kTransactionActive, Runtime::Current()->IsActiveTransaction()); } if (kTransactionActive) { Runtime::Current()->RecordWriteField32(this, field_offset, old_value, true); } if (kVerifyFlags & kVerifyThis) { VerifyObject(this); } uint8_t* raw_addr = reinterpret_cast(this) + field_offset.Int32Value(); AtomicInteger* atomic_addr = reinterpret_cast(raw_addr); return atomic_addr->CompareExchangeWeakRelaxed(old_value, new_value); } inline bool Object::CasLockWordWeakRelaxed(LockWord old_val, LockWord new_val) { // Force use of non-transactional mode and do not check. return CasFieldWeakRelaxed32( OFFSET_OF_OBJECT_MEMBER(Object, monitor_), old_val.GetValue(), new_val.GetValue()); } inline bool Object::CasLockWordWeakRelease(LockWord old_val, LockWord new_val) { // Force use of non-transactional mode and do not check. return CasFieldWeakRelease32( OFFSET_OF_OBJECT_MEMBER(Object, monitor_), old_val.GetValue(), new_val.GetValue()); } inline uint32_t Object::GetReadBarrierState(uintptr_t* fake_address_dependency) { if (!kUseBakerReadBarrier) { LOG(FATAL) << "Unreachable"; UNREACHABLE(); } #if defined(__arm__) uintptr_t obj = reinterpret_cast(this); uintptr_t result; DCHECK_EQ(OFFSETOF_MEMBER(Object, monitor_), 4U); // Use inline assembly to prevent the compiler from optimizing away the false dependency. __asm__ __volatile__( "ldr %[result], [%[obj], #4]\n\t" // This instruction is enough to "fool the compiler and the CPU" by having `fad` always be // null, without them being able to assume that fact. "eor %[fad], %[result], %[result]\n\t" : [result] "+r" (result), [fad] "=r" (*fake_address_dependency) : [obj] "r" (obj)); DCHECK_EQ(*fake_address_dependency, 0U); LockWord lw(static_cast(result)); uint32_t rb_state = lw.ReadBarrierState(); return rb_state; #elif defined(__aarch64__) uintptr_t obj = reinterpret_cast(this); uintptr_t result; DCHECK_EQ(OFFSETOF_MEMBER(Object, monitor_), 4U); // Use inline assembly to prevent the compiler from optimizing away the false dependency. __asm__ __volatile__( "ldr %w[result], [%[obj], #4]\n\t" // This instruction is enough to "fool the compiler and the CPU" by having `fad` always be // null, without them being able to assume that fact. "eor %[fad], %[result], %[result]\n\t" : [result] "+r" (result), [fad] "=r" (*fake_address_dependency) : [obj] "r" (obj)); DCHECK_EQ(*fake_address_dependency, 0U); LockWord lw(static_cast(result)); uint32_t rb_state = lw.ReadBarrierState(); return rb_state; #elif defined(__i386__) || defined(__x86_64__) LockWord lw = GetLockWord(false); // i386/x86_64 don't need fake address dependency. Use a compiler fence to avoid compiler // reordering. *fake_address_dependency = 0; std::atomic_signal_fence(std::memory_order_acquire); uint32_t rb_state = lw.ReadBarrierState(); return rb_state; #else // MIPS32/MIPS64: use a memory barrier to prevent load-load reordering. LockWord lw = GetLockWord(false); *fake_address_dependency = 0; std::atomic_thread_fence(std::memory_order_acquire); uint32_t rb_state = lw.ReadBarrierState(); return rb_state; #endif } inline uint32_t Object::GetReadBarrierState() { if (!kUseBakerReadBarrier) { LOG(FATAL) << "Unreachable"; UNREACHABLE(); } DCHECK(kUseBakerReadBarrier); LockWord lw(GetField(OFFSET_OF_OBJECT_MEMBER(Object, monitor_))); uint32_t rb_state = lw.ReadBarrierState(); DCHECK(ReadBarrier::IsValidReadBarrierState(rb_state)) << rb_state; return rb_state; } inline uint32_t Object::GetReadBarrierStateAcquire() { if (!kUseBakerReadBarrier) { LOG(FATAL) << "Unreachable"; UNREACHABLE(); } LockWord lw(GetFieldAcquire(OFFSET_OF_OBJECT_MEMBER(Object, monitor_))); uint32_t rb_state = lw.ReadBarrierState(); DCHECK(ReadBarrier::IsValidReadBarrierState(rb_state)) << rb_state; return rb_state; } template inline bool Object::AtomicSetReadBarrierState(uint32_t expected_rb_state, uint32_t rb_state) { if (!kUseBakerReadBarrier) { LOG(FATAL) << "Unreachable"; UNREACHABLE(); } DCHECK(ReadBarrier::IsValidReadBarrierState(expected_rb_state)) << expected_rb_state; DCHECK(ReadBarrier::IsValidReadBarrierState(rb_state)) << rb_state; LockWord expected_lw; LockWord new_lw; do { LockWord lw = GetLockWord(false); if (UNLIKELY(lw.ReadBarrierState() != expected_rb_state)) { // Lost the race. return false; } expected_lw = lw; expected_lw.SetReadBarrierState(expected_rb_state); new_lw = lw; new_lw.SetReadBarrierState(rb_state); // ConcurrentCopying::ProcessMarkStackRef uses this with kCasRelease == true. // If kCasRelease == true, use a CAS release so that when GC updates all the fields of // an object and then changes the object from gray to black, the field updates (stores) will be // visible (won't be reordered after this CAS.) } while (!(kCasRelease ? CasLockWordWeakRelease(expected_lw, new_lw) : CasLockWordWeakRelaxed(expected_lw, new_lw))); return true; } inline bool Object::AtomicSetMarkBit(uint32_t expected_mark_bit, uint32_t mark_bit) { LockWord expected_lw; LockWord new_lw; do { LockWord lw = GetLockWord(false); if (UNLIKELY(lw.MarkBitState() != expected_mark_bit)) { // Lost the race. return false; } expected_lw = lw; new_lw = lw; new_lw.SetMarkBitState(mark_bit); // Since this is only set from the mutator, we can use the non release Cas. } while (!CasLockWordWeakRelaxed(expected_lw, new_lw)); return true; } template inline bool Object::CasFieldStrongRelaxedObjectWithoutWriteBarrier( MemberOffset field_offset, ObjPtr old_value, ObjPtr new_value) { if (kCheckTransaction) { DCHECK_EQ(kTransactionActive, Runtime::Current()->IsActiveTransaction()); } if (kVerifyFlags & kVerifyThis) { VerifyObject(this); } if (kVerifyFlags & kVerifyWrites) { VerifyObject(new_value); } if (kVerifyFlags & kVerifyReads) { VerifyObject(old_value); } if (kTransactionActive) { Runtime::Current()->RecordWriteFieldReference(this, field_offset, old_value, true); } HeapReference old_ref(HeapReference::FromObjPtr(old_value)); HeapReference new_ref(HeapReference::FromObjPtr(new_value)); uint8_t* raw_addr = reinterpret_cast(this) + field_offset.Int32Value(); Atomic* atomic_addr = reinterpret_cast*>(raw_addr); bool success = atomic_addr->CompareExchangeStrongRelaxed(old_ref.reference_, new_ref.reference_); return success; } template inline bool Object::CasFieldStrongReleaseObjectWithoutWriteBarrier( MemberOffset field_offset, ObjPtr old_value, ObjPtr new_value) { if (kCheckTransaction) { DCHECK_EQ(kTransactionActive, Runtime::Current()->IsActiveTransaction()); } if (kVerifyFlags & kVerifyThis) { VerifyObject(this); } if (kVerifyFlags & kVerifyWrites) { VerifyObject(new_value); } if (kVerifyFlags & kVerifyReads) { VerifyObject(old_value); } if (kTransactionActive) { Runtime::Current()->RecordWriteFieldReference(this, field_offset, old_value, true); } HeapReference old_ref(HeapReference::FromObjPtr(old_value)); HeapReference new_ref(HeapReference::FromObjPtr(new_value)); uint8_t* raw_addr = reinterpret_cast(this) + field_offset.Int32Value(); Atomic* atomic_addr = reinterpret_cast*>(raw_addr); bool success = atomic_addr->CompareExchangeStrongRelease(old_ref.reference_, new_ref.reference_); return success; } } // namespace mirror } // namespace art #endif // ART_RUNTIME_MIRROR_OBJECT_READBARRIER_INL_H_