/* * Copyright (C) 2014 The Android Open Source Project * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ #ifndef ART_RUNTIME_HANDLE_SCOPE_H_ #define ART_RUNTIME_HANDLE_SCOPE_H_ #include #include "base/enums.h" #include "base/logging.h" #include "base/macros.h" #include "base/mutex.h" #include "handle.h" #include "stack_reference.h" #include "verify_object.h" namespace art { class HandleScope; template class ObjPtr; class Thread; class VariableSizedHandleScope; namespace mirror { class Object; } // Basic handle scope, tracked by a list. May be variable sized. class PACKED(4) BaseHandleScope { public: bool IsVariableSized() const { return number_of_references_ == kNumReferencesVariableSized; } // Number of references contained within this handle scope. ALWAYS_INLINE uint32_t NumberOfReferences() const; ALWAYS_INLINE bool Contains(StackReference* handle_scope_entry) const; template ALWAYS_INLINE void VisitRoots(Visitor& visitor) REQUIRES_SHARED(Locks::mutator_lock_); // Link to previous BaseHandleScope or null. BaseHandleScope* GetLink() const { return link_; } ALWAYS_INLINE VariableSizedHandleScope* AsVariableSized(); ALWAYS_INLINE HandleScope* AsHandleScope(); ALWAYS_INLINE const VariableSizedHandleScope* AsVariableSized() const; ALWAYS_INLINE const HandleScope* AsHandleScope() const; protected: BaseHandleScope(BaseHandleScope* link, uint32_t num_references) : link_(link), number_of_references_(num_references) {} // Variable sized constructor. explicit BaseHandleScope(BaseHandleScope* link) : link_(link), number_of_references_(kNumReferencesVariableSized) {} static constexpr int32_t kNumReferencesVariableSized = -1; // Link-list of handle scopes. The root is held by a Thread. BaseHandleScope* const link_; // Number of handlerized references. -1 for variable sized handle scopes. const int32_t number_of_references_; private: DISALLOW_COPY_AND_ASSIGN(BaseHandleScope); }; // HandleScopes are scoped objects containing a number of Handles. They are used to allocate // handles, for these handles (and the objects contained within them) to be visible/roots for the // GC. It is most common to stack allocate HandleScopes using StackHandleScope. class PACKED(4) HandleScope : public BaseHandleScope { public: ~HandleScope() {} // We have versions with and without explicit pointer size of the following. The first two are // used at runtime, so OFFSETOF_MEMBER computes the right offsets automatically. The last one // takes the pointer size explicitly so that at compile time we can cross-compile correctly. // Returns the size of a HandleScope containing num_references handles. static size_t SizeOf(uint32_t num_references); // Returns the size of a HandleScope containing num_references handles. static size_t SizeOf(PointerSize pointer_size, uint32_t num_references); ALWAYS_INLINE mirror::Object* GetReference(size_t i) const REQUIRES_SHARED(Locks::mutator_lock_); ALWAYS_INLINE Handle GetHandle(size_t i); ALWAYS_INLINE MutableHandle GetMutableHandle(size_t i) REQUIRES_SHARED(Locks::mutator_lock_); ALWAYS_INLINE void SetReference(size_t i, mirror::Object* object) REQUIRES_SHARED(Locks::mutator_lock_); ALWAYS_INLINE bool Contains(StackReference* handle_scope_entry) const; // Offset of link within HandleScope, used by generated code. static constexpr size_t LinkOffset(PointerSize pointer_size ATTRIBUTE_UNUSED) { return 0; } // Offset of length within handle scope, used by generated code. static constexpr size_t NumberOfReferencesOffset(PointerSize pointer_size) { return static_cast(pointer_size); } // Offset of link within handle scope, used by generated code. static constexpr size_t ReferencesOffset(PointerSize pointer_size) { return NumberOfReferencesOffset(pointer_size) + sizeof(number_of_references_); } // Placement new creation. static HandleScope* Create(void* storage, BaseHandleScope* link, uint32_t num_references) WARN_UNUSED { return new (storage) HandleScope(link, num_references); } // Number of references contained within this handle scope. ALWAYS_INLINE uint32_t NumberOfReferences() const { DCHECK_GE(number_of_references_, 0); return static_cast(number_of_references_); } template void VisitRoots(Visitor& visitor) REQUIRES_SHARED(Locks::mutator_lock_) { for (size_t i = 0, count = NumberOfReferences(); i < count; ++i) { // GetReference returns a pointer to the stack reference within the handle scope. If this // needs to be updated, it will be done by the root visitor. visitor.VisitRootIfNonNull(GetHandle(i).GetReference()); } } protected: // Return backing storage used for references. ALWAYS_INLINE StackReference* GetReferences() const { uintptr_t address = reinterpret_cast(this) + ReferencesOffset(kRuntimePointerSize); return reinterpret_cast*>(address); } explicit HandleScope(size_t number_of_references) : HandleScope(nullptr, number_of_references) {} // Semi-hidden constructor. Construction expected by generated code and StackHandleScope. HandleScope(BaseHandleScope* link, uint32_t num_references) : BaseHandleScope(link, num_references) {} // Storage for references. // StackReference references_[number_of_references_] private: DISALLOW_COPY_AND_ASSIGN(HandleScope); }; // A wrapper which wraps around Object** and restores the pointer in the destructor. // TODO: Delete template class HandleWrapper : public MutableHandle { public: HandleWrapper(T** obj, const MutableHandle& handle) : MutableHandle(handle), obj_(obj) { } HandleWrapper(const HandleWrapper&) = default; ~HandleWrapper() { *obj_ = MutableHandle::Get(); } private: T** const obj_; }; // A wrapper which wraps around ObjPtr* and restores the pointer in the destructor. // TODO: Add more functionality. template class HandleWrapperObjPtr : public MutableHandle { public: HandleWrapperObjPtr(ObjPtr* obj, const MutableHandle& handle) : MutableHandle(handle), obj_(obj) {} HandleWrapperObjPtr(const HandleWrapperObjPtr&) = default; ~HandleWrapperObjPtr() { *obj_ = ObjPtr(MutableHandle::Get()); } private: ObjPtr* const obj_; }; // Fixed size handle scope that is not necessarily linked in the thread. template class PACKED(4) FixedSizeHandleScope : public HandleScope { public: template ALWAYS_INLINE MutableHandle NewHandle(T* object) REQUIRES_SHARED(Locks::mutator_lock_); template ALWAYS_INLINE HandleWrapper NewHandleWrapper(T** object) REQUIRES_SHARED(Locks::mutator_lock_); template ALWAYS_INLINE HandleWrapperObjPtr NewHandleWrapper(ObjPtr* object) REQUIRES_SHARED(Locks::mutator_lock_); template ALWAYS_INLINE MutableHandle NewHandle(ObjPtr object) REQUIRES_SHARED(Locks::mutator_lock_); ALWAYS_INLINE void SetReference(size_t i, mirror::Object* object) REQUIRES_SHARED(Locks::mutator_lock_); size_t RemainingSlots() const { return kNumReferences - pos_; } private: explicit ALWAYS_INLINE FixedSizeHandleScope(BaseHandleScope* link, mirror::Object* fill_value = nullptr); ALWAYS_INLINE ~FixedSizeHandleScope() {} template ALWAYS_INLINE MutableHandle GetHandle(size_t i) REQUIRES_SHARED(Locks::mutator_lock_) { DCHECK_LT(i, kNumReferences); return MutableHandle(&GetReferences()[i]); } // Reference storage needs to be first as expected by the HandleScope layout. StackReference storage_[kNumReferences]; // Position new handles will be created. uint32_t pos_ = 0; template friend class StackHandleScope; friend class VariableSizedHandleScope; }; // Scoped handle storage of a fixed size that is stack allocated. template class PACKED(4) StackHandleScope FINAL : public FixedSizeHandleScope { public: explicit ALWAYS_INLINE StackHandleScope(Thread* self, mirror::Object* fill_value = nullptr); ALWAYS_INLINE ~StackHandleScope(); Thread* Self() const { return self_; } private: // The thread that the stack handle scope is a linked list upon. The stack handle scope will // push and pop itself from this thread. Thread* const self_; }; // Utility class to manage a variable sized handle scope by having a list of fixed size handle // scopes. // Calls to NewHandle will create a new handle inside the current FixedSizeHandleScope. // When the current handle scope becomes full a new one is created and put at the front of the // list. class VariableSizedHandleScope : public BaseHandleScope { public: explicit VariableSizedHandleScope(Thread* const self); ~VariableSizedHandleScope(); template MutableHandle NewHandle(T* object) REQUIRES_SHARED(Locks::mutator_lock_); template MutableHandle NewHandle(ObjPtr ptr) REQUIRES_SHARED(Locks::mutator_lock_); // Number of references contained within this handle scope. ALWAYS_INLINE uint32_t NumberOfReferences() const; ALWAYS_INLINE bool Contains(StackReference* handle_scope_entry) const; template void VisitRoots(Visitor& visitor) REQUIRES_SHARED(Locks::mutator_lock_); private: static constexpr size_t kLocalScopeSize = 64u; static constexpr size_t kSizeOfReferencesPerScope = kLocalScopeSize - /* BaseHandleScope::link_ */ sizeof(BaseHandleScope*) - /* BaseHandleScope::number_of_references_ */ sizeof(int32_t) - /* FixedSizeHandleScope<>::pos_ */ sizeof(uint32_t); static constexpr size_t kNumReferencesPerScope = kSizeOfReferencesPerScope / sizeof(StackReference); Thread* const self_; // Linked list of fixed size handle scopes. using LocalScopeType = FixedSizeHandleScope; static_assert(sizeof(LocalScopeType) == kLocalScopeSize, "Unexpected size of LocalScopeType"); LocalScopeType* current_scope_; DISALLOW_COPY_AND_ASSIGN(VariableSizedHandleScope); }; } // namespace art #endif // ART_RUNTIME_HANDLE_SCOPE_H_