1 /* 2 * Copyright (C) 2014 The Android Open Source Project 3 * 4 * Licensed under the Apache License, Version 2.0 (the "License"); 5 * you may not use this file except in compliance with the License. 6 * You may obtain a copy of the License at 7 * 8 * http://www.apache.org/licenses/LICENSE-2.0 9 * 10 * Unless required by applicable law or agreed to in writing, software 11 * distributed under the License is distributed on an "AS IS" BASIS, 12 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. 13 * See the License for the specific language governing permissions and 14 * limitations under the License. 15 */ 16 17 #ifndef ART_RUNTIME_HANDLE_SCOPE_H_ 18 #define ART_RUNTIME_HANDLE_SCOPE_H_ 19 20 #include <stack> 21 22 #include <android-base/logging.h> 23 24 #include "base/enums.h" 25 #include "base/locks.h" 26 #include "base/macros.h" 27 #include "stack_reference.h" 28 29 namespace art { 30 31 template<class T> class Handle; 32 class HandleScope; 33 template<class T> class HandleWrapper; 34 template<class T> class HandleWrapperObjPtr; 35 template<class T> class MutableHandle; 36 template<class MirrorType> class ObjPtr; 37 class Thread; 38 class VariableSizedHandleScope; 39 40 namespace mirror { 41 class Object; 42 } // namespace mirror 43 44 // Basic handle scope, tracked by a list. May be variable sized. 45 class PACKED(4) BaseHandleScope { 46 public: IsVariableSized()47 bool IsVariableSized() const { 48 return number_of_references_ == kNumReferencesVariableSized; 49 } 50 51 // Number of references contained within this handle scope. 52 ALWAYS_INLINE uint32_t NumberOfReferences() const; 53 54 ALWAYS_INLINE bool Contains(StackReference<mirror::Object>* handle_scope_entry) const; 55 56 template <typename Visitor> 57 ALWAYS_INLINE void VisitRoots(Visitor& visitor) REQUIRES_SHARED(Locks::mutator_lock_); 58 59 // Link to previous BaseHandleScope or null. GetLink()60 BaseHandleScope* GetLink() const { 61 return link_; 62 } 63 64 ALWAYS_INLINE VariableSizedHandleScope* AsVariableSized(); 65 ALWAYS_INLINE HandleScope* AsHandleScope(); 66 ALWAYS_INLINE const VariableSizedHandleScope* AsVariableSized() const; 67 ALWAYS_INLINE const HandleScope* AsHandleScope() const; 68 69 protected: BaseHandleScope(BaseHandleScope * link,uint32_t num_references)70 BaseHandleScope(BaseHandleScope* link, uint32_t num_references) 71 : link_(link), 72 number_of_references_(num_references) {} 73 74 // Variable sized constructor. BaseHandleScope(BaseHandleScope * link)75 explicit BaseHandleScope(BaseHandleScope* link) 76 : link_(link), 77 number_of_references_(kNumReferencesVariableSized) {} 78 79 static constexpr int32_t kNumReferencesVariableSized = -1; 80 81 // Link-list of handle scopes. The root is held by a Thread. 82 BaseHandleScope* const link_; 83 84 // Number of handlerized references. -1 for variable sized handle scopes. 85 const int32_t number_of_references_; 86 87 private: 88 DISALLOW_COPY_AND_ASSIGN(BaseHandleScope); 89 }; 90 91 // HandleScopes are scoped objects containing a number of Handles. They are used to allocate 92 // handles, for these handles (and the objects contained within them) to be visible/roots for the 93 // GC. It is most common to stack allocate HandleScopes using StackHandleScope. 94 class PACKED(4) HandleScope : public BaseHandleScope { 95 public: ~HandleScope()96 ~HandleScope() {} 97 98 // We have versions with and without explicit pointer size of the following. The first two are 99 // used at runtime, so OFFSETOF_MEMBER computes the right offsets automatically. The last one 100 // takes the pointer size explicitly so that at compile time we can cross-compile correctly. 101 102 // Returns the size of a HandleScope containing num_references handles. 103 static size_t SizeOf(uint32_t num_references); 104 105 // Returns the size of a HandleScope containing num_references handles. 106 static size_t SizeOf(PointerSize pointer_size, uint32_t num_references); 107 108 ALWAYS_INLINE mirror::Object* GetReference(size_t i) const 109 REQUIRES_SHARED(Locks::mutator_lock_); 110 111 ALWAYS_INLINE Handle<mirror::Object> GetHandle(size_t i); 112 113 ALWAYS_INLINE MutableHandle<mirror::Object> GetMutableHandle(size_t i) 114 REQUIRES_SHARED(Locks::mutator_lock_); 115 116 ALWAYS_INLINE void SetReference(size_t i, mirror::Object* object) 117 REQUIRES_SHARED(Locks::mutator_lock_); 118 119 ALWAYS_INLINE bool Contains(StackReference<mirror::Object>* handle_scope_entry) const; 120 121 // Offset of link within HandleScope, used by generated code. LinkOffset(PointerSize pointer_size ATTRIBUTE_UNUSED)122 static constexpr size_t LinkOffset(PointerSize pointer_size ATTRIBUTE_UNUSED) { 123 return 0; 124 } 125 126 // Offset of length within handle scope, used by generated code. NumberOfReferencesOffset(PointerSize pointer_size)127 static constexpr size_t NumberOfReferencesOffset(PointerSize pointer_size) { 128 return static_cast<size_t>(pointer_size); 129 } 130 131 // Offset of link within handle scope, used by generated code. ReferencesOffset(PointerSize pointer_size)132 static constexpr size_t ReferencesOffset(PointerSize pointer_size) { 133 return NumberOfReferencesOffset(pointer_size) + sizeof(number_of_references_); 134 } 135 136 // Placement new creation. Create(void * storage,BaseHandleScope * link,uint32_t num_references)137 static HandleScope* Create(void* storage, BaseHandleScope* link, uint32_t num_references) 138 WARN_UNUSED { 139 return new (storage) HandleScope(link, num_references); 140 } 141 142 // Number of references contained within this handle scope. NumberOfReferences()143 ALWAYS_INLINE uint32_t NumberOfReferences() const { 144 DCHECK_GE(number_of_references_, 0); 145 return static_cast<uint32_t>(number_of_references_); 146 } 147 148 template <typename Visitor> 149 ALWAYS_INLINE void VisitRoots(Visitor& visitor) REQUIRES_SHARED(Locks::mutator_lock_); 150 151 protected: 152 // Return backing storage used for references. GetReferences()153 ALWAYS_INLINE StackReference<mirror::Object>* GetReferences() const { 154 uintptr_t address = reinterpret_cast<uintptr_t>(this) + ReferencesOffset(kRuntimePointerSize); 155 return reinterpret_cast<StackReference<mirror::Object>*>(address); 156 } 157 HandleScope(size_t number_of_references)158 explicit HandleScope(size_t number_of_references) : HandleScope(nullptr, number_of_references) {} 159 160 // Semi-hidden constructor. Construction expected by generated code and StackHandleScope. HandleScope(BaseHandleScope * link,uint32_t num_references)161 HandleScope(BaseHandleScope* link, uint32_t num_references) 162 : BaseHandleScope(link, num_references) {} 163 164 // Storage for references. 165 // StackReference<mirror::Object> references_[number_of_references_] 166 167 private: 168 DISALLOW_COPY_AND_ASSIGN(HandleScope); 169 }; 170 171 // Fixed size handle scope that is not necessarily linked in the thread. 172 template<size_t kNumReferences> 173 class PACKED(4) FixedSizeHandleScope : public HandleScope { 174 public: 175 template<class T> 176 ALWAYS_INLINE MutableHandle<T> NewHandle(T* object) REQUIRES_SHARED(Locks::mutator_lock_); 177 178 template<class T> 179 ALWAYS_INLINE HandleWrapper<T> NewHandleWrapper(T** object) 180 REQUIRES_SHARED(Locks::mutator_lock_); 181 182 template<class T> 183 ALWAYS_INLINE HandleWrapperObjPtr<T> NewHandleWrapper(ObjPtr<T>* object) 184 REQUIRES_SHARED(Locks::mutator_lock_); 185 186 template<class MirrorType> 187 ALWAYS_INLINE MutableHandle<MirrorType> NewHandle(ObjPtr<MirrorType> object) 188 REQUIRES_SHARED(Locks::mutator_lock_); 189 190 ALWAYS_INLINE void SetReference(size_t i, mirror::Object* object) 191 REQUIRES_SHARED(Locks::mutator_lock_); 192 RemainingSlots()193 size_t RemainingSlots() const { 194 return kNumReferences - pos_; 195 } 196 197 private: 198 explicit ALWAYS_INLINE FixedSizeHandleScope(BaseHandleScope* link, 199 mirror::Object* fill_value = nullptr); ~FixedSizeHandleScope()200 ALWAYS_INLINE ~FixedSizeHandleScope() {} 201 202 template<class T> GetHandle(size_t i)203 ALWAYS_INLINE MutableHandle<T> GetHandle(size_t i) REQUIRES_SHARED(Locks::mutator_lock_) { 204 DCHECK_LT(i, kNumReferences); 205 return MutableHandle<T>(&GetReferences()[i]); 206 } 207 208 // Reference storage needs to be first as expected by the HandleScope layout. 209 StackReference<mirror::Object> storage_[kNumReferences]; 210 211 // Position new handles will be created. 212 uint32_t pos_ = 0; 213 214 template<size_t kNumRefs> friend class StackHandleScope; 215 friend class VariableSizedHandleScope; 216 }; 217 218 // Scoped handle storage of a fixed size that is stack allocated. 219 template<size_t kNumReferences> 220 class PACKED(4) StackHandleScope final : public FixedSizeHandleScope<kNumReferences> { 221 public: 222 explicit ALWAYS_INLINE StackHandleScope(Thread* self, mirror::Object* fill_value = nullptr); 223 ALWAYS_INLINE ~StackHandleScope(); 224 Self()225 Thread* Self() const { 226 return self_; 227 } 228 229 private: 230 // The thread that the stack handle scope is a linked list upon. The stack handle scope will 231 // push and pop itself from this thread. 232 Thread* const self_; 233 }; 234 235 // Utility class to manage a variable sized handle scope by having a list of fixed size handle 236 // scopes. 237 // Calls to NewHandle will create a new handle inside the current FixedSizeHandleScope. 238 // When the current handle scope becomes full a new one is created and put at the front of the 239 // list. 240 class VariableSizedHandleScope : public BaseHandleScope { 241 public: 242 explicit VariableSizedHandleScope(Thread* const self); 243 ~VariableSizedHandleScope(); 244 245 template<class T> 246 MutableHandle<T> NewHandle(T* object) REQUIRES_SHARED(Locks::mutator_lock_); 247 248 template<class MirrorType> 249 MutableHandle<MirrorType> NewHandle(ObjPtr<MirrorType> ptr) 250 REQUIRES_SHARED(Locks::mutator_lock_); 251 252 // Number of references contained within this handle scope. 253 ALWAYS_INLINE uint32_t NumberOfReferences() const; 254 255 ALWAYS_INLINE bool Contains(StackReference<mirror::Object>* handle_scope_entry) const; 256 257 template <typename Visitor> 258 void VisitRoots(Visitor& visitor) REQUIRES_SHARED(Locks::mutator_lock_); 259 260 private: 261 static constexpr size_t kMaxLocalScopeSize = 64u; 262 // In order to have consistent compilation with both 32bit and 64bit dex2oat 263 // binaries we need this to be an actual constant. We picked this because it 264 // will ensure that we use <64bit internal scopes. 265 static constexpr size_t kNumReferencesPerScope = 12u; 266 267 Thread* const self_; 268 269 // Linked list of fixed size handle scopes. 270 using LocalScopeType = FixedSizeHandleScope<kNumReferencesPerScope>; 271 static_assert(sizeof(LocalScopeType) <= kMaxLocalScopeSize, "Unexpected size of LocalScopeType"); 272 LocalScopeType* current_scope_; 273 274 DISALLOW_COPY_AND_ASSIGN(VariableSizedHandleScope); 275 }; 276 277 } // namespace art 278 279 #endif // ART_RUNTIME_HANDLE_SCOPE_H_ 280