/art/runtime/ |
D | gc_root-inl.h | 31 inline MirrorType* GcRoot<MirrorType>::Read(GcRootSource* gc_root_source) const { in Read() 37 inline GcRoot<MirrorType>::GcRoot(MirrorType* ref) in GcRoot() function 41 inline GcRoot<MirrorType>::GcRoot(ObjPtr<MirrorType> ref) in GcRoot() function 42 : GcRoot(ref.Ptr()) { } in GcRoot()
|
D | intern_table-inl.h | 31 inline std::size_t InternTable::StringHash::operator()(const GcRoot<mirror::String>& root) const { in operator() 40 inline bool InternTable::StringEquals::operator()(const GcRoot<mirror::String>& a, in operator() 41 const GcRoot<mirror::String>& b) const { in operator() 48 inline bool InternTable::StringEquals::operator()(const GcRoot<mirror::String>& a, in operator() 114 for (GcRoot<mirror::String>& string : intern_strings) { in AddInternStrings()
|
D | gc_root.h | 187 class GcRoot { 217 ALWAYS_INLINE GcRoot() {} in GcRoot() function 218 explicit ALWAYS_INLINE GcRoot(MirrorType* ref) 220 explicit ALWAYS_INLINE GcRoot(ObjPtr<MirrorType> ref) 247 ALWAYS_INLINE void VisitRootIfNonNull(GcRoot<MirrorType>& root) in VisitRootIfNonNull() 263 void VisitRoot(GcRoot<MirrorType>& root) REQUIRES_SHARED(Locks::mutator_lock_) { in VisitRoot() 294 ALWAYS_INLINE void VisitRootIfNonNull(GcRoot<MirrorType>& root) const in VisitRootIfNonNull() 310 void VisitRoot(GcRoot<MirrorType>& root) const REQUIRES_SHARED(Locks::mutator_lock_) { in VisitRoot()
|
D | intern_table.h | 77 std::size_t operator()(const GcRoot<mirror::String>& root) const NO_THREAD_SAFETY_ANALYSIS; 88 bool operator()(const GcRoot<mirror::String>& a, const GcRoot<mirror::String>& b) const 92 bool operator()(const GcRoot<mirror::String>& a, const Utf8String& b) const 98 void MakeEmpty(GcRoot<mirror::String>& item) const { in MakeEmpty() 99 item = GcRoot<mirror::String>(); in MakeEmpty() 101 bool IsEmpty(const GcRoot<mirror::String>& item) const { in IsEmpty() 106 using UnorderedSet = HashSet<GcRoot<mirror::String>, 110 TrackingAllocator<GcRoot<mirror::String>, kAllocatorTagInternTable>>; 330 std::vector<GcRoot<mirror::String>> new_strong_intern_roots_
|
D | reference_table.cc | 56 entries_.push_back(GcRoot<mirror::Object>(obj)); in Add() 128 bool operator()(GcRoot<mirror::Object> root1, GcRoot<mirror::Object> root2) const in Dump() 222 GcRoot<mirror::Object>& stack_for_object = it->first; in Dump() 263 for (GcRoot<mirror::Object>& root : entries) { in Dump() 275 GcRoot<mirror::Object> root; in Dump() 288 void Reset(GcRoot<mirror::Object>& _root) { in Dump() 298 for (GcRoot<mirror::Object>& root : sorted_entries) { in Dump() 358 for (GcRoot<mirror::Object>& root : entries_) { in VisitRoots()
|
D | class_table-inl.h | 77 for (GcRoot<mirror::Object>& root : strong_roots_) { in VisitRoots() 81 for (GcRoot<mirror::Object>& root : oat_file->GetBssGcRoots()) { in VisitRoots() 95 for (GcRoot<mirror::Object>& root : strong_roots_) { in VisitRoots() 99 for (GcRoot<mirror::Object>& root : oat_file->GetBssGcRoots()) { in VisitRoots() 140 GcRoot<mirror::Class>(before_ptr).Read<kReadBarrierOption>()); in Read() 153 GcRoot<mirror::Class> root(before_ptr); in VisitRoot()
|
D | reference_table.h | 58 typedef std::vector<GcRoot<mirror::Object>, 59 TrackingAllocator<GcRoot<mirror::Object>, kAllocatorTagReferenceTable>> Table;
|
D | indirect_reference_table.h | 161 GcRoot<mirror::Object>* GetReference() { in GetReference() 166 const GcRoot<mirror::Object>* GetReference() const { in GetReference() 179 GcRoot<mirror::Object> references_[kIRTPrevCount]; 198 GcRoot<mirror::Object>* operator*() REQUIRES_SHARED(Locks::mutator_lock_) { in REQUIRES_SHARED()
|
D | indirect_reference_table-inl.h | 107 references_[serial_] = GcRoot<mirror::Object>(obj); in Add() 112 references_[serial_] = GcRoot<mirror::Object>(obj); in SetReference()
|
D | intern_table.cc | 133 new_strong_intern_roots_.push_back(GcRoot<mirror::String>(s)); in InsertStrong() 281 for (GcRoot<mirror::String>& entry : weak_interns_.tables_.front().set_) { in PromoteWeakToStrong() 312 auto it = table.set_.find(GcRoot<mirror::String>(s)); in Remove() 324 auto it = table.set_.find(GcRoot<mirror::String>(s)); in Find() 351 tables_.back().set_.insert(GcRoot<mirror::String>(s)); in Insert() 378 *it = GcRoot<mirror::String>(new_object->AsString()); in SweepWeaks()
|
D | art_method-inl.h | 70 declaring_class_ = GcRoot<mirror::Class>(new_declaring_class); in SetDeclaringClass() 75 GcRoot<mirror::Class> expected_root(expected_class); in CASDeclaringClass() 76 GcRoot<mirror::Class> desired_root(desired_class); in CASDeclaringClass() 77 auto atomic_root_class = reinterpret_cast<Atomic<GcRoot<mirror::Class>>*>(&declaring_class_); in CASDeclaringClass()
|
D | class_table_test.cc | 39 ALWAYS_INLINE void VisitRootIfNonNull(GcRoot<MirrorType>& root) const in VisitRootIfNonNull() 55 void VisitRoot(GcRoot<MirrorType>& root) const REQUIRES_SHARED(Locks::mutator_lock_) { in VisitRoot()
|
D | runtime.h | 801 GcRoot<mirror::Object> GetSentinel() REQUIRES_SHARED(Locks::mutator_lock_); 812 static void ProcessWeakClass(GcRoot<mirror::Class>* root_ptr, 1105 GcRoot<mirror::Throwable> pre_allocated_OutOfMemoryError_when_throwing_exception_; 1106 GcRoot<mirror::Throwable> pre_allocated_OutOfMemoryError_when_throwing_oome_; 1107 GcRoot<mirror::Throwable> pre_allocated_OutOfMemoryError_when_handling_stack_overflow_; 1108 GcRoot<mirror::Throwable> pre_allocated_NoClassDefFoundError_; 1117 GcRoot<mirror::Object> sentinel_;
|
D | class_table.cc | 139 for (GcRoot<mirror::Object>& root : strong_roots_) { in InsertStrongRoot() 144 strong_roots_.push_back(GcRoot<mirror::Object>(obj)); in InsertStrongRoot()
|
/art/runtime/entrypoints/quick/ |
D | quick_dexcache_entrypoints.cc | 40 static_assert(sizeof(GcRoot<mirror::Class>) == sizeof(GcRoot<mirror::Object>), "Size check."); in StoreObjectInBss() 41 static_assert(sizeof(GcRoot<mirror::String>) == sizeof(GcRoot<mirror::Object>), "Size check."); in StoreObjectInBss() 43 DCHECK_ALIGNED(bss_offset, sizeof(GcRoot<mirror::Object>)); in StoreObjectInBss() 50 GcRoot<mirror::Object>* slot = reinterpret_cast<GcRoot<mirror::Object>*>( in StoreObjectInBss() 56 std::atomic<GcRoot<mirror::Object>>* atomic_slot = in StoreObjectInBss() 57 reinterpret_cast<std::atomic<GcRoot<mirror::Object>>*>(slot); in StoreObjectInBss() 59 atomic_slot->store(GcRoot<mirror::Object>(object), std::memory_order_release); in StoreObjectInBss() 92 sizeof(GcRoot<mirror::Class>)); in StoreTypeInBss() 119 sizeof(GcRoot<mirror::Class>)); in StoreStringInBss()
|
D | quick_entrypoints.h | 41 template<class MirrorType> class GcRoot; variable 134 extern "C" mirror::Object* artReadBarrierForRootSlow(GcRoot<mirror::Object>* root)
|
/art/runtime/mirror/ |
D | dex_cache-inl.h | 51 first_elem.object = GcRoot<T>(nullptr); in Initialize() 88 GcRoot<mirror::String>* preresolved_strings = GetPreResolvedStrings(); in GetResolvedString() 120 GetPreResolvedStrings()[string_idx.index_] = GcRoot<mirror::String>(resolved); in SetPreResolvedString() 132 GcRoot<mirror::String>*>(PreResolvedStringsOffset(), nullptr); in ClearPreResolvedStrings() 212 GcRoot<mirror::CallSite>& target = GetResolvedCallSites()[call_site_idx]; in GetResolvedCallSite() 213 Atomic<GcRoot<mirror::CallSite>>& ref = in GetResolvedCallSite() 214 reinterpret_cast<Atomic<GcRoot<mirror::CallSite>>&>(target); in GetResolvedCallSite() 223 GcRoot<mirror::CallSite> null_call_site(nullptr); in SetResolvedCallSite() 224 GcRoot<mirror::CallSite> candidate(call_site); in SetResolvedCallSite() 225 GcRoot<mirror::CallSite>& target = GetResolvedCallSites()[call_site_idx]; in SetResolvedCallSite() [all …]
|
D | dex_cache.h | 52 GcRoot<T> object; 212 void FixupResolvedCallSites(GcRoot<mirror::CallSite>* dest, const Visitor& visitor) 274 return alignof(GcRoot<mirror::String>); in PreResolvedStringsAlignment() 335 GcRoot<mirror::String>* GetPreResolvedStrings() ALWAYS_INLINE in GetPreResolvedStrings() 337 return GetFieldPtr64<GcRoot<mirror::String>*, kVerifyFlags>(PreResolvedStringsOffset()); in GetPreResolvedStrings() 344 void SetPreResolvedStrings(GcRoot<mirror::String>* strings) in SetPreResolvedStrings() 393 GcRoot<CallSite>* GetResolvedCallSites() in GetResolvedCallSites() 396 return GetFieldPtr<GcRoot<CallSite>*, kVerifyFlags>(ResolvedCallSitesOffset()); in GetResolvedCallSites() 399 void SetResolvedCallSites(GcRoot<CallSite>* resolved_call_sites) in SetResolvedCallSites() 460 return sizeof(GcRoot<mirror::String>) * num_strings; in PreResolvedStringsSize() [all …]
|
D | dex_cache.cc | 76 GcRoot<mirror::CallSite>* call_sites = in InitializeNativeFields() 77 AllocArray<GcRoot<CallSite>>(self, linear_alloc, num_call_sites); in InitializeNativeFields() 193 GcRoot<mirror::String>* strings = in AddPreResolvedStringsArray() 194 linear_alloc->AllocArray<GcRoot<mirror::String>>(self, num_strings); in AddPreResolvedStringsArray() 220 GcRoot<CallSite>* resolved_call_sites, in SetNativeArrays()
|
/art/runtime/gc/ |
D | system_weak_test.cc | 79 weak_ = GcRoot<mirror::Object>(new_object); in Sweep() 84 GcRoot<mirror::Object> Get() in Get() 94 void Set(GcRoot<mirror::Object> obj) in Set() 107 GcRoot<mirror::Object> weak_ GUARDED_BY(allow_disallow_lock_); 144 cswh.Set(GcRoot<mirror::Object>(s.Get())); in TEST_F() 165 cswh.Set(GcRoot<mirror::Object>(mirror::String::AllocFromModifiedUtf8(soa.Self(), "ABC"))); in TEST_F() 189 cswh.Set(GcRoot<mirror::Object>(s.Get())); in TEST_F()
|
D | allocation_record.cc | 78 GcRoot<mirror::Class>& klass = record->GetClassGcRoot(); in SweepClassObject() 87 klass = GcRoot<mirror::Class>(new_object->AsClass()); in SweepClassObject() 105 it->first = GcRoot<mirror::Object>(nullptr); in SweepAllocationRecords() 114 it->first = GcRoot<mirror::Object>(new_object); in SweepAllocationRecords()
|
D | allocation_record.h | 185 GcRoot<mirror::Class>& GetClassGcRoot() REQUIRES_SHARED(Locks::mutator_lock_) { in GetClassGcRoot() 196 GcRoot<mirror::Class> klass_; 213 using EntryPair = std::pair<GcRoot<mirror::Object>, AllocRecord>; 235 entries_.push_back(EntryPair(GcRoot<mirror::Object>(obj), std::move(record))); in Put()
|
/art/openjdkjvmti/ |
D | jvmti_weak_table.h | 147 auto it = tagged_objects_.find(art::GcRoot<art::mirror::Object>(obj)); in GetTagLocked() 199 size_t operator()(const art::GcRoot<art::mirror::Object>& r) const in operator() 206 bool operator()(const art::GcRoot<art::mirror::Object>& r1, in operator() 207 const art::GcRoot<art::mirror::Object>& r2) const in operator() 213 using TagAllocator = JvmtiAllocator<std::pair<const art::GcRoot<art::mirror::Object>, T>>; 214 std::unordered_map<art::GcRoot<art::mirror::Object>,
|
D | jvmti_weak_table-inl.h | 70 auto WithReadBarrierUpdater = [&](const art::GcRoot<art::mirror::Object>& original_root, in UpdateTableWithReadBarrier() 108 auto it = tagged_objects_.find(art::GcRoot<art::mirror::Object>(obj)); in RemoveLocked() 153 auto it = tagged_objects_.find(art::GcRoot<art::mirror::Object>(obj)); in SetLocked() 173 auto insert_it = tagged_objects_.emplace(art::GcRoot<art::mirror::Object>(obj), new_tag); in SetLocked() 201 auto IsMarkedUpdater = [&](const art::GcRoot<art::mirror::Object>& original_root ATTRIBUTE_UNUSED, in SweepImpl() 232 tagged_objects_.emplace(art::GcRoot<art::mirror::Object>(target_obj), tag); in UpdateTableWith()
|
/art/runtime/jit/ |
D | profiling_info.cc | 92 GcRoot<mirror::Class> expected_root(existing); in AddInvokeInfo() 93 GcRoot<mirror::Class> desired_root(cls); in AddInvokeInfo() 94 auto atomic_root = reinterpret_cast<Atomic<GcRoot<mirror::Class>>*>(&cache->classes_[i]); in AddInvokeInfo()
|