/art/runtime/jit/ |
D | jit_code_cache.h | 67 size_t CodeCacheSize() REQUIRES(!lock_); 70 size_t DataCacheSize() REQUIRES(!lock_); 74 REQUIRES(!lock_); 82 REQUIRES(!lock_); 86 REQUIRES(!lock_); 90 REQUIRES(!lock_); 103 REQUIRES(!lock_); 109 bool ContainsMethod(ArtMethod* method) REQUIRES(!lock_); 114 REQUIRES(!lock_); 119 REQUIRES(!lock_); [all …]
|
D | jit_code_cache.cc | 123 : lock_("Jit code cache", kJitCodeCacheLock), in JitCodeCache() 124 lock_cond_("Jit code cache variable", lock_), in JitCodeCache() 169 MutexLock mu(Thread::Current(), lock_); in ContainsMethod() 263 MutexLock mu(self, lock_); in RemoveMethodsIn() 299 MutexLock mu(self, lock_); in ClearGcRootsInInlineCaches() 326 MutexLock mu(self, lock_); in CommitCodeInternal() 352 MutexLock mu(self, lock_); in CommitCodeInternal() 387 MutexLock mu(Thread::Current(), lock_); in CodeCacheSize() 396 MutexLock mu(Thread::Current(), lock_); in DataCacheSize() 405 MutexLock mu(self, lock_); in ClearData() [all …]
|
/art/runtime/ |
D | class_table.h | 74 REQUIRES(!lock_) 80 REQUIRES(!lock_) 84 size_t NumZygoteClasses() const REQUIRES(!lock_); 87 size_t NumNonZygoteClasses() const REQUIRES(!lock_); 91 REQUIRES(!lock_) 98 REQUIRES(!lock_) 104 REQUIRES(!lock_) 110 REQUIRES(!lock_) 115 REQUIRES(!lock_) 120 REQUIRES(!lock_) [all …]
|
D | barrier.h | 42 void Pass(Thread* self) REQUIRES(!lock_); 45 void Wait(Thread* self) REQUIRES(!lock_); 54 void Increment(Thread* self, int delta) REQUIRES(!lock_); 58 bool Increment(Thread* self, int delta, uint32_t timeout_ms) REQUIRES(!lock_); 62 void Init(Thread* self, int count) REQUIRES(!lock_); 65 void SetCountLocked(Thread* self, int count) REQUIRES(lock_); 68 int count_ GUARDED_BY(lock_); 70 Mutex lock_ ACQUIRED_AFTER(Locks::abort_lock_); 71 ConditionVariable condition_ GUARDED_BY(lock_);
|
D | signal_catcher.h | 48 void SetHaltFlag(bool new_value) REQUIRES(!lock_); 49 bool ShouldHalt() REQUIRES(!lock_); 50 int WaitForSignal(Thread* self, SignalSet& signals) REQUIRES(!lock_); 54 mutable Mutex lock_ DEFAULT_MUTEX_ACQUIRED_AFTER; 55 ConditionVariable cond_ GUARDED_BY(lock_); 56 bool halt_ GUARDED_BY(lock_); 57 pthread_t pthread_ GUARDED_BY(lock_); 58 Thread* thread_ GUARDED_BY(lock_);
|
D | linear_alloc.h | 31 void* Alloc(Thread* self, size_t size) REQUIRES(!lock_); 34 void* Realloc(Thread* self, void* ptr, size_t old_size, size_t new_size) REQUIRES(!lock_); 38 T* AllocArray(Thread* self, size_t elements) REQUIRES(!lock_) { in AllocArray() 43 size_t GetUsedMemory() const REQUIRES(!lock_); 45 ArenaPool* GetArenaPool() REQUIRES(!lock_); 48 bool Contains(void* ptr) const REQUIRES(!lock_); 55 mutable Mutex lock_ DEFAULT_MUTEX_ACQUIRED_AFTER; 56 ArenaAllocator allocator_ GUARDED_BY(lock_);
|
D | linear_alloc.cc | 23 LinearAlloc::LinearAlloc(ArenaPool* pool) : lock_("linear alloc"), allocator_(pool) { in LinearAlloc() 27 MutexLock mu(self, lock_); in Realloc() 32 MutexLock mu(self, lock_); in Alloc() 37 MutexLock mu(Thread::Current(), lock_); in GetUsedMemory() 42 MutexLock mu(Thread::Current(), lock_); in GetArenaPool() 47 MutexLock mu(Thread::Current(), lock_); in Contains()
|
D | class_table.cc | 23 ClassTable::ClassTable() : lock_("Class loader classes", kClassLoaderClassesLock) { in ClassTable() 30 WriterMutexLock mu(Thread::Current(), lock_); in FreezeSnapshot() 35 ReaderMutexLock mu(Thread::Current(), lock_); in Contains() 46 ReaderMutexLock mu(Thread::Current(), lock_); in LookupByDescriptor() 57 WriterMutexLock mu(Thread::Current(), lock_); in UpdateClass() 81 ReaderMutexLock mu(Thread::Current(), lock_); in NumZygoteClasses() 90 ReaderMutexLock mu(Thread::Current(), lock_); in NumNonZygoteClasses() 95 ReaderMutexLock mu(Thread::Current(), lock_); in Lookup() 106 WriterMutexLock mu(Thread::Current(), lock_); in Insert() 111 WriterMutexLock mu(Thread::Current(), lock_); in InsertWithHash() [all …]
|
D | barrier.cc | 28 lock_("GC barrier lock", kThreadSuspendCountLock), in Barrier() 29 condition_("GC barrier condition", lock_) { in Barrier() 33 MutexLock mu(self, lock_); in Pass() 42 MutexLock mu(self, lock_); in Init() 47 MutexLock mu(self, lock_); in Increment() 63 MutexLock mu(self, lock_); in Increment()
|
D | signal_catcher.cc | 69 lock_("SignalCatcher lock"), in SignalCatcher() 70 cond_("SignalCatcher::cond_", lock_), in SignalCatcher() 78 MutexLock mu(self, lock_); in SignalCatcher() 93 MutexLock mu(Thread::Current(), lock_); in SetHaltFlag() 98 MutexLock mu(Thread::Current(), lock_); in ShouldHalt() 192 MutexLock mu(self, signal_catcher->lock_); in Run()
|
D | class_table-inl.h | 26 ReaderMutexLock mu(Thread::Current(), lock_); in VisitRoots() 39 ReaderMutexLock mu(Thread::Current(), lock_); in VisitRoots() 52 ReaderMutexLock mu(Thread::Current(), lock_); in Visit()
|
/art/runtime/gc/ |
D | task_processor.h | 57 void AddTask(Thread* self, HeapTask* task) REQUIRES(!*lock_); 58 HeapTask* GetTask(Thread* self) REQUIRES(!*lock_); 59 void Start(Thread* self) REQUIRES(!*lock_); 62 void Stop(Thread* self) REQUIRES(!*lock_); 63 void RunAllTasks(Thread* self) REQUIRES(!*lock_); 64 bool IsRunning() const REQUIRES(!*lock_); 66 REQUIRES(!*lock_); 67 Thread* GetRunningThread() const REQUIRES(!*lock_); 77 mutable Mutex* lock_ DEFAULT_MUTEX_ACQUIRED_AFTER; 78 bool is_running_ GUARDED_BY(lock_); [all …]
|
D | task_processor.cc | 26 : lock_(new Mutex("Task processor lock", kReferenceProcessorLock)), is_running_(false), in TaskProcessor() 29 cond_.reset(new ConditionVariable("Task processor condition", *lock_)); in TaskProcessor() 33 delete lock_; in ~TaskProcessor() 38 MutexLock mu(self, *lock_); in AddTask() 45 MutexLock mu(self, *lock_); in GetTask() 75 MutexLock mu(self, *lock_); in UpdateTargetRunTime() 97 MutexLock mu(Thread::Current(), *lock_); in IsRunning() 102 MutexLock mu(Thread::Current(), *lock_); in GetRunningThread() 107 MutexLock mu(self, *lock_); in Stop() 114 MutexLock mu(self, *lock_); in Start()
|
/art/runtime/jdwp/ |
D | object_registry.h | 67 REQUIRES(!Locks::thread_list_lock_, !Locks::thread_suspend_count_lock_, !lock_); 71 REQUIRES(!Locks::thread_list_lock_, !Locks::thread_suspend_count_lock_, !lock_); 76 REQUIRES(!Locks::thread_list_lock_, !Locks::thread_suspend_count_lock_, !lock_); 80 REQUIRES(!Locks::thread_list_lock_, !Locks::thread_suspend_count_lock_, !lock_); 83 SHARED_REQUIRES(Locks::mutator_lock_) REQUIRES(!lock_) { in Get() 91 void Clear() SHARED_REQUIRES(Locks::mutator_lock_) REQUIRES(!lock_); 94 SHARED_REQUIRES(Locks::mutator_lock_) REQUIRES(!lock_); 97 SHARED_REQUIRES(Locks::mutator_lock_) REQUIRES(!lock_); 100 SHARED_REQUIRES(Locks::mutator_lock_) REQUIRES(!lock_); 103 SHARED_REQUIRES(Locks::mutator_lock_) REQUIRES(!lock_); [all …]
|
D | object_registry.cc | 35 : lock_("ObjectRegistry lock", kJdwpObjectRegistryLock), next_id_(1) { in ObjectRegistry() 90 MutexLock mu(soa.Self(), lock_); in InternalAdd() 152 MutexLock mu(self, lock_); in Clear() 172 MutexLock mu(self, lock_); in InternalGet() 188 MutexLock mu(self, lock_); in GetJObject() 197 MutexLock mu(self, lock_); in DisableCollection() 205 MutexLock mu(self, lock_); in EnableCollection() 235 MutexLock mu(self, lock_); in IsCollected() 249 MutexLock mu(self, lock_); in DisposeObject()
|
/art/runtime/gc/space/ |
D | large_object_space.h | 133 size_t AllocationSize(mirror::Object* obj, size_t* usable_size) REQUIRES(!lock_); 136 REQUIRES(!lock_); 137 size_t Free(Thread* self, mirror::Object* ptr) REQUIRES(!lock_); 138 void Walk(DlMallocSpace::WalkCallback, void* arg) OVERRIDE REQUIRES(!lock_); 150 bool IsZygoteLargeObject(Thread* self, mirror::Object* obj) const OVERRIDE REQUIRES(!lock_); 151 void SetAllLargeObjectsAsZygoteObjects(Thread* self) OVERRIDE REQUIRES(!lock_); 154 mutable Mutex lock_ DEFAULT_MUTEX_ACQUIRED_AFTER; 156 GUARDED_BY(lock_); 167 REQUIRES(lock_); 170 OVERRIDE REQUIRES(!lock_); [all …]
|
D | dlmalloc_space.h | 53 OVERRIDE REQUIRES(!lock_); 57 OVERRIDE REQUIRES(!lock_) { in Alloc() 67 REQUIRES(!lock_) 71 REQUIRES(!lock_) 89 REQUIRES(!lock_); 107 void Walk(WalkCallback callback, void* arg) OVERRIDE REQUIRES(!lock_); 150 REQUIRES(lock_);
|
D | dlmalloc_space.cc | 133 MutexLock mu(self, lock_); in AllocWithGrowth() 168 MutexLock mu(self, lock_); in Free() 197 MutexLock mu(self, lock_); in FreeList() 218 MutexLock mu(self, lock_); in FreeList() 225 MutexLock mu(Thread::Current(), lock_); in Trim() 236 MutexLock mu(Thread::Current(), lock_); in Walk() 242 MutexLock mu(Thread::Current(), lock_); in GetFootprint() 247 MutexLock mu(Thread::Current(), lock_); in GetFootprintLimit() 252 MutexLock mu(Thread::Current(), lock_); in SetFootprintLimit() 265 MutexLock mu(Thread::Current(), lock_); in GetBytesAllocated() [all …]
|
/art/runtime/base/ |
D | timing_logger.h | 36 void End() REQUIRES(!lock_); 37 void Reset() REQUIRES(!lock_); 38 void Dump(std::ostream& os) const REQUIRES(!lock_); 44 void SetName(const std::string& name) REQUIRES(!lock_); 45 void AddLogger(const TimingLogger& logger) REQUIRES(!lock_); 46 size_t GetIterations() const REQUIRES(!lock_); 61 REQUIRES(lock_); 62 void DumpHistogram(std::ostream &os) const REQUIRES(lock_); 67 std::set<Histogram<uint64_t>*, HistogramComparator> histograms_ GUARDED_BY(lock_); 70 mutable Mutex lock_ DEFAULT_MUTEX_ACQUIRED_AFTER; [all …]
|
D | timing_logger.cc | 41 lock_(lock_name_.c_str(), kDefaultMutexLevel, true) { in CumulativeLogger() 50 MutexLock mu(Thread::Current(), lock_); in SetName() 58 MutexLock mu(Thread::Current(), lock_); in End() 63 MutexLock mu(Thread::Current(), lock_); in Reset() 70 MutexLock mu(Thread::Current(), lock_); in AddLogger() 82 MutexLock mu(Thread::Current(), lock_); in GetIterations() 87 MutexLock mu(Thread::Current(), lock_); in Dump()
|
/art/compiler/dex/quick/ |
D | dex_file_to_method_inliner_map.cc | 31 : lock_("DexFileToMethodInlinerMap lock", kDexFileToMethodInlinerMapLock) { in DexFileToMethodInlinerMap() 43 ReaderMutexLock mu(self, lock_); in GetMethodInliner() 58 WriterMutexLock mu(self, lock_); in GetMethodInliner() 66 locked_inliner->lock_.ExclusiveLock(self); // Acquire inliner's lock_ before releasing lock_. in GetMethodInliner() 69 locked_inliner->lock_.ExclusiveUnlock(self); in GetMethodInliner()
|
D | dex_file_method_inliner.h | 59 SHARED_REQUIRES(Locks::mutator_lock_) REQUIRES(!lock_); 64 InlineMethodFlags IsIntrinsicOrSpecial(uint32_t method_index) REQUIRES(!lock_); 69 bool IsIntrinsic(uint32_t method_index, InlineMethod* intrinsic) REQUIRES(!lock_); 74 bool IsSpecial(uint32_t method_index) REQUIRES(!lock_); 80 REQUIRES(!lock_); 85 bool IsStringInitMethodIndex(uint32_t method_index) REQUIRES(!lock_); 380 void FindIntrinsics(const DexFile* dex_file) REQUIRES(lock_); 384 bool AddInlineMethod(int32_t method_idx, const InlineMethod& method) REQUIRES(!lock_); 386 ReaderWriterMutex lock_; variable 390 SafeMap<uint32_t, InlineMethod> inline_methods_ GUARDED_BY(lock_);
|
D | dex_file_to_method_inliner_map.h | 48 ReaderWriterMutex lock_; 49 std::map<const DexFile*, DexFileMethodInliner*> inliners_ GUARDED_BY(lock_);
|
/art/runtime/gc/allocator/ |
D | rosalloc.h | 54 size_t ByteSize(RosAlloc* rosalloc) const REQUIRES(rosalloc->lock_) { in ByteSize() 63 REQUIRES(rosalloc->lock_) { in SetByteSize() 72 void* End(RosAlloc* rosalloc) REQUIRES(rosalloc->lock_) { in End() 78 REQUIRES(rosalloc->lock_) { in IsLargerThanPageReleaseThreshold() 82 REQUIRES(rosalloc->lock_) { in IsAtEndOfSpace() 85 bool ShouldReleasePages(RosAlloc* rosalloc) REQUIRES(rosalloc->lock_) { in ShouldReleasePages() 102 void ReleasePages(RosAlloc* rosalloc) REQUIRES(rosalloc->lock_) { in ReleasePages() 718 AllocationTrackingSet<FreePageRun*, kAllocatorTagRosAlloc> free_page_runs_ GUARDED_BY(lock_); 751 GUARDED_BY(lock_); 754 Mutex lock_ DEFAULT_MUTEX_ACQUIRED_AFTER; [all …]
|
/art/compiler/utils/ |
D | swap_space.h | 38 void* Alloc(size_t size) REQUIRES(!lock_); 39 void Free(void* ptr, size_t size) REQUIRES(!lock_); 81 SpaceChunk NewFileChunk(size_t min_size) REQUIRES(lock_); 83 void RemoveChunk(FreeBySizeSet::const_iterator free_by_size_pos) REQUIRES(lock_); 84 void InsertChunk(const SpaceChunk& chunk) REQUIRES(lock_); 92 FreeByStartSet free_by_start_ GUARDED_BY(lock_); 94 FreeBySizeSet free_by_size_ GUARDED_BY(lock_); 96 mutable Mutex lock_ DEFAULT_MUTEX_ACQUIRED_AFTER;
|