/art/runtime/native/ |
D | dalvik_system_VMDebug.cc | 260 gc::Heap* heap = Runtime::Current()->GetHeap(); in VMDebug_countInstancesOfClass() local 263 heap->CollectGarbage(false); in VMDebug_countInstancesOfClass() 271 heap->CountInstances(classes, countAssignable, &count); in VMDebug_countInstancesOfClass() 292 gc::Heap* heap = Runtime::Current()->GetHeap(); in VMDebug_getHeapSpaceStats() local 293 for (gc::space::ContinuousSpace* space : heap->GetContinuousSpaces()) { in VMDebug_getHeapSpaceStats() 312 for (gc::space::DiscontinuousSpace* space : heap->GetDiscontinuousSpaces()) { in VMDebug_getHeapSpaceStats() 350 gc::Heap* heap = Runtime::Current()->GetHeap(); in VMDebug_getRuntimeStatInternal() local 353 std::string output = std::to_string(heap->GetGcCount()); in VMDebug_getRuntimeStatInternal() 357 std::string output = std::to_string(NsToMs(heap->GetGcTime())); in VMDebug_getRuntimeStatInternal() 361 std::string output = std::to_string(heap->GetBytesAllocatedEver()); in VMDebug_getRuntimeStatInternal() [all …]
|
/art/runtime/gc/ |
D | heap_test.cc | 33 Heap* heap = Runtime::Current()->GetHeap(); in TEST_F() local 34 int64_t max_memory_before = heap->GetMaxMemory(); in TEST_F() 35 int64_t total_memory_before = heap->GetTotalMemory(); in TEST_F() 36 heap->ClearGrowthLimit(); in TEST_F() 37 int64_t max_memory_after = heap->GetMaxMemory(); in TEST_F() 38 int64_t total_memory_after = heap->GetTotalMemory(); in TEST_F()
|
D | reference_queue.cc | 89 Heap* heap = Runtime::Current()->GetHeap(); in DequeuePendingReference() local 90 if (kUseBakerOrBrooksReadBarrier && heap->CurrentCollectorType() == kCollectorTypeCC && in DequeuePendingReference() 91 heap->ConcurrentCopyingCollector()->IsActive()) { in DequeuePendingReference() 97 if (heap->ConcurrentCopyingCollector()->RegionSpace()->IsInToSpace(ref)) { in DequeuePendingReference()
|
/art/runtime/ |
D | read_barrier-inl.h | 61 gc::Heap* heap = Runtime::Current()->GetHeap(); in Barrier() local 62 if (heap != nullptr && heap->GetReadBarrierTable()->IsSet(old_ref)) { in Barrier() 158 gc::Heap* heap = Runtime::Current()->GetHeap(); in IsDuringStartup() local 159 if (heap == nullptr) { in IsDuringStartup() 163 if (heap->CurrentCollectorType() != gc::kCollectorTypeCC) { in IsDuringStartup() 167 gc::collector::ConcurrentCopying* collector = heap->ConcurrentCopyingCollector(); in IsDuringStartup()
|
D | runtime_android.cc | 48 gc::Heap* heap = runtime->GetHeap(); in HandleUnexpectedSignal() local 49 if (kDumpHeapObjectOnSigsevg && heap != nullptr && info != nullptr) { in HandleUnexpectedSignal() 51 heap->DumpObject(LOG(INTERNAL_FATAL), reinterpret_cast<mirror::Object*>(info->si_addr)); in HandleUnexpectedSignal()
|
/art/runtime/mirror/ |
D | object.cc | 85 gc::Heap* heap = Runtime::Current()->GetHeap(); in CopyObject() local 91 heap->WriteBarrierArray(dest, 0, array->GetLength()); in CopyObject() 94 heap->WriteBarrierEveryFieldOf(dest); in CopyObject() 97 heap->AddFinalizerReference(self, &dest); in CopyObject() 125 gc::Heap* heap = Runtime::Current()->GetHeap(); in Clone() local 131 if (heap->IsMovableObject(this)) { in Clone() 132 copy = heap->AllocObject<true>(self, GetClass(), num_bytes, visitor); in Clone() 134 copy = heap->AllocNonMovableObject<true>(self, GetClass(), num_bytes, visitor); in Clone()
|
D | array.cc | 134 auto* heap = Runtime::Current()->GetHeap(); in CopyOf() local 135 gc::AllocatorType allocator_type = heap->IsMovableObject(this) ? heap->GetCurrentAllocator() : in CopyOf() 136 heap->GetCurrentNonMovingAllocator(); in CopyOf()
|
/art/runtime/gc/accounting/ |
D | mod_union_table.h | 60 explicit ModUnionTable(const std::string& name, Heap* heap, space::ContinuousSpace* space) in ModUnionTable() argument 62 heap_(heap), in ModUnionTable() 111 explicit ModUnionTableReferenceCache(const std::string& name, Heap* heap, in ModUnionTableReferenceCache() argument 113 : ModUnionTable(name, heap, space) {} in ModUnionTableReferenceCache() 152 explicit ModUnionTableCardCache(const std::string& name, Heap* heap,
|
D | mod_union_table-inl.h | 31 explicit ModUnionTableToZygoteAllocspace(const std::string& name, Heap* heap, in ModUnionTableToZygoteAllocspace() argument 33 : ModUnionTableReferenceCache(name, heap, space) {} in ModUnionTableToZygoteAllocspace()
|
D | mod_union_table_test.cc | 107 const std::string& name, Heap* heap, space::ContinuousSpace* space, in ModUnionTableRefCacheToSpace() argument 109 : ModUnionTableReferenceCache(name, heap, space), target_space_(target_space) {} in ModUnionTableRefCacheToSpace() 167 gc::Heap* const heap = runtime->GetHeap(); in RunTest() local 169 auto* space = heap->GetNonMovingSpace(); in RunTest() 175 heap->AddSpace(other_space.get()); in RunTest() 242 heap->RemoveSpace(other_space.get()); in RunTest()
|
D | remembered_set.h | 49 explicit RememberedSet(const std::string& name, Heap* heap, space::ContinuousSpace* space) in RememberedSet() argument 50 : name_(name), heap_(heap), space_(space) {} in RememberedSet()
|
D | mod_union_table.cc | 223 Heap* heap = mod_union_table_->GetHeap(); in operator ()() local 224 space::ContinuousSpace* from_space = heap->FindContinuousSpaceFromObject(obj, false); in operator ()() 225 space::ContinuousSpace* to_space = heap->FindContinuousSpaceFromObject(ref, false); in operator ()() 233 heap->DumpSpaces(LOG(INFO)); in operator ()() 345 if (VLOG_IS_ON(heap)) { in UpdateAndMarkReferences() 350 ModUnionTableCardCache::ModUnionTableCardCache(const std::string& name, Heap* heap, in ModUnionTableCardCache() argument 352 : ModUnionTable(name, heap, space) { in ModUnionTableCardCache()
|
D | heap_bitmap.h | 67 explicit HeapBitmap(Heap* heap) : heap_(heap) {} in HeapBitmap() argument
|
/art/test/104-growth-limit/ |
D | info.txt | 2 conventions, can be cleared and the resulting heap is at least as large 3 as the growth limited heap.
|
/art/runtime/gc/space/ |
D | malloc_space.cc | 134 VLOG(heap) << "MallocSpace::MoreCore " << PrettySize(increment); in MoreCore() 178 VLOG(heap) << "Begin " << reinterpret_cast<const void*>(begin_) << "\n" in CreateZygoteSpace() 186 VLOG(heap) << "Creating new AllocSpace: "; in CreateZygoteSpace() 187 VLOG(heap) << "Size " << GetMemMap()->Size(); in CreateZygoteSpace() 188 VLOG(heap) << "GrowthLimit " << PrettySize(growth_limit); in CreateZygoteSpace() 189 VLOG(heap) << "Capacity " << PrettySize(capacity); in CreateZygoteSpace() 214 VLOG(heap) << "Failed creating zygote space from space " << GetName(); in CreateZygoteSpace() 216 VLOG(heap) << "zygote space creation done"; in CreateZygoteSpace()
|
D | dlmalloc_space.cc | 79 if (VLOG_IS_ON(heap) || VLOG_IS_ON(startup)) { in Create() 103 if (VLOG_IS_ON(heap) || VLOG_IS_ON(startup)) { in Create() 251 VLOG(heap) << "DlMallocSpace::SetFootprintLimit " << PrettySize(new_size); in SetFootprintLimit() 323 Heap* heap = Runtime::Current()->GetHeap(); in ArtDlMallocMoreCore() local 324 ::art::gc::space::DlMallocSpace* dlmalloc_space = heap->GetDlMallocSpace(); in ArtDlMallocMoreCore() 328 for (space::ContinuousSpace* space : heap->GetContinuousSpaces()) { in ArtDlMallocMoreCore()
|
D | rosalloc_space.cc | 101 if (VLOG_IS_ON(heap) || VLOG_IS_ON(startup)) { in Create() 127 if (VLOG_IS_ON(heap) || VLOG_IS_ON(startup)) { in Create() 248 VLOG(heap) << "RosAllocSpace::Trim() "; in Trim() 278 VLOG(heap) << "RosAllocSpace::SetFootprintLimit " << PrettySize(new_size); in SetFootprintLimit() 383 Heap* heap = Runtime::Current()->GetHeap(); in ArtRosAllocMoreCore() local 384 art::gc::space::RosAllocSpace* rosalloc_space = heap->GetRosAllocSpace(rosalloc); in ArtRosAllocMoreCore()
|
/art/compiler/ |
D | image_test.cc | 119 gc::Heap* heap = Runtime::Current()->GetHeap(); in TEST_F() local 120 ASSERT_TRUE(!heap->GetContinuousSpaces().empty()); in TEST_F() 121 gc::space::ContinuousSpace* space = heap->GetNonMovingSpace(); in TEST_F() 167 gc::Heap* heap = Runtime::Current()->GetHeap(); in TEST_F() local 168 ASSERT_TRUE(heap->HasImageSpace()); in TEST_F() 169 ASSERT_TRUE(heap->GetNonMovingSpace()->IsMallocSpace()); in TEST_F() 171 gc::space::ImageSpace* image_space = heap->GetImageSpace(); in TEST_F()
|
/art/runtime/gc/collector/ |
D | partial_mark_sweep.cc | 28 PartialMarkSweep::PartialMarkSweep(Heap* heap, bool is_concurrent, const std::string& name_prefix) in PartialMarkSweep() argument 29 : MarkSweep(heap, is_concurrent, name_prefix.empty() ? "partial " : name_prefix) { in PartialMarkSweep()
|
D | sticky_mark_sweep.cc | 27 StickyMarkSweep::StickyMarkSweep(Heap* heap, bool is_concurrent, const std::string& name_prefix) in StickyMarkSweep() argument 28 : PartialMarkSweep(heap, is_concurrent, in StickyMarkSweep()
|
D | garbage_collector.cc | 41 : duration_ns_(0), timings_("GC iteration timing logger", true, VLOG_IS_ON(heap)) { in Iteration() 61 GarbageCollector::GarbageCollector(Heap* heap, const std::string& name) in GarbageCollector() argument 62 : heap_(heap), in GarbageCollector()
|
D | partial_mark_sweep.h | 33 explicit PartialMarkSweep(Heap* heap, bool is_concurrent, const std::string& name_prefix = "");
|
D | semi_space.cc | 89 SemiSpace::SemiSpace(Heap* heap, bool generational, const std::string& name_prefix) in SemiSpace() argument 90 : GarbageCollector(heap, in SemiSpace() 191 VLOG(heap) << "Whole heap collection"; in MarkingPhase() 194 VLOG(heap) << "Bump pointer space only collection"; in MarkingPhase() 259 VLOG(heap) << "Protecting from_space_ : " << *from_space_; in MarkingPhase() 265 VLOG(heap) << "Protecting from_space_ with PROT_READ : " << *from_space_; in MarkingPhase() 414 VLOG(heap) << "Avoided dirtying " << PrettySize(saved_bytes_); in ReclaimPhase() 775 VLOG(heap) << "Protecting from_space_ with PROT_NONE : " << *from_space_; in FinishPhase()
|
/art/patchoat/ |
D | patchoat.h | 66 MemMap* heap, off_t delta, TimingLogger* timings) in PatchOat() argument 67 : image_(image), bitmap_(bitmap), heap_(heap), in PatchOat() 70 gc::accounting::ContinuousSpaceBitmap* bitmap, MemMap* heap, off_t delta, in PatchOat() argument 72 : oat_file_(oat_file), image_(image), bitmap_(bitmap), heap_(heap), in PatchOat()
|
/art/test/130-hprof/ |
D | info.txt | 1 Dump the heap for this test.
|