Lines Matching refs:Page
87 void VerifyMarkingOnPage(const Page* page, Address start, Address end);
98 void MarkingVerifier::VerifyMarkingOnPage(const Page* page, Address start, in VerifyMarkingOnPage()
136 Page* page = *(it++); in VerifyMarking()
144 for (Page* p : *space) { in VerifyMarking()
264 Page* page = *(it++); in VerifyEvacuation()
273 for (Page* p : *space) { in VerifyEvacuation()
351 static_cast<size_t>(tasks * Page::kPageSize))) { in NumberOfParallelCompactionTasks()
415 void MarkCompactCollector::AddEvacuationCandidate(Page* p) { in AddEvacuationCandidate()
483 for (Page* p : *space) { in VerifyMarkbitsAreClean()
491 for (Page* p : PageRange(space->first_allocatable_address(), space->top())) { in VerifyMarkbitsAreClean()
518 for (Page* p : *space) { in ClearMarkbitsInPagedSpace()
524 for (Page* p : *space) { in ClearMarkbitsInNewSpace()
609 typedef std::pair<size_t, Page*> LiveBytesPagePair; in CollectEvacuationCandidates()
614 Page* owner_of_linear_allocation_area = in CollectEvacuationCandidates()
617 : Page::FromAllocationAreaAddress(space->top()); in CollectEvacuationCandidates()
618 for (Page* p : *space) { in CollectEvacuationCandidates()
640 Page* p = pages[i].second; in CollectEvacuationCandidates()
660 Page* p = pages[i].second; in CollectEvacuationCandidates()
747 for (Page* p : evacuation_candidates_) { in AbortCompaction()
1116 Page* p = Page::FromAddress(reinterpret_cast<Address>(value)); in RecordMigratedSlot()
1119 p->IsFlagSet(Page::PAGE_NEW_NEW_PROMOTION)); in RecordMigratedSlot()
1121 Page::FromAddress(slot), slot); in RecordMigratedSlot()
1124 Page::FromAddress(slot), slot); in RecordMigratedSlot()
1254 Page* page = Page::FromAddress(object->address()); in AbortCompactionForTesting()
1255 if (page->IsFlagSet(Page::COMPACTION_WAS_ABORTED_FOR_TESTING)) { in AbortCompactionForTesting()
1256 page->ClearFlag(Page::COMPACTION_WAS_ABORTED_FOR_TESTING); in AbortCompactionForTesting()
1258 page->SetFlag(Page::COMPACTION_WAS_ABORTED_FOR_TESTING); in AbortCompactionForTesting()
1374 static void Move(Page* page) { in Move()
1378 page->SetFlag(Page::PAGE_NEW_NEW_PROMOTION); in Move()
1382 Page* new_page = Page::ConvertNewToOld(page); in Move()
1384 new_page->SetFlag(Page::PAGE_NEW_OLD_PROMOTION); in Move()
1419 Page::FromAddress(object->address())->owner()->identity(), object, in Visit()
2123 Page* target_page = Page::FromAddress(reinterpret_cast<Address>(target)); in RecordRelocSlot()
2124 Page* source_page = Page::FromAddress(reinterpret_cast<Address>(host)); in RecordRelocSlot()
2153 Page::FromAddress(heap_obj->address()) in UpdateSlot()
2154 ->IsFlagSet(Page::COMPACTION_WAS_ABORTED)); in UpdateSlot()
2278 Page::FromAddress(reinterpret_cast<Address>(*p)), in UpdateReferenceInExternalStringTableEntry()
2279 Page::FromHeapObject(new_string), ExternalString::cast(new_string)); in UpdateReferenceInExternalStringTableEntry()
2291 for (Page* p : in EvacuatePrologue()
2317 for (Page* p : *heap()->old_space()) { in EvacuateEpilogue()
2348 return FLAG_page_promotion_threshold * Page::kAllocatableMemory / 100; in PageEvacuationThreshold()
2349 return Page::kAllocatableMemory + kPointerSize; in PageEvacuationThreshold()
2369 void EvacuatePage(Page* page);
2386 virtual void RawEvacuatePage(Page* page, intptr_t* saved_live_bytes) = 0;
2414 void Evacuator::EvacuatePage(Page* page) { in EvacuatePage()
2432 page->IsFlagSet(Page::PAGE_NEW_OLD_PROMOTION) || in EvacuatePage()
2433 page->IsFlagSet(Page::PAGE_NEW_NEW_PROMOTION), in EvacuatePage()
2436 evacuation_time, page->IsFlagSet(Page::COMPACTION_WAS_ABORTED)); in EvacuatePage()
2467 void RawEvacuatePage(Page* page, intptr_t* live_bytes) override;
2472 void FullEvacuator::RawEvacuatePage(Page* page, intptr_t* live_bytes) { in RawEvacuatePage()
2522 explicit PageEvacuationItem(Page* page) : page_(page) {} in PageEvacuationItem()
2524 Page* page() const { return page_; } in page()
2527 Page* page_;
2599 bool MarkCompactCollectorBase::ShouldMovePage(Page* p, intptr_t live_bytes) { in ShouldMovePage()
2612 for (Page* page : old_space_evacuation_pages_) { in EvacuatePagesInParallel()
2617 for (Page* page : new_space_evacuation_pages_) { in EvacuatePagesInParallel()
2662 void MarkCompactCollector::RecordLiveSlotsOnPage(Page* page) { in RecordLiveSlotsOnPage()
2780 for (Page* p : new_space_evacuation_pages_) { in Evacuate()
2781 if (p->IsFlagSet(Page::PAGE_NEW_NEW_PROMOTION)) { in Evacuate()
2782 p->ClearFlag(Page::PAGE_NEW_NEW_PROMOTION); in Evacuate()
2784 } else if (p->IsFlagSet(Page::PAGE_NEW_OLD_PROMOTION)) { in Evacuate()
2785 p->ClearFlag(Page::PAGE_NEW_OLD_PROMOTION); in Evacuate()
2792 for (Page* p : old_space_evacuation_pages_) { in Evacuate()
2798 if (p->IsFlagSet(Page::COMPACTION_WAS_ABORTED)) { in Evacuate()
2800 p->ClearFlag(Page::COMPACTION_WAS_ABORTED); in Evacuate()
2858 if (chunk_->IsFlagSet(Page::PAGE_NEW_NEW_PROMOTION)) { in Process()
2950 if (Page::FromAddress(heap_object->address()) in CheckAndUpdateOldToNewSlot()
2951 ->IsFlagSet(Page::PAGE_NEW_NEW_PROMOTION)) { in CheckAndUpdateOldToNewSlot()
3082 explicit ArrayBufferTrackerUpdatingItem(Page* page, EvacuationState state) in ArrayBufferTrackerUpdatingItem()
3103 Page* const page_;
3113 for (Page* page : PageRange(space_start, space_end)) { in CollectToSpaceUpdatingItems()
3153 for (Page* p : new_space_evacuation_pages_) { in CollectNewSpaceArrayBufferTrackerItems()
3168 for (Page* p : old_space_evacuation_pages_) { in CollectOldSpaceArrayBufferTrackerItems()
3179 Page* p = object_and_page.second; in CollectOldSpaceArrayBufferTrackerItems()
3275 HeapObject* failed_object, Page* page) { in ReportAbortedEvacuationCandidate()
3284 Page* page = object_and_page.second; in PostProcessEvacuationCandidates()
3285 page->SetFlag(Page::COMPACTION_WAS_ABORTED); in PostProcessEvacuationCandidates()
3307 for (Page* p : old_space_evacuation_pages_) { in PostProcessEvacuationCandidates()
3308 if (p->IsFlagSet(Page::COMPACTION_WAS_ABORTED)) { in PostProcessEvacuationCandidates()
3327 for (Page* p : old_space_evacuation_pages_) { in ReleaseEvacuationCandidates()
3346 Page* p = *(it++); in StartSweepSpace()
3355 if (p->IsFlagSet(Page::NEVER_ALLOCATE_ON_PAGE)) { in StartSweepSpace()
3360 p->set_concurrent_sweeping_state(Page::kSweepingInProgress); in StartSweepSpace()
3647 for (Page* p : sweep_to_iterate_pages_) { in CleanupSweepToIteratePages()
3648 if (p->IsFlagSet(Page::SWEEP_TO_ITERATE)) { in CleanupSweepToIteratePages()
3649 p->ClearFlag(Page::SWEEP_TO_ITERATE); in CleanupSweepToIteratePages()
3701 Page* p = Page::FromAddress(reinterpret_cast<Address>(value)); in RecordMigratedSlot()
3704 p->IsFlagSet(Page::PAGE_NEW_NEW_PROMOTION)); in RecordMigratedSlot()
3706 Page::FromAddress(slot), slot); in RecordMigratedSlot()
3709 Page::FromAddress(slot), slot); in RecordMigratedSlot()
3832 for (Page* p : in CollectGarbage()
3834 DCHECK(!p->IsFlagSet(Page::SWEEP_TO_ITERATE)); in CollectGarbage()
3857 Page* p, MarkingTreatmentMode marking_mode, in MakeIterable()
3902 p->ClearFlag(Page::SWEEP_TO_ITERATE); in MakeIterable()
3993 for (Page* p : in EvacuatePrologue()
4042 Page::kPageSize); in YoungGenerationMarkingTask()
4094 local_live_bytes_[Page::FromAddress(reinterpret_cast<Address>(object))] += in IncrementLiveBytes()
4108 std::unordered_map<Page*, intptr_t, Page::Hasher> local_live_bytes_;
4240 static_cast<int>(heap()->new_space()->Capacity()) / Page::kPageSize; in MarkRootSetInParallel()
4321 for (Page* p : new_space_evacuation_pages_) { in Evacuate()
4322 if (p->IsFlagSet(Page::PAGE_NEW_NEW_PROMOTION) || in Evacuate()
4323 p->IsFlagSet(Page::PAGE_NEW_OLD_PROMOTION)) { in Evacuate()
4324 p->ClearFlag(Page::PAGE_NEW_NEW_PROMOTION); in Evacuate()
4325 p->ClearFlag(Page::PAGE_NEW_OLD_PROMOTION); in Evacuate()
4326 p->SetFlag(Page::SWEEP_TO_ITERATE); in Evacuate()
4352 void RawEvacuatePage(Page* page, intptr_t* live_bytes) override;
4357 void YoungGenerationEvacuator::RawEvacuatePage(Page* page, in RawEvacuatePage()
4424 for (Page* page : new_space_evacuation_pages_) { in EvacuatePagesInParallel()
4450 for (Page* p : new_space_evacuation_pages_) { in CollectNewSpaceArrayBufferTrackerItems()