Lines Matching refs:Page
47 class Page; variable
125 DCHECK((Page::kObjectStartOffset <= offset) && (offset <= Page::kPageSize))
160 FreeListCategory(FreeList* free_list, Page* page) in FreeListCategory()
199 inline Page* page() const { return page_; } in page()
227 Page* const page_;
764 class Page : public MemoryChunk {
777 static Page* FromAddress(Address addr) { in FromAddress()
778 return reinterpret_cast<Page*>(OffsetFrom(addr) & ~kPageAlignmentMask); in FromAddress()
780 static Page* FromHeapObject(const HeapObject* o) { in FromHeapObject()
781 return reinterpret_cast<Page*>(reinterpret_cast<Address>(o) & in FromHeapObject()
789 static Page* FromAllocationAreaAddress(Address address) { in FromAllocationAreaAddress()
790 return Page::FromAddress(address - kPointerSize); in FromAllocationAreaAddress()
795 return Page::FromAddress(address1) == Page::FromAddress(address2); in OnSamePage()
807 static Page* ConvertNewToOld(Page* old_page);
813 Page* next_page() { return static_cast<Page*>(list_node_.next()); } in next_page()
814 Page* prev_page() { return static_cast<Page*>(list_node_.prev()); } in prev_page()
892 class ReadOnlyPage : public Page {
1073 STATIC_ASSERT(sizeof(Page) <= MemoryChunk::kHeaderSize);
1119 DCHECK(size >= static_cast<size_t>(Page::kPageSize)); in FreeBlock()
1124 DCHECK(size >= static_cast<size_t>(Page::kPageSize)); in FreeBlock()
1218 Page* page = Page::FromAddress(addr); in Update()
1231 static const int kSize = Page::kPageSize / kRegionSize;
1233 STATIC_ASSERT(Page::kPageSize % kRegionSize == 0);
1372 : Page::kAllocatableMemory; in PageAreaSize()
1387 Page* AllocatePage(size_t size, SpaceType* owner, Executability executable);
1409 return (Available() / Page::kPageSize) * Page::kAllocatableMemory; in MaxAvailable()
1490 Page* InitializePagesInChunk(int chunk_id, int pages_in_chunk,
1550 extern template Page*
1553 extern template Page*
1556 extern template Page*
1594 typedef PageIteratorImpl<Page> PageIterator;
1600 PageRange(Page* begin, Page* end) : begin_(begin), end_(end) {} in PageRange()
1601 explicit PageRange(Page* page) : PageRange(page, page->next_page()) {} in PageRange()
1608 Page* begin_;
1609 Page* end_;
1625 explicit HeapObjectIterator(Page* page);
1684 return (Page::FromAllocationAreaAddress(top_) == in VerifyPagedAllocation()
1685 Page::FromAllocationAreaAddress(limit_)) && in VerifyPagedAllocation()
1727 size_t AllocatedOnPage(Page* page) { return allocated_on_page_[page]; } in AllocatedOnPage()
1730 void IncreaseAllocatedBytes(size_t bytes, Page* page) { in IncreaseAllocatedBytes()
1738 void DecreaseAllocatedBytes(size_t bytes, Page* page) { in DecreaseAllocatedBytes()
1775 std::unordered_map<Page*, size_t, Page::Hasher> allocated_on_page_;
1882 size_t EvictFreeListItems(Page* page);
1883 bool ContainsPageFreeListItems(Page* page);
1909 inline Page* GetPageForCategoryType(FreeListCategoryType type);
1937 static const size_t kMaxBlockSize = Page::kAllocatableMemory;
2203 Page* page = Page::FromAddress(start); in NON_EXPORTED_BASE()
2225 void DecreaseAllocatedBytes(size_t bytes, Page* page) { in NON_EXPORTED_BASE()
2228 void IncreaseAllocatedBytes(size_t bytes, Page* page) { in NON_EXPORTED_BASE()
2238 void RefineAllocatedBytesAfterSweeping(Page* page); in NON_EXPORTED_BASE()
2240 Page* InitializePage(MemoryChunk* chunk, Executability executable); in NON_EXPORTED_BASE()
2242 void ReleasePage(Page* page); in NON_EXPORTED_BASE()
2246 size_t AddPage(Page* page); in NON_EXPORTED_BASE()
2247 void RemovePage(Page* page); in NON_EXPORTED_BASE()
2250 Page* RemovePageSafe(int size_in_bytes); in NON_EXPORTED_BASE()
2299 inline void UnlinkFreeListCategories(Page* page); in NON_EXPORTED_BASE()
2300 inline size_t RelinkFreeListCategories(Page* page); in NON_EXPORTED_BASE()
2302 Page* first_page() { return reinterpret_cast<Page*>(Space::first_page()); } in NON_EXPORTED_BASE()
2311 size_t ShrinkPageToHighWaterMark(Page* page); in NON_EXPORTED_BASE()
2321 Page::FromAddress(top) == Page::FromAddress(limit - 1)); in NON_EXPORTED_BASE()
2456 Page* current_page() { return current_page_; } in current_page()
2466 Page* next_page = current_page_->next_page(); in AdvancePage()
2482 void RemovePage(Page* page);
2483 void PrependPage(Page* page);
2485 Page* InitializePage(MemoryChunk* chunk, Executability executable);
2518 Page* first_page() { return reinterpret_cast<Page*>(Space::first_page()); } in first_page()
2519 Page* last_page() { return reinterpret_cast<Page*>(Space::last_page()); } in last_page()
2545 return static_cast<int>(current_capacity_ / Page::kPageSize); in max_pages()
2567 Page* current_page_;
2633 return to_space_.pages_used() * Page::kAllocatableMemory + in Size()
2642 return (to_space_.current_capacity() / Page::kPageSize) * in Capacity()
2643 Page::kAllocatableMemory; in Capacity()
2683 Page* const age_mark_page = Page::FromAllocationAreaAddress(age_mark); in AllocatedSinceLastGC()
2684 Page* const last_page = Page::FromAllocationAreaAddress(top()); in AllocatedSinceLastGC()
2685 Page* current_page = age_mark_page; in AllocatedSinceLastGC()
2698 allocated += Page::kAllocatableMemory; in AllocatedSinceLastGC()
2707 void MovePageFromSpaceToSpace(Page* page) { in MovePageFromSpaceToSpace()
2806 Page* first_page() { return to_space_.first_page(); } in first_page()
2807 Page* last_page() { return to_space_.last_page(); } in last_page()
3001 if (chunk_size <= (Page::kPageSize + Page::kObjectStartOffset)) return 0; in ObjectSizeFor()
3002 return chunk_size - Page::kPageSize - Page::kObjectStartOffset; in ObjectSizeFor()