Lines Matching refs:MemoryChunk

46 class MemoryChunk;  variable
252 class MemoryChunk {
256 size_t operator()(MemoryChunk* const chunk) const { in operator()
418 static MemoryChunk* FromAddress(Address a) { in FromAddress()
419 return reinterpret_cast<MemoryChunk*>(OffsetFrom(a) & ~kAlignmentMask); in FromAddress()
422 static MemoryChunk* FromHeapObject(const HeapObject* o) { in FromHeapObject()
423 return reinterpret_cast<MemoryChunk*>(reinterpret_cast<Address>(o) & in FromHeapObject()
430 static inline MemoryChunk* FromAnyPointerAddress(Heap* heap, Address addr);
437 MemoryChunk* chunk = MemoryChunk::FromAddress(mark - 1); in UpdateHighWaterMark()
448 return reinterpret_cast<Address>(const_cast<MemoryChunk*>(this)); in address()
548 if (IsFlagSet(MemoryChunk::HAS_PROGRESS_BAR)) { in ResetProgressBar()
649 base::ListNode<MemoryChunk>& list_node() { return list_node_; } in list_node()
652 static MemoryChunk* Initialize(Heap* heap, Address base, size_t size,
727 base::ListNode<MemoryChunk> list_node_;
754 static_assert(kMaxRegularHeapObjectSize <= MemoryChunk::kAllocatableMemory,
764 class Page : public MemoryChunk {
770 static_cast<intptr_t>(MemoryChunk::POINTERS_TO_HERE_ARE_INTERESTING) |
771 static_cast<intptr_t>(MemoryChunk::POINTERS_FROM_HERE_ARE_INTERESTING) |
772 static_cast<intptr_t>(MemoryChunk::INCREMENTAL_MARKING);
899 class LargePage : public MemoryChunk {
920 static LargePage* Initialize(Heap* heap, MemoryChunk* chunk,
1030 MemoryChunk* first_page() { return memory_chunk_list_.front(); } in first_page()
1031 MemoryChunk* last_page() { return memory_chunk_list_.back(); } in last_page()
1033 base::List<MemoryChunk>& memory_chunk_list() { return memory_chunk_list_; } in memory_chunk_list()
1048 base::List<MemoryChunk> memory_chunk_list_;
1068 STATIC_ASSERT(MemoryChunk::kSizeOffset == offsetof(MemoryChunk, size_));
1071 STATIC_ASSERT(sizeof(MemoryChunk) <= MemoryChunk::kHeaderSize);
1072 STATIC_ASSERT(sizeof(LargePage) <= MemoryChunk::kHeaderSize);
1073 STATIC_ASSERT(sizeof(Page) <= MemoryChunk::kHeaderSize);
1118 DCHECK(IsAddressAligned(start, MemoryChunk::kAlignment)); in FreeBlock()
1123 DCHECK(IsAddressAligned(start, MemoryChunk::kAlignment)); in FreeBlock()
1261 void AddMemoryChunkSafe(MemoryChunk* chunk) { in AddMemoryChunkSafe()
1269 MemoryChunk* TryGetPooledMemoryChunkSafe() { in TryGetPooledMemoryChunkSafe()
1275 MemoryChunk* chunk = GetMemoryChunkSafe<kPooled>(); in TryGetPooledMemoryChunkSafe()
1312 void AddMemoryChunkSafe(MemoryChunk* chunk) { in AddMemoryChunkSafe()
1318 MemoryChunk* GetMemoryChunkSafe() { in GetMemoryChunkSafe()
1321 MemoryChunk* chunk = chunks_[type].back(); in GetMemoryChunkSafe()
1336 std::vector<MemoryChunk*> chunks_[kNumberOfChunkQueues];
1393 void Free(MemoryChunk* chunk);
1422 MemoryChunk* AllocateChunk(size_t reserve_area_size, size_t commit_area_size,
1440 void PartialFreeMemory(MemoryChunk* chunk, Address start_free,
1451 bool IsMemoryChunkExecutable(MemoryChunk* chunk) { in IsMemoryChunkExecutable()
1476 void PreFreeMemory(MemoryChunk* chunk);
1479 void PerformFreeMemory(MemoryChunk* chunk);
1484 MemoryChunk* AllocatePagePooled(SpaceType* owner);
1508 void RegisterExecutableMemoryChunk(MemoryChunk* chunk) { in RegisterExecutableMemoryChunk()
1509 DCHECK(chunk->IsFlagSet(MemoryChunk::IS_EXECUTABLE)); in RegisterExecutableMemoryChunk()
1514 void UnregisterExecutableMemoryChunk(MemoryChunk* chunk) { in UnregisterExecutableMemoryChunk()
1543 std::unordered_set<MemoryChunk*> executable_memory_;
2240 Page* InitializePage(MemoryChunk* chunk, Executability executable); in NON_EXPORTED_BASE()
2322 MemoryChunk::UpdateHighWaterMark(allocation_info_.top()); in NON_EXPORTED_BASE()
2485 Page* InitializePage(MemoryChunk* chunk, Executability executable);
3116 inline MemoryChunk* next();