Searched refs:kRegionSize (Results 1 – 8 of 8) sorted by relevance
/art/runtime/gc/accounting/ |
D | read_barrier_table.h | 38 size_t capacity = static_cast<size_t>(kHeapCapacity / kRegionSize); in ReadBarrierTable() 39 DCHECK_EQ(kHeapCapacity / kRegionSize, in ReadBarrierTable() 40 static_cast<uint64_t>(static_cast<size_t>(kHeapCapacity / kRegionSize))); in ReadBarrierTable() 58 DCHECK_ALIGNED(start_addr, kRegionSize); in Clear() 59 DCHECK_ALIGNED(end_addr, kRegionSize); in Clear() 87 static constexpr size_t kRegionSize = 256 * KB; variable 95 uint8_t* entry_addr = mem_map_.Begin() + reinterpret_cast<uintptr_t>(heap_addr) / kRegionSize; in EntryFromAddr()
|
/art/runtime/gc/space/ |
D | region_space.cc | 53 CHECK_ALIGNED(capacity, kRegionSize); in CreateMemMap() 61 capacity + kRegionSize, in CreateMemMap() 80 CHECK_EQ(mem_map.Size(), capacity + kRegionSize); in CreateMemMap() 83 if (IsAlignedParam(mem_map.Begin(), kRegionSize)) { in CreateMemMap() 89 mem_map.AlignBy(kRegionSize); in CreateMemMap() 91 CHECK_ALIGNED(mem_map.Begin(), kRegionSize); in CreateMemMap() 92 CHECK_ALIGNED(mem_map.End(), kRegionSize); in CreateMemMap() 112 num_regions_(mem_map_.Size() / kRegionSize), in RegionSpace() 121 CHECK_ALIGNED(mem_map_.Size(), kRegionSize); in RegionSpace() 122 CHECK_ALIGNED(mem_map_.Begin(), kRegionSize); in RegionSpace() [all …]
|
D | region_space-inl.h | 55 if (LIKELY(num_bytes <= kRegionSize)) { in AllocNonvirtual() 316 DCHECK_GT(num_bytes, kRegionSize); in AllocLarge() 317 size_t num_regs_in_large_region = RoundUp(num_bytes, kRegionSize) / kRegionSize; in AllocLarge() 319 DCHECK_LT((num_regs_in_large_region - 1) * kRegionSize, num_bytes); in AllocLarge() 320 DCHECK_LE(num_bytes, num_regs_in_large_region * kRegionSize); in AllocLarge() 416 size_t allocated = num_regs_in_large_region * kRegionSize; in AllocLargeInRange() 461 DCHECK_ALIGNED(large_obj, kRegionSize); in FreeLarge() 464 uint8_t* end_addr = AlignUp(reinterpret_cast<uint8_t*>(large_obj) + bytes_allocated, kRegionSize); in FreeLarge() 466 for (uint8_t* addr = begin_addr; addr < end_addr; addr += kRegionSize) { in FreeLarge() 489 DCHECK_LT(begin_ + kRegionSize, Top()); in BytesAllocated() [all …]
|
D | region_space.h | 236 static constexpr size_t kRegionSize = 256 * KB; variable 354 size_t reg_idx = offset / kRegionSize; in RegionIdxForRefUnchecked() 381 return num_evac_regions_ * kRegionSize; in EvacBytes() 423 DCHECK_EQ(static_cast<size_t>(end - begin), kRegionSize); in Init() 479 DCHECK_LT(begin_ + kRegionSize, Top()); in IsLarge() 677 size_t reg_idx = offset / kRegionSize; in RefToRegionLocked()
|
/art/test/1001-app-image-regions/ |
D | app_image_regions.cc | 33 return gc::space::RegionSpace::kRegionSize; in Java_Main_getRegionSize()
|
/art/runtime/gc/collector/ |
D | concurrent_copying.cc | 121 static_assert(space::RegionSpace::kRegionSize == accounting::ReadBarrierTable::kRegionSize, in ConcurrentCopying() 3470 CHECK_LE(region_space_alloc_size, space::RegionSpace::kRegionSize); in Copy() 3914 << PrettySize(region_space_->GetMaxPeakNumNonFreeRegions() * space::RegionSpace::kRegionSize) in DumpPerformanceInfo() 3916 << PrettySize(region_space_->GetNumRegions() * space::RegionSpace::kRegionSize / 2) in DumpPerformanceInfo()
|
/art/runtime/gc/ |
D | heap.cc | 4596 if (space::RegionSpace::kRegionSize >= alloc_size) { in AllocWithNewTLAB() 4599 space::RegionSpace::kRegionSize, in AllocWithNewTLAB() 4602 kUsePartialTlabs ? kPartialTlabSize : gc::space::RegionSpace::kRegionSize; in AllocWithNewTLAB()
|
/art/dex2oat/linker/ |
D | image_writer.cc | 3818 region_size_ = gc::space::RegionSpace::kRegionSize; in ImageWriter()
|