Home
last modified time | relevance | path

Searched refs:alloc_size (Results 1 – 13 of 13) sorted by relevance

/art/runtime/gc/
Dheap-inl.h303 size_t alloc_size, in TryToAllocate() argument
310 UNLIKELY(IsOutOfMemoryOnAllocation(allocator_type, alloc_size, kGrow))) { in TryToAllocate()
317 alloc_size = RoundUp(alloc_size, space::BumpPointerSpace::kAlignment); in TryToAllocate()
318 ret = bump_pointer_space_->AllocNonvirtual(alloc_size); in TryToAllocate()
320 *bytes_allocated = alloc_size; in TryToAllocate()
321 *usable_size = alloc_size; in TryToAllocate()
322 *bytes_tl_bulk_allocated = alloc_size; in TryToAllocate()
329 size_t max_bytes_tl_bulk_allocated = rosalloc_space_->MaxBytesBulkAllocatedFor(alloc_size); in TryToAllocate()
335 ret = rosalloc_space_->Alloc(self, alloc_size, bytes_allocated, usable_size, in TryToAllocate()
340 rosalloc_space_->MaxBytesBulkAllocatedForNonvirtual(alloc_size); in TryToAllocate()
[all …]
Dheap.cc1939 size_t alloc_size, in AllocateInternalWithGc() argument
1957 l->PreObjectAllocated(self, h_klass, &alloc_size); in AllocateInternalWithGc()
1982 mirror::Object* ptr = TryToAllocate<true, false>(self, allocator, alloc_size, bytes_allocated, in AllocateInternalWithGc()
1992 alloc_size, in AllocateInternalWithGc()
2029 alloc_size, bytes_allocated, in AllocateInternalWithGc()
2040 VLOG(gc) << "Forcing collection of SoftReferences for " << PrettySize(alloc_size) in AllocateInternalWithGc()
2050 static_cast<int64_t>(kMinFreedHeapAfterGcForAlloc * growth_limit_ + alloc_size); in AllocateInternalWithGc()
2098 self, allocator, alloc_size, bytes_allocated, usable_size, bytes_tl_bulk_allocated); in AllocateInternalWithGc()
2116 ThrowOutOfMemoryError(self, alloc_size, allocator); in AllocateInternalWithGc()
2409 size_t alloc_size = RoundUp(obj_size, kObjectAlignment); in MarkNonForwardedObject() local
[all …]
Dheap.h921 EXPORT void JHPCheckNonTlabSampleAllocation(Thread* self, mirror::Object* ret, size_t alloc_size);
926 size_t alloc_size,
1141 size_t alloc_size,
1149 size_t alloc_size,
1165 size_t alloc_size,
1324 size_t alloc_size);
/art/runtime/javaheapprof/
Djavaheapsampler.cc77 size_t HeapSampler::GetSampleOffset(size_t alloc_size, in GetSampleOffset() argument
81 size_t exhausted_size = alloc_size + tlab_used; in GetSampleOffset()
110 size_t next_bytes_until_sample = bytes_until_sample - alloc_size; in GetSampleOffset()
113 << next_bytes_until_sample << " alloc= " << alloc_size; in GetSampleOffset()
Djavaheapsampler.h63 size_t GetSampleOffset(size_t alloc_size,
/art/runtime/gc/space/
Dspace_test.h180 size_t alloc_size; in SizeFootPrintGrowthLimitAndTrimBody() local
182 alloc_size = object_size; in SizeFootPrintGrowthLimitAndTrimBody()
184 alloc_size = test_rand(&rand_seed) % static_cast<size_t>(-object_size); in SizeFootPrintGrowthLimitAndTrimBody()
187 if (alloc_size < size_of_zero_length_byte_array) { in SizeFootPrintGrowthLimitAndTrimBody()
188 alloc_size = size_of_zero_length_byte_array; in SizeFootPrintGrowthLimitAndTrimBody()
196 object.Assign(Alloc(space, self, alloc_size, &bytes_allocated, nullptr, in SizeFootPrintGrowthLimitAndTrimBody()
199 object.Assign(AllocWithGrowth(space, self, alloc_size, &bytes_allocated, nullptr, in SizeFootPrintGrowthLimitAndTrimBody()
Dlarge_object_space.cc217 size_t alloc_size = it->second.mem_map.BaseSize(); in AllocationSize() local
219 *usable_size = alloc_size; in AllocationSize()
221 return alloc_size; in AllocationSize()
433 size_t alloc_size = cur_info->ByteSize(); in Walk() local
435 uint8_t* byte_end = byte_start + alloc_size; in Walk()
436 callback(byte_start, byte_end, alloc_size, arg); in Walk()
528 size_t alloc_size = info->ByteSize(); in AllocationSize() local
530 *usable_size = alloc_size; in AllocationSize()
532 return alloc_size; in AllocationSize()
Dlarge_object_space_test.cc133 size_t alloc_size, bytes_tl_bulk_allocated; in Run() local
134 mirror::Object* ptr = los_->Alloc(self, size_, &alloc_size, nullptr, in Run()
Dregion_space.h320 void AddLiveBytes(mirror::Object* ref, size_t alloc_size) { in AddLiveBytes() argument
322 reg->AddLiveBytes(alloc_size); in AddLiveBytes()
Dregion_space.cc658 size_t alloc_size = RoundUp(obj_size, space::RegionSpace::kAlignment); in CheckLiveBytesAgainstRegionBitmap() local
659 live_bytes_recount += alloc_size; in CheckLiveBytesAgainstRegionBitmap()
/art/runtime/gc/collector/
Dconcurrent_copying.cc1182 size_t alloc_size = RoundUp(obj_size, space::RegionSpace::kAlignment); in AddLiveBytesAndScanRef() local
1183 region_space_->AddLiveBytes(ref, alloc_size); in AddLiveBytesAndScanRef()
3393 mirror::Object* ConcurrentCopying::AllocateInSkippedBlock(Thread* const self, size_t alloc_size) { in AllocateInSkippedBlock() argument
3395 CHECK_ALIGNED(alloc_size, space::RegionSpace::kAlignment); in AllocateInSkippedBlock()
3401 auto it = skipped_blocks_map_.lower_bound(alloc_size); in AllocateInSkippedBlock()
3407 CHECK_GE(byte_size, alloc_size); in AllocateInSkippedBlock()
3408 if (byte_size > alloc_size && byte_size - alloc_size < min_object_size) { in AllocateInSkippedBlock()
3410 it = skipped_blocks_map_.lower_bound(alloc_size + min_object_size); in AllocateInSkippedBlock()
3415 CHECK_ALIGNED(it->first - alloc_size, space::RegionSpace::kAlignment); in AllocateInSkippedBlock()
3416 CHECK_GE(it->first - alloc_size, min_object_size) in AllocateInSkippedBlock()
[all …]
Dconcurrent_copying.h274 mirror::Object* AllocateInSkippedBlock(Thread* const self, size_t alloc_size)
/art/runtime/mirror/
Dstring-alloc-inl.h210 size_t alloc_size = RoundUp(size, kObjectAlignment); in Alloc() local
233 alloc_size, in Alloc()