Home
last modified time | relevance | path

Searched refs:RoundUp (Results 1 – 25 of 109) sorted by relevance

12345

/art/runtime/gc/space/
Dbump_pointer_space-inl.h32 num_bytes = RoundUp(num_bytes, kAlignment); in Alloc()
49 num_bytes = RoundUp(num_bytes, kAlignment); in AllocThreadUnsafe()
95 return reinterpret_cast<mirror::Object*>(RoundUp(position, kAlignment)); in GetNextObject()
Dmalloc_space.cc105 *growth_limit = RoundUp(*growth_limit, gPageSize); in CreateMemMap()
106 *capacity = RoundUp(*capacity, gPageSize); in CreateMemMap()
143 growth_limit = RoundUp(growth_limit, gPageSize); in SetGrowthLimit()
186 SetEnd(reinterpret_cast<uint8_t*>(RoundUp(reinterpret_cast<uintptr_t>(End()), gPageSize))); in CreateZygoteSpace()
191 size_t size = RoundUp(Size(), gPageSize); in CreateZygoteSpace()
203 SetGrowthLimit(RoundUp(size, gPageSize)); in CreateZygoteSpace()
Dbump_pointer_space.cc28 capacity = RoundUp(capacity, gPageSize); in Create()
225 bytes = RoundUp(bytes, kAlignment); in AllocNewTlab()
254 *usable_size = RoundUp(num_bytes, kAlignment); in AllocationSizeNonvirtual()
/art/runtime/arch/arm/
Djni_frame_arm.h70 return RoundUp(size, kAapcsStackAlignment); in GetCriticalNativeStubFrameSize()
80 return RoundUp(size, kAapcsStackAlignment); in GetCriticalNativeDirectCallFrameSize()
/art/runtime/arch/x86/
Djni_frame_x86.h66 return RoundUp(size, kNativeStackAlignment); in GetCriticalNativeStubFrameSize()
76 return RoundUp(size, kNativeStackAlignment); in GetCriticalNativeDirectCallFrameSize()
/art/runtime/oat/
Dimage.cc72 CHECK_EQ(image_begin, RoundUp(image_begin, kElfSegmentAlignment)); in ImageHeader()
74 CHECK_EQ(oat_file_begin, RoundUp(oat_file_begin, kElfSegmentAlignment)); in ImageHeader()
75 CHECK_EQ(oat_data_begin, RoundUp(oat_data_begin, kElfSegmentAlignment)); in ImageHeader()
133 return image_reservation_size_ == RoundUp(image_size_, kElfSegmentAlignment); in IsAppImage()
191 static const size_t kStartPos = RoundUp(sizeof(ImageHeader), kObjectAlignment); in VisitObjects()
195 pos += RoundUp(object->SizeOf(), kObjectAlignment); in VisitObjects()
408 out_offset = RoundUp(out_offset, alignof(ImageHeader::Block)); in WriteData()
428 out_offset = RoundUp(out_offset, kElfSegmentAlignment); in WriteData()
/art/runtime/arch/arm64/
Djni_frame_arm64.h77 return RoundUp(size, kAapcs64StackAlignment); in GetCriticalNativeStubFrameSize()
87 return RoundUp(size, kAapcs64StackAlignment); in GetCriticalNativeDirectCallFrameSize()
/art/runtime/arch/riscv64/
Djni_frame_riscv64.h80 return RoundUp(size, kNativeStackAlignment); in GetCriticalNativeStubFrameSize()
90 return RoundUp(size, kNativeStackAlignment); in GetCriticalNativeDirectCallFrameSize()
/art/runtime/arch/x86_64/
Djni_frame_x86_64.h83 return RoundUp(size, kNativeStackAlignment); in GetCriticalNativeStubFrameSize()
93 return RoundUp(size, kNativeStackAlignment); in GetCriticalNativeDirectCallFrameSize()
/art/runtime/
Dnterp_helpers.cc176 return RoundUp(NterpGetFrameSizeWithoutPadding(method, isa), kStackAlignment); in NterpGetFrameSize()
199 RoundUp(out_regs * kVRegSize, kPointerSize) + // out arguments and pointer alignment in NterpGetReferenceArray()
208 RoundUp(out_regs * kVRegSize, kPointerSize); // out arguments and pointer alignment in NterpGetDexPC()
242 DCHECK_EQ(NterpGetFrameSize(method, isa), RoundUp(frame_size_without_padding, kStackAlignment)); in CanMethodUseNterp()
Dlinear_alloc-inl.h109 RoundUp(reinterpret_cast<uintptr_t>(ptr), 16) - reinterpret_cast<uintptr_t>(ptr); in AllocAlign16()
118 required_size = size + RoundUp(sizeof(TrackingHeader), 16); in AllocAlign16()
/art/odrefresh/
Dodr_fs_utils_test.cc145 uint64_t expected_bytes_used = RoundUp(kFirstFileBytes, sb.st_blocks * kBytesPerBlock) + in TEST_F()
146 RoundUp(kSecondFileBytes, sb.st_blocks * kBytesPerBlock); in TEST_F()
154 expected_bytes_used += RoundUp(i, sb.st_blocks * kBytesPerBlock); in TEST_F()
/art/libartbase/base/metrics/
Dmetrics.h317 static_assert(RoundUp(sizeof(*this), sizeof(uint64_t))
318 == RoundUp(sizeof(intptr_t) + sizeof(value_t), sizeof(uint64_t)));
355 static_assert(RoundUp(sizeof(*this), sizeof(uint64_t))
356 == RoundUp(sizeof(intptr_t) + sizeof(value_t) + sizeof(count_t),
408 static_assert(RoundUp(sizeof(*this), sizeof(uint64_t)) ==
409 RoundUp(sizeof(intptr_t) + sizeof(value_t), sizeof(uint64_t)));
452 static_assert(RoundUp(sizeof(*this), sizeof(uint64_t))
453 == RoundUp(sizeof(intptr_t) + sizeof(value_t) * num_buckets_, sizeof(uint64_t)));
514 static_assert(RoundUp(sizeof(*this), sizeof(uint64_t)) ==
515 RoundUp(sizeof(intptr_t) + sizeof(T), sizeof(uint64_t)));
/art/dex2oat/utils/
Dswap_space.cc111 size = RoundUp(size, 8U); in Alloc()
151 size_t next_part = std::max(RoundUp(min_size, page_size), RoundUp(kMinimumMapSize, page_size)); in NewFileChunk()
179 size = RoundUp(size, 8U); in Free()
/art/runtime/gc/accounting/
Dbitmap.cc49 const size_t bitmap_size = RoundUp( in AllocateMemMap()
50 RoundUp(num_bits, kBitsPerBitmapWord) / kBitsPerBitmapWord * sizeof(uintptr_t), gPageSize); in AllocateMemMap()
Dspace_bitmap_test.cc143 {0, RoundUp(10 * KB, gObjectAlignment) + gObjectAlignment}, in TYPED_TEST()
147 {RoundUp(1 * KB, gObjectAlignment) + gObjectAlignment, in TYPED_TEST()
148 RoundUp(2 * KB, gObjectAlignment) + 5 * gObjectAlignment}, in TYPED_TEST()
/art/dex2oat/linker/arm/
Drelative_patcher_thumb2_test.cc753 uint32_t thunk_offset = GetMethodOffset(method_idx) + RoundUp(kMethodCodeSize, kArmCodeAlignment); in TestBakerFieldWide()
819 thunk_offset += RoundUp(expected_thunk.size(), kArmCodeAlignment); in TestBakerFieldWide()
854 uint32_t thunk_offset = GetMethodOffset(method_idx) + RoundUp(kMethodCodeSize, kArmCodeAlignment); in TestBakerFieldNarrow()
923 thunk_offset += RoundUp(expected_thunk.size(), kArmCodeAlignment); in TestBakerFieldNarrow()
983 RoundUp(raw_code1.size() + sizeof(OatQuickMethodHeader), kArmCodeAlignment); in TEST_F()
1004 - RoundUp(thunk_size + sizeof(OatQuickMethodHeader), kArmCodeAlignment) in TEST_F()
1005 - RoundUp(kNopCode.size() + sizeof(OatQuickMethodHeader), kArmCodeAlignment) in TEST_F()
1052 RoundUp(raw_code1.size() + sizeof(OatQuickMethodHeader), kArmCodeAlignment); in TEST_F()
1060 BneWWithOffset(kLiteralOffset1, RoundUp(raw_code1.size(), kArmCodeAlignment)); in TEST_F()
1084 RoundUp(raw_code1.size() + sizeof(OatQuickMethodHeader), kArmCodeAlignment); in TEST_F()
[all …]
/art/compiler/jni/quick/arm/
Dcalling_convention_arm.cc238 gpr_index_ = RoundUp(gpr_index_, 2u) + 2u; in Next()
256 return RoundUp(gpr_index_, 2u) + 1u < kHFCoreArgumentRegistersCount; in IsCurrentParamInRegister()
279 CHECK_EQ(RoundUp(gpr_index_, 2u), 2u); in CurrentParamRegister()
442 return RoundUp(total_size, kStackAlignment); in FrameSize()
463 size_t out_args_size = RoundUp(size, kAapcsStackAlignment); in OutFrameSize()
/art/libartbase/base/
Darena_allocator.h273 bytes = RoundUp(bytes, kAlignment);
292 RoundUp(reinterpret_cast<uintptr_t>(ptr_), 16) - reinterpret_cast<uintptr_t>(ptr_);
313 const size_t aligned_ptr_size = RoundUp(ptr_size, kAlignment);
319 const size_t aligned_new_size = RoundUp(new_size, kAlignment);
Dscoped_arena_allocator.cc96 size_t rounded_bytes = RoundUp(bytes + kMemoryToolRedZoneBytes, 8); in AllocWithMemoryTool()
152 arena_stack_->top_ptr_ = mark_ptr_ + RoundUp(sizeof(ScopedArenaAllocator), 8); in Reset()
Dbit_utils.h171 constexpr T RoundUp(T x, std::remove_reference_t<T> n) WARN_UNUSED;
174 constexpr T RoundUp(T x, std::remove_reference_t<T> n) { in RoundUp() function
181 return RoundUp(x, n); in CondRoundUp()
201 return reinterpret_cast<T*>(RoundUp(reinterpret_cast<uintptr_t>(x), n)); in AlignUp()
513 return RoundUp(num_bits, kBitsPerByte) / kBitsPerByte; in BitsToBytesRoundUp()
/art/runtime/mirror/
Darray.h84 size_t data_offset = RoundUp(OFFSETOF_MEMBER(Array, first_element_), component_size); in DataOffset()
85 DCHECK_EQ(RoundUp(data_offset, component_size), data_offset) in DataOffset()
92 constexpr size_t data_offset = RoundUp(kFirstElementOffset, kComponentSize); in DataOffset()
93 static_assert(RoundUp(data_offset, kComponentSize) == data_offset, "RoundUp fail"); in DataOffset()
/art/dex2oat/linker/arm64/
Drelative_patcher_arm64_test.cc1088 GetMethodOffset(method_idx) + RoundUp(kMethodCodeSize, kArm64CodeAlignment); in TestBakerField()
1141 thunk_offset += RoundUp(expected_thunk.size(), kArm64CodeAlignment); in TestBakerField()
1180 1 * MB - RoundUp(raw_code1.size() + sizeof(OatQuickMethodHeader), kArm64CodeAlignment); in TEST_F()
1195 1 * MB - RoundUp(thunk_size + sizeof(OatQuickMethodHeader), kArm64CodeAlignment) in TEST_F()
1196 - RoundUp(kNopCode.size() + sizeof(OatQuickMethodHeader), kArm64CodeAlignment) in TEST_F()
1240 1 * MB - RoundUp(raw_code1.size() + sizeof(OatQuickMethodHeader), kArm64CodeAlignment); in TEST_F()
1247 const uint32_t cbnz_offset = RoundUp(raw_code1.size(), kArm64CodeAlignment) - kLiteralOffset1; in TEST_F()
1269 1 * MB - RoundUp(raw_code1.size() + sizeof(OatQuickMethodHeader), kArm64CodeAlignment); in TEST_F()
1284 1 * MB - RoundUp(thunk_size + sizeof(OatQuickMethodHeader), kArm64CodeAlignment) in TEST_F()
1285 - RoundUp(kNopCode.size() + sizeof(OatQuickMethodHeader), kArm64CodeAlignment) in TEST_F()
[all …]
/art/compiler/optimizing/
Dcritical_native_abi_fixup_arm.cc35 reg = RoundUp(reg, 2u); in FixUpArguments()
/art/runtime/gc/allocator/
Drosalloc.h548 return RoundUp(size, kThreadLocalBracketQuantumSize); in RoundToBracketSize()
550 return RoundUp(size, kBracketQuantumSize); in RoundToBracketSize()
562 return RoundUp(size, kThreadLocalBracketQuantumSize) / kThreadLocalBracketQuantumSize - 1; in SizeToIndex()
564 return (RoundUp(size, kBracketQuantumSize) - kMaxThreadLocalBracketSize) / kBracketQuantumSize in SizeToIndex()
579 bracket_size = RoundUp(size, kThreadLocalBracketQuantumSize); in SizeToIndexAndBracketSize()
582 bracket_size = RoundUp(size, kBracketQuantumSize); in SizeToIndexAndBracketSize()
877 return RoundUp(bytes, gPageSize); in UsableSize()

12345