/art/runtime/gc/space/ |
D | bump_pointer_space-inl.h | 32 num_bytes = RoundUp(num_bytes, kAlignment); in Alloc() 49 num_bytes = RoundUp(num_bytes, kAlignment); in AllocThreadUnsafe() 95 return reinterpret_cast<mirror::Object*>(RoundUp(position, kAlignment)); in GetNextObject()
|
D | malloc_space.cc | 105 *growth_limit = RoundUp(*growth_limit, gPageSize); in CreateMemMap() 106 *capacity = RoundUp(*capacity, gPageSize); in CreateMemMap() 143 growth_limit = RoundUp(growth_limit, gPageSize); in SetGrowthLimit() 186 SetEnd(reinterpret_cast<uint8_t*>(RoundUp(reinterpret_cast<uintptr_t>(End()), gPageSize))); in CreateZygoteSpace() 191 size_t size = RoundUp(Size(), gPageSize); in CreateZygoteSpace() 203 SetGrowthLimit(RoundUp(size, gPageSize)); in CreateZygoteSpace()
|
D | bump_pointer_space.cc | 28 capacity = RoundUp(capacity, gPageSize); in Create() 225 bytes = RoundUp(bytes, kAlignment); in AllocNewTlab() 254 *usable_size = RoundUp(num_bytes, kAlignment); in AllocationSizeNonvirtual()
|
/art/runtime/arch/arm/ |
D | jni_frame_arm.h | 70 return RoundUp(size, kAapcsStackAlignment); in GetCriticalNativeStubFrameSize() 80 return RoundUp(size, kAapcsStackAlignment); in GetCriticalNativeDirectCallFrameSize()
|
/art/runtime/arch/x86/ |
D | jni_frame_x86.h | 66 return RoundUp(size, kNativeStackAlignment); in GetCriticalNativeStubFrameSize() 76 return RoundUp(size, kNativeStackAlignment); in GetCriticalNativeDirectCallFrameSize()
|
/art/runtime/oat/ |
D | image.cc | 72 CHECK_EQ(image_begin, RoundUp(image_begin, kElfSegmentAlignment)); in ImageHeader() 74 CHECK_EQ(oat_file_begin, RoundUp(oat_file_begin, kElfSegmentAlignment)); in ImageHeader() 75 CHECK_EQ(oat_data_begin, RoundUp(oat_data_begin, kElfSegmentAlignment)); in ImageHeader() 133 return image_reservation_size_ == RoundUp(image_size_, kElfSegmentAlignment); in IsAppImage() 191 static const size_t kStartPos = RoundUp(sizeof(ImageHeader), kObjectAlignment); in VisitObjects() 195 pos += RoundUp(object->SizeOf(), kObjectAlignment); in VisitObjects() 408 out_offset = RoundUp(out_offset, alignof(ImageHeader::Block)); in WriteData() 428 out_offset = RoundUp(out_offset, kElfSegmentAlignment); in WriteData()
|
/art/runtime/arch/arm64/ |
D | jni_frame_arm64.h | 77 return RoundUp(size, kAapcs64StackAlignment); in GetCriticalNativeStubFrameSize() 87 return RoundUp(size, kAapcs64StackAlignment); in GetCriticalNativeDirectCallFrameSize()
|
/art/runtime/arch/riscv64/ |
D | jni_frame_riscv64.h | 80 return RoundUp(size, kNativeStackAlignment); in GetCriticalNativeStubFrameSize() 90 return RoundUp(size, kNativeStackAlignment); in GetCriticalNativeDirectCallFrameSize()
|
/art/runtime/arch/x86_64/ |
D | jni_frame_x86_64.h | 83 return RoundUp(size, kNativeStackAlignment); in GetCriticalNativeStubFrameSize() 93 return RoundUp(size, kNativeStackAlignment); in GetCriticalNativeDirectCallFrameSize()
|
/art/runtime/ |
D | nterp_helpers.cc | 176 return RoundUp(NterpGetFrameSizeWithoutPadding(method, isa), kStackAlignment); in NterpGetFrameSize() 199 RoundUp(out_regs * kVRegSize, kPointerSize) + // out arguments and pointer alignment in NterpGetReferenceArray() 208 RoundUp(out_regs * kVRegSize, kPointerSize); // out arguments and pointer alignment in NterpGetDexPC() 242 DCHECK_EQ(NterpGetFrameSize(method, isa), RoundUp(frame_size_without_padding, kStackAlignment)); in CanMethodUseNterp()
|
D | linear_alloc-inl.h | 109 RoundUp(reinterpret_cast<uintptr_t>(ptr), 16) - reinterpret_cast<uintptr_t>(ptr); in AllocAlign16() 118 required_size = size + RoundUp(sizeof(TrackingHeader), 16); in AllocAlign16()
|
/art/odrefresh/ |
D | odr_fs_utils_test.cc | 145 uint64_t expected_bytes_used = RoundUp(kFirstFileBytes, sb.st_blocks * kBytesPerBlock) + in TEST_F() 146 RoundUp(kSecondFileBytes, sb.st_blocks * kBytesPerBlock); in TEST_F() 154 expected_bytes_used += RoundUp(i, sb.st_blocks * kBytesPerBlock); in TEST_F()
|
/art/libartbase/base/metrics/ |
D | metrics.h | 317 static_assert(RoundUp(sizeof(*this), sizeof(uint64_t)) 318 == RoundUp(sizeof(intptr_t) + sizeof(value_t), sizeof(uint64_t))); 355 static_assert(RoundUp(sizeof(*this), sizeof(uint64_t)) 356 == RoundUp(sizeof(intptr_t) + sizeof(value_t) + sizeof(count_t), 408 static_assert(RoundUp(sizeof(*this), sizeof(uint64_t)) == 409 RoundUp(sizeof(intptr_t) + sizeof(value_t), sizeof(uint64_t))); 452 static_assert(RoundUp(sizeof(*this), sizeof(uint64_t)) 453 == RoundUp(sizeof(intptr_t) + sizeof(value_t) * num_buckets_, sizeof(uint64_t))); 514 static_assert(RoundUp(sizeof(*this), sizeof(uint64_t)) == 515 RoundUp(sizeof(intptr_t) + sizeof(T), sizeof(uint64_t)));
|
/art/dex2oat/utils/ |
D | swap_space.cc | 111 size = RoundUp(size, 8U); in Alloc() 151 size_t next_part = std::max(RoundUp(min_size, page_size), RoundUp(kMinimumMapSize, page_size)); in NewFileChunk() 179 size = RoundUp(size, 8U); in Free()
|
/art/runtime/gc/accounting/ |
D | bitmap.cc | 49 const size_t bitmap_size = RoundUp( in AllocateMemMap() 50 RoundUp(num_bits, kBitsPerBitmapWord) / kBitsPerBitmapWord * sizeof(uintptr_t), gPageSize); in AllocateMemMap()
|
D | space_bitmap_test.cc | 143 {0, RoundUp(10 * KB, gObjectAlignment) + gObjectAlignment}, in TYPED_TEST() 147 {RoundUp(1 * KB, gObjectAlignment) + gObjectAlignment, in TYPED_TEST() 148 RoundUp(2 * KB, gObjectAlignment) + 5 * gObjectAlignment}, in TYPED_TEST()
|
/art/dex2oat/linker/arm/ |
D | relative_patcher_thumb2_test.cc | 753 uint32_t thunk_offset = GetMethodOffset(method_idx) + RoundUp(kMethodCodeSize, kArmCodeAlignment); in TestBakerFieldWide() 819 thunk_offset += RoundUp(expected_thunk.size(), kArmCodeAlignment); in TestBakerFieldWide() 854 uint32_t thunk_offset = GetMethodOffset(method_idx) + RoundUp(kMethodCodeSize, kArmCodeAlignment); in TestBakerFieldNarrow() 923 thunk_offset += RoundUp(expected_thunk.size(), kArmCodeAlignment); in TestBakerFieldNarrow() 983 RoundUp(raw_code1.size() + sizeof(OatQuickMethodHeader), kArmCodeAlignment); in TEST_F() 1004 - RoundUp(thunk_size + sizeof(OatQuickMethodHeader), kArmCodeAlignment) in TEST_F() 1005 - RoundUp(kNopCode.size() + sizeof(OatQuickMethodHeader), kArmCodeAlignment) in TEST_F() 1052 RoundUp(raw_code1.size() + sizeof(OatQuickMethodHeader), kArmCodeAlignment); in TEST_F() 1060 BneWWithOffset(kLiteralOffset1, RoundUp(raw_code1.size(), kArmCodeAlignment)); in TEST_F() 1084 RoundUp(raw_code1.size() + sizeof(OatQuickMethodHeader), kArmCodeAlignment); in TEST_F() [all …]
|
/art/compiler/jni/quick/arm/ |
D | calling_convention_arm.cc | 238 gpr_index_ = RoundUp(gpr_index_, 2u) + 2u; in Next() 256 return RoundUp(gpr_index_, 2u) + 1u < kHFCoreArgumentRegistersCount; in IsCurrentParamInRegister() 279 CHECK_EQ(RoundUp(gpr_index_, 2u), 2u); in CurrentParamRegister() 442 return RoundUp(total_size, kStackAlignment); in FrameSize() 463 size_t out_args_size = RoundUp(size, kAapcsStackAlignment); in OutFrameSize()
|
/art/libartbase/base/ |
D | arena_allocator.h | 273 bytes = RoundUp(bytes, kAlignment); 292 RoundUp(reinterpret_cast<uintptr_t>(ptr_), 16) - reinterpret_cast<uintptr_t>(ptr_); 313 const size_t aligned_ptr_size = RoundUp(ptr_size, kAlignment); 319 const size_t aligned_new_size = RoundUp(new_size, kAlignment);
|
D | scoped_arena_allocator.cc | 96 size_t rounded_bytes = RoundUp(bytes + kMemoryToolRedZoneBytes, 8); in AllocWithMemoryTool() 152 arena_stack_->top_ptr_ = mark_ptr_ + RoundUp(sizeof(ScopedArenaAllocator), 8); in Reset()
|
D | bit_utils.h | 171 constexpr T RoundUp(T x, std::remove_reference_t<T> n) WARN_UNUSED; 174 constexpr T RoundUp(T x, std::remove_reference_t<T> n) { in RoundUp() function 181 return RoundUp(x, n); in CondRoundUp() 201 return reinterpret_cast<T*>(RoundUp(reinterpret_cast<uintptr_t>(x), n)); in AlignUp() 513 return RoundUp(num_bits, kBitsPerByte) / kBitsPerByte; in BitsToBytesRoundUp()
|
/art/runtime/mirror/ |
D | array.h | 84 size_t data_offset = RoundUp(OFFSETOF_MEMBER(Array, first_element_), component_size); in DataOffset() 85 DCHECK_EQ(RoundUp(data_offset, component_size), data_offset) in DataOffset() 92 constexpr size_t data_offset = RoundUp(kFirstElementOffset, kComponentSize); in DataOffset() 93 static_assert(RoundUp(data_offset, kComponentSize) == data_offset, "RoundUp fail"); in DataOffset()
|
/art/dex2oat/linker/arm64/ |
D | relative_patcher_arm64_test.cc | 1088 GetMethodOffset(method_idx) + RoundUp(kMethodCodeSize, kArm64CodeAlignment); in TestBakerField() 1141 thunk_offset += RoundUp(expected_thunk.size(), kArm64CodeAlignment); in TestBakerField() 1180 1 * MB - RoundUp(raw_code1.size() + sizeof(OatQuickMethodHeader), kArm64CodeAlignment); in TEST_F() 1195 1 * MB - RoundUp(thunk_size + sizeof(OatQuickMethodHeader), kArm64CodeAlignment) in TEST_F() 1196 - RoundUp(kNopCode.size() + sizeof(OatQuickMethodHeader), kArm64CodeAlignment) in TEST_F() 1240 1 * MB - RoundUp(raw_code1.size() + sizeof(OatQuickMethodHeader), kArm64CodeAlignment); in TEST_F() 1247 const uint32_t cbnz_offset = RoundUp(raw_code1.size(), kArm64CodeAlignment) - kLiteralOffset1; in TEST_F() 1269 1 * MB - RoundUp(raw_code1.size() + sizeof(OatQuickMethodHeader), kArm64CodeAlignment); in TEST_F() 1284 1 * MB - RoundUp(thunk_size + sizeof(OatQuickMethodHeader), kArm64CodeAlignment) in TEST_F() 1285 - RoundUp(kNopCode.size() + sizeof(OatQuickMethodHeader), kArm64CodeAlignment) in TEST_F() [all …]
|
/art/compiler/optimizing/ |
D | critical_native_abi_fixup_arm.cc | 35 reg = RoundUp(reg, 2u); in FixUpArguments()
|
/art/runtime/gc/allocator/ |
D | rosalloc.h | 548 return RoundUp(size, kThreadLocalBracketQuantumSize); in RoundToBracketSize() 550 return RoundUp(size, kBracketQuantumSize); in RoundToBracketSize() 562 return RoundUp(size, kThreadLocalBracketQuantumSize) / kThreadLocalBracketQuantumSize - 1; in SizeToIndex() 564 return (RoundUp(size, kBracketQuantumSize) - kMaxThreadLocalBracketSize) / kBracketQuantumSize in SizeToIndex() 579 bracket_size = RoundUp(size, kThreadLocalBracketQuantumSize); in SizeToIndexAndBracketSize() 582 bracket_size = RoundUp(size, kBracketQuantumSize); in SizeToIndexAndBracketSize() 877 return RoundUp(bytes, gPageSize); in UsableSize()
|