Home
last modified time | relevance | path

Searched refs:aligned_size (Results 1 – 25 of 40) sorted by relevance

12

/external/llvm-project/clang/lib/AST/Interp/
DInterpStack.h31 new (grow(aligned_size<T>())) T(std::forward<Tys>(Args)...); in push()
39 shrink(aligned_size<T>()); in pop()
47 shrink(aligned_size<T>()); in discard()
52 return *reinterpret_cast<T *>(peek(aligned_size<T>())); in peek()
67 template <typename T> constexpr size_t aligned_size() const { in aligned_size() function
/external/arm-trusted-firmware/plat/qti/common/src/
Dqti_common.c124 size_t *aligned_size) in qti_align_mem_region() argument
127 *aligned_size = round_up(addr - *aligned_addr + size, PAGE_SIZE); in qti_align_mem_region()
134 size_t aligned_size; in qti_mmap_add_dynamic_region() local
136 qti_align_mem_region(base_pa, size, &aligned_pa, &aligned_size); in qti_mmap_add_dynamic_region()
143 return mmap_add_dynamic_region(aligned_pa, aligned_pa, aligned_size, in qti_mmap_add_dynamic_region()
/external/libvpx/libvpx/vpx_mem/
Dvpx_mem.c59 const uint64_t aligned_size = get_aligned_malloc_size(size, align); in vpx_memalign() local
60 if (!check_size_argument_overflow(1, aligned_size)) return NULL; in vpx_memalign()
62 addr = malloc((size_t)aligned_size); in vpx_memalign()
/external/libaom/libaom/aom_mem/
Daom_mem.c51 const size_t aligned_size = GetAlignedMallocSize(size, align); in aom_memalign() local
53 if (!check_size_argument_overflow(1, aligned_size)) return NULL; in aom_memalign()
55 void *const addr = malloc(aligned_size); in aom_memalign()
/external/llvm-project/compiler-rt/lib/asan/
Dasan_poisoning.h39 ALWAYS_INLINE void FastPoisonShadow(uptr aligned_beg, uptr aligned_size, in FastPoisonShadow() argument
43 __sanitizer_fill_shadow(aligned_beg, aligned_size, value, in FastPoisonShadow()
48 aligned_beg + aligned_size - SHADOW_GRANULARITY) + 1; in FastPoisonShadow()
Dasan_poisoning.cpp293 uptr aligned_size = size & ~(SHADOW_GRANULARITY - 1); in PoisonAlignedStackMemory() local
294 PoisonShadow(addr, aligned_size, in PoisonAlignedStackMemory()
296 if (size == aligned_size) in PoisonAlignedStackMemory()
298 s8 end_offset = (s8)(size - aligned_size); in PoisonAlignedStackMemory()
299 s8* shadow_end = (s8*)MemToShadow(addr + aligned_size); in PoisonAlignedStackMemory()
Dasan_globals.cpp64 uptr aligned_size = RoundUpTo(g.size, SHADOW_GRANULARITY); in PoisonRedZones() local
65 FastPoisonShadow(g.beg + aligned_size, g.size_with_redzone - aligned_size, in PoisonRedZones()
67 if (g.size != aligned_size) { in PoisonRedZones()
/external/tensorflow/tensorflow/lite/delegates/gpu/common/task/
Dtensor_linear_desc.cc100 int aligned_size) { in UploadLinearData() argument
101 size = aligned_size == 0 ? DivideRoundUp(src.shape.v, 4) : aligned_size; in UploadLinearData()
Dtensor_linear_desc.h49 int aligned_size = 0);
/external/libchrome/base/files/
Dmemory_mapped_file_posix.cc53 size_t aligned_size = 0; in MapFileRegionToMemory() local
57 &aligned_size, in MapFileRegionToMemory()
68 map_size = aligned_size; in MapFileRegionToMemory()
Dmemory_mapped_file.cc119 size_t* aligned_size, in CalculateVMAlignedBoundaries() argument
126 *aligned_size = (size + *offset + mask) & ~mask; in CalculateVMAlignedBoundaries()
/external/google-breakpad/src/client/
Dminidump_file_writer.cc287 size_t aligned_size = (size + 7) & ~7; // 64-bit alignment in Allocate() local
289 if (position_ + aligned_size > size_) { in Allocate()
290 size_t growth = aligned_size; in Allocate()
305 position_ += static_cast<MDRVA>(aligned_size); in Allocate()
/external/mesa3d/src/mesa/main/
Dglthread_marshal.h68 const int aligned_size = align(size, 8); in _mesa_glthread_allocate_command() local
70 next->used += aligned_size; in _mesa_glthread_allocate_command()
72 cmd_base->cmd_size = aligned_size; in _mesa_glthread_allocate_command()
/external/compiler-rt/lib/asan/
Dasan_poisoning.h39 ALWAYS_INLINE void FastPoisonShadow(uptr aligned_beg, uptr aligned_size, in FastPoisonShadow() argument
44 aligned_beg + aligned_size - SHADOW_GRANULARITY) + 1; in FastPoisonShadow()
Dasan_poisoning.cc296 uptr aligned_size = size & ~(SHADOW_GRANULARITY - 1); in PoisonAlignedStackMemory() local
297 PoisonShadow(addr, aligned_size, in PoisonAlignedStackMemory()
299 if (size == aligned_size) in PoisonAlignedStackMemory()
301 s8 end_offset = (s8)(size - aligned_size); in PoisonAlignedStackMemory()
302 s8* shadow_end = (s8*)MemToShadow(addr + aligned_size); in PoisonAlignedStackMemory()
Dasan_globals.cc64 uptr aligned_size = RoundUpTo(g.size, SHADOW_GRANULARITY); in PoisonRedZones() local
65 FastPoisonShadow(g.beg + aligned_size, g.size_with_redzone - aligned_size, in PoisonRedZones()
67 if (g.size != aligned_size) { in PoisonRedZones()
/external/python/cpython2/Modules/_ctypes/libffi/src/microblaze/
Dffi.c77 int aligned_size = WORD_ALIGN(size); in ffi_prep_args() local
80 stack_args_p += aligned_size; in ffi_prep_args()
133 memcpy(addr, value, aligned_size); in ffi_prep_args()
/external/libffi/src/microblaze/
Dffi.c77 int aligned_size = WORD_FFI_ALIGN(size); in ffi_prep_args() local
80 stack_args_p += aligned_size; in ffi_prep_args()
133 memcpy(addr, value, aligned_size); in ffi_prep_args()
/external/skqp/src/core/
DSkMask.cpp41 size_t aligned_size = SkSafeMath::Align4(size); in AllocImage() local
46 return static_cast<uint8_t*>(sk_malloc_flags(aligned_size, flags)); in AllocImage()
/external/skia/src/core/
DSkMask.cpp41 size_t aligned_size = SkSafeMath::Align4(size); in AllocImage() local
46 return static_cast<uint8_t*>(sk_malloc_flags(aligned_size, flags)); in AllocImage()
/external/mesa3d/src/broadcom/compiler/
Dv3d_nir_lower_robust_buffer_access.c97 nir_ssa_def *aligned_size = in lower_shared() local
99 nir_ssa_def *offset = nir_umin(b, instr->src[0].ssa, aligned_size); in lower_shared()
/external/tensorflow/tensorflow/lite/micro/
Dmemory_helpers.cc45 size_t aligned_size = (((size + (alignment - 1)) / alignment) * alignment); in AlignSizeUp() local
46 return aligned_size; in AlignSizeUp()
/external/llvm-project/compiler-rt/lib/memprof/
Dmemprof_malloc_linux.cpp73 uptr aligned_size = RoundUpTo(size_in_bytes, kWordSize); in PosixMemalignFromLocalPool() local
75 uptr *end_mem = (uptr *)(aligned_addr + aligned_size); in PosixMemalignFromLocalPool()
/external/libchrome/base/
Dpickle.cc44 size_t aligned_size = bits::Align(size, sizeof(uint32_t)); in Advance() local
45 if (end_index_ - read_index_ < aligned_size) { in Advance()
48 read_index_ += aligned_size; in Advance()
/external/mesa3d/src/gallium/drivers/radeonsi/
Dsi_compute_blit.c328 uint64_t aligned_size = size & ~3ull; in si_clear_buffer() local
329 if (aligned_size >= 4) { in si_clear_buffer()
355 si_compute_do_clear_or_copy(sctx, dst, offset, NULL, 0, aligned_size, clear_value, in si_clear_buffer()
359 si_cp_dma_clear_buffer(sctx, sctx->gfx_cs, dst, offset, aligned_size, *clear_value, 0, in si_clear_buffer()
363 offset += aligned_size; in si_clear_buffer()
364 size -= aligned_size; in si_clear_buffer()

12