Searched refs:CHUNK_CEILING (Results 1 – 7 of 7) sorted by relevance
268 CHUNK_CEILING(oldsize) - oldsize); in huge_ralloc_no_move_expand()272 CHUNK_CEILING(oldsize)), 0, usize - in huge_ralloc_no_move_expand()273 CHUNK_CEILING(oldsize)); in huge_ralloc_no_move_expand()304 if (CHUNK_CEILING(oldsize) >= CHUNK_CEILING(usize) in huge_ralloc_no_move()305 && CHUNK_CEILING(oldsize) <= CHUNK_CEILING(size+extra)) { in huge_ralloc_no_move()312 if (CHUNK_CEILING(oldsize) >= CHUNK_CEILING(usize)) { in huge_ralloc_no_move()
52 csize = CHUNK_CEILING(minsize + nsize); in base_chunk_alloc()
343 ssize_t cactive_diff = CHUNK_CEILING((arena->nactive + add_pages in arena_cactive_update()344 - sub_pages) << LG_PAGE) - CHUNK_CEILING(arena->nactive << in arena_cactive_update()811 size_t csize = CHUNK_CEILING(usize); in arena_chunk_alloc_huge()845 csize = CHUNK_CEILING(usize); in arena_chunk_dalloc_huge()869 assert(CHUNK_CEILING(oldsize) == CHUNK_CEILING(usize)); in arena_chunk_ralloc_huge_similar()894 size_t cdiff = CHUNK_CEILING(oldsize) - CHUNK_CEILING(usize); in arena_chunk_ralloc_huge_shrink()909 CHUNK_CEILING(usize)); in arena_chunk_ralloc_huge_shrink()952 void *nchunk = (void *)((uintptr_t)chunk + CHUNK_CEILING(oldsize)); in arena_chunk_ralloc_huge_expand()954 size_t cdiff = CHUNK_CEILING(usize) - CHUNK_CEILING(oldsize); in arena_chunk_ralloc_huge_expand()
75 assert(size == CHUNK_CEILING(size)); in chunk_first_best_fit()94 alloc_size = CHUNK_CEILING(s2u(size + alignment - chunksize)); in chunk_recycle()
19 #define CHUNK_CEILING(s) \ macro
724 alignment = CHUNK_CEILING(alignment); in sa2u()
724 alignment = CHUNK_CEILING(alignment);