Home
last modified time | relevance | path

Searched refs:CHUNK_CEILING (Results 1 – 7 of 7) sorted by relevance

/external/jemalloc/src/
Dhuge.c268 CHUNK_CEILING(oldsize) - oldsize); in huge_ralloc_no_move_expand()
272 CHUNK_CEILING(oldsize)), 0, usize - in huge_ralloc_no_move_expand()
273 CHUNK_CEILING(oldsize)); in huge_ralloc_no_move_expand()
304 if (CHUNK_CEILING(oldsize) >= CHUNK_CEILING(usize) in huge_ralloc_no_move()
305 && CHUNK_CEILING(oldsize) <= CHUNK_CEILING(size+extra)) { in huge_ralloc_no_move()
312 if (CHUNK_CEILING(oldsize) >= CHUNK_CEILING(usize)) { in huge_ralloc_no_move()
Dbase.c52 csize = CHUNK_CEILING(minsize + nsize); in base_chunk_alloc()
Darena.c343 ssize_t cactive_diff = CHUNK_CEILING((arena->nactive + add_pages in arena_cactive_update()
344 - sub_pages) << LG_PAGE) - CHUNK_CEILING(arena->nactive << in arena_cactive_update()
811 size_t csize = CHUNK_CEILING(usize); in arena_chunk_alloc_huge()
845 csize = CHUNK_CEILING(usize); in arena_chunk_dalloc_huge()
869 assert(CHUNK_CEILING(oldsize) == CHUNK_CEILING(usize)); in arena_chunk_ralloc_huge_similar()
894 size_t cdiff = CHUNK_CEILING(oldsize) - CHUNK_CEILING(usize); in arena_chunk_ralloc_huge_shrink()
909 CHUNK_CEILING(usize)); in arena_chunk_ralloc_huge_shrink()
952 void *nchunk = (void *)((uintptr_t)chunk + CHUNK_CEILING(oldsize)); in arena_chunk_ralloc_huge_expand()
954 size_t cdiff = CHUNK_CEILING(usize) - CHUNK_CEILING(oldsize); in arena_chunk_ralloc_huge_expand()
Dchunk.c75 assert(size == CHUNK_CEILING(size)); in chunk_first_best_fit()
94 alloc_size = CHUNK_CEILING(s2u(size + alignment - chunksize)); in chunk_recycle()
/external/jemalloc/include/jemalloc/internal/
Dchunk.h19 #define CHUNK_CEILING(s) \ macro
Djemalloc_internal.h724 alignment = CHUNK_CEILING(alignment); in sa2u()
Djemalloc_internal.h.in724 alignment = CHUNK_CEILING(alignment);