Home
last modified time | relevance | path

Searched refs:arena_t (Results 1 – 25 of 29) sorted by relevance

12

/external/jemalloc_new/include/jemalloc/internal/
Darena_externs.h19 void arena_basic_stats_merge(tsdn_t *tsdn, arena_t *arena,
22 void arena_stats_merge(tsdn_t *tsdn, arena_t *arena, unsigned *nthreads,
26 void arena_extents_dirty_dalloc(tsdn_t *tsdn, arena_t *arena,
31 extent_t *arena_extent_alloc_large(tsdn_t *tsdn, arena_t *arena,
33 void arena_extent_dalloc_large_prep(tsdn_t *tsdn, arena_t *arena,
35 void arena_extent_ralloc_large_shrink(tsdn_t *tsdn, arena_t *arena,
37 void arena_extent_ralloc_large_expand(tsdn_t *tsdn, arena_t *arena,
39 ssize_t arena_dirty_decay_ms_get(arena_t *arena);
40 bool arena_dirty_decay_ms_set(tsdn_t *tsdn, arena_t *arena, ssize_t decay_ms);
41 ssize_t arena_muzzy_decay_ms_get(arena_t *arena);
[all …]
Dextent_externs.h15 extent_t *extent_alloc(tsdn_t *tsdn, arena_t *arena);
16 void extent_dalloc(tsdn_t *tsdn, arena_t *arena, extent_t *extent);
18 extent_hooks_t *extent_hooks_get(arena_t *arena);
19 extent_hooks_t *extent_hooks_set(tsd_t *tsd, arena_t *arena,
34 extent_t *extents_alloc(tsdn_t *tsdn, arena_t *arena,
38 void extents_dalloc(tsdn_t *tsdn, arena_t *arena,
40 extent_t *extents_evict(tsdn_t *tsdn, arena_t *arena,
45 extent_t *extent_alloc_wrapper(tsdn_t *tsdn, arena_t *arena,
48 void extent_dalloc_gap(tsdn_t *tsdn, arena_t *arena, extent_t *extent);
49 void extent_dalloc_wrapper(tsdn_t *tsdn, arena_t *arena,
[all …]
Djemalloc_internal_inlines_b.h7 static inline arena_t *
8 arena_choose_impl(tsd_t *tsd, arena_t *arena, bool internal) { in arena_choose_impl()
9 arena_t *ret; in arena_choose_impl()
61 static inline arena_t *
62 arena_choose(tsd_t *tsd, arena_t *arena) { in arena_choose()
66 static inline arena_t *
67 arena_ichoose(tsd_t *tsd, arena_t *arena) { in arena_ichoose()
72 arena_is_auto(arena_t *arena) { in arena_is_auto()
Darena_inlines_a.h5 arena_ind_get(const arena_t *arena) { in arena_ind_get()
10 arena_internal_add(arena_t *arena, size_t size) { in arena_internal_add()
15 arena_internal_sub(arena_t *arena, size_t size) { in arena_internal_sub()
20 arena_internal_get(arena_t *arena) { in arena_internal_get()
25 arena_prof_accum(tsdn_t *tsdn, arena_t *arena, uint64_t accumbytes) { in arena_prof_accum()
38 arena_t *oldarena = tsd_arena_get(tsd); in percpu_arena_update()
44 arena_t *newarena = arena_get(tsd_tsdn(tsd), newind, true); in percpu_arena_update()
Dtcache_externs.h32 void *tcache_alloc_small_hard(tsdn_t *tsdn, arena_t *arena, tcache_t *tcache,
39 arena_t *arena);
42 void tcache_stats_merge(tsdn_t *tsdn, tcache_t *tcache, arena_t *arena);
47 void tcache_arena_associate(tsdn_t *tsdn, tcache_t *tcache, arena_t *arena);
Djemalloc_internal_externs.h39 void arena_set(unsigned ind, arena_t *arena);
41 arena_t *arena_init(tsdn_t *tsdn, unsigned ind, extent_hooks_t *extent_hooks);
43 arena_t *arena_choose_hard(tsd_t *tsd, bool internal);
Dlarge_externs.h4 void *large_malloc(tsdn_t *tsdn, arena_t *arena, size_t usize, bool zero);
5 void *large_palloc(tsdn_t *tsdn, arena_t *arena, size_t usize, size_t alignment,
9 void *large_ralloc(tsdn_t *tsdn, arena_t *arena, extent_t *extent, size_t usize,
Djemalloc_internal_inlines_a.h82 static inline arena_t *
84 arena_t *ret; in arena_get()
88 ret = (arena_t *)atomic_load_p(&arenas[ind], ATOMIC_ACQUIRE); in arena_get()
149 pre_reentrancy(tsd_t *tsd, arena_t *arena) { in pre_reentrancy()
Djemalloc_internal_inlines_c.h26 JEMALLOC_ALWAYS_INLINE arena_t *
42 bool is_internal, arena_t *arena, bool slow_path) { in iallocztm()
68 tcache_t *tcache, bool is_internal, arena_t *arena) { in ipallocztm()
88 tcache_t *tcache, arena_t *arena) { in ipalloct()
137 arena_t *arena) { in iralloct_realign()
174 bool zero, tcache_t *tcache, arena_t *arena) { in iralloct()
Dtsd.h73 O(iarena, arena_t *, arena_t *) \
74 O(arena, arena_t *, arena_t *) \
Dbackground_thread_inlines.h16 arena_background_thread_info_get(arena_t *arena) { in arena_background_thread_info_get()
44 arena_background_thread_inactivity_check(tsdn_t *tsdn, arena_t *arena, in arena_background_thread_inactivity_check()
Darena_inlines_b.h61 arena_decay_ticks(tsdn_t *tsdn, arena_t *arena, unsigned nticks) { in arena_decay_ticks()
79 arena_decay_tick(tsdn_t *tsdn, arena_t *arena) { in arena_decay_tick()
87 arena_malloc(tsdn_t *tsdn, arena_t *arena, size_t size, szind_t ind, bool zero, in arena_malloc()
108 JEMALLOC_ALWAYS_INLINE arena_t *
Dextent_inlines.h37 static inline arena_t *
49 return (arena_t *)atomic_load_p(&arenas[arena_ind], ATOMIC_ACQUIRE); in extent_arena_get()
180 extent_arena_set(extent_t *extent, arena_t *arena) { in extent_arena_set()
303 extent_init(extent_t *extent, arena_t *arena, void *addr, size_t size, in extent_init()
Dtcache_structs.h40 arena_t *arena;
Dextent_dss.h20 void *extent_alloc_dss(tsdn_t *tsdn, arena_t *arena, void *new_addr,
Darena_types.h21 typedef struct arena_s arena_t; typedef
Dtcache_inlines.h42 tcache_alloc_small(tsd_t *tsd, arena_t *arena, tcache_t *tcache, in tcache_alloc_small()
106 tcache_alloc_large(tsd_t *tsd, arena_t *arena, tcache_t *tcache, size_t size, in tcache_alloc_large()
Dbackground_thread_externs.h16 void background_thread_interval_check(tsdn_t *tsdn, arena_t *arena,
/external/jemalloc_new/src/
Darena.c51 static void arena_decay_to_limit(tsdn_t *tsdn, arena_t *arena,
54 static bool arena_decay_dirty(tsdn_t *tsdn, arena_t *arena,
56 static void arena_dalloc_bin_slab(tsdn_t *tsdn, arena_t *arena, extent_t *slab,
58 static void arena_bin_lower_slab(tsdn_t *tsdn, arena_t *arena, extent_t *slab,
64 arena_basic_stats_merge(UNUSED tsdn_t *tsdn, arena_t *arena, unsigned *nthreads, in arena_basic_stats_merge()
77 arena_stats_merge(tsdn_t *tsdn, arena_t *arena, unsigned *nthreads, in arena_stats_merge()
209 arena_extents_dirty_dalloc(tsdn_t *tsdn, arena_t *arena, in arena_extents_dirty_dalloc()
278 arena_nactive_add(arena_t *arena, size_t add_pages) { in arena_nactive_add()
283 arena_nactive_sub(arena_t *arena, size_t sub_pages) { in arena_nactive_sub()
289 arena_large_malloc_stats_update(tsdn_t *tsdn, arena_t *arena, size_t usize) { in arena_large_malloc_stats_update()
[all …]
Dextent.c34 static bool extent_commit_impl(tsdn_t *tsdn, arena_t *arena,
43 static bool extent_purge_lazy_impl(tsdn_t *tsdn, arena_t *arena,
50 static bool extent_purge_forced_impl(tsdn_t *tsdn, arena_t *arena,
58 static extent_t *extent_split_impl(tsdn_t *tsdn, arena_t *arena,
67 static bool extent_merge_impl(tsdn_t *tsdn, arena_t *arena,
109 static extent_t *extent_recycle(tsdn_t *tsdn, arena_t *arena,
113 static extent_t *extent_try_coalesce(tsdn_t *tsdn, arena_t *arena,
116 static void extent_record(tsdn_t *tsdn, arena_t *arena,
177 extent_alloc(tsdn_t *tsdn, arena_t *arena) { in extent_alloc()
190 extent_dalloc(tsdn_t *tsdn, arena_t *arena, extent_t *extent) { in extent_dalloc()
[all …]
Dandroid_je_mallinfo.c17 static size_t accumulate_large_allocs(arena_t* arena) { in accumulate_large_allocs()
34 static size_t accumulate_small_allocs(arena_t* arena) { in accumulate_small_allocs()
57 arena_t* arena = atomic_load_p(&arenas[i], ATOMIC_ACQUIRE); in je_mallinfo()
85 arena_t* arena = atomic_load_p(&arenas[aidx], ATOMIC_ACQUIRE); in je_mallinfo_arena_info()
102 arena_t* arena = atomic_load_p(&arenas[aidx], ATOMIC_ACQUIRE); in je_mallinfo_bin_info()
Dlarge.c14 large_malloc(tsdn_t *tsdn, arena_t *arena, size_t usize, bool zero) { in large_malloc()
21 large_palloc(tsdn_t *tsdn, arena_t *arena, size_t usize, size_t alignment, in large_palloc()
105 arena_t *arena = extent_arena_get(extent); in large_ralloc_no_move_shrink()
141 arena_t *arena = extent_arena_get(extent); in large_ralloc_no_move_expand()
272 large_ralloc_move_helper(tsdn_t *tsdn, arena_t *arena, size_t usize, in large_ralloc_move_helper()
281 large_ralloc(tsdn_t *tsdn, arena_t *arena, extent_t *extent, size_t usize, in large_ralloc()
317 large_dalloc_prep_impl(tsdn_t *tsdn, arena_t *arena, extent_t *extent, in large_dalloc_prep_impl()
338 large_dalloc_finish_impl(tsdn_t *tsdn, arena_t *arena, extent_t *extent) { in large_dalloc_finish_impl()
355 arena_t *arena = extent_arena_get(extent); in large_dalloc()
Dtcache.c92 tcache_alloc_small_hard(tsdn_t *tsdn, arena_t *arena, tcache_t *tcache, in tcache_alloc_small_hard()
115 arena_t *arena = tcache->arena; in tcache_bin_flush_small()
127 arena_t *bin_arena = extent_arena_get(extent); in tcache_bin_flush_small()
206 arena_t *arena = tcache->arena; in tcache_bin_flush_large()
218 arena_t *locked_arena = extent_arena_get(extent); in tcache_bin_flush_large()
301 tcache_arena_associate(tsdn_t *tsdn, tcache_t *tcache, arena_t *arena) { in tcache_arena_associate()
323 arena_t *arena = tcache->arena; in tcache_arena_dissociate()
349 tcache_arena_reassociate(tsdn_t *tsdn, tcache_t *tcache, arena_t *arena) { in tcache_arena_reassociate()
427 arena_t *arena; in tsd_tcache_data_init()
547 tcache_stats_merge(tsdn_t *tsdn, tcache_t *tcache, arena_t *arena) { in tcache_stats_merge()
Djemalloc.c88 static arena_t *a0; /* arenas[0]; read-only after initialization. */
291 arena_set(unsigned ind, arena_t *arena) { in arena_set()
311 static arena_t *
313 arena_t *arena; in arena_init_locked()
357 arena_t *
359 arena_t *arena; in arena_init()
372 arena_t *arena = arena_get(tsd_tsdn(tsd), ind, false); in arena_bind()
384 arena_t *oldarena, *newarena; in arena_migrate()
395 arena_t *arena; in arena_unbind()
484 arena_t *
[all …]
Dctl.c707 ctl_arena_stats_amerge(tsdn_t *tsdn, ctl_arena_t *ctl_arena, arena_t *arena) { in ctl_arena_stats_amerge()
874 ctl_arena_refresh(tsdn_t *tsdn, arena_t *arena, ctl_arena_t *ctl_sdarena, in ctl_arena_refresh()
928 VARIABLE_ARRAY(arena_t *, tarenas, ctl_arenas->narenas); in ctl_refresh()
1696 arena_t *oldarena; in CTL_RO_CONFIG_GEN()
1708 arena_t *newarena; in CTL_RO_CONFIG_GEN()
1953 VARIABLE_ARRAY(arena_t *, tarenas, narenas); in arena_i_decay()
1972 arena_t *tarena; in arena_i_decay()
2023 arena_t **arena) { in arena_i_reset_destroy_helper()
2079 arena_t *arena; in arena_i_reset_ctl()
2099 arena_t *arena; in arena_i_destroy_ctl()
[all …]

12