/external/jemalloc/src/ |
D | tcache.c | 33 tcache_event_hard(tsd_t *tsd, tcache_t *tcache) in tcache_event_hard() argument 35 index_t binind = tcache->next_gc_bin; in tcache_event_hard() 36 tcache_bin_t *tbin = &tcache->tbins[binind]; in tcache_event_hard() 44 tcache_bin_flush_small(tsd, tcache, tbin, binind, in tcache_event_hard() 49 - tbin->low_water + (tbin->low_water >> 2), tcache); in tcache_event_hard() 67 tcache->next_gc_bin++; in tcache_event_hard() 68 if (tcache->next_gc_bin == nhbins) in tcache_event_hard() 69 tcache->next_gc_bin = 0; in tcache_event_hard() 70 tcache->ev_cnt = 0; in tcache_event_hard() 74 tcache_alloc_small_hard(tsd_t *tsd, arena_t *arena, tcache_t *tcache, in tcache_alloc_small_hard() argument [all …]
|
D | huge.c | 35 tcache_t *tcache) in huge_malloc() argument 45 return (huge_palloc(tsd, arena, usize, chunksize, zero, tcache)); in huge_malloc() 50 bool zero, tcache_t *tcache) in huge_palloc() argument 60 CACHELINE, false, tcache, true, arena); in huge_palloc() 82 idalloctm(tsd, node, tcache, true); in huge_palloc() 90 idalloctm(tsd, node, tcache, true); in huge_palloc() 330 size_t extra, size_t alignment, bool zero, tcache_t *tcache) in huge_ralloc() argument 346 tcache); in huge_ralloc() 348 ret = huge_malloc(tsd, arena, size + extra, zero, tcache); in huge_ralloc() 356 tcache); in huge_ralloc() [all …]
|
D | jemalloc.c | 691 tcache_t *tcache; in stats_print_atexit() local 700 ql_foreach(tcache, &arena->tcache_ql, link) { in stats_print_atexit() 701 tcache_stats_merge(tcache, arena); in stats_print_atexit() 1710 ifree(tsd_t *tsd, void *ptr, tcache_t *tcache) argument 1727 iqalloc(tsd, ptr, tcache); 1732 isfree(tsd_t *tsd, void *ptr, size_t usize, tcache_t *tcache) argument 1745 isqalloc(tsd, ptr, usize, tcache); 1891 size_t *alignment, bool *zero, tcache_t **tcache, arena_t **arena) argument 1904 *tcache = NULL; 1906 *tcache = tcaches_get(tsd, MALLOCX_TCACHE_GET(flags)); [all …]
|
D | prof.c | 770 tcache_t *tcache; in prof_lookup() local 784 tcache = tcache_get(tsd, true); in prof_lookup() 785 ret.v = iallocztm(tsd, sizeof(prof_tctx_t), false, tcache, true, in prof_lookup() 805 idalloctm(tsd, ret.v, tcache, true); in prof_lookup() 1691 tcache_t *tcache; in prof_tdata_init_impl() local 1696 tcache = tcache_get(tsd, true); in prof_tdata_init_impl() 1698 tcache, true, NULL); in prof_tdata_init_impl() 1712 idalloctm(tsd, tdata, tcache, true); in prof_tdata_init_impl() 1758 tcache_t *tcache; in prof_tdata_destroy_locked() local 1765 tcache = tcache_get(tsd, false); in prof_tdata_destroy_locked() [all …]
|
D | arena.c | 2150 bool zero, tcache_t *tcache) in arena_palloc() argument 2157 ret = arena_malloc(tsd, arena, usize, zero, tcache); in arena_palloc() 2165 ret = arena_malloc(tsd, arena, usize, zero, tcache); in arena_palloc() 2173 ret = huge_malloc(tsd, arena, usize, zero, tcache); in arena_palloc() 2176 tcache); in arena_palloc() 2625 size_t extra, size_t alignment, bool zero, tcache_t *tcache) in arena_ralloc() argument 2645 ret = ipalloct(tsd, usize, alignment, zero, tcache, in arena_ralloc() 2649 tcache); in arena_ralloc() 2661 tcache, arena); in arena_ralloc() 2664 tcache); in arena_ralloc() [all …]
|
D | ctl.c | 486 {NAME("tcache"), CHILD(named, tcache)}, 1330 tcache_t *tcache = tsd_tcache_get(tsd); in CTL_RO_BOOL_CONFIG_GEN() local 1331 if (tcache != NULL) { in CTL_RO_BOOL_CONFIG_GEN() 1332 tcache_arena_reassociate(tcache, oldarena, in CTL_RO_BOOL_CONFIG_GEN()
|
D | stats.c | 496 OPT_WRITE_BOOL(tcache) in stats_print()
|
/external/jemalloc/include/jemalloc/internal/ |
D | tcache.h | 105 tcache_t *tcache; member 139 void tcache_event_hard(tsd_t *tsd, tcache_t *tcache); 140 void *tcache_alloc_small_hard(tsd_t *tsd, arena_t *arena, tcache_t *tcache, 142 void tcache_bin_flush_small(tsd_t *tsd, tcache_t *tcache, tcache_bin_t *tbin, 145 unsigned rem, tcache_t *tcache); 146 void tcache_arena_associate(tcache_t *tcache, arena_t *arena); 147 void tcache_arena_reassociate(tcache_t *tcache, arena_t *oldarena, 149 void tcache_arena_dissociate(tcache_t *tcache, arena_t *arena); 154 void tcache_stats_merge(tcache_t *tcache, arena_t *arena); 165 void tcache_event(tsd_t *tsd, tcache_t *tcache); [all …]
|
D | jemalloc_internal.h | 814 void *iallocztm(tsd_t *tsd, size_t size, bool zero, tcache_t *tcache, 816 void *imalloct(tsd_t *tsd, size_t size, tcache_t *tcache, arena_t *arena); 818 void *icalloct(tsd_t *tsd, size_t size, tcache_t *tcache, arena_t *arena); 821 tcache_t *tcache, bool is_metadata, arena_t *arena); 823 tcache_t *tcache, arena_t *arena); 828 void idalloctm(tsd_t *tsd, void *ptr, tcache_t *tcache, bool is_metadata); 829 void idalloct(tsd_t *tsd, void *ptr, tcache_t *tcache); 831 void iqalloc(tsd_t *tsd, void *ptr, tcache_t *tcache); 832 void isdalloct(tsd_t *tsd, void *ptr, size_t size, tcache_t *tcache); 833 void isqalloc(tsd_t *tsd, void *ptr, size_t size, tcache_t *tcache); [all …]
|
D | jemalloc_internal.h.in | 162 * jemalloc can conceptually be broken into components (arena, tcache, etc.), 186 * t: tcache 365 #include "jemalloc/internal/tcache.h" 396 #include "jemalloc/internal/tcache.h" 475 #include "jemalloc/internal/tcache.h" 798 * Include portions of arena.h interleaved with tcache.h in order to resolve 804 #include "jemalloc/internal/tcache.h" 814 void *iallocztm(tsd_t *tsd, size_t size, bool zero, tcache_t *tcache, 816 void *imalloct(tsd_t *tsd, size_t size, tcache_t *tcache, arena_t *arena); 818 void *icalloct(tsd_t *tsd, size_t size, tcache_t *tcache, arena_t *arena); [all …]
|
D | huge.h | 13 tcache_t *tcache); 15 bool zero, tcache_t *tcache); 20 tcache_t *tcache); 25 void huge_dalloc(tsd_t *tsd, void *ptr, tcache_t *tcache);
|
D | arena.h | 461 size_t alignment, bool zero, tcache_t *tcache); 485 size_t size, size_t extra, size_t alignment, bool zero, tcache_t *tcache); 551 tcache_t *tcache); 554 void arena_dalloc(tsd_t *tsd, void *ptr, tcache_t *tcache); 555 void arena_sdalloc(tsd_t *tsd, void *ptr, size_t size, tcache_t *tcache); 1063 tcache_t *tcache) in arena_malloc() argument 1073 if (likely(tcache != NULL)) { in arena_malloc() 1074 return (tcache_alloc_small(tsd, arena, tcache, size, in arena_malloc() 1083 if (likely(tcache != NULL) && size <= tcache_maxclass) { in arena_malloc() 1084 return (tcache_alloc_large(tsd, arena, tcache, size, in arena_malloc() [all …]
|
D | tsd.h | 535 O(tcache, tcache_t *) \
|
/external/jemalloc/include/jemalloc/ |
D | jemalloc_macros.h.in | 23 * Bias tcache index bits so that 0 encodes "automatic tcache management", and 1
|
/external/jemalloc/ |
D | ChangeLog | 31 - Add support for explicit tcaches. The "tcache.create", "tcache.flush", and 32 "tcache.destroy" mallctls control tcache lifetime and flushing, and the 34 control which tcache is used for each operation. 133 MALLOCX_TCACHE(tc) and MALLOCX_TCACHE_NONE flags to control tcache usage. 269 + internal zero-initialized data structures (relevant to tcache and prof 363 - Disable tcache by default if running inside Valgrind, in order to avoid 373 - Fix error return value for "thread.tcache.enabled" mallctl. 401 - Add the "thread.tcache.enabled" mallctl. 413 - Rename the "tcache.flush" mallctl to "thread.tcache.flush". 461 - Fix build issues for --disable-tcache. [all …]
|
D | Android.mk | 73 src/tcache.c \
|
D | INSTALL | 143 --disable-tcache 146 the "opt.tcache" option for usage details.
|
D | configure.ac | 811 AC_ARG_ENABLE([tcache], optenable 812 [AS_HELP_STRING([--disable-tcache], [Disable per thread caches])], 1669 AC_MSG_RESULT([tcache : ${enable_tcache}])
|
D | Makefile.in | 85 $(srcroot)src/tcache.c $(srcroot)src/util.c $(srcroot)src/tsd.c
|
/external/jemalloc/test/unit/ |
D | mallctl.c | 137 TEST_MALLCTL_CONFIG(tcache); in TEST_BEGIN() 173 TEST_MALLCTL_OPT(bool, tcache, tcache); in TEST_BEGIN() 174 TEST_MALLCTL_OPT(size_t, lg_tcache_max, tcache); in TEST_BEGIN()
|
/external/valgrind/coregrind/m_syswrap/ |
D | syswrap-linux.c | 2997 unsigned *, cpu, unsigned *, node, struct vki_getcpu_cache *, tcache); in PRE()
|