/external/jemalloc/include/jemalloc/internal/ |
D | bitmap.h | 6 #define BITMAP_MAXBITS (ZU(1) << LG_BITMAP_MAXBITS) 15 #define BITMAP_GROUP_NBITS (ZU(1) << LG_BITMAP_GROUP_NBITS) 162 return (!(g & (ZU(1) << (bit & BITMAP_GROUP_NBITS_MASK)))); in bitmap_get() 177 assert(g & (ZU(1) << (bit & BITMAP_GROUP_NBITS_MASK))); in bitmap_set() 178 g ^= ZU(1) << (bit & BITMAP_GROUP_NBITS_MASK); in bitmap_set() 190 assert(g & (ZU(1) << (bit & BITMAP_GROUP_NBITS_MASK))); in bitmap_set() 191 g ^= ZU(1) << (bit & BITMAP_GROUP_NBITS_MASK); in bitmap_set() 246 assert((g & (ZU(1) << (bit & BITMAP_GROUP_NBITS_MASK))) == 0); in bitmap_unset() 247 g ^= ZU(1) << (bit & BITMAP_GROUP_NBITS_MASK); in bitmap_unset() 260 assert((g & (ZU(1) << (bit & BITMAP_GROUP_NBITS_MASK))) in bitmap_unset() [all …]
|
D | jemalloc_internal.h | 214 (ZU(1) << (flags & MALLOCX_LG_ALIGN_MASK)) 603 ((ZU(1) << LG_SIZE_CLASS_GROUP) - 1); in psz2ind() 616 size_t mod = pind & ((ZU(1) << LG_SIZE_CLASS_GROUP) - 1); in pind2sz_compute() 619 size_t grp_size = ((ZU(1) << (LG_PAGE + in pind2sz_compute() 657 size_t delta = ZU(1) << lg_delta; in psz2u() 671 if (size <= (ZU(1) << LG_TINY_MAXCLASS)) { in size2index_compute() 688 ((ZU(1) << LG_SIZE_CLASS_GROUP) - 1); in size2index_compute() 723 return (ZU(1) << (LG_TINY_MAXCLASS - NTBINS + 1 + index)); in index2size_compute() 728 size_t mod = reduced_index & ((ZU(1) << LG_SIZE_CLASS_GROUP) - in index2size_compute() 732 size_t grp_size = ((ZU(1) << (LG_QUANTUM + in index2size_compute() [all …]
|
D | jemalloc_internal_macros.h | 41 #define ZU(z) ((size_t)z) macro 46 #define KZU(z) ZU(z##ULL)
|
D | prng.h | 130 assert(lg_range <= ZU(1) << (3 + LG_SIZEOF_PTR)); in prng_lg_range_zu() 143 ret = state1 >> ((ZU(1) << (3 + LG_SIZEOF_PTR)) - lg_range); in prng_lg_range_zu()
|
D | jemalloc_internal.h.in | 214 (ZU(1) << (flags & MALLOCX_LG_ALIGN_MASK)) 603 ((ZU(1) << LG_SIZE_CLASS_GROUP) - 1); 616 size_t mod = pind & ((ZU(1) << LG_SIZE_CLASS_GROUP) - 1); 619 size_t grp_size = ((ZU(1) << (LG_PAGE + 657 size_t delta = ZU(1) << lg_delta; 671 if (size <= (ZU(1) << LG_TINY_MAXCLASS)) { 688 ((ZU(1) << LG_SIZE_CLASS_GROUP) - 1); 723 return (ZU(1) << (LG_TINY_MAXCLASS - NTBINS + 1 + index)); 728 size_t mod = reduced_index & ((ZU(1) << LG_SIZE_CLASS_GROUP) - 732 size_t grp_size = ((ZU(1) << (LG_QUANTUM + [all …]
|
D | quarantine.h | 8 #define JEMALLOC_VALGRIND_QUARANTINE_DEFAULT (ZU(1) << 24)
|
D | rtree.h | 151 return ((key >> ((ZU(1) << (LG_SIZEOF_PTR+3)) - in rtree_subkey() 152 rtree->levels[level].cumbits)) & ((ZU(1) << in rtree_subkey()
|
/external/jemalloc/test/unit/ |
D | prng.c | 84 ra = prng_lg_range_zu(&sa, ZU(1) << (3 + LG_SIZEOF_PTR), atomic); in test_prng_lg_range_zu() 86 rb = prng_lg_range_zu(&sa, ZU(1) << (3 + LG_SIZEOF_PTR), atomic); in test_prng_lg_range_zu() 91 rb = prng_lg_range_zu(&sb, ZU(1) << (3 + LG_SIZEOF_PTR), atomic); in test_prng_lg_range_zu() 96 ra = prng_lg_range_zu(&sa, ZU(1) << (3 + LG_SIZEOF_PTR), atomic); in test_prng_lg_range_zu() 97 rb = prng_lg_range_zu(&sa, ZU(1) << (3 + LG_SIZEOF_PTR), atomic); in test_prng_lg_range_zu() 102 ra = prng_lg_range_zu(&sa, ZU(1) << (3 + LG_SIZEOF_PTR), atomic); in test_prng_lg_range_zu() 103 for (lg_range = (ZU(1) << (3 + LG_SIZEOF_PTR)) - 1; lg_range > 0; in test_prng_lg_range_zu() 109 assert_zu_eq(rb, (ra >> ((ZU(1) << (3 + LG_SIZEOF_PTR)) - in test_prng_lg_range_zu()
|
D | size_classes.c | 148 assert_u_eq(size2index(ZU(PTRDIFF_MAX)+1), NSIZES, in TEST_BEGIN() 155 assert_zu_eq(s2u(ZU(PTRDIFF_MAX)+1), 0, in TEST_BEGIN() 162 assert_u_eq(psz2ind(ZU(PTRDIFF_MAX)+1), NPSIZES, in TEST_BEGIN() 169 assert_zu_eq(psz2u(ZU(PTRDIFF_MAX)+1), 0, in TEST_BEGIN()
|
D | ckh.c | 38 "ckh_count() should return %zu, but it returned %zu", ZU(0), in TEST_BEGIN() 110 #define NITEMS ZU(1000) in TEST_BEGIN() 200 ZU(0), ckh_count(&ckh)); in TEST_BEGIN()
|
D | zero.c | 24 ZU(0), sz_prev); in test_zero()
|
D | quarantine.c | 24 #define SZ ZU(256) in TEST_BEGIN()
|
D | pack.c | 15 #define SZ (ZU(1) << (LG_PAGE - 2))
|
D | junk.c | 93 ZU(0), sz_prev); in test_junk()
|
/external/jemalloc/test/integration/ |
D | rallocx.c | 57 #define MAXSZ ZU(12 * 1024 * 1024) in TEST_BEGIN() 163 #define MAX_ALIGN (ZU(1) << 25) in TEST_BEGIN() 165 align = ZU(1); in TEST_BEGIN() 190 #define MAX_VALIDATE (ZU(1) << 22) in TEST_BEGIN() 201 (void *)((uintptr_t)q & ((ZU(1) << lg_align)-1)), in TEST_BEGIN() 235 assert_ptr_null(rallocx(p, ZU(PTRDIFF_MAX)+1, 0), in TEST_BEGIN() 236 "Expected OOM for rallocx(p, size=%#zx, 0)", ZU(PTRDIFF_MAX)+1); in TEST_BEGIN() 241 assert_ptr_null(rallocx(p, 1, MALLOCX_ALIGN(ZU(PTRDIFF_MAX)+1)), in TEST_BEGIN() 243 ZU(PTRDIFF_MAX)+1); in TEST_BEGIN()
|
D | mallocx.c | 75 assert_ptr_null(mallocx(ZU(PTRDIFF_MAX)+1, 0), in TEST_BEGIN() 76 "Expected OOM for mallocx(size=%#zx, 0)", ZU(PTRDIFF_MAX)+1); in TEST_BEGIN() 81 assert_ptr_null(mallocx(1, MALLOCX_ALIGN(ZU(PTRDIFF_MAX)+1)), in TEST_BEGIN() 83 ZU(PTRDIFF_MAX)+1); in TEST_BEGIN()
|
/external/jemalloc/src/ |
D | quarantine.c | 28 size = offsetof(quarantine_t, objs) + ((ZU(1) << lg_maxobjs) * in quarantine_init() 74 if (quarantine->first + quarantine->curobjs <= (ZU(1) << in quarantine_grow() 81 size_t ncopy_a = (ZU(1) << quarantine->lg_maxobjs) - in quarantine_grow() 104 quarantine->first = (quarantine->first + 1) & ((ZU(1) << in quarantine_drain_one() 139 if (quarantine->curobjs == (ZU(1) << quarantine->lg_maxobjs)) in quarantine() 142 assert(quarantine->curobjs < (ZU(1) << quarantine->lg_maxobjs)); in quarantine() 146 ((ZU(1) << quarantine->lg_maxobjs) - 1); in quarantine()
|
D | ckh.c | 58 for (i = 0; i < (ZU(1) << LG_CKH_BUCKET_CELLS); i++) { in ckh_bucket_search() 80 bucket = hashes[0] & ((ZU(1) << ckh->lg_curbuckets) - 1); in ckh_isearch() 86 bucket = hashes[1] & ((ZU(1) << ckh->lg_curbuckets) - 1); in ckh_isearch() 104 for (i = 0; i < (ZU(1) << LG_CKH_BUCKET_CELLS); i++) { in ckh_try_bucket_insert() 106 ((i + offset) & ((ZU(1) << LG_CKH_BUCKET_CELLS) - 1))]; in ckh_try_bucket_insert() 161 tbucket = hashes[1] & ((ZU(1) << ckh->lg_curbuckets) - 1); in ckh_evict_reloc_insert() 163 tbucket = hashes[0] & ((ZU(1) << ckh->lg_curbuckets) in ckh_evict_reloc_insert() 205 bucket = hashes[0] & ((ZU(1) << ckh->lg_curbuckets) - 1); in ckh_try_insert() 210 bucket = hashes[1] & ((ZU(1) << ckh->lg_curbuckets) - 1); in ckh_try_insert() 382 (ZU(1) << lg_mincells) < mincells; in ckh_new() [all …]
|
D | rtree.c | 18 assert(RTREE_HEIGHT_MAX == ((ZU(1) << (LG_SIZEOF_PTR+3)) / in rtree_new() 71 nchildren = ZU(1) << rtree->levels[level].bits; in rtree_delete_subtree() 111 node = rtree->alloc(ZU(1) << rtree->levels[level].bits); in rtree_node_init()
|
D | bitmap.c | 15 assert(nbits <= (ZU(1) << LG_BITMAP_MAXBITS)); in bitmap_info_init() 79 assert(nbits <= (ZU(1) << LG_BITMAP_MAXBITS)); in bitmap_info_init()
|
D | tcache.c | 559 if (opt_lg_tcache_max < 0 || (ZU(1) << opt_lg_tcache_max) < SMALL_MAXCLASS) in tcache_boot() 561 else if ((ZU(1) << opt_lg_tcache_max) > large_maxclass) in tcache_boot() 564 tcache_maxclass = (ZU(1) << opt_lg_tcache_max); in tcache_boot()
|
D | jemalloc.c | 42 size_t opt_quarantine = ZU(0); 95 (((ZU(1)<<lg_grp) + (ZU(ndelta)<<lg_delta))), 108 ((ZU(1)<<lg_grp) + (ZU(ndelta)<<lg_delta)), 1748 assert(((uintptr_t)result & (alignment - 1)) == ZU(0)); 2217 assert(alignment == 0 || ((uintptr_t)p & (alignment - 1)) == ZU(0)); 2236 assert(alignment == 0 || ((uintptr_t)p & (alignment - 1)) == ZU(0)); 2427 assert(alignment == 0 || ((uintptr_t)p & (alignment - 1)) == ZU(0));
|
/external/ImageMagick/PerlMagick/t/reference/filter/ |
D | Level.miff | 340 #FHg4A_5XS:XS:XS:ZU<ZU<\S:_PDbWAXZ2_nf���������������������������������������������������������… 347 &DX72F/WR9XS:ZU<ZU<ZU<\S:bPD_XAU_5Xqd����������������������������������������������������…
|
D | Minify.miff | 15 …VJ;TH:WL?hbp�����̜x��Ui�BG�@>�?:�>4�;,�7-�3.�1/�23�6/�60�9/�;2�>7�?3�</�?3�ZU�ls�fq�sk����ӽ�ط�ܾ���…
|
D | Scale.miff | 15 ….NC5q=.�@1�E>�??�@7�?,�>-�>+�=,�<-�>/�B5�MF`NC;5153./3..2/11/4..^?9�z�ZX�E@ZU][Z_S,.1-...,),***)*,…
|