/external/valgrind/tests/ |
D | malloc.h | 14 static void* memalign16(size_t szB) in memalign16() argument 19 posix_memalign((void **)&x, 16, szB); in memalign16() 21 x = memalign(16, szB); in memalign16() 30 static void* memalign32(size_t szB) in memalign32() argument 35 posix_memalign((void **)&x, 32, szB); in memalign32() 37 x = memalign(32, szB); in memalign32() 46 static void* memalign64(size_t szB) in memalign64() argument 51 posix_memalign((void **)&x, 64, szB); in memalign64() 53 x = memalign(64, szB); in memalign64()
|
/external/valgrind/exp-sgcheck/ |
D | sg_main.c | 174 if (fb1->szB < fb2->szB) return -1; in StackBlock__cmp() 175 if (fb1->szB > fb2->szB) return 1; in StackBlock__cmp() 230 sb->base, sb->szB, sb->spRel ? 'Y' : 'N', in pp_StackBlocks() 314 tl_assert(sb0->szB < sb1->szB); in StackBlocks__find_and_dealloc__or_add() 315 sb0->szB = sb1->szB; in StackBlocks__find_and_dealloc__or_add() 343 || sb1->base + sb1->szB > sb2->base)) { in StackBlocks__find_and_dealloc__or_add() 403 if (gb1->szB < gb2->szB) return -1; in GlobalBlock__cmp() 404 if (gb1->szB > gb2->szB) return 1; in GlobalBlock__cmp() 473 SizeT szB; /* copied from .descr->szB */ member 486 VG_(printf)(" [%#lx,+%lu) descr=%p %s %lu\n", nd->addr, nd->szB, in pp_StackTree() [all …]
|
D | h_main.c | 95 SizeT szB; /* may be zero */ member 110 } else if (a < seg->addr + seg->szB && seg->szB > 0) { in Seg__cmp() 115 *n = a - (seg->addr + seg->szB); in Seg__cmp() 136 return seg->szB; in Seg__size() 291 seg->szB = size; in add_new_segment() 419 if (new_size <= seg->szB) { in h_replace_realloc() 436 VG_(memcpy)((void*)p_new, p_old, seg->szB); in h_replace_realloc() 458 return ( seg ? seg->szB : 0 ); in h_replace_malloc_usable_size()
|
/external/valgrind/memcheck/ |
D | mc_malloc_wrappers.c | 78 MC_Chunk* create_MC_Chunk ( ThreadId tid, Addr p, SizeT szB, 102 const int l = (mc->szB >= MC_(clo_freelist_big_blocks) ? 0 : 1); in add_to_freed_queue() 113 if (mc->szB >= MC_(clo_freelist_vol)) { in add_to_freed_queue() 122 VG_(free_queue_volume) += (Long)mc->szB; in add_to_freed_queue() 149 VG_(free_queue_volume) -= (Long)mc1->szB; in release_oldest_block() 178 if (VG_(addr_is_in_block)( a, mc->data, mc->szB, in MC_() 190 MC_Chunk* create_MC_Chunk ( ThreadId tid, Addr p, SizeT szB, in create_MC_Chunk() argument 195 mc->szB = szB; in create_MC_Chunk() 241 if (found_mc->szB != mc->szB in in_block_list() 340 Addr p, SizeT szB, SizeT alignB, in MC_() [all …]
|
D | mc_leakcheck.c | 272 a_hi = ((Addr)chunks[i]->data) + chunks[i]->szB; in find_chunk_for_OLD() 300 a_mid_hi = chunks[mid]->data + chunks[mid]->szB; in find_chunk_for() 307 if (chunks[mid]->szB == 0) in find_chunk_for() 379 if (mc->szB > 1) { in find_active_chunks() 380 m = find_chunk_for(mc->data + (mc->szB - 1), mallocs, n_mallocs); in find_active_chunks() 547 tl_assert(ptr < ch->data + ch->szB + (ch->szB==0 ? 1 : 0)); in lc_is_a_chunk_ptr() 566 VG_(printf)("pushing %#lx-%#lx\n", ch->data, ch->data + ch->szB); in lc_push() 779 if (3 * sizeof(SizeT) + capacity + 1 == ch->szB in heuristic_reachedness() 808 if (size > 0 && (ch->szB - sizeof(ULong)) == size) { in heuristic_reachedness() 834 if (nr_elts > 0 && (ch->szB - sizeof(SizeT)) % nr_elts == 0) { in heuristic_reachedness() [all …]
|
D | mc_errors.c | 95 SizeT szB; // size of value in bytes member 117 SizeT szB; // not used for exec (jump) errors member 168 SizeT szB; // Size in bytes; 0 if unused. member 336 lr->szB + lr->indirect_szB, in pp_LossRecord() 340 lr->szB, in pp_LossRecord() 359 lr->szB + lr->indirect_szB, d_bytes, in pp_LossRecord() 360 lr->szB, d_direct_bytes, in pp_LossRecord() 367 lr->szB + lr->indirect_szB ); in pp_LossRecord() 374 lr->szB, d_direct_bytes, in pp_LossRecord() 378 emit( " <leakedbytes>%lu</leakedbytes>\n", lr->szB); in pp_LossRecord() [all …]
|
D | mc_machine.c | 100 static Int get_otrack_shadow_offset_wrk ( Int offset, Int szB ); /*fwds*/ 102 Int MC_(get_otrack_shadow_offset) ( Int offset, Int szB ) in MC_() 104 Int cand = get_otrack_shadow_offset_wrk( offset, szB ); in MC_() 113 static Int get_otrack_shadow_offset_wrk ( Int offset, Int szB ) in get_otrack_shadow_offset_wrk() argument 124 Int sz = szB; in get_otrack_shadow_offset_wrk() 328 offset,szB); in get_otrack_shadow_offset_wrk() 342 Int sz = szB; in get_otrack_shadow_offset_wrk() 528 offset,szB); in get_otrack_shadow_offset_wrk() 542 Int sz = szB; in get_otrack_shadow_offset_wrk() 589 if (o == 1+ GOF(RAX) && szB == 1) return GOF(CC_OP); in get_otrack_shadow_offset_wrk() [all …]
|
D | mc_main.c | 1205 SSizeT szB = nBits / 8; in mc_LOADV_128_or_256_slow() local 1206 SSizeT szL = szB / 8; /* Size in Longs (64-bit units) */ in mc_LOADV_128_or_256_slow() 1215 tl_assert((szB & (szB-1)) == 0 && szL > 0); in mc_LOADV_128_or_256_slow() 1260 MC_(record_address_error)( VG_(get_running_tid)(), a, szB, False ); in mc_LOADV_128_or_256_slow() 1293 if (0 == (a & (szB - 1)) && n_addrs_bad < szB) { in mc_LOADV_128_or_256_slow() 1309 MC_(record_address_error)( VG_(get_running_tid)(), a, szB, False ); in mc_LOADV_128_or_256_slow() 1356 SSizeT szB = nBits / 8; in mc_LOADVn_slow() local 1373 for (i = szB-1; i >= 0; i--) { in mc_LOADVn_slow() 1375 ai = a + byte_offset_w(szB, bigendian, i); in mc_LOADVn_slow() 1392 MC_(record_address_error)( VG_(get_running_tid)(), a, szB, False ); in mc_LOADVn_slow() [all …]
|
D | mc_include.h | 68 SizeT szB : (sizeof(SizeT)*8)-2; // Size requested; 30 or 62 bits. 166 Int MC_(get_otrack_shadow_offset) ( Int offset, Int szB ); 433 SizeT szB; // Sum of all MC_Chunk.szB values. member 476 void MC_(who_points_at) ( Addr address, SizeT szB); 526 void MC_(record_address_error) ( ThreadId tid, Addr a, Int szB, 529 void MC_(record_value_error) ( ThreadId tid, Int szB, UInt otag ); 537 Addr src, Addr dst, SizeT szB );
|
/external/valgrind/coregrind/m_debuginfo/ |
D | readdwarf3.c | 209 return c->sli_next >= c->sli.ioff + c->sli.szB; in is_at_end_Cursor() 227 return c->sli.ioff + c->sli.szB - c->sli_next; in get_remaining_length_Cursor() 244 if (c->sli_next + sizeof(UChar) > c->sli.ioff + c->sli.szB) { in get_UChar() 256 if (c->sli_next + sizeof(UShort) > c->sli.ioff + c->sli.szB) { in get_UShort() 268 if (c->sli_next + sizeof(UInt) > c->sli.ioff + c->sli.szB) { in get_UInt() 280 if (c->sli_next + sizeof(ULong) > c->sli.ioff + c->sli.szB) { in get_ULong() 533 if (die >= cc->escn_debug_info.szB) { in uncook_die() 534 if (die >= cc->escn_debug_info.szB + cc->escn_debug_types.szB) { in uncook_die() 536 die -= cc->escn_debug_info.szB + cc->escn_debug_types.szB; in uncook_die() 539 die -= cc->escn_debug_info.szB; in uncook_die() [all …]
|
D | readmacho.c | 98 Bool ML_(is_macho_object_file)( const void* buf, SizeT szB ) in ML_() 117 if (szB < sizeof(struct fat_header)) in ML_() 122 if (szB < sizeof(struct MACH_HEADER)) in ML_() 225 sli.szB = arch.size; in map_image_aboard() 241 if (sli.szB < sizeof(struct MACH_HEADER)) { in map_image_aboard() 246 if (sli.szB > ML_(img_size)(sli.img)) { in map_image_aboard() 251 if (sli.ioff >= 0 && sli.ioff + sli.szB <= ML_(img_size)(sli.img)) { in map_image_aboard() 268 if (sli.szB < sizeof(struct MACH_HEADER) + mh.sizeofcmds) { in map_image_aboard() 278 vg_assert(sli.szB > 0); in map_image_aboard() 279 vg_assert(sli.ioff + sli.szB <= ML_(img_size)(sli.img)); in map_image_aboard() [all …]
|
D | misc.c | 45 void* ML_(dinfo_zalloc) ( const HChar* cc, SizeT szB ) { in ML_() 47 vg_assert(szB > 0); in ML_() 48 v = VG_(arena_malloc)( VG_AR_DINFO, cc, szB ); in ML_() 49 VG_(memset)(v, 0, szB); in ML_() 53 void ML_(dinfo_shrink_block)( void* ptr, SizeT szB ) { in ML_() 54 VG_(arena_realloc_shrink)( VG_AR_DINFO, ptr, szB ); in ML_()
|
D | tytypes.c | 127 te->Te.TyBase.szB, te->Te.TyBase.enc, in ML_() 132 VG_(printf)("Te_TyPtr(%d,0x%05lx)", te->Te.TyPorR.szB, in ML_() 136 VG_(printf)("Te_TyRef(%d,0x%05lx)", te->Te.TyPorR.szB, in ML_() 140 VG_(printf)("Te_TyMbr(%d,0x%05lx)", te->Te.TyPorR.szB, in ML_() 144 VG_(printf)("Te_TyRvalRef(%d,0x%05lx)", te->Te.TyPorR.szB, in ML_() 156 te->Te.TyStOrUn.szB, in ML_() 169 te->Te.TyEnum.szB, te->Te.TyEnum.atomRs, in ML_() 525 r = Int__cmp(te1->Te.TyBase.szB, te2->Te.TyBase.szB); in ML_() 533 r = Int__cmp(te1->Te.TyPorR.szB, te2->Te.TyPorR.szB); in ML_() 547 r = UWord__cmp(te1->Te.TyStOrUn.szB, te2->Te.TyStOrUn.szB); in ML_() [all …]
|
D | priv_tytypes.h | 105 Int szB; member 109 Int szB; member 118 UWord szB; member 126 Int szB; member
|
/external/valgrind/memcheck/tests/ |
D | calloc-overflow.c | 13 size_t szB = 0x1000000010000001ULL; in main() local 15 size_t szB = 0x10000001UL; in main() local 17 x = calloc(szB, 0x10); in main()
|
/external/valgrind/massif/ |
D | ms_main.c | 535 SizeT szB; member 558 SizeT szB; // memory size for the node, be it Sig or Insig member 589 xpt->szB = 0; in new_XPt() 633 return ( sxpt1->szB < sxpt2->szB ? 1 in SXPt_revcmp_szB() 634 : sxpt1->szB > sxpt2->szB ? -1 in SXPt_revcmp_szB() 673 if (xpt->children[i]->szB >= sig_child_threshold_szB) { in dup_XTree() 684 sxpt->szB = xpt->szB; in dup_XTree() 699 if (xpt->children[i]->szB >= sig_child_threshold_szB) { in dup_XTree() 701 sig_children_szB += xpt->children[i]->szB; in dup_XTree() 703 insig_children_szB += xpt->children[i]->szB; in dup_XTree() [all …]
|
D | ms_print.in | 263 my ($szB, $szB_scaled) = @_; 277 if ($szB < 1000) { return sprintf("%5d", $szB); } 307 my ($szB) = @_; 311 my $szB_scaled = $szB; 325 return (max_label_2($szB, $szB_scaled), $unit); 331 my ($szB) = @_; 336 my $szB_scaled = $szB; 341 return (max_label_2($szB, $szB_scaled), $unit);
|
/external/valgrind/exp-dhat/ |
D | dh_main.c | 620 static void* dh_malloc ( ThreadId tid, SizeT szB ) in dh_malloc() argument 622 return new_block( tid, NULL, szB, VG_(clo_alignment), /*is_zeroed*/False ); in dh_malloc() 625 static void* dh___builtin_new ( ThreadId tid, SizeT szB ) in dh___builtin_new() argument 627 return new_block( tid, NULL, szB, VG_(clo_alignment), /*is_zeroed*/False ); in dh___builtin_new() 630 static void* dh___builtin_vec_new ( ThreadId tid, SizeT szB ) in dh___builtin_vec_new() argument 632 return new_block( tid, NULL, szB, VG_(clo_alignment), /*is_zeroed*/False ); in dh___builtin_vec_new() 635 static void* dh_calloc ( ThreadId tid, SizeT m, SizeT szB ) in dh_calloc() argument 637 return new_block( tid, NULL, m*szB, VG_(clo_alignment), /*is_zeroed*/True ); in dh_calloc() 640 static void *dh_memalign ( ThreadId tid, SizeT alignB, SizeT szB ) in dh_memalign() argument 642 return new_block( tid, NULL, szB, alignB, False ); in dh_memalign() [all …]
|
/external/valgrind/coregrind/m_initimg/ |
D | initimg-solaris.c | 910 SizeT szB = (SizeT)VG_(client_rlimit_stack).rlim_cur; in VG_() local 911 if (szB < m1) in VG_() 912 szB = m1; in VG_() 913 if (szB > m16) in VG_() 914 szB = m16; in VG_() 917 szB = VG_(clo_main_stacksize); in VG_() 918 if (szB < m1) in VG_() 919 szB = m1; in VG_() 921 szB = VG_PGROUNDUP(szB); in VG_() 923 "Setup client stack: size will be %ld\n", szB); in VG_() [all …]
|
D | initimg-linux.c | 959 SizeT szB = (SizeT)VG_(client_rlimit_stack).rlim_cur; in VG_() local 960 if (szB < m1) szB = m1; in VG_() 961 if (szB > m16) szB = m16; in VG_() 962 if (VG_(clo_main_stacksize) > 0) szB = VG_(clo_main_stacksize); in VG_() 963 if (szB < m1) szB = m1; in VG_() 964 szB = VG_PGROUNDUP(szB); in VG_() 966 "Setup client stack: size will be %lu\n", szB); in VG_() 968 iifii.clstack_max_size = szB; in VG_()
|
/external/valgrind/helgrind/ |
D | libhb_core.c | 427 static inline Bool address_in_range (Addr a, Addr start, SizeT szB) in address_in_range() argument 433 tl_assert ((a - start < szB) in address_in_range() 435 && a < start + szB)); in address_in_range() 436 return a - start < szB; in address_in_range() 1648 static void shmem__invalidate_scache_range (Addr ga, SizeT szB) in shmem__invalidate_scache_range() argument 1655 tl_assert (0 == (szB & (N_LINE_ARANGE - 1))); in shmem__invalidate_scache_range() 1659 Word nwix = szB / N_LINE_ARANGE; in shmem__invalidate_scache_range() 1665 if (address_in_range(cache_shmem.tags0[ga_ix], ga, szB)) in shmem__invalidate_scache_range() 4445 UInt szB : 32 - SCALARTS_N_THRBITS - 1; member 4582 static void event_map_bind ( Addr a, SizeT szB, Bool isW, Thr* thr ) in event_map_bind() argument [all …]
|
D | hg_errors.c | 309 Int szB; member 433 Int acc_szB = xe->XE.Race.szB; in HG_() 470 Addr data_addr, Int szB, Bool isWrite, in HG_() 500 xe.XE.Race.szB = szB; in HG_() 504 tl_assert(szB == 8 || szB == 4 || szB == 2 || szB == 1); in HG_() 671 return xe1->XE.Race.szB == xe2->XE.Race.szB in HG_() 1163 Int szB; in HG_() local 1165 szB = xe->XE.Race.szB; in HG_() 1178 what, szB, (void*)err_ga, (Int)xe->XE.Race.thr->errmsg_index ); in HG_() 1227 what, szB, (void*)err_ga, (Int)xe->XE.Race.thr->errmsg_index ); in HG_()
|
D | hg_main.c | 4109 SizeT szB; /* size requested */ member 4137 SizeT szB, SizeT alignB, Bool is_zeroed ) in handle_alloc() argument 4142 tl_assert( ((SSizeT)szB) >= 0 ); in handle_alloc() 4143 p = (Addr)VG_(cli_malloc)(alignB, szB); in handle_alloc() 4148 VG_(memset)((void*)p, 0, szB); in handle_alloc() 4155 md->szB = szB; in handle_alloc() 4162 evh__new_mem_heap( p, szB, is_zeroed ); in handle_alloc() 4205 SizeT szB; in handle_free() local 4213 szB = md->szB; in handle_free() 4226 evh__die_mem_heap( (Addr)p, szB ); in handle_free() [all …]
|
/external/valgrind/include/ |
D | pub_tool_gdbserver.h | 110 extern Bool VG_(is_watched)(PointKind kind, Addr addr, Int szB); 188 SizeT* szB,
|
/external/valgrind/coregrind/m_aspacemgr/ |
D | aspacemgr-common.c | 430 Int szB; in VG_() local 437 szB = VG_STACK_GUARD_SZB in VG_() 440 sres = VG_(am_mmap_anon_float_valgrind)( szB ); in VG_() 446 aspacem_assert(VG_IS_PAGE_ALIGNED(szB)); in VG_() 483 (ULong)(Addr)stack, szB); in VG_() 490 (void)ML_(am_do_munmap_NO_NOTIFY)( (Addr)stack, szB ); in VG_()
|