Home
last modified time | relevance | path

Searched refs:szB (Results 1 – 25 of 61) sorted by relevance

123

/external/valgrind/tests/
Dmalloc.h14 static void* memalign16(size_t szB) in memalign16() argument
19 posix_memalign((void **)&x, 16, szB); in memalign16()
21 x = memalign(16, szB); in memalign16()
30 static void* memalign32(size_t szB) in memalign32() argument
35 posix_memalign((void **)&x, 32, szB); in memalign32()
37 x = memalign(32, szB); in memalign32()
46 static void* memalign64(size_t szB) in memalign64() argument
51 posix_memalign((void **)&x, 64, szB); in memalign64()
53 x = memalign(64, szB); in memalign64()
/external/valgrind/exp-sgcheck/
Dsg_main.c174 if (fb1->szB < fb2->szB) return -1; in StackBlock__cmp()
175 if (fb1->szB > fb2->szB) return 1; in StackBlock__cmp()
230 sb->base, sb->szB, sb->spRel ? 'Y' : 'N', in pp_StackBlocks()
314 tl_assert(sb0->szB < sb1->szB); in StackBlocks__find_and_dealloc__or_add()
315 sb0->szB = sb1->szB; in StackBlocks__find_and_dealloc__or_add()
343 || sb1->base + sb1->szB > sb2->base)) { in StackBlocks__find_and_dealloc__or_add()
403 if (gb1->szB < gb2->szB) return -1; in GlobalBlock__cmp()
404 if (gb1->szB > gb2->szB) return 1; in GlobalBlock__cmp()
473 SizeT szB; /* copied from .descr->szB */ member
486 VG_(printf)(" [%#lx,+%lu) descr=%p %s %lu\n", nd->addr, nd->szB, in pp_StackTree()
[all …]
Dh_main.c95 SizeT szB; /* may be zero */ member
110 } else if (a < seg->addr + seg->szB && seg->szB > 0) { in Seg__cmp()
115 *n = a - (seg->addr + seg->szB); in Seg__cmp()
136 return seg->szB; in Seg__size()
291 seg->szB = size; in add_new_segment()
419 if (new_size <= seg->szB) { in h_replace_realloc()
436 VG_(memcpy)((void*)p_new, p_old, seg->szB); in h_replace_realloc()
458 return ( seg ? seg->szB : 0 ); in h_replace_malloc_usable_size()
/external/valgrind/memcheck/
Dmc_malloc_wrappers.c78 MC_Chunk* create_MC_Chunk ( ThreadId tid, Addr p, SizeT szB,
102 const int l = (mc->szB >= MC_(clo_freelist_big_blocks) ? 0 : 1); in add_to_freed_queue()
113 if (mc->szB >= MC_(clo_freelist_vol)) { in add_to_freed_queue()
122 VG_(free_queue_volume) += (Long)mc->szB; in add_to_freed_queue()
149 VG_(free_queue_volume) -= (Long)mc1->szB; in release_oldest_block()
178 if (VG_(addr_is_in_block)( a, mc->data, mc->szB, in MC_()
190 MC_Chunk* create_MC_Chunk ( ThreadId tid, Addr p, SizeT szB, in create_MC_Chunk() argument
195 mc->szB = szB; in create_MC_Chunk()
241 if (found_mc->szB != mc->szB in in_block_list()
340 Addr p, SizeT szB, SizeT alignB, in MC_()
[all …]
Dmc_leakcheck.c272 a_hi = ((Addr)chunks[i]->data) + chunks[i]->szB; in find_chunk_for_OLD()
300 a_mid_hi = chunks[mid]->data + chunks[mid]->szB; in find_chunk_for()
307 if (chunks[mid]->szB == 0) in find_chunk_for()
379 if (mc->szB > 1) { in find_active_chunks()
380 m = find_chunk_for(mc->data + (mc->szB - 1), mallocs, n_mallocs); in find_active_chunks()
547 tl_assert(ptr < ch->data + ch->szB + (ch->szB==0 ? 1 : 0)); in lc_is_a_chunk_ptr()
566 VG_(printf)("pushing %#lx-%#lx\n", ch->data, ch->data + ch->szB); in lc_push()
779 if (3 * sizeof(SizeT) + capacity + 1 == ch->szB in heuristic_reachedness()
808 if (size > 0 && (ch->szB - sizeof(ULong)) == size) { in heuristic_reachedness()
834 if (nr_elts > 0 && (ch->szB - sizeof(SizeT)) % nr_elts == 0) { in heuristic_reachedness()
[all …]
Dmc_errors.c95 SizeT szB; // size of value in bytes member
117 SizeT szB; // not used for exec (jump) errors member
168 SizeT szB; // Size in bytes; 0 if unused. member
336 lr->szB + lr->indirect_szB, in pp_LossRecord()
340 lr->szB, in pp_LossRecord()
359 lr->szB + lr->indirect_szB, d_bytes, in pp_LossRecord()
360 lr->szB, d_direct_bytes, in pp_LossRecord()
367 lr->szB + lr->indirect_szB ); in pp_LossRecord()
374 lr->szB, d_direct_bytes, in pp_LossRecord()
378 emit( " <leakedbytes>%lu</leakedbytes>\n", lr->szB); in pp_LossRecord()
[all …]
Dmc_machine.c100 static Int get_otrack_shadow_offset_wrk ( Int offset, Int szB ); /*fwds*/
102 Int MC_(get_otrack_shadow_offset) ( Int offset, Int szB ) in MC_()
104 Int cand = get_otrack_shadow_offset_wrk( offset, szB ); in MC_()
113 static Int get_otrack_shadow_offset_wrk ( Int offset, Int szB ) in get_otrack_shadow_offset_wrk() argument
124 Int sz = szB; in get_otrack_shadow_offset_wrk()
328 offset,szB); in get_otrack_shadow_offset_wrk()
342 Int sz = szB; in get_otrack_shadow_offset_wrk()
528 offset,szB); in get_otrack_shadow_offset_wrk()
542 Int sz = szB; in get_otrack_shadow_offset_wrk()
589 if (o == 1+ GOF(RAX) && szB == 1) return GOF(CC_OP); in get_otrack_shadow_offset_wrk()
[all …]
Dmc_main.c1205 SSizeT szB = nBits / 8; in mc_LOADV_128_or_256_slow() local
1206 SSizeT szL = szB / 8; /* Size in Longs (64-bit units) */ in mc_LOADV_128_or_256_slow()
1215 tl_assert((szB & (szB-1)) == 0 && szL > 0); in mc_LOADV_128_or_256_slow()
1260 MC_(record_address_error)( VG_(get_running_tid)(), a, szB, False ); in mc_LOADV_128_or_256_slow()
1293 if (0 == (a & (szB - 1)) && n_addrs_bad < szB) { in mc_LOADV_128_or_256_slow()
1309 MC_(record_address_error)( VG_(get_running_tid)(), a, szB, False ); in mc_LOADV_128_or_256_slow()
1356 SSizeT szB = nBits / 8; in mc_LOADVn_slow() local
1373 for (i = szB-1; i >= 0; i--) { in mc_LOADVn_slow()
1375 ai = a + byte_offset_w(szB, bigendian, i); in mc_LOADVn_slow()
1392 MC_(record_address_error)( VG_(get_running_tid)(), a, szB, False ); in mc_LOADVn_slow()
[all …]
Dmc_include.h68 SizeT szB : (sizeof(SizeT)*8)-2; // Size requested; 30 or 62 bits.
166 Int MC_(get_otrack_shadow_offset) ( Int offset, Int szB );
433 SizeT szB; // Sum of all MC_Chunk.szB values. member
476 void MC_(who_points_at) ( Addr address, SizeT szB);
526 void MC_(record_address_error) ( ThreadId tid, Addr a, Int szB,
529 void MC_(record_value_error) ( ThreadId tid, Int szB, UInt otag );
537 Addr src, Addr dst, SizeT szB );
/external/valgrind/coregrind/m_debuginfo/
Dreaddwarf3.c209 return c->sli_next >= c->sli.ioff + c->sli.szB; in is_at_end_Cursor()
227 return c->sli.ioff + c->sli.szB - c->sli_next; in get_remaining_length_Cursor()
244 if (c->sli_next + sizeof(UChar) > c->sli.ioff + c->sli.szB) { in get_UChar()
256 if (c->sli_next + sizeof(UShort) > c->sli.ioff + c->sli.szB) { in get_UShort()
268 if (c->sli_next + sizeof(UInt) > c->sli.ioff + c->sli.szB) { in get_UInt()
280 if (c->sli_next + sizeof(ULong) > c->sli.ioff + c->sli.szB) { in get_ULong()
533 if (die >= cc->escn_debug_info.szB) { in uncook_die()
534 if (die >= cc->escn_debug_info.szB + cc->escn_debug_types.szB) { in uncook_die()
536 die -= cc->escn_debug_info.szB + cc->escn_debug_types.szB; in uncook_die()
539 die -= cc->escn_debug_info.szB; in uncook_die()
[all …]
Dreadmacho.c98 Bool ML_(is_macho_object_file)( const void* buf, SizeT szB ) in ML_()
117 if (szB < sizeof(struct fat_header)) in ML_()
122 if (szB < sizeof(struct MACH_HEADER)) in ML_()
225 sli.szB = arch.size; in map_image_aboard()
241 if (sli.szB < sizeof(struct MACH_HEADER)) { in map_image_aboard()
246 if (sli.szB > ML_(img_size)(sli.img)) { in map_image_aboard()
251 if (sli.ioff >= 0 && sli.ioff + sli.szB <= ML_(img_size)(sli.img)) { in map_image_aboard()
268 if (sli.szB < sizeof(struct MACH_HEADER) + mh.sizeofcmds) { in map_image_aboard()
278 vg_assert(sli.szB > 0); in map_image_aboard()
279 vg_assert(sli.ioff + sli.szB <= ML_(img_size)(sli.img)); in map_image_aboard()
[all …]
Dmisc.c45 void* ML_(dinfo_zalloc) ( const HChar* cc, SizeT szB ) { in ML_()
47 vg_assert(szB > 0); in ML_()
48 v = VG_(arena_malloc)( VG_AR_DINFO, cc, szB ); in ML_()
49 VG_(memset)(v, 0, szB); in ML_()
53 void ML_(dinfo_shrink_block)( void* ptr, SizeT szB ) { in ML_()
54 VG_(arena_realloc_shrink)( VG_AR_DINFO, ptr, szB ); in ML_()
Dtytypes.c127 te->Te.TyBase.szB, te->Te.TyBase.enc, in ML_()
132 VG_(printf)("Te_TyPtr(%d,0x%05lx)", te->Te.TyPorR.szB, in ML_()
136 VG_(printf)("Te_TyRef(%d,0x%05lx)", te->Te.TyPorR.szB, in ML_()
140 VG_(printf)("Te_TyMbr(%d,0x%05lx)", te->Te.TyPorR.szB, in ML_()
144 VG_(printf)("Te_TyRvalRef(%d,0x%05lx)", te->Te.TyPorR.szB, in ML_()
156 te->Te.TyStOrUn.szB, in ML_()
169 te->Te.TyEnum.szB, te->Te.TyEnum.atomRs, in ML_()
525 r = Int__cmp(te1->Te.TyBase.szB, te2->Te.TyBase.szB); in ML_()
533 r = Int__cmp(te1->Te.TyPorR.szB, te2->Te.TyPorR.szB); in ML_()
547 r = UWord__cmp(te1->Te.TyStOrUn.szB, te2->Te.TyStOrUn.szB); in ML_()
[all …]
Dpriv_tytypes.h105 Int szB; member
109 Int szB; member
118 UWord szB; member
126 Int szB; member
/external/valgrind/memcheck/tests/
Dcalloc-overflow.c13 size_t szB = 0x1000000010000001ULL; in main() local
15 size_t szB = 0x10000001UL; in main() local
17 x = calloc(szB, 0x10); in main()
/external/valgrind/massif/
Dms_main.c535 SizeT szB; member
558 SizeT szB; // memory size for the node, be it Sig or Insig member
589 xpt->szB = 0; in new_XPt()
633 return ( sxpt1->szB < sxpt2->szB ? 1 in SXPt_revcmp_szB()
634 : sxpt1->szB > sxpt2->szB ? -1 in SXPt_revcmp_szB()
673 if (xpt->children[i]->szB >= sig_child_threshold_szB) { in dup_XTree()
684 sxpt->szB = xpt->szB; in dup_XTree()
699 if (xpt->children[i]->szB >= sig_child_threshold_szB) { in dup_XTree()
701 sig_children_szB += xpt->children[i]->szB; in dup_XTree()
703 insig_children_szB += xpt->children[i]->szB; in dup_XTree()
[all …]
Dms_print.in263 my ($szB, $szB_scaled) = @_;
277 if ($szB < 1000) { return sprintf("%5d", $szB); }
307 my ($szB) = @_;
311 my $szB_scaled = $szB;
325 return (max_label_2($szB, $szB_scaled), $unit);
331 my ($szB) = @_;
336 my $szB_scaled = $szB;
341 return (max_label_2($szB, $szB_scaled), $unit);
/external/valgrind/exp-dhat/
Ddh_main.c620 static void* dh_malloc ( ThreadId tid, SizeT szB ) in dh_malloc() argument
622 return new_block( tid, NULL, szB, VG_(clo_alignment), /*is_zeroed*/False ); in dh_malloc()
625 static void* dh___builtin_new ( ThreadId tid, SizeT szB ) in dh___builtin_new() argument
627 return new_block( tid, NULL, szB, VG_(clo_alignment), /*is_zeroed*/False ); in dh___builtin_new()
630 static void* dh___builtin_vec_new ( ThreadId tid, SizeT szB ) in dh___builtin_vec_new() argument
632 return new_block( tid, NULL, szB, VG_(clo_alignment), /*is_zeroed*/False ); in dh___builtin_vec_new()
635 static void* dh_calloc ( ThreadId tid, SizeT m, SizeT szB ) in dh_calloc() argument
637 return new_block( tid, NULL, m*szB, VG_(clo_alignment), /*is_zeroed*/True ); in dh_calloc()
640 static void *dh_memalign ( ThreadId tid, SizeT alignB, SizeT szB ) in dh_memalign() argument
642 return new_block( tid, NULL, szB, alignB, False ); in dh_memalign()
[all …]
/external/valgrind/coregrind/m_initimg/
Dinitimg-solaris.c910 SizeT szB = (SizeT)VG_(client_rlimit_stack).rlim_cur; in VG_() local
911 if (szB < m1) in VG_()
912 szB = m1; in VG_()
913 if (szB > m16) in VG_()
914 szB = m16; in VG_()
917 szB = VG_(clo_main_stacksize); in VG_()
918 if (szB < m1) in VG_()
919 szB = m1; in VG_()
921 szB = VG_PGROUNDUP(szB); in VG_()
923 "Setup client stack: size will be %ld\n", szB); in VG_()
[all …]
Dinitimg-linux.c959 SizeT szB = (SizeT)VG_(client_rlimit_stack).rlim_cur; in VG_() local
960 if (szB < m1) szB = m1; in VG_()
961 if (szB > m16) szB = m16; in VG_()
962 if (VG_(clo_main_stacksize) > 0) szB = VG_(clo_main_stacksize); in VG_()
963 if (szB < m1) szB = m1; in VG_()
964 szB = VG_PGROUNDUP(szB); in VG_()
966 "Setup client stack: size will be %lu\n", szB); in VG_()
968 iifii.clstack_max_size = szB; in VG_()
/external/valgrind/helgrind/
Dlibhb_core.c427 static inline Bool address_in_range (Addr a, Addr start, SizeT szB) in address_in_range() argument
433 tl_assert ((a - start < szB) in address_in_range()
435 && a < start + szB)); in address_in_range()
436 return a - start < szB; in address_in_range()
1648 static void shmem__invalidate_scache_range (Addr ga, SizeT szB) in shmem__invalidate_scache_range() argument
1655 tl_assert (0 == (szB & (N_LINE_ARANGE - 1))); in shmem__invalidate_scache_range()
1659 Word nwix = szB / N_LINE_ARANGE; in shmem__invalidate_scache_range()
1665 if (address_in_range(cache_shmem.tags0[ga_ix], ga, szB)) in shmem__invalidate_scache_range()
4445 UInt szB : 32 - SCALARTS_N_THRBITS - 1; member
4582 static void event_map_bind ( Addr a, SizeT szB, Bool isW, Thr* thr ) in event_map_bind() argument
[all …]
Dhg_errors.c309 Int szB; member
433 Int acc_szB = xe->XE.Race.szB; in HG_()
470 Addr data_addr, Int szB, Bool isWrite, in HG_()
500 xe.XE.Race.szB = szB; in HG_()
504 tl_assert(szB == 8 || szB == 4 || szB == 2 || szB == 1); in HG_()
671 return xe1->XE.Race.szB == xe2->XE.Race.szB in HG_()
1163 Int szB; in HG_() local
1165 szB = xe->XE.Race.szB; in HG_()
1178 what, szB, (void*)err_ga, (Int)xe->XE.Race.thr->errmsg_index ); in HG_()
1227 what, szB, (void*)err_ga, (Int)xe->XE.Race.thr->errmsg_index ); in HG_()
Dhg_main.c4109 SizeT szB; /* size requested */ member
4137 SizeT szB, SizeT alignB, Bool is_zeroed ) in handle_alloc() argument
4142 tl_assert( ((SSizeT)szB) >= 0 ); in handle_alloc()
4143 p = (Addr)VG_(cli_malloc)(alignB, szB); in handle_alloc()
4148 VG_(memset)((void*)p, 0, szB); in handle_alloc()
4155 md->szB = szB; in handle_alloc()
4162 evh__new_mem_heap( p, szB, is_zeroed ); in handle_alloc()
4205 SizeT szB; in handle_free() local
4213 szB = md->szB; in handle_free()
4226 evh__die_mem_heap( (Addr)p, szB ); in handle_free()
[all …]
/external/valgrind/include/
Dpub_tool_gdbserver.h110 extern Bool VG_(is_watched)(PointKind kind, Addr addr, Int szB);
188 SizeT* szB,
/external/valgrind/coregrind/m_aspacemgr/
Daspacemgr-common.c430 Int szB; in VG_() local
437 szB = VG_STACK_GUARD_SZB in VG_()
440 sres = VG_(am_mmap_anon_float_valgrind)( szB ); in VG_()
446 aspacem_assert(VG_IS_PAGE_ALIGNED(szB)); in VG_()
483 (ULong)(Addr)stack, szB); in VG_()
490 (void)ML_(am_do_munmap_NO_NOTIFY)( (Addr)stack, szB ); in VG_()

123