Lines Matching refs:bm_name
78 def collect_latency(bm_name, args): argument
84 heading('Latency Profiles: %s' % bm_name)
86 'make', bm_name, 'CONFIG=basicprof', '-j',
90 ['bins/basicprof/%s' % bm_name, '--benchmark_list_tests']).splitlines():
95 'bins/basicprof/%s' % bm_name,
135 def collect_perf(bm_name, args): argument
137 heading('Flamegraphs: %s' % bm_name)
139 'make', bm_name, 'CONFIG=mutrace', '-j',
146 ['bins/mutrace/%s' % bm_name, '--benchmark_list_tests']).splitlines():
153 'bins/mutrace/%s' % bm_name,
188 def run_summary(bm_name, cfg, base_json_name): argument
190 'make', bm_name,
195 'bins/%s/%s' % (cfg, bm_name),
204 def collect_summary(bm_name, args): argument
205 heading('Summary: %s [no counters]' % bm_name)
206 text(run_summary(bm_name, 'opt', bm_name))
207 heading('Summary: %s [with counters]' % bm_name)
208 text(run_summary(bm_name, 'counters', bm_name))
210 with open('%s.csv' % bm_name, 'w') as f:
214 '%s.counters.json' % bm_name,
215 '%s.opt.json' % bm_name
219 '%s.csv' % bm_name
260 for bm_name in args.benchmarks:
261 collectors[collect](bm_name, args)