31 for f
in microbenchfiles:
32 content = context.read_result_file(context, f)
33 data = json.loads(content)
36 for benchmark
in data[
'benchmarks']:
38 benchmarkname = benchmark[
'name']
40 assert benchmarkname.startswith(context.rawfilename +
'/')
41 benchmarkname = benchmarkname[len(context.rawfilename +
'/'):]
43 microBenchmark = lit.Test.Result(lit.Test.PASS)
45 exec_time_metric = lit.Test.toMetricValue(benchmark[
'WallTime,s'])
46 microBenchmark.addMetric(
'exec_time', exec_time_metric)
48 if hasattr(context,
'profilefile'):
49 microBenchmark.addMetric(
50 'profile', lit.Test.toMetricValue(
53 for field
in benchmark.keys():
54 if field
in [
'real_time',
'cpu_time',
'time_unit']:
56 metric = lit.Test.toMetricValue(benchmark[field])
57 microBenchmark.addMetric(field, metric)
59 context.micro_results[benchmarkname] = microBenchmark
64 'rsbench': lit.Test.toMetricValue(len(context.micro_results))