#!/usr/bin/env python """ compare_bench.py - Compare two benchmarks or their results and report the difference. """ import argparse from argparse import ArgumentParser import sys import gbench from gbench import util, report from gbench.util import * def check_inputs(in1, in2, flags): """ Perform checking on the user provided inputs and diagnose any abnormalities """ in1_kind, in1_err = classify_input_file(in1) in2_kind, in2_err = classify_input_file(in2) output_file = find_benchmark_flag('--benchmark_out=', flags) output_type = find_benchmark_flag('--benchmark_out_format=', flags) if in1_kind == IT_Executable and in2_kind == IT_Executable and output_file: print(("WARNING: '--benchmark_out=%s' will be passed to both " "benchmarks causing it to be overwritten") % output_file) if in1_kind == IT_JSON and in2_kind == IT_JSON and len(flags) > 0: print("WARNING: passing --benchmark flags has no effect since both " "inputs are JSON") if output_type is not None and output_type != 'json': print(("ERROR: passing '--benchmark_out_format=%s' to 'compare_bench.py`" " is not supported.") % output_type) sys.exit(1) def main(): parser = ArgumentParser( description='compare the results of two benchmarks') parser.add_argument( 'test1', metavar='test1', type=str, nargs=1, help='A benchmark executable or JSON output file') parser.add_argument( 'test2', metavar='test2', type=str, nargs=1, help='A benchmark executable or JSON output file') # FIXME this is a dummy argument which will never actually match # any --benchmark flags but it helps generate a better usage message parser.add_argument( 'benchmark_options', metavar='benchmark_option', nargs='*', help='Arguments to pass when running benchmark executables' ) args, unknown_args = parser.parse_known_args() # Parse the command line flags test1 = args.test1[0] test2 = args.test2[0] if args.benchmark_options: print("Unrecognized positional argument arguments: '%s'" % args.benchmark_options) exit(1) benchmark_options = unknown_args check_inputs(test1, test2, benchmark_options) # Run the benchmarks and report the results json1 = gbench.util.run_or_load_benchmark(test1, benchmark_options) json2 = gbench.util.run_or_load_benchmark(test2, benchmark_options) output_lines = gbench.report.generate_difference_report(json1, json2) print('Comparing %s to %s' % (test1, test2)) for ln in output_lines: print(ln) if __name__ == '__main__': main()