123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316 |
- #!/usr/bin/env python
- """
- compare.py - versatile benchmark output compare tool
- """
- import argparse
- from argparse import ArgumentParser
- import sys
- import gbench
- from gbench import util, report
- from gbench.util import *
- def check_inputs(in1, in2, flags):
- """
- Perform checking on the user provided inputs and diagnose any abnormalities
- """
- in1_kind, in1_err = classify_input_file(in1)
- in2_kind, in2_err = classify_input_file(in2)
- output_file = find_benchmark_flag('--benchmark_out=', flags)
- output_type = find_benchmark_flag('--benchmark_out_format=', flags)
- if in1_kind == IT_Executable and in2_kind == IT_Executable and output_file:
- print(("WARNING: '--benchmark_out=%s' will be passed to both "
- "benchmarks causing it to be overwritten") % output_file)
- if in1_kind == IT_JSON and in2_kind == IT_JSON and len(flags) > 0:
- print("WARNING: passing optional flags has no effect since both "
- "inputs are JSON")
- if output_type is not None and output_type != 'json':
- print(("ERROR: passing '--benchmark_out_format=%s' to 'compare.py`"
- " is not supported.") % output_type)
- sys.exit(1)
- def create_parser():
- parser = ArgumentParser(
- description='versatile benchmark output compare tool')
- subparsers = parser.add_subparsers(
- help='This tool has multiple modes of operation:',
- dest='mode')
- parser_a = subparsers.add_parser(
- 'benchmarks',
- help='The most simple use-case, compare all the output of these two benchmarks')
- baseline = parser_a.add_argument_group(
- 'baseline', 'The benchmark baseline')
- baseline.add_argument(
- 'test_baseline',
- metavar='test_baseline',
- type=argparse.FileType('r'),
- nargs=1,
- help='A benchmark executable or JSON output file')
- contender = parser_a.add_argument_group(
- 'contender', 'The benchmark that will be compared against the baseline')
- contender.add_argument(
- 'test_contender',
- metavar='test_contender',
- type=argparse.FileType('r'),
- nargs=1,
- help='A benchmark executable or JSON output file')
- parser_a.add_argument(
- 'benchmark_options',
- metavar='benchmark_options',
- nargs=argparse.REMAINDER,
- help='Arguments to pass when running benchmark executables')
- parser_b = subparsers.add_parser(
- 'filters', help='Compare filter one with the filter two of benchmark')
- baseline = parser_b.add_argument_group(
- 'baseline', 'The benchmark baseline')
- baseline.add_argument(
- 'test',
- metavar='test',
- type=argparse.FileType('r'),
- nargs=1,
- help='A benchmark executable or JSON output file')
- baseline.add_argument(
- 'filter_baseline',
- metavar='filter_baseline',
- type=str,
- nargs=1,
- help='The first filter, that will be used as baseline')
- contender = parser_b.add_argument_group(
- 'contender', 'The benchmark that will be compared against the baseline')
- contender.add_argument(
- 'filter_contender',
- metavar='filter_contender',
- type=str,
- nargs=1,
- help='The second filter, that will be compared against the baseline')
- parser_b.add_argument(
- 'benchmark_options',
- metavar='benchmark_options',
- nargs=argparse.REMAINDER,
- help='Arguments to pass when running benchmark executables')
- parser_c = subparsers.add_parser(
- 'benchmarksfiltered',
- help='Compare filter one of first benchmark with filter two of the second benchmark')
- baseline = parser_c.add_argument_group(
- 'baseline', 'The benchmark baseline')
- baseline.add_argument(
- 'test_baseline',
- metavar='test_baseline',
- type=argparse.FileType('r'),
- nargs=1,
- help='A benchmark executable or JSON output file')
- baseline.add_argument(
- 'filter_baseline',
- metavar='filter_baseline',
- type=str,
- nargs=1,
- help='The first filter, that will be used as baseline')
- contender = parser_c.add_argument_group(
- 'contender', 'The benchmark that will be compared against the baseline')
- contender.add_argument(
- 'test_contender',
- metavar='test_contender',
- type=argparse.FileType('r'),
- nargs=1,
- help='The second benchmark executable or JSON output file, that will be compared against the baseline')
- contender.add_argument(
- 'filter_contender',
- metavar='filter_contender',
- type=str,
- nargs=1,
- help='The second filter, that will be compared against the baseline')
- parser_c.add_argument(
- 'benchmark_options',
- metavar='benchmark_options',
- nargs=argparse.REMAINDER,
- help='Arguments to pass when running benchmark executables')
- return parser
- def main():
- # Parse the command line flags
- parser = create_parser()
- args, unknown_args = parser.parse_known_args()
- if args.mode is None:
- parser.print_help()
- exit(1)
- assert not unknown_args
- benchmark_options = args.benchmark_options
- if args.mode == 'benchmarks':
- test_baseline = args.test_baseline[0].name
- test_contender = args.test_contender[0].name
- filter_baseline = ''
- filter_contender = ''
- # NOTE: if test_baseline == test_contender, you are analyzing the stdev
- description = 'Comparing %s to %s' % (test_baseline, test_contender)
- elif args.mode == 'filters':
- test_baseline = args.test[0].name
- test_contender = args.test[0].name
- filter_baseline = args.filter_baseline[0]
- filter_contender = args.filter_contender[0]
- # NOTE: if filter_baseline == filter_contender, you are analyzing the
- # stdev
- description = 'Comparing %s to %s (from %s)' % (
- filter_baseline, filter_contender, args.test[0].name)
- elif args.mode == 'benchmarksfiltered':
- test_baseline = args.test_baseline[0].name
- test_contender = args.test_contender[0].name
- filter_baseline = args.filter_baseline[0]
- filter_contender = args.filter_contender[0]
- # NOTE: if test_baseline == test_contender and
- # filter_baseline == filter_contender, you are analyzing the stdev
- description = 'Comparing %s (from %s) to %s (from %s)' % (
- filter_baseline, test_baseline, filter_contender, test_contender)
- else:
- # should never happen
- print("Unrecognized mode of operation: '%s'" % args.mode)
- parser.print_help()
- exit(1)
- check_inputs(test_baseline, test_contender, benchmark_options)
- options_baseline = []
- options_contender = []
- if filter_baseline and filter_contender:
- options_baseline = ['--benchmark_filter=%s' % filter_baseline]
- options_contender = ['--benchmark_filter=%s' % filter_contender]
- # Run the benchmarks and report the results
- json1 = json1_orig = gbench.util.run_or_load_benchmark(
- test_baseline, benchmark_options + options_baseline)
- json2 = json2_orig = gbench.util.run_or_load_benchmark(
- test_contender, benchmark_options + options_contender)
- # Now, filter the benchmarks so that the difference report can work
- if filter_baseline and filter_contender:
- replacement = '[%s vs. %s]' % (filter_baseline, filter_contender)
- json1 = gbench.report.filter_benchmark(
- json1_orig, filter_baseline, replacement)
- json2 = gbench.report.filter_benchmark(
- json2_orig, filter_contender, replacement)
- # Diff and output
- output_lines = gbench.report.generate_difference_report(json1, json2)
- print(description)
- for ln in output_lines:
- print(ln)
- import unittest
- class TestParser(unittest.TestCase):
- def setUp(self):
- self.parser = create_parser()
- testInputs = os.path.join(
- os.path.dirname(
- os.path.realpath(__file__)),
- 'gbench',
- 'Inputs')
- self.testInput0 = os.path.join(testInputs, 'test1_run1.json')
- self.testInput1 = os.path.join(testInputs, 'test1_run2.json')
- def test_benchmarks_basic(self):
- parsed = self.parser.parse_args(
- ['benchmarks', self.testInput0, self.testInput1])
- self.assertEqual(parsed.mode, 'benchmarks')
- self.assertEqual(parsed.test_baseline[0].name, self.testInput0)
- self.assertEqual(parsed.test_contender[0].name, self.testInput1)
- self.assertFalse(parsed.benchmark_options)
- def test_benchmarks_with_remainder(self):
- parsed = self.parser.parse_args(
- ['benchmarks', self.testInput0, self.testInput1, 'd'])
- self.assertEqual(parsed.mode, 'benchmarks')
- self.assertEqual(parsed.test_baseline[0].name, self.testInput0)
- self.assertEqual(parsed.test_contender[0].name, self.testInput1)
- self.assertEqual(parsed.benchmark_options, ['d'])
- def test_benchmarks_with_remainder_after_doubleminus(self):
- parsed = self.parser.parse_args(
- ['benchmarks', self.testInput0, self.testInput1, '--', 'e'])
- self.assertEqual(parsed.mode, 'benchmarks')
- self.assertEqual(parsed.test_baseline[0].name, self.testInput0)
- self.assertEqual(parsed.test_contender[0].name, self.testInput1)
- self.assertEqual(parsed.benchmark_options, ['e'])
- def test_filters_basic(self):
- parsed = self.parser.parse_args(
- ['filters', self.testInput0, 'c', 'd'])
- self.assertEqual(parsed.mode, 'filters')
- self.assertEqual(parsed.test[0].name, self.testInput0)
- self.assertEqual(parsed.filter_baseline[0], 'c')
- self.assertEqual(parsed.filter_contender[0], 'd')
- self.assertFalse(parsed.benchmark_options)
- def test_filters_with_remainder(self):
- parsed = self.parser.parse_args(
- ['filters', self.testInput0, 'c', 'd', 'e'])
- self.assertEqual(parsed.mode, 'filters')
- self.assertEqual(parsed.test[0].name, self.testInput0)
- self.assertEqual(parsed.filter_baseline[0], 'c')
- self.assertEqual(parsed.filter_contender[0], 'd')
- self.assertEqual(parsed.benchmark_options, ['e'])
- def test_filters_with_remainder_after_doubleminus(self):
- parsed = self.parser.parse_args(
- ['filters', self.testInput0, 'c', 'd', '--', 'f'])
- self.assertEqual(parsed.mode, 'filters')
- self.assertEqual(parsed.test[0].name, self.testInput0)
- self.assertEqual(parsed.filter_baseline[0], 'c')
- self.assertEqual(parsed.filter_contender[0], 'd')
- self.assertEqual(parsed.benchmark_options, ['f'])
- def test_benchmarksfiltered_basic(self):
- parsed = self.parser.parse_args(
- ['benchmarksfiltered', self.testInput0, 'c', self.testInput1, 'e'])
- self.assertEqual(parsed.mode, 'benchmarksfiltered')
- self.assertEqual(parsed.test_baseline[0].name, self.testInput0)
- self.assertEqual(parsed.filter_baseline[0], 'c')
- self.assertEqual(parsed.test_contender[0].name, self.testInput1)
- self.assertEqual(parsed.filter_contender[0], 'e')
- self.assertFalse(parsed.benchmark_options)
- def test_benchmarksfiltered_with_remainder(self):
- parsed = self.parser.parse_args(
- ['benchmarksfiltered', self.testInput0, 'c', self.testInput1, 'e', 'f'])
- self.assertEqual(parsed.mode, 'benchmarksfiltered')
- self.assertEqual(parsed.test_baseline[0].name, self.testInput0)
- self.assertEqual(parsed.filter_baseline[0], 'c')
- self.assertEqual(parsed.test_contender[0].name, self.testInput1)
- self.assertEqual(parsed.filter_contender[0], 'e')
- self.assertEqual(parsed.benchmark_options[0], 'f')
- def test_benchmarksfiltered_with_remainder_after_doubleminus(self):
- parsed = self.parser.parse_args(
- ['benchmarksfiltered', self.testInput0, 'c', self.testInput1, 'e', '--', 'g'])
- self.assertEqual(parsed.mode, 'benchmarksfiltered')
- self.assertEqual(parsed.test_baseline[0].name, self.testInput0)
- self.assertEqual(parsed.filter_baseline[0], 'c')
- self.assertEqual(parsed.test_contender[0].name, self.testInput1)
- self.assertEqual(parsed.filter_contender[0], 'e')
- self.assertEqual(parsed.benchmark_options[0], 'g')
- if __name__ == '__main__':
- # unittest.main()
- main()
- # vim: tabstop=4 expandtab shiftwidth=4 softtabstop=4
- # kate: tab-width: 4; replace-tabs on; indent-width 4; tab-indents: off;
- # kate: indent-mode python; remove-trailing-spaces modified;
|