forked from OSchip/llvm-project
				
			
		
			
				
	
	
		
			317 lines
		
	
	
		
			12 KiB
		
	
	
	
		
			Python
		
	
	
	
			
		
		
	
	
			317 lines
		
	
	
		
			12 KiB
		
	
	
	
		
			Python
		
	
	
	
| #!/usr/bin/env python
 | |
| 
 | |
| """
 | |
| compare.py - versatile benchmark output compare tool
 | |
| """
 | |
| 
 | |
| import argparse
 | |
| from argparse import ArgumentParser
 | |
| import sys
 | |
| import gbench
 | |
| from gbench import util, report
 | |
| from gbench.util import *
 | |
| 
 | |
| 
 | |
| def check_inputs(in1, in2, flags):
 | |
|     """
 | |
|     Perform checking on the user provided inputs and diagnose any abnormalities
 | |
|     """
 | |
|     in1_kind, in1_err = classify_input_file(in1)
 | |
|     in2_kind, in2_err = classify_input_file(in2)
 | |
|     output_file = find_benchmark_flag('--benchmark_out=', flags)
 | |
|     output_type = find_benchmark_flag('--benchmark_out_format=', flags)
 | |
|     if in1_kind == IT_Executable and in2_kind == IT_Executable and output_file:
 | |
|         print(("WARNING: '--benchmark_out=%s' will be passed to both "
 | |
|                "benchmarks causing it to be overwritten") % output_file)
 | |
|     if in1_kind == IT_JSON and in2_kind == IT_JSON and len(flags) > 0:
 | |
|         print("WARNING: passing optional flags has no effect since both "
 | |
|               "inputs are JSON")
 | |
|     if output_type is not None and output_type != 'json':
 | |
|         print(("ERROR: passing '--benchmark_out_format=%s' to 'compare.py`"
 | |
|                " is not supported.") % output_type)
 | |
|         sys.exit(1)
 | |
| 
 | |
| 
 | |
| def create_parser():
 | |
|     parser = ArgumentParser(
 | |
|         description='versatile benchmark output compare tool')
 | |
|     subparsers = parser.add_subparsers(
 | |
|         help='This tool has multiple modes of operation:',
 | |
|         dest='mode')
 | |
| 
 | |
|     parser_a = subparsers.add_parser(
 | |
|         'benchmarks',
 | |
|         help='The most simple use-case, compare all the output of these two benchmarks')
 | |
|     baseline = parser_a.add_argument_group(
 | |
|         'baseline', 'The benchmark baseline')
 | |
|     baseline.add_argument(
 | |
|         'test_baseline',
 | |
|         metavar='test_baseline',
 | |
|         type=argparse.FileType('r'),
 | |
|         nargs=1,
 | |
|         help='A benchmark executable or JSON output file')
 | |
|     contender = parser_a.add_argument_group(
 | |
|         'contender', 'The benchmark that will be compared against the baseline')
 | |
|     contender.add_argument(
 | |
|         'test_contender',
 | |
|         metavar='test_contender',
 | |
|         type=argparse.FileType('r'),
 | |
|         nargs=1,
 | |
|         help='A benchmark executable or JSON output file')
 | |
|     parser_a.add_argument(
 | |
|         'benchmark_options',
 | |
|         metavar='benchmark_options',
 | |
|         nargs=argparse.REMAINDER,
 | |
|         help='Arguments to pass when running benchmark executables')
 | |
| 
 | |
|     parser_b = subparsers.add_parser(
 | |
|         'filters', help='Compare filter one with the filter two of benchmark')
 | |
|     baseline = parser_b.add_argument_group(
 | |
|         'baseline', 'The benchmark baseline')
 | |
|     baseline.add_argument(
 | |
|         'test',
 | |
|         metavar='test',
 | |
|         type=argparse.FileType('r'),
 | |
|         nargs=1,
 | |
|         help='A benchmark executable or JSON output file')
 | |
|     baseline.add_argument(
 | |
|         'filter_baseline',
 | |
|         metavar='filter_baseline',
 | |
|         type=str,
 | |
|         nargs=1,
 | |
|         help='The first filter, that will be used as baseline')
 | |
|     contender = parser_b.add_argument_group(
 | |
|         'contender', 'The benchmark that will be compared against the baseline')
 | |
|     contender.add_argument(
 | |
|         'filter_contender',
 | |
|         metavar='filter_contender',
 | |
|         type=str,
 | |
|         nargs=1,
 | |
|         help='The second filter, that will be compared against the baseline')
 | |
|     parser_b.add_argument(
 | |
|         'benchmark_options',
 | |
|         metavar='benchmark_options',
 | |
|         nargs=argparse.REMAINDER,
 | |
|         help='Arguments to pass when running benchmark executables')
 | |
| 
 | |
|     parser_c = subparsers.add_parser(
 | |
|         'benchmarksfiltered',
 | |
|         help='Compare filter one of first benchmark with filter two of the second benchmark')
 | |
|     baseline = parser_c.add_argument_group(
 | |
|         'baseline', 'The benchmark baseline')
 | |
|     baseline.add_argument(
 | |
|         'test_baseline',
 | |
|         metavar='test_baseline',
 | |
|         type=argparse.FileType('r'),
 | |
|         nargs=1,
 | |
|         help='A benchmark executable or JSON output file')
 | |
|     baseline.add_argument(
 | |
|         'filter_baseline',
 | |
|         metavar='filter_baseline',
 | |
|         type=str,
 | |
|         nargs=1,
 | |
|         help='The first filter, that will be used as baseline')
 | |
|     contender = parser_c.add_argument_group(
 | |
|         'contender', 'The benchmark that will be compared against the baseline')
 | |
|     contender.add_argument(
 | |
|         'test_contender',
 | |
|         metavar='test_contender',
 | |
|         type=argparse.FileType('r'),
 | |
|         nargs=1,
 | |
|         help='The second benchmark executable or JSON output file, that will be compared against the baseline')
 | |
|     contender.add_argument(
 | |
|         'filter_contender',
 | |
|         metavar='filter_contender',
 | |
|         type=str,
 | |
|         nargs=1,
 | |
|         help='The second filter, that will be compared against the baseline')
 | |
|     parser_c.add_argument(
 | |
|         'benchmark_options',
 | |
|         metavar='benchmark_options',
 | |
|         nargs=argparse.REMAINDER,
 | |
|         help='Arguments to pass when running benchmark executables')
 | |
| 
 | |
|     return parser
 | |
| 
 | |
| 
 | |
| def main():
 | |
|     # Parse the command line flags
 | |
|     parser = create_parser()
 | |
|     args, unknown_args = parser.parse_known_args()
 | |
|     if args.mode is None:
 | |
|       parser.print_help()
 | |
|       exit(1)
 | |
|     assert not unknown_args
 | |
|     benchmark_options = args.benchmark_options
 | |
| 
 | |
|     if args.mode == 'benchmarks':
 | |
|         test_baseline = args.test_baseline[0].name
 | |
|         test_contender = args.test_contender[0].name
 | |
|         filter_baseline = ''
 | |
|         filter_contender = ''
 | |
| 
 | |
|         # NOTE: if test_baseline == test_contender, you are analyzing the stdev
 | |
| 
 | |
|         description = 'Comparing %s to %s' % (test_baseline, test_contender)
 | |
|     elif args.mode == 'filters':
 | |
|         test_baseline = args.test[0].name
 | |
|         test_contender = args.test[0].name
 | |
|         filter_baseline = args.filter_baseline[0]
 | |
|         filter_contender = args.filter_contender[0]
 | |
| 
 | |
|         # NOTE: if filter_baseline == filter_contender, you are analyzing the
 | |
|         # stdev
 | |
| 
 | |
|         description = 'Comparing %s to %s (from %s)' % (
 | |
|             filter_baseline, filter_contender, args.test[0].name)
 | |
|     elif args.mode == 'benchmarksfiltered':
 | |
|         test_baseline = args.test_baseline[0].name
 | |
|         test_contender = args.test_contender[0].name
 | |
|         filter_baseline = args.filter_baseline[0]
 | |
|         filter_contender = args.filter_contender[0]
 | |
| 
 | |
|         # NOTE: if test_baseline == test_contender and
 | |
|         # filter_baseline == filter_contender, you are analyzing the stdev
 | |
| 
 | |
|         description = 'Comparing %s (from %s) to %s (from %s)' % (
 | |
|             filter_baseline, test_baseline, filter_contender, test_contender)
 | |
|     else:
 | |
|         # should never happen
 | |
|         print("Unrecognized mode of operation: '%s'" % args.mode)
 | |
|         parser.print_help()
 | |
|         exit(1)
 | |
| 
 | |
|     check_inputs(test_baseline, test_contender, benchmark_options)
 | |
| 
 | |
|     options_baseline = []
 | |
|     options_contender = []
 | |
| 
 | |
|     if filter_baseline and filter_contender:
 | |
|         options_baseline = ['--benchmark_filter=%s' % filter_baseline]
 | |
|         options_contender = ['--benchmark_filter=%s' % filter_contender]
 | |
| 
 | |
|     # Run the benchmarks and report the results
 | |
|     json1 = json1_orig = gbench.util.run_or_load_benchmark(
 | |
|         test_baseline, benchmark_options + options_baseline)
 | |
|     json2 = json2_orig = gbench.util.run_or_load_benchmark(
 | |
|         test_contender, benchmark_options + options_contender)
 | |
| 
 | |
|     # Now, filter the benchmarks so that the difference report can work
 | |
|     if filter_baseline and filter_contender:
 | |
|         replacement = '[%s vs. %s]' % (filter_baseline, filter_contender)
 | |
|         json1 = gbench.report.filter_benchmark(
 | |
|             json1_orig, filter_baseline, replacement)
 | |
|         json2 = gbench.report.filter_benchmark(
 | |
|             json2_orig, filter_contender, replacement)
 | |
| 
 | |
|     # Diff and output
 | |
|     output_lines = gbench.report.generate_difference_report(json1, json2)
 | |
|     print(description)
 | |
|     for ln in output_lines:
 | |
|         print(ln)
 | |
| 
 | |
| 
 | |
| import unittest
 | |
| 
 | |
| 
 | |
| class TestParser(unittest.TestCase):
 | |
|     def setUp(self):
 | |
|         self.parser = create_parser()
 | |
|         testInputs = os.path.join(
 | |
|             os.path.dirname(
 | |
|                 os.path.realpath(__file__)),
 | |
|             'gbench',
 | |
|             'Inputs')
 | |
|         self.testInput0 = os.path.join(testInputs, 'test1_run1.json')
 | |
|         self.testInput1 = os.path.join(testInputs, 'test1_run2.json')
 | |
| 
 | |
|     def test_benchmarks_basic(self):
 | |
|         parsed = self.parser.parse_args(
 | |
|             ['benchmarks', self.testInput0, self.testInput1])
 | |
|         self.assertEqual(parsed.mode, 'benchmarks')
 | |
|         self.assertEqual(parsed.test_baseline[0].name, self.testInput0)
 | |
|         self.assertEqual(parsed.test_contender[0].name, self.testInput1)
 | |
|         self.assertFalse(parsed.benchmark_options)
 | |
| 
 | |
|     def test_benchmarks_with_remainder(self):
 | |
|         parsed = self.parser.parse_args(
 | |
|             ['benchmarks', self.testInput0, self.testInput1, 'd'])
 | |
|         self.assertEqual(parsed.mode, 'benchmarks')
 | |
|         self.assertEqual(parsed.test_baseline[0].name, self.testInput0)
 | |
|         self.assertEqual(parsed.test_contender[0].name, self.testInput1)
 | |
|         self.assertEqual(parsed.benchmark_options, ['d'])
 | |
| 
 | |
|     def test_benchmarks_with_remainder_after_doubleminus(self):
 | |
|         parsed = self.parser.parse_args(
 | |
|             ['benchmarks', self.testInput0, self.testInput1, '--', 'e'])
 | |
|         self.assertEqual(parsed.mode, 'benchmarks')
 | |
|         self.assertEqual(parsed.test_baseline[0].name, self.testInput0)
 | |
|         self.assertEqual(parsed.test_contender[0].name, self.testInput1)
 | |
|         self.assertEqual(parsed.benchmark_options, ['e'])
 | |
| 
 | |
|     def test_filters_basic(self):
 | |
|         parsed = self.parser.parse_args(
 | |
|             ['filters', self.testInput0, 'c', 'd'])
 | |
|         self.assertEqual(parsed.mode, 'filters')
 | |
|         self.assertEqual(parsed.test[0].name, self.testInput0)
 | |
|         self.assertEqual(parsed.filter_baseline[0], 'c')
 | |
|         self.assertEqual(parsed.filter_contender[0], 'd')
 | |
|         self.assertFalse(parsed.benchmark_options)
 | |
| 
 | |
|     def test_filters_with_remainder(self):
 | |
|         parsed = self.parser.parse_args(
 | |
|             ['filters', self.testInput0, 'c', 'd', 'e'])
 | |
|         self.assertEqual(parsed.mode, 'filters')
 | |
|         self.assertEqual(parsed.test[0].name, self.testInput0)
 | |
|         self.assertEqual(parsed.filter_baseline[0], 'c')
 | |
|         self.assertEqual(parsed.filter_contender[0], 'd')
 | |
|         self.assertEqual(parsed.benchmark_options, ['e'])
 | |
| 
 | |
|     def test_filters_with_remainder_after_doubleminus(self):
 | |
|         parsed = self.parser.parse_args(
 | |
|             ['filters', self.testInput0, 'c', 'd', '--', 'f'])
 | |
|         self.assertEqual(parsed.mode, 'filters')
 | |
|         self.assertEqual(parsed.test[0].name, self.testInput0)
 | |
|         self.assertEqual(parsed.filter_baseline[0], 'c')
 | |
|         self.assertEqual(parsed.filter_contender[0], 'd')
 | |
|         self.assertEqual(parsed.benchmark_options, ['f'])
 | |
| 
 | |
|     def test_benchmarksfiltered_basic(self):
 | |
|         parsed = self.parser.parse_args(
 | |
|             ['benchmarksfiltered', self.testInput0, 'c', self.testInput1, 'e'])
 | |
|         self.assertEqual(parsed.mode, 'benchmarksfiltered')
 | |
|         self.assertEqual(parsed.test_baseline[0].name, self.testInput0)
 | |
|         self.assertEqual(parsed.filter_baseline[0], 'c')
 | |
|         self.assertEqual(parsed.test_contender[0].name, self.testInput1)
 | |
|         self.assertEqual(parsed.filter_contender[0], 'e')
 | |
|         self.assertFalse(parsed.benchmark_options)
 | |
| 
 | |
|     def test_benchmarksfiltered_with_remainder(self):
 | |
|         parsed = self.parser.parse_args(
 | |
|             ['benchmarksfiltered', self.testInput0, 'c', self.testInput1, 'e', 'f'])
 | |
|         self.assertEqual(parsed.mode, 'benchmarksfiltered')
 | |
|         self.assertEqual(parsed.test_baseline[0].name, self.testInput0)
 | |
|         self.assertEqual(parsed.filter_baseline[0], 'c')
 | |
|         self.assertEqual(parsed.test_contender[0].name, self.testInput1)
 | |
|         self.assertEqual(parsed.filter_contender[0], 'e')
 | |
|         self.assertEqual(parsed.benchmark_options[0], 'f')
 | |
| 
 | |
|     def test_benchmarksfiltered_with_remainder_after_doubleminus(self):
 | |
|         parsed = self.parser.parse_args(
 | |
|             ['benchmarksfiltered', self.testInput0, 'c', self.testInput1, 'e', '--', 'g'])
 | |
|         self.assertEqual(parsed.mode, 'benchmarksfiltered')
 | |
|         self.assertEqual(parsed.test_baseline[0].name, self.testInput0)
 | |
|         self.assertEqual(parsed.filter_baseline[0], 'c')
 | |
|         self.assertEqual(parsed.test_contender[0].name, self.testInput1)
 | |
|         self.assertEqual(parsed.filter_contender[0], 'e')
 | |
|         self.assertEqual(parsed.benchmark_options[0], 'g')
 | |
| 
 | |
| 
 | |
| if __name__ == '__main__':
 | |
|     # unittest.main()
 | |
|     main()
 | |
| 
 | |
| # vim: tabstop=4 expandtab shiftwidth=4 softtabstop=4
 | |
| # kate: tab-width: 4; replace-tabs on; indent-width 4; tab-indents: off;
 | |
| # kate: indent-mode python; remove-trailing-spaces modified;
 |