mirror of
https://github.com/google/benchmark.git
synced 2024-12-01 07:16:07 +00:00
6bc17754f6
Previously compare.py ignored the --benchmarks_filter argument when loading JSON. This defeated any workflow when a single run of the benchmark was run, followed by multiple "subset reports" run against it with the 'benchmarks' command. Concretely this came up with the simple case: compare.py benchmarks a.json b.json --benchmarks_filter=BM_Example This has no practical impact on the 'filters' and 'benchmarkfiltered' comand, which do their thing at a later stage. Fixes #1484 Co-authored-by: dominic <510002+dmah42@users.noreply.github.com>
204 lines
6.7 KiB
Python
204 lines
6.7 KiB
Python
"""util.py - General utilities for running, loading, and processing benchmarks
|
|
"""
|
|
import json
|
|
import os
|
|
import re
|
|
import subprocess
|
|
import sys
|
|
import tempfile
|
|
|
|
|
|
# Input file type enumeration
|
|
IT_Invalid = 0
|
|
IT_JSON = 1
|
|
IT_Executable = 2
|
|
|
|
_num_magic_bytes = 2 if sys.platform.startswith('win') else 4
|
|
|
|
|
|
def is_executable_file(filename):
|
|
"""
|
|
Return 'True' if 'filename' names a valid file which is likely
|
|
an executable. A file is considered an executable if it starts with the
|
|
magic bytes for a EXE, Mach O, or ELF file.
|
|
"""
|
|
if not os.path.isfile(filename):
|
|
return False
|
|
with open(filename, mode='rb') as f:
|
|
magic_bytes = f.read(_num_magic_bytes)
|
|
if sys.platform == 'darwin':
|
|
return magic_bytes in [
|
|
b'\xfe\xed\xfa\xce', # MH_MAGIC
|
|
b'\xce\xfa\xed\xfe', # MH_CIGAM
|
|
b'\xfe\xed\xfa\xcf', # MH_MAGIC_64
|
|
b'\xcf\xfa\xed\xfe', # MH_CIGAM_64
|
|
b'\xca\xfe\xba\xbe', # FAT_MAGIC
|
|
b'\xbe\xba\xfe\xca' # FAT_CIGAM
|
|
]
|
|
elif sys.platform.startswith('win'):
|
|
return magic_bytes == b'MZ'
|
|
else:
|
|
return magic_bytes == b'\x7FELF'
|
|
|
|
|
|
def is_json_file(filename):
|
|
"""
|
|
Returns 'True' if 'filename' names a valid JSON output file.
|
|
'False' otherwise.
|
|
"""
|
|
try:
|
|
with open(filename, 'r') as f:
|
|
json.load(f)
|
|
return True
|
|
except BaseException:
|
|
pass
|
|
return False
|
|
|
|
|
|
def classify_input_file(filename):
|
|
"""
|
|
Return a tuple (type, msg) where 'type' specifies the classified type
|
|
of 'filename'. If 'type' is 'IT_Invalid' then 'msg' is a human readable
|
|
string representing the error.
|
|
"""
|
|
ftype = IT_Invalid
|
|
err_msg = None
|
|
if not os.path.exists(filename):
|
|
err_msg = "'%s' does not exist" % filename
|
|
elif not os.path.isfile(filename):
|
|
err_msg = "'%s' does not name a file" % filename
|
|
elif is_executable_file(filename):
|
|
ftype = IT_Executable
|
|
elif is_json_file(filename):
|
|
ftype = IT_JSON
|
|
else:
|
|
err_msg = "'%s' does not name a valid benchmark executable or JSON file" % filename
|
|
return ftype, err_msg
|
|
|
|
|
|
def check_input_file(filename):
|
|
"""
|
|
Classify the file named by 'filename' and return the classification.
|
|
If the file is classified as 'IT_Invalid' print an error message and exit
|
|
the program.
|
|
"""
|
|
ftype, msg = classify_input_file(filename)
|
|
if ftype == IT_Invalid:
|
|
print("Invalid input file: %s" % msg)
|
|
sys.exit(1)
|
|
return ftype
|
|
|
|
|
|
def find_benchmark_flag(prefix, benchmark_flags):
|
|
"""
|
|
Search the specified list of flags for a flag matching `<prefix><arg>` and
|
|
if it is found return the arg it specifies. If specified more than once the
|
|
last value is returned. If the flag is not found None is returned.
|
|
"""
|
|
assert prefix.startswith('--') and prefix.endswith('=')
|
|
result = None
|
|
for f in benchmark_flags:
|
|
if f.startswith(prefix):
|
|
result = f[len(prefix):]
|
|
return result
|
|
|
|
|
|
def remove_benchmark_flags(prefix, benchmark_flags):
|
|
"""
|
|
Return a new list containing the specified benchmark_flags except those
|
|
with the specified prefix.
|
|
"""
|
|
assert prefix.startswith('--') and prefix.endswith('=')
|
|
return [f for f in benchmark_flags if not f.startswith(prefix)]
|
|
|
|
|
|
def load_benchmark_results(fname, benchmark_filter):
|
|
"""
|
|
Read benchmark output from a file and return the JSON object.
|
|
|
|
Apply benchmark_filter, a regular expression, with nearly the same
|
|
semantics of the --benchmark_filter argument. May be None.
|
|
Note: the Python regular expression engine is used instead of the
|
|
one used by the C++ code, which may produce different results
|
|
in complex cases.
|
|
|
|
REQUIRES: 'fname' names a file containing JSON benchmark output.
|
|
"""
|
|
def benchmark_wanted(benchmark):
|
|
if benchmark_filter is None:
|
|
return True
|
|
name = benchmark.get('run_name', None) or benchmark['name']
|
|
if re.search(benchmark_filter, name):
|
|
return True
|
|
return False
|
|
|
|
with open(fname, 'r') as f:
|
|
results = json.load(f)
|
|
if 'benchmarks' in results:
|
|
results['benchmarks'] = list(filter(benchmark_wanted,
|
|
results['benchmarks']))
|
|
return results
|
|
|
|
|
|
def sort_benchmark_results(result):
|
|
benchmarks = result['benchmarks']
|
|
|
|
# From inner key to the outer key!
|
|
benchmarks = sorted(
|
|
benchmarks, key=lambda benchmark: benchmark['repetition_index'] if 'repetition_index' in benchmark else -1)
|
|
benchmarks = sorted(
|
|
benchmarks, key=lambda benchmark: 1 if 'run_type' in benchmark and benchmark['run_type'] == "aggregate" else 0)
|
|
benchmarks = sorted(
|
|
benchmarks, key=lambda benchmark: benchmark['per_family_instance_index'] if 'per_family_instance_index' in benchmark else -1)
|
|
benchmarks = sorted(
|
|
benchmarks, key=lambda benchmark: benchmark['family_index'] if 'family_index' in benchmark else -1)
|
|
|
|
result['benchmarks'] = benchmarks
|
|
return result
|
|
|
|
|
|
def run_benchmark(exe_name, benchmark_flags):
|
|
"""
|
|
Run a benchmark specified by 'exe_name' with the specified
|
|
'benchmark_flags'. The benchmark is run directly as a subprocess to preserve
|
|
real time console output.
|
|
RETURNS: A JSON object representing the benchmark output
|
|
"""
|
|
output_name = find_benchmark_flag('--benchmark_out=',
|
|
benchmark_flags)
|
|
is_temp_output = False
|
|
if output_name is None:
|
|
is_temp_output = True
|
|
thandle, output_name = tempfile.mkstemp()
|
|
os.close(thandle)
|
|
benchmark_flags = list(benchmark_flags) + \
|
|
['--benchmark_out=%s' % output_name]
|
|
|
|
cmd = [exe_name] + benchmark_flags
|
|
print("RUNNING: %s" % ' '.join(cmd))
|
|
exitCode = subprocess.call(cmd)
|
|
if exitCode != 0:
|
|
print('TEST FAILED...')
|
|
sys.exit(exitCode)
|
|
json_res = load_benchmark_results(output_name, None)
|
|
if is_temp_output:
|
|
os.unlink(output_name)
|
|
return json_res
|
|
|
|
|
|
def run_or_load_benchmark(filename, benchmark_flags):
|
|
"""
|
|
Get the results for a specified benchmark. If 'filename' specifies
|
|
an executable benchmark then the results are generated by running the
|
|
benchmark. Otherwise 'filename' must name a valid JSON output file,
|
|
which is loaded and the result returned.
|
|
"""
|
|
ftype = check_input_file(filename)
|
|
if ftype == IT_JSON:
|
|
benchmark_filter = find_benchmark_flag('--benchmark_filter=',
|
|
benchmark_flags)
|
|
return load_benchmark_results(filename, benchmark_filter)
|
|
if ftype == IT_Executable:
|
|
return run_benchmark(filename, benchmark_flags)
|
|
raise ValueError('Unknown file type %s' % ftype)
|