Add BenchmarkType to TestResults proto and fix logging of C++ microbenchmarks.

C++ microbenchmarks results are now normalized by iters (like python benchmarks).
Change: 149045367
This commit is contained in:
Eugene Brevdo 2017-03-02 14:08:44 -08:00 committed by TensorFlower Gardener
parent 356e599ce5
commit 54c0ff4a73
5 changed files with 40 additions and 15 deletions

View File

@ -41,8 +41,8 @@ Status TestReporter::Benchmark(int64 iters, double cpu_time, double wall_time,
double throughput) { double throughput) {
if (closed_) return Status::OK(); if (closed_) return Status::OK();
benchmark_entry_.set_iters(iters); benchmark_entry_.set_iters(iters);
benchmark_entry_.set_cpu_time(cpu_time); benchmark_entry_.set_cpu_time(cpu_time / iters);
benchmark_entry_.set_wall_time(wall_time); benchmark_entry_.set_wall_time(wall_time / iters);
benchmark_entry_.set_throughput(throughput); benchmark_entry_.set_throughput(throughput);
return Status::OK(); return Status::OK();
} }

View File

@ -174,4 +174,13 @@ message TestResults {
// Benchmark target identifier. // Benchmark target identifier.
string name = 9; string name = 9;
// The type of benchmark.
enum BenchmarkType {
UNKNOWN = 0; // Fallback for protos written before Type was introduced.
CPP_MICROBENCHMARK = 1;
PYTHON_BENCHMARK = 2;
ANDROID_BENCHMARK = 3;
}
BenchmarkType benchmark_type = 10;
}; };

View File

@ -8,7 +8,8 @@ def tf_cc_logged_benchmark(
target=None, target=None,
benchmarks="..", benchmarks="..",
tags=[], tags=[],
test_log_output_prefix=""): test_log_output_prefix="",
benchmark_type="cpp_microbenchmark"):
if not name: if not name:
fail("Must provide a name") fail("Must provide a name")
if not target: if not target:
@ -31,7 +32,8 @@ def tf_cc_logged_benchmark(
args = [ args = [
"--name=//%s:%s" % (PACKAGE_NAME, name), "--name=//%s:%s" % (PACKAGE_NAME, name),
"--test_name=" + target, "--test_name=" + target,
"--test_args=--benchmarks=%s" % benchmarks "--test_args=--benchmarks=%s" % benchmarks,
"--benchmark_type=%s" % benchmark_type,
], ],
data = [ data = [
target, target,
@ -56,4 +58,5 @@ def tf_py_logged_benchmark(
target=target, target=target,
benchmarks=benchmarks, benchmarks=benchmarks,
tags=tags, tags=tags,
test_log_output_prefix=test_log_output_prefix) test_log_output_prefix=test_log_output_prefix,
benchmark_type="python_benchmark")

View File

@ -68,8 +68,10 @@ def main(unused_args):
name = FLAGS.name name = FLAGS.name
test_name = FLAGS.test_name test_name = FLAGS.test_name
test_args = FLAGS.test_args test_args = FLAGS.test_args
test_results, _ = run_and_gather_logs_lib.run_and_gather_logs(name, test_name, benchmark_type = FLAGS.benchmark_type
test_args) test_results, _ = run_and_gather_logs_lib.run_and_gather_logs(
name, test_name=test_name, test_args=test_args,
benchmark_type=benchmark_type)
# Additional bits we receive from bazel # Additional bits we receive from bazel
test_results.build_configuration.CopyFrom(gather_build_configuration()) test_results.build_configuration.CopyFrom(gather_build_configuration())
@ -102,6 +104,11 @@ if __name__ == "__main__":
"--name", type=str, default="", help="Benchmark target identifier.") "--name", type=str, default="", help="Benchmark target identifier.")
parser.add_argument( parser.add_argument(
"--test_name", type=str, default="", help="Test target to run.") "--test_name", type=str, default="", help="Test target to run.")
parser.add_argument(
"--benchmark_type",
type=str,
default="",
help="BenchmarkType enum string (benchmark type).")
parser.add_argument( parser.add_argument(
"--test_args", "--test_args",
type=str, type=str,

View File

@ -46,15 +46,16 @@ def get_git_commit_sha():
return os.getenv("GIT_COMMIT") return os.getenv("GIT_COMMIT")
def process_test_logs(name, test_name, test_args, start_time, run_time, def process_test_logs(name, test_name, test_args, benchmark_type,
log_files): start_time, run_time, log_files):
"""Gather test information and put it in a TestResults proto. """Gather test information and put it in a TestResults proto.
Args: Args:
name: Benchmark target identifier. name: Benchmark target identifier.
test_name: A unique bazel target, e.g. "//path/to:test" test_name: A unique bazel target, e.g. "//path/to:test"
test_args: A string containing all arguments to run the target with. test_args: A string containing all arguments to run the target with.
benchmark_type: A string representing the BenchmarkType enum; the
benchmark type for this target.
start_time: Test starting time (epoch) start_time: Test starting time (epoch)
run_time: Wall time that the test ran for run_time: Wall time that the test ran for
log_files: Paths to the log files log_files: Paths to the log files
@ -68,6 +69,8 @@ def process_test_logs(name, test_name, test_args, start_time, run_time,
results.target = test_name results.target = test_name
results.start_time = start_time results.start_time = start_time
results.run_time = run_time results.run_time = run_time
results.benchmark_type = test_log_pb2.TestResults.BenchmarkType.Value(
benchmark_type.upper())
# Gather source code information # Gather source code information
git_sha = get_git_commit_sha() git_sha = get_git_commit_sha()
@ -90,13 +93,15 @@ def process_benchmarks(log_files):
return benchmarks return benchmarks
def run_and_gather_logs(name, test_name, test_args): def run_and_gather_logs(name, test_name, test_args, benchmark_type):
"""Run the bazel test given by test_name. Gather and return the logs. """Run the bazel test given by test_name. Gather and return the logs.
Args: Args:
name: Benchmark target identifier. name: Benchmark target identifier.
test_name: A unique bazel target, e.g. "//path/to:test" test_name: A unique bazel target, e.g. "//path/to:test"
test_args: A string containing all arguments to run the target with. test_args: A string containing all arguments to run the target with.
benchmark_type: A string representing the BenchmarkType enum; the
benchmark type for this target.
Returns: Returns:
A tuple (test_results, mangled_test_name), where A tuple (test_results, mangled_test_name), where
@ -145,8 +150,9 @@ def run_and_gather_logs(name, test_name, test_args):
return (process_test_logs( return (process_test_logs(
name, name,
test_name, test_name=test_name,
test_args, test_args=test_args,
benchmark_type=benchmark_type,
start_time=int(start_time), start_time=int(start_time),
run_time=run_time, run_time=run_time,
log_files=log_files), mangled_test_name) log_files=log_files), mangled_test_name)