Fix broken Guitar benchmark target //learning/brain/contrib/data:csv_dataset_benchmark and //learning/brain/contrib/data:map_and_batch_benchmark.

PiperOrigin-RevId: 286513162
Change-Id: I74441fd6b478af8ecc33a5cd1a3ef1b3ed2cda14
This commit is contained in:
Hye Soo Yang 2019-12-19 20:34:45 -08:00 committed by TensorFlower Gardener
parent 67d1f027b1
commit dd6df67b9b
2 changed files with 2 additions and 2 deletions

View File

@ -51,7 +51,7 @@ class CsvDatasetBenchmark(test.Benchmark):
self._filenames = []
for n in self._num_cols:
fn = os.path.join(self._temp_dir, 'file%d.csv' % n)
with open(fn, 'wb') as f:
with open(fn, 'w') as f:
# Just write 100 rows and use `repeat`... Assumes the cost
# of creating an iterator is not significant
row = ','.join(str_val for _ in range(n))

View File

@ -116,7 +116,7 @@ class MapAndBatchBenchmark(test.Benchmark):
def name(method, label, num_calls, inter_op, element_size, batch_size):
return ("%s_id_%s_num_calls_%d_inter_op_%d_elem_size_%d_batch_size_%d" % (
method,
hashlib.sha1(label).hexdigest()[:8],
hashlib.sha1((label).encode("utf-8")).hexdigest()[:8],
num_calls,
inter_op,
element_size,