Merge pull request #29193 from chosungmann:fix-typos

PiperOrigin-RevId: 251835242
This commit is contained in:
TensorFlower Gardener 2019-06-06 05:59:19 -07:00
commit 05d496e412
5 changed files with 11 additions and 11 deletions

View File

@ -267,10 +267,10 @@ TEST(CAPI, MonitoringMultipleSampler) {
TFE_MonitoringSamplerCellAdd(cell1, 2.0);
TF_Buffer* result1 = TF_NewBuffer();
TFE_MonitoringSamplerCellValue(cell1, result1);
tensorflow::HistogramProto hitogram1;
EXPECT_TRUE(hitogram1.ParseFromString(
tensorflow::HistogramProto histogram1;
EXPECT_TRUE(histogram1.ParseFromString(
{reinterpret_cast<const char*>(result1->data), result1->length}));
EXPECT_EQ(hitogram1.sum(), 3.0);
EXPECT_EQ(histogram1.sum(), 3.0);
delete result1;
auto* sampler2 = TFE_MonitoringNewSampler2("test/sampler2", buckets, status,
@ -281,10 +281,10 @@ TEST(CAPI, MonitoringMultipleSampler) {
TFE_MonitoringSamplerCellAdd(cell2, 3.0);
TF_Buffer* result2 = TF_NewBuffer();
TFE_MonitoringSamplerCellValue(cell2, result2);
tensorflow::HistogramProto hitogram2;
EXPECT_TRUE(hitogram2.ParseFromString(
tensorflow::HistogramProto histogram2;
EXPECT_TRUE(histogram2.ParseFromString(
{reinterpret_cast<const char*>(result2->data), result2->length}));
EXPECT_EQ(hitogram2.sum(), 5.0);
EXPECT_EQ(histogram2.sum(), 5.0);
delete result2;
TFE_MonitoringDeleteBuckets(buckets);

View File

@ -42,7 +42,7 @@ class Client {
// Compile the computation with the given argument shapes and returns the
// handle to the compiled executable. The compiled executable is cached on the
// service, and the returned handle can be used for exection without
// service, and the returned handle can be used for execution without
// re-compile.
// * The shape and layout of the arguments being executed with will affect how
// the computation is compiled. If argument_shapes is empty, the parameters'

View File

@ -110,7 +110,7 @@ __xla_cpu_runtime_MKLSingleThreadedMatMulF32(const void* run_options_ptr,
int64 m, int64 n, int64 k,
int32 transpose_lhs,
int32 transpose_rhs) {
// Set the thread number to 1 for single threaded excution.
// Set the thread number to 1 for single threaded execution.
int prev_num_threads = mkl_set_num_threads_local(1);
MatMulF32(nullptr, out, lhs, rhs, m, n, k, transpose_lhs, transpose_rhs);
// Set thread number back to the previous number.
@ -123,7 +123,7 @@ __xla_cpu_runtime_MKLSingleThreadedMatMulF64(const void* run_options_ptr,
double* rhs, int64 m, int64 n,
int64 k, int32 transpose_lhs,
int32 transpose_rhs) {
// Set the thread number to 1 for single threaded excution.
// Set the thread number to 1 for single threaded execution.
int prev_num_threads = mkl_set_num_threads_local(1);
MatMulF64(nullptr, out, lhs, rhs, m, n, k, transpose_lhs, transpose_rhs);
// Set thread number back to the previous number.

View File

@ -189,7 +189,7 @@
"\n",
"Implement a python function `bisecting_line_search(f, a, b, epsilon)` which returns a value such that `tf.abs(f(value)) < epsilon`.\n",
"\n",
"One thing to keep in mind: python's `==` opertor is not overloaded on Tensors, so you need to use `tf.equal` to compare for equality."
"One thing to keep in mind: python's `==` operator is not overloaded on Tensors, so you need to use `tf.equal` to compare for equality."
]
},
{

View File

@ -893,7 +893,7 @@ def _make_execution_function_with_cloning(model, mode):
distributed_function = _make_graph_execution_function(model, mode)
# We cache the distributed execution function on the model since creating
# distributed models and exection functions are expensive.
# distributed models and execution functions are expensive.
distributed_model._distributed_function = distributed_function
distributed_model._recompile_exec_function = False
return distributed_function