diff --git a/tensorflow/contrib/cudnn_rnn/python/kernel_tests/cudnn_rnn_ops_test.py b/tensorflow/contrib/cudnn_rnn/python/kernel_tests/cudnn_rnn_ops_test.py index 2e70d2d5ec8..aebdcea10b5 100644 --- a/tensorflow/contrib/cudnn_rnn/python/kernel_tests/cudnn_rnn_ops_test.py +++ b/tensorflow/contrib/cudnn_rnn/python/kernel_tests/cudnn_rnn_ops_test.py @@ -286,14 +286,14 @@ class CudnnRNNTestSaveRestore(TensorFlowTestCase): save_path = os.path.join(self.get_temp_dir(), "save-restore-variable-test") saver = saver_lib.Saver(write_version=saver_pb2.SaverDef.V2) - # Passing graph explictly, otherwise an old sess would be reused. + # Passing graph explicitly, otherwise an old sess would be reused. with self.test_session( use_gpu=True, graph=ops.get_default_graph()) as sess: sess.run(variables.global_variables_initializer()) params_v = sess.run(params) val = saver.save(sess, save_path) self.assertEqual(save_path, val) - # Passing graph explictly, otherwise an old sess would be reused. + # Passing graph explicitly, otherwise an old sess would be reused. with self.test_session( use_gpu=True, graph=ops.get_default_graph()) as sess: reset_params = state_ops.assign( @@ -328,14 +328,14 @@ class CudnnRNNTestSaveRestore(TensorFlowTestCase): save_path = os.path.join(self.get_temp_dir(), "save-restore-variable-test") saver = saver_lib.Saver(write_version=saver_pb2.SaverDef.V2) - # Passing graph explictly, otherwise an old sess would be reused. + # Passing graph explicitly, otherwise an old sess would be reused. with self.test_session( use_gpu=True, graph=ops.get_default_graph()) as sess: sess.run(variables.global_variables_initializer()) params_v = sess.run(param_vars) val = saver.save(sess, save_path) self.assertEqual(save_path, val) - # Passing graph explictly, otherwise an old sess would be reused. + # Passing graph explicitly, otherwise an old sess would be reused. with self.test_session( use_gpu=True, graph=ops.get_default_graph()) as sess: reset_params = [ @@ -398,14 +398,14 @@ class CudnnRNNTestSaveRestore(TensorFlowTestCase): params=params, is_training=False) total_sum = sum(map(math_ops.reduce_sum, outputs)) - # Passing graph explictly, otherwise an old sess would be reused. + # Passing graph explicitly, otherwise an old sess would be reused. with self.test_session( use_gpu=True, graph=ops.get_default_graph()) as sess: sess.run(variables.global_variables_initializer()) total_sum_v = sess.run(total_sum) val = saver.save(sess, save_path) self.assertEqual(save_path, val) - # Passing graph explictly, otherwise an old sess would be reused. + # Passing graph explicitly, otherwise an old sess would be reused. with self.test_session( use_gpu=True, graph=ops.get_default_graph()) as sess: reset_params = state_ops.assign( diff --git a/tensorflow/contrib/mpi/mpi_utils.cc b/tensorflow/contrib/mpi/mpi_utils.cc index b8e7d1a274f..8184b856264 100644 --- a/tensorflow/contrib/mpi/mpi_utils.cc +++ b/tensorflow/contrib/mpi/mpi_utils.cc @@ -61,7 +61,7 @@ void MPIUtils::InitMPI() { MPI_CHECK(MPI_Comm_size(MPI_COMM_WORLD, &number_of_procs)); MPI_CHECK(MPI_Get_processor_name(my_host_name, &len)); fprintf(stderr, - "MPI Environment initialised. Process id: %d Total processes: %d " + "MPI Environment initialized. Process id: %d Total processes: %d " "|| Hostname: %s \n", proc_id, number_of_procs, my_host_name); } diff --git a/tensorflow/core/common_runtime/simple_placer.cc b/tensorflow/core/common_runtime/simple_placer.cc index 5e6c3d164b1..f3406ac850b 100644 --- a/tensorflow/core/common_runtime/simple_placer.cc +++ b/tensorflow/core/common_runtime/simple_placer.cc @@ -682,7 +682,7 @@ Status SimplePlacer::Run() { int dst_root_id = colocation_graph.FindRoot(dst->id()); auto& src_root = colocation_graph.members_[src_root_id]; auto& dst_root = colocation_graph.members_[dst_root_id]; - // If both the source node and this node have paritally + // If both the source node and this node have partially // specified a device, then 'node's device should be // cleared: the reference edge forces 'node' to be on the // same device as the source node. diff --git a/tensorflow/core/protobuf/rewriter_config.proto b/tensorflow/core/protobuf/rewriter_config.proto index b71edb4568e..aea00b17d9b 100644 --- a/tensorflow/core/protobuf/rewriter_config.proto +++ b/tensorflow/core/protobuf/rewriter_config.proto @@ -30,7 +30,7 @@ message RewriterConfig { // Fold constants (default is OFF) Toggle constant_folding = 3; - // If true, don't remove unecessary ops from the graph + // If true, don't remove unnecessary ops from the graph bool disable_model_pruning = 2; enum MemOptType { diff --git a/tensorflow/python/framework/function.py b/tensorflow/python/framework/function.py index 34295d8c200..7220f85dc41 100644 --- a/tensorflow/python/framework/function.py +++ b/tensorflow/python/framework/function.py @@ -584,7 +584,7 @@ class _FuncGraph(ops.Graph): _FuncGraph overrides ops.Graph's create_op() so that we can keep track of all inputs into every op created inside the function. If any input is from other graphs, we keep track of it in self.capture - and substitue the input with a place holder. + and substitute the input with a place holder. Each captured input's corresponding place holder is converted into a function argument and the caller passes in the captured tensor. diff --git a/tensorflow/tools/docker/parameterized_docker_build.sh b/tensorflow/tools/docker/parameterized_docker_build.sh index b320a6222dd..830e3dcd32e 100755 --- a/tensorflow/tools/docker/parameterized_docker_build.sh +++ b/tensorflow/tools/docker/parameterized_docker_build.sh @@ -13,7 +13,7 @@ # See the License for the specific language governing permissions and # limitations under the License. # ============================================================================== -# Paramterized build and test for TensorFlow Docker images. +# Parameterized build and test for TensorFlow Docker images. # # Usage: # parameterized_docker_build.sh