From 4e1221f01b157b1954d031d8e7faff6bd0f95bae Mon Sep 17 00:00:00 2001
From: Simon DENEL <simondenel1@gmail.com>
Date: Fri, 1 Jul 2016 19:56:39 +0200
Subject: [PATCH] Small typo (#3150)

* Small typo

* Previous score is 92%, not 91%.
---
 tensorflow/g3doc/tutorials/mnist/pros/index.md | 4 ++--
 1 file changed, 2 insertions(+), 2 deletions(-)

diff --git a/tensorflow/g3doc/tutorials/mnist/pros/index.md b/tensorflow/g3doc/tutorials/mnist/pros/index.md
index 12de1df66cb..324a29c02eb 100644
--- a/tensorflow/g3doc/tutorials/mnist/pros/index.md
+++ b/tensorflow/g3doc/tutorials/mnist/pros/index.md
@@ -232,7 +232,7 @@ print(accuracy.eval(feed_dict={x: mnist.test.images, y_: mnist.test.labels}))
 
 ## Build a Multilayer Convolutional Network
 
-Getting 91% accuracy on MNIST is bad. It's almost embarrassingly bad. In this
+Getting 92% accuracy on MNIST is bad. It's almost embarrassingly bad. In this
 section, we'll fix that, jumping from a very simple model to something
 moderately sophisticated: a small convolutional neural network. This will get us
 to around 99.2% accuracy -- not state of the art, but respectable.
@@ -243,7 +243,7 @@ To create this model, we're going to need to create a lot of weights and biases.
 One should generally initialize weights with a small amount of noise for
 symmetry breaking, and to prevent 0 gradients. Since we're using ReLU neurons,
 it is also good practice to initialize them with a slightly positive initial
-bias to avoid "dead neurons." Instead of doing this repeatedly while we build
+bias to avoid "dead neurons". Instead of doing this repeatedly while we build
 the model, let's create two handy functions to do it for us.
 
 ```python