aboutsummaryrefslogtreecommitdiffhomepage
path: root/tensorflow/examples/tutorials/mnist/mnist_with_summaries.py
diff options
context:
space:
mode:
Diffstat (limited to 'tensorflow/examples/tutorials/mnist/mnist_with_summaries.py')
-rw-r--r--tensorflow/examples/tutorials/mnist/mnist_with_summaries.py4
1 files changed, 2 insertions, 2 deletions
diff --git a/tensorflow/examples/tutorials/mnist/mnist_with_summaries.py b/tensorflow/examples/tutorials/mnist/mnist_with_summaries.py
index 75ea0b9c67..698c97ca1d 100644
--- a/tensorflow/examples/tutorials/mnist/mnist_with_summaries.py
+++ b/tensorflow/examples/tutorials/mnist/mnist_with_summaries.py
@@ -14,7 +14,7 @@
# ==============================================================================
"""A simple MNIST classifier which displays summaries in TensorBoard.
- This is an unimpressive MNIST model, but it is a good example of using
+This is an unimpressive MNIST model, but it is a good example of using
tf.name_scope to make a graph legible in the TensorBoard graph explorer, and of
naming summary tags so that they are grouped meaningfully in TensorBoard.
@@ -78,7 +78,7 @@ def train():
def nn_layer(input_tensor, input_dim, output_dim, layer_name, act=tf.nn.relu):
"""Reusable code for making a simple neural net layer.
- It does a matrix multiply, bias add, and then uses relu to nonlinearize.
+ It does a matrix multiply, bias add, and then uses ReLU to nonlinearize.
It also sets up name scoping so that the resultant graph is easy to read,
and adds a number of summary ops.
"""