aboutsummaryrefslogtreecommitdiffhomepage
path: root/tensorflow/python/kernel_tests/logging_ops_test.py
blob: 18ca441b23c09382dd1a6cafec74a0d3523fc220 (plain)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
"""Tests for tensorflow.kernels.logging_ops."""

import tensorflow.python.platform

import tensorflow as tf


class LoggingOpsTest(tf.test.TestCase):

  def testAssertDivideByZero(self):
    with self.test_session() as sess:
      epsilon = tf.convert_to_tensor(1e-20)
      x = tf.convert_to_tensor(0.0)
      y = tf.convert_to_tensor(1.0)
      z = tf.convert_to_tensor(2.0)
      # assert(epsilon < y)
      # z / y
      with sess.graph.control_dependencies(
          [tf.Assert(tf.less(epsilon, y), ["Divide-by-zero"])]):
        out = tf.div(z, y)
      self.assertAllEqual(2.0, out.eval())
      # assert(epsilon < x)
      # z / x
      #
      # This tests printing out multiple tensors
      with sess.graph.control_dependencies(
          [tf.Assert(tf.less(epsilon, x),
                     ["Divide-by-zero", "less than x"])]):
        out = tf.div(z, x)
      with self.assertRaisesOpError("less than x"):
        out.eval()


class PrintGradientTest(tf.test.TestCase):

  def testPrintGradient(self):
    with self.test_session():
      inp = tf.constant(2.0, shape=[100, 32], name="in")
      w = tf.constant(4.0, shape=[10, 100], name="w")
      wx = tf.matmul(w, inp, name="wx")
      wx_print = tf.Print(wx, [w, w, w])
      wx_grad = tf.gradients(wx, w)[0]
      wx_print_grad = tf.gradients(wx_print, w)[0]
      wxg = wx_grad.eval()
      wxpg = wx_print_grad.eval()
    self.assertAllEqual(wxg, wxpg)


if __name__ == "__main__":
  tf.test.main()