aboutsummaryrefslogtreecommitdiffhomepage
path: root/tensorflow/python/layers
diff options
context:
space:
mode:
authorGravatar Anna R <annarev@google.com>2018-03-28 16:52:39 -0700
committerGravatar TensorFlower Gardener <gardener@tensorflow.org>2018-03-28 16:55:15 -0700
commit108178da2a20ea2d3899417ee932d46ba1a5c652 (patch)
tree313bd8cec176f8c9ef67b25c6484a650d1f2092a /tensorflow/python/layers
parent390e19ab990f5656e09d98624c92b3c80e52937d (diff)
Automated g4 rollback of changelist 190835392
PiperOrigin-RevId: 190858242
Diffstat (limited to 'tensorflow/python/layers')
-rw-r--r--tensorflow/python/layers/convolutional.py2
-rw-r--r--tensorflow/python/layers/convolutional_test.py6
2 files changed, 0 insertions, 8 deletions
diff --git a/tensorflow/python/layers/convolutional.py b/tensorflow/python/layers/convolutional.py
index 2d99b1688f..74e7c63fb3 100644
--- a/tensorflow/python/layers/convolutional.py
+++ b/tensorflow/python/layers/convolutional.py
@@ -180,8 +180,6 @@ class _Conv(base.Layer):
# bias_add when computing gradients. To use bias_add, we collapse Z
# and Y into a single dimension to obtain a 4D input tensor.
outputs_shape = outputs.shape.as_list()
- if outputs_shape[0] is None:
- outputs_shape[0] = -1
outputs_4d = array_ops.reshape(outputs,
[outputs_shape[0], outputs_shape[1],
outputs_shape[2] * outputs_shape[3],
diff --git a/tensorflow/python/layers/convolutional_test.py b/tensorflow/python/layers/convolutional_test.py
index cdb42f5bd1..160e732b67 100644
--- a/tensorflow/python/layers/convolutional_test.py
+++ b/tensorflow/python/layers/convolutional_test.py
@@ -325,12 +325,6 @@ class ConvTest(test.TestCase):
self.assertEqual(conv3d.kernel_constraint, k_constraint)
self.assertEqual(conv3d.bias_constraint, b_constraint)
- def testConv3DChannelsFirst(self):
- # Test case for GitHub issue 15655
- images = array_ops.placeholder(
- dtype=dtypes.float32, shape=[None, 1, 32, 32, 32])
- conv_layers.conv3d(images, 32, 9, data_format='channels_first')
-
@test_util.with_c_api
class SeparableConv1DTest(test.TestCase):