aboutsummaryrefslogtreecommitdiffhomepage
path: root/tensorflow/contrib/seq2seq
diff options
context:
space:
mode:
authorGravatar Adam Roberts <adarob@google.com>2018-01-26 13:43:47 -0800
committerGravatar TensorFlower Gardener <gardener@tensorflow.org>2018-01-26 13:47:29 -0800
commit1bd5b2e6fada5334b04e2db87cf246d1a83fa533 (patch)
tree5f1a476ec11dcece49215d3fa98b7bae31160c89 /tensorflow/contrib/seq2seq
parent079a0e53311b4cf913be5ec5bd26bbb2b0649e93 (diff)
Automated g4 rollback of changelist 183321394
PiperOrigin-RevId: 183438398
Diffstat (limited to 'tensorflow/contrib/seq2seq')
-rw-r--r--tensorflow/contrib/seq2seq/python/ops/helper.py36
1 files changed, 0 insertions, 36 deletions
diff --git a/tensorflow/contrib/seq2seq/python/ops/helper.py b/tensorflow/contrib/seq2seq/python/ops/helper.py
index 6d8f786223..ef3722ee41 100644
--- a/tensorflow/contrib/seq2seq/python/ops/helper.py
+++ b/tensorflow/contrib/seq2seq/python/ops/helper.py
@@ -73,14 +73,6 @@ class Helper(object):
raise NotImplementedError("batch_size has not been implemented")
@abc.abstractproperty
- def input_shape(self):
- """Shape of each input element in batch.
-
- Returns a `TensorShape`.
- """
- raise NotImplementedError("input_shape has not been implemented")
-
- @abc.abstractproperty
def sample_ids_shape(self):
"""Shape of tensor returned by `sample`, excluding the batch dimension.
@@ -135,7 +127,6 @@ class CustomHelper(Helper):
self._sample_fn = sample_fn
self._next_inputs_fn = next_inputs_fn
self._batch_size = None
- self._input_shape = None
self._sample_ids_shape = tensor_shape.TensorShape(sample_ids_shape or [])
self._sample_ids_dtype = sample_ids_dtype or dtypes.int32
@@ -158,8 +149,6 @@ class CustomHelper(Helper):
(finished, next_inputs) = self._initialize_fn()
if self._batch_size is None:
self._batch_size = array_ops.size(finished)
- if self._input_shape is None:
- self._input_shape = next_inputs.shape[1:]
return (finished, next_inputs)
def sample(self, time, outputs, state, name=None):
@@ -195,7 +184,6 @@ class TrainingHelper(Helper):
"""
with ops.name_scope(name, "TrainingHelper", [inputs, sequence_length]):
inputs = ops.convert_to_tensor(inputs, name="inputs")
- self._inputs = inputs
if not time_major:
inputs = nest.map_structure(_transpose_batch_time, inputs)
@@ -211,17 +199,12 @@ class TrainingHelper(Helper):
lambda inp: array_ops.zeros_like(inp[0, :]), inputs)
self._batch_size = array_ops.size(sequence_length)
- self._input_shape = inputs.shape[2:]
@property
def batch_size(self):
return self._batch_size
@property
- def input_shape(self):
- return self._input_shape
-
- @property
def sample_ids_shape(self):
return tensor_shape.TensorShape([])
@@ -229,14 +212,6 @@ class TrainingHelper(Helper):
def sample_ids_dtype(self):
return dtypes.int32
- @property
- def inputs(self):
- return self._inputs
-
- @property
- def sequence_length(self):
- return self._sequence_length
-
def initialize(self, name=None):
with ops.name_scope(name, "TrainingHelperInitialize"):
finished = math_ops.equal(0, self._sequence_length)
@@ -541,17 +516,12 @@ class GreedyEmbeddingHelper(Helper):
if self._end_token.get_shape().ndims != 0:
raise ValueError("end_token must be a scalar")
self._start_inputs = self._embedding_fn(self._start_tokens)
- self._input_shape = self._start_inputs.shape[1:]
@property
def batch_size(self):
return self._batch_size
@property
- def input_shape(self):
- return self._input_shape
-
- @property
def sample_ids_shape(self):
return tensor_shape.TensorShape([])
@@ -662,8 +632,6 @@ class InferenceHelper(Helper):
self._sample_dtype = sample_dtype
self._next_inputs_fn = next_inputs_fn
self._batch_size = array_ops.shape(start_inputs)[0]
- self._input_shape = start_inputs.shape[1:]
-
self._start_inputs = ops.convert_to_tensor(
start_inputs, name="start_inputs")
@@ -672,10 +640,6 @@ class InferenceHelper(Helper):
return self._batch_size
@property
- def input_shape(self):
- return self._input_shape
-
- @property
def sample_ids_shape(self):
return self._sample_shape