# Copyright 2015 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # ============================================================================== """Ops for building neural network seq2seq decoders and losses. See the [Contrib Seq2seq](https://tensorflow.org/api_guides/python/contrib.seq2seq) guide. """ from __future__ import absolute_import from __future__ import division from __future__ import print_function # pylint: disable=unused-import,wildcard-import,line-too-long from tensorflow.contrib.seq2seq.python.ops.attention_wrapper import * from tensorflow.contrib.seq2seq.python.ops.basic_decoder import * from tensorflow.contrib.seq2seq.python.ops.beam_search_decoder import * from tensorflow.contrib.seq2seq.python.ops.beam_search_ops import * from tensorflow.contrib.seq2seq.python.ops.decoder import * from tensorflow.contrib.seq2seq.python.ops.helper import * from tensorflow.contrib.seq2seq.python.ops.loss import * from tensorflow.python.util.all_util import remove_undocumented # pylint: enable=unused-import,widcard-import,line-too-long _allowed_symbols = [ "sequence_loss", "Decoder", "dynamic_decode", "BasicDecoder", "BasicDecoderOutput", "BeamSearchDecoder", "BeamSearchDecoderOutput", "BeamSearchDecoderState", "Helper", "CustomHelper", "FinalBeamSearchDecoderOutput", "gather_tree", "GreedyEmbeddingHelper", "InferenceHelper", "SampleEmbeddingHelper", "ScheduledEmbeddingTrainingHelper", "ScheduledOutputTrainingHelper", "TrainingHelper", "BahdanauAttention", "LuongAttention", "hardmax", "AttentionWrapperState", "AttentionWrapper", "AttentionMechanism", "tile_batch", "safe_cumprod", "monotonic_attention", "monotonic_probability_fn", "BahdanauMonotonicAttention", "LuongMonotonicAttention", ] remove_undocumented(__name__, _allowed_symbols)