aboutsummaryrefslogtreecommitdiffhomepage
path: root/tensorflow/python
diff options
context:
space:
mode:
authorGravatar Anna R <annarev@google.com>2018-02-08 13:43:20 -0800
committerGravatar TensorFlower Gardener <gardener@tensorflow.org>2018-02-08 13:49:36 -0800
commitfa3fb289ba6a1718f9c76b2277a58f95f5e878ab (patch)
treeda19b938d1afc7246cee0aae68ab265d74740ce3 /tensorflow/python
parent597377fca28c76306e749f78f8073f55726d54c9 (diff)
Adding tf_export decorators/calls to TensorFlow functions and constants.
PiperOrigin-RevId: 185044705
Diffstat (limited to 'tensorflow/python')
-rw-r--r--tensorflow/python/estimator/estimator.py10
-rw-r--r--tensorflow/python/estimator/exporter.py4
-rw-r--r--tensorflow/python/estimator/model_fn.py3
-rw-r--r--tensorflow/python/estimator/run_config.py2
-rw-r--r--tensorflow/python/estimator/training.py4
-rw-r--r--tensorflow/python/estimator/warm_starting_util.py3
-rw-r--r--tensorflow/python/keras/_impl/keras/datasets/boston_housing.py2
-rw-r--r--tensorflow/python/keras/_impl/keras/datasets/cifar10.py2
-rw-r--r--tensorflow/python/keras/_impl/keras/datasets/cifar100.py2
-rw-r--r--tensorflow/python/keras/_impl/keras/datasets/imdb.py3
-rw-r--r--tensorflow/python/keras/_impl/keras/datasets/mnist.py2
-rw-r--r--tensorflow/python/keras/_impl/keras/datasets/reuters.py3
-rw-r--r--tensorflow/python/layers/base.py3
-rw-r--r--tensorflow/python/layers/convolutional.py15
-rw-r--r--tensorflow/python/layers/core.py7
-rw-r--r--tensorflow/python/layers/network.py2
-rw-r--r--tensorflow/python/layers/normalization.py3
-rw-r--r--tensorflow/python/layers/pooling.py13
18 files changed, 80 insertions, 3 deletions
diff --git a/tensorflow/python/estimator/estimator.py b/tensorflow/python/estimator/estimator.py
index 17fab3df4d..5d36108bbf 100644
--- a/tensorflow/python/estimator/estimator.py
+++ b/tensorflow/python/estimator/estimator.py
@@ -57,12 +57,14 @@ from tensorflow.python.training import training_util
from tensorflow.python.util import compat
from tensorflow.python.util import compat_internal
from tensorflow.python.util import nest
+from tensorflow.python.util.tf_export import tf_export
_VALID_MODEL_FN_ARGS = set(
['features', 'labels', 'mode', 'params', 'self', 'config'])
+@tf_export('estimator.Estimator')
class Estimator(object):
"""Estimator class to train and evaluate TensorFlow models.
@@ -502,9 +504,11 @@ class Estimator(object):
def _assert_members_are_not_overridden(self):
"""Asserts members of `Estimator` are not overridden."""
- allowed_overrides = set(['_call_input_fn', '_create_global_step',
- '_convert_train_steps_to_hooks',
- '_convert_eval_steps_to_hooks'])
+ allowed_overrides = set([
+ '_call_input_fn', '_create_global_step',
+ '_convert_train_steps_to_hooks', '_convert_eval_steps_to_hooks',
+ '_tf_api_names'
+ ])
estimator_members = set([m for m in Estimator.__dict__.keys()
if not m.startswith('__')])
subclass_members = set(self.__class__.__dict__.keys())
diff --git a/tensorflow/python/estimator/exporter.py b/tensorflow/python/estimator/exporter.py
index ba522f396d..a3f04626d1 100644
--- a/tensorflow/python/estimator/exporter.py
+++ b/tensorflow/python/estimator/exporter.py
@@ -25,8 +25,10 @@ from tensorflow.python.estimator import gc
from tensorflow.python.framework import errors_impl
from tensorflow.python.platform import gfile
from tensorflow.python.platform import tf_logging
+from tensorflow.python.util.tf_export import tf_export
+@tf_export('estimator.Exporter')
class Exporter(object):
"""A class representing a type of model export."""
@@ -123,6 +125,7 @@ class _SavedModelExporter(Exporter):
return export_result
+@tf_export('estimator.FinalExporter')
class FinalExporter(Exporter):
"""This class exports the serving graph and checkpoints in the end.
@@ -174,6 +177,7 @@ class FinalExporter(Exporter):
is_the_final_export)
+@tf_export('estimator.LatestExporter')
class LatestExporter(Exporter):
"""This class regularly exports the serving graph and checkpoints.
diff --git a/tensorflow/python/estimator/model_fn.py b/tensorflow/python/estimator/model_fn.py
index b08f83fc56..8111ab564c 100644
--- a/tensorflow/python/estimator/model_fn.py
+++ b/tensorflow/python/estimator/model_fn.py
@@ -31,8 +31,10 @@ from tensorflow.python.saved_model import signature_constants
from tensorflow.python.training import monitored_session
from tensorflow.python.training import session_run_hook
from tensorflow.python.util import nest
+from tensorflow.python.util.tf_export import tf_export
+@tf_export('estimator.ModeKeys')
class ModeKeys(object):
"""Standard names for model modes.
@@ -52,6 +54,7 @@ LOSS_METRIC_KEY = 'loss'
AVERAGE_LOSS_METRIC_KEY = 'average_loss'
+@tf_export('estimator.EstimatorSpec')
class EstimatorSpec(
collections.namedtuple('EstimatorSpec', [
'mode', 'predictions', 'loss', 'train_op', 'eval_metric_ops',
diff --git a/tensorflow/python/estimator/run_config.py b/tensorflow/python/estimator/run_config.py
index 0c636a8da1..3e021242c4 100644
--- a/tensorflow/python/estimator/run_config.py
+++ b/tensorflow/python/estimator/run_config.py
@@ -28,6 +28,7 @@ from tensorflow.core.protobuf import config_pb2
from tensorflow.python.platform import tf_logging as logging
from tensorflow.python.training import server_lib
from tensorflow.python.util import compat_internal
+from tensorflow.python.util.tf_export import tf_export
_USE_DEFAULT = object()
@@ -286,6 +287,7 @@ class TaskType(object):
EVALUATOR = 'evaluator'
+@tf_export('estimator.RunConfig')
class RunConfig(object):
"""This class specifies the configurations for an `Estimator` run."""
diff --git a/tensorflow/python/estimator/training.py b/tensorflow/python/estimator/training.py
index 2e84c5014f..63328dcfb5 100644
--- a/tensorflow/python/estimator/training.py
+++ b/tensorflow/python/estimator/training.py
@@ -35,6 +35,7 @@ from tensorflow.python.training import basic_session_run_hooks
from tensorflow.python.training import server_lib
from tensorflow.python.training import session_run_hook
from tensorflow.python.util import compat
+from tensorflow.python.util.tf_export import tf_export
_MAX_DELAY_SECS = 60
_DELAY_SECS_PER_WORKER = 5
@@ -114,6 +115,7 @@ def _is_google_env():
return tf_config.get(_ENVIRONMENT_KEY) == _ENVIRONMENT_GOOGLE_VALUE
+@tf_export('estimator.TrainSpec')
class TrainSpec(
collections.namedtuple('TrainSpec', ['input_fn', 'max_steps', 'hooks'])):
"""Configuration for the "train" part for the `train_and_evaluate` call.
@@ -158,6 +160,7 @@ class TrainSpec(
cls, input_fn=input_fn, max_steps=max_steps, hooks=hooks)
+@tf_export('estimator.EvalSpec')
class EvalSpec(
collections.namedtuple('EvalSpec', [
'input_fn', 'steps', 'name', 'hooks', 'exporters', 'start_delay_secs',
@@ -246,6 +249,7 @@ class EvalSpec(
throttle_secs=throttle_secs)
+@tf_export('estimator.train_and_evaluate')
def train_and_evaluate(estimator, train_spec, eval_spec):
"""Train and evaluate the `estimator`.
diff --git a/tensorflow/python/estimator/warm_starting_util.py b/tensorflow/python/estimator/warm_starting_util.py
index 57db968d56..adb013f5c6 100644
--- a/tensorflow/python/estimator/warm_starting_util.py
+++ b/tensorflow/python/estimator/warm_starting_util.py
@@ -30,8 +30,10 @@ from tensorflow.python.platform import tf_logging as logging
from tensorflow.python.training import checkpoint_ops
from tensorflow.python.training import checkpoint_utils
from tensorflow.python.training import saver
+from tensorflow.python.util.tf_export import tf_export
+@tf_export("estimator.VocabInfo")
class VocabInfo(
collections.namedtuple("VocabInfo", [
"new_vocab",
@@ -81,6 +83,7 @@ class VocabInfo(
)
+@tf_export("estimator.WarmStartSettings")
class WarmStartSettings(
collections.namedtuple("WarmStartSettings", [
"ckpt_to_initialize_from",
diff --git a/tensorflow/python/keras/_impl/keras/datasets/boston_housing.py b/tensorflow/python/keras/_impl/keras/datasets/boston_housing.py
index cfd7df61d5..13fa9aed2b 100644
--- a/tensorflow/python/keras/_impl/keras/datasets/boston_housing.py
+++ b/tensorflow/python/keras/_impl/keras/datasets/boston_housing.py
@@ -21,8 +21,10 @@ from __future__ import print_function
import numpy as np
from tensorflow.python.keras._impl.keras.utils.data_utils import get_file
+from tensorflow.python.util.tf_export import tf_export
+@tf_export('keras.datasets.boston_housing.load_data')
def load_data(path='boston_housing.npz', test_split=0.2, seed=113):
"""Loads the Boston Housing dataset.
diff --git a/tensorflow/python/keras/_impl/keras/datasets/cifar10.py b/tensorflow/python/keras/_impl/keras/datasets/cifar10.py
index fb9d98d42c..6b77243382 100644
--- a/tensorflow/python/keras/_impl/keras/datasets/cifar10.py
+++ b/tensorflow/python/keras/_impl/keras/datasets/cifar10.py
@@ -25,8 +25,10 @@ import numpy as np
from tensorflow.python.keras._impl.keras import backend as K
from tensorflow.python.keras._impl.keras.datasets.cifar import load_batch
from tensorflow.python.keras._impl.keras.utils.data_utils import get_file
+from tensorflow.python.util.tf_export import tf_export
+@tf_export('keras.datasets.cifar10.load_data')
def load_data():
"""Loads CIFAR10 dataset.
diff --git a/tensorflow/python/keras/_impl/keras/datasets/cifar100.py b/tensorflow/python/keras/_impl/keras/datasets/cifar100.py
index 95aace599a..28d74116a5 100644
--- a/tensorflow/python/keras/_impl/keras/datasets/cifar100.py
+++ b/tensorflow/python/keras/_impl/keras/datasets/cifar100.py
@@ -25,8 +25,10 @@ import numpy as np
from tensorflow.python.keras._impl.keras import backend as K
from tensorflow.python.keras._impl.keras.datasets.cifar import load_batch
from tensorflow.python.keras._impl.keras.utils.data_utils import get_file
+from tensorflow.python.util.tf_export import tf_export
+@tf_export('keras.datasets.cifar100.load_data')
def load_data(label_mode='fine'):
"""Loads CIFAR100 dataset.
diff --git a/tensorflow/python/keras/_impl/keras/datasets/imdb.py b/tensorflow/python/keras/_impl/keras/datasets/imdb.py
index 880c9c821b..e2dddf7730 100644
--- a/tensorflow/python/keras/_impl/keras/datasets/imdb.py
+++ b/tensorflow/python/keras/_impl/keras/datasets/imdb.py
@@ -25,8 +25,10 @@ import numpy as np
from tensorflow.python.keras._impl.keras.preprocessing.sequence import _remove_long_seq
from tensorflow.python.keras._impl.keras.utils.data_utils import get_file
from tensorflow.python.platform import tf_logging as logging
+from tensorflow.python.util.tf_export import tf_export
+@tf_export('keras.datasets.imdb.load_data')
def load_data(path='imdb.npz',
num_words=None,
skip_top=0,
@@ -128,6 +130,7 @@ def load_data(path='imdb.npz',
return (x_train, y_train), (x_test, y_test)
+@tf_export('keras.datasets.imdb.get_word_index')
def get_word_index(path='imdb_word_index.json'):
"""Retrieves the dictionary mapping word indices back to words.
diff --git a/tensorflow/python/keras/_impl/keras/datasets/mnist.py b/tensorflow/python/keras/_impl/keras/datasets/mnist.py
index ec12a31dcf..e30691373e 100644
--- a/tensorflow/python/keras/_impl/keras/datasets/mnist.py
+++ b/tensorflow/python/keras/_impl/keras/datasets/mnist.py
@@ -21,8 +21,10 @@ from __future__ import print_function
import numpy as np
from tensorflow.python.keras._impl.keras.utils.data_utils import get_file
+from tensorflow.python.util.tf_export import tf_export
+@tf_export('keras.datasets.mnist.load_data')
def load_data(path='mnist.npz'):
"""Loads the MNIST dataset.
diff --git a/tensorflow/python/keras/_impl/keras/datasets/reuters.py b/tensorflow/python/keras/_impl/keras/datasets/reuters.py
index 95cf8852a9..b711696b5e 100644
--- a/tensorflow/python/keras/_impl/keras/datasets/reuters.py
+++ b/tensorflow/python/keras/_impl/keras/datasets/reuters.py
@@ -25,8 +25,10 @@ import numpy as np
from tensorflow.python.keras._impl.keras.preprocessing.sequence import _remove_long_seq
from tensorflow.python.keras._impl.keras.utils.data_utils import get_file
from tensorflow.python.platform import tf_logging as logging
+from tensorflow.python.util.tf_export import tf_export
+@tf_export('keras.datasets.reuters.load_data')
def load_data(path='reuters.npz',
num_words=None,
skip_top=0,
@@ -112,6 +114,7 @@ def load_data(path='reuters.npz',
return (x_train, y_train), (x_test, y_test)
+@tf_export('keras.datasets.reuters.get_word_index')
def get_word_index(path='reuters_word_index.json'):
"""Retrieves the dictionary mapping word indices back to words.
diff --git a/tensorflow/python/layers/base.py b/tensorflow/python/layers/base.py
index 5dea732cba..3a3c559541 100644
--- a/tensorflow/python/layers/base.py
+++ b/tensorflow/python/layers/base.py
@@ -37,8 +37,10 @@ from tensorflow.python.ops import variable_scope as vs
from tensorflow.python.ops import variables as tf_variables
from tensorflow.python.platform import tf_logging as logging
from tensorflow.python.util import nest
+from tensorflow.python.util.tf_export import tf_export
+@tf_export('layers.Layer')
class Layer(object):
"""Base layer class.
@@ -1228,6 +1230,7 @@ class Layer(object):
', found shape=' + str(shape))
+@tf_export('keras.layers.InputSpec', 'layers.InputSpec')
class InputSpec(object):
"""Specifies the ndim, dtype and shape of every input to a layer.
diff --git a/tensorflow/python/layers/convolutional.py b/tensorflow/python/layers/convolutional.py
index e8dba3cea3..689046fe78 100644
--- a/tensorflow/python/layers/convolutional.py
+++ b/tensorflow/python/layers/convolutional.py
@@ -29,6 +29,7 @@ from tensorflow.python.ops import array_ops
from tensorflow.python.ops import init_ops
from tensorflow.python.ops import nn
from tensorflow.python.ops import nn_ops
+from tensorflow.python.util.tf_export import tf_export
class _Conv(base.Layer):
@@ -222,6 +223,7 @@ class _Conv(base.Layer):
new_space)
+@tf_export('layers.Conv1D')
class Conv1D(_Conv):
"""1D convolution layer (e.g. temporal convolution).
@@ -311,6 +313,7 @@ class Conv1D(_Conv):
name=name, **kwargs)
+@tf_export('layers.conv1d')
def conv1d(inputs,
filters,
kernel_size,
@@ -411,6 +414,7 @@ def conv1d(inputs,
return layer.apply(inputs)
+@tf_export('layers.Conv2D')
class Conv2D(_Conv):
"""2D convolution layer (e.g. spatial convolution over images).
@@ -507,6 +511,7 @@ class Conv2D(_Conv):
name=name, **kwargs)
+@tf_export('layers.conv2d')
def conv2d(inputs,
filters,
kernel_size,
@@ -614,6 +619,7 @@ def conv2d(inputs,
return layer.apply(inputs)
+@tf_export('layers.Conv3D')
class Conv3D(_Conv):
"""3D convolution layer (e.g. spatial convolution over volumes).
@@ -711,6 +717,7 @@ class Conv3D(_Conv):
name=name, **kwargs)
+@tf_export('layers.conv3d')
def conv3d(inputs,
filters,
kernel_size,
@@ -980,6 +987,7 @@ class _SeparableConv(_Conv):
raise NotImplementedError
+@tf_export('layers.SeparableConv1D')
class SeparableConv1D(_SeparableConv):
"""Depthwise separable 1D convolution.
@@ -1123,6 +1131,7 @@ class SeparableConv1D(_SeparableConv):
return outputs
+@tf_export('layers.SeparableConv2D')
class SeparableConv2D(_SeparableConv):
"""Depthwise separable 2D convolution.
@@ -1260,6 +1269,7 @@ class SeparableConv2D(_SeparableConv):
return outputs
+@tf_export('layers.separable_conv1d')
def separable_conv1d(inputs,
filters,
kernel_size,
@@ -1376,6 +1386,7 @@ def separable_conv1d(inputs,
return layer.apply(inputs)
+@tf_export('layers.separable_conv2d')
def separable_conv2d(inputs,
filters,
kernel_size,
@@ -1497,6 +1508,7 @@ def separable_conv2d(inputs,
return layer.apply(inputs)
+@tf_export('layers.Conv2DTranspose')
class Conv2DTranspose(Conv2D):
"""Transposed 2D convolution layer (sometimes called 2D Deconvolution).
@@ -1695,6 +1707,7 @@ class Conv2DTranspose(Conv2D):
return tensor_shape.TensorShape(output_shape)
+@tf_export('layers.conv2d_transpose')
def conv2d_transpose(inputs,
filters,
kernel_size,
@@ -1790,6 +1803,7 @@ def conv2d_transpose(inputs,
return layer.apply(inputs)
+@tf_export('layers.Conv3DTranspose')
class Conv3DTranspose(Conv3D):
"""Transposed 3D convolution layer (sometimes called 3D Deconvolution).
@@ -2018,6 +2032,7 @@ class Conv3DTranspose(Conv3D):
return tensor_shape.TensorShape(output_shape)
+@tf_export('layers.conv3d_transpose')
def conv3d_transpose(inputs,
filters,
kernel_size,
diff --git a/tensorflow/python/layers/core.py b/tensorflow/python/layers/core.py
index 7bf62d45b8..ec4fca78f0 100644
--- a/tensorflow/python/layers/core.py
+++ b/tensorflow/python/layers/core.py
@@ -37,8 +37,10 @@ from tensorflow.python.ops import init_ops
from tensorflow.python.ops import math_ops
from tensorflow.python.ops import nn
from tensorflow.python.ops import standard_ops
+from tensorflow.python.util.tf_export import tf_export
+@tf_export('layers.Dense')
class Dense(base.Layer):
"""Densely-connected layer class.
@@ -173,6 +175,7 @@ class Dense(base.Layer):
return input_shape[:-1].concatenate(self.units)
+@tf_export('layers.dense')
def dense(
inputs, units,
activation=None,
@@ -248,6 +251,7 @@ def dense(
return layer.apply(inputs)
+@tf_export('layers.Dropout')
class Dropout(base.Layer):
"""Applies Dropout to the input.
@@ -309,6 +313,7 @@ class Dropout(base.Layer):
return input_shape
+@tf_export('layers.dropout')
def dropout(inputs,
rate=0.5,
noise_shape=None,
@@ -350,6 +355,7 @@ def dropout(inputs,
return layer.apply(inputs, training=training)
+@tf_export('layers.Flatten')
class Flatten(base.Layer):
"""Flattens an input tensor while preserving the batch axis (axis 0).
@@ -386,6 +392,7 @@ class Flatten(base.Layer):
return tensor_shape.TensorShape(output_shape)
+@tf_export('layers.flatten')
def flatten(inputs, name=None):
"""Flattens an input tensor while preserving the batch axis (axis 0).
diff --git a/tensorflow/python/layers/network.py b/tensorflow/python/layers/network.py
index 7bcf25064c..6de8f35502 100644
--- a/tensorflow/python/layers/network.py
+++ b/tensorflow/python/layers/network.py
@@ -30,6 +30,7 @@ from tensorflow.python.ops import array_ops
from tensorflow.python.ops import variable_scope as vs
from tensorflow.python.platform import tf_logging as logging
from tensorflow.python.util import nest
+from tensorflow.python.util.tf_export import tf_export
class InputLayer(base.Layer):
@@ -117,6 +118,7 @@ class InputLayer(base.Layer):
output_tensors=[input_tensor])
+@tf_export('layers.Input')
def Input( # pylint: disable=invalid-name
shape=None,
batch_size=None,
diff --git a/tensorflow/python/layers/normalization.py b/tensorflow/python/layers/normalization.py
index 890c12f6e0..656d566ab5 100644
--- a/tensorflow/python/layers/normalization.py
+++ b/tensorflow/python/layers/normalization.py
@@ -39,8 +39,10 @@ from tensorflow.python.ops import math_ops
from tensorflow.python.ops import init_ops
from tensorflow.python.ops import state_ops
from tensorflow.python.training import moving_averages
+from tensorflow.python.util.tf_export import tf_export
+@tf_export('layers.BatchNormalization')
class BatchNormalization(base.Layer):
"""Batch Normalization layer from http://arxiv.org/abs/1502.03167.
@@ -629,6 +631,7 @@ class BatchNormalization(base.Layer):
return input_shape
+@tf_export('layers.batch_normalization')
def batch_normalization(inputs,
axis=-1,
momentum=0.99,
diff --git a/tensorflow/python/layers/pooling.py b/tensorflow/python/layers/pooling.py
index ab06a3a408..50503ce093 100644
--- a/tensorflow/python/layers/pooling.py
+++ b/tensorflow/python/layers/pooling.py
@@ -26,6 +26,7 @@ from tensorflow.python.layers import base
from tensorflow.python.layers import utils
from tensorflow.python.ops import array_ops
from tensorflow.python.ops import nn
+from tensorflow.python.util.tf_export import tf_export
class _Pooling1D(base.Layer):
@@ -96,6 +97,7 @@ class _Pooling1D(base.Layer):
return tensor_shape.TensorShape([input_shape[0], length, input_shape[2]])
+@tf_export('layers.AveragePooling1D')
class AveragePooling1D(_Pooling1D):
"""Average Pooling layer for 1D inputs.
@@ -127,6 +129,7 @@ class AveragePooling1D(_Pooling1D):
**kwargs)
+@tf_export('layers.average_pooling1d')
def average_pooling1d(inputs, pool_size, strides,
padding='valid', data_format='channels_last',
name=None):
@@ -161,6 +164,7 @@ def average_pooling1d(inputs, pool_size, strides,
return layer.apply(inputs)
+@tf_export('layers.MaxPooling1D')
class MaxPooling1D(_Pooling1D):
"""Max Pooling layer for 1D inputs.
@@ -192,6 +196,7 @@ class MaxPooling1D(_Pooling1D):
**kwargs)
+@tf_export('layers.max_pooling1d')
def max_pooling1d(inputs, pool_size, strides,
padding='valid', data_format='channels_last',
name=None):
@@ -297,6 +302,7 @@ class _Pooling2D(base.Layer):
[input_shape[0], rows, cols, input_shape[3]])
+@tf_export('layers.AveragePooling2D')
class AveragePooling2D(_Pooling2D):
"""Average pooling layer for 2D inputs (e.g. images).
@@ -328,6 +334,7 @@ class AveragePooling2D(_Pooling2D):
padding=padding, data_format=data_format, name=name, **kwargs)
+@tf_export('layers.average_pooling2d')
def average_pooling2d(inputs,
pool_size, strides,
padding='valid', data_format='channels_last',
@@ -365,6 +372,7 @@ def average_pooling2d(inputs,
return layer.apply(inputs)
+@tf_export('layers.MaxPooling2D')
class MaxPooling2D(_Pooling2D):
"""Max pooling layer for 2D inputs (e.g. images).
@@ -396,6 +404,7 @@ class MaxPooling2D(_Pooling2D):
padding=padding, data_format=data_format, name=name, **kwargs)
+@tf_export('layers.max_pooling2d')
def max_pooling2d(inputs,
pool_size, strides,
padding='valid', data_format='channels_last',
@@ -515,6 +524,7 @@ class _Pooling3D(base.Layer):
[input_shape[0], len_dim1, len_dim2, len_dim3, input_shape[4]])
+@tf_export('layers.AveragePooling3D')
class AveragePooling3D(_Pooling3D):
"""Average pooling layer for 3D inputs (e.g. volumes).
@@ -548,6 +558,7 @@ class AveragePooling3D(_Pooling3D):
padding=padding, data_format=data_format, name=name, **kwargs)
+@tf_export('layers.average_pooling3d')
def average_pooling3d(inputs,
pool_size, strides,
padding='valid', data_format='channels_last',
@@ -587,6 +598,7 @@ def average_pooling3d(inputs,
return layer.apply(inputs)
+@tf_export('layers.MaxPooling3D')
class MaxPooling3D(_Pooling3D):
"""Max pooling layer for 3D inputs (e.g. volumes).
@@ -620,6 +632,7 @@ class MaxPooling3D(_Pooling3D):
padding=padding, data_format=data_format, name=name, **kwargs)
+@tf_export('layers.max_pooling3d')
def max_pooling3d(inputs,
pool_size, strides,
padding='valid', data_format='channels_last',