diff options
author | Manjunath Kudlur <keveman@gmail.com> | 2015-11-06 16:27:58 -0800 |
---|---|---|
committer | Manjunath Kudlur <keveman@gmail.com> | 2015-11-06 16:27:58 -0800 |
commit | f41959ccb2d9d4c722fe8fc3351401d53bcf4900 (patch) | |
tree | ef0ca22cb2a5ac4bdec9d080d8e0788a53ed496d /tensorflow/core/kernels/softplus_op.h |
TensorFlow: Initial commit of TensorFlow library.
TensorFlow is an open source software library for numerical computation
using data flow graphs.
Base CL: 107276108
Diffstat (limited to 'tensorflow/core/kernels/softplus_op.h')
-rw-r--r-- | tensorflow/core/kernels/softplus_op.h | 46 |
1 files changed, 46 insertions, 0 deletions
diff --git a/tensorflow/core/kernels/softplus_op.h b/tensorflow/core/kernels/softplus_op.h new file mode 100644 index 0000000000..3545a78246 --- /dev/null +++ b/tensorflow/core/kernels/softplus_op.h @@ -0,0 +1,46 @@ +#ifndef TENSORFLOW_KERNELS_SOFTPLUS_OP_H_ +#define TENSORFLOW_KERNELS_SOFTPLUS_OP_H_ +// Functor definition for SoftplusOp and SoftplusGradOp, must be compilable by +// nvcc. + +#include "tensorflow/core/framework/tensor_types.h" +#include "third_party/eigen3/unsupported/Eigen/CXX11/Tensor" + +namespace tensorflow { +namespace functor { + +// Functor used by SoftplusOp to do the computations. +template <typename Device, typename T> +struct Softplus { + // Computes Softplus activation. + // + // features: any shape. + // activations: same shape as "features". + void operator()(const Device& d, typename TTypes<T>::ConstTensor features, + typename TTypes<T>::Tensor activations) { + activations.device(d) = + (features > features.constant(30.f)) + .select(features, (features.exp() + features.constant(1.0f)).log()); + } +}; + +// Functor used by SoftplusGradOp to do the computations. +template <typename Device, typename T> +struct SoftplusGrad { + // Computes SoftplusGrad backprops. + // + // gradients: gradients backpropagated to the Softplus op. + // features: inputs that where passed to the Softplus op. + // backprops: gradients to backpropagate to the Softplus inputs. + void operator()(const Device& d, typename TTypes<T>::ConstTensor gradients, + typename TTypes<T>::ConstTensor features, + typename TTypes<T>::Tensor backprops) { + backprops.device(d) = + gradients / ((-features).exp() + features.constant(1.0f)); + } +}; + +} // namespace functor +} // namespace tensorflow + +#endif // TENSORFLOW_KERNELS_SOFTPLUS_OP_H_ |