aboutsummaryrefslogtreecommitdiffhomepage
path: root/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
diff options
context:
space:
mode:
authorGravatar Rasmus Munk Larsen <rmlarsen@google.com>2019-08-07 12:12:52 -0700
committerGravatar Rasmus Munk Larsen <rmlarsen@google.com>2019-08-07 12:12:52 -0700
commit09871261653b4a373b2aed1561c38a7f5d21a21e (patch)
tree39649af61bc3c41dda09c3142846281d8d0ba2e2 /unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
parent0050644b23b3d211090f26c6e52c99068c51d651 (diff)
Clean up unnecessary namespace specifiers in TensorBlock.h.
Diffstat (limited to 'unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h')
-rw-r--r--unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h40
1 files changed, 20 insertions, 20 deletions
diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h b/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
index 38c06aba2..f8942d527 100644
--- a/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
+++ b/unsupported/Eigen/CXX11/src/Tensor/TensorBlock.h
@@ -147,8 +147,8 @@ struct TensorBlockCopyOp {
typedef typename packet_traits<Scalar>::type Packet;
enum {
- Vectorizable = internal::packet_traits<Scalar>::Vectorizable,
- PacketSize = internal::packet_traits<Scalar>::size
+ Vectorizable = packet_traits<Scalar>::Vectorizable,
+ PacketSize = packet_traits<Scalar>::size
};
static EIGEN_DEVICE_FUNC EIGEN_STRONG_INLINE void Run(
@@ -171,8 +171,8 @@ struct TensorBlockCopyOp {
if (dst_stride == 1) {
// LINEAR
for (StorageIndex i = 0; i < vectorized_size; i += PacketSize) {
- Packet p = internal::ploadu<Packet>(src + i);
- internal::pstoreu<Scalar, Packet>(dst + i, p);
+ Packet p = ploadu<Packet>(src + i);
+ pstoreu<Scalar, Packet>(dst + i, p);
}
for (StorageIndex i = vectorized_size; i < num_coeff_to_copy; ++i) {
dst[i] = src[i];
@@ -180,8 +180,8 @@ struct TensorBlockCopyOp {
} else {
// SCATTER
for (StorageIndex i = 0; i < vectorized_size; i += PacketSize) {
- Packet p = internal::ploadu<Packet>(src + i);
- internal::pscatter<Scalar, Packet>(dst + i * dst_stride, p, dst_stride);
+ Packet p = ploadu<Packet>(src + i);
+ pscatter<Scalar, Packet>(dst + i * dst_stride, p, dst_stride);
}
for (StorageIndex i = vectorized_size; i < num_coeff_to_copy; ++i) {
dst[i * dst_stride] = src[i];
@@ -192,8 +192,8 @@ struct TensorBlockCopyOp {
if (dst_stride == 1) {
// LINEAR
for (StorageIndex i = 0; i < vectorized_size; i += PacketSize) {
- Packet p = internal::pload1<Packet>(src);
- internal::pstoreu<Scalar, Packet>(dst + i, p);
+ Packet p = pload1<Packet>(src);
+ pstoreu<Scalar, Packet>(dst + i, p);
}
for (StorageIndex i = vectorized_size; i < num_coeff_to_copy; ++i) {
dst[i] = *src;
@@ -201,8 +201,8 @@ struct TensorBlockCopyOp {
} else {
// SCATTER
for (StorageIndex i = 0; i < vectorized_size; i += PacketSize) {
- Packet p = internal::pload1<Packet>(src);
- internal::pscatter<Scalar, Packet>(dst + i * dst_stride, p, dst_stride);
+ Packet p = pload1<Packet>(src);
+ pscatter<Scalar, Packet>(dst + i * dst_stride, p, dst_stride);
}
for (StorageIndex i = vectorized_size; i < num_coeff_to_copy; ++i) {
dst[i * dst_stride] = *src;
@@ -213,8 +213,8 @@ struct TensorBlockCopyOp {
// GATHER
const StorageIndex vectorized_size = (num_coeff_to_copy / PacketSize) * PacketSize;
for (StorageIndex i = 0; i < vectorized_size; i += PacketSize) {
- Packet p = internal::pgather<Scalar, Packet>(src + i * src_stride, src_stride);
- internal::pstoreu<Scalar, Packet>(dst + i, p);
+ Packet p = pgather<Scalar, Packet>(src + i * src_stride, src_stride);
+ pstoreu<Scalar, Packet>(dst + i, p);
}
for (StorageIndex i = vectorized_size; i < num_coeff_to_copy; ++i) {
dst[i] = src[i * src_stride];
@@ -491,11 +491,11 @@ struct TensorBlockCwiseUnaryOp {
const StorageIndex output_index, const StorageIndex output_stride,
OutputScalar* output_data, const StorageIndex input_index,
const StorageIndex input_stride, const InputScalar* input_data) {
- typedef const Eigen::Array<InputScalar, Dynamic, 1> Input;
- typedef Eigen::Array<OutputScalar, Dynamic, 1> Output;
+ typedef const Array<InputScalar, Dynamic, 1> Input;
+ typedef Array<OutputScalar, Dynamic, 1> Output;
- typedef Eigen::Map<Input, 0, InnerStride<> > InputMap;
- typedef Eigen::Map<Output, 0, InnerStride<> > OutputMap;
+ typedef Map<Input, 0, InnerStride<> > InputMap;
+ typedef Map<Output, 0, InnerStride<> > OutputMap;
const InputScalar* input_base = &input_data[input_index];
OutputScalar* output_base = &output_data[output_index];
@@ -503,7 +503,7 @@ struct TensorBlockCwiseUnaryOp {
const InputMap input(input_base, num_coeff, InnerStride<>(input_stride));
OutputMap output(output_base, num_coeff, InnerStride<>(output_stride));
- output = Eigen::CwiseUnaryOp<UnaryFunctor, InputMap>(input, functor);
+ output = CwiseUnaryOp<UnaryFunctor, InputMap>(input, functor);
}
};
@@ -518,7 +518,7 @@ struct TensorBlockCwiseUnaryOp {
template <typename UnaryFunctor, typename StorageIndex, typename OutputScalar,
int NumDims, int Layout>
struct TensorBlockCwiseUnaryIO {
- typedef typename internal::TensorBlock<OutputScalar, StorageIndex, NumDims,
+ typedef typename TensorBlock<OutputScalar, StorageIndex, NumDims,
Layout>::Dimensions Dimensions;
struct BlockIteratorState {
@@ -865,7 +865,7 @@ class TensorBlockMapper {
const TensorBlockShapeType block_shape,
Index min_target_size)
: m_dimensions(dims),
- m_block_dim_sizes(BlockDimensions(dims, block_shape, internal::convert_index<StorageIndex>(min_target_size))) {
+ m_block_dim_sizes(BlockDimensions(dims, block_shape, convert_index<StorageIndex>(min_target_size))) {
// Calculate block counts by dimension and total block count.
DSizes<StorageIndex, NumDims> block_count;
for (Index i = 0; i < block_count.rank(); ++i) {
@@ -974,7 +974,7 @@ class TensorBlockMapper {
if (block_shape == kUniformAllDims) {
// Tensor will not fit within 'min_target_size' budget: calculate tensor
// block dimension sizes based on "square" dimension size target.
- const StorageIndex dim_size_target = internal::convert_index<StorageIndex>(
+ const StorageIndex dim_size_target = convert_index<StorageIndex>(
std::pow(static_cast<float>(min_target_size),
1.0f / static_cast<float>(block_dim_sizes.rank())));
for (Index i = 0; i < block_dim_sizes.rank(); ++i) {