diff options
Diffstat (limited to 'tensorflow/core/kernels/cwise_ops_test.cc')
-rw-r--r-- | tensorflow/core/kernels/cwise_ops_test.cc | 51 |
1 files changed, 51 insertions, 0 deletions
diff --git a/tensorflow/core/kernels/cwise_ops_test.cc b/tensorflow/core/kernels/cwise_ops_test.cc index 6250928aca..92018ec871 100644 --- a/tensorflow/core/kernels/cwise_ops_test.cc +++ b/tensorflow/core/kernels/cwise_ops_test.cc @@ -51,18 +51,38 @@ static int ColsFromArg(int arg) { return (arg % kRows); } BENCHMARK(BM_##DEVICE##_##FUNC##_##TYPE)->Range(4 << 10, 1 << 20); BM_UNARY(cpu, Floor, float, DT_FLOAT); +#if GOOGLE_CUDA BM_UNARY(gpu, Floor, float, DT_FLOAT); +#endif // GOOGLE_CUDA +#ifdef TENSORFLOW_USE_SYCL +BM_UNARY(sycl, Floor, float, DT_FLOAT); +#endif // TENSORFLOW_USE_SYCL + BM_UNARY(cpu, Floor, double, DT_DOUBLE); +#if GOOGLE_CUDA BM_UNARY(gpu, Floor, double, DT_DOUBLE); +#endif // GOOGLE_CUDA +#ifdef TENSORFLOW_USE_SYCL +BM_UNARY(sycl, Floor, double, DT_DOUBLE); +#endif // TENSORFLOW_USE_SYCL + BM_UNARY(cpu, Conj, std::complex<float>, DT_COMPLEX64); +#if GOOGLE_CUDA BM_UNARY(gpu, Conj, std::complex<float>, DT_COMPLEX64); +#endif // GOOGLE_CUDA BM_UNARY(cpu, Conj, std::complex<double>, DT_COMPLEX128); +#if GOOGLE_CUDA BM_UNARY(gpu, Conj, std::complex<double>, DT_COMPLEX128); +#endif // GOOGLE_CUDA BM_UNARY(cpu, Rint, double, DT_DOUBLE); +#if GOOGLE_CUDA BM_UNARY(gpu, Rint, double, DT_DOUBLE); +#endif // GOOGLE_CUDA BM_UNARY(cpu, Rint, float, DT_FLOAT); +#if GOOGLE_CUDA BM_UNARY(gpu, Rint, float, DT_FLOAT); +#endif // GOOGLE_CUDA // data func scalar. static Graph* BinaryScalar(int num, const string& func) { @@ -90,9 +110,20 @@ static Graph* BinaryScalar(int num, const string& func) { ->Arg(1048576); BM_BINARY_SCALAR(cpu, Less); +#if GOOGLE_CUDA BM_BINARY_SCALAR(gpu, Less); +#endif // GOOGLE_CUDA +#ifdef TENSORFLOW_USE_SYCL +BM_BINARY_SCALAR(sycl, Less); +#endif // TENSORFLOW_USE_SYCL + BM_BINARY_SCALAR(cpu, Add); +#if GOOGLE_CUDA BM_BINARY_SCALAR(gpu, Add); +#endif // GOOGLE_CUDA +#ifdef TENSORFLOW_USE_SYCL +BM_BINARY_SCALAR(sycl, Add); +#endif // TENSORFLOW_USE_SYCL #undef BM_BINARY_SCALAR template <class T> @@ -130,9 +161,13 @@ static Graph* BiasAdd(int rows, int cols, DataType type) { using Eigen::half; BM_BIAS_ADD_ALL(cpu, float, DT_FLOAT); +#if GOOGLE_CUDA BM_BIAS_ADD_ALL(gpu, float, DT_FLOAT); +#endif // GOOGLE_CUDA BM_BIAS_ADD_ALL(cpu, half, DT_HALF); +#if GOOGLE_CUDA BM_BIAS_ADD_ALL(gpu, half, DT_HALF); +#endif // GOOGLE_CUDA #undef BM_BIAS_ADD_ALL #undef BM_BIAS_ADD @@ -180,12 +215,18 @@ static Graph* BiasAddGrad(int rows, int cols, int channels, DataType type, BM_BIAS_ADD_GRAD(DEVICE, FORMAT, C_TYPE, TF_TYPE, 4096, 4096, 1); using Eigen::half; +#if GOOGLE_CUDA BM_BIAS_ADD_GRAD_ALL(gpu, NCHW, float, DT_FLOAT); BM_BIAS_ADD_GRAD_ALL(gpu, NCHW, half, DT_HALF); +#endif // GOOGLE_CUDA BM_BIAS_ADD_GRAD_ALL(cpu, NHWC, float, DT_FLOAT); +#if GOOGLE_CUDA BM_BIAS_ADD_GRAD_ALL(gpu, NHWC, float, DT_FLOAT); +#endif // GOOGLE_CUDA BM_BIAS_ADD_GRAD_ALL(cpu, NHWC, half, DT_HALF); +#if GOOGLE_CUDA BM_BIAS_ADD_GRAD_ALL(gpu, NHWC, half, DT_HALF); +#endif // GOOGLE_CUDA #undef BM_BIAS_ADD_GRAD_ALL #undef BM_BIAS_ADD_GRAD @@ -223,7 +264,12 @@ static Graph* BcastAdd(int rows, int cols, int dim) { BM_BCAST_ADD_ROW(DEVICE, 2048, 512); \ BM_BCAST_ADD_ROW(DEVICE, 4096, 512); BM_BCAST_ADD_ROW_ALL(cpu); +#if GOOGLE_CUDA BM_BCAST_ADD_ROW_ALL(gpu); +#endif // GOOGLE_CUDA +#ifdef TENSORFLOW_USE_SYCL +BM_BCAST_ADD_ROW_ALL(sycl); +#endif // TENSORFLOW_USE_SYCL #undef BM_BCAST_ADD_ROW_ALL #undef BM_BCAST_ADD_ROW @@ -244,7 +290,12 @@ BM_BCAST_ADD_ROW_ALL(gpu); BM_BCAST_ADD_COL(DEVICE, 2048, 512); \ BM_BCAST_ADD_COL(DEVICE, 4096, 512); BM_BCAST_ADD_COL_ALL(cpu); +#if GOOGLE_CUDA BM_BCAST_ADD_COL_ALL(gpu); +#endif // GOOGLE_CUDA +#ifdef TENSORFLOW_USE_SYCL +BM_BCAST_ADD_COL_ALL(sycl); +#endif // TENSORFLOW_USE_SYCL #undef BM_BCAST_ADD_COL_ALL #undef BM_BCAST_ADD_COL |