From e892524efe7e8adbd43bf4c1c150f4f4ebf27d1d Mon Sep 17 00:00:00 2001 From: Benoit Steiner Date: Wed, 15 Jul 2015 12:38:34 -0700 Subject: Added support for multi gpu configuration to the GpuDevice class --- unsupported/Eigen/CXX11/src/Tensor/TensorConvolution.h | 8 ++++---- 1 file changed, 4 insertions(+), 4 deletions(-) (limited to 'unsupported/Eigen/CXX11/src/Tensor/TensorConvolution.h') diff --git a/unsupported/Eigen/CXX11/src/Tensor/TensorConvolution.h b/unsupported/Eigen/CXX11/src/Tensor/TensorConvolution.h index 88db9d410..64b0315ee 100644 --- a/unsupported/Eigen/CXX11/src/Tensor/TensorConvolution.h +++ b/unsupported/Eigen/CXX11/src/Tensor/TensorConvolution.h @@ -835,10 +835,10 @@ struct TensorEvaluator::Dimensions InputDims; - const int maxSharedMem = sharedMemPerBlock(); - const int maxThreadsPerBlock = maxCudaThreadsPerBlock(); - const int maxBlocksPerProcessor = maxCudaThreadsPerMultiProcessor() / maxThreadsPerBlock; - const int numMultiProcessors = getNumCudaMultiProcessors(); + const int maxSharedMem = m_device.sharedMemPerBlock(); + const int maxThreadsPerBlock = m_device.maxCudaThreadsPerBlock(); + const int maxBlocksPerProcessor = m_device.maxCudaThreadsPerMultiProcessor() / maxThreadsPerBlock; + const int numMultiProcessors = m_device.getNumCudaMultiProcessors(); const int warpSize = 32; switch (NumKernelDims) { -- cgit v1.2.3