aboutsummaryrefslogtreecommitdiffhomepage
path: root/unsupported/Eigen/CXX11
Commit message (Collapse)AuthorAge
* Fixed the partial evaluation of non vectorizable tensor subexpressionsGravatar Benoit Steiner2016-04-25
|
* Refined the cost of the striding operation.Gravatar Benoit Steiner2016-04-25
|
* Provide access to the base threadpool classesGravatar Benoit Steiner2016-04-21
|
* Added the ability to switch to the new thread pool with a #defineGravatar Benoit Steiner2016-04-21
|
* Fixed several compilation warningsGravatar Benoit Steiner2016-04-21
|
* Don't crash when attempting to reduce empty tensors.Gravatar Benoit Steiner2016-04-20
|
* Started to implement a portable way to yield.Gravatar Benoit Steiner2016-04-19
|
* Implemented a more portable version of thread local variablesGravatar Benoit Steiner2016-04-19
|
* Fixed a compilation error with nvcc 7.Gravatar Benoit Steiner2016-04-19
|
* Simplified the code that launches cuda kernels.Gravatar Benoit Steiner2016-04-19
|
* Don't take the address of a kernel on CUDA devices that don't support this ↵Gravatar Benoit Steiner2016-04-19
| | | | feature.
* Use numext::ceil instead of std::ceilGravatar Benoit Steiner2016-04-19
|
* Avoid an unnecessary copy of the evaluator.Gravatar Benoit Steiner2016-04-19
|
* Use DenseIndex in the MeanReducer to avoid overflows when processing very ↵Gravatar Benoit Steiner2016-04-19
| | | | large tensors.
* Move the evalGemm method into the TensorContractionEvaluatorBase class to ↵Gravatar Benoit Steiner2016-04-15
| | | | make it accessible from both the single and multithreaded contraction evaluators.
* Deleted unnecessary variableGravatar Benoit Steiner2016-04-15
|
* Fixed a few compilation warningsGravatar Benoit Steiner2016-04-15
|
* Merged in rmlarsen/eigen (pull request PR-178)Gravatar Benoit Steiner2016-04-15
|\ | | | | | | Eigen Tensor cost model part 2: Thread scheduling for standard evaluators and reductions.
| * Get rid of void* casting when calling EvalRange::run.Gravatar Rasmus Munk Larsen2016-04-15
| |
* | Added ability to access the cache sizes from the tensor devicesGravatar Benoit Steiner2016-04-14
| |
* | Added support for exclusive orGravatar Benoit Steiner2016-04-14
| |
| * Eigen Tensor cost model part 2: Thread scheduling for standard evaluators ↵Gravatar Rasmus Munk Larsen2016-04-14
| | | | | | | | and reductions. The cost model is turned off by default.
* | Added missing definition of PacketSize in the gpu evaluator of convolutionGravatar Benoit Steiner2016-04-14
| |
* | Merged in rmlarsen/eigen (pull request PR-177)Gravatar Benoit Steiner2016-04-14
|\| | | | | | | Eigen Tensor cost model part 1.
* | CleanupGravatar Benoit Steiner2016-04-14
| |
* | Prepared the migration to the new non blocking thread poolGravatar Benoit Steiner2016-04-14
| |
| * Improvements to cost model.Gravatar Rasmus Munk Larsen2016-04-14
| |
* | Added a more scalable non blocking thread poolGravatar Benoit Steiner2016-04-14
| |
| * Merge upstream updates.Gravatar Rasmus Munk Larsen2016-04-14
| |\ | |/ |/|
| * Eigen cost model part 1. This implements a basic recursive framework to ↵Gravatar Rasmus Munk Larsen2016-04-14
| | | | | | | | estimate the cost of evaluating tensor expressions.
* | Silenced a compilation warningGravatar Benoit Steiner2016-04-14
| |
* | Added support for fp16 to the sigmoid functionGravatar Benoit Steiner2016-04-14
|/
* Defer the decision to vectorize tensor CUDA code to the meta kernel. This ↵Gravatar Benoit Steiner2016-04-12
| | | | makes it possible to decide to vectorize or not depending on the capability of the target cuda architecture. In particular, this enables us to vectorize the processing of fp16 when running on device of capability >= 5.3
* Added missing EIGEN_DEVICE_FUNC to the tensor conversion code.Gravatar Benoit Steiner2016-04-07
|
* Added support for isinf, isnan, and isfinite checks to the tensor apiGravatar Benoit Steiner2016-04-07
|
* Fixed typos in the implementation of the zeta and polygamma ops.Gravatar Benoit Steiner2016-04-06
|
* Merge upstream.Gravatar Till Hoffmann2016-04-01
|\
* | Fixed CUDA signature.Gravatar Till Hoffmann2016-04-01
| |
| * Merged eigen/eigen into defaultGravatar tillahoffmann2016-04-01
|/|
* | Added polygamma function.Gravatar Till Hoffmann2016-04-01
| |
* | Added zeta function.Gravatar Till Hoffmann2016-04-01
| |
| * Relaxed the condition used to gate the fft code.Gravatar Benoit Steiner2016-03-31
| |
| * Properly gate the fft codeGravatar Benoit Steiner2016-03-31
|/
* Fixed a off-by-one bug in a debug assertionGravatar Benoit Steiner2016-03-30
|
* Added NumTraits for type2index.Gravatar Benoit Steiner2016-03-30
|
* Fixed compilation warningGravatar Benoit Steiner2016-03-30
|
* Added missing assignment operator to the TensorUInt128 class, and made misc ↵Gravatar Benoit Steiner2016-03-30
| | | | small improvements
* Fixed the formatting of the README.Gravatar Benoit Steiner2016-03-29
|
* Attempt to fix the formatting of the READMEGravatar Benoit Steiner2016-03-29
|
* Added support for fmodGravatar Benoit Steiner2016-03-28
|