| Commit message (Collapse) | Author | Age |
... | |
|
|
|
| |
Started to improve multithreaded contractions
|
| |
|
| |
|
| |
|
| |
|
| |
|
| |
|
|
|
|
| |
number on gpu.
|
|
|
|
| |
Added a functor to encapsulate the generation of random numbers on cpu and gpu.
|
| |
|
| |
|
| |
|
|
|
|
| |
Improved contractions on GPU
|
|
|
|
|
|
| |
support std::pair so we can't use them when targeting GPUs.
Improved the performance on tensor contractions
|
|
|
|
| |
More tests
|
| |
|
| |
|
| |
|
| |
|
| |
|
|
|
|
|
|
|
| |
declare and initialize a tensor on the same line, as in:
Tensor<bla> T = A + B; or
Tensor<bla> T(A.reshape(new_shape));
|
| |
|
| |
|
| |
|
| |
|
| |
|
| |
|
|
|
|
| |
Sped up tensor slicing by a factor of 3 by using these fast integer divisions.
|
| |
|
|\ |
|
| | |
|
| | |
|
| | |
|
| |
| |
| |
| | |
EvalSubExprsIfNeeded apu.
|
| | |
|
| |
| |
| |
| | |
reshaping operations
|
| |
| |
| |
| | |
Use mempy to speedup tensor copies whenever possible.
|
| | |
|
| | |
|
| | |
|
| | |
|
| | |
|
| | |
|
| | |
|
| | |
|
| | |
|
| | |
|
| | |
|
| | |
|
| | |
|