Commit message (Collapse) | Author | Age | |
---|---|---|---|
* | Made it possible to limit the number of blocks that will be used to evaluate ↵ | 2016-02-01 | |
| | | | | a tensor expression on a CUDA device. This makesit possible to set aside streaming multiprocessors for other computations. | ||
* | Silenced a few compilation warnings. | 2016-01-11 | |
| | |||
* | Silenced several compilation warnings triggered by nvcc. | 2016-01-11 | |
| | |||
* | Cleaned up double-defined macro from last commit | 2016-01-10 | |
| | |||
* | Alternative way of forcing instantiation of device kernels without | 2016-01-10 | |
| | | | | | | causing warnings or requiring device to device kernel invocations. This allows Tensorflow to work on SM 3.0 (ie, Amazon EC2) machines. | ||
* | Prevent nvcc from miscompiling the cuda metakernel. Unfortunately this ↵ | 2016-01-08 | |
| | | | | reintroduces some compulation warnings but it's much better than having to deal with random assertion failures. | ||
* | Silenced some compilation warnings triggered by nvcc | 2015-12-17 | |
| | |||
* | Made it possible to refer t oa GPUDevice from code compile with a regular ↵ | 2015-11-23 | |
| | | | | C++ compiler | ||
* | Split TensorDeviceType.h in 3 files to make it more manageable | 2015-11-20 | |