aboutsummaryrefslogtreecommitdiffhomepage
path: root/third_party
Commit message (Collapse)AuthorAge
...
* Switched to the latest version of Eigen that provides significant performanceGravatar Benoit Steiner2016-07-12
| | | | | | improvements for fp16 Added SpecialFunctions to the list of eigen headers TensorFlow depends on Change: 127264575
* Automated rollback of change 127233960Gravatar Vijay Vasudevan2016-07-12
| | | | Change: 127253427
* Switched to the latest version of Eigen that provides significant performanceGravatar Benoit Steiner2016-07-12
| | | | | improvements for fp16 Change: 127233960
* Merge changes from github.Gravatar A. Unique TensorFlower2016-06-30
| | | | Change: 126335170
* Adds a "currentThreadIndex" method to Eigen's ThreadPoolDevice. Use it to ↵Gravatar A. Unique TensorFlower2016-06-27
| | | | | | handle per-thread buffer allocation for the tileable executor without resorting to thread_local that is not fully supported on Android. Change: 126009029
* Upgraded Eigen to the latest version that provides new scan operations. This ↵Gravatar Benoit Steiner2016-06-23
| | | | | | will enable the implementation of the cumsum operation in TensorFlow Change: 125697517
* Enable the vectorization of adds and mult on fp16s. This improves theGravatar Benoit Steiner2016-06-08
| | | | | performance of the toy mnist training by 1 order of magnitude Change: 124374286
* Improved the performance of full reductions on GPU.Gravatar Benoit Steiner2016-06-07
| | | | | | | | | | | | | | | | | NEW BM_fullReduction/10 4591 4595 153149 20.8M items/s BM_fullReduction/64 5073 5075 100000 770.0M items/s BM_fullReduction/512 9067 9070 75263 26.9G items/s BM_fullReduction/4k 243984 244125 2868 64.0G items/s BM_fullReduction/5k 359125 359273 1951 64.8G items/s OLD BM_fullReduction/10 9085 9087 74395 10.5M items/s BM_fullReduction/64 9478 9478 72014 412.1M items/s BM_fullReduction/512 14643 14646 46902 16.7G items/s BM_fullReduction/4k 260338 260384 2678 60.0G items/s BM_fullReduction/5k 385076 385178 1818 60.5G items/s Change: 124290852
* Enable fp16 for most of the pooling ops (MaxPool, AvgPool, associatedGravatar Benoit Steiner2016-06-06
| | | | | gradients, some variants etc.). Change: 124197406
* Merge changes from github.Gravatar Martin Wicke2016-06-06
| | | | Change: 124183870
* Enable fp16 for most of the pooling ops (MaxPool, AvgPool, associatedGravatar A. Unique TensorFlower2016-06-03
| | | | | gradients, some variants etc.). Change: 123967787
* Enable fp16 for most of the pooling ops (MaxPool, AvgPool, associatedGravatar Benoit Steiner2016-06-03
| | | | | gradients, some variants etc.). Change: 123967117
* Update copyright for 3p/tf.Gravatar A. Unique TensorFlower2016-06-02
| | | | Change: 123901292
* Added support for convolutions of 16bit floats on CPUGravatar Benoit Steiner2016-05-31
| | | | Change: 123659102
* Fix Tensorflow compilation error that comes up when TF is used as a ↵Gravatar Kiril Gorovoy2016-05-27
| | | | | | submodule and compiled for GPU. Change: 123468144
* Add a new contrib/avro directory and the initial schema that resemblesGravatar A. Unique TensorFlower2016-05-27
| | | | | tf.train.Example. Change: 123445810
* Pass -O3 when building tensorflow with clang -c opt. (This isn't hookedGravatar A. Unique TensorFlower2016-05-25
| | | | | up to the OSS build yet, we're working on it.) Change: 123248081
* Upgraded to the latest version of Eigen that supports convolutions on fp16Gravatar Benoit Steiner2016-05-25
| | | | Change: 123238579
* Merge changes from github.Gravatar Derek Murray2016-05-23
| | | | Change: 123026122
* Google authentication for GCS file system.Gravatar Alexey Surkov2016-05-19
| | | | | | | Implements an authentication mechanism based on Application Default Credentials: https://developers.google.com/identity/protocols/application-default-credentials https://developers.google.com/identity/protocols/OAuth2ServiceAccount Change: 122741738
* Switched to the latest version of Eigen that performs much better on machinesGravatar Benoit Steiner2016-05-16
| | | | | | | | with many cpu cores For example, the wall time for the following tutorial went down from 13m35 to 5m27: bazel run -c opt --copt=-mavx tensorflow/examples/tutorials/word2vec/word2vec_basic Change: 122462177
* Move if_cuda build macro into //third_party/gpus/cuda/build_defs.bzl,Gravatar A. Unique TensorFlower2016-05-13
| | | | | | and remove the cuda_crosstool_condition build condition. Now if_cuda is just using_nvcc || using_gcudacc. Change: 122291892
* Add //third_party/gpus/cuda:using_clang build condition.Gravatar A. Unique TensorFlower2016-05-13
| | | | | | This is unused at the moment, but will eventually let us build CUDA code with vanilla clang. Change: 122289910
* Pass --define=using_cuda_nvcc to CUDA builds.Gravatar A. Unique TensorFlower2016-05-13
| | | | | | | | This has no practical effect, as CUDA builds are always with nvcc, but it lets us modify the build config rule //third_party/gpus/cuda:using_nvcc so it returns true, rather than false, for CUDA builds. Change: 122288952
* Made the contraction code compatible with fp16Gravatar Benoit Steiner2016-05-12
| | | | Change: 122192081
* Upgraded to the latest version of Eigen that speeds up full reductions on fp16Gravatar Benoit Steiner2016-05-12
| | | | | by about 3 orders of magnitude as well as some partial reductions by 30% when using cuda 7.5 or above Change: 122191448
* Merge changes from github.Gravatar A. Unique TensorFlower2016-05-05
| | | | Change: 121586635
* Improved support for min and max on 16 bit floats when running on recent cudaGravatar Benoit Steiner2016-04-27
| | | | | | gpus Updated the check numerics code to make it compatible with fp16 Change: 120980302
* Made it possible to compute the cross entropy using 16 bit floatsGravatar Benoit Steiner2016-04-25
| | | | Change: 120739269
* Rollback of rollback of cl/120366069:Gravatar A. Unique TensorFlower2016-04-21
| | | | | | | | tensorflow: switch to eigen thread pool This is first step of switching tensorflow to the new non-blocking thread pool in eigen. Change: 120510292
* Prevent TensorFlow from crashing when attempting to reduce an empty tensor ↵Gravatar Benoit Steiner2016-04-21
| | | | | | on GPU Change: 120505517
* Fixed a compilation error when targeting cuda 3.0 devices such as the onesGravatar Benoit Steiner2016-04-20
| | | | | offered by AWS Change: 120369420
* Upgraded to the latest version of Eigen that adds support for computing theGravatar Benoit Steiner2016-04-14
| | | | | sigmoid of fp16 and introduces a condition estimator. Change: 119907721
* Added support for trigonometric and transcendental functions of half floatsGravatar Benoit Steiner2016-04-14
| | | | Change: 119850987
* Upgraded to the latest version of Eigen that provides significant performanceGravatar Benoit Steiner2016-04-13
| | | | | improvements for fp16 Change: 119771118
* Minimal open source CUPTI GPU Tracer.Gravatar A. Unique TensorFlower2016-04-13
| | | | Change: 119768540
* Made isinf, isnan, isfinite, ceil and floor work with 16 bit floats.Gravatar Benoit Steiner2016-04-09
| | | | Change: 119458778
* Upgraded to the latest version of Eigen that has bug fixes for complex numbersGravatar Benoit Steiner2016-04-08
| | | | | as well as fp16 Change: 119398881
* Fix genrules that didn't work when TensorFlow was imported as a submodule ↵Gravatar Kiril Gorovoy2016-04-07
| | | | | | and compiled with --config=cuda. Change: 119318629
* Upgraded to the latest version of eigen that introduces implementations of ↵Gravatar Benoit Steiner2016-04-07
| | | | | | | the zeta and polygamma functions, as well as improved support for float16. Change: 119279101
* Upgrade to the latest version of Eigen that provides better support for float16Gravatar Benoit Steiner2016-04-04
| | | | | and fixes the computation of absolute values on gpu. Change: 119001808
* Merge changes from github.Gravatar Vijay Vasudevan2016-03-29
| | | | Change: 118532471
* Upgraded to the latest version of EigenGravatar Benoit Steiner2016-03-28
| | | | Change: 118414762
* Upgraded to the latest version of Eigen that provides better support for fp16.Gravatar Benoit Steiner2016-03-28
| | | | | Use Eigen mod functors directly instead of duplicating them. Change: 118362359
* Move the NeuralNetwork code out of third_party/eigen3 and intoGravatar Benoit Steiner2016-03-23
| | | | | tensorflow/core/kernel. Change: 117941211
* Update Eigen NN headers in staging repo to match public contents.Gravatar Vijay Vasudevan2016-03-22
|
* Re-rollback of "TensorFlow: move eigen some NN code from our ↵Gravatar Vijay Vasudevan2016-03-18
| | | | | | | third_party/eigen3 copy to being part of TF, add tests." Change: 117608627
* Rollforward of "TensorFlow: move eigen some NN code from our ↵Gravatar Vijay Vasudevan2016-03-18
| | | | | | | third_party/eigen3 copy to being part of TF, add tests." Change: 117587217
* TensorFlow: update eigen to latest change to fix TensorChippingGravatar Vijay Vasudevan2016-03-18
| | | | Change: 117570343
* Rollback of "TensorFlow: move eigen some NN code from our third_party/eigen3 ↵Gravatar Vijay Vasudevan2016-03-18
| | | | | | | copy to being part of TF, add tests." Change: 117519243