aboutsummaryrefslogtreecommitdiffhomepage
diff options
context:
space:
mode:
-rw-r--r--CONTRIBUTING.md10
-rw-r--r--RELEASE.md114
-rw-r--r--configure.py28
-rw-r--r--tensorflow/compiler/tf2xla/graph_compiler.h2
-rw-r--r--tensorflow/compiler/xla/index_util.h2
-rw-r--r--tensorflow/compiler/xla/python/xla_client.py2
-rw-r--r--tensorflow/compiler/xla/service/BUILD2
-rw-r--r--tensorflow/compiler/xla/service/buffer_assignment_test.cc2
-rw-r--r--tensorflow/compiler/xla/service/cpu/ir_emitter.cc2
-rw-r--r--tensorflow/compiler/xla/service/cpu/ir_function.cc4
-rw-r--r--tensorflow/compiler/xla/service/hlo_evaluator.cc2
-rw-r--r--tensorflow/contrib/all_reduce/python/all_reduce.py2
-rw-r--r--tensorflow/contrib/cmake/CMakeLists.txt16
-rw-r--r--tensorflow/contrib/cmake/external/boringssl.cmake1
-rw-r--r--tensorflow/contrib/cmake/external/farmhash.cmake1
-rw-r--r--tensorflow/contrib/cmake/external/fft2d.cmake1
-rw-r--r--tensorflow/contrib/cmake/external/gif.cmake1
-rw-r--r--tensorflow/contrib/cmake/external/googletest.cmake10
-rw-r--r--tensorflow/contrib/cmake/external/grpc.cmake16
-rw-r--r--tensorflow/contrib/cmake/external/highwayhash.cmake1
-rw-r--r--tensorflow/contrib/cmake/external/jemalloc.cmake15
-rw-r--r--tensorflow/contrib/cmake/external/jpeg.cmake1
-rw-r--r--tensorflow/contrib/cmake/external/jsoncpp.cmake7
-rw-r--r--tensorflow/contrib/cmake/external/lmdb.cmake13
-rw-r--r--tensorflow/contrib/cmake/external/nsync.cmake1
-rw-r--r--tensorflow/contrib/cmake/external/png.cmake17
-rw-r--r--tensorflow/contrib/cmake/external/protobuf.cmake44
-rw-r--r--tensorflow/contrib/cmake/external/re2.cmake7
-rw-r--r--tensorflow/contrib/cmake/external/snappy.cmake7
-rw-r--r--tensorflow/contrib/cmake/external/sqlite.cmake1
-rw-r--r--tensorflow/contrib/cmake/external/zlib.cmake17
-rw-r--r--tensorflow/contrib/cmake/python_modules.txt6
-rw-r--r--tensorflow/contrib/cmake/tests/cuda/compatibility_test.c22
-rw-r--r--tensorflow/contrib/cmake/tests/cuda/compatibility_test.cc20
-rw-r--r--tensorflow/contrib/cmake/tf_cc_ops.cmake6
-rwxr-xr-xtensorflow/contrib/cmake/tf_python.cmake25
-rw-r--r--tensorflow/contrib/cmake/tf_shared_lib.cmake6
-rw-r--r--tensorflow/contrib/cmake/tf_tests.cmake2
-rw-r--r--tensorflow/contrib/cmake/tf_tools.cmake3
-rw-r--r--tensorflow/contrib/data/python/ops/dataset_ops.py691
-rw-r--r--tensorflow/contrib/data/python/ops/grouping.py2
-rw-r--r--tensorflow/contrib/distributions/__init__.py1
-rw-r--r--tensorflow/contrib/eager/python/examples/gan/README.md2
-rw-r--r--tensorflow/contrib/eager/python/examples/mnist/mnist.py2
-rw-r--r--tensorflow/contrib/framework/python/ops/accumulate_n_v2.py2
-rw-r--r--tensorflow/contrib/gan/python/eval/python/summaries_impl.py7
-rw-r--r--tensorflow/contrib/gan/python/eval/python/summaries_test.py14
-rw-r--r--tensorflow/contrib/gdr/README.md2
-rw-r--r--tensorflow/contrib/layers/__init__.py2
-rw-r--r--tensorflow/contrib/layers/python/layers/layers.py69
-rw-r--r--tensorflow/contrib/layers/python/layers/layers_test.py50
-rw-r--r--tensorflow/contrib/learn/python/learn/monitors.py2
-rw-r--r--tensorflow/contrib/lite/README.md2
-rw-r--r--tensorflow/contrib/lite/examples/label_image/bitmap_helpers_impl.h6
-rw-r--r--tensorflow/contrib/lite/g3doc/custom_operators.md2
-rw-r--r--tensorflow/contrib/lite/graph_info.cc2
-rw-r--r--tensorflow/contrib/lite/kernels/internal/optimized/tensor_utils_impl.h2
-rw-r--r--tensorflow/contrib/lite/kernels/internal/reference/portable_tensor_utils.h2
-rw-r--r--tensorflow/contrib/lite/simple_memory_arena.h4
-rw-r--r--tensorflow/contrib/lite/testing/generate_examples.py2
-rw-r--r--tensorflow/contrib/lite/toco/graph_transformations/resolve_constant_concatenation.cc2
-rw-r--r--tensorflow/contrib/lite/toco/tensorflow_graph_matching/resolve_svdf_test.cc12
-rw-r--r--tensorflow/contrib/makefile/README.md2
-rwxr-xr-xtensorflow/contrib/makefile/build_all_android.sh2
-rw-r--r--tensorflow/contrib/metrics/python/ops/metric_ops.py4
-rw-r--r--tensorflow/contrib/nearest_neighbor/kernels/heap.h2
-rw-r--r--tensorflow/contrib/opt/python/training/elastic_average_optimizer.py2
-rw-r--r--tensorflow/contrib/opt/python/training/model_average_optimizer_test.py1
-rw-r--r--tensorflow/contrib/py2tf/converters/BUILD1
-rw-r--r--tensorflow/contrib/rnn/python/kernel_tests/rnn_cell_test.py1
-rw-r--r--tensorflow/contrib/seq2seq/python/ops/attention_wrapper.py2
-rw-r--r--tensorflow/contrib/slim/README.md2
-rw-r--r--tensorflow/contrib/solvers/python/ops/linear_equations.py1
-rw-r--r--tensorflow/contrib/tpu/profiler/pip_package/setup.py2
-rw-r--r--tensorflow/contrib/tpu/tpu_estimator.md2
-rw-r--r--tensorflow/contrib/verbs/README.md12
-rw-r--r--tensorflow/core/distributed_runtime/rpc/grpc_serialization_traits.h2
-rw-r--r--tensorflow/core/framework/op_def_util.h2
-rw-r--r--tensorflow/core/framework/op_kernel.cc2
-rw-r--r--tensorflow/core/kernels/constant_op_test.cc2
-rw-r--r--tensorflow/core/kernels/cwise_op_gpu_invert.cu.cc2
-rw-r--r--tensorflow/core/kernels/cwise_op_invert.cc10
-rw-r--r--tensorflow/core/kernels/cwise_ops_gpu_common.cu.h3
-rw-r--r--tensorflow/core/lib/gif/gif_io.cc42
-rw-r--r--tensorflow/core/lib/gtl/optional.h2
-rw-r--r--tensorflow/core/lib/io/record_reader.cc2
-rw-r--r--tensorflow/core/ops/image_ops.cc24
-rw-r--r--tensorflow/core/platform/env.cc37
-rw-r--r--tensorflow/core/platform/env.h8
-rw-r--r--tensorflow/core/platform/file_system.cc4
-rw-r--r--tensorflow/core/platform/file_system.h3
-rw-r--r--tensorflow/core/platform/posix/posix_file_system.cc72
-rw-r--r--tensorflow/core/platform/posix/posix_file_system.h2
-rw-r--r--tensorflow/core/public/version.h4
-rw-r--r--tensorflow/docs_src/api_guides/python/contrib.distributions.md1
-rw-r--r--tensorflow/docs_src/api_guides/python/regression_examples.md2
-rw-r--r--tensorflow/docs_src/community/welcome.md2
-rw-r--r--tensorflow/docs_src/get_started/get_started_for_beginners.md2
-rw-r--r--tensorflow/docs_src/install/install_c.md2
-rw-r--r--tensorflow/docs_src/install/install_go.md2
-rw-r--r--tensorflow/docs_src/install/install_java.md22
-rw-r--r--tensorflow/docs_src/install/install_linux.md22
-rw-r--r--tensorflow/docs_src/install/install_mac.md10
-rw-r--r--tensorflow/docs_src/install/install_sources.md12
-rw-r--r--tensorflow/docs_src/programmers_guide/debugger.md2
-rw-r--r--tensorflow/docs_src/programmers_guide/index.md2
-rw-r--r--tensorflow/examples/image_retraining/retrain.py4
-rw-r--r--tensorflow/java/src/main/java/org/tensorflow/NativeLibrary.java2
-rw-r--r--tensorflow/python/client/session.py18
-rw-r--r--tensorflow/python/data/ops/dataset_ops.py2
-rw-r--r--tensorflow/python/debug/cli/cli_shared.py2
-rw-r--r--tensorflow/python/debug/cli/tensor_format.py2
-rw-r--r--tensorflow/python/framework/test_util.py96
-rw-r--r--tensorflow/python/keras/_impl/keras/backend.py8
-rw-r--r--tensorflow/python/keras/_impl/keras/backend_test.py9
-rw-r--r--tensorflow/python/keras/_impl/keras/layers/lstm_test.py16
-rw-r--r--tensorflow/python/keras/_impl/keras/layers/recurrent.py3
-rw-r--r--tensorflow/python/kernel_tests/conv2d_transpose_test.py6
-rw-r--r--tensorflow/python/kernel_tests/distributions/bernoulli_test.py7
-rw-r--r--tensorflow/python/kernel_tests/distributions/beta_test.py4
-rw-r--r--tensorflow/python/kernel_tests/neon_depthwise_conv_op_test.py4
-rw-r--r--tensorflow/python/kernel_tests/py_func_test.py11
-rw-r--r--tensorflow/python/layers/maxout.py2
-rw-r--r--tensorflow/python/lib/io/file_io.i16
-rw-r--r--tensorflow/python/ops/bitwise_ops_test.py6
-rw-r--r--tensorflow/python/ops/distributions/bernoulli.py20
-rw-r--r--tensorflow/python/ops/distributions/beta.py9
-rw-r--r--tensorflow/python/ops/image_ops_test.py42
-rw-r--r--tensorflow/python/ops/linalg/linear_operator.py1
-rw-r--r--tensorflow/python/ops/losses/losses_impl.py8
-rw-r--r--tensorflow/python/ops/script_ops.py4
-rw-r--r--tensorflow/python/ops/standard_ops.py1
-rw-r--r--tensorflow/python/ops/state_ops.py2
-rw-r--r--tensorflow/python/util/compat_internal.py2
-rw-r--r--tensorflow/stream_executor/dnn.h2
-rwxr-xr-xtensorflow/tools/ci_build/ci_sanity.sh2
-rw-r--r--tensorflow/tools/docker/Dockerfile.devel2
-rw-r--r--tensorflow/tools/docker/Dockerfile.devel-cpu-mkl2
-rw-r--r--tensorflow/tools/docker/Dockerfile.devel-gpu2
-rw-r--r--tensorflow/tools/graph_transforms/BUILD11
-rw-r--r--tensorflow/tools/pip_package/setup.py2
-rw-r--r--third_party/repo.bzl2
142 files changed, 1673 insertions, 320 deletions
diff --git a/CONTRIBUTING.md b/CONTRIBUTING.md
index de4fded6ae..3dad41a88c 100644
--- a/CONTRIBUTING.md
+++ b/CONTRIBUTING.md
@@ -41,7 +41,7 @@ TensorFlow coding style.
#### General guidelines and philosophy for contribution
* Include unit tests when you contribute new features, as they help to
- a) prove that your code works correctly, b) guard against future breaking
+ a) prove that your code works correctly, and b) guard against future breaking
changes to lower the maintenance cost.
* Bug fixes also generally require unit tests, because the presence of bugs
usually indicates insufficient test coverage.
@@ -51,7 +51,7 @@ TensorFlow coding style.
non-backward-compatible API changes without a major release. Reviewers of your
pull request will comment on any API compatibility issues.
* When you contribute a new feature to TensorFlow, the maintenance burden is (by
- default) transferred to the TensorFlow team. This means that benefit of
+ default) transferred to the TensorFlow team. This means that benefit of the
contribution must be compared against the cost of maintaining the feature.
* Full new features (e.g., a new op implementing a cutting-edge algorithm)
typically will live in
@@ -68,8 +68,8 @@ Include a license at the top of new files.
* [Java license example](https://github.com/tensorflow/tensorflow/blob/master/tensorflow/java/src/main/java/org/tensorflow/Graph.java#L1)
* [Go license example](https://github.com/tensorflow/tensorflow/blob/master/tensorflow/go/operation.go#L1)
* [Bash license example](https://github.com/tensorflow/tensorflow/blob/master/tensorflow/tools/ci_build/ci_sanity.sh#L2)
-* [HTML license example](https://github.com/tensorflow/tensorflow/blob/master/tensorflow/tensorboard/dist/index.html#L2)
-* [JavaScript/TypeScript license example](https://github.com/tensorflow/tensorflow/blob/master/tensorflow/tensorboard/components/tf_backend/backend.ts#L1)
+* [HTML license example](https://github.com/tensorflow/tensorboard/blob/master/tensorboard/components/tf_backend/tf-backend.html#L2)
+* [JavaScript/TypeScript license example](https://github.com/tensorflow/tensorboard/blob/master/tensorboard/components/tf_backend/backend.ts#L1)
Bazel BUILD files also need to include a license section, e.g.,
[BUILD example](https://github.com/tensorflow/tensorflow/blob/master/tensorflow/core/BUILD#L61).
@@ -163,7 +163,7 @@ There are two ways to run TensorFlow unit tests.
bazel test ${flags} //tensorflow/python/...
```
-2. Using [Docker](www.docker.com) and TensorFlow's CI scripts.
+2. Using [Docker](https://www.docker.com) and TensorFlow's CI scripts.
```bash
# Install Docker first, then this will build and run cpu tests
diff --git a/RELEASE.md b/RELEASE.md
index b11b1e40db..0720a8c639 100644
--- a/RELEASE.md
+++ b/RELEASE.md
@@ -1,9 +1,98 @@
+# Release 1.6.0
+
+## Breaking Changes
+* Prebuilt binaries are now built against CUDA 9.0 and cuDNN 7.
+* Prebuilt binaries will use AVX instructions. This may break TF on older CPUs.
+
+## Major Features And Improvements
+* New Optimizer internal API for non-slot variables. Descendants of AdamOptimizer that access _beta[12]_power will need to be updated.
+* `tf.estimator.{FinalExporter,LatestExporter}` now export stripped SavedModels. This improves forward compatibility of the SavedModel.
+* FFT support added to XLA CPU/GPU.
+
+## Bug Fixes and Other Changes
+* Documentation updates:
+ * Added a second version of Getting Started, which is aimed at ML
+newcomers.
+ * Clarified documentation on `resize_images.align_corners` parameter.
+ * Additional documentation for TPUs.
+* Google Cloud Storage (GCS):
+ * Add client-side throttle.
+ * Add a `FlushCaches()` method to the FileSystem interface, with an implementation for GcsFileSystem.
+* Other:
+ * Add `tf.contrib.distributions.Kumaraswamy`.
+ * `RetryingFileSystem::FlushCaches()` calls the base FileSystem's `FlushCaches()`.
+ * Add auto_correlation to distributions.
+ * Add `tf.contrib.distributions.Autoregressive`.
+ * Add SeparableConv1D layer.
+ * Add convolutional Flipout layers.
+ * When both inputs of `tf.matmul` are bfloat16, it returns bfloat16, instead of float32.
+ * Added `tf.contrib.image.connected_components`.
+ * Add `tf.contrib.framework.CriticalSection` that allows atomic variable access.
+ * Output variance over trees predictions for classifications tasks.
+ * For `pt` and `eval` commands, allow writing tensor values to filesystem as numpy files.
+ * gRPC: Propagate truncated errors (instead of returning gRPC internal error).
+ * Augment parallel_interleave to support 2 kinds of prefetching.
+ * Improved XLA support for C64-related ops log, pow, atan2, tanh.
+ * Add probabilistic convolutional layers.
+
+## API Changes
+* Introducing prepare_variance boolean with default setting to False for backward compatibility.
+* Move `layers_dense_variational_impl.py` to `layers_dense_variational.py`.
+
+## Known Bugs
+* Using XLA:GPU with CUDA 9 and CUDA 9.1 results in garbage results and/or
+ `CUDA_ILLEGAL_ADDRESS` failures.
+
+ Google discovered in mid-December 2017 that the PTX-to-SASS compiler in CUDA 9
+ and CUDA 9.1 sometimes does not properly compute the carry bit when
+ decomposing 64-bit address calculations with large offsets (e.g. `load [x +
+ large_constant]`) into 32-bit arithmetic in SASS.
+
+ As a result, these versions of `ptxas` miscompile most XLA programs which use
+ more than 4GB of temp memory. This results in garbage results and/or
+ `CUDA_ERROR_ILLEGAL_ADDRESS` failures.
+
+ A fix in CUDA 9.1.121 is expected in late February 2018. We do not expect a
+ fix for CUDA 9.0.x. Until the fix is available, the only workaround is to
+ [downgrade](https://developer.nvidia.com/cuda-toolkit-archive) to CUDA 8.0.x
+ or disable XLA:GPU.
+
+ TensorFlow will print a warning if you use XLA:GPU with a known-bad version of
+ CUDA; see e00ba24c4038e7644da417ddc639169b6ea59122.
+
+## Thanks to our Contributors
+
+This release contains contributions from many people at Google, as well as:
+
+4d55397500, Ag Ramesh, Aiden Scandella, Akimasa Kimura, Alex Rothberg, Allen Goodman,
+amilioto, Andrei Costinescu, Andrei Nigmatulin, Anjum Sayed, Anthony Platanios,
+Anush Elangovan, Armando Fandango, Ashish Kumar Ram, Ashwini Shukla, Ben, Bhavani Subramanian,
+Brett Koonce, Carl Thomé, cclauss, Cesc, Changming Sun, Christoph Boeddeker, Clayne Robison,
+Clemens Schulz, Clint (Woonhyuk Baek), codrut3, Cole Gerdemann, Colin Raffel, Daniel Trebbien,
+Daniel Ylitalo, Daniel Zhang, Daniyar, Darjan Salaj, Dave Maclachlan, David Norman, Dong--Jian,
+dongsamb, dssgsra, Edward H, eladweiss, elilienstein, Eric Lilienstein, error.d, Eunji Jeong, fanlu,
+Florian Courtial, fo40225, Fred, Gregg Helt, Guozhong Zhuang, Hanchen Li, hsm207, hyunyoung2,
+ImSheridan, Ishant Mrinal Haloi, Jacky Ko, Jay Young, Jean Flaherty, Jerome, JerrikEph, Jesse
+Kinkead, jfaath, Jian Lin, jinghuangintel, Jiongyan Zhang, Joel Hestness, Joel Shor, Johnny Chan,
+Julian Niedermeier, Julian Wolff, JxKing, K-W-W, Karl Lessard, Kasper Marstal, Keiji Ariyama,
+Koan-Sin Tan, Loki Der Quaeler, Loo Rong Jie, Luke Schaefer, Lynn Jackson, ManHyuk, Matt Basta,
+Matt Smith, Matthew Schulkind, Michael, michaelkhan3, Miguel Piedrafita, Mikalai Drabovich,
+Mike Knapp, mjwen, mktozk, Mohamed Aly, Mohammad Ashraf Bhuiyan, Myungjoo Ham, Naman Bhalla,
+Namrata-Ibm, Nathan Luehr, nathansilberman, Netzeband, Niranjan Hasabnis, Omar Aflak, Ozge
+Yalcinkaya, Parth P Panchal, patrickzzy, Patryk Chrabaszcz, Paul Van Eck, Paweł Kapica, Peng Yu,
+Philip Yang, Pierre Blondeau, Po-Hsien Chu, powderluv, Puyu Wang, Rajendra Arora, Rasmus, Renat
+Idrisov, resec, Robin Richtsfeld, Ronald Eddy Jr, Sahil Singh, Sam Matzek, Sami Kama, sandipmgiri,
+Santiago Castro, Sayed Hadi Hashemi, Scott Tseng, Sergii Khomenko, Shahid, Shengpeng Liu, Shreyash
+Sharma, Shrinidhi Kl, Simone Cirillo, simsicon, Stanislav Levental, starsblinking, Stephen Lumenta,
+Steven Hickson, Su Tang, Taehoon Lee, Takuya Wakisaka, Ted Chang, Ted Ying, Tijmen Verhulsdonck,
+Timofey Kondrashov, vade, vaibhav, Valentin Khrulkov, vchigrin, Victor Costan, Viraj Navkal,
+Vivek Rane, wagonhelm, Yan Facai (颜发才), Yanbo Liang, Yaroslav Bulatov, yegord, Yong Tang,
+Yoni Tsafir, yordun, Yuan (Terry) Tang, Yuxin Wu, zhengdi, Zhengsheng Wei, 田传武
+
# Release 1.5.0
## Breaking Changes
* Prebuilt binaries are now built against CUDA 9.0 and cuDNN 7.
-* Our Linux binaries are built using ubuntu 16 containers, potentially
- introducing glibc incompatibility issues with ubuntu 14.
* Starting from 1.6 release, our prebuilt binaries will use AVX instructions.
This may break TF on older CPUs.
@@ -146,6 +235,27 @@
* Minor refactor: move stats files from `stochastic` to `common` and remove
`stochastic`.
+## Known Bugs
+* Using XLA:GPU with CUDA 9 and CUDA 9.1 results in garbage results and/or
+ `CUDA_ILLEGAL_ADDRESS` failures.
+
+ Google discovered in mid-December 2017 that the PTX-to-SASS compiler in CUDA 9
+ and CUDA 9.1 sometimes does not properly compute the carry bit when
+ decomposing 64-bit address calculations with large offsets (e.g. `load [x +
+ large_constant]`) into 32-bit arithmetic in SASS.
+
+ As a result, these versions of `ptxas` miscompile most XLA programs which use
+ more than 4GB of temp memory. This results in garbage results and/or
+ `CUDA_ERROR_ILLEGAL_ADDRESS` failures.
+
+ A fix in CUDA 9.1.121 is expected in late February 2018. We do not expect a
+ fix for CUDA 9.0.x. Until the fix is available, the only workaround is to
+ [downgrade](https://developer.nvidia.com/cuda-toolkit-archive) to CUDA 8.0.x
+ or disable XLA:GPU.
+
+ TensorFlow will print a warning if you use XLA:GPU with a known-bad version of
+ CUDA; see e00ba24c4038e7644da417ddc639169b6ea59122.
+
## Thanks to our Contributors
This release contains contributions from many people at Google, as well as:
diff --git a/configure.py b/configure.py
index 27519b4aba..6b1fa7f1a8 100644
--- a/configure.py
+++ b/configure.py
@@ -827,6 +827,28 @@ def set_gcc_host_compiler_path(environ_cp):
write_action_env_to_bazelrc('GCC_HOST_COMPILER_PATH', gcc_host_compiler_path)
+def reformat_version_sequence(version_str, sequence_count):
+ """Reformat the version string to have the given number of sequences.
+
+ For example:
+ Given (7, 2) -> 7.0
+ (7.0.1, 2) -> 7.0
+ (5, 1) -> 5
+ (5.0.3.2, 1) -> 5
+
+ Args:
+ version_str: String, the version string.
+ sequence_count: int, an integer.
+ Returns:
+ string, reformatted version string.
+ """
+ v = version_str.split('.')
+ if len(v) < sequence_count:
+ v = v + (['0'] * (sequence_count - len(v)))
+
+ return '.'.join(v[:sequence_count])
+
+
def set_tf_cuda_version(environ_cp):
"""Set CUDA_TOOLKIT_PATH and TF_CUDA_VERSION."""
ask_cuda_version = (
@@ -837,6 +859,7 @@ def set_tf_cuda_version(environ_cp):
# Configure the Cuda SDK version to use.
tf_cuda_version = get_from_env_or_user_or_default(
environ_cp, 'TF_CUDA_VERSION', ask_cuda_version, _DEFAULT_CUDA_VERSION)
+ tf_cuda_version = reformat_version_sequence(str(tf_cuda_version), 2)
# Find out where the CUDA toolkit is installed
default_cuda_path = _DEFAULT_CUDA_PATH
@@ -893,6 +916,7 @@ def set_tf_cudnn_version(environ_cp):
tf_cudnn_version = get_from_env_or_user_or_default(
environ_cp, 'TF_CUDNN_VERSION', ask_cudnn_version,
_DEFAULT_CUDNN_VERSION)
+ tf_cudnn_version = reformat_version_sequence(str(tf_cudnn_version), 1)
default_cudnn_path = environ_cp.get('CUDA_TOOLKIT_PATH')
ask_cudnn_path = (r'Please specify the location where cuDNN %s library is '
@@ -1400,6 +1424,10 @@ def main():
if is_linux():
set_tf_tensorrt_install_path(environ_cp)
set_tf_cuda_compute_capabilities(environ_cp)
+ if 'LD_LIBRARY_PATH' in environ_cp and environ_cp.get(
+ 'LD_LIBRARY_PATH') != '1':
+ write_action_env_to_bazelrc('LD_LIBRARY_PATH',
+ environ_cp.get('LD_LIBRARY_PATH'))
set_tf_cuda_clang(environ_cp)
if environ_cp.get('TF_CUDA_CLANG') == '1':
diff --git a/tensorflow/compiler/tf2xla/graph_compiler.h b/tensorflow/compiler/tf2xla/graph_compiler.h
index ba00160b6d..127562eb23 100644
--- a/tensorflow/compiler/tf2xla/graph_compiler.h
+++ b/tensorflow/compiler/tf2xla/graph_compiler.h
@@ -70,7 +70,7 @@ class GraphCompiler {
private:
// Partially sets params. This partially set params can be reused
- // across multple nodes visit.
+ // across multiple nodes visit.
void PartiallySetupParams(OpKernelContext::Params* params);
// Tests if a node is a functional node. A functional node represents a
diff --git a/tensorflow/compiler/xla/index_util.h b/tensorflow/compiler/xla/index_util.h
index 0b9188e852..142006f262 100644
--- a/tensorflow/compiler/xla/index_util.h
+++ b/tensorflow/compiler/xla/index_util.h
@@ -37,7 +37,7 @@ class IndexUtil {
static int64 MultidimensionalIndexToLinearIndex(
const Shape& shape, tensorflow::gtl::ArraySlice<int64> multi_index);
- // Coverts a linear index into multidimensional index (eg {x, y, z}) based on
+ // Converts a linear index into multidimensional index (eg {x, y, z}) based on
// the shape and its layout. The first index in the returned multidimensional
// index is dimension 0.
static std::vector<int64> LinearIndexToMultidimensionalIndex(
diff --git a/tensorflow/compiler/xla/python/xla_client.py b/tensorflow/compiler/xla/python/xla_client.py
index bcff9508fb..9bda9d0929 100644
--- a/tensorflow/compiler/xla/python/xla_client.py
+++ b/tensorflow/compiler/xla/python/xla_client.py
@@ -1128,7 +1128,7 @@ def initialize_replica_count(replica_count):
Args:
replica_count: number of replicas that are desired for set up during XLA
- initalization.
+ initialization.
Raises:
A runtime exception if the XLA service has already been initialized.
diff --git a/tensorflow/compiler/xla/service/BUILD b/tensorflow/compiler/xla/service/BUILD
index 22d1c5a9bd..4a076ac090 100644
--- a/tensorflow/compiler/xla/service/BUILD
+++ b/tensorflow/compiler/xla/service/BUILD
@@ -171,8 +171,6 @@ cc_library(
":shape_inference",
"//tensorflow/compiler/xla:literal_util",
"//tensorflow/compiler/xla:shape_util",
- "//tensorflow/compiler/xla:status",
- "//tensorflow/compiler/xla:status_macros",
"//tensorflow/compiler/xla:statusor",
"//tensorflow/compiler/xla:types",
"//tensorflow/compiler/xla:util",
diff --git a/tensorflow/compiler/xla/service/buffer_assignment_test.cc b/tensorflow/compiler/xla/service/buffer_assignment_test.cc
index ef067cc31f..cd73654b8f 100644
--- a/tensorflow/compiler/xla/service/buffer_assignment_test.cc
+++ b/tensorflow/compiler/xla/service/buffer_assignment_test.cc
@@ -614,7 +614,7 @@ TEST_F(BufferAssignmentTest, TrivialMap) {
BufferAllocation map_buffer = GetAssignedOutputAllocation(*buffers, map);
EXPECT_NE(param0_buffer.index(), map_buffer.index());
- // The final computation node of the map is an add of an f32 parm and a
+ // The final computation node of the map is an add of an f32 param and a
// constant.
EXPECT_EQ(HloOpcode::kAdd, inner_last->opcode());
const BufferAllocation& inner_add_buffer =
diff --git a/tensorflow/compiler/xla/service/cpu/ir_emitter.cc b/tensorflow/compiler/xla/service/cpu/ir_emitter.cc
index 496aea051c..4dffaee87f 100644
--- a/tensorflow/compiler/xla/service/cpu/ir_emitter.cc
+++ b/tensorflow/compiler/xla/service/cpu/ir_emitter.cc
@@ -1337,7 +1337,7 @@ IrEmitter::ReductionGenerator IrEmitter::MatchReductionGenerator(
if (ShapeUtil::ElementIsComplex(root_shape)) {
// TODO(b/65408531): Complex add could by done via bitcast to <float x [2N]>
// Complex multiply would be more challenging. We could perhaps use a
- // strided load to get all reals in a vector, all imags in a vector, or use
+ // strided load to get all reals in a vector, all images in a vector, or use
// CreateShuffleVector on a bitcast to float x [2N].
*failure_reason = "complex values not supported";
return nullptr;
diff --git a/tensorflow/compiler/xla/service/cpu/ir_function.cc b/tensorflow/compiler/xla/service/cpu/ir_function.cc
index ca8c290dd1..2d6f2f3818 100644
--- a/tensorflow/compiler/xla/service/cpu/ir_function.cc
+++ b/tensorflow/compiler/xla/service/cpu/ir_function.cc
@@ -209,9 +209,9 @@ std::vector<llvm::Value*> GetArrayFunctionCallArguments(
parameter_addresses[i], ir_builder->getInt8PtrTy(),
AsStringRef(tensorflow::strings::StrCat(name, "_parameter_", i,
"_address_as_i8ptr")));
- llvm::Value* slot_in_param_adresses = ir_builder->CreateInBoundsGEP(
+ llvm::Value* slot_in_param_addresses = ir_builder->CreateInBoundsGEP(
parameter_addresses_buffer, {ir_builder->getInt64(i)});
- ir_builder->CreateStore(parameter_as_i8ptr, slot_in_param_adresses);
+ ir_builder->CreateStore(parameter_as_i8ptr, slot_in_param_addresses);
}
const auto to_int8_ptr = [=](llvm::Value* ptr) {
diff --git a/tensorflow/compiler/xla/service/hlo_evaluator.cc b/tensorflow/compiler/xla/service/hlo_evaluator.cc
index 8016b38d15..296f010a92 100644
--- a/tensorflow/compiler/xla/service/hlo_evaluator.cc
+++ b/tensorflow/compiler/xla/service/hlo_evaluator.cc
@@ -34,8 +34,6 @@ limitations under the License.
#include "tensorflow/compiler/xla/service/hlo_query.h"
#include "tensorflow/compiler/xla/service/shape_inference.h"
#include "tensorflow/compiler/xla/shape_util.h"
-#include "tensorflow/compiler/xla/status.h"
-#include "tensorflow/compiler/xla/status_macros.h"
#include "tensorflow/compiler/xla/types.h"
#include "tensorflow/compiler/xla/util.h"
#include "tensorflow/compiler/xla/window_util.h"
diff --git a/tensorflow/contrib/all_reduce/python/all_reduce.py b/tensorflow/contrib/all_reduce/python/all_reduce.py
index 16617b7266..6658f0d9c1 100644
--- a/tensorflow/contrib/all_reduce/python/all_reduce.py
+++ b/tensorflow/contrib/all_reduce/python/all_reduce.py
@@ -758,7 +758,7 @@ def _build_nccl_hybrid(input_tensors, red_op, upper_level_f):
def _reduce_non_singleton(input_tensors, red_f, un_op):
- """If input_tenors has more than one element apply red_f, else apply un_op."""
+ """If input_tensors has more than one element apply red_f, else apply un_op."""
if len(input_tensors) > 1:
return red_f(input_tensors)
else:
diff --git a/tensorflow/contrib/cmake/CMakeLists.txt b/tensorflow/contrib/cmake/CMakeLists.txt
index c9ba1f4cc0..524946a9a5 100644
--- a/tensorflow/contrib/cmake/CMakeLists.txt
+++ b/tensorflow/contrib/cmake/CMakeLists.txt
@@ -286,7 +286,21 @@ if (tensorflow_ENABLE_GPU)
list(APPEND CMAKE_LIBRARY_PATH "${tensorflow_CUDA_LIBRARY_PATH}/stubs")
endif (NOT WIN32)
- find_package(CUDA ${tensorflow_CUDA_VERSION} REQUIRED)
+ # later command will make use of the value in tensorflow_CUDA_VERSION
+ find_package(CUDA ${tensorflow_CUDA_VERSION} REQUIRED EXACT)
+
+ # Test compatibility of compiler on CUDA
+ try_compile(CUDA_TEST_COMPILE_C
+ ${CMAKE_CURRENT_BINARY_DIR}/tests/cuda
+ ${CMAKE_CURRENT_SOURCE_DIR}/tests/cuda/compatibility_test.c
+ CMAKE_FLAGS -DINCLUDE_DIRECTORIES=${CUDA_INCLUDE_DIRS})
+ try_compile(CUDA_TEST_COMPILE_CXX
+ ${CMAKE_CURRENT_BINARY_DIR}/tests/cuda
+ ${CMAKE_CURRENT_SOURCE_DIR}/tests/cuda/compatibility_test.cc
+ CMAKE_FLAGS -DINCLUDE_DIRECTORIES=${CUDA_INCLUDE_DIRS})
+ if(NOT (CUDA_TEST_COMPILE_C AND CUDA_TEST_COMPILE_CXX))
+ message(FATAL_ERROR "Selected compiler (or version) is not supported for CUDA")
+ endif()
# by default we assume compute cabability 3.5 and 5.2. If you change this change it in
# CUDA_NVCC_FLAGS and cuda_config.h below
diff --git a/tensorflow/contrib/cmake/external/boringssl.cmake b/tensorflow/contrib/cmake/external/boringssl.cmake
index 5ad477fdff..3c4bb01e24 100644
--- a/tensorflow/contrib/cmake/external/boringssl.cmake
+++ b/tensorflow/contrib/cmake/external/boringssl.cmake
@@ -37,6 +37,7 @@ ExternalProject_Add(boringssl
GIT_TAG ${boringssl_TAG}
DOWNLOAD_DIR "${DOWNLOAD_LOCATION}"
# BUILD_IN_SOURCE 1
+ BUILD_BYPRODUCTS ${boringssl_STATIC_LIBRARIES}
INSTALL_COMMAND ""
CMAKE_CACHE_ARGS
-DCMAKE_POSITION_INDEPENDENT_CODE:BOOL=${tensorflow_ENABLE_POSITION_INDEPENDENT_CODE}
diff --git a/tensorflow/contrib/cmake/external/farmhash.cmake b/tensorflow/contrib/cmake/external/farmhash.cmake
index 0cd0c1030c..d51569bc21 100644
--- a/tensorflow/contrib/cmake/external/farmhash.cmake
+++ b/tensorflow/contrib/cmake/external/farmhash.cmake
@@ -33,6 +33,7 @@ if(WIN32)
URL_HASH ${farmhash_HASH}
DOWNLOAD_DIR "${DOWNLOAD_LOCATION}"
BUILD_IN_SOURCE 1
+ BUILD_BYPRODUCTS ${farmhash_STATIC_LIBRARIES}
PATCH_COMMAND ${CMAKE_COMMAND} -E copy_if_different ${CMAKE_CURRENT_SOURCE_DIR}/patches/farmhash/CMakeLists.txt ${farmhash_BUILD}
INSTALL_DIR ${farmhash_INSTALL}
CMAKE_CACHE_ARGS
diff --git a/tensorflow/contrib/cmake/external/fft2d.cmake b/tensorflow/contrib/cmake/external/fft2d.cmake
index d3af2a4676..a7bc50d5bc 100644
--- a/tensorflow/contrib/cmake/external/fft2d.cmake
+++ b/tensorflow/contrib/cmake/external/fft2d.cmake
@@ -29,6 +29,7 @@ if(WIN32)
URL_HASH ${fft2d_HASH}
DOWNLOAD_DIR "${DOWNLOAD_LOCATION}"
BUILD_IN_SOURCE 1
+ BUILD_BYPRODUCTS ${fft2d_STATIC_LIBRARIES}
PATCH_COMMAND ${CMAKE_COMMAND} -E copy_if_different ${CMAKE_CURRENT_SOURCE_DIR}/patches/fft2d/CMakeLists.txt ${fft2d_BUILD}/src/fft2d/CMakeLists.txt
INSTALL_DIR ${fft2d_INSTALL}
CMAKE_CACHE_ARGS
diff --git a/tensorflow/contrib/cmake/external/gif.cmake b/tensorflow/contrib/cmake/external/gif.cmake
index 3d53c51fff..e1f8d13f8e 100644
--- a/tensorflow/contrib/cmake/external/gif.cmake
+++ b/tensorflow/contrib/cmake/external/gif.cmake
@@ -33,6 +33,7 @@ if(WIN32)
PREFIX gif
URL ${gif_URL}
URL_HASH ${gif_HASH}
+ BUILD_BYPRODUCTS ${gif_STATIC_LIBRARIES}
PATCH_COMMAND ${CMAKE_COMMAND} -E copy_if_different ${CMAKE_SOURCE_DIR}/patches/gif/CMakeLists.txt ${gif_BUILD}
INSTALL_DIR ${gif_INSTALL}
DOWNLOAD_DIR "${DOWNLOAD_LOCATION}"
diff --git a/tensorflow/contrib/cmake/external/googletest.cmake b/tensorflow/contrib/cmake/external/googletest.cmake
index d09bb02890..7cc5ae6390 100644
--- a/tensorflow/contrib/cmake/external/googletest.cmake
+++ b/tensorflow/contrib/cmake/external/googletest.cmake
@@ -20,8 +20,13 @@ set(googletest_BUILD ${CMAKE_CURRENT_BINARY_DIR}/googletest/)
set(googletest_TAG ec44c6c1675c25b9827aacd08c02433cccde7780)
if(WIN32)
- set(googletest_STATIC_LIBRARIES
- ${CMAKE_CURRENT_BINARY_DIR}/googletest/src/googletest/googletest/$(Configuration)/gtest.lib)
+ if(${CMAKE_GENERATOR} MATCHES "Visual Studio.*")
+ set(googletest_STATIC_LIBRARIES
+ ${CMAKE_CURRENT_BINARY_DIR}/googletest/src/googletest/googletest/$(Configuration)/gtest.lib)
+ else()
+ set(googletest_STATIC_LIBRARIES
+ ${CMAKE_CURRENT_BINARY_DIR}/googletest/src/googletest/googletest/gtest.lib)
+ endif()
else()
set(googletest_STATIC_LIBRARIES
${CMAKE_CURRENT_BINARY_DIR}/googletest/src/googletest/googletest/${CMAKE_BUILD_TYPE}/gtest.a)
@@ -33,6 +38,7 @@ ExternalProject_Add(googletest
GIT_TAG ${googletest_TAG}
DOWNLOAD_DIR "${DOWNLOAD_LOCATION}"
BUILD_IN_SOURCE 1
+ BUILD_BYPRODUCTS ${googletest_STATIC_LIBRARIES}
#PATCH_COMMAND ${CMAKE_COMMAND} -E copy_if_different ${CMAKE_SOURCE_DIR}/patches/grpc/CMakeLists.txt ${GRPC_BUILD}
INSTALL_COMMAND ""
CMAKE_CACHE_ARGS
diff --git a/tensorflow/contrib/cmake/external/grpc.cmake b/tensorflow/contrib/cmake/external/grpc.cmake
index 28adb4fe84..a9f43a3ecb 100644
--- a/tensorflow/contrib/cmake/external/grpc.cmake
+++ b/tensorflow/contrib/cmake/external/grpc.cmake
@@ -20,10 +20,17 @@ set(GRPC_BUILD ${CMAKE_CURRENT_BINARY_DIR}/grpc/src/grpc)
set(GRPC_TAG 730b778632e79cc3c96ad237f282d687ee325ce7)
if(WIN32)
- set(grpc_STATIC_LIBRARIES
- ${CMAKE_CURRENT_BINARY_DIR}/grpc/src/grpc/Release/grpc++_unsecure.lib
- ${CMAKE_CURRENT_BINARY_DIR}/grpc/src/grpc/Release/grpc_unsecure.lib
- ${CMAKE_CURRENT_BINARY_DIR}/grpc/src/grpc/Release/gpr.lib)
+ if(${CMAKE_GENERATOR} MATCHES "Visual Studio.*")
+ set(grpc_STATIC_LIBRARIES
+ ${CMAKE_CURRENT_BINARY_DIR}/grpc/src/grpc/Release/grpc++_unsecure.lib
+ ${CMAKE_CURRENT_BINARY_DIR}/grpc/src/grpc/Release/grpc_unsecure.lib
+ ${CMAKE_CURRENT_BINARY_DIR}/grpc/src/grpc/Release/gpr.lib)
+ else()
+ set(grpc_STATIC_LIBRARIES
+ ${CMAKE_CURRENT_BINARY_DIR}/grpc/src/grpc/grpc++_unsecure.lib
+ ${CMAKE_CURRENT_BINARY_DIR}/grpc/src/grpc/grpc_unsecure.lib
+ ${CMAKE_CURRENT_BINARY_DIR}/grpc/src/grpc/gpr.lib)
+ endif()
else()
set(grpc_STATIC_LIBRARIES
${CMAKE_CURRENT_BINARY_DIR}/grpc/src/grpc/libgrpc++_unsecure.a
@@ -40,6 +47,7 @@ ExternalProject_Add(grpc
GIT_TAG ${GRPC_TAG}
DOWNLOAD_DIR "${DOWNLOAD_LOCATION}"
BUILD_IN_SOURCE 1
+ BUILD_BYPRODUCTS ${grpc_STATIC_LIBRARIES}
BUILD_COMMAND ${CMAKE_COMMAND} --build . --config Release --target grpc++_unsecure
COMMAND ${CMAKE_COMMAND} --build . --config Release --target grpc_cpp_plugin
INSTALL_COMMAND ""
diff --git a/tensorflow/contrib/cmake/external/highwayhash.cmake b/tensorflow/contrib/cmake/external/highwayhash.cmake
index 2c23bef8a3..a6e8a38d8c 100644
--- a/tensorflow/contrib/cmake/external/highwayhash.cmake
+++ b/tensorflow/contrib/cmake/external/highwayhash.cmake
@@ -42,6 +42,7 @@ ExternalProject_Add(highwayhash
GIT_TAG ${highwayhash_TAG}
DOWNLOAD_DIR "${DOWNLOAD_LOCATION}"
BUILD_IN_SOURCE 1
+ BUILD_BYPRODUCTS ${highwayhash_STATIC_LIBRARIES}
PATCH_COMMAND ${CMAKE_COMMAND} -E copy_if_different ${CMAKE_CURRENT_SOURCE_DIR}/patches/highwayhash/CMakeLists.txt ${highwayhash_BUILD}
INSTALL_DIR ${highwayhash_INSTALL}
CMAKE_CACHE_ARGS
diff --git a/tensorflow/contrib/cmake/external/jemalloc.cmake b/tensorflow/contrib/cmake/external/jemalloc.cmake
index 198ba13e64..afadcc007d 100644
--- a/tensorflow/contrib/cmake/external/jemalloc.cmake
+++ b/tensorflow/contrib/cmake/external/jemalloc.cmake
@@ -24,8 +24,11 @@ if (WIN32)
${jemalloc_INCLUDE_DIRS}
${CMAKE_CURRENT_BINARY_DIR}/jemalloc/src/jemalloc/include/msvc_compat
)
- set(jemalloc_ADDITIONAL_CMAKE_OPTIONS -A x64)
- set(jemalloc_STATIC_LIBRARIES ${jemalloc_BUILD}/Release/jemalloc.lib)
+ if(${CMAKE_GENERATOR} MATCHES "Visual Studio.*")
+ set(jemalloc_STATIC_LIBRARIES ${jemalloc_BUILD}/Release/jemalloc.lib)
+ else()
+ set(jemalloc_STATIC_LIBRARIES ${jemalloc_BUILD}/jemalloc.lib)
+ endif()
else()
set(jemalloc_STATIC_LIBRARIES ${jemalloc_BUILD}/Release/jemalloc.a)
endif()
@@ -36,12 +39,12 @@ ExternalProject_Add(jemalloc
URL_HASH ${jemalloc_HASH}
DOWNLOAD_DIR "${DOWNLOAD_LOCATION}"
BUILD_IN_SOURCE 1
- CONFIGURE_COMMAND ${CMAKE_COMMAND}
+ BUILD_BYPRODUCTS ${jemalloc_STATIC_LIBRARIES}
+ BUILD_COMMAND ${CMAKE_COMMAND} --build . --config Release --target jemalloc
+ INSTALL_COMMAND ${CMAKE_COMMAND} -E echo "Skipping install step."
+ CMAKE_CACHE_ARGS
-DCMAKE_BUILD_TYPE:STRING=Release
-DCMAKE_VERBOSE_MAKEFILE:BOOL=OFF
-Dwith-jemalloc-prefix:STRING=jemalloc_
-Dwithout-export:BOOL=ON
- ${jemalloc_ADDITIONAL_CMAKE_OPTIONS}
- BUILD_COMMAND ${CMAKE_COMMAND} --build . --config Release --target jemalloc
- INSTALL_COMMAND ${CMAKE_COMMAND} -E echo "Skipping install step."
)
diff --git a/tensorflow/contrib/cmake/external/jpeg.cmake b/tensorflow/contrib/cmake/external/jpeg.cmake
index d9a165e856..c1c5842aa4 100644
--- a/tensorflow/contrib/cmake/external/jpeg.cmake
+++ b/tensorflow/contrib/cmake/external/jpeg.cmake
@@ -46,6 +46,7 @@ if (WIN32)
PREFIX jpeg
URL ${jpeg_URL}
URL_HASH ${jpeg_HASH}
+ BUILD_BYPRODUCTS ${jpeg_STATIC_LIBRARIES}
PATCH_COMMAND ${CMAKE_COMMAND} -E copy_if_different ${CMAKE_CURRENT_SOURCE_DIR}/patches/jpeg/CMakeLists.txt ${jpeg_BUILD}
INSTALL_DIR ${jpeg_INSTALL}
DOWNLOAD_DIR "${DOWNLOAD_LOCATION}"
diff --git a/tensorflow/contrib/cmake/external/jsoncpp.cmake b/tensorflow/contrib/cmake/external/jsoncpp.cmake
index 861201f97e..84c52e3652 100644
--- a/tensorflow/contrib/cmake/external/jsoncpp.cmake
+++ b/tensorflow/contrib/cmake/external/jsoncpp.cmake
@@ -23,7 +23,11 @@ set(jsoncpp_LIBRARIES ${jsoncpp_BUILD}/obj/so/libjsoncpp.so)
set(jsoncpp_INCLUDES ${jsoncpp_BUILD})
if(WIN32)
- set(jsoncpp_STATIC_LIBRARIES ${jsoncpp_BUILD}/$(Configuration)/jsoncpp.lib)
+ if(${CMAKE_GENERATOR} MATCHES "Visual Studio.*")
+ set(jsoncpp_STATIC_LIBRARIES ${jsoncpp_BUILD}/$(Configuration)/jsoncpp.lib)
+ else()
+ set(jsoncpp_STATIC_LIBRARIES ${jsoncpp_BUILD}/jsoncpp.lib)
+ endif()
else()
set(jsoncpp_STATIC_LIBRARIES ${jsoncpp_BUILD}/libjsoncpp.a)
endif()
@@ -40,6 +44,7 @@ ExternalProject_Add(jsoncpp
GIT_TAG ${jsoncpp_TAG}
DOWNLOAD_DIR "${DOWNLOAD_LOCATION}"
BUILD_IN_SOURCE 1
+ BUILD_BYPRODUCTS ${jsoncpp_STATIC_LIBRARIES}
INSTALL_COMMAND ""
CMAKE_CACHE_ARGS
-DCMAKE_POSITION_INDEPENDENT_CODE:BOOL=${tensorflow_ENABLE_POSITION_INDEPENDENT_CODE}
diff --git a/tensorflow/contrib/cmake/external/lmdb.cmake b/tensorflow/contrib/cmake/external/lmdb.cmake
index 41b314e285..ed5ab788ac 100644
--- a/tensorflow/contrib/cmake/external/lmdb.cmake
+++ b/tensorflow/contrib/cmake/external/lmdb.cmake
@@ -20,10 +20,17 @@ set(lmdb_HASH SHA256=108532fb94c6f227558d45be3f3347b52539f0f58290a7bb31ec06c462d
set(lmdb_BUILD ${CMAKE_BINARY_DIR}/lmdb/src/lmdb)
set(lmdb_INSTALL ${CMAKE_BINARY_DIR}/lmdb/install)
+if(WIN32)
+ set(lmdb_STATIC_LIBRARIES ${lmdb_INSTALL}/lib/lmdb.lib)
+else()
+ set(lmdb_STATIC_LIBRARIES ${lmdb_INSTALL}/lib/liblmdb.a)
+endif()
+
ExternalProject_Add(lmdb
PREFIX lmdb
URL ${lmdb_URL}
URL_HASH ${lmdb_HASH}
+ BUILD_BYPRODUCTS ${lmdb_STATIC_LIBRARIES}
PATCH_COMMAND ${CMAKE_COMMAND} -E copy_if_different
${CMAKE_CURRENT_SOURCE_DIR}/patches/lmdb/CMakeLists.txt ${lmdb_BUILD}
INSTALL_DIR ${lmdb_INSTALL}
@@ -35,12 +42,6 @@ ExternalProject_Add(lmdb
-DCMAKE_INSTALL_PREFIX:STRING=${lmdb_INSTALL}
)
-if(WIN32)
- set(lmdb_STATIC_LIBRARIES ${lmdb_INSTALL}/lib/lmdb.lib)
-else()
- set(lmdb_STATIC_LIBRARIES ${lmdb_INSTALL}/lib/liblmdb.a)
-endif()
-
set(lmdb_HEADERS
"${lmdb_INSTALL}/include/lmdb.h"
"${lmdb_INSTALL}/include/midl.h"
diff --git a/tensorflow/contrib/cmake/external/nsync.cmake b/tensorflow/contrib/cmake/external/nsync.cmake
index 0508006047..f3a37ff508 100644
--- a/tensorflow/contrib/cmake/external/nsync.cmake
+++ b/tensorflow/contrib/cmake/external/nsync.cmake
@@ -42,6 +42,7 @@ ExternalProject_Add(nsync
GIT_TAG ${nsync_TAG}
DOWNLOAD_DIR "${DOWNLOAD_LOCATION}"
BUILD_IN_SOURCE 1
+ BUILD_BYPRODUCTS ${nsync_STATIC_LIBRARIES}
PATCH_COMMAND ${CMAKE_COMMAND} -E copy_if_different ${CMAKE_CURRENT_SOURCE_DIR}/patches/nsync/CMakeLists.txt ${nsync_BUILD}
INSTALL_DIR ${nsync_INSTALL}
CMAKE_CACHE_ARGS
diff --git a/tensorflow/contrib/cmake/external/png.cmake b/tensorflow/contrib/cmake/external/png.cmake
index b277be5690..6cd66a6599 100644
--- a/tensorflow/contrib/cmake/external/png.cmake
+++ b/tensorflow/contrib/cmake/external/png.cmake
@@ -21,9 +21,19 @@ set(png_BUILD ${CMAKE_BINARY_DIR}/png/src/png)
set(png_INSTALL ${CMAKE_BINARY_DIR}/png/install)
if(WIN32)
- set(png_STATIC_LIBRARIES
- debug ${CMAKE_BINARY_DIR}/png/install/lib/libpng12_staticd.lib
- optimized ${CMAKE_BINARY_DIR}/png/install/lib/libpng12_static.lib)
+ if(${CMAKE_GENERATOR} MATCHES "Visual Studio.*")
+ set(png_STATIC_LIBRARIES
+ debug ${CMAKE_BINARY_DIR}/png/install/lib/libpng12_staticd.lib
+ optimized ${CMAKE_BINARY_DIR}/png/install/lib/libpng12_static.lib)
+ else()
+ if(CMAKE_BUILD_TYPE EQUAL Debug)
+ set(png_STATIC_LIBRARIES
+ ${CMAKE_BINARY_DIR}/png/install/lib/libpng12_staticd.lib)
+ else()
+ set(png_STATIC_LIBRARIES
+ ${CMAKE_BINARY_DIR}/png/install/lib/libpng12_static.lib)
+ endif()
+ endif()
else()
set(png_STATIC_LIBRARIES ${CMAKE_BINARY_DIR}/png/install/lib/libpng12.a)
endif()
@@ -38,6 +48,7 @@ ExternalProject_Add(png
DEPENDS zlib
URL ${png_URL}
URL_HASH ${png_HASH}
+ BUILD_BYPRODUCTS ${png_STATIC_LIBRARIES}
INSTALL_DIR ${png_INSTALL}
DOWNLOAD_DIR "${DOWNLOAD_LOCATION}"
CMAKE_CACHE_ARGS
diff --git a/tensorflow/contrib/cmake/external/protobuf.cmake b/tensorflow/contrib/cmake/external/protobuf.cmake
index fd05fa6d47..aba8a5244e 100644
--- a/tensorflow/contrib/cmake/external/protobuf.cmake
+++ b/tensorflow/contrib/cmake/external/protobuf.cmake
@@ -19,11 +19,34 @@ set(PROTOBUF_URL https://github.com/google/protobuf.git)
set(PROTOBUF_TAG 396336eb961b75f03b25824fe86cf6490fb75e3a)
if(WIN32)
- set(protobuf_STATIC_LIBRARIES
- debug ${CMAKE_CURRENT_BINARY_DIR}/protobuf/src/protobuf/$(Configuration)/libprotobufd.lib
- optimized ${CMAKE_CURRENT_BINARY_DIR}/protobuf/src/protobuf/$(Configuration)/libprotobuf.lib)
- set(PROTOBUF_PROTOC_EXECUTABLE ${CMAKE_CURRENT_BINARY_DIR}/protobuf/src/protobuf/$(Configuration)/protoc.exe)
- set(PROTOBUF_ADDITIONAL_CMAKE_OPTIONS -Dprotobuf_MSVC_STATIC_RUNTIME:BOOL=OFF -A x64)
+ if(${CMAKE_GENERATOR} MATCHES "Visual Studio.*")
+ set(protobuf_STATIC_LIBRARIES
+ debug ${CMAKE_CURRENT_BINARY_DIR}/protobuf/src/protobuf/$(Configuration)/libprotobufd.lib
+ optimized ${CMAKE_CURRENT_BINARY_DIR}/protobuf/src/protobuf/$(Configuration)/libprotobuf.lib)
+ set(PROTOBUF_PROTOC_EXECUTABLE ${CMAKE_CURRENT_BINARY_DIR}/protobuf/src/protobuf/$(Configuration)/protoc.exe)
+ else()
+ if(CMAKE_BUILD_TYPE EQUAL Debug)
+ set(protobuf_STATIC_LIBRARIES
+ ${CMAKE_CURRENT_BINARY_DIR}/protobuf/src/protobuf/libprotobufd.lib)
+ else()
+ set(protobuf_STATIC_LIBRARIES
+ ${CMAKE_CURRENT_BINARY_DIR}/protobuf/src/protobuf/libprotobuf.lib)
+ endif()
+ set(PROTOBUF_PROTOC_EXECUTABLE ${CMAKE_CURRENT_BINARY_DIR}/protobuf/src/protobuf/protoc.exe)
+ endif()
+
+ # This section is to make sure CONFIGURE_COMMAND use the same generator settings
+ set(PROTOBUF_GENERATOR_PLATFORM)
+ if (CMAKE_GENERATOR_PLATFORM)
+ set(PROTOBUF_GENERATOR_PLATFORM -A ${CMAKE_GENERATOR_PLATFORM})
+ endif()
+ set(PROTOBUF_GENERATOR_TOOLSET)
+ if (CMAKE_GENERATOR_TOOLSET)
+ set(PROTOBUF_GENERATOR_TOOLSET -T ${CMAKE_GENERATOR_TOOLSET})
+ endif()
+ set(PROTOBUF_ADDITIONAL_CMAKE_OPTIONS -Dprotobuf_MSVC_STATIC_RUNTIME:BOOL=OFF
+ -G${CMAKE_GENERATOR} ${PROTOBUF_GENERATOR_PLATFORM} ${PROTOBUF_GENERATOR_TOOLSET})
+ # End of section
else()
set(protobuf_STATIC_LIBRARIES ${CMAKE_CURRENT_BINARY_DIR}/protobuf/src/protobuf/libprotobuf.a)
set(PROTOBUF_PROTOC_EXECUTABLE ${CMAKE_CURRENT_BINARY_DIR}/protobuf/src/protobuf/protoc)
@@ -36,10 +59,15 @@ ExternalProject_Add(protobuf
GIT_TAG ${PROTOBUF_TAG}
DOWNLOAD_DIR "${DOWNLOAD_LOCATION}"
BUILD_IN_SOURCE 1
+ BUILD_BYPRODUCTS ${PROTOBUF_PROTOC_EXECUTABLE} ${protobuf_STATIC_LIBRARIES}
SOURCE_DIR ${CMAKE_CURRENT_BINARY_DIR}/protobuf/src/protobuf
+ # SOURCE_SUBDIR cmake/ # Requires CMake 3.7, this will allow removal of CONFIGURE_COMMAND
+ # CONFIGURE_COMMAND resets some settings made in CMAKE_CACHE_ARGS and the generator used
CONFIGURE_COMMAND ${CMAKE_COMMAND} cmake/
- -Dprotobuf_BUILD_TESTS=OFF
- -DCMAKE_POSITION_INDEPENDENT_CODE=ON
+ -DCMAKE_POSITION_INDEPENDENT_CODE:BOOL=${tensorflow_ENABLE_POSITION_INDEPENDENT_CODE}
+ -DCMAKE_BUILD_TYPE:STRING=Release
+ -DCMAKE_VERBOSE_MAKEFILE:BOOL=OFF
+ -Dprotobuf_BUILD_TESTS:BOOL=OFF
-DZLIB_ROOT=${ZLIB_INSTALL}
${PROTOBUF_ADDITIONAL_CMAKE_OPTIONS}
INSTALL_COMMAND ""
@@ -47,5 +75,7 @@ ExternalProject_Add(protobuf
-DCMAKE_POSITION_INDEPENDENT_CODE:BOOL=${tensorflow_ENABLE_POSITION_INDEPENDENT_CODE}
-DCMAKE_BUILD_TYPE:STRING=Release
-DCMAKE_VERBOSE_MAKEFILE:BOOL=OFF
+ -Dprotobuf_BUILD_TESTS:BOOL=OFF
+ -Dprotobuf_MSVC_STATIC_RUNTIME:BOOL=OFF
-DZLIB_ROOT:STRING=${ZLIB_INSTALL}
)
diff --git a/tensorflow/contrib/cmake/external/re2.cmake b/tensorflow/contrib/cmake/external/re2.cmake
index 371d8447f9..c4bc0b1707 100644
--- a/tensorflow/contrib/cmake/external/re2.cmake
+++ b/tensorflow/contrib/cmake/external/re2.cmake
@@ -21,7 +21,11 @@ set(re2_INSTALL ${CMAKE_CURRENT_BINARY_DIR}/re2/install)
set(re2_TAG e7efc48)
if(WIN32)
- set(re2_STATIC_LIBRARIES ${re2_BUILD}/$(Configuration)/re2.lib)
+ if(${CMAKE_GENERATOR} MATCHES "Visual Studio.*")
+ set(re2_STATIC_LIBRARIES ${re2_BUILD}/$(Configuration)/re2.lib)
+ else()
+ set(re2_STATIC_LIBRARIES ${re2_BUILD}/re2.lib)
+ endif()
else()
set(re2_STATIC_LIBRARIES ${re2_BUILD}/libre2.a)
endif()
@@ -36,6 +40,7 @@ ExternalProject_Add(re2
GIT_TAG ${re2_TAG}
INSTALL_DIR ${re2_INSTALL}
BUILD_IN_SOURCE 1
+ BUILD_BYPRODUCTS ${re2_STATIC_LIBRARIES}
DOWNLOAD_DIR "${DOWNLOAD_LOCATION}"
CMAKE_CACHE_ARGS
-DCMAKE_POSITION_INDEPENDENT_CODE:BOOL=${tensorflow_ENABLE_POSITION_INDEPENDENT_CODE}
diff --git a/tensorflow/contrib/cmake/external/snappy.cmake b/tensorflow/contrib/cmake/external/snappy.cmake
index fd57734298..f54197643b 100644
--- a/tensorflow/contrib/cmake/external/snappy.cmake
+++ b/tensorflow/contrib/cmake/external/snappy.cmake
@@ -20,7 +20,11 @@ set(snappy_BUILD ${CMAKE_CURRENT_BINARY_DIR}/snappy/src/snappy)
set(snappy_INCLUDE_DIR ${CMAKE_CURRENT_BINARY_DIR}/snappy/src/snappy)
if(WIN32)
- set(snappy_STATIC_LIBRARIES ${snappy_BUILD}/$(Configuration)/snappy.lib)
+ if(${CMAKE_GENERATOR} MATCHES "Visual Studio.*")
+ set(snappy_STATIC_LIBRARIES ${snappy_BUILD}/$(Configuration)/snappy.lib)
+ else()
+ set(snappy_STATIC_LIBRARIES ${snappy_BUILD}/snappy.lib)
+ endif()
else()
set(snappy_STATIC_LIBRARIES ${snappy_BUILD}/libsnappy.a)
endif()
@@ -35,6 +39,7 @@ ExternalProject_Add(snappy
GIT_TAG ${snappy_TAG}
DOWNLOAD_DIR "${DOWNLOAD_LOCATION}"
BUILD_IN_SOURCE 1
+ BUILD_BYPRODUCTS ${snappy_STATIC_LIBRARIES}
INSTALL_COMMAND ""
LOG_DOWNLOAD ON
LOG_CONFIGURE ON
diff --git a/tensorflow/contrib/cmake/external/sqlite.cmake b/tensorflow/contrib/cmake/external/sqlite.cmake
index 8297c60712..57c4ae7651 100644
--- a/tensorflow/contrib/cmake/external/sqlite.cmake
+++ b/tensorflow/contrib/cmake/external/sqlite.cmake
@@ -36,6 +36,7 @@ if (WIN32)
PREFIX sqlite
URL ${sqlite_URL}
URL_HASH ${sqlite_HASH}
+ BUILD_BYPRODUCTS ${sqlite_STATIC_LIBRARIES}
PATCH_COMMAND ${CMAKE_COMMAND} -E copy_if_different ${CMAKE_CURRENT_SOURCE_DIR}/patches/sqlite/CMakeLists.txt ${sqlite_BUILD}
INSTALL_DIR ${sqlite_INSTALL}
DOWNLOAD_DIR "${DOWNLOAD_LOCATION}"
diff --git a/tensorflow/contrib/cmake/external/zlib.cmake b/tensorflow/contrib/cmake/external/zlib.cmake
index 5bec14fb00..c5eb0cbcc7 100644
--- a/tensorflow/contrib/cmake/external/zlib.cmake
+++ b/tensorflow/contrib/cmake/external/zlib.cmake
@@ -21,9 +21,19 @@ set(ZLIB_INSTALL ${CMAKE_CURRENT_BINARY_DIR}/zlib/install)
set(ZLIB_TAG 50893291621658f355bc5b4d450a8d06a563053d)
if(WIN32)
- set(zlib_STATIC_LIBRARIES
- debug ${CMAKE_CURRENT_BINARY_DIR}/zlib/install/lib/zlibstaticd.lib
- optimized ${CMAKE_CURRENT_BINARY_DIR}/zlib/install/lib/zlibstatic.lib)
+ if(${CMAKE_GENERATOR} MATCHES "Visual Studio.*")
+ set(zlib_STATIC_LIBRARIES
+ debug ${CMAKE_CURRENT_BINARY_DIR}/zlib/install/lib/zlibstaticd.lib
+ optimized ${CMAKE_CURRENT_BINARY_DIR}/zlib/install/lib/zlibstatic.lib)
+ else()
+ if(CMAKE_BUILD_TYPE EQUAL Debug)
+ set(zlib_STATIC_LIBRARIES
+ ${CMAKE_CURRENT_BINARY_DIR}/zlib/install/lib/zlibstaticd.lib)
+ else()
+ set(zlib_STATIC_LIBRARIES
+ ${CMAKE_CURRENT_BINARY_DIR}/zlib/install/lib/zlibstatic.lib)
+ endif()
+ endif()
else()
set(zlib_STATIC_LIBRARIES
${CMAKE_CURRENT_BINARY_DIR}/zlib/install/lib/libz.a)
@@ -40,6 +50,7 @@ ExternalProject_Add(zlib
GIT_TAG ${ZLIB_TAG}
INSTALL_DIR ${ZLIB_INSTALL}
BUILD_IN_SOURCE 1
+ BUILD_BYPRODUCTS ${zlib_STATIC_LIBRARIES}
DOWNLOAD_DIR "${DOWNLOAD_LOCATION}"
CMAKE_CACHE_ARGS
-DCMAKE_POSITION_INDEPENDENT_CODE:BOOL=${tensorflow_ENABLE_POSITION_INDEPENDENT_CODE}
diff --git a/tensorflow/contrib/cmake/python_modules.txt b/tensorflow/contrib/cmake/python_modules.txt
index 2720c43b78..f55043c93d 100644
--- a/tensorflow/contrib/cmake/python_modules.txt
+++ b/tensorflow/contrib/cmake/python_modules.txt
@@ -299,7 +299,9 @@ tensorflow/contrib/linear_optimizer/kernels/g3doc
tensorflow/contrib/linear_optimizer/python
tensorflow/contrib/linear_optimizer/python/ops
# TODO(drpngx): Fix failing imports
+# tensorflow/contrib/lite
# tensorflow/contrib/lite/python
+# tensorflow/contrib/lite/toco
# tensorflow/contrib/lite/toco/python
tensorflow/contrib/lookup
tensorflow/contrib/losses
@@ -360,6 +362,7 @@ tensorflow/contrib/reduce_slice_ops/kernels
tensorflow/contrib/reduce_slice_ops/ops
tensorflow/contrib/reduce_slice_ops/python
tensorflow/contrib/reduce_slice_ops/python/ops
+tensorflow/contrib/remote_fused_graph
tensorflow/contrib/remote_fused_graph/pylib
tensorflow/contrib/remote_fused_graph/pylib/python
tensorflow/contrib/remote_fused_graph/pylib/python/ops
@@ -409,6 +412,7 @@ tensorflow/contrib/summary
tensorflow/contrib/tensorboard
tensorflow/contrib/tensorboard/plugins
tensorflow/contrib/tensorboard/plugins/projector
+tensorflow/contrib/tensorboard/plugins/trace
tensorflow/contrib/tensor_forest
tensorflow/contrib/tensor_forest/client
tensorflow/contrib/tensor_forest/hybrid
@@ -419,6 +423,7 @@ tensorflow/contrib/tensor_forest/hybrid/python/layers
tensorflow/contrib/tensor_forest/hybrid/python/models
tensorflow/contrib/tensor_forest/hybrid/python/ops
tensorflow/contrib/tensor_forest/kernels
+tensorflow/contrib/tensor_forest/proto
tensorflow/contrib/tensor_forest/python
tensorflow/contrib/tensor_forest/python/ops
tensorflow/contrib/testing
@@ -439,6 +444,7 @@ tensorflow/contrib/timeseries/python/timeseries/state_space_models
tensorflow/contrib/tpu
tensorflow/contrib/tpu/ops
tensorflow/contrib/tpu/profiler
+tensorflow/contrib/tpu/proto
tensorflow/contrib/tpu/python
tensorflow/contrib/tpu/python/ops
tensorflow/contrib/tpu/python/profiler
diff --git a/tensorflow/contrib/cmake/tests/cuda/compatibility_test.c b/tensorflow/contrib/cmake/tests/cuda/compatibility_test.c
new file mode 100644
index 0000000000..9e355da33a
--- /dev/null
+++ b/tensorflow/contrib/cmake/tests/cuda/compatibility_test.c
@@ -0,0 +1,22 @@
+/* Copyright 2018 The TensorFlow Authors. All Rights Reserved.
+
+Licensed under the Apache License, Version 2.0 (the "License");
+you may not use this file except in compliance with the License.
+You may obtain a copy of the License at
+
+ http://www.apache.org/licenses/LICENSE-2.0
+
+Unless required by applicable law or agreed to in writing, software
+distributed under the License is distributed on an "AS IS" BASIS,
+WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+See the License for the specific language governing permissions and
+limitations under the License.
+==============================================================================*/
+
+// This is a program to test if compiler is compatible with CUDA.
+#define __CUDACC__
+#include "crt/host_config.h"
+
+int main(void) {
+ return 0;
+}
diff --git a/tensorflow/contrib/cmake/tests/cuda/compatibility_test.cc b/tensorflow/contrib/cmake/tests/cuda/compatibility_test.cc
new file mode 100644
index 0000000000..a50461cafd
--- /dev/null
+++ b/tensorflow/contrib/cmake/tests/cuda/compatibility_test.cc
@@ -0,0 +1,20 @@
+/* Copyright 2018 The TensorFlow Authors. All Rights Reserved.
+
+Licensed under the Apache License, Version 2.0 (the "License");
+you may not use this file except in compliance with the License.
+You may obtain a copy of the License at
+
+ http://www.apache.org/licenses/LICENSE-2.0
+
+Unless required by applicable law or agreed to in writing, software
+distributed under the License is distributed on an "AS IS" BASIS,
+WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+See the License for the specific language governing permissions and
+limitations under the License.
+============================================================================*/
+
+// This is a program to test if compiler is compatible with CUDA.
+#define __CUDACC__
+#include "crt/host_config.h"
+
+int main(void) { return 0; }
diff --git a/tensorflow/contrib/cmake/tf_cc_ops.cmake b/tensorflow/contrib/cmake/tf_cc_ops.cmake
index f3cf3e7044..f73da0b8ab 100644
--- a/tensorflow/contrib/cmake/tf_cc_ops.cmake
+++ b/tensorflow/contrib/cmake/tf_cc_ops.cmake
@@ -149,7 +149,11 @@ add_library(tf_cc OBJECT ${tf_cc_srcs})
add_dependencies(tf_cc tf_cc_framework tf_cc_ops)
if (WIN32)
- set (pywrap_tensorflow_lib "${CMAKE_CURRENT_BINARY_DIR}/${CMAKE_BUILD_TYPE}/pywrap_tensorflow_internal.lib")
+ if(${CMAKE_GENERATOR} MATCHES "Visual Studio.*")
+ set (pywrap_tensorflow_lib "${CMAKE_CURRENT_BINARY_DIR}/${CMAKE_BUILD_TYPE}/pywrap_tensorflow_internal.lib")
+ else()
+ set (pywrap_tensorflow_lib "${CMAKE_CURRENT_BINARY_DIR}/pywrap_tensorflow_internal.lib")
+ endif()
else (WIN32)
set (pywrap_tensorflow_lib "${CMAKE_CURRENT_BINARY_DIR}/libpywrap_tensorflow_internal.so")
endif (WIN32)
diff --git a/tensorflow/contrib/cmake/tf_python.cmake b/tensorflow/contrib/cmake/tf_python.cmake
index b3c6663b8b..b730ebd3ba 100755
--- a/tensorflow/contrib/cmake/tf_python.cmake
+++ b/tensorflow/contrib/cmake/tf_python.cmake
@@ -541,7 +541,11 @@ if(WIN32)
${nsync_STATIC_LIBRARIES}
)
- set(pywrap_tensorflow_deffile "${CMAKE_CURRENT_BINARY_DIR}/${CMAKE_BUILD_TYPE}/pywrap_tensorflow.def")
+ if(${CMAKE_GENERATOR} MATCHES "Visual Studio.*")
+ set(pywrap_tensorflow_deffile "${CMAKE_CURRENT_BINARY_DIR}/${CMAKE_BUILD_TYPE}/pywrap_tensorflow.def")
+ else()
+ set(pywrap_tensorflow_deffile "${CMAKE_CURRENT_BINARY_DIR}/pywrap_tensorflow.def")
+ endif()
set_source_files_properties(${pywrap_tensorflow_deffile} PROPERTIES GENERATED TRUE)
add_custom_command(TARGET pywrap_tensorflow_internal_static POST_BUILD
@@ -549,6 +553,7 @@ if(WIN32)
--input "${pywrap_tensorflow_internal_static_dependencies}"
--output "${pywrap_tensorflow_deffile}"
--target _pywrap_tensorflow_internal.pyd
+ BYPRODUCTS ${pywrap_tensorflow_deffile} # Required for Ninja
)
endif(WIN32)
@@ -702,11 +707,19 @@ add_custom_command(TARGET tf_python_copy_scripts_to_destination PRE_BUILD
${CMAKE_CURRENT_BINARY_DIR}/tf_python/tensorflow/contrib/testing/python/framework/)
if(WIN32)
- add_custom_command(TARGET tf_python_build_pip_package POST_BUILD
- COMMAND ${CMAKE_COMMAND} -E copy ${CMAKE_CURRENT_BINARY_DIR}/$(Configuration)/pywrap_tensorflow_internal.dll
- ${CMAKE_CURRENT_BINARY_DIR}/tf_python/tensorflow/python/_pywrap_tensorflow_internal.pyd
- COMMAND ${CMAKE_COMMAND} -E copy ${CMAKE_CURRENT_BINARY_DIR}/$(Configuration)/pywrap_tensorflow_internal.lib
- ${CMAKE_CURRENT_BINARY_DIR}/tf_python/tensorflow/python/)
+ if(${CMAKE_GENERATOR} MATCHES "Visual Studio.*")
+ add_custom_command(TARGET tf_python_build_pip_package POST_BUILD
+ COMMAND ${CMAKE_COMMAND} -E copy ${CMAKE_CURRENT_BINARY_DIR}/$(Configuration)/pywrap_tensorflow_internal.dll
+ ${CMAKE_CURRENT_BINARY_DIR}/tf_python/tensorflow/python/_pywrap_tensorflow_internal.pyd
+ COMMAND ${CMAKE_COMMAND} -E copy ${CMAKE_CURRENT_BINARY_DIR}/$(Configuration)/pywrap_tensorflow_internal.lib
+ ${CMAKE_CURRENT_BINARY_DIR}/tf_python/tensorflow/python/)
+ else()
+ add_custom_command(TARGET tf_python_build_pip_package POST_BUILD
+ COMMAND ${CMAKE_COMMAND} -E copy ${CMAKE_CURRENT_BINARY_DIR}/pywrap_tensorflow_internal.dll
+ ${CMAKE_CURRENT_BINARY_DIR}/tf_python/tensorflow/python/_pywrap_tensorflow_internal.pyd
+ COMMAND ${CMAKE_COMMAND} -E copy ${CMAKE_CURRENT_BINARY_DIR}/pywrap_tensorflow_internal.lib
+ ${CMAKE_CURRENT_BINARY_DIR}/tf_python/tensorflow/python/)
+ endif()
else()
add_custom_command(TARGET tf_python_build_pip_package POST_BUILD
COMMAND ${CMAKE_COMMAND} -E copy ${CMAKE_CURRENT_BINARY_DIR}/libpywrap_tensorflow_internal.so
diff --git a/tensorflow/contrib/cmake/tf_shared_lib.cmake b/tensorflow/contrib/cmake/tf_shared_lib.cmake
index 571d2b0dec..6d36d5fc5c 100644
--- a/tensorflow/contrib/cmake/tf_shared_lib.cmake
+++ b/tensorflow/contrib/cmake/tf_shared_lib.cmake
@@ -46,7 +46,11 @@ if(WIN32)
$<TARGET_FILE:tf_protos_cc>
)
- set(tensorflow_deffile "${CMAKE_CURRENT_BINARY_DIR}/${CMAKE_BUILD_TYPE}/tensorflow.def")
+ if(${CMAKE_GENERATOR} MATCHES "Visual Studio.*")
+ set(tensorflow_deffile "${CMAKE_CURRENT_BINARY_DIR}/${CMAKE_BUILD_TYPE}/tensorflow.def")
+ else()
+ set(tensorflow_deffile "${CMAKE_CURRENT_BINARY_DIR}/tensorflow.def")
+ endif()
set_source_files_properties(${tensorflow_deffile} PROPERTIES GENERATED TRUE)
add_custom_command(TARGET tensorflow_static POST_BUILD
diff --git a/tensorflow/contrib/cmake/tf_tests.cmake b/tensorflow/contrib/cmake/tf_tests.cmake
index 0a8d691f32..1c4ebd7f0c 100644
--- a/tensorflow/contrib/cmake/tf_tests.cmake
+++ b/tensorflow/contrib/cmake/tf_tests.cmake
@@ -310,6 +310,8 @@ if (tensorflow_BUILD_PYTHON_TESTS)
"${tensorflow_source_dir}/tensorflow/python/kernel_tests/control_flow_util_test.py"
# Flaky replicate_model_fn_test
"${tensorflow_source_dir}/tensorflow/contrib/estimator/python/estimator/replicate_model_fn_test.py" # b/71901810
+ # Broken io_utils_test
+ "${tensorflow_source_dir}/tensorflow/python/keras/_impl/keras/utils/io_utils_test.py" # b/72894325
)
endif()
list(REMOVE_ITEM tf_test_src_py ${tf_test_src_py_exclude})
diff --git a/tensorflow/contrib/cmake/tf_tools.cmake b/tensorflow/contrib/cmake/tf_tools.cmake
index cb58a2e7df..58c7df95c8 100644
--- a/tensorflow/contrib/cmake/tf_tools.cmake
+++ b/tensorflow/contrib/cmake/tf_tools.cmake
@@ -48,9 +48,6 @@ file(GLOB_RECURSE tf_tools_transform_graph_lib_exclude_srcs
"${tensorflow_source_dir}/tensorflow/tools/graph_transforms/compare_graphs.cc"
"${tensorflow_source_dir}/tensorflow/tools/graph_transforms/summarize_graph_main.cc"
"${tensorflow_source_dir}/tensorflow/tools/graph_transforms/transform_graph_main.cc"
- "${tensorflow_source_dir}/tensorflow/tools/graph_transforms/quantize_nodes.cc"
- "${tensorflow_source_dir}/tensorflow/tools/graph_transforms/quantize_weights.cc"
- "${tensorflow_source_dir}/tensorflow/tools/graph_transforms/round_weights.cc"
)
list(REMOVE_ITEM tf_tools_transform_graph_lib_srcs ${tf_tools_transform_graph_lib_exclude_srcs})
diff --git a/tensorflow/contrib/data/python/ops/dataset_ops.py b/tensorflow/contrib/data/python/ops/dataset_ops.py
new file mode 100644
index 0000000000..bb6b049694
--- /dev/null
+++ b/tensorflow/contrib/data/python/ops/dataset_ops.py
@@ -0,0 +1,691 @@
+# Copyright 2017 The TensorFlow Authors. All Rights Reserved.
+#
+# Licensed under the Apache License, Version 2.0 (the "License");
+# you may not use this file except in compliance with the License.
+# You may obtain a copy of the License at
+#
+# http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS,
+# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+# See the License for the specific language governing permissions and
+# limitations under the License.
+# ==============================================================================
+"""Python wrappers for Datasets and Iterators."""
+from __future__ import absolute_import
+from __future__ import division
+from __future__ import print_function
+
+from tensorflow.contrib.data.python.ops import batching
+from tensorflow.contrib.data.python.ops import enumerate_ops
+from tensorflow.contrib.data.python.ops import error_ops
+from tensorflow.contrib.data.python.ops import grouping
+from tensorflow.python.data.ops import dataset_ops
+from tensorflow.python.data.util import nest
+from tensorflow.python.ops import gen_dataset_ops
+from tensorflow.python.ops import gen_io_ops
+from tensorflow.python.util import deprecation
+
+
+class Dataset(dataset_ops.Dataset):
+ """Represents a potentially large set of elements.
+
+ A `Dataset` can be used to represent an input pipeline as a
+ collection of elements (nested structures of tensors) and a "logical
+ plan" of transformations that act on those elements.
+ """
+
+ def __init__(self, dataset):
+ super(Dataset, self).__init__()
+ self._dataset = dataset
+
+ @deprecation.deprecated(None, "Use `ds._as_variant_tensor()`.")
+ def make_dataset_resource(self):
+ return self._as_variant_tensor()
+
+ def _as_variant_tensor(self):
+ return self._dataset._as_variant_tensor() # pylint: disable=protected-access
+
+ @property
+ def output_classes(self):
+ return self._dataset.output_classes
+
+ @property
+ def output_shapes(self):
+ return self._dataset.output_shapes
+
+ @property
+ def output_types(self):
+ return self._dataset.output_types
+
+ @staticmethod
+ @deprecation.deprecated(None, "Use `tf.data.Dataset.from_tensors()`.")
+ def from_tensors(tensors):
+ """Creates a `Dataset` with a single element, comprising the given tensors.
+
+ Args:
+ tensors: A nested structure of tensors.
+
+ Returns:
+ A `Dataset`.
+ """
+ return Dataset(dataset_ops.TensorDataset(tensors))
+
+ @staticmethod
+ @deprecation.deprecated(None, "Use `tf.data.Dataset.from_tensor_slices()`.")
+ def from_tensor_slices(tensors):
+ """Creates a `Dataset` whose elements are slices of the given tensors.
+
+ Args:
+ tensors: A nested structure of tensors, each having the same size in the
+ 0th dimension.
+
+ Returns:
+ A `Dataset`.
+ """
+ return Dataset(dataset_ops.TensorSliceDataset(tensors))
+
+ @staticmethod
+ @deprecation.deprecated(None,
+ "Use `tf.data.Dataset.from_sparse_tensor_slices()`.")
+ def from_sparse_tensor_slices(sparse_tensor):
+ """Splits each rank-N `tf.SparseTensor` in this dataset row-wise.
+
+ Args:
+ sparse_tensor: A `tf.SparseTensor`.
+
+ Returns:
+ A `Dataset` of rank-(N-1) sparse tensors.
+ """
+ return Dataset(dataset_ops.SparseTensorSliceDataset(sparse_tensor))
+
+ @staticmethod
+ @deprecation.deprecated(None, "Use `tf.data.Dataset.from_generator()`.")
+ def from_generator(generator, output_types, output_shapes=None):
+ """Creates a `Dataset` whose elements are generated by `generator`.
+
+ The `generator` argument must be a callable object that returns
+ an object that support the `iter()` protocol (e.g. a generator function).
+ The elements generated by `generator` must be compatible with the given
+ `output_types` and (optional) `output_shapes` arguments.
+
+ For example:
+
+ ```python
+ import itertools
+
+ def gen():
+ for i in itertools.count(1):
+ yield (i, [1] * i)
+
+ ds = Dataset.from_generator(
+ gen, (tf.int64, tf.int64), (tf.TensorShape([]), tf.TensorShape([None])))
+ value = ds.make_one_shot_iterator().get_next()
+
+ sess.run(value) # (1, array([1]))
+ sess.run(value) # (2, array([1, 1]))
+ ```
+
+ Args:
+ generator: A callable object that takes no arguments and returns an
+ object that supports the `iter()` protocol.
+ output_types: A nested structure of `tf.DType` objects corresponding to
+ each component of an element yielded by `generator`.
+ output_shapes: (Optional.) A nested structure of `tf.TensorShape`
+ objects corresponding to each component of an element yielded by
+ `generator`.
+
+ Returns:
+ A `Dataset`.
+ """
+ return Dataset(
+ dataset_ops.Dataset.from_generator(generator, output_types,
+ output_shapes))
+
+ @staticmethod
+ @deprecation.deprecated(None, "Use `tf.data.Dataset.range()`.")
+ def range(*args):
+ """Creates a `Dataset` of a step-separated range of values.
+
+ For example:
+
+ ```python
+ Dataset.range(5) == [0, 1, 2, 3, 4]
+ Dataset.range(2, 5) == [2, 3, 4]
+ Dataset.range(1, 5, 2) == [1, 3]
+ Dataset.range(1, 5, -2) == []
+ Dataset.range(5, 1) == []
+ Dataset.range(5, 1, -2) == [5, 3]
+ ```
+
+ Args:
+ *args: follow same semantics as python's xrange.
+ len(args) == 1 -> start = 0, stop = args[0], step = 1
+ len(args) == 2 -> start = args[0], stop = args[1], step = 1
+ len(args) == 3 -> start = args[0], stop = args[1, stop = args[2]
+
+ Returns:
+ A `RangeDataset`.
+
+ Raises:
+ ValueError: if len(args) == 0.
+ """
+ return Dataset(dataset_ops.RangeDataset(*args))
+
+ @staticmethod
+ @deprecation.deprecated(None, "Use `tf.data.Dataset.zip()`.")
+ def zip(datasets):
+ """Creates a `Dataset` by zipping together the given datasets.
+
+ This method has similar semantics to the built-in `zip()` function
+ in Python, with the main difference being that the `datasets`
+ argument can be an arbitrary nested structure of `Dataset` objects.
+ For example:
+
+ ```python
+ # NOTE: The following examples use `{ ... }` to represent the
+ # contents of a dataset.
+ a = { 1, 2, 3 }
+ b = { 4, 5, 6 }
+ c = { (7, 8), (9, 10), (11, 12) }
+ d = { 13, 14 }
+
+ # The nested structure of the `datasets` argument determines the
+ # structure of elements in the resulting dataset.
+ Dataset.zip((a, b)) == { (1, 4), (2, 5), (3, 6) }
+ Dataset.zip((b, a)) == { (4, 1), (5, 2), (6, 3) }
+
+ # The `datasets` argument may contain an arbitrary number of
+ # datasets.
+ Dataset.zip((a, b, c)) == { (1, 4, (7, 8)),
+ (2, 5, (9, 10)),
+ (3, 6, (11, 12)) }
+
+ # The number of elements in the resulting dataset is the same as
+ # the size of the smallest dataset in `datasets`.
+ Dataset.zip((a, d)) == { (1, 13), (2, 14) }
+ ```
+
+ Args:
+ datasets: A nested structure of datasets.
+
+ Returns:
+ A `Dataset`.
+ """
+ return Dataset(dataset_ops.ZipDataset(datasets))
+
+ def concatenate(self, dataset):
+ """Creates a `Dataset` by concatenating given dataset with this dataset.
+
+ ```python
+ # NOTE: The following examples use `{ ... }` to represent the
+ # contents of a dataset.
+ a = { 1, 2, 3 }
+ b = { 4, 5, 6, 7 }
+
+ # Input dataset and dataset to be concatenated should have same
+ # nested structures and output types.
+ # c = { (8, 9), (10, 11), (12, 13) }
+ # d = { 14.0, 15.0, 16.0 }
+ # a.concatenate(c) and a.concatenate(d) would result in error.
+
+ a.concatenate(b) == { 1, 2, 3, 4, 5, 6, 7 }
+ ```
+
+ Args:
+ dataset: `Dataset` to be concatenated.
+
+ Returns:
+ A `Dataset`.
+ """
+ return Dataset(dataset_ops.ConcatenateDataset(self._dataset, dataset))
+
+ def prefetch(self, buffer_size):
+ """Creates a `Dataset` that prefetches elements from this dataset.
+
+ Args:
+ buffer_size: A `tf.int64` scalar `tf.Tensor`, representing the
+ maximum number elements that will be buffered when prefetching.
+
+ Returns:
+ A `Dataset`.
+ """
+ return Dataset(dataset_ops.PrefetchDataset(self._dataset, buffer_size))
+
+ @staticmethod
+ @deprecation.deprecated(None, "Use `tf.data.Dataset.list_files()`.")
+ def list_files(file_pattern):
+ """A dataset of all files matching a pattern.
+
+ Example:
+ If we had the following files on our filesystem:
+ - /path/to/dir/a.txt
+ - /path/to/dir/b.py
+ - /path/to/dir/c.py
+ If we pass "/path/to/dir/*.py" as the directory, the dataset would
+ produce:
+ - /path/to/dir/b.py
+ - /path/to/dir/c.py
+
+ Args:
+ file_pattern: A string or scalar string `tf.Tensor`, representing
+ the filename pattern that will be matched.
+
+ Returns:
+ A `Dataset` of strings corresponding to file names.
+ """
+ return Dataset.from_tensor_slices(gen_io_ops.matching_files(file_pattern))
+
+ def repeat(self, count=None):
+ """Repeats this dataset `count` times.
+
+ Args:
+ count: (Optional.) A `tf.int64` scalar `tf.Tensor`, representing the
+ number of times the elements of this dataset should be repeated. The
+ default behavior (if `count` is `None` or `-1`) is for the elements to
+ be repeated indefinitely.
+
+ Returns:
+ A `Dataset`.
+ """
+ return Dataset(dataset_ops.RepeatDataset(self._dataset, count))
+
+ @deprecation.deprecated(
+ None, "Use `ds.apply(tf.contrib.data.enumerate_dataset())`.")
+ def enumerate(self, start=0):
+ """Deprecated: Use `Dataset.apply(tf.contrib.data.enumerate_dataset(..)`."""
+
+ return self.apply(enumerate_ops.enumerate_dataset(start))
+
+ def shuffle(self, buffer_size, seed=None):
+ """Randomly shuffles the elements of this dataset.
+
+ Args:
+ buffer_size: A `tf.int64` scalar `tf.Tensor`, representing the
+ number of elements from this dataset from which the new
+ dataset will sample.
+ seed: (Optional.) A `tf.int64` scalar `tf.Tensor`, representing the
+ random seed that will be used to create the distribution. See
+ @{tf.set_random_seed} for behavior.
+
+ Returns:
+ A `Dataset`.
+ """
+ return Dataset(dataset_ops.ShuffleDataset(self._dataset, buffer_size, seed))
+
+ def cache(self, filename=""):
+ """Caches the elements in this dataset.
+
+ Args:
+ filename: A `tf.string` scalar `tf.Tensor`, representing the name of a
+ directory on the filesystem to use for caching tensors in this Dataset.
+ If a filename is not provided, the dataset will be cached in memory.
+
+ Returns:
+ A `Dataset`.
+ """
+ return Dataset(dataset_ops.CacheDataset(self._dataset, filename))
+
+ def take(self, count):
+ """Creates a `Dataset` with at most `count` elements from this dataset.
+
+ Args:
+ count: A `tf.int64` scalar `tf.Tensor`, representing the number of
+ elements of this dataset that should be taken to form the new dataset.
+ If `count` is -1, or if `count` is greater than the size of this
+ dataset, the new dataset will contain all elements of this dataset.
+
+ Returns:
+ A `Dataset`.
+ """
+ return Dataset(dataset_ops.TakeDataset(self._dataset, count))
+
+ def skip(self, count):
+ """Creates a `Dataset` that skips `count` elements from this dataset.
+
+ Args:
+ count: A `tf.int64` scalar `tf.Tensor`, representing the number
+ of elements of this dataset that should be skipped to form the
+ new dataset. If `count` is greater than the size of this
+ dataset, the new dataset will contain no elements. If `count`
+ is -1, skips the entire dataset.
+
+ Returns:
+ A `Dataset`.
+ """
+ return Dataset(dataset_ops.SkipDataset(self._dataset, count))
+
+ def shard(self, num_shards, index):
+ """Creates a `Dataset` that includes only 1/`num_shards` of this dataset.
+
+ This dataset operator is very useful when running distributed training, as
+ it allows each worker to read a unique subset.
+
+ When reading a single input file, you can skip elements as follows:
+
+ ```python
+ d = tf.data.TFRecordDataset(FLAGS.input_file)
+ d = d.shard(FLAGS.num_workers, FLAGS.worker_index)
+ d = d.repeat(FLAGS.num_epochs)
+ d = d.shuffle(FLAGS.shuffle_buffer_size)
+ d = d.map(parser_fn, num_parallel_calls=FLAGS.num_map_threads)
+ ```
+
+ Important caveats:
+
+ - Be sure to shard before you use any randomizing operator (such as
+ shuffle).
+ - Generally it is best if the shard operator is used early in the dataset
+ pipeline. For example, when reading from a set of TFRecord files, shard
+ before converting the dataset to input samples. This avoids reading every
+ file on every worker. The following is an example of an efficient
+ sharding strategy within a complete pipeline:
+
+ ```python
+ d = tf.data.Dataset.list_files(FLAGS.pattern)
+ d = d.shard(FLAGS.num_workers, FLAGS.worker_index)
+ d = d.repeat(FLAGS.num_epochs)
+ d = d.shuffle(FLAGS.shuffle_buffer_size)
+ d = d.interleave(tf.data.TFRecordDataset,
+ cycle_length=FLAGS.num_readers, block_length=1)
+ d = d.map(parser_fn, num_parallel_calls=FLAGS.num_map_threads)
+ ```
+
+ Args:
+ num_shards: A `tf.int64` scalar `tf.Tensor`, representing the number of
+ shards operating in parallel.
+ index: A `tf.int64` scalar `tf.Tensor`, representing the worker index.
+
+ Returns:
+ A `Dataset`.
+
+ Raises:
+ ValueError: if `num_shards` or `index` are illegal values. Note: error
+ checking is done on a best-effort basis, and aren't guaranteed to be
+ caught upon dataset creation. (e.g. providing in a placeholder tensor
+ bypasses the early checking, and will instead result in an error during
+ a session.run call.)
+ """
+ return Dataset(self._dataset.shard(num_shards, index))
+
+ @deprecation.deprecated(None,
+ "Use `ds.apply(tf.contrib.data.ignore_errors())`.")
+ def ignore_errors(self):
+ """Deprecated: Use `Dataset.apply(tf.contrib.data.ignore_errors())`."""
+
+ return self.apply(error_ops.ignore_errors())
+
+ def batch(self, batch_size):
+ """Combines consecutive elements of this dataset into batches.
+
+ Args:
+ batch_size: A `tf.int64` scalar `tf.Tensor`, representing the number of
+ consecutive elements of this dataset to combine in a single batch.
+
+ Returns:
+ A `Dataset`.
+ """
+ return Dataset(dataset_ops.BatchDataset(self._dataset, batch_size))
+
+ def padded_batch(self, batch_size, padded_shapes, padding_values=None):
+ """Combines consecutive elements of this dataset into padded batches.
+
+ Like `Dataset.dense_to_sparse_batch()`, this method combines
+ multiple consecutive elements of this dataset, which might have
+ different shapes, into a single element. The tensors in the
+ resulting element have an additional outer dimension, and are
+ padded to the respective shape in `padded_shapes`.
+
+ Args:
+ batch_size: A `tf.int64` scalar `tf.Tensor`, representing the number of
+ consecutive elements of this dataset to combine in a single batch.
+ padded_shapes: A nested structure of `tf.TensorShape` or
+ `tf.int64` vector tensor-like objects representing the shape
+ to which the respective component of each input element should
+ be padded prior to batching. Any unknown dimensions
+ (e.g. `tf.Dimension(None)` in a `tf.TensorShape` or `-1` in a
+ tensor-like object) will be padded to the maximum size of that
+ dimension in each batch.
+ padding_values: (Optional.) A nested structure of scalar-shaped
+ `tf.Tensor`, representing the padding values to use for the
+ respective components. Defaults are `0` for numeric types and
+ the empty string for string types.
+
+ Returns:
+ A `Dataset`.
+ """
+ return Dataset(
+ dataset_ops.PaddedBatchDataset(self._dataset, batch_size, padded_shapes,
+ padding_values))
+
+ @deprecation.deprecated(
+ None, "Use `ds.apply(tf.contrib.data.dense_to_sparse_batch())`.")
+ def dense_to_sparse_batch(self, batch_size, row_shape):
+ """Use: `Dataset.apply(tf.contrib.data.dense_to_sparse_batch(...))`."""
+
+ return self.apply(batching.dense_to_sparse_batch(batch_size, row_shape))
+
+ @deprecation.deprecated(None,
+ "Use `ds.apply(tf.contrib.data.group_by_window())`.")
+ def group_by_window(self, key_func, reduce_func, window_size):
+ """Deprecated: Use `Dataset.apply(tf.contrib.data.group_by_window(...))`."""
+
+ return self.apply(
+ grouping.group_by_window(key_func, reduce_func, window_size))
+
+ @deprecation.deprecated_args(
+ None, "Replace `num_threads=T` with `num_parallel_calls=T`. Replace "
+ "`output_buffer_size=N` with `ds.prefetch(N)` on the returned dataset.",
+ "num_threads", "output_buffer_size")
+ def map(self,
+ map_func,
+ num_threads=None,
+ output_buffer_size=None,
+ num_parallel_calls=None):
+ """Maps `map_func` across this dataset.
+
+ Args:
+ map_func: A function mapping a nested structure of tensors (having
+ shapes and types defined by `self.output_shapes` and
+ `self.output_types`) to another nested structure of tensors.
+ num_threads: (Optional.) Deprecated, use `num_parallel_calls` instead.
+ output_buffer_size: (Optional.) A `tf.int64` scalar `tf.Tensor`,
+ representing the maximum number of processed elements that will be
+ buffered.
+ num_parallel_calls: (Optional.) A `tf.int32` scalar `tf.Tensor`,
+ representing the number elements to process in parallel. If not
+ specified, elements will be processed sequentially.
+
+ Returns:
+ A `Dataset`.
+ """
+ if num_threads is None and num_parallel_calls is None:
+ ret = Dataset(dataset_ops.MapDataset(self._dataset, map_func))
+ else:
+ if num_threads is None:
+ ret = Dataset(
+ dataset_ops.ParallelMapDataset(self._dataset, map_func,
+ num_parallel_calls))
+ else:
+ ret = Dataset(
+ dataset_ops.ParallelMapDataset(self._dataset, map_func,
+ num_threads))
+ if output_buffer_size is not None:
+ ret = ret.prefetch(output_buffer_size)
+ return ret
+
+ def flat_map(self, map_func):
+ """Maps `map_func` across this dataset and flattens the result.
+
+ Args:
+ map_func: A function mapping a nested structure of tensors (having shapes
+ and types defined by `self.output_shapes` and `self.output_types`) to a
+ `Dataset`.
+
+ Returns:
+ A `Dataset`.
+ """
+ return Dataset(dataset_ops.FlatMapDataset(self._dataset, map_func))
+
+ def interleave(self, map_func, cycle_length, block_length=1):
+ """Maps `map_func` across this dataset, and interleaves the results.
+
+ For example, you can use `Dataset.interleave()` to process many input files
+ concurrently:
+
+ ```python
+ # Preprocess 4 files concurrently, and interleave blocks of 16 records from
+ # each file.
+ filenames = ["/var/data/file1.txt", "/var/data/file2.txt", ...]
+ dataset = (Dataset.from_tensor_slices(filenames)
+ .interleave(lambda x:
+ TextLineDataset(x).map(parse_fn, num_parallel_calls=1),
+ cycle_length=4, block_length=16))
+ ```
+
+ The `cycle_length` and `block_length` arguments control the order in which
+ elements are produced. `cycle_length` controls the number of input elements
+ that are processed concurrently. If you set `cycle_length` to 1, this
+ transformation will handle one input element at a time, and will produce
+ identical results = to @{tf.data.Dataset.flat_map}. In general,
+ this transformation will apply `map_func` to `cycle_length` input elements,
+ open iterators on the returned `Dataset` objects, and cycle through them
+ producing `block_length` consecutive elements from each iterator, and
+ consuming the next input element each time it reaches the end of an
+ iterator.
+
+ For example:
+
+ ```python
+ # NOTE: The following examples use `{ ... }` to represent the
+ # contents of a dataset.
+ a = { 1, 2, 3, 4, 5 }
+
+ # NOTE: New lines indicate "block" boundaries.
+ a.interleave(lambda x: Dataset.from_tensors(x).repeat(6),
+ cycle_length=2, block_length=4) == {
+ 1, 1, 1, 1,
+ 2, 2, 2, 2,
+ 1, 1,
+ 2, 2,
+ 3, 3, 3, 3,
+ 4, 4, 4, 4,
+ 3, 3,
+ 4, 4,
+ 5, 5, 5, 5,
+ 5, 5,
+ }
+ ```
+
+ NOTE: The order of elements yielded by this transformation is
+ deterministic, as long as `map_func` is a pure function. If
+ `map_func` contains any stateful operations, the order in which
+ that state is accessed is undefined.
+
+ Args:
+ map_func: A function mapping a nested structure of tensors (having shapes
+ and types defined by `self.output_shapes` and `self.output_types`) to a
+ `Dataset`.
+ cycle_length: The number of elements from this dataset that will be
+ processed concurrently.
+ block_length: The number of consecutive elements to produce from each
+ input element before cycling to another input element.
+
+ Returns:
+ A `Dataset`.
+ """
+ return Dataset(
+ dataset_ops.InterleaveDataset(self._dataset, map_func, cycle_length,
+ block_length))
+
+ @deprecation.deprecated(None, "Use `ds.apply(tf.contrib.data.unbatch())`.")
+ def unbatch(self):
+ """Deprecated: Use `Dataset.apply(tf.contrib.data.unbatch()`."""
+
+ return self.apply(batching.unbatch())
+
+ def filter(self, predicate):
+ """Filters this dataset according to `predicate`.
+
+ Args:
+ predicate: A function mapping a nested structure of tensors (having shapes
+ and types defined by `self.output_shapes` and `self.output_types`) to a
+ scalar `tf.bool` tensor.
+
+ Returns:
+ A `Dataset`.
+ """
+ return Dataset(dataset_ops.FilterDataset(self._dataset, predicate))
+
+ def apply(self, transformation_func):
+ """Apply a transformation function to this dataset.
+
+ `apply` enables chaining of custom `Dataset` transformations, which are
+ represented as functions that take one `Dataset` argument and return a
+ transformed `Dataset`.
+
+ For example:
+
+ ```
+ dataset = (dataset.map(lambda x: x ** 2)
+ .(group_by_window(key_func, reduce_func, window_size))
+ .map(lambda x: x ** 3))
+ ```
+
+ Args:
+ transformation_func: A function that takes one `Dataset` argument and
+ returns a `Dataset`.
+
+ Returns:
+ The `Dataset` returned by applying `transformation_func` to this dataset.
+ """
+ dataset = transformation_func(self)
+ if not isinstance(dataset, dataset_ops.Dataset):
+ raise TypeError("`transformation_func` must return a Dataset.")
+ return Dataset(dataset)
+
+
+def get_single_element(dataset):
+ """Returns the single element in `dataset` as a nested structure of tensors.
+
+ This function enables you to use a @{tf.data.Dataset} in a stateless
+ "tensor-in tensor-out" expression, without creating a @{tf.data.Iterator}.
+ This can be useful when your preprocessing transformations are expressed
+ as a `Dataset`, and you want to use the transformation at serving time.
+ For example:
+
+ ```python
+ input_batch = tf.placeholder(tf.string, shape=[BATCH_SIZE])
+
+ def preprocessing_fn(input_str):
+ # ...
+ return image, label
+
+ dataset = (tf.data.Dataset.from_tensor_slices(input_batch)
+ .map(preprocessing_fn, num_parallel_calls=BATCH_SIZE)
+ .batch(BATCH_SIZE))
+
+ image_batch, label_batch = tf.contrib.data.get_single_element(dataset)
+ ```
+
+ Args:
+ dataset: A @{tf.data.Dataset} object containing a single element.
+
+ Returns:
+ A nested structure of @{tf.Tensor} objects, corresponding to the single
+ element of `dataset`.
+
+ Raises:
+ TypeError: if `dataset` is not a `tf.data.Dataset` object.
+ InvalidArgumentError (at runtime): if `dataset` does not contain exactly
+ one element.
+ """
+ if not isinstance(dataset, dataset_ops.Dataset):
+ raise TypeError("`dataset` must be a `tf.data.Dataset` object.")
+ return nest.pack_sequence_as(
+ dataset.output_types,
+ gen_dataset_ops.dataset_to_single_element(
+ dataset._as_variant_tensor(), # pylint: disable=protected-access
+ output_types=nest.flatten(dataset.output_types),
+ output_shapes=nest.flatten(dataset.output_shapes)))
diff --git a/tensorflow/contrib/data/python/ops/grouping.py b/tensorflow/contrib/data/python/ops/grouping.py
index ef91c56726..67b085002a 100644
--- a/tensorflow/contrib/data/python/ops/grouping.py
+++ b/tensorflow/contrib/data/python/ops/grouping.py
@@ -45,7 +45,7 @@ def group_by_window(key_func,
key_func: A function mapping a nested structure of tensors
(having shapes and types defined by `self.output_shapes` and
`self.output_types`) to a scalar `tf.int64` tensor.
- reduce_func: A function mapping a key and a dataset of up to `batch_size`
+ reduce_func: A function mapping a key and a dataset of up to `window_size`
consecutive elements matching that key to another dataset.
window_size: A `tf.int64` scalar `tf.Tensor`, representing the number of
consecutive elements matching the same key to combine in a single
diff --git a/tensorflow/contrib/distributions/__init__.py b/tensorflow/contrib/distributions/__init__.py
index f0517142ab..61c411271d 100644
--- a/tensorflow/contrib/distributions/__init__.py
+++ b/tensorflow/contrib/distributions/__init__.py
@@ -98,7 +98,6 @@ _allowed_symbols = [
'Autoregressive',
'Binomial',
'Bernoulli',
- 'BernoulliWithSigmoidProbs',
'Beta',
'BetaWithSoftplusConcentration',
'Categorical',
diff --git a/tensorflow/contrib/eager/python/examples/gan/README.md b/tensorflow/contrib/eager/python/examples/gan/README.md
index e8c9db1a1e..208a64b05d 100644
--- a/tensorflow/contrib/eager/python/examples/gan/README.md
+++ b/tensorflow/contrib/eager/python/examples/gan/README.md
@@ -11,7 +11,7 @@ Other eager execution examples can be found under the parent directory.
- `mnist.py`: Model definitions and training routines.
- `mnist_test.py`: Benchmarks for training and using the models using eager
execution.
-- `mnist_graph_test.py`: Benchmarks for trainig and using the models using
+- `mnist_graph_test.py`: Benchmarks for training and using the models using
graph execution. The same model definitions and loss functions are used in
all benchmarks.
diff --git a/tensorflow/contrib/eager/python/examples/mnist/mnist.py b/tensorflow/contrib/eager/python/examples/mnist/mnist.py
index 241eb23ce9..58b1e89d15 100644
--- a/tensorflow/contrib/eager/python/examples/mnist/mnist.py
+++ b/tensorflow/contrib/eager/python/examples/mnist/mnist.py
@@ -39,7 +39,7 @@ class MNISTModel(tf.keras.Model):
"""MNIST Network.
Network structure is equivalent to:
- https://github.com/tensorflow/tensorflow/blob/r1.5/tensorflow/examples/tutorials/mnist/mnist_deep.py
+ https://github.com/tensorflow/tensorflow/blob/r1.6/tensorflow/examples/tutorials/mnist/mnist_deep.py
and
https://github.com/tensorflow/models/blob/master/tutorials/image/mnist/convolutional.py
diff --git a/tensorflow/contrib/framework/python/ops/accumulate_n_v2.py b/tensorflow/contrib/framework/python/ops/accumulate_n_v2.py
index 2375ee4f55..476528b0dd 100644
--- a/tensorflow/contrib/framework/python/ops/accumulate_n_v2.py
+++ b/tensorflow/contrib/framework/python/ops/accumulate_n_v2.py
@@ -22,6 +22,7 @@ from __future__ import print_function
from tensorflow.python.eager import context
from tensorflow.python.framework import ops
from tensorflow.python.framework import tensor_shape
+from tensorflow.python.ops import array_ops
from tensorflow.python.ops import gen_math_ops
from tensorflow.python.ops import math_ops
@@ -108,4 +109,3 @@ def _AddNGrad(op, grad):
"""Same as gradient for AddN. Copies the gradient to all inputs."""
# Not broadcasting.
return [grad] * len(op.inputs)
-
diff --git a/tensorflow/contrib/gan/python/eval/python/summaries_impl.py b/tensorflow/contrib/gan/python/eval/python/summaries_impl.py
index 74811ff409..0d1afad72d 100644
--- a/tensorflow/contrib/gan/python/eval/python/summaries_impl.py
+++ b/tensorflow/contrib/gan/python/eval/python/summaries_impl.py
@@ -39,12 +39,13 @@ def _assert_is_image(data):
data.shape[1:].assert_is_fully_defined()
-def add_gan_model_image_summaries(gan_model, grid_size=4):
+def add_gan_model_image_summaries(gan_model, grid_size=4, model_summaries=True):
"""Adds image summaries for real and fake images.
Args:
gan_model: A GANModel tuple.
grid_size: The size of an image grid.
+ model_summaries: Also add summaries of the model.
Raises:
ValueError: If real and generated data aren't images.
@@ -83,7 +84,9 @@ def add_gan_model_image_summaries(gan_model, grid_size=4):
image_shape=generated_image_shape,
num_channels=generated_channels),
max_outputs=1)
- add_gan_model_summaries(gan_model)
+
+ if model_summaries:
+ add_gan_model_summaries(gan_model)
def add_image_comparison_summaries(gan_model, num_comparisons=2,
diff --git a/tensorflow/contrib/gan/python/eval/python/summaries_test.py b/tensorflow/contrib/gan/python/eval/python/summaries_test.py
index a02d8772e1..5549df971d 100644
--- a/tensorflow/contrib/gan/python/eval/python/summaries_test.py
+++ b/tensorflow/contrib/gan/python/eval/python/summaries_test.py
@@ -71,9 +71,10 @@ def get_cyclegan_model():
class SummariesTest(test.TestCase):
- def _test_add_gan_model_image_summaries_impl(self, get_model_fn,
- expected_num_summary_ops):
- summaries.add_gan_model_image_summaries(get_model_fn(), grid_size=2)
+ def _test_add_gan_model_image_summaries_impl(
+ self, get_model_fn, expected_num_summary_ops, model_summaries):
+ summaries.add_gan_model_image_summaries(
+ get_model_fn(), grid_size=2, model_summaries=model_summaries)
self.assertEquals(expected_num_summary_ops,
len(ops.get_collection(ops.GraphKeys.SUMMARIES)))
@@ -82,10 +83,13 @@ class SummariesTest(test.TestCase):
summary.merge_all().eval()
def test_add_gan_model_image_summaries(self):
- self._test_add_gan_model_image_summaries_impl(get_gan_model, 5)
+ self._test_add_gan_model_image_summaries_impl(get_gan_model, 5, True)
+
+ def test_add_gan_model_image_summaries_no_model(self):
+ self._test_add_gan_model_image_summaries_impl(get_gan_model, 2, False)
def test_add_gan_model_image_summaries_for_cyclegan(self):
- self._test_add_gan_model_image_summaries_impl(get_cyclegan_model, 10)
+ self._test_add_gan_model_image_summaries_impl(get_cyclegan_model, 10, True)
def _test_add_gan_model_summaries_impl(self, get_model_fn,
expected_num_summary_ops):
diff --git a/tensorflow/contrib/gdr/README.md b/tensorflow/contrib/gdr/README.md
index 34ce60b360..8242d93f12 100644
--- a/tensorflow/contrib/gdr/README.md
+++ b/tensorflow/contrib/gdr/README.md
@@ -119,4 +119,4 @@ In the original design (as in the reference), tensor buffers are only registered
Reference
===
-Bairen Yi, Jiacheng Xia, Li Chen, and Kai Chen. 2017. Towards Zero Copy Dataflows using RDMA. In Proceedings of SIGCOMM Posters and Demos'17, Los Angeles, CA, USA, August 22-24, 2017, 3 pages. https://doi.org/10.1145/3123878.3123907
+Bairen Yi, Jiacheng Xia, Li Chen, and Kai Chen. 2017. Towards Zero Copy Dataflows using RDMA. In Proceedings of SIGCOMM Posters and Demos'17, Los Angeles, CA, USA, August 22-24, 2017, 3 pages. https://doi.org/10.1145/3123878.3131975
diff --git a/tensorflow/contrib/layers/__init__.py b/tensorflow/contrib/layers/__init__.py
index ef419862b4..337c9e06b8 100644
--- a/tensorflow/contrib/layers/__init__.py
+++ b/tensorflow/contrib/layers/__init__.py
@@ -35,6 +35,7 @@ See the @{$python/contrib.layers} guide.
@@fully_connected
@@GDN
@@gdn
+@@images_to_sequence
@@layer_norm
@@linear
@@max_pool2d
@@ -50,6 +51,7 @@ See the @{$python/contrib.layers} guide.
@@scale_gradient
@@separable_conv2d
@@separable_convolution2d
+@@sequence_to_images
@@softmax
@@spatial_softmax
@@stack
diff --git a/tensorflow/contrib/layers/python/layers/layers.py b/tensorflow/contrib/layers/python/layers/layers.py
index 1c3af19a6c..dcee775337 100644
--- a/tensorflow/contrib/layers/python/layers/layers.py
+++ b/tensorflow/contrib/layers/python/layers/layers.py
@@ -60,9 +60,10 @@ __all__ = [
'conv2d_in_plane', 'conv2d_transpose', 'conv3d_transpose', 'convolution',
'convolution2d', 'convolution2d_in_plane', 'convolution2d_transpose',
'convolution3d', 'convolution3d_transpose', 'dense_to_sparse', 'dropout',
- 'elu', 'flatten', 'fully_connected', 'GDN', 'gdn', 'layer_norm', 'linear',
- 'pool', 'max_pool2d', 'max_pool3d', 'one_hot_encoding', 'relu', 'relu6',
- 'repeat', 'scale_gradient', 'separable_conv2d', 'separable_convolution2d',
+ 'elu', 'flatten', 'fully_connected', 'GDN', 'gdn', 'images_to_sequence',
+ 'layer_norm', 'linear', 'pool', 'max_pool2d', 'max_pool3d',
+ 'one_hot_encoding', 'relu', 'relu6', 'repeat', 'scale_gradient',
+ 'separable_conv2d', 'separable_convolution2d', 'sequence_to_images',
'softmax', 'spatial_softmax', 'stack', 'unit_norm',
'legacy_fully_connected', 'legacy_linear', 'legacy_relu', 'maxout'
]
@@ -2186,6 +2187,36 @@ def layer_norm(inputs,
@add_arg_scope
+def images_to_sequence(inputs,
+ data_format=DATA_FORMAT_NHWC,
+ outputs_collections=None,
+ scope=None):
+ """Convert a batch of images into a batch of sequences.
+ Args:
+ inputs: a (num_images, height, width, depth) tensor
+ data_format: A string. `NHWC` (default) and `NCHW` are supported.
+ outputs_collections: The collections to which the outputs are added.
+ scope: Optional scope for name_scope.
+ Returns:
+ (width, num_images*height, depth) sequence tensor
+ """
+ if data_format not in (DATA_FORMAT_NCHW, DATA_FORMAT_NHWC):
+ raise ValueError('data_format has to be either NCHW or NHWC.')
+ with ops.name_scope(scope, 'ImagesToSequence', [inputs]) as sc:
+ inputs = ops.convert_to_tensor(inputs)
+ df = ('channels_first'
+ if data_format and data_format.startswith('NC') else 'channels_last')
+ if df == 'channels_first':
+ inputs = array_ops.transpose(inputs, [0, 2, 3, 1])
+ _, _, width, depth = inputs.get_shape().as_list()
+ s = array_ops.shape(inputs)
+ batch_size, height = s[0], s[1]
+ transposed = array_ops.transpose(inputs, [2, 0, 1, 3])
+ outputs = array_ops.reshape(transposed, [width, batch_size * height, depth])
+ return utils.collect_named_outputs(outputs_collections, sc, outputs)
+
+
+@add_arg_scope
def max_pool2d(inputs,
kernel_size,
stride=2,
@@ -2665,6 +2696,38 @@ def separable_convolution2d(
@add_arg_scope
+def sequence_to_images(inputs,
+ height,
+ output_data_format='channels_last',
+ outputs_collections=None,
+ scope=None):
+ """Convert a batch of sequences into a batch of images.
+ Args:
+ inputs: (num_steps, num_batches, depth) sequence tensor
+ height: the height of the images
+ output_data_format: Format of output tensor.
+ Currently supports `'channels_first'` and `'channels_last'`.
+ outputs_collections: The collections to which the outputs are added.
+ scope: Optional scope for name_scope.
+ Returns:
+ A tensor representing the output of the operation.
+ """
+ with ops.name_scope(scope, 'SequenceToImages', [inputs]) as sc:
+ inputs = ops.convert_to_tensor(inputs)
+ width, num_batches, depth = inputs.get_shape().as_list()
+ if num_batches is None:
+ num_batches = -1
+ else:
+ num_batches = num_batches // height
+ reshaped = array_ops.reshape(inputs, [width, num_batches, height, depth])
+ if output_data_format == 'channels_first':
+ outputs = array_ops.transpose(reshaped, [1, 3, 2, 0])
+ else:
+ outputs = array_ops.transpose(reshaped, [1, 2, 0, 3])
+ return utils.collect_named_outputs(outputs_collections, sc, outputs)
+
+
+@add_arg_scope
def softmax(logits, scope=None):
"""Performs softmax on Nth dimension of N-dimensional logit tensor.
diff --git a/tensorflow/contrib/layers/python/layers/layers_test.py b/tensorflow/contrib/layers/python/layers/layers_test.py
index 972ff10bf9..5c0ae9a3f1 100644
--- a/tensorflow/contrib/layers/python/layers/layers_test.py
+++ b/tensorflow/contrib/layers/python/layers/layers_test.py
@@ -708,7 +708,7 @@ class Convolution2dTransposeTests(test.TestCase):
_layers.convolution2d_transpose(images, 32, 3, data_format='CHWN')
def testOutputSizeWithStrideOneSamePaddingNCHW(self):
- # `NCHW` data fomat is only supported for `GPU` device.
+ # `NCHW` data format is only supported for `GPU` device.
if test.is_gpu_available(cuda_only=True):
with self.test_session(use_gpu=True) as sess:
num_filters = 32
@@ -2196,7 +2196,7 @@ class BatchNormTest(test.TestCase):
# After initialization moving_mean == 0 and moving_variance == 1.
self.assertAllClose(mean, [0] * 3)
self.assertAllClose(variance, [1] * 3)
- # Simulate assigment from saver restore.
+ # Simulate assignment from saver restore.
init_assigns = [
state_ops.assign(moving_mean, expected_mean),
state_ops.assign(moving_variance, expected_var)
@@ -2950,6 +2950,28 @@ class GDNTest(test.TestCase):
self.assertAllClose(y, x * np.sqrt(1 + .1 * (x**2)), rtol=0, atol=1e-6)
+class ImagesToSequenceTest(test.TestCase):
+
+ def testInvalidDataFormat(self):
+ height, width = 7, 11
+ images = np.random.uniform(size=(5, height, width, 2))
+ with self.assertRaisesRegexp(ValueError,
+ 'data_format has to be either NCHW or NHWC.'):
+ _layers.images_to_sequence(images, data_format='CHWN')
+
+ def testImagesToSequenceDims(self):
+ height, width = 7, 11
+ images = np.random.uniform(size=(2, height, width, 5)).astype(np.float32)
+ output = _layers.images_to_sequence(images)
+ self.assertListEqual(output.get_shape().as_list(), [11, 14, 5])
+
+ def testImagesToSequenceNCHW(self):
+ height, width = 7, 11
+ images = np.random.uniform(size=(2, 5, height, width)).astype(np.float32)
+ output = _layers.images_to_sequence(images, data_format='NCHW')
+ self.assertListEqual(output.get_shape().as_list(), [11, 14, 5])
+
+
class MaxPool2DTest(test.TestCase):
def testInvalidDataFormat(self):
@@ -3418,6 +3440,30 @@ class ScaleGradientTests(test.TestCase):
np.testing.assert_array_equal([3 * 2], g_x.eval())
+class SequenceToImagesTest(test.TestCase):
+
+ def testImagesToSequenceDims(self):
+ num_batches = 14
+ num_time_steps = 11
+ num_channels = 5
+ desired_height = 7
+ sequence = np.random.uniform(
+ size=(num_time_steps, num_batches, num_channels)).astype(np.float32)
+ output = _layers.sequence_to_images(sequence, desired_height)
+ self.assertListEqual(output.get_shape().as_list(), [2, 7, 11, 5])
+
+ def testImagesToSequenceNCHW(self):
+ num_batches = 14
+ num_time_steps = 11
+ num_channels = 5
+ desired_height = 7
+ sequence = np.random.uniform(
+ size=(num_time_steps, num_batches, num_channels)).astype(np.float32)
+ output = _layers.sequence_to_images(
+ sequence, desired_height, output_data_format='channels_first')
+ self.assertListEqual(output.get_shape().as_list(), [2, 5, 7, 11])
+
+
class SoftmaxTests(test.TestCase):
def setUp(self):
diff --git a/tensorflow/contrib/learn/python/learn/monitors.py b/tensorflow/contrib/learn/python/learn/monitors.py
index 0948dee7e2..51381a7427 100644
--- a/tensorflow/contrib/learn/python/learn/monitors.py
+++ b/tensorflow/contrib/learn/python/learn/monitors.py
@@ -879,7 +879,7 @@ class GraphDump(BaseMonitor):
this_output = self.data[step] if step in self.data else {}
other_output = other_dump.data[step] if step in other_dump.data else {}
for key in this_output:
- if not isinstance(key, str) and not isinstance(key, unicode):
+ if not isinstance(key, six.string_types):
continue
if key not in other_output:
raise ValueError("%s missing at step %s.", (key, step))
diff --git a/tensorflow/contrib/lite/README.md b/tensorflow/contrib/lite/README.md
index 7cdf91da88..3e55d2a496 100644
--- a/tensorflow/contrib/lite/README.md
+++ b/tensorflow/contrib/lite/README.md
@@ -172,7 +172,7 @@ Here is a sample command line to convert the frozen Graphdef to '.tflite' format
```
bazel build tensorflow/contrib/lite/toco:toco
-bazel-bin/tensorflow/contrib/lite/toco/toco -- \
+bazel-bin/tensorflow/contrib/lite/toco/toco \
--input_file=$(pwd)/mobilenet_v1_1.0_224/frozen_graph.pb \
--input_format=TENSORFLOW_GRAPHDEF --output_format=TFLITE \
--output_file=/tmp/mobilenet_v1_1.0_224.tflite --inference_type=FLOAT \
diff --git a/tensorflow/contrib/lite/examples/label_image/bitmap_helpers_impl.h b/tensorflow/contrib/lite/examples/label_image/bitmap_helpers_impl.h
index a908e0c5ee..2a64c1de72 100644
--- a/tensorflow/contrib/lite/examples/label_image/bitmap_helpers_impl.h
+++ b/tensorflow/contrib/lite/examples/label_image/bitmap_helpers_impl.h
@@ -22,6 +22,12 @@ limitations under the License.
#include "tensorflow/contrib/lite/string_util.h"
#include "tensorflow/contrib/lite/version.h"
+#include "tensorflow/contrib/lite/builtin_op_data.h"
+#include "tensorflow/contrib/lite/interpreter.h"
+#include "tensorflow/contrib/lite/kernels/register.h"
+#include "tensorflow/contrib/lite/string_util.h"
+#include "tensorflow/contrib/lite/version.h"
+
#include "tensorflow/contrib/lite/examples/label_image/label_image.h"
namespace tflite {
diff --git a/tensorflow/contrib/lite/g3doc/custom_operators.md b/tensorflow/contrib/lite/g3doc/custom_operators.md
index 204a489a93..d7cc854eba 100644
--- a/tensorflow/contrib/lite/g3doc/custom_operators.md
+++ b/tensorflow/contrib/lite/g3doc/custom_operators.md
@@ -73,7 +73,7 @@ TfLiteStatus SinEval(TfLiteContext* context, TfLiteNode* node) {
}
TfLiteRegistration* Register_SIN() {
- static TfLiteRegistration r = {nullptr, nullptr, SinResize, SinEval};
+ static TfLiteRegistration r = {nullptr, nullptr, SinPrepare, SinEval};
return &r;
}
```
diff --git a/tensorflow/contrib/lite/graph_info.cc b/tensorflow/contrib/lite/graph_info.cc
index 186b80fe82..e60ed2c246 100644
--- a/tensorflow/contrib/lite/graph_info.cc
+++ b/tensorflow/contrib/lite/graph_info.cc
@@ -1,4 +1,4 @@
-/* Copyright 2017 The TensorFlow Authors. All Rights Reserved.
+/* Copyright 2018 The TensorFlow Authors. All Rights Reserved.
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
diff --git a/tensorflow/contrib/lite/kernels/internal/optimized/tensor_utils_impl.h b/tensorflow/contrib/lite/kernels/internal/optimized/tensor_utils_impl.h
index f8be99e82f..4e324a5e10 100644
--- a/tensorflow/contrib/lite/kernels/internal/optimized/tensor_utils_impl.h
+++ b/tensorflow/contrib/lite/kernels/internal/optimized/tensor_utils_impl.h
@@ -15,7 +15,7 @@ limitations under the License.
#ifndef TF_LITE_KERNELS_INTERNAL_OPTIMIZED_TENSOR_UTILS_IMPL_H_
#define TF_LITE_KERNELS_INTERNAL_OPTIMIZED_TENSOR_UTILS_IMPL_H_
-// TDOD(ghodrat): Remove this header file and the dependency to internal data
+// TODO(ghodrat): Remove this header file and the dependency to internal data
// structure.
#include "tensorflow/contrib/lite/builtin_op_data.h"
diff --git a/tensorflow/contrib/lite/kernels/internal/reference/portable_tensor_utils.h b/tensorflow/contrib/lite/kernels/internal/reference/portable_tensor_utils.h
index afc3e26e79..c05c21b472 100644
--- a/tensorflow/contrib/lite/kernels/internal/reference/portable_tensor_utils.h
+++ b/tensorflow/contrib/lite/kernels/internal/reference/portable_tensor_utils.h
@@ -15,7 +15,7 @@ limitations under the License.
#ifndef TENSORFLOW_CONTRIB_LITE_KERNELS_INTERNAL_REFERENCE_PORTABLE_TENSOR_UTILS_H_
#define TENSORFLOW_CONTRIB_LITE_KERNELS_INTERNAL_REFERENCE_PORTABLE_TENSOR_UTILS_H_
-// TDOD(ghodrat): Remove this header file and the dependency to internal data
+// TODO(ghodrat): Remove this header file and the dependency to internal data
// structure.
#include "tensorflow/contrib/lite/builtin_op_data.h"
diff --git a/tensorflow/contrib/lite/simple_memory_arena.h b/tensorflow/contrib/lite/simple_memory_arena.h
index 0c5e00a1f2..0535522374 100644
--- a/tensorflow/contrib/lite/simple_memory_arena.h
+++ b/tensorflow/contrib/lite/simple_memory_arena.h
@@ -36,9 +36,9 @@ struct ArenaAlloc {
}
};
-// This small class is responsible for allocating, dealocating and reusing
+// This small class is responsible for allocating, deallocating and reusing
// dynamic memory from a common underlying buffer. The arena can be used in
-// scenarios when the pattern of memory allocations and dealocations is
+// scenarios when the pattern of memory allocations and deallocations is
// repetitive, e.g. running NN inference in multiple iterations.
class SimpleMemoryArena {
public:
diff --git a/tensorflow/contrib/lite/testing/generate_examples.py b/tensorflow/contrib/lite/testing/generate_examples.py
index 944031da24..b6c09306d6 100644
--- a/tensorflow/contrib/lite/testing/generate_examples.py
+++ b/tensorflow/contrib/lite/testing/generate_examples.py
@@ -629,7 +629,7 @@ def make_constant_tests(zip_path):
def build_graph(parameters):
# Since Toco & Tflite can't have a single constant op in the entire graph,
- # this test adds a zero tesnor with a constant op tensor.
+ # this test adds a zero tensor with a constant op tensor.
input1 = tf.placeholder(dtype=parameters["dtype"], name="input1",
shape=parameters["input_shape"])
out = tf.ones(parameters["input_shape"], dtype=parameters["dtype"]) + input1
diff --git a/tensorflow/contrib/lite/toco/graph_transformations/resolve_constant_concatenation.cc b/tensorflow/contrib/lite/toco/graph_transformations/resolve_constant_concatenation.cc
index e79e2a32fc..064810b53e 100644
--- a/tensorflow/contrib/lite/toco/graph_transformations/resolve_constant_concatenation.cc
+++ b/tensorflow/contrib/lite/toco/graph_transformations/resolve_constant_concatenation.cc
@@ -73,7 +73,7 @@ void CopyTensorSegments(const std::vector<Array*>& input_arrays,
// Receives a series of input arrays of type Array and an integer showing the
// axis on which those arrays will be concatenated. It returns the concatenated
-// arrray.
+// array.
template <ArrayDataType A>
void ConcatenateTensorBuffers(const std::vector<Array*>& input_arrays,
int concatenation_axis,
diff --git a/tensorflow/contrib/lite/toco/tensorflow_graph_matching/resolve_svdf_test.cc b/tensorflow/contrib/lite/toco/tensorflow_graph_matching/resolve_svdf_test.cc
index 664e828c19..646d048496 100644
--- a/tensorflow/contrib/lite/toco/tensorflow_graph_matching/resolve_svdf_test.cc
+++ b/tensorflow/contrib/lite/toco/tensorflow_graph_matching/resolve_svdf_test.cc
@@ -103,11 +103,11 @@ class ResolveSvdfTest : public ::testing::Test {
// Add the float vector as an attribute to the node.
(*node->mutable_attr())["dtype"].set_type(tensorflow::DT_FLOAT);
tensorflow::TensorProto* allocated_tensor = new tensorflow::TensorProto;
- tensorflow::TensorShapeProto* allocated_tesnor_shape =
+ tensorflow::TensorShapeProto* allocated_tensor_shape =
new tensorflow::TensorShapeProto;
- auto tensor_shape_dim0 = allocated_tesnor_shape->add_dim();
+ auto tensor_shape_dim0 = allocated_tensor_shape->add_dim();
tensor_shape_dim0->set_size(values.size());
- allocated_tensor->set_allocated_tensor_shape(allocated_tesnor_shape);
+ allocated_tensor->set_allocated_tensor_shape(allocated_tensor_shape);
allocated_tensor->set_tensor_content(
string(reinterpret_cast<const char*>(values.data()),
values.size() * sizeof(float)));
@@ -122,11 +122,11 @@ class ResolveSvdfTest : public ::testing::Test {
// Add the float vector as an attribute to the node.
(*node->mutable_attr())["dtype"].set_type(tensorflow::DT_INT32);
tensorflow::TensorProto* allocated_tensor = new tensorflow::TensorProto;
- tensorflow::TensorShapeProto* allocated_tesnor_shape =
+ tensorflow::TensorShapeProto* allocated_tensor_shape =
new tensorflow::TensorShapeProto;
- auto tensor_shape_dim0 = allocated_tesnor_shape->add_dim();
+ auto tensor_shape_dim0 = allocated_tensor_shape->add_dim();
tensor_shape_dim0->set_size(values.size());
- allocated_tensor->set_allocated_tensor_shape(allocated_tesnor_shape);
+ allocated_tensor->set_allocated_tensor_shape(allocated_tensor_shape);
allocated_tensor->set_tensor_content(
string(reinterpret_cast<const char*>(values.data()),
values.size() * sizeof(int)));
diff --git a/tensorflow/contrib/makefile/README.md b/tensorflow/contrib/makefile/README.md
index 0613de2cab..6959ca344f 100644
--- a/tensorflow/contrib/makefile/README.md
+++ b/tensorflow/contrib/makefile/README.md
@@ -268,7 +268,7 @@ selectively register only for the operators used in your graph.
```bash
tensorflow/contrib/makefile/build_all_ios.sh -a arm64 -g $HOME/graphs/inception/tensorflow_inception_graph.pb
```
-Please note this is an aggresive optimization of the operators and the resulting library may not work with other graphs but will reduce the size of the final library.
+Please note this is an aggressive optimization of the operators and the resulting library may not work with other graphs but will reduce the size of the final library.
The `compile_ios_tensorflow.sh` script can take optional command-line arguments.
The first argument will be passed as a C++ optimization flag and defaults to
diff --git a/tensorflow/contrib/makefile/build_all_android.sh b/tensorflow/contrib/makefile/build_all_android.sh
index f67c516186..fc88f59e09 100755
--- a/tensorflow/contrib/makefile/build_all_android.sh
+++ b/tensorflow/contrib/makefile/build_all_android.sh
@@ -52,7 +52,7 @@ shift $((OPTIND - 1))
if [ "$ARCH" == "tegra" ]; then
if [[ -z "${JETPACK}" ]]; then
- export JETPACK="$HOME/JetPack_Android_3.0"
+ export JETPACK="$HOME/JetPack_Android_3.2"
fi
if [ ! -d ${JETPACK} ]; then
echo "Can't find Jetpack at ${JETPACK}"
diff --git a/tensorflow/contrib/metrics/python/ops/metric_ops.py b/tensorflow/contrib/metrics/python/ops/metric_ops.py
index fc12bfd2b7..31e274c5fd 100644
--- a/tensorflow/contrib/metrics/python/ops/metric_ops.py
+++ b/tensorflow/contrib/metrics/python/ops/metric_ops.py
@@ -740,7 +740,7 @@ def _streaming_confusion_matrix_at_thresholds(predictions,
else:
for include in includes:
if include not in all_includes:
- raise ValueError('Invaild key: %s.' % include)
+ raise ValueError('Invalid key: %s.' % include)
predictions, labels, weights = metrics_impl._remove_squeezable_dimensions( # pylint: disable=protected-access
predictions, labels, weights)
@@ -1516,7 +1516,7 @@ def precision_recall_at_equal_thresholds(labels,
predictions: A floating point `Tensor` of arbitrary shape and whose values
are in the range `[0, 1]`.
weights: Optional; If provided, a `Tensor` that has the same dtype as,
- and broadcastable to, `predictions`. This tensor is multplied by counts.
+ and broadcastable to, `predictions`. This tensor is multiplied by counts.
num_thresholds: Optional; Number of thresholds, evenly distributed in
`[0, 1]`. Should be `>= 2`. Defaults to 201. Note that the number of bins
is 1 less than `num_thresholds`. Using an even `num_thresholds` value
diff --git a/tensorflow/contrib/nearest_neighbor/kernels/heap.h b/tensorflow/contrib/nearest_neighbor/kernels/heap.h
index 32925569a8..a2dbb8052b 100644
--- a/tensorflow/contrib/nearest_neighbor/kernels/heap.h
+++ b/tensorflow/contrib/nearest_neighbor/kernels/heap.h
@@ -56,7 +56,7 @@ class HeapBase {
// This method adds an element at the end of the internal array without
// "heapifying" the array afterwards. This is useful for setting up a heap
- // where a single call to heapify at the end of the inital insertion
+ // where a single call to heapify at the end of the initial insertion
// operations suffices.
void InsertUnsorted(const KeyType& key, const DataType& data) {
if (v_.size() == static_cast<size_t>(num_elements_)) {
diff --git a/tensorflow/contrib/opt/python/training/elastic_average_optimizer.py b/tensorflow/contrib/opt/python/training/elastic_average_optimizer.py
index 716ee9cdf7..5763593b81 100644
--- a/tensorflow/contrib/opt/python/training/elastic_average_optimizer.py
+++ b/tensorflow/contrib/opt/python/training/elastic_average_optimizer.py
@@ -150,7 +150,7 @@ class ElasticAverageOptimizer(optimizer.Optimizer):
self._global_map = ea_custom_getter._global_map
if moving_rate is None:
- self._moving_rate = BETA / communication_period / num_worker
+ self._moving_rate = self.BETA / communication_period / num_worker
else:
self._moving_rate = moving_rate
if rho is None:
diff --git a/tensorflow/contrib/opt/python/training/model_average_optimizer_test.py b/tensorflow/contrib/opt/python/training/model_average_optimizer_test.py
index 29ecd22839..6cca0a8a00 100644
--- a/tensorflow/contrib/opt/python/training/model_average_optimizer_test.py
+++ b/tensorflow/contrib/opt/python/training/model_average_optimizer_test.py
@@ -110,7 +110,6 @@ def _get_workers(num_workers, steps, workers):
class ModelAverageOptimizerTest(test.TestCase):
-
def _run(self, train_op, sess):
sess.run(train_op)
diff --git a/tensorflow/contrib/py2tf/converters/BUILD b/tensorflow/contrib/py2tf/converters/BUILD
index 93c751b28d..e9a96ec8d1 100644
--- a/tensorflow/contrib/py2tf/converters/BUILD
+++ b/tensorflow/contrib/py2tf/converters/BUILD
@@ -132,6 +132,7 @@ py_test(
py_test(
name = "for_loops_test",
srcs = ["for_loops_test.py"],
+ srcs_version = "PY2AND3",
deps = [
":test_lib",
"//tensorflow/contrib/py2tf/pyct",
diff --git a/tensorflow/contrib/rnn/python/kernel_tests/rnn_cell_test.py b/tensorflow/contrib/rnn/python/kernel_tests/rnn_cell_test.py
index 51933be29d..eef1ae25e9 100644
--- a/tensorflow/contrib/rnn/python/kernel_tests/rnn_cell_test.py
+++ b/tensorflow/contrib/rnn/python/kernel_tests/rnn_cell_test.py
@@ -1605,6 +1605,7 @@ class WeightNormLSTMCellTest(test.TestCase):
self.assertAllClose(expected_c, actual_c, 1e-5)
self.assertAllClose(expected_h, actual_h, 1e-5)
+
def testBasicCellWithNorm(self):
"""Tests cell w/o peepholes and with normalisation"""
diff --git a/tensorflow/contrib/seq2seq/python/ops/attention_wrapper.py b/tensorflow/contrib/seq2seq/python/ops/attention_wrapper.py
index d6b5eceb47..0a53fd66db 100644
--- a/tensorflow/contrib/seq2seq/python/ops/attention_wrapper.py
+++ b/tensorflow/contrib/seq2seq/python/ops/attention_wrapper.py
@@ -331,7 +331,7 @@ def _luong_score(query, keys, scale):
# batched matmul on:
# [batch_size, 1, depth] . [batch_size, depth, max_time]
# resulting in an output shape of:
- # [batch_time, 1, max_time].
+ # [batch_size, 1, max_time].
# we then squeeze out the center singleton dimension.
score = math_ops.matmul(query, keys, transpose_b=True)
score = array_ops.squeeze(score, [1])
diff --git a/tensorflow/contrib/slim/README.md b/tensorflow/contrib/slim/README.md
index c7a54cb9a2..2d9df8f27e 100644
--- a/tensorflow/contrib/slim/README.md
+++ b/tensorflow/contrib/slim/README.md
@@ -145,7 +145,7 @@ regular_variables_and_model_variables = slim.get_variables()
How does this work? When you create a model variable via TF-Slim's layers or
directly via the `slim.model_variable` function, TF-Slim adds the variable to
-a the `tf.GraphKeys.MODEL_VARIABLES` collection. What if you have your own
+the `tf.GraphKeys.MODEL_VARIABLES` collection. What if you have your own
custom layers or variable creation routine but still want TF-Slim to manage or
be aware of your model variables? TF-Slim provides a convenience function for
adding the model variable to its collection:
diff --git a/tensorflow/contrib/solvers/python/ops/linear_equations.py b/tensorflow/contrib/solvers/python/ops/linear_equations.py
index 2395707257..d791d46763 100644
--- a/tensorflow/contrib/solvers/python/ops/linear_equations.py
+++ b/tensorflow/contrib/solvers/python/ops/linear_equations.py
@@ -28,6 +28,7 @@ from tensorflow.python.ops import array_ops
from tensorflow.python.ops import control_flow_ops
from tensorflow.python.ops import linalg_ops
from tensorflow.python.ops import math_ops
+from tensorflow.python.ops import linalg_ops
def conjugate_gradient(operator,
diff --git a/tensorflow/contrib/tpu/profiler/pip_package/setup.py b/tensorflow/contrib/tpu/profiler/pip_package/setup.py
index 33ade16003..76f1dd2a56 100644
--- a/tensorflow/contrib/tpu/profiler/pip_package/setup.py
+++ b/tensorflow/contrib/tpu/profiler/pip_package/setup.py
@@ -20,7 +20,7 @@ from __future__ import print_function
from setuptools import setup
-_VERSION = '1.5.0-rc1'
+_VERSION = '1.6.0-rc0'
CONSOLE_SCRIPTS = [
'capture_tpu_profile=cloud_tpu_profiler.main:run_main',
diff --git a/tensorflow/contrib/tpu/tpu_estimator.md b/tensorflow/contrib/tpu/tpu_estimator.md
index ca1255b16b..4ef8f9eebd 100644
--- a/tensorflow/contrib/tpu/tpu_estimator.md
+++ b/tensorflow/contrib/tpu/tpu_estimator.md
@@ -231,7 +231,7 @@ Refer to this link for all [Cloud TPU documentation](https://cloud.google.com/tp
### Profiling
-You can profile the `worker` by using instructions as spcified in the [Cloud TPU Tools](https://cloud.google.com/tpu/docs/cloud-tpu-tools).
+You can profile the `worker` by using instructions as specified in the [Cloud TPU Tools](https://cloud.google.com/tpu/docs/cloud-tpu-tools).
### Is `int64` supported?
diff --git a/tensorflow/contrib/verbs/README.md b/tensorflow/contrib/verbs/README.md
index 1b99f4ce4f..58fed4e5cb 100644
--- a/tensorflow/contrib/verbs/README.md
+++ b/tensorflow/contrib/verbs/README.md
@@ -25,9 +25,9 @@ The design is based on TensorFlow r1.0. An RDMA path is added between servers fo
During the server setup, an RDMA manager is created to manage low-level RDMA components such as RDMA channel and RDMA adapter, an RDMA rendezvous manager is created to oversee send/recv operations between servers. Following the distributed TensorFlow design philosophy, the send operation is passive, i.e. merely placing a tensor in the local out-going table. It is the receive operation that actually initiates the tensor transfer.
TensorFlow dynamically allocates memory for tensors that are to be sent or received. This causes difficulty for RDMA operations where pinned memory is required. Few remedies are possible:
-1. The memory is pinned, transfered, then unpinned for each and every tensor to be transferred. This incurs significant operation overhead since pinning and unpinning memory for each dynamically generated tensor is slow.
+1. The memory is pinned, transferred, then unpinned for each and every tensor to be transferred. This incurs significant operation overhead since pinning and unpinning memory for each dynamically generated tensor is slow.
2. Buffer is pre-allocated and pinned for each tensor. This incurs large memory overhead and extra copying from the tensor to its pinned buffer, but may still be faster than the former.
-3. Following HKUST research on the use of GPU direct, and their [GDR implementation](https://github.com/tensorflow/tensorflow/blob/master/tensorflow/contrib/gdr/README.md), there is a smart way to benefit from the TensorFlow allocation theme which is mostly pool based, i.e allocators pre-allocate a large memory block, and allocate the tensors from there. By attaching a custom Visitor to relevant alloactors, we can do a single registration of the entire memory block, which zeros the registration overhead. Once the block is registered, each new tensor allocated will be at a registred address, which will allow us to do direct RDMA writes to it.
+3. Following HKUST research on the use of GPU direct, and their [GDR implementation](https://github.com/tensorflow/tensorflow/blob/master/tensorflow/contrib/gdr/README.md), there is a smart way to benefit from the TensorFlow allocation theme which is mostly pool based, i.e allocators pre-allocate a large memory block, and allocate the tensors from there. By attaching a custom Visitor to relevant allocators, we can do a single registration of the entire memory block, which zeros the registration overhead. Once the block is registered, each new tensor allocated will be at a registered address, which will allow us to do direct RDMA writes to it.
For best performance, we will adopt HKUST 0 copies approach in our solution. This means:
@@ -77,7 +77,7 @@ When the receiver receives the **RDMA_MESSAGE_META_DATA_RESPONSE**, it will loca
1. Update the local meta-data cache.
2. Reallocate the result/proxy tensors.
-3. Re-send the tensor request. For tracability, the new message has a different name: **RDMA_MESSAGE_TENSOR_RE_REQUEST**.
+3. Re-send the tensor request. For traceability, the new message has a different name: **RDMA_MESSAGE_TENSOR_RE_REQUEST**.
When the sender receives a **RDMA_MESSAGE_TENSOR_RE_REQUEST**, it will locate the relevant **RdmaTensorResponse** using the request index specified in the message, and invoke its **Resume()** method, which will RDMA write the contents of the tensor that was cloned earlier, to the new remote address specified in the re-request.
@@ -93,7 +93,7 @@ When the receiver receives the RDMA write, it will locate the relevant **RdmaTen
1. When the sender receives a tensor request, the source tensor may or may not be ready yet. The situation is handled through a process of tag matching:
* If the request arrives before the tensor is ready, then a callback is put in a local table, and will be invoked once the tensor arrives.
- * If the tensor is ready before the request arives, than the tensor is put in a local table. When the request arrives, it will invoke the callback immediatly.
+ * If the tensor is ready before the request arives, than the tensor is put in a local table. When the request arrives, it will invoke the callback immediately.
In code it is done by calling **RecvLocalAsync()**, which receives the tensor's key, step-id, and the callback.
2. When the callback is invoked, the relevant tensor is removed from the tag matching table. In the case where we need to send the tensor's meta-data, the **RdmaTensorResponse** will store a copy of the tensor until the re-request arrives.
3. The sending of protocol messages (**RDMA_MESSAGE_TENSOR_REQUEST**, **RDMA_MESSAGE_META_DATA_RESPONSE** and **RDMA_MESSAGE_TENSOR_RE_REQUEST**) is done by the class **RdmaMessageBuffer**. All messages are sent using RDMA writes from/to fixed messages buffers. This implies that we cannot send on a specific channel more than one message at a time. In order to synchronize the messages, the **RdmaMessageBuffer** holds the a local and remote buffer statuses which can be either busy or idle. When a write is issued, both statuses will be changed to busy. When the write-complete event is received, the local status is changed to idle. When the write is received on the remote side, the remote side will parse the message, and return an ACK back to the sending side on which the sending side will update the remote status to idle. When both the local and remote statuses are idle, the next message can be sent.
@@ -115,7 +115,7 @@ When the receiver receives the RDMA write, it will locate the relevant **RdmaTen
* Reallocate the result tensor (and proxy tensor if required).
* Re-send the request to the remote side.
* **RecvTensorContent()** - Receive tensor content from the remote side (RDMA write was completed).
- * Decode proto if required and/or move to GPU if the content was not written to it directly (GPU direct is not avaliable).
+ * Decode proto if required and/or move to GPU if the content was not written to it directly (GPU direct is not available).
* Invoke the done callback.
* **class RdmaTensorResponse** - Holds and manages information for a single tensor response throughout the entire send cycle. API:
* **Start()** - Start the response sequence.
@@ -153,7 +153,7 @@ When the receiver receives the RDMA write, it will locate the relevant **RdmaTen
* request_index - Request index.
* is_dead/data_type/tensor_shape/tensor_bytes - The up-to-date meta-data.
* checksum - In data validation mode, this will hold the checksum of the source tensor.
-* **RDMA_MESSAGE_TENSOR_RE_REQUEST** - (receiver ==> sender) Tensor re-requset after meta-data update and reallocation of result/proxy tensors.
+* **RDMA_MESSAGE_TENSOR_RE_REQUEST** - (receiver ==> sender) Tensor re-request after meta-data update and reallocation of result/proxy tensors.
* type - The message type.
* name (name_size) - Name of the requested tensor.
* step_id - Step ID.
diff --git a/tensorflow/core/distributed_runtime/rpc/grpc_serialization_traits.h b/tensorflow/core/distributed_runtime/rpc/grpc_serialization_traits.h
index dd114d39c6..730124c25e 100644
--- a/tensorflow/core/distributed_runtime/rpc/grpc_serialization_traits.h
+++ b/tensorflow/core/distributed_runtime/rpc/grpc_serialization_traits.h
@@ -66,7 +66,7 @@ class GrpcBufferWriter final
}
// It's dangerous to keep an inlined grpc_slice as the backup slice, since
// on a following Next() call, a reference will be returned to this slice
- // via GRPC_SLICE_START_PTR, which will not be an adddress held by
+ // via GRPC_SLICE_START_PTR, which will not be an address held by
// slice_buffer_.
have_backup_ = backup_slice_.refcount != NULL;
byte_count_ -= count;
diff --git a/tensorflow/core/framework/op_def_util.h b/tensorflow/core/framework/op_def_util.h
index d1613ee89b..0ba1325a03 100644
--- a/tensorflow/core/framework/op_def_util.h
+++ b/tensorflow/core/framework/op_def_util.h
@@ -82,7 +82,7 @@ bool AttrDefEqual(const OpDef::AttrDef& a1, const OpDef::AttrDef& a2);
uint64 AttrDefHash(const OpDef::AttrDef& a);
// Returns true if all AttrDefs in `a1` equal corresponding AttrDefs in
-// `a2`. Corrspondence is established by name.
+// `a2`. Correspondence is established by name.
bool RepeatedAttrDefEqual(const protobuf::RepeatedPtrField<OpDef::AttrDef>& a1,
const protobuf::RepeatedPtrField<OpDef::AttrDef>& a2);
diff --git a/tensorflow/core/framework/op_kernel.cc b/tensorflow/core/framework/op_kernel.cc
index 5d58d3e78e..8654437059 100644
--- a/tensorflow/core/framework/op_kernel.cc
+++ b/tensorflow/core/framework/op_kernel.cc
@@ -474,7 +474,7 @@ std::unique_ptr<Tensor> OpKernelContext::forward_input(
return nullptr;
}
// Check that input and output memory types match, i.e.
- // that they either both live in host or both live in device memmory.
+ // that they either both live in host or both live in device memory.
if (input_memory_type(input_index) != output_memory_type) {
return nullptr;
}
diff --git a/tensorflow/core/kernels/constant_op_test.cc b/tensorflow/core/kernels/constant_op_test.cc
index 7a05d9371d..a6baae73d8 100644
--- a/tensorflow/core/kernels/constant_op_test.cc
+++ b/tensorflow/core/kernels/constant_op_test.cc
@@ -77,7 +77,7 @@ void ConstantOpTest::PersistentMemoryTrackingTest(bool on_gpu) {
EXPECT_EQ(ctx.persistent_memory_allocated(), 480);
}
- // Remove memry leak errors.
+ // Remove memory leak errors.
for (auto allocator_pair : ctx.wrapped_allocators()) {
allocator_pair.second->GetRecordsAndUnRef();
}
diff --git a/tensorflow/core/kernels/cwise_op_gpu_invert.cu.cc b/tensorflow/core/kernels/cwise_op_gpu_invert.cu.cc
index 62f33612db..1072ef3aa6 100644
--- a/tensorflow/core/kernels/cwise_op_gpu_invert.cu.cc
+++ b/tensorflow/core/kernels/cwise_op_gpu_invert.cu.cc
@@ -19,7 +19,7 @@ limitations under the License.
namespace tensorflow {
namespace functor {
-DEFINE_UNARY6(invert, int8, int16, int32, int64, uint8, uint16);
+DEFINE_UNARY8(invert, int8, int16, int32, int64, uint8, uint16, uint32, uint64);
} // namespace functor
} // namespace tensorflow
diff --git a/tensorflow/core/kernels/cwise_op_invert.cc b/tensorflow/core/kernels/cwise_op_invert.cc
index f5cafcc780..98c8d7e9b2 100644
--- a/tensorflow/core/kernels/cwise_op_invert.cc
+++ b/tensorflow/core/kernels/cwise_op_invert.cc
@@ -16,17 +16,17 @@ limitations under the License.
#include "tensorflow/core/kernels/cwise_ops_common.h"
namespace tensorflow {
-REGISTER6(UnaryOp, CPU, "Invert", functor::invert, int8, int16, int32, int64,
- uint8, uint16);
+REGISTER8(UnaryOp, CPU, "Invert", functor::invert, int8, int16, int32, int64,
+ uint8, uint16, uint32, uint64);
#ifdef TENSORFLOW_USE_SYCL
REGISTER6(UnaryOp, SYCL, "Invert", functor::invert, int8, int16, int32, int64,
- uint8, uint16);
+ uint8, uint16, uint32, uint64);
#endif // TENSORFLOW_USE_SYCL
#if GOOGLE_CUDA
-REGISTER6(UnaryOp, GPU, "Invert", functor::invert, int8, int16, int32, int64,
- uint8, uint16);
+REGISTER8(UnaryOp, GPU, "Invert", functor::invert, int8, int16, int32, int64,
+ uint8, uint16, uint32, uint64);
#endif // GOOGLE_CUDA
} // namespace tensorflow
diff --git a/tensorflow/core/kernels/cwise_ops_gpu_common.cu.h b/tensorflow/core/kernels/cwise_ops_gpu_common.cu.h
index 6dd108f722..965e42dcce 100644
--- a/tensorflow/core/kernels/cwise_ops_gpu_common.cu.h
+++ b/tensorflow/core/kernels/cwise_ops_gpu_common.cu.h
@@ -136,6 +136,9 @@ struct ApproximateEqual<GPUDevice, T> {
#define DEFINE_UNARY7(F, T0, T1, T2, T3, T4, T5, T6) \
DEFINE_UNARY2(F, T0, T1); \
DEFINE_UNARY5(F, T2, T3, T4, T5, T6)
+#define DEFINE_UNARY8(F, T0, T1, T2, T3, T4, T5, T6, T7) \
+ DEFINE_UNARY4(F, T0, T1, T2, T3); \
+ DEFINE_UNARY4(F, T4, T5, T6, T7)
// Macros to explicitly instantiate kernels on GPU for multiple types
// (T0, T1, etc.) for BinaryFunctor.
diff --git a/tensorflow/core/lib/gif/gif_io.cc b/tensorflow/core/lib/gif/gif_io.cc
index e5deb2b873..9a5215320f 100644
--- a/tensorflow/core/lib/gif/gif_io.cc
+++ b/tensorflow/core/lib/gif/gif_io.cc
@@ -16,6 +16,7 @@ limitations under the License.
// Functions to read images in GIF format.
#include "tensorflow/core/lib/gif/gif_io.h"
+#include <algorithm>
#include "tensorflow/core/lib/gtl/cleanup.h"
#include "tensorflow/core/lib/strings/strcat.h"
#include "tensorflow/core/platform/gif.h"
@@ -89,23 +90,52 @@ uint8* Decode(const void* srcdata, int datasize,
uint8* const dstdata = allocate_output(num_frames, width, height, channel);
if (!dstdata) return nullptr;
for (int k = 0; k < num_frames; k++) {
+ uint8* this_dst = dstdata + k * width * channel * height;
+
SavedImage* this_image = &gif_file->SavedImages[k];
GifImageDesc* img_desc = &this_image->ImageDesc;
+
+ int imgLeft = img_desc->Left;
+ int imgTop = img_desc->Top;
+ int imgRight = img_desc->Left + img_desc->Width;
+ int imgBottom = img_desc->Top + img_desc->Height;
+
if (img_desc->Left != 0 || img_desc->Top != 0 || img_desc->Width != width ||
img_desc->Height != height) {
- *error_string = strings::StrCat("can't process optimized gif");
- return nullptr;
+ // If the first frame does not fill the entire canvas then return error.
+ if (k == 0) {
+ *error_string =
+ strings::StrCat("the first frame does not fill the canvas");
+ return nullptr;
+ }
+ // Otherwise previous frame will be reused to fill the unoccupied canvas.
+ imgLeft = std::max(imgLeft, 0);
+ imgTop = std::max(imgTop, 0);
+ imgRight = std::min(imgRight, width);
+ imgBottom = std::min(imgBottom, height);
+
+ uint8* last_dst = dstdata + (k - 1) * width * channel * height;
+ for (int i = 0; i < height; ++i) {
+ uint8* p_dst = this_dst + i * width * channel;
+ uint8* l_dst = last_dst + i * width * channel;
+ for (int j = 0; j < width; ++j) {
+ p_dst[j * channel + 0] = l_dst[j * channel + 0];
+ p_dst[j * channel + 1] = l_dst[j * channel + 1];
+ p_dst[j * channel + 2] = l_dst[j * channel + 2];
+ }
+ }
}
ColorMapObject* color_map = this_image->ImageDesc.ColorMap
? this_image->ImageDesc.ColorMap
: gif_file->SColorMap;
- uint8* this_dst = dstdata + k * width * channel * height;
- for (int i = 0; i < height; ++i) {
+ for (int i = imgTop; i < imgBottom; ++i) {
uint8* p_dst = this_dst + i * width * channel;
- for (int j = 0; j < width; ++j) {
- GifByteType color_index = this_image->RasterBits[i * width + j];
+ for (int j = imgLeft; j < imgRight; ++j) {
+ GifByteType color_index =
+ this_image->RasterBits[(i - img_desc->Top) * (img_desc->Width) +
+ (j - img_desc->Left)];
const GifColorType& gif_color = color_map->Colors[color_index];
p_dst[j * channel + 0] = gif_color.Red;
p_dst[j * channel + 1] = gif_color.Green;
diff --git a/tensorflow/core/lib/gtl/optional.h b/tensorflow/core/lib/gtl/optional.h
index fa33c24c0c..4ee3f88d18 100644
--- a/tensorflow/core/lib/gtl/optional.h
+++ b/tensorflow/core/lib/gtl/optional.h
@@ -478,7 +478,7 @@ class optional : private internal_optional::optional_data<T>,
return *this;
}
- // Copy assigment, standard semantics.
+ // Copy assignment, standard semantics.
optional& operator=(const optional& src) = default;
// Move assignment, standard semantics.
diff --git a/tensorflow/core/lib/io/record_reader.cc b/tensorflow/core/lib/io/record_reader.cc
index 9cc6c4034f..254fdf115d 100644
--- a/tensorflow/core/lib/io/record_reader.cc
+++ b/tensorflow/core/lib/io/record_reader.cc
@@ -49,7 +49,7 @@ RecordReaderOptions RecordReaderOptions::CreateRecordReaderOptions(
#endif // IS_SLIM_BUILD
} else if (compression_type != compression::kNone) {
LOG(ERROR) << "Unsupported compression_type:" << compression_type
- << ". No comprression will be used.";
+ << ". No compression will be used.";
}
return options;
}
diff --git a/tensorflow/core/ops/image_ops.cc b/tensorflow/core/ops/image_ops.cc
index a62e2d782b..4c05b274fe 100644
--- a/tensorflow/core/ops/image_ops.cc
+++ b/tensorflow/core/ops/image_ops.cc
@@ -455,6 +455,17 @@ REGISTER_OP("SampleDistortedBoundingBox")
.Attr("use_image_if_no_bounding_boxes: bool = false")
.SetIsStateful()
.SetShapeFn([](InferenceContext* c) {
+ // Get inputs and validate ranks.
+ ShapeHandle image_size;
+ TF_RETURN_IF_ERROR(c->WithRank(c->input(0), 1, &image_size));
+ ShapeHandle bounding_boxes;
+ TF_RETURN_IF_ERROR(c->WithRank(c->input(1), 3, &bounding_boxes));
+ // image_size: 1-D with [height, width, channels]
+ // bounding_boxes: 3-D with shape [batch, N, 4]
+ DimensionHandle unused;
+ TF_RETURN_IF_ERROR(c->WithValue(c->Dim(image_size, 0), 3, &unused));
+ TF_RETURN_IF_ERROR(c->WithValue(c->Dim(bounding_boxes, 2), 4, &unused));
+
c->set_output(0, c->Vector(3));
c->set_output(1, c->Vector(3));
c->set_output(2, c->MakeShape({1, 1, 4}));
@@ -477,6 +488,19 @@ REGISTER_OP("SampleDistortedBoundingBoxV2")
.Attr("use_image_if_no_bounding_boxes: bool = false")
.SetIsStateful()
.SetShapeFn([](InferenceContext* c) {
+ // Get inputs and validate ranks.
+ ShapeHandle image_size;
+ TF_RETURN_IF_ERROR(c->WithRank(c->input(0), 1, &image_size));
+ ShapeHandle bounding_boxes;
+ TF_RETURN_IF_ERROR(c->WithRank(c->input(1), 3, &bounding_boxes));
+ ShapeHandle min_object_covered;
+ TF_RETURN_IF_ERROR(c->WithRank(c->input(2), 0, &min_object_covered));
+ // image_size: 1-D with [height, width, channels]
+ // bounding_boxes: 3-D with shape [batch, N, 4]
+ DimensionHandle unused;
+ TF_RETURN_IF_ERROR(c->WithValue(c->Dim(image_size, 0), 3, &unused));
+ TF_RETURN_IF_ERROR(c->WithValue(c->Dim(bounding_boxes, 2), 4, &unused));
+
c->set_output(0, c->Vector(3));
c->set_output(1, c->Vector(3));
c->set_output(2, c->MakeShape({1, 1, 4}));
diff --git a/tensorflow/core/platform/env.cc b/tensorflow/core/platform/env.cc
index 1bcca1243f..12509c250e 100644
--- a/tensorflow/core/platform/env.cc
+++ b/tensorflow/core/platform/env.cc
@@ -44,6 +44,9 @@ limitations under the License.
namespace tensorflow {
+// 128KB copy buffer
+constexpr size_t kCopyFileBufferSize = 128 * 1024;
+
class FileSystemRegistryImpl : public FileSystemRegistry {
public:
Status Register(const string& scheme, Factory factory) override;
@@ -278,6 +281,17 @@ Status Env::RenameFile(const string& src, const string& target) {
return src_fs->RenameFile(src, target);
}
+Status Env::CopyFile(const string& src, const string& target) {
+ FileSystem* src_fs;
+ FileSystem* target_fs;
+ TF_RETURN_IF_ERROR(GetFileSystemForFile(src, &src_fs));
+ TF_RETURN_IF_ERROR(GetFileSystemForFile(target, &target_fs));
+ if (src_fs == target_fs) {
+ return src_fs->CopyFile(src, target);
+ }
+ return FileSystemCopyFile(src_fs, src, target_fs, target);
+}
+
string Env::GetExecutablePath() {
char exe_path[PATH_MAX] = {0};
#ifdef __APPLE__
@@ -406,6 +420,29 @@ Status WriteStringToFile(Env* env, const string& fname,
return s;
}
+Status FileSystemCopyFile(FileSystem* src_fs, const string& src,
+ FileSystem* target_fs, const string& target) {
+ std::unique_ptr<RandomAccessFile> src_file;
+ TF_RETURN_IF_ERROR(src_fs->NewRandomAccessFile(src, &src_file));
+
+ std::unique_ptr<WritableFile> target_file;
+ TF_RETURN_IF_ERROR(target_fs->NewWritableFile(target, &target_file));
+
+ uint64 offset = 0;
+ std::unique_ptr<char[]> scratch(new char[kCopyFileBufferSize]);
+ Status s = Status::OK();
+ while (s.ok()) {
+ StringPiece result;
+ s = src_file->Read(offset, kCopyFileBufferSize, &result, scratch.get());
+ if (!(s.ok() || s.code() == error::OUT_OF_RANGE)) {
+ return s;
+ }
+ TF_RETURN_IF_ERROR(target_file->Append(result));
+ offset += result.size();
+ }
+ return target_file->Close();
+}
+
// A ZeroCopyInputStream on a RandomAccessFile.
namespace {
class FileStream : public ::tensorflow::protobuf::io::ZeroCopyInputStream {
diff --git a/tensorflow/core/platform/env.h b/tensorflow/core/platform/env.h
index 34aaf3f78b..4ce4e0b4e0 100644
--- a/tensorflow/core/platform/env.h
+++ b/tensorflow/core/platform/env.h
@@ -214,6 +214,9 @@ class Env {
/// replaced.
Status RenameFile(const string& src, const string& target);
+ /// \brief Copy the src to target.
+ Status CopyFile(const string& src, const string& target);
+
/// \brief Returns the absolute path of the current executable. It resolves
/// symlinks if there is any.
string GetExecutablePath();
@@ -381,6 +384,11 @@ struct ThreadOptions {
size_t guard_size = 0; // 0: use system default value
};
+/// A utility routine: copy contents of `src` in file system `src_fs`
+/// to `target` in file system `target_fs`.
+Status FileSystemCopyFile(FileSystem* src_fs, const string& src,
+ FileSystem* target_fs, const string& target);
+
/// A utility routine: reads contents of named file into `*data`
Status ReadFileToString(Env* env, const string& fname, string* data);
diff --git a/tensorflow/core/platform/file_system.cc b/tensorflow/core/platform/file_system.cc
index b9866cf641..271d73f5f1 100644
--- a/tensorflow/core/platform/file_system.cc
+++ b/tensorflow/core/platform/file_system.cc
@@ -265,4 +265,8 @@ Status FileSystem::RecursivelyCreateDir(const string& dirname) {
return Status::OK();
}
+Status FileSystem::CopyFile(const string& src, const string& target) {
+ return FileSystemCopyFile(this, src, this, target);
+}
+
} // namespace tensorflow
diff --git a/tensorflow/core/platform/file_system.h b/tensorflow/core/platform/file_system.h
index d32efcea09..3085b6958f 100644
--- a/tensorflow/core/platform/file_system.h
+++ b/tensorflow/core/platform/file_system.h
@@ -189,6 +189,9 @@ class FileSystem {
/// \brief Overwrites the target if it exists.
virtual Status RenameFile(const string& src, const string& target) = 0;
+ /// \brief Copy the src to target.
+ virtual Status CopyFile(const string& src, const string& target);
+
/// \brief Translate an URI to a filename for the FileSystem implementation.
///
/// The implementation in this class cleans up the path, removing
diff --git a/tensorflow/core/platform/posix/posix_file_system.cc b/tensorflow/core/platform/posix/posix_file_system.cc
index fb7a5a9995..9a8021565c 100644
--- a/tensorflow/core/platform/posix/posix_file_system.cc
+++ b/tensorflow/core/platform/posix/posix_file_system.cc
@@ -18,6 +18,9 @@ limitations under the License.
#include <fcntl.h>
#include <stdio.h>
#include <sys/mman.h>
+#if !defined(__APPLE__)
+#include <sys/sendfile.h>
+#endif
#include <sys/stat.h>
#include <sys/time.h>
#include <sys/types.h>
@@ -34,6 +37,9 @@ limitations under the License.
namespace tensorflow {
+// 128KB of copy buffer
+constexpr size_t kPosixCopyFileBufferSize = 128 * 1024;
+
// pread() based random-access
class PosixRandomAccessFile : public RandomAccessFile {
private:
@@ -276,4 +282,70 @@ Status PosixFileSystem::RenameFile(const string& src, const string& target) {
return result;
}
+Status PosixFileSystem::CopyFile(const string& src, const string& target) {
+ string translated_src = TranslateName(src);
+ struct stat sbuf;
+ if (stat(translated_src.c_str(), &sbuf) != 0) {
+ return IOError(src, errno);
+ }
+ int src_fd = open(translated_src.c_str(), O_RDONLY);
+ if (src_fd < 0) {
+ return IOError(src, errno);
+ }
+ string translated_target = TranslateName(target);
+ // O_WRONLY | O_CREAT:
+ // Open file for write and if file does not exist, create the file.
+ // S_IRUSR | S_IWUSR | S_IRGRP | S_IROTH:
+ // Create the file with permission of 0644
+ int target_fd = open(translated_target.c_str(), O_WRONLY | O_CREAT,
+ S_IRUSR | S_IWUSR | S_IRGRP | S_IROTH);
+ if (target_fd < 0) {
+ close(src_fd);
+ return IOError(target, errno);
+ }
+ int rc = 0;
+ off_t offset = 0;
+ std::unique_ptr<char[]> buffer(new char[kPosixCopyFileBufferSize]);
+ while (offset < sbuf.st_size) {
+ // Use uint64 for safe compare SSIZE_MAX
+ uint64 chunk = sbuf.st_size - offset;
+ if (chunk > SSIZE_MAX) {
+ chunk = SSIZE_MAX;
+ }
+#if defined(__linux__) && !defined(__ANDROID__)
+ rc = sendfile(target_fd, src_fd, &offset, static_cast<size_t>(chunk));
+#else
+ if (chunk > kPosixCopyFileBufferSize) {
+ chunk = kPosixCopyFileBufferSize;
+ }
+ rc = read(src_fd, buffer.get(), static_cast<size_t>(chunk));
+ if (rc <= 0) {
+ break;
+ }
+ rc = write(target_fd, buffer.get(), static_cast<size_t>(chunk));
+ offset += chunk;
+#endif
+ if (rc <= 0) {
+ break;
+ }
+ }
+
+ Status result = Status::OK();
+ if (rc < 0) {
+ result = IOError(target, errno);
+ }
+
+ // Keep the error code
+ rc = close(target_fd);
+ if (rc < 0 && result == Status::OK()) {
+ result = IOError(target, errno);
+ }
+ rc = close(src_fd);
+ if (rc < 0 && result == Status::OK()) {
+ result = IOError(target, errno);
+ }
+
+ return result;
+}
+
} // namespace tensorflow
diff --git a/tensorflow/core/platform/posix/posix_file_system.h b/tensorflow/core/platform/posix/posix_file_system.h
index fe050fd5a0..98ffa43b8a 100644
--- a/tensorflow/core/platform/posix/posix_file_system.h
+++ b/tensorflow/core/platform/posix/posix_file_system.h
@@ -56,6 +56,8 @@ class PosixFileSystem : public FileSystem {
Status GetFileSize(const string& fname, uint64* size) override;
Status RenameFile(const string& src, const string& target) override;
+
+ Status CopyFile(const string& src, const string& target) override;
};
Status IOError(const string& context, int err_number);
diff --git a/tensorflow/core/public/version.h b/tensorflow/core/public/version.h
index b02f899b87..50bfa91267 100644
--- a/tensorflow/core/public/version.h
+++ b/tensorflow/core/public/version.h
@@ -19,12 +19,12 @@ limitations under the License.
// TensorFlow uses semantic versioning, see http://semver.org/.
#define TF_MAJOR_VERSION 1
-#define TF_MINOR_VERSION 5
+#define TF_MINOR_VERSION 6
#define TF_PATCH_VERSION 0
// TF_VERSION_SUFFIX is non-empty for pre-releases (e.g. "-alpha", "-alpha.1",
// "-beta", "-rc", "-rc.1")
-#define TF_VERSION_SUFFIX ""
+#define TF_VERSION_SUFFIX "-rc0"
#define TF_STR_HELPER(x) #x
#define TF_STR(x) TF_STR_HELPER(x)
diff --git a/tensorflow/docs_src/api_guides/python/contrib.distributions.md b/tensorflow/docs_src/api_guides/python/contrib.distributions.md
index 7a3d509b75..533d7dac13 100644
--- a/tensorflow/docs_src/api_guides/python/contrib.distributions.md
+++ b/tensorflow/docs_src/api_guides/python/contrib.distributions.md
@@ -17,7 +17,6 @@ initialized with parameters that define the distributions.
* @{tf.contrib.distributions.Binomial}
* @{tf.contrib.distributions.Bernoulli}
-* @{tf.contrib.distributions.BernoulliWithSigmoidProbs}
* @{tf.contrib.distributions.Beta}
* @{tf.contrib.distributions.Categorical}
* @{tf.contrib.distributions.Chi2}
diff --git a/tensorflow/docs_src/api_guides/python/regression_examples.md b/tensorflow/docs_src/api_guides/python/regression_examples.md
index dae50a8f03..7de2be0552 100644
--- a/tensorflow/docs_src/api_guides/python/regression_examples.md
+++ b/tensorflow/docs_src/api_guides/python/regression_examples.md
@@ -38,7 +38,7 @@ The preceding examples rely on the following data set utility:
<tr> <th>Utility</th> <th>Description</th></tr>
<tr>
- <td><a href="../../examples/get_started/regression/imports85.py">imports85.py</a></td>
+ <td><a href="https://www.tensorflow.org/code/tensorflow/examples/get_started/regression/imports85.py">imports85.py</a></td>
<td>This program provides utility functions that load the
<tt>imports85</tt> data set into formats that other TensorFlow
programs (for example, <tt>linear_regression.py</tt> and
diff --git a/tensorflow/docs_src/community/welcome.md b/tensorflow/docs_src/community/welcome.md
index d2d3f9edae..9f6fe91b14 100644
--- a/tensorflow/docs_src/community/welcome.md
+++ b/tensorflow/docs_src/community/welcome.md
@@ -65,5 +65,5 @@ please read the following list carefully:
on GitHub. For example, use the issue tracker to request a
new operation in TensorFlow.
* To report vulnerabilities, please follow our
- [vulnerability disclosure guidelines](https://github.com/tensorflow/tensorflow/blob/master/SECURITY.md).
+ [vulnerability disclosure guidelines](https://github.com/tensorflow/tensorflow/blob/master/tensorflow/SECURITY.md).
diff --git a/tensorflow/docs_src/get_started/get_started_for_beginners.md b/tensorflow/docs_src/get_started/get_started_for_beginners.md
index e680234d4d..b88483be69 100644
--- a/tensorflow/docs_src/get_started/get_started_for_beginners.md
+++ b/tensorflow/docs_src/get_started/get_started_for_beginners.md
@@ -705,7 +705,7 @@ for pred_dict, expec in zip(predictions, expected):
class_id = pred_dict['class_ids'][0]
probability = pred_dict['probabilities'][class_id]
- print(template.format(SPECIES[class_id], 100 * probability, expec))
+ print(template.format(iris_data.SPECIES[class_id], 100 * probability, expec))
```
Running the program yields the following output:
diff --git a/tensorflow/docs_src/install/install_c.md b/tensorflow/docs_src/install/install_c.md
index 14add7c77e..a783205b4a 100644
--- a/tensorflow/docs_src/install/install_c.md
+++ b/tensorflow/docs_src/install/install_c.md
@@ -38,7 +38,7 @@ enable TensorFlow for C:
OS="linux" # Change to "darwin" for macOS
TARGET_DIRECTORY="/usr/local"
curl -L \
- "https://storage.googleapis.com/tensorflow/libtensorflow/libtensorflow-${TF_TYPE}-${OS}-x86_64-1.5.0.tar.gz" |
+ "https://storage.googleapis.com/tensorflow/libtensorflow/libtensorflow-${TF_TYPE}-${OS}-x86_64-1.6.0-rc0.tar.gz" |
sudo tar -C $TARGET_DIRECTORY -xz
The `tar` command extracts the TensorFlow C library into the `lib`
diff --git a/tensorflow/docs_src/install/install_go.md b/tensorflow/docs_src/install/install_go.md
index d2af9d9843..5249e04615 100644
--- a/tensorflow/docs_src/install/install_go.md
+++ b/tensorflow/docs_src/install/install_go.md
@@ -38,7 +38,7 @@ steps to install this library and enable TensorFlow for Go:
TF_TYPE="cpu" # Change to "gpu" for GPU support
TARGET_DIRECTORY='/usr/local'
curl -L \
- "https://storage.googleapis.com/tensorflow/libtensorflow/libtensorflow-${TF_TYPE}-$(go env GOOS)-x86_64-1.5.0.tar.gz" |
+ "https://storage.googleapis.com/tensorflow/libtensorflow/libtensorflow-${TF_TYPE}-$(go env GOOS)-x86_64-1.6.0-rc0.tar.gz" |
sudo tar -C $TARGET_DIRECTORY -xz
The `tar` command extracts the TensorFlow C library into the `lib`
diff --git a/tensorflow/docs_src/install/install_java.md b/tensorflow/docs_src/install/install_java.md
index e5388c4b1e..0c6c773e62 100644
--- a/tensorflow/docs_src/install/install_java.md
+++ b/tensorflow/docs_src/install/install_java.md
@@ -36,7 +36,7 @@ following to the project's `pom.xml` to use the TensorFlow Java APIs:
<dependency>
<groupId>org.tensorflow</groupId>
<artifactId>tensorflow</artifactId>
- <version>1.5.0</version>
+ <version>1.6.0-rc0</version>
</dependency>
```
@@ -65,7 +65,7 @@ As an example, these steps will create a Maven project that uses TensorFlow:
<dependency>
<groupId>org.tensorflow</groupId>
<artifactId>tensorflow</artifactId>
- <version>1.5.0</version>
+ <version>1.6.0-rc0</version>
</dependency>
</dependencies>
</project>
@@ -123,12 +123,12 @@ instead:
<dependency>
<groupId>org.tensorflow</groupId>
<artifactId>libtensorflow</artifactId>
- <version>1.5.0</version>
+ <version>1.6.0-rc0</version>
</dependency>
<dependency>
<groupId>org.tensorflow</groupId>
<artifactId>libtensorflow_jni_gpu</artifactId>
- <version>1.5.0</version>
+ <version>1.6.0-rc0</version>
</dependency>
```
@@ -147,7 +147,7 @@ refer to the simpler instructions above instead.
Take the following steps to install TensorFlow for Java on Linux or macOS:
1. Download
- [libtensorflow.jar](https://storage.googleapis.com/tensorflow/libtensorflow/libtensorflow-1.5.0.jar),
+ [libtensorflow.jar](https://storage.googleapis.com/tensorflow/libtensorflow/libtensorflow-1.6.0-rc0.jar),
which is the TensorFlow Java Archive (JAR).
2. Decide whether you will run TensorFlow for Java on CPU(s) only or with
@@ -166,7 +166,7 @@ Take the following steps to install TensorFlow for Java on Linux or macOS:
OS=$(uname -s | tr '[:upper:]' '[:lower:]')
mkdir -p ./jni
curl -L \
- "https://storage.googleapis.com/tensorflow/libtensorflow/libtensorflow_jni-${TF_TYPE}-${OS}-x86_64-1.5.0.tar.gz" |
+ "https://storage.googleapis.com/tensorflow/libtensorflow/libtensorflow_jni-${TF_TYPE}-${OS}-x86_64-1.6.0-rc0.tar.gz" |
tar -xz -C ./jni
### Install on Windows
@@ -174,10 +174,10 @@ Take the following steps to install TensorFlow for Java on Linux or macOS:
Take the following steps to install TensorFlow for Java on Windows:
1. Download
- [libtensorflow.jar](https://storage.googleapis.com/tensorflow/libtensorflow/libtensorflow-1.5.0.jar),
+ [libtensorflow.jar](https://storage.googleapis.com/tensorflow/libtensorflow/libtensorflow-1.6.0-rc0.jar),
which is the TensorFlow Java Archive (JAR).
2. Download the following Java Native Interface (JNI) file appropriate for
- [TensorFlow for Java on Windows](https://storage.googleapis.com/tensorflow/libtensorflow/libtensorflow_jni-cpu-windows-x86_64-1.5.0.zip).
+ [TensorFlow for Java on Windows](https://storage.googleapis.com/tensorflow/libtensorflow/libtensorflow_jni-cpu-windows-x86_64-1.6.0-rc0.zip).
3. Extract this .zip file.
@@ -225,7 +225,7 @@ must be part of your `classpath`. For example, you can include the
downloaded `.jar` in your `classpath` by using the `-cp` compilation flag
as follows:
-<pre><b>javac -cp libtensorflow-1.5.0.jar HelloTF.java</b></pre>
+<pre><b>javac -cp libtensorflow-1.6.0-rc0.jar HelloTF.java</b></pre>
### Running
@@ -239,11 +239,11 @@ two files are available to the JVM:
For example, the following command line executes the `HelloTF` program on Linux
and macOS X:
-<pre><b>java -cp libtensorflow-1.5.0.jar:. -Djava.library.path=./jni HelloTF</b></pre>
+<pre><b>java -cp libtensorflow-1.6.0-rc0.jar:. -Djava.library.path=./jni HelloTF</b></pre>
And the following command line executes the `HelloTF` program on Windows:
-<pre><b>java -cp libtensorflow-1.5.0.jar;. -Djava.library.path=jni HelloTF</b></pre>
+<pre><b>java -cp libtensorflow-1.6.0-rc0.jar;. -Djava.library.path=jni HelloTF</b></pre>
If the program prints <tt>Hello from <i>version</i></tt>, you've successfully
installed TensorFlow for Java and are ready to use the API. If the program
diff --git a/tensorflow/docs_src/install/install_linux.md b/tensorflow/docs_src/install/install_linux.md
index cd8c14599f..105b225177 100644
--- a/tensorflow/docs_src/install/install_linux.md
+++ b/tensorflow/docs_src/install/install_linux.md
@@ -188,7 +188,7 @@ Take the following steps to install TensorFlow with Virtualenv:
Virtualenv environment:
<pre>(tensorflow)$ <b>pip3 install --upgrade \
- https://storage.googleapis.com/tensorflow/linux/cpu/tensorflow-1.5.0-cp34-cp34m-linux_x86_64.whl</b></pre>
+ https://storage.googleapis.com/tensorflow/linux/cpu/tensorflow-1.6.0rc0-cp34-cp34m-linux_x86_64.whl</b></pre>
If you encounter installation problems, see
[Common Installation Problems](#common_installation_problems).
@@ -293,7 +293,7 @@ take the following steps:
<pre>
$ <b>sudo pip3 install --upgrade \
- https://storage.googleapis.com/tensorflow/linux/cpu/tensorflow-1.5.0-cp34-cp34m-linux_x86_64.whl</b>
+ https://storage.googleapis.com/tensorflow/linux/cpu/tensorflow-1.6.0rc0-cp34-cp34m-linux_x86_64.whl</b>
</pre>
If this step fails, see
@@ -480,7 +480,7 @@ Take the following steps to install TensorFlow in an Anaconda environment:
<pre>
(tensorflow)$ <b>pip install --ignore-installed --upgrade \
- https://storage.googleapis.com/tensorflow/linux/cpu/tensorflow-1.5.0-cp34-cp34m-linux_x86_64.whl</b></pre>
+ https://storage.googleapis.com/tensorflow/linux/cpu/tensorflow-1.6.0rc0-cp34-cp34m-linux_x86_64.whl</b></pre>
<a name="ValidateYourInstallation"></a>
@@ -648,14 +648,14 @@ This section documents the relevant values for Linux installations.
CPU only:
<pre>
-https://storage.googleapis.com/tensorflow/linux/cpu/tensorflow-1.5.0-cp27-none-linux_x86_64.whl
+https://storage.googleapis.com/tensorflow/linux/cpu/tensorflow-1.6.0rc0-cp27-none-linux_x86_64.whl
</pre>
GPU support:
<pre>
-https://storage.googleapis.com/tensorflow/linux/gpu/tensorflow_gpu-1.5.0-cp27-none-linux_x86_64.whl
+https://storage.googleapis.com/tensorflow/linux/gpu/tensorflow_gpu-1.6.0rc0-cp27-none-linux_x86_64.whl
</pre>
Note that GPU support requires the NVIDIA hardware and software described in
@@ -667,14 +667,14 @@ Note that GPU support requires the NVIDIA hardware and software described in
CPU only:
<pre>
-https://storage.googleapis.com/tensorflow/linux/cpu/tensorflow-1.5.0-cp34-cp34m-linux_x86_64.whl
+https://storage.googleapis.com/tensorflow/linux/cpu/tensorflow-1.6.0rc0-cp34-cp34m-linux_x86_64.whl
</pre>
GPU support:
<pre>
-https://storage.googleapis.com/tensorflow/linux/gpu/tensorflow_gpu-1.5.0-cp34-cp34m-linux_x86_64.whl
+https://storage.googleapis.com/tensorflow/linux/gpu/tensorflow_gpu-1.6.0rc0-cp34-cp34m-linux_x86_64.whl
</pre>
Note that GPU support requires the NVIDIA hardware and software described in
@@ -686,14 +686,14 @@ Note that GPU support requires the NVIDIA hardware and software described in
CPU only:
<pre>
-https://storage.googleapis.com/tensorflow/linux/cpu/tensorflow-1.5.0-cp35-cp35m-linux_x86_64.whl
+https://storage.googleapis.com/tensorflow/linux/cpu/tensorflow-1.6.0rc0-cp35-cp35m-linux_x86_64.whl
</pre>
GPU support:
<pre>
-https://storage.googleapis.com/tensorflow/linux/gpu/tensorflow_gpu-1.5.0-cp35-cp35m-linux_x86_64.whl
+https://storage.googleapis.com/tensorflow/linux/gpu/tensorflow_gpu-1.6.0rc0-cp35-cp35m-linux_x86_64.whl
</pre>
@@ -705,14 +705,14 @@ Note that GPU support requires the NVIDIA hardware and software described in
CPU only:
<pre>
-https://storage.googleapis.com/tensorflow/linux/cpu/tensorflow-1.5.0-cp36-cp36m-linux_x86_64.whl
+https://storage.googleapis.com/tensorflow/linux/cpu/tensorflow-1.6.0rc0-cp36-cp36m-linux_x86_64.whl
</pre>
GPU support:
<pre>
-https://storage.googleapis.com/tensorflow/linux/gpu/tensorflow_gpu-1.5.0-cp36-cp36m-linux_x86_64.whl
+https://storage.googleapis.com/tensorflow/linux/gpu/tensorflow_gpu-1.6.0rc0-cp36-cp36m-linux_x86_64.whl
</pre>
diff --git a/tensorflow/docs_src/install/install_mac.md b/tensorflow/docs_src/install/install_mac.md
index f49d3a2f08..a6ea548cfb 100644
--- a/tensorflow/docs_src/install/install_mac.md
+++ b/tensorflow/docs_src/install/install_mac.md
@@ -115,7 +115,7 @@ Take the following steps to install TensorFlow with Virtualenv:
TensorFlow in the active Virtualenv is as follows:
<pre> $ <b>pip3 install --upgrade \
- https://storage.googleapis.com/tensorflow/mac/cpu/tensorflow-1.5.0-py3-none-any.whl</b></pre>
+ https://storage.googleapis.com/tensorflow/mac/cpu/tensorflow-1.6.0rc0-py3-none-any.whl</b></pre>
If you encounter installation problems, see
[Common Installation Problems](#common-installation-problems).
@@ -238,7 +238,7 @@ take the following steps:
issue the following command:
<pre> $ <b>sudo pip3 install --upgrade \
- https://storage.googleapis.com/tensorflow/mac/cpu/tensorflow-1.5.0-py3-none-any.whl</b> </pre>
+ https://storage.googleapis.com/tensorflow/mac/cpu/tensorflow-1.6.0rc0-py3-none-any.whl</b> </pre>
If the preceding command fails, see
[installation problems](#common-installation-problems).
@@ -347,7 +347,7 @@ Take the following steps to install TensorFlow in an Anaconda environment:
TensorFlow for Python 2.7:
<pre> (<i>targetDirectory</i>)$ <b>pip install --ignore-installed --upgrade \
- https://storage.googleapis.com/tensorflow/mac/cpu/tensorflow-1.5.0-py2-none-any.whl</b></pre>
+ https://storage.googleapis.com/tensorflow/mac/cpu/tensorflow-1.6.0rc0-py2-none-any.whl</b></pre>
<a name="ValidateYourInstallation"></a>
@@ -520,7 +520,7 @@ This section documents the relevant values for Mac OS installations.
<pre>
-https://storage.googleapis.com/tensorflow/mac/cpu/tensorflow-1.5.0-py2-none-any.whl
+https://storage.googleapis.com/tensorflow/mac/cpu/tensorflow-1.6.0rc0-py2-none-any.whl
</pre>
@@ -528,5 +528,5 @@ https://storage.googleapis.com/tensorflow/mac/cpu/tensorflow-1.5.0-py2-none-any.
<pre>
-https://storage.googleapis.com/tensorflow/mac/cpu/tensorflow-1.5.0-py3-none-any.whl
+https://storage.googleapis.com/tensorflow/mac/cpu/tensorflow-1.6.0rc0-py3-none-any.whl
</pre>
diff --git a/tensorflow/docs_src/install/install_sources.md b/tensorflow/docs_src/install/install_sources.md
index 175b5c4402..90031b4b5e 100644
--- a/tensorflow/docs_src/install/install_sources.md
+++ b/tensorflow/docs_src/install/install_sources.md
@@ -221,7 +221,7 @@ problem, do either of the following:
* Download Xcode 7.2 and select it as your default by issuing the following
command:
- <pre> $ <b>sudo xcode-select -s /Application/Xcode-7.2/Xcode.app</b></pre>
+ <pre> $ <b>sudo xcode-select -s /Applications/Xcode-7.2/Xcode.app</b></pre>
**NOTE:** Your system must fulfill the NVIDIA software requirements described
in one of the following documents:
@@ -359,10 +359,10 @@ Invoke `pip install` to install that pip package.
The filename of the `.whl` file depends on your platform.
For example, the following command will install the pip package
-for TensorFlow 1.5.0 on Linux:
+for TensorFlow 1.6.0rc0 on Linux:
<pre>
-$ <b>sudo pip install /tmp/tensorflow_pkg/tensorflow-1.5.0-py2-none-any.whl</b>
+$ <b>sudo pip install /tmp/tensorflow_pkg/tensorflow-1.6.0rc0-py2-none-any.whl</b>
</pre>
## Validate your installation
@@ -460,7 +460,8 @@ Stack Overflow and specify the `tensorflow` tag.
**Linux**
<table>
<tr><th>Version:</th><th>CPU/GPU:</th><th>Python Version:</th><th>Compiler:</th><th>Build Tools:</th><th>cuDNN:</th><th>CUDA:</th></tr>
-
+<tr><td>tensorflow-1.6.0rc0</td><td>CPU</td><td>2.7, 3.3-3.6</td><td>GCC 4.8</td><td>Bazel 0.9.0</td><td>N/A</td><td>N/A</td></tr>
+<tr><td>tensorflow_gpu-1.6.0rc0</td><td>GPU</td><td>2.7, 3.3-3.6</td><td>GCC 4.8</td><td>Bazel 0.9.0</td><td>7</td><td>9</td></tr>
<tr><td>tensorflow-1.5.0</td><td>CPU</td><td>2.7, 3.3-3.6</td><td>GCC 4.8</td><td>Bazel 0.8.0</td><td>N/A</td><td>N/A</td></tr>
<tr><td>tensorflow_gpu-1.5.0</td><td>GPU</td><td>2.7, 3.3-3.6</td><td>GCC 4.8</td><td>Bazel 0.8.0</td><td>7</td><td>9</td></tr>
<tr><td>tensorflow-1.4.0</td><td>CPU</td><td>2.7, 3.3-3.6</td><td>GCC 4.8</td><td>Bazel 0.5.4</td><td>N/A</td><td>N/A</td></tr>
@@ -478,6 +479,7 @@ Stack Overflow and specify the `tensorflow` tag.
**Mac**
<table>
<tr><th>Version:</th><th>CPU/GPU:</th><th>Python Version:</th><th>Compiler:</th><th>Build Tools:</th><th>cuDNN:</th><th>CUDA:</th></tr>
+<tr><td>tensorflow-1.6.0rc0</td><td>CPU</td><td>2.7, 3.3-3.6</td><td>Clang from xcode</td><td>Bazel 0.8.1</td><td>N/A</td><td>N/A</td></tr>
<tr><td>tensorflow-1.5.0</td><td>CPU</td><td>2.7, 3.3-3.6</td><td>Clang from xcode</td><td>Bazel 0.8.1</td><td>N/A</td><td>N/A</td></tr>
<tr><td>tensorflow-1.4.0</td><td>CPU</td><td>2.7, 3.3-3.6</td><td>Clang from xcode</td><td>Bazel 0.5.4</td><td>N/A</td><td>N/A</td></tr>
<tr><td>tensorflow-1.3.0</td><td>CPU</td><td>2.7, 3.3-3.6</td><td>Clang from xcode</td><td>Bazel 0.4.5</td><td>N/A</td><td>N/A</td></tr>
@@ -491,6 +493,8 @@ Stack Overflow and specify the `tensorflow` tag.
**Windows**
<table>
<tr><th>Version:</th><th>CPU/GPU:</th><th>Python Version:</th><th>Compiler:</th><th>Build Tools:</th><th>cuDNN:</th><th>CUDA:</th></tr>
+<tr><td>tensorflow-1.6.0rc0</td><td>CPU</td><td>3.5-3.6</td><td>MSVC 2015 update 3</td><td>Cmake v3.6.3</td><td>N/A</td><td>N/A</td></tr>
+<tr><td>tensorflow_gpu-1.6.0rc0</td><td>GPU</td><td>3.5-3.6</td><td>MSVC 2015 update 3</td><td>Cmake v3.6.3</td><td>7</td><td>9</td></tr>
<tr><td>tensorflow-1.5.0</td><td>CPU</td><td>3.5-3.6</td><td>MSVC 2015 update 3</td><td>Cmake v3.6.3</td><td>N/A</td><td>N/A</td></tr>
<tr><td>tensorflow_gpu-1.5.0</td><td>GPU</td><td>3.5-3.6</td><td>MSVC 2015 update 3</td><td>Cmake v3.6.3</td><td>7</td><td>9</td></tr>
<tr><td>tensorflow-1.4.0</td><td>CPU</td><td>3.5-3.6</td><td>MSVC 2015 update 3</td><td>Cmake v3.6.3</td><td>N/A</td><td>N/A</td></tr>
diff --git a/tensorflow/docs_src/programmers_guide/debugger.md b/tensorflow/docs_src/programmers_guide/debugger.md
index dbc4517087..c8fdae6f60 100644
--- a/tensorflow/docs_src/programmers_guide/debugger.md
+++ b/tensorflow/docs_src/programmers_guide/debugger.md
@@ -214,7 +214,7 @@ navigate between these screens by clicking the `<--` and
### Other Features of the tfdbg CLI
In addition to the commands listed above, the tfdbg CLI provides the following
-addditional features:
+additional features:
* To navigate through previous tfdbg commands, type in a few characters
followed by the Up or Down arrow keys. tfdbg will show you the history of
diff --git a/tensorflow/docs_src/programmers_guide/index.md b/tensorflow/docs_src/programmers_guide/index.md
index d45e666ce7..7a5e90081d 100644
--- a/tensorflow/docs_src/programmers_guide/index.md
+++ b/tensorflow/docs_src/programmers_guide/index.md
@@ -13,7 +13,7 @@ works. The units are as follows:
## Low Level APIs
* @{$programmers_guide/low_level_intro}, which introduces the
- basics of how you can to use TensorFlow outside of the high Level APIs.
+ basics of how you can use TensorFlow outside of the high Level APIs.
* @{$programmers_guide/tensors}, which explains how to create,
manipulate, and access Tensors--the fundamental object in TensorFlow.
* @{$programmers_guide/variables}, which details how
diff --git a/tensorflow/examples/image_retraining/retrain.py b/tensorflow/examples/image_retraining/retrain.py
index ec22684eaf..868310cbc0 100644
--- a/tensorflow/examples/image_retraining/retrain.py
+++ b/tensorflow/examples/image_retraining/retrain.py
@@ -344,8 +344,8 @@ def maybe_download_and_extract(data_url):
filepath, _ = urllib.request.urlretrieve(data_url, filepath, _progress)
print()
statinfo = os.stat(filepath)
- tf.logging.info('Successfully downloaded', filename, statinfo.st_size,
- 'bytes.')
+ tf.logging.info('Successfully downloaded %s %d bytes.', filename,
+ statinfo.st_size)
print('Extracting file from ', filepath)
tarfile.open(filepath, 'r:gz').extractall(dest_directory)
else:
diff --git a/tensorflow/java/src/main/java/org/tensorflow/NativeLibrary.java b/tensorflow/java/src/main/java/org/tensorflow/NativeLibrary.java
index 499757e8cf..cf773e1686 100644
--- a/tensorflow/java/src/main/java/org/tensorflow/NativeLibrary.java
+++ b/tensorflow/java/src/main/java/org/tensorflow/NativeLibrary.java
@@ -88,7 +88,7 @@ final class NativeLibrary {
// Deletions are in the reverse order of requests, so we need to request that the directory be
// deleted first, so that it is empty when the request is fulfilled.
tempPath.deleteOnExit();
- final String tempDirectory = tempPath.toString();
+ final String tempDirectory = tempPath.getCanonicalPath();
if (frameworkResource != null) {
extractResource(frameworkResource, frameworkLibName, tempDirectory);
} else {
diff --git a/tensorflow/python/client/session.py b/tensorflow/python/client/session.py
index 6befeb846d..f3c4fecdc0 100644
--- a/tensorflow/python/client/session.py
+++ b/tensorflow/python/client/session.py
@@ -1539,8 +1539,22 @@ class Session(BaseSession):
def __exit__(self, exec_type, exec_value, exec_tb):
if exec_type is errors.OpError:
logging.error('Session closing due to OpError: %s', (exec_value,))
- self._default_session_context_manager.__exit__(exec_type, exec_value,
- exec_tb)
+ try:
+ self._default_session_context_manager.__exit__(exec_type, exec_value,
+ exec_tb)
+ except RuntimeError as error:
+ if error == exec_value:
+ # NOTE(skyewm): for some reason, in Python3,
+ # _default_session_context_manager.__exit__ will re-raise the "not
+ # re-entrant" exception raised in __enter__ above (note that if we're
+ # here, we're in the outer session context manager, since __exit__ is
+ # not called when __enter__ raises an exception). We still want to
+ # continue cleaning up this context manager before the exception is
+ # further propagated, so we ignore it here (note that it'll continue
+ # being propagated after this method completes).
+ pass
+ else:
+ raise
self._default_graph_context_manager.__exit__(exec_type, exec_value, exec_tb)
self._default_session_context_manager = None
diff --git a/tensorflow/python/data/ops/dataset_ops.py b/tensorflow/python/data/ops/dataset_ops.py
index c4b7e4919b..b665443b7a 100644
--- a/tensorflow/python/data/ops/dataset_ops.py
+++ b/tensorflow/python/data/ops/dataset_ops.py
@@ -773,7 +773,7 @@ class Dataset(object):
return PaddedBatchDataset(self, batch_size, padded_shapes, padding_values)
def map(self, map_func, num_parallel_calls=None):
- """Maps `map_func` across this datset.
+ """Maps `map_func` across this dataset.
Args:
map_func: A function mapping a nested structure of tensors (having
diff --git a/tensorflow/python/debug/cli/cli_shared.py b/tensorflow/python/debug/cli/cli_shared.py
index a0fe6066ac..dea019fef5 100644
--- a/tensorflow/python/debug/cli/cli_shared.py
+++ b/tensorflow/python/debug/cli/cli_shared.py
@@ -175,7 +175,7 @@ def format_tensor(tensor,
include_numeric_summary: Whether a text summary of the numeric values (if
applicable) will be included.
write_path: A path to save the tensor value (after any slicing) to
- (optinal). `numpy.save()` is used to save the value.
+ (optional). `numpy.save()` is used to save the value.
Returns:
An instance of `debugger_cli_common.RichTextLines` representing the
diff --git a/tensorflow/python/debug/cli/tensor_format.py b/tensorflow/python/debug/cli/tensor_format.py
index e0759a8bc1..9ba84e3f22 100644
--- a/tensorflow/python/debug/cli/tensor_format.py
+++ b/tensorflow/python/debug/cli/tensor_format.py
@@ -134,7 +134,7 @@ def format_tensor(tensor,
if include_metadata:
lines.append(" dtype: %s" % str(tensor.dtype))
- lines.append(" shape: %s" % str(tensor.shape))
+ lines.append(" shape: %s" % str(tensor.shape).replace("L", ""))
if lines:
lines.append("")
diff --git a/tensorflow/python/framework/test_util.py b/tensorflow/python/framework/test_util.py
index c09e2d8084..1560766fc9 100644
--- a/tensorflow/python/framework/test_util.py
+++ b/tensorflow/python/framework/test_util.py
@@ -691,7 +691,7 @@ class TensorFlowTestCase(googletest.TestCase):
self._tempdir = tempfile.mkdtemp(dir=googletest.GetTempDir())
return self._tempdir
- def _AssertProtoEquals(self, a, b):
+ def _AssertProtoEquals(self, a, b, msg=None):
"""Asserts that a and b are the same proto.
Uses ProtoEq() first, as it returns correct results
@@ -701,11 +701,12 @@ class TensorFlowTestCase(googletest.TestCase):
Args:
a: a proto.
b: another proto.
+ msg: Optional message to report on failure.
"""
if not compare.ProtoEq(a, b):
- compare.assertProtoEqual(self, a, b, normalize_numbers=True)
+ compare.assertProtoEqual(self, a, b, normalize_numbers=True, msg=msg)
- def assertProtoEquals(self, expected_message_maybe_ascii, message):
+ def assertProtoEquals(self, expected_message_maybe_ascii, message, msg=None):
"""Asserts that message is same as parsed expected_message_ascii.
Creates another prototype of message, reads the ascii message into it and
@@ -714,8 +715,9 @@ class TensorFlowTestCase(googletest.TestCase):
Args:
expected_message_maybe_ascii: proto message in original or ascii form.
message: the message to validate.
+ msg: Optional message to report on failure.
"""
-
+ msg = msg if msg else ""
if isinstance(expected_message_maybe_ascii, type(message)):
expected_message = expected_message_maybe_ascii
self._AssertProtoEquals(expected_message, message)
@@ -725,20 +727,21 @@ class TensorFlowTestCase(googletest.TestCase):
expected_message_maybe_ascii,
expected_message,
descriptor_pool=descriptor_pool.Default())
- self._AssertProtoEquals(expected_message, message)
+ self._AssertProtoEquals(expected_message, message, msg=msg)
else:
- assert False, ("Can't compare protos of type %s and %s" %
- (type(expected_message_maybe_ascii), type(message)))
+ assert False, ("Can't compare protos of type %s and %s. %s" %
+ (type(expected_message_maybe_ascii), type(message), msg))
def assertProtoEqualsVersion(
self,
expected,
actual,
producer=versions.GRAPH_DEF_VERSION,
- min_consumer=versions.GRAPH_DEF_VERSION_MIN_CONSUMER):
+ min_consumer=versions.GRAPH_DEF_VERSION_MIN_CONSUMER,
+ msg=None):
expected = "versions { producer: %d min_consumer: %d };\n%s" % (
producer, min_consumer, expected)
- self.assertProtoEquals(expected, actual)
+ self.assertProtoEquals(expected, actual, msg=msg)
def assertStartsWith(self, actual, expected_start, msg=None):
"""Assert that actual.startswith(expected_start) is True.
@@ -1028,7 +1031,7 @@ class TensorFlowTestCase(googletest.TestCase):
"%f != %f +/- %f%s" % (f1, f2, err, " (%s)" % msg
if msg is not None else ""))
- def assertArrayNear(self, farray1, farray2, err):
+ def assertArrayNear(self, farray1, farray2, err, msg=None):
"""Asserts that two float arrays are near each other.
Checks that for all elements of farray1 and farray2
@@ -1038,23 +1041,25 @@ class TensorFlowTestCase(googletest.TestCase):
farray1: a list of float values.
farray2: a list of float values.
err: a float value.
+ msg: Optional message to report on failure.
"""
- self.assertEqual(len(farray1), len(farray2))
+ self.assertEqual(len(farray1), len(farray2), msg=msg)
for f1, f2 in zip(farray1, farray2):
- self.assertNear(float(f1), float(f2), err)
+ self.assertNear(float(f1), float(f2), err, msg=msg)
def _NDArrayNear(self, ndarray1, ndarray2, err):
return np.linalg.norm(ndarray1 - ndarray2) < err
- def assertNDArrayNear(self, ndarray1, ndarray2, err):
+ def assertNDArrayNear(self, ndarray1, ndarray2, err, msg=None):
"""Asserts that two numpy arrays have near values.
Args:
ndarray1: a numpy ndarray.
ndarray2: a numpy ndarray.
err: a float. The maximum absolute difference allowed.
+ msg: Optional message to report on failure.
"""
- self.assertTrue(self._NDArrayNear(ndarray1, ndarray2, err))
+ self.assertTrue(self._NDArrayNear(ndarray1, ndarray2, err), msg=msg)
def _GetNdArray(self, a):
if not isinstance(a, np.ndarray):
@@ -1096,9 +1101,16 @@ class TensorFlowTestCase(googletest.TestCase):
np.testing.assert_allclose(
a, b, rtol=rtol, atol=atol, err_msg=msg, equal_nan=True)
- def _assertAllCloseRecursive(self, a, b, rtol=1e-6, atol=1e-6, path=None):
+ def _assertAllCloseRecursive(self,
+ a,
+ b,
+ rtol=1e-6,
+ atol=1e-6,
+ path=None,
+ msg=None):
path = path or []
path_str = (("[" + "][".join([str(p) for p in path]) + "]") if path else "")
+ msg = msg if msg else ""
# Check if a and/or b are namedtuples.
if hasattr(a, "_asdict"):
@@ -1107,18 +1119,18 @@ class TensorFlowTestCase(googletest.TestCase):
b = b._asdict()
a_is_dict = isinstance(a, dict)
if a_is_dict != isinstance(b, dict):
- raise ValueError("Can't compare dict to non-dict, a%s vs b%s." %
- (path_str, path_str))
+ raise ValueError("Can't compare dict to non-dict, a%s vs b%s. %s" %
+ (path_str, path_str, msg))
if a_is_dict:
self.assertItemsEqual(
a.keys(),
b.keys(),
- msg="mismatched keys: a%s has keys %s, but b%s has keys %s" %
- (path_str, a.keys(), path_str, b.keys()))
+ msg="mismatched keys: a%s has keys %s, but b%s has keys %s. %s" %
+ (path_str, a.keys(), path_str, b.keys(), msg))
for k in a:
path.append(k)
self._assertAllCloseRecursive(
- a[k], b[k], rtol=rtol, atol=atol, path=path)
+ a[k], b[k], rtol=rtol, atol=atol, path=path, msg=msg)
del path[-1]
elif isinstance(a, (list, tuple)):
# Try to directly compare a, b as ndarrays; if not work, then traverse
@@ -1131,17 +1143,17 @@ class TensorFlowTestCase(googletest.TestCase):
b_as_ndarray,
rtol=rtol,
atol=atol,
- msg="Mismatched value: a%s is different from b%s." % (path_str,
- path_str))
+ msg="Mismatched value: a%s is different from b%s. %s" %
+ (path_str, path_str, msg))
except (ValueError, TypeError) as e:
if len(a) != len(b):
raise ValueError(
- "Mismatched length: a%s has %d items, but b%s has %d items" %
- (path_str, len(a), path_str, len(b)))
+ "Mismatched length: a%s has %d items, but b%s has %d items. %s" %
+ (path_str, len(a), path_str, len(b), msg))
for idx, (a_ele, b_ele) in enumerate(zip(a, b)):
path.append(str(idx))
self._assertAllCloseRecursive(
- a_ele, b_ele, rtol=rtol, atol=atol, path=path)
+ a_ele, b_ele, rtol=rtol, atol=atol, path=path, msg=msg)
del path[-1]
# a and b are ndarray like objects
else:
@@ -1159,7 +1171,7 @@ class TensorFlowTestCase(googletest.TestCase):
e.args = ((e.args[0] + ' : ' + msg,) + e.args[1:])
raise
- def assertAllClose(self, a, b, rtol=1e-6, atol=1e-6):
+ def assertAllClose(self, a, b, rtol=1e-6, atol=1e-6, msg=None):
"""Asserts that two structures of numpy arrays, have near values.
`a` and `b` can be arbitrarily nested structures. A layer of a nested
@@ -1172,6 +1184,7 @@ class TensorFlowTestCase(googletest.TestCase):
numpy `ndarray`, or any arbitrarily nested of structure of these.
rtol: relative tolerance.
atol: absolute tolerance.
+ msg: Optional message to report on failure.
Raises:
ValueError: if only one of `a[p]` and `b[p]` is a dict or
@@ -1179,7 +1192,7 @@ class TensorFlowTestCase(googletest.TestCase):
to the nested structure, e.g. given `a = [(1, 1), {'d': (6, 7)}]` and
`[p] = [1]['d']`, then `a[p] = (6, 7)`.
"""
- self._assertAllCloseRecursive(a, b, rtol=rtol, atol=atol)
+ self._assertAllCloseRecursive(a, b, rtol=rtol, atol=atol, msg=msg)
def assertAllCloseAccordingToType(self,
a,
@@ -1191,7 +1204,8 @@ class TensorFlowTestCase(googletest.TestCase):
half_rtol=1e-3,
half_atol=1e-3,
bfloat16_rtol=1e-2,
- bfloat16_atol=1e-2):
+ bfloat16_atol=1e-2,
+ msg=None):
"""Like assertAllClose, but also suitable for comparing fp16 arrays.
In particular, the tolerance is reduced to 1e-3 if at least
@@ -1208,6 +1222,7 @@ class TensorFlowTestCase(googletest.TestCase):
half_atol: absolute tolerance for float16.
bfloat16_rtol: relative tolerance for bfloat16.
bfloat16_atol: absolute tolerance for bfloat16.
+ msg: Optional message to report on failure.
"""
a = self._GetNdArray(a)
b = self._GetNdArray(b)
@@ -1224,19 +1239,21 @@ class TensorFlowTestCase(googletest.TestCase):
rtol = max(rtol, bfloat16_rtol)
atol = max(atol, bfloat16_atol)
- self.assertAllClose(a, b, rtol=rtol, atol=atol)
+ self.assertAllClose(a, b, rtol=rtol, atol=atol, msg=msg)
- def assertAllEqual(self, a, b):
+ def assertAllEqual(self, a, b, msg=None):
"""Asserts that two numpy arrays have the same values.
Args:
a: the expected numpy ndarray or anything can be converted to one.
b: the actual numpy ndarray or anything can be converted to one.
+ msg: Optional message to report on failure.
"""
+ msg = msg if msg else ""
a = self._GetNdArray(a)
b = self._GetNdArray(b)
- self.assertEqual(a.shape, b.shape, "Shape mismatch: expected %s, got %s." %
- (a.shape, b.shape))
+ self.assertEqual(a.shape, b.shape, "Shape mismatch: expected %s, got %s."
+ " %s" % (a.shape, b.shape, msg))
same = (a == b)
if a.dtype == np.float32 or a.dtype == np.float64:
@@ -1253,7 +1270,7 @@ class TensorFlowTestCase(googletest.TestCase):
x, y = a, b
print("not equal lhs = ", x)
print("not equal rhs = ", y)
- np.testing.assert_array_equal(a, b)
+ np.testing.assert_array_equal(a, b, err_msg=msg)
# pylint: disable=g-doc-return-or-yield
@contextlib.contextmanager
@@ -1303,12 +1320,13 @@ class TensorFlowTestCase(googletest.TestCase):
return self.assertRaisesWithPredicateMatch(errors.OpError,
expected_err_re_or_predicate)
- def assertShapeEqual(self, np_array, tf_tensor):
+ def assertShapeEqual(self, np_array, tf_tensor, msg=None):
"""Asserts that a Numpy ndarray and a TensorFlow tensor have the same shape.
Args:
np_array: A Numpy ndarray or Numpy scalar.
tf_tensor: A Tensor.
+ msg: Optional message to report on failure.
Raises:
TypeError: If the arguments have the wrong type.
@@ -1317,19 +1335,21 @@ class TensorFlowTestCase(googletest.TestCase):
raise TypeError("np_array must be a Numpy ndarray or Numpy scalar")
if not isinstance(tf_tensor, ops.Tensor):
raise TypeError("tf_tensor must be a Tensor")
- self.assertAllEqual(np_array.shape, tf_tensor.get_shape().as_list())
+ self.assertAllEqual(
+ np_array.shape, tf_tensor.get_shape().as_list(), msg=msg)
- def assertDeviceEqual(self, device1, device2):
+ def assertDeviceEqual(self, device1, device2, msg=None):
"""Asserts that the two given devices are the same.
Args:
device1: A string device name or TensorFlow `DeviceSpec` object.
device2: A string device name or TensorFlow `DeviceSpec` object.
+ msg: Optional message to report on failure.
"""
device1 = pydev.canonical_name(device1)
device2 = pydev.canonical_name(device2)
- self.assertEqual(device1, device2,
- "Devices %s and %s are not equal" % (device1, device2))
+ self.assertEqual(device1, device2, "Devices %s and %s are not equal. %s" %
+ (device1, device2, msg))
# Fix Python 3 compatibility issues
if six.PY3:
diff --git a/tensorflow/python/keras/_impl/keras/backend.py b/tensorflow/python/keras/_impl/keras/backend.py
index 1fa264660d..a238a3f748 100644
--- a/tensorflow/python/keras/_impl/keras/backend.py
+++ b/tensorflow/python/keras/_impl/keras/backend.py
@@ -2895,6 +2895,7 @@ def rnn(step_function,
ndim = len(inputs.get_shape())
if ndim < 3:
raise ValueError('Input should be at least 3D.')
+ inputs_shape = inputs.get_shape()
axes = [1, 0] + list(range(2, ndim))
inputs = array_ops.transpose(inputs, (axes))
@@ -3079,6 +3080,13 @@ def rnn(step_function,
axes = [1, 0] + list(range(2, len(outputs.get_shape())))
outputs = array_ops.transpose(outputs, axes)
+
+ # Static shape inference: (samples, time, ...)
+ outputs_shape = outputs.get_shape().as_list()
+ outputs_shape[0] = inputs_shape[0]
+ outputs_shape[1] = inputs_shape[1]
+ outputs.set_shape(outputs_shape)
+
last_output._uses_learning_phase = uses_learning_phase
return last_output, outputs, new_states
diff --git a/tensorflow/python/keras/_impl/keras/backend_test.py b/tensorflow/python/keras/_impl/keras/backend_test.py
index 27833e368d..f29ca49378 100644
--- a/tensorflow/python/keras/_impl/keras/backend_test.py
+++ b/tensorflow/python/keras/_impl/keras/backend_test.py
@@ -915,6 +915,15 @@ class BackendNNOpsTest(test.TestCase):
last_output, outputs, new_states = keras.backend.rnn(rnn_fn, inputs,
initial_states,
**kwargs)
+ # check static shape inference
+ self.assertEquals(last_output.get_shape().as_list(),
+ [num_samples, output_dim])
+ self.assertEquals(outputs.get_shape().as_list(),
+ [num_samples, timesteps, output_dim])
+ for state in new_states:
+ self.assertEquals(state.get_shape().as_list(),
+ [num_samples, output_dim])
+
last_output_list[i].append(keras.backend.eval(last_output))
outputs_list[i].append(keras.backend.eval(outputs))
self.assertEqual(len(new_states), 1)
diff --git a/tensorflow/python/keras/_impl/keras/layers/lstm_test.py b/tensorflow/python/keras/_impl/keras/layers/lstm_test.py
index 8d359bf17c..1de5485179 100644
--- a/tensorflow/python/keras/_impl/keras/layers/lstm_test.py
+++ b/tensorflow/python/keras/_impl/keras/layers/lstm_test.py
@@ -39,6 +39,22 @@ class LSTMLayerTest(test.TestCase):
'return_sequences': True},
input_shape=(num_samples, timesteps, embedding_dim))
+ def test_static_shape_inference_LSTM(self):
+ # Github issue: 15165
+ num_samples = 2
+ timesteps = 3
+ embedding_dim = 4
+ units = 2
+
+ model = keras.models.Sequential()
+ inputs = keras.layers.Dense(
+ embedding_dim, input_shape=(timesteps, embedding_dim))
+ model.add(inputs)
+ layer = keras.layers.LSTM(units, return_sequences=True)
+ model.add(layer)
+ outputs = model.layers[-1].output
+ self.assertEquals(outputs.get_shape().as_list(), [None, timesteps, units])
+
def test_dynamic_behavior_LSTM(self):
num_samples = 2
timesteps = 3
diff --git a/tensorflow/python/keras/_impl/keras/layers/recurrent.py b/tensorflow/python/keras/_impl/keras/layers/recurrent.py
index b34b92c763..2e9003f52d 100644
--- a/tensorflow/python/keras/_impl/keras/layers/recurrent.py
+++ b/tensorflow/python/keras/_impl/keras/layers/recurrent.py
@@ -19,6 +19,7 @@ from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
+import numbers
import numpy as np
from tensorflow.python.framework import tensor_shape
@@ -413,7 +414,7 @@ class RNN(Layer):
@property
def states(self):
if self._states is None:
- if isinstance(self.cell.state_size, int):
+ if isinstance(self.cell.state_size, numbers.Integral):
num_states = 1
else:
num_states = len(self.cell.state_size)
diff --git a/tensorflow/python/kernel_tests/conv2d_transpose_test.py b/tensorflow/python/kernel_tests/conv2d_transpose_test.py
index 1a65c3f429..b692d3da60 100644
--- a/tensorflow/python/kernel_tests/conv2d_transpose_test.py
+++ b/tensorflow/python/kernel_tests/conv2d_transpose_test.py
@@ -174,7 +174,7 @@ class Conv2DTransposeTest(test.TestCase):
self.assertLess(err, err_tolerance)
def testConv2DTransposeSingleStrideNCHW(self):
- # `NCHW` data fomat is only supported for CUDA device.
+ # `NCHW` data format is only supported for CUDA device.
if test.is_gpu_available(cuda_only=True):
with self.test_session(use_gpu=True):
strides = [1, 1, 1, 1]
@@ -209,7 +209,7 @@ class Conv2DTransposeTest(test.TestCase):
self.assertAllClose(target, value[n, k, h, w])
def testConv2DTransposeSameNCHW(self):
- # `NCHW` data fomat is only supported for CUDA device.
+ # `NCHW` data format is only supported for CUDA device.
if test.is_gpu_available(cuda_only=True):
with self.test_session(use_gpu=True):
strides = [1, 1, 2, 2]
@@ -245,7 +245,7 @@ class Conv2DTransposeTest(test.TestCase):
self.assertAllClose(target, value[n, k, h, w])
def testConv2DTransposeValidNCHW(self):
- # `NCHW` data fomat is only supported for CUDA device.
+ # `NCHW` data format is only supported for CUDA device.
if test.is_gpu_available(cuda_only=True):
with self.test_session(use_gpu=True):
strides = [1, 1, 2, 2]
diff --git a/tensorflow/python/kernel_tests/distributions/bernoulli_test.py b/tensorflow/python/kernel_tests/distributions/bernoulli_test.py
index a269d72273..09812db816 100644
--- a/tensorflow/python/kernel_tests/distributions/bernoulli_test.py
+++ b/tensorflow/python/kernel_tests/distributions/bernoulli_test.py
@@ -25,7 +25,6 @@ import numpy as np
from tensorflow.python.framework import constant_op
from tensorflow.python.framework import dtypes
from tensorflow.python.ops import array_ops
-from tensorflow.python.ops import math_ops
from tensorflow.python.ops.distributions import bernoulli
from tensorflow.python.ops.distributions import kullback_leibler
from tensorflow.python.platform import test
@@ -291,12 +290,6 @@ class BernoulliTest(test.TestCase):
[np.sqrt(var(0.5)), np.sqrt(var(0.4))]],
dtype=np.float32))
- def testBernoulliWithSigmoidProbs(self):
- p = np.array([8.3, 4.2])
- dist = bernoulli.BernoulliWithSigmoidProbs(logits=p)
- with self.test_session():
- self.assertAllClose(math_ops.sigmoid(p).eval(), dist.probs.eval())
-
def testBernoulliBernoulliKL(self):
with self.test_session() as sess:
batch_size = 6
diff --git a/tensorflow/python/kernel_tests/distributions/beta_test.py b/tensorflow/python/kernel_tests/distributions/beta_test.py
index 91a451f033..ab5041a6eb 100644
--- a/tensorflow/python/kernel_tests/distributions/beta_test.py
+++ b/tensorflow/python/kernel_tests/distributions/beta_test.py
@@ -107,8 +107,10 @@ class BetaTest(test.TestCase):
dist.prob([-1., 0.1, 0.5]).eval()
with self.assertRaisesOpError("sample must be positive"):
dist.prob([0., 0.1, 0.5]).eval()
- with self.assertRaisesOpError("sample must be no larger than `1`"):
+ with self.assertRaisesOpError("sample must be less than `1`"):
dist.prob([.1, .2, 1.2]).eval()
+ with self.assertRaisesOpError("sample must be less than `1`"):
+ dist.prob([.1, .2, 1.0]).eval()
def testPdfTwoBatches(self):
with self.test_session():
diff --git a/tensorflow/python/kernel_tests/neon_depthwise_conv_op_test.py b/tensorflow/python/kernel_tests/neon_depthwise_conv_op_test.py
index 30795eed8a..d8ce9fffbd 100644
--- a/tensorflow/python/kernel_tests/neon_depthwise_conv_op_test.py
+++ b/tensorflow/python/kernel_tests/neon_depthwise_conv_op_test.py
@@ -148,7 +148,7 @@ class DepthwiseConv2DTest(test.TestCase):
print("depthwise conv_2d: ", tensor_in_sizes, "*", filter_in_sizes,
", stride:", stride, ", padding: ", padding, ", max diff: ",
np.amax(np.absolute(native_result - interface_result)))
- self.assertArrayNear(
+ self.assertAllClose(
np.ravel(native_result), np.ravel(interface_result), 1e-5)
self.assertShapeEqual(native_result, conv_native)
self.assertShapeEqual(native_result, conv_interface)
@@ -213,7 +213,7 @@ class DepthwiseConv2DTest(test.TestCase):
t1, t2, strides=[1, stride, stride, 1], padding=padding)
value = sess.run(conv)
print("value = ", value)
- self.assertArrayNear(expected, np.ravel(value), 1e-5)
+ self.assertAllClose(expected, np.ravel(value), 1e-5)
self.assertShapeEqual(value, conv)
def testConv2D2x2Filter(self):
diff --git a/tensorflow/python/kernel_tests/py_func_test.py b/tensorflow/python/kernel_tests/py_func_test.py
index c7181497d8..61fb3f12e4 100644
--- a/tensorflow/python/kernel_tests/py_func_test.py
+++ b/tensorflow/python/kernel_tests/py_func_test.py
@@ -1,3 +1,4 @@
+# -*- coding: utf-8 -*-
# Copyright 2015 The TensorFlow Authors. All Rights Reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
@@ -212,6 +213,16 @@ class PyFuncTest(test.TestCase):
value.op.run()
self.assertAllEqual(np_array, [1.0, 2.0])
+ def testReturnUnicodeString(self):
+ with self.test_session():
+ correct = u"你好 世界"
+
+ def unicode_string():
+ return correct
+
+ z, = script_ops.py_func(unicode_string, [], [dtypes.string])
+ self.assertEqual(z.eval(), correct.encode("utf8"))
+
def testBadNumpyReturnType(self):
with self.test_session():
diff --git a/tensorflow/python/layers/maxout.py b/tensorflow/python/layers/maxout.py
index 20ce6c9770..765a1c4fda 100644
--- a/tensorflow/python/layers/maxout.py
+++ b/tensorflow/python/layers/maxout.py
@@ -106,6 +106,6 @@ class MaxOut(base.Layer):
if shape[i] is None:
shape[i] = gen_array_ops.shape(inputs)[i]
outputs = math_ops.reduce_max(
- gen_array_ops.reshape(inputs, shape), -1, keep_dims=False)
+ gen_array_ops.reshape(inputs, shape), -1, keepdims=False)
return outputs
diff --git a/tensorflow/python/lib/io/file_io.i b/tensorflow/python/lib/io/file_io.i
index c0c4e035fc..891a7b0fd0 100644
--- a/tensorflow/python/lib/io/file_io.i
+++ b/tensorflow/python/lib/io/file_io.i
@@ -110,21 +110,15 @@ void RecursivelyCreateDir(const string& dirname, TF_Status* out_status) {
}
}
-void CopyFile(const string& oldpath, const string& newpath, bool overwrite,
+void CopyFile(const string& src, const string& target, bool overwrite,
TF_Status* out_status) {
- // If overwrite is false and the newpath file exists then it's an error.
- if (!overwrite && tensorflow::Env::Default()->FileExists(newpath).ok()) {
+ // If overwrite is false and the target file exists then its an error.
+ if (!overwrite && tensorflow::Env::Default()->FileExists(target).ok()) {
TF_SetStatus(out_status, TF_ALREADY_EXISTS, "file already exists");
return;
}
- string file_content;
- tensorflow::Status status = ReadFileToString(tensorflow::Env::Default(),
- oldpath, &file_content);
- if (!status.ok()) {
- Set_TF_Status_from_Status(out_status, status);
- return;
- }
- status = WriteStringToFile(tensorflow::Env::Default(), newpath, file_content);
+ tensorflow::Status status =
+ tensorflow::Env::Default()->CopyFile(src, target);
if (!status.ok()) {
Set_TF_Status_from_Status(out_status, status);
}
diff --git a/tensorflow/python/ops/bitwise_ops_test.py b/tensorflow/python/ops/bitwise_ops_test.py
index f9b025b787..0a2af3716b 100644
--- a/tensorflow/python/ops/bitwise_ops_test.py
+++ b/tensorflow/python/ops/bitwise_ops_test.py
@@ -70,8 +70,10 @@ class BitwiseOpTest(test_util.TensorFlowTestCase):
self.assertAllEqual(truth, popcnt_result)
def testInvertOp(self):
- dtype_list = [dtypes.int8, dtypes.int16, dtypes.int32, dtypes.int64,
- dtypes.uint8, dtypes.uint16]
+ dtype_list = [
+ dtypes.int8, dtypes.int16, dtypes.int32, dtypes.int64, dtypes.uint8,
+ dtypes.uint16, dtypes.uint32, dtypes.uint64
+ ]
inputs = [0, 5, 3, 14]
with self.test_session(use_gpu=True) as sess:
for dtype in dtype_list:
diff --git a/tensorflow/python/ops/distributions/bernoulli.py b/tensorflow/python/ops/distributions/bernoulli.py
index 4c16d62e9a..68aaf3815e 100644
--- a/tensorflow/python/ops/distributions/bernoulli.py
+++ b/tensorflow/python/ops/distributions/bernoulli.py
@@ -157,26 +157,6 @@ class Bernoulli(distribution.Distribution):
return math_ops.cast(self.probs > 0.5, self.dtype)
-class BernoulliWithSigmoidProbs(Bernoulli):
- """Bernoulli with `probs = nn.sigmoid(logits)`."""
-
- def __init__(self,
- logits=None,
- dtype=dtypes.int32,
- validate_args=False,
- allow_nan_stats=True,
- name="BernoulliWithSigmoidProbs"):
- parameters = locals()
- with ops.name_scope(name):
- super(BernoulliWithSigmoidProbs, self).__init__(
- probs=nn.sigmoid(logits, name="sigmoid_probs"),
- dtype=dtype,
- validate_args=validate_args,
- allow_nan_stats=allow_nan_stats,
- name=name)
- self._parameters = parameters
-
-
@kullback_leibler.RegisterKL(Bernoulli, Bernoulli)
def _kl_bernoulli_bernoulli(a, b, name=None):
"""Calculate the batched KL divergence KL(a || b) with a and b Bernoulli.
diff --git a/tensorflow/python/ops/distributions/beta.py b/tensorflow/python/ops/distributions/beta.py
index 6d6b40b045..469bcadb8e 100644
--- a/tensorflow/python/ops/distributions/beta.py
+++ b/tensorflow/python/ops/distributions/beta.py
@@ -304,12 +304,11 @@ class Beta(distribution.Distribution):
if not self.validate_args:
return x
return control_flow_ops.with_dependencies([
- check_ops.assert_positive(
- x,
- message="sample must be positive"),
+ check_ops.assert_positive(x, message="sample must be positive"),
check_ops.assert_less(
- x, array_ops.ones([], self.dtype),
- message="sample must be no larger than `1`."),
+ x,
+ array_ops.ones([], self.dtype),
+ message="sample must be less than `1`."),
], x)
diff --git a/tensorflow/python/ops/image_ops_test.py b/tensorflow/python/ops/image_ops_test.py
index b12bd3d5b0..18625293e0 100644
--- a/tensorflow/python/ops/image_ops_test.py
+++ b/tensorflow/python/ops/image_ops_test.py
@@ -1869,8 +1869,8 @@ class SelectDistortedCropBoxTest(test_util.TensorFlowTestCase):
image_size = constant_op.constant(
[40, 50, 1], shape=[3], dtype=dtypes.int32)
bounding_box = constant_op.constant(
- [0.0, 0.0, 1.0, 1.0],
- shape=[4],
+ [[[0.0, 0.0, 1.0, 1.0]]],
+ shape=[1, 1, 4],
dtype=dtypes.float32,
)
begin, end, bbox_for_drawing = image_ops.sample_distorted_bounding_box(
@@ -1884,6 +1884,10 @@ class SelectDistortedCropBoxTest(test_util.TensorFlowTestCase):
self.assertAllEqual([3], begin.get_shape().as_list())
self.assertAllEqual([3], end.get_shape().as_list())
self.assertAllEqual([1, 1, 4], bbox_for_drawing.get_shape().as_list())
+ # Actual run to make sure shape is correct inside Compute().
+ begin = begin.eval()
+ end = end.eval()
+ bbox_for_drawing = bbox_for_drawing.eval()
begin, end, bbox_for_drawing = image_ops.sample_distorted_bounding_box(
image_size=image_size,
@@ -1903,8 +1907,8 @@ class SelectDistortedCropBoxTest(test_util.TensorFlowTestCase):
image_size = constant_op.constant(
[40, 50, 1], shape=[3], dtype=dtypes.int32)
bounding_box = constant_op.constant(
- [0.0, 0.0, 1.0, 1.0],
- shape=[4],
+ [[[0.0, 0.0, 1.0, 1.0]]],
+ shape=[1, 1, 4],
dtype=dtypes.float32,
)
begin, end, bbox_for_drawing = image_ops.sample_distorted_bounding_box(
@@ -1916,6 +1920,10 @@ class SelectDistortedCropBoxTest(test_util.TensorFlowTestCase):
self.assertAllEqual([3], begin.get_shape().as_list())
self.assertAllEqual([3], end.get_shape().as_list())
self.assertAllEqual([1, 1, 4], bbox_for_drawing.get_shape().as_list())
+ # Actual run to make sure shape is correct inside Compute().
+ begin = begin.eval()
+ end = end.eval()
+ bbox_for_drawing = bbox_for_drawing.eval()
class ResizeImagesTest(test_util.TensorFlowTestCase):
@@ -2822,20 +2830,9 @@ class PngTest(test_util.TensorFlowTestCase):
class GifTest(test_util.TensorFlowTestCase):
- def testOptimizedGifErrorString(self):
- filename = "tensorflow/core/lib/gif/testdata/optimized.gif"
-
- with self.test_session(use_gpu=True) as sess:
- gif = io_ops.read_file(filename)
- image = image_ops.decode_gif(gif)
- with self.assertRaisesRegexp(errors.InvalidArgumentError,
- "can't process optimized gif"):
- gif, image = sess.run([gif, image])
-
- def testValid(self):
+ def _testValid(self, filename):
# Read some real GIFs
prefix = "tensorflow/core/lib/gif/testdata/"
- filename = "scan.gif"
WIDTH = 20
HEIGHT = 40
STRIDE = 5
@@ -2862,16 +2859,9 @@ class GifTest(test_util.TensorFlowTestCase):
self.assertAllClose(frame, gt)
- def testInValid(self):
- # Read some real GIFs
- prefix = "tensorflow/core/lib/gif/testdata/"
- filename = "optimized.gif"
-
- with self.test_session(use_gpu=True) as sess:
- gif0 = io_ops.read_file(prefix + filename)
- image0 = image_ops.decode_gif(gif0)
- with self.assertRaises(errors.InvalidArgumentError):
- gif0, image0 = sess.run([gif0, image0])
+ def testValid(self):
+ self._testValid("scan.gif")
+ self._testValid("optimized.gif")
def testShape(self):
with self.test_session(use_gpu=True) as sess:
diff --git a/tensorflow/python/ops/linalg/linear_operator.py b/tensorflow/python/ops/linalg/linear_operator.py
index 2c24a29567..957a795918 100644
--- a/tensorflow/python/ops/linalg/linear_operator.py
+++ b/tensorflow/python/ops/linalg/linear_operator.py
@@ -480,7 +480,6 @@ class LinearOperator(object):
cond,
self._max_condition_number_to_be_non_singular(),
message="Singular matrix up to precision epsilon.")
- raise NotImplementedError("assert_non_singular is not implemented.")
def _max_condition_number_to_be_non_singular(self):
"""Return the maximum condition number that we consider nonsingular."""
diff --git a/tensorflow/python/ops/losses/losses_impl.py b/tensorflow/python/ops/losses/losses_impl.py
index 5222333d7e..ca408988dd 100644
--- a/tensorflow/python/ops/losses/losses_impl.py
+++ b/tensorflow/python/ops/losses/losses_impl.py
@@ -726,9 +726,11 @@ def softmax_cross_entropy(
smooth_negatives = label_smoothing / num_classes
onehot_labels = onehot_labels * smooth_positives + smooth_negatives
- losses = nn.softmax_cross_entropy_with_logits(labels=onehot_labels,
- logits=logits,
- name="xentropy")
+ onehot_labels = array_ops.stop_gradient(
+ onehot_labels, name="labels_stop_gradient")
+ losses = nn.softmax_cross_entropy_with_logits_v2(
+ labels=onehot_labels, logits=logits, name="xentropy")
+
return compute_weighted_loss(
losses, weights, scope, loss_collection, reduction=reduction)
diff --git a/tensorflow/python/ops/script_ops.py b/tensorflow/python/ops/script_ops.py
index c7e8c28efd..6fe2f61016 100644
--- a/tensorflow/python/ops/script_ops.py
+++ b/tensorflow/python/ops/script_ops.py
@@ -98,7 +98,7 @@ class FuncRegistry(object):
components of a tensor have different lengths. This is bad: ignoring the
padding is wrong for text data, and removing the padding is wrong for binary
data. To avoid this bug, we redo the conversion using an object dtype.
- Additionally, we convert unicode strings to (byte-)strings for Python3
+ Additionally, we convert unicode strings to (byte-)strings for
compatibility.
Args:
@@ -112,7 +112,7 @@ class FuncRegistry(object):
if result.dtype.char == "S" and result is not value:
return np.asarray(value, order="C", dtype=object)
elif result.dtype.char == "U" and result is not value:
- value = np.vectorize(lambda x: x.encode())(value)
+ value = np.vectorize(lambda x: x.encode("utf8"))(value)
return np.asarray(value, order="C", dtype=object)
elif result.dtype.char == "U":
return result.astype(np.bytes_)
diff --git a/tensorflow/python/ops/standard_ops.py b/tensorflow/python/ops/standard_ops.py
index a2164f78b8..f6d9111009 100644
--- a/tensorflow/python/ops/standard_ops.py
+++ b/tensorflow/python/ops/standard_ops.py
@@ -27,6 +27,7 @@ from tensorflow.python.ops import array_grad
from tensorflow.python.ops import data_flow_grad
from tensorflow.python.ops import manip_grad
from tensorflow.python.ops import math_grad
+from tensorflow.python.ops import manip_grad
from tensorflow.python.ops import sparse_grad
from tensorflow.python.ops import spectral_grad
from tensorflow.python.ops import state_grad
diff --git a/tensorflow/python/ops/state_ops.py b/tensorflow/python/ops/state_ops.py
index 1323df5a17..6c0a090d16 100644
--- a/tensorflow/python/ops/state_ops.py
+++ b/tensorflow/python/ops/state_ops.py
@@ -278,7 +278,7 @@ def assign(ref, value, validate_shape=None, use_locking=None, name=None):
return gen_state_ops.assign(
ref, value, use_locking=use_locking, name=name,
validate_shape=validate_shape)
- return ref.assign(value)
+ return ref.assign(value, name=name)
@tf_export("count_up_to")
diff --git a/tensorflow/python/util/compat_internal.py b/tensorflow/python/util/compat_internal.py
index fee1d6fab7..1905c3e383 100644
--- a/tensorflow/python/util/compat_internal.py
+++ b/tensorflow/python/util/compat_internal.py
@@ -18,6 +18,8 @@ from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
+from tensorflow.python.util.compat import as_str_any
+
def path_to_str(path):
"""Returns the file system path representation of a `PathLike` object,
diff --git a/tensorflow/stream_executor/dnn.h b/tensorflow/stream_executor/dnn.h
index f4162b0962..aa88fe770f 100644
--- a/tensorflow/stream_executor/dnn.h
+++ b/tensorflow/stream_executor/dnn.h
@@ -896,7 +896,7 @@ class DnnSupport {
// offset: offset parameters.
// estimated_mean: population mean estimated during training.
// Used for inference only; empty for training.
- // estimated_variance: population variance estimated during traning,
+ // estimated_variance: population variance estimated during training,
// used for inference only; empty for training.
// x_desc: dimensions of the input data, which is the same as the dimensions
// of the output.
diff --git a/tensorflow/tools/ci_build/ci_sanity.sh b/tensorflow/tools/ci_build/ci_sanity.sh
index f980ced2e4..aeac085d30 100755
--- a/tensorflow/tools/ci_build/ci_sanity.sh
+++ b/tensorflow/tools/ci_build/ci_sanity.sh
@@ -353,7 +353,7 @@ do_external_licenses_check(){
# Whitelist
echo ${EXTRA_LICENSE_FILE}
- grep -e "@bazel_tools//src/" -e "@bazel_tools//tools/" -e "@com_google_absl//" -e "//external" -e "@local" -v ${EXTRA_LICENSES_FILE} > temp.txt
+ grep -e "@bazel_tools//src" -e "@bazel_tools//tools/" -e "@com_google_absl//" -e "//external" -e "@local" -v ${EXTRA_LICENSES_FILE} > temp.txt
mv temp.txt ${EXTRA_LICENSES_FILE}
diff --git a/tensorflow/tools/docker/Dockerfile.devel b/tensorflow/tools/docker/Dockerfile.devel
index 5dc4a053fd..d16761c367 100644
--- a/tensorflow/tools/docker/Dockerfile.devel
+++ b/tensorflow/tools/docker/Dockerfile.devel
@@ -70,7 +70,7 @@ RUN mkdir /bazel && \
# Download and build TensorFlow.
WORKDIR /tensorflow
-RUN git clone --branch=r1.5 --depth=1 https://github.com/tensorflow/tensorflow.git .
+RUN git clone --branch=r1.6 --depth=1 https://github.com/tensorflow/tensorflow.git .
# TODO(craigcitro): Don't install the pip package, since it makes it
# more difficult to experiment with local changes. Instead, just add
diff --git a/tensorflow/tools/docker/Dockerfile.devel-cpu-mkl b/tensorflow/tools/docker/Dockerfile.devel-cpu-mkl
index 96b260ad3a..3690e7dfe5 100644
--- a/tensorflow/tools/docker/Dockerfile.devel-cpu-mkl
+++ b/tensorflow/tools/docker/Dockerfile.devel-cpu-mkl
@@ -3,7 +3,7 @@ FROM tensorflow/tensorflow:latest-devel
LABEL maintainer="Clayne Robison<clayne.b.robison@intel.com>"
# These arguments are parameterized. Use --build-args to override.
-ARG TF_BRANCH=r1.5
+ARG TF_BRANCH=r1.6
ARG WHL_DIR=/whl
RUN apt-get update && apt-get install -y --no-install-recommends \
diff --git a/tensorflow/tools/docker/Dockerfile.devel-gpu b/tensorflow/tools/docker/Dockerfile.devel-gpu
index 07ffd3839a..4ef37881bc 100644
--- a/tensorflow/tools/docker/Dockerfile.devel-gpu
+++ b/tensorflow/tools/docker/Dockerfile.devel-gpu
@@ -79,7 +79,7 @@ RUN mkdir /bazel && \
# Download and build TensorFlow.
WORKDIR /tensorflow
-RUN git clone --branch=r1.5 --depth=1 https://github.com/tensorflow/tensorflow.git .
+RUN git clone --branch=r1.6 --depth=1 https://github.com/tensorflow/tensorflow.git .
# Configure the build for our CUDA configuration.
ENV CI_BUILD_PYTHON python
diff --git a/tensorflow/tools/graph_transforms/BUILD b/tensorflow/tools/graph_transforms/BUILD
index b5465b7fb3..8601b3d0f1 100644
--- a/tensorflow/tools/graph_transforms/BUILD
+++ b/tensorflow/tools/graph_transforms/BUILD
@@ -99,22 +99,21 @@ cc_library(
"freeze_requantization_ranges.cc",
"fuse_convolutions.cc",
"insert_logging.cc",
- "remove_ema.cc",
"obfuscate_names.cc",
+ "quantize_nodes.cc",
+ "quantize_weights.cc",
"remove_attribute.cc",
"remove_device.cc",
+ "remove_ema.cc",
"remove_nodes.cc",
"rename_attribute.cc",
"rename_op.cc",
+ "round_weights.cc",
"set_device.cc",
"sort_by_execution_order.cc",
"sparsify_gather.cc",
"strip_unused_nodes.cc",
- ] + if_not_windows([
- "quantize_nodes.cc",
- "quantize_weights.cc",
- "round_weights.cc",
- ]),
+ ],
hdrs = [
"fold_constants_lib.h",
],
diff --git a/tensorflow/tools/pip_package/setup.py b/tensorflow/tools/pip_package/setup.py
index bc4315c600..0e6b32bb49 100644
--- a/tensorflow/tools/pip_package/setup.py
+++ b/tensorflow/tools/pip_package/setup.py
@@ -29,7 +29,7 @@ from setuptools.dist import Distribution
# This version string is semver compatible, but incompatible with pip.
# For pip, we will remove all '-' characters from this string, and use the
# result for pip.
-_VERSION = '1.5.0'
+_VERSION = '1.6.0-rc0'
REQUIRED_PACKAGES = [
'absl-py >= 0.1.6',
diff --git a/third_party/repo.bzl b/third_party/repo.bzl
index 11e9c842d2..aa178fa8ca 100644
--- a/third_party/repo.bzl
+++ b/third_party/repo.bzl
@@ -27,7 +27,7 @@ def _wrap_bash_cmd(ctx, cmd):
bazel_sh = _get_env_var(ctx, "BAZEL_SH")
if not bazel_sh:
fail("BAZEL_SH environment variable is not set")
- cmd = [bazel_sh, "-c", " ".join(cmd)]
+ cmd = [bazel_sh, "-l", "-c", " ".join(cmd)]
return cmd
def _get_env_var(ctx, name):