aboutsummaryrefslogtreecommitdiffhomepage
diff options
context:
space:
mode:
authorGravatar Jonathan Hseu <jhseu@google.com>2016-09-19 11:14:58 -0800
committerGravatar TensorFlower Gardener <gardener@tensorflow.org>2016-09-19 12:17:00 -0700
commit9f5b098f77e5e51a2fc681369dba8ea5dbba5356 (patch)
tree24c75c9ba4ca43829b86902a7754374104a56acf
parent7efb6906a1d6d146ff84ea4391891d243ee2c5d2 (diff)
HDFS support
Notes: - The test is tagged as manual, and you must download the Hadoop distribution to run it. - We ask during ./configure whether to include HDFS support. - Copied hdfs.h from Hadoop here in third_party. It's licensed Apache 2.0. Change: 133615494
-rwxr-xr-xconfigure24
-rw-r--r--tensorflow/BUILD2
-rw-r--r--tensorflow/contrib/makefile/proto_text_cc_files.txt1
-rw-r--r--tensorflow/core/BUILD1
-rw-r--r--tensorflow/core/platform/default/build_config.bzl11
-rw-r--r--tensorflow/core/platform/hadoop/BUILD69
-rw-r--r--tensorflow/core/platform/hadoop/hadoop_file_system.cc431
-rw-r--r--tensorflow/core/platform/hadoop/hadoop_file_system.h73
-rw-r--r--tensorflow/core/platform/hadoop/hadoop_file_system_test.cc186
-rw-r--r--tensorflow/core/platform/posix/error.cc163
-rw-r--r--tensorflow/core/platform/posix/error.h27
-rw-r--r--tensorflow/core/platform/posix/posix_file_system.cc143
-rw-r--r--tensorflow/third_party/hadoop/BUILD20
-rw-r--r--tensorflow/third_party/hadoop/hdfs.h911
-rwxr-xr-xtensorflow/tools/ci_build/builds/configured2
15 files changed, 1918 insertions, 146 deletions
diff --git a/configure b/configure
index ce75bb490a..feac140664 100755
--- a/configure
+++ b/configure
@@ -1,7 +1,5 @@
#!/usr/bin/env bash
-DO_NOT_SUBMIT_WARNING="Unofficial setting. DO NOT SUBMIT!!!"
-
# Find out the absolute path to where ./configure resides
pushd `dirname $0` #> /dev/null
SOURCE_BASE_DIR=`pwd -P`
@@ -60,6 +58,28 @@ else
perl -pi -e "s,WITH_GCP_SUPPORT = (False|True),WITH_GCP_SUPPORT = False,s" tensorflow/core/platform/default/build_config.bzl
fi
+while [ "$TF_NEED_HDFS" == "" ]; do
+ read -p "Do you wish to build TensorFlow with "\
+"Hadoop File System support? [y/N] " INPUT
+ case $INPUT in
+ [Yy]* ) echo "Hadoop File System support will be enabled for "\
+"TensorFlow"; TF_NEED_HDFS=1;;
+ [Nn]* ) echo "No Hadoop File System support will be enabled for "\
+"TensorFlow"; TF_NEED_HDFS=0;;
+ "" ) echo "No Hadoop File System support will be enabled for "\
+"TensorFlow"; TF_NEED_HDFS=0;;
+ * ) echo "Invalid selection: " $INPUT;;
+ esac
+done
+
+if [ "$TF_NEED_HDFS" == "1" ]; then
+ # Update Bazel build configuration.
+ perl -pi -e "s,WITH_HDFS_SUPPORT = (False|True),WITH_HDFS_SUPPORT = True,s" tensorflow/core/platform/default/build_config.bzl
+else
+ # Update Bazel build configuration.
+ perl -pi -e "s,WITH_HDFS_SUPPORT = (False|True),WITH_HDFS_SUPPORT = False,s" tensorflow/core/platform/default/build_config.bzl
+fi
+
## Find swig path
if [ -z "$SWIG_PATH" ]; then
SWIG_PATH=`type -p swig 2> /dev/null`
diff --git a/tensorflow/BUILD b/tensorflow/BUILD
index 5b88f1eaa5..39b8868d52 100644
--- a/tensorflow/BUILD
+++ b/tensorflow/BUILD
@@ -131,6 +131,7 @@ filegroup(
"//tensorflow/core/ops/compat:all_files",
"//tensorflow/core/platform/cloud:all_files",
"//tensorflow/core/platform/default/build_config:all_files",
+ "//tensorflow/core/platform/hadoop:all_files",
"//tensorflow/core/util/ctc:all_files",
"//tensorflow/examples/android:all_files",
"//tensorflow/examples/how_tos/reading_data:all_files",
@@ -166,6 +167,7 @@ filegroup(
"//tensorflow/tensorboard/lib:all_files",
"//tensorflow/tensorboard/lib/python:all_files",
"//tensorflow/tensorboard/scripts:all_files",
+ "//tensorflow/third_party/hadoop:all_files",
"//tensorflow/tools/dist_test/server:all_files",
"//tensorflow/tools/docker:all_files",
"//tensorflow/tools/docker/notebooks:all_files",
diff --git a/tensorflow/contrib/makefile/proto_text_cc_files.txt b/tensorflow/contrib/makefile/proto_text_cc_files.txt
index 648be42db5..9b0a0942cd 100644
--- a/tensorflow/contrib/makefile/proto_text_cc_files.txt
+++ b/tensorflow/contrib/makefile/proto_text_cc_files.txt
@@ -5,6 +5,7 @@ tensorflow/core/platform/tensor_coding.cc
tensorflow/core/platform/protobuf_util.cc
tensorflow/core/platform/posix/posix_file_system.cc
tensorflow/core/platform/posix/port.cc
+tensorflow/core/platform/posix/error.cc
tensorflow/core/platform/posix/env.cc
tensorflow/core/platform/load_library.cc
tensorflow/core/platform/file_system.cc
diff --git a/tensorflow/core/BUILD b/tensorflow/core/BUILD
index 2ad40d0cf3..fa5441976d 100644
--- a/tensorflow/core/BUILD
+++ b/tensorflow/core/BUILD
@@ -630,6 +630,7 @@ filegroup(
"platform/default/test_benchmark.*",
"platform/cuda.h",
"platform/google/**/*",
+ "platform/hadoop/**/*",
"platform/jpeg.*",
"platform/png.*",
"platform/gif.*",
diff --git a/tensorflow/core/platform/default/build_config.bzl b/tensorflow/core/platform/default/build_config.bzl
index c357e3d23e..8a8dfaef60 100644
--- a/tensorflow/core/platform/default/build_config.bzl
+++ b/tensorflow/core/platform/default/build_config.bzl
@@ -3,8 +3,9 @@
load("@protobuf//:protobuf.bzl", "cc_proto_library")
load("@protobuf//:protobuf.bzl", "py_proto_library")
-# configure may change the following line to True
+# configure may change the following lines to True
WITH_GCP_SUPPORT = False
+WITH_HDFS_SUPPORT = False
# Appends a suffix to a list of deps.
def tf_deps(deps, suffix):
@@ -123,5 +124,9 @@ def tf_kernel_tests_linkstatic():
return 0
def tf_additional_lib_deps():
- return (["//tensorflow/core/platform/cloud:gcs_file_system"]
- if WITH_GCP_SUPPORT else [])
+ deps = []
+ if WITH_GCP_SUPPORT:
+ deps.append("//tensorflow/core/platform/cloud:gcs_file_system")
+ if WITH_HDFS_SUPPORT:
+ deps.append("//tensorflow/core/platform/hadoop:hadoop_file_system")
+ return deps
diff --git a/tensorflow/core/platform/hadoop/BUILD b/tensorflow/core/platform/hadoop/BUILD
new file mode 100644
index 0000000000..457b4fe14c
--- /dev/null
+++ b/tensorflow/core/platform/hadoop/BUILD
@@ -0,0 +1,69 @@
+# Description:
+# Hadoop file system implementation.
+
+package(
+ default_visibility = ["//visibility:public"],
+)
+
+licenses(["notice"]) # Apache 2.0
+
+load(
+ "//tensorflow:tensorflow.bzl",
+ "tf_cc_test",
+)
+
+filegroup(
+ name = "all_files",
+ srcs = glob(
+ ["**/*"],
+ exclude = [
+ "**/METADATA",
+ "**/OWNERS",
+ ],
+ ),
+ visibility = ["//tensorflow:__subpackages__"],
+)
+
+cc_library(
+ name = "hadoop_file_system",
+ srcs = ["hadoop_file_system.cc"],
+ hdrs = ["hadoop_file_system.h"],
+ deps = [
+ "//tensorflow/core:lib",
+ "//tensorflow/core:lib_internal",
+ "//tensorflow/third_party/hadoop:hdfs",
+ ],
+ alwayslink = 1,
+)
+
+# This test is set to manual because it requires downloading the Hadoop
+# distribution to run. To run this test:
+# 1. Ensure $JAVA_HOME is set.
+# 2. Download the binary Hadoop distribution from:
+# http://hadoop.apache.org/releases.html
+# 3. Extract the Hadoop distribution and run:
+# source libexec/hadoop-config.sh
+# 4. bazel test \
+# --test_env=LD_LIBRARY_PATH=$JAVA_HOME/jre/lib/amd64/server \
+# --test_env=HADOOP_HDFS_HOME=$HADOOP_HDFS_HOME \
+# --test_env=CLASSPATH=$($HADOOP_HDFS_HOME/bin/hadoop classpath --glob) \
+# --test_strategy=local \
+# :hadoop_file_system_test
+tf_cc_test(
+ name = "hadoop_file_system_test",
+ size = "small",
+ srcs = [
+ "hadoop_file_system_test.cc",
+ ],
+ tags = [
+ "manual",
+ "notap",
+ ],
+ deps = [
+ ":hadoop_file_system",
+ "//tensorflow/core:lib",
+ "//tensorflow/core:lib_internal",
+ "//tensorflow/core:test",
+ "//tensorflow/core:test_main",
+ ],
+)
diff --git a/tensorflow/core/platform/hadoop/hadoop_file_system.cc b/tensorflow/core/platform/hadoop/hadoop_file_system.cc
new file mode 100644
index 0000000000..662818cc9a
--- /dev/null
+++ b/tensorflow/core/platform/hadoop/hadoop_file_system.cc
@@ -0,0 +1,431 @@
+/* Copyright 2016 The TensorFlow Authors. All Rights Reserved.
+
+Licensed under the Apache License, Version 2.0 (the "License");
+you may not use this file except in compliance with the License.
+You may obtain a copy of the License at
+
+ http://www.apache.org/licenses/LICENSE-2.0
+
+Unless required by applicable law or agreed to in writing, software
+distributed under the License is distributed on an "AS IS" BASIS,
+WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+See the License for the specific language governing permissions and
+limitations under the License.
+==============================================================================*/
+
+#include "tensorflow/core/platform/hadoop/hadoop_file_system.h"
+
+#include <errno.h>
+
+#include "tensorflow/core/lib/core/status.h"
+#include "tensorflow/core/lib/io/path.h"
+#include "tensorflow/core/lib/strings/strcat.h"
+#include "tensorflow/core/platform/env.h"
+#include "tensorflow/core/platform/logging.h"
+#include "tensorflow/core/platform/posix/error.h"
+#include "tensorflow/third_party/hadoop/hdfs.h"
+
+namespace tensorflow {
+
+template <typename R, typename... Args>
+Status BindFunc(void* handle, const char* name,
+ std::function<R(Args...)>* func) {
+ void* symbol_ptr = nullptr;
+ TF_RETURN_IF_ERROR(
+ Env::Default()->GetSymbolFromLibrary(handle, name, &symbol_ptr));
+ *func = reinterpret_cast<R (*)(Args...)>(symbol_ptr);
+ return Status::OK();
+}
+
+class LibHDFS {
+ public:
+ static LibHDFS* Load() {
+ static LibHDFS* lib = []() -> LibHDFS* {
+ LibHDFS* lib = new LibHDFS;
+ lib->LoadAndBind();
+ return lib;
+ }();
+
+ return lib;
+ }
+
+ // The status, if any, from failure to load.
+ Status status() { return status_; }
+
+ std::function<hdfsFS(hdfsBuilder*)> hdfsBuilderConnect;
+ std::function<hdfsBuilder*()> hdfsNewBuilder;
+ std::function<void(hdfsBuilder*, const char*)> hdfsBuilderSetNameNode;
+ std::function<int(hdfsFS, hdfsFile)> hdfsCloseFile;
+ std::function<tSize(hdfsFS, hdfsFile, tOffset, void*, tSize)> hdfsPread;
+ std::function<tSize(hdfsFS, hdfsFile, const void*, tSize)> hdfsWrite;
+ std::function<int(hdfsFS, hdfsFile)> hdfsFlush;
+ std::function<int(hdfsFS, hdfsFile)> hdfsHSync;
+ std::function<hdfsFile(hdfsFS, const char*, int, int, short, tSize)>
+ hdfsOpenFile;
+ std::function<int(hdfsFS, const char*)> hdfsExists;
+ std::function<hdfsFileInfo*(hdfsFS, const char*, int*)> hdfsListDirectory;
+ std::function<void(hdfsFileInfo*, int)> hdfsFreeFileInfo;
+ std::function<int(hdfsFS, const char*, int recursive)> hdfsDelete;
+ std::function<int(hdfsFS, const char*)> hdfsCreateDirectory;
+ std::function<hdfsFileInfo*(hdfsFS, const char*)> hdfsGetPathInfo;
+ std::function<int(hdfsFS, const char*, const char*)> hdfsRename;
+
+ private:
+ void LoadAndBind() {
+ auto TryLoadAndBind = [this](const char* name, void** handle) -> Status {
+ TF_RETURN_IF_ERROR(Env::Default()->LoadLibrary(name, handle));
+#define BIND_HDFS_FUNC(function) \
+ TF_RETURN_IF_ERROR(BindFunc(*handle, #function, &function));
+
+ BIND_HDFS_FUNC(hdfsBuilderConnect);
+ BIND_HDFS_FUNC(hdfsNewBuilder);
+ BIND_HDFS_FUNC(hdfsBuilderSetNameNode);
+ BIND_HDFS_FUNC(hdfsCloseFile);
+ BIND_HDFS_FUNC(hdfsPread);
+ BIND_HDFS_FUNC(hdfsWrite);
+ BIND_HDFS_FUNC(hdfsFlush);
+ BIND_HDFS_FUNC(hdfsHSync);
+ BIND_HDFS_FUNC(hdfsOpenFile);
+ BIND_HDFS_FUNC(hdfsExists);
+ BIND_HDFS_FUNC(hdfsListDirectory);
+ BIND_HDFS_FUNC(hdfsFreeFileInfo);
+ BIND_HDFS_FUNC(hdfsDelete);
+ BIND_HDFS_FUNC(hdfsCreateDirectory);
+ BIND_HDFS_FUNC(hdfsGetPathInfo);
+ BIND_HDFS_FUNC(hdfsRename);
+#undef BIND_HDFS_FUNC
+ return Status::OK();
+ };
+
+ // libhdfs.so won't be in the standard locations. Use the path as specified
+ // in the libhdfs documentation.
+ char* hdfs_home = getenv("HADOOP_HDFS_HOME");
+ if (hdfs_home == nullptr) {
+ status_ = errors::FailedPrecondition(
+ "Environment variable HADOOP_HDFS_HOME not set");
+ return;
+ }
+ string path = io::JoinPath(hdfs_home, "lib", "native", "libhdfs.so");
+ status_ = TryLoadAndBind(path.c_str(), &handle_);
+ return;
+ }
+
+ Status status_;
+ void* handle_ = nullptr;
+};
+
+HadoopFileSystem::HadoopFileSystem() : hdfs_(LibHDFS::Load()) {}
+
+HadoopFileSystem::~HadoopFileSystem() {}
+
+// We rely on HDFS connection caching here. The HDFS client calls
+// org.apache.hadoop.fs.FileSystem.get(), which caches the connection
+// internally.
+Status HadoopFileSystem::Connect(StringPiece fname, hdfsFS* fs) {
+ TF_RETURN_IF_ERROR(hdfs_->status());
+
+ if (!fname.Consume("hdfs://")) {
+ return errors::InvalidArgument("HDFS path must start with hdfs://");
+ }
+ auto first_slash = fname.find('/');
+ string namenode;
+ if (first_slash == string::npos) {
+ namenode = fname.ToString();
+ } else {
+ namenode = fname.substr(0, first_slash).ToString();
+ }
+
+ hdfsBuilder* builder = hdfs_->hdfsNewBuilder();
+ if (namenode == "localfilesystem") {
+ hdfs_->hdfsBuilderSetNameNode(builder, nullptr);
+ } else {
+ hdfs_->hdfsBuilderSetNameNode(builder, namenode.c_str());
+ }
+ *fs = hdfs_->hdfsBuilderConnect(builder);
+ if (*fs == nullptr) {
+ return errors::NotFound(strerror(errno));
+ }
+ return Status::OK();
+}
+
+string HadoopFileSystem::TranslateName(const string& name) const {
+ StringPiece sp = name;
+ sp.Consume("hdfs://");
+ auto first_slash = sp.find('/');
+ if (first_slash == string::npos) {
+ return string();
+ }
+ sp.remove_prefix(first_slash);
+ return sp.ToString();
+}
+
+class HDFSRandomAccessFile : public RandomAccessFile {
+ public:
+ HDFSRandomAccessFile(const string& fname, LibHDFS* hdfs, hdfsFS fs,
+ hdfsFile file)
+ : filename_(fname), hdfs_(hdfs), fs_(fs), file_(file) {}
+
+ ~HDFSRandomAccessFile() override { hdfs_->hdfsCloseFile(fs_, file_); }
+
+ Status Read(uint64 offset, size_t n, StringPiece* result,
+ char* scratch) const override {
+ Status s;
+ char* dst = scratch;
+ while (n > 0 && s.ok()) {
+ tSize r = hdfs_->hdfsPread(fs_, file_, static_cast<tOffset>(offset), dst,
+ static_cast<tSize>(n));
+ if (r > 0) {
+ dst += r;
+ n -= r;
+ offset += r;
+ } else if (r == 0) {
+ s = Status(error::OUT_OF_RANGE, "Read less bytes than requested");
+ } else if (errno == EINTR || errno == EAGAIN) {
+ // hdfsPread may return EINTR too. Just retry.
+ } else {
+ s = IOError(filename_, errno);
+ }
+ }
+ *result = StringPiece(scratch, dst - scratch);
+ return s;
+ }
+
+ private:
+ string filename_;
+ LibHDFS* hdfs_;
+ hdfsFS fs_;
+ hdfsFile file_;
+};
+
+Status HadoopFileSystem::NewRandomAccessFile(
+ const string& fname, std::unique_ptr<RandomAccessFile>* result) {
+ hdfsFS fs = nullptr;
+ TF_RETURN_IF_ERROR(Connect(fname, &fs));
+
+ hdfsFile file =
+ hdfs_->hdfsOpenFile(fs, TranslateName(fname).c_str(), O_RDONLY, 0, 0, 0);
+ if (file == nullptr) {
+ return IOError(fname, errno);
+ }
+ result->reset(new HDFSRandomAccessFile(fname, hdfs_, fs, file));
+ return Status::OK();
+}
+
+class HDFSWritableFile : public WritableFile {
+ public:
+ HDFSWritableFile(const string& fname, LibHDFS* hdfs, hdfsFS fs, hdfsFile file)
+ : filename_(fname), hdfs_(hdfs), fs_(fs), file_(file) {}
+
+ ~HDFSWritableFile() override {
+ if (file_ != nullptr) {
+ Close();
+ }
+ }
+
+ Status Append(const StringPiece& data) override {
+ if (hdfs_->hdfsWrite(fs_, file_, data.data(),
+ static_cast<tSize>(data.size())) == -1) {
+ return IOError(filename_, errno);
+ }
+ return Status::OK();
+ }
+
+ Status Close() override {
+ Status result;
+ if (hdfs_->hdfsCloseFile(fs_, file_) != 0) {
+ result = IOError(filename_, errno);
+ }
+ hdfs_ = nullptr;
+ fs_ = nullptr;
+ file_ = nullptr;
+ return result;
+ }
+
+ Status Flush() override {
+ if (hdfs_->hdfsFlush(fs_, file_) != 0) {
+ return IOError(filename_, errno);
+ }
+ return Status::OK();
+ }
+
+ Status Sync() override {
+ if (hdfs_->hdfsHSync(fs_, file_) != 0) {
+ return IOError(filename_, errno);
+ }
+ return Status::OK();
+ }
+
+ private:
+ string filename_;
+ LibHDFS* hdfs_;
+ hdfsFS fs_;
+ hdfsFile file_;
+};
+
+Status HadoopFileSystem::NewWritableFile(
+ const string& fname, std::unique_ptr<WritableFile>* result) {
+ hdfsFS fs = nullptr;
+ TF_RETURN_IF_ERROR(Connect(fname, &fs));
+
+ hdfsFile file =
+ hdfs_->hdfsOpenFile(fs, TranslateName(fname).c_str(), O_WRONLY, 0, 0, 0);
+ if (file == nullptr) {
+ return IOError(fname, errno);
+ }
+ result->reset(new HDFSWritableFile(fname, hdfs_, fs, file));
+ return Status::OK();
+}
+
+Status HadoopFileSystem::NewAppendableFile(
+ const string& fname, std::unique_ptr<WritableFile>* result) {
+ hdfsFS fs = nullptr;
+ TF_RETURN_IF_ERROR(Connect(fname, &fs));
+
+ hdfsFile file = hdfs_->hdfsOpenFile(fs, TranslateName(fname).c_str(),
+ O_WRONLY | O_APPEND, 0, 0, 0);
+ if (file == nullptr) {
+ return IOError(fname, errno);
+ }
+ result->reset(new HDFSWritableFile(fname, hdfs_, fs, file));
+ return Status::OK();
+}
+
+Status HadoopFileSystem::NewReadOnlyMemoryRegionFromFile(
+ const string& fname, std::unique_ptr<ReadOnlyMemoryRegion>* result) {
+ // hadoopReadZero() technically supports this call with the following
+ // caveats:
+ // - It only works up to 2 GB. We'd have to Stat() the file to ensure that
+ // it fits.
+ // - If not on the local filesystem, the entire file will be read, making
+ // it inefficient for callers that assume typical mmap() behavior.
+ return errors::Unimplemented("HDFS does not support ReadOnlyMemoryRegion");
+}
+
+bool HadoopFileSystem::FileExists(const string& fname) {
+ hdfsFS fs = nullptr;
+ Status status = Connect(fname, &fs);
+ if (!status.ok()) {
+ LOG(ERROR) << "Connect failed: " << status.error_message();
+ return false;
+ }
+
+ return hdfs_->hdfsExists(fs, TranslateName(fname).c_str()) == 0;
+}
+
+Status HadoopFileSystem::GetChildren(const string& dir,
+ std::vector<string>* result) {
+ result->clear();
+ hdfsFS fs = nullptr;
+ TF_RETURN_IF_ERROR(Connect(dir, &fs));
+
+ // hdfsListDirectory returns nullptr if the directory is empty. Do a separate
+ // check to verify the directory exists first.
+ FileStatistics stat;
+ TF_RETURN_IF_ERROR(Stat(dir, &stat));
+
+ int entries = 0;
+ hdfsFileInfo* info =
+ hdfs_->hdfsListDirectory(fs, TranslateName(dir).c_str(), &entries);
+ if (info == nullptr) {
+ if (stat.is_directory) {
+ // Assume it's an empty directory.
+ return Status::OK();
+ }
+ return IOError(dir, errno);
+ }
+ for (int i = 0; i < entries; i++) {
+ result->push_back(io::Basename(info[i].mName).ToString());
+ }
+ hdfs_->hdfsFreeFileInfo(info, entries);
+ return Status::OK();
+}
+
+Status HadoopFileSystem::DeleteFile(const string& fname) {
+ hdfsFS fs = nullptr;
+ TF_RETURN_IF_ERROR(Connect(fname, &fs));
+
+ if (hdfs_->hdfsDelete(fs, TranslateName(fname).c_str(),
+ /*recursive=*/0) != 0) {
+ return IOError(fname, errno);
+ }
+ return Status::OK();
+}
+
+Status HadoopFileSystem::CreateDir(const string& dir) {
+ hdfsFS fs = nullptr;
+ TF_RETURN_IF_ERROR(Connect(dir, &fs));
+
+ if (hdfs_->hdfsCreateDirectory(fs, TranslateName(dir).c_str()) != 0) {
+ return IOError(dir, errno);
+ }
+ return Status::OK();
+}
+
+Status HadoopFileSystem::DeleteDir(const string& dir) {
+ hdfsFS fs = nullptr;
+ TF_RETURN_IF_ERROR(Connect(dir, &fs));
+
+ // Count the number of entries in the directory, and only delete if it's
+ // non-empty. This is consistent with the interface, but note that there's
+ // a race condition where a file may be added after this check, in which
+ // case the directory will still be deleted.
+ int entries = 0;
+ hdfsFileInfo* info =
+ hdfs_->hdfsListDirectory(fs, TranslateName(dir).c_str(), &entries);
+ if (info != nullptr) {
+ return IOError(dir, errno);
+ }
+ hdfs_->hdfsFreeFileInfo(info, entries);
+
+ if (entries > 0) {
+ return errors::FailedPrecondition("Cannot delete a non-empty directory.");
+ }
+ if (hdfs_->hdfsDelete(fs, TranslateName(dir).c_str(),
+ /*recursive=*/1) != 0) {
+ return IOError(dir, errno);
+ }
+ return Status::OK();
+}
+
+Status HadoopFileSystem::GetFileSize(const string& fname, uint64* size) {
+ hdfsFS fs = nullptr;
+ TF_RETURN_IF_ERROR(Connect(fname, &fs));
+
+ hdfsFileInfo* info = hdfs_->hdfsGetPathInfo(fs, TranslateName(fname).c_str());
+ if (info == nullptr) {
+ return IOError(fname, errno);
+ }
+ *size = static_cast<uint64>(info->mSize);
+ hdfs_->hdfsFreeFileInfo(info, 1);
+ return Status::OK();
+}
+
+Status HadoopFileSystem::RenameFile(const string& src, const string& target) {
+ hdfsFS fs = nullptr;
+ TF_RETURN_IF_ERROR(Connect(src, &fs));
+
+ if (hdfs_->hdfsRename(fs, TranslateName(src).c_str(),
+ TranslateName(target).c_str()) != 0) {
+ return IOError(src, errno);
+ }
+ return Status::OK();
+}
+
+Status HadoopFileSystem::Stat(const string& fname, FileStatistics* stats) {
+ hdfsFS fs = nullptr;
+ TF_RETURN_IF_ERROR(Connect(fname, &fs));
+
+ hdfsFileInfo* info = hdfs_->hdfsGetPathInfo(fs, TranslateName(fname).c_str());
+ if (info == nullptr) {
+ return IOError(fname, errno);
+ }
+ stats->length = static_cast<int64>(info->mSize);
+ stats->mtime_nsec = static_cast<int64>(info->mLastMod) * 1e9;
+ stats->is_directory = info->mKind == kObjectKindDirectory;
+ hdfs_->hdfsFreeFileInfo(info, 1);
+ return Status::OK();
+}
+
+REGISTER_FILE_SYSTEM("hdfs", HadoopFileSystem);
+
+} // namespace tensorflow
diff --git a/tensorflow/core/platform/hadoop/hadoop_file_system.h b/tensorflow/core/platform/hadoop/hadoop_file_system.h
new file mode 100644
index 0000000000..182d6995fc
--- /dev/null
+++ b/tensorflow/core/platform/hadoop/hadoop_file_system.h
@@ -0,0 +1,73 @@
+/* Copyright 2016 The TensorFlow Authors. All Rights Reserved.
+
+Licensed under the Apache License, Version 2.0 (the "License");
+you may not use this file except in compliance with the License.
+You may obtain a copy of the License at
+
+ http://www.apache.org/licenses/LICENSE-2.0
+
+Unless required by applicable law or agreed to in writing, software
+distributed under the License is distributed on an "AS IS" BASIS,
+WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+See the License for the specific language governing permissions and
+limitations under the License.
+==============================================================================*/
+
+#ifndef THIRD_PARTY_TENSORFLOW_CORE_PLATFORM_HADOOP_HADOOP_FILE_SYSTEM_H_
+#define THIRD_PARTY_TENSORFLOW_CORE_PLATFORM_HADOOP_HADOOP_FILE_SYSTEM_H_
+
+#include "tensorflow/core/platform/env.h"
+
+extern "C" {
+struct hdfs_internal;
+typedef hdfs_internal* hdfsFS;
+}
+
+namespace tensorflow {
+
+class LibHDFS;
+
+class HadoopFileSystem : public FileSystem {
+ public:
+ HadoopFileSystem();
+ ~HadoopFileSystem();
+
+ Status NewRandomAccessFile(
+ const string& fname, std::unique_ptr<RandomAccessFile>* result) override;
+
+ Status NewWritableFile(const string& fname,
+ std::unique_ptr<WritableFile>* result) override;
+
+ Status NewAppendableFile(const string& fname,
+ std::unique_ptr<WritableFile>* result) override;
+
+ Status NewReadOnlyMemoryRegionFromFile(
+ const string& fname,
+ std::unique_ptr<ReadOnlyMemoryRegion>* result) override;
+
+ bool FileExists(const string& fname) override;
+
+ Status GetChildren(const string& dir, std::vector<string>* result) override;
+
+ Status DeleteFile(const string& fname) override;
+
+ Status CreateDir(const string& name) override;
+
+ Status DeleteDir(const string& name) override;
+
+ Status GetFileSize(const string& fname, uint64* size) override;
+
+ Status RenameFile(const string& src, const string& target) override;
+
+ Status Stat(const string& fname, FileStatistics* stat) override;
+
+ string TranslateName(const string& name) const override;
+
+ private:
+ Status Connect(StringPiece fname, hdfsFS* fs);
+ LibHDFS* hdfs_;
+};
+
+} // namespace tensorflow
+
+#endif // THIRD_PARTY_TENSORFLOW_CORE_PLATFORM_HADOOP_HADOOP_FILE_SYSTEM_H_
diff --git a/tensorflow/core/platform/hadoop/hadoop_file_system_test.cc b/tensorflow/core/platform/hadoop/hadoop_file_system_test.cc
new file mode 100644
index 0000000000..f927e15752
--- /dev/null
+++ b/tensorflow/core/platform/hadoop/hadoop_file_system_test.cc
@@ -0,0 +1,186 @@
+/* Copyright 2016 The TensorFlow Authors. All Rights Reserved.
+
+Licensed under the Apache License, Version 2.0 (the "License");
+you may not use this file except in compliance with the License.
+You may obtain a copy of the License at
+
+ http://www.apache.org/licenses/LICENSE-2.0
+
+Unless required by applicable law or agreed to in writing, software
+distributed under the License is distributed on an "AS IS" BASIS,
+WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+See the License for the specific language governing permissions and
+limitations under the License.
+==============================================================================*/
+
+#include "tensorflow/core/platform/hadoop/hadoop_file_system.h"
+
+#include "tensorflow/core/lib/core/status_test_util.h"
+#include "tensorflow/core/lib/gtl/stl_util.h"
+#include "tensorflow/core/lib/io/path.h"
+#include "tensorflow/core/platform/file_system.h"
+#include "tensorflow/core/platform/test.h"
+
+namespace tensorflow {
+namespace {
+
+class HadoopFileSystemTest : public ::testing::Test {
+ protected:
+ HadoopFileSystemTest() {}
+
+ Status WriteString(const string& fname, const string& content) {
+ std::unique_ptr<WritableFile> writer;
+ TF_RETURN_IF_ERROR(hdfs.NewWritableFile(fname, &writer));
+ TF_RETURN_IF_ERROR(writer->Append(content));
+ TF_RETURN_IF_ERROR(writer->Close());
+ return Status::OK();
+ }
+
+ Status ReadAll(const string& fname, string* content) {
+ std::unique_ptr<RandomAccessFile> reader;
+ TF_RETURN_IF_ERROR(hdfs.NewRandomAccessFile(fname, &reader));
+
+ uint64 file_size = 0;
+ TF_RETURN_IF_ERROR(hdfs.GetFileSize(fname, &file_size));
+
+ content->resize(file_size);
+ StringPiece result;
+ TF_RETURN_IF_ERROR(
+ reader->Read(0, file_size, &result, gtl::string_as_array(content)));
+ if (file_size != result.size()) {
+ return errors::DataLoss("expected ", file_size, " got ", result.size(),
+ " bytes");
+ }
+ return Status::OK();
+ }
+
+ HadoopFileSystem hdfs;
+};
+
+TEST_F(HadoopFileSystemTest, RandomAccessFile) {
+ const string fname = io::JoinPath("hdfs://localfilesystem", testing::TmpDir(),
+ "RandomAccessFile");
+ const string content = "abcdefghijklmn";
+ TF_ASSERT_OK(WriteString(fname, content));
+
+ std::unique_ptr<RandomAccessFile> reader;
+ TF_EXPECT_OK(hdfs.NewRandomAccessFile(fname, &reader));
+
+ string got;
+ got.resize(content.size());
+ StringPiece result;
+ TF_EXPECT_OK(
+ reader->Read(0, content.size(), &result, gtl::string_as_array(&got)));
+ EXPECT_EQ(content.size(), result.size());
+ EXPECT_EQ(content, result);
+
+ got.clear();
+ got.resize(4);
+ TF_EXPECT_OK(reader->Read(2, 4, &result, gtl::string_as_array(&got)));
+ EXPECT_EQ(4, result.size());
+ EXPECT_EQ(content.substr(2, 4), result);
+}
+
+TEST_F(HadoopFileSystemTest, WritableFile) {
+ std::unique_ptr<WritableFile> writer;
+ const string fname =
+ io::JoinPath("hdfs://localfilesystem", testing::TmpDir(), "WritableFile");
+ TF_EXPECT_OK(hdfs.NewWritableFile(fname, &writer));
+ TF_EXPECT_OK(writer->Append("content1,"));
+ TF_EXPECT_OK(writer->Append("content2"));
+ TF_EXPECT_OK(writer->Flush());
+ TF_EXPECT_OK(writer->Sync());
+ TF_EXPECT_OK(writer->Close());
+
+ string content;
+ TF_EXPECT_OK(ReadAll(fname, &content));
+ EXPECT_EQ("content1,content2", content);
+}
+
+TEST_F(HadoopFileSystemTest, FileExists) {
+ const string fname =
+ io::JoinPath("hdfs://localfilesystem", testing::TmpDir(), "FileExists");
+ EXPECT_FALSE(hdfs.FileExists(fname));
+ TF_ASSERT_OK(WriteString(fname, "test"));
+ EXPECT_TRUE(hdfs.FileExists(fname));
+}
+
+TEST_F(HadoopFileSystemTest, GetChildren) {
+ const string base =
+ io::JoinPath("hdfs://localfilesystem", testing::TmpDir(), "GetChildren");
+ TF_EXPECT_OK(hdfs.CreateDir(base));
+
+ const string file = io::JoinPath(base, "testfile.csv");
+ TF_EXPECT_OK(WriteString(file, "blah"));
+ const string subdir = io::JoinPath(base, "subdir");
+ TF_EXPECT_OK(hdfs.CreateDir(subdir));
+
+ vector<string> children;
+ TF_EXPECT_OK(hdfs.GetChildren(base, &children));
+ std::sort(children.begin(), children.end());
+ EXPECT_EQ(vector<string>({"subdir", "testfile.csv"}), children);
+}
+
+TEST_F(HadoopFileSystemTest, DeleteFile) {
+ const string fname =
+ io::JoinPath("hdfs://localfilesystem", testing::TmpDir(), "DeleteFile");
+ EXPECT_FALSE(hdfs.DeleteFile(fname).ok());
+ TF_ASSERT_OK(WriteString(fname, "test"));
+ TF_EXPECT_OK(hdfs.DeleteFile(fname));
+}
+
+TEST_F(HadoopFileSystemTest, GetFileSize) {
+ const string fname =
+ io::JoinPath("hdfs://localfilesystem", testing::TmpDir(), "GetFileSize");
+ TF_ASSERT_OK(WriteString(fname, "test"));
+ uint64 file_size = 0;
+ TF_EXPECT_OK(hdfs.GetFileSize(fname, &file_size));
+ EXPECT_EQ(4, file_size);
+}
+
+TEST_F(HadoopFileSystemTest, CreateDirStat) {
+ const string dir = io::JoinPath("hdfs://localfilesystem", testing::TmpDir(),
+ "CreateDirStat");
+ TF_EXPECT_OK(hdfs.CreateDir(dir));
+ FileStatistics stat;
+ TF_EXPECT_OK(hdfs.Stat(dir, &stat));
+ EXPECT_TRUE(stat.is_directory);
+}
+
+TEST_F(HadoopFileSystemTest, DeleteDir) {
+ const string dir =
+ io::JoinPath("hdfs://localfilesystem", testing::TmpDir(), "DeleteDir");
+ EXPECT_FALSE(hdfs.DeleteDir(dir).ok());
+ TF_EXPECT_OK(hdfs.CreateDir(dir));
+ TF_EXPECT_OK(hdfs.DeleteDir(dir));
+ FileStatistics stat;
+ EXPECT_FALSE(hdfs.Stat(dir, &stat).ok());
+}
+
+TEST_F(HadoopFileSystemTest, RenameFile) {
+ const string fname1 =
+ io::JoinPath("hdfs://localfilesystem", testing::TmpDir(), "RenameFile1");
+ const string fname2 =
+ io::JoinPath("hdfs://localfilesystem", testing::TmpDir(), "RenameFile2");
+ TF_ASSERT_OK(WriteString(fname1, "test"));
+ TF_EXPECT_OK(hdfs.RenameFile(fname1, fname2));
+ string content;
+ TF_EXPECT_OK(ReadAll(fname2, &content));
+ EXPECT_EQ("test", content);
+}
+
+TEST_F(HadoopFileSystemTest, StatFile) {
+ const string fname =
+ io::JoinPath("hdfs://localfilesystem", testing::TmpDir(), "StatFile");
+ TF_ASSERT_OK(WriteString(fname, "test"));
+ FileStatistics stat;
+ TF_EXPECT_OK(hdfs.Stat(fname, &stat));
+ EXPECT_EQ(4, stat.length);
+ EXPECT_FALSE(stat.is_directory);
+}
+
+// NewAppendableFile() is not testable. Local filesystem maps to
+// ChecksumFileSystem in Hadoop, where appending is an unsupported operation.
+
+} // namespace
+} // namespace tensorflow
diff --git a/tensorflow/core/platform/posix/error.cc b/tensorflow/core/platform/posix/error.cc
new file mode 100644
index 0000000000..30d135a255
--- /dev/null
+++ b/tensorflow/core/platform/posix/error.cc
@@ -0,0 +1,163 @@
+/* Copyright 2016 The TensorFlow Authors. All Rights Reserved.
+
+Licensed under the Apache License, Version 2.0 (the "License");
+you may not use this file except in compliance with the License.
+You may obtain a copy of the License at
+
+ http://www.apache.org/licenses/LICENSE-2.0
+
+Unless required by applicable law or agreed to in writing, software
+distributed under the License is distributed on an "AS IS" BASIS,
+WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+See the License for the specific language governing permissions and
+limitations under the License.
+==============================================================================*/
+
+#include "tensorflow/core/platform/posix/error.h"
+
+#include <errno.h>
+#include <string.h>
+
+#include "tensorflow/core/lib/core/status.h"
+#include "tensorflow/core/lib/strings/strcat.h"
+
+namespace tensorflow {
+
+error::Code ErrnoToCode(int err_number) {
+ error::Code code;
+ switch (err_number) {
+ case 0:
+ code = error::OK;
+ break;
+ case EINVAL: // Invalid argument
+ case ENAMETOOLONG: // Filename too long
+ case E2BIG: // Argument list too long
+ case EDESTADDRREQ: // Destination address required
+ case EDOM: // Mathematics argument out of domain of function
+ case EFAULT: // Bad address
+ case EILSEQ: // Illegal byte sequence
+ case ENOPROTOOPT: // Protocol not available
+ case ENOSTR: // Not a STREAM
+ case ENOTSOCK: // Not a socket
+ case ENOTTY: // Inappropriate I/O control operation
+ case EPROTOTYPE: // Protocol wrong type for socket
+ case ESPIPE: // Invalid seek
+ code = error::INVALID_ARGUMENT;
+ break;
+ case ETIMEDOUT: // Connection timed out
+ case ETIME: // Timer expired
+ code = error::DEADLINE_EXCEEDED;
+ break;
+ case ENODEV: // No such device
+ case ENOENT: // No such file or directory
+ case ENXIO: // No such device or address
+ case ESRCH: // No such process
+ code = error::NOT_FOUND;
+ break;
+ case EEXIST: // File exists
+ case EADDRNOTAVAIL: // Address not available
+ case EALREADY: // Connection already in progress
+ code = error::ALREADY_EXISTS;
+ break;
+ case EPERM: // Operation not permitted
+ case EACCES: // Permission denied
+ case EROFS: // Read only file system
+ code = error::PERMISSION_DENIED;
+ break;
+ case ENOTEMPTY: // Directory not empty
+ case EISDIR: // Is a directory
+ case ENOTDIR: // Not a directory
+ case EADDRINUSE: // Address already in use
+ case EBADF: // Invalid file descriptor
+ case EBUSY: // Device or resource busy
+ case ECHILD: // No child processes
+ case EISCONN: // Socket is connected
+ case ENOTBLK: // Block device required
+ case ENOTCONN: // The socket is not connected
+ case EPIPE: // Broken pipe
+ case ESHUTDOWN: // Cannot send after transport endpoint shutdown
+ case ETXTBSY: // Text file busy
+ code = error::FAILED_PRECONDITION;
+ break;
+ case ENOSPC: // No space left on device
+ case EDQUOT: // Disk quota exceeded
+ case EMFILE: // Too many open files
+ case EMLINK: // Too many links
+ case ENFILE: // Too many open files in system
+ case ENOBUFS: // No buffer space available
+ case ENODATA: // No message is available on the STREAM read queue
+ case ENOMEM: // Not enough space
+ case ENOSR: // No STREAM resources
+ case EUSERS: // Too many users
+ code = error::RESOURCE_EXHAUSTED;
+ break;
+ case EFBIG: // File too large
+ case EOVERFLOW: // Value too large to be stored in data type
+ case ERANGE: // Result too large
+ code = error::OUT_OF_RANGE;
+ break;
+ case ENOSYS: // Function not implemented
+ case ENOTSUP: // Operation not supported
+ case EAFNOSUPPORT: // Address family not supported
+ case EPFNOSUPPORT: // Protocol family not supported
+ case EPROTONOSUPPORT: // Protocol not supported
+ case ESOCKTNOSUPPORT: // Socket type not supported
+ case EXDEV: // Improper link
+ code = error::UNIMPLEMENTED;
+ break;
+ case EAGAIN: // Resource temporarily unavailable
+ case ECONNREFUSED: // Connection refused
+ case ECONNABORTED: // Connection aborted
+ case ECONNRESET: // Connection reset
+ case EINTR: // Interrupted function call
+ case EHOSTDOWN: // Host is down
+ case EHOSTUNREACH: // Host is unreachable
+ case ENETDOWN: // Network is down
+ case ENETRESET: // Connection aborted by network
+ case ENETUNREACH: // Network unreachable
+ case ENOLCK: // No locks available
+ case ENOLINK: // Link has been severed
+#if !defined(__APPLE__)
+ case ENONET: // Machine is not on the network
+#endif
+ code = error::UNAVAILABLE;
+ break;
+ case EDEADLK: // Resource deadlock avoided
+ case ESTALE: // Stale file handle
+ code = error::ABORTED;
+ break;
+ case ECANCELED: // Operation cancelled
+ code = error::CANCELLED;
+ break;
+ // NOTE: If you get any of the following (especially in a
+ // reproducible way) and can propose a better mapping,
+ // please email the owners about updating this mapping.
+ case EBADMSG: // Bad message
+ case EIDRM: // Identifier removed
+ case EINPROGRESS: // Operation in progress
+ case EIO: // I/O error
+ case ELOOP: // Too many levels of symbolic links
+ case ENOEXEC: // Exec format error
+ case ENOMSG: // No message of the desired type
+ case EPROTO: // Protocol error
+ case EREMOTE: // Object is remote
+ code = error::UNKNOWN;
+ break;
+ default: {
+ code = error::UNKNOWN;
+ break;
+ }
+ }
+ return code;
+}
+
+Status IOError(const string& context, int err_number) {
+ auto code = ErrnoToCode(err_number);
+ if (code == error::UNKNOWN) {
+ return Status(code, strings::StrCat(context, "; ", strerror(err_number)));
+ } else {
+ return Status(code, context);
+ }
+}
+
+} // namespace tensorflow
diff --git a/tensorflow/core/platform/posix/error.h b/tensorflow/core/platform/posix/error.h
new file mode 100644
index 0000000000..9b614d0f70
--- /dev/null
+++ b/tensorflow/core/platform/posix/error.h
@@ -0,0 +1,27 @@
+/* Copyright 2016 The TensorFlow Authors. All Rights Reserved.
+
+Licensed under the Apache License, Version 2.0 (the "License");
+you may not use this file except in compliance with the License.
+You may obtain a copy of the License at
+
+ http://www.apache.org/licenses/LICENSE-2.0
+
+Unless required by applicable law or agreed to in writing, software
+distributed under the License is distributed on an "AS IS" BASIS,
+WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+See the License for the specific language governing permissions and
+limitations under the License.
+==============================================================================*/
+
+#ifndef TENSORFLOW_CORE_PLATFORM_POSIX_ERROR_H_
+#define TENSORFLOW_CORE_PLATFORM_POSIX_ERROR_H_
+
+#include "tensorflow/core/lib/core/status.h"
+
+namespace tensorflow {
+
+Status IOError(const string& context, int err_number);
+
+} // namespace tensorflow
+
+#endif // TENSORFLOW_CORE_PLATFORM_POSIX_POSIX_FILE_SYSTEM_H_
diff --git a/tensorflow/core/platform/posix/posix_file_system.cc b/tensorflow/core/platform/posix/posix_file_system.cc
index 7ec8d16d33..4bd8d84d6c 100644
--- a/tensorflow/core/platform/posix/posix_file_system.cc
+++ b/tensorflow/core/platform/posix/posix_file_system.cc
@@ -25,143 +25,15 @@ limitations under the License.
#include <unistd.h>
#include "tensorflow/core/lib/core/error_codes.pb.h"
+#include "tensorflow/core/lib/core/status.h"
#include "tensorflow/core/lib/strings/strcat.h"
#include "tensorflow/core/platform/env.h"
#include "tensorflow/core/platform/logging.h"
+#include "tensorflow/core/platform/posix/error.h"
#include "tensorflow/core/platform/posix/posix_file_system.h"
namespace tensorflow {
-namespace {
-
-error::Code ErrnoToCode(int err_number) {
- error::Code code;
- switch (err_number) {
- case 0:
- code = error::OK;
- break;
- case EINVAL: // Invalid argument
- case ENAMETOOLONG: // Filename too long
- case E2BIG: // Argument list too long
- case EDESTADDRREQ: // Destination address required
- case EDOM: // Mathematics argument out of domain of function
- case EFAULT: // Bad address
- case EILSEQ: // Illegal byte sequence
- case ENOPROTOOPT: // Protocol not available
- case ENOSTR: // Not a STREAM
- case ENOTSOCK: // Not a socket
- case ENOTTY: // Inappropriate I/O control operation
- case EPROTOTYPE: // Protocol wrong type for socket
- case ESPIPE: // Invalid seek
- code = error::INVALID_ARGUMENT;
- break;
- case ETIMEDOUT: // Connection timed out
- case ETIME: // Timer expired
- code = error::DEADLINE_EXCEEDED;
- break;
- case ENODEV: // No such device
- case ENOENT: // No such file or directory
- case ENXIO: // No such device or address
- case ESRCH: // No such process
- code = error::NOT_FOUND;
- break;
- case EEXIST: // File exists
- case EADDRNOTAVAIL: // Address not available
- case EALREADY: // Connection already in progress
- code = error::ALREADY_EXISTS;
- break;
- case EPERM: // Operation not permitted
- case EACCES: // Permission denied
- case EROFS: // Read only file system
- code = error::PERMISSION_DENIED;
- break;
- case ENOTEMPTY: // Directory not empty
- case EISDIR: // Is a directory
- case ENOTDIR: // Not a directory
- case EADDRINUSE: // Address already in use
- case EBADF: // Invalid file descriptor
- case EBUSY: // Device or resource busy
- case ECHILD: // No child processes
- case EISCONN: // Socket is connected
- case ENOTBLK: // Block device required
- case ENOTCONN: // The socket is not connected
- case EPIPE: // Broken pipe
- case ESHUTDOWN: // Cannot send after transport endpoint shutdown
- case ETXTBSY: // Text file busy
- code = error::FAILED_PRECONDITION;
- break;
- case ENOSPC: // No space left on device
- case EDQUOT: // Disk quota exceeded
- case EMFILE: // Too many open files
- case EMLINK: // Too many links
- case ENFILE: // Too many open files in system
- case ENOBUFS: // No buffer space available
- case ENODATA: // No message is available on the STREAM read queue
- case ENOMEM: // Not enough space
- case ENOSR: // No STREAM resources
- case EUSERS: // Too many users
- code = error::RESOURCE_EXHAUSTED;
- break;
- case EFBIG: // File too large
- case EOVERFLOW: // Value too large to be stored in data type
- case ERANGE: // Result too large
- code = error::OUT_OF_RANGE;
- break;
- case ENOSYS: // Function not implemented
- case ENOTSUP: // Operation not supported
- case EAFNOSUPPORT: // Address family not supported
- case EPFNOSUPPORT: // Protocol family not supported
- case EPROTONOSUPPORT: // Protocol not supported
- case ESOCKTNOSUPPORT: // Socket type not supported
- case EXDEV: // Improper link
- code = error::UNIMPLEMENTED;
- break;
- case EAGAIN: // Resource temporarily unavailable
- case ECONNREFUSED: // Connection refused
- case ECONNABORTED: // Connection aborted
- case ECONNRESET: // Connection reset
- case EINTR: // Interrupted function call
- case EHOSTDOWN: // Host is down
- case EHOSTUNREACH: // Host is unreachable
- case ENETDOWN: // Network is down
- case ENETRESET: // Connection aborted by network
- case ENETUNREACH: // Network unreachable
- case ENOLCK: // No locks available
- case ENOLINK: // Link has been severed
-#if !defined(__APPLE__)
- case ENONET: // Machine is not on the network
-#endif
- code = error::UNAVAILABLE;
- break;
- case EDEADLK: // Resource deadlock avoided
- case ESTALE: // Stale file handle
- code = error::ABORTED;
- break;
- case ECANCELED: // Operation cancelled
- code = error::CANCELLED;
- break;
- // NOTE: If you get any of the following (especially in a
- // reproducible way) and can propose a better mapping,
- // please email the owners about updating this mapping.
- case EBADMSG: // Bad message
- case EIDRM: // Identifier removed
- case EINPROGRESS: // Operation in progress
- case EIO: // I/O error
- case ELOOP: // Too many levels of symbolic links
- case ENOEXEC: // Exec format error
- case ENOMSG: // No message of the desired type
- case EPROTO: // Protocol error
- case EREMOTE: // Object is remote
- code = error::UNKNOWN;
- break;
- default: {
- code = error::UNKNOWN;
- break;
- }
- }
- return code;
-}
-
// pread() based random-access
class PosixRandomAccessFile : public RandomAccessFile {
private:
@@ -258,8 +130,6 @@ class PosixReadOnlyMemoryRegion : public ReadOnlyMemoryRegion {
const uint64 length_;
};
-} // namespace
-
Status PosixFileSystem::NewRandomAccessFile(
const string& fname, std::unique_ptr<RandomAccessFile>* result) {
string translated_fname = TranslateName(fname);
@@ -401,13 +271,4 @@ Status PosixFileSystem::RenameFile(const string& src, const string& target) {
return result;
}
-Status IOError(const string& context, int err_number) {
- auto code = ErrnoToCode(err_number);
- if (code == error::UNKNOWN) {
- return Status(code, strings::StrCat(context, "; ", strerror(err_number)));
- } else {
- return Status(code, context);
- }
-}
-
} // namespace tensorflow
diff --git a/tensorflow/third_party/hadoop/BUILD b/tensorflow/third_party/hadoop/BUILD
new file mode 100644
index 0000000000..f25208c416
--- /dev/null
+++ b/tensorflow/third_party/hadoop/BUILD
@@ -0,0 +1,20 @@
+package(default_visibility = ["//visibility:public"])
+
+licenses(["notice"]) # Apache 2.0
+
+filegroup(
+ name = "all_files",
+ srcs = glob(
+ ["**/*"],
+ exclude = [
+ "**/METADATA",
+ "**/OWNERS",
+ ],
+ ),
+ visibility = ["//tensorflow:__subpackages__"],
+)
+
+cc_library(
+ name = "hdfs",
+ hdrs = ["hdfs.h"],
+)
diff --git a/tensorflow/third_party/hadoop/hdfs.h b/tensorflow/third_party/hadoop/hdfs.h
new file mode 100644
index 0000000000..560d8bba0e
--- /dev/null
+++ b/tensorflow/third_party/hadoop/hdfs.h
@@ -0,0 +1,911 @@
+/**
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements. See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership. The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License. You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+#ifndef LIBHDFS_HDFS_H
+#define LIBHDFS_HDFS_H
+
+#include <errno.h> /* for EINTERNAL, etc. */
+#include <fcntl.h> /* for O_RDONLY, O_WRONLY */
+#include <stdint.h> /* for uint64_t, etc. */
+#include <time.h> /* for time_t */
+
+/*
+ * Support export of DLL symbols during libhdfs build, and import of DLL symbols
+ * during client application build. A client application may optionally define
+ * symbol LIBHDFS_DLL_IMPORT in its build. This is not strictly required, but
+ * the compiler can produce more efficient code with it.
+ */
+#ifdef WIN32
+#ifdef LIBHDFS_DLL_EXPORT
+#define LIBHDFS_EXTERNAL __declspec(dllexport)
+#elif LIBHDFS_DLL_IMPORT
+#define LIBHDFS_EXTERNAL __declspec(dllimport)
+#else
+#define LIBHDFS_EXTERNAL
+#endif
+#else
+#ifdef LIBHDFS_DLL_EXPORT
+#define LIBHDFS_EXTERNAL __attribute__((visibility("default")))
+#elif LIBHDFS_DLL_IMPORT
+#define LIBHDFS_EXTERNAL __attribute__((visibility("default")))
+#else
+#define LIBHDFS_EXTERNAL
+#endif
+#endif
+
+#ifndef O_RDONLY
+#define O_RDONLY 1
+#endif
+
+#ifndef O_WRONLY
+#define O_WRONLY 2
+#endif
+
+#ifndef EINTERNAL
+#define EINTERNAL 255
+#endif
+
+#define ELASTIC_BYTE_BUFFER_POOL_CLASS \
+ "org/apache/hadoop/io/ElasticByteBufferPool"
+
+/** All APIs set errno to meaningful values */
+
+#ifdef __cplusplus
+extern "C" {
+#endif
+/**
+ * Some utility decls used in libhdfs.
+ */
+struct hdfsBuilder;
+typedef int32_t tSize; /// size of data for read/write io ops
+typedef time_t tTime; /// time type in seconds
+typedef int64_t tOffset; /// offset within the file
+typedef uint16_t tPort; /// port
+typedef enum tObjectKind {
+ kObjectKindFile = 'F',
+ kObjectKindDirectory = 'D',
+} tObjectKind;
+
+/**
+ * The C reflection of org.apache.org.hadoop.FileSystem .
+ */
+struct hdfs_internal;
+typedef struct hdfs_internal *hdfsFS;
+
+struct hdfsFile_internal;
+typedef struct hdfsFile_internal *hdfsFile;
+
+struct hadoopRzOptions;
+
+struct hadoopRzBuffer;
+
+/**
+ * Determine if a file is open for read.
+ *
+ * @param file The HDFS file
+ * @return 1 if the file is open for read; 0 otherwise
+ */
+LIBHDFS_EXTERNAL
+int hdfsFileIsOpenForRead(hdfsFile file);
+
+/**
+ * Determine if a file is open for write.
+ *
+ * @param file The HDFS file
+ * @return 1 if the file is open for write; 0 otherwise
+ */
+LIBHDFS_EXTERNAL
+int hdfsFileIsOpenForWrite(hdfsFile file);
+
+struct hdfsReadStatistics {
+ uint64_t totalBytesRead;
+ uint64_t totalLocalBytesRead;
+ uint64_t totalShortCircuitBytesRead;
+ uint64_t totalZeroCopyBytesRead;
+};
+
+/**
+ * Get read statistics about a file. This is only applicable to files
+ * opened for reading.
+ *
+ * @param file The HDFS file
+ * @param stats (out parameter) on a successful return, the read
+ * statistics. Unchanged otherwise. You must free the
+ * returned statistics with hdfsFileFreeReadStatistics.
+ * @return 0 if the statistics were successfully returned,
+ * -1 otherwise. On a failure, please check errno against
+ * ENOTSUP. webhdfs, LocalFilesystem, and so forth may
+ * not support read statistics.
+ */
+LIBHDFS_EXTERNAL
+int hdfsFileGetReadStatistics(hdfsFile file, struct hdfsReadStatistics **stats);
+
+/**
+ * @param stats HDFS read statistics for a file.
+ *
+ * @return the number of remote bytes read.
+ */
+LIBHDFS_EXTERNAL
+int64_t hdfsReadStatisticsGetRemoteBytesRead(
+ const struct hdfsReadStatistics *stats);
+
+/**
+ * Clear the read statistics for a file.
+ *
+ * @param file The file to clear the read statistics of.
+ *
+ * @return 0 on success; the error code otherwise.
+ * EINVAL: the file is not open for reading.
+ * ENOTSUP: the file does not support clearing the read
+ * statistics.
+ * Errno will also be set to this code on failure.
+ */
+LIBHDFS_EXTERNAL
+int hdfsFileClearReadStatistics(hdfsFile file);
+
+/**
+ * Free some HDFS read statistics.
+ *
+ * @param stats The HDFS read statistics to free.
+ */
+LIBHDFS_EXTERNAL
+void hdfsFileFreeReadStatistics(struct hdfsReadStatistics *stats);
+
+/**
+ * hdfsConnectAsUser - Connect to a hdfs file system as a specific user
+ * Connect to the hdfs.
+ * @param nn The NameNode. See hdfsBuilderSetNameNode for details.
+ * @param port The port on which the server is listening.
+ * @param user the user name (this is hadoop domain user). Or NULL is equivelant
+ * to hhdfsConnect(host, port)
+ * @return Returns a handle to the filesystem or NULL on error.
+ * @deprecated Use hdfsBuilderConnect instead.
+ */
+LIBHDFS_EXTERNAL
+hdfsFS hdfsConnectAsUser(const char *nn, tPort port, const char *user);
+
+/**
+ * hdfsConnect - Connect to a hdfs file system.
+ * Connect to the hdfs.
+ * @param nn The NameNode. See hdfsBuilderSetNameNode for details.
+ * @param port The port on which the server is listening.
+ * @return Returns a handle to the filesystem or NULL on error.
+ * @deprecated Use hdfsBuilderConnect instead.
+ */
+LIBHDFS_EXTERNAL
+hdfsFS hdfsConnect(const char *nn, tPort port);
+
+/**
+ * hdfsConnect - Connect to an hdfs file system.
+ *
+ * Forces a new instance to be created
+ *
+ * @param nn The NameNode. See hdfsBuilderSetNameNode for details.
+ * @param port The port on which the server is listening.
+ * @param user The user name to use when connecting
+ * @return Returns a handle to the filesystem or NULL on error.
+ * @deprecated Use hdfsBuilderConnect instead.
+ */
+LIBHDFS_EXTERNAL
+hdfsFS hdfsConnectAsUserNewInstance(const char *nn, tPort port,
+ const char *user);
+
+/**
+ * hdfsConnect - Connect to an hdfs file system.
+ *
+ * Forces a new instance to be created
+ *
+ * @param nn The NameNode. See hdfsBuilderSetNameNode for details.
+ * @param port The port on which the server is listening.
+ * @return Returns a handle to the filesystem or NULL on error.
+ * @deprecated Use hdfsBuilderConnect instead.
+ */
+LIBHDFS_EXTERNAL
+hdfsFS hdfsConnectNewInstance(const char *nn, tPort port);
+
+/**
+ * Connect to HDFS using the parameters defined by the builder.
+ *
+ * The HDFS builder will be freed, whether or not the connection was
+ * successful.
+ *
+ * Every successful call to hdfsBuilderConnect should be matched with a call
+ * to hdfsDisconnect, when the hdfsFS is no longer needed.
+ *
+ * @param bld The HDFS builder
+ * @return Returns a handle to the filesystem, or NULL on error.
+ */
+LIBHDFS_EXTERNAL
+hdfsFS hdfsBuilderConnect(struct hdfsBuilder *bld);
+
+/**
+ * Create an HDFS builder.
+ *
+ * @return The HDFS builder, or NULL on error.
+ */
+LIBHDFS_EXTERNAL
+struct hdfsBuilder *hdfsNewBuilder(void);
+
+/**
+ * Force the builder to always create a new instance of the FileSystem,
+ * rather than possibly finding one in the cache.
+ *
+ * @param bld The HDFS builder
+ */
+LIBHDFS_EXTERNAL
+void hdfsBuilderSetForceNewInstance(struct hdfsBuilder *bld);
+
+/**
+ * Set the HDFS NameNode to connect to.
+ *
+ * @param bld The HDFS builder
+ * @param nn The NameNode to use.
+ *
+ * If the string given is 'default', the default NameNode
+ * configuration will be used (from the XML configuration files)
+ *
+ * If NULL is given, a LocalFileSystem will be created.
+ *
+ * If the string starts with a protocol type such as file:// or
+ * hdfs://, this protocol type will be used. If not, the
+ * hdfs:// protocol type will be used.
+ *
+ * You may specify a NameNode port in the usual way by
+ * passing a string of the format hdfs://<hostname>:<port>.
+ * Alternately, you may set the port with
+ * hdfsBuilderSetNameNodePort. However, you must not pass the
+ * port in two different ways.
+ */
+LIBHDFS_EXTERNAL
+void hdfsBuilderSetNameNode(struct hdfsBuilder *bld, const char *nn);
+
+/**
+ * Set the port of the HDFS NameNode to connect to.
+ *
+ * @param bld The HDFS builder
+ * @param port The port.
+ */
+LIBHDFS_EXTERNAL
+void hdfsBuilderSetNameNodePort(struct hdfsBuilder *bld, tPort port);
+
+/**
+ * Set the username to use when connecting to the HDFS cluster.
+ *
+ * @param bld The HDFS builder
+ * @param userName The user name. The string will be shallow-copied.
+ */
+LIBHDFS_EXTERNAL
+void hdfsBuilderSetUserName(struct hdfsBuilder *bld, const char *userName);
+
+/**
+ * Set the path to the Kerberos ticket cache to use when connecting to
+ * the HDFS cluster.
+ *
+ * @param bld The HDFS builder
+ * @param kerbTicketCachePath The Kerberos ticket cache path. The string
+ * will be shallow-copied.
+ */
+LIBHDFS_EXTERNAL
+void hdfsBuilderSetKerbTicketCachePath(struct hdfsBuilder *bld,
+ const char *kerbTicketCachePath);
+
+/**
+ * Free an HDFS builder.
+ *
+ * It is normally not necessary to call this function since
+ * hdfsBuilderConnect frees the builder.
+ *
+ * @param bld The HDFS builder
+ */
+LIBHDFS_EXTERNAL
+void hdfsFreeBuilder(struct hdfsBuilder *bld);
+
+/**
+ * Set a configuration string for an HdfsBuilder.
+ *
+ * @param key The key to set.
+ * @param val The value, or NULL to set no value.
+ * This will be shallow-copied. You are responsible for
+ * ensuring that it remains valid until the builder is
+ * freed.
+ *
+ * @return 0 on success; nonzero error code otherwise.
+ */
+LIBHDFS_EXTERNAL
+int hdfsBuilderConfSetStr(struct hdfsBuilder *bld, const char *key,
+ const char *val);
+
+/**
+ * Get a configuration string.
+ *
+ * @param key The key to find
+ * @param val (out param) The value. This will be set to NULL if the
+ * key isn't found. You must free this string with
+ * hdfsConfStrFree.
+ *
+ * @return 0 on success; nonzero error code otherwise.
+ * Failure to find the key is not an error.
+ */
+LIBHDFS_EXTERNAL
+int hdfsConfGetStr(const char *key, char **val);
+
+/**
+ * Get a configuration integer.
+ *
+ * @param key The key to find
+ * @param val (out param) The value. This will NOT be changed if the
+ * key isn't found.
+ *
+ * @return 0 on success; nonzero error code otherwise.
+ * Failure to find the key is not an error.
+ */
+LIBHDFS_EXTERNAL
+int hdfsConfGetInt(const char *key, int32_t *val);
+
+/**
+ * Free a configuration string found with hdfsConfGetStr.
+ *
+ * @param val A configuration string obtained from hdfsConfGetStr
+ */
+LIBHDFS_EXTERNAL
+void hdfsConfStrFree(char *val);
+
+/**
+ * hdfsDisconnect - Disconnect from the hdfs file system.
+ * Disconnect from hdfs.
+ * @param fs The configured filesystem handle.
+ * @return Returns 0 on success, -1 on error.
+ * Even if there is an error, the resources associated with the
+ * hdfsFS will be freed.
+ */
+LIBHDFS_EXTERNAL
+int hdfsDisconnect(hdfsFS fs);
+
+/**
+ * hdfsOpenFile - Open a hdfs file in given mode.
+ * @param fs The configured filesystem handle.
+ * @param path The full path to the file.
+ * @param flags - an | of bits/fcntl.h file flags - supported flags are
+ * O_RDONLY, O_WRONLY (meaning create or overwrite i.e., implies O_TRUNCAT),
+ * O_WRONLY|O_APPEND. Other flags are generally ignored other than (O_RDWR ||
+ * (O_EXCL & O_CREAT)) which return NULL and set errno equal ENOTSUP.
+ * @param bufferSize Size of buffer for read/write - pass 0 if you want
+ * to use the default configured values.
+ * @param replication Block replication - pass 0 if you want to use
+ * the default configured values.
+ * @param blocksize Size of block - pass 0 if you want to use the
+ * default configured values.
+ * @return Returns the handle to the open file or NULL on error.
+ */
+LIBHDFS_EXTERNAL
+hdfsFile hdfsOpenFile(hdfsFS fs, const char *path, int flags, int bufferSize,
+ short replication, tSize blocksize);
+
+/**
+ * hdfsTruncateFile - Truncate a hdfs file to given lenght.
+ * @param fs The configured filesystem handle.
+ * @param path The full path to the file.
+ * @param newlength The size the file is to be truncated to
+ * @return 1 if the file has been truncated to the desired newlength
+ * and is immediately available to be reused for write operations
+ * such as append.
+ * 0 if a background process of adjusting the length of the last
+ * block has been started, and clients should wait for it to
+ * complete before proceeding with further file updates.
+ * -1 on error.
+ */
+int hdfsTruncateFile(hdfsFS fs, const char *path, tOffset newlength);
+
+/**
+ * hdfsUnbufferFile - Reduce the buffering done on a file.
+ *
+ * @param file The file to unbuffer.
+ * @return 0 on success
+ * ENOTSUP if the file does not support unbuffering
+ * Errno will also be set to this value.
+ */
+LIBHDFS_EXTERNAL
+int hdfsUnbufferFile(hdfsFile file);
+
+/**
+ * hdfsCloseFile - Close an open file.
+ * @param fs The configured filesystem handle.
+ * @param file The file handle.
+ * @return Returns 0 on success, -1 on error.
+ * On error, errno will be set appropriately.
+ * If the hdfs file was valid, the memory associated with it will
+ * be freed at the end of this call, even if there was an I/O
+ * error.
+ */
+LIBHDFS_EXTERNAL
+int hdfsCloseFile(hdfsFS fs, hdfsFile file);
+
+/**
+ * hdfsExists - Checks if a given path exsits on the filesystem
+ * @param fs The configured filesystem handle.
+ * @param path The path to look for
+ * @return Returns 0 on success, -1 on error.
+ */
+LIBHDFS_EXTERNAL
+int hdfsExists(hdfsFS fs, const char *path);
+
+/**
+ * hdfsSeek - Seek to given offset in file.
+ * This works only for files opened in read-only mode.
+ * @param fs The configured filesystem handle.
+ * @param file The file handle.
+ * @param desiredPos Offset into the file to seek into.
+ * @return Returns 0 on success, -1 on error.
+ */
+LIBHDFS_EXTERNAL
+int hdfsSeek(hdfsFS fs, hdfsFile file, tOffset desiredPos);
+
+/**
+ * hdfsTell - Get the current offset in the file, in bytes.
+ * @param fs The configured filesystem handle.
+ * @param file The file handle.
+ * @return Current offset, -1 on error.
+ */
+LIBHDFS_EXTERNAL
+tOffset hdfsTell(hdfsFS fs, hdfsFile file);
+
+/**
+ * hdfsRead - Read data from an open file.
+ * @param fs The configured filesystem handle.
+ * @param file The file handle.
+ * @param buffer The buffer to copy read bytes into.
+ * @param length The length of the buffer.
+ * @return On success, a positive number indicating how many bytes
+ * were read.
+ * On end-of-file, 0.
+ * On error, -1. Errno will be set to the error code.
+ * Just like the POSIX read function, hdfsRead will return -1
+ * and set errno to EINTR if data is temporarily unavailable,
+ * but we are not yet at the end of the file.
+ */
+LIBHDFS_EXTERNAL
+tSize hdfsRead(hdfsFS fs, hdfsFile file, void *buffer, tSize length);
+
+/**
+ * hdfsPread - Positional read of data from an open file.
+ * @param fs The configured filesystem handle.
+ * @param file The file handle.
+ * @param position Position from which to read
+ * @param buffer The buffer to copy read bytes into.
+ * @param length The length of the buffer.
+ * @return See hdfsRead
+ */
+LIBHDFS_EXTERNAL
+tSize hdfsPread(hdfsFS fs, hdfsFile file, tOffset position, void *buffer,
+ tSize length);
+
+/**
+ * hdfsWrite - Write data into an open file.
+ * @param fs The configured filesystem handle.
+ * @param file The file handle.
+ * @param buffer The data.
+ * @param length The no. of bytes to write.
+ * @return Returns the number of bytes written, -1 on error.
+ */
+LIBHDFS_EXTERNAL
+tSize hdfsWrite(hdfsFS fs, hdfsFile file, const void *buffer, tSize length);
+
+/**
+ * hdfsWrite - Flush the data.
+ * @param fs The configured filesystem handle.
+ * @param file The file handle.
+ * @return Returns 0 on success, -1 on error.
+ */
+LIBHDFS_EXTERNAL
+int hdfsFlush(hdfsFS fs, hdfsFile file);
+
+/**
+ * hdfsHFlush - Flush out the data in client's user buffer. After the
+ * return of this call, new readers will see the data.
+ * @param fs configured filesystem handle
+ * @param file file handle
+ * @return 0 on success, -1 on error and sets errno
+ */
+LIBHDFS_EXTERNAL
+int hdfsHFlush(hdfsFS fs, hdfsFile file);
+
+/**
+ * hdfsHSync - Similar to posix fsync, Flush out the data in client's
+ * user buffer. all the way to the disk device (but the disk may have
+ * it in its cache).
+ * @param fs configured filesystem handle
+ * @param file file handle
+ * @return 0 on success, -1 on error and sets errno
+ */
+LIBHDFS_EXTERNAL
+int hdfsHSync(hdfsFS fs, hdfsFile file);
+
+/**
+ * hdfsAvailable - Number of bytes that can be read from this
+ * input stream without blocking.
+ * @param fs The configured filesystem handle.
+ * @param file The file handle.
+ * @return Returns available bytes; -1 on error.
+ */
+LIBHDFS_EXTERNAL
+int hdfsAvailable(hdfsFS fs, hdfsFile file);
+
+/**
+ * hdfsCopy - Copy file from one filesystem to another.
+ * @param srcFS The handle to source filesystem.
+ * @param src The path of source file.
+ * @param dstFS The handle to destination filesystem.
+ * @param dst The path of destination file.
+ * @return Returns 0 on success, -1 on error.
+ */
+LIBHDFS_EXTERNAL
+int hdfsCopy(hdfsFS srcFS, const char *src, hdfsFS dstFS, const char *dst);
+
+/**
+ * hdfsMove - Move file from one filesystem to another.
+ * @param srcFS The handle to source filesystem.
+ * @param src The path of source file.
+ * @param dstFS The handle to destination filesystem.
+ * @param dst The path of destination file.
+ * @return Returns 0 on success, -1 on error.
+ */
+LIBHDFS_EXTERNAL
+int hdfsMove(hdfsFS srcFS, const char *src, hdfsFS dstFS, const char *dst);
+
+/**
+ * hdfsDelete - Delete file.
+ * @param fs The configured filesystem handle.
+ * @param path The path of the file.
+ * @param recursive if path is a directory and set to
+ * non-zero, the directory is deleted else throws an exception. In
+ * case of a file the recursive argument is irrelevant.
+ * @return Returns 0 on success, -1 on error.
+ */
+LIBHDFS_EXTERNAL
+int hdfsDelete(hdfsFS fs, const char *path, int recursive);
+
+/**
+ * hdfsRename - Rename file.
+ * @param fs The configured filesystem handle.
+ * @param oldPath The path of the source file.
+ * @param newPath The path of the destination file.
+ * @return Returns 0 on success, -1 on error.
+ */
+LIBHDFS_EXTERNAL
+int hdfsRename(hdfsFS fs, const char *oldPath, const char *newPath);
+
+/**
+ * hdfsGetWorkingDirectory - Get the current working directory for
+ * the given filesystem.
+ * @param fs The configured filesystem handle.
+ * @param buffer The user-buffer to copy path of cwd into.
+ * @param bufferSize The length of user-buffer.
+ * @return Returns buffer, NULL on error.
+ */
+LIBHDFS_EXTERNAL
+char *hdfsGetWorkingDirectory(hdfsFS fs, char *buffer, size_t bufferSize);
+
+/**
+ * hdfsSetWorkingDirectory - Set the working directory. All relative
+ * paths will be resolved relative to it.
+ * @param fs The configured filesystem handle.
+ * @param path The path of the new 'cwd'.
+ * @return Returns 0 on success, -1 on error.
+ */
+LIBHDFS_EXTERNAL
+int hdfsSetWorkingDirectory(hdfsFS fs, const char *path);
+
+/**
+ * hdfsCreateDirectory - Make the given file and all non-existent
+ * parents into directories.
+ * @param fs The configured filesystem handle.
+ * @param path The path of the directory.
+ * @return Returns 0 on success, -1 on error.
+ */
+LIBHDFS_EXTERNAL
+int hdfsCreateDirectory(hdfsFS fs, const char *path);
+
+/**
+ * hdfsSetReplication - Set the replication of the specified
+ * file to the supplied value
+ * @param fs The configured filesystem handle.
+ * @param path The path of the file.
+ * @return Returns 0 on success, -1 on error.
+ */
+LIBHDFS_EXTERNAL
+int hdfsSetReplication(hdfsFS fs, const char *path, int16_t replication);
+
+/**
+ * hdfsFileInfo - Information about a file/directory.
+ */
+typedef struct {
+ tObjectKind mKind; /* file or directory */
+ char *mName; /* the name of the file */
+ tTime mLastMod; /* the last modification time for the file in seconds */
+ tOffset mSize; /* the size of the file in bytes */
+ short mReplication; /* the count of replicas */
+ tOffset mBlockSize; /* the block size for the file */
+ char *mOwner; /* the owner of the file */
+ char *mGroup; /* the group associated with the file */
+ short mPermissions; /* the permissions associated with the file */
+ tTime mLastAccess; /* the last access time for the file in seconds */
+} hdfsFileInfo;
+
+/**
+ * hdfsListDirectory - Get list of files/directories for a given
+ * directory-path. hdfsFreeFileInfo should be called to deallocate memory.
+ * @param fs The configured filesystem handle.
+ * @param path The path of the directory.
+ * @param numEntries Set to the number of files/directories in path.
+ * @return Returns a dynamically-allocated array of hdfsFileInfo
+ * objects; NULL on error.
+ */
+LIBHDFS_EXTERNAL
+hdfsFileInfo *hdfsListDirectory(hdfsFS fs, const char *path, int *numEntries);
+
+/**
+ * hdfsGetPathInfo - Get information about a path as a (dynamically
+ * allocated) single hdfsFileInfo struct. hdfsFreeFileInfo should be
+ * called when the pointer is no longer needed.
+ * @param fs The configured filesystem handle.
+ * @param path The path of the file.
+ * @return Returns a dynamically-allocated hdfsFileInfo object;
+ * NULL on error.
+ */
+LIBHDFS_EXTERNAL
+hdfsFileInfo *hdfsGetPathInfo(hdfsFS fs, const char *path);
+
+/**
+ * hdfsFreeFileInfo - Free up the hdfsFileInfo array (including fields)
+ * @param hdfsFileInfo The array of dynamically-allocated hdfsFileInfo
+ * objects.
+ * @param numEntries The size of the array.
+ */
+LIBHDFS_EXTERNAL
+void hdfsFreeFileInfo(hdfsFileInfo *hdfsFileInfo, int numEntries);
+
+/**
+ * hdfsFileIsEncrypted: determine if a file is encrypted based on its
+ * hdfsFileInfo.
+ * @return -1 if there was an error (errno will be set), 0 if the file is
+ * not encrypted, 1 if the file is encrypted.
+ */
+LIBHDFS_EXTERNAL
+int hdfsFileIsEncrypted(hdfsFileInfo *hdfsFileInfo);
+
+/**
+ * hdfsGetHosts - Get hostnames where a particular block (determined by
+ * pos & blocksize) of a file is stored. The last element in the array
+ * is NULL. Due to replication, a single block could be present on
+ * multiple hosts.
+ * @param fs The configured filesystem handle.
+ * @param path The path of the file.
+ * @param start The start of the block.
+ * @param length The length of the block.
+ * @return Returns a dynamically-allocated 2-d array of blocks-hosts;
+ * NULL on error.
+ */
+LIBHDFS_EXTERNAL
+char ***hdfsGetHosts(hdfsFS fs, const char *path, tOffset start,
+ tOffset length);
+
+/**
+ * hdfsFreeHosts - Free up the structure returned by hdfsGetHosts
+ * @param hdfsFileInfo The array of dynamically-allocated hdfsFileInfo
+ * objects.
+ * @param numEntries The size of the array.
+ */
+LIBHDFS_EXTERNAL
+void hdfsFreeHosts(char ***blockHosts);
+
+/**
+ * hdfsGetDefaultBlockSize - Get the default blocksize.
+ *
+ * @param fs The configured filesystem handle.
+ * @deprecated Use hdfsGetDefaultBlockSizeAtPath instead.
+ *
+ * @return Returns the default blocksize, or -1 on error.
+ */
+LIBHDFS_EXTERNAL
+tOffset hdfsGetDefaultBlockSize(hdfsFS fs);
+
+/**
+ * hdfsGetDefaultBlockSizeAtPath - Get the default blocksize at the
+ * filesystem indicated by a given path.
+ *
+ * @param fs The configured filesystem handle.
+ * @param path The given path will be used to locate the actual
+ * filesystem. The full path does not have to exist.
+ *
+ * @return Returns the default blocksize, or -1 on error.
+ */
+LIBHDFS_EXTERNAL
+tOffset hdfsGetDefaultBlockSizeAtPath(hdfsFS fs, const char *path);
+
+/**
+ * hdfsGetCapacity - Return the raw capacity of the filesystem.
+ * @param fs The configured filesystem handle.
+ * @return Returns the raw-capacity; -1 on error.
+ */
+LIBHDFS_EXTERNAL
+tOffset hdfsGetCapacity(hdfsFS fs);
+
+/**
+ * hdfsGetUsed - Return the total raw size of all files in the filesystem.
+ * @param fs The configured filesystem handle.
+ * @return Returns the total-size; -1 on error.
+ */
+LIBHDFS_EXTERNAL
+tOffset hdfsGetUsed(hdfsFS fs);
+
+/**
+ * Change the user and/or group of a file or directory.
+ *
+ * @param fs The configured filesystem handle.
+ * @param path the path to the file or directory
+ * @param owner User string. Set to NULL for 'no change'
+ * @param group Group string. Set to NULL for 'no change'
+ * @return 0 on success else -1
+ */
+LIBHDFS_EXTERNAL
+int hdfsChown(hdfsFS fs, const char *path, const char *owner,
+ const char *group);
+
+/**
+ * hdfsChmod
+ * @param fs The configured filesystem handle.
+ * @param path the path to the file or directory
+ * @param mode the bitmask to set it to
+ * @return 0 on success else -1
+ */
+LIBHDFS_EXTERNAL
+int hdfsChmod(hdfsFS fs, const char *path, short mode);
+
+/**
+ * hdfsUtime
+ * @param fs The configured filesystem handle.
+ * @param path the path to the file or directory
+ * @param mtime new modification time or -1 for no change
+ * @param atime new access time or -1 for no change
+ * @return 0 on success else -1
+ */
+LIBHDFS_EXTERNAL
+int hdfsUtime(hdfsFS fs, const char *path, tTime mtime, tTime atime);
+
+/**
+ * Allocate a zero-copy options structure.
+ *
+ * You must free all options structures allocated with this function using
+ * hadoopRzOptionsFree.
+ *
+ * @return A zero-copy options structure, or NULL if one could
+ * not be allocated. If NULL is returned, errno will
+ * contain the error number.
+ */
+LIBHDFS_EXTERNAL
+struct hadoopRzOptions *hadoopRzOptionsAlloc(void);
+
+/**
+ * Determine whether we should skip checksums in read0.
+ *
+ * @param opts The options structure.
+ * @param skip Nonzero to skip checksums sometimes; zero to always
+ * check them.
+ *
+ * @return 0 on success; -1 plus errno on failure.
+ */
+LIBHDFS_EXTERNAL
+int hadoopRzOptionsSetSkipChecksum(struct hadoopRzOptions *opts, int skip);
+
+/**
+ * Set the ByteBufferPool to use with read0.
+ *
+ * @param opts The options structure.
+ * @param className If this is NULL, we will not use any
+ * ByteBufferPool. If this is non-NULL, it will be
+ * treated as the name of the pool class to use.
+ * For example, you can use
+ * ELASTIC_BYTE_BUFFER_POOL_CLASS.
+ *
+ * @return 0 if the ByteBufferPool class was found and
+ * instantiated;
+ * -1 plus errno otherwise.
+ */
+LIBHDFS_EXTERNAL
+int hadoopRzOptionsSetByteBufferPool(struct hadoopRzOptions *opts,
+ const char *className);
+
+/**
+ * Free a hadoopRzOptionsFree structure.
+ *
+ * @param opts The options structure to free.
+ * Any associated ByteBufferPool will also be freed.
+ */
+LIBHDFS_EXTERNAL
+void hadoopRzOptionsFree(struct hadoopRzOptions *opts);
+
+/**
+ * Perform a byte buffer read.
+ * If possible, this will be a zero-copy (mmap) read.
+ *
+ * @param file The file to read from.
+ * @param opts An options structure created by hadoopRzOptionsAlloc.
+ * @param maxLength The maximum length to read. We may read fewer bytes
+ * than this length.
+ *
+ * @return On success, we will return a new hadoopRzBuffer.
+ * This buffer will continue to be valid and readable
+ * until it is released by readZeroBufferFree. Failure to
+ * release a buffer will lead to a memory leak.
+ * You can access the data within the hadoopRzBuffer with
+ * hadoopRzBufferGet. If you have reached EOF, the data
+ * within the hadoopRzBuffer will be NULL. You must still
+ * free hadoopRzBuffer instances containing NULL.
+ *
+ * On failure, we will return NULL plus an errno code.
+ * errno = EOPNOTSUPP indicates that we could not do a
+ * zero-copy read, and there was no ByteBufferPool
+ * supplied.
+ */
+LIBHDFS_EXTERNAL
+struct hadoopRzBuffer *hadoopReadZero(hdfsFile file,
+ struct hadoopRzOptions *opts,
+ int32_t maxLength);
+
+/**
+ * Determine the length of the buffer returned from readZero.
+ *
+ * @param buffer a buffer returned from readZero.
+ * @return the length of the buffer.
+ */
+LIBHDFS_EXTERNAL
+int32_t hadoopRzBufferLength(const struct hadoopRzBuffer *buffer);
+
+/**
+ * Get a pointer to the raw buffer returned from readZero.
+ *
+ * To find out how many bytes this buffer contains, call
+ * hadoopRzBufferLength.
+ *
+ * @param buffer a buffer returned from readZero.
+ * @return a pointer to the start of the buffer. This will be
+ * NULL when end-of-file has been reached.
+ */
+LIBHDFS_EXTERNAL
+const void *hadoopRzBufferGet(const struct hadoopRzBuffer *buffer);
+
+/**
+ * Release a buffer obtained through readZero.
+ *
+ * @param file The hdfs stream that created this buffer. This must be
+ * the same stream you called hadoopReadZero on.
+ * @param buffer The buffer to release.
+ */
+LIBHDFS_EXTERNAL
+void hadoopRzBufferFree(hdfsFile file, struct hadoopRzBuffer *buffer);
+
+#ifdef __cplusplus
+}
+#endif
+
+#undef LIBHDFS_EXTERNAL
+#endif /*LIBHDFS_HDFS_H*/
+
+/**
+ * vim: ts=4: sw=4: et
+ */
diff --git a/tensorflow/tools/ci_build/builds/configured b/tensorflow/tools/ci_build/builds/configured
index a8842acdf4..e62a6ffe4d 100755
--- a/tensorflow/tools/ci_build/builds/configured
+++ b/tensorflow/tools/ci_build/builds/configured
@@ -30,6 +30,8 @@ shift 1
# Enable support for Google Cloud Platform (GCP)
export TF_NEED_GCP=1
+# Enable support for HDFS
+export TF_NEED_HDFS=1
if [[ "$1" == "--disable-gcp" ]]; then
export TF_NEED_GCP=0