aboutsummaryrefslogtreecommitdiffhomepage
path: root/tensorflow/compiler
diff options
context:
space:
mode:
authorGravatar Sanjoy Das <sanjoy@google.com>2018-10-03 14:09:05 -0700
committerGravatar TensorFlower Gardener <gardener@tensorflow.org>2018-10-03 14:12:33 -0700
commit94267ccc14516ad9df67897bea8ede20cbad24ca (patch)
tree103969eb919e53f750f7f23e72d9e492c3310f84 /tensorflow/compiler
parented904611009a74ae530335d3bd16b7070238cec3 (diff)
Move out-params to end of argument list and add an out_ prefix; NFC
PiperOrigin-RevId: 215624875
Diffstat (limited to 'tensorflow/compiler')
-rw-r--r--tensorflow/compiler/jit/kernels/xla_ops.cc2
-rw-r--r--tensorflow/compiler/jit/xla_compilation_cache.cc33
-rw-r--r--tensorflow/compiler/jit/xla_compilation_cache.h29
-rw-r--r--tensorflow/compiler/jit/xla_compile_on_demand_op.cc2
4 files changed, 33 insertions, 33 deletions
diff --git a/tensorflow/compiler/jit/kernels/xla_ops.cc b/tensorflow/compiler/jit/kernels/xla_ops.cc
index cfd27a6510..accc86a86d 100644
--- a/tensorflow/compiler/jit/kernels/xla_ops.cc
+++ b/tensorflow/compiler/jit/kernels/xla_ops.cc
@@ -277,7 +277,7 @@ static Status CompileToLocalExecutable(
compile_options.always_return_tuple = false;
return cache->Compile(options, function, constant_args, *variables, ctx,
- kernel, executable, compile_options);
+ compile_options, kernel, executable);
}
void XlaLocalLaunchBase::Compute(OpKernelContext* ctx) {
diff --git a/tensorflow/compiler/jit/xla_compilation_cache.cc b/tensorflow/compiler/jit/xla_compilation_cache.cc
index 3aa9e9c7ed..0471995015 100644
--- a/tensorflow/compiler/jit/xla_compilation_cache.cc
+++ b/tensorflow/compiler/jit/xla_compilation_cache.cc
@@ -228,37 +228,38 @@ Status XlaCompilationCache::Compile(
const XlaCompiler::Options& options, const NameAttrList& function,
const std::map<int, Tensor>& constant_args,
const std::map<int, OptionalTensor>& variable_args, OpKernelContext* ctx,
- const XlaCompiler::CompilationResult** compilation_result,
- xla::LocalExecutable** executable,
- const XlaCompiler::CompileOptions& compile_options) {
+ const XlaCompiler::CompileOptions& compile_options,
+ const XlaCompiler::CompilationResult** out_compilation_result,
+ xla::LocalExecutable** out_executable) {
return CompileImpl(options, function, constant_args, variable_args, ctx,
- compilation_result, executable, compile_options, false);
+ compile_options, /*compile_single_op=*/false,
+ out_compilation_result, out_executable);
}
Status XlaCompilationCache::CompileSingleOp(
const XlaCompiler::Options& options,
const std::map<int, Tensor>& constant_args,
const std::map<int, OptionalTensor>& variable_args, OpKernelContext* ctx,
- const XlaCompiler::CompilationResult** compilation_result,
- xla::LocalExecutable** executable,
- const XlaCompiler::CompileOptions& compile_options) {
+ const XlaCompiler::CompileOptions& compile_options,
+ const XlaCompiler::CompilationResult** out_compilation_result,
+ xla::LocalExecutable** out_executable) {
const NodeDef& def = ctx->op_kernel().def();
NameAttrList name;
name.set_name(def.op());
*name.mutable_attr() = def.attr();
- return CompileImpl(options, name, constant_args, variable_args, ctx,
- compilation_result, executable, compile_options, true);
+ return CompileImpl(
+ options, name, constant_args, variable_args, ctx, compile_options,
+ /*compile_single_op=*/true, out_compilation_result, out_executable);
}
Status XlaCompilationCache::CompileImpl(
const XlaCompiler::Options& options, const NameAttrList& function,
const std::map<int, Tensor>& constant_args,
const std::map<int, OptionalTensor>& variable_args, OpKernelContext* ctx,
- const XlaCompiler::CompilationResult** compilation_result,
- xla::LocalExecutable** executable,
- const XlaCompiler::CompileOptions& compile_options,
- bool compile_single_op) {
- CHECK_NE(executable, nullptr);
+ const XlaCompiler::CompileOptions& compile_options, bool compile_single_op,
+ const XlaCompiler::CompilationResult** out_compilation_result,
+ xla::LocalExecutable** out_executable) {
+ DCHECK_NE(out_executable, nullptr);
VLOG(2) << "XlaCompilationCache::Compile " << DebugString();
if (VLOG_IS_ON(2)) {
@@ -357,8 +358,8 @@ Status XlaCompilationCache::CompileImpl(
}
}
TF_RETURN_IF_ERROR(entry->compilation_status);
- *compilation_result = &entry->compilation_result;
- *executable = entry->executable.get();
+ *out_compilation_result = &entry->compilation_result;
+ *out_executable = entry->executable.get();
return Status::OK();
}
diff --git a/tensorflow/compiler/jit/xla_compilation_cache.h b/tensorflow/compiler/jit/xla_compilation_cache.h
index 17c0321c1e..75c7758f73 100644
--- a/tensorflow/compiler/jit/xla_compilation_cache.h
+++ b/tensorflow/compiler/jit/xla_compilation_cache.h
@@ -68,9 +68,9 @@ class XlaCompilationCache : public ResourceBase {
const std::map<int, Tensor>& constant_args,
const std::map<int, OptionalTensor>& variable_args,
OpKernelContext* ctx,
- const XlaCompiler::CompilationResult** compilation_result,
- xla::LocalExecutable** executable,
- const XlaCompiler::CompileOptions& compile_options);
+ const XlaCompiler::CompileOptions& compile_options,
+ const XlaCompiler::CompilationResult** out_compilation_result,
+ xla::LocalExecutable** out_executable);
// As above, but calls XlaCompiler::CompileSingleOp instead of
// XlaCompiler::CompileFunction.
@@ -78,9 +78,9 @@ class XlaCompilationCache : public ResourceBase {
const XlaCompiler::Options& options,
const std::map<int, Tensor>& constant_args,
const std::map<int, OptionalTensor>& variable_args, OpKernelContext* ctx,
- const XlaCompiler::CompilationResult** compilation_result,
- xla::LocalExecutable** executable,
- const XlaCompiler::CompileOptions& compile_options);
+ const XlaCompiler::CompileOptions& compile_options,
+ const XlaCompiler::CompilationResult** out_compilation_result,
+ xla::LocalExecutable** out_executable);
xla::LocalClient* client() const { return client_; }
const DeviceType& device_type() const { return device_type_; }
@@ -89,15 +89,14 @@ class XlaCompilationCache : public ResourceBase {
private:
// Common implementation of Compile and CompileSingleOp.
- Status CompileImpl(const XlaCompiler::Options& options,
- const NameAttrList& function,
- const std::map<int, Tensor>& constant_args,
- const std::map<int, OptionalTensor>& variable_args,
- OpKernelContext* ctx,
- const XlaCompiler::CompilationResult** compilation_result,
- xla::LocalExecutable** executable,
- const XlaCompiler::CompileOptions& compile_options,
- bool compile_single_op);
+ Status CompileImpl(
+ const XlaCompiler::Options& options, const NameAttrList& function,
+ const std::map<int, Tensor>& constant_args,
+ const std::map<int, OptionalTensor>& variable_args, OpKernelContext* ctx,
+ const XlaCompiler::CompileOptions& compile_options,
+ bool compile_single_op,
+ const XlaCompiler::CompilationResult** out_compilation_result,
+ xla::LocalExecutable** out_executable);
// Takes `result` which has been compiled from a Tensorflow subgraph to a
// XLA computation already, and generates an XLA LocalExecutable `executable`.
diff --git a/tensorflow/compiler/jit/xla_compile_on_demand_op.cc b/tensorflow/compiler/jit/xla_compile_on_demand_op.cc
index b98c0cb028..79976c85df 100644
--- a/tensorflow/compiler/jit/xla_compile_on_demand_op.cc
+++ b/tensorflow/compiler/jit/xla_compile_on_demand_op.cc
@@ -180,7 +180,7 @@ Status XlaCompileOnDemandOp::Compile(
std::map<int, OptionalTensor> variable_args = GetVariables(ctx);
return cache->CompileSingleOp(options, constant_arguments, variable_args, ctx,
- result, executable, compile_options);
+ compile_options, result, executable);
}
void XlaCompileOnDemandOp::Compute(OpKernelContext* ctx) {