aboutsummaryrefslogtreecommitdiffhomepage
path: root/tensorflow/compiler/xla/service/gpu/gpu_executable.cc
diff options
context:
space:
mode:
authorGravatar Peter Hawkins <phawkins@google.com>2017-05-14 19:31:26 -0700
committerGravatar TensorFlower Gardener <gardener@tensorflow.org>2017-05-14 19:35:08 -0700
commit45f4d963f218340cdc3620a9fc0350655023f87d (patch)
treee79be3decdc7e521f1c4c121d6cd5b38ad46a960 /tensorflow/compiler/xla/service/gpu/gpu_executable.cc
parente4633442278246a175ea51db01eb16fcbc01342e (diff)
Automated g4 rollback of changelist 155305963
PiperOrigin-RevId: 156011931
Diffstat (limited to 'tensorflow/compiler/xla/service/gpu/gpu_executable.cc')
-rw-r--r--tensorflow/compiler/xla/service/gpu/gpu_executable.cc13
1 files changed, 1 insertions, 12 deletions
diff --git a/tensorflow/compiler/xla/service/gpu/gpu_executable.cc b/tensorflow/compiler/xla/service/gpu/gpu_executable.cc
index b4b788162f..32f0368b4b 100644
--- a/tensorflow/compiler/xla/service/gpu/gpu_executable.cc
+++ b/tensorflow/compiler/xla/service/gpu/gpu_executable.cc
@@ -107,17 +107,13 @@ class HloExecutionProfiler {
// Implementation note: HLO profiling is always enabled for GPU executables,
// since we can use timers around thunks.
-GpuExecutable::GpuExecutable(tensorflow::StringPiece cubin,
- tensorflow::StringPiece ptx,
- std::pair<int, int> compute_capability,
+GpuExecutable::GpuExecutable(tensorflow::StringPiece ptx,
std::unique_ptr<ThunkSchedule> thunk_schedule,
std::unique_ptr<HloModule> hlo_module,
std::unique_ptr<HloModuleConfig> module_config,
std::unique_ptr<BufferAssignment> assignment)
: Executable(std::move(hlo_module), std::move(module_config)),
- cubin_(cubin),
ptx_(ptx),
- compute_capability_(compute_capability),
thunk_schedule_(std::move(thunk_schedule)),
assignment_(std::move(assignment)) {}
@@ -190,13 +186,6 @@ StatusOr<se::DeviceMemoryBase> GpuExecutable::ExecuteOnStream(
// false.
TF_RET_CHECK(!module_config().has_hybrid_result());
- // Ensure the compute capability of the cubin and the stream match.
- std::pair<int, int> stream_compute_compatibility;
- stream->parent()->GetDeviceDescription().cuda_compute_capability(
- &stream_compute_compatibility.first,
- &stream_compute_compatibility.second);
- TF_RET_CHECK(stream_compute_compatibility == compute_capability_);
-
BufferAllocations::Builder buffer_allocations_builder;
for (BufferAllocation::Index i = 0; i < assignment_->Allocations().size();
++i) {