/* Copyright 2017 The TensorFlow Authors. All Rights Reserved. Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License. ==============================================================================*/ #ifndef TENSORFLOW_COMPILER_XLA_SERVICE_GPU_IR_EMITTER_H_ #define TENSORFLOW_COMPILER_XLA_SERVICE_GPU_IR_EMITTER_H_ #include #include #include #include #include #include "absl/strings/string_view.h" #include "absl/types/span.h" #include "llvm/IR/Function.h" #include "llvm/IR/IRBuilder.h" #include "llvm/IR/Value.h" #include "tensorflow/compiler/xla/service/buffer_assignment.h" #include "tensorflow/compiler/xla/service/dfs_hlo_visitor_with_default.h" #include "tensorflow/compiler/xla/service/gpu/elemental_ir_emitter.h" #include "tensorflow/compiler/xla/service/gpu/hlo_to_ir_bindings.h" #include "tensorflow/compiler/xla/service/gpu/ir_emitter_context.h" #include "tensorflow/compiler/xla/service/gpu/kernel_thunk.h" #include "tensorflow/compiler/xla/service/gpu/thunk.h" #include "tensorflow/compiler/xla/service/hlo_computation.h" #include "tensorflow/compiler/xla/service/hlo_instruction.h" #include "tensorflow/compiler/xla/service/llvm_ir/ir_array.h" #include "tensorflow/compiler/xla/service/llvm_ir/ir_builder_mixin.h" #include "tensorflow/compiler/xla/service/llvm_ir/llvm_loop.h" #include "tensorflow/compiler/xla/service/llvm_ir/loop_emitter.h" #include "tensorflow/compiler/xla/statusor.h" #include "tensorflow/compiler/xla/types.h" #include "tensorflow/compiler/xla/xla_data.pb.h" #include "tensorflow/core/platform/types.h" namespace xla { namespace gpu { // Abstract base class for translating HLO graphs to LLVM IR for a GPU. // // There are two concrete subclasses of IrEmitter: IrEmitterNested and // IrEmitterUnnested. In the unnested variety, each HLO gets its own kernel // function, whereas in the nested version the whole computation is emitted as // one *non-kernel* function. // // In XLA, kernel functions never call other kernel functions. This means that // if we have a kernel -- e.g. implementing a kReduce HLO -- that wants to use // an HLO computation as a "subroutine" -- e.g. the HLO computation that // specifies how to reduce two elements -- then the subroutine computation must // be emitted using IrEmitterNested. // // Fusion nodes are a special case. A fusion node is emitted using // IrEmitterUnnested, but the code is generated using FusedIrEmitter, which is // not a subclass of gpu::IrEmitter, and in fact is better understood as an IR // generator generator. See comments on that class. class IrEmitter : public DfsHloVisitorWithDefault, public IrBuilderMixin { public: IrEmitter(const IrEmitter&) = delete; IrEmitter& operator=(const IrEmitter&) = delete; Status DefaultAction(HloInstruction* hlo) override; Status HandleConstant(HloInstruction* constant) override; Status HandleBitcast(HloInstruction* bitcast) override; Status HandleGetTupleElement(HloInstruction* get_tuple_element) override; Status HandleDot(HloInstruction* dot) override; Status HandleConvolution(HloInstruction* convolution) override; Status HandleFft(HloInstruction* fft) override; Status HandleCrossReplicaSum(HloInstruction* crs) override; Status HandleInfeed(HloInstruction* infeed) override; Status HandleOutfeed(HloInstruction* outfeed) override; Status HandleSend(HloInstruction* send) override; Status HandleSendDone(HloInstruction* send_done) override; Status HandleRecv(HloInstruction* recv) override; Status HandleRecvDone(HloInstruction* recv_done) override; Status HandleParameter(HloInstruction* parameter) override; Status HandleReduce(HloInstruction* reduce) override; Status HandleTuple(HloInstruction* tuple) override; Status HandleScatter(HloInstruction* scatter) override; Status HandleSelect(HloInstruction* select) override; Status HandleTupleSelect(HloInstruction* tuple_select) override; Status HandleFusion(HloInstruction* fusion) override; Status HandleCall(HloInstruction* call) override; Status HandleCustomCall(HloInstruction* custom_call) override; Status HandleBatchNormInference(HloInstruction* batch_norm) override; Status HandleBatchNormTraining(HloInstruction* batch_norm) override; Status HandleBatchNormGrad(HloInstruction* batch_norm) override; Status FinishVisit(HloInstruction* root) override { return Status::OK(); } llvm::IRBuilder<>* builder() { return &b_; } protected: // Constructs an IrEmitter with the given IrEmitter context. // ir_emitter_context is owned by the caller and should outlive the IrEmitter // object. explicit IrEmitter(const HloModuleConfig& hlo_module_config, IrEmitterContext* ir_emitter_context, bool is_nested); // Helper for calling HloToIrBindings::GetIrArray. // // Gets the IrArray which contains inst. This array has metadata that makes // it valid only within the IR that implements consumer. If you are // implementing an HLO and want to get its own output buffer, call // GetIrArray(hlo, hlo). llvm_ir::IrArray GetIrArray(const HloInstruction& inst, const HloInstruction& consumer, const ShapeIndex& shape_index = {}) { return bindings_.GetIrArray(inst, consumer, shape_index); } // A convenient helper for calling HloToIrBindings::GetBasePointer. llvm::Value* GetBasePointer(const HloInstruction& inst) const { return bindings_.GetBasePointer(inst); } // Generates the IrArray for each output of an hlo instruction and returns // a vector containing such IrArrays. std::vector ConstructIrArrayForOutputs( const HloInstruction& hlo); // A convenient helper for calling BufferAssignment::GetUniqueSlice. BufferAllocation::Slice GetAllocationSlice( const HloInstruction& hlo, const ShapeIndex& index = {}) const { return ir_emitter_context_->buffer_assignment() .GetUniqueSlice(&hlo, index) .ConsumeValueOrDie(); } // Emit a singlethreaded or multithreaded loop that computes every element in // the result of the given HLO instruction. This produces a series of nested // loops (e.g. one for each dimension of the `hlo`'s shape). The body of the // inner-most loop is provided by the body_emitter function. virtual Status EmitTargetElementLoop( const HloInstruction& hlo, const llvm_ir::ElementGenerator& body_emitter) = 0; // Emits a call in IR to the given nested computation with the given operands // and output. If no IR function has been previously emitted for the // computation, also emits such a function. Status EmitCallToNestedComputation(const HloComputation& nested_computation, absl::Span operands, llvm::Value* output); // Emits an atomic operation that implements `nested_computation` in the // sequentially consistent memory model. `output_address` and `source_address` // are the arguments of the nested computation. For example, // atomicAdd(output_address, *source_address). Status EmitAtomicOperationForNestedComputation( const HloComputation& nested_computation, llvm::Value* output_address, llvm::Value* source_address); GpuElementalIrEmitter::NestedComputer GetNestedComputer() { return std::bind(&IrEmitter::ComputeNestedElement, this, std::placeholders::_1, std::placeholders::_2); } IrEmitterContext* ir_emitter_context_; llvm::Module* module_; // The following fields track the IR emission state. According to LLVM memory // management rules, their memory is owned by the module. llvm::IRBuilder<> b_; // Mapping from HLO to its underlying LLVM value. HloToIrBindings bindings_; // Hlo configuration data used during code generation. const HloModuleConfig& hlo_module_config_; private: // A helper method for EmitAtomicOperationForNestedComputation. Certain // computations, such as floating-point addition and integer maximization, can // be simply implemented using an LLVM atomic instruction. If "computation" is // one of this kind, emits code to do that and returns true; otherwise, // returns false. bool MaybeEmitDirectAtomicOperation(const HloComputation& computation, llvm::Value* output_address, llvm::Value* source_address); // A helper method for EmitAtomicOperationForNestedComputation. It implements // binary atomic operations using atomicCAS with special handling to support // small data types. Status EmitAtomicOperationUsingCAS(const HloComputation& computation, llvm::Value* output_address, llvm::Value* source_address); // A helper method for HandleSort(). It adds the inner comparison loop where // we compare elements pointed to by 'keys_index' and 'compare_keys_index'. void EmitCompareLoop(int64 dimension_to_sort, const llvm_ir::IrArray::Index& keys_index, const llvm_ir::IrArray::Index& compare_keys_index, const llvm_ir::IrArray& keys_array); StatusOr ComputeNestedElement( const HloComputation& computation, absl::Span parameter_elements); // Emits an atomic operation that implements `nested_computation` in the // sequentially consistent memory model. `output_address` and `source_address` // are the arguments of the nested computation. For example, // atomicAdd(output_address, *source_address). StatusOr EmitAtomicFunctionForNestedComputation( const HloComputation& nested_computation, llvm::Type* element_ir_type); // Map nested computations to emitted IR functions. This serves as a cache so // that IrEmitter does not emit multiple functions for the same // HloComputation. std::map computation_to_ir_function_; }; } // namespace gpu } // namespace xla #endif // TENSORFLOW_COMPILER_XLA_SERVICE_GPU_IR_EMITTER_H_