/* Copyright 2017 The TensorFlow Authors. All Rights Reserved. Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License. ==============================================================================*/ #ifndef TENSORFLOW_CONTRIB_LITE_KERNELS_TEST_UTIL_H_ #define TENSORFLOW_CONTRIB_LITE_KERNELS_TEST_UTIL_H_ #include #include #include #include "tensorflow/contrib/lite/interpreter.h" #include "tensorflow/contrib/lite/kernels/internal/tensor_utils.h" #include "tensorflow/contrib/lite/kernels/register.h" #include "tensorflow/contrib/lite/model.h" #include "tensorflow/contrib/lite/string_util.h" #include "tensorflow/contrib/lite/testing/util.h" #include "tensorflow/core/platform/logging.h" namespace tflite { // A gmock matcher that check that elements of a float vector match to a given // tolerance. std::vector<::testing::Matcher> ArrayFloatNear( const std::vector& values, float max_abs_error = 1e-5); template inline std::vector Quantize(const std::vector& data, float scale, int32_t zero_point) { std::vector q; for (float f : data) { q.push_back(static_cast(std::max( std::numeric_limits::min(), std::min(std::numeric_limits::max(), std::round(zero_point + (f / scale)))))); } return q; } template inline std::vector Dequantize(const std::vector& data, float scale, int32_t zero_point) { std::vector f; for (T q : data) { f.push_back(scale * (q - zero_point)); } return f; } // A test model that contains a single operator. All operator inputs and // output are external to the model, so the tests can directly access them. // Typical usage: // SingleOpModel m; // int a = m.AddInput({TensorType_FLOAT32, a_shape}); // int b = m.AddInput({TensorType_FLOAT32, b_shape}); // int c = m.AddOutput({TensorType_FLOAT32, {}}); // m.SetBuiltinOp(...); // m.BuildInterpreter({GetShape(a), GetShape(b)}); // m.PopulateTensor(a, {...}); // m.PopulateTensor(b, {...}); // m.Invoke(); // EXPECT_THAT(m.ExtractVector(c), ArrayFloatNear({...})); // // A helper struct to construct test tensors. This is particularly useful for // quantized tensor which must have their scale and zero_point defined before // the actual data is known. This mimics what happens in practice: quantization // parameters are calculate during training. struct TensorData { TensorType type; std::vector shape; float min; float max; float scale; int32_t zero_point; }; class SingleOpResolver : public OpResolver { public: SingleOpResolver(const BuiltinOperator op, TfLiteRegistration* registration) : op_(op), registration_(*registration) { registration_.builtin_code = static_cast(op); registration_.version = 1; } const TfLiteRegistration* FindOp(BuiltinOperator op, int version) const override { if (op == op_) { return ®istration_; } return nullptr; } const TfLiteRegistration* FindOp(const char* op, int version) const override { return nullptr; } private: const BuiltinOperator op_; TfLiteRegistration registration_; }; class SingleOpModel { public: SingleOpModel() {} ~SingleOpModel() {} // Set a function callback that is run right after graph is prepared // that allows applying external delegates. This is useful for testing // other runtimes like NN API or GPU. void SetApplyDelegate(std::function apply_delegate_fn) { apply_delegate_fn_ = apply_delegate_fn; } // Copying or assignment is disallowed to simplify ownership semantics. SingleOpModel(const SingleOpModel&) = delete; SingleOpModel& operator=(const SingleOpModel&) = delete; // Add a TensorType input tensor and return its index. int AddInput(TensorType type, bool is_variable = false) { return AddInput(TensorData{type}, is_variable); } int AddInput(const TensorData& t, bool is_variable = false); // Templated version of AddConstInput(). template int AddConstInput(TensorType type, std::initializer_list data, std::initializer_list shape) { int id = AddTensor(TensorData{type, shape}, data); inputs_.push_back(id); return id; } // Add a null input tensor (optional input) and return kOptionalTensor. int AddNullInput(); // Add a TensorType output tensor and return its index. int AddOutput(TensorType type) { return AddOutput(TensorData{type}); } int AddOutput(const TensorData& t); template void QuantizeAndPopulate(int index, const std::vector& data) { TfLiteTensor* t = interpreter_->tensor(index); auto q = Quantize(data, t->params.scale, t->params.zero_point); PopulateTensor(index, 0, q.data(), q.data() + q.size()); } void SymmetricQuantizeAndPopulate(int index, const std::vector& data) { TfLiteTensor* t = interpreter_->tensor(index); const int length = data.size(); std::vector q(length); float min, max, scaling_factor; tensor_utils::SymmetricQuantizeFloats(data.data(), length, q.data(), &min, &max, &scaling_factor); // Update quantization params. t->params.scale = scaling_factor; t->params.zero_point = 0; PopulateTensor(index, /*offset=*/0, reinterpret_cast(q.data()), reinterpret_cast(q.data() + q.size())); } const std::vector& GetShape(int id) { return tensor_data_.at(id).shape; } float GetScale(int id) { return tensor_data_.at(id).scale; } int32_t GetZeroPoint(int id) { return tensor_data_.at(id).zero_point; } // Define the operator in this model. void SetBuiltinOp(BuiltinOperator type, BuiltinOptions builtin_options_type, flatbuffers::Offset builtin_options); void SetCustomOp(const string& name, const std::vector& custom_option, const std::function& registeration); // Build the interpreter for this model. Also, resize and allocate all // tensors given the shapes of the inputs. void BuildInterpreter(std::vector> input_shapes, bool allow_fp32_relax_to_fp16 = false); void Invoke(); void PopulateStringTensor(int index, const std::vector& content) { auto tensor = interpreter_->tensor(index); DynamicBuffer buf; for (const string& s : content) { buf.AddString(s.data(), s.length()); } buf.WriteToTensor(tensor); } // Populate the tensor given its index. // TODO(b/110696148) clean up and merge with vector-taking variant below. template void PopulateTensor(int index, const std::initializer_list& data) { T* v = interpreter_->typed_tensor(index); CHECK(v) << "No tensor with index '" << index << "'."; for (T f : data) { *v = f; ++v; } } // Populate the tensor given its index. // TODO(b/110696148) clean up and merge with initializer_list-taking variant // above. template void PopulateTensor(int index, const std::vector& data) { T* v = interpreter_->typed_tensor(index); CHECK(v) << "No tensor with index '" << index << "'."; for (T f : data) { *v = f; ++v; } } // Partially populate the tensor, starting at the given offset. template void PopulateTensor(int index, int offset, T* begin, T* end) { T* v = interpreter_->typed_tensor(index); memcpy(v + offset, begin, (end - begin) * sizeof(T)); } // Return a vector with the flattened contents of a tensor. template std::vector ExtractVector(int index) { T* v = interpreter_->typed_tensor(index); CHECK(v); return std::vector(v, v + GetTensorSize(index)); } std::vector GetTensorShape(int index) { std::vector result; TfLiteTensor* t = interpreter_->tensor(index); for (int i = 0; i < t->dims->size; ++i) { result.push_back(t->dims->data[i]); } return result; } void SetResolver(std::unique_ptr resolver) { resolver_ = std::move(resolver); } protected: int32_t GetTensorSize(int index) const; flatbuffers::FlatBufferBuilder builder_; std::unique_ptr interpreter_; std::unique_ptr resolver_; private: // TODO(gavinbelson): sync this method with // //tensorflow/contrib/lite/kernels/internal/quantization_util.h?l=31 template std::pair QuantizationParams(float f_min, float f_max) { // These are required by many quantized operations. CHECK_LE(f_min, 0); CHECK_GE(f_max, 0); T q_min = std::numeric_limits::min(); T q_max = std::numeric_limits::max(); float range = q_max - q_min; float scale = (f_max - f_min) / range; int32_t zero_point = std::min( q_max, std::max(q_min, static_cast(std::round(q_min - f_min / scale)))); return {scale, zero_point}; } template int AddTensor(TensorData t, std::initializer_list data, bool is_variable = false) { int id = tensors_.size(); // This is slightly different depending on whether we are adding a // quantized or a regular tensor. bool is_quantized = (t.min != 0 || t.max != 0 || t.scale != 0); flatbuffers::Offset q_params = 0; if (is_quantized) { if (t.min != 0 || t.max != 0) { if (t.type == TensorType_UINT8) { std::tie(t.scale, t.zero_point) = QuantizationParams(t.min, t.max); } else if (t.type == TensorType_INT32) { std::tie(t.scale, t.zero_point) = QuantizationParams(t.min, t.max); } else if (t.type == TensorType_INT16) { std::tie(t.scale, t.zero_point) = QuantizationParams(t.min, t.max); } else { LOG(FATAL) << "No support for the requested quantized type"; } t.min = 0; t.max = 0; } q_params = CreateQuantizationParameters( builder_, /*min=*/0, /*max=*/0, builder_.CreateVector({t.scale}), builder_.CreateVector({t.zero_point})); } int buffer_id = 0; if (data.size()) { // Initialize buffers list with empty buffer to allow for non-const // tensors. if (buffers_.empty()) { buffers_.push_back(CreateBuffer(builder_, builder_.CreateVector({}))); } // Add data as a Buffer to buffers list. buffer_id = buffers_.size(); auto data_buffer = builder_.CreateVector(reinterpret_cast(data.begin()), sizeof(T) * data.size()); buffers_.push_back(CreateBuffer(builder_, data_buffer)); } tensors_.push_back(CreateTensor(builder_, builder_.CreateVector(t.shape), t.type, /*buffer=*/buffer_id, /*name=*/0, q_params, is_variable)); tensor_data_[id] = t; return id; } std::map tensor_data_; std::vector inputs_; std::vector outputs_; std::vector> tensors_; std::vector> opcodes_; std::vector> operators_; std::vector> buffers_; std::map> custom_registrations_; // A function pointer that gets called after the interpreter is created but // before evaluation happens. This is useful for applying a delegate. std::function apply_delegate_fn_; }; // Base class for single op unit tests. // The tests are parameterized to test multiple kernels for a single op. // The parameters are strings like "optimized" and "reference" to have better // readability in test reports. // // To use this class: // * Define a constant map from strings to TfLiteRegistration. // * Implement a test class that inherits SingleOpTest. // * Instantiate the test cases with SingleOpTest::GetKernelTags helper // function. // * Call GetRegistration to get the TfLiteRegistration to be used before // building the interpreter. class SingleOpTest : public ::testing::TestWithParam { public: static std::vector GetKernelTags( const std::map& kernel_map) { std::vector tags; for (auto it : kernel_map) { tags.push_back(it.first); } return tags; } protected: virtual const std::map& GetKernelMap() = 0; TfLiteRegistration* GetRegistration() { return GetKernelMap().at(GetParam()); } }; // Strings have a special implementation that is in test_util.cc template <> std::vector SingleOpModel::ExtractVector(int index); } // namespace tflite #endif // TENSORFLOW_CONTRIB_LITE_KERNELS_TEST_UTIL_H_