/* Copyright 2016 The TensorFlow Authors. All Rights Reserved. Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License. ==============================================================================*/ #include "tensorflow/cc/saved_model/loader.h" #include #include "tensorflow/cc/saved_model/constants.h" #include "tensorflow/cc/saved_model/reader.h" #include "tensorflow/core/lib/io/path.h" #include "tensorflow/core/lib/monitoring/counter.h" #include "tensorflow/core/lib/strings/str_util.h" #include "tensorflow/core/lib/strings/strcat.h" #include "tensorflow/core/platform/env.h" #include "tensorflow/core/platform/protobuf_internal.h" #include "tensorflow/core/protobuf/saved_model.pb.h" #include "tensorflow/core/protobuf/saver.pb.h" #include "tensorflow/core/public/session.h" #include "tensorflow/core/public/session_options.h" #include "tensorflow/core/util/tensor_bundle/naming.h" namespace tensorflow { namespace { auto* load_attempt_count = monitoring::Counter<2>::New( "/tensorflow/cc/saved_model/load_attempt_count", "The number of times a SavedModel was successfully loaded.", "model_path", "status"); auto* load_latency = monitoring::Counter<1>::New( "/tensorflow/cc/saved_model/load_latency", "Latency in microseconds for SavedModels that were successfully loaded.", "model_path"); constexpr char kLoadAttemptFail[] = "fail"; constexpr char kLoadAttemptSuccess[] = "success"; Status LoadMetaGraphIntoSession(const MetaGraphDef& meta_graph_def, const SessionOptions& session_options, std::unique_ptr* session) { Session* session_p = nullptr; TF_RETURN_IF_ERROR(NewSession(session_options, &session_p)); session->reset(session_p); return (*session)->Create(meta_graph_def.graph_def()); } Tensor CreateStringTensor(const string& value) { Tensor tensor(DT_STRING, TensorShape({})); tensor.scalar()() = value; return tensor; } void AddAssetsTensorsToInputs(const StringPiece export_dir, const std::vector& asset_file_defs, std::vector>* inputs) { if (asset_file_defs.empty()) { return; } for (auto& asset_file_def : asset_file_defs) { Tensor assets_file_path_tensor = CreateStringTensor(io::JoinPath( export_dir, kSavedModelAssetsDirectory, asset_file_def.filename())); inputs->push_back( {asset_file_def.tensor_info().name(), assets_file_path_tensor}); } } bool HasMainOp(const MetaGraphDef& meta_graph_def) { const auto& collection_def_map = meta_graph_def.collection_def(); if (collection_def_map.find(kSavedModelMainOpKey) != collection_def_map.end()) { return true; } return false; } Status RunMainOp(const RunOptions& run_options, const string& export_dir, const MetaGraphDef& meta_graph_def, const std::vector& asset_file_defs, Session* session, const string& main_op_key) { LOG(INFO) << "Running MainOp with key " << main_op_key << " on SavedModel bundle."; const auto& collection_def_map = meta_graph_def.collection_def(); const auto main_op_it = collection_def_map.find(main_op_key); if (main_op_it != collection_def_map.end()) { if (main_op_it->second.node_list().value_size() != 1) { return errors::FailedPrecondition( strings::StrCat("Expected exactly one main op in : ", export_dir)); } std::vector> inputs; AddAssetsTensorsToInputs(export_dir, asset_file_defs, &inputs); RunMetadata run_metadata; const StringPiece main_op_name = main_op_it->second.node_list().value(0); return session->Run(run_options, inputs, {}, {main_op_name.ToString()}, nullptr /* outputs */, &run_metadata); } return Status::OK(); } Status RunRestore(const RunOptions& run_options, const string& export_dir, const StringPiece restore_op_name, const StringPiece variable_filename_const_op_name, const std::vector& asset_file_defs, Session* session) { LOG(INFO) << "Restoring SavedModel bundle."; // Find path to variables to be restored in export directory. const string variables_directory = io::JoinPath(export_dir, kSavedModelVariablesDirectory); // Check for saver checkpoints in v2 format. Models exported in the checkpoint // v2 format will have a variables.index file. The corresponding // variables are stored in the variables.data-?????-of-????? files. const string variables_index_path = io::JoinPath( variables_directory, MetaFilename(kSavedModelVariablesFilename)); if (!Env::Default()->FileExists(variables_index_path).ok()) { LOG(INFO) << "The specified SavedModel has no variables; no checkpoints " "were restored."; return Status::OK(); } const string variables_path = io::JoinPath(variables_directory, kSavedModelVariablesFilename); // Add variables to the graph. Tensor variables_path_tensor(DT_STRING, TensorShape({})); variables_path_tensor.scalar()() = variables_path; std::vector> inputs = { {variable_filename_const_op_name.ToString(), variables_path_tensor}}; AddAssetsTensorsToInputs(export_dir, asset_file_defs, &inputs); RunMetadata run_metadata; return session->Run(run_options, inputs, {}, {restore_op_name.ToString()}, nullptr /* outputs */, &run_metadata); } Status GetAssetFileDefs(const MetaGraphDef& meta_graph_def, std::vector* asset_file_defs) { const auto& collection_def_map = meta_graph_def.collection_def(); const auto assets_it = collection_def_map.find(kSavedModelAssetsKey); if (assets_it == collection_def_map.end()) { return Status::OK(); } const auto& any_assets = assets_it->second.any_list().value(); for (const auto& any_asset : any_assets) { AssetFileDef asset_file_def; TF_RETURN_IF_ERROR( ParseAny(any_asset, &asset_file_def, "tensorflow.AssetFileDef")); asset_file_defs->push_back(asset_file_def); } return Status::OK(); } Status LoadSavedModelInternal(const SessionOptions& session_options, const RunOptions& run_options, const string& export_dir, const std::unordered_set& tags, SavedModelBundle* const bundle) { TF_RETURN_IF_ERROR(ReadMetaGraphDefFromSavedModel(export_dir, tags, &bundle->meta_graph_def)); TF_RETURN_IF_ERROR(LoadMetaGraphIntoSession( bundle->meta_graph_def, session_options, &bundle->session)); std::vector asset_file_defs; TF_RETURN_IF_ERROR( GetAssetFileDefs(bundle->meta_graph_def, &asset_file_defs)); TF_RETURN_IF_ERROR( RunRestore(run_options, export_dir, bundle->meta_graph_def.saver_def().restore_op_name(), bundle->meta_graph_def.saver_def().filename_tensor_name(), asset_file_defs, bundle->session.get())); if (HasMainOp(bundle->meta_graph_def)) { TF_RETURN_IF_ERROR(RunMainOp(run_options, export_dir, bundle->meta_graph_def, asset_file_defs, bundle->session.get(), kSavedModelMainOpKey)); } else { TF_RETURN_IF_ERROR(RunMainOp( run_options, export_dir, bundle->meta_graph_def, asset_file_defs, bundle->session.get(), kSavedModelLegacyInitOpKey)); } return Status::OK(); } } // namespace Status LoadSavedModel(const SessionOptions& session_options, const RunOptions& run_options, const string& export_dir, const std::unordered_set& tags, SavedModelBundle* const bundle) { // TODO(robson): Add tests for the counters. const uint64 start_microseconds = Env::Default()->NowMicros(); const Status status = LoadSavedModelInternal(session_options, run_options, export_dir, tags, bundle); const uint64 load_latency_microsecs = [&]() -> uint64 { const uint64 end_microseconds = Env::Default()->NowMicros(); // Avoid clock skew. if (end_microseconds < start_microseconds) return 0; return end_microseconds - start_microseconds; }(); auto log_and_count = [&](const string& status_str) { LOG(INFO) << "SavedModel load for tags { " << str_util::Join(tags, " ") << " }; Status: " << status_str << ". Took " << load_latency_microsecs << " microseconds."; load_attempt_count->GetCell(export_dir, status_str)->IncrementBy(1); }; if (status.ok()) { log_and_count(kLoadAttemptSuccess); } else { log_and_count(kLoadAttemptFail); } load_latency->GetCell(export_dir)->IncrementBy(load_latency_microsecs); return status; } bool MaybeSavedModelDirectory(const string& export_dir) { const string saved_model_pb_path = io::JoinPath(export_dir, kSavedModelFilenamePb); const string saved_model_pbtxt_path = io::JoinPath(export_dir, kSavedModelFilenamePbTxt); return Env::Default()->FileExists(saved_model_pb_path).ok() || Env::Default()->FileExists(saved_model_pbtxt_path).ok(); } } // namespace tensorflow