- Update build rules for tflite diff to compile for iOS, and avoid compiling flex delegate for iOS.
- Add new method for GenerateTestSpec to generate testspec from TFLite. PiperOrigin-RevId: 285884997 Change-Id: I803bd619013f7410bd56283a715e46c8719d4810
This commit is contained in:
parent
f7f7b48e42
commit
3d42bdcff7
@ -219,16 +219,18 @@ cc_library(
|
||||
":join",
|
||||
":split",
|
||||
":test_runner",
|
||||
"@com_google_absl//absl/strings",
|
||||
"//tensorflow/lite:builtin_op_data",
|
||||
"//tensorflow/lite:framework",
|
||||
"//tensorflow/lite:string_util",
|
||||
"//tensorflow/lite/delegates/flex:delegate",
|
||||
"//tensorflow/lite/kernels:builtin_ops",
|
||||
"//tensorflow/lite/kernels:custom_ops",
|
||||
"//tensorflow/lite/kernels:reference_ops",
|
||||
"//tensorflow/lite/tools/evaluation:utils",
|
||||
"@com_google_absl//absl/strings",
|
||||
],
|
||||
] + select({
|
||||
"//tensorflow:ios": [],
|
||||
"//conditions:default": ["//tensorflow/lite/delegates/flex:delegate"],
|
||||
}),
|
||||
)
|
||||
|
||||
tf_cc_test(
|
||||
@ -355,6 +357,7 @@ cc_library(
|
||||
":join",
|
||||
":split",
|
||||
":tf_driver",
|
||||
":tflite_driver",
|
||||
"//tensorflow/lite:string",
|
||||
] + select({
|
||||
"//conditions:default": [
|
||||
@ -403,6 +406,9 @@ cc_library(
|
||||
"//tensorflow:android": [
|
||||
"//tensorflow/core:android_tensorflow_lib",
|
||||
],
|
||||
"//tensorflow:ios": [
|
||||
"//tensorflow/core:ios_tensorflow_lib",
|
||||
],
|
||||
}),
|
||||
)
|
||||
|
||||
@ -435,6 +441,9 @@ cc_library(
|
||||
"//tensorflow:android": [
|
||||
"//tensorflow/core:android_tensorflow_lib",
|
||||
],
|
||||
"//tensorflow:ios": [
|
||||
"//tensorflow/core:ios_tensorflow_lib",
|
||||
],
|
||||
}),
|
||||
)
|
||||
|
||||
|
@ -22,6 +22,7 @@ limitations under the License.
|
||||
#include "tensorflow/lite/testing/join.h"
|
||||
#include "tensorflow/lite/testing/split.h"
|
||||
#include "tensorflow/lite/testing/tf_driver.h"
|
||||
#include "tensorflow/lite/testing/tflite_driver.h"
|
||||
|
||||
namespace tflite {
|
||||
namespace testing {
|
||||
@ -83,6 +84,68 @@ std::vector<string> GenerateInputValues(
|
||||
return input_values;
|
||||
}
|
||||
|
||||
bool GenerateTestSpecFromRunner(std::iostream& stream, int num_invocations,
|
||||
const std::vector<string>& input_layer,
|
||||
const std::vector<string>& input_layer_type,
|
||||
const std::vector<string>& input_layer_shape,
|
||||
const std::vector<string>& output_layer,
|
||||
TestRunner* runner) {
|
||||
stream << "reshape {\n";
|
||||
for (const auto& shape : input_layer_shape) {
|
||||
stream << " input: \"" << shape << "\"\n";
|
||||
}
|
||||
stream << "}\n";
|
||||
|
||||
// Generate inputs.
|
||||
std::mt19937 random_engine;
|
||||
for (int i = 0; i < num_invocations; ++i) {
|
||||
// Note that the input values are random, so each invocation will have a
|
||||
// different set.
|
||||
std::vector<string> input_values = GenerateInputValues(
|
||||
&random_engine, input_layer, input_layer_type, input_layer_shape);
|
||||
if (input_values.empty()) {
|
||||
std::cerr << "Unable to generate input values for the TensorFlow model. "
|
||||
"Make sure the correct values are defined for "
|
||||
"input_layer, input_layer_type, and input_layer_shape."
|
||||
<< std::endl;
|
||||
return false;
|
||||
}
|
||||
|
||||
// Run TensorFlow.
|
||||
auto inputs = runner->GetInputs();
|
||||
for (int j = 0; j < input_values.size(); j++) {
|
||||
runner->SetInput(inputs[j], input_values[j]);
|
||||
if (!runner->IsValid()) {
|
||||
std::cerr << runner->GetErrorMessage() << std::endl;
|
||||
return false;
|
||||
}
|
||||
}
|
||||
|
||||
runner->Invoke();
|
||||
if (!runner->IsValid()) {
|
||||
std::cerr << runner->GetErrorMessage() << std::endl;
|
||||
return false;
|
||||
}
|
||||
|
||||
// Write second part of test spec, with inputs and outputs.
|
||||
stream << "invoke {\n";
|
||||
for (const auto& value : input_values) {
|
||||
stream << " input: \"" << value << "\"\n";
|
||||
}
|
||||
auto outputs = runner->GetOutputs();
|
||||
for (int j = 0; j < output_layer.size(); j++) {
|
||||
stream << " output: \"" << runner->ReadOutput(outputs[j]) << "\"\n";
|
||||
if (!runner->IsValid()) {
|
||||
std::cerr << runner->GetErrorMessage() << std::endl;
|
||||
return false;
|
||||
}
|
||||
}
|
||||
stream << "}\n";
|
||||
}
|
||||
|
||||
return true;
|
||||
}
|
||||
|
||||
} // namespace
|
||||
|
||||
bool GenerateTestSpecFromTensorflowModel(
|
||||
@ -108,61 +171,29 @@ bool GenerateTestSpecFromTensorflowModel(
|
||||
std::cerr << runner.GetErrorMessage() << std::endl;
|
||||
return false;
|
||||
}
|
||||
|
||||
// Write first part of test spec, defining model and input shapes.
|
||||
stream << "load_model: " << tflite_model_path << "\n";
|
||||
stream << "reshape {\n";
|
||||
for (const auto& shape : input_layer_shape) {
|
||||
stream << " input: \"" << shape << "\"\n";
|
||||
return GenerateTestSpecFromRunner(stream, num_invocations, input_layer,
|
||||
input_layer_type, input_layer_shape,
|
||||
output_layer, &runner);
|
||||
}
|
||||
|
||||
bool GenerateTestSpecFromTFLiteModel(
|
||||
std::iostream& stream, const string& tflite_model_path, int num_invocations,
|
||||
const std::vector<string>& input_layer,
|
||||
const std::vector<string>& input_layer_type,
|
||||
const std::vector<string>& input_layer_shape,
|
||||
const std::vector<string>& output_layer) {
|
||||
TfLiteDriver runner;
|
||||
runner.LoadModel(tflite_model_path);
|
||||
if (!runner.IsValid()) {
|
||||
std::cerr << runner.GetErrorMessage() << std::endl;
|
||||
return false;
|
||||
}
|
||||
stream << "}\n";
|
||||
|
||||
// Generate inputs.
|
||||
std::mt19937 random_engine;
|
||||
for (int i = 0; i < num_invocations; ++i) {
|
||||
// Note that the input values are random, so each invocation will have a
|
||||
// different set.
|
||||
std::vector<string> input_values = GenerateInputValues(
|
||||
&random_engine, input_layer, input_layer_type, input_layer_shape);
|
||||
if (input_values.empty()) {
|
||||
std::cerr << "Unable to generate input values for the TensorFlow model. "
|
||||
"Make sure the correct values are defined for "
|
||||
"input_layer, input_layer_type, and input_layer_shape."
|
||||
<< std::endl;
|
||||
return false;
|
||||
}
|
||||
|
||||
// Run TensorFlow.
|
||||
for (int j = 0; j < input_values.size(); j++) {
|
||||
runner.SetInput(j, input_values[j]);
|
||||
if (!runner.IsValid()) {
|
||||
std::cerr << runner.GetErrorMessage() << std::endl;
|
||||
return false;
|
||||
}
|
||||
}
|
||||
|
||||
runner.Invoke();
|
||||
if (!runner.IsValid()) {
|
||||
std::cerr << runner.GetErrorMessage() << std::endl;
|
||||
return false;
|
||||
}
|
||||
|
||||
// Write second part of test spec, with inputs and outputs.
|
||||
stream << "invoke {\n";
|
||||
for (const auto& value : input_values) {
|
||||
stream << " input: \"" << value << "\"\n";
|
||||
}
|
||||
for (int j = 0; j < output_layer.size(); j++) {
|
||||
stream << " output: \"" << runner.ReadOutput(j) << "\"\n";
|
||||
if (!runner.IsValid()) {
|
||||
std::cerr << runner.GetErrorMessage() << std::endl;
|
||||
return false;
|
||||
}
|
||||
}
|
||||
stream << "}\n";
|
||||
}
|
||||
|
||||
return true;
|
||||
runner.AllocateTensors();
|
||||
return GenerateTestSpecFromRunner(stream, num_invocations, input_layer,
|
||||
input_layer_type, input_layer_shape,
|
||||
output_layer, &runner);
|
||||
}
|
||||
|
||||
} // namespace testing
|
||||
|
@ -46,6 +46,14 @@ bool GenerateTestSpecFromTensorflowModel(
|
||||
const std::vector<string>& input_layer_shape,
|
||||
const std::vector<string>& output_layer);
|
||||
|
||||
// Generate test spec by executing TFLite model on random inputs.
|
||||
bool GenerateTestSpecFromTFLiteModel(
|
||||
std::iostream& stream, const string& tflite_model_path, int num_invocations,
|
||||
const std::vector<string>& input_layer,
|
||||
const std::vector<string>& input_layer_type,
|
||||
const std::vector<string>& input_layer_shape,
|
||||
const std::vector<string>& output_layer);
|
||||
|
||||
// Generates random values that are filled into the tensor.
|
||||
template <typename T, typename RandomFunction>
|
||||
std::vector<T> GenerateRandomTensor(const std::vector<int>& shape,
|
||||
|
@ -18,9 +18,12 @@ limitations under the License.
|
||||
#include <complex>
|
||||
#include <memory>
|
||||
#include <vector>
|
||||
|
||||
#include "absl/strings/escaping.h"
|
||||
#include "tensorflow/lite/builtin_op_data.h"
|
||||
#if !defined(__APPLE__)
|
||||
#include "tensorflow/lite/delegates/flex/delegate.h"
|
||||
#endif
|
||||
#include "tensorflow/lite/kernels/custom_ops_register.h"
|
||||
#include "tensorflow/lite/kernels/register.h"
|
||||
#include "tensorflow/lite/kernels/register_ref.h"
|
||||
@ -331,10 +334,12 @@ TfLiteDriver::TfLiteDriver(DelegateType delegate_type, bool reference_kernel)
|
||||
delegate_ = evaluation::CreateGPUDelegate(/*model=*/nullptr);
|
||||
break;
|
||||
case DelegateType::kFlex:
|
||||
#if !defined(__APPLE__)
|
||||
delegate_ = Interpreter::TfLiteDelegatePtr(
|
||||
FlexDelegate::Create().release(), [](TfLiteDelegate* delegate) {
|
||||
delete static_cast<tflite::FlexDelegate*>(delegate);
|
||||
});
|
||||
#endif
|
||||
break;
|
||||
}
|
||||
}
|
||||
|
@ -18,7 +18,9 @@ limitations under the License.
|
||||
#include <map>
|
||||
#include <memory>
|
||||
|
||||
#if !defined(__APPLE__)
|
||||
#include "tensorflow/lite/delegates/flex/delegate.h"
|
||||
#endif
|
||||
#include "tensorflow/lite/interpreter.h"
|
||||
#include "tensorflow/lite/kernels/register.h"
|
||||
#include "tensorflow/lite/kernels/register_ref.h"
|
||||
|
Loading…
Reference in New Issue
Block a user