niobures's picture
ffmpeg (v6.0/v7.x), gensim, numpy, opencv
be94e5d verified
// This file is part of OpenCV project.
// It is subject to the license terms in the LICENSE file found in the top-level directory
// of this distribution and at http://opencv.org/license.html.
#include "../precomp.hpp"
#ifdef HAVE_FLATBUFFERS
#include "schema_generated.h"
#include "builtin_op_data.h"
#endif
#include <opencv2/core/utils/logger.defines.hpp>
#undef CV_LOG_STRIP_LEVEL
#define CV_LOG_STRIP_LEVEL CV_LOG_LEVEL_VERBOSE + 1
#include <opencv2/core/utils/logger.hpp>
namespace cv {
namespace dnn {
CV__DNN_INLINE_NS_BEGIN
#ifdef HAVE_FLATBUFFERS
using namespace opencv_tflite;
class TFLiteImporter {
public:
TFLiteImporter(Net& net, const char* modelBuffer, size_t bufSize);
private:
const opencv_tflite::Model* model;
const flatbuffers::Vector<flatbuffers::Offset<opencv_tflite::Tensor> >* modelTensors;
std::map<int, Mat> allTensors;
Net& dstNet;
// This is a vector of pairs (layerId, outputId) where we iterate over
// indices from TFLite notation and get created OpenCV layers.
std::map<int, std::pair<int, int> > layerIds;
// Tracking of layouts for layers outputs.
std::vector<DataLayout> layouts;
void populateNet();
// Wrap TFLite Tensor to OpenCV Mat without data copying
Mat parseTensor(const Tensor& tensor);
typedef void (TFLiteImporter::*TFLiteImporterNodeParser)(const Operator&, const std::string&, LayerParams&);
typedef std::map<std::string, TFLiteImporterNodeParser> DispatchMap;
const DispatchMap dispatch;
static DispatchMap buildDispatchMap();
void parseConvolution(const Operator& op, const std::string& opcode, LayerParams& layerParams);
void parseDWConvolution(const Operator& op, const std::string& opcode, LayerParams& layerParams);
void parsePadding(const Operator& op, const std::string& opcode, LayerParams& layerParams);
void parseEltwise(const Operator& op, const std::string& opcode, LayerParams& layerParams);
void parsePooling(const Operator& op, const std::string& opcode, LayerParams& layerParams);
void parsePoolingWithArgmax(const Operator& op, const std::string& opcode, LayerParams& layerParams);
void parseUnpooling(const Operator& op, const std::string& opcode, LayerParams& layerParams);
void parseReshape(const Operator& op, const std::string& opcode, LayerParams& layerParams);
void parseConcat(const Operator& op, const std::string& opcode, LayerParams& layerParams);
void parsePack(const Operator& op, const std::string& opcode, LayerParams& layerParams);
void parseResize(const Operator& op, const std::string& opcode, LayerParams& layerParams);
void parseDeconvolution(const Operator& op, const std::string& opcode, LayerParams& layerParams);
void parseQuantize(const Operator& op, const std::string& opcode, LayerParams& layerParams);
void parseDequantize(const Operator& op, const std::string& opcode, LayerParams& layerParams);
void parseDetectionPostProcess(const Operator& op, const std::string& opcode, LayerParams& layerParams);
void parseActivation(const Operator& op, const std::string& opcode, LayerParams& layerParams);
void parseSplit(const Operator& op, const std::string& opcode, LayerParams& layerParams);
void parseFullyConnected(const Operator& op, const std::string& opcode, LayerParams& layerParams);
void parseSoftmax(const Operator& op, const std::string& opcode, LayerParams& layerParams);
void parseCast(const Operator& op, const std::string& opcode, LayerParams& layerParams);
void parseTranspose(const Operator& op, const std::string& opcode, LayerParams& layerParams);
void parseGlobalPooling(const Operator& op, const std::string& opcode, LayerParams& layerParams);
void parseFusedActivation(const Operator& op, ActivationFunctionType activ);
void parseActivation(const Operator& op, const std::string& opcode, LayerParams& layerParams, bool isFused);
void addLayer(LayerParams& layerParams, const Operator& op);
int addPermuteLayer(const std::vector<int>& order, const std::string& permName, const std::pair<int, int>& inpId, int dtype);
int addReshapeLayer(const std::vector<int>& shape, int axis, int num_axes,
const std::string& name, const std::pair<int, int>& inpId, int dtype);
int addFlattenLayer(int axis, int end_axis, const std::string& name, const std::pair<int, int>& inpId, int dtype);
inline bool isInt8(const Operator& op);
inline void getQuantParams(const Operator& op, float& inpScale, int& inpZero, float& outScale, int& outZero);
};
Mat TFLiteImporter::parseTensor(const Tensor& tensor)
{
const auto tensor_shape = tensor.shape();
CV_Assert(tensor_shape);
std::vector<int> shape(tensor_shape->begin(), tensor_shape->end());
int bufferIdx = tensor.buffer();
CV_Assert(bufferIdx != 0); // 0th buffer is a no-data buffer
const Buffer* buffer = model->buffers()->Get(bufferIdx);
CV_Assert(buffer);
const auto buffer_data = buffer->data();
if (!buffer_data)
return Mat();
const void* data = buffer_data->data();
int dtype = -1;
switch (tensor.type()) {
case TensorType_FLOAT32:
dtype = CV_32F;
break;
case TensorType_INT32:
dtype = CV_32S;
break;
case TensorType_FLOAT16:
dtype = CV_16F;
break;
case TensorType_INT8:
dtype = CV_8S;
break;
default:
CV_Error(Error::StsNotImplemented, format("Parse tensor with type %s", EnumNameTensorType(tensor.type())));
}
return shape.empty() ? Mat() : Mat(shape, dtype, const_cast<void*>(data));
}
TFLiteImporter::TFLiteImporter(Net& dstNet, const char* modelBuffer, size_t bufSize)
: dstNet(dstNet), dispatch(buildDispatchMap())
{
flatbuffers::Verifier verifier((const uint8_t*)modelBuffer, bufSize);
if (!VerifyModelBuffer(verifier)) {
CV_Error(Error::StsError, "DNN/TFLite: model is incorrect");
}
model = GetModel(modelBuffer);
CV_Assert(model);
CV_Assert(model->subgraphs());
CV_Assert(model->buffers());
CV_CheckEQ((size_t)model->subgraphs()->size(), 1u, "");
modelTensors = model->subgraphs()->Get(0)->tensors();
CV_Assert(modelTensors);
for (int i = 0; i < modelTensors->size(); ++i) {
const Tensor* tensor = modelTensors->Get(i);
CV_Assert(tensor);
if (tensor->buffer() != 0) {
allTensors[i] = parseTensor(*tensor);
}
}
populateNet();
}
DataLayout estimateLayout(const Tensor& t)
{
const auto t_shape = t.shape();
CV_Assert(t_shape);
switch (t_shape->size()) {
case 5: return DNN_LAYOUT_NDHWC;
case 4: return DNN_LAYOUT_NHWC;
case 2: return DNN_LAYOUT_PLANAR;
default: return DNN_LAYOUT_UNKNOWN;
}
}
void TFLiteImporter::populateNet()
{
CV_Assert(model);
const auto model_subgraphs = model->subgraphs();
CV_Assert(model_subgraphs);
const SubGraph* subgraph = model_subgraphs->Get(0);
CV_Assert(subgraph);
const auto subgraph_inputs = subgraph->inputs();
CV_Assert(subgraph_inputs);
const auto subgraph_operators = subgraph->operators();
CV_Assert(subgraph_operators);
const auto opCodes = model->operator_codes();
CV_Assert(opCodes);
CV_Assert(modelTensors);
layouts.resize(modelTensors->size(), DNN_LAYOUT_UNKNOWN);
size_t subgraph_inputs_size = subgraph_inputs->size();
std::vector<std::string> inputsNames(subgraph_inputs_size);
std::vector<MatShape> inputsShapes(subgraph_inputs_size);
for (size_t i = 0; i < subgraph_inputs_size; ++i)
{
int idx = subgraph_inputs->Get(i);
layerIds[idx] = std::make_pair(0, i);
const auto tensor = modelTensors->Get(idx);
if (!tensor)
CV_Error(Error::StsError, cv::format("DNN/TFLite: subgraph input %d (%d) is NULL", (int)i, idx));
layouts[idx] = estimateLayout(*tensor);
// Keep info about origin inputs names and shapes
inputsNames[i] = tensor->name()->str();
std::vector<int> shape(tensor->shape()->begin(), tensor->shape()->end());
if (layouts[idx] == DNN_LAYOUT_NHWC) {
CV_CheckEQ(shape.size(), (size_t)4, "");
std::swap(shape[2], shape[3]);
std::swap(shape[1], shape[2]);
}
inputsShapes[i] = shape;
}
dstNet.setInputsNames(inputsNames);
for (size_t i = 0; i < subgraph_inputs_size; ++i)
{
dstNet.setInputShape(inputsNames[i], inputsShapes[i]);
}
const auto& all_operators = *subgraph_operators;
const size_t all_operators_size = all_operators.size();
for (size_t op_idx = 0; op_idx < all_operators_size; ++op_idx)
{
const auto op = all_operators[op_idx];
CV_Assert(op);
const auto op_inputs = op->inputs();
CV_Assert(op_inputs);
const auto op_outputs = op->outputs();
CV_Assert(op_outputs);
int idx = op->opcode_index();
LayerParams layerParams;
layerParams.name = modelTensors->Get(op_outputs->Get(0))->name()->str();
std::string type = EnumNameBuiltinOperator(BuiltinOperator(opCodes->Get(idx)->deprecated_builtin_code()));
if (type == "CUSTOM") {
type = opCodes->Get(idx)->custom_code()->str();
}
CV_LOG_DEBUG(NULL, "DNN/TFLite: processing operator (" << op_idx << "/" << all_operators_size << ") with " << op_inputs->size() << " inputs: "
<< cv::format("[%s]:(%s)", type.c_str(), layerParams.name.c_str()));
try
{
if (type == "DEQUANTIZE") {
// Convert from FP16 to FP32
Mat data = allTensors[op_inputs->Get(0)];
if (!data.empty()) {
// Dequantize a buffer
Mat dataFP32;
data.convertTo(dataFP32, CV_32F);
allTensors[op_outputs->Get(0)] = dataFP32;
continue;
}
}
DispatchMap::const_iterator iter = dispatch.find(type);
if (iter == dispatch.end())
CV_Error(Error::StsNotImplemented, "Unsupported operator type " + type);
CALL_MEMBER_FN(*this, iter->second)(*op, type, layerParams);
}
catch (const cv::Exception& e)
{
CV_LOG_ERROR(NULL, "DNN/TFLite: Problem during import of operator "
<< cv::format("[%s]:(%s)", type.c_str(), layerParams.name.c_str())
<< " (" << op_idx << "/" << all_operators_size << "). Exception: " << e.what());
if (DNN_DIAGNOSTICS_RUN)
{
continue;
}
throw;
}
}
}
TFLiteImporter::DispatchMap TFLiteImporter::buildDispatchMap()
{
static DispatchMap dispatch;
if (!dispatch.empty())
return dispatch;
dispatch["CONV_2D"] = &TFLiteImporter::parseConvolution;
dispatch["DEPTHWISE_CONV_2D"] = &TFLiteImporter::parseDWConvolution;
dispatch["ADD"] = dispatch["MUL"] = &TFLiteImporter::parseEltwise;
dispatch["RELU"] = dispatch["PRELU"] = dispatch["HARD_SWISH"] =
dispatch["LOGISTIC"] = &TFLiteImporter::parseActivation;
dispatch["MAX_POOL_2D"] = dispatch["AVERAGE_POOL_2D"] = &TFLiteImporter::parsePooling;
dispatch["MaxPoolingWithArgmax2D"] = &TFLiteImporter::parsePoolingWithArgmax;
dispatch["MaxUnpooling2D"] = &TFLiteImporter::parseUnpooling;
dispatch["PAD"] = &TFLiteImporter::parsePadding;
dispatch["RESHAPE"] = &TFLiteImporter::parseReshape;
dispatch["CONCATENATION"] = &TFLiteImporter::parseConcat;
dispatch["PACK"] = &TFLiteImporter::parsePack;
dispatch["RESIZE_BILINEAR"] = dispatch["RESIZE_NEAREST_NEIGHBOR"] = &TFLiteImporter::parseResize;
dispatch["Convolution2DTransposeBias"] = &TFLiteImporter::parseDeconvolution;
dispatch["QUANTIZE"] = &TFLiteImporter::parseQuantize;
dispatch["DEQUANTIZE"] = &TFLiteImporter::parseDequantize;
dispatch["SPLIT"] = &TFLiteImporter::parseSplit;
dispatch["FULLY_CONNECTED"] = &TFLiteImporter::parseFullyConnected;
dispatch["SOFTMAX"] = &TFLiteImporter::parseSoftmax;
dispatch["CAST"] = &TFLiteImporter::parseCast;
dispatch["TFLite_Detection_PostProcess"] = &TFLiteImporter::parseDetectionPostProcess;
dispatch["TRANSPOSE"] = &TFLiteImporter::parseTranspose;
dispatch["MEAN"] = dispatch["REDUCE_MAX"] = &TFLiteImporter::parseGlobalPooling;
return dispatch;
}
void TFLiteImporter::addLayer(LayerParams& layerParams, const Operator& op) {
const auto op_inputs = op.inputs();
const auto op_outputs = op.outputs();
// Collect input blobs
if (layerParams.blobs.empty()) {
for (int idx : *op_inputs) {
if (layerIds.find(idx) != layerIds.end()) {
continue; // Output from a different layer
}
Mat blob = allTensors[idx];
layerParams.blobs.push_back(blob.u ? blob : blob.clone()); // some tensors are owned by OpenCV
}
} else {
for (auto& blob : layerParams.blobs) {
CV_Assert(blob.u);
}
}
int dtype = CV_32F;
if (isInt8(op)) {
dtype = CV_8S;
if (layerParams.type != "Quantize")
layerParams.type += "Int8";
if (!layerParams.has("zeropoints")) {
float inpScale, outScale;
int inpZero, outZero;
getQuantParams(op, inpScale, inpZero, outScale, outZero);
layerParams.set("input_scale", inpScale);
layerParams.set("input_zeropoint", inpZero);
layerParams.set("scales", outScale);
layerParams.set("zeropoints", outZero);
}
}
int layerId = dstNet.addLayer(layerParams.name, layerParams.type, dtype, layerParams);
// Connect layer to inputs
int i = 0;
std::vector<DataLayout> inpLayouts;
for (int idx : *op_inputs) {
if (layerIds.find(idx) == layerIds.end()) {
continue; // Const input
}
inpLayouts.push_back(layouts[idx]);
auto it = layerIds.find(idx);
CV_Assert(it != layerIds.end());
dstNet.connect(it->second.first, it->second.second, layerId, i++);
}
// Predict output layout. Some layer-specific parsers may set them explicitly.
// Otherwise, propagate input layout.
if (layouts[op_outputs->Get(0)] == DNN_LAYOUT_UNKNOWN) {
DataLayout predictedLayout = DNN_LAYOUT_UNKNOWN;
for (auto layout : inpLayouts) {
if (layout != DNN_LAYOUT_UNKNOWN) {
if (predictedLayout == DNN_LAYOUT_UNKNOWN)
predictedLayout = layout;
else if (predictedLayout != layout) {
predictedLayout = DNN_LAYOUT_UNKNOWN;
break;
}
}
}
layouts[op_outputs->Get(0)] = predictedLayout;
}
// Register outputs
i = 0;
for (int idx : *op_outputs) {
layerIds[idx] = std::make_pair(layerId, i++);
}
}
void TFLiteImporter::parseConvolution(const Operator& op, const std::string& opcode, LayerParams& layerParams) {
layerParams.type = "Convolution";
auto options = reinterpret_cast<const Conv2DOptions*>(op.builtin_options());
layerParams.set("pad_mode", EnumNamePadding(options->padding()));
layerParams.set("stride_w", options->stride_w());
layerParams.set("stride_h", options->stride_h());
layerParams.set("dilation_w", options->dilation_w_factor());
layerParams.set("dilation_h", options->dilation_h_factor());
// Get filter size
int filterIdx = op.inputs()->Get(1);
Mat filter = allTensors[filterIdx];
int oc = filter.size[0];
int kh = filter.size[1];
int kw = filter.size[2];
layerParams.set("kernel_w", kw);
layerParams.set("kernel_h", kh);
layerParams.set("num_output", oc);
bool isInt8 = filter.depth() == CV_8S;
// Fill convolutions blobs here because of two reasons:
// 1. Kernel transposition
// 2. Extra blob with kernel scales in case of INT8 mode
bool hasBias = op.inputs()->size() > 2;
layerParams.blobs.resize(1 + (int)hasBias + (int)isInt8);
if (hasBias) {
Mat bias = allTensors[op.inputs()->Get(2)];
layerParams.blobs[1] = bias.u ? bias : bias.clone();
}
// Reorder filter data from OHWI to OIHW and change shape correspondingly.
transposeND(filter, {0, 3, 1, 2}, layerParams.blobs[0]);
if (isInt8) {
float inpScale, outScale;
int inpZero, outZero;
getQuantParams(op, inpScale, inpZero, outScale, outZero);
layerParams.blobs[2] = Mat(oc, 1, CV_32F);
auto filterScales = modelTensors->Get(filterIdx)->quantization()->scale();
if (filterScales->size() == 1) {
layerParams.blobs[2].setTo(inpScale * filterScales->Get(0) / outScale);
} else {
for (size_t i = 0; i < filterScales->size(); ++i) {
layerParams.blobs[2].at<float>(i) = inpScale * filterScales->Get(i) / outScale;
}
}
if (hasBias) {
Mat bias = layerParams.blobs[1].reshape(1, oc);
Mat weights_2d = layerParams.blobs[0].reshape(1, oc);
for (int i = 0; i < oc; i++)
{
bias.at<int>(i) -= inpZero * (cv::sum(weights_2d.row(i))[0]);
}
}
}
addLayer(layerParams, op);
parseFusedActivation(op, options->fused_activation_function());
}
void TFLiteImporter::parseDWConvolution(const Operator& op, const std::string& opcode, LayerParams& layerParams) {
layerParams.type = "Convolution";
auto options = reinterpret_cast<const DepthwiseConv2DOptions*>(op.builtin_options());
layerParams.set("pad_mode", EnumNamePadding(options->padding()));
layerParams.set("stride_w", options->stride_w());
layerParams.set("stride_h", options->stride_h());
layerParams.set("dilation_w", options->dilation_w_factor());
layerParams.set("dilation_h", options->dilation_h_factor());
int filterIdx = op.inputs()->Get(1);
Mat filter = allTensors[filterIdx];
int kh = filter.size[1];
int kw = filter.size[2];
int oc = filter.size[3];
layerParams.set("kernel_w", kw);
layerParams.set("kernel_h", kh);
layerParams.set("num_output", oc);
layerParams.set("group", oc);
bool isInt8 = filter.depth() == CV_8S;
// Fill convolutions blobs here because of two reasons:
// 1. Kernel transposition
// 2. Extra blob with kernel scales in case of INT8 mode
bool hasBias = op.inputs()->size() > 2;
layerParams.blobs.resize(1 + (int)hasBias + (int)isInt8);
if (hasBias) {
Mat bias = allTensors[op.inputs()->Get(2)];
layerParams.blobs[1] = bias.u ? bias : bias.clone();
}
transposeND(filter, {3, 0, 1, 2}, layerParams.blobs[0]);
if (isInt8) {
float inpScale, outScale;
int inpZero, outZero;
getQuantParams(op, inpScale, inpZero, outScale, outZero);
layerParams.blobs[2] = Mat(oc, 1, CV_32F);
auto filterScales = modelTensors->Get(filterIdx)->quantization()->scale();
if (filterScales->size() == 1) {
layerParams.blobs[2].setTo(inpScale * filterScales->Get(0) / outScale);
} else {
for (size_t i = 0; i < filterScales->size(); ++i) {
layerParams.blobs[2].at<float>(i) = inpScale * filterScales->Get(i) / outScale;
}
}
if (hasBias) {
Mat bias = layerParams.blobs[1].reshape(1, oc);
Mat weights_2d = layerParams.blobs[0].reshape(1, oc);
for (int i = 0; i < oc; i++)
{
bias.at<int>(i) -= inpZero * (cv::sum(weights_2d.row(i))[0]);
}
}
}
addLayer(layerParams, op);
parseFusedActivation(op, options->fused_activation_function());
}
void TFLiteImporter::parsePadding(const Operator& op, const std::string& opcode, LayerParams& layerParams) {
layerParams.type = "Padding";
Mat paddings = allTensors[op.inputs()->Get(1)].clone();
CV_CheckTypeEQ(paddings.type(), CV_32S, "");
// N H W C
// 0 1 2 3 4 5 6 7
std::swap(paddings.at<int32_t>(2), paddings.at<int32_t>(6));
std::swap(paddings.at<int32_t>(3), paddings.at<int32_t>(7));
// N C W H
// 0 1 2 3 4 5 6 7
std::swap(paddings.at<int32_t>(4), paddings.at<int32_t>(6));
std::swap(paddings.at<int32_t>(5), paddings.at<int32_t>(7));
// N C H W
// 0 1 2 3 4 5 6 7
layerParams.set("paddings", DictValue::arrayInt<int32_t*>((int32_t*)paddings.data, paddings.total()));
addLayer(layerParams, op);
}
void TFLiteImporter::parseEltwise(const Operator& op, const std::string& opcode, LayerParams& layerParams) {
ActivationFunctionType activ = ActivationFunctionType_NONE;
layerParams.type = "Eltwise";
if (opcode == "ADD") {
auto options = reinterpret_cast<const AddOptions*>(op.builtin_options());
activ = options->fused_activation_function();
layerParams.set("operation", "sum");
}
else if (opcode == "MUL") {
auto options = reinterpret_cast<const MulOptions*>(op.builtin_options());
activ = options->fused_activation_function();
layerParams.set("operation", "prod");
} else {
CV_Error(Error::StsNotImplemented, "Unknown opcode for Eltwise layer: " + opcode);
}
if (isInt8(op)) {
const Tensor* out = modelTensors->Get(op.outputs()->Get(0));
float outScale = out->quantization()->scale()->Get(0);
int outZero = out->quantization()->zero_point()->Get(0);
const size_t numInps = op.inputs()->size();
std::vector<float> inputScales(numInps);
std::vector<int> inputZeros(numInps);
std::vector<float> coeffs(numInps);
float offset = outZero;
for (int i = 0; i < numInps; ++i) {
const Tensor* inp = modelTensors->Get(op.inputs()->Get(i));
float inpScale = inp->quantization()->scale()->Get(0);
int inpZero = inp->quantization()->zero_point()->Get(0);
inputScales[i] = inpScale;
inputZeros[i] = inpZero;
coeffs[i] = inpScale / outScale;
offset -= coeffs[i] * inpZero;
}
layerParams.set("input_scales", DictValue::arrayReal(inputScales.data(), numInps));
layerParams.set("input_zeropoints", DictValue::arrayInt(inputZeros.data(), numInps));
layerParams.set("coeff", DictValue::arrayReal(coeffs.data(), numInps));
layerParams.set("offset", offset);
layerParams.set("scales", outScale);
layerParams.set("zeropoints", outZero);
}
addLayer(layerParams, op);
parseFusedActivation(op, activ);
}
void TFLiteImporter::parsePooling(const Operator& op, const std::string& opcode, LayerParams& layerParams) {
layerParams.type = "Pooling";
auto options = reinterpret_cast<const Pool2DOptions*>(op.builtin_options());
layerParams.set("pad_mode", EnumNamePadding(options->padding()));
layerParams.set("stride_w", options->stride_w());
layerParams.set("stride_h", options->stride_h());
layerParams.set("kernel_w", options->filter_width());
layerParams.set("kernel_h", options->filter_height());
if (opcode == "MAX_POOL_2D")
layerParams.set("pool", "max");
else if (opcode == "AVERAGE_POOL_2D")
layerParams.set("pool", "ave");
else
CV_Error(Error::StsNotImplemented, "Pool type selection for " + opcode);
addLayer(layerParams, op);
parseFusedActivation(op, options->fused_activation_function());
}
void TFLiteImporter::parsePoolingWithArgmax(const Operator& op, const std::string& opcode, LayerParams& layerParams) {
layerParams.type = "Pooling";
CV_CheckLE(op.custom_options()->size(), sizeof(TfLitePoolParams), "");
const auto* params = reinterpret_cast<const TfLitePoolParams*>(op.custom_options()->Data());
if (params->activation != kTfLiteActNone) {
CV_Error(Error::StsNotImplemented, "Argmax pooling with fused activation");
}
if (params->padding != kTfLitePaddingUnknown)
layerParams.set("pad_mode", params->padding == kTfLitePaddingSame ? "SAME" : "VALID");
layerParams.set("stride_w", params->stride_width);
layerParams.set("stride_h", params->stride_height);
layerParams.set("kernel_w", params->filter_width);
layerParams.set("kernel_h", params->filter_height);
layerParams.set("pool", "max");
addLayer(layerParams, op);
}
void TFLiteImporter::parseUnpooling(const Operator& op, const std::string& opcode, LayerParams& layerParams) {
layerParams.type = "MaxUnpool";
CV_CheckLE(op.custom_options()->size(), sizeof(TfLitePoolParams), "");
const auto* params = reinterpret_cast<const TfLitePoolParams*>(op.custom_options()->Data());
if (params->activation != kTfLiteActNone) {
CV_Error(Error::StsNotImplemented, "Unpooling with fused activation");
}
layerParams.set("pool_stride_w", params->stride_width);
layerParams.set("pool_stride_h", params->stride_height);
layerParams.set("pool_k_w", params->filter_width);
layerParams.set("pool_k_h", params->filter_height);
layerParams.set("pool_pad_w", 0);
layerParams.set("pool_pad_h", 0);
addLayer(layerParams, op);
}
void TFLiteImporter::parseReshape(const Operator& op, const std::string& opcode, LayerParams& layerParams) {
DataLayout inpLayout = layouts[op.inputs()->Get(0)];
layerParams.type = "Reshape";
std::vector<int> shape;
if (op.inputs()->size() > 1) {
shape = allTensors[op.inputs()->Get(1)];
} else {
auto options = op.builtin_options_as_ReshapeOptions();
CV_Assert(options);
shape.assign(options->new_shape()->begin(), options->new_shape()->end());
}
if (inpLayout == DNN_LAYOUT_NHWC) {
if (shape.size() == 4) {
// Keep data but change a shape to OpenCV's NCHW order
std::swap(shape[2], shape[3]);
std::swap(shape[1], shape[2]);
} else {
// Permute to NCHW entire data and reshape to given a shape
std::vector<int> order = {0, 2, 3, 1};
const std::string name = layerParams.name + "/permute";
auto inpId = layerIds[op.inputs()->Get(0)];
int permId = addPermuteLayer(order, name, inpId, isInt8(op) ? CV_8S : CV_32F); // NCHW -> NHWC
layerIds[op.inputs()->Get(0)] = std::make_pair(permId, 0);
layouts[op.outputs()->Get(0)] = DNN_LAYOUT_NCHW;
}
}
layerParams.set("dim", DictValue::arrayInt<int*>(shape.data(), shape.size()));
addLayer(layerParams, op);
}
void TFLiteImporter::parseConcat(const Operator& op, const std::string& opcode, LayerParams& layerParams) {
layerParams.type = "Concat";
auto options = reinterpret_cast<const ConcatenationOptions*>(op.builtin_options());
int axis = options->axis();
DataLayout inpLayout = layouts[op.inputs()->Get(0)];
if (inpLayout == DNN_LAYOUT_NHWC) {
// OpenCV works in NCHW data layout. So change the axis correspondingly.
axis = normalize_axis(axis, 4);
static const int remap[] = {0, 2, 3, 1};
axis = remap[axis];
}
layerParams.set("axis", axis);
addLayer(layerParams, op);
parseFusedActivation(op, options->fused_activation_function());
}
void TFLiteImporter::parsePack(const Operator& op, const std::string& opcode, LayerParams& layerParams) {
auto options = reinterpret_cast<const PackOptions*>(op.builtin_options());
int axis = options->axis();
DataLayout inpLayout = layouts[op.inputs()->Get(0)];
if (inpLayout == DNN_LAYOUT_NHWC) {
// OpenCV works in NCHW data layout. So change the axis correspondingly.
axis = normalize_axis(axis, 5); // 5 because Pack adds a new axis so -1 would mean 4
static const int remap[] = {0, 1, 3, 4, 2};
axis = remap[axis];
}
// Replace Pack layer to Reshape + Concat
// Use a set because there are models which replicate single layer data by Pack.
std::set<int> op_inputs(op.inputs()->begin(), op.inputs()->end());
std::map<int, std::pair<int, int> > originLayerIds;
for (int inp : op_inputs) {
auto inpId = layerIds[inp];
int dims = modelTensors->Get(inp)->shape()->size();
std::vector<int> shape{1, -1};
if (axis == dims) {
std::swap(shape[0], shape[1]);
}
const auto name = modelTensors->Get(inp)->name()->str() + "/reshape";
int reshapeId = addReshapeLayer(shape, axis == dims ? dims - 1 : axis, 1,
name, inpId, isInt8(op) ? CV_8S : CV_32F);
originLayerIds[inp] = layerIds[inp];
layerIds[inp] = std::make_pair(reshapeId, 0);
}
layerParams.type = "Concat";
layerParams.set("axis", axis);
addLayer(layerParams, op);
// Restore origin layer inputs
for (const auto& ids : originLayerIds) {
layerIds[ids.first] = ids.second;
}
}
void TFLiteImporter::parseResize(const Operator& op, const std::string& opcode, LayerParams& layerParams) {
layerParams.type = "Resize";
if (opcode == "RESIZE_BILINEAR") {
auto options = op.builtin_options_as_ResizeBilinearOptions();
layerParams.set("interpolation", "bilinear");
layerParams.set("align_corners", options->align_corners());
layerParams.set("half_pixel_centers", options->half_pixel_centers());
} else if (opcode == "RESIZE_NEAREST_NEIGHBOR") {
auto options = op.builtin_options_as_ResizeNearestNeighborOptions();
layerParams.set("interpolation", "nearest");
layerParams.set("align_corners", options->align_corners());
layerParams.set("half_pixel_centers", options->half_pixel_centers());
}
Mat shape = allTensors[op.inputs()->Get(1)].reshape(1, 1);
layerParams.set("height", shape.at<int>(0, 0));
layerParams.set("width", shape.at<int>(0, 1));
addLayer(layerParams, op);
}
void TFLiteImporter::parseTranspose(const Operator& op, const std::string& opcode, LayerParams& layerParams)
{
layerParams.type = "Permute";
std::vector<int> perm = allTensors[op.inputs()->Get(1)];
DataLayout inpLayout = layouts[op.inputs()->Get(0)];
if (inpLayout == DNN_LAYOUT_NHWC && perm.size() == 4) {
// OpenCV operates under the assumption that NCHW format, whereas TFLite defaults to NHWC.
// Therfore, to align these layouts, the axes of the permutation vector should be adjusted accordingly.
// For implementation details, please refer to the disscusion:
// https://github.com/opencv/opencv/pull/25297#issuecomment-2049762298
if (perm[0] != 0) {
CV_Error(Error::StsParseError, "The first axis should not be permuted.");
}
if (perm[1] == 1 && perm[2] == 2 && perm[3] == 3) {
std::vector<int> orderLP = {0, 1, 2, 3};
layerParams.set("order", DictValue::arrayInt<int*>(orderLP.data(), orderLP.size()));
layouts[op.outputs()->Get(0)] = DNN_LAYOUT_NCHW;
}
else if (perm[1] == 1 && perm[2] == 3 && perm[3] == 2) {
std::vector<int> orderLP = {0, 3, 2, 1};
layerParams.set("order", DictValue::arrayInt<int*>(orderLP.data(), orderLP.size()));
}
else if (perm[1] == 2 && perm[2] == 1 && perm[3] == 3) {
std::vector<int> orderLP = {0, 1, 3, 2};
layerParams.set("order", DictValue::arrayInt<int*>(orderLP.data(), orderLP.size()));
layouts[op.outputs()->Get(0)] = DNN_LAYOUT_NCHW;
}
else if (perm[1] == 2 && perm[2] == 3 && perm[3] == 1) {
std::vector<int> orderLP = {0, 2, 3, 1};
layerParams.set("order", DictValue::arrayInt<int*>(orderLP.data(), orderLP.size()));
}
}
else {
layerParams.set("order", DictValue::arrayInt<int*>(perm.data(), perm.size()));
}
addLayer(layerParams, op);
}
void TFLiteImporter::parseGlobalPooling(const Operator& op, const std::string& opcode, LayerParams& layerParams)
{
layerParams.type = "Pooling";
if(opcode == "MEAN") {
layerParams.set("pool", "ave");
}
else if (opcode == "REDUCE_MAX") {
layerParams.set("pool", "max");
}
else {
CV_Error(Error::StsNotImplemented, "Unsupported pooling " + opcode);
}
layerParams.set("global_pooling", true);
auto options = op.builtin_options_as_ReducerOptions();
bool keep_dims = options->keep_dims();
if (!keep_dims) {
const auto name = layerParams.name;
layerParams.name += "/global_pooling";
addLayer(layerParams, op);
int out = op.outputs()->Get(0);
auto outId = layerIds[out];
int flattenId = addFlattenLayer(1, -1, name, outId, isInt8(op) ? CV_8S : CV_32F);
layerIds[out] = std::make_pair(flattenId, 0);
}
else {
addLayer(layerParams, op);
}
}
int TFLiteImporter::addPermuteLayer(const std::vector<int>& order, const std::string& permName,
const std::pair<int, int>& inpId, int dtype)
{
LayerParams permLP;
permLP.set("order", DictValue::arrayInt<const int*>(order.data(), order.size()));
int permId = dstNet.addLayer(permName, "Permute", dtype, permLP);
dstNet.connect(inpId.first, inpId.second, permId, 0);
return permId;
}
int TFLiteImporter::addReshapeLayer(const std::vector<int>& shape, int axis, int num_axes,
const std::string& name, const std::pair<int, int>& inpId, int dtype)
{
LayerParams lp;
lp.set("axis", axis);
lp.set("dim", DictValue::arrayInt<const int*>(shape.data(), shape.size()));
lp.set("num_axes", num_axes);
int id = dstNet.addLayer(name, "Reshape", dtype, lp);
dstNet.connect(inpId.first, inpId.second, id, 0);
return id;
}
int TFLiteImporter::addFlattenLayer(int axis, int end_axis, const std::string& name, const std::pair<int, int>& inpId, int dtype)
{
LayerParams lp;
lp.set("axis", axis);
lp.set("end_axis", end_axis);
int id = dstNet.addLayer(name, "Flatten", dtype, lp);
dstNet.connect(inpId.first, inpId.second, id, 0);
return id;
}
void TFLiteImporter::parseDeconvolution(const Operator& op, const std::string& opcode, LayerParams& layerParams) {
layerParams.type = "Deconvolution";
CV_CheckLE(op.custom_options()->size(), sizeof(TfLiteTransposeConvParams), "");
const auto* params = reinterpret_cast<const TfLiteTransposeConvParams*>(op.custom_options()->Data());
if (params->padding != kTfLitePaddingUnknown)
layerParams.set("pad_mode", params->padding == kTfLitePaddingSame ? "SAME" : "VALID");
layerParams.set("stride_w", params->stride_width);
layerParams.set("stride_h", params->stride_height);
// Get filter size
int filterIdx = op.inputs()->Get(1);
Mat filter = allTensors[filterIdx];
int oc = filter.size[0];
int kh = filter.size[1];
int kw = filter.size[2];
int ic = filter.size[3];
layerParams.set("kernel_w", kw);
layerParams.set("kernel_h", kh);
layerParams.set("num_output", oc);
// Add adjust padding similar to TensorFlow (see tf_importer)
const auto* outShape = modelTensors->Get(op.outputs()->Get(0))->shape();
const int outH = outShape->Get(1);
const int outW = outShape->Get(2);
if (params->padding == kTfLitePaddingSame)
{
layerParams.set("adj_w", (outW - 1) % params->stride_width);
layerParams.set("adj_h", (outH - 1) % params->stride_height);
}
else if (params->padding == kTfLitePaddingValid)
{
layerParams.set("adj_w", (outW - kw) % params->stride_width);
layerParams.set("adj_h", (outH - kh) % params->stride_height);
}
// Reorder filter data from OHWI to IOHW and change shape correspondingly.
filter = allTensors[filterIdx] = filter.reshape(1, {ic, oc, kh, kw});
CV_CheckTypeEQ(filter.type(), CV_32F, "");
Mat filterCopy = filter.clone();
float* data = filterCopy.ptr<float>();
float* dstData = filter.ptr<float>();
int total = oc * ic * kh * kw;
for (int i_oc = 0; i_oc < oc; i_oc++) {
for (int i_ic = 0; i_ic < ic; i_ic++) {
for (int i_h = 0; i_h < kh; i_h++) {
for (int i_w = 0; i_w < kw; i_w++) {
int dst_i = kw * (kh * (oc * i_ic + i_oc) + i_h) + i_w;
int src_i = ic * (kw * (kh * i_oc + i_h) + i_w) + i_ic;
CV_CheckLT(dst_i, total, "");
CV_CheckLT(src_i, total, "");
dstData[dst_i] = data[src_i];
}
}
}
}
addLayer(layerParams, op);
}
void TFLiteImporter::parseQuantize(const Operator& op, const std::string& opcode, LayerParams& layerParams) {
layerParams.type = "Quantize";
layerParams.set("scales", 1);
layerParams.set("zeropoints", -128);
addLayer(layerParams, op);
}
void TFLiteImporter::parseDequantize(const Operator& op, const std::string& opcode, LayerParams& layerParams) {
layerParams.type = "Dequantize";
float inpScale, outScale;
int inpZero, outZero;
getQuantParams(op, inpScale, inpZero, outScale, outZero);
layerParams.set("scales", inpScale);
layerParams.set("zeropoints", inpZero);
addLayer(layerParams, op);
}
void TFLiteImporter::parseSplit(const Operator& op, const std::string& opcode, LayerParams& layerParams) {
layerParams.type = "Slice";
auto options = op.builtin_options_as_SplitOptions();
CV_Assert(options);
layerParams.set("num_split", options->num_splits());
addLayer(layerParams, op);
}
void TFLiteImporter::parseFullyConnected(const Operator& op, const std::string& opcode, LayerParams& layerParams) {
layerParams.type = "Gemm";
auto options = op.builtin_options_as_FullyConnectedOptions();
CV_Assert(options);
layerParams.set("transB", true);
layerParams.set("constB", true);
addLayer(layerParams, op);
parseFusedActivation(op, options->fused_activation_function());
}
void TFLiteImporter::parseSoftmax(const Operator& op, const std::string& opcode, LayerParams& layerParams) {
layerParams.type = "Softmax";
addLayer(layerParams, op);
}
void TFLiteImporter::parseCast(const Operator& op, const std::string& opcode, LayerParams& layerParams) {
layerParams.type = "Identity";
addLayer(layerParams, op);
}
void TFLiteImporter::parseDetectionPostProcess(const Operator& op, const std::string& opcode, LayerParams& layerParams) {
// Parse parameters;
std::vector<std::string> keys(1, "");
const uint8_t* data = op.custom_options()->Data();
int offset = 0;
// Read zero delimited keys
while (data[offset] != 10 && offset < op.custom_options()->size()) {
if (data[offset]) {
keys.back() += data[offset];
} else {
keys.emplace_back("");
}
offset += 1;
}
keys.pop_back();
std::sort(keys.begin(), keys.end());
// TODO: Replace empirical offset to something more reliable.
offset += 25;
std::map<std::string, uint32_t> parameters;
for (int i = 0; i < keys.size(); ++i) {
parameters[keys[i]] = *reinterpret_cast<const uint32_t*>(data + offset + i * 4);
}
parameters["num_classes"] = modelTensors->Get(op.inputs()->Get(1))->shape()->Get(2);
layerParams.type = "DetectionOutput";
layerParams.set("num_classes", parameters["num_classes"]);
layerParams.set("share_location", true);
layerParams.set("background_label_id", parameters["num_classes"] + 1);
layerParams.set("nms_threshold", *(float*)&parameters["nms_iou_threshold"]);
layerParams.set("confidence_threshold", *(float*)&parameters["nms_score_threshold"]);
layerParams.set("top_k", parameters["max_detections"]);
layerParams.set("keep_top_k", parameters["max_detections"]);
layerParams.set("code_type", "CENTER_SIZE");
layerParams.set("loc_pred_transposed", true);
// Replace third input from tensor to Const layer with the priors
Mat priors = allTensors[op.inputs()->Get(2)].clone();
// Change priors data from (ycenter, xcenter, h, w) to (xmin, ymin, xmax, ymax)
priors = priors.reshape(1, priors.total() / 4);
Mat tmp = priors.col(0).clone();
priors.col(0) = priors.col(1) - 0.5 * priors.col(3);
priors.col(1) = tmp - 0.5 * priors.col(2);
tmp = priors.col(2).clone();
priors.col(2) = priors.col(0) + priors.col(3);
priors.col(3) = priors.col(1) + tmp;
float x_scale = *(float*)&parameters["x_scale"];
float y_scale = *(float*)&parameters["y_scale"];
float w_scale = *(float*)&parameters["w_scale"];
float h_scale = *(float*)&parameters["h_scale"];
if (x_scale != 1.0f || y_scale != 1.0f || w_scale != 1.0f || h_scale != 1.0f) {
int numPriors = priors.rows;
priors.resize(numPriors * 2);
Mat_<float> scales({1, 4}, {1.f / x_scale, 1.f / y_scale,
1.f / w_scale, 1.f / h_scale});
repeat(scales, numPriors, 1, priors.rowRange(numPriors, priors.rows));
priors = priors.reshape(1, {1, 2, (int)priors.total() / 2});
layerParams.set("variance_encoded_in_target", false);
} else {
priors = priors.reshape(1, {1, 1, (int)priors.total()});
layerParams.set("variance_encoded_in_target", true);
}
LayerParams priorsLP;
priorsLP.name = layerParams.name + "/priors";
priorsLP.type = "Const";
priorsLP.blobs.resize(1, priors);
int priorsId = dstNet.addLayer(priorsLP.name, priorsLP.type, priorsLP);
layerIds[op.inputs()->Get(2)] = std::make_pair(priorsId, 0);
addLayer(layerParams, op);
}
void TFLiteImporter::parseFusedActivation(const Operator& op, ActivationFunctionType activ) {
LayerParams activParams;
activParams.name = modelTensors->Get(op.outputs()->Get(0))->name()->str() + "/activ";
parseActivation(op, EnumNameActivationFunctionType(activ), activParams, true);
}
void TFLiteImporter::parseActivation(const Operator& op, const std::string& opcode, LayerParams& activParams) {
parseActivation(op, opcode, activParams, false);
}
void TFLiteImporter::parseActivation(const Operator& op, const std::string& opcode, LayerParams& activParams, bool isFused) {
if (opcode == "NONE")
return;
else if (opcode == "RELU6")
activParams.type = "ReLU6";
else if (opcode == "PRELU")
activParams.type = "PReLU";
else if (opcode == "RELU")
activParams.type = "ReLU";
else if (opcode == "HARD_SWISH")
activParams.type = "HardSwish";
else if (opcode == "LOGISTIC")
activParams.type = "Sigmoid";
else
CV_Error(Error::StsNotImplemented, "Unsupported activation " + opcode);
if (isInt8(op)) {
float inpScale, outScale;
int inpZero, outZero;
getQuantParams(op, inpScale, inpZero, outScale, outZero);
if (isFused) {
activParams.type += "Int8";
activParams.set("input_scale", outScale);
activParams.set("input_zeropoint", outZero);
activParams.set("scales", outScale);
activParams.set("zeropoints", outZero);
}
Mat lookUpTable(1, 256, CV_8S);
int8_t* table = lookUpTable.ptr<int8_t>();
for (int i = -128; i < 128; i++) {
float x, y = i;
if (isFused)
x = outScale * (i - outZero);
else
x = inpScale * (i - inpZero);
if (opcode == "RELU6")
y = std::min(std::max(x, 0.f), 6.f);
else if (opcode == "LOGISTIC")
y = 1.0f / (1.0f + std::exp(-x));
else if (opcode == "HARD_SWISH")
y = x * max(0.f, min(1.f, x / 6.f + 0.5f));
else
CV_Error(Error::StsNotImplemented, "Lookup table for " + opcode);
int quantized = outZero + cvRound(y / outScale);
table[i + 128] = saturate_cast<int8_t>(quantized);
}
activParams.blobs.resize(1, lookUpTable);
}
if (isFused) {
int dtype = isInt8(op) ? CV_8S : CV_32F;
int layerId = dstNet.addLayerToPrev(activParams.name, activParams.type, dtype, activParams);
// Override layer ids mapping
int i = 0;
for (int idx : *op.outputs()) {
layerIds[idx] = std::make_pair(layerId, i++);
}
} else {
addLayer(activParams, op);
}
}
bool TFLiteImporter::isInt8(const Operator& op) {
const Tensor* out = modelTensors->Get(op.outputs()->Get(0));
return out->type() == TensorType_INT8;
}
void TFLiteImporter::getQuantParams(const Operator& op, float& inpScale, int& inpZero, float& outScale, int& outZero) {
const Tensor* inp = modelTensors->Get(op.inputs()->Get(0));
const Tensor* out = modelTensors->Get(op.outputs()->Get(0));
inpScale = outScale = inpZero = outZero = 0;
if (inp->quantization()) {
if (inp->quantization()->scale()) {
CV_Assert(inp->quantization()->scale()->size() == 1);
inpScale = inp->quantization()->scale()->Get(0);
}
if (inp->quantization()->zero_point()) {
CV_Assert(inp->quantization()->zero_point()->size() == 1);
inpZero = inp->quantization()->zero_point()->Get(0);
}
}
if (out->quantization()) {
if (out->quantization()->scale()) {
CV_Assert(out->quantization()->scale()->size() == 1);
outScale = out->quantization()->scale()->Get(0);
}
if (out->quantization()->zero_point()) {
CV_Assert(out->quantization()->zero_point()->size() == 1);
outZero = out->quantization()->zero_point()->Get(0);
}
}
}
Net readNetFromTFLite(const String &modelPath) {
Net net;
std::vector<char> content;
const std::ios::openmode mode = std::ios::in | std::ios::binary;
std::ifstream ifs(modelPath, mode);
if (!ifs.is_open())
CV_Error(Error::StsError, cv::format("DNN/TFLite: can't open model file '%s'", modelPath.c_str()));
ifs.seekg(0, std::ios::end);
const size_t sz = ifs.tellg();
CV_Assert(sz > 0);
content.resize(sz);
ifs.seekg(0, std::ios::beg);
ifs.read(content.data(), sz);
CV_Assert(!ifs.bad());
TFLiteImporter(net, content.data(), content.size());
return net;
}
Net readNetFromTFLite(const std::vector<uchar>& bufferModel) {
return readNetFromTFLite((const char*)bufferModel.data(), bufferModel.size());
}
Net readNetFromTFLite(const char *bufferModel, size_t bufSize) {
Net net;
TFLiteImporter(net, bufferModel, bufSize);
return net;
}
#else // HAVE_FLATBUFFERS
#define DNN_TFLITE_UNSUPPORTED() CV_Error(Error::StsError, "DNN/TFLite: Build OpenCV with FlatBuffers to import TFLite models: https://github.com/opencv/opencv/pull/23161")
Net readNetFromTFLite(const String &) {
DNN_TFLITE_UNSUPPORTED();
}
Net readNetFromTFLite(const std::vector<uchar>&) {
DNN_TFLITE_UNSUPPORTED();
}
Net readNetFromTFLite(const char *, size_t) {
DNN_TFLITE_UNSUPPORTED();
}
#endif // HAVE_FLATBUFFERS
CV__DNN_INLINE_NS_END
}} // namespace cv::dnn