// This file is part of OpenCV project. // It is subject to the license terms in the LICENSE file found in the top-level directory // of this distribution and at http://opencv.org/license.html. #include "../precomp.hpp" #ifdef HAVE_FLATBUFFERS #include "schema_generated.h" #include "builtin_op_data.h" #endif #include #undef CV_LOG_STRIP_LEVEL #define CV_LOG_STRIP_LEVEL CV_LOG_LEVEL_VERBOSE + 1 #include namespace cv { namespace dnn { CV__DNN_INLINE_NS_BEGIN #ifdef HAVE_FLATBUFFERS using namespace opencv_tflite; class TFLiteImporter { public: TFLiteImporter(Net& net, const char* modelBuffer, size_t bufSize); private: const opencv_tflite::Model* model; const flatbuffers::Vector >* modelTensors; std::map allTensors; Net& dstNet; // This is a vector of pairs (layerId, outputId) where we iterate over // indices from TFLite notation and get created OpenCV layers. std::map > layerIds; // Tracking of layouts for layers outputs. std::vector layouts; void populateNet(); // Wrap TFLite Tensor to OpenCV Mat without data copying Mat parseTensor(const Tensor& tensor); typedef void (TFLiteImporter::*TFLiteImporterNodeParser)(const Operator&, const std::string&, LayerParams&); typedef std::map DispatchMap; const DispatchMap dispatch; static DispatchMap buildDispatchMap(); void parseConvolution(const Operator& op, const std::string& opcode, LayerParams& layerParams); void parseDWConvolution(const Operator& op, const std::string& opcode, LayerParams& layerParams); void parsePadding(const Operator& op, const std::string& opcode, LayerParams& layerParams); void parseEltwise(const Operator& op, const std::string& opcode, LayerParams& layerParams); void parsePooling(const Operator& op, const std::string& opcode, LayerParams& layerParams); void parsePoolingWithArgmax(const Operator& op, const std::string& opcode, LayerParams& layerParams); void parseUnpooling(const Operator& op, const std::string& opcode, LayerParams& layerParams); void parseReshape(const Operator& op, const std::string& opcode, LayerParams& layerParams); void parseConcat(const Operator& op, const std::string& opcode, LayerParams& layerParams); void parsePack(const Operator& op, const std::string& opcode, LayerParams& layerParams); void parseResize(const Operator& op, const std::string& opcode, LayerParams& layerParams); void parseDeconvolution(const Operator& op, const std::string& opcode, LayerParams& layerParams); void parseQuantize(const Operator& op, const std::string& opcode, LayerParams& layerParams); void parseDequantize(const Operator& op, const std::string& opcode, LayerParams& layerParams); void parseDetectionPostProcess(const Operator& op, const std::string& opcode, LayerParams& layerParams); void parseActivation(const Operator& op, const std::string& opcode, LayerParams& layerParams); void parseSplit(const Operator& op, const std::string& opcode, LayerParams& layerParams); void parseFullyConnected(const Operator& op, const std::string& opcode, LayerParams& layerParams); void parseSoftmax(const Operator& op, const std::string& opcode, LayerParams& layerParams); void parseCast(const Operator& op, const std::string& opcode, LayerParams& layerParams); void parseTranspose(const Operator& op, const std::string& opcode, LayerParams& layerParams); void parseGlobalPooling(const Operator& op, const std::string& opcode, LayerParams& layerParams); void parseFusedActivation(const Operator& op, ActivationFunctionType activ); void parseActivation(const Operator& op, const std::string& opcode, LayerParams& layerParams, bool isFused); void addLayer(LayerParams& layerParams, const Operator& op); int addPermuteLayer(const std::vector& order, const std::string& permName, const std::pair& inpId, int dtype); int addReshapeLayer(const std::vector& shape, int axis, int num_axes, const std::string& name, const std::pair& inpId, int dtype); int addFlattenLayer(int axis, int end_axis, const std::string& name, const std::pair& inpId, int dtype); inline bool isInt8(const Operator& op); inline void getQuantParams(const Operator& op, float& inpScale, int& inpZero, float& outScale, int& outZero); }; Mat TFLiteImporter::parseTensor(const Tensor& tensor) { const auto tensor_shape = tensor.shape(); CV_Assert(tensor_shape); std::vector shape(tensor_shape->begin(), tensor_shape->end()); int bufferIdx = tensor.buffer(); CV_Assert(bufferIdx != 0); // 0th buffer is a no-data buffer const Buffer* buffer = model->buffers()->Get(bufferIdx); CV_Assert(buffer); const auto buffer_data = buffer->data(); if (!buffer_data) return Mat(); const void* data = buffer_data->data(); int dtype = -1; switch (tensor.type()) { case TensorType_FLOAT32: dtype = CV_32F; break; case TensorType_INT32: dtype = CV_32S; break; case TensorType_FLOAT16: dtype = CV_16F; break; case TensorType_INT8: dtype = CV_8S; break; default: CV_Error(Error::StsNotImplemented, format("Parse tensor with type %s", EnumNameTensorType(tensor.type()))); } return shape.empty() ? Mat() : Mat(shape, dtype, const_cast(data)); } TFLiteImporter::TFLiteImporter(Net& dstNet, const char* modelBuffer, size_t bufSize) : dstNet(dstNet), dispatch(buildDispatchMap()) { flatbuffers::Verifier verifier((const uint8_t*)modelBuffer, bufSize); if (!VerifyModelBuffer(verifier)) { CV_Error(Error::StsError, "DNN/TFLite: model is incorrect"); } model = GetModel(modelBuffer); CV_Assert(model); CV_Assert(model->subgraphs()); CV_Assert(model->buffers()); CV_CheckEQ((size_t)model->subgraphs()->size(), 1u, ""); modelTensors = model->subgraphs()->Get(0)->tensors(); CV_Assert(modelTensors); for (int i = 0; i < modelTensors->size(); ++i) { const Tensor* tensor = modelTensors->Get(i); CV_Assert(tensor); if (tensor->buffer() != 0) { allTensors[i] = parseTensor(*tensor); } } populateNet(); } DataLayout estimateLayout(const Tensor& t) { const auto t_shape = t.shape(); CV_Assert(t_shape); switch (t_shape->size()) { case 5: return DNN_LAYOUT_NDHWC; case 4: return DNN_LAYOUT_NHWC; case 2: return DNN_LAYOUT_PLANAR; default: return DNN_LAYOUT_UNKNOWN; } } void TFLiteImporter::populateNet() { CV_Assert(model); const auto model_subgraphs = model->subgraphs(); CV_Assert(model_subgraphs); const SubGraph* subgraph = model_subgraphs->Get(0); CV_Assert(subgraph); const auto subgraph_inputs = subgraph->inputs(); CV_Assert(subgraph_inputs); const auto subgraph_operators = subgraph->operators(); CV_Assert(subgraph_operators); const auto opCodes = model->operator_codes(); CV_Assert(opCodes); CV_Assert(modelTensors); layouts.resize(modelTensors->size(), DNN_LAYOUT_UNKNOWN); size_t subgraph_inputs_size = subgraph_inputs->size(); std::vector inputsNames(subgraph_inputs_size); std::vector inputsShapes(subgraph_inputs_size); for (size_t i = 0; i < subgraph_inputs_size; ++i) { int idx = subgraph_inputs->Get(i); layerIds[idx] = std::make_pair(0, i); const auto tensor = modelTensors->Get(idx); if (!tensor) CV_Error(Error::StsError, cv::format("DNN/TFLite: subgraph input %d (%d) is NULL", (int)i, idx)); layouts[idx] = estimateLayout(*tensor); // Keep info about origin inputs names and shapes inputsNames[i] = tensor->name()->str(); std::vector shape(tensor->shape()->begin(), tensor->shape()->end()); if (layouts[idx] == DNN_LAYOUT_NHWC) { CV_CheckEQ(shape.size(), (size_t)4, ""); std::swap(shape[2], shape[3]); std::swap(shape[1], shape[2]); } inputsShapes[i] = shape; } dstNet.setInputsNames(inputsNames); for (size_t i = 0; i < subgraph_inputs_size; ++i) { dstNet.setInputShape(inputsNames[i], inputsShapes[i]); } const auto& all_operators = *subgraph_operators; const size_t all_operators_size = all_operators.size(); for (size_t op_idx = 0; op_idx < all_operators_size; ++op_idx) { const auto op = all_operators[op_idx]; CV_Assert(op); const auto op_inputs = op->inputs(); CV_Assert(op_inputs); const auto op_outputs = op->outputs(); CV_Assert(op_outputs); int idx = op->opcode_index(); LayerParams layerParams; layerParams.name = modelTensors->Get(op_outputs->Get(0))->name()->str(); std::string type = EnumNameBuiltinOperator(BuiltinOperator(opCodes->Get(idx)->deprecated_builtin_code())); if (type == "CUSTOM") { type = opCodes->Get(idx)->custom_code()->str(); } CV_LOG_DEBUG(NULL, "DNN/TFLite: processing operator (" << op_idx << "/" << all_operators_size << ") with " << op_inputs->size() << " inputs: " << cv::format("[%s]:(%s)", type.c_str(), layerParams.name.c_str())); try { if (type == "DEQUANTIZE") { // Convert from FP16 to FP32 Mat data = allTensors[op_inputs->Get(0)]; if (!data.empty()) { // Dequantize a buffer Mat dataFP32; data.convertTo(dataFP32, CV_32F); allTensors[op_outputs->Get(0)] = dataFP32; continue; } } DispatchMap::const_iterator iter = dispatch.find(type); if (iter == dispatch.end()) CV_Error(Error::StsNotImplemented, "Unsupported operator type " + type); CALL_MEMBER_FN(*this, iter->second)(*op, type, layerParams); } catch (const cv::Exception& e) { CV_LOG_ERROR(NULL, "DNN/TFLite: Problem during import of operator " << cv::format("[%s]:(%s)", type.c_str(), layerParams.name.c_str()) << " (" << op_idx << "/" << all_operators_size << "). Exception: " << e.what()); if (DNN_DIAGNOSTICS_RUN) { continue; } throw; } } } TFLiteImporter::DispatchMap TFLiteImporter::buildDispatchMap() { static DispatchMap dispatch; if (!dispatch.empty()) return dispatch; dispatch["CONV_2D"] = &TFLiteImporter::parseConvolution; dispatch["DEPTHWISE_CONV_2D"] = &TFLiteImporter::parseDWConvolution; dispatch["ADD"] = dispatch["MUL"] = &TFLiteImporter::parseEltwise; dispatch["RELU"] = dispatch["PRELU"] = dispatch["HARD_SWISH"] = dispatch["LOGISTIC"] = &TFLiteImporter::parseActivation; dispatch["MAX_POOL_2D"] = dispatch["AVERAGE_POOL_2D"] = &TFLiteImporter::parsePooling; dispatch["MaxPoolingWithArgmax2D"] = &TFLiteImporter::parsePoolingWithArgmax; dispatch["MaxUnpooling2D"] = &TFLiteImporter::parseUnpooling; dispatch["PAD"] = &TFLiteImporter::parsePadding; dispatch["RESHAPE"] = &TFLiteImporter::parseReshape; dispatch["CONCATENATION"] = &TFLiteImporter::parseConcat; dispatch["PACK"] = &TFLiteImporter::parsePack; dispatch["RESIZE_BILINEAR"] = dispatch["RESIZE_NEAREST_NEIGHBOR"] = &TFLiteImporter::parseResize; dispatch["Convolution2DTransposeBias"] = &TFLiteImporter::parseDeconvolution; dispatch["QUANTIZE"] = &TFLiteImporter::parseQuantize; dispatch["DEQUANTIZE"] = &TFLiteImporter::parseDequantize; dispatch["SPLIT"] = &TFLiteImporter::parseSplit; dispatch["FULLY_CONNECTED"] = &TFLiteImporter::parseFullyConnected; dispatch["SOFTMAX"] = &TFLiteImporter::parseSoftmax; dispatch["CAST"] = &TFLiteImporter::parseCast; dispatch["TFLite_Detection_PostProcess"] = &TFLiteImporter::parseDetectionPostProcess; dispatch["TRANSPOSE"] = &TFLiteImporter::parseTranspose; dispatch["MEAN"] = dispatch["REDUCE_MAX"] = &TFLiteImporter::parseGlobalPooling; return dispatch; } void TFLiteImporter::addLayer(LayerParams& layerParams, const Operator& op) { const auto op_inputs = op.inputs(); const auto op_outputs = op.outputs(); // Collect input blobs if (layerParams.blobs.empty()) { for (int idx : *op_inputs) { if (layerIds.find(idx) != layerIds.end()) { continue; // Output from a different layer } Mat blob = allTensors[idx]; layerParams.blobs.push_back(blob.u ? blob : blob.clone()); // some tensors are owned by OpenCV } } else { for (auto& blob : layerParams.blobs) { CV_Assert(blob.u); } } int dtype = CV_32F; if (isInt8(op)) { dtype = CV_8S; if (layerParams.type != "Quantize") layerParams.type += "Int8"; if (!layerParams.has("zeropoints")) { float inpScale, outScale; int inpZero, outZero; getQuantParams(op, inpScale, inpZero, outScale, outZero); layerParams.set("input_scale", inpScale); layerParams.set("input_zeropoint", inpZero); layerParams.set("scales", outScale); layerParams.set("zeropoints", outZero); } } int layerId = dstNet.addLayer(layerParams.name, layerParams.type, dtype, layerParams); // Connect layer to inputs int i = 0; std::vector inpLayouts; for (int idx : *op_inputs) { if (layerIds.find(idx) == layerIds.end()) { continue; // Const input } inpLayouts.push_back(layouts[idx]); auto it = layerIds.find(idx); CV_Assert(it != layerIds.end()); dstNet.connect(it->second.first, it->second.second, layerId, i++); } // Predict output layout. Some layer-specific parsers may set them explicitly. // Otherwise, propagate input layout. if (layouts[op_outputs->Get(0)] == DNN_LAYOUT_UNKNOWN) { DataLayout predictedLayout = DNN_LAYOUT_UNKNOWN; for (auto layout : inpLayouts) { if (layout != DNN_LAYOUT_UNKNOWN) { if (predictedLayout == DNN_LAYOUT_UNKNOWN) predictedLayout = layout; else if (predictedLayout != layout) { predictedLayout = DNN_LAYOUT_UNKNOWN; break; } } } layouts[op_outputs->Get(0)] = predictedLayout; } // Register outputs i = 0; for (int idx : *op_outputs) { layerIds[idx] = std::make_pair(layerId, i++); } } void TFLiteImporter::parseConvolution(const Operator& op, const std::string& opcode, LayerParams& layerParams) { layerParams.type = "Convolution"; auto options = reinterpret_cast(op.builtin_options()); layerParams.set("pad_mode", EnumNamePadding(options->padding())); layerParams.set("stride_w", options->stride_w()); layerParams.set("stride_h", options->stride_h()); layerParams.set("dilation_w", options->dilation_w_factor()); layerParams.set("dilation_h", options->dilation_h_factor()); // Get filter size int filterIdx = op.inputs()->Get(1); Mat filter = allTensors[filterIdx]; int oc = filter.size[0]; int kh = filter.size[1]; int kw = filter.size[2]; layerParams.set("kernel_w", kw); layerParams.set("kernel_h", kh); layerParams.set("num_output", oc); bool isInt8 = filter.depth() == CV_8S; // Fill convolutions blobs here because of two reasons: // 1. Kernel transposition // 2. Extra blob with kernel scales in case of INT8 mode bool hasBias = op.inputs()->size() > 2; layerParams.blobs.resize(1 + (int)hasBias + (int)isInt8); if (hasBias) { Mat bias = allTensors[op.inputs()->Get(2)]; layerParams.blobs[1] = bias.u ? bias : bias.clone(); } // Reorder filter data from OHWI to OIHW and change shape correspondingly. transposeND(filter, {0, 3, 1, 2}, layerParams.blobs[0]); if (isInt8) { float inpScale, outScale; int inpZero, outZero; getQuantParams(op, inpScale, inpZero, outScale, outZero); layerParams.blobs[2] = Mat(oc, 1, CV_32F); auto filterScales = modelTensors->Get(filterIdx)->quantization()->scale(); if (filterScales->size() == 1) { layerParams.blobs[2].setTo(inpScale * filterScales->Get(0) / outScale); } else { for (size_t i = 0; i < filterScales->size(); ++i) { layerParams.blobs[2].at(i) = inpScale * filterScales->Get(i) / outScale; } } if (hasBias) { Mat bias = layerParams.blobs[1].reshape(1, oc); Mat weights_2d = layerParams.blobs[0].reshape(1, oc); for (int i = 0; i < oc; i++) { bias.at(i) -= inpZero * (cv::sum(weights_2d.row(i))[0]); } } } addLayer(layerParams, op); parseFusedActivation(op, options->fused_activation_function()); } void TFLiteImporter::parseDWConvolution(const Operator& op, const std::string& opcode, LayerParams& layerParams) { layerParams.type = "Convolution"; auto options = reinterpret_cast(op.builtin_options()); layerParams.set("pad_mode", EnumNamePadding(options->padding())); layerParams.set("stride_w", options->stride_w()); layerParams.set("stride_h", options->stride_h()); layerParams.set("dilation_w", options->dilation_w_factor()); layerParams.set("dilation_h", options->dilation_h_factor()); int filterIdx = op.inputs()->Get(1); Mat filter = allTensors[filterIdx]; int kh = filter.size[1]; int kw = filter.size[2]; int oc = filter.size[3]; layerParams.set("kernel_w", kw); layerParams.set("kernel_h", kh); layerParams.set("num_output", oc); layerParams.set("group", oc); bool isInt8 = filter.depth() == CV_8S; // Fill convolutions blobs here because of two reasons: // 1. Kernel transposition // 2. Extra blob with kernel scales in case of INT8 mode bool hasBias = op.inputs()->size() > 2; layerParams.blobs.resize(1 + (int)hasBias + (int)isInt8); if (hasBias) { Mat bias = allTensors[op.inputs()->Get(2)]; layerParams.blobs[1] = bias.u ? bias : bias.clone(); } transposeND(filter, {3, 0, 1, 2}, layerParams.blobs[0]); if (isInt8) { float inpScale, outScale; int inpZero, outZero; getQuantParams(op, inpScale, inpZero, outScale, outZero); layerParams.blobs[2] = Mat(oc, 1, CV_32F); auto filterScales = modelTensors->Get(filterIdx)->quantization()->scale(); if (filterScales->size() == 1) { layerParams.blobs[2].setTo(inpScale * filterScales->Get(0) / outScale); } else { for (size_t i = 0; i < filterScales->size(); ++i) { layerParams.blobs[2].at(i) = inpScale * filterScales->Get(i) / outScale; } } if (hasBias) { Mat bias = layerParams.blobs[1].reshape(1, oc); Mat weights_2d = layerParams.blobs[0].reshape(1, oc); for (int i = 0; i < oc; i++) { bias.at(i) -= inpZero * (cv::sum(weights_2d.row(i))[0]); } } } addLayer(layerParams, op); parseFusedActivation(op, options->fused_activation_function()); } void TFLiteImporter::parsePadding(const Operator& op, const std::string& opcode, LayerParams& layerParams) { layerParams.type = "Padding"; Mat paddings = allTensors[op.inputs()->Get(1)].clone(); CV_CheckTypeEQ(paddings.type(), CV_32S, ""); // N H W C // 0 1 2 3 4 5 6 7 std::swap(paddings.at(2), paddings.at(6)); std::swap(paddings.at(3), paddings.at(7)); // N C W H // 0 1 2 3 4 5 6 7 std::swap(paddings.at(4), paddings.at(6)); std::swap(paddings.at(5), paddings.at(7)); // N C H W // 0 1 2 3 4 5 6 7 layerParams.set("paddings", DictValue::arrayInt((int32_t*)paddings.data, paddings.total())); addLayer(layerParams, op); } void TFLiteImporter::parseEltwise(const Operator& op, const std::string& opcode, LayerParams& layerParams) { ActivationFunctionType activ = ActivationFunctionType_NONE; layerParams.type = "Eltwise"; if (opcode == "ADD") { auto options = reinterpret_cast(op.builtin_options()); activ = options->fused_activation_function(); layerParams.set("operation", "sum"); } else if (opcode == "MUL") { auto options = reinterpret_cast(op.builtin_options()); activ = options->fused_activation_function(); layerParams.set("operation", "prod"); } else { CV_Error(Error::StsNotImplemented, "Unknown opcode for Eltwise layer: " + opcode); } if (isInt8(op)) { const Tensor* out = modelTensors->Get(op.outputs()->Get(0)); float outScale = out->quantization()->scale()->Get(0); int outZero = out->quantization()->zero_point()->Get(0); const size_t numInps = op.inputs()->size(); std::vector inputScales(numInps); std::vector inputZeros(numInps); std::vector coeffs(numInps); float offset = outZero; for (int i = 0; i < numInps; ++i) { const Tensor* inp = modelTensors->Get(op.inputs()->Get(i)); float inpScale = inp->quantization()->scale()->Get(0); int inpZero = inp->quantization()->zero_point()->Get(0); inputScales[i] = inpScale; inputZeros[i] = inpZero; coeffs[i] = inpScale / outScale; offset -= coeffs[i] * inpZero; } layerParams.set("input_scales", DictValue::arrayReal(inputScales.data(), numInps)); layerParams.set("input_zeropoints", DictValue::arrayInt(inputZeros.data(), numInps)); layerParams.set("coeff", DictValue::arrayReal(coeffs.data(), numInps)); layerParams.set("offset", offset); layerParams.set("scales", outScale); layerParams.set("zeropoints", outZero); } addLayer(layerParams, op); parseFusedActivation(op, activ); } void TFLiteImporter::parsePooling(const Operator& op, const std::string& opcode, LayerParams& layerParams) { layerParams.type = "Pooling"; auto options = reinterpret_cast(op.builtin_options()); layerParams.set("pad_mode", EnumNamePadding(options->padding())); layerParams.set("stride_w", options->stride_w()); layerParams.set("stride_h", options->stride_h()); layerParams.set("kernel_w", options->filter_width()); layerParams.set("kernel_h", options->filter_height()); if (opcode == "MAX_POOL_2D") layerParams.set("pool", "max"); else if (opcode == "AVERAGE_POOL_2D") layerParams.set("pool", "ave"); else CV_Error(Error::StsNotImplemented, "Pool type selection for " + opcode); addLayer(layerParams, op); parseFusedActivation(op, options->fused_activation_function()); } void TFLiteImporter::parsePoolingWithArgmax(const Operator& op, const std::string& opcode, LayerParams& layerParams) { layerParams.type = "Pooling"; CV_CheckLE(op.custom_options()->size(), sizeof(TfLitePoolParams), ""); const auto* params = reinterpret_cast(op.custom_options()->Data()); if (params->activation != kTfLiteActNone) { CV_Error(Error::StsNotImplemented, "Argmax pooling with fused activation"); } if (params->padding != kTfLitePaddingUnknown) layerParams.set("pad_mode", params->padding == kTfLitePaddingSame ? "SAME" : "VALID"); layerParams.set("stride_w", params->stride_width); layerParams.set("stride_h", params->stride_height); layerParams.set("kernel_w", params->filter_width); layerParams.set("kernel_h", params->filter_height); layerParams.set("pool", "max"); addLayer(layerParams, op); } void TFLiteImporter::parseUnpooling(const Operator& op, const std::string& opcode, LayerParams& layerParams) { layerParams.type = "MaxUnpool"; CV_CheckLE(op.custom_options()->size(), sizeof(TfLitePoolParams), ""); const auto* params = reinterpret_cast(op.custom_options()->Data()); if (params->activation != kTfLiteActNone) { CV_Error(Error::StsNotImplemented, "Unpooling with fused activation"); } layerParams.set("pool_stride_w", params->stride_width); layerParams.set("pool_stride_h", params->stride_height); layerParams.set("pool_k_w", params->filter_width); layerParams.set("pool_k_h", params->filter_height); layerParams.set("pool_pad_w", 0); layerParams.set("pool_pad_h", 0); addLayer(layerParams, op); } void TFLiteImporter::parseReshape(const Operator& op, const std::string& opcode, LayerParams& layerParams) { DataLayout inpLayout = layouts[op.inputs()->Get(0)]; layerParams.type = "Reshape"; std::vector shape; if (op.inputs()->size() > 1) { shape = allTensors[op.inputs()->Get(1)]; } else { auto options = op.builtin_options_as_ReshapeOptions(); CV_Assert(options); shape.assign(options->new_shape()->begin(), options->new_shape()->end()); } if (inpLayout == DNN_LAYOUT_NHWC) { if (shape.size() == 4) { // Keep data but change a shape to OpenCV's NCHW order std::swap(shape[2], shape[3]); std::swap(shape[1], shape[2]); } else { // Permute to NCHW entire data and reshape to given a shape std::vector order = {0, 2, 3, 1}; const std::string name = layerParams.name + "/permute"; auto inpId = layerIds[op.inputs()->Get(0)]; int permId = addPermuteLayer(order, name, inpId, isInt8(op) ? CV_8S : CV_32F); // NCHW -> NHWC layerIds[op.inputs()->Get(0)] = std::make_pair(permId, 0); layouts[op.outputs()->Get(0)] = DNN_LAYOUT_NCHW; } } layerParams.set("dim", DictValue::arrayInt(shape.data(), shape.size())); addLayer(layerParams, op); } void TFLiteImporter::parseConcat(const Operator& op, const std::string& opcode, LayerParams& layerParams) { layerParams.type = "Concat"; auto options = reinterpret_cast(op.builtin_options()); int axis = options->axis(); DataLayout inpLayout = layouts[op.inputs()->Get(0)]; if (inpLayout == DNN_LAYOUT_NHWC) { // OpenCV works in NCHW data layout. So change the axis correspondingly. axis = normalize_axis(axis, 4); static const int remap[] = {0, 2, 3, 1}; axis = remap[axis]; } layerParams.set("axis", axis); addLayer(layerParams, op); parseFusedActivation(op, options->fused_activation_function()); } void TFLiteImporter::parsePack(const Operator& op, const std::string& opcode, LayerParams& layerParams) { auto options = reinterpret_cast(op.builtin_options()); int axis = options->axis(); DataLayout inpLayout = layouts[op.inputs()->Get(0)]; if (inpLayout == DNN_LAYOUT_NHWC) { // OpenCV works in NCHW data layout. So change the axis correspondingly. axis = normalize_axis(axis, 5); // 5 because Pack adds a new axis so -1 would mean 4 static const int remap[] = {0, 1, 3, 4, 2}; axis = remap[axis]; } // Replace Pack layer to Reshape + Concat // Use a set because there are models which replicate single layer data by Pack. std::set op_inputs(op.inputs()->begin(), op.inputs()->end()); std::map > originLayerIds; for (int inp : op_inputs) { auto inpId = layerIds[inp]; int dims = modelTensors->Get(inp)->shape()->size(); std::vector shape{1, -1}; if (axis == dims) { std::swap(shape[0], shape[1]); } const auto name = modelTensors->Get(inp)->name()->str() + "/reshape"; int reshapeId = addReshapeLayer(shape, axis == dims ? dims - 1 : axis, 1, name, inpId, isInt8(op) ? CV_8S : CV_32F); originLayerIds[inp] = layerIds[inp]; layerIds[inp] = std::make_pair(reshapeId, 0); } layerParams.type = "Concat"; layerParams.set("axis", axis); addLayer(layerParams, op); // Restore origin layer inputs for (const auto& ids : originLayerIds) { layerIds[ids.first] = ids.second; } } void TFLiteImporter::parseResize(const Operator& op, const std::string& opcode, LayerParams& layerParams) { layerParams.type = "Resize"; if (opcode == "RESIZE_BILINEAR") { auto options = op.builtin_options_as_ResizeBilinearOptions(); layerParams.set("interpolation", "bilinear"); layerParams.set("align_corners", options->align_corners()); layerParams.set("half_pixel_centers", options->half_pixel_centers()); } else if (opcode == "RESIZE_NEAREST_NEIGHBOR") { auto options = op.builtin_options_as_ResizeNearestNeighborOptions(); layerParams.set("interpolation", "nearest"); layerParams.set("align_corners", options->align_corners()); layerParams.set("half_pixel_centers", options->half_pixel_centers()); } Mat shape = allTensors[op.inputs()->Get(1)].reshape(1, 1); layerParams.set("height", shape.at(0, 0)); layerParams.set("width", shape.at(0, 1)); addLayer(layerParams, op); } void TFLiteImporter::parseTranspose(const Operator& op, const std::string& opcode, LayerParams& layerParams) { layerParams.type = "Permute"; std::vector perm = allTensors[op.inputs()->Get(1)]; DataLayout inpLayout = layouts[op.inputs()->Get(0)]; if (inpLayout == DNN_LAYOUT_NHWC && perm.size() == 4) { // OpenCV operates under the assumption that NCHW format, whereas TFLite defaults to NHWC. // Therfore, to align these layouts, the axes of the permutation vector should be adjusted accordingly. // For implementation details, please refer to the disscusion: // https://github.com/opencv/opencv/pull/25297#issuecomment-2049762298 if (perm[0] != 0) { CV_Error(Error::StsParseError, "The first axis should not be permuted."); } if (perm[1] == 1 && perm[2] == 2 && perm[3] == 3) { std::vector orderLP = {0, 1, 2, 3}; layerParams.set("order", DictValue::arrayInt(orderLP.data(), orderLP.size())); layouts[op.outputs()->Get(0)] = DNN_LAYOUT_NCHW; } else if (perm[1] == 1 && perm[2] == 3 && perm[3] == 2) { std::vector orderLP = {0, 3, 2, 1}; layerParams.set("order", DictValue::arrayInt(orderLP.data(), orderLP.size())); } else if (perm[1] == 2 && perm[2] == 1 && perm[3] == 3) { std::vector orderLP = {0, 1, 3, 2}; layerParams.set("order", DictValue::arrayInt(orderLP.data(), orderLP.size())); layouts[op.outputs()->Get(0)] = DNN_LAYOUT_NCHW; } else if (perm[1] == 2 && perm[2] == 3 && perm[3] == 1) { std::vector orderLP = {0, 2, 3, 1}; layerParams.set("order", DictValue::arrayInt(orderLP.data(), orderLP.size())); } } else { layerParams.set("order", DictValue::arrayInt(perm.data(), perm.size())); } addLayer(layerParams, op); } void TFLiteImporter::parseGlobalPooling(const Operator& op, const std::string& opcode, LayerParams& layerParams) { layerParams.type = "Pooling"; if(opcode == "MEAN") { layerParams.set("pool", "ave"); } else if (opcode == "REDUCE_MAX") { layerParams.set("pool", "max"); } else { CV_Error(Error::StsNotImplemented, "Unsupported pooling " + opcode); } layerParams.set("global_pooling", true); auto options = op.builtin_options_as_ReducerOptions(); bool keep_dims = options->keep_dims(); if (!keep_dims) { const auto name = layerParams.name; layerParams.name += "/global_pooling"; addLayer(layerParams, op); int out = op.outputs()->Get(0); auto outId = layerIds[out]; int flattenId = addFlattenLayer(1, -1, name, outId, isInt8(op) ? CV_8S : CV_32F); layerIds[out] = std::make_pair(flattenId, 0); } else { addLayer(layerParams, op); } } int TFLiteImporter::addPermuteLayer(const std::vector& order, const std::string& permName, const std::pair& inpId, int dtype) { LayerParams permLP; permLP.set("order", DictValue::arrayInt(order.data(), order.size())); int permId = dstNet.addLayer(permName, "Permute", dtype, permLP); dstNet.connect(inpId.first, inpId.second, permId, 0); return permId; } int TFLiteImporter::addReshapeLayer(const std::vector& shape, int axis, int num_axes, const std::string& name, const std::pair& inpId, int dtype) { LayerParams lp; lp.set("axis", axis); lp.set("dim", DictValue::arrayInt(shape.data(), shape.size())); lp.set("num_axes", num_axes); int id = dstNet.addLayer(name, "Reshape", dtype, lp); dstNet.connect(inpId.first, inpId.second, id, 0); return id; } int TFLiteImporter::addFlattenLayer(int axis, int end_axis, const std::string& name, const std::pair& inpId, int dtype) { LayerParams lp; lp.set("axis", axis); lp.set("end_axis", end_axis); int id = dstNet.addLayer(name, "Flatten", dtype, lp); dstNet.connect(inpId.first, inpId.second, id, 0); return id; } void TFLiteImporter::parseDeconvolution(const Operator& op, const std::string& opcode, LayerParams& layerParams) { layerParams.type = "Deconvolution"; CV_CheckLE(op.custom_options()->size(), sizeof(TfLiteTransposeConvParams), ""); const auto* params = reinterpret_cast(op.custom_options()->Data()); if (params->padding != kTfLitePaddingUnknown) layerParams.set("pad_mode", params->padding == kTfLitePaddingSame ? "SAME" : "VALID"); layerParams.set("stride_w", params->stride_width); layerParams.set("stride_h", params->stride_height); // Get filter size int filterIdx = op.inputs()->Get(1); Mat filter = allTensors[filterIdx]; int oc = filter.size[0]; int kh = filter.size[1]; int kw = filter.size[2]; int ic = filter.size[3]; layerParams.set("kernel_w", kw); layerParams.set("kernel_h", kh); layerParams.set("num_output", oc); // Add adjust padding similar to TensorFlow (see tf_importer) const auto* outShape = modelTensors->Get(op.outputs()->Get(0))->shape(); const int outH = outShape->Get(1); const int outW = outShape->Get(2); if (params->padding == kTfLitePaddingSame) { layerParams.set("adj_w", (outW - 1) % params->stride_width); layerParams.set("adj_h", (outH - 1) % params->stride_height); } else if (params->padding == kTfLitePaddingValid) { layerParams.set("adj_w", (outW - kw) % params->stride_width); layerParams.set("adj_h", (outH - kh) % params->stride_height); } // Reorder filter data from OHWI to IOHW and change shape correspondingly. filter = allTensors[filterIdx] = filter.reshape(1, {ic, oc, kh, kw}); CV_CheckTypeEQ(filter.type(), CV_32F, ""); Mat filterCopy = filter.clone(); float* data = filterCopy.ptr(); float* dstData = filter.ptr(); int total = oc * ic * kh * kw; for (int i_oc = 0; i_oc < oc; i_oc++) { for (int i_ic = 0; i_ic < ic; i_ic++) { for (int i_h = 0; i_h < kh; i_h++) { for (int i_w = 0; i_w < kw; i_w++) { int dst_i = kw * (kh * (oc * i_ic + i_oc) + i_h) + i_w; int src_i = ic * (kw * (kh * i_oc + i_h) + i_w) + i_ic; CV_CheckLT(dst_i, total, ""); CV_CheckLT(src_i, total, ""); dstData[dst_i] = data[src_i]; } } } } addLayer(layerParams, op); } void TFLiteImporter::parseQuantize(const Operator& op, const std::string& opcode, LayerParams& layerParams) { layerParams.type = "Quantize"; layerParams.set("scales", 1); layerParams.set("zeropoints", -128); addLayer(layerParams, op); } void TFLiteImporter::parseDequantize(const Operator& op, const std::string& opcode, LayerParams& layerParams) { layerParams.type = "Dequantize"; float inpScale, outScale; int inpZero, outZero; getQuantParams(op, inpScale, inpZero, outScale, outZero); layerParams.set("scales", inpScale); layerParams.set("zeropoints", inpZero); addLayer(layerParams, op); } void TFLiteImporter::parseSplit(const Operator& op, const std::string& opcode, LayerParams& layerParams) { layerParams.type = "Slice"; auto options = op.builtin_options_as_SplitOptions(); CV_Assert(options); layerParams.set("num_split", options->num_splits()); addLayer(layerParams, op); } void TFLiteImporter::parseFullyConnected(const Operator& op, const std::string& opcode, LayerParams& layerParams) { layerParams.type = "Gemm"; auto options = op.builtin_options_as_FullyConnectedOptions(); CV_Assert(options); layerParams.set("transB", true); layerParams.set("constB", true); addLayer(layerParams, op); parseFusedActivation(op, options->fused_activation_function()); } void TFLiteImporter::parseSoftmax(const Operator& op, const std::string& opcode, LayerParams& layerParams) { layerParams.type = "Softmax"; addLayer(layerParams, op); } void TFLiteImporter::parseCast(const Operator& op, const std::string& opcode, LayerParams& layerParams) { layerParams.type = "Identity"; addLayer(layerParams, op); } void TFLiteImporter::parseDetectionPostProcess(const Operator& op, const std::string& opcode, LayerParams& layerParams) { // Parse parameters; std::vector keys(1, ""); const uint8_t* data = op.custom_options()->Data(); int offset = 0; // Read zero delimited keys while (data[offset] != 10 && offset < op.custom_options()->size()) { if (data[offset]) { keys.back() += data[offset]; } else { keys.emplace_back(""); } offset += 1; } keys.pop_back(); std::sort(keys.begin(), keys.end()); // TODO: Replace empirical offset to something more reliable. offset += 25; std::map parameters; for (int i = 0; i < keys.size(); ++i) { parameters[keys[i]] = *reinterpret_cast(data + offset + i * 4); } parameters["num_classes"] = modelTensors->Get(op.inputs()->Get(1))->shape()->Get(2); layerParams.type = "DetectionOutput"; layerParams.set("num_classes", parameters["num_classes"]); layerParams.set("share_location", true); layerParams.set("background_label_id", parameters["num_classes"] + 1); layerParams.set("nms_threshold", *(float*)¶meters["nms_iou_threshold"]); layerParams.set("confidence_threshold", *(float*)¶meters["nms_score_threshold"]); layerParams.set("top_k", parameters["max_detections"]); layerParams.set("keep_top_k", parameters["max_detections"]); layerParams.set("code_type", "CENTER_SIZE"); layerParams.set("loc_pred_transposed", true); // Replace third input from tensor to Const layer with the priors Mat priors = allTensors[op.inputs()->Get(2)].clone(); // Change priors data from (ycenter, xcenter, h, w) to (xmin, ymin, xmax, ymax) priors = priors.reshape(1, priors.total() / 4); Mat tmp = priors.col(0).clone(); priors.col(0) = priors.col(1) - 0.5 * priors.col(3); priors.col(1) = tmp - 0.5 * priors.col(2); tmp = priors.col(2).clone(); priors.col(2) = priors.col(0) + priors.col(3); priors.col(3) = priors.col(1) + tmp; float x_scale = *(float*)¶meters["x_scale"]; float y_scale = *(float*)¶meters["y_scale"]; float w_scale = *(float*)¶meters["w_scale"]; float h_scale = *(float*)¶meters["h_scale"]; if (x_scale != 1.0f || y_scale != 1.0f || w_scale != 1.0f || h_scale != 1.0f) { int numPriors = priors.rows; priors.resize(numPriors * 2); Mat_ scales({1, 4}, {1.f / x_scale, 1.f / y_scale, 1.f / w_scale, 1.f / h_scale}); repeat(scales, numPriors, 1, priors.rowRange(numPriors, priors.rows)); priors = priors.reshape(1, {1, 2, (int)priors.total() / 2}); layerParams.set("variance_encoded_in_target", false); } else { priors = priors.reshape(1, {1, 1, (int)priors.total()}); layerParams.set("variance_encoded_in_target", true); } LayerParams priorsLP; priorsLP.name = layerParams.name + "/priors"; priorsLP.type = "Const"; priorsLP.blobs.resize(1, priors); int priorsId = dstNet.addLayer(priorsLP.name, priorsLP.type, priorsLP); layerIds[op.inputs()->Get(2)] = std::make_pair(priorsId, 0); addLayer(layerParams, op); } void TFLiteImporter::parseFusedActivation(const Operator& op, ActivationFunctionType activ) { LayerParams activParams; activParams.name = modelTensors->Get(op.outputs()->Get(0))->name()->str() + "/activ"; parseActivation(op, EnumNameActivationFunctionType(activ), activParams, true); } void TFLiteImporter::parseActivation(const Operator& op, const std::string& opcode, LayerParams& activParams) { parseActivation(op, opcode, activParams, false); } void TFLiteImporter::parseActivation(const Operator& op, const std::string& opcode, LayerParams& activParams, bool isFused) { if (opcode == "NONE") return; else if (opcode == "RELU6") activParams.type = "ReLU6"; else if (opcode == "PRELU") activParams.type = "PReLU"; else if (opcode == "RELU") activParams.type = "ReLU"; else if (opcode == "HARD_SWISH") activParams.type = "HardSwish"; else if (opcode == "LOGISTIC") activParams.type = "Sigmoid"; else CV_Error(Error::StsNotImplemented, "Unsupported activation " + opcode); if (isInt8(op)) { float inpScale, outScale; int inpZero, outZero; getQuantParams(op, inpScale, inpZero, outScale, outZero); if (isFused) { activParams.type += "Int8"; activParams.set("input_scale", outScale); activParams.set("input_zeropoint", outZero); activParams.set("scales", outScale); activParams.set("zeropoints", outZero); } Mat lookUpTable(1, 256, CV_8S); int8_t* table = lookUpTable.ptr(); for (int i = -128; i < 128; i++) { float x, y = i; if (isFused) x = outScale * (i - outZero); else x = inpScale * (i - inpZero); if (opcode == "RELU6") y = std::min(std::max(x, 0.f), 6.f); else if (opcode == "LOGISTIC") y = 1.0f / (1.0f + std::exp(-x)); else if (opcode == "HARD_SWISH") y = x * max(0.f, min(1.f, x / 6.f + 0.5f)); else CV_Error(Error::StsNotImplemented, "Lookup table for " + opcode); int quantized = outZero + cvRound(y / outScale); table[i + 128] = saturate_cast(quantized); } activParams.blobs.resize(1, lookUpTable); } if (isFused) { int dtype = isInt8(op) ? CV_8S : CV_32F; int layerId = dstNet.addLayerToPrev(activParams.name, activParams.type, dtype, activParams); // Override layer ids mapping int i = 0; for (int idx : *op.outputs()) { layerIds[idx] = std::make_pair(layerId, i++); } } else { addLayer(activParams, op); } } bool TFLiteImporter::isInt8(const Operator& op) { const Tensor* out = modelTensors->Get(op.outputs()->Get(0)); return out->type() == TensorType_INT8; } void TFLiteImporter::getQuantParams(const Operator& op, float& inpScale, int& inpZero, float& outScale, int& outZero) { const Tensor* inp = modelTensors->Get(op.inputs()->Get(0)); const Tensor* out = modelTensors->Get(op.outputs()->Get(0)); inpScale = outScale = inpZero = outZero = 0; if (inp->quantization()) { if (inp->quantization()->scale()) { CV_Assert(inp->quantization()->scale()->size() == 1); inpScale = inp->quantization()->scale()->Get(0); } if (inp->quantization()->zero_point()) { CV_Assert(inp->quantization()->zero_point()->size() == 1); inpZero = inp->quantization()->zero_point()->Get(0); } } if (out->quantization()) { if (out->quantization()->scale()) { CV_Assert(out->quantization()->scale()->size() == 1); outScale = out->quantization()->scale()->Get(0); } if (out->quantization()->zero_point()) { CV_Assert(out->quantization()->zero_point()->size() == 1); outZero = out->quantization()->zero_point()->Get(0); } } } Net readNetFromTFLite(const String &modelPath) { Net net; std::vector content; const std::ios::openmode mode = std::ios::in | std::ios::binary; std::ifstream ifs(modelPath, mode); if (!ifs.is_open()) CV_Error(Error::StsError, cv::format("DNN/TFLite: can't open model file '%s'", modelPath.c_str())); ifs.seekg(0, std::ios::end); const size_t sz = ifs.tellg(); CV_Assert(sz > 0); content.resize(sz); ifs.seekg(0, std::ios::beg); ifs.read(content.data(), sz); CV_Assert(!ifs.bad()); TFLiteImporter(net, content.data(), content.size()); return net; } Net readNetFromTFLite(const std::vector& bufferModel) { return readNetFromTFLite((const char*)bufferModel.data(), bufferModel.size()); } Net readNetFromTFLite(const char *bufferModel, size_t bufSize) { Net net; TFLiteImporter(net, bufferModel, bufSize); return net; } #else // HAVE_FLATBUFFERS #define DNN_TFLITE_UNSUPPORTED() CV_Error(Error::StsError, "DNN/TFLite: Build OpenCV with FlatBuffers to import TFLite models: https://github.com/opencv/opencv/pull/23161") Net readNetFromTFLite(const String &) { DNN_TFLITE_UNSUPPORTED(); } Net readNetFromTFLite(const std::vector&) { DNN_TFLITE_UNSUPPORTED(); } Net readNetFromTFLite(const char *, size_t) { DNN_TFLITE_UNSUPPORTED(); } #endif // HAVE_FLATBUFFERS CV__DNN_INLINE_NS_END }} // namespace cv::dnn