You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Model API is a set of wrapper classes for particular tasks and model architectures, simplifying data preprocess and postprocess as well as routine procedures (model loading, asynchronous execution, etc.). It is aimed at simplifying end-to-end model inference for different deployment scenarios, including local execution and serving. The Model API is based on the OpenVINO inference API.
How it works
Model API searches for additional information required for model inference, data, pre/postprocessing, label names, etc. directly in OpenVINO Intermediate Representation. This information is used to prepare the inference data, process and output the inference results in a human-readable format.
Currently, ModelAPI supports models trained in OpenVINO Training Extensions framework.
Training Extensions embed all the metadata required for inference into model file. For models coming from other than Training Extensions frameworks metadata generation step is required before using ModelAPI.
OpenCV and OpenVINO locations are optional. In most cases, these dependencies are discovered by cmake without extra guidance.
Build:
cmake --build . -j
To build a .tar.gz package with the library, run:
cmake --build . --target package -j
Usage
Python
frommodel_api.modelsimportModel# Create a model wrapper from a compatible model generated by OpenVINO Training Extensions# Use URL to work with OVMS-served model, e.g. "localhost:9000/models/ssdlite_mobilenet_v2"model=Model.create_model("model.xml")
# Run synchronous inference locallyresult=model(image) # image is numpy.ndarray# Print results in model-specific formatprint(f"Inference result: {result}")
C++
In C++ we have to specify model type in advance, let's set it to detection model.
#include<models/detection_model.h>
#include<models/results.h>// Load the modelauto model = Model::create_model("model.xml");
// Run synchronous inference locallyauto result = model->infer(image); // image is cv::Mat// Iterate over the vector of DetectedObject with box coordinates, confidence and label stringfor (auto& obj : result->objects) {
std::cout << obj.label << " | " << obj.confidence << " | " << int(obj.x) << " | " << int(obj.y) << " | "
<< int(obj.x + obj.width) << " | " << int(obj.y + obj.height) << std::endl;
}
Model's static method create_model() has two overloads. One constructs the model from a string (a path or a model name) (shown above) and the other takes an already constructed InferenceAdapter.
Prepare a model for InferenceAdapter
There are usecases when it is not possible to modify an internal ov::Model and it is hidden behind InferenceAdapter. For example the model can be served using OVMS. create_model() can construct a model from a given InferenceAdapter. That approach assumes that the model in InferenceAdapter was already configured by create_model() called with a string (a path or a model name). It is possible to prepare such model using C++ or Python:
C++
auto model = DetectionModel::create_model("~/.cache/omz/public/ssdlite_mobilenet_v2/FP16/ssdlite_mobilenet_v2.xml");
const std::shared_ptr<ov::Model>& ov_model = model->getModel();
ov::serialize(ov_model, "serialized.xml");