You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
A high-performance, on-device neural network inference framework.
Goal
This project ONE aims at providing a high-performance, on-device neural network (NN) inference
framework that performs inference of a given NN model on processors, such as CPU, GPU, DSP or NPU.
We develop a runtime that runs on a Linux kernel-based OS platform such as Ubuntu, Tizen, or
Android, and a compiler toolchain to support NN models created using various NN training frameworks such
as Tensorflow or PyTorch in a unified form at runtime.
If the feature you want is on the list, 👍 to the body of the issue. The feature with the most
👍 is placed at the top of the list. When adding new features, we will prioritize them with this reference.
Of course, it is good to add an additional comment which describes your request in detail.
For features not listed, create a new issue.
Sooner or later, the maintainer will tag the FEATURE_REQUEST label and appear on the list.
We expect one of the most frequent feature requests would be the operator kernel implementation.
It is good to make a request, but it is better if you contribute by yourself. See the following guide,
How to add a new operation, for help.
We are looking forward to your participation.
Thank you in advance!
How to Contact
Please post questions, issues, or suggestions into Issues. This is the best way to communicate with the developer.
You can also have an open discussion with community members through gitter.im channel.