MicroFlow is a robust and efficient TinyML inference engine designed for deploying machine learning models on embedded systems. It was developed by Matteo Carnelos as part of his master's thesis project at the University of Padova in collaboration with Grepit AB.
MicroFlow uses a compiler-based approach, resulting in the following engine structure:
graph LR
subgraph host[Host]
model(Neural Network Model) --> compiler(MicroFlow Compiler)
end
subgraph target[Target]
code(Generated Source Code) --- weights[(Weights)]
code --- runtime(MicroFlow Runtime)
end
compiler --> code
compiler --> weights
MicroFlow consists of two primary components: the compiler, represented by the microflow-macros
crate, and the runtime, represented by the microflow
crate.
The compiler, which runs prior to the Rust compiler, is responsible for parsing and pre-processing the model.
It generates the necessary source code to enable inference on the model.
On the other hand, the runtime is a [no_std]
component designed to run on the target MCU.
It encompasses the implementation of operators, activation functions, and quantization procedures.
MicroFlow utilizes Rust Procedural Macros as its user interface.
By applying the model
macro to a struct
and providing the model's path, the MicroFlow compiler generates a predict()
method.
This method can be called to perform inference on the given model.
Currently, MicroFlow only supports models in the TensorFlow Lite format (.tflite
).
Here is a minimal example showcasing the usage of MicroFlow:
use microflow::model;
#[model("path/to/model.tflite")]
struct MyModel;
fn main() {
let prediction = MyModel::predict(input_data);
}
The examples provided with MicroFlow can be found in the examples
folder.
To run an example on a target board, cd
into the board directory for the example (e.g. examples/arduino-uno
) and run the command:
cargo run --example <example-name>
Otherwise, to run the example locally, just run the above command in the root directory.
[!NOTE] For board examples, you might need to install additional tools and configure the runner to make the example work for your setup.
Currently, MicroFlow supports the following operators and activation functions:
Operator | Quantized | Tensor Type |
---|---|---|
FullyConnected |
✓ | Tensor2D |
Conv2D |
✓ | Tensor4D |
DepthwiseConv2D |
✓ | Tensor4D |
AveragePool2D |
✓ | Tensor4D |
Reshape |
✓ | Tensor2D , Tensor4D |
Activation Function | Quantized |
---|---|
ReLU |
✓ |
ReLU6 |
✓ |
Softmax |
✓ |
These operators and activation functions cover common building blocks for neural networks and enable efficient inference with reduced memory and computational requirements. However, MicroFlow's development roadmap includes plans for implementing additional operators and activation functions to expand the range of supported models.
The examples
folder contains the code used to test MicroFlow on different MCUs, including:
The models ued to test the inference engines can be found in the models
directory.
These models include:
Contributors are welcome. For major changes, please open an issue first to discuss what you would like to change. Please make sure to update tests as appropriate.
MicroFlow has been published on arxiv.org and can be cited as follows:
@misc{carnelos2024microflowefficientrustbasedinference,
title={MicroFlow: An Efficient Rust-Based Inference Engine for TinyML},
author={Matteo Carnelos and Francesco Pasti and Nicola Bellotto},
year={2024},
eprint={2409.19432},
archivePrefix={arXiv},
primaryClass={cs.LG},
url={https://arxiv.org/abs/2409.19432},
}
Licensed under either of
at your option.
Copyright © 2024, Matteo Carnelos