Closed zengpeng629 closed 2 years ago
You might have noticed that it is possible to use ONNXRT with C++ as well and there are examples for it (you are not restricted to Python). Those high level opensource runtimes are not just for experiments - they are for deployment as well. Efficiency is very good - as long as the graph is fully offloaded to underlying TIDL there is not much efficiency loss.
Your question whether to use the high level gstreamer interface in StarterKit or the low level TIOVX interface. I think every company will have their own strategy regarding the tools to use - it's hard to generalize. If your company uses Linux then I think it is a good idea to use the high level runtimes. But if you only run RTOS in the SoC then your only option is to use TIOVX.
Hi Manu,
Thank you again for your kind help!
It is more clear now about the deployment process! Another question is that could you share some document links where I can learn about how the graph offloading works?
Hi Zeng,
Have you looked at the TIDL documentation? For the 8.0 release, it is here: https://software-dl.ti.com/jacinto7/esd/processor-sdk-rtos-jacinto7/08_00_00_12/exports/docs/tidl_j7_08_00_00_10/ti_dl/docs/user_guide_html/index.html This should give you some details.
The exact offloading mechanism is specific to the runtime. For example, see these links: https://onnxruntime.ai/docs/execution-providers/ https://www.tensorflow.org/lite/performance/delegates
Thank you so much!
I will check the documents, those are very helpful to me!
Best regards, Zeng
Hi TI team,
I am currently working on deploying a detection model on tda4 and aiming to make it work for AD usage in the future (for cars). I followed edgeai tutorials and documents and successfully ran inference on TDA4VM with ONNX Runtime Interface using python, so here is my question:
Best, Zeng