onnx / tensorflow-onnx

Convert TensorFlow, Keras, Tensorflow.js and Tflite models to ONNX
Apache License 2.0
2.31k stars 433 forks source link

Much Higher inference time using converted onnx model #2114

Open purvang3 opened 1 year ago

purvang3 commented 1 year ago

Ask a Question

Question

I am converting tensorflow 1.14 trained tensorflow object detection model to onnx and running inference.
I am using tensorflow/tensorflow:1.14.0-gpu-py3 docker image. 
using onnxruntime-gpu==1.1 or latest, in both I am getting 5x higher inference time than frozen_graph.pb.

Actuallly onnxruntime using cpu instead of Gpu. is this because of wrong onnxruntime-gpu version?

if yes, then which onnx version should I install? I already tested with 1.1 and latest release.

Further information

Model opset: 13

Notes

fatcat-z commented 1 year ago

Could you please open an issue in onnxruntime repo for a suggestion about inference performance?