Open fg-mindee opened 3 years ago
ONNX conversion seems to be incompatible with TF 2.4.* as per https://github.com/onnx/keras-onnx/issues/662. I tried on my end and encountered the same problem. Moving this to the next release until this gets fixed!
A good lead for ONNX support would be to use https://github.com/onnx/tensorflow-onnx (might have to create a savemodel to use it but it's worth a look)
@frgfm i think we can remove the tensorrt point If we support onnx wdyt ?
Yes sure! We'll need to take a look at pruning at some point
yeah pruning is fine but tensorrt is a bit to much (should do the user on his own side if we can provide onnx this should be not so tricky)
Add a
doctr.models.utils
module to compress existing models and improve their latency / memory load for inference purposes on CPU. Some interesting leads to investigate:Optional: TensorRT export (cf. https://developer.nvidia.com/blog/speeding-up-deep-learning-inference-using-tensorflow-onnx-and-tensorrt/)