Open aaditya-srivathsan opened 6 months ago
[8] Assertion failed: (axes.allValuesKnown()) && "This version of TensorRT does not support dynamic axes."
Check [Slice -> "onnx::Slice_934"] and see if the dynamic axes can be eliminated by constant folding or other ways.
Check [Slice -> "onnx::Slice_934"] and see if the dynamic axes can be eliminated by constant folding or other ways.
@zerollzeng I did try folding the model using polygraph surgeon
but that doesnt work either. Is there some other ways you can tell me about other than this approach?
Normally it requires modify the original model code and re-export the onnx, although sometimes it's impossible if dynamic shapes is really needed...
@zerollzeng is there any roadmap plans to support dynamic axes in the times to come?
@aaditya-srivathsan We’ve added dynamic axes support in 10.3 which is the next release.
@Nanthini10 Do you know when 10.3 will get released?
Description
I have an onnx model i would like to convert to a trt engine to run some perf testing and see the differences in performance. For context, this is a DINO model generated by the MMDEPLOY packages and also a dependency on a shared object file. The onnx backend itself works as expected at inference time.
But for some reason while trying to convert the model using trtexec like so
I get the following error
I have tried using a static deploy config for the same DINO model config but that doesnt work wither. Any idea how to potentially fix this issue? I am running the trtexec commands on a 23.08 version release of the tensortRT container
Environment
TensorRT Version: 8.6.1
NVIDIA GPU: T4
NVIDIA Driver Version: 515
CUDA Version: 11.7
CUDNN Version:
Operating System:Ubuntu 20.7
Python Version (if applicable): 3.8
Tensorflow Version (if applicable): None
PyTorch Version (if applicable): None
Baremetal or Container (if so, version): TensorRT-23.08-py
Relevant Files
Model link: For privacy reasons, cannot share the onnx file
Steps To Reproduce
Commands or scripts:
Have you tried the latest release?: yes, that doesnt work either
Can this model run on other frameworks? For example run ONNX model with ONNXRuntime (
polygraphy run <model.onnx> --onnxrt
): No polygraph throws the same error as well. Moreover folding constants on this file doesnt work either