Closed alsozatch closed 6 months ago
Could you please share a reproduce? Thanks!
Here is a reproducible example. Run 'python3 reproducible_int8_exporter.py'. reproducible_int8_exporter.zip
I can also reproduce the issue on TRT 8.6 on x86. But in my test it's been fixed in TRT 10.
9.2 also pass.
Thanks. I’m on a Jetson AGX Orin though, so aarch64 not x86 and JetPack 6.0 is the latest for at least half a year probably more, so I’m stuck at TensorRT 8.6. Is there any more information you can obtain to help with this?
Sorry for the late reply, I'm checking internally.
Could you please try mark /model.0/conv/Conv
or /model.0/act/Sigmoid
as network output so that the layer fusion can be break?
this can be done quickly with polygraphy run model.onnx --mark ouptut_layer_name
to test, for details please check polygraphy run -h
closing since no activity for more than 3 weeks, pls reopen if you still have question. Thanks!
Any update on this?
Same issue here on jetpack 6.0
.....
On jetpack 5.1.0
works fine. Is this related to cuda 12
? (same issue when using Tensorrt 10.4
and cuda 12.2
on a x86 machine)
Description
INT8 quantization fails with error in title on Jetson AGX Orin in latest JetPack 6.0 container. I also tried this previously on JetPack 5.1.2 and it failed with the same issue so I tried the latest version of TensorRT 8.6.2. Can I get some help with this? Thanks. Full log below.
Environment
TensorRT Version: 8.6.2
NVIDIA GPU: Ampere on Jetson AGX Orin
NVIDIA Driver Version:
CUDA Version: 12.2
CUDNN Version:
Operating System:
Python Version (if applicable): 3.10.12
PyTorch Version (if applicable): 2.1.0
Baremetal or Container (if so, version): dustynv/l4t-pytorch:r36.2.0