-
Hi, when I convert a quantized onnx model to trt engine using `trtexec --fp16 --int8 --onnx=xxx.onnx`, I met a error like below:
```
......
[03/01/2023-16:52:23] [TRT] [V] QuantizeDoubleInputNodes:…
-
### Checklist
- [X] I have searched related issues but cannot get the expected help.
- [X] 2. I have read the [FAQ documentation](https://github.com/open-mmlab/mmdeploy/tree/main/docs/en/faq.md) b…
-
### Search before asking
- [X] I have searched the YOLOv5 [issues](https://github.com/ultralytics/yolov5/issues) and [discussions](https://github.com/ultralytics/yolov5/discussions) and found no si…
-
Hi,
After doing quantization on the yolov7-tiny model with the recommended settings, I am getting a lower throughput with the result (qat-tiny.pt) on the benchmark when compared to doing the same ben…
-
## Description
Part of my network structure is as follows,
![YCC9pMhvae](https://user-images.githubusercontent.com/30153920/216005113-ddf9e4dc-0c26-479a-a4f8-1e5d3ae86aad.png)
log when genera…
-
## ❓ Question
When I'm not using TensorRT, I run my model through an FX interpreter that times each call op (by inserting CUDA events before/after and measuring the elapsed time). I'd like to do so…
-
**Describe the bug**
ONNX to ORT conversion works when device=CPU, but does not with Direct ML (exact same code)
Low level details:
- I tried with 7 networks and it happens in all of them (includ…
-
## Description
when i try to use bert demo on jetson-orin-nx-8g, some error occurs.
## Environment
**TensorRT Version**: 8.5.2.2
**NVIDIA GPU**:
**NVIDIA Driver Version**:
**CUDA Versi…
-
hello, i asked the same question before,i am using tensorrt8.5 to convert my onnx model to int8 model,
the special thing is that my onnx model contains a custom DCNV2 plugin, The python implementatio…
-
### Checklist
- [X] I have searched related issues but cannot get the expected help.
- [X] 2. I have read the [FAQ documentation](https://github.com/open-mmlab/mmdeploy/blob/master/docs/en/faq.md) bu…