-
根据文档,使用advanced ptq对torch官方mobilenet v2模型进行定点化,但是在测试过程中遇到了如下问题:
1. 文档中给出的ptq_reconstruction_config会报错
文档中的ptq_reconstruction_config为字典类型,但是实际代码在消费这个config时既用到了字典访问,又用到了attribute访问,导致报错:
在advanced_p…
-
## Description
## Environment
**TensorRT Version**: 8.4.1.5
**NVIDIA GPU**: 1080ti
**NVIDIA Driver Version**: 450
**CUDA Version**: 11.0
**CUDNN Version**: 8.1.0
**Operating System…
pangr updated
2 years ago
-
#117
-
### Environment
- Python: 32-bit, 3.6
- pyodbc: 4.0.23
- OS: Windows 10
- DB: Access / SQL
- driver: Timberline Office
### Issue
This is my query:
```sql
SELECT CURRENT_JCT_TRANSACTI…
-
hello,
I want use the mobilenetv2-10.onnx to convert to pytorch and then use PostQuantizer to int8 tflite.
I use onnx2pytorch (https://github.com/ToriML/onnx2pytorch) package to convert model to…
-
## Bug Description
Compiling torchscript modules to tensorrt was working fine in last release.
Now the simple script below fails with the following error message.
I have tried with tracing, sc…
-
I am try to run the deeplab segmentation model in Directory “neural-compressor/examples/tensorflow/semantic_image_segmentation/deeplab/quantization/ptq/main.py”
Data is preferred correctly, which is …
-
hello, i am using tensorrt8.3 to convert my onnx model to int8 model,
the special thing is that my onnx model contains a custom DCNV2 plugin, The python implementation of DCNV2 is based on the follow…
-
Recently, I quantized a pre-trained ResNet50 model from fp32 to int8, and I noticed that the performance isn't what I expected. The performance is only about 2x compared to the equivalent fp32 model. …
-
## Description
## Environment
**TensorRT Version**: 8.4.1
**NVIDIA GPU**: A10
**NVIDIA Driver Version**: 510.47.03
**CUDA Version**: 11.6:
After int8 quantization of the model, I fo…