-
**Describe the solution you'd like**
I found that the latest release of tensorrt 8.0 is support for the int8 quantization on GPU, which is great accelerate inference speed.
And now onnxruntime is …
-
在试用dipoorlet PTQ量化 torch 导出的onnx模型时报错: ValueError: cannot reshape array of size 172800 into shape (0,0,3,180,320)。
1. torch.onnx.export() 导出时指定了dynamic_axes, 具体如下:
```
torch.onnx.export(
…
-
Can it be used in sd1.5 and can it be combined with other acceleration methods such as ByteDance/Hyper SD
-
首先感谢大佬无私的开源, 整套工具都很全, 从训练到部署, 支持的也比较全, 但是我看没有做8bits量化,大佬请问是有尝试过么, 做量化会损失精度么?
-
I tried to run the exported onnx file on both RTX3070 and RTX 4090, but can not see speed improvement (even slower than the unquantized model). Here is the warning of onnxruntime:
`2024-09-20 19:58:0…
-
### 1. System information
- Occurs in Google Colab w/ TF 2.14
- Have also verified w. TF 2.7 (Anaconda) on Windows 10
### 2. Code
[Colab to reproduce issue](https://colab.research.google.com…
-
It was announced that there will be a Team Trios Pro Tour in 2018, as well as an increase in team Grand Prix and team PTQs. It seems like it would be a great option to have on xmage.
Team Construct…
-
## Motivation
1. To design and implement the better quantization part of MMRazor with community.
1. Collect more requirements and suggestions before releasing quantization by RFC (Request for C…
-
## **Summary**
This is a design discussion RFC for contributing some device-agnostic compression algorithms, like the post training quantization(QDQ quant format) and structural sparsity supported …
-
Hi, maintainer of the `swi-prolog` Arch Linux package here.
When upgrading the package from 9.2.4 to 9.2.7, one of the tests stopped working:
```
ctest --test-dir build --output-on-failure
Int…