-
## Description
the model after pytorch_quantization qat, accuracy descend relative to before pytorch_quantization qat
## Environment
**TensorRT Version**: 8.5.3.1
**NVIDIA GPU**: TIT…
-
### Description of the Feature:
torch QAT supports three mode.
* Eager Mode Quantization
* FX Graph Mode Quantization
* PyTorch 2 Export Quantization
detail in https://pytorch.org/docs/stabl…
-
### Search before asking
- [X] I have searched the Ultralytics YOLO [issues](https://github.com/ultralytics/ultralytics/issues) and found no similar bug report.
### Ultralytics YOLO Component
_No …
-
### System Info / 系統信息
torch 2.4.1 / diffuser 0.30.2 / Ubuntu 22.04.4 LTS / Cuda driver 12.6
### Information / 问题信息
- [X] The official example scripts / 官方的示例脚本
- [ ] My own modified scripts / 我自己…
-
I tested the `torchao.quantize_` and `torchao.autoquant` for the first time on a toy model and wanted to summarize my usability/readability comments, in case that is helpful.
# torchao.quantize_ fe…
vkuzo updated
2 weeks ago
-
https://github.com/intel/neural-compressor/blob/master/docs/source/quantization_weight_only.md#examples
how to set eval_func?
https://github.com/intel/neural-compressor/blob/master/examples/3…
-
I have tried to quantize a model by following the guide ([PyTorch Quantization — Model Optimizer 0.15.0](https://nvidia.github.io/TensorRT-Model-Optimizer/guides/_pytorch_quantization.html)), and I ca…
-
I found a statement which said **3. Better support for vision transformers.** in link https://nvidia.github.io/TensorRT-Model-Optimizer/guides/_onnx_quantization.html.
I'm working on quantizing VIT n…
-
A lot of code for tensor subclasses can likely be conslidated together into a base class that other classes can utilize
_get_to_kwargs:
https://github.com/pytorch/ao/blob/main/torchao/dtypes/affin…
-
I am accelerating a custom pytorch network using Vitis-AI. After following the steps below the model is quantized and the .xmodel is compiled, however the accuracy of the model takes a huge hit going …