-
This is to track progress of fx graph mode quantization tutorials:
- [x] post training static quantization
- [x] post training dynamic quantization
- [ ] quantization aware training
cc @jianyu…
-
## Description
I am trying to figure out if TensoRT and the `pytorch_quantization` module support post-training quantization for vision transformers.
The following piece of code follows the `pyt…
RuRo updated
2 months ago
-
Hello!
I'm looking into adding diffusion based image generation into a video game for fairly simple stuff like decals. The game will be running under some pretty heavy resource constraints. If I …
-
### System Info
CPU-X86
GPU-H100
Server XE9640
Code: TensorRT-LLM 0.8.0 release
### Who can help?
@Tracin @juney-nvidia
Regarding the [FP8 Post Quantization]((https://github.com/NVIDIA/Tenso…
-
Hi there
**Describe the bug**
I used [Convert TF Object Detection API model to TFLite](https://colab.research.google.com/github/tensorflow/models/blob/master/research/object_detection/colab_tutori…
-
Post-training quantization (PTQ) - without finetune and Quantization aware training (QAT) works fine but
get error in Post-training quantization (PTQ) - fast finetune:
activation = layer.layer.acti…
-
https://tensorflow.google.cn/lite/performance/quantization_spec
from this document we can see the post-training quantization do asymmetric quantization for activations, do symmetric quantization for …
-
Inspired by a recent back and forth with @gau-nernst we should add some quantized training recipes in AO for small models (600M param range)
Character.ai recently shared that they're working on qua…
-
Hello, I build Bolt(tag: v1.5.1) with the linux-x86_64_avx512 version, and convert onnx model to PTQ version by X2bolt.Then try post_training_quantization to quantize it to int8 precision. I follow th…
-
### 🐛 Describe the bug
from torch.ao.quantization.quantizer import (
XNNPACKQuantizer,
get_symmetric_quantization_config,
)
the code abve report error:
ImportError: cannot import name 'X…