-
Hi,
I posted the error message in the repo of Tensorrt and they refered to this repo, so I open an issue here. The problem is that when I quantize the model in pytorch with modelopt and export it t…
-
While running this example:
```
$ cd TensorRT-Model-Optimizer/llm_ptq
$ scripts/huggingface_example.sh --type llama --model $model --quant fp8 --tp 2
```
there was a non-fatal failure:
```
[8ad0971d…
-
I use the AIMET PTQ to quantize the CLIP text model.
But I encounter this error [KeyError: 'Graph has no buffer /text_model/encoder/layers.0/layer_norm1/Constant_output_0, referred to as input for …
-
I search in the doc and find setting for inference for model of QAT. Is there a function for inference of PTQ model?
-
Hi, can you share best practices for quantization for CNN models?
Are the modelopt quantized PTQ is the way to go with tensorrt for cnn models (resnet retinanet etc)? I was able to quantize retinanet…
-
# PTQ | Faster Coding With Vim
Bài viết dành cho những người chưa từng sử dụng vim! Động lực Nếu bạn gõ 10 ngón thì bàn phím như 1 cánh tay nối dài hay 1 phần cơ thể của bạn vậy. Lúc đó bạn chỉ tập t…
-
When I try to ptq mobilenetv2, I get an error: "ImportError: cannot import name 'ConvBNReLUFusion' from 'torch.quantization.fx.fusion_patterns'"
-
onnx_ptq/evaluate_vit.py error: ValueError: Runtime TRT is not supported.
![企业微信截图_17225064037714](https://github.com/user-attachments/assets/b1ad1ffc-9744-46ac-8d2e-ed6aeb5584a2)
-
Here we keep track of what part of `quantize` in `ptq_common.py` are tested and what are still missing.
-
When I runned the [ptq.py](https://github.com/open-mmlab/mmrazor/blob/main/tools/ptq.py), Unfortunately It threw an error, and the error message is as follows. The reason for the error is most likely …