-
I want to use qat method for my model, but i can only find ptq quantizer in executorch, are there some examples of how to implement Quantization Aware Training (QAT) for qnn backend?
-
When I runned the [ptq.py](https://github.com/open-mmlab/mmrazor/blob/main/tools/ptq.py), Unfortunately It threw an error, and the error message is as follows. The reason for the error is most likely …
-
### Search before asking
- [X] I have searched the YOLOv6 [issues](https://github.com/meituan/YOLOv6/issues) and found no similar feature requests.
### Description
post training quantization using…
-
Can you release centerpoint.scn.PTQ.onnx? THANKS!
-
Greetings🐱! As a part of https://github.com/openvinotoolkit/nncf/issues/2766 TorchFX PTQ backend support, we are gladly presenting to you following issue
### Context
The task is to cover FXModel…
-
### 🐛 Describe the bug
from torch.ao.quantization.quantizer import (
XNNPACKQuantizer,
get_symmetric_quantization_config,
)
the code abve report error:
ImportError: cannot import name 'X…
-
devicec : nvidia NX
1.using trt --fp16
` /usr/src/tensorrt/bin/trtexec --onnx=best.onnx --workspace=4096 --saveEngine=best.engine --fp16`
the result of infer speed is 36.8ms
2. using pytorch_qua…
-
I have used PTQ for int8 export from pytorch model and despite attempts at calibration, there is a significant drop in detection accuracy.
I am moving to quantization aware training to improve the…
-
I used mtq.INT8_default_CFG as recommended for CNN networks (mtq.quantize(model, config, forward_loop). My initial model ran at 80FPS after quantization it dropped to 40FPS? I checked the model struct…
-
Hi neuralmagic team !
Very nice work with AutoFP8 ! We were thinking of integrating AutoFP8 in transformers, so that users can run your checkpoints directly with transformers ! We would simply rep…