-
What is the config file I need to parse into the tools ptq or qat file? is it the config file for the model like the one in mmdetection? Can you give me an example of how to quantize yolox to onnx bac…
-
Hello! After reading your paper carefully, the method proposed in the paper is effective for PTQ, and I am very interested in your method. As for the module capacity calculation method mentioned in th…
-
I use the AIMET PTQ to quantize the CLIP text model.
But I encounter this error [KeyError: 'Graph has no buffer /text_model/encoder/layers.0/layer_norm1/Constant_output_0, referred to as input for …
-
### Checklist
- I have searched related issues but cannot get the expected help.
- I have read related documents and don't know what to do.
### Describe the question you meet
Does mmrazor su…
-
Hi,
I am looking at quantization of a SlowFast model. The input, according to netron is 1x1x3x32x256x256. May I know if it is possible to perform PTQ on the model?
Do I just implement the Calibr…
MrOCW updated
2 years ago
-
## Description
I generated calibration cache for Vision Transformer onnx model using EntropyCalibration2 method. When trying to generate engine file using cache file for INT8 precision using trte…
-
### Is there an existing issue for this?
- [X] I have searched the existing issues
### Current Behavior
我是用t4进行推理,它是支持int8和int4的
输入长度为1000,int4需要32s,fp16只需要12s
### Expected Behavior
_No respon…
-
There's something wrong with my verification, could you give me some reference data?
My experiment is as follows:
1, ptq —> 2, qat —> 3, bn reestimation —> 4, fold_all_batch_norms_to_scale
afte…
-
## 📝 Description
While going through your YouTube video explanation on Quantisation. I came across this doubt when I was validating the formulas of `scales` and `zero_point` for Asymmetric and Symmet…
-
Does Minicpmv2.6 currently support int8/fp8 quantization?
thanks~