-
i have completed stable diffusion quantization in txt2img as demo shows.
the result is very good.
when i want to transfer sd quantization in inpainting task, i meet the problem that the quantization r…
-
## Bug Description
When doing Post-training quantization using the INT8 calibration API, the model export works fine when using the `ptq.DataLoaderCalibrator` but there is a runtime error when loa…
-
### 🐛 Describe the bug
When trying to compile a model with the QNN partitioner with the GPU or DSP i get the following error:
```
[ERROR] [Qnn ExecuTorch]: Cannot Open QNN library libQnnDsp.so, w…
-
When trying to repeat your code, we find that when inferencing using default fp16, the peak memory goes with:
about 9800MB
But when inferencing using W8A8(after PTQ), the peak memory goes with:
…
-
### Discussed in https://github.com/openvinotoolkit/nncf/discussions/2547
Originally posted by **MinGiSa** March 5, 2024
I've been working on converting Torch models into OpenVINO models rece…
-
## ❓ Question
I'm trying to run the `examples/dynamo/vgg16_fp8_ptq.y` example but got following error:
```
Traceback (most recent call last):
File "/home/wh/generative_action/SynHSI/vgg_quat.p…
-
Hi,
I faced the issue when I tried to run **6.1 normal inference** and **6.2 inference with mixed precision** as your indications. But something was wrong:
**For 6.1 normal inference:**
(viditq…
-
# Quantize the model
model_prepared = tq.prepare(model_fused)
model_quantized = tq.convert(model_prepared)
# Define the quantization configuration
quant_config = tq.get_default_qconfig('fbge…
-
1. X2bolt -d onnx -m model -i PTQ #输出为model_ptq_input.bolt
2. ./post_training_quantization -p model_ptq_input.bolt -i INT8_FP32 -b true -q NOQUANT -c 0 -o false
3. 推理报错如下:
[ERROR] thread 121948 fil…
-
Are the code and parameters in this repository consistent with the parameters used in the experiments described in the paper? I conducted an experiment on an A100 using the provided command "bash 10_o…