Open azhuvath opened 1 month ago
@azhuvath could you clarify:
optimum-cli export openvino -m stabilityai/stable-diffusion-xl-base-1.0 --weight-format int8 {model_dir}
as implemented in the openvino-notebook, to get the INT8 format?@azhuvath could you clarify:
- You mentioned other models, is it the same model with different precision or they are completely different model?
- Did you use this command
optimum-cli export openvino -m stabilityai/stable-diffusion-xl-base-1.0 --weight-format int8 {model_dir}
as implemented in the openvino-notebook, to get the INT8 format?- Did you convert and run the model using OV 2024? (this is to confirm whether you use converted model files from older version into newer)
Quantized the model with Optimum CLI and used OV 2024.4 version which is the latest. Will test it once more after getting access to the system.
OpenVINO Version
2024.4
Operating System
Windows System
Device used for inference
GPU
Framework
None
Model used
stabilityai/stable-diffusion-xl-base-1.0
Issue description
Trying to follow the below notebook on a LNL Windows machine targeting iGPU. https://github.com/openvinotoolkit/openvino_notebooks/blob/latest/notebooks/stable-diffusion-xl/stable-diffusion-xl.ipynb
Getting the following error. Other models works on the GPU and hence the issue is not with GPU or the drivers. This is a INT8 quantized model.
Step-by-step reproduction
Relevant log output
No response
Issue submission checklist