Open mengfei25 opened 1 month ago
Looks like hf_distil_whisper is regression ./torchbench/amp_bf16/inductor_torchbench_amp_bf16_inference_xpu_performance_all.log:xpu eval hf_distil_whisper running benchmark: 100%|āāāāāāāāāā| 10/10 [00:00<00:00, 10.61it/s]erformance_all.log- ./torchbench/amp_bf16/inductor_torchbench_amp_bf16_inference_xpu_performance_all.log-1.656x pytorch: https://github.com/pytorch/pytorch/commit/dadc0ed torch-xpu-ops: https://github.com/intel/torch-xpu-ops/commit/45e55a3
Looks like hf_distil_whisper is regression ./torchbench/amp_bf16/inductor_torchbench_amp_bf16_inference_xpu_performance_all.log:xpu eval hf_distil_whisper running benchmark: 100%|āāāāāāāāāā| 10/10 [00:00<00:00, 10.61it/s]erformance_all.log- ./torchbench/amp_bf16/inductor_torchbench_amp_bf16_inference_xpu_performance_all.log-1.656x pytorch: pytorch/pytorch@dadc0ed torch-xpu-ops: 45e55a3
Hi @retonym, this is a regression issue, can we double check it?
Recollect the model test on pytorch/pytorch@dadc0ed torch-xpu-ops: 45e55a3 on my local pvc 1100, this issue exists. Besides, this model also fails with out-of-memory on the CUDA backend.
š Describe the bug
Out of memory in weekly test, https://github.com/intel/torch-xpu-ops/actions/runs/10218591763
Model list:
GPTJForCausalLM
GPTJForQuestionAnswering
hf_distil_whisper
hf_T5_base
llava
stable_diffusion_unet
Versions
torch-xpu-ops: https://github.com/intel/torch-xpu-ops/commit/1d70431c072db889d9a47ea4956049fe340a426d pytorch: d224857b3af5c9d5a3c7a48401475c09d90db296 device: pvc 1100, bundle: 0.5.3, driver: 803.61