-
@fengyuan14 - The commit https://github.com/intel/torch-xpu-ops/commit/5bf9e0cc768f7a3b13d829118683275f324399f1 muted debug logs of "explicit" CPU fallbacks. This complicated debug for 3d party contri…
-
```
loading model: 0it [00:03, ?it/s]
xpu train AllenaiLongformerBase
File "/localdisk/abaden/Projects/intel-xpu-backend-for-triton/python/triton/compiler/compiler.py", line 376, in _init_…
-
Some benchmarks are failing when subgroup size is set to 16:
- `huggingface YituTechConvBert amp_bf16 training`
- `huggingface YituTechConvBert bf16 training`
Found in: https://github.com/intel…
-
Hi @echarlaix . I want to enable all model utils in ipex ([modeling_utils](https://github.com/huggingface/optimum-intel/blob/main/optimum/exporters/ipex/modeling_utils.py)) on XPU; it may need some ch…
-
grid_sample is currently not implemented for XPU backend, see
https://github.com/intel/torch-xpu-ops/blob/8cc6d5102878cefd3c71245fc7509a4548d07da2/src/aten/XPUFallback.template#L249
The problem is…
-
### Describe the issue
python -m pip install torch==2.0.0a0 intel_extension_for_pytorch==2.0.110+gitba7f6c1 -f https://developer.intel.com/ipex-whl-stable-xpu
wheel files seem to include torch 2.0…
-
We should be able to support intel GPUs! We are using the intel developer cloud. Please advise.
Distributor ID: Ubuntu
Description: Ubuntu 22.04.4 LTS
Release: 22.04
Codename: …
-
After upgrade triton commit pin from `b8c64f64c18d8cac598b3adb355c21e7439c21de`( currently stock Pytorch in used) to `514e4cdf004278c82216364d1f8534b940cd4238` (2.4 release candidate),
We found the …
-
All I need is to run ollama3 on an Intel GPU (Arc™ A750) and I follow the steps as described in the IPEX-LLM documentation, but it runs on the CPU. Search engines can't find a solution to the problem.…
-
As @mfrancepillois said in https://github.com/intel/intel-xpu-backend-for-triton/pull/1395#pullrequestreview-2128572018, this pass in itself won't lead to better performance. Check the pipeline is cap…