-
Hi,
I am interested in the NPU inference for this project.
I tried to run llama on NPU with python\llm\example\NPU\HF-Transformers-AutoModels\Model\llama2\generate.py.
I used interface `model.save_…
-
I have tested the inference speed and memory usage of Qwen1.5-14b on my machine using the example in ipex-llm. The peek cpu usage to load Qwen1.5-14b in 4-bit is about 24GB. The peek GPU usage is abou…
-
On NVIDIA GPUs, there is a relation between `nvidia-smi` and PyTorch, `nvidia-smi`, which is similar to `xpu-smi` is used to detect and monitor GPU telemetry. However, absence of `nvidia-smi` on the …
-
### Your current environment
Collecting environment information...
/home/sdp/miniforge3/envs/liangan1/lib/python3.10/site-packages/transformers/utils/hub.py:128: FutureWarning: Using `TRANSFORMERS_C…
-
`ror occurred when executing IPAdapter:
Could not run 'aten::_upsample_bicubic2d_aa.out' with arguments from the 'XPU' backend. This could be because the operator doesn't exist for this backend, or…
-
### 🐛 Describe the bug
# Cases of XPU supported OP
- [x] adaptive_max_pool2d: Expected out tensor to have dtype c10::BFloat16/c10::Half/float/double, but got long int instead
"test_dispatch_meta_o…
-
### Describe the bug
Wuerstchen and Wuerstchen based Stable Cascade models generates corrupted images.
- Happens with any resolution.
- Happens with both Intel ARC and Intel Datacenter GPU Ma…
-
### Describe the bug
When following README.md to install the GPU version of IPEX:
```
python -m pip install torch==2.0.1a0 torchvision==0.15.2a0 intel_extension_for_pytorch==2.0.110+xpu -f https:…
-
when I use Pyinstaller to package and run the Python demo code, The .exe process exit when
`model = model.to('xpu')`
There is no problem running the demo directly using Python.
anyone know …
-
### 🐛 Describe the bug
torchbench_amp_bf16_training
xpu train opacus_cifar10
Traceback (most recent call last):
File "/home/sdp/actions-runner/_work/torch-xpu-ops/pytorch/benc…