-
### 🐛 Describe the bug
torchbench_amp_bf16_training
- [ ] `detectron2_fasterrcnn_r_101_c4`
- [ ] `detectron2_fasterrcnn_r_101_dc5`
- [ ] `detectron2_fasterrcnn_r_101_fpn`
- [ ] `detectron2_fast…
-
I am trying to implement a concurrent asynchronous send and receive between multiple processes. This results in deadlock. Minimum code to reproduce this is as follows:
```
import torch.nn.parallel…
-
New 2D load variants are used in FP8 load. https://github.com/intel/intel-xpu-backend-for-triton/commit/5f088429c8aeb0082c262d812945c2b4e7b86233
- 8b, 8 row, 16 column, 4 blocks, 2D load. GenISA si…
-
### 🚀 The feature, motivation and pitch
The issue is introduced in codegen pr https://github.com/intel/torch-xpu-ops/pull/310.
The FAILED UT throw errors like
```bash
RuntimeError: scatter_…
-
### Describe the issue
during import intel_extension_for_pytorch as ipex
7 │
│ ❱ 8 import intel_ex…
-
### Is there an existing issue for this?
- [X] I have searched the existing issues and checked the recent builds/commits
### What would your feature do ?
1. torch: 2.2.0+cu121
2. more perfomance
…
-
### Checklist
- [X] The issue exists after disabling all extensions
- [X] The issue exists on a clean installation of webui
- [ ] The issue is caused by an extension, but I believe it is caused by a …
-
### Describe the issue
Following this guide(https://intel.github.io/intel-extension-for-pytorch/xpu/2.0.110+xpu/tutorials/installations/windows.html), I successfully install ipex-xpu on native Wind…
-
### Describe the bug
I found in jupyter notebook, `to('xpu')` makes the Jupyter kernel die.
### Notebook to reproduce
![image](https://github.com/intel/intel-extension-for-pytorch/assets/105281…
-
When I run GEMM benchmark on A770 I get about ~`0.3 TFLOPs`, while 1550 can get about `250 TFLOPs`
Performance table:
![image](https://github.com/user-attachments/assets/366947f8-82ce-4454-83ae-f…