-
CUDA add one kernel for performance optimization to support partial channel last case.
https://github.com/pytorch/pytorch/blob/df85f34a14dd30f784418624b05bd52b12ab8b0b/aten/src/ATen/native/cuda/UpSa…
-
I'm not able to run Triton GEMM tutorial example using intel-xpu-backend-for-triton. Below are the local changes done to "triton/python/tutorials/03-matrix-multiplication.py" to use XPU. Please scroll…
-
HW: MTL with ARC iGPU
OS: Ubuntu 22.04
Kernel: 6.5.0-41-generic
Ref: https://github.com/intel-analytics/ipex-llm/blob/main/docs/mddocs/Quickstart/install_linux_gpu.md
Problem1: cannot find GPU dr…
-
`torch.utils.data.DataLoader(pin_memory_device='xpu')` is currently not supported with upstream PyTorch XPU backend. I know that with IPEX this feature was supported. Please, support the feature if it…
-
The latest Triton refactoring removed the Intel Triton backend from the third-party, `llvm-target` branch is a fork of `openai/Triton` with in-tree modifications.
To upstream Intel XPU Triton backend…
-
### 🐛 Describe the bug
A case fail due to oneDNN matmul implementation,
Skip the case temporarily.
### Versions
Latest pytorch.
-
### Describe the bug
After building docker image following [[RECOMMENDED] Docker-based environment setup with pre-built wheels](https://github.com/intel/intel-extension-for-pytorch/tree/v2.3.0%2Bcpu/…
-
I'm creating POC for testing microservices architecture in GPU system provided by opea.
[Build Mega Service of ChatQnA on Xeon](https://github.com/opea-project/GenAIExamples/blob/main/ChatQnA/docker/…
-
Using ipex-llm docker version for inferencing, but during inference time it experiences errors from util files
below is the log:
```
------------------------------------------------------------…
-
### Describe the bug
Repeated calls into `torch.dist.reduce_scatter_tensor` eventually raise a
`ZE_RESULT_ERROR_OUT_OF_DEVICE_MEMORY` error in multi-node setups. Similar behavior is found when
usin…