-
Validating and improving how profiling is done. Internal [Ipex profiling example](https://git.cels.anl.gov/blusch/ipex_profiling) works in isolation, but with this Megatron-DeepSpeed we are missing th…
-
### 🐛 Describe the bug
torchbench_amp_bf16_training
xpu train torchrec_dlrm
ERROR:common:
Traceback (most recent call last):
File "/home/sdp/actions-runner/_work/torch-xpu-op…
-
### Describe the bug
Intel compute runtime doesn't allow allocating a buffer bigger than 4 GB.
https://github.com/intel/compute-runtime/issues/627
When you allocate an array in intel-extensio…
-
基于intelanalytics/ipex-llm-xpu:latest镜像制作的镜像chatdoc_intel:1.1
启动容器test,无法加载Embedder模型。
dockerfile:
FROM intelanalytics/ipex-llm-xpu:latest
RUN export DEBIAN_FRONTEND=noninteractive && \
se…
-
-
Use "--extra-index-url https://pytorch-extension.intel.com/release-whl/stable/xpu/us/ "
instead of "-f https://developer.intel.com/ipex-whl-stable-xpu"
in the instruction guide for bigdl-llm install…
-
2024-07-03 22:05:41.445430: I external/local_tsl/tsl/cuda/cudart_stub.cc:31] Could not find cuda drivers on your machine, GPU will not be used.
2024-07-03 22:05:41.446433: I tensorflow/core/platform/…
-
There is no stand along profiler tools for Triton XPU now.
We used to use:
1. the Torch legacy profiler with the IPEX extension. (This is going to be removed by IPEX)
2. The new torch profiler wi…
-
When I run `python -m fastchat.serve.model_worker --model-path lmsys/vicuna-7b-v1.5 --device xpu --gpus X`, whatever the X is, or without `--gpus`, the memory will be always loaded to both CPU and int…
-
# Motivation
This RFC aims to propose a design for a series of generic memory-related APIs tailored for stream-based accelerators to help users simplify the runtime code written for different devices…