Open xiaowangintel opened 8 months ago
Please add to the PR description 1) how to build/install the pre-requisite software components; 2) how to run inference with and without tensor parallel.
@Chillee This is the initial PR to support Intel GPU. Most needed code changes should be there. Further performance optimizations will be applied inside IPEX. May I ask your review? Thanks!
This PR adds the initial Intel GPU support in GPT-fast with the device option "xpu" (i.e., --device "xpu"). Both single device and multi-device via tensor parallel are supported functionally while performance is still being improved. Refer to the following steps to run the generation on Intel GPU. We will update the tutorial later with improved performance.
Installation
How to run gpt-fast code on intel GPUs?
Note: