openvinotoolkit / openvino

OpenVINO™ is an open-source toolkit for optimizing and deploying AI inference
https://docs.openvino.ai
Apache License 2.0
7.36k stars 2.31k forks source link

Looking forward to providing OpenVINO backend support for Llama.cpp! #27736

Open Torinlq opened 4 days ago

Torinlq commented 4 days ago

Request Description

Llama.cpp is a very popular and excellent LLM/VLM inference deployment framework, implemented in pure C/C++, without any dependencies, and cross-platform. Based on SYCL and Vulkan, it can support some Intel integrated graphics cards for inference acceleration, but there are many compatibility issues and it does not support NPU at all. Can Intel provide OpenVINO backend support for this project?

Feature Use Case

No response

Issue submission checklist

rkazants commented 3 days ago

@ynimmaga, please take a look as well

Regards, Roman