intel-analytics / ipex-llm

Accelerate local LLM inference and finetuning (LLaMA, Mistral, ChatGLM, Qwen, Baichuan, Mixtral, Gemma, Phi, MiniCPM, etc.) on Intel XPU (e.g., local PC with iGPU and NPU, discrete GPU such as Arc, Flex and Max); seamlessly integrate with llama.cpp, Ollama, HuggingFace, LangChain, LlamaIndex, GraphRAG, DeepSpeed, vLLM, FastChat, Axolotl, etc.
Apache License 2.0
6.55k stars 1.25k forks source link

gemma 2 support #11456

Open manfye opened 3 months ago

manfye commented 3 months ago

hi, any update on gemma 2 support and speed up?

https://blog.google/technology/developers/google-gemma-2/

qiuxin2012 commented 3 months ago

We will support gemma 2 in the future.