intel-analytics / ipex-llm

Accelerate local LLM inference and finetuning (LLaMA, Mistral, ChatGLM, Qwen, Baichuan, Mixtral, Gemma, Phi, etc.) on Intel CPU and GPU (e.g., local PC with iGPU, discrete GPU such as Arc, Flex and Max); seamlessly integrate with llama.cpp, Ollama, HuggingFace, LangChain, LlamaIndex, DeepSpeed, vLLM, FastChat, Axolotl, etc.
Apache License 2.0
6.31k stars 1.23k forks source link

Feature Request: RoSA and QRoSA #10755

Open ElliottDyson opened 3 months ago

ElliottDyson commented 3 months ago

It would be brilliant if we could get implemented fine-tuning methods for robust adaptation given how much better it is than LoRA and QLoRA methods.

Uxito-Ada commented 3 months ago

Hi @ElliottDyson , thanks for your proposal.

Currently we provide many fine-tuning options e.g. ReLoRA, axolotl and DPO etc. as shown here, as well as Galore and LISA on way, where some can outperform LoRA.

We are going to investigate and evaluate whether to support RoSA and QRoSA.