axolotl-ai-cloud / axolotl

Go ahead and axolotl questions
https://axolotl-ai-cloud.github.io/axolotl/
Apache License 2.0
7.84k stars 863 forks source link

LongLora suport #623

Open generalsvr opened 1 year ago

generalsvr commented 1 year ago

⚠️ Please check that this feature request hasn't been suggested before.

🔖 Feature description

Can you implement this new LoRA method? That would be great to have 32k+ LoRA models. Looks promising.

✔️ Solution

https://github.com/dvlab-research/LongLoRA

http://arxiv.org/abs/2309.12307

❓ Alternatives

No response

📝 Additional Context

No response

Acknowledgements

NanoCode012 commented 1 year ago

Seems like they provided a patch for llama in their repo.

Parts I've noticed:

winglian commented 1 year ago

May want to keep track of https://github.com/huggingface/peft/issues/958 in case it is supported there.

winglian commented 1 year ago

looking at the shift/unshift code, it seems it's not packed sequence length aware, so that would need some modification (or simply not allow packed sequences to work w this features)

DhruvaBansal00 commented 8 months ago

Is this something that is on the roadmap?