Open luyuntao92 opened 8 months ago
如果有更新,建议再follow一下这篇工作:https://arxiv.org/abs/2402.08562《Higher Layers Need More LoRA Experts》 允许自定义设置每一层的专家个数
mark:)
mark
mark
mark
mark
mark
請問是否有更新,增加該方法
十分期待在LLaMA-Factory的基础上,增加LoRA-Moe的设置,这将增加了LLaMA的适用性。
十分期待能够增加这个训练
Reminder
Reproduction
https://arxiv.org/pdf/2312.09979.pdf 《LoRAMoE: Alleviate World Knowledge Forgetting in Large Language Models via MoE-Style Plugin》
Expected behavior
看论文,效果好像挺不错的
System Info
No response
Others
No response