AIoT-MLSys-Lab / Efficient-LLMs-Survey

[TMLR 2024] Efficient Large Language Models: A Survey
https://arxiv.org/abs/2312.03863
970 stars 82 forks source link

A NeurIPS paper on efficient architecture #3

Closed renll closed 9 months ago

renll commented 9 months ago

Thanks for the great survey! Could you please include a discussion of this work from Microsoft and UIUC? It proposes a general modular activation mechanism, SMA, that unifies previous works on MoE, adaptive computation, dynamic routing and sparse attention, and further applies SMA to develop a novel architecture, SeqBoat, to achieve SoTA quality-efficiency trade-off on Long Range Arena.

SUSTechBruce commented 9 months ago

Thanks for the suggestion and congrats on your NeurIPS paper! We've added it to the github paperlist and will update it to survey in next version~