A curated list of exceptional papers and resources on Mixture of Experts and related topics.
News: Our Mixture of Experts survey has been released. The Evolution of Mixture of Experts: A Survey from Basics to Breakthroughs
Mendeley | ResearchGate | PDF If our work has been of assistance to you, please feel free to cite our survey. Thank you.
@article{article,
author = {Vats, Arpita and Raja, Rahul and Jain, Vinija and Chadha, Aman},
year = {2024},
month = {08},
pages = {12},
title = {THE EVOLUTION OF MIXTURE OF EXPERTS: A SURVEY FROM BASICS TO BREAKTHROUGHS}
}
Name | Paper | Venue | Year |
---|---|---|---|
The Sparsely-Gated Mixture-of-Experts Layer | Outrageously Large Neural Networks: The Sparsely-Gated Mixture-of-Experts Layer | arXiv | 2017 |
Name | Paper | Venue | Year |
---|---|---|---|
MoME | MoME: Mixture of Multimodal Experts for Generalist Multimodal Large Language Models | arXiv | 2024 |
CAME | CAME: Competitively Learning a Mixture-of-Experts Model for First-stage Retrieval | ACM | 2024 |
SummaReranker | SummaReranker: A Multi-Task Mixture-of-Experts Re-ranking Framework for Abstractive Summarization | arXiv | 2022 |
MDFEND | MDFEND: Multi-domain Fake News Detection | arXiv | 2022 |
PLE | PLE outperforming state-of-the-art MTL models | RecSys | 2021 |
Name | Paper | Venue | Year |
---|---|---|---|
MoE-Infinity | MoE-Infinity: Offloading-Efficient MoE Model Serving | arXiv | 2024 |
SMT 2.0 | SMT 2.0: A Surrogate Modeling Toolbox with a focus on Hierarchical and Mixed Variables Gaussian Processes | arXiv | 2023 |
Hope our survey with collection of all the recent MoE can help your work.