dmccloskey / EvoNet

MIT License
2 stars 0 forks source link

MoE example #91

Open dmccloskey opened 5 years ago

dmccloskey commented 5 years ago

References:

oam Shazeer, Azalia Mirhoseini, Krzysztof Maziarz, Andy Davis, Quoc Le, Geoffrey Hinton, and Jeff Dean. Outrageously large neural networks: The sparsely-gated mixture-of-experts layer. arXiv preprint arXiv:1701.06538, 2017.

Objectives:

ModelBuilder::addMoE