issues
search
allenai
/
OLMoE
OLMoE: Open Mixture-of-Experts Language Models
https://arxiv.org/abs/2409.02060
Apache License 2.0
468
stars
37
forks
source link
issues
Newest
Newest
Most commented
Recently updated
Oldest
Least commented
Least recently updated
Update README.md
#20
jacob-morrison
closed
6 days ago
0
Can't open link in README Adaptation “Clone this open-instruct branch & follow its setup instructions.”
#19
He-JYang
closed
6 days ago
2
Update README.md
#18
schmmd
closed
3 weeks ago
0
Add llama.cpp usage instructions to README
#17
2015aroras
closed
1 month ago
0
Config Naming
#16
mchorton
closed
1 month ago
1
Early loss divergence for upcycling
#15
yazdayy
opened
1 month ago
5
Dropout Regularization in expert modules
#14
taehyunzzz
closed
1 month ago
3
Supported generative tasks
#13
taehyunzzz
closed
1 month ago
1
recommended conf
#12
raingart
closed
1 month ago
1
How to get the MMLU results in Table 4?
#11
mathfinder
opened
2 months ago
1
Tokenized dataset?
#10
joelburget
opened
2 months ago
1
Implementing MoE Sparse Upcycling
#9
adumans
opened
2 months ago
13
MOE Export Parallelism Training Script
#8
wdlctc
closed
2 months ago
5
llama.cpp / GGUF support
#7
sammcj
opened
2 months ago
7
Davidw/dclm eval
#6
dwadden
closed
3 months ago
0
routing_analysis
#5
swj0419
closed
3 months ago
1
Increase plot font size; remove spines; add mixtral; change colors; a…
#4
Muennighoff
closed
2 months ago
0
add scripts for MoE analysis
#3
shmsw25
closed
3 months ago
0
Davidw/dclm eval
#2
dwadden
closed
3 months ago
0
Davidw/dclm eval
#1
dwadden
closed
3 months ago
0