issues
search
kyegomez
/
MoE-Mamba
Implementation of MoE Mamba from the paper: "MoE-Mamba: Efficient Selective State Space Models with Mixture of Experts" in Pytorch and Zeta
https://discord.gg/GYbXvDGevY
MIT License
84
stars
5
forks
source link
issues
Newest
Newest
Most commented
Recently updated
Oldest
Least commented
Least recently updated
Bump pypa/gh-action-pypi-publish from 1.8.11 to 1.12.2
#14
dependabot[bot]
opened
6 days ago
0
Bump pypa/gh-action-pypi-publish from 1.8.11 to 1.11.0
#13
dependabot[bot]
closed
6 days ago
1
Hello, How much GPU using when train Mamba-MoE?
#12
zhipengY
opened
3 weeks ago
1
Bump pypa/gh-action-pypi-publish from 1.8.11 to 1.10.3
#11
dependabot[bot]
closed
1 week ago
1
Bump pypa/gh-action-pypi-publish from 1.8.11 to 1.10.2
#10
dependabot[bot]
closed
1 month ago
1
Torch Version
#9
YongLD
opened
1 month ago
2
Bump pypa/gh-action-pypi-publish from 1.8.11 to 1.10.1
#8
dependabot[bot]
closed
1 month ago
1
Bump pypa/gh-action-pypi-publish from 1.8.11 to 1.10.0
#7
dependabot[bot]
closed
2 months ago
1
Is Class SwitchMixtureOfExperts unused in main model?
#6
lunaaa95
closed
1 month ago
2
Bump pypa/gh-action-pypi-publish from 1.8.11 to 1.9.0
#5
dependabot[bot]
closed
2 months ago
2
[BUG] I tried to run example.py as is but it fails
#4
arelkeselbri
closed
7 months ago
2
Bump pypa/gh-action-pypi-publish from 1.8.11 to 1.8.14
#3
dependabot[bot]
closed
6 months ago
2
Bump pypa/gh-action-pypi-publish from 1.8.11 to 1.8.12
#2
dependabot[bot]
closed
8 months ago
1
Bump actions/checkout from 2 to 4
#1
dependabot[bot]
closed
10 months ago
0