MR-GREEN1337 / Mistral-7b-PyTorch

Implementation of Mistral 7b using PyTorch
Apache License 2.0
1 stars 0 forks source link

Shouldn't we use MAMBA? #1

Closed Mr-Harry closed 2 weeks ago

MR-GREEN1337 commented 1 month ago

The trend in transformer-based LLMs is shifting towards smaller models with a greater emphasis on the quality of training data. This results in high-performing models with fewer parameters. Additionally, various techniques have been developed to efficiently run these models on CPUs. Moreover, significant investments have been made in developing these types of models. So I wouldn't see them being replaced any time soon.