Open KevinRyu opened 11 months ago
Hey, it probably doesn't make sense to run this on Windows as you'll need a GPU (which I assume you don't have locally). It's probably best to use some cloud GPU service or just run on Google Colab. You can find a demo Mamba-Chat on Google Colab here
Hey, it probably doesn't make sense to run this on Windows as you'll need a GPU (which I assume you don't have locally). It's probably best to use some cloud GPU service or just run on Google Colab. You can find a demo Mamba-Chat on Google Colab here
Does Mamba Chat only run on a GPU?
@SzaremehrjardiMT Currently yes. There's an open issue in llama.cpp to support the mamba architecture, though, which would make it possible to run without a GPU: https://github.com/ggerganov/llama.cpp/issues/4353
Hey, it probably doesn't make sense to run this on Windows as you'll need a GPU (which I assume you don't have locally). It's probably best to use some cloud GPU service or just run on Google Colab. You can find a demo Mamba-Chat on Google Colab here
@KevinRyu It won't be optimized, but you can try mamba-minimal
Hello,
When I tried to install packages with
requirements.txt
, I got the following error.As I know,
triton
package supports something like linux only. What should I do?