Closed simkimsia closed 3 months ago
My first option is to run quantized versions.
I read this https://github.com/databricks/dbrx#mlx
and then went to https://huggingface.co/mlx-community/dbrx-instruct-4bit
I read this
On my Macbook Pro M2 with 96GB of Unified Memory, DBRX Instruct in 4-bit for the above prompt it eats 70.2GB of RAM.
I am on a macbook pro M1 Max with 64Gb memory.
I guess that's not enough?
My next version is to figure out what's a cheap way to run the model but the details confuse me.
Can help?
sorry i shifted this to the community discussion on hugging face https://huggingface.co/databricks/dbrx-instruct/discussions/38
My first option is to run quantized versions.
Quantized
I read this https://github.com/databricks/dbrx#mlx
and then went to https://huggingface.co/mlx-community/dbrx-instruct-4bit
I read this
I am on a macbook pro M1 Max with 64Gb memory.
I guess that's not enough?
Computing
My next version is to figure out what's a cheap way to run the model but the details confuse me.
Can help?