open-mmlab / Multimodal-GPT

Multimodal-GPT
Apache License 2.0
1.48k stars 126 forks source link

Will this work with quantized versions of llama models? #9

Open thistleknot opened 1 year ago

thistleknot commented 1 year ago

What about alpaca?

For example. I have a smaller version of llama that runs locally

python server.py --model ggml-alpaca-7b-q4 --listen

I'm going to give it a shot/go with this reduced model and get back to you over the weekend