Official inference framework for 1-bit LLMs
11.39k
stars
768
forks
source link
where can i have a converted i2_s.gguf model, to convert it to gguf needs quite a lot of GPU RAM,which makes it not a good choice to run Locally. #95
Closed
ssdutliuhaibo closed 3 weeks ago