issues
search
turboderp
/
exllamav2
A fast inference library for running LLMs locally on modern consumer-class GPUs
MIT License
3.2k
stars
235
forks
source link
quanting community
#365
Closed
Kerushii
closed
3 months ago