issues
search
kvcache-ai
/
ktransformers
A Flexible Framework for Experiencing Cutting-edge LLM Inference Optimizations
Apache License 2.0
741
stars
39
forks
source link
Release v0.1.2
#36
Closed
UnicornChan
closed
3 months ago
UnicornChan
commented
3 months ago
Support windows native.
Support multiple GPU
Support qlen > 1.
Support new model: mixtral 8
7B and 8
22B
Support q2k, q3k dequant on gpu.
Support github action to create pre compile package
Fix some bugs