mit-han-lab / TinyChatEngine

TinyChatEngine: On-Device LLM Inference Library
https://mit-han-lab.github.io/TinyChatEngine/
MIT License
760 stars 73 forks source link

Cleanup and fix bugs for CUDA #41

Closed RaymondWang0 closed 1 year ago