issues
search
usyd-fsalab
/
fp6_llm
An efficient GPU support for LLM inference with x-bit quantization (e.g. FP6,FP5).
Apache License 2.0
172
stars
14
forks
source link
Refine README.
#1
Closed
JamesTheZ
closed
6 months ago