issues
search
usyd-fsalab
/
fp6_llm
An efficient GPU support for LLM inference with x-bit quantization (e.g. FP6,FP5).
Apache License 2.0
188
stars
15
forks
source link
Update README.md
#9
Closed
tpoisonooo
closed
5 months ago
tpoisonooo
commented
5 months ago
typo
typo