SOTA low-bit LLM quantization (INT8/FP8/INT4/FP4/NF4) & sparsity; leading model compression techniques on TensorFlow, PyTorch, and ONNX Runtime
2.23k
stars
257
forks
source link
fix welcome.html link issue #1962
Closed
NeoZhangJianyu closed 3 months ago
Type of Change
documentation
Description
fix the links issues in welcome.html
Expected Behavior & Potential Risk
the expected behavior that triggered by this PR
How has this PR been tested?
how to reproduce the test (including hardware information)
Dependency Change?
any library dependency introduced or removed