SOTA low-bit LLM quantization (INT8/FP8/INT4/FP4/NF4) & sparsity; leading model compression techniques on TensorFlow, PyTorch, and ONNX Runtime
2.18k
stars
252
forks
source link
Set `low_gpu_mem_usage=False` for AutoRound #1960
Closed
Kaihui-intel closed 2 months ago
Type of Change
feature or bug fix or documentation or validation or others
API changed or not
Description
detail description
Expected Behavior & Potential Risk
the expected behavior that triggered by this PR
How has this PR been tested?
how to reproduce the test (including hardware information)
Dependency Change?
any library dependency introduced or removed