Open ArlanCooper opened 2 months ago
i see now awq only support 4-bit quantization, can it supports 2-bit,3-bit, 8-bit quantization?
Efficient and accurate low-bit weight quantization (INT3/4) for LLMs, supporting instruction-tuned models and multi-modal LMs.
:thinking:
i see now awq only support 4-bit quantization, can it supports 2-bit,3-bit, 8-bit quantization?