mit-han-lab / llm-awq

[MLSys 2024 Best Paper Award] AWQ: Activation-aware Weight Quantization for LLM Compression and Acceleration
MIT License
2.38k stars 184 forks source link

Some question about example #136

Open YIHUASHAO opened 8 months ago

YIHUASHAO commented 8 months ago

When using transformers 4.32.0,here is a question (https://github.com/mit-han-lab/llm-awq/assets/83972796/2294b59d-b211-4a3b-9fa6-3d211bd2ca34)