mlc-ai / mlc-llm

Universal LLM Deployment Engine with ML Compilation
https://llm.mlc.ai/
Apache License 2.0
19.23k stars 1.58k forks source link

[Model] Keep vision encoder weights unquantized to maintain accuracy #3028

Closed mengshyu closed 6 days ago

mengshyu commented 1 week ago

This PR ensures that vision encoder layers are excluded from quantization, improving accuracy for models with vision components.