codelion / optillm

Optimizing inference proxy for LLMs
Apache License 2.0
1.6k stars 128 forks source link

Update optillm.py #94

Closed codelion closed 1 week ago