codefuse-ai / ModelCache

A LLM semantic caching system aiming to enhance user experience by reducing response time via cached query-result pairs.
Other
874 stars 43 forks source link

Modelcache for mm #27

Closed peng3307165 closed 6 months ago

peng3307165 commented 6 months ago

Added caching capabilities for multi-modal scenarios.