codefuse-ai / ModelCache

A LLM semantic caching system aiming to enhance user experience by reducing response time via cached query-result pairs.
Other
780 stars 40 forks source link

Merge pull request #1 from codefuse-ai/modelcache_dev #2

Closed peng3307165 closed 10 months ago

peng3307165 commented 10 months ago

Bert model's ability to handle long texts and address the 512 token limit